Skip to content

Q: model support v0.9.3 #1202

@kalle07

Description

@kalle07

Jan
https://huggingface.co/janhq/Jan-v1-2509
works fine with no error export as gguf
(llama-version 11/25 - last month)

next two models i only changed the model path...

Mistral-Magistral-Devstral is supported?
https://huggingface.co/DavidAU/Mistral-Magistral-Devstral-Instruct-FUSED-CODER-Reasoning-36B
...
2025-12-27 16:20:38 WARNING modeling_utils.py L4670: torch_dtype is deprecated! Use dtype instead!
Loading checkpoint shards: 100%|██████████| 15/15 [00:00<00:00, 183.83it/s]
2025-12-27 16:20:39 INFO base.py L390: using torch.bfloat16 for quantization tuning
2025-12-27 16:20:39 WARNING base.py L419: using algorithm extension for quantization.
Traceback (most recent call last):
File "C:\Users\xxx\Documents\python\autoround\auto_next.py", line 66, in
autoround.quantize_and_save(output_dir, format='gguf:q3_k_s')
File "auto_round/alg_ext.py", line 140, in alg_ext.wrapper_func
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 1055, in quantize_and_save
format_list = self._parse_format_to_list(format)
File "auto_round/alg_ext.py", line 140, in alg_ext.wrapper_func
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 835, in _parse_format_to_list
gguf_args_check(self, formats, model_type=ModelType.TEXT)
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\utils.py", line 535, in gguf_args_check
from auto_round.export.export_to_gguf.convert_hf_to_gguf import ( # pylint: disable=E0401
ImportError: cannot import name 'ModelBase' from 'auto_round.export.export_to_gguf.convert_hf_to_gguf' (C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\export\export_to_gguf\convert_hf_to_gguf.py)

Granite small is supported?
https://huggingface.co/ibm-granite/granite-4.0-h-small
...
2025-12-27 16:32:40 INFO main.py L524: start to quantize e:\granite_small
2025-12-27 16:32:40 WARNING modeling_utils.py L4670: torch_dtype is deprecated! Use dtype instead!
2025-12-27 16:32:40 WARNING modeling_granitemoehybrid.py L460: The fast path is not available because one of (selective_state_update, causal_conv1d_fn, causal_conv1d_update) is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation and https://github.com/Dao-AILab/causal-conv1d

Loading checkpoint shards: 0%| | 0/14 [00:00<?, ?it/s]
Loading checkpoint shards: 100%|██████████| 14/14 [00:00<00:00, 172.62it/s]
2025-12-27 16:35:34 INFO base.py L390: using torch.float16 for quantization tuning
Traceback (most recent call last):
File "", line 198, in run_module_as_main
File "", line 88, in run_code
File "c:\Users\xxx\Documents\python\autoround\venv\Scripts\auto_round_best.exe_main
.py", line 6, in
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round_main
.py", line 894, in run_best
tune(args)
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round_main_.py", line 667, in tune
model, folders = autoround.quantize_and_save(export_dir, format=args.format) # pylint: disable=E1101
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 1055, in quantize_and_save
format_list = self._parse_format_to_list(format)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 835, in _parse_format_to_list
gguf_args_check(self, formats, model_type=ModelType.TEXT)
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\utils.py", line 535, in gguf_args_check
from auto_round.export.export_to_gguf.convert_hf_to_gguf import ( # pylint: disable=E0401
ImportError: cannot import name 'ModelBase' from 'auto_round.export.export_to_gguf.convert_hf_to_gguf' (C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\export\export_to_gguf\convert_hf_to_gguf.py)

i can not give you more logging ... or you give me a python code for your needs

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions