-
Notifications
You must be signed in to change notification settings - Fork 66
Description
Jan
https://huggingface.co/janhq/Jan-v1-2509
works fine with no error export as gguf
(llama-version 11/25 - last month)
next two models i only changed the model path...
Mistral-Magistral-Devstral is supported?
https://huggingface.co/DavidAU/Mistral-Magistral-Devstral-Instruct-FUSED-CODER-Reasoning-36B
...
2025-12-27 16:20:38 WARNING modeling_utils.py L4670: torch_dtype is deprecated! Use dtype instead!
Loading checkpoint shards: 100%|██████████| 15/15 [00:00<00:00, 183.83it/s]
2025-12-27 16:20:39 INFO base.py L390: using torch.bfloat16 for quantization tuning
2025-12-27 16:20:39 WARNING base.py L419: using algorithm extension for quantization.
Traceback (most recent call last):
File "C:\Users\xxx\Documents\python\autoround\auto_next.py", line 66, in
autoround.quantize_and_save(output_dir, format='gguf:q3_k_s')
File "auto_round/alg_ext.py", line 140, in alg_ext.wrapper_func
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 1055, in quantize_and_save
format_list = self._parse_format_to_list(format)
File "auto_round/alg_ext.py", line 140, in alg_ext.wrapper_func
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 835, in _parse_format_to_list
gguf_args_check(self, formats, model_type=ModelType.TEXT)
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\utils.py", line 535, in gguf_args_check
from auto_round.export.export_to_gguf.convert_hf_to_gguf import ( # pylint: disable=E0401
ImportError: cannot import name 'ModelBase' from 'auto_round.export.export_to_gguf.convert_hf_to_gguf' (C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\export\export_to_gguf\convert_hf_to_gguf.py)
Granite small is supported?
https://huggingface.co/ibm-granite/granite-4.0-h-small
...
2025-12-27 16:32:40 INFO main.py L524: start to quantize e:\granite_small
2025-12-27 16:32:40 WARNING modeling_utils.py L4670: torch_dtype is deprecated! Use dtype instead!
2025-12-27 16:32:40 WARNING modeling_granitemoehybrid.py L460: The fast path is not available because one of (selective_state_update, causal_conv1d_fn, causal_conv1d_update) is None. Falling back to the naive implementation. To install follow https://github.com/state-spaces/mamba/#installation and https://github.com/Dao-AILab/causal-conv1d
Loading checkpoint shards: 0%| | 0/14 [00:00<?, ?it/s]
Loading checkpoint shards: 100%|██████████| 14/14 [00:00<00:00, 172.62it/s]
2025-12-27 16:35:34 INFO base.py L390: using torch.float16 for quantization tuning
Traceback (most recent call last):
File "", line 198, in run_module_as_main
File "", line 88, in run_code
File "c:\Users\xxx\Documents\python\autoround\venv\Scripts\auto_round_best.exe_main.py", line 6, in
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round_main.py", line 894, in run_best
tune(args)
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round_main_.py", line 667, in tune
model, folders = autoround.quantize_and_save(export_dir, format=args.format) # pylint: disable=E1101
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 1055, in quantize_and_save
format_list = self._parse_format_to_list(format)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\base.py", line 835, in _parse_format_to_list
gguf_args_check(self, formats, model_type=ModelType.TEXT)
File "C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\compressors\utils.py", line 535, in gguf_args_check
from auto_round.export.export_to_gguf.convert_hf_to_gguf import ( # pylint: disable=E0401
ImportError: cannot import name 'ModelBase' from 'auto_round.export.export_to_gguf.convert_hf_to_gguf' (C:\Users\xxx\Documents\python\autoround\venv\Lib\site-packages\auto_round\export\export_to_gguf\convert_hf_to_gguf.py)
i can not give you more logging ... or you give me a python code for your needs