ariG23498 HF Staff commited on
Commit
1ffd517
·
verified ·
1 Parent(s): 9599134

Upload cyankiwi_GLM-4.7-Flash-AWQ-4bit_0.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. cyankiwi_GLM-4.7-Flash-AWQ-4bit_0.txt +43 -20
cyankiwi_GLM-4.7-Flash-AWQ-4bit_0.txt CHANGED
@@ -11,26 +11,49 @@ pipe(messages)
11
 
12
  ERROR:
13
  Traceback (most recent call last):
14
- File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1360, in from_pretrained
15
- config_class = CONFIG_MAPPING[config_dict["model_type"]]
16
- ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^
17
- File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1048, in __getitem__
18
- raise KeyError(key)
19
- KeyError: 'glm4_moe_lite'
20
-
21
- During handling of the above exception, another exception occurred:
22
-
23
- Traceback (most recent call last):
24
- File "/tmp/cyankiwi_GLM-4.7-Flash-AWQ-4bit_0gpVaou.py", line 26, in <module>
25
  pipe = pipeline("text-generation", model="cyankiwi/GLM-4.7-Flash-AWQ-4bit")
26
- File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 922, in pipeline
27
- config = AutoConfig.from_pretrained(
28
- model, _from_pipeline=task, code_revision=code_revision, **hub_kwargs, **model_kwargs
 
 
29
  )
30
- File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/models/auto/configuration_auto.py", line 1362, in from_pretrained
31
- raise ValueError(
32
- ...<8 lines>...
 
 
 
 
33
  )
34
- ValueError: The checkpoint you are trying to load has model type `glm4_moe_lite` but Transformers does not recognize this architecture. This could be because of an issue with the checkpoint, or because your version of Transformers is out of date.
35
-
36
- You can update Transformers with the command `pip install --upgrade transformers`. If this does not work, and the checkpoint is very new, then there may not be a release version that supports this model yet. In this case, you can get the most up-to-date code by installing Transformers from source with the command `pip install git+https://github.com/huggingface/transformers.git`
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
  ERROR:
13
  Traceback (most recent call last):
14
+ File "/tmp/cyankiwi_GLM-4.7-Flash-AWQ-4bit_04Cpv3J.py", line 26, in <module>
 
 
 
 
 
 
 
 
 
 
15
  pipe = pipeline("text-generation", model="cyankiwi/GLM-4.7-Flash-AWQ-4bit")
16
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/pipelines/__init__.py", line 836, in pipeline
17
+ model = load_model(
18
+ adapter_path if adapter_path is not None else model,
19
+ ...<4 lines>...
20
+ **model_kwargs,
21
  )
22
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/pipelines/base.py", line 232, in load_model
23
+ model = model_class.from_pretrained(model, **kwargs)
24
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 372, in from_pretrained
25
+ return model_class.from_pretrained(
26
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~^
27
+ pretrained_model_name_or_path, *model_args, config=config, **hub_kwargs, **kwargs
28
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
29
  )
30
+ ^
31
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/modeling_utils.py", line 4015, in from_pretrained
32
+ hf_quantizer, config, device_map = get_hf_quantizer(
33
+ ~~~~~~~~~~~~~~~~^
34
+ config, quantization_config, device_map, weights_only, user_agent
35
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
36
+ )
37
+ ^
38
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 312, in get_hf_quantizer
39
+ config.quantization_config = AutoHfQuantizer.merge_quantization_configs(
40
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
41
+ config.quantization_config, quantization_config
42
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
43
+ )
44
+ ^
45
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 213, in merge_quantization_configs
46
+ quantization_config = AutoQuantizationConfig.from_dict(quantization_config)
47
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/quantizers/auto.py", line 139, in from_dict
48
+ return target_cls.from_dict(quantization_config_dict)
49
+ ~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^
50
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 1203, in from_dict
51
+ return super().from_dict(config_dict, return_unused_kwargs=return_unused_kwargs, **kwargs)
52
+ ~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
53
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 115, in from_dict
54
+ config = cls(**config_dict)
55
+ File "/tmp/.cache/uv/environments-v2/4a272a7d010dadc8/lib/python3.13/site-packages/transformers/utils/quantization_config.py", line 1133, in __init__
56
+ raise ImportError(
57
+ "compressed_tensors is not installed and is required for compressed-tensors quantization. Please install it with `pip install compressed-tensors`."
58
+ )
59
+ ImportError: compressed_tensors is not installed and is required for compressed-tensors quantization. Please install it with `pip install compressed-tensors`.