Upload moonshotai_Kimi-K2.5_1.txt with huggingface_hub
Browse files
moonshotai_Kimi-K2.5_1.txt
CHANGED
|
@@ -6,7 +6,7 @@ model = AutoModel.from_pretrained("moonshotai/Kimi-K2.5", trust_remote_code=True
|
|
| 6 |
|
| 7 |
ERROR:
|
| 8 |
Traceback (most recent call last):
|
| 9 |
-
File "/tmp/moonshotai_Kimi-K2.
|
| 10 |
model = AutoModel.from_pretrained("moonshotai/Kimi-K2.5", trust_remote_code=True, dtype="auto")
|
| 11 |
File "/tmp/.cache/uv/environments-v2/1fa8411085c3a50c/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 354, in from_pretrained
|
| 12 |
model_class = get_class_from_dynamic_module(
|
|
@@ -19,8 +19,8 @@ Traceback (most recent call last):
|
|
| 19 |
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^
|
| 20 |
File "<frozen importlib._bootstrap_external>", line 1026, in exec_module
|
| 21 |
File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
|
| 22 |
-
File "/tmp/.cache/huggingface/modules/transformers_modules/moonshotai/Kimi_hyphen_K2_dot_5/
|
| 23 |
from .modeling_deepseek import DeepseekV3ForCausalLM
|
| 24 |
-
File "/tmp/.cache/huggingface/modules/transformers_modules/moonshotai/Kimi_hyphen_K2_dot_5/
|
| 25 |
from transformers.utils.import_utils import is_torch_fx_available
|
| 26 |
ImportError: cannot import name 'is_torch_fx_available' from 'transformers.utils.import_utils' (/tmp/.cache/uv/environments-v2/1fa8411085c3a50c/lib/python3.13/site-packages/transformers/utils/import_utils.py)
|
|
|
|
| 6 |
|
| 7 |
ERROR:
|
| 8 |
Traceback (most recent call last):
|
| 9 |
+
File "/tmp/moonshotai_Kimi-K2.5_1DsYEr5.py", line 25, in <module>
|
| 10 |
model = AutoModel.from_pretrained("moonshotai/Kimi-K2.5", trust_remote_code=True, dtype="auto")
|
| 11 |
File "/tmp/.cache/uv/environments-v2/1fa8411085c3a50c/lib/python3.13/site-packages/transformers/models/auto/auto_factory.py", line 354, in from_pretrained
|
| 12 |
model_class = get_class_from_dynamic_module(
|
|
|
|
| 19 |
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^
|
| 20 |
File "<frozen importlib._bootstrap_external>", line 1026, in exec_module
|
| 21 |
File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
|
| 22 |
+
File "/tmp/.cache/huggingface/modules/transformers_modules/moonshotai/Kimi_hyphen_K2_dot_5/907b3ee811fde9c0f6011b4a9f4050a506b3c566/modeling_kimi_k25.py", line 67, in <module>
|
| 23 |
from .modeling_deepseek import DeepseekV3ForCausalLM
|
| 24 |
+
File "/tmp/.cache/huggingface/modules/transformers_modules/moonshotai/Kimi_hyphen_K2_dot_5/907b3ee811fde9c0f6011b4a9f4050a506b3c566/modeling_deepseek.py", line 47, in <module>
|
| 25 |
from transformers.utils.import_utils import is_torch_fx_available
|
| 26 |
ImportError: cannot import name 'is_torch_fx_available' from 'transformers.utils.import_utils' (/tmp/.cache/uv/environments-v2/1fa8411085c3a50c/lib/python3.13/site-packages/transformers/utils/import_utils.py)
|