Upload LiquidAI_LFM2.5-VL-1.6B_1.txt with huggingface_hub
Browse files
LiquidAI_LFM2.5-VL-1.6B_1.txt
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
```CODE:
|
| 2 |
# Load model directly
|
| 3 |
-
from transformers import AutoProcessor,
|
| 4 |
|
| 5 |
processor = AutoProcessor.from_pretrained("LiquidAI/LFM2.5-VL-1.6B")
|
| 6 |
-
model =
|
| 7 |
messages = [
|
| 8 |
{
|
| 9 |
"role": "user",
|
|
@@ -27,6 +27,22 @@ print(processor.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
|
|
| 27 |
|
| 28 |
ERROR:
|
| 29 |
Traceback (most recent call last):
|
| 30 |
-
File "/tmp/LiquidAI_LFM2.5-VL-1.
|
| 31 |
-
|
| 32 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
```CODE:
|
| 2 |
# Load model directly
|
| 3 |
+
from transformers import AutoProcessor, AutoModelForImageTextToText
|
| 4 |
|
| 5 |
processor = AutoProcessor.from_pretrained("LiquidAI/LFM2.5-VL-1.6B")
|
| 6 |
+
model = AutoModelForImageTextToText.from_pretrained("LiquidAI/LFM2.5-VL-1.6B")
|
| 7 |
messages = [
|
| 8 |
{
|
| 9 |
"role": "user",
|
|
|
|
| 27 |
|
| 28 |
ERROR:
|
| 29 |
Traceback (most recent call last):
|
| 30 |
+
File "/tmp/LiquidAI_LFM2.5-VL-1.6B_1Hgbe78.py", line 26, in <module>
|
| 31 |
+
processor = AutoProcessor.from_pretrained("LiquidAI/LFM2.5-VL-1.6B")
|
| 32 |
+
File "/tmp/.cache/uv/environments-v2/27a07e03a063efb6/lib/python3.13/site-packages/transformers/models/auto/processing_auto.py", line 396, in from_pretrained
|
| 33 |
+
return processor_class.from_pretrained(
|
| 34 |
+
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
|
| 35 |
+
pretrained_model_name_or_path, trust_remote_code=trust_remote_code, **kwargs
|
| 36 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 37 |
+
)
|
| 38 |
+
^
|
| 39 |
+
File "/tmp/.cache/uv/environments-v2/27a07e03a063efb6/lib/python3.13/site-packages/transformers/processing_utils.py", line 1394, in from_pretrained
|
| 40 |
+
args = cls._get_arguments_from_pretrained(pretrained_model_name_or_path, **kwargs)
|
| 41 |
+
File "/tmp/.cache/uv/environments-v2/27a07e03a063efb6/lib/python3.13/site-packages/transformers/processing_utils.py", line 1453, in _get_arguments_from_pretrained
|
| 42 |
+
args.append(attribute_class.from_pretrained(pretrained_model_name_or_path, **kwargs))
|
| 43 |
+
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 44 |
+
File "/tmp/.cache/uv/environments-v2/27a07e03a063efb6/lib/python3.13/site-packages/transformers/models/auto/tokenization_auto.py", line 1153, in from_pretrained
|
| 45 |
+
raise ValueError(
|
| 46 |
+
f"Tokenizer class {tokenizer_class_candidate} does not exist or is not currently imported."
|
| 47 |
+
)
|
| 48 |
+
ValueError: Tokenizer class TokenizersBackend does not exist or is not currently imported.
|