ariG23498 HF Staff commited on
Commit
9454dc8
·
verified ·
1 Parent(s): ddaed64

Upload google_translategemma-4b-it_2.txt with huggingface_hub

Browse files
Files changed (1) hide show
  1. google_translategemma-4b-it_2.txt +71 -0
google_translategemma-4b-it_2.txt ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ```CODE:
2
+ # Load model directly
3
+ from transformers import AutoProcessor, AutoModelForImageTextToText
4
+
5
+ processor = AutoProcessor.from_pretrained("google/translategemma-4b-it")
6
+ model = AutoModelForImageTextToText.from_pretrained("google/translategemma-4b-it")
7
+ messages = [
8
+ {
9
+ "role": "user",
10
+ "content": [
11
+ {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/p-blog/candy.JPG"},
12
+ {"type": "text", "text": "What animal is on the candy?"}
13
+ ]
14
+ },
15
+ ]
16
+ inputs = processor.apply_chat_template(
17
+ messages,
18
+ add_generation_prompt=True,
19
+ tokenize=True,
20
+ return_dict=True,
21
+ return_tensors="pt",
22
+ ).to(model.device)
23
+
24
+ outputs = model.generate(**inputs, max_new_tokens=40)
25
+ print(processor.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
26
+ ```
27
+
28
+ ERROR:
29
+ Traceback (most recent call last):
30
+ File "/tmp/google_translategemma-4b-it_2j8QQQp.py", line 37, in <module>
31
+ inputs = processor.apply_chat_template(
32
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^
33
+ messages,
34
+ ^^^^^^^^^
35
+ ...<3 lines>...
36
+ return_tensors="pt",
37
+ ^^^^^^^^^^^^^^^^^^^^
38
+ ).to(model.device)
39
+ ^
40
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
41
+ return func(*args, **kwargs)
42
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func
43
+ return func(*args, **kwargs)
44
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/transformers/processing_utils.py", line 1675, in apply_chat_template
45
+ prompt, generation_indices = render_jinja_template(
46
+ ~~~~~~~~~~~~~~~~~~~~~^
47
+ conversations=conversations,
48
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
49
+ ...<2 lines>...
50
+ **self.tokenizer.special_tokens_map, # tokenizer special tokens are used by some templates
51
+ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
52
+ )
53
+ ^
54
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/transformers/utils/chat_template_utils.py", line 539, in render_jinja_template
55
+ rendered_chat = compiled_template.render(
56
+ messages=chat,
57
+ ...<3 lines>...
58
+ **kwargs,
59
+ )
60
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/jinja2/environment.py", line 1295, in render
61
+ self.environment.handle_exception()
62
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^
63
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/jinja2/environment.py", line 942, in handle_exception
64
+ raise rewrite_traceback_stack(source=source)
65
+ File "<template>", line 601, in top-level template code
66
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/jinja2/sandbox.py", line 401, in call
67
+ return __context.call(__obj, *args, **kwargs)
68
+ ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^
69
+ File "/tmp/.cache/uv/environments-v2/472dbc3769380f36/lib/python3.13/site-packages/transformers/utils/chat_template_utils.py", line 447, in raise_exception
70
+ raise jinja2.exceptions.TemplateError(message)
71
+ jinja2.exceptions.TemplateError: User role must provide `content` as an iterable with exactly one item. That item must be a `mapping(type:'text' | 'image', source_lang_code:string, target_lang_code:string, text:string | none, image:string | none)`.