Upload modeling_mic21.py with huggingface_hub
Browse files- modeling_mic21.py +1 -1
modeling_mic21.py
CHANGED
|
@@ -78,7 +78,7 @@ class MIC21SummarizerModel(PreTrainedModel):
|
|
| 78 |
{"role":"system","content":"Generate title and description for the provided image. The image features are: "},
|
| 79 |
{"role":"user","content":"Generate a title:"}]
|
| 80 |
|
| 81 |
-
tokenized_messages = self.components["tokenizer"].apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
|
| 82 |
#.to(self.in_device)
|
| 83 |
vectorized_messages = self.components["llm"].model.embed_tokens(tokenized_messages[0]).unsqueeze(0)
|
| 84 |
vectorized_messages = vectorized_messages.repeat(batch_size,1,1)
|
|
|
|
| 78 |
{"role":"system","content":"Generate title and description for the provided image. The image features are: "},
|
| 79 |
{"role":"user","content":"Generate a title:"}]
|
| 80 |
|
| 81 |
+
tokenized_messages = self.components["tokenizer"].apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt").cuda()
|
| 82 |
#.to(self.in_device)
|
| 83 |
vectorized_messages = self.components["llm"].model.embed_tokens(tokenized_messages[0]).unsqueeze(0)
|
| 84 |
vectorized_messages = vectorized_messages.repeat(batch_size,1,1)
|