Spaces:
Sleeping
Sleeping
Update LLMPipeline.py
Browse files- LLMPipeline.py +1 -1
LLMPipeline.py
CHANGED
|
@@ -78,7 +78,7 @@ def generate_image_prompt_viejo(summary):
|
|
| 78 |
prompt = f"Crea un prompt para imagen tipo meme sobre este titular de noticias:\n{summary}\nEl prompt debe ser corto, visual y gracioso. Devuelve solo el prompt."
|
| 79 |
result = text_generator(prompt, max_new_tokens=400, do_sample=True, temperature=0.7)[0]['generated_text']
|
| 80 |
return result.strip()
|
| 81 |
-
|
| 82 |
tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-small")
|
| 83 |
model = T5ForConditionalGeneration.from_pretrained("google/flan-t5-small")
|
| 84 |
|
|
|
|
| 78 |
prompt = f"Crea un prompt para imagen tipo meme sobre este titular de noticias:\n{summary}\nEl prompt debe ser corto, visual y gracioso. Devuelve solo el prompt."
|
| 79 |
result = text_generator(prompt, max_new_tokens=400, do_sample=True, temperature=0.7)[0]['generated_text']
|
| 80 |
return result.strip()
|
| 81 |
+
from transformers import T5Tokenizer, T5ForConditionalGeneration
|
| 82 |
tokenizer = T5Tokenizer.from_pretrained("google/flan-t5-small")
|
| 83 |
model = T5ForConditionalGeneration.from_pretrained("google/flan-t5-small")
|
| 84 |
|