tebicap commited on
Commit
4aaf034
·
1 Parent(s): fee7bb8

prueba como en la documentación

Browse files
Files changed (1) hide show
  1. app.py +5 -7
app.py CHANGED
@@ -9,16 +9,14 @@ tokenizer = T5Tokenizer.from_pretrained(model_name)
9
  # Define a function to generate text using T5
10
  def generate_text(prompt):
11
  # Tokenize input and generate output
12
- input_ids = tokenizer.encode(prompt, return_tensors="pt", max_length=512, truncation=True)
 
 
13
  output_ids = model.generate(input_ids)
14
 
15
- for i in range(len(output_ids)):
16
- print(f"-----partes del output: ({i})")
17
- print(tokenizer.decode(output_ids[i], skip_special_tokens=True))
18
- print("---")
19
-
20
  # Decode the generated output
21
- generated_text = tokenizer.decode(output_ids[0], skip_special_tokens=True)
 
22
 
23
  return generated_text
24
 
 
9
  # Define a function to generate text using T5
10
  def generate_text(prompt):
11
  # Tokenize input and generate output
12
+ #input_ids = tokenizer.encode(prompt, return_tensors="pt", max_length=512, truncation=True)
13
+ input_ids = tokenizer.encode(prompt, return_tensors="pt").input_ids
14
+
15
  output_ids = model.generate(input_ids)
16
 
 
 
 
 
 
17
  # Decode the generated output
18
+ #generated_text = tokenizer.decode(output_ids[0], skip_special_tokens=True)
19
+ generated_text = tokenizer.decode(outputs[0])
20
 
21
  return generated_text
22