Spaces:
Sleeping
Sleeping
Upload folder using huggingface_hub
Browse files
app.py
CHANGED
|
@@ -1,12 +1,11 @@
|
|
| 1 |
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
|
| 2 |
import gradio as gr
|
| 3 |
|
| 4 |
-
model = AutoModelForSeq2SeqLM.from_pretrained("google/flan-t5-
|
| 5 |
-
tokenizer = AutoTokenizer.from_pretrained("google/flan-t5-
|
| 6 |
|
| 7 |
def index(prompt):
|
| 8 |
-
|
| 9 |
-
inputs = tokenizer(prompt, return_tensors="pt", max_length=512, truncation=True)
|
| 10 |
outputs = model.generate(**inputs)
|
| 11 |
|
| 12 |
decoded_outputs = tokenizer.batch_decode(outputs, skip_special_tokens=True)
|
|
|
|
| 1 |
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
|
| 2 |
import gradio as gr
|
| 3 |
|
| 4 |
+
model = AutoModelForSeq2SeqLM.from_pretrained("google/flan-t5-large")
|
| 5 |
+
tokenizer = AutoTokenizer.from_pretrained("google/flan-t5-large")
|
| 6 |
|
| 7 |
def index(prompt):
|
| 8 |
+
inputs = tokenizer(prompt, return_tensors="pt")
|
|
|
|
| 9 |
outputs = model.generate(**inputs)
|
| 10 |
|
| 11 |
decoded_outputs = tokenizer.batch_decode(outputs, skip_special_tokens=True)
|