poopoo
Browse files
app.py
CHANGED
|
@@ -6,16 +6,12 @@ from PyPDF2 import PdfReader
|
|
| 6 |
# Models and Tokenizers Setup
|
| 7 |
models = {
|
| 8 |
"Text Generator (Bloom)": {
|
| 9 |
-
"model": AutoModelForCausalLM.from_pretrained(
|
| 10 |
-
"bigscience/bloom-560m",
|
| 11 |
-
device_map="auto",
|
| 12 |
-
torch_dtype="auto"
|
| 13 |
-
),
|
| 14 |
"tokenizer": AutoTokenizer.from_pretrained("bigscience/bloom-560m"),
|
| 15 |
},
|
| 16 |
"PDF Summarizer (T5)": {
|
| 17 |
-
"model": AutoModelForSeq2SeqLM.from_pretrained("t5-small"),
|
| 18 |
-
"tokenizer": AutoTokenizer.from_pretrained("t5-small", use_fast=False),
|
| 19 |
},
|
| 20 |
"Broken Answer (T0pp)": {
|
| 21 |
"model": AutoModelForSeq2SeqLM.from_pretrained("bigscience/T0pp"),
|
|
@@ -37,11 +33,11 @@ def chat_with_model(model_choice, user_message, chat_history, file=None):
|
|
| 37 |
model = model_info["model"]
|
| 38 |
|
| 39 |
# Tokenize Input
|
| 40 |
-
inputs = tokenizer(user_message, return_tensors="pt", truncation=True, max_length=
|
| 41 |
|
| 42 |
# Adjust max_length and parameters for the PDF summarizer
|
| 43 |
-
max_length =
|
| 44 |
-
num_beams =
|
| 45 |
outputs = model.generate(
|
| 46 |
**inputs,
|
| 47 |
max_length=max_length,
|
|
@@ -117,4 +113,4 @@ def create_chat_interface():
|
|
| 117 |
|
| 118 |
if __name__ == "__main__":
|
| 119 |
interface = create_chat_interface()
|
| 120 |
-
interface.launch(server_name="0.0.0.0", server_port=7860)
|
|
|
|
| 6 |
# Models and Tokenizers Setup
|
| 7 |
models = {
|
| 8 |
"Text Generator (Bloom)": {
|
| 9 |
+
"model": AutoModelForCausalLM.from_pretrained("bigscience/bloom-560m"),
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
"tokenizer": AutoTokenizer.from_pretrained("bigscience/bloom-560m"),
|
| 11 |
},
|
| 12 |
"PDF Summarizer (T5)": {
|
| 13 |
+
"model": AutoModelForSeq2SeqLM.from_pretrained("aaliyaan/t5-small-finetuned-career"),
|
| 14 |
+
"tokenizer": AutoTokenizer.from_pretrained("aaliyaan/t5-small-finetuned-career", use_fast=False),
|
| 15 |
},
|
| 16 |
"Broken Answer (T0pp)": {
|
| 17 |
"model": AutoModelForSeq2SeqLM.from_pretrained("bigscience/T0pp"),
|
|
|
|
| 33 |
model = model_info["model"]
|
| 34 |
|
| 35 |
# Tokenize Input
|
| 36 |
+
inputs = tokenizer(user_message, return_tensors="pt", padding=True, truncation=True, max_length=512)
|
| 37 |
|
| 38 |
# Adjust max_length and parameters for the PDF summarizer
|
| 39 |
+
max_length = 150
|
| 40 |
+
num_beams = 5
|
| 41 |
outputs = model.generate(
|
| 42 |
**inputs,
|
| 43 |
max_length=max_length,
|
|
|
|
| 113 |
|
| 114 |
if __name__ == "__main__":
|
| 115 |
interface = create_chat_interface()
|
| 116 |
+
interface.launch(server_name="0.0.0.0", server_port=7860)
|