Update app.py
Browse files
app.py
CHANGED
|
@@ -33,11 +33,11 @@ model_specs = [
|
|
| 33 |
# "tokenizer": AutoTokenizer.from_pretrained("google/flan-t5-xl"),
|
| 34 |
# "model": AutoModelForSeq2SeqLM.from_pretrained("google/flan-t5-xl"),
|
| 35 |
# },
|
| 36 |
-
{
|
| 37 |
-
"name": "LLama",
|
| 38 |
-
"tokenizer": AutoTokenizer.from_pretrained("decapoda-research/llama-7b-hf"),
|
| 39 |
-
"model": AutoModelForCausalLM.from_pretrained("decapoda-research/llama-7b-hf"),
|
| 40 |
-
},
|
| 41 |
]
|
| 42 |
|
| 43 |
def generate_response(model_name, input_text, speak_output):
|
|
|
|
| 33 |
# "tokenizer": AutoTokenizer.from_pretrained("google/flan-t5-xl"),
|
| 34 |
# "model": AutoModelForSeq2SeqLM.from_pretrained("google/flan-t5-xl"),
|
| 35 |
# },
|
| 36 |
+
# {
|
| 37 |
+
# "name": "LLama",
|
| 38 |
+
# "tokenizer": AutoTokenizer.from_pretrained("decapoda-research/llama-7b-hf"),
|
| 39 |
+
# "model": AutoModelForCausalLM.from_pretrained("decapoda-research/llama-7b-hf"),
|
| 40 |
+
# },
|
| 41 |
]
|
| 42 |
|
| 43 |
def generate_response(model_name, input_text, speak_output):
|