Update app.py
Browse files
app.py
CHANGED
|
@@ -9,7 +9,7 @@ dtype = torch.bfloat16
|
|
| 9 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 10 |
model = AutoModelForCausalLM.from_pretrained(
|
| 11 |
model_id,
|
| 12 |
-
device_map="cpu",
|
| 13 |
torch_dtype=dtype,
|
| 14 |
)
|
| 15 |
|
|
@@ -44,7 +44,7 @@ if prompt := st.chat_input():
|
|
| 44 |
|
| 45 |
##Get response to the message using client
|
| 46 |
inputs = tokenizer.encode(text, add_special_tokens=False, return_tensors="pt")
|
| 47 |
-
outputs = model.generate(
|
| 48 |
|
| 49 |
|
| 50 |
|
|
|
|
| 9 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
| 10 |
model = AutoModelForCausalLM.from_pretrained(
|
| 11 |
model_id,
|
| 12 |
+
#device_map="cpu",
|
| 13 |
torch_dtype=dtype,
|
| 14 |
)
|
| 15 |
|
|
|
|
| 44 |
|
| 45 |
##Get response to the message using client
|
| 46 |
inputs = tokenizer.encode(text, add_special_tokens=False, return_tensors="pt")
|
| 47 |
+
outputs = model.generate(**inputs, max_new_tokens=150)
|
| 48 |
|
| 49 |
|
| 50 |
|