try to improve the inference
Browse files
app.py
CHANGED
|
@@ -46,7 +46,7 @@ def chat(message, history):
|
|
| 46 |
with torch.no_grad():
|
| 47 |
outputs = model.generate(
|
| 48 |
**inputs,
|
| 49 |
-
max_new_tokens=
|
| 50 |
temperature=0.7,
|
| 51 |
do_sample=True,
|
| 52 |
use_cache=True,
|
|
|
|
| 46 |
with torch.no_grad():
|
| 47 |
outputs = model.generate(
|
| 48 |
**inputs,
|
| 49 |
+
max_new_tokens=256,
|
| 50 |
temperature=0.7,
|
| 51 |
do_sample=True,
|
| 52 |
use_cache=True,
|