reduced tokens to 150
Browse files
app.py
CHANGED
|
@@ -107,7 +107,7 @@ def respond(message, history):
|
|
| 107 |
with torch.no_grad():
|
| 108 |
chat_history_ids = model.generate(
|
| 109 |
input_ids,
|
| 110 |
-
max_length=
|
| 111 |
temperature=0.9,
|
| 112 |
top_p=0.9,
|
| 113 |
repetition_penalty=1.2,
|
|
|
|
| 107 |
with torch.no_grad():
|
| 108 |
chat_history_ids = model.generate(
|
| 109 |
input_ids,
|
| 110 |
+
max_length=150,
|
| 111 |
temperature=0.9,
|
| 112 |
top_p=0.9,
|
| 113 |
repetition_penalty=1.2,
|