Spaces:
Sleeping
Sleeping
change max tokens
Browse files
app.py
CHANGED
|
@@ -199,7 +199,7 @@ def generate_response(audio_path, message, chat_history=None):
|
|
| 199 |
with torch.no_grad():
|
| 200 |
generate_ids = model.generate(
|
| 201 |
**inputs,
|
| 202 |
-
max_new_tokens=
|
| 203 |
do_sample=True,
|
| 204 |
temperature=0.7,
|
| 205 |
top_p=0.9,
|
|
|
|
| 199 |
with torch.no_grad():
|
| 200 |
generate_ids = model.generate(
|
| 201 |
**inputs,
|
| 202 |
+
max_new_tokens=300,
|
| 203 |
do_sample=True,
|
| 204 |
temperature=0.7,
|
| 205 |
top_p=0.9,
|