Reducing default token generation to 15 tokens to prevent the model from taking too long to respond.
Browse files
app.py
CHANGED
|
@@ -70,7 +70,7 @@ demo = gr.ChatInterface(
|
|
| 70 |
type='messages',
|
| 71 |
additional_inputs=[
|
| 72 |
gr.Dropdown(choices=["DialoGPT-med-FT", "DialoGPT-medium"], value="DialoGPT-med-FT", label="Model"),
|
| 73 |
-
gr.Slider(minimum=1, maximum=
|
| 74 |
gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
|
| 75 |
gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
|
| 76 |
],
|
|
|
|
| 70 |
type='messages',
|
| 71 |
additional_inputs=[
|
| 72 |
gr.Dropdown(choices=["DialoGPT-med-FT", "DialoGPT-medium"], value="DialoGPT-med-FT", label="Model"),
|
| 73 |
+
gr.Slider(minimum=1, maximum=100, value=15, step=1, label="Max new tokens"),
|
| 74 |
gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
|
| 75 |
gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p (nucleus sampling)"),
|
| 76 |
],
|