Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -28,16 +28,15 @@ def predict(message, history):
|
|
| 28 |
top_p=0.2,
|
| 29 |
top_k=20,
|
| 30 |
temperature=0.1,
|
|
|
|
|
|
|
| 31 |
num_beams=1
|
| 32 |
)
|
| 33 |
t = Thread(target=model.generate, kwargs=generate_kwargs)
|
| 34 |
t.start() # Starting the generation in a separate thread.
|
| 35 |
partial_message = ""
|
| 36 |
for new_token in streamer:
|
| 37 |
-
|
| 38 |
-
if '</s>' in partial_message: # Breaking the loop if the stop token is generated.
|
| 39 |
-
break
|
| 40 |
-
yield partial_message
|
| 41 |
|
| 42 |
# Setting up the Gradio chat interface.
|
| 43 |
gr.ChatInterface(predict,
|
|
|
|
| 28 |
top_p=0.2,
|
| 29 |
top_k=20,
|
| 30 |
temperature=0.1,
|
| 31 |
+
repetition_penalty=2.0,
|
| 32 |
+
length_penalty=-0.5,
|
| 33 |
num_beams=1
|
| 34 |
)
|
| 35 |
t = Thread(target=model.generate, kwargs=generate_kwargs)
|
| 36 |
t.start() # Starting the generation in a separate thread.
|
| 37 |
partial_message = ""
|
| 38 |
for new_token in streamer:
|
| 39 |
+
yield new_token
|
|
|
|
|
|
|
|
|
|
| 40 |
|
| 41 |
# Setting up the Gradio chat interface.
|
| 42 |
gr.ChatInterface(predict,
|