Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -271,9 +271,7 @@ def predict(
|
|
| 271 |
response = strip_stop_words(full_response, stop_words)
|
| 272 |
conversation.update_last_message(response)
|
| 273 |
gradio_chatbot_output[-1][1] = response
|
| 274 |
-
yield gradio_chatbot_output, to_gradio_history(
|
| 275 |
-
conversation
|
| 276 |
-
), "Generating..."
|
| 277 |
|
| 278 |
print("flushed result to gradio")
|
| 279 |
torch.cuda.empty_cache()
|
|
@@ -284,7 +282,7 @@ def predict(
|
|
| 284 |
f"temperature: {temperature}, top_p: {top_p}, repetition_penalty: {repetition_penalty}, max_length_tokens: {max_length_tokens}"
|
| 285 |
)
|
| 286 |
|
| 287 |
-
yield gradio_chatbot_output
|
| 288 |
|
| 289 |
|
| 290 |
def retry(
|
|
|
|
| 271 |
response = strip_stop_words(full_response, stop_words)
|
| 272 |
conversation.update_last_message(response)
|
| 273 |
gradio_chatbot_output[-1][1] = response
|
| 274 |
+
yield gradio_chatbot_output, #to_gradio_history(conversation), "Generating..."
|
|
|
|
|
|
|
| 275 |
|
| 276 |
print("flushed result to gradio")
|
| 277 |
torch.cuda.empty_cache()
|
|
|
|
| 282 |
f"temperature: {temperature}, top_p: {top_p}, repetition_penalty: {repetition_penalty}, max_length_tokens: {max_length_tokens}"
|
| 283 |
)
|
| 284 |
|
| 285 |
+
yield gradio_chatbot_output,, "Generate: Success",# to_gradio_history(conversation)
|
| 286 |
|
| 287 |
|
| 288 |
def retry(
|