Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -338,14 +338,15 @@ def generate(text, history, rag_option, model_option, k=3, top_p=0.6, temperatu
|
|
| 338 |
result = response.json()
|
| 339 |
print("result:------------------")
|
| 340 |
chatbot_response = result[0]['generated_text']
|
| 341 |
-
print(chatbot_response)
|
| 342 |
print("anzahl tokens gesamt antwort:------------------")
|
| 343 |
print (len(chatbot_response.split()))
|
| 344 |
except Exception as e:
|
| 345 |
raise gr.Error(e)
|
| 346 |
|
| 347 |
chatbot_message = chatbot_response[len(prompt):].strip()
|
| 348 |
-
|
|
|
|
|
|
|
| 349 |
|
| 350 |
#Antwort als Stream ausgeben...
|
| 351 |
for i in range(len(chatbot_message)):
|
|
|
|
| 338 |
result = response.json()
|
| 339 |
print("result:------------------")
|
| 340 |
chatbot_response = result[0]['generated_text']
|
|
|
|
| 341 |
print("anzahl tokens gesamt antwort:------------------")
|
| 342 |
print (len(chatbot_response.split()))
|
| 343 |
except Exception as e:
|
| 344 |
raise gr.Error(e)
|
| 345 |
|
| 346 |
chatbot_message = chatbot_response[len(prompt):].strip()
|
| 347 |
+
print("history/chatbot_rsponse:--------------------------------")
|
| 348 |
+
print(history)
|
| 349 |
+
print(chatbot_message)
|
| 350 |
|
| 351 |
#Antwort als Stream ausgeben...
|
| 352 |
for i in range(len(chatbot_message)):
|