Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -144,8 +144,8 @@ def chat_with_model(prompt, document_section, model_choice='gpt-3.5-turbo'):
|
|
| 144 |
chunk_message = chunk['choices'][0]['delta'] # extract the message
|
| 145 |
collected_messages.append(chunk_message) # save the message
|
| 146 |
content=chunk["choices"][0].get("delta",{}).get("content")
|
| 147 |
-
st.write(f'*{content}*')
|
| 148 |
-
st.write(f"Full response received {chunk_time:.2f} seconds after request")
|
| 149 |
full_reply_content = ''.join([m.get('content', '') for m in collected_messages])
|
| 150 |
st.write(f"Full conversation received: {full_reply_content}")
|
| 151 |
return full_reply_content
|
|
|
|
| 144 |
chunk_message = chunk['choices'][0]['delta'] # extract the message
|
| 145 |
collected_messages.append(chunk_message) # save the message
|
| 146 |
content=chunk["choices"][0].get("delta",{}).get("content")
|
| 147 |
+
#st.write(f'*{content}*')
|
| 148 |
+
#st.write(f"Full response received {chunk_time:.2f} seconds after request")
|
| 149 |
full_reply_content = ''.join([m.get('content', '') for m in collected_messages])
|
| 150 |
st.write(f"Full conversation received: {full_reply_content}")
|
| 151 |
return full_reply_content
|