Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -40,11 +40,8 @@ def get_ai_response(prompt, chat_history):
|
|
| 40 |
st.session_state.is_streaming = True
|
| 41 |
st.session_state.response = ""
|
| 42 |
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
system_tokens = num_tokens_from_string(system_message["content"])
|
| 47 |
-
user_tokens = num_tokens_from_string(user_message["content"])
|
| 48 |
|
| 49 |
# Maximum allowed tokens
|
| 50 |
context_length = 4096 - max_tokens
|
|
@@ -63,7 +60,7 @@ def get_ai_response(prompt, chat_history):
|
|
| 63 |
break
|
| 64 |
|
| 65 |
# Construct final message list
|
| 66 |
-
final_messages = [
|
| 67 |
|
| 68 |
try:
|
| 69 |
with st.chat_message("assistant", avatar=st.session_state.assistant_avatar):
|
|
|
|
| 40 |
st.session_state.is_streaming = True
|
| 41 |
st.session_state.response = ""
|
| 42 |
|
| 43 |
+
system_tokens = num_tokens_from_string(system_prompt)
|
| 44 |
+
user_tokens = num_tokens_from_string(prompt)
|
|
|
|
|
|
|
|
|
|
| 45 |
|
| 46 |
# Maximum allowed tokens
|
| 47 |
context_length = 4096 - max_tokens
|
|
|
|
| 60 |
break
|
| 61 |
|
| 62 |
# Construct final message list
|
| 63 |
+
final_messages = [{"role": "system", "content": system_prompt}] + trimmed_history + [{"role": "user", "content": prompt}]
|
| 64 |
|
| 65 |
try:
|
| 66 |
with st.chat_message("assistant", avatar=st.session_state.assistant_avatar):
|