Spaces:
Sleeping
Sleeping
clean code
Browse files
app.py
CHANGED
|
@@ -963,16 +963,6 @@ def respond(user_message, data, chat_history, socratic_mode=False):
|
|
| 963 |
如果學生的問題與資料文本無關,請告訴學生你無法回答超出範圍的問題
|
| 964 |
"""
|
| 965 |
|
| 966 |
-
print("=== socratic_mode ===")
|
| 967 |
-
print(socratic_mode)
|
| 968 |
-
print("=== socratic_mode ===")
|
| 969 |
-
|
| 970 |
-
print("=== sys_content ===")
|
| 971 |
-
print(sys_content)
|
| 972 |
-
print("=== sys_content ===")
|
| 973 |
-
|
| 974 |
-
|
| 975 |
-
|
| 976 |
messages = [
|
| 977 |
{"role": "system", "content": sys_content}
|
| 978 |
]
|
|
@@ -991,26 +981,16 @@ def respond(user_message, data, chat_history, socratic_mode=False):
|
|
| 991 |
{"role": "assistant", "content": chat[1]}
|
| 992 |
]
|
| 993 |
messages += old_messages
|
| 994 |
-
|
| 995 |
else:
|
| 996 |
pass
|
| 997 |
|
| 998 |
messages.append({"role": "user", "content": user_message})
|
| 999 |
-
|
| 1000 |
-
print("=====messages=====")
|
| 1001 |
-
print(messages)
|
| 1002 |
-
print("=====messages=====")
|
| 1003 |
-
|
| 1004 |
-
|
| 1005 |
request_payload = {
|
| 1006 |
"model": "gpt-4-1106-preview",
|
| 1007 |
"messages": messages,
|
| 1008 |
"max_tokens": 4000 # 設定一個較大的值,可根據需要調整
|
| 1009 |
}
|
| 1010 |
-
|
| 1011 |
response = client.chat.completions.create(**request_payload)
|
| 1012 |
-
print(response)
|
| 1013 |
-
|
| 1014 |
response_text = response.choices[0].message.content.strip()
|
| 1015 |
|
| 1016 |
# 更新聊天历史
|
|
|
|
| 963 |
如果學生的問題與資料文本無關,請告訴學生你無法回答超出範圍的問題
|
| 964 |
"""
|
| 965 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 966 |
messages = [
|
| 967 |
{"role": "system", "content": sys_content}
|
| 968 |
]
|
|
|
|
| 981 |
{"role": "assistant", "content": chat[1]}
|
| 982 |
]
|
| 983 |
messages += old_messages
|
|
|
|
| 984 |
else:
|
| 985 |
pass
|
| 986 |
|
| 987 |
messages.append({"role": "user", "content": user_message})
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 988 |
request_payload = {
|
| 989 |
"model": "gpt-4-1106-preview",
|
| 990 |
"messages": messages,
|
| 991 |
"max_tokens": 4000 # 設定一個較大的值,可根據需要調整
|
| 992 |
}
|
|
|
|
| 993 |
response = client.chat.completions.create(**request_payload)
|
|
|
|
|
|
|
| 994 |
response_text = response.choices[0].message.content.strip()
|
| 995 |
|
| 996 |
# 更新聊天历史
|