Spaces:
Sleeping
Sleeping
change Max Token
Browse files
app.py
CHANGED
|
@@ -38,7 +38,7 @@ def chat_with_gpt(user_message, history):
|
|
| 38 |
response = client.chat.completions.create(
|
| 39 |
model="gpt-4o-mini",
|
| 40 |
messages=messages,
|
| 41 |
-
max_tokens=
|
| 42 |
)
|
| 43 |
bot_response = response.choices[0].message.content.strip()
|
| 44 |
|
|
@@ -68,7 +68,7 @@ def generate_summary_feedback():
|
|
| 68 |
response = client.chat.completions.create(
|
| 69 |
model="gpt-4o-mini",
|
| 70 |
messages=messages,
|
| 71 |
-
max_tokens=
|
| 72 |
)
|
| 73 |
return "Status: Selesai\n\n" + response.choices[0].message.content.strip()
|
| 74 |
except Exception as e:
|
|
|
|
| 38 |
response = client.chat.completions.create(
|
| 39 |
model="gpt-4o-mini",
|
| 40 |
messages=messages,
|
| 41 |
+
max_tokens=300
|
| 42 |
)
|
| 43 |
bot_response = response.choices[0].message.content.strip()
|
| 44 |
|
|
|
|
| 68 |
response = client.chat.completions.create(
|
| 69 |
model="gpt-4o-mini",
|
| 70 |
messages=messages,
|
| 71 |
+
max_tokens=1000
|
| 72 |
)
|
| 73 |
return "Status: Selesai\n\n" + response.choices[0].message.content.strip()
|
| 74 |
except Exception as e:
|