Spaces:
Sleeping
Sleeping
Update config.py
Browse files
config.py
CHANGED
|
@@ -102,24 +102,14 @@ You should acknowledge and welcome the student to the conversation. Introduce yo
|
|
| 102 |
# - Input Cost per 1M Tokens: $0.15
|
| 103 |
# - Output Cost per 1M Tokens: $0.60
|
| 104 |
#
|
| 105 |
-
# - **Model:** o1
|
| 106 |
-
# - Context Length: 128K
|
| 107 |
-
# - Input Cost per 1M Tokens: $15.00
|
| 108 |
-
# - Output Cost per 1M Tokens: $60.00
|
| 109 |
-
#
|
| 110 |
-
# - **Model:** o1-mini
|
| 111 |
-
# - Context Length: 128K
|
| 112 |
-
# - Input Cost per 1M Tokens: $3.00
|
| 113 |
-
# - Output Cost per 1M Tokens: $12.00
|
| 114 |
-
|
| 115 |
# The model_name refers to the name of the model you want to use. You can choose from the following models:
|
| 116 |
ai_model = "gpt-4o"
|
| 117 |
|
| 118 |
# Temperature refers to the randomness/creativity of the responses. A higher temperature will result in more random/creative responses. It varies between 0 and 1.
|
| 119 |
-
temperature = 0.
|
| 120 |
|
| 121 |
# Max_tokens refers to the maximum number of tokens (words) the AI can generate. The higher the number, the longer the response. It varies between 1 and 2048.
|
| 122 |
-
max_tokens =
|
| 123 |
|
| 124 |
# Frequency penalty parameter for the response. Higher penalty will result in more diverse responses. It varies between 0 and 1.
|
| 125 |
frequency_penalty = 0.5
|
|
|
|
| 102 |
# - Input Cost per 1M Tokens: $0.15
|
| 103 |
# - Output Cost per 1M Tokens: $0.60
|
| 104 |
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 105 |
# The model_name refers to the name of the model you want to use. You can choose from the following models:
|
| 106 |
ai_model = "gpt-4o"
|
| 107 |
|
| 108 |
# Temperature refers to the randomness/creativity of the responses. A higher temperature will result in more random/creative responses. It varies between 0 and 1.
|
| 109 |
+
temperature = 0.3
|
| 110 |
|
| 111 |
# Max_tokens refers to the maximum number of tokens (words) the AI can generate. The higher the number, the longer the response. It varies between 1 and 2048.
|
| 112 |
+
max_tokens = 1000
|
| 113 |
|
| 114 |
# Frequency penalty parameter for the response. Higher penalty will result in more diverse responses. It varies between 0 and 1.
|
| 115 |
frequency_penalty = 0.5
|