Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -67,7 +67,7 @@ class Resume_Overall:
|
|
| 67 |
# Generate a response from the GPT-3 model
|
| 68 |
chat_completion = self.client.chat.completions.create(
|
| 69 |
model = "ChatGPT",
|
| 70 |
-
|
| 71 |
max_tokens=200,
|
| 72 |
temperature=0,
|
| 73 |
n=1,
|
|
@@ -94,7 +94,7 @@ class Resume_Overall:
|
|
| 94 |
# Generate a response from the GPT-3 model
|
| 95 |
chat_completion = self.client.chat.completions.create(
|
| 96 |
model = "ChatGPT",
|
| 97 |
-
|
| 98 |
max_tokens=200,
|
| 99 |
temperature=0,
|
| 100 |
n=1,
|
|
@@ -123,7 +123,7 @@ class Resume_Overall:
|
|
| 123 |
# Generate a response from the GPT-3 model
|
| 124 |
chat_completion = self.client.chat.completions.create(
|
| 125 |
model = "ChatGPT", # Choose the GPT-3 engine you want to use
|
| 126 |
-
|
| 127 |
max_tokens=100, # Set the maximum number of tokens in the generated response
|
| 128 |
temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
|
| 129 |
n=1, # Generate a single response
|
|
@@ -143,7 +143,7 @@ class Resume_Overall:
|
|
| 143 |
|
| 144 |
chat_completion = self.client.chat.completions.create(
|
| 145 |
model = "ChatGPT",
|
| 146 |
-
|
| 147 |
max_tokens=100,
|
| 148 |
temperature=0,
|
| 149 |
n=1,
|
|
@@ -176,7 +176,7 @@ class Resume_Overall:
|
|
| 176 |
|
| 177 |
chat_completion = self.client.chat.completions.create(
|
| 178 |
model = "ChatGPT", # Choose the GPT-3 engine you want to use
|
| 179 |
-
|
| 180 |
max_tokens=500, # Set the maximum number of tokens in the generated response
|
| 181 |
temperature=0.5, # Controls the randomness of the output. Higher values = more random, lower values = more focused
|
| 182 |
)
|
|
@@ -203,7 +203,7 @@ class Resume_Overall:
|
|
| 203 |
# Generate a response from the GPT-3 model
|
| 204 |
chat_completion = self.client.chat.completions.create(
|
| 205 |
model = "ChatGPT", # Choose the GPT-3 engine you want to use
|
| 206 |
-
|
| 207 |
max_tokens=200, # Set the maximum number of tokens in the generated response
|
| 208 |
temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
|
| 209 |
n=1, # Generate a single response
|
|
|
|
| 67 |
# Generate a response from the GPT-3 model
|
| 68 |
chat_completion = self.client.chat.completions.create(
|
| 69 |
model = "ChatGPT",
|
| 70 |
+
messages = conversation,
|
| 71 |
max_tokens=200,
|
| 72 |
temperature=0,
|
| 73 |
n=1,
|
|
|
|
| 94 |
# Generate a response from the GPT-3 model
|
| 95 |
chat_completion = self.client.chat.completions.create(
|
| 96 |
model = "ChatGPT",
|
| 97 |
+
messages = conversation,
|
| 98 |
max_tokens=200,
|
| 99 |
temperature=0,
|
| 100 |
n=1,
|
|
|
|
| 123 |
# Generate a response from the GPT-3 model
|
| 124 |
chat_completion = self.client.chat.completions.create(
|
| 125 |
model = "ChatGPT", # Choose the GPT-3 engine you want to use
|
| 126 |
+
messages = conversation,
|
| 127 |
max_tokens=100, # Set the maximum number of tokens in the generated response
|
| 128 |
temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
|
| 129 |
n=1, # Generate a single response
|
|
|
|
| 143 |
|
| 144 |
chat_completion = self.client.chat.completions.create(
|
| 145 |
model = "ChatGPT",
|
| 146 |
+
messages = conversation,
|
| 147 |
max_tokens=100,
|
| 148 |
temperature=0,
|
| 149 |
n=1,
|
|
|
|
| 176 |
|
| 177 |
chat_completion = self.client.chat.completions.create(
|
| 178 |
model = "ChatGPT", # Choose the GPT-3 engine you want to use
|
| 179 |
+
messages = conversation,
|
| 180 |
max_tokens=500, # Set the maximum number of tokens in the generated response
|
| 181 |
temperature=0.5, # Controls the randomness of the output. Higher values = more random, lower values = more focused
|
| 182 |
)
|
|
|
|
| 203 |
# Generate a response from the GPT-3 model
|
| 204 |
chat_completion = self.client.chat.completions.create(
|
| 205 |
model = "ChatGPT", # Choose the GPT-3 engine you want to use
|
| 206 |
+
messages = conversation,
|
| 207 |
max_tokens=200, # Set the maximum number of tokens in the generated response
|
| 208 |
temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
|
| 209 |
n=1, # Generate a single response
|