Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
|
@@ -126,8 +126,7 @@ def generate(prompt, history, agent_name=agents[0], sys_prompt="", temperature=0
|
|
| 126 |
|
| 127 |
if VERBOSE:
|
| 128 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
| 129 |
-
return resp
|
| 130 |
-
|
| 131 |
|
| 132 |
def compress_history(purpose, task, history, directory):
|
| 133 |
resp = run_gpt(
|
|
@@ -332,8 +331,7 @@ def generate(
|
|
| 332 |
|
| 333 |
if VERBOSE:
|
| 334 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
| 335 |
-
return resp
|
| 336 |
-
|
| 337 |
|
| 338 |
def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
|
| 339 |
"""Generates text in chunks to avoid token limit errors."""
|
|
|
|
| 126 |
|
| 127 |
if VERBOSE:
|
| 128 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
| 129 |
+
return resp # Return resp here
|
|
|
|
| 130 |
|
| 131 |
def compress_history(purpose, task, history, directory):
|
| 132 |
resp = run_gpt(
|
|
|
|
| 331 |
|
| 332 |
if VERBOSE:
|
| 333 |
logging.info(LOG_RESPONSE.format(resp)) # Log the response
|
| 334 |
+
return resp # Return resp here
|
|
|
|
| 335 |
|
| 336 |
def generate_text_chunked(input_text, model, generation_parameters, max_tokens_to_generate):
|
| 337 |
"""Generates text in chunks to avoid token limit errors."""
|