Update process_data.py
Browse files- process_data.py +15 -13
process_data.py
CHANGED
|
@@ -31,6 +31,8 @@ def generate_json(input_data, parameters):
|
|
| 31 |
farm_prompt = input_data["input_context"] + farm_prompt
|
| 32 |
interactions_prompt = input_data["input_context"] + interactions_prompt
|
| 33 |
trial_prompt = input_data["input_context"] + trial_prompt
|
|
|
|
|
|
|
| 34 |
|
| 35 |
try:
|
| 36 |
#Call OpenAI API to generate structured output based on prompt
|
|
@@ -209,20 +211,20 @@ def pre_processing(input_data, parameters):
|
|
| 209 |
pre_processing_list = [parameters["context_pre_prompt"], parameters["summary_pre_prompt"], parameters["conversation_pre_prompt"], parameters["example_pre_prompt"]]
|
| 210 |
|
| 211 |
for pre_prompt in pre_processing_list:
|
| 212 |
-
|
| 213 |
-
response = client.chat.completions.create(
|
| 214 |
-
model=parameters["model_version"],
|
| 215 |
-
messages=[
|
| 216 |
-
{"role": "system", "content": pre_prompt},
|
| 217 |
-
{"role": "user", "content": input_text}
|
| 218 |
-
]
|
| 219 |
-
)
|
| 220 |
-
|
| 221 |
-
response_text = response['choices'][0]['message']['content']
|
| 222 |
-
|
| 223 |
try:
|
| 224 |
-
|
| 225 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 226 |
|
| 227 |
except Exception as e:
|
| 228 |
print("Failed to parse response as JSON. Error was:")
|
|
|
|
| 31 |
farm_prompt = input_data["input_context"] + farm_prompt
|
| 32 |
interactions_prompt = input_data["input_context"] + interactions_prompt
|
| 33 |
trial_prompt = input_data["input_context"] + trial_prompt
|
| 34 |
+
|
| 35 |
+
|
| 36 |
|
| 37 |
try:
|
| 38 |
#Call OpenAI API to generate structured output based on prompt
|
|
|
|
| 211 |
pre_processing_list = [parameters["context_pre_prompt"], parameters["summary_pre_prompt"], parameters["conversation_pre_prompt"], parameters["example_pre_prompt"]]
|
| 212 |
|
| 213 |
for pre_prompt in pre_processing_list:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 214 |
try:
|
| 215 |
+
|
| 216 |
+
if pre_prompt:
|
| 217 |
+
response = client.chat.completions.create(
|
| 218 |
+
model=parameters["model_version"],
|
| 219 |
+
messages=[
|
| 220 |
+
{"role": "system", "content": pre_prompt},
|
| 221 |
+
{"role": "user", "content": input_text}
|
| 222 |
+
]
|
| 223 |
+
)
|
| 224 |
+
|
| 225 |
+
response_text = response['choices'][0]['message']['content']
|
| 226 |
+
partially_processed_input = json.loads(response_text)
|
| 227 |
+
input_text = partially_processed_input
|
| 228 |
|
| 229 |
except Exception as e:
|
| 230 |
print("Failed to parse response as JSON. Error was:")
|