Update process_data.py
Browse files- process_data.py +9 -24
process_data.py
CHANGED
|
@@ -209,6 +209,15 @@ def pre_processing(input_data, parameters):
|
|
| 209 |
|
| 210 |
response_text = response.choices[0].message.content
|
| 211 |
partially_processed_input = json.loads(response_text)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 212 |
input_text = partially_processed_input
|
| 213 |
|
| 214 |
except Exception as e:
|
|
@@ -239,30 +248,6 @@ def parse_survey_stack_parameters(data):
|
|
| 239 |
|
| 240 |
processed_data["model_version"] = data[0]['data']['modelversion']['value'][0]
|
| 241 |
|
| 242 |
-
# this is a comment
|
| 243 |
-
#if data['group_2']['preprompt']['value'][0] == 'continue_preprompts':
|
| 244 |
-
# processed_data["pre_prompt"] = True
|
| 245 |
-
# processed_data["context_pre_prompt"] = data[0]['data']['contextpreprompt']['value']
|
| 246 |
-
# processed_data["summary_pre_prompt"] = data[0]['data']['summarypreprompt']['value']
|
| 247 |
-
# processed_data["conversation_pre_prompt"] = data[0]['data']['conversationpreprompt']['value']
|
| 248 |
-
# processed_data["example_pre_prompt"] = data[0]['data']['examplepreprompt']['value']
|
| 249 |
-
# processed_data["chaining"] = data[0]['data']['prepromptchaining']['value'][0]
|
| 250 |
-
# if processed_data["chaining"] == "no":
|
| 251 |
-
# combined_prompt = " ".join(filter(None, [processed_data["context_pre_prompt"], processed_data["summary_pre_prompt"], processed_data["conversation_pre_prompt"], processed_data["example_pre_prompt"]]))
|
| 252 |
-
# processed_data["combined_prompt"] = combined_prompt
|
| 253 |
-
# processed_data["chaining"] = False
|
| 254 |
-
# else:
|
| 255 |
-
# processed_data["chaining"] = True
|
| 256 |
-
# processed_data["combined_pre_prompt"] = None
|
| 257 |
-
#else:
|
| 258 |
-
# processed_data["pre_prompt"] = False
|
| 259 |
-
# processed_data["context_pre_prompt"] = None
|
| 260 |
-
# processed_data["summary_pre_prompt"] = None
|
| 261 |
-
# processed_data["conversation_pre_prompt"] = None
|
| 262 |
-
# processed_data["example_pre_prompt"] = None
|
| 263 |
-
# processed_data["chaining"] = None
|
| 264 |
-
# processed_data["combined_pre_prompt"] = None
|
| 265 |
-
|
| 266 |
print("DATA: ")
|
| 267 |
print(data)
|
| 268 |
|
|
|
|
| 209 |
|
| 210 |
response_text = response.choices[0].message.content
|
| 211 |
partially_processed_input = json.loads(response_text)
|
| 212 |
+
|
| 213 |
+
print("Pre-Processing: ")
|
| 214 |
+
print("Prompt: ")
|
| 215 |
+
print(pre_prompt)
|
| 216 |
+
print("Input Text: ")
|
| 217 |
+
print(input_text)
|
| 218 |
+
print("Result: ")
|
| 219 |
+
print(partially_processed_input)
|
| 220 |
+
|
| 221 |
input_text = partially_processed_input
|
| 222 |
|
| 223 |
except Exception as e:
|
|
|
|
| 248 |
|
| 249 |
processed_data["model_version"] = data[0]['data']['modelversion']['value'][0]
|
| 250 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 251 |
print("DATA: ")
|
| 252 |
print(data)
|
| 253 |
|