Update process_data.py
Browse files- process_data.py +19 -19
process_data.py
CHANGED
|
@@ -100,9 +100,9 @@ def generate_json(processed_data):
|
|
| 100 |
input_text = processed_data["input_text"]
|
| 101 |
model_version = processed_data["parameters"]["model_version"]
|
| 102 |
|
| 103 |
-
farm_prompt = processed_data["prompts"]["
|
| 104 |
-
interactions_prompt = processed_data["prompts"]["
|
| 105 |
-
trial_prompt = processed_data["prompts"]["
|
| 106 |
|
| 107 |
|
| 108 |
try:
|
|
@@ -476,13 +476,13 @@ def pre_processing(processed_data):
|
|
| 476 |
print("Starting preprocessing")
|
| 477 |
|
| 478 |
|
| 479 |
-
pre_processing_list = [processed_data.get('parameters', {}).get('
|
| 480 |
|
| 481 |
print("Preprocessing list")
|
| 482 |
print(pre_processing_list)
|
| 483 |
|
| 484 |
print("Model Version")
|
| 485 |
-
model_version = processed_data["parameters"]["
|
| 486 |
print(model_version)
|
| 487 |
|
| 488 |
if processed_data["inputstyle"] == "individual-pieces-input-text":
|
|
@@ -669,9 +669,9 @@ def parse_survey_stack(data):
|
|
| 669 |
print(processed_data["input_text"])
|
| 670 |
|
| 671 |
print("NOW ONTO THE PROMPTS from Input Item")
|
| 672 |
-
processed_data["prompts"]["
|
| 673 |
-
processed_data["prompts"]["
|
| 674 |
-
processed_data["prompts"]["
|
| 675 |
|
| 676 |
|
| 677 |
# parameters
|
|
@@ -679,17 +679,17 @@ def parse_survey_stack(data):
|
|
| 679 |
print("PARAMETERS" + str(parameter_data))
|
| 680 |
|
| 681 |
processed_data["parameters"] = {}
|
| 682 |
-
processed_data["parameters"]["
|
| 683 |
-
processed_data["parameters"]["
|
| 684 |
-
processed_data["parameters"]["
|
| 685 |
-
|
| 686 |
-
if processed_data["parameters"]["
|
| 687 |
-
processed_data["parameters"]["
|
| 688 |
-
processed_data["parameters"]["
|
| 689 |
-
processed_data["parameters"]["
|
| 690 |
-
processed_data["parameters"]["
|
| 691 |
-
processed_data["parameters"]["
|
| 692 |
-
processed_data["parameters"]["
|
| 693 |
|
| 694 |
print("RETURNING DATA")
|
| 695 |
print(processed_data)
|
|
|
|
| 100 |
input_text = processed_data["input_text"]
|
| 101 |
model_version = processed_data["parameters"]["model_version"]
|
| 102 |
|
| 103 |
+
farm_prompt = processed_data["prompts"]["firstschemaprompt"]
|
| 104 |
+
interactions_prompt = processed_data["prompts"]["secondschemaprompt"]
|
| 105 |
+
trial_prompt = processed_data["prompts"]["thirdschemaprompt"]
|
| 106 |
|
| 107 |
|
| 108 |
try:
|
|
|
|
| 476 |
print("Starting preprocessing")
|
| 477 |
|
| 478 |
|
| 479 |
+
pre_processing_list = [processed_data.get('parameters', {}).get('preprocessingprompt1', None), processed_data.get('parameters', {}).get('preprocessingprompt2', None), processed_data.get('parameters', {}).get('preprocessingprompt3', None)]
|
| 480 |
|
| 481 |
print("Preprocessing list")
|
| 482 |
print(pre_processing_list)
|
| 483 |
|
| 484 |
print("Model Version")
|
| 485 |
+
model_version = processed_data["parameters"]["preprocessmodelversion"]
|
| 486 |
print(model_version)
|
| 487 |
|
| 488 |
if processed_data["inputstyle"] == "individual-pieces-input-text":
|
|
|
|
| 669 |
print(processed_data["input_text"])
|
| 670 |
|
| 671 |
print("NOW ONTO THE PROMPTS from Input Item")
|
| 672 |
+
processed_data["prompts"]["firstschemaprompt"] = input_item.get('firstschemaprompt', {}).get('value')
|
| 673 |
+
processed_data["prompts"]["secondschemaprompt"] = input_item.get('secondschemaprompt', {}).get('value')
|
| 674 |
+
processed_data["prompts"]["thirdschemaprompt"] = input_item.get('thirdschemaprompt', {}).get('value')
|
| 675 |
|
| 676 |
|
| 677 |
# parameters
|
|
|
|
| 679 |
print("PARAMETERS" + str(parameter_data))
|
| 680 |
|
| 681 |
processed_data["parameters"] = {}
|
| 682 |
+
processed_data["parameters"]["modelversion"] = parameter_data.get('modelversion', {}).get('value')[0]
|
| 683 |
+
processed_data["parameters"]["preprocessdata"] = parameter_data.get('preprocessdata', {}).get('value')[0]
|
| 684 |
+
processed_data["parameters"]["promptstyle"] = parameter_data.get('promptstyle', {}).get('value')
|
| 685 |
+
|
| 686 |
+
if processed_data["parameters"]["preprocessdata"] == "yes":
|
| 687 |
+
processed_data["parameters"]["preprocessmodelversion"] = parameter_data.get('preprocessmodelversion', {}).get('value')[0]
|
| 688 |
+
processed_data["parameters"]["multiplepreprompts"] = parameter_data.get('multiplepreprompts', {}).get('value')
|
| 689 |
+
processed_data["parameters"]["prepromptstyle"] = parameter_data.get('prepromptstyle', {}).get('value')
|
| 690 |
+
processed_data["parameters"]["preprocessingprompt1"] = parameter_data.get('preprocessingprompt1', {}).get('value')
|
| 691 |
+
processed_data["parameters"]["preprocessingprompt2"] = parameter_data.get('preprocessingprompt2', {}).get('value')
|
| 692 |
+
processed_data["parameters"]["preprocessingprompt3"] = parameter_data.get('preprocessingprompt3', {}).get('value')
|
| 693 |
|
| 694 |
print("RETURNING DATA")
|
| 695 |
print(processed_data)
|