Spaces:
Sleeping
Sleeping
tweak to prompt
Browse files
appStore/__pycache__/rag.cpython-310.pyc
CHANGED
|
Binary files a/appStore/__pycache__/rag.cpython-310.pyc and b/appStore/__pycache__/rag.cpython-310.pyc differ
|
|
|
appStore/__pycache__/target.cpython-310.pyc
CHANGED
|
Binary files a/appStore/__pycache__/target.cpython-310.pyc and b/appStore/__pycache__/target.cpython-310.pyc differ
|
|
|
appStore/rag.py
CHANGED
|
@@ -14,9 +14,9 @@ model_select = "gpt-3.5-turbo-1106"
|
|
| 14 |
|
| 15 |
|
| 16 |
# define a special function for putting the prompt together (as we can't use haystack)
|
| 17 |
-
def get_prompt(context):
|
| 18 |
base_prompt="Summarize the following context efficiently in bullet points, the less the better. \
|
| 19 |
-
Summarize only activities that address the vulnerability of
|
| 20 |
Formatting example: \
|
| 21 |
- Collect and utilize gender-disaggregated data to inform and improve climate change adaptation efforts. \
|
| 22 |
- Prioritize gender sensitivity in adaptation options, ensuring participation and benefits for women, who are more vulnerable to climate impacts. \
|
|
@@ -53,9 +53,7 @@ def completion_with_backoff(**kwargs):
|
|
| 53 |
|
| 54 |
|
| 55 |
# construct RAG query, send to openai and process response
|
| 56 |
-
def run_query(
|
| 57 |
-
docs = df
|
| 58 |
-
|
| 59 |
'''
|
| 60 |
For non-streamed completion, enable the following 2 lines and comment out the code below
|
| 61 |
'''
|
|
@@ -63,7 +61,7 @@ def run_query(df):
|
|
| 63 |
# result = res.choices[0].message.content
|
| 64 |
|
| 65 |
# instantiate ChatCompletion as a generator object (stream is set to True)
|
| 66 |
-
response = completion_with_backoff(model=model_select, messages=[{"role": "user", "content": get_prompt(
|
| 67 |
# iterate through the streamed output
|
| 68 |
report = []
|
| 69 |
res_box = st.empty()
|
|
|
|
| 14 |
|
| 15 |
|
| 16 |
# define a special function for putting the prompt together (as we can't use haystack)
|
| 17 |
+
def get_prompt(context, label):
|
| 18 |
base_prompt="Summarize the following context efficiently in bullet points, the less the better. \
|
| 19 |
+
Summarize only activities that address the vulnerability of "+label+" to climate change. \
|
| 20 |
Formatting example: \
|
| 21 |
- Collect and utilize gender-disaggregated data to inform and improve climate change adaptation efforts. \
|
| 22 |
- Prioritize gender sensitivity in adaptation options, ensuring participation and benefits for women, who are more vulnerable to climate impacts. \
|
|
|
|
| 53 |
|
| 54 |
|
| 55 |
# construct RAG query, send to openai and process response
|
| 56 |
+
def run_query(context, label):
|
|
|
|
|
|
|
| 57 |
'''
|
| 58 |
For non-streamed completion, enable the following 2 lines and comment out the code below
|
| 59 |
'''
|
|
|
|
| 61 |
# result = res.choices[0].message.content
|
| 62 |
|
| 63 |
# instantiate ChatCompletion as a generator object (stream is set to True)
|
| 64 |
+
response = completion_with_backoff(model=model_select, messages=[{"role": "user", "content": get_prompt(context, label)}], stream=True)
|
| 65 |
# iterate through the streamed output
|
| 66 |
report = []
|
| 67 |
res_box = st.empty()
|
appStore/target.py
CHANGED
|
@@ -102,7 +102,7 @@ def target_display():
|
|
| 102 |
# construct RAG query for each label, send to openai and process response
|
| 103 |
for i in range(0,len(df_agg)):
|
| 104 |
st.write(df_agg['Vulnerability Label'].iloc[i])
|
| 105 |
-
run_query(df_agg['text'].iloc[i])
|
| 106 |
# st.write(df_agg['text'].iloc[i])
|
| 107 |
|
| 108 |
|
|
|
|
| 102 |
# construct RAG query for each label, send to openai and process response
|
| 103 |
for i in range(0,len(df_agg)):
|
| 104 |
st.write(df_agg['Vulnerability Label'].iloc[i])
|
| 105 |
+
run_query(context = df_agg['text'].iloc[i], label = df_agg['Vulnerability Label'].iloc[i])
|
| 106 |
# st.write(df_agg['text'].iloc[i])
|
| 107 |
|
| 108 |
|