Update app.py
Browse files
app.py
CHANGED
|
@@ -16,19 +16,24 @@ bg = st.text_area("Background information",
|
|
| 16 |
|
| 17 |
if 'ft' in st.session_state or st.button('Next step'):
|
| 18 |
if 'ft' not in st.session_state:
|
|
|
|
| 19 |
messages = [
|
| 20 |
{
|
| 21 |
"role": "user",
|
| 22 |
"content": f"""Generate possible factors related to the goal
|
| 23 |
-
based on the information provided. List the factors in markdown list
|
|
|
|
|
|
|
| 24 |
goals:\n\n{gl}\n\n
|
| 25 |
-
background:\n\n{bg}\n\n
|
|
|
|
|
|
|
| 26 |
}
|
| 27 |
]
|
| 28 |
|
| 29 |
chat_completion = client.chat.completions.create(
|
| 30 |
messages=messages,
|
| 31 |
-
model="llama3-
|
| 32 |
)
|
| 33 |
|
| 34 |
ft = st.text_area("Factors to be considered in questions",
|
|
@@ -36,7 +41,28 @@ if 'ft' in st.session_state or st.button('Next step'):
|
|
| 36 |
st.session_state['ft'] = ft
|
| 37 |
st.write('Edit above factors, add or remove if needed')
|
| 38 |
|
|
|
|
|
|
|
| 39 |
if st.button('Generate questions'):
|
| 40 |
-
|
| 41 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
# dump questions into excel and allow download
|
|
|
|
| 16 |
|
| 17 |
if 'ft' in st.session_state or st.button('Next step'):
|
| 18 |
if 'ft' not in st.session_state:
|
| 19 |
+
rf = os.environ.get("DESIGN_PROMPT")
|
| 20 |
messages = [
|
| 21 |
{
|
| 22 |
"role": "user",
|
| 23 |
"content": f"""Generate possible factors related to the goal
|
| 24 |
+
based on the information provided. List the factors in markdown list.
|
| 25 |
+
You can give advice to human to edit their goal if you cannot define
|
| 26 |
+
more than 3 factors.\n\n
|
| 27 |
goals:\n\n{gl}\n\n
|
| 28 |
+
background:\n\n{bg}\n\n
|
| 29 |
+
some design principles for your reference:\n\n
|
| 30 |
+
reference:\n\n```markdown\n{rf}\n\n```\n\n""",
|
| 31 |
}
|
| 32 |
]
|
| 33 |
|
| 34 |
chat_completion = client.chat.completions.create(
|
| 35 |
messages=messages,
|
| 36 |
+
model="llama3-70b-8192",
|
| 37 |
)
|
| 38 |
|
| 39 |
ft = st.text_area("Factors to be considered in questions",
|
|
|
|
| 41 |
st.session_state['ft'] = ft
|
| 42 |
st.write('Edit above factors, add or remove if needed')
|
| 43 |
|
| 44 |
+
rf = os.environ.get("DESIGN_PROMPT")
|
| 45 |
+
|
| 46 |
if st.button('Generate questions'):
|
| 47 |
+
rf += f'''\n\n---\n\n{os.environ.get("DESIGN_PROMPT_SIXPOINT")}'''
|
| 48 |
+
messages = [
|
| 49 |
+
{
|
| 50 |
+
"role": "user",
|
| 51 |
+
"content": f"""Extract all factors from the text to create context.
|
| 52 |
+
Generate 5 question based on the text and context.
|
| 53 |
+
Write the questionnaire form in markdown\n\n
|
| 54 |
+
text:\n\n{ft}\n\n
|
| 55 |
+
some design principles for your reference:\n\n
|
| 56 |
+
reference:\n\n```markdown\n{rf}\n```\n\n""",
|
| 57 |
+
}
|
| 58 |
+
]
|
| 59 |
+
|
| 60 |
+
chat_completion = client.chat.completions.create(
|
| 61 |
+
messages=messages,
|
| 62 |
+
model="llama3-70b-8192",
|
| 63 |
+
)
|
| 64 |
+
|
| 65 |
+
md = chat_completion.choices[0].message.content
|
| 66 |
+
suggested_sampling = 5 * len(md.split("\n"))
|
| 67 |
+
st.markdown(f"{md}\n\n**suggested number of sampling**: {suggested_sampling}")
|
| 68 |
# dump questions into excel and allow download
|