Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -113,7 +113,7 @@ if "documents" in st.session_state:
|
|
| 113 |
"Identify research gaps",
|
| 114 |
"Suggest research ideas",
|
| 115 |
"Simulate a debate",
|
| 116 |
-
"Translate
|
| 117 |
"Generate citation"
|
| 118 |
])
|
| 119 |
|
|
@@ -134,7 +134,7 @@ if "documents" in st.session_state:
|
|
| 134 |
if task == "Summarize document":
|
| 135 |
chain = create_stuff_documents_chain(llm, summary_prompt)
|
| 136 |
summary = run_chain(chain, {"context": docs})
|
| 137 |
-
|
| 138 |
st.markdown("### π Summary")
|
| 139 |
st.write(summary)
|
| 140 |
|
|
@@ -144,7 +144,7 @@ if "documents" in st.session_state:
|
|
| 144 |
summary = run_chain(chain1, {"context": docs})
|
| 145 |
chain2 = LLMChain(llm=llm, prompt=gap_prompt)
|
| 146 |
gaps = run_chain(chain2, {"summary": summary})
|
| 147 |
-
|
| 148 |
st.markdown("### π Identified Gaps")
|
| 149 |
st.write(gaps)
|
| 150 |
|
|
@@ -156,6 +156,7 @@ if "documents" in st.session_state:
|
|
| 156 |
gaps = run_chain(chain2, {"summary": summary})
|
| 157 |
chain3 = LLMChain(llm=llm, prompt=idea_prompt)
|
| 158 |
ideas = run_chain(chain3, {"gaps": gaps})
|
|
|
|
| 159 |
st.markdown("### π‘ Research Ideas")
|
| 160 |
st.write(ideas)
|
| 161 |
|
|
@@ -165,21 +166,25 @@ if "documents" in st.session_state:
|
|
| 165 |
summary = run_chain(chain, {"context": docs})
|
| 166 |
debate_chain = LLMChain(llm=llm, prompt=debate_prompt)
|
| 167 |
debate = run_chain(debate_chain, {"summary": summary})
|
|
|
|
| 168 |
st.markdown("### π Debate")
|
| 169 |
st.write(debate)
|
| 170 |
|
| 171 |
# Translate agent
|
| 172 |
-
elif task == "Translate
|
| 173 |
-
|
| 174 |
-
|
| 175 |
-
|
| 176 |
-
|
| 177 |
-
|
| 178 |
-
|
|
|
|
|
|
|
| 179 |
|
| 180 |
# Citation agent
|
| 181 |
elif task == "Generate citation":
|
| 182 |
citation_chain = create_stuff_documents_chain(llm, citation_prompt)
|
| 183 |
citation = run_chain(citation_chain, {"context": docs})
|
|
|
|
| 184 |
st.markdown("### π APA Citation")
|
| 185 |
st.code(citation, language="markdown")
|
|
|
|
| 113 |
"Identify research gaps",
|
| 114 |
"Suggest research ideas",
|
| 115 |
"Simulate a debate",
|
| 116 |
+
"Translate response",
|
| 117 |
"Generate citation"
|
| 118 |
])
|
| 119 |
|
|
|
|
| 134 |
if task == "Summarize document":
|
| 135 |
chain = create_stuff_documents_chain(llm, summary_prompt)
|
| 136 |
summary = run_chain(chain, {"context": docs})
|
| 137 |
+
st.session_state["last_agent_output"] = summary
|
| 138 |
st.markdown("### π Summary")
|
| 139 |
st.write(summary)
|
| 140 |
|
|
|
|
| 144 |
summary = run_chain(chain1, {"context": docs})
|
| 145 |
chain2 = LLMChain(llm=llm, prompt=gap_prompt)
|
| 146 |
gaps = run_chain(chain2, {"summary": summary})
|
| 147 |
+
st.session_state["last_agent_output"] = gaps
|
| 148 |
st.markdown("### π Identified Gaps")
|
| 149 |
st.write(gaps)
|
| 150 |
|
|
|
|
| 156 |
gaps = run_chain(chain2, {"summary": summary})
|
| 157 |
chain3 = LLMChain(llm=llm, prompt=idea_prompt)
|
| 158 |
ideas = run_chain(chain3, {"gaps": gaps})
|
| 159 |
+
st.session_state["last_agent_output"] = ideas
|
| 160 |
st.markdown("### π‘ Research Ideas")
|
| 161 |
st.write(ideas)
|
| 162 |
|
|
|
|
| 166 |
summary = run_chain(chain, {"context": docs})
|
| 167 |
debate_chain = LLMChain(llm=llm, prompt=debate_prompt)
|
| 168 |
debate = run_chain(debate_chain, {"summary": summary})
|
| 169 |
+
st.session_state["last_agent_output"] = debate
|
| 170 |
st.markdown("### π Debate")
|
| 171 |
st.write(debate)
|
| 172 |
|
| 173 |
# Translate agent
|
| 174 |
+
elif task == "Translate response":
|
| 175 |
+
if "last_agent_output" in st.session_state:
|
| 176 |
+
to_translate = st.session_state["last_agent_output"]
|
| 177 |
+
translate_chain = LLMChain(llm=llm, prompt=translate_prompt)
|
| 178 |
+
translated = translate_chain.invoke({"language": user_language, "content": to_translate})
|
| 179 |
+
st.markdown(f"### π Translated Response ({user_language})")
|
| 180 |
+
st.write(translated)
|
| 181 |
+
else:
|
| 182 |
+
st.warning("Please run another agent first (e.g., Summary, Gap, Debate) to generate something to translate.")
|
| 183 |
|
| 184 |
# Citation agent
|
| 185 |
elif task == "Generate citation":
|
| 186 |
citation_chain = create_stuff_documents_chain(llm, citation_prompt)
|
| 187 |
citation = run_chain(citation_chain, {"context": docs})
|
| 188 |
+
st.session_state["last_agent_output"] = citation
|
| 189 |
st.markdown("### π APA Citation")
|
| 190 |
st.code(citation, language="markdown")
|