Spaces:
Sleeping
Sleeping
Update src/rag_system.py
Browse files- src/rag_system.py +4 -3
src/rag_system.py
CHANGED
|
@@ -70,7 +70,7 @@ C:{context}
|
|
| 70 |
A:"""
|
| 71 |
|
| 72 |
message = HumanMessage(content=prompt)
|
| 73 |
-
response = self.llm([message])
|
| 74 |
answer = response.content
|
| 75 |
|
| 76 |
# Cache response
|
|
@@ -96,7 +96,7 @@ A:"""
|
|
| 96 |
{text}"""
|
| 97 |
|
| 98 |
message = HumanMessage(content=prompt)
|
| 99 |
-
response = self.llm([message])
|
| 100 |
summary = response.content
|
| 101 |
|
| 102 |
# Cache
|
|
@@ -120,7 +120,8 @@ A:"""
|
|
| 120 |
qa_prompt += f"Context:{context}\nAnswers:"
|
| 121 |
|
| 122 |
message = HumanMessage(content=qa_prompt)
|
| 123 |
-
response = self.llm([message])
|
|
|
|
| 124 |
|
| 125 |
# Parse responses
|
| 126 |
answers = response.content.split('\n')
|
|
|
|
| 70 |
A:"""
|
| 71 |
|
| 72 |
message = HumanMessage(content=prompt)
|
| 73 |
+
response = self.llm.invoke([message])
|
| 74 |
answer = response.content
|
| 75 |
|
| 76 |
# Cache response
|
|
|
|
| 96 |
{text}"""
|
| 97 |
|
| 98 |
message = HumanMessage(content=prompt)
|
| 99 |
+
response = self.llm.invoke([message])
|
| 100 |
summary = response.content
|
| 101 |
|
| 102 |
# Cache
|
|
|
|
| 120 |
qa_prompt += f"Context:{context}\nAnswers:"
|
| 121 |
|
| 122 |
message = HumanMessage(content=qa_prompt)
|
| 123 |
+
response = self.llm.invoke([message])
|
| 124 |
+
|
| 125 |
|
| 126 |
# Parse responses
|
| 127 |
answers = response.content.split('\n')
|