Spaces:
Sleeping
Sleeping
Update src/qa/chain.py
Browse files- src/qa/chain.py +4 -2
src/qa/chain.py
CHANGED
|
@@ -27,12 +27,12 @@ class QAChain:
|
|
| 27 |
self,
|
| 28 |
persist_dir: str = "data/vector_store",
|
| 29 |
model_name: str = "text-embedding-3-small",
|
| 30 |
-
k: int =
|
| 31 |
verbose: bool = False,
|
| 32 |
llm_model: str = "gpt-4o-mini",
|
| 33 |
llm_temperature: float = 0.3,
|
| 34 |
llm_max_tokens: Optional[int] = None,
|
| 35 |
-
max_history_turns: int =
|
| 36 |
max_history_chars: int = 10000
|
| 37 |
):
|
| 38 |
"""
|
|
@@ -251,8 +251,10 @@ class QAChain:
|
|
| 251 |
'source_urls': urls_text
|
| 252 |
}
|
| 253 |
qa_prompt_with_history = ChatPromptTemplate.from_messages([
|
|
|
|
| 254 |
("system", CHARACTER_TEMPLATE),
|
| 255 |
("system", QA_TEMPLATE_WITH_HISTORY),
|
|
|
|
| 256 |
])
|
| 257 |
|
| 258 |
# γγγ³γγγγ³γγ¬γΌγγι©η¨γγ¦εηγηζ
|
|
|
|
| 27 |
self,
|
| 28 |
persist_dir: str = "data/vector_store",
|
| 29 |
model_name: str = "text-embedding-3-small",
|
| 30 |
+
k: int = 10,
|
| 31 |
verbose: bool = False,
|
| 32 |
llm_model: str = "gpt-4o-mini",
|
| 33 |
llm_temperature: float = 0.3,
|
| 34 |
llm_max_tokens: Optional[int] = None,
|
| 35 |
+
max_history_turns: int = 3,
|
| 36 |
max_history_chars: int = 10000
|
| 37 |
):
|
| 38 |
"""
|
|
|
|
| 251 |
'source_urls': urls_text
|
| 252 |
}
|
| 253 |
qa_prompt_with_history = ChatPromptTemplate.from_messages([
|
| 254 |
+
("user", "context: {context}"),
|
| 255 |
("system", CHARACTER_TEMPLATE),
|
| 256 |
("system", QA_TEMPLATE_WITH_HISTORY),
|
| 257 |
+
("user", "θ³ͺε: {question}")
|
| 258 |
])
|
| 259 |
|
| 260 |
# γγγ³γγγγ³γγ¬γΌγγι©η¨γγ¦εηγηζ
|