Update utils.py
Browse files
utils.py
CHANGED
|
@@ -400,17 +400,17 @@ def rag_chain(llm, prompt, retriever):
|
|
| 400 |
outputs = modell_rag.generate(inputs['input_ids'], max_length=150, num_beams=2, early_stopping=True)
|
| 401 |
answer = tokenizer_rag.decode(outputs[0], skip_special_tokens=True)
|
| 402 |
"""
|
| 403 |
-
|
| 404 |
-
|
| 405 |
-
|
| 406 |
-
#
|
|
|
|
| 407 |
|
| 408 |
# Erstelle das Ergebnis-Dictionary
|
| 409 |
result = {
|
| 410 |
"answer": answer,
|
| 411 |
"relevant_docs": most_relevant_docs
|
| 412 |
}
|
| 413 |
-
|
| 414 |
else:
|
| 415 |
# keine relevanten Dokumente gefunden
|
| 416 |
result = {
|
|
|
|
| 400 |
outputs = modell_rag.generate(inputs['input_ids'], max_length=150, num_beams=2, early_stopping=True)
|
| 401 |
answer = tokenizer_rag.decode(outputs[0], skip_special_tokens=True)
|
| 402 |
"""
|
| 403 |
+
#Alternative, wenn llm direkt übergeben....................................
|
| 404 |
+
#llm_chain = LLMChain(llm = llm, prompt = RAG_CHAIN_PROMPT)
|
| 405 |
+
#answer = llm_chain.run({"context": combined_content, "question": prompt})
|
| 406 |
+
#Alternative, wenn mit API_URL ...........................................
|
| 407 |
+
answer = query(llm, {"inputs": input_text,})
|
| 408 |
|
| 409 |
# Erstelle das Ergebnis-Dictionary
|
| 410 |
result = {
|
| 411 |
"answer": answer,
|
| 412 |
"relevant_docs": most_relevant_docs
|
| 413 |
}
|
|
|
|
| 414 |
else:
|
| 415 |
# keine relevanten Dokumente gefunden
|
| 416 |
result = {
|