Update app.py
Browse files
app.py
CHANGED
|
@@ -80,22 +80,7 @@ import os
|
|
| 80 |
from langchain_core.runnables import RunnableLambda
|
| 81 |
|
| 82 |
|
| 83 |
-
|
| 84 |
-
You are an expert in League of Legends (LoL) lore. You will only answer questions related to the champions and their stories within the game.
|
| 85 |
-
|
| 86 |
-
Instructions:
|
| 87 |
-
1. **Only use the context provided below** to answer the question. You should reference the context directly to ensure your answer is as relevant as possible.
|
| 88 |
-
2. If the question is outside the scope of League of Legends lore, respond by saying: *"Please ask something related to League of Legends lore."*
|
| 89 |
-
3. If the provided context does not provide a clear answer or you're unsure, respond by saying: *"I'm unsure based on the provided context."*
|
| 90 |
-
|
| 91 |
-
Context: {context}
|
| 92 |
-
|
| 93 |
-
Question: {question}
|
| 94 |
-
|
| 95 |
-
Answer:
|
| 96 |
-
|
| 97 |
-
"""
|
| 98 |
-
prompt_template = ChatPromptTemplate.from_template(prompt)
|
| 99 |
llm = HuggingFaceHub(
|
| 100 |
repo_id="HuggingFaceH4/zephyr-7b-beta",
|
| 101 |
model_kwargs={"temperature": 0.1, "max_length": 50, "return_full_text" : False}
|
|
|
|
| 80 |
from langchain_core.runnables import RunnableLambda
|
| 81 |
|
| 82 |
|
| 83 |
+
#prompt_template = ChatPromptTemplate.from_template(prompt)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 84 |
llm = HuggingFaceHub(
|
| 85 |
repo_id="HuggingFaceH4/zephyr-7b-beta",
|
| 86 |
model_kwargs={"temperature": 0.1, "max_length": 50, "return_full_text" : False}
|