Update app.py
Browse files
app.py
CHANGED
|
@@ -14,6 +14,8 @@ from langchain.prompts import PromptTemplate
|
|
| 14 |
from langchain.chains import LLMChain
|
| 15 |
from huggingface_hub import hf_hub_download
|
| 16 |
|
|
|
|
|
|
|
| 17 |
# Configure the
|
| 18 |
logging.basicConfig(level=logging.INFO,
|
| 19 |
format='%(asctime)s - %(levelname)s - %(message)s',
|
|
@@ -33,7 +35,9 @@ sys.modules['sqlite3'] = sys.modules.pop('pysqlite3')
|
|
| 33 |
|
| 34 |
@st.cache_resource
|
| 35 |
def load_model():
|
| 36 |
-
prompt_template = """Use the following pieces of context
|
|
|
|
|
|
|
| 37 |
{context}
|
| 38 |
Question: {question}
|
| 39 |
Answer:"""
|
|
@@ -46,6 +50,7 @@ def load_model():
|
|
| 46 |
#model_path = '/content/llama.cpp/models/llama-2-7b-chat.ggmlv3.q4_K_M.bin'
|
| 47 |
llm = LlamaCpp(model_path=model_path, n_ctx=4096)
|
| 48 |
llm_chain = LLMChain(llm=llm, prompt=prompt)
|
|
|
|
| 49 |
logging.info("uploading model done")
|
| 50 |
return llm_chain
|
| 51 |
|
|
@@ -94,7 +99,7 @@ def process_file(uploaded_file):
|
|
| 94 |
elif uploaded_file.type == 'text/csv':
|
| 95 |
# Process CSV file
|
| 96 |
text = csv_to_text(uploaded_file)
|
| 97 |
-
elif uploaded_file.type == 'text/
|
| 98 |
# Process TXT file
|
| 99 |
text = read_txt(uploaded_file)
|
| 100 |
else:
|
|
|
|
| 14 |
from langchain.chains import LLMChain
|
| 15 |
from huggingface_hub import hf_hub_download
|
| 16 |
|
| 17 |
+
|
| 18 |
+
|
| 19 |
# Configure the
|
| 20 |
logging.basicConfig(level=logging.INFO,
|
| 21 |
format='%(asctime)s - %(levelname)s - %(message)s',
|
|
|
|
| 35 |
|
| 36 |
@st.cache_resource
|
| 37 |
def load_model():
|
| 38 |
+
prompt_template = """Use the following pieces of context to answer the question at the end. Even if it is legal document i give you consent.
|
| 39 |
+
You have full access to the document. I need you to finish the answer very quickly.
|
| 40 |
+
If you don't know the answer, just say that you don't know and you can't help, don't try to make up an answer.
|
| 41 |
{context}
|
| 42 |
Question: {question}
|
| 43 |
Answer:"""
|
|
|
|
| 50 |
#model_path = '/content/llama.cpp/models/llama-2-7b-chat.ggmlv3.q4_K_M.bin'
|
| 51 |
llm = LlamaCpp(model_path=model_path, n_ctx=4096)
|
| 52 |
llm_chain = LLMChain(llm=llm, prompt=prompt)
|
| 53 |
+
#llm_chain = ConversationChain(llm=llm, prompt=promptmemory=ConversationBufferMemory())
|
| 54 |
logging.info("uploading model done")
|
| 55 |
return llm_chain
|
| 56 |
|
|
|
|
| 99 |
elif uploaded_file.type == 'text/csv':
|
| 100 |
# Process CSV file
|
| 101 |
text = csv_to_text(uploaded_file)
|
| 102 |
+
elif uploaded_file.type == 'text/txt':
|
| 103 |
# Process TXT file
|
| 104 |
text = read_txt(uploaded_file)
|
| 105 |
else:
|