some changes
Browse files
app.py
CHANGED
|
@@ -11,6 +11,7 @@ from langchain.text_splitter import RecursiveCharacterTextSplitter
|
|
| 11 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 12 |
|
| 13 |
|
|
|
|
| 14 |
persist_directory = 'db'
|
| 15 |
embeddings_model_name = 'sentence-transformers/all-MiniLM-L6-v2'
|
| 16 |
|
|
@@ -34,6 +35,8 @@ def get_vector_store(target_source_chunks):
|
|
| 34 |
return retriver
|
| 35 |
|
| 36 |
def get_conversation_chain(retriever):
|
|
|
|
|
|
|
| 37 |
memory = ConversationBufferMemory(memory_key='chat_history', return_messages=True,)
|
| 38 |
chain = RetrievalQA.from_llm(
|
| 39 |
llm=model,
|
|
|
|
| 11 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 12 |
|
| 13 |
|
| 14 |
+
# Updated Prompt Template
|
| 15 |
persist_directory = 'db'
|
| 16 |
embeddings_model_name = 'sentence-transformers/all-MiniLM-L6-v2'
|
| 17 |
|
|
|
|
| 35 |
return retriver
|
| 36 |
|
| 37 |
def get_conversation_chain(retriever):
|
| 38 |
+
tokenizer = AutoTokenizer.from_pretrained("red1xe/Llama-2-7B-codeGPT")
|
| 39 |
+
model = AutoModelForCausalLM.from_pretrained("red1xe/Llama-2-7B-codeGPT")
|
| 40 |
memory = ConversationBufferMemory(memory_key='chat_history', return_messages=True,)
|
| 41 |
chain = RetrievalQA.from_llm(
|
| 42 |
llm=model,
|