Spaces:
Runtime error
Runtime error
Commit ·
75c6673
1
Parent(s): 2e480c2
fixed lama function
Browse files- .gitignore +0 -1
- main.py +7 -7
- models/llama-2-7b-chat.ggmlv3.q4_0.bin +3 -0
.gitignore
CHANGED
|
@@ -3,7 +3,6 @@ venv/*
|
|
| 3 |
.env
|
| 4 |
|
| 5 |
#LLMs
|
| 6 |
-
models/*
|
| 7 |
|
| 8 |
#Database
|
| 9 |
db/*
|
|
|
|
| 3 |
.env
|
| 4 |
|
| 5 |
#LLMs
|
|
|
|
| 6 |
|
| 7 |
#Database
|
| 8 |
db/*
|
main.py
CHANGED
|
@@ -81,7 +81,7 @@ class VectorDatabaseInitializer:
|
|
| 81 |
text_splitter = CharacterTextSplitter(chunk_size=500, chunk_overlap=50)
|
| 82 |
text_chunks = text_splitter.split_documents(raw_documents)
|
| 83 |
|
| 84 |
-
if selected_embeddings == "
|
| 85 |
embeddings = OpenAIEmbeddings()
|
| 86 |
else:
|
| 87 |
embeddings = HuggingFaceEmbeddings(model_name='sentence-transformers/all-MiniLM-L6-v2', model_kwargs={'device': 'cpu'})
|
|
@@ -94,7 +94,7 @@ class VectorDatabaseInitializer:
|
|
| 94 |
class LLMHandler:
|
| 95 |
@staticmethod
|
| 96 |
def initialize_llm(llm_type, temperature, max_tokens):
|
| 97 |
-
if llm_type == "
|
| 98 |
llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=temperature, max_tokens=max_tokens)
|
| 99 |
else:
|
| 100 |
llm = CTransformers(model="models/llama-2-7b-chat.ggmlv3.q4_0.bin",
|
|
@@ -113,7 +113,7 @@ def handle_user_selection(upload_document, user_query, selected_llms, temperatur
|
|
| 113 |
timestamp = datetime.datetime.now().strftime("%Y%m%d%H%M%S")
|
| 114 |
|
| 115 |
vector_db_initializer = VectorDatabaseInitializer()
|
| 116 |
-
vectordb = vector_db_initializer.initialize_vector_database(raw_documents, selected_embeddings, f'db_{timestamp}')
|
| 117 |
|
| 118 |
llm_handler = LLMHandler()
|
| 119 |
llm = llm_handler.initialize_llm(selected_llms, temperature, max_tokens)
|
|
@@ -169,11 +169,11 @@ with gr.Blocks(theme=gr.Theme.from_hub('finlaymacklon/smooth_slate'), title="LLM
|
|
| 169 |
query = gr.Textbox(label="Query", placeholder="Type a question and Enter", lines=1)
|
| 170 |
with gr.Column(scale=0.1):
|
| 171 |
template_input = gr.Textbox(label="Custom Template (optional)", placeholder="Enter your template or leave blank for default", lines=3, scale=1)
|
| 172 |
-
selected_llm = gr.Radio(["
|
| 173 |
temperature = gr.Slider(0.0,1.0, label="temperature", value=0.8)
|
| 174 |
-
max_tokens = gr.Slider(50, 1500, label="Max_Tokens", value =
|
| 175 |
-
selected_embeddings = gr.Radio(["
|
| 176 |
-
selected_vector_db = gr.Radio(["Chroma"], label="Select Vector Databases")
|
| 177 |
clear_vector_button = gr.Button("Clear Vector store")
|
| 178 |
clear_vector_output = gr.Textbox(label = "Vector store output")
|
| 179 |
clear_history_button = gr.Button("Clear history")
|
|
|
|
| 81 |
text_splitter = CharacterTextSplitter(chunk_size=500, chunk_overlap=50)
|
| 82 |
text_chunks = text_splitter.split_documents(raw_documents)
|
| 83 |
|
| 84 |
+
if selected_embeddings == "OpenAI":
|
| 85 |
embeddings = OpenAIEmbeddings()
|
| 86 |
else:
|
| 87 |
embeddings = HuggingFaceEmbeddings(model_name='sentence-transformers/all-MiniLM-L6-v2', model_kwargs={'device': 'cpu'})
|
|
|
|
| 94 |
class LLMHandler:
|
| 95 |
@staticmethod
|
| 96 |
def initialize_llm(llm_type, temperature, max_tokens):
|
| 97 |
+
if llm_type == "gpt-3.5-turbo":
|
| 98 |
llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=temperature, max_tokens=max_tokens)
|
| 99 |
else:
|
| 100 |
llm = CTransformers(model="models/llama-2-7b-chat.ggmlv3.q4_0.bin",
|
|
|
|
| 113 |
timestamp = datetime.datetime.now().strftime("%Y%m%d%H%M%S")
|
| 114 |
|
| 115 |
vector_db_initializer = VectorDatabaseInitializer()
|
| 116 |
+
vectordb = vector_db_initializer.initialize_vector_database(raw_documents, selected_embeddings, f'VectorDB/db_{timestamp}')
|
| 117 |
|
| 118 |
llm_handler = LLMHandler()
|
| 119 |
llm = llm_handler.initialize_llm(selected_llms, temperature, max_tokens)
|
|
|
|
| 169 |
query = gr.Textbox(label="Query", placeholder="Type a question and Enter", lines=1)
|
| 170 |
with gr.Column(scale=0.1):
|
| 171 |
template_input = gr.Textbox(label="Custom Template (optional)", placeholder="Enter your template or leave blank for default", lines=3, scale=1)
|
| 172 |
+
selected_llm = gr.Radio(choices = ["gpt-3.5-turbo", "llama-2-7b-chat"], label="Select LLMs",value = 'gpt-3.5-turbo')
|
| 173 |
temperature = gr.Slider(0.0,1.0, label="temperature", value=0.8)
|
| 174 |
+
max_tokens = gr.Slider(50, 1500, label="Max_Tokens", value = 100)
|
| 175 |
+
selected_embeddings = gr.Radio(choices = ["OpenAI", "MiniLM-L6-v2"], label="Select Embeddings",value = 'OpenAI')
|
| 176 |
+
selected_vector_db = gr.Radio(choices = ["Chroma"], label="Select Vector Databases",value = 'Chroma')
|
| 177 |
clear_vector_button = gr.Button("Clear Vector store")
|
| 178 |
clear_vector_output = gr.Textbox(label = "Vector store output")
|
| 179 |
clear_history_button = gr.Button("Clear history")
|
models/llama-2-7b-chat.ggmlv3.q4_0.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8daa9615cce30c259a9555b1cc250d461d1bc69980a274b44d7eda0be78076d8
|
| 3 |
+
size 3791725184
|