Spaces:
Sleeping
Sleeping
Commit ·
8ed84d6
1
Parent(s): a7ad8f5
Updated app.py
Browse files
app.py
CHANGED
|
@@ -6,11 +6,11 @@ import gradio as gr
|
|
| 6 |
|
| 7 |
# imports for langchain, plotly and Chroma
|
| 8 |
|
| 9 |
-
from
|
|
|
|
| 10 |
from langchain.text_splitter import CharacterTextSplitter
|
| 11 |
from langchain.schema import Document
|
| 12 |
from langchain_openai import OpenAIEmbeddings, ChatOpenAI
|
| 13 |
-
from langchain.embeddings import HuggingFaceEmbeddings
|
| 14 |
from langchain_chroma import Chroma
|
| 15 |
from langchain.memory import ConversationBufferMemory
|
| 16 |
from langchain.chains import ConversationalRetrievalChain
|
|
@@ -23,16 +23,8 @@ import matplotlib.pyplot as plt
|
|
| 23 |
from random import randint
|
| 24 |
import shutil
|
| 25 |
|
| 26 |
-
MODEL = "gpt-4o-mini"
|
| 27 |
db_name = "vector_db"
|
| 28 |
-
|
| 29 |
-
# Load environment variables in a file called .env
|
| 30 |
-
|
| 31 |
-
load_dotenv(override=True)
|
| 32 |
-
os.environ['OPENAI_API_KEY'] = os.getenv('OPENAI_API_KEY')
|
| 33 |
-
|
| 34 |
folder = "my-knowledge-base/"
|
| 35 |
-
db_name = "vectorstore_db"
|
| 36 |
MODEL = "llama3.2:latest"
|
| 37 |
|
| 38 |
def process_files(files):
|
|
|
|
| 6 |
|
| 7 |
# imports for langchain, plotly and Chroma
|
| 8 |
|
| 9 |
+
from langchain_community.document_loaders import DirectoryLoader, TextLoader
|
| 10 |
+
from langchain_community.embeddings import HuggingFaceEmbeddings
|
| 11 |
from langchain.text_splitter import CharacterTextSplitter
|
| 12 |
from langchain.schema import Document
|
| 13 |
from langchain_openai import OpenAIEmbeddings, ChatOpenAI
|
|
|
|
| 14 |
from langchain_chroma import Chroma
|
| 15 |
from langchain.memory import ConversationBufferMemory
|
| 16 |
from langchain.chains import ConversationalRetrievalChain
|
|
|
|
| 23 |
from random import randint
|
| 24 |
import shutil
|
| 25 |
|
|
|
|
| 26 |
db_name = "vector_db"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
folder = "my-knowledge-base/"
|
|
|
|
| 28 |
MODEL = "llama3.2:latest"
|
| 29 |
|
| 30 |
def process_files(files):
|