mingmingmom888 commited on
Commit
edf5bb0
·
1 Parent(s): eb3426c

Upload 3 files

Browse files
Files changed (3) hide show
  1. app.py +94 -0
  2. guide1.txt +0 -0
  3. requirements.txt +8 -0
app.py ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from langchain.embeddings.openai import OpenAIEmbeddings
2
+ from langchain.vectorstores import Chroma
3
+ from langchain.text_splitter import CharacterTextSplitter
4
+ from langchain.chains.question_answering import load_qa_chain
5
+ from langchain.llms import OpenAI
6
+ import os
7
+ from glob import glob
8
+ import shutil
9
+
10
+
11
+ files = glob("./shakespeare/**/*.html")
12
+
13
+ os.mkdir('./data')
14
+ destination_folder = './data/'
15
+
16
+ for html_file in files:
17
+ shutil.move(html_file, destination_folder + html_file.split("/")[-1])
18
+
19
+
20
+
21
+ from langchain.document_loaders import BSHTMLLoader, DirectoryLoader
22
+
23
+ bshtml_dir_loader = DirectoryLoader('./data/', loader_cls=BSHTMLLoader)
24
+
25
+ data = bshtml_dir_loader.load()
26
+
27
+ text_splitter = CharacterTextSplitter.from_huggingface_tokenizer(bloomz_tokenizer, chunk_size=100, chunk_overlap=0, separator="\n")
28
+
29
+ documents = text_splitter.split_documents(data)
30
+
31
+ from langchain.embeddings import HuggingFaceEmbeddings
32
+
33
+ embeddings = HuggingFaceEmbeddings()
34
+
35
+
36
+ persist_directory = "vector_db"
37
+
38
+ vectordb = Chroma.from_documents(documents=documents, embedding=embeddings, persist_directory=persist_directory)
39
+ vectordb = Chroma(persist_directory=persist_directory, embedding_function=embeddings)
40
+
41
+
42
+ vectordb.persist()
43
+ vectordb = None
44
+
45
+
46
+ from langchain import HuggingFacePipeline
47
+
48
+ llm = HuggingFacePipeline.from_model_id(
49
+ model_id="bigscience/bloomz-1b7",
50
+ task="text-generation",
51
+ model_kwargs={"temperature" : 0, "max_length" : 500})
52
+
53
+ doc_retriever = vectordb.as_retriever()
54
+
55
+ from langchain.chains import RetrievalQA
56
+
57
+ shakespeare_qa = RetrievalQA.from_chain_type(llm=llm, chain_type="stuff", retriever=doc_retriever)
58
+
59
+
60
+ """
61
+
62
+ with open("guide1.txt") as f:
63
+ hitchhikersguide = f.read()
64
+
65
+ text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=0, separator = "\n")
66
+ texts = text_splitter.split_text(hitchhikersguide)
67
+
68
+
69
+ from langchain.embeddings.openai import OpenAIEmbeddings
70
+ #os.environ["OPENAI_API_KEY"] = openai.api_key
71
+ embeddings = OpenAIEmbeddings()
72
+ embeddings = OpenAIEmbeddings()
73
+
74
+ docsearch = Chroma.from_texts(texts, embeddings, metadatas=[{"source": str(i)} for i in range(len(texts))]).as_retriever()
75
+ """
76
+ chain = load_qa_chain(OpenAI(temperature=0), chain_type="stuff")
77
+
78
+ def make_inference(query):
79
+ docs = shakespeare_qa.get_relevant_documents(query)
80
+ return(chain.run(input_documents=docs, question=query))
81
+
82
+ if __name__ == "__main__":
83
+ # make a gradio interface
84
+ import gradio as gr
85
+
86
+ gr.Interface(
87
+ make_inference,
88
+ [
89
+ gr.inputs.Textbox(lines=2, label="Query"),
90
+ ],
91
+ gr.outputs.Textbox(label="Response"),
92
+ title="🗣️TalkToMyDoc📄",
93
+ description="🗣️TalkToMyDoc📄 is a tool that allows you to ask questions about a document. In this case - Hitch Hitchhiker's Guide to the Galaxy.",
94
+ ).launch()
guide1.txt ADDED
The diff for this file is too large to render. See raw diff
 
requirements.txt ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ langchain
2
+ openai
3
+ tiktoken
4
+ beautifulsoup4
5
+ transformers
6
+ huggingface-hub
7
+ sentence_transformers
8
+ chromadb