mingmingmom888 commited on
Commit
04d974b
·
1 Parent(s): 878840f

Upload app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -13
app.py CHANGED
@@ -36,24 +36,12 @@ text_splitter = RecursiveCharacterTextSplitter(
36
  documents = text_splitter.split_documents(data)
37
 
38
 
39
- #from transformers import AutoTokenizer
40
-
41
- #bloomz_tokenizer = AutoTokenizer.from_pretrained("bigscience/bloomz-1b7")
42
-
43
  from langchain.embeddings.openai import OpenAIEmbeddings
44
 
45
  embeddings = OpenAIEmbeddings()
46
 
47
 
48
 
49
- #text_splitter = CharacterTextSplitter.from_huggingface_tokenizer(bloomz_tokenizer, chunk_size=100, chunk_overlap=0, separator="\n")
50
-
51
- #documents = text_splitter.split_documents(data)
52
-
53
- #from langchain.embeddings import HuggingFaceEmbeddings
54
-
55
- #embeddings = HuggingFaceEmbeddings()
56
-
57
  from langchain.vectorstores import Chroma
58
 
59
  persist_directory = "vector_db"
@@ -78,7 +66,7 @@ shakespeare_qa = RetrievalQA.from_chain_type(llm=llm, chain_type="stuff", retrie
78
 
79
 
80
 
81
- chain = load_qa_chain(OpenAI(temperature=0), chain_type="stuff")
82
 
83
  def make_inference(query):
84
  docs = shakespeare_qa.get_relevant_documents(query)
 
36
  documents = text_splitter.split_documents(data)
37
 
38
 
 
 
 
 
39
  from langchain.embeddings.openai import OpenAIEmbeddings
40
 
41
  embeddings = OpenAIEmbeddings()
42
 
43
 
44
 
 
 
 
 
 
 
 
 
45
  from langchain.vectorstores import Chroma
46
 
47
  persist_directory = "vector_db"
 
66
 
67
 
68
 
69
+ #chain = load_qa_chain(OpenAI(temperature=0), chain_type="stuff")
70
 
71
  def make_inference(query):
72
  docs = shakespeare_qa.get_relevant_documents(query)