Update app.py
Browse files
app.py
CHANGED
|
@@ -6,12 +6,13 @@ import nest_asyncio
|
|
| 6 |
from llama_index import ServiceContext, LLMPredictor, PromptHelper
|
| 7 |
from llama_index.text_splitter import TokenTextSplitter
|
| 8 |
from llama_index.node_parser import SimpleNodeParser
|
| 9 |
-
from langchain.embeddings import HuggingFaceEmbeddings
|
| 10 |
from llama_index import SimpleDirectoryReader, VectorStoreIndex
|
| 11 |
from gradio import Interface
|
| 12 |
nest_asyncio.apply()
|
| 13 |
|
| 14 |
-
embed_model =
|
|
|
|
| 15 |
"""
|
| 16 |
node_parser = SimpleNodeParser.from_defaults(text_splitter=TokenTextSplitter(chunk_size=1024, chunk_overlap=20))
|
| 17 |
prompt_helper = PromptHelper(
|
|
|
|
| 6 |
from llama_index import ServiceContext, LLMPredictor, PromptHelper
|
| 7 |
from llama_index.text_splitter import TokenTextSplitter
|
| 8 |
from llama_index.node_parser import SimpleNodeParser
|
| 9 |
+
from langchain.embeddings import HuggingFaceEmbeddings, HuggingFaceInstructEmbeddings
|
| 10 |
from llama_index import SimpleDirectoryReader, VectorStoreIndex
|
| 11 |
from gradio import Interface
|
| 12 |
nest_asyncio.apply()
|
| 13 |
|
| 14 |
+
embed_model = HuggingFaceInstructEmbeddings(model_name="hkunlp/instructor-xl",
|
| 15 |
+
model_kwargs={"device": "cpu"})
|
| 16 |
"""
|
| 17 |
node_parser = SimpleNodeParser.from_defaults(text_splitter=TokenTextSplitter(chunk_size=1024, chunk_overlap=20))
|
| 18 |
prompt_helper = PromptHelper(
|