from `llama_index.embeddings.huggingface_api` to `llama_index.embeddings.huggingface`
#68
by
shern - opened
unit2/llama-index/components.ipynb
CHANGED
|
@@ -142,7 +142,7 @@
|
|
| 142 |
}
|
| 143 |
],
|
| 144 |
"source": [
|
| 145 |
-
"from llama_index.embeddings.
|
| 146 |
"from llama_index.core.node_parser import SentenceSplitter\n",
|
| 147 |
"from llama_index.core.ingestion import IngestionPipeline\n",
|
| 148 |
"\n",
|
|
@@ -223,7 +223,7 @@
|
|
| 223 |
"outputs": [],
|
| 224 |
"source": [
|
| 225 |
"from llama_index.core import VectorStoreIndex\n",
|
| 226 |
-
"from llama_index.embeddings.
|
| 227 |
"\n",
|
| 228 |
"embed_model = HuggingFaceInferenceAPIEmbedding(model_name=\"BAAI/bge-small-en-v1.5\")\n",
|
| 229 |
"index = VectorStoreIndex.from_vector_store(\n",
|
|
|
|
| 142 |
}
|
| 143 |
],
|
| 144 |
"source": [
|
| 145 |
+
"from llama_index.embeddings.huggingface import HuggingFaceInferenceAPIEmbedding\n",
|
| 146 |
"from llama_index.core.node_parser import SentenceSplitter\n",
|
| 147 |
"from llama_index.core.ingestion import IngestionPipeline\n",
|
| 148 |
"\n",
|
|
|
|
| 223 |
"outputs": [],
|
| 224 |
"source": [
|
| 225 |
"from llama_index.core import VectorStoreIndex\n",
|
| 226 |
+
"from llama_index.embeddings.huggingface import HuggingFaceInferenceAPIEmbedding\n",
|
| 227 |
"\n",
|
| 228 |
"embed_model = HuggingFaceInferenceAPIEmbedding(model_name=\"BAAI/bge-small-en-v1.5\")\n",
|
| 229 |
"index = VectorStoreIndex.from_vector_store(\n",
|