use optimum ONNX MiniLM (no disk, no internet)
Browse files
rag.py
CHANGED
|
@@ -105,21 +105,16 @@ def get_texts() -> List[str]:
|
|
| 105 |
|
| 106 |
## ------------------------------------------------------------------------------rtutu
|
| 107 |
# ------------------------------------------------------------------
|
|
|
|
| 108 |
@lru_cache(maxsize=1)
|
| 109 |
def get_vectorstore() -> FAISS:
|
| 110 |
texts = get_texts()
|
| 111 |
|
| 112 |
-
# --- FINAL:
|
| 113 |
-
import
|
| 114 |
-
|
| 115 |
-
|
| 116 |
)
|
| 117 |
-
|
| 118 |
-
from sentence_transformers import SentenceTransformer
|
| 119 |
-
model = SentenceTransformer(local_model_path, device="cpu", cache_folder=None)
|
| 120 |
-
|
| 121 |
-
from langchain.embeddings import SentenceTransformerEmbeddings
|
| 122 |
-
embeddings = SentenceTransformerEmbeddings(model=model)
|
| 123 |
# ------------------------------------------------------------------------
|
| 124 |
|
| 125 |
if not texts: # no data → empty FAISS
|
|
|
|
| 105 |
|
| 106 |
## ------------------------------------------------------------------------------rtutu
|
| 107 |
# ------------------------------------------------------------------
|
| 108 |
+
# ------------------------------------------------------------------
|
| 109 |
@lru_cache(maxsize=1)
|
| 110 |
def get_vectorstore() -> FAISS:
|
| 111 |
texts = get_texts()
|
| 112 |
|
| 113 |
+
# --- FINAL: use optimum ONNX MiniLM (already on disk) -----------------
|
| 114 |
+
from langchain_community.embeddings import OptimumSentenceEmbeddings
|
| 115 |
+
embeddings = OptimumSentenceEmbeddings(
|
| 116 |
+
model_name="optimum/all-MiniLM-L6-v2"
|
| 117 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 118 |
# ------------------------------------------------------------------------
|
| 119 |
|
| 120 |
if not texts: # no data → empty FAISS
|