Update src/qa.py
Browse files
src/qa.py
CHANGED
|
@@ -5,14 +5,14 @@ from vectorstore import search_faiss
|
|
| 5 |
|
| 6 |
print("✅ qa.py loaded from:", __file__)
|
| 7 |
|
| 8 |
-
#
|
| 9 |
CACHE_DIR = "/tmp/huggingface"
|
| 10 |
os.environ["HF_HOME"] = CACHE_DIR
|
| 11 |
os.environ["TRANSFORMERS_CACHE"] = CACHE_DIR
|
| 12 |
os.environ["HF_DATASETS_CACHE"] = CACHE_DIR
|
| 13 |
|
| 14 |
# ----------------------------
|
| 15 |
-
#
|
| 16 |
# ----------------------------
|
| 17 |
_query_model = SentenceTransformer(
|
| 18 |
"sentence-transformers/all-MiniLM-L6-v2",
|
|
@@ -23,25 +23,13 @@ _query_model = SentenceTransformer(
|
|
| 23 |
# LLM for answers
|
| 24 |
# ----------------------------
|
| 25 |
MODEL_NAME = "google/flan-t5-small"
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
cache_dir=CACHE_DIR
|
| 34 |
-
)
|
| 35 |
-
# Save pipeline model locally
|
| 36 |
-
_answer_model.model.save_pretrained(MODEL_PATH)
|
| 37 |
-
_answer_model.tokenizer.save_pretrained(MODEL_PATH)
|
| 38 |
-
else:
|
| 39 |
-
print(f"✅ Loading {MODEL_NAME} from {MODEL_PATH}")
|
| 40 |
-
_answer_model = pipeline(
|
| 41 |
-
"text2text-generation",
|
| 42 |
-
model=MODEL_PATH,
|
| 43 |
-
cache_dir=CACHE_DIR
|
| 44 |
-
)
|
| 45 |
|
| 46 |
# ----------------------------
|
| 47 |
# Functions
|
|
|
|
| 5 |
|
| 6 |
print("✅ qa.py loaded from:", __file__)
|
| 7 |
|
| 8 |
+
# Force Hugging Face to use /tmp for cache
|
| 9 |
CACHE_DIR = "/tmp/huggingface"
|
| 10 |
os.environ["HF_HOME"] = CACHE_DIR
|
| 11 |
os.environ["TRANSFORMERS_CACHE"] = CACHE_DIR
|
| 12 |
os.environ["HF_DATASETS_CACHE"] = CACHE_DIR
|
| 13 |
|
| 14 |
# ----------------------------
|
| 15 |
+
# Query embedding model
|
| 16 |
# ----------------------------
|
| 17 |
_query_model = SentenceTransformer(
|
| 18 |
"sentence-transformers/all-MiniLM-L6-v2",
|
|
|
|
| 23 |
# LLM for answers
|
| 24 |
# ----------------------------
|
| 25 |
MODEL_NAME = "google/flan-t5-small"
|
| 26 |
+
|
| 27 |
+
# Make sure model downloads into /tmp
|
| 28 |
+
_answer_model = pipeline(
|
| 29 |
+
"text2text-generation",
|
| 30 |
+
model=MODEL_NAME,
|
| 31 |
+
cache_dir=CACHE_DIR
|
| 32 |
+
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
# ----------------------------
|
| 35 |
# Functions
|