Shubham170793 commited on
Commit
93a72c6
·
verified ·
1 Parent(s): 641185f

Update src/qa.py

Browse files
Files changed (1) hide show
  1. src/qa.py +9 -21
src/qa.py CHANGED
@@ -5,14 +5,14 @@ from vectorstore import search_faiss
5
 
6
  print("✅ qa.py loaded from:", __file__)
7
 
8
- # Always redirect Hugging Face caches to /tmp
9
  CACHE_DIR = "/tmp/huggingface"
10
  os.environ["HF_HOME"] = CACHE_DIR
11
  os.environ["TRANSFORMERS_CACHE"] = CACHE_DIR
12
  os.environ["HF_DATASETS_CACHE"] = CACHE_DIR
13
 
14
  # ----------------------------
15
- # Embedding model for queries
16
  # ----------------------------
17
  _query_model = SentenceTransformer(
18
  "sentence-transformers/all-MiniLM-L6-v2",
@@ -23,25 +23,13 @@ _query_model = SentenceTransformer(
23
  # LLM for answers
24
  # ----------------------------
25
  MODEL_NAME = "google/flan-t5-small"
26
- MODEL_PATH = os.path.join(CACHE_DIR, MODEL_NAME)
27
-
28
- if not os.path.exists(MODEL_PATH):
29
- print(f"⬇️ Downloading {MODEL_NAME} to {MODEL_PATH}")
30
- _answer_model = pipeline(
31
- "text2text-generation",
32
- model=MODEL_NAME,
33
- cache_dir=CACHE_DIR
34
- )
35
- # Save pipeline model locally
36
- _answer_model.model.save_pretrained(MODEL_PATH)
37
- _answer_model.tokenizer.save_pretrained(MODEL_PATH)
38
- else:
39
- print(f"✅ Loading {MODEL_NAME} from {MODEL_PATH}")
40
- _answer_model = pipeline(
41
- "text2text-generation",
42
- model=MODEL_PATH,
43
- cache_dir=CACHE_DIR
44
- )
45
 
46
  # ----------------------------
47
  # Functions
 
5
 
6
  print("✅ qa.py loaded from:", __file__)
7
 
8
+ # Force Hugging Face to use /tmp for cache
9
  CACHE_DIR = "/tmp/huggingface"
10
  os.environ["HF_HOME"] = CACHE_DIR
11
  os.environ["TRANSFORMERS_CACHE"] = CACHE_DIR
12
  os.environ["HF_DATASETS_CACHE"] = CACHE_DIR
13
 
14
  # ----------------------------
15
+ # Query embedding model
16
  # ----------------------------
17
  _query_model = SentenceTransformer(
18
  "sentence-transformers/all-MiniLM-L6-v2",
 
23
  # LLM for answers
24
  # ----------------------------
25
  MODEL_NAME = "google/flan-t5-small"
26
+
27
+ # Make sure model downloads into /tmp
28
+ _answer_model = pipeline(
29
+ "text2text-generation",
30
+ model=MODEL_NAME,
31
+ cache_dir=CACHE_DIR
32
+ )
 
 
 
 
 
 
 
 
 
 
 
 
33
 
34
  # ----------------------------
35
  # Functions