Spaces:
Running on Zero
Running on Zero
Upload src/veris_classifier/classifier.py with huggingface_hub
Browse files
src/veris_classifier/classifier.py
CHANGED
|
@@ -33,8 +33,8 @@ QA_SYSTEM_PROMPT = (
|
|
| 33 |
|
| 34 |
# ββ HF Model Backend βββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 35 |
|
| 36 |
-
HF_MODEL_ID = "vibesecurityguy/veris-classifier-
|
| 37 |
-
BASE_MODEL_ID = "
|
| 38 |
_hf_pipeline = None
|
| 39 |
_hf_tokenizer = None
|
| 40 |
|
|
@@ -43,7 +43,7 @@ def load_hf_model():
|
|
| 43 |
"""Load the base model + LoRA adapter from HF Hub. Called once on first request.
|
| 44 |
|
| 45 |
The model repo only contains LoRA adapter weights (162 MB), not a full model.
|
| 46 |
-
We load the base
|
| 47 |
"""
|
| 48 |
global _hf_pipeline, _hf_tokenizer
|
| 49 |
|
|
|
|
| 33 |
|
| 34 |
# ββ HF Model Backend βββββββββββββββββββββββββββββββββββββββββββββββββββββ
|
| 35 |
|
| 36 |
+
HF_MODEL_ID = "vibesecurityguy/veris-classifier-v2" # LoRA adapter repo
|
| 37 |
+
BASE_MODEL_ID = "mistralai/Mistral-7B-Instruct-v0.3" # Base model
|
| 38 |
_hf_pipeline = None
|
| 39 |
_hf_tokenizer = None
|
| 40 |
|
|
|
|
| 43 |
"""Load the base model + LoRA adapter from HF Hub. Called once on first request.
|
| 44 |
|
| 45 |
The model repo only contains LoRA adapter weights (162 MB), not a full model.
|
| 46 |
+
We load the base Mistral-7B-Instruct model, then merge the adapter on top.
|
| 47 |
"""
|
| 48 |
global _hf_pipeline, _hf_tokenizer
|
| 49 |
|