Spaces:
Running on Zero
Running on Zero
Upload src/veris_classifier/classifier.py with huggingface_hub
Browse files
src/veris_classifier/classifier.py
CHANGED
|
@@ -54,6 +54,15 @@ def load_hf_model():
|
|
| 54 |
from peft import PeftModel
|
| 55 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
| 56 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 57 |
logger.info(f"Loading base model: {BASE_MODEL_ID}")
|
| 58 |
_hf_tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL_ID, trust_remote_code=True)
|
| 59 |
if _hf_tokenizer.pad_token is None:
|
|
|
|
| 54 |
from peft import PeftModel
|
| 55 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
| 56 |
|
| 57 |
+
# This model path expects GPU execution (ZeroGPU on Spaces). On CPU-only
|
| 58 |
+
# runtimes, transformers can fail with opaque disk offload errors.
|
| 59 |
+
if not torch.cuda.is_available():
|
| 60 |
+
raise RuntimeError(
|
| 61 |
+
"Fine-tuned model requires GPU. This Space appears to be on CPU-only "
|
| 62 |
+
"(no CUDA device available). Request ZeroGPU (A10G) or provide an "
|
| 63 |
+
"OpenAI API key to use fallback inference."
|
| 64 |
+
)
|
| 65 |
+
|
| 66 |
logger.info(f"Loading base model: {BASE_MODEL_ID}")
|
| 67 |
_hf_tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL_ID, trust_remote_code=True)
|
| 68 |
if _hf_tokenizer.pad_token is None:
|