Update app.py
Browse files
app.py
CHANGED
|
@@ -31,13 +31,13 @@ def load_models():
|
|
| 31 |
print("📥 Initialisation CatSense v12.13 (Vision Pure Mode)...")
|
| 32 |
|
| 33 |
# Modèle VLM (seulement le modèle, pas le processor)
|
| 34 |
-
vlm_id = "HuggingFaceTB/
|
| 35 |
vlm_model = AutoModelForImageTextToText.from_pretrained(
|
| 36 |
vlm_id, torch_dtype=torch.bfloat16 if torch.cuda.is_available() else torch.float32
|
| 37 |
).to(DEVICE).eval()
|
| 38 |
|
| 39 |
# LLM
|
| 40 |
-
llm_id = "HuggingFaceTB/SmolLM2-
|
| 41 |
llm_tok = AutoTokenizer.from_pretrained(llm_id)
|
| 42 |
llm_model = AutoModelForCausalLM.from_pretrained(
|
| 43 |
llm_id, torch_dtype=torch.bfloat16 if torch.cuda.is_available() else torch.float32
|
|
|
|
| 31 |
print("📥 Initialisation CatSense v12.13 (Vision Pure Mode)...")
|
| 32 |
|
| 33 |
# Modèle VLM (seulement le modèle, pas le processor)
|
| 34 |
+
vlm_id = "HuggingFaceTB/SmolVLM2-256M-Video-Instruct"
|
| 35 |
vlm_model = AutoModelForImageTextToText.from_pretrained(
|
| 36 |
vlm_id, torch_dtype=torch.bfloat16 if torch.cuda.is_available() else torch.float32
|
| 37 |
).to(DEVICE).eval()
|
| 38 |
|
| 39 |
# LLM
|
| 40 |
+
llm_id = "HuggingFaceTB/SmolLM2-135M-Instruct"
|
| 41 |
llm_tok = AutoTokenizer.from_pretrained(llm_id)
|
| 42 |
llm_model = AutoModelForCausalLM.from_pretrained(
|
| 43 |
llm_id, torch_dtype=torch.bfloat16 if torch.cuda.is_available() else torch.float32
|