testdocker / app.py
KavinduHansaka's picture
Update app.py
6e67f7a verified
raw
history blame contribute delete
429 Bytes
import torch
from transformers import AutoTokenizer, Mistral3ForConditionalGeneration, FineGrainedFP8Config
model_id = "mistralai/Ministral-3-14B-Instruct-2512"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = Mistral3ForConditionalGeneration.from_pretrained(
model_id,
device_map="auto",
quantization_config=FineGrainedFP8Config(dequantize=True),
torch_dtype=torch.bfloat16
)
print("Model loaded")