How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("text-generation", model="ig1/BioMistral-7B-FP8-Dynamic")
messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe(messages)
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("ig1/BioMistral-7B-FP8-Dynamic")
model = AutoModelForCausalLM.from_pretrained("ig1/BioMistral-7B-FP8-Dynamic")
messages = [
    {"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
	messages,
	add_generation_prompt=True,
	tokenize=True,
	return_dict=True,
	return_tensors="pt",
).to(model.device)

outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))
Quick Links

BioMistral-7B-FP8-Dynamic

Overview

BioMistral-7B-FP8-Dynamic is an FP8 Dynamic–quantized version of the BioMistral-7B model, designed for high-performance inference while maintaining strong quality on biomedical and medical NLP tasks.

This model is primarily intended for deployment with vLLM on modern GPUs (Hopper / Ada architectures).


Base Model

  • Base model: BioMistral-7B
  • Architecture: Mistral-style decoder-only Transformer
  • Domain: Biomedical / Medical Natural Language Processing

Quantization

  • Method: FP8 Dynamic
  • Scope: Linear layers
  • Objective: Reduce VRAM usage and improve inference throughput

Notes

  • The weights are already quantized.
  • Do not apply additional runtime quantization.

Intended Use

  • Biomedical and medical text generation
  • Medical writing assistance
  • Summarization and analysis of scientific literature
  • Medical RAG pipelines (clinical notes, research papers)

Deployment (vLLM)

Recommended

vllm serve ig1/BioMistral-7B-FP8-Dynamic \
  --served-model-name biomistral-7b-fp8 \
  --dtype auto
Downloads last month
4
Safetensors
Model size
7B params
Tensor type
BF16
·
F8_E4M3
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support