Voxtral-Mini-4B-Realtime-2602-TurboQuant-MLX-2bit

2-bit MLX weight-quantized build of mistralai/Voxtral-Mini-4B-Realtime-2602 with TurboQuant KV-cache. Extreme-compression variant for memory-constrained real-time ASR.

Overview

  • Base: mistralai/Voxtral-Mini-4B-Realtime-2602 โ€” 4B real-time ASR model
  • Weight precision: 2-bit (group-wise)
  • KV-cache profile: TurboQuant
  • Approx. on-disk size: ~1.2 GB
  • Runtime: MLX on Apple Silicon

Best used with clean, single-speaker audio. For noisier streams consider the RotorQuant 2-bit variant.

Quickstart

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("majentik/Voxtral-Mini-4B-Realtime-2602-TurboQuant-MLX-2bit")

for chunk in audio_stream():
    prompt = tokenizer.apply_chat_template(
        [{"role": "user", "content": [{"type": "audio", "path": chunk}]}],
        add_generation_prompt=True,
    )
    emit(generate(model, tokenizer, prompt=prompt, max_tokens=32))

Model specs

Field Value
Parameters 4B
Weight bits 2
Group size 32
Cache profile TurboQuant
Size on disk ~1.2 GB
Target hardware Apple Silicon (M1/M2/M3/M4)
License Apache 2.0

RotorQuant vs TurboQuant

TurboQuant RotorQuant
Strategy Per-head static calibration Rotational online re-basis
Memory reduction ~3.5x on KV-cache ~4x on KV-cache
Best for Predictable domains Noisy/multi-speaker streams

See also

Downloads last month
12
MLX
Hardware compatibility
Log In to add your hardware

2-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for majentik/Voxtral-Mini-4B-Realtime-2602-TurboQuant-MLX-2bit