Qwen3-ASR-1.7B-Q8
8-bit quantized version of Qwen/Qwen3-ASR-1.7B for MLX.
Quantization
- Bits: 8
- Group size: 64
- Format: MLX safetensors
- Quantized using mlx-qwen3-asr
Usage
from mlx_qwen3_asr import Session
session = Session("illitan/Qwen3-ASR-1.7B-Q8")
result = session.transcribe("audio.wav")
print(result.text)
Source
Quantized from Qwen/Qwen3-ASR-1.7B using nn.quantize(model, bits=8, group_size=64).
- Downloads last month
- 27
Hardware compatibility
Log In to add your hardware
Quantized
Model tree for illitan/Qwen3-ASR-1.7B-Q8
Base model
Qwen/Qwen3-ASR-1.7B