Qwen3-ASR-1.7B-Q8

8-bit quantized version of Qwen/Qwen3-ASR-1.7B for MLX.

Quantization

  • Bits: 8
  • Group size: 64
  • Format: MLX safetensors
  • Quantized using mlx-qwen3-asr

Usage

from mlx_qwen3_asr import Session

session = Session("illitan/Qwen3-ASR-1.7B-Q8")
result = session.transcribe("audio.wav")
print(result.text)

Source

Quantized from Qwen/Qwen3-ASR-1.7B using nn.quantize(model, bits=8, group_size=64).

Downloads last month
27
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for illitan/Qwen3-ASR-1.7B-Q8

Finetuned
(6)
this model