Model Card for Model ID

This model is a quantized version of Qwen3-32B converted to AWQ Q4 format using the mlx library for efficient inference. It retains the core capabilities of Qwen3 while optimizing for resource constraints.

Downloads last month
19
Safetensors
Model size
5B params
Tensor type
BF16
·
U32
·
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Goraint/Qwen3-32B-4bit-AWQ-MLX

Base model

Qwen/Qwen3-32B
Finetuned
(470)
this model