Whisper Small โ MLX 4-bit Quantized
Custom MLX 4-bit quantization of OpenAI Whisper Small optimized for MetalRT GPU inference on Apple Silicon.
Usage
Used by RCLI with the MetalRT engine for speech-to-text:
rcli setup # select MetalRT or Both engines
Note: Whisper Small is in GPU beta. Whisper Tiny is recommended for production use.
License
Model weights: MIT (OpenAI) MetalRT engine: Proprietary (RunAnywhere, Inc.)
Contact
- Downloads last month
- 42
Hardware compatibility
Log In to add your hardware
Quantized
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support