Whisper Tiny - MLX FP16

This is the OpenAI Whisper Tiny model converted to MLX format with FP16 precision, optimized for Apple Silicon inference.

Model Details

Property Value
Base Model openai/whisper-tiny
Parameters ~39M
Format MLX SafeTensors (FP16)
Model Size 70.94 MB
Sample Rate 16,000 Hz
Audio Layers 4
Text Layers 4
Hidden Size 384
Attention Heads 6
Vocabulary Size 51,865

Intended Use

This model is optimized for on-device automatic speech recognition (ASR) on Apple Silicon devices (Mac, iPhone, iPad). It is designed for use with the WhisperKit or MLX frameworks.

Files

  • config.json - Model configuration
  • model.safetensors - Model weights in SafeTensors format (FP16)
  • multilingual.tiktoken - Tokenizer

Usage

import mlx_whisper

result = mlx_whisper.transcribe(
    "audio.mp3",
    path_or_hf_repo="aitytech/Whisper-Tiny-MLX-FP16",
)
print(result["text"])

Original Model

Downloads last month
-
MLX
Hardware compatibility
Log In to add your hardware

Quantized

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for aitytech/Whisper-Tiny-MLX-FP16

Finetuned
(1701)
this model

Paper for aitytech/Whisper-Tiny-MLX-FP16