mlx-community/whisper-medium-asr-fp16

This model was converted to MLX format from openai/whisper-medium using mlx-audio version 0.2.10. Refer to the original model card for more details on the model.

Use with mlx-audio

pip install -U mlx-audio

CLI Example:

python -m mlx_audio.stt.generate --model mlx-community/whisper-medium-asr-fp16 --audio "audio.wav"

Python Example:

from mlx_audio.stt.utils import load_model
from mlx_audio.stt.generate import generate_transcription
model = load_model("mlx-community/whisper-medium-asr-fp16")
transcription = generate_transcription(
    model=model,
    audio_path="path_to_audio.wav",
    output_path="path_to_output.txt",
    format="txt",
    verbose=True,
)
print(transcription.text)
Downloads last month
10
Safetensors
Model size
0.8B params
Tensor type
F16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Evaluation results