div0-space's picture
Upload README.md with huggingface_hub
e1eebdf verified
---
license: mit
language:
- en
- multilingual
tags:
- whisper
- speech-to-text
- mlx
- quantized
- q8
- apple-silicon
library_name: transformers
pipeline_tag: automatic-speech-recognition
base_model: openai/whisper-medium
---
# Whisper Medium - MLX Q8 Quantized
8-bit quantized version of OpenAI's Whisper Medium, optimized for Apple Silicon with MLX.
## Model Details
| Property | Value |
|----------|-------|
| Original Model | [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) |
| Parameters | ~769M |
| Quantization | INT8 (Q8) |
| Size | ~800MB |
## Other Whisper Models
| Model | Size | Link |
|-------|------|------|
| small | ~300MB | [LibraxisAI/whisper-small-mlx-q8](https://huggingface.co/LibraxisAI/whisper-small-mlx-q8) |
| **medium** (this) | ~800MB | - |
| large-v3 | ~1.6GB | [LibraxisAI/whisper-large-v3-mlx-q8](https://huggingface.co/LibraxisAI/whisper-large-v3-mlx-q8) |
| large-v3-turbo | ~900MB | [LibraxisAI/whisper-large-v3-turbo-mlx-q8](https://huggingface.co/LibraxisAI/whisper-large-v3-turbo-mlx-q8) |
## Usage
```python
import mlx_whisper
result = mlx_whisper.transcribe(
"audio.wav",
path_or_hf_repo="LibraxisAI/whisper-medium-mlx-q8"
)
print(result["text"])
```
## Hardware Requirements
- Apple Silicon Mac (M1/M2/M3/M4)
- Minimum 8GB RAM
## License
MIT - inherited from OpenAI Whisper.
---
*Converted by [LibraxisAI](https://huggingface.co/LibraxisAI) using mlx-whisper*