mudler's picture
Upload README.md with huggingface_hub
75f17d5 verified
---
language: it
license: mit
tags:
- whisper
- whisper.cpp
- ggml
- automatic-speech-recognition
- italian
- localai
- quantized
datasets:
- mozilla-foundation/common_voice_25_0
- facebook/multilingual_librispeech
- facebook/voxpopuli
- google/fleurs
base_model: openai/whisper-tiny
pipeline_tag: automatic-speech-recognition
---
# whisper-tiny-it-multi-ggml
[whisper.cpp](https://github.com/ggerganov/whisper.cpp) GGML quantizations of [LocalAI-io/whisper-tiny-it-multi](https://huggingface.co/LocalAI-io/whisper-tiny-it-multi) for fast CPU/GPU inference.
**Author:** Ettore Di Giacinto
Brought to you by the [LocalAI](https://github.com/mudler/LocalAI) team. These models can be used directly with [LocalAI](https://localai.io) and any whisper.cpp-based runtime.
## Files
| File | Quantization | Description |
|------|--------------|-------------|
| `ggml-model-f16.bin` | float16 | Full precision (no quantization) — highest quality |
| `ggml-model-q8_0.bin` | int8 | 8-bit quantization — minimal quality loss |
| `ggml-model-q5_0.bin` | int5 | 5-bit quantization — good quality/size tradeoff |
| `ggml-model-q4_0.bin` | int4 | 4-bit quantization — smallest size, fastest |
## Training
Fine-tuned [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) (39M params) on **Common Voice 25.0 + MLS + VoxPopuli + FLEURS Italian**.
See [LocalAI-io/whisper-tiny-it-multi](https://huggingface.co/LocalAI-io/whisper-tiny-it-multi) for the full safetensors model and detailed WER results.
## Usage
### whisper.cpp
```bash
# Download a quant
huggingface-cli download LocalAI-io/whisper-tiny-it-multi-ggml ggml-model-q5_0.bin --local-dir .
# Run
./whisper-cli -m ggml-model-q5_0.bin -f audio.wav -l it
```
### whisper.cpp Python bindings (pywhispercpp)
```python
from pywhispercpp.model import Model
model = Model("ggml-model-q5_0.bin", language="it")
segments = model.transcribe("audio.wav")
for seg in segments:
print(seg.text)
```
### LocalAI
```yaml
# In your LocalAI model config
name: whisper-tiny-it-multi
backend: whisper
parameters:
model: ggml-model-q5_0.bin
```
## Links
- **HF Safetensors:** [LocalAI-io/whisper-tiny-it-multi](https://huggingface.co/LocalAI-io/whisper-tiny-it-multi)
- **CTranslate2 INT8:** [LocalAI-io/whisper-tiny-it-multi-ct2-int8](https://huggingface.co/LocalAI-io/whisper-tiny-it-multi-ct2-int8)
- **Code:** [github.com/localai-org/whisper-it](https://github.com/localai-org/whisper-it)
- **whisper.cpp:** [github.com/ggerganov/whisper.cpp](https://github.com/ggerganov/whisper.cpp)
- **LocalAI:** [github.com/mudler/LocalAI](https://github.com/mudler/LocalAI)