File size: 1,380 Bytes
bd8b4ab |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 |
---
language:
- es
- ca
license: apache-2.0
---
# OVOS - Whisper Large v3 Tiny Caesar
This model is an ONNX-format export of the model available at [projecte-aina/whisper-large-v3-tiny-caesar](https://huggingface.co/projecte-aina/whisper-large-v3-tiny-caesar), for ease of use in edge devices and CPU-based inference environments.
# Requirements
The export is based on:
- [optimum](https://github.com/huggingface/optimum) for exporting the model
- [onnx-asr](https://github.com/istupakov/onnx-asr) for inference
The requirements can be installed as
```bash
$ pip install optimum[onnxruntime] onnx-asr
```
# Usage
```python
import onnx_asr
model = onnx_asr.load_model("OpenVoiceOS/whisper-large-v3-tiny-caesar-onnx")
print(model.recognize("test.wav"))
```
# Export
According to [onnx-asr/convert-model-to-onnx](https://github.com/istupakov/onnx-asr?tab=readme-ov-file#convert-model-to-onnx)):
```bash
$ optimum-cli export onnx --task automatic-speech-recognition-with-past --model projecte-aina/whisper-large-v3-tiny-caesar whisper-onnx
$ cd whisper-onnx && rm decoder.onnx* decoder_with_past_model.onnx* # only the merged decoder is needed
```
# Licensing
The license is derived from the original model: Apache 2.0. For more details, please refer to [projecte-aina/whisper-large-v3-tiny-caesar](https://huggingface.co/projecte-aina/whisper-large-v3-tiny-caesar).
|