File size: 1,380 Bytes
bd8b4ab
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46

---
language: 
- es
- ca

license: apache-2.0
---

# OVOS - Whisper Large v3 Tiny Caesar

This model is an ONNX-format export of the model available at [projecte-aina/whisper-large-v3-tiny-caesar](https://huggingface.co/projecte-aina/whisper-large-v3-tiny-caesar), for ease of use in edge devices and CPU-based inference environments.

# Requirements

The export is based on:
- [optimum](https://github.com/huggingface/optimum) for exporting the model
- [onnx-asr](https://github.com/istupakov/onnx-asr) for inference

The requirements can be installed as

```bash
$ pip install optimum[onnxruntime] onnx-asr
```

# Usage

```python
import onnx_asr
model = onnx_asr.load_model("OpenVoiceOS/whisper-large-v3-tiny-caesar-onnx")
print(model.recognize("test.wav"))
```

# Export

According to [onnx-asr/convert-model-to-onnx](https://github.com/istupakov/onnx-asr?tab=readme-ov-file#convert-model-to-onnx)):

```bash
$ optimum-cli export onnx --task automatic-speech-recognition-with-past --model projecte-aina/whisper-large-v3-tiny-caesar whisper-onnx
$ cd whisper-onnx && rm decoder.onnx* decoder_with_past_model.onnx*  # only the merged decoder is needed
```

# Licensing

The license is derived from the original model: Apache 2.0. For more details, please refer to [projecte-aina/whisper-large-v3-tiny-caesar](https://huggingface.co/projecte-aina/whisper-large-v3-tiny-caesar).