File size: 1,495 Bytes
76d9c4f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
---
title: Whisper V3 Playground
description: Translate audio snippets into text on a Streamlit playground.
version: EN
---

## Try out this model on [VESSL Hub](https://vessl.ai/hub).

This example runs a general-purpose speech recognition model,  [Whisper V3](https://github.com/openai/whisper). It is trained on a 680k hours of diverse labelled audio. Whisper is also a multitasking model that can perform multilingual speech recognition, speech translation, and language identification. It can generalize to many domains without additional fine-tuning.

<img
  className="rounded-md"
  src="/images/whisper-results.png"
/>

## Running the model

You can run the model with our quick command.
```sh
vessl run create -f whisper.yaml
```

If you open log pages, you can see the result of inference for first 5 data in [Librispeech_asr dataset](https://www.openslr.org/12).


Here's a rundown of the `whisper.yaml` file.
```yaml
name: whisper-v3
description: A template Run for inference of whisper v3 on librispeech_asr test set
resources:
  cluster: vessl-gcp-oregon
  preset: v1.l4-1.mem-42
image: quay.io/vessl-ai/hub:torch2.1.0-cuda12.2-202312070053
import:
  /model/: hf://huggingface.co/VESSL/Whisper-large-v3
  /dataset/: hf://huggingface.co/datasets/VESSL/librispeech_asr_clean_test
  /code/:
    git:
      url: https://github.com/vessl-ai/hub-model
      ref: main
run:
  - command: |-
      pip install -r requirements.txt
      python inference.py
    workdir: /code/whisper-v3
```