File size: 5,031 Bytes
76397f6
 
 
 
 
 
 
 
 
 
 
9609dd2
 
 
 
 
76397f6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
07521ab
76397f6
 
07521ab
76397f6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9609dd2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
---
library_name: transformers
license: apache-2.0
base_model: openai/whisper-small
tags:
- generated_from_trainer
metrics:
- wer
model-index:
- name: Amit65/whisper-small-multilingual
  results: []
language:
- en
- mr
- hi
pipeline_tag: automatic-speech-recognition
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# Amit65/whisper-small-multilingual

This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6283
- Wer: 80.0691

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

full fine tuning on custom data and evaluate on word error rate(WER)

## Training procedure
Apply full fine tuning using hugging face trainer API


### Training results

| Training Loss | Epoch  | Step | Validation Loss | Wer      |
|:-------------:|:------:|:----:|:---------------:|:--------:|
| 3.4481        | 0.0480 | 25   | 1.7935          | 138.3641 |
| 1.494         | 0.0960 | 50   | 1.3053          | 105.6452 |
| 1.4092        | 0.1440 | 75   | 1.1546          | 102.6498 |
| 1.1367        | 0.1919 | 100  | 1.0424          | 105.4147 |
| 0.9748        | 0.2399 | 125  | 1.0038          | 116.7051 |
| 0.9522        | 0.2879 | 150  | 1.0032          | 140.6682 |
| 0.9114        | 0.3359 | 175  | 0.9329          | 126.2673 |
| 0.9498        | 0.3839 | 200  | 0.9077          | 117.0507 |
| 0.8762        | 0.4319 | 225  | 0.9359          | 97.4654  |
| 0.9051        | 0.4798 | 250  | 0.8390          | 88.5945  |
| 0.7941        | 0.5278 | 275  | 0.8869          | 105.2995 |
| 0.8417        | 0.5758 | 300  | 0.8299          | 109.7926 |
| 0.9244        | 0.6238 | 325  | 0.8105          | 79.9539  |
| 0.855         | 0.6718 | 350  | 0.7960          | 87.5576  |
| 0.7516        | 0.7198 | 375  | 0.7844          | 88.9401  |
| 0.9119        | 0.7678 | 400  | 0.8116          | 87.4424  |
| 0.7478        | 0.8157 | 425  | 0.7593          | 79.0323  |
| 0.7125        | 0.8637 | 450  | 0.7280          | 84.2166  |
| 0.8235        | 0.9117 | 475  | 0.7171          | 88.9401  |
| 0.6975        | 0.9597 | 500  | 0.7029          | 74.8848  |
| 0.5599        | 1.0077 | 525  | 0.7060          | 76.6129  |
| 0.4681        | 1.0557 | 550  | 0.6891          | 100.8065 |
| 0.3496        | 1.1036 | 575  | 0.6995          | 104.9539 |
| 0.4196        | 1.1516 | 600  | 0.7102          | 82.4885  |
| 0.3884        | 1.1996 | 625  | 0.6856          | 104.7235 |
| 0.4788        | 1.2476 | 650  | 0.6745          | 81.6820  |
| 0.4237        | 1.2956 | 675  | 0.6722          | 81.9124  |
| 0.4001        | 1.3436 | 700  | 0.6740          | 83.2949  |
| 0.3909        | 1.3916 | 725  | 0.6823          | 71.8894  |
| 0.3435        | 1.4395 | 750  | 0.6934          | 75.1152  |
| 0.344         | 1.4875 | 775  | 0.6810          | 72.0046  |
| 0.3071        | 1.5355 | 800  | 0.6704          | 71.1982  |
| 0.3392        | 1.5835 | 825  | 0.6589          | 88.3641  |
| 0.3742        | 1.6315 | 850  | 0.6532          | 77.9954  |
| 0.4153        | 1.6795 | 875  | 0.6363          | 79.8387  |
| 0.3416        | 1.7274 | 900  | 0.6560          | 79.4931  |
| 0.3121        | 1.7754 | 925  | 0.6320          | 82.0276  |
| 0.2986        | 1.8234 | 950  | 0.6447          | 76.9585  |
| 0.3761        | 1.8714 | 975  | 0.6420          | 75.8065  |
| 0.4394        | 1.9194 | 1000 | 0.6234          | 77.5346  |
| 0.3094        | 1.9674 | 1025 | 0.6430          | 81.5668  |
| 0.3468        | 2.0154 | 1050 | 0.6266          | 78.5714  |
| 0.25          | 2.0633 | 1075 | 0.6251          | 79.0323  |
| 0.1969        | 2.1113 | 1100 | 0.6337          | 81.2212  |
| 0.157         | 2.1593 | 1125 | 0.6367          | 76.8433  |
| 0.2118        | 2.2073 | 1150 | 0.6414          | 74.4240  |
| 0.2207        | 2.2553 | 1175 | 0.6345          | 77.4194  |
| 0.1965        | 2.3033 | 1200 | 0.6414          | 76.9585  |
| 0.1959        | 2.3512 | 1225 | 0.6322          | 79.6083  |
| 0.1668        | 2.3992 | 1250 | 0.6394          | 81.5668  |
| 0.2128        | 2.4472 | 1275 | 0.6361          | 80.4147  |
| 0.173         | 2.4952 | 1300 | 0.6322          | 74.8848  |
| 0.152         | 2.5432 | 1325 | 0.6312          | 73.3871  |
| 0.1897        | 2.5912 | 1350 | 0.6334          | 79.0323  |
| 0.1666        | 2.6392 | 1375 | 0.6339          | 81.1060  |
| 0.202         | 2.6871 | 1400 | 0.6283          | 77.9954  |
| 0.1511        | 2.7351 | 1425 | 0.6296          | 80.8756  |
| 0.1616        | 2.7831 | 1450 | 0.6313          | 80.4147  |
| 0.1482        | 2.8311 | 1475 | 0.6289          | 80.5300  |
| 0.1672        | 2.8791 | 1500 | 0.6283          | 80.0691  |


### Framework versions

- Transformers 4.52.4
- Pytorch 2.6.0+cu124
- Datasets 3.6.0
- Tokenizers 0.21.1