rbcurzon commited on
Commit
cce268f
·
verified ·
1 Parent(s): f2a4bc1

Model save

Browse files
Files changed (2) hide show
  1. README.md +21 -29
  2. generation_config.json +27 -13
README.md CHANGED
@@ -1,27 +1,14 @@
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: openai/whisper-medium
5
  tags:
6
  - generated_from_trainer
7
- datasets:
8
- - rbcurzon/ph_dialect_asr
9
  metrics:
10
  - wer
11
  model-index:
12
  - name: whisper-medium-test
13
- results:
14
- - task:
15
- name: Automatic Speech Recognition
16
- type: automatic-speech-recognition
17
- dataset:
18
- name: rbcurzon/ph_dialect_asr all
19
- type: rbcurzon/ph_dialect_asr
20
- args: all
21
- metrics:
22
- - name: Wer
23
- type: wer
24
- value: 0.12251221112488153
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -29,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
29
 
30
  # whisper-medium-test
31
 
32
- This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the rbcurzon/ph_dialect_asr all dataset.
33
  It achieves the following results on the evaluation set:
34
- - Loss: 0.2824
35
- - Wer: 0.1225
36
 
37
  ## Model description
38
 
@@ -55,26 +42,31 @@ The following hyperparameters were used during training:
55
  - train_batch_size: 4
56
  - eval_batch_size: 4
57
  - seed: 42
 
 
58
  - gradient_accumulation_steps: 4
59
- - total_train_batch_size: 16
60
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
 
61
  - lr_scheduler_type: linear
62
  - lr_scheduler_warmup_steps: 500
63
- - training_steps: 3000
64
  - mixed_precision_training: Native AMP
65
 
66
  ### Training results
67
 
68
- | Training Loss | Epoch | Step | Validation Loss | Wer |
69
- |:-------------:|:------:|:----:|:---------------:|:------:|
70
- | 0.1981 | 1.6453 | 1000 | 0.2715 | 0.1455 |
71
- | 0.0236 | 3.2897 | 2000 | 0.2709 | 0.1267 |
72
- | 0.0066 | 4.9350 | 3000 | 0.2824 | 0.1225 |
 
 
73
 
74
 
75
  ### Framework versions
76
 
77
- - Transformers 4.54.0.dev0
78
- - Pytorch 2.7.1+cu126
79
- - Datasets 4.0.0
80
  - Tokenizers 0.21.2
 
1
  ---
2
  library_name: transformers
3
  license: apache-2.0
4
+ base_model: openai/whisper-small
5
  tags:
6
  - generated_from_trainer
 
 
7
  metrics:
8
  - wer
9
  model-index:
10
  - name: whisper-medium-test
11
+ results: []
 
 
 
 
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
16
 
17
  # whisper-medium-test
18
 
19
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.4044
22
+ - Wer: 0.1675
23
 
24
  ## Model description
25
 
 
42
  - train_batch_size: 4
43
  - eval_batch_size: 4
44
  - seed: 42
45
+ - distributed_type: multi-GPU
46
+ - num_devices: 2
47
  - gradient_accumulation_steps: 4
48
+ - total_train_batch_size: 32
49
+ - total_eval_batch_size: 8
50
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
  - lr_scheduler_type: linear
52
  - lr_scheduler_warmup_steps: 500
53
+ - training_steps: 5000
54
  - mixed_precision_training: Native AMP
55
 
56
  ### Training results
57
 
58
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
59
+ |:-------------:|:-------:|:----:|:---------------:|:------:|
60
+ | 0.2023 | 2.9607 | 1000 | 0.3294 | 0.1874 |
61
+ | 0.0309 | 5.9192 | 2000 | 0.3414 | 0.1717 |
62
+ | 0.0037 | 8.8777 | 3000 | 0.3782 | 0.1708 |
63
+ | 0.0018 | 11.8362 | 4000 | 0.3964 | 0.1680 |
64
+ | 0.0013 | 14.7947 | 5000 | 0.4044 | 0.1675 |
65
 
66
 
67
  ### Framework versions
68
 
69
+ - Transformers 4.56.0.dev0
70
+ - Pytorch 2.6.0+cu124
71
+ - Datasets 3.6.0
72
  - Tokenizers 0.21.2
generation_config.json CHANGED
@@ -1,28 +1,44 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 13,
5
- 15
6
  ],
7
  [
8
- 15,
 
 
 
 
 
 
 
 
9
  4
10
  ],
11
  [
12
- 15,
13
- 15
14
  ],
15
  [
16
- 16,
17
- 1
18
  ],
19
  [
20
- 20,
21
  0
22
  ],
23
  [
24
- 23,
25
- 4
 
 
 
 
 
 
 
 
26
  ]
27
  ],
28
  "begin_suppress_tokens": [
@@ -226,8 +242,6 @@
226
  49870,
227
  50254,
228
  50258,
229
- 50358,
230
- 50359,
231
  50360,
232
  50361,
233
  50362
@@ -237,5 +251,5 @@
237
  "transcribe": 50359,
238
  "translate": 50358
239
  },
240
- "transformers_version": "4.54.0.dev0"
241
  }
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 5,
5
+ 3
6
  ],
7
  [
8
+ 5,
9
+ 9
10
+ ],
11
+ [
12
+ 8,
13
+ 0
14
+ ],
15
+ [
16
+ 8,
17
  4
18
  ],
19
  [
20
+ 8,
21
+ 7
22
  ],
23
  [
24
+ 8,
25
+ 8
26
  ],
27
  [
28
+ 9,
29
  0
30
  ],
31
  [
32
+ 9,
33
+ 7
34
+ ],
35
+ [
36
+ 9,
37
+ 9
38
+ ],
39
+ [
40
+ 10,
41
+ 5
42
  ]
43
  ],
44
  "begin_suppress_tokens": [
 
242
  49870,
243
  50254,
244
  50258,
 
 
245
  50360,
246
  50361,
247
  50362
 
251
  "transcribe": 50359,
252
  "translate": 50358
253
  },
254
+ "transformers_version": "4.56.0.dev0"
255
  }