jwh1449 commited on
Commit
c3b03c2
·
verified ·
1 Parent(s): 77291a3

End of training

Browse files
README.md CHANGED
@@ -21,8 +21,8 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the AIhub_foreign_dataset4 dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 1.0461
25
- - Cer: 38.7247
26
 
27
  ## Model description
28
 
@@ -41,12 +41,12 @@ More information needed
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
- - learning_rate: 0.0001
45
  - train_batch_size: 16
46
  - eval_batch_size: 8
47
  - seed: 42
48
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
- - lr_scheduler_type: linear
50
  - lr_scheduler_warmup_steps: 500
51
  - training_steps: 6000
52
  - mixed_precision_training: Native AMP
@@ -55,17 +55,17 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Cer |
57
  |:-------------:|:------:|:----:|:---------------:|:-------:|
58
- | 1.1331 | 0.7299 | 1000 | 1.0860 | 51.9778 |
59
- | 1.0137 | 1.4599 | 2000 | 1.0169 | 52.3109 |
60
- | 0.6556 | 2.1898 | 3000 | 1.0076 | 46.6288 |
61
- | 0.7336 | 2.9197 | 4000 | 0.9763 | 44.7615 |
62
- | 0.5157 | 3.6496 | 5000 | 1.0181 | 40.7681 |
63
- | 0.3103 | 4.3796 | 6000 | 1.0461 | 38.7247 |
64
 
65
 
66
  ### Framework versions
67
 
68
  - Transformers 4.51.0.dev0
69
  - Pytorch 2.6.0+cu124
70
- - Datasets 3.4.1
71
  - Tokenizers 0.21.1
 
21
 
22
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the AIhub_foreign_dataset4 dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.9395
25
+ - Cer: 48.1817
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 2e-05
45
  - train_batch_size: 16
46
  - eval_batch_size: 8
47
  - seed: 42
48
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
49
+ - lr_scheduler_type: cosine
50
  - lr_scheduler_warmup_steps: 500
51
  - training_steps: 6000
52
  - mixed_precision_training: Native AMP
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Cer |
57
  |:-------------:|:------:|:----:|:---------------:|:-------:|
58
+ | 1.0479 | 0.7299 | 1000 | 0.9942 | 53.5695 |
59
+ | 0.983 | 1.4599 | 2000 | 0.9553 | 50.5824 |
60
+ | 0.7329 | 2.1898 | 3000 | 0.9432 | 56.3785 |
61
+ | 0.7979 | 2.9197 | 4000 | 0.9299 | 48.6001 |
62
+ | 0.6956 | 3.6496 | 5000 | 0.9368 | 47.2723 |
63
+ | 0.5971 | 4.3796 | 6000 | 0.9395 | 48.1817 |
64
 
65
 
66
  ### Framework versions
67
 
68
  - Transformers 4.51.0.dev0
69
  - Pytorch 2.6.0+cu124
70
+ - Datasets 3.5.0
71
  - Tokenizers 0.21.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec9af9bcccc317ef9cb9577fff1fa7f2d30b6fae0c4df202b38d347f7e5a383a
3
  size 290403936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ea36c228cebc740100e73f4a97f3a7fc887d3fb54eb9b97b31a60fec4836ba8
3
  size 290403936
runs/Mar28_15-51-54_f32a79138847/events.out.tfevents.1743177119.f32a79138847.1038.5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fab71ae114f048e243157b93cf9062f3bc35b7afcf43fa1a24adaec676473341
3
- size 58316
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f19411916e9d3b2f17019d1dd1838b2522ce7a7115abba660890c389486b4a33
3
+ size 58670