abedNa commited on
Commit
4a81eec
·
verified ·
1 Parent(s): 58cfbc3

Model save

Browse files
Files changed (1) hide show
  1. README.md +14 -14
README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
2
  library_name: peft
3
  license: apache-2.0
4
- base_model: openai/whisper-large-v3
5
  tags:
6
- - base_model:adapter:openai/whisper-large-v3
7
  - lora
8
  - transformers
9
  metrics:
@@ -18,11 +18,11 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # whisper-v4
20
 
21
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.3231
24
- - Wer Ortho: 0.8194
25
- - Wer: 0.7037
26
 
27
  ## Model description
28
 
@@ -42,24 +42,24 @@ More information needed
42
 
43
  The following hyperparameters were used during training:
44
  - learning_rate: 1e-05
45
- - train_batch_size: 4
46
- - eval_batch_size: 4
47
  - seed: 42
48
  - gradient_accumulation_steps: 16
49
- - total_train_batch_size: 64
50
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
- - lr_scheduler_type: constant_with_warmup
52
  - lr_scheduler_warmup_ratio: 0.1
53
- - lr_scheduler_warmup_steps: 500
54
- - num_epochs: 10
55
  - mixed_precision_training: Native AMP
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
60
  |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|
61
- | 1.3381 | 4.5664 | 100 | 1.3943 | 0.7563 | 0.6406 |
62
- | 1.2976 | 9.0944 | 200 | 1.3231 | 0.8194 | 0.7037 |
63
 
64
 
65
  ### Framework versions
 
1
  ---
2
  library_name: peft
3
  license: apache-2.0
4
+ base_model: ivrit-ai/whisper-large-v3-turbo
5
  tags:
6
+ - base_model:adapter:ivrit-ai/whisper-large-v3-turbo
7
  - lora
8
  - transformers
9
  metrics:
 
18
 
19
  # whisper-v4
20
 
21
+ This model is a fine-tuned version of [ivrit-ai/whisper-large-v3-turbo](https://huggingface.co/ivrit-ai/whisper-large-v3-turbo) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.3019
24
+ - Wer Ortho: 0.1547
25
+ - Wer: 0.1048
26
 
27
  ## Model description
28
 
 
42
 
43
  The following hyperparameters were used during training:
44
  - learning_rate: 1e-05
45
+ - train_batch_size: 8
46
+ - eval_batch_size: 8
47
  - seed: 42
48
  - gradient_accumulation_steps: 16
49
+ - total_train_batch_size: 128
50
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
+ - lr_scheduler_type: linear
52
  - lr_scheduler_warmup_ratio: 0.1
53
+ - lr_scheduler_warmup_steps: 800
54
+ - num_epochs: 5
55
  - mixed_precision_training: Native AMP
56
 
57
  ### Training results
58
 
59
  | Training Loss | Epoch | Step | Validation Loss | Wer Ortho | Wer |
60
  |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|
61
+ | 0.3076 | 1.9608 | 100 | 0.3021 | 0.1497 | 0.0996 |
62
+ | 0.308 | 3.9216 | 200 | 0.3019 | 0.1547 | 0.1048 |
63
 
64
 
65
  ### Framework versions