Bgeorge commited on
Commit
fef7966
·
verified ·
1 Parent(s): 1554cdb

End of training

Browse files
Files changed (1) hide show
  1. README.md +29 -32
README.md CHANGED
@@ -1,11 +1,14 @@
1
  ---
2
  library_name: transformers
3
- license: apache-2.0
4
- base_model: facebook/wav2vec2-base
5
  tags:
6
  - generated_from_trainer
7
  metrics:
8
  - accuracy
 
 
 
9
  model-index:
10
  - name: model_dialect
11
  results: []
@@ -16,10 +19,13 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # model_dialect
18
 
19
- This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.8299
22
- - Accuracy: 0.7182
 
 
 
23
 
24
  ## Model description
25
 
@@ -38,42 +44,33 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 4e-05
42
- - train_batch_size: 32
43
- - eval_batch_size: 32
44
  - seed: 42
45
- - gradient_accumulation_steps: 4
46
- - total_train_batch_size: 128
47
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
- - lr_scheduler_warmup_ratio: 0.1
50
- - num_epochs: 16
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Validation Loss | Accuracy |
55
- |:-------------:|:-------:|:----:|:---------------:|:--------:|
56
- | 6.4248 | 0.9455 | 13 | 1.5911 | 0.2748 |
57
- | 6.3186 | 1.9636 | 27 | 1.4754 | 0.4249 |
58
- | 5.5404 | 2.9818 | 41 | 1.2967 | 0.4850 |
59
- | 5.136 | 4.0 | 55 | 1.1871 | 0.5289 |
60
- | 4.8487 | 4.9455 | 68 | 1.1144 | 0.5612 |
61
- | 4.337 | 5.9636 | 82 | 1.0689 | 0.5820 |
62
- | 4.1023 | 6.9818 | 96 | 0.9931 | 0.6628 |
63
- | 3.762 | 8.0 | 110 | 0.9842 | 0.6305 |
64
- | 3.4974 | 8.9455 | 123 | 0.9379 | 0.6420 |
65
- | 3.4265 | 9.9636 | 137 | 0.9304 | 0.6559 |
66
- | 3.111 | 10.9818 | 151 | 0.8909 | 0.6767 |
67
- | 2.9471 | 12.0 | 165 | 0.8744 | 0.6859 |
68
- | 2.8061 | 12.9455 | 178 | 0.8533 | 0.7021 |
69
- | 2.7032 | 13.9636 | 192 | 0.8541 | 0.7067 |
70
- | 2.5328 | 14.9818 | 206 | 0.8299 | 0.7136 |
71
- | 2.5328 | 15.1273 | 208 | 0.8299 | 0.7182 |
72
 
73
 
74
  ### Framework versions
75
 
76
- - Transformers 4.46.0
77
  - Pytorch 2.4.0
78
  - Datasets 3.0.1
79
  - Tokenizers 0.20.0
 
1
  ---
2
  library_name: transformers
3
+ license: bsd-3-clause
4
+ base_model: MIT/ast-finetuned-audioset-10-10-0.4593
5
  tags:
6
  - generated_from_trainer
7
  metrics:
8
  - accuracy
9
+ - precision
10
+ - recall
11
+ - f1
12
  model-index:
13
  - name: model_dialect
14
  results: []
 
19
 
20
  # model_dialect
21
 
22
+ This model is a fine-tuned version of [MIT/ast-finetuned-audioset-10-10-0.4593](https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593) on an unknown dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.6309
25
+ - Accuracy: 0.7529
26
+ - Precision: 0.7608
27
+ - Recall: 0.7661
28
+ - F1: 0.7623
29
 
30
  ## Model description
31
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
+ - learning_rate: 5e-05
48
+ - train_batch_size: 8
49
+ - eval_batch_size: 8
50
  - seed: 42
51
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
 
 
52
  - lr_scheduler_type: linear
53
+ - num_epochs: 10
 
54
 
55
  ### Training results
56
 
57
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:---------:|:------:|:------:|
59
+ | 1.3499 | 1.0 | 217 | 1.4002 | 0.4065 | 0.2570 | 0.3730 | 0.2609 |
60
+ | 1.2635 | 2.0 | 434 | 1.0922 | 0.5242 | 0.6868 | 0.4970 | 0.4952 |
61
+ | 1.0379 | 3.0 | 651 | 1.0788 | 0.5335 | 0.7047 | 0.5137 | 0.5053 |
62
+ | 0.9536 | 4.0 | 868 | 0.8706 | 0.6513 | 0.6804 | 0.6631 | 0.6668 |
63
+ | 0.9321 | 5.0 | 1085 | 0.9052 | 0.6397 | 0.6693 | 0.6474 | 0.6307 |
64
+ | 0.9395 | 6.0 | 1302 | 0.8028 | 0.6767 | 0.7494 | 0.6692 | 0.6845 |
65
+ | 0.752 | 7.0 | 1519 | 0.7386 | 0.7344 | 0.7650 | 0.7386 | 0.7441 |
66
+ | 0.7984 | 8.0 | 1736 | 0.7100 | 0.7206 | 0.7370 | 0.7383 | 0.7323 |
67
+ | 0.5841 | 9.0 | 1953 | 0.6509 | 0.7413 | 0.7525 | 0.7494 | 0.7503 |
68
+ | 0.6678 | 10.0 | 2170 | 0.6309 | 0.7529 | 0.7608 | 0.7661 | 0.7623 |
 
 
 
 
 
 
69
 
70
 
71
  ### Framework versions
72
 
73
+ - Transformers 4.45.1
74
  - Pytorch 2.4.0
75
  - Datasets 3.0.1
76
  - Tokenizers 0.20.0