JaratX commited on
Commit
e0cf9f1
·
1 Parent(s): 730d32c

Add/Update model artifacts

Browse files
Files changed (31) hide show
  1. agent/acoustic/acoustic_scaler.joblib +3 -0
  2. agent/acoustic/randomforest_acoustic_model.joblib +3 -0
  3. agent/acoustic/svm_acoustic_model.joblib +3 -0
  4. agent/text/best_model/config.json +45 -0
  5. agent/text/best_model/pytorch_model.bin +3 -0
  6. agent/text/best_model/training_args.bin +3 -0
  7. client/acoustic/acoustic_scaler.joblib +3 -0
  8. client/acoustic/randomforest_acoustic_model.joblib +3 -0
  9. client/acoustic/svm_acoustic_model.joblib +3 -0
  10. client/text/best_model/config.json +45 -0
  11. client/text/best_model/pytorch_model.bin +3 -0
  12. client/text/best_model/training_args.bin +3 -0
  13. transcription/w2v-bert-darija-finetuned-clean/added_tokens.json +4 -0
  14. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/config.json +82 -0
  15. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/preprocessor_config.json +10 -0
  16. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/rng_state.pth +3 -0
  17. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/scheduler.pt +3 -0
  18. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/trainer_state.json +220 -0
  19. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/training_args.bin +3 -0
  20. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/config.json +82 -0
  21. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/preprocessor_config.json +10 -0
  22. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/rng_state.pth +3 -0
  23. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/scheduler.pt +3 -0
  24. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/trainer_state.json +237 -0
  25. transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/training_args.bin +3 -0
  26. transcription/w2v-bert-darija-finetuned-clean/config.json +82 -0
  27. transcription/w2v-bert-darija-finetuned-clean/preprocessor_config.json +11 -0
  28. transcription/w2v-bert-darija-finetuned-clean/special_tokens_map.json +6 -0
  29. transcription/w2v-bert-darija-finetuned-clean/tokenizer_config.json +48 -0
  30. transcription/w2v-bert-darija-finetuned-clean/training_args.bin +3 -0
  31. transcription/w2v-bert-darija-finetuned-clean/vocab.json +58 -0
agent/acoustic/acoustic_scaler.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a293589603ee75ad1b4d3c522db0eadc8536bd22d8c1cdb58e96ec6400523a65
3
+ size 1527
agent/acoustic/randomforest_acoustic_model.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82c81d3adaaa40046c9402898bc7f675b90b5f0d02cf655817af0acb20efbc33
3
+ size 5724433
agent/acoustic/svm_acoustic_model.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3e0e7f53578abfb47d8babf69ed7cc831a5a0b7208a51014052dbdec07dbb27
3
+ size 579555
agent/text/best_model/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "SI2M-Lab/DarijaBERT",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "cls_token": "[CLS]",
9
+ "do_lower_case": true,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "aggressive",
16
+ "1": "courtois",
17
+ "2": "sec"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "aggressive": 0,
23
+ "courtois": 1,
24
+ "sec": 2
25
+ },
26
+ "layer_norm_eps": 1e-12,
27
+ "mask_token": "[MASK]",
28
+ "max_len": 128,
29
+ "max_position_embeddings": 512,
30
+ "model_max_length": 128,
31
+ "model_type": "bert",
32
+ "num_attention_heads": 12,
33
+ "num_hidden_layers": 12,
34
+ "pad_token": "[PAD]",
35
+ "pad_token_id": 0,
36
+ "position_embedding_type": "absolute",
37
+ "problem_type": "single_label_classification",
38
+ "sep_token": "[SEP]",
39
+ "torch_dtype": "float32",
40
+ "transformers_version": "4.43.3",
41
+ "type_vocab_size": 2,
42
+ "unk_token": "[UNK]",
43
+ "use_cache": true,
44
+ "vocab_size": 80000
45
+ }
agent/text/best_model/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b07682e083d17d041bdbc77bdc19a2ae576eaf349fdc58cf7e30820818096ec
3
+ size 590003438
agent/text/best_model/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebc421668ec59fcd0f8e105fe824fa99e64cae00e5192669714e2a6da8bbbba0
3
+ size 5368
client/acoustic/acoustic_scaler.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:161d0c8cdcfdc7c3fc229606cabadb0c75c7e446a5b4a77cf1dab7d4253c579b
3
+ size 1527
client/acoustic/randomforest_acoustic_model.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18fab4d53235941d54e10742717be6c151a2232d207a2b2b10bb670c4380a7a4
3
+ size 4156625
client/acoustic/svm_acoustic_model.joblib ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6aa4d84363bd11e7a8c321677e404a637ce6c321e2d0ea40b39a0f0dc7bb7066
3
+ size 437459
client/text/best_model/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "SI2M-Lab/DarijaBERT",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "cls_token": "[CLS]",
9
+ "do_lower_case": true,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "id2label": {
15
+ "0": "content",
16
+ "1": "neutre",
17
+ "2": "tres mecontent"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "content": 0,
23
+ "neutre": 1,
24
+ "tres mecontent": 2
25
+ },
26
+ "layer_norm_eps": 1e-12,
27
+ "mask_token": "[MASK]",
28
+ "max_len": 128,
29
+ "max_position_embeddings": 512,
30
+ "model_max_length": 128,
31
+ "model_type": "bert",
32
+ "num_attention_heads": 12,
33
+ "num_hidden_layers": 12,
34
+ "pad_token": "[PAD]",
35
+ "pad_token_id": 0,
36
+ "position_embedding_type": "absolute",
37
+ "problem_type": "single_label_classification",
38
+ "sep_token": "[SEP]",
39
+ "torch_dtype": "float32",
40
+ "transformers_version": "4.43.3",
41
+ "type_vocab_size": 2,
42
+ "unk_token": "[UNK]",
43
+ "use_cache": true,
44
+ "vocab_size": 80000
45
+ }
client/text/best_model/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a97f23ea9c953d26eb314bbcf11ab292eaef85577353e4f71358ffdcd79b2929
3
+ size 590003438
client/text/best_model/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf2a6b9904c864e3fb34ed5b56de4cc2aa5627737d86fe02bb78a4067e59fe5c
3
+ size 5368
transcription/w2v-bert-darija-finetuned-clean/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "</s>": 57,
3
+ "<s>": 56
4
+ }
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/config.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/w2v-bert-2.0",
3
+ "activation_dropout": 0.0,
4
+ "adapter_act": "relu",
5
+ "adapter_kernel_size": 3,
6
+ "adapter_stride": 2,
7
+ "add_adapter": true,
8
+ "apply_spec_augment": false,
9
+ "architectures": [
10
+ "Wav2Vec2BertForCTC"
11
+ ],
12
+ "attention_dropout": 0.0,
13
+ "bos_token_id": 1,
14
+ "classifier_proj_size": 768,
15
+ "codevector_dim": 768,
16
+ "conformer_conv_dropout": 0.1,
17
+ "contrastive_logits_temperature": 0.1,
18
+ "conv_depthwise_kernel_size": 31,
19
+ "ctc_loss_reduction": "mean",
20
+ "ctc_zero_infinity": false,
21
+ "diversity_loss_weight": 0.1,
22
+ "eos_token_id": 2,
23
+ "feat_proj_dropout": 0.0,
24
+ "feat_quantizer_dropout": 0.0,
25
+ "feature_projection_input_dim": 160,
26
+ "final_dropout": 0.1,
27
+ "hidden_act": "swish",
28
+ "hidden_dropout": 0.0,
29
+ "hidden_size": 1024,
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 4096,
32
+ "layer_norm_eps": 1e-05,
33
+ "layerdrop": 0.0,
34
+ "left_max_position_embeddings": 64,
35
+ "mask_feature_length": 10,
36
+ "mask_feature_min_masks": 0,
37
+ "mask_feature_prob": 0.0,
38
+ "mask_time_length": 10,
39
+ "mask_time_min_masks": 2,
40
+ "mask_time_prob": 0.0,
41
+ "max_source_positions": 5000,
42
+ "model_type": "wav2vec2-bert",
43
+ "num_adapter_layers": 1,
44
+ "num_attention_heads": 16,
45
+ "num_codevector_groups": 2,
46
+ "num_codevectors_per_group": 320,
47
+ "num_hidden_layers": 24,
48
+ "num_negatives": 100,
49
+ "output_hidden_size": 1024,
50
+ "pad_token_id": 55,
51
+ "position_embeddings_type": "relative_key",
52
+ "proj_codevector_dim": 768,
53
+ "right_max_position_embeddings": 8,
54
+ "rotary_embedding_base": 10000,
55
+ "tdnn_dilation": [
56
+ 1,
57
+ 2,
58
+ 3,
59
+ 1,
60
+ 1
61
+ ],
62
+ "tdnn_dim": [
63
+ 512,
64
+ 512,
65
+ 512,
66
+ 512,
67
+ 1500
68
+ ],
69
+ "tdnn_kernel": [
70
+ 5,
71
+ 3,
72
+ 3,
73
+ 1,
74
+ 1
75
+ ],
76
+ "torch_dtype": "float32",
77
+ "transformers_version": "4.46.3",
78
+ "use_intermediate_ffn_before_adapter": false,
79
+ "use_weighted_layer_sum": false,
80
+ "vocab_size": 58,
81
+ "xvector_output_dim": 512
82
+ }
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/preprocessor_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "feature_extractor_type": "SeamlessM4TFeatureExtractor",
3
+ "feature_size": 80,
4
+ "num_mel_bins": 80,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "return_attention_mask": true,
8
+ "sampling_rate": 16000,
9
+ "stride": 2
10
+ }
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2a812a278c72cb787298245f122ab589285b867cb9fb13812819e443341549a
3
+ size 14244
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a83b4666c22523603f0a60b34d13d8692110ae1c6107f863de3447d27acb008b
3
+ size 1064
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/trainer_state.json ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 10.99630996309963,
5
+ "eval_steps": 500,
6
+ "global_step": 1490,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.996309963099631,
13
+ "grad_norm": 6.060778617858887,
14
+ "learning_rate": 1.3100000000000002e-05,
15
+ "loss": 4.0745,
16
+ "step": 135
17
+ },
18
+ {
19
+ "epoch": 0.996309963099631,
20
+ "eval_loss": 2.8919060230255127,
21
+ "eval_pen_wer": 0.9288469971028115,
22
+ "eval_runtime": 35.4574,
23
+ "eval_samples_per_second": 7.643,
24
+ "eval_steps_per_second": 0.959,
25
+ "eval_wer": 0.9983822042467139,
26
+ "step": 135
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "grad_norm": 32.26631164550781,
31
+ "learning_rate": 2.6700000000000002e-05,
32
+ "loss": 1.9077,
33
+ "step": 271
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_loss": 1.279911756515503,
38
+ "eval_pen_wer": 0.4244393978114909,
39
+ "eval_runtime": 36.3495,
40
+ "eval_samples_per_second": 7.455,
41
+ "eval_steps_per_second": 0.935,
42
+ "eval_wer": 0.6604651162790698,
43
+ "step": 271
44
+ },
45
+ {
46
+ "epoch": 2.9963099630996313,
47
+ "grad_norm": 5.846123218536377,
48
+ "learning_rate": 4.02e-05,
49
+ "loss": 1.2007,
50
+ "step": 406
51
+ },
52
+ {
53
+ "epoch": 2.9963099630996313,
54
+ "eval_loss": 1.1114612817764282,
55
+ "eval_pen_wer": 0.36216589798996246,
56
+ "eval_runtime": 36.7217,
57
+ "eval_samples_per_second": 7.38,
58
+ "eval_steps_per_second": 0.926,
59
+ "eval_wer": 0.5937984496124031,
60
+ "step": 406
61
+ },
62
+ {
63
+ "epoch": 4.0,
64
+ "grad_norm": 7.9398193359375,
65
+ "learning_rate": 4.8303571428571434e-05,
66
+ "loss": 1.0249,
67
+ "step": 542
68
+ },
69
+ {
70
+ "epoch": 4.0,
71
+ "eval_loss": 1.0737940073013306,
72
+ "eval_pen_wer": 0.34027092257628144,
73
+ "eval_runtime": 36.6597,
74
+ "eval_samples_per_second": 7.392,
75
+ "eval_steps_per_second": 0.927,
76
+ "eval_wer": 0.5542298618132794,
77
+ "step": 542
78
+ },
79
+ {
80
+ "epoch": 4.996309963099631,
81
+ "grad_norm": 5.927522659301758,
82
+ "learning_rate": 4.2276785714285714e-05,
83
+ "loss": 0.8813,
84
+ "step": 677
85
+ },
86
+ {
87
+ "epoch": 4.996309963099631,
88
+ "eval_loss": 1.0277714729309082,
89
+ "eval_pen_wer": 0.34514653099587383,
90
+ "eval_runtime": 36.4973,
91
+ "eval_samples_per_second": 7.425,
92
+ "eval_steps_per_second": 0.932,
93
+ "eval_wer": 0.5530165149983148,
94
+ "step": 677
95
+ },
96
+ {
97
+ "epoch": 6.0,
98
+ "grad_norm": 11.025650978088379,
99
+ "learning_rate": 3.620535714285714e-05,
100
+ "loss": 0.7769,
101
+ "step": 813
102
+ },
103
+ {
104
+ "epoch": 6.0,
105
+ "eval_loss": 0.9775798320770264,
106
+ "eval_pen_wer": 0.32710190850231297,
107
+ "eval_runtime": 36.7174,
108
+ "eval_samples_per_second": 7.381,
109
+ "eval_steps_per_second": 0.926,
110
+ "eval_wer": 0.5332659251769464,
111
+ "step": 813
112
+ },
113
+ {
114
+ "epoch": 6.996309963099631,
115
+ "grad_norm": 6.112392425537109,
116
+ "learning_rate": 3.0178571428571427e-05,
117
+ "loss": 0.7,
118
+ "step": 948
119
+ },
120
+ {
121
+ "epoch": 6.996309963099631,
122
+ "eval_loss": 0.9915932416915894,
123
+ "eval_pen_wer": 0.3124926266483397,
124
+ "eval_runtime": 36.3976,
125
+ "eval_samples_per_second": 7.446,
126
+ "eval_steps_per_second": 0.934,
127
+ "eval_wer": 0.5078530502190766,
128
+ "step": 948
129
+ },
130
+ {
131
+ "epoch": 8.0,
132
+ "grad_norm": 17.379560470581055,
133
+ "learning_rate": 2.4107142857142858e-05,
134
+ "loss": 0.6065,
135
+ "step": 1084
136
+ },
137
+ {
138
+ "epoch": 8.0,
139
+ "eval_loss": 0.992123544216156,
140
+ "eval_pen_wer": 0.31970698851386425,
141
+ "eval_runtime": 36.1353,
142
+ "eval_samples_per_second": 7.5,
143
+ "eval_steps_per_second": 0.941,
144
+ "eval_wer": 0.5029322548028311,
145
+ "step": 1084
146
+ },
147
+ {
148
+ "epoch": 8.99630996309963,
149
+ "grad_norm": 7.368852615356445,
150
+ "learning_rate": 1.8080357142857144e-05,
151
+ "loss": 0.5375,
152
+ "step": 1219
153
+ },
154
+ {
155
+ "epoch": 8.99630996309963,
156
+ "eval_loss": 0.984150230884552,
157
+ "eval_pen_wer": 0.302135588075932,
158
+ "eval_runtime": 36.9038,
159
+ "eval_samples_per_second": 7.343,
160
+ "eval_steps_per_second": 0.921,
161
+ "eval_wer": 0.4912032355915066,
162
+ "step": 1219
163
+ },
164
+ {
165
+ "epoch": 10.0,
166
+ "grad_norm": 9.9882230758667,
167
+ "learning_rate": 1.2008928571428573e-05,
168
+ "loss": 0.4681,
169
+ "step": 1355
170
+ },
171
+ {
172
+ "epoch": 10.0,
173
+ "eval_loss": 1.0245816707611084,
174
+ "eval_pen_wer": 0.3019771813401745,
175
+ "eval_runtime": 36.9691,
176
+ "eval_samples_per_second": 7.33,
177
+ "eval_steps_per_second": 0.92,
178
+ "eval_wer": 0.4895180316818335,
179
+ "step": 1355
180
+ },
181
+ {
182
+ "epoch": 10.99630996309963,
183
+ "grad_norm": 5.356689453125,
184
+ "learning_rate": 5.982142857142858e-06,
185
+ "loss": 0.415,
186
+ "step": 1490
187
+ },
188
+ {
189
+ "epoch": 10.99630996309963,
190
+ "eval_loss": 1.0347157716751099,
191
+ "eval_pen_wer": 0.29928239999524137,
192
+ "eval_runtime": 36.6987,
193
+ "eval_samples_per_second": 7.384,
194
+ "eval_steps_per_second": 0.926,
195
+ "eval_wer": 0.48230535894843274,
196
+ "step": 1490
197
+ }
198
+ ],
199
+ "logging_steps": 500,
200
+ "max_steps": 1620,
201
+ "num_input_tokens_seen": 0,
202
+ "num_train_epochs": 12,
203
+ "save_steps": 500,
204
+ "stateful_callbacks": {
205
+ "TrainerControl": {
206
+ "args": {
207
+ "should_epoch_stop": false,
208
+ "should_evaluate": false,
209
+ "should_log": false,
210
+ "should_save": true,
211
+ "should_training_stop": false
212
+ },
213
+ "attributes": {}
214
+ }
215
+ },
216
+ "total_flos": 8.432449882085846e+18,
217
+ "train_batch_size": 4,
218
+ "trial_name": null,
219
+ "trial_params": null
220
+ }
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1490/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2793aaf5af1a42cb015bfef656c097f2acac87a48fffa9d3175eaeceb802e54
3
+ size 5304
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/config.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/w2v-bert-2.0",
3
+ "activation_dropout": 0.0,
4
+ "adapter_act": "relu",
5
+ "adapter_kernel_size": 3,
6
+ "adapter_stride": 2,
7
+ "add_adapter": true,
8
+ "apply_spec_augment": false,
9
+ "architectures": [
10
+ "Wav2Vec2BertForCTC"
11
+ ],
12
+ "attention_dropout": 0.0,
13
+ "bos_token_id": 1,
14
+ "classifier_proj_size": 768,
15
+ "codevector_dim": 768,
16
+ "conformer_conv_dropout": 0.1,
17
+ "contrastive_logits_temperature": 0.1,
18
+ "conv_depthwise_kernel_size": 31,
19
+ "ctc_loss_reduction": "mean",
20
+ "ctc_zero_infinity": false,
21
+ "diversity_loss_weight": 0.1,
22
+ "eos_token_id": 2,
23
+ "feat_proj_dropout": 0.0,
24
+ "feat_quantizer_dropout": 0.0,
25
+ "feature_projection_input_dim": 160,
26
+ "final_dropout": 0.1,
27
+ "hidden_act": "swish",
28
+ "hidden_dropout": 0.0,
29
+ "hidden_size": 1024,
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 4096,
32
+ "layer_norm_eps": 1e-05,
33
+ "layerdrop": 0.0,
34
+ "left_max_position_embeddings": 64,
35
+ "mask_feature_length": 10,
36
+ "mask_feature_min_masks": 0,
37
+ "mask_feature_prob": 0.0,
38
+ "mask_time_length": 10,
39
+ "mask_time_min_masks": 2,
40
+ "mask_time_prob": 0.0,
41
+ "max_source_positions": 5000,
42
+ "model_type": "wav2vec2-bert",
43
+ "num_adapter_layers": 1,
44
+ "num_attention_heads": 16,
45
+ "num_codevector_groups": 2,
46
+ "num_codevectors_per_group": 320,
47
+ "num_hidden_layers": 24,
48
+ "num_negatives": 100,
49
+ "output_hidden_size": 1024,
50
+ "pad_token_id": 55,
51
+ "position_embeddings_type": "relative_key",
52
+ "proj_codevector_dim": 768,
53
+ "right_max_position_embeddings": 8,
54
+ "rotary_embedding_base": 10000,
55
+ "tdnn_dilation": [
56
+ 1,
57
+ 2,
58
+ 3,
59
+ 1,
60
+ 1
61
+ ],
62
+ "tdnn_dim": [
63
+ 512,
64
+ 512,
65
+ 512,
66
+ 512,
67
+ 1500
68
+ ],
69
+ "tdnn_kernel": [
70
+ 5,
71
+ 3,
72
+ 3,
73
+ 1,
74
+ 1
75
+ ],
76
+ "torch_dtype": "float32",
77
+ "transformers_version": "4.46.3",
78
+ "use_intermediate_ffn_before_adapter": false,
79
+ "use_weighted_layer_sum": false,
80
+ "vocab_size": 58,
81
+ "xvector_output_dim": 512
82
+ }
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/preprocessor_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "feature_extractor_type": "SeamlessM4TFeatureExtractor",
3
+ "feature_size": 80,
4
+ "num_mel_bins": 80,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "return_attention_mask": true,
8
+ "sampling_rate": 16000,
9
+ "stride": 2
10
+ }
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a995beb400c6c1030bc370a69431ecc53b71182f594e2d6947b1b9a8e2bf06a7
3
+ size 14244
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8678820546c1a63031103789e84d8dddaf3251d385b714efbe8413d1d83de7f5
3
+ size 1064
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/trainer_state.json ADDED
@@ -0,0 +1,237 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 11.955719557195572,
5
+ "eval_steps": 500,
6
+ "global_step": 1620,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.996309963099631,
13
+ "grad_norm": 6.060778617858887,
14
+ "learning_rate": 1.3100000000000002e-05,
15
+ "loss": 4.0745,
16
+ "step": 135
17
+ },
18
+ {
19
+ "epoch": 0.996309963099631,
20
+ "eval_loss": 2.8919060230255127,
21
+ "eval_pen_wer": 0.9288469971028115,
22
+ "eval_runtime": 35.4574,
23
+ "eval_samples_per_second": 7.643,
24
+ "eval_steps_per_second": 0.959,
25
+ "eval_wer": 0.9983822042467139,
26
+ "step": 135
27
+ },
28
+ {
29
+ "epoch": 2.0,
30
+ "grad_norm": 32.26631164550781,
31
+ "learning_rate": 2.6700000000000002e-05,
32
+ "loss": 1.9077,
33
+ "step": 271
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_loss": 1.279911756515503,
38
+ "eval_pen_wer": 0.4244393978114909,
39
+ "eval_runtime": 36.3495,
40
+ "eval_samples_per_second": 7.455,
41
+ "eval_steps_per_second": 0.935,
42
+ "eval_wer": 0.6604651162790698,
43
+ "step": 271
44
+ },
45
+ {
46
+ "epoch": 2.9963099630996313,
47
+ "grad_norm": 5.846123218536377,
48
+ "learning_rate": 4.02e-05,
49
+ "loss": 1.2007,
50
+ "step": 406
51
+ },
52
+ {
53
+ "epoch": 2.9963099630996313,
54
+ "eval_loss": 1.1114612817764282,
55
+ "eval_pen_wer": 0.36216589798996246,
56
+ "eval_runtime": 36.7217,
57
+ "eval_samples_per_second": 7.38,
58
+ "eval_steps_per_second": 0.926,
59
+ "eval_wer": 0.5937984496124031,
60
+ "step": 406
61
+ },
62
+ {
63
+ "epoch": 4.0,
64
+ "grad_norm": 7.9398193359375,
65
+ "learning_rate": 4.8303571428571434e-05,
66
+ "loss": 1.0249,
67
+ "step": 542
68
+ },
69
+ {
70
+ "epoch": 4.0,
71
+ "eval_loss": 1.0737940073013306,
72
+ "eval_pen_wer": 0.34027092257628144,
73
+ "eval_runtime": 36.6597,
74
+ "eval_samples_per_second": 7.392,
75
+ "eval_steps_per_second": 0.927,
76
+ "eval_wer": 0.5542298618132794,
77
+ "step": 542
78
+ },
79
+ {
80
+ "epoch": 4.996309963099631,
81
+ "grad_norm": 5.927522659301758,
82
+ "learning_rate": 4.2276785714285714e-05,
83
+ "loss": 0.8813,
84
+ "step": 677
85
+ },
86
+ {
87
+ "epoch": 4.996309963099631,
88
+ "eval_loss": 1.0277714729309082,
89
+ "eval_pen_wer": 0.34514653099587383,
90
+ "eval_runtime": 36.4973,
91
+ "eval_samples_per_second": 7.425,
92
+ "eval_steps_per_second": 0.932,
93
+ "eval_wer": 0.5530165149983148,
94
+ "step": 677
95
+ },
96
+ {
97
+ "epoch": 6.0,
98
+ "grad_norm": 11.025650978088379,
99
+ "learning_rate": 3.620535714285714e-05,
100
+ "loss": 0.7769,
101
+ "step": 813
102
+ },
103
+ {
104
+ "epoch": 6.0,
105
+ "eval_loss": 0.9775798320770264,
106
+ "eval_pen_wer": 0.32710190850231297,
107
+ "eval_runtime": 36.7174,
108
+ "eval_samples_per_second": 7.381,
109
+ "eval_steps_per_second": 0.926,
110
+ "eval_wer": 0.5332659251769464,
111
+ "step": 813
112
+ },
113
+ {
114
+ "epoch": 6.996309963099631,
115
+ "grad_norm": 6.112392425537109,
116
+ "learning_rate": 3.0178571428571427e-05,
117
+ "loss": 0.7,
118
+ "step": 948
119
+ },
120
+ {
121
+ "epoch": 6.996309963099631,
122
+ "eval_loss": 0.9915932416915894,
123
+ "eval_pen_wer": 0.3124926266483397,
124
+ "eval_runtime": 36.3976,
125
+ "eval_samples_per_second": 7.446,
126
+ "eval_steps_per_second": 0.934,
127
+ "eval_wer": 0.5078530502190766,
128
+ "step": 948
129
+ },
130
+ {
131
+ "epoch": 8.0,
132
+ "grad_norm": 17.379560470581055,
133
+ "learning_rate": 2.4107142857142858e-05,
134
+ "loss": 0.6065,
135
+ "step": 1084
136
+ },
137
+ {
138
+ "epoch": 8.0,
139
+ "eval_loss": 0.992123544216156,
140
+ "eval_pen_wer": 0.31970698851386425,
141
+ "eval_runtime": 36.1353,
142
+ "eval_samples_per_second": 7.5,
143
+ "eval_steps_per_second": 0.941,
144
+ "eval_wer": 0.5029322548028311,
145
+ "step": 1084
146
+ },
147
+ {
148
+ "epoch": 8.99630996309963,
149
+ "grad_norm": 7.368852615356445,
150
+ "learning_rate": 1.8080357142857144e-05,
151
+ "loss": 0.5375,
152
+ "step": 1219
153
+ },
154
+ {
155
+ "epoch": 8.99630996309963,
156
+ "eval_loss": 0.984150230884552,
157
+ "eval_pen_wer": 0.302135588075932,
158
+ "eval_runtime": 36.9038,
159
+ "eval_samples_per_second": 7.343,
160
+ "eval_steps_per_second": 0.921,
161
+ "eval_wer": 0.4912032355915066,
162
+ "step": 1219
163
+ },
164
+ {
165
+ "epoch": 10.0,
166
+ "grad_norm": 9.9882230758667,
167
+ "learning_rate": 1.2008928571428573e-05,
168
+ "loss": 0.4681,
169
+ "step": 1355
170
+ },
171
+ {
172
+ "epoch": 10.0,
173
+ "eval_loss": 1.0245816707611084,
174
+ "eval_pen_wer": 0.3019771813401745,
175
+ "eval_runtime": 36.9691,
176
+ "eval_samples_per_second": 7.33,
177
+ "eval_steps_per_second": 0.92,
178
+ "eval_wer": 0.4895180316818335,
179
+ "step": 1355
180
+ },
181
+ {
182
+ "epoch": 10.99630996309963,
183
+ "grad_norm": 5.356689453125,
184
+ "learning_rate": 5.982142857142858e-06,
185
+ "loss": 0.415,
186
+ "step": 1490
187
+ },
188
+ {
189
+ "epoch": 10.99630996309963,
190
+ "eval_loss": 1.0347157716751099,
191
+ "eval_pen_wer": 0.29928239999524137,
192
+ "eval_runtime": 36.6987,
193
+ "eval_samples_per_second": 7.384,
194
+ "eval_steps_per_second": 0.926,
195
+ "eval_wer": 0.48230535894843274,
196
+ "step": 1490
197
+ },
198
+ {
199
+ "epoch": 11.955719557195572,
200
+ "grad_norm": 6.189606666564941,
201
+ "learning_rate": 1.7857142857142858e-07,
202
+ "loss": 0.3668,
203
+ "step": 1620
204
+ },
205
+ {
206
+ "epoch": 11.955719557195572,
207
+ "eval_loss": 1.0610431432724,
208
+ "eval_pen_wer": 0.2970119003941046,
209
+ "eval_runtime": 36.7697,
210
+ "eval_samples_per_second": 7.37,
211
+ "eval_steps_per_second": 0.925,
212
+ "eval_wer": 0.4813616447590158,
213
+ "step": 1620
214
+ }
215
+ ],
216
+ "logging_steps": 500,
217
+ "max_steps": 1620,
218
+ "num_input_tokens_seen": 0,
219
+ "num_train_epochs": 12,
220
+ "save_steps": 500,
221
+ "stateful_callbacks": {
222
+ "TrainerControl": {
223
+ "args": {
224
+ "should_epoch_stop": false,
225
+ "should_evaluate": false,
226
+ "should_log": false,
227
+ "should_save": true,
228
+ "should_training_stop": true
229
+ },
230
+ "attributes": {}
231
+ }
232
+ },
233
+ "total_flos": 9.176643836587086e+18,
234
+ "train_batch_size": 4,
235
+ "trial_name": null,
236
+ "trial_params": null
237
+ }
transcription/w2v-bert-darija-finetuned-clean/checkpoint-1620/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2793aaf5af1a42cb015bfef656c097f2acac87a48fffa9d3175eaeceb802e54
3
+ size 5304
transcription/w2v-bert-darija-finetuned-clean/config.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/w2v-bert-2.0",
3
+ "activation_dropout": 0.0,
4
+ "adapter_act": "relu",
5
+ "adapter_kernel_size": 3,
6
+ "adapter_stride": 2,
7
+ "add_adapter": true,
8
+ "apply_spec_augment": false,
9
+ "architectures": [
10
+ "Wav2Vec2BertForCTC"
11
+ ],
12
+ "attention_dropout": 0.0,
13
+ "bos_token_id": 1,
14
+ "classifier_proj_size": 768,
15
+ "codevector_dim": 768,
16
+ "conformer_conv_dropout": 0.1,
17
+ "contrastive_logits_temperature": 0.1,
18
+ "conv_depthwise_kernel_size": 31,
19
+ "ctc_loss_reduction": "mean",
20
+ "ctc_zero_infinity": false,
21
+ "diversity_loss_weight": 0.1,
22
+ "eos_token_id": 2,
23
+ "feat_proj_dropout": 0.0,
24
+ "feat_quantizer_dropout": 0.0,
25
+ "feature_projection_input_dim": 160,
26
+ "final_dropout": 0.1,
27
+ "hidden_act": "swish",
28
+ "hidden_dropout": 0.0,
29
+ "hidden_size": 1024,
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 4096,
32
+ "layer_norm_eps": 1e-05,
33
+ "layerdrop": 0.0,
34
+ "left_max_position_embeddings": 64,
35
+ "mask_feature_length": 10,
36
+ "mask_feature_min_masks": 0,
37
+ "mask_feature_prob": 0.0,
38
+ "mask_time_length": 10,
39
+ "mask_time_min_masks": 2,
40
+ "mask_time_prob": 0.0,
41
+ "max_source_positions": 5000,
42
+ "model_type": "wav2vec2-bert",
43
+ "num_adapter_layers": 1,
44
+ "num_attention_heads": 16,
45
+ "num_codevector_groups": 2,
46
+ "num_codevectors_per_group": 320,
47
+ "num_hidden_layers": 24,
48
+ "num_negatives": 100,
49
+ "output_hidden_size": 1024,
50
+ "pad_token_id": 55,
51
+ "position_embeddings_type": "relative_key",
52
+ "proj_codevector_dim": 768,
53
+ "right_max_position_embeddings": 8,
54
+ "rotary_embedding_base": 10000,
55
+ "tdnn_dilation": [
56
+ 1,
57
+ 2,
58
+ 3,
59
+ 1,
60
+ 1
61
+ ],
62
+ "tdnn_dim": [
63
+ 512,
64
+ 512,
65
+ 512,
66
+ 512,
67
+ 1500
68
+ ],
69
+ "tdnn_kernel": [
70
+ 5,
71
+ 3,
72
+ 3,
73
+ 1,
74
+ 1
75
+ ],
76
+ "torch_dtype": "float32",
77
+ "transformers_version": "4.46.3",
78
+ "use_intermediate_ffn_before_adapter": false,
79
+ "use_weighted_layer_sum": false,
80
+ "vocab_size": 58,
81
+ "xvector_output_dim": 512
82
+ }
transcription/w2v-bert-darija-finetuned-clean/preprocessor_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "feature_extractor_type": "SeamlessM4TFeatureExtractor",
3
+ "feature_size": 80,
4
+ "num_mel_bins": 80,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "processor_class": "Wav2Vec2BertProcessor",
8
+ "return_attention_mask": true,
9
+ "sampling_rate": 16000,
10
+ "stride": 2
11
+ }
transcription/w2v-bert-darija-finetuned-clean/special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "[PAD]",
5
+ "unk_token": "[UNK]"
6
+ }
transcription/w2v-bert-darija-finetuned-clean/tokenizer_config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "54": {
4
+ "content": "[UNK]",
5
+ "lstrip": true,
6
+ "normalized": false,
7
+ "rstrip": true,
8
+ "single_word": false,
9
+ "special": false
10
+ },
11
+ "55": {
12
+ "content": "[PAD]",
13
+ "lstrip": true,
14
+ "normalized": false,
15
+ "rstrip": true,
16
+ "single_word": false,
17
+ "special": false
18
+ },
19
+ "56": {
20
+ "content": "<s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "57": {
28
+ "content": "</s>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ }
35
+ },
36
+ "bos_token": "<s>",
37
+ "clean_up_tokenization_spaces": false,
38
+ "do_lower_case": false,
39
+ "eos_token": "</s>",
40
+ "model_max_length": 1000000000000000019884624838656,
41
+ "pad_token": "[PAD]",
42
+ "processor_class": "Wav2Vec2BertProcessor",
43
+ "replace_word_delimiter_char": " ",
44
+ "target_lang": null,
45
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
46
+ "unk_token": "[UNK]",
47
+ "word_delimiter_token": "|"
48
+ }
transcription/w2v-bert-darija-finetuned-clean/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2793aaf5af1a42cb015bfef656c097f2acac87a48fffa9d3175eaeceb802e54
3
+ size 5304
transcription/w2v-bert-darija-finetuned-clean/vocab.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "\t": 0,
3
+ "\n": 1,
4
+ "!": 3,
5
+ "(": 4,
6
+ ")": 5,
7
+ ",": 6,
8
+ ".": 7,
9
+ "?": 8,
10
+ "[PAD]": 55,
11
+ "[UNK]": 54,
12
+ "|": 2,
13
+ "،": 9,
14
+ "؟": 10,
15
+ "ء": 11,
16
+ "آ": 12,
17
+ "أ": 13,
18
+ "ؤ": 14,
19
+ "إ": 15,
20
+ "ئ": 16,
21
+ "ا": 17,
22
+ "ب": 18,
23
+ "ة": 19,
24
+ "ت": 20,
25
+ "ث": 21,
26
+ "ج": 22,
27
+ "ح": 23,
28
+ "خ": 24,
29
+ "د": 25,
30
+ "ذ": 26,
31
+ "ر": 27,
32
+ "ز": 28,
33
+ "س": 29,
34
+ "ش": 30,
35
+ "ص": 31,
36
+ "ض": 32,
37
+ "ط": 33,
38
+ "ظ": 34,
39
+ "ع": 35,
40
+ "غ": 36,
41
+ "ف": 37,
42
+ "ق": 38,
43
+ "ك": 39,
44
+ "ل": 40,
45
+ "م": 41,
46
+ "ن": 42,
47
+ "ه": 43,
48
+ "و": 44,
49
+ "ى": 45,
50
+ "ي": 46,
51
+ "ً": 47,
52
+ "ٍ": 48,
53
+ "ّ": 49,
54
+ "ٱ": 50,
55
+ "پ": 51,
56
+ "ڤ": 52,
57
+ "ڭ": 53
58
+ }