binhtx commited on
Commit
3f043c1
·
1 Parent(s): 4661415

train success v4 2000 step

Browse files
config.json CHANGED
@@ -11,7 +11,10 @@
11
  "MBartModel"
12
  ],
13
  "attention_dropout": 0.0,
 
 
14
  "classifier_dropout": 0.0,
 
15
  "d_model": 768,
16
  "decoder_attention_heads": 12,
17
  "decoder_ffn_dim": 3072,
@@ -24,15 +27,39 @@
24
  "encoder_ffn_dim": 3072,
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 6,
 
 
27
  "gradient_checkpointing": false,
 
 
 
 
28
  "init_std": 0.02,
29
  "is_decoder": true,
30
  "is_encoder_decoder": false,
 
 
 
 
31
  "max_position_embeddings": 1024,
32
  "model_type": "mbart",
33
  "num_hidden_layers": 6,
 
 
 
 
 
 
 
34
  "scale_embedding": false,
 
 
 
 
 
35
  "tokenizer_class": "BartphoTokenizer",
 
 
36
  "use_cache": true,
37
  "vocab_size": 41031
38
  },
@@ -45,11 +72,14 @@
45
  "adapter_kernel_size": 3,
46
  "adapter_stride": 2,
47
  "add_adapter": true,
 
48
  "apply_spec_augment": true,
49
  "architectures": [
50
  "Wav2Vec2ForPreTraining"
51
  ],
52
  "attention_dropout": 0.1,
 
 
53
  "classifier_proj_size": 256,
54
  "codevector_dim": 768,
55
  "contrastive_logits_temperature": 0.1,
@@ -81,23 +111,37 @@
81
  2,
82
  2
83
  ],
 
84
  "ctc_loss_reduction": "sum",
85
  "ctc_zero_infinity": false,
 
86
  "diversity_loss_weight": 0.1,
87
  "do_stable_layer_norm": true,
88
  "dtype": "float32",
 
89
  "feat_extract_activation": "gelu",
90
  "feat_extract_dropout": 0.0,
91
  "feat_extract_norm": "layer",
92
  "feat_proj_dropout": 0.1,
93
  "feat_quantizer_dropout": 0.0,
94
  "final_dropout": 0.0,
 
95
  "gradient_checkpointing": false,
96
  "hidden_act": "gelu",
97
  "hidden_dropout": 0.1,
98
  "hidden_size": 1024,
 
 
 
 
99
  "initializer_range": 0.02,
100
  "intermediate_size": 4096,
 
 
 
 
 
 
101
  "layer_norm_eps": 1e-05,
102
  "layerdrop": 0.1,
103
  "mask_channel_length": 10,
@@ -124,8 +168,17 @@
124
  "num_feat_extract_layers": 7,
125
  "num_hidden_layers": 24,
126
  "num_negatives": 100,
 
127
  "output_hidden_size": 1024,
 
 
 
 
128
  "proj_codevector_dim": 768,
 
 
 
 
129
  "tdnn_dilation": [
130
  1,
131
  2,
@@ -147,6 +200,12 @@
147
  1,
148
  1
149
  ],
 
 
 
 
 
 
150
  "use_weighted_layer_sum": false,
151
  "vocab_size": 96,
152
  "xvector_output_dim": 512
@@ -156,5 +215,5 @@
156
  "model_type": "speech-encoder-decoder",
157
  "pad_token_id": 1,
158
  "tie_word_embeddings": false,
159
- "transformers_version": "4.57.5"
160
  }
 
11
  "MBartModel"
12
  ],
13
  "attention_dropout": 0.0,
14
+ "bos_token_id": 0,
15
+ "chunk_size_feed_forward": 0,
16
  "classifier_dropout": 0.0,
17
+ "cross_attention_hidden_size": null,
18
  "d_model": 768,
19
  "decoder_attention_heads": 12,
20
  "decoder_ffn_dim": 3072,
 
27
  "encoder_ffn_dim": 3072,
28
  "encoder_layerdrop": 0.0,
29
  "encoder_layers": 6,
30
+ "eos_token_id": 2,
31
+ "finetuning_task": null,
32
  "gradient_checkpointing": false,
33
+ "id2label": {
34
+ "0": "LABEL_0",
35
+ "1": "LABEL_1"
36
+ },
37
  "init_std": 0.02,
38
  "is_decoder": true,
39
  "is_encoder_decoder": false,
40
+ "label2id": {
41
+ "LABEL_0": 0,
42
+ "LABEL_1": 1
43
+ },
44
  "max_position_embeddings": 1024,
45
  "model_type": "mbart",
46
  "num_hidden_layers": 6,
47
+ "output_attentions": false,
48
+ "output_hidden_states": false,
49
+ "pad_token_id": 1,
50
+ "prefix": null,
51
+ "problem_type": null,
52
+ "pruned_heads": {},
53
+ "return_dict": true,
54
  "scale_embedding": false,
55
+ "sep_token_id": null,
56
+ "task_specific_params": null,
57
+ "tf_legacy_loss": false,
58
+ "tie_encoder_decoder": false,
59
+ "tie_word_embeddings": true,
60
  "tokenizer_class": "BartphoTokenizer",
61
+ "torchscript": false,
62
+ "use_bfloat16": false,
63
  "use_cache": true,
64
  "vocab_size": 41031
65
  },
 
72
  "adapter_kernel_size": 3,
73
  "adapter_stride": 2,
74
  "add_adapter": true,
75
+ "add_cross_attention": false,
76
  "apply_spec_augment": true,
77
  "architectures": [
78
  "Wav2Vec2ForPreTraining"
79
  ],
80
  "attention_dropout": 0.1,
81
+ "bos_token_id": 1,
82
+ "chunk_size_feed_forward": 0,
83
  "classifier_proj_size": 256,
84
  "codevector_dim": 768,
85
  "contrastive_logits_temperature": 0.1,
 
111
  2,
112
  2
113
  ],
114
+ "cross_attention_hidden_size": null,
115
  "ctc_loss_reduction": "sum",
116
  "ctc_zero_infinity": false,
117
+ "decoder_start_token_id": null,
118
  "diversity_loss_weight": 0.1,
119
  "do_stable_layer_norm": true,
120
  "dtype": "float32",
121
+ "eos_token_id": 2,
122
  "feat_extract_activation": "gelu",
123
  "feat_extract_dropout": 0.0,
124
  "feat_extract_norm": "layer",
125
  "feat_proj_dropout": 0.1,
126
  "feat_quantizer_dropout": 0.0,
127
  "final_dropout": 0.0,
128
+ "finetuning_task": null,
129
  "gradient_checkpointing": false,
130
  "hidden_act": "gelu",
131
  "hidden_dropout": 0.1,
132
  "hidden_size": 1024,
133
+ "id2label": {
134
+ "0": "LABEL_0",
135
+ "1": "LABEL_1"
136
+ },
137
  "initializer_range": 0.02,
138
  "intermediate_size": 4096,
139
+ "is_decoder": false,
140
+ "is_encoder_decoder": false,
141
+ "label2id": {
142
+ "LABEL_0": 0,
143
+ "LABEL_1": 1
144
+ },
145
  "layer_norm_eps": 1e-05,
146
  "layerdrop": 0.1,
147
  "mask_channel_length": 10,
 
168
  "num_feat_extract_layers": 7,
169
  "num_hidden_layers": 24,
170
  "num_negatives": 100,
171
+ "output_attentions": false,
172
  "output_hidden_size": 1024,
173
+ "output_hidden_states": false,
174
+ "pad_token_id": 0,
175
+ "prefix": null,
176
+ "problem_type": null,
177
  "proj_codevector_dim": 768,
178
+ "pruned_heads": {},
179
+ "return_dict": true,
180
+ "sep_token_id": null,
181
+ "task_specific_params": null,
182
  "tdnn_dilation": [
183
  1,
184
  2,
 
200
  1,
201
  1
202
  ],
203
+ "tf_legacy_loss": false,
204
+ "tie_encoder_decoder": false,
205
+ "tie_word_embeddings": true,
206
+ "tokenizer_class": null,
207
+ "torchscript": false,
208
+ "use_bfloat16": false,
209
  "use_weighted_layer_sum": false,
210
  "vocab_size": 96,
211
  "xvector_output_dim": 512
 
215
  "model_type": "speech-encoder-decoder",
216
  "pad_token_id": 1,
217
  "tie_word_embeddings": false,
218
+ "transformers_version": "5.1.0"
219
  }
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "eos_token_id": 2,
6
  "forced_eos_token_id": 2,
7
  "pad_token_id": 1,
8
- "transformers_version": "4.57.5"
9
  }
 
5
  "eos_token_id": 2,
6
  "forced_eos_token_id": 2,
7
  "pad_token_id": 1,
8
+ "transformers_version": "5.1.0"
9
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1234c9bde0f2d753397db6c0a4a5da72eff1ba124a615a1b309507bb2ae31ab2
3
- size 1696553104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e132ef90cdcfdbed22c366cf4701b775777a5ecfcf5a1f3ca27afdda9f64fa
3
+ size 1822600472
preprocessor_config.json CHANGED
@@ -4,7 +4,6 @@
4
  "feature_size": 1,
5
  "padding_side": "right",
6
  "padding_value": 0.0,
7
- "processor_class": "Wav2Vec2Processor",
8
  "return_attention_mask": true,
9
  "sampling_rate": 16000
10
  }
 
4
  "feature_size": 1,
5
  "padding_side": "right",
6
  "padding_value": 0.0,
 
7
  "return_attention_mask": true,
8
  "sampling_rate": 16000
9
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a174c57ede6c576b3f0d90f19a76ba11856d3c885364b7b57c559e477407203e
3
- size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90e578251d8e4b5e298fb1d9fcad08843a7bc3f1ab47d8570c8be8e5b8e29d70
3
+ size 5329