ljcamargo commited on
Commit
819db7a
·
verified ·
1 Parent(s): dbd7e44

Training in progress, step 300, checkpoint

Browse files
last-checkpoint/config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.0,
3
+ "activation_function": "gelu",
4
+ "apply_spec_augment": false,
5
+ "architectures": [
6
+ "WhisperEncoderOnlyForClassification"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "begin_suppress_tokens": [
10
+ 220,
11
+ 50256
12
+ ],
13
+ "bos_token_id": 50257,
14
+ "classifier_proj_size": 256,
15
+ "d_model": 1280,
16
+ "decoder_attention_heads": 20,
17
+ "decoder_ffn_dim": 5120,
18
+ "decoder_layerdrop": 0.0,
19
+ "decoder_layers": 4,
20
+ "decoder_start_token_id": 50258,
21
+ "dropout": 0.0,
22
+ "dtype": "float32",
23
+ "encoder_attention_heads": 20,
24
+ "encoder_ffn_dim": 5120,
25
+ "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 32,
27
+ "eos_token_id": 50257,
28
+ "init_std": 0.02,
29
+ "is_encoder_decoder": true,
30
+ "mask_feature_length": 10,
31
+ "mask_feature_min_masks": 0,
32
+ "mask_feature_prob": 0.0,
33
+ "mask_time_length": 10,
34
+ "mask_time_min_masks": 2,
35
+ "mask_time_prob": 0.05,
36
+ "max_source_positions": 1500,
37
+ "max_target_positions": 448,
38
+ "median_filter_width": 7,
39
+ "model_type": "whisper_encoder_classifier",
40
+ "n_code": 1965,
41
+ "n_fam": 11,
42
+ "n_super": 68,
43
+ "num_hidden_layers": 32,
44
+ "num_mel_bins": 128,
45
+ "pad_token_id": 50257,
46
+ "scale_embedding": false,
47
+ "transformers_version": "4.56.2",
48
+ "unsloth_version": "2025.12.8",
49
+ "use_cache": true,
50
+ "use_weighted_layer_sum": false,
51
+ "vocab_size": 51866
52
+ }
last-checkpoint/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:389139e8293cf0265bd0dd8a65a2482d21cc51a5fa821642143b7daf82f108b0
3
+ size 2558403928
last-checkpoint/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7507fef582f26fa6dd26593efcff1fe00c92de5de40a9e14d4bcfae3d4660cf5
3
+ size 1313044361
last-checkpoint/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6bfeac7f2161ffd59c7dec4e192c42f83cdf28d8494d04d02013314243750b1
3
+ size 14645
last-checkpoint/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30af866df24edce708e1eb20700878b402fa05707fa9bc5f332496baf440dbbb
3
+ size 1383
last-checkpoint/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e7f847589ff5ebc49461e971738b280435583eb06a67d9e67099912165ff9ed
3
+ size 1465
last-checkpoint/trainer_state.json ADDED
@@ -0,0 +1,251 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.12,
6
+ "eval_steps": 500,
7
+ "global_step": 300,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.0004,
14
+ "grad_norm": 8.119571685791016,
15
+ "learning_rate": 0.0,
16
+ "loss": 3.4677,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.004,
21
+ "grad_norm": 8.098031997680664,
22
+ "learning_rate": 9e-06,
23
+ "loss": 4.4682,
24
+ "step": 10
25
+ },
26
+ {
27
+ "epoch": 0.008,
28
+ "grad_norm": 8.463135719299316,
29
+ "learning_rate": 1.9e-05,
30
+ "loss": 4.6132,
31
+ "step": 20
32
+ },
33
+ {
34
+ "epoch": 0.012,
35
+ "grad_norm": 7.452148914337158,
36
+ "learning_rate": 2.9e-05,
37
+ "loss": 4.223,
38
+ "step": 30
39
+ },
40
+ {
41
+ "epoch": 0.016,
42
+ "grad_norm": 5.572236061096191,
43
+ "learning_rate": 3.9000000000000006e-05,
44
+ "loss": 4.4111,
45
+ "step": 40
46
+ },
47
+ {
48
+ "epoch": 0.02,
49
+ "grad_norm": 7.39447021484375,
50
+ "learning_rate": 4.9e-05,
51
+ "loss": 4.0543,
52
+ "step": 50
53
+ },
54
+ {
55
+ "epoch": 0.024,
56
+ "grad_norm": 9.954078674316406,
57
+ "learning_rate": 4.999833521640187e-05,
58
+ "loss": 4.5007,
59
+ "step": 60
60
+ },
61
+ {
62
+ "epoch": 0.028,
63
+ "grad_norm": 5.994736194610596,
64
+ "learning_rate": 4.9992580693557054e-05,
65
+ "loss": 4.6204,
66
+ "step": 70
67
+ },
68
+ {
69
+ "epoch": 0.032,
70
+ "grad_norm": 7.18951940536499,
71
+ "learning_rate": 4.998271682453017e-05,
72
+ "loss": 4.2467,
73
+ "step": 80
74
+ },
75
+ {
76
+ "epoch": 0.036,
77
+ "grad_norm": 9.591761589050293,
78
+ "learning_rate": 4.996874523116464e-05,
79
+ "loss": 4.4063,
80
+ "step": 90
81
+ },
82
+ {
83
+ "epoch": 0.04,
84
+ "grad_norm": 7.566534519195557,
85
+ "learning_rate": 4.995066821070679e-05,
86
+ "loss": 4.0773,
87
+ "step": 100
88
+ },
89
+ {
90
+ "epoch": 0.044,
91
+ "grad_norm": 5.793056488037109,
92
+ "learning_rate": 4.9928488735428105e-05,
93
+ "loss": 4.063,
94
+ "step": 110
95
+ },
96
+ {
97
+ "epoch": 0.048,
98
+ "grad_norm": 7.917051792144775,
99
+ "learning_rate": 4.990221045213652e-05,
100
+ "loss": 4.2533,
101
+ "step": 120
102
+ },
103
+ {
104
+ "epoch": 0.052,
105
+ "grad_norm": 7.804363250732422,
106
+ "learning_rate": 4.987183768157686e-05,
107
+ "loss": 4.0497,
108
+ "step": 130
109
+ },
110
+ {
111
+ "epoch": 0.056,
112
+ "grad_norm": 7.605108261108398,
113
+ "learning_rate": 4.983737541772033e-05,
114
+ "loss": 4.4334,
115
+ "step": 140
116
+ },
117
+ {
118
+ "epoch": 0.06,
119
+ "grad_norm": 9.323838233947754,
120
+ "learning_rate": 4.979882932694346e-05,
121
+ "loss": 4.0412,
122
+ "step": 150
123
+ },
124
+ {
125
+ "epoch": 0.064,
126
+ "grad_norm": 9.219818115234375,
127
+ "learning_rate": 4.9756205747096385e-05,
128
+ "loss": 3.9774,
129
+ "step": 160
130
+ },
131
+ {
132
+ "epoch": 0.068,
133
+ "grad_norm": 8.375937461853027,
134
+ "learning_rate": 4.9709511686460775e-05,
135
+ "loss": 4.0021,
136
+ "step": 170
137
+ },
138
+ {
139
+ "epoch": 0.072,
140
+ "grad_norm": 8.215436935424805,
141
+ "learning_rate": 4.96587548225975e-05,
142
+ "loss": 4.4227,
143
+ "step": 180
144
+ },
145
+ {
146
+ "epoch": 0.076,
147
+ "grad_norm": 9.561295509338379,
148
+ "learning_rate": 4.960394350108429e-05,
149
+ "loss": 4.1091,
150
+ "step": 190
151
+ },
152
+ {
153
+ "epoch": 0.08,
154
+ "grad_norm": 9.014528274536133,
155
+ "learning_rate": 4.954508673414351e-05,
156
+ "loss": 3.9428,
157
+ "step": 200
158
+ },
159
+ {
160
+ "epoch": 0.084,
161
+ "grad_norm": 7.69431209564209,
162
+ "learning_rate": 4.948219419916037e-05,
163
+ "loss": 4.368,
164
+ "step": 210
165
+ },
166
+ {
167
+ "epoch": 0.088,
168
+ "grad_norm": 11.925583839416504,
169
+ "learning_rate": 4.941527623709172e-05,
170
+ "loss": 3.6757,
171
+ "step": 220
172
+ },
173
+ {
174
+ "epoch": 0.092,
175
+ "grad_norm": 8.191117286682129,
176
+ "learning_rate": 4.934434385076576e-05,
177
+ "loss": 4.1905,
178
+ "step": 230
179
+ },
180
+ {
181
+ "epoch": 0.096,
182
+ "grad_norm": 7.863613128662109,
183
+ "learning_rate": 4.926940870307296e-05,
184
+ "loss": 4.0099,
185
+ "step": 240
186
+ },
187
+ {
188
+ "epoch": 0.1,
189
+ "grad_norm": 10.665002822875977,
190
+ "learning_rate": 4.9190483115048375e-05,
191
+ "loss": 3.9059,
192
+ "step": 250
193
+ },
194
+ {
195
+ "epoch": 0.104,
196
+ "grad_norm": 10.385906219482422,
197
+ "learning_rate": 4.910758006384583e-05,
198
+ "loss": 3.9221,
199
+ "step": 260
200
+ },
201
+ {
202
+ "epoch": 0.108,
203
+ "grad_norm": 8.544922828674316,
204
+ "learning_rate": 4.9020713180604126e-05,
205
+ "loss": 3.9398,
206
+ "step": 270
207
+ },
208
+ {
209
+ "epoch": 0.112,
210
+ "grad_norm": 7.989080429077148,
211
+ "learning_rate": 4.892989674820585e-05,
212
+ "loss": 3.7757,
213
+ "step": 280
214
+ },
215
+ {
216
+ "epoch": 0.116,
217
+ "grad_norm": 6.576107025146484,
218
+ "learning_rate": 4.8835145698928856e-05,
219
+ "loss": 3.5309,
220
+ "step": 290
221
+ },
222
+ {
223
+ "epoch": 0.12,
224
+ "grad_norm": 9.80089282989502,
225
+ "learning_rate": 4.873647561199115e-05,
226
+ "loss": 4.1776,
227
+ "step": 300
228
+ }
229
+ ],
230
+ "logging_steps": 10,
231
+ "max_steps": 2500,
232
+ "num_input_tokens_seen": 0,
233
+ "num_train_epochs": 1,
234
+ "save_steps": 300,
235
+ "stateful_callbacks": {
236
+ "TrainerControl": {
237
+ "args": {
238
+ "should_epoch_stop": false,
239
+ "should_evaluate": false,
240
+ "should_log": false,
241
+ "should_save": true,
242
+ "should_training_stop": false
243
+ },
244
+ "attributes": {}
245
+ }
246
+ },
247
+ "total_flos": 1.41041809391616e+19,
248
+ "train_batch_size": 8,
249
+ "trial_name": null,
250
+ "trial_params": null
251
+ }
last-checkpoint/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96555ac3c0979f5dce122c8f700ea90690bfa57f9771c6c52bfcefa74d7b8abe
3
+ size 5841