besimray commited on
Commit
88bc464
·
verified ·
1 Parent(s): 082a36c

Training in progress, step 40, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c262af3316d92f1e6a59b3f3fce338e187fac9175949a02bc80bdc70674949db
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ff347f2a4ff28924131388644033ac74b2be4fef6beba204132fa14740d367c
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d7abfd469db57334042c7cf1fda0a239d09983f7d72fbc0bfe26e16575bc765
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3640f11e9bfa375e1b65f931e090ec5e24dfd0440bca1d38b3176296ce111f2e
3
  size 23159290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eaefb486716e7cdd53aa7207d6e8d9df693e032d4b9b882b4154f9c46bb10d61
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:161d6a87de01bfcacb8d5d93875329a9cc29a12cdde97e0bfe85d9cdc27baa92
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c297c5cf11a27c75d9f99f1df69752f78c3ad41b0275adf50cdd1b67f9d0bb3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:062016b917f4cd81cef6ab15bfe81df4e94586c0afba8905b655deaa6fff468f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 1.1702154874801636,
3
- "best_model_checkpoint": "miner_id_besimray/checkpoint-20",
4
- "epoch": 0.42105263157894735,
5
  "eval_steps": 20,
6
- "global_step": 20,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -163,6 +163,154 @@
163
  "eval_samples_per_second": 47.875,
164
  "eval_steps_per_second": 4.788,
165
  "step": 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
166
  }
167
  ],
168
  "logging_steps": 1,
@@ -191,7 +339,7 @@
191
  "attributes": {}
192
  }
193
  },
194
- "total_flos": 2033729462599680.0,
195
  "train_batch_size": 10,
196
  "trial_name": null,
197
  "trial_params": null
 
1
  {
2
+ "best_metric": 1.153311014175415,
3
+ "best_model_checkpoint": "miner_id_besimray/checkpoint-40",
4
+ "epoch": 0.8421052631578947,
5
  "eval_steps": 20,
6
+ "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
163
  "eval_samples_per_second": 47.875,
164
  "eval_steps_per_second": 4.788,
165
  "step": 20
166
+ },
167
+ {
168
+ "epoch": 0.4421052631578947,
169
+ "grad_norm": 0.4274454414844513,
170
+ "learning_rate": 0.0001969689520376687,
171
+ "loss": 1.0972,
172
+ "step": 21
173
+ },
174
+ {
175
+ "epoch": 0.4631578947368421,
176
+ "grad_norm": 0.3145941197872162,
177
+ "learning_rate": 0.00019639628606958533,
178
+ "loss": 1.1746,
179
+ "step": 22
180
+ },
181
+ {
182
+ "epoch": 0.4842105263157895,
183
+ "grad_norm": 0.34570032358169556,
184
+ "learning_rate": 0.00019577508166849304,
185
+ "loss": 1.2273,
186
+ "step": 23
187
+ },
188
+ {
189
+ "epoch": 0.5052631578947369,
190
+ "grad_norm": 0.4847642481327057,
191
+ "learning_rate": 0.00019510565162951537,
192
+ "loss": 1.2152,
193
+ "step": 24
194
+ },
195
+ {
196
+ "epoch": 0.5263157894736842,
197
+ "grad_norm": 0.3375917375087738,
198
+ "learning_rate": 0.00019438833303083678,
199
+ "loss": 1.1977,
200
+ "step": 25
201
+ },
202
+ {
203
+ "epoch": 0.5473684210526316,
204
+ "grad_norm": 0.42273998260498047,
205
+ "learning_rate": 0.00019362348706397373,
206
+ "loss": 1.24,
207
+ "step": 26
208
+ },
209
+ {
210
+ "epoch": 0.5684210526315789,
211
+ "grad_norm": 0.4141988456249237,
212
+ "learning_rate": 0.0001928114988519039,
213
+ "loss": 1.2562,
214
+ "step": 27
215
+ },
216
+ {
217
+ "epoch": 0.5894736842105263,
218
+ "grad_norm": 0.382915198802948,
219
+ "learning_rate": 0.0001919527772551451,
220
+ "loss": 1.226,
221
+ "step": 28
222
+ },
223
+ {
224
+ "epoch": 0.6105263157894737,
225
+ "grad_norm": 0.37382128834724426,
226
+ "learning_rate": 0.00019104775466588161,
227
+ "loss": 1.3248,
228
+ "step": 29
229
+ },
230
+ {
231
+ "epoch": 0.631578947368421,
232
+ "grad_norm": 0.3107808232307434,
233
+ "learning_rate": 0.0001900968867902419,
234
+ "loss": 1.1354,
235
+ "step": 30
236
+ },
237
+ {
238
+ "epoch": 0.6526315789473685,
239
+ "grad_norm": 0.34155750274658203,
240
+ "learning_rate": 0.0001891006524188368,
241
+ "loss": 1.0803,
242
+ "step": 31
243
+ },
244
+ {
245
+ "epoch": 0.6736842105263158,
246
+ "grad_norm": 0.3141622543334961,
247
+ "learning_rate": 0.0001880595531856738,
248
+ "loss": 1.1457,
249
+ "step": 32
250
+ },
251
+ {
252
+ "epoch": 0.6947368421052632,
253
+ "grad_norm": 0.35257869958877563,
254
+ "learning_rate": 0.00018697411331556956,
255
+ "loss": 1.2322,
256
+ "step": 33
257
+ },
258
+ {
259
+ "epoch": 0.7157894736842105,
260
+ "grad_norm": 0.42415115237236023,
261
+ "learning_rate": 0.00018584487936018661,
262
+ "loss": 1.179,
263
+ "step": 34
264
+ },
265
+ {
266
+ "epoch": 0.7368421052631579,
267
+ "grad_norm": 0.3805026710033417,
268
+ "learning_rate": 0.00018467241992282843,
269
+ "loss": 1.126,
270
+ "step": 35
271
+ },
272
+ {
273
+ "epoch": 0.7578947368421053,
274
+ "grad_norm": 0.3650873601436615,
275
+ "learning_rate": 0.00018345732537213027,
276
+ "loss": 1.2333,
277
+ "step": 36
278
+ },
279
+ {
280
+ "epoch": 0.7789473684210526,
281
+ "grad_norm": 0.3254134953022003,
282
+ "learning_rate": 0.00018220020754479102,
283
+ "loss": 1.2721,
284
+ "step": 37
285
+ },
286
+ {
287
+ "epoch": 0.8,
288
+ "grad_norm": 0.36688175797462463,
289
+ "learning_rate": 0.00018090169943749476,
290
+ "loss": 1.2615,
291
+ "step": 38
292
+ },
293
+ {
294
+ "epoch": 0.8210526315789474,
295
+ "grad_norm": 0.3320186734199524,
296
+ "learning_rate": 0.00017956245488817812,
297
+ "loss": 1.1474,
298
+ "step": 39
299
+ },
300
+ {
301
+ "epoch": 0.8421052631578947,
302
+ "grad_norm": 0.37685626745224,
303
+ "learning_rate": 0.000178183148246803,
304
+ "loss": 1.1977,
305
+ "step": 40
306
+ },
307
+ {
308
+ "epoch": 0.8421052631578947,
309
+ "eval_loss": 1.153311014175415,
310
+ "eval_runtime": 2.0851,
311
+ "eval_samples_per_second": 47.96,
312
+ "eval_steps_per_second": 4.796,
313
+ "step": 40
314
  }
315
  ],
316
  "logging_steps": 1,
 
339
  "attributes": {}
340
  }
341
  },
342
+ "total_flos": 4078799424061440.0,
343
  "train_batch_size": 10,
344
  "trial_name": null,
345
  "trial_params": null