besimray commited on
Commit
d1fba60
·
verified ·
1 Parent(s): 83fb4a8

Training in progress, step 60, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ff347f2a4ff28924131388644033ac74b2be4fef6beba204132fa14740d367c
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87f7d1cab1f1c1f94445cd22369257aaa21529875106c6c56c58b5307c1bc477
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3640f11e9bfa375e1b65f931e090ec5e24dfd0440bca1d38b3176296ce111f2e
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa7b7ae7c5bb58cdf42f58142b085e62f4e7a4f192e231d24d38bd4992f8be42
3
  size 23159290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:161d6a87de01bfcacb8d5d93875329a9cc29a12cdde97e0bfe85d9cdc27baa92
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cf47f03b057010c0300b7a3c8bdc5b1074d0d042fbd2a7295866ded90ee1395
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:062016b917f4cd81cef6ab15bfe81df4e94586c0afba8905b655deaa6fff468f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50b80d146b8b37f82a1962ba385b8329fd6c2c35ba5116c53e131bb661ab681b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 1.153311014175415,
3
- "best_model_checkpoint": "miner_id_besimray/checkpoint-40",
4
- "epoch": 0.8421052631578947,
5
  "eval_steps": 20,
6
- "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -311,6 +311,154 @@
311
  "eval_samples_per_second": 47.96,
312
  "eval_steps_per_second": 4.796,
313
  "step": 40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
314
  }
315
  ],
316
  "logging_steps": 1,
@@ -339,7 +487,7 @@
339
  "attributes": {}
340
  }
341
  },
342
- "total_flos": 4078799424061440.0,
343
  "train_batch_size": 10,
344
  "trial_name": null,
345
  "trial_params": null
 
1
  {
2
+ "best_metric": 1.1519354581832886,
3
+ "best_model_checkpoint": "miner_id_besimray/checkpoint-60",
4
+ "epoch": 1.263157894736842,
5
  "eval_steps": 20,
6
+ "global_step": 60,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
311
  "eval_samples_per_second": 47.96,
312
  "eval_steps_per_second": 4.796,
313
  "step": 40
314
+ },
315
+ {
316
+ "epoch": 0.8631578947368421,
317
+ "grad_norm": 0.43908432126045227,
318
+ "learning_rate": 0.0001767644740358011,
319
+ "loss": 1.2448,
320
+ "step": 41
321
+ },
322
+ {
323
+ "epoch": 0.8842105263157894,
324
+ "grad_norm": 0.3212919235229492,
325
+ "learning_rate": 0.00017530714660036112,
326
+ "loss": 1.1699,
327
+ "step": 42
328
+ },
329
+ {
330
+ "epoch": 0.9052631578947369,
331
+ "grad_norm": 0.339679479598999,
332
+ "learning_rate": 0.00017381189974873407,
333
+ "loss": 1.1572,
334
+ "step": 43
335
+ },
336
+ {
337
+ "epoch": 0.9263157894736842,
338
+ "grad_norm": 0.3269651234149933,
339
+ "learning_rate": 0.00017227948638273916,
340
+ "loss": 1.1494,
341
+ "step": 44
342
+ },
343
+ {
344
+ "epoch": 0.9473684210526315,
345
+ "grad_norm": 0.34232962131500244,
346
+ "learning_rate": 0.00017071067811865476,
347
+ "loss": 1.2262,
348
+ "step": 45
349
+ },
350
+ {
351
+ "epoch": 0.968421052631579,
352
+ "grad_norm": 0.34240803122520447,
353
+ "learning_rate": 0.00016910626489868649,
354
+ "loss": 1.1834,
355
+ "step": 46
356
+ },
357
+ {
358
+ "epoch": 0.9894736842105263,
359
+ "grad_norm": 0.35933125019073486,
360
+ "learning_rate": 0.00016746705459320745,
361
+ "loss": 1.0391,
362
+ "step": 47
363
+ },
364
+ {
365
+ "epoch": 1.0105263157894737,
366
+ "grad_norm": 0.3355937898159027,
367
+ "learning_rate": 0.00016579387259397127,
368
+ "loss": 1.2872,
369
+ "step": 48
370
+ },
371
+ {
372
+ "epoch": 1.0315789473684212,
373
+ "grad_norm": 0.3706349730491638,
374
+ "learning_rate": 0.0001640875613985024,
375
+ "loss": 1.0775,
376
+ "step": 49
377
+ },
378
+ {
379
+ "epoch": 1.0526315789473684,
380
+ "grad_norm": 0.427852988243103,
381
+ "learning_rate": 0.00016234898018587337,
382
+ "loss": 1.1524,
383
+ "step": 50
384
+ },
385
+ {
386
+ "epoch": 1.0736842105263158,
387
+ "grad_norm": 0.3533117473125458,
388
+ "learning_rate": 0.000160579004384082,
389
+ "loss": 1.1181,
390
+ "step": 51
391
+ },
392
+ {
393
+ "epoch": 1.0947368421052632,
394
+ "grad_norm": 0.3712696135044098,
395
+ "learning_rate": 0.00015877852522924732,
396
+ "loss": 1.0986,
397
+ "step": 52
398
+ },
399
+ {
400
+ "epoch": 1.1157894736842104,
401
+ "grad_norm": 0.3790956437587738,
402
+ "learning_rate": 0.0001569484493168452,
403
+ "loss": 1.1749,
404
+ "step": 53
405
+ },
406
+ {
407
+ "epoch": 1.1368421052631579,
408
+ "grad_norm": 0.3779037296772003,
409
+ "learning_rate": 0.00015508969814521025,
410
+ "loss": 1.1089,
411
+ "step": 54
412
+ },
413
+ {
414
+ "epoch": 1.1578947368421053,
415
+ "grad_norm": 0.36196696758270264,
416
+ "learning_rate": 0.00015320320765153367,
417
+ "loss": 1.0186,
418
+ "step": 55
419
+ },
420
+ {
421
+ "epoch": 1.1789473684210527,
422
+ "grad_norm": 0.3449699282646179,
423
+ "learning_rate": 0.00015128992774059063,
424
+ "loss": 1.064,
425
+ "step": 56
426
+ },
427
+ {
428
+ "epoch": 1.2,
429
+ "grad_norm": 0.43372786045074463,
430
+ "learning_rate": 0.0001493508218064347,
431
+ "loss": 1.007,
432
+ "step": 57
433
+ },
434
+ {
435
+ "epoch": 1.2210526315789474,
436
+ "grad_norm": 0.35580453276634216,
437
+ "learning_rate": 0.00014738686624729986,
438
+ "loss": 1.0295,
439
+ "step": 58
440
+ },
441
+ {
442
+ "epoch": 1.2421052631578948,
443
+ "grad_norm": 0.47308239340782166,
444
+ "learning_rate": 0.00014539904997395468,
445
+ "loss": 1.1361,
446
+ "step": 59
447
+ },
448
+ {
449
+ "epoch": 1.263157894736842,
450
+ "grad_norm": 0.3692001402378082,
451
+ "learning_rate": 0.00014338837391175582,
452
+ "loss": 1.099,
453
+ "step": 60
454
+ },
455
+ {
456
+ "epoch": 1.263157894736842,
457
+ "eval_loss": 1.1519354581832886,
458
+ "eval_runtime": 2.053,
459
+ "eval_samples_per_second": 48.71,
460
+ "eval_steps_per_second": 4.871,
461
+ "step": 60
462
  }
463
  ],
464
  "logging_steps": 1,
 
487
  "attributes": {}
488
  }
489
  },
490
+ "total_flos": 6036925560913920.0,
491
  "train_batch_size": 10,
492
  "trial_name": null,
493
  "trial_params": null