Wilsonwin commited on
Commit
f9ed321
·
verified ·
1 Parent(s): b1fbf4c

Training in progress, step 1000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87ff9a10f70d200b2c856dee103d85c0b4769807377164ae058876c9b2f8cdf4
3
  size 328277848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6e3a61755ce1ee753f0e8d104d120e50755452bd8cbf6e05ff0a62e9faa8d42
3
  size 328277848
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3e3d47782f2acc46ebaf4d2a3bfa3f0989afdba6a7eed3d7d950f2b764c9119
3
  size 318646859
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91744d22093227f73ff956561d21826bda5cf09b77b7d63e333247291b5a2ff3
3
  size 318646859
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a9464afe392ee166598218bbc5440e93bb5f6227028ba0c26fc8e4fdcf73ae8
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:577f0b7cee8114520712167e366c7fd1da45e6cc96d21bb243dccf2193c0d60f
3
  size 14645
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d6f69db244377919464a655cd8dc38a3ecd36738e8ab7fd789261995d39f57a
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:641712abd35039f810da46b5ecace55e8c31f5b5a7d2cfa0aaa8182597f8aad6
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 0.08447372867038351,
6
  "eval_steps": 500,
7
- "global_step": 500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -373,6 +373,364 @@
373
  "eval_samples_per_second": 116.708,
374
  "eval_steps_per_second": 2.451,
375
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
376
  }
377
  ],
378
  "logging_steps": 10,
@@ -392,7 +750,7 @@
392
  "attributes": {}
393
  }
394
  },
395
- "total_flos": 1.6722841042944e+16,
396
  "train_batch_size": 48,
397
  "trial_name": null,
398
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 0.16894745734076702,
6
  "eval_steps": 500,
7
+ "global_step": 1000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
373
  "eval_samples_per_second": 116.708,
374
  "eval_steps_per_second": 2.451,
375
  "step": 500
376
+ },
377
+ {
378
+ "epoch": 0.08616320324379118,
379
+ "grad_norm": 0.7771898508071899,
380
+ "learning_rate": 7.635e-05,
381
+ "loss": 7.393017578125,
382
+ "step": 510
383
+ },
384
+ {
385
+ "epoch": 0.08785267781719885,
386
+ "grad_norm": 0.5564572811126709,
387
+ "learning_rate": 7.785e-05,
388
+ "loss": 7.3636009216308596,
389
+ "step": 520
390
+ },
391
+ {
392
+ "epoch": 0.08954215239060652,
393
+ "grad_norm": 0.5228703022003174,
394
+ "learning_rate": 7.934999999999999e-05,
395
+ "loss": 7.33685073852539,
396
+ "step": 530
397
+ },
398
+ {
399
+ "epoch": 0.0912316269640142,
400
+ "grad_norm": 0.8311993479728699,
401
+ "learning_rate": 8.085e-05,
402
+ "loss": 7.2986083984375,
403
+ "step": 540
404
+ },
405
+ {
406
+ "epoch": 0.09292110153742186,
407
+ "grad_norm": 0.7295158505439758,
408
+ "learning_rate": 8.235e-05,
409
+ "loss": 7.309038543701172,
410
+ "step": 550
411
+ },
412
+ {
413
+ "epoch": 0.09461057611082953,
414
+ "grad_norm": 0.7679043412208557,
415
+ "learning_rate": 8.385e-05,
416
+ "loss": 7.274485015869141,
417
+ "step": 560
418
+ },
419
+ {
420
+ "epoch": 0.0963000506842372,
421
+ "grad_norm": 0.6973268389701843,
422
+ "learning_rate": 8.534999999999999e-05,
423
+ "loss": 7.272534942626953,
424
+ "step": 570
425
+ },
426
+ {
427
+ "epoch": 0.09798952525764487,
428
+ "grad_norm": 0.5742895007133484,
429
+ "learning_rate": 8.684999999999998e-05,
430
+ "loss": 7.2565864562988285,
431
+ "step": 580
432
+ },
433
+ {
434
+ "epoch": 0.09967899983105254,
435
+ "grad_norm": 0.820361316204071,
436
+ "learning_rate": 8.834999999999999e-05,
437
+ "loss": 7.229536437988282,
438
+ "step": 590
439
+ },
440
+ {
441
+ "epoch": 0.10136847440446022,
442
+ "grad_norm": 0.7243834733963013,
443
+ "learning_rate": 8.984999999999999e-05,
444
+ "loss": 7.2052772521972654,
445
+ "step": 600
446
+ },
447
+ {
448
+ "epoch": 0.10305794897786788,
449
+ "grad_norm": 0.6259952187538147,
450
+ "learning_rate": 9.134999999999998e-05,
451
+ "loss": 7.208747863769531,
452
+ "step": 610
453
+ },
454
+ {
455
+ "epoch": 0.10474742355127556,
456
+ "grad_norm": 0.6420820951461792,
457
+ "learning_rate": 9.285e-05,
458
+ "loss": 7.16771240234375,
459
+ "step": 620
460
+ },
461
+ {
462
+ "epoch": 0.10643689812468322,
463
+ "grad_norm": 0.6610616445541382,
464
+ "learning_rate": 9.434999999999999e-05,
465
+ "loss": 7.163240051269531,
466
+ "step": 630
467
+ },
468
+ {
469
+ "epoch": 0.10812637269809089,
470
+ "grad_norm": 0.9573401212692261,
471
+ "learning_rate": 9.585e-05,
472
+ "loss": 7.147833251953125,
473
+ "step": 640
474
+ },
475
+ {
476
+ "epoch": 0.10981584727149857,
477
+ "grad_norm": 0.771244466304779,
478
+ "learning_rate": 9.735e-05,
479
+ "loss": 7.108192443847656,
480
+ "step": 650
481
+ },
482
+ {
483
+ "epoch": 0.11150532184490623,
484
+ "grad_norm": 0.8439337611198425,
485
+ "learning_rate": 9.884999999999999e-05,
486
+ "loss": 7.119499206542969,
487
+ "step": 660
488
+ },
489
+ {
490
+ "epoch": 0.1131947964183139,
491
+ "grad_norm": 0.8653160333633423,
492
+ "learning_rate": 0.00010035,
493
+ "loss": 7.084934234619141,
494
+ "step": 670
495
+ },
496
+ {
497
+ "epoch": 0.11488427099172158,
498
+ "grad_norm": 0.9962228536605835,
499
+ "learning_rate": 0.00010185,
500
+ "loss": 7.0695549011230465,
501
+ "step": 680
502
+ },
503
+ {
504
+ "epoch": 0.11657374556512924,
505
+ "grad_norm": 0.9841294288635254,
506
+ "learning_rate": 0.00010334999999999998,
507
+ "loss": 7.0598899841308596,
508
+ "step": 690
509
+ },
510
+ {
511
+ "epoch": 0.11826322013853692,
512
+ "grad_norm": 0.8194125890731812,
513
+ "learning_rate": 0.00010484999999999999,
514
+ "loss": 7.0336753845214846,
515
+ "step": 700
516
+ },
517
+ {
518
+ "epoch": 0.11995269471194459,
519
+ "grad_norm": 1.038486123085022,
520
+ "learning_rate": 0.00010634999999999998,
521
+ "loss": 7.024503326416015,
522
+ "step": 710
523
+ },
524
+ {
525
+ "epoch": 0.12164216928535225,
526
+ "grad_norm": 0.8891283273696899,
527
+ "learning_rate": 0.00010784999999999999,
528
+ "loss": 6.99432601928711,
529
+ "step": 720
530
+ },
531
+ {
532
+ "epoch": 0.12333164385875993,
533
+ "grad_norm": 0.8836078643798828,
534
+ "learning_rate": 0.00010934999999999999,
535
+ "loss": 6.993320465087891,
536
+ "step": 730
537
+ },
538
+ {
539
+ "epoch": 0.1250211184321676,
540
+ "grad_norm": 1.1054128408432007,
541
+ "learning_rate": 0.00011084999999999998,
542
+ "loss": 6.96356201171875,
543
+ "step": 740
544
+ },
545
+ {
546
+ "epoch": 0.12671059300557527,
547
+ "grad_norm": 0.7598650455474854,
548
+ "learning_rate": 0.00011235,
549
+ "loss": 6.965660095214844,
550
+ "step": 750
551
+ },
552
+ {
553
+ "epoch": 0.12840006757898292,
554
+ "grad_norm": 0.7678954005241394,
555
+ "learning_rate": 0.00011384999999999999,
556
+ "loss": 6.939690399169922,
557
+ "step": 760
558
+ },
559
+ {
560
+ "epoch": 0.1300895421523906,
561
+ "grad_norm": 0.9568160772323608,
562
+ "learning_rate": 0.00011535,
563
+ "loss": 6.922018432617188,
564
+ "step": 770
565
+ },
566
+ {
567
+ "epoch": 0.13177901672579828,
568
+ "grad_norm": 0.853614330291748,
569
+ "learning_rate": 0.00011685,
570
+ "loss": 6.915154266357422,
571
+ "step": 780
572
+ },
573
+ {
574
+ "epoch": 0.13346849129920596,
575
+ "grad_norm": 1.0338886976242065,
576
+ "learning_rate": 0.00011834999999999999,
577
+ "loss": 6.894224548339844,
578
+ "step": 790
579
+ },
580
+ {
581
+ "epoch": 0.1351579658726136,
582
+ "grad_norm": 1.0952171087265015,
583
+ "learning_rate": 0.00011985,
584
+ "loss": 6.892420959472656,
585
+ "step": 800
586
+ },
587
+ {
588
+ "epoch": 0.1368474404460213,
589
+ "grad_norm": 1.3561885356903076,
590
+ "learning_rate": 0.00012135,
591
+ "loss": 6.864668273925782,
592
+ "step": 810
593
+ },
594
+ {
595
+ "epoch": 0.13853691501942897,
596
+ "grad_norm": 0.9499754309654236,
597
+ "learning_rate": 0.00012284999999999998,
598
+ "loss": 6.874826049804687,
599
+ "step": 820
600
+ },
601
+ {
602
+ "epoch": 0.14022638959283662,
603
+ "grad_norm": 0.8440932035446167,
604
+ "learning_rate": 0.00012435,
605
+ "loss": 6.846308898925781,
606
+ "step": 830
607
+ },
608
+ {
609
+ "epoch": 0.1419158641662443,
610
+ "grad_norm": 0.8613818883895874,
611
+ "learning_rate": 0.00012585,
612
+ "loss": 6.831328582763672,
613
+ "step": 840
614
+ },
615
+ {
616
+ "epoch": 0.14360533873965198,
617
+ "grad_norm": 0.9079796075820923,
618
+ "learning_rate": 0.00012734999999999998,
619
+ "loss": 6.828762054443359,
620
+ "step": 850
621
+ },
622
+ {
623
+ "epoch": 0.14529481331305963,
624
+ "grad_norm": 0.9620388150215149,
625
+ "learning_rate": 0.00012885,
626
+ "loss": 6.814015197753906,
627
+ "step": 860
628
+ },
629
+ {
630
+ "epoch": 0.1469842878864673,
631
+ "grad_norm": 0.937064528465271,
632
+ "learning_rate": 0.00013035,
633
+ "loss": 6.802629852294922,
634
+ "step": 870
635
+ },
636
+ {
637
+ "epoch": 0.148673762459875,
638
+ "grad_norm": 0.915363609790802,
639
+ "learning_rate": 0.00013184999999999998,
640
+ "loss": 6.805186462402344,
641
+ "step": 880
642
+ },
643
+ {
644
+ "epoch": 0.15036323703328264,
645
+ "grad_norm": 0.9192573428153992,
646
+ "learning_rate": 0.00013335,
647
+ "loss": 6.74979248046875,
648
+ "step": 890
649
+ },
650
+ {
651
+ "epoch": 0.15205271160669032,
652
+ "grad_norm": 1.0629407167434692,
653
+ "learning_rate": 0.00013485,
654
+ "loss": 6.771487426757813,
655
+ "step": 900
656
+ },
657
+ {
658
+ "epoch": 0.153742186180098,
659
+ "grad_norm": 1.3261147737503052,
660
+ "learning_rate": 0.00013634999999999998,
661
+ "loss": 6.740608978271484,
662
+ "step": 910
663
+ },
664
+ {
665
+ "epoch": 0.15543166075350565,
666
+ "grad_norm": 1.0575299263000488,
667
+ "learning_rate": 0.00013785,
668
+ "loss": 6.7410430908203125,
669
+ "step": 920
670
+ },
671
+ {
672
+ "epoch": 0.15712113532691332,
673
+ "grad_norm": 1.2365992069244385,
674
+ "learning_rate": 0.00013935,
675
+ "loss": 6.7108512878417965,
676
+ "step": 930
677
+ },
678
+ {
679
+ "epoch": 0.158810609900321,
680
+ "grad_norm": 1.1486926078796387,
681
+ "learning_rate": 0.00014084999999999998,
682
+ "loss": 6.697355651855469,
683
+ "step": 940
684
+ },
685
+ {
686
+ "epoch": 0.16050008447372868,
687
+ "grad_norm": 1.2013252973556519,
688
+ "learning_rate": 0.00014235,
689
+ "loss": 6.674120330810547,
690
+ "step": 950
691
+ },
692
+ {
693
+ "epoch": 0.16218955904713633,
694
+ "grad_norm": 1.2197651863098145,
695
+ "learning_rate": 0.00014384999999999997,
696
+ "loss": 6.670748138427735,
697
+ "step": 960
698
+ },
699
+ {
700
+ "epoch": 0.163879033620544,
701
+ "grad_norm": 1.031817078590393,
702
+ "learning_rate": 0.00014534999999999998,
703
+ "loss": 6.645486450195312,
704
+ "step": 970
705
+ },
706
+ {
707
+ "epoch": 0.1655685081939517,
708
+ "grad_norm": 1.0909827947616577,
709
+ "learning_rate": 0.00014685,
710
+ "loss": 6.659033203125,
711
+ "step": 980
712
+ },
713
+ {
714
+ "epoch": 0.16725798276735934,
715
+ "grad_norm": 0.9495222568511963,
716
+ "learning_rate": 0.00014834999999999997,
717
+ "loss": 6.633167266845703,
718
+ "step": 990
719
+ },
720
+ {
721
+ "epoch": 0.16894745734076702,
722
+ "grad_norm": 1.0306235551834106,
723
+ "learning_rate": 0.00014984999999999998,
724
+ "loss": 6.637188720703125,
725
+ "step": 1000
726
+ },
727
+ {
728
+ "epoch": 0.16894745734076702,
729
+ "eval_loss": 6.614048957824707,
730
+ "eval_runtime": 3.666,
731
+ "eval_samples_per_second": 272.774,
732
+ "eval_steps_per_second": 5.728,
733
+ "step": 1000
734
  }
735
  ],
736
  "logging_steps": 10,
 
750
  "attributes": {}
751
  }
752
  },
753
+ "total_flos": 3.3445682085888e+16,
754
  "train_batch_size": 48,
755
  "trial_name": null,
756
  "trial_params": null