irishprancer commited on
Commit
4c7227b
·
verified ·
1 Parent(s): a7a1323

Training in progress, step 750, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ec2673b7ed27fb69ad96241bab728428b1bef12d399bb3f46a06da7f0da0ec9
3
  size 527048968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c199b9e3b512305005f55eb68da324ab293b3b8950e1f5a272d3930921d33d9b
3
  size 527048968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0654a3762b3e1bf30b2c72f526af9c46661bf182e89d1f3eb32221c06275d32c
3
  size 1054135994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b022a2a2d53b6a44873d81b2a118fb3a243719db4c3693bec47989187fe8332
3
  size 1054135994
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f76c4c2dca19bfb0a463693a0e409b44510488650d816e566bccd2a2851e9524
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ca9142fcbd976a2b9880762578e5776f18d9cad34016a627060f41ab78ec47d
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da0b9a1e2fa3da24e9af8b74787d70ff4c95d9bc9b74eeab135df0350a00b462
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1436fdc963b029d14915fc9410fa4da5908bac7c909f67f4bfe1e0807f32ecef
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7167752981185913,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
- "epoch": 19.565217391304348,
5
  "eval_steps": 150,
6
- "global_step": 450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -442,6 +442,296 @@
442
  "eval_samples_per_second": 25.331,
443
  "eval_steps_per_second": 25.331,
444
  "step": 450
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
445
  }
446
  ],
447
  "logging_steps": 10,
@@ -461,7 +751,7 @@
461
  "attributes": {}
462
  }
463
  },
464
- "total_flos": 1.1591800814075904e+16,
465
  "train_batch_size": 4,
466
  "trial_name": null,
467
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7167752981185913,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
+ "epoch": 32.608695652173914,
5
  "eval_steps": 150,
6
+ "global_step": 750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
442
  "eval_samples_per_second": 25.331,
443
  "eval_steps_per_second": 25.331,
444
  "step": 450
445
+ },
446
+ {
447
+ "epoch": 20.0,
448
+ "grad_norm": 2.1191132068634033,
449
+ "learning_rate": 2.9996147467351856e-05,
450
+ "loss": 0.515,
451
+ "step": 460
452
+ },
453
+ {
454
+ "epoch": 20.434782608695652,
455
+ "grad_norm": 1.2782835960388184,
456
+ "learning_rate": 2.9995930474939773e-05,
457
+ "loss": 0.4784,
458
+ "step": 470
459
+ },
460
+ {
461
+ "epoch": 20.869565217391305,
462
+ "grad_norm": 1.4754245281219482,
463
+ "learning_rate": 2.9995707538619975e-05,
464
+ "loss": 0.5705,
465
+ "step": 480
466
+ },
467
+ {
468
+ "epoch": 21.304347826086957,
469
+ "grad_norm": 1.322965383529663,
470
+ "learning_rate": 2.9995478658480822e-05,
471
+ "loss": 0.5162,
472
+ "step": 490
473
+ },
474
+ {
475
+ "epoch": 21.73913043478261,
476
+ "grad_norm": 1.2421406507492065,
477
+ "learning_rate": 2.9995243834613043e-05,
478
+ "loss": 0.5209,
479
+ "step": 500
480
+ },
481
+ {
482
+ "epoch": 22.17391304347826,
483
+ "grad_norm": 1.7905986309051514,
484
+ "learning_rate": 2.9995003067109707e-05,
485
+ "loss": 0.4834,
486
+ "step": 510
487
+ },
488
+ {
489
+ "epoch": 22.608695652173914,
490
+ "grad_norm": 1.5309374332427979,
491
+ "learning_rate": 2.9994756356066246e-05,
492
+ "loss": 0.5617,
493
+ "step": 520
494
+ },
495
+ {
496
+ "epoch": 23.043478260869566,
497
+ "grad_norm": 1.7404286861419678,
498
+ "learning_rate": 2.999450370158046e-05,
499
+ "loss": 0.4927,
500
+ "step": 530
501
+ },
502
+ {
503
+ "epoch": 23.47826086956522,
504
+ "grad_norm": 1.3119419813156128,
505
+ "learning_rate": 2.9994245103752478e-05,
506
+ "loss": 0.4384,
507
+ "step": 540
508
+ },
509
+ {
510
+ "epoch": 23.91304347826087,
511
+ "grad_norm": 1.2318671941757202,
512
+ "learning_rate": 2.999398056268481e-05,
513
+ "loss": 0.5268,
514
+ "step": 550
515
+ },
516
+ {
517
+ "epoch": 24.347826086956523,
518
+ "grad_norm": 1.4077017307281494,
519
+ "learning_rate": 2.9993710078482306e-05,
520
+ "loss": 0.5206,
521
+ "step": 560
522
+ },
523
+ {
524
+ "epoch": 24.782608695652176,
525
+ "grad_norm": 0.9560300707817078,
526
+ "learning_rate": 2.9993433651252185e-05,
527
+ "loss": 0.4429,
528
+ "step": 570
529
+ },
530
+ {
531
+ "epoch": 25.217391304347824,
532
+ "grad_norm": 1.7175792455673218,
533
+ "learning_rate": 2.9993151281104006e-05,
534
+ "loss": 0.5326,
535
+ "step": 580
536
+ },
537
+ {
538
+ "epoch": 25.652173913043477,
539
+ "grad_norm": 1.1363499164581299,
540
+ "learning_rate": 2.9992862968149695e-05,
541
+ "loss": 0.4734,
542
+ "step": 590
543
+ },
544
+ {
545
+ "epoch": 26.08695652173913,
546
+ "grad_norm": 1.1709671020507812,
547
+ "learning_rate": 2.9992568712503533e-05,
548
+ "loss": 0.4608,
549
+ "step": 600
550
+ },
551
+ {
552
+ "epoch": 26.08695652173913,
553
+ "eval_loss": 0.7204815149307251,
554
+ "eval_runtime": 0.4403,
555
+ "eval_samples_per_second": 22.71,
556
+ "eval_steps_per_second": 22.71,
557
+ "step": 600
558
+ },
559
+ {
560
+ "epoch": 26.08695652173913,
561
+ "eval_loss": 0.8609212040901184,
562
+ "eval_runtime": 0.5019,
563
+ "eval_samples_per_second": 19.926,
564
+ "eval_steps_per_second": 19.926,
565
+ "step": 600
566
+ },
567
+ {
568
+ "epoch": 26.08695652173913,
569
+ "eval_loss": 0.7204815149307251,
570
+ "eval_runtime": 0.5015,
571
+ "eval_samples_per_second": 19.942,
572
+ "eval_steps_per_second": 19.942,
573
+ "step": 600
574
+ },
575
+ {
576
+ "epoch": 26.08695652173913,
577
+ "eval_loss": 0.7344802618026733,
578
+ "eval_runtime": 0.4734,
579
+ "eval_samples_per_second": 21.126,
580
+ "eval_steps_per_second": 21.126,
581
+ "step": 600
582
+ },
583
+ {
584
+ "epoch": 26.08695652173913,
585
+ "eval_loss": 0.8617879748344421,
586
+ "eval_runtime": 0.4638,
587
+ "eval_samples_per_second": 21.559,
588
+ "eval_steps_per_second": 21.559,
589
+ "step": 600
590
+ },
591
+ {
592
+ "epoch": 26.52173913043478,
593
+ "grad_norm": 1.0739339590072632,
594
+ "learning_rate": 2.171901642542767e-06,
595
+ "loss": 0.5121,
596
+ "step": 610
597
+ },
598
+ {
599
+ "epoch": 26.956521739130434,
600
+ "grad_norm": 1.3133119344711304,
601
+ "learning_rate": 4.343803285085534e-06,
602
+ "loss": 0.4308,
603
+ "step": 620
604
+ },
605
+ {
606
+ "epoch": 27.391304347826086,
607
+ "grad_norm": 1.1951584815979004,
608
+ "learning_rate": 6.5157049276283e-06,
609
+ "loss": 0.4829,
610
+ "step": 630
611
+ },
612
+ {
613
+ "epoch": 27.82608695652174,
614
+ "grad_norm": 1.3615652322769165,
615
+ "learning_rate": 8.687606570171068e-06,
616
+ "loss": 0.5071,
617
+ "step": 640
618
+ },
619
+ {
620
+ "epoch": 28.26086956521739,
621
+ "grad_norm": 1.3934813737869263,
622
+ "learning_rate": 1.0859508212713834e-05,
623
+ "loss": 0.4195,
624
+ "step": 650
625
+ },
626
+ {
627
+ "epoch": 28.695652173913043,
628
+ "grad_norm": 1.5364168882369995,
629
+ "learning_rate": 1.30314098552566e-05,
630
+ "loss": 0.5198,
631
+ "step": 660
632
+ },
633
+ {
634
+ "epoch": 29.130434782608695,
635
+ "grad_norm": 1.291927695274353,
636
+ "learning_rate": 1.5203311497799366e-05,
637
+ "loss": 0.4474,
638
+ "step": 670
639
+ },
640
+ {
641
+ "epoch": 29.565217391304348,
642
+ "grad_norm": 1.0491178035736084,
643
+ "learning_rate": 1.7375213140342136e-05,
644
+ "loss": 0.489,
645
+ "step": 680
646
+ },
647
+ {
648
+ "epoch": 30.0,
649
+ "grad_norm": 2.413468360900879,
650
+ "learning_rate": 1.9547114782884902e-05,
651
+ "loss": 0.4297,
652
+ "step": 690
653
+ },
654
+ {
655
+ "epoch": 30.434782608695652,
656
+ "grad_norm": 1.3009700775146484,
657
+ "learning_rate": 2.1719016425427668e-05,
658
+ "loss": 0.498,
659
+ "step": 700
660
+ },
661
+ {
662
+ "epoch": 30.869565217391305,
663
+ "grad_norm": 1.5959917306900024,
664
+ "learning_rate": 2.1719014273246623e-05,
665
+ "loss": 0.4323,
666
+ "step": 710
667
+ },
668
+ {
669
+ "epoch": 31.304347826086957,
670
+ "grad_norm": 0.9005913138389587,
671
+ "learning_rate": 2.171900781670434e-05,
672
+ "loss": 0.4452,
673
+ "step": 720
674
+ },
675
+ {
676
+ "epoch": 31.73913043478261,
677
+ "grad_norm": 1.5817768573760986,
678
+ "learning_rate": 2.1718997055803376e-05,
679
+ "loss": 0.5078,
680
+ "step": 730
681
+ },
682
+ {
683
+ "epoch": 32.17391304347826,
684
+ "grad_norm": 1.16550874710083,
685
+ "learning_rate": 2.1718981990548e-05,
686
+ "loss": 0.4618,
687
+ "step": 740
688
+ },
689
+ {
690
+ "epoch": 32.608695652173914,
691
+ "grad_norm": 1.7140876054763794,
692
+ "learning_rate": 2.1718962620944183e-05,
693
+ "loss": 0.5004,
694
+ "step": 750
695
+ },
696
+ {
697
+ "epoch": 32.608695652173914,
698
+ "eval_loss": 0.7238079309463501,
699
+ "eval_runtime": 0.3947,
700
+ "eval_samples_per_second": 25.334,
701
+ "eval_steps_per_second": 25.334,
702
+ "step": 750
703
+ },
704
+ {
705
+ "epoch": 32.608695652173914,
706
+ "eval_loss": 0.8609212040901184,
707
+ "eval_runtime": 0.3993,
708
+ "eval_samples_per_second": 25.042,
709
+ "eval_steps_per_second": 25.042,
710
+ "step": 750
711
+ },
712
+ {
713
+ "epoch": 32.608695652173914,
714
+ "eval_loss": 0.7238079309463501,
715
+ "eval_runtime": 0.389,
716
+ "eval_samples_per_second": 25.705,
717
+ "eval_steps_per_second": 25.705,
718
+ "step": 750
719
+ },
720
+ {
721
+ "epoch": 32.608695652173914,
722
+ "eval_loss": 0.729554295539856,
723
+ "eval_runtime": 0.3906,
724
+ "eval_samples_per_second": 25.605,
725
+ "eval_steps_per_second": 25.605,
726
+ "step": 750
727
+ },
728
+ {
729
+ "epoch": 32.608695652173914,
730
+ "eval_loss": 0.8608381152153015,
731
+ "eval_runtime": 0.391,
732
+ "eval_samples_per_second": 25.574,
733
+ "eval_steps_per_second": 25.574,
734
+ "step": 750
735
  }
736
  ],
737
  "logging_steps": 10,
 
751
  "attributes": {}
752
  }
753
  },
754
+ "total_flos": 1.9231766387195904e+16,
755
  "train_batch_size": 4,
756
  "trial_name": null,
757
  "trial_params": null