Rakhman16 commited on
Commit
d9c943b
·
verified ·
1 Parent(s): 5e09225

Training in progress, step 3000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a6445366fb72dd64a30a112d0cfda302649ce8c29f7108f12ace64035585c67
3
  size 891558696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8050d0ba4cc80419a583c93968e6bf55216baf5e1371c2c9e7133fa0b1464ed
3
  size 891558696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ebbdc69304c59243d4149daaa710b1485f7a643fa3ccf2294412ebf232abae9
3
  size 1783272762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b05d6d3a797d87d258d9bcbc672b81e5538c72ed7f79e2144c48b29830d47b1
3
  size 1783272762
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca2a334725c06878338083a325b1190cf2f41e51415271d4e791a39c574e91d1
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee21e9bc6023190a0db96d08050b5e5b20632d2971675b4236e71d6b2aa60903
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3eedbad439d02a3d9e01c697324385512054df489e38825f0e26bf6d99fe88ac
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4a50e6bc9278ed49513d0c8109e953279561a58261b90ebf3bd479594596325
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.20481644570827484,
3
- "best_model_checkpoint": "./fine-tuned/checkpoint-2500",
4
- "epoch": 1.7562346329469616,
5
  "eval_steps": 100,
6
- "global_step": 2500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -557,6 +557,116 @@
557
  "eval_samples_per_second": 66.462,
558
  "eval_steps_per_second": 2.086,
559
  "step": 2500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
560
  }
561
  ],
562
  "logging_steps": 50,
@@ -576,7 +686,7 @@
576
  "attributes": {}
577
  }
578
  },
579
- "total_flos": 2.43560320303104e+16,
580
  "train_batch_size": 32,
581
  "trial_name": null,
582
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.2037852257490158,
3
+ "best_model_checkpoint": "./fine-tuned/checkpoint-3000",
4
+ "epoch": 2.107481559536354,
5
  "eval_steps": 100,
6
+ "global_step": 3000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
557
  "eval_samples_per_second": 66.462,
558
  "eval_steps_per_second": 2.086,
559
  "step": 2500
560
+ },
561
+ {
562
+ "epoch": 1.7913593256059008,
563
+ "grad_norm": 20207.35546875,
564
+ "learning_rate": 1.6560084328882643e-05,
565
+ "loss": 0.2049,
566
+ "step": 2550
567
+ },
568
+ {
569
+ "epoch": 1.82648401826484,
570
+ "grad_norm": 17453.359375,
571
+ "learning_rate": 1.6296556570625438e-05,
572
+ "loss": 0.2101,
573
+ "step": 2600
574
+ },
575
+ {
576
+ "epoch": 1.82648401826484,
577
+ "eval_loss": 0.20485134422779083,
578
+ "eval_runtime": 67.202,
579
+ "eval_samples_per_second": 66.367,
580
+ "eval_steps_per_second": 2.083,
581
+ "step": 2600
582
+ },
583
+ {
584
+ "epoch": 1.8616087109237793,
585
+ "grad_norm": 24568.439453125,
586
+ "learning_rate": 1.603302881236824e-05,
587
+ "loss": 0.2081,
588
+ "step": 2650
589
+ },
590
+ {
591
+ "epoch": 1.8967334035827186,
592
+ "grad_norm": 22425.1875,
593
+ "learning_rate": 1.5769501054111034e-05,
594
+ "loss": 0.2032,
595
+ "step": 2700
596
+ },
597
+ {
598
+ "epoch": 1.8967334035827186,
599
+ "eval_loss": 0.2041337788105011,
600
+ "eval_runtime": 67.2372,
601
+ "eval_samples_per_second": 66.332,
602
+ "eval_steps_per_second": 2.082,
603
+ "step": 2700
604
+ },
605
+ {
606
+ "epoch": 1.9318580962416578,
607
+ "grad_norm": 21858.3828125,
608
+ "learning_rate": 1.550597329585383e-05,
609
+ "loss": 0.2074,
610
+ "step": 2750
611
+ },
612
+ {
613
+ "epoch": 1.966982788900597,
614
+ "grad_norm": 17712.39453125,
615
+ "learning_rate": 1.5242445537596626e-05,
616
+ "loss": 0.205,
617
+ "step": 2800
618
+ },
619
+ {
620
+ "epoch": 1.966982788900597,
621
+ "eval_loss": 0.20371712744235992,
622
+ "eval_runtime": 67.1299,
623
+ "eval_samples_per_second": 66.438,
624
+ "eval_steps_per_second": 2.086,
625
+ "step": 2800
626
+ },
627
+ {
628
+ "epoch": 2.0021074815595363,
629
+ "grad_norm": 20413.91796875,
630
+ "learning_rate": 1.4978917779339424e-05,
631
+ "loss": 0.203,
632
+ "step": 2850
633
+ },
634
+ {
635
+ "epoch": 2.0372321742184756,
636
+ "grad_norm": 21380.130859375,
637
+ "learning_rate": 1.471539002108222e-05,
638
+ "loss": 0.199,
639
+ "step": 2900
640
+ },
641
+ {
642
+ "epoch": 2.0372321742184756,
643
+ "eval_loss": 0.20416177809238434,
644
+ "eval_runtime": 67.1771,
645
+ "eval_samples_per_second": 66.392,
646
+ "eval_steps_per_second": 2.084,
647
+ "step": 2900
648
+ },
649
+ {
650
+ "epoch": 2.072356866877415,
651
+ "grad_norm": 28436.697265625,
652
+ "learning_rate": 1.4451862262825019e-05,
653
+ "loss": 0.1989,
654
+ "step": 2950
655
+ },
656
+ {
657
+ "epoch": 2.107481559536354,
658
+ "grad_norm": 18739.8359375,
659
+ "learning_rate": 1.4188334504567815e-05,
660
+ "loss": 0.1982,
661
+ "step": 3000
662
+ },
663
+ {
664
+ "epoch": 2.107481559536354,
665
+ "eval_loss": 0.2037852257490158,
666
+ "eval_runtime": 67.2417,
667
+ "eval_samples_per_second": 66.328,
668
+ "eval_steps_per_second": 2.082,
669
+ "step": 3000
670
  }
671
  ],
672
  "logging_steps": 50,
 
686
  "attributes": {}
687
  }
688
  },
689
+ "total_flos": 2.92254115627008e+16,
690
  "train_batch_size": 32,
691
  "trial_name": null,
692
  "trial_params": null