smirki commited on
Commit
a84d5f0
·
verified ·
1 Parent(s): 709f3da

Training in progress, step 550, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b251bf8f717d2eaad015081b982f4929a58a8d9661edee6c24f307df8efcdf3
3
  size 479005064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d02a8faa5615c04b1d54c94aeef6dc48105f780e4600e7eb5c4867bb627bdeab
3
  size 479005064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9cbe10b1599bcc43e19eaf8c8e8e5a8697c1f3e4e5b61977144271a6a8f7fed3
3
  size 958299770
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c5b0c837c7e20527f21e9fa9165cd1531678ee468a19f2840ff680cde906a4a
3
  size 958299770
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a0f93278d443d2a03483a44ae35cd09803c07ecb28455d1fbcbcd3d1ff96845
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d8dad906bd8d5d7d412b2e015c576ead9f79eec52fb6b8380d832bbdd892497
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a08ec2cb61042f6a4c8de11b52973f8ea774d1a9c9a71eab79348a724979acbf
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4855e84feafd6cb690dd0a0e35b0faeffec05476b505ea8cc93bf6ce3a96076f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.003982128208599804,
5
  "eval_steps": 500,
6
- "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -607,6 +607,66 @@
607
  "reward_std": 0.10076134353876114,
608
  "rewards/custom_reward_logic_v2": 0.13374999910593033,
609
  "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
610
  }
611
  ],
612
  "logging_steps": 10,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.004380341029459785,
5
  "eval_steps": 500,
6
+ "global_step": 550,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
607
  "reward_std": 0.10076134353876114,
608
  "rewards/custom_reward_logic_v2": 0.13374999910593033,
609
  "step": 500
610
+ },
611
+ {
612
+ "completion_length": 18.61875,
613
+ "epoch": 0.0040617707727718,
614
+ "grad_norm": 1.673584222793579,
615
+ "kl": 0.3592236742377281,
616
+ "learning_rate": 1.7482355012393177e-06,
617
+ "loss": 0.0144,
618
+ "reward": 0.14000000134110452,
619
+ "reward_std": 0.09233622029423713,
620
+ "rewards/custom_reward_logic_v2": 0.14000000134110452,
621
+ "step": 510
622
+ },
623
+ {
624
+ "completion_length": 20.41875,
625
+ "epoch": 0.004141413336943796,
626
+ "grad_norm": 1.1183210611343384,
627
+ "kl": 0.33706687912344935,
628
+ "learning_rate": 1.6449496416858285e-06,
629
+ "loss": 0.0135,
630
+ "reward": 0.10500000081956387,
631
+ "reward_std": 0.09869231358170509,
632
+ "rewards/custom_reward_logic_v2": 0.10500000081956387,
633
+ "step": 520
634
+ },
635
+ {
636
+ "completion_length": 20.8125,
637
+ "epoch": 0.004221055901115793,
638
+ "grad_norm": 0.4382721185684204,
639
+ "kl": 0.3610161267220974,
640
+ "learning_rate": 1.5432914190872757e-06,
641
+ "loss": 0.0144,
642
+ "reward": 0.19562500044703485,
643
+ "reward_std": 0.1188055507838726,
644
+ "rewards/custom_reward_logic_v2": 0.19562500044703485,
645
+ "step": 530
646
+ },
647
+ {
648
+ "completion_length": 19.46875,
649
+ "epoch": 0.004300698465287789,
650
+ "grad_norm": 1.3095043897628784,
651
+ "kl": 0.34188042730093005,
652
+ "learning_rate": 1.443454345648252e-06,
653
+ "loss": 0.0137,
654
+ "reward": 0.20749999955296516,
655
+ "reward_std": 0.19307591021060944,
656
+ "rewards/custom_reward_logic_v2": 0.20749999955296516,
657
+ "step": 540
658
+ },
659
+ {
660
+ "completion_length": 21.63125,
661
+ "epoch": 0.004380341029459785,
662
+ "grad_norm": 0.04259713739156723,
663
+ "kl": 0.3633933149278164,
664
+ "learning_rate": 1.3456284669124159e-06,
665
+ "loss": 0.0145,
666
+ "reward": 0.11875000149011612,
667
+ "reward_std": 0.1355846919119358,
668
+ "rewards/custom_reward_logic_v2": 0.11875000149011612,
669
+ "step": 550
670
  }
671
  ],
672
  "logging_steps": 10,