smirki commited on
Commit
c62280e
·
verified ·
1 Parent(s): 817c483

Training in progress, step 600, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d02a8faa5615c04b1d54c94aeef6dc48105f780e4600e7eb5c4867bb627bdeab
3
  size 479005064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8552145847bfeab4636144433c43db62dcc91675e6d80bf8a61ef9290e9fc440
3
  size 479005064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c5b0c837c7e20527f21e9fa9165cd1531678ee468a19f2840ff680cde906a4a
3
  size 958299770
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eea82d271c17fd83a4656876cadcc33c1faa5ce883f21449769fbb145e1864a9
3
  size 958299770
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d8dad906bd8d5d7d412b2e015c576ead9f79eec52fb6b8380d832bbdd892497
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee66f0b7a716e07b51f02c7e7cb42ecc8627dd065518adcf55ce909d0ef818d8
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4855e84feafd6cb690dd0a0e35b0faeffec05476b505ea8cc93bf6ce3a96076f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c12eb55eb955504de5fa1d8a599d32de65813670b5001ef30e45d7a6b99dbe1f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.004380341029459785,
5
  "eval_steps": 500,
6
- "global_step": 550,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -667,6 +667,66 @@
667
  "reward_std": 0.1355846919119358,
668
  "rewards/custom_reward_logic_v2": 0.11875000149011612,
669
  "step": 550
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
670
  }
671
  ],
672
  "logging_steps": 10,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.004778553850319765,
5
  "eval_steps": 500,
6
+ "global_step": 600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
667
  "reward_std": 0.1355846919119358,
668
  "rewards/custom_reward_logic_v2": 0.11875000149011612,
669
  "step": 550
670
+ },
671
+ {
672
+ "completion_length": 19.2625,
673
+ "epoch": 0.004459983593631781,
674
+ "grad_norm": 0.5744329690933228,
675
+ "kl": 0.3458960048854351,
676
+ "learning_rate": 1.2500000000000007e-06,
677
+ "loss": 0.0138,
678
+ "reward": 0.3293750025331974,
679
+ "reward_std": 0.104243653267622,
680
+ "rewards/custom_reward_logic_v2": 0.3293750025331974,
681
+ "step": 560
682
+ },
683
+ {
684
+ "completion_length": 20.05625,
685
+ "epoch": 0.004539626157803777,
686
+ "grad_norm": 0.1549508273601532,
687
+ "kl": 0.346449576318264,
688
+ "learning_rate": 1.1567509791329402e-06,
689
+ "loss": 0.0139,
690
+ "reward": 0.17625000029802323,
691
+ "reward_std": 0.1429968483746052,
692
+ "rewards/custom_reward_logic_v2": 0.17625000029802323,
693
+ "step": 570
694
+ },
695
+ {
696
+ "completion_length": 19.86875,
697
+ "epoch": 0.004619268721975773,
698
+ "grad_norm": 0.14351911842823029,
699
+ "kl": 0.39532790407538415,
700
+ "learning_rate": 1.0660589091223854e-06,
701
+ "loss": 0.0158,
702
+ "reward": 0.17437500022351743,
703
+ "reward_std": 0.13940104842185974,
704
+ "rewards/custom_reward_logic_v2": 0.17437500022351743,
705
+ "step": 580
706
+ },
707
+ {
708
+ "completion_length": 19.28125,
709
+ "epoch": 0.004698911286147769,
710
+ "grad_norm": 1.1975979804992676,
711
+ "kl": 0.3690756544470787,
712
+ "learning_rate": 9.780964274781984e-07,
713
+ "loss": 0.0148,
714
+ "reward": 0.20562500059604644,
715
+ "reward_std": 0.11339747980237007,
716
+ "rewards/custom_reward_logic_v2": 0.20562500059604644,
717
+ "step": 590
718
+ },
719
+ {
720
+ "completion_length": 18.3125,
721
+ "epoch": 0.004778553850319765,
722
+ "grad_norm": 0.03664500266313553,
723
+ "kl": 0.34169030636548997,
724
+ "learning_rate": 8.930309757836517e-07,
725
+ "loss": 0.0137,
726
+ "reward": 0.12624999806284903,
727
+ "reward_std": 0.1257291093468666,
728
+ "rewards/custom_reward_logic_v2": 0.12624999806284903,
729
+ "step": 600
730
  }
731
  ],
732
  "logging_steps": 10,