smirki commited on
Commit
52fd054
·
verified ·
1 Parent(s): 1a4f269

Training in progress, step 500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff241d617f452617d3c88f6cf63325935d4668c5daf5e939cf72829f92a221d4
3
  size 479005064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b251bf8f717d2eaad015081b982f4929a58a8d9661edee6c24f307df8efcdf3
3
  size 479005064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1870f17179349635d5e9f8dc264de59804f82c37fb07d8c3e59aa2307ca2761a
3
  size 958299770
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cbe10b1599bcc43e19eaf8c8e8e5a8697c1f3e4e5b61977144271a6a8f7fed3
3
  size 958299770
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb5fd2bad97b0505323e94353254cc1ebf7b1bbfc6106a6a54db604a83d738b3
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a0f93278d443d2a03483a44ae35cd09803c07ecb28455d1fbcbcd3d1ff96845
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf5539fcaa2354df9ab3f6e01030e649c726b26d278c21ce5b06138c07200265
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a08ec2cb61042f6a4c8de11b52973f8ea774d1a9c9a71eab79348a724979acbf
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0035839153877398238,
5
  "eval_steps": 500,
6
- "global_step": 450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -547,6 +547,66 @@
547
  "reward_std": 0.10089804157614708,
548
  "rewards/custom_reward_logic_v2": 0.11125000230967999,
549
  "step": 450
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
550
  }
551
  ],
552
  "logging_steps": 10,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.003982128208599804,
5
  "eval_steps": 500,
6
+ "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
547
  "reward_std": 0.10089804157614708,
548
  "rewards/custom_reward_logic_v2": 0.11125000230967999,
549
  "step": 450
550
+ },
551
+ {
552
+ "completion_length": 20.14375,
553
+ "epoch": 0.0036635579519118198,
554
+ "grad_norm": 0.7745999693870544,
555
+ "kl": 0.3447819516062737,
556
+ "learning_rate": 2.2821106431308546e-06,
557
+ "loss": 0.0138,
558
+ "reward": 0.11187500096857547,
559
+ "reward_std": 0.12871785834431648,
560
+ "rewards/custom_reward_logic_v2": 0.11187500096857547,
561
+ "step": 460
562
+ },
563
+ {
564
+ "completion_length": 27.35,
565
+ "epoch": 0.003743200516083816,
566
+ "grad_norm": 1.4974488019943237,
567
+ "kl": 0.3425402037799358,
568
+ "learning_rate": 2.173684519449872e-06,
569
+ "loss": 0.0137,
570
+ "reward": 0.10542500019073486,
571
+ "reward_std": 0.22862085253000258,
572
+ "rewards/custom_reward_logic_v2": 0.10542500019073486,
573
+ "step": 470
574
+ },
575
+ {
576
+ "completion_length": 19.40625,
577
+ "epoch": 0.003822843080255812,
578
+ "grad_norm": 1.1579034328460693,
579
+ "kl": 0.3382424309849739,
580
+ "learning_rate": 2.0658795558326745e-06,
581
+ "loss": 0.0135,
582
+ "reward": 0.15562499947845937,
583
+ "reward_std": 0.1210292175412178,
584
+ "rewards/custom_reward_logic_v2": 0.15562499947845937,
585
+ "step": 480
586
+ },
587
+ {
588
+ "completion_length": 37.24375,
589
+ "epoch": 0.003902485644427808,
590
+ "grad_norm": 0.7052723169326782,
591
+ "kl": 0.35214473977684974,
592
+ "learning_rate": 1.958900965154743e-06,
593
+ "loss": 0.0141,
594
+ "reward": 0.07051250115036964,
595
+ "reward_std": 0.20176818892359732,
596
+ "rewards/custom_reward_logic_v2": 0.07051250115036964,
597
+ "step": 490
598
+ },
599
+ {
600
+ "completion_length": 19.23125,
601
+ "epoch": 0.003982128208599804,
602
+ "grad_norm": 0.062097422778606415,
603
+ "kl": 0.3568013899028301,
604
+ "learning_rate": 1.852952387243698e-06,
605
+ "loss": 0.0143,
606
+ "reward": 0.13374999910593033,
607
+ "reward_std": 0.10076134353876114,
608
+ "rewards/custom_reward_logic_v2": 0.13374999910593033,
609
+ "step": 500
610
  }
611
  ],
612
  "logging_steps": 10,