smirki commited on
Commit
a38be7e
·
verified ·
1 Parent(s): 3d1ab2a

Training in progress, step 450, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a656f61ae1ccd60f5456798788c8e6ded39b418346d6fe6071eb8ce468298a83
3
  size 479005064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff241d617f452617d3c88f6cf63325935d4668c5daf5e939cf72829f92a221d4
3
  size 479005064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf849eb44752b2c9a42282b0505698ef4c3181196d5a7e2c98af9389f0ba6ef6
3
  size 958299770
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1870f17179349635d5e9f8dc264de59804f82c37fb07d8c3e59aa2307ca2761a
3
  size 958299770
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd3ccbab0f5afd279e81643827121534f4d8480da4449b3d948c3974c92f9bbd
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb5fd2bad97b0505323e94353254cc1ebf7b1bbfc6106a6a54db604a83d738b3
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8241f030a944f045c9ceca4728f16fdc2a35629a475d8a3997b745440061990
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf5539fcaa2354df9ab3f6e01030e649c726b26d278c21ce5b06138c07200265
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0031857025668798433,
5
  "eval_steps": 500,
6
- "global_step": 400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -487,6 +487,66 @@
487
  "reward_std": 0.09467698186635971,
488
  "rewards/custom_reward_logic_v2": 0.09312500022351741,
489
  "step": 400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
490
  }
491
  ],
492
  "logging_steps": 10,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.0035839153877398238,
5
  "eval_steps": 500,
6
+ "global_step": 450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
487
  "reward_std": 0.09467698186635971,
488
  "rewards/custom_reward_logic_v2": 0.09312500022351741,
489
  "step": 400
490
+ },
491
+ {
492
+ "completion_length": 21.9125,
493
+ "epoch": 0.0032653451310518393,
494
+ "grad_norm": 0.12919628620147705,
495
+ "kl": 0.3515960440039635,
496
+ "learning_rate": 2.82631548055013e-06,
497
+ "loss": 0.0141,
498
+ "reward": 0.07376250103116036,
499
+ "reward_std": 0.15707473903894426,
500
+ "rewards/custom_reward_logic_v2": 0.07376250103116036,
501
+ "step": 410
502
+ },
503
+ {
504
+ "completion_length": 20.325,
505
+ "epoch": 0.0033449876952238353,
506
+ "grad_norm": 0.08202961087226868,
507
+ "kl": 0.34852803200483323,
508
+ "learning_rate": 2.717889356869146e-06,
509
+ "loss": 0.0139,
510
+ "reward": 0.2074999999254942,
511
+ "reward_std": 0.11485048606991768,
512
+ "rewards/custom_reward_logic_v2": 0.2074999999254942,
513
+ "step": 420
514
+ },
515
+ {
516
+ "completion_length": 19.7375,
517
+ "epoch": 0.0034246302593958313,
518
+ "grad_norm": 1.5309367179870605,
519
+ "kl": 0.339575307816267,
520
+ "learning_rate": 2.6090484684133406e-06,
521
+ "loss": 0.0136,
522
+ "reward": 0.03375000059604645,
523
+ "reward_std": 0.08364979848265648,
524
+ "rewards/custom_reward_logic_v2": 0.03375000059604645,
525
+ "step": 430
526
+ },
527
+ {
528
+ "completion_length": 29.5375,
529
+ "epoch": 0.0035042728235678278,
530
+ "grad_norm": 0.08438611030578613,
531
+ "kl": 0.34286700189113617,
532
+ "learning_rate": 2.5e-06,
533
+ "loss": 0.0137,
534
+ "reward": 0.08044999912381172,
535
+ "reward_std": 0.16319628208875656,
536
+ "rewards/custom_reward_logic_v2": 0.08044999912381172,
537
+ "step": 440
538
+ },
539
+ {
540
+ "completion_length": 23.2125,
541
+ "epoch": 0.0035839153877398238,
542
+ "grad_norm": 0.11283387243747711,
543
+ "kl": 0.3263735562562943,
544
+ "learning_rate": 2.3909515315866606e-06,
545
+ "loss": 0.0131,
546
+ "reward": 0.11125000230967999,
547
+ "reward_std": 0.10089804157614708,
548
+ "rewards/custom_reward_logic_v2": 0.11125000230967999,
549
+ "step": 450
550
  }
551
  ],
552
  "logging_steps": 10,