smirki commited on
Commit
86da38c
·
verified ·
1 Parent(s): 98f21ab

Training in progress, step 1700, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2113bb192f2baa5cca5eaf82cdc29cc112a196f0d353c65e596bbc93a9194868
3
  size 479005064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb4a26b564992bfe1b0c0cbf9877d71a7e4c4cabeff18b9fcdc775fdba1b254f
3
  size 479005064
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d39914c7d80371a45e6dfe8ca4f4df8415bdc590cf39528b2981db04c2d9b5ab
3
  size 958299770
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bda5b6a4da0480b964d69c900d61e56ab25544d0c3fbb12fa357689206e9625d
3
  size 958299770
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3abc69817bfc7e8384f353860e5ac5df7fe0bc25efa8cefc94df527ed53e449a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21e6e0ff9935bb592aa01d75fb327a21b659c1135b9f0f607483a3ede13d0d3b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3aa9ee72eac8c93f8b4aad2310d8b5495ea694aa75e620528535e72bb4aabf2e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6a11a087092ce528653bb5e030a3a014f4a88c8ff1a2a24bc7e40980baafb9d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.013340129498809344,
5
  "eval_steps": 500,
6
- "global_step": 1675,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2011,6 +2011,42 @@
2011
  "reward_std": 0.26727318242192266,
2012
  "rewards/custom_reward_simplified_v7_dblog": 0.6375,
2013
  "step": 1670
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2014
  }
2015
  ],
2016
  "logging_steps": 10,
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.013539235909239333,
5
  "eval_steps": 500,
6
+ "global_step": 1700,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2011
  "reward_std": 0.26727318242192266,
2012
  "rewards/custom_reward_simplified_v7_dblog": 0.6375,
2013
  "step": 1670
2014
+ },
2015
+ {
2016
+ "completion_length": 655.70625,
2017
+ "epoch": 0.013379950780895341,
2018
+ "grad_norm": 0.2751936614513397,
2019
+ "kl": 0.01622524333652109,
2020
+ "learning_rate": 1.2500000000000007e-06,
2021
+ "loss": 0.0006,
2022
+ "reward": 0.878125,
2023
+ "reward_std": 0.284642493724823,
2024
+ "rewards/custom_reward_simplified_v7_dblog": 0.878125,
2025
+ "step": 1680
2026
+ },
2027
+ {
2028
+ "completion_length": 684.98125,
2029
+ "epoch": 0.013459593345067338,
2030
+ "grad_norm": 0.23118546605110168,
2031
+ "kl": 0.01642036633566022,
2032
+ "learning_rate": 1.218643748000337e-06,
2033
+ "loss": 0.0007,
2034
+ "reward": 0.85625,
2035
+ "reward_std": 0.339317075163126,
2036
+ "rewards/custom_reward_simplified_v7_dblog": 0.85625,
2037
+ "step": 1690
2038
+ },
2039
+ {
2040
+ "completion_length": 743.51875,
2041
+ "epoch": 0.013539235909239333,
2042
+ "grad_norm": 0.22867274284362793,
2043
+ "kl": 0.01721250016707927,
2044
+ "learning_rate": 1.1875585491636e-06,
2045
+ "loss": 0.0007,
2046
+ "reward": 0.653125,
2047
+ "reward_std": 0.277196903526783,
2048
+ "rewards/custom_reward_simplified_v7_dblog": 0.653125,
2049
+ "step": 1700
2050
  }
2051
  ],
2052
  "logging_steps": 10,