bitsoko commited on
Commit
ea2f81f
·
verified ·
1 Parent(s): 27786db

Training in progress, step 250, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c09ea6ce50093c41f2f5fb8b4e6414b0c577151e86f47fe18188690e6492eb39
3
  size 119597408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb1045a6588100f77da2fc367c78dce331d79e23ec98420f4290693e5657de94
3
  size 119597408
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39ca64088c14fde06f9661cede8c9942fee2ff16eb48934cb42c77609210f140
3
  size 60386772
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f72f392d6528b4317aef9fefb0c2d13f84965b56617fbc00a465411e33a0ab1
3
  size 60386772
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:caf99753320d5a922d21785c9acff8aad90fb47892dad1f221c94717bf3a04b1
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8d87e9c1f179372f9242367ad770c30c7bd7eeb48af101c530d591d3c29dd85
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21226719b9e25fe4304ec632d533e056eb568a7272feb0ef7163c162da3912fa
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b00c955337268a520a0035ea66591799a038b74d25acad9619f1a1a47155f3ce
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.012270691453463402,
5
  "eval_steps": 20,
6
- "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -157,6 +157,36 @@
157
  "eval_samples_per_second": 5.074,
158
  "eval_steps_per_second": 0.66,
159
  "step": 200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
160
  }
161
  ],
162
  "logging_steps": 20,
@@ -164,7 +194,7 @@
164
  "num_input_tokens_seen": 0,
165
  "num_train_epochs": 3,
166
  "save_steps": 50,
167
- "total_flos": 1.25490630177792e+16,
168
  "train_batch_size": 2,
169
  "trial_name": null,
170
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.015338364316829254,
5
  "eval_steps": 20,
6
+ "global_step": 250,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
157
  "eval_samples_per_second": 5.074,
158
  "eval_steps_per_second": 0.66,
159
  "step": 200
160
+ },
161
+ {
162
+ "epoch": 0.013497760598809742,
163
+ "grad_norm": 0.8630465269088745,
164
+ "learning_rate": 0.00019912051051296735,
165
+ "loss": 1.64,
166
+ "step": 220
167
+ },
168
+ {
169
+ "epoch": 0.013497760598809742,
170
+ "eval_loss": 1.5652003288269043,
171
+ "eval_runtime": 19.9102,
172
+ "eval_samples_per_second": 5.023,
173
+ "eval_steps_per_second": 0.653,
174
+ "step": 220
175
+ },
176
+ {
177
+ "epoch": 0.014724829744156084,
178
+ "grad_norm": 0.7266297936439514,
179
+ "learning_rate": 0.00019903869753742944,
180
+ "loss": 1.6705,
181
+ "step": 240
182
+ },
183
+ {
184
+ "epoch": 0.014724829744156084,
185
+ "eval_loss": 1.5418590307235718,
186
+ "eval_runtime": 19.7135,
187
+ "eval_samples_per_second": 5.073,
188
+ "eval_steps_per_second": 0.659,
189
+ "step": 240
190
  }
191
  ],
192
  "logging_steps": 20,
 
194
  "num_input_tokens_seen": 0,
195
  "num_train_epochs": 3,
196
  "save_steps": 50,
197
+ "total_flos": 1.568681283999744e+16,
198
  "train_batch_size": 2,
199
  "trial_name": null,
200
  "trial_params": null