bitsoko commited on
Commit
d14d2b6
·
verified ·
1 Parent(s): 5b80a3c

Training in progress, step 350, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b984dfd07820164946065ae23ca85bbdb9fd648664a601e3bf20a76b6d205e18
3
  size 1912664024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f55186aef0f4403758d9ce629cd7dca80608ab1157777195225659ae9cab27c
3
  size 1912664024
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7645326d3fd71a6ea46ef71475c21611715db65d24641885ad9b51b20c73a24d
3
  size 958697812
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ac6d376a99dca2c862d044dc6bf54e25ea4b0e9b7d5fb64d21cc77371fcbd7e
3
  size 958697812
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c22b464bb859e79260348791ff89706887d2ac8a613c40a3bf13d707af961dcf
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:facdb6d766a802a4f41ae7ca1bf9ed0da8d62bf0691baea9ce44e43fff7b8488
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fccd625c4165c9d4a7f080a65a0472641bb2f8b11a70f1674d014cc7c7ba718
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d41c5e0dcdb54227b9ba743f10aa7627f76f4bc3177dd08d51a11e9d2ee11e9
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.018406037180195105,
5
  "eval_steps": 20,
6
- "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -232,6 +232,36 @@
232
  "eval_samples_per_second": 4.254,
233
  "eval_steps_per_second": 0.553,
234
  "step": 300
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
235
  }
236
  ],
237
  "logging_steps": 20,
@@ -239,7 +269,7 @@
239
  "num_input_tokens_seen": 0,
240
  "num_train_epochs": 1,
241
  "save_steps": 50,
242
- "total_flos": 2.13106037133312e+16,
243
  "train_batch_size": 2,
244
  "trial_name": null,
245
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.021473710043560953,
5
  "eval_steps": 20,
6
+ "global_step": 350,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
232
  "eval_samples_per_second": 4.254,
233
  "eval_steps_per_second": 0.553,
234
  "step": 300
235
+ },
236
+ {
237
+ "epoch": 0.019633106325541445,
238
+ "grad_norm": 0.09555982798337936,
239
+ "learning_rate": 0.00019613354609058549,
240
+ "loss": 1.6213,
241
+ "step": 320
242
+ },
243
+ {
244
+ "epoch": 0.019633106325541445,
245
+ "eval_loss": 1.5556869506835938,
246
+ "eval_runtime": 23.5239,
247
+ "eval_samples_per_second": 4.251,
248
+ "eval_steps_per_second": 0.553,
249
+ "step": 320
250
+ },
251
+ {
252
+ "epoch": 0.020860175470887785,
253
+ "grad_norm": 0.13320715725421906,
254
+ "learning_rate": 0.00019588805695347982,
255
+ "loss": 1.6956,
256
+ "step": 340
257
+ },
258
+ {
259
+ "epoch": 0.020860175470887785,
260
+ "eval_loss": 1.5424914360046387,
261
+ "eval_runtime": 23.6064,
262
+ "eval_samples_per_second": 4.236,
263
+ "eval_steps_per_second": 0.551,
264
+ "step": 340
265
  }
266
  ],
267
  "logging_steps": 20,
 
269
  "num_input_tokens_seen": 0,
270
  "num_train_epochs": 1,
271
  "save_steps": 50,
272
+ "total_flos": 2.51593625816064e+16,
273
  "train_batch_size": 2,
274
  "trial_name": null,
275
  "trial_params": null