PEFT
Safetensors
Generated from Trainer
paulrichmond commited on
Commit
0ac76de
·
verified ·
1 Parent(s): 572c71b

Training in progress, step 18444, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d74881490c58204c809ecc63d49988dff67283b41b5ba8928513c1e758c01a6
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:835444f9e28ac5484b00aeb974fc2781eea0baeb83bb2ac9f97a6f07fc4d81b1
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9422bd31ecf2a5d6b1c8c13fe817398668cecd58b962bb20e0a6029f9ac8d3d
3
  size 168150738
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8420915da398af22a8b097d398ff247cdfd9d112d48e473387b322aad7d73248
3
  size 168150738
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e03e06adeb2983022141cbb3499da248eadd0a0a927d6e04f8eed941c0d3d686
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf0c4d80b282260ef34ccddc4ebdfb0ff868efdec857e174744b9741a7c9964f
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82ab3f4c71b15efcf419ae699bcc09c12d9d69fda628dcac52fdb8059760dd34
3
  size 1000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112b0e3f3850ccaa762f9eb894c6c56a98cff86a4eab4f8d883ff8a96dbc8748
3
  size 1000
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.9973975276512688,
5
  "eval_steps": 2048,
6
- "global_step": 18432,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -129211,6 +129211,90 @@
129211
  "eval_test_samples_per_second": 13.302,
129212
  "eval_test_steps_per_second": 0.831,
129213
  "step": 18432
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129214
  }
129215
  ],
129216
  "logging_steps": 1,
@@ -129225,12 +129309,12 @@
129225
  "should_evaluate": false,
129226
  "should_log": false,
129227
  "should_save": true,
129228
- "should_training_stop": false
129229
  },
129230
  "attributes": {}
129231
  }
129232
  },
129233
- "total_flos": 4.425362788270473e+18,
129234
  "train_batch_size": 16,
129235
  "trial_name": null,
129236
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 4.0,
5
  "eval_steps": 2048,
6
+ "global_step": 18444,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
129211
  "eval_test_samples_per_second": 13.302,
129212
  "eval_test_steps_per_second": 0.831,
129213
  "step": 18432
129214
+ },
129215
+ {
129216
+ "epoch": 3.9976144003469964,
129217
+ "grad_norm": 1.5703761577606201,
129218
+ "learning_rate": 3.0003218233616098e-06,
129219
+ "loss": 1.8993,
129220
+ "step": 18433
129221
+ },
129222
+ {
129223
+ "epoch": 3.997831273042724,
129224
+ "grad_norm": 1.6585696935653687,
129225
+ "learning_rate": 3.000265969737023e-06,
129226
+ "loss": 1.8753,
129227
+ "step": 18434
129228
+ },
129229
+ {
129230
+ "epoch": 3.9980481457384514,
129231
+ "grad_norm": 1.722380518913269,
129232
+ "learning_rate": 3.000215435499207e-06,
129233
+ "loss": 1.9192,
129234
+ "step": 18435
129235
+ },
129236
+ {
129237
+ "epoch": 3.998265018434179,
129238
+ "grad_norm": 1.6945613622665405,
129239
+ "learning_rate": 3.000170220649976e-06,
129240
+ "loss": 1.7501,
129241
+ "step": 18436
129242
+ },
129243
+ {
129244
+ "epoch": 3.9984818911299067,
129245
+ "grad_norm": 1.587354302406311,
129246
+ "learning_rate": 3.000130325190978e-06,
129247
+ "loss": 1.6834,
129248
+ "step": 18437
129249
+ },
129250
+ {
129251
+ "epoch": 3.9986987638256344,
129252
+ "grad_norm": 1.530950903892517,
129253
+ "learning_rate": 3.0000957491236143e-06,
129254
+ "loss": 1.8401,
129255
+ "step": 18438
129256
+ },
129257
+ {
129258
+ "epoch": 3.998915636521362,
129259
+ "grad_norm": 1.571050763130188,
129260
+ "learning_rate": 3.000066492449139e-06,
129261
+ "loss": 1.7994,
129262
+ "step": 18439
129263
+ },
129264
+ {
129265
+ "epoch": 3.9991325092170893,
129266
+ "grad_norm": 1.3360202312469482,
129267
+ "learning_rate": 3.00004255516859e-06,
129268
+ "loss": 1.7106,
129269
+ "step": 18440
129270
+ },
129271
+ {
129272
+ "epoch": 3.9993493819128174,
129273
+ "grad_norm": 1.9069377183914185,
129274
+ "learning_rate": 3.0000239372828404e-06,
129275
+ "loss": 2.0346,
129276
+ "step": 18441
129277
+ },
129278
+ {
129279
+ "epoch": 3.9995662546085446,
129280
+ "grad_norm": 1.6238856315612793,
129281
+ "learning_rate": 3.0000106387925348e-06,
129282
+ "loss": 1.8038,
129283
+ "step": 18442
129284
+ },
129285
+ {
129286
+ "epoch": 3.9997831273042723,
129287
+ "grad_norm": 1.4525717496871948,
129288
+ "learning_rate": 3.00000265969815e-06,
129289
+ "loss": 1.9284,
129290
+ "step": 18443
129291
+ },
129292
+ {
129293
+ "epoch": 4.0,
129294
+ "grad_norm": 2.0924839973449707,
129295
+ "learning_rate": 2.9999999999999997e-06,
129296
+ "loss": 1.8805,
129297
+ "step": 18444
129298
  }
129299
  ],
129300
  "logging_steps": 1,
 
129309
  "should_evaluate": false,
129310
  "should_log": false,
129311
  "should_save": true,
129312
+ "should_training_stop": true
129313
  },
129314
  "attributes": {}
129315
  }
129316
  },
129317
+ "total_flos": 4.42814002458683e+18,
129318
  "train_batch_size": 16,
129319
  "trial_name": null,
129320
  "trial_params": null