PEFT
Safetensors
Generated from Trainer
cpapa79 commited on
Commit
f9c9bc6
·
verified ·
1 Parent(s): 184688a

Training in progress, step 18444, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8267f04b6ad643fe76621b515f6ef4a89574aeb2874b7c9ea7900800d3cbcafa
3
  size 18899856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e23fbabee21f3e140e3b3444ce9c1357a35aa394fba3f86e89b9d2359d8f434
3
  size 18899856
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68754aed55e70b60e47104a0f3d7786bfd3a484332f76c108dfa140bff640ce9
3
  size 37911546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59480740c38bdb92e1247b615148ace85eb247a65b4c1157843ee5dae6ae1151
3
  size 37911546
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32fe7f27596afd7931277b7e50b18e54d28b82a6b242c7ca22b1114b4d30423e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e591aeb9b33177e25fcd649ccfc82933f3329eda6ee273b1a0bfe406c384c55
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82ab3f4c71b15efcf419ae699bcc09c12d9d69fda628dcac52fdb8059760dd34
3
  size 1000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112b0e3f3850ccaa762f9eb894c6c56a98cff86a4eab4f8d883ff8a96dbc8748
3
  size 1000
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.9973975276512688,
5
  "eval_steps": 2048,
6
- "global_step": 18432,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -129211,6 +129211,90 @@
129211
  "eval_test_samples_per_second": 14.935,
129212
  "eval_test_steps_per_second": 0.933,
129213
  "step": 18432
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129214
  }
129215
  ],
129216
  "logging_steps": 1,
@@ -129225,12 +129309,12 @@
129225
  "should_evaluate": false,
129226
  "should_log": false,
129227
  "should_save": true,
129228
- "should_training_stop": false
129229
  },
129230
  "attributes": {}
129231
  }
129232
  },
129233
- "total_flos": 4.4158057732736287e+18,
129234
  "train_batch_size": 16,
129235
  "trial_name": null,
129236
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 4.0,
5
  "eval_steps": 2048,
6
+ "global_step": 18444,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
129211
  "eval_test_samples_per_second": 14.935,
129212
  "eval_test_steps_per_second": 0.933,
129213
  "step": 18432
129214
+ },
129215
+ {
129216
+ "epoch": 3.9976144003469964,
129217
+ "grad_norm": 0.7179842591285706,
129218
+ "learning_rate": 3.0003218233616098e-06,
129219
+ "loss": 2.1294,
129220
+ "step": 18433
129221
+ },
129222
+ {
129223
+ "epoch": 3.997831273042724,
129224
+ "grad_norm": 0.7400171756744385,
129225
+ "learning_rate": 3.000265969737023e-06,
129226
+ "loss": 2.1279,
129227
+ "step": 18434
129228
+ },
129229
+ {
129230
+ "epoch": 3.9980481457384514,
129231
+ "grad_norm": 0.816265881061554,
129232
+ "learning_rate": 3.000215435499207e-06,
129233
+ "loss": 2.2085,
129234
+ "step": 18435
129235
+ },
129236
+ {
129237
+ "epoch": 3.998265018434179,
129238
+ "grad_norm": 0.6999260783195496,
129239
+ "learning_rate": 3.000170220649976e-06,
129240
+ "loss": 1.9719,
129241
+ "step": 18436
129242
+ },
129243
+ {
129244
+ "epoch": 3.9984818911299067,
129245
+ "grad_norm": 0.7754216194152832,
129246
+ "learning_rate": 3.000130325190978e-06,
129247
+ "loss": 1.9501,
129248
+ "step": 18437
129249
+ },
129250
+ {
129251
+ "epoch": 3.9986987638256344,
129252
+ "grad_norm": 0.7585975527763367,
129253
+ "learning_rate": 3.0000957491236143e-06,
129254
+ "loss": 2.0977,
129255
+ "step": 18438
129256
+ },
129257
+ {
129258
+ "epoch": 3.998915636521362,
129259
+ "grad_norm": 0.7343212962150574,
129260
+ "learning_rate": 3.000066492449139e-06,
129261
+ "loss": 2.0824,
129262
+ "step": 18439
129263
+ },
129264
+ {
129265
+ "epoch": 3.9991325092170893,
129266
+ "grad_norm": 0.7170373797416687,
129267
+ "learning_rate": 3.00004255516859e-06,
129268
+ "loss": 2.0028,
129269
+ "step": 18440
129270
+ },
129271
+ {
129272
+ "epoch": 3.9993493819128174,
129273
+ "grad_norm": 0.9392771124839783,
129274
+ "learning_rate": 3.0000239372828404e-06,
129275
+ "loss": 2.3121,
129276
+ "step": 18441
129277
+ },
129278
+ {
129279
+ "epoch": 3.9995662546085446,
129280
+ "grad_norm": 0.6705074310302734,
129281
+ "learning_rate": 3.0000106387925348e-06,
129282
+ "loss": 2.0513,
129283
+ "step": 18442
129284
+ },
129285
+ {
129286
+ "epoch": 3.9997831273042723,
129287
+ "grad_norm": 0.719359815120697,
129288
+ "learning_rate": 3.00000265969815e-06,
129289
+ "loss": 2.2008,
129290
+ "step": 18443
129291
+ },
129292
+ {
129293
+ "epoch": 4.0,
129294
+ "grad_norm": 1.0389312505722046,
129295
+ "learning_rate": 2.9999999999999997e-06,
129296
+ "loss": 2.1296,
129297
+ "step": 18444
129298
  }
129299
  ],
129300
  "logging_steps": 1,
 
129309
  "should_evaluate": false,
129310
  "should_log": false,
129311
  "should_save": true,
129312
+ "should_training_stop": true
129313
  },
129314
  "attributes": {}
129315
  }
129316
  },
129317
+ "total_flos": 4.418577011869483e+18,
129318
  "train_batch_size": 16,
129319
  "trial_name": null,
129320
  "trial_params": null