PEFT
Safetensors
Generated from Trainer
paulrichmond commited on
Commit
148ca75
·
verified ·
1 Parent(s): fda3746

Training in progress, step 18444, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:195b15be2e8b1b7904e5d8749601220537331772934d950fce781fd8d7e5a001
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6286f714f60e819357e2ff83ccb680fc1edeec36b5d22da800765db964908b43
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:62d10b215bfbe1dc56494fcfc3b74aa2b5605bd7a48d6e5385ac19955bcfc045
3
  size 168150738
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce6c7abaa15b1e3a6af628219ef7a13b0763dd109560c1371c4a105e59597ca2
3
  size 168150738
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:328bb4e59c1aa99386194ef35d5ac9c293719a8db28e42ca89f2abe0d2bcd607
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61ef5f4a6e53cc6a6774b0b359452317c4b2b55b42e8f0b94b1c3ab8e4f3dd39
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82ab3f4c71b15efcf419ae699bcc09c12d9d69fda628dcac52fdb8059760dd34
3
  size 1000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112b0e3f3850ccaa762f9eb894c6c56a98cff86a4eab4f8d883ff8a96dbc8748
3
  size 1000
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.9973975276512688,
5
  "eval_steps": 2048,
6
- "global_step": 18432,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -129211,6 +129211,90 @@
129211
  "eval_test_samples_per_second": 13.089,
129212
  "eval_test_steps_per_second": 0.818,
129213
  "step": 18432
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129214
  }
129215
  ],
129216
  "logging_steps": 1,
@@ -129225,12 +129309,12 @@
129225
  "should_evaluate": false,
129226
  "should_log": false,
129227
  "should_save": true,
129228
- "should_training_stop": false
129229
  },
129230
  "attributes": {}
129231
  }
129232
  },
129233
- "total_flos": 4.458545126859866e+18,
129234
  "train_batch_size": 16,
129235
  "trial_name": null,
129236
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 4.0,
5
  "eval_steps": 2048,
6
+ "global_step": 18444,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
129211
  "eval_test_samples_per_second": 13.089,
129212
  "eval_test_steps_per_second": 0.818,
129213
  "step": 18432
129214
+ },
129215
+ {
129216
+ "epoch": 3.9976144003469964,
129217
+ "grad_norm": 1.635145664215088,
129218
+ "learning_rate": 3.0003218233616098e-06,
129219
+ "loss": 1.8547,
129220
+ "step": 18433
129221
+ },
129222
+ {
129223
+ "epoch": 3.997831273042724,
129224
+ "grad_norm": 1.6694488525390625,
129225
+ "learning_rate": 3.000265969737023e-06,
129226
+ "loss": 1.9154,
129227
+ "step": 18434
129228
+ },
129229
+ {
129230
+ "epoch": 3.9980481457384514,
129231
+ "grad_norm": 1.7029920816421509,
129232
+ "learning_rate": 3.000215435499207e-06,
129233
+ "loss": 1.8898,
129234
+ "step": 18435
129235
+ },
129236
+ {
129237
+ "epoch": 3.998265018434179,
129238
+ "grad_norm": 1.4422721862792969,
129239
+ "learning_rate": 3.000170220649976e-06,
129240
+ "loss": 1.6105,
129241
+ "step": 18436
129242
+ },
129243
+ {
129244
+ "epoch": 3.9984818911299067,
129245
+ "grad_norm": 1.5800944566726685,
129246
+ "learning_rate": 3.000130325190978e-06,
129247
+ "loss": 1.7082,
129248
+ "step": 18437
129249
+ },
129250
+ {
129251
+ "epoch": 3.9986987638256344,
129252
+ "grad_norm": 1.4681060314178467,
129253
+ "learning_rate": 3.0000957491236143e-06,
129254
+ "loss": 1.8266,
129255
+ "step": 18438
129256
+ },
129257
+ {
129258
+ "epoch": 3.998915636521362,
129259
+ "grad_norm": 1.8325995206832886,
129260
+ "learning_rate": 3.000066492449139e-06,
129261
+ "loss": 1.8093,
129262
+ "step": 18439
129263
+ },
129264
+ {
129265
+ "epoch": 3.9991325092170893,
129266
+ "grad_norm": 1.5083396434783936,
129267
+ "learning_rate": 3.00004255516859e-06,
129268
+ "loss": 1.7805,
129269
+ "step": 18440
129270
+ },
129271
+ {
129272
+ "epoch": 3.9993493819128174,
129273
+ "grad_norm": 1.8943501710891724,
129274
+ "learning_rate": 3.0000239372828404e-06,
129275
+ "loss": 1.9769,
129276
+ "step": 18441
129277
+ },
129278
+ {
129279
+ "epoch": 3.9995662546085446,
129280
+ "grad_norm": 1.4337064027786255,
129281
+ "learning_rate": 3.0000106387925348e-06,
129282
+ "loss": 1.8074,
129283
+ "step": 18442
129284
+ },
129285
+ {
129286
+ "epoch": 3.9997831273042723,
129287
+ "grad_norm": 1.4833385944366455,
129288
+ "learning_rate": 3.00000265969815e-06,
129289
+ "loss": 1.8618,
129290
+ "step": 18443
129291
+ },
129292
+ {
129293
+ "epoch": 4.0,
129294
+ "grad_norm": 1.9734277725219727,
129295
+ "learning_rate": 2.9999999999999997e-06,
129296
+ "loss": 1.8028,
129297
+ "step": 18444
129298
  }
129299
  ],
129300
  "logging_steps": 1,
 
129309
  "should_evaluate": false,
129310
  "should_log": false,
129311
  "should_save": true,
129312
+ "should_training_stop": true
129313
  },
129314
  "attributes": {}
129315
  }
129316
  },
129317
+ "total_flos": 4.461328143064498e+18,
129318
  "train_batch_size": 16,
129319
  "trial_name": null,
129320
  "trial_params": null