cools commited on
Commit
91388b4
·
1 Parent(s): 131bc29

Training in progress, step 50, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -217,23 +217,4 @@ The following `bitsandbytes` quantization config was used during training:
217
  ### Framework versions
218
 
219
 
220
- - PEFT 0.6.2
221
- ## Training procedure
222
-
223
-
224
- The following `bitsandbytes` quantization config was used during training:
225
- - quant_method: bitsandbytes
226
- - load_in_8bit: False
227
- - load_in_4bit: True
228
- - llm_int8_threshold: 6.0
229
- - llm_int8_skip_modules: None
230
- - llm_int8_enable_fp32_cpu_offload: False
231
- - llm_int8_has_fp16_weight: True
232
- - bnb_4bit_quant_type: nf4
233
- - bnb_4bit_use_double_quant: True
234
- - bnb_4bit_compute_dtype: float16
235
-
236
- ### Framework versions
237
-
238
-
239
  - PEFT 0.6.2
 
217
  ### Framework versions
218
 
219
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
220
  - PEFT 0.6.2
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97cba999fb720c4d176ef8a8d3c8d6809705b6f07a15dc41a16e639cba3a4151
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:269e02dc931865a4d7c1981eca33d2486373d043dd99da5f05186ac7501a7648
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:346e9b48608e1d59b317e52559a01620207035d8ff8260d0025b99fef9d3a988
3
  size 42545748
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b219215a0cdf22b74e690dedd537471ed5ec27d5a848a0c00dd9462c08dea9c
3
  size 42545748
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:377540a51df182761cc61ba1eb907a8a946aa1a4db87591c9078f8e011fe7b38
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58973b63eb41c8b34c51f7eb73abffa4ae0d2af806c0e342de338e3f48596cb0
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cba506068924f81fe6223c30197a8cb6d84cb2904e97497dc664ebe8dc4a033f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02173fc97b82f2e000f1d5084934ed87f720106f5dddb2874a3633a6c2964dab
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 25.0,
5
  "eval_steps": 1000,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -157,13 +157,163 @@
157
  "learning_rate": 0.00016631578947368423,
158
  "loss": 0.0127,
159
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
160
  }
161
  ],
162
  "logging_steps": 1,
163
  "max_steps": 100,
164
  "num_train_epochs": 100,
165
  "save_steps": 25,
166
- "total_flos": 267436185600000.0,
167
  "trial_name": null,
168
  "trial_params": null
169
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 50.0,
5
  "eval_steps": 1000,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
157
  "learning_rate": 0.00016631578947368423,
158
  "loss": 0.0127,
159
  "step": 25
160
+ },
161
+ {
162
+ "epoch": 26.0,
163
+ "learning_rate": 0.00016421052631578948,
164
+ "loss": 0.0049,
165
+ "step": 26
166
+ },
167
+ {
168
+ "epoch": 27.0,
169
+ "learning_rate": 0.00016210526315789473,
170
+ "loss": 0.0042,
171
+ "step": 27
172
+ },
173
+ {
174
+ "epoch": 28.0,
175
+ "learning_rate": 0.00016,
176
+ "loss": 0.0052,
177
+ "step": 28
178
+ },
179
+ {
180
+ "epoch": 29.0,
181
+ "learning_rate": 0.00015789473684210527,
182
+ "loss": 0.0041,
183
+ "step": 29
184
+ },
185
+ {
186
+ "epoch": 30.0,
187
+ "learning_rate": 0.00015578947368421052,
188
+ "loss": 0.0044,
189
+ "step": 30
190
+ },
191
+ {
192
+ "epoch": 31.0,
193
+ "learning_rate": 0.0001536842105263158,
194
+ "loss": 0.0044,
195
+ "step": 31
196
+ },
197
+ {
198
+ "epoch": 32.0,
199
+ "learning_rate": 0.00015157894736842108,
200
+ "loss": 0.0043,
201
+ "step": 32
202
+ },
203
+ {
204
+ "epoch": 33.0,
205
+ "learning_rate": 0.00014947368421052633,
206
+ "loss": 0.0043,
207
+ "step": 33
208
+ },
209
+ {
210
+ "epoch": 34.0,
211
+ "learning_rate": 0.00014736842105263158,
212
+ "loss": 0.0043,
213
+ "step": 34
214
+ },
215
+ {
216
+ "epoch": 35.0,
217
+ "learning_rate": 0.00014526315789473686,
218
+ "loss": 0.0044,
219
+ "step": 35
220
+ },
221
+ {
222
+ "epoch": 36.0,
223
+ "learning_rate": 0.0001431578947368421,
224
+ "loss": 0.0035,
225
+ "step": 36
226
+ },
227
+ {
228
+ "epoch": 37.0,
229
+ "learning_rate": 0.00014105263157894736,
230
+ "loss": 0.004,
231
+ "step": 37
232
+ },
233
+ {
234
+ "epoch": 38.0,
235
+ "learning_rate": 0.00013894736842105264,
236
+ "loss": 0.0044,
237
+ "step": 38
238
+ },
239
+ {
240
+ "epoch": 39.0,
241
+ "learning_rate": 0.0001368421052631579,
242
+ "loss": 0.0041,
243
+ "step": 39
244
+ },
245
+ {
246
+ "epoch": 40.0,
247
+ "learning_rate": 0.00013473684210526317,
248
+ "loss": 0.0043,
249
+ "step": 40
250
+ },
251
+ {
252
+ "epoch": 41.0,
253
+ "learning_rate": 0.00013263157894736842,
254
+ "loss": 0.0041,
255
+ "step": 41
256
+ },
257
+ {
258
+ "epoch": 42.0,
259
+ "learning_rate": 0.0001305263157894737,
260
+ "loss": 0.0042,
261
+ "step": 42
262
+ },
263
+ {
264
+ "epoch": 43.0,
265
+ "learning_rate": 0.00012842105263157895,
266
+ "loss": 0.0036,
267
+ "step": 43
268
+ },
269
+ {
270
+ "epoch": 44.0,
271
+ "learning_rate": 0.0001263157894736842,
272
+ "loss": 0.0038,
273
+ "step": 44
274
+ },
275
+ {
276
+ "epoch": 45.0,
277
+ "learning_rate": 0.00012421052631578949,
278
+ "loss": 0.0039,
279
+ "step": 45
280
+ },
281
+ {
282
+ "epoch": 46.0,
283
+ "learning_rate": 0.00012210526315789474,
284
+ "loss": 0.0045,
285
+ "step": 46
286
+ },
287
+ {
288
+ "epoch": 47.0,
289
+ "learning_rate": 0.00012,
290
+ "loss": 0.0053,
291
+ "step": 47
292
+ },
293
+ {
294
+ "epoch": 48.0,
295
+ "learning_rate": 0.00011789473684210525,
296
+ "loss": 0.0042,
297
+ "step": 48
298
+ },
299
+ {
300
+ "epoch": 49.0,
301
+ "learning_rate": 0.00011578947368421053,
302
+ "loss": 0.0029,
303
+ "step": 49
304
+ },
305
+ {
306
+ "epoch": 50.0,
307
+ "learning_rate": 0.0001136842105263158,
308
+ "loss": 0.004,
309
+ "step": 50
310
  }
311
  ],
312
  "logging_steps": 1,
313
  "max_steps": 100,
314
  "num_train_epochs": 100,
315
  "save_steps": 25,
316
+ "total_flos": 534872371200000.0,
317
  "trial_name": null,
318
  "trial_params": null
319
  }