robertou2 commited on
Commit
e291733
·
verified ·
1 Parent(s): ffbe7b6

Upload folder using huggingface_hub

Browse files
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90f012c442accfa6f38a4282c517cdcc4287c8d382b12c8202d20a6d21141576
3
  size 738232680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a304d75910dd0421e95eb1bd8b9470584dd6a41faaf2f2456b1247f80dddf925
3
  size 738232680
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d92483e7ef82d62e4cefd6d7ae69e21b950d46ba883869dbfcced86267a001e
3
  size 1476611275
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c04d67455664030c431864810ecb7f420e978f4745cfd77fed357a7d337cbf3
3
  size 1476611275
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d1bc75e2fb0afb4cdce6286143bee936f37febf68cc744d52dbaaafbfa7bcf7
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac8091a108f9f7cd19dbe7d91d51c39045d769dcf6a14a778cf5cfeee0ee761e
3
  size 14645
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7db8e7bffe73040ea5cd93ece4d9c142345880217e710213ff0948759dd1246e
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f05b53c6c8b5fae09e3d46adb0f5cf2ced6d37c2dbb609562af4961873c13040
3
  size 1465
trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 405,
3
- "best_metric": 0.0006800881819799542,
4
- "best_model_checkpoint": "/content/drive/MyDrive/lora_model/outputs/task15_microsoft/Phi-4-mini-instruct/checkpoint-405",
5
- "epoch": 27.0,
6
  "eval_steps": 500,
7
- "global_step": 405,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -338,81 +338,6 @@
338
  "eval_samples_per_second": 8.812,
339
  "eval_steps_per_second": 1.175,
340
  "step": 330
341
- },
342
- {
343
- "epoch": 23.0,
344
- "grad_norm": 0.021359996870160103,
345
- "learning_rate": 1.1713892356143239e-05,
346
- "loss": 0.0009,
347
- "step": 345
348
- },
349
- {
350
- "epoch": 23.0,
351
- "eval_loss": 0.0008289433317258954,
352
- "eval_runtime": 3.4102,
353
- "eval_samples_per_second": 8.797,
354
- "eval_steps_per_second": 1.173,
355
- "step": 345
356
- },
357
- {
358
- "epoch": 24.0,
359
- "grad_norm": 0.01979278400540352,
360
- "learning_rate": 9.722147576623743e-06,
361
- "loss": 0.0008,
362
- "step": 360
363
- },
364
- {
365
- "epoch": 24.0,
366
- "eval_loss": 0.0007673827931284904,
367
- "eval_runtime": 3.4084,
368
- "eval_samples_per_second": 8.802,
369
- "eval_steps_per_second": 1.174,
370
- "step": 360
371
- },
372
- {
373
- "epoch": 25.0,
374
- "grad_norm": 0.015766171738505363,
375
- "learning_rate": 7.87452118782363e-06,
376
- "loss": 0.0007,
377
- "step": 375
378
- },
379
- {
380
- "epoch": 25.0,
381
- "eval_loss": 0.0007245048182085156,
382
- "eval_runtime": 3.413,
383
- "eval_samples_per_second": 8.79,
384
- "eval_steps_per_second": 1.172,
385
- "step": 375
386
- },
387
- {
388
- "epoch": 26.0,
389
- "grad_norm": 0.01694435626268387,
390
- "learning_rate": 6.188442140777742e-06,
391
- "loss": 0.0007,
392
- "step": 390
393
- },
394
- {
395
- "epoch": 26.0,
396
- "eval_loss": 0.0006994114373810589,
397
- "eval_runtime": 3.4088,
398
- "eval_samples_per_second": 8.801,
399
- "eval_steps_per_second": 1.173,
400
- "step": 390
401
- },
402
- {
403
- "epoch": 27.0,
404
- "grad_norm": 0.014102422632277012,
405
- "learning_rate": 4.679815484975505e-06,
406
- "loss": 0.0007,
407
- "step": 405
408
- },
409
- {
410
- "epoch": 27.0,
411
- "eval_loss": 0.0006800881819799542,
412
- "eval_runtime": 3.4112,
413
- "eval_samples_per_second": 8.795,
414
- "eval_steps_per_second": 1.173,
415
- "step": 405
416
  }
417
  ],
418
  "logging_steps": 10,
@@ -432,7 +357,7 @@
432
  "attributes": {}
433
  }
434
  },
435
- "total_flos": 1.846419663323136e+16,
436
  "train_batch_size": 2,
437
  "trial_name": null,
438
  "trial_params": null
 
1
  {
2
+ "best_global_step": 330,
3
+ "best_metric": 0.0009259060025215149,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/lora_model/outputs/task15_microsoft/Phi-4-mini-instruct/checkpoint-330",
5
+ "epoch": 22.0,
6
  "eval_steps": 500,
7
+ "global_step": 330,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
338
  "eval_samples_per_second": 8.812,
339
  "eval_steps_per_second": 1.175,
340
  "step": 330
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
341
  }
342
  ],
343
  "logging_steps": 10,
 
357
  "attributes": {}
358
  }
359
  },
360
+ "total_flos": 1.5057254808465408e+16,
361
  "train_batch_size": 2,
362
  "trial_name": null,
363
  "trial_params": null