irishprancer commited on
Commit
8884869
·
verified ·
1 Parent(s): 99ca4e1

Training in progress, step 4800, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:094a7e88ac190113472d8c01b2df08de094f04602e59899cd894797763a1f29d
3
  size 1054440872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4ced61eae70f21c5c63f1107feef14d1705683fed216ab47273965c0e2494b1
3
  size 1054440872
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70bb1c9ae5586829ee6cf66544b0def35379e96db4143482afabf8f97b0e51a5
3
  size 2041777658
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1eda538d96287c20519cd7975c5f861a14271b86e5a675b2f234149c6482a2d
3
  size 2041777658
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f23dbcd70ff2919050fcc9cba72fcb6491579fd36451328703f82dae74022f11
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94d3cd3d9f0d1fea2d03498cf97fd97401d6a6a7de7fa934b8538bc7c7ec16fb
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:417483f84722a4bfad9b8eb4c5866f76ce06d223609156fc27071200c2ff1069
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da49fd85c9470977f77e555b7807d1ad2e6d8a428e34963bbd75ab42e9b66434
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.6787537932395935,
3
  "best_model_checkpoint": "./output/checkpoint-4050",
4
- "epoch": 0.014683777741287625,
5
  "eval_steps": 150,
6
- "global_step": 4650,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -3510,6 +3510,119 @@
3510
  "eval_samples_per_second": 11.407,
3511
  "eval_steps_per_second": 11.407,
3512
  "step": 4650
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3513
  }
3514
  ],
3515
  "logging_steps": 10,
@@ -3529,7 +3642,7 @@
3529
  "attributes": {}
3530
  }
3531
  },
3532
- "total_flos": 1.5524478093223526e+17,
3533
  "train_batch_size": 8,
3534
  "trial_name": null,
3535
  "trial_params": null
 
1
  {
2
  "best_metric": 0.6787537932395935,
3
  "best_model_checkpoint": "./output/checkpoint-4050",
4
+ "epoch": 0.015157447991006582,
5
  "eval_steps": 150,
6
+ "global_step": 4800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
3510
  "eval_samples_per_second": 11.407,
3511
  "eval_steps_per_second": 11.407,
3512
  "step": 4650
3513
+ },
3514
+ {
3515
+ "epoch": 0.014715355757935556,
3516
+ "grad_norm": 13.99572467803955,
3517
+ "learning_rate": 3.277183051523893e-07,
3518
+ "loss": 0.8312,
3519
+ "step": 4660
3520
+ },
3521
+ {
3522
+ "epoch": 0.014746933774583486,
3523
+ "grad_norm": 7.099660873413086,
3524
+ "learning_rate": 3.0879516905387914e-07,
3525
+ "loss": 0.9549,
3526
+ "step": 4670
3527
+ },
3528
+ {
3529
+ "epoch": 0.014778511791231417,
3530
+ "grad_norm": 8.244866371154785,
3531
+ "learning_rate": 2.904285748875281e-07,
3532
+ "loss": 0.7216,
3533
+ "step": 4680
3534
+ },
3535
+ {
3536
+ "epoch": 0.014810089807879347,
3537
+ "grad_norm": 18.085098266601562,
3538
+ "learning_rate": 2.7261927763208694e-07,
3539
+ "loss": 0.7163,
3540
+ "step": 4690
3541
+ },
3542
+ {
3543
+ "epoch": 0.014841667824527277,
3544
+ "grad_norm": 11.467937469482422,
3545
+ "learning_rate": 2.553680093580127e-07,
3546
+ "loss": 0.8257,
3547
+ "step": 4700
3548
+ },
3549
+ {
3550
+ "epoch": 0.014873245841175208,
3551
+ "grad_norm": 9.333969116210938,
3552
+ "learning_rate": 2.3867547919737676e-07,
3553
+ "loss": 0.8658,
3554
+ "step": 4710
3555
+ },
3556
+ {
3557
+ "epoch": 0.014904823857823138,
3558
+ "grad_norm": 10.59934139251709,
3559
+ "learning_rate": 2.225423733147088e-07,
3560
+ "loss": 0.8157,
3561
+ "step": 4720
3562
+ },
3563
+ {
3564
+ "epoch": 0.014936401874471069,
3565
+ "grad_norm": 11.30579948425293,
3566
+ "learning_rate": 2.069693548788066e-07,
3567
+ "loss": 0.9591,
3568
+ "step": 4730
3569
+ },
3570
+ {
3571
+ "epoch": 0.014967979891118999,
3572
+ "grad_norm": 9.181888580322266,
3573
+ "learning_rate": 1.9195706403546185e-07,
3574
+ "loss": 0.9756,
3575
+ "step": 4740
3576
+ },
3577
+ {
3578
+ "epoch": 0.01499955790776693,
3579
+ "grad_norm": 10.071789741516113,
3580
+ "learning_rate": 1.775061178811501e-07,
3581
+ "loss": 0.8792,
3582
+ "step": 4750
3583
+ },
3584
+ {
3585
+ "epoch": 0.01503113592441486,
3586
+ "grad_norm": 11.81341552734375,
3587
+ "learning_rate": 1.636171104376709e-07,
3588
+ "loss": 0.8474,
3589
+ "step": 4760
3590
+ },
3591
+ {
3592
+ "epoch": 0.01506271394106279,
3593
+ "grad_norm": 9.52579116821289,
3594
+ "learning_rate": 1.5029061262771796e-07,
3595
+ "loss": 0.815,
3596
+ "step": 4770
3597
+ },
3598
+ {
3599
+ "epoch": 0.01509429195771072,
3600
+ "grad_norm": 13.034448623657227,
3601
+ "learning_rate": 1.375271722514209e-07,
3602
+ "loss": 0.7754,
3603
+ "step": 4780
3604
+ },
3605
+ {
3606
+ "epoch": 0.015125869974358651,
3607
+ "grad_norm": 11.439579963684082,
3608
+ "learning_rate": 1.253273139638221e-07,
3609
+ "loss": 0.7933,
3610
+ "step": 4790
3611
+ },
3612
+ {
3613
+ "epoch": 0.015157447991006582,
3614
+ "grad_norm": 7.407265663146973,
3615
+ "learning_rate": 1.1369153925331249e-07,
3616
+ "loss": 0.9544,
3617
+ "step": 4800
3618
+ },
3619
+ {
3620
+ "epoch": 0.015157447991006582,
3621
+ "eval_loss": 0.680745005607605,
3622
+ "eval_runtime": 43.3846,
3623
+ "eval_samples_per_second": 11.525,
3624
+ "eval_steps_per_second": 11.525,
3625
+ "step": 4800
3626
  }
3627
  ],
3628
  "logging_steps": 10,
 
3642
  "attributes": {}
3643
  }
3644
  },
3645
+ "total_flos": 1.604290409996452e+17,
3646
  "train_batch_size": 8,
3647
  "trial_name": null,
3648
  "trial_params": null