TweedleDeepLearnings commited on
Commit
496b107
·
verified ·
1 Parent(s): 0a49a05

Training in progress, step 4800, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09a967aa6e8a06933ec10fe64749fe2a5ebd8f476503db63c96a4eeb363a0d0f
3
  size 2231685024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1bc3b3e143f9c5cf9f286dd374831316522298c162cc7d9aee812c4494e2502
3
  size 2231685024
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5cda6d8a0ed828c466cfe33411954edb901baed050118b024f1d8ad1c69b11e1
3
  size 4383116999
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3f31c32a5cc6868da9e0787a7682ce0c0935d63cee861a343159b380d70512b
3
  size 4383116999
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0eebef1e3232e8f2e3f4e49a633d4bbbbd205f77ce4cf6d894d7356c90d284d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4413b3c1c89ddb3d57ddabb40868d44f33618613ca6b14588ab3262fd959fe1a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5aa7843fee19f986423b7658b656356c03da99a77a7d470397314a84a80a47a
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:536269e746606d89d7092eefc1d242400387136c0a490c65f4003318a00e54ce
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7991831302642822,
3
  "best_model_checkpoint": "./output/checkpoint-4650",
4
- "epoch": 0.3342197944368576,
5
  "eval_steps": 150,
6
- "global_step": 4650,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -3510,6 +3510,119 @@
3510
  "eval_samples_per_second": 6.721,
3511
  "eval_steps_per_second": 6.721,
3512
  "step": 4650
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3513
  }
3514
  ],
3515
  "logging_steps": 10,
@@ -3529,7 +3642,7 @@
3529
  "attributes": {}
3530
  }
3531
  },
3532
- "total_flos": 3.9545543201425e+17,
3533
  "train_batch_size": 8,
3534
  "trial_name": null,
3535
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7991831302642822,
3
  "best_model_checkpoint": "./output/checkpoint-4650",
4
+ "epoch": 0.34500107812836917,
5
  "eval_steps": 150,
6
+ "global_step": 4800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
3510
  "eval_samples_per_second": 6.721,
3511
  "eval_steps_per_second": 6.721,
3512
  "step": 4650
3513
+ },
3514
+ {
3515
+ "epoch": 0.33493854668295836,
3516
+ "grad_norm": 4.487733364105225,
3517
+ "learning_rate": 9.466184635518361e-07,
3518
+ "loss": 0.6828,
3519
+ "step": 4660
3520
+ },
3521
+ {
3522
+ "epoch": 0.33565729892905916,
3523
+ "grad_norm": 3.950061082839966,
3524
+ "learning_rate": 8.919587459299461e-07,
3525
+ "loss": 0.6465,
3526
+ "step": 4670
3527
+ },
3528
+ {
3529
+ "epoch": 0.3363760511751599,
3530
+ "grad_norm": 4.053449630737305,
3531
+ "learning_rate": 8.389066067082852e-07,
3532
+ "loss": 0.6912,
3533
+ "step": 4680
3534
+ },
3535
+ {
3536
+ "epoch": 0.3370948034212607,
3537
+ "grad_norm": 3.2221829891204834,
3538
+ "learning_rate": 7.874642266525102e-07,
3539
+ "loss": 0.7184,
3540
+ "step": 4690
3541
+ },
3542
+ {
3543
+ "epoch": 0.3378135556673615,
3544
+ "grad_norm": 3.5493359565734863,
3545
+ "learning_rate": 7.376337203573824e-07,
3546
+ "loss": 0.6761,
3547
+ "step": 4700
3548
+ },
3549
+ {
3550
+ "epoch": 0.3385323079134622,
3551
+ "grad_norm": 3.3036763668060303,
3552
+ "learning_rate": 6.894171361598467e-07,
3553
+ "loss": 0.6892,
3554
+ "step": 4710
3555
+ },
3556
+ {
3557
+ "epoch": 0.339251060159563,
3558
+ "grad_norm": 4.632784366607666,
3559
+ "learning_rate": 6.428164560548134e-07,
3560
+ "loss": 0.7149,
3561
+ "step": 4720
3562
+ },
3563
+ {
3564
+ "epoch": 0.33996981240566376,
3565
+ "grad_norm": 4.091222763061523,
3566
+ "learning_rate": 5.978335956137304e-07,
3567
+ "loss": 0.7652,
3568
+ "step": 4730
3569
+ },
3570
+ {
3571
+ "epoch": 0.34068856465176456,
3572
+ "grad_norm": 4.298058032989502,
3573
+ "learning_rate": 5.544704039058025e-07,
3574
+ "loss": 0.6956,
3575
+ "step": 4740
3576
+ },
3577
+ {
3578
+ "epoch": 0.3414073168978653,
3579
+ "grad_norm": 3.7245140075683594,
3580
+ "learning_rate": 5.12728663421993e-07,
3581
+ "loss": 0.6434,
3582
+ "step": 4750
3583
+ },
3584
+ {
3585
+ "epoch": 0.3421260691439661,
3586
+ "grad_norm": 3.273090124130249,
3587
+ "learning_rate": 4.7261009000177274e-07,
3588
+ "loss": 0.72,
3589
+ "step": 4760
3590
+ },
3591
+ {
3592
+ "epoch": 0.34284482139006683,
3593
+ "grad_norm": 4.249372959136963,
3594
+ "learning_rate": 4.3411633276255307e-07,
3595
+ "loss": 0.7028,
3596
+ "step": 4770
3597
+ },
3598
+ {
3599
+ "epoch": 0.34356357363616763,
3600
+ "grad_norm": 3.6272311210632324,
3601
+ "learning_rate": 3.972489740319274e-07,
3602
+ "loss": 0.731,
3603
+ "step": 4780
3604
+ },
3605
+ {
3606
+ "epoch": 0.34428232588226837,
3607
+ "grad_norm": 4.036685943603516,
3608
+ "learning_rate": 3.62009529282612e-07,
3609
+ "loss": 0.7919,
3610
+ "step": 4790
3611
+ },
3612
+ {
3613
+ "epoch": 0.34500107812836917,
3614
+ "grad_norm": 4.364447116851807,
3615
+ "learning_rate": 3.283994470701579e-07,
3616
+ "loss": 0.7161,
3617
+ "step": 4800
3618
+ },
3619
+ {
3620
+ "epoch": 0.34500107812836917,
3621
+ "eval_loss": 0.7992691397666931,
3622
+ "eval_runtime": 77.2387,
3623
+ "eval_samples_per_second": 6.473,
3624
+ "eval_steps_per_second": 6.473,
3625
+ "step": 4800
3626
  }
3627
  ],
3628
  "logging_steps": 10,
 
3642
  "attributes": {}
3643
  }
3644
  },
3645
+ "total_flos": 4.088392743032586e+17,
3646
  "train_batch_size": 8,
3647
  "trial_name": null,
3648
  "trial_params": null