Alawy21 commited on
Commit
e19bcfa
·
verified ·
1 Parent(s): e6f57e0

Training in progress, step 800, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09d9cb540c5a860ee18428d97e9c6a2c9413c826ef861b10bfe2273b4ba87918
3
  size 295488936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f06a94a8fd2110630b29f1a0fd61ece5d491e081468de01dce6bcb49a29196b3
3
  size 295488936
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb303aed8062be2e28e6c28799dc86de07c6cd39ee42a654f3ed5866ddcc7f5a
3
  size 591203178
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:583aa4e7918fb15bddf477cd5bcb1d9cae4199ecc8a16903e5bbd6fb6f6c04a5
3
  size 591203178
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9899ccda7f0d8d9511991180b93aab508ce6e8489de708c88ad1188e7e1d90d6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d245e05e72192c132e0f2edb6fdcae0c578c890f0fe912f17ec7b0bba2d38cc3
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4be6285b51eb8628843ebcf538440997e9983602416b820387811937bd0b2278
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:822082b213c0642410a7baeaf5135c14aa4c940d0ddb23409fe69b75772aee28
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 1.2369912416280269,
6
  "eval_steps": 100,
7
- "global_step": 600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -476,6 +476,162 @@
476
  "eval_samples_per_second": 0.408,
477
  "eval_steps_per_second": 0.408,
478
  "step": 600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
479
  }
480
  ],
481
  "logging_steps": 10,
@@ -495,7 +651,7 @@
495
  "attributes": {}
496
  }
497
  },
498
- "total_flos": 3.221911898996736e+16,
499
  "train_batch_size": 1,
500
  "trial_name": null,
501
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 1.6491499227202473,
6
  "eval_steps": 100,
7
+ "global_step": 800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
476
  "eval_samples_per_second": 0.408,
477
  "eval_steps_per_second": 0.408,
478
  "step": 600
479
+ },
480
+ {
481
+ "epoch": 1.2575991756826377,
482
+ "grad_norm": 0.5280514359474182,
483
+ "learning_rate": 7.217943436218728e-05,
484
+ "loss": 0.1515,
485
+ "step": 610
486
+ },
487
+ {
488
+ "epoch": 1.2782071097372487,
489
+ "grad_norm": 0.5676441788673401,
490
+ "learning_rate": 7.109767620035689e-05,
491
+ "loss": 0.1345,
492
+ "step": 620
493
+ },
494
+ {
495
+ "epoch": 1.2988150437918597,
496
+ "grad_norm": 0.5746245980262756,
497
+ "learning_rate": 7.000376641716133e-05,
498
+ "loss": 0.1387,
499
+ "step": 630
500
+ },
501
+ {
502
+ "epoch": 1.3194229778464708,
503
+ "grad_norm": 0.6514201164245605,
504
+ "learning_rate": 6.889833507144532e-05,
505
+ "loss": 0.1185,
506
+ "step": 640
507
+ },
508
+ {
509
+ "epoch": 1.3400309119010818,
510
+ "grad_norm": 0.7305589318275452,
511
+ "learning_rate": 6.778201885812404e-05,
512
+ "loss": 0.1569,
513
+ "step": 650
514
+ },
515
+ {
516
+ "epoch": 1.3606388459556928,
517
+ "grad_norm": 0.41217121481895447,
518
+ "learning_rate": 6.66554607414661e-05,
519
+ "loss": 0.1469,
520
+ "step": 660
521
+ },
522
+ {
523
+ "epoch": 1.3812467800103039,
524
+ "grad_norm": 0.4568374454975128,
525
+ "learning_rate": 6.551930958476571e-05,
526
+ "loss": 0.1249,
527
+ "step": 670
528
+ },
529
+ {
530
+ "epoch": 1.401854714064915,
531
+ "grad_norm": 0.567737340927124,
532
+ "learning_rate": 6.437421977661709e-05,
533
+ "loss": 0.1348,
534
+ "step": 680
535
+ },
536
+ {
537
+ "epoch": 1.422462648119526,
538
+ "grad_norm": 0.5140897035598755,
539
+ "learning_rate": 6.322085085400644e-05,
540
+ "loss": 0.1274,
541
+ "step": 690
542
+ },
543
+ {
544
+ "epoch": 1.443070582174137,
545
+ "grad_norm": 0.6037400960922241,
546
+ "learning_rate": 6.205986712243875e-05,
547
+ "loss": 0.1437,
548
+ "step": 700
549
+ },
550
+ {
551
+ "epoch": 1.443070582174137,
552
+ "eval_loss": 0.1935451626777649,
553
+ "eval_runtime": 146.9127,
554
+ "eval_samples_per_second": 0.408,
555
+ "eval_steps_per_second": 0.408,
556
+ "step": 700
557
+ },
558
+ {
559
+ "epoch": 1.463678516228748,
560
+ "grad_norm": 0.5562382340431213,
561
+ "learning_rate": 6.0891937273317935e-05,
562
+ "loss": 0.1407,
563
+ "step": 710
564
+ },
565
+ {
566
+ "epoch": 1.484286450283359,
567
+ "grad_norm": 0.4450112581253052,
568
+ "learning_rate": 5.9717733998800803e-05,
569
+ "loss": 0.1474,
570
+ "step": 720
571
+ },
572
+ {
573
+ "epoch": 1.50489438433797,
574
+ "grad_norm": 0.5950725674629211,
575
+ "learning_rate": 5.853793360434687e-05,
576
+ "loss": 0.1393,
577
+ "step": 730
578
+ },
579
+ {
580
+ "epoch": 1.525502318392581,
581
+ "grad_norm": 0.5703924298286438,
582
+ "learning_rate": 5.735321561918697e-05,
583
+ "loss": 0.1396,
584
+ "step": 740
585
+ },
586
+ {
587
+ "epoch": 1.5461102524471921,
588
+ "grad_norm": 0.5534270405769348,
589
+ "learning_rate": 5.61642624049349e-05,
590
+ "loss": 0.1296,
591
+ "step": 750
592
+ },
593
+ {
594
+ "epoch": 1.5667181865018032,
595
+ "grad_norm": 0.5007007122039795,
596
+ "learning_rate": 5.497175876256796e-05,
597
+ "loss": 0.1261,
598
+ "step": 760
599
+ },
600
+ {
601
+ "epoch": 1.5873261205564142,
602
+ "grad_norm": 0.6271963715553284,
603
+ "learning_rate": 5.377639153800229e-05,
604
+ "loss": 0.1513,
605
+ "step": 770
606
+ },
607
+ {
608
+ "epoch": 1.6079340546110252,
609
+ "grad_norm": 0.5237583518028259,
610
+ "learning_rate": 5.2578849226490525e-05,
611
+ "loss": 0.1465,
612
+ "step": 780
613
+ },
614
+ {
615
+ "epoch": 1.6285419886656363,
616
+ "grad_norm": 0.4081502854824066,
617
+ "learning_rate": 5.137982157606937e-05,
618
+ "loss": 0.1214,
619
+ "step": 790
620
+ },
621
+ {
622
+ "epoch": 1.6491499227202473,
623
+ "grad_norm": 0.670275866985321,
624
+ "learning_rate": 5.017999919028566e-05,
625
+ "loss": 0.1371,
626
+ "step": 800
627
+ },
628
+ {
629
+ "epoch": 1.6491499227202473,
630
+ "eval_loss": 0.19273868203163147,
631
+ "eval_runtime": 146.7244,
632
+ "eval_samples_per_second": 0.409,
633
+ "eval_steps_per_second": 0.409,
634
+ "step": 800
635
  }
636
  ],
637
  "logging_steps": 10,
 
651
  "attributes": {}
652
  }
653
  },
654
+ "total_flos": 4.309328094007296e+16,
655
  "train_batch_size": 1,
656
  "trial_name": null,
657
  "trial_params": null