alicegoesdown commited on
Commit
cd724a0
·
verified ·
1 Parent(s): c4ea699

Training in progress, step 750, checkpoint

Browse files
last-checkpoint/lora_top/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74c935a4db11d9954248a9ac8c7f1ed129a2d53ead6a4137b79d300258c6b6fd
3
  size 12591264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5999d72c1222f72a4968e4d8e4eb3c8728c8d56481f4f0e1e0c0a1f3c5ff835
3
  size 12591264
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c794acd8ed44a50952d429716499bc5f43af1db5a3592ccabf56d02c4146e251
3
  size 25206586
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:611b03eef46d707027dae04702e351acde7c973d8bf2ea65e1a47d8aa941d2d4
3
  size 25206586
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe75e0f6d5068d897d45b2b460d1ca19c86e7c11996a5acadcd9e0af782efe20
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6b729b9def64816500ff70c349edf6d0ea9335a9d5fd4690519f3e17c4cb8e3
3
  size 14180
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e6f142d93dbd6e62ebe01d74ed2ceb68210344a36e52cda5c9422f1f0eeb827
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4bd676352a42650ddef3ebd028841320ee20dd5d83b6e34b1d1463e0d94bcf5
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 3.631777286529541,
3
- "best_model_checkpoint": "./output/checkpoint-600",
4
- "epoch": 0.9259259259259259,
5
  "eval_steps": 150,
6
- "global_step": 600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -459,6 +459,119 @@
459
  "eval_samples_per_second": 34.423,
460
  "eval_steps_per_second": 34.423,
461
  "step": 600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
462
  }
463
  ],
464
  "logging_steps": 10,
@@ -478,7 +591,7 @@
478
  "attributes": {}
479
  }
480
  },
481
- "total_flos": 2268906529357824.0,
482
  "train_batch_size": 16,
483
  "trial_name": null,
484
  "trial_params": null
 
1
  {
2
+ "best_metric": 3.404348134994507,
3
+ "best_model_checkpoint": "./output/checkpoint-750",
4
+ "epoch": 1.1574074074074074,
5
  "eval_steps": 150,
6
+ "global_step": 750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
459
  "eval_samples_per_second": 34.423,
460
  "eval_steps_per_second": 34.423,
461
  "step": 600
462
+ },
463
+ {
464
+ "epoch": 0.941358024691358,
465
+ "grad_norm": 3.062471866607666,
466
+ "learning_rate": 0.0002920523948642432,
467
+ "loss": 3.4515,
468
+ "step": 610
469
+ },
470
+ {
471
+ "epoch": 0.9567901234567902,
472
+ "grad_norm": 2.787814140319824,
473
+ "learning_rate": 0.00029174058845931434,
474
+ "loss": 3.6189,
475
+ "step": 620
476
+ },
477
+ {
478
+ "epoch": 0.9722222222222222,
479
+ "grad_norm": 2.71366810798645,
480
+ "learning_rate": 0.0002914229556539539,
481
+ "loss": 3.4178,
482
+ "step": 630
483
+ },
484
+ {
485
+ "epoch": 0.9876543209876543,
486
+ "grad_norm": 2.544954776763916,
487
+ "learning_rate": 0.00029109950950480235,
488
+ "loss": 3.5334,
489
+ "step": 640
490
+ },
491
+ {
492
+ "epoch": 1.0030864197530864,
493
+ "grad_norm": 2.5218825340270996,
494
+ "learning_rate": 0.00029077026330746403,
495
+ "loss": 3.3136,
496
+ "step": 650
497
+ },
498
+ {
499
+ "epoch": 1.0185185185185186,
500
+ "grad_norm": 3.377516746520996,
501
+ "learning_rate": 0.00029043523059596053,
502
+ "loss": 3.0759,
503
+ "step": 660
504
+ },
505
+ {
506
+ "epoch": 1.0339506172839505,
507
+ "grad_norm": 3.7312910556793213,
508
+ "learning_rate": 0.00029009442514217447,
509
+ "loss": 3.0688,
510
+ "step": 670
511
+ },
512
+ {
513
+ "epoch": 1.0493827160493827,
514
+ "grad_norm": 3.517127513885498,
515
+ "learning_rate": 0.00028974786095528306,
516
+ "loss": 3.1925,
517
+ "step": 680
518
+ },
519
+ {
520
+ "epoch": 1.0648148148148149,
521
+ "grad_norm": 3.9042561054229736,
522
+ "learning_rate": 0.00028939555228118277,
523
+ "loss": 3.1548,
524
+ "step": 690
525
+ },
526
+ {
527
+ "epoch": 1.0802469135802468,
528
+ "grad_norm": 3.7608659267425537,
529
+ "learning_rate": 0.0002890375136019032,
530
+ "loss": 3.0335,
531
+ "step": 700
532
+ },
533
+ {
534
+ "epoch": 1.095679012345679,
535
+ "grad_norm": 3.591569662094116,
536
+ "learning_rate": 0.0002886737596350122,
537
+ "loss": 3.1341,
538
+ "step": 710
539
+ },
540
+ {
541
+ "epoch": 1.1111111111111112,
542
+ "grad_norm": 3.2849345207214355,
543
+ "learning_rate": 0.0002883043053330106,
544
+ "loss": 2.9512,
545
+ "step": 720
546
+ },
547
+ {
548
+ "epoch": 1.126543209876543,
549
+ "grad_norm": 3.238452196121216,
550
+ "learning_rate": 0.0002879291658827176,
551
+ "loss": 2.9095,
552
+ "step": 730
553
+ },
554
+ {
555
+ "epoch": 1.1419753086419753,
556
+ "grad_norm": 4.236405849456787,
557
+ "learning_rate": 0.00028754835670464677,
558
+ "loss": 2.9601,
559
+ "step": 740
560
+ },
561
+ {
562
+ "epoch": 1.1574074074074074,
563
+ "grad_norm": 3.696310043334961,
564
+ "learning_rate": 0.00028716189345237184,
565
+ "loss": 2.9871,
566
+ "step": 750
567
+ },
568
+ {
569
+ "epoch": 1.1574074074074074,
570
+ "eval_loss": 3.404348134994507,
571
+ "eval_runtime": 13.6336,
572
+ "eval_samples_per_second": 36.674,
573
+ "eval_steps_per_second": 36.674,
574
+ "step": 750
575
  }
576
  ],
577
  "logging_steps": 10,
 
591
  "attributes": {}
592
  }
593
  },
594
+ "total_flos": 2848396981714944.0,
595
  "train_batch_size": 16,
596
  "trial_name": null,
597
  "trial_params": null