rootxhacker commited on
Commit
f94e0c1
·
verified ·
1 Parent(s): 3815efd

Training in progress, step 4000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24082cd1025a9c16d6bc6449cbc8ef1f1d242da73316dafd4f2689fc8c860917
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75410142e7d88776abfc3caa1d65963cd8d85da0ff4f4ef2a8275d75bab13fc7
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4d61ee4cf28e305ccbd9c0cd63b04a590ac2c163f4c12096fd42f05674ed054
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:543307a68c41ab7eab4206e5219ee92650f34cf85fdde0ae16f2041a7d666f3f
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32c4abfc978e57dde0e9122ce6e0a211a14cdb3e9d7e1f8e79d82dbf8734a5a1
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ba374ceeab3c18ecca1a51999f0124017113854b9f36714acb5b3cbd2a2d7ce
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:835fd47c5d5c5f39bf487b0cc415ae7b687f6d1d1601a7df5275be58f8fd603c
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51fee8eb233312b3e7cebdb2d20739865391ab32cb8dc65d3edc7a1f54a4611d
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b0c041eadc91e24d174b435fcb1ec46a0c56f36b09b7dbbf49009145cd7fa33
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c31d829461e38fcd04f37e8cd6d881ff82a519c1416dbfe781c31b6217218abc
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 3500,
3
  "best_metric": 4.488556861877441,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-3500",
5
- "epoch": 0.26921006076455656,
6
  "eval_steps": 250,
7
- "global_step": 3500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -610,6 +610,92 @@
610
  "eval_samples_per_second": 52.851,
611
  "eval_steps_per_second": 13.213,
612
  "step": 3500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
613
  }
614
  ],
615
  "logging_steps": 50,
 
2
  "best_global_step": 3500,
3
  "best_metric": 4.488556861877441,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-3500",
5
+ "epoch": 0.30766864087377893,
6
  "eval_steps": 250,
7
+ "global_step": 4000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
610
  "eval_samples_per_second": 52.851,
611
  "eval_steps_per_second": 13.213,
612
  "step": 3500
613
+ },
614
+ {
615
+ "epoch": 0.2730559187754788,
616
+ "grad_norm": 3.794307231903076,
617
+ "learning_rate": 0.00012227418109628322,
618
+ "loss": 4.4994,
619
+ "step": 3550
620
+ },
621
+ {
622
+ "epoch": 0.27690177678640104,
623
+ "grad_norm": 3.5770812034606934,
624
+ "learning_rate": 0.00012201683761626409,
625
+ "loss": 4.5888,
626
+ "step": 3600
627
+ },
628
+ {
629
+ "epoch": 0.2807476347973233,
630
+ "grad_norm": 4.770874500274658,
631
+ "learning_rate": 0.00012175949413624498,
632
+ "loss": 4.5644,
633
+ "step": 3650
634
+ },
635
+ {
636
+ "epoch": 0.28459349280824553,
637
+ "grad_norm": 3.4447147846221924,
638
+ "learning_rate": 0.00012150215065622586,
639
+ "loss": 4.5301,
640
+ "step": 3700
641
+ },
642
+ {
643
+ "epoch": 0.2884393508191678,
644
+ "grad_norm": 4.76978063583374,
645
+ "learning_rate": 0.00012124480717620675,
646
+ "loss": 4.5563,
647
+ "step": 3750
648
+ },
649
+ {
650
+ "epoch": 0.2884393508191678,
651
+ "eval_loss": 4.53049898147583,
652
+ "eval_runtime": 18.9074,
653
+ "eval_samples_per_second": 52.889,
654
+ "eval_steps_per_second": 13.222,
655
+ "step": 3750
656
+ },
657
+ {
658
+ "epoch": 0.29228520883009,
659
+ "grad_norm": 5.7456512451171875,
660
+ "learning_rate": 0.00012098746369618763,
661
+ "loss": 4.5612,
662
+ "step": 3800
663
+ },
664
+ {
665
+ "epoch": 0.2961310668410122,
666
+ "grad_norm": 5.577849864959717,
667
+ "learning_rate": 0.00012073012021616851,
668
+ "loss": 4.4629,
669
+ "step": 3850
670
+ },
671
+ {
672
+ "epoch": 0.29997692485193445,
673
+ "grad_norm": 4.432284832000732,
674
+ "learning_rate": 0.00012047277673614939,
675
+ "loss": 4.6661,
676
+ "step": 3900
677
+ },
678
+ {
679
+ "epoch": 0.3038227828628567,
680
+ "grad_norm": 5.174475193023682,
681
+ "learning_rate": 0.00012021543325613027,
682
+ "loss": 4.4835,
683
+ "step": 3950
684
+ },
685
+ {
686
+ "epoch": 0.30766864087377893,
687
+ "grad_norm": 3.5657413005828857,
688
+ "learning_rate": 0.00011995808977611117,
689
+ "loss": 4.4894,
690
+ "step": 4000
691
+ },
692
+ {
693
+ "epoch": 0.30766864087377893,
694
+ "eval_loss": 4.5068535804748535,
695
+ "eval_runtime": 18.7951,
696
+ "eval_samples_per_second": 53.205,
697
+ "eval_steps_per_second": 13.301,
698
+ "step": 4000
699
  }
700
  ],
701
  "logging_steps": 50,