rootxhacker commited on
Commit
a30f9f4
·
verified ·
1 Parent(s): b6371ed

Training in progress, step 16000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6067e5da0c9bf8b752e4ae8c8bf60db131ffcb5bdc72c1bef0c50cf034756793
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72bbf1639eb1e5bb9c61f65f54a93961c6ba815ff973db8268065d37b1392548
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70b3ce160c7b776e775a166f1cb68bd1487d55fdb2e7ccd66b23097de2464f87
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca4486e1d2b5552334fa9da91cfad2dd01fc4b1d85bf70178d94c3a9a06c15d0
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e443e4b8c749db54f7826bd4573803e0f05009235443f881fa9932376e1aba8
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2268c750118e17cd7866635da6e44831e071d618455f938ab68da6b4daa1e485
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c46d93b856d73467b22e6683f2d8e760cc9d35c9e6261e9397f7ae15ef306957
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d55844b7da2e1071a7670fddb811c2bc2d24a926bedcc2bd08e1031b2ef69bf6
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:682c1f39bde20c90a4373af21bd5402ac0e8e55951fcfe8f05caeac3b074ab9a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1ee899fe281691d25f12760d42a5b17bb85c5d0104daf152b3a9bdedb20881a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 15500,
3
- "best_metric": 1.4925825595855713,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-15500",
5
- "epoch": 1.1922159833858934,
6
  "eval_steps": 250,
7
- "global_step": 15500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2674,6 +2674,92 @@
2674
  "eval_samples_per_second": 54.91,
2675
  "eval_steps_per_second": 13.727,
2676
  "step": 15500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2677
  }
2678
  ],
2679
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 15750,
3
+ "best_metric": 1.4893933534622192,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-15500",
5
+ "epoch": 1.2306745634951157,
6
  "eval_steps": 250,
7
+ "global_step": 16000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2674
  "eval_samples_per_second": 54.91,
2675
  "eval_steps_per_second": 13.727,
2676
  "step": 15500
2677
+ },
2678
+ {
2679
+ "epoch": 1.1960618413968156,
2680
+ "grad_norm": 2.7062911987304688,
2681
+ "learning_rate": 6.053748023969707e-05,
2682
+ "loss": 1.4145,
2683
+ "step": 15550
2684
+ },
2685
+ {
2686
+ "epoch": 1.1999076994077378,
2687
+ "grad_norm": 1.5163620710372925,
2688
+ "learning_rate": 6.028013675967794e-05,
2689
+ "loss": 1.4322,
2690
+ "step": 15600
2691
+ },
2692
+ {
2693
+ "epoch": 1.2037535574186602,
2694
+ "grad_norm": 1.342063546180725,
2695
+ "learning_rate": 6.002279327965883e-05,
2696
+ "loss": 1.4696,
2697
+ "step": 15650
2698
+ },
2699
+ {
2700
+ "epoch": 1.2075994154295824,
2701
+ "grad_norm": 1.8180099725723267,
2702
+ "learning_rate": 5.9765449799639715e-05,
2703
+ "loss": 1.4647,
2704
+ "step": 15700
2705
+ },
2706
+ {
2707
+ "epoch": 1.2114452734405046,
2708
+ "grad_norm": 1.951982855796814,
2709
+ "learning_rate": 5.9508106319620595e-05,
2710
+ "loss": 1.5141,
2711
+ "step": 15750
2712
+ },
2713
+ {
2714
+ "epoch": 1.2114452734405046,
2715
+ "eval_loss": 1.4893933534622192,
2716
+ "eval_runtime": 18.1951,
2717
+ "eval_samples_per_second": 54.96,
2718
+ "eval_steps_per_second": 13.74,
2719
+ "step": 15750
2720
+ },
2721
+ {
2722
+ "epoch": 1.2152911314514268,
2723
+ "grad_norm": 1.7536894083023071,
2724
+ "learning_rate": 5.925076283960148e-05,
2725
+ "loss": 1.514,
2726
+ "step": 15800
2727
+ },
2728
+ {
2729
+ "epoch": 1.2191369894623492,
2730
+ "grad_norm": 1.1857939958572388,
2731
+ "learning_rate": 5.899341935958237e-05,
2732
+ "loss": 1.4745,
2733
+ "step": 15850
2734
+ },
2735
+ {
2736
+ "epoch": 1.2229828474732714,
2737
+ "grad_norm": 1.2500842809677124,
2738
+ "learning_rate": 5.873607587956324e-05,
2739
+ "loss": 1.4325,
2740
+ "step": 15900
2741
+ },
2742
+ {
2743
+ "epoch": 1.2268287054841935,
2744
+ "grad_norm": 2.025336742401123,
2745
+ "learning_rate": 5.847873239954413e-05,
2746
+ "loss": 1.4913,
2747
+ "step": 15950
2748
+ },
2749
+ {
2750
+ "epoch": 1.2306745634951157,
2751
+ "grad_norm": 1.1440426111221313,
2752
+ "learning_rate": 5.8221388919525014e-05,
2753
+ "loss": 1.451,
2754
+ "step": 16000
2755
+ },
2756
+ {
2757
+ "epoch": 1.2306745634951157,
2758
+ "eval_loss": 1.492313265800476,
2759
+ "eval_runtime": 18.0024,
2760
+ "eval_samples_per_second": 55.548,
2761
+ "eval_steps_per_second": 13.887,
2762
+ "step": 16000
2763
  }
2764
  ],
2765
  "logging_steps": 50,