rootxhacker commited on
Commit
fd2fe06
·
verified ·
1 Parent(s): 0de08a0

Training in progress, step 17000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d2bb37f0bb2536ca33711abe7b3143ef3974ef2ca0c218977672374d7ca8e27
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b80b3583e2afb611f4c9adf5dda97e41fa3abaeec3239ac74a43ddde70a05578
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3916cf9823a85ba87d1391c608e5091a2091feb324e91b9bb2a76ddf86c0982d
3
  size 1544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfba693cf1d282bbea5c304bdd309307768a9ce459d448905954c71c856485be
3
  size 1544
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:518b6ef8fc143d463491edf1260763c10b5e11e873cd36abf76a046091e577c9
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba8347dc980899d1fbf4c0d68fa8a8cf6d3395e6ddd6aee4b5d0bea2791716a0
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d28d5f3a49ada942ec355f276b33d6e4bff345075b872b6e2b651f5666a06f0a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a9d5359cc143881ecdb88768e105d086fd2336ed10ecf8c85dd03dc0e505da1
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b340848445b4f5b2d5567ddb4c818db5bf89137348148d19bd0b15ff0d69c43d
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3333c3c82f9e4ec8ba5829fe0ea7dc977fdffe09f41702806e9515db920536e
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d72187e2b5b8f8cffa32a0fbea90f82e23c165518eb95af21f7e18bd9e8c4a42
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14c33e4d9235be55afaa758d537d857af11e671e33d9253fad9c2c79b3373700
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 16250,
3
- "best_metric": 1.4893407821655273,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-15500",
5
- "epoch": 1.2691331436043383,
6
  "eval_steps": 250,
7
- "global_step": 16500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2846,6 +2846,92 @@
2846
  "eval_samples_per_second": 55.967,
2847
  "eval_steps_per_second": 13.992,
2848
  "step": 16500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2849
  }
2850
  ],
2851
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 16750,
3
+ "best_metric": 1.4798808097839355,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-15500",
5
+ "epoch": 1.3075917237135606,
6
  "eval_steps": 250,
7
+ "global_step": 17000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2846
  "eval_samples_per_second": 55.967,
2847
  "eval_steps_per_second": 13.992,
2848
  "step": 16500
2849
+ },
2850
+ {
2851
+ "epoch": 1.2729790016152602,
2852
+ "grad_norm": 1.3523976802825928,
2853
+ "learning_rate": 0.00011668181700127263,
2854
+ "loss": 1.4796,
2855
+ "step": 16550
2856
+ },
2857
+ {
2858
+ "epoch": 1.2768248596261826,
2859
+ "grad_norm": 1.9770869016647339,
2860
+ "learning_rate": 0.00011642209697945614,
2861
+ "loss": 1.4756,
2862
+ "step": 16600
2863
+ },
2864
+ {
2865
+ "epoch": 1.2806707176371048,
2866
+ "grad_norm": 1.7973159551620483,
2867
+ "learning_rate": 0.00011616237695763967,
2868
+ "loss": 1.4163,
2869
+ "step": 16650
2870
+ },
2871
+ {
2872
+ "epoch": 1.284516575648027,
2873
+ "grad_norm": 1.3054739236831665,
2874
+ "learning_rate": 0.00011590265693582318,
2875
+ "loss": 1.4105,
2876
+ "step": 16700
2877
+ },
2878
+ {
2879
+ "epoch": 1.2883624336589494,
2880
+ "grad_norm": 1.457047939300537,
2881
+ "learning_rate": 0.00011564293691400672,
2882
+ "loss": 1.4698,
2883
+ "step": 16750
2884
+ },
2885
+ {
2886
+ "epoch": 1.2883624336589494,
2887
+ "eval_loss": 1.4798808097839355,
2888
+ "eval_runtime": 17.8696,
2889
+ "eval_samples_per_second": 55.961,
2890
+ "eval_steps_per_second": 13.99,
2891
+ "step": 16750
2892
+ },
2893
+ {
2894
+ "epoch": 1.2922082916698716,
2895
+ "grad_norm": 2.117663860321045,
2896
+ "learning_rate": 0.00011538841129262656,
2897
+ "loss": 1.4252,
2898
+ "step": 16800
2899
+ },
2900
+ {
2901
+ "epoch": 1.2960541496807938,
2902
+ "grad_norm": 3.7321341037750244,
2903
+ "learning_rate": 0.00011512869127081007,
2904
+ "loss": 1.4818,
2905
+ "step": 16850
2906
+ },
2907
+ {
2908
+ "epoch": 1.299900007691716,
2909
+ "grad_norm": 2.497528553009033,
2910
+ "learning_rate": 0.00011486897124899358,
2911
+ "loss": 1.4176,
2912
+ "step": 16900
2913
+ },
2914
+ {
2915
+ "epoch": 1.3037458657026382,
2916
+ "grad_norm": 1.6085398197174072,
2917
+ "learning_rate": 0.00011460925122717712,
2918
+ "loss": 1.4373,
2919
+ "step": 16950
2920
+ },
2921
+ {
2922
+ "epoch": 1.3075917237135606,
2923
+ "grad_norm": 1.8937525749206543,
2924
+ "learning_rate": 0.00011434953120536062,
2925
+ "loss": 1.4823,
2926
+ "step": 17000
2927
+ },
2928
+ {
2929
+ "epoch": 1.3075917237135606,
2930
+ "eval_loss": 1.4854488372802734,
2931
+ "eval_runtime": 17.8761,
2932
+ "eval_samples_per_second": 55.941,
2933
+ "eval_steps_per_second": 13.985,
2934
+ "step": 17000
2935
  }
2936
  ],
2937
  "logging_steps": 50,