rootxhacker commited on
Commit
ab806e1
·
verified ·
1 Parent(s): a108ada

Training in progress, step 28500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2b2a09b0c718d7a491a19203c596e4ee868ab1c484877d294e6c4faf6f30efb
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe6e2fa1318dd3ad1205203e4fff88f82e693270053626bb1d002f9b23c9b24c
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84d5f993c46a18f90bdec40ac57cbb61c0044ca158da830a2a251cb1b9f73c26
3
  size 1544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:599432259819e93dad17855a204a171f8d5bf2703ad47dd88ce1d465a13b2f0b
3
  size 1544
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f145355c23ee5fb538578cd9954c12d1f33dfb98ccfc124985339d68356ee942
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48da092ff21d2489fb67e59a49519237044b07edc1b34a941910a561850d4e26
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a374ffed3d3d8a7f9634cdb9b7207d347f6957d04522010a4d50d73e4c623442
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e10c23934142af0d6a52cc080658216fed6f296b2eaa84b6cfd6a75c3f4683b
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13f83252b2ae8a53cd4f408d431cb4cec02e9d65b9dadef7fdff56180c3d5cd3
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:575e45baa8bffe406e93f42f00b83861120a81ba67710e20edf9cb739fb34380
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:969919a5c5fc35784da996c8ca0a4cb297b59fb4c9aadbd1a769e70ab51007e7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:841075f91d905b0f965b453df9b09a92889e8e8880ee6a19f7c088b54a9e11c6
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 28000,
3
- "best_metric": 1.008616328239441,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-28000",
5
- "epoch": 2.1536804861164525,
6
  "eval_steps": 250,
7
- "global_step": 28000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -4824,6 +4824,92 @@
4824
  "eval_samples_per_second": 56.661,
4825
  "eval_steps_per_second": 14.165,
4826
  "step": 28000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4827
  }
4828
  ],
4829
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 28500,
3
+ "best_metric": 1.0052642822265625,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-28500",
5
+ "epoch": 2.192139066225675,
6
  "eval_steps": 250,
7
+ "global_step": 28500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
4824
  "eval_samples_per_second": 56.661,
4825
  "eval_steps_per_second": 14.165,
4826
  "step": 28000
4827
+ },
4828
+ {
4829
+ "epoch": 2.157526344127375,
4830
+ "grad_norm": 1.3012685775756836,
4831
+ "learning_rate": 5.696698958522713e-05,
4832
+ "loss": 1.0703,
4833
+ "step": 28050
4834
+ },
4835
+ {
4836
+ "epoch": 2.1613722021382973,
4837
+ "grad_norm": 0.8459142446517944,
4838
+ "learning_rate": 5.6707269563410646e-05,
4839
+ "loss": 1.0507,
4840
+ "step": 28100
4841
+ },
4842
+ {
4843
+ "epoch": 2.1652180601492192,
4844
+ "grad_norm": 0.6889505982398987,
4845
+ "learning_rate": 5.6447549541594166e-05,
4846
+ "loss": 1.0056,
4847
+ "step": 28150
4848
+ },
4849
+ {
4850
+ "epoch": 2.1690639181601417,
4851
+ "grad_norm": 1.2236456871032715,
4852
+ "learning_rate": 5.618782951977768e-05,
4853
+ "loss": 0.988,
4854
+ "step": 28200
4855
+ },
4856
+ {
4857
+ "epoch": 2.1729097761710636,
4858
+ "grad_norm": 1.3419203758239746,
4859
+ "learning_rate": 5.59281094979612e-05,
4860
+ "loss": 0.9467,
4861
+ "step": 28250
4862
+ },
4863
+ {
4864
+ "epoch": 2.1729097761710636,
4865
+ "eval_loss": 1.0140153169631958,
4866
+ "eval_runtime": 17.7299,
4867
+ "eval_samples_per_second": 56.402,
4868
+ "eval_steps_per_second": 14.1,
4869
+ "step": 28250
4870
+ },
4871
+ {
4872
+ "epoch": 2.176755634181986,
4873
+ "grad_norm": 0.8394871354103088,
4874
+ "learning_rate": 5.5668389476144725e-05,
4875
+ "loss": 0.9824,
4876
+ "step": 28300
4877
+ },
4878
+ {
4879
+ "epoch": 2.1806014921929084,
4880
+ "grad_norm": 0.7906908392906189,
4881
+ "learning_rate": 5.540866945432823e-05,
4882
+ "loss": 0.9915,
4883
+ "step": 28350
4884
+ },
4885
+ {
4886
+ "epoch": 2.1844473502038304,
4887
+ "grad_norm": 1.0852785110473633,
4888
+ "learning_rate": 5.514894943251175e-05,
4889
+ "loss": 0.9473,
4890
+ "step": 28400
4891
+ },
4892
+ {
4893
+ "epoch": 2.188293208214753,
4894
+ "grad_norm": 0.9187583923339844,
4895
+ "learning_rate": 5.488922941069528e-05,
4896
+ "loss": 0.9418,
4897
+ "step": 28450
4898
+ },
4899
+ {
4900
+ "epoch": 2.192139066225675,
4901
+ "grad_norm": 0.7200838923454285,
4902
+ "learning_rate": 5.4629509388878797e-05,
4903
+ "loss": 0.9776,
4904
+ "step": 28500
4905
+ },
4906
+ {
4907
+ "epoch": 2.192139066225675,
4908
+ "eval_loss": 1.0052642822265625,
4909
+ "eval_runtime": 17.5069,
4910
+ "eval_samples_per_second": 57.12,
4911
+ "eval_steps_per_second": 14.28,
4912
+ "step": 28500
4913
  }
4914
  ],
4915
  "logging_steps": 50,