rootxhacker commited on
Commit
cce6dcc
·
verified ·
1 Parent(s): 17e4788

Training in progress, step 22500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7eaab29466400eb2a657f3436b2818198ebc3b9fd87cb963c622b5d81bbf4d9
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48890795503039966e7c3f8690ea5e69f170d6d775e60a7c19814730a3ff98f0
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55f3c6089dbca62c8b2bebd6faf4edc5b1bb8a8a960c2f857f9b4984ed1a89d0
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:323d035573abc68ea1fb28a0dbd169b42c919244458b9d48c24a8196ef4f1fbd
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:625da12a62ce1ec712e7ea0b4fc06eed36ff20b9634198d1bc6029989e807f8d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0edeec917b285b0e88139f206de61b3e301e9b677aa51d87d86f36aba6dd8e7e
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46c97cfd73b672df8a5afb8d418c975440124b784cf73259cde1c1d1e99e483d
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b66cb5271b1546b480a6c7621e1addbb2863a83f6bfe9ced4a812a79ec729e68
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:137a43b60deca6dd3da445bd48fbb2a7c159344a7a642ca673dc22a6f6abb915
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee960841273de5fa43863b36a2f8d901df0b6ff85b61e6b4992f108a85a45d9c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 22000,
3
- "best_metric": 1.4664525985717773,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-22000",
5
- "epoch": 1.6921775248057842,
6
  "eval_steps": 250,
7
- "global_step": 22000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3792,6 +3792,92 @@
3792
  "eval_samples_per_second": 55.454,
3793
  "eval_steps_per_second": 13.863,
3794
  "step": 22000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3795
  }
3796
  ],
3797
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 22250,
3
+ "best_metric": 1.457463264465332,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-22000",
5
+ "epoch": 1.7306361049150065,
6
  "eval_steps": 250,
7
+ "global_step": 22500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3792
  "eval_samples_per_second": 55.454,
3793
  "eval_steps_per_second": 13.863,
3794
  "step": 22000
3795
+ },
3796
+ {
3797
+ "epoch": 1.6960233828167064,
3798
+ "grad_norm": 0.8262001276016235,
3799
+ "learning_rate": 2.7087974706812248e-05,
3800
+ "loss": 1.4304,
3801
+ "step": 22050
3802
+ },
3803
+ {
3804
+ "epoch": 1.6998692408276286,
3805
+ "grad_norm": 1.6224443912506104,
3806
+ "learning_rate": 2.6830631226793128e-05,
3807
+ "loss": 1.4127,
3808
+ "step": 22100
3809
+ },
3810
+ {
3811
+ "epoch": 1.703715098838551,
3812
+ "grad_norm": 1.3338160514831543,
3813
+ "learning_rate": 2.6573287746774014e-05,
3814
+ "loss": 1.4842,
3815
+ "step": 22150
3816
+ },
3817
+ {
3818
+ "epoch": 1.707560956849473,
3819
+ "grad_norm": 1.940238356590271,
3820
+ "learning_rate": 2.6315944266754897e-05,
3821
+ "loss": 1.4279,
3822
+ "step": 22200
3823
+ },
3824
+ {
3825
+ "epoch": 1.7114068148603954,
3826
+ "grad_norm": 2.091132164001465,
3827
+ "learning_rate": 2.6058600786735777e-05,
3828
+ "loss": 1.3779,
3829
+ "step": 22250
3830
+ },
3831
+ {
3832
+ "epoch": 1.7114068148603954,
3833
+ "eval_loss": 1.457463264465332,
3834
+ "eval_runtime": 18.1835,
3835
+ "eval_samples_per_second": 54.995,
3836
+ "eval_steps_per_second": 13.749,
3837
+ "step": 22250
3838
+ },
3839
+ {
3840
+ "epoch": 1.7152526728713176,
3841
+ "grad_norm": 1.4367913007736206,
3842
+ "learning_rate": 2.5801257306716663e-05,
3843
+ "loss": 1.4821,
3844
+ "step": 22300
3845
+ },
3846
+ {
3847
+ "epoch": 1.7190985308822397,
3848
+ "grad_norm": 1.9735435247421265,
3849
+ "learning_rate": 2.5543913826697546e-05,
3850
+ "loss": 1.3754,
3851
+ "step": 22350
3852
+ },
3853
+ {
3854
+ "epoch": 1.7229443888931621,
3855
+ "grad_norm": 1.4968055486679077,
3856
+ "learning_rate": 2.5286570346678426e-05,
3857
+ "loss": 1.4045,
3858
+ "step": 22400
3859
+ },
3860
+ {
3861
+ "epoch": 1.7267902469040843,
3862
+ "grad_norm": 1.0449949502944946,
3863
+ "learning_rate": 2.5029226866659312e-05,
3864
+ "loss": 1.4458,
3865
+ "step": 22450
3866
+ },
3867
+ {
3868
+ "epoch": 1.7306361049150065,
3869
+ "grad_norm": 1.164890170097351,
3870
+ "learning_rate": 2.4771883386640196e-05,
3871
+ "loss": 1.4407,
3872
+ "step": 22500
3873
+ },
3874
+ {
3875
+ "epoch": 1.7306361049150065,
3876
+ "eval_loss": 1.4607012271881104,
3877
+ "eval_runtime": 18.2079,
3878
+ "eval_samples_per_second": 54.921,
3879
+ "eval_steps_per_second": 13.73,
3880
+ "step": 22500
3881
  }
3882
  ],
3883
  "logging_steps": 50,