rootxhacker commited on
Commit
1276257
·
verified ·
1 Parent(s): 38f9aad

Training in progress, step 12500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04d930c5c4a881dfd26ea978e652c702b6adb1b5b7ed1eed9ca7135e21510049
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd49fdfe9bdd857be5419dddbdbaa6f93701dd16afd12b8202d9abb0f608e217
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59ca03346afd36214198fb9410564c8a86b40b37890c4e424de07767c10a30ec
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8066ffe6e1fa6681b38c047067ca29a6eed18da1b1acf88431943e295d78e330
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:205a7550a4c055924e0126c3501b4b656ca55ae3243c46218f86d29fc2fb9758
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a5ae118ec673d57a7f62a6284956ef83acf5113a9420dcd7ba5fd941799b59c
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:699adc1847435a16f1605601a8df2b43b7960dcbb43148ad8a4600f5fe2cb9d4
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63eee2f78c948ec330b424961c4bdfb29608b15bd25064fec9d957a2f7fcbcd2
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:638497dfdf5eb5c99bf0b6f11b605f213450ed644d6ebdee360506f0a10a4ae0
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:334ae361472f93791190a5614ebfa0a90c8bcb9632490f070b08594d58ab788b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 12000,
3
- "best_metric": 1.5159597396850586,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-12000",
5
- "epoch": 0.9230059226213368,
6
  "eval_steps": 250,
7
- "global_step": 12000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2072,6 +2072,92 @@
2072
  "eval_samples_per_second": 55.894,
2073
  "eval_steps_per_second": 13.974,
2074
  "step": 12000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2075
  }
2076
  ],
2077
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 12250,
3
+ "best_metric": 1.5108226537704468,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-12000",
5
+ "epoch": 0.9614645027305592,
6
  "eval_steps": 250,
7
+ "global_step": 12500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2072
  "eval_samples_per_second": 55.894,
2073
  "eval_steps_per_second": 13.974,
2074
  "step": 12000
2075
+ },
2076
+ {
2077
+ "epoch": 0.926851780632259,
2078
+ "grad_norm": 1.736138939857483,
2079
+ "learning_rate": 7.855152384103525e-05,
2080
+ "loss": 1.5304,
2081
+ "step": 12050
2082
+ },
2083
+ {
2084
+ "epoch": 0.9306976386431813,
2085
+ "grad_norm": 1.807916283607483,
2086
+ "learning_rate": 7.829418036101613e-05,
2087
+ "loss": 1.4984,
2088
+ "step": 12100
2089
+ },
2090
+ {
2091
+ "epoch": 0.9345434966541035,
2092
+ "grad_norm": 1.1977109909057617,
2093
+ "learning_rate": 7.803683688099702e-05,
2094
+ "loss": 1.4307,
2095
+ "step": 12150
2096
+ },
2097
+ {
2098
+ "epoch": 0.9383893546650258,
2099
+ "grad_norm": 0.8386535048484802,
2100
+ "learning_rate": 7.77794934009779e-05,
2101
+ "loss": 1.444,
2102
+ "step": 12200
2103
+ },
2104
+ {
2105
+ "epoch": 0.942235212675948,
2106
+ "grad_norm": 1.395053744316101,
2107
+ "learning_rate": 7.752214992095878e-05,
2108
+ "loss": 1.4866,
2109
+ "step": 12250
2110
+ },
2111
+ {
2112
+ "epoch": 0.942235212675948,
2113
+ "eval_loss": 1.5108226537704468,
2114
+ "eval_runtime": 18.0888,
2115
+ "eval_samples_per_second": 55.283,
2116
+ "eval_steps_per_second": 13.821,
2117
+ "step": 12250
2118
+ },
2119
+ {
2120
+ "epoch": 0.9460810706868702,
2121
+ "grad_norm": 1.5271111726760864,
2122
+ "learning_rate": 7.726480644093966e-05,
2123
+ "loss": 1.4849,
2124
+ "step": 12300
2125
+ },
2126
+ {
2127
+ "epoch": 0.9499269286977925,
2128
+ "grad_norm": 3.0610506534576416,
2129
+ "learning_rate": 7.700746296092054e-05,
2130
+ "loss": 1.4613,
2131
+ "step": 12350
2132
+ },
2133
+ {
2134
+ "epoch": 0.9537727867087147,
2135
+ "grad_norm": 1.8968026638031006,
2136
+ "learning_rate": 7.675011948090143e-05,
2137
+ "loss": 1.591,
2138
+ "step": 12400
2139
+ },
2140
+ {
2141
+ "epoch": 0.957618644719637,
2142
+ "grad_norm": 1.748979926109314,
2143
+ "learning_rate": 7.649277600088231e-05,
2144
+ "loss": 1.4781,
2145
+ "step": 12450
2146
+ },
2147
+ {
2148
+ "epoch": 0.9614645027305592,
2149
+ "grad_norm": 1.6586661338806152,
2150
+ "learning_rate": 7.62354325208632e-05,
2151
+ "loss": 1.4668,
2152
+ "step": 12500
2153
+ },
2154
+ {
2155
+ "epoch": 0.9614645027305592,
2156
+ "eval_loss": 1.5503162145614624,
2157
+ "eval_runtime": 17.9222,
2158
+ "eval_samples_per_second": 55.797,
2159
+ "eval_steps_per_second": 13.949,
2160
+ "step": 12500
2161
  }
2162
  ],
2163
  "logging_steps": 50,