rootxhacker commited on
Commit
02d6d70
·
verified ·
1 Parent(s): 7fa9751

Training in progress, step 13000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a623f2e366d56a653f11e3f01819889dc1645d4d1136c5b510bf0c1c92dab66
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34351883b0f6e42fd271ed7ebc5095493f25642a1a75cc341d7789a6aeca36d0
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6015f4fec963cdcdbd9534ce342304afb8094ad5f761558a674ddf3068ab0335
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39353fd13db5daeeb97a98bc4902673ea2228d2bc5b271b25f3916da657ebef6
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f314be3a1ba2ce8f5f725cf645fa00681145b061c07cf8ad59e1e13356924cb
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8448ab33e8a6201d670f6c9fb2d20010c66be0d08d4deb6a1a6fb49d8b9c406d
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:082b149e63f0f47e9e60d37040a9d015e2fa6a3d80388f6fa7d859aa3870fd21
3
- size 14308
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9893f35eae5cc019e9b33dde5eb3fb2bd041941900e50675b17b7d1262060d8f
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0678e6da71fe39cdc83977acb14900d45184a734b31b0b8a20ead5262367273c
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2b1ebe381b38de043f6835dd6952f12b83db6f06939f85a5aefa0eb3268de14
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7680f728f95fcd934cd32ef342f02b8f2b507472a27e9e8de734cc5468e1428
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adb9317783409b5e27bc46a88956d26ababa0364a288c25ea22c620413a62bd0
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 12500,
3
- "best_metric": -30.455406188964844,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-12500",
5
- "epoch": 0.9614645027305592,
6
  "eval_steps": 250,
7
- "global_step": 12500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2158,6 +2158,92 @@
2158
  "eval_samples_per_second": 61.423,
2159
  "eval_steps_per_second": 15.356,
2160
  "step": 12500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2161
  }
2162
  ],
2163
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 13000,
3
+ "best_metric": -30.469120025634766,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-13000",
5
+ "epoch": 0.9999230828397816,
6
  "eval_steps": 250,
7
+ "global_step": 13000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2158
  "eval_samples_per_second": 61.423,
2159
  "eval_steps_per_second": 15.356,
2160
  "step": 12500
2161
+ },
2162
+ {
2163
+ "epoch": 0.9653103607414815,
2164
+ "grad_norm": 0.1690986603498459,
2165
+ "learning_rate": 0.00013744902994571853,
2166
+ "loss": -30.4565,
2167
+ "step": 12550
2168
+ },
2169
+ {
2170
+ "epoch": 0.9691562187524037,
2171
+ "grad_norm": 0.15991179645061493,
2172
+ "learning_rate": 0.00013718930992390204,
2173
+ "loss": -30.4596,
2174
+ "step": 12600
2175
+ },
2176
+ {
2177
+ "epoch": 0.973002076763326,
2178
+ "grad_norm": 0.3283814787864685,
2179
+ "learning_rate": 0.00013692958990208556,
2180
+ "loss": -30.4609,
2181
+ "step": 12650
2182
+ },
2183
+ {
2184
+ "epoch": 0.9768479347742481,
2185
+ "grad_norm": 0.09776933491230011,
2186
+ "learning_rate": 0.0001366698698802691,
2187
+ "loss": -30.4628,
2188
+ "step": 12700
2189
+ },
2190
+ {
2191
+ "epoch": 0.9806937927851703,
2192
+ "grad_norm": 0.16940400004386902,
2193
+ "learning_rate": 0.0001364101498584526,
2194
+ "loss": -30.4646,
2195
+ "step": 12750
2196
+ },
2197
+ {
2198
+ "epoch": 0.9806937927851703,
2199
+ "eval_loss": -30.46562385559082,
2200
+ "eval_runtime": 16.3759,
2201
+ "eval_samples_per_second": 61.065,
2202
+ "eval_steps_per_second": 15.266,
2203
+ "step": 12750
2204
+ },
2205
+ {
2206
+ "epoch": 0.9845396507960926,
2207
+ "grad_norm": 0.1123572438955307,
2208
+ "learning_rate": 0.0001361504298366361,
2209
+ "loss": -30.4657,
2210
+ "step": 12800
2211
+ },
2212
+ {
2213
+ "epoch": 0.9883855088070148,
2214
+ "grad_norm": 0.19518135488033295,
2215
+ "learning_rate": 0.00013589070981481963,
2216
+ "loss": -30.4662,
2217
+ "step": 12850
2218
+ },
2219
+ {
2220
+ "epoch": 0.9922313668179371,
2221
+ "grad_norm": 0.09862416982650757,
2222
+ "learning_rate": 0.00013563098979300315,
2223
+ "loss": -30.4669,
2224
+ "step": 12900
2225
+ },
2226
+ {
2227
+ "epoch": 0.9960772248288593,
2228
+ "grad_norm": 0.08367747068405151,
2229
+ "learning_rate": 0.00013537126977118666,
2230
+ "loss": -30.4679,
2231
+ "step": 12950
2232
+ },
2233
+ {
2234
+ "epoch": 0.9999230828397816,
2235
+ "grad_norm": 0.110712930560112,
2236
+ "learning_rate": 0.0001351115497493702,
2237
+ "loss": -30.4689,
2238
+ "step": 13000
2239
+ },
2240
+ {
2241
+ "epoch": 0.9999230828397816,
2242
+ "eval_loss": -30.469120025634766,
2243
+ "eval_runtime": 16.2879,
2244
+ "eval_samples_per_second": 61.395,
2245
+ "eval_steps_per_second": 15.349,
2246
+ "step": 13000
2247
  }
2248
  ],
2249
  "logging_steps": 50,