rootxhacker commited on
Commit
a34fe6b
·
verified ·
1 Parent(s): 1876b73

Training in progress, step 13000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4563a36b3299867126e2f2872e4b74a5609a7c6cb213d2085d095900f9215fa3
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2d7fcaf2efe42ffb5f3c9e028de2d4020155b47fd00969c045a7b96f9a500e0
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9ead48f8f886b45b9f89bef7b002712a2b38bb6cc69a0971816a685b1318dd1
3
  size 1544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:828b8db8ad6923c1cf715c1424f34fe8d45d1942c07709d2ed3e3b91c7b41825
3
  size 1544
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea5c1fbeda3adc57688fb9a2ac22bbf43523ff5c2f34a45a7dd94247c865d2ac
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81731a2aece3051885d7289cef2d696a33b091ba71fd2ad97dba80506614db51
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a5ae118ec673d57a7f62a6284956ef83acf5113a9420dcd7ba5fd941799b59c
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbf8948534182fde8a4da31776a53cc78337e1e18ed4b526e2ee8ca99eff5731
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe8cf9d8bac2dd71fd820efd0a6b09790c369caf5e00f7d2853b196fd0826191
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1a2f8fe642f01391ef2e54fcf96a41d1f8399b7293ea1f181058d0539b5b33d
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fab19116031cb09fcbccb12fc7d5785df4d500dae544895e7f9a43d44ebf2019
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69b23dccdca16565cb13b99c2e00b6827579e6c2cc8759c8f2106a0579d096c8
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 12250,
3
- "best_metric": 1.5220181941986084,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-12000",
5
- "epoch": 0.9614645027305592,
6
  "eval_steps": 250,
7
- "global_step": 12500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2158,6 +2158,92 @@
2158
  "eval_samples_per_second": 56.369,
2159
  "eval_steps_per_second": 14.092,
2160
  "step": 12500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2161
  }
2162
  ],
2163
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 13000,
3
+ "best_metric": 1.5119102001190186,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-13000",
5
+ "epoch": 0.9999230828397816,
6
  "eval_steps": 250,
7
+ "global_step": 13000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2158
  "eval_samples_per_second": 56.369,
2159
  "eval_steps_per_second": 14.092,
2160
  "step": 12500
2161
+ },
2162
+ {
2163
+ "epoch": 0.9653103607414815,
2164
+ "grad_norm": 1.888836145401001,
2165
+ "learning_rate": 0.00013745941874659118,
2166
+ "loss": 1.5681,
2167
+ "step": 12550
2168
+ },
2169
+ {
2170
+ "epoch": 0.9691562187524037,
2171
+ "grad_norm": 1.4329860210418701,
2172
+ "learning_rate": 0.0001371996987247747,
2173
+ "loss": 1.5018,
2174
+ "step": 12600
2175
+ },
2176
+ {
2177
+ "epoch": 0.973002076763326,
2178
+ "grad_norm": 1.969533920288086,
2179
+ "learning_rate": 0.00013693997870295823,
2180
+ "loss": 1.5494,
2181
+ "step": 12650
2182
+ },
2183
+ {
2184
+ "epoch": 0.9768479347742481,
2185
+ "grad_norm": 2.1219890117645264,
2186
+ "learning_rate": 0.00013668025868114175,
2187
+ "loss": 1.5386,
2188
+ "step": 12700
2189
+ },
2190
+ {
2191
+ "epoch": 0.9806937927851703,
2192
+ "grad_norm": 1.6632941961288452,
2193
+ "learning_rate": 0.00013642053865932526,
2194
+ "loss": 1.5253,
2195
+ "step": 12750
2196
+ },
2197
+ {
2198
+ "epoch": 0.9806937927851703,
2199
+ "eval_loss": 1.5147372484207153,
2200
+ "eval_runtime": 17.8072,
2201
+ "eval_samples_per_second": 56.157,
2202
+ "eval_steps_per_second": 14.039,
2203
+ "step": 12750
2204
+ },
2205
+ {
2206
+ "epoch": 0.9845396507960926,
2207
+ "grad_norm": 1.292913794517517,
2208
+ "learning_rate": 0.00013616081863750877,
2209
+ "loss": 1.4413,
2210
+ "step": 12800
2211
+ },
2212
+ {
2213
+ "epoch": 0.9883855088070148,
2214
+ "grad_norm": 1.1377824544906616,
2215
+ "learning_rate": 0.00013590109861569228,
2216
+ "loss": 1.5705,
2217
+ "step": 12850
2218
+ },
2219
+ {
2220
+ "epoch": 0.9922313668179371,
2221
+ "grad_norm": 1.656996726989746,
2222
+ "learning_rate": 0.0001356413785938758,
2223
+ "loss": 1.5311,
2224
+ "step": 12900
2225
+ },
2226
+ {
2227
+ "epoch": 0.9960772248288593,
2228
+ "grad_norm": 1.6639357805252075,
2229
+ "learning_rate": 0.00013538165857205934,
2230
+ "loss": 1.5676,
2231
+ "step": 12950
2232
+ },
2233
+ {
2234
+ "epoch": 0.9999230828397816,
2235
+ "grad_norm": 1.0893466472625732,
2236
+ "learning_rate": 0.00013512193855024285,
2237
+ "loss": 1.5215,
2238
+ "step": 13000
2239
+ },
2240
+ {
2241
+ "epoch": 0.9999230828397816,
2242
+ "eval_loss": 1.5119102001190186,
2243
+ "eval_runtime": 17.7291,
2244
+ "eval_samples_per_second": 56.404,
2245
+ "eval_steps_per_second": 14.101,
2246
+ "step": 13000
2247
  }
2248
  ],
2249
  "logging_steps": 50,