rootxhacker commited on
Commit
ba7d108
·
verified ·
1 Parent(s): 859dc6a

Training in progress, step 2000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45253cd6a22264fcd7ca6604e87717a83e19ed0d2d4435aba88387d4537d6529
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37e7ac0622b34136a1adc5dda670dd8f06fe9699327bbef64e2b847e41d62d45
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b17ffeb44d19aff3ac9c819aa738ee5ab85c93083a940599d75b2a7ffda4f79
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d5aaec88b2f521c4d19280ee5a843dca115995f362a0926fa7ff5637694cc6d
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ad00466a0acdb40237dae4c0b4283060a706960c994c5977eb55d8a845c8d60
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc0f99b2f4c37e6bd3da61564f89d37dedf31c888089686d20000b137615abc8
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97a66f532d27e837c455454684371b16caffae096e4b153335efaed54e929716
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4da9c3920b2d28ca98d344913f643342e405b080bfc26edfe880e644c5be7370
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8167a7a628c5d6d5061531d993a99d8970329d2f198daf9e2bd5698793632aed
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43c2f44953ece167e5bd34ec80ca4bdfcda9e945ce63675d98939409afdb1cb3
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8648ee281a6907a159c6d2a012bfd17e90caa435b3ae05a6accbf2bacc5a6ed9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:951a13f6b06bce52a2ba9f7d2d8f60254ac5cc4c021e67355429cc5c41426dfb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 1500,
3
- "best_metric": 1.6279098987579346,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-1500",
5
- "epoch": 0.1153757403276671,
6
  "eval_steps": 250,
7
- "global_step": 1500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -266,6 +266,92 @@
266
  "eval_samples_per_second": 58.859,
267
  "eval_steps_per_second": 14.715,
268
  "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
269
  }
270
  ],
271
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 2000,
3
+ "best_metric": 1.60262131690979,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-2000",
5
+ "epoch": 0.15383432043688947,
6
  "eval_steps": 250,
7
+ "global_step": 2000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
266
  "eval_samples_per_second": 58.859,
267
  "eval_steps_per_second": 14.715,
268
  "step": 1500
269
+ },
270
+ {
271
+ "epoch": 0.11922159833858934,
272
+ "grad_norm": 1.4586989879608154,
273
+ "learning_rate": 0.00019457185154403555,
274
+ "loss": 1.6701,
275
+ "step": 1550
276
+ },
277
+ {
278
+ "epoch": 0.12306745634951158,
279
+ "grad_norm": 2.564028024673462,
280
+ "learning_rate": 0.00019431213152221907,
281
+ "loss": 1.6413,
282
+ "step": 1600
283
+ },
284
+ {
285
+ "epoch": 0.12691331436043382,
286
+ "grad_norm": 1.8470406532287598,
287
+ "learning_rate": 0.00019405241150040255,
288
+ "loss": 1.6075,
289
+ "step": 1650
290
+ },
291
+ {
292
+ "epoch": 0.13075917237135604,
293
+ "grad_norm": 2.475013017654419,
294
+ "learning_rate": 0.0001937926914785861,
295
+ "loss": 1.6882,
296
+ "step": 1700
297
+ },
298
+ {
299
+ "epoch": 0.13460503038227828,
300
+ "grad_norm": 1.0003979206085205,
301
+ "learning_rate": 0.0001935329714567696,
302
+ "loss": 1.6552,
303
+ "step": 1750
304
+ },
305
+ {
306
+ "epoch": 0.13460503038227828,
307
+ "eval_loss": 1.610037088394165,
308
+ "eval_runtime": 16.9155,
309
+ "eval_samples_per_second": 59.117,
310
+ "eval_steps_per_second": 14.779,
311
+ "step": 1750
312
+ },
313
+ {
314
+ "epoch": 0.13845088839320052,
315
+ "grad_norm": 0.8518445491790771,
316
+ "learning_rate": 0.00019327325143495314,
317
+ "loss": 1.6631,
318
+ "step": 1800
319
+ },
320
+ {
321
+ "epoch": 0.14229674640412276,
322
+ "grad_norm": 1.305446743965149,
323
+ "learning_rate": 0.00019301353141313666,
324
+ "loss": 1.5437,
325
+ "step": 1850
326
+ },
327
+ {
328
+ "epoch": 0.146142604415045,
329
+ "grad_norm": 1.6358033418655396,
330
+ "learning_rate": 0.00019275381139132017,
331
+ "loss": 1.6463,
332
+ "step": 1900
333
+ },
334
+ {
335
+ "epoch": 0.14998846242596722,
336
+ "grad_norm": 1.0969074964523315,
337
+ "learning_rate": 0.00019249409136950368,
338
+ "loss": 1.5903,
339
+ "step": 1950
340
+ },
341
+ {
342
+ "epoch": 0.15383432043688947,
343
+ "grad_norm": 1.289174199104309,
344
+ "learning_rate": 0.0001922343713476872,
345
+ "loss": 1.6313,
346
+ "step": 2000
347
+ },
348
+ {
349
+ "epoch": 0.15383432043688947,
350
+ "eval_loss": 1.60262131690979,
351
+ "eval_runtime": 16.8495,
352
+ "eval_samples_per_second": 59.349,
353
+ "eval_steps_per_second": 14.837,
354
+ "step": 2000
355
  }
356
  ],
357
  "logging_steps": 50,