rootxhacker commited on
Commit
d1fe8d0
·
verified ·
1 Parent(s): 15d568e

Training in progress, step 8500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8cf14ec041a3170fc93bd13962e58c3433da275c26e3ddcd248f44fe9261656
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64d306725c80312e5129c700db5f1b446afe8e7ae7dd6316ba0e260f92b175cc
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5afd0a70049fd01ffc5b0b2af6dbbfcaccc76963457c541c72cb41caeec8cf4
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aded19b16b58d2665657ed9436c8f8202f2f226989fa9f44064f3004c4c9f352
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1383466fc8ec131ed5aea99415fe444f7a641621908ee8efda1f95cd39e73d3d
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef627c205c49617bbd8eff17e12d329ee9da18b36aa3d369121d14fed0afb2b
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:565e9a1f20b4e27434c54601d243feb5ec67c27b2d167d2e9e70202e4cc1bb10
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca8ab81e621162f8a9202e1abd3bdf1fc40e3d6f45f592533b7f8b7c3cb0c2df
3
  size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4875a6ebf2ebe9f7eec596109836a4abaa1f225d707484c4b3129f5845c484e
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54672d09f1aec146185491fd461111d38d14cab878b7fb3a8d83bbaa34824c7e
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:763a8e91c3a700b449d0501649e8a0112790ca7b7dbb67a7e5e5dcfd39db55d8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d578571b1ae2f1a6a8668f7ff332c5420175cf78129d97a69b053a775064e46
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 8000,
3
- "best_metric": 1.410463809967041,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-8000",
5
- "epoch": 0.6153372817475579,
6
  "eval_steps": 250,
7
- "global_step": 8000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1384,6 +1384,92 @@
1384
  "eval_samples_per_second": 59.34,
1385
  "eval_steps_per_second": 14.835,
1386
  "step": 8000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1387
  }
1388
  ],
1389
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 8500,
3
+ "best_metric": 1.4011365175247192,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-8500",
5
+ "epoch": 0.6537958618567803,
6
  "eval_steps": 250,
7
+ "global_step": 8500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1384
  "eval_samples_per_second": 59.34,
1385
  "eval_steps_per_second": 14.835,
1386
  "step": 8000
1387
+ },
1388
+ {
1389
+ "epoch": 0.6191831397584802,
1390
+ "grad_norm": 1.6925851106643677,
1391
+ "learning_rate": 0.0001608134431083292,
1392
+ "loss": 1.3827,
1393
+ "step": 8050
1394
+ },
1395
+ {
1396
+ "epoch": 0.6230289977694023,
1397
+ "grad_norm": 1.419701337814331,
1398
+ "learning_rate": 0.00016055372308651275,
1399
+ "loss": 1.4598,
1400
+ "step": 8100
1401
+ },
1402
+ {
1403
+ "epoch": 0.6268748557803245,
1404
+ "grad_norm": 1.6113284826278687,
1405
+ "learning_rate": 0.00016029400306469626,
1406
+ "loss": 1.4369,
1407
+ "step": 8150
1408
+ },
1409
+ {
1410
+ "epoch": 0.6307207137912468,
1411
+ "grad_norm": 1.8418887853622437,
1412
+ "learning_rate": 0.00016003428304287978,
1413
+ "loss": 1.405,
1414
+ "step": 8200
1415
+ },
1416
+ {
1417
+ "epoch": 0.634566571802169,
1418
+ "grad_norm": 0.9828066825866699,
1419
+ "learning_rate": 0.00015977456302106332,
1420
+ "loss": 1.3865,
1421
+ "step": 8250
1422
+ },
1423
+ {
1424
+ "epoch": 0.634566571802169,
1425
+ "eval_loss": 1.4110686779022217,
1426
+ "eval_runtime": 16.91,
1427
+ "eval_samples_per_second": 59.137,
1428
+ "eval_steps_per_second": 14.784,
1429
+ "step": 8250
1430
+ },
1431
+ {
1432
+ "epoch": 0.6384124298130913,
1433
+ "grad_norm": 1.8815706968307495,
1434
+ "learning_rate": 0.00015951484299924683,
1435
+ "loss": 1.3965,
1436
+ "step": 8300
1437
+ },
1438
+ {
1439
+ "epoch": 0.6422582878240135,
1440
+ "grad_norm": 0.975824773311615,
1441
+ "learning_rate": 0.00015925512297743034,
1442
+ "loss": 1.4273,
1443
+ "step": 8350
1444
+ },
1445
+ {
1446
+ "epoch": 0.6461041458349358,
1447
+ "grad_norm": 1.3958059549331665,
1448
+ "learning_rate": 0.00015899540295561385,
1449
+ "loss": 1.4204,
1450
+ "step": 8400
1451
+ },
1452
+ {
1453
+ "epoch": 0.649950003845858,
1454
+ "grad_norm": 1.059217929840088,
1455
+ "learning_rate": 0.00015873568293379737,
1456
+ "loss": 1.4439,
1457
+ "step": 8450
1458
+ },
1459
+ {
1460
+ "epoch": 0.6537958618567803,
1461
+ "grad_norm": 1.491314172744751,
1462
+ "learning_rate": 0.0001584759629119809,
1463
+ "loss": 1.5128,
1464
+ "step": 8500
1465
+ },
1466
+ {
1467
+ "epoch": 0.6537958618567803,
1468
+ "eval_loss": 1.4011365175247192,
1469
+ "eval_runtime": 16.9623,
1470
+ "eval_samples_per_second": 58.954,
1471
+ "eval_steps_per_second": 14.739,
1472
+ "step": 8500
1473
  }
1474
  ],
1475
  "logging_steps": 50,