rootxhacker commited on
Commit
0947d48
·
verified ·
1 Parent(s): 911760f

Training in progress, step 9500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c025d03493052da11f2e276acb3d21a6d929e8421e1c7592ebeacd87d114cd85
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46059868a56444f774675df268eb914452898f29b7df82787796af4c7fa5d48d
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed85d0da178bc8a897d04aed9fccd3af1551001178952fa7b7d9b07378424006
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d22244b9bcee513c0db5fa0dd4382afaa1310731018b62ed1ec48732dc32403d
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79658bc05ff8e823814b286e695d1366121826201ccee57ea590843dcddf9763
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24d290d02846044f2e528ff22c1dc13c928158f90f0b3902a3633693e875f22d
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9fe45a2e23b5067bfb664c7c87d67578826fdd3c8538fe28491021efb33c2e2
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:330590f2dd931f25cb518a6afcc3373fe0b53fe3178be67482b5a6355a86a689
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ed1ff55cffa4c7997d35d9e9aec781ef5a02c8194294d8fa15d89ecfc4185f6
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90087180835653f1b7a3f029c8126085005e238a2fd17bd1f7457696339c6b13
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ea10c4ee31d77ef3033b6c398543d815e76e7ca17a3e28c3601be5a6eebbb15
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:121ffc46f9f634ab597be8168f47dd7955b764f9421f382e55c4e5758e2a1624
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 9000,
3
- "best_metric": 1.902275800704956,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-9000",
5
- "epoch": 0.6922544419660026,
6
  "eval_steps": 250,
7
- "global_step": 9000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1556,6 +1556,92 @@
1556
  "eval_samples_per_second": 58.972,
1557
  "eval_steps_per_second": 14.743,
1558
  "step": 9000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1559
  }
1560
  ],
1561
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 9500,
3
+ "best_metric": 1.8725571632385254,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-9500",
5
+ "epoch": 0.7307130220752249,
6
  "eval_steps": 250,
7
+ "global_step": 9500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1556
  "eval_samples_per_second": 58.972,
1557
  "eval_steps_per_second": 14.743,
1558
  "step": 9000
1559
+ },
1560
+ {
1561
+ "epoch": 0.6961002999769248,
1562
+ "grad_norm": 3.6892011165618896,
1563
+ "learning_rate": 0.00015562423707243591,
1564
+ "loss": 1.9551,
1565
+ "step": 9050
1566
+ },
1567
+ {
1568
+ "epoch": 0.6999461579878471,
1569
+ "grad_norm": 1.524671196937561,
1570
+ "learning_rate": 0.00015536451705061943,
1571
+ "loss": 1.9109,
1572
+ "step": 9100
1573
+ },
1574
+ {
1575
+ "epoch": 0.7037920159987693,
1576
+ "grad_norm": 1.5293575525283813,
1577
+ "learning_rate": 0.00015510479702880297,
1578
+ "loss": 1.8891,
1579
+ "step": 9150
1580
+ },
1581
+ {
1582
+ "epoch": 0.7076378740096916,
1583
+ "grad_norm": 2.321476697921753,
1584
+ "learning_rate": 0.00015484507700698648,
1585
+ "loss": 1.9021,
1586
+ "step": 9200
1587
+ },
1588
+ {
1589
+ "epoch": 0.7114837320206138,
1590
+ "grad_norm": 2.710942029953003,
1591
+ "learning_rate": 0.00015458535698517,
1592
+ "loss": 1.9414,
1593
+ "step": 9250
1594
+ },
1595
+ {
1596
+ "epoch": 0.7114837320206138,
1597
+ "eval_loss": 1.9107917547225952,
1598
+ "eval_runtime": 16.997,
1599
+ "eval_samples_per_second": 58.834,
1600
+ "eval_steps_per_second": 14.708,
1601
+ "step": 9250
1602
+ },
1603
+ {
1604
+ "epoch": 0.7153295900315361,
1605
+ "grad_norm": 1.9385954141616821,
1606
+ "learning_rate": 0.0001543256369633535,
1607
+ "loss": 1.9401,
1608
+ "step": 9300
1609
+ },
1610
+ {
1611
+ "epoch": 0.7191754480424583,
1612
+ "grad_norm": 2.589629650115967,
1613
+ "learning_rate": 0.00015406591694153702,
1614
+ "loss": 1.9054,
1615
+ "step": 9350
1616
+ },
1617
+ {
1618
+ "epoch": 0.7230213060533806,
1619
+ "grad_norm": 1.6431207656860352,
1620
+ "learning_rate": 0.00015380619691972056,
1621
+ "loss": 1.9324,
1622
+ "step": 9400
1623
+ },
1624
+ {
1625
+ "epoch": 0.7268671640643027,
1626
+ "grad_norm": 4.840892314910889,
1627
+ "learning_rate": 0.00015354647689790407,
1628
+ "loss": 1.993,
1629
+ "step": 9450
1630
+ },
1631
+ {
1632
+ "epoch": 0.7307130220752249,
1633
+ "grad_norm": 0.9328492879867554,
1634
+ "learning_rate": 0.00015328675687608758,
1635
+ "loss": 1.8637,
1636
+ "step": 9500
1637
+ },
1638
+ {
1639
+ "epoch": 0.7307130220752249,
1640
+ "eval_loss": 1.8725571632385254,
1641
+ "eval_runtime": 16.9918,
1642
+ "eval_samples_per_second": 58.852,
1643
+ "eval_steps_per_second": 14.713,
1644
+ "step": 9500
1645
  }
1646
  ],
1647
  "logging_steps": 50,