rootxhacker commited on
Commit
d037f24
·
verified ·
1 Parent(s): 4cdc045

Training in progress, step 9500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d41b0a6b7b93326606bd1adb6fe554c4f3a4896093e8d2cda0d85615659fbea9
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b9baf2c6ba5ebbe672256c852fecb5f216fb3f2e6bf3f464d3b0a62f927a601
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61e6388870be8d5091de8ee4171a40d8d34fd90416acec89a4530ab1810d1d11
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d40db6775badf11b35e39be49c6fef0bc883f3580d3d852ee11aee4261d7eedd
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8114c7619958e289ae818af20bbbb40ca496ec99d1d2cf5336f332be768bd676
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71923ca4e5165f8d69e24064ca7c57c4feab7a075d2d60f03dd76eb79c8d3ad8
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f82ab7544ab0f2c7299f888437f6d772f90becf0f776876608740887a96023f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:632e9f7e063df82f0a0f7f24d19a972e733d8fb69f46a15890d826ee60504fae
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c86e9949dede8e9039d4479cc9a7b6de2de62acc5bc9bce167cc78f0df5d789
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:599596a3cb79dad5f0353c782d77b8e23494fd9a0c28c4b3bdb9f32604aafb58
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:67134d509d59e4154961ca89be855a6a8dd7ecee21023e7a214d3f3e40df8ac6
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37b10e7fbb18174f6f934d4c5e8eb4aeea59a6280951bc7b58e46146366fc7ea
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 8750,
3
- "best_metric": 1.3920339345932007,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-8500",
5
- "epoch": 0.6922544419660026,
6
  "eval_steps": 250,
7
- "global_step": 9000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1556,6 +1556,92 @@
1556
  "eval_samples_per_second": 59.652,
1557
  "eval_steps_per_second": 14.913,
1558
  "step": 9000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1559
  }
1560
  ],
1561
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 9500,
3
+ "best_metric": 1.382208228111267,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-9500",
5
+ "epoch": 0.7307130220752249,
6
  "eval_steps": 250,
7
+ "global_step": 9500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1556
  "eval_samples_per_second": 59.652,
1557
  "eval_steps_per_second": 14.913,
1558
  "step": 9000
1559
+ },
1560
+ {
1561
+ "epoch": 0.6961002999769248,
1562
+ "grad_norm": 2.0058114528656006,
1563
+ "learning_rate": 0.00015561904267199958,
1564
+ "loss": 1.4288,
1565
+ "step": 9050
1566
+ },
1567
+ {
1568
+ "epoch": 0.6999461579878471,
1569
+ "grad_norm": 0.9105901122093201,
1570
+ "learning_rate": 0.00015535932265018312,
1571
+ "loss": 1.4271,
1572
+ "step": 9100
1573
+ },
1574
+ {
1575
+ "epoch": 0.7037920159987693,
1576
+ "grad_norm": 1.241396188735962,
1577
+ "learning_rate": 0.00015509960262836663,
1578
+ "loss": 1.3543,
1579
+ "step": 9150
1580
+ },
1581
+ {
1582
+ "epoch": 0.7076378740096916,
1583
+ "grad_norm": 1.2965749502182007,
1584
+ "learning_rate": 0.00015483988260655017,
1585
+ "loss": 1.3501,
1586
+ "step": 9200
1587
+ },
1588
+ {
1589
+ "epoch": 0.7114837320206138,
1590
+ "grad_norm": 1.4165613651275635,
1591
+ "learning_rate": 0.00015458016258473365,
1592
+ "loss": 1.4465,
1593
+ "step": 9250
1594
+ },
1595
+ {
1596
+ "epoch": 0.7114837320206138,
1597
+ "eval_loss": 1.3871207237243652,
1598
+ "eval_runtime": 16.7909,
1599
+ "eval_samples_per_second": 59.556,
1600
+ "eval_steps_per_second": 14.889,
1601
+ "step": 9250
1602
+ },
1603
+ {
1604
+ "epoch": 0.7153295900315361,
1605
+ "grad_norm": 1.7752403020858765,
1606
+ "learning_rate": 0.00015432044256291717,
1607
+ "loss": 1.4282,
1608
+ "step": 9300
1609
+ },
1610
+ {
1611
+ "epoch": 0.7191754480424583,
1612
+ "grad_norm": 1.6717677116394043,
1613
+ "learning_rate": 0.0001540607225411007,
1614
+ "loss": 1.3986,
1615
+ "step": 9350
1616
+ },
1617
+ {
1618
+ "epoch": 0.7230213060533806,
1619
+ "grad_norm": 1.290183186531067,
1620
+ "learning_rate": 0.00015380100251928422,
1621
+ "loss": 1.3891,
1622
+ "step": 9400
1623
+ },
1624
+ {
1625
+ "epoch": 0.7268671640643027,
1626
+ "grad_norm": 1.6836214065551758,
1627
+ "learning_rate": 0.00015354128249746773,
1628
+ "loss": 1.4028,
1629
+ "step": 9450
1630
+ },
1631
+ {
1632
+ "epoch": 0.7307130220752249,
1633
+ "grad_norm": 0.5705507397651672,
1634
+ "learning_rate": 0.00015328156247565127,
1635
+ "loss": 1.365,
1636
+ "step": 9500
1637
+ },
1638
+ {
1639
+ "epoch": 0.7307130220752249,
1640
+ "eval_loss": 1.382208228111267,
1641
+ "eval_runtime": 16.8905,
1642
+ "eval_samples_per_second": 59.205,
1643
+ "eval_steps_per_second": 14.801,
1644
+ "step": 9500
1645
  }
1646
  ],
1647
  "logging_steps": 50,