rootxhacker commited on
Commit
a947326
·
verified ·
1 Parent(s): a490837

Training in progress, step 38500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc9e082dc3b2b599ece245c68d836aa3280cd4e7a9b0918ad403250af895d5e3
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9fccd6e8cf77b198a4f4b87f312750517824202a02166eb04e56577ecd2ba00
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c350b485739b4fddb9d5148423e5dfbced50fc251b92973c875bcddb4d66a775
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66d49ec8e5680b4ea0cac8d1930f19a010f9ae0bd4d5ca4b158ebc77b3d54acd
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54a8acb2b80e4297ad7cc3210bf7b7be72618237e5040794a64b58c69faec7fc
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38f23e9c9a2a443780a64b0cf18b93c2bd78a442ad08f1e37293ea92f7c88993
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2c0cded846d640f3719120429980641a6206dc92844fb42e8f946ff3bea6b2e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1b4df89a65698a9335c43463fa8aa3cc49c1acdfae5f7a285af000ba78e95b1
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fc19191502cd55e651632979b36f722e8851e564a429a904a6573212b31486c
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a1d56165e0890bb9432c217b765bdd79a8b4d93dcde80b4671874513824409b
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:541cb6dea9c80e2359e5134078f161ec83a5de3af8e29b3e9b6b880e5dd9058f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3dea87c40748698e240c93bc3eea5482b3f627ab3a1ed391f6cbe2fe5d7da89
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 37750,
3
- "best_metric": -30.47093391418457,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
5
- "epoch": 2.9228520883008997,
6
  "eval_steps": 250,
7
- "global_step": 38000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -6544,6 +6544,92 @@
6544
  "eval_samples_per_second": 59.431,
6545
  "eval_steps_per_second": 14.858,
6546
  "step": 38000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6547
  }
6548
  ],
6549
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 38250,
3
+ "best_metric": -30.471132278442383,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
5
+ "epoch": 2.961310668410122,
6
  "eval_steps": 250,
7
+ "global_step": 38500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
6544
  "eval_samples_per_second": 59.431,
6545
  "eval_steps_per_second": 14.858,
6546
  "step": 38000
6547
+ },
6548
+ {
6549
+ "epoch": 2.926697946311822,
6550
+ "grad_norm": 0.9110496044158936,
6551
+ "learning_rate": 5.033374022803418e-06,
6552
+ "loss": -30.4668,
6553
+ "step": 38050
6554
+ },
6555
+ {
6556
+ "epoch": 2.9305438043227445,
6557
+ "grad_norm": 0.7347224950790405,
6558
+ "learning_rate": 4.773654000986936e-06,
6559
+ "loss": -30.4668,
6560
+ "step": 38100
6561
+ },
6562
+ {
6563
+ "epoch": 2.9343896623336665,
6564
+ "grad_norm": 0.946358859539032,
6565
+ "learning_rate": 4.513933979170454e-06,
6566
+ "loss": -30.4669,
6567
+ "step": 38150
6568
+ },
6569
+ {
6570
+ "epoch": 2.938235520344589,
6571
+ "grad_norm": 0.5716750621795654,
6572
+ "learning_rate": 4.254213957353972e-06,
6573
+ "loss": -30.4669,
6574
+ "step": 38200
6575
+ },
6576
+ {
6577
+ "epoch": 2.9420813783555113,
6578
+ "grad_norm": 1.8405364751815796,
6579
+ "learning_rate": 3.994493935537491e-06,
6580
+ "loss": -30.4667,
6581
+ "step": 38250
6582
+ },
6583
+ {
6584
+ "epoch": 2.9420813783555113,
6585
+ "eval_loss": -30.471132278442383,
6586
+ "eval_runtime": 16.8489,
6587
+ "eval_samples_per_second": 59.351,
6588
+ "eval_steps_per_second": 14.838,
6589
+ "step": 38250
6590
+ },
6591
+ {
6592
+ "epoch": 2.9459272363664333,
6593
+ "grad_norm": 0.6454173922538757,
6594
+ "learning_rate": 3.734773913721009e-06,
6595
+ "loss": -30.467,
6596
+ "step": 38300
6597
+ },
6598
+ {
6599
+ "epoch": 2.9497730943773557,
6600
+ "grad_norm": 0.7232244610786438,
6601
+ "learning_rate": 3.475053891904527e-06,
6602
+ "loss": -30.4668,
6603
+ "step": 38350
6604
+ },
6605
+ {
6606
+ "epoch": 2.953618952388278,
6607
+ "grad_norm": 1.0730299949645996,
6608
+ "learning_rate": 3.215333870088045e-06,
6609
+ "loss": -30.4668,
6610
+ "step": 38400
6611
+ },
6612
+ {
6613
+ "epoch": 2.9574648103992,
6614
+ "grad_norm": 0.4886613190174103,
6615
+ "learning_rate": 2.9556138482715634e-06,
6616
+ "loss": -30.4669,
6617
+ "step": 38450
6618
+ },
6619
+ {
6620
+ "epoch": 2.961310668410122,
6621
+ "grad_norm": 0.5141581296920776,
6622
+ "learning_rate": 2.6958938264550813e-06,
6623
+ "loss": -30.4669,
6624
+ "step": 38500
6625
+ },
6626
+ {
6627
+ "epoch": 2.961310668410122,
6628
+ "eval_loss": -30.471107482910156,
6629
+ "eval_runtime": 16.867,
6630
+ "eval_samples_per_second": 59.287,
6631
+ "eval_steps_per_second": 14.822,
6632
+ "step": 38500
6633
  }
6634
  ],
6635
  "logging_steps": 50,