rootxhacker commited on
Commit
5931e3e
·
verified ·
1 Parent(s): 28a740f

Training in progress, step 3500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84adf5fb4a3947b35993a790cf4973ef5593a87e6e164d9578f03b8b9c27e099
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11a062f0d5247b9505627696a295b4feca04c6c45dd688fcbc2c07d7828e414c
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:308f424d188d86d1fcd72962eb2fa00bace74a0cc7f5cd9ae44b93fa7b48ebe5
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ce273710431703c604c9d640c23434ce3f2b038664d7b3df1e4a13e933764d0
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65a04555108b26fbf193a851c8c842cf49a4c844bc1961b985cfc83f43bcde61
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb347b9f0c8f6a0d9a0e6b08949ae99d2459f1f662e3a7dd16f39b8ba0ddf69c
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:238412d43efe8b606030d3301ede6818fb1a62974cd0cbb655e125124222d862
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cf21f101c997b0643b73ca8f669be23f0abb9e5a2c6fedb07e17298d9cdd268
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04b1366b3719985cdc1dc0eb0fa220f7b35ea714d4ff368885830073c629f4d9
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:028b7d42ce0dbafb2a0126ad830dd1957166a4ca85043c6644487cae8315bfe9
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2976163d4a5cd674bbf6d82ca442cb1ceba9d5f9d444c557ab914dab5858c6ce
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d16b13e1d35412f9085034fa2481674486c7b966c61cae81848c36587b827f3d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 3000,
3
- "best_metric": 1.5423797369003296,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-3000",
5
- "epoch": 0.2307514806553342,
6
  "eval_steps": 250,
7
- "global_step": 3000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -524,6 +524,92 @@
524
  "eval_samples_per_second": 59.161,
525
  "eval_steps_per_second": 14.79,
526
  "step": 3000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
527
  }
528
  ],
529
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 3500,
3
+ "best_metric": 1.5354665517807007,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-3500",
5
+ "epoch": 0.26921006076455656,
6
  "eval_steps": 250,
7
+ "global_step": 3500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
524
  "eval_samples_per_second": 59.161,
525
  "eval_steps_per_second": 14.79,
526
  "step": 3000
527
+ },
528
+ {
529
+ "epoch": 0.23459733866625646,
530
+ "grad_norm": 3.606706142425537,
531
+ "learning_rate": 0.00018678025088954107,
532
+ "loss": 1.5998,
533
+ "step": 3050
534
+ },
535
+ {
536
+ "epoch": 0.23844319667717867,
537
+ "grad_norm": 1.2053996324539185,
538
+ "learning_rate": 0.0001865205308677246,
539
+ "loss": 1.5512,
540
+ "step": 3100
541
+ },
542
+ {
543
+ "epoch": 0.2422890546881009,
544
+ "grad_norm": 1.5675193071365356,
545
+ "learning_rate": 0.00018626081084590813,
546
+ "loss": 1.5495,
547
+ "step": 3150
548
+ },
549
+ {
550
+ "epoch": 0.24613491269902316,
551
+ "grad_norm": 1.1827441453933716,
552
+ "learning_rate": 0.00018600109082409164,
553
+ "loss": 1.6509,
554
+ "step": 3200
555
+ },
556
+ {
557
+ "epoch": 0.2499807707099454,
558
+ "grad_norm": 1.7992998361587524,
559
+ "learning_rate": 0.00018574137080227515,
560
+ "loss": 1.5617,
561
+ "step": 3250
562
+ },
563
+ {
564
+ "epoch": 0.2499807707099454,
565
+ "eval_loss": 1.5403118133544922,
566
+ "eval_runtime": 16.9627,
567
+ "eval_samples_per_second": 58.953,
568
+ "eval_steps_per_second": 14.738,
569
+ "step": 3250
570
+ },
571
+ {
572
+ "epoch": 0.25382662872086764,
573
+ "grad_norm": 1.3044147491455078,
574
+ "learning_rate": 0.00018548165078045866,
575
+ "loss": 1.5226,
576
+ "step": 3300
577
+ },
578
+ {
579
+ "epoch": 0.2576724867317899,
580
+ "grad_norm": 1.4057557582855225,
581
+ "learning_rate": 0.0001852219307586422,
582
+ "loss": 1.517,
583
+ "step": 3350
584
+ },
585
+ {
586
+ "epoch": 0.26151834474271207,
587
+ "grad_norm": 0.9969176650047302,
588
+ "learning_rate": 0.00018496221073682572,
589
+ "loss": 1.5157,
590
+ "step": 3400
591
+ },
592
+ {
593
+ "epoch": 0.2653642027536343,
594
+ "grad_norm": 1.0183119773864746,
595
+ "learning_rate": 0.00018470249071500923,
596
+ "loss": 1.5012,
597
+ "step": 3450
598
+ },
599
+ {
600
+ "epoch": 0.26921006076455656,
601
+ "grad_norm": 1.1896519660949707,
602
+ "learning_rate": 0.00018444277069319277,
603
+ "loss": 1.4655,
604
+ "step": 3500
605
+ },
606
+ {
607
+ "epoch": 0.26921006076455656,
608
+ "eval_loss": 1.5354665517807007,
609
+ "eval_runtime": 16.8964,
610
+ "eval_samples_per_second": 59.184,
611
+ "eval_steps_per_second": 14.796,
612
+ "step": 3500
613
  }
614
  ],
615
  "logging_steps": 50,