rootxhacker commited on
Commit
fdda422
·
verified ·
1 Parent(s): 3f3377a

Training in progress, step 3500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f77dbb7c6bf6487a48a714e4c4e440e388d69a8afa7310e97fc320615f989a6
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24082cd1025a9c16d6bc6449cbc8ef1f1d242da73316dafd4f2689fc8c860917
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1814592ba74cdcca20fb08c3ea7e5ee573986aff77c3810b20a982bd1be5d78
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4d61ee4cf28e305ccbd9c0cd63b04a590ac2c163f4c12096fd42f05674ed054
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cdad770e6415de25f2503b80a04125ce36c191e7ebbe40cc70935c15b3079b5
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32c4abfc978e57dde0e9122ce6e0a211a14cdb3e9d7e1f8e79d82dbf8734a5a1
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9aa9825ab5c6a64b78405ab88bd271bc07d34d8fb6641eef252a11c55bfe386
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:835fd47c5d5c5f39bf487b0cc415ae7b687f6d1d1601a7df5275be58f8fd603c
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0276c75f304794326dfbdfec2d07b2898318c16d7455e1133773ea583662ada9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b0c041eadc91e24d174b435fcb1ec46a0c56f36b09b7dbbf49009145cd7fa33
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 3000,
3
- "best_metric": 4.55267858505249,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-3000",
5
- "epoch": 0.2307514806553342,
6
  "eval_steps": 250,
7
- "global_step": 3000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -524,6 +524,92 @@
524
  "eval_samples_per_second": 52.705,
525
  "eval_steps_per_second": 13.176,
526
  "step": 3000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
527
  }
528
  ],
529
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 3500,
3
+ "best_metric": 4.488556861877441,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-3500",
5
+ "epoch": 0.26921006076455656,
6
  "eval_steps": 250,
7
+ "global_step": 3500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
524
  "eval_samples_per_second": 52.705,
525
  "eval_steps_per_second": 13.176,
526
  "step": 3000
527
+ },
528
+ {
529
+ "epoch": 0.23459733866625646,
530
+ "grad_norm": 4.74845027923584,
531
+ "learning_rate": 0.00012484761589647437,
532
+ "loss": 4.6248,
533
+ "step": 3050
534
+ },
535
+ {
536
+ "epoch": 0.23844319667717867,
537
+ "grad_norm": 6.299524784088135,
538
+ "learning_rate": 0.00012459027241645528,
539
+ "loss": 4.5457,
540
+ "step": 3100
541
+ },
542
+ {
543
+ "epoch": 0.2422890546881009,
544
+ "grad_norm": 5.853606700897217,
545
+ "learning_rate": 0.00012433292893643616,
546
+ "loss": 4.5135,
547
+ "step": 3150
548
+ },
549
+ {
550
+ "epoch": 0.24613491269902316,
551
+ "grad_norm": 3.1514365673065186,
552
+ "learning_rate": 0.00012407558545641704,
553
+ "loss": 4.672,
554
+ "step": 3200
555
+ },
556
+ {
557
+ "epoch": 0.2499807707099454,
558
+ "grad_norm": 8.455827713012695,
559
+ "learning_rate": 0.00012381824197639792,
560
+ "loss": 4.6545,
561
+ "step": 3250
562
+ },
563
+ {
564
+ "epoch": 0.2499807707099454,
565
+ "eval_loss": 4.550297737121582,
566
+ "eval_runtime": 18.9801,
567
+ "eval_samples_per_second": 52.687,
568
+ "eval_steps_per_second": 13.172,
569
+ "step": 3250
570
+ },
571
+ {
572
+ "epoch": 0.25382662872086764,
573
+ "grad_norm": 2.8094310760498047,
574
+ "learning_rate": 0.0001235608984963788,
575
+ "loss": 4.5392,
576
+ "step": 3300
577
+ },
578
+ {
579
+ "epoch": 0.2576724867317899,
580
+ "grad_norm": 3.2565436363220215,
581
+ "learning_rate": 0.00012330355501635968,
582
+ "loss": 4.481,
583
+ "step": 3350
584
+ },
585
+ {
586
+ "epoch": 0.26151834474271207,
587
+ "grad_norm": 3.5588488578796387,
588
+ "learning_rate": 0.00012304621153634056,
589
+ "loss": 4.5543,
590
+ "step": 3400
591
+ },
592
+ {
593
+ "epoch": 0.2653642027536343,
594
+ "grad_norm": 3.0696310997009277,
595
+ "learning_rate": 0.00012278886805632146,
596
+ "loss": 4.5858,
597
+ "step": 3450
598
+ },
599
+ {
600
+ "epoch": 0.26921006076455656,
601
+ "grad_norm": 3.886117935180664,
602
+ "learning_rate": 0.00012253152457630234,
603
+ "loss": 4.4694,
604
+ "step": 3500
605
+ },
606
+ {
607
+ "epoch": 0.26921006076455656,
608
+ "eval_loss": 4.488556861877441,
609
+ "eval_runtime": 18.9212,
610
+ "eval_samples_per_second": 52.851,
611
+ "eval_steps_per_second": 13.213,
612
+ "step": 3500
613
  }
614
  ],
615
  "logging_steps": 50,