rootxhacker commited on
Commit
cba18be
·
verified ·
1 Parent(s): 50d065b

Training in progress, step 15500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9310d7742dd19e663f6402741ea20db935d3b2e0e01a71d5a768ac7db25bc72
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9f1ed52cf80a5ab8d5ed12c9555137201579cc86f9a2f068ecfc94b91627d37
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c1a1611c8e46f685348b5dfe48ae762df7c9c3b98434f8da77fcc5f9f81a751
3
  size 1544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5311da3aa8361f2e32c9e1dadd3f99fb776577e3edbc936e8d5e4a247f82b48
3
  size 1544
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f525658a7b4e5ff255750bc03647522ca336d6f40ea836aaa028d9a5bebc0152
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ac8a732d897ca3b87406efba9af34062b6403d9fcd06247903138f1a20ce830
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c298ef220b2440e2c8688c2918aba70637e9e1c8a2951767772c4969343135c8
3
- size 14308
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e443e4b8c749db54f7826bd4573803e0f05009235443f881fa9932376e1aba8
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:453f146988a415e18f9bddb29836fc724515457272291d8aed123845d69e5c29
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce992341f57d419e4ec2c7637893018b15690261bc216bb5e0545131a4a5ab1d
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73ddc28e63ea8c6b2f70c524e6eba5501abe22b3d62b8aa4437458f65de50c75
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b94bf7f68d57ae943a99758bd0c736315f81c408fbb84a10b3ea5a5a1956ed5
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 13750,
3
- "best_metric": 1.5073590278625488,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-13000",
5
- "epoch": 1.153757403276671,
6
  "eval_steps": 250,
7
- "global_step": 15000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2588,6 +2588,92 @@
2588
  "eval_samples_per_second": 55.77,
2589
  "eval_steps_per_second": 13.943,
2590
  "step": 15000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2591
  }
2592
  ],
2593
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 15500,
3
+ "best_metric": 1.5013692378997803,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-15500",
5
+ "epoch": 1.1922159833858934,
6
  "eval_steps": 250,
7
+ "global_step": 15500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2588
  "eval_samples_per_second": 55.77,
2589
  "eval_steps_per_second": 13.943,
2590
  "step": 15000
2591
+ },
2592
+ {
2593
+ "epoch": 1.1576032612875933,
2594
+ "grad_norm": 1.5970553159713745,
2595
+ "learning_rate": 0.0001244734176557671,
2596
+ "loss": 1.4515,
2597
+ "step": 15050
2598
+ },
2599
+ {
2600
+ "epoch": 1.1614491192985155,
2601
+ "grad_norm": 1.0201988220214844,
2602
+ "learning_rate": 0.0001242136976339506,
2603
+ "loss": 1.5431,
2604
+ "step": 15100
2605
+ },
2606
+ {
2607
+ "epoch": 1.1652949773094377,
2608
+ "grad_norm": 0.9830596446990967,
2609
+ "learning_rate": 0.00012395397761213412,
2610
+ "loss": 1.4581,
2611
+ "step": 15150
2612
+ },
2613
+ {
2614
+ "epoch": 1.16914083532036,
2615
+ "grad_norm": 2.3363943099975586,
2616
+ "learning_rate": 0.00012369425759031766,
2617
+ "loss": 1.5147,
2618
+ "step": 15200
2619
+ },
2620
+ {
2621
+ "epoch": 1.1729866933312822,
2622
+ "grad_norm": 1.8213731050491333,
2623
+ "learning_rate": 0.00012343453756850117,
2624
+ "loss": 1.5497,
2625
+ "step": 15250
2626
+ },
2627
+ {
2628
+ "epoch": 1.1729866933312822,
2629
+ "eval_loss": 1.5053696632385254,
2630
+ "eval_runtime": 17.7358,
2631
+ "eval_samples_per_second": 56.383,
2632
+ "eval_steps_per_second": 14.096,
2633
+ "step": 15250
2634
+ },
2635
+ {
2636
+ "epoch": 1.1768325513422044,
2637
+ "grad_norm": 1.1468195915222168,
2638
+ "learning_rate": 0.00012317481754668468,
2639
+ "loss": 1.4292,
2640
+ "step": 15300
2641
+ },
2642
+ {
2643
+ "epoch": 1.1806784093531266,
2644
+ "grad_norm": 1.2235878705978394,
2645
+ "learning_rate": 0.0001229150975248682,
2646
+ "loss": 1.4428,
2647
+ "step": 15350
2648
+ },
2649
+ {
2650
+ "epoch": 1.1845242673640488,
2651
+ "grad_norm": 1.5518691539764404,
2652
+ "learning_rate": 0.0001226553775030517,
2653
+ "loss": 1.4889,
2654
+ "step": 15400
2655
+ },
2656
+ {
2657
+ "epoch": 1.1883701253749712,
2658
+ "grad_norm": 1.217755913734436,
2659
+ "learning_rate": 0.00012239565748123522,
2660
+ "loss": 1.4841,
2661
+ "step": 15450
2662
+ },
2663
+ {
2664
+ "epoch": 1.1922159833858934,
2665
+ "grad_norm": 1.6093647480010986,
2666
+ "learning_rate": 0.00012213593745941876,
2667
+ "loss": 1.5461,
2668
+ "step": 15500
2669
+ },
2670
+ {
2671
+ "epoch": 1.1922159833858934,
2672
+ "eval_loss": 1.5013692378997803,
2673
+ "eval_runtime": 18.0052,
2674
+ "eval_samples_per_second": 55.539,
2675
+ "eval_steps_per_second": 13.885,
2676
+ "step": 15500
2677
  }
2678
  ],
2679
  "logging_steps": 50,