rootxhacker commited on
Commit
69a53a8
·
verified ·
1 Parent(s): 026888d

Training in progress, step 15500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35e8b88b22335eff4aa95a5db7b6615364abd358d14317af2dec16bcdb2efc61
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6067e5da0c9bf8b752e4ae8c8bf60db131ffcb5bdc72c1bef0c50cf034756793
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8a93a0264eb6605e0d626be93ba5f8dee38234bcac32f7e6d5d1cccc2a1057f
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70b3ce160c7b776e775a166f1cb68bd1487d55fdb2e7ccd66b23097de2464f87
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c298ef220b2440e2c8688c2918aba70637e9e1c8a2951767772c4969343135c8
3
- size 14308
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e443e4b8c749db54f7826bd4573803e0f05009235443f881fa9932376e1aba8
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08eda93be407e2e8125df91c801018bcedfb4cde17a8cf02b9f76837f9ade1c1
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c46d93b856d73467b22e6683f2d8e760cc9d35c9e6261e9397f7ae15ef306957
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a00183886b48331e49cfb9b953bfa8a92696629d3ac41c1b27b8636569368855
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:682c1f39bde20c90a4373af21bd5402ac0e8e55951fcfe8f05caeac3b074ab9a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 14750,
3
- "best_metric": 1.4990500211715698,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-13000",
5
- "epoch": 1.153757403276671,
6
  "eval_steps": 250,
7
- "global_step": 15000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2588,6 +2588,92 @@
2588
  "eval_samples_per_second": 55.799,
2589
  "eval_steps_per_second": 13.95,
2590
  "step": 15000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2591
  }
2592
  ],
2593
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 15500,
3
+ "best_metric": 1.4925825595855713,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-15500",
5
+ "epoch": 1.1922159833858934,
6
  "eval_steps": 250,
7
+ "global_step": 15500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2588
  "eval_samples_per_second": 55.799,
2589
  "eval_steps_per_second": 13.95,
2590
  "step": 15000
2591
+ },
2592
+ {
2593
+ "epoch": 1.1576032612875933,
2594
+ "grad_norm": 1.5460007190704346,
2595
+ "learning_rate": 6.311091503988823e-05,
2596
+ "loss": 1.442,
2597
+ "step": 15050
2598
+ },
2599
+ {
2600
+ "epoch": 1.1614491192985155,
2601
+ "grad_norm": 1.001439094543457,
2602
+ "learning_rate": 6.285357155986911e-05,
2603
+ "loss": 1.5308,
2604
+ "step": 15100
2605
+ },
2606
+ {
2607
+ "epoch": 1.1652949773094377,
2608
+ "grad_norm": 0.8740602731704712,
2609
+ "learning_rate": 6.259622807985e-05,
2610
+ "loss": 1.455,
2611
+ "step": 15150
2612
+ },
2613
+ {
2614
+ "epoch": 1.16914083532036,
2615
+ "grad_norm": 2.034207820892334,
2616
+ "learning_rate": 6.233888459983088e-05,
2617
+ "loss": 1.5089,
2618
+ "step": 15200
2619
+ },
2620
+ {
2621
+ "epoch": 1.1729866933312822,
2622
+ "grad_norm": 1.8656599521636963,
2623
+ "learning_rate": 6.208154111981176e-05,
2624
+ "loss": 1.5368,
2625
+ "step": 15250
2626
+ },
2627
+ {
2628
+ "epoch": 1.1729866933312822,
2629
+ "eval_loss": 1.4986381530761719,
2630
+ "eval_runtime": 18.1736,
2631
+ "eval_samples_per_second": 55.025,
2632
+ "eval_steps_per_second": 13.756,
2633
+ "step": 15250
2634
+ },
2635
+ {
2636
+ "epoch": 1.1768325513422044,
2637
+ "grad_norm": 1.2697277069091797,
2638
+ "learning_rate": 6.182419763979266e-05,
2639
+ "loss": 1.4239,
2640
+ "step": 15300
2641
+ },
2642
+ {
2643
+ "epoch": 1.1806784093531266,
2644
+ "grad_norm": 1.1131771802902222,
2645
+ "learning_rate": 6.156685415977354e-05,
2646
+ "loss": 1.4309,
2647
+ "step": 15350
2648
+ },
2649
+ {
2650
+ "epoch": 1.1845242673640488,
2651
+ "grad_norm": 1.5322145223617554,
2652
+ "learning_rate": 6.130951067975442e-05,
2653
+ "loss": 1.4793,
2654
+ "step": 15400
2655
+ },
2656
+ {
2657
+ "epoch": 1.1883701253749712,
2658
+ "grad_norm": 1.1703407764434814,
2659
+ "learning_rate": 6.10521671997353e-05,
2660
+ "loss": 1.4761,
2661
+ "step": 15450
2662
+ },
2663
+ {
2664
+ "epoch": 1.1922159833858934,
2665
+ "grad_norm": 1.4056655168533325,
2666
+ "learning_rate": 6.079482371971618e-05,
2667
+ "loss": 1.5311,
2668
+ "step": 15500
2669
+ },
2670
+ {
2671
+ "epoch": 1.1922159833858934,
2672
+ "eval_loss": 1.4925825595855713,
2673
+ "eval_runtime": 18.2116,
2674
+ "eval_samples_per_second": 54.91,
2675
+ "eval_steps_per_second": 13.727,
2676
+ "step": 15500
2677
  }
2678
  ],
2679
  "logging_steps": 50,