rootxhacker commited on
Commit
8d72071
·
verified ·
1 Parent(s): 43aaf0b

Training in progress, step 3000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1615a27846e2fc7e29e90ffc163230511731ab45b760a836e71cf204236e624d
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f77dbb7c6bf6487a48a714e4c4e440e388d69a8afa7310e97fc320615f989a6
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6df96e6756fb0b0f050ff10208655f563c96fccd6a701602a5e90ccbd1e235e0
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1814592ba74cdcca20fb08c3ea7e5ee573986aff77c3810b20a982bd1be5d78
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74818f2f3daacf3c0fca3dde38a211872e713dacdefaa18d7435559b30dc197d
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cdad770e6415de25f2503b80a04125ce36c191e7ebbe40cc70935c15b3079b5
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc76fea5e634e306eb2d4aa830b8059981049eb83bd4681631ccd02732a47898
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9aa9825ab5c6a64b78405ab88bd271bc07d34d8fb6641eef252a11c55bfe386
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d8c94e41b1ac1e8ac4ea58294a516d2e00679d859d84c40c165bc9d685fc309
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0276c75f304794326dfbdfec2d07b2898318c16d7455e1133773ea583662ada9
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 2500,
3
- "best_metric": 4.558788776397705,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-2500",
5
- "epoch": 0.19229290054611184,
6
  "eval_steps": 250,
7
- "global_step": 2500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -438,6 +438,92 @@
438
  "eval_samples_per_second": 52.664,
439
  "eval_steps_per_second": 13.166,
440
  "step": 2500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
441
  }
442
  ],
443
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 3000,
3
+ "best_metric": 4.55267858505249,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-3000",
5
+ "epoch": 0.2307514806553342,
6
  "eval_steps": 250,
7
+ "global_step": 3000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
438
  "eval_samples_per_second": 52.664,
439
  "eval_steps_per_second": 13.166,
440
  "step": 2500
441
+ },
442
+ {
443
+ "epoch": 0.19613875855703408,
444
+ "grad_norm": 2.456908702850342,
445
+ "learning_rate": 0.00012742105069666555,
446
+ "loss": 4.3847,
447
+ "step": 2550
448
+ },
449
+ {
450
+ "epoch": 0.19998461656795632,
451
+ "grad_norm": 5.154629707336426,
452
+ "learning_rate": 0.00012716370721664645,
453
+ "loss": 4.6019,
454
+ "step": 2600
455
+ },
456
+ {
457
+ "epoch": 0.20383047457887854,
458
+ "grad_norm": 3.0423479080200195,
459
+ "learning_rate": 0.0001269063637366273,
460
+ "loss": 4.4796,
461
+ "step": 2650
462
+ },
463
+ {
464
+ "epoch": 0.20767633258980078,
465
+ "grad_norm": 4.218437194824219,
466
+ "learning_rate": 0.00012664902025660819,
467
+ "loss": 4.5566,
468
+ "step": 2700
469
+ },
470
+ {
471
+ "epoch": 0.21152219060072303,
472
+ "grad_norm": 5.20380163192749,
473
+ "learning_rate": 0.0001263916767765891,
474
+ "loss": 4.3311,
475
+ "step": 2750
476
+ },
477
+ {
478
+ "epoch": 0.21152219060072303,
479
+ "eval_loss": 4.574987888336182,
480
+ "eval_runtime": 18.8565,
481
+ "eval_samples_per_second": 53.032,
482
+ "eval_steps_per_second": 13.258,
483
+ "step": 2750
484
+ },
485
+ {
486
+ "epoch": 0.21536804861164527,
487
+ "grad_norm": 4.369246482849121,
488
+ "learning_rate": 0.00012613433329656997,
489
+ "loss": 4.4131,
490
+ "step": 2800
491
+ },
492
+ {
493
+ "epoch": 0.21921390662256748,
494
+ "grad_norm": 5.0442376136779785,
495
+ "learning_rate": 0.00012587698981655085,
496
+ "loss": 4.5027,
497
+ "step": 2850
498
+ },
499
+ {
500
+ "epoch": 0.22305976463348973,
501
+ "grad_norm": 3.6387200355529785,
502
+ "learning_rate": 0.00012561964633653173,
503
+ "loss": 4.6659,
504
+ "step": 2900
505
+ },
506
+ {
507
+ "epoch": 0.22690562264441197,
508
+ "grad_norm": 3.7960562705993652,
509
+ "learning_rate": 0.0001253623028565126,
510
+ "loss": 4.4826,
511
+ "step": 2950
512
+ },
513
+ {
514
+ "epoch": 0.2307514806553342,
515
+ "grad_norm": 4.273965835571289,
516
+ "learning_rate": 0.0001251049593764935,
517
+ "loss": 4.5869,
518
+ "step": 3000
519
+ },
520
+ {
521
+ "epoch": 0.2307514806553342,
522
+ "eval_loss": 4.55267858505249,
523
+ "eval_runtime": 18.9735,
524
+ "eval_samples_per_second": 52.705,
525
+ "eval_steps_per_second": 13.176,
526
+ "step": 3000
527
  }
528
  ],
529
  "logging_steps": 50,