rootxhacker commited on
Commit
b7455fd
·
verified ·
1 Parent(s): 2407f62

Training in progress, step 3000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e10a7c8bb0d516328a4f08fb821bc0e53102e1b0a0d8ea1d6c8cc3bce9dca90
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84adf5fb4a3947b35993a790cf4973ef5593a87e6e164d9578f03b8b9c27e099
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fdb38bb607b4da045b51a34cfbf7428a4bf481a4b7dcfcd083d72647f444c879
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:308f424d188d86d1fcd72962eb2fa00bace74a0cc7f5cd9ae44b93fa7b48ebe5
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d026d37c2efc698996a40494d349ec54a4ff40c81599ec63f15410e52a0e87f
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65a04555108b26fbf193a851c8c842cf49a4c844bc1961b985cfc83f43bcde61
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:100aab8649e7180e1ab3ef8feaab197f99e31e9f728f5189c5431d80dbc2eacd
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:238412d43efe8b606030d3301ede6818fb1a62974cd0cbb655e125124222d862
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:996af29fdb07e722f1f70dd9ba54354e5a4baf4f673aef8bef5139d74d107a73
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04b1366b3719985cdc1dc0eb0fa220f7b35ea714d4ff368885830073c629f4d9
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d57c57de49427b666f955c0cc0173da1ca9651de809c40ac268e017db320996
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2976163d4a5cd674bbf6d82ca442cb1ceba9d5f9d444c557ab914dab5858c6ce
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 2500,
3
- "best_metric": 1.5644338130950928,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-2500",
5
- "epoch": 0.19229290054611184,
6
  "eval_steps": 250,
7
- "global_step": 2500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -438,6 +438,92 @@
438
  "eval_samples_per_second": 59.426,
439
  "eval_steps_per_second": 14.856,
440
  "step": 2500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
441
  }
442
  ],
443
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 3000,
3
+ "best_metric": 1.5423797369003296,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-3000",
5
+ "epoch": 0.2307514806553342,
6
  "eval_steps": 250,
7
+ "global_step": 3000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
438
  "eval_samples_per_second": 59.426,
439
  "eval_steps_per_second": 14.856,
440
  "step": 2500
441
+ },
442
+ {
443
+ "epoch": 0.19613875855703408,
444
+ "grad_norm": 0.8316934704780579,
445
+ "learning_rate": 0.00018937745110770592,
446
+ "loss": 1.5182,
447
+ "step": 2550
448
+ },
449
+ {
450
+ "epoch": 0.19998461656795632,
451
+ "grad_norm": 1.3362255096435547,
452
+ "learning_rate": 0.0001891177310858894,
453
+ "loss": 1.5557,
454
+ "step": 2600
455
+ },
456
+ {
457
+ "epoch": 0.20383047457887854,
458
+ "grad_norm": 1.2854304313659668,
459
+ "learning_rate": 0.00018885801106407294,
460
+ "loss": 1.5417,
461
+ "step": 2650
462
+ },
463
+ {
464
+ "epoch": 0.20767633258980078,
465
+ "grad_norm": 1.7450816631317139,
466
+ "learning_rate": 0.00018859829104225646,
467
+ "loss": 1.5489,
468
+ "step": 2700
469
+ },
470
+ {
471
+ "epoch": 0.21152219060072303,
472
+ "grad_norm": 1.546242117881775,
473
+ "learning_rate": 0.00018833857102043997,
474
+ "loss": 1.4724,
475
+ "step": 2750
476
+ },
477
+ {
478
+ "epoch": 0.21152219060072303,
479
+ "eval_loss": 1.5916800498962402,
480
+ "eval_runtime": 16.9148,
481
+ "eval_samples_per_second": 59.12,
482
+ "eval_steps_per_second": 14.78,
483
+ "step": 2750
484
+ },
485
+ {
486
+ "epoch": 0.21536804861164527,
487
+ "grad_norm": 1.8118102550506592,
488
+ "learning_rate": 0.0001880788509986235,
489
+ "loss": 1.511,
490
+ "step": 2800
491
+ },
492
+ {
493
+ "epoch": 0.21921390662256748,
494
+ "grad_norm": 1.7519344091415405,
495
+ "learning_rate": 0.00018781913097680702,
496
+ "loss": 1.5161,
497
+ "step": 2850
498
+ },
499
+ {
500
+ "epoch": 0.22305976463348973,
501
+ "grad_norm": 1.7779595851898193,
502
+ "learning_rate": 0.00018755941095499053,
503
+ "loss": 1.6146,
504
+ "step": 2900
505
+ },
506
+ {
507
+ "epoch": 0.22690562264441197,
508
+ "grad_norm": 1.0836044549942017,
509
+ "learning_rate": 0.00018729969093317405,
510
+ "loss": 1.5074,
511
+ "step": 2950
512
+ },
513
+ {
514
+ "epoch": 0.2307514806553342,
515
+ "grad_norm": 1.4328197240829468,
516
+ "learning_rate": 0.00018703997091135756,
517
+ "loss": 1.5865,
518
+ "step": 3000
519
+ },
520
+ {
521
+ "epoch": 0.2307514806553342,
522
+ "eval_loss": 1.5423797369003296,
523
+ "eval_runtime": 16.9029,
524
+ "eval_samples_per_second": 59.161,
525
+ "eval_steps_per_second": 14.79,
526
+ "step": 3000
527
  }
528
  ],
529
  "logging_steps": 50,