rootxhacker commited on
Commit
88fb3ef
·
verified ·
1 Parent(s): d9f087f

Training in progress, step 38000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c27b235ffd7eb3533febc9af0f60470e3dba0b8e45360272535091fdccd177b
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc9e082dc3b2b599ece245c68d836aa3280cd4e7a9b0918ad403250af895d5e3
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:696e56b33a026fc5d8957bdce7458fb9a6f7ad74969dc7cd1a68cbc7a0b070a7
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c350b485739b4fddb9d5148423e5dfbced50fc251b92973c875bcddb4d66a775
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31f016b3f65efa39acc365f2cd200e250e466d276146e2b7b6697bb3bf4c7a78
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54a8acb2b80e4297ad7cc3210bf7b7be72618237e5040794a64b58c69faec7fc
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97f5694df4cc55cfb23211b8e8f8f54247ffb944ba00f7e779e21697183c2f1f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2c0cded846d640f3719120429980641a6206dc92844fb42e8f946ff3bea6b2e
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2cc58fd6b30bc41899d1238111aee5cdb3d8eeebbfe25f934de223f53728d54d
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fc19191502cd55e651632979b36f722e8851e564a429a904a6573212b31486c
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23ce717bb1ff7d19ce8b39673c5e006d14b3fec124190d834c88a63ab05da6d0
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:541cb6dea9c80e2359e5134078f161ec83a5de3af8e29b3e9b6b880e5dd9058f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 37500,
3
- "best_metric": -30.470531463623047,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
5
- "epoch": 2.8843935081916774,
6
  "eval_steps": 250,
7
- "global_step": 37500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -6458,6 +6458,92 @@
6458
  "eval_samples_per_second": 59.544,
6459
  "eval_steps_per_second": 14.886,
6460
  "step": 37500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6461
  }
6462
  ],
6463
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 37750,
3
+ "best_metric": -30.47093391418457,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
5
+ "epoch": 2.9228520883008997,
6
  "eval_steps": 250,
7
+ "global_step": 38000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
6458
  "eval_samples_per_second": 59.544,
6459
  "eval_steps_per_second": 14.886,
6460
  "step": 37500
6461
+ },
6462
+ {
6463
+ "epoch": 2.8882393662026,
6464
+ "grad_norm": 0.6514129042625427,
6465
+ "learning_rate": 7.630574240968237e-06,
6466
+ "loss": -30.4666,
6467
+ "step": 37550
6468
+ },
6469
+ {
6470
+ "epoch": 2.892085224213522,
6471
+ "grad_norm": 1.0104998350143433,
6472
+ "learning_rate": 7.3708542191517545e-06,
6473
+ "loss": -30.4667,
6474
+ "step": 37600
6475
+ },
6476
+ {
6477
+ "epoch": 2.895931082224444,
6478
+ "grad_norm": 0.713027834892273,
6479
+ "learning_rate": 7.111134197335273e-06,
6480
+ "loss": -30.4665,
6481
+ "step": 37650
6482
+ },
6483
+ {
6484
+ "epoch": 2.8997769402353666,
6485
+ "grad_norm": 0.6001731157302856,
6486
+ "learning_rate": 6.851414175518792e-06,
6487
+ "loss": -30.4666,
6488
+ "step": 37700
6489
+ },
6490
+ {
6491
+ "epoch": 2.903622798246289,
6492
+ "grad_norm": 0.6344665288925171,
6493
+ "learning_rate": 6.591694153702309e-06,
6494
+ "loss": -30.4664,
6495
+ "step": 37750
6496
+ },
6497
+ {
6498
+ "epoch": 2.903622798246289,
6499
+ "eval_loss": -30.47093391418457,
6500
+ "eval_runtime": 16.9622,
6501
+ "eval_samples_per_second": 58.955,
6502
+ "eval_steps_per_second": 14.739,
6503
+ "step": 37750
6504
+ },
6505
+ {
6506
+ "epoch": 2.907468656257211,
6507
+ "grad_norm": 0.5279808640480042,
6508
+ "learning_rate": 6.331974131885828e-06,
6509
+ "loss": -30.4667,
6510
+ "step": 37800
6511
+ },
6512
+ {
6513
+ "epoch": 2.9113145142681334,
6514
+ "grad_norm": 0.9036094546318054,
6515
+ "learning_rate": 6.072254110069346e-06,
6516
+ "loss": -30.4667,
6517
+ "step": 37850
6518
+ },
6519
+ {
6520
+ "epoch": 2.9151603722790553,
6521
+ "grad_norm": 0.601686954498291,
6522
+ "learning_rate": 5.812534088252864e-06,
6523
+ "loss": -30.4668,
6524
+ "step": 37900
6525
+ },
6526
+ {
6527
+ "epoch": 2.9190062302899777,
6528
+ "grad_norm": 1.0191963911056519,
6529
+ "learning_rate": 5.552814066436382e-06,
6530
+ "loss": -30.4667,
6531
+ "step": 37950
6532
+ },
6533
+ {
6534
+ "epoch": 2.9228520883008997,
6535
+ "grad_norm": 0.9260501265525818,
6536
+ "learning_rate": 5.2930940446198996e-06,
6537
+ "loss": -30.4667,
6538
+ "step": 38000
6539
+ },
6540
+ {
6541
+ "epoch": 2.9228520883008997,
6542
+ "eval_loss": -30.47071075439453,
6543
+ "eval_runtime": 16.8261,
6544
+ "eval_samples_per_second": 59.431,
6545
+ "eval_steps_per_second": 14.858,
6546
+ "step": 38000
6547
  }
6548
  ],
6549
  "logging_steps": 50,