rootxhacker commited on
Commit
4fa58b3
·
verified ·
1 Parent(s): 285c634

Training in progress, step 20000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:879ead83bd2c1fe46b7e68211ea0caa2723b2414cf5d08c8c7b5b29823750c25
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bff1d47cdb13e83e8fe6a9d20b6c173dfae48b029aecf5c92bf458772133f5b2
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a6fddc34b7b7d3235b23dac04ebf76b113434f948440bce4f7bae9c4f95c22a
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c26f8a7499e23aa16477b380bf134293878cb8936b98f8cad1d8a606fc561e1a
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa7c19bbb017976e63ac5e9fec3eea0eda0cf5868e0275e6b00fde078e2c850d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b205d8106465470d82a9f668e113b1a4c937f3fe768b385e78f85eb171e49cc2
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:abdb0a20afab4dc0c80b1c5f76a237a0247272f8d00055102fe692e4d58c1f05
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:344e181f9677a4d18097c7085524ad711f8dfb151d42735faf9658fba6102bed
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:488aaa88c1afe9c9c3279271cc45e297a0d20ece7902b965486e19f7d9811602
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fc471d6cffa6d4b3db986d1e54c37de8f3b7f607849783a6e6b32bce97519eb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 18750,
3
  "best_metric": 1.4680087566375732,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-15500",
5
- "epoch": 1.4998846242596724,
6
  "eval_steps": 250,
7
- "global_step": 19500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -3362,6 +3362,92 @@
3362
  "eval_samples_per_second": 55.491,
3363
  "eval_steps_per_second": 13.873,
3364
  "step": 19500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3365
  }
3366
  ],
3367
  "logging_steps": 50,
 
2
  "best_global_step": 18750,
3
  "best_metric": 1.4680087566375732,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-15500",
5
+ "epoch": 1.5383432043688947,
6
  "eval_steps": 250,
7
+ "global_step": 20000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
3362
  "eval_samples_per_second": 55.491,
3363
  "eval_steps_per_second": 13.873,
3364
  "step": 19500
3365
+ },
3366
+ {
3367
+ "epoch": 1.5037304822705946,
3368
+ "grad_norm": 0.8779070377349854,
3369
+ "learning_rate": 3.995000183816771e-05,
3370
+ "loss": 1.3728,
3371
+ "step": 19550
3372
+ },
3373
+ {
3374
+ "epoch": 1.5075763402815168,
3375
+ "grad_norm": 1.6068123579025269,
3376
+ "learning_rate": 3.9692658358148595e-05,
3377
+ "loss": 1.5204,
3378
+ "step": 19600
3379
+ },
3380
+ {
3381
+ "epoch": 1.511422198292439,
3382
+ "grad_norm": 1.7712832689285278,
3383
+ "learning_rate": 3.9435314878129475e-05,
3384
+ "loss": 1.514,
3385
+ "step": 19650
3386
+ },
3387
+ {
3388
+ "epoch": 1.5152680563033614,
3389
+ "grad_norm": 1.2519572973251343,
3390
+ "learning_rate": 3.917797139811036e-05,
3391
+ "loss": 1.3953,
3392
+ "step": 19700
3393
+ },
3394
+ {
3395
+ "epoch": 1.5191139143142836,
3396
+ "grad_norm": 1.5644786357879639,
3397
+ "learning_rate": 3.892062791809125e-05,
3398
+ "loss": 1.4772,
3399
+ "step": 19750
3400
+ },
3401
+ {
3402
+ "epoch": 1.5191139143142836,
3403
+ "eval_loss": 1.4710900783538818,
3404
+ "eval_runtime": 18.4205,
3405
+ "eval_samples_per_second": 54.287,
3406
+ "eval_steps_per_second": 13.572,
3407
+ "step": 19750
3408
+ },
3409
+ {
3410
+ "epoch": 1.5229597723252057,
3411
+ "grad_norm": 1.6755670309066772,
3412
+ "learning_rate": 3.866328443807212e-05,
3413
+ "loss": 1.4148,
3414
+ "step": 19800
3415
+ },
3416
+ {
3417
+ "epoch": 1.5268056303361282,
3418
+ "grad_norm": 1.7168843746185303,
3419
+ "learning_rate": 3.840594095805301e-05,
3420
+ "loss": 1.4211,
3421
+ "step": 19850
3422
+ },
3423
+ {
3424
+ "epoch": 1.5306514883470501,
3425
+ "grad_norm": 1.5205817222595215,
3426
+ "learning_rate": 3.8148597478033894e-05,
3427
+ "loss": 1.4663,
3428
+ "step": 19900
3429
+ },
3430
+ {
3431
+ "epoch": 1.5344973463579725,
3432
+ "grad_norm": 1.608231544494629,
3433
+ "learning_rate": 3.789125399801477e-05,
3434
+ "loss": 1.3634,
3435
+ "step": 19950
3436
+ },
3437
+ {
3438
+ "epoch": 1.5383432043688947,
3439
+ "grad_norm": 1.5260729789733887,
3440
+ "learning_rate": 3.763391051799566e-05,
3441
+ "loss": 1.4114,
3442
+ "step": 20000
3443
+ },
3444
+ {
3445
+ "epoch": 1.5383432043688947,
3446
+ "eval_loss": 1.4733539819717407,
3447
+ "eval_runtime": 18.105,
3448
+ "eval_samples_per_second": 55.233,
3449
+ "eval_steps_per_second": 13.808,
3450
+ "step": 20000
3451
  }
3452
  ],
3453
  "logging_steps": 50,