Training in progress, step 3500, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 36730224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:11a062f0d5247b9505627696a295b4feca04c6c45dd688fcbc2c07d7828e414c
|
| 3 |
size 36730224
|
last-checkpoint/ar_diffusion_info.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1736
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1ce273710431703c604c9d640c23434ce3f2b038664d7b3df1e4a13e933764d0
|
| 3 |
size 1736
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 73588346
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eb347b9f0c8f6a0d9a0e6b08949ae99d2459f1f662e3a7dd16f39b8ba0ddf69c
|
| 3 |
size 73588346
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3cf21f101c997b0643b73ca8f669be23f0abb9e5a2c6fedb07e17298d9cdd268
|
| 3 |
+
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:028b7d42ce0dbafb2a0126ad830dd1957166a4ca85043c6644487cae8315bfe9
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d16b13e1d35412f9085034fa2481674486c7b966c61cae81848c36587b827f3d
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"best_global_step":
|
| 3 |
-
"best_metric": 1.
|
| 4 |
-
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -524,6 +524,92 @@
|
|
| 524 |
"eval_samples_per_second": 59.161,
|
| 525 |
"eval_steps_per_second": 14.79,
|
| 526 |
"step": 3000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 527 |
}
|
| 528 |
],
|
| 529 |
"logging_steps": 50,
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_global_step": 3500,
|
| 3 |
+
"best_metric": 1.5354665517807007,
|
| 4 |
+
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-3500",
|
| 5 |
+
"epoch": 0.26921006076455656,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 3500,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 524 |
"eval_samples_per_second": 59.161,
|
| 525 |
"eval_steps_per_second": 14.79,
|
| 526 |
"step": 3000
|
| 527 |
+
},
|
| 528 |
+
{
|
| 529 |
+
"epoch": 0.23459733866625646,
|
| 530 |
+
"grad_norm": 3.606706142425537,
|
| 531 |
+
"learning_rate": 0.00018678025088954107,
|
| 532 |
+
"loss": 1.5998,
|
| 533 |
+
"step": 3050
|
| 534 |
+
},
|
| 535 |
+
{
|
| 536 |
+
"epoch": 0.23844319667717867,
|
| 537 |
+
"grad_norm": 1.2053996324539185,
|
| 538 |
+
"learning_rate": 0.0001865205308677246,
|
| 539 |
+
"loss": 1.5512,
|
| 540 |
+
"step": 3100
|
| 541 |
+
},
|
| 542 |
+
{
|
| 543 |
+
"epoch": 0.2422890546881009,
|
| 544 |
+
"grad_norm": 1.5675193071365356,
|
| 545 |
+
"learning_rate": 0.00018626081084590813,
|
| 546 |
+
"loss": 1.5495,
|
| 547 |
+
"step": 3150
|
| 548 |
+
},
|
| 549 |
+
{
|
| 550 |
+
"epoch": 0.24613491269902316,
|
| 551 |
+
"grad_norm": 1.1827441453933716,
|
| 552 |
+
"learning_rate": 0.00018600109082409164,
|
| 553 |
+
"loss": 1.6509,
|
| 554 |
+
"step": 3200
|
| 555 |
+
},
|
| 556 |
+
{
|
| 557 |
+
"epoch": 0.2499807707099454,
|
| 558 |
+
"grad_norm": 1.7992998361587524,
|
| 559 |
+
"learning_rate": 0.00018574137080227515,
|
| 560 |
+
"loss": 1.5617,
|
| 561 |
+
"step": 3250
|
| 562 |
+
},
|
| 563 |
+
{
|
| 564 |
+
"epoch": 0.2499807707099454,
|
| 565 |
+
"eval_loss": 1.5403118133544922,
|
| 566 |
+
"eval_runtime": 16.9627,
|
| 567 |
+
"eval_samples_per_second": 58.953,
|
| 568 |
+
"eval_steps_per_second": 14.738,
|
| 569 |
+
"step": 3250
|
| 570 |
+
},
|
| 571 |
+
{
|
| 572 |
+
"epoch": 0.25382662872086764,
|
| 573 |
+
"grad_norm": 1.3044147491455078,
|
| 574 |
+
"learning_rate": 0.00018548165078045866,
|
| 575 |
+
"loss": 1.5226,
|
| 576 |
+
"step": 3300
|
| 577 |
+
},
|
| 578 |
+
{
|
| 579 |
+
"epoch": 0.2576724867317899,
|
| 580 |
+
"grad_norm": 1.4057557582855225,
|
| 581 |
+
"learning_rate": 0.0001852219307586422,
|
| 582 |
+
"loss": 1.517,
|
| 583 |
+
"step": 3350
|
| 584 |
+
},
|
| 585 |
+
{
|
| 586 |
+
"epoch": 0.26151834474271207,
|
| 587 |
+
"grad_norm": 0.9969176650047302,
|
| 588 |
+
"learning_rate": 0.00018496221073682572,
|
| 589 |
+
"loss": 1.5157,
|
| 590 |
+
"step": 3400
|
| 591 |
+
},
|
| 592 |
+
{
|
| 593 |
+
"epoch": 0.2653642027536343,
|
| 594 |
+
"grad_norm": 1.0183119773864746,
|
| 595 |
+
"learning_rate": 0.00018470249071500923,
|
| 596 |
+
"loss": 1.5012,
|
| 597 |
+
"step": 3450
|
| 598 |
+
},
|
| 599 |
+
{
|
| 600 |
+
"epoch": 0.26921006076455656,
|
| 601 |
+
"grad_norm": 1.1896519660949707,
|
| 602 |
+
"learning_rate": 0.00018444277069319277,
|
| 603 |
+
"loss": 1.4655,
|
| 604 |
+
"step": 3500
|
| 605 |
+
},
|
| 606 |
+
{
|
| 607 |
+
"epoch": 0.26921006076455656,
|
| 608 |
+
"eval_loss": 1.5354665517807007,
|
| 609 |
+
"eval_runtime": 16.8964,
|
| 610 |
+
"eval_samples_per_second": 59.184,
|
| 611 |
+
"eval_steps_per_second": 14.796,
|
| 612 |
+
"step": 3500
|
| 613 |
}
|
| 614 |
],
|
| 615 |
"logging_steps": 50,
|