Training in progress, step 38500, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 36730224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d9fccd6e8cf77b198a4f4b87f312750517824202a02166eb04e56577ecd2ba00
|
| 3 |
size 36730224
|
last-checkpoint/ar_diffusion_info.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1736
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:66d49ec8e5680b4ea0cac8d1930f19a010f9ae0bd4d5ca4b158ebc77b3d54acd
|
| 3 |
size 1736
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 73588346
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:38f23e9c9a2a443780a64b0cf18b93c2bd78a442ad08f1e37293ea92f7c88993
|
| 3 |
size 73588346
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1b4df89a65698a9335c43463fa8aa3cc49c1acdfae5f7a285af000ba78e95b1
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a1d56165e0890bb9432c217b765bdd79a8b4d93dcde80b4671874513824409b
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f3dea87c40748698e240c93bc3eea5482b3f627ab3a1ed391f6cbe2fe5d7da89
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"best_global_step":
|
| 3 |
-
"best_metric": -30.
|
| 4 |
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
|
| 5 |
-
"epoch": 2.
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -6544,6 +6544,92 @@
|
|
| 6544 |
"eval_samples_per_second": 59.431,
|
| 6545 |
"eval_steps_per_second": 14.858,
|
| 6546 |
"step": 38000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6547 |
}
|
| 6548 |
],
|
| 6549 |
"logging_steps": 50,
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_global_step": 38250,
|
| 3 |
+
"best_metric": -30.471132278442383,
|
| 4 |
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
|
| 5 |
+
"epoch": 2.961310668410122,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 38500,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 6544 |
"eval_samples_per_second": 59.431,
|
| 6545 |
"eval_steps_per_second": 14.858,
|
| 6546 |
"step": 38000
|
| 6547 |
+
},
|
| 6548 |
+
{
|
| 6549 |
+
"epoch": 2.926697946311822,
|
| 6550 |
+
"grad_norm": 0.9110496044158936,
|
| 6551 |
+
"learning_rate": 5.033374022803418e-06,
|
| 6552 |
+
"loss": -30.4668,
|
| 6553 |
+
"step": 38050
|
| 6554 |
+
},
|
| 6555 |
+
{
|
| 6556 |
+
"epoch": 2.9305438043227445,
|
| 6557 |
+
"grad_norm": 0.7347224950790405,
|
| 6558 |
+
"learning_rate": 4.773654000986936e-06,
|
| 6559 |
+
"loss": -30.4668,
|
| 6560 |
+
"step": 38100
|
| 6561 |
+
},
|
| 6562 |
+
{
|
| 6563 |
+
"epoch": 2.9343896623336665,
|
| 6564 |
+
"grad_norm": 0.946358859539032,
|
| 6565 |
+
"learning_rate": 4.513933979170454e-06,
|
| 6566 |
+
"loss": -30.4669,
|
| 6567 |
+
"step": 38150
|
| 6568 |
+
},
|
| 6569 |
+
{
|
| 6570 |
+
"epoch": 2.938235520344589,
|
| 6571 |
+
"grad_norm": 0.5716750621795654,
|
| 6572 |
+
"learning_rate": 4.254213957353972e-06,
|
| 6573 |
+
"loss": -30.4669,
|
| 6574 |
+
"step": 38200
|
| 6575 |
+
},
|
| 6576 |
+
{
|
| 6577 |
+
"epoch": 2.9420813783555113,
|
| 6578 |
+
"grad_norm": 1.8405364751815796,
|
| 6579 |
+
"learning_rate": 3.994493935537491e-06,
|
| 6580 |
+
"loss": -30.4667,
|
| 6581 |
+
"step": 38250
|
| 6582 |
+
},
|
| 6583 |
+
{
|
| 6584 |
+
"epoch": 2.9420813783555113,
|
| 6585 |
+
"eval_loss": -30.471132278442383,
|
| 6586 |
+
"eval_runtime": 16.8489,
|
| 6587 |
+
"eval_samples_per_second": 59.351,
|
| 6588 |
+
"eval_steps_per_second": 14.838,
|
| 6589 |
+
"step": 38250
|
| 6590 |
+
},
|
| 6591 |
+
{
|
| 6592 |
+
"epoch": 2.9459272363664333,
|
| 6593 |
+
"grad_norm": 0.6454173922538757,
|
| 6594 |
+
"learning_rate": 3.734773913721009e-06,
|
| 6595 |
+
"loss": -30.467,
|
| 6596 |
+
"step": 38300
|
| 6597 |
+
},
|
| 6598 |
+
{
|
| 6599 |
+
"epoch": 2.9497730943773557,
|
| 6600 |
+
"grad_norm": 0.7232244610786438,
|
| 6601 |
+
"learning_rate": 3.475053891904527e-06,
|
| 6602 |
+
"loss": -30.4668,
|
| 6603 |
+
"step": 38350
|
| 6604 |
+
},
|
| 6605 |
+
{
|
| 6606 |
+
"epoch": 2.953618952388278,
|
| 6607 |
+
"grad_norm": 1.0730299949645996,
|
| 6608 |
+
"learning_rate": 3.215333870088045e-06,
|
| 6609 |
+
"loss": -30.4668,
|
| 6610 |
+
"step": 38400
|
| 6611 |
+
},
|
| 6612 |
+
{
|
| 6613 |
+
"epoch": 2.9574648103992,
|
| 6614 |
+
"grad_norm": 0.4886613190174103,
|
| 6615 |
+
"learning_rate": 2.9556138482715634e-06,
|
| 6616 |
+
"loss": -30.4669,
|
| 6617 |
+
"step": 38450
|
| 6618 |
+
},
|
| 6619 |
+
{
|
| 6620 |
+
"epoch": 2.961310668410122,
|
| 6621 |
+
"grad_norm": 0.5141581296920776,
|
| 6622 |
+
"learning_rate": 2.6958938264550813e-06,
|
| 6623 |
+
"loss": -30.4669,
|
| 6624 |
+
"step": 38500
|
| 6625 |
+
},
|
| 6626 |
+
{
|
| 6627 |
+
"epoch": 2.961310668410122,
|
| 6628 |
+
"eval_loss": -30.471107482910156,
|
| 6629 |
+
"eval_runtime": 16.867,
|
| 6630 |
+
"eval_samples_per_second": 59.287,
|
| 6631 |
+
"eval_steps_per_second": 14.822,
|
| 6632 |
+
"step": 38500
|
| 6633 |
}
|
| 6634 |
],
|
| 6635 |
"logging_steps": 50,
|