Training in progress, step 39000, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 36730224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aec93bf3af378bc69e279d3da1f22a2356cd072cc873629676a0de486c684213
|
| 3 |
size 36730224
|
last-checkpoint/ar_diffusion_info.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1736
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1e17b4b57a9838b7215832af6318f536b3b6acee7c730845278b1a8f5d0cb17
|
| 3 |
size 1736
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 73588346
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e8f0bbc447de59f2033199d6477b7b4db91aeaa9c681dfd30f7ca1df276b9b1
|
| 3 |
size 73588346
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:60919181f100a38a6bc8dcdeb135786c172525306422517244d9c17e263e8748
|
| 3 |
+
size 14180
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5cd25c48749969bd5bf972a3edb2b91874acf34f950f422e77399a6589ee1816
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:06d9f9352b3f9aa43ec304f953eb511fb6c42be93c910bdb749b28823ddd5892
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": 38250,
|
| 3 |
"best_metric": -30.471132278442383,
|
| 4 |
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
|
| 5 |
-
"epoch": 2.
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -6630,6 +6630,92 @@
|
|
| 6630 |
"eval_samples_per_second": 59.287,
|
| 6631 |
"eval_steps_per_second": 14.822,
|
| 6632 |
"step": 38500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6633 |
}
|
| 6634 |
],
|
| 6635 |
"logging_steps": 50,
|
|
|
|
| 2 |
"best_global_step": 38250,
|
| 3 |
"best_metric": -30.471132278442383,
|
| 4 |
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
|
| 5 |
+
"epoch": 2.999769248519345,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 39000,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 6630 |
"eval_samples_per_second": 59.287,
|
| 6631 |
"eval_steps_per_second": 14.822,
|
| 6632 |
"step": 38500
|
| 6633 |
+
},
|
| 6634 |
+
{
|
| 6635 |
+
"epoch": 2.9651565264210444,
|
| 6636 |
+
"grad_norm": 0.5108075737953186,
|
| 6637 |
+
"learning_rate": 2.4361738046385997e-06,
|
| 6638 |
+
"loss": -30.467,
|
| 6639 |
+
"step": 38550
|
| 6640 |
+
},
|
| 6641 |
+
{
|
| 6642 |
+
"epoch": 2.969002384431967,
|
| 6643 |
+
"grad_norm": 0.608711838722229,
|
| 6644 |
+
"learning_rate": 2.1764537828221176e-06,
|
| 6645 |
+
"loss": -30.467,
|
| 6646 |
+
"step": 38600
|
| 6647 |
+
},
|
| 6648 |
+
{
|
| 6649 |
+
"epoch": 2.972848242442889,
|
| 6650 |
+
"grad_norm": 0.9723970890045166,
|
| 6651 |
+
"learning_rate": 1.916733761005636e-06,
|
| 6652 |
+
"loss": -30.4672,
|
| 6653 |
+
"step": 38650
|
| 6654 |
+
},
|
| 6655 |
+
{
|
| 6656 |
+
"epoch": 2.976694100453811,
|
| 6657 |
+
"grad_norm": 0.4361710846424103,
|
| 6658 |
+
"learning_rate": 1.6570137391891543e-06,
|
| 6659 |
+
"loss": -30.467,
|
| 6660 |
+
"step": 38700
|
| 6661 |
+
},
|
| 6662 |
+
{
|
| 6663 |
+
"epoch": 2.9805399584647336,
|
| 6664 |
+
"grad_norm": 1.1543878316879272,
|
| 6665 |
+
"learning_rate": 1.3972937173726722e-06,
|
| 6666 |
+
"loss": -30.4669,
|
| 6667 |
+
"step": 38750
|
| 6668 |
+
},
|
| 6669 |
+
{
|
| 6670 |
+
"epoch": 2.9805399584647336,
|
| 6671 |
+
"eval_loss": -30.47084617614746,
|
| 6672 |
+
"eval_runtime": 16.7842,
|
| 6673 |
+
"eval_samples_per_second": 59.58,
|
| 6674 |
+
"eval_steps_per_second": 14.895,
|
| 6675 |
+
"step": 38750
|
| 6676 |
+
},
|
| 6677 |
+
{
|
| 6678 |
+
"epoch": 2.9843858164756556,
|
| 6679 |
+
"grad_norm": 0.4299401640892029,
|
| 6680 |
+
"learning_rate": 1.1375736955561905e-06,
|
| 6681 |
+
"loss": -30.467,
|
| 6682 |
+
"step": 38800
|
| 6683 |
+
},
|
| 6684 |
+
{
|
| 6685 |
+
"epoch": 2.988231674486578,
|
| 6686 |
+
"grad_norm": 0.8476674556732178,
|
| 6687 |
+
"learning_rate": 8.778536737397087e-07,
|
| 6688 |
+
"loss": -30.467,
|
| 6689 |
+
"step": 38850
|
| 6690 |
+
},
|
| 6691 |
+
{
|
| 6692 |
+
"epoch": 2.9920775324975004,
|
| 6693 |
+
"grad_norm": 1.1194632053375244,
|
| 6694 |
+
"learning_rate": 6.181336519232268e-07,
|
| 6695 |
+
"loss": -30.4671,
|
| 6696 |
+
"step": 38900
|
| 6697 |
+
},
|
| 6698 |
+
{
|
| 6699 |
+
"epoch": 2.9959233905084224,
|
| 6700 |
+
"grad_norm": 0.7896573543548584,
|
| 6701 |
+
"learning_rate": 3.584136301067449e-07,
|
| 6702 |
+
"loss": -30.4671,
|
| 6703 |
+
"step": 38950
|
| 6704 |
+
},
|
| 6705 |
+
{
|
| 6706 |
+
"epoch": 2.999769248519345,
|
| 6707 |
+
"grad_norm": 0.32893452048301697,
|
| 6708 |
+
"learning_rate": 9.869360829026311e-08,
|
| 6709 |
+
"loss": -30.4671,
|
| 6710 |
+
"step": 39000
|
| 6711 |
+
},
|
| 6712 |
+
{
|
| 6713 |
+
"epoch": 2.999769248519345,
|
| 6714 |
+
"eval_loss": -30.471084594726562,
|
| 6715 |
+
"eval_runtime": 16.7188,
|
| 6716 |
+
"eval_samples_per_second": 59.813,
|
| 6717 |
+
"eval_steps_per_second": 14.953,
|
| 6718 |
+
"step": 39000
|
| 6719 |
}
|
| 6720 |
],
|
| 6721 |
"logging_steps": 50,
|