Training in progress, step 2320, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff6279bb72a1a41fb48956b4feb1a93393d18d4e6392ad78211fc04c9f099df2
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:74ccf80491212ab3533b791bb0a94ac69d6d47751d6deeebce7061fdefd9b30f
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:19bb1874a3bf4e8cd76c40201bfecc47f7333e535e927638bb37015a23fd28aa
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:830ffa36a2cc85b79b79c8f68dead7aaec7fb58f5ebad1169970ef47fdf22b5f
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1847,14 +1847,30 @@
|
|
| 1847 |
"loss": 0.395,
|
| 1848 |
"num_input_tokens_seen": 1559137,
|
| 1849 |
"step": 2300
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1850 |
}
|
| 1851 |
],
|
| 1852 |
"logging_steps": 10,
|
| 1853 |
"max_steps": 2795,
|
| 1854 |
-
"num_input_tokens_seen":
|
| 1855 |
"num_train_epochs": 1,
|
| 1856 |
"save_steps": 20,
|
| 1857 |
-
"total_flos": 3.
|
| 1858 |
"train_batch_size": 1,
|
| 1859 |
"trial_name": null,
|
| 1860 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.8300536672629696,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 2320,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1847 |
"loss": 0.395,
|
| 1848 |
"num_input_tokens_seen": 1559137,
|
| 1849 |
"step": 2300
|
| 1850 |
+
},
|
| 1851 |
+
{
|
| 1852 |
+
"epoch": 0.8264758497316637,
|
| 1853 |
+
"grad_norm": 0.25360018014907837,
|
| 1854 |
+
"learning_rate": 3.4704830053667264e-05,
|
| 1855 |
+
"loss": 0.3743,
|
| 1856 |
+
"num_input_tokens_seen": 1565661,
|
| 1857 |
+
"step": 2310
|
| 1858 |
+
},
|
| 1859 |
+
{
|
| 1860 |
+
"epoch": 0.8300536672629696,
|
| 1861 |
+
"grad_norm": 0.27616095542907715,
|
| 1862 |
+
"learning_rate": 3.3989266547406084e-05,
|
| 1863 |
+
"loss": 0.3971,
|
| 1864 |
+
"num_input_tokens_seen": 1572070,
|
| 1865 |
+
"step": 2320
|
| 1866 |
}
|
| 1867 |
],
|
| 1868 |
"logging_steps": 10,
|
| 1869 |
"max_steps": 2795,
|
| 1870 |
+
"num_input_tokens_seen": 1572070,
|
| 1871 |
"num_train_epochs": 1,
|
| 1872 |
"save_steps": 20,
|
| 1873 |
+
"total_flos": 3.535030270457856e+16,
|
| 1874 |
"train_batch_size": 1,
|
| 1875 |
"trial_name": null,
|
| 1876 |
"trial_params": null
|