Training in progress, step 2360, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bd0aba29ef3bfaae1191218954a63d90da222aed22aa2200b611cf74f049d222
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d8cd6e4f50429cda27a28378dce5408f87f0a91096366b51efe3dd54bb9f0e0
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3eea0d25f64360cae73a775ba2c6c88f883be49aa9244b8f3d005a3b98fa5466
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:328a8cc6caa8e9e02ff012a126152b40a2427772a86d6a8867f005290455d35c
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1879,14 +1879,30 @@
|
|
| 1879 |
"loss": 0.362,
|
| 1880 |
"num_input_tokens_seen": 1586863,
|
| 1881 |
"step": 2340
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1882 |
}
|
| 1883 |
],
|
| 1884 |
"logging_steps": 10,
|
| 1885 |
"max_steps": 2795,
|
| 1886 |
-
"num_input_tokens_seen":
|
| 1887 |
"num_train_epochs": 1,
|
| 1888 |
"save_steps": 20,
|
| 1889 |
-
"total_flos": 3.
|
| 1890 |
"train_batch_size": 1,
|
| 1891 |
"trial_name": null,
|
| 1892 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.8443649373881932,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 2360,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1879 |
"loss": 0.362,
|
| 1880 |
"num_input_tokens_seen": 1586863,
|
| 1881 |
"step": 2340
|
| 1882 |
+
},
|
| 1883 |
+
{
|
| 1884 |
+
"epoch": 0.8407871198568873,
|
| 1885 |
+
"grad_norm": 0.3128429353237152,
|
| 1886 |
+
"learning_rate": 3.184257602862254e-05,
|
| 1887 |
+
"loss": 0.373,
|
| 1888 |
+
"num_input_tokens_seen": 1593227,
|
| 1889 |
+
"step": 2350
|
| 1890 |
+
},
|
| 1891 |
+
{
|
| 1892 |
+
"epoch": 0.8443649373881932,
|
| 1893 |
+
"grad_norm": 0.3150210678577423,
|
| 1894 |
+
"learning_rate": 3.112701252236136e-05,
|
| 1895 |
+
"loss": 0.3722,
|
| 1896 |
+
"num_input_tokens_seen": 1601444,
|
| 1897 |
+
"step": 2360
|
| 1898 |
}
|
| 1899 |
],
|
| 1900 |
"logging_steps": 10,
|
| 1901 |
"max_steps": 2795,
|
| 1902 |
+
"num_input_tokens_seen": 1601444,
|
| 1903 |
"num_train_epochs": 1,
|
| 1904 |
"save_steps": 20,
|
| 1905 |
+
"total_flos": 3.601082023346995e+16,
|
| 1906 |
"train_batch_size": 1,
|
| 1907 |
"trial_name": null,
|
| 1908 |
"trial_params": null
|