Training in progress, step 2380, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3595b12f86459195f8d09179d7c0dbf8792ab5da530439f3a09a0bba7f7a42a1
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:58fbfe5c5eaca18bcc1c80efd605fd210e276ac78ec3e25b0dac1bfe7a2ee3d1
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c99e1271397b81fc889fdb77bb7999277542aa1b382e9e22a5a0bd0cd0a478e3
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d5e00b537e62e61b7fa13230cc02b759e6748bf2bcdb20f27dd8576bd1d7344
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1895,14 +1895,30 @@
|
|
| 1895 |
"loss": 0.3722,
|
| 1896 |
"num_input_tokens_seen": 1601444,
|
| 1897 |
"step": 2360
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1898 |
}
|
| 1899 |
],
|
| 1900 |
"logging_steps": 10,
|
| 1901 |
"max_steps": 2795,
|
| 1902 |
-
"num_input_tokens_seen":
|
| 1903 |
"num_train_epochs": 1,
|
| 1904 |
"save_steps": 20,
|
| 1905 |
-
"total_flos": 3.
|
| 1906 |
"train_batch_size": 1,
|
| 1907 |
"trial_name": null,
|
| 1908 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.851520572450805,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 2380,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1895 |
"loss": 0.3722,
|
| 1896 |
"num_input_tokens_seen": 1601444,
|
| 1897 |
"step": 2360
|
| 1898 |
+
},
|
| 1899 |
+
{
|
| 1900 |
+
"epoch": 0.8479427549194991,
|
| 1901 |
+
"grad_norm": 0.44683215022087097,
|
| 1902 |
+
"learning_rate": 3.041144901610018e-05,
|
| 1903 |
+
"loss": 0.4009,
|
| 1904 |
+
"num_input_tokens_seen": 1607535,
|
| 1905 |
+
"step": 2370
|
| 1906 |
+
},
|
| 1907 |
+
{
|
| 1908 |
+
"epoch": 0.851520572450805,
|
| 1909 |
+
"grad_norm": 0.3482876121997833,
|
| 1910 |
+
"learning_rate": 2.9695885509839e-05,
|
| 1911 |
+
"loss": 0.3792,
|
| 1912 |
+
"num_input_tokens_seen": 1615204,
|
| 1913 |
+
"step": 2380
|
| 1914 |
}
|
| 1915 |
],
|
| 1916 |
"logging_steps": 10,
|
| 1917 |
"max_steps": 2795,
|
| 1918 |
+
"num_input_tokens_seen": 1615204,
|
| 1919 |
"num_train_epochs": 1,
|
| 1920 |
"save_steps": 20,
|
| 1921 |
+
"total_flos": 3.632023404151603e+16,
|
| 1922 |
"train_batch_size": 1,
|
| 1923 |
"trial_name": null,
|
| 1924 |
"trial_params": null
|