Training in progress, step 2280, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b4d0e6f6e105ca8708dd58c58cac0d38ef3d7e182e4118c1d7c3bbbb4cfbb8e5
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fe2060f54f8854e34b115c7210a27ece708af54ff47a9aed67412ee4f2d4639
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1817c09855fd07da9d0e441c81f4b0b9ce05c1e1787ca7dcb5d996cd39c530fd
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:36cd02824fbe35ffd9c39b1765aeeef980a727adc03234507c83ce617d1b746f
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1815,14 +1815,30 @@
|
|
| 1815 |
"loss": 0.369,
|
| 1816 |
"num_input_tokens_seen": 1530483,
|
| 1817 |
"step": 2260
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1818 |
}
|
| 1819 |
],
|
| 1820 |
"logging_steps": 10,
|
| 1821 |
"max_steps": 2795,
|
| 1822 |
-
"num_input_tokens_seen":
|
| 1823 |
"num_train_epochs": 1,
|
| 1824 |
"save_steps": 20,
|
| 1825 |
-
"total_flos": 3.
|
| 1826 |
"train_batch_size": 1,
|
| 1827 |
"trial_name": null,
|
| 1828 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.815742397137746,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 2280,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1815 |
"loss": 0.369,
|
| 1816 |
"num_input_tokens_seen": 1530483,
|
| 1817 |
"step": 2260
|
| 1818 |
+
},
|
| 1819 |
+
{
|
| 1820 |
+
"epoch": 0.8121645796064401,
|
| 1821 |
+
"grad_norm": 0.31918245553970337,
|
| 1822 |
+
"learning_rate": 3.756708407871199e-05,
|
| 1823 |
+
"loss": 0.4161,
|
| 1824 |
+
"num_input_tokens_seen": 1538556,
|
| 1825 |
+
"step": 2270
|
| 1826 |
+
},
|
| 1827 |
+
{
|
| 1828 |
+
"epoch": 0.815742397137746,
|
| 1829 |
+
"grad_norm": 0.36853113770484924,
|
| 1830 |
+
"learning_rate": 3.685152057245081e-05,
|
| 1831 |
+
"loss": 0.3834,
|
| 1832 |
+
"num_input_tokens_seen": 1545040,
|
| 1833 |
+
"step": 2280
|
| 1834 |
}
|
| 1835 |
],
|
| 1836 |
"logging_steps": 10,
|
| 1837 |
"max_steps": 2795,
|
| 1838 |
+
"num_input_tokens_seen": 1545040,
|
| 1839 |
"num_train_epochs": 1,
|
| 1840 |
"save_steps": 20,
|
| 1841 |
+
"total_flos": 3.474249345810432e+16,
|
| 1842 |
"train_batch_size": 1,
|
| 1843 |
"trial_name": null,
|
| 1844 |
"trial_params": null
|