Training in progress, step 2660, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ac4c2e27c7954ba8f32458e86bce305764577f448e1ef7b05942fdf1638b8d3
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:281f3e1e3de94056c7322594abe70311fe1bd96ae510bce52c2cda2200510403
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:61c76e152102faadd3e0723a7b1e98668fc77dc3b57498a40931f275e4f17e53
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ae0efadb0abbc637034bafcf665143567ca2a6d58001c0ddda8fd674ac11d7db
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -2119,14 +2119,30 @@
|
|
| 2119 |
"loss": 0.3877,
|
| 2120 |
"num_input_tokens_seen": 1790778,
|
| 2121 |
"step": 2640
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2122 |
}
|
| 2123 |
],
|
| 2124 |
"logging_steps": 10,
|
| 2125 |
"max_steps": 2795,
|
| 2126 |
-
"num_input_tokens_seen":
|
| 2127 |
"num_train_epochs": 1,
|
| 2128 |
"save_steps": 20,
|
| 2129 |
-
"total_flos": 4.
|
| 2130 |
"train_batch_size": 1,
|
| 2131 |
"trial_name": null,
|
| 2132 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.9516994633273703,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 2660,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 2119 |
"loss": 0.3877,
|
| 2120 |
"num_input_tokens_seen": 1790778,
|
| 2121 |
"step": 2640
|
| 2122 |
+
},
|
| 2123 |
+
{
|
| 2124 |
+
"epoch": 0.9481216457960644,
|
| 2125 |
+
"grad_norm": 0.3405713140964508,
|
| 2126 |
+
"learning_rate": 1.037567084078712e-05,
|
| 2127 |
+
"loss": 0.3674,
|
| 2128 |
+
"num_input_tokens_seen": 1797992,
|
| 2129 |
+
"step": 2650
|
| 2130 |
+
},
|
| 2131 |
+
{
|
| 2132 |
+
"epoch": 0.9516994633273703,
|
| 2133 |
+
"grad_norm": 0.4525390565395355,
|
| 2134 |
+
"learning_rate": 9.660107334525939e-06,
|
| 2135 |
+
"loss": 0.376,
|
| 2136 |
+
"num_input_tokens_seen": 1804337,
|
| 2137 |
+
"step": 2660
|
| 2138 |
}
|
| 2139 |
],
|
| 2140 |
"logging_steps": 10,
|
| 2141 |
"max_steps": 2795,
|
| 2142 |
+
"num_input_tokens_seen": 1804337,
|
| 2143 |
"num_train_epochs": 1,
|
| 2144 |
"save_steps": 20,
|
| 2145 |
+
"total_flos": 4.05731673087529e+16,
|
| 2146 |
"train_batch_size": 1,
|
| 2147 |
"trial_name": null,
|
| 2148 |
"trial_params": null
|