Training in progress, step 280, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e4cede093a047dae1d9b39f98f01ebd38cf9627aaf8ecf1e2945971471563b4
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f858b048df2c4990c74df403bfc1d645b453dacbbe71b54093671cdea7cfc9f
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:155906040ca3a73448d35f864001cf3a30094785911dc4949af7556c64d5da61
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aba7537afe16c3362d871f59e72c1a9387d8f379b5c058726a66f4dd3222a4f6
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -215,14 +215,30 @@
|
|
| 215 |
"loss": 0.4297,
|
| 216 |
"num_input_tokens_seen": 174828,
|
| 217 |
"step": 260
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 218 |
}
|
| 219 |
],
|
| 220 |
"logging_steps": 10,
|
| 221 |
"max_steps": 2795,
|
| 222 |
-
"num_input_tokens_seen":
|
| 223 |
"num_train_epochs": 1,
|
| 224 |
"save_steps": 20,
|
| 225 |
-
"total_flos":
|
| 226 |
"train_batch_size": 1,
|
| 227 |
"trial_name": null,
|
| 228 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.1001788908765653,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 280,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 215 |
"loss": 0.4297,
|
| 216 |
"num_input_tokens_seen": 174828,
|
| 217 |
"step": 260
|
| 218 |
+
},
|
| 219 |
+
{
|
| 220 |
+
"epoch": 0.09660107334525939,
|
| 221 |
+
"grad_norm": 0.37481924891471863,
|
| 222 |
+
"learning_rate": 0.00018067978533094814,
|
| 223 |
+
"loss": 0.4314,
|
| 224 |
+
"num_input_tokens_seen": 181578,
|
| 225 |
+
"step": 270
|
| 226 |
+
},
|
| 227 |
+
{
|
| 228 |
+
"epoch": 0.1001788908765653,
|
| 229 |
+
"grad_norm": 0.28545519709587097,
|
| 230 |
+
"learning_rate": 0.00017996422182468694,
|
| 231 |
+
"loss": 0.4332,
|
| 232 |
+
"num_input_tokens_seen": 187842,
|
| 233 |
+
"step": 280
|
| 234 |
}
|
| 235 |
],
|
| 236 |
"logging_steps": 10,
|
| 237 |
"max_steps": 2795,
|
| 238 |
+
"num_input_tokens_seen": 187842,
|
| 239 |
"num_train_epochs": 1,
|
| 240 |
"save_steps": 20,
|
| 241 |
+
"total_flos": 4223903236263936.0,
|
| 242 |
"train_batch_size": 1,
|
| 243 |
"trial_name": null,
|
| 244 |
"trial_params": null
|