Training in progress, step 1640, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f10257b7014e9aac14048d4c25d8edac512b0d9b68ecd3b472b859543cdef14b
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b9380c503afe8a570e3b4d2d7682a1be66462065c80d82ec8ac42553b0ce37b6
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e4bae246be55e0c5c479cc3228ea806f5dad96bfbd997b22379ceb5166207ba
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a7b5d3fd1283600a406407d024141f0dd9978a7bb53ae522b56311cc64308157
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1303,14 +1303,30 @@
|
|
| 1303 |
"loss": 0.4022,
|
| 1304 |
"num_input_tokens_seen": 1087604,
|
| 1305 |
"step": 1620
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1306 |
}
|
| 1307 |
],
|
| 1308 |
"logging_steps": 10,
|
| 1309 |
"max_steps": 2795,
|
| 1310 |
-
"num_input_tokens_seen":
|
| 1311 |
"num_train_epochs": 1,
|
| 1312 |
"save_steps": 20,
|
| 1313 |
-
"total_flos": 2.
|
| 1314 |
"train_batch_size": 1,
|
| 1315 |
"trial_name": null,
|
| 1316 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.5867620751341681,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 1640,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1303 |
"loss": 0.4022,
|
| 1304 |
"num_input_tokens_seen": 1087604,
|
| 1305 |
"step": 1620
|
| 1306 |
+
},
|
| 1307 |
+
{
|
| 1308 |
+
"epoch": 0.5831842576028623,
|
| 1309 |
+
"grad_norm": 0.36473724246025085,
|
| 1310 |
+
"learning_rate": 8.336314847942756e-05,
|
| 1311 |
+
"loss": 0.3857,
|
| 1312 |
+
"num_input_tokens_seen": 1094532,
|
| 1313 |
+
"step": 1630
|
| 1314 |
+
},
|
| 1315 |
+
{
|
| 1316 |
+
"epoch": 0.5867620751341681,
|
| 1317 |
+
"grad_norm": 0.29369673132896423,
|
| 1318 |
+
"learning_rate": 8.264758497316637e-05,
|
| 1319 |
+
"loss": 0.405,
|
| 1320 |
+
"num_input_tokens_seen": 1100973,
|
| 1321 |
+
"step": 1640
|
| 1322 |
}
|
| 1323 |
],
|
| 1324 |
"logging_steps": 10,
|
| 1325 |
"max_steps": 2795,
|
| 1326 |
+
"num_input_tokens_seen": 1100973,
|
| 1327 |
"num_train_epochs": 1,
|
| 1328 |
"save_steps": 20,
|
| 1329 |
+
"total_flos": 2.4756994802755584e+16,
|
| 1330 |
"train_batch_size": 1,
|
| 1331 |
"trial_name": null,
|
| 1332 |
"trial_params": null
|