Training in progress, step 9840, checkpoint
Browse files
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1657
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:52fa49c9aee53eccacfc20fa0a53e81c66a2d9d79fadbf39a861f03d505a8dc4
|
| 3 |
size 1657
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14709
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c16a6ae3dbdee2c3d74d8c96a643a478a246c38052a582fabe164c53a454d5b0
|
| 3 |
size 14709
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1465
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c7f8d076674adacf33d6e96dcffeac08e85cb7bfeb9c3481b82797b68933114a
|
| 3 |
size 1465
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch": 75.
|
| 6 |
"eval_steps": 500,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -17702,6 +17702,24 @@
|
|
| 17702 |
"mean_token_accuracy": 0.6422079712152481,
|
| 17703 |
"num_tokens": 35850712.0,
|
| 17704 |
"step": 9830
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17705 |
}
|
| 17706 |
],
|
| 17707 |
"logging_steps": 5,
|
|
@@ -17721,7 +17739,7 @@
|
|
| 17721 |
"attributes": {}
|
| 17722 |
}
|
| 17723 |
},
|
| 17724 |
-
"total_flos": 1.
|
| 17725 |
"train_batch_size": 4,
|
| 17726 |
"trial_name": null,
|
| 17727 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 75.6923076923077,
|
| 6 |
"eval_steps": 500,
|
| 7 |
+
"global_step": 9840,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 17702 |
"mean_token_accuracy": 0.6422079712152481,
|
| 17703 |
"num_tokens": 35850712.0,
|
| 17704 |
"step": 9830
|
| 17705 |
+
},
|
| 17706 |
+
{
|
| 17707 |
+
"epoch": 75.65384615384616,
|
| 17708 |
+
"grad_norm": 0.0,
|
| 17709 |
+
"learning_rate": 8.300000000000001e-07,
|
| 17710 |
+
"loss": 1.7786,
|
| 17711 |
+
"mean_token_accuracy": 0.6427565932273864,
|
| 17712 |
+
"num_tokens": 35868743.0,
|
| 17713 |
+
"step": 9835
|
| 17714 |
+
},
|
| 17715 |
+
{
|
| 17716 |
+
"epoch": 75.6923076923077,
|
| 17717 |
+
"grad_norm": 0.0,
|
| 17718 |
+
"learning_rate": 8.05e-07,
|
| 17719 |
+
"loss": 1.7353,
|
| 17720 |
+
"mean_token_accuracy": 0.657978093624115,
|
| 17721 |
+
"num_tokens": 35886399.0,
|
| 17722 |
+
"step": 9840
|
| 17723 |
}
|
| 17724 |
],
|
| 17725 |
"logging_steps": 5,
|
|
|
|
| 17739 |
"attributes": {}
|
| 17740 |
}
|
| 17741 |
},
|
| 17742 |
+
"total_flos": 1.8198268608990413e+18,
|
| 17743 |
"train_batch_size": 4,
|
| 17744 |
"trial_name": null,
|
| 17745 |
"trial_params": null
|