Training in progress, step 9880, checkpoint
Browse files
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1657
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fec2e4937aab544bc4a7a289c34393ecd56103032bc0d0d26b02e58276b1dbbc
|
| 3 |
size 1657
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14709
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ec370349279813f830889071811ab5efe4b9e7cff1f539189d3d3eccba3be68
|
| 3 |
size 14709
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1465
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:75fb6279d7a90f2eb40ea21d57c0a3b3b9e3c8c93e25ada48d38c03d7fd5446d
|
| 3 |
size 1465
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch":
|
| 6 |
"eval_steps": 500,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -17774,6 +17774,24 @@
|
|
| 17774 |
"mean_token_accuracy": 0.6595763146877289,
|
| 17775 |
"num_tokens": 35996459.0,
|
| 17776 |
"step": 9870
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17777 |
}
|
| 17778 |
],
|
| 17779 |
"logging_steps": 5,
|
|
@@ -17793,7 +17811,7 @@
|
|
| 17793 |
"attributes": {}
|
| 17794 |
}
|
| 17795 |
},
|
| 17796 |
-
"total_flos": 1.
|
| 17797 |
"train_batch_size": 4,
|
| 17798 |
"trial_name": null,
|
| 17799 |
"trial_params": null
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 76.0,
|
| 6 |
"eval_steps": 500,
|
| 7 |
+
"global_step": 9880,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 17774 |
"mean_token_accuracy": 0.6595763146877289,
|
| 17775 |
"num_tokens": 35996459.0,
|
| 17776 |
"step": 9870
|
| 17777 |
+
},
|
| 17778 |
+
{
|
| 17779 |
+
"epoch": 75.96153846153847,
|
| 17780 |
+
"grad_norm": 0.0,
|
| 17781 |
+
"learning_rate": 6.3e-07,
|
| 17782 |
+
"loss": 1.7751,
|
| 17783 |
+
"mean_token_accuracy": 0.6440996766090393,
|
| 17784 |
+
"num_tokens": 36013795.0,
|
| 17785 |
+
"step": 9875
|
| 17786 |
+
},
|
| 17787 |
+
{
|
| 17788 |
+
"epoch": 76.0,
|
| 17789 |
+
"grad_norm": 0.0,
|
| 17790 |
+
"learning_rate": 6.05e-07,
|
| 17791 |
+
"loss": 1.719,
|
| 17792 |
+
"mean_token_accuracy": 0.6536471992731094,
|
| 17793 |
+
"num_tokens": 36031980.0,
|
| 17794 |
+
"step": 9880
|
| 17795 |
}
|
| 17796 |
],
|
| 17797 |
"logging_steps": 5,
|
|
|
|
| 17811 |
"attributes": {}
|
| 17812 |
}
|
| 17813 |
},
|
| 17814 |
+
"total_flos": 1.8271429276203418e+18,
|
| 17815 |
"train_batch_size": 4,
|
| 17816 |
"trial_name": null,
|
| 17817 |
"trial_params": null
|