Training in progress, step 14400, checkpoint
Browse files
last-checkpoint/README.md
CHANGED
|
@@ -1311,6 +1311,10 @@ You can finetune this model on your own dataset.
|
|
| 1311 |
| 0.2492 | 14100 | 0.3853 |
|
| 1312 |
| 0.2500 | 14150 | 0.4068 |
|
| 1313 |
| 0.2509 | 14200 | 0.4071 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1314 |
|
| 1315 |
</details>
|
| 1316 |
|
|
|
|
| 1311 |
| 0.2492 | 14100 | 0.3853 |
|
| 1312 |
| 0.2500 | 14150 | 0.4068 |
|
| 1313 |
| 0.2509 | 14200 | 0.4071 |
|
| 1314 |
+
| 0.2518 | 14250 | 0.3588 |
|
| 1315 |
+
| 0.2527 | 14300 | 0.4161 |
|
| 1316 |
+
| 0.2536 | 14350 | 0.3431 |
|
| 1317 |
+
| 0.2545 | 14400 | 0.3576 |
|
| 1318 |
|
| 1319 |
</details>
|
| 1320 |
|
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 90864192
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c70d1baeacb306a6e3e4e7da1daf6668ef9a846b36b36844f97182da49b7189d
|
| 3 |
size 90864192
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 180609210
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c7cc231dc6a04e0cb1fde6371ac1117217c084c2ff5861f99992f6271b0b2ae
|
| 3 |
size 180609210
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1575cd53f4b8b110d92e8c23aa79a7bcf9dc9c336882843030d5971034763624
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:97511a6060aacb39a58a4067ef60a5d731c4117afaf40a314802e695ac048d50
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:696c7475f37eb1308f8246ec63f84866eb3d506203673b38e04a231cb4eacd6d
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 500,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -1996,6 +1996,34 @@
|
|
| 1996 |
"learning_rate": 4.161905322887829e-05,
|
| 1997 |
"loss": 0.4071,
|
| 1998 |
"step": 14200
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1999 |
}
|
| 2000 |
],
|
| 2001 |
"logging_steps": 50,
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 0.25445742255835735,
|
| 6 |
"eval_steps": 500,
|
| 7 |
+
"global_step": 14400,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 1996 |
"learning_rate": 4.161905322887829e-05,
|
| 1997 |
"loss": 0.4071,
|
| 1998 |
"step": 14200
|
| 1999 |
+
},
|
| 2000 |
+
{
|
| 2001 |
+
"epoch": 0.25180682440670776,
|
| 2002 |
+
"grad_norm": 1.7028473615646362,
|
| 2003 |
+
"learning_rate": 4.156996721053975e-05,
|
| 2004 |
+
"loss": 0.3588,
|
| 2005 |
+
"step": 14250
|
| 2006 |
+
},
|
| 2007 |
+
{
|
| 2008 |
+
"epoch": 0.2526903571239243,
|
| 2009 |
+
"grad_norm": 1.6032434701919556,
|
| 2010 |
+
"learning_rate": 4.1520881192201214e-05,
|
| 2011 |
+
"loss": 0.4161,
|
| 2012 |
+
"step": 14300
|
| 2013 |
+
},
|
| 2014 |
+
{
|
| 2015 |
+
"epoch": 0.2535738898411408,
|
| 2016 |
+
"grad_norm": 1.6103026866912842,
|
| 2017 |
+
"learning_rate": 4.147179517386268e-05,
|
| 2018 |
+
"loss": 0.3431,
|
| 2019 |
+
"step": 14350
|
| 2020 |
+
},
|
| 2021 |
+
{
|
| 2022 |
+
"epoch": 0.25445742255835735,
|
| 2023 |
+
"grad_norm": 3.727078914642334,
|
| 2024 |
+
"learning_rate": 4.142270915552414e-05,
|
| 2025 |
+
"loss": 0.3576,
|
| 2026 |
+
"step": 14400
|
| 2027 |
}
|
| 2028 |
],
|
| 2029 |
"logging_steps": 50,
|