Training in progress, step 2520, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6346a70cf6093679d4d5b93fd3421bc7a3e8ca9f233962b574cc1e605da07a63
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2970350d5affaa1384aa18b2fcabe3041d0121f55ef78024ff71924e2b9bd97d
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9eccae15e2cae4c77b5aefc4760075dc2fa9e695d63c8bcac7425551f60d749f
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8a5c37c503fee750e3269106b2ec8e747f83501e7f65e6d8e35bc17a08fbb85e
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -2007,14 +2007,30 @@
|
|
| 2007 |
"loss": 0.391,
|
| 2008 |
"num_input_tokens_seen": 1694486,
|
| 2009 |
"step": 2500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2010 |
}
|
| 2011 |
],
|
| 2012 |
"logging_steps": 10,
|
| 2013 |
"max_steps": 2795,
|
| 2014 |
-
"num_input_tokens_seen":
|
| 2015 |
"num_train_epochs": 1,
|
| 2016 |
"save_steps": 20,
|
| 2017 |
-
"total_flos": 3.
|
| 2018 |
"train_batch_size": 1,
|
| 2019 |
"trial_name": null,
|
| 2020 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.9016100178890877,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 2520,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 2007 |
"loss": 0.391,
|
| 2008 |
"num_input_tokens_seen": 1694486,
|
| 2009 |
"step": 2500
|
| 2010 |
+
},
|
| 2011 |
+
{
|
| 2012 |
+
"epoch": 0.8980322003577818,
|
| 2013 |
+
"grad_norm": 0.2778749167919159,
|
| 2014 |
+
"learning_rate": 2.039355992844365e-05,
|
| 2015 |
+
"loss": 0.3755,
|
| 2016 |
+
"num_input_tokens_seen": 1701728,
|
| 2017 |
+
"step": 2510
|
| 2018 |
+
},
|
| 2019 |
+
{
|
| 2020 |
+
"epoch": 0.9016100178890877,
|
| 2021 |
+
"grad_norm": 0.45739054679870605,
|
| 2022 |
+
"learning_rate": 1.967799642218247e-05,
|
| 2023 |
+
"loss": 0.3857,
|
| 2024 |
+
"num_input_tokens_seen": 1707744,
|
| 2025 |
+
"step": 2520
|
| 2026 |
}
|
| 2027 |
],
|
| 2028 |
"logging_steps": 10,
|
| 2029 |
"max_steps": 2795,
|
| 2030 |
+
"num_input_tokens_seen": 1707744,
|
| 2031 |
"num_train_epochs": 1,
|
| 2032 |
"save_steps": 20,
|
| 2033 |
+
"total_flos": 3.840113184650035e+16,
|
| 2034 |
"train_batch_size": 1,
|
| 2035 |
"trial_name": null,
|
| 2036 |
"trial_params": null
|