Training in progress, step 220, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6f96b7c71372582451a8856d4292de48642291473692ec963deff5ab833b885c
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d3cd83379ee9a0b75766a6d2fa05de4b0412e196e40c20ae17db8468a2a4f74
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:982ef721154bc6d11d2c93bcbf72ab00515e1bf0940ca62b8577a5bfc9ad31eb
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d1b11a25e8c3f6d77fc413c8d381545b6c097ced995430c3981be2b0781bd184
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -167,14 +167,30 @@
|
|
| 167 |
"loss": 0.4489,
|
| 168 |
"num_input_tokens_seen": 135571,
|
| 169 |
"step": 200
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 170 |
}
|
| 171 |
],
|
| 172 |
"logging_steps": 10,
|
| 173 |
"max_steps": 2795,
|
| 174 |
-
"num_input_tokens_seen":
|
| 175 |
"num_train_epochs": 1,
|
| 176 |
"save_steps": 20,
|
| 177 |
-
"total_flos":
|
| 178 |
"train_batch_size": 1,
|
| 179 |
"trial_name": null,
|
| 180 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.07871198568872988,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 220,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 167 |
"loss": 0.4489,
|
| 168 |
"num_input_tokens_seen": 135571,
|
| 169 |
"step": 200
|
| 170 |
+
},
|
| 171 |
+
{
|
| 172 |
+
"epoch": 0.07513416815742398,
|
| 173 |
+
"grad_norm": 0.5354058742523193,
|
| 174 |
+
"learning_rate": 0.00018497316636851522,
|
| 175 |
+
"loss": 0.4634,
|
| 176 |
+
"num_input_tokens_seen": 141479,
|
| 177 |
+
"step": 210
|
| 178 |
+
},
|
| 179 |
+
{
|
| 180 |
+
"epoch": 0.07871198568872988,
|
| 181 |
+
"grad_norm": 0.4328760504722595,
|
| 182 |
+
"learning_rate": 0.00018425760286225405,
|
| 183 |
+
"loss": 0.4545,
|
| 184 |
+
"num_input_tokens_seen": 147677,
|
| 185 |
+
"step": 220
|
| 186 |
}
|
| 187 |
],
|
| 188 |
"logging_steps": 10,
|
| 189 |
"max_steps": 2795,
|
| 190 |
+
"num_input_tokens_seen": 147677,
|
| 191 |
"num_train_epochs": 1,
|
| 192 |
"save_steps": 20,
|
| 193 |
+
"total_flos": 3320734224623616.0,
|
| 194 |
"train_batch_size": 1,
|
| 195 |
"trial_name": null,
|
| 196 |
"trial_params": null
|