Training in progress, step 160, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2ded6ea93dca5b528b272b56d0dba78bfb48686ef712e6aee97e12ad754ec51a
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f7fcfe6bdcb982add10fed5d7c4599c6d6b8da496a1bdce1e099c1f92ef34189
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6e40ba6383d47c8ec509ed3abd3668fd616332aaf4bb58238305f857b037e170
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7815a0f425c279b798b8a01de0a05716135e29d85893e062dbb79cc33f526eb8
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -119,14 +119,30 @@
|
|
| 119 |
"loss": 0.456,
|
| 120 |
"num_input_tokens_seen": 96192,
|
| 121 |
"step": 140
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 122 |
}
|
| 123 |
],
|
| 124 |
"logging_steps": 10,
|
| 125 |
"max_steps": 2795,
|
| 126 |
-
"num_input_tokens_seen":
|
| 127 |
"num_train_epochs": 1,
|
| 128 |
"save_steps": 20,
|
| 129 |
-
"total_flos":
|
| 130 |
"train_batch_size": 1,
|
| 131 |
"trial_name": null,
|
| 132 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.057245080500894455,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 160,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 119 |
"loss": 0.456,
|
| 120 |
"num_input_tokens_seen": 96192,
|
| 121 |
"step": 140
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 0.05366726296958855,
|
| 125 |
+
"grad_norm": 0.4566495418548584,
|
| 126 |
+
"learning_rate": 0.0001892665474060823,
|
| 127 |
+
"loss": 0.429,
|
| 128 |
+
"num_input_tokens_seen": 101609,
|
| 129 |
+
"step": 150
|
| 130 |
+
},
|
| 131 |
+
{
|
| 132 |
+
"epoch": 0.057245080500894455,
|
| 133 |
+
"grad_norm": 0.4657338559627533,
|
| 134 |
+
"learning_rate": 0.0001885509838998211,
|
| 135 |
+
"loss": 0.4445,
|
| 136 |
+
"num_input_tokens_seen": 107467,
|
| 137 |
+
"step": 160
|
| 138 |
}
|
| 139 |
],
|
| 140 |
"logging_steps": 10,
|
| 141 |
"max_steps": 2795,
|
| 142 |
+
"num_input_tokens_seen": 107467,
|
| 143 |
"num_train_epochs": 1,
|
| 144 |
"save_steps": 20,
|
| 145 |
+
"total_flos": 2416553321895936.0,
|
| 146 |
"train_batch_size": 1,
|
| 147 |
"trial_name": null,
|
| 148 |
"trial_params": null
|