Training in progress, step 978, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100059752
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2af79b87cdba2f6ee09202e5806579a3a64b21691a917cde1ab005a67018b158
|
| 3 |
size 100059752
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 51244404
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:90ad1ea0daa57f559d92fd099a3ead6026fc5b03b3c5d094a521dbf126d1985a
|
| 3 |
size 51244404
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b53b526ca46b50c78b2d293051395fcfc3faf741ae9849365dcec0d68cfe9398
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe6ddabd32cdc41f4c81c1d53be4d94bd11320d8500ca8b0b69995845949f304
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 326,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -4595,6 +4595,2296 @@
|
|
| 4595 |
"eval_samples_per_second": 27.784,
|
| 4596 |
"eval_steps_per_second": 13.917,
|
| 4597 |
"step": 652
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4598 |
}
|
| 4599 |
],
|
| 4600 |
"logging_steps": 1,
|
|
@@ -4614,7 +6904,7 @@
|
|
| 4614 |
"attributes": {}
|
| 4615 |
}
|
| 4616 |
},
|
| 4617 |
-
"total_flos": 1.
|
| 4618 |
"train_batch_size": 2,
|
| 4619 |
"trial_name": null,
|
| 4620 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.751007870992513,
|
| 5 |
"eval_steps": 326,
|
| 6 |
+
"global_step": 978,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 4595 |
"eval_samples_per_second": 27.784,
|
| 4596 |
"eval_steps_per_second": 13.917,
|
| 4597 |
"step": 652
|
| 4598 |
+
},
|
| 4599 |
+
{
|
| 4600 |
+
"epoch": 0.50143981570359,
|
| 4601 |
+
"grad_norm": 1.54447603225708,
|
| 4602 |
+
"learning_rate": 0.00010085038219340393,
|
| 4603 |
+
"loss": 1.296,
|
| 4604 |
+
"step": 653
|
| 4605 |
+
},
|
| 4606 |
+
{
|
| 4607 |
+
"epoch": 0.5022077174121713,
|
| 4608 |
+
"grad_norm": 1.8366080522537231,
|
| 4609 |
+
"learning_rate": 0.00010060741943827776,
|
| 4610 |
+
"loss": 1.6111,
|
| 4611 |
+
"step": 654
|
| 4612 |
+
},
|
| 4613 |
+
{
|
| 4614 |
+
"epoch": 0.5029756191207525,
|
| 4615 |
+
"grad_norm": 1.22926926612854,
|
| 4616 |
+
"learning_rate": 0.00010036445309730944,
|
| 4617 |
+
"loss": 1.0101,
|
| 4618 |
+
"step": 655
|
| 4619 |
+
},
|
| 4620 |
+
{
|
| 4621 |
+
"epoch": 0.5037435208293338,
|
| 4622 |
+
"grad_norm": 1.0615041255950928,
|
| 4623 |
+
"learning_rate": 0.00010012148460482738,
|
| 4624 |
+
"loss": 0.8333,
|
| 4625 |
+
"step": 656
|
| 4626 |
+
},
|
| 4627 |
+
{
|
| 4628 |
+
"epoch": 0.5045114225379151,
|
| 4629 |
+
"grad_norm": 1.7094695568084717,
|
| 4630 |
+
"learning_rate": 9.987851539517262e-05,
|
| 4631 |
+
"loss": 1.094,
|
| 4632 |
+
"step": 657
|
| 4633 |
+
},
|
| 4634 |
+
{
|
| 4635 |
+
"epoch": 0.5052793242464965,
|
| 4636 |
+
"grad_norm": 1.4576027393341064,
|
| 4637 |
+
"learning_rate": 9.963554690269058e-05,
|
| 4638 |
+
"loss": 0.957,
|
| 4639 |
+
"step": 658
|
| 4640 |
+
},
|
| 4641 |
+
{
|
| 4642 |
+
"epoch": 0.5060472259550778,
|
| 4643 |
+
"grad_norm": 1.4120879173278809,
|
| 4644 |
+
"learning_rate": 9.939258056172225e-05,
|
| 4645 |
+
"loss": 1.1924,
|
| 4646 |
+
"step": 659
|
| 4647 |
+
},
|
| 4648 |
+
{
|
| 4649 |
+
"epoch": 0.5068151276636591,
|
| 4650 |
+
"grad_norm": 1.4560344219207764,
|
| 4651 |
+
"learning_rate": 9.914961780659609e-05,
|
| 4652 |
+
"loss": 1.0673,
|
| 4653 |
+
"step": 660
|
| 4654 |
+
},
|
| 4655 |
+
{
|
| 4656 |
+
"epoch": 0.5075830293722403,
|
| 4657 |
+
"grad_norm": 1.4893819093704224,
|
| 4658 |
+
"learning_rate": 9.890666007161929e-05,
|
| 4659 |
+
"loss": 1.1459,
|
| 4660 |
+
"step": 661
|
| 4661 |
+
},
|
| 4662 |
+
{
|
| 4663 |
+
"epoch": 0.5083509310808216,
|
| 4664 |
+
"grad_norm": 1.5299136638641357,
|
| 4665 |
+
"learning_rate": 9.866370879106947e-05,
|
| 4666 |
+
"loss": 1.2983,
|
| 4667 |
+
"step": 662
|
| 4668 |
+
},
|
| 4669 |
+
{
|
| 4670 |
+
"epoch": 0.509118832789403,
|
| 4671 |
+
"grad_norm": 1.6000056266784668,
|
| 4672 |
+
"learning_rate": 9.84207653991861e-05,
|
| 4673 |
+
"loss": 1.0902,
|
| 4674 |
+
"step": 663
|
| 4675 |
+
},
|
| 4676 |
+
{
|
| 4677 |
+
"epoch": 0.5098867344979843,
|
| 4678 |
+
"grad_norm": 1.5117030143737793,
|
| 4679 |
+
"learning_rate": 9.817783133016206e-05,
|
| 4680 |
+
"loss": 1.2475,
|
| 4681 |
+
"step": 664
|
| 4682 |
+
},
|
| 4683 |
+
{
|
| 4684 |
+
"epoch": 0.5106546362065656,
|
| 4685 |
+
"grad_norm": 1.7616103887557983,
|
| 4686 |
+
"learning_rate": 9.793490801813528e-05,
|
| 4687 |
+
"loss": 1.4153,
|
| 4688 |
+
"step": 665
|
| 4689 |
+
},
|
| 4690 |
+
{
|
| 4691 |
+
"epoch": 0.5114225379151469,
|
| 4692 |
+
"grad_norm": 1.6097863912582397,
|
| 4693 |
+
"learning_rate": 9.769199689718009e-05,
|
| 4694 |
+
"loss": 1.3362,
|
| 4695 |
+
"step": 666
|
| 4696 |
+
},
|
| 4697 |
+
{
|
| 4698 |
+
"epoch": 0.5121904396237281,
|
| 4699 |
+
"grad_norm": 1.431277871131897,
|
| 4700 |
+
"learning_rate": 9.744909940129895e-05,
|
| 4701 |
+
"loss": 1.2254,
|
| 4702 |
+
"step": 667
|
| 4703 |
+
},
|
| 4704 |
+
{
|
| 4705 |
+
"epoch": 0.5129583413323094,
|
| 4706 |
+
"grad_norm": 1.4436259269714355,
|
| 4707 |
+
"learning_rate": 9.720621696441378e-05,
|
| 4708 |
+
"loss": 1.1335,
|
| 4709 |
+
"step": 668
|
| 4710 |
+
},
|
| 4711 |
+
{
|
| 4712 |
+
"epoch": 0.5137262430408908,
|
| 4713 |
+
"grad_norm": 1.7994132041931152,
|
| 4714 |
+
"learning_rate": 9.69633510203577e-05,
|
| 4715 |
+
"loss": 1.3912,
|
| 4716 |
+
"step": 669
|
| 4717 |
+
},
|
| 4718 |
+
{
|
| 4719 |
+
"epoch": 0.5144941447494721,
|
| 4720 |
+
"grad_norm": 1.828708529472351,
|
| 4721 |
+
"learning_rate": 9.672050300286636e-05,
|
| 4722 |
+
"loss": 1.5837,
|
| 4723 |
+
"step": 670
|
| 4724 |
+
},
|
| 4725 |
+
{
|
| 4726 |
+
"epoch": 0.5152620464580534,
|
| 4727 |
+
"grad_norm": 1.592246651649475,
|
| 4728 |
+
"learning_rate": 9.64776743455697e-05,
|
| 4729 |
+
"loss": 1.3176,
|
| 4730 |
+
"step": 671
|
| 4731 |
+
},
|
| 4732 |
+
{
|
| 4733 |
+
"epoch": 0.5160299481666347,
|
| 4734 |
+
"grad_norm": 1.7357617616653442,
|
| 4735 |
+
"learning_rate": 9.623486648198326e-05,
|
| 4736 |
+
"loss": 1.47,
|
| 4737 |
+
"step": 672
|
| 4738 |
+
},
|
| 4739 |
+
{
|
| 4740 |
+
"epoch": 0.5167978498752159,
|
| 4741 |
+
"grad_norm": 1.6944161653518677,
|
| 4742 |
+
"learning_rate": 9.599208084549993e-05,
|
| 4743 |
+
"loss": 1.4586,
|
| 4744 |
+
"step": 673
|
| 4745 |
+
},
|
| 4746 |
+
{
|
| 4747 |
+
"epoch": 0.5175657515837973,
|
| 4748 |
+
"grad_norm": 1.827109456062317,
|
| 4749 |
+
"learning_rate": 9.574931886938128e-05,
|
| 4750 |
+
"loss": 1.2451,
|
| 4751 |
+
"step": 674
|
| 4752 |
+
},
|
| 4753 |
+
{
|
| 4754 |
+
"epoch": 0.5183336532923786,
|
| 4755 |
+
"grad_norm": 1.3678934574127197,
|
| 4756 |
+
"learning_rate": 9.550658198674931e-05,
|
| 4757 |
+
"loss": 1.2403,
|
| 4758 |
+
"step": 675
|
| 4759 |
+
},
|
| 4760 |
+
{
|
| 4761 |
+
"epoch": 0.5191015550009599,
|
| 4762 |
+
"grad_norm": 0.994289755821228,
|
| 4763 |
+
"learning_rate": 9.526387163057777e-05,
|
| 4764 |
+
"loss": 0.8628,
|
| 4765 |
+
"step": 676
|
| 4766 |
+
},
|
| 4767 |
+
{
|
| 4768 |
+
"epoch": 0.5198694567095412,
|
| 4769 |
+
"grad_norm": 1.3589410781860352,
|
| 4770 |
+
"learning_rate": 9.502118923368388e-05,
|
| 4771 |
+
"loss": 0.8397,
|
| 4772 |
+
"step": 677
|
| 4773 |
+
},
|
| 4774 |
+
{
|
| 4775 |
+
"epoch": 0.5206373584181225,
|
| 4776 |
+
"grad_norm": 1.3971327543258667,
|
| 4777 |
+
"learning_rate": 9.477853622871984e-05,
|
| 4778 |
+
"loss": 1.1653,
|
| 4779 |
+
"step": 678
|
| 4780 |
+
},
|
| 4781 |
+
{
|
| 4782 |
+
"epoch": 0.5214052601267037,
|
| 4783 |
+
"grad_norm": 1.3670016527175903,
|
| 4784 |
+
"learning_rate": 9.453591404816423e-05,
|
| 4785 |
+
"loss": 1.0544,
|
| 4786 |
+
"step": 679
|
| 4787 |
+
},
|
| 4788 |
+
{
|
| 4789 |
+
"epoch": 0.5221731618352851,
|
| 4790 |
+
"grad_norm": 1.3910092115402222,
|
| 4791 |
+
"learning_rate": 9.429332412431377e-05,
|
| 4792 |
+
"loss": 1.2027,
|
| 4793 |
+
"step": 680
|
| 4794 |
+
},
|
| 4795 |
+
{
|
| 4796 |
+
"epoch": 0.5229410635438664,
|
| 4797 |
+
"grad_norm": 1.4565633535385132,
|
| 4798 |
+
"learning_rate": 9.405076788927469e-05,
|
| 4799 |
+
"loss": 1.3962,
|
| 4800 |
+
"step": 681
|
| 4801 |
+
},
|
| 4802 |
+
{
|
| 4803 |
+
"epoch": 0.5237089652524477,
|
| 4804 |
+
"grad_norm": 1.4770212173461914,
|
| 4805 |
+
"learning_rate": 9.380824677495441e-05,
|
| 4806 |
+
"loss": 1.3789,
|
| 4807 |
+
"step": 682
|
| 4808 |
+
},
|
| 4809 |
+
{
|
| 4810 |
+
"epoch": 0.524476866961029,
|
| 4811 |
+
"grad_norm": 1.5948761701583862,
|
| 4812 |
+
"learning_rate": 9.356576221305289e-05,
|
| 4813 |
+
"loss": 1.4457,
|
| 4814 |
+
"step": 683
|
| 4815 |
+
},
|
| 4816 |
+
{
|
| 4817 |
+
"epoch": 0.5252447686696103,
|
| 4818 |
+
"grad_norm": 1.4264615774154663,
|
| 4819 |
+
"learning_rate": 9.332331563505444e-05,
|
| 4820 |
+
"loss": 1.0041,
|
| 4821 |
+
"step": 684
|
| 4822 |
+
},
|
| 4823 |
+
{
|
| 4824 |
+
"epoch": 0.5260126703781915,
|
| 4825 |
+
"grad_norm": 1.5794204473495483,
|
| 4826 |
+
"learning_rate": 9.308090847221905e-05,
|
| 4827 |
+
"loss": 1.3674,
|
| 4828 |
+
"step": 685
|
| 4829 |
+
},
|
| 4830 |
+
{
|
| 4831 |
+
"epoch": 0.5267805720867729,
|
| 4832 |
+
"grad_norm": 1.5819778442382812,
|
| 4833 |
+
"learning_rate": 9.283854215557409e-05,
|
| 4834 |
+
"loss": 1.5201,
|
| 4835 |
+
"step": 686
|
| 4836 |
+
},
|
| 4837 |
+
{
|
| 4838 |
+
"epoch": 0.5275484737953542,
|
| 4839 |
+
"grad_norm": 1.3187499046325684,
|
| 4840 |
+
"learning_rate": 9.259621811590578e-05,
|
| 4841 |
+
"loss": 1.1523,
|
| 4842 |
+
"step": 687
|
| 4843 |
+
},
|
| 4844 |
+
{
|
| 4845 |
+
"epoch": 0.5283163755039355,
|
| 4846 |
+
"grad_norm": 1.585579514503479,
|
| 4847 |
+
"learning_rate": 9.235393778375068e-05,
|
| 4848 |
+
"loss": 1.2612,
|
| 4849 |
+
"step": 688
|
| 4850 |
+
},
|
| 4851 |
+
{
|
| 4852 |
+
"epoch": 0.5290842772125168,
|
| 4853 |
+
"grad_norm": 1.7242546081542969,
|
| 4854 |
+
"learning_rate": 9.211170258938747e-05,
|
| 4855 |
+
"loss": 1.374,
|
| 4856 |
+
"step": 689
|
| 4857 |
+
},
|
| 4858 |
+
{
|
| 4859 |
+
"epoch": 0.5298521789210981,
|
| 4860 |
+
"grad_norm": 1.4706342220306396,
|
| 4861 |
+
"learning_rate": 9.18695139628282e-05,
|
| 4862 |
+
"loss": 1.264,
|
| 4863 |
+
"step": 690
|
| 4864 |
+
},
|
| 4865 |
+
{
|
| 4866 |
+
"epoch": 0.5306200806296794,
|
| 4867 |
+
"grad_norm": 1.520382046699524,
|
| 4868 |
+
"learning_rate": 9.162737333381019e-05,
|
| 4869 |
+
"loss": 1.4948,
|
| 4870 |
+
"step": 691
|
| 4871 |
+
},
|
| 4872 |
+
{
|
| 4873 |
+
"epoch": 0.5313879823382607,
|
| 4874 |
+
"grad_norm": 1.598055362701416,
|
| 4875 |
+
"learning_rate": 9.138528213178727e-05,
|
| 4876 |
+
"loss": 1.4628,
|
| 4877 |
+
"step": 692
|
| 4878 |
+
},
|
| 4879 |
+
{
|
| 4880 |
+
"epoch": 0.532155884046842,
|
| 4881 |
+
"grad_norm": 1.2860671281814575,
|
| 4882 |
+
"learning_rate": 9.11432417859216e-05,
|
| 4883 |
+
"loss": 1.0033,
|
| 4884 |
+
"step": 693
|
| 4885 |
+
},
|
| 4886 |
+
{
|
| 4887 |
+
"epoch": 0.5329237857554233,
|
| 4888 |
+
"grad_norm": 1.4247864484786987,
|
| 4889 |
+
"learning_rate": 9.090125372507492e-05,
|
| 4890 |
+
"loss": 1.1321,
|
| 4891 |
+
"step": 694
|
| 4892 |
+
},
|
| 4893 |
+
{
|
| 4894 |
+
"epoch": 0.5336916874640046,
|
| 4895 |
+
"grad_norm": 1.3131632804870605,
|
| 4896 |
+
"learning_rate": 9.065931937780059e-05,
|
| 4897 |
+
"loss": 0.9775,
|
| 4898 |
+
"step": 695
|
| 4899 |
+
},
|
| 4900 |
+
{
|
| 4901 |
+
"epoch": 0.534459589172586,
|
| 4902 |
+
"grad_norm": 1.3943092823028564,
|
| 4903 |
+
"learning_rate": 9.041744017233462e-05,
|
| 4904 |
+
"loss": 1.1079,
|
| 4905 |
+
"step": 696
|
| 4906 |
+
},
|
| 4907 |
+
{
|
| 4908 |
+
"epoch": 0.5352274908811672,
|
| 4909 |
+
"grad_norm": 1.3534505367279053,
|
| 4910 |
+
"learning_rate": 9.017561753658764e-05,
|
| 4911 |
+
"loss": 0.9166,
|
| 4912 |
+
"step": 697
|
| 4913 |
+
},
|
| 4914 |
+
{
|
| 4915 |
+
"epoch": 0.5359953925897485,
|
| 4916 |
+
"grad_norm": 1.6207268238067627,
|
| 4917 |
+
"learning_rate": 8.993385289813627e-05,
|
| 4918 |
+
"loss": 1.1386,
|
| 4919 |
+
"step": 698
|
| 4920 |
+
},
|
| 4921 |
+
{
|
| 4922 |
+
"epoch": 0.5367632942983298,
|
| 4923 |
+
"grad_norm": 1.4252740144729614,
|
| 4924 |
+
"learning_rate": 8.969214768421483e-05,
|
| 4925 |
+
"loss": 1.2191,
|
| 4926 |
+
"step": 699
|
| 4927 |
+
},
|
| 4928 |
+
{
|
| 4929 |
+
"epoch": 0.5375311960069111,
|
| 4930 |
+
"grad_norm": 1.3587523698806763,
|
| 4931 |
+
"learning_rate": 8.945050332170672e-05,
|
| 4932 |
+
"loss": 1.1997,
|
| 4933 |
+
"step": 700
|
| 4934 |
+
},
|
| 4935 |
+
{
|
| 4936 |
+
"epoch": 0.5382990977154924,
|
| 4937 |
+
"grad_norm": 1.596205234527588,
|
| 4938 |
+
"learning_rate": 8.920892123713614e-05,
|
| 4939 |
+
"loss": 1.1552,
|
| 4940 |
+
"step": 701
|
| 4941 |
+
},
|
| 4942 |
+
{
|
| 4943 |
+
"epoch": 0.5390669994240738,
|
| 4944 |
+
"grad_norm": 1.779982089996338,
|
| 4945 |
+
"learning_rate": 8.89674028566597e-05,
|
| 4946 |
+
"loss": 1.46,
|
| 4947 |
+
"step": 702
|
| 4948 |
+
},
|
| 4949 |
+
{
|
| 4950 |
+
"epoch": 0.539834901132655,
|
| 4951 |
+
"grad_norm": 1.493944525718689,
|
| 4952 |
+
"learning_rate": 8.872594960605785e-05,
|
| 4953 |
+
"loss": 1.238,
|
| 4954 |
+
"step": 703
|
| 4955 |
+
},
|
| 4956 |
+
{
|
| 4957 |
+
"epoch": 0.5406028028412363,
|
| 4958 |
+
"grad_norm": 1.5526620149612427,
|
| 4959 |
+
"learning_rate": 8.848456291072666e-05,
|
| 4960 |
+
"loss": 1.3413,
|
| 4961 |
+
"step": 704
|
| 4962 |
+
},
|
| 4963 |
+
{
|
| 4964 |
+
"epoch": 0.5413707045498176,
|
| 4965 |
+
"grad_norm": 1.4706637859344482,
|
| 4966 |
+
"learning_rate": 8.82432441956692e-05,
|
| 4967 |
+
"loss": 1.2897,
|
| 4968 |
+
"step": 705
|
| 4969 |
+
},
|
| 4970 |
+
{
|
| 4971 |
+
"epoch": 0.5421386062583989,
|
| 4972 |
+
"grad_norm": 1.2133674621582031,
|
| 4973 |
+
"learning_rate": 8.80019948854873e-05,
|
| 4974 |
+
"loss": 0.9593,
|
| 4975 |
+
"step": 706
|
| 4976 |
+
},
|
| 4977 |
+
{
|
| 4978 |
+
"epoch": 0.5429065079669803,
|
| 4979 |
+
"grad_norm": 1.2873600721359253,
|
| 4980 |
+
"learning_rate": 8.776081640437294e-05,
|
| 4981 |
+
"loss": 1.118,
|
| 4982 |
+
"step": 707
|
| 4983 |
+
},
|
| 4984 |
+
{
|
| 4985 |
+
"epoch": 0.5436744096755616,
|
| 4986 |
+
"grad_norm": 1.5350018739700317,
|
| 4987 |
+
"learning_rate": 8.751971017610012e-05,
|
| 4988 |
+
"loss": 1.3241,
|
| 4989 |
+
"step": 708
|
| 4990 |
+
},
|
| 4991 |
+
{
|
| 4992 |
+
"epoch": 0.5444423113841428,
|
| 4993 |
+
"grad_norm": 1.4526116847991943,
|
| 4994 |
+
"learning_rate": 8.727867762401623e-05,
|
| 4995 |
+
"loss": 1.2734,
|
| 4996 |
+
"step": 709
|
| 4997 |
+
},
|
| 4998 |
+
{
|
| 4999 |
+
"epoch": 0.5452102130927241,
|
| 5000 |
+
"grad_norm": 1.551858901977539,
|
| 5001 |
+
"learning_rate": 8.703772017103372e-05,
|
| 5002 |
+
"loss": 1.3445,
|
| 5003 |
+
"step": 710
|
| 5004 |
+
},
|
| 5005 |
+
{
|
| 5006 |
+
"epoch": 0.5459781148013054,
|
| 5007 |
+
"grad_norm": 1.0594290494918823,
|
| 5008 |
+
"learning_rate": 8.679683923962174e-05,
|
| 5009 |
+
"loss": 0.891,
|
| 5010 |
+
"step": 711
|
| 5011 |
+
},
|
| 5012 |
+
{
|
| 5013 |
+
"epoch": 0.5467460165098867,
|
| 5014 |
+
"grad_norm": 1.4673049449920654,
|
| 5015 |
+
"learning_rate": 8.655603625179759e-05,
|
| 5016 |
+
"loss": 1.3731,
|
| 5017 |
+
"step": 712
|
| 5018 |
+
},
|
| 5019 |
+
{
|
| 5020 |
+
"epoch": 0.5475139182184681,
|
| 5021 |
+
"grad_norm": 1.2594002485275269,
|
| 5022 |
+
"learning_rate": 8.631531262911857e-05,
|
| 5023 |
+
"loss": 1.0889,
|
| 5024 |
+
"step": 713
|
| 5025 |
+
},
|
| 5026 |
+
{
|
| 5027 |
+
"epoch": 0.5482818199270494,
|
| 5028 |
+
"grad_norm": 1.3199065923690796,
|
| 5029 |
+
"learning_rate": 8.607466979267338e-05,
|
| 5030 |
+
"loss": 1.1011,
|
| 5031 |
+
"step": 714
|
| 5032 |
+
},
|
| 5033 |
+
{
|
| 5034 |
+
"epoch": 0.5490497216356306,
|
| 5035 |
+
"grad_norm": 1.752083659172058,
|
| 5036 |
+
"learning_rate": 8.583410916307386e-05,
|
| 5037 |
+
"loss": 1.4096,
|
| 5038 |
+
"step": 715
|
| 5039 |
+
},
|
| 5040 |
+
{
|
| 5041 |
+
"epoch": 0.5498176233442119,
|
| 5042 |
+
"grad_norm": 1.4408888816833496,
|
| 5043 |
+
"learning_rate": 8.559363216044647e-05,
|
| 5044 |
+
"loss": 1.2785,
|
| 5045 |
+
"step": 716
|
| 5046 |
+
},
|
| 5047 |
+
{
|
| 5048 |
+
"epoch": 0.5505855250527932,
|
| 5049 |
+
"grad_norm": 1.881710410118103,
|
| 5050 |
+
"learning_rate": 8.53532402044241e-05,
|
| 5051 |
+
"loss": 1.854,
|
| 5052 |
+
"step": 717
|
| 5053 |
+
},
|
| 5054 |
+
{
|
| 5055 |
+
"epoch": 0.5513534267613746,
|
| 5056 |
+
"grad_norm": 1.3627890348434448,
|
| 5057 |
+
"learning_rate": 8.51129347141374e-05,
|
| 5058 |
+
"loss": 1.097,
|
| 5059 |
+
"step": 718
|
| 5060 |
+
},
|
| 5061 |
+
{
|
| 5062 |
+
"epoch": 0.5521213284699559,
|
| 5063 |
+
"grad_norm": 1.4654157161712646,
|
| 5064 |
+
"learning_rate": 8.487271710820681e-05,
|
| 5065 |
+
"loss": 1.3027,
|
| 5066 |
+
"step": 719
|
| 5067 |
+
},
|
| 5068 |
+
{
|
| 5069 |
+
"epoch": 0.5528892301785372,
|
| 5070 |
+
"grad_norm": 1.7268285751342773,
|
| 5071 |
+
"learning_rate": 8.463258880473373e-05,
|
| 5072 |
+
"loss": 1.3451,
|
| 5073 |
+
"step": 720
|
| 5074 |
+
},
|
| 5075 |
+
{
|
| 5076 |
+
"epoch": 0.5536571318871184,
|
| 5077 |
+
"grad_norm": 1.112960934638977,
|
| 5078 |
+
"learning_rate": 8.439255122129254e-05,
|
| 5079 |
+
"loss": 0.9304,
|
| 5080 |
+
"step": 721
|
| 5081 |
+
},
|
| 5082 |
+
{
|
| 5083 |
+
"epoch": 0.5544250335956997,
|
| 5084 |
+
"grad_norm": 1.4320957660675049,
|
| 5085 |
+
"learning_rate": 8.415260577492195e-05,
|
| 5086 |
+
"loss": 1.2068,
|
| 5087 |
+
"step": 722
|
| 5088 |
+
},
|
| 5089 |
+
{
|
| 5090 |
+
"epoch": 0.555192935304281,
|
| 5091 |
+
"grad_norm": 1.6056063175201416,
|
| 5092 |
+
"learning_rate": 8.391275388211684e-05,
|
| 5093 |
+
"loss": 1.1656,
|
| 5094 |
+
"step": 723
|
| 5095 |
+
},
|
| 5096 |
+
{
|
| 5097 |
+
"epoch": 0.5559608370128624,
|
| 5098 |
+
"grad_norm": 1.415808081626892,
|
| 5099 |
+
"learning_rate": 8.367299695881973e-05,
|
| 5100 |
+
"loss": 1.3507,
|
| 5101 |
+
"step": 724
|
| 5102 |
+
},
|
| 5103 |
+
{
|
| 5104 |
+
"epoch": 0.5567287387214437,
|
| 5105 |
+
"grad_norm": 1.5225753784179688,
|
| 5106 |
+
"learning_rate": 8.34333364204125e-05,
|
| 5107 |
+
"loss": 1.4994,
|
| 5108 |
+
"step": 725
|
| 5109 |
+
},
|
| 5110 |
+
{
|
| 5111 |
+
"epoch": 0.557496640430025,
|
| 5112 |
+
"grad_norm": 1.4271262884140015,
|
| 5113 |
+
"learning_rate": 8.319377368170808e-05,
|
| 5114 |
+
"loss": 1.2742,
|
| 5115 |
+
"step": 726
|
| 5116 |
+
},
|
| 5117 |
+
{
|
| 5118 |
+
"epoch": 0.5582645421386062,
|
| 5119 |
+
"grad_norm": 1.217391848564148,
|
| 5120 |
+
"learning_rate": 8.295431015694202e-05,
|
| 5121 |
+
"loss": 1.0471,
|
| 5122 |
+
"step": 727
|
| 5123 |
+
},
|
| 5124 |
+
{
|
| 5125 |
+
"epoch": 0.5590324438471875,
|
| 5126 |
+
"grad_norm": 1.5793488025665283,
|
| 5127 |
+
"learning_rate": 8.271494725976418e-05,
|
| 5128 |
+
"loss": 1.3258,
|
| 5129 |
+
"step": 728
|
| 5130 |
+
},
|
| 5131 |
+
{
|
| 5132 |
+
"epoch": 0.5598003455557689,
|
| 5133 |
+
"grad_norm": 1.4085965156555176,
|
| 5134 |
+
"learning_rate": 8.247568640323036e-05,
|
| 5135 |
+
"loss": 1.2015,
|
| 5136 |
+
"step": 729
|
| 5137 |
+
},
|
| 5138 |
+
{
|
| 5139 |
+
"epoch": 0.5605682472643502,
|
| 5140 |
+
"grad_norm": 1.4623831510543823,
|
| 5141 |
+
"learning_rate": 8.223652899979402e-05,
|
| 5142 |
+
"loss": 1.1587,
|
| 5143 |
+
"step": 730
|
| 5144 |
+
},
|
| 5145 |
+
{
|
| 5146 |
+
"epoch": 0.5613361489729315,
|
| 5147 |
+
"grad_norm": 1.187821626663208,
|
| 5148 |
+
"learning_rate": 8.199747646129775e-05,
|
| 5149 |
+
"loss": 1.0068,
|
| 5150 |
+
"step": 731
|
| 5151 |
+
},
|
| 5152 |
+
{
|
| 5153 |
+
"epoch": 0.5621040506815128,
|
| 5154 |
+
"grad_norm": 1.485092282295227,
|
| 5155 |
+
"learning_rate": 8.175853019896534e-05,
|
| 5156 |
+
"loss": 1.3343,
|
| 5157 |
+
"step": 732
|
| 5158 |
+
},
|
| 5159 |
+
{
|
| 5160 |
+
"epoch": 0.562871952390094,
|
| 5161 |
+
"grad_norm": 1.3452789783477783,
|
| 5162 |
+
"learning_rate": 8.15196916233929e-05,
|
| 5163 |
+
"loss": 1.2402,
|
| 5164 |
+
"step": 733
|
| 5165 |
+
},
|
| 5166 |
+
{
|
| 5167 |
+
"epoch": 0.5636398540986753,
|
| 5168 |
+
"grad_norm": 1.7040382623672485,
|
| 5169 |
+
"learning_rate": 8.128096214454105e-05,
|
| 5170 |
+
"loss": 1.1712,
|
| 5171 |
+
"step": 734
|
| 5172 |
+
},
|
| 5173 |
+
{
|
| 5174 |
+
"epoch": 0.5644077558072567,
|
| 5175 |
+
"grad_norm": 1.5438884496688843,
|
| 5176 |
+
"learning_rate": 8.104234317172621e-05,
|
| 5177 |
+
"loss": 1.1095,
|
| 5178 |
+
"step": 735
|
| 5179 |
+
},
|
| 5180 |
+
{
|
| 5181 |
+
"epoch": 0.565175657515838,
|
| 5182 |
+
"grad_norm": 1.173060417175293,
|
| 5183 |
+
"learning_rate": 8.080383611361254e-05,
|
| 5184 |
+
"loss": 0.9693,
|
| 5185 |
+
"step": 736
|
| 5186 |
+
},
|
| 5187 |
+
{
|
| 5188 |
+
"epoch": 0.5659435592244193,
|
| 5189 |
+
"grad_norm": 1.4293849468231201,
|
| 5190 |
+
"learning_rate": 8.056544237820351e-05,
|
| 5191 |
+
"loss": 1.0731,
|
| 5192 |
+
"step": 737
|
| 5193 |
+
},
|
| 5194 |
+
{
|
| 5195 |
+
"epoch": 0.5667114609330006,
|
| 5196 |
+
"grad_norm": 1.4654078483581543,
|
| 5197 |
+
"learning_rate": 8.03271633728335e-05,
|
| 5198 |
+
"loss": 1.4003,
|
| 5199 |
+
"step": 738
|
| 5200 |
+
},
|
| 5201 |
+
{
|
| 5202 |
+
"epoch": 0.5674793626415818,
|
| 5203 |
+
"grad_norm": 1.4996401071548462,
|
| 5204 |
+
"learning_rate": 8.008900050415973e-05,
|
| 5205 |
+
"loss": 1.117,
|
| 5206 |
+
"step": 739
|
| 5207 |
+
},
|
| 5208 |
+
{
|
| 5209 |
+
"epoch": 0.5682472643501632,
|
| 5210 |
+
"grad_norm": 1.5051236152648926,
|
| 5211 |
+
"learning_rate": 7.985095517815371e-05,
|
| 5212 |
+
"loss": 1.5012,
|
| 5213 |
+
"step": 740
|
| 5214 |
+
},
|
| 5215 |
+
{
|
| 5216 |
+
"epoch": 0.5690151660587445,
|
| 5217 |
+
"grad_norm": 1.4953557252883911,
|
| 5218 |
+
"learning_rate": 7.961302880009314e-05,
|
| 5219 |
+
"loss": 1.2279,
|
| 5220 |
+
"step": 741
|
| 5221 |
+
},
|
| 5222 |
+
{
|
| 5223 |
+
"epoch": 0.5697830677673258,
|
| 5224 |
+
"grad_norm": 1.3291648626327515,
|
| 5225 |
+
"learning_rate": 7.937522277455343e-05,
|
| 5226 |
+
"loss": 1.1494,
|
| 5227 |
+
"step": 742
|
| 5228 |
+
},
|
| 5229 |
+
{
|
| 5230 |
+
"epoch": 0.5705509694759071,
|
| 5231 |
+
"grad_norm": 1.496016025543213,
|
| 5232 |
+
"learning_rate": 7.913753850539964e-05,
|
| 5233 |
+
"loss": 1.2484,
|
| 5234 |
+
"step": 743
|
| 5235 |
+
},
|
| 5236 |
+
{
|
| 5237 |
+
"epoch": 0.5713188711844884,
|
| 5238 |
+
"grad_norm": 1.4508578777313232,
|
| 5239 |
+
"learning_rate": 7.889997739577783e-05,
|
| 5240 |
+
"loss": 1.1512,
|
| 5241 |
+
"step": 744
|
| 5242 |
+
},
|
| 5243 |
+
{
|
| 5244 |
+
"epoch": 0.5720867728930696,
|
| 5245 |
+
"grad_norm": 1.3858839273452759,
|
| 5246 |
+
"learning_rate": 7.866254084810724e-05,
|
| 5247 |
+
"loss": 1.3323,
|
| 5248 |
+
"step": 745
|
| 5249 |
+
},
|
| 5250 |
+
{
|
| 5251 |
+
"epoch": 0.572854674601651,
|
| 5252 |
+
"grad_norm": 1.4042750597000122,
|
| 5253 |
+
"learning_rate": 7.842523026407159e-05,
|
| 5254 |
+
"loss": 0.9763,
|
| 5255 |
+
"step": 746
|
| 5256 |
+
},
|
| 5257 |
+
{
|
| 5258 |
+
"epoch": 0.5736225763102323,
|
| 5259 |
+
"grad_norm": 1.4263511896133423,
|
| 5260 |
+
"learning_rate": 7.818804704461108e-05,
|
| 5261 |
+
"loss": 1.1,
|
| 5262 |
+
"step": 747
|
| 5263 |
+
},
|
| 5264 |
+
{
|
| 5265 |
+
"epoch": 0.5743904780188136,
|
| 5266 |
+
"grad_norm": 1.4217872619628906,
|
| 5267 |
+
"learning_rate": 7.795099258991404e-05,
|
| 5268 |
+
"loss": 1.001,
|
| 5269 |
+
"step": 748
|
| 5270 |
+
},
|
| 5271 |
+
{
|
| 5272 |
+
"epoch": 0.5751583797273949,
|
| 5273 |
+
"grad_norm": 1.5163894891738892,
|
| 5274 |
+
"learning_rate": 7.771406829940852e-05,
|
| 5275 |
+
"loss": 1.1633,
|
| 5276 |
+
"step": 749
|
| 5277 |
+
},
|
| 5278 |
+
{
|
| 5279 |
+
"epoch": 0.5759262814359762,
|
| 5280 |
+
"grad_norm": 1.423274278640747,
|
| 5281 |
+
"learning_rate": 7.747727557175434e-05,
|
| 5282 |
+
"loss": 1.4725,
|
| 5283 |
+
"step": 750
|
| 5284 |
+
},
|
| 5285 |
+
{
|
| 5286 |
+
"epoch": 0.5766941831445574,
|
| 5287 |
+
"grad_norm": 1.5022209882736206,
|
| 5288 |
+
"learning_rate": 7.724061580483449e-05,
|
| 5289 |
+
"loss": 1.1757,
|
| 5290 |
+
"step": 751
|
| 5291 |
+
},
|
| 5292 |
+
{
|
| 5293 |
+
"epoch": 0.5774620848531388,
|
| 5294 |
+
"grad_norm": 1.3035588264465332,
|
| 5295 |
+
"learning_rate": 7.700409039574717e-05,
|
| 5296 |
+
"loss": 1.0345,
|
| 5297 |
+
"step": 752
|
| 5298 |
+
},
|
| 5299 |
+
{
|
| 5300 |
+
"epoch": 0.5782299865617201,
|
| 5301 |
+
"grad_norm": 1.6852291822433472,
|
| 5302 |
+
"learning_rate": 7.676770074079732e-05,
|
| 5303 |
+
"loss": 1.1958,
|
| 5304 |
+
"step": 753
|
| 5305 |
+
},
|
| 5306 |
+
{
|
| 5307 |
+
"epoch": 0.5789978882703014,
|
| 5308 |
+
"grad_norm": 1.4031083583831787,
|
| 5309 |
+
"learning_rate": 7.653144823548852e-05,
|
| 5310 |
+
"loss": 1.3684,
|
| 5311 |
+
"step": 754
|
| 5312 |
+
},
|
| 5313 |
+
{
|
| 5314 |
+
"epoch": 0.5797657899788827,
|
| 5315 |
+
"grad_norm": 1.4951286315917969,
|
| 5316 |
+
"learning_rate": 7.62953342745146e-05,
|
| 5317 |
+
"loss": 1.4351,
|
| 5318 |
+
"step": 755
|
| 5319 |
+
},
|
| 5320 |
+
{
|
| 5321 |
+
"epoch": 0.580533691687464,
|
| 5322 |
+
"grad_norm": 1.273272156715393,
|
| 5323 |
+
"learning_rate": 7.605936025175174e-05,
|
| 5324 |
+
"loss": 1.1404,
|
| 5325 |
+
"step": 756
|
| 5326 |
+
},
|
| 5327 |
+
{
|
| 5328 |
+
"epoch": 0.5813015933960453,
|
| 5329 |
+
"grad_norm": 1.4025629758834839,
|
| 5330 |
+
"learning_rate": 7.582352756024971e-05,
|
| 5331 |
+
"loss": 1.4564,
|
| 5332 |
+
"step": 757
|
| 5333 |
+
},
|
| 5334 |
+
{
|
| 5335 |
+
"epoch": 0.5820694951046266,
|
| 5336 |
+
"grad_norm": 1.344117283821106,
|
| 5337 |
+
"learning_rate": 7.558783759222417e-05,
|
| 5338 |
+
"loss": 1.1624,
|
| 5339 |
+
"step": 758
|
| 5340 |
+
},
|
| 5341 |
+
{
|
| 5342 |
+
"epoch": 0.5828373968132079,
|
| 5343 |
+
"grad_norm": 1.3355436325073242,
|
| 5344 |
+
"learning_rate": 7.535229173904811e-05,
|
| 5345 |
+
"loss": 1.043,
|
| 5346 |
+
"step": 759
|
| 5347 |
+
},
|
| 5348 |
+
{
|
| 5349 |
+
"epoch": 0.5836052985217892,
|
| 5350 |
+
"grad_norm": 1.2501863241195679,
|
| 5351 |
+
"learning_rate": 7.511689139124382e-05,
|
| 5352 |
+
"loss": 1.2641,
|
| 5353 |
+
"step": 760
|
| 5354 |
+
},
|
| 5355 |
+
{
|
| 5356 |
+
"epoch": 0.5843732002303705,
|
| 5357 |
+
"grad_norm": 1.3736534118652344,
|
| 5358 |
+
"learning_rate": 7.488163793847458e-05,
|
| 5359 |
+
"loss": 1.1726,
|
| 5360 |
+
"step": 761
|
| 5361 |
+
},
|
| 5362 |
+
{
|
| 5363 |
+
"epoch": 0.5851411019389519,
|
| 5364 |
+
"grad_norm": 1.429854154586792,
|
| 5365 |
+
"learning_rate": 7.464653276953643e-05,
|
| 5366 |
+
"loss": 1.2585,
|
| 5367 |
+
"step": 762
|
| 5368 |
+
},
|
| 5369 |
+
{
|
| 5370 |
+
"epoch": 0.5859090036475331,
|
| 5371 |
+
"grad_norm": 1.7089576721191406,
|
| 5372 |
+
"learning_rate": 7.441157727235015e-05,
|
| 5373 |
+
"loss": 1.5036,
|
| 5374 |
+
"step": 763
|
| 5375 |
+
},
|
| 5376 |
+
{
|
| 5377 |
+
"epoch": 0.5866769053561144,
|
| 5378 |
+
"grad_norm": 1.3912054300308228,
|
| 5379 |
+
"learning_rate": 7.417677283395284e-05,
|
| 5380 |
+
"loss": 1.0993,
|
| 5381 |
+
"step": 764
|
| 5382 |
+
},
|
| 5383 |
+
{
|
| 5384 |
+
"epoch": 0.5874448070646957,
|
| 5385 |
+
"grad_norm": 1.7586991786956787,
|
| 5386 |
+
"learning_rate": 7.394212084048995e-05,
|
| 5387 |
+
"loss": 1.2544,
|
| 5388 |
+
"step": 765
|
| 5389 |
+
},
|
| 5390 |
+
{
|
| 5391 |
+
"epoch": 0.588212708773277,
|
| 5392 |
+
"grad_norm": 1.2912665605545044,
|
| 5393 |
+
"learning_rate": 7.370762267720685e-05,
|
| 5394 |
+
"loss": 1.0896,
|
| 5395 |
+
"step": 766
|
| 5396 |
+
},
|
| 5397 |
+
{
|
| 5398 |
+
"epoch": 0.5889806104818583,
|
| 5399 |
+
"grad_norm": 1.5114494562149048,
|
| 5400 |
+
"learning_rate": 7.347327972844096e-05,
|
| 5401 |
+
"loss": 1.1424,
|
| 5402 |
+
"step": 767
|
| 5403 |
+
},
|
| 5404 |
+
{
|
| 5405 |
+
"epoch": 0.5897485121904397,
|
| 5406 |
+
"grad_norm": 1.5202648639678955,
|
| 5407 |
+
"learning_rate": 7.323909337761317e-05,
|
| 5408 |
+
"loss": 1.2709,
|
| 5409 |
+
"step": 768
|
| 5410 |
+
},
|
| 5411 |
+
{
|
| 5412 |
+
"epoch": 0.5905164138990209,
|
| 5413 |
+
"grad_norm": 1.5682910680770874,
|
| 5414 |
+
"learning_rate": 7.30050650072202e-05,
|
| 5415 |
+
"loss": 1.5064,
|
| 5416 |
+
"step": 769
|
| 5417 |
+
},
|
| 5418 |
+
{
|
| 5419 |
+
"epoch": 0.5912843156076022,
|
| 5420 |
+
"grad_norm": 1.7039157152175903,
|
| 5421 |
+
"learning_rate": 7.277119599882586e-05,
|
| 5422 |
+
"loss": 1.1595,
|
| 5423 |
+
"step": 770
|
| 5424 |
+
},
|
| 5425 |
+
{
|
| 5426 |
+
"epoch": 0.5920522173161835,
|
| 5427 |
+
"grad_norm": 6.636146068572998,
|
| 5428 |
+
"learning_rate": 7.25374877330534e-05,
|
| 5429 |
+
"loss": 0.9431,
|
| 5430 |
+
"step": 771
|
| 5431 |
+
},
|
| 5432 |
+
{
|
| 5433 |
+
"epoch": 0.5928201190247648,
|
| 5434 |
+
"grad_norm": 1.5151164531707764,
|
| 5435 |
+
"learning_rate": 7.230394158957705e-05,
|
| 5436 |
+
"loss": 1.2479,
|
| 5437 |
+
"step": 772
|
| 5438 |
+
},
|
| 5439 |
+
{
|
| 5440 |
+
"epoch": 0.5935880207333462,
|
| 5441 |
+
"grad_norm": 1.6289341449737549,
|
| 5442 |
+
"learning_rate": 7.20705589471139e-05,
|
| 5443 |
+
"loss": 1.2524,
|
| 5444 |
+
"step": 773
|
| 5445 |
+
},
|
| 5446 |
+
{
|
| 5447 |
+
"epoch": 0.5943559224419275,
|
| 5448 |
+
"grad_norm": 1.5332694053649902,
|
| 5449 |
+
"learning_rate": 7.1837341183416e-05,
|
| 5450 |
+
"loss": 1.4389,
|
| 5451 |
+
"step": 774
|
| 5452 |
+
},
|
| 5453 |
+
{
|
| 5454 |
+
"epoch": 0.5951238241505087,
|
| 5455 |
+
"grad_norm": 1.4293265342712402,
|
| 5456 |
+
"learning_rate": 7.160428967526187e-05,
|
| 5457 |
+
"loss": 1.1761,
|
| 5458 |
+
"step": 775
|
| 5459 |
+
},
|
| 5460 |
+
{
|
| 5461 |
+
"epoch": 0.59589172585909,
|
| 5462 |
+
"grad_norm": 1.1657414436340332,
|
| 5463 |
+
"learning_rate": 7.137140579844871e-05,
|
| 5464 |
+
"loss": 0.9015,
|
| 5465 |
+
"step": 776
|
| 5466 |
+
},
|
| 5467 |
+
{
|
| 5468 |
+
"epoch": 0.5966596275676713,
|
| 5469 |
+
"grad_norm": 1.196415662765503,
|
| 5470 |
+
"learning_rate": 7.1138690927784e-05,
|
| 5471 |
+
"loss": 1.0494,
|
| 5472 |
+
"step": 777
|
| 5473 |
+
},
|
| 5474 |
+
{
|
| 5475 |
+
"epoch": 0.5974275292762526,
|
| 5476 |
+
"grad_norm": 1.522498369216919,
|
| 5477 |
+
"learning_rate": 7.090614643707762e-05,
|
| 5478 |
+
"loss": 1.2778,
|
| 5479 |
+
"step": 778
|
| 5480 |
+
},
|
| 5481 |
+
{
|
| 5482 |
+
"epoch": 0.598195430984834,
|
| 5483 |
+
"grad_norm": 1.7265686988830566,
|
| 5484 |
+
"learning_rate": 7.067377369913352e-05,
|
| 5485 |
+
"loss": 1.3857,
|
| 5486 |
+
"step": 779
|
| 5487 |
+
},
|
| 5488 |
+
{
|
| 5489 |
+
"epoch": 0.5989633326934153,
|
| 5490 |
+
"grad_norm": 1.781242847442627,
|
| 5491 |
+
"learning_rate": 7.044157408574185e-05,
|
| 5492 |
+
"loss": 1.8548,
|
| 5493 |
+
"step": 780
|
| 5494 |
+
},
|
| 5495 |
+
{
|
| 5496 |
+
"epoch": 0.5997312344019965,
|
| 5497 |
+
"grad_norm": 1.2553101778030396,
|
| 5498 |
+
"learning_rate": 7.020954896767058e-05,
|
| 5499 |
+
"loss": 1.0755,
|
| 5500 |
+
"step": 781
|
| 5501 |
+
},
|
| 5502 |
+
{
|
| 5503 |
+
"epoch": 0.6004991361105778,
|
| 5504 |
+
"grad_norm": 1.6484979391098022,
|
| 5505 |
+
"learning_rate": 6.997769971465769e-05,
|
| 5506 |
+
"loss": 1.3693,
|
| 5507 |
+
"step": 782
|
| 5508 |
+
},
|
| 5509 |
+
{
|
| 5510 |
+
"epoch": 0.6012670378191591,
|
| 5511 |
+
"grad_norm": 2.0107421875,
|
| 5512 |
+
"learning_rate": 6.974602769540289e-05,
|
| 5513 |
+
"loss": 1.3331,
|
| 5514 |
+
"step": 783
|
| 5515 |
+
},
|
| 5516 |
+
{
|
| 5517 |
+
"epoch": 0.6020349395277405,
|
| 5518 |
+
"grad_norm": 1.945434331893921,
|
| 5519 |
+
"learning_rate": 6.951453427755968e-05,
|
| 5520 |
+
"loss": 1.468,
|
| 5521 |
+
"step": 784
|
| 5522 |
+
},
|
| 5523 |
+
{
|
| 5524 |
+
"epoch": 0.6028028412363218,
|
| 5525 |
+
"grad_norm": 1.637454628944397,
|
| 5526 |
+
"learning_rate": 6.928322082772712e-05,
|
| 5527 |
+
"loss": 1.3608,
|
| 5528 |
+
"step": 785
|
| 5529 |
+
},
|
| 5530 |
+
{
|
| 5531 |
+
"epoch": 0.6035707429449031,
|
| 5532 |
+
"grad_norm": 1.5328805446624756,
|
| 5533 |
+
"learning_rate": 6.905208871144187e-05,
|
| 5534 |
+
"loss": 1.2353,
|
| 5535 |
+
"step": 786
|
| 5536 |
+
},
|
| 5537 |
+
{
|
| 5538 |
+
"epoch": 0.6043386446534843,
|
| 5539 |
+
"grad_norm": 1.452931523323059,
|
| 5540 |
+
"learning_rate": 6.882113929317015e-05,
|
| 5541 |
+
"loss": 1.1362,
|
| 5542 |
+
"step": 787
|
| 5543 |
+
},
|
| 5544 |
+
{
|
| 5545 |
+
"epoch": 0.6051065463620656,
|
| 5546 |
+
"grad_norm": 1.3005062341690063,
|
| 5547 |
+
"learning_rate": 6.859037393629957e-05,
|
| 5548 |
+
"loss": 1.0392,
|
| 5549 |
+
"step": 788
|
| 5550 |
+
},
|
| 5551 |
+
{
|
| 5552 |
+
"epoch": 0.6058744480706469,
|
| 5553 |
+
"grad_norm": 1.3638523817062378,
|
| 5554 |
+
"learning_rate": 6.835979400313122e-05,
|
| 5555 |
+
"loss": 1.2327,
|
| 5556 |
+
"step": 789
|
| 5557 |
+
},
|
| 5558 |
+
{
|
| 5559 |
+
"epoch": 0.6066423497792283,
|
| 5560 |
+
"grad_norm": 1.7218124866485596,
|
| 5561 |
+
"learning_rate": 6.81294008548715e-05,
|
| 5562 |
+
"loss": 1.3321,
|
| 5563 |
+
"step": 790
|
| 5564 |
+
},
|
| 5565 |
+
{
|
| 5566 |
+
"epoch": 0.6074102514878096,
|
| 5567 |
+
"grad_norm": 1.0565749406814575,
|
| 5568 |
+
"learning_rate": 6.789919585162423e-05,
|
| 5569 |
+
"loss": 0.9172,
|
| 5570 |
+
"step": 791
|
| 5571 |
+
},
|
| 5572 |
+
{
|
| 5573 |
+
"epoch": 0.6081781531963909,
|
| 5574 |
+
"grad_norm": 1.1628111600875854,
|
| 5575 |
+
"learning_rate": 6.766918035238237e-05,
|
| 5576 |
+
"loss": 1.0269,
|
| 5577 |
+
"step": 792
|
| 5578 |
+
},
|
| 5579 |
+
{
|
| 5580 |
+
"epoch": 0.6089460549049721,
|
| 5581 |
+
"grad_norm": 1.5882484912872314,
|
| 5582 |
+
"learning_rate": 6.743935571502038e-05,
|
| 5583 |
+
"loss": 1.281,
|
| 5584 |
+
"step": 793
|
| 5585 |
+
},
|
| 5586 |
+
{
|
| 5587 |
+
"epoch": 0.6097139566135534,
|
| 5588 |
+
"grad_norm": 1.0573451519012451,
|
| 5589 |
+
"learning_rate": 6.720972329628577e-05,
|
| 5590 |
+
"loss": 0.681,
|
| 5591 |
+
"step": 794
|
| 5592 |
+
},
|
| 5593 |
+
{
|
| 5594 |
+
"epoch": 0.6104818583221348,
|
| 5595 |
+
"grad_norm": 1.6153432130813599,
|
| 5596 |
+
"learning_rate": 6.698028445179148e-05,
|
| 5597 |
+
"loss": 1.3266,
|
| 5598 |
+
"step": 795
|
| 5599 |
+
},
|
| 5600 |
+
{
|
| 5601 |
+
"epoch": 0.6112497600307161,
|
| 5602 |
+
"grad_norm": 1.3120388984680176,
|
| 5603 |
+
"learning_rate": 6.675104053600763e-05,
|
| 5604 |
+
"loss": 1.101,
|
| 5605 |
+
"step": 796
|
| 5606 |
+
},
|
| 5607 |
+
{
|
| 5608 |
+
"epoch": 0.6120176617392974,
|
| 5609 |
+
"grad_norm": 1.544906735420227,
|
| 5610 |
+
"learning_rate": 6.65219929022535e-05,
|
| 5611 |
+
"loss": 1.2774,
|
| 5612 |
+
"step": 797
|
| 5613 |
+
},
|
| 5614 |
+
{
|
| 5615 |
+
"epoch": 0.6127855634478787,
|
| 5616 |
+
"grad_norm": 1.5639188289642334,
|
| 5617 |
+
"learning_rate": 6.629314290268987e-05,
|
| 5618 |
+
"loss": 1.301,
|
| 5619 |
+
"step": 798
|
| 5620 |
+
},
|
| 5621 |
+
{
|
| 5622 |
+
"epoch": 0.6135534651564599,
|
| 5623 |
+
"grad_norm": 1.116890549659729,
|
| 5624 |
+
"learning_rate": 6.606449188831057e-05,
|
| 5625 |
+
"loss": 1.0978,
|
| 5626 |
+
"step": 799
|
| 5627 |
+
},
|
| 5628 |
+
{
|
| 5629 |
+
"epoch": 0.6143213668650412,
|
| 5630 |
+
"grad_norm": 1.459924578666687,
|
| 5631 |
+
"learning_rate": 6.583604120893488e-05,
|
| 5632 |
+
"loss": 1.1193,
|
| 5633 |
+
"step": 800
|
| 5634 |
+
},
|
| 5635 |
+
{
|
| 5636 |
+
"epoch": 0.6150892685736226,
|
| 5637 |
+
"grad_norm": 1.2557390928268433,
|
| 5638 |
+
"learning_rate": 6.560779221319938e-05,
|
| 5639 |
+
"loss": 1.0979,
|
| 5640 |
+
"step": 801
|
| 5641 |
+
},
|
| 5642 |
+
{
|
| 5643 |
+
"epoch": 0.6158571702822039,
|
| 5644 |
+
"grad_norm": 1.281129240989685,
|
| 5645 |
+
"learning_rate": 6.537974624855003e-05,
|
| 5646 |
+
"loss": 1.3004,
|
| 5647 |
+
"step": 802
|
| 5648 |
+
},
|
| 5649 |
+
{
|
| 5650 |
+
"epoch": 0.6166250719907852,
|
| 5651 |
+
"grad_norm": 1.4236611127853394,
|
| 5652 |
+
"learning_rate": 6.51519046612342e-05,
|
| 5653 |
+
"loss": 1.1254,
|
| 5654 |
+
"step": 803
|
| 5655 |
+
},
|
| 5656 |
+
{
|
| 5657 |
+
"epoch": 0.6173929736993665,
|
| 5658 |
+
"grad_norm": 1.4232025146484375,
|
| 5659 |
+
"learning_rate": 6.492426879629282e-05,
|
| 5660 |
+
"loss": 1.0886,
|
| 5661 |
+
"step": 804
|
| 5662 |
+
},
|
| 5663 |
+
{
|
| 5664 |
+
"epoch": 0.6181608754079477,
|
| 5665 |
+
"grad_norm": 1.569633960723877,
|
| 5666 |
+
"learning_rate": 6.46968399975522e-05,
|
| 5667 |
+
"loss": 1.1388,
|
| 5668 |
+
"step": 805
|
| 5669 |
+
},
|
| 5670 |
+
{
|
| 5671 |
+
"epoch": 0.618928777116529,
|
| 5672 |
+
"grad_norm": 1.7293319702148438,
|
| 5673 |
+
"learning_rate": 6.446961960761643e-05,
|
| 5674 |
+
"loss": 1.4027,
|
| 5675 |
+
"step": 806
|
| 5676 |
+
},
|
| 5677 |
+
{
|
| 5678 |
+
"epoch": 0.6196966788251104,
|
| 5679 |
+
"grad_norm": 1.5421292781829834,
|
| 5680 |
+
"learning_rate": 6.424260896785914e-05,
|
| 5681 |
+
"loss": 1.2518,
|
| 5682 |
+
"step": 807
|
| 5683 |
+
},
|
| 5684 |
+
{
|
| 5685 |
+
"epoch": 0.6204645805336917,
|
| 5686 |
+
"grad_norm": 1.5703529119491577,
|
| 5687 |
+
"learning_rate": 6.40158094184158e-05,
|
| 5688 |
+
"loss": 1.2429,
|
| 5689 |
+
"step": 808
|
| 5690 |
+
},
|
| 5691 |
+
{
|
| 5692 |
+
"epoch": 0.621232482242273,
|
| 5693 |
+
"grad_norm": 1.575032353401184,
|
| 5694 |
+
"learning_rate": 6.378922229817575e-05,
|
| 5695 |
+
"loss": 1.3882,
|
| 5696 |
+
"step": 809
|
| 5697 |
+
},
|
| 5698 |
+
{
|
| 5699 |
+
"epoch": 0.6220003839508543,
|
| 5700 |
+
"grad_norm": 1.1647250652313232,
|
| 5701 |
+
"learning_rate": 6.356284894477412e-05,
|
| 5702 |
+
"loss": 1.031,
|
| 5703 |
+
"step": 810
|
| 5704 |
+
},
|
| 5705 |
+
{
|
| 5706 |
+
"epoch": 0.6227682856594355,
|
| 5707 |
+
"grad_norm": 1.8725273609161377,
|
| 5708 |
+
"learning_rate": 6.333669069458432e-05,
|
| 5709 |
+
"loss": 1.4888,
|
| 5710 |
+
"step": 811
|
| 5711 |
+
},
|
| 5712 |
+
{
|
| 5713 |
+
"epoch": 0.6235361873680169,
|
| 5714 |
+
"grad_norm": 1.4924973249435425,
|
| 5715 |
+
"learning_rate": 6.311074888270971e-05,
|
| 5716 |
+
"loss": 1.3342,
|
| 5717 |
+
"step": 812
|
| 5718 |
+
},
|
| 5719 |
+
{
|
| 5720 |
+
"epoch": 0.6243040890765982,
|
| 5721 |
+
"grad_norm": 1.2791497707366943,
|
| 5722 |
+
"learning_rate": 6.288502484297607e-05,
|
| 5723 |
+
"loss": 1.1041,
|
| 5724 |
+
"step": 813
|
| 5725 |
+
},
|
| 5726 |
+
{
|
| 5727 |
+
"epoch": 0.6250719907851795,
|
| 5728 |
+
"grad_norm": 1.1811076402664185,
|
| 5729 |
+
"learning_rate": 6.265951990792347e-05,
|
| 5730 |
+
"loss": 0.838,
|
| 5731 |
+
"step": 814
|
| 5732 |
+
},
|
| 5733 |
+
{
|
| 5734 |
+
"epoch": 0.6258398924937608,
|
| 5735 |
+
"grad_norm": 1.6005576848983765,
|
| 5736 |
+
"learning_rate": 6.243423540879865e-05,
|
| 5737 |
+
"loss": 1.3403,
|
| 5738 |
+
"step": 815
|
| 5739 |
+
},
|
| 5740 |
+
{
|
| 5741 |
+
"epoch": 0.6266077942023421,
|
| 5742 |
+
"grad_norm": 1.2689980268478394,
|
| 5743 |
+
"learning_rate": 6.220917267554686e-05,
|
| 5744 |
+
"loss": 1.0595,
|
| 5745 |
+
"step": 816
|
| 5746 |
+
},
|
| 5747 |
+
{
|
| 5748 |
+
"epoch": 0.6273756959109233,
|
| 5749 |
+
"grad_norm": 1.4566655158996582,
|
| 5750 |
+
"learning_rate": 6.198433303680439e-05,
|
| 5751 |
+
"loss": 1.2665,
|
| 5752 |
+
"step": 817
|
| 5753 |
+
},
|
| 5754 |
+
{
|
| 5755 |
+
"epoch": 0.6281435976195047,
|
| 5756 |
+
"grad_norm": 1.9303637742996216,
|
| 5757 |
+
"learning_rate": 6.175971781989025e-05,
|
| 5758 |
+
"loss": 0.9127,
|
| 5759 |
+
"step": 818
|
| 5760 |
+
},
|
| 5761 |
+
{
|
| 5762 |
+
"epoch": 0.628911499328086,
|
| 5763 |
+
"grad_norm": 1.7557438611984253,
|
| 5764 |
+
"learning_rate": 6.153532835079886e-05,
|
| 5765 |
+
"loss": 1.4541,
|
| 5766 |
+
"step": 819
|
| 5767 |
+
},
|
| 5768 |
+
{
|
| 5769 |
+
"epoch": 0.6296794010366673,
|
| 5770 |
+
"grad_norm": 1.5808215141296387,
|
| 5771 |
+
"learning_rate": 6.131116595419178e-05,
|
| 5772 |
+
"loss": 1.4305,
|
| 5773 |
+
"step": 820
|
| 5774 |
+
},
|
| 5775 |
+
{
|
| 5776 |
+
"epoch": 0.6304473027452486,
|
| 5777 |
+
"grad_norm": 1.6107852458953857,
|
| 5778 |
+
"learning_rate": 6.108723195339011e-05,
|
| 5779 |
+
"loss": 1.1872,
|
| 5780 |
+
"step": 821
|
| 5781 |
+
},
|
| 5782 |
+
{
|
| 5783 |
+
"epoch": 0.63121520445383,
|
| 5784 |
+
"grad_norm": 1.4639952182769775,
|
| 5785 |
+
"learning_rate": 6.086352767036673e-05,
|
| 5786 |
+
"loss": 0.8872,
|
| 5787 |
+
"step": 822
|
| 5788 |
+
},
|
| 5789 |
+
{
|
| 5790 |
+
"epoch": 0.6319831061624112,
|
| 5791 |
+
"grad_norm": 1.5953764915466309,
|
| 5792 |
+
"learning_rate": 6.064005442573824e-05,
|
| 5793 |
+
"loss": 1.3455,
|
| 5794 |
+
"step": 823
|
| 5795 |
+
},
|
| 5796 |
+
{
|
| 5797 |
+
"epoch": 0.6327510078709925,
|
| 5798 |
+
"grad_norm": 1.7272642850875854,
|
| 5799 |
+
"learning_rate": 6.041681353875746e-05,
|
| 5800 |
+
"loss": 1.273,
|
| 5801 |
+
"step": 824
|
| 5802 |
+
},
|
| 5803 |
+
{
|
| 5804 |
+
"epoch": 0.6335189095795738,
|
| 5805 |
+
"grad_norm": 1.5354080200195312,
|
| 5806 |
+
"learning_rate": 6.019380632730546e-05,
|
| 5807 |
+
"loss": 1.1075,
|
| 5808 |
+
"step": 825
|
| 5809 |
+
},
|
| 5810 |
+
{
|
| 5811 |
+
"epoch": 0.6342868112881551,
|
| 5812 |
+
"grad_norm": 1.4213950634002686,
|
| 5813 |
+
"learning_rate": 5.997103410788385e-05,
|
| 5814 |
+
"loss": 1.2971,
|
| 5815 |
+
"step": 826
|
| 5816 |
+
},
|
| 5817 |
+
{
|
| 5818 |
+
"epoch": 0.6350547129967364,
|
| 5819 |
+
"grad_norm": 1.8925020694732666,
|
| 5820 |
+
"learning_rate": 5.9748498195606925e-05,
|
| 5821 |
+
"loss": 1.3272,
|
| 5822 |
+
"step": 827
|
| 5823 |
+
},
|
| 5824 |
+
{
|
| 5825 |
+
"epoch": 0.6358226147053178,
|
| 5826 |
+
"grad_norm": 1.3079675436019897,
|
| 5827 |
+
"learning_rate": 5.952619990419408e-05,
|
| 5828 |
+
"loss": 1.1373,
|
| 5829 |
+
"step": 828
|
| 5830 |
+
},
|
| 5831 |
+
{
|
| 5832 |
+
"epoch": 0.636590516413899,
|
| 5833 |
+
"grad_norm": 1.5753669738769531,
|
| 5834 |
+
"learning_rate": 5.9304140545961784e-05,
|
| 5835 |
+
"loss": 1.0857,
|
| 5836 |
+
"step": 829
|
| 5837 |
+
},
|
| 5838 |
+
{
|
| 5839 |
+
"epoch": 0.6373584181224803,
|
| 5840 |
+
"grad_norm": 1.3841052055358887,
|
| 5841 |
+
"learning_rate": 5.9082321431816156e-05,
|
| 5842 |
+
"loss": 1.2683,
|
| 5843 |
+
"step": 830
|
| 5844 |
+
},
|
| 5845 |
+
{
|
| 5846 |
+
"epoch": 0.6381263198310616,
|
| 5847 |
+
"grad_norm": 1.3311750888824463,
|
| 5848 |
+
"learning_rate": 5.88607438712449e-05,
|
| 5849 |
+
"loss": 0.9628,
|
| 5850 |
+
"step": 831
|
| 5851 |
+
},
|
| 5852 |
+
{
|
| 5853 |
+
"epoch": 0.6388942215396429,
|
| 5854 |
+
"grad_norm": 1.4135347604751587,
|
| 5855 |
+
"learning_rate": 5.863940917230986e-05,
|
| 5856 |
+
"loss": 1.0973,
|
| 5857 |
+
"step": 832
|
| 5858 |
+
},
|
| 5859 |
+
{
|
| 5860 |
+
"epoch": 0.6396621232482242,
|
| 5861 |
+
"grad_norm": 1.665837049484253,
|
| 5862 |
+
"learning_rate": 5.841831864163909e-05,
|
| 5863 |
+
"loss": 1.5469,
|
| 5864 |
+
"step": 833
|
| 5865 |
+
},
|
| 5866 |
+
{
|
| 5867 |
+
"epoch": 0.6404300249568056,
|
| 5868 |
+
"grad_norm": 1.3354551792144775,
|
| 5869 |
+
"learning_rate": 5.8197473584419184e-05,
|
| 5870 |
+
"loss": 1.1408,
|
| 5871 |
+
"step": 834
|
| 5872 |
+
},
|
| 5873 |
+
{
|
| 5874 |
+
"epoch": 0.6411979266653869,
|
| 5875 |
+
"grad_norm": 1.3838506937026978,
|
| 5876 |
+
"learning_rate": 5.7976875304387756e-05,
|
| 5877 |
+
"loss": 1.0322,
|
| 5878 |
+
"step": 835
|
| 5879 |
+
},
|
| 5880 |
+
{
|
| 5881 |
+
"epoch": 0.6419658283739681,
|
| 5882 |
+
"grad_norm": 1.6376174688339233,
|
| 5883 |
+
"learning_rate": 5.7756525103825474e-05,
|
| 5884 |
+
"loss": 1.4984,
|
| 5885 |
+
"step": 836
|
| 5886 |
+
},
|
| 5887 |
+
{
|
| 5888 |
+
"epoch": 0.6427337300825494,
|
| 5889 |
+
"grad_norm": 1.2751094102859497,
|
| 5890 |
+
"learning_rate": 5.753642428354852e-05,
|
| 5891 |
+
"loss": 1.2019,
|
| 5892 |
+
"step": 837
|
| 5893 |
+
},
|
| 5894 |
+
{
|
| 5895 |
+
"epoch": 0.6435016317911307,
|
| 5896 |
+
"grad_norm": 1.3204684257507324,
|
| 5897 |
+
"learning_rate": 5.731657414290085e-05,
|
| 5898 |
+
"loss": 1.1769,
|
| 5899 |
+
"step": 838
|
| 5900 |
+
},
|
| 5901 |
+
{
|
| 5902 |
+
"epoch": 0.644269533499712,
|
| 5903 |
+
"grad_norm": 1.444831132888794,
|
| 5904 |
+
"learning_rate": 5.7096975979746704e-05,
|
| 5905 |
+
"loss": 1.2601,
|
| 5906 |
+
"step": 839
|
| 5907 |
+
},
|
| 5908 |
+
{
|
| 5909 |
+
"epoch": 0.6450374352082934,
|
| 5910 |
+
"grad_norm": 1.5533994436264038,
|
| 5911 |
+
"learning_rate": 5.687763109046255e-05,
|
| 5912 |
+
"loss": 1.2037,
|
| 5913 |
+
"step": 840
|
| 5914 |
+
},
|
| 5915 |
+
{
|
| 5916 |
+
"epoch": 0.6458053369168747,
|
| 5917 |
+
"grad_norm": 1.3148772716522217,
|
| 5918 |
+
"learning_rate": 5.665854076992991e-05,
|
| 5919 |
+
"loss": 1.0888,
|
| 5920 |
+
"step": 841
|
| 5921 |
+
},
|
| 5922 |
+
{
|
| 5923 |
+
"epoch": 0.6465732386254559,
|
| 5924 |
+
"grad_norm": 1.465082049369812,
|
| 5925 |
+
"learning_rate": 5.643970631152735e-05,
|
| 5926 |
+
"loss": 1.2246,
|
| 5927 |
+
"step": 842
|
| 5928 |
+
},
|
| 5929 |
+
{
|
| 5930 |
+
"epoch": 0.6473411403340372,
|
| 5931 |
+
"grad_norm": 1.5530979633331299,
|
| 5932 |
+
"learning_rate": 5.622112900712304e-05,
|
| 5933 |
+
"loss": 1.4689,
|
| 5934 |
+
"step": 843
|
| 5935 |
+
},
|
| 5936 |
+
{
|
| 5937 |
+
"epoch": 0.6481090420426185,
|
| 5938 |
+
"grad_norm": 1.1694343090057373,
|
| 5939 |
+
"learning_rate": 5.600281014706703e-05,
|
| 5940 |
+
"loss": 0.9216,
|
| 5941 |
+
"step": 844
|
| 5942 |
+
},
|
| 5943 |
+
{
|
| 5944 |
+
"epoch": 0.6488769437511999,
|
| 5945 |
+
"grad_norm": 1.592970371246338,
|
| 5946 |
+
"learning_rate": 5.57847510201837e-05,
|
| 5947 |
+
"loss": 1.1406,
|
| 5948 |
+
"step": 845
|
| 5949 |
+
},
|
| 5950 |
+
{
|
| 5951 |
+
"epoch": 0.6496448454597812,
|
| 5952 |
+
"grad_norm": 1.7581945657730103,
|
| 5953 |
+
"learning_rate": 5.556695291376406e-05,
|
| 5954 |
+
"loss": 1.4523,
|
| 5955 |
+
"step": 846
|
| 5956 |
+
},
|
| 5957 |
+
{
|
| 5958 |
+
"epoch": 0.6504127471683625,
|
| 5959 |
+
"grad_norm": 1.7311755418777466,
|
| 5960 |
+
"learning_rate": 5.5349417113558254e-05,
|
| 5961 |
+
"loss": 1.3606,
|
| 5962 |
+
"step": 847
|
| 5963 |
+
},
|
| 5964 |
+
{
|
| 5965 |
+
"epoch": 0.6511806488769437,
|
| 5966 |
+
"grad_norm": 1.291117548942566,
|
| 5967 |
+
"learning_rate": 5.5132144903768e-05,
|
| 5968 |
+
"loss": 1.0444,
|
| 5969 |
+
"step": 848
|
| 5970 |
+
},
|
| 5971 |
+
{
|
| 5972 |
+
"epoch": 0.651948550585525,
|
| 5973 |
+
"grad_norm": 1.4802870750427246,
|
| 5974 |
+
"learning_rate": 5.491513756703881e-05,
|
| 5975 |
+
"loss": 1.1835,
|
| 5976 |
+
"step": 849
|
| 5977 |
+
},
|
| 5978 |
+
{
|
| 5979 |
+
"epoch": 0.6527164522941064,
|
| 5980 |
+
"grad_norm": 1.0915478467941284,
|
| 5981 |
+
"learning_rate": 5.46983963844526e-05,
|
| 5982 |
+
"loss": 0.9936,
|
| 5983 |
+
"step": 850
|
| 5984 |
+
},
|
| 5985 |
+
{
|
| 5986 |
+
"epoch": 0.6534843540026877,
|
| 5987 |
+
"grad_norm": 1.842965006828308,
|
| 5988 |
+
"learning_rate": 5.448192263552006e-05,
|
| 5989 |
+
"loss": 1.2762,
|
| 5990 |
+
"step": 851
|
| 5991 |
+
},
|
| 5992 |
+
{
|
| 5993 |
+
"epoch": 0.654252255711269,
|
| 5994 |
+
"grad_norm": 1.5491669178009033,
|
| 5995 |
+
"learning_rate": 5.426571759817314e-05,
|
| 5996 |
+
"loss": 1.2144,
|
| 5997 |
+
"step": 852
|
| 5998 |
+
},
|
| 5999 |
+
{
|
| 6000 |
+
"epoch": 0.6550201574198503,
|
| 6001 |
+
"grad_norm": 1.6982448101043701,
|
| 6002 |
+
"learning_rate": 5.4049782548757386e-05,
|
| 6003 |
+
"loss": 1.0843,
|
| 6004 |
+
"step": 853
|
| 6005 |
+
},
|
| 6006 |
+
{
|
| 6007 |
+
"epoch": 0.6557880591284315,
|
| 6008 |
+
"grad_norm": 1.5886989831924438,
|
| 6009 |
+
"learning_rate": 5.383411876202464e-05,
|
| 6010 |
+
"loss": 1.393,
|
| 6011 |
+
"step": 854
|
| 6012 |
+
},
|
| 6013 |
+
{
|
| 6014 |
+
"epoch": 0.6565559608370128,
|
| 6015 |
+
"grad_norm": 1.3143181800842285,
|
| 6016 |
+
"learning_rate": 5.36187275111253e-05,
|
| 6017 |
+
"loss": 1.1188,
|
| 6018 |
+
"step": 855
|
| 6019 |
+
},
|
| 6020 |
+
{
|
| 6021 |
+
"epoch": 0.6573238625455942,
|
| 6022 |
+
"grad_norm": 1.499030590057373,
|
| 6023 |
+
"learning_rate": 5.340361006760082e-05,
|
| 6024 |
+
"loss": 1.1549,
|
| 6025 |
+
"step": 856
|
| 6026 |
+
},
|
| 6027 |
+
{
|
| 6028 |
+
"epoch": 0.6580917642541755,
|
| 6029 |
+
"grad_norm": 1.2520569562911987,
|
| 6030 |
+
"learning_rate": 5.318876770137634e-05,
|
| 6031 |
+
"loss": 1.1459,
|
| 6032 |
+
"step": 857
|
| 6033 |
+
},
|
| 6034 |
+
{
|
| 6035 |
+
"epoch": 0.6588596659627568,
|
| 6036 |
+
"grad_norm": 1.5099189281463623,
|
| 6037 |
+
"learning_rate": 5.297420168075307e-05,
|
| 6038 |
+
"loss": 1.2704,
|
| 6039 |
+
"step": 858
|
| 6040 |
+
},
|
| 6041 |
+
{
|
| 6042 |
+
"epoch": 0.6596275676713381,
|
| 6043 |
+
"grad_norm": 1.6787750720977783,
|
| 6044 |
+
"learning_rate": 5.275991327240082e-05,
|
| 6045 |
+
"loss": 1.2595,
|
| 6046 |
+
"step": 859
|
| 6047 |
+
},
|
| 6048 |
+
{
|
| 6049 |
+
"epoch": 0.6603954693799193,
|
| 6050 |
+
"grad_norm": 1.695313811302185,
|
| 6051 |
+
"learning_rate": 5.254590374135058e-05,
|
| 6052 |
+
"loss": 1.3029,
|
| 6053 |
+
"step": 860
|
| 6054 |
+
},
|
| 6055 |
+
{
|
| 6056 |
+
"epoch": 0.6611633710885007,
|
| 6057 |
+
"grad_norm": 1.2656611204147339,
|
| 6058 |
+
"learning_rate": 5.233217435098707e-05,
|
| 6059 |
+
"loss": 1.2773,
|
| 6060 |
+
"step": 861
|
| 6061 |
+
},
|
| 6062 |
+
{
|
| 6063 |
+
"epoch": 0.661931272797082,
|
| 6064 |
+
"grad_norm": 1.3123670816421509,
|
| 6065 |
+
"learning_rate": 5.2118726363041036e-05,
|
| 6066 |
+
"loss": 1.653,
|
| 6067 |
+
"step": 862
|
| 6068 |
+
},
|
| 6069 |
+
{
|
| 6070 |
+
"epoch": 0.6626991745056633,
|
| 6071 |
+
"grad_norm": 1.6967337131500244,
|
| 6072 |
+
"learning_rate": 5.190556103758223e-05,
|
| 6073 |
+
"loss": 1.356,
|
| 6074 |
+
"step": 863
|
| 6075 |
+
},
|
| 6076 |
+
{
|
| 6077 |
+
"epoch": 0.6634670762142446,
|
| 6078 |
+
"grad_norm": 1.0755648612976074,
|
| 6079 |
+
"learning_rate": 5.1692679633011564e-05,
|
| 6080 |
+
"loss": 0.8091,
|
| 6081 |
+
"step": 864
|
| 6082 |
+
},
|
| 6083 |
+
{
|
| 6084 |
+
"epoch": 0.6642349779228259,
|
| 6085 |
+
"grad_norm": 1.0523223876953125,
|
| 6086 |
+
"learning_rate": 5.148008340605392e-05,
|
| 6087 |
+
"loss": 0.854,
|
| 6088 |
+
"step": 865
|
| 6089 |
+
},
|
| 6090 |
+
{
|
| 6091 |
+
"epoch": 0.6650028796314071,
|
| 6092 |
+
"grad_norm": 1.3313195705413818,
|
| 6093 |
+
"learning_rate": 5.1267773611750624e-05,
|
| 6094 |
+
"loss": 1.0723,
|
| 6095 |
+
"step": 866
|
| 6096 |
+
},
|
| 6097 |
+
{
|
| 6098 |
+
"epoch": 0.6657707813399885,
|
| 6099 |
+
"grad_norm": 1.5703046321868896,
|
| 6100 |
+
"learning_rate": 5.1055751503452144e-05,
|
| 6101 |
+
"loss": 1.3341,
|
| 6102 |
+
"step": 867
|
| 6103 |
+
},
|
| 6104 |
+
{
|
| 6105 |
+
"epoch": 0.6665386830485698,
|
| 6106 |
+
"grad_norm": 1.6033220291137695,
|
| 6107 |
+
"learning_rate": 5.0844018332810594e-05,
|
| 6108 |
+
"loss": 1.4145,
|
| 6109 |
+
"step": 868
|
| 6110 |
+
},
|
| 6111 |
+
{
|
| 6112 |
+
"epoch": 0.6673065847571511,
|
| 6113 |
+
"grad_norm": 1.1940321922302246,
|
| 6114 |
+
"learning_rate": 5.0632575349772225e-05,
|
| 6115 |
+
"loss": 1.0971,
|
| 6116 |
+
"step": 869
|
| 6117 |
+
},
|
| 6118 |
+
{
|
| 6119 |
+
"epoch": 0.6680744864657324,
|
| 6120 |
+
"grad_norm": 1.534477710723877,
|
| 6121 |
+
"learning_rate": 5.0421423802570454e-05,
|
| 6122 |
+
"loss": 1.2956,
|
| 6123 |
+
"step": 870
|
| 6124 |
+
},
|
| 6125 |
+
{
|
| 6126 |
+
"epoch": 0.6688423881743137,
|
| 6127 |
+
"grad_norm": 1.457879900932312,
|
| 6128 |
+
"learning_rate": 5.0210564937718055e-05,
|
| 6129 |
+
"loss": 1.3399,
|
| 6130 |
+
"step": 871
|
| 6131 |
+
},
|
| 6132 |
+
{
|
| 6133 |
+
"epoch": 0.669610289882895,
|
| 6134 |
+
"grad_norm": 1.6395093202590942,
|
| 6135 |
+
"learning_rate": 5.000000000000002e-05,
|
| 6136 |
+
"loss": 1.2267,
|
| 6137 |
+
"step": 872
|
| 6138 |
+
},
|
| 6139 |
+
{
|
| 6140 |
+
"epoch": 0.6703781915914763,
|
| 6141 |
+
"grad_norm": 1.403594970703125,
|
| 6142 |
+
"learning_rate": 4.978973023246616e-05,
|
| 6143 |
+
"loss": 1.2561,
|
| 6144 |
+
"step": 873
|
| 6145 |
+
},
|
| 6146 |
+
{
|
| 6147 |
+
"epoch": 0.6711460933000576,
|
| 6148 |
+
"grad_norm": 1.3514751195907593,
|
| 6149 |
+
"learning_rate": 4.957975687642389e-05,
|
| 6150 |
+
"loss": 0.8501,
|
| 6151 |
+
"step": 874
|
| 6152 |
+
},
|
| 6153 |
+
{
|
| 6154 |
+
"epoch": 0.6719139950086389,
|
| 6155 |
+
"grad_norm": 1.7166774272918701,
|
| 6156 |
+
"learning_rate": 4.937008117143055e-05,
|
| 6157 |
+
"loss": 1.2769,
|
| 6158 |
+
"step": 875
|
| 6159 |
+
},
|
| 6160 |
+
{
|
| 6161 |
+
"epoch": 0.6726818967172202,
|
| 6162 |
+
"grad_norm": 1.637560486793518,
|
| 6163 |
+
"learning_rate": 4.9160704355286577e-05,
|
| 6164 |
+
"loss": 1.2707,
|
| 6165 |
+
"step": 876
|
| 6166 |
+
},
|
| 6167 |
+
{
|
| 6168 |
+
"epoch": 0.6734497984258015,
|
| 6169 |
+
"grad_norm": 1.7437028884887695,
|
| 6170 |
+
"learning_rate": 4.895162766402781e-05,
|
| 6171 |
+
"loss": 1.356,
|
| 6172 |
+
"step": 877
|
| 6173 |
+
},
|
| 6174 |
+
{
|
| 6175 |
+
"epoch": 0.6742177001343828,
|
| 6176 |
+
"grad_norm": 1.4267345666885376,
|
| 6177 |
+
"learning_rate": 4.8742852331918364e-05,
|
| 6178 |
+
"loss": 1.2037,
|
| 6179 |
+
"step": 878
|
| 6180 |
+
},
|
| 6181 |
+
{
|
| 6182 |
+
"epoch": 0.6749856018429641,
|
| 6183 |
+
"grad_norm": 2.0313918590545654,
|
| 6184 |
+
"learning_rate": 4.8534379591443246e-05,
|
| 6185 |
+
"loss": 1.5035,
|
| 6186 |
+
"step": 879
|
| 6187 |
+
},
|
| 6188 |
+
{
|
| 6189 |
+
"epoch": 0.6757535035515454,
|
| 6190 |
+
"grad_norm": 1.4802058935165405,
|
| 6191 |
+
"learning_rate": 4.8326210673301284e-05,
|
| 6192 |
+
"loss": 1.1481,
|
| 6193 |
+
"step": 880
|
| 6194 |
+
},
|
| 6195 |
+
{
|
| 6196 |
+
"epoch": 0.6765214052601267,
|
| 6197 |
+
"grad_norm": 1.6689741611480713,
|
| 6198 |
+
"learning_rate": 4.811834680639765e-05,
|
| 6199 |
+
"loss": 1.2703,
|
| 6200 |
+
"step": 881
|
| 6201 |
+
},
|
| 6202 |
+
{
|
| 6203 |
+
"epoch": 0.677289306968708,
|
| 6204 |
+
"grad_norm": 1.475077748298645,
|
| 6205 |
+
"learning_rate": 4.791078921783653e-05,
|
| 6206 |
+
"loss": 1.0892,
|
| 6207 |
+
"step": 882
|
| 6208 |
+
},
|
| 6209 |
+
{
|
| 6210 |
+
"epoch": 0.6780572086772894,
|
| 6211 |
+
"grad_norm": 1.7580574750900269,
|
| 6212 |
+
"learning_rate": 4.770353913291428e-05,
|
| 6213 |
+
"loss": 1.2977,
|
| 6214 |
+
"step": 883
|
| 6215 |
+
},
|
| 6216 |
+
{
|
| 6217 |
+
"epoch": 0.6788251103858706,
|
| 6218 |
+
"grad_norm": 1.482071876525879,
|
| 6219 |
+
"learning_rate": 4.749659777511177e-05,
|
| 6220 |
+
"loss": 1.2857,
|
| 6221 |
+
"step": 884
|
| 6222 |
+
},
|
| 6223 |
+
{
|
| 6224 |
+
"epoch": 0.6795930120944519,
|
| 6225 |
+
"grad_norm": 1.2742493152618408,
|
| 6226 |
+
"learning_rate": 4.728996636608738e-05,
|
| 6227 |
+
"loss": 1.0448,
|
| 6228 |
+
"step": 885
|
| 6229 |
+
},
|
| 6230 |
+
{
|
| 6231 |
+
"epoch": 0.6803609138030332,
|
| 6232 |
+
"grad_norm": 1.8095487356185913,
|
| 6233 |
+
"learning_rate": 4.708364612566969e-05,
|
| 6234 |
+
"loss": 1.4402,
|
| 6235 |
+
"step": 886
|
| 6236 |
+
},
|
| 6237 |
+
{
|
| 6238 |
+
"epoch": 0.6811288155116145,
|
| 6239 |
+
"grad_norm": 1.4991480112075806,
|
| 6240 |
+
"learning_rate": 4.6877638271850485e-05,
|
| 6241 |
+
"loss": 1.0895,
|
| 6242 |
+
"step": 887
|
| 6243 |
+
},
|
| 6244 |
+
{
|
| 6245 |
+
"epoch": 0.6818967172201958,
|
| 6246 |
+
"grad_norm": 1.588757038116455,
|
| 6247 |
+
"learning_rate": 4.667194402077714e-05,
|
| 6248 |
+
"loss": 1.472,
|
| 6249 |
+
"step": 888
|
| 6250 |
+
},
|
| 6251 |
+
{
|
| 6252 |
+
"epoch": 0.6826646189287772,
|
| 6253 |
+
"grad_norm": 1.5407018661499023,
|
| 6254 |
+
"learning_rate": 4.646656458674595e-05,
|
| 6255 |
+
"loss": 1.0996,
|
| 6256 |
+
"step": 889
|
| 6257 |
+
},
|
| 6258 |
+
{
|
| 6259 |
+
"epoch": 0.6834325206373584,
|
| 6260 |
+
"grad_norm": 1.4947824478149414,
|
| 6261 |
+
"learning_rate": 4.62615011821946e-05,
|
| 6262 |
+
"loss": 1.1829,
|
| 6263 |
+
"step": 890
|
| 6264 |
+
},
|
| 6265 |
+
{
|
| 6266 |
+
"epoch": 0.6842004223459397,
|
| 6267 |
+
"grad_norm": 1.6118358373641968,
|
| 6268 |
+
"learning_rate": 4.6056755017695155e-05,
|
| 6269 |
+
"loss": 1.26,
|
| 6270 |
+
"step": 891
|
| 6271 |
+
},
|
| 6272 |
+
{
|
| 6273 |
+
"epoch": 0.684968324054521,
|
| 6274 |
+
"grad_norm": 1.4456740617752075,
|
| 6275 |
+
"learning_rate": 4.585232730194682e-05,
|
| 6276 |
+
"loss": 1.1356,
|
| 6277 |
+
"step": 892
|
| 6278 |
+
},
|
| 6279 |
+
{
|
| 6280 |
+
"epoch": 0.6857362257631023,
|
| 6281 |
+
"grad_norm": 1.421918272972107,
|
| 6282 |
+
"learning_rate": 4.5648219241769054e-05,
|
| 6283 |
+
"loss": 1.1096,
|
| 6284 |
+
"step": 893
|
| 6285 |
+
},
|
| 6286 |
+
{
|
| 6287 |
+
"epoch": 0.6865041274716837,
|
| 6288 |
+
"grad_norm": 1.182092547416687,
|
| 6289 |
+
"learning_rate": 4.5444432042093996e-05,
|
| 6290 |
+
"loss": 0.9751,
|
| 6291 |
+
"step": 894
|
| 6292 |
+
},
|
| 6293 |
+
{
|
| 6294 |
+
"epoch": 0.687272029180265,
|
| 6295 |
+
"grad_norm": 1.425451636314392,
|
| 6296 |
+
"learning_rate": 4.524096690595978e-05,
|
| 6297 |
+
"loss": 1.3076,
|
| 6298 |
+
"step": 895
|
| 6299 |
+
},
|
| 6300 |
+
{
|
| 6301 |
+
"epoch": 0.6880399308888462,
|
| 6302 |
+
"grad_norm": 1.116645097732544,
|
| 6303 |
+
"learning_rate": 4.5037825034503304e-05,
|
| 6304 |
+
"loss": 0.7194,
|
| 6305 |
+
"step": 896
|
| 6306 |
+
},
|
| 6307 |
+
{
|
| 6308 |
+
"epoch": 0.6888078325974275,
|
| 6309 |
+
"grad_norm": 1.3570024967193604,
|
| 6310 |
+
"learning_rate": 4.4835007626953e-05,
|
| 6311 |
+
"loss": 0.9747,
|
| 6312 |
+
"step": 897
|
| 6313 |
+
},
|
| 6314 |
+
{
|
| 6315 |
+
"epoch": 0.6895757343060088,
|
| 6316 |
+
"grad_norm": 1.3995678424835205,
|
| 6317 |
+
"learning_rate": 4.4632515880621894e-05,
|
| 6318 |
+
"loss": 1.0773,
|
| 6319 |
+
"step": 898
|
| 6320 |
+
},
|
| 6321 |
+
{
|
| 6322 |
+
"epoch": 0.6903436360145901,
|
| 6323 |
+
"grad_norm": 1.5802603960037231,
|
| 6324 |
+
"learning_rate": 4.443035099090048e-05,
|
| 6325 |
+
"loss": 1.3399,
|
| 6326 |
+
"step": 899
|
| 6327 |
+
},
|
| 6328 |
+
{
|
| 6329 |
+
"epoch": 0.6911115377231715,
|
| 6330 |
+
"grad_norm": 1.563367247581482,
|
| 6331 |
+
"learning_rate": 4.4228514151249825e-05,
|
| 6332 |
+
"loss": 1.0937,
|
| 6333 |
+
"step": 900
|
| 6334 |
+
},
|
| 6335 |
+
{
|
| 6336 |
+
"epoch": 0.6918794394317528,
|
| 6337 |
+
"grad_norm": 1.6263221502304077,
|
| 6338 |
+
"learning_rate": 4.4027006553194115e-05,
|
| 6339 |
+
"loss": 1.258,
|
| 6340 |
+
"step": 901
|
| 6341 |
+
},
|
| 6342 |
+
{
|
| 6343 |
+
"epoch": 0.692647341140334,
|
| 6344 |
+
"grad_norm": 1.682242751121521,
|
| 6345 |
+
"learning_rate": 4.3825829386314166e-05,
|
| 6346 |
+
"loss": 1.3231,
|
| 6347 |
+
"step": 902
|
| 6348 |
+
},
|
| 6349 |
+
{
|
| 6350 |
+
"epoch": 0.6934152428489153,
|
| 6351 |
+
"grad_norm": 1.3335479497909546,
|
| 6352 |
+
"learning_rate": 4.362498383823996e-05,
|
| 6353 |
+
"loss": 1.0193,
|
| 6354 |
+
"step": 903
|
| 6355 |
+
},
|
| 6356 |
+
{
|
| 6357 |
+
"epoch": 0.6941831445574966,
|
| 6358 |
+
"grad_norm": 1.619031548500061,
|
| 6359 |
+
"learning_rate": 4.342447109464385e-05,
|
| 6360 |
+
"loss": 1.2696,
|
| 6361 |
+
"step": 904
|
| 6362 |
+
},
|
| 6363 |
+
{
|
| 6364 |
+
"epoch": 0.694951046266078,
|
| 6365 |
+
"grad_norm": 1.4861383438110352,
|
| 6366 |
+
"learning_rate": 4.322429233923351e-05,
|
| 6367 |
+
"loss": 1.3105,
|
| 6368 |
+
"step": 905
|
| 6369 |
+
},
|
| 6370 |
+
{
|
| 6371 |
+
"epoch": 0.6957189479746593,
|
| 6372 |
+
"grad_norm": 1.4006625413894653,
|
| 6373 |
+
"learning_rate": 4.3024448753744925e-05,
|
| 6374 |
+
"loss": 1.3573,
|
| 6375 |
+
"step": 906
|
| 6376 |
+
},
|
| 6377 |
+
{
|
| 6378 |
+
"epoch": 0.6964868496832406,
|
| 6379 |
+
"grad_norm": 1.2410006523132324,
|
| 6380 |
+
"learning_rate": 4.282494151793548e-05,
|
| 6381 |
+
"loss": 0.9566,
|
| 6382 |
+
"step": 907
|
| 6383 |
+
},
|
| 6384 |
+
{
|
| 6385 |
+
"epoch": 0.6972547513918218,
|
| 6386 |
+
"grad_norm": 1.0770704746246338,
|
| 6387 |
+
"learning_rate": 4.2625771809576874e-05,
|
| 6388 |
+
"loss": 0.7789,
|
| 6389 |
+
"step": 908
|
| 6390 |
+
},
|
| 6391 |
+
{
|
| 6392 |
+
"epoch": 0.6980226531004031,
|
| 6393 |
+
"grad_norm": 1.2837942838668823,
|
| 6394 |
+
"learning_rate": 4.242694080444837e-05,
|
| 6395 |
+
"loss": 1.1047,
|
| 6396 |
+
"step": 909
|
| 6397 |
+
},
|
| 6398 |
+
{
|
| 6399 |
+
"epoch": 0.6987905548089844,
|
| 6400 |
+
"grad_norm": 1.1040459871292114,
|
| 6401 |
+
"learning_rate": 4.2228449676329616e-05,
|
| 6402 |
+
"loss": 0.9763,
|
| 6403 |
+
"step": 910
|
| 6404 |
+
},
|
| 6405 |
+
{
|
| 6406 |
+
"epoch": 0.6995584565175658,
|
| 6407 |
+
"grad_norm": 1.6829538345336914,
|
| 6408 |
+
"learning_rate": 4.2030299596993883e-05,
|
| 6409 |
+
"loss": 1.4803,
|
| 6410 |
+
"step": 911
|
| 6411 |
+
},
|
| 6412 |
+
{
|
| 6413 |
+
"epoch": 0.7003263582261471,
|
| 6414 |
+
"grad_norm": 1.4828580617904663,
|
| 6415 |
+
"learning_rate": 4.1832491736201077e-05,
|
| 6416 |
+
"loss": 1.2561,
|
| 6417 |
+
"step": 912
|
| 6418 |
+
},
|
| 6419 |
+
{
|
| 6420 |
+
"epoch": 0.7010942599347284,
|
| 6421 |
+
"grad_norm": 1.568756341934204,
|
| 6422 |
+
"learning_rate": 4.1635027261690827e-05,
|
| 6423 |
+
"loss": 1.4303,
|
| 6424 |
+
"step": 913
|
| 6425 |
+
},
|
| 6426 |
+
{
|
| 6427 |
+
"epoch": 0.7018621616433096,
|
| 6428 |
+
"grad_norm": 1.3178819417953491,
|
| 6429 |
+
"learning_rate": 4.143790733917564e-05,
|
| 6430 |
+
"loss": 0.8952,
|
| 6431 |
+
"step": 914
|
| 6432 |
+
},
|
| 6433 |
+
{
|
| 6434 |
+
"epoch": 0.7026300633518909,
|
| 6435 |
+
"grad_norm": 1.4817698001861572,
|
| 6436 |
+
"learning_rate": 4.124113313233404e-05,
|
| 6437 |
+
"loss": 1.3478,
|
| 6438 |
+
"step": 915
|
| 6439 |
+
},
|
| 6440 |
+
{
|
| 6441 |
+
"epoch": 0.7033979650604723,
|
| 6442 |
+
"grad_norm": 1.4555271863937378,
|
| 6443 |
+
"learning_rate": 4.1044705802803574e-05,
|
| 6444 |
+
"loss": 1.3088,
|
| 6445 |
+
"step": 916
|
| 6446 |
+
},
|
| 6447 |
+
{
|
| 6448 |
+
"epoch": 0.7041658667690536,
|
| 6449 |
+
"grad_norm": 1.4719526767730713,
|
| 6450 |
+
"learning_rate": 4.084862651017406e-05,
|
| 6451 |
+
"loss": 1.2125,
|
| 6452 |
+
"step": 917
|
| 6453 |
+
},
|
| 6454 |
+
{
|
| 6455 |
+
"epoch": 0.7049337684776349,
|
| 6456 |
+
"grad_norm": 1.2816683053970337,
|
| 6457 |
+
"learning_rate": 4.065289641198073e-05,
|
| 6458 |
+
"loss": 1.208,
|
| 6459 |
+
"step": 918
|
| 6460 |
+
},
|
| 6461 |
+
{
|
| 6462 |
+
"epoch": 0.7057016701862162,
|
| 6463 |
+
"grad_norm": 1.3525464534759521,
|
| 6464 |
+
"learning_rate": 4.045751666369736e-05,
|
| 6465 |
+
"loss": 1.1444,
|
| 6466 |
+
"step": 919
|
| 6467 |
+
},
|
| 6468 |
+
{
|
| 6469 |
+
"epoch": 0.7064695718947974,
|
| 6470 |
+
"grad_norm": 1.4733887910842896,
|
| 6471 |
+
"learning_rate": 4.026248841872946e-05,
|
| 6472 |
+
"loss": 1.2619,
|
| 6473 |
+
"step": 920
|
| 6474 |
+
},
|
| 6475 |
+
{
|
| 6476 |
+
"epoch": 0.7072374736033787,
|
| 6477 |
+
"grad_norm": 1.3341400623321533,
|
| 6478 |
+
"learning_rate": 4.006781282840748e-05,
|
| 6479 |
+
"loss": 1.153,
|
| 6480 |
+
"step": 921
|
| 6481 |
+
},
|
| 6482 |
+
{
|
| 6483 |
+
"epoch": 0.7080053753119601,
|
| 6484 |
+
"grad_norm": 1.5136420726776123,
|
| 6485 |
+
"learning_rate": 3.987349104198007e-05,
|
| 6486 |
+
"loss": 1.1891,
|
| 6487 |
+
"step": 922
|
| 6488 |
+
},
|
| 6489 |
+
{
|
| 6490 |
+
"epoch": 0.7087732770205414,
|
| 6491 |
+
"grad_norm": 1.2748438119888306,
|
| 6492 |
+
"learning_rate": 3.9679524206607156e-05,
|
| 6493 |
+
"loss": 1.0631,
|
| 6494 |
+
"step": 923
|
| 6495 |
+
},
|
| 6496 |
+
{
|
| 6497 |
+
"epoch": 0.7095411787291227,
|
| 6498 |
+
"grad_norm": 1.4721794128417969,
|
| 6499 |
+
"learning_rate": 3.948591346735325e-05,
|
| 6500 |
+
"loss": 1.2841,
|
| 6501 |
+
"step": 924
|
| 6502 |
+
},
|
| 6503 |
+
{
|
| 6504 |
+
"epoch": 0.710309080437704,
|
| 6505 |
+
"grad_norm": 1.9709136486053467,
|
| 6506 |
+
"learning_rate": 3.929265996718072e-05,
|
| 6507 |
+
"loss": 1.478,
|
| 6508 |
+
"step": 925
|
| 6509 |
+
},
|
| 6510 |
+
{
|
| 6511 |
+
"epoch": 0.7110769821462852,
|
| 6512 |
+
"grad_norm": 2.1271793842315674,
|
| 6513 |
+
"learning_rate": 3.9099764846943e-05,
|
| 6514 |
+
"loss": 1.1028,
|
| 6515 |
+
"step": 926
|
| 6516 |
+
},
|
| 6517 |
+
{
|
| 6518 |
+
"epoch": 0.7118448838548666,
|
| 6519 |
+
"grad_norm": 1.5257985591888428,
|
| 6520 |
+
"learning_rate": 3.89072292453778e-05,
|
| 6521 |
+
"loss": 0.9435,
|
| 6522 |
+
"step": 927
|
| 6523 |
+
},
|
| 6524 |
+
{
|
| 6525 |
+
"epoch": 0.7126127855634479,
|
| 6526 |
+
"grad_norm": 1.2246668338775635,
|
| 6527 |
+
"learning_rate": 3.871505429910057e-05,
|
| 6528 |
+
"loss": 0.9961,
|
| 6529 |
+
"step": 928
|
| 6530 |
+
},
|
| 6531 |
+
{
|
| 6532 |
+
"epoch": 0.7133806872720292,
|
| 6533 |
+
"grad_norm": 1.3091059923171997,
|
| 6534 |
+
"learning_rate": 3.85232411425976e-05,
|
| 6535 |
+
"loss": 1.1994,
|
| 6536 |
+
"step": 929
|
| 6537 |
+
},
|
| 6538 |
+
{
|
| 6539 |
+
"epoch": 0.7141485889806105,
|
| 6540 |
+
"grad_norm": 1.6306805610656738,
|
| 6541 |
+
"learning_rate": 3.833179090821929e-05,
|
| 6542 |
+
"loss": 1.4055,
|
| 6543 |
+
"step": 930
|
| 6544 |
+
},
|
| 6545 |
+
{
|
| 6546 |
+
"epoch": 0.7149164906891918,
|
| 6547 |
+
"grad_norm": 1.6316722631454468,
|
| 6548 |
+
"learning_rate": 3.814070472617375e-05,
|
| 6549 |
+
"loss": 1.4363,
|
| 6550 |
+
"step": 931
|
| 6551 |
+
},
|
| 6552 |
+
{
|
| 6553 |
+
"epoch": 0.715684392397773,
|
| 6554 |
+
"grad_norm": 1.474591612815857,
|
| 6555 |
+
"learning_rate": 3.794998372451981e-05,
|
| 6556 |
+
"loss": 1.1251,
|
| 6557 |
+
"step": 932
|
| 6558 |
+
},
|
| 6559 |
+
{
|
| 6560 |
+
"epoch": 0.7164522941063544,
|
| 6561 |
+
"grad_norm": 1.4235069751739502,
|
| 6562 |
+
"learning_rate": 3.775962902916056e-05,
|
| 6563 |
+
"loss": 1.0631,
|
| 6564 |
+
"step": 933
|
| 6565 |
+
},
|
| 6566 |
+
{
|
| 6567 |
+
"epoch": 0.7172201958149357,
|
| 6568 |
+
"grad_norm": 1.580439567565918,
|
| 6569 |
+
"learning_rate": 3.756964176383655e-05,
|
| 6570 |
+
"loss": 1.3518,
|
| 6571 |
+
"step": 934
|
| 6572 |
+
},
|
| 6573 |
+
{
|
| 6574 |
+
"epoch": 0.717988097523517,
|
| 6575 |
+
"grad_norm": 1.3070342540740967,
|
| 6576 |
+
"learning_rate": 3.7380023050119415e-05,
|
| 6577 |
+
"loss": 1.2074,
|
| 6578 |
+
"step": 935
|
| 6579 |
+
},
|
| 6580 |
+
{
|
| 6581 |
+
"epoch": 0.7187559992320983,
|
| 6582 |
+
"grad_norm": 1.4719895124435425,
|
| 6583 |
+
"learning_rate": 3.7190774007404835e-05,
|
| 6584 |
+
"loss": 1.0298,
|
| 6585 |
+
"step": 936
|
| 6586 |
+
},
|
| 6587 |
+
{
|
| 6588 |
+
"epoch": 0.7195239009406796,
|
| 6589 |
+
"grad_norm": 1.440446376800537,
|
| 6590 |
+
"learning_rate": 3.700189575290641e-05,
|
| 6591 |
+
"loss": 1.1669,
|
| 6592 |
+
"step": 937
|
| 6593 |
+
},
|
| 6594 |
+
{
|
| 6595 |
+
"epoch": 0.7202918026492608,
|
| 6596 |
+
"grad_norm": 1.3623183965682983,
|
| 6597 |
+
"learning_rate": 3.681338940164868e-05,
|
| 6598 |
+
"loss": 1.1463,
|
| 6599 |
+
"step": 938
|
| 6600 |
+
},
|
| 6601 |
+
{
|
| 6602 |
+
"epoch": 0.7210597043578422,
|
| 6603 |
+
"grad_norm": 1.303728461265564,
|
| 6604 |
+
"learning_rate": 3.6625256066460735e-05,
|
| 6605 |
+
"loss": 1.0379,
|
| 6606 |
+
"step": 939
|
| 6607 |
+
},
|
| 6608 |
+
{
|
| 6609 |
+
"epoch": 0.7218276060664235,
|
| 6610 |
+
"grad_norm": 1.3376387357711792,
|
| 6611 |
+
"learning_rate": 3.6437496857969566e-05,
|
| 6612 |
+
"loss": 1.2042,
|
| 6613 |
+
"step": 940
|
| 6614 |
+
},
|
| 6615 |
+
{
|
| 6616 |
+
"epoch": 0.7225955077750048,
|
| 6617 |
+
"grad_norm": 1.6355212926864624,
|
| 6618 |
+
"learning_rate": 3.625011288459365e-05,
|
| 6619 |
+
"loss": 1.3353,
|
| 6620 |
+
"step": 941
|
| 6621 |
+
},
|
| 6622 |
+
{
|
| 6623 |
+
"epoch": 0.7233634094835861,
|
| 6624 |
+
"grad_norm": 1.3309329748153687,
|
| 6625 |
+
"learning_rate": 3.606310525253621e-05,
|
| 6626 |
+
"loss": 0.765,
|
| 6627 |
+
"step": 942
|
| 6628 |
+
},
|
| 6629 |
+
{
|
| 6630 |
+
"epoch": 0.7241313111921674,
|
| 6631 |
+
"grad_norm": 1.7754614353179932,
|
| 6632 |
+
"learning_rate": 3.5876475065778715e-05,
|
| 6633 |
+
"loss": 1.1589,
|
| 6634 |
+
"step": 943
|
| 6635 |
+
},
|
| 6636 |
+
{
|
| 6637 |
+
"epoch": 0.7248992129007487,
|
| 6638 |
+
"grad_norm": 1.2315130233764648,
|
| 6639 |
+
"learning_rate": 3.5690223426074576e-05,
|
| 6640 |
+
"loss": 0.9012,
|
| 6641 |
+
"step": 944
|
| 6642 |
+
},
|
| 6643 |
+
{
|
| 6644 |
+
"epoch": 0.72566711460933,
|
| 6645 |
+
"grad_norm": 1.3348158597946167,
|
| 6646 |
+
"learning_rate": 3.550435143294238e-05,
|
| 6647 |
+
"loss": 1.0467,
|
| 6648 |
+
"step": 945
|
| 6649 |
+
},
|
| 6650 |
+
{
|
| 6651 |
+
"epoch": 0.7264350163179113,
|
| 6652 |
+
"grad_norm": 1.4577149152755737,
|
| 6653 |
+
"learning_rate": 3.531886018365954e-05,
|
| 6654 |
+
"loss": 1.0583,
|
| 6655 |
+
"step": 946
|
| 6656 |
+
},
|
| 6657 |
+
{
|
| 6658 |
+
"epoch": 0.7272029180264926,
|
| 6659 |
+
"grad_norm": 1.927155613899231,
|
| 6660 |
+
"learning_rate": 3.513375077325575e-05,
|
| 6661 |
+
"loss": 1.2384,
|
| 6662 |
+
"step": 947
|
| 6663 |
+
},
|
| 6664 |
+
{
|
| 6665 |
+
"epoch": 0.7279708197350739,
|
| 6666 |
+
"grad_norm": 2.0631394386291504,
|
| 6667 |
+
"learning_rate": 3.4949024294506674e-05,
|
| 6668 |
+
"loss": 1.3989,
|
| 6669 |
+
"step": 948
|
| 6670 |
+
},
|
| 6671 |
+
{
|
| 6672 |
+
"epoch": 0.7287387214436553,
|
| 6673 |
+
"grad_norm": 1.7409682273864746,
|
| 6674 |
+
"learning_rate": 3.476468183792716e-05,
|
| 6675 |
+
"loss": 1.3022,
|
| 6676 |
+
"step": 949
|
| 6677 |
+
},
|
| 6678 |
+
{
|
| 6679 |
+
"epoch": 0.7295066231522365,
|
| 6680 |
+
"grad_norm": 1.0105476379394531,
|
| 6681 |
+
"learning_rate": 3.458072449176525e-05,
|
| 6682 |
+
"loss": 0.902,
|
| 6683 |
+
"step": 950
|
| 6684 |
+
},
|
| 6685 |
+
{
|
| 6686 |
+
"epoch": 0.7302745248608178,
|
| 6687 |
+
"grad_norm": 1.6116336584091187,
|
| 6688 |
+
"learning_rate": 3.439715334199538e-05,
|
| 6689 |
+
"loss": 1.2317,
|
| 6690 |
+
"step": 951
|
| 6691 |
+
},
|
| 6692 |
+
{
|
| 6693 |
+
"epoch": 0.7310424265693991,
|
| 6694 |
+
"grad_norm": 1.6682941913604736,
|
| 6695 |
+
"learning_rate": 3.4213969472312154e-05,
|
| 6696 |
+
"loss": 1.4236,
|
| 6697 |
+
"step": 952
|
| 6698 |
+
},
|
| 6699 |
+
{
|
| 6700 |
+
"epoch": 0.7318103282779804,
|
| 6701 |
+
"grad_norm": 1.2750146389007568,
|
| 6702 |
+
"learning_rate": 3.403117396412391e-05,
|
| 6703 |
+
"loss": 0.8671,
|
| 6704 |
+
"step": 953
|
| 6705 |
+
},
|
| 6706 |
+
{
|
| 6707 |
+
"epoch": 0.7325782299865617,
|
| 6708 |
+
"grad_norm": 1.2994352579116821,
|
| 6709 |
+
"learning_rate": 3.384876789654631e-05,
|
| 6710 |
+
"loss": 1.1212,
|
| 6711 |
+
"step": 954
|
| 6712 |
+
},
|
| 6713 |
+
{
|
| 6714 |
+
"epoch": 0.7333461316951431,
|
| 6715 |
+
"grad_norm": 2.186424732208252,
|
| 6716 |
+
"learning_rate": 3.366675234639601e-05,
|
| 6717 |
+
"loss": 1.4304,
|
| 6718 |
+
"step": 955
|
| 6719 |
+
},
|
| 6720 |
+
{
|
| 6721 |
+
"epoch": 0.7341140334037243,
|
| 6722 |
+
"grad_norm": 1.4062567949295044,
|
| 6723 |
+
"learning_rate": 3.348512838818425e-05,
|
| 6724 |
+
"loss": 1.1381,
|
| 6725 |
+
"step": 956
|
| 6726 |
+
},
|
| 6727 |
+
{
|
| 6728 |
+
"epoch": 0.7348819351123056,
|
| 6729 |
+
"grad_norm": 1.4242571592330933,
|
| 6730 |
+
"learning_rate": 3.3303897094110636e-05,
|
| 6731 |
+
"loss": 1.0323,
|
| 6732 |
+
"step": 957
|
| 6733 |
+
},
|
| 6734 |
+
{
|
| 6735 |
+
"epoch": 0.7356498368208869,
|
| 6736 |
+
"grad_norm": 1.2858667373657227,
|
| 6737 |
+
"learning_rate": 3.3123059534056634e-05,
|
| 6738 |
+
"loss": 1.0643,
|
| 6739 |
+
"step": 958
|
| 6740 |
+
},
|
| 6741 |
+
{
|
| 6742 |
+
"epoch": 0.7364177385294682,
|
| 6743 |
+
"grad_norm": 1.6064081192016602,
|
| 6744 |
+
"learning_rate": 3.294261677557935e-05,
|
| 6745 |
+
"loss": 1.2303,
|
| 6746 |
+
"step": 959
|
| 6747 |
+
},
|
| 6748 |
+
{
|
| 6749 |
+
"epoch": 0.7371856402380496,
|
| 6750 |
+
"grad_norm": 1.5934802293777466,
|
| 6751 |
+
"learning_rate": 3.2762569883905205e-05,
|
| 6752 |
+
"loss": 1.2093,
|
| 6753 |
+
"step": 960
|
| 6754 |
+
},
|
| 6755 |
+
{
|
| 6756 |
+
"epoch": 0.7379535419466309,
|
| 6757 |
+
"grad_norm": 1.4172946214675903,
|
| 6758 |
+
"learning_rate": 3.258291992192377e-05,
|
| 6759 |
+
"loss": 1.2843,
|
| 6760 |
+
"step": 961
|
| 6761 |
+
},
|
| 6762 |
+
{
|
| 6763 |
+
"epoch": 0.7387214436552121,
|
| 6764 |
+
"grad_norm": 1.6268061399459839,
|
| 6765 |
+
"learning_rate": 3.240366795018117e-05,
|
| 6766 |
+
"loss": 1.4062,
|
| 6767 |
+
"step": 962
|
| 6768 |
+
},
|
| 6769 |
+
{
|
| 6770 |
+
"epoch": 0.7394893453637934,
|
| 6771 |
+
"grad_norm": 1.4560757875442505,
|
| 6772 |
+
"learning_rate": 3.222481502687425e-05,
|
| 6773 |
+
"loss": 1.3195,
|
| 6774 |
+
"step": 963
|
| 6775 |
+
},
|
| 6776 |
+
{
|
| 6777 |
+
"epoch": 0.7402572470723747,
|
| 6778 |
+
"grad_norm": 1.716341257095337,
|
| 6779 |
+
"learning_rate": 3.2046362207844e-05,
|
| 6780 |
+
"loss": 1.3937,
|
| 6781 |
+
"step": 964
|
| 6782 |
+
},
|
| 6783 |
+
{
|
| 6784 |
+
"epoch": 0.741025148780956,
|
| 6785 |
+
"grad_norm": 1.4211591482162476,
|
| 6786 |
+
"learning_rate": 3.1868310546569424e-05,
|
| 6787 |
+
"loss": 1.3429,
|
| 6788 |
+
"step": 965
|
| 6789 |
+
},
|
| 6790 |
+
{
|
| 6791 |
+
"epoch": 0.7417930504895374,
|
| 6792 |
+
"grad_norm": 1.4726101160049438,
|
| 6793 |
+
"learning_rate": 3.1690661094161364e-05,
|
| 6794 |
+
"loss": 1.2199,
|
| 6795 |
+
"step": 966
|
| 6796 |
+
},
|
| 6797 |
+
{
|
| 6798 |
+
"epoch": 0.7425609521981187,
|
| 6799 |
+
"grad_norm": 1.6528246402740479,
|
| 6800 |
+
"learning_rate": 3.151341489935627e-05,
|
| 6801 |
+
"loss": 1.4508,
|
| 6802 |
+
"step": 967
|
| 6803 |
+
},
|
| 6804 |
+
{
|
| 6805 |
+
"epoch": 0.7433288539066999,
|
| 6806 |
+
"grad_norm": 1.4311929941177368,
|
| 6807 |
+
"learning_rate": 3.133657300850995e-05,
|
| 6808 |
+
"loss": 1.3215,
|
| 6809 |
+
"step": 968
|
| 6810 |
+
},
|
| 6811 |
+
{
|
| 6812 |
+
"epoch": 0.7440967556152812,
|
| 6813 |
+
"grad_norm": 1.6301379203796387,
|
| 6814 |
+
"learning_rate": 3.116013646559146e-05,
|
| 6815 |
+
"loss": 1.1288,
|
| 6816 |
+
"step": 969
|
| 6817 |
+
},
|
| 6818 |
+
{
|
| 6819 |
+
"epoch": 0.7448646573238625,
|
| 6820 |
+
"grad_norm": 1.50910222530365,
|
| 6821 |
+
"learning_rate": 3.0984106312177e-05,
|
| 6822 |
+
"loss": 1.139,
|
| 6823 |
+
"step": 970
|
| 6824 |
+
},
|
| 6825 |
+
{
|
| 6826 |
+
"epoch": 0.7456325590324439,
|
| 6827 |
+
"grad_norm": 1.4490171670913696,
|
| 6828 |
+
"learning_rate": 3.0808483587443595e-05,
|
| 6829 |
+
"loss": 1.1863,
|
| 6830 |
+
"step": 971
|
| 6831 |
+
},
|
| 6832 |
+
{
|
| 6833 |
+
"epoch": 0.7464004607410252,
|
| 6834 |
+
"grad_norm": 1.835550308227539,
|
| 6835 |
+
"learning_rate": 3.063326932816307e-05,
|
| 6836 |
+
"loss": 1.2192,
|
| 6837 |
+
"step": 972
|
| 6838 |
+
},
|
| 6839 |
+
{
|
| 6840 |
+
"epoch": 0.7471683624496065,
|
| 6841 |
+
"grad_norm": 1.5930982828140259,
|
| 6842 |
+
"learning_rate": 3.045846456869592e-05,
|
| 6843 |
+
"loss": 1.148,
|
| 6844 |
+
"step": 973
|
| 6845 |
+
},
|
| 6846 |
+
{
|
| 6847 |
+
"epoch": 0.7479362641581877,
|
| 6848 |
+
"grad_norm": 1.7530690431594849,
|
| 6849 |
+
"learning_rate": 3.0284070340985295e-05,
|
| 6850 |
+
"loss": 1.2803,
|
| 6851 |
+
"step": 974
|
| 6852 |
+
},
|
| 6853 |
+
{
|
| 6854 |
+
"epoch": 0.748704165866769,
|
| 6855 |
+
"grad_norm": 1.397155523300171,
|
| 6856 |
+
"learning_rate": 3.011008767455059e-05,
|
| 6857 |
+
"loss": 1.2601,
|
| 6858 |
+
"step": 975
|
| 6859 |
+
},
|
| 6860 |
+
{
|
| 6861 |
+
"epoch": 0.7494720675753503,
|
| 6862 |
+
"grad_norm": 1.3860681056976318,
|
| 6863 |
+
"learning_rate": 2.9936517596481818e-05,
|
| 6864 |
+
"loss": 1.314,
|
| 6865 |
+
"step": 976
|
| 6866 |
+
},
|
| 6867 |
+
{
|
| 6868 |
+
"epoch": 0.7502399692839317,
|
| 6869 |
+
"grad_norm": 1.445178747177124,
|
| 6870 |
+
"learning_rate": 2.9763361131433208e-05,
|
| 6871 |
+
"loss": 0.9815,
|
| 6872 |
+
"step": 977
|
| 6873 |
+
},
|
| 6874 |
+
{
|
| 6875 |
+
"epoch": 0.751007870992513,
|
| 6876 |
+
"grad_norm": 1.291642427444458,
|
| 6877 |
+
"learning_rate": 2.9590619301617183e-05,
|
| 6878 |
+
"loss": 1.0057,
|
| 6879 |
+
"step": 978
|
| 6880 |
+
},
|
| 6881 |
+
{
|
| 6882 |
+
"epoch": 0.751007870992513,
|
| 6883 |
+
"eval_loss": 1.251602292060852,
|
| 6884 |
+
"eval_runtime": 19.752,
|
| 6885 |
+
"eval_samples_per_second": 27.795,
|
| 6886 |
+
"eval_steps_per_second": 13.923,
|
| 6887 |
+
"step": 978
|
| 6888 |
}
|
| 6889 |
],
|
| 6890 |
"logging_steps": 1,
|
|
|
|
| 6904 |
"attributes": {}
|
| 6905 |
}
|
| 6906 |
},
|
| 6907 |
+
"total_flos": 1.9488354717401088e+17,
|
| 6908 |
"train_batch_size": 2,
|
| 6909 |
"trial_name": null,
|
| 6910 |
"trial_params": null
|