Training in progress, step 17800, checkpoint
Browse files
last-checkpoint/README.md
CHANGED
|
@@ -1379,6 +1379,10 @@ You can finetune this model on your own dataset.
|
|
| 1379 |
| 0.3092 | 17500 | 0.3045 |
|
| 1380 |
| 0.3101 | 17550 | 0.3271 |
|
| 1381 |
| 0.3110 | 17600 | 0.519 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1382 |
|
| 1383 |
</details>
|
| 1384 |
|
|
|
|
| 1379 |
| 0.3092 | 17500 | 0.3045 |
|
| 1380 |
| 0.3101 | 17550 | 0.3271 |
|
| 1381 |
| 0.3110 | 17600 | 0.519 |
|
| 1382 |
+
| 0.3119 | 17650 | 0.341 |
|
| 1383 |
+
| 0.3128 | 17700 | 0.3352 |
|
| 1384 |
+
| 0.3137 | 17750 | 0.4406 |
|
| 1385 |
+
| 0.3145 | 17800 | 0.3822 |
|
| 1386 |
|
| 1387 |
</details>
|
| 1388 |
|
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 90864192
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d23dcacb737e2af42314c6ab6136446e4982fee8745c6c03e350fd15717250aa
|
| 3 |
size 90864192
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 180609210
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4eb8f9bc1642eae53fb943b388f6c8e06377f86d8a4ed25c9948af7605742194
|
| 3 |
size 180609210
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:116d3f323503837c4f5074c61dce27aa590a7a4580eb34472179dff367b43737
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54d3b315ce89145171b35cb9993d6540d84cc0fc327f65af910fcf8f238d2633
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3bb6755a2a28890a2c297eec26254b464de2f60f49387af25b290236695ca25d
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 500,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -2472,6 +2472,34 @@
|
|
| 2472 |
"learning_rate": 3.828316742259135e-05,
|
| 2473 |
"loss": 0.519,
|
| 2474 |
"step": 17600
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2475 |
}
|
| 2476 |
],
|
| 2477 |
"logging_steps": 50,
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 0.31453764732908057,
|
| 6 |
"eval_steps": 500,
|
| 7 |
+
"global_step": 17800,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 2472 |
"learning_rate": 3.828316742259135e-05,
|
| 2473 |
"loss": 0.519,
|
| 2474 |
"step": 17600
|
| 2475 |
+
},
|
| 2476 |
+
{
|
| 2477 |
+
"epoch": 0.31188704917743104,
|
| 2478 |
+
"grad_norm": 4.507582187652588,
|
| 2479 |
+
"learning_rate": 3.823408140425281e-05,
|
| 2480 |
+
"loss": 0.341,
|
| 2481 |
+
"step": 17650
|
| 2482 |
+
},
|
| 2483 |
+
{
|
| 2484 |
+
"epoch": 0.31277058189464757,
|
| 2485 |
+
"grad_norm": 1.3272327184677124,
|
| 2486 |
+
"learning_rate": 3.8184995385914276e-05,
|
| 2487 |
+
"loss": 0.3352,
|
| 2488 |
+
"step": 17700
|
| 2489 |
+
},
|
| 2490 |
+
{
|
| 2491 |
+
"epoch": 0.3136541146118641,
|
| 2492 |
+
"grad_norm": 2.516676664352417,
|
| 2493 |
+
"learning_rate": 3.813590936757574e-05,
|
| 2494 |
+
"loss": 0.4406,
|
| 2495 |
+
"step": 17750
|
| 2496 |
+
},
|
| 2497 |
+
{
|
| 2498 |
+
"epoch": 0.31453764732908057,
|
| 2499 |
+
"grad_norm": 1.8230887651443481,
|
| 2500 |
+
"learning_rate": 3.8086823349237206e-05,
|
| 2501 |
+
"loss": 0.3822,
|
| 2502 |
+
"step": 17800
|
| 2503 |
}
|
| 2504 |
],
|
| 2505 |
"logging_steps": 50,
|