Training in progress, step 17200, checkpoint
Browse files
last-checkpoint/README.md
CHANGED
|
@@ -1367,6 +1367,10 @@ You can finetune this model on your own dataset.
|
|
| 1367 |
| 0.2986 | 16900 | 0.3352 |
|
| 1368 |
| 0.2995 | 16950 | 0.3812 |
|
| 1369 |
| 0.3004 | 17000 | 0.3332 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1370 |
|
| 1371 |
</details>
|
| 1372 |
|
|
|
|
| 1367 |
| 0.2986 | 16900 | 0.3352 |
|
| 1368 |
| 0.2995 | 16950 | 0.3812 |
|
| 1369 |
| 0.3004 | 17000 | 0.3332 |
|
| 1370 |
+
| 0.3013 | 17050 | 0.3472 |
|
| 1371 |
+
| 0.3022 | 17100 | 0.4207 |
|
| 1372 |
+
| 0.3031 | 17150 | 0.3849 |
|
| 1373 |
+
| 0.3039 | 17200 | 0.3272 |
|
| 1374 |
|
| 1375 |
</details>
|
| 1376 |
|
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 90864192
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c8aa8eabf0bd95daa104946b416f3292c9a92c6053a47e54e8098083f61f3874
|
| 3 |
size 90864192
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 180609210
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:550870691770e56b635695a47badd3d85105a2abddd01d2f68ac682eb4779cfb
|
| 3 |
size 180609210
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9642b48564b74a9ef3023b3677305007ed89dfb95240067ad9d1be72f7d2692a
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4c157b3717ebad1d1c4deed2f4a1e630570eeaf89e281bb553ac41881d02a28e
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fcdcffdcd8983b3d6481c27ecbd4b8783b8846e8b39268fee55aae8ba8b88e41
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -2,9 +2,9 @@
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 500,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -2388,6 +2388,34 @@
|
|
| 2388 |
"learning_rate": 3.887219964265379e-05,
|
| 2389 |
"loss": 0.3332,
|
| 2390 |
"step": 17000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2391 |
}
|
| 2392 |
],
|
| 2393 |
"logging_steps": 50,
|
|
|
|
| 2 |
"best_global_step": null,
|
| 3 |
"best_metric": null,
|
| 4 |
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 0.3039352547224824,
|
| 6 |
"eval_steps": 500,
|
| 7 |
+
"global_step": 17200,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 2388 |
"learning_rate": 3.887219964265379e-05,
|
| 2389 |
"loss": 0.3332,
|
| 2390 |
"step": 17000
|
| 2391 |
+
},
|
| 2392 |
+
{
|
| 2393 |
+
"epoch": 0.3012846565708328,
|
| 2394 |
+
"grad_norm": 1.9923433065414429,
|
| 2395 |
+
"learning_rate": 3.882311362431525e-05,
|
| 2396 |
+
"loss": 0.3472,
|
| 2397 |
+
"step": 17050
|
| 2398 |
+
},
|
| 2399 |
+
{
|
| 2400 |
+
"epoch": 0.30216818928804934,
|
| 2401 |
+
"grad_norm": 1.5736125707626343,
|
| 2402 |
+
"learning_rate": 3.8774027605976714e-05,
|
| 2403 |
+
"loss": 0.4207,
|
| 2404 |
+
"step": 17100
|
| 2405 |
+
},
|
| 2406 |
+
{
|
| 2407 |
+
"epoch": 0.30305172200526587,
|
| 2408 |
+
"grad_norm": 2.2181496620178223,
|
| 2409 |
+
"learning_rate": 3.872494158763818e-05,
|
| 2410 |
+
"loss": 0.3849,
|
| 2411 |
+
"step": 17150
|
| 2412 |
+
},
|
| 2413 |
+
{
|
| 2414 |
+
"epoch": 0.3039352547224824,
|
| 2415 |
+
"grad_norm": 1.5112169981002808,
|
| 2416 |
+
"learning_rate": 3.867585556929964e-05,
|
| 2417 |
+
"loss": 0.3272,
|
| 2418 |
+
"step": 17200
|
| 2419 |
}
|
| 2420 |
],
|
| 2421 |
"logging_steps": 50,
|