Training in progress, step 1820, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 100697728
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28ddd758223901c924b4a39d2b9adc57ddd56c3afa8fb973d5ec1abe128fca3b
|
| 3 |
size 100697728
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 201541754
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7e299b48ed8c7654ff0cc635ef02bd506683a7ab5a1e444aa5178257f66c3942
|
| 3 |
size 201541754
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8cda418843404e2782fe6f5f002b4cdccdcb0443cc8062fcf67c27f24ef06b82
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:314de47bf6b1f60450b370f7a9cc3a90b8e2ec4f652380a82046763165a3b095
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1447,14 +1447,30 @@
|
|
| 1447 |
"loss": 0.3885,
|
| 1448 |
"num_input_tokens_seen": 1217841,
|
| 1449 |
"step": 1800
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1450 |
}
|
| 1451 |
],
|
| 1452 |
"logging_steps": 10,
|
| 1453 |
"max_steps": 2795,
|
| 1454 |
-
"num_input_tokens_seen":
|
| 1455 |
"num_train_epochs": 1,
|
| 1456 |
"save_steps": 20,
|
| 1457 |
-
"total_flos": 2.
|
| 1458 |
"train_batch_size": 1,
|
| 1459 |
"trial_name": null,
|
| 1460 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": null,
|
| 3 |
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.6511627906976745,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 1820,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1447 |
"loss": 0.3885,
|
| 1448 |
"num_input_tokens_seen": 1217841,
|
| 1449 |
"step": 1800
|
| 1450 |
+
},
|
| 1451 |
+
{
|
| 1452 |
+
"epoch": 0.6475849731663685,
|
| 1453 |
+
"grad_norm": 0.43780773878097534,
|
| 1454 |
+
"learning_rate": 7.048300536672629e-05,
|
| 1455 |
+
"loss": 0.3893,
|
| 1456 |
+
"num_input_tokens_seen": 1225095,
|
| 1457 |
+
"step": 1810
|
| 1458 |
+
},
|
| 1459 |
+
{
|
| 1460 |
+
"epoch": 0.6511627906976745,
|
| 1461 |
+
"grad_norm": 0.381865918636322,
|
| 1462 |
+
"learning_rate": 6.976744186046513e-05,
|
| 1463 |
+
"loss": 0.3884,
|
| 1464 |
+
"num_input_tokens_seen": 1234033,
|
| 1465 |
+
"step": 1820
|
| 1466 |
}
|
| 1467 |
],
|
| 1468 |
"logging_steps": 10,
|
| 1469 |
"max_steps": 2795,
|
| 1470 |
+
"num_input_tokens_seen": 1234033,
|
| 1471 |
"num_train_epochs": 1,
|
| 1472 |
"save_steps": 20,
|
| 1473 |
+
"total_flos": 2.7749044315736064e+16,
|
| 1474 |
"train_batch_size": 1,
|
| 1475 |
"trial_name": null,
|
| 1476 |
"trial_params": null
|