Training in progress, step 4800, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 21919400
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9b8e7ec63f00c08a0d78de870428c4cdb987deac18b80648eecdc9bcf7f2db6b
|
| 3 |
size 21919400
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4954713
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8a86be7b6fbcc19bcc5ffa9bb3027746f006a1b13ceed203486c580889aee16f
|
| 3 |
size 4954713
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14180
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:267a692e774fa423b3e8019601554a0456febdd9da57d9f82e4289f21b314860
|
| 3 |
size 14180
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1256
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54129884a43fac233d16f8f024313b97a23fce2899d2b6122913d1d94ddf2e29
|
| 3 |
size 1256
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 11.487832069396973,
|
| 3 |
"best_model_checkpoint": "./output/checkpoint-3900",
|
| 4 |
-
"epoch": 15.
|
| 5 |
"eval_steps": 150,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -3510,6 +3510,119 @@
|
|
| 3510 |
"eval_samples_per_second": 123.62,
|
| 3511 |
"eval_steps_per_second": 123.62,
|
| 3512 |
"step": 4650
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3513 |
}
|
| 3514 |
],
|
| 3515 |
"logging_steps": 10,
|
|
@@ -3529,7 +3642,7 @@
|
|
| 3529 |
"attributes": {}
|
| 3530 |
}
|
| 3531 |
},
|
| 3532 |
-
"total_flos":
|
| 3533 |
"train_batch_size": 8,
|
| 3534 |
"trial_name": null,
|
| 3535 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 11.487832069396973,
|
| 3 |
"best_model_checkpoint": "./output/checkpoint-3900",
|
| 4 |
+
"epoch": 15.894039735099337,
|
| 5 |
"eval_steps": 150,
|
| 6 |
+
"global_step": 4800,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 3510 |
"eval_samples_per_second": 123.62,
|
| 3511 |
"eval_steps_per_second": 123.62,
|
| 3512 |
"step": 4650
|
| 3513 |
+
},
|
| 3514 |
+
{
|
| 3515 |
+
"epoch": 15.43046357615894,
|
| 3516 |
+
"grad_norm": 0.5005173087120056,
|
| 3517 |
+
"learning_rate": 1.1832730794397948e-06,
|
| 3518 |
+
"loss": 11.4594,
|
| 3519 |
+
"step": 4660
|
| 3520 |
+
},
|
| 3521 |
+
{
|
| 3522 |
+
"epoch": 15.463576158940397,
|
| 3523 |
+
"grad_norm": 0.3516858220100403,
|
| 3524 |
+
"learning_rate": 1.1149484324124322e-06,
|
| 3525 |
+
"loss": 11.4586,
|
| 3526 |
+
"step": 4670
|
| 3527 |
+
},
|
| 3528 |
+
{
|
| 3529 |
+
"epoch": 15.496688741721854,
|
| 3530 |
+
"grad_norm": 0.43980923295021057,
|
| 3531 |
+
"learning_rate": 1.048633258385356e-06,
|
| 3532 |
+
"loss": 11.4724,
|
| 3533 |
+
"step": 4680
|
| 3534 |
+
},
|
| 3535 |
+
{
|
| 3536 |
+
"epoch": 15.52980132450331,
|
| 3537 |
+
"grad_norm": 0.391050785779953,
|
| 3538 |
+
"learning_rate": 9.843302833156374e-07,
|
| 3539 |
+
"loss": 11.47,
|
| 3540 |
+
"step": 4690
|
| 3541 |
+
},
|
| 3542 |
+
{
|
| 3543 |
+
"epoch": 15.562913907284768,
|
| 3544 |
+
"grad_norm": 0.4212386906147003,
|
| 3545 |
+
"learning_rate": 9.220421504467278e-07,
|
| 3546 |
+
"loss": 11.4647,
|
| 3547 |
+
"step": 4700
|
| 3548 |
+
},
|
| 3549 |
+
{
|
| 3550 |
+
"epoch": 15.596026490066226,
|
| 3551 |
+
"grad_norm": 0.4143437445163727,
|
| 3552 |
+
"learning_rate": 8.617714201998081e-07,
|
| 3553 |
+
"loss": 11.4658,
|
| 3554 |
+
"step": 4710
|
| 3555 |
+
},
|
| 3556 |
+
{
|
| 3557 |
+
"epoch": 15.629139072847682,
|
| 3558 |
+
"grad_norm": 0.46835222840309143,
|
| 3559 |
+
"learning_rate": 8.035205700685165e-07,
|
| 3560 |
+
"loss": 11.4813,
|
| 3561 |
+
"step": 4720
|
| 3562 |
+
},
|
| 3563 |
+
{
|
| 3564 |
+
"epoch": 15.66225165562914,
|
| 3565 |
+
"grad_norm": 0.38814619183540344,
|
| 3566 |
+
"learning_rate": 7.472919945171628e-07,
|
| 3567 |
+
"loss": 11.4636,
|
| 3568 |
+
"step": 4730
|
| 3569 |
+
},
|
| 3570 |
+
{
|
| 3571 |
+
"epoch": 15.695364238410596,
|
| 3572 |
+
"grad_norm": 0.3894265592098236,
|
| 3573 |
+
"learning_rate": 6.930880048822528e-07,
|
| 3574 |
+
"loss": 11.4568,
|
| 3575 |
+
"step": 4740
|
| 3576 |
+
},
|
| 3577 |
+
{
|
| 3578 |
+
"epoch": 15.728476821192054,
|
| 3579 |
+
"grad_norm": 0.3959190249443054,
|
| 3580 |
+
"learning_rate": 6.40910829277491e-07,
|
| 3581 |
+
"loss": 11.4678,
|
| 3582 |
+
"step": 4750
|
| 3583 |
+
},
|
| 3584 |
+
{
|
| 3585 |
+
"epoch": 15.76158940397351,
|
| 3586 |
+
"grad_norm": 0.42889001965522766,
|
| 3587 |
+
"learning_rate": 5.907626125022157e-07,
|
| 3588 |
+
"loss": 11.4679,
|
| 3589 |
+
"step": 4760
|
| 3590 |
+
},
|
| 3591 |
+
{
|
| 3592 |
+
"epoch": 15.794701986754967,
|
| 3593 |
+
"grad_norm": 0.34252163767814636,
|
| 3594 |
+
"learning_rate": 5.426454159531912e-07,
|
| 3595 |
+
"loss": 11.4672,
|
| 3596 |
+
"step": 4770
|
| 3597 |
+
},
|
| 3598 |
+
{
|
| 3599 |
+
"epoch": 15.827814569536423,
|
| 3600 |
+
"grad_norm": 0.5112998485565186,
|
| 3601 |
+
"learning_rate": 4.965612175399091e-07,
|
| 3602 |
+
"loss": 11.4672,
|
| 3603 |
+
"step": 4780
|
| 3604 |
+
},
|
| 3605 |
+
{
|
| 3606 |
+
"epoch": 15.860927152317881,
|
| 3607 |
+
"grad_norm": 0.38379812240600586,
|
| 3608 |
+
"learning_rate": 4.525119116032649e-07,
|
| 3609 |
+
"loss": 11.4695,
|
| 3610 |
+
"step": 4790
|
| 3611 |
+
},
|
| 3612 |
+
{
|
| 3613 |
+
"epoch": 15.894039735099337,
|
| 3614 |
+
"grad_norm": 0.40494051575660706,
|
| 3615 |
+
"learning_rate": 4.104993088376973e-07,
|
| 3616 |
+
"loss": 11.4611,
|
| 3617 |
+
"step": 4800
|
| 3618 |
+
},
|
| 3619 |
+
{
|
| 3620 |
+
"epoch": 15.894039735099337,
|
| 3621 |
+
"eval_loss": 11.487883567810059,
|
| 3622 |
+
"eval_runtime": 2.1723,
|
| 3623 |
+
"eval_samples_per_second": 123.373,
|
| 3624 |
+
"eval_steps_per_second": 123.373,
|
| 3625 |
+
"step": 4800
|
| 3626 |
}
|
| 3627 |
],
|
| 3628 |
"logging_steps": 10,
|
|
|
|
| 3642 |
"attributes": {}
|
| 3643 |
}
|
| 3644 |
},
|
| 3645 |
+
"total_flos": 86880464660304.0,
|
| 3646 |
"train_batch_size": 8,
|
| 3647 |
"trial_name": null,
|
| 3648 |
"trial_params": null
|