Training in progress, step 4800, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 2231685024
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c1bc3b3e143f9c5cf9f286dd374831316522298c162cc7d9aee812c4494e2502
|
| 3 |
size 2231685024
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4383116999
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b3f31c32a5cc6868da9e0787a7682ce0c0935d63cee861a343159b380d70512b
|
| 3 |
size 4383116999
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4413b3c1c89ddb3d57ddabb40868d44f33618613ca6b14588ab3262fd959fe1a
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1256
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:536269e746606d89d7092eefc1d242400387136c0a490c65f4003318a00e54ce
|
| 3 |
size 1256
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.7991831302642822,
|
| 3 |
"best_model_checkpoint": "./output/checkpoint-4650",
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 150,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -3510,6 +3510,119 @@
|
|
| 3510 |
"eval_samples_per_second": 6.721,
|
| 3511 |
"eval_steps_per_second": 6.721,
|
| 3512 |
"step": 4650
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3513 |
}
|
| 3514 |
],
|
| 3515 |
"logging_steps": 10,
|
|
@@ -3529,7 +3642,7 @@
|
|
| 3529 |
"attributes": {}
|
| 3530 |
}
|
| 3531 |
},
|
| 3532 |
-
"total_flos":
|
| 3533 |
"train_batch_size": 8,
|
| 3534 |
"trial_name": null,
|
| 3535 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.7991831302642822,
|
| 3 |
"best_model_checkpoint": "./output/checkpoint-4650",
|
| 4 |
+
"epoch": 0.34500107812836917,
|
| 5 |
"eval_steps": 150,
|
| 6 |
+
"global_step": 4800,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 3510 |
"eval_samples_per_second": 6.721,
|
| 3511 |
"eval_steps_per_second": 6.721,
|
| 3512 |
"step": 4650
|
| 3513 |
+
},
|
| 3514 |
+
{
|
| 3515 |
+
"epoch": 0.33493854668295836,
|
| 3516 |
+
"grad_norm": 4.487733364105225,
|
| 3517 |
+
"learning_rate": 9.466184635518361e-07,
|
| 3518 |
+
"loss": 0.6828,
|
| 3519 |
+
"step": 4660
|
| 3520 |
+
},
|
| 3521 |
+
{
|
| 3522 |
+
"epoch": 0.33565729892905916,
|
| 3523 |
+
"grad_norm": 3.950061082839966,
|
| 3524 |
+
"learning_rate": 8.919587459299461e-07,
|
| 3525 |
+
"loss": 0.6465,
|
| 3526 |
+
"step": 4670
|
| 3527 |
+
},
|
| 3528 |
+
{
|
| 3529 |
+
"epoch": 0.3363760511751599,
|
| 3530 |
+
"grad_norm": 4.053449630737305,
|
| 3531 |
+
"learning_rate": 8.389066067082852e-07,
|
| 3532 |
+
"loss": 0.6912,
|
| 3533 |
+
"step": 4680
|
| 3534 |
+
},
|
| 3535 |
+
{
|
| 3536 |
+
"epoch": 0.3370948034212607,
|
| 3537 |
+
"grad_norm": 3.2221829891204834,
|
| 3538 |
+
"learning_rate": 7.874642266525102e-07,
|
| 3539 |
+
"loss": 0.7184,
|
| 3540 |
+
"step": 4690
|
| 3541 |
+
},
|
| 3542 |
+
{
|
| 3543 |
+
"epoch": 0.3378135556673615,
|
| 3544 |
+
"grad_norm": 3.5493359565734863,
|
| 3545 |
+
"learning_rate": 7.376337203573824e-07,
|
| 3546 |
+
"loss": 0.6761,
|
| 3547 |
+
"step": 4700
|
| 3548 |
+
},
|
| 3549 |
+
{
|
| 3550 |
+
"epoch": 0.3385323079134622,
|
| 3551 |
+
"grad_norm": 3.3036763668060303,
|
| 3552 |
+
"learning_rate": 6.894171361598467e-07,
|
| 3553 |
+
"loss": 0.6892,
|
| 3554 |
+
"step": 4710
|
| 3555 |
+
},
|
| 3556 |
+
{
|
| 3557 |
+
"epoch": 0.339251060159563,
|
| 3558 |
+
"grad_norm": 4.632784366607666,
|
| 3559 |
+
"learning_rate": 6.428164560548134e-07,
|
| 3560 |
+
"loss": 0.7149,
|
| 3561 |
+
"step": 4720
|
| 3562 |
+
},
|
| 3563 |
+
{
|
| 3564 |
+
"epoch": 0.33996981240566376,
|
| 3565 |
+
"grad_norm": 4.091222763061523,
|
| 3566 |
+
"learning_rate": 5.978335956137304e-07,
|
| 3567 |
+
"loss": 0.7652,
|
| 3568 |
+
"step": 4730
|
| 3569 |
+
},
|
| 3570 |
+
{
|
| 3571 |
+
"epoch": 0.34068856465176456,
|
| 3572 |
+
"grad_norm": 4.298058032989502,
|
| 3573 |
+
"learning_rate": 5.544704039058025e-07,
|
| 3574 |
+
"loss": 0.6956,
|
| 3575 |
+
"step": 4740
|
| 3576 |
+
},
|
| 3577 |
+
{
|
| 3578 |
+
"epoch": 0.3414073168978653,
|
| 3579 |
+
"grad_norm": 3.7245140075683594,
|
| 3580 |
+
"learning_rate": 5.12728663421993e-07,
|
| 3581 |
+
"loss": 0.6434,
|
| 3582 |
+
"step": 4750
|
| 3583 |
+
},
|
| 3584 |
+
{
|
| 3585 |
+
"epoch": 0.3421260691439661,
|
| 3586 |
+
"grad_norm": 3.273090124130249,
|
| 3587 |
+
"learning_rate": 4.7261009000177274e-07,
|
| 3588 |
+
"loss": 0.72,
|
| 3589 |
+
"step": 4760
|
| 3590 |
+
},
|
| 3591 |
+
{
|
| 3592 |
+
"epoch": 0.34284482139006683,
|
| 3593 |
+
"grad_norm": 4.249372959136963,
|
| 3594 |
+
"learning_rate": 4.3411633276255307e-07,
|
| 3595 |
+
"loss": 0.7028,
|
| 3596 |
+
"step": 4770
|
| 3597 |
+
},
|
| 3598 |
+
{
|
| 3599 |
+
"epoch": 0.34356357363616763,
|
| 3600 |
+
"grad_norm": 3.6272311210632324,
|
| 3601 |
+
"learning_rate": 3.972489740319274e-07,
|
| 3602 |
+
"loss": 0.731,
|
| 3603 |
+
"step": 4780
|
| 3604 |
+
},
|
| 3605 |
+
{
|
| 3606 |
+
"epoch": 0.34428232588226837,
|
| 3607 |
+
"grad_norm": 4.036685943603516,
|
| 3608 |
+
"learning_rate": 3.62009529282612e-07,
|
| 3609 |
+
"loss": 0.7919,
|
| 3610 |
+
"step": 4790
|
| 3611 |
+
},
|
| 3612 |
+
{
|
| 3613 |
+
"epoch": 0.34500107812836917,
|
| 3614 |
+
"grad_norm": 4.364447116851807,
|
| 3615 |
+
"learning_rate": 3.283994470701579e-07,
|
| 3616 |
+
"loss": 0.7161,
|
| 3617 |
+
"step": 4800
|
| 3618 |
+
},
|
| 3619 |
+
{
|
| 3620 |
+
"epoch": 0.34500107812836917,
|
| 3621 |
+
"eval_loss": 0.7992691397666931,
|
| 3622 |
+
"eval_runtime": 77.2387,
|
| 3623 |
+
"eval_samples_per_second": 6.473,
|
| 3624 |
+
"eval_steps_per_second": 6.473,
|
| 3625 |
+
"step": 4800
|
| 3626 |
}
|
| 3627 |
],
|
| 3628 |
"logging_steps": 10,
|
|
|
|
| 3642 |
"attributes": {}
|
| 3643 |
}
|
| 3644 |
},
|
| 3645 |
+
"total_flos": 4.088392743032586e+17,
|
| 3646 |
"train_batch_size": 8,
|
| 3647 |
"trial_name": null,
|
| 3648 |
"trial_params": null
|