Training in progress, step 76, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 335604696
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:17a688d72965a31498835e19a3d50bc535a4301ec1529cae220b288f76255751
|
| 3 |
size 335604696
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 671466706
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:14a37368ea40e4c82c5f268ea72a4c729bee978393ed7f62b728dd5129a46d12
|
| 3 |
size 671466706
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47acafa232bcc28ab5833d79ff3104d3e9a72de50268a1ee1ac71395fc84423e
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8ddabf60630e29c65bf15d4d38a70b53e1096fcbf23378832f8b10226895763
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 1.9277753829956055,
|
| 3 |
"best_model_checkpoint": "miner_id_24/checkpoint-50",
|
| 4 |
-
"epoch": 0.
|
| 5 |
"eval_steps": 25,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -381,6 +381,196 @@
|
|
| 381 |
"eval_samples_per_second": 3.509,
|
| 382 |
"eval_steps_per_second": 3.509,
|
| 383 |
"step": 50
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 384 |
}
|
| 385 |
],
|
| 386 |
"logging_steps": 1,
|
|
@@ -404,12 +594,12 @@
|
|
| 404 |
"should_evaluate": false,
|
| 405 |
"should_log": false,
|
| 406 |
"should_save": true,
|
| 407 |
-
"should_training_stop":
|
| 408 |
},
|
| 409 |
"attributes": {}
|
| 410 |
}
|
| 411 |
},
|
| 412 |
-
"total_flos":
|
| 413 |
"train_batch_size": 1,
|
| 414 |
"trial_name": null,
|
| 415 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 1.9277753829956055,
|
| 3 |
"best_model_checkpoint": "miner_id_24/checkpoint-50",
|
| 4 |
+
"epoch": 0.0324825366296697,
|
| 5 |
"eval_steps": 25,
|
| 6 |
+
"global_step": 76,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 381 |
"eval_samples_per_second": 3.509,
|
| 382 |
"eval_steps_per_second": 3.509,
|
| 383 |
"step": 50
|
| 384 |
+
},
|
| 385 |
+
{
|
| 386 |
+
"epoch": 0.021797491685699402,
|
| 387 |
+
"grad_norm": 73.92220306396484,
|
| 388 |
+
"learning_rate": 9.916118258515936e-05,
|
| 389 |
+
"loss": 60.9493,
|
| 390 |
+
"step": 51
|
| 391 |
+
},
|
| 392 |
+
{
|
| 393 |
+
"epoch": 0.022224893483458213,
|
| 394 |
+
"grad_norm": 65.9156265258789,
|
| 395 |
+
"learning_rate": 9.421851671977372e-05,
|
| 396 |
+
"loss": 60.0054,
|
| 397 |
+
"step": 52
|
| 398 |
+
},
|
| 399 |
+
{
|
| 400 |
+
"epoch": 0.022652295281217028,
|
| 401 |
+
"grad_norm": 49.42222595214844,
|
| 402 |
+
"learning_rate": 8.940340387142462e-05,
|
| 403 |
+
"loss": 57.0988,
|
| 404 |
+
"step": 53
|
| 405 |
+
},
|
| 406 |
+
{
|
| 407 |
+
"epoch": 0.02307969707897584,
|
| 408 |
+
"grad_norm": 31.272438049316406,
|
| 409 |
+
"learning_rate": 8.47245211989213e-05,
|
| 410 |
+
"loss": 59.2207,
|
| 411 |
+
"step": 54
|
| 412 |
+
},
|
| 413 |
+
{
|
| 414 |
+
"epoch": 0.02350709887673465,
|
| 415 |
+
"grad_norm": 19.92361831665039,
|
| 416 |
+
"learning_rate": 8.019030036508827e-05,
|
| 417 |
+
"loss": 58.0069,
|
| 418 |
+
"step": 55
|
| 419 |
+
},
|
| 420 |
+
{
|
| 421 |
+
"epoch": 0.023934500674493462,
|
| 422 |
+
"grad_norm": 33.7132453918457,
|
| 423 |
+
"learning_rate": 7.580891234233904e-05,
|
| 424 |
+
"loss": 56.2188,
|
| 425 |
+
"step": 56
|
| 426 |
+
},
|
| 427 |
+
{
|
| 428 |
+
"epoch": 0.024361902472252273,
|
| 429 |
+
"grad_norm": 23.189586639404297,
|
| 430 |
+
"learning_rate": 7.158825268803127e-05,
|
| 431 |
+
"loss": 56.9692,
|
| 432 |
+
"step": 57
|
| 433 |
+
},
|
| 434 |
+
{
|
| 435 |
+
"epoch": 0.024789304270011085,
|
| 436 |
+
"grad_norm": 21.800243377685547,
|
| 437 |
+
"learning_rate": 6.75359273161392e-05,
|
| 438 |
+
"loss": 61.5117,
|
| 439 |
+
"step": 58
|
| 440 |
+
},
|
| 441 |
+
{
|
| 442 |
+
"epoch": 0.0252167060677699,
|
| 443 |
+
"grad_norm": 20.54871940612793,
|
| 444 |
+
"learning_rate": 6.365923879088219e-05,
|
| 445 |
+
"loss": 57.6141,
|
| 446 |
+
"step": 59
|
| 447 |
+
},
|
| 448 |
+
{
|
| 449 |
+
"epoch": 0.02564410786552871,
|
| 450 |
+
"grad_norm": 23.886445999145508,
|
| 451 |
+
"learning_rate": 5.996517316701069e-05,
|
| 452 |
+
"loss": 56.7736,
|
| 453 |
+
"step": 60
|
| 454 |
+
},
|
| 455 |
+
{
|
| 456 |
+
"epoch": 0.026071509663287522,
|
| 457 |
+
"grad_norm": 25.174198150634766,
|
| 458 |
+
"learning_rate": 5.646038740046304e-05,
|
| 459 |
+
"loss": 55.5788,
|
| 460 |
+
"step": 61
|
| 461 |
+
},
|
| 462 |
+
{
|
| 463 |
+
"epoch": 0.026498911461046333,
|
| 464 |
+
"grad_norm": 25.409194946289062,
|
| 465 |
+
"learning_rate": 5.315119735208132e-05,
|
| 466 |
+
"loss": 58.9748,
|
| 467 |
+
"step": 62
|
| 468 |
+
},
|
| 469 |
+
{
|
| 470 |
+
"epoch": 0.026926313258805144,
|
| 471 |
+
"grad_norm": 25.356067657470703,
|
| 472 |
+
"learning_rate": 5.004356640600297e-05,
|
| 473 |
+
"loss": 55.9752,
|
| 474 |
+
"step": 63
|
| 475 |
+
},
|
| 476 |
+
{
|
| 477 |
+
"epoch": 0.027353715056563956,
|
| 478 |
+
"grad_norm": 30.15899085998535,
|
| 479 |
+
"learning_rate": 4.71430947232396e-05,
|
| 480 |
+
"loss": 60.8193,
|
| 481 |
+
"step": 64
|
| 482 |
+
},
|
| 483 |
+
{
|
| 484 |
+
"epoch": 0.027781116854322767,
|
| 485 |
+
"grad_norm": 25.970443725585938,
|
| 486 |
+
"learning_rate": 4.4455009149808265e-05,
|
| 487 |
+
"loss": 58.5226,
|
| 488 |
+
"step": 65
|
| 489 |
+
},
|
| 490 |
+
{
|
| 491 |
+
"epoch": 0.02820851865208158,
|
| 492 |
+
"grad_norm": 26.306665420532227,
|
| 493 |
+
"learning_rate": 4.1984153797601665e-05,
|
| 494 |
+
"loss": 58.1114,
|
| 495 |
+
"step": 66
|
| 496 |
+
},
|
| 497 |
+
{
|
| 498 |
+
"epoch": 0.028635920449840393,
|
| 499 |
+
"grad_norm": 25.005186080932617,
|
| 500 |
+
"learning_rate": 3.9734981314971234e-05,
|
| 501 |
+
"loss": 58.226,
|
| 502 |
+
"step": 67
|
| 503 |
+
},
|
| 504 |
+
{
|
| 505 |
+
"epoch": 0.029063322247599204,
|
| 506 |
+
"grad_norm": 25.699512481689453,
|
| 507 |
+
"learning_rate": 3.771154486275363e-05,
|
| 508 |
+
"loss": 55.8354,
|
| 509 |
+
"step": 68
|
| 510 |
+
},
|
| 511 |
+
{
|
| 512 |
+
"epoch": 0.029490724045358015,
|
| 513 |
+
"grad_norm": 25.881290435791016,
|
| 514 |
+
"learning_rate": 3.591749081020113e-05,
|
| 515 |
+
"loss": 55.4757,
|
| 516 |
+
"step": 69
|
| 517 |
+
},
|
| 518 |
+
{
|
| 519 |
+
"epoch": 0.029918125843116827,
|
| 520 |
+
"grad_norm": 29.308368682861328,
|
| 521 |
+
"learning_rate": 3.435605216397765e-05,
|
| 522 |
+
"loss": 60.0276,
|
| 523 |
+
"step": 70
|
| 524 |
+
},
|
| 525 |
+
{
|
| 526 |
+
"epoch": 0.030345527640875638,
|
| 527 |
+
"grad_norm": 27.00686264038086,
|
| 528 |
+
"learning_rate": 3.303004274206237e-05,
|
| 529 |
+
"loss": 59.6951,
|
| 530 |
+
"step": 71
|
| 531 |
+
},
|
| 532 |
+
{
|
| 533 |
+
"epoch": 0.030772929438634453,
|
| 534 |
+
"grad_norm": 27.934207916259766,
|
| 535 |
+
"learning_rate": 3.194185210305936e-05,
|
| 536 |
+
"loss": 56.2346,
|
| 537 |
+
"step": 72
|
| 538 |
+
},
|
| 539 |
+
{
|
| 540 |
+
"epoch": 0.031200331236393264,
|
| 541 |
+
"grad_norm": 28.110820770263672,
|
| 542 |
+
"learning_rate": 3.1093441240051626e-05,
|
| 543 |
+
"loss": 59.4317,
|
| 544 |
+
"step": 73
|
| 545 |
+
},
|
| 546 |
+
{
|
| 547 |
+
"epoch": 0.031627733034152075,
|
| 548 |
+
"grad_norm": 28.240201950073242,
|
| 549 |
+
"learning_rate": 3.048633904675892e-05,
|
| 550 |
+
"loss": 58.142,
|
| 551 |
+
"step": 74
|
| 552 |
+
},
|
| 553 |
+
{
|
| 554 |
+
"epoch": 0.032055134831910886,
|
| 555 |
+
"grad_norm": 30.365863800048828,
|
| 556 |
+
"learning_rate": 3.012163956236801e-05,
|
| 557 |
+
"loss": 56.8291,
|
| 558 |
+
"step": 75
|
| 559 |
+
},
|
| 560 |
+
{
|
| 561 |
+
"epoch": 0.032055134831910886,
|
| 562 |
+
"eval_loss": 1.7859845161437988,
|
| 563 |
+
"eval_runtime": 14.2376,
|
| 564 |
+
"eval_samples_per_second": 3.512,
|
| 565 |
+
"eval_steps_per_second": 3.512,
|
| 566 |
+
"step": 75
|
| 567 |
+
},
|
| 568 |
+
{
|
| 569 |
+
"epoch": 0.0324825366296697,
|
| 570 |
+
"grad_norm": 31.245229721069336,
|
| 571 |
+
"learning_rate": 2.9999999999999997e-05,
|
| 572 |
+
"loss": 63.3069,
|
| 573 |
+
"step": 76
|
| 574 |
}
|
| 575 |
],
|
| 576 |
"logging_steps": 1,
|
|
|
|
| 594 |
"should_evaluate": false,
|
| 595 |
"should_log": false,
|
| 596 |
"should_save": true,
|
| 597 |
+
"should_training_stop": true
|
| 598 |
},
|
| 599 |
"attributes": {}
|
| 600 |
}
|
| 601 |
},
|
| 602 |
+
"total_flos": 2.2066295782087066e+17,
|
| 603 |
"train_batch_size": 1,
|
| 604 |
"trial_name": null,
|
| 605 |
"trial_params": null
|