| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.03548406174226743, | |
| "eval_steps": 180, | |
| "global_step": 900, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.444444444444445e-07, | |
| "loss": 1.4612, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 1.4717, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4.444444444444444e-06, | |
| "loss": 1.4887, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 1.4126, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 8.888888888888888e-06, | |
| "loss": 1.2639, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.1111111111111113e-05, | |
| "loss": 1.1005, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 1.0341, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.555555555555556e-05, | |
| "loss": 1.0016, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.7777777777777777e-05, | |
| "loss": 0.9879, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2e-05, | |
| "loss": 0.947, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.2222222222222227e-05, | |
| "loss": 0.9558, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.444444444444445e-05, | |
| "loss": 0.9208, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 0.9096, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 2.888888888888889e-05, | |
| "loss": 0.8924, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.111111111111112e-05, | |
| "loss": 0.8883, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 0.879, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.555555555555555e-05, | |
| "loss": 0.8643, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.777777777777778e-05, | |
| "loss": 0.8491, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 4e-05, | |
| "loss": 0.8485, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.999623940897003e-05, | |
| "loss": 0.8734, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.998495905008461e-05, | |
| "loss": 0.8522, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.996616316542537e-05, | |
| "loss": 0.8434, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.993985882335584e-05, | |
| "loss": 0.8312, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.9906055915863316e-05, | |
| "loss": 0.8248, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.9864767154838864e-05, | |
| "loss": 0.8387, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.9816008067296905e-05, | |
| "loss": 0.8191, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.9759796989536185e-05, | |
| "loss": 0.8047, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.9696155060244166e-05, | |
| "loss": 0.7999, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.9625106212547696e-05, | |
| "loss": 0.7841, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.9546677165012714e-05, | |
| "loss": 0.7546, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.946089741159648e-05, | |
| "loss": 0.7539, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.9367799210556124e-05, | |
| "loss": 0.7346, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.926741757231761e-05, | |
| "loss": 0.7315, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.915979024630978e-05, | |
| "loss": 0.744, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.904495770676831e-05, | |
| "loss": 0.7298, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.892296313751502e-05, | |
| "loss": 0.7385, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.879385241571817e-05, | |
| "loss": 0.7541, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "eval_loss": 0.7222402691841125, | |
| "eval_runtime": 118.1542, | |
| "eval_samples_per_second": 16.927, | |
| "eval_steps_per_second": 4.232, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.865767409464002e-05, | |
| "loss": 0.7311, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.8514479385377813e-05, | |
| "loss": 0.7376, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.8364322137605484e-05, | |
| "loss": 0.727, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.8207258819322936e-05, | |
| "loss": 0.739, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.804334849562076e-05, | |
| "loss": 0.741, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.787265280646825e-05, | |
| "loss": 0.7216, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.7695235943533155e-05, | |
| "loss": 0.7115, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.7511164626041823e-05, | |
| "loss": 0.731, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.732050807568878e-05, | |
| "loss": 0.7185, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.7123337990605335e-05, | |
| "loss": 0.729, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.691972851839682e-05, | |
| "loss": 0.7485, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.6709756228258735e-05, | |
| "loss": 0.7344, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.649350008218214e-05, | |
| "loss": 0.7527, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.6271041405259354e-05, | |
| "loss": 0.7381, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.604246385510088e-05, | |
| "loss": 0.7512, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.580785339037519e-05, | |
| "loss": 0.7368, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.5567298238483206e-05, | |
| "loss": 0.7228, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.532088886237956e-05, | |
| "loss": 0.7361, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.506871792655321e-05, | |
| "loss": 0.727, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.48108802621801e-05, | |
| "loss": 0.7198, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.4547472831460976e-05, | |
| "loss": 0.736, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.4278594691157985e-05, | |
| "loss": 0.7552, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.400434695534337e-05, | |
| "loss": 0.7316, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.372483275737468e-05, | |
| "loss": 0.7235, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.3440157211110454e-05, | |
| "loss": 0.7189, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.315042737138128e-05, | |
| "loss": 0.7358, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.285575219373079e-05, | |
| "loss": 0.7252, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.255624249344198e-05, | |
| "loss": 0.7157, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.2252010903864057e-05, | |
| "loss": 0.7175, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.194317183405573e-05, | |
| "loss": 0.7417, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.1629841425760534e-05, | |
| "loss": 0.736, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.1312137509730776e-05, | |
| "loss": 0.7194, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.0990179561416124e-05, | |
| "loss": 0.7222, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.066408865603383e-05, | |
| "loss": 0.7196, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.0333987423037262e-05, | |
| "loss": 0.714, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 3.0000000000000004e-05, | |
| "loss": 0.7126, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "eval_loss": 0.7118021249771118, | |
| "eval_runtime": 117.7706, | |
| "eval_samples_per_second": 16.982, | |
| "eval_steps_per_second": 4.246, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.9662251985932773e-05, | |
| "loss": 0.7159, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.9320870394050783e-05, | |
| "loss": 0.7249, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.897598360400925e-05, | |
| "loss": 0.7182, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 2.8627721313625073e-05, | |
| "loss": 0.7083, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.8276214490102788e-05, | |
| "loss": 0.7238, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.792159532078314e-05, | |
| "loss": 0.7137, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.7563997163432853e-05, | |
| "loss": 0.7182, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.720355449609421e-05, | |
| "loss": 0.7285, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.684040286651338e-05, | |
| "loss": 0.6988, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.6474678841166426e-05, | |
| "loss": 0.7217, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.6106519953902268e-05, | |
| "loss": 0.7185, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.5736064654221808e-05, | |
| "loss": 0.7303, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.536345225521275e-05, | |
| "loss": 0.7134, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.4988822881159627e-05, | |
| "loss": 0.7215, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.4612317414848804e-05, | |
| "loss": 0.7219, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.423407744458822e-05, | |
| "loss": 0.7253, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.3854245210961798e-05, | |
| "loss": 0.7251, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.3472963553338614e-05, | |
| "loss": 0.7259, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.3090375856156813e-05, | |
| "loss": 0.7412, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.2706625995002626e-05, | |
| "loss": 0.7316, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.2321858282504606e-05, | |
| "loss": 0.7166, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.1936217414063584e-05, | |
| "loss": 0.7453, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.154984841343862e-05, | |
| "loss": 0.7135, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.1162896578209517e-05, | |
| "loss": 0.712, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.077550742513634e-05, | |
| "loss": 0.7185, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.038782663543649e-05, | |
| "loss": 0.7366, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2e-05, | |
| "loss": 0.7302, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9612173364563517e-05, | |
| "loss": 0.7393, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9224492574863663e-05, | |
| "loss": 0.7296, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.8837103421790486e-05, | |
| "loss": 0.7109, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.8450151586561386e-05, | |
| "loss": 0.7092, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.806378258593642e-05, | |
| "loss": 0.7381, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.7678141717495394e-05, | |
| "loss": 0.733, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.7293374004997384e-05, | |
| "loss": 0.7068, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.6909624143843193e-05, | |
| "loss": 0.7062, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.6527036446661396e-05, | |
| "loss": 0.7342, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "eval_loss": 0.7100361585617065, | |
| "eval_runtime": 117.785, | |
| "eval_samples_per_second": 16.98, | |
| "eval_steps_per_second": 4.245, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.6145754789038205e-05, | |
| "loss": 0.7193, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.5765922555411793e-05, | |
| "loss": 0.7087, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.53876825851512e-05, | |
| "loss": 0.7228, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.5011177118840376e-05, | |
| "loss": 0.7254, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.4636547744787251e-05, | |
| "loss": 0.732, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.4263935345778202e-05, | |
| "loss": 0.7186, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.389348004609774e-05, | |
| "loss": 0.7091, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.3525321158833582e-05, | |
| "loss": 0.7278, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.3159597133486628e-05, | |
| "loss": 0.7088, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.2796445503905797e-05, | |
| "loss": 0.7188, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.2436002836567154e-05, | |
| "loss": 0.7311, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.2078404679216864e-05, | |
| "loss": 0.727, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.1723785509897219e-05, | |
| "loss": 0.7089, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.1372278686374935e-05, | |
| "loss": 0.7108, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.1024016395990758e-05, | |
| "loss": 0.7189, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.067912960594923e-05, | |
| "loss": 0.717, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.033774801406723e-05, | |
| "loss": 0.713, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.0000000000000006e-05, | |
| "loss": 0.7217, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.666012576962743e-06, | |
| "loss": 0.7099, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.33591134396618e-06, | |
| "loss": 0.6947, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.009820438583881e-06, | |
| "loss": 0.7051, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 8.687862490269232e-06, | |
| "loss": 0.7244, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 8.370158574239466e-06, | |
| "loss": 0.7233, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 8.056828165944282e-06, | |
| "loss": 0.712, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 7.747989096135943e-06, | |
| "loss": 0.7236, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 7.443757506558033e-06, | |
| "loss": 0.7134, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 7.1442478062692135e-06, | |
| "loss": 0.714, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.84957262861873e-06, | |
| "loss": 0.7058, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.559842788889552e-06, | |
| "loss": 0.7265, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.275167242625331e-06, | |
| "loss": 0.7119, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.9956530446566305e-06, | |
| "loss": 0.7263, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.721405308842023e-06, | |
| "loss": 0.7274, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.452527168539026e-06, | |
| "loss": 0.7166, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.189119737819912e-06, | |
| "loss": 0.713, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.9312820734467855e-06, | |
| "loss": 0.7253, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.679111137620442e-06, | |
| "loss": 0.7216, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "eval_loss": 0.708259105682373, | |
| "eval_runtime": 117.7068, | |
| "eval_samples_per_second": 16.991, | |
| "eval_steps_per_second": 4.248, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.4327017615168e-06, | |
| "loss": 0.7165, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.1921466096248164e-06, | |
| "loss": 0.7303, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.957536144899123e-06, | |
| "loss": 0.7198, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.7289585947406504e-06, | |
| "loss": 0.7186, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.5064999178178648e-06, | |
| "loss": 0.7438, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.290243771741275e-06, | |
| "loss": 0.7191, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.0802714816031787e-06, | |
| "loss": 0.7132, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.876662009394673e-06, | |
| "loss": 0.7278, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.679491924311226e-06, | |
| "loss": 0.7085, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.488835373958185e-06, | |
| "loss": 0.7115, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.304764056466844e-06, | |
| "loss": 0.7118, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.127347193531757e-06, | |
| "loss": 0.7121, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.9566515043792455e-06, | |
| "loss": 0.7174, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.792741180677069e-06, | |
| "loss": 0.6913, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.6356778623945223e-06, | |
| "loss": 0.7038, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.4855206146221934e-06, | |
| "loss": 0.7156, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.3423259053599891e-06, | |
| "loss": 0.7113, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.2061475842818337e-06, | |
| "loss": 0.727, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.0770368624849947e-06, | |
| "loss": 0.7186, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.550422932316938e-07, | |
| "loss": 0.7088, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 8.402097536902221e-07, | |
| "loss": 0.7125, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 7.325824276823934e-07, | |
| "loss": 0.7076, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.322007894438842e-07, | |
| "loss": 0.7099, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 5.391025884035239e-07, | |
| "loss": 0.7238, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.533228349872887e-07, | |
| "loss": 0.7241, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.748937874523062e-07, | |
| "loss": 0.7248, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.038449397558396e-07, | |
| "loss": 0.7083, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.402030104638198e-07, | |
| "loss": 0.7167, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.839919327030937e-07, | |
| "loss": 0.7277, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.3523284516113955e-07, | |
| "loss": 0.7075, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 9.394408413668343e-08, | |
| "loss": 0.7134, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 6.014117664415953e-08, | |
| "loss": 0.7147, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.383683457463649e-08, | |
| "loss": 0.7287, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 1.5040949915399173e-08, | |
| "loss": 0.7152, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.760591029973171e-09, | |
| "loss": 0.7195, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 0.0, | |
| "loss": 0.7171, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "eval_loss": 0.7084137797355652, | |
| "eval_runtime": 117.7342, | |
| "eval_samples_per_second": 16.987, | |
| "eval_steps_per_second": 4.247, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "step": 900, | |
| "total_flos": 1.1929120479106826e+18, | |
| "train_loss": 0.7602509791321225, | |
| "train_runtime": 22175.0023, | |
| "train_samples_per_second": 0.649, | |
| "train_steps_per_second": 0.041 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 900, | |
| "num_train_epochs": 1, | |
| "save_steps": 180, | |
| "total_flos": 1.1929120479106826e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |