| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 1746, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 1.142857142857143e-07, | |
| "loss": 2.7168, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 5.714285714285715e-07, | |
| "loss": 2.5728, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.142857142857143e-06, | |
| "loss": 2.5422, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 1.7142857142857145e-06, | |
| "loss": 2.3792, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.285714285714286e-06, | |
| "loss": 2.2176, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 2.8571428571428573e-06, | |
| "loss": 1.9282, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 3.428571428571429e-06, | |
| "loss": 1.7601, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.6887, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 4.571428571428572e-06, | |
| "loss": 1.6597, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 5.142857142857142e-06, | |
| "loss": 1.5152, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 5.7142857142857145e-06, | |
| "loss": 1.5448, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 6.285714285714286e-06, | |
| "loss": 1.5993, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 6.857142857142858e-06, | |
| "loss": 1.5941, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 7.428571428571429e-06, | |
| "loss": 1.5832, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.5369, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 8.571428571428571e-06, | |
| "loss": 1.4847, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 9.142857142857144e-06, | |
| "loss": 1.544, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 9.714285714285715e-06, | |
| "loss": 1.5631, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.0285714285714285e-05, | |
| "loss": 1.5124, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.0857142857142858e-05, | |
| "loss": 1.4762, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.1428571428571429e-05, | |
| "loss": 1.5304, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.2e-05, | |
| "loss": 1.5102, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.2571428571428572e-05, | |
| "loss": 1.477, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 1.3142857142857145e-05, | |
| "loss": 1.5082, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.3714285714285716e-05, | |
| "loss": 1.4684, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 1.4285714285714287e-05, | |
| "loss": 1.5267, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 1.4857142857142858e-05, | |
| "loss": 1.4454, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 1.542857142857143e-05, | |
| "loss": 1.4974, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 1.5087, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 1.6571428571428574e-05, | |
| "loss": 1.5172, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 1.7142857142857142e-05, | |
| "loss": 1.5034, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.7714285714285717e-05, | |
| "loss": 1.4841, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 1.8285714285714288e-05, | |
| "loss": 1.5138, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 1.885714285714286e-05, | |
| "loss": 1.5664, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 1.942857142857143e-05, | |
| "loss": 1.5134, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 2e-05, | |
| "loss": 1.5214, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 1.9999500133801672e-05, | |
| "loss": 1.502, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 1.9998000585179915e-05, | |
| "loss": 1.4694, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 1.9995501504049474e-05, | |
| "loss": 1.5153, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.9992003140251582e-05, | |
| "loss": 1.5056, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 1.9987505843529004e-05, | |
| "loss": 1.4324, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 1.998201006349106e-05, | |
| "loss": 1.4889, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 1.997551634956868e-05, | |
| "loss": 1.5285, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.9968025350959497e-05, | |
| "loss": 1.5191, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 1.9959537816562897e-05, | |
| "loss": 1.5196, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 1.9950054594905196e-05, | |
| "loss": 1.5304, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.9939576634054782e-05, | |
| "loss": 1.4943, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.992810498152735e-05, | |
| "loss": 1.5335, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 1.9915640784181163e-05, | |
| "loss": 1.4391, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 1.990218528810242e-05, | |
| "loss": 1.494, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 1.988773983848064e-05, | |
| "loss": 1.4153, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 1.9872305879474234e-05, | |
| "loss": 1.4412, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 1.9855884954066077e-05, | |
| "loss": 1.4328, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.983847870390928e-05, | |
| "loss": 1.4412, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 1.9820088869163066e-05, | |
| "loss": 1.4981, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 1.9800717288318792e-05, | |
| "loss": 1.4064, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 1.9780365898016147e-05, | |
| "loss": 1.4006, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 1.9759036732849552e-05, | |
| "loss": 1.4635, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 1.9736731925164755e-05, | |
| "loss": 1.4569, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 1.971345370484563e-05, | |
| "loss": 1.5035, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.9689204399091285e-05, | |
| "loss": 1.3976, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 1.9663986432183374e-05, | |
| "loss": 1.4406, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 1.963780232524374e-05, | |
| "loss": 1.4725, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 1.961065469598239e-05, | |
| "loss": 1.4824, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 1.958254625843577e-05, | |
| "loss": 1.5003, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.9553479822695435e-05, | |
| "loss": 1.4694, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 1.952345829462714e-05, | |
| "loss": 1.481, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.9492484675580302e-05, | |
| "loss": 1.4117, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 1.9460562062087963e-05, | |
| "loss": 1.4311, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 1.942769364555721e-05, | |
| "loss": 1.43, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.939388271195013e-05, | |
| "loss": 1.5002, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 1.935913264145529e-05, | |
| "loss": 1.4256, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 1.932344690814981e-05, | |
| "loss": 1.4574, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 1.928682907965207e-05, | |
| "loss": 1.4027, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.924928281676501e-05, | |
| "loss": 1.3318, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 1.9210811873110164e-05, | |
| "loss": 1.4903, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 1.9171420094752404e-05, | |
| "loss": 1.4705, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 1.9131111419815428e-05, | |
| "loss": 1.4515, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 1.9089889878088063e-05, | |
| "loss": 1.3972, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 1.904775959062137e-05, | |
| "loss": 1.4329, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 1.9004724769316682e-05, | |
| "loss": 1.4771, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.8960789716504502e-05, | |
| "loss": 1.4283, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 1.8915958824514388e-05, | |
| "loss": 1.4044, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 1.887023657523586e-05, | |
| "loss": 1.4577, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 1.8823627539670296e-05, | |
| "loss": 1.4222, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 1.877613637747398e-05, | |
| "loss": 1.4151, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.8727767836492256e-05, | |
| "loss": 1.425, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 1.8678526752284857e-05, | |
| "loss": 1.3253, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.8628418047642505e-05, | |
| "loss": 1.3933, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.8577446732094732e-05, | |
| "loss": 1.4658, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 1.8525617901409082e-05, | |
| "loss": 1.4172, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.8472936737081673e-05, | |
| "loss": 1.4866, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.841940850581917e-05, | |
| "loss": 1.4293, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 1.8365038559012263e-05, | |
| "loss": 1.4085, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.8309832332200677e-05, | |
| "loss": 1.421, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.8253795344529758e-05, | |
| "loss": 1.3329, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 1.81969331981987e-05, | |
| "loss": 1.3832, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.813925157790049e-05, | |
| "loss": 1.2692, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 1.8080756250253575e-05, | |
| "loss": 1.4182, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 1.802145306322537e-05, | |
| "loss": 1.3419, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 1.79613479455476e-05, | |
| "loss": 1.356, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 1.7900446906123604e-05, | |
| "loss": 1.3546, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.7838756033427585e-05, | |
| "loss": 1.3617, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 1.7776281494895956e-05, | |
| "loss": 1.3494, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 1.7713029536310722e-05, | |
| "loss": 1.382, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 1.76490064811751e-05, | |
| "loss": 1.3565, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 1.7584218730081325e-05, | |
| "loss": 1.3842, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 1.7518672760070764e-05, | |
| "loss": 1.369, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.7452375123986386e-05, | |
| "loss": 1.3536, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 1.7385332449817655e-05, | |
| "loss": 1.3606, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.73175514400379e-05, | |
| "loss": 1.3455, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 1.7249038870934263e-05, | |
| "loss": 1.3637, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 1.717980159193023e-05, | |
| "loss": 1.3794, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 1.7109846524900885e-05, | |
| "loss": 1.3189, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.7039180663480917e-05, | |
| "loss": 1.3252, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 1.6967811072365423e-05, | |
| "loss": 1.342, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 1.689574488660364e-05, | |
| "loss": 1.3497, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.4483790397644043, | |
| "eval_runtime": 133.8837, | |
| "eval_samples_per_second": 10.18, | |
| "eval_steps_per_second": 0.112, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.682298931088563e-05, | |
| "loss": 1.2855, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 1.6749551618822006e-05, | |
| "loss": 1.2097, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.6675439152216747e-05, | |
| "loss": 1.3221, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 1.6600659320333247e-05, | |
| "loss": 1.229, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 1.652521959915356e-05, | |
| "loss": 1.1311, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 1.6449127530631005e-05, | |
| "loss": 1.2322, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 1.6372390721936198e-05, | |
| "loss": 1.2631, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.6295016844696494e-05, | |
| "loss": 1.1868, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.621701363422907e-05, | |
| "loss": 1.1462, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 1.6138388888767587e-05, | |
| "loss": 1.0917, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 1.6059150468682558e-05, | |
| "loss": 1.0819, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 1.5979306295695547e-05, | |
| "loss": 1.1196, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.58988643520872e-05, | |
| "loss": 1.0547, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 1.5817832679899222e-05, | |
| "loss": 1.0307, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.5736219380130395e-05, | |
| "loss": 0.9733, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 1.5654032611926698e-05, | |
| "loss": 1.0215, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 1.5571280591765613e-05, | |
| "loss": 0.9941, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 1.548797159263467e-05, | |
| "loss": 0.9208, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 1.540411394320442e-05, | |
| "loss": 0.9202, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 1.531971602699574e-05, | |
| "loss": 0.9048, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.5234786281541736e-05, | |
| "loss": 0.9502, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.5149333197544203e-05, | |
| "loss": 0.8838, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.5063365318024791e-05, | |
| "loss": 0.9124, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 1.497689123747092e-05, | |
| "loss": 0.8446, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.4889919600976573e-05, | |
| "loss": 0.8778, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 1.4802459103378008e-05, | |
| "loss": 0.7858, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.4714518488384512e-05, | |
| "loss": 0.8087, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.4626106547704262e-05, | |
| "loss": 0.8072, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.4537232120165401e-05, | |
| "loss": 0.7525, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 1.4447904090832366e-05, | |
| "loss": 0.8278, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 1.4358131390117645e-05, | |
| "loss": 0.8115, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 1.4267922992888961e-05, | |
| "loss": 0.7743, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 1.4177287917572032e-05, | |
| "loss": 0.8078, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 1.4086235225248966e-05, | |
| "loss": 0.7173, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.3994774018752389e-05, | |
| "loss": 0.7239, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.3902913441755422e-05, | |
| "loss": 0.704, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 1.3810662677857548e-05, | |
| "loss": 0.7421, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 1.3718030949666488e-05, | |
| "loss": 0.7604, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 1.3625027517876216e-05, | |
| "loss": 0.7587, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 1.3531661680341098e-05, | |
| "loss": 0.7139, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 1.3437942771146388e-05, | |
| "loss": 0.715, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 1.3343880159675062e-05, | |
| "loss": 0.7455, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.3249483249671117e-05, | |
| "loss": 0.7436, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 1.3154761478299461e-05, | |
| "loss": 0.7172, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 1.3059724315202444e-05, | |
| "loss": 0.741, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 1.296438126155315e-05, | |
| "loss": 0.6807, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 1.286874184910553e-05, | |
| "loss": 0.7379, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 1.2772815639241491e-05, | |
| "loss": 0.7086, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 1.2676612222014999e-05, | |
| "loss": 0.7675, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.258014121519334e-05, | |
| "loss": 0.7193, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 1.2483412263295602e-05, | |
| "loss": 0.6671, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 1.2386435036628474e-05, | |
| "loss": 0.6785, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 1.228921923031948e-05, | |
| "loss": 0.6683, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 1.2191774563347727e-05, | |
| "loss": 0.7199, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 1.2094110777572257e-05, | |
| "loss": 0.7024, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.1996237636758133e-05, | |
| "loss": 0.6914, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.1898164925600316e-05, | |
| "loss": 0.6947, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 1.1799902448745466e-05, | |
| "loss": 0.6846, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 1.1701460029811733e-05, | |
| "loss": 0.6788, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 1.1602847510406657e-05, | |
| "loss": 0.6986, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 1.150407474914327e-05, | |
| "loss": 0.7083, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 1.1405151620654513e-05, | |
| "loss": 0.6188, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 1.1306088014606018e-05, | |
| "loss": 0.7138, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.1206893834707415e-05, | |
| "loss": 0.6972, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 1.110757899772222e-05, | |
| "loss": 0.7069, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 1.1008153432476436e-05, | |
| "loss": 0.7038, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 1.0908627078865928e-05, | |
| "loss": 0.6523, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 1.0809009886862695e-05, | |
| "loss": 0.6411, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 1.0709311815520151e-05, | |
| "loss": 0.6691, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 1.0609542831977478e-05, | |
| "loss": 0.6952, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 1.0509712910463175e-05, | |
| "loss": 0.7437, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.0409832031297912e-05, | |
| "loss": 0.7274, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 1.0309910179896762e-05, | |
| "loss": 0.6345, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 1.0209957345770924e-05, | |
| "loss": 0.6331, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 1.0109983521529036e-05, | |
| "loss": 0.6683, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 1.000999870187819e-05, | |
| "loss": 0.6999, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 9.910012882624726e-06, | |
| "loss": 0.6881, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 9.810036059674914e-06, | |
| "loss": 0.6735, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.710078228035635e-06, | |
| "loss": 0.6506, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 9.610149380815148e-06, | |
| "loss": 0.6896, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 9.510259508224052e-06, | |
| "loss": 0.6672, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 9.41041859657651e-06, | |
| "loss": 0.6851, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 9.310636627291915e-06, | |
| "loss": 0.6618, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 9.210923575897001e-06, | |
| "loss": 0.6249, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 9.111289411028545e-06, | |
| "loss": 0.631, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 9.0117440934368e-06, | |
| "loss": 0.6352, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 8.912297574989649e-06, | |
| "loss": 0.6613, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 8.812959797677718e-06, | |
| "loss": 0.6452, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 8.713740692620432e-06, | |
| "loss": 0.6671, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 8.614650179073152e-06, | |
| "loss": 0.6833, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 8.515698163435542e-06, | |
| "loss": 0.7135, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 8.41689453826118e-06, | |
| "loss": 0.6784, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 8.318249181268556e-06, | |
| "loss": 0.6886, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 8.219771954353597e-06, | |
| "loss": 0.6221, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 8.121472702603703e-06, | |
| "loss": 0.677, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 8.023361253313533e-06, | |
| "loss": 0.631, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 7.925447415002523e-06, | |
| "loss": 0.5897, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 7.827740976434287e-06, | |
| "loss": 0.6003, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 7.730251705638034e-06, | |
| "loss": 0.6457, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.632989348931993e-06, | |
| "loss": 0.6322, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 7.535963629949071e-06, | |
| "loss": 0.5849, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 7.439184248664713e-06, | |
| "loss": 0.6439, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 7.34266088042721e-06, | |
| "loss": 0.641, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 7.246403174990379e-06, | |
| "loss": 0.6496, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 7.15042075554888e-06, | |
| "loss": 0.6245, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 7.054723217776138e-06, | |
| "loss": 0.6483, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.959320128865026e-06, | |
| "loss": 0.6532, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 6.86422102657142e-06, | |
| "loss": 0.6116, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 6.76943541826067e-06, | |
| "loss": 0.5523, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 6.6749727799571016e-06, | |
| "loss": 0.6321, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 6.580842555396706e-06, | |
| "loss": 0.6502, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 6.487054155082973e-06, | |
| "loss": 0.6254, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 6.393616955346129e-06, | |
| "loss": 0.6547, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.3005402974057415e-06, | |
| "loss": 0.6172, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 6.207833486436832e-06, | |
| "loss": 0.6557, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 6.115505790639642e-06, | |
| "loss": 0.6709, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 1.5766805410385132, | |
| "eval_runtime": 133.3882, | |
| "eval_samples_per_second": 10.218, | |
| "eval_steps_per_second": 0.112, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6.023566440313023e-06, | |
| "loss": 0.559, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 5.932024626931689e-06, | |
| "loss": 0.5316, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 5.840889502227285e-06, | |
| "loss": 0.6037, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 5.750170177273471e-06, | |
| "loss": 0.5708, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 5.659875721575065e-06, | |
| "loss": 0.5343, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 5.570015162161331e-06, | |
| "loss": 0.5318, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 5.480597482683507e-06, | |
| "loss": 0.5536, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 5.391631622516697e-06, | |
| "loss": 0.5613, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 5.3031264758661595e-06, | |
| "loss": 0.5205, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 5.2150908908781335e-06, | |
| "loss": 0.5141, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 5.1275336687552554e-06, | |
| "loss": 0.4753, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 5.040463562876678e-06, | |
| "loss": 0.5046, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.9538892779229645e-06, | |
| "loss": 0.509, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 4.867819469005851e-06, | |
| "loss": 0.4434, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 4.782262740802975e-06, | |
| "loss": 0.4493, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 4.697227646697627e-06, | |
| "loss": 0.4474, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.612722687923658e-06, | |
| "loss": 0.4455, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 4.5287563127155545e-06, | |
| "loss": 0.43, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.445336915463874e-06, | |
| "loss": 0.4147, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 4.362472835876011e-06, | |
| "loss": 0.3961, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.280172358142448e-06, | |
| "loss": 0.4416, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 4.198443710108583e-06, | |
| "loss": 0.3929, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 4.117295062452122e-06, | |
| "loss": 0.4044, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 4.036734527866266e-06, | |
| "loss": 0.4289, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 3.956770160248658e-06, | |
| "loss": 0.383, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 3.8774099538961705e-06, | |
| "loss": 0.391, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.79866184270575e-06, | |
| "loss": 0.3592, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.720533699381177e-06, | |
| "loss": 0.3803, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 3.64303333464605e-06, | |
| "loss": 0.3274, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 3.5661684964629174e-06, | |
| "loss": 0.3652, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 3.4899468692586537e-06, | |
| "loss": 0.3681, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 3.414376073156275e-06, | |
| "loss": 0.3589, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 3.3394636632130893e-06, | |
| "loss": 0.3431, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 3.2652171286654032e-06, | |
| "loss": 0.3458, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 3.191643892179821e-06, | |
| "loss": 0.3163, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 3.1187513091111345e-06, | |
| "loss": 0.2961, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 3.0465466667670286e-06, | |
| "loss": 0.3045, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 2.975037183679519e-06, | |
| "loss": 0.3366, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 2.9042300088832877e-06, | |
| "loss": 0.3511, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 2.834132221200998e-06, | |
| "loss": 0.3254, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.7647508285355774e-06, | |
| "loss": 0.3208, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 2.6960927671696213e-06, | |
| "loss": 0.3209, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.6281649010719577e-06, | |
| "loss": 0.3605, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 2.5609740212114154e-06, | |
| "loss": 0.3405, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 2.494526844877938e-06, | |
| "loss": 0.3352, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 2.4288300150110077e-06, | |
| "loss": 0.307, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 2.3638900995355418e-06, | |
| "loss": 0.3171, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 2.299713590705275e-06, | |
| "loss": 0.3189, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 2.2363069044537044e-06, | |
| "loss": 0.3225, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 2.173676379752672e-06, | |
| "loss": 0.3645, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 2.1118282779786325e-06, | |
| "loss": 0.3216, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 2.0507687822866897e-06, | |
| "loss": 0.3094, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.9905039969924388e-06, | |
| "loss": 0.3119, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.9310399469617048e-06, | |
| "loss": 0.3193, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.8723825770082116e-06, | |
| "loss": 0.3117, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.8145377512992646e-06, | |
| "loss": 0.3199, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.7575112527694871e-06, | |
| "loss": 0.2935, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.7013087825426855e-06, | |
| "loss": 0.3258, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 1.645935959361884e-06, | |
| "loss": 0.3007, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 1.591398319027605e-06, | |
| "loss": 0.3276, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 1.5377013138444386e-06, | |
| "loss": 0.3204, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.4848503120759527e-06, | |
| "loss": 0.3157, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 1.432850597408012e-06, | |
| "loss": 0.2953, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 1.3817073684205562e-06, | |
| "loss": 0.2959, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 1.331425738067872e-06, | |
| "loss": 0.3428, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 1.2820107331674426e-06, | |
| "loss": 0.3364, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 1.2334672938973958e-06, | |
| "loss": 0.3162, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 1.1858002733026196e-06, | |
| "loss": 0.2815, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 1.139014436809589e-06, | |
| "loss": 0.3064, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 1.093114461749948e-06, | |
| "loss": 0.3043, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 1.048104936892902e-06, | |
| "loss": 0.3775, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 1.003990361986472e-06, | |
| "loss": 0.3461, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 9.607751473076232e-07, | |
| "loss": 0.3095, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 9.18463613221372e-07, | |
| "loss": 0.2902, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 8.770599897488563e-07, | |
| "loss": 0.3263, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 8.365684161444454e-07, | |
| "loss": 0.3094, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.969929404819377e-07, | |
| "loss": 0.3099, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 7.583375192498422e-07, | |
| "loss": 0.3441, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 7.206060169558482e-07, | |
| "loss": 0.3085, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.838022057404848e-07, | |
| "loss": 0.2788, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.479297649999839e-07, | |
| "loss": 0.3381, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 6.129922810184663e-07, | |
| "loss": 0.3003, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.789932466093884e-07, | |
| "loss": 0.3267, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.459360607663644e-07, | |
| "loss": 0.3103, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.138240283233631e-07, | |
| "loss": 0.2711, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 4.826603596242918e-07, | |
| "loss": 0.3007, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.5244817020207775e-07, | |
| "loss": 0.3087, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.231904804671727e-07, | |
| "loss": 0.3132, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 3.9489021540560114e-07, | |
| "loss": 0.3119, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.6755020428654974e-07, | |
| "loss": 0.2992, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 3.411731803794982e-07, | |
| "loss": 0.3365, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 3.1576178068098205e-07, | |
| "loss": 0.3505, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 2.913185456509537e-07, | |
| "loss": 0.3634, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 2.678459189588045e-07, | |
| "loss": 0.3027, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 2.453462472390711e-07, | |
| "loss": 0.3239, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 2.238217798568254e-07, | |
| "loss": 0.3222, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 2.0327466868280354e-07, | |
| "loss": 0.3058, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.8370696787827658e-07, | |
| "loss": 0.3013, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 1.6512063368968488e-07, | |
| "loss": 0.2995, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 1.4751752425307243e-07, | |
| "loss": 0.325, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 1.3089939940831854e-07, | |
| "loss": 0.3107, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 1.1526792052319769e-07, | |
| "loss": 0.3109, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 1.0062465032729896e-07, | |
| "loss": 0.3114, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 8.697105275577989e-08, | |
| "loss": 0.3613, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 7.430849280302465e-08, | |
| "loss": 0.3221, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 6.263823638617328e-08, | |
| "loss": 0.286, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 5.196145021856836e-08, | |
| "loss": 0.3689, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 4.2279201693110615e-08, | |
| "loss": 0.3032, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 3.359245877555428e-08, | |
| "loss": 0.3079, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 2.5902089907730022e-08, | |
| "loss": 0.3242, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.920886392072885e-08, | |
| "loss": 0.3382, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 1.3513449958034718e-08, | |
| "loss": 0.329, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 8.816417408634704e-09, | |
| "loss": 0.3636, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.11823585008675e-09, | |
| "loss": 0.3181, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 2.4192750015850088e-09, | |
| "loss": 0.3656, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 7.198046869882902e-10, | |
| "loss": 0.3502, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 1.999480784942165e-11, | |
| "loss": 0.3156, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 1.7387598752975464, | |
| "eval_runtime": 133.3136, | |
| "eval_samples_per_second": 10.224, | |
| "eval_steps_per_second": 0.113, | |
| "step": 1746 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 1746, | |
| "total_flos": 1230956927057920.0, | |
| "train_loss": 0.872046828952715, | |
| "train_runtime": 30104.3205, | |
| "train_samples_per_second": 2.783, | |
| "train_steps_per_second": 0.058 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 1746, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "total_flos": 1230956927057920.0, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |