| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.11470520761642579, |
| "eval_steps": 5, |
| "global_step": 500, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0011470520761642578, |
| "eval_loss": 2.755366325378418, |
| "eval_runtime": 14.0088, |
| "eval_samples_per_second": 469.989, |
| "eval_steps_per_second": 3.712, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.0022941041523285156, |
| "eval_loss": 2.750617027282715, |
| "eval_runtime": 14.0654, |
| "eval_samples_per_second": 468.099, |
| "eval_steps_per_second": 3.697, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0034411562284927736, |
| "eval_loss": 2.7424304485321045, |
| "eval_runtime": 14.1186, |
| "eval_samples_per_second": 466.336, |
| "eval_steps_per_second": 3.683, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.004588208304657031, |
| "eval_loss": 2.730889081954956, |
| "eval_runtime": 14.2451, |
| "eval_samples_per_second": 462.193, |
| "eval_steps_per_second": 3.65, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.005735260380821289, |
| "eval_loss": 2.7159576416015625, |
| "eval_runtime": 14.245, |
| "eval_samples_per_second": 462.199, |
| "eval_steps_per_second": 3.65, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.006882312456985547, |
| "eval_loss": 2.6975343227386475, |
| "eval_runtime": 14.3448, |
| "eval_samples_per_second": 458.981, |
| "eval_steps_per_second": 3.625, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.008029364533149804, |
| "eval_loss": 2.675693988800049, |
| "eval_runtime": 16.965, |
| "eval_samples_per_second": 388.094, |
| "eval_steps_per_second": 3.065, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.009176416609314062, |
| "eval_loss": 2.650242328643799, |
| "eval_runtime": 14.2414, |
| "eval_samples_per_second": 462.315, |
| "eval_steps_per_second": 3.651, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.01032346868547832, |
| "eval_loss": 2.6213560104370117, |
| "eval_runtime": 14.4168, |
| "eval_samples_per_second": 456.689, |
| "eval_steps_per_second": 3.607, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.011470520761642579, |
| "eval_loss": 2.5892951488494873, |
| "eval_runtime": 14.3332, |
| "eval_samples_per_second": 459.353, |
| "eval_steps_per_second": 3.628, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.012617572837806837, |
| "eval_loss": 2.5538384914398193, |
| "eval_runtime": 14.4067, |
| "eval_samples_per_second": 457.01, |
| "eval_steps_per_second": 3.609, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.013764624913971095, |
| "eval_loss": 2.51452374458313, |
| "eval_runtime": 14.4948, |
| "eval_samples_per_second": 454.233, |
| "eval_steps_per_second": 3.588, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.014911676990135353, |
| "eval_loss": 2.4726295471191406, |
| "eval_runtime": 14.3162, |
| "eval_samples_per_second": 459.898, |
| "eval_steps_per_second": 3.632, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.01605872906629961, |
| "eval_loss": 2.428196668624878, |
| "eval_runtime": 14.3204, |
| "eval_samples_per_second": 459.763, |
| "eval_steps_per_second": 3.631, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.017205781142463867, |
| "eval_loss": 2.379453659057617, |
| "eval_runtime": 14.3796, |
| "eval_samples_per_second": 457.872, |
| "eval_steps_per_second": 3.616, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.018352833218628125, |
| "eval_loss": 2.327221393585205, |
| "eval_runtime": 14.3415, |
| "eval_samples_per_second": 459.089, |
| "eval_steps_per_second": 3.626, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.019499885294792383, |
| "eval_loss": 2.2712388038635254, |
| "eval_runtime": 14.462, |
| "eval_samples_per_second": 455.263, |
| "eval_steps_per_second": 3.596, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.02064693737095664, |
| "eval_loss": 2.2120068073272705, |
| "eval_runtime": 14.334, |
| "eval_samples_per_second": 459.326, |
| "eval_steps_per_second": 3.628, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.0217939894471209, |
| "eval_loss": 2.1501331329345703, |
| "eval_runtime": 14.3506, |
| "eval_samples_per_second": 458.796, |
| "eval_steps_per_second": 3.624, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.022941041523285157, |
| "grad_norm": 19.895355224609375, |
| "learning_rate": 7.645259938837921e-07, |
| "loss": 3.6197, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.022941041523285157, |
| "eval_loss": 2.086596727371216, |
| "eval_runtime": 16.6465, |
| "eval_samples_per_second": 395.519, |
| "eval_steps_per_second": 3.124, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.024088093599449415, |
| "eval_loss": 2.0223236083984375, |
| "eval_runtime": 14.471, |
| "eval_samples_per_second": 454.979, |
| "eval_steps_per_second": 3.593, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.025235145675613673, |
| "eval_loss": 1.9571231603622437, |
| "eval_runtime": 14.3783, |
| "eval_samples_per_second": 457.913, |
| "eval_steps_per_second": 3.617, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.02638219775177793, |
| "eval_loss": 1.890655517578125, |
| "eval_runtime": 14.288, |
| "eval_samples_per_second": 460.805, |
| "eval_steps_per_second": 3.639, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.02752924982794219, |
| "eval_loss": 1.823920488357544, |
| "eval_runtime": 14.3764, |
| "eval_samples_per_second": 457.972, |
| "eval_steps_per_second": 3.617, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.028676301904106447, |
| "eval_loss": 1.758331298828125, |
| "eval_runtime": 14.2666, |
| "eval_samples_per_second": 461.499, |
| "eval_steps_per_second": 3.645, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.029823353980270705, |
| "eval_loss": 1.6937522888183594, |
| "eval_runtime": 14.2522, |
| "eval_samples_per_second": 461.963, |
| "eval_steps_per_second": 3.649, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.030970406056434963, |
| "eval_loss": 1.6316019296646118, |
| "eval_runtime": 16.3104, |
| "eval_samples_per_second": 403.67, |
| "eval_steps_per_second": 3.188, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.03211745813259922, |
| "eval_loss": 1.571895718574524, |
| "eval_runtime": 14.2422, |
| "eval_samples_per_second": 462.288, |
| "eval_steps_per_second": 3.651, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.033264510208763476, |
| "eval_loss": 1.5148202180862427, |
| "eval_runtime": 14.3624, |
| "eval_samples_per_second": 458.418, |
| "eval_steps_per_second": 3.621, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.034411562284927734, |
| "eval_loss": 1.459762454032898, |
| "eval_runtime": 14.3035, |
| "eval_samples_per_second": 460.308, |
| "eval_steps_per_second": 3.635, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.03555861436109199, |
| "eval_loss": 1.4081143140792847, |
| "eval_runtime": 14.2988, |
| "eval_samples_per_second": 460.46, |
| "eval_steps_per_second": 3.637, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.03670566643725625, |
| "eval_loss": 1.3612124919891357, |
| "eval_runtime": 14.2569, |
| "eval_samples_per_second": 461.812, |
| "eval_steps_per_second": 3.647, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.03785271851342051, |
| "eval_loss": 1.318212866783142, |
| "eval_runtime": 14.282, |
| "eval_samples_per_second": 460.999, |
| "eval_steps_per_second": 3.641, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.038999770589584766, |
| "eval_loss": 1.2802687883377075, |
| "eval_runtime": 14.4058, |
| "eval_samples_per_second": 457.038, |
| "eval_steps_per_second": 3.61, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.040146822665749024, |
| "eval_loss": 1.246294617652893, |
| "eval_runtime": 14.2804, |
| "eval_samples_per_second": 461.051, |
| "eval_steps_per_second": 3.641, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.04129387474191328, |
| "eval_loss": 1.2160167694091797, |
| "eval_runtime": 14.3449, |
| "eval_samples_per_second": 458.977, |
| "eval_steps_per_second": 3.625, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.04244092681807754, |
| "eval_loss": 1.189509630203247, |
| "eval_runtime": 14.2361, |
| "eval_samples_per_second": 462.486, |
| "eval_steps_per_second": 3.653, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.0435879788942418, |
| "eval_loss": 1.1653709411621094, |
| "eval_runtime": 14.3748, |
| "eval_samples_per_second": 458.025, |
| "eval_steps_per_second": 3.617, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.044735030970406056, |
| "eval_loss": 1.143513798713684, |
| "eval_runtime": 14.25, |
| "eval_samples_per_second": 462.035, |
| "eval_steps_per_second": 3.649, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.045882083046570314, |
| "grad_norm": 8.192963600158691, |
| "learning_rate": 1.5290519877675841e-06, |
| "loss": 2.292, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.045882083046570314, |
| "eval_loss": 1.1239805221557617, |
| "eval_runtime": 16.3012, |
| "eval_samples_per_second": 403.896, |
| "eval_steps_per_second": 3.19, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.04702913512273457, |
| "eval_loss": 1.1064891815185547, |
| "eval_runtime": 14.2673, |
| "eval_samples_per_second": 461.476, |
| "eval_steps_per_second": 3.645, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.04817618719889883, |
| "eval_loss": 1.0907284021377563, |
| "eval_runtime": 14.2601, |
| "eval_samples_per_second": 461.707, |
| "eval_steps_per_second": 3.647, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.04932323927506309, |
| "eval_loss": 1.076059341430664, |
| "eval_runtime": 14.2845, |
| "eval_samples_per_second": 460.919, |
| "eval_steps_per_second": 3.64, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.050470291351227346, |
| "eval_loss": 1.062280535697937, |
| "eval_runtime": 14.217, |
| "eval_samples_per_second": 463.109, |
| "eval_steps_per_second": 3.658, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.051617343427391604, |
| "eval_loss": 1.0492550134658813, |
| "eval_runtime": 14.2233, |
| "eval_samples_per_second": 462.901, |
| "eval_steps_per_second": 3.656, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.05276439550355586, |
| "eval_loss": 1.0374095439910889, |
| "eval_runtime": 14.5693, |
| "eval_samples_per_second": 451.909, |
| "eval_steps_per_second": 3.569, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.05391144757972012, |
| "eval_loss": 1.0259910821914673, |
| "eval_runtime": 15.9876, |
| "eval_samples_per_second": 411.819, |
| "eval_steps_per_second": 3.253, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.05505849965588438, |
| "eval_loss": 1.0147359371185303, |
| "eval_runtime": 14.2101, |
| "eval_samples_per_second": 463.332, |
| "eval_steps_per_second": 3.659, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.056205551732048636, |
| "eval_loss": 1.0042893886566162, |
| "eval_runtime": 14.2547, |
| "eval_samples_per_second": 461.882, |
| "eval_steps_per_second": 3.648, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.057352603808212894, |
| "eval_loss": 0.9941452741622925, |
| "eval_runtime": 14.2125, |
| "eval_samples_per_second": 463.253, |
| "eval_steps_per_second": 3.659, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.05849965588437715, |
| "eval_loss": 0.9848644733428955, |
| "eval_runtime": 14.2628, |
| "eval_samples_per_second": 461.621, |
| "eval_steps_per_second": 3.646, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.05964670796054141, |
| "eval_loss": 0.9763049483299255, |
| "eval_runtime": 14.2014, |
| "eval_samples_per_second": 463.617, |
| "eval_steps_per_second": 3.662, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.06079376003670567, |
| "eval_loss": 0.9682185649871826, |
| "eval_runtime": 16.4598, |
| "eval_samples_per_second": 400.004, |
| "eval_steps_per_second": 3.159, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.06194081211286993, |
| "eval_loss": 0.9602033495903015, |
| "eval_runtime": 14.3432, |
| "eval_samples_per_second": 459.031, |
| "eval_steps_per_second": 3.625, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.06308786418903418, |
| "eval_loss": 0.952538251876831, |
| "eval_runtime": 14.2143, |
| "eval_samples_per_second": 463.195, |
| "eval_steps_per_second": 3.658, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.06423491626519844, |
| "eval_loss": 0.9450673460960388, |
| "eval_runtime": 14.2195, |
| "eval_samples_per_second": 463.025, |
| "eval_steps_per_second": 3.657, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.0653819683413627, |
| "eval_loss": 0.937529981136322, |
| "eval_runtime": 14.2726, |
| "eval_samples_per_second": 461.303, |
| "eval_steps_per_second": 3.643, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.06652902041752695, |
| "eval_loss": 0.930277407169342, |
| "eval_runtime": 14.5409, |
| "eval_samples_per_second": 452.791, |
| "eval_steps_per_second": 3.576, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.06767607249369122, |
| "eval_loss": 0.9230740666389465, |
| "eval_runtime": 14.2552, |
| "eval_samples_per_second": 461.868, |
| "eval_steps_per_second": 3.648, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.06882312456985547, |
| "grad_norm": 7.243612766265869, |
| "learning_rate": 2.2935779816513764e-06, |
| "loss": 1.5711, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.06882312456985547, |
| "eval_loss": 0.915981650352478, |
| "eval_runtime": 16.3997, |
| "eval_samples_per_second": 401.471, |
| "eval_steps_per_second": 3.171, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.06997017664601973, |
| "eval_loss": 0.9087598323822021, |
| "eval_runtime": 14.1993, |
| "eval_samples_per_second": 463.685, |
| "eval_steps_per_second": 3.662, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.07111722872218398, |
| "eval_loss": 0.9022247791290283, |
| "eval_runtime": 14.3064, |
| "eval_samples_per_second": 460.214, |
| "eval_steps_per_second": 3.635, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.07226428079834825, |
| "eval_loss": 0.8950537443161011, |
| "eval_runtime": 14.2087, |
| "eval_samples_per_second": 463.377, |
| "eval_steps_per_second": 3.66, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.0734113328745125, |
| "eval_loss": 0.8874984383583069, |
| "eval_runtime": 14.2295, |
| "eval_samples_per_second": 462.699, |
| "eval_steps_per_second": 3.654, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.07455838495067676, |
| "eval_loss": 0.8809635043144226, |
| "eval_runtime": 14.2266, |
| "eval_samples_per_second": 462.796, |
| "eval_steps_per_second": 3.655, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.07570543702684102, |
| "eval_loss": 0.8745627403259277, |
| "eval_runtime": 14.2942, |
| "eval_samples_per_second": 460.608, |
| "eval_steps_per_second": 3.638, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.07685248910300528, |
| "eval_loss": 0.8683872818946838, |
| "eval_runtime": 14.1413, |
| "eval_samples_per_second": 465.586, |
| "eval_steps_per_second": 3.677, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.07799954117916953, |
| "eval_loss": 0.8624699115753174, |
| "eval_runtime": 14.1694, |
| "eval_samples_per_second": 464.662, |
| "eval_steps_per_second": 3.67, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.0791465932553338, |
| "eval_loss": 0.8569262027740479, |
| "eval_runtime": 14.2132, |
| "eval_samples_per_second": 463.231, |
| "eval_steps_per_second": 3.659, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.08029364533149805, |
| "eval_loss": 0.8515614867210388, |
| "eval_runtime": 14.3011, |
| "eval_samples_per_second": 460.383, |
| "eval_steps_per_second": 3.636, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.08144069740766231, |
| "eval_loss": 0.8465690612792969, |
| "eval_runtime": 14.2483, |
| "eval_samples_per_second": 462.09, |
| "eval_steps_per_second": 3.65, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.08258774948382656, |
| "eval_loss": 0.8418980836868286, |
| "eval_runtime": 14.1978, |
| "eval_samples_per_second": 463.735, |
| "eval_steps_per_second": 3.663, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.08373480155999083, |
| "eval_loss": 0.836972177028656, |
| "eval_runtime": 16.4774, |
| "eval_samples_per_second": 399.577, |
| "eval_steps_per_second": 3.156, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.08488185363615508, |
| "eval_loss": 0.8320812582969666, |
| "eval_runtime": 14.3098, |
| "eval_samples_per_second": 460.103, |
| "eval_steps_per_second": 3.634, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.08602890571231935, |
| "eval_loss": 0.8273819088935852, |
| "eval_runtime": 14.2755, |
| "eval_samples_per_second": 461.209, |
| "eval_steps_per_second": 3.643, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.0871759577884836, |
| "eval_loss": 0.8223117589950562, |
| "eval_runtime": 14.1567, |
| "eval_samples_per_second": 465.081, |
| "eval_steps_per_second": 3.673, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.08832300986464786, |
| "eval_loss": 0.8169983625411987, |
| "eval_runtime": 14.1912, |
| "eval_samples_per_second": 463.948, |
| "eval_steps_per_second": 3.664, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.08947006194081211, |
| "eval_loss": 0.811504602432251, |
| "eval_runtime": 14.1791, |
| "eval_samples_per_second": 464.347, |
| "eval_steps_per_second": 3.667, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.09061711401697638, |
| "eval_loss": 0.8054670095443726, |
| "eval_runtime": 14.2016, |
| "eval_samples_per_second": 463.611, |
| "eval_steps_per_second": 3.662, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.09176416609314063, |
| "grad_norm": 5.852241516113281, |
| "learning_rate": 3.0581039755351682e-06, |
| "loss": 1.3859, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.09176416609314063, |
| "eval_loss": 0.8003625273704529, |
| "eval_runtime": 16.3667, |
| "eval_samples_per_second": 402.281, |
| "eval_steps_per_second": 3.177, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.0929112181693049, |
| "eval_loss": 0.7954948544502258, |
| "eval_runtime": 14.1443, |
| "eval_samples_per_second": 465.489, |
| "eval_steps_per_second": 3.676, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.09405827024546914, |
| "eval_loss": 0.7906287312507629, |
| "eval_runtime": 14.2485, |
| "eval_samples_per_second": 462.082, |
| "eval_steps_per_second": 3.649, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.0952053223216334, |
| "eval_loss": 0.7856321930885315, |
| "eval_runtime": 14.2745, |
| "eval_samples_per_second": 461.242, |
| "eval_steps_per_second": 3.643, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.09635237439779766, |
| "eval_loss": 0.7809005379676819, |
| "eval_runtime": 14.2039, |
| "eval_samples_per_second": 463.536, |
| "eval_steps_per_second": 3.661, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.09749942647396191, |
| "eval_loss": 0.7759343981742859, |
| "eval_runtime": 14.2261, |
| "eval_samples_per_second": 462.813, |
| "eval_steps_per_second": 3.655, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.09864647855012618, |
| "eval_loss": 0.7706864476203918, |
| "eval_runtime": 16.5161, |
| "eval_samples_per_second": 398.642, |
| "eval_steps_per_second": 3.148, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.09979353062629043, |
| "eval_loss": 0.765408992767334, |
| "eval_runtime": 14.193, |
| "eval_samples_per_second": 463.89, |
| "eval_steps_per_second": 3.664, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.10094058270245469, |
| "eval_loss": 0.7602970600128174, |
| "eval_runtime": 14.2387, |
| "eval_samples_per_second": 462.402, |
| "eval_steps_per_second": 3.652, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.10208763477861894, |
| "eval_loss": 0.7555930018424988, |
| "eval_runtime": 14.1713, |
| "eval_samples_per_second": 464.601, |
| "eval_steps_per_second": 3.669, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.10323468685478321, |
| "eval_loss": 0.7515881061553955, |
| "eval_runtime": 14.2211, |
| "eval_samples_per_second": 462.973, |
| "eval_steps_per_second": 3.657, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.10438173893094746, |
| "eval_loss": 0.7474184036254883, |
| "eval_runtime": 14.1814, |
| "eval_samples_per_second": 464.27, |
| "eval_steps_per_second": 3.667, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.10552879100711172, |
| "eval_loss": 0.7430188059806824, |
| "eval_runtime": 14.2099, |
| "eval_samples_per_second": 463.34, |
| "eval_steps_per_second": 3.659, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.10667584308327598, |
| "eval_loss": 0.7378737330436707, |
| "eval_runtime": 16.2296, |
| "eval_samples_per_second": 405.68, |
| "eval_steps_per_second": 3.204, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.10782289515944024, |
| "eval_loss": 0.7332107424736023, |
| "eval_runtime": 14.1689, |
| "eval_samples_per_second": 464.68, |
| "eval_steps_per_second": 3.67, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.10896994723560449, |
| "eval_loss": 0.7287834286689758, |
| "eval_runtime": 14.4646, |
| "eval_samples_per_second": 455.18, |
| "eval_steps_per_second": 3.595, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.11011699931176876, |
| "eval_loss": 0.7247486710548401, |
| "eval_runtime": 14.1683, |
| "eval_samples_per_second": 464.699, |
| "eval_steps_per_second": 3.67, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.11126405138793301, |
| "eval_loss": 0.7211238741874695, |
| "eval_runtime": 14.235, |
| "eval_samples_per_second": 462.523, |
| "eval_steps_per_second": 3.653, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.11241110346409727, |
| "eval_loss": 0.7170650959014893, |
| "eval_runtime": 14.2367, |
| "eval_samples_per_second": 462.466, |
| "eval_steps_per_second": 3.653, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.11355815554026152, |
| "eval_loss": 0.7122666835784912, |
| "eval_runtime": 14.2574, |
| "eval_samples_per_second": 461.796, |
| "eval_steps_per_second": 3.647, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.11470520761642579, |
| "grad_norm": 5.7844438552856445, |
| "learning_rate": 3.8226299694189605e-06, |
| "loss": 1.2632, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.11470520761642579, |
| "eval_loss": 0.7078971862792969, |
| "eval_runtime": 14.238, |
| "eval_samples_per_second": 462.423, |
| "eval_steps_per_second": 3.652, |
| "step": 500 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 13077, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 3, |
| "save_steps": 5, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 128, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|