{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9980866629150253, "eval_steps": 500, "global_step": 1480, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0013505908835115363, "grad_norm": 2.148791476511835, "learning_rate": 6.756756756756757e-08, "loss": 2.1725, "step": 1 }, { "epoch": 0.006752954417557681, "grad_norm": 2.0693435263055955, "learning_rate": 3.378378378378379e-07, "loss": 2.1697, "step": 5 }, { "epoch": 0.013505908835115363, "grad_norm": 1.698535954485606, "learning_rate": 6.756756756756758e-07, "loss": 2.0659, "step": 10 }, { "epoch": 0.020258863252673044, "grad_norm": 0.9452543971000898, "learning_rate": 1.0135135135135136e-06, "loss": 1.8928, "step": 15 }, { "epoch": 0.027011817670230726, "grad_norm": 0.8038988537572226, "learning_rate": 1.3513513513513515e-06, "loss": 1.5709, "step": 20 }, { "epoch": 0.03376477208778841, "grad_norm": 0.471508019359059, "learning_rate": 1.6891891891891894e-06, "loss": 1.1256, "step": 25 }, { "epoch": 0.04051772650534609, "grad_norm": 0.252352654342694, "learning_rate": 2.0270270270270273e-06, "loss": 0.9097, "step": 30 }, { "epoch": 0.04727068092290377, "grad_norm": 0.24974204597161465, "learning_rate": 2.364864864864865e-06, "loss": 0.7799, "step": 35 }, { "epoch": 0.05402363534046145, "grad_norm": 0.20795816160607852, "learning_rate": 2.702702702702703e-06, "loss": 0.7137, "step": 40 }, { "epoch": 0.06077658975801913, "grad_norm": 0.17844205306863006, "learning_rate": 3.040540540540541e-06, "loss": 0.5982, "step": 45 }, { "epoch": 0.06752954417557681, "grad_norm": 0.14449057063114276, "learning_rate": 3.3783783783783788e-06, "loss": 0.5388, "step": 50 }, { "epoch": 0.0742824985931345, "grad_norm": 0.12180944441742095, "learning_rate": 3.7162162162162162e-06, "loss": 0.5621, "step": 55 }, { "epoch": 0.08103545301069218, "grad_norm": 0.11672624417203623, "learning_rate": 4.0540540540540545e-06, "loss": 0.4885, "step": 60 }, { "epoch": 0.08778840742824986, "grad_norm": 0.12365779514235553, "learning_rate": 4.391891891891892e-06, "loss": 0.4725, "step": 65 }, { "epoch": 0.09454136184580754, "grad_norm": 0.13848942929552646, "learning_rate": 4.72972972972973e-06, "loss": 0.4737, "step": 70 }, { "epoch": 0.10129431626336523, "grad_norm": 0.13947287893715526, "learning_rate": 5.067567567567568e-06, "loss": 0.4686, "step": 75 }, { "epoch": 0.1080472706809229, "grad_norm": 0.10498499488363304, "learning_rate": 5.405405405405406e-06, "loss": 0.4607, "step": 80 }, { "epoch": 0.11480022509848059, "grad_norm": 0.1263112338839764, "learning_rate": 5.743243243243244e-06, "loss": 0.471, "step": 85 }, { "epoch": 0.12155317951603826, "grad_norm": 0.1082136353357591, "learning_rate": 6.081081081081082e-06, "loss": 0.432, "step": 90 }, { "epoch": 0.12830613393359594, "grad_norm": 0.08923160665434038, "learning_rate": 6.41891891891892e-06, "loss": 0.4433, "step": 95 }, { "epoch": 0.13505908835115363, "grad_norm": 0.11807106476588053, "learning_rate": 6.7567567567567575e-06, "loss": 0.4398, "step": 100 }, { "epoch": 0.14181204276871132, "grad_norm": 0.09395401671145927, "learning_rate": 7.0945945945945946e-06, "loss": 0.3899, "step": 105 }, { "epoch": 0.148564997186269, "grad_norm": 0.09428015614908754, "learning_rate": 7.4324324324324324e-06, "loss": 0.3983, "step": 110 }, { "epoch": 0.15531795160382666, "grad_norm": 0.08605902096557952, "learning_rate": 7.77027027027027e-06, "loss": 0.415, "step": 115 }, { "epoch": 0.16207090602138435, "grad_norm": 0.08713323888522238, "learning_rate": 8.108108108108109e-06, "loss": 0.4075, "step": 120 }, { "epoch": 0.16882386043894204, "grad_norm": 0.07858143188282206, "learning_rate": 8.445945945945948e-06, "loss": 0.4089, "step": 125 }, { "epoch": 0.17557681485649973, "grad_norm": 0.07427563021554226, "learning_rate": 8.783783783783785e-06, "loss": 0.4291, "step": 130 }, { "epoch": 0.1823297692740574, "grad_norm": 0.09144835191366023, "learning_rate": 9.121621621621622e-06, "loss": 0.3895, "step": 135 }, { "epoch": 0.18908272369161508, "grad_norm": 0.07052586919720526, "learning_rate": 9.45945945945946e-06, "loss": 0.4029, "step": 140 }, { "epoch": 0.19583567810917277, "grad_norm": 0.0769561636008079, "learning_rate": 9.797297297297298e-06, "loss": 0.4148, "step": 145 }, { "epoch": 0.20258863252673046, "grad_norm": 0.07216832297932871, "learning_rate": 9.999944372378571e-06, "loss": 0.3801, "step": 150 }, { "epoch": 0.20934158694428812, "grad_norm": 0.06883498336890326, "learning_rate": 9.999318575852451e-06, "loss": 0.3728, "step": 155 }, { "epoch": 0.2160945413618458, "grad_norm": 0.07007134671747867, "learning_rate": 9.99799753559161e-06, "loss": 0.3831, "step": 160 }, { "epoch": 0.2228474957794035, "grad_norm": 0.0638944841323381, "learning_rate": 9.995981435310078e-06, "loss": 0.3852, "step": 165 }, { "epoch": 0.22960045019696118, "grad_norm": 0.0732472641559478, "learning_rate": 9.993270555382283e-06, "loss": 0.4058, "step": 170 }, { "epoch": 0.23635340461451884, "grad_norm": 0.06688107743191332, "learning_rate": 9.989865272804064e-06, "loss": 0.3981, "step": 175 }, { "epoch": 0.24310635903207653, "grad_norm": 0.0750542374720311, "learning_rate": 9.985766061140233e-06, "loss": 0.4034, "step": 180 }, { "epoch": 0.24985931344963422, "grad_norm": 0.06261496896671846, "learning_rate": 9.980973490458728e-06, "loss": 0.3611, "step": 185 }, { "epoch": 0.2566122678671919, "grad_norm": 0.07059993646852766, "learning_rate": 9.97548822725133e-06, "loss": 0.3603, "step": 190 }, { "epoch": 0.26336522228474957, "grad_norm": 0.06560938196431165, "learning_rate": 9.969311034340977e-06, "loss": 0.3853, "step": 195 }, { "epoch": 0.27011817670230726, "grad_norm": 0.05548643801780881, "learning_rate": 9.962442770775675e-06, "loss": 0.3659, "step": 200 }, { "epoch": 0.27687113111986494, "grad_norm": 0.06003482909108544, "learning_rate": 9.954884391709043e-06, "loss": 0.3907, "step": 205 }, { "epoch": 0.28362408553742263, "grad_norm": 0.04978856362198666, "learning_rate": 9.946636948267468e-06, "loss": 0.349, "step": 210 }, { "epoch": 0.2903770399549803, "grad_norm": 0.056553641270721766, "learning_rate": 9.937701587403941e-06, "loss": 0.4041, "step": 215 }, { "epoch": 0.297129994372538, "grad_norm": 0.06048923977199324, "learning_rate": 9.928079551738542e-06, "loss": 0.3704, "step": 220 }, { "epoch": 0.30388294879009564, "grad_norm": 0.055403005803656744, "learning_rate": 9.91777217938564e-06, "loss": 0.3765, "step": 225 }, { "epoch": 0.31063590320765333, "grad_norm": 0.05878900335636033, "learning_rate": 9.906780903767799e-06, "loss": 0.3927, "step": 230 }, { "epoch": 0.317388857625211, "grad_norm": 0.05009538620005328, "learning_rate": 9.895107253416434e-06, "loss": 0.3826, "step": 235 }, { "epoch": 0.3241418120427687, "grad_norm": 0.06193069860962348, "learning_rate": 9.882752851759247e-06, "loss": 0.3694, "step": 240 }, { "epoch": 0.3308947664603264, "grad_norm": 0.05658978100208689, "learning_rate": 9.869719416894462e-06, "loss": 0.3729, "step": 245 }, { "epoch": 0.3376477208778841, "grad_norm": 0.060376802387617515, "learning_rate": 9.856008761351882e-06, "loss": 0.373, "step": 250 }, { "epoch": 0.34440067529544177, "grad_norm": 0.047418407201082834, "learning_rate": 9.841622791840839e-06, "loss": 0.3343, "step": 255 }, { "epoch": 0.35115362971299946, "grad_norm": 0.062039985277230605, "learning_rate": 9.826563508985017e-06, "loss": 0.3897, "step": 260 }, { "epoch": 0.3579065841305571, "grad_norm": 0.04965955489714088, "learning_rate": 9.810833007044247e-06, "loss": 0.3925, "step": 265 }, { "epoch": 0.3646595385481148, "grad_norm": 0.05316606675382464, "learning_rate": 9.794433473623249e-06, "loss": 0.377, "step": 270 }, { "epoch": 0.37141249296567247, "grad_norm": 0.0467476855297103, "learning_rate": 9.777367189367412e-06, "loss": 0.3779, "step": 275 }, { "epoch": 0.37816544738323016, "grad_norm": 0.04978059836724344, "learning_rate": 9.759636527645633e-06, "loss": 0.3521, "step": 280 }, { "epoch": 0.38491840180078785, "grad_norm": 0.05999211649054394, "learning_rate": 9.74124395422025e-06, "loss": 0.3725, "step": 285 }, { "epoch": 0.39167135621834553, "grad_norm": 0.04365453179095536, "learning_rate": 9.722192026904145e-06, "loss": 0.3402, "step": 290 }, { "epoch": 0.3984243106359032, "grad_norm": 0.04616864773915026, "learning_rate": 9.702483395205023e-06, "loss": 0.3655, "step": 295 }, { "epoch": 0.4051772650534609, "grad_norm": 0.04782080283525874, "learning_rate": 9.682120799956961e-06, "loss": 0.3263, "step": 300 }, { "epoch": 0.41193021947101854, "grad_norm": 0.0498656978185221, "learning_rate": 9.661107072939244e-06, "loss": 0.3387, "step": 305 }, { "epoch": 0.41868317388857623, "grad_norm": 0.04142545069347494, "learning_rate": 9.639445136482549e-06, "loss": 0.3388, "step": 310 }, { "epoch": 0.4254361283061339, "grad_norm": 0.04183998822986712, "learning_rate": 9.61713800306255e-06, "loss": 0.3615, "step": 315 }, { "epoch": 0.4321890827236916, "grad_norm": 0.045933316046675596, "learning_rate": 9.594188774880981e-06, "loss": 0.3607, "step": 320 }, { "epoch": 0.4389420371412493, "grad_norm": 0.04805726955974, "learning_rate": 9.570600643434217e-06, "loss": 0.3371, "step": 325 }, { "epoch": 0.445694991558807, "grad_norm": 0.04569635226134762, "learning_rate": 9.546376889069442e-06, "loss": 0.3582, "step": 330 }, { "epoch": 0.4524479459763647, "grad_norm": 0.050347567122321094, "learning_rate": 9.521520880528453e-06, "loss": 0.3831, "step": 335 }, { "epoch": 0.45920090039392236, "grad_norm": 0.051704382901289696, "learning_rate": 9.496036074479184e-06, "loss": 0.3604, "step": 340 }, { "epoch": 0.46595385481148, "grad_norm": 0.04717995332220621, "learning_rate": 9.46992601503499e-06, "loss": 0.3498, "step": 345 }, { "epoch": 0.4727068092290377, "grad_norm": 0.04394756827213909, "learning_rate": 9.44319433326178e-06, "loss": 0.3711, "step": 350 }, { "epoch": 0.47945976364659537, "grad_norm": 0.05455789140928886, "learning_rate": 9.415844746673047e-06, "loss": 0.3395, "step": 355 }, { "epoch": 0.48621271806415306, "grad_norm": 0.0486039405429815, "learning_rate": 9.387881058712888e-06, "loss": 0.3596, "step": 360 }, { "epoch": 0.49296567248171075, "grad_norm": 0.049127971484592976, "learning_rate": 9.359307158227067e-06, "loss": 0.3759, "step": 365 }, { "epoch": 0.49971862689926844, "grad_norm": 0.05804448606366045, "learning_rate": 9.330127018922195e-06, "loss": 0.3741, "step": 370 }, { "epoch": 0.5064715813168261, "grad_norm": 0.05121055195790688, "learning_rate": 9.300344698813124e-06, "loss": 0.3978, "step": 375 }, { "epoch": 0.5132245357343838, "grad_norm": 0.05072015885221285, "learning_rate": 9.269964339658605e-06, "loss": 0.349, "step": 380 }, { "epoch": 0.5199774901519415, "grad_norm": 0.0446955462817659, "learning_rate": 9.238990166385304e-06, "loss": 0.3783, "step": 385 }, { "epoch": 0.5267304445694991, "grad_norm": 0.047869039864268624, "learning_rate": 9.207426486500252e-06, "loss": 0.3587, "step": 390 }, { "epoch": 0.5334833989870569, "grad_norm": 0.04183937339521655, "learning_rate": 9.175277689491804e-06, "loss": 0.3385, "step": 395 }, { "epoch": 0.5402363534046145, "grad_norm": 0.04804994911136886, "learning_rate": 9.142548246219212e-06, "loss": 0.3696, "step": 400 }, { "epoch": 0.5469893078221723, "grad_norm": 0.056116289553118946, "learning_rate": 9.109242708290864e-06, "loss": 0.3427, "step": 405 }, { "epoch": 0.5537422622397299, "grad_norm": 0.04610990276019498, "learning_rate": 9.075365707431311e-06, "loss": 0.3684, "step": 410 }, { "epoch": 0.5604952166572875, "grad_norm": 0.04711791241060553, "learning_rate": 9.040921954837139e-06, "loss": 0.3262, "step": 415 }, { "epoch": 0.5672481710748453, "grad_norm": 0.0436305028584867, "learning_rate": 9.005916240521788e-06, "loss": 0.3466, "step": 420 }, { "epoch": 0.5740011254924029, "grad_norm": 0.043193879987331325, "learning_rate": 8.97035343264943e-06, "loss": 0.3213, "step": 425 }, { "epoch": 0.5807540799099606, "grad_norm": 0.04807145092887773, "learning_rate": 8.93423847685795e-06, "loss": 0.3553, "step": 430 }, { "epoch": 0.5875070343275183, "grad_norm": 0.04707322735993806, "learning_rate": 8.89757639557118e-06, "loss": 0.3611, "step": 435 }, { "epoch": 0.594259988745076, "grad_norm": 0.04694876715909799, "learning_rate": 8.860372287300432e-06, "loss": 0.3703, "step": 440 }, { "epoch": 0.6010129431626337, "grad_norm": 0.051713244981409834, "learning_rate": 8.822631325935463e-06, "loss": 0.3763, "step": 445 }, { "epoch": 0.6077658975801913, "grad_norm": 0.07337197692235135, "learning_rate": 8.78435876002496e-06, "loss": 0.3867, "step": 450 }, { "epoch": 0.614518851997749, "grad_norm": 0.03984879157392941, "learning_rate": 8.745559912046625e-06, "loss": 0.3292, "step": 455 }, { "epoch": 0.6212718064153067, "grad_norm": 0.042340122733838116, "learning_rate": 8.706240177667003e-06, "loss": 0.3476, "step": 460 }, { "epoch": 0.6280247608328644, "grad_norm": 0.04565714030234626, "learning_rate": 8.666405024991105e-06, "loss": 0.3361, "step": 465 }, { "epoch": 0.634777715250422, "grad_norm": 0.04169016093573778, "learning_rate": 8.626059993801988e-06, "loss": 0.339, "step": 470 }, { "epoch": 0.6415306696679798, "grad_norm": 0.04325992397568874, "learning_rate": 8.585210694790333e-06, "loss": 0.3366, "step": 475 }, { "epoch": 0.6482836240855374, "grad_norm": 0.0415410087146755, "learning_rate": 8.543862808774193e-06, "loss": 0.3489, "step": 480 }, { "epoch": 0.6550365785030952, "grad_norm": 0.04286421536314813, "learning_rate": 8.502022085908963e-06, "loss": 0.3471, "step": 485 }, { "epoch": 0.6617895329206528, "grad_norm": 0.04783002216381401, "learning_rate": 8.459694344887732e-06, "loss": 0.4098, "step": 490 }, { "epoch": 0.6685424873382104, "grad_norm": 0.04187887777828664, "learning_rate": 8.416885472132077e-06, "loss": 0.3462, "step": 495 }, { "epoch": 0.6752954417557682, "grad_norm": 0.05373723455097553, "learning_rate": 8.373601420973464e-06, "loss": 0.344, "step": 500 }, { "epoch": 0.6820483961733258, "grad_norm": 0.04300150710853549, "learning_rate": 8.329848210825322e-06, "loss": 0.3647, "step": 505 }, { "epoch": 0.6888013505908835, "grad_norm": 0.043097229603458374, "learning_rate": 8.285631926345943e-06, "loss": 0.3693, "step": 510 }, { "epoch": 0.6955543050084412, "grad_norm": 0.040768739194901454, "learning_rate": 8.240958716592304e-06, "loss": 0.3645, "step": 515 }, { "epoch": 0.7023072594259989, "grad_norm": 0.046324281574150986, "learning_rate": 8.195834794164925e-06, "loss": 0.3501, "step": 520 }, { "epoch": 0.7090602138435566, "grad_norm": 0.04048770768699556, "learning_rate": 8.150266434343904e-06, "loss": 0.3296, "step": 525 }, { "epoch": 0.7158131682611142, "grad_norm": 0.04249502701153697, "learning_rate": 8.104259974216219e-06, "loss": 0.3489, "step": 530 }, { "epoch": 0.7225661226786719, "grad_norm": 0.03877822496803113, "learning_rate": 8.057821811794457e-06, "loss": 0.3673, "step": 535 }, { "epoch": 0.7293190770962296, "grad_norm": 0.048769655901584476, "learning_rate": 8.010958405127048e-06, "loss": 0.3313, "step": 540 }, { "epoch": 0.7360720315137873, "grad_norm": 0.07133171500587031, "learning_rate": 7.963676271400158e-06, "loss": 0.3472, "step": 545 }, { "epoch": 0.7428249859313449, "grad_norm": 0.04064256549350507, "learning_rate": 7.915981986031367e-06, "loss": 0.3478, "step": 550 }, { "epoch": 0.7495779403489027, "grad_norm": 0.04549548056733051, "learning_rate": 7.86788218175523e-06, "loss": 0.3342, "step": 555 }, { "epoch": 0.7563308947664603, "grad_norm": 0.03599813075890018, "learning_rate": 7.819383547700889e-06, "loss": 0.3354, "step": 560 }, { "epoch": 0.7630838491840181, "grad_norm": 0.04316680817634258, "learning_rate": 7.770492828461824e-06, "loss": 0.3622, "step": 565 }, { "epoch": 0.7698368036015757, "grad_norm": 0.04908046724152037, "learning_rate": 7.721216823157896e-06, "loss": 0.3559, "step": 570 }, { "epoch": 0.7765897580191333, "grad_norm": 0.040354310023923715, "learning_rate": 7.671562384489819e-06, "loss": 0.3629, "step": 575 }, { "epoch": 0.7833427124366911, "grad_norm": 0.043579845331084145, "learning_rate": 7.621536417786159e-06, "loss": 0.3404, "step": 580 }, { "epoch": 0.7900956668542487, "grad_norm": 0.04040670293206675, "learning_rate": 7.571145880043036e-06, "loss": 0.3672, "step": 585 }, { "epoch": 0.7968486212718064, "grad_norm": 0.04244690517322693, "learning_rate": 7.520397778956623e-06, "loss": 0.3499, "step": 590 }, { "epoch": 0.8036015756893641, "grad_norm": 0.05433806018620802, "learning_rate": 7.469299171948608e-06, "loss": 0.3704, "step": 595 }, { "epoch": 0.8103545301069218, "grad_norm": 0.04428365952002865, "learning_rate": 7.417857165184723e-06, "loss": 0.3474, "step": 600 }, { "epoch": 0.8171074845244795, "grad_norm": 0.03948151196873183, "learning_rate": 7.366078912586523e-06, "loss": 0.3426, "step": 605 }, { "epoch": 0.8238604389420371, "grad_norm": 0.03804562874118621, "learning_rate": 7.313971614836496e-06, "loss": 0.3418, "step": 610 }, { "epoch": 0.8306133933595948, "grad_norm": 0.04327583528687119, "learning_rate": 7.261542518376677e-06, "loss": 0.3408, "step": 615 }, { "epoch": 0.8373663477771525, "grad_norm": 0.038581581579252, "learning_rate": 7.208798914400916e-06, "loss": 0.3429, "step": 620 }, { "epoch": 0.8441193021947102, "grad_norm": 0.040274824329860945, "learning_rate": 7.155748137840892e-06, "loss": 0.385, "step": 625 }, { "epoch": 0.8508722566122678, "grad_norm": 0.043944187715932226, "learning_rate": 7.102397566346073e-06, "loss": 0.3366, "step": 630 }, { "epoch": 0.8576252110298256, "grad_norm": 0.04145675205661903, "learning_rate": 7.048754619257716e-06, "loss": 0.3755, "step": 635 }, { "epoch": 0.8643781654473832, "grad_norm": 0.041045163518307656, "learning_rate": 6.994826756577082e-06, "loss": 0.3722, "step": 640 }, { "epoch": 0.871131119864941, "grad_norm": 0.05061812503051895, "learning_rate": 6.940621477927988e-06, "loss": 0.3687, "step": 645 }, { "epoch": 0.8778840742824986, "grad_norm": 0.03703919699696153, "learning_rate": 6.88614632151385e-06, "loss": 0.3454, "step": 650 }, { "epoch": 0.8846370287000562, "grad_norm": 0.039635652764389524, "learning_rate": 6.831408863069364e-06, "loss": 0.365, "step": 655 }, { "epoch": 0.891389983117614, "grad_norm": 0.0428338033091146, "learning_rate": 6.7764167148069695e-06, "loss": 0.3251, "step": 660 }, { "epoch": 0.8981429375351716, "grad_norm": 0.0393255629299852, "learning_rate": 6.721177524358226e-06, "loss": 0.3519, "step": 665 }, { "epoch": 0.9048958919527293, "grad_norm": 0.03605877433184958, "learning_rate": 6.665698973710289e-06, "loss": 0.3586, "step": 670 }, { "epoch": 0.911648846370287, "grad_norm": 0.0418094941760349, "learning_rate": 6.609988778137582e-06, "loss": 0.333, "step": 675 }, { "epoch": 0.9184018007878447, "grad_norm": 0.04338138841178037, "learning_rate": 6.554054685128857e-06, "loss": 0.334, "step": 680 }, { "epoch": 0.9251547552054024, "grad_norm": 0.04225445092421321, "learning_rate": 6.497904473309766e-06, "loss": 0.3519, "step": 685 }, { "epoch": 0.93190770962296, "grad_norm": 0.04826057110998316, "learning_rate": 6.44154595136111e-06, "loss": 0.3598, "step": 690 }, { "epoch": 0.9386606640405177, "grad_norm": 0.04381442124287324, "learning_rate": 6.384986956932897e-06, "loss": 0.3339, "step": 695 }, { "epoch": 0.9454136184580754, "grad_norm": 0.0414974847898415, "learning_rate": 6.328235355554382e-06, "loss": 0.3645, "step": 700 }, { "epoch": 0.9521665728756331, "grad_norm": 0.04276356905517851, "learning_rate": 6.271299039540228e-06, "loss": 0.3645, "step": 705 }, { "epoch": 0.9589195272931907, "grad_norm": 0.046676300677672276, "learning_rate": 6.214185926892936e-06, "loss": 0.3551, "step": 710 }, { "epoch": 0.9656724817107485, "grad_norm": 0.04006191125365821, "learning_rate": 6.156903960201709e-06, "loss": 0.3395, "step": 715 }, { "epoch": 0.9724254361283061, "grad_norm": 0.047147919741739285, "learning_rate": 6.099461105537889e-06, "loss": 0.3321, "step": 720 }, { "epoch": 0.9791783905458639, "grad_norm": 0.041770572631853936, "learning_rate": 6.041865351347146e-06, "loss": 0.3421, "step": 725 }, { "epoch": 0.9859313449634215, "grad_norm": 0.04239922397171486, "learning_rate": 5.984124707338528e-06, "loss": 0.3373, "step": 730 }, { "epoch": 0.9926842993809791, "grad_norm": 0.046194589474773985, "learning_rate": 5.926247203370583e-06, "loss": 0.3405, "step": 735 }, { "epoch": 0.9994372537985369, "grad_norm": 0.03944962664459983, "learning_rate": 5.8682408883346535e-06, "loss": 0.2909, "step": 740 }, { "epoch": 1.0, "eval_loss": 0.3324998915195465, "eval_runtime": 18.1309, "eval_samples_per_second": 19.8, "eval_steps_per_second": 4.964, "step": 741 }, { "epoch": 1.005402363534046, "grad_norm": 0.040121304388969325, "learning_rate": 5.810113829035544e-06, "loss": 0.355, "step": 745 }, { "epoch": 1.0121553179516039, "grad_norm": 0.047689669773125845, "learning_rate": 5.751874109069685e-06, "loss": 0.3611, "step": 750 }, { "epoch": 1.0189082723691616, "grad_norm": 0.03912739831058518, "learning_rate": 5.693529827700967e-06, "loss": 0.2899, "step": 755 }, { "epoch": 1.0256612267867191, "grad_norm": 0.04377909856345864, "learning_rate": 5.635089098734394e-06, "loss": 0.3197, "step": 760 }, { "epoch": 1.0324141812042769, "grad_norm": 0.037393171944725745, "learning_rate": 5.576560049387717e-06, "loss": 0.339, "step": 765 }, { "epoch": 1.0391671356218346, "grad_norm": 0.042256918616508245, "learning_rate": 5.517950819161196e-06, "loss": 0.328, "step": 770 }, { "epoch": 1.0459200900393921, "grad_norm": 0.040529502097932234, "learning_rate": 5.459269558705667e-06, "loss": 0.3259, "step": 775 }, { "epoch": 1.0526730444569499, "grad_norm": 0.04379220562588465, "learning_rate": 5.400524428689035e-06, "loss": 0.3464, "step": 780 }, { "epoch": 1.0594259988745076, "grad_norm": 0.04725991662965075, "learning_rate": 5.341723598661409e-06, "loss": 0.3267, "step": 785 }, { "epoch": 1.0661789532920654, "grad_norm": 0.036832603953267457, "learning_rate": 5.282875245918963e-06, "loss": 0.3261, "step": 790 }, { "epoch": 1.0729319077096229, "grad_norm": 0.04175828917214628, "learning_rate": 5.2239875543667465e-06, "loss": 0.3346, "step": 795 }, { "epoch": 1.0796848621271806, "grad_norm": 0.03618010757790516, "learning_rate": 5.165068713380568e-06, "loss": 0.3182, "step": 800 }, { "epoch": 1.0864378165447384, "grad_norm": 0.04097386666815235, "learning_rate": 5.106126916668118e-06, "loss": 0.3192, "step": 805 }, { "epoch": 1.093190770962296, "grad_norm": 0.040300039979862916, "learning_rate": 5.047170361129484e-06, "loss": 0.3138, "step": 810 }, { "epoch": 1.0999437253798536, "grad_norm": 0.0413703774570022, "learning_rate": 4.988207245717232e-06, "loss": 0.3558, "step": 815 }, { "epoch": 1.1066966797974114, "grad_norm": 0.03938201981503368, "learning_rate": 4.929245770296191e-06, "loss": 0.3436, "step": 820 }, { "epoch": 1.1134496342149691, "grad_norm": 0.043903225983181984, "learning_rate": 4.870294134503123e-06, "loss": 0.3301, "step": 825 }, { "epoch": 1.1202025886325266, "grad_norm": 0.04208713272188743, "learning_rate": 4.811360536606416e-06, "loss": 0.3108, "step": 830 }, { "epoch": 1.1269555430500844, "grad_norm": 0.03660393810864938, "learning_rate": 4.752453172365966e-06, "loss": 0.333, "step": 835 }, { "epoch": 1.1337084974676421, "grad_norm": 0.04407131833701405, "learning_rate": 4.69358023389342e-06, "loss": 0.3605, "step": 840 }, { "epoch": 1.1404614518851997, "grad_norm": 0.04616999836259198, "learning_rate": 4.634749908512907e-06, "loss": 0.3439, "step": 845 }, { "epoch": 1.1472144063027574, "grad_norm": 0.04325365689860756, "learning_rate": 4.575970377622456e-06, "loss": 0.3049, "step": 850 }, { "epoch": 1.1539673607203151, "grad_norm": 0.03846831221970018, "learning_rate": 4.517249815556219e-06, "loss": 0.3511, "step": 855 }, { "epoch": 1.1607203151378729, "grad_norm": 0.04569268985609919, "learning_rate": 4.458596388447691e-06, "loss": 0.3209, "step": 860 }, { "epoch": 1.1674732695554304, "grad_norm": 0.0409252349150715, "learning_rate": 4.400018253094065e-06, "loss": 0.308, "step": 865 }, { "epoch": 1.1742262239729881, "grad_norm": 0.04133167625196089, "learning_rate": 4.341523555821882e-06, "loss": 0.3377, "step": 870 }, { "epoch": 1.180979178390546, "grad_norm": 0.043097502447341854, "learning_rate": 4.283120431354137e-06, "loss": 0.3117, "step": 875 }, { "epoch": 1.1877321328081036, "grad_norm": 0.042543275375795706, "learning_rate": 4.224817001679011e-06, "loss": 0.3344, "step": 880 }, { "epoch": 1.1944850872256612, "grad_norm": 0.04304997898343135, "learning_rate": 4.1666213749203545e-06, "loss": 0.3237, "step": 885 }, { "epoch": 1.201238041643219, "grad_norm": 0.04100222680465783, "learning_rate": 4.1085416442101205e-06, "loss": 0.3641, "step": 890 }, { "epoch": 1.2079909960607766, "grad_norm": 0.04523396323757555, "learning_rate": 4.050585886562858e-06, "loss": 0.3494, "step": 895 }, { "epoch": 1.2147439504783342, "grad_norm": 0.042513327919157136, "learning_rate": 3.992762161752474e-06, "loss": 0.343, "step": 900 }, { "epoch": 1.221496904895892, "grad_norm": 0.046594930027348866, "learning_rate": 3.935078511191368e-06, "loss": 0.3235, "step": 905 }, { "epoch": 1.2282498593134497, "grad_norm": 0.03906041128591151, "learning_rate": 3.877542956812137e-06, "loss": 0.3221, "step": 910 }, { "epoch": 1.2350028137310074, "grad_norm": 0.044515080037505254, "learning_rate": 3.820163499951984e-06, "loss": 0.3579, "step": 915 }, { "epoch": 1.241755768148565, "grad_norm": 0.04205841651552162, "learning_rate": 3.7629481202399886e-06, "loss": 0.3375, "step": 920 }, { "epoch": 1.2485087225661227, "grad_norm": 0.04152789361719496, "learning_rate": 3.705904774487396e-06, "loss": 0.2971, "step": 925 }, { "epoch": 1.2552616769836804, "grad_norm": 0.05137094607363216, "learning_rate": 3.64904139558109e-06, "loss": 0.3343, "step": 930 }, { "epoch": 1.2620146314012382, "grad_norm": 0.04499346319735512, "learning_rate": 3.5923658913803726e-06, "loss": 0.3274, "step": 935 }, { "epoch": 1.2687675858187957, "grad_norm": 0.04194049880241552, "learning_rate": 3.5358861436172487e-06, "loss": 0.3436, "step": 940 }, { "epoch": 1.2755205402363534, "grad_norm": 0.04739038652189242, "learning_rate": 3.47961000680032e-06, "loss": 0.3395, "step": 945 }, { "epoch": 1.2822734946539112, "grad_norm": 0.04407332378962217, "learning_rate": 3.4235453071224882e-06, "loss": 0.3283, "step": 950 }, { "epoch": 1.2890264490714687, "grad_norm": 0.046768041779961654, "learning_rate": 3.3676998413725726e-06, "loss": 0.3457, "step": 955 }, { "epoch": 1.2957794034890264, "grad_norm": 0.04405892619650361, "learning_rate": 3.3120813758510385e-06, "loss": 0.3446, "step": 960 }, { "epoch": 1.3025323579065842, "grad_norm": 0.04917585118916133, "learning_rate": 3.2566976452899507e-06, "loss": 0.3267, "step": 965 }, { "epoch": 1.3092853123241417, "grad_norm": 0.04327329891699119, "learning_rate": 3.2015563517773214e-06, "loss": 0.3382, "step": 970 }, { "epoch": 1.3160382667416994, "grad_norm": 0.04751650625794217, "learning_rate": 3.1466651636860025e-06, "loss": 0.3325, "step": 975 }, { "epoch": 1.3227912211592572, "grad_norm": 0.05680333767716324, "learning_rate": 3.0920317146072577e-06, "loss": 0.3207, "step": 980 }, { "epoch": 1.329544175576815, "grad_norm": 0.0497656718123479, "learning_rate": 3.0376636022891813e-06, "loss": 0.329, "step": 985 }, { "epoch": 1.3362971299943727, "grad_norm": 0.048529124206900136, "learning_rate": 2.983568387580094e-06, "loss": 0.3322, "step": 990 }, { "epoch": 1.3430500844119302, "grad_norm": 0.04475425551771912, "learning_rate": 2.9297535933770732e-06, "loss": 0.3278, "step": 995 }, { "epoch": 1.349803038829488, "grad_norm": 0.043952297509005415, "learning_rate": 2.8762267035797607e-06, "loss": 0.2961, "step": 1000 }, { "epoch": 1.3565559932470457, "grad_norm": 0.04672219098205047, "learning_rate": 2.822995162049599e-06, "loss": 0.3618, "step": 1005 }, { "epoch": 1.3633089476646032, "grad_norm": 0.04260321337685024, "learning_rate": 2.7700663715746213e-06, "loss": 0.3382, "step": 1010 }, { "epoch": 1.370061902082161, "grad_norm": 0.048008697667558865, "learning_rate": 2.7174476928399685e-06, "loss": 0.3466, "step": 1015 }, { "epoch": 1.3768148564997187, "grad_norm": 0.04892067375449484, "learning_rate": 2.66514644340426e-06, "loss": 0.3425, "step": 1020 }, { "epoch": 1.3835678109172762, "grad_norm": 0.04689882121063575, "learning_rate": 2.613169896681949e-06, "loss": 0.2877, "step": 1025 }, { "epoch": 1.390320765334834, "grad_norm": 0.054157177491435074, "learning_rate": 2.5615252809318287e-06, "loss": 0.3306, "step": 1030 }, { "epoch": 1.3970737197523917, "grad_norm": 0.048672772164928886, "learning_rate": 2.5102197782518145e-06, "loss": 0.3216, "step": 1035 }, { "epoch": 1.4038266741699492, "grad_norm": 0.06752970884761282, "learning_rate": 2.4592605235801544e-06, "loss": 0.3276, "step": 1040 }, { "epoch": 1.410579628587507, "grad_norm": 0.04710574868399225, "learning_rate": 2.4086546037031734e-06, "loss": 0.3173, "step": 1045 }, { "epoch": 1.4173325830050647, "grad_norm": 0.045184769590470804, "learning_rate": 2.3584090562697427e-06, "loss": 0.3164, "step": 1050 }, { "epoch": 1.4240855374226224, "grad_norm": 0.043485542556231396, "learning_rate": 2.3085308688125695e-06, "loss": 0.3027, "step": 1055 }, { "epoch": 1.4308384918401802, "grad_norm": 0.05260603355819636, "learning_rate": 2.2590269777764516e-06, "loss": 0.3004, "step": 1060 }, { "epoch": 1.4375914462577377, "grad_norm": 0.051099932583330926, "learning_rate": 2.2099042675536437e-06, "loss": 0.3368, "step": 1065 }, { "epoch": 1.4443444006752955, "grad_norm": 0.045328406677686464, "learning_rate": 2.161169569526461e-06, "loss": 0.3056, "step": 1070 }, { "epoch": 1.4510973550928532, "grad_norm": 0.05676490778567245, "learning_rate": 2.1128296611172593e-06, "loss": 0.3052, "step": 1075 }, { "epoch": 1.4578503095104107, "grad_norm": 0.044797358674842205, "learning_rate": 2.0648912648459072e-06, "loss": 0.3447, "step": 1080 }, { "epoch": 1.4646032639279685, "grad_norm": 0.049289994897252776, "learning_rate": 2.0173610473949048e-06, "loss": 0.3375, "step": 1085 }, { "epoch": 1.4713562183455262, "grad_norm": 0.04706434397429944, "learning_rate": 1.9702456186822595e-06, "loss": 0.3464, "step": 1090 }, { "epoch": 1.4781091727630837, "grad_norm": 0.05030965485900401, "learning_rate": 1.9235515309422685e-06, "loss": 0.3119, "step": 1095 }, { "epoch": 1.4848621271806415, "grad_norm": 0.04413503044032754, "learning_rate": 1.8772852778143064e-06, "loss": 0.3371, "step": 1100 }, { "epoch": 1.4916150815981992, "grad_norm": 0.06233556882470439, "learning_rate": 1.831453293439771e-06, "loss": 0.3253, "step": 1105 }, { "epoch": 1.498368036015757, "grad_norm": 0.05282578176334435, "learning_rate": 1.7860619515673034e-06, "loss": 0.3032, "step": 1110 }, { "epoch": 1.5051209904333147, "grad_norm": 0.05220647521633855, "learning_rate": 1.7411175646664103e-06, "loss": 0.3161, "step": 1115 }, { "epoch": 1.5118739448508722, "grad_norm": 0.047487304887968124, "learning_rate": 1.6966263830495939e-06, "loss": 0.3104, "step": 1120 }, { "epoch": 1.51862689926843, "grad_norm": 0.05098969364978786, "learning_rate": 1.6525945940031407e-06, "loss": 0.3692, "step": 1125 }, { "epoch": 1.5253798536859877, "grad_norm": 0.04810126110395747, "learning_rate": 1.6090283209266682e-06, "loss": 0.3383, "step": 1130 }, { "epoch": 1.5321328081035452, "grad_norm": 0.06656533903224827, "learning_rate": 1.5659336224815642e-06, "loss": 0.3413, "step": 1135 }, { "epoch": 1.538885762521103, "grad_norm": 0.057091145465475585, "learning_rate": 1.5233164917484117e-06, "loss": 0.3247, "step": 1140 }, { "epoch": 1.5456387169386607, "grad_norm": 0.05098329100385167, "learning_rate": 1.4811828553935498e-06, "loss": 0.3284, "step": 1145 }, { "epoch": 1.5523916713562183, "grad_norm": 0.04852633539753427, "learning_rate": 1.439538572844873e-06, "loss": 0.307, "step": 1150 }, { "epoch": 1.559144625773776, "grad_norm": 0.05935932625720792, "learning_rate": 1.3983894354769616e-06, "loss": 0.3247, "step": 1155 }, { "epoch": 1.5658975801913337, "grad_norm": 0.06306295104638487, "learning_rate": 1.3577411658056965e-06, "loss": 0.3286, "step": 1160 }, { "epoch": 1.5726505346088913, "grad_norm": 0.06348638606124773, "learning_rate": 1.3175994166924394e-06, "loss": 0.3247, "step": 1165 }, { "epoch": 1.5794034890264492, "grad_norm": 0.051811719215914646, "learning_rate": 1.2779697705579058e-06, "loss": 0.3246, "step": 1170 }, { "epoch": 1.5861564434440067, "grad_norm": 0.06830322751813911, "learning_rate": 1.2388577386058248e-06, "loss": 0.3326, "step": 1175 }, { "epoch": 1.5929093978615643, "grad_norm": 0.04922857249304829, "learning_rate": 1.2002687600565138e-06, "loss": 0.3061, "step": 1180 }, { "epoch": 1.5996623522791222, "grad_norm": 0.047108708806732846, "learning_rate": 1.1622082013904535e-06, "loss": 0.3118, "step": 1185 }, { "epoch": 1.6064153066966798, "grad_norm": 0.059687932246599115, "learning_rate": 1.1246813556019925e-06, "loss": 0.3372, "step": 1190 }, { "epoch": 1.6131682611142375, "grad_norm": 0.05955615759336213, "learning_rate": 1.0876934414632523e-06, "loss": 0.2956, "step": 1195 }, { "epoch": 1.6199212155317952, "grad_norm": 0.06542544159973171, "learning_rate": 1.0512496027983715e-06, "loss": 0.3118, "step": 1200 }, { "epoch": 1.6266741699493528, "grad_norm": 0.05198203822638065, "learning_rate": 1.0153549077681617e-06, "loss": 0.346, "step": 1205 }, { "epoch": 1.6334271243669105, "grad_norm": 0.06101598768362396, "learning_rate": 9.80014348165298e-07, "loss": 0.2952, "step": 1210 }, { "epoch": 1.6401800787844683, "grad_norm": 0.05552589226769097, "learning_rate": 9.452328387201104e-07, "loss": 0.2903, "step": 1215 }, { "epoch": 1.6469330332020258, "grad_norm": 0.06472026824610987, "learning_rate": 9.110152164171127e-07, "loss": 0.3081, "step": 1220 }, { "epoch": 1.6536859876195835, "grad_norm": 0.07431179400574185, "learning_rate": 8.773662398223276e-07, "loss": 0.3095, "step": 1225 }, { "epoch": 1.6604389420371413, "grad_norm": 0.06752269915658678, "learning_rate": 8.44290588421533e-07, "loss": 0.3187, "step": 1230 }, { "epoch": 1.6671918964546988, "grad_norm": 0.06578494925550707, "learning_rate": 8.117928619694848e-07, "loss": 0.3036, "step": 1235 }, { "epoch": 1.6739448508722568, "grad_norm": 0.06180051968265024, "learning_rate": 7.798775798502484e-07, "loss": 0.3023, "step": 1240 }, { "epoch": 1.6806978052898143, "grad_norm": 0.07560845712778133, "learning_rate": 7.485491804486972e-07, "loss": 0.2951, "step": 1245 }, { "epoch": 1.687450759707372, "grad_norm": 0.07194010404747912, "learning_rate": 7.178120205332717e-07, "loss": 0.3372, "step": 1250 }, { "epoch": 1.6942037141249298, "grad_norm": 0.06976609157173444, "learning_rate": 6.876703746500984e-07, "loss": 0.2973, "step": 1255 }, { "epoch": 1.7009566685424873, "grad_norm": 0.07659553993307627, "learning_rate": 6.581284345285371e-07, "loss": 0.284, "step": 1260 }, { "epoch": 1.707709622960045, "grad_norm": 0.05961865699755965, "learning_rate": 6.291903084982481e-07, "loss": 0.3153, "step": 1265 }, { "epoch": 1.7144625773776028, "grad_norm": 0.0812482207068001, "learning_rate": 6.008600209178539e-07, "loss": 0.342, "step": 1270 }, { "epoch": 1.7212155317951603, "grad_norm": 0.06939464438812673, "learning_rate": 5.7314151161528e-07, "loss": 0.3287, "step": 1275 }, { "epoch": 1.727968486212718, "grad_norm": 0.06615067862128174, "learning_rate": 5.460386353398583e-07, "loss": 0.3036, "step": 1280 }, { "epoch": 1.7347214406302758, "grad_norm": 0.08487351817752944, "learning_rate": 5.195551612262478e-07, "loss": 0.3059, "step": 1285 }, { "epoch": 1.7414743950478333, "grad_norm": 0.0660084040820942, "learning_rate": 4.936947722702762e-07, "loss": 0.3004, "step": 1290 }, { "epoch": 1.7482273494653913, "grad_norm": 0.07500490253176417, "learning_rate": 4.6846106481675035e-07, "loss": 0.3163, "step": 1295 }, { "epoch": 1.7549803038829488, "grad_norm": 0.07341914462851476, "learning_rate": 4.43857548059321e-07, "loss": 0.3028, "step": 1300 }, { "epoch": 1.7617332583005063, "grad_norm": 0.05936547346570651, "learning_rate": 4.198876435524718e-07, "loss": 0.3071, "step": 1305 }, { "epoch": 1.7684862127180643, "grad_norm": 0.06361743408706552, "learning_rate": 3.9655468473568436e-07, "loss": 0.3151, "step": 1310 }, { "epoch": 1.7752391671356218, "grad_norm": 0.07650389023350329, "learning_rate": 3.7386191646987094e-07, "loss": 0.333, "step": 1315 }, { "epoch": 1.7819921215531795, "grad_norm": 0.06885824668728105, "learning_rate": 3.51812494586114e-07, "loss": 0.2898, "step": 1320 }, { "epoch": 1.7887450759707373, "grad_norm": 0.0645999509118337, "learning_rate": 3.3040948544679817e-07, "loss": 0.3091, "step": 1325 }, { "epoch": 1.7954980303882948, "grad_norm": 0.07732866321608588, "learning_rate": 3.096558655191706e-07, "loss": 0.3148, "step": 1330 }, { "epoch": 1.8022509848058526, "grad_norm": 0.0733331982280887, "learning_rate": 2.895545209614176e-07, "loss": 0.3101, "step": 1335 }, { "epoch": 1.8090039392234103, "grad_norm": 0.06997792694817997, "learning_rate": 2.701082472212879e-07, "loss": 0.3302, "step": 1340 }, { "epoch": 1.8157568936409678, "grad_norm": 0.07837574138979145, "learning_rate": 2.5131974864734063e-07, "loss": 0.3425, "step": 1345 }, { "epoch": 1.8225098480585256, "grad_norm": 0.07239388554199513, "learning_rate": 2.331916381128535e-07, "loss": 0.3054, "step": 1350 }, { "epoch": 1.8292628024760833, "grad_norm": 0.0674988666661417, "learning_rate": 2.157264366524603e-07, "loss": 0.28, "step": 1355 }, { "epoch": 1.8360157568936408, "grad_norm": 0.07181164697688498, "learning_rate": 1.989265731115525e-07, "loss": 0.3307, "step": 1360 }, { "epoch": 1.8427687113111988, "grad_norm": 0.07548202922443094, "learning_rate": 1.827943838085111e-07, "loss": 0.306, "step": 1365 }, { "epoch": 1.8495216657287563, "grad_norm": 0.05982075375718766, "learning_rate": 1.6733211220979316e-07, "loss": 0.2981, "step": 1370 }, { "epoch": 1.856274620146314, "grad_norm": 0.08355451454560252, "learning_rate": 1.5254190861794415e-07, "loss": 0.3083, "step": 1375 }, { "epoch": 1.8630275745638718, "grad_norm": 0.07033662219869818, "learning_rate": 1.3842582987255494e-07, "loss": 0.293, "step": 1380 }, { "epoch": 1.8697805289814293, "grad_norm": 0.06615537508166416, "learning_rate": 1.2498583906422779e-07, "loss": 0.3072, "step": 1385 }, { "epoch": 1.876533483398987, "grad_norm": 0.059237664909188886, "learning_rate": 1.1222380526156929e-07, "loss": 0.313, "step": 1390 }, { "epoch": 1.8832864378165448, "grad_norm": 0.08185957993390121, "learning_rate": 1.0014150325126315e-07, "loss": 0.3071, "step": 1395 }, { "epoch": 1.8900393922341023, "grad_norm": 0.06113218790264024, "learning_rate": 8.874061329125939e-08, "loss": 0.3116, "step": 1400 }, { "epoch": 1.89679234665166, "grad_norm": 0.08593985138096456, "learning_rate": 7.802272087709951e-08, "loss": 0.3106, "step": 1405 }, { "epoch": 1.9035453010692178, "grad_norm": 0.0753339474248732, "learning_rate": 6.798931652142738e-08, "loss": 0.3118, "step": 1410 }, { "epoch": 1.9102982554867753, "grad_norm": 0.06406340322576629, "learning_rate": 5.864179554670724e-08, "loss": 0.3384, "step": 1415 }, { "epoch": 1.9170512099043333, "grad_norm": 0.06739406393286214, "learning_rate": 4.998145789118114e-08, "loss": 0.2755, "step": 1420 }, { "epoch": 1.9238041643218908, "grad_norm": 0.07101405306951455, "learning_rate": 4.2009507928084116e-08, "loss": 0.2802, "step": 1425 }, { "epoch": 1.9305571187394484, "grad_norm": 0.07509273619426268, "learning_rate": 3.4727054298161475e-08, "loss": 0.2778, "step": 1430 }, { "epoch": 1.9373100731570063, "grad_norm": 0.0719074660592959, "learning_rate": 2.8135109755487723e-08, "loss": 0.3238, "step": 1435 }, { "epoch": 1.9440630275745638, "grad_norm": 0.07042162359075328, "learning_rate": 2.223459102662695e-08, "loss": 0.3235, "step": 1440 }, { "epoch": 1.9508159819921216, "grad_norm": 0.07906944364159181, "learning_rate": 1.7026318683147082e-08, "loss": 0.3115, "step": 1445 }, { "epoch": 1.9575689364096793, "grad_norm": 0.08613249297162148, "learning_rate": 1.2511017027501682e-08, "loss": 0.3032, "step": 1450 }, { "epoch": 1.9643218908272368, "grad_norm": 0.07379471104805907, "learning_rate": 8.689313992306104e-09, "loss": 0.303, "step": 1455 }, { "epoch": 1.9710748452447946, "grad_norm": 0.07085802534414035, "learning_rate": 5.561741053010661e-09, "loss": 0.3059, "step": 1460 }, { "epoch": 1.9778277996623523, "grad_norm": 0.07455102650353135, "learning_rate": 3.1287331539903155e-09, "loss": 0.3165, "step": 1465 }, { "epoch": 1.9845807540799099, "grad_norm": 0.07699115693937987, "learning_rate": 1.3906286480563913e-09, "loss": 0.3108, "step": 1470 }, { "epoch": 1.9913337084974676, "grad_norm": 0.06672257472875329, "learning_rate": 3.4766924940476954e-10, "loss": 0.3134, "step": 1475 }, { "epoch": 1.9980866629150253, "grad_norm": 0.06196469052553269, "learning_rate": 0.0, "loss": 0.3054, "step": 1480 }, { "epoch": 1.9980866629150253, "eval_loss": 0.3086739182472229, "eval_runtime": 17.7246, "eval_samples_per_second": 20.254, "eval_steps_per_second": 5.078, "step": 1480 }, { "epoch": 1.9980866629150253, "step": 1480, "total_flos": 3.6719988787579453e+18, "train_loss": 0.37467364362768224, "train_runtime": 13260.7751, "train_samples_per_second": 5.36, "train_steps_per_second": 0.112 } ], "logging_steps": 5, "max_steps": 1480, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.6719988787579453e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }