{ "best_global_step": 38000, "best_metric": 0.98431396484375, "best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-38000", "epoch": 2.999769248519345, "eval_steps": 250, "global_step": 39000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003845858010922237, "grad_norm": 10.111166000366211, "learning_rate": 1.76e-05, "loss": 13.1594, "step": 50 }, { "epoch": 0.007691716021844474, "grad_norm": 5.981163501739502, "learning_rate": 3.76e-05, "loss": 7.7407, "step": 100 }, { "epoch": 0.01153757403276671, "grad_norm": 7.169628143310547, "learning_rate": 5.76e-05, "loss": 6.5841, "step": 150 }, { "epoch": 0.015383432043688947, "grad_norm": 9.481369972229004, "learning_rate": 7.76e-05, "loss": 5.8929, "step": 200 }, { "epoch": 0.019229290054611183, "grad_norm": 5.348191261291504, "learning_rate": 9.76e-05, "loss": 5.8651, "step": 250 }, { "epoch": 0.019229290054611183, "eval_loss": 5.691133975982666, "eval_runtime": 19.1859, "eval_samples_per_second": 52.122, "eval_steps_per_second": 13.03, "step": 250 }, { "epoch": 0.02307514806553342, "grad_norm": 9.455660820007324, "learning_rate": 0.0001176, "loss": 5.3783, "step": 300 }, { "epoch": 0.02692100607645566, "grad_norm": 11.51925277709961, "learning_rate": 0.00013759999999999998, "loss": 5.0764, "step": 350 }, { "epoch": 0.030766864087377895, "grad_norm": 5.797102928161621, "learning_rate": 0.0001576, "loss": 5.2967, "step": 400 }, { "epoch": 0.03461272209830013, "grad_norm": 3.018831253051758, "learning_rate": 0.0001776, "loss": 5.1947, "step": 450 }, { "epoch": 0.038458580109222366, "grad_norm": 5.181191921234131, "learning_rate": 0.0001976, "loss": 5.2159, "step": 500 }, { "epoch": 0.038458580109222366, "eval_loss": 5.163902759552002, "eval_runtime": 19.7141, "eval_samples_per_second": 50.725, "eval_steps_per_second": 12.681, "step": 500 }, { "epoch": 0.0423044381201446, "grad_norm": 6.368561744689941, "learning_rate": 0.0001997714463808015, "loss": 5.0723, "step": 550 }, { "epoch": 0.04615029613106684, "grad_norm": 4.666464328765869, "learning_rate": 0.00019951172635898504, "loss": 5.0754, "step": 600 }, { "epoch": 0.04999615414198908, "grad_norm": 9.395013809204102, "learning_rate": 0.00019925200633716855, "loss": 4.7718, "step": 650 }, { "epoch": 0.05384201215291132, "grad_norm": 4.279386520385742, "learning_rate": 0.00019899228631535204, "loss": 4.781, "step": 700 }, { "epoch": 0.05768787016383355, "grad_norm": 4.274901866912842, "learning_rate": 0.00019873256629353558, "loss": 4.9529, "step": 750 }, { "epoch": 0.05768787016383355, "eval_loss": 4.844481945037842, "eval_runtime": 18.9171, "eval_samples_per_second": 52.862, "eval_steps_per_second": 13.216, "step": 750 }, { "epoch": 0.06153372817475579, "grad_norm": 3.241445541381836, "learning_rate": 0.0001984728462717191, "loss": 5.0586, "step": 800 }, { "epoch": 0.06537958618567802, "grad_norm": 5.880845546722412, "learning_rate": 0.00019821312624990263, "loss": 4.952, "step": 850 }, { "epoch": 0.06922544419660026, "grad_norm": 3.81998610496521, "learning_rate": 0.00019795340622808614, "loss": 5.0012, "step": 900 }, { "epoch": 0.0730713022075225, "grad_norm": 4.52741003036499, "learning_rate": 0.00019769368620626966, "loss": 4.7775, "step": 950 }, { "epoch": 0.07691716021844473, "grad_norm": 27.68866729736328, "learning_rate": 0.00019743916058488948, "loss": 4.9852, "step": 1000 }, { "epoch": 0.07691716021844473, "eval_loss": 4.836514472961426, "eval_runtime": 19.7103, "eval_samples_per_second": 50.735, "eval_steps_per_second": 12.684, "step": 1000 }, { "epoch": 0.08076301822936698, "grad_norm": 5.79191255569458, "learning_rate": 0.00019717944056307302, "loss": 4.6731, "step": 1050 }, { "epoch": 0.0846088762402892, "grad_norm": 4.957877159118652, "learning_rate": 0.00019691972054125653, "loss": 4.7966, "step": 1100 }, { "epoch": 0.08845473425121145, "grad_norm": 3.2968597412109375, "learning_rate": 0.00019666000051944005, "loss": 4.752, "step": 1150 }, { "epoch": 0.09230059226213368, "grad_norm": 6.059363842010498, "learning_rate": 0.00019640028049762359, "loss": 4.7368, "step": 1200 }, { "epoch": 0.09614645027305592, "grad_norm": 5.9793171882629395, "learning_rate": 0.0001961405604758071, "loss": 4.9613, "step": 1250 }, { "epoch": 0.09614645027305592, "eval_loss": 4.76948356628418, "eval_runtime": 18.6703, "eval_samples_per_second": 53.561, "eval_steps_per_second": 13.39, "step": 1250 }, { "epoch": 0.09999230828397816, "grad_norm": 7.604544162750244, "learning_rate": 0.0001958808404539906, "loss": 4.8365, "step": 1300 }, { "epoch": 0.10383816629490039, "grad_norm": 6.4756083488464355, "learning_rate": 0.00019562112043217412, "loss": 4.8381, "step": 1350 }, { "epoch": 0.10768402430582263, "grad_norm": 3.605341672897339, "learning_rate": 0.00019536140041035764, "loss": 4.7004, "step": 1400 }, { "epoch": 0.11152988231674486, "grad_norm": 3.4805853366851807, "learning_rate": 0.00019510168038854115, "loss": 4.7293, "step": 1450 }, { "epoch": 0.1153757403276671, "grad_norm": 4.5123796463012695, "learning_rate": 0.0001948419603667247, "loss": 4.8315, "step": 1500 }, { "epoch": 0.1153757403276671, "eval_loss": 4.784451961517334, "eval_runtime": 18.5259, "eval_samples_per_second": 53.979, "eval_steps_per_second": 13.495, "step": 1500 }, { "epoch": 0.11922159833858934, "grad_norm": 3.392902135848999, "learning_rate": 0.0001945822403449082, "loss": 4.859, "step": 1550 }, { "epoch": 0.12306745634951158, "grad_norm": 4.783381938934326, "learning_rate": 0.00019432252032309172, "loss": 4.7791, "step": 1600 }, { "epoch": 0.12691331436043382, "grad_norm": 3.96708345413208, "learning_rate": 0.00019406280030127523, "loss": 4.7743, "step": 1650 }, { "epoch": 0.13075917237135604, "grad_norm": 5.381892681121826, "learning_rate": 0.00019380308027945874, "loss": 4.8915, "step": 1700 }, { "epoch": 0.13460503038227828, "grad_norm": 3.269899606704712, "learning_rate": 0.00019354336025764228, "loss": 4.8823, "step": 1750 }, { "epoch": 0.13460503038227828, "eval_loss": 4.689135551452637, "eval_runtime": 18.4665, "eval_samples_per_second": 54.152, "eval_steps_per_second": 13.538, "step": 1750 }, { "epoch": 0.13845088839320052, "grad_norm": 2.5163233280181885, "learning_rate": 0.0001932836402358258, "loss": 4.8513, "step": 1800 }, { "epoch": 0.14229674640412276, "grad_norm": 4.244402885437012, "learning_rate": 0.0001930239202140093, "loss": 4.6855, "step": 1850 }, { "epoch": 0.146142604415045, "grad_norm": 6.551025867462158, "learning_rate": 0.00019276420019219285, "loss": 4.8173, "step": 1900 }, { "epoch": 0.14998846242596722, "grad_norm": 2.791435956954956, "learning_rate": 0.00019250448017037633, "loss": 4.6983, "step": 1950 }, { "epoch": 0.15383432043688947, "grad_norm": 3.4012036323547363, "learning_rate": 0.00019224476014855984, "loss": 4.8178, "step": 2000 }, { "epoch": 0.15383432043688947, "eval_loss": 4.666446208953857, "eval_runtime": 18.3775, "eval_samples_per_second": 54.414, "eval_steps_per_second": 13.604, "step": 2000 }, { "epoch": 0.1576801784478117, "grad_norm": 6.46567964553833, "learning_rate": 0.00019198504012674338, "loss": 4.8124, "step": 2050 }, { "epoch": 0.16152603645873395, "grad_norm": 4.300732135772705, "learning_rate": 0.0001917253201049269, "loss": 4.7302, "step": 2100 }, { "epoch": 0.16537189446965617, "grad_norm": 4.140190601348877, "learning_rate": 0.0001914656000831104, "loss": 4.6498, "step": 2150 }, { "epoch": 0.1692177524805784, "grad_norm": 3.221662998199463, "learning_rate": 0.00019120588006129395, "loss": 4.6428, "step": 2200 }, { "epoch": 0.17306361049150065, "grad_norm": 2.6079111099243164, "learning_rate": 0.00019094616003947746, "loss": 4.6115, "step": 2250 }, { "epoch": 0.17306361049150065, "eval_loss": 4.661706924438477, "eval_runtime": 18.7413, "eval_samples_per_second": 53.358, "eval_steps_per_second": 13.34, "step": 2250 }, { "epoch": 0.1769094685024229, "grad_norm": 4.190133094787598, "learning_rate": 0.00019068644001766095, "loss": 4.7041, "step": 2300 }, { "epoch": 0.18075532651334514, "grad_norm": 5.242035388946533, "learning_rate": 0.0001904267199958445, "loss": 4.6656, "step": 2350 }, { "epoch": 0.18460118452426735, "grad_norm": 6.203541278839111, "learning_rate": 0.000190166999974028, "loss": 4.6835, "step": 2400 }, { "epoch": 0.1884470425351896, "grad_norm": 2.8591034412384033, "learning_rate": 0.00018990727995221151, "loss": 4.6751, "step": 2450 }, { "epoch": 0.19229290054611184, "grad_norm": 5.204433441162109, "learning_rate": 0.00018964755993039505, "loss": 4.7301, "step": 2500 }, { "epoch": 0.19229290054611184, "eval_loss": 4.597048282623291, "eval_runtime": 18.5098, "eval_samples_per_second": 54.025, "eval_steps_per_second": 13.506, "step": 2500 }, { "epoch": 0.19613875855703408, "grad_norm": 2.59525465965271, "learning_rate": 0.00018938783990857857, "loss": 4.4364, "step": 2550 }, { "epoch": 0.19998461656795632, "grad_norm": 3.844686985015869, "learning_rate": 0.00018912811988676208, "loss": 4.6437, "step": 2600 }, { "epoch": 0.20383047457887854, "grad_norm": 3.4633946418762207, "learning_rate": 0.0001888683998649456, "loss": 4.5016, "step": 2650 }, { "epoch": 0.20767633258980078, "grad_norm": 3.7852296829223633, "learning_rate": 0.0001886086798431291, "loss": 4.5845, "step": 2700 }, { "epoch": 0.21152219060072303, "grad_norm": 3.8716065883636475, "learning_rate": 0.00018834895982131265, "loss": 4.3669, "step": 2750 }, { "epoch": 0.21152219060072303, "eval_loss": 4.602295875549316, "eval_runtime": 18.4747, "eval_samples_per_second": 54.128, "eval_steps_per_second": 13.532, "step": 2750 }, { "epoch": 0.21536804861164527, "grad_norm": 3.9932167530059814, "learning_rate": 0.00018808923979949616, "loss": 4.4545, "step": 2800 }, { "epoch": 0.21921390662256748, "grad_norm": 4.182176113128662, "learning_rate": 0.00018782951977767967, "loss": 4.5355, "step": 2850 }, { "epoch": 0.22305976463348973, "grad_norm": 3.776895523071289, "learning_rate": 0.00018756979975586318, "loss": 4.6859, "step": 2900 }, { "epoch": 0.22690562264441197, "grad_norm": 3.9219324588775635, "learning_rate": 0.0001873100797340467, "loss": 4.5186, "step": 2950 }, { "epoch": 0.2307514806553342, "grad_norm": 3.7879323959350586, "learning_rate": 0.0001870503597122302, "loss": 4.612, "step": 3000 }, { "epoch": 0.2307514806553342, "eval_loss": 4.579595565795898, "eval_runtime": 18.476, "eval_samples_per_second": 54.124, "eval_steps_per_second": 13.531, "step": 3000 }, { "epoch": 0.23459733866625646, "grad_norm": 6.408226490020752, "learning_rate": 0.00018679063969041375, "loss": 4.6614, "step": 3050 }, { "epoch": 0.23844319667717867, "grad_norm": 6.529899597167969, "learning_rate": 0.00018653091966859726, "loss": 4.5576, "step": 3100 }, { "epoch": 0.2422890546881009, "grad_norm": 4.417705059051514, "learning_rate": 0.00018627119964678078, "loss": 4.5466, "step": 3150 }, { "epoch": 0.24613491269902316, "grad_norm": 3.052746295928955, "learning_rate": 0.0001860114796249643, "loss": 4.6959, "step": 3200 }, { "epoch": 0.2499807707099454, "grad_norm": 5.121955871582031, "learning_rate": 0.0001857517596031478, "loss": 4.6796, "step": 3250 }, { "epoch": 0.2499807707099454, "eval_loss": 4.576270580291748, "eval_runtime": 18.8896, "eval_samples_per_second": 52.939, "eval_steps_per_second": 13.235, "step": 3250 }, { "epoch": 0.25382662872086764, "grad_norm": 2.772292375564575, "learning_rate": 0.00018549203958133131, "loss": 4.534, "step": 3300 }, { "epoch": 0.2576724867317899, "grad_norm": 6.105401039123535, "learning_rate": 0.00018523231955951485, "loss": 4.4958, "step": 3350 }, { "epoch": 0.26151834474271207, "grad_norm": 2.742037057876587, "learning_rate": 0.00018497259953769837, "loss": 4.5835, "step": 3400 }, { "epoch": 0.2653642027536343, "grad_norm": 3.6125235557556152, "learning_rate": 0.0001847128795158819, "loss": 4.5917, "step": 3450 }, { "epoch": 0.26921006076455656, "grad_norm": 3.8596699237823486, "learning_rate": 0.00018445315949406542, "loss": 4.4909, "step": 3500 }, { "epoch": 0.26921006076455656, "eval_loss": 4.520543098449707, "eval_runtime": 18.7111, "eval_samples_per_second": 53.444, "eval_steps_per_second": 13.361, "step": 3500 }, { "epoch": 0.2730559187754788, "grad_norm": 3.639690399169922, "learning_rate": 0.0001841934394722489, "loss": 4.5127, "step": 3550 }, { "epoch": 0.27690177678640104, "grad_norm": 3.864473342895508, "learning_rate": 0.00018393371945043245, "loss": 4.599, "step": 3600 }, { "epoch": 0.2807476347973233, "grad_norm": 4.662705421447754, "learning_rate": 0.00018367399942861596, "loss": 4.5951, "step": 3650 }, { "epoch": 0.28459349280824553, "grad_norm": 3.066333532333374, "learning_rate": 0.00018341427940679947, "loss": 4.5583, "step": 3700 }, { "epoch": 0.2884393508191678, "grad_norm": 4.054274082183838, "learning_rate": 0.000183154559384983, "loss": 4.5732, "step": 3750 }, { "epoch": 0.2884393508191678, "eval_loss": 4.569947242736816, "eval_runtime": 18.5242, "eval_samples_per_second": 53.983, "eval_steps_per_second": 13.496, "step": 3750 }, { "epoch": 0.29228520883009, "grad_norm": 6.440130710601807, "learning_rate": 0.00018289483936316652, "loss": 4.6025, "step": 3800 }, { "epoch": 0.2961310668410122, "grad_norm": 5.849060535430908, "learning_rate": 0.00018263511934135004, "loss": 4.4932, "step": 3850 }, { "epoch": 0.29997692485193445, "grad_norm": 7.0537800788879395, "learning_rate": 0.00018237539931953355, "loss": 4.6986, "step": 3900 }, { "epoch": 0.3038227828628567, "grad_norm": 5.134806156158447, "learning_rate": 0.00018211567929771706, "loss": 4.5074, "step": 3950 }, { "epoch": 0.30766864087377893, "grad_norm": 2.872307777404785, "learning_rate": 0.00018185595927590058, "loss": 4.547, "step": 4000 }, { "epoch": 0.30766864087377893, "eval_loss": 4.5459160804748535, "eval_runtime": 18.4232, "eval_samples_per_second": 54.279, "eval_steps_per_second": 13.57, "step": 4000 }, { "epoch": 0.3115144988847012, "grad_norm": 5.374965190887451, "learning_rate": 0.00018159623925408412, "loss": 4.5327, "step": 4050 }, { "epoch": 0.3153603568956234, "grad_norm": 4.99652624130249, "learning_rate": 0.00018133651923226763, "loss": 4.3966, "step": 4100 }, { "epoch": 0.31920621490654566, "grad_norm": 6.257124423980713, "learning_rate": 0.00018107679921045114, "loss": 4.5792, "step": 4150 }, { "epoch": 0.3230520729174679, "grad_norm": 8.056533813476562, "learning_rate": 0.00018081707918863465, "loss": 4.6401, "step": 4200 }, { "epoch": 0.32689793092839015, "grad_norm": 4.024567127227783, "learning_rate": 0.00018055735916681817, "loss": 4.5516, "step": 4250 }, { "epoch": 0.32689793092839015, "eval_loss": 4.558110237121582, "eval_runtime": 18.5343, "eval_samples_per_second": 53.954, "eval_steps_per_second": 13.488, "step": 4250 }, { "epoch": 0.33074378893931233, "grad_norm": 2.4326066970825195, "learning_rate": 0.0001802976391450017, "loss": 4.5761, "step": 4300 }, { "epoch": 0.3345896469502346, "grad_norm": 2.7615299224853516, "learning_rate": 0.00018003791912318522, "loss": 4.4291, "step": 4350 }, { "epoch": 0.3384355049611568, "grad_norm": 3.9387362003326416, "learning_rate": 0.00017977819910136873, "loss": 4.5422, "step": 4400 }, { "epoch": 0.34228136297207906, "grad_norm": 23.72602653503418, "learning_rate": 0.00017951847907955227, "loss": 4.4379, "step": 4450 }, { "epoch": 0.3461272209830013, "grad_norm": 2.968930959701538, "learning_rate": 0.0001792639534581721, "loss": 4.433, "step": 4500 }, { "epoch": 0.3461272209830013, "eval_loss": 4.5359063148498535, "eval_runtime": 18.5422, "eval_samples_per_second": 53.931, "eval_steps_per_second": 13.483, "step": 4500 }, { "epoch": 0.34997307899392355, "grad_norm": 6.404330730438232, "learning_rate": 0.0001790042334363556, "loss": 4.5673, "step": 4550 }, { "epoch": 0.3538189370048458, "grad_norm": 10.212136268615723, "learning_rate": 0.00017874451341453912, "loss": 4.6249, "step": 4600 }, { "epoch": 0.35766479501576803, "grad_norm": 4.401816368103027, "learning_rate": 0.00017848479339272266, "loss": 4.6305, "step": 4650 }, { "epoch": 0.3615106530266903, "grad_norm": 4.710996150970459, "learning_rate": 0.00017822507337090617, "loss": 4.3731, "step": 4700 }, { "epoch": 0.36535651103761246, "grad_norm": 3.150613307952881, "learning_rate": 0.0001779653533490897, "loss": 4.4491, "step": 4750 }, { "epoch": 0.36535651103761246, "eval_loss": 4.564510345458984, "eval_runtime": 18.5575, "eval_samples_per_second": 53.886, "eval_steps_per_second": 13.472, "step": 4750 }, { "epoch": 0.3692023690485347, "grad_norm": 4.828207492828369, "learning_rate": 0.0001777056333272732, "loss": 4.5923, "step": 4800 }, { "epoch": 0.37304822705945695, "grad_norm": 3.780848264694214, "learning_rate": 0.0001774459133054567, "loss": 4.5544, "step": 4850 }, { "epoch": 0.3768940850703792, "grad_norm": 4.04913854598999, "learning_rate": 0.00017718619328364023, "loss": 4.5271, "step": 4900 }, { "epoch": 0.38073994308130144, "grad_norm": 4.097137451171875, "learning_rate": 0.00017692647326182377, "loss": 4.4929, "step": 4950 }, { "epoch": 0.3845858010922237, "grad_norm": 4.65788459777832, "learning_rate": 0.00017666675324000728, "loss": 4.5888, "step": 5000 }, { "epoch": 0.3845858010922237, "eval_loss": 4.465761661529541, "eval_runtime": 18.6518, "eval_samples_per_second": 53.614, "eval_steps_per_second": 13.404, "step": 5000 }, { "epoch": 0.3884316591031459, "grad_norm": 3.0376453399658203, "learning_rate": 0.00017640703321819082, "loss": 4.574, "step": 5050 }, { "epoch": 0.39227751711406816, "grad_norm": 2.6457693576812744, "learning_rate": 0.00017614731319637433, "loss": 4.5782, "step": 5100 }, { "epoch": 0.3961233751249904, "grad_norm": 4.438416957855225, "learning_rate": 0.00017588759317455782, "loss": 4.5586, "step": 5150 }, { "epoch": 0.39996923313591265, "grad_norm": 5.325882911682129, "learning_rate": 0.00017562787315274136, "loss": 4.5136, "step": 5200 }, { "epoch": 0.40381509114683484, "grad_norm": 1.816029667854309, "learning_rate": 0.00017536815313092487, "loss": 4.4819, "step": 5250 }, { "epoch": 0.40381509114683484, "eval_loss": 4.561609268188477, "eval_runtime": 18.5623, "eval_samples_per_second": 53.873, "eval_steps_per_second": 13.468, "step": 5250 }, { "epoch": 0.4076609491577571, "grad_norm": 4.011863708496094, "learning_rate": 0.00017510843310910838, "loss": 4.4284, "step": 5300 }, { "epoch": 0.4115068071686793, "grad_norm": 10.169037818908691, "learning_rate": 0.00017484871308729192, "loss": 4.5414, "step": 5350 }, { "epoch": 0.41535266517960157, "grad_norm": 3.9185502529144287, "learning_rate": 0.00017458899306547544, "loss": 4.6659, "step": 5400 }, { "epoch": 0.4191985231905238, "grad_norm": 2.873530864715576, "learning_rate": 0.00017432927304365895, "loss": 4.4482, "step": 5450 }, { "epoch": 0.42304438120144605, "grad_norm": 4.379590034484863, "learning_rate": 0.00017406955302184246, "loss": 4.5501, "step": 5500 }, { "epoch": 0.42304438120144605, "eval_loss": 4.513929843902588, "eval_runtime": 18.6196, "eval_samples_per_second": 53.707, "eval_steps_per_second": 13.427, "step": 5500 }, { "epoch": 0.4268902392123683, "grad_norm": 4.2113752365112305, "learning_rate": 0.00017380983300002597, "loss": 4.5454, "step": 5550 }, { "epoch": 0.43073609722329054, "grad_norm": 4.782048225402832, "learning_rate": 0.0001735501129782095, "loss": 4.4705, "step": 5600 }, { "epoch": 0.4345819552342128, "grad_norm": 2.6753036975860596, "learning_rate": 0.00017329039295639303, "loss": 4.4592, "step": 5650 }, { "epoch": 0.43842781324513497, "grad_norm": 5.012415885925293, "learning_rate": 0.00017303067293457654, "loss": 4.623, "step": 5700 }, { "epoch": 0.4422736712560572, "grad_norm": 3.0666699409484863, "learning_rate": 0.00017277095291276005, "loss": 4.4325, "step": 5750 }, { "epoch": 0.4422736712560572, "eval_loss": 4.489352703094482, "eval_runtime": 18.4948, "eval_samples_per_second": 54.069, "eval_steps_per_second": 13.517, "step": 5750 }, { "epoch": 0.44611952926697945, "grad_norm": 6.8570876121521, "learning_rate": 0.00017251123289094357, "loss": 4.5179, "step": 5800 }, { "epoch": 0.4499653872779017, "grad_norm": 7.190755844116211, "learning_rate": 0.00017225151286912708, "loss": 4.5877, "step": 5850 }, { "epoch": 0.45381124528882394, "grad_norm": 4.404886722564697, "learning_rate": 0.00017199179284731062, "loss": 4.4072, "step": 5900 }, { "epoch": 0.4576571032997462, "grad_norm": 3.0543084144592285, "learning_rate": 0.00017173207282549413, "loss": 4.4656, "step": 5950 }, { "epoch": 0.4615029613106684, "grad_norm": 6.7454514503479, "learning_rate": 0.00017147235280367764, "loss": 4.5688, "step": 6000 }, { "epoch": 0.4615029613106684, "eval_loss": 4.480144023895264, "eval_runtime": 18.5584, "eval_samples_per_second": 53.884, "eval_steps_per_second": 13.471, "step": 6000 }, { "epoch": 0.46534881932159067, "grad_norm": 5.5196661949157715, "learning_rate": 0.00017121263278186118, "loss": 4.5703, "step": 6050 }, { "epoch": 0.4691946773325129, "grad_norm": 4.253966331481934, "learning_rate": 0.00017095291276004467, "loss": 4.6012, "step": 6100 }, { "epoch": 0.4730405353434351, "grad_norm": 2.459376096725464, "learning_rate": 0.00017069319273822818, "loss": 4.5002, "step": 6150 }, { "epoch": 0.47688639335435734, "grad_norm": 4.933450698852539, "learning_rate": 0.00017043347271641172, "loss": 4.5703, "step": 6200 }, { "epoch": 0.4807322513652796, "grad_norm": 4.511186599731445, "learning_rate": 0.00017017375269459524, "loss": 4.4665, "step": 6250 }, { "epoch": 0.4807322513652796, "eval_loss": 4.485811233520508, "eval_runtime": 18.522, "eval_samples_per_second": 53.99, "eval_steps_per_second": 13.497, "step": 6250 }, { "epoch": 0.4845781093762018, "grad_norm": 5.634074687957764, "learning_rate": 0.00016991403267277875, "loss": 4.4616, "step": 6300 }, { "epoch": 0.48842396738712407, "grad_norm": 3.319650650024414, "learning_rate": 0.0001696543126509623, "loss": 4.4836, "step": 6350 }, { "epoch": 0.4922698253980463, "grad_norm": 3.306976079940796, "learning_rate": 0.00016939459262914577, "loss": 4.5256, "step": 6400 }, { "epoch": 0.49611568340896856, "grad_norm": 4.1797308921813965, "learning_rate": 0.0001691348726073293, "loss": 4.3822, "step": 6450 }, { "epoch": 0.4999615414198908, "grad_norm": 3.2349929809570312, "learning_rate": 0.00016887515258551283, "loss": 4.4384, "step": 6500 }, { "epoch": 0.4999615414198908, "eval_loss": 4.485826015472412, "eval_runtime": 18.614, "eval_samples_per_second": 53.723, "eval_steps_per_second": 13.431, "step": 6500 }, { "epoch": 0.503807399430813, "grad_norm": 3.325056791305542, "learning_rate": 0.00016861543256369634, "loss": 4.618, "step": 6550 }, { "epoch": 0.5076532574417353, "grad_norm": 4.026259899139404, "learning_rate": 0.0001683609069423162, "loss": 4.5485, "step": 6600 }, { "epoch": 0.5114991154526575, "grad_norm": 3.1270413398742676, "learning_rate": 0.00016810118692049973, "loss": 4.4426, "step": 6650 }, { "epoch": 0.5153449734635798, "grad_norm": 5.264435768127441, "learning_rate": 0.00016784146689868324, "loss": 4.3544, "step": 6700 }, { "epoch": 0.519190831474502, "grad_norm": 4.349465847015381, "learning_rate": 0.00016758694127730307, "loss": 4.5912, "step": 6750 }, { "epoch": 0.519190831474502, "eval_loss": 4.503914833068848, "eval_runtime": 18.5932, "eval_samples_per_second": 53.783, "eval_steps_per_second": 13.446, "step": 6750 }, { "epoch": 0.5230366894854241, "grad_norm": 2.5264992713928223, "learning_rate": 0.00016732722125548658, "loss": 4.3703, "step": 6800 }, { "epoch": 0.5268825474963464, "grad_norm": 4.934820175170898, "learning_rate": 0.00016706750123367012, "loss": 4.5652, "step": 6850 }, { "epoch": 0.5307284055072686, "grad_norm": 4.21425724029541, "learning_rate": 0.00016680778121185363, "loss": 4.4684, "step": 6900 }, { "epoch": 0.5345742635181909, "grad_norm": 5.111146926879883, "learning_rate": 0.00016654806119003715, "loss": 4.5119, "step": 6950 }, { "epoch": 0.5384201215291131, "grad_norm": 4.563775062561035, "learning_rate": 0.00016628834116822069, "loss": 4.489, "step": 7000 }, { "epoch": 0.5384201215291131, "eval_loss": 4.429446220397949, "eval_runtime": 18.6345, "eval_samples_per_second": 53.664, "eval_steps_per_second": 13.416, "step": 7000 }, { "epoch": 0.5422659795400354, "grad_norm": 6.055607795715332, "learning_rate": 0.00016602862114640417, "loss": 4.3805, "step": 7050 }, { "epoch": 0.5461118375509576, "grad_norm": 3.190605878829956, "learning_rate": 0.00016576890112458768, "loss": 4.561, "step": 7100 }, { "epoch": 0.5499576955618799, "grad_norm": 2.95857834815979, "learning_rate": 0.00016550918110277122, "loss": 4.5957, "step": 7150 }, { "epoch": 0.5538035535728021, "grad_norm": 5.055838108062744, "learning_rate": 0.00016524946108095474, "loss": 4.3916, "step": 7200 }, { "epoch": 0.5576494115837243, "grad_norm": 6.25083589553833, "learning_rate": 0.00016498974105913825, "loss": 4.6649, "step": 7250 }, { "epoch": 0.5576494115837243, "eval_loss": 4.42822265625, "eval_runtime": 18.5876, "eval_samples_per_second": 53.799, "eval_steps_per_second": 13.45, "step": 7250 }, { "epoch": 0.5614952695946466, "grad_norm": 3.6657636165618896, "learning_rate": 0.0001647300210373218, "loss": 4.5059, "step": 7300 }, { "epoch": 0.5653411276055688, "grad_norm": 3.6100645065307617, "learning_rate": 0.0001644703010155053, "loss": 4.4379, "step": 7350 }, { "epoch": 0.5691869856164911, "grad_norm": 3.535804271697998, "learning_rate": 0.0001642105809936888, "loss": 4.3881, "step": 7400 }, { "epoch": 0.5730328436274132, "grad_norm": 2.9636013507843018, "learning_rate": 0.00016395086097187233, "loss": 4.491, "step": 7450 }, { "epoch": 0.5768787016383355, "grad_norm": 2.9023678302764893, "learning_rate": 0.00016369114095005584, "loss": 4.5567, "step": 7500 }, { "epoch": 0.5768787016383355, "eval_loss": 4.459234237670898, "eval_runtime": 18.5868, "eval_samples_per_second": 53.802, "eval_steps_per_second": 13.45, "step": 7500 }, { "epoch": 0.5807245596492577, "grad_norm": 4.381130218505859, "learning_rate": 0.00016343142092823938, "loss": 4.521, "step": 7550 }, { "epoch": 0.58457041766018, "grad_norm": 2.533957004547119, "learning_rate": 0.0001631717009064229, "loss": 4.3233, "step": 7600 }, { "epoch": 0.5884162756711022, "grad_norm": 3.277646541595459, "learning_rate": 0.0001629119808846064, "loss": 4.5503, "step": 7650 }, { "epoch": 0.5922621336820244, "grad_norm": 3.871952772140503, "learning_rate": 0.00016265226086278992, "loss": 4.4196, "step": 7700 }, { "epoch": 0.5961079916929467, "grad_norm": 3.191589832305908, "learning_rate": 0.00016239254084097343, "loss": 4.479, "step": 7750 }, { "epoch": 0.5961079916929467, "eval_loss": 4.44810152053833, "eval_runtime": 18.425, "eval_samples_per_second": 54.274, "eval_steps_per_second": 13.569, "step": 7750 }, { "epoch": 0.5999538497038689, "grad_norm": 3.7022933959960938, "learning_rate": 0.00016213282081915695, "loss": 4.5402, "step": 7800 }, { "epoch": 0.6037997077147912, "grad_norm": 2.883859395980835, "learning_rate": 0.00016187310079734049, "loss": 4.5457, "step": 7850 }, { "epoch": 0.6076455657257134, "grad_norm": 3.5229415893554688, "learning_rate": 0.000161613380775524, "loss": 4.5185, "step": 7900 }, { "epoch": 0.6114914237366357, "grad_norm": 4.529599666595459, "learning_rate": 0.0001613536607537075, "loss": 4.4489, "step": 7950 }, { "epoch": 0.6153372817475579, "grad_norm": 5.1023850440979, "learning_rate": 0.00016109394073189102, "loss": 4.6043, "step": 8000 }, { "epoch": 0.6153372817475579, "eval_loss": 4.478011608123779, "eval_runtime": 18.576, "eval_samples_per_second": 53.833, "eval_steps_per_second": 13.458, "step": 8000 }, { "epoch": 0.6191831397584802, "grad_norm": 4.5992255210876465, "learning_rate": 0.00016083422071007454, "loss": 4.3146, "step": 8050 }, { "epoch": 0.6230289977694023, "grad_norm": 5.412031650543213, "learning_rate": 0.00016057450068825805, "loss": 4.5841, "step": 8100 }, { "epoch": 0.6268748557803245, "grad_norm": 4.4531779289245605, "learning_rate": 0.0001603147806664416, "loss": 4.5084, "step": 8150 }, { "epoch": 0.6307207137912468, "grad_norm": 3.913174867630005, "learning_rate": 0.0001600550606446251, "loss": 4.4183, "step": 8200 }, { "epoch": 0.634566571802169, "grad_norm": 3.3952407836914062, "learning_rate": 0.00015979534062280862, "loss": 4.3496, "step": 8250 }, { "epoch": 0.634566571802169, "eval_loss": 4.46174430847168, "eval_runtime": 18.5381, "eval_samples_per_second": 53.943, "eval_steps_per_second": 13.486, "step": 8250 }, { "epoch": 0.6384124298130913, "grad_norm": 4.156221389770508, "learning_rate": 0.00015953562060099213, "loss": 4.361, "step": 8300 }, { "epoch": 0.6422582878240135, "grad_norm": 3.2021920680999756, "learning_rate": 0.00015927590057917564, "loss": 4.473, "step": 8350 }, { "epoch": 0.6461041458349358, "grad_norm": 5.048036575317383, "learning_rate": 0.00015901618055735918, "loss": 4.4142, "step": 8400 }, { "epoch": 0.649950003845858, "grad_norm": 3.91768217086792, "learning_rate": 0.0001587564605355427, "loss": 4.4672, "step": 8450 }, { "epoch": 0.6537958618567803, "grad_norm": 9.229452133178711, "learning_rate": 0.0001584967405137262, "loss": 4.7837, "step": 8500 }, { "epoch": 0.6537958618567803, "eval_loss": 4.409055709838867, "eval_runtime": 18.5943, "eval_samples_per_second": 53.78, "eval_steps_per_second": 13.445, "step": 8500 }, { "epoch": 0.6576417198677025, "grad_norm": 2.7313661575317383, "learning_rate": 0.00015823702049190975, "loss": 4.4721, "step": 8550 }, { "epoch": 0.6614875778786247, "grad_norm": 4.160475730895996, "learning_rate": 0.00015797730047009326, "loss": 4.501, "step": 8600 }, { "epoch": 0.665333435889547, "grad_norm": 11.54045581817627, "learning_rate": 0.00015771758044827675, "loss": 4.4433, "step": 8650 }, { "epoch": 0.6691792939004692, "grad_norm": 4.087617874145508, "learning_rate": 0.00015745786042646029, "loss": 4.4981, "step": 8700 }, { "epoch": 0.6730251519113915, "grad_norm": 4.155121803283691, "learning_rate": 0.0001571981404046438, "loss": 4.3874, "step": 8750 }, { "epoch": 0.6730251519113915, "eval_loss": 4.418811321258545, "eval_runtime": 18.6306, "eval_samples_per_second": 53.675, "eval_steps_per_second": 13.419, "step": 8750 }, { "epoch": 0.6768710099223136, "grad_norm": 4.071916580200195, "learning_rate": 0.0001569384203828273, "loss": 4.5531, "step": 8800 }, { "epoch": 0.6807168679332359, "grad_norm": 3.395460605621338, "learning_rate": 0.00015667870036101085, "loss": 4.4609, "step": 8850 }, { "epoch": 0.6845627259441581, "grad_norm": 3.4933230876922607, "learning_rate": 0.00015641898033919436, "loss": 4.4536, "step": 8900 }, { "epoch": 0.6884085839550804, "grad_norm": 6.921072483062744, "learning_rate": 0.00015615926031737788, "loss": 4.3478, "step": 8950 }, { "epoch": 0.6922544419660026, "grad_norm": 3.920626401901245, "learning_rate": 0.0001558995402955614, "loss": 4.3761, "step": 9000 }, { "epoch": 0.6922544419660026, "eval_loss": 4.415992259979248, "eval_runtime": 18.5147, "eval_samples_per_second": 54.011, "eval_steps_per_second": 13.503, "step": 9000 }, { "epoch": 0.6961002999769248, "grad_norm": 7.213745594024658, "learning_rate": 0.0001556398202737449, "loss": 4.323, "step": 9050 }, { "epoch": 0.6999461579878471, "grad_norm": 3.2426984310150146, "learning_rate": 0.00015538010025192842, "loss": 4.3922, "step": 9100 }, { "epoch": 0.7037920159987693, "grad_norm": 3.256950855255127, "learning_rate": 0.00015512038023011196, "loss": 4.2602, "step": 9150 }, { "epoch": 0.7076378740096916, "grad_norm": 6.132264614105225, "learning_rate": 0.00015486066020829547, "loss": 4.3734, "step": 9200 }, { "epoch": 0.7114837320206138, "grad_norm": 3.921595573425293, "learning_rate": 0.00015460094018647898, "loss": 4.4776, "step": 9250 }, { "epoch": 0.7114837320206138, "eval_loss": 4.3921356201171875, "eval_runtime": 18.5511, "eval_samples_per_second": 53.905, "eval_steps_per_second": 13.476, "step": 9250 }, { "epoch": 0.7153295900315361, "grad_norm": 5.416064739227295, "learning_rate": 0.0001543412201646625, "loss": 4.5659, "step": 9300 }, { "epoch": 0.7191754480424583, "grad_norm": 4.542217254638672, "learning_rate": 0.000154081500142846, "loss": 4.2557, "step": 9350 }, { "epoch": 0.7230213060533806, "grad_norm": 3.7075681686401367, "learning_rate": 0.00015382178012102955, "loss": 4.453, "step": 9400 }, { "epoch": 0.7268671640643027, "grad_norm": 4.457496166229248, "learning_rate": 0.00015356206009921306, "loss": 4.3861, "step": 9450 }, { "epoch": 0.7307130220752249, "grad_norm": 1.7784981727600098, "learning_rate": 0.00015330234007739657, "loss": 4.3258, "step": 9500 }, { "epoch": 0.7307130220752249, "eval_loss": 4.357193470001221, "eval_runtime": 18.6372, "eval_samples_per_second": 53.656, "eval_steps_per_second": 13.414, "step": 9500 }, { "epoch": 0.7345588800861472, "grad_norm": 3.8532371520996094, "learning_rate": 0.0001530426200555801, "loss": 4.3473, "step": 9550 }, { "epoch": 0.7384047380970694, "grad_norm": 4.654659271240234, "learning_rate": 0.0001527829000337636, "loss": 4.457, "step": 9600 }, { "epoch": 0.7422505961079917, "grad_norm": 2.420182228088379, "learning_rate": 0.0001525231800119471, "loss": 4.4521, "step": 9650 }, { "epoch": 0.7460964541189139, "grad_norm": 4.189414978027344, "learning_rate": 0.00015226345999013065, "loss": 4.2569, "step": 9700 }, { "epoch": 0.7499423121298362, "grad_norm": 2.824084997177124, "learning_rate": 0.00015200373996831416, "loss": 4.3409, "step": 9750 }, { "epoch": 0.7499423121298362, "eval_loss": 4.378731727600098, "eval_runtime": 18.4857, "eval_samples_per_second": 54.096, "eval_steps_per_second": 13.524, "step": 9750 }, { "epoch": 0.7537881701407584, "grad_norm": 6.379781723022461, "learning_rate": 0.00015174401994649768, "loss": 4.3041, "step": 9800 }, { "epoch": 0.7576340281516807, "grad_norm": 1.7334113121032715, "learning_rate": 0.00015148429992468122, "loss": 4.3096, "step": 9850 }, { "epoch": 0.7614798861626029, "grad_norm": 4.287415027618408, "learning_rate": 0.00015122457990286473, "loss": 4.4411, "step": 9900 }, { "epoch": 0.7653257441735252, "grad_norm": 3.3184821605682373, "learning_rate": 0.00015096485988104821, "loss": 4.3992, "step": 9950 }, { "epoch": 0.7691716021844474, "grad_norm": 4.698968887329102, "learning_rate": 0.00015070513985923175, "loss": 4.4726, "step": 10000 }, { "epoch": 0.7691716021844474, "eval_loss": 4.408615589141846, "eval_runtime": 18.5408, "eval_samples_per_second": 53.935, "eval_steps_per_second": 13.484, "step": 10000 }, { "epoch": 0.7730174601953695, "grad_norm": 3.882775068283081, "learning_rate": 0.00015044541983741527, "loss": 4.5207, "step": 10050 }, { "epoch": 0.7768633182062918, "grad_norm": 5.814795017242432, "learning_rate": 0.0001501856998155988, "loss": 4.2462, "step": 10100 }, { "epoch": 0.780709176217214, "grad_norm": 4.733581066131592, "learning_rate": 0.00014992597979378232, "loss": 4.5563, "step": 10150 }, { "epoch": 0.7845550342281363, "grad_norm": 4.805403232574463, "learning_rate": 0.00014966625977196583, "loss": 4.4353, "step": 10200 }, { "epoch": 0.7884008922390585, "grad_norm": 5.814332008361816, "learning_rate": 0.00014940653975014935, "loss": 4.4004, "step": 10250 }, { "epoch": 0.7884008922390585, "eval_loss": 4.41144323348999, "eval_runtime": 18.5402, "eval_samples_per_second": 53.937, "eval_steps_per_second": 13.484, "step": 10250 }, { "epoch": 0.7922467502499808, "grad_norm": 5.321393013000488, "learning_rate": 0.00014914681972833286, "loss": 4.4383, "step": 10300 }, { "epoch": 0.796092608260903, "grad_norm": 3.681452751159668, "learning_rate": 0.00014888709970651637, "loss": 4.5094, "step": 10350 }, { "epoch": 0.7999384662718253, "grad_norm": 4.766401767730713, "learning_rate": 0.0001486273796846999, "loss": 4.3104, "step": 10400 }, { "epoch": 0.8037843242827475, "grad_norm": 4.676774024963379, "learning_rate": 0.00014836765966288342, "loss": 4.516, "step": 10450 }, { "epoch": 0.8076301822936697, "grad_norm": 3.623643159866333, "learning_rate": 0.00014810793964106694, "loss": 4.5146, "step": 10500 }, { "epoch": 0.8076301822936697, "eval_loss": 4.398375034332275, "eval_runtime": 18.4969, "eval_samples_per_second": 54.063, "eval_steps_per_second": 13.516, "step": 10500 }, { "epoch": 0.811476040304592, "grad_norm": 4.379317760467529, "learning_rate": 0.00014784821961925045, "loss": 4.6715, "step": 10550 }, { "epoch": 0.8153218983155142, "grad_norm": 3.034796714782715, "learning_rate": 0.00014758849959743396, "loss": 4.4511, "step": 10600 }, { "epoch": 0.8191677563264365, "grad_norm": 3.8016927242279053, "learning_rate": 0.00014732877957561748, "loss": 4.2966, "step": 10650 }, { "epoch": 0.8230136143373586, "grad_norm": 4.330080509185791, "learning_rate": 0.00014706905955380102, "loss": 4.6496, "step": 10700 }, { "epoch": 0.8268594723482809, "grad_norm": 8.032389640808105, "learning_rate": 0.00014680933953198453, "loss": 4.2758, "step": 10750 }, { "epoch": 0.8268594723482809, "eval_loss": 4.388455867767334, "eval_runtime": 18.481, "eval_samples_per_second": 54.11, "eval_steps_per_second": 13.527, "step": 10750 }, { "epoch": 0.8307053303592031, "grad_norm": 3.1724319458007812, "learning_rate": 0.00014654961951016807, "loss": 4.405, "step": 10800 }, { "epoch": 0.8345511883701254, "grad_norm": 3.529196262359619, "learning_rate": 0.00014628989948835155, "loss": 4.4433, "step": 10850 }, { "epoch": 0.8383970463810476, "grad_norm": 4.758362293243408, "learning_rate": 0.00014603017946653507, "loss": 4.5747, "step": 10900 }, { "epoch": 0.8422429043919698, "grad_norm": 3.524068832397461, "learning_rate": 0.0001457704594447186, "loss": 4.3469, "step": 10950 }, { "epoch": 0.8460887624028921, "grad_norm": 4.452401161193848, "learning_rate": 0.00014551073942290212, "loss": 4.3767, "step": 11000 }, { "epoch": 0.8460887624028921, "eval_loss": 4.352676868438721, "eval_runtime": 18.5817, "eval_samples_per_second": 53.816, "eval_steps_per_second": 13.454, "step": 11000 }, { "epoch": 0.8499346204138143, "grad_norm": 6.128251075744629, "learning_rate": 0.00014525101940108563, "loss": 4.3993, "step": 11050 }, { "epoch": 0.8537804784247366, "grad_norm": 3.9961323738098145, "learning_rate": 0.00014499129937926917, "loss": 4.406, "step": 11100 }, { "epoch": 0.8576263364356588, "grad_norm": 3.889711856842041, "learning_rate": 0.00014473157935745269, "loss": 4.2991, "step": 11150 }, { "epoch": 0.8614721944465811, "grad_norm": 3.5852463245391846, "learning_rate": 0.00014447185933563617, "loss": 4.2967, "step": 11200 }, { "epoch": 0.8653180524575033, "grad_norm": 3.343247652053833, "learning_rate": 0.0001442121393138197, "loss": 4.314, "step": 11250 }, { "epoch": 0.8653180524575033, "eval_loss": 4.387504577636719, "eval_runtime": 18.554, "eval_samples_per_second": 53.897, "eval_steps_per_second": 13.474, "step": 11250 }, { "epoch": 0.8691639104684256, "grad_norm": 3.5600407123565674, "learning_rate": 0.00014395241929200322, "loss": 4.2958, "step": 11300 }, { "epoch": 0.8730097684793477, "grad_norm": 4.299932956695557, "learning_rate": 0.00014369269927018674, "loss": 4.2966, "step": 11350 }, { "epoch": 0.8768556264902699, "grad_norm": 2.5763466358184814, "learning_rate": 0.00014343297924837028, "loss": 4.3739, "step": 11400 }, { "epoch": 0.8807014845011922, "grad_norm": 3.670653820037842, "learning_rate": 0.0001431732592265538, "loss": 4.3932, "step": 11450 }, { "epoch": 0.8845473425121144, "grad_norm": 5.472078800201416, "learning_rate": 0.0001429135392047373, "loss": 4.3183, "step": 11500 }, { "epoch": 0.8845473425121144, "eval_loss": 4.377117156982422, "eval_runtime": 18.5574, "eval_samples_per_second": 53.887, "eval_steps_per_second": 13.472, "step": 11500 }, { "epoch": 0.8883932005230367, "grad_norm": 4.711415767669678, "learning_rate": 0.00014265381918292082, "loss": 4.5701, "step": 11550 }, { "epoch": 0.8922390585339589, "grad_norm": 3.1737523078918457, "learning_rate": 0.00014239409916110433, "loss": 4.4522, "step": 11600 }, { "epoch": 0.8960849165448812, "grad_norm": 4.876018047332764, "learning_rate": 0.00014213437913928787, "loss": 4.3937, "step": 11650 }, { "epoch": 0.8999307745558034, "grad_norm": 7.117967128753662, "learning_rate": 0.00014187465911747138, "loss": 4.3585, "step": 11700 }, { "epoch": 0.9037766325667257, "grad_norm": 2.587160587310791, "learning_rate": 0.0001416149390956549, "loss": 1.5054, "step": 11750 }, { "epoch": 0.9037766325667257, "eval_loss": 1.5468424558639526, "eval_runtime": 17.9563, "eval_samples_per_second": 55.691, "eval_steps_per_second": 13.923, "step": 11750 }, { "epoch": 0.9076224905776479, "grad_norm": 1.4935526847839355, "learning_rate": 0.0001413552190738384, "loss": 1.5719, "step": 11800 }, { "epoch": 0.9114683485885701, "grad_norm": 1.4879201650619507, "learning_rate": 0.00014109549905202192, "loss": 1.4862, "step": 11850 }, { "epoch": 0.9153142065994924, "grad_norm": 1.7936193943023682, "learning_rate": 0.00014083577903020543, "loss": 1.488, "step": 11900 }, { "epoch": 0.9191600646104146, "grad_norm": 2.14953875541687, "learning_rate": 0.00014057605900838897, "loss": 1.5383, "step": 11950 }, { "epoch": 0.9230059226213368, "grad_norm": 1.5575013160705566, "learning_rate": 0.00014031633898657249, "loss": 1.4814, "step": 12000 }, { "epoch": 0.9230059226213368, "eval_loss": 1.5273067951202393, "eval_runtime": 17.7259, "eval_samples_per_second": 56.414, "eval_steps_per_second": 14.104, "step": 12000 }, { "epoch": 0.926851780632259, "grad_norm": 1.9142653942108154, "learning_rate": 0.000140056618964756, "loss": 1.5436, "step": 12050 }, { "epoch": 0.9306976386431813, "grad_norm": 1.9730989933013916, "learning_rate": 0.00013979689894293954, "loss": 1.5175, "step": 12100 }, { "epoch": 0.9345434966541035, "grad_norm": 0.9956797361373901, "learning_rate": 0.00013953717892112302, "loss": 1.4491, "step": 12150 }, { "epoch": 0.9383893546650258, "grad_norm": 1.0358608961105347, "learning_rate": 0.00013927745889930654, "loss": 1.4488, "step": 12200 }, { "epoch": 0.942235212675948, "grad_norm": 1.5184404850006104, "learning_rate": 0.00013901773887749008, "loss": 1.5037, "step": 12250 }, { "epoch": 0.942235212675948, "eval_loss": 1.5220181941986084, "eval_runtime": 17.7647, "eval_samples_per_second": 56.291, "eval_steps_per_second": 14.073, "step": 12250 }, { "epoch": 0.9460810706868702, "grad_norm": 1.3881593942642212, "learning_rate": 0.0001387580188556736, "loss": 1.4952, "step": 12300 }, { "epoch": 0.9499269286977925, "grad_norm": 2.319173574447632, "learning_rate": 0.0001384982988338571, "loss": 1.4667, "step": 12350 }, { "epoch": 0.9537727867087147, "grad_norm": 1.9042879343032837, "learning_rate": 0.00013823857881204064, "loss": 1.6038, "step": 12400 }, { "epoch": 0.957618644719637, "grad_norm": 1.9162698984146118, "learning_rate": 0.00013797885879022415, "loss": 1.4904, "step": 12450 }, { "epoch": 0.9614645027305592, "grad_norm": 2.0601863861083984, "learning_rate": 0.00013771913876840767, "loss": 1.4758, "step": 12500 }, { "epoch": 0.9614645027305592, "eval_loss": 1.5356587171554565, "eval_runtime": 17.7404, "eval_samples_per_second": 56.369, "eval_steps_per_second": 14.092, "step": 12500 }, { "epoch": 0.9653103607414815, "grad_norm": 1.888836145401001, "learning_rate": 0.00013745941874659118, "loss": 1.5681, "step": 12550 }, { "epoch": 0.9691562187524037, "grad_norm": 1.4329860210418701, "learning_rate": 0.0001371996987247747, "loss": 1.5018, "step": 12600 }, { "epoch": 0.973002076763326, "grad_norm": 1.969533920288086, "learning_rate": 0.00013693997870295823, "loss": 1.5494, "step": 12650 }, { "epoch": 0.9768479347742481, "grad_norm": 2.1219890117645264, "learning_rate": 0.00013668025868114175, "loss": 1.5386, "step": 12700 }, { "epoch": 0.9806937927851703, "grad_norm": 1.6632941961288452, "learning_rate": 0.00013642053865932526, "loss": 1.5253, "step": 12750 }, { "epoch": 0.9806937927851703, "eval_loss": 1.5147372484207153, "eval_runtime": 17.8072, "eval_samples_per_second": 56.157, "eval_steps_per_second": 14.039, "step": 12750 }, { "epoch": 0.9845396507960926, "grad_norm": 1.292913794517517, "learning_rate": 0.00013616081863750877, "loss": 1.4413, "step": 12800 }, { "epoch": 0.9883855088070148, "grad_norm": 1.1377824544906616, "learning_rate": 0.00013590109861569228, "loss": 1.5705, "step": 12850 }, { "epoch": 0.9922313668179371, "grad_norm": 1.656996726989746, "learning_rate": 0.0001356413785938758, "loss": 1.5311, "step": 12900 }, { "epoch": 0.9960772248288593, "grad_norm": 1.6639357805252075, "learning_rate": 0.00013538165857205934, "loss": 1.5676, "step": 12950 }, { "epoch": 0.9999230828397816, "grad_norm": 1.0893466472625732, "learning_rate": 0.00013512193855024285, "loss": 1.5215, "step": 13000 }, { "epoch": 0.9999230828397816, "eval_loss": 1.5119102001190186, "eval_runtime": 17.7291, "eval_samples_per_second": 56.404, "eval_steps_per_second": 14.101, "step": 13000 }, { "epoch": 1.0037689408507038, "grad_norm": 1.4789248704910278, "learning_rate": 0.00013486221852842636, "loss": 1.4551, "step": 13050 }, { "epoch": 1.007614798861626, "grad_norm": 1.3270663022994995, "learning_rate": 0.00013460249850660988, "loss": 1.4776, "step": 13100 }, { "epoch": 1.0114606568725482, "grad_norm": 1.3546854257583618, "learning_rate": 0.0001343427784847934, "loss": 1.5807, "step": 13150 }, { "epoch": 1.0153065148834706, "grad_norm": 1.303915023803711, "learning_rate": 0.0001340830584629769, "loss": 1.4309, "step": 13200 }, { "epoch": 1.0191523728943928, "grad_norm": 0.8854748606681824, "learning_rate": 0.00013382333844116044, "loss": 1.4395, "step": 13250 }, { "epoch": 1.0191523728943928, "eval_loss": 1.5238608121871948, "eval_runtime": 17.7588, "eval_samples_per_second": 56.31, "eval_steps_per_second": 14.077, "step": 13250 }, { "epoch": 1.022998230905315, "grad_norm": 1.5649653673171997, "learning_rate": 0.00013356361841934395, "loss": 1.5022, "step": 13300 }, { "epoch": 1.0268440889162371, "grad_norm": 1.6031616926193237, "learning_rate": 0.0001333038983975275, "loss": 1.4315, "step": 13350 }, { "epoch": 1.0306899469271595, "grad_norm": 1.3788844347000122, "learning_rate": 0.00013304417837571098, "loss": 1.578, "step": 13400 }, { "epoch": 1.0345358049380817, "grad_norm": 1.4347171783447266, "learning_rate": 0.0001327844583538945, "loss": 1.3993, "step": 13450 }, { "epoch": 1.038381662949004, "grad_norm": 1.9777193069458008, "learning_rate": 0.00013252473833207803, "loss": 1.4331, "step": 13500 }, { "epoch": 1.038381662949004, "eval_loss": 1.5231057405471802, "eval_runtime": 17.8114, "eval_samples_per_second": 56.144, "eval_steps_per_second": 14.036, "step": 13500 }, { "epoch": 1.042227520959926, "grad_norm": 2.056574583053589, "learning_rate": 0.00013226501831026155, "loss": 1.5387, "step": 13550 }, { "epoch": 1.0460733789708483, "grad_norm": 1.41805899143219, "learning_rate": 0.00013200529828844506, "loss": 1.4194, "step": 13600 }, { "epoch": 1.0499192369817707, "grad_norm": 1.5727626085281372, "learning_rate": 0.0001317455782666286, "loss": 1.4763, "step": 13650 }, { "epoch": 1.0537650949926929, "grad_norm": 1.8175796270370483, "learning_rate": 0.0001314858582448121, "loss": 1.5232, "step": 13700 }, { "epoch": 1.057610953003615, "grad_norm": 1.459721565246582, "learning_rate": 0.0001312261382229956, "loss": 1.4926, "step": 13750 }, { "epoch": 1.057610953003615, "eval_loss": 1.5073590278625488, "eval_runtime": 17.8208, "eval_samples_per_second": 56.114, "eval_steps_per_second": 14.029, "step": 13750 }, { "epoch": 1.0614568110145373, "grad_norm": 1.7236889600753784, "learning_rate": 0.00013096641820117914, "loss": 1.4485, "step": 13800 }, { "epoch": 1.0653026690254597, "grad_norm": 1.1652172803878784, "learning_rate": 0.00013070669817936265, "loss": 1.4706, "step": 13850 }, { "epoch": 1.0691485270363819, "grad_norm": 1.1279985904693604, "learning_rate": 0.00013044697815754616, "loss": 1.5507, "step": 13900 }, { "epoch": 1.072994385047304, "grad_norm": 2.2368061542510986, "learning_rate": 0.0001301872581357297, "loss": 1.5184, "step": 13950 }, { "epoch": 1.0768402430582262, "grad_norm": 1.1515541076660156, "learning_rate": 0.00012992753811391322, "loss": 1.5184, "step": 14000 }, { "epoch": 1.0768402430582262, "eval_loss": 1.5123000144958496, "eval_runtime": 17.8325, "eval_samples_per_second": 56.077, "eval_steps_per_second": 14.019, "step": 14000 }, { "epoch": 1.0806861010691486, "grad_norm": 2.1172475814819336, "learning_rate": 0.00012966781809209673, "loss": 1.4055, "step": 14050 }, { "epoch": 1.0845319590800708, "grad_norm": 1.196999430656433, "learning_rate": 0.00012940809807028024, "loss": 1.464, "step": 14100 }, { "epoch": 1.088377817090993, "grad_norm": 1.3582040071487427, "learning_rate": 0.00012914837804846375, "loss": 1.4502, "step": 14150 }, { "epoch": 1.0922236751019152, "grad_norm": 1.6588162183761597, "learning_rate": 0.0001288886580266473, "loss": 1.5174, "step": 14200 }, { "epoch": 1.0960695331128374, "grad_norm": 1.7531650066375732, "learning_rate": 0.0001286289380048308, "loss": 1.505, "step": 14250 }, { "epoch": 1.0960695331128374, "eval_loss": 1.5160688161849976, "eval_runtime": 17.764, "eval_samples_per_second": 56.294, "eval_steps_per_second": 14.073, "step": 14250 }, { "epoch": 1.0999153911237598, "grad_norm": 1.868784785270691, "learning_rate": 0.00012836921798301432, "loss": 1.5544, "step": 14300 }, { "epoch": 1.103761249134682, "grad_norm": 1.9493080377578735, "learning_rate": 0.00012810949796119783, "loss": 1.533, "step": 14350 }, { "epoch": 1.1076071071456042, "grad_norm": 0.7309526801109314, "learning_rate": 0.00012784977793938135, "loss": 1.4672, "step": 14400 }, { "epoch": 1.1114529651565264, "grad_norm": 1.3281447887420654, "learning_rate": 0.00012759005791756486, "loss": 1.3874, "step": 14450 }, { "epoch": 1.1152988231674485, "grad_norm": 1.0158611536026, "learning_rate": 0.0001273303378957484, "loss": 1.4966, "step": 14500 }, { "epoch": 1.1152988231674485, "eval_loss": 1.5160739421844482, "eval_runtime": 17.6672, "eval_samples_per_second": 56.602, "eval_steps_per_second": 14.151, "step": 14500 }, { "epoch": 1.119144681178371, "grad_norm": 1.6422228813171387, "learning_rate": 0.0001270706178739319, "loss": 1.5494, "step": 14550 }, { "epoch": 1.1229905391892931, "grad_norm": 0.7187716960906982, "learning_rate": 0.00012681089785211542, "loss": 1.4221, "step": 14600 }, { "epoch": 1.1268363972002153, "grad_norm": 1.2605098485946655, "learning_rate": 0.00012655117783029896, "loss": 1.4299, "step": 14650 }, { "epoch": 1.1306822552111375, "grad_norm": 2.598015069961548, "learning_rate": 0.00012629145780848245, "loss": 1.5757, "step": 14700 }, { "epoch": 1.13452811322206, "grad_norm": 1.4004614353179932, "learning_rate": 0.00012603173778666596, "loss": 1.482, "step": 14750 }, { "epoch": 1.13452811322206, "eval_loss": 1.5089725255966187, "eval_runtime": 17.9036, "eval_samples_per_second": 55.855, "eval_steps_per_second": 13.964, "step": 14750 }, { "epoch": 1.1383739712329821, "grad_norm": 1.3800735473632812, "learning_rate": 0.0001257720177648495, "loss": 1.5285, "step": 14800 }, { "epoch": 1.1422198292439043, "grad_norm": 1.3741459846496582, "learning_rate": 0.00012551229774303301, "loss": 1.5242, "step": 14850 }, { "epoch": 1.1460656872548265, "grad_norm": 2.232680559158325, "learning_rate": 0.00012525257772121653, "loss": 1.4483, "step": 14900 }, { "epoch": 1.149911545265749, "grad_norm": 1.4408409595489502, "learning_rate": 0.00012499285769940007, "loss": 1.5414, "step": 14950 }, { "epoch": 1.153757403276671, "grad_norm": 1.5221819877624512, "learning_rate": 0.00012473313767758355, "loss": 1.5246, "step": 15000 }, { "epoch": 1.153757403276671, "eval_loss": 1.516871452331543, "eval_runtime": 17.9308, "eval_samples_per_second": 55.77, "eval_steps_per_second": 13.943, "step": 15000 }, { "epoch": 1.1576032612875933, "grad_norm": 1.5970553159713745, "learning_rate": 0.0001244734176557671, "loss": 1.4515, "step": 15050 }, { "epoch": 1.1614491192985155, "grad_norm": 1.0201988220214844, "learning_rate": 0.0001242136976339506, "loss": 1.5431, "step": 15100 }, { "epoch": 1.1652949773094377, "grad_norm": 0.9830596446990967, "learning_rate": 0.00012395397761213412, "loss": 1.4581, "step": 15150 }, { "epoch": 1.16914083532036, "grad_norm": 2.3363943099975586, "learning_rate": 0.00012369425759031766, "loss": 1.5147, "step": 15200 }, { "epoch": 1.1729866933312822, "grad_norm": 1.8213731050491333, "learning_rate": 0.00012343453756850117, "loss": 1.5497, "step": 15250 }, { "epoch": 1.1729866933312822, "eval_loss": 1.5053696632385254, "eval_runtime": 17.7358, "eval_samples_per_second": 56.383, "eval_steps_per_second": 14.096, "step": 15250 }, { "epoch": 1.1768325513422044, "grad_norm": 1.1468195915222168, "learning_rate": 0.00012317481754668468, "loss": 1.4292, "step": 15300 }, { "epoch": 1.1806784093531266, "grad_norm": 1.2235878705978394, "learning_rate": 0.0001229150975248682, "loss": 1.4428, "step": 15350 }, { "epoch": 1.1845242673640488, "grad_norm": 1.5518691539764404, "learning_rate": 0.0001226553775030517, "loss": 1.4889, "step": 15400 }, { "epoch": 1.1883701253749712, "grad_norm": 1.217755913734436, "learning_rate": 0.00012239565748123522, "loss": 1.4841, "step": 15450 }, { "epoch": 1.1922159833858934, "grad_norm": 1.6093647480010986, "learning_rate": 0.00012213593745941876, "loss": 1.5461, "step": 15500 }, { "epoch": 1.1922159833858934, "eval_loss": 1.5013692378997803, "eval_runtime": 18.0052, "eval_samples_per_second": 55.539, "eval_steps_per_second": 13.885, "step": 15500 }, { "epoch": 1.1960618413968156, "grad_norm": 3.367638349533081, "learning_rate": 0.00012187621743760226, "loss": 1.4261, "step": 15550 }, { "epoch": 1.1999076994077378, "grad_norm": 1.6329169273376465, "learning_rate": 0.00012161649741578578, "loss": 1.4475, "step": 15600 }, { "epoch": 1.2037535574186602, "grad_norm": 1.397910475730896, "learning_rate": 0.00012135677739396932, "loss": 1.4747, "step": 15650 }, { "epoch": 1.2075994154295824, "grad_norm": 1.7463736534118652, "learning_rate": 0.00012109705737215283, "loss": 1.4745, "step": 15700 }, { "epoch": 1.2114452734405046, "grad_norm": 1.8097542524337769, "learning_rate": 0.00012083733735033633, "loss": 1.5276, "step": 15750 }, { "epoch": 1.2114452734405046, "eval_loss": 1.4955236911773682, "eval_runtime": 17.9054, "eval_samples_per_second": 55.849, "eval_steps_per_second": 13.962, "step": 15750 }, { "epoch": 1.2152911314514268, "grad_norm": 1.745730996131897, "learning_rate": 0.00012057761732851987, "loss": 1.5282, "step": 15800 }, { "epoch": 1.2191369894623492, "grad_norm": 1.3859128952026367, "learning_rate": 0.00012031789730670338, "loss": 1.4795, "step": 15850 }, { "epoch": 1.2229828474732714, "grad_norm": 1.5086127519607544, "learning_rate": 0.0001200581772848869, "loss": 1.437, "step": 15900 }, { "epoch": 1.2268287054841935, "grad_norm": 2.088292121887207, "learning_rate": 0.00011979845726307042, "loss": 1.5011, "step": 15950 }, { "epoch": 1.2306745634951157, "grad_norm": 1.2746011018753052, "learning_rate": 0.00011953873724125393, "loss": 1.4596, "step": 16000 }, { "epoch": 1.2306745634951157, "eval_loss": 1.4984314441680908, "eval_runtime": 17.8066, "eval_samples_per_second": 56.159, "eval_steps_per_second": 14.04, "step": 16000 }, { "epoch": 1.234520421506038, "grad_norm": 6.292486667633057, "learning_rate": 0.00011927901721943746, "loss": 1.3934, "step": 16050 }, { "epoch": 1.2383662795169603, "grad_norm": 1.6574532985687256, "learning_rate": 0.00011901929719762097, "loss": 1.4821, "step": 16100 }, { "epoch": 1.2422121375278825, "grad_norm": 1.8651037216186523, "learning_rate": 0.00011875957717580448, "loss": 1.49, "step": 16150 }, { "epoch": 1.2460579955388047, "grad_norm": 1.3768175840377808, "learning_rate": 0.00011849985715398801, "loss": 1.4579, "step": 16200 }, { "epoch": 1.2499038535497269, "grad_norm": 1.1569020748138428, "learning_rate": 0.00011824013713217152, "loss": 1.4029, "step": 16250 }, { "epoch": 1.2499038535497269, "eval_loss": 1.4893407821655273, "eval_runtime": 17.9552, "eval_samples_per_second": 55.694, "eval_steps_per_second": 13.924, "step": 16250 }, { "epoch": 1.253749711560649, "grad_norm": 1.8632296323776245, "learning_rate": 0.00011798041711035504, "loss": 1.4592, "step": 16300 }, { "epoch": 1.2575955695715715, "grad_norm": 1.8080470561981201, "learning_rate": 0.00011772069708853856, "loss": 1.4678, "step": 16350 }, { "epoch": 1.2614414275824937, "grad_norm": 1.4193981885910034, "learning_rate": 0.00011746097706672208, "loss": 1.5031, "step": 16400 }, { "epoch": 1.2652872855934159, "grad_norm": 1.5050238370895386, "learning_rate": 0.00011720125704490559, "loss": 1.4107, "step": 16450 }, { "epoch": 1.2691331436043383, "grad_norm": 1.12454092502594, "learning_rate": 0.00011694153702308911, "loss": 1.4572, "step": 16500 }, { "epoch": 1.2691331436043383, "eval_loss": 1.4972718954086304, "eval_runtime": 17.8677, "eval_samples_per_second": 55.967, "eval_steps_per_second": 13.992, "step": 16500 }, { "epoch": 1.2729790016152602, "grad_norm": 1.3523976802825928, "learning_rate": 0.00011668181700127263, "loss": 1.4796, "step": 16550 }, { "epoch": 1.2768248596261826, "grad_norm": 1.9770869016647339, "learning_rate": 0.00011642209697945614, "loss": 1.4756, "step": 16600 }, { "epoch": 1.2806707176371048, "grad_norm": 1.7973159551620483, "learning_rate": 0.00011616237695763967, "loss": 1.4163, "step": 16650 }, { "epoch": 1.284516575648027, "grad_norm": 1.3054739236831665, "learning_rate": 0.00011590265693582318, "loss": 1.4105, "step": 16700 }, { "epoch": 1.2883624336589494, "grad_norm": 1.457047939300537, "learning_rate": 0.00011564293691400672, "loss": 1.4698, "step": 16750 }, { "epoch": 1.2883624336589494, "eval_loss": 1.4798808097839355, "eval_runtime": 17.8696, "eval_samples_per_second": 55.961, "eval_steps_per_second": 13.99, "step": 16750 }, { "epoch": 1.2922082916698716, "grad_norm": 2.117663860321045, "learning_rate": 0.00011538841129262656, "loss": 1.4252, "step": 16800 }, { "epoch": 1.2960541496807938, "grad_norm": 3.7321341037750244, "learning_rate": 0.00011512869127081007, "loss": 1.4818, "step": 16850 }, { "epoch": 1.299900007691716, "grad_norm": 2.497528553009033, "learning_rate": 0.00011486897124899358, "loss": 1.4176, "step": 16900 }, { "epoch": 1.3037458657026382, "grad_norm": 1.6085398197174072, "learning_rate": 0.00011460925122717712, "loss": 1.4373, "step": 16950 }, { "epoch": 1.3075917237135606, "grad_norm": 1.8937525749206543, "learning_rate": 0.00011434953120536062, "loss": 1.4823, "step": 17000 }, { "epoch": 1.3075917237135606, "eval_loss": 1.4854488372802734, "eval_runtime": 17.8761, "eval_samples_per_second": 55.941, "eval_steps_per_second": 13.985, "step": 17000 }, { "epoch": 1.3114375817244828, "grad_norm": 1.6637665033340454, "learning_rate": 0.00011408981118354414, "loss": 1.5071, "step": 17050 }, { "epoch": 1.315283439735405, "grad_norm": 2.0815582275390625, "learning_rate": 0.00011383009116172768, "loss": 1.4606, "step": 17100 }, { "epoch": 1.3191292977463271, "grad_norm": 1.6492595672607422, "learning_rate": 0.00011357037113991117, "loss": 1.4408, "step": 17150 }, { "epoch": 1.3229751557572493, "grad_norm": 0.8617509603500366, "learning_rate": 0.00011331065111809469, "loss": 1.497, "step": 17200 }, { "epoch": 1.3268210137681717, "grad_norm": 1.6395294666290283, "learning_rate": 0.00011305093109627823, "loss": 1.4774, "step": 17250 }, { "epoch": 1.3268210137681717, "eval_loss": 1.4835026264190674, "eval_runtime": 18.0136, "eval_samples_per_second": 55.514, "eval_steps_per_second": 13.878, "step": 17250 }, { "epoch": 1.330666871779094, "grad_norm": 2.7765560150146484, "learning_rate": 0.00011279121107446174, "loss": 1.4864, "step": 17300 }, { "epoch": 1.3345127297900161, "grad_norm": 1.2104064226150513, "learning_rate": 0.00011253149105264524, "loss": 1.4075, "step": 17350 }, { "epoch": 1.3383585878009385, "grad_norm": 1.6772801876068115, "learning_rate": 0.00011227177103082878, "loss": 1.4629, "step": 17400 }, { "epoch": 1.3422044458118605, "grad_norm": 2.254371404647827, "learning_rate": 0.00011201205100901229, "loss": 1.4447, "step": 17450 }, { "epoch": 1.346050303822783, "grad_norm": 2.2015669345855713, "learning_rate": 0.00011175233098719582, "loss": 1.4664, "step": 17500 }, { "epoch": 1.346050303822783, "eval_loss": 1.496685266494751, "eval_runtime": 17.7963, "eval_samples_per_second": 56.191, "eval_steps_per_second": 14.048, "step": 17500 }, { "epoch": 1.349896161833705, "grad_norm": 1.740045428276062, "learning_rate": 0.00011149261096537933, "loss": 1.4486, "step": 17550 }, { "epoch": 1.3537420198446273, "grad_norm": 1.299919605255127, "learning_rate": 0.00011123289094356284, "loss": 1.4868, "step": 17600 }, { "epoch": 1.3575878778555497, "grad_norm": 1.6288009881973267, "learning_rate": 0.00011097317092174637, "loss": 1.4595, "step": 17650 }, { "epoch": 1.3614337358664719, "grad_norm": 0.8747851252555847, "learning_rate": 0.00011071345089992988, "loss": 1.4241, "step": 17700 }, { "epoch": 1.365279593877394, "grad_norm": 1.9510573148727417, "learning_rate": 0.0001104537308781134, "loss": 1.496, "step": 17750 }, { "epoch": 1.365279593877394, "eval_loss": 1.4812238216400146, "eval_runtime": 17.8955, "eval_samples_per_second": 55.88, "eval_steps_per_second": 13.97, "step": 17750 }, { "epoch": 1.3691254518883162, "grad_norm": 1.4853876829147339, "learning_rate": 0.00011019401085629692, "loss": 1.4645, "step": 17800 }, { "epoch": 1.3729713098992384, "grad_norm": 1.5125057697296143, "learning_rate": 0.00010993429083448044, "loss": 1.4052, "step": 17850 }, { "epoch": 1.3768171679101608, "grad_norm": 0.7320863008499146, "learning_rate": 0.00010967457081266395, "loss": 1.5016, "step": 17900 }, { "epoch": 1.380663025921083, "grad_norm": 1.9995285272598267, "learning_rate": 0.00010941485079084747, "loss": 1.4234, "step": 17950 }, { "epoch": 1.3845088839320052, "grad_norm": 0.8304823637008667, "learning_rate": 0.00010915513076903099, "loss": 1.4068, "step": 18000 }, { "epoch": 1.3845088839320052, "eval_loss": 1.4845945835113525, "eval_runtime": 17.7402, "eval_samples_per_second": 56.369, "eval_steps_per_second": 14.092, "step": 18000 }, { "epoch": 1.3883547419429274, "grad_norm": 1.8189436197280884, "learning_rate": 0.0001088954107472145, "loss": 1.4872, "step": 18050 }, { "epoch": 1.3922005999538496, "grad_norm": 1.4212762117385864, "learning_rate": 0.00010863569072539803, "loss": 1.4067, "step": 18100 }, { "epoch": 1.396046457964772, "grad_norm": 1.9733264446258545, "learning_rate": 0.00010837597070358154, "loss": 1.4787, "step": 18150 }, { "epoch": 1.3998923159756942, "grad_norm": 1.320064663887024, "learning_rate": 0.00010811625068176505, "loss": 1.5242, "step": 18200 }, { "epoch": 1.4037381739866164, "grad_norm": 1.4177141189575195, "learning_rate": 0.00010785653065994858, "loss": 1.4719, "step": 18250 }, { "epoch": 1.4037381739866164, "eval_loss": 1.4992233514785767, "eval_runtime": 17.8574, "eval_samples_per_second": 55.999, "eval_steps_per_second": 14.0, "step": 18250 }, { "epoch": 1.4075840319975388, "grad_norm": 1.8219791650772095, "learning_rate": 0.00010759681063813209, "loss": 1.4392, "step": 18300 }, { "epoch": 1.4114298900084608, "grad_norm": 1.3196603059768677, "learning_rate": 0.00010733709061631563, "loss": 1.4837, "step": 18350 }, { "epoch": 1.4152757480193832, "grad_norm": 1.01405668258667, "learning_rate": 0.00010707737059449914, "loss": 1.455, "step": 18400 }, { "epoch": 1.4191216060303053, "grad_norm": 1.8538917303085327, "learning_rate": 0.00010681765057268264, "loss": 1.4209, "step": 18450 }, { "epoch": 1.4229674640412275, "grad_norm": 0.8785907030105591, "learning_rate": 0.00010655793055086618, "loss": 1.4716, "step": 18500 }, { "epoch": 1.4229674640412275, "eval_loss": 1.485592246055603, "eval_runtime": 17.7246, "eval_samples_per_second": 56.419, "eval_steps_per_second": 14.105, "step": 18500 }, { "epoch": 1.42681332205215, "grad_norm": 1.0700381994247437, "learning_rate": 0.0001062982105290497, "loss": 1.3846, "step": 18550 }, { "epoch": 1.4306591800630721, "grad_norm": 1.2848351001739502, "learning_rate": 0.0001060384905072332, "loss": 1.4874, "step": 18600 }, { "epoch": 1.4345050380739943, "grad_norm": 1.4261386394500732, "learning_rate": 0.00010577877048541674, "loss": 1.5771, "step": 18650 }, { "epoch": 1.4383508960849165, "grad_norm": 1.3613426685333252, "learning_rate": 0.00010551905046360025, "loss": 1.4445, "step": 18700 }, { "epoch": 1.4421967540958387, "grad_norm": 0.8625685572624207, "learning_rate": 0.00010525933044178375, "loss": 1.3845, "step": 18750 }, { "epoch": 1.4421967540958387, "eval_loss": 1.4757392406463623, "eval_runtime": 17.8903, "eval_samples_per_second": 55.896, "eval_steps_per_second": 13.974, "step": 18750 }, { "epoch": 1.446042612106761, "grad_norm": 1.5575672388076782, "learning_rate": 0.00010499961041996729, "loss": 1.5387, "step": 18800 }, { "epoch": 1.4498884701176833, "grad_norm": 2.2173306941986084, "learning_rate": 0.0001047398903981508, "loss": 1.459, "step": 18850 }, { "epoch": 1.4537343281286055, "grad_norm": 1.904809832572937, "learning_rate": 0.00010448017037633431, "loss": 1.5229, "step": 18900 }, { "epoch": 1.4575801861395277, "grad_norm": 1.8832893371582031, "learning_rate": 0.00010422045035451784, "loss": 1.4093, "step": 18950 }, { "epoch": 1.4614260441504499, "grad_norm": 1.852971076965332, "learning_rate": 0.00010396073033270135, "loss": 1.389, "step": 19000 }, { "epoch": 1.4614260441504499, "eval_loss": 1.4788576364517212, "eval_runtime": 17.7751, "eval_samples_per_second": 56.258, "eval_steps_per_second": 14.065, "step": 19000 }, { "epoch": 1.4652719021613723, "grad_norm": 1.2875189781188965, "learning_rate": 0.00010370101031088487, "loss": 1.4826, "step": 19050 }, { "epoch": 1.4691177601722945, "grad_norm": 1.7036223411560059, "learning_rate": 0.00010344129028906839, "loss": 1.4392, "step": 19100 }, { "epoch": 1.4729636181832166, "grad_norm": 1.537514328956604, "learning_rate": 0.0001031815702672519, "loss": 1.4846, "step": 19150 }, { "epoch": 1.476809476194139, "grad_norm": 0.9159242510795593, "learning_rate": 0.00010292185024543543, "loss": 1.4668, "step": 19200 }, { "epoch": 1.480655334205061, "grad_norm": 3.47868013381958, "learning_rate": 0.00010266213022361894, "loss": 1.493, "step": 19250 }, { "epoch": 1.480655334205061, "eval_loss": 1.4858986139297485, "eval_runtime": 17.7738, "eval_samples_per_second": 56.263, "eval_steps_per_second": 14.066, "step": 19250 }, { "epoch": 1.4845011922159834, "grad_norm": 1.467437505722046, "learning_rate": 0.00010240241020180246, "loss": 1.4639, "step": 19300 }, { "epoch": 1.4883470502269056, "grad_norm": 1.2710049152374268, "learning_rate": 0.00010214269017998598, "loss": 1.4244, "step": 19350 }, { "epoch": 1.4921929082378278, "grad_norm": 2.0059661865234375, "learning_rate": 0.0001018829701581695, "loss": 1.4258, "step": 19400 }, { "epoch": 1.4960387662487502, "grad_norm": 1.7536308765411377, "learning_rate": 0.00010162325013635301, "loss": 1.396, "step": 19450 }, { "epoch": 1.4998846242596724, "grad_norm": 0.9684279561042786, "learning_rate": 0.00010136353011453655, "loss": 1.4598, "step": 19500 }, { "epoch": 1.4998846242596724, "eval_loss": 1.4841110706329346, "eval_runtime": 17.9151, "eval_samples_per_second": 55.819, "eval_steps_per_second": 13.955, "step": 19500 }, { "epoch": 1.5037304822705946, "grad_norm": 0.9244908690452576, "learning_rate": 0.00010110381009272005, "loss": 1.3834, "step": 19550 }, { "epoch": 1.5075763402815168, "grad_norm": 1.6488862037658691, "learning_rate": 0.00010084409007090356, "loss": 1.5311, "step": 19600 }, { "epoch": 1.511422198292439, "grad_norm": 1.9130067825317383, "learning_rate": 0.0001005843700490871, "loss": 1.5212, "step": 19650 }, { "epoch": 1.5152680563033614, "grad_norm": 1.326277256011963, "learning_rate": 0.0001003246500272706, "loss": 1.4067, "step": 19700 }, { "epoch": 1.5191139143142836, "grad_norm": 1.7258195877075195, "learning_rate": 0.00010006493000545411, "loss": 1.4844, "step": 19750 }, { "epoch": 1.5191139143142836, "eval_loss": 1.4816969633102417, "eval_runtime": 17.826, "eval_samples_per_second": 56.098, "eval_steps_per_second": 14.024, "step": 19750 }, { "epoch": 1.5229597723252057, "grad_norm": 1.8164838552474976, "learning_rate": 9.980520998363765e-05, "loss": 1.4233, "step": 19800 }, { "epoch": 1.5268056303361282, "grad_norm": 1.5884016752243042, "learning_rate": 9.954548996182115e-05, "loss": 1.4313, "step": 19850 }, { "epoch": 1.5306514883470501, "grad_norm": 1.5381648540496826, "learning_rate": 9.928576994000468e-05, "loss": 1.4789, "step": 19900 }, { "epoch": 1.5344973463579725, "grad_norm": 1.6448626518249512, "learning_rate": 9.90260499181882e-05, "loss": 1.3756, "step": 19950 }, { "epoch": 1.5383432043688947, "grad_norm": 1.6137230396270752, "learning_rate": 9.876632989637172e-05, "loss": 1.4171, "step": 20000 }, { "epoch": 1.5383432043688947, "eval_loss": 1.4770597219467163, "eval_runtime": 17.9365, "eval_samples_per_second": 55.752, "eval_steps_per_second": 13.938, "step": 20000 }, { "epoch": 1.542189062379817, "grad_norm": 1.7058050632476807, "learning_rate": 9.850660987455523e-05, "loss": 1.4646, "step": 20050 }, { "epoch": 1.5460349203907393, "grad_norm": 2.2624917030334473, "learning_rate": 9.824688985273876e-05, "loss": 1.4586, "step": 20100 }, { "epoch": 1.5498807784016613, "grad_norm": 2.216883420944214, "learning_rate": 9.798716983092227e-05, "loss": 1.4376, "step": 20150 }, { "epoch": 1.5537266364125837, "grad_norm": 1.9749584197998047, "learning_rate": 9.772744980910578e-05, "loss": 1.47, "step": 20200 }, { "epoch": 1.5575724944235059, "grad_norm": 2.185480833053589, "learning_rate": 9.746772978728931e-05, "loss": 1.39, "step": 20250 }, { "epoch": 1.5575724944235059, "eval_loss": 1.4724150896072388, "eval_runtime": 17.8191, "eval_samples_per_second": 56.12, "eval_steps_per_second": 14.03, "step": 20250 }, { "epoch": 1.561418352434428, "grad_norm": 1.5810290575027466, "learning_rate": 9.720800976547284e-05, "loss": 1.4267, "step": 20300 }, { "epoch": 1.5652642104453505, "grad_norm": 2.0329344272613525, "learning_rate": 9.694828974365633e-05, "loss": 1.5066, "step": 20350 }, { "epoch": 1.5691100684562724, "grad_norm": 2.7335126399993896, "learning_rate": 9.668856972183986e-05, "loss": 1.4749, "step": 20400 }, { "epoch": 1.5729559264671948, "grad_norm": 1.0576220750808716, "learning_rate": 9.642884970002339e-05, "loss": 1.4318, "step": 20450 }, { "epoch": 1.576801784478117, "grad_norm": 0.6857870817184448, "learning_rate": 9.616912967820689e-05, "loss": 1.3567, "step": 20500 }, { "epoch": 1.576801784478117, "eval_loss": 1.479669213294983, "eval_runtime": 17.8378, "eval_samples_per_second": 56.061, "eval_steps_per_second": 14.015, "step": 20500 }, { "epoch": 1.5806476424890392, "grad_norm": 1.430114507675171, "learning_rate": 9.590940965639041e-05, "loss": 1.3714, "step": 20550 }, { "epoch": 1.5844935004999616, "grad_norm": 1.7613717317581177, "learning_rate": 9.564968963457394e-05, "loss": 1.4285, "step": 20600 }, { "epoch": 1.5883393585108838, "grad_norm": 1.3678529262542725, "learning_rate": 9.538996961275745e-05, "loss": 1.4463, "step": 20650 }, { "epoch": 1.592185216521806, "grad_norm": 3.1135504245758057, "learning_rate": 9.513024959094097e-05, "loss": 1.4287, "step": 20700 }, { "epoch": 1.5960310745327284, "grad_norm": 1.7866570949554443, "learning_rate": 9.487052956912449e-05, "loss": 1.4515, "step": 20750 }, { "epoch": 1.5960310745327284, "eval_loss": 1.4817472696304321, "eval_runtime": 17.9267, "eval_samples_per_second": 55.783, "eval_steps_per_second": 13.946, "step": 20750 }, { "epoch": 1.5998769325436504, "grad_norm": 1.2406786680221558, "learning_rate": 9.4610809547308e-05, "loss": 1.5364, "step": 20800 }, { "epoch": 1.6037227905545728, "grad_norm": 1.9183951616287231, "learning_rate": 9.435108952549152e-05, "loss": 1.5169, "step": 20850 }, { "epoch": 1.607568648565495, "grad_norm": 1.852089524269104, "learning_rate": 9.409136950367504e-05, "loss": 1.4668, "step": 20900 }, { "epoch": 1.6114145065764172, "grad_norm": 2.3430335521698, "learning_rate": 9.383164948185856e-05, "loss": 1.4297, "step": 20950 }, { "epoch": 1.6152603645873396, "grad_norm": 1.9646743535995483, "learning_rate": 9.357192946004207e-05, "loss": 1.5143, "step": 21000 }, { "epoch": 1.6152603645873396, "eval_loss": 1.4690666198730469, "eval_runtime": 17.7563, "eval_samples_per_second": 56.318, "eval_steps_per_second": 14.079, "step": 21000 }, { "epoch": 1.6191062225982615, "grad_norm": 1.8204776048660278, "learning_rate": 9.33122094382256e-05, "loss": 1.4423, "step": 21050 }, { "epoch": 1.622952080609184, "grad_norm": 1.6150448322296143, "learning_rate": 9.305248941640912e-05, "loss": 1.4368, "step": 21100 }, { "epoch": 1.6267979386201061, "grad_norm": 2.4393765926361084, "learning_rate": 9.279276939459264e-05, "loss": 1.6115, "step": 21150 }, { "epoch": 1.6306437966310283, "grad_norm": 1.217399001121521, "learning_rate": 9.253304937277615e-05, "loss": 1.5792, "step": 21200 }, { "epoch": 1.6344896546419507, "grad_norm": 1.4467514753341675, "learning_rate": 9.227332935095967e-05, "loss": 1.4838, "step": 21250 }, { "epoch": 1.6344896546419507, "eval_loss": 1.4741238355636597, "eval_runtime": 17.7656, "eval_samples_per_second": 56.289, "eval_steps_per_second": 14.072, "step": 21250 }, { "epoch": 1.6383355126528727, "grad_norm": 1.4274511337280273, "learning_rate": 9.201360932914319e-05, "loss": 1.4747, "step": 21300 }, { "epoch": 1.642181370663795, "grad_norm": 1.5080102682113647, "learning_rate": 9.175908370776304e-05, "loss": 1.4262, "step": 21350 }, { "epoch": 1.6460272286747173, "grad_norm": 1.2717032432556152, "learning_rate": 9.149936368594655e-05, "loss": 1.4418, "step": 21400 }, { "epoch": 1.6498730866856395, "grad_norm": 1.3594080209732056, "learning_rate": 9.123964366413006e-05, "loss": 1.5184, "step": 21450 }, { "epoch": 1.6537189446965619, "grad_norm": 1.896607518196106, "learning_rate": 9.097992364231359e-05, "loss": 1.4702, "step": 21500 }, { "epoch": 1.6537189446965619, "eval_loss": 1.4777735471725464, "eval_runtime": 17.8161, "eval_samples_per_second": 56.129, "eval_steps_per_second": 14.032, "step": 21500 }, { "epoch": 1.657564802707484, "grad_norm": 1.3839844465255737, "learning_rate": 9.072020362049712e-05, "loss": 1.522, "step": 21550 }, { "epoch": 1.6614106607184063, "grad_norm": 1.187853455543518, "learning_rate": 9.046048359868063e-05, "loss": 1.4469, "step": 21600 }, { "epoch": 1.6652565187293287, "grad_norm": 2.8398866653442383, "learning_rate": 9.020076357686414e-05, "loss": 1.4875, "step": 21650 }, { "epoch": 1.6691023767402506, "grad_norm": 1.801963448524475, "learning_rate": 8.994104355504767e-05, "loss": 1.3892, "step": 21700 }, { "epoch": 1.672948234751173, "grad_norm": 1.7727116346359253, "learning_rate": 8.968132353323118e-05, "loss": 1.3886, "step": 21750 }, { "epoch": 1.672948234751173, "eval_loss": 1.4662117958068848, "eval_runtime": 17.7809, "eval_samples_per_second": 56.24, "eval_steps_per_second": 14.06, "step": 21750 }, { "epoch": 1.6767940927620952, "grad_norm": 1.7052053213119507, "learning_rate": 8.94216035114147e-05, "loss": 1.4483, "step": 21800 }, { "epoch": 1.6806399507730174, "grad_norm": 1.9812465906143188, "learning_rate": 8.916188348959822e-05, "loss": 1.4325, "step": 21850 }, { "epoch": 1.6844858087839398, "grad_norm": 1.2499721050262451, "learning_rate": 8.890216346778173e-05, "loss": 1.5246, "step": 21900 }, { "epoch": 1.6883316667948618, "grad_norm": 1.2503259181976318, "learning_rate": 8.864244344596525e-05, "loss": 1.4661, "step": 21950 }, { "epoch": 1.6921775248057842, "grad_norm": 2.201223134994507, "learning_rate": 8.838272342414877e-05, "loss": 1.5103, "step": 22000 }, { "epoch": 1.6921775248057842, "eval_loss": 1.470120906829834, "eval_runtime": 17.7715, "eval_samples_per_second": 56.27, "eval_steps_per_second": 14.067, "step": 22000 }, { "epoch": 1.6960233828167064, "grad_norm": 0.8225556015968323, "learning_rate": 8.81230034023323e-05, "loss": 1.4372, "step": 22050 }, { "epoch": 1.6998692408276286, "grad_norm": 1.495335578918457, "learning_rate": 8.78632833805158e-05, "loss": 1.4231, "step": 22100 }, { "epoch": 1.703715098838551, "grad_norm": 1.6913652420043945, "learning_rate": 8.760356335869933e-05, "loss": 1.4859, "step": 22150 }, { "epoch": 1.707560956849473, "grad_norm": 1.9825598001480103, "learning_rate": 8.734384333688285e-05, "loss": 1.4354, "step": 22200 }, { "epoch": 1.7114068148603954, "grad_norm": 2.212759017944336, "learning_rate": 8.708412331506635e-05, "loss": 1.3861, "step": 22250 }, { "epoch": 1.7114068148603954, "eval_loss": 1.4596961736679077, "eval_runtime": 17.8176, "eval_samples_per_second": 56.124, "eval_steps_per_second": 14.031, "step": 22250 }, { "epoch": 1.7152526728713176, "grad_norm": 1.4325975179672241, "learning_rate": 8.682440329324988e-05, "loss": 1.4805, "step": 22300 }, { "epoch": 1.7190985308822397, "grad_norm": 1.9796292781829834, "learning_rate": 8.65646832714334e-05, "loss": 1.3836, "step": 22350 }, { "epoch": 1.7229443888931621, "grad_norm": 1.6221562623977661, "learning_rate": 8.630496324961692e-05, "loss": 1.4176, "step": 22400 }, { "epoch": 1.7267902469040843, "grad_norm": 1.1431959867477417, "learning_rate": 8.604524322780043e-05, "loss": 1.4553, "step": 22450 }, { "epoch": 1.7306361049150065, "grad_norm": 1.1562083959579468, "learning_rate": 8.578552320598396e-05, "loss": 1.4489, "step": 22500 }, { "epoch": 1.7306361049150065, "eval_loss": 1.464021921157837, "eval_runtime": 17.8166, "eval_samples_per_second": 56.127, "eval_steps_per_second": 14.032, "step": 22500 }, { "epoch": 1.734481962925929, "grad_norm": 0.8531803488731384, "learning_rate": 8.552580318416747e-05, "loss": 1.4298, "step": 22550 }, { "epoch": 1.738327820936851, "grad_norm": 1.3987632989883423, "learning_rate": 8.526608316235098e-05, "loss": 1.4629, "step": 22600 }, { "epoch": 1.7421736789477733, "grad_norm": 1.4521870613098145, "learning_rate": 8.500636314053451e-05, "loss": 1.4005, "step": 22650 }, { "epoch": 1.7460195369586955, "grad_norm": 1.0557054281234741, "learning_rate": 8.474664311871803e-05, "loss": 1.4079, "step": 22700 }, { "epoch": 1.7498653949696177, "grad_norm": 1.5067927837371826, "learning_rate": 8.448692309690155e-05, "loss": 1.4664, "step": 22750 }, { "epoch": 1.7498653949696177, "eval_loss": 1.4728831052780151, "eval_runtime": 17.7456, "eval_samples_per_second": 56.352, "eval_steps_per_second": 14.088, "step": 22750 }, { "epoch": 1.75371125298054, "grad_norm": 1.3237221240997314, "learning_rate": 8.422720307508506e-05, "loss": 1.5266, "step": 22800 }, { "epoch": 1.757557110991462, "grad_norm": 1.4342700242996216, "learning_rate": 8.396748305326859e-05, "loss": 1.4929, "step": 22850 }, { "epoch": 1.7614029690023845, "grad_norm": 5.2202534675598145, "learning_rate": 8.37077630314521e-05, "loss": 1.457, "step": 22900 }, { "epoch": 1.7652488270133067, "grad_norm": 2.7584545612335205, "learning_rate": 8.344804300963561e-05, "loss": 1.4523, "step": 22950 }, { "epoch": 1.7690946850242288, "grad_norm": 1.8208624124526978, "learning_rate": 8.318832298781914e-05, "loss": 1.4746, "step": 23000 }, { "epoch": 1.7690946850242288, "eval_loss": 1.467396855354309, "eval_runtime": 17.663, "eval_samples_per_second": 56.616, "eval_steps_per_second": 14.154, "step": 23000 }, { "epoch": 1.7729405430351513, "grad_norm": 1.6651790142059326, "learning_rate": 8.292860296600265e-05, "loss": 1.4537, "step": 23050 }, { "epoch": 1.7767864010460732, "grad_norm": 1.5910587310791016, "learning_rate": 8.266888294418618e-05, "loss": 1.5305, "step": 23100 }, { "epoch": 1.7806322590569956, "grad_norm": 1.2711199522018433, "learning_rate": 8.240916292236969e-05, "loss": 1.4877, "step": 23150 }, { "epoch": 1.7844781170679178, "grad_norm": 1.7133463621139526, "learning_rate": 8.21494429005532e-05, "loss": 1.4547, "step": 23200 }, { "epoch": 1.78832397507884, "grad_norm": 1.6850301027297974, "learning_rate": 8.188972287873673e-05, "loss": 1.4228, "step": 23250 }, { "epoch": 1.78832397507884, "eval_loss": 1.4735645055770874, "eval_runtime": 17.7983, "eval_samples_per_second": 56.185, "eval_steps_per_second": 14.046, "step": 23250 }, { "epoch": 1.7921698330897624, "grad_norm": 1.1047898530960083, "learning_rate": 8.163000285692024e-05, "loss": 1.4817, "step": 23300 }, { "epoch": 1.7960156911006846, "grad_norm": 1.8970743417739868, "learning_rate": 8.137028283510376e-05, "loss": 1.4884, "step": 23350 }, { "epoch": 1.7998615491116068, "grad_norm": 1.4730740785598755, "learning_rate": 8.111056281328728e-05, "loss": 1.4436, "step": 23400 }, { "epoch": 1.8037074071225292, "grad_norm": 1.8288697004318237, "learning_rate": 8.08508427914708e-05, "loss": 1.4614, "step": 23450 }, { "epoch": 1.8075532651334512, "grad_norm": 1.2339516878128052, "learning_rate": 8.059112276965432e-05, "loss": 1.4276, "step": 23500 }, { "epoch": 1.8075532651334512, "eval_loss": 1.4655163288116455, "eval_runtime": 17.9781, "eval_samples_per_second": 55.623, "eval_steps_per_second": 13.906, "step": 23500 }, { "epoch": 1.8113991231443736, "grad_norm": 1.8125578165054321, "learning_rate": 8.033140274783783e-05, "loss": 1.4318, "step": 23550 }, { "epoch": 1.8152449811552958, "grad_norm": 1.941846489906311, "learning_rate": 8.007168272602136e-05, "loss": 1.4569, "step": 23600 }, { "epoch": 1.819090839166218, "grad_norm": 2.054161787033081, "learning_rate": 7.981196270420487e-05, "loss": 1.4426, "step": 23650 }, { "epoch": 1.8229366971771404, "grad_norm": 1.6102700233459473, "learning_rate": 7.955224268238839e-05, "loss": 1.4407, "step": 23700 }, { "epoch": 1.8267825551880623, "grad_norm": 1.7302616834640503, "learning_rate": 7.929252266057191e-05, "loss": 1.4432, "step": 23750 }, { "epoch": 1.8267825551880623, "eval_loss": 1.4570631980895996, "eval_runtime": 17.8893, "eval_samples_per_second": 55.899, "eval_steps_per_second": 13.975, "step": 23750 }, { "epoch": 1.8306284131989847, "grad_norm": 9.521604537963867, "learning_rate": 7.903280263875543e-05, "loss": 1.3995, "step": 23800 }, { "epoch": 1.834474271209907, "grad_norm": 1.9546847343444824, "learning_rate": 7.877308261693894e-05, "loss": 1.4957, "step": 23850 }, { "epoch": 1.838320129220829, "grad_norm": 1.173722505569458, "learning_rate": 7.851336259512246e-05, "loss": 1.4157, "step": 23900 }, { "epoch": 1.8421659872317515, "grad_norm": 1.5322128534317017, "learning_rate": 7.825364257330599e-05, "loss": 1.4453, "step": 23950 }, { "epoch": 1.8460118452426735, "grad_norm": 1.0376055240631104, "learning_rate": 7.799392255148949e-05, "loss": 1.4646, "step": 24000 }, { "epoch": 1.8460118452426735, "eval_loss": 1.4612687826156616, "eval_runtime": 17.984, "eval_samples_per_second": 55.605, "eval_steps_per_second": 13.901, "step": 24000 }, { "epoch": 1.8498577032535959, "grad_norm": 1.5047483444213867, "learning_rate": 7.773420252967302e-05, "loss": 1.4546, "step": 24050 }, { "epoch": 1.853703561264518, "grad_norm": 1.0463405847549438, "learning_rate": 7.747448250785654e-05, "loss": 1.5014, "step": 24100 }, { "epoch": 1.8575494192754403, "grad_norm": 1.8368524312973022, "learning_rate": 7.721476248604004e-05, "loss": 1.4616, "step": 24150 }, { "epoch": 1.8613952772863627, "grad_norm": 1.4084677696228027, "learning_rate": 7.695504246422357e-05, "loss": 1.4255, "step": 24200 }, { "epoch": 1.8652411352972849, "grad_norm": 1.2279951572418213, "learning_rate": 7.66953224424071e-05, "loss": 1.4254, "step": 24250 }, { "epoch": 1.8652411352972849, "eval_loss": 1.460336685180664, "eval_runtime": 17.7185, "eval_samples_per_second": 56.438, "eval_steps_per_second": 14.11, "step": 24250 }, { "epoch": 1.869086993308207, "grad_norm": 1.9729641675949097, "learning_rate": 7.643560242059061e-05, "loss": 1.4656, "step": 24300 }, { "epoch": 1.8729328513191295, "grad_norm": 0.9121168255805969, "learning_rate": 7.617588239877412e-05, "loss": 1.3949, "step": 24350 }, { "epoch": 1.8767787093300514, "grad_norm": 1.8953206539154053, "learning_rate": 7.591616237695765e-05, "loss": 1.4006, "step": 24400 }, { "epoch": 1.8806245673409738, "grad_norm": 1.5828944444656372, "learning_rate": 7.565644235514116e-05, "loss": 1.5085, "step": 24450 }, { "epoch": 1.884470425351896, "grad_norm": 2.027841329574585, "learning_rate": 7.539672233332467e-05, "loss": 1.3978, "step": 24500 }, { "epoch": 1.884470425351896, "eval_loss": 1.445096731185913, "eval_runtime": 18.4795, "eval_samples_per_second": 54.114, "eval_steps_per_second": 13.528, "step": 24500 }, { "epoch": 1.8883162833628182, "grad_norm": 3.0785481929779053, "learning_rate": 7.51370023115082e-05, "loss": 1.493, "step": 24550 }, { "epoch": 1.8921621413737406, "grad_norm": 1.2793898582458496, "learning_rate": 7.487728228969173e-05, "loss": 1.3865, "step": 24600 }, { "epoch": 1.8960079993846626, "grad_norm": 1.474137544631958, "learning_rate": 7.461756226787522e-05, "loss": 1.4316, "step": 24650 }, { "epoch": 1.899853857395585, "grad_norm": 1.270415186882019, "learning_rate": 7.435784224605875e-05, "loss": 1.3844, "step": 24700 }, { "epoch": 1.9036997154065072, "grad_norm": 1.3681602478027344, "learning_rate": 7.409812222424228e-05, "loss": 1.4202, "step": 24750 }, { "epoch": 1.9036997154065072, "eval_loss": 1.4705748558044434, "eval_runtime": 18.5933, "eval_samples_per_second": 53.783, "eval_steps_per_second": 13.446, "step": 24750 }, { "epoch": 1.9075455734174294, "grad_norm": 1.028225302696228, "learning_rate": 7.384359660286212e-05, "loss": 1.4658, "step": 24800 }, { "epoch": 1.9113914314283518, "grad_norm": 1.1178765296936035, "learning_rate": 7.358387658104564e-05, "loss": 1.4616, "step": 24850 }, { "epoch": 1.9152372894392737, "grad_norm": 2.1088242530822754, "learning_rate": 7.332415655922915e-05, "loss": 1.4916, "step": 24900 }, { "epoch": 1.9190831474501961, "grad_norm": 0.7096924781799316, "learning_rate": 7.306443653741267e-05, "loss": 1.3687, "step": 24950 }, { "epoch": 1.9229290054611183, "grad_norm": 1.4435713291168213, "learning_rate": 7.28047165155962e-05, "loss": 1.5273, "step": 25000 }, { "epoch": 1.9229290054611183, "eval_loss": 1.4638206958770752, "eval_runtime": 18.4856, "eval_samples_per_second": 54.096, "eval_steps_per_second": 13.524, "step": 25000 }, { "epoch": 1.9267748634720405, "grad_norm": 1.4893878698349, "learning_rate": 7.25449964937797e-05, "loss": 1.5067, "step": 25050 }, { "epoch": 1.930620721482963, "grad_norm": 0.8735935091972351, "learning_rate": 7.228527647196322e-05, "loss": 1.4671, "step": 25100 }, { "epoch": 1.9344665794938851, "grad_norm": 1.6086535453796387, "learning_rate": 7.202555645014675e-05, "loss": 1.4551, "step": 25150 }, { "epoch": 1.9383124375048073, "grad_norm": 0.683675229549408, "learning_rate": 7.176583642833027e-05, "loss": 1.4673, "step": 25200 }, { "epoch": 1.9421582955157297, "grad_norm": 1.9318158626556396, "learning_rate": 7.150611640651378e-05, "loss": 1.4199, "step": 25250 }, { "epoch": 1.9421582955157297, "eval_loss": 1.4574114084243774, "eval_runtime": 18.5222, "eval_samples_per_second": 53.989, "eval_steps_per_second": 13.497, "step": 25250 }, { "epoch": 1.9460041535266517, "grad_norm": 1.9871971607208252, "learning_rate": 7.12463963846973e-05, "loss": 1.5002, "step": 25300 }, { "epoch": 1.949850011537574, "grad_norm": 1.4302830696105957, "learning_rate": 7.098667636288082e-05, "loss": 1.46, "step": 25350 }, { "epoch": 1.9536958695484963, "grad_norm": 1.8389050960540771, "learning_rate": 7.072695634106434e-05, "loss": 1.4025, "step": 25400 }, { "epoch": 1.9575417275594185, "grad_norm": 1.7089191675186157, "learning_rate": 7.046723631924785e-05, "loss": 1.4507, "step": 25450 }, { "epoch": 1.9613875855703409, "grad_norm": 1.3698766231536865, "learning_rate": 7.020751629743138e-05, "loss": 1.4954, "step": 25500 }, { "epoch": 1.9613875855703409, "eval_loss": 1.454710841178894, "eval_runtime": 18.5708, "eval_samples_per_second": 53.848, "eval_steps_per_second": 13.462, "step": 25500 }, { "epoch": 1.9652334435812628, "grad_norm": 1.808030128479004, "learning_rate": 6.994779627561489e-05, "loss": 1.4254, "step": 25550 }, { "epoch": 1.9690793015921852, "grad_norm": 1.634099006652832, "learning_rate": 6.96880762537984e-05, "loss": 1.3325, "step": 25600 }, { "epoch": 1.9729251596031074, "grad_norm": 0.8857108354568481, "learning_rate": 6.942835623198193e-05, "loss": 1.3983, "step": 25650 }, { "epoch": 1.9767710176140296, "grad_norm": 1.6114498376846313, "learning_rate": 6.916863621016545e-05, "loss": 1.3348, "step": 25700 }, { "epoch": 1.980616875624952, "grad_norm": 1.4415462017059326, "learning_rate": 6.890891618834895e-05, "loss": 1.4617, "step": 25750 }, { "epoch": 1.980616875624952, "eval_loss": 1.4457746744155884, "eval_runtime": 17.9567, "eval_samples_per_second": 55.689, "eval_steps_per_second": 13.922, "step": 25750 }, { "epoch": 1.984462733635874, "grad_norm": 1.9614554643630981, "learning_rate": 6.864919616653248e-05, "loss": 1.3813, "step": 25800 }, { "epoch": 1.9883085916467964, "grad_norm": 1.2938437461853027, "learning_rate": 6.8389476144716e-05, "loss": 1.4047, "step": 25850 }, { "epoch": 1.9921544496577186, "grad_norm": 2.1129326820373535, "learning_rate": 6.812975612289952e-05, "loss": 1.4362, "step": 25900 }, { "epoch": 1.9960003076686408, "grad_norm": 0.8634279370307922, "learning_rate": 6.787003610108303e-05, "loss": 1.4805, "step": 25950 }, { "epoch": 1.9998461656795632, "grad_norm": 2.995699405670166, "learning_rate": 6.761031607926656e-05, "loss": 1.5073, "step": 26000 }, { "epoch": 1.9998461656795632, "eval_loss": 1.453719973564148, "eval_runtime": 18.0595, "eval_samples_per_second": 55.373, "eval_steps_per_second": 13.843, "step": 26000 }, { "epoch": 2.003692023690485, "grad_norm": 1.621793508529663, "learning_rate": 6.73557904578864e-05, "loss": 1.369, "step": 26050 }, { "epoch": 2.0075378817014076, "grad_norm": 2.223520278930664, "learning_rate": 6.709607043606992e-05, "loss": 1.4269, "step": 26100 }, { "epoch": 2.01138373971233, "grad_norm": 1.4860827922821045, "learning_rate": 6.683635041425344e-05, "loss": 1.3634, "step": 26150 }, { "epoch": 2.015229597723252, "grad_norm": 2.0796148777008057, "learning_rate": 6.657663039243696e-05, "loss": 1.4233, "step": 26200 }, { "epoch": 2.0190754557341744, "grad_norm": 1.6398444175720215, "learning_rate": 6.631691037062047e-05, "loss": 1.4058, "step": 26250 }, { "epoch": 2.0190754557341744, "eval_loss": 1.4733901023864746, "eval_runtime": 18.0349, "eval_samples_per_second": 55.448, "eval_steps_per_second": 13.862, "step": 26250 }, { "epoch": 2.0229213137450963, "grad_norm": 1.7550077438354492, "learning_rate": 6.605719034880399e-05, "loss": 1.4436, "step": 26300 }, { "epoch": 2.0267671717560187, "grad_norm": 2.3273561000823975, "learning_rate": 6.579747032698751e-05, "loss": 1.356, "step": 26350 }, { "epoch": 2.030613029766941, "grad_norm": 1.1432509422302246, "learning_rate": 6.553775030517103e-05, "loss": 1.4116, "step": 26400 }, { "epoch": 2.034458887777863, "grad_norm": 1.2345376014709473, "learning_rate": 6.527803028335455e-05, "loss": 1.4465, "step": 26450 }, { "epoch": 2.0383047457887855, "grad_norm": 1.485564112663269, "learning_rate": 6.501831026153807e-05, "loss": 1.3896, "step": 26500 }, { "epoch": 2.0383047457887855, "eval_loss": 1.4492217302322388, "eval_runtime": 17.9114, "eval_samples_per_second": 55.83, "eval_steps_per_second": 13.958, "step": 26500 }, { "epoch": 2.042150603799708, "grad_norm": 1.0810669660568237, "learning_rate": 6.475859023972158e-05, "loss": 1.2936, "step": 26550 }, { "epoch": 2.04599646181063, "grad_norm": 1.23382568359375, "learning_rate": 6.44988702179051e-05, "loss": 1.3922, "step": 26600 }, { "epoch": 2.0498423198215523, "grad_norm": 1.8218950033187866, "learning_rate": 6.423915019608862e-05, "loss": 1.4041, "step": 26650 }, { "epoch": 2.0536881778324743, "grad_norm": 0.6482899785041809, "learning_rate": 6.397943017427213e-05, "loss": 1.3609, "step": 26700 }, { "epoch": 2.0575340358433967, "grad_norm": 1.9538156986236572, "learning_rate": 6.371971015245566e-05, "loss": 1.4558, "step": 26750 }, { "epoch": 2.0575340358433967, "eval_loss": 1.454428791999817, "eval_runtime": 18.2441, "eval_samples_per_second": 54.812, "eval_steps_per_second": 13.703, "step": 26750 }, { "epoch": 2.061379893854319, "grad_norm": 1.087234377861023, "learning_rate": 6.345999013063917e-05, "loss": 1.3967, "step": 26800 }, { "epoch": 2.065225751865241, "grad_norm": 1.638533592224121, "learning_rate": 6.32002701088227e-05, "loss": 1.4558, "step": 26850 }, { "epoch": 2.0690716098761635, "grad_norm": 1.4552900791168213, "learning_rate": 6.294055008700621e-05, "loss": 1.4553, "step": 26900 }, { "epoch": 2.0729174678870854, "grad_norm": 1.6330054998397827, "learning_rate": 6.268083006518974e-05, "loss": 1.4977, "step": 26950 }, { "epoch": 2.076763325898008, "grad_norm": 1.497938632965088, "learning_rate": 6.242111004337325e-05, "loss": 1.4449, "step": 27000 }, { "epoch": 2.076763325898008, "eval_loss": 1.4610888957977295, "eval_runtime": 17.9657, "eval_samples_per_second": 55.662, "eval_steps_per_second": 13.915, "step": 27000 }, { "epoch": 2.0806091839089302, "grad_norm": 0.9695401191711426, "learning_rate": 6.216139002155676e-05, "loss": 1.4943, "step": 27050 }, { "epoch": 2.084455041919852, "grad_norm": 1.243717074394226, "learning_rate": 6.190166999974029e-05, "loss": 1.4199, "step": 27100 }, { "epoch": 2.0883008999307746, "grad_norm": 1.6151024103164673, "learning_rate": 6.16419499779238e-05, "loss": 1.4248, "step": 27150 }, { "epoch": 2.0921467579416966, "grad_norm": 1.7448607683181763, "learning_rate": 6.138222995610731e-05, "loss": 1.3944, "step": 27200 }, { "epoch": 2.095992615952619, "grad_norm": 3.0997345447540283, "learning_rate": 6.112250993429084e-05, "loss": 1.4174, "step": 27250 }, { "epoch": 2.095992615952619, "eval_loss": 1.4659229516983032, "eval_runtime": 18.0653, "eval_samples_per_second": 55.355, "eval_steps_per_second": 13.839, "step": 27250 }, { "epoch": 2.0998384739635414, "grad_norm": 1.637845754623413, "learning_rate": 6.086278991247436e-05, "loss": 1.446, "step": 27300 }, { "epoch": 2.1036843319744634, "grad_norm": 1.4263664484024048, "learning_rate": 6.060306989065787e-05, "loss": 0.9793, "step": 27350 }, { "epoch": 2.1075301899853858, "grad_norm": 0.8709418773651123, "learning_rate": 6.034334986884139e-05, "loss": 0.9715, "step": 27400 }, { "epoch": 2.111376047996308, "grad_norm": 0.8483341336250305, "learning_rate": 6.008362984702491e-05, "loss": 0.9163, "step": 27450 }, { "epoch": 2.11522190600723, "grad_norm": 2.070937156677246, "learning_rate": 5.9823909825208425e-05, "loss": 1.0047, "step": 27500 }, { "epoch": 2.11522190600723, "eval_loss": 1.0159448385238647, "eval_runtime": 17.7021, "eval_samples_per_second": 56.49, "eval_steps_per_second": 14.123, "step": 27500 }, { "epoch": 2.1190677640181526, "grad_norm": 0.9638277292251587, "learning_rate": 5.9564189803391944e-05, "loss": 0.9753, "step": 27550 }, { "epoch": 2.1229136220290745, "grad_norm": 1.1322181224822998, "learning_rate": 5.9304469781575464e-05, "loss": 0.948, "step": 27600 }, { "epoch": 2.126759480039997, "grad_norm": 1.144047737121582, "learning_rate": 5.904474975975898e-05, "loss": 1.009, "step": 27650 }, { "epoch": 2.1306053380509193, "grad_norm": 2.713625431060791, "learning_rate": 5.8785029737942496e-05, "loss": 1.0069, "step": 27700 }, { "epoch": 2.1344511960618413, "grad_norm": 1.1035822629928589, "learning_rate": 5.852530971612602e-05, "loss": 1.0013, "step": 27750 }, { "epoch": 2.1344511960618413, "eval_loss": 1.0144418478012085, "eval_runtime": 17.6881, "eval_samples_per_second": 56.535, "eval_steps_per_second": 14.134, "step": 27750 }, { "epoch": 2.1382970540727637, "grad_norm": 1.2658100128173828, "learning_rate": 5.826558969430954e-05, "loss": 1.0185, "step": 27800 }, { "epoch": 2.1421429120836857, "grad_norm": 0.9421238303184509, "learning_rate": 5.800586967249305e-05, "loss": 0.9892, "step": 27850 }, { "epoch": 2.145988770094608, "grad_norm": 0.9409565925598145, "learning_rate": 5.7746149650676575e-05, "loss": 0.9985, "step": 27900 }, { "epoch": 2.1498346281055305, "grad_norm": 1.445890188217163, "learning_rate": 5.7486429628860094e-05, "loss": 1.0147, "step": 27950 }, { "epoch": 2.1536804861164525, "grad_norm": 1.109020709991455, "learning_rate": 5.722670960704361e-05, "loss": 1.0093, "step": 28000 }, { "epoch": 2.1536804861164525, "eval_loss": 1.008616328239441, "eval_runtime": 17.6489, "eval_samples_per_second": 56.661, "eval_steps_per_second": 14.165, "step": 28000 }, { "epoch": 2.157526344127375, "grad_norm": 1.3012685775756836, "learning_rate": 5.696698958522713e-05, "loss": 1.0703, "step": 28050 }, { "epoch": 2.1613722021382973, "grad_norm": 0.8459142446517944, "learning_rate": 5.6707269563410646e-05, "loss": 1.0507, "step": 28100 }, { "epoch": 2.1652180601492192, "grad_norm": 0.6889505982398987, "learning_rate": 5.6447549541594166e-05, "loss": 1.0056, "step": 28150 }, { "epoch": 2.1690639181601417, "grad_norm": 1.2236456871032715, "learning_rate": 5.618782951977768e-05, "loss": 0.988, "step": 28200 }, { "epoch": 2.1729097761710636, "grad_norm": 1.3419203758239746, "learning_rate": 5.59281094979612e-05, "loss": 0.9467, "step": 28250 }, { "epoch": 2.1729097761710636, "eval_loss": 1.0140153169631958, "eval_runtime": 17.7299, "eval_samples_per_second": 56.402, "eval_steps_per_second": 14.1, "step": 28250 }, { "epoch": 2.176755634181986, "grad_norm": 0.8394871354103088, "learning_rate": 5.5668389476144725e-05, "loss": 0.9824, "step": 28300 }, { "epoch": 2.1806014921929084, "grad_norm": 0.7906908392906189, "learning_rate": 5.540866945432823e-05, "loss": 0.9915, "step": 28350 }, { "epoch": 2.1844473502038304, "grad_norm": 1.0852785110473633, "learning_rate": 5.514894943251175e-05, "loss": 0.9473, "step": 28400 }, { "epoch": 2.188293208214753, "grad_norm": 0.9187583923339844, "learning_rate": 5.488922941069528e-05, "loss": 0.9418, "step": 28450 }, { "epoch": 2.192139066225675, "grad_norm": 0.7200838923454285, "learning_rate": 5.4629509388878797e-05, "loss": 0.9776, "step": 28500 }, { "epoch": 2.192139066225675, "eval_loss": 1.0052642822265625, "eval_runtime": 17.5069, "eval_samples_per_second": 57.12, "eval_steps_per_second": 14.28, "step": 28500 }, { "epoch": 2.195984924236597, "grad_norm": 1.00751793384552, "learning_rate": 5.436978936706231e-05, "loss": 0.9479, "step": 28550 }, { "epoch": 2.1998307822475196, "grad_norm": 0.9239784479141235, "learning_rate": 5.411006934524583e-05, "loss": 0.9896, "step": 28600 }, { "epoch": 2.2036766402584416, "grad_norm": 0.9650816917419434, "learning_rate": 5.385034932342935e-05, "loss": 1.0081, "step": 28650 }, { "epoch": 2.207522498269364, "grad_norm": 1.455723524093628, "learning_rate": 5.359062930161286e-05, "loss": 0.9717, "step": 28700 }, { "epoch": 2.211368356280286, "grad_norm": 0.609380304813385, "learning_rate": 5.333090927979638e-05, "loss": 0.9774, "step": 28750 }, { "epoch": 2.211368356280286, "eval_loss": 1.0018259286880493, "eval_runtime": 17.3475, "eval_samples_per_second": 57.645, "eval_steps_per_second": 14.411, "step": 28750 }, { "epoch": 2.2152142142912084, "grad_norm": 0.9981700778007507, "learning_rate": 5.30711892579799e-05, "loss": 0.9873, "step": 28800 }, { "epoch": 2.2190600723021308, "grad_norm": 0.5452422499656677, "learning_rate": 5.2811469236163414e-05, "loss": 1.0078, "step": 28850 }, { "epoch": 2.2229059303130527, "grad_norm": 1.4739840030670166, "learning_rate": 5.255174921434693e-05, "loss": 0.9739, "step": 28900 }, { "epoch": 2.226751788323975, "grad_norm": 0.9250359535217285, "learning_rate": 5.229202919253045e-05, "loss": 0.9723, "step": 28950 }, { "epoch": 2.230597646334897, "grad_norm": 1.0843122005462646, "learning_rate": 5.203230917071398e-05, "loss": 0.989, "step": 29000 }, { "epoch": 2.230597646334897, "eval_loss": 0.9999537467956543, "eval_runtime": 17.2417, "eval_samples_per_second": 57.999, "eval_steps_per_second": 14.5, "step": 29000 }, { "epoch": 2.2344435043458195, "grad_norm": 0.5559306740760803, "learning_rate": 5.1772589148897485e-05, "loss": 0.9749, "step": 29050 }, { "epoch": 2.238289362356742, "grad_norm": 0.6720598936080933, "learning_rate": 5.151286912708101e-05, "loss": 0.9662, "step": 29100 }, { "epoch": 2.242135220367664, "grad_norm": 1.117200493812561, "learning_rate": 5.125314910526453e-05, "loss": 0.9552, "step": 29150 }, { "epoch": 2.2459810783785863, "grad_norm": 0.6818645000457764, "learning_rate": 5.0993429083448044e-05, "loss": 0.9032, "step": 29200 }, { "epoch": 2.2498269363895087, "grad_norm": 0.9796412587165833, "learning_rate": 5.0733709061631564e-05, "loss": 0.9429, "step": 29250 }, { "epoch": 2.2498269363895087, "eval_loss": 1.0121312141418457, "eval_runtime": 17.3211, "eval_samples_per_second": 57.733, "eval_steps_per_second": 14.433, "step": 29250 }, { "epoch": 2.2536727944004307, "grad_norm": 1.021713137626648, "learning_rate": 5.047398903981508e-05, "loss": 0.979, "step": 29300 }, { "epoch": 2.257518652411353, "grad_norm": 1.1321250200271606, "learning_rate": 5.02142690179986e-05, "loss": 1.0327, "step": 29350 }, { "epoch": 2.261364510422275, "grad_norm": 0.7670277953147888, "learning_rate": 4.9954548996182116e-05, "loss": 0.9668, "step": 29400 }, { "epoch": 2.2652103684331975, "grad_norm": 1.447698712348938, "learning_rate": 4.9694828974365635e-05, "loss": 0.9288, "step": 29450 }, { "epoch": 2.26905622644412, "grad_norm": 1.0438776016235352, "learning_rate": 4.9435108952549155e-05, "loss": 1.0154, "step": 29500 }, { "epoch": 2.26905622644412, "eval_loss": 1.0100510120391846, "eval_runtime": 17.4293, "eval_samples_per_second": 57.375, "eval_steps_per_second": 14.344, "step": 29500 }, { "epoch": 2.272902084455042, "grad_norm": 0.6814424991607666, "learning_rate": 4.9175388930732675e-05, "loss": 0.9563, "step": 29550 }, { "epoch": 2.2767479424659642, "grad_norm": 0.85086989402771, "learning_rate": 4.891566890891619e-05, "loss": 0.9778, "step": 29600 }, { "epoch": 2.280593800476886, "grad_norm": 1.1364527940750122, "learning_rate": 4.8655948887099714e-05, "loss": 0.996, "step": 29650 }, { "epoch": 2.2844396584878086, "grad_norm": 0.7810873985290527, "learning_rate": 4.839622886528323e-05, "loss": 1.0141, "step": 29700 }, { "epoch": 2.288285516498731, "grad_norm": 1.02638578414917, "learning_rate": 4.8136508843466746e-05, "loss": 0.97, "step": 29750 }, { "epoch": 2.288285516498731, "eval_loss": 1.0048705339431763, "eval_runtime": 17.487, "eval_samples_per_second": 57.185, "eval_steps_per_second": 14.296, "step": 29750 }, { "epoch": 2.292131374509653, "grad_norm": 1.334876298904419, "learning_rate": 4.7876788821650266e-05, "loss": 0.9533, "step": 29800 }, { "epoch": 2.2959772325205754, "grad_norm": 1.3690305948257446, "learning_rate": 4.761706879983378e-05, "loss": 0.992, "step": 29850 }, { "epoch": 2.299823090531498, "grad_norm": 1.525981068611145, "learning_rate": 4.7357348778017305e-05, "loss": 0.949, "step": 29900 }, { "epoch": 2.3036689485424198, "grad_norm": 0.6612236499786377, "learning_rate": 4.709762875620082e-05, "loss": 1.0112, "step": 29950 }, { "epoch": 2.307514806553342, "grad_norm": 0.8452871441841125, "learning_rate": 4.683790873438433e-05, "loss": 0.9858, "step": 30000 }, { "epoch": 2.307514806553342, "eval_loss": 0.9945911169052124, "eval_runtime": 17.2501, "eval_samples_per_second": 57.971, "eval_steps_per_second": 14.493, "step": 30000 }, { "epoch": 2.311360664564264, "grad_norm": 1.3450604677200317, "learning_rate": 4.657818871256786e-05, "loss": 0.973, "step": 30050 }, { "epoch": 2.3152065225751866, "grad_norm": 1.1462957859039307, "learning_rate": 4.631846869075137e-05, "loss": 0.9789, "step": 30100 }, { "epoch": 2.3190523805861085, "grad_norm": 1.0691301822662354, "learning_rate": 4.605874866893489e-05, "loss": 0.9937, "step": 30150 }, { "epoch": 2.322898238597031, "grad_norm": 1.527723789215088, "learning_rate": 4.579902864711841e-05, "loss": 0.9276, "step": 30200 }, { "epoch": 2.3267440966079533, "grad_norm": 0.7396986484527588, "learning_rate": 4.553930862530192e-05, "loss": 0.9648, "step": 30250 }, { "epoch": 2.3267440966079533, "eval_loss": 1.0029717683792114, "eval_runtime": 17.2662, "eval_samples_per_second": 57.917, "eval_steps_per_second": 14.479, "step": 30250 }, { "epoch": 2.3305899546188753, "grad_norm": 0.9694539308547974, "learning_rate": 4.527958860348545e-05, "loss": 0.9378, "step": 30300 }, { "epoch": 2.3344358126297977, "grad_norm": 1.161028504371643, "learning_rate": 4.501986858166896e-05, "loss": 0.9807, "step": 30350 }, { "epoch": 2.33828167064072, "grad_norm": 1.6975845098495483, "learning_rate": 4.476014855985248e-05, "loss": 0.9809, "step": 30400 }, { "epoch": 2.342127528651642, "grad_norm": 1.3228962421417236, "learning_rate": 4.4500428538036e-05, "loss": 1.0339, "step": 30450 }, { "epoch": 2.3459733866625645, "grad_norm": 1.0168397426605225, "learning_rate": 4.424070851621952e-05, "loss": 0.9544, "step": 30500 }, { "epoch": 2.3459733866625645, "eval_loss": 1.002646565437317, "eval_runtime": 17.3363, "eval_samples_per_second": 57.682, "eval_steps_per_second": 14.421, "step": 30500 }, { "epoch": 2.3498192446734865, "grad_norm": 0.9951680302619934, "learning_rate": 4.398098849440303e-05, "loss": 1.0062, "step": 30550 }, { "epoch": 2.353665102684409, "grad_norm": 0.5752933025360107, "learning_rate": 4.372126847258655e-05, "loss": 0.9649, "step": 30600 }, { "epoch": 2.3575109606953313, "grad_norm": 1.0051320791244507, "learning_rate": 4.346154845077007e-05, "loss": 0.9958, "step": 30650 }, { "epoch": 2.3613568187062532, "grad_norm": 0.7760717868804932, "learning_rate": 4.320182842895359e-05, "loss": 1.0125, "step": 30700 }, { "epoch": 2.3652026767171757, "grad_norm": 0.852301836013794, "learning_rate": 4.294210840713711e-05, "loss": 0.9523, "step": 30750 }, { "epoch": 2.3652026767171757, "eval_loss": 1.003655195236206, "eval_runtime": 17.3241, "eval_samples_per_second": 57.723, "eval_steps_per_second": 14.431, "step": 30750 }, { "epoch": 2.3690485347280976, "grad_norm": 0.9062100648880005, "learning_rate": 4.2682388385320624e-05, "loss": 0.9735, "step": 30800 }, { "epoch": 2.37289439273902, "grad_norm": 1.309615969657898, "learning_rate": 4.2427862763940476e-05, "loss": 0.9668, "step": 30850 }, { "epoch": 2.3767402507499424, "grad_norm": 1.0907591581344604, "learning_rate": 4.2168142742123995e-05, "loss": 0.9502, "step": 30900 }, { "epoch": 2.3805861087608644, "grad_norm": 1.0946288108825684, "learning_rate": 4.190842272030751e-05, "loss": 0.9545, "step": 30950 }, { "epoch": 2.384431966771787, "grad_norm": 1.225540280342102, "learning_rate": 4.164870269849103e-05, "loss": 0.9635, "step": 31000 }, { "epoch": 2.384431966771787, "eval_loss": 1.0031681060791016, "eval_runtime": 17.2062, "eval_samples_per_second": 58.119, "eval_steps_per_second": 14.53, "step": 31000 }, { "epoch": 2.3882778247827092, "grad_norm": 0.8717153668403625, "learning_rate": 4.138898267667455e-05, "loss": 1.0038, "step": 31050 }, { "epoch": 2.392123682793631, "grad_norm": 1.007270097732544, "learning_rate": 4.112926265485807e-05, "loss": 0.9858, "step": 31100 }, { "epoch": 2.3959695408045536, "grad_norm": 1.9409807920455933, "learning_rate": 4.0869542633041587e-05, "loss": 1.004, "step": 31150 }, { "epoch": 2.3998153988154756, "grad_norm": 0.6027572154998779, "learning_rate": 4.06098226112251e-05, "loss": 0.9958, "step": 31200 }, { "epoch": 2.403661256826398, "grad_norm": 0.8274515867233276, "learning_rate": 4.035010258940862e-05, "loss": 0.9852, "step": 31250 }, { "epoch": 2.403661256826398, "eval_loss": 1.0001976490020752, "eval_runtime": 17.2233, "eval_samples_per_second": 58.061, "eval_steps_per_second": 14.515, "step": 31250 }, { "epoch": 2.4075071148373204, "grad_norm": 0.7238942384719849, "learning_rate": 4.009038256759214e-05, "loss": 1.0062, "step": 31300 }, { "epoch": 2.4113529728482423, "grad_norm": 0.8912849426269531, "learning_rate": 3.983066254577565e-05, "loss": 0.9802, "step": 31350 }, { "epoch": 2.4151988308591648, "grad_norm": 1.1922829151153564, "learning_rate": 3.957094252395918e-05, "loss": 1.0317, "step": 31400 }, { "epoch": 2.4190446888700867, "grad_norm": 1.3773999214172363, "learning_rate": 3.931122250214269e-05, "loss": 0.9773, "step": 31450 }, { "epoch": 2.422890546881009, "grad_norm": 1.0747745037078857, "learning_rate": 3.905150248032621e-05, "loss": 1.0208, "step": 31500 }, { "epoch": 2.422890546881009, "eval_loss": 1.0009056329727173, "eval_runtime": 17.1069, "eval_samples_per_second": 58.456, "eval_steps_per_second": 14.614, "step": 31500 }, { "epoch": 2.4267364048919315, "grad_norm": 1.3409994840621948, "learning_rate": 3.879178245850973e-05, "loss": 0.9546, "step": 31550 }, { "epoch": 2.4305822629028535, "grad_norm": 1.482633352279663, "learning_rate": 3.853206243669324e-05, "loss": 1.0155, "step": 31600 }, { "epoch": 2.434428120913776, "grad_norm": 1.0458152294158936, "learning_rate": 3.827234241487677e-05, "loss": 0.989, "step": 31650 }, { "epoch": 2.4382739789246983, "grad_norm": 0.9805555939674377, "learning_rate": 3.801262239306028e-05, "loss": 0.9125, "step": 31700 }, { "epoch": 2.4421198369356203, "grad_norm": 1.5680670738220215, "learning_rate": 3.77529023712438e-05, "loss": 0.9713, "step": 31750 }, { "epoch": 2.4421198369356203, "eval_loss": 0.9981379508972168, "eval_runtime": 17.3284, "eval_samples_per_second": 57.709, "eval_steps_per_second": 14.427, "step": 31750 }, { "epoch": 2.4459656949465427, "grad_norm": 1.0316798686981201, "learning_rate": 3.749318234942732e-05, "loss": 0.9689, "step": 31800 }, { "epoch": 2.4498115529574647, "grad_norm": 0.6630721092224121, "learning_rate": 3.7233462327610834e-05, "loss": 0.9772, "step": 31850 }, { "epoch": 2.453657410968387, "grad_norm": 1.1662702560424805, "learning_rate": 3.6973742305794354e-05, "loss": 1.0165, "step": 31900 }, { "epoch": 2.457503268979309, "grad_norm": 1.0451244115829468, "learning_rate": 3.671402228397787e-05, "loss": 1.0233, "step": 31950 }, { "epoch": 2.4613491269902315, "grad_norm": 0.9077771306037903, "learning_rate": 3.645430226216139e-05, "loss": 0.9797, "step": 32000 }, { "epoch": 2.4613491269902315, "eval_loss": 0.9961766004562378, "eval_runtime": 17.2799, "eval_samples_per_second": 57.871, "eval_steps_per_second": 14.468, "step": 32000 }, { "epoch": 2.465194985001154, "grad_norm": 0.999718964099884, "learning_rate": 3.619458224034491e-05, "loss": 0.9725, "step": 32050 }, { "epoch": 2.469040843012076, "grad_norm": 2.740297794342041, "learning_rate": 3.5934862218528425e-05, "loss": 0.9838, "step": 32100 }, { "epoch": 2.4728867010229982, "grad_norm": 1.199425458908081, "learning_rate": 3.5675142196711945e-05, "loss": 0.9807, "step": 32150 }, { "epoch": 2.4767325590339206, "grad_norm": 0.7113758325576782, "learning_rate": 3.5415422174895465e-05, "loss": 1.0046, "step": 32200 }, { "epoch": 2.4805784170448426, "grad_norm": 0.9929390549659729, "learning_rate": 3.5155702153078984e-05, "loss": 0.9692, "step": 32250 }, { "epoch": 2.4805784170448426, "eval_loss": 1.002519965171814, "eval_runtime": 17.3042, "eval_samples_per_second": 57.789, "eval_steps_per_second": 14.447, "step": 32250 }, { "epoch": 2.484424275055765, "grad_norm": 0.8170703649520874, "learning_rate": 3.48959821312625e-05, "loss": 0.982, "step": 32300 }, { "epoch": 2.488270133066687, "grad_norm": 0.8909692168235779, "learning_rate": 3.463626210944602e-05, "loss": 0.9953, "step": 32350 }, { "epoch": 2.4921159910776094, "grad_norm": 1.806539535522461, "learning_rate": 3.4376542087629536e-05, "loss": 1.0069, "step": 32400 }, { "epoch": 2.495961849088532, "grad_norm": 1.4509518146514893, "learning_rate": 3.4116822065813056e-05, "loss": 0.9642, "step": 32450 }, { "epoch": 2.4998077070994538, "grad_norm": 1.0312175750732422, "learning_rate": 3.3857102043996575e-05, "loss": 0.9921, "step": 32500 }, { "epoch": 2.4998077070994538, "eval_loss": 0.9993879199028015, "eval_runtime": 17.3389, "eval_samples_per_second": 57.674, "eval_steps_per_second": 14.418, "step": 32500 }, { "epoch": 2.503653565110376, "grad_norm": 0.867513120174408, "learning_rate": 3.359738202218009e-05, "loss": 0.994, "step": 32550 }, { "epoch": 2.507499423121298, "grad_norm": 1.7425885200500488, "learning_rate": 3.3337662000363615e-05, "loss": 1.0068, "step": 32600 }, { "epoch": 2.5113452811322206, "grad_norm": 0.9053608775138855, "learning_rate": 3.307794197854713e-05, "loss": 0.9704, "step": 32650 }, { "epoch": 2.515191139143143, "grad_norm": 1.0533051490783691, "learning_rate": 3.281822195673064e-05, "loss": 0.9506, "step": 32700 }, { "epoch": 2.519036997154065, "grad_norm": 1.2495230436325073, "learning_rate": 3.255850193491417e-05, "loss": 0.9936, "step": 32750 }, { "epoch": 2.519036997154065, "eval_loss": 0.9990929961204529, "eval_runtime": 17.239, "eval_samples_per_second": 58.008, "eval_steps_per_second": 14.502, "step": 32750 }, { "epoch": 2.5228828551649873, "grad_norm": 0.8339760303497314, "learning_rate": 3.229878191309768e-05, "loss": 1.0008, "step": 32800 }, { "epoch": 2.5267287131759097, "grad_norm": 1.2839399576187134, "learning_rate": 3.20390618912812e-05, "loss": 1.0109, "step": 32850 }, { "epoch": 2.5305745711868317, "grad_norm": 1.2146100997924805, "learning_rate": 3.177934186946472e-05, "loss": 0.9834, "step": 32900 }, { "epoch": 2.534420429197754, "grad_norm": 0.7952923774719238, "learning_rate": 3.151962184764823e-05, "loss": 1.0013, "step": 32950 }, { "epoch": 2.5382662872086765, "grad_norm": 1.67001211643219, "learning_rate": 3.125990182583176e-05, "loss": 0.9797, "step": 33000 }, { "epoch": 2.5382662872086765, "eval_loss": 1.001037359237671, "eval_runtime": 17.347, "eval_samples_per_second": 57.647, "eval_steps_per_second": 14.412, "step": 33000 }, { "epoch": 2.5421121452195985, "grad_norm": 0.7536977529525757, "learning_rate": 3.100018180401527e-05, "loss": 0.9818, "step": 33050 }, { "epoch": 2.5459580032305205, "grad_norm": 1.3777302503585815, "learning_rate": 3.074046178219879e-05, "loss": 0.953, "step": 33100 }, { "epoch": 2.549803861241443, "grad_norm": 0.9286239743232727, "learning_rate": 3.048074176038231e-05, "loss": 1.028, "step": 33150 }, { "epoch": 2.5536497192523653, "grad_norm": 0.6741893291473389, "learning_rate": 3.022102173856583e-05, "loss": 0.9693, "step": 33200 }, { "epoch": 2.5574955772632872, "grad_norm": 1.3020586967468262, "learning_rate": 2.9961301716749346e-05, "loss": 0.9673, "step": 33250 }, { "epoch": 2.5574955772632872, "eval_loss": 1.0041394233703613, "eval_runtime": 17.32, "eval_samples_per_second": 57.737, "eval_steps_per_second": 14.434, "step": 33250 }, { "epoch": 2.5613414352742097, "grad_norm": 1.4816234111785889, "learning_rate": 2.9701581694932862e-05, "loss": 0.9964, "step": 33300 }, { "epoch": 2.565187293285132, "grad_norm": 0.8813285827636719, "learning_rate": 2.9441861673116382e-05, "loss": 0.966, "step": 33350 }, { "epoch": 2.569033151296054, "grad_norm": 1.0042293071746826, "learning_rate": 2.9182141651299898e-05, "loss": 0.9254, "step": 33400 }, { "epoch": 2.5728790093069764, "grad_norm": 1.5775707960128784, "learning_rate": 2.892242162948342e-05, "loss": 1.0215, "step": 33450 }, { "epoch": 2.576724867317899, "grad_norm": 1.0266311168670654, "learning_rate": 2.8662701607666937e-05, "loss": 0.9323, "step": 33500 }, { "epoch": 2.576724867317899, "eval_loss": 1.0002070665359497, "eval_runtime": 17.2317, "eval_samples_per_second": 58.033, "eval_steps_per_second": 14.508, "step": 33500 }, { "epoch": 2.580570725328821, "grad_norm": 0.5430648922920227, "learning_rate": 2.8402981585850453e-05, "loss": 0.9658, "step": 33550 }, { "epoch": 2.5844165833397432, "grad_norm": 0.6678454279899597, "learning_rate": 2.8143261564033973e-05, "loss": 1.0292, "step": 33600 }, { "epoch": 2.588262441350665, "grad_norm": 0.7208724021911621, "learning_rate": 2.788354154221749e-05, "loss": 0.9505, "step": 33650 }, { "epoch": 2.5921082993615876, "grad_norm": 1.2248526811599731, "learning_rate": 2.7623821520401012e-05, "loss": 0.9728, "step": 33700 }, { "epoch": 2.5959541573725096, "grad_norm": 1.0026588439941406, "learning_rate": 2.7364101498584525e-05, "loss": 0.9783, "step": 33750 }, { "epoch": 2.5959541573725096, "eval_loss": 0.9881900548934937, "eval_runtime": 17.2742, "eval_samples_per_second": 57.89, "eval_steps_per_second": 14.472, "step": 33750 }, { "epoch": 2.599800015383432, "grad_norm": 0.9579987525939941, "learning_rate": 2.710438147676804e-05, "loss": 0.9882, "step": 33800 }, { "epoch": 2.6036458733943544, "grad_norm": 1.0152076482772827, "learning_rate": 2.6844661454951564e-05, "loss": 1.041, "step": 33850 }, { "epoch": 2.6074917314052763, "grad_norm": 1.1370351314544678, "learning_rate": 2.658494143313508e-05, "loss": 0.9518, "step": 33900 }, { "epoch": 2.6113375894161988, "grad_norm": 0.9851937890052795, "learning_rate": 2.63252214113186e-05, "loss": 0.9125, "step": 33950 }, { "epoch": 2.615183447427121, "grad_norm": 0.8480270504951477, "learning_rate": 2.6065501389502116e-05, "loss": 0.9736, "step": 34000 }, { "epoch": 2.615183447427121, "eval_loss": 0.987713634967804, "eval_runtime": 17.2436, "eval_samples_per_second": 57.993, "eval_steps_per_second": 14.498, "step": 34000 }, { "epoch": 2.619029305438043, "grad_norm": 0.5307362079620361, "learning_rate": 2.580578136768564e-05, "loss": 1.0031, "step": 34050 }, { "epoch": 2.6228751634489655, "grad_norm": 1.1112557649612427, "learning_rate": 2.5546061345869156e-05, "loss": 0.9928, "step": 34100 }, { "epoch": 2.626721021459888, "grad_norm": 0.646759569644928, "learning_rate": 2.5286341324052672e-05, "loss": 1.0173, "step": 34150 }, { "epoch": 2.63056687947081, "grad_norm": 1.0114878416061401, "learning_rate": 2.502662130223619e-05, "loss": 0.9765, "step": 34200 }, { "epoch": 2.6344127374817323, "grad_norm": 0.8782021403312683, "learning_rate": 2.4766901280419708e-05, "loss": 0.9926, "step": 34250 }, { "epoch": 2.6344127374817323, "eval_loss": 0.9932020306587219, "eval_runtime": 17.1921, "eval_samples_per_second": 58.166, "eval_steps_per_second": 14.542, "step": 34250 }, { "epoch": 2.6382585954926543, "grad_norm": 1.0792268514633179, "learning_rate": 2.4507181258603227e-05, "loss": 1.0541, "step": 34300 }, { "epoch": 2.6421044535035767, "grad_norm": 0.9647793173789978, "learning_rate": 2.4247461236786744e-05, "loss": 0.9779, "step": 34350 }, { "epoch": 2.6459503115144987, "grad_norm": 1.0052498579025269, "learning_rate": 2.3987741214970263e-05, "loss": 0.9798, "step": 34400 }, { "epoch": 2.649796169525421, "grad_norm": 0.8122023344039917, "learning_rate": 2.3728021193153783e-05, "loss": 1.0205, "step": 34450 }, { "epoch": 2.6536420275363435, "grad_norm": 1.452087163925171, "learning_rate": 2.3468301171337302e-05, "loss": 0.9745, "step": 34500 }, { "epoch": 2.6536420275363435, "eval_loss": 0.9959968328475952, "eval_runtime": 17.2025, "eval_samples_per_second": 58.131, "eval_steps_per_second": 14.533, "step": 34500 }, { "epoch": 2.6574878855472654, "grad_norm": 1.3329054117202759, "learning_rate": 2.3208581149520815e-05, "loss": 1.0035, "step": 34550 }, { "epoch": 2.661333743558188, "grad_norm": 0.8142715096473694, "learning_rate": 2.2948861127704335e-05, "loss": 0.9783, "step": 34600 }, { "epoch": 2.6651796015691103, "grad_norm": 0.5424798130989075, "learning_rate": 2.2689141105887855e-05, "loss": 0.9754, "step": 34650 }, { "epoch": 2.6690254595800322, "grad_norm": 0.8890462517738342, "learning_rate": 2.2429421084071374e-05, "loss": 0.9707, "step": 34700 }, { "epoch": 2.6728713175909546, "grad_norm": 1.0329838991165161, "learning_rate": 2.216970106225489e-05, "loss": 0.9845, "step": 34750 }, { "epoch": 2.6728713175909546, "eval_loss": 0.9899721741676331, "eval_runtime": 17.3008, "eval_samples_per_second": 57.801, "eval_steps_per_second": 14.45, "step": 34750 }, { "epoch": 2.676717175601877, "grad_norm": 1.0697598457336426, "learning_rate": 2.190998104043841e-05, "loss": 1.0167, "step": 34800 }, { "epoch": 2.680563033612799, "grad_norm": 0.8134399056434631, "learning_rate": 2.1650261018621926e-05, "loss": 0.9001, "step": 34850 }, { "epoch": 2.684408891623721, "grad_norm": 1.0405962467193604, "learning_rate": 2.1390540996805446e-05, "loss": 1.0184, "step": 34900 }, { "epoch": 2.6882547496346434, "grad_norm": 1.3229318857192993, "learning_rate": 2.1130820974988962e-05, "loss": 0.9556, "step": 34950 }, { "epoch": 2.692100607645566, "grad_norm": 1.2907413244247437, "learning_rate": 2.087110095317248e-05, "loss": 0.9355, "step": 35000 }, { "epoch": 2.692100607645566, "eval_loss": 0.9918172359466553, "eval_runtime": 17.2562, "eval_samples_per_second": 57.95, "eval_steps_per_second": 14.488, "step": 35000 }, { "epoch": 2.6959464656564878, "grad_norm": 1.3044216632843018, "learning_rate": 2.0611380931356e-05, "loss": 0.9787, "step": 35050 }, { "epoch": 2.69979232366741, "grad_norm": 1.4273097515106201, "learning_rate": 2.0351660909539518e-05, "loss": 0.9775, "step": 35100 }, { "epoch": 2.7036381816783326, "grad_norm": 1.6097638607025146, "learning_rate": 2.0091940887723034e-05, "loss": 0.9926, "step": 35150 }, { "epoch": 2.7074840396892546, "grad_norm": 1.4893895387649536, "learning_rate": 1.9832220865906553e-05, "loss": 1.0435, "step": 35200 }, { "epoch": 2.711329897700177, "grad_norm": 1.1620233058929443, "learning_rate": 1.9572500844090073e-05, "loss": 0.9525, "step": 35250 }, { "epoch": 2.711329897700177, "eval_loss": 0.9926208257675171, "eval_runtime": 17.3977, "eval_samples_per_second": 57.479, "eval_steps_per_second": 14.37, "step": 35250 }, { "epoch": 2.7151757557110994, "grad_norm": 1.2087517976760864, "learning_rate": 1.931278082227359e-05, "loss": 1.0005, "step": 35300 }, { "epoch": 2.7190216137220213, "grad_norm": 0.8885460495948792, "learning_rate": 1.9058255200893437e-05, "loss": 0.9911, "step": 35350 }, { "epoch": 2.7228674717329437, "grad_norm": 0.9543077945709229, "learning_rate": 1.8798535179076957e-05, "loss": 1.0437, "step": 35400 }, { "epoch": 2.7267133297438657, "grad_norm": 1.2362306118011475, "learning_rate": 1.8538815157260476e-05, "loss": 0.9766, "step": 35450 }, { "epoch": 2.730559187754788, "grad_norm": 1.1946227550506592, "learning_rate": 1.8279095135443993e-05, "loss": 1.0333, "step": 35500 }, { "epoch": 2.730559187754788, "eval_loss": 0.9953948855400085, "eval_runtime": 17.1888, "eval_samples_per_second": 58.178, "eval_steps_per_second": 14.544, "step": 35500 }, { "epoch": 2.73440504576571, "grad_norm": 0.66939777135849, "learning_rate": 1.801937511362751e-05, "loss": 0.8892, "step": 35550 }, { "epoch": 2.7382509037766325, "grad_norm": 1.0852998495101929, "learning_rate": 1.775965509181103e-05, "loss": 0.9502, "step": 35600 }, { "epoch": 2.742096761787555, "grad_norm": 0.7603423595428467, "learning_rate": 1.7499935069994548e-05, "loss": 1.0034, "step": 35650 }, { "epoch": 2.745942619798477, "grad_norm": 0.4625702202320099, "learning_rate": 1.7240215048178064e-05, "loss": 1.0123, "step": 35700 }, { "epoch": 2.7497884778093993, "grad_norm": 1.4387953281402588, "learning_rate": 1.6980495026361584e-05, "loss": 0.9704, "step": 35750 }, { "epoch": 2.7497884778093993, "eval_loss": 0.9952225685119629, "eval_runtime": 17.4458, "eval_samples_per_second": 57.32, "eval_steps_per_second": 14.33, "step": 35750 }, { "epoch": 2.7536343358203217, "grad_norm": 0.6899126172065735, "learning_rate": 1.67207750045451e-05, "loss": 0.9627, "step": 35800 }, { "epoch": 2.7574801938312437, "grad_norm": 1.0329424142837524, "learning_rate": 1.646105498272862e-05, "loss": 0.9207, "step": 35850 }, { "epoch": 2.761326051842166, "grad_norm": 1.1055504083633423, "learning_rate": 1.6201334960912136e-05, "loss": 0.9834, "step": 35900 }, { "epoch": 2.7651719098530885, "grad_norm": 0.7458188533782959, "learning_rate": 1.5941614939095655e-05, "loss": 1.003, "step": 35950 }, { "epoch": 2.7690177678640104, "grad_norm": 1.112021803855896, "learning_rate": 1.5681894917279175e-05, "loss": 1.0001, "step": 36000 }, { "epoch": 2.7690177678640104, "eval_loss": 0.9910063147544861, "eval_runtime": 17.2718, "eval_samples_per_second": 57.898, "eval_steps_per_second": 14.474, "step": 36000 }, { "epoch": 2.772863625874933, "grad_norm": 0.5958703756332397, "learning_rate": 1.5422174895462695e-05, "loss": 0.9503, "step": 36050 }, { "epoch": 2.776709483885855, "grad_norm": 1.1087392568588257, "learning_rate": 1.516245487364621e-05, "loss": 0.9725, "step": 36100 }, { "epoch": 2.780555341896777, "grad_norm": 1.2012007236480713, "learning_rate": 1.4902734851829727e-05, "loss": 1.0098, "step": 36150 }, { "epoch": 2.784401199907699, "grad_norm": 0.9333285093307495, "learning_rate": 1.4643014830013247e-05, "loss": 0.9764, "step": 36200 }, { "epoch": 2.7882470579186216, "grad_norm": 0.9407594203948975, "learning_rate": 1.4383294808196765e-05, "loss": 0.9404, "step": 36250 }, { "epoch": 2.7882470579186216, "eval_loss": 0.9894633889198303, "eval_runtime": 17.3239, "eval_samples_per_second": 57.724, "eval_steps_per_second": 14.431, "step": 36250 }, { "epoch": 2.792092915929544, "grad_norm": 0.9794307947158813, "learning_rate": 1.4123574786380284e-05, "loss": 0.9445, "step": 36300 }, { "epoch": 2.795938773940466, "grad_norm": 1.1447358131408691, "learning_rate": 1.3863854764563799e-05, "loss": 1.0056, "step": 36350 }, { "epoch": 2.7997846319513884, "grad_norm": 0.8741857409477234, "learning_rate": 1.3604134742747318e-05, "loss": 0.9511, "step": 36400 }, { "epoch": 2.803630489962311, "grad_norm": 1.0769715309143066, "learning_rate": 1.3344414720930836e-05, "loss": 0.9789, "step": 36450 }, { "epoch": 2.8074763479732328, "grad_norm": 1.1045129299163818, "learning_rate": 1.3084694699114356e-05, "loss": 0.9958, "step": 36500 }, { "epoch": 2.8074763479732328, "eval_loss": 0.9914500117301941, "eval_runtime": 17.3793, "eval_samples_per_second": 57.54, "eval_steps_per_second": 14.385, "step": 36500 }, { "epoch": 2.811322205984155, "grad_norm": 0.9856983423233032, "learning_rate": 1.2824974677297874e-05, "loss": 0.9923, "step": 36550 }, { "epoch": 2.8151680639950776, "grad_norm": 1.2112038135528564, "learning_rate": 1.2565254655481392e-05, "loss": 0.9388, "step": 36600 }, { "epoch": 2.8190139220059995, "grad_norm": 1.1824342012405396, "learning_rate": 1.230553463366491e-05, "loss": 0.984, "step": 36650 }, { "epoch": 2.8228597800169215, "grad_norm": 1.3278725147247314, "learning_rate": 1.204581461184843e-05, "loss": 0.9835, "step": 36700 }, { "epoch": 2.826705638027844, "grad_norm": 1.4297553300857544, "learning_rate": 1.1786094590031946e-05, "loss": 0.9999, "step": 36750 }, { "epoch": 2.826705638027844, "eval_loss": 0.9847651720046997, "eval_runtime": 17.5097, "eval_samples_per_second": 57.111, "eval_steps_per_second": 14.278, "step": 36750 }, { "epoch": 2.8305514960387663, "grad_norm": 1.1944117546081543, "learning_rate": 1.1526374568215465e-05, "loss": 0.9758, "step": 36800 }, { "epoch": 2.8343973540496883, "grad_norm": 1.1629287004470825, "learning_rate": 1.1266654546398983e-05, "loss": 0.988, "step": 36850 }, { "epoch": 2.8382432120606107, "grad_norm": 1.598382592201233, "learning_rate": 1.1006934524582501e-05, "loss": 0.9679, "step": 36900 }, { "epoch": 2.842089070071533, "grad_norm": 0.9534172415733337, "learning_rate": 1.0747214502766019e-05, "loss": 0.9921, "step": 36950 }, { "epoch": 2.845934928082455, "grad_norm": 1.5716655254364014, "learning_rate": 1.0487494480949537e-05, "loss": 0.9898, "step": 37000 }, { "epoch": 2.845934928082455, "eval_loss": 0.9878412485122681, "eval_runtime": 17.2808, "eval_samples_per_second": 57.868, "eval_steps_per_second": 14.467, "step": 37000 }, { "epoch": 2.8497807860933775, "grad_norm": 1.3013421297073364, "learning_rate": 1.0227774459133055e-05, "loss": 0.9717, "step": 37050 }, { "epoch": 2.8536266441043, "grad_norm": 1.263071060180664, "learning_rate": 9.968054437316573e-06, "loss": 1.0259, "step": 37100 }, { "epoch": 2.857472502115222, "grad_norm": 1.1513851881027222, "learning_rate": 9.70833441550009e-06, "loss": 1.0015, "step": 37150 }, { "epoch": 2.8613183601261443, "grad_norm": 0.7431422472000122, "learning_rate": 9.448614393683609e-06, "loss": 0.999, "step": 37200 }, { "epoch": 2.8651642181370662, "grad_norm": 0.6744217872619629, "learning_rate": 9.188894371867128e-06, "loss": 0.9285, "step": 37250 }, { "epoch": 2.8651642181370662, "eval_loss": 0.994976282119751, "eval_runtime": 17.3294, "eval_samples_per_second": 57.705, "eval_steps_per_second": 14.426, "step": 37250 }, { "epoch": 2.8690100761479886, "grad_norm": 1.2962367534637451, "learning_rate": 8.929174350050646e-06, "loss": 0.9433, "step": 37300 }, { "epoch": 2.8728559341589106, "grad_norm": 0.9955423474311829, "learning_rate": 8.669454328234164e-06, "loss": 1.0447, "step": 37350 }, { "epoch": 2.876701792169833, "grad_norm": 0.5840064287185669, "learning_rate": 8.409734306417682e-06, "loss": 0.9504, "step": 37400 }, { "epoch": 2.8805476501807554, "grad_norm": 1.0777620077133179, "learning_rate": 8.1500142846012e-06, "loss": 0.9635, "step": 37450 }, { "epoch": 2.8843935081916774, "grad_norm": 0.9312844276428223, "learning_rate": 7.890294262784718e-06, "loss": 1.005, "step": 37500 }, { "epoch": 2.8843935081916774, "eval_loss": 0.9864250421524048, "eval_runtime": 17.2238, "eval_samples_per_second": 58.059, "eval_steps_per_second": 14.515, "step": 37500 }, { "epoch": 2.8882393662026, "grad_norm": 0.6800574660301208, "learning_rate": 7.630574240968237e-06, "loss": 0.983, "step": 37550 }, { "epoch": 2.892085224213522, "grad_norm": 0.9417561292648315, "learning_rate": 7.3708542191517545e-06, "loss": 0.973, "step": 37600 }, { "epoch": 2.895931082224444, "grad_norm": 0.6454310417175293, "learning_rate": 7.111134197335273e-06, "loss": 1.0493, "step": 37650 }, { "epoch": 2.8997769402353666, "grad_norm": 1.0287562608718872, "learning_rate": 6.851414175518792e-06, "loss": 0.9821, "step": 37700 }, { "epoch": 2.903622798246289, "grad_norm": 1.6874816417694092, "learning_rate": 6.591694153702309e-06, "loss": 0.9761, "step": 37750 }, { "epoch": 2.903622798246289, "eval_loss": 0.9886119961738586, "eval_runtime": 17.2542, "eval_samples_per_second": 57.957, "eval_steps_per_second": 14.489, "step": 37750 }, { "epoch": 2.907468656257211, "grad_norm": 0.7677489519119263, "learning_rate": 6.331974131885828e-06, "loss": 1.0421, "step": 37800 }, { "epoch": 2.9113145142681334, "grad_norm": 0.9784395098686218, "learning_rate": 6.072254110069346e-06, "loss": 0.9288, "step": 37850 }, { "epoch": 2.9151603722790553, "grad_norm": 0.6236763000488281, "learning_rate": 5.812534088252864e-06, "loss": 0.9777, "step": 37900 }, { "epoch": 2.9190062302899777, "grad_norm": 1.0675079822540283, "learning_rate": 5.552814066436382e-06, "loss": 0.9659, "step": 37950 }, { "epoch": 2.9228520883008997, "grad_norm": 0.9310262203216553, "learning_rate": 5.2930940446198996e-06, "loss": 0.9334, "step": 38000 }, { "epoch": 2.9228520883008997, "eval_loss": 0.98431396484375, "eval_runtime": 17.3967, "eval_samples_per_second": 57.482, "eval_steps_per_second": 14.371, "step": 38000 }, { "epoch": 2.926697946311822, "grad_norm": 0.9615415930747986, "learning_rate": 5.038568423239748e-06, "loss": 0.9715, "step": 38050 }, { "epoch": 2.9305438043227445, "grad_norm": 0.5959100127220154, "learning_rate": 4.778848401423266e-06, "loss": 0.936, "step": 38100 }, { "epoch": 2.9343896623336665, "grad_norm": 1.0181951522827148, "learning_rate": 4.519128379606784e-06, "loss": 0.969, "step": 38150 }, { "epoch": 2.938235520344589, "grad_norm": 0.8716458082199097, "learning_rate": 4.259408357790302e-06, "loss": 1.002, "step": 38200 }, { "epoch": 2.9420813783555113, "grad_norm": 1.6100435256958008, "learning_rate": 3.99968833597382e-06, "loss": 0.9425, "step": 38250 }, { "epoch": 2.9420813783555113, "eval_loss": 0.9931854009628296, "eval_runtime": 17.3406, "eval_samples_per_second": 57.668, "eval_steps_per_second": 14.417, "step": 38250 }, { "epoch": 2.9459272363664333, "grad_norm": 1.1552485227584839, "learning_rate": 3.7399683141573383e-06, "loss": 1.0424, "step": 38300 }, { "epoch": 2.9497730943773557, "grad_norm": 1.441148042678833, "learning_rate": 3.480248292340857e-06, "loss": 1.0688, "step": 38350 }, { "epoch": 2.953618952388278, "grad_norm": 0.8127447366714478, "learning_rate": 3.220528270524375e-06, "loss": 0.9497, "step": 38400 }, { "epoch": 2.9574648103992, "grad_norm": 0.9566346406936646, "learning_rate": 2.9608082487078933e-06, "loss": 0.9628, "step": 38450 }, { "epoch": 2.961310668410122, "grad_norm": 0.7955614924430847, "learning_rate": 2.7010882268914113e-06, "loss": 1.016, "step": 38500 }, { "epoch": 2.961310668410122, "eval_loss": 0.9919273257255554, "eval_runtime": 17.3829, "eval_samples_per_second": 57.528, "eval_steps_per_second": 14.382, "step": 38500 }, { "epoch": 2.9651565264210444, "grad_norm": 0.8041125535964966, "learning_rate": 2.4413682050749296e-06, "loss": 0.9631, "step": 38550 }, { "epoch": 2.969002384431967, "grad_norm": 1.0755919218063354, "learning_rate": 2.1816481832584475e-06, "loss": 0.9689, "step": 38600 }, { "epoch": 2.972848242442889, "grad_norm": 0.8630362153053284, "learning_rate": 1.9219281614419654e-06, "loss": 0.9822, "step": 38650 }, { "epoch": 2.976694100453811, "grad_norm": 0.6994553804397583, "learning_rate": 1.662208139625484e-06, "loss": 0.9706, "step": 38700 }, { "epoch": 2.9805399584647336, "grad_norm": 1.1746189594268799, "learning_rate": 1.4024881178090021e-06, "loss": 0.9695, "step": 38750 }, { "epoch": 2.9805399584647336, "eval_loss": 0.9868382215499878, "eval_runtime": 17.3783, "eval_samples_per_second": 57.543, "eval_steps_per_second": 14.386, "step": 38750 }, { "epoch": 2.9843858164756556, "grad_norm": 1.0636792182922363, "learning_rate": 1.14276809599252e-06, "loss": 1.0233, "step": 38800 }, { "epoch": 2.988231674486578, "grad_norm": 0.8173992037773132, "learning_rate": 8.830480741760383e-07, "loss": 0.963, "step": 38850 }, { "epoch": 2.9920775324975004, "grad_norm": 1.591539978981018, "learning_rate": 6.233280523595564e-07, "loss": 0.949, "step": 38900 }, { "epoch": 2.9959233905084224, "grad_norm": 0.6467046737670898, "learning_rate": 3.636080305430746e-07, "loss": 0.9649, "step": 38950 }, { "epoch": 2.999769248519345, "grad_norm": 0.9910799860954285, "learning_rate": 1.0388800872659275e-07, "loss": 1.0502, "step": 39000 }, { "epoch": 2.999769248519345, "eval_loss": 0.9894677400588989, "eval_runtime": 17.4751, "eval_samples_per_second": 57.224, "eval_steps_per_second": 14.306, "step": 39000 } ], "logging_steps": 50, "max_steps": 39003, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }