{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.05012165450121654, "eval_steps": 500, "global_step": 515, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 9.732360097323601e-05, "grad_norm": 16.13226202148005, "learning_rate": 3.2362459546925574e-08, "loss": 1.1997, "step": 1 }, { "epoch": 0.00019464720194647202, "grad_norm": 15.765097511926365, "learning_rate": 6.472491909385115e-08, "loss": 1.384, "step": 2 }, { "epoch": 0.00029197080291970805, "grad_norm": 16.64113665586635, "learning_rate": 9.70873786407767e-08, "loss": 1.2291, "step": 3 }, { "epoch": 0.00038929440389294404, "grad_norm": 20.34864047521242, "learning_rate": 1.294498381877023e-07, "loss": 0.9025, "step": 4 }, { "epoch": 0.00048661800486618007, "grad_norm": 28.710711096046108, "learning_rate": 1.6181229773462782e-07, "loss": 1.0305, "step": 5 }, { "epoch": 0.0005839416058394161, "grad_norm": 21.945801992582915, "learning_rate": 1.941747572815534e-07, "loss": 1.0979, "step": 6 }, { "epoch": 0.0006812652068126521, "grad_norm": 23.947905905644966, "learning_rate": 2.26537216828479e-07, "loss": 1.1909, "step": 7 }, { "epoch": 0.0007785888077858881, "grad_norm": 19.835016686730835, "learning_rate": 2.588996763754046e-07, "loss": 1.2083, "step": 8 }, { "epoch": 0.0008759124087591241, "grad_norm": 16.926846352507788, "learning_rate": 2.9126213592233014e-07, "loss": 1.2369, "step": 9 }, { "epoch": 0.0009732360097323601, "grad_norm": 21.349924470647284, "learning_rate": 3.2362459546925565e-07, "loss": 1.0052, "step": 10 }, { "epoch": 0.0010705596107055961, "grad_norm": 25.127579741628022, "learning_rate": 3.5598705501618125e-07, "loss": 1.2631, "step": 11 }, { "epoch": 0.0011678832116788322, "grad_norm": 12.524049131196549, "learning_rate": 3.883495145631068e-07, "loss": 1.0884, "step": 12 }, { "epoch": 0.001265206812652068, "grad_norm": 20.706648432487587, "learning_rate": 4.207119741100324e-07, "loss": 1.1469, "step": 13 }, { "epoch": 0.0013625304136253042, "grad_norm": 17.655230197318655, "learning_rate": 4.53074433656958e-07, "loss": 1.2922, "step": 14 }, { "epoch": 0.00145985401459854, "grad_norm": 16.550170455008725, "learning_rate": 4.854368932038835e-07, "loss": 1.1792, "step": 15 }, { "epoch": 0.0015571776155717761, "grad_norm": 24.456798845425887, "learning_rate": 5.177993527508092e-07, "loss": 1.0804, "step": 16 }, { "epoch": 0.0016545012165450122, "grad_norm": 14.659117460865279, "learning_rate": 5.501618122977346e-07, "loss": 1.0973, "step": 17 }, { "epoch": 0.0017518248175182481, "grad_norm": 15.324823146378344, "learning_rate": 5.825242718446603e-07, "loss": 0.9791, "step": 18 }, { "epoch": 0.0018491484184914842, "grad_norm": 12.483869597287145, "learning_rate": 6.148867313915858e-07, "loss": 1.0829, "step": 19 }, { "epoch": 0.0019464720194647203, "grad_norm": 11.921211994178957, "learning_rate": 6.472491909385113e-07, "loss": 0.6862, "step": 20 }, { "epoch": 0.0020437956204379564, "grad_norm": 14.53279456676939, "learning_rate": 6.79611650485437e-07, "loss": 0.7814, "step": 21 }, { "epoch": 0.0021411192214111923, "grad_norm": 15.68359520937104, "learning_rate": 7.119741100323625e-07, "loss": 0.883, "step": 22 }, { "epoch": 0.002238442822384428, "grad_norm": 14.062468532950906, "learning_rate": 7.443365695792882e-07, "loss": 1.0087, "step": 23 }, { "epoch": 0.0023357664233576644, "grad_norm": 11.150778403716444, "learning_rate": 7.766990291262136e-07, "loss": 0.4884, "step": 24 }, { "epoch": 0.0024330900243309003, "grad_norm": 7.740982223602688, "learning_rate": 8.090614886731392e-07, "loss": 0.8543, "step": 25 }, { "epoch": 0.002530413625304136, "grad_norm": 6.4338060169141915, "learning_rate": 8.414239482200648e-07, "loss": 0.7948, "step": 26 }, { "epoch": 0.002627737226277372, "grad_norm": 6.227022582398367, "learning_rate": 8.737864077669904e-07, "loss": 0.7814, "step": 27 }, { "epoch": 0.0027250608272506084, "grad_norm": 7.989531820662516, "learning_rate": 9.06148867313916e-07, "loss": 0.5645, "step": 28 }, { "epoch": 0.0028223844282238442, "grad_norm": 6.4745089193753, "learning_rate": 9.385113268608415e-07, "loss": 0.6802, "step": 29 }, { "epoch": 0.00291970802919708, "grad_norm": 8.23650018531745, "learning_rate": 9.70873786407767e-07, "loss": 0.6218, "step": 30 }, { "epoch": 0.0030170316301703164, "grad_norm": 4.915479010119541, "learning_rate": 1.0032362459546926e-06, "loss": 0.8879, "step": 31 }, { "epoch": 0.0031143552311435523, "grad_norm": 4.288138757396447, "learning_rate": 1.0355987055016184e-06, "loss": 0.5917, "step": 32 }, { "epoch": 0.003211678832116788, "grad_norm": 4.230901102531741, "learning_rate": 1.0679611650485437e-06, "loss": 0.7373, "step": 33 }, { "epoch": 0.0033090024330900245, "grad_norm": 4.714303656539792, "learning_rate": 1.1003236245954693e-06, "loss": 0.5886, "step": 34 }, { "epoch": 0.0034063260340632603, "grad_norm": 4.1204943469600925, "learning_rate": 1.132686084142395e-06, "loss": 0.5991, "step": 35 }, { "epoch": 0.0035036496350364962, "grad_norm": 3.124375547961107, "learning_rate": 1.1650485436893206e-06, "loss": 0.432, "step": 36 }, { "epoch": 0.0036009732360097325, "grad_norm": 3.741153837090354, "learning_rate": 1.197411003236246e-06, "loss": 0.6379, "step": 37 }, { "epoch": 0.0036982968369829684, "grad_norm": 3.7740270813504506, "learning_rate": 1.2297734627831717e-06, "loss": 0.5595, "step": 38 }, { "epoch": 0.0037956204379562043, "grad_norm": 4.783986424289694, "learning_rate": 1.2621359223300972e-06, "loss": 0.8717, "step": 39 }, { "epoch": 0.0038929440389294406, "grad_norm": 4.242597978097827, "learning_rate": 1.2944983818770226e-06, "loss": 0.6632, "step": 40 }, { "epoch": 0.0039902676399026765, "grad_norm": 4.309602952976607, "learning_rate": 1.3268608414239483e-06, "loss": 0.7191, "step": 41 }, { "epoch": 0.004087591240875913, "grad_norm": 4.136462382872819, "learning_rate": 1.359223300970874e-06, "loss": 0.6782, "step": 42 }, { "epoch": 0.004184914841849148, "grad_norm": 4.2148643401229, "learning_rate": 1.3915857605177997e-06, "loss": 0.8932, "step": 43 }, { "epoch": 0.0042822384428223845, "grad_norm": 3.829331188520966, "learning_rate": 1.423948220064725e-06, "loss": 0.4697, "step": 44 }, { "epoch": 0.004379562043795621, "grad_norm": 3.4564347781684557, "learning_rate": 1.4563106796116506e-06, "loss": 0.3377, "step": 45 }, { "epoch": 0.004476885644768856, "grad_norm": 3.319649807488789, "learning_rate": 1.4886731391585763e-06, "loss": 0.4589, "step": 46 }, { "epoch": 0.0045742092457420926, "grad_norm": 3.8856546910308034, "learning_rate": 1.5210355987055017e-06, "loss": 0.8413, "step": 47 }, { "epoch": 0.004671532846715329, "grad_norm": 3.7955924171570605, "learning_rate": 1.5533980582524272e-06, "loss": 0.588, "step": 48 }, { "epoch": 0.004768856447688564, "grad_norm": 4.5762685715882805, "learning_rate": 1.585760517799353e-06, "loss": 0.6472, "step": 49 }, { "epoch": 0.004866180048661801, "grad_norm": 4.284420204063246, "learning_rate": 1.6181229773462783e-06, "loss": 0.5233, "step": 50 }, { "epoch": 0.004963503649635037, "grad_norm": 4.0399534913964645, "learning_rate": 1.650485436893204e-06, "loss": 0.6737, "step": 51 }, { "epoch": 0.005060827250608272, "grad_norm": 4.850258079033273, "learning_rate": 1.6828478964401297e-06, "loss": 0.5017, "step": 52 }, { "epoch": 0.005158150851581509, "grad_norm": 3.289730774319516, "learning_rate": 1.715210355987055e-06, "loss": 0.6378, "step": 53 }, { "epoch": 0.005255474452554744, "grad_norm": 3.116783938182044, "learning_rate": 1.7475728155339808e-06, "loss": 0.5681, "step": 54 }, { "epoch": 0.00535279805352798, "grad_norm": 3.5896487509946677, "learning_rate": 1.7799352750809063e-06, "loss": 0.5222, "step": 55 }, { "epoch": 0.005450121654501217, "grad_norm": 3.3627737905222146, "learning_rate": 1.812297734627832e-06, "loss": 0.351, "step": 56 }, { "epoch": 0.005547445255474452, "grad_norm": 3.405981770724818, "learning_rate": 1.8446601941747574e-06, "loss": 0.5832, "step": 57 }, { "epoch": 0.0056447688564476885, "grad_norm": 3.231134680455488, "learning_rate": 1.877022653721683e-06, "loss": 0.5558, "step": 58 }, { "epoch": 0.005742092457420925, "grad_norm": 4.2963387449464605, "learning_rate": 1.9093851132686085e-06, "loss": 0.7544, "step": 59 }, { "epoch": 0.00583941605839416, "grad_norm": 3.3678084152804315, "learning_rate": 1.941747572815534e-06, "loss": 0.554, "step": 60 }, { "epoch": 0.0059367396593673965, "grad_norm": 3.635756089652443, "learning_rate": 1.9741100323624596e-06, "loss": 0.5312, "step": 61 }, { "epoch": 0.006034063260340633, "grad_norm": 3.91764256649437, "learning_rate": 2.006472491909385e-06, "loss": 0.4329, "step": 62 }, { "epoch": 0.006131386861313868, "grad_norm": 3.4866607421863565, "learning_rate": 2.0388349514563107e-06, "loss": 0.4453, "step": 63 }, { "epoch": 0.006228710462287105, "grad_norm": 2.9369425143161147, "learning_rate": 2.0711974110032367e-06, "loss": 0.467, "step": 64 }, { "epoch": 0.006326034063260341, "grad_norm": 3.0906723589687024, "learning_rate": 2.103559870550162e-06, "loss": 0.3917, "step": 65 }, { "epoch": 0.006423357664233576, "grad_norm": 3.5121616512799747, "learning_rate": 2.1359223300970874e-06, "loss": 0.6428, "step": 66 }, { "epoch": 0.006520681265206813, "grad_norm": 3.470270871630247, "learning_rate": 2.1682847896440134e-06, "loss": 0.586, "step": 67 }, { "epoch": 0.006618004866180049, "grad_norm": 2.8689679430782498, "learning_rate": 2.2006472491909385e-06, "loss": 0.2938, "step": 68 }, { "epoch": 0.006715328467153284, "grad_norm": 4.115573400175418, "learning_rate": 2.2330097087378645e-06, "loss": 0.3855, "step": 69 }, { "epoch": 0.006812652068126521, "grad_norm": 3.903319335204406, "learning_rate": 2.26537216828479e-06, "loss": 0.6272, "step": 70 }, { "epoch": 0.006909975669099757, "grad_norm": 2.649165320750572, "learning_rate": 2.297734627831715e-06, "loss": 0.5229, "step": 71 }, { "epoch": 0.0070072992700729924, "grad_norm": 2.8543884488184235, "learning_rate": 2.330097087378641e-06, "loss": 0.4006, "step": 72 }, { "epoch": 0.007104622871046229, "grad_norm": 2.9817247056794134, "learning_rate": 2.3624595469255667e-06, "loss": 0.2331, "step": 73 }, { "epoch": 0.007201946472019465, "grad_norm": 3.592880940053797, "learning_rate": 2.394822006472492e-06, "loss": 0.4889, "step": 74 }, { "epoch": 0.0072992700729927005, "grad_norm": 2.89844013224274, "learning_rate": 2.427184466019418e-06, "loss": 0.4711, "step": 75 }, { "epoch": 0.007396593673965937, "grad_norm": 2.6071345596032134, "learning_rate": 2.4595469255663434e-06, "loss": 0.4844, "step": 76 }, { "epoch": 0.007493917274939173, "grad_norm": 2.9053930844585776, "learning_rate": 2.491909385113269e-06, "loss": 0.5163, "step": 77 }, { "epoch": 0.0075912408759124085, "grad_norm": 3.4016540038418115, "learning_rate": 2.5242718446601945e-06, "loss": 0.5852, "step": 78 }, { "epoch": 0.007688564476885645, "grad_norm": 2.7133170026932887, "learning_rate": 2.55663430420712e-06, "loss": 0.4934, "step": 79 }, { "epoch": 0.007785888077858881, "grad_norm": 3.2321439410345585, "learning_rate": 2.588996763754045e-06, "loss": 0.62, "step": 80 }, { "epoch": 0.007883211678832117, "grad_norm": 2.6835948161160545, "learning_rate": 2.621359223300971e-06, "loss": 0.4689, "step": 81 }, { "epoch": 0.007980535279805353, "grad_norm": 4.716894934604404, "learning_rate": 2.6537216828478967e-06, "loss": 0.3364, "step": 82 }, { "epoch": 0.00807785888077859, "grad_norm": 2.6507857723180646, "learning_rate": 2.686084142394822e-06, "loss": 0.3785, "step": 83 }, { "epoch": 0.008175182481751826, "grad_norm": 2.356714630861861, "learning_rate": 2.718446601941748e-06, "loss": 0.2591, "step": 84 }, { "epoch": 0.00827250608272506, "grad_norm": 2.755477478688418, "learning_rate": 2.7508090614886734e-06, "loss": 0.4762, "step": 85 }, { "epoch": 0.008369829683698296, "grad_norm": 3.7771581783688837, "learning_rate": 2.7831715210355993e-06, "loss": 0.4627, "step": 86 }, { "epoch": 0.008467153284671533, "grad_norm": 2.8568450908810257, "learning_rate": 2.8155339805825245e-06, "loss": 0.4322, "step": 87 }, { "epoch": 0.008564476885644769, "grad_norm": 2.914756058289183, "learning_rate": 2.84789644012945e-06, "loss": 0.4835, "step": 88 }, { "epoch": 0.008661800486618005, "grad_norm": 2.414182197047686, "learning_rate": 2.880258899676376e-06, "loss": 0.493, "step": 89 }, { "epoch": 0.008759124087591242, "grad_norm": 2.8597853736106975, "learning_rate": 2.912621359223301e-06, "loss": 0.6063, "step": 90 }, { "epoch": 0.008856447688564476, "grad_norm": 2.4567808863650007, "learning_rate": 2.9449838187702267e-06, "loss": 0.5874, "step": 91 }, { "epoch": 0.008953771289537713, "grad_norm": 2.819434031784131, "learning_rate": 2.9773462783171527e-06, "loss": 0.552, "step": 92 }, { "epoch": 0.009051094890510949, "grad_norm": 1.9840396387462764, "learning_rate": 3.0097087378640778e-06, "loss": 0.3736, "step": 93 }, { "epoch": 0.009148418491484185, "grad_norm": 2.52047300259283, "learning_rate": 3.0420711974110033e-06, "loss": 0.407, "step": 94 }, { "epoch": 0.009245742092457421, "grad_norm": 3.140839526692518, "learning_rate": 3.0744336569579293e-06, "loss": 0.6513, "step": 95 }, { "epoch": 0.009343065693430658, "grad_norm": 3.1368865731879554, "learning_rate": 3.1067961165048544e-06, "loss": 0.4804, "step": 96 }, { "epoch": 0.009440389294403892, "grad_norm": 2.6987222968513196, "learning_rate": 3.13915857605178e-06, "loss": 0.4228, "step": 97 }, { "epoch": 0.009537712895377129, "grad_norm": 2.5779408707034026, "learning_rate": 3.171521035598706e-06, "loss": 0.4654, "step": 98 }, { "epoch": 0.009635036496350365, "grad_norm": 2.5189587792888934, "learning_rate": 3.2038834951456315e-06, "loss": 0.5465, "step": 99 }, { "epoch": 0.009732360097323601, "grad_norm": 2.457408493992738, "learning_rate": 3.2362459546925567e-06, "loss": 0.5077, "step": 100 }, { "epoch": 0.009829683698296838, "grad_norm": 2.445932328031196, "learning_rate": 3.2686084142394826e-06, "loss": 0.492, "step": 101 }, { "epoch": 0.009927007299270074, "grad_norm": 2.3199141960061915, "learning_rate": 3.300970873786408e-06, "loss": 0.4432, "step": 102 }, { "epoch": 0.010024330900243308, "grad_norm": 3.88769555780582, "learning_rate": 3.3333333333333333e-06, "loss": 0.3684, "step": 103 }, { "epoch": 0.010121654501216545, "grad_norm": 2.63905676146042, "learning_rate": 3.3656957928802593e-06, "loss": 0.4238, "step": 104 }, { "epoch": 0.010218978102189781, "grad_norm": 3.0073749174392885, "learning_rate": 3.398058252427185e-06, "loss": 0.4655, "step": 105 }, { "epoch": 0.010316301703163017, "grad_norm": 2.613524831872459, "learning_rate": 3.43042071197411e-06, "loss": 0.4948, "step": 106 }, { "epoch": 0.010413625304136254, "grad_norm": 2.4293628733346764, "learning_rate": 3.462783171521036e-06, "loss": 0.3717, "step": 107 }, { "epoch": 0.010510948905109488, "grad_norm": 3.3036504610837016, "learning_rate": 3.4951456310679615e-06, "loss": 0.4939, "step": 108 }, { "epoch": 0.010608272506082725, "grad_norm": 2.6808221933664846, "learning_rate": 3.5275080906148866e-06, "loss": 0.4809, "step": 109 }, { "epoch": 0.01070559610705596, "grad_norm": 2.853958419293739, "learning_rate": 3.5598705501618126e-06, "loss": 0.4066, "step": 110 }, { "epoch": 0.010802919708029197, "grad_norm": 5.3412930378250145, "learning_rate": 3.592233009708738e-06, "loss": 0.3599, "step": 111 }, { "epoch": 0.010900243309002433, "grad_norm": 2.983669976646381, "learning_rate": 3.624595469255664e-06, "loss": 0.6187, "step": 112 }, { "epoch": 0.01099756690997567, "grad_norm": 3.388543821878077, "learning_rate": 3.6569579288025893e-06, "loss": 0.717, "step": 113 }, { "epoch": 0.011094890510948904, "grad_norm": 3.0720120062792127, "learning_rate": 3.689320388349515e-06, "loss": 0.5057, "step": 114 }, { "epoch": 0.01119221411192214, "grad_norm": 2.521868238475485, "learning_rate": 3.721682847896441e-06, "loss": 0.4308, "step": 115 }, { "epoch": 0.011289537712895377, "grad_norm": 2.641085251645149, "learning_rate": 3.754045307443366e-06, "loss": 0.4047, "step": 116 }, { "epoch": 0.011386861313868613, "grad_norm": 2.6936547530255828, "learning_rate": 3.7864077669902915e-06, "loss": 0.5548, "step": 117 }, { "epoch": 0.01148418491484185, "grad_norm": 5.599830434139348, "learning_rate": 3.818770226537217e-06, "loss": 0.5338, "step": 118 }, { "epoch": 0.011581508515815086, "grad_norm": 2.6372065340185378, "learning_rate": 3.851132686084142e-06, "loss": 0.4833, "step": 119 }, { "epoch": 0.01167883211678832, "grad_norm": 2.555049765563167, "learning_rate": 3.883495145631068e-06, "loss": 0.4295, "step": 120 }, { "epoch": 0.011776155717761557, "grad_norm": 2.22725048478721, "learning_rate": 3.915857605177994e-06, "loss": 0.4074, "step": 121 }, { "epoch": 0.011873479318734793, "grad_norm": 3.0093045583939984, "learning_rate": 3.948220064724919e-06, "loss": 0.7168, "step": 122 }, { "epoch": 0.01197080291970803, "grad_norm": 2.8800338131191223, "learning_rate": 3.980582524271845e-06, "loss": 0.3826, "step": 123 }, { "epoch": 0.012068126520681266, "grad_norm": 2.3197904571086974, "learning_rate": 4.01294498381877e-06, "loss": 0.2584, "step": 124 }, { "epoch": 0.012165450121654502, "grad_norm": 2.929540360888414, "learning_rate": 4.045307443365696e-06, "loss": 0.4617, "step": 125 }, { "epoch": 0.012262773722627737, "grad_norm": 2.5602803735383137, "learning_rate": 4.0776699029126215e-06, "loss": 0.2561, "step": 126 }, { "epoch": 0.012360097323600973, "grad_norm": 2.676345297957673, "learning_rate": 4.1100323624595475e-06, "loss": 0.2996, "step": 127 }, { "epoch": 0.01245742092457421, "grad_norm": 1.9047794610871986, "learning_rate": 4.1423948220064734e-06, "loss": 0.3475, "step": 128 }, { "epoch": 0.012554744525547445, "grad_norm": 2.9014607006450555, "learning_rate": 4.1747572815533986e-06, "loss": 0.4748, "step": 129 }, { "epoch": 0.012652068126520682, "grad_norm": 2.2992367182815987, "learning_rate": 4.207119741100324e-06, "loss": 0.3465, "step": 130 }, { "epoch": 0.012749391727493918, "grad_norm": 2.668874383033437, "learning_rate": 4.23948220064725e-06, "loss": 0.6119, "step": 131 }, { "epoch": 0.012846715328467153, "grad_norm": 2.69106703615133, "learning_rate": 4.271844660194175e-06, "loss": 0.4743, "step": 132 }, { "epoch": 0.012944038929440389, "grad_norm": 2.972314561813759, "learning_rate": 4.304207119741101e-06, "loss": 0.5766, "step": 133 }, { "epoch": 0.013041362530413625, "grad_norm": 2.7487017428059635, "learning_rate": 4.336569579288027e-06, "loss": 0.5818, "step": 134 }, { "epoch": 0.013138686131386862, "grad_norm": 3.1117207482379663, "learning_rate": 4.368932038834952e-06, "loss": 0.6468, "step": 135 }, { "epoch": 0.013236009732360098, "grad_norm": 2.781796948090657, "learning_rate": 4.401294498381877e-06, "loss": 0.7209, "step": 136 }, { "epoch": 0.013333333333333334, "grad_norm": 2.5480533986327556, "learning_rate": 4.433656957928803e-06, "loss": 0.5907, "step": 137 }, { "epoch": 0.013430656934306569, "grad_norm": 2.054397852683208, "learning_rate": 4.466019417475729e-06, "loss": 0.4079, "step": 138 }, { "epoch": 0.013527980535279805, "grad_norm": 2.2564046621809037, "learning_rate": 4.498381877022654e-06, "loss": 0.4, "step": 139 }, { "epoch": 0.013625304136253041, "grad_norm": 2.8739841159071022, "learning_rate": 4.53074433656958e-06, "loss": 0.5819, "step": 140 }, { "epoch": 0.013722627737226278, "grad_norm": 2.6418540847993657, "learning_rate": 4.563106796116505e-06, "loss": 0.589, "step": 141 }, { "epoch": 0.013819951338199514, "grad_norm": 2.431908870746442, "learning_rate": 4.59546925566343e-06, "loss": 0.5468, "step": 142 }, { "epoch": 0.013917274939172749, "grad_norm": 4.44933942542394, "learning_rate": 4.627831715210356e-06, "loss": 0.3846, "step": 143 }, { "epoch": 0.014014598540145985, "grad_norm": 2.2469929628351126, "learning_rate": 4.660194174757282e-06, "loss": 0.3047, "step": 144 }, { "epoch": 0.014111922141119221, "grad_norm": 2.8361034502388205, "learning_rate": 4.6925566343042074e-06, "loss": 0.4186, "step": 145 }, { "epoch": 0.014209245742092457, "grad_norm": 2.485184255788147, "learning_rate": 4.724919093851133e-06, "loss": 0.455, "step": 146 }, { "epoch": 0.014306569343065694, "grad_norm": 2.677307495548506, "learning_rate": 4.7572815533980585e-06, "loss": 0.6346, "step": 147 }, { "epoch": 0.01440389294403893, "grad_norm": 2.9440091029213034, "learning_rate": 4.789644012944984e-06, "loss": 0.4961, "step": 148 }, { "epoch": 0.014501216545012165, "grad_norm": 2.6810327828724723, "learning_rate": 4.82200647249191e-06, "loss": 0.3754, "step": 149 }, { "epoch": 0.014598540145985401, "grad_norm": 2.519257002697837, "learning_rate": 4.854368932038836e-06, "loss": 0.249, "step": 150 }, { "epoch": 0.014695863746958637, "grad_norm": 2.8041238457488578, "learning_rate": 4.886731391585761e-06, "loss": 0.3117, "step": 151 }, { "epoch": 0.014793187347931874, "grad_norm": 2.363481194731433, "learning_rate": 4.919093851132687e-06, "loss": 0.3325, "step": 152 }, { "epoch": 0.01489051094890511, "grad_norm": 3.078347599868747, "learning_rate": 4.951456310679612e-06, "loss": 0.3569, "step": 153 }, { "epoch": 0.014987834549878346, "grad_norm": 3.2926461094535515, "learning_rate": 4.983818770226538e-06, "loss": 0.716, "step": 154 }, { "epoch": 0.01508515815085158, "grad_norm": 2.340052421830345, "learning_rate": 5.016181229773464e-06, "loss": 0.2642, "step": 155 }, { "epoch": 0.015182481751824817, "grad_norm": 1.8915730140906823, "learning_rate": 5.048543689320389e-06, "loss": 0.3523, "step": 156 }, { "epoch": 0.015279805352798053, "grad_norm": 4.2448533254564484, "learning_rate": 5.080906148867314e-06, "loss": 0.5185, "step": 157 }, { "epoch": 0.01537712895377129, "grad_norm": 2.1172922256300333, "learning_rate": 5.11326860841424e-06, "loss": 0.3341, "step": 158 }, { "epoch": 0.015474452554744526, "grad_norm": 2.7414250631657113, "learning_rate": 5.145631067961165e-06, "loss": 0.5965, "step": 159 }, { "epoch": 0.015571776155717762, "grad_norm": 1.977804344185745, "learning_rate": 5.17799352750809e-06, "loss": 0.239, "step": 160 }, { "epoch": 0.015669099756690997, "grad_norm": 2.771807640315191, "learning_rate": 5.210355987055017e-06, "loss": 0.4122, "step": 161 }, { "epoch": 0.015766423357664233, "grad_norm": 1.9977073642008174, "learning_rate": 5.242718446601942e-06, "loss": 0.3423, "step": 162 }, { "epoch": 0.01586374695863747, "grad_norm": 3.222730527079622, "learning_rate": 5.275080906148867e-06, "loss": 0.5647, "step": 163 }, { "epoch": 0.015961070559610706, "grad_norm": 2.95441646694508, "learning_rate": 5.307443365695793e-06, "loss": 0.5198, "step": 164 }, { "epoch": 0.016058394160583942, "grad_norm": 2.3346384576429116, "learning_rate": 5.3398058252427185e-06, "loss": 0.3516, "step": 165 }, { "epoch": 0.01615571776155718, "grad_norm": 2.089159587923689, "learning_rate": 5.372168284789644e-06, "loss": 0.3704, "step": 166 }, { "epoch": 0.016253041362530415, "grad_norm": 2.8135820638465088, "learning_rate": 5.4045307443365705e-06, "loss": 0.3729, "step": 167 }, { "epoch": 0.01635036496350365, "grad_norm": 2.991259557993277, "learning_rate": 5.436893203883496e-06, "loss": 0.5622, "step": 168 }, { "epoch": 0.016447688564476887, "grad_norm": 3.1512644455187857, "learning_rate": 5.4692556634304216e-06, "loss": 0.5915, "step": 169 }, { "epoch": 0.01654501216545012, "grad_norm": 2.616126184062516, "learning_rate": 5.501618122977347e-06, "loss": 0.4252, "step": 170 }, { "epoch": 0.016642335766423356, "grad_norm": 1.9958281517625203, "learning_rate": 5.533980582524272e-06, "loss": 0.3704, "step": 171 }, { "epoch": 0.016739659367396593, "grad_norm": 2.470731302334384, "learning_rate": 5.566343042071199e-06, "loss": 0.4373, "step": 172 }, { "epoch": 0.01683698296836983, "grad_norm": 2.583270308023139, "learning_rate": 5.598705501618124e-06, "loss": 0.4125, "step": 173 }, { "epoch": 0.016934306569343065, "grad_norm": 1.9644684632241667, "learning_rate": 5.631067961165049e-06, "loss": 0.3522, "step": 174 }, { "epoch": 0.0170316301703163, "grad_norm": 2.4207097357376046, "learning_rate": 5.663430420711975e-06, "loss": 0.3579, "step": 175 }, { "epoch": 0.017128953771289538, "grad_norm": 2.3511041847292034, "learning_rate": 5.6957928802589e-06, "loss": 0.5412, "step": 176 }, { "epoch": 0.017226277372262774, "grad_norm": 2.274427899539275, "learning_rate": 5.728155339805825e-06, "loss": 0.5353, "step": 177 }, { "epoch": 0.01732360097323601, "grad_norm": 2.133749284526256, "learning_rate": 5.760517799352752e-06, "loss": 0.4392, "step": 178 }, { "epoch": 0.017420924574209247, "grad_norm": 2.3097462109285787, "learning_rate": 5.792880258899677e-06, "loss": 0.4442, "step": 179 }, { "epoch": 0.017518248175182483, "grad_norm": 2.2128802818602056, "learning_rate": 5.825242718446602e-06, "loss": 0.5635, "step": 180 }, { "epoch": 0.017615571776155716, "grad_norm": 2.103405792854256, "learning_rate": 5.857605177993528e-06, "loss": 0.4533, "step": 181 }, { "epoch": 0.017712895377128952, "grad_norm": 2.0565661990183597, "learning_rate": 5.889967637540453e-06, "loss": 0.3806, "step": 182 }, { "epoch": 0.01781021897810219, "grad_norm": 2.179649872267064, "learning_rate": 5.9223300970873785e-06, "loss": 0.3842, "step": 183 }, { "epoch": 0.017907542579075425, "grad_norm": 3.8333244047199146, "learning_rate": 5.954692556634305e-06, "loss": 0.3876, "step": 184 }, { "epoch": 0.01800486618004866, "grad_norm": 2.2893517217095716, "learning_rate": 5.9870550161812304e-06, "loss": 0.4781, "step": 185 }, { "epoch": 0.018102189781021898, "grad_norm": 1.6022498167897639, "learning_rate": 6.0194174757281556e-06, "loss": 0.2306, "step": 186 }, { "epoch": 0.018199513381995134, "grad_norm": 2.32863493589546, "learning_rate": 6.0517799352750815e-06, "loss": 0.5139, "step": 187 }, { "epoch": 0.01829683698296837, "grad_norm": 2.0789478938631314, "learning_rate": 6.084142394822007e-06, "loss": 0.2824, "step": 188 }, { "epoch": 0.018394160583941607, "grad_norm": 1.7544615955949223, "learning_rate": 6.116504854368932e-06, "loss": 0.4172, "step": 189 }, { "epoch": 0.018491484184914843, "grad_norm": 1.931043696572374, "learning_rate": 6.148867313915859e-06, "loss": 0.3584, "step": 190 }, { "epoch": 0.01858880778588808, "grad_norm": 2.467258437370788, "learning_rate": 6.181229773462784e-06, "loss": 0.462, "step": 191 }, { "epoch": 0.018686131386861315, "grad_norm": 2.1541091684996965, "learning_rate": 6.213592233009709e-06, "loss": 0.3967, "step": 192 }, { "epoch": 0.01878345498783455, "grad_norm": 2.2330486922808395, "learning_rate": 6.245954692556635e-06, "loss": 0.5316, "step": 193 }, { "epoch": 0.018880778588807785, "grad_norm": 2.3498262097642395, "learning_rate": 6.27831715210356e-06, "loss": 0.4815, "step": 194 }, { "epoch": 0.01897810218978102, "grad_norm": 1.7045092076002246, "learning_rate": 6.310679611650487e-06, "loss": 0.3, "step": 195 }, { "epoch": 0.019075425790754257, "grad_norm": 2.5703331850837023, "learning_rate": 6.343042071197412e-06, "loss": 0.4143, "step": 196 }, { "epoch": 0.019172749391727494, "grad_norm": 2.6940646171495133, "learning_rate": 6.375404530744337e-06, "loss": 0.5463, "step": 197 }, { "epoch": 0.01927007299270073, "grad_norm": 2.4185580273524847, "learning_rate": 6.407766990291263e-06, "loss": 0.5215, "step": 198 }, { "epoch": 0.019367396593673966, "grad_norm": 2.6509824694985946, "learning_rate": 6.440129449838188e-06, "loss": 0.5286, "step": 199 }, { "epoch": 0.019464720194647202, "grad_norm": 2.4807219128312767, "learning_rate": 6.472491909385113e-06, "loss": 0.3996, "step": 200 }, { "epoch": 0.01956204379562044, "grad_norm": 2.651883834043772, "learning_rate": 6.50485436893204e-06, "loss": 0.3499, "step": 201 }, { "epoch": 0.019659367396593675, "grad_norm": 2.670759179984812, "learning_rate": 6.537216828478965e-06, "loss": 0.552, "step": 202 }, { "epoch": 0.01975669099756691, "grad_norm": 2.51305850829245, "learning_rate": 6.56957928802589e-06, "loss": 0.3806, "step": 203 }, { "epoch": 0.019854014598540148, "grad_norm": 2.435954851305265, "learning_rate": 6.601941747572816e-06, "loss": 0.6093, "step": 204 }, { "epoch": 0.01995133819951338, "grad_norm": 2.091315833022872, "learning_rate": 6.6343042071197415e-06, "loss": 0.3573, "step": 205 }, { "epoch": 0.020048661800486617, "grad_norm": 2.205515437184344, "learning_rate": 6.666666666666667e-06, "loss": 0.2892, "step": 206 }, { "epoch": 0.020145985401459853, "grad_norm": 2.314981932930035, "learning_rate": 6.6990291262135935e-06, "loss": 0.4184, "step": 207 }, { "epoch": 0.02024330900243309, "grad_norm": 1.9102474885146974, "learning_rate": 6.731391585760519e-06, "loss": 0.2287, "step": 208 }, { "epoch": 0.020340632603406326, "grad_norm": 1.9408029275065433, "learning_rate": 6.763754045307444e-06, "loss": 0.3958, "step": 209 }, { "epoch": 0.020437956204379562, "grad_norm": 2.1006467731485823, "learning_rate": 6.79611650485437e-06, "loss": 0.3764, "step": 210 }, { "epoch": 0.0205352798053528, "grad_norm": 2.0927447282795146, "learning_rate": 6.828478964401295e-06, "loss": 0.531, "step": 211 }, { "epoch": 0.020632603406326035, "grad_norm": 3.4830081465453633, "learning_rate": 6.86084142394822e-06, "loss": 0.4887, "step": 212 }, { "epoch": 0.02072992700729927, "grad_norm": 2.253360993066953, "learning_rate": 6.893203883495147e-06, "loss": 0.4587, "step": 213 }, { "epoch": 0.020827250608272507, "grad_norm": 3.3751096354443852, "learning_rate": 6.925566343042072e-06, "loss": 0.3427, "step": 214 }, { "epoch": 0.020924574209245744, "grad_norm": 1.9729713112803993, "learning_rate": 6.957928802588997e-06, "loss": 0.384, "step": 215 }, { "epoch": 0.021021897810218976, "grad_norm": 2.761285167796522, "learning_rate": 6.990291262135923e-06, "loss": 0.3512, "step": 216 }, { "epoch": 0.021119221411192213, "grad_norm": 2.431882400442612, "learning_rate": 7.022653721682848e-06, "loss": 0.3971, "step": 217 }, { "epoch": 0.02121654501216545, "grad_norm": 3.659254877088116, "learning_rate": 7.055016181229773e-06, "loss": 0.4115, "step": 218 }, { "epoch": 0.021313868613138685, "grad_norm": 2.5501534359714655, "learning_rate": 7.0873786407767e-06, "loss": 0.4963, "step": 219 }, { "epoch": 0.02141119221411192, "grad_norm": 4.296894309260591, "learning_rate": 7.119741100323625e-06, "loss": 0.5203, "step": 220 }, { "epoch": 0.021508515815085158, "grad_norm": 2.5489854552137237, "learning_rate": 7.152103559870551e-06, "loss": 0.4343, "step": 221 }, { "epoch": 0.021605839416058394, "grad_norm": 2.00955207958064, "learning_rate": 7.184466019417476e-06, "loss": 0.3603, "step": 222 }, { "epoch": 0.02170316301703163, "grad_norm": 2.2675038932590224, "learning_rate": 7.2168284789644015e-06, "loss": 0.3968, "step": 223 }, { "epoch": 0.021800486618004867, "grad_norm": 2.4690586331753277, "learning_rate": 7.249190938511328e-06, "loss": 0.5883, "step": 224 }, { "epoch": 0.021897810218978103, "grad_norm": 2.141328682063472, "learning_rate": 7.2815533980582534e-06, "loss": 0.3547, "step": 225 }, { "epoch": 0.02199513381995134, "grad_norm": 2.223927434368622, "learning_rate": 7.3139158576051786e-06, "loss": 0.5031, "step": 226 }, { "epoch": 0.022092457420924576, "grad_norm": 2.8602320319532346, "learning_rate": 7.3462783171521046e-06, "loss": 0.4226, "step": 227 }, { "epoch": 0.02218978102189781, "grad_norm": 2.8852449405031835, "learning_rate": 7.37864077669903e-06, "loss": 0.4298, "step": 228 }, { "epoch": 0.022287104622871045, "grad_norm": 1.7370522944561966, "learning_rate": 7.411003236245955e-06, "loss": 0.3827, "step": 229 }, { "epoch": 0.02238442822384428, "grad_norm": 2.3907908463140584, "learning_rate": 7.443365695792882e-06, "loss": 0.4139, "step": 230 }, { "epoch": 0.022481751824817518, "grad_norm": 2.27581306432663, "learning_rate": 7.475728155339807e-06, "loss": 0.4736, "step": 231 }, { "epoch": 0.022579075425790754, "grad_norm": 2.1861094823645675, "learning_rate": 7.508090614886732e-06, "loss": 0.4809, "step": 232 }, { "epoch": 0.02267639902676399, "grad_norm": 1.9626208371421419, "learning_rate": 7.540453074433658e-06, "loss": 0.3436, "step": 233 }, { "epoch": 0.022773722627737226, "grad_norm": 1.7092390993202267, "learning_rate": 7.572815533980583e-06, "loss": 0.3224, "step": 234 }, { "epoch": 0.022871046228710463, "grad_norm": 3.0168693228526546, "learning_rate": 7.605177993527508e-06, "loss": 0.6366, "step": 235 }, { "epoch": 0.0229683698296837, "grad_norm": 2.424919921496664, "learning_rate": 7.637540453074434e-06, "loss": 0.4483, "step": 236 }, { "epoch": 0.023065693430656935, "grad_norm": 2.4586833984787626, "learning_rate": 7.66990291262136e-06, "loss": 0.4031, "step": 237 }, { "epoch": 0.02316301703163017, "grad_norm": 2.092010230715883, "learning_rate": 7.702265372168284e-06, "loss": 0.4257, "step": 238 }, { "epoch": 0.023260340632603408, "grad_norm": 2.3360188447701655, "learning_rate": 7.734627831715211e-06, "loss": 0.4684, "step": 239 }, { "epoch": 0.02335766423357664, "grad_norm": 2.087175894606599, "learning_rate": 7.766990291262136e-06, "loss": 0.4272, "step": 240 }, { "epoch": 0.023454987834549877, "grad_norm": 2.598684557686617, "learning_rate": 7.799352750809061e-06, "loss": 0.5401, "step": 241 }, { "epoch": 0.023552311435523113, "grad_norm": 2.025117037181364, "learning_rate": 7.831715210355988e-06, "loss": 0.372, "step": 242 }, { "epoch": 0.02364963503649635, "grad_norm": 2.2467324584398405, "learning_rate": 7.864077669902913e-06, "loss": 0.5891, "step": 243 }, { "epoch": 0.023746958637469586, "grad_norm": 2.38036373195977, "learning_rate": 7.896440129449839e-06, "loss": 0.5133, "step": 244 }, { "epoch": 0.023844282238442822, "grad_norm": 2.052700924442009, "learning_rate": 7.928802588996765e-06, "loss": 0.5161, "step": 245 }, { "epoch": 0.02394160583941606, "grad_norm": 3.4299018810240254, "learning_rate": 7.96116504854369e-06, "loss": 0.5314, "step": 246 }, { "epoch": 0.024038929440389295, "grad_norm": 1.3903956706369247, "learning_rate": 7.993527508090616e-06, "loss": 0.3539, "step": 247 }, { "epoch": 0.02413625304136253, "grad_norm": 2.4599878810180873, "learning_rate": 8.02588996763754e-06, "loss": 0.4876, "step": 248 }, { "epoch": 0.024233576642335768, "grad_norm": 2.4053308291912083, "learning_rate": 8.058252427184466e-06, "loss": 0.5185, "step": 249 }, { "epoch": 0.024330900243309004, "grad_norm": 1.6624263546342495, "learning_rate": 8.090614886731393e-06, "loss": 0.2909, "step": 250 }, { "epoch": 0.024428223844282237, "grad_norm": 2.4091367373679597, "learning_rate": 8.122977346278318e-06, "loss": 0.6192, "step": 251 }, { "epoch": 0.024525547445255473, "grad_norm": 2.4595313520548427, "learning_rate": 8.155339805825243e-06, "loss": 0.3444, "step": 252 }, { "epoch": 0.02462287104622871, "grad_norm": 2.3200411140153174, "learning_rate": 8.18770226537217e-06, "loss": 0.6112, "step": 253 }, { "epoch": 0.024720194647201946, "grad_norm": 2.029624875741936, "learning_rate": 8.220064724919095e-06, "loss": 0.4524, "step": 254 }, { "epoch": 0.024817518248175182, "grad_norm": 1.8862765408033388, "learning_rate": 8.25242718446602e-06, "loss": 0.2173, "step": 255 }, { "epoch": 0.02491484184914842, "grad_norm": 2.575687620331568, "learning_rate": 8.284789644012947e-06, "loss": 0.4599, "step": 256 }, { "epoch": 0.025012165450121655, "grad_norm": 2.373530485379713, "learning_rate": 8.317152103559872e-06, "loss": 0.5326, "step": 257 }, { "epoch": 0.02510948905109489, "grad_norm": 2.4086353319447262, "learning_rate": 8.349514563106797e-06, "loss": 0.6275, "step": 258 }, { "epoch": 0.025206812652068127, "grad_norm": 2.1075725625285697, "learning_rate": 8.381877022653722e-06, "loss": 0.44, "step": 259 }, { "epoch": 0.025304136253041364, "grad_norm": 2.0285700798989614, "learning_rate": 8.414239482200647e-06, "loss": 0.3489, "step": 260 }, { "epoch": 0.0254014598540146, "grad_norm": 2.5592973746241, "learning_rate": 8.446601941747573e-06, "loss": 0.4403, "step": 261 }, { "epoch": 0.025498783454987836, "grad_norm": 2.470930078509074, "learning_rate": 8.4789644012945e-06, "loss": 0.4985, "step": 262 }, { "epoch": 0.02559610705596107, "grad_norm": 2.099638103909556, "learning_rate": 8.511326860841424e-06, "loss": 0.4194, "step": 263 }, { "epoch": 0.025693430656934305, "grad_norm": 1.6030834140551835, "learning_rate": 8.54368932038835e-06, "loss": 0.3382, "step": 264 }, { "epoch": 0.02579075425790754, "grad_norm": 1.8960928547169034, "learning_rate": 8.576051779935276e-06, "loss": 0.2838, "step": 265 }, { "epoch": 0.025888077858880778, "grad_norm": 2.4306930963261966, "learning_rate": 8.608414239482202e-06, "loss": 0.4956, "step": 266 }, { "epoch": 0.025985401459854014, "grad_norm": 2.374430136325354, "learning_rate": 8.640776699029127e-06, "loss": 0.5083, "step": 267 }, { "epoch": 0.02608272506082725, "grad_norm": 2.410095115145934, "learning_rate": 8.673139158576054e-06, "loss": 0.4247, "step": 268 }, { "epoch": 0.026180048661800487, "grad_norm": 2.41271065696519, "learning_rate": 8.705501618122979e-06, "loss": 0.6946, "step": 269 }, { "epoch": 0.026277372262773723, "grad_norm": 1.752688930628829, "learning_rate": 8.737864077669904e-06, "loss": 0.2662, "step": 270 }, { "epoch": 0.02637469586374696, "grad_norm": 1.9842034213162434, "learning_rate": 8.770226537216829e-06, "loss": 0.3611, "step": 271 }, { "epoch": 0.026472019464720196, "grad_norm": 2.4137979998327497, "learning_rate": 8.802588996763754e-06, "loss": 0.501, "step": 272 }, { "epoch": 0.026569343065693432, "grad_norm": 2.929650064864996, "learning_rate": 8.834951456310681e-06, "loss": 0.6153, "step": 273 }, { "epoch": 0.02666666666666667, "grad_norm": 2.281738020025263, "learning_rate": 8.867313915857606e-06, "loss": 0.5395, "step": 274 }, { "epoch": 0.0267639902676399, "grad_norm": 2.1406726692627975, "learning_rate": 8.899676375404531e-06, "loss": 0.4039, "step": 275 }, { "epoch": 0.026861313868613138, "grad_norm": 3.2366954201371523, "learning_rate": 8.932038834951458e-06, "loss": 0.5414, "step": 276 }, { "epoch": 0.026958637469586374, "grad_norm": 2.1900667662872513, "learning_rate": 8.964401294498383e-06, "loss": 0.3815, "step": 277 }, { "epoch": 0.02705596107055961, "grad_norm": 2.5301939091612216, "learning_rate": 8.996763754045308e-06, "loss": 0.8016, "step": 278 }, { "epoch": 0.027153284671532846, "grad_norm": 2.2552758985680907, "learning_rate": 9.029126213592233e-06, "loss": 0.4133, "step": 279 }, { "epoch": 0.027250608272506083, "grad_norm": 2.309545536997134, "learning_rate": 9.06148867313916e-06, "loss": 0.5346, "step": 280 }, { "epoch": 0.02734793187347932, "grad_norm": 2.585578916644781, "learning_rate": 9.093851132686085e-06, "loss": 0.407, "step": 281 }, { "epoch": 0.027445255474452555, "grad_norm": 1.8503464194025006, "learning_rate": 9.12621359223301e-06, "loss": 0.4674, "step": 282 }, { "epoch": 0.02754257907542579, "grad_norm": 2.431490115980846, "learning_rate": 9.158576051779936e-06, "loss": 0.6026, "step": 283 }, { "epoch": 0.027639902676399028, "grad_norm": 1.916233248702735, "learning_rate": 9.19093851132686e-06, "loss": 0.4949, "step": 284 }, { "epoch": 0.027737226277372264, "grad_norm": 2.2160236640245072, "learning_rate": 9.223300970873788e-06, "loss": 0.4765, "step": 285 }, { "epoch": 0.027834549878345497, "grad_norm": 2.0764827118780143, "learning_rate": 9.255663430420713e-06, "loss": 0.472, "step": 286 }, { "epoch": 0.027931873479318733, "grad_norm": 2.638286661284288, "learning_rate": 9.288025889967638e-06, "loss": 0.6312, "step": 287 }, { "epoch": 0.02802919708029197, "grad_norm": 1.940011273577467, "learning_rate": 9.320388349514565e-06, "loss": 0.4555, "step": 288 }, { "epoch": 0.028126520681265206, "grad_norm": 1.8760624736314784, "learning_rate": 9.35275080906149e-06, "loss": 0.3625, "step": 289 }, { "epoch": 0.028223844282238442, "grad_norm": 1.3468692859077058, "learning_rate": 9.385113268608415e-06, "loss": 0.2442, "step": 290 }, { "epoch": 0.02832116788321168, "grad_norm": 2.1497394847504014, "learning_rate": 9.41747572815534e-06, "loss": 0.5227, "step": 291 }, { "epoch": 0.028418491484184915, "grad_norm": 2.1233743171190014, "learning_rate": 9.449838187702267e-06, "loss": 0.6184, "step": 292 }, { "epoch": 0.02851581508515815, "grad_norm": 2.337806183860394, "learning_rate": 9.482200647249192e-06, "loss": 0.5491, "step": 293 }, { "epoch": 0.028613138686131388, "grad_norm": 2.015000594070385, "learning_rate": 9.514563106796117e-06, "loss": 0.5137, "step": 294 }, { "epoch": 0.028710462287104624, "grad_norm": 2.0267324830753766, "learning_rate": 9.546925566343042e-06, "loss": 0.4117, "step": 295 }, { "epoch": 0.02880778588807786, "grad_norm": 1.732639028192012, "learning_rate": 9.579288025889967e-06, "loss": 0.3156, "step": 296 }, { "epoch": 0.028905109489051097, "grad_norm": 2.1204056159243923, "learning_rate": 9.611650485436894e-06, "loss": 0.6056, "step": 297 }, { "epoch": 0.02900243309002433, "grad_norm": 1.7868071753968195, "learning_rate": 9.64401294498382e-06, "loss": 0.3417, "step": 298 }, { "epoch": 0.029099756690997566, "grad_norm": 1.9477439300595292, "learning_rate": 9.676375404530746e-06, "loss": 0.3631, "step": 299 }, { "epoch": 0.029197080291970802, "grad_norm": 1.7688147839655162, "learning_rate": 9.708737864077671e-06, "loss": 0.3605, "step": 300 }, { "epoch": 0.029294403892944038, "grad_norm": 1.9162335597538034, "learning_rate": 9.741100323624596e-06, "loss": 0.2498, "step": 301 }, { "epoch": 0.029391727493917275, "grad_norm": 2.9282579520055756, "learning_rate": 9.773462783171522e-06, "loss": 0.4286, "step": 302 }, { "epoch": 0.02948905109489051, "grad_norm": 1.9744499285549086, "learning_rate": 9.805825242718447e-06, "loss": 0.3391, "step": 303 }, { "epoch": 0.029586374695863747, "grad_norm": 2.2116032868392455, "learning_rate": 9.838187702265373e-06, "loss": 0.3414, "step": 304 }, { "epoch": 0.029683698296836983, "grad_norm": 1.9159144570242486, "learning_rate": 9.870550161812299e-06, "loss": 0.2915, "step": 305 }, { "epoch": 0.02978102189781022, "grad_norm": 2.671718838238437, "learning_rate": 9.902912621359224e-06, "loss": 0.79, "step": 306 }, { "epoch": 0.029878345498783456, "grad_norm": 2.093937424199301, "learning_rate": 9.935275080906149e-06, "loss": 0.576, "step": 307 }, { "epoch": 0.029975669099756692, "grad_norm": 1.895574286512308, "learning_rate": 9.967637540453076e-06, "loss": 0.4223, "step": 308 }, { "epoch": 0.03007299270072993, "grad_norm": 2.142643554578675, "learning_rate": 1e-05, "loss": 0.4719, "step": 309 }, { "epoch": 0.03017031630170316, "grad_norm": 2.2548613483238378, "learning_rate": 9.999999751573464e-06, "loss": 0.5547, "step": 310 }, { "epoch": 0.030267639902676398, "grad_norm": 2.375146158639999, "learning_rate": 9.99999900629388e-06, "loss": 0.3864, "step": 311 }, { "epoch": 0.030364963503649634, "grad_norm": 1.558937895217452, "learning_rate": 9.99999776416132e-06, "loss": 0.3409, "step": 312 }, { "epoch": 0.03046228710462287, "grad_norm": 2.7508940543848115, "learning_rate": 9.99999602517591e-06, "loss": 0.3641, "step": 313 }, { "epoch": 0.030559610705596107, "grad_norm": 2.228096737889712, "learning_rate": 9.99999378933782e-06, "loss": 0.6464, "step": 314 }, { "epoch": 0.030656934306569343, "grad_norm": 1.5612763763472, "learning_rate": 9.999991056647274e-06, "loss": 0.3124, "step": 315 }, { "epoch": 0.03075425790754258, "grad_norm": 2.3203527787434104, "learning_rate": 9.999987827104544e-06, "loss": 0.5893, "step": 316 }, { "epoch": 0.030851581508515816, "grad_norm": 1.8472611567410342, "learning_rate": 9.999984100709951e-06, "loss": 0.3732, "step": 317 }, { "epoch": 0.030948905109489052, "grad_norm": 2.269778108549014, "learning_rate": 9.999979877463866e-06, "loss": 0.5537, "step": 318 }, { "epoch": 0.03104622871046229, "grad_norm": 2.381498581134022, "learning_rate": 9.999975157366705e-06, "loss": 0.7179, "step": 319 }, { "epoch": 0.031143552311435525, "grad_norm": 1.7030655036823346, "learning_rate": 9.99996994041894e-06, "loss": 0.4256, "step": 320 }, { "epoch": 0.031240875912408757, "grad_norm": 1.8361141038730153, "learning_rate": 9.999964226621089e-06, "loss": 0.4648, "step": 321 }, { "epoch": 0.031338199513381994, "grad_norm": 1.7985459229558753, "learning_rate": 9.99995801597372e-06, "loss": 0.3031, "step": 322 }, { "epoch": 0.031435523114355234, "grad_norm": 2.4309020119442915, "learning_rate": 9.99995130847745e-06, "loss": 0.5011, "step": 323 }, { "epoch": 0.031532846715328466, "grad_norm": 2.048514022969095, "learning_rate": 9.999944104132944e-06, "loss": 0.6152, "step": 324 }, { "epoch": 0.031630170316301706, "grad_norm": 1.8892667320795724, "learning_rate": 9.99993640294092e-06, "loss": 0.4738, "step": 325 }, { "epoch": 0.03172749391727494, "grad_norm": 2.081179331819785, "learning_rate": 9.999928204902141e-06, "loss": 0.5192, "step": 326 }, { "epoch": 0.03182481751824817, "grad_norm": 2.410280889073595, "learning_rate": 9.999919510017424e-06, "loss": 0.3314, "step": 327 }, { "epoch": 0.03192214111922141, "grad_norm": 1.663034255724975, "learning_rate": 9.999910318287632e-06, "loss": 0.3342, "step": 328 }, { "epoch": 0.032019464720194644, "grad_norm": 1.7874391345352068, "learning_rate": 9.999900629713679e-06, "loss": 0.3189, "step": 329 }, { "epoch": 0.032116788321167884, "grad_norm": 2.1098429973097805, "learning_rate": 9.999890444296528e-06, "loss": 0.4561, "step": 330 }, { "epoch": 0.03221411192214112, "grad_norm": 2.4678279265353558, "learning_rate": 9.999879762037187e-06, "loss": 0.5831, "step": 331 }, { "epoch": 0.03231143552311436, "grad_norm": 1.6643716587630457, "learning_rate": 9.999868582936726e-06, "loss": 0.4371, "step": 332 }, { "epoch": 0.03240875912408759, "grad_norm": 2.088466639768523, "learning_rate": 9.999856906996246e-06, "loss": 0.3904, "step": 333 }, { "epoch": 0.03250608272506083, "grad_norm": 2.0023651443392256, "learning_rate": 9.999844734216914e-06, "loss": 0.4802, "step": 334 }, { "epoch": 0.03260340632603406, "grad_norm": 2.161282844007076, "learning_rate": 9.99983206459994e-06, "loss": 0.5187, "step": 335 }, { "epoch": 0.0327007299270073, "grad_norm": 2.10212671583593, "learning_rate": 9.999818898146576e-06, "loss": 0.4618, "step": 336 }, { "epoch": 0.032798053527980535, "grad_norm": 2.2142899508809286, "learning_rate": 9.999805234858137e-06, "loss": 0.2387, "step": 337 }, { "epoch": 0.032895377128953775, "grad_norm": 2.1084763484693023, "learning_rate": 9.999791074735981e-06, "loss": 0.5652, "step": 338 }, { "epoch": 0.03299270072992701, "grad_norm": 2.261838498017328, "learning_rate": 9.99977641778151e-06, "loss": 0.7224, "step": 339 }, { "epoch": 0.03309002433090024, "grad_norm": 1.612816030006559, "learning_rate": 9.999761263996184e-06, "loss": 0.377, "step": 340 }, { "epoch": 0.03318734793187348, "grad_norm": 2.1209830295615832, "learning_rate": 9.999745613381507e-06, "loss": 0.614, "step": 341 }, { "epoch": 0.03328467153284671, "grad_norm": 1.7938764015879674, "learning_rate": 9.999729465939036e-06, "loss": 0.3983, "step": 342 }, { "epoch": 0.03338199513381995, "grad_norm": 1.943418875698731, "learning_rate": 9.999712821670375e-06, "loss": 0.4708, "step": 343 }, { "epoch": 0.033479318734793186, "grad_norm": 1.9787546900237571, "learning_rate": 9.99969568057718e-06, "loss": 0.578, "step": 344 }, { "epoch": 0.033576642335766425, "grad_norm": 1.4798263726328331, "learning_rate": 9.99967804266115e-06, "loss": 0.394, "step": 345 }, { "epoch": 0.03367396593673966, "grad_norm": 2.1936298043995484, "learning_rate": 9.99965990792404e-06, "loss": 0.6316, "step": 346 }, { "epoch": 0.0337712895377129, "grad_norm": 2.2799650780195133, "learning_rate": 9.99964127636765e-06, "loss": 0.3985, "step": 347 }, { "epoch": 0.03386861313868613, "grad_norm": 1.8519049219191819, "learning_rate": 9.999622147993837e-06, "loss": 0.3853, "step": 348 }, { "epoch": 0.03396593673965937, "grad_norm": 1.5111895282974241, "learning_rate": 9.999602522804497e-06, "loss": 0.4201, "step": 349 }, { "epoch": 0.0340632603406326, "grad_norm": 1.8605769784283237, "learning_rate": 9.99958240080158e-06, "loss": 0.5225, "step": 350 }, { "epoch": 0.034160583941605836, "grad_norm": 1.6063240538866903, "learning_rate": 9.999561781987087e-06, "loss": 0.3165, "step": 351 }, { "epoch": 0.034257907542579076, "grad_norm": 1.4751976204077173, "learning_rate": 9.999540666363068e-06, "loss": 0.3156, "step": 352 }, { "epoch": 0.03435523114355231, "grad_norm": 2.1029966771511757, "learning_rate": 9.99951905393162e-06, "loss": 0.5336, "step": 353 }, { "epoch": 0.03445255474452555, "grad_norm": 2.1419054642874267, "learning_rate": 9.99949694469489e-06, "loss": 0.5253, "step": 354 }, { "epoch": 0.03454987834549878, "grad_norm": 2.169397271826959, "learning_rate": 9.999474338655075e-06, "loss": 0.5567, "step": 355 }, { "epoch": 0.03464720194647202, "grad_norm": 2.2972412855327797, "learning_rate": 9.999451235814422e-06, "loss": 0.5233, "step": 356 }, { "epoch": 0.034744525547445254, "grad_norm": 1.830377999961128, "learning_rate": 9.999427636175228e-06, "loss": 0.4297, "step": 357 }, { "epoch": 0.034841849148418494, "grad_norm": 2.1217123292302875, "learning_rate": 9.999403539739837e-06, "loss": 0.3605, "step": 358 }, { "epoch": 0.03493917274939173, "grad_norm": 2.001599625802253, "learning_rate": 9.999378946510642e-06, "loss": 0.5237, "step": 359 }, { "epoch": 0.035036496350364967, "grad_norm": 1.6719956399048532, "learning_rate": 9.99935385649009e-06, "loss": 0.424, "step": 360 }, { "epoch": 0.0351338199513382, "grad_norm": 1.5962062682133515, "learning_rate": 9.99932826968067e-06, "loss": 0.4228, "step": 361 }, { "epoch": 0.03523114355231143, "grad_norm": 1.9754274750693919, "learning_rate": 9.999302186084929e-06, "loss": 0.4333, "step": 362 }, { "epoch": 0.03532846715328467, "grad_norm": 1.8248617929879183, "learning_rate": 9.999275605705457e-06, "loss": 0.4985, "step": 363 }, { "epoch": 0.035425790754257905, "grad_norm": 2.5923075514224982, "learning_rate": 9.999248528544895e-06, "loss": 0.4829, "step": 364 }, { "epoch": 0.035523114355231145, "grad_norm": 1.9900801938638135, "learning_rate": 9.999220954605932e-06, "loss": 0.587, "step": 365 }, { "epoch": 0.03562043795620438, "grad_norm": 1.731558772897005, "learning_rate": 9.999192883891314e-06, "loss": 0.3299, "step": 366 }, { "epoch": 0.03571776155717762, "grad_norm": 2.339577788711278, "learning_rate": 9.999164316403823e-06, "loss": 0.4845, "step": 367 }, { "epoch": 0.03581508515815085, "grad_norm": 1.9784113864985955, "learning_rate": 9.999135252146302e-06, "loss": 0.5776, "step": 368 }, { "epoch": 0.03591240875912409, "grad_norm": 1.5555461256937277, "learning_rate": 9.999105691121638e-06, "loss": 0.3563, "step": 369 }, { "epoch": 0.03600973236009732, "grad_norm": 1.7905677559908044, "learning_rate": 9.99907563333277e-06, "loss": 0.546, "step": 370 }, { "epoch": 0.03610705596107056, "grad_norm": 2.0490894714600287, "learning_rate": 9.999045078782684e-06, "loss": 0.6836, "step": 371 }, { "epoch": 0.036204379562043795, "grad_norm": 2.216601446334751, "learning_rate": 9.999014027474413e-06, "loss": 0.5237, "step": 372 }, { "epoch": 0.036301703163017035, "grad_norm": 1.5937926342815392, "learning_rate": 9.998982479411047e-06, "loss": 0.3539, "step": 373 }, { "epoch": 0.03639902676399027, "grad_norm": 2.3941848280266864, "learning_rate": 9.99895043459572e-06, "loss": 0.6249, "step": 374 }, { "epoch": 0.0364963503649635, "grad_norm": 2.072859669066288, "learning_rate": 9.998917893031615e-06, "loss": 0.5415, "step": 375 }, { "epoch": 0.03659367396593674, "grad_norm": 1.670908711065728, "learning_rate": 9.998884854721968e-06, "loss": 0.3034, "step": 376 }, { "epoch": 0.03669099756690997, "grad_norm": 1.9880303784818283, "learning_rate": 9.998851319670057e-06, "loss": 0.5025, "step": 377 }, { "epoch": 0.03678832116788321, "grad_norm": 1.3517666701087396, "learning_rate": 9.99881728787922e-06, "loss": 0.2775, "step": 378 }, { "epoch": 0.036885644768856446, "grad_norm": 1.8952553535268069, "learning_rate": 9.998782759352839e-06, "loss": 0.5306, "step": 379 }, { "epoch": 0.036982968369829686, "grad_norm": 1.8730537486024816, "learning_rate": 9.998747734094338e-06, "loss": 0.386, "step": 380 }, { "epoch": 0.03708029197080292, "grad_norm": 2.058996056292158, "learning_rate": 9.998712212107205e-06, "loss": 0.5641, "step": 381 }, { "epoch": 0.03717761557177616, "grad_norm": 1.9837834234853275, "learning_rate": 9.998676193394966e-06, "loss": 0.2628, "step": 382 }, { "epoch": 0.03727493917274939, "grad_norm": 2.189700953999047, "learning_rate": 9.998639677961203e-06, "loss": 0.6024, "step": 383 }, { "epoch": 0.03737226277372263, "grad_norm": 2.060696593716547, "learning_rate": 9.99860266580954e-06, "loss": 0.5377, "step": 384 }, { "epoch": 0.037469586374695864, "grad_norm": 2.0831966609629227, "learning_rate": 9.99856515694366e-06, "loss": 0.5063, "step": 385 }, { "epoch": 0.0375669099756691, "grad_norm": 2.2950496556846227, "learning_rate": 9.998527151367288e-06, "loss": 0.6484, "step": 386 }, { "epoch": 0.037664233576642336, "grad_norm": 2.2597922123273873, "learning_rate": 9.9984886490842e-06, "loss": 0.6617, "step": 387 }, { "epoch": 0.03776155717761557, "grad_norm": 2.071575887731456, "learning_rate": 9.99844965009822e-06, "loss": 0.5405, "step": 388 }, { "epoch": 0.03785888077858881, "grad_norm": 2.004249587957457, "learning_rate": 9.99841015441323e-06, "loss": 0.4306, "step": 389 }, { "epoch": 0.03795620437956204, "grad_norm": 1.9297023880727862, "learning_rate": 9.99837016203315e-06, "loss": 0.4083, "step": 390 }, { "epoch": 0.03805352798053528, "grad_norm": 2.001337081282171, "learning_rate": 9.998329672961952e-06, "loss": 0.4999, "step": 391 }, { "epoch": 0.038150851581508514, "grad_norm": 1.7630230797021285, "learning_rate": 9.998288687203665e-06, "loss": 0.4267, "step": 392 }, { "epoch": 0.038248175182481754, "grad_norm": 1.4413546421147376, "learning_rate": 9.998247204762358e-06, "loss": 0.3028, "step": 393 }, { "epoch": 0.03834549878345499, "grad_norm": 2.032450629241147, "learning_rate": 9.998205225642154e-06, "loss": 0.4216, "step": 394 }, { "epoch": 0.03844282238442823, "grad_norm": 1.8288270303352272, "learning_rate": 9.998162749847224e-06, "loss": 0.451, "step": 395 }, { "epoch": 0.03854014598540146, "grad_norm": 1.5869427581540143, "learning_rate": 9.998119777381791e-06, "loss": 0.4896, "step": 396 }, { "epoch": 0.03863746958637469, "grad_norm": 1.9312614168983935, "learning_rate": 9.998076308250122e-06, "loss": 0.351, "step": 397 }, { "epoch": 0.03873479318734793, "grad_norm": 2.182734939846557, "learning_rate": 9.99803234245654e-06, "loss": 0.4456, "step": 398 }, { "epoch": 0.038832116788321165, "grad_norm": 1.6075130172605856, "learning_rate": 9.997987880005412e-06, "loss": 0.3333, "step": 399 }, { "epoch": 0.038929440389294405, "grad_norm": 2.0206579020801048, "learning_rate": 9.997942920901154e-06, "loss": 0.4662, "step": 400 }, { "epoch": 0.03902676399026764, "grad_norm": 2.0019154912621246, "learning_rate": 9.997897465148236e-06, "loss": 0.588, "step": 401 }, { "epoch": 0.03912408759124088, "grad_norm": 1.9556688755730123, "learning_rate": 9.997851512751178e-06, "loss": 0.5364, "step": 402 }, { "epoch": 0.03922141119221411, "grad_norm": 2.1735940620422687, "learning_rate": 9.997805063714541e-06, "loss": 0.4155, "step": 403 }, { "epoch": 0.03931873479318735, "grad_norm": 1.893104755523836, "learning_rate": 9.997758118042945e-06, "loss": 0.2835, "step": 404 }, { "epoch": 0.03941605839416058, "grad_norm": 1.892857392200546, "learning_rate": 9.99771067574105e-06, "loss": 0.317, "step": 405 }, { "epoch": 0.03951338199513382, "grad_norm": 2.194365925195629, "learning_rate": 9.997662736813575e-06, "loss": 0.5972, "step": 406 }, { "epoch": 0.039610705596107056, "grad_norm": 2.3359516870584547, "learning_rate": 9.997614301265281e-06, "loss": 0.3505, "step": 407 }, { "epoch": 0.039708029197080295, "grad_norm": 1.8041349283411827, "learning_rate": 9.997565369100983e-06, "loss": 0.4003, "step": 408 }, { "epoch": 0.03980535279805353, "grad_norm": 2.2199870140108273, "learning_rate": 9.997515940325542e-06, "loss": 0.4428, "step": 409 }, { "epoch": 0.03990267639902676, "grad_norm": 2.193796849633566, "learning_rate": 9.997466014943871e-06, "loss": 0.3906, "step": 410 }, { "epoch": 0.04, "grad_norm": 2.7309920828616168, "learning_rate": 9.99741559296093e-06, "loss": 0.6283, "step": 411 }, { "epoch": 0.040097323600973234, "grad_norm": 2.220745639846989, "learning_rate": 9.99736467438173e-06, "loss": 0.4568, "step": 412 }, { "epoch": 0.04019464720194647, "grad_norm": 1.905067765139487, "learning_rate": 9.99731325921133e-06, "loss": 0.3198, "step": 413 }, { "epoch": 0.040291970802919706, "grad_norm": 2.0461180940034116, "learning_rate": 9.997261347454841e-06, "loss": 0.3783, "step": 414 }, { "epoch": 0.040389294403892946, "grad_norm": 1.9732614929529544, "learning_rate": 9.99720893911742e-06, "loss": 0.5211, "step": 415 }, { "epoch": 0.04048661800486618, "grad_norm": 2.341156401873798, "learning_rate": 9.997156034204276e-06, "loss": 0.5094, "step": 416 }, { "epoch": 0.04058394160583942, "grad_norm": 2.2588135503158138, "learning_rate": 9.997102632720664e-06, "loss": 0.591, "step": 417 }, { "epoch": 0.04068126520681265, "grad_norm": 2.187795564574772, "learning_rate": 9.997048734671893e-06, "loss": 0.3811, "step": 418 }, { "epoch": 0.04077858880778589, "grad_norm": 2.2570398189900938, "learning_rate": 9.996994340063314e-06, "loss": 0.4494, "step": 419 }, { "epoch": 0.040875912408759124, "grad_norm": 2.3267878846596597, "learning_rate": 9.996939448900341e-06, "loss": 0.5254, "step": 420 }, { "epoch": 0.04097323600973236, "grad_norm": 1.9149387144635641, "learning_rate": 9.99688406118842e-06, "loss": 0.4281, "step": 421 }, { "epoch": 0.0410705596107056, "grad_norm": 2.4052095021382285, "learning_rate": 9.996828176933062e-06, "loss": 0.61, "step": 422 }, { "epoch": 0.04116788321167883, "grad_norm": 2.8744864627123237, "learning_rate": 9.996771796139814e-06, "loss": 0.4708, "step": 423 }, { "epoch": 0.04126520681265207, "grad_norm": 2.0334953222734513, "learning_rate": 9.996714918814284e-06, "loss": 0.2697, "step": 424 }, { "epoch": 0.0413625304136253, "grad_norm": 2.1314093477075486, "learning_rate": 9.996657544962119e-06, "loss": 0.3026, "step": 425 }, { "epoch": 0.04145985401459854, "grad_norm": 1.7241742631767316, "learning_rate": 9.996599674589022e-06, "loss": 0.3624, "step": 426 }, { "epoch": 0.041557177615571775, "grad_norm": 2.417754377928955, "learning_rate": 9.996541307700746e-06, "loss": 0.6682, "step": 427 }, { "epoch": 0.041654501216545015, "grad_norm": 2.2126055245100256, "learning_rate": 9.99648244430309e-06, "loss": 0.3705, "step": 428 }, { "epoch": 0.04175182481751825, "grad_norm": 1.8224510106748588, "learning_rate": 9.996423084401901e-06, "loss": 0.4318, "step": 429 }, { "epoch": 0.04184914841849149, "grad_norm": 1.6786428352287364, "learning_rate": 9.996363228003079e-06, "loss": 0.4662, "step": 430 }, { "epoch": 0.04194647201946472, "grad_norm": 1.9342922605897592, "learning_rate": 9.99630287511257e-06, "loss": 0.4874, "step": 431 }, { "epoch": 0.04204379562043795, "grad_norm": 1.9444011100602645, "learning_rate": 9.996242025736377e-06, "loss": 0.3711, "step": 432 }, { "epoch": 0.04214111922141119, "grad_norm": 3.114184163688958, "learning_rate": 9.99618067988054e-06, "loss": 0.5342, "step": 433 }, { "epoch": 0.042238442822384425, "grad_norm": 1.993932460938173, "learning_rate": 9.99611883755116e-06, "loss": 0.465, "step": 434 }, { "epoch": 0.042335766423357665, "grad_norm": 1.5062408953506277, "learning_rate": 9.99605649875438e-06, "loss": 0.3862, "step": 435 }, { "epoch": 0.0424330900243309, "grad_norm": 2.5287447175721733, "learning_rate": 9.995993663496394e-06, "loss": 0.5638, "step": 436 }, { "epoch": 0.04253041362530414, "grad_norm": 1.7215400937807486, "learning_rate": 9.995930331783448e-06, "loss": 0.3507, "step": 437 }, { "epoch": 0.04262773722627737, "grad_norm": 1.5105936757865817, "learning_rate": 9.995866503621834e-06, "loss": 0.4086, "step": 438 }, { "epoch": 0.04272506082725061, "grad_norm": 1.828501540310894, "learning_rate": 9.995802179017893e-06, "loss": 0.3477, "step": 439 }, { "epoch": 0.04282238442822384, "grad_norm": 1.6658361590948114, "learning_rate": 9.995737357978022e-06, "loss": 0.4006, "step": 440 }, { "epoch": 0.04291970802919708, "grad_norm": 1.6434395036324305, "learning_rate": 9.995672040508656e-06, "loss": 0.4349, "step": 441 }, { "epoch": 0.043017031630170316, "grad_norm": 1.9913424027071125, "learning_rate": 9.99560622661629e-06, "loss": 0.3415, "step": 442 }, { "epoch": 0.043114355231143556, "grad_norm": 1.6487474195389296, "learning_rate": 9.995539916307463e-06, "loss": 0.4804, "step": 443 }, { "epoch": 0.04321167883211679, "grad_norm": 1.4861266391850032, "learning_rate": 9.995473109588764e-06, "loss": 0.411, "step": 444 }, { "epoch": 0.04330900243309002, "grad_norm": 1.4390762643228305, "learning_rate": 9.995405806466831e-06, "loss": 0.3806, "step": 445 }, { "epoch": 0.04340632603406326, "grad_norm": 1.7775332171720517, "learning_rate": 9.995338006948353e-06, "loss": 0.3332, "step": 446 }, { "epoch": 0.043503649635036494, "grad_norm": 1.7312883283317864, "learning_rate": 9.995269711040067e-06, "loss": 0.2736, "step": 447 }, { "epoch": 0.043600973236009734, "grad_norm": 1.7973901424872405, "learning_rate": 9.995200918748759e-06, "loss": 0.5597, "step": 448 }, { "epoch": 0.04369829683698297, "grad_norm": 2.0409413301370334, "learning_rate": 9.995131630081265e-06, "loss": 0.6045, "step": 449 }, { "epoch": 0.043795620437956206, "grad_norm": 3.2708903670147347, "learning_rate": 9.995061845044473e-06, "loss": 0.6245, "step": 450 }, { "epoch": 0.04389294403892944, "grad_norm": 1.744466889932859, "learning_rate": 9.994991563645314e-06, "loss": 0.4129, "step": 451 }, { "epoch": 0.04399026763990268, "grad_norm": 1.8775864246251477, "learning_rate": 9.994920785890771e-06, "loss": 0.414, "step": 452 }, { "epoch": 0.04408759124087591, "grad_norm": 1.3868286948878126, "learning_rate": 9.994849511787881e-06, "loss": 0.3164, "step": 453 }, { "epoch": 0.04418491484184915, "grad_norm": 1.6888257223301795, "learning_rate": 9.994777741343727e-06, "loss": 0.3241, "step": 454 }, { "epoch": 0.044282238442822384, "grad_norm": 1.5029594314338663, "learning_rate": 9.994705474565436e-06, "loss": 0.4148, "step": 455 }, { "epoch": 0.04437956204379562, "grad_norm": 1.7159996915963702, "learning_rate": 9.994632711460193e-06, "loss": 0.3387, "step": 456 }, { "epoch": 0.04447688564476886, "grad_norm": 1.7717997513120352, "learning_rate": 9.994559452035228e-06, "loss": 0.4547, "step": 457 }, { "epoch": 0.04457420924574209, "grad_norm": 1.887765282184233, "learning_rate": 9.99448569629782e-06, "loss": 0.5919, "step": 458 }, { "epoch": 0.04467153284671533, "grad_norm": 2.0151049512314585, "learning_rate": 9.994411444255298e-06, "loss": 0.4556, "step": 459 }, { "epoch": 0.04476885644768856, "grad_norm": 1.5706463359289826, "learning_rate": 9.994336695915041e-06, "loss": 0.3443, "step": 460 }, { "epoch": 0.0448661800486618, "grad_norm": 1.9067884841542395, "learning_rate": 9.994261451284477e-06, "loss": 0.5862, "step": 461 }, { "epoch": 0.044963503649635035, "grad_norm": 1.7346846845298518, "learning_rate": 9.994185710371083e-06, "loss": 0.3588, "step": 462 }, { "epoch": 0.045060827250608275, "grad_norm": 1.5593715629463312, "learning_rate": 9.994109473182385e-06, "loss": 0.2891, "step": 463 }, { "epoch": 0.04515815085158151, "grad_norm": 2.326736753149576, "learning_rate": 9.994032739725959e-06, "loss": 0.6517, "step": 464 }, { "epoch": 0.04525547445255475, "grad_norm": 2.2142852132770305, "learning_rate": 9.99395551000943e-06, "loss": 0.3571, "step": 465 }, { "epoch": 0.04535279805352798, "grad_norm": 1.7351954813390544, "learning_rate": 9.993877784040474e-06, "loss": 0.3849, "step": 466 }, { "epoch": 0.04545012165450121, "grad_norm": 1.3962336815381617, "learning_rate": 9.993799561826811e-06, "loss": 0.311, "step": 467 }, { "epoch": 0.04554744525547445, "grad_norm": 1.878958465421645, "learning_rate": 9.993720843376216e-06, "loss": 0.5602, "step": 468 }, { "epoch": 0.045644768856447686, "grad_norm": 1.519160992933857, "learning_rate": 9.993641628696513e-06, "loss": 0.2379, "step": 469 }, { "epoch": 0.045742092457420926, "grad_norm": 2.5345930464298885, "learning_rate": 9.99356191779557e-06, "loss": 0.4239, "step": 470 }, { "epoch": 0.04583941605839416, "grad_norm": 1.3153911718041251, "learning_rate": 9.993481710681314e-06, "loss": 0.3454, "step": 471 }, { "epoch": 0.0459367396593674, "grad_norm": 2.16208125563947, "learning_rate": 9.993401007361707e-06, "loss": 0.5386, "step": 472 }, { "epoch": 0.04603406326034063, "grad_norm": 1.8150842593472827, "learning_rate": 9.993319807844775e-06, "loss": 0.3077, "step": 473 }, { "epoch": 0.04613138686131387, "grad_norm": 1.6656864462678063, "learning_rate": 9.993238112138584e-06, "loss": 0.4927, "step": 474 }, { "epoch": 0.046228710462287104, "grad_norm": 1.3429917702468868, "learning_rate": 9.993155920251252e-06, "loss": 0.2433, "step": 475 }, { "epoch": 0.04632603406326034, "grad_norm": 1.3651155739367906, "learning_rate": 9.993073232190949e-06, "loss": 0.2947, "step": 476 }, { "epoch": 0.046423357664233576, "grad_norm": 1.7815516701613203, "learning_rate": 9.992990047965887e-06, "loss": 0.5372, "step": 477 }, { "epoch": 0.046520681265206816, "grad_norm": 1.846696342179327, "learning_rate": 9.992906367584337e-06, "loss": 0.5127, "step": 478 }, { "epoch": 0.04661800486618005, "grad_norm": 1.7511253825578088, "learning_rate": 9.992822191054612e-06, "loss": 0.4074, "step": 479 }, { "epoch": 0.04671532846715328, "grad_norm": 1.8105635986872588, "learning_rate": 9.992737518385076e-06, "loss": 0.4998, "step": 480 }, { "epoch": 0.04681265206812652, "grad_norm": 2.2743597617900746, "learning_rate": 9.992652349584147e-06, "loss": 0.6249, "step": 481 }, { "epoch": 0.046909975669099754, "grad_norm": 1.93948496382319, "learning_rate": 9.992566684660282e-06, "loss": 0.5411, "step": 482 }, { "epoch": 0.047007299270072994, "grad_norm": 1.4073760716303516, "learning_rate": 9.992480523621999e-06, "loss": 0.3506, "step": 483 }, { "epoch": 0.04710462287104623, "grad_norm": 1.388293079160528, "learning_rate": 9.992393866477856e-06, "loss": 0.3304, "step": 484 }, { "epoch": 0.04720194647201947, "grad_norm": 2.082643572745618, "learning_rate": 9.992306713236467e-06, "loss": 0.5653, "step": 485 }, { "epoch": 0.0472992700729927, "grad_norm": 1.7104664332606834, "learning_rate": 9.992219063906492e-06, "loss": 0.3317, "step": 486 }, { "epoch": 0.04739659367396594, "grad_norm": 1.7575848919482624, "learning_rate": 9.992130918496638e-06, "loss": 0.4109, "step": 487 }, { "epoch": 0.04749391727493917, "grad_norm": 1.7351379091271637, "learning_rate": 9.992042277015668e-06, "loss": 0.5065, "step": 488 }, { "epoch": 0.04759124087591241, "grad_norm": 1.4444570948381004, "learning_rate": 9.991953139472387e-06, "loss": 0.4023, "step": 489 }, { "epoch": 0.047688564476885645, "grad_norm": 1.4697709289140384, "learning_rate": 9.991863505875656e-06, "loss": 0.3364, "step": 490 }, { "epoch": 0.04778588807785888, "grad_norm": 1.9428205960506804, "learning_rate": 9.99177337623438e-06, "loss": 0.4303, "step": 491 }, { "epoch": 0.04788321167883212, "grad_norm": 1.931152158561148, "learning_rate": 9.991682750557516e-06, "loss": 0.2857, "step": 492 }, { "epoch": 0.04798053527980535, "grad_norm": 1.9301394655308035, "learning_rate": 9.991591628854067e-06, "loss": 0.5998, "step": 493 }, { "epoch": 0.04807785888077859, "grad_norm": 1.7788293016868693, "learning_rate": 9.99150001113309e-06, "loss": 0.4595, "step": 494 }, { "epoch": 0.04817518248175182, "grad_norm": 2.0641225732440134, "learning_rate": 9.99140789740369e-06, "loss": 0.3848, "step": 495 }, { "epoch": 0.04827250608272506, "grad_norm": 2.2832955373527044, "learning_rate": 9.99131528767502e-06, "loss": 0.6396, "step": 496 }, { "epoch": 0.048369829683698295, "grad_norm": 1.6658790952812916, "learning_rate": 9.99122218195628e-06, "loss": 0.5429, "step": 497 }, { "epoch": 0.048467153284671535, "grad_norm": 1.6568038302360257, "learning_rate": 9.991128580256725e-06, "loss": 0.4532, "step": 498 }, { "epoch": 0.04856447688564477, "grad_norm": 1.8451659374514144, "learning_rate": 9.991034482585656e-06, "loss": 0.5845, "step": 499 }, { "epoch": 0.04866180048661801, "grad_norm": 1.9103948838029656, "learning_rate": 9.99093988895242e-06, "loss": 0.5508, "step": 500 }, { "epoch": 0.04875912408759124, "grad_norm": 1.9691733858712537, "learning_rate": 9.990844799366422e-06, "loss": 0.6374, "step": 501 }, { "epoch": 0.048856447688564474, "grad_norm": 2.1278472226161846, "learning_rate": 9.990749213837108e-06, "loss": 0.572, "step": 502 }, { "epoch": 0.04895377128953771, "grad_norm": 1.9704028865885994, "learning_rate": 9.990653132373977e-06, "loss": 0.6282, "step": 503 }, { "epoch": 0.049051094890510946, "grad_norm": 1.8965741341561362, "learning_rate": 9.990556554986577e-06, "loss": 0.5749, "step": 504 }, { "epoch": 0.049148418491484186, "grad_norm": 1.5425018763105707, "learning_rate": 9.990459481684504e-06, "loss": 0.4236, "step": 505 }, { "epoch": 0.04924574209245742, "grad_norm": 1.736669998068125, "learning_rate": 9.990361912477405e-06, "loss": 0.4275, "step": 506 }, { "epoch": 0.04934306569343066, "grad_norm": 2.049335776858506, "learning_rate": 9.990263847374976e-06, "loss": 0.6897, "step": 507 }, { "epoch": 0.04944038929440389, "grad_norm": 1.8544975871268152, "learning_rate": 9.990165286386961e-06, "loss": 0.4811, "step": 508 }, { "epoch": 0.04953771289537713, "grad_norm": 1.5709178763522822, "learning_rate": 9.990066229523155e-06, "loss": 0.4585, "step": 509 }, { "epoch": 0.049635036496350364, "grad_norm": 2.1410068811754153, "learning_rate": 9.989966676793399e-06, "loss": 0.4773, "step": 510 }, { "epoch": 0.049732360097323604, "grad_norm": 1.760724042734433, "learning_rate": 9.989866628207589e-06, "loss": 0.3144, "step": 511 }, { "epoch": 0.04982968369829684, "grad_norm": 1.8521560168370175, "learning_rate": 9.989766083775662e-06, "loss": 0.4656, "step": 512 }, { "epoch": 0.049927007299270076, "grad_norm": 1.544987615640627, "learning_rate": 9.989665043507616e-06, "loss": 0.4089, "step": 513 }, { "epoch": 0.05002433090024331, "grad_norm": 1.9122960249889975, "learning_rate": 9.989563507413487e-06, "loss": 0.4535, "step": 514 }, { "epoch": 0.05012165450121654, "grad_norm": 1.5187134098621655, "learning_rate": 9.989461475503363e-06, "loss": 0.31, "step": 515 } ], "logging_steps": 1.0, "max_steps": 10275, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 515, "total_flos": 65890827296768.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }