{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 2199, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00136542072025943, "grad_norm": 3.8247339725494385, "learning_rate": 0.0, "loss": 2.0623, "num_tokens": 810917.0, "step": 1 }, { "epoch": 0.00273084144051886, "grad_norm": 3.6301887035369873, "learning_rate": 6.060606060606061e-07, "loss": 2.1473, "num_tokens": 1637352.0, "step": 2 }, { "epoch": 0.00409626216077829, "grad_norm": 3.466407299041748, "learning_rate": 1.2121212121212122e-06, "loss": 1.9383, "num_tokens": 2400944.0, "step": 3 }, { "epoch": 0.00546168288103772, "grad_norm": 2.7405834197998047, "learning_rate": 1.8181818181818183e-06, "loss": 1.6934, "num_tokens": 3074555.0, "step": 4 }, { "epoch": 0.006827103601297149, "grad_norm": 3.5803465843200684, "learning_rate": 2.4242424242424244e-06, "loss": 2.0142, "num_tokens": 3896440.0, "step": 5 }, { "epoch": 0.00819252432155658, "grad_norm": 3.3620965480804443, "learning_rate": 3.0303030303030305e-06, "loss": 2.0289, "num_tokens": 4696744.0, "step": 6 }, { "epoch": 0.009557945041816009, "grad_norm": 3.392517328262329, "learning_rate": 3.6363636363636366e-06, "loss": 2.0567, "num_tokens": 5449532.0, "step": 7 }, { "epoch": 0.01092336576207544, "grad_norm": 3.623455286026001, "learning_rate": 4.242424242424243e-06, "loss": 2.1689, "num_tokens": 6279890.0, "step": 8 }, { "epoch": 0.01228878648233487, "grad_norm": 3.158599376678467, "learning_rate": 4.848484848484849e-06, "loss": 2.1514, "num_tokens": 7141872.0, "step": 9 }, { "epoch": 0.013654207202594299, "grad_norm": 2.516094207763672, "learning_rate": 5.4545454545454545e-06, "loss": 1.8426, "num_tokens": 7944488.0, "step": 10 }, { "epoch": 0.01501962792285373, "grad_norm": 2.396498918533325, "learning_rate": 6.060606060606061e-06, "loss": 1.829, "num_tokens": 8687257.0, "step": 11 }, { "epoch": 0.01638504864311316, "grad_norm": 1.9637004137039185, "learning_rate": 6.666666666666667e-06, "loss": 1.7448, "num_tokens": 9460642.0, "step": 12 }, { "epoch": 0.01775046936337259, "grad_norm": 1.7760802507400513, "learning_rate": 7.272727272727273e-06, "loss": 1.6932, "num_tokens": 10243996.0, "step": 13 }, { "epoch": 0.019115890083632018, "grad_norm": 1.5945855379104614, "learning_rate": 7.87878787878788e-06, "loss": 1.6004, "num_tokens": 10963865.0, "step": 14 }, { "epoch": 0.02048131080389145, "grad_norm": 1.715968132019043, "learning_rate": 8.484848484848486e-06, "loss": 1.7994, "num_tokens": 11786756.0, "step": 15 }, { "epoch": 0.02184673152415088, "grad_norm": 1.2287795543670654, "learning_rate": 9.090909090909091e-06, "loss": 1.4206, "num_tokens": 12463989.0, "step": 16 }, { "epoch": 0.023212152244410308, "grad_norm": 1.5592275857925415, "learning_rate": 9.696969696969698e-06, "loss": 1.6503, "num_tokens": 13328035.0, "step": 17 }, { "epoch": 0.02457757296466974, "grad_norm": 1.3033058643341064, "learning_rate": 1.0303030303030304e-05, "loss": 1.3737, "num_tokens": 14090112.0, "step": 18 }, { "epoch": 0.02594299368492917, "grad_norm": 1.3272504806518555, "learning_rate": 1.0909090909090909e-05, "loss": 1.4976, "num_tokens": 14858605.0, "step": 19 }, { "epoch": 0.027308414405188598, "grad_norm": 1.2131528854370117, "learning_rate": 1.1515151515151517e-05, "loss": 1.5291, "num_tokens": 15678311.0, "step": 20 }, { "epoch": 0.02867383512544803, "grad_norm": 1.0684608221054077, "learning_rate": 1.2121212121212122e-05, "loss": 1.5024, "num_tokens": 16492154.0, "step": 21 }, { "epoch": 0.03003925584570746, "grad_norm": 1.0076466798782349, "learning_rate": 1.2727272727272728e-05, "loss": 1.4566, "num_tokens": 17329278.0, "step": 22 }, { "epoch": 0.03140467656596689, "grad_norm": 0.9865179061889648, "learning_rate": 1.3333333333333333e-05, "loss": 1.1633, "num_tokens": 18047779.0, "step": 23 }, { "epoch": 0.03277009728622632, "grad_norm": 0.8135017156600952, "learning_rate": 1.3939393939393942e-05, "loss": 1.2537, "num_tokens": 18845314.0, "step": 24 }, { "epoch": 0.034135518006485746, "grad_norm": 0.7201207280158997, "learning_rate": 1.4545454545454546e-05, "loss": 1.1672, "num_tokens": 19608652.0, "step": 25 }, { "epoch": 0.03550093872674518, "grad_norm": 0.7128145098686218, "learning_rate": 1.5151515151515153e-05, "loss": 1.187, "num_tokens": 20391088.0, "step": 26 }, { "epoch": 0.03686635944700461, "grad_norm": 0.7086578607559204, "learning_rate": 1.575757575757576e-05, "loss": 1.1558, "num_tokens": 21209788.0, "step": 27 }, { "epoch": 0.038231780167264036, "grad_norm": 1.3576252460479736, "learning_rate": 1.6363636363636366e-05, "loss": 1.069, "num_tokens": 21957235.0, "step": 28 }, { "epoch": 0.03959720088752347, "grad_norm": 0.529454231262207, "learning_rate": 1.6969696969696972e-05, "loss": 1.0148, "num_tokens": 22717810.0, "step": 29 }, { "epoch": 0.0409626216077829, "grad_norm": 0.48387089371681213, "learning_rate": 1.7575757575757576e-05, "loss": 1.0073, "num_tokens": 23483922.0, "step": 30 }, { "epoch": 0.042328042328042326, "grad_norm": 0.5119532346725464, "learning_rate": 1.8181818181818182e-05, "loss": 1.064, "num_tokens": 24260390.0, "step": 31 }, { "epoch": 0.04369346304830176, "grad_norm": 0.5288600325584412, "learning_rate": 1.8787878787878792e-05, "loss": 1.1868, "num_tokens": 25131243.0, "step": 32 }, { "epoch": 0.04505888376856119, "grad_norm": 0.4310317635536194, "learning_rate": 1.9393939393939395e-05, "loss": 1.0206, "num_tokens": 25931856.0, "step": 33 }, { "epoch": 0.046424304488820615, "grad_norm": 0.4431857168674469, "learning_rate": 2e-05, "loss": 1.0552, "num_tokens": 26778000.0, "step": 34 }, { "epoch": 0.04778972520908005, "grad_norm": 0.44794008135795593, "learning_rate": 2.0606060606060608e-05, "loss": 1.0103, "num_tokens": 27622031.0, "step": 35 }, { "epoch": 0.04915514592933948, "grad_norm": 0.3783922791481018, "learning_rate": 2.121212121212121e-05, "loss": 0.8411, "num_tokens": 28357435.0, "step": 36 }, { "epoch": 0.050520566649598905, "grad_norm": 0.366346538066864, "learning_rate": 2.1818181818181818e-05, "loss": 0.8788, "num_tokens": 29103274.0, "step": 37 }, { "epoch": 0.05188598736985834, "grad_norm": 0.374860554933548, "learning_rate": 2.2424242424242424e-05, "loss": 0.9103, "num_tokens": 29917333.0, "step": 38 }, { "epoch": 0.05325140809011777, "grad_norm": 0.3877836763858795, "learning_rate": 2.3030303030303034e-05, "loss": 0.9096, "num_tokens": 30783071.0, "step": 39 }, { "epoch": 0.054616828810377195, "grad_norm": 0.32811984419822693, "learning_rate": 2.363636363636364e-05, "loss": 0.83, "num_tokens": 31543642.0, "step": 40 }, { "epoch": 0.05598224953063663, "grad_norm": 0.30647575855255127, "learning_rate": 2.4242424242424244e-05, "loss": 0.8248, "num_tokens": 32329323.0, "step": 41 }, { "epoch": 0.05734767025089606, "grad_norm": 0.3121391236782074, "learning_rate": 2.484848484848485e-05, "loss": 0.8266, "num_tokens": 33125319.0, "step": 42 }, { "epoch": 0.058713090971155485, "grad_norm": 0.30968454480171204, "learning_rate": 2.5454545454545457e-05, "loss": 0.8373, "num_tokens": 33954951.0, "step": 43 }, { "epoch": 0.06007851169141492, "grad_norm": 0.25962692499160767, "learning_rate": 2.6060606060606063e-05, "loss": 0.7813, "num_tokens": 34727155.0, "step": 44 }, { "epoch": 0.06144393241167435, "grad_norm": 0.2479821741580963, "learning_rate": 2.6666666666666667e-05, "loss": 0.7432, "num_tokens": 35495619.0, "step": 45 }, { "epoch": 0.06280935313193378, "grad_norm": 0.23943600058555603, "learning_rate": 2.7272727272727273e-05, "loss": 0.729, "num_tokens": 36253648.0, "step": 46 }, { "epoch": 0.06417477385219321, "grad_norm": 0.24141737818717957, "learning_rate": 2.7878787878787883e-05, "loss": 0.742, "num_tokens": 37013786.0, "step": 47 }, { "epoch": 0.06554019457245264, "grad_norm": 0.27584323287010193, "learning_rate": 2.848484848484849e-05, "loss": 0.7498, "num_tokens": 37777217.0, "step": 48 }, { "epoch": 0.06690561529271206, "grad_norm": 0.23000292479991913, "learning_rate": 2.9090909090909093e-05, "loss": 0.7578, "num_tokens": 38609357.0, "step": 49 }, { "epoch": 0.06827103601297149, "grad_norm": 0.22273904085159302, "learning_rate": 2.96969696969697e-05, "loss": 0.6681, "num_tokens": 39364320.0, "step": 50 }, { "epoch": 0.06963645673323093, "grad_norm": 0.23272264003753662, "learning_rate": 3.0303030303030306e-05, "loss": 0.7417, "num_tokens": 40168688.0, "step": 51 }, { "epoch": 0.07100187745349036, "grad_norm": 0.18300127983093262, "learning_rate": 3.090909090909091e-05, "loss": 0.6284, "num_tokens": 40887680.0, "step": 52 }, { "epoch": 0.07236729817374979, "grad_norm": 0.21360281109809875, "learning_rate": 3.151515151515152e-05, "loss": 0.7153, "num_tokens": 41717726.0, "step": 53 }, { "epoch": 0.07373271889400922, "grad_norm": 0.21551239490509033, "learning_rate": 3.212121212121212e-05, "loss": 0.6743, "num_tokens": 42491245.0, "step": 54 }, { "epoch": 0.07509813961426864, "grad_norm": 0.397121787071228, "learning_rate": 3.272727272727273e-05, "loss": 0.725, "num_tokens": 43374525.0, "step": 55 }, { "epoch": 0.07646356033452807, "grad_norm": 0.2243133932352066, "learning_rate": 3.3333333333333335e-05, "loss": 0.6478, "num_tokens": 44119621.0, "step": 56 }, { "epoch": 0.07782898105478751, "grad_norm": 0.1898639053106308, "learning_rate": 3.3939393939393945e-05, "loss": 0.6396, "num_tokens": 44884679.0, "step": 57 }, { "epoch": 0.07919440177504694, "grad_norm": 0.1998988837003708, "learning_rate": 3.454545454545455e-05, "loss": 0.6668, "num_tokens": 45638858.0, "step": 58 }, { "epoch": 0.08055982249530637, "grad_norm": 0.19098971784114838, "learning_rate": 3.515151515151515e-05, "loss": 0.6443, "num_tokens": 46418214.0, "step": 59 }, { "epoch": 0.0819252432155658, "grad_norm": 0.18564876914024353, "learning_rate": 3.575757575757576e-05, "loss": 0.6522, "num_tokens": 47224980.0, "step": 60 }, { "epoch": 0.08329066393582522, "grad_norm": 0.21455052495002747, "learning_rate": 3.6363636363636364e-05, "loss": 0.6391, "num_tokens": 47952103.0, "step": 61 }, { "epoch": 0.08465608465608465, "grad_norm": 0.1984366625547409, "learning_rate": 3.6969696969696974e-05, "loss": 0.6326, "num_tokens": 48790470.0, "step": 62 }, { "epoch": 0.08602150537634409, "grad_norm": 0.16422998905181885, "learning_rate": 3.7575757575757584e-05, "loss": 0.6489, "num_tokens": 49622523.0, "step": 63 }, { "epoch": 0.08738692609660352, "grad_norm": 0.19873517751693726, "learning_rate": 3.818181818181819e-05, "loss": 0.6284, "num_tokens": 50395407.0, "step": 64 }, { "epoch": 0.08875234681686295, "grad_norm": 0.1674010306596756, "learning_rate": 3.878787878787879e-05, "loss": 0.6249, "num_tokens": 51189042.0, "step": 65 }, { "epoch": 0.09011776753712238, "grad_norm": 0.20986004173755646, "learning_rate": 3.93939393939394e-05, "loss": 0.628, "num_tokens": 51945385.0, "step": 66 }, { "epoch": 0.0914831882573818, "grad_norm": 0.1877584308385849, "learning_rate": 4e-05, "loss": 0.5938, "num_tokens": 52685034.0, "step": 67 }, { "epoch": 0.09284860897764123, "grad_norm": 0.20853351056575775, "learning_rate": 3.999998047637464e-05, "loss": 0.6305, "num_tokens": 53493841.0, "step": 68 }, { "epoch": 0.09421402969790067, "grad_norm": 0.1890617311000824, "learning_rate": 3.99999219055409e-05, "loss": 0.5903, "num_tokens": 54207375.0, "step": 69 }, { "epoch": 0.0955794504181601, "grad_norm": 0.20418766140937805, "learning_rate": 3.9999824287625855e-05, "loss": 0.618, "num_tokens": 54969994.0, "step": 70 }, { "epoch": 0.09694487113841953, "grad_norm": 0.1915808618068695, "learning_rate": 3.999968762284124e-05, "loss": 0.5866, "num_tokens": 55762965.0, "step": 71 }, { "epoch": 0.09831029185867896, "grad_norm": 0.20797483623027802, "learning_rate": 3.999951191148356e-05, "loss": 0.6034, "num_tokens": 56561625.0, "step": 72 }, { "epoch": 0.09967571257893838, "grad_norm": 0.15858304500579834, "learning_rate": 3.999929715393395e-05, "loss": 0.5758, "num_tokens": 57354498.0, "step": 73 }, { "epoch": 0.10104113329919781, "grad_norm": 0.1874593198299408, "learning_rate": 3.999904335065829e-05, "loss": 0.6218, "num_tokens": 58152251.0, "step": 74 }, { "epoch": 0.10240655401945725, "grad_norm": 0.17945021390914917, "learning_rate": 3.999875050220715e-05, "loss": 0.6213, "num_tokens": 59011068.0, "step": 75 }, { "epoch": 0.10377197473971668, "grad_norm": 0.1742994785308838, "learning_rate": 3.9998418609215826e-05, "loss": 0.5526, "num_tokens": 59784522.0, "step": 76 }, { "epoch": 0.10513739545997611, "grad_norm": 0.1795405000448227, "learning_rate": 3.9998047672404265e-05, "loss": 0.5734, "num_tokens": 60587642.0, "step": 77 }, { "epoch": 0.10650281618023553, "grad_norm": 0.1702861189842224, "learning_rate": 3.9997637692577146e-05, "loss": 0.573, "num_tokens": 61327592.0, "step": 78 }, { "epoch": 0.10786823690049496, "grad_norm": 0.18214769661426544, "learning_rate": 3.999718867062384e-05, "loss": 0.5556, "num_tokens": 62099441.0, "step": 79 }, { "epoch": 0.10923365762075439, "grad_norm": 0.15847574174404144, "learning_rate": 3.99967006075184e-05, "loss": 0.5876, "num_tokens": 62888444.0, "step": 80 }, { "epoch": 0.11059907834101383, "grad_norm": 0.15977297723293304, "learning_rate": 3.999617350431958e-05, "loss": 0.5621, "num_tokens": 63682672.0, "step": 81 }, { "epoch": 0.11196449906127326, "grad_norm": 0.16256140172481537, "learning_rate": 3.9995607362170823e-05, "loss": 0.5618, "num_tokens": 64573108.0, "step": 82 }, { "epoch": 0.11332991978153269, "grad_norm": 0.15531837940216064, "learning_rate": 3.999500218230025e-05, "loss": 0.5712, "num_tokens": 65411389.0, "step": 83 }, { "epoch": 0.11469534050179211, "grad_norm": 0.17020969092845917, "learning_rate": 3.999435796602068e-05, "loss": 0.5551, "num_tokens": 66170606.0, "step": 84 }, { "epoch": 0.11606076122205154, "grad_norm": 0.1668613702058792, "learning_rate": 3.99936747147296e-05, "loss": 0.5662, "num_tokens": 66943810.0, "step": 85 }, { "epoch": 0.11742618194231097, "grad_norm": 0.16691303253173828, "learning_rate": 3.999295242990918e-05, "loss": 0.5685, "num_tokens": 67765182.0, "step": 86 }, { "epoch": 0.11879160266257041, "grad_norm": 0.19694039225578308, "learning_rate": 3.999219111312627e-05, "loss": 0.5793, "num_tokens": 68529643.0, "step": 87 }, { "epoch": 0.12015702338282984, "grad_norm": 0.1393001228570938, "learning_rate": 3.99913907660324e-05, "loss": 0.5627, "num_tokens": 69306401.0, "step": 88 }, { "epoch": 0.12152244410308927, "grad_norm": 0.20182672142982483, "learning_rate": 3.9990551390363735e-05, "loss": 0.5952, "num_tokens": 70150195.0, "step": 89 }, { "epoch": 0.1228878648233487, "grad_norm": 0.14460156857967377, "learning_rate": 3.998967298794113e-05, "loss": 0.5633, "num_tokens": 70946010.0, "step": 90 }, { "epoch": 0.12425328554360812, "grad_norm": 0.1627425104379654, "learning_rate": 3.998875556067011e-05, "loss": 0.5765, "num_tokens": 71769941.0, "step": 91 }, { "epoch": 0.12561870626386756, "grad_norm": 0.15942293405532837, "learning_rate": 3.9987799110540836e-05, "loss": 0.5781, "num_tokens": 72564346.0, "step": 92 }, { "epoch": 0.12698412698412698, "grad_norm": 0.17416979372501373, "learning_rate": 3.9986803639628124e-05, "loss": 0.5706, "num_tokens": 73334106.0, "step": 93 }, { "epoch": 0.12834954770438642, "grad_norm": 0.16509035229682922, "learning_rate": 3.9985769150091434e-05, "loss": 0.5952, "num_tokens": 74100260.0, "step": 94 }, { "epoch": 0.12971496842464583, "grad_norm": 0.17479553818702698, "learning_rate": 3.998469564417489e-05, "loss": 0.5408, "num_tokens": 74906036.0, "step": 95 }, { "epoch": 0.13108038914490527, "grad_norm": 0.14281851053237915, "learning_rate": 3.998358312420723e-05, "loss": 0.5245, "num_tokens": 75710176.0, "step": 96 }, { "epoch": 0.13244580986516472, "grad_norm": 0.17302556335926056, "learning_rate": 3.998243159260184e-05, "loss": 0.5717, "num_tokens": 76589725.0, "step": 97 }, { "epoch": 0.13381123058542413, "grad_norm": 0.14468969404697418, "learning_rate": 3.9981241051856734e-05, "loss": 0.5334, "num_tokens": 77363332.0, "step": 98 }, { "epoch": 0.13517665130568357, "grad_norm": 0.1566748321056366, "learning_rate": 3.998001150455452e-05, "loss": 0.5626, "num_tokens": 78157499.0, "step": 99 }, { "epoch": 0.13654207202594298, "grad_norm": 0.1352996677160263, "learning_rate": 3.997874295336246e-05, "loss": 0.5613, "num_tokens": 78950774.0, "step": 100 }, { "epoch": 0.13790749274620243, "grad_norm": 0.15847869217395782, "learning_rate": 3.9977435401032415e-05, "loss": 0.56, "num_tokens": 79771176.0, "step": 101 }, { "epoch": 0.13927291346646187, "grad_norm": 0.14283543825149536, "learning_rate": 3.997608885040083e-05, "loss": 0.5636, "num_tokens": 80574592.0, "step": 102 }, { "epoch": 0.14063833418672128, "grad_norm": 0.1420551836490631, "learning_rate": 3.997470330438879e-05, "loss": 0.5651, "num_tokens": 81458112.0, "step": 103 }, { "epoch": 0.14200375490698072, "grad_norm": 0.14828644692897797, "learning_rate": 3.997327876600194e-05, "loss": 0.5618, "num_tokens": 82283035.0, "step": 104 }, { "epoch": 0.14336917562724014, "grad_norm": 0.14222629368305206, "learning_rate": 3.997181523833051e-05, "loss": 0.5612, "num_tokens": 83107054.0, "step": 105 }, { "epoch": 0.14473459634749958, "grad_norm": 0.170798659324646, "learning_rate": 3.9970312724549325e-05, "loss": 0.5348, "num_tokens": 83870064.0, "step": 106 }, { "epoch": 0.146100017067759, "grad_norm": 0.13094858825206757, "learning_rate": 3.996877122791778e-05, "loss": 0.5552, "num_tokens": 84629069.0, "step": 107 }, { "epoch": 0.14746543778801843, "grad_norm": 0.16360780596733093, "learning_rate": 3.9967190751779815e-05, "loss": 0.5652, "num_tokens": 85399380.0, "step": 108 }, { "epoch": 0.14883085850827787, "grad_norm": 0.14088203012943268, "learning_rate": 3.9965571299563966e-05, "loss": 0.5583, "num_tokens": 86218939.0, "step": 109 }, { "epoch": 0.1501962792285373, "grad_norm": 0.14044348895549774, "learning_rate": 3.9963912874783276e-05, "loss": 0.5373, "num_tokens": 86997053.0, "step": 110 }, { "epoch": 0.15156169994879673, "grad_norm": 0.150618776679039, "learning_rate": 3.9962215481035374e-05, "loss": 0.5221, "num_tokens": 87689120.0, "step": 111 }, { "epoch": 0.15292712066905614, "grad_norm": 0.1493023931980133, "learning_rate": 3.9960479122002384e-05, "loss": 0.5253, "num_tokens": 88488438.0, "step": 112 }, { "epoch": 0.15429254138931559, "grad_norm": 0.13517378270626068, "learning_rate": 3.995870380145098e-05, "loss": 0.5483, "num_tokens": 89273873.0, "step": 113 }, { "epoch": 0.15565796210957503, "grad_norm": 0.18321827054023743, "learning_rate": 3.995688952323236e-05, "loss": 0.5937, "num_tokens": 90108832.0, "step": 114 }, { "epoch": 0.15702338282983444, "grad_norm": 0.1421629786491394, "learning_rate": 3.99550362912822e-05, "loss": 0.5582, "num_tokens": 90928601.0, "step": 115 }, { "epoch": 0.15838880355009388, "grad_norm": 0.16815102100372314, "learning_rate": 3.995314410962073e-05, "loss": 0.5288, "num_tokens": 91689223.0, "step": 116 }, { "epoch": 0.1597542242703533, "grad_norm": 0.14444385468959808, "learning_rate": 3.995121298235262e-05, "loss": 0.5236, "num_tokens": 92516553.0, "step": 117 }, { "epoch": 0.16111964499061274, "grad_norm": 0.15844464302062988, "learning_rate": 3.9949242913667066e-05, "loss": 0.5522, "num_tokens": 93292360.0, "step": 118 }, { "epoch": 0.16248506571087215, "grad_norm": 0.16474685072898865, "learning_rate": 3.994723390783771e-05, "loss": 0.5159, "num_tokens": 94084233.0, "step": 119 }, { "epoch": 0.1638504864311316, "grad_norm": 0.1486508548259735, "learning_rate": 3.994518596922267e-05, "loss": 0.5455, "num_tokens": 94912958.0, "step": 120 }, { "epoch": 0.16521590715139103, "grad_norm": 0.1605079025030136, "learning_rate": 3.994309910226453e-05, "loss": 0.5314, "num_tokens": 95684230.0, "step": 121 }, { "epoch": 0.16658132787165045, "grad_norm": 0.16724231839179993, "learning_rate": 3.994097331149031e-05, "loss": 0.5597, "num_tokens": 96567534.0, "step": 122 }, { "epoch": 0.1679467485919099, "grad_norm": 0.1558239758014679, "learning_rate": 3.9938808601511475e-05, "loss": 0.5183, "num_tokens": 97354656.0, "step": 123 }, { "epoch": 0.1693121693121693, "grad_norm": 0.15532882511615753, "learning_rate": 3.99366049770239e-05, "loss": 0.5712, "num_tokens": 98139538.0, "step": 124 }, { "epoch": 0.17067759003242874, "grad_norm": 0.1702343225479126, "learning_rate": 3.9934362442807904e-05, "loss": 0.5521, "num_tokens": 98963905.0, "step": 125 }, { "epoch": 0.17204301075268819, "grad_norm": 0.1348073035478592, "learning_rate": 3.993208100372818e-05, "loss": 0.516, "num_tokens": 99674602.0, "step": 126 }, { "epoch": 0.1734084314729476, "grad_norm": 0.1658230423927307, "learning_rate": 3.992976066473386e-05, "loss": 0.5651, "num_tokens": 100467258.0, "step": 127 }, { "epoch": 0.17477385219320704, "grad_norm": 0.16453611850738525, "learning_rate": 3.992740143085842e-05, "loss": 0.5416, "num_tokens": 101369120.0, "step": 128 }, { "epoch": 0.17613927291346645, "grad_norm": 0.14305415749549866, "learning_rate": 3.992500330721973e-05, "loss": 0.5447, "num_tokens": 102174323.0, "step": 129 }, { "epoch": 0.1775046936337259, "grad_norm": 0.14612653851509094, "learning_rate": 3.9922566299020015e-05, "loss": 0.5266, "num_tokens": 103057513.0, "step": 130 }, { "epoch": 0.1788701143539853, "grad_norm": 0.14261004328727722, "learning_rate": 3.9920090411545855e-05, "loss": 0.5463, "num_tokens": 103874672.0, "step": 131 }, { "epoch": 0.18023553507424475, "grad_norm": 0.13625173270702362, "learning_rate": 3.991757565016818e-05, "loss": 0.5308, "num_tokens": 104702175.0, "step": 132 }, { "epoch": 0.1816009557945042, "grad_norm": 0.12169279903173447, "learning_rate": 3.991502202034225e-05, "loss": 0.5353, "num_tokens": 105508889.0, "step": 133 }, { "epoch": 0.1829663765147636, "grad_norm": 0.14052985608577728, "learning_rate": 3.991242952760761e-05, "loss": 0.5309, "num_tokens": 106286985.0, "step": 134 }, { "epoch": 0.18433179723502305, "grad_norm": 0.12482933700084686, "learning_rate": 3.990979817758815e-05, "loss": 0.5193, "num_tokens": 107034477.0, "step": 135 }, { "epoch": 0.18569721795528246, "grad_norm": 0.15964215993881226, "learning_rate": 3.990712797599202e-05, "loss": 0.5445, "num_tokens": 107815602.0, "step": 136 }, { "epoch": 0.1870626386755419, "grad_norm": 0.13896864652633667, "learning_rate": 3.9904418928611686e-05, "loss": 0.5436, "num_tokens": 108544217.0, "step": 137 }, { "epoch": 0.18842805939580134, "grad_norm": 0.13342343270778656, "learning_rate": 3.9901671041323856e-05, "loss": 0.5277, "num_tokens": 109292340.0, "step": 138 }, { "epoch": 0.18979348011606076, "grad_norm": 0.1488216519355774, "learning_rate": 3.989888432008948e-05, "loss": 0.5238, "num_tokens": 110016634.0, "step": 139 }, { "epoch": 0.1911589008363202, "grad_norm": 0.13699519634246826, "learning_rate": 3.98960587709538e-05, "loss": 0.5592, "num_tokens": 110851610.0, "step": 140 }, { "epoch": 0.19252432155657961, "grad_norm": 0.13484938442707062, "learning_rate": 3.989319440004623e-05, "loss": 0.4911, "num_tokens": 111596782.0, "step": 141 }, { "epoch": 0.19388974227683906, "grad_norm": 0.1357431858778, "learning_rate": 3.9890291213580444e-05, "loss": 0.5215, "num_tokens": 112403338.0, "step": 142 }, { "epoch": 0.19525516299709847, "grad_norm": 0.13509908318519592, "learning_rate": 3.988734921785429e-05, "loss": 0.53, "num_tokens": 113269004.0, "step": 143 }, { "epoch": 0.1966205837173579, "grad_norm": 0.12600664794445038, "learning_rate": 3.988436841924982e-05, "loss": 0.5185, "num_tokens": 114083746.0, "step": 144 }, { "epoch": 0.19798600443761735, "grad_norm": 0.13923096656799316, "learning_rate": 3.988134882423326e-05, "loss": 0.5361, "num_tokens": 114930696.0, "step": 145 }, { "epoch": 0.19935142515787677, "grad_norm": 0.15443995594978333, "learning_rate": 3.987829043935498e-05, "loss": 0.5223, "num_tokens": 115682051.0, "step": 146 }, { "epoch": 0.2007168458781362, "grad_norm": 0.14599281549453735, "learning_rate": 3.987519327124951e-05, "loss": 0.5273, "num_tokens": 116451759.0, "step": 147 }, { "epoch": 0.20208226659839562, "grad_norm": 0.16006502509117126, "learning_rate": 3.9872057326635525e-05, "loss": 0.5486, "num_tokens": 117254700.0, "step": 148 }, { "epoch": 0.20344768731865506, "grad_norm": 0.13126929104328156, "learning_rate": 3.986888261231579e-05, "loss": 0.5186, "num_tokens": 118030789.0, "step": 149 }, { "epoch": 0.2048131080389145, "grad_norm": 0.1480434387922287, "learning_rate": 3.98656691351772e-05, "loss": 0.5156, "num_tokens": 118743654.0, "step": 150 }, { "epoch": 0.20617852875917392, "grad_norm": 0.14261175692081451, "learning_rate": 3.9862416902190706e-05, "loss": 0.5292, "num_tokens": 119532223.0, "step": 151 }, { "epoch": 0.20754394947943336, "grad_norm": 0.127353698015213, "learning_rate": 3.985912592041136e-05, "loss": 0.5013, "num_tokens": 120322536.0, "step": 152 }, { "epoch": 0.20890937019969277, "grad_norm": 0.13723164796829224, "learning_rate": 3.9855796196978265e-05, "loss": 0.5098, "num_tokens": 121125785.0, "step": 153 }, { "epoch": 0.21027479091995221, "grad_norm": 0.15193486213684082, "learning_rate": 3.985242773911456e-05, "loss": 0.5196, "num_tokens": 121927858.0, "step": 154 }, { "epoch": 0.21164021164021163, "grad_norm": 0.13050471246242523, "learning_rate": 3.9849020554127415e-05, "loss": 0.5101, "num_tokens": 122808775.0, "step": 155 }, { "epoch": 0.21300563236047107, "grad_norm": 0.15082880854606628, "learning_rate": 3.9845574649407996e-05, "loss": 0.538, "num_tokens": 123584876.0, "step": 156 }, { "epoch": 0.2143710530807305, "grad_norm": 0.14649607241153717, "learning_rate": 3.984209003243149e-05, "loss": 0.4893, "num_tokens": 124317070.0, "step": 157 }, { "epoch": 0.21573647380098993, "grad_norm": 0.12898193299770355, "learning_rate": 3.9838566710757044e-05, "loss": 0.5563, "num_tokens": 125173487.0, "step": 158 }, { "epoch": 0.21710189452124937, "grad_norm": 0.13576172292232513, "learning_rate": 3.983500469202776e-05, "loss": 0.5322, "num_tokens": 126034658.0, "step": 159 }, { "epoch": 0.21846731524150878, "grad_norm": 0.13193383812904358, "learning_rate": 3.983140398397071e-05, "loss": 0.5036, "num_tokens": 126822301.0, "step": 160 }, { "epoch": 0.21983273596176822, "grad_norm": 0.11993744969367981, "learning_rate": 3.982776459439687e-05, "loss": 0.5144, "num_tokens": 127596381.0, "step": 161 }, { "epoch": 0.22119815668202766, "grad_norm": 0.12352003902196884, "learning_rate": 3.982408653120115e-05, "loss": 0.5188, "num_tokens": 128376625.0, "step": 162 }, { "epoch": 0.22256357740228708, "grad_norm": 0.14056414365768433, "learning_rate": 3.9820369802362325e-05, "loss": 0.5504, "num_tokens": 129165656.0, "step": 163 }, { "epoch": 0.22392899812254652, "grad_norm": 0.1410672962665558, "learning_rate": 3.981661441594308e-05, "loss": 0.527, "num_tokens": 129962103.0, "step": 164 }, { "epoch": 0.22529441884280593, "grad_norm": 0.13969191908836365, "learning_rate": 3.981282038008993e-05, "loss": 0.4969, "num_tokens": 130743246.0, "step": 165 }, { "epoch": 0.22665983956306537, "grad_norm": 0.1447666436433792, "learning_rate": 3.9808987703033256e-05, "loss": 0.5001, "num_tokens": 131566020.0, "step": 166 }, { "epoch": 0.2280252602833248, "grad_norm": 0.13351190090179443, "learning_rate": 3.980511639308724e-05, "loss": 0.5127, "num_tokens": 132411125.0, "step": 167 }, { "epoch": 0.22939068100358423, "grad_norm": 0.13083189725875854, "learning_rate": 3.980120645864991e-05, "loss": 0.5249, "num_tokens": 133168196.0, "step": 168 }, { "epoch": 0.23075610172384367, "grad_norm": 0.13658730685710907, "learning_rate": 3.979725790820302e-05, "loss": 0.5223, "num_tokens": 134042762.0, "step": 169 }, { "epoch": 0.23212152244410308, "grad_norm": 0.14834824204444885, "learning_rate": 3.979327075031215e-05, "loss": 0.5383, "num_tokens": 134763575.0, "step": 170 }, { "epoch": 0.23348694316436253, "grad_norm": 0.14315855503082275, "learning_rate": 3.97892449936266e-05, "loss": 0.5141, "num_tokens": 135525003.0, "step": 171 }, { "epoch": 0.23485236388462194, "grad_norm": 0.13643085956573486, "learning_rate": 3.9785180646879406e-05, "loss": 0.522, "num_tokens": 136288630.0, "step": 172 }, { "epoch": 0.23621778460488138, "grad_norm": 0.13491426408290863, "learning_rate": 3.978107771888733e-05, "loss": 0.5616, "num_tokens": 137111606.0, "step": 173 }, { "epoch": 0.23758320532514082, "grad_norm": 0.12633995711803436, "learning_rate": 3.9776936218550825e-05, "loss": 0.5258, "num_tokens": 137958048.0, "step": 174 }, { "epoch": 0.23894862604540024, "grad_norm": 0.12897689640522003, "learning_rate": 3.977275615485401e-05, "loss": 0.5292, "num_tokens": 138707213.0, "step": 175 }, { "epoch": 0.24031404676565968, "grad_norm": 0.12580054998397827, "learning_rate": 3.9768537536864646e-05, "loss": 0.4924, "num_tokens": 139433684.0, "step": 176 }, { "epoch": 0.2416794674859191, "grad_norm": 0.13030987977981567, "learning_rate": 3.9764280373734165e-05, "loss": 0.506, "num_tokens": 140184059.0, "step": 177 }, { "epoch": 0.24304488820617853, "grad_norm": 0.1296316385269165, "learning_rate": 3.9759984674697584e-05, "loss": 0.496, "num_tokens": 140920103.0, "step": 178 }, { "epoch": 0.24441030892643795, "grad_norm": 0.12144947052001953, "learning_rate": 3.975565044907354e-05, "loss": 0.5024, "num_tokens": 141782005.0, "step": 179 }, { "epoch": 0.2457757296466974, "grad_norm": 0.13304942846298218, "learning_rate": 3.975127770626422e-05, "loss": 0.5452, "num_tokens": 142604020.0, "step": 180 }, { "epoch": 0.24714115036695683, "grad_norm": 0.12659525871276855, "learning_rate": 3.9746866455755385e-05, "loss": 0.5028, "num_tokens": 143376701.0, "step": 181 }, { "epoch": 0.24850657108721624, "grad_norm": 0.136659637093544, "learning_rate": 3.974241670711632e-05, "loss": 0.5022, "num_tokens": 144138031.0, "step": 182 }, { "epoch": 0.24987199180747569, "grad_norm": 0.13254864513874054, "learning_rate": 3.973792846999984e-05, "loss": 0.5098, "num_tokens": 144965455.0, "step": 183 }, { "epoch": 0.2512374125277351, "grad_norm": 0.13324053585529327, "learning_rate": 3.973340175414222e-05, "loss": 0.4829, "num_tokens": 145646325.0, "step": 184 }, { "epoch": 0.25260283324799454, "grad_norm": 0.13832193613052368, "learning_rate": 3.972883656936325e-05, "loss": 0.5123, "num_tokens": 146411168.0, "step": 185 }, { "epoch": 0.25396825396825395, "grad_norm": 0.12322260439395905, "learning_rate": 3.972423292556613e-05, "loss": 0.5037, "num_tokens": 147313146.0, "step": 186 }, { "epoch": 0.2553336746885134, "grad_norm": 0.15065120160579681, "learning_rate": 3.9719590832737516e-05, "loss": 0.5646, "num_tokens": 148150636.0, "step": 187 }, { "epoch": 0.25669909540877284, "grad_norm": 0.12956693768501282, "learning_rate": 3.9714910300947454e-05, "loss": 0.5264, "num_tokens": 148961511.0, "step": 188 }, { "epoch": 0.25806451612903225, "grad_norm": 0.13886499404907227, "learning_rate": 3.9710191340349385e-05, "loss": 0.4974, "num_tokens": 149754009.0, "step": 189 }, { "epoch": 0.25942993684929166, "grad_norm": 0.12982018291950226, "learning_rate": 3.9705433961180115e-05, "loss": 0.4961, "num_tokens": 150492877.0, "step": 190 }, { "epoch": 0.26079535756955113, "grad_norm": 0.12580277025699615, "learning_rate": 3.970063817375979e-05, "loss": 0.5026, "num_tokens": 151284637.0, "step": 191 }, { "epoch": 0.26216077828981055, "grad_norm": 0.1302463561296463, "learning_rate": 3.969580398849186e-05, "loss": 0.5049, "num_tokens": 152097333.0, "step": 192 }, { "epoch": 0.26352619901006996, "grad_norm": 0.13056300580501556, "learning_rate": 3.96909314158631e-05, "loss": 0.4933, "num_tokens": 152807242.0, "step": 193 }, { "epoch": 0.26489161973032943, "grad_norm": 0.12933456897735596, "learning_rate": 3.9686020466443514e-05, "loss": 0.4986, "num_tokens": 153601328.0, "step": 194 }, { "epoch": 0.26625704045058884, "grad_norm": 0.129929780960083, "learning_rate": 3.968107115088641e-05, "loss": 0.5245, "num_tokens": 154434231.0, "step": 195 }, { "epoch": 0.26762246117084826, "grad_norm": 0.1340140849351883, "learning_rate": 3.96760834799283e-05, "loss": 0.5263, "num_tokens": 155275964.0, "step": 196 }, { "epoch": 0.26898788189110767, "grad_norm": 0.12070321291685104, "learning_rate": 3.967105746438887e-05, "loss": 0.5131, "num_tokens": 156100553.0, "step": 197 }, { "epoch": 0.27035330261136714, "grad_norm": 0.13051171600818634, "learning_rate": 3.966599311517103e-05, "loss": 0.4992, "num_tokens": 156837817.0, "step": 198 }, { "epoch": 0.27171872333162655, "grad_norm": 0.1300807148218155, "learning_rate": 3.9660890443260836e-05, "loss": 0.5121, "num_tokens": 157651844.0, "step": 199 }, { "epoch": 0.27308414405188597, "grad_norm": 0.12717124819755554, "learning_rate": 3.9655749459727465e-05, "loss": 0.5091, "num_tokens": 158442369.0, "step": 200 }, { "epoch": 0.27444956477214544, "grad_norm": 0.12676075100898743, "learning_rate": 3.965057017572321e-05, "loss": 0.479, "num_tokens": 159225029.0, "step": 201 }, { "epoch": 0.27581498549240485, "grad_norm": 0.12440596520900726, "learning_rate": 3.9645352602483456e-05, "loss": 0.4962, "num_tokens": 159992340.0, "step": 202 }, { "epoch": 0.27718040621266427, "grad_norm": 0.13880044221878052, "learning_rate": 3.964009675132662e-05, "loss": 0.5329, "num_tokens": 160873144.0, "step": 203 }, { "epoch": 0.27854582693292373, "grad_norm": 0.12567774951457977, "learning_rate": 3.963480263365421e-05, "loss": 0.5019, "num_tokens": 161617185.0, "step": 204 }, { "epoch": 0.27991124765318315, "grad_norm": 0.12985511124134064, "learning_rate": 3.962947026095069e-05, "loss": 0.5218, "num_tokens": 162454763.0, "step": 205 }, { "epoch": 0.28127666837344256, "grad_norm": 0.1546543836593628, "learning_rate": 3.9624099644783526e-05, "loss": 0.5189, "num_tokens": 163196513.0, "step": 206 }, { "epoch": 0.282642089093702, "grad_norm": 0.1389068067073822, "learning_rate": 3.9618690796803155e-05, "loss": 0.4874, "num_tokens": 163950509.0, "step": 207 }, { "epoch": 0.28400750981396145, "grad_norm": 0.13041161000728607, "learning_rate": 3.961324372874296e-05, "loss": 0.4917, "num_tokens": 164755296.0, "step": 208 }, { "epoch": 0.28537293053422086, "grad_norm": 0.1695312112569809, "learning_rate": 3.960775845241921e-05, "loss": 0.5138, "num_tokens": 165537274.0, "step": 209 }, { "epoch": 0.2867383512544803, "grad_norm": 0.1249571293592453, "learning_rate": 3.960223497973108e-05, "loss": 0.5235, "num_tokens": 166343431.0, "step": 210 }, { "epoch": 0.28810377197473974, "grad_norm": 0.15330882370471954, "learning_rate": 3.959667332266057e-05, "loss": 0.5019, "num_tokens": 167109702.0, "step": 211 }, { "epoch": 0.28946919269499916, "grad_norm": 0.1437467336654663, "learning_rate": 3.959107349327256e-05, "loss": 0.5051, "num_tokens": 167939138.0, "step": 212 }, { "epoch": 0.29083461341525857, "grad_norm": 0.13683173060417175, "learning_rate": 3.95854355037147e-05, "loss": 0.5176, "num_tokens": 168719781.0, "step": 213 }, { "epoch": 0.292200034135518, "grad_norm": 0.16967898607254028, "learning_rate": 3.9579759366217444e-05, "loss": 0.5263, "num_tokens": 169538687.0, "step": 214 }, { "epoch": 0.29356545485577745, "grad_norm": 0.14437046647071838, "learning_rate": 3.9574045093093983e-05, "loss": 0.5058, "num_tokens": 170421781.0, "step": 215 }, { "epoch": 0.29493087557603687, "grad_norm": 0.1256384253501892, "learning_rate": 3.9568292696740236e-05, "loss": 0.4832, "num_tokens": 171173402.0, "step": 216 }, { "epoch": 0.2962962962962963, "grad_norm": 0.13323193788528442, "learning_rate": 3.9562502189634846e-05, "loss": 0.4882, "num_tokens": 171991522.0, "step": 217 }, { "epoch": 0.29766171701655575, "grad_norm": 0.12156341969966888, "learning_rate": 3.95566735843391e-05, "loss": 0.4909, "num_tokens": 172832867.0, "step": 218 }, { "epoch": 0.29902713773681516, "grad_norm": 0.1267954558134079, "learning_rate": 3.9550806893496944e-05, "loss": 0.4705, "num_tokens": 173628377.0, "step": 219 }, { "epoch": 0.3003925584570746, "grad_norm": 0.12566301226615906, "learning_rate": 3.9544902129834945e-05, "loss": 0.5132, "num_tokens": 174421681.0, "step": 220 }, { "epoch": 0.301757979177334, "grad_norm": 0.12414814531803131, "learning_rate": 3.953895930616226e-05, "loss": 0.5046, "num_tokens": 175253289.0, "step": 221 }, { "epoch": 0.30312339989759346, "grad_norm": 0.12783204019069672, "learning_rate": 3.9532978435370594e-05, "loss": 0.5322, "num_tokens": 176094038.0, "step": 222 }, { "epoch": 0.3044888206178529, "grad_norm": 0.12932175397872925, "learning_rate": 3.952695953043422e-05, "loss": 0.5115, "num_tokens": 176868819.0, "step": 223 }, { "epoch": 0.3058542413381123, "grad_norm": 0.12589162588119507, "learning_rate": 3.952090260440989e-05, "loss": 0.4879, "num_tokens": 177629175.0, "step": 224 }, { "epoch": 0.30721966205837176, "grad_norm": 0.1593281626701355, "learning_rate": 3.951480767043683e-05, "loss": 0.4897, "num_tokens": 178471729.0, "step": 225 }, { "epoch": 0.30858508277863117, "grad_norm": 0.12014173716306686, "learning_rate": 3.950867474173675e-05, "loss": 0.4985, "num_tokens": 179322474.0, "step": 226 }, { "epoch": 0.3099505034988906, "grad_norm": 0.13787603378295898, "learning_rate": 3.950250383161375e-05, "loss": 0.5154, "num_tokens": 180109609.0, "step": 227 }, { "epoch": 0.31131592421915005, "grad_norm": 0.11992870271205902, "learning_rate": 3.9496294953454336e-05, "loss": 0.5022, "num_tokens": 180923069.0, "step": 228 }, { "epoch": 0.31268134493940947, "grad_norm": 0.12946543097496033, "learning_rate": 3.949004812072738e-05, "loss": 0.511, "num_tokens": 181814846.0, "step": 229 }, { "epoch": 0.3140467656596689, "grad_norm": 0.12610578536987305, "learning_rate": 3.948376334698409e-05, "loss": 0.4813, "num_tokens": 182571366.0, "step": 230 }, { "epoch": 0.3154121863799283, "grad_norm": 0.13610830903053284, "learning_rate": 3.947744064585795e-05, "loss": 0.4969, "num_tokens": 183413056.0, "step": 231 }, { "epoch": 0.31677760710018776, "grad_norm": 0.12810571491718292, "learning_rate": 3.9471080031064765e-05, "loss": 0.4642, "num_tokens": 184197165.0, "step": 232 }, { "epoch": 0.3181430278204472, "grad_norm": 0.15171320736408234, "learning_rate": 3.946468151640255e-05, "loss": 0.4826, "num_tokens": 184963380.0, "step": 233 }, { "epoch": 0.3195084485407066, "grad_norm": 0.11839660257101059, "learning_rate": 3.945824511575157e-05, "loss": 0.4877, "num_tokens": 185712248.0, "step": 234 }, { "epoch": 0.32087386926096606, "grad_norm": 0.13386324048042297, "learning_rate": 3.945177084307423e-05, "loss": 0.507, "num_tokens": 186543537.0, "step": 235 }, { "epoch": 0.3222392899812255, "grad_norm": 0.13439667224884033, "learning_rate": 3.944525871241514e-05, "loss": 0.5153, "num_tokens": 187344004.0, "step": 236 }, { "epoch": 0.3236047107014849, "grad_norm": 0.14512810111045837, "learning_rate": 3.9438708737900986e-05, "loss": 0.5311, "num_tokens": 188181249.0, "step": 237 }, { "epoch": 0.3249701314217443, "grad_norm": 0.12478740513324738, "learning_rate": 3.94321209337406e-05, "loss": 0.4745, "num_tokens": 188952112.0, "step": 238 }, { "epoch": 0.32633555214200377, "grad_norm": 0.1324058175086975, "learning_rate": 3.942549531422483e-05, "loss": 0.5003, "num_tokens": 189709028.0, "step": 239 }, { "epoch": 0.3277009728622632, "grad_norm": 0.1250074803829193, "learning_rate": 3.941883189372659e-05, "loss": 0.4989, "num_tokens": 190513403.0, "step": 240 }, { "epoch": 0.3290663935825226, "grad_norm": 0.13516570627689362, "learning_rate": 3.941213068670078e-05, "loss": 0.491, "num_tokens": 191274413.0, "step": 241 }, { "epoch": 0.33043181430278207, "grad_norm": 0.12889507412910461, "learning_rate": 3.940539170768426e-05, "loss": 0.5155, "num_tokens": 192106428.0, "step": 242 }, { "epoch": 0.3317972350230415, "grad_norm": 0.14338460564613342, "learning_rate": 3.939861497129586e-05, "loss": 0.5329, "num_tokens": 192972804.0, "step": 243 }, { "epoch": 0.3331626557433009, "grad_norm": 0.1224081963300705, "learning_rate": 3.939180049223629e-05, "loss": 0.4964, "num_tokens": 193797629.0, "step": 244 }, { "epoch": 0.3345280764635603, "grad_norm": 0.15926790237426758, "learning_rate": 3.9384948285288145e-05, "loss": 0.5154, "num_tokens": 194603220.0, "step": 245 }, { "epoch": 0.3358934971838198, "grad_norm": 0.13762278854846954, "learning_rate": 3.937805836531585e-05, "loss": 0.4882, "num_tokens": 195362273.0, "step": 246 }, { "epoch": 0.3372589179040792, "grad_norm": 0.13821974396705627, "learning_rate": 3.937113074726567e-05, "loss": 0.4982, "num_tokens": 196180166.0, "step": 247 }, { "epoch": 0.3386243386243386, "grad_norm": 0.12870004773139954, "learning_rate": 3.9364165446165615e-05, "loss": 0.4687, "num_tokens": 196918032.0, "step": 248 }, { "epoch": 0.3399897593445981, "grad_norm": 0.1432657688856125, "learning_rate": 3.935716247712546e-05, "loss": 0.5127, "num_tokens": 197792328.0, "step": 249 }, { "epoch": 0.3413551800648575, "grad_norm": 0.1537325233221054, "learning_rate": 3.9350121855336694e-05, "loss": 0.4892, "num_tokens": 198560912.0, "step": 250 }, { "epoch": 0.3427206007851169, "grad_norm": 0.126512348651886, "learning_rate": 3.9343043596072464e-05, "loss": 0.4867, "num_tokens": 199383511.0, "step": 251 }, { "epoch": 0.34408602150537637, "grad_norm": 0.14470696449279785, "learning_rate": 3.9335927714687584e-05, "loss": 0.4925, "num_tokens": 200146973.0, "step": 252 }, { "epoch": 0.3454514422256358, "grad_norm": 0.1352502852678299, "learning_rate": 3.9328774226618485e-05, "loss": 0.4911, "num_tokens": 200943710.0, "step": 253 }, { "epoch": 0.3468168629458952, "grad_norm": 0.1337669938802719, "learning_rate": 3.932158314738317e-05, "loss": 0.4973, "num_tokens": 201723952.0, "step": 254 }, { "epoch": 0.3481822836661546, "grad_norm": 0.13206641376018524, "learning_rate": 3.931435449258118e-05, "loss": 0.4979, "num_tokens": 202554154.0, "step": 255 }, { "epoch": 0.3495477043864141, "grad_norm": 0.14025771617889404, "learning_rate": 3.930708827789357e-05, "loss": 0.4743, "num_tokens": 203349315.0, "step": 256 }, { "epoch": 0.3509131251066735, "grad_norm": 0.135168194770813, "learning_rate": 3.9299784519082894e-05, "loss": 0.4814, "num_tokens": 204137301.0, "step": 257 }, { "epoch": 0.3522785458269329, "grad_norm": 0.1312488466501236, "learning_rate": 3.929244323199313e-05, "loss": 0.503, "num_tokens": 204932127.0, "step": 258 }, { "epoch": 0.3536439665471924, "grad_norm": 0.14493899047374725, "learning_rate": 3.928506443254967e-05, "loss": 0.4854, "num_tokens": 205725045.0, "step": 259 }, { "epoch": 0.3550093872674518, "grad_norm": 0.11534607410430908, "learning_rate": 3.927764813675928e-05, "loss": 0.4641, "num_tokens": 206541541.0, "step": 260 }, { "epoch": 0.3563748079877112, "grad_norm": 0.13069704174995422, "learning_rate": 3.9270194360710064e-05, "loss": 0.485, "num_tokens": 207346244.0, "step": 261 }, { "epoch": 0.3577402287079706, "grad_norm": 0.13459208607673645, "learning_rate": 3.926270312057145e-05, "loss": 0.5353, "num_tokens": 208117201.0, "step": 262 }, { "epoch": 0.3591056494282301, "grad_norm": 0.12582483887672424, "learning_rate": 3.925517443259412e-05, "loss": 0.492, "num_tokens": 208848612.0, "step": 263 }, { "epoch": 0.3604710701484895, "grad_norm": 0.13332468271255493, "learning_rate": 3.924760831310999e-05, "loss": 0.5085, "num_tokens": 209629191.0, "step": 264 }, { "epoch": 0.3618364908687489, "grad_norm": 0.13498535752296448, "learning_rate": 3.924000477853218e-05, "loss": 0.4774, "num_tokens": 210368355.0, "step": 265 }, { "epoch": 0.3632019115890084, "grad_norm": 0.12421321123838425, "learning_rate": 3.9232363845354974e-05, "loss": 0.4769, "num_tokens": 211064793.0, "step": 266 }, { "epoch": 0.3645673323092678, "grad_norm": 0.1342562437057495, "learning_rate": 3.922468553015379e-05, "loss": 0.4883, "num_tokens": 211871464.0, "step": 267 }, { "epoch": 0.3659327530295272, "grad_norm": 0.1267426311969757, "learning_rate": 3.9216969849585135e-05, "loss": 0.4992, "num_tokens": 212717177.0, "step": 268 }, { "epoch": 0.3672981737497866, "grad_norm": 0.13620273768901825, "learning_rate": 3.920921682038656e-05, "loss": 0.4852, "num_tokens": 213529589.0, "step": 269 }, { "epoch": 0.3686635944700461, "grad_norm": 0.1233750581741333, "learning_rate": 3.9201426459376655e-05, "loss": 0.4936, "num_tokens": 214363269.0, "step": 270 }, { "epoch": 0.3700290151903055, "grad_norm": 0.12385567277669907, "learning_rate": 3.9193598783454984e-05, "loss": 0.487, "num_tokens": 215172877.0, "step": 271 }, { "epoch": 0.3713944359105649, "grad_norm": 0.1324305534362793, "learning_rate": 3.918573380960206e-05, "loss": 0.4904, "num_tokens": 216019108.0, "step": 272 }, { "epoch": 0.3727598566308244, "grad_norm": 0.12108734995126724, "learning_rate": 3.91778315548793e-05, "loss": 0.48, "num_tokens": 216834505.0, "step": 273 }, { "epoch": 0.3741252773510838, "grad_norm": 0.12073052674531937, "learning_rate": 3.9169892036429014e-05, "loss": 0.4905, "num_tokens": 217575081.0, "step": 274 }, { "epoch": 0.3754906980713432, "grad_norm": 0.1345428228378296, "learning_rate": 3.916191527147431e-05, "loss": 0.4721, "num_tokens": 218330904.0, "step": 275 }, { "epoch": 0.3768561187916027, "grad_norm": 0.1265539973974228, "learning_rate": 3.915390127731914e-05, "loss": 0.4855, "num_tokens": 219058547.0, "step": 276 }, { "epoch": 0.3782215395118621, "grad_norm": 0.12938545644283295, "learning_rate": 3.91458500713482e-05, "loss": 0.4952, "num_tokens": 219874010.0, "step": 277 }, { "epoch": 0.3795869602321215, "grad_norm": 0.13269777595996857, "learning_rate": 3.913776167102688e-05, "loss": 0.4888, "num_tokens": 220660471.0, "step": 278 }, { "epoch": 0.38095238095238093, "grad_norm": 0.12637197971343994, "learning_rate": 3.9129636093901284e-05, "loss": 0.4673, "num_tokens": 221372658.0, "step": 279 }, { "epoch": 0.3823178016726404, "grad_norm": 0.13545191287994385, "learning_rate": 3.9121473357598175e-05, "loss": 0.5057, "num_tokens": 222145732.0, "step": 280 }, { "epoch": 0.3836832223928998, "grad_norm": 0.1305961310863495, "learning_rate": 3.91132734798249e-05, "loss": 0.4864, "num_tokens": 222988223.0, "step": 281 }, { "epoch": 0.38504864311315923, "grad_norm": 0.1402752548456192, "learning_rate": 3.910503647836939e-05, "loss": 0.5246, "num_tokens": 223759249.0, "step": 282 }, { "epoch": 0.3864140638334187, "grad_norm": 0.11946433037519455, "learning_rate": 3.9096762371100094e-05, "loss": 0.4987, "num_tokens": 224582759.0, "step": 283 }, { "epoch": 0.3877794845536781, "grad_norm": 0.14142248034477234, "learning_rate": 3.908845117596597e-05, "loss": 0.493, "num_tokens": 225371765.0, "step": 284 }, { "epoch": 0.3891449052739375, "grad_norm": 0.1451454758644104, "learning_rate": 3.908010291099643e-05, "loss": 0.4931, "num_tokens": 226168287.0, "step": 285 }, { "epoch": 0.39051032599419694, "grad_norm": 0.13596823811531067, "learning_rate": 3.907171759430128e-05, "loss": 0.5002, "num_tokens": 226997883.0, "step": 286 }, { "epoch": 0.3918757467144564, "grad_norm": 0.1453455537557602, "learning_rate": 3.9063295244070743e-05, "loss": 0.4815, "num_tokens": 227839102.0, "step": 287 }, { "epoch": 0.3932411674347158, "grad_norm": 0.13035890460014343, "learning_rate": 3.9054835878575336e-05, "loss": 0.4754, "num_tokens": 228610228.0, "step": 288 }, { "epoch": 0.39460658815497524, "grad_norm": 0.12686631083488464, "learning_rate": 3.904633951616589e-05, "loss": 0.4956, "num_tokens": 229452496.0, "step": 289 }, { "epoch": 0.3959720088752347, "grad_norm": 0.13363999128341675, "learning_rate": 3.90378061752735e-05, "loss": 0.5059, "num_tokens": 230218550.0, "step": 290 }, { "epoch": 0.3973374295954941, "grad_norm": 0.14305762946605682, "learning_rate": 3.902923587440947e-05, "loss": 0.5011, "num_tokens": 231028034.0, "step": 291 }, { "epoch": 0.39870285031575353, "grad_norm": 0.13325920701026917, "learning_rate": 3.902062863216528e-05, "loss": 0.4768, "num_tokens": 231746624.0, "step": 292 }, { "epoch": 0.40006827103601295, "grad_norm": 0.152951180934906, "learning_rate": 3.901198446721256e-05, "loss": 0.488, "num_tokens": 232528366.0, "step": 293 }, { "epoch": 0.4014336917562724, "grad_norm": 0.12882177531719208, "learning_rate": 3.900330339830301e-05, "loss": 0.4735, "num_tokens": 233298687.0, "step": 294 }, { "epoch": 0.40279911247653183, "grad_norm": 0.13907700777053833, "learning_rate": 3.899458544426841e-05, "loss": 0.4849, "num_tokens": 234130385.0, "step": 295 }, { "epoch": 0.40416453319679124, "grad_norm": 0.1272072046995163, "learning_rate": 3.8985830624020533e-05, "loss": 0.4549, "num_tokens": 234846290.0, "step": 296 }, { "epoch": 0.4055299539170507, "grad_norm": 0.15328718721866608, "learning_rate": 3.897703895655115e-05, "loss": 0.5099, "num_tokens": 235669164.0, "step": 297 }, { "epoch": 0.4068953746373101, "grad_norm": 0.13334152102470398, "learning_rate": 3.8968210460931957e-05, "loss": 0.4728, "num_tokens": 236474167.0, "step": 298 }, { "epoch": 0.40826079535756954, "grad_norm": 0.13858328759670258, "learning_rate": 3.8959345156314524e-05, "loss": 0.4943, "num_tokens": 237347244.0, "step": 299 }, { "epoch": 0.409626216077829, "grad_norm": 0.11589275300502777, "learning_rate": 3.8950443061930295e-05, "loss": 0.4815, "num_tokens": 238150290.0, "step": 300 }, { "epoch": 0.4109916367980884, "grad_norm": 0.14060929417610168, "learning_rate": 3.89415041970905e-05, "loss": 0.4725, "num_tokens": 238979776.0, "step": 301 }, { "epoch": 0.41235705751834784, "grad_norm": 0.12716488540172577, "learning_rate": 3.8932528581186145e-05, "loss": 0.5006, "num_tokens": 239735656.0, "step": 302 }, { "epoch": 0.41372247823860725, "grad_norm": 0.15183211863040924, "learning_rate": 3.892351623368796e-05, "loss": 0.4968, "num_tokens": 240549603.0, "step": 303 }, { "epoch": 0.4150878989588667, "grad_norm": 0.1288321167230606, "learning_rate": 3.891446717414635e-05, "loss": 0.5057, "num_tokens": 241375585.0, "step": 304 }, { "epoch": 0.41645331967912613, "grad_norm": 0.13644681870937347, "learning_rate": 3.8905381422191385e-05, "loss": 0.5164, "num_tokens": 242195243.0, "step": 305 }, { "epoch": 0.41781874039938555, "grad_norm": 0.12351369857788086, "learning_rate": 3.8896258997532685e-05, "loss": 0.4753, "num_tokens": 242980105.0, "step": 306 }, { "epoch": 0.419184161119645, "grad_norm": 0.13307064771652222, "learning_rate": 3.8887099919959475e-05, "loss": 0.5136, "num_tokens": 243750825.0, "step": 307 }, { "epoch": 0.42054958183990443, "grad_norm": 0.1285872757434845, "learning_rate": 3.887790420934045e-05, "loss": 0.4733, "num_tokens": 244494702.0, "step": 308 }, { "epoch": 0.42191500256016384, "grad_norm": 0.11997794359922409, "learning_rate": 3.886867188562379e-05, "loss": 0.5002, "num_tokens": 245349162.0, "step": 309 }, { "epoch": 0.42328042328042326, "grad_norm": 0.13487938046455383, "learning_rate": 3.8859402968837106e-05, "loss": 0.5057, "num_tokens": 246109777.0, "step": 310 }, { "epoch": 0.4246458440006827, "grad_norm": 0.12995147705078125, "learning_rate": 3.885009747908737e-05, "loss": 0.4854, "num_tokens": 246917096.0, "step": 311 }, { "epoch": 0.42601126472094214, "grad_norm": 0.1249474361538887, "learning_rate": 3.884075543656092e-05, "loss": 0.4983, "num_tokens": 247691991.0, "step": 312 }, { "epoch": 0.42737668544120155, "grad_norm": 0.12506252527236938, "learning_rate": 3.883137686152336e-05, "loss": 0.4818, "num_tokens": 248462046.0, "step": 313 }, { "epoch": 0.428742106161461, "grad_norm": 0.12926730513572693, "learning_rate": 3.8821961774319555e-05, "loss": 0.4872, "num_tokens": 249240298.0, "step": 314 }, { "epoch": 0.43010752688172044, "grad_norm": 0.11830523610115051, "learning_rate": 3.881251019537358e-05, "loss": 0.49, "num_tokens": 249970068.0, "step": 315 }, { "epoch": 0.43147294760197985, "grad_norm": 0.12135085463523865, "learning_rate": 3.880302214518866e-05, "loss": 0.5108, "num_tokens": 250746191.0, "step": 316 }, { "epoch": 0.43283836832223926, "grad_norm": 0.13148951530456543, "learning_rate": 3.879349764434716e-05, "loss": 0.5099, "num_tokens": 251526613.0, "step": 317 }, { "epoch": 0.43420378904249873, "grad_norm": 0.12720270454883575, "learning_rate": 3.878393671351048e-05, "loss": 0.4943, "num_tokens": 252361320.0, "step": 318 }, { "epoch": 0.43556920976275815, "grad_norm": 0.11525409668684006, "learning_rate": 3.877433937341908e-05, "loss": 0.4826, "num_tokens": 253162257.0, "step": 319 }, { "epoch": 0.43693463048301756, "grad_norm": 0.12461768090724945, "learning_rate": 3.876470564489239e-05, "loss": 0.4548, "num_tokens": 253965539.0, "step": 320 }, { "epoch": 0.43830005120327703, "grad_norm": 0.14165204763412476, "learning_rate": 3.875503554882878e-05, "loss": 0.5104, "num_tokens": 254851788.0, "step": 321 }, { "epoch": 0.43966547192353644, "grad_norm": 0.16590973734855652, "learning_rate": 3.8745329106205486e-05, "loss": 0.5313, "num_tokens": 255674573.0, "step": 322 }, { "epoch": 0.44103089264379586, "grad_norm": 0.13366802036762238, "learning_rate": 3.873558633807865e-05, "loss": 0.4869, "num_tokens": 256448118.0, "step": 323 }, { "epoch": 0.4423963133640553, "grad_norm": 0.13845571875572205, "learning_rate": 3.872580726558315e-05, "loss": 0.4576, "num_tokens": 257145651.0, "step": 324 }, { "epoch": 0.44376173408431474, "grad_norm": 0.12902379035949707, "learning_rate": 3.871599190993266e-05, "loss": 0.4843, "num_tokens": 257887720.0, "step": 325 }, { "epoch": 0.44512715480457415, "grad_norm": 0.13291458785533905, "learning_rate": 3.8706140292419546e-05, "loss": 0.489, "num_tokens": 258702814.0, "step": 326 }, { "epoch": 0.44649257552483357, "grad_norm": 0.1350133717060089, "learning_rate": 3.869625243441484e-05, "loss": 0.4904, "num_tokens": 259480497.0, "step": 327 }, { "epoch": 0.44785799624509304, "grad_norm": 0.12515218555927277, "learning_rate": 3.868632835736819e-05, "loss": 0.4759, "num_tokens": 260243718.0, "step": 328 }, { "epoch": 0.44922341696535245, "grad_norm": 0.1322977989912033, "learning_rate": 3.867636808280781e-05, "loss": 0.5025, "num_tokens": 261058418.0, "step": 329 }, { "epoch": 0.45058883768561186, "grad_norm": 0.1619795858860016, "learning_rate": 3.866637163234045e-05, "loss": 0.4616, "num_tokens": 261882902.0, "step": 330 }, { "epoch": 0.45195425840587133, "grad_norm": 0.12863211333751678, "learning_rate": 3.865633902765133e-05, "loss": 0.5065, "num_tokens": 262715524.0, "step": 331 }, { "epoch": 0.45331967912613075, "grad_norm": 0.13143686950206757, "learning_rate": 3.8646270290504085e-05, "loss": 0.4685, "num_tokens": 263420416.0, "step": 332 }, { "epoch": 0.45468509984639016, "grad_norm": 0.1165337860584259, "learning_rate": 3.8636165442740746e-05, "loss": 0.4824, "num_tokens": 264172768.0, "step": 333 }, { "epoch": 0.4560505205666496, "grad_norm": 0.14321830868721008, "learning_rate": 3.862602450628167e-05, "loss": 0.4844, "num_tokens": 265016520.0, "step": 334 }, { "epoch": 0.45741594128690904, "grad_norm": 0.13957498967647552, "learning_rate": 3.861584750312552e-05, "loss": 0.4994, "num_tokens": 265835105.0, "step": 335 }, { "epoch": 0.45878136200716846, "grad_norm": 0.14110423624515533, "learning_rate": 3.860563445534918e-05, "loss": 0.4638, "num_tokens": 266617436.0, "step": 336 }, { "epoch": 0.46014678272742787, "grad_norm": 0.129932701587677, "learning_rate": 3.859538538510772e-05, "loss": 0.4904, "num_tokens": 267457455.0, "step": 337 }, { "epoch": 0.46151220344768734, "grad_norm": 0.11453530937433243, "learning_rate": 3.858510031463438e-05, "loss": 0.4664, "num_tokens": 268242258.0, "step": 338 }, { "epoch": 0.46287762416794676, "grad_norm": 0.14709436893463135, "learning_rate": 3.857477926624046e-05, "loss": 0.4944, "num_tokens": 269131393.0, "step": 339 }, { "epoch": 0.46424304488820617, "grad_norm": 0.12136434018611908, "learning_rate": 3.8564422262315336e-05, "loss": 0.5085, "num_tokens": 270011275.0, "step": 340 }, { "epoch": 0.4656084656084656, "grad_norm": 0.12732860445976257, "learning_rate": 3.855402932532637e-05, "loss": 0.4681, "num_tokens": 270767380.0, "step": 341 }, { "epoch": 0.46697388632872505, "grad_norm": 0.13329514861106873, "learning_rate": 3.854360047781886e-05, "loss": 0.4643, "num_tokens": 271529966.0, "step": 342 }, { "epoch": 0.46833930704898447, "grad_norm": 0.12536756694316864, "learning_rate": 3.8533135742416045e-05, "loss": 0.4679, "num_tokens": 272309263.0, "step": 343 }, { "epoch": 0.4697047277692439, "grad_norm": 0.1255759447813034, "learning_rate": 3.852263514181897e-05, "loss": 0.4737, "num_tokens": 273129564.0, "step": 344 }, { "epoch": 0.47107014848950335, "grad_norm": 0.13341021537780762, "learning_rate": 3.8512098698806506e-05, "loss": 0.5111, "num_tokens": 273989173.0, "step": 345 }, { "epoch": 0.47243556920976276, "grad_norm": 0.11491604149341583, "learning_rate": 3.850152643623527e-05, "loss": 0.4776, "num_tokens": 274740053.0, "step": 346 }, { "epoch": 0.4738009899300222, "grad_norm": 0.13949774205684662, "learning_rate": 3.8490918377039576e-05, "loss": 0.4936, "num_tokens": 275525377.0, "step": 347 }, { "epoch": 0.47516641065028165, "grad_norm": 0.12117448449134827, "learning_rate": 3.8480274544231415e-05, "loss": 0.5006, "num_tokens": 276343288.0, "step": 348 }, { "epoch": 0.47653183137054106, "grad_norm": 0.1340339034795761, "learning_rate": 3.846959496090036e-05, "loss": 0.469, "num_tokens": 277146721.0, "step": 349 }, { "epoch": 0.4778972520908005, "grad_norm": 0.11864349991083145, "learning_rate": 3.8458879650213534e-05, "loss": 0.4862, "num_tokens": 277938948.0, "step": 350 }, { "epoch": 0.4792626728110599, "grad_norm": 0.14227856695652008, "learning_rate": 3.844812863541557e-05, "loss": 0.4969, "num_tokens": 278729851.0, "step": 351 }, { "epoch": 0.48062809353131936, "grad_norm": 0.13014350831508636, "learning_rate": 3.843734193982857e-05, "loss": 0.4967, "num_tokens": 279556826.0, "step": 352 }, { "epoch": 0.48199351425157877, "grad_norm": 0.13454239070415497, "learning_rate": 3.842651958685202e-05, "loss": 0.4837, "num_tokens": 280362910.0, "step": 353 }, { "epoch": 0.4833589349718382, "grad_norm": 0.12066609412431717, "learning_rate": 3.8415661599962746e-05, "loss": 0.4993, "num_tokens": 281155713.0, "step": 354 }, { "epoch": 0.48472435569209765, "grad_norm": 0.13374511897563934, "learning_rate": 3.8404768002714904e-05, "loss": 0.4917, "num_tokens": 281956614.0, "step": 355 }, { "epoch": 0.48608977641235707, "grad_norm": 0.12322619557380676, "learning_rate": 3.8393838818739874e-05, "loss": 0.5049, "num_tokens": 282810430.0, "step": 356 }, { "epoch": 0.4874551971326165, "grad_norm": 0.12862402200698853, "learning_rate": 3.838287407174626e-05, "loss": 0.4977, "num_tokens": 283636081.0, "step": 357 }, { "epoch": 0.4888206178528759, "grad_norm": 0.12247277796268463, "learning_rate": 3.8371873785519785e-05, "loss": 0.4934, "num_tokens": 284474091.0, "step": 358 }, { "epoch": 0.49018603857313536, "grad_norm": 0.1291111558675766, "learning_rate": 3.836083798392327e-05, "loss": 0.4796, "num_tokens": 285219505.0, "step": 359 }, { "epoch": 0.4915514592933948, "grad_norm": 0.13791251182556152, "learning_rate": 3.83497666908966e-05, "loss": 0.5003, "num_tokens": 285977446.0, "step": 360 }, { "epoch": 0.4929168800136542, "grad_norm": 0.12347041070461273, "learning_rate": 3.833865993045665e-05, "loss": 0.4596, "num_tokens": 286710703.0, "step": 361 }, { "epoch": 0.49428230073391366, "grad_norm": 0.1313571184873581, "learning_rate": 3.8327517726697194e-05, "loss": 0.4545, "num_tokens": 287451186.0, "step": 362 }, { "epoch": 0.4956477214541731, "grad_norm": 0.1284482479095459, "learning_rate": 3.8316340103788947e-05, "loss": 0.4848, "num_tokens": 288254717.0, "step": 363 }, { "epoch": 0.4970131421744325, "grad_norm": 0.13200753927230835, "learning_rate": 3.830512708597942e-05, "loss": 0.4717, "num_tokens": 289009025.0, "step": 364 }, { "epoch": 0.4983785628946919, "grad_norm": 0.11362718790769577, "learning_rate": 3.829387869759293e-05, "loss": 0.4867, "num_tokens": 289767068.0, "step": 365 }, { "epoch": 0.49974398361495137, "grad_norm": 0.1497984528541565, "learning_rate": 3.82825949630305e-05, "loss": 0.4792, "num_tokens": 290633366.0, "step": 366 }, { "epoch": 0.5011094043352108, "grad_norm": 0.1118970662355423, "learning_rate": 3.827127590676986e-05, "loss": 0.4619, "num_tokens": 291387788.0, "step": 367 }, { "epoch": 0.5024748250554703, "grad_norm": 0.1377825289964676, "learning_rate": 3.8259921553365324e-05, "loss": 0.478, "num_tokens": 292209554.0, "step": 368 }, { "epoch": 0.5038402457757296, "grad_norm": 0.13069762289524078, "learning_rate": 3.8248531927447805e-05, "loss": 0.4964, "num_tokens": 293091921.0, "step": 369 }, { "epoch": 0.5052056664959891, "grad_norm": 0.1204625815153122, "learning_rate": 3.823710705372473e-05, "loss": 0.4647, "num_tokens": 293866345.0, "step": 370 }, { "epoch": 0.5065710872162486, "grad_norm": 0.1387002319097519, "learning_rate": 3.822564695697998e-05, "loss": 0.4495, "num_tokens": 294644548.0, "step": 371 }, { "epoch": 0.5079365079365079, "grad_norm": 0.1187334656715393, "learning_rate": 3.8214151662073854e-05, "loss": 0.4902, "num_tokens": 295474477.0, "step": 372 }, { "epoch": 0.5093019286567674, "grad_norm": 0.14254598319530487, "learning_rate": 3.8202621193942976e-05, "loss": 0.5061, "num_tokens": 296307758.0, "step": 373 }, { "epoch": 0.5106673493770268, "grad_norm": 0.12825170159339905, "learning_rate": 3.819105557760032e-05, "loss": 0.4681, "num_tokens": 297089759.0, "step": 374 }, { "epoch": 0.5120327700972862, "grad_norm": 0.1315608024597168, "learning_rate": 3.8179454838135066e-05, "loss": 0.5087, "num_tokens": 297915462.0, "step": 375 }, { "epoch": 0.5133981908175457, "grad_norm": 0.11585547029972076, "learning_rate": 3.8167819000712615e-05, "loss": 0.4531, "num_tokens": 298677243.0, "step": 376 }, { "epoch": 0.514763611537805, "grad_norm": 0.12849685549736023, "learning_rate": 3.815614809057448e-05, "loss": 0.455, "num_tokens": 299450788.0, "step": 377 }, { "epoch": 0.5161290322580645, "grad_norm": 0.11222395300865173, "learning_rate": 3.814444213303828e-05, "loss": 0.4887, "num_tokens": 300260294.0, "step": 378 }, { "epoch": 0.517494452978324, "grad_norm": 0.11868235468864441, "learning_rate": 3.8132701153497655e-05, "loss": 0.47, "num_tokens": 301013018.0, "step": 379 }, { "epoch": 0.5188598736985833, "grad_norm": 0.13232892751693726, "learning_rate": 3.8120925177422196e-05, "loss": 0.4842, "num_tokens": 301735232.0, "step": 380 }, { "epoch": 0.5202252944188428, "grad_norm": 0.12665817141532898, "learning_rate": 3.810911423035745e-05, "loss": 0.5095, "num_tokens": 302537556.0, "step": 381 }, { "epoch": 0.5215907151391023, "grad_norm": 0.12276158481836319, "learning_rate": 3.809726833792479e-05, "loss": 0.4649, "num_tokens": 303362755.0, "step": 382 }, { "epoch": 0.5229561358593616, "grad_norm": 0.12200341373682022, "learning_rate": 3.808538752582142e-05, "loss": 0.4874, "num_tokens": 304173699.0, "step": 383 }, { "epoch": 0.5243215565796211, "grad_norm": 0.11856434494256973, "learning_rate": 3.8073471819820294e-05, "loss": 0.4735, "num_tokens": 304909615.0, "step": 384 }, { "epoch": 0.5256869772998806, "grad_norm": 0.11634600162506104, "learning_rate": 3.8061521245770044e-05, "loss": 0.4646, "num_tokens": 305695795.0, "step": 385 }, { "epoch": 0.5270523980201399, "grad_norm": 0.13224948942661285, "learning_rate": 3.8049535829594956e-05, "loss": 0.4986, "num_tokens": 306501621.0, "step": 386 }, { "epoch": 0.5284178187403994, "grad_norm": 0.11825838685035706, "learning_rate": 3.803751559729489e-05, "loss": 0.4608, "num_tokens": 307268646.0, "step": 387 }, { "epoch": 0.5297832394606589, "grad_norm": 0.13386474549770355, "learning_rate": 3.802546057494524e-05, "loss": 0.4993, "num_tokens": 308088492.0, "step": 388 }, { "epoch": 0.5311486601809182, "grad_norm": 0.12466966360807419, "learning_rate": 3.801337078869686e-05, "loss": 0.4814, "num_tokens": 308904659.0, "step": 389 }, { "epoch": 0.5325140809011777, "grad_norm": 0.11707313358783722, "learning_rate": 3.8001246264776025e-05, "loss": 0.4822, "num_tokens": 309630742.0, "step": 390 }, { "epoch": 0.5338795016214372, "grad_norm": 0.12110523134469986, "learning_rate": 3.798908702948437e-05, "loss": 0.4854, "num_tokens": 310445167.0, "step": 391 }, { "epoch": 0.5352449223416965, "grad_norm": 0.11914714425802231, "learning_rate": 3.797689310919883e-05, "loss": 0.4815, "num_tokens": 311256291.0, "step": 392 }, { "epoch": 0.536610343061956, "grad_norm": 0.11186397075653076, "learning_rate": 3.7964664530371565e-05, "loss": 0.4678, "num_tokens": 311991214.0, "step": 393 }, { "epoch": 0.5379757637822153, "grad_norm": 0.1148197203874588, "learning_rate": 3.795240131952992e-05, "loss": 0.4537, "num_tokens": 312728136.0, "step": 394 }, { "epoch": 0.5393411845024748, "grad_norm": 0.11663847416639328, "learning_rate": 3.7940103503276383e-05, "loss": 0.4486, "num_tokens": 313484090.0, "step": 395 }, { "epoch": 0.5407066052227343, "grad_norm": 1.368383765220642, "learning_rate": 3.792777110828852e-05, "loss": 0.4892, "num_tokens": 314299944.0, "step": 396 }, { "epoch": 0.5420720259429936, "grad_norm": 0.13734261691570282, "learning_rate": 3.791540416131887e-05, "loss": 0.5067, "num_tokens": 315074368.0, "step": 397 }, { "epoch": 0.5434374466632531, "grad_norm": 0.12051980942487717, "learning_rate": 3.7903002689194976e-05, "loss": 0.4745, "num_tokens": 315835589.0, "step": 398 }, { "epoch": 0.5448028673835126, "grad_norm": 0.1339758038520813, "learning_rate": 3.7890566718819216e-05, "loss": 0.489, "num_tokens": 316600285.0, "step": 399 }, { "epoch": 0.5461682881037719, "grad_norm": 0.12199810892343521, "learning_rate": 3.787809627716887e-05, "loss": 0.474, "num_tokens": 317360708.0, "step": 400 }, { "epoch": 0.5475337088240314, "grad_norm": 0.13229268789291382, "learning_rate": 3.786559139129594e-05, "loss": 0.4961, "num_tokens": 318183562.0, "step": 401 }, { "epoch": 0.5488991295442909, "grad_norm": 0.14335356652736664, "learning_rate": 3.785305208832717e-05, "loss": 0.4966, "num_tokens": 318924725.0, "step": 402 }, { "epoch": 0.5502645502645502, "grad_norm": 0.12446943670511246, "learning_rate": 3.784047839546399e-05, "loss": 0.4932, "num_tokens": 319715516.0, "step": 403 }, { "epoch": 0.5516299709848097, "grad_norm": 0.1267673820257187, "learning_rate": 3.7827870339982385e-05, "loss": 0.4706, "num_tokens": 320524087.0, "step": 404 }, { "epoch": 0.5529953917050692, "grad_norm": 0.11786322295665741, "learning_rate": 3.781522794923291e-05, "loss": 0.4672, "num_tokens": 321320871.0, "step": 405 }, { "epoch": 0.5543608124253285, "grad_norm": 0.13830965757369995, "learning_rate": 3.780255125064061e-05, "loss": 0.4529, "num_tokens": 322123574.0, "step": 406 }, { "epoch": 0.555726233145588, "grad_norm": 0.1261206418275833, "learning_rate": 3.7789840271704925e-05, "loss": 0.4946, "num_tokens": 322942375.0, "step": 407 }, { "epoch": 0.5570916538658475, "grad_norm": 0.1320163458585739, "learning_rate": 3.777709503999968e-05, "loss": 0.464, "num_tokens": 323711214.0, "step": 408 }, { "epoch": 0.5584570745861068, "grad_norm": 0.15263430774211884, "learning_rate": 3.776431558317302e-05, "loss": 0.4783, "num_tokens": 324410229.0, "step": 409 }, { "epoch": 0.5598224953063663, "grad_norm": 0.12495387345552444, "learning_rate": 3.775150192894729e-05, "loss": 0.4815, "num_tokens": 325227916.0, "step": 410 }, { "epoch": 0.5611879160266257, "grad_norm": 0.14345838129520416, "learning_rate": 3.773865410511906e-05, "loss": 0.4837, "num_tokens": 326060576.0, "step": 411 }, { "epoch": 0.5625533367468851, "grad_norm": 0.12346039712429047, "learning_rate": 3.7725772139559e-05, "loss": 0.4644, "num_tokens": 326847413.0, "step": 412 }, { "epoch": 0.5639187574671446, "grad_norm": 0.13675923645496368, "learning_rate": 3.771285606021184e-05, "loss": 0.4865, "num_tokens": 327647516.0, "step": 413 }, { "epoch": 0.565284178187404, "grad_norm": 0.1362939029932022, "learning_rate": 3.769990589509635e-05, "loss": 0.4845, "num_tokens": 328483286.0, "step": 414 }, { "epoch": 0.5666495989076634, "grad_norm": 0.12718060612678528, "learning_rate": 3.768692167230519e-05, "loss": 0.4587, "num_tokens": 329171505.0, "step": 415 }, { "epoch": 0.5680150196279229, "grad_norm": 0.12941154837608337, "learning_rate": 3.7673903420004946e-05, "loss": 0.4883, "num_tokens": 329980802.0, "step": 416 }, { "epoch": 0.5693804403481822, "grad_norm": 0.12722451984882355, "learning_rate": 3.7660851166436e-05, "loss": 0.4895, "num_tokens": 330854004.0, "step": 417 }, { "epoch": 0.5707458610684417, "grad_norm": 0.12728151679039001, "learning_rate": 3.764776493991249e-05, "loss": 0.4466, "num_tokens": 331543497.0, "step": 418 }, { "epoch": 0.5721112817887012, "grad_norm": 0.13428033888339996, "learning_rate": 3.763464476882226e-05, "loss": 0.5026, "num_tokens": 332409815.0, "step": 419 }, { "epoch": 0.5734767025089605, "grad_norm": 0.13157378137111664, "learning_rate": 3.76214906816268e-05, "loss": 0.4624, "num_tokens": 333197876.0, "step": 420 }, { "epoch": 0.57484212322922, "grad_norm": 0.12552690505981445, "learning_rate": 3.7608302706861154e-05, "loss": 0.452, "num_tokens": 333955160.0, "step": 421 }, { "epoch": 0.5762075439494795, "grad_norm": 0.13645872473716736, "learning_rate": 3.759508087313388e-05, "loss": 0.4804, "num_tokens": 334753387.0, "step": 422 }, { "epoch": 0.5775729646697388, "grad_norm": 0.12514165043830872, "learning_rate": 3.7581825209127e-05, "loss": 0.4704, "num_tokens": 335521237.0, "step": 423 }, { "epoch": 0.5789383853899983, "grad_norm": 0.13832393288612366, "learning_rate": 3.756853574359592e-05, "loss": 0.5018, "num_tokens": 336317735.0, "step": 424 }, { "epoch": 0.5803038061102577, "grad_norm": 0.12883363664150238, "learning_rate": 3.7555212505369355e-05, "loss": 0.4524, "num_tokens": 337027844.0, "step": 425 }, { "epoch": 0.5816692268305171, "grad_norm": 0.1348278522491455, "learning_rate": 3.7541855523349306e-05, "loss": 0.4413, "num_tokens": 337737812.0, "step": 426 }, { "epoch": 0.5830346475507766, "grad_norm": 0.13043542206287384, "learning_rate": 3.7528464826510965e-05, "loss": 0.4984, "num_tokens": 338501994.0, "step": 427 }, { "epoch": 0.584400068271036, "grad_norm": 0.1313769370317459, "learning_rate": 3.751504044390264e-05, "loss": 0.4882, "num_tokens": 339341284.0, "step": 428 }, { "epoch": 0.5857654889912954, "grad_norm": 0.13678865134716034, "learning_rate": 3.750158240464576e-05, "loss": 0.4935, "num_tokens": 340175933.0, "step": 429 }, { "epoch": 0.5871309097115549, "grad_norm": 0.14056245982646942, "learning_rate": 3.7488090737934714e-05, "loss": 0.4635, "num_tokens": 340897665.0, "step": 430 }, { "epoch": 0.5884963304318143, "grad_norm": 0.14164437353610992, "learning_rate": 3.7474565473036886e-05, "loss": 0.4965, "num_tokens": 341765677.0, "step": 431 }, { "epoch": 0.5898617511520737, "grad_norm": 0.12716221809387207, "learning_rate": 3.74610066392925e-05, "loss": 0.4587, "num_tokens": 342537361.0, "step": 432 }, { "epoch": 0.5912271718723332, "grad_norm": 0.13122911751270294, "learning_rate": 3.744741426611464e-05, "loss": 0.4715, "num_tokens": 343396334.0, "step": 433 }, { "epoch": 0.5925925925925926, "grad_norm": 0.11437292397022247, "learning_rate": 3.7433788382989106e-05, "loss": 0.4406, "num_tokens": 344178200.0, "step": 434 }, { "epoch": 0.593958013312852, "grad_norm": 0.15477678179740906, "learning_rate": 3.742012901947443e-05, "loss": 0.4868, "num_tokens": 344973474.0, "step": 435 }, { "epoch": 0.5953234340331115, "grad_norm": 0.12759697437286377, "learning_rate": 3.7406436205201755e-05, "loss": 0.4857, "num_tokens": 345799515.0, "step": 436 }, { "epoch": 0.5966888547533709, "grad_norm": 0.11280731111764908, "learning_rate": 3.7392709969874785e-05, "loss": 0.4719, "num_tokens": 346597671.0, "step": 437 }, { "epoch": 0.5980542754736303, "grad_norm": 0.12547516822814941, "learning_rate": 3.737895034326973e-05, "loss": 0.4714, "num_tokens": 347369229.0, "step": 438 }, { "epoch": 0.5994196961938898, "grad_norm": 0.12590670585632324, "learning_rate": 3.736515735523523e-05, "loss": 0.4597, "num_tokens": 348179454.0, "step": 439 }, { "epoch": 0.6007851169141492, "grad_norm": 0.12164514511823654, "learning_rate": 3.735133103569231e-05, "loss": 0.4364, "num_tokens": 348940907.0, "step": 440 }, { "epoch": 0.6021505376344086, "grad_norm": 0.1175864115357399, "learning_rate": 3.733747141463428e-05, "loss": 0.4433, "num_tokens": 349655955.0, "step": 441 }, { "epoch": 0.603515958354668, "grad_norm": 0.1369248479604721, "learning_rate": 3.7323578522126704e-05, "loss": 0.4882, "num_tokens": 350450446.0, "step": 442 }, { "epoch": 0.6048813790749274, "grad_norm": 0.13582779467105865, "learning_rate": 3.730965238830732e-05, "loss": 0.4923, "num_tokens": 351250119.0, "step": 443 }, { "epoch": 0.6062467997951869, "grad_norm": 0.15286728739738464, "learning_rate": 3.729569304338597e-05, "loss": 0.4761, "num_tokens": 352064552.0, "step": 444 }, { "epoch": 0.6076122205154463, "grad_norm": 0.139332577586174, "learning_rate": 3.7281700517644554e-05, "loss": 0.4755, "num_tokens": 352843500.0, "step": 445 }, { "epoch": 0.6089776412357057, "grad_norm": 0.14917761087417603, "learning_rate": 3.726767484143694e-05, "loss": 0.4901, "num_tokens": 353626154.0, "step": 446 }, { "epoch": 0.6103430619559652, "grad_norm": 0.1418592482805252, "learning_rate": 3.72536160451889e-05, "loss": 0.479, "num_tokens": 354390317.0, "step": 447 }, { "epoch": 0.6117084826762246, "grad_norm": 0.11798273772001266, "learning_rate": 3.7239524159398076e-05, "loss": 0.4545, "num_tokens": 355143363.0, "step": 448 }, { "epoch": 0.613073903396484, "grad_norm": 0.12445493042469025, "learning_rate": 3.722539921463387e-05, "loss": 0.4555, "num_tokens": 355990666.0, "step": 449 }, { "epoch": 0.6144393241167435, "grad_norm": 0.13971227407455444, "learning_rate": 3.7211241241537414e-05, "loss": 0.4963, "num_tokens": 356868742.0, "step": 450 }, { "epoch": 0.6158047448370029, "grad_norm": 0.12297403812408447, "learning_rate": 3.7197050270821485e-05, "loss": 0.4739, "num_tokens": 357657534.0, "step": 451 }, { "epoch": 0.6171701655572623, "grad_norm": 0.13458892703056335, "learning_rate": 3.7182826333270436e-05, "loss": 0.513, "num_tokens": 358519845.0, "step": 452 }, { "epoch": 0.6185355862775218, "grad_norm": 0.13525551557540894, "learning_rate": 3.7168569459740126e-05, "loss": 0.4828, "num_tokens": 359296634.0, "step": 453 }, { "epoch": 0.6199010069977812, "grad_norm": 0.11983805149793625, "learning_rate": 3.715427968115789e-05, "loss": 0.4573, "num_tokens": 360083788.0, "step": 454 }, { "epoch": 0.6212664277180406, "grad_norm": 0.1301286667585373, "learning_rate": 3.713995702852241e-05, "loss": 0.4777, "num_tokens": 360850110.0, "step": 455 }, { "epoch": 0.6226318484383001, "grad_norm": 0.12501905858516693, "learning_rate": 3.71256015329037e-05, "loss": 0.4945, "num_tokens": 361661721.0, "step": 456 }, { "epoch": 0.6239972691585595, "grad_norm": 0.12088735401630402, "learning_rate": 3.711121322544302e-05, "loss": 0.4569, "num_tokens": 362368444.0, "step": 457 }, { "epoch": 0.6253626898788189, "grad_norm": 0.11755267530679703, "learning_rate": 3.709679213735283e-05, "loss": 0.4621, "num_tokens": 363131376.0, "step": 458 }, { "epoch": 0.6267281105990783, "grad_norm": 0.11962898820638657, "learning_rate": 3.708233829991663e-05, "loss": 0.4895, "num_tokens": 363968705.0, "step": 459 }, { "epoch": 0.6280935313193378, "grad_norm": 0.11750291287899017, "learning_rate": 3.706785174448905e-05, "loss": 0.4423, "num_tokens": 364723139.0, "step": 460 }, { "epoch": 0.6294589520395972, "grad_norm": 0.11464496701955795, "learning_rate": 3.7053332502495636e-05, "loss": 0.467, "num_tokens": 365546943.0, "step": 461 }, { "epoch": 0.6308243727598566, "grad_norm": 0.12138552218675613, "learning_rate": 3.703878060543286e-05, "loss": 0.4662, "num_tokens": 366323327.0, "step": 462 }, { "epoch": 0.6321897934801161, "grad_norm": 0.11078216880559921, "learning_rate": 3.702419608486803e-05, "loss": 0.4701, "num_tokens": 367094074.0, "step": 463 }, { "epoch": 0.6335552142003755, "grad_norm": 0.1131124272942543, "learning_rate": 3.700957897243924e-05, "loss": 0.4512, "num_tokens": 367895419.0, "step": 464 }, { "epoch": 0.6349206349206349, "grad_norm": 0.11850471049547195, "learning_rate": 3.6994929299855254e-05, "loss": 0.4706, "num_tokens": 368683099.0, "step": 465 }, { "epoch": 0.6362860556408944, "grad_norm": 0.12135849893093109, "learning_rate": 3.6980247098895485e-05, "loss": 0.4627, "num_tokens": 369514780.0, "step": 466 }, { "epoch": 0.6376514763611538, "grad_norm": 0.11464785784482956, "learning_rate": 3.696553240140993e-05, "loss": 0.4778, "num_tokens": 370222242.0, "step": 467 }, { "epoch": 0.6390168970814132, "grad_norm": 0.12643593549728394, "learning_rate": 3.695078523931903e-05, "loss": 0.5156, "num_tokens": 371051620.0, "step": 468 }, { "epoch": 0.6403823178016727, "grad_norm": 0.12640857696533203, "learning_rate": 3.693600564461371e-05, "loss": 0.4889, "num_tokens": 371832582.0, "step": 469 }, { "epoch": 0.6417477385219321, "grad_norm": 0.11404629051685333, "learning_rate": 3.6921193649355206e-05, "loss": 0.4729, "num_tokens": 372625383.0, "step": 470 }, { "epoch": 0.6431131592421915, "grad_norm": 0.11478064954280853, "learning_rate": 3.6906349285675065e-05, "loss": 0.4395, "num_tokens": 373417309.0, "step": 471 }, { "epoch": 0.644478579962451, "grad_norm": 0.11802941560745239, "learning_rate": 3.6891472585775034e-05, "loss": 0.47, "num_tokens": 374217424.0, "step": 472 }, { "epoch": 0.6458440006827103, "grad_norm": 0.13621556758880615, "learning_rate": 3.687656358192702e-05, "loss": 0.4601, "num_tokens": 374934019.0, "step": 473 }, { "epoch": 0.6472094214029698, "grad_norm": 0.11538510769605637, "learning_rate": 3.6861622306473e-05, "loss": 0.477, "num_tokens": 375737112.0, "step": 474 }, { "epoch": 0.6485748421232292, "grad_norm": 0.1322227567434311, "learning_rate": 3.684664879182496e-05, "loss": 0.4655, "num_tokens": 376514011.0, "step": 475 }, { "epoch": 0.6499402628434886, "grad_norm": 0.12792377173900604, "learning_rate": 3.6831643070464806e-05, "loss": 0.4779, "num_tokens": 377275352.0, "step": 476 }, { "epoch": 0.6513056835637481, "grad_norm": 0.12464753538370132, "learning_rate": 3.681660517494436e-05, "loss": 0.4669, "num_tokens": 378083152.0, "step": 477 }, { "epoch": 0.6526711042840075, "grad_norm": 0.11663442105054855, "learning_rate": 3.680153513788517e-05, "loss": 0.4565, "num_tokens": 378819068.0, "step": 478 }, { "epoch": 0.6540365250042669, "grad_norm": 0.11345703154802322, "learning_rate": 3.678643299197855e-05, "loss": 0.4656, "num_tokens": 379545024.0, "step": 479 }, { "epoch": 0.6554019457245264, "grad_norm": 0.14854824542999268, "learning_rate": 3.677129876998548e-05, "loss": 0.4567, "num_tokens": 380337327.0, "step": 480 }, { "epoch": 0.6567673664447858, "grad_norm": 0.11623409390449524, "learning_rate": 3.675613250473648e-05, "loss": 0.4729, "num_tokens": 381179608.0, "step": 481 }, { "epoch": 0.6581327871650452, "grad_norm": 0.11213913559913635, "learning_rate": 3.6740934229131606e-05, "loss": 0.4828, "num_tokens": 381977225.0, "step": 482 }, { "epoch": 0.6594982078853047, "grad_norm": 0.11585624516010284, "learning_rate": 3.672570397614036e-05, "loss": 0.4704, "num_tokens": 382738745.0, "step": 483 }, { "epoch": 0.6608636286055641, "grad_norm": 0.12673477828502655, "learning_rate": 3.67104417788016e-05, "loss": 0.4819, "num_tokens": 383526241.0, "step": 484 }, { "epoch": 0.6622290493258235, "grad_norm": 0.1205187737941742, "learning_rate": 3.669514767022349e-05, "loss": 0.4712, "num_tokens": 384332792.0, "step": 485 }, { "epoch": 0.663594470046083, "grad_norm": 0.1131928414106369, "learning_rate": 3.6679821683583396e-05, "loss": 0.4413, "num_tokens": 385039566.0, "step": 486 }, { "epoch": 0.6649598907663424, "grad_norm": 0.10829973965883255, "learning_rate": 3.666446385212787e-05, "loss": 0.458, "num_tokens": 385819984.0, "step": 487 }, { "epoch": 0.6663253114866018, "grad_norm": 0.11053313314914703, "learning_rate": 3.664907420917252e-05, "loss": 0.4628, "num_tokens": 386634372.0, "step": 488 }, { "epoch": 0.6676907322068613, "grad_norm": 0.1296968013048172, "learning_rate": 3.6633652788101975e-05, "loss": 0.4892, "num_tokens": 387500745.0, "step": 489 }, { "epoch": 0.6690561529271206, "grad_norm": 0.10958334803581238, "learning_rate": 3.66181996223698e-05, "loss": 0.4746, "num_tokens": 388318419.0, "step": 490 }, { "epoch": 0.6704215736473801, "grad_norm": 0.11833487451076508, "learning_rate": 3.66027147454984e-05, "loss": 0.4673, "num_tokens": 389132686.0, "step": 491 }, { "epoch": 0.6717869943676396, "grad_norm": 0.12071003019809723, "learning_rate": 3.6587198191079014e-05, "loss": 0.4484, "num_tokens": 389859616.0, "step": 492 }, { "epoch": 0.6731524150878989, "grad_norm": 0.10776004940271378, "learning_rate": 3.657164999277156e-05, "loss": 0.4532, "num_tokens": 390677062.0, "step": 493 }, { "epoch": 0.6745178358081584, "grad_norm": 0.115529865026474, "learning_rate": 3.6556070184304624e-05, "loss": 0.4396, "num_tokens": 391459546.0, "step": 494 }, { "epoch": 0.6758832565284179, "grad_norm": 0.1182776466012001, "learning_rate": 3.654045879947535e-05, "loss": 0.4833, "num_tokens": 392221878.0, "step": 495 }, { "epoch": 0.6772486772486772, "grad_norm": 0.12117262929677963, "learning_rate": 3.652481587214939e-05, "loss": 0.4587, "num_tokens": 393026231.0, "step": 496 }, { "epoch": 0.6786140979689367, "grad_norm": 0.1088140606880188, "learning_rate": 3.650914143626081e-05, "loss": 0.4279, "num_tokens": 393796556.0, "step": 497 }, { "epoch": 0.6799795186891961, "grad_norm": 0.11418397724628448, "learning_rate": 3.6493435525812034e-05, "loss": 0.4686, "num_tokens": 394576697.0, "step": 498 }, { "epoch": 0.6813449394094555, "grad_norm": 0.1318536400794983, "learning_rate": 3.647769817487377e-05, "loss": 0.5031, "num_tokens": 395371861.0, "step": 499 }, { "epoch": 0.682710360129715, "grad_norm": 0.11794960498809814, "learning_rate": 3.6461929417584915e-05, "loss": 0.4446, "num_tokens": 396138886.0, "step": 500 }, { "epoch": 0.6840757808499744, "grad_norm": 0.11284691840410233, "learning_rate": 3.6446129288152514e-05, "loss": 0.4951, "num_tokens": 396977902.0, "step": 501 }, { "epoch": 0.6854412015702338, "grad_norm": 0.12376514822244644, "learning_rate": 3.643029782085165e-05, "loss": 0.4823, "num_tokens": 397750134.0, "step": 502 }, { "epoch": 0.6868066222904933, "grad_norm": 0.11219038814306259, "learning_rate": 3.64144350500254e-05, "loss": 0.4541, "num_tokens": 398488595.0, "step": 503 }, { "epoch": 0.6881720430107527, "grad_norm": 0.11093121021986008, "learning_rate": 3.639854101008472e-05, "loss": 0.4605, "num_tokens": 399316702.0, "step": 504 }, { "epoch": 0.6895374637310121, "grad_norm": 0.12141437083482742, "learning_rate": 3.638261573550844e-05, "loss": 0.4627, "num_tokens": 400012857.0, "step": 505 }, { "epoch": 0.6909028844512716, "grad_norm": 0.11223065108060837, "learning_rate": 3.636665926084313e-05, "loss": 0.4756, "num_tokens": 400820194.0, "step": 506 }, { "epoch": 0.6922683051715309, "grad_norm": 0.12112702429294586, "learning_rate": 3.6350671620703025e-05, "loss": 0.476, "num_tokens": 401637120.0, "step": 507 }, { "epoch": 0.6936337258917904, "grad_norm": 0.11513993889093399, "learning_rate": 3.633465284976998e-05, "loss": 0.4448, "num_tokens": 402317695.0, "step": 508 }, { "epoch": 0.6949991466120499, "grad_norm": 0.1166527196764946, "learning_rate": 3.631860298279338e-05, "loss": 0.471, "num_tokens": 403145111.0, "step": 509 }, { "epoch": 0.6963645673323092, "grad_norm": 0.10724858939647675, "learning_rate": 3.630252205459009e-05, "loss": 0.4477, "num_tokens": 403908423.0, "step": 510 }, { "epoch": 0.6977299880525687, "grad_norm": 0.1304098665714264, "learning_rate": 3.628641010004431e-05, "loss": 0.4545, "num_tokens": 404690467.0, "step": 511 }, { "epoch": 0.6990954087728282, "grad_norm": 0.1180436760187149, "learning_rate": 3.6270267154107576e-05, "loss": 0.498, "num_tokens": 405551385.0, "step": 512 }, { "epoch": 0.7004608294930875, "grad_norm": 0.11220570653676987, "learning_rate": 3.625409325179865e-05, "loss": 0.4853, "num_tokens": 406383570.0, "step": 513 }, { "epoch": 0.701826250213347, "grad_norm": 0.11797350645065308, "learning_rate": 3.623788842820344e-05, "loss": 0.4559, "num_tokens": 407156016.0, "step": 514 }, { "epoch": 0.7031916709336065, "grad_norm": 0.1125183179974556, "learning_rate": 3.622165271847494e-05, "loss": 0.4596, "num_tokens": 407892245.0, "step": 515 }, { "epoch": 0.7045570916538658, "grad_norm": 0.12060261517763138, "learning_rate": 3.6205386157833144e-05, "loss": 0.4871, "num_tokens": 408630930.0, "step": 516 }, { "epoch": 0.7059225123741253, "grad_norm": 0.11833718419075012, "learning_rate": 3.618908878156496e-05, "loss": 0.4586, "num_tokens": 409404146.0, "step": 517 }, { "epoch": 0.7072879330943848, "grad_norm": 0.12842777371406555, "learning_rate": 3.617276062502415e-05, "loss": 0.4899, "num_tokens": 410233454.0, "step": 518 }, { "epoch": 0.7086533538146441, "grad_norm": 0.12680326402187347, "learning_rate": 3.615640172363126e-05, "loss": 0.4622, "num_tokens": 410984862.0, "step": 519 }, { "epoch": 0.7100187745349036, "grad_norm": 0.12710969150066376, "learning_rate": 3.61400121128735e-05, "loss": 0.499, "num_tokens": 411761902.0, "step": 520 }, { "epoch": 0.7113841952551629, "grad_norm": 0.11364305764436722, "learning_rate": 3.6123591828304726e-05, "loss": 0.4642, "num_tokens": 412588366.0, "step": 521 }, { "epoch": 0.7127496159754224, "grad_norm": 0.1171727254986763, "learning_rate": 3.610714090554533e-05, "loss": 0.441, "num_tokens": 413372551.0, "step": 522 }, { "epoch": 0.7141150366956819, "grad_norm": 0.12130892276763916, "learning_rate": 3.609065938028216e-05, "loss": 0.4686, "num_tokens": 414123255.0, "step": 523 }, { "epoch": 0.7154804574159412, "grad_norm": 0.12494514882564545, "learning_rate": 3.6074147288268456e-05, "loss": 0.476, "num_tokens": 414911593.0, "step": 524 }, { "epoch": 0.7168458781362007, "grad_norm": 0.12262006103992462, "learning_rate": 3.605760466532374e-05, "loss": 0.4769, "num_tokens": 415667737.0, "step": 525 }, { "epoch": 0.7182112988564602, "grad_norm": 0.1096077412366867, "learning_rate": 3.604103154733382e-05, "loss": 0.4574, "num_tokens": 416432559.0, "step": 526 }, { "epoch": 0.7195767195767195, "grad_norm": 0.11143909394741058, "learning_rate": 3.602442797025061e-05, "loss": 0.4437, "num_tokens": 417204561.0, "step": 527 }, { "epoch": 0.720942140296979, "grad_norm": 0.11174468696117401, "learning_rate": 3.6007793970092095e-05, "loss": 0.4732, "num_tokens": 417988051.0, "step": 528 }, { "epoch": 0.7223075610172385, "grad_norm": 0.11228536069393158, "learning_rate": 3.599112958294229e-05, "loss": 0.4486, "num_tokens": 418779302.0, "step": 529 }, { "epoch": 0.7236729817374978, "grad_norm": 0.11849799007177353, "learning_rate": 3.597443484495112e-05, "loss": 0.4663, "num_tokens": 419567457.0, "step": 530 }, { "epoch": 0.7250384024577573, "grad_norm": 0.11003316938877106, "learning_rate": 3.595770979233433e-05, "loss": 0.4559, "num_tokens": 420368051.0, "step": 531 }, { "epoch": 0.7264038231780168, "grad_norm": 0.11033588647842407, "learning_rate": 3.594095446137343e-05, "loss": 0.4595, "num_tokens": 421099868.0, "step": 532 }, { "epoch": 0.7277692438982761, "grad_norm": 0.12446051836013794, "learning_rate": 3.592416888841564e-05, "loss": 0.4991, "num_tokens": 421942092.0, "step": 533 }, { "epoch": 0.7291346646185356, "grad_norm": 0.11612110584974289, "learning_rate": 3.5907353109873745e-05, "loss": 0.4675, "num_tokens": 422785753.0, "step": 534 }, { "epoch": 0.7305000853387951, "grad_norm": 0.11136294156312943, "learning_rate": 3.589050716222609e-05, "loss": 0.4619, "num_tokens": 423606299.0, "step": 535 }, { "epoch": 0.7318655060590544, "grad_norm": 0.11682203412055969, "learning_rate": 3.5873631082016445e-05, "loss": 0.4848, "num_tokens": 424360117.0, "step": 536 }, { "epoch": 0.7332309267793139, "grad_norm": 0.11769674718379974, "learning_rate": 3.585672490585394e-05, "loss": 0.4691, "num_tokens": 425166636.0, "step": 537 }, { "epoch": 0.7345963474995733, "grad_norm": 0.1243009939789772, "learning_rate": 3.583978867041302e-05, "loss": 0.4784, "num_tokens": 425973427.0, "step": 538 }, { "epoch": 0.7359617682198327, "grad_norm": 0.10507700592279434, "learning_rate": 3.5822822412433304e-05, "loss": 0.4657, "num_tokens": 426720036.0, "step": 539 }, { "epoch": 0.7373271889400922, "grad_norm": 0.10975233465433121, "learning_rate": 3.5805826168719564e-05, "loss": 0.47, "num_tokens": 427545403.0, "step": 540 }, { "epoch": 0.7386926096603516, "grad_norm": 0.11738420277833939, "learning_rate": 3.578879997614161e-05, "loss": 0.4901, "num_tokens": 428384320.0, "step": 541 }, { "epoch": 0.740058030380611, "grad_norm": 0.11441399157047272, "learning_rate": 3.577174387163421e-05, "loss": 0.4764, "num_tokens": 429204114.0, "step": 542 }, { "epoch": 0.7414234511008705, "grad_norm": 0.12232547253370285, "learning_rate": 3.5754657892197056e-05, "loss": 0.4685, "num_tokens": 429964635.0, "step": 543 }, { "epoch": 0.7427888718211298, "grad_norm": 0.12009816616773605, "learning_rate": 3.5737542074894595e-05, "loss": 0.458, "num_tokens": 430734728.0, "step": 544 }, { "epoch": 0.7441542925413893, "grad_norm": 0.11319591850042343, "learning_rate": 3.572039645685603e-05, "loss": 0.4882, "num_tokens": 431673075.0, "step": 545 }, { "epoch": 0.7455197132616488, "grad_norm": 0.1104775071144104, "learning_rate": 3.5703221075275227e-05, "loss": 0.4754, "num_tokens": 432436049.0, "step": 546 }, { "epoch": 0.7468851339819081, "grad_norm": 0.11132311075925827, "learning_rate": 3.5686015967410584e-05, "loss": 0.4619, "num_tokens": 433225490.0, "step": 547 }, { "epoch": 0.7482505547021676, "grad_norm": 0.11578576266765594, "learning_rate": 3.5668781170584996e-05, "loss": 0.4421, "num_tokens": 434000184.0, "step": 548 }, { "epoch": 0.7496159754224271, "grad_norm": 0.12022195756435394, "learning_rate": 3.565151672218578e-05, "loss": 0.4457, "num_tokens": 434704866.0, "step": 549 }, { "epoch": 0.7509813961426864, "grad_norm": 0.12704306840896606, "learning_rate": 3.563422265966454e-05, "loss": 0.4714, "num_tokens": 435493023.0, "step": 550 }, { "epoch": 0.7523468168629459, "grad_norm": 0.12024980783462524, "learning_rate": 3.5616899020537166e-05, "loss": 0.4481, "num_tokens": 436250412.0, "step": 551 }, { "epoch": 0.7537122375832054, "grad_norm": 0.1099313497543335, "learning_rate": 3.559954584238366e-05, "loss": 0.4523, "num_tokens": 437007876.0, "step": 552 }, { "epoch": 0.7550776583034647, "grad_norm": 0.11078646779060364, "learning_rate": 3.558216316284815e-05, "loss": 0.4686, "num_tokens": 437732327.0, "step": 553 }, { "epoch": 0.7564430790237242, "grad_norm": 0.11917616426944733, "learning_rate": 3.556475101963872e-05, "loss": 0.4504, "num_tokens": 438486918.0, "step": 554 }, { "epoch": 0.7578084997439836, "grad_norm": 0.11478155106306076, "learning_rate": 3.5547309450527394e-05, "loss": 0.4616, "num_tokens": 439267058.0, "step": 555 }, { "epoch": 0.759173920464243, "grad_norm": 0.11705777794122696, "learning_rate": 3.5529838493350034e-05, "loss": 0.4624, "num_tokens": 440066254.0, "step": 556 }, { "epoch": 0.7605393411845025, "grad_norm": 0.11049804091453552, "learning_rate": 3.551233818600623e-05, "loss": 0.4347, "num_tokens": 440813273.0, "step": 557 }, { "epoch": 0.7619047619047619, "grad_norm": 0.11581044644117355, "learning_rate": 3.5494808566459267e-05, "loss": 0.4591, "num_tokens": 441545387.0, "step": 558 }, { "epoch": 0.7632701826250213, "grad_norm": 0.11534260958433151, "learning_rate": 3.5477249672735976e-05, "loss": 0.4522, "num_tokens": 442378053.0, "step": 559 }, { "epoch": 0.7646356033452808, "grad_norm": 0.11141332238912582, "learning_rate": 3.5459661542926756e-05, "loss": 0.4476, "num_tokens": 443126194.0, "step": 560 }, { "epoch": 0.7660010240655402, "grad_norm": 0.12198268622159958, "learning_rate": 3.5442044215185363e-05, "loss": 0.454, "num_tokens": 443881072.0, "step": 561 }, { "epoch": 0.7673664447857996, "grad_norm": 0.10830750316381454, "learning_rate": 3.542439772772893e-05, "loss": 0.4557, "num_tokens": 444662023.0, "step": 562 }, { "epoch": 0.7687318655060591, "grad_norm": 0.11014734953641891, "learning_rate": 3.540672211883783e-05, "loss": 0.4646, "num_tokens": 445555182.0, "step": 563 }, { "epoch": 0.7700972862263185, "grad_norm": 0.12126334756612778, "learning_rate": 3.538901742685563e-05, "loss": 0.4556, "num_tokens": 446347138.0, "step": 564 }, { "epoch": 0.7714627069465779, "grad_norm": 0.11031445860862732, "learning_rate": 3.537128369018895e-05, "loss": 0.4593, "num_tokens": 447154379.0, "step": 565 }, { "epoch": 0.7728281276668374, "grad_norm": 0.11559077352285385, "learning_rate": 3.535352094730746e-05, "loss": 0.4552, "num_tokens": 447926931.0, "step": 566 }, { "epoch": 0.7741935483870968, "grad_norm": 0.11568443477153778, "learning_rate": 3.533572923674372e-05, "loss": 0.4747, "num_tokens": 448755226.0, "step": 567 }, { "epoch": 0.7755589691073562, "grad_norm": 0.10809071362018585, "learning_rate": 3.531790859709313e-05, "loss": 0.4658, "num_tokens": 449550443.0, "step": 568 }, { "epoch": 0.7769243898276156, "grad_norm": 0.12662017345428467, "learning_rate": 3.5300059067013885e-05, "loss": 0.4632, "num_tokens": 450363415.0, "step": 569 }, { "epoch": 0.778289810547875, "grad_norm": 0.10905206948518753, "learning_rate": 3.528218068522679e-05, "loss": 0.4351, "num_tokens": 451085842.0, "step": 570 }, { "epoch": 0.7796552312681345, "grad_norm": 0.11559968441724777, "learning_rate": 3.526427349051528e-05, "loss": 0.4683, "num_tokens": 451920933.0, "step": 571 }, { "epoch": 0.7810206519883939, "grad_norm": 0.11436466872692108, "learning_rate": 3.52463375217253e-05, "loss": 0.4626, "num_tokens": 452759021.0, "step": 572 }, { "epoch": 0.7823860727086533, "grad_norm": 0.11434222757816315, "learning_rate": 3.522837281776519e-05, "loss": 0.4436, "num_tokens": 453509716.0, "step": 573 }, { "epoch": 0.7837514934289128, "grad_norm": 0.11853418499231339, "learning_rate": 3.521037941760562e-05, "loss": 0.4607, "num_tokens": 454339195.0, "step": 574 }, { "epoch": 0.7851169141491722, "grad_norm": 0.11093449592590332, "learning_rate": 3.5192357360279536e-05, "loss": 0.4664, "num_tokens": 455156513.0, "step": 575 }, { "epoch": 0.7864823348694316, "grad_norm": 0.12498512864112854, "learning_rate": 3.517430668488203e-05, "loss": 0.485, "num_tokens": 456063494.0, "step": 576 }, { "epoch": 0.7878477555896911, "grad_norm": 0.12736006081104279, "learning_rate": 3.5156227430570294e-05, "loss": 0.4732, "num_tokens": 456916086.0, "step": 577 }, { "epoch": 0.7892131763099505, "grad_norm": 0.11338227987289429, "learning_rate": 3.5138119636563495e-05, "loss": 0.4672, "num_tokens": 457760380.0, "step": 578 }, { "epoch": 0.7905785970302099, "grad_norm": 0.12746867537498474, "learning_rate": 3.5119983342142726e-05, "loss": 0.4784, "num_tokens": 458598544.0, "step": 579 }, { "epoch": 0.7919440177504694, "grad_norm": 0.13174785673618317, "learning_rate": 3.51018185866509e-05, "loss": 0.4698, "num_tokens": 459338363.0, "step": 580 }, { "epoch": 0.7933094384707288, "grad_norm": 0.11068687587976456, "learning_rate": 3.508362540949266e-05, "loss": 0.4557, "num_tokens": 460114221.0, "step": 581 }, { "epoch": 0.7946748591909882, "grad_norm": 0.11722614616155624, "learning_rate": 3.506540385013433e-05, "loss": 0.4658, "num_tokens": 460897942.0, "step": 582 }, { "epoch": 0.7960402799112477, "grad_norm": 0.12795697152614594, "learning_rate": 3.5047153948103775e-05, "loss": 0.4677, "num_tokens": 461654501.0, "step": 583 }, { "epoch": 0.7974057006315071, "grad_norm": 0.11356557905673981, "learning_rate": 3.502887574299036e-05, "loss": 0.4718, "num_tokens": 462514849.0, "step": 584 }, { "epoch": 0.7987711213517665, "grad_norm": 0.11711650341749191, "learning_rate": 3.501056927444485e-05, "loss": 0.4477, "num_tokens": 463281466.0, "step": 585 }, { "epoch": 0.8001365420720259, "grad_norm": 0.12362288683652878, "learning_rate": 3.499223458217931e-05, "loss": 0.4527, "num_tokens": 464011338.0, "step": 586 }, { "epoch": 0.8015019627922854, "grad_norm": 0.11874205619096756, "learning_rate": 3.497387170596703e-05, "loss": 0.4702, "num_tokens": 464778947.0, "step": 587 }, { "epoch": 0.8028673835125448, "grad_norm": 0.11979168653488159, "learning_rate": 3.4955480685642454e-05, "loss": 0.4658, "num_tokens": 465587117.0, "step": 588 }, { "epoch": 0.8042328042328042, "grad_norm": 0.1319931149482727, "learning_rate": 3.493706156110107e-05, "loss": 0.489, "num_tokens": 466364468.0, "step": 589 }, { "epoch": 0.8055982249530637, "grad_norm": 0.1215633749961853, "learning_rate": 3.491861437229933e-05, "loss": 0.497, "num_tokens": 467263988.0, "step": 590 }, { "epoch": 0.8069636456733231, "grad_norm": 0.13243712484836578, "learning_rate": 3.490013915925456e-05, "loss": 0.4638, "num_tokens": 468074575.0, "step": 591 }, { "epoch": 0.8083290663935825, "grad_norm": 0.1296439915895462, "learning_rate": 3.4881635962044894e-05, "loss": 0.4729, "num_tokens": 468856827.0, "step": 592 }, { "epoch": 0.809694487113842, "grad_norm": 0.10735549032688141, "learning_rate": 3.486310482080917e-05, "loss": 0.4479, "num_tokens": 469617883.0, "step": 593 }, { "epoch": 0.8110599078341014, "grad_norm": 0.12577994167804718, "learning_rate": 3.4844545775746835e-05, "loss": 0.4503, "num_tokens": 470454979.0, "step": 594 }, { "epoch": 0.8124253285543608, "grad_norm": 0.13775967061519623, "learning_rate": 3.482595886711786e-05, "loss": 0.4645, "num_tokens": 471212857.0, "step": 595 }, { "epoch": 0.8137907492746203, "grad_norm": 0.11470581591129303, "learning_rate": 3.480734413524268e-05, "loss": 0.4538, "num_tokens": 471992155.0, "step": 596 }, { "epoch": 0.8151561699948797, "grad_norm": 0.12183646112680435, "learning_rate": 3.478870162050209e-05, "loss": 0.4368, "num_tokens": 472772198.0, "step": 597 }, { "epoch": 0.8165215907151391, "grad_norm": 0.11565186828374863, "learning_rate": 3.4770031363337126e-05, "loss": 0.4568, "num_tokens": 473563244.0, "step": 598 }, { "epoch": 0.8178870114353985, "grad_norm": 0.11791719496250153, "learning_rate": 3.4751333404249026e-05, "loss": 0.4683, "num_tokens": 474356498.0, "step": 599 }, { "epoch": 0.819252432155658, "grad_norm": 0.11324689537286758, "learning_rate": 3.4732607783799125e-05, "loss": 0.4731, "num_tokens": 475192924.0, "step": 600 }, { "epoch": 0.8206178528759174, "grad_norm": 0.14114902913570404, "learning_rate": 3.4713854542608744e-05, "loss": 0.4714, "num_tokens": 475964234.0, "step": 601 }, { "epoch": 0.8219832735961768, "grad_norm": 0.12459276616573334, "learning_rate": 3.469507372135914e-05, "loss": 0.4739, "num_tokens": 476765905.0, "step": 602 }, { "epoch": 0.8233486943164362, "grad_norm": 0.1444176435470581, "learning_rate": 3.4676265360791396e-05, "loss": 0.468, "num_tokens": 477497369.0, "step": 603 }, { "epoch": 0.8247141150366957, "grad_norm": 0.11855731904506683, "learning_rate": 3.465742950170633e-05, "loss": 0.4483, "num_tokens": 478247746.0, "step": 604 }, { "epoch": 0.8260795357569551, "grad_norm": 0.125793918967247, "learning_rate": 3.4638566184964414e-05, "loss": 0.447, "num_tokens": 479061479.0, "step": 605 }, { "epoch": 0.8274449564772145, "grad_norm": 0.11707732081413269, "learning_rate": 3.4619675451485686e-05, "loss": 0.4728, "num_tokens": 479936071.0, "step": 606 }, { "epoch": 0.828810377197474, "grad_norm": 0.12941159307956696, "learning_rate": 3.460075734224966e-05, "loss": 0.4417, "num_tokens": 480702703.0, "step": 607 }, { "epoch": 0.8301757979177334, "grad_norm": 0.1327252984046936, "learning_rate": 3.458181189829523e-05, "loss": 0.4628, "num_tokens": 481474943.0, "step": 608 }, { "epoch": 0.8315412186379928, "grad_norm": 0.12540404498577118, "learning_rate": 3.4562839160720586e-05, "loss": 0.4549, "num_tokens": 482259302.0, "step": 609 }, { "epoch": 0.8329066393582523, "grad_norm": 0.12677837908267975, "learning_rate": 3.4543839170683135e-05, "loss": 0.4728, "num_tokens": 483122182.0, "step": 610 }, { "epoch": 0.8342720600785117, "grad_norm": 0.11998961120843887, "learning_rate": 3.45248119693994e-05, "loss": 0.4728, "num_tokens": 483942345.0, "step": 611 }, { "epoch": 0.8356374807987711, "grad_norm": 0.1441909819841385, "learning_rate": 3.4505757598144926e-05, "loss": 0.4694, "num_tokens": 484789857.0, "step": 612 }, { "epoch": 0.8370029015190306, "grad_norm": 0.10763027518987656, "learning_rate": 3.4486676098254206e-05, "loss": 0.417, "num_tokens": 485485142.0, "step": 613 }, { "epoch": 0.83836832223929, "grad_norm": 0.141633540391922, "learning_rate": 3.4467567511120577e-05, "loss": 0.4537, "num_tokens": 486274385.0, "step": 614 }, { "epoch": 0.8397337429595494, "grad_norm": 0.11985165625810623, "learning_rate": 3.444843187819614e-05, "loss": 0.4599, "num_tokens": 487076335.0, "step": 615 }, { "epoch": 0.8410991636798089, "grad_norm": 0.14960482716560364, "learning_rate": 3.442926924099167e-05, "loss": 0.4437, "num_tokens": 487799125.0, "step": 616 }, { "epoch": 0.8424645844000682, "grad_norm": 0.11227377504110336, "learning_rate": 3.441007964107651e-05, "loss": 0.4571, "num_tokens": 488613366.0, "step": 617 }, { "epoch": 0.8438300051203277, "grad_norm": 0.12462513893842697, "learning_rate": 3.43908631200785e-05, "loss": 0.4658, "num_tokens": 489356121.0, "step": 618 }, { "epoch": 0.8451954258405872, "grad_norm": 0.1364748477935791, "learning_rate": 3.437161971968388e-05, "loss": 0.4638, "num_tokens": 490218511.0, "step": 619 }, { "epoch": 0.8465608465608465, "grad_norm": 0.12243972718715668, "learning_rate": 3.435234948163721e-05, "loss": 0.4452, "num_tokens": 491003541.0, "step": 620 }, { "epoch": 0.847926267281106, "grad_norm": 0.12146487832069397, "learning_rate": 3.433305244774124e-05, "loss": 0.4778, "num_tokens": 491859418.0, "step": 621 }, { "epoch": 0.8492916880013655, "grad_norm": 0.11502723395824432, "learning_rate": 3.4313728659856873e-05, "loss": 0.4588, "num_tokens": 492623443.0, "step": 622 }, { "epoch": 0.8506571087216248, "grad_norm": 0.12581685185432434, "learning_rate": 3.4294378159903044e-05, "loss": 0.4786, "num_tokens": 493417315.0, "step": 623 }, { "epoch": 0.8520225294418843, "grad_norm": 0.1123691201210022, "learning_rate": 3.4275000989856634e-05, "loss": 0.452, "num_tokens": 494139379.0, "step": 624 }, { "epoch": 0.8533879501621437, "grad_norm": 0.12129925191402435, "learning_rate": 3.4255597191752365e-05, "loss": 0.4817, "num_tokens": 494971383.0, "step": 625 }, { "epoch": 0.8547533708824031, "grad_norm": 0.12296999990940094, "learning_rate": 3.423616680768276e-05, "loss": 0.4632, "num_tokens": 495752309.0, "step": 626 }, { "epoch": 0.8561187916026626, "grad_norm": 0.13785012066364288, "learning_rate": 3.421670987979796e-05, "loss": 0.4607, "num_tokens": 496580248.0, "step": 627 }, { "epoch": 0.857484212322922, "grad_norm": 0.11436482518911362, "learning_rate": 3.419722645030574e-05, "loss": 0.4848, "num_tokens": 497477553.0, "step": 628 }, { "epoch": 0.8588496330431814, "grad_norm": 0.11406676471233368, "learning_rate": 3.417771656147133e-05, "loss": 0.4622, "num_tokens": 498286581.0, "step": 629 }, { "epoch": 0.8602150537634409, "grad_norm": 0.11792094260454178, "learning_rate": 3.4158180255617365e-05, "loss": 0.4694, "num_tokens": 499052435.0, "step": 630 }, { "epoch": 0.8615804744837003, "grad_norm": 0.11038850992918015, "learning_rate": 3.413861757512381e-05, "loss": 0.4481, "num_tokens": 499816873.0, "step": 631 }, { "epoch": 0.8629458952039597, "grad_norm": 0.11146378517150879, "learning_rate": 3.411902856242782e-05, "loss": 0.4603, "num_tokens": 500612310.0, "step": 632 }, { "epoch": 0.8643113159242192, "grad_norm": 0.11141832172870636, "learning_rate": 3.4099413260023654e-05, "loss": 0.4596, "num_tokens": 501372246.0, "step": 633 }, { "epoch": 0.8656767366444785, "grad_norm": 0.11945119500160217, "learning_rate": 3.407977171046267e-05, "loss": 0.4868, "num_tokens": 502236216.0, "step": 634 }, { "epoch": 0.867042157364738, "grad_norm": 0.11724088340997696, "learning_rate": 3.406010395635307e-05, "loss": 0.4795, "num_tokens": 503012542.0, "step": 635 }, { "epoch": 0.8684075780849975, "grad_norm": 0.1085352972149849, "learning_rate": 3.404041004035998e-05, "loss": 0.4542, "num_tokens": 503810590.0, "step": 636 }, { "epoch": 0.8697729988052568, "grad_norm": 0.10321240872144699, "learning_rate": 3.4020690005205245e-05, "loss": 0.4587, "num_tokens": 504650413.0, "step": 637 }, { "epoch": 0.8711384195255163, "grad_norm": 0.12369442731142044, "learning_rate": 3.400094389366737e-05, "loss": 0.4561, "num_tokens": 505380996.0, "step": 638 }, { "epoch": 0.8725038402457758, "grad_norm": 0.10374335944652557, "learning_rate": 3.3981171748581434e-05, "loss": 0.4446, "num_tokens": 506230685.0, "step": 639 }, { "epoch": 0.8738692609660351, "grad_norm": 0.11708637326955795, "learning_rate": 3.396137361283898e-05, "loss": 0.4448, "num_tokens": 507038971.0, "step": 640 }, { "epoch": 0.8752346816862946, "grad_norm": 0.11119847744703293, "learning_rate": 3.3941549529387936e-05, "loss": 0.4585, "num_tokens": 507830039.0, "step": 641 }, { "epoch": 0.8766001024065541, "grad_norm": 0.10514701902866364, "learning_rate": 3.392169954123253e-05, "loss": 0.4451, "num_tokens": 508602614.0, "step": 642 }, { "epoch": 0.8779655231268134, "grad_norm": 0.10346753150224686, "learning_rate": 3.390182369143319e-05, "loss": 0.4498, "num_tokens": 509459114.0, "step": 643 }, { "epoch": 0.8793309438470729, "grad_norm": 0.11535058170557022, "learning_rate": 3.3881922023106414e-05, "loss": 0.4561, "num_tokens": 510286640.0, "step": 644 }, { "epoch": 0.8806963645673324, "grad_norm": 0.1082838699221611, "learning_rate": 3.3861994579424725e-05, "loss": 0.4545, "num_tokens": 511009781.0, "step": 645 }, { "epoch": 0.8820617852875917, "grad_norm": 0.11458565294742584, "learning_rate": 3.384204140361657e-05, "loss": 0.4674, "num_tokens": 511872083.0, "step": 646 }, { "epoch": 0.8834272060078512, "grad_norm": 0.11582283675670624, "learning_rate": 3.38220625389662e-05, "loss": 0.4868, "num_tokens": 512695192.0, "step": 647 }, { "epoch": 0.8847926267281107, "grad_norm": 0.11874864995479584, "learning_rate": 3.380205802881362e-05, "loss": 0.4748, "num_tokens": 513581628.0, "step": 648 }, { "epoch": 0.88615804744837, "grad_norm": 0.10694429278373718, "learning_rate": 3.378202791655441e-05, "loss": 0.4313, "num_tokens": 514382407.0, "step": 649 }, { "epoch": 0.8875234681686295, "grad_norm": 0.10390761494636536, "learning_rate": 3.376197224563977e-05, "loss": 0.4581, "num_tokens": 515166821.0, "step": 650 }, { "epoch": 0.8888888888888888, "grad_norm": 0.13072054088115692, "learning_rate": 3.374189105957627e-05, "loss": 0.4626, "num_tokens": 515974328.0, "step": 651 }, { "epoch": 0.8902543096091483, "grad_norm": 0.11440446972846985, "learning_rate": 3.372178440192588e-05, "loss": 0.4702, "num_tokens": 516806639.0, "step": 652 }, { "epoch": 0.8916197303294078, "grad_norm": 0.10949018597602844, "learning_rate": 3.3701652316305784e-05, "loss": 0.4557, "num_tokens": 517660816.0, "step": 653 }, { "epoch": 0.8929851510496671, "grad_norm": 0.11035680025815964, "learning_rate": 3.3681494846388365e-05, "loss": 0.4578, "num_tokens": 518459802.0, "step": 654 }, { "epoch": 0.8943505717699266, "grad_norm": 0.11676550656557083, "learning_rate": 3.366131203590106e-05, "loss": 0.4517, "num_tokens": 519265238.0, "step": 655 }, { "epoch": 0.8957159924901861, "grad_norm": 0.10526785999536514, "learning_rate": 3.364110392862625e-05, "loss": 0.4752, "num_tokens": 520153394.0, "step": 656 }, { "epoch": 0.8970814132104454, "grad_norm": 0.1325811743736267, "learning_rate": 3.362087056840123e-05, "loss": 0.462, "num_tokens": 520924635.0, "step": 657 }, { "epoch": 0.8984468339307049, "grad_norm": 0.11251124739646912, "learning_rate": 3.360061199911807e-05, "loss": 0.4565, "num_tokens": 521809493.0, "step": 658 }, { "epoch": 0.8998122546509644, "grad_norm": 0.11627322435379028, "learning_rate": 3.358032826472349e-05, "loss": 0.4632, "num_tokens": 522574517.0, "step": 659 }, { "epoch": 0.9011776753712237, "grad_norm": 0.11396261304616928, "learning_rate": 3.3560019409218845e-05, "loss": 0.449, "num_tokens": 523350113.0, "step": 660 }, { "epoch": 0.9025430960914832, "grad_norm": 0.11369147151708603, "learning_rate": 3.3539685476659964e-05, "loss": 0.4572, "num_tokens": 524203752.0, "step": 661 }, { "epoch": 0.9039085168117427, "grad_norm": 0.10942088812589645, "learning_rate": 3.351932651115708e-05, "loss": 0.4614, "num_tokens": 524969300.0, "step": 662 }, { "epoch": 0.905273937532002, "grad_norm": 0.11247974634170532, "learning_rate": 3.3498942556874724e-05, "loss": 0.4549, "num_tokens": 525763253.0, "step": 663 }, { "epoch": 0.9066393582522615, "grad_norm": 0.10730231553316116, "learning_rate": 3.3478533658031645e-05, "loss": 0.45, "num_tokens": 526548055.0, "step": 664 }, { "epoch": 0.9080047789725209, "grad_norm": 0.11326640844345093, "learning_rate": 3.345809985890069e-05, "loss": 0.4474, "num_tokens": 527285929.0, "step": 665 }, { "epoch": 0.9093701996927803, "grad_norm": 0.11290981620550156, "learning_rate": 3.343764120380873e-05, "loss": 0.46, "num_tokens": 528039458.0, "step": 666 }, { "epoch": 0.9107356204130398, "grad_norm": 0.10783292353153229, "learning_rate": 3.341715773713657e-05, "loss": 0.4564, "num_tokens": 528869150.0, "step": 667 }, { "epoch": 0.9121010411332992, "grad_norm": 0.12166076898574829, "learning_rate": 3.339664950331881e-05, "loss": 0.4584, "num_tokens": 529679423.0, "step": 668 }, { "epoch": 0.9134664618535586, "grad_norm": 0.11464548856019974, "learning_rate": 3.3376116546843784e-05, "loss": 0.4614, "num_tokens": 530493277.0, "step": 669 }, { "epoch": 0.9148318825738181, "grad_norm": 0.12845425307750702, "learning_rate": 3.335555891225349e-05, "loss": 0.4948, "num_tokens": 531301060.0, "step": 670 }, { "epoch": 0.9161973032940774, "grad_norm": 0.11455569416284561, "learning_rate": 3.333497664414341e-05, "loss": 0.4539, "num_tokens": 532055240.0, "step": 671 }, { "epoch": 0.9175627240143369, "grad_norm": 0.10505792498588562, "learning_rate": 3.33143697871625e-05, "loss": 0.4246, "num_tokens": 532836986.0, "step": 672 }, { "epoch": 0.9189281447345964, "grad_norm": 0.10845118016004562, "learning_rate": 3.3293738386013046e-05, "loss": 0.4495, "num_tokens": 533621672.0, "step": 673 }, { "epoch": 0.9202935654548557, "grad_norm": 0.11200494319200516, "learning_rate": 3.327308248545057e-05, "loss": 0.4459, "num_tokens": 534369867.0, "step": 674 }, { "epoch": 0.9216589861751152, "grad_norm": 0.1077418178319931, "learning_rate": 3.325240213028374e-05, "loss": 0.4959, "num_tokens": 535181213.0, "step": 675 }, { "epoch": 0.9230244068953747, "grad_norm": 0.10418538004159927, "learning_rate": 3.323169736537429e-05, "loss": 0.4484, "num_tokens": 535969004.0, "step": 676 }, { "epoch": 0.924389827615634, "grad_norm": 0.11140068620443344, "learning_rate": 3.321096823563689e-05, "loss": 0.4855, "num_tokens": 536746754.0, "step": 677 }, { "epoch": 0.9257552483358935, "grad_norm": 0.11707666516304016, "learning_rate": 3.3190214786039065e-05, "loss": 0.4638, "num_tokens": 537545303.0, "step": 678 }, { "epoch": 0.927120669056153, "grad_norm": 0.10054581612348557, "learning_rate": 3.316943706160111e-05, "loss": 0.4435, "num_tokens": 538290548.0, "step": 679 }, { "epoch": 0.9284860897764123, "grad_norm": 0.11315910518169403, "learning_rate": 3.314863510739597e-05, "loss": 0.4775, "num_tokens": 539068033.0, "step": 680 }, { "epoch": 0.9298515104966718, "grad_norm": 0.11001914739608765, "learning_rate": 3.312780896854915e-05, "loss": 0.4357, "num_tokens": 539811481.0, "step": 681 }, { "epoch": 0.9312169312169312, "grad_norm": 0.11392703652381897, "learning_rate": 3.310695869023861e-05, "loss": 0.4556, "num_tokens": 540597155.0, "step": 682 }, { "epoch": 0.9325823519371906, "grad_norm": 0.11599786579608917, "learning_rate": 3.308608431769469e-05, "loss": 0.4505, "num_tokens": 541380221.0, "step": 683 }, { "epoch": 0.9339477726574501, "grad_norm": 0.10585466772317886, "learning_rate": 3.30651858962e-05, "loss": 0.431, "num_tokens": 542184640.0, "step": 684 }, { "epoch": 0.9353131933777095, "grad_norm": 0.11063703149557114, "learning_rate": 3.304426347108932e-05, "loss": 0.4734, "num_tokens": 543036383.0, "step": 685 }, { "epoch": 0.9366786140979689, "grad_norm": 0.11668778955936432, "learning_rate": 3.302331708774947e-05, "loss": 0.4597, "num_tokens": 543806341.0, "step": 686 }, { "epoch": 0.9380440348182284, "grad_norm": 0.12304060161113739, "learning_rate": 3.300234679161928e-05, "loss": 0.491, "num_tokens": 544661203.0, "step": 687 }, { "epoch": 0.9394094555384878, "grad_norm": 0.10760016739368439, "learning_rate": 3.298135262818944e-05, "loss": 0.4583, "num_tokens": 545504542.0, "step": 688 }, { "epoch": 0.9407748762587472, "grad_norm": 0.11730523407459259, "learning_rate": 3.296033464300241e-05, "loss": 0.4556, "num_tokens": 546304403.0, "step": 689 }, { "epoch": 0.9421402969790067, "grad_norm": 0.11088015139102936, "learning_rate": 3.293929288165234e-05, "loss": 0.4574, "num_tokens": 547060229.0, "step": 690 }, { "epoch": 0.943505717699266, "grad_norm": 0.1110675185918808, "learning_rate": 3.291822738978494e-05, "loss": 0.4536, "num_tokens": 547916549.0, "step": 691 }, { "epoch": 0.9448711384195255, "grad_norm": 0.11330344527959824, "learning_rate": 3.289713821309741e-05, "loss": 0.4765, "num_tokens": 548722315.0, "step": 692 }, { "epoch": 0.946236559139785, "grad_norm": 0.11284665018320084, "learning_rate": 3.287602539733832e-05, "loss": 0.4571, "num_tokens": 549493024.0, "step": 693 }, { "epoch": 0.9476019798600444, "grad_norm": 0.11617136746644974, "learning_rate": 3.2854888988307536e-05, "loss": 0.4724, "num_tokens": 550260246.0, "step": 694 }, { "epoch": 0.9489674005803038, "grad_norm": 0.11260492354631424, "learning_rate": 3.2833729031856086e-05, "loss": 0.4452, "num_tokens": 550998721.0, "step": 695 }, { "epoch": 0.9503328213005633, "grad_norm": 0.10697690397500992, "learning_rate": 3.281254557388609e-05, "loss": 0.4656, "num_tokens": 551779998.0, "step": 696 }, { "epoch": 0.9516982420208226, "grad_norm": 0.11812393367290497, "learning_rate": 3.279133866035065e-05, "loss": 0.4726, "num_tokens": 552566383.0, "step": 697 }, { "epoch": 0.9530636627410821, "grad_norm": 0.13856619596481323, "learning_rate": 3.277010833725375e-05, "loss": 0.4665, "num_tokens": 553369153.0, "step": 698 }, { "epoch": 0.9544290834613415, "grad_norm": 0.11128226667642593, "learning_rate": 3.2748854650650156e-05, "loss": 0.4573, "num_tokens": 554144592.0, "step": 699 }, { "epoch": 0.955794504181601, "grad_norm": 0.12070026993751526, "learning_rate": 3.27275776466453e-05, "loss": 0.4426, "num_tokens": 554927484.0, "step": 700 }, { "epoch": 0.9571599249018604, "grad_norm": 0.1045927107334137, "learning_rate": 3.270627737139522e-05, "loss": 0.4265, "num_tokens": 555711746.0, "step": 701 }, { "epoch": 0.9585253456221198, "grad_norm": 0.1139330342411995, "learning_rate": 3.268495387110644e-05, "loss": 0.4643, "num_tokens": 556568564.0, "step": 702 }, { "epoch": 0.9598907663423792, "grad_norm": 0.12209394574165344, "learning_rate": 3.266360719203582e-05, "loss": 0.444, "num_tokens": 557390732.0, "step": 703 }, { "epoch": 0.9612561870626387, "grad_norm": 0.11047131568193436, "learning_rate": 3.264223738049056e-05, "loss": 0.4747, "num_tokens": 558211621.0, "step": 704 }, { "epoch": 0.9626216077828981, "grad_norm": 0.10255604237318039, "learning_rate": 3.262084448282801e-05, "loss": 0.4311, "num_tokens": 558954534.0, "step": 705 }, { "epoch": 0.9639870285031575, "grad_norm": 0.11644338816404343, "learning_rate": 3.259942854545559e-05, "loss": 0.4668, "num_tokens": 559792223.0, "step": 706 }, { "epoch": 0.965352449223417, "grad_norm": 0.11366729438304901, "learning_rate": 3.257798961483074e-05, "loss": 0.435, "num_tokens": 560465831.0, "step": 707 }, { "epoch": 0.9667178699436764, "grad_norm": 0.11976787447929382, "learning_rate": 3.2556527737460744e-05, "loss": 0.4747, "num_tokens": 561230502.0, "step": 708 }, { "epoch": 0.9680832906639358, "grad_norm": 0.10525017231702805, "learning_rate": 3.2535042959902663e-05, "loss": 0.4509, "num_tokens": 561984119.0, "step": 709 }, { "epoch": 0.9694487113841953, "grad_norm": 0.12042440474033356, "learning_rate": 3.2513535328763264e-05, "loss": 0.4506, "num_tokens": 562775181.0, "step": 710 }, { "epoch": 0.9708141321044547, "grad_norm": 0.11168006807565689, "learning_rate": 3.249200489069886e-05, "loss": 0.4562, "num_tokens": 563597668.0, "step": 711 }, { "epoch": 0.9721795528247141, "grad_norm": 0.11944103240966797, "learning_rate": 3.247045169241526e-05, "loss": 0.5028, "num_tokens": 564399768.0, "step": 712 }, { "epoch": 0.9735449735449735, "grad_norm": 0.11592871695756912, "learning_rate": 3.2448875780667637e-05, "loss": 0.4398, "num_tokens": 565164650.0, "step": 713 }, { "epoch": 0.974910394265233, "grad_norm": 0.11733651906251907, "learning_rate": 3.242727720226044e-05, "loss": 0.4534, "num_tokens": 565870717.0, "step": 714 }, { "epoch": 0.9762758149854924, "grad_norm": 0.12647582590579987, "learning_rate": 3.240565600404727e-05, "loss": 0.469, "num_tokens": 566665118.0, "step": 715 }, { "epoch": 0.9776412357057518, "grad_norm": 0.11627350747585297, "learning_rate": 3.2384012232930835e-05, "loss": 0.4413, "num_tokens": 567407714.0, "step": 716 }, { "epoch": 0.9790066564260113, "grad_norm": 0.10581623017787933, "learning_rate": 3.236234593586278e-05, "loss": 0.4632, "num_tokens": 568286277.0, "step": 717 }, { "epoch": 0.9803720771462707, "grad_norm": 0.12117189168930054, "learning_rate": 3.234065715984361e-05, "loss": 0.4603, "num_tokens": 569034498.0, "step": 718 }, { "epoch": 0.9817374978665301, "grad_norm": 0.11610922962427139, "learning_rate": 3.2318945951922626e-05, "loss": 0.4535, "num_tokens": 569782220.0, "step": 719 }, { "epoch": 0.9831029185867896, "grad_norm": 0.11887255311012268, "learning_rate": 3.2297212359197765e-05, "loss": 0.4524, "num_tokens": 570604213.0, "step": 720 }, { "epoch": 0.984468339307049, "grad_norm": 0.120740607380867, "learning_rate": 3.227545642881553e-05, "loss": 0.44, "num_tokens": 571330997.0, "step": 721 }, { "epoch": 0.9858337600273084, "grad_norm": 0.12224970012903214, "learning_rate": 3.2253678207970874e-05, "loss": 0.4792, "num_tokens": 572227227.0, "step": 722 }, { "epoch": 0.9871991807475679, "grad_norm": 0.1129923164844513, "learning_rate": 3.223187774390712e-05, "loss": 0.4446, "num_tokens": 572980517.0, "step": 723 }, { "epoch": 0.9885646014678273, "grad_norm": 0.11229727417230606, "learning_rate": 3.221005508391583e-05, "loss": 0.4412, "num_tokens": 573776549.0, "step": 724 }, { "epoch": 0.9899300221880867, "grad_norm": 0.10871855169534683, "learning_rate": 3.218821027533671e-05, "loss": 0.4516, "num_tokens": 574592007.0, "step": 725 }, { "epoch": 0.9912954429083461, "grad_norm": 0.10619313269853592, "learning_rate": 3.2166343365557534e-05, "loss": 0.4489, "num_tokens": 575339898.0, "step": 726 }, { "epoch": 0.9926608636286056, "grad_norm": 0.11651577055454254, "learning_rate": 3.214445440201401e-05, "loss": 0.4707, "num_tokens": 576108560.0, "step": 727 }, { "epoch": 0.994026284348865, "grad_norm": 0.11979848146438599, "learning_rate": 3.2122543432189685e-05, "loss": 0.4862, "num_tokens": 576901545.0, "step": 728 }, { "epoch": 0.9953917050691244, "grad_norm": 0.10968338698148727, "learning_rate": 3.210061050361583e-05, "loss": 0.4467, "num_tokens": 577667593.0, "step": 729 }, { "epoch": 0.9967571257893838, "grad_norm": 0.11466460675001144, "learning_rate": 3.207865566387138e-05, "loss": 0.4511, "num_tokens": 578505417.0, "step": 730 }, { "epoch": 0.9981225465096433, "grad_norm": 0.10736589878797531, "learning_rate": 3.2056678960582776e-05, "loss": 0.4586, "num_tokens": 579272925.0, "step": 731 }, { "epoch": 0.9994879672299027, "grad_norm": 0.10721560567617416, "learning_rate": 3.203468044142391e-05, "loss": 0.4492, "num_tokens": 580063320.0, "step": 732 }, { "epoch": 1.0, "grad_norm": 0.10721560567617416, "learning_rate": 3.2012660154115974e-05, "loss": 0.4664, "num_tokens": 580282057.0, "step": 733 }, { "epoch": 1.0013654207202594, "grad_norm": 0.1727520227432251, "learning_rate": 3.19906181464274e-05, "loss": 0.4284, "num_tokens": 581067400.0, "step": 734 }, { "epoch": 1.002730841440519, "grad_norm": 0.1289745420217514, "learning_rate": 3.1968554466173736e-05, "loss": 0.4489, "num_tokens": 581868341.0, "step": 735 }, { "epoch": 1.0040962621607783, "grad_norm": 0.12519559264183044, "learning_rate": 3.1946469161217534e-05, "loss": 0.4262, "num_tokens": 582584415.0, "step": 736 }, { "epoch": 1.0054616828810377, "grad_norm": 0.10837225615978241, "learning_rate": 3.192436227946828e-05, "loss": 0.4473, "num_tokens": 583391225.0, "step": 737 }, { "epoch": 1.0068271036012972, "grad_norm": 0.12996280193328857, "learning_rate": 3.190223386888223e-05, "loss": 0.4276, "num_tokens": 584163598.0, "step": 738 }, { "epoch": 1.0081925243215566, "grad_norm": 0.12366433441638947, "learning_rate": 3.188008397746236e-05, "loss": 0.4215, "num_tokens": 584922674.0, "step": 739 }, { "epoch": 1.009557945041816, "grad_norm": 0.11067067831754684, "learning_rate": 3.185791265325826e-05, "loss": 0.4424, "num_tokens": 585730191.0, "step": 740 }, { "epoch": 1.0109233657620755, "grad_norm": 0.13117508590221405, "learning_rate": 3.1835719944365994e-05, "loss": 0.4444, "num_tokens": 586537712.0, "step": 741 }, { "epoch": 1.012288786482335, "grad_norm": 0.11535347998142242, "learning_rate": 3.1813505898928017e-05, "loss": 0.4232, "num_tokens": 587407480.0, "step": 742 }, { "epoch": 1.0136542072025942, "grad_norm": 0.11509476602077484, "learning_rate": 3.179127056513309e-05, "loss": 0.4733, "num_tokens": 588195784.0, "step": 743 }, { "epoch": 1.0150196279228538, "grad_norm": 0.13017131388187408, "learning_rate": 3.176901399121611e-05, "loss": 0.4312, "num_tokens": 588969328.0, "step": 744 }, { "epoch": 1.0163850486431132, "grad_norm": 0.11499501019716263, "learning_rate": 3.1746736225458104e-05, "loss": 0.4348, "num_tokens": 589710448.0, "step": 745 }, { "epoch": 1.0177504693633725, "grad_norm": 0.11986736208200455, "learning_rate": 3.172443731618602e-05, "loss": 0.4184, "num_tokens": 590567266.0, "step": 746 }, { "epoch": 1.0191158900836321, "grad_norm": 0.10962419956922531, "learning_rate": 3.1702117311772725e-05, "loss": 0.4355, "num_tokens": 591362903.0, "step": 747 }, { "epoch": 1.0204813108038915, "grad_norm": 0.13433480262756348, "learning_rate": 3.16797762606368e-05, "loss": 0.4471, "num_tokens": 592198261.0, "step": 748 }, { "epoch": 1.0218467315241508, "grad_norm": 0.12164328247308731, "learning_rate": 3.165741421124251e-05, "loss": 0.4592, "num_tokens": 593018491.0, "step": 749 }, { "epoch": 1.0232121522444102, "grad_norm": 0.11811253428459167, "learning_rate": 3.163503121209966e-05, "loss": 0.4453, "num_tokens": 593832905.0, "step": 750 }, { "epoch": 1.0245775729646698, "grad_norm": 0.11906921863555908, "learning_rate": 3.161262731176351e-05, "loss": 0.4299, "num_tokens": 594630145.0, "step": 751 }, { "epoch": 1.0259429936849291, "grad_norm": 0.11274818331003189, "learning_rate": 3.159020255883464e-05, "loss": 0.4157, "num_tokens": 595353817.0, "step": 752 }, { "epoch": 1.0273084144051885, "grad_norm": 0.1162695363163948, "learning_rate": 3.15677570019589e-05, "loss": 0.4311, "num_tokens": 596113722.0, "step": 753 }, { "epoch": 1.028673835125448, "grad_norm": 0.1071300134062767, "learning_rate": 3.154529068982724e-05, "loss": 0.4603, "num_tokens": 596946163.0, "step": 754 }, { "epoch": 1.0300392558457074, "grad_norm": 0.1145242378115654, "learning_rate": 3.152280367117565e-05, "loss": 0.4343, "num_tokens": 597795740.0, "step": 755 }, { "epoch": 1.0314046765659668, "grad_norm": 0.10494213551282883, "learning_rate": 3.150029599478502e-05, "loss": 0.4312, "num_tokens": 598620339.0, "step": 756 }, { "epoch": 1.0327700972862264, "grad_norm": 0.1111769899725914, "learning_rate": 3.147776770948108e-05, "loss": 0.4346, "num_tokens": 599444342.0, "step": 757 }, { "epoch": 1.0341355180064857, "grad_norm": 0.10525674372911453, "learning_rate": 3.145521886413425e-05, "loss": 0.4544, "num_tokens": 600228411.0, "step": 758 }, { "epoch": 1.035500938726745, "grad_norm": 0.11027923226356506, "learning_rate": 3.143264950765954e-05, "loss": 0.4286, "num_tokens": 601133346.0, "step": 759 }, { "epoch": 1.0368663594470047, "grad_norm": 0.10458922386169434, "learning_rate": 3.1410059689016485e-05, "loss": 0.4312, "num_tokens": 601917178.0, "step": 760 }, { "epoch": 1.038231780167264, "grad_norm": 0.11405925452709198, "learning_rate": 3.138744945720899e-05, "loss": 0.4353, "num_tokens": 602679326.0, "step": 761 }, { "epoch": 1.0395972008875234, "grad_norm": 0.11770788580179214, "learning_rate": 3.136481886128523e-05, "loss": 0.4635, "num_tokens": 603469224.0, "step": 762 }, { "epoch": 1.040962621607783, "grad_norm": 0.12108530849218369, "learning_rate": 3.134216795033758e-05, "loss": 0.4243, "num_tokens": 604267016.0, "step": 763 }, { "epoch": 1.0423280423280423, "grad_norm": 0.10753355920314789, "learning_rate": 3.131949677350248e-05, "loss": 0.42, "num_tokens": 605042770.0, "step": 764 }, { "epoch": 1.0436934630483017, "grad_norm": 0.11862706393003464, "learning_rate": 3.12968053799603e-05, "loss": 0.433, "num_tokens": 605829265.0, "step": 765 }, { "epoch": 1.0450588837685613, "grad_norm": 0.1181521788239479, "learning_rate": 3.127409381893531e-05, "loss": 0.4382, "num_tokens": 606628098.0, "step": 766 }, { "epoch": 1.0464243044888206, "grad_norm": 0.1060904935002327, "learning_rate": 3.125136213969552e-05, "loss": 0.4583, "num_tokens": 607451366.0, "step": 767 }, { "epoch": 1.04778972520908, "grad_norm": 0.10791075974702835, "learning_rate": 3.1228610391552556e-05, "loss": 0.4446, "num_tokens": 608238948.0, "step": 768 }, { "epoch": 1.0491551459293396, "grad_norm": 0.11885756254196167, "learning_rate": 3.12058386238616e-05, "loss": 0.4232, "num_tokens": 608989128.0, "step": 769 }, { "epoch": 1.050520566649599, "grad_norm": 0.12080185860395432, "learning_rate": 3.1183046886021266e-05, "loss": 0.4631, "num_tokens": 609812552.0, "step": 770 }, { "epoch": 1.0518859873698583, "grad_norm": 0.10882670432329178, "learning_rate": 3.116023522747348e-05, "loss": 0.4102, "num_tokens": 610528969.0, "step": 771 }, { "epoch": 1.0532514080901179, "grad_norm": 0.13234961032867432, "learning_rate": 3.1137403697703376e-05, "loss": 0.4207, "num_tokens": 611323222.0, "step": 772 }, { "epoch": 1.0546168288103772, "grad_norm": 0.1150064542889595, "learning_rate": 3.111455234623921e-05, "loss": 0.4281, "num_tokens": 612116787.0, "step": 773 }, { "epoch": 1.0559822495306366, "grad_norm": 0.12247920036315918, "learning_rate": 3.109168122265222e-05, "loss": 0.4194, "num_tokens": 612940436.0, "step": 774 }, { "epoch": 1.0573476702508962, "grad_norm": 0.12803204357624054, "learning_rate": 3.106879037655656e-05, "loss": 0.4267, "num_tokens": 613738093.0, "step": 775 }, { "epoch": 1.0587130909711555, "grad_norm": 0.11331775784492493, "learning_rate": 3.1045879857609146e-05, "loss": 0.4038, "num_tokens": 614440034.0, "step": 776 }, { "epoch": 1.0600785116914149, "grad_norm": 0.11099238693714142, "learning_rate": 3.102294971550957e-05, "loss": 0.4726, "num_tokens": 615320263.0, "step": 777 }, { "epoch": 1.0614439324116745, "grad_norm": 0.1271698772907257, "learning_rate": 3.1e-05, "loss": 0.4349, "num_tokens": 616161957.0, "step": 778 }, { "epoch": 1.0628093531319338, "grad_norm": 0.10497069358825684, "learning_rate": 3.097703076086508e-05, "loss": 0.4156, "num_tokens": 616983170.0, "step": 779 }, { "epoch": 1.0641747738521932, "grad_norm": 0.14704136550426483, "learning_rate": 3.095404204793176e-05, "loss": 0.4425, "num_tokens": 617763729.0, "step": 780 }, { "epoch": 1.0655401945724527, "grad_norm": 0.1050507053732872, "learning_rate": 3.093103391106929e-05, "loss": 0.4444, "num_tokens": 618562540.0, "step": 781 }, { "epoch": 1.066905615292712, "grad_norm": 0.12067756056785583, "learning_rate": 3.090800640018903e-05, "loss": 0.4434, "num_tokens": 619458953.0, "step": 782 }, { "epoch": 1.0682710360129715, "grad_norm": 0.10909315943717957, "learning_rate": 3.088495956524435e-05, "loss": 0.4336, "num_tokens": 620291119.0, "step": 783 }, { "epoch": 1.0696364567332308, "grad_norm": 0.12649162113666534, "learning_rate": 3.086189345623058e-05, "loss": 0.4479, "num_tokens": 621102388.0, "step": 784 }, { "epoch": 1.0710018774534904, "grad_norm": 0.10843208432197571, "learning_rate": 3.083880812318482e-05, "loss": 0.4151, "num_tokens": 621881412.0, "step": 785 }, { "epoch": 1.0723672981737498, "grad_norm": 0.13368567824363708, "learning_rate": 3.0815703616185886e-05, "loss": 0.4222, "num_tokens": 622587541.0, "step": 786 }, { "epoch": 1.0737327188940091, "grad_norm": 0.10663442313671112, "learning_rate": 3.079257998535423e-05, "loss": 0.4177, "num_tokens": 623380285.0, "step": 787 }, { "epoch": 1.0750981396142687, "grad_norm": 0.11533872038125992, "learning_rate": 3.076943728085171e-05, "loss": 0.4449, "num_tokens": 624120468.0, "step": 788 }, { "epoch": 1.076463560334528, "grad_norm": 0.11374733597040176, "learning_rate": 3.074627555288163e-05, "loss": 0.4399, "num_tokens": 624926232.0, "step": 789 }, { "epoch": 1.0778289810547874, "grad_norm": 0.10639557242393494, "learning_rate": 3.072309485168853e-05, "loss": 0.4587, "num_tokens": 625745715.0, "step": 790 }, { "epoch": 1.079194401775047, "grad_norm": 0.09936220943927765, "learning_rate": 3.0699895227558105e-05, "loss": 0.4397, "num_tokens": 626535481.0, "step": 791 }, { "epoch": 1.0805598224953064, "grad_norm": 0.11979438364505768, "learning_rate": 3.067667673081711e-05, "loss": 0.4218, "num_tokens": 627292591.0, "step": 792 }, { "epoch": 1.0819252432155657, "grad_norm": 0.10676059126853943, "learning_rate": 3.065343941183324e-05, "loss": 0.4442, "num_tokens": 628060331.0, "step": 793 }, { "epoch": 1.0832906639358253, "grad_norm": 0.11738094687461853, "learning_rate": 3.063018332101502e-05, "loss": 0.439, "num_tokens": 628821908.0, "step": 794 }, { "epoch": 1.0846560846560847, "grad_norm": 0.1117849126458168, "learning_rate": 3.060690850881169e-05, "loss": 0.4221, "num_tokens": 629639920.0, "step": 795 }, { "epoch": 1.086021505376344, "grad_norm": 0.12161971628665924, "learning_rate": 3.058361502571312e-05, "loss": 0.4208, "num_tokens": 630372733.0, "step": 796 }, { "epoch": 1.0873869260966036, "grad_norm": 0.10490982979536057, "learning_rate": 3.056030292224964e-05, "loss": 0.4441, "num_tokens": 631146173.0, "step": 797 }, { "epoch": 1.088752346816863, "grad_norm": 0.11648774147033691, "learning_rate": 3.053697224899203e-05, "loss": 0.4509, "num_tokens": 631939303.0, "step": 798 }, { "epoch": 1.0901177675371223, "grad_norm": 0.12287832051515579, "learning_rate": 3.0513623056551323e-05, "loss": 0.4469, "num_tokens": 632761525.0, "step": 799 }, { "epoch": 1.0914831882573819, "grad_norm": 0.1051541417837143, "learning_rate": 3.0490255395578713e-05, "loss": 0.4109, "num_tokens": 633492889.0, "step": 800 }, { "epoch": 1.0928486089776412, "grad_norm": 0.1187848299741745, "learning_rate": 3.0466869316765487e-05, "loss": 0.4238, "num_tokens": 634231405.0, "step": 801 }, { "epoch": 1.0942140296979006, "grad_norm": 0.10816425830125809, "learning_rate": 3.0443464870842873e-05, "loss": 0.4159, "num_tokens": 635034883.0, "step": 802 }, { "epoch": 1.0955794504181602, "grad_norm": 0.12002389878034592, "learning_rate": 3.0420042108581934e-05, "loss": 0.4387, "num_tokens": 635851850.0, "step": 803 }, { "epoch": 1.0969448711384195, "grad_norm": 0.10779226571321487, "learning_rate": 3.039660108079348e-05, "loss": 0.4605, "num_tokens": 636660356.0, "step": 804 }, { "epoch": 1.098310291858679, "grad_norm": 0.12015695869922638, "learning_rate": 3.037314183832793e-05, "loss": 0.4138, "num_tokens": 637411850.0, "step": 805 }, { "epoch": 1.0996757125789385, "grad_norm": 0.10190171748399734, "learning_rate": 3.0349664432075227e-05, "loss": 0.4427, "num_tokens": 638337522.0, "step": 806 }, { "epoch": 1.1010411332991978, "grad_norm": 0.11264593154191971, "learning_rate": 3.0326168912964725e-05, "loss": 0.4062, "num_tokens": 639129354.0, "step": 807 }, { "epoch": 1.1024065540194572, "grad_norm": 0.10695818066596985, "learning_rate": 3.0302655331965044e-05, "loss": 0.4336, "num_tokens": 639930835.0, "step": 808 }, { "epoch": 1.1037719747397168, "grad_norm": 0.11429625749588013, "learning_rate": 3.0279123740084005e-05, "loss": 0.4496, "num_tokens": 640722963.0, "step": 809 }, { "epoch": 1.1051373954599761, "grad_norm": 0.1193954274058342, "learning_rate": 3.02555741883685e-05, "loss": 0.4287, "num_tokens": 641485325.0, "step": 810 }, { "epoch": 1.1065028161802355, "grad_norm": 0.11619765311479568, "learning_rate": 3.0232006727904368e-05, "loss": 0.4212, "num_tokens": 642206871.0, "step": 811 }, { "epoch": 1.107868236900495, "grad_norm": 0.10819490253925323, "learning_rate": 3.0208421409816303e-05, "loss": 0.4458, "num_tokens": 643062873.0, "step": 812 }, { "epoch": 1.1092336576207544, "grad_norm": 0.11174608767032623, "learning_rate": 3.018481828526776e-05, "loss": 0.428, "num_tokens": 643898720.0, "step": 813 }, { "epoch": 1.1105990783410138, "grad_norm": 0.11436447501182556, "learning_rate": 3.0161197405460774e-05, "loss": 0.4248, "num_tokens": 644625027.0, "step": 814 }, { "epoch": 1.1119644990612731, "grad_norm": 0.12968292832374573, "learning_rate": 3.013755882163595e-05, "loss": 0.4255, "num_tokens": 645391645.0, "step": 815 }, { "epoch": 1.1133299197815327, "grad_norm": 0.1049925833940506, "learning_rate": 3.0113902585072252e-05, "loss": 0.4146, "num_tokens": 646171211.0, "step": 816 }, { "epoch": 1.114695340501792, "grad_norm": 0.12944838404655457, "learning_rate": 3.0090228747086968e-05, "loss": 0.4584, "num_tokens": 646974194.0, "step": 817 }, { "epoch": 1.1160607612220514, "grad_norm": 0.12091442197561264, "learning_rate": 3.0066537359035554e-05, "loss": 0.4412, "num_tokens": 647853935.0, "step": 818 }, { "epoch": 1.117426181942311, "grad_norm": 0.14090082049369812, "learning_rate": 3.0042828472311548e-05, "loss": 0.4539, "num_tokens": 648683508.0, "step": 819 }, { "epoch": 1.1187916026625704, "grad_norm": 0.10180014371871948, "learning_rate": 3.0019102138346427e-05, "loss": 0.441, "num_tokens": 649516172.0, "step": 820 }, { "epoch": 1.1201570233828297, "grad_norm": 0.1147070825099945, "learning_rate": 2.9995358408609547e-05, "loss": 0.4159, "num_tokens": 650271262.0, "step": 821 }, { "epoch": 1.1215224441030893, "grad_norm": 0.12652705609798431, "learning_rate": 2.9971597334607977e-05, "loss": 0.4376, "num_tokens": 651069702.0, "step": 822 }, { "epoch": 1.1228878648233487, "grad_norm": 0.10852767527103424, "learning_rate": 2.9947818967886416e-05, "loss": 0.4325, "num_tokens": 651875285.0, "step": 823 }, { "epoch": 1.124253285543608, "grad_norm": 0.11566869914531708, "learning_rate": 2.9924023360027078e-05, "loss": 0.4337, "num_tokens": 652646822.0, "step": 824 }, { "epoch": 1.1256187062638676, "grad_norm": 0.12914276123046875, "learning_rate": 2.9900210562649587e-05, "loss": 0.4129, "num_tokens": 653452644.0, "step": 825 }, { "epoch": 1.126984126984127, "grad_norm": 0.11778885126113892, "learning_rate": 2.9876380627410832e-05, "loss": 0.4551, "num_tokens": 654200606.0, "step": 826 }, { "epoch": 1.1283495477043863, "grad_norm": 0.10677200555801392, "learning_rate": 2.98525336060049e-05, "loss": 0.4271, "num_tokens": 654981082.0, "step": 827 }, { "epoch": 1.129714968424646, "grad_norm": 0.11517840623855591, "learning_rate": 2.9828669550162942e-05, "loss": 0.4265, "num_tokens": 655759983.0, "step": 828 }, { "epoch": 1.1310803891449053, "grad_norm": 0.11427383869886398, "learning_rate": 2.9804788511653054e-05, "loss": 0.4093, "num_tokens": 656463429.0, "step": 829 }, { "epoch": 1.1324458098651646, "grad_norm": 0.10794113576412201, "learning_rate": 2.9780890542280173e-05, "loss": 0.3918, "num_tokens": 657203221.0, "step": 830 }, { "epoch": 1.1338112305854242, "grad_norm": 0.10753177851438522, "learning_rate": 2.975697569388596e-05, "loss": 0.4094, "num_tokens": 657915956.0, "step": 831 }, { "epoch": 1.1351766513056836, "grad_norm": 0.11716162413358688, "learning_rate": 2.9733044018348707e-05, "loss": 0.4249, "num_tokens": 658671188.0, "step": 832 }, { "epoch": 1.136542072025943, "grad_norm": 0.11125216633081436, "learning_rate": 2.9709095567583194e-05, "loss": 0.4285, "num_tokens": 659417387.0, "step": 833 }, { "epoch": 1.1379074927462025, "grad_norm": 0.12217295169830322, "learning_rate": 2.96851303935406e-05, "loss": 0.4417, "num_tokens": 660166645.0, "step": 834 }, { "epoch": 1.1392729134664619, "grad_norm": 0.1153254434466362, "learning_rate": 2.966114854820837e-05, "loss": 0.4468, "num_tokens": 660921528.0, "step": 835 }, { "epoch": 1.1406383341867212, "grad_norm": 0.11246630549430847, "learning_rate": 2.9637150083610136e-05, "loss": 0.4218, "num_tokens": 661723902.0, "step": 836 }, { "epoch": 1.1420037549069808, "grad_norm": 0.10594592243432999, "learning_rate": 2.961313505180556e-05, "loss": 0.43, "num_tokens": 662547258.0, "step": 837 }, { "epoch": 1.1433691756272402, "grad_norm": 0.11457372456789017, "learning_rate": 2.9589103504890254e-05, "loss": 0.4251, "num_tokens": 663342588.0, "step": 838 }, { "epoch": 1.1447345963474995, "grad_norm": 0.12373913824558258, "learning_rate": 2.9565055494995645e-05, "loss": 0.4416, "num_tokens": 664195675.0, "step": 839 }, { "epoch": 1.1461000170677589, "grad_norm": 0.102802574634552, "learning_rate": 2.9540991074288895e-05, "loss": 0.4391, "num_tokens": 664964128.0, "step": 840 }, { "epoch": 1.1474654377880185, "grad_norm": 0.11161854863166809, "learning_rate": 2.9516910294972745e-05, "loss": 0.4308, "num_tokens": 665778002.0, "step": 841 }, { "epoch": 1.1488308585082778, "grad_norm": 0.10945606976747513, "learning_rate": 2.9492813209285428e-05, "loss": 0.4285, "num_tokens": 666591923.0, "step": 842 }, { "epoch": 1.1501962792285374, "grad_norm": 0.09887667745351791, "learning_rate": 2.9468699869500552e-05, "loss": 0.4337, "num_tokens": 667407564.0, "step": 843 }, { "epoch": 1.1515616999487968, "grad_norm": 0.12268222868442535, "learning_rate": 2.9444570327926983e-05, "loss": 0.4542, "num_tokens": 668220160.0, "step": 844 }, { "epoch": 1.1529271206690561, "grad_norm": 0.10903208702802658, "learning_rate": 2.9420424636908755e-05, "loss": 0.4435, "num_tokens": 669076230.0, "step": 845 }, { "epoch": 1.1542925413893155, "grad_norm": 0.09536011517047882, "learning_rate": 2.9396262848824887e-05, "loss": 0.4227, "num_tokens": 669905056.0, "step": 846 }, { "epoch": 1.155657962109575, "grad_norm": 0.10977378487586975, "learning_rate": 2.937208501608936e-05, "loss": 0.4319, "num_tokens": 670758927.0, "step": 847 }, { "epoch": 1.1570233828298344, "grad_norm": 0.10482428967952728, "learning_rate": 2.934789119115095e-05, "loss": 0.4288, "num_tokens": 671499026.0, "step": 848 }, { "epoch": 1.158388803550094, "grad_norm": 0.1125073954463005, "learning_rate": 2.9323681426493117e-05, "loss": 0.4333, "num_tokens": 672272538.0, "step": 849 }, { "epoch": 1.1597542242703534, "grad_norm": 0.1028812974691391, "learning_rate": 2.9299455774633894e-05, "loss": 0.4418, "num_tokens": 673064366.0, "step": 850 }, { "epoch": 1.1611196449906127, "grad_norm": 0.12862148880958557, "learning_rate": 2.9275214288125794e-05, "loss": 0.4454, "num_tokens": 673825106.0, "step": 851 }, { "epoch": 1.162485065710872, "grad_norm": 0.11415711045265198, "learning_rate": 2.9250957019555674e-05, "loss": 0.4403, "num_tokens": 674615478.0, "step": 852 }, { "epoch": 1.1638504864311316, "grad_norm": 0.11945604532957077, "learning_rate": 2.9226684021544627e-05, "loss": 0.4226, "num_tokens": 675369115.0, "step": 853 }, { "epoch": 1.165215907151391, "grad_norm": 0.1197180226445198, "learning_rate": 2.9202395346747865e-05, "loss": 0.433, "num_tokens": 676141893.0, "step": 854 }, { "epoch": 1.1665813278716504, "grad_norm": 0.12143833935260773, "learning_rate": 2.9178091047854603e-05, "loss": 0.4507, "num_tokens": 676887538.0, "step": 855 }, { "epoch": 1.16794674859191, "grad_norm": 0.10418221354484558, "learning_rate": 2.915377117758796e-05, "loss": 0.4325, "num_tokens": 677687153.0, "step": 856 }, { "epoch": 1.1693121693121693, "grad_norm": 0.11161814630031586, "learning_rate": 2.9129435788704827e-05, "loss": 0.4544, "num_tokens": 678491181.0, "step": 857 }, { "epoch": 1.1706775900324287, "grad_norm": 0.12205932289361954, "learning_rate": 2.9105084933995763e-05, "loss": 0.4202, "num_tokens": 679314820.0, "step": 858 }, { "epoch": 1.1720430107526882, "grad_norm": 0.10707756876945496, "learning_rate": 2.908071866628487e-05, "loss": 0.4262, "num_tokens": 680096424.0, "step": 859 }, { "epoch": 1.1734084314729476, "grad_norm": 0.10238593816757202, "learning_rate": 2.9056337038429708e-05, "loss": 0.445, "num_tokens": 680903665.0, "step": 860 }, { "epoch": 1.174773852193207, "grad_norm": 0.12830592691898346, "learning_rate": 2.9031940103321115e-05, "loss": 0.4525, "num_tokens": 681712526.0, "step": 861 }, { "epoch": 1.1761392729134665, "grad_norm": 0.10117005556821823, "learning_rate": 2.9007527913883176e-05, "loss": 0.4317, "num_tokens": 682565997.0, "step": 862 }, { "epoch": 1.177504693633726, "grad_norm": 0.10368222743272781, "learning_rate": 2.8983100523073044e-05, "loss": 0.4275, "num_tokens": 683391965.0, "step": 863 }, { "epoch": 1.1788701143539853, "grad_norm": 0.10828981548547745, "learning_rate": 2.895865798388086e-05, "loss": 0.4049, "num_tokens": 684162155.0, "step": 864 }, { "epoch": 1.1802355350742448, "grad_norm": 0.10821667313575745, "learning_rate": 2.8934200349329623e-05, "loss": 0.4247, "num_tokens": 684948818.0, "step": 865 }, { "epoch": 1.1816009557945042, "grad_norm": 0.0981997400522232, "learning_rate": 2.8909727672475072e-05, "loss": 0.4605, "num_tokens": 685791713.0, "step": 866 }, { "epoch": 1.1829663765147636, "grad_norm": 0.11923803389072418, "learning_rate": 2.888524000640558e-05, "loss": 0.4425, "num_tokens": 686591354.0, "step": 867 }, { "epoch": 1.1843317972350231, "grad_norm": 0.09878721833229065, "learning_rate": 2.886073740424205e-05, "loss": 0.4519, "num_tokens": 687372756.0, "step": 868 }, { "epoch": 1.1856972179552825, "grad_norm": 0.11289258301258087, "learning_rate": 2.8836219919137764e-05, "loss": 0.4134, "num_tokens": 688122631.0, "step": 869 }, { "epoch": 1.1870626386755418, "grad_norm": 0.09934485703706741, "learning_rate": 2.8811687604278297e-05, "loss": 0.4326, "num_tokens": 688884022.0, "step": 870 }, { "epoch": 1.1884280593958014, "grad_norm": 0.11091029644012451, "learning_rate": 2.8787140512881398e-05, "loss": 0.4347, "num_tokens": 689784422.0, "step": 871 }, { "epoch": 1.1897934801160608, "grad_norm": 0.10345850139856339, "learning_rate": 2.8762578698196876e-05, "loss": 0.4281, "num_tokens": 690579553.0, "step": 872 }, { "epoch": 1.1911589008363201, "grad_norm": 0.1121901348233223, "learning_rate": 2.8738002213506462e-05, "loss": 0.4587, "num_tokens": 691313002.0, "step": 873 }, { "epoch": 1.1925243215565797, "grad_norm": 0.1007329598069191, "learning_rate": 2.8713411112123722e-05, "loss": 0.4166, "num_tokens": 692071531.0, "step": 874 }, { "epoch": 1.193889742276839, "grad_norm": 0.11033616214990616, "learning_rate": 2.868880544739394e-05, "loss": 0.4282, "num_tokens": 692858578.0, "step": 875 }, { "epoch": 1.1952551629970984, "grad_norm": 0.09891016036272049, "learning_rate": 2.866418527269396e-05, "loss": 0.4091, "num_tokens": 693638406.0, "step": 876 }, { "epoch": 1.1966205837173578, "grad_norm": 0.10362882912158966, "learning_rate": 2.8639550641432145e-05, "loss": 0.4457, "num_tokens": 694516711.0, "step": 877 }, { "epoch": 1.1979860044376174, "grad_norm": 0.13386715948581696, "learning_rate": 2.861490160704818e-05, "loss": 0.443, "num_tokens": 695304100.0, "step": 878 }, { "epoch": 1.1993514251578767, "grad_norm": 0.09622975438833237, "learning_rate": 2.859023822301302e-05, "loss": 0.4152, "num_tokens": 696084594.0, "step": 879 }, { "epoch": 1.2007168458781363, "grad_norm": 0.10887964814901352, "learning_rate": 2.856556054282873e-05, "loss": 0.436, "num_tokens": 696922679.0, "step": 880 }, { "epoch": 1.2020822665983957, "grad_norm": 0.1115502268075943, "learning_rate": 2.854086862002841e-05, "loss": 0.4312, "num_tokens": 697733834.0, "step": 881 }, { "epoch": 1.203447687318655, "grad_norm": 0.11597379297018051, "learning_rate": 2.8516162508176034e-05, "loss": 0.4443, "num_tokens": 698545731.0, "step": 882 }, { "epoch": 1.2048131080389144, "grad_norm": 0.10612155497074127, "learning_rate": 2.8491442260866372e-05, "loss": 0.4205, "num_tokens": 699257938.0, "step": 883 }, { "epoch": 1.206178528759174, "grad_norm": 0.10816977918148041, "learning_rate": 2.8466707931724836e-05, "loss": 0.4144, "num_tokens": 700080664.0, "step": 884 }, { "epoch": 1.2075439494794333, "grad_norm": 0.125789076089859, "learning_rate": 2.8441959574407415e-05, "loss": 0.4213, "num_tokens": 700868458.0, "step": 885 }, { "epoch": 1.2089093701996927, "grad_norm": 0.1027161106467247, "learning_rate": 2.841719724260052e-05, "loss": 0.4545, "num_tokens": 701616434.0, "step": 886 }, { "epoch": 1.2102747909199523, "grad_norm": 0.11393710970878601, "learning_rate": 2.8392420990020864e-05, "loss": 0.449, "num_tokens": 702347503.0, "step": 887 }, { "epoch": 1.2116402116402116, "grad_norm": 0.11594937741756439, "learning_rate": 2.8367630870415364e-05, "loss": 0.4195, "num_tokens": 703076915.0, "step": 888 }, { "epoch": 1.213005632360471, "grad_norm": 0.11371751129627228, "learning_rate": 2.834282693756102e-05, "loss": 0.4493, "num_tokens": 703901744.0, "step": 889 }, { "epoch": 1.2143710530807306, "grad_norm": 0.10476959496736526, "learning_rate": 2.8318009245264814e-05, "loss": 0.4328, "num_tokens": 704671295.0, "step": 890 }, { "epoch": 1.21573647380099, "grad_norm": 0.11183628439903259, "learning_rate": 2.829317784736353e-05, "loss": 0.435, "num_tokens": 705384296.0, "step": 891 }, { "epoch": 1.2171018945212493, "grad_norm": 0.11892073601484299, "learning_rate": 2.826833279772375e-05, "loss": 0.4359, "num_tokens": 706281848.0, "step": 892 }, { "epoch": 1.2184673152415089, "grad_norm": 0.10097572207450867, "learning_rate": 2.8243474150241602e-05, "loss": 0.4406, "num_tokens": 707091739.0, "step": 893 }, { "epoch": 1.2198327359617682, "grad_norm": 0.11000510305166245, "learning_rate": 2.821860195884276e-05, "loss": 0.4294, "num_tokens": 707873489.0, "step": 894 }, { "epoch": 1.2211981566820276, "grad_norm": 0.10158401727676392, "learning_rate": 2.819371627748226e-05, "loss": 0.4171, "num_tokens": 708615231.0, "step": 895 }, { "epoch": 1.2225635774022872, "grad_norm": 0.11211096495389938, "learning_rate": 2.816881716014441e-05, "loss": 0.4311, "num_tokens": 709382455.0, "step": 896 }, { "epoch": 1.2239289981225465, "grad_norm": 0.10477086156606674, "learning_rate": 2.814390466084264e-05, "loss": 0.4359, "num_tokens": 710214956.0, "step": 897 }, { "epoch": 1.2252944188428059, "grad_norm": 0.11373650282621384, "learning_rate": 2.8118978833619458e-05, "loss": 0.434, "num_tokens": 710936541.0, "step": 898 }, { "epoch": 1.2266598395630655, "grad_norm": 0.1080748587846756, "learning_rate": 2.8094039732546225e-05, "loss": 0.4417, "num_tokens": 711824581.0, "step": 899 }, { "epoch": 1.2280252602833248, "grad_norm": 0.11180043965578079, "learning_rate": 2.806908741172314e-05, "loss": 0.4095, "num_tokens": 712627527.0, "step": 900 }, { "epoch": 1.2293906810035842, "grad_norm": 0.10685121268033981, "learning_rate": 2.8044121925279068e-05, "loss": 0.4331, "num_tokens": 713373680.0, "step": 901 }, { "epoch": 1.2307561017238438, "grad_norm": 0.1211898997426033, "learning_rate": 2.801914332737143e-05, "loss": 0.4305, "num_tokens": 714163581.0, "step": 902 }, { "epoch": 1.2321215224441031, "grad_norm": 0.10798826813697815, "learning_rate": 2.799415167218608e-05, "loss": 0.437, "num_tokens": 714902320.0, "step": 903 }, { "epoch": 1.2334869431643625, "grad_norm": 0.10918522626161575, "learning_rate": 2.7969147013937227e-05, "loss": 0.4539, "num_tokens": 715689161.0, "step": 904 }, { "epoch": 1.234852363884622, "grad_norm": 0.12351608276367188, "learning_rate": 2.794412940686725e-05, "loss": 0.4434, "num_tokens": 716469432.0, "step": 905 }, { "epoch": 1.2362177846048814, "grad_norm": 0.11523031443357468, "learning_rate": 2.7919098905246645e-05, "loss": 0.4475, "num_tokens": 717283736.0, "step": 906 }, { "epoch": 1.2375832053251408, "grad_norm": 0.10629519075155258, "learning_rate": 2.7894055563373883e-05, "loss": 0.4313, "num_tokens": 717997368.0, "step": 907 }, { "epoch": 1.2389486260454001, "grad_norm": 0.11949961632490158, "learning_rate": 2.7868999435575256e-05, "loss": 0.4288, "num_tokens": 718901181.0, "step": 908 }, { "epoch": 1.2403140467656597, "grad_norm": 0.10314188152551651, "learning_rate": 2.7843930576204833e-05, "loss": 0.4437, "num_tokens": 719746864.0, "step": 909 }, { "epoch": 1.241679467485919, "grad_norm": 0.10666344314813614, "learning_rate": 2.7818849039644272e-05, "loss": 0.4192, "num_tokens": 720532690.0, "step": 910 }, { "epoch": 1.2430448882061786, "grad_norm": 0.11229496449232101, "learning_rate": 2.7793754880302747e-05, "loss": 0.4302, "num_tokens": 721320795.0, "step": 911 }, { "epoch": 1.244410308926438, "grad_norm": 0.10500967502593994, "learning_rate": 2.7768648152616802e-05, "loss": 0.4269, "num_tokens": 722158114.0, "step": 912 }, { "epoch": 1.2457757296466974, "grad_norm": 0.10424060374498367, "learning_rate": 2.7743528911050267e-05, "loss": 0.4215, "num_tokens": 722988391.0, "step": 913 }, { "epoch": 1.2471411503669567, "grad_norm": 0.11117185652256012, "learning_rate": 2.7718397210094092e-05, "loss": 0.4147, "num_tokens": 723752755.0, "step": 914 }, { "epoch": 1.2485065710872163, "grad_norm": 0.10296066105365753, "learning_rate": 2.7693253104266276e-05, "loss": 0.4172, "num_tokens": 724534617.0, "step": 915 }, { "epoch": 1.2498719918074757, "grad_norm": 0.10204853862524033, "learning_rate": 2.7668096648111714e-05, "loss": 0.4038, "num_tokens": 725269594.0, "step": 916 }, { "epoch": 1.2512374125277352, "grad_norm": 0.11298200488090515, "learning_rate": 2.764292789620209e-05, "loss": 0.406, "num_tokens": 726008186.0, "step": 917 }, { "epoch": 1.2526028332479946, "grad_norm": 0.09750685095787048, "learning_rate": 2.761774690313579e-05, "loss": 0.4516, "num_tokens": 726869901.0, "step": 918 }, { "epoch": 1.253968253968254, "grad_norm": 0.12105249613523483, "learning_rate": 2.7592553723537718e-05, "loss": 0.4023, "num_tokens": 727611984.0, "step": 919 }, { "epoch": 1.2553336746885133, "grad_norm": 0.10312298685312271, "learning_rate": 2.7567348412059226e-05, "loss": 0.4094, "num_tokens": 728398784.0, "step": 920 }, { "epoch": 1.256699095408773, "grad_norm": 0.11662142723798752, "learning_rate": 2.7542131023378004e-05, "loss": 0.4166, "num_tokens": 729218165.0, "step": 921 }, { "epoch": 1.2580645161290323, "grad_norm": 0.11374817788600922, "learning_rate": 2.751690161219791e-05, "loss": 0.4344, "num_tokens": 730044037.0, "step": 922 }, { "epoch": 1.2594299368492916, "grad_norm": 0.10284482687711716, "learning_rate": 2.74916602332489e-05, "loss": 0.4369, "num_tokens": 730836404.0, "step": 923 }, { "epoch": 1.2607953575695512, "grad_norm": 0.1080724224448204, "learning_rate": 2.7466406941286886e-05, "loss": 0.4402, "num_tokens": 731671400.0, "step": 924 }, { "epoch": 1.2621607782898105, "grad_norm": 0.11090350896120071, "learning_rate": 2.7441141791093627e-05, "loss": 0.4469, "num_tokens": 732455095.0, "step": 925 }, { "epoch": 1.26352619901007, "grad_norm": 0.11164332181215286, "learning_rate": 2.7415864837476614e-05, "loss": 0.4141, "num_tokens": 733130498.0, "step": 926 }, { "epoch": 1.2648916197303295, "grad_norm": 0.11161181330680847, "learning_rate": 2.7390576135268915e-05, "loss": 0.4381, "num_tokens": 733961193.0, "step": 927 }, { "epoch": 1.2662570404505888, "grad_norm": 0.10609298199415207, "learning_rate": 2.7365275739329112e-05, "loss": 0.4133, "num_tokens": 734717776.0, "step": 928 }, { "epoch": 1.2676224611708482, "grad_norm": 0.10293814539909363, "learning_rate": 2.7339963704541143e-05, "loss": 0.4414, "num_tokens": 735571959.0, "step": 929 }, { "epoch": 1.2689878818911078, "grad_norm": 0.10730300843715668, "learning_rate": 2.7314640085814196e-05, "loss": 0.429, "num_tokens": 736403603.0, "step": 930 }, { "epoch": 1.2703533026113671, "grad_norm": 0.10636942088603973, "learning_rate": 2.7289304938082583e-05, "loss": 0.4168, "num_tokens": 737150162.0, "step": 931 }, { "epoch": 1.2717187233316265, "grad_norm": 0.1057177484035492, "learning_rate": 2.7263958316305627e-05, "loss": 0.4412, "num_tokens": 737898062.0, "step": 932 }, { "epoch": 1.2730841440518859, "grad_norm": 0.10473381727933884, "learning_rate": 2.723860027546756e-05, "loss": 0.4029, "num_tokens": 738680801.0, "step": 933 }, { "epoch": 1.2744495647721454, "grad_norm": 0.11922014504671097, "learning_rate": 2.721323087057735e-05, "loss": 0.4247, "num_tokens": 739488521.0, "step": 934 }, { "epoch": 1.2758149854924048, "grad_norm": 0.10803649574518204, "learning_rate": 2.7187850156668648e-05, "loss": 0.4481, "num_tokens": 740340395.0, "step": 935 }, { "epoch": 1.2771804062126644, "grad_norm": 0.11122105270624161, "learning_rate": 2.716245818879963e-05, "loss": 0.4251, "num_tokens": 741163522.0, "step": 936 }, { "epoch": 1.2785458269329237, "grad_norm": 0.11244485527276993, "learning_rate": 2.713705502205287e-05, "loss": 0.4375, "num_tokens": 741924879.0, "step": 937 }, { "epoch": 1.279911247653183, "grad_norm": 0.10340699553489685, "learning_rate": 2.711164071153525e-05, "loss": 0.424, "num_tokens": 742750165.0, "step": 938 }, { "epoch": 1.2812766683734425, "grad_norm": 0.11647756397724152, "learning_rate": 2.7086215312377836e-05, "loss": 0.4286, "num_tokens": 743546062.0, "step": 939 }, { "epoch": 1.282642089093702, "grad_norm": 0.09964065253734589, "learning_rate": 2.7060778879735714e-05, "loss": 0.4214, "num_tokens": 744321077.0, "step": 940 }, { "epoch": 1.2840075098139614, "grad_norm": 0.11527177691459656, "learning_rate": 2.7035331468787945e-05, "loss": 0.4528, "num_tokens": 745081867.0, "step": 941 }, { "epoch": 1.285372930534221, "grad_norm": 0.11940614879131317, "learning_rate": 2.7009873134737375e-05, "loss": 0.4332, "num_tokens": 745967640.0, "step": 942 }, { "epoch": 1.2867383512544803, "grad_norm": 0.11036977171897888, "learning_rate": 2.698440393281056e-05, "loss": 0.4019, "num_tokens": 746740620.0, "step": 943 }, { "epoch": 1.2881037719747397, "grad_norm": 0.11523044854402542, "learning_rate": 2.6958923918257634e-05, "loss": 0.4377, "num_tokens": 747507874.0, "step": 944 }, { "epoch": 1.289469192694999, "grad_norm": 0.10984805226325989, "learning_rate": 2.693343314635217e-05, "loss": 0.4147, "num_tokens": 748287145.0, "step": 945 }, { "epoch": 1.2908346134152586, "grad_norm": 0.11636126786470413, "learning_rate": 2.6907931672391106e-05, "loss": 0.4285, "num_tokens": 749079098.0, "step": 946 }, { "epoch": 1.292200034135518, "grad_norm": 0.10377080738544464, "learning_rate": 2.688241955169456e-05, "loss": 0.4133, "num_tokens": 749815874.0, "step": 947 }, { "epoch": 1.2935654548557776, "grad_norm": 0.1056763082742691, "learning_rate": 2.6856896839605772e-05, "loss": 0.4614, "num_tokens": 750661799.0, "step": 948 }, { "epoch": 1.294930875576037, "grad_norm": 0.11057127267122269, "learning_rate": 2.6831363591490945e-05, "loss": 0.4314, "num_tokens": 751496654.0, "step": 949 }, { "epoch": 1.2962962962962963, "grad_norm": 0.10512425750494003, "learning_rate": 2.6805819862739146e-05, "loss": 0.4506, "num_tokens": 752329975.0, "step": 950 }, { "epoch": 1.2976617170165556, "grad_norm": 0.1037357822060585, "learning_rate": 2.6780265708762177e-05, "loss": 0.437, "num_tokens": 753126327.0, "step": 951 }, { "epoch": 1.2990271377368152, "grad_norm": 0.10946936160326004, "learning_rate": 2.6754701184994445e-05, "loss": 0.4525, "num_tokens": 753979352.0, "step": 952 }, { "epoch": 1.3003925584570746, "grad_norm": 0.12194236367940903, "learning_rate": 2.6729126346892863e-05, "loss": 0.4496, "num_tokens": 754737562.0, "step": 953 }, { "epoch": 1.301757979177334, "grad_norm": 0.09943599998950958, "learning_rate": 2.670354124993672e-05, "loss": 0.4329, "num_tokens": 755479271.0, "step": 954 }, { "epoch": 1.3031233998975935, "grad_norm": 0.11142334342002869, "learning_rate": 2.6677945949627547e-05, "loss": 0.4424, "num_tokens": 756245955.0, "step": 955 }, { "epoch": 1.3044888206178529, "grad_norm": 0.11167591065168381, "learning_rate": 2.6652340501489026e-05, "loss": 0.42, "num_tokens": 757100865.0, "step": 956 }, { "epoch": 1.3058542413381122, "grad_norm": 0.10453712195158005, "learning_rate": 2.662672496106683e-05, "loss": 0.43, "num_tokens": 757863941.0, "step": 957 }, { "epoch": 1.3072196620583718, "grad_norm": 0.1268044412136078, "learning_rate": 2.6601099383928545e-05, "loss": 0.4293, "num_tokens": 758655530.0, "step": 958 }, { "epoch": 1.3085850827786312, "grad_norm": 0.09853125363588333, "learning_rate": 2.657546382566353e-05, "loss": 0.4495, "num_tokens": 759470723.0, "step": 959 }, { "epoch": 1.3099505034988905, "grad_norm": 0.12078141421079636, "learning_rate": 2.6549818341882776e-05, "loss": 0.4039, "num_tokens": 760164706.0, "step": 960 }, { "epoch": 1.31131592421915, "grad_norm": 0.10448627173900604, "learning_rate": 2.652416298821883e-05, "loss": 0.4427, "num_tokens": 760993351.0, "step": 961 }, { "epoch": 1.3126813449394095, "grad_norm": 0.10524991899728775, "learning_rate": 2.6498497820325625e-05, "loss": 0.4513, "num_tokens": 761830986.0, "step": 962 }, { "epoch": 1.3140467656596688, "grad_norm": 0.09629581868648529, "learning_rate": 2.6472822893878415e-05, "loss": 0.4424, "num_tokens": 762657356.0, "step": 963 }, { "epoch": 1.3154121863799282, "grad_norm": 0.11350353062152863, "learning_rate": 2.644713826457359e-05, "loss": 0.4196, "num_tokens": 763394392.0, "step": 964 }, { "epoch": 1.3167776071001878, "grad_norm": 0.11060746759176254, "learning_rate": 2.6421443988128607e-05, "loss": 0.4464, "num_tokens": 764218209.0, "step": 965 }, { "epoch": 1.3181430278204471, "grad_norm": 0.11917940527200699, "learning_rate": 2.639574012028185e-05, "loss": 0.4409, "num_tokens": 764963041.0, "step": 966 }, { "epoch": 1.3195084485407067, "grad_norm": 0.09922454506158829, "learning_rate": 2.6370026716792505e-05, "loss": 0.4328, "num_tokens": 765821849.0, "step": 967 }, { "epoch": 1.320873869260966, "grad_norm": 0.11456112563610077, "learning_rate": 2.6344303833440443e-05, "loss": 0.4386, "num_tokens": 766639170.0, "step": 968 }, { "epoch": 1.3222392899812254, "grad_norm": 0.11576081067323685, "learning_rate": 2.63185715260261e-05, "loss": 0.4574, "num_tokens": 767470588.0, "step": 969 }, { "epoch": 1.3236047107014848, "grad_norm": 0.10726109147071838, "learning_rate": 2.6292829850370355e-05, "loss": 0.4117, "num_tokens": 768196145.0, "step": 970 }, { "epoch": 1.3249701314217444, "grad_norm": 0.11310656368732452, "learning_rate": 2.626707886231443e-05, "loss": 0.4498, "num_tokens": 769017646.0, "step": 971 }, { "epoch": 1.3263355521420037, "grad_norm": 0.09882940351963043, "learning_rate": 2.6241318617719703e-05, "loss": 0.3946, "num_tokens": 769768116.0, "step": 972 }, { "epoch": 1.3277009728622633, "grad_norm": 0.101749487221241, "learning_rate": 2.621554917246768e-05, "loss": 0.4217, "num_tokens": 770621828.0, "step": 973 }, { "epoch": 1.3290663935825227, "grad_norm": 0.10765894502401352, "learning_rate": 2.618977058245979e-05, "loss": 0.4337, "num_tokens": 771403348.0, "step": 974 }, { "epoch": 1.330431814302782, "grad_norm": 0.10739171504974365, "learning_rate": 2.616398290361733e-05, "loss": 0.4473, "num_tokens": 772176413.0, "step": 975 }, { "epoch": 1.3317972350230414, "grad_norm": 0.10923103243112564, "learning_rate": 2.613818619188129e-05, "loss": 0.4273, "num_tokens": 772963982.0, "step": 976 }, { "epoch": 1.333162655743301, "grad_norm": 0.10686330497264862, "learning_rate": 2.6112380503212255e-05, "loss": 0.4458, "num_tokens": 773820909.0, "step": 977 }, { "epoch": 1.3345280764635603, "grad_norm": 0.10851866751909256, "learning_rate": 2.6086565893590302e-05, "loss": 0.411, "num_tokens": 774613420.0, "step": 978 }, { "epoch": 1.33589349718382, "grad_norm": 0.10899483412504196, "learning_rate": 2.606074241901484e-05, "loss": 0.436, "num_tokens": 775458462.0, "step": 979 }, { "epoch": 1.3372589179040792, "grad_norm": 0.10930947214365005, "learning_rate": 2.6034910135504534e-05, "loss": 0.4137, "num_tokens": 776185415.0, "step": 980 }, { "epoch": 1.3386243386243386, "grad_norm": 0.09565704315900803, "learning_rate": 2.6009069099097125e-05, "loss": 0.4659, "num_tokens": 777041110.0, "step": 981 }, { "epoch": 1.339989759344598, "grad_norm": 0.11016254127025604, "learning_rate": 2.5983219365849367e-05, "loss": 0.4261, "num_tokens": 777862793.0, "step": 982 }, { "epoch": 1.3413551800648575, "grad_norm": 0.10499899834394455, "learning_rate": 2.595736099183688e-05, "loss": 0.4147, "num_tokens": 778615797.0, "step": 983 }, { "epoch": 1.342720600785117, "grad_norm": 0.10532669723033905, "learning_rate": 2.5931494033154002e-05, "loss": 0.4288, "num_tokens": 779370721.0, "step": 984 }, { "epoch": 1.3440860215053765, "grad_norm": 0.10797038674354553, "learning_rate": 2.5905618545913728e-05, "loss": 0.4209, "num_tokens": 780153197.0, "step": 985 }, { "epoch": 1.3454514422256358, "grad_norm": 0.10069423168897629, "learning_rate": 2.5879734586247534e-05, "loss": 0.4222, "num_tokens": 780904279.0, "step": 986 }, { "epoch": 1.3468168629458952, "grad_norm": 0.11785499006509781, "learning_rate": 2.5853842210305268e-05, "loss": 0.4472, "num_tokens": 781684419.0, "step": 987 }, { "epoch": 1.3481822836661546, "grad_norm": 0.10913155972957611, "learning_rate": 2.582794147425506e-05, "loss": 0.4497, "num_tokens": 782601802.0, "step": 988 }, { "epoch": 1.3495477043864141, "grad_norm": 0.09946605563163757, "learning_rate": 2.5802032434283156e-05, "loss": 0.3978, "num_tokens": 783290883.0, "step": 989 }, { "epoch": 1.3509131251066735, "grad_norm": 0.10821465402841568, "learning_rate": 2.5776115146593827e-05, "loss": 0.437, "num_tokens": 784140587.0, "step": 990 }, { "epoch": 1.3522785458269329, "grad_norm": 0.10317730158567429, "learning_rate": 2.5750189667409222e-05, "loss": 0.4241, "num_tokens": 784926011.0, "step": 991 }, { "epoch": 1.3536439665471924, "grad_norm": 0.11781519651412964, "learning_rate": 2.572425605296928e-05, "loss": 0.4139, "num_tokens": 785678668.0, "step": 992 }, { "epoch": 1.3550093872674518, "grad_norm": 0.107591912150383, "learning_rate": 2.5698314359531557e-05, "loss": 0.4257, "num_tokens": 786441586.0, "step": 993 }, { "epoch": 1.3563748079877112, "grad_norm": 0.11734827607870102, "learning_rate": 2.567236464337117e-05, "loss": 0.4216, "num_tokens": 787165696.0, "step": 994 }, { "epoch": 1.3577402287079705, "grad_norm": 0.13131317496299744, "learning_rate": 2.5646406960780614e-05, "loss": 0.429, "num_tokens": 788005743.0, "step": 995 }, { "epoch": 1.35910564942823, "grad_norm": 0.11139479279518127, "learning_rate": 2.5620441368069675e-05, "loss": 0.4138, "num_tokens": 788767131.0, "step": 996 }, { "epoch": 1.3604710701484894, "grad_norm": 0.12356792390346527, "learning_rate": 2.5594467921565308e-05, "loss": 0.4301, "num_tokens": 789535844.0, "step": 997 }, { "epoch": 1.361836490868749, "grad_norm": 0.11917587369680405, "learning_rate": 2.5568486677611487e-05, "loss": 0.4362, "num_tokens": 790268727.0, "step": 998 }, { "epoch": 1.3632019115890084, "grad_norm": 0.1146593987941742, "learning_rate": 2.5542497692569103e-05, "loss": 0.4247, "num_tokens": 791069579.0, "step": 999 }, { "epoch": 1.3645673323092677, "grad_norm": 0.11165698617696762, "learning_rate": 2.5516501022815857e-05, "loss": 0.4457, "num_tokens": 791937177.0, "step": 1000 }, { "epoch": 1.365932753029527, "grad_norm": 0.11099539697170258, "learning_rate": 2.5490496724746104e-05, "loss": 0.4433, "num_tokens": 792738424.0, "step": 1001 }, { "epoch": 1.3672981737497867, "grad_norm": 0.1142379492521286, "learning_rate": 2.5464484854770755e-05, "loss": 0.4064, "num_tokens": 793499030.0, "step": 1002 }, { "epoch": 1.368663594470046, "grad_norm": 0.10760563611984253, "learning_rate": 2.5438465469317142e-05, "loss": 0.4367, "num_tokens": 794307172.0, "step": 1003 }, { "epoch": 1.3700290151903056, "grad_norm": 0.11333650350570679, "learning_rate": 2.54124386248289e-05, "loss": 0.4401, "num_tokens": 795173341.0, "step": 1004 }, { "epoch": 1.371394435910565, "grad_norm": 0.0997990146279335, "learning_rate": 2.538640437776585e-05, "loss": 0.4302, "num_tokens": 795916087.0, "step": 1005 }, { "epoch": 1.3727598566308243, "grad_norm": 0.10008875280618668, "learning_rate": 2.5360362784603866e-05, "loss": 0.4316, "num_tokens": 796716961.0, "step": 1006 }, { "epoch": 1.3741252773510837, "grad_norm": 0.11423381417989731, "learning_rate": 2.533431390183476e-05, "loss": 0.4364, "num_tokens": 797494957.0, "step": 1007 }, { "epoch": 1.3754906980713433, "grad_norm": 0.10408934205770493, "learning_rate": 2.530825778596616e-05, "loss": 0.4326, "num_tokens": 798310851.0, "step": 1008 }, { "epoch": 1.3768561187916026, "grad_norm": 0.10851862281560898, "learning_rate": 2.528219449352138e-05, "loss": 0.4256, "num_tokens": 799054528.0, "step": 1009 }, { "epoch": 1.3782215395118622, "grad_norm": 0.0996582880616188, "learning_rate": 2.5256124081039307e-05, "loss": 0.4262, "num_tokens": 799887367.0, "step": 1010 }, { "epoch": 1.3795869602321216, "grad_norm": 0.12108288705348969, "learning_rate": 2.5230046605074262e-05, "loss": 0.4427, "num_tokens": 800762199.0, "step": 1011 }, { "epoch": 1.380952380952381, "grad_norm": 0.11542606353759766, "learning_rate": 2.5203962122195915e-05, "loss": 0.456, "num_tokens": 801571455.0, "step": 1012 }, { "epoch": 1.3823178016726403, "grad_norm": 0.11006367951631546, "learning_rate": 2.517787068898909e-05, "loss": 0.4375, "num_tokens": 802428953.0, "step": 1013 }, { "epoch": 1.3836832223928999, "grad_norm": 0.1359989196062088, "learning_rate": 2.5151772362053742e-05, "loss": 0.4294, "num_tokens": 803220231.0, "step": 1014 }, { "epoch": 1.3850486431131592, "grad_norm": 0.10926545411348343, "learning_rate": 2.512566719800475e-05, "loss": 0.427, "num_tokens": 804011086.0, "step": 1015 }, { "epoch": 1.3864140638334188, "grad_norm": 0.10893993824720383, "learning_rate": 2.5099555253471824e-05, "loss": 0.4261, "num_tokens": 804763110.0, "step": 1016 }, { "epoch": 1.3877794845536782, "grad_norm": 0.10825891047716141, "learning_rate": 2.5073436585099384e-05, "loss": 0.4315, "num_tokens": 805564626.0, "step": 1017 }, { "epoch": 1.3891449052739375, "grad_norm": 0.10739433765411377, "learning_rate": 2.5047311249546464e-05, "loss": 0.4117, "num_tokens": 806330230.0, "step": 1018 }, { "epoch": 1.3905103259941969, "grad_norm": 0.11548051238059998, "learning_rate": 2.502117930348651e-05, "loss": 0.4275, "num_tokens": 807151952.0, "step": 1019 }, { "epoch": 1.3918757467144565, "grad_norm": 0.10339847207069397, "learning_rate": 2.499504080360735e-05, "loss": 0.4292, "num_tokens": 807990493.0, "step": 1020 }, { "epoch": 1.3932411674347158, "grad_norm": 0.1048726812005043, "learning_rate": 2.4968895806611014e-05, "loss": 0.4286, "num_tokens": 808822125.0, "step": 1021 }, { "epoch": 1.3946065881549752, "grad_norm": 0.10640150308609009, "learning_rate": 2.4942744369213626e-05, "loss": 0.4226, "num_tokens": 809561550.0, "step": 1022 }, { "epoch": 1.3959720088752348, "grad_norm": 0.1056184396147728, "learning_rate": 2.4916586548145286e-05, "loss": 0.422, "num_tokens": 810378731.0, "step": 1023 }, { "epoch": 1.3973374295954941, "grad_norm": 0.11018586903810501, "learning_rate": 2.489042240014993e-05, "loss": 0.4364, "num_tokens": 811231037.0, "step": 1024 }, { "epoch": 1.3987028503157535, "grad_norm": 0.10734830051660538, "learning_rate": 2.4864251981985233e-05, "loss": 0.4316, "num_tokens": 812050064.0, "step": 1025 }, { "epoch": 1.4000682710360128, "grad_norm": 0.11405730247497559, "learning_rate": 2.483807535042246e-05, "loss": 0.4668, "num_tokens": 812916074.0, "step": 1026 }, { "epoch": 1.4014336917562724, "grad_norm": 0.10311237722635269, "learning_rate": 2.4811892562246373e-05, "loss": 0.42, "num_tokens": 813764303.0, "step": 1027 }, { "epoch": 1.4027991124765318, "grad_norm": 0.11048302054405212, "learning_rate": 2.4785703674255062e-05, "loss": 0.421, "num_tokens": 814490556.0, "step": 1028 }, { "epoch": 1.4041645331967914, "grad_norm": 0.13120746612548828, "learning_rate": 2.475950874325988e-05, "loss": 0.4361, "num_tokens": 815366820.0, "step": 1029 }, { "epoch": 1.4055299539170507, "grad_norm": 0.09605155140161514, "learning_rate": 2.4733307826085265e-05, "loss": 0.4157, "num_tokens": 816195802.0, "step": 1030 }, { "epoch": 1.40689537463731, "grad_norm": 0.11790984869003296, "learning_rate": 2.4707100979568646e-05, "loss": 0.4371, "num_tokens": 816983376.0, "step": 1031 }, { "epoch": 1.4082607953575694, "grad_norm": 0.11094418913125992, "learning_rate": 2.468088826056032e-05, "loss": 0.4417, "num_tokens": 817813401.0, "step": 1032 }, { "epoch": 1.409626216077829, "grad_norm": 0.10623343288898468, "learning_rate": 2.465466972592333e-05, "loss": 0.4193, "num_tokens": 818645727.0, "step": 1033 }, { "epoch": 1.4109916367980884, "grad_norm": 0.11408126354217529, "learning_rate": 2.4628445432533315e-05, "loss": 0.4167, "num_tokens": 819384258.0, "step": 1034 }, { "epoch": 1.412357057518348, "grad_norm": 0.12455479800701141, "learning_rate": 2.4602215437278422e-05, "loss": 0.4487, "num_tokens": 820154980.0, "step": 1035 }, { "epoch": 1.4137224782386073, "grad_norm": 0.11390459537506104, "learning_rate": 2.4575979797059165e-05, "loss": 0.4303, "num_tokens": 820885309.0, "step": 1036 }, { "epoch": 1.4150878989588667, "grad_norm": 0.10783706605434418, "learning_rate": 2.4549738568788293e-05, "loss": 0.4167, "num_tokens": 821614200.0, "step": 1037 }, { "epoch": 1.416453319679126, "grad_norm": 0.1114407554268837, "learning_rate": 2.4523491809390686e-05, "loss": 0.4327, "num_tokens": 822436539.0, "step": 1038 }, { "epoch": 1.4178187403993856, "grad_norm": 0.11341927200555801, "learning_rate": 2.4497239575803233e-05, "loss": 0.4344, "num_tokens": 823179229.0, "step": 1039 }, { "epoch": 1.419184161119645, "grad_norm": 0.11667715758085251, "learning_rate": 2.447098192497467e-05, "loss": 0.4253, "num_tokens": 823894968.0, "step": 1040 }, { "epoch": 1.4205495818399045, "grad_norm": 0.11697737872600555, "learning_rate": 2.4444718913865506e-05, "loss": 0.4494, "num_tokens": 824610375.0, "step": 1041 }, { "epoch": 1.421915002560164, "grad_norm": 0.09855484217405319, "learning_rate": 2.441845059944789e-05, "loss": 0.4272, "num_tokens": 825447947.0, "step": 1042 }, { "epoch": 1.4232804232804233, "grad_norm": 0.12015564739704132, "learning_rate": 2.4392177038705432e-05, "loss": 0.4362, "num_tokens": 826245089.0, "step": 1043 }, { "epoch": 1.4246458440006826, "grad_norm": 0.10571084916591644, "learning_rate": 2.436589828863317e-05, "loss": 0.423, "num_tokens": 827058741.0, "step": 1044 }, { "epoch": 1.4260112647209422, "grad_norm": 0.11008228361606598, "learning_rate": 2.433961440623737e-05, "loss": 0.4229, "num_tokens": 827873766.0, "step": 1045 }, { "epoch": 1.4273766854412016, "grad_norm": 0.11864599585533142, "learning_rate": 2.431332544853544e-05, "loss": 0.4091, "num_tokens": 828704087.0, "step": 1046 }, { "epoch": 1.4287421061614611, "grad_norm": 0.10300114750862122, "learning_rate": 2.4287031472555796e-05, "loss": 0.4312, "num_tokens": 829533650.0, "step": 1047 }, { "epoch": 1.4301075268817205, "grad_norm": 0.09964679926633835, "learning_rate": 2.426073253533775e-05, "loss": 0.4261, "num_tokens": 830354588.0, "step": 1048 }, { "epoch": 1.4314729476019799, "grad_norm": 0.12031849473714828, "learning_rate": 2.4234428693931355e-05, "loss": 0.4422, "num_tokens": 831157835.0, "step": 1049 }, { "epoch": 1.4328383683222392, "grad_norm": 0.10048427432775497, "learning_rate": 2.420812000539734e-05, "loss": 0.4026, "num_tokens": 831929780.0, "step": 1050 }, { "epoch": 1.4342037890424988, "grad_norm": 0.11776667833328247, "learning_rate": 2.4181806526806893e-05, "loss": 0.4291, "num_tokens": 832797366.0, "step": 1051 }, { "epoch": 1.4355692097627581, "grad_norm": 0.1011839210987091, "learning_rate": 2.415548831524165e-05, "loss": 0.4212, "num_tokens": 833716848.0, "step": 1052 }, { "epoch": 1.4369346304830175, "grad_norm": 0.10930094122886658, "learning_rate": 2.4129165427793475e-05, "loss": 0.4361, "num_tokens": 834524447.0, "step": 1053 }, { "epoch": 1.438300051203277, "grad_norm": 0.10795430839061737, "learning_rate": 2.41028379215644e-05, "loss": 0.4552, "num_tokens": 835303579.0, "step": 1054 }, { "epoch": 1.4396654719235364, "grad_norm": 0.09194496273994446, "learning_rate": 2.4076505853666458e-05, "loss": 0.4148, "num_tokens": 836053825.0, "step": 1055 }, { "epoch": 1.4410308926437958, "grad_norm": 0.11195285618305206, "learning_rate": 2.405016928122159e-05, "loss": 0.4195, "num_tokens": 836795144.0, "step": 1056 }, { "epoch": 1.4423963133640554, "grad_norm": 0.10386230796575546, "learning_rate": 2.40238282613615e-05, "loss": 0.4458, "num_tokens": 837573897.0, "step": 1057 }, { "epoch": 1.4437617340843147, "grad_norm": 0.11411534249782562, "learning_rate": 2.399748285122755e-05, "loss": 0.4411, "num_tokens": 838425493.0, "step": 1058 }, { "epoch": 1.445127154804574, "grad_norm": 0.09108033776283264, "learning_rate": 2.3971133107970617e-05, "loss": 0.4022, "num_tokens": 839236719.0, "step": 1059 }, { "epoch": 1.4464925755248337, "grad_norm": 0.10645338147878647, "learning_rate": 2.3944779088750975e-05, "loss": 0.4244, "num_tokens": 840025819.0, "step": 1060 }, { "epoch": 1.447857996245093, "grad_norm": 0.10921335965394974, "learning_rate": 2.391842085073818e-05, "loss": 0.4399, "num_tokens": 840804892.0, "step": 1061 }, { "epoch": 1.4492234169653524, "grad_norm": 0.101954884827137, "learning_rate": 2.389205845111095e-05, "loss": 0.4192, "num_tokens": 841637742.0, "step": 1062 }, { "epoch": 1.4505888376856118, "grad_norm": 0.11108341813087463, "learning_rate": 2.386569194705701e-05, "loss": 0.4155, "num_tokens": 842420275.0, "step": 1063 }, { "epoch": 1.4519542584058713, "grad_norm": 0.10921622067689896, "learning_rate": 2.383932139577301e-05, "loss": 0.4264, "num_tokens": 843195985.0, "step": 1064 }, { "epoch": 1.4533196791261307, "grad_norm": 0.10597191005945206, "learning_rate": 2.381294685446435e-05, "loss": 0.4286, "num_tokens": 843971035.0, "step": 1065 }, { "epoch": 1.4546850998463903, "grad_norm": 0.0980968102812767, "learning_rate": 2.378656838034511e-05, "loss": 0.4269, "num_tokens": 844761965.0, "step": 1066 }, { "epoch": 1.4560505205666496, "grad_norm": 0.09648693352937698, "learning_rate": 2.3760186030637905e-05, "loss": 0.4163, "num_tokens": 845538267.0, "step": 1067 }, { "epoch": 1.457415941286909, "grad_norm": 0.11654429137706757, "learning_rate": 2.3733799862573746e-05, "loss": 0.448, "num_tokens": 846387110.0, "step": 1068 }, { "epoch": 1.4587813620071683, "grad_norm": 0.095599465072155, "learning_rate": 2.3707409933391918e-05, "loss": 0.426, "num_tokens": 847148130.0, "step": 1069 }, { "epoch": 1.460146782727428, "grad_norm": 0.10824096947908401, "learning_rate": 2.3681016300339875e-05, "loss": 0.4226, "num_tokens": 847919331.0, "step": 1070 }, { "epoch": 1.4615122034476873, "grad_norm": 0.10332082957029343, "learning_rate": 2.3654619020673124e-05, "loss": 0.4343, "num_tokens": 848697077.0, "step": 1071 }, { "epoch": 1.4628776241679469, "grad_norm": 0.1033865362405777, "learning_rate": 2.3628218151655043e-05, "loss": 0.398, "num_tokens": 849444665.0, "step": 1072 }, { "epoch": 1.4642430448882062, "grad_norm": 0.10221332311630249, "learning_rate": 2.3601813750556834e-05, "loss": 0.4035, "num_tokens": 850227439.0, "step": 1073 }, { "epoch": 1.4656084656084656, "grad_norm": 0.10571034997701645, "learning_rate": 2.3575405874657343e-05, "loss": 0.4255, "num_tokens": 851079480.0, "step": 1074 }, { "epoch": 1.466973886328725, "grad_norm": 0.10364726185798645, "learning_rate": 2.3548994581242948e-05, "loss": 0.4638, "num_tokens": 851931761.0, "step": 1075 }, { "epoch": 1.4683393070489845, "grad_norm": 0.10997550934553146, "learning_rate": 2.352257992760745e-05, "loss": 0.4166, "num_tokens": 852702836.0, "step": 1076 }, { "epoch": 1.4697047277692439, "grad_norm": 0.10522348433732986, "learning_rate": 2.349616197105195e-05, "loss": 0.4516, "num_tokens": 853502013.0, "step": 1077 }, { "epoch": 1.4710701484895035, "grad_norm": 0.10616324841976166, "learning_rate": 2.346974076888469e-05, "loss": 0.4009, "num_tokens": 854227415.0, "step": 1078 }, { "epoch": 1.4724355692097628, "grad_norm": 0.11846499145030975, "learning_rate": 2.3443316378420975e-05, "loss": 0.4232, "num_tokens": 854974112.0, "step": 1079 }, { "epoch": 1.4738009899300222, "grad_norm": 0.1030370369553566, "learning_rate": 2.3416888856983003e-05, "loss": 0.4264, "num_tokens": 855738735.0, "step": 1080 }, { "epoch": 1.4751664106502815, "grad_norm": 0.10405363887548447, "learning_rate": 2.3390458261899783e-05, "loss": 0.4339, "num_tokens": 856538382.0, "step": 1081 }, { "epoch": 1.4765318313705411, "grad_norm": 0.11005817353725433, "learning_rate": 2.3364024650506986e-05, "loss": 0.4438, "num_tokens": 857328763.0, "step": 1082 }, { "epoch": 1.4778972520908005, "grad_norm": 0.1163555234670639, "learning_rate": 2.333758808014684e-05, "loss": 0.4075, "num_tokens": 858085107.0, "step": 1083 }, { "epoch": 1.4792626728110598, "grad_norm": 0.11217539012432098, "learning_rate": 2.3311148608167944e-05, "loss": 0.4251, "num_tokens": 858880828.0, "step": 1084 }, { "epoch": 1.4806280935313194, "grad_norm": 0.11735870689153671, "learning_rate": 2.3284706291925257e-05, "loss": 0.4298, "num_tokens": 859693219.0, "step": 1085 }, { "epoch": 1.4819935142515788, "grad_norm": 0.10702086240053177, "learning_rate": 2.3258261188779863e-05, "loss": 0.4556, "num_tokens": 860535848.0, "step": 1086 }, { "epoch": 1.4833589349718381, "grad_norm": 0.11275660991668701, "learning_rate": 2.323181335609891e-05, "loss": 0.4491, "num_tokens": 861328797.0, "step": 1087 }, { "epoch": 1.4847243556920977, "grad_norm": 0.09833118319511414, "learning_rate": 2.3205362851255453e-05, "loss": 0.4083, "num_tokens": 862032708.0, "step": 1088 }, { "epoch": 1.486089776412357, "grad_norm": 0.11304361373186111, "learning_rate": 2.3178909731628365e-05, "loss": 0.3934, "num_tokens": 862726194.0, "step": 1089 }, { "epoch": 1.4874551971326164, "grad_norm": 0.1036527082324028, "learning_rate": 2.3152454054602166e-05, "loss": 0.4366, "num_tokens": 863565182.0, "step": 1090 }, { "epoch": 1.488820617852876, "grad_norm": 0.10629989206790924, "learning_rate": 2.312599587756695e-05, "loss": 0.4356, "num_tokens": 864360781.0, "step": 1091 }, { "epoch": 1.4901860385731354, "grad_norm": 0.1064317524433136, "learning_rate": 2.3099535257918212e-05, "loss": 0.4298, "num_tokens": 865189088.0, "step": 1092 }, { "epoch": 1.4915514592933947, "grad_norm": 0.10382280498743057, "learning_rate": 2.307307225305676e-05, "loss": 0.4249, "num_tokens": 865982262.0, "step": 1093 }, { "epoch": 1.492916880013654, "grad_norm": 0.10707318782806396, "learning_rate": 2.304660692038857e-05, "loss": 0.4435, "num_tokens": 866800240.0, "step": 1094 }, { "epoch": 1.4942823007339137, "grad_norm": 0.10118279606103897, "learning_rate": 2.302013931732466e-05, "loss": 0.4634, "num_tokens": 867602247.0, "step": 1095 }, { "epoch": 1.495647721454173, "grad_norm": 0.1076766774058342, "learning_rate": 2.2993669501280995e-05, "loss": 0.433, "num_tokens": 868454523.0, "step": 1096 }, { "epoch": 1.4970131421744326, "grad_norm": 0.10012049227952957, "learning_rate": 2.2967197529678323e-05, "loss": 0.4449, "num_tokens": 869281295.0, "step": 1097 }, { "epoch": 1.498378562894692, "grad_norm": 0.10443660616874695, "learning_rate": 2.2940723459942067e-05, "loss": 0.432, "num_tokens": 870123860.0, "step": 1098 }, { "epoch": 1.4997439836149513, "grad_norm": 0.09993372112512589, "learning_rate": 2.2914247349502206e-05, "loss": 0.4299, "num_tokens": 870914316.0, "step": 1099 }, { "epoch": 1.5011094043352107, "grad_norm": 0.10264316946268082, "learning_rate": 2.2887769255793165e-05, "loss": 0.442, "num_tokens": 871795860.0, "step": 1100 }, { "epoch": 1.5024748250554703, "grad_norm": 0.10182645916938782, "learning_rate": 2.2861289236253627e-05, "loss": 0.438, "num_tokens": 872593862.0, "step": 1101 }, { "epoch": 1.5038402457757296, "grad_norm": 0.10481473058462143, "learning_rate": 2.2834807348326497e-05, "loss": 0.4141, "num_tokens": 873328198.0, "step": 1102 }, { "epoch": 1.5052056664959892, "grad_norm": 0.11086317151784897, "learning_rate": 2.2808323649458703e-05, "loss": 0.4282, "num_tokens": 874148499.0, "step": 1103 }, { "epoch": 1.5065710872162486, "grad_norm": 0.09038373082876205, "learning_rate": 2.278183819710113e-05, "loss": 0.4176, "num_tokens": 874923135.0, "step": 1104 }, { "epoch": 1.507936507936508, "grad_norm": 0.10969164967536926, "learning_rate": 2.2755351048708428e-05, "loss": 0.4209, "num_tokens": 875690940.0, "step": 1105 }, { "epoch": 1.5093019286567673, "grad_norm": 0.10826202481985092, "learning_rate": 2.2728862261738972e-05, "loss": 0.4304, "num_tokens": 876534127.0, "step": 1106 }, { "epoch": 1.5106673493770268, "grad_norm": 0.10513294488191605, "learning_rate": 2.2702371893654646e-05, "loss": 0.4456, "num_tokens": 877329693.0, "step": 1107 }, { "epoch": 1.5120327700972862, "grad_norm": 0.09945349395275116, "learning_rate": 2.2675880001920806e-05, "loss": 0.4138, "num_tokens": 878154771.0, "step": 1108 }, { "epoch": 1.5133981908175458, "grad_norm": 0.10369694232940674, "learning_rate": 2.264938664400608e-05, "loss": 0.4374, "num_tokens": 878890917.0, "step": 1109 }, { "epoch": 1.5147636115378051, "grad_norm": 0.11033739149570465, "learning_rate": 2.2622891877382285e-05, "loss": 0.4341, "num_tokens": 879621489.0, "step": 1110 }, { "epoch": 1.5161290322580645, "grad_norm": 0.1120574027299881, "learning_rate": 2.2596395759524316e-05, "loss": 0.4158, "num_tokens": 880343260.0, "step": 1111 }, { "epoch": 1.5174944529783239, "grad_norm": 0.09995916485786438, "learning_rate": 2.2569898347909967e-05, "loss": 0.433, "num_tokens": 881186694.0, "step": 1112 }, { "epoch": 1.5188598736985832, "grad_norm": 0.10880868136882782, "learning_rate": 2.2543399700019858e-05, "loss": 0.4312, "num_tokens": 881982089.0, "step": 1113 }, { "epoch": 1.5202252944188428, "grad_norm": 0.11715039610862732, "learning_rate": 2.2516899873337286e-05, "loss": 0.4242, "num_tokens": 882802813.0, "step": 1114 }, { "epoch": 1.5215907151391024, "grad_norm": 0.10645370185375214, "learning_rate": 2.2490398925348108e-05, "loss": 0.4411, "num_tokens": 883575179.0, "step": 1115 }, { "epoch": 1.5229561358593617, "grad_norm": 0.11245811730623245, "learning_rate": 2.2463896913540605e-05, "loss": 0.4321, "num_tokens": 884389083.0, "step": 1116 }, { "epoch": 1.524321556579621, "grad_norm": 0.10634024441242218, "learning_rate": 2.2437393895405377e-05, "loss": 0.4397, "num_tokens": 885191883.0, "step": 1117 }, { "epoch": 1.5256869772998805, "grad_norm": 0.12098433822393417, "learning_rate": 2.24108899284352e-05, "loss": 0.4299, "num_tokens": 885934055.0, "step": 1118 }, { "epoch": 1.5270523980201398, "grad_norm": 0.10963091254234314, "learning_rate": 2.2384385070124907e-05, "loss": 0.4349, "num_tokens": 886736653.0, "step": 1119 }, { "epoch": 1.5284178187403994, "grad_norm": 0.10462947934865952, "learning_rate": 2.2357879377971265e-05, "loss": 0.4029, "num_tokens": 887460928.0, "step": 1120 }, { "epoch": 1.529783239460659, "grad_norm": 0.11493690311908722, "learning_rate": 2.2331372909472863e-05, "loss": 0.4359, "num_tokens": 888271581.0, "step": 1121 }, { "epoch": 1.5311486601809183, "grad_norm": 0.10729134827852249, "learning_rate": 2.2304865722129948e-05, "loss": 0.442, "num_tokens": 889134574.0, "step": 1122 }, { "epoch": 1.5325140809011777, "grad_norm": 0.11409099400043488, "learning_rate": 2.227835787344436e-05, "loss": 0.4365, "num_tokens": 889961661.0, "step": 1123 }, { "epoch": 1.533879501621437, "grad_norm": 0.11054748296737671, "learning_rate": 2.2251849420919342e-05, "loss": 0.4076, "num_tokens": 890727978.0, "step": 1124 }, { "epoch": 1.5352449223416964, "grad_norm": 0.113284632563591, "learning_rate": 2.2225340422059463e-05, "loss": 0.4542, "num_tokens": 891530109.0, "step": 1125 }, { "epoch": 1.536610343061956, "grad_norm": 0.11223120987415314, "learning_rate": 2.2198830934370474e-05, "loss": 0.412, "num_tokens": 892325031.0, "step": 1126 }, { "epoch": 1.5379757637822153, "grad_norm": 0.11619285494089127, "learning_rate": 2.2172321015359196e-05, "loss": 0.452, "num_tokens": 893220154.0, "step": 1127 }, { "epoch": 1.539341184502475, "grad_norm": 0.10065492987632751, "learning_rate": 2.214581072253336e-05, "loss": 0.4127, "num_tokens": 893942080.0, "step": 1128 }, { "epoch": 1.5407066052227343, "grad_norm": 0.10353072732686996, "learning_rate": 2.2119300113401527e-05, "loss": 0.4191, "num_tokens": 894735447.0, "step": 1129 }, { "epoch": 1.5420720259429936, "grad_norm": 0.11721158772706985, "learning_rate": 2.209278924547296e-05, "loss": 0.4419, "num_tokens": 895549785.0, "step": 1130 }, { "epoch": 1.543437446663253, "grad_norm": 0.09938321262598038, "learning_rate": 2.206627817625744e-05, "loss": 0.4147, "num_tokens": 896380901.0, "step": 1131 }, { "epoch": 1.5448028673835126, "grad_norm": 0.10967298597097397, "learning_rate": 2.2039766963265237e-05, "loss": 0.4385, "num_tokens": 897207503.0, "step": 1132 }, { "epoch": 1.546168288103772, "grad_norm": 0.10768026113510132, "learning_rate": 2.201325566400688e-05, "loss": 0.4342, "num_tokens": 897996296.0, "step": 1133 }, { "epoch": 1.5475337088240315, "grad_norm": 0.0999634712934494, "learning_rate": 2.1986744335993126e-05, "loss": 0.4239, "num_tokens": 898804208.0, "step": 1134 }, { "epoch": 1.5488991295442909, "grad_norm": 0.11623304337263107, "learning_rate": 2.1960233036734766e-05, "loss": 0.4404, "num_tokens": 899564990.0, "step": 1135 }, { "epoch": 1.5502645502645502, "grad_norm": 0.1038210317492485, "learning_rate": 2.1933721823742554e-05, "loss": 0.4614, "num_tokens": 900377004.0, "step": 1136 }, { "epoch": 1.5516299709848096, "grad_norm": 0.11308035254478455, "learning_rate": 2.190721075452705e-05, "loss": 0.4219, "num_tokens": 901116101.0, "step": 1137 }, { "epoch": 1.5529953917050692, "grad_norm": 0.10411029309034348, "learning_rate": 2.188069988659848e-05, "loss": 0.4578, "num_tokens": 901976746.0, "step": 1138 }, { "epoch": 1.5543608124253285, "grad_norm": 0.10899616032838821, "learning_rate": 2.185418927746665e-05, "loss": 0.4292, "num_tokens": 902769540.0, "step": 1139 }, { "epoch": 1.5557262331455881, "grad_norm": 0.09949219971895218, "learning_rate": 2.1827678984640816e-05, "loss": 0.4417, "num_tokens": 903610115.0, "step": 1140 }, { "epoch": 1.5570916538658475, "grad_norm": 0.10675116628408432, "learning_rate": 2.180116906562953e-05, "loss": 0.4171, "num_tokens": 904481626.0, "step": 1141 }, { "epoch": 1.5584570745861068, "grad_norm": 0.10316082835197449, "learning_rate": 2.177465957794054e-05, "loss": 0.4479, "num_tokens": 905342129.0, "step": 1142 }, { "epoch": 1.5598224953063662, "grad_norm": 0.09931397438049316, "learning_rate": 2.1748150579080667e-05, "loss": 0.427, "num_tokens": 906100800.0, "step": 1143 }, { "epoch": 1.5611879160266255, "grad_norm": 0.10551901161670685, "learning_rate": 2.1721642126555647e-05, "loss": 0.4107, "num_tokens": 906867834.0, "step": 1144 }, { "epoch": 1.5625533367468851, "grad_norm": 0.11511816829442978, "learning_rate": 2.1695134277870054e-05, "loss": 0.4499, "num_tokens": 907606847.0, "step": 1145 }, { "epoch": 1.5639187574671447, "grad_norm": 0.11347976326942444, "learning_rate": 2.1668627090527143e-05, "loss": 0.4531, "num_tokens": 908418404.0, "step": 1146 }, { "epoch": 1.565284178187404, "grad_norm": 0.10698295384645462, "learning_rate": 2.1642120622028737e-05, "loss": 0.4359, "num_tokens": 909256031.0, "step": 1147 }, { "epoch": 1.5666495989076634, "grad_norm": 0.10423040390014648, "learning_rate": 2.16156149298751e-05, "loss": 0.4206, "num_tokens": 910012655.0, "step": 1148 }, { "epoch": 1.5680150196279228, "grad_norm": 0.09503877907991409, "learning_rate": 2.158911007156481e-05, "loss": 0.4115, "num_tokens": 910830084.0, "step": 1149 }, { "epoch": 1.5693804403481821, "grad_norm": 0.11356286704540253, "learning_rate": 2.1562606104594633e-05, "loss": 0.4414, "num_tokens": 911651424.0, "step": 1150 }, { "epoch": 1.5707458610684417, "grad_norm": 0.10226497054100037, "learning_rate": 2.15361030864594e-05, "loss": 0.4122, "num_tokens": 912334147.0, "step": 1151 }, { "epoch": 1.5721112817887013, "grad_norm": 0.10798600316047668, "learning_rate": 2.1509601074651898e-05, "loss": 0.4452, "num_tokens": 913148278.0, "step": 1152 }, { "epoch": 1.5734767025089607, "grad_norm": 0.09717244654893875, "learning_rate": 2.1483100126662716e-05, "loss": 0.4361, "num_tokens": 913952788.0, "step": 1153 }, { "epoch": 1.57484212322922, "grad_norm": 0.10743214190006256, "learning_rate": 2.1456600299980148e-05, "loss": 0.4303, "num_tokens": 914772360.0, "step": 1154 }, { "epoch": 1.5762075439494794, "grad_norm": 0.1024937629699707, "learning_rate": 2.1430101652090042e-05, "loss": 0.4305, "num_tokens": 915562794.0, "step": 1155 }, { "epoch": 1.5775729646697387, "grad_norm": 0.1062883734703064, "learning_rate": 2.140360424047569e-05, "loss": 0.4541, "num_tokens": 916296484.0, "step": 1156 }, { "epoch": 1.5789383853899983, "grad_norm": 0.10940805822610855, "learning_rate": 2.1377108122617717e-05, "loss": 0.4567, "num_tokens": 917129759.0, "step": 1157 }, { "epoch": 1.5803038061102577, "grad_norm": 0.1017025038599968, "learning_rate": 2.135061335599393e-05, "loss": 0.4343, "num_tokens": 917924737.0, "step": 1158 }, { "epoch": 1.5816692268305172, "grad_norm": 0.11112738400697708, "learning_rate": 2.1324119998079196e-05, "loss": 0.4623, "num_tokens": 918677239.0, "step": 1159 }, { "epoch": 1.5830346475507766, "grad_norm": 0.10579797625541687, "learning_rate": 2.1297628106345357e-05, "loss": 0.4432, "num_tokens": 919493236.0, "step": 1160 }, { "epoch": 1.584400068271036, "grad_norm": 0.1019376888871193, "learning_rate": 2.1271137738261033e-05, "loss": 0.4093, "num_tokens": 920237123.0, "step": 1161 }, { "epoch": 1.5857654889912953, "grad_norm": 0.10801465064287186, "learning_rate": 2.1244648951291578e-05, "loss": 0.4204, "num_tokens": 921026469.0, "step": 1162 }, { "epoch": 1.587130909711555, "grad_norm": 0.09357950836420059, "learning_rate": 2.121816180289888e-05, "loss": 0.423, "num_tokens": 921871168.0, "step": 1163 }, { "epoch": 1.5884963304318143, "grad_norm": 0.10583366453647614, "learning_rate": 2.11916763505413e-05, "loss": 0.4209, "num_tokens": 922680274.0, "step": 1164 }, { "epoch": 1.5898617511520738, "grad_norm": 0.10824963450431824, "learning_rate": 2.116519265167351e-05, "loss": 0.4297, "num_tokens": 923455809.0, "step": 1165 }, { "epoch": 1.5912271718723332, "grad_norm": 0.11430494487285614, "learning_rate": 2.1138710763746385e-05, "loss": 0.4288, "num_tokens": 924237147.0, "step": 1166 }, { "epoch": 1.5925925925925926, "grad_norm": 0.10338452458381653, "learning_rate": 2.1112230744206847e-05, "loss": 0.428, "num_tokens": 925142574.0, "step": 1167 }, { "epoch": 1.593958013312852, "grad_norm": 0.09948620200157166, "learning_rate": 2.1085752650497796e-05, "loss": 0.4252, "num_tokens": 925929169.0, "step": 1168 }, { "epoch": 1.5953234340331115, "grad_norm": 0.10861583054065704, "learning_rate": 2.105927654005794e-05, "loss": 0.4113, "num_tokens": 926694420.0, "step": 1169 }, { "epoch": 1.5966888547533709, "grad_norm": 0.10676854848861694, "learning_rate": 2.1032802470321686e-05, "loss": 0.4336, "num_tokens": 927479350.0, "step": 1170 }, { "epoch": 1.5980542754736304, "grad_norm": 0.10202588886022568, "learning_rate": 2.1006330498719008e-05, "loss": 0.4137, "num_tokens": 928242692.0, "step": 1171 }, { "epoch": 1.5994196961938898, "grad_norm": 0.09805753827095032, "learning_rate": 2.0979860682675344e-05, "loss": 0.4307, "num_tokens": 929044980.0, "step": 1172 }, { "epoch": 1.6007851169141492, "grad_norm": 0.10406523942947388, "learning_rate": 2.095339307961144e-05, "loss": 0.4344, "num_tokens": 929865577.0, "step": 1173 }, { "epoch": 1.6021505376344085, "grad_norm": 0.10491067916154861, "learning_rate": 2.0926927746943243e-05, "loss": 0.4333, "num_tokens": 930689945.0, "step": 1174 }, { "epoch": 1.6035159583546679, "grad_norm": 0.09647490084171295, "learning_rate": 2.0900464742081793e-05, "loss": 0.4257, "num_tokens": 931497109.0, "step": 1175 }, { "epoch": 1.6048813790749274, "grad_norm": 0.10221977531909943, "learning_rate": 2.087400412243305e-05, "loss": 0.4196, "num_tokens": 932269630.0, "step": 1176 }, { "epoch": 1.606246799795187, "grad_norm": 0.09734781831502914, "learning_rate": 2.0847545945397836e-05, "loss": 0.4244, "num_tokens": 933024387.0, "step": 1177 }, { "epoch": 1.6076122205154464, "grad_norm": 0.11743166297674179, "learning_rate": 2.0821090268371648e-05, "loss": 0.4406, "num_tokens": 933831869.0, "step": 1178 }, { "epoch": 1.6089776412357057, "grad_norm": 0.10071767121553421, "learning_rate": 2.0794637148744553e-05, "loss": 0.4244, "num_tokens": 934627778.0, "step": 1179 }, { "epoch": 1.610343061955965, "grad_norm": 0.10254139453172684, "learning_rate": 2.0768186643901094e-05, "loss": 0.4321, "num_tokens": 935360207.0, "step": 1180 }, { "epoch": 1.6117084826762245, "grad_norm": 0.09870327264070511, "learning_rate": 2.0741738811220142e-05, "loss": 0.4286, "num_tokens": 936090972.0, "step": 1181 }, { "epoch": 1.613073903396484, "grad_norm": 0.1082155704498291, "learning_rate": 2.0715293708074745e-05, "loss": 0.4306, "num_tokens": 936943765.0, "step": 1182 }, { "epoch": 1.6144393241167436, "grad_norm": 0.10144319385290146, "learning_rate": 2.0688851391832058e-05, "loss": 0.4242, "num_tokens": 937720450.0, "step": 1183 }, { "epoch": 1.615804744837003, "grad_norm": 0.09733365476131439, "learning_rate": 2.0662411919853177e-05, "loss": 0.4525, "num_tokens": 938528702.0, "step": 1184 }, { "epoch": 1.6171701655572623, "grad_norm": 0.10772135853767395, "learning_rate": 2.063597534949302e-05, "loss": 0.4141, "num_tokens": 939328579.0, "step": 1185 }, { "epoch": 1.6185355862775217, "grad_norm": 0.11240187287330627, "learning_rate": 2.0609541738100223e-05, "loss": 0.4508, "num_tokens": 940098976.0, "step": 1186 }, { "epoch": 1.619901006997781, "grad_norm": 0.09764362871646881, "learning_rate": 2.0583111143017007e-05, "loss": 0.4059, "num_tokens": 940869277.0, "step": 1187 }, { "epoch": 1.6212664277180406, "grad_norm": 0.10536693781614304, "learning_rate": 2.0556683621579037e-05, "loss": 0.4437, "num_tokens": 941730902.0, "step": 1188 }, { "epoch": 1.6226318484383002, "grad_norm": 0.09326350688934326, "learning_rate": 2.0530259231115308e-05, "loss": 0.435, "num_tokens": 942580096.0, "step": 1189 }, { "epoch": 1.6239972691585596, "grad_norm": 0.10353576391935349, "learning_rate": 2.0503838028948055e-05, "loss": 0.4357, "num_tokens": 943384404.0, "step": 1190 }, { "epoch": 1.625362689878819, "grad_norm": 0.10457667708396912, "learning_rate": 2.0477420072392555e-05, "loss": 0.4363, "num_tokens": 944187628.0, "step": 1191 }, { "epoch": 1.6267281105990783, "grad_norm": 0.10501637309789658, "learning_rate": 2.0451005418757058e-05, "loss": 0.4195, "num_tokens": 944996026.0, "step": 1192 }, { "epoch": 1.6280935313193377, "grad_norm": 0.09834054857492447, "learning_rate": 2.0424594125342666e-05, "loss": 0.4256, "num_tokens": 945830258.0, "step": 1193 }, { "epoch": 1.6294589520395972, "grad_norm": 0.09918879717588425, "learning_rate": 2.0398186249443172e-05, "loss": 0.4181, "num_tokens": 946630900.0, "step": 1194 }, { "epoch": 1.6308243727598566, "grad_norm": 0.09979560971260071, "learning_rate": 2.037178184834496e-05, "loss": 0.4409, "num_tokens": 947433484.0, "step": 1195 }, { "epoch": 1.6321897934801162, "grad_norm": 0.1088239848613739, "learning_rate": 2.034538097932689e-05, "loss": 0.4457, "num_tokens": 948339313.0, "step": 1196 }, { "epoch": 1.6335552142003755, "grad_norm": 0.10808560252189636, "learning_rate": 2.031898369966013e-05, "loss": 0.434, "num_tokens": 949106309.0, "step": 1197 }, { "epoch": 1.6349206349206349, "grad_norm": 0.0944829136133194, "learning_rate": 2.029259006660809e-05, "loss": 0.4188, "num_tokens": 949885415.0, "step": 1198 }, { "epoch": 1.6362860556408942, "grad_norm": 0.10671178251504898, "learning_rate": 2.0266200137426263e-05, "loss": 0.4443, "num_tokens": 950717090.0, "step": 1199 }, { "epoch": 1.6376514763611538, "grad_norm": 0.09412442147731781, "learning_rate": 2.0239813969362097e-05, "loss": 0.4134, "num_tokens": 951505245.0, "step": 1200 }, { "epoch": 1.6390168970814132, "grad_norm": 0.09649308025836945, "learning_rate": 2.0213431619654894e-05, "loss": 0.4034, "num_tokens": 952250524.0, "step": 1201 }, { "epoch": 1.6403823178016728, "grad_norm": 0.11565569788217545, "learning_rate": 2.0187053145535663e-05, "loss": 0.4173, "num_tokens": 952984198.0, "step": 1202 }, { "epoch": 1.6417477385219321, "grad_norm": 0.09909109771251678, "learning_rate": 2.0160678604227e-05, "loss": 0.4465, "num_tokens": 953833244.0, "step": 1203 }, { "epoch": 1.6431131592421915, "grad_norm": 0.11315872520208359, "learning_rate": 2.0134308052942995e-05, "loss": 0.4249, "num_tokens": 954537461.0, "step": 1204 }, { "epoch": 1.6444785799624508, "grad_norm": 0.10907097160816193, "learning_rate": 2.0107941548889056e-05, "loss": 0.4124, "num_tokens": 955386831.0, "step": 1205 }, { "epoch": 1.6458440006827102, "grad_norm": 0.09925055503845215, "learning_rate": 2.008157914926182e-05, "loss": 0.4303, "num_tokens": 956215017.0, "step": 1206 }, { "epoch": 1.6472094214029698, "grad_norm": 0.09871881455183029, "learning_rate": 2.0055220911249034e-05, "loss": 0.434, "num_tokens": 957009672.0, "step": 1207 }, { "epoch": 1.6485748421232294, "grad_norm": 0.10397522151470184, "learning_rate": 2.0028866892029395e-05, "loss": 0.4217, "num_tokens": 957769231.0, "step": 1208 }, { "epoch": 1.6499402628434887, "grad_norm": 0.1142621785402298, "learning_rate": 2.000251714877246e-05, "loss": 0.4362, "num_tokens": 958550093.0, "step": 1209 }, { "epoch": 1.651305683563748, "grad_norm": 0.10088682174682617, "learning_rate": 1.9976171738638504e-05, "loss": 0.4182, "num_tokens": 959279883.0, "step": 1210 }, { "epoch": 1.6526711042840074, "grad_norm": 0.10685870796442032, "learning_rate": 1.9949830718778415e-05, "loss": 0.4268, "num_tokens": 960157449.0, "step": 1211 }, { "epoch": 1.6540365250042668, "grad_norm": 0.10144811123609543, "learning_rate": 1.9923494146333544e-05, "loss": 0.4196, "num_tokens": 960970246.0, "step": 1212 }, { "epoch": 1.6554019457245264, "grad_norm": 0.10008132457733154, "learning_rate": 1.9897162078435613e-05, "loss": 0.4347, "num_tokens": 961782156.0, "step": 1213 }, { "epoch": 1.656767366444786, "grad_norm": 0.1032542884349823, "learning_rate": 1.9870834572206527e-05, "loss": 0.4248, "num_tokens": 962607862.0, "step": 1214 }, { "epoch": 1.6581327871650453, "grad_norm": 0.09882889688014984, "learning_rate": 1.984451168475836e-05, "loss": 0.4253, "num_tokens": 963402788.0, "step": 1215 }, { "epoch": 1.6594982078853047, "grad_norm": 0.11136434972286224, "learning_rate": 1.9818193473193113e-05, "loss": 0.3962, "num_tokens": 964120512.0, "step": 1216 }, { "epoch": 1.660863628605564, "grad_norm": 0.09195875376462936, "learning_rate": 1.9791879994602674e-05, "loss": 0.4279, "num_tokens": 964941873.0, "step": 1217 }, { "epoch": 1.6622290493258234, "grad_norm": 0.12059548497200012, "learning_rate": 1.9765571306068647e-05, "loss": 0.4378, "num_tokens": 965698041.0, "step": 1218 }, { "epoch": 1.663594470046083, "grad_norm": 0.10328560322523117, "learning_rate": 1.973926746466226e-05, "loss": 0.4085, "num_tokens": 966446979.0, "step": 1219 }, { "epoch": 1.6649598907663425, "grad_norm": 0.10857341438531876, "learning_rate": 1.9712968527444216e-05, "loss": 0.4122, "num_tokens": 967230354.0, "step": 1220 }, { "epoch": 1.666325311486602, "grad_norm": 0.09499169886112213, "learning_rate": 1.9686674551464566e-05, "loss": 0.4253, "num_tokens": 967994746.0, "step": 1221 }, { "epoch": 1.6676907322068613, "grad_norm": 0.11111078411340714, "learning_rate": 1.966038559376264e-05, "loss": 0.4313, "num_tokens": 968765497.0, "step": 1222 }, { "epoch": 1.6690561529271206, "grad_norm": 0.107989102602005, "learning_rate": 1.9634101711366833e-05, "loss": 0.4198, "num_tokens": 969495054.0, "step": 1223 }, { "epoch": 1.67042157364738, "grad_norm": 0.10619103908538818, "learning_rate": 1.9607822961294567e-05, "loss": 0.4203, "num_tokens": 970254151.0, "step": 1224 }, { "epoch": 1.6717869943676396, "grad_norm": 0.10418576747179031, "learning_rate": 1.958154940055212e-05, "loss": 0.3901, "num_tokens": 970933870.0, "step": 1225 }, { "epoch": 1.673152415087899, "grad_norm": 0.11449429392814636, "learning_rate": 1.9555281086134496e-05, "loss": 0.4329, "num_tokens": 971766168.0, "step": 1226 }, { "epoch": 1.6745178358081585, "grad_norm": 0.09717535227537155, "learning_rate": 1.952901807502533e-05, "loss": 0.4265, "num_tokens": 972619577.0, "step": 1227 }, { "epoch": 1.6758832565284179, "grad_norm": 0.11108912527561188, "learning_rate": 1.9502760424196773e-05, "loss": 0.4292, "num_tokens": 973452752.0, "step": 1228 }, { "epoch": 1.6772486772486772, "grad_norm": 0.09589552879333496, "learning_rate": 1.9476508190609316e-05, "loss": 0.4133, "num_tokens": 974254390.0, "step": 1229 }, { "epoch": 1.6786140979689366, "grad_norm": 0.09917382150888443, "learning_rate": 1.945026143121171e-05, "loss": 0.4088, "num_tokens": 974988187.0, "step": 1230 }, { "epoch": 1.6799795186891961, "grad_norm": 0.11548617482185364, "learning_rate": 1.9424020202940844e-05, "loss": 0.4203, "num_tokens": 975742863.0, "step": 1231 }, { "epoch": 1.6813449394094555, "grad_norm": 0.10977286845445633, "learning_rate": 1.939778456272158e-05, "loss": 0.4348, "num_tokens": 976557115.0, "step": 1232 }, { "epoch": 1.682710360129715, "grad_norm": 0.09885360300540924, "learning_rate": 1.9371554567466687e-05, "loss": 0.4388, "num_tokens": 977330538.0, "step": 1233 }, { "epoch": 1.6840757808499744, "grad_norm": 0.11047488451004028, "learning_rate": 1.9345330274076675e-05, "loss": 0.4084, "num_tokens": 978106762.0, "step": 1234 }, { "epoch": 1.6854412015702338, "grad_norm": 0.10932295024394989, "learning_rate": 1.931911173943968e-05, "loss": 0.4501, "num_tokens": 978908908.0, "step": 1235 }, { "epoch": 1.6868066222904932, "grad_norm": 0.09673037379980087, "learning_rate": 1.9292899020431356e-05, "loss": 0.4171, "num_tokens": 979757591.0, "step": 1236 }, { "epoch": 1.6881720430107527, "grad_norm": 0.10407116264104843, "learning_rate": 1.9266692173914747e-05, "loss": 0.4312, "num_tokens": 980626835.0, "step": 1237 }, { "epoch": 1.689537463731012, "grad_norm": 0.09228197485208511, "learning_rate": 1.9240491256740127e-05, "loss": 0.4037, "num_tokens": 981404618.0, "step": 1238 }, { "epoch": 1.6909028844512717, "grad_norm": 0.10878165811300278, "learning_rate": 1.921429632574494e-05, "loss": 0.4244, "num_tokens": 982201485.0, "step": 1239 }, { "epoch": 1.692268305171531, "grad_norm": 0.09236898273229599, "learning_rate": 1.9188107437753633e-05, "loss": 0.4389, "num_tokens": 983013541.0, "step": 1240 }, { "epoch": 1.6936337258917904, "grad_norm": 0.10652187466621399, "learning_rate": 1.9161924649577542e-05, "loss": 0.3975, "num_tokens": 983742301.0, "step": 1241 }, { "epoch": 1.6949991466120498, "grad_norm": 0.1030779629945755, "learning_rate": 1.913574801801478e-05, "loss": 0.4257, "num_tokens": 984490034.0, "step": 1242 }, { "epoch": 1.6963645673323091, "grad_norm": 0.1025134027004242, "learning_rate": 1.9109577599850083e-05, "loss": 0.4161, "num_tokens": 985271325.0, "step": 1243 }, { "epoch": 1.6977299880525687, "grad_norm": 0.10931537300348282, "learning_rate": 1.908341345185472e-05, "loss": 0.4299, "num_tokens": 986102589.0, "step": 1244 }, { "epoch": 1.6990954087728283, "grad_norm": 0.10324174165725708, "learning_rate": 1.9057255630786376e-05, "loss": 0.4197, "num_tokens": 986905815.0, "step": 1245 }, { "epoch": 1.7004608294930876, "grad_norm": 0.08915841579437256, "learning_rate": 1.903110419338899e-05, "loss": 0.4336, "num_tokens": 987683747.0, "step": 1246 }, { "epoch": 1.701826250213347, "grad_norm": 0.11005804687738419, "learning_rate": 1.9004959196392652e-05, "loss": 0.4339, "num_tokens": 988499119.0, "step": 1247 }, { "epoch": 1.7031916709336063, "grad_norm": 0.11490438133478165, "learning_rate": 1.89788206965135e-05, "loss": 0.4336, "num_tokens": 989278731.0, "step": 1248 }, { "epoch": 1.7045570916538657, "grad_norm": 0.09970224648714066, "learning_rate": 1.895268875045355e-05, "loss": 0.416, "num_tokens": 990123804.0, "step": 1249 }, { "epoch": 1.7059225123741253, "grad_norm": 0.10445905476808548, "learning_rate": 1.892656341490062e-05, "loss": 0.4234, "num_tokens": 990912624.0, "step": 1250 }, { "epoch": 1.7072879330943849, "grad_norm": 0.12103025615215302, "learning_rate": 1.890044474652818e-05, "loss": 0.4155, "num_tokens": 991641761.0, "step": 1251 }, { "epoch": 1.7086533538146442, "grad_norm": 0.1032596006989479, "learning_rate": 1.8874332801995258e-05, "loss": 0.437, "num_tokens": 992427975.0, "step": 1252 }, { "epoch": 1.7100187745349036, "grad_norm": 0.11127634346485138, "learning_rate": 1.8848227637946257e-05, "loss": 0.4053, "num_tokens": 993173677.0, "step": 1253 }, { "epoch": 1.711384195255163, "grad_norm": 0.10870165377855301, "learning_rate": 1.8822129311010918e-05, "loss": 0.4222, "num_tokens": 993969869.0, "step": 1254 }, { "epoch": 1.7127496159754223, "grad_norm": 0.11180712282657623, "learning_rate": 1.8796037877804098e-05, "loss": 0.4233, "num_tokens": 994735532.0, "step": 1255 }, { "epoch": 1.7141150366956819, "grad_norm": 0.09520142525434494, "learning_rate": 1.8769953394925747e-05, "loss": 0.3875, "num_tokens": 995472010.0, "step": 1256 }, { "epoch": 1.7154804574159412, "grad_norm": 0.10195495188236237, "learning_rate": 1.8743875918960702e-05, "loss": 0.4256, "num_tokens": 996229055.0, "step": 1257 }, { "epoch": 1.7168458781362008, "grad_norm": 0.11490486562252045, "learning_rate": 1.8717805506478624e-05, "loss": 0.4373, "num_tokens": 997075833.0, "step": 1258 }, { "epoch": 1.7182112988564602, "grad_norm": 0.09933003783226013, "learning_rate": 1.869174221403385e-05, "loss": 0.408, "num_tokens": 997834457.0, "step": 1259 }, { "epoch": 1.7195767195767195, "grad_norm": 0.09807156771421432, "learning_rate": 1.866568609816525e-05, "loss": 0.4273, "num_tokens": 998687422.0, "step": 1260 }, { "epoch": 1.720942140296979, "grad_norm": 0.1102372333407402, "learning_rate": 1.863963721539614e-05, "loss": 0.4097, "num_tokens": 999459349.0, "step": 1261 }, { "epoch": 1.7223075610172385, "grad_norm": 0.09855660051107407, "learning_rate": 1.8613595622234158e-05, "loss": 0.4384, "num_tokens": 1000275200.0, "step": 1262 }, { "epoch": 1.7236729817374978, "grad_norm": 0.09799787402153015, "learning_rate": 1.8587561375171104e-05, "loss": 0.4251, "num_tokens": 1001022661.0, "step": 1263 }, { "epoch": 1.7250384024577574, "grad_norm": 0.10651054233312607, "learning_rate": 1.856153453068286e-05, "loss": 0.4136, "num_tokens": 1001795746.0, "step": 1264 }, { "epoch": 1.7264038231780168, "grad_norm": 0.09899908304214478, "learning_rate": 1.8535515145229247e-05, "loss": 0.4331, "num_tokens": 1002641080.0, "step": 1265 }, { "epoch": 1.7277692438982761, "grad_norm": 0.10645182430744171, "learning_rate": 1.85095032752539e-05, "loss": 0.4159, "num_tokens": 1003431199.0, "step": 1266 }, { "epoch": 1.7291346646185355, "grad_norm": 0.09632181376218796, "learning_rate": 1.8483498977184152e-05, "loss": 0.4014, "num_tokens": 1004199642.0, "step": 1267 }, { "epoch": 1.730500085338795, "grad_norm": 0.09573043882846832, "learning_rate": 1.84575023074309e-05, "loss": 0.403, "num_tokens": 1004955479.0, "step": 1268 }, { "epoch": 1.7318655060590544, "grad_norm": 0.09624087810516357, "learning_rate": 1.8431513322388522e-05, "loss": 0.4053, "num_tokens": 1005731414.0, "step": 1269 }, { "epoch": 1.733230926779314, "grad_norm": 0.10934878140687943, "learning_rate": 1.8405532078434695e-05, "loss": 0.4165, "num_tokens": 1006506654.0, "step": 1270 }, { "epoch": 1.7345963474995734, "grad_norm": 0.10288947075605392, "learning_rate": 1.8379558631930324e-05, "loss": 0.4221, "num_tokens": 1007327791.0, "step": 1271 }, { "epoch": 1.7359617682198327, "grad_norm": 0.10705222189426422, "learning_rate": 1.835359303921939e-05, "loss": 0.4557, "num_tokens": 1008149678.0, "step": 1272 }, { "epoch": 1.737327188940092, "grad_norm": 0.1046479195356369, "learning_rate": 1.832763535662884e-05, "loss": 0.429, "num_tokens": 1008946649.0, "step": 1273 }, { "epoch": 1.7386926096603514, "grad_norm": 0.10460752993822098, "learning_rate": 1.830168564046845e-05, "loss": 0.4305, "num_tokens": 1009730603.0, "step": 1274 }, { "epoch": 1.740058030380611, "grad_norm": 0.10355538129806519, "learning_rate": 1.827574394703073e-05, "loss": 0.4384, "num_tokens": 1010490882.0, "step": 1275 }, { "epoch": 1.7414234511008706, "grad_norm": 0.11373353749513626, "learning_rate": 1.824981033259078e-05, "loss": 0.4402, "num_tokens": 1011213415.0, "step": 1276 }, { "epoch": 1.74278887182113, "grad_norm": 0.09712375700473785, "learning_rate": 1.8223884853406172e-05, "loss": 0.4194, "num_tokens": 1011991059.0, "step": 1277 }, { "epoch": 1.7441542925413893, "grad_norm": 0.08968063443899155, "learning_rate": 1.8197967565716847e-05, "loss": 0.4189, "num_tokens": 1012823541.0, "step": 1278 }, { "epoch": 1.7455197132616487, "grad_norm": 0.10207287967205048, "learning_rate": 1.8172058525744948e-05, "loss": 0.4339, "num_tokens": 1013657151.0, "step": 1279 }, { "epoch": 1.746885133981908, "grad_norm": 0.1041182354092598, "learning_rate": 1.8146157789694737e-05, "loss": 0.4313, "num_tokens": 1014457375.0, "step": 1280 }, { "epoch": 1.7482505547021676, "grad_norm": 0.10371288657188416, "learning_rate": 1.8120265413752475e-05, "loss": 0.409, "num_tokens": 1015316709.0, "step": 1281 }, { "epoch": 1.7496159754224272, "grad_norm": 0.1026940569281578, "learning_rate": 1.8094381454086278e-05, "loss": 0.4247, "num_tokens": 1016091820.0, "step": 1282 }, { "epoch": 1.7509813961426866, "grad_norm": 0.09848587960004807, "learning_rate": 1.8068505966846007e-05, "loss": 0.4351, "num_tokens": 1016938282.0, "step": 1283 }, { "epoch": 1.752346816862946, "grad_norm": 0.10114136338233948, "learning_rate": 1.8042639008163133e-05, "loss": 0.4269, "num_tokens": 1017732359.0, "step": 1284 }, { "epoch": 1.7537122375832053, "grad_norm": 0.1001061201095581, "learning_rate": 1.8016780634150635e-05, "loss": 0.424, "num_tokens": 1018603105.0, "step": 1285 }, { "epoch": 1.7550776583034646, "grad_norm": 0.10559725761413574, "learning_rate": 1.7990930900902884e-05, "loss": 0.4424, "num_tokens": 1019496785.0, "step": 1286 }, { "epoch": 1.7564430790237242, "grad_norm": 0.09608396142721176, "learning_rate": 1.7965089864495472e-05, "loss": 0.4205, "num_tokens": 1020244646.0, "step": 1287 }, { "epoch": 1.7578084997439836, "grad_norm": 0.10634123533964157, "learning_rate": 1.7939257580985162e-05, "loss": 0.4417, "num_tokens": 1021067981.0, "step": 1288 }, { "epoch": 1.7591739204642431, "grad_norm": 0.10522617399692535, "learning_rate": 1.791343410640971e-05, "loss": 0.4201, "num_tokens": 1021773796.0, "step": 1289 }, { "epoch": 1.7605393411845025, "grad_norm": 0.10655199736356735, "learning_rate": 1.7887619496787754e-05, "loss": 0.4462, "num_tokens": 1022592246.0, "step": 1290 }, { "epoch": 1.7619047619047619, "grad_norm": 0.10434888303279877, "learning_rate": 1.786181380811872e-05, "loss": 0.4178, "num_tokens": 1023286591.0, "step": 1291 }, { "epoch": 1.7632701826250212, "grad_norm": 0.10192246735095978, "learning_rate": 1.7836017096382673e-05, "loss": 0.4318, "num_tokens": 1024030669.0, "step": 1292 }, { "epoch": 1.7646356033452808, "grad_norm": 0.09775141626596451, "learning_rate": 1.7810229417540213e-05, "loss": 0.4261, "num_tokens": 1024788545.0, "step": 1293 }, { "epoch": 1.7660010240655402, "grad_norm": 0.10783270746469498, "learning_rate": 1.7784450827532323e-05, "loss": 0.4104, "num_tokens": 1025578420.0, "step": 1294 }, { "epoch": 1.7673664447857997, "grad_norm": 0.10182036459445953, "learning_rate": 1.7758681382280302e-05, "loss": 0.4404, "num_tokens": 1026412530.0, "step": 1295 }, { "epoch": 1.768731865506059, "grad_norm": 0.10119203478097916, "learning_rate": 1.773292113768558e-05, "loss": 0.4397, "num_tokens": 1027337955.0, "step": 1296 }, { "epoch": 1.7700972862263185, "grad_norm": 0.09669125825166702, "learning_rate": 1.770717014962965e-05, "loss": 0.438, "num_tokens": 1028168892.0, "step": 1297 }, { "epoch": 1.7714627069465778, "grad_norm": 0.11458604037761688, "learning_rate": 1.7681428473973908e-05, "loss": 0.4133, "num_tokens": 1028944499.0, "step": 1298 }, { "epoch": 1.7728281276668374, "grad_norm": 0.09683628380298615, "learning_rate": 1.7655696166559563e-05, "loss": 0.4077, "num_tokens": 1029673074.0, "step": 1299 }, { "epoch": 1.7741935483870968, "grad_norm": 0.10090797394514084, "learning_rate": 1.76299732832075e-05, "loss": 0.4005, "num_tokens": 1030458876.0, "step": 1300 }, { "epoch": 1.7755589691073563, "grad_norm": 0.12856976687908173, "learning_rate": 1.760425987971816e-05, "loss": 0.4571, "num_tokens": 1031241488.0, "step": 1301 }, { "epoch": 1.7769243898276157, "grad_norm": 0.09310390055179596, "learning_rate": 1.7578556011871398e-05, "loss": 0.4222, "num_tokens": 1032011071.0, "step": 1302 }, { "epoch": 1.778289810547875, "grad_norm": 0.10239263623952866, "learning_rate": 1.7552861735426415e-05, "loss": 0.4144, "num_tokens": 1032764805.0, "step": 1303 }, { "epoch": 1.7796552312681344, "grad_norm": 0.11468919366598129, "learning_rate": 1.7527177106121588e-05, "loss": 0.4083, "num_tokens": 1033471368.0, "step": 1304 }, { "epoch": 1.7810206519883938, "grad_norm": 0.10556650906801224, "learning_rate": 1.7501502179674378e-05, "loss": 0.4074, "num_tokens": 1034188086.0, "step": 1305 }, { "epoch": 1.7823860727086533, "grad_norm": 0.10582953691482544, "learning_rate": 1.7475837011781175e-05, "loss": 0.4341, "num_tokens": 1034978588.0, "step": 1306 }, { "epoch": 1.783751493428913, "grad_norm": 0.10499069839715958, "learning_rate": 1.745018165811723e-05, "loss": 0.4262, "num_tokens": 1035721266.0, "step": 1307 }, { "epoch": 1.7851169141491723, "grad_norm": 0.09730007499456406, "learning_rate": 1.7424536174336478e-05, "loss": 0.4084, "num_tokens": 1036502290.0, "step": 1308 }, { "epoch": 1.7864823348694316, "grad_norm": 0.10208161175251007, "learning_rate": 1.7398900616071457e-05, "loss": 0.4115, "num_tokens": 1037255022.0, "step": 1309 }, { "epoch": 1.787847755589691, "grad_norm": 0.1013292595744133, "learning_rate": 1.737327503893318e-05, "loss": 0.407, "num_tokens": 1038019772.0, "step": 1310 }, { "epoch": 1.7892131763099504, "grad_norm": 0.10515838116407394, "learning_rate": 1.7347659498510983e-05, "loss": 0.4236, "num_tokens": 1038775929.0, "step": 1311 }, { "epoch": 1.79057859703021, "grad_norm": 0.11257584393024445, "learning_rate": 1.7322054050372455e-05, "loss": 0.4251, "num_tokens": 1039558152.0, "step": 1312 }, { "epoch": 1.7919440177504695, "grad_norm": 0.09948402643203735, "learning_rate": 1.7296458750063285e-05, "loss": 0.4598, "num_tokens": 1040345161.0, "step": 1313 }, { "epoch": 1.7933094384707289, "grad_norm": 0.11289824545383453, "learning_rate": 1.7270873653107143e-05, "loss": 0.4124, "num_tokens": 1041099101.0, "step": 1314 }, { "epoch": 1.7946748591909882, "grad_norm": 0.14673297107219696, "learning_rate": 1.724529881500556e-05, "loss": 0.425, "num_tokens": 1041891210.0, "step": 1315 }, { "epoch": 1.7960402799112476, "grad_norm": 0.09155464172363281, "learning_rate": 1.7219734291237832e-05, "loss": 0.4135, "num_tokens": 1042594045.0, "step": 1316 }, { "epoch": 1.797405700631507, "grad_norm": 0.11391695588827133, "learning_rate": 1.7194180137260853e-05, "loss": 0.4343, "num_tokens": 1043352629.0, "step": 1317 }, { "epoch": 1.7987711213517665, "grad_norm": 0.11216048896312714, "learning_rate": 1.7168636408509064e-05, "loss": 0.4148, "num_tokens": 1044151121.0, "step": 1318 }, { "epoch": 1.800136542072026, "grad_norm": 0.10885075479745865, "learning_rate": 1.714310316039424e-05, "loss": 0.4304, "num_tokens": 1044859590.0, "step": 1319 }, { "epoch": 1.8015019627922855, "grad_norm": 0.09847783297300339, "learning_rate": 1.7117580448305454e-05, "loss": 0.4095, "num_tokens": 1045606465.0, "step": 1320 }, { "epoch": 1.8028673835125448, "grad_norm": 0.10576991736888885, "learning_rate": 1.7092068327608903e-05, "loss": 0.4146, "num_tokens": 1046290193.0, "step": 1321 }, { "epoch": 1.8042328042328042, "grad_norm": 0.11168473213911057, "learning_rate": 1.706656685364783e-05, "loss": 0.4087, "num_tokens": 1046976064.0, "step": 1322 }, { "epoch": 1.8055982249530635, "grad_norm": 0.09181300550699234, "learning_rate": 1.704107608174237e-05, "loss": 0.4345, "num_tokens": 1047747108.0, "step": 1323 }, { "epoch": 1.8069636456733231, "grad_norm": 0.09620440006256104, "learning_rate": 1.7015596067189444e-05, "loss": 0.4295, "num_tokens": 1048553007.0, "step": 1324 }, { "epoch": 1.8083290663935825, "grad_norm": 0.10641511529684067, "learning_rate": 1.6990126865262634e-05, "loss": 0.4475, "num_tokens": 1049325794.0, "step": 1325 }, { "epoch": 1.809694487113842, "grad_norm": 0.11064237356185913, "learning_rate": 1.6964668531212064e-05, "loss": 0.4123, "num_tokens": 1050040355.0, "step": 1326 }, { "epoch": 1.8110599078341014, "grad_norm": 0.10759802162647247, "learning_rate": 1.693922112026429e-05, "loss": 0.4194, "num_tokens": 1050864685.0, "step": 1327 }, { "epoch": 1.8124253285543608, "grad_norm": 0.09407397359609604, "learning_rate": 1.691378468762217e-05, "loss": 0.4261, "num_tokens": 1051685988.0, "step": 1328 }, { "epoch": 1.8137907492746201, "grad_norm": 0.10178737342357635, "learning_rate": 1.688835928846475e-05, "loss": 0.4105, "num_tokens": 1052438936.0, "step": 1329 }, { "epoch": 1.8151561699948797, "grad_norm": 0.10596182197332382, "learning_rate": 1.686294497794714e-05, "loss": 0.4325, "num_tokens": 1053223670.0, "step": 1330 }, { "epoch": 1.816521590715139, "grad_norm": 0.10111228376626968, "learning_rate": 1.6837541811200382e-05, "loss": 0.4101, "num_tokens": 1053981128.0, "step": 1331 }, { "epoch": 1.8178870114353987, "grad_norm": 0.09170404821634293, "learning_rate": 1.6812149843331358e-05, "loss": 0.413, "num_tokens": 1054765567.0, "step": 1332 }, { "epoch": 1.819252432155658, "grad_norm": 0.09445077180862427, "learning_rate": 1.678676912942266e-05, "loss": 0.4366, "num_tokens": 1055586607.0, "step": 1333 }, { "epoch": 1.8206178528759174, "grad_norm": 0.10222826898097992, "learning_rate": 1.6761399724532446e-05, "loss": 0.4169, "num_tokens": 1056409531.0, "step": 1334 }, { "epoch": 1.8219832735961767, "grad_norm": 0.09889331459999084, "learning_rate": 1.6736041683694372e-05, "loss": 0.44, "num_tokens": 1057182961.0, "step": 1335 }, { "epoch": 1.823348694316436, "grad_norm": 0.10644587874412537, "learning_rate": 1.671069506191743e-05, "loss": 0.4129, "num_tokens": 1057969537.0, "step": 1336 }, { "epoch": 1.8247141150366957, "grad_norm": 0.09862502664327621, "learning_rate": 1.668535991418582e-05, "loss": 0.4134, "num_tokens": 1058771263.0, "step": 1337 }, { "epoch": 1.8260795357569553, "grad_norm": 0.09676501154899597, "learning_rate": 1.6660036295458863e-05, "loss": 0.4161, "num_tokens": 1059585123.0, "step": 1338 }, { "epoch": 1.8274449564772146, "grad_norm": 0.0959937572479248, "learning_rate": 1.6634724260670894e-05, "loss": 0.4194, "num_tokens": 1060371810.0, "step": 1339 }, { "epoch": 1.828810377197474, "grad_norm": 0.09867382794618607, "learning_rate": 1.660942386473109e-05, "loss": 0.4136, "num_tokens": 1061181439.0, "step": 1340 }, { "epoch": 1.8301757979177333, "grad_norm": 0.09332245588302612, "learning_rate": 1.658413516252339e-05, "loss": 0.432, "num_tokens": 1061943758.0, "step": 1341 }, { "epoch": 1.8315412186379927, "grad_norm": 0.1014634445309639, "learning_rate": 1.6558858208906375e-05, "loss": 0.4462, "num_tokens": 1062808721.0, "step": 1342 }, { "epoch": 1.8329066393582523, "grad_norm": 0.10539007931947708, "learning_rate": 1.653359305871312e-05, "loss": 0.4249, "num_tokens": 1063615244.0, "step": 1343 }, { "epoch": 1.8342720600785118, "grad_norm": 0.09658123552799225, "learning_rate": 1.6508339766751106e-05, "loss": 0.4508, "num_tokens": 1064386752.0, "step": 1344 }, { "epoch": 1.8356374807987712, "grad_norm": 0.1064629778265953, "learning_rate": 1.6483098387802098e-05, "loss": 0.4097, "num_tokens": 1065113745.0, "step": 1345 }, { "epoch": 1.8370029015190306, "grad_norm": 0.09992269426584244, "learning_rate": 1.6457868976622e-05, "loss": 0.4144, "num_tokens": 1065929542.0, "step": 1346 }, { "epoch": 1.83836832223929, "grad_norm": 0.09313240647315979, "learning_rate": 1.643265158794077e-05, "loss": 0.4391, "num_tokens": 1066742446.0, "step": 1347 }, { "epoch": 1.8397337429595493, "grad_norm": 0.09255487471818924, "learning_rate": 1.640744627646229e-05, "loss": 0.4099, "num_tokens": 1067526789.0, "step": 1348 }, { "epoch": 1.8410991636798089, "grad_norm": 0.10632722079753876, "learning_rate": 1.6382253096864217e-05, "loss": 0.4452, "num_tokens": 1068292661.0, "step": 1349 }, { "epoch": 1.8424645844000682, "grad_norm": 0.09883720427751541, "learning_rate": 1.635707210379791e-05, "loss": 0.4279, "num_tokens": 1069124833.0, "step": 1350 }, { "epoch": 1.8438300051203278, "grad_norm": 0.09659432619810104, "learning_rate": 1.6331903351888295e-05, "loss": 0.4303, "num_tokens": 1069941702.0, "step": 1351 }, { "epoch": 1.8451954258405872, "grad_norm": 0.09616973251104355, "learning_rate": 1.630674689573373e-05, "loss": 0.4279, "num_tokens": 1070713181.0, "step": 1352 }, { "epoch": 1.8465608465608465, "grad_norm": 0.10171636193990707, "learning_rate": 1.628160278990591e-05, "loss": 0.4266, "num_tokens": 1071520917.0, "step": 1353 }, { "epoch": 1.8479262672811059, "grad_norm": 0.10320540517568588, "learning_rate": 1.6256471088949742e-05, "loss": 0.4369, "num_tokens": 1072306312.0, "step": 1354 }, { "epoch": 1.8492916880013655, "grad_norm": 0.10071463137865067, "learning_rate": 1.62313518473832e-05, "loss": 0.4136, "num_tokens": 1073069991.0, "step": 1355 }, { "epoch": 1.8506571087216248, "grad_norm": 0.09289583563804626, "learning_rate": 1.6206245119697262e-05, "loss": 0.4485, "num_tokens": 1073889448.0, "step": 1356 }, { "epoch": 1.8520225294418844, "grad_norm": 0.10843529552221298, "learning_rate": 1.6181150960355734e-05, "loss": 0.4325, "num_tokens": 1074693586.0, "step": 1357 }, { "epoch": 1.8533879501621437, "grad_norm": 0.09780066460371017, "learning_rate": 1.615606942379517e-05, "loss": 0.4318, "num_tokens": 1075502984.0, "step": 1358 }, { "epoch": 1.854753370882403, "grad_norm": 0.10227876901626587, "learning_rate": 1.6131000564424747e-05, "loss": 0.4225, "num_tokens": 1076236906.0, "step": 1359 }, { "epoch": 1.8561187916026625, "grad_norm": 0.09862654656171799, "learning_rate": 1.6105944436626123e-05, "loss": 0.4463, "num_tokens": 1077113329.0, "step": 1360 }, { "epoch": 1.857484212322922, "grad_norm": 0.09246192872524261, "learning_rate": 1.6080901094753357e-05, "loss": 0.4226, "num_tokens": 1077909759.0, "step": 1361 }, { "epoch": 1.8588496330431814, "grad_norm": 0.10295914858579636, "learning_rate": 1.6055870593132756e-05, "loss": 0.4285, "num_tokens": 1078762966.0, "step": 1362 }, { "epoch": 1.860215053763441, "grad_norm": 0.09733414649963379, "learning_rate": 1.6030852986062782e-05, "loss": 0.4121, "num_tokens": 1079529774.0, "step": 1363 }, { "epoch": 1.8615804744837003, "grad_norm": 0.09988225996494293, "learning_rate": 1.6005848327813924e-05, "loss": 0.4332, "num_tokens": 1080252465.0, "step": 1364 }, { "epoch": 1.8629458952039597, "grad_norm": 0.0956203043460846, "learning_rate": 1.5980856672628583e-05, "loss": 0.4239, "num_tokens": 1081043315.0, "step": 1365 }, { "epoch": 1.864311315924219, "grad_norm": 0.10383440554141998, "learning_rate": 1.5955878074720938e-05, "loss": 0.4216, "num_tokens": 1081790261.0, "step": 1366 }, { "epoch": 1.8656767366444784, "grad_norm": 0.09701799601316452, "learning_rate": 1.5930912588276863e-05, "loss": 0.4285, "num_tokens": 1082588616.0, "step": 1367 }, { "epoch": 1.867042157364738, "grad_norm": 0.09449412673711777, "learning_rate": 1.590596026745378e-05, "loss": 0.4366, "num_tokens": 1083400860.0, "step": 1368 }, { "epoch": 1.8684075780849976, "grad_norm": 0.11165119707584381, "learning_rate": 1.5881021166380547e-05, "loss": 0.4399, "num_tokens": 1084171509.0, "step": 1369 }, { "epoch": 1.869772998805257, "grad_norm": 0.0934496596455574, "learning_rate": 1.5856095339157357e-05, "loss": 0.4136, "num_tokens": 1084999465.0, "step": 1370 }, { "epoch": 1.8711384195255163, "grad_norm": 0.10673777014017105, "learning_rate": 1.58311828398556e-05, "loss": 0.4193, "num_tokens": 1085785619.0, "step": 1371 }, { "epoch": 1.8725038402457757, "grad_norm": 0.096223384141922, "learning_rate": 1.5806283722517747e-05, "loss": 0.4203, "num_tokens": 1086561771.0, "step": 1372 }, { "epoch": 1.873869260966035, "grad_norm": 0.10207728296518326, "learning_rate": 1.5781398041157245e-05, "loss": 0.4347, "num_tokens": 1087387505.0, "step": 1373 }, { "epoch": 1.8752346816862946, "grad_norm": 0.10143262892961502, "learning_rate": 1.5756525849758407e-05, "loss": 0.4312, "num_tokens": 1088192198.0, "step": 1374 }, { "epoch": 1.8766001024065542, "grad_norm": 0.0937027782201767, "learning_rate": 1.573166720227626e-05, "loss": 0.4264, "num_tokens": 1089050181.0, "step": 1375 }, { "epoch": 1.8779655231268135, "grad_norm": 0.10106776654720306, "learning_rate": 1.5706822152636472e-05, "loss": 0.4232, "num_tokens": 1089789085.0, "step": 1376 }, { "epoch": 1.8793309438470729, "grad_norm": 0.10197136551141739, "learning_rate": 1.56819907547352e-05, "loss": 0.4342, "num_tokens": 1090478922.0, "step": 1377 }, { "epoch": 1.8806963645673322, "grad_norm": 0.10028499364852905, "learning_rate": 1.5657173062438986e-05, "loss": 0.4052, "num_tokens": 1091281811.0, "step": 1378 }, { "epoch": 1.8820617852875916, "grad_norm": 0.12268929183483124, "learning_rate": 1.5632369129584645e-05, "loss": 0.41, "num_tokens": 1092062502.0, "step": 1379 }, { "epoch": 1.8834272060078512, "grad_norm": 0.10172151774168015, "learning_rate": 1.5607579009979145e-05, "loss": 0.4245, "num_tokens": 1092844030.0, "step": 1380 }, { "epoch": 1.8847926267281108, "grad_norm": 0.09577567130327225, "learning_rate": 1.5582802757399485e-05, "loss": 0.4483, "num_tokens": 1093703152.0, "step": 1381 }, { "epoch": 1.8861580474483701, "grad_norm": 0.09673799574375153, "learning_rate": 1.5558040425592584e-05, "loss": 0.412, "num_tokens": 1094496330.0, "step": 1382 }, { "epoch": 1.8875234681686295, "grad_norm": 0.09749764949083328, "learning_rate": 1.5533292068275173e-05, "loss": 0.4246, "num_tokens": 1095287277.0, "step": 1383 }, { "epoch": 1.8888888888888888, "grad_norm": 0.09495698660612106, "learning_rate": 1.5508557739133647e-05, "loss": 0.4026, "num_tokens": 1096043409.0, "step": 1384 }, { "epoch": 1.8902543096091482, "grad_norm": 0.09388016164302826, "learning_rate": 1.5483837491823975e-05, "loss": 0.4102, "num_tokens": 1096778795.0, "step": 1385 }, { "epoch": 1.8916197303294078, "grad_norm": 0.09931745380163193, "learning_rate": 1.545913137997159e-05, "loss": 0.4175, "num_tokens": 1097587021.0, "step": 1386 }, { "epoch": 1.8929851510496671, "grad_norm": 0.10708718001842499, "learning_rate": 1.5434439457171274e-05, "loss": 0.4151, "num_tokens": 1098390570.0, "step": 1387 }, { "epoch": 1.8943505717699267, "grad_norm": 0.08802707493305206, "learning_rate": 1.540976177698698e-05, "loss": 0.4237, "num_tokens": 1099202976.0, "step": 1388 }, { "epoch": 1.895715992490186, "grad_norm": 0.10204858332872391, "learning_rate": 1.5385098392951826e-05, "loss": 0.4212, "num_tokens": 1100046373.0, "step": 1389 }, { "epoch": 1.8970814132104454, "grad_norm": 0.10473590344190598, "learning_rate": 1.5360449358567864e-05, "loss": 0.3984, "num_tokens": 1100791999.0, "step": 1390 }, { "epoch": 1.8984468339307048, "grad_norm": 0.09322867542505264, "learning_rate": 1.5335814727306044e-05, "loss": 0.4179, "num_tokens": 1101480290.0, "step": 1391 }, { "epoch": 1.8998122546509644, "grad_norm": 0.10474120080471039, "learning_rate": 1.5311194552606067e-05, "loss": 0.4364, "num_tokens": 1102342832.0, "step": 1392 }, { "epoch": 1.9011776753712237, "grad_norm": 0.09817834198474884, "learning_rate": 1.528658888787628e-05, "loss": 0.4399, "num_tokens": 1103150856.0, "step": 1393 }, { "epoch": 1.9025430960914833, "grad_norm": 0.09619283676147461, "learning_rate": 1.5261997786493547e-05, "loss": 0.4301, "num_tokens": 1103992434.0, "step": 1394 }, { "epoch": 1.9039085168117427, "grad_norm": 0.10345003753900528, "learning_rate": 1.5237421301803136e-05, "loss": 0.4258, "num_tokens": 1104776554.0, "step": 1395 }, { "epoch": 1.905273937532002, "grad_norm": 0.10167142003774643, "learning_rate": 1.521285948711861e-05, "loss": 0.426, "num_tokens": 1105572947.0, "step": 1396 }, { "epoch": 1.9066393582522614, "grad_norm": 0.08946793526411057, "learning_rate": 1.5188312395721712e-05, "loss": 0.4241, "num_tokens": 1106400789.0, "step": 1397 }, { "epoch": 1.9080047789725207, "grad_norm": 0.10430753976106644, "learning_rate": 1.5163780080862247e-05, "loss": 0.4215, "num_tokens": 1107197584.0, "step": 1398 }, { "epoch": 1.9093701996927803, "grad_norm": 0.09312907606363297, "learning_rate": 1.5139262595757954e-05, "loss": 0.4083, "num_tokens": 1107991173.0, "step": 1399 }, { "epoch": 1.91073562041304, "grad_norm": 0.10167186707258224, "learning_rate": 1.5114759993594419e-05, "loss": 0.4341, "num_tokens": 1108789190.0, "step": 1400 }, { "epoch": 1.9121010411332993, "grad_norm": 0.10425346344709396, "learning_rate": 1.5090272327524935e-05, "loss": 0.4256, "num_tokens": 1109622837.0, "step": 1401 }, { "epoch": 1.9134664618535586, "grad_norm": 0.10787136107683182, "learning_rate": 1.5065799650670383e-05, "loss": 0.3961, "num_tokens": 1110366039.0, "step": 1402 }, { "epoch": 1.914831882573818, "grad_norm": 0.09407512098550797, "learning_rate": 1.5041342016119145e-05, "loss": 0.4267, "num_tokens": 1111197489.0, "step": 1403 }, { "epoch": 1.9161973032940773, "grad_norm": 0.11724904179573059, "learning_rate": 1.5016899476926962e-05, "loss": 0.4387, "num_tokens": 1111995722.0, "step": 1404 }, { "epoch": 1.917562724014337, "grad_norm": 0.11706633865833282, "learning_rate": 1.4992472086116828e-05, "loss": 0.4191, "num_tokens": 1112746863.0, "step": 1405 }, { "epoch": 1.9189281447345965, "grad_norm": 0.106146439909935, "learning_rate": 1.4968059896678898e-05, "loss": 0.4436, "num_tokens": 1113560910.0, "step": 1406 }, { "epoch": 1.9202935654548559, "grad_norm": 0.1055377945303917, "learning_rate": 1.4943662961570304e-05, "loss": 0.4529, "num_tokens": 1114384817.0, "step": 1407 }, { "epoch": 1.9216589861751152, "grad_norm": 0.11571165919303894, "learning_rate": 1.4919281333715136e-05, "loss": 0.4149, "num_tokens": 1115141182.0, "step": 1408 }, { "epoch": 1.9230244068953746, "grad_norm": 0.10280241817235947, "learning_rate": 1.489491506600424e-05, "loss": 0.4101, "num_tokens": 1115905625.0, "step": 1409 }, { "epoch": 1.924389827615634, "grad_norm": 0.0992712527513504, "learning_rate": 1.487056421129518e-05, "loss": 0.4173, "num_tokens": 1116731071.0, "step": 1410 }, { "epoch": 1.9257552483358935, "grad_norm": 0.10596533119678497, "learning_rate": 1.4846228822412046e-05, "loss": 0.4141, "num_tokens": 1117529232.0, "step": 1411 }, { "epoch": 1.927120669056153, "grad_norm": 0.10475833714008331, "learning_rate": 1.4821908952145409e-05, "loss": 0.42, "num_tokens": 1118303056.0, "step": 1412 }, { "epoch": 1.9284860897764124, "grad_norm": 0.09152897447347641, "learning_rate": 1.4797604653252142e-05, "loss": 0.4148, "num_tokens": 1119127789.0, "step": 1413 }, { "epoch": 1.9298515104966718, "grad_norm": 0.1027945801615715, "learning_rate": 1.477331597845538e-05, "loss": 0.4006, "num_tokens": 1119863146.0, "step": 1414 }, { "epoch": 1.9312169312169312, "grad_norm": 0.09834238141775131, "learning_rate": 1.4749042980444332e-05, "loss": 0.4352, "num_tokens": 1120643766.0, "step": 1415 }, { "epoch": 1.9325823519371905, "grad_norm": 0.10886780172586441, "learning_rate": 1.4724785711874208e-05, "loss": 0.4395, "num_tokens": 1121446634.0, "step": 1416 }, { "epoch": 1.93394777265745, "grad_norm": 0.10399835556745529, "learning_rate": 1.4700544225366114e-05, "loss": 0.4259, "num_tokens": 1122153577.0, "step": 1417 }, { "epoch": 1.9353131933777095, "grad_norm": 0.09844211488962173, "learning_rate": 1.4676318573506896e-05, "loss": 0.4325, "num_tokens": 1122982545.0, "step": 1418 }, { "epoch": 1.936678614097969, "grad_norm": 0.09757062792778015, "learning_rate": 1.465210880884906e-05, "loss": 0.4169, "num_tokens": 1123741571.0, "step": 1419 }, { "epoch": 1.9380440348182284, "grad_norm": 0.10535794496536255, "learning_rate": 1.4627914983910643e-05, "loss": 0.431, "num_tokens": 1124493814.0, "step": 1420 }, { "epoch": 1.9394094555384878, "grad_norm": 0.09897228330373764, "learning_rate": 1.4603737151175122e-05, "loss": 0.4294, "num_tokens": 1125295050.0, "step": 1421 }, { "epoch": 1.9407748762587471, "grad_norm": 0.09626423567533493, "learning_rate": 1.4579575363091252e-05, "loss": 0.4047, "num_tokens": 1126090909.0, "step": 1422 }, { "epoch": 1.9421402969790067, "grad_norm": 0.09447965025901794, "learning_rate": 1.4555429672073014e-05, "loss": 0.4386, "num_tokens": 1126965037.0, "step": 1423 }, { "epoch": 1.943505717699266, "grad_norm": 0.10495449602603912, "learning_rate": 1.4531300130499455e-05, "loss": 0.4374, "num_tokens": 1127809105.0, "step": 1424 }, { "epoch": 1.9448711384195256, "grad_norm": 0.09577663987874985, "learning_rate": 1.4507186790714581e-05, "loss": 0.4242, "num_tokens": 1128581251.0, "step": 1425 }, { "epoch": 1.946236559139785, "grad_norm": 0.09336096048355103, "learning_rate": 1.4483089705027267e-05, "loss": 0.4225, "num_tokens": 1129368507.0, "step": 1426 }, { "epoch": 1.9476019798600444, "grad_norm": 0.10478734970092773, "learning_rate": 1.4459008925711104e-05, "loss": 0.4221, "num_tokens": 1130109086.0, "step": 1427 }, { "epoch": 1.9489674005803037, "grad_norm": 0.10016389936208725, "learning_rate": 1.4434944505004355e-05, "loss": 0.4218, "num_tokens": 1130920279.0, "step": 1428 }, { "epoch": 1.9503328213005633, "grad_norm": 0.0936041921377182, "learning_rate": 1.4410896495109752e-05, "loss": 0.4129, "num_tokens": 1131734499.0, "step": 1429 }, { "epoch": 1.9516982420208226, "grad_norm": 0.09493396431207657, "learning_rate": 1.4386864948194447e-05, "loss": 0.4388, "num_tokens": 1132580298.0, "step": 1430 }, { "epoch": 1.9530636627410822, "grad_norm": 0.10224965214729309, "learning_rate": 1.4362849916389867e-05, "loss": 0.4249, "num_tokens": 1133365599.0, "step": 1431 }, { "epoch": 1.9544290834613416, "grad_norm": 0.10786716639995575, "learning_rate": 1.4338851451791631e-05, "loss": 0.4583, "num_tokens": 1134183758.0, "step": 1432 }, { "epoch": 1.955794504181601, "grad_norm": 0.0868016853928566, "learning_rate": 1.4314869606459408e-05, "loss": 0.4383, "num_tokens": 1135048421.0, "step": 1433 }, { "epoch": 1.9571599249018603, "grad_norm": 0.104495108127594, "learning_rate": 1.4290904432416813e-05, "loss": 0.4198, "num_tokens": 1135754552.0, "step": 1434 }, { "epoch": 1.9585253456221197, "grad_norm": 0.08921115100383759, "learning_rate": 1.42669559816513e-05, "loss": 0.4162, "num_tokens": 1136541054.0, "step": 1435 }, { "epoch": 1.9598907663423792, "grad_norm": 0.09871520102024078, "learning_rate": 1.4243024306114048e-05, "loss": 0.4027, "num_tokens": 1137292289.0, "step": 1436 }, { "epoch": 1.9612561870626388, "grad_norm": 0.09752541035413742, "learning_rate": 1.4219109457719839e-05, "loss": 0.4279, "num_tokens": 1138155806.0, "step": 1437 }, { "epoch": 1.9626216077828982, "grad_norm": 0.10382602363824844, "learning_rate": 1.4195211488346953e-05, "loss": 0.424, "num_tokens": 1138982292.0, "step": 1438 }, { "epoch": 1.9639870285031575, "grad_norm": 0.09439077973365784, "learning_rate": 1.4171330449837063e-05, "loss": 0.4263, "num_tokens": 1139809406.0, "step": 1439 }, { "epoch": 1.965352449223417, "grad_norm": 0.09448818117380142, "learning_rate": 1.4147466393995109e-05, "loss": 0.4267, "num_tokens": 1140647980.0, "step": 1440 }, { "epoch": 1.9667178699436763, "grad_norm": 0.09933383762836456, "learning_rate": 1.412361937258917e-05, "loss": 0.4289, "num_tokens": 1141389332.0, "step": 1441 }, { "epoch": 1.9680832906639358, "grad_norm": 0.10131408274173737, "learning_rate": 1.4099789437350424e-05, "loss": 0.4122, "num_tokens": 1142186348.0, "step": 1442 }, { "epoch": 1.9694487113841954, "grad_norm": 0.102419912815094, "learning_rate": 1.4075976639972928e-05, "loss": 0.4122, "num_tokens": 1142893343.0, "step": 1443 }, { "epoch": 1.9708141321044548, "grad_norm": 0.11110963672399521, "learning_rate": 1.4052181032113596e-05, "loss": 0.4414, "num_tokens": 1143720657.0, "step": 1444 }, { "epoch": 1.9721795528247141, "grad_norm": 0.09212847799062729, "learning_rate": 1.4028402665392027e-05, "loss": 0.416, "num_tokens": 1144589859.0, "step": 1445 }, { "epoch": 1.9735449735449735, "grad_norm": 0.09953609108924866, "learning_rate": 1.4004641591390455e-05, "loss": 0.4235, "num_tokens": 1145370654.0, "step": 1446 }, { "epoch": 1.9749103942652328, "grad_norm": 0.10138403624296188, "learning_rate": 1.3980897861653575e-05, "loss": 0.4039, "num_tokens": 1146127463.0, "step": 1447 }, { "epoch": 1.9762758149854924, "grad_norm": 0.09742390364408493, "learning_rate": 1.3957171527688466e-05, "loss": 0.4213, "num_tokens": 1146922781.0, "step": 1448 }, { "epoch": 1.9776412357057518, "grad_norm": 0.09856224805116653, "learning_rate": 1.393346264096445e-05, "loss": 0.4292, "num_tokens": 1147702644.0, "step": 1449 }, { "epoch": 1.9790066564260114, "grad_norm": 0.09630004316568375, "learning_rate": 1.3909771252913038e-05, "loss": 0.3938, "num_tokens": 1148484245.0, "step": 1450 }, { "epoch": 1.9803720771462707, "grad_norm": 0.09308762103319168, "learning_rate": 1.3886097414927754e-05, "loss": 0.4372, "num_tokens": 1149298374.0, "step": 1451 }, { "epoch": 1.98173749786653, "grad_norm": 0.10601536929607391, "learning_rate": 1.386244117836406e-05, "loss": 0.4348, "num_tokens": 1150101704.0, "step": 1452 }, { "epoch": 1.9831029185867894, "grad_norm": 0.10297173261642456, "learning_rate": 1.3838802594539228e-05, "loss": 0.4263, "num_tokens": 1150912652.0, "step": 1453 }, { "epoch": 1.984468339307049, "grad_norm": 0.08929338306188583, "learning_rate": 1.3815181714732251e-05, "loss": 0.4194, "num_tokens": 1151705083.0, "step": 1454 }, { "epoch": 1.9858337600273084, "grad_norm": 0.09480933845043182, "learning_rate": 1.3791578590183701e-05, "loss": 0.438, "num_tokens": 1152570247.0, "step": 1455 }, { "epoch": 1.987199180747568, "grad_norm": 0.10004927963018417, "learning_rate": 1.3767993272095643e-05, "loss": 0.4245, "num_tokens": 1153286126.0, "step": 1456 }, { "epoch": 1.9885646014678273, "grad_norm": 0.0996708944439888, "learning_rate": 1.3744425811631512e-05, "loss": 0.43, "num_tokens": 1154062224.0, "step": 1457 }, { "epoch": 1.9899300221880867, "grad_norm": 0.10280726104974747, "learning_rate": 1.3720876259915996e-05, "loss": 0.4283, "num_tokens": 1154896258.0, "step": 1458 }, { "epoch": 1.991295442908346, "grad_norm": 0.10244744271039963, "learning_rate": 1.3697344668034967e-05, "loss": 0.4241, "num_tokens": 1155692079.0, "step": 1459 }, { "epoch": 1.9926608636286056, "grad_norm": 0.0969838947057724, "learning_rate": 1.3673831087035286e-05, "loss": 0.4205, "num_tokens": 1156490029.0, "step": 1460 }, { "epoch": 1.994026284348865, "grad_norm": 0.09454359114170074, "learning_rate": 1.3650335567924779e-05, "loss": 0.4145, "num_tokens": 1157269822.0, "step": 1461 }, { "epoch": 1.9953917050691246, "grad_norm": 0.09734661877155304, "learning_rate": 1.3626858161672074e-05, "loss": 0.408, "num_tokens": 1158094009.0, "step": 1462 }, { "epoch": 1.996757125789384, "grad_norm": 0.10791545361280441, "learning_rate": 1.3603398919206523e-05, "loss": 0.4229, "num_tokens": 1158893953.0, "step": 1463 }, { "epoch": 1.9981225465096433, "grad_norm": 0.09216755628585815, "learning_rate": 1.3579957891418068e-05, "loss": 0.4254, "num_tokens": 1159659316.0, "step": 1464 }, { "epoch": 1.9994879672299026, "grad_norm": 0.10077279061079025, "learning_rate": 1.3556535129157138e-05, "loss": 0.4092, "num_tokens": 1160383886.0, "step": 1465 }, { "epoch": 2.0, "grad_norm": 0.10077279061079025, "learning_rate": 1.3533130683234515e-05, "loss": 0.4193, "num_tokens": 1160565010.0, "step": 1466 }, { "epoch": 2.0013654207202594, "grad_norm": 0.18715286254882812, "learning_rate": 1.3509744604421289e-05, "loss": 0.4111, "num_tokens": 1161377956.0, "step": 1467 }, { "epoch": 2.0027308414405187, "grad_norm": 0.10137300938367844, "learning_rate": 1.3486376943448686e-05, "loss": 0.3954, "num_tokens": 1162182485.0, "step": 1468 }, { "epoch": 2.004096262160778, "grad_norm": 0.10925179719924927, "learning_rate": 1.3463027751007975e-05, "loss": 0.396, "num_tokens": 1162949504.0, "step": 1469 }, { "epoch": 2.005461682881038, "grad_norm": 0.1124008446931839, "learning_rate": 1.3439697077750367e-05, "loss": 0.4035, "num_tokens": 1163748131.0, "step": 1470 }, { "epoch": 2.0068271036012972, "grad_norm": 0.10214317589998245, "learning_rate": 1.3416384974286895e-05, "loss": 0.4093, "num_tokens": 1164541124.0, "step": 1471 }, { "epoch": 2.0081925243215566, "grad_norm": 0.10282789170742035, "learning_rate": 1.3393091491188317e-05, "loss": 0.3835, "num_tokens": 1165317960.0, "step": 1472 }, { "epoch": 2.009557945041816, "grad_norm": 0.10354765504598618, "learning_rate": 1.3369816678984986e-05, "loss": 0.3955, "num_tokens": 1166135026.0, "step": 1473 }, { "epoch": 2.0109233657620753, "grad_norm": 0.11022363603115082, "learning_rate": 1.3346560588166764e-05, "loss": 0.4254, "num_tokens": 1167000518.0, "step": 1474 }, { "epoch": 2.0122887864823347, "grad_norm": 0.09984241425991058, "learning_rate": 1.3323323269182891e-05, "loss": 0.3839, "num_tokens": 1167778823.0, "step": 1475 }, { "epoch": 2.0136542072025945, "grad_norm": 0.10889684408903122, "learning_rate": 1.3300104772441897e-05, "loss": 0.4262, "num_tokens": 1168618434.0, "step": 1476 }, { "epoch": 2.015019627922854, "grad_norm": 0.09522828459739685, "learning_rate": 1.3276905148311481e-05, "loss": 0.4081, "num_tokens": 1169400262.0, "step": 1477 }, { "epoch": 2.016385048643113, "grad_norm": 0.10460998117923737, "learning_rate": 1.325372444711838e-05, "loss": 0.4129, "num_tokens": 1170260551.0, "step": 1478 }, { "epoch": 2.0177504693633725, "grad_norm": 0.10600446909666061, "learning_rate": 1.3230562719148292e-05, "loss": 0.3933, "num_tokens": 1170997996.0, "step": 1479 }, { "epoch": 2.019115890083632, "grad_norm": 0.09731242060661316, "learning_rate": 1.320742001464578e-05, "loss": 0.3971, "num_tokens": 1171705341.0, "step": 1480 }, { "epoch": 2.0204813108038913, "grad_norm": 0.10351541638374329, "learning_rate": 1.3184296383814114e-05, "loss": 0.3904, "num_tokens": 1172436562.0, "step": 1481 }, { "epoch": 2.021846731524151, "grad_norm": 0.10083338618278503, "learning_rate": 1.3161191876815194e-05, "loss": 0.4141, "num_tokens": 1173298027.0, "step": 1482 }, { "epoch": 2.0232121522444104, "grad_norm": 0.10191808640956879, "learning_rate": 1.3138106543769433e-05, "loss": 0.418, "num_tokens": 1174097622.0, "step": 1483 }, { "epoch": 2.02457757296467, "grad_norm": 0.10276789963245392, "learning_rate": 1.3115040434755656e-05, "loss": 0.3922, "num_tokens": 1174801389.0, "step": 1484 }, { "epoch": 2.025942993684929, "grad_norm": 0.10319360345602036, "learning_rate": 1.309199359981098e-05, "loss": 0.3985, "num_tokens": 1175644296.0, "step": 1485 }, { "epoch": 2.0273084144051885, "grad_norm": 0.09495635330677032, "learning_rate": 1.3068966088930713e-05, "loss": 0.4135, "num_tokens": 1176456618.0, "step": 1486 }, { "epoch": 2.028673835125448, "grad_norm": 0.10647090524435043, "learning_rate": 1.3045957952068242e-05, "loss": 0.4222, "num_tokens": 1177271381.0, "step": 1487 }, { "epoch": 2.0300392558457077, "grad_norm": 0.10156340897083282, "learning_rate": 1.3022969239134927e-05, "loss": 0.422, "num_tokens": 1178072478.0, "step": 1488 }, { "epoch": 2.031404676565967, "grad_norm": 0.10821613669395447, "learning_rate": 1.3000000000000006e-05, "loss": 0.4293, "num_tokens": 1178929939.0, "step": 1489 }, { "epoch": 2.0327700972862264, "grad_norm": 0.10066142678260803, "learning_rate": 1.297705028449044e-05, "loss": 0.3994, "num_tokens": 1179775652.0, "step": 1490 }, { "epoch": 2.0341355180064857, "grad_norm": 0.0914134830236435, "learning_rate": 1.2954120142390863e-05, "loss": 0.4013, "num_tokens": 1180579777.0, "step": 1491 }, { "epoch": 2.035500938726745, "grad_norm": 0.10027533024549484, "learning_rate": 1.293120962344344e-05, "loss": 0.4013, "num_tokens": 1181397450.0, "step": 1492 }, { "epoch": 2.0368663594470044, "grad_norm": 0.09783679246902466, "learning_rate": 1.2908318777347778e-05, "loss": 0.4261, "num_tokens": 1182211514.0, "step": 1493 }, { "epoch": 2.0382317801672643, "grad_norm": 0.10575064271688461, "learning_rate": 1.2885447653760804e-05, "loss": 0.3967, "num_tokens": 1182958855.0, "step": 1494 }, { "epoch": 2.0395972008875236, "grad_norm": 0.09552831947803497, "learning_rate": 1.2862596302296635e-05, "loss": 0.392, "num_tokens": 1183729426.0, "step": 1495 }, { "epoch": 2.040962621607783, "grad_norm": 0.09887399524450302, "learning_rate": 1.2839764772526527e-05, "loss": 0.3838, "num_tokens": 1184508441.0, "step": 1496 }, { "epoch": 2.0423280423280423, "grad_norm": 0.09516591578722, "learning_rate": 1.2816953113978738e-05, "loss": 0.387, "num_tokens": 1185245526.0, "step": 1497 }, { "epoch": 2.0436934630483017, "grad_norm": 0.10134809464216232, "learning_rate": 1.2794161376138403e-05, "loss": 0.4018, "num_tokens": 1186010738.0, "step": 1498 }, { "epoch": 2.045058883768561, "grad_norm": 0.09626221656799316, "learning_rate": 1.2771389608447452e-05, "loss": 0.397, "num_tokens": 1186795066.0, "step": 1499 }, { "epoch": 2.0464243044888204, "grad_norm": 0.09192902594804764, "learning_rate": 1.2748637860304486e-05, "loss": 0.4033, "num_tokens": 1187561632.0, "step": 1500 }, { "epoch": 2.04778972520908, "grad_norm": 0.1101919412612915, "learning_rate": 1.2725906181064693e-05, "loss": 0.4191, "num_tokens": 1188476085.0, "step": 1501 }, { "epoch": 2.0491551459293396, "grad_norm": 0.0893695279955864, "learning_rate": 1.2703194620039706e-05, "loss": 0.41, "num_tokens": 1189306269.0, "step": 1502 }, { "epoch": 2.050520566649599, "grad_norm": 0.10124479979276657, "learning_rate": 1.2680503226497533e-05, "loss": 0.4256, "num_tokens": 1190136857.0, "step": 1503 }, { "epoch": 2.0518859873698583, "grad_norm": 0.11104397475719452, "learning_rate": 1.2657832049662426e-05, "loss": 0.394, "num_tokens": 1190899604.0, "step": 1504 }, { "epoch": 2.0532514080901176, "grad_norm": 0.10146714001893997, "learning_rate": 1.2635181138714771e-05, "loss": 0.3977, "num_tokens": 1191671835.0, "step": 1505 }, { "epoch": 2.054616828810377, "grad_norm": 0.09603208303451538, "learning_rate": 1.261255054279102e-05, "loss": 0.3857, "num_tokens": 1192452454.0, "step": 1506 }, { "epoch": 2.055982249530637, "grad_norm": 0.09947937726974487, "learning_rate": 1.2589940310983519e-05, "loss": 0.3896, "num_tokens": 1193201420.0, "step": 1507 }, { "epoch": 2.057347670250896, "grad_norm": 0.11060870438814163, "learning_rate": 1.2567350492340466e-05, "loss": 0.4114, "num_tokens": 1193980512.0, "step": 1508 }, { "epoch": 2.0587130909711555, "grad_norm": 0.09556976705789566, "learning_rate": 1.2544781135865755e-05, "loss": 0.4129, "num_tokens": 1194783119.0, "step": 1509 }, { "epoch": 2.060078511691415, "grad_norm": 0.1040840595960617, "learning_rate": 1.2522232290518924e-05, "loss": 0.3936, "num_tokens": 1195551226.0, "step": 1510 }, { "epoch": 2.0614439324116742, "grad_norm": 0.09847235679626465, "learning_rate": 1.2499704005214983e-05, "loss": 0.4302, "num_tokens": 1196349889.0, "step": 1511 }, { "epoch": 2.0628093531319336, "grad_norm": 0.09874686598777771, "learning_rate": 1.2477196328824364e-05, "loss": 0.4211, "num_tokens": 1197304362.0, "step": 1512 }, { "epoch": 2.0641747738521934, "grad_norm": 0.09354843199253082, "learning_rate": 1.2454709310172763e-05, "loss": 0.3961, "num_tokens": 1198117058.0, "step": 1513 }, { "epoch": 2.0655401945724527, "grad_norm": 0.09405104070901871, "learning_rate": 1.2432242998041101e-05, "loss": 0.3909, "num_tokens": 1198910387.0, "step": 1514 }, { "epoch": 2.066905615292712, "grad_norm": 0.09199946373701096, "learning_rate": 1.2409797441165363e-05, "loss": 0.3958, "num_tokens": 1199701646.0, "step": 1515 }, { "epoch": 2.0682710360129715, "grad_norm": 0.09918805956840515, "learning_rate": 1.2387372688236499e-05, "loss": 0.4125, "num_tokens": 1200499867.0, "step": 1516 }, { "epoch": 2.069636456733231, "grad_norm": 0.09719354659318924, "learning_rate": 1.2364968787900344e-05, "loss": 0.4026, "num_tokens": 1201291264.0, "step": 1517 }, { "epoch": 2.07100187745349, "grad_norm": 0.10316963493824005, "learning_rate": 1.2342585788757492e-05, "loss": 0.4142, "num_tokens": 1202090256.0, "step": 1518 }, { "epoch": 2.07236729817375, "grad_norm": 0.09896045178174973, "learning_rate": 1.2320223739363204e-05, "loss": 0.3898, "num_tokens": 1202781901.0, "step": 1519 }, { "epoch": 2.0737327188940093, "grad_norm": 0.09228277206420898, "learning_rate": 1.229788268822728e-05, "loss": 0.4231, "num_tokens": 1203600558.0, "step": 1520 }, { "epoch": 2.0750981396142687, "grad_norm": 0.09974472224712372, "learning_rate": 1.2275562683813984e-05, "loss": 0.4103, "num_tokens": 1204405366.0, "step": 1521 }, { "epoch": 2.076463560334528, "grad_norm": 0.09792926907539368, "learning_rate": 1.2253263774541901e-05, "loss": 0.3933, "num_tokens": 1205196198.0, "step": 1522 }, { "epoch": 2.0778289810547874, "grad_norm": 0.09631495177745819, "learning_rate": 1.2230986008783891e-05, "loss": 0.4358, "num_tokens": 1205984404.0, "step": 1523 }, { "epoch": 2.0791944017750468, "grad_norm": 0.10573236644268036, "learning_rate": 1.2208729434866923e-05, "loss": 0.3896, "num_tokens": 1206712276.0, "step": 1524 }, { "epoch": 2.0805598224953066, "grad_norm": 0.09373833239078522, "learning_rate": 1.218649410107199e-05, "loss": 0.4176, "num_tokens": 1207531544.0, "step": 1525 }, { "epoch": 2.081925243215566, "grad_norm": 0.09758885204792023, "learning_rate": 1.216428005563401e-05, "loss": 0.3913, "num_tokens": 1208324596.0, "step": 1526 }, { "epoch": 2.0832906639358253, "grad_norm": 0.09951991587877274, "learning_rate": 1.2142087346741742e-05, "loss": 0.3967, "num_tokens": 1209076513.0, "step": 1527 }, { "epoch": 2.0846560846560847, "grad_norm": 0.09367062896490097, "learning_rate": 1.2119916022537644e-05, "loss": 0.4037, "num_tokens": 1209911090.0, "step": 1528 }, { "epoch": 2.086021505376344, "grad_norm": 0.09444288909435272, "learning_rate": 1.2097766131117781e-05, "loss": 0.4147, "num_tokens": 1210692356.0, "step": 1529 }, { "epoch": 2.0873869260966034, "grad_norm": 0.09753865003585815, "learning_rate": 1.207563772053173e-05, "loss": 0.4029, "num_tokens": 1211500251.0, "step": 1530 }, { "epoch": 2.0887523468168627, "grad_norm": 0.10317228734493256, "learning_rate": 1.2053530838782468e-05, "loss": 0.4027, "num_tokens": 1212266823.0, "step": 1531 }, { "epoch": 2.0901177675371225, "grad_norm": 0.09413281083106995, "learning_rate": 1.203144553382627e-05, "loss": 0.3971, "num_tokens": 1213006548.0, "step": 1532 }, { "epoch": 2.091483188257382, "grad_norm": 0.0980541855096817, "learning_rate": 1.2009381853572608e-05, "loss": 0.4156, "num_tokens": 1213854037.0, "step": 1533 }, { "epoch": 2.0928486089776412, "grad_norm": 0.09670227766036987, "learning_rate": 1.1987339845884036e-05, "loss": 0.3965, "num_tokens": 1214593373.0, "step": 1534 }, { "epoch": 2.0942140296979006, "grad_norm": 0.09512105584144592, "learning_rate": 1.1965319558576103e-05, "loss": 0.4185, "num_tokens": 1215378962.0, "step": 1535 }, { "epoch": 2.09557945041816, "grad_norm": 0.10186047852039337, "learning_rate": 1.1943321039417234e-05, "loss": 0.4107, "num_tokens": 1216151621.0, "step": 1536 }, { "epoch": 2.0969448711384193, "grad_norm": 0.09818754345178604, "learning_rate": 1.1921344336128633e-05, "loss": 0.4379, "num_tokens": 1216980686.0, "step": 1537 }, { "epoch": 2.098310291858679, "grad_norm": 0.09560073167085648, "learning_rate": 1.189938949638418e-05, "loss": 0.3983, "num_tokens": 1217770716.0, "step": 1538 }, { "epoch": 2.0996757125789385, "grad_norm": 0.09929785132408142, "learning_rate": 1.187745656781032e-05, "loss": 0.4071, "num_tokens": 1218606727.0, "step": 1539 }, { "epoch": 2.101041133299198, "grad_norm": 0.09448471665382385, "learning_rate": 1.1855545597985989e-05, "loss": 0.4094, "num_tokens": 1219348248.0, "step": 1540 }, { "epoch": 2.102406554019457, "grad_norm": 0.10522617399692535, "learning_rate": 1.183365663444247e-05, "loss": 0.4015, "num_tokens": 1220160533.0, "step": 1541 }, { "epoch": 2.1037719747397166, "grad_norm": 0.09850455075502396, "learning_rate": 1.1811789724663298e-05, "loss": 0.3983, "num_tokens": 1220948277.0, "step": 1542 }, { "epoch": 2.105137395459976, "grad_norm": 0.09858173131942749, "learning_rate": 1.1789944916084177e-05, "loss": 0.404, "num_tokens": 1221737008.0, "step": 1543 }, { "epoch": 2.1065028161802357, "grad_norm": 0.09535755962133408, "learning_rate": 1.1768122256092887e-05, "loss": 0.3943, "num_tokens": 1222490829.0, "step": 1544 }, { "epoch": 2.107868236900495, "grad_norm": 0.09770728647708893, "learning_rate": 1.174632179202913e-05, "loss": 0.3976, "num_tokens": 1223275614.0, "step": 1545 }, { "epoch": 2.1092336576207544, "grad_norm": 0.09400632232427597, "learning_rate": 1.1724543571184476e-05, "loss": 0.4161, "num_tokens": 1224113420.0, "step": 1546 }, { "epoch": 2.110599078341014, "grad_norm": 0.100069560110569, "learning_rate": 1.1702787640802238e-05, "loss": 0.3941, "num_tokens": 1224913037.0, "step": 1547 }, { "epoch": 2.111964499061273, "grad_norm": 0.09886769950389862, "learning_rate": 1.1681054048077378e-05, "loss": 0.4271, "num_tokens": 1225756916.0, "step": 1548 }, { "epoch": 2.1133299197815325, "grad_norm": 0.09174753725528717, "learning_rate": 1.1659342840156394e-05, "loss": 0.4179, "num_tokens": 1226539614.0, "step": 1549 }, { "epoch": 2.1146953405017923, "grad_norm": 0.10905648022890091, "learning_rate": 1.163765406413723e-05, "loss": 0.3973, "num_tokens": 1227293759.0, "step": 1550 }, { "epoch": 2.1160607612220517, "grad_norm": 0.09578517079353333, "learning_rate": 1.1615987767069173e-05, "loss": 0.4078, "num_tokens": 1228115787.0, "step": 1551 }, { "epoch": 2.117426181942311, "grad_norm": 0.10092911124229431, "learning_rate": 1.1594343995952728e-05, "loss": 0.423, "num_tokens": 1228847049.0, "step": 1552 }, { "epoch": 2.1187916026625704, "grad_norm": 0.1085011288523674, "learning_rate": 1.1572722797739572e-05, "loss": 0.4276, "num_tokens": 1229655500.0, "step": 1553 }, { "epoch": 2.1201570233828297, "grad_norm": 0.10378377139568329, "learning_rate": 1.1551124219332367e-05, "loss": 0.4008, "num_tokens": 1230409124.0, "step": 1554 }, { "epoch": 2.121522444103089, "grad_norm": 0.10518782585859299, "learning_rate": 1.1529548307584748e-05, "loss": 0.4278, "num_tokens": 1231159279.0, "step": 1555 }, { "epoch": 2.122887864823349, "grad_norm": 0.09774231910705566, "learning_rate": 1.150799510930114e-05, "loss": 0.426, "num_tokens": 1232015199.0, "step": 1556 }, { "epoch": 2.1242532855436083, "grad_norm": 0.10320428758859634, "learning_rate": 1.1486464671236738e-05, "loss": 0.3941, "num_tokens": 1232744565.0, "step": 1557 }, { "epoch": 2.1256187062638676, "grad_norm": 0.09619291871786118, "learning_rate": 1.1464957040097337e-05, "loss": 0.3829, "num_tokens": 1233520955.0, "step": 1558 }, { "epoch": 2.126984126984127, "grad_norm": 0.09326492249965668, "learning_rate": 1.144347226253927e-05, "loss": 0.3737, "num_tokens": 1234228944.0, "step": 1559 }, { "epoch": 2.1283495477043863, "grad_norm": 0.10086032748222351, "learning_rate": 1.1422010385169262e-05, "loss": 0.4032, "num_tokens": 1235023266.0, "step": 1560 }, { "epoch": 2.1297149684246457, "grad_norm": 0.09269057214260101, "learning_rate": 1.1400571454544412e-05, "loss": 0.4336, "num_tokens": 1235882722.0, "step": 1561 }, { "epoch": 2.1310803891449055, "grad_norm": 0.09564582258462906, "learning_rate": 1.1379155517172e-05, "loss": 0.4017, "num_tokens": 1236660063.0, "step": 1562 }, { "epoch": 2.132445809865165, "grad_norm": 0.10141199082136154, "learning_rate": 1.1357762619509445e-05, "loss": 0.4005, "num_tokens": 1237456857.0, "step": 1563 }, { "epoch": 2.133811230585424, "grad_norm": 0.1050870344042778, "learning_rate": 1.1336392807964185e-05, "loss": 0.3825, "num_tokens": 1238189659.0, "step": 1564 }, { "epoch": 2.1351766513056836, "grad_norm": 0.0982542634010315, "learning_rate": 1.1315046128893571e-05, "loss": 0.4032, "num_tokens": 1238942355.0, "step": 1565 }, { "epoch": 2.136542072025943, "grad_norm": 0.09714022278785706, "learning_rate": 1.1293722628604782e-05, "loss": 0.3939, "num_tokens": 1239785025.0, "step": 1566 }, { "epoch": 2.1379074927462023, "grad_norm": 0.10043782740831375, "learning_rate": 1.1272422353354706e-05, "loss": 0.395, "num_tokens": 1240514295.0, "step": 1567 }, { "epoch": 2.1392729134664616, "grad_norm": 0.10150834918022156, "learning_rate": 1.1251145349349855e-05, "loss": 0.4106, "num_tokens": 1241336593.0, "step": 1568 }, { "epoch": 2.1406383341867214, "grad_norm": 0.10023326426744461, "learning_rate": 1.1229891662746248e-05, "loss": 0.3999, "num_tokens": 1242211551.0, "step": 1569 }, { "epoch": 2.142003754906981, "grad_norm": 0.09345634281635284, "learning_rate": 1.1208661339649348e-05, "loss": 0.3979, "num_tokens": 1242913890.0, "step": 1570 }, { "epoch": 2.14336917562724, "grad_norm": 0.10114452242851257, "learning_rate": 1.1187454426113917e-05, "loss": 0.4211, "num_tokens": 1243810208.0, "step": 1571 }, { "epoch": 2.1447345963474995, "grad_norm": 0.09386526793241501, "learning_rate": 1.1166270968143923e-05, "loss": 0.4173, "num_tokens": 1244601601.0, "step": 1572 }, { "epoch": 2.146100017067759, "grad_norm": 0.09636539965867996, "learning_rate": 1.1145111011692471e-05, "loss": 0.3838, "num_tokens": 1245345849.0, "step": 1573 }, { "epoch": 2.1474654377880182, "grad_norm": 0.1019698828458786, "learning_rate": 1.1123974602661685e-05, "loss": 0.419, "num_tokens": 1246156817.0, "step": 1574 }, { "epoch": 2.148830858508278, "grad_norm": 0.1017390638589859, "learning_rate": 1.1102861786902595e-05, "loss": 0.4024, "num_tokens": 1246964727.0, "step": 1575 }, { "epoch": 2.1501962792285374, "grad_norm": 0.09533166885375977, "learning_rate": 1.108177261021507e-05, "loss": 0.3946, "num_tokens": 1247775569.0, "step": 1576 }, { "epoch": 2.1515616999487968, "grad_norm": 0.09515392780303955, "learning_rate": 1.1060707118347662e-05, "loss": 0.4217, "num_tokens": 1248599657.0, "step": 1577 }, { "epoch": 2.152927120669056, "grad_norm": 0.10441562533378601, "learning_rate": 1.103966535699759e-05, "loss": 0.408, "num_tokens": 1249392768.0, "step": 1578 }, { "epoch": 2.1542925413893155, "grad_norm": 0.10444799810647964, "learning_rate": 1.1018647371810562e-05, "loss": 0.4157, "num_tokens": 1250194838.0, "step": 1579 }, { "epoch": 2.155657962109575, "grad_norm": 0.10099997371435165, "learning_rate": 1.0997653208380723e-05, "loss": 0.4109, "num_tokens": 1251072381.0, "step": 1580 }, { "epoch": 2.1570233828298346, "grad_norm": 0.09785196930170059, "learning_rate": 1.0976682912250537e-05, "loss": 0.3813, "num_tokens": 1251828392.0, "step": 1581 }, { "epoch": 2.158388803550094, "grad_norm": 0.09404036402702332, "learning_rate": 1.0955736528910692e-05, "loss": 0.3892, "num_tokens": 1252549978.0, "step": 1582 }, { "epoch": 2.1597542242703534, "grad_norm": 0.09750542789697647, "learning_rate": 1.0934814103800004e-05, "loss": 0.4186, "num_tokens": 1253372571.0, "step": 1583 }, { "epoch": 2.1611196449906127, "grad_norm": 0.09923666715621948, "learning_rate": 1.0913915682305319e-05, "loss": 0.4203, "num_tokens": 1254250220.0, "step": 1584 }, { "epoch": 2.162485065710872, "grad_norm": 0.10125962644815445, "learning_rate": 1.0893041309761403e-05, "loss": 0.3915, "num_tokens": 1254999065.0, "step": 1585 }, { "epoch": 2.1638504864311314, "grad_norm": 0.0976707860827446, "learning_rate": 1.0872191031450861e-05, "loss": 0.3917, "num_tokens": 1255778563.0, "step": 1586 }, { "epoch": 2.1652159071513912, "grad_norm": 0.09801660478115082, "learning_rate": 1.0851364892604032e-05, "loss": 0.4195, "num_tokens": 1256500330.0, "step": 1587 }, { "epoch": 2.1665813278716506, "grad_norm": 0.10114692151546478, "learning_rate": 1.0830562938398893e-05, "loss": 0.4071, "num_tokens": 1257358876.0, "step": 1588 }, { "epoch": 2.16794674859191, "grad_norm": 0.10210877656936646, "learning_rate": 1.080978521396094e-05, "loss": 0.4096, "num_tokens": 1258113679.0, "step": 1589 }, { "epoch": 2.1693121693121693, "grad_norm": 0.0998600497841835, "learning_rate": 1.0789031764363114e-05, "loss": 0.4054, "num_tokens": 1258887915.0, "step": 1590 }, { "epoch": 2.1706775900324287, "grad_norm": 0.10441919416189194, "learning_rate": 1.0768302634625716e-05, "loss": 0.4405, "num_tokens": 1259728618.0, "step": 1591 }, { "epoch": 2.172043010752688, "grad_norm": 0.09311792999505997, "learning_rate": 1.0747597869716265e-05, "loss": 0.4093, "num_tokens": 1260546418.0, "step": 1592 }, { "epoch": 2.1734084314729474, "grad_norm": 0.09695204347372055, "learning_rate": 1.072691751454944e-05, "loss": 0.4289, "num_tokens": 1261372269.0, "step": 1593 }, { "epoch": 2.174773852193207, "grad_norm": 0.10256882756948471, "learning_rate": 1.0706261613986961e-05, "loss": 0.3978, "num_tokens": 1262097079.0, "step": 1594 }, { "epoch": 2.1761392729134665, "grad_norm": 0.10069379955530167, "learning_rate": 1.0685630212837503e-05, "loss": 0.3889, "num_tokens": 1262868819.0, "step": 1595 }, { "epoch": 2.177504693633726, "grad_norm": 0.09579601138830185, "learning_rate": 1.0665023355856596e-05, "loss": 0.4153, "num_tokens": 1263651092.0, "step": 1596 }, { "epoch": 2.1788701143539853, "grad_norm": 0.10160060971975327, "learning_rate": 1.064444108774652e-05, "loss": 0.4179, "num_tokens": 1264404731.0, "step": 1597 }, { "epoch": 2.1802355350742446, "grad_norm": 0.0991467759013176, "learning_rate": 1.0623883453156218e-05, "loss": 0.4364, "num_tokens": 1265270335.0, "step": 1598 }, { "epoch": 2.1816009557945044, "grad_norm": 0.09742279350757599, "learning_rate": 1.0603350496681197e-05, "loss": 0.3797, "num_tokens": 1266096089.0, "step": 1599 }, { "epoch": 2.1829663765147638, "grad_norm": 0.09754429757595062, "learning_rate": 1.0582842262863442e-05, "loss": 0.4078, "num_tokens": 1266883383.0, "step": 1600 }, { "epoch": 2.184331797235023, "grad_norm": 0.09237004071474075, "learning_rate": 1.0562358796191274e-05, "loss": 0.4034, "num_tokens": 1267665521.0, "step": 1601 }, { "epoch": 2.1856972179552825, "grad_norm": 0.0905998945236206, "learning_rate": 1.0541900141099319e-05, "loss": 0.3802, "num_tokens": 1268426436.0, "step": 1602 }, { "epoch": 2.187062638675542, "grad_norm": 0.09777001291513443, "learning_rate": 1.052146634196836e-05, "loss": 0.383, "num_tokens": 1269159212.0, "step": 1603 }, { "epoch": 2.188428059395801, "grad_norm": 0.10268130898475647, "learning_rate": 1.0501057443125277e-05, "loss": 0.3912, "num_tokens": 1269970955.0, "step": 1604 }, { "epoch": 2.1897934801160606, "grad_norm": 0.09783319383859634, "learning_rate": 1.0480673488842923e-05, "loss": 0.3997, "num_tokens": 1270735994.0, "step": 1605 }, { "epoch": 2.1911589008363204, "grad_norm": 0.10340055078268051, "learning_rate": 1.0460314523340045e-05, "loss": 0.389, "num_tokens": 1271505147.0, "step": 1606 }, { "epoch": 2.1925243215565797, "grad_norm": 0.08844783157110214, "learning_rate": 1.0439980590781161e-05, "loss": 0.4113, "num_tokens": 1272347644.0, "step": 1607 }, { "epoch": 2.193889742276839, "grad_norm": 0.09120872616767883, "learning_rate": 1.0419671735276518e-05, "loss": 0.3985, "num_tokens": 1273111001.0, "step": 1608 }, { "epoch": 2.1952551629970984, "grad_norm": 0.0976913645863533, "learning_rate": 1.0399388000881942e-05, "loss": 0.417, "num_tokens": 1273905616.0, "step": 1609 }, { "epoch": 2.196620583717358, "grad_norm": 0.09868104755878448, "learning_rate": 1.0379129431598775e-05, "loss": 0.3923, "num_tokens": 1274679387.0, "step": 1610 }, { "epoch": 2.197986004437617, "grad_norm": 0.09754499793052673, "learning_rate": 1.0358896071373758e-05, "loss": 0.4251, "num_tokens": 1275521234.0, "step": 1611 }, { "epoch": 2.199351425157877, "grad_norm": 0.10130767524242401, "learning_rate": 1.0338687964098953e-05, "loss": 0.3919, "num_tokens": 1276273212.0, "step": 1612 }, { "epoch": 2.2007168458781363, "grad_norm": 0.09586264938116074, "learning_rate": 1.0318505153611642e-05, "loss": 0.4007, "num_tokens": 1276998518.0, "step": 1613 }, { "epoch": 2.2020822665983957, "grad_norm": 0.09694807976484299, "learning_rate": 1.0298347683694224e-05, "loss": 0.4001, "num_tokens": 1277792468.0, "step": 1614 }, { "epoch": 2.203447687318655, "grad_norm": 0.0962553396821022, "learning_rate": 1.0278215598074133e-05, "loss": 0.426, "num_tokens": 1278599468.0, "step": 1615 }, { "epoch": 2.2048131080389144, "grad_norm": 0.10469597578048706, "learning_rate": 1.025810894042373e-05, "loss": 0.4096, "num_tokens": 1279392460.0, "step": 1616 }, { "epoch": 2.2061785287591738, "grad_norm": 0.09032826870679855, "learning_rate": 1.0238027754360238e-05, "loss": 0.4056, "num_tokens": 1280195113.0, "step": 1617 }, { "epoch": 2.2075439494794336, "grad_norm": 0.09862501919269562, "learning_rate": 1.0217972083445591e-05, "loss": 0.4102, "num_tokens": 1281009568.0, "step": 1618 }, { "epoch": 2.208909370199693, "grad_norm": 0.10134632140398026, "learning_rate": 1.0197941971186393e-05, "loss": 0.4126, "num_tokens": 1281778806.0, "step": 1619 }, { "epoch": 2.2102747909199523, "grad_norm": 0.09289058297872543, "learning_rate": 1.0177937461033798e-05, "loss": 0.407, "num_tokens": 1282611524.0, "step": 1620 }, { "epoch": 2.2116402116402116, "grad_norm": 0.09921691566705704, "learning_rate": 1.0157958596383433e-05, "loss": 0.3892, "num_tokens": 1283340447.0, "step": 1621 }, { "epoch": 2.213005632360471, "grad_norm": 0.09506433457136154, "learning_rate": 1.0138005420575281e-05, "loss": 0.425, "num_tokens": 1284151754.0, "step": 1622 }, { "epoch": 2.2143710530807303, "grad_norm": 0.0955457091331482, "learning_rate": 1.01180779768936e-05, "loss": 0.4258, "num_tokens": 1285004244.0, "step": 1623 }, { "epoch": 2.21573647380099, "grad_norm": 0.09424754977226257, "learning_rate": 1.0098176308566817e-05, "loss": 0.4064, "num_tokens": 1285760951.0, "step": 1624 }, { "epoch": 2.2171018945212495, "grad_norm": 0.09152563661336899, "learning_rate": 1.0078300458767467e-05, "loss": 0.3837, "num_tokens": 1286556350.0, "step": 1625 }, { "epoch": 2.218467315241509, "grad_norm": 0.09262444078922272, "learning_rate": 1.0058450470612068e-05, "loss": 0.4026, "num_tokens": 1287319034.0, "step": 1626 }, { "epoch": 2.219832735961768, "grad_norm": 0.09885437041521072, "learning_rate": 1.003862638716103e-05, "loss": 0.4128, "num_tokens": 1288110356.0, "step": 1627 }, { "epoch": 2.2211981566820276, "grad_norm": 0.09445696324110031, "learning_rate": 1.0018828251418576e-05, "loss": 0.3972, "num_tokens": 1288908202.0, "step": 1628 }, { "epoch": 2.222563577402287, "grad_norm": 0.0938974916934967, "learning_rate": 9.999056106332635e-06, "loss": 0.4188, "num_tokens": 1289765305.0, "step": 1629 }, { "epoch": 2.2239289981225463, "grad_norm": 0.0939706414937973, "learning_rate": 9.979309994794757e-06, "loss": 0.4037, "num_tokens": 1290581671.0, "step": 1630 }, { "epoch": 2.225294418842806, "grad_norm": 0.09601099044084549, "learning_rate": 9.959589959640021e-06, "loss": 0.4165, "num_tokens": 1291330271.0, "step": 1631 }, { "epoch": 2.2266598395630655, "grad_norm": 0.10425194352865219, "learning_rate": 9.939896043646932e-06, "loss": 0.4259, "num_tokens": 1292135959.0, "step": 1632 }, { "epoch": 2.228025260283325, "grad_norm": 0.09601804614067078, "learning_rate": 9.920228289537337e-06, "loss": 0.4168, "num_tokens": 1293014531.0, "step": 1633 }, { "epoch": 2.229390681003584, "grad_norm": 0.092473603785038, "learning_rate": 9.90058673997634e-06, "loss": 0.4218, "num_tokens": 1293863410.0, "step": 1634 }, { "epoch": 2.2307561017238435, "grad_norm": 0.09709075093269348, "learning_rate": 9.880971437572194e-06, "loss": 0.3937, "num_tokens": 1294653550.0, "step": 1635 }, { "epoch": 2.232121522444103, "grad_norm": 0.093082956969738, "learning_rate": 9.861382424876196e-06, "loss": 0.3965, "num_tokens": 1295400034.0, "step": 1636 }, { "epoch": 2.2334869431643627, "grad_norm": 0.09785421192646027, "learning_rate": 9.841819744382634e-06, "loss": 0.4111, "num_tokens": 1296147375.0, "step": 1637 }, { "epoch": 2.234852363884622, "grad_norm": 0.09823501855134964, "learning_rate": 9.822283438528677e-06, "loss": 0.3797, "num_tokens": 1296839819.0, "step": 1638 }, { "epoch": 2.2362177846048814, "grad_norm": 0.09553851932287216, "learning_rate": 9.802773549694267e-06, "loss": 0.4138, "num_tokens": 1297638888.0, "step": 1639 }, { "epoch": 2.2375832053251408, "grad_norm": 0.11889755725860596, "learning_rate": 9.783290120202043e-06, "loss": 0.407, "num_tokens": 1298370118.0, "step": 1640 }, { "epoch": 2.2389486260454, "grad_norm": 0.10373746603727341, "learning_rate": 9.763833192317248e-06, "loss": 0.4038, "num_tokens": 1299173013.0, "step": 1641 }, { "epoch": 2.2403140467656595, "grad_norm": 0.10018904507160187, "learning_rate": 9.744402808247635e-06, "loss": 0.4137, "num_tokens": 1300039414.0, "step": 1642 }, { "epoch": 2.2416794674859193, "grad_norm": 0.09447652846574783, "learning_rate": 9.724999010143372e-06, "loss": 0.4018, "num_tokens": 1300829563.0, "step": 1643 }, { "epoch": 2.2430448882061786, "grad_norm": 0.10066372901201248, "learning_rate": 9.70562184009696e-06, "loss": 0.4186, "num_tokens": 1301608766.0, "step": 1644 }, { "epoch": 2.244410308926438, "grad_norm": 0.09525424242019653, "learning_rate": 9.686271340143132e-06, "loss": 0.4035, "num_tokens": 1302370100.0, "step": 1645 }, { "epoch": 2.2457757296466974, "grad_norm": 0.09915542602539062, "learning_rate": 9.666947552258764e-06, "loss": 0.3838, "num_tokens": 1303237479.0, "step": 1646 }, { "epoch": 2.2471411503669567, "grad_norm": 0.09861354529857635, "learning_rate": 9.647650518362799e-06, "loss": 0.4251, "num_tokens": 1304079944.0, "step": 1647 }, { "epoch": 2.248506571087216, "grad_norm": 0.09522774815559387, "learning_rate": 9.628380280316124e-06, "loss": 0.4101, "num_tokens": 1304851869.0, "step": 1648 }, { "epoch": 2.249871991807476, "grad_norm": 0.09866340458393097, "learning_rate": 9.609136879921506e-06, "loss": 0.4075, "num_tokens": 1305663248.0, "step": 1649 }, { "epoch": 2.2512374125277352, "grad_norm": 0.0976831465959549, "learning_rate": 9.589920358923492e-06, "loss": 0.4284, "num_tokens": 1306472256.0, "step": 1650 }, { "epoch": 2.2526028332479946, "grad_norm": 0.09768351912498474, "learning_rate": 9.570730759008331e-06, "loss": 0.3919, "num_tokens": 1307281646.0, "step": 1651 }, { "epoch": 2.253968253968254, "grad_norm": 0.09716468304395676, "learning_rate": 9.551568121803859e-06, "loss": 0.3929, "num_tokens": 1308089995.0, "step": 1652 }, { "epoch": 2.2553336746885133, "grad_norm": 0.10004810988903046, "learning_rate": 9.53243248887943e-06, "loss": 0.4066, "num_tokens": 1308832199.0, "step": 1653 }, { "epoch": 2.2566990954087727, "grad_norm": 0.09702539443969727, "learning_rate": 9.5133239017458e-06, "loss": 0.4102, "num_tokens": 1309604520.0, "step": 1654 }, { "epoch": 2.258064516129032, "grad_norm": 0.09941393882036209, "learning_rate": 9.494242401855081e-06, "loss": 0.42, "num_tokens": 1310433095.0, "step": 1655 }, { "epoch": 2.259429936849292, "grad_norm": 0.09873232990503311, "learning_rate": 9.475188030600608e-06, "loss": 0.4257, "num_tokens": 1311197541.0, "step": 1656 }, { "epoch": 2.260795357569551, "grad_norm": 0.09672843664884567, "learning_rate": 9.456160829316872e-06, "loss": 0.385, "num_tokens": 1311989075.0, "step": 1657 }, { "epoch": 2.2621607782898105, "grad_norm": 0.09278816729784012, "learning_rate": 9.437160839279422e-06, "loss": 0.3981, "num_tokens": 1312776150.0, "step": 1658 }, { "epoch": 2.26352619901007, "grad_norm": 0.09568900614976883, "learning_rate": 9.418188101704779e-06, "loss": 0.4025, "num_tokens": 1313537368.0, "step": 1659 }, { "epoch": 2.2648916197303293, "grad_norm": 0.09646489471197128, "learning_rate": 9.399242657750346e-06, "loss": 0.405, "num_tokens": 1314334621.0, "step": 1660 }, { "epoch": 2.266257040450589, "grad_norm": 0.10129553079605103, "learning_rate": 9.380324548514316e-06, "loss": 0.402, "num_tokens": 1315175054.0, "step": 1661 }, { "epoch": 2.2676224611708484, "grad_norm": 0.09621522575616837, "learning_rate": 9.361433815035588e-06, "loss": 0.3932, "num_tokens": 1315941092.0, "step": 1662 }, { "epoch": 2.268987881891108, "grad_norm": 0.09434306621551514, "learning_rate": 9.342570498293668e-06, "loss": 0.3856, "num_tokens": 1316719695.0, "step": 1663 }, { "epoch": 2.270353302611367, "grad_norm": 0.09932339936494827, "learning_rate": 9.323734639208607e-06, "loss": 0.3856, "num_tokens": 1317408352.0, "step": 1664 }, { "epoch": 2.2717187233316265, "grad_norm": 0.09530244022607803, "learning_rate": 9.304926278640865e-06, "loss": 0.4013, "num_tokens": 1318242125.0, "step": 1665 }, { "epoch": 2.273084144051886, "grad_norm": 0.09522101283073425, "learning_rate": 9.286145457391265e-06, "loss": 0.4064, "num_tokens": 1319031497.0, "step": 1666 }, { "epoch": 2.274449564772145, "grad_norm": 0.09409365803003311, "learning_rate": 9.26739221620088e-06, "loss": 0.3945, "num_tokens": 1319864254.0, "step": 1667 }, { "epoch": 2.275814985492405, "grad_norm": 0.09421293437480927, "learning_rate": 9.248666595750976e-06, "loss": 0.3935, "num_tokens": 1320645796.0, "step": 1668 }, { "epoch": 2.2771804062126644, "grad_norm": 0.09486691653728485, "learning_rate": 9.229968636662878e-06, "loss": 0.4155, "num_tokens": 1321551414.0, "step": 1669 }, { "epoch": 2.2785458269329237, "grad_norm": 0.10786010324954987, "learning_rate": 9.21129837949792e-06, "loss": 0.4026, "num_tokens": 1322299632.0, "step": 1670 }, { "epoch": 2.279911247653183, "grad_norm": 0.09283756464719772, "learning_rate": 9.192655864757318e-06, "loss": 0.3867, "num_tokens": 1323086753.0, "step": 1671 }, { "epoch": 2.2812766683734425, "grad_norm": 0.0939759761095047, "learning_rate": 9.174041132882142e-06, "loss": 0.4037, "num_tokens": 1323797011.0, "step": 1672 }, { "epoch": 2.282642089093702, "grad_norm": 0.10076319426298141, "learning_rate": 9.155454224253172e-06, "loss": 0.4299, "num_tokens": 1324680335.0, "step": 1673 }, { "epoch": 2.2840075098139616, "grad_norm": 0.09774956107139587, "learning_rate": 9.136895179190832e-06, "loss": 0.4127, "num_tokens": 1325491488.0, "step": 1674 }, { "epoch": 2.285372930534221, "grad_norm": 0.09333214163780212, "learning_rate": 9.118364037955105e-06, "loss": 0.4074, "num_tokens": 1326235879.0, "step": 1675 }, { "epoch": 2.2867383512544803, "grad_norm": 0.09747271239757538, "learning_rate": 9.099860840745443e-06, "loss": 0.3956, "num_tokens": 1327108376.0, "step": 1676 }, { "epoch": 2.2881037719747397, "grad_norm": 0.0900663435459137, "learning_rate": 9.081385627700679e-06, "loss": 0.4146, "num_tokens": 1327864471.0, "step": 1677 }, { "epoch": 2.289469192694999, "grad_norm": 0.09995405375957489, "learning_rate": 9.062938438898936e-06, "loss": 0.4041, "num_tokens": 1328657954.0, "step": 1678 }, { "epoch": 2.2908346134152584, "grad_norm": 0.09262382984161377, "learning_rate": 9.04451931435755e-06, "loss": 0.4106, "num_tokens": 1329427349.0, "step": 1679 }, { "epoch": 2.2922000341355178, "grad_norm": 0.09872937947511673, "learning_rate": 9.02612829403297e-06, "loss": 0.4139, "num_tokens": 1330249911.0, "step": 1680 }, { "epoch": 2.2935654548557776, "grad_norm": 0.09606411308050156, "learning_rate": 9.007765417820693e-06, "loss": 0.3999, "num_tokens": 1331098066.0, "step": 1681 }, { "epoch": 2.294930875576037, "grad_norm": 0.09903182834386826, "learning_rate": 8.989430725555157e-06, "loss": 0.4215, "num_tokens": 1331899449.0, "step": 1682 }, { "epoch": 2.2962962962962963, "grad_norm": 0.09553909301757812, "learning_rate": 8.971124257009644e-06, "loss": 0.3984, "num_tokens": 1332668878.0, "step": 1683 }, { "epoch": 2.2976617170165556, "grad_norm": 0.09784558415412903, "learning_rate": 8.952846051896229e-06, "loss": 0.4007, "num_tokens": 1333492846.0, "step": 1684 }, { "epoch": 2.299027137736815, "grad_norm": 0.09710320085287094, "learning_rate": 8.934596149865675e-06, "loss": 0.3967, "num_tokens": 1334350439.0, "step": 1685 }, { "epoch": 2.300392558457075, "grad_norm": 0.09889866411685944, "learning_rate": 8.916374590507344e-06, "loss": 0.4027, "num_tokens": 1335094810.0, "step": 1686 }, { "epoch": 2.301757979177334, "grad_norm": 0.10243453085422516, "learning_rate": 8.898181413349107e-06, "loss": 0.4001, "num_tokens": 1335890238.0, "step": 1687 }, { "epoch": 2.3031233998975935, "grad_norm": 0.09387733787298203, "learning_rate": 8.880016657857278e-06, "loss": 0.4035, "num_tokens": 1336672281.0, "step": 1688 }, { "epoch": 2.304488820617853, "grad_norm": 0.09659291058778763, "learning_rate": 8.861880363436507e-06, "loss": 0.4223, "num_tokens": 1337526505.0, "step": 1689 }, { "epoch": 2.3058542413381122, "grad_norm": 0.09423097968101501, "learning_rate": 8.843772569429712e-06, "loss": 0.4141, "num_tokens": 1338313843.0, "step": 1690 }, { "epoch": 2.3072196620583716, "grad_norm": 0.10207855701446533, "learning_rate": 8.825693315117974e-06, "loss": 0.4188, "num_tokens": 1339134220.0, "step": 1691 }, { "epoch": 2.308585082778631, "grad_norm": 0.10592233389616013, "learning_rate": 8.807642639720475e-06, "loss": 0.4005, "num_tokens": 1339905417.0, "step": 1692 }, { "epoch": 2.3099505034988908, "grad_norm": 0.0970434620976448, "learning_rate": 8.789620582394385e-06, "loss": 0.4019, "num_tokens": 1340747998.0, "step": 1693 }, { "epoch": 2.31131592421915, "grad_norm": 0.09269697219133377, "learning_rate": 8.771627182234824e-06, "loss": 0.3876, "num_tokens": 1341502591.0, "step": 1694 }, { "epoch": 2.3126813449394095, "grad_norm": 0.098619244992733, "learning_rate": 8.753662478274704e-06, "loss": 0.4081, "num_tokens": 1342274617.0, "step": 1695 }, { "epoch": 2.314046765659669, "grad_norm": 0.09772706776857376, "learning_rate": 8.735726509484722e-06, "loss": 0.4052, "num_tokens": 1343027036.0, "step": 1696 }, { "epoch": 2.315412186379928, "grad_norm": 0.09785567224025726, "learning_rate": 8.717819314773213e-06, "loss": 0.3915, "num_tokens": 1343869808.0, "step": 1697 }, { "epoch": 2.316777607100188, "grad_norm": 0.09396972507238388, "learning_rate": 8.699940932986121e-06, "loss": 0.4, "num_tokens": 1344656396.0, "step": 1698 }, { "epoch": 2.3181430278204473, "grad_norm": 0.09574185311794281, "learning_rate": 8.682091402906871e-06, "loss": 0.3846, "num_tokens": 1345433620.0, "step": 1699 }, { "epoch": 2.3195084485407067, "grad_norm": 0.09349527209997177, "learning_rate": 8.66427076325629e-06, "loss": 0.3964, "num_tokens": 1346138604.0, "step": 1700 }, { "epoch": 2.320873869260966, "grad_norm": 0.09816993027925491, "learning_rate": 8.646479052692541e-06, "loss": 0.4047, "num_tokens": 1346962455.0, "step": 1701 }, { "epoch": 2.3222392899812254, "grad_norm": 0.10581393539905548, "learning_rate": 8.62871630981105e-06, "loss": 0.4045, "num_tokens": 1347709933.0, "step": 1702 }, { "epoch": 2.3236047107014848, "grad_norm": 0.09043489396572113, "learning_rate": 8.610982573144376e-06, "loss": 0.4245, "num_tokens": 1348556651.0, "step": 1703 }, { "epoch": 2.324970131421744, "grad_norm": 0.09663306176662445, "learning_rate": 8.593277881162171e-06, "loss": 0.4218, "num_tokens": 1349271651.0, "step": 1704 }, { "epoch": 2.326335552142004, "grad_norm": 0.10015175491571426, "learning_rate": 8.575602272271077e-06, "loss": 0.4118, "num_tokens": 1350145163.0, "step": 1705 }, { "epoch": 2.3277009728622633, "grad_norm": 0.09526615589857101, "learning_rate": 8.557955784814647e-06, "loss": 0.431, "num_tokens": 1350961322.0, "step": 1706 }, { "epoch": 2.3290663935825227, "grad_norm": 0.10025918483734131, "learning_rate": 8.540338457073253e-06, "loss": 0.4004, "num_tokens": 1351705998.0, "step": 1707 }, { "epoch": 2.330431814302782, "grad_norm": 0.1089571937918663, "learning_rate": 8.522750327264024e-06, "loss": 0.3965, "num_tokens": 1352435688.0, "step": 1708 }, { "epoch": 2.3317972350230414, "grad_norm": 0.09926284104585648, "learning_rate": 8.505191433540743e-06, "loss": 0.4012, "num_tokens": 1353186083.0, "step": 1709 }, { "epoch": 2.3331626557433007, "grad_norm": 0.09825371950864792, "learning_rate": 8.487661813993767e-06, "loss": 0.4219, "num_tokens": 1354000253.0, "step": 1710 }, { "epoch": 2.3345280764635605, "grad_norm": 0.09669969230890274, "learning_rate": 8.47016150664997e-06, "loss": 0.4019, "num_tokens": 1354802805.0, "step": 1711 }, { "epoch": 2.33589349718382, "grad_norm": 0.09824495762586594, "learning_rate": 8.452690549472608e-06, "loss": 0.3849, "num_tokens": 1355569674.0, "step": 1712 }, { "epoch": 2.3372589179040792, "grad_norm": 0.09518352150917053, "learning_rate": 8.43524898036129e-06, "loss": 0.3911, "num_tokens": 1356348293.0, "step": 1713 }, { "epoch": 2.3386243386243386, "grad_norm": 0.09892245382070541, "learning_rate": 8.417836837151857e-06, "loss": 0.4155, "num_tokens": 1357133277.0, "step": 1714 }, { "epoch": 2.339989759344598, "grad_norm": 0.09305068105459213, "learning_rate": 8.400454157616343e-06, "loss": 0.396, "num_tokens": 1357935470.0, "step": 1715 }, { "epoch": 2.3413551800648573, "grad_norm": 0.09468959271907806, "learning_rate": 8.383100979462843e-06, "loss": 0.4071, "num_tokens": 1358657615.0, "step": 1716 }, { "epoch": 2.3427206007851167, "grad_norm": 0.09446854889392853, "learning_rate": 8.365777340335466e-06, "loss": 0.4093, "num_tokens": 1359451699.0, "step": 1717 }, { "epoch": 2.3440860215053765, "grad_norm": 0.09599003940820694, "learning_rate": 8.348483277814228e-06, "loss": 0.401, "num_tokens": 1360207796.0, "step": 1718 }, { "epoch": 2.345451442225636, "grad_norm": 0.0982680395245552, "learning_rate": 8.331218829415008e-06, "loss": 0.3989, "num_tokens": 1360991712.0, "step": 1719 }, { "epoch": 2.346816862945895, "grad_norm": 0.09348499029874802, "learning_rate": 8.313984032589423e-06, "loss": 0.388, "num_tokens": 1361687998.0, "step": 1720 }, { "epoch": 2.3481822836661546, "grad_norm": 0.0956382304430008, "learning_rate": 8.296778924724777e-06, "loss": 0.4174, "num_tokens": 1362478611.0, "step": 1721 }, { "epoch": 2.349547704386414, "grad_norm": 0.0981563925743103, "learning_rate": 8.27960354314397e-06, "loss": 0.3933, "num_tokens": 1363219247.0, "step": 1722 }, { "epoch": 2.3509131251066737, "grad_norm": 0.09659777581691742, "learning_rate": 8.262457925105414e-06, "loss": 0.3776, "num_tokens": 1363929189.0, "step": 1723 }, { "epoch": 2.352278545826933, "grad_norm": 0.10304844379425049, "learning_rate": 8.245342107802954e-06, "loss": 0.4154, "num_tokens": 1364782567.0, "step": 1724 }, { "epoch": 2.3536439665471924, "grad_norm": 0.09249619394540787, "learning_rate": 8.22825612836579e-06, "loss": 0.3969, "num_tokens": 1365621876.0, "step": 1725 }, { "epoch": 2.355009387267452, "grad_norm": 0.09000667184591293, "learning_rate": 8.2112000238584e-06, "loss": 0.3992, "num_tokens": 1366376013.0, "step": 1726 }, { "epoch": 2.356374807987711, "grad_norm": 0.10186122357845306, "learning_rate": 8.194173831280438e-06, "loss": 0.4011, "num_tokens": 1367095704.0, "step": 1727 }, { "epoch": 2.3577402287079705, "grad_norm": 0.09162911027669907, "learning_rate": 8.177177587566698e-06, "loss": 0.3917, "num_tokens": 1367826636.0, "step": 1728 }, { "epoch": 2.35910564942823, "grad_norm": 0.09847848117351532, "learning_rate": 8.160211329586991e-06, "loss": 0.3969, "num_tokens": 1368667661.0, "step": 1729 }, { "epoch": 2.3604710701484897, "grad_norm": 0.09857863187789917, "learning_rate": 8.143275094146068e-06, "loss": 0.4067, "num_tokens": 1369420989.0, "step": 1730 }, { "epoch": 2.361836490868749, "grad_norm": 0.09178880602121353, "learning_rate": 8.12636891798356e-06, "loss": 0.3918, "num_tokens": 1370254348.0, "step": 1731 }, { "epoch": 2.3632019115890084, "grad_norm": 0.09568242728710175, "learning_rate": 8.109492837773912e-06, "loss": 0.384, "num_tokens": 1371029173.0, "step": 1732 }, { "epoch": 2.3645673323092677, "grad_norm": 0.09005656093358994, "learning_rate": 8.092646890126255e-06, "loss": 0.4089, "num_tokens": 1371752256.0, "step": 1733 }, { "epoch": 2.365932753029527, "grad_norm": 0.1021956130862236, "learning_rate": 8.075831111584364e-06, "loss": 0.4236, "num_tokens": 1372569219.0, "step": 1734 }, { "epoch": 2.3672981737497865, "grad_norm": 0.0977121889591217, "learning_rate": 8.05904553862657e-06, "loss": 0.4206, "num_tokens": 1373430175.0, "step": 1735 }, { "epoch": 2.3686635944700463, "grad_norm": 0.10290434211492538, "learning_rate": 8.042290207665677e-06, "loss": 0.3748, "num_tokens": 1374187197.0, "step": 1736 }, { "epoch": 2.3700290151903056, "grad_norm": 0.0911070704460144, "learning_rate": 8.025565155048884e-06, "loss": 0.3953, "num_tokens": 1374917106.0, "step": 1737 }, { "epoch": 2.371394435910565, "grad_norm": 0.09373777359724045, "learning_rate": 8.008870417057709e-06, "loss": 0.4177, "num_tokens": 1375736124.0, "step": 1738 }, { "epoch": 2.3727598566308243, "grad_norm": 0.09695442765951157, "learning_rate": 7.99220602990791e-06, "loss": 0.4023, "num_tokens": 1376519049.0, "step": 1739 }, { "epoch": 2.3741252773510837, "grad_norm": 0.10201204568147659, "learning_rate": 7.975572029749402e-06, "loss": 0.3844, "num_tokens": 1377326429.0, "step": 1740 }, { "epoch": 2.375490698071343, "grad_norm": 0.09315043687820435, "learning_rate": 7.958968452666184e-06, "loss": 0.4001, "num_tokens": 1378091477.0, "step": 1741 }, { "epoch": 2.376856118791603, "grad_norm": 0.10300703346729279, "learning_rate": 7.94239533467626e-06, "loss": 0.4066, "num_tokens": 1378869160.0, "step": 1742 }, { "epoch": 2.378221539511862, "grad_norm": 0.09678801149129868, "learning_rate": 7.925852711731557e-06, "loss": 0.3882, "num_tokens": 1379639295.0, "step": 1743 }, { "epoch": 2.3795869602321216, "grad_norm": 0.09189815074205399, "learning_rate": 7.909340619717848e-06, "loss": 0.3835, "num_tokens": 1380422846.0, "step": 1744 }, { "epoch": 2.380952380952381, "grad_norm": 0.0885414108633995, "learning_rate": 7.892859094454672e-06, "loss": 0.4055, "num_tokens": 1381294741.0, "step": 1745 }, { "epoch": 2.3823178016726403, "grad_norm": 0.09266644716262817, "learning_rate": 7.87640817169528e-06, "loss": 0.3875, "num_tokens": 1382017672.0, "step": 1746 }, { "epoch": 2.3836832223928996, "grad_norm": 0.10692784190177917, "learning_rate": 7.859987887126512e-06, "loss": 0.4099, "num_tokens": 1382835216.0, "step": 1747 }, { "epoch": 2.3850486431131595, "grad_norm": 0.10279741883277893, "learning_rate": 7.843598276368756e-06, "loss": 0.3891, "num_tokens": 1383522147.0, "step": 1748 }, { "epoch": 2.386414063833419, "grad_norm": 0.09932444244623184, "learning_rate": 7.827239374975853e-06, "loss": 0.4213, "num_tokens": 1384307226.0, "step": 1749 }, { "epoch": 2.387779484553678, "grad_norm": 0.09532559663057327, "learning_rate": 7.810911218435044e-06, "loss": 0.4046, "num_tokens": 1385145567.0, "step": 1750 }, { "epoch": 2.3891449052739375, "grad_norm": 0.095296211540699, "learning_rate": 7.794613842166859e-06, "loss": 0.4183, "num_tokens": 1385939791.0, "step": 1751 }, { "epoch": 2.390510325994197, "grad_norm": 0.09857485443353653, "learning_rate": 7.778347281525058e-06, "loss": 0.4104, "num_tokens": 1386752292.0, "step": 1752 }, { "epoch": 2.3918757467144562, "grad_norm": 0.09792789816856384, "learning_rate": 7.762111571796562e-06, "loss": 0.3807, "num_tokens": 1387522814.0, "step": 1753 }, { "epoch": 2.3932411674347156, "grad_norm": 0.09649821370840073, "learning_rate": 7.745906748201354e-06, "loss": 0.4063, "num_tokens": 1388308036.0, "step": 1754 }, { "epoch": 2.3946065881549754, "grad_norm": 0.09559693932533264, "learning_rate": 7.72973284589243e-06, "loss": 0.4009, "num_tokens": 1389069283.0, "step": 1755 }, { "epoch": 2.3959720088752348, "grad_norm": 0.0972905158996582, "learning_rate": 7.713589899955698e-06, "loss": 0.4137, "num_tokens": 1389868793.0, "step": 1756 }, { "epoch": 2.397337429595494, "grad_norm": 0.09548412263393402, "learning_rate": 7.69747794540992e-06, "loss": 0.4291, "num_tokens": 1390709024.0, "step": 1757 }, { "epoch": 2.3987028503157535, "grad_norm": 0.10413064807653427, "learning_rate": 7.681397017206621e-06, "loss": 0.4078, "num_tokens": 1391504759.0, "step": 1758 }, { "epoch": 2.400068271036013, "grad_norm": 0.09684674441814423, "learning_rate": 7.665347150230029e-06, "loss": 0.4221, "num_tokens": 1392285188.0, "step": 1759 }, { "epoch": 2.4014336917562726, "grad_norm": 0.09484835714101791, "learning_rate": 7.649328379296984e-06, "loss": 0.4127, "num_tokens": 1393105374.0, "step": 1760 }, { "epoch": 2.402799112476532, "grad_norm": 0.09593571722507477, "learning_rate": 7.633340739156879e-06, "loss": 0.4033, "num_tokens": 1393823494.0, "step": 1761 }, { "epoch": 2.4041645331967914, "grad_norm": 0.10123632103204727, "learning_rate": 7.617384264491558e-06, "loss": 0.4107, "num_tokens": 1394667198.0, "step": 1762 }, { "epoch": 2.4055299539170507, "grad_norm": 0.09611580520868301, "learning_rate": 7.601458989915283e-06, "loss": 0.4308, "num_tokens": 1395563568.0, "step": 1763 }, { "epoch": 2.40689537463731, "grad_norm": 0.09757862985134125, "learning_rate": 7.585564949974615e-06, "loss": 0.4179, "num_tokens": 1396329196.0, "step": 1764 }, { "epoch": 2.4082607953575694, "grad_norm": 0.09676212817430496, "learning_rate": 7.569702179148357e-06, "loss": 0.4048, "num_tokens": 1397104045.0, "step": 1765 }, { "epoch": 2.409626216077829, "grad_norm": 0.09432803094387054, "learning_rate": 7.553870711847489e-06, "loss": 0.4223, "num_tokens": 1397901951.0, "step": 1766 }, { "epoch": 2.4109916367980886, "grad_norm": 0.09891921281814575, "learning_rate": 7.538070582415087e-06, "loss": 0.4222, "num_tokens": 1398816207.0, "step": 1767 }, { "epoch": 2.412357057518348, "grad_norm": 0.09480462223291397, "learning_rate": 7.522301825126235e-06, "loss": 0.3823, "num_tokens": 1399594570.0, "step": 1768 }, { "epoch": 2.4137224782386073, "grad_norm": 0.09379947930574417, "learning_rate": 7.506564474187972e-06, "loss": 0.4036, "num_tokens": 1400391587.0, "step": 1769 }, { "epoch": 2.4150878989588667, "grad_norm": 0.0955926701426506, "learning_rate": 7.490858563739198e-06, "loss": 0.3909, "num_tokens": 1401148721.0, "step": 1770 }, { "epoch": 2.416453319679126, "grad_norm": 0.10154493153095245, "learning_rate": 7.475184127850619e-06, "loss": 0.4014, "num_tokens": 1401943821.0, "step": 1771 }, { "epoch": 2.4178187403993854, "grad_norm": 0.09139566868543625, "learning_rate": 7.459541200524655e-06, "loss": 0.4013, "num_tokens": 1402668535.0, "step": 1772 }, { "epoch": 2.419184161119645, "grad_norm": 0.09163139760494232, "learning_rate": 7.443929815695381e-06, "loss": 0.4072, "num_tokens": 1403473500.0, "step": 1773 }, { "epoch": 2.4205495818399045, "grad_norm": 0.09533286839723587, "learning_rate": 7.428350007228442e-06, "loss": 0.4257, "num_tokens": 1404316948.0, "step": 1774 }, { "epoch": 2.421915002560164, "grad_norm": 0.0950283408164978, "learning_rate": 7.412801808920987e-06, "loss": 0.3958, "num_tokens": 1405042674.0, "step": 1775 }, { "epoch": 2.4232804232804233, "grad_norm": 0.10051173716783524, "learning_rate": 7.397285254501603e-06, "loss": 0.4081, "num_tokens": 1405839141.0, "step": 1776 }, { "epoch": 2.4246458440006826, "grad_norm": 0.08750990778207779, "learning_rate": 7.3818003776302105e-06, "loss": 0.423, "num_tokens": 1406631957.0, "step": 1777 }, { "epoch": 2.426011264720942, "grad_norm": 0.09720689058303833, "learning_rate": 7.366347211898031e-06, "loss": 0.3904, "num_tokens": 1407385791.0, "step": 1778 }, { "epoch": 2.4273766854412013, "grad_norm": 0.10062023997306824, "learning_rate": 7.350925790827481e-06, "loss": 0.3933, "num_tokens": 1408181054.0, "step": 1779 }, { "epoch": 2.428742106161461, "grad_norm": 0.09341268986463547, "learning_rate": 7.335536147872131e-06, "loss": 0.426, "num_tokens": 1409032244.0, "step": 1780 }, { "epoch": 2.4301075268817205, "grad_norm": 0.09167987108230591, "learning_rate": 7.320178316416604e-06, "loss": 0.4025, "num_tokens": 1409879239.0, "step": 1781 }, { "epoch": 2.43147294760198, "grad_norm": 0.09318269789218903, "learning_rate": 7.30485232977652e-06, "loss": 0.4212, "num_tokens": 1410690046.0, "step": 1782 }, { "epoch": 2.432838368322239, "grad_norm": 0.09486595541238785, "learning_rate": 7.289558221198401e-06, "loss": 0.3984, "num_tokens": 1411505870.0, "step": 1783 }, { "epoch": 2.4342037890424986, "grad_norm": 0.09608538448810577, "learning_rate": 7.2742960238596415e-06, "loss": 0.3848, "num_tokens": 1412260793.0, "step": 1784 }, { "epoch": 2.4355692097627584, "grad_norm": 0.09319755434989929, "learning_rate": 7.259065770868397e-06, "loss": 0.4076, "num_tokens": 1413061116.0, "step": 1785 }, { "epoch": 2.4369346304830177, "grad_norm": 0.09459555149078369, "learning_rate": 7.243867495263527e-06, "loss": 0.4102, "num_tokens": 1413907558.0, "step": 1786 }, { "epoch": 2.438300051203277, "grad_norm": 0.09714709222316742, "learning_rate": 7.228701230014528e-06, "loss": 0.3906, "num_tokens": 1414652146.0, "step": 1787 }, { "epoch": 2.4396654719235364, "grad_norm": 0.09034030139446259, "learning_rate": 7.2135670080214516e-06, "loss": 0.4188, "num_tokens": 1415502490.0, "step": 1788 }, { "epoch": 2.441030892643796, "grad_norm": 0.09504076093435287, "learning_rate": 7.198464862114838e-06, "loss": 0.4157, "num_tokens": 1416299429.0, "step": 1789 }, { "epoch": 2.442396313364055, "grad_norm": 0.09970548748970032, "learning_rate": 7.183394825055652e-06, "loss": 0.4203, "num_tokens": 1417177003.0, "step": 1790 }, { "epoch": 2.4437617340843145, "grad_norm": 0.09054993093013763, "learning_rate": 7.168356929535196e-06, "loss": 0.4052, "num_tokens": 1417968963.0, "step": 1791 }, { "epoch": 2.4451271548045743, "grad_norm": 0.09485835582017899, "learning_rate": 7.153351208175048e-06, "loss": 0.3996, "num_tokens": 1418716921.0, "step": 1792 }, { "epoch": 2.4464925755248337, "grad_norm": 0.09628629684448242, "learning_rate": 7.138377693527011e-06, "loss": 0.4107, "num_tokens": 1419547958.0, "step": 1793 }, { "epoch": 2.447857996245093, "grad_norm": 0.09505634009838104, "learning_rate": 7.123436418072989e-06, "loss": 0.4122, "num_tokens": 1420349607.0, "step": 1794 }, { "epoch": 2.4492234169653524, "grad_norm": 0.09078752994537354, "learning_rate": 7.108527414224974e-06, "loss": 0.3953, "num_tokens": 1421121810.0, "step": 1795 }, { "epoch": 2.4505888376856118, "grad_norm": 0.10481933504343033, "learning_rate": 7.09365071432494e-06, "loss": 0.427, "num_tokens": 1421910199.0, "step": 1796 }, { "epoch": 2.4519542584058716, "grad_norm": 0.10021737962961197, "learning_rate": 7.078806350644796e-06, "loss": 0.3749, "num_tokens": 1422706147.0, "step": 1797 }, { "epoch": 2.453319679126131, "grad_norm": 0.09530755877494812, "learning_rate": 7.063994355386293e-06, "loss": 0.4136, "num_tokens": 1423529479.0, "step": 1798 }, { "epoch": 2.4546850998463903, "grad_norm": 0.09331899136304855, "learning_rate": 7.049214760680973e-06, "loss": 0.4139, "num_tokens": 1424345864.0, "step": 1799 }, { "epoch": 2.4560505205666496, "grad_norm": 0.09697050601243973, "learning_rate": 7.03446759859008e-06, "loss": 0.3999, "num_tokens": 1425043344.0, "step": 1800 }, { "epoch": 2.457415941286909, "grad_norm": 0.10173619538545609, "learning_rate": 7.0197529011045165e-06, "loss": 0.4139, "num_tokens": 1425827500.0, "step": 1801 }, { "epoch": 2.4587813620071683, "grad_norm": 0.09823158383369446, "learning_rate": 7.005070700144754e-06, "loss": 0.3907, "num_tokens": 1426578235.0, "step": 1802 }, { "epoch": 2.4601467827274277, "grad_norm": 0.0910264328122139, "learning_rate": 6.990421027560767e-06, "loss": 0.3853, "num_tokens": 1427308406.0, "step": 1803 }, { "epoch": 2.4615122034476875, "grad_norm": 0.09311746060848236, "learning_rate": 6.975803915131972e-06, "loss": 0.397, "num_tokens": 1428065875.0, "step": 1804 }, { "epoch": 2.462877624167947, "grad_norm": 0.10079707950353622, "learning_rate": 6.961219394567148e-06, "loss": 0.3944, "num_tokens": 1428838663.0, "step": 1805 }, { "epoch": 2.4642430448882062, "grad_norm": 0.08551298081874847, "learning_rate": 6.946667497504371e-06, "loss": 0.4013, "num_tokens": 1429636120.0, "step": 1806 }, { "epoch": 2.4656084656084656, "grad_norm": 0.09375662356615067, "learning_rate": 6.932148255510955e-06, "loss": 0.3919, "num_tokens": 1430422929.0, "step": 1807 }, { "epoch": 2.466973886328725, "grad_norm": 0.09600969403982162, "learning_rate": 6.917661700083371e-06, "loss": 0.3769, "num_tokens": 1431212078.0, "step": 1808 }, { "epoch": 2.4683393070489843, "grad_norm": 0.0947382003068924, "learning_rate": 6.9032078626471765e-06, "loss": 0.4117, "num_tokens": 1432025261.0, "step": 1809 }, { "epoch": 2.469704727769244, "grad_norm": 0.09546377509832382, "learning_rate": 6.888786774556972e-06, "loss": 0.4352, "num_tokens": 1432828284.0, "step": 1810 }, { "epoch": 2.4710701484895035, "grad_norm": 0.09574444591999054, "learning_rate": 6.8743984670963045e-06, "loss": 0.3765, "num_tokens": 1433533612.0, "step": 1811 }, { "epoch": 2.472435569209763, "grad_norm": 0.0878109261393547, "learning_rate": 6.860042971477601e-06, "loss": 0.4139, "num_tokens": 1434344180.0, "step": 1812 }, { "epoch": 2.473800989930022, "grad_norm": 0.10002578049898148, "learning_rate": 6.845720318842117e-06, "loss": 0.3976, "num_tokens": 1435153086.0, "step": 1813 }, { "epoch": 2.4751664106502815, "grad_norm": 0.0894789770245552, "learning_rate": 6.831430540259875e-06, "loss": 0.3853, "num_tokens": 1435959542.0, "step": 1814 }, { "epoch": 2.476531831370541, "grad_norm": 0.0940762609243393, "learning_rate": 6.8171736667295695e-06, "loss": 0.4039, "num_tokens": 1436752252.0, "step": 1815 }, { "epoch": 2.4778972520908003, "grad_norm": 0.10347861051559448, "learning_rate": 6.802949729178516e-06, "loss": 0.4124, "num_tokens": 1437495623.0, "step": 1816 }, { "epoch": 2.47926267281106, "grad_norm": 0.09600193053483963, "learning_rate": 6.788758758462586e-06, "loss": 0.4194, "num_tokens": 1438314049.0, "step": 1817 }, { "epoch": 2.4806280935313194, "grad_norm": 0.09283630549907684, "learning_rate": 6.774600785366134e-06, "loss": 0.3906, "num_tokens": 1439121216.0, "step": 1818 }, { "epoch": 2.4819935142515788, "grad_norm": 0.08823809772729874, "learning_rate": 6.7604758406019325e-06, "loss": 0.3956, "num_tokens": 1439910675.0, "step": 1819 }, { "epoch": 2.483358934971838, "grad_norm": 0.09449886530637741, "learning_rate": 6.746383954811109e-06, "loss": 0.3981, "num_tokens": 1440689562.0, "step": 1820 }, { "epoch": 2.4847243556920975, "grad_norm": 0.09484851360321045, "learning_rate": 6.732325158563071e-06, "loss": 0.4394, "num_tokens": 1441520996.0, "step": 1821 }, { "epoch": 2.4860897764123573, "grad_norm": 0.09852176904678345, "learning_rate": 6.718299482355448e-06, "loss": 0.3846, "num_tokens": 1442245109.0, "step": 1822 }, { "epoch": 2.4874551971326166, "grad_norm": 0.0956614539027214, "learning_rate": 6.704306956614035e-06, "loss": 0.3848, "num_tokens": 1443015929.0, "step": 1823 }, { "epoch": 2.488820617852876, "grad_norm": 0.09699973464012146, "learning_rate": 6.690347611692688e-06, "loss": 0.4189, "num_tokens": 1443804949.0, "step": 1824 }, { "epoch": 2.4901860385731354, "grad_norm": 0.0890147015452385, "learning_rate": 6.676421477873305e-06, "loss": 0.3863, "num_tokens": 1444571298.0, "step": 1825 }, { "epoch": 2.4915514592933947, "grad_norm": 0.09558888524770737, "learning_rate": 6.662528585365724e-06, "loss": 0.3865, "num_tokens": 1445285635.0, "step": 1826 }, { "epoch": 2.492916880013654, "grad_norm": 0.09582627564668655, "learning_rate": 6.648668964307694e-06, "loss": 0.4002, "num_tokens": 1446119786.0, "step": 1827 }, { "epoch": 2.4942823007339134, "grad_norm": 0.0911812111735344, "learning_rate": 6.6348426447647756e-06, "loss": 0.3935, "num_tokens": 1446924803.0, "step": 1828 }, { "epoch": 2.4956477214541732, "grad_norm": 0.09683258831501007, "learning_rate": 6.621049656730278e-06, "loss": 0.4186, "num_tokens": 1447701397.0, "step": 1829 }, { "epoch": 2.4970131421744326, "grad_norm": 0.09953998029232025, "learning_rate": 6.607290030125219e-06, "loss": 0.3943, "num_tokens": 1448526996.0, "step": 1830 }, { "epoch": 2.498378562894692, "grad_norm": 0.09429541230201721, "learning_rate": 6.5935637947982485e-06, "loss": 0.379, "num_tokens": 1449287847.0, "step": 1831 }, { "epoch": 2.4997439836149513, "grad_norm": 0.09509303420782089, "learning_rate": 6.579870980525573e-06, "loss": 0.4283, "num_tokens": 1450188772.0, "step": 1832 }, { "epoch": 2.5011094043352107, "grad_norm": 0.0907871276140213, "learning_rate": 6.566211617010899e-06, "loss": 0.4028, "num_tokens": 1450935495.0, "step": 1833 }, { "epoch": 2.5024748250554705, "grad_norm": 0.09417210519313812, "learning_rate": 6.552585733885369e-06, "loss": 0.3902, "num_tokens": 1451703993.0, "step": 1834 }, { "epoch": 2.50384024577573, "grad_norm": 0.10014346241950989, "learning_rate": 6.538993360707504e-06, "loss": 0.4186, "num_tokens": 1452414043.0, "step": 1835 }, { "epoch": 2.505205666495989, "grad_norm": 0.09827961027622223, "learning_rate": 6.525434526963119e-06, "loss": 0.4075, "num_tokens": 1453237503.0, "step": 1836 }, { "epoch": 2.5065710872162486, "grad_norm": 0.09835567325353622, "learning_rate": 6.511909262065287e-06, "loss": 0.404, "num_tokens": 1454064859.0, "step": 1837 }, { "epoch": 2.507936507936508, "grad_norm": 0.08895473182201385, "learning_rate": 6.498417595354247e-06, "loss": 0.4009, "num_tokens": 1454823949.0, "step": 1838 }, { "epoch": 2.5093019286567673, "grad_norm": 0.09347252547740936, "learning_rate": 6.484959556097359e-06, "loss": 0.4144, "num_tokens": 1455623248.0, "step": 1839 }, { "epoch": 2.5106673493770266, "grad_norm": 0.0972462072968483, "learning_rate": 6.4715351734890465e-06, "loss": 0.4149, "num_tokens": 1456436991.0, "step": 1840 }, { "epoch": 2.512032770097286, "grad_norm": 0.09585370868444443, "learning_rate": 6.458144476650699e-06, "loss": 0.4142, "num_tokens": 1457253396.0, "step": 1841 }, { "epoch": 2.513398190817546, "grad_norm": 0.09620252251625061, "learning_rate": 6.44478749463065e-06, "loss": 0.4181, "num_tokens": 1458010697.0, "step": 1842 }, { "epoch": 2.514763611537805, "grad_norm": 0.09454936534166336, "learning_rate": 6.431464256404084e-06, "loss": 0.3784, "num_tokens": 1458764233.0, "step": 1843 }, { "epoch": 2.5161290322580645, "grad_norm": 0.09057123959064484, "learning_rate": 6.418174790873002e-06, "loss": 0.3833, "num_tokens": 1459454120.0, "step": 1844 }, { "epoch": 2.517494452978324, "grad_norm": 0.09649854153394699, "learning_rate": 6.404919126866124e-06, "loss": 0.4038, "num_tokens": 1460285800.0, "step": 1845 }, { "epoch": 2.518859873698583, "grad_norm": 0.09683135896921158, "learning_rate": 6.391697293138856e-06, "loss": 0.4135, "num_tokens": 1461130936.0, "step": 1846 }, { "epoch": 2.520225294418843, "grad_norm": 0.09567543864250183, "learning_rate": 6.378509318373205e-06, "loss": 0.4032, "num_tokens": 1461946526.0, "step": 1847 }, { "epoch": 2.5215907151391024, "grad_norm": 0.09870240837335587, "learning_rate": 6.36535523117774e-06, "loss": 0.4139, "num_tokens": 1462687772.0, "step": 1848 }, { "epoch": 2.5229561358593617, "grad_norm": 0.09247472882270813, "learning_rate": 6.352235060087516e-06, "loss": 0.413, "num_tokens": 1463617569.0, "step": 1849 }, { "epoch": 2.524321556579621, "grad_norm": 0.09459920227527618, "learning_rate": 6.3391488335640075e-06, "loss": 0.4225, "num_tokens": 1464370713.0, "step": 1850 }, { "epoch": 2.5256869772998805, "grad_norm": 0.09836330264806747, "learning_rate": 6.326096579995058e-06, "loss": 0.4253, "num_tokens": 1465114050.0, "step": 1851 }, { "epoch": 2.52705239802014, "grad_norm": 0.0973476767539978, "learning_rate": 6.313078327694814e-06, "loss": 0.4023, "num_tokens": 1465884916.0, "step": 1852 }, { "epoch": 2.528417818740399, "grad_norm": 0.09233789145946503, "learning_rate": 6.30009410490366e-06, "loss": 0.4223, "num_tokens": 1466689313.0, "step": 1853 }, { "epoch": 2.529783239460659, "grad_norm": 0.09793077409267426, "learning_rate": 6.2871439397881664e-06, "loss": 0.3717, "num_tokens": 1467423211.0, "step": 1854 }, { "epoch": 2.5311486601809183, "grad_norm": 0.09201990813016891, "learning_rate": 6.2742278604410135e-06, "loss": 0.4325, "num_tokens": 1468245492.0, "step": 1855 }, { "epoch": 2.5325140809011777, "grad_norm": 0.10148776322603226, "learning_rate": 6.261345894880947e-06, "loss": 0.3849, "num_tokens": 1468983572.0, "step": 1856 }, { "epoch": 2.533879501621437, "grad_norm": 0.09654538333415985, "learning_rate": 6.248498071052713e-06, "loss": 0.3886, "num_tokens": 1469781507.0, "step": 1857 }, { "epoch": 2.5352449223416964, "grad_norm": 0.093331478536129, "learning_rate": 6.235684416826989e-06, "loss": 0.3985, "num_tokens": 1470568649.0, "step": 1858 }, { "epoch": 2.536610343061956, "grad_norm": 0.08825855702161789, "learning_rate": 6.222904960000322e-06, "loss": 0.4191, "num_tokens": 1471480425.0, "step": 1859 }, { "epoch": 2.5379757637822156, "grad_norm": 0.09126836061477661, "learning_rate": 6.210159728295081e-06, "loss": 0.4082, "num_tokens": 1472318093.0, "step": 1860 }, { "epoch": 2.539341184502475, "grad_norm": 0.09164980798959732, "learning_rate": 6.197448749359399e-06, "loss": 0.4234, "num_tokens": 1473069778.0, "step": 1861 }, { "epoch": 2.5407066052227343, "grad_norm": 0.10349740833044052, "learning_rate": 6.184772050767094e-06, "loss": 0.3933, "num_tokens": 1473920662.0, "step": 1862 }, { "epoch": 2.5420720259429936, "grad_norm": 0.09195978194475174, "learning_rate": 6.172129660017622e-06, "loss": 0.4222, "num_tokens": 1474762728.0, "step": 1863 }, { "epoch": 2.543437446663253, "grad_norm": 0.09737087041139603, "learning_rate": 6.159521604536017e-06, "loss": 0.4095, "num_tokens": 1475525542.0, "step": 1864 }, { "epoch": 2.5448028673835124, "grad_norm": 0.09696702659130096, "learning_rate": 6.1469479116728315e-06, "loss": 0.3995, "num_tokens": 1476322593.0, "step": 1865 }, { "epoch": 2.5461682881037717, "grad_norm": 0.09194034337997437, "learning_rate": 6.13440860870407e-06, "loss": 0.4026, "num_tokens": 1477148378.0, "step": 1866 }, { "epoch": 2.5475337088240315, "grad_norm": 0.09153051674365997, "learning_rate": 6.121903722831139e-06, "loss": 0.3937, "num_tokens": 1478083181.0, "step": 1867 }, { "epoch": 2.548899129544291, "grad_norm": 0.09269209206104279, "learning_rate": 6.109433281180784e-06, "loss": 0.3923, "num_tokens": 1478838008.0, "step": 1868 }, { "epoch": 2.5502645502645502, "grad_norm": 0.08581818640232086, "learning_rate": 6.096997310805033e-06, "loss": 0.4107, "num_tokens": 1479703039.0, "step": 1869 }, { "epoch": 2.5516299709848096, "grad_norm": 0.09672318398952484, "learning_rate": 6.0845958386811285e-06, "loss": 0.4182, "num_tokens": 1480533293.0, "step": 1870 }, { "epoch": 2.5529953917050694, "grad_norm": 0.08904803544282913, "learning_rate": 6.072228891711487e-06, "loss": 0.3855, "num_tokens": 1481310785.0, "step": 1871 }, { "epoch": 2.5543608124253288, "grad_norm": 0.08800652623176575, "learning_rate": 6.05989649672362e-06, "loss": 0.4073, "num_tokens": 1482058492.0, "step": 1872 }, { "epoch": 2.555726233145588, "grad_norm": 0.09128383547067642, "learning_rate": 6.047598680470086e-06, "loss": 0.3955, "num_tokens": 1482870871.0, "step": 1873 }, { "epoch": 2.5570916538658475, "grad_norm": 0.094191774725914, "learning_rate": 6.035335469628445e-06, "loss": 0.3889, "num_tokens": 1483675014.0, "step": 1874 }, { "epoch": 2.558457074586107, "grad_norm": 0.09238852560520172, "learning_rate": 6.023106890801177e-06, "loss": 0.4173, "num_tokens": 1484492262.0, "step": 1875 }, { "epoch": 2.559822495306366, "grad_norm": 0.09014425426721573, "learning_rate": 6.010912970515631e-06, "loss": 0.4081, "num_tokens": 1485364477.0, "step": 1876 }, { "epoch": 2.5611879160266255, "grad_norm": 0.09259936958551407, "learning_rate": 5.9987537352239755e-06, "loss": 0.3969, "num_tokens": 1486108199.0, "step": 1877 }, { "epoch": 2.562553336746885, "grad_norm": 0.0938514843583107, "learning_rate": 5.986629211303147e-06, "loss": 0.4037, "num_tokens": 1486872220.0, "step": 1878 }, { "epoch": 2.5639187574671447, "grad_norm": 0.09032351523637772, "learning_rate": 5.9745394250547695e-06, "loss": 0.4009, "num_tokens": 1487716900.0, "step": 1879 }, { "epoch": 2.565284178187404, "grad_norm": 0.08772122114896774, "learning_rate": 5.962484402705118e-06, "loss": 0.4076, "num_tokens": 1488516312.0, "step": 1880 }, { "epoch": 2.5666495989076634, "grad_norm": 0.09692647308111191, "learning_rate": 5.950464170405052e-06, "loss": 0.4063, "num_tokens": 1489308517.0, "step": 1881 }, { "epoch": 2.568015019627923, "grad_norm": 0.0898628979921341, "learning_rate": 5.938478754229964e-06, "loss": 0.4153, "num_tokens": 1490120630.0, "step": 1882 }, { "epoch": 2.569380440348182, "grad_norm": 0.09441210329532623, "learning_rate": 5.926528180179712e-06, "loss": 0.395, "num_tokens": 1490924723.0, "step": 1883 }, { "epoch": 2.570745861068442, "grad_norm": 0.09696640819311142, "learning_rate": 5.914612474178582e-06, "loss": 0.4236, "num_tokens": 1491683948.0, "step": 1884 }, { "epoch": 2.5721112817887013, "grad_norm": 0.09465841203927994, "learning_rate": 5.902731662075215e-06, "loss": 0.3996, "num_tokens": 1492459019.0, "step": 1885 }, { "epoch": 2.5734767025089607, "grad_norm": 0.09624532610177994, "learning_rate": 5.890885769642553e-06, "loss": 0.4049, "num_tokens": 1493237631.0, "step": 1886 }, { "epoch": 2.57484212322922, "grad_norm": 0.09668421745300293, "learning_rate": 5.879074822577806e-06, "loss": 0.4047, "num_tokens": 1494015382.0, "step": 1887 }, { "epoch": 2.5762075439494794, "grad_norm": 0.087858185172081, "learning_rate": 5.8672988465023515e-06, "loss": 0.4179, "num_tokens": 1494835052.0, "step": 1888 }, { "epoch": 2.5775729646697387, "grad_norm": 0.0970911756157875, "learning_rate": 5.855557866961721e-06, "loss": 0.3934, "num_tokens": 1495583344.0, "step": 1889 }, { "epoch": 2.578938385389998, "grad_norm": 0.0900985598564148, "learning_rate": 5.843851909425517e-06, "loss": 0.3962, "num_tokens": 1496377377.0, "step": 1890 }, { "epoch": 2.5803038061102574, "grad_norm": 0.09294350445270538, "learning_rate": 5.8321809992873865e-06, "loss": 0.4095, "num_tokens": 1497205363.0, "step": 1891 }, { "epoch": 2.5816692268305172, "grad_norm": 0.09490440785884857, "learning_rate": 5.820545161864934e-06, "loss": 0.3867, "num_tokens": 1497974242.0, "step": 1892 }, { "epoch": 2.5830346475507766, "grad_norm": 0.09491667151451111, "learning_rate": 5.808944422399689e-06, "loss": 0.4174, "num_tokens": 1498773633.0, "step": 1893 }, { "epoch": 2.584400068271036, "grad_norm": 0.09046527743339539, "learning_rate": 5.797378806057029e-06, "loss": 0.4003, "num_tokens": 1499550674.0, "step": 1894 }, { "epoch": 2.5857654889912953, "grad_norm": 0.09246797114610672, "learning_rate": 5.785848337926157e-06, "loss": 0.4314, "num_tokens": 1500399990.0, "step": 1895 }, { "epoch": 2.587130909711555, "grad_norm": 0.10155907273292542, "learning_rate": 5.774353043020022e-06, "loss": 0.4155, "num_tokens": 1501211656.0, "step": 1896 }, { "epoch": 2.5884963304318145, "grad_norm": 0.09509172290563583, "learning_rate": 5.762892946275271e-06, "loss": 0.4089, "num_tokens": 1501994740.0, "step": 1897 }, { "epoch": 2.589861751152074, "grad_norm": 0.09087436646223068, "learning_rate": 5.751468072552196e-06, "loss": 0.4168, "num_tokens": 1502754279.0, "step": 1898 }, { "epoch": 2.591227171872333, "grad_norm": 0.09958524256944656, "learning_rate": 5.740078446634683e-06, "loss": 0.4106, "num_tokens": 1503517437.0, "step": 1899 }, { "epoch": 2.5925925925925926, "grad_norm": 0.093954898416996, "learning_rate": 5.72872409323015e-06, "loss": 0.4166, "num_tokens": 1504363030.0, "step": 1900 }, { "epoch": 2.593958013312852, "grad_norm": 0.08738911896944046, "learning_rate": 5.717405036969504e-06, "loss": 0.3924, "num_tokens": 1505148136.0, "step": 1901 }, { "epoch": 2.5953234340331113, "grad_norm": 0.09143788367509842, "learning_rate": 5.706121302407077e-06, "loss": 0.4105, "num_tokens": 1505920947.0, "step": 1902 }, { "epoch": 2.5966888547533706, "grad_norm": 0.09348037838935852, "learning_rate": 5.694872914020581e-06, "loss": 0.4088, "num_tokens": 1506701932.0, "step": 1903 }, { "epoch": 2.5980542754736304, "grad_norm": 0.09961096197366714, "learning_rate": 5.683659896211058e-06, "loss": 0.3955, "num_tokens": 1507438519.0, "step": 1904 }, { "epoch": 2.59941969619389, "grad_norm": 0.0925658568739891, "learning_rate": 5.672482273302811e-06, "loss": 0.4315, "num_tokens": 1508194430.0, "step": 1905 }, { "epoch": 2.600785116914149, "grad_norm": 0.09268490970134735, "learning_rate": 5.6613400695433605e-06, "loss": 0.422, "num_tokens": 1509043296.0, "step": 1906 }, { "epoch": 2.6021505376344085, "grad_norm": 0.09789250791072845, "learning_rate": 5.6502333091033975e-06, "loss": 0.3871, "num_tokens": 1509785666.0, "step": 1907 }, { "epoch": 2.603515958354668, "grad_norm": 0.10204152017831802, "learning_rate": 5.6391620160767315e-06, "loss": 0.4137, "num_tokens": 1510664592.0, "step": 1908 }, { "epoch": 2.6048813790749277, "grad_norm": 0.09404512494802475, "learning_rate": 5.628126214480222e-06, "loss": 0.4114, "num_tokens": 1511459184.0, "step": 1909 }, { "epoch": 2.606246799795187, "grad_norm": 0.09007330238819122, "learning_rate": 5.617125928253745e-06, "loss": 0.4007, "num_tokens": 1512239123.0, "step": 1910 }, { "epoch": 2.6076122205154464, "grad_norm": 0.10262413322925568, "learning_rate": 5.606161181260123e-06, "loss": 0.3884, "num_tokens": 1513001998.0, "step": 1911 }, { "epoch": 2.6089776412357057, "grad_norm": 0.08777247369289398, "learning_rate": 5.595231997285099e-06, "loss": 0.3954, "num_tokens": 1513808401.0, "step": 1912 }, { "epoch": 2.610343061955965, "grad_norm": 0.09884387999773026, "learning_rate": 5.584338400037258e-06, "loss": 0.3892, "num_tokens": 1514534790.0, "step": 1913 }, { "epoch": 2.6117084826762245, "grad_norm": 0.09843716770410538, "learning_rate": 5.5734804131479894e-06, "loss": 0.3981, "num_tokens": 1515289686.0, "step": 1914 }, { "epoch": 2.613073903396484, "grad_norm": 0.0932624340057373, "learning_rate": 5.562658060171433e-06, "loss": 0.3992, "num_tokens": 1516068561.0, "step": 1915 }, { "epoch": 2.6144393241167436, "grad_norm": 0.09369459003210068, "learning_rate": 5.551871364584433e-06, "loss": 0.3866, "num_tokens": 1516912171.0, "step": 1916 }, { "epoch": 2.615804744837003, "grad_norm": 0.09081745892763138, "learning_rate": 5.541120349786477e-06, "loss": 0.4264, "num_tokens": 1517728395.0, "step": 1917 }, { "epoch": 2.6171701655572623, "grad_norm": 0.09968625009059906, "learning_rate": 5.53040503909965e-06, "loss": 0.3982, "num_tokens": 1518471172.0, "step": 1918 }, { "epoch": 2.6185355862775217, "grad_norm": 0.10264986008405685, "learning_rate": 5.51972545576859e-06, "loss": 0.3721, "num_tokens": 1519131142.0, "step": 1919 }, { "epoch": 2.619901006997781, "grad_norm": 0.09678686410188675, "learning_rate": 5.509081622960423e-06, "loss": 0.4055, "num_tokens": 1519969270.0, "step": 1920 }, { "epoch": 2.621266427718041, "grad_norm": 0.09031660854816437, "learning_rate": 5.498473563764734e-06, "loss": 0.4046, "num_tokens": 1520796973.0, "step": 1921 }, { "epoch": 2.6226318484383, "grad_norm": 0.08469882607460022, "learning_rate": 5.487901301193501e-06, "loss": 0.3934, "num_tokens": 1521570678.0, "step": 1922 }, { "epoch": 2.6239972691585596, "grad_norm": 0.09647819399833679, "learning_rate": 5.477364858181034e-06, "loss": 0.4182, "num_tokens": 1522383638.0, "step": 1923 }, { "epoch": 2.625362689878819, "grad_norm": 0.09949373453855515, "learning_rate": 5.466864257583956e-06, "loss": 0.3933, "num_tokens": 1523180444.0, "step": 1924 }, { "epoch": 2.6267281105990783, "grad_norm": 0.09711216390132904, "learning_rate": 5.456399522181137e-06, "loss": 0.4086, "num_tokens": 1524012063.0, "step": 1925 }, { "epoch": 2.6280935313193377, "grad_norm": 0.09419972449541092, "learning_rate": 5.445970674673636e-06, "loss": 0.3967, "num_tokens": 1524768115.0, "step": 1926 }, { "epoch": 2.629458952039597, "grad_norm": 0.09041792154312134, "learning_rate": 5.435577737684668e-06, "loss": 0.3941, "num_tokens": 1525624462.0, "step": 1927 }, { "epoch": 2.6308243727598564, "grad_norm": 0.0896880030632019, "learning_rate": 5.4252207337595445e-06, "loss": 0.4029, "num_tokens": 1526424038.0, "step": 1928 }, { "epoch": 2.632189793480116, "grad_norm": 0.09013330191373825, "learning_rate": 5.414899685365627e-06, "loss": 0.3956, "num_tokens": 1527181680.0, "step": 1929 }, { "epoch": 2.6335552142003755, "grad_norm": 0.09870900958776474, "learning_rate": 5.404614614892283e-06, "loss": 0.3968, "num_tokens": 1527900395.0, "step": 1930 }, { "epoch": 2.634920634920635, "grad_norm": 0.0937613919377327, "learning_rate": 5.394365544650826e-06, "loss": 0.3944, "num_tokens": 1528673960.0, "step": 1931 }, { "epoch": 2.6362860556408942, "grad_norm": 0.09375488758087158, "learning_rate": 5.384152496874483e-06, "loss": 0.4282, "num_tokens": 1529529147.0, "step": 1932 }, { "epoch": 2.637651476361154, "grad_norm": 0.09262928366661072, "learning_rate": 5.373975493718329e-06, "loss": 0.3901, "num_tokens": 1530348806.0, "step": 1933 }, { "epoch": 2.6390168970814134, "grad_norm": 0.09570545703172684, "learning_rate": 5.363834557259262e-06, "loss": 0.4009, "num_tokens": 1531139701.0, "step": 1934 }, { "epoch": 2.6403823178016728, "grad_norm": 0.09804388880729675, "learning_rate": 5.3537297094959236e-06, "loss": 0.4085, "num_tokens": 1531899693.0, "step": 1935 }, { "epoch": 2.641747738521932, "grad_norm": 0.09978711605072021, "learning_rate": 5.3436609723486765e-06, "loss": 0.4391, "num_tokens": 1532694767.0, "step": 1936 }, { "epoch": 2.6431131592421915, "grad_norm": 0.09757533669471741, "learning_rate": 5.333628367659548e-06, "loss": 0.4102, "num_tokens": 1533531871.0, "step": 1937 }, { "epoch": 2.644478579962451, "grad_norm": 0.09257442504167557, "learning_rate": 5.32363191719219e-06, "loss": 0.4099, "num_tokens": 1534292481.0, "step": 1938 }, { "epoch": 2.64584400068271, "grad_norm": 0.09919571876525879, "learning_rate": 5.313671642631817e-06, "loss": 0.4134, "num_tokens": 1535087317.0, "step": 1939 }, { "epoch": 2.6472094214029696, "grad_norm": 0.10052475333213806, "learning_rate": 5.303747565585169e-06, "loss": 0.4043, "num_tokens": 1535892499.0, "step": 1940 }, { "epoch": 2.6485748421232294, "grad_norm": 0.09220122545957565, "learning_rate": 5.2938597075804606e-06, "loss": 0.3962, "num_tokens": 1536664872.0, "step": 1941 }, { "epoch": 2.6499402628434887, "grad_norm": 0.0939878523349762, "learning_rate": 5.284008090067348e-06, "loss": 0.4127, "num_tokens": 1537360127.0, "step": 1942 }, { "epoch": 2.651305683563748, "grad_norm": 0.0976935476064682, "learning_rate": 5.274192734416857e-06, "loss": 0.3933, "num_tokens": 1538105143.0, "step": 1943 }, { "epoch": 2.6526711042840074, "grad_norm": 0.09012950956821442, "learning_rate": 5.2644136619213595e-06, "loss": 0.4145, "num_tokens": 1538957311.0, "step": 1944 }, { "epoch": 2.654036525004267, "grad_norm": 0.094614177942276, "learning_rate": 5.254670893794515e-06, "loss": 0.4147, "num_tokens": 1539786554.0, "step": 1945 }, { "epoch": 2.6554019457245266, "grad_norm": 0.09006557613611221, "learning_rate": 5.2449644511712325e-06, "loss": 0.3752, "num_tokens": 1540582121.0, "step": 1946 }, { "epoch": 2.656767366444786, "grad_norm": 0.0902918204665184, "learning_rate": 5.235294355107616e-06, "loss": 0.4026, "num_tokens": 1541370901.0, "step": 1947 }, { "epoch": 2.6581327871650453, "grad_norm": 0.09573540091514587, "learning_rate": 5.225660626580924e-06, "loss": 0.407, "num_tokens": 1542144660.0, "step": 1948 }, { "epoch": 2.6594982078853047, "grad_norm": 0.09641249477863312, "learning_rate": 5.216063286489525e-06, "loss": 0.3841, "num_tokens": 1542896064.0, "step": 1949 }, { "epoch": 2.660863628605564, "grad_norm": 0.10188838839530945, "learning_rate": 5.206502355652844e-06, "loss": 0.3965, "num_tokens": 1543668149.0, "step": 1950 }, { "epoch": 2.6622290493258234, "grad_norm": 0.09118275344371796, "learning_rate": 5.196977854811341e-06, "loss": 0.3947, "num_tokens": 1544485209.0, "step": 1951 }, { "epoch": 2.6635944700460827, "grad_norm": 0.0989047959446907, "learning_rate": 5.1874898046264255e-06, "loss": 0.4289, "num_tokens": 1545336605.0, "step": 1952 }, { "epoch": 2.6649598907663425, "grad_norm": 0.09386307001113892, "learning_rate": 5.178038225680451e-06, "loss": 0.4165, "num_tokens": 1546100006.0, "step": 1953 }, { "epoch": 2.666325311486602, "grad_norm": 0.09479722380638123, "learning_rate": 5.168623138476645e-06, "loss": 0.3794, "num_tokens": 1546833199.0, "step": 1954 }, { "epoch": 2.6676907322068613, "grad_norm": 0.08828211575746536, "learning_rate": 5.159244563439085e-06, "loss": 0.4094, "num_tokens": 1547641926.0, "step": 1955 }, { "epoch": 2.6690561529271206, "grad_norm": 0.09676945209503174, "learning_rate": 5.149902520912631e-06, "loss": 0.3841, "num_tokens": 1548390921.0, "step": 1956 }, { "epoch": 2.67042157364738, "grad_norm": 0.09259799122810364, "learning_rate": 5.140597031162901e-06, "loss": 0.3899, "num_tokens": 1549217833.0, "step": 1957 }, { "epoch": 2.67178699436764, "grad_norm": 0.08961710333824158, "learning_rate": 5.131328114376213e-06, "loss": 0.4131, "num_tokens": 1550005315.0, "step": 1958 }, { "epoch": 2.673152415087899, "grad_norm": 0.1000472828745842, "learning_rate": 5.1220957906595556e-06, "loss": 0.3791, "num_tokens": 1550760419.0, "step": 1959 }, { "epoch": 2.6745178358081585, "grad_norm": 0.09008089452981949, "learning_rate": 5.112900080040529e-06, "loss": 0.4003, "num_tokens": 1551515702.0, "step": 1960 }, { "epoch": 2.675883256528418, "grad_norm": 0.10186358541250229, "learning_rate": 5.103741002467314e-06, "loss": 0.388, "num_tokens": 1552258243.0, "step": 1961 }, { "epoch": 2.677248677248677, "grad_norm": 0.09509968757629395, "learning_rate": 5.094618577808621e-06, "loss": 0.3988, "num_tokens": 1553044171.0, "step": 1962 }, { "epoch": 2.6786140979689366, "grad_norm": 0.09126241505146027, "learning_rate": 5.085532825853651e-06, "loss": 0.4386, "num_tokens": 1553820306.0, "step": 1963 }, { "epoch": 2.679979518689196, "grad_norm": 0.10007580369710922, "learning_rate": 5.076483766312049e-06, "loss": 0.4104, "num_tokens": 1554657148.0, "step": 1964 }, { "epoch": 2.6813449394094553, "grad_norm": 0.08790992945432663, "learning_rate": 5.0674714188138645e-06, "loss": 0.4104, "num_tokens": 1555521429.0, "step": 1965 }, { "epoch": 2.682710360129715, "grad_norm": 0.08914686739444733, "learning_rate": 5.05849580290951e-06, "loss": 0.4096, "num_tokens": 1556331459.0, "step": 1966 }, { "epoch": 2.6840757808499744, "grad_norm": 0.09171886742115021, "learning_rate": 5.0495569380697085e-06, "loss": 0.3905, "num_tokens": 1557032842.0, "step": 1967 }, { "epoch": 2.685441201570234, "grad_norm": 0.09723248332738876, "learning_rate": 5.040654843685476e-06, "loss": 0.4177, "num_tokens": 1557883362.0, "step": 1968 }, { "epoch": 2.686806622290493, "grad_norm": 0.08785977959632874, "learning_rate": 5.031789539068047e-06, "loss": 0.4171, "num_tokens": 1558757501.0, "step": 1969 }, { "epoch": 2.688172043010753, "grad_norm": 0.09138452261686325, "learning_rate": 5.022961043448851e-06, "loss": 0.3937, "num_tokens": 1559486217.0, "step": 1970 }, { "epoch": 2.6895374637310123, "grad_norm": 0.09416215121746063, "learning_rate": 5.014169375979469e-06, "loss": 0.3881, "num_tokens": 1560164807.0, "step": 1971 }, { "epoch": 2.6909028844512717, "grad_norm": 0.10719852149486542, "learning_rate": 5.005414555731598e-06, "loss": 0.4121, "num_tokens": 1560932496.0, "step": 1972 }, { "epoch": 2.692268305171531, "grad_norm": 0.09036031365394592, "learning_rate": 4.996696601696996e-06, "loss": 0.3715, "num_tokens": 1561606265.0, "step": 1973 }, { "epoch": 2.6936337258917904, "grad_norm": 0.09729386121034622, "learning_rate": 4.9880155327874445e-06, "loss": 0.4158, "num_tokens": 1562405033.0, "step": 1974 }, { "epoch": 2.6949991466120498, "grad_norm": 0.09700602293014526, "learning_rate": 4.97937136783472e-06, "loss": 0.3919, "num_tokens": 1563247776.0, "step": 1975 }, { "epoch": 2.696364567332309, "grad_norm": 0.09253328293561935, "learning_rate": 4.97076412559053e-06, "loss": 0.3734, "num_tokens": 1563973578.0, "step": 1976 }, { "epoch": 2.6977299880525685, "grad_norm": 0.09939900040626526, "learning_rate": 4.9621938247265e-06, "loss": 0.3877, "num_tokens": 1564751727.0, "step": 1977 }, { "epoch": 2.6990954087728283, "grad_norm": 0.08860796689987183, "learning_rate": 4.953660483834112e-06, "loss": 0.4166, "num_tokens": 1565569657.0, "step": 1978 }, { "epoch": 2.7004608294930876, "grad_norm": 0.09382149577140808, "learning_rate": 4.9451641214246696e-06, "loss": 0.3956, "num_tokens": 1566323577.0, "step": 1979 }, { "epoch": 2.701826250213347, "grad_norm": 0.08633459359407425, "learning_rate": 4.936704755929259e-06, "loss": 0.4168, "num_tokens": 1567106515.0, "step": 1980 }, { "epoch": 2.7031916709336063, "grad_norm": 0.09902051836252213, "learning_rate": 4.928282405698719e-06, "loss": 0.3943, "num_tokens": 1567954858.0, "step": 1981 }, { "epoch": 2.7045570916538657, "grad_norm": 0.0989512950181961, "learning_rate": 4.919897089003578e-06, "loss": 0.3786, "num_tokens": 1568726012.0, "step": 1982 }, { "epoch": 2.7059225123741255, "grad_norm": 0.08764886856079102, "learning_rate": 4.911548824034037e-06, "loss": 0.4242, "num_tokens": 1569646395.0, "step": 1983 }, { "epoch": 2.707287933094385, "grad_norm": 0.09326142072677612, "learning_rate": 4.903237628899912e-06, "loss": 0.4091, "num_tokens": 1570404897.0, "step": 1984 }, { "epoch": 2.7086533538146442, "grad_norm": 0.09448375552892685, "learning_rate": 4.8949635216306176e-06, "loss": 0.4232, "num_tokens": 1571212375.0, "step": 1985 }, { "epoch": 2.7100187745349036, "grad_norm": 0.09336896985769272, "learning_rate": 4.886726520175104e-06, "loss": 0.3761, "num_tokens": 1571975179.0, "step": 1986 }, { "epoch": 2.711384195255163, "grad_norm": 0.0889391377568245, "learning_rate": 4.878526642401829e-06, "loss": 0.4096, "num_tokens": 1572783156.0, "step": 1987 }, { "epoch": 2.7127496159754223, "grad_norm": 0.09144853800535202, "learning_rate": 4.870363906098718e-06, "loss": 0.393, "num_tokens": 1573570494.0, "step": 1988 }, { "epoch": 2.7141150366956817, "grad_norm": 0.09347197413444519, "learning_rate": 4.862238328973129e-06, "loss": 0.4044, "num_tokens": 1574401284.0, "step": 1989 }, { "epoch": 2.715480457415941, "grad_norm": 0.09447088092565536, "learning_rate": 4.854149928651809e-06, "loss": 0.4079, "num_tokens": 1575246582.0, "step": 1990 }, { "epoch": 2.716845878136201, "grad_norm": 0.08733005821704865, "learning_rate": 4.846098722680857e-06, "loss": 0.403, "num_tokens": 1575994728.0, "step": 1991 }, { "epoch": 2.71821129885646, "grad_norm": 0.09372181445360184, "learning_rate": 4.838084728525688e-06, "loss": 0.406, "num_tokens": 1576820749.0, "step": 1992 }, { "epoch": 2.7195767195767195, "grad_norm": 0.09063650667667389, "learning_rate": 4.830107963570993e-06, "loss": 0.4065, "num_tokens": 1577624753.0, "step": 1993 }, { "epoch": 2.720942140296979, "grad_norm": 0.09505254775285721, "learning_rate": 4.822168445120702e-06, "loss": 0.4081, "num_tokens": 1578381136.0, "step": 1994 }, { "epoch": 2.7223075610172387, "grad_norm": 0.09137237071990967, "learning_rate": 4.814266190397946e-06, "loss": 0.4012, "num_tokens": 1579170655.0, "step": 1995 }, { "epoch": 2.723672981737498, "grad_norm": 0.09401339292526245, "learning_rate": 4.806401216545022e-06, "loss": 0.403, "num_tokens": 1580011988.0, "step": 1996 }, { "epoch": 2.7250384024577574, "grad_norm": 0.09516818821430206, "learning_rate": 4.798573540623349e-06, "loss": 0.3761, "num_tokens": 1580705371.0, "step": 1997 }, { "epoch": 2.7264038231780168, "grad_norm": 0.0952184647321701, "learning_rate": 4.790783179613446e-06, "loss": 0.389, "num_tokens": 1581452864.0, "step": 1998 }, { "epoch": 2.727769243898276, "grad_norm": 0.09491953253746033, "learning_rate": 4.783030150414873e-06, "loss": 0.4167, "num_tokens": 1582211888.0, "step": 1999 }, { "epoch": 2.7291346646185355, "grad_norm": 0.08810961246490479, "learning_rate": 4.775314469846213e-06, "loss": 0.4227, "num_tokens": 1583069974.0, "step": 2000 }, { "epoch": 2.730500085338795, "grad_norm": 0.0928787961602211, "learning_rate": 4.767636154645027e-06, "loss": 0.4044, "num_tokens": 1583861996.0, "step": 2001 }, { "epoch": 2.731865506059054, "grad_norm": 0.09371338784694672, "learning_rate": 4.759995221467824e-06, "loss": 0.4107, "num_tokens": 1584609658.0, "step": 2002 }, { "epoch": 2.733230926779314, "grad_norm": 0.0996243804693222, "learning_rate": 4.752391686890016e-06, "loss": 0.4024, "num_tokens": 1585374258.0, "step": 2003 }, { "epoch": 2.7345963474995734, "grad_norm": 0.09057795256376266, "learning_rate": 4.744825567405886e-06, "loss": 0.4253, "num_tokens": 1586256964.0, "step": 2004 }, { "epoch": 2.7359617682198327, "grad_norm": 0.09075921773910522, "learning_rate": 4.737296879428551e-06, "loss": 0.4217, "num_tokens": 1587082041.0, "step": 2005 }, { "epoch": 2.737327188940092, "grad_norm": 0.09139140695333481, "learning_rate": 4.729805639289938e-06, "loss": 0.4122, "num_tokens": 1587909953.0, "step": 2006 }, { "epoch": 2.7386926096603514, "grad_norm": 0.09419417381286621, "learning_rate": 4.7223518632407304e-06, "loss": 0.4113, "num_tokens": 1588735122.0, "step": 2007 }, { "epoch": 2.7400580303806112, "grad_norm": 0.09040486812591553, "learning_rate": 4.714935567450339e-06, "loss": 0.4275, "num_tokens": 1589595156.0, "step": 2008 }, { "epoch": 2.7414234511008706, "grad_norm": 0.08999887108802795, "learning_rate": 4.7075567680068766e-06, "loss": 0.3847, "num_tokens": 1590388813.0, "step": 2009 }, { "epoch": 2.74278887182113, "grad_norm": 0.09301687031984329, "learning_rate": 4.700215480917109e-06, "loss": 0.3837, "num_tokens": 1591118640.0, "step": 2010 }, { "epoch": 2.7441542925413893, "grad_norm": 0.09677398949861526, "learning_rate": 4.692911722106433e-06, "loss": 0.4011, "num_tokens": 1591870386.0, "step": 2011 }, { "epoch": 2.7455197132616487, "grad_norm": 0.09561135619878769, "learning_rate": 4.685645507418829e-06, "loss": 0.3985, "num_tokens": 1592619405.0, "step": 2012 }, { "epoch": 2.746885133981908, "grad_norm": 0.08917643129825592, "learning_rate": 4.678416852616837e-06, "loss": 0.3828, "num_tokens": 1593377849.0, "step": 2013 }, { "epoch": 2.7482505547021674, "grad_norm": 0.09636924415826797, "learning_rate": 4.671225773381514e-06, "loss": 0.3871, "num_tokens": 1594145767.0, "step": 2014 }, { "epoch": 2.749615975422427, "grad_norm": 0.08828398585319519, "learning_rate": 4.664072285312418e-06, "loss": 0.3843, "num_tokens": 1594937059.0, "step": 2015 }, { "epoch": 2.7509813961426866, "grad_norm": 0.08945630490779877, "learning_rate": 4.656956403927546e-06, "loss": 0.4032, "num_tokens": 1595875001.0, "step": 2016 }, { "epoch": 2.752346816862946, "grad_norm": 0.08904142677783966, "learning_rate": 4.649878144663317e-06, "loss": 0.4154, "num_tokens": 1596733404.0, "step": 2017 }, { "epoch": 2.7537122375832053, "grad_norm": 0.09310203045606613, "learning_rate": 4.642837522874542e-06, "loss": 0.4225, "num_tokens": 1597639120.0, "step": 2018 }, { "epoch": 2.7550776583034646, "grad_norm": 0.09172581881284714, "learning_rate": 4.635834553834389e-06, "loss": 0.4049, "num_tokens": 1598434991.0, "step": 2019 }, { "epoch": 2.7564430790237244, "grad_norm": 0.09331541508436203, "learning_rate": 4.628869252734334e-06, "loss": 0.4183, "num_tokens": 1599240307.0, "step": 2020 }, { "epoch": 2.757808499743984, "grad_norm": 0.09317054599523544, "learning_rate": 4.621941634684151e-06, "loss": 0.3919, "num_tokens": 1600031957.0, "step": 2021 }, { "epoch": 2.759173920464243, "grad_norm": 0.09479717165231705, "learning_rate": 4.615051714711862e-06, "loss": 0.4037, "num_tokens": 1600774714.0, "step": 2022 }, { "epoch": 2.7605393411845025, "grad_norm": 0.09513933956623077, "learning_rate": 4.608199507763714e-06, "loss": 0.4232, "num_tokens": 1601611106.0, "step": 2023 }, { "epoch": 2.761904761904762, "grad_norm": 0.09755943715572357, "learning_rate": 4.6013850287041444e-06, "loss": 0.4202, "num_tokens": 1602451588.0, "step": 2024 }, { "epoch": 2.763270182625021, "grad_norm": 0.08889942616224289, "learning_rate": 4.594608292315743e-06, "loss": 0.3954, "num_tokens": 1603324342.0, "step": 2025 }, { "epoch": 2.7646356033452806, "grad_norm": 0.09201060980558395, "learning_rate": 4.5878693132992285e-06, "loss": 0.3991, "num_tokens": 1604063459.0, "step": 2026 }, { "epoch": 2.76600102406554, "grad_norm": 0.09707494080066681, "learning_rate": 4.581168106273413e-06, "loss": 0.3962, "num_tokens": 1604814176.0, "step": 2027 }, { "epoch": 2.7673664447857997, "grad_norm": 0.09260115772485733, "learning_rate": 4.574504685775173e-06, "loss": 0.4063, "num_tokens": 1605661321.0, "step": 2028 }, { "epoch": 2.768731865506059, "grad_norm": 0.09111277014017105, "learning_rate": 4.5678790662594055e-06, "loss": 0.3983, "num_tokens": 1606531858.0, "step": 2029 }, { "epoch": 2.7700972862263185, "grad_norm": 0.08657192438840866, "learning_rate": 4.5612912620990145e-06, "loss": 0.4201, "num_tokens": 1607369362.0, "step": 2030 }, { "epoch": 2.771462706946578, "grad_norm": 0.09224317967891693, "learning_rate": 4.5547412875848645e-06, "loss": 0.4118, "num_tokens": 1608190617.0, "step": 2031 }, { "epoch": 2.7728281276668376, "grad_norm": 0.09337349236011505, "learning_rate": 4.548229156925768e-06, "loss": 0.4156, "num_tokens": 1609030832.0, "step": 2032 }, { "epoch": 2.774193548387097, "grad_norm": 0.08948948979377747, "learning_rate": 4.541754884248436e-06, "loss": 0.4285, "num_tokens": 1609917716.0, "step": 2033 }, { "epoch": 2.7755589691073563, "grad_norm": 0.0859433114528656, "learning_rate": 4.535318483597449e-06, "loss": 0.4117, "num_tokens": 1610745993.0, "step": 2034 }, { "epoch": 2.7769243898276157, "grad_norm": 0.09645164012908936, "learning_rate": 4.52891996893524e-06, "loss": 0.3857, "num_tokens": 1611456094.0, "step": 2035 }, { "epoch": 2.778289810547875, "grad_norm": 0.09129565954208374, "learning_rate": 4.522559354142054e-06, "loss": 0.3978, "num_tokens": 1612211571.0, "step": 2036 }, { "epoch": 2.7796552312681344, "grad_norm": 0.09490007162094116, "learning_rate": 4.51623665301592e-06, "loss": 0.3905, "num_tokens": 1613006719.0, "step": 2037 }, { "epoch": 2.7810206519883938, "grad_norm": 0.0881430059671402, "learning_rate": 4.50995187927262e-06, "loss": 0.3981, "num_tokens": 1613829458.0, "step": 2038 }, { "epoch": 2.782386072708653, "grad_norm": 0.09500205516815186, "learning_rate": 4.503705046545664e-06, "loss": 0.4041, "num_tokens": 1614634838.0, "step": 2039 }, { "epoch": 2.783751493428913, "grad_norm": 0.09084132313728333, "learning_rate": 4.497496168386252e-06, "loss": 0.3816, "num_tokens": 1615427165.0, "step": 2040 }, { "epoch": 2.7851169141491723, "grad_norm": 0.08830274641513824, "learning_rate": 4.491325258263255e-06, "loss": 0.4055, "num_tokens": 1616196805.0, "step": 2041 }, { "epoch": 2.7864823348694316, "grad_norm": 0.09827621281147003, "learning_rate": 4.485192329563172e-06, "loss": 0.4139, "num_tokens": 1616985104.0, "step": 2042 }, { "epoch": 2.787847755589691, "grad_norm": 0.09381141513586044, "learning_rate": 4.479097395590118e-06, "loss": 0.406, "num_tokens": 1617769020.0, "step": 2043 }, { "epoch": 2.7892131763099504, "grad_norm": 0.09323320537805557, "learning_rate": 4.4730404695657845e-06, "loss": 0.4135, "num_tokens": 1618598837.0, "step": 2044 }, { "epoch": 2.79057859703021, "grad_norm": 0.09418671578168869, "learning_rate": 4.467021564629407e-06, "loss": 0.3987, "num_tokens": 1619424184.0, "step": 2045 }, { "epoch": 2.7919440177504695, "grad_norm": 0.09328383207321167, "learning_rate": 4.461040693837746e-06, "loss": 0.4137, "num_tokens": 1620226696.0, "step": 2046 }, { "epoch": 2.793309438470729, "grad_norm": 0.09878713637590408, "learning_rate": 4.455097870165058e-06, "loss": 0.4215, "num_tokens": 1621053327.0, "step": 2047 }, { "epoch": 2.7946748591909882, "grad_norm": 0.09423433989286423, "learning_rate": 4.449193106503059e-06, "loss": 0.4029, "num_tokens": 1621848103.0, "step": 2048 }, { "epoch": 2.7960402799112476, "grad_norm": 0.09238555282354355, "learning_rate": 4.4433264156609035e-06, "loss": 0.4091, "num_tokens": 1622612972.0, "step": 2049 }, { "epoch": 2.797405700631507, "grad_norm": 0.0946132019162178, "learning_rate": 4.437497810365158e-06, "loss": 0.4146, "num_tokens": 1623418948.0, "step": 2050 }, { "epoch": 2.7987711213517663, "grad_norm": 0.09438643604516983, "learning_rate": 4.431707303259766e-06, "loss": 0.4055, "num_tokens": 1624145656.0, "step": 2051 }, { "epoch": 2.8001365420720257, "grad_norm": 0.09127840399742126, "learning_rate": 4.425954906906022e-06, "loss": 0.4064, "num_tokens": 1624914231.0, "step": 2052 }, { "epoch": 2.8015019627922855, "grad_norm": 0.09927979856729507, "learning_rate": 4.420240633782562e-06, "loss": 0.4377, "num_tokens": 1625750164.0, "step": 2053 }, { "epoch": 2.802867383512545, "grad_norm": 0.09673245996236801, "learning_rate": 4.414564496285303e-06, "loss": 0.4342, "num_tokens": 1626530965.0, "step": 2054 }, { "epoch": 2.804232804232804, "grad_norm": 0.0999177023768425, "learning_rate": 4.408926506727444e-06, "loss": 0.3983, "num_tokens": 1627352395.0, "step": 2055 }, { "epoch": 2.8055982249530635, "grad_norm": 0.08903316408395767, "learning_rate": 4.403326677339432e-06, "loss": 0.3998, "num_tokens": 1628148642.0, "step": 2056 }, { "epoch": 2.8069636456733233, "grad_norm": 0.08505116403102875, "learning_rate": 4.397765020268929e-06, "loss": 0.3902, "num_tokens": 1629027708.0, "step": 2057 }, { "epoch": 2.8083290663935827, "grad_norm": 0.09638615697622299, "learning_rate": 4.392241547580792e-06, "loss": 0.4069, "num_tokens": 1629810178.0, "step": 2058 }, { "epoch": 2.809694487113842, "grad_norm": 0.09382128715515137, "learning_rate": 4.386756271257042e-06, "loss": 0.4074, "num_tokens": 1630624847.0, "step": 2059 }, { "epoch": 2.8110599078341014, "grad_norm": 0.08805206418037415, "learning_rate": 4.381309203196847e-06, "loss": 0.396, "num_tokens": 1631419698.0, "step": 2060 }, { "epoch": 2.812425328554361, "grad_norm": 0.08849799633026123, "learning_rate": 4.3759003552164834e-06, "loss": 0.39, "num_tokens": 1632214402.0, "step": 2061 }, { "epoch": 2.81379074927462, "grad_norm": 0.09452936798334122, "learning_rate": 4.3705297390493205e-06, "loss": 0.4074, "num_tokens": 1633037954.0, "step": 2062 }, { "epoch": 2.8151561699948795, "grad_norm": 0.09069782495498657, "learning_rate": 4.365197366345796e-06, "loss": 0.4116, "num_tokens": 1633906816.0, "step": 2063 }, { "epoch": 2.816521590715139, "grad_norm": 0.10188496112823486, "learning_rate": 4.359903248673378e-06, "loss": 0.3975, "num_tokens": 1634701822.0, "step": 2064 }, { "epoch": 2.8178870114353987, "grad_norm": 0.09093333780765533, "learning_rate": 4.354647397516551e-06, "loss": 0.421, "num_tokens": 1635485228.0, "step": 2065 }, { "epoch": 2.819252432155658, "grad_norm": 0.09407094866037369, "learning_rate": 4.349429824276789e-06, "loss": 0.3973, "num_tokens": 1636262854.0, "step": 2066 }, { "epoch": 2.8206178528759174, "grad_norm": 0.09620767831802368, "learning_rate": 4.344250540272537e-06, "loss": 0.4048, "num_tokens": 1637049203.0, "step": 2067 }, { "epoch": 2.8219832735961767, "grad_norm": 0.09660089761018753, "learning_rate": 4.339109556739164e-06, "loss": 0.3946, "num_tokens": 1637795380.0, "step": 2068 }, { "epoch": 2.823348694316436, "grad_norm": 0.09138870984315872, "learning_rate": 4.334006884828971e-06, "loss": 0.4267, "num_tokens": 1638625099.0, "step": 2069 }, { "epoch": 2.824714115036696, "grad_norm": 0.09680335968732834, "learning_rate": 4.328942535611134e-06, "loss": 0.4157, "num_tokens": 1639424844.0, "step": 2070 }, { "epoch": 2.8260795357569553, "grad_norm": 0.11171326041221619, "learning_rate": 4.323916520071709e-06, "loss": 0.3928, "num_tokens": 1640115370.0, "step": 2071 }, { "epoch": 2.8274449564772146, "grad_norm": 0.1020292118191719, "learning_rate": 4.318928849113587e-06, "loss": 0.3957, "num_tokens": 1640843999.0, "step": 2072 }, { "epoch": 2.828810377197474, "grad_norm": 0.09293735772371292, "learning_rate": 4.313979533556485e-06, "loss": 0.3918, "num_tokens": 1641623960.0, "step": 2073 }, { "epoch": 2.8301757979177333, "grad_norm": 0.0906636044383049, "learning_rate": 4.309068584136909e-06, "loss": 0.397, "num_tokens": 1642414713.0, "step": 2074 }, { "epoch": 2.8315412186379927, "grad_norm": 0.0924590453505516, "learning_rate": 4.304196011508142e-06, "loss": 0.3898, "num_tokens": 1643200280.0, "step": 2075 }, { "epoch": 2.832906639358252, "grad_norm": 0.0928448885679245, "learning_rate": 4.299361826240214e-06, "loss": 0.3766, "num_tokens": 1643990454.0, "step": 2076 }, { "epoch": 2.834272060078512, "grad_norm": 0.09456855803728104, "learning_rate": 4.2945660388198866e-06, "loss": 0.415, "num_tokens": 1644708896.0, "step": 2077 }, { "epoch": 2.835637480798771, "grad_norm": 0.10004931688308716, "learning_rate": 4.289808659650617e-06, "loss": 0.3979, "num_tokens": 1645539266.0, "step": 2078 }, { "epoch": 2.8370029015190306, "grad_norm": 0.08817711472511292, "learning_rate": 4.285089699052549e-06, "loss": 0.4026, "num_tokens": 1646352931.0, "step": 2079 }, { "epoch": 2.83836832223929, "grad_norm": 0.09613456577062607, "learning_rate": 4.28040916726249e-06, "loss": 0.378, "num_tokens": 1647087485.0, "step": 2080 }, { "epoch": 2.8397337429595493, "grad_norm": 0.09716501086950302, "learning_rate": 4.275767074433873e-06, "loss": 0.3961, "num_tokens": 1647777620.0, "step": 2081 }, { "epoch": 2.841099163679809, "grad_norm": 0.09134925156831741, "learning_rate": 4.271163430636755e-06, "loss": 0.4045, "num_tokens": 1648511710.0, "step": 2082 }, { "epoch": 2.8424645844000684, "grad_norm": 0.11628374457359314, "learning_rate": 4.2665982458577795e-06, "loss": 0.4195, "num_tokens": 1649334843.0, "step": 2083 }, { "epoch": 2.843830005120328, "grad_norm": 0.0996968224644661, "learning_rate": 4.262071530000165e-06, "loss": 0.4293, "num_tokens": 1650222339.0, "step": 2084 }, { "epoch": 2.845195425840587, "grad_norm": 0.09137681126594543, "learning_rate": 4.2575832928836795e-06, "loss": 0.3968, "num_tokens": 1651003198.0, "step": 2085 }, { "epoch": 2.8465608465608465, "grad_norm": 0.09371490776538849, "learning_rate": 4.253133544244619e-06, "loss": 0.4049, "num_tokens": 1651763189.0, "step": 2086 }, { "epoch": 2.847926267281106, "grad_norm": 0.09612946212291718, "learning_rate": 4.2487222937357844e-06, "loss": 0.4072, "num_tokens": 1652508824.0, "step": 2087 }, { "epoch": 2.8492916880013652, "grad_norm": 0.09517689794301987, "learning_rate": 4.2443495509264666e-06, "loss": 0.3858, "num_tokens": 1653246535.0, "step": 2088 }, { "epoch": 2.8506571087216246, "grad_norm": 0.08995765447616577, "learning_rate": 4.2400153253024175e-06, "loss": 0.4027, "num_tokens": 1654071255.0, "step": 2089 }, { "epoch": 2.8520225294418844, "grad_norm": 0.09080367535352707, "learning_rate": 4.235719626265841e-06, "loss": 0.3971, "num_tokens": 1654872631.0, "step": 2090 }, { "epoch": 2.8533879501621437, "grad_norm": 0.09881149232387543, "learning_rate": 4.2314624631353575e-06, "loss": 0.4127, "num_tokens": 1655625565.0, "step": 2091 }, { "epoch": 2.854753370882403, "grad_norm": 0.09460882097482681, "learning_rate": 4.227243845146e-06, "loss": 0.3938, "num_tokens": 1656424783.0, "step": 2092 }, { "epoch": 2.8561187916026625, "grad_norm": 0.09198158234357834, "learning_rate": 4.223063781449177e-06, "loss": 0.4232, "num_tokens": 1657248064.0, "step": 2093 }, { "epoch": 2.8574842123229223, "grad_norm": 0.09066101163625717, "learning_rate": 4.218922281112668e-06, "loss": 0.4073, "num_tokens": 1658111288.0, "step": 2094 }, { "epoch": 2.8588496330431816, "grad_norm": 0.09621741622686386, "learning_rate": 4.214819353120597e-06, "loss": 0.3884, "num_tokens": 1658962442.0, "step": 2095 }, { "epoch": 2.860215053763441, "grad_norm": 0.08896927535533905, "learning_rate": 4.210755006373406e-06, "loss": 0.4087, "num_tokens": 1659802633.0, "step": 2096 }, { "epoch": 2.8615804744837003, "grad_norm": 0.08354035019874573, "learning_rate": 4.2067292496878556e-06, "loss": 0.43, "num_tokens": 1660718232.0, "step": 2097 }, { "epoch": 2.8629458952039597, "grad_norm": 0.09707864373922348, "learning_rate": 4.202742091796985e-06, "loss": 0.4011, "num_tokens": 1661510885.0, "step": 2098 }, { "epoch": 2.864311315924219, "grad_norm": 0.09442807734012604, "learning_rate": 4.198793541350099e-06, "loss": 0.4121, "num_tokens": 1662363145.0, "step": 2099 }, { "epoch": 2.8656767366444784, "grad_norm": 0.09239194542169571, "learning_rate": 4.194883606912757e-06, "loss": 0.4045, "num_tokens": 1663105905.0, "step": 2100 }, { "epoch": 2.8670421573647378, "grad_norm": 0.09297695010900497, "learning_rate": 4.191012296966747e-06, "loss": 0.4089, "num_tokens": 1663971513.0, "step": 2101 }, { "epoch": 2.8684075780849976, "grad_norm": 0.08861865103244781, "learning_rate": 4.187179619910074e-06, "loss": 0.4108, "num_tokens": 1664794511.0, "step": 2102 }, { "epoch": 2.869772998805257, "grad_norm": 0.09513044357299805, "learning_rate": 4.183385584056927e-06, "loss": 0.4081, "num_tokens": 1665578051.0, "step": 2103 }, { "epoch": 2.8711384195255163, "grad_norm": 0.09409821778535843, "learning_rate": 4.1796301976376785e-06, "loss": 0.4315, "num_tokens": 1666449291.0, "step": 2104 }, { "epoch": 2.8725038402457757, "grad_norm": 0.09183581173419952, "learning_rate": 4.175913468798858e-06, "loss": 0.4048, "num_tokens": 1667238143.0, "step": 2105 }, { "epoch": 2.873869260966035, "grad_norm": 0.09400010854005814, "learning_rate": 4.172235405603132e-06, "loss": 0.3951, "num_tokens": 1667995602.0, "step": 2106 }, { "epoch": 2.875234681686295, "grad_norm": 0.0950804203748703, "learning_rate": 4.168596016029295e-06, "loss": 0.4174, "num_tokens": 1668839510.0, "step": 2107 }, { "epoch": 2.876600102406554, "grad_norm": 0.09047482162714005, "learning_rate": 4.164995307972244e-06, "loss": 0.4091, "num_tokens": 1669647788.0, "step": 2108 }, { "epoch": 2.8779655231268135, "grad_norm": 0.08775763213634491, "learning_rate": 4.161433289242963e-06, "loss": 0.415, "num_tokens": 1670454175.0, "step": 2109 }, { "epoch": 2.879330943847073, "grad_norm": 0.09450545161962509, "learning_rate": 4.157909967568514e-06, "loss": 0.4069, "num_tokens": 1671243937.0, "step": 2110 }, { "epoch": 2.8806963645673322, "grad_norm": 0.09100879728794098, "learning_rate": 4.154425350592006e-06, "loss": 0.4024, "num_tokens": 1672014627.0, "step": 2111 }, { "epoch": 2.8820617852875916, "grad_norm": 0.09718020260334015, "learning_rate": 4.150979445872593e-06, "loss": 0.3991, "num_tokens": 1672809026.0, "step": 2112 }, { "epoch": 2.883427206007851, "grad_norm": 0.08769235759973526, "learning_rate": 4.1475722608854405e-06, "loss": 0.3933, "num_tokens": 1673566281.0, "step": 2113 }, { "epoch": 2.8847926267281108, "grad_norm": 0.09560836106538773, "learning_rate": 4.144203803021736e-06, "loss": 0.3984, "num_tokens": 1674360354.0, "step": 2114 }, { "epoch": 2.88615804744837, "grad_norm": 0.09532773494720459, "learning_rate": 4.140874079588641e-06, "loss": 0.4015, "num_tokens": 1675158182.0, "step": 2115 }, { "epoch": 2.8875234681686295, "grad_norm": 0.09468579292297363, "learning_rate": 4.137583097809299e-06, "loss": 0.3949, "num_tokens": 1675951277.0, "step": 2116 }, { "epoch": 2.888888888888889, "grad_norm": 0.09642697125673294, "learning_rate": 4.134330864822808e-06, "loss": 0.3996, "num_tokens": 1676751624.0, "step": 2117 }, { "epoch": 2.890254309609148, "grad_norm": 0.0840301364660263, "learning_rate": 4.131117387684213e-06, "loss": 0.401, "num_tokens": 1677535455.0, "step": 2118 }, { "epoch": 2.891619730329408, "grad_norm": 0.09197047352790833, "learning_rate": 4.127942673364479e-06, "loss": 0.3859, "num_tokens": 1678286661.0, "step": 2119 }, { "epoch": 2.8929851510496674, "grad_norm": 0.09475096315145493, "learning_rate": 4.124806728750492e-06, "loss": 0.4007, "num_tokens": 1679045179.0, "step": 2120 }, { "epoch": 2.8943505717699267, "grad_norm": 0.09494233876466751, "learning_rate": 4.121709560645026e-06, "loss": 0.4104, "num_tokens": 1679839310.0, "step": 2121 }, { "epoch": 2.895715992490186, "grad_norm": 0.08756506443023682, "learning_rate": 4.118651175766747e-06, "loss": 0.405, "num_tokens": 1680599507.0, "step": 2122 }, { "epoch": 2.8970814132104454, "grad_norm": 0.10284225642681122, "learning_rate": 4.115631580750181e-06, "loss": 0.4284, "num_tokens": 1681327300.0, "step": 2123 }, { "epoch": 2.898446833930705, "grad_norm": 0.09609955549240112, "learning_rate": 4.112650782145711e-06, "loss": 0.406, "num_tokens": 1682163714.0, "step": 2124 }, { "epoch": 2.899812254650964, "grad_norm": 0.09525071084499359, "learning_rate": 4.1097087864195576e-06, "loss": 0.405, "num_tokens": 1682951560.0, "step": 2125 }, { "epoch": 2.9011776753712235, "grad_norm": 0.09249138087034225, "learning_rate": 4.106805599953769e-06, "loss": 0.4113, "num_tokens": 1683701534.0, "step": 2126 }, { "epoch": 2.9025430960914833, "grad_norm": 0.09081511944532394, "learning_rate": 4.103941229046205e-06, "loss": 0.3984, "num_tokens": 1684492435.0, "step": 2127 }, { "epoch": 2.9039085168117427, "grad_norm": 0.08997368812561035, "learning_rate": 4.101115679910519e-06, "loss": 0.4033, "num_tokens": 1685308869.0, "step": 2128 }, { "epoch": 2.905273937532002, "grad_norm": 0.0954335555434227, "learning_rate": 4.098328958676152e-06, "loss": 0.4278, "num_tokens": 1686142500.0, "step": 2129 }, { "epoch": 2.9066393582522614, "grad_norm": 0.08735547214746475, "learning_rate": 4.095581071388315e-06, "loss": 0.3955, "num_tokens": 1686974546.0, "step": 2130 }, { "epoch": 2.9080047789725207, "grad_norm": 0.09224262088537216, "learning_rate": 4.092872024007979e-06, "loss": 0.4115, "num_tokens": 1687727196.0, "step": 2131 }, { "epoch": 2.9093701996927805, "grad_norm": 0.10246105492115021, "learning_rate": 4.090201822411856e-06, "loss": 0.3979, "num_tokens": 1688493639.0, "step": 2132 }, { "epoch": 2.91073562041304, "grad_norm": 0.09128902852535248, "learning_rate": 4.087570472392394e-06, "loss": 0.3836, "num_tokens": 1689249823.0, "step": 2133 }, { "epoch": 2.9121010411332993, "grad_norm": 0.09504935890436172, "learning_rate": 4.084977979657756e-06, "loss": 0.3887, "num_tokens": 1690059560.0, "step": 2134 }, { "epoch": 2.9134664618535586, "grad_norm": 0.09151072800159454, "learning_rate": 4.082424349831818e-06, "loss": 0.4114, "num_tokens": 1690828208.0, "step": 2135 }, { "epoch": 2.914831882573818, "grad_norm": 0.09162576496601105, "learning_rate": 4.079909588454145e-06, "loss": 0.4275, "num_tokens": 1691684162.0, "step": 2136 }, { "epoch": 2.9161973032940773, "grad_norm": 0.10094135254621506, "learning_rate": 4.077433700979991e-06, "loss": 0.4037, "num_tokens": 1692517952.0, "step": 2137 }, { "epoch": 2.9175627240143367, "grad_norm": 0.09236815571784973, "learning_rate": 4.074996692780277e-06, "loss": 0.4141, "num_tokens": 1693340282.0, "step": 2138 }, { "epoch": 2.9189281447345965, "grad_norm": 0.09626796841621399, "learning_rate": 4.072598569141582e-06, "loss": 0.4171, "num_tokens": 1694142182.0, "step": 2139 }, { "epoch": 2.920293565454856, "grad_norm": 0.09043467789888382, "learning_rate": 4.070239335266142e-06, "loss": 0.3704, "num_tokens": 1694923772.0, "step": 2140 }, { "epoch": 2.921658986175115, "grad_norm": 0.08931101113557816, "learning_rate": 4.067918996271815e-06, "loss": 0.412, "num_tokens": 1695667705.0, "step": 2141 }, { "epoch": 2.9230244068953746, "grad_norm": 0.08742230385541916, "learning_rate": 4.0656375571921006e-06, "loss": 0.3854, "num_tokens": 1696553776.0, "step": 2142 }, { "epoch": 2.924389827615634, "grad_norm": 0.08221264183521271, "learning_rate": 4.063395022976103e-06, "loss": 0.3978, "num_tokens": 1697315471.0, "step": 2143 }, { "epoch": 2.9257552483358937, "grad_norm": 0.11607661843299866, "learning_rate": 4.061191398488533e-06, "loss": 0.3918, "num_tokens": 1698081277.0, "step": 2144 }, { "epoch": 2.927120669056153, "grad_norm": 0.09390314668416977, "learning_rate": 4.059026688509696e-06, "loss": 0.3985, "num_tokens": 1698831346.0, "step": 2145 }, { "epoch": 2.9284860897764124, "grad_norm": 0.09260386228561401, "learning_rate": 4.0569008977354756e-06, "loss": 0.4116, "num_tokens": 1699651656.0, "step": 2146 }, { "epoch": 2.929851510496672, "grad_norm": 0.0972478985786438, "learning_rate": 4.054814030777336e-06, "loss": 0.3859, "num_tokens": 1700380507.0, "step": 2147 }, { "epoch": 2.931216931216931, "grad_norm": 0.09140291064977646, "learning_rate": 4.052766092162299e-06, "loss": 0.4289, "num_tokens": 1701208034.0, "step": 2148 }, { "epoch": 2.9325823519371905, "grad_norm": 0.08816833049058914, "learning_rate": 4.050757086332941e-06, "loss": 0.3988, "num_tokens": 1701992867.0, "step": 2149 }, { "epoch": 2.93394777265745, "grad_norm": 0.095639668405056, "learning_rate": 4.048787017647381e-06, "loss": 0.4246, "num_tokens": 1702791943.0, "step": 2150 }, { "epoch": 2.9353131933777092, "grad_norm": 0.09317848831415176, "learning_rate": 4.046855890379276e-06, "loss": 0.4067, "num_tokens": 1703597518.0, "step": 2151 }, { "epoch": 2.936678614097969, "grad_norm": 0.09753479808568954, "learning_rate": 4.0449637087178e-06, "loss": 0.3954, "num_tokens": 1704359180.0, "step": 2152 }, { "epoch": 2.9380440348182284, "grad_norm": 0.09088147431612015, "learning_rate": 4.043110476767649e-06, "loss": 0.3851, "num_tokens": 1705051695.0, "step": 2153 }, { "epoch": 2.9394094555384878, "grad_norm": 0.09331274032592773, "learning_rate": 4.041296198549024e-06, "loss": 0.3943, "num_tokens": 1705792572.0, "step": 2154 }, { "epoch": 2.940774876258747, "grad_norm": 0.09522411227226257, "learning_rate": 4.0395208779976204e-06, "loss": 0.4051, "num_tokens": 1706623182.0, "step": 2155 }, { "epoch": 2.942140296979007, "grad_norm": 0.08710099011659622, "learning_rate": 4.037784518964631e-06, "loss": 0.3987, "num_tokens": 1707457312.0, "step": 2156 }, { "epoch": 2.9435057176992663, "grad_norm": 0.0934676080942154, "learning_rate": 4.0360871252167235e-06, "loss": 0.3854, "num_tokens": 1708194415.0, "step": 2157 }, { "epoch": 2.9448711384195256, "grad_norm": 0.08873570710420609, "learning_rate": 4.03442870043604e-06, "loss": 0.3997, "num_tokens": 1708998796.0, "step": 2158 }, { "epoch": 2.946236559139785, "grad_norm": 0.09487366676330566, "learning_rate": 4.032809248220187e-06, "loss": 0.41, "num_tokens": 1709861251.0, "step": 2159 }, { "epoch": 2.9476019798600444, "grad_norm": 0.09194494783878326, "learning_rate": 4.031228772082228e-06, "loss": 0.4066, "num_tokens": 1710645651.0, "step": 2160 }, { "epoch": 2.9489674005803037, "grad_norm": 0.09275718778371811, "learning_rate": 4.0296872754506785e-06, "loss": 0.4176, "num_tokens": 1711455263.0, "step": 2161 }, { "epoch": 2.950332821300563, "grad_norm": 0.08732987940311432, "learning_rate": 4.028184761669494e-06, "loss": 0.3923, "num_tokens": 1712205340.0, "step": 2162 }, { "epoch": 2.9516982420208224, "grad_norm": 0.09425903111696243, "learning_rate": 4.026721233998066e-06, "loss": 0.3994, "num_tokens": 1712971039.0, "step": 2163 }, { "epoch": 2.9530636627410822, "grad_norm": 0.09121242165565491, "learning_rate": 4.02529669561121e-06, "loss": 0.3987, "num_tokens": 1713756974.0, "step": 2164 }, { "epoch": 2.9544290834613416, "grad_norm": 0.09776400029659271, "learning_rate": 4.023911149599168e-06, "loss": 0.4156, "num_tokens": 1714520674.0, "step": 2165 }, { "epoch": 2.955794504181601, "grad_norm": 0.09825741499662399, "learning_rate": 4.022564598967591e-06, "loss": 0.4019, "num_tokens": 1715338937.0, "step": 2166 }, { "epoch": 2.9571599249018603, "grad_norm": 0.09575600922107697, "learning_rate": 4.021257046637543e-06, "loss": 0.42, "num_tokens": 1716076794.0, "step": 2167 }, { "epoch": 2.9585253456221197, "grad_norm": 0.10192405432462692, "learning_rate": 4.019988495445487e-06, "loss": 0.4027, "num_tokens": 1716892891.0, "step": 2168 }, { "epoch": 2.9598907663423795, "grad_norm": 0.09007620066404343, "learning_rate": 4.018758948143276e-06, "loss": 0.3852, "num_tokens": 1717663195.0, "step": 2169 }, { "epoch": 2.961256187062639, "grad_norm": 0.08797558397054672, "learning_rate": 4.017568407398161e-06, "loss": 0.3809, "num_tokens": 1718386031.0, "step": 2170 }, { "epoch": 2.962621607782898, "grad_norm": 0.0926051139831543, "learning_rate": 4.016416875792769e-06, "loss": 0.4085, "num_tokens": 1719146222.0, "step": 2171 }, { "epoch": 2.9639870285031575, "grad_norm": 0.09780365973711014, "learning_rate": 4.015304355825112e-06, "loss": 0.3946, "num_tokens": 1719918485.0, "step": 2172 }, { "epoch": 2.965352449223417, "grad_norm": 0.09424018114805222, "learning_rate": 4.014230849908567e-06, "loss": 0.4113, "num_tokens": 1720692597.0, "step": 2173 }, { "epoch": 2.9667178699436763, "grad_norm": 0.09893258661031723, "learning_rate": 4.0131963603718825e-06, "loss": 0.3804, "num_tokens": 1721426604.0, "step": 2174 }, { "epoch": 2.9680832906639356, "grad_norm": 0.10091055184602737, "learning_rate": 4.012200889459169e-06, "loss": 0.4249, "num_tokens": 1722280213.0, "step": 2175 }, { "epoch": 2.9694487113841954, "grad_norm": 0.09061630070209503, "learning_rate": 4.01124443932989e-06, "loss": 0.4026, "num_tokens": 1723103374.0, "step": 2176 }, { "epoch": 2.9708141321044548, "grad_norm": 0.0957452654838562, "learning_rate": 4.01032701205887e-06, "loss": 0.3943, "num_tokens": 1723862459.0, "step": 2177 }, { "epoch": 2.972179552824714, "grad_norm": 0.09087466448545456, "learning_rate": 4.0094486096362705e-06, "loss": 0.4166, "num_tokens": 1724626937.0, "step": 2178 }, { "epoch": 2.9735449735449735, "grad_norm": 0.09446859359741211, "learning_rate": 4.008609233967607e-06, "loss": 0.4045, "num_tokens": 1725453645.0, "step": 2179 }, { "epoch": 2.974910394265233, "grad_norm": 0.08850158005952835, "learning_rate": 4.007808886873731e-06, "loss": 0.3804, "num_tokens": 1726229406.0, "step": 2180 }, { "epoch": 2.9762758149854927, "grad_norm": 0.09173580259084702, "learning_rate": 4.007047570090823e-06, "loss": 0.4061, "num_tokens": 1727025429.0, "step": 2181 }, { "epoch": 2.977641235705752, "grad_norm": 0.09286835789680481, "learning_rate": 4.006325285270407e-06, "loss": 0.419, "num_tokens": 1727929729.0, "step": 2182 }, { "epoch": 2.9790066564260114, "grad_norm": 0.09282168745994568, "learning_rate": 4.0056420339793265e-06, "loss": 0.3955, "num_tokens": 1728716985.0, "step": 2183 }, { "epoch": 2.9803720771462707, "grad_norm": 0.09299791604280472, "learning_rate": 4.004997817699755e-06, "loss": 0.4066, "num_tokens": 1729525589.0, "step": 2184 }, { "epoch": 2.98173749786653, "grad_norm": 0.09397131949663162, "learning_rate": 4.004392637829182e-06, "loss": 0.4023, "num_tokens": 1730283312.0, "step": 2185 }, { "epoch": 2.9831029185867894, "grad_norm": 0.0946667492389679, "learning_rate": 4.003826495680423e-06, "loss": 0.4009, "num_tokens": 1731154422.0, "step": 2186 }, { "epoch": 2.984468339307049, "grad_norm": 0.08708909153938293, "learning_rate": 4.003299392481602e-06, "loss": 0.4132, "num_tokens": 1732022738.0, "step": 2187 }, { "epoch": 2.985833760027308, "grad_norm": 0.09451030194759369, "learning_rate": 4.002811329376164e-06, "loss": 0.4064, "num_tokens": 1732813003.0, "step": 2188 }, { "epoch": 2.987199180747568, "grad_norm": 0.08798997104167938, "learning_rate": 4.002362307422857e-06, "loss": 0.3884, "num_tokens": 1733578594.0, "step": 2189 }, { "epoch": 2.9885646014678273, "grad_norm": 0.09080372750759125, "learning_rate": 4.00195232759574e-06, "loss": 0.4092, "num_tokens": 1734401521.0, "step": 2190 }, { "epoch": 2.9899300221880867, "grad_norm": 0.09099128842353821, "learning_rate": 4.00158139078418e-06, "loss": 0.4211, "num_tokens": 1735202088.0, "step": 2191 }, { "epoch": 2.991295442908346, "grad_norm": 0.09328500181436539, "learning_rate": 4.001249497792849e-06, "loss": 0.4083, "num_tokens": 1736034050.0, "step": 2192 }, { "epoch": 2.992660863628606, "grad_norm": 0.09119760245084763, "learning_rate": 4.000956649341716e-06, "loss": 0.3952, "num_tokens": 1736815019.0, "step": 2193 }, { "epoch": 2.994026284348865, "grad_norm": 0.09160342067480087, "learning_rate": 4.000702846066058e-06, "loss": 0.4057, "num_tokens": 1737625996.0, "step": 2194 }, { "epoch": 2.9953917050691246, "grad_norm": 0.0919833704829216, "learning_rate": 4.000488088516446e-06, "loss": 0.3915, "num_tokens": 1738415506.0, "step": 2195 }, { "epoch": 2.996757125789384, "grad_norm": 0.09145677089691162, "learning_rate": 4.000312377158754e-06, "loss": 0.3965, "num_tokens": 1739198066.0, "step": 2196 }, { "epoch": 2.9981225465096433, "grad_norm": 0.0916244313120842, "learning_rate": 4.000175712374149e-06, "loss": 0.4042, "num_tokens": 1739947353.0, "step": 2197 }, { "epoch": 2.9994879672299026, "grad_norm": 0.09559661149978638, "learning_rate": 4.000078094459098e-06, "loss": 0.396, "num_tokens": 1740648522.0, "step": 2198 }, { "epoch": 3.0, "grad_norm": 0.145839661359787, "learning_rate": 4.000019523625363e-06, "loss": 0.3991, "num_tokens": 1740826744.0, "step": 2199 }, { "epoch": 3.0, "step": 2199, "total_flos": 1.8348642473640722e+19, "train_loss": 0.4601271056543431, "train_runtime": 22288.4565, "train_samples_per_second": 12.616, "train_steps_per_second": 0.099 } ], "logging_steps": 1, "max_steps": 2199, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.8348642473640722e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }