diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,22687 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 3235, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0015455950540958269, + "grad_norm": 5.26565455238215, + "learning_rate": 2.469135802469136e-07, + "loss": 0.8853, + "step": 1 + }, + { + "epoch": 0.0030911901081916537, + "grad_norm": 5.282847760707683, + "learning_rate": 4.938271604938272e-07, + "loss": 0.8908, + "step": 2 + }, + { + "epoch": 0.00463678516228748, + "grad_norm": 5.179577811088595, + "learning_rate": 7.407407407407407e-07, + "loss": 0.8789, + "step": 3 + }, + { + "epoch": 0.0061823802163833074, + "grad_norm": 5.14965313767903, + "learning_rate": 9.876543209876544e-07, + "loss": 0.8745, + "step": 4 + }, + { + "epoch": 0.0077279752704791345, + "grad_norm": 5.0926621078802095, + "learning_rate": 1.234567901234568e-06, + "loss": 0.8697, + "step": 5 + }, + { + "epoch": 0.00927357032457496, + "grad_norm": 4.936823049639591, + "learning_rate": 1.4814814814814815e-06, + "loss": 0.8677, + "step": 6 + }, + { + "epoch": 0.010819165378670788, + "grad_norm": 4.800318401151413, + "learning_rate": 1.7283950617283952e-06, + "loss": 0.8667, + "step": 7 + }, + { + "epoch": 0.012364760432766615, + "grad_norm": 3.9449223561641844, + "learning_rate": 1.9753086419753087e-06, + "loss": 0.8235, + "step": 8 + }, + { + "epoch": 0.013910355486862442, + "grad_norm": 3.7105633047486504, + "learning_rate": 2.222222222222222e-06, + "loss": 0.8259, + "step": 9 + }, + { + "epoch": 0.015455950540958269, + "grad_norm": 2.198523539340047, + "learning_rate": 2.469135802469136e-06, + "loss": 0.7827, + "step": 10 + }, + { + "epoch": 0.017001545595054096, + "grad_norm": 2.0963348737509295, + "learning_rate": 2.7160493827160496e-06, + "loss": 0.776, + "step": 11 + }, + { + "epoch": 0.01854714064914992, + "grad_norm": 1.8699088033519315, + "learning_rate": 2.962962962962963e-06, + "loss": 0.7768, + "step": 12 + }, + { + "epoch": 0.02009273570324575, + "grad_norm": 2.7286736237238034, + "learning_rate": 3.2098765432098767e-06, + "loss": 0.7374, + "step": 13 + }, + { + "epoch": 0.021638330757341576, + "grad_norm": 3.3504502146745563, + "learning_rate": 3.4567901234567904e-06, + "loss": 0.7616, + "step": 14 + }, + { + "epoch": 0.023183925811437404, + "grad_norm": 3.3641769676772864, + "learning_rate": 3.7037037037037037e-06, + "loss": 0.7633, + "step": 15 + }, + { + "epoch": 0.02472952086553323, + "grad_norm": 3.147552036611318, + "learning_rate": 3.9506172839506175e-06, + "loss": 0.7376, + "step": 16 + }, + { + "epoch": 0.02627511591962906, + "grad_norm": 2.8976148119744427, + "learning_rate": 4.197530864197531e-06, + "loss": 0.736, + "step": 17 + }, + { + "epoch": 0.027820710973724884, + "grad_norm": 2.081476960627112, + "learning_rate": 4.444444444444444e-06, + "loss": 0.6908, + "step": 18 + }, + { + "epoch": 0.02936630602782071, + "grad_norm": 1.806761897373877, + "learning_rate": 4.691358024691358e-06, + "loss": 0.689, + "step": 19 + }, + { + "epoch": 0.030911901081916538, + "grad_norm": 1.390550633941569, + "learning_rate": 4.938271604938272e-06, + "loss": 0.6851, + "step": 20 + }, + { + "epoch": 0.03245749613601236, + "grad_norm": 1.1260914755524882, + "learning_rate": 5.185185185185185e-06, + "loss": 0.6689, + "step": 21 + }, + { + "epoch": 0.03400309119010819, + "grad_norm": 1.1828994926159446, + "learning_rate": 5.432098765432099e-06, + "loss": 0.6484, + "step": 22 + }, + { + "epoch": 0.03554868624420402, + "grad_norm": 1.2488339030387958, + "learning_rate": 5.6790123456790125e-06, + "loss": 0.6608, + "step": 23 + }, + { + "epoch": 0.03709428129829984, + "grad_norm": 1.1994124081567479, + "learning_rate": 5.925925925925926e-06, + "loss": 0.6476, + "step": 24 + }, + { + "epoch": 0.03863987635239567, + "grad_norm": 1.0594113595099088, + "learning_rate": 6.17283950617284e-06, + "loss": 0.6384, + "step": 25 + }, + { + "epoch": 0.0401854714064915, + "grad_norm": 0.8136334420177265, + "learning_rate": 6.419753086419753e-06, + "loss": 0.6363, + "step": 26 + }, + { + "epoch": 0.04173106646058733, + "grad_norm": 0.8515726939745646, + "learning_rate": 6.666666666666667e-06, + "loss": 0.6326, + "step": 27 + }, + { + "epoch": 0.04327666151468315, + "grad_norm": 0.8875870370322879, + "learning_rate": 6.913580246913581e-06, + "loss": 0.6229, + "step": 28 + }, + { + "epoch": 0.04482225656877898, + "grad_norm": 0.768439654173595, + "learning_rate": 7.160493827160494e-06, + "loss": 0.6167, + "step": 29 + }, + { + "epoch": 0.04636785162287481, + "grad_norm": 0.5189099049678193, + "learning_rate": 7.4074074074074075e-06, + "loss": 0.5938, + "step": 30 + }, + { + "epoch": 0.04791344667697063, + "grad_norm": 0.6395256514127601, + "learning_rate": 7.654320987654322e-06, + "loss": 0.6152, + "step": 31 + }, + { + "epoch": 0.04945904173106646, + "grad_norm": 0.746617550878664, + "learning_rate": 7.901234567901235e-06, + "loss": 0.5969, + "step": 32 + }, + { + "epoch": 0.05100463678516229, + "grad_norm": 0.5858126224692574, + "learning_rate": 8.148148148148148e-06, + "loss": 0.6024, + "step": 33 + }, + { + "epoch": 0.05255023183925812, + "grad_norm": 0.39467639483180406, + "learning_rate": 8.395061728395062e-06, + "loss": 0.5956, + "step": 34 + }, + { + "epoch": 0.05409582689335394, + "grad_norm": 0.5640256567279153, + "learning_rate": 8.641975308641975e-06, + "loss": 0.5821, + "step": 35 + }, + { + "epoch": 0.05564142194744977, + "grad_norm": 0.5987573689680895, + "learning_rate": 8.888888888888888e-06, + "loss": 0.5833, + "step": 36 + }, + { + "epoch": 0.0571870170015456, + "grad_norm": 0.44066358802919997, + "learning_rate": 9.135802469135803e-06, + "loss": 0.591, + "step": 37 + }, + { + "epoch": 0.05873261205564142, + "grad_norm": 0.3852388525791168, + "learning_rate": 9.382716049382717e-06, + "loss": 0.5778, + "step": 38 + }, + { + "epoch": 0.06027820710973725, + "grad_norm": 0.5098271176851854, + "learning_rate": 9.62962962962963e-06, + "loss": 0.5713, + "step": 39 + }, + { + "epoch": 0.061823802163833076, + "grad_norm": 0.39503936645354354, + "learning_rate": 9.876543209876543e-06, + "loss": 0.5714, + "step": 40 + }, + { + "epoch": 0.0633693972179289, + "grad_norm": 0.32905075142191664, + "learning_rate": 1.0123456790123458e-05, + "loss": 0.5728, + "step": 41 + }, + { + "epoch": 0.06491499227202473, + "grad_norm": 0.3840798410090591, + "learning_rate": 1.037037037037037e-05, + "loss": 0.5661, + "step": 42 + }, + { + "epoch": 0.06646058732612056, + "grad_norm": 0.38897396040812976, + "learning_rate": 1.0617283950617285e-05, + "loss": 0.5822, + "step": 43 + }, + { + "epoch": 0.06800618238021638, + "grad_norm": 0.29495303808913936, + "learning_rate": 1.0864197530864198e-05, + "loss": 0.5543, + "step": 44 + }, + { + "epoch": 0.0695517774343122, + "grad_norm": 0.3253252000829262, + "learning_rate": 1.1111111111111113e-05, + "loss": 0.5593, + "step": 45 + }, + { + "epoch": 0.07109737248840804, + "grad_norm": 0.34116573815145546, + "learning_rate": 1.1358024691358025e-05, + "loss": 0.5491, + "step": 46 + }, + { + "epoch": 0.07264296754250386, + "grad_norm": 0.24621781702597118, + "learning_rate": 1.160493827160494e-05, + "loss": 0.5619, + "step": 47 + }, + { + "epoch": 0.07418856259659969, + "grad_norm": 0.369173395538977, + "learning_rate": 1.1851851851851852e-05, + "loss": 0.5563, + "step": 48 + }, + { + "epoch": 0.07573415765069552, + "grad_norm": 0.2782501293355875, + "learning_rate": 1.2098765432098767e-05, + "loss": 0.5456, + "step": 49 + }, + { + "epoch": 0.07727975270479134, + "grad_norm": 0.29617283022535873, + "learning_rate": 1.234567901234568e-05, + "loss": 0.5553, + "step": 50 + }, + { + "epoch": 0.07882534775888717, + "grad_norm": 0.3110918507531697, + "learning_rate": 1.2592592592592593e-05, + "loss": 0.5483, + "step": 51 + }, + { + "epoch": 0.080370942812983, + "grad_norm": 0.27084532621368845, + "learning_rate": 1.2839506172839507e-05, + "loss": 0.5547, + "step": 52 + }, + { + "epoch": 0.08191653786707882, + "grad_norm": 0.2457944469090477, + "learning_rate": 1.3086419753086422e-05, + "loss": 0.5347, + "step": 53 + }, + { + "epoch": 0.08346213292117466, + "grad_norm": 0.2541436303825332, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.5428, + "step": 54 + }, + { + "epoch": 0.08500772797527048, + "grad_norm": 0.2376504860719097, + "learning_rate": 1.3580246913580248e-05, + "loss": 0.5388, + "step": 55 + }, + { + "epoch": 0.0865533230293663, + "grad_norm": 0.24597957657408892, + "learning_rate": 1.3827160493827162e-05, + "loss": 0.5513, + "step": 56 + }, + { + "epoch": 0.08809891808346214, + "grad_norm": 0.24278390126556001, + "learning_rate": 1.4074074074074075e-05, + "loss": 0.5549, + "step": 57 + }, + { + "epoch": 0.08964451313755796, + "grad_norm": 0.20020826855660456, + "learning_rate": 1.4320987654320988e-05, + "loss": 0.5328, + "step": 58 + }, + { + "epoch": 0.09119010819165378, + "grad_norm": 0.25383977984023004, + "learning_rate": 1.4567901234567903e-05, + "loss": 0.5505, + "step": 59 + }, + { + "epoch": 0.09273570324574962, + "grad_norm": 0.2340720421695047, + "learning_rate": 1.4814814814814815e-05, + "loss": 0.5386, + "step": 60 + }, + { + "epoch": 0.09428129829984544, + "grad_norm": 0.25602875868447883, + "learning_rate": 1.506172839506173e-05, + "loss": 0.5367, + "step": 61 + }, + { + "epoch": 0.09582689335394126, + "grad_norm": 0.27002467070309155, + "learning_rate": 1.5308641975308643e-05, + "loss": 0.5397, + "step": 62 + }, + { + "epoch": 0.0973724884080371, + "grad_norm": 0.3234378757167277, + "learning_rate": 1.555555555555556e-05, + "loss": 0.5358, + "step": 63 + }, + { + "epoch": 0.09891808346213292, + "grad_norm": 0.28118242669639826, + "learning_rate": 1.580246913580247e-05, + "loss": 0.5366, + "step": 64 + }, + { + "epoch": 0.10046367851622875, + "grad_norm": 0.24122559938222943, + "learning_rate": 1.6049382716049385e-05, + "loss": 0.5361, + "step": 65 + }, + { + "epoch": 0.10200927357032458, + "grad_norm": 0.28007867616979454, + "learning_rate": 1.6296296296296297e-05, + "loss": 0.5466, + "step": 66 + }, + { + "epoch": 0.1035548686244204, + "grad_norm": 0.251129640949984, + "learning_rate": 1.654320987654321e-05, + "loss": 0.5303, + "step": 67 + }, + { + "epoch": 0.10510046367851623, + "grad_norm": 0.22565320702718436, + "learning_rate": 1.6790123456790123e-05, + "loss": 0.5244, + "step": 68 + }, + { + "epoch": 0.10664605873261206, + "grad_norm": 0.24920566874778105, + "learning_rate": 1.7037037037037038e-05, + "loss": 0.5332, + "step": 69 + }, + { + "epoch": 0.10819165378670788, + "grad_norm": 0.29209867356120256, + "learning_rate": 1.728395061728395e-05, + "loss": 0.5302, + "step": 70 + }, + { + "epoch": 0.10973724884080371, + "grad_norm": 0.27057400482624655, + "learning_rate": 1.7530864197530865e-05, + "loss": 0.5195, + "step": 71 + }, + { + "epoch": 0.11128284389489954, + "grad_norm": 0.35167113888602347, + "learning_rate": 1.7777777777777777e-05, + "loss": 0.5163, + "step": 72 + }, + { + "epoch": 0.11282843894899536, + "grad_norm": 0.39760550684028373, + "learning_rate": 1.802469135802469e-05, + "loss": 0.5178, + "step": 73 + }, + { + "epoch": 0.1143740340030912, + "grad_norm": 0.6784824026653671, + "learning_rate": 1.8271604938271607e-05, + "loss": 0.5341, + "step": 74 + }, + { + "epoch": 0.11591962905718702, + "grad_norm": 0.9618391802990648, + "learning_rate": 1.851851851851852e-05, + "loss": 0.5337, + "step": 75 + }, + { + "epoch": 0.11746522411128284, + "grad_norm": 1.207818633909923, + "learning_rate": 1.8765432098765433e-05, + "loss": 0.5251, + "step": 76 + }, + { + "epoch": 0.11901081916537867, + "grad_norm": 0.4633663432837794, + "learning_rate": 1.901234567901235e-05, + "loss": 0.5189, + "step": 77 + }, + { + "epoch": 0.1205564142194745, + "grad_norm": 0.8571543214971838, + "learning_rate": 1.925925925925926e-05, + "loss": 0.5213, + "step": 78 + }, + { + "epoch": 0.12210200927357033, + "grad_norm": 1.0464179386041605, + "learning_rate": 1.9506172839506175e-05, + "loss": 0.5327, + "step": 79 + }, + { + "epoch": 0.12364760432766615, + "grad_norm": 0.6030795060182889, + "learning_rate": 1.9753086419753087e-05, + "loss": 0.5108, + "step": 80 + }, + { + "epoch": 0.125193199381762, + "grad_norm": 0.8282902108432075, + "learning_rate": 2e-05, + "loss": 0.5248, + "step": 81 + }, + { + "epoch": 0.1267387944358578, + "grad_norm": 0.6834029854964587, + "learning_rate": 2.0246913580246917e-05, + "loss": 0.5272, + "step": 82 + }, + { + "epoch": 0.12828438948995363, + "grad_norm": 0.9070124398991254, + "learning_rate": 2.049382716049383e-05, + "loss": 0.5181, + "step": 83 + }, + { + "epoch": 0.12982998454404945, + "grad_norm": 0.7509309688845062, + "learning_rate": 2.074074074074074e-05, + "loss": 0.5108, + "step": 84 + }, + { + "epoch": 0.13137557959814528, + "grad_norm": 0.5603999059650706, + "learning_rate": 2.0987654320987655e-05, + "loss": 0.5101, + "step": 85 + }, + { + "epoch": 0.13292117465224113, + "grad_norm": 0.8636069290455202, + "learning_rate": 2.123456790123457e-05, + "loss": 0.5146, + "step": 86 + }, + { + "epoch": 0.13446676970633695, + "grad_norm": 0.8045336877644039, + "learning_rate": 2.1481481481481485e-05, + "loss": 0.5154, + "step": 87 + }, + { + "epoch": 0.13601236476043277, + "grad_norm": 0.6023948125967806, + "learning_rate": 2.1728395061728397e-05, + "loss": 0.5083, + "step": 88 + }, + { + "epoch": 0.1375579598145286, + "grad_norm": 0.7704598071209162, + "learning_rate": 2.197530864197531e-05, + "loss": 0.5169, + "step": 89 + }, + { + "epoch": 0.1391035548686244, + "grad_norm": 0.741700249163537, + "learning_rate": 2.2222222222222227e-05, + "loss": 0.5198, + "step": 90 + }, + { + "epoch": 0.14064914992272023, + "grad_norm": 0.4950278030178324, + "learning_rate": 2.246913580246914e-05, + "loss": 0.5078, + "step": 91 + }, + { + "epoch": 0.14219474497681608, + "grad_norm": 0.5837929907445218, + "learning_rate": 2.271604938271605e-05, + "loss": 0.5036, + "step": 92 + }, + { + "epoch": 0.1437403400309119, + "grad_norm": 0.6238323145413748, + "learning_rate": 2.2962962962962965e-05, + "loss": 0.4985, + "step": 93 + }, + { + "epoch": 0.14528593508500773, + "grad_norm": 0.5532970702371948, + "learning_rate": 2.320987654320988e-05, + "loss": 0.5134, + "step": 94 + }, + { + "epoch": 0.14683153013910355, + "grad_norm": 0.8342787251297591, + "learning_rate": 2.3456790123456795e-05, + "loss": 0.5168, + "step": 95 + }, + { + "epoch": 0.14837712519319937, + "grad_norm": 0.8824818854861024, + "learning_rate": 2.3703703703703703e-05, + "loss": 0.5091, + "step": 96 + }, + { + "epoch": 0.14992272024729522, + "grad_norm": 1.288218263701589, + "learning_rate": 2.3950617283950618e-05, + "loss": 0.5228, + "step": 97 + }, + { + "epoch": 0.15146831530139104, + "grad_norm": 0.7289192134121079, + "learning_rate": 2.4197530864197533e-05, + "loss": 0.5027, + "step": 98 + }, + { + "epoch": 0.15301391035548687, + "grad_norm": 0.6904973969603343, + "learning_rate": 2.444444444444445e-05, + "loss": 0.5073, + "step": 99 + }, + { + "epoch": 0.1545595054095827, + "grad_norm": 1.0630399683570595, + "learning_rate": 2.469135802469136e-05, + "loss": 0.5096, + "step": 100 + }, + { + "epoch": 0.1561051004636785, + "grad_norm": 1.143477539924475, + "learning_rate": 2.4938271604938275e-05, + "loss": 0.5082, + "step": 101 + }, + { + "epoch": 0.15765069551777433, + "grad_norm": 0.8679214652629397, + "learning_rate": 2.5185185185185187e-05, + "loss": 0.5074, + "step": 102 + }, + { + "epoch": 0.15919629057187018, + "grad_norm": 1.1787592552800283, + "learning_rate": 2.54320987654321e-05, + "loss": 0.4968, + "step": 103 + }, + { + "epoch": 0.160741885625966, + "grad_norm": 1.0427291703445587, + "learning_rate": 2.5679012345679013e-05, + "loss": 0.5009, + "step": 104 + }, + { + "epoch": 0.16228748068006182, + "grad_norm": 0.847469777194458, + "learning_rate": 2.5925925925925928e-05, + "loss": 0.5145, + "step": 105 + }, + { + "epoch": 0.16383307573415765, + "grad_norm": 1.032596111298762, + "learning_rate": 2.6172839506172843e-05, + "loss": 0.4992, + "step": 106 + }, + { + "epoch": 0.16537867078825347, + "grad_norm": 0.7112034821814768, + "learning_rate": 2.641975308641976e-05, + "loss": 0.5096, + "step": 107 + }, + { + "epoch": 0.16692426584234932, + "grad_norm": 0.7754993141540937, + "learning_rate": 2.6666666666666667e-05, + "loss": 0.4907, + "step": 108 + }, + { + "epoch": 0.16846986089644514, + "grad_norm": 0.7610858355351189, + "learning_rate": 2.691358024691358e-05, + "loss": 0.5065, + "step": 109 + }, + { + "epoch": 0.17001545595054096, + "grad_norm": 0.7407566934069416, + "learning_rate": 2.7160493827160497e-05, + "loss": 0.5025, + "step": 110 + }, + { + "epoch": 0.17156105100463678, + "grad_norm": 0.8022655281725392, + "learning_rate": 2.740740740740741e-05, + "loss": 0.4963, + "step": 111 + }, + { + "epoch": 0.1731066460587326, + "grad_norm": 0.7631247318207706, + "learning_rate": 2.7654320987654323e-05, + "loss": 0.5018, + "step": 112 + }, + { + "epoch": 0.17465224111282843, + "grad_norm": 1.078272532566959, + "learning_rate": 2.7901234567901235e-05, + "loss": 0.4933, + "step": 113 + }, + { + "epoch": 0.17619783616692428, + "grad_norm": 0.9770429338531161, + "learning_rate": 2.814814814814815e-05, + "loss": 0.5108, + "step": 114 + }, + { + "epoch": 0.1777434312210201, + "grad_norm": 1.209883015711933, + "learning_rate": 2.8395061728395065e-05, + "loss": 0.5048, + "step": 115 + }, + { + "epoch": 0.17928902627511592, + "grad_norm": 0.6568854476685991, + "learning_rate": 2.8641975308641977e-05, + "loss": 0.5063, + "step": 116 + }, + { + "epoch": 0.18083462132921174, + "grad_norm": 0.6369454626125819, + "learning_rate": 2.888888888888889e-05, + "loss": 0.49, + "step": 117 + }, + { + "epoch": 0.18238021638330756, + "grad_norm": 0.6945493811214182, + "learning_rate": 2.9135802469135807e-05, + "loss": 0.5008, + "step": 118 + }, + { + "epoch": 0.1839258114374034, + "grad_norm": 0.7084610539302401, + "learning_rate": 2.938271604938272e-05, + "loss": 0.4928, + "step": 119 + }, + { + "epoch": 0.18547140649149924, + "grad_norm": 1.0347139753934802, + "learning_rate": 2.962962962962963e-05, + "loss": 0.503, + "step": 120 + }, + { + "epoch": 0.18701700154559506, + "grad_norm": 1.0981919160573457, + "learning_rate": 2.9876543209876545e-05, + "loss": 0.5032, + "step": 121 + }, + { + "epoch": 0.18856259659969088, + "grad_norm": 0.8082068804313435, + "learning_rate": 3.012345679012346e-05, + "loss": 0.5002, + "step": 122 + }, + { + "epoch": 0.1901081916537867, + "grad_norm": 0.9238973243615479, + "learning_rate": 3.0370370370370375e-05, + "loss": 0.4982, + "step": 123 + }, + { + "epoch": 0.19165378670788252, + "grad_norm": 0.9631424104539678, + "learning_rate": 3.061728395061729e-05, + "loss": 0.5021, + "step": 124 + }, + { + "epoch": 0.19319938176197837, + "grad_norm": 1.1086567942544703, + "learning_rate": 3.08641975308642e-05, + "loss": 0.4983, + "step": 125 + }, + { + "epoch": 0.1947449768160742, + "grad_norm": 1.1628943248183918, + "learning_rate": 3.111111111111112e-05, + "loss": 0.5024, + "step": 126 + }, + { + "epoch": 0.19629057187017002, + "grad_norm": 0.8257350753088295, + "learning_rate": 3.135802469135803e-05, + "loss": 0.4998, + "step": 127 + }, + { + "epoch": 0.19783616692426584, + "grad_norm": 0.8618783547766707, + "learning_rate": 3.160493827160494e-05, + "loss": 0.4922, + "step": 128 + }, + { + "epoch": 0.19938176197836166, + "grad_norm": 1.241169327864676, + "learning_rate": 3.185185185185185e-05, + "loss": 0.4999, + "step": 129 + }, + { + "epoch": 0.2009273570324575, + "grad_norm": 0.828132554003403, + "learning_rate": 3.209876543209877e-05, + "loss": 0.5026, + "step": 130 + }, + { + "epoch": 0.20247295208655333, + "grad_norm": 0.7793563201875642, + "learning_rate": 3.234567901234568e-05, + "loss": 0.4886, + "step": 131 + }, + { + "epoch": 0.20401854714064915, + "grad_norm": 0.9675464742848618, + "learning_rate": 3.259259259259259e-05, + "loss": 0.4915, + "step": 132 + }, + { + "epoch": 0.20556414219474498, + "grad_norm": 0.8431076894503557, + "learning_rate": 3.2839506172839505e-05, + "loss": 0.5025, + "step": 133 + }, + { + "epoch": 0.2071097372488408, + "grad_norm": 0.8812947158826483, + "learning_rate": 3.308641975308642e-05, + "loss": 0.4903, + "step": 134 + }, + { + "epoch": 0.20865533230293662, + "grad_norm": 1.3213441436820172, + "learning_rate": 3.3333333333333335e-05, + "loss": 0.4997, + "step": 135 + }, + { + "epoch": 0.21020092735703247, + "grad_norm": 0.5642206183809721, + "learning_rate": 3.3580246913580247e-05, + "loss": 0.4982, + "step": 136 + }, + { + "epoch": 0.2117465224111283, + "grad_norm": 1.3173420862661687, + "learning_rate": 3.3827160493827165e-05, + "loss": 0.502, + "step": 137 + }, + { + "epoch": 0.2132921174652241, + "grad_norm": 0.8513440164344224, + "learning_rate": 3.4074074074074077e-05, + "loss": 0.4971, + "step": 138 + }, + { + "epoch": 0.21483771251931993, + "grad_norm": 0.8672711786334716, + "learning_rate": 3.4320987654320995e-05, + "loss": 0.489, + "step": 139 + }, + { + "epoch": 0.21638330757341576, + "grad_norm": 1.0621643679180979, + "learning_rate": 3.45679012345679e-05, + "loss": 0.4991, + "step": 140 + }, + { + "epoch": 0.21792890262751158, + "grad_norm": 0.7921164646769278, + "learning_rate": 3.481481481481482e-05, + "loss": 0.4833, + "step": 141 + }, + { + "epoch": 0.21947449768160743, + "grad_norm": 1.1313070889701453, + "learning_rate": 3.506172839506173e-05, + "loss": 0.4983, + "step": 142 + }, + { + "epoch": 0.22102009273570325, + "grad_norm": 0.979446436116847, + "learning_rate": 3.530864197530865e-05, + "loss": 0.4888, + "step": 143 + }, + { + "epoch": 0.22256568778979907, + "grad_norm": 0.9631416113357064, + "learning_rate": 3.555555555555555e-05, + "loss": 0.4973, + "step": 144 + }, + { + "epoch": 0.2241112828438949, + "grad_norm": 0.8774746523842448, + "learning_rate": 3.580246913580247e-05, + "loss": 0.4909, + "step": 145 + }, + { + "epoch": 0.22565687789799072, + "grad_norm": 0.9473774265327042, + "learning_rate": 3.604938271604938e-05, + "loss": 0.4867, + "step": 146 + }, + { + "epoch": 0.22720247295208656, + "grad_norm": 1.2382081588447356, + "learning_rate": 3.62962962962963e-05, + "loss": 0.4895, + "step": 147 + }, + { + "epoch": 0.2287480680061824, + "grad_norm": 0.9373600980716077, + "learning_rate": 3.654320987654321e-05, + "loss": 0.4906, + "step": 148 + }, + { + "epoch": 0.2302936630602782, + "grad_norm": 1.1641931805143952, + "learning_rate": 3.6790123456790125e-05, + "loss": 0.493, + "step": 149 + }, + { + "epoch": 0.23183925811437403, + "grad_norm": 0.6192917083811479, + "learning_rate": 3.703703703703704e-05, + "loss": 0.5018, + "step": 150 + }, + { + "epoch": 0.23338485316846985, + "grad_norm": 0.9568940117279771, + "learning_rate": 3.7283950617283955e-05, + "loss": 0.4859, + "step": 151 + }, + { + "epoch": 0.23493044822256567, + "grad_norm": 0.7558753315433521, + "learning_rate": 3.7530864197530867e-05, + "loss": 0.4963, + "step": 152 + }, + { + "epoch": 0.23647604327666152, + "grad_norm": 0.9212501384989878, + "learning_rate": 3.777777777777778e-05, + "loss": 0.4819, + "step": 153 + }, + { + "epoch": 0.23802163833075735, + "grad_norm": 1.1697128542170077, + "learning_rate": 3.80246913580247e-05, + "loss": 0.4823, + "step": 154 + }, + { + "epoch": 0.23956723338485317, + "grad_norm": 1.1974076561986966, + "learning_rate": 3.827160493827161e-05, + "loss": 0.4898, + "step": 155 + }, + { + "epoch": 0.241112828438949, + "grad_norm": 0.5721357290416084, + "learning_rate": 3.851851851851852e-05, + "loss": 0.4883, + "step": 156 + }, + { + "epoch": 0.2426584234930448, + "grad_norm": 0.8279428890745845, + "learning_rate": 3.876543209876543e-05, + "loss": 0.4913, + "step": 157 + }, + { + "epoch": 0.24420401854714066, + "grad_norm": 0.7905766725922289, + "learning_rate": 3.901234567901235e-05, + "loss": 0.4999, + "step": 158 + }, + { + "epoch": 0.24574961360123648, + "grad_norm": 1.1663187217487774, + "learning_rate": 3.925925925925926e-05, + "loss": 0.4985, + "step": 159 + }, + { + "epoch": 0.2472952086553323, + "grad_norm": 1.044068287922803, + "learning_rate": 3.950617283950617e-05, + "loss": 0.4805, + "step": 160 + }, + { + "epoch": 0.24884080370942813, + "grad_norm": 0.9571058865876857, + "learning_rate": 3.975308641975309e-05, + "loss": 0.4991, + "step": 161 + }, + { + "epoch": 0.250386398763524, + "grad_norm": 1.1566279067789822, + "learning_rate": 4e-05, + "loss": 0.4964, + "step": 162 + }, + { + "epoch": 0.25193199381761977, + "grad_norm": 1.1448282250848643, + "learning_rate": 4.0246913580246915e-05, + "loss": 0.498, + "step": 163 + }, + { + "epoch": 0.2534775888717156, + "grad_norm": 1.1203356595739242, + "learning_rate": 4.049382716049383e-05, + "loss": 0.4873, + "step": 164 + }, + { + "epoch": 0.2550231839258114, + "grad_norm": 1.0233529441473044, + "learning_rate": 4.0740740740740745e-05, + "loss": 0.4926, + "step": 165 + }, + { + "epoch": 0.25656877897990726, + "grad_norm": 0.9118723834779173, + "learning_rate": 4.098765432098766e-05, + "loss": 0.4803, + "step": 166 + }, + { + "epoch": 0.2581143740340031, + "grad_norm": 1.1742293487040738, + "learning_rate": 4.123456790123457e-05, + "loss": 0.4891, + "step": 167 + }, + { + "epoch": 0.2596599690880989, + "grad_norm": 0.9090936632777925, + "learning_rate": 4.148148148148148e-05, + "loss": 0.4814, + "step": 168 + }, + { + "epoch": 0.26120556414219476, + "grad_norm": 1.0102483704824479, + "learning_rate": 4.17283950617284e-05, + "loss": 0.4787, + "step": 169 + }, + { + "epoch": 0.26275115919629055, + "grad_norm": 1.0211857026705857, + "learning_rate": 4.197530864197531e-05, + "loss": 0.488, + "step": 170 + }, + { + "epoch": 0.2642967542503864, + "grad_norm": 1.1146681594826109, + "learning_rate": 4.222222222222223e-05, + "loss": 0.4888, + "step": 171 + }, + { + "epoch": 0.26584234930448225, + "grad_norm": 0.9537477496730035, + "learning_rate": 4.246913580246914e-05, + "loss": 0.4734, + "step": 172 + }, + { + "epoch": 0.26738794435857804, + "grad_norm": 1.043638066624262, + "learning_rate": 4.271604938271606e-05, + "loss": 0.4742, + "step": 173 + }, + { + "epoch": 0.2689335394126739, + "grad_norm": 1.4422937022582138, + "learning_rate": 4.296296296296297e-05, + "loss": 0.4959, + "step": 174 + }, + { + "epoch": 0.2704791344667697, + "grad_norm": 1.0299464445089963, + "learning_rate": 4.3209876543209875e-05, + "loss": 0.481, + "step": 175 + }, + { + "epoch": 0.27202472952086554, + "grad_norm": 1.1818391836210722, + "learning_rate": 4.345679012345679e-05, + "loss": 0.4847, + "step": 176 + }, + { + "epoch": 0.2735703245749614, + "grad_norm": 0.7618135468190563, + "learning_rate": 4.3703703703703705e-05, + "loss": 0.4759, + "step": 177 + }, + { + "epoch": 0.2751159196290572, + "grad_norm": 1.6171899862729473, + "learning_rate": 4.395061728395062e-05, + "loss": 0.4822, + "step": 178 + }, + { + "epoch": 0.27666151468315303, + "grad_norm": 0.5985280763844202, + "learning_rate": 4.4197530864197535e-05, + "loss": 0.4903, + "step": 179 + }, + { + "epoch": 0.2782071097372488, + "grad_norm": 1.014644632490285, + "learning_rate": 4.444444444444445e-05, + "loss": 0.4777, + "step": 180 + }, + { + "epoch": 0.2797527047913447, + "grad_norm": 1.4790130521335796, + "learning_rate": 4.4691358024691365e-05, + "loss": 0.4924, + "step": 181 + }, + { + "epoch": 0.28129829984544047, + "grad_norm": 0.9371239153663216, + "learning_rate": 4.493827160493828e-05, + "loss": 0.4903, + "step": 182 + }, + { + "epoch": 0.2828438948995363, + "grad_norm": 0.8819988307673926, + "learning_rate": 4.518518518518519e-05, + "loss": 0.4857, + "step": 183 + }, + { + "epoch": 0.28438948995363217, + "grad_norm": 1.2980448181187785, + "learning_rate": 4.54320987654321e-05, + "loss": 0.4895, + "step": 184 + }, + { + "epoch": 0.28593508500772796, + "grad_norm": 1.0937651336408682, + "learning_rate": 4.567901234567901e-05, + "loss": 0.4776, + "step": 185 + }, + { + "epoch": 0.2874806800618238, + "grad_norm": 1.1682285456048371, + "learning_rate": 4.592592592592593e-05, + "loss": 0.4917, + "step": 186 + }, + { + "epoch": 0.2890262751159196, + "grad_norm": 1.1571200068187466, + "learning_rate": 4.617283950617284e-05, + "loss": 0.4876, + "step": 187 + }, + { + "epoch": 0.29057187017001546, + "grad_norm": 0.728643662456322, + "learning_rate": 4.641975308641976e-05, + "loss": 0.4752, + "step": 188 + }, + { + "epoch": 0.2921174652241113, + "grad_norm": 1.5497374490680917, + "learning_rate": 4.666666666666667e-05, + "loss": 0.4821, + "step": 189 + }, + { + "epoch": 0.2936630602782071, + "grad_norm": 0.6173816215753983, + "learning_rate": 4.691358024691359e-05, + "loss": 0.4824, + "step": 190 + }, + { + "epoch": 0.29520865533230295, + "grad_norm": 1.3179137124249143, + "learning_rate": 4.7160493827160495e-05, + "loss": 0.4944, + "step": 191 + }, + { + "epoch": 0.29675425038639874, + "grad_norm": 0.7994479560157165, + "learning_rate": 4.7407407407407407e-05, + "loss": 0.4809, + "step": 192 + }, + { + "epoch": 0.2982998454404946, + "grad_norm": 1.2701249362355918, + "learning_rate": 4.7654320987654325e-05, + "loss": 0.482, + "step": 193 + }, + { + "epoch": 0.29984544049459044, + "grad_norm": 0.7512432473101126, + "learning_rate": 4.7901234567901237e-05, + "loss": 0.4839, + "step": 194 + }, + { + "epoch": 0.30139103554868624, + "grad_norm": 0.929048702574481, + "learning_rate": 4.8148148148148155e-05, + "loss": 0.4824, + "step": 195 + }, + { + "epoch": 0.3029366306027821, + "grad_norm": 1.1727570739816218, + "learning_rate": 4.8395061728395067e-05, + "loss": 0.4844, + "step": 196 + }, + { + "epoch": 0.3044822256568779, + "grad_norm": 1.2592385923726135, + "learning_rate": 4.8641975308641985e-05, + "loss": 0.4883, + "step": 197 + }, + { + "epoch": 0.30602782071097373, + "grad_norm": 1.0274500278429952, + "learning_rate": 4.88888888888889e-05, + "loss": 0.4737, + "step": 198 + }, + { + "epoch": 0.3075734157650695, + "grad_norm": 1.0222774019121448, + "learning_rate": 4.91358024691358e-05, + "loss": 0.4799, + "step": 199 + }, + { + "epoch": 0.3091190108191654, + "grad_norm": 1.0732342293445107, + "learning_rate": 4.938271604938272e-05, + "loss": 0.4808, + "step": 200 + }, + { + "epoch": 0.3106646058732612, + "grad_norm": 1.1148226181132161, + "learning_rate": 4.962962962962963e-05, + "loss": 0.4764, + "step": 201 + }, + { + "epoch": 0.312210200927357, + "grad_norm": 0.8997498855655075, + "learning_rate": 4.987654320987655e-05, + "loss": 0.4872, + "step": 202 + }, + { + "epoch": 0.31375579598145287, + "grad_norm": 1.1821249594599401, + "learning_rate": 5.012345679012346e-05, + "loss": 0.4697, + "step": 203 + }, + { + "epoch": 0.31530139103554866, + "grad_norm": 1.0971633801119358, + "learning_rate": 5.037037037037037e-05, + "loss": 0.485, + "step": 204 + }, + { + "epoch": 0.3168469860896445, + "grad_norm": 0.8255153623316014, + "learning_rate": 5.061728395061729e-05, + "loss": 0.4744, + "step": 205 + }, + { + "epoch": 0.31839258114374036, + "grad_norm": 1.2668927846380267, + "learning_rate": 5.08641975308642e-05, + "loss": 0.4811, + "step": 206 + }, + { + "epoch": 0.31993817619783615, + "grad_norm": 1.2156929653202533, + "learning_rate": 5.111111111111111e-05, + "loss": 0.4823, + "step": 207 + }, + { + "epoch": 0.321483771251932, + "grad_norm": 0.8745911104546992, + "learning_rate": 5.1358024691358027e-05, + "loss": 0.4777, + "step": 208 + }, + { + "epoch": 0.3230293663060278, + "grad_norm": 1.7497488486354917, + "learning_rate": 5.160493827160494e-05, + "loss": 0.4847, + "step": 209 + }, + { + "epoch": 0.32457496136012365, + "grad_norm": 0.8228927010361928, + "learning_rate": 5.1851851851851857e-05, + "loss": 0.4717, + "step": 210 + }, + { + "epoch": 0.3261205564142195, + "grad_norm": 1.9889828694359233, + "learning_rate": 5.209876543209877e-05, + "loss": 0.4835, + "step": 211 + }, + { + "epoch": 0.3276661514683153, + "grad_norm": 1.287587405121619, + "learning_rate": 5.234567901234569e-05, + "loss": 0.4724, + "step": 212 + }, + { + "epoch": 0.32921174652241114, + "grad_norm": 1.6948902446976986, + "learning_rate": 5.25925925925926e-05, + "loss": 0.4756, + "step": 213 + }, + { + "epoch": 0.33075734157650694, + "grad_norm": 1.4119177281832886, + "learning_rate": 5.283950617283952e-05, + "loss": 0.4826, + "step": 214 + }, + { + "epoch": 0.3323029366306028, + "grad_norm": 1.5167781358660195, + "learning_rate": 5.308641975308642e-05, + "loss": 0.4811, + "step": 215 + }, + { + "epoch": 0.33384853168469864, + "grad_norm": 1.1308110283174948, + "learning_rate": 5.333333333333333e-05, + "loss": 0.4816, + "step": 216 + }, + { + "epoch": 0.33539412673879443, + "grad_norm": 1.2278896437711684, + "learning_rate": 5.358024691358025e-05, + "loss": 0.474, + "step": 217 + }, + { + "epoch": 0.3369397217928903, + "grad_norm": 0.8622626370979445, + "learning_rate": 5.382716049382716e-05, + "loss": 0.4744, + "step": 218 + }, + { + "epoch": 0.3384853168469861, + "grad_norm": 1.1986006504365891, + "learning_rate": 5.407407407407408e-05, + "loss": 0.4798, + "step": 219 + }, + { + "epoch": 0.3400309119010819, + "grad_norm": 0.6179519214399118, + "learning_rate": 5.432098765432099e-05, + "loss": 0.4683, + "step": 220 + }, + { + "epoch": 0.3415765069551777, + "grad_norm": 0.9582362210598186, + "learning_rate": 5.4567901234567905e-05, + "loss": 0.4721, + "step": 221 + }, + { + "epoch": 0.34312210200927357, + "grad_norm": 0.8775268539297884, + "learning_rate": 5.481481481481482e-05, + "loss": 0.4838, + "step": 222 + }, + { + "epoch": 0.3446676970633694, + "grad_norm": 1.0835329353629608, + "learning_rate": 5.506172839506173e-05, + "loss": 0.4836, + "step": 223 + }, + { + "epoch": 0.3462132921174652, + "grad_norm": 1.1439631376935295, + "learning_rate": 5.5308641975308647e-05, + "loss": 0.4767, + "step": 224 + }, + { + "epoch": 0.34775888717156106, + "grad_norm": 1.2716091535791791, + "learning_rate": 5.555555555555556e-05, + "loss": 0.4846, + "step": 225 + }, + { + "epoch": 0.34930448222565685, + "grad_norm": 0.8542190983120236, + "learning_rate": 5.580246913580247e-05, + "loss": 0.4736, + "step": 226 + }, + { + "epoch": 0.3508500772797527, + "grad_norm": 0.7750218925853332, + "learning_rate": 5.604938271604939e-05, + "loss": 0.4782, + "step": 227 + }, + { + "epoch": 0.35239567233384855, + "grad_norm": 0.8410137777186895, + "learning_rate": 5.62962962962963e-05, + "loss": 0.4659, + "step": 228 + }, + { + "epoch": 0.35394126738794435, + "grad_norm": 1.1347741397027802, + "learning_rate": 5.654320987654322e-05, + "loss": 0.4842, + "step": 229 + }, + { + "epoch": 0.3554868624420402, + "grad_norm": 1.2640994747112773, + "learning_rate": 5.679012345679013e-05, + "loss": 0.4811, + "step": 230 + }, + { + "epoch": 0.357032457496136, + "grad_norm": 1.0358540531545488, + "learning_rate": 5.7037037037037035e-05, + "loss": 0.4815, + "step": 231 + }, + { + "epoch": 0.35857805255023184, + "grad_norm": 1.1731444912874869, + "learning_rate": 5.728395061728395e-05, + "loss": 0.4824, + "step": 232 + }, + { + "epoch": 0.3601236476043277, + "grad_norm": 1.0963715252451798, + "learning_rate": 5.7530864197530865e-05, + "loss": 0.4819, + "step": 233 + }, + { + "epoch": 0.3616692426584235, + "grad_norm": 1.1234656582964098, + "learning_rate": 5.777777777777778e-05, + "loss": 0.4788, + "step": 234 + }, + { + "epoch": 0.36321483771251933, + "grad_norm": 1.0039863248210223, + "learning_rate": 5.8024691358024695e-05, + "loss": 0.4848, + "step": 235 + }, + { + "epoch": 0.36476043276661513, + "grad_norm": 0.8562522614103883, + "learning_rate": 5.827160493827161e-05, + "loss": 0.4713, + "step": 236 + }, + { + "epoch": 0.366306027820711, + "grad_norm": 1.0344076244948395, + "learning_rate": 5.8518518518518525e-05, + "loss": 0.4677, + "step": 237 + }, + { + "epoch": 0.3678516228748068, + "grad_norm": 1.034057279913179, + "learning_rate": 5.876543209876544e-05, + "loss": 0.4768, + "step": 238 + }, + { + "epoch": 0.3693972179289026, + "grad_norm": 1.7367687508499554, + "learning_rate": 5.901234567901235e-05, + "loss": 0.4852, + "step": 239 + }, + { + "epoch": 0.37094281298299847, + "grad_norm": 0.8118841495079787, + "learning_rate": 5.925925925925926e-05, + "loss": 0.4634, + "step": 240 + }, + { + "epoch": 0.37248840803709427, + "grad_norm": 1.471099919135472, + "learning_rate": 5.950617283950618e-05, + "loss": 0.4887, + "step": 241 + }, + { + "epoch": 0.3740340030911901, + "grad_norm": 1.1587198015371982, + "learning_rate": 5.975308641975309e-05, + "loss": 0.4717, + "step": 242 + }, + { + "epoch": 0.3755795981452859, + "grad_norm": 1.758659389813336, + "learning_rate": 6.000000000000001e-05, + "loss": 0.4895, + "step": 243 + }, + { + "epoch": 0.37712519319938176, + "grad_norm": 1.3315013513741276, + "learning_rate": 6.024691358024692e-05, + "loss": 0.4873, + "step": 244 + }, + { + "epoch": 0.3786707882534776, + "grad_norm": 1.7550563966474997, + "learning_rate": 6.049382716049383e-05, + "loss": 0.4754, + "step": 245 + }, + { + "epoch": 0.3802163833075734, + "grad_norm": 1.4584548318573913, + "learning_rate": 6.074074074074075e-05, + "loss": 0.4889, + "step": 246 + }, + { + "epoch": 0.38176197836166925, + "grad_norm": 1.7243876781798975, + "learning_rate": 6.098765432098766e-05, + "loss": 0.4765, + "step": 247 + }, + { + "epoch": 0.38330757341576505, + "grad_norm": 1.167432430205111, + "learning_rate": 6.123456790123457e-05, + "loss": 0.4726, + "step": 248 + }, + { + "epoch": 0.3848531684698609, + "grad_norm": 1.4858200526822005, + "learning_rate": 6.148148148148148e-05, + "loss": 0.4828, + "step": 249 + }, + { + "epoch": 0.38639876352395675, + "grad_norm": 0.9500856749304806, + "learning_rate": 6.17283950617284e-05, + "loss": 0.48, + "step": 250 + }, + { + "epoch": 0.38794435857805254, + "grad_norm": 0.8681770559849791, + "learning_rate": 6.197530864197531e-05, + "loss": 0.4703, + "step": 251 + }, + { + "epoch": 0.3894899536321484, + "grad_norm": 1.0989972142230926, + "learning_rate": 6.222222222222223e-05, + "loss": 0.4695, + "step": 252 + }, + { + "epoch": 0.3910355486862442, + "grad_norm": 0.9402230268151549, + "learning_rate": 6.246913580246914e-05, + "loss": 0.4758, + "step": 253 + }, + { + "epoch": 0.39258114374034003, + "grad_norm": 0.9848484342795987, + "learning_rate": 6.271604938271606e-05, + "loss": 0.4668, + "step": 254 + }, + { + "epoch": 0.3941267387944359, + "grad_norm": 1.1574464067866512, + "learning_rate": 6.296296296296298e-05, + "loss": 0.4657, + "step": 255 + }, + { + "epoch": 0.3956723338485317, + "grad_norm": 1.1322864127449526, + "learning_rate": 6.320987654320988e-05, + "loss": 0.4766, + "step": 256 + }, + { + "epoch": 0.3972179289026275, + "grad_norm": 1.196737268139593, + "learning_rate": 6.34567901234568e-05, + "loss": 0.468, + "step": 257 + }, + { + "epoch": 0.3987635239567233, + "grad_norm": 0.8615114262443972, + "learning_rate": 6.37037037037037e-05, + "loss": 0.4786, + "step": 258 + }, + { + "epoch": 0.40030911901081917, + "grad_norm": 1.0155583184427355, + "learning_rate": 6.395061728395062e-05, + "loss": 0.476, + "step": 259 + }, + { + "epoch": 0.401854714064915, + "grad_norm": 0.8705021038685753, + "learning_rate": 6.419753086419754e-05, + "loss": 0.4814, + "step": 260 + }, + { + "epoch": 0.4034003091190108, + "grad_norm": 1.0551108697985605, + "learning_rate": 6.444444444444446e-05, + "loss": 0.4698, + "step": 261 + }, + { + "epoch": 0.40494590417310666, + "grad_norm": 1.2069387128869065, + "learning_rate": 6.469135802469136e-05, + "loss": 0.4743, + "step": 262 + }, + { + "epoch": 0.40649149922720246, + "grad_norm": 0.7605544444447253, + "learning_rate": 6.493827160493828e-05, + "loss": 0.4543, + "step": 263 + }, + { + "epoch": 0.4080370942812983, + "grad_norm": 0.8205540585121743, + "learning_rate": 6.518518518518519e-05, + "loss": 0.4684, + "step": 264 + }, + { + "epoch": 0.4095826893353941, + "grad_norm": 1.0619805628658, + "learning_rate": 6.54320987654321e-05, + "loss": 0.4668, + "step": 265 + }, + { + "epoch": 0.41112828438948995, + "grad_norm": 1.232094986638869, + "learning_rate": 6.567901234567901e-05, + "loss": 0.4657, + "step": 266 + }, + { + "epoch": 0.4126738794435858, + "grad_norm": 1.2059342270172937, + "learning_rate": 6.592592592592593e-05, + "loss": 0.4627, + "step": 267 + }, + { + "epoch": 0.4142194744976816, + "grad_norm": 0.9923650404733554, + "learning_rate": 6.617283950617285e-05, + "loss": 0.4823, + "step": 268 + }, + { + "epoch": 0.41576506955177744, + "grad_norm": 0.9006494596530477, + "learning_rate": 6.641975308641977e-05, + "loss": 0.468, + "step": 269 + }, + { + "epoch": 0.41731066460587324, + "grad_norm": 0.9325765412437474, + "learning_rate": 6.666666666666667e-05, + "loss": 0.4784, + "step": 270 + }, + { + "epoch": 0.4188562596599691, + "grad_norm": 1.047774171198826, + "learning_rate": 6.691358024691359e-05, + "loss": 0.4671, + "step": 271 + }, + { + "epoch": 0.42040185471406494, + "grad_norm": 1.2038403751218132, + "learning_rate": 6.716049382716049e-05, + "loss": 0.4831, + "step": 272 + }, + { + "epoch": 0.42194744976816073, + "grad_norm": 0.848323979703427, + "learning_rate": 6.740740740740741e-05, + "loss": 0.4768, + "step": 273 + }, + { + "epoch": 0.4234930448222566, + "grad_norm": 1.468772962629579, + "learning_rate": 6.765432098765433e-05, + "loss": 0.4699, + "step": 274 + }, + { + "epoch": 0.4250386398763524, + "grad_norm": 0.7234151499833007, + "learning_rate": 6.790123456790123e-05, + "loss": 0.4709, + "step": 275 + }, + { + "epoch": 0.4265842349304482, + "grad_norm": 0.7864446856107487, + "learning_rate": 6.814814814814815e-05, + "loss": 0.4707, + "step": 276 + }, + { + "epoch": 0.4281298299845441, + "grad_norm": 1.5551025678321875, + "learning_rate": 6.839506172839507e-05, + "loss": 0.4828, + "step": 277 + }, + { + "epoch": 0.42967542503863987, + "grad_norm": 0.759156333480939, + "learning_rate": 6.864197530864199e-05, + "loss": 0.4602, + "step": 278 + }, + { + "epoch": 0.4312210200927357, + "grad_norm": 0.9162967605158892, + "learning_rate": 6.88888888888889e-05, + "loss": 0.4676, + "step": 279 + }, + { + "epoch": 0.4327666151468315, + "grad_norm": 1.3177533448409955, + "learning_rate": 6.91358024691358e-05, + "loss": 0.4771, + "step": 280 + }, + { + "epoch": 0.43431221020092736, + "grad_norm": 1.0065569840708148, + "learning_rate": 6.938271604938272e-05, + "loss": 0.4654, + "step": 281 + }, + { + "epoch": 0.43585780525502316, + "grad_norm": 1.0099019290554392, + "learning_rate": 6.962962962962964e-05, + "loss": 0.4736, + "step": 282 + }, + { + "epoch": 0.437403400309119, + "grad_norm": 0.9542599858654569, + "learning_rate": 6.987654320987655e-05, + "loss": 0.4697, + "step": 283 + }, + { + "epoch": 0.43894899536321486, + "grad_norm": 1.0993378431582324, + "learning_rate": 7.012345679012346e-05, + "loss": 0.4646, + "step": 284 + }, + { + "epoch": 0.44049459041731065, + "grad_norm": 0.9091639988441267, + "learning_rate": 7.037037037037038e-05, + "loss": 0.4731, + "step": 285 + }, + { + "epoch": 0.4420401854714065, + "grad_norm": 1.1115775076593608, + "learning_rate": 7.06172839506173e-05, + "loss": 0.4735, + "step": 286 + }, + { + "epoch": 0.4435857805255023, + "grad_norm": 1.1534031048862625, + "learning_rate": 7.08641975308642e-05, + "loss": 0.4748, + "step": 287 + }, + { + "epoch": 0.44513137557959814, + "grad_norm": 1.1500223572788075, + "learning_rate": 7.11111111111111e-05, + "loss": 0.4778, + "step": 288 + }, + { + "epoch": 0.446676970633694, + "grad_norm": 1.0303778729219273, + "learning_rate": 7.135802469135802e-05, + "loss": 0.4688, + "step": 289 + }, + { + "epoch": 0.4482225656877898, + "grad_norm": 0.9689431632471549, + "learning_rate": 7.160493827160494e-05, + "loss": 0.4762, + "step": 290 + }, + { + "epoch": 0.44976816074188564, + "grad_norm": 0.9233987172654912, + "learning_rate": 7.185185185185186e-05, + "loss": 0.4696, + "step": 291 + }, + { + "epoch": 0.45131375579598143, + "grad_norm": 1.4934762550254848, + "learning_rate": 7.209876543209877e-05, + "loss": 0.4752, + "step": 292 + }, + { + "epoch": 0.4528593508500773, + "grad_norm": 0.7502253552129426, + "learning_rate": 7.234567901234568e-05, + "loss": 0.4657, + "step": 293 + }, + { + "epoch": 0.45440494590417313, + "grad_norm": 1.117072200683503, + "learning_rate": 7.25925925925926e-05, + "loss": 0.4663, + "step": 294 + }, + { + "epoch": 0.4559505409582689, + "grad_norm": 0.9331429681836702, + "learning_rate": 7.283950617283952e-05, + "loss": 0.4766, + "step": 295 + }, + { + "epoch": 0.4574961360123648, + "grad_norm": 1.068844821636654, + "learning_rate": 7.308641975308643e-05, + "loss": 0.4699, + "step": 296 + }, + { + "epoch": 0.45904173106646057, + "grad_norm": 1.1226087523721424, + "learning_rate": 7.333333333333333e-05, + "loss": 0.4606, + "step": 297 + }, + { + "epoch": 0.4605873261205564, + "grad_norm": 1.0261605921557848, + "learning_rate": 7.358024691358025e-05, + "loss": 0.4664, + "step": 298 + }, + { + "epoch": 0.46213292117465227, + "grad_norm": 1.040973288606845, + "learning_rate": 7.382716049382717e-05, + "loss": 0.4696, + "step": 299 + }, + { + "epoch": 0.46367851622874806, + "grad_norm": 0.9783172608381693, + "learning_rate": 7.407407407407409e-05, + "loss": 0.4599, + "step": 300 + }, + { + "epoch": 0.4652241112828439, + "grad_norm": 1.2289274446920366, + "learning_rate": 7.432098765432099e-05, + "loss": 0.4676, + "step": 301 + }, + { + "epoch": 0.4667697063369397, + "grad_norm": 0.9551069659281435, + "learning_rate": 7.456790123456791e-05, + "loss": 0.4621, + "step": 302 + }, + { + "epoch": 0.46831530139103555, + "grad_norm": 0.9271166884767883, + "learning_rate": 7.481481481481483e-05, + "loss": 0.4612, + "step": 303 + }, + { + "epoch": 0.46986089644513135, + "grad_norm": 1.004864766717887, + "learning_rate": 7.506172839506173e-05, + "loss": 0.4683, + "step": 304 + }, + { + "epoch": 0.4714064914992272, + "grad_norm": 1.128627756522529, + "learning_rate": 7.530864197530865e-05, + "loss": 0.4709, + "step": 305 + }, + { + "epoch": 0.47295208655332305, + "grad_norm": 0.9321358642047468, + "learning_rate": 7.555555555555556e-05, + "loss": 0.4626, + "step": 306 + }, + { + "epoch": 0.47449768160741884, + "grad_norm": 1.3278999945984602, + "learning_rate": 7.580246913580247e-05, + "loss": 0.4748, + "step": 307 + }, + { + "epoch": 0.4760432766615147, + "grad_norm": 0.944025022165049, + "learning_rate": 7.60493827160494e-05, + "loss": 0.4606, + "step": 308 + }, + { + "epoch": 0.4775888717156105, + "grad_norm": 1.0799574982338942, + "learning_rate": 7.62962962962963e-05, + "loss": 0.4624, + "step": 309 + }, + { + "epoch": 0.47913446676970634, + "grad_norm": 0.8620477915536247, + "learning_rate": 7.654320987654322e-05, + "loss": 0.4704, + "step": 310 + }, + { + "epoch": 0.4806800618238022, + "grad_norm": 0.6727562335388572, + "learning_rate": 7.679012345679013e-05, + "loss": 0.4539, + "step": 311 + }, + { + "epoch": 0.482225656877898, + "grad_norm": 0.7125580919951122, + "learning_rate": 7.703703703703704e-05, + "loss": 0.465, + "step": 312 + }, + { + "epoch": 0.48377125193199383, + "grad_norm": 0.8850579621624475, + "learning_rate": 7.728395061728396e-05, + "loss": 0.4651, + "step": 313 + }, + { + "epoch": 0.4853168469860896, + "grad_norm": 1.089814285070669, + "learning_rate": 7.753086419753086e-05, + "loss": 0.4719, + "step": 314 + }, + { + "epoch": 0.4868624420401855, + "grad_norm": 1.0347330093917404, + "learning_rate": 7.777777777777778e-05, + "loss": 0.469, + "step": 315 + }, + { + "epoch": 0.4884080370942813, + "grad_norm": 1.033824468869202, + "learning_rate": 7.80246913580247e-05, + "loss": 0.4739, + "step": 316 + }, + { + "epoch": 0.4899536321483771, + "grad_norm": 1.0214619727971168, + "learning_rate": 7.827160493827162e-05, + "loss": 0.4636, + "step": 317 + }, + { + "epoch": 0.49149922720247297, + "grad_norm": 0.9265658300824663, + "learning_rate": 7.851851851851852e-05, + "loss": 0.4576, + "step": 318 + }, + { + "epoch": 0.49304482225656876, + "grad_norm": 0.9757407254253527, + "learning_rate": 7.876543209876544e-05, + "loss": 0.4696, + "step": 319 + }, + { + "epoch": 0.4945904173106646, + "grad_norm": 1.102852921361581, + "learning_rate": 7.901234567901235e-05, + "loss": 0.4649, + "step": 320 + }, + { + "epoch": 0.49613601236476046, + "grad_norm": 0.9164031059399039, + "learning_rate": 7.925925925925926e-05, + "loss": 0.468, + "step": 321 + }, + { + "epoch": 0.49768160741885625, + "grad_norm": 1.001279432380049, + "learning_rate": 7.950617283950618e-05, + "loss": 0.4604, + "step": 322 + }, + { + "epoch": 0.4992272024729521, + "grad_norm": 1.3988013409604274, + "learning_rate": 7.975308641975309e-05, + "loss": 0.4725, + "step": 323 + }, + { + "epoch": 0.500772797527048, + "grad_norm": 0.9225532294408699, + "learning_rate": 8e-05, + "loss": 0.4672, + "step": 324 + }, + { + "epoch": 0.5023183925811437, + "grad_norm": 1.2366631094121028, + "learning_rate": 7.999997670593474e-05, + "loss": 0.4743, + "step": 325 + }, + { + "epoch": 0.5038639876352395, + "grad_norm": 0.9985754515306058, + "learning_rate": 7.999990682376607e-05, + "loss": 0.4742, + "step": 326 + }, + { + "epoch": 0.5054095826893354, + "grad_norm": 1.77167590709832, + "learning_rate": 7.999979035357536e-05, + "loss": 0.4716, + "step": 327 + }, + { + "epoch": 0.5069551777434312, + "grad_norm": 0.9268029187821334, + "learning_rate": 7.99996272954983e-05, + "loss": 0.4537, + "step": 328 + }, + { + "epoch": 0.508500772797527, + "grad_norm": 2.2199583912571446, + "learning_rate": 7.99994176497248e-05, + "loss": 0.4732, + "step": 329 + }, + { + "epoch": 0.5100463678516228, + "grad_norm": 1.6042414500936655, + "learning_rate": 7.999916141649901e-05, + "loss": 0.4829, + "step": 330 + }, + { + "epoch": 0.5115919629057187, + "grad_norm": 1.4455600037251086, + "learning_rate": 7.99988585961194e-05, + "loss": 0.4691, + "step": 331 + }, + { + "epoch": 0.5131375579598145, + "grad_norm": 1.022348436352116, + "learning_rate": 7.999850918893863e-05, + "loss": 0.4641, + "step": 332 + }, + { + "epoch": 0.5146831530139103, + "grad_norm": 1.1934129252938614, + "learning_rate": 7.999811319536369e-05, + "loss": 0.4701, + "step": 333 + }, + { + "epoch": 0.5162287480680062, + "grad_norm": 0.7743917973611034, + "learning_rate": 7.999767061585576e-05, + "loss": 0.461, + "step": 334 + }, + { + "epoch": 0.517774343122102, + "grad_norm": 0.9949229362004298, + "learning_rate": 7.999718145093036e-05, + "loss": 0.4625, + "step": 335 + }, + { + "epoch": 0.5193199381761978, + "grad_norm": 0.5977321063031327, + "learning_rate": 7.999664570115717e-05, + "loss": 0.4738, + "step": 336 + }, + { + "epoch": 0.5208655332302936, + "grad_norm": 0.8157527643511809, + "learning_rate": 7.99960633671602e-05, + "loss": 0.4657, + "step": 337 + }, + { + "epoch": 0.5224111282843895, + "grad_norm": 0.7880571467871172, + "learning_rate": 7.999543444961772e-05, + "loss": 0.4755, + "step": 338 + }, + { + "epoch": 0.5239567233384853, + "grad_norm": 0.639800468540418, + "learning_rate": 7.999475894926221e-05, + "loss": 0.4589, + "step": 339 + }, + { + "epoch": 0.5255023183925811, + "grad_norm": 0.7688262011165063, + "learning_rate": 7.999403686688042e-05, + "loss": 0.4667, + "step": 340 + }, + { + "epoch": 0.527047913446677, + "grad_norm": 0.9245531439939358, + "learning_rate": 7.999326820331338e-05, + "loss": 0.4642, + "step": 341 + }, + { + "epoch": 0.5285935085007728, + "grad_norm": 1.322503342624274, + "learning_rate": 7.999245295945634e-05, + "loss": 0.4551, + "step": 342 + }, + { + "epoch": 0.5301391035548686, + "grad_norm": 0.6402251237059537, + "learning_rate": 7.999159113625881e-05, + "loss": 0.4567, + "step": 343 + }, + { + "epoch": 0.5316846986089645, + "grad_norm": 0.7517897833225893, + "learning_rate": 7.999068273472459e-05, + "loss": 0.4661, + "step": 344 + }, + { + "epoch": 0.5332302936630603, + "grad_norm": 1.0070553668165176, + "learning_rate": 7.998972775591167e-05, + "loss": 0.4711, + "step": 345 + }, + { + "epoch": 0.5347758887171561, + "grad_norm": 1.2274192513747975, + "learning_rate": 7.998872620093235e-05, + "loss": 0.466, + "step": 346 + }, + { + "epoch": 0.5363214837712519, + "grad_norm": 0.7117374365803301, + "learning_rate": 7.99876780709531e-05, + "loss": 0.4681, + "step": 347 + }, + { + "epoch": 0.5378670788253478, + "grad_norm": 0.8188752537996355, + "learning_rate": 7.99865833671947e-05, + "loss": 0.465, + "step": 348 + }, + { + "epoch": 0.5394126738794436, + "grad_norm": 0.875214173758601, + "learning_rate": 7.998544209093217e-05, + "loss": 0.4734, + "step": 349 + }, + { + "epoch": 0.5409582689335394, + "grad_norm": 0.659624722972033, + "learning_rate": 7.998425424349473e-05, + "loss": 0.4637, + "step": 350 + }, + { + "epoch": 0.5425038639876353, + "grad_norm": 0.7057348490226953, + "learning_rate": 7.99830198262659e-05, + "loss": 0.4623, + "step": 351 + }, + { + "epoch": 0.5440494590417311, + "grad_norm": 0.7050244930924285, + "learning_rate": 7.998173884068338e-05, + "loss": 0.454, + "step": 352 + }, + { + "epoch": 0.5455950540958269, + "grad_norm": 0.6020745367214642, + "learning_rate": 7.998041128823916e-05, + "loss": 0.4559, + "step": 353 + }, + { + "epoch": 0.5471406491499228, + "grad_norm": 0.6218265771052139, + "learning_rate": 7.997903717047944e-05, + "loss": 0.468, + "step": 354 + }, + { + "epoch": 0.5486862442040186, + "grad_norm": 0.7498605169579337, + "learning_rate": 7.997761648900466e-05, + "loss": 0.4568, + "step": 355 + }, + { + "epoch": 0.5502318392581144, + "grad_norm": 0.7386153906816254, + "learning_rate": 7.997614924546948e-05, + "loss": 0.4622, + "step": 356 + }, + { + "epoch": 0.5517774343122102, + "grad_norm": 0.8615583934080182, + "learning_rate": 7.997463544158282e-05, + "loss": 0.4632, + "step": 357 + }, + { + "epoch": 0.5533230293663061, + "grad_norm": 1.2489575879713575, + "learning_rate": 7.99730750791078e-05, + "loss": 0.4627, + "step": 358 + }, + { + "epoch": 0.5548686244204019, + "grad_norm": 0.8146213378195578, + "learning_rate": 7.997146815986179e-05, + "loss": 0.4568, + "step": 359 + }, + { + "epoch": 0.5564142194744977, + "grad_norm": 0.7922732907357454, + "learning_rate": 7.996981468571636e-05, + "loss": 0.4565, + "step": 360 + }, + { + "epoch": 0.5579598145285936, + "grad_norm": 0.6941937896489192, + "learning_rate": 7.996811465859733e-05, + "loss": 0.4517, + "step": 361 + }, + { + "epoch": 0.5595054095826894, + "grad_norm": 0.7332247912397324, + "learning_rate": 7.996636808048472e-05, + "loss": 0.463, + "step": 362 + }, + { + "epoch": 0.5610510046367851, + "grad_norm": 0.7508020875142575, + "learning_rate": 7.996457495341277e-05, + "loss": 0.4602, + "step": 363 + }, + { + "epoch": 0.5625965996908809, + "grad_norm": 0.756110893237292, + "learning_rate": 7.996273527946995e-05, + "loss": 0.4528, + "step": 364 + }, + { + "epoch": 0.5641421947449768, + "grad_norm": 0.7127020025169759, + "learning_rate": 7.996084906079895e-05, + "loss": 0.4545, + "step": 365 + }, + { + "epoch": 0.5656877897990726, + "grad_norm": 0.6905141182528117, + "learning_rate": 7.995891629959662e-05, + "loss": 0.4604, + "step": 366 + }, + { + "epoch": 0.5672333848531684, + "grad_norm": 0.8143923681191669, + "learning_rate": 7.995693699811409e-05, + "loss": 0.452, + "step": 367 + }, + { + "epoch": 0.5687789799072643, + "grad_norm": 1.0747779771959909, + "learning_rate": 7.995491115865662e-05, + "loss": 0.4471, + "step": 368 + }, + { + "epoch": 0.5703245749613601, + "grad_norm": 1.1193263847381563, + "learning_rate": 7.995283878358374e-05, + "loss": 0.462, + "step": 369 + }, + { + "epoch": 0.5718701700154559, + "grad_norm": 0.8604609875694763, + "learning_rate": 7.995071987530914e-05, + "loss": 0.4547, + "step": 370 + }, + { + "epoch": 0.5734157650695518, + "grad_norm": 0.8277466123842292, + "learning_rate": 7.994855443630072e-05, + "loss": 0.4564, + "step": 371 + }, + { + "epoch": 0.5749613601236476, + "grad_norm": 0.9834357821168109, + "learning_rate": 7.994634246908058e-05, + "loss": 0.4588, + "step": 372 + }, + { + "epoch": 0.5765069551777434, + "grad_norm": 1.1250087341862804, + "learning_rate": 7.994408397622501e-05, + "loss": 0.4592, + "step": 373 + }, + { + "epoch": 0.5780525502318392, + "grad_norm": 0.8218693286115759, + "learning_rate": 7.994177896036446e-05, + "loss": 0.4585, + "step": 374 + }, + { + "epoch": 0.5795981452859351, + "grad_norm": 0.8271377679531405, + "learning_rate": 7.993942742418362e-05, + "loss": 0.4606, + "step": 375 + }, + { + "epoch": 0.5811437403400309, + "grad_norm": 1.0470920301230913, + "learning_rate": 7.993702937042131e-05, + "loss": 0.4532, + "step": 376 + }, + { + "epoch": 0.5826893353941267, + "grad_norm": 1.1461008670711745, + "learning_rate": 7.993458480187057e-05, + "loss": 0.4711, + "step": 377 + }, + { + "epoch": 0.5842349304482226, + "grad_norm": 0.8038032056131119, + "learning_rate": 7.993209372137858e-05, + "loss": 0.4605, + "step": 378 + }, + { + "epoch": 0.5857805255023184, + "grad_norm": 0.9733659124496695, + "learning_rate": 7.99295561318467e-05, + "loss": 0.4585, + "step": 379 + }, + { + "epoch": 0.5873261205564142, + "grad_norm": 1.1224303019272235, + "learning_rate": 7.99269720362305e-05, + "loss": 0.4624, + "step": 380 + }, + { + "epoch": 0.58887171561051, + "grad_norm": 0.6177695872757162, + "learning_rate": 7.992434143753968e-05, + "loss": 0.4459, + "step": 381 + }, + { + "epoch": 0.5904173106646059, + "grad_norm": 0.5624887565167997, + "learning_rate": 7.992166433883807e-05, + "loss": 0.4531, + "step": 382 + }, + { + "epoch": 0.5919629057187017, + "grad_norm": 0.6382536533975867, + "learning_rate": 7.991894074324374e-05, + "loss": 0.4565, + "step": 383 + }, + { + "epoch": 0.5935085007727975, + "grad_norm": 0.7222552352500594, + "learning_rate": 7.991617065392884e-05, + "loss": 0.4451, + "step": 384 + }, + { + "epoch": 0.5950540958268934, + "grad_norm": 0.8115172194271355, + "learning_rate": 7.99133540741197e-05, + "loss": 0.4521, + "step": 385 + }, + { + "epoch": 0.5965996908809892, + "grad_norm": 0.9417727734786303, + "learning_rate": 7.991049100709684e-05, + "loss": 0.4543, + "step": 386 + }, + { + "epoch": 0.598145285935085, + "grad_norm": 0.9514296085746278, + "learning_rate": 7.990758145619484e-05, + "loss": 0.4561, + "step": 387 + }, + { + "epoch": 0.5996908809891809, + "grad_norm": 0.9714733033456525, + "learning_rate": 7.99046254248025e-05, + "loss": 0.451, + "step": 388 + }, + { + "epoch": 0.6012364760432767, + "grad_norm": 0.9844237398266047, + "learning_rate": 7.990162291636269e-05, + "loss": 0.4504, + "step": 389 + }, + { + "epoch": 0.6027820710973725, + "grad_norm": 0.9369492734406104, + "learning_rate": 7.989857393437245e-05, + "loss": 0.4585, + "step": 390 + }, + { + "epoch": 0.6043276661514683, + "grad_norm": 0.7551208261075399, + "learning_rate": 7.989547848238294e-05, + "loss": 0.4452, + "step": 391 + }, + { + "epoch": 0.6058732612055642, + "grad_norm": 0.6232383668619544, + "learning_rate": 7.989233656399944e-05, + "loss": 0.4587, + "step": 392 + }, + { + "epoch": 0.60741885625966, + "grad_norm": 0.4901931334082307, + "learning_rate": 7.988914818288136e-05, + "loss": 0.4572, + "step": 393 + }, + { + "epoch": 0.6089644513137558, + "grad_norm": 0.47374939525246373, + "learning_rate": 7.988591334274221e-05, + "loss": 0.4446, + "step": 394 + }, + { + "epoch": 0.6105100463678517, + "grad_norm": 0.4596712513868155, + "learning_rate": 7.988263204734962e-05, + "loss": 0.4524, + "step": 395 + }, + { + "epoch": 0.6120556414219475, + "grad_norm": 0.4282353169289228, + "learning_rate": 7.987930430052533e-05, + "loss": 0.4577, + "step": 396 + }, + { + "epoch": 0.6136012364760433, + "grad_norm": 0.5099867320381566, + "learning_rate": 7.98759301061452e-05, + "loss": 0.4485, + "step": 397 + }, + { + "epoch": 0.615146831530139, + "grad_norm": 0.539515319591019, + "learning_rate": 7.987250946813911e-05, + "loss": 0.4493, + "step": 398 + }, + { + "epoch": 0.616692426584235, + "grad_norm": 0.6309528596296303, + "learning_rate": 7.986904239049113e-05, + "loss": 0.4644, + "step": 399 + }, + { + "epoch": 0.6182380216383307, + "grad_norm": 0.7946789370337108, + "learning_rate": 7.986552887723937e-05, + "loss": 0.4513, + "step": 400 + }, + { + "epoch": 0.6197836166924265, + "grad_norm": 0.9655674223669141, + "learning_rate": 7.986196893247603e-05, + "loss": 0.4518, + "step": 401 + }, + { + "epoch": 0.6213292117465224, + "grad_norm": 1.1337471741612535, + "learning_rate": 7.985836256034737e-05, + "loss": 0.4604, + "step": 402 + }, + { + "epoch": 0.6228748068006182, + "grad_norm": 0.8273182843501121, + "learning_rate": 7.985470976505378e-05, + "loss": 0.4567, + "step": 403 + }, + { + "epoch": 0.624420401854714, + "grad_norm": 0.6777953490489269, + "learning_rate": 7.985101055084963e-05, + "loss": 0.4463, + "step": 404 + }, + { + "epoch": 0.6259659969088099, + "grad_norm": 0.7153814353017776, + "learning_rate": 7.984726492204346e-05, + "loss": 0.4412, + "step": 405 + }, + { + "epoch": 0.6275115919629057, + "grad_norm": 0.7425537671761728, + "learning_rate": 7.98434728829978e-05, + "loss": 0.4569, + "step": 406 + }, + { + "epoch": 0.6290571870170015, + "grad_norm": 0.839336233607707, + "learning_rate": 7.983963443812924e-05, + "loss": 0.446, + "step": 407 + }, + { + "epoch": 0.6306027820710973, + "grad_norm": 0.9586866438462934, + "learning_rate": 7.983574959190843e-05, + "loss": 0.4535, + "step": 408 + }, + { + "epoch": 0.6321483771251932, + "grad_norm": 0.9765270261026754, + "learning_rate": 7.983181834886006e-05, + "loss": 0.4475, + "step": 409 + }, + { + "epoch": 0.633693972179289, + "grad_norm": 0.759582420091956, + "learning_rate": 7.982784071356289e-05, + "loss": 0.452, + "step": 410 + }, + { + "epoch": 0.6352395672333848, + "grad_norm": 0.6353310663111404, + "learning_rate": 7.982381669064965e-05, + "loss": 0.4421, + "step": 411 + }, + { + "epoch": 0.6367851622874807, + "grad_norm": 0.7361906306338643, + "learning_rate": 7.981974628480715e-05, + "loss": 0.4438, + "step": 412 + }, + { + "epoch": 0.6383307573415765, + "grad_norm": 0.6339865173716746, + "learning_rate": 7.981562950077619e-05, + "loss": 0.4485, + "step": 413 + }, + { + "epoch": 0.6398763523956723, + "grad_norm": 0.5435774755527445, + "learning_rate": 7.981146634335163e-05, + "loss": 0.4536, + "step": 414 + }, + { + "epoch": 0.6414219474497682, + "grad_norm": 0.5710868607662681, + "learning_rate": 7.980725681738229e-05, + "loss": 0.4548, + "step": 415 + }, + { + "epoch": 0.642967542503864, + "grad_norm": 0.5911308867835612, + "learning_rate": 7.980300092777102e-05, + "loss": 0.451, + "step": 416 + }, + { + "epoch": 0.6445131375579598, + "grad_norm": 0.7225191067183553, + "learning_rate": 7.979869867947468e-05, + "loss": 0.4417, + "step": 417 + }, + { + "epoch": 0.6460587326120556, + "grad_norm": 0.7521628176249268, + "learning_rate": 7.97943500775041e-05, + "loss": 0.4402, + "step": 418 + }, + { + "epoch": 0.6476043276661515, + "grad_norm": 0.678780273863019, + "learning_rate": 7.978995512692412e-05, + "loss": 0.456, + "step": 419 + }, + { + "epoch": 0.6491499227202473, + "grad_norm": 0.7182629873142787, + "learning_rate": 7.978551383285354e-05, + "loss": 0.4551, + "step": 420 + }, + { + "epoch": 0.6506955177743431, + "grad_norm": 0.8393864782623631, + "learning_rate": 7.978102620046517e-05, + "loss": 0.4504, + "step": 421 + }, + { + "epoch": 0.652241112828439, + "grad_norm": 1.133016706790667, + "learning_rate": 7.977649223498575e-05, + "loss": 0.4514, + "step": 422 + }, + { + "epoch": 0.6537867078825348, + "grad_norm": 1.03878671480764, + "learning_rate": 7.977191194169603e-05, + "loss": 0.4639, + "step": 423 + }, + { + "epoch": 0.6553323029366306, + "grad_norm": 0.8444433492995927, + "learning_rate": 7.976728532593067e-05, + "loss": 0.4491, + "step": 424 + }, + { + "epoch": 0.6568778979907264, + "grad_norm": 0.6960863030640867, + "learning_rate": 7.97626123930783e-05, + "loss": 0.4493, + "step": 425 + }, + { + "epoch": 0.6584234930448223, + "grad_norm": 0.7005072676048999, + "learning_rate": 7.975789314858153e-05, + "loss": 0.4476, + "step": 426 + }, + { + "epoch": 0.6599690880989181, + "grad_norm": 0.836746348938439, + "learning_rate": 7.975312759793685e-05, + "loss": 0.4512, + "step": 427 + }, + { + "epoch": 0.6615146831530139, + "grad_norm": 1.0689325787549664, + "learning_rate": 7.974831574669472e-05, + "loss": 0.4627, + "step": 428 + }, + { + "epoch": 0.6630602782071098, + "grad_norm": 1.1186581408986205, + "learning_rate": 7.974345760045953e-05, + "loss": 0.4488, + "step": 429 + }, + { + "epoch": 0.6646058732612056, + "grad_norm": 0.6777940465736243, + "learning_rate": 7.973855316488957e-05, + "loss": 0.4492, + "step": 430 + }, + { + "epoch": 0.6661514683153014, + "grad_norm": 0.5338973438235535, + "learning_rate": 7.973360244569705e-05, + "loss": 0.4431, + "step": 431 + }, + { + "epoch": 0.6676970633693973, + "grad_norm": 0.6210050484906554, + "learning_rate": 7.972860544864809e-05, + "loss": 0.4541, + "step": 432 + }, + { + "epoch": 0.6692426584234931, + "grad_norm": 0.7162208131594553, + "learning_rate": 7.972356217956272e-05, + "loss": 0.4578, + "step": 433 + }, + { + "epoch": 0.6707882534775889, + "grad_norm": 0.708321211019045, + "learning_rate": 7.971847264431482e-05, + "loss": 0.4498, + "step": 434 + }, + { + "epoch": 0.6723338485316847, + "grad_norm": 0.5917552072415284, + "learning_rate": 7.971333684883223e-05, + "loss": 0.4521, + "step": 435 + }, + { + "epoch": 0.6738794435857806, + "grad_norm": 0.5189396718295533, + "learning_rate": 7.970815479909661e-05, + "loss": 0.4433, + "step": 436 + }, + { + "epoch": 0.6754250386398764, + "grad_norm": 0.544043936139772, + "learning_rate": 7.970292650114351e-05, + "loss": 0.4478, + "step": 437 + }, + { + "epoch": 0.6769706336939721, + "grad_norm": 0.4283212741759581, + "learning_rate": 7.969765196106233e-05, + "loss": 0.4454, + "step": 438 + }, + { + "epoch": 0.678516228748068, + "grad_norm": 0.35731818057965875, + "learning_rate": 7.969233118499636e-05, + "loss": 0.4521, + "step": 439 + }, + { + "epoch": 0.6800618238021638, + "grad_norm": 0.49696678982887327, + "learning_rate": 7.968696417914274e-05, + "loss": 0.4401, + "step": 440 + }, + { + "epoch": 0.6816074188562596, + "grad_norm": 0.496159232355586, + "learning_rate": 7.968155094975243e-05, + "loss": 0.4458, + "step": 441 + }, + { + "epoch": 0.6831530139103554, + "grad_norm": 0.3930491418328142, + "learning_rate": 7.96760915031302e-05, + "loss": 0.4505, + "step": 442 + }, + { + "epoch": 0.6846986089644513, + "grad_norm": 0.41749812940055175, + "learning_rate": 7.967058584563471e-05, + "loss": 0.4625, + "step": 443 + }, + { + "epoch": 0.6862442040185471, + "grad_norm": 0.5552461791440614, + "learning_rate": 7.966503398367845e-05, + "loss": 0.4489, + "step": 444 + }, + { + "epoch": 0.6877897990726429, + "grad_norm": 0.539358431362722, + "learning_rate": 7.965943592372766e-05, + "loss": 0.4422, + "step": 445 + }, + { + "epoch": 0.6893353941267388, + "grad_norm": 0.5055158333988253, + "learning_rate": 7.965379167230241e-05, + "loss": 0.4557, + "step": 446 + }, + { + "epoch": 0.6908809891808346, + "grad_norm": 0.6860788834653874, + "learning_rate": 7.964810123597659e-05, + "loss": 0.4424, + "step": 447 + }, + { + "epoch": 0.6924265842349304, + "grad_norm": 1.011035092699708, + "learning_rate": 7.964236462137788e-05, + "loss": 0.4522, + "step": 448 + }, + { + "epoch": 0.6939721792890263, + "grad_norm": 1.3481812963210864, + "learning_rate": 7.963658183518772e-05, + "loss": 0.4637, + "step": 449 + }, + { + "epoch": 0.6955177743431221, + "grad_norm": 0.5038082193164464, + "learning_rate": 7.963075288414134e-05, + "loss": 0.4402, + "step": 450 + }, + { + "epoch": 0.6970633693972179, + "grad_norm": 0.8801962675300073, + "learning_rate": 7.962487777502773e-05, + "loss": 0.4537, + "step": 451 + }, + { + "epoch": 0.6986089644513137, + "grad_norm": 1.3239962632072204, + "learning_rate": 7.961895651468967e-05, + "loss": 0.4641, + "step": 452 + }, + { + "epoch": 0.7001545595054096, + "grad_norm": 0.36296382947469763, + "learning_rate": 7.961298911002367e-05, + "loss": 0.4539, + "step": 453 + }, + { + "epoch": 0.7017001545595054, + "grad_norm": 1.229972375753171, + "learning_rate": 7.960697556797997e-05, + "loss": 0.4538, + "step": 454 + }, + { + "epoch": 0.7032457496136012, + "grad_norm": 0.7635879409253314, + "learning_rate": 7.960091589556257e-05, + "loss": 0.4514, + "step": 455 + }, + { + "epoch": 0.7047913446676971, + "grad_norm": 0.6558156495845362, + "learning_rate": 7.959481009982919e-05, + "loss": 0.4442, + "step": 456 + }, + { + "epoch": 0.7063369397217929, + "grad_norm": 1.0384695968547808, + "learning_rate": 7.958865818789127e-05, + "loss": 0.4493, + "step": 457 + }, + { + "epoch": 0.7078825347758887, + "grad_norm": 1.0235891024753911, + "learning_rate": 7.958246016691396e-05, + "loss": 0.4405, + "step": 458 + }, + { + "epoch": 0.7094281298299846, + "grad_norm": 0.9251493068561275, + "learning_rate": 7.957621604411612e-05, + "loss": 0.4545, + "step": 459 + }, + { + "epoch": 0.7109737248840804, + "grad_norm": 0.8685368960150491, + "learning_rate": 7.95699258267703e-05, + "loss": 0.4431, + "step": 460 + }, + { + "epoch": 0.7125193199381762, + "grad_norm": 0.9488476837307872, + "learning_rate": 7.956358952220273e-05, + "loss": 0.4419, + "step": 461 + }, + { + "epoch": 0.714064914992272, + "grad_norm": 1.2420512500174485, + "learning_rate": 7.955720713779332e-05, + "loss": 0.4504, + "step": 462 + }, + { + "epoch": 0.7156105100463679, + "grad_norm": 0.6758251019013465, + "learning_rate": 7.955077868097568e-05, + "loss": 0.444, + "step": 463 + }, + { + "epoch": 0.7171561051004637, + "grad_norm": 0.6734311630977253, + "learning_rate": 7.954430415923703e-05, + "loss": 0.4499, + "step": 464 + }, + { + "epoch": 0.7187017001545595, + "grad_norm": 0.90449672539994, + "learning_rate": 7.953778358011827e-05, + "loss": 0.449, + "step": 465 + }, + { + "epoch": 0.7202472952086554, + "grad_norm": 1.3061506463069403, + "learning_rate": 7.953121695121395e-05, + "loss": 0.4499, + "step": 466 + }, + { + "epoch": 0.7217928902627512, + "grad_norm": 0.5869778083359343, + "learning_rate": 7.952460428017223e-05, + "loss": 0.4469, + "step": 467 + }, + { + "epoch": 0.723338485316847, + "grad_norm": 0.41164098744813804, + "learning_rate": 7.951794557469492e-05, + "loss": 0.4514, + "step": 468 + }, + { + "epoch": 0.7248840803709428, + "grad_norm": 0.7840446972608712, + "learning_rate": 7.951124084253745e-05, + "loss": 0.4445, + "step": 469 + }, + { + "epoch": 0.7264296754250387, + "grad_norm": 1.1636968066294135, + "learning_rate": 7.95044900915088e-05, + "loss": 0.4482, + "step": 470 + }, + { + "epoch": 0.7279752704791345, + "grad_norm": 0.7797641335885679, + "learning_rate": 7.949769332947162e-05, + "loss": 0.4338, + "step": 471 + }, + { + "epoch": 0.7295208655332303, + "grad_norm": 0.5014085938497415, + "learning_rate": 7.949085056434213e-05, + "loss": 0.4622, + "step": 472 + }, + { + "epoch": 0.7310664605873262, + "grad_norm": 0.7769090079710187, + "learning_rate": 7.948396180409012e-05, + "loss": 0.4544, + "step": 473 + }, + { + "epoch": 0.732612055641422, + "grad_norm": 1.1171659903459716, + "learning_rate": 7.947702705673893e-05, + "loss": 0.448, + "step": 474 + }, + { + "epoch": 0.7341576506955177, + "grad_norm": 0.7619870379947891, + "learning_rate": 7.947004633036548e-05, + "loss": 0.4439, + "step": 475 + }, + { + "epoch": 0.7357032457496137, + "grad_norm": 0.6685466568508059, + "learning_rate": 7.946301963310028e-05, + "loss": 0.4541, + "step": 476 + }, + { + "epoch": 0.7372488408037094, + "grad_norm": 0.643490097890329, + "learning_rate": 7.945594697312732e-05, + "loss": 0.4511, + "step": 477 + }, + { + "epoch": 0.7387944358578052, + "grad_norm": 0.5896476046955875, + "learning_rate": 7.944882835868417e-05, + "loss": 0.4476, + "step": 478 + }, + { + "epoch": 0.740340030911901, + "grad_norm": 0.7100375800765096, + "learning_rate": 7.944166379806186e-05, + "loss": 0.453, + "step": 479 + }, + { + "epoch": 0.7418856259659969, + "grad_norm": 0.8256463904622523, + "learning_rate": 7.943445329960503e-05, + "loss": 0.4529, + "step": 480 + }, + { + "epoch": 0.7434312210200927, + "grad_norm": 0.6962258117516658, + "learning_rate": 7.942719687171175e-05, + "loss": 0.4492, + "step": 481 + }, + { + "epoch": 0.7449768160741885, + "grad_norm": 0.6381012396696243, + "learning_rate": 7.941989452283359e-05, + "loss": 0.4483, + "step": 482 + }, + { + "epoch": 0.7465224111282844, + "grad_norm": 0.8129367906824256, + "learning_rate": 7.941254626147565e-05, + "loss": 0.4431, + "step": 483 + }, + { + "epoch": 0.7480680061823802, + "grad_norm": 0.7781368768779827, + "learning_rate": 7.940515209619644e-05, + "loss": 0.438, + "step": 484 + }, + { + "epoch": 0.749613601236476, + "grad_norm": 0.6249515855342817, + "learning_rate": 7.939771203560799e-05, + "loss": 0.4431, + "step": 485 + }, + { + "epoch": 0.7511591962905718, + "grad_norm": 0.5094482079908089, + "learning_rate": 7.939022608837576e-05, + "loss": 0.4405, + "step": 486 + }, + { + "epoch": 0.7527047913446677, + "grad_norm": 0.575246536145531, + "learning_rate": 7.938269426321867e-05, + "loss": 0.4394, + "step": 487 + }, + { + "epoch": 0.7542503863987635, + "grad_norm": 0.725728800009646, + "learning_rate": 7.937511656890903e-05, + "loss": 0.4352, + "step": 488 + }, + { + "epoch": 0.7557959814528593, + "grad_norm": 0.8217792413157236, + "learning_rate": 7.936749301427264e-05, + "loss": 0.4437, + "step": 489 + }, + { + "epoch": 0.7573415765069552, + "grad_norm": 0.9119629982005905, + "learning_rate": 7.935982360818863e-05, + "loss": 0.439, + "step": 490 + }, + { + "epoch": 0.758887171561051, + "grad_norm": 0.9928588690645079, + "learning_rate": 7.935210835958963e-05, + "loss": 0.4581, + "step": 491 + }, + { + "epoch": 0.7604327666151468, + "grad_norm": 0.9954906017773287, + "learning_rate": 7.934434727746161e-05, + "loss": 0.4481, + "step": 492 + }, + { + "epoch": 0.7619783616692427, + "grad_norm": 0.7651782309999712, + "learning_rate": 7.93365403708439e-05, + "loss": 0.4524, + "step": 493 + }, + { + "epoch": 0.7635239567233385, + "grad_norm": 0.4537098778667196, + "learning_rate": 7.932868764882926e-05, + "loss": 0.4452, + "step": 494 + }, + { + "epoch": 0.7650695517774343, + "grad_norm": 0.41264029448347583, + "learning_rate": 7.932078912056377e-05, + "loss": 0.446, + "step": 495 + }, + { + "epoch": 0.7666151468315301, + "grad_norm": 0.5624666992251612, + "learning_rate": 7.931284479524685e-05, + "loss": 0.4421, + "step": 496 + }, + { + "epoch": 0.768160741885626, + "grad_norm": 0.6132371242178343, + "learning_rate": 7.930485468213131e-05, + "loss": 0.4372, + "step": 497 + }, + { + "epoch": 0.7697063369397218, + "grad_norm": 0.5700958776156412, + "learning_rate": 7.929681879052325e-05, + "loss": 0.4441, + "step": 498 + }, + { + "epoch": 0.7712519319938176, + "grad_norm": 0.6025219052546477, + "learning_rate": 7.92887371297821e-05, + "loss": 0.4442, + "step": 499 + }, + { + "epoch": 0.7727975270479135, + "grad_norm": 0.7623541245497347, + "learning_rate": 7.92806097093206e-05, + "loss": 0.4405, + "step": 500 + }, + { + "epoch": 0.7743431221020093, + "grad_norm": 0.8822025006890616, + "learning_rate": 7.927243653860478e-05, + "loss": 0.4574, + "step": 501 + }, + { + "epoch": 0.7758887171561051, + "grad_norm": 0.82008990123641, + "learning_rate": 7.926421762715395e-05, + "loss": 0.4386, + "step": 502 + }, + { + "epoch": 0.7774343122102009, + "grad_norm": 0.6836316562948456, + "learning_rate": 7.925595298454073e-05, + "loss": 0.4471, + "step": 503 + }, + { + "epoch": 0.7789799072642968, + "grad_norm": 0.6645314827937467, + "learning_rate": 7.924764262039095e-05, + "loss": 0.4456, + "step": 504 + }, + { + "epoch": 0.7805255023183926, + "grad_norm": 0.6024413666963008, + "learning_rate": 7.923928654438371e-05, + "loss": 0.4477, + "step": 505 + }, + { + "epoch": 0.7820710973724884, + "grad_norm": 0.6067806497405188, + "learning_rate": 7.92308847662514e-05, + "loss": 0.4482, + "step": 506 + }, + { + "epoch": 0.7836166924265843, + "grad_norm": 0.8022751574120819, + "learning_rate": 7.922243729577957e-05, + "loss": 0.4496, + "step": 507 + }, + { + "epoch": 0.7851622874806801, + "grad_norm": 1.0187980123821614, + "learning_rate": 7.9213944142807e-05, + "loss": 0.4443, + "step": 508 + }, + { + "epoch": 0.7867078825347759, + "grad_norm": 0.9559480149445991, + "learning_rate": 7.920540531722572e-05, + "loss": 0.4455, + "step": 509 + }, + { + "epoch": 0.7882534775888718, + "grad_norm": 0.653004430829866, + "learning_rate": 7.919682082898092e-05, + "loss": 0.4399, + "step": 510 + }, + { + "epoch": 0.7897990726429676, + "grad_norm": 0.4991146707421027, + "learning_rate": 7.9188190688071e-05, + "loss": 0.4457, + "step": 511 + }, + { + "epoch": 0.7913446676970634, + "grad_norm": 0.7431054178473085, + "learning_rate": 7.917951490454748e-05, + "loss": 0.4496, + "step": 512 + }, + { + "epoch": 0.7928902627511591, + "grad_norm": 0.8725524077390574, + "learning_rate": 7.917079348851509e-05, + "loss": 0.4477, + "step": 513 + }, + { + "epoch": 0.794435857805255, + "grad_norm": 0.6178648850982248, + "learning_rate": 7.916202645013169e-05, + "loss": 0.4518, + "step": 514 + }, + { + "epoch": 0.7959814528593508, + "grad_norm": 0.4298646959576983, + "learning_rate": 7.915321379960829e-05, + "loss": 0.4499, + "step": 515 + }, + { + "epoch": 0.7975270479134466, + "grad_norm": 0.617294586390934, + "learning_rate": 7.9144355547209e-05, + "loss": 0.4519, + "step": 516 + }, + { + "epoch": 0.7990726429675425, + "grad_norm": 0.7476939585049011, + "learning_rate": 7.913545170325105e-05, + "loss": 0.4412, + "step": 517 + }, + { + "epoch": 0.8006182380216383, + "grad_norm": 0.7296163770474515, + "learning_rate": 7.912650227810478e-05, + "loss": 0.4353, + "step": 518 + }, + { + "epoch": 0.8021638330757341, + "grad_norm": 0.7674361690186301, + "learning_rate": 7.911750728219362e-05, + "loss": 0.4375, + "step": 519 + }, + { + "epoch": 0.80370942812983, + "grad_norm": 0.8319004424044151, + "learning_rate": 7.910846672599408e-05, + "loss": 0.4402, + "step": 520 + }, + { + "epoch": 0.8052550231839258, + "grad_norm": 0.7935095223866093, + "learning_rate": 7.909938062003572e-05, + "loss": 0.4432, + "step": 521 + }, + { + "epoch": 0.8068006182380216, + "grad_norm": 0.7275342789621305, + "learning_rate": 7.909024897490114e-05, + "loss": 0.4422, + "step": 522 + }, + { + "epoch": 0.8083462132921174, + "grad_norm": 0.6875408362152193, + "learning_rate": 7.9081071801226e-05, + "loss": 0.4446, + "step": 523 + }, + { + "epoch": 0.8098918083462133, + "grad_norm": 0.5489098475992172, + "learning_rate": 7.9071849109699e-05, + "loss": 0.4453, + "step": 524 + }, + { + "epoch": 0.8114374034003091, + "grad_norm": 0.4948536954876545, + "learning_rate": 7.906258091106184e-05, + "loss": 0.4439, + "step": 525 + }, + { + "epoch": 0.8129829984544049, + "grad_norm": 0.4895935666593852, + "learning_rate": 7.90532672161092e-05, + "loss": 0.4532, + "step": 526 + }, + { + "epoch": 0.8145285935085008, + "grad_norm": 0.4699864148512322, + "learning_rate": 7.904390803568879e-05, + "loss": 0.4379, + "step": 527 + }, + { + "epoch": 0.8160741885625966, + "grad_norm": 0.5369887995106534, + "learning_rate": 7.903450338070126e-05, + "loss": 0.4465, + "step": 528 + }, + { + "epoch": 0.8176197836166924, + "grad_norm": 0.6633108394754598, + "learning_rate": 7.902505326210028e-05, + "loss": 0.4384, + "step": 529 + }, + { + "epoch": 0.8191653786707882, + "grad_norm": 0.5780383800688729, + "learning_rate": 7.90155576908924e-05, + "loss": 0.4446, + "step": 530 + }, + { + "epoch": 0.8207109737248841, + "grad_norm": 0.5181731618383103, + "learning_rate": 7.900601667813714e-05, + "loss": 0.4508, + "step": 531 + }, + { + "epoch": 0.8222565687789799, + "grad_norm": 0.6003510684977706, + "learning_rate": 7.899643023494695e-05, + "loss": 0.4379, + "step": 532 + }, + { + "epoch": 0.8238021638330757, + "grad_norm": 0.6489577208509044, + "learning_rate": 7.898679837248723e-05, + "loss": 0.4401, + "step": 533 + }, + { + "epoch": 0.8253477588871716, + "grad_norm": 0.7106579431009961, + "learning_rate": 7.89771211019762e-05, + "loss": 0.4481, + "step": 534 + }, + { + "epoch": 0.8268933539412674, + "grad_norm": 0.8120044732059997, + "learning_rate": 7.896739843468502e-05, + "loss": 0.4339, + "step": 535 + }, + { + "epoch": 0.8284389489953632, + "grad_norm": 0.8242079009990684, + "learning_rate": 7.895763038193771e-05, + "loss": 0.4476, + "step": 536 + }, + { + "epoch": 0.8299845440494591, + "grad_norm": 0.9028256658800938, + "learning_rate": 7.894781695511117e-05, + "loss": 0.4407, + "step": 537 + }, + { + "epoch": 0.8315301391035549, + "grad_norm": 0.8808561340581151, + "learning_rate": 7.893795816563511e-05, + "loss": 0.4309, + "step": 538 + }, + { + "epoch": 0.8330757341576507, + "grad_norm": 0.8058197940502757, + "learning_rate": 7.89280540249921e-05, + "loss": 0.4406, + "step": 539 + }, + { + "epoch": 0.8346213292117465, + "grad_norm": 0.7700753495449788, + "learning_rate": 7.891810454471753e-05, + "loss": 0.4502, + "step": 540 + }, + { + "epoch": 0.8361669242658424, + "grad_norm": 0.7626009050964863, + "learning_rate": 7.890810973639958e-05, + "loss": 0.4425, + "step": 541 + }, + { + "epoch": 0.8377125193199382, + "grad_norm": 0.5970485832623469, + "learning_rate": 7.889806961167924e-05, + "loss": 0.4389, + "step": 542 + }, + { + "epoch": 0.839258114374034, + "grad_norm": 0.4132167503544843, + "learning_rate": 7.88879841822503e-05, + "loss": 0.4484, + "step": 543 + }, + { + "epoch": 0.8408037094281299, + "grad_norm": 0.561299960687465, + "learning_rate": 7.887785345985926e-05, + "loss": 0.4493, + "step": 544 + }, + { + "epoch": 0.8423493044822257, + "grad_norm": 0.8945919322396804, + "learning_rate": 7.886767745630542e-05, + "loss": 0.4432, + "step": 545 + }, + { + "epoch": 0.8438948995363215, + "grad_norm": 1.128259316099117, + "learning_rate": 7.88574561834408e-05, + "loss": 0.4399, + "step": 546 + }, + { + "epoch": 0.8454404945904173, + "grad_norm": 0.6004881396794288, + "learning_rate": 7.884718965317016e-05, + "loss": 0.4307, + "step": 547 + }, + { + "epoch": 0.8469860896445132, + "grad_norm": 0.501549489983025, + "learning_rate": 7.883687787745095e-05, + "loss": 0.436, + "step": 548 + }, + { + "epoch": 0.848531684698609, + "grad_norm": 0.8233296170013034, + "learning_rate": 7.882652086829333e-05, + "loss": 0.4281, + "step": 549 + }, + { + "epoch": 0.8500772797527048, + "grad_norm": 0.9394743239402724, + "learning_rate": 7.881611863776014e-05, + "loss": 0.4493, + "step": 550 + }, + { + "epoch": 0.8516228748068007, + "grad_norm": 0.9023855999242737, + "learning_rate": 7.880567119796692e-05, + "loss": 0.4328, + "step": 551 + }, + { + "epoch": 0.8531684698608965, + "grad_norm": 0.8111759904343439, + "learning_rate": 7.879517856108178e-05, + "loss": 0.441, + "step": 552 + }, + { + "epoch": 0.8547140649149922, + "grad_norm": 0.6977912406931622, + "learning_rate": 7.87846407393256e-05, + "loss": 0.4514, + "step": 553 + }, + { + "epoch": 0.8562596599690881, + "grad_norm": 0.5293324244486036, + "learning_rate": 7.877405774497176e-05, + "loss": 0.4334, + "step": 554 + }, + { + "epoch": 0.8578052550231839, + "grad_norm": 0.4023348521021668, + "learning_rate": 7.87634295903463e-05, + "loss": 0.4426, + "step": 555 + }, + { + "epoch": 0.8593508500772797, + "grad_norm": 0.48562563002908576, + "learning_rate": 7.875275628782794e-05, + "loss": 0.4395, + "step": 556 + }, + { + "epoch": 0.8608964451313755, + "grad_norm": 0.5600880982432517, + "learning_rate": 7.874203784984783e-05, + "loss": 0.451, + "step": 557 + }, + { + "epoch": 0.8624420401854714, + "grad_norm": 0.4867775949500963, + "learning_rate": 7.873127428888982e-05, + "loss": 0.4443, + "step": 558 + }, + { + "epoch": 0.8639876352395672, + "grad_norm": 0.45670488658685837, + "learning_rate": 7.872046561749023e-05, + "loss": 0.44, + "step": 559 + }, + { + "epoch": 0.865533230293663, + "grad_norm": 0.4761398613593943, + "learning_rate": 7.870961184823799e-05, + "loss": 0.4342, + "step": 560 + }, + { + "epoch": 0.8670788253477589, + "grad_norm": 0.46409040634999327, + "learning_rate": 7.869871299377449e-05, + "loss": 0.4304, + "step": 561 + }, + { + "epoch": 0.8686244204018547, + "grad_norm": 0.5109193076120637, + "learning_rate": 7.868776906679369e-05, + "loss": 0.4364, + "step": 562 + }, + { + "epoch": 0.8701700154559505, + "grad_norm": 0.6469465230363364, + "learning_rate": 7.867678008004199e-05, + "loss": 0.4416, + "step": 563 + }, + { + "epoch": 0.8717156105100463, + "grad_norm": 0.789980590549452, + "learning_rate": 7.866574604631832e-05, + "loss": 0.4425, + "step": 564 + }, + { + "epoch": 0.8732612055641422, + "grad_norm": 0.8933023925413848, + "learning_rate": 7.865466697847402e-05, + "loss": 0.4346, + "step": 565 + }, + { + "epoch": 0.874806800618238, + "grad_norm": 0.8968704012364771, + "learning_rate": 7.864354288941296e-05, + "loss": 0.4496, + "step": 566 + }, + { + "epoch": 0.8763523956723338, + "grad_norm": 0.8582673246978159, + "learning_rate": 7.863237379209137e-05, + "loss": 0.4329, + "step": 567 + }, + { + "epoch": 0.8778979907264297, + "grad_norm": 0.8087136075140331, + "learning_rate": 7.862115969951796e-05, + "loss": 0.4371, + "step": 568 + }, + { + "epoch": 0.8794435857805255, + "grad_norm": 0.7542353694260568, + "learning_rate": 7.860990062475382e-05, + "loss": 0.4365, + "step": 569 + }, + { + "epoch": 0.8809891808346213, + "grad_norm": 0.9014794292836983, + "learning_rate": 7.85985965809124e-05, + "loss": 0.4342, + "step": 570 + }, + { + "epoch": 0.8825347758887172, + "grad_norm": 1.070357819758905, + "learning_rate": 7.858724758115959e-05, + "loss": 0.4502, + "step": 571 + }, + { + "epoch": 0.884080370942813, + "grad_norm": 0.8931616566277254, + "learning_rate": 7.857585363871359e-05, + "loss": 0.4325, + "step": 572 + }, + { + "epoch": 0.8856259659969088, + "grad_norm": 0.7106036244064576, + "learning_rate": 7.856441476684495e-05, + "loss": 0.434, + "step": 573 + }, + { + "epoch": 0.8871715610510046, + "grad_norm": 0.6349342889735747, + "learning_rate": 7.85529309788766e-05, + "loss": 0.4458, + "step": 574 + }, + { + "epoch": 0.8887171561051005, + "grad_norm": 0.7282774160754655, + "learning_rate": 7.854140228818371e-05, + "loss": 0.4418, + "step": 575 + }, + { + "epoch": 0.8902627511591963, + "grad_norm": 0.715674599165362, + "learning_rate": 7.85298287081938e-05, + "loss": 0.442, + "step": 576 + }, + { + "epoch": 0.8918083462132921, + "grad_norm": 0.6093688140005056, + "learning_rate": 7.851821025238665e-05, + "loss": 0.4389, + "step": 577 + }, + { + "epoch": 0.893353941267388, + "grad_norm": 0.4640001330868731, + "learning_rate": 7.850654693429432e-05, + "loss": 0.4467, + "step": 578 + }, + { + "epoch": 0.8948995363214838, + "grad_norm": 0.4329185407466596, + "learning_rate": 7.84948387675011e-05, + "loss": 0.4422, + "step": 579 + }, + { + "epoch": 0.8964451313755796, + "grad_norm": 0.5522433341281833, + "learning_rate": 7.848308576564356e-05, + "loss": 0.4402, + "step": 580 + }, + { + "epoch": 0.8979907264296755, + "grad_norm": 0.5371418584269162, + "learning_rate": 7.847128794241041e-05, + "loss": 0.4408, + "step": 581 + }, + { + "epoch": 0.8995363214837713, + "grad_norm": 0.48480106420137625, + "learning_rate": 7.845944531154267e-05, + "loss": 0.4348, + "step": 582 + }, + { + "epoch": 0.9010819165378671, + "grad_norm": 0.5334433638893552, + "learning_rate": 7.844755788683345e-05, + "loss": 0.4403, + "step": 583 + }, + { + "epoch": 0.9026275115919629, + "grad_norm": 0.5014659034435851, + "learning_rate": 7.843562568212809e-05, + "loss": 0.4289, + "step": 584 + }, + { + "epoch": 0.9041731066460588, + "grad_norm": 0.5686974467379184, + "learning_rate": 7.842364871132407e-05, + "loss": 0.4374, + "step": 585 + }, + { + "epoch": 0.9057187017001546, + "grad_norm": 0.6733417899674484, + "learning_rate": 7.841162698837099e-05, + "loss": 0.4396, + "step": 586 + }, + { + "epoch": 0.9072642967542504, + "grad_norm": 0.46013610351286977, + "learning_rate": 7.839956052727061e-05, + "loss": 0.4464, + "step": 587 + }, + { + "epoch": 0.9088098918083463, + "grad_norm": 0.4602822817401326, + "learning_rate": 7.838744934207676e-05, + "loss": 0.4349, + "step": 588 + }, + { + "epoch": 0.910355486862442, + "grad_norm": 0.6095025922435364, + "learning_rate": 7.83752934468954e-05, + "loss": 0.4312, + "step": 589 + }, + { + "epoch": 0.9119010819165378, + "grad_norm": 0.5896066011664475, + "learning_rate": 7.836309285588452e-05, + "loss": 0.4355, + "step": 590 + }, + { + "epoch": 0.9134466769706336, + "grad_norm": 0.5421806312279646, + "learning_rate": 7.835084758325417e-05, + "loss": 0.43, + "step": 591 + }, + { + "epoch": 0.9149922720247295, + "grad_norm": 0.6621099500827932, + "learning_rate": 7.83385576432665e-05, + "loss": 0.4289, + "step": 592 + }, + { + "epoch": 0.9165378670788253, + "grad_norm": 0.7808387078845036, + "learning_rate": 7.832622305023564e-05, + "loss": 0.4405, + "step": 593 + }, + { + "epoch": 0.9180834621329211, + "grad_norm": 0.8519171092319919, + "learning_rate": 7.83138438185277e-05, + "loss": 0.4557, + "step": 594 + }, + { + "epoch": 0.919629057187017, + "grad_norm": 0.8521530486540375, + "learning_rate": 7.830141996256083e-05, + "loss": 0.4383, + "step": 595 + }, + { + "epoch": 0.9211746522411128, + "grad_norm": 0.736575807908396, + "learning_rate": 7.828895149680513e-05, + "loss": 0.4417, + "step": 596 + }, + { + "epoch": 0.9227202472952086, + "grad_norm": 0.5494295213432342, + "learning_rate": 7.827643843578268e-05, + "loss": 0.4302, + "step": 597 + }, + { + "epoch": 0.9242658423493045, + "grad_norm": 0.5331247439334492, + "learning_rate": 7.826388079406746e-05, + "loss": 0.4393, + "step": 598 + }, + { + "epoch": 0.9258114374034003, + "grad_norm": 0.6935679018267995, + "learning_rate": 7.825127858628541e-05, + "loss": 0.4483, + "step": 599 + }, + { + "epoch": 0.9273570324574961, + "grad_norm": 0.7595133695662849, + "learning_rate": 7.823863182711436e-05, + "loss": 0.4381, + "step": 600 + }, + { + "epoch": 0.9289026275115919, + "grad_norm": 0.6857450673605338, + "learning_rate": 7.822594053128403e-05, + "loss": 0.4389, + "step": 601 + }, + { + "epoch": 0.9304482225656878, + "grad_norm": 0.6824829999836448, + "learning_rate": 7.8213204713576e-05, + "loss": 0.4417, + "step": 602 + }, + { + "epoch": 0.9319938176197836, + "grad_norm": 0.6650583983572376, + "learning_rate": 7.820042438882374e-05, + "loss": 0.4371, + "step": 603 + }, + { + "epoch": 0.9335394126738794, + "grad_norm": 0.4745479272047121, + "learning_rate": 7.818759957191253e-05, + "loss": 0.435, + "step": 604 + }, + { + "epoch": 0.9350850077279753, + "grad_norm": 0.3498563813008576, + "learning_rate": 7.817473027777948e-05, + "loss": 0.4311, + "step": 605 + }, + { + "epoch": 0.9366306027820711, + "grad_norm": 0.445017616125712, + "learning_rate": 7.816181652141348e-05, + "loss": 0.4308, + "step": 606 + }, + { + "epoch": 0.9381761978361669, + "grad_norm": 0.4274818418189949, + "learning_rate": 7.814885831785525e-05, + "loss": 0.4303, + "step": 607 + }, + { + "epoch": 0.9397217928902627, + "grad_norm": 0.3867818825049829, + "learning_rate": 7.813585568219723e-05, + "loss": 0.4365, + "step": 608 + }, + { + "epoch": 0.9412673879443586, + "grad_norm": 0.39201184835391145, + "learning_rate": 7.812280862958364e-05, + "loss": 0.4356, + "step": 609 + }, + { + "epoch": 0.9428129829984544, + "grad_norm": 0.4730224744291155, + "learning_rate": 7.810971717521043e-05, + "loss": 0.43, + "step": 610 + }, + { + "epoch": 0.9443585780525502, + "grad_norm": 0.42281022746677577, + "learning_rate": 7.809658133432526e-05, + "loss": 0.4397, + "step": 611 + }, + { + "epoch": 0.9459041731066461, + "grad_norm": 0.4616372163135228, + "learning_rate": 7.808340112222748e-05, + "loss": 0.4404, + "step": 612 + }, + { + "epoch": 0.9474497681607419, + "grad_norm": 0.4906932188952762, + "learning_rate": 7.807017655426812e-05, + "loss": 0.4348, + "step": 613 + }, + { + "epoch": 0.9489953632148377, + "grad_norm": 0.5433982101934519, + "learning_rate": 7.80569076458499e-05, + "loss": 0.4351, + "step": 614 + }, + { + "epoch": 0.9505409582689336, + "grad_norm": 0.6181933088082244, + "learning_rate": 7.804359441242714e-05, + "loss": 0.4361, + "step": 615 + }, + { + "epoch": 0.9520865533230294, + "grad_norm": 0.7533568986921175, + "learning_rate": 7.803023686950582e-05, + "loss": 0.4398, + "step": 616 + }, + { + "epoch": 0.9536321483771252, + "grad_norm": 0.9369337286385903, + "learning_rate": 7.80168350326435e-05, + "loss": 0.4495, + "step": 617 + }, + { + "epoch": 0.955177743431221, + "grad_norm": 1.1515472785597607, + "learning_rate": 7.800338891744934e-05, + "loss": 0.4381, + "step": 618 + }, + { + "epoch": 0.9567233384853169, + "grad_norm": 0.8492454483579691, + "learning_rate": 7.798989853958412e-05, + "loss": 0.4355, + "step": 619 + }, + { + "epoch": 0.9582689335394127, + "grad_norm": 0.5008569867912441, + "learning_rate": 7.797636391476005e-05, + "loss": 0.4327, + "step": 620 + }, + { + "epoch": 0.9598145285935085, + "grad_norm": 0.42790561003786926, + "learning_rate": 7.796278505874102e-05, + "loss": 0.4377, + "step": 621 + }, + { + "epoch": 0.9613601236476044, + "grad_norm": 0.6704670742916456, + "learning_rate": 7.794916198734232e-05, + "loss": 0.4408, + "step": 622 + }, + { + "epoch": 0.9629057187017002, + "grad_norm": 0.7315174010174182, + "learning_rate": 7.793549471643083e-05, + "loss": 0.4344, + "step": 623 + }, + { + "epoch": 0.964451313755796, + "grad_norm": 0.5180116517867882, + "learning_rate": 7.792178326192483e-05, + "loss": 0.44, + "step": 624 + }, + { + "epoch": 0.9659969088098919, + "grad_norm": 0.3682930959457848, + "learning_rate": 7.790802763979411e-05, + "loss": 0.4422, + "step": 625 + }, + { + "epoch": 0.9675425038639877, + "grad_norm": 0.4827151927848899, + "learning_rate": 7.789422786605988e-05, + "loss": 0.4414, + "step": 626 + }, + { + "epoch": 0.9690880989180835, + "grad_norm": 0.525652375349752, + "learning_rate": 7.788038395679479e-05, + "loss": 0.4396, + "step": 627 + }, + { + "epoch": 0.9706336939721792, + "grad_norm": 0.47711550987418944, + "learning_rate": 7.78664959281229e-05, + "loss": 0.4417, + "step": 628 + }, + { + "epoch": 0.9721792890262752, + "grad_norm": 0.4521350816222783, + "learning_rate": 7.78525637962196e-05, + "loss": 0.4329, + "step": 629 + }, + { + "epoch": 0.973724884080371, + "grad_norm": 0.5311829664198351, + "learning_rate": 7.783858757731174e-05, + "loss": 0.4328, + "step": 630 + }, + { + "epoch": 0.9752704791344667, + "grad_norm": 0.5404316116308204, + "learning_rate": 7.782456728767744e-05, + "loss": 0.4275, + "step": 631 + }, + { + "epoch": 0.9768160741885626, + "grad_norm": 0.6410717302232665, + "learning_rate": 7.781050294364615e-05, + "loss": 0.435, + "step": 632 + }, + { + "epoch": 0.9783616692426584, + "grad_norm": 0.7640852772159162, + "learning_rate": 7.779639456159872e-05, + "loss": 0.4311, + "step": 633 + }, + { + "epoch": 0.9799072642967542, + "grad_norm": 0.8147701250526043, + "learning_rate": 7.778224215796719e-05, + "loss": 0.4348, + "step": 634 + }, + { + "epoch": 0.98145285935085, + "grad_norm": 0.8184848854239374, + "learning_rate": 7.77680457492349e-05, + "loss": 0.4402, + "step": 635 + }, + { + "epoch": 0.9829984544049459, + "grad_norm": 0.8728094725370318, + "learning_rate": 7.775380535193649e-05, + "loss": 0.4289, + "step": 636 + }, + { + "epoch": 0.9845440494590417, + "grad_norm": 0.8091290006835459, + "learning_rate": 7.773952098265778e-05, + "loss": 0.4368, + "step": 637 + }, + { + "epoch": 0.9860896445131375, + "grad_norm": 0.6942530585191017, + "learning_rate": 7.772519265803581e-05, + "loss": 0.4372, + "step": 638 + }, + { + "epoch": 0.9876352395672334, + "grad_norm": 0.5165228454411432, + "learning_rate": 7.771082039475884e-05, + "loss": 0.4381, + "step": 639 + }, + { + "epoch": 0.9891808346213292, + "grad_norm": 0.37777307745002114, + "learning_rate": 7.769640420956627e-05, + "loss": 0.4326, + "step": 640 + }, + { + "epoch": 0.990726429675425, + "grad_norm": 0.36547077198781047, + "learning_rate": 7.768194411924873e-05, + "loss": 0.4294, + "step": 641 + }, + { + "epoch": 0.9922720247295209, + "grad_norm": 0.4503781104554527, + "learning_rate": 7.766744014064786e-05, + "loss": 0.4378, + "step": 642 + }, + { + "epoch": 0.9938176197836167, + "grad_norm": 0.5481159449044525, + "learning_rate": 7.765289229065657e-05, + "loss": 0.4375, + "step": 643 + }, + { + "epoch": 0.9953632148377125, + "grad_norm": 0.6296919301305356, + "learning_rate": 7.763830058621873e-05, + "loss": 0.4462, + "step": 644 + }, + { + "epoch": 0.9969088098918083, + "grad_norm": 0.6488083143648484, + "learning_rate": 7.762366504432937e-05, + "loss": 0.4436, + "step": 645 + }, + { + "epoch": 0.9984544049459042, + "grad_norm": 0.6477864119314514, + "learning_rate": 7.760898568203453e-05, + "loss": 0.4355, + "step": 646 + }, + { + "epoch": 1.0, + "grad_norm": 0.7056159823492316, + "learning_rate": 7.759426251643135e-05, + "loss": 0.4445, + "step": 647 + }, + { + "epoch": 1.001545595054096, + "grad_norm": 0.729906652382838, + "learning_rate": 7.757949556466794e-05, + "loss": 0.4175, + "step": 648 + }, + { + "epoch": 1.0030911901081916, + "grad_norm": 0.744332000011287, + "learning_rate": 7.756468484394339e-05, + "loss": 0.4175, + "step": 649 + }, + { + "epoch": 1.0046367851622875, + "grad_norm": 0.9054700744643306, + "learning_rate": 7.75498303715078e-05, + "loss": 0.4243, + "step": 650 + }, + { + "epoch": 1.0061823802163834, + "grad_norm": 0.8567705197263996, + "learning_rate": 7.753493216466225e-05, + "loss": 0.4251, + "step": 651 + }, + { + "epoch": 1.007727975270479, + "grad_norm": 0.8080573567421983, + "learning_rate": 7.751999024075871e-05, + "loss": 0.4259, + "step": 652 + }, + { + "epoch": 1.009273570324575, + "grad_norm": 0.8497810423019736, + "learning_rate": 7.750500461720008e-05, + "loss": 0.4295, + "step": 653 + }, + { + "epoch": 1.010819165378671, + "grad_norm": 0.963231768182949, + "learning_rate": 7.748997531144019e-05, + "loss": 0.4202, + "step": 654 + }, + { + "epoch": 1.0123647604327666, + "grad_norm": 0.9131327521435064, + "learning_rate": 7.747490234098371e-05, + "loss": 0.4234, + "step": 655 + }, + { + "epoch": 1.0139103554868625, + "grad_norm": 3.233993309506214, + "learning_rate": 7.745978572338615e-05, + "loss": 0.4208, + "step": 656 + }, + { + "epoch": 1.0154559505409582, + "grad_norm": 1.1478534927734931, + "learning_rate": 7.744462547625393e-05, + "loss": 0.4313, + "step": 657 + }, + { + "epoch": 1.017001545595054, + "grad_norm": 1.2500819953566402, + "learning_rate": 7.742942161724422e-05, + "loss": 0.4317, + "step": 658 + }, + { + "epoch": 1.01854714064915, + "grad_norm": 0.7276767086353634, + "learning_rate": 7.7414174164065e-05, + "loss": 0.4177, + "step": 659 + }, + { + "epoch": 1.0200927357032457, + "grad_norm": 1.169391244438201, + "learning_rate": 7.739888313447502e-05, + "loss": 0.4258, + "step": 660 + }, + { + "epoch": 1.0216383307573416, + "grad_norm": 0.5246199930754715, + "learning_rate": 7.738354854628381e-05, + "loss": 0.4214, + "step": 661 + }, + { + "epoch": 1.0231839258114375, + "grad_norm": 0.8263151377221576, + "learning_rate": 7.73681704173516e-05, + "loss": 0.4205, + "step": 662 + }, + { + "epoch": 1.0247295208655331, + "grad_norm": 0.6551303044770097, + "learning_rate": 7.735274876558937e-05, + "loss": 0.4266, + "step": 663 + }, + { + "epoch": 1.026275115919629, + "grad_norm": 0.5149694864670407, + "learning_rate": 7.733728360895875e-05, + "loss": 0.4174, + "step": 664 + }, + { + "epoch": 1.027820710973725, + "grad_norm": 0.5586430541966018, + "learning_rate": 7.732177496547204e-05, + "loss": 0.4286, + "step": 665 + }, + { + "epoch": 1.0293663060278206, + "grad_norm": 0.5739556085683527, + "learning_rate": 7.730622285319225e-05, + "loss": 0.4207, + "step": 666 + }, + { + "epoch": 1.0309119010819165, + "grad_norm": 0.5090479419066287, + "learning_rate": 7.729062729023294e-05, + "loss": 0.4195, + "step": 667 + }, + { + "epoch": 1.0324574961360125, + "grad_norm": 0.44214443641614104, + "learning_rate": 7.727498829475835e-05, + "loss": 0.4226, + "step": 668 + }, + { + "epoch": 1.0340030911901081, + "grad_norm": 0.4301128111577211, + "learning_rate": 7.725930588498324e-05, + "loss": 0.424, + "step": 669 + }, + { + "epoch": 1.035548686244204, + "grad_norm": 0.4399222127023858, + "learning_rate": 7.724358007917296e-05, + "loss": 0.4185, + "step": 670 + }, + { + "epoch": 1.0370942812983, + "grad_norm": 0.48272414168544736, + "learning_rate": 7.722781089564342e-05, + "loss": 0.4266, + "step": 671 + }, + { + "epoch": 1.0386398763523956, + "grad_norm": 0.44917139741794737, + "learning_rate": 7.721199835276105e-05, + "loss": 0.4212, + "step": 672 + }, + { + "epoch": 1.0401854714064915, + "grad_norm": 0.5647705885835368, + "learning_rate": 7.719614246894274e-05, + "loss": 0.4228, + "step": 673 + }, + { + "epoch": 1.0417310664605872, + "grad_norm": 0.5300385509440503, + "learning_rate": 7.718024326265591e-05, + "loss": 0.424, + "step": 674 + }, + { + "epoch": 1.0432766615146831, + "grad_norm": 0.5033099549666273, + "learning_rate": 7.716430075241842e-05, + "loss": 0.4135, + "step": 675 + }, + { + "epoch": 1.044822256568779, + "grad_norm": 0.5281034352711879, + "learning_rate": 7.714831495679857e-05, + "loss": 0.4239, + "step": 676 + }, + { + "epoch": 1.0463678516228747, + "grad_norm": 0.4854866818040774, + "learning_rate": 7.713228589441505e-05, + "loss": 0.4269, + "step": 677 + }, + { + "epoch": 1.0479134466769706, + "grad_norm": 0.4753894723383696, + "learning_rate": 7.711621358393697e-05, + "loss": 0.4185, + "step": 678 + }, + { + "epoch": 1.0494590417310665, + "grad_norm": 0.4597262766834238, + "learning_rate": 7.71000980440838e-05, + "loss": 0.4218, + "step": 679 + }, + { + "epoch": 1.0510046367851622, + "grad_norm": 0.4835682943503437, + "learning_rate": 7.708393929362535e-05, + "loss": 0.423, + "step": 680 + }, + { + "epoch": 1.052550231839258, + "grad_norm": 0.42980910216068385, + "learning_rate": 7.70677373513818e-05, + "loss": 0.4077, + "step": 681 + }, + { + "epoch": 1.054095826893354, + "grad_norm": 0.4474633900583748, + "learning_rate": 7.705149223622357e-05, + "loss": 0.412, + "step": 682 + }, + { + "epoch": 1.0556414219474497, + "grad_norm": 2.4069816314810333, + "learning_rate": 7.703520396707142e-05, + "loss": 0.427, + "step": 683 + }, + { + "epoch": 1.0571870170015456, + "grad_norm": 93.04219750838124, + "learning_rate": 7.701887256289635e-05, + "loss": 0.7967, + "step": 684 + }, + { + "epoch": 1.0587326120556415, + "grad_norm": 2.1906687985628297, + "learning_rate": 7.70024980427196e-05, + "loss": 0.4442, + "step": 685 + }, + { + "epoch": 1.0602782071097372, + "grad_norm": 272.835975471966, + "learning_rate": 7.698608042561262e-05, + "loss": 5.4936, + "step": 686 + }, + { + "epoch": 1.061823802163833, + "grad_norm": 4.3404427976272535, + "learning_rate": 7.696961973069707e-05, + "loss": 0.579, + "step": 687 + }, + { + "epoch": 1.063369397217929, + "grad_norm": 10.760573360273582, + "learning_rate": 7.695311597714476e-05, + "loss": 0.7585, + "step": 688 + }, + { + "epoch": 1.0649149922720247, + "grad_norm": 188.1831077117581, + "learning_rate": 7.693656918417768e-05, + "loss": 1.8238, + "step": 689 + }, + { + "epoch": 1.0664605873261206, + "grad_norm": 197.49638552527358, + "learning_rate": 7.691997937106794e-05, + "loss": 2.2759, + "step": 690 + }, + { + "epoch": 1.0680061823802163, + "grad_norm": 206.56798798122128, + "learning_rate": 7.690334655713772e-05, + "loss": 2.0006, + "step": 691 + }, + { + "epoch": 1.0695517774343122, + "grad_norm": 100.6833875702295, + "learning_rate": 7.688667076175934e-05, + "loss": 8.2281, + "step": 692 + }, + { + "epoch": 1.071097372488408, + "grad_norm": 105.295675147931, + "learning_rate": 7.686995200435513e-05, + "loss": 4.6001, + "step": 693 + }, + { + "epoch": 1.0726429675425038, + "grad_norm": 14.645438032599293, + "learning_rate": 7.685319030439753e-05, + "loss": 1.1132, + "step": 694 + }, + { + "epoch": 1.0741885625965997, + "grad_norm": 23.371900123313832, + "learning_rate": 7.683638568140888e-05, + "loss": 0.884, + "step": 695 + }, + { + "epoch": 1.0757341576506956, + "grad_norm": 5.890050408513227, + "learning_rate": 7.681953815496163e-05, + "loss": 0.6911, + "step": 696 + }, + { + "epoch": 1.0772797527047913, + "grad_norm": 3.959864486181502, + "learning_rate": 7.680264774467813e-05, + "loss": 0.6442, + "step": 697 + }, + { + "epoch": 1.0788253477588872, + "grad_norm": 13.66560312210619, + "learning_rate": 7.678571447023069e-05, + "loss": 0.7786, + "step": 698 + }, + { + "epoch": 1.080370942812983, + "grad_norm": 3.3446009588671073, + "learning_rate": 7.676873835134157e-05, + "loss": 0.6308, + "step": 699 + }, + { + "epoch": 1.0819165378670788, + "grad_norm": 3.1766326038920907, + "learning_rate": 7.675171940778288e-05, + "loss": 0.6062, + "step": 700 + }, + { + "epoch": 1.0834621329211747, + "grad_norm": 2.8537048610899776, + "learning_rate": 7.673465765937667e-05, + "loss": 0.5381, + "step": 701 + }, + { + "epoch": 1.0850077279752706, + "grad_norm": 15.15550131506517, + "learning_rate": 7.67175531259948e-05, + "loss": 0.7292, + "step": 702 + }, + { + "epoch": 1.0865533230293662, + "grad_norm": 97.51530829433813, + "learning_rate": 7.670040582755897e-05, + "loss": 9.4639, + "step": 703 + }, + { + "epoch": 1.0880989180834622, + "grad_norm": 54.95438868978781, + "learning_rate": 7.668321578404071e-05, + "loss": 9.5178, + "step": 704 + }, + { + "epoch": 1.089644513137558, + "grad_norm": 420.2876754740128, + "learning_rate": 7.666598301546132e-05, + "loss": 28.2843, + "step": 705 + }, + { + "epoch": 1.0911901081916537, + "grad_norm": 52.35848535158038, + "learning_rate": 7.664870754189183e-05, + "loss": 9.2741, + "step": 706 + }, + { + "epoch": 1.0927357032457496, + "grad_norm": 97.93519552001106, + "learning_rate": 7.663138938345309e-05, + "loss": 13.0573, + "step": 707 + }, + { + "epoch": 1.0942812982998453, + "grad_norm": 39.499276377622536, + "learning_rate": 7.661402856031557e-05, + "loss": 16.4409, + "step": 708 + }, + { + "epoch": 1.0958268933539412, + "grad_norm": 70.20801371263504, + "learning_rate": 7.659662509269948e-05, + "loss": 9.4351, + "step": 709 + }, + { + "epoch": 1.0973724884080371, + "grad_norm": 259.3118119719558, + "learning_rate": 7.657917900087475e-05, + "loss": 22.6132, + "step": 710 + }, + { + "epoch": 1.0989180834621328, + "grad_norm": 63.40624302863192, + "learning_rate": 7.656169030516085e-05, + "loss": 10.2828, + "step": 711 + }, + { + "epoch": 1.1004636785162287, + "grad_norm": 24.90925284691798, + "learning_rate": 7.654415902592693e-05, + "loss": 8.0415, + "step": 712 + }, + { + "epoch": 1.1020092735703246, + "grad_norm": 151.42831740527757, + "learning_rate": 7.652658518359172e-05, + "loss": 8.9223, + "step": 713 + }, + { + "epoch": 1.1035548686244203, + "grad_norm": 347.70782039381953, + "learning_rate": 7.650896879862355e-05, + "loss": 8.2149, + "step": 714 + }, + { + "epoch": 1.1051004636785162, + "grad_norm": 48.74238085739299, + "learning_rate": 7.649130989154028e-05, + "loss": 8.6492, + "step": 715 + }, + { + "epoch": 1.1066460587326121, + "grad_norm": 28.495354706861466, + "learning_rate": 7.64736084829093e-05, + "loss": 7.6341, + "step": 716 + }, + { + "epoch": 1.1081916537867078, + "grad_norm": 9.757410329201203, + "learning_rate": 7.645586459334746e-05, + "loss": 6.7279, + "step": 717 + }, + { + "epoch": 1.1097372488408037, + "grad_norm": 13.581780015633562, + "learning_rate": 7.643807824352119e-05, + "loss": 6.7868, + "step": 718 + }, + { + "epoch": 1.1112828438948996, + "grad_norm": 33.35818290868364, + "learning_rate": 7.642024945414625e-05, + "loss": 7.0932, + "step": 719 + }, + { + "epoch": 1.1128284389489953, + "grad_norm": 35.833130776074306, + "learning_rate": 7.640237824598792e-05, + "loss": 6.8967, + "step": 720 + }, + { + "epoch": 1.1143740340030912, + "grad_norm": 17.593556359959898, + "learning_rate": 7.638446463986085e-05, + "loss": 6.7292, + "step": 721 + }, + { + "epoch": 1.1159196290571871, + "grad_norm": 11.628269896069664, + "learning_rate": 7.636650865662907e-05, + "loss": 6.5477, + "step": 722 + }, + { + "epoch": 1.1174652241112828, + "grad_norm": 25.27507341382732, + "learning_rate": 7.634851031720598e-05, + "loss": 6.5934, + "step": 723 + }, + { + "epoch": 1.1190108191653787, + "grad_norm": 22.411429742342815, + "learning_rate": 7.633046964255431e-05, + "loss": 6.614, + "step": 724 + }, + { + "epoch": 1.1205564142194744, + "grad_norm": 9.461470565160505, + "learning_rate": 7.631238665368606e-05, + "loss": 6.3066, + "step": 725 + }, + { + "epoch": 1.1221020092735703, + "grad_norm": 9.156472187617643, + "learning_rate": 7.629426137166259e-05, + "loss": 6.2147, + "step": 726 + }, + { + "epoch": 1.1236476043276662, + "grad_norm": 8.476840375076899, + "learning_rate": 7.627609381759445e-05, + "loss": 6.1244, + "step": 727 + }, + { + "epoch": 1.125193199381762, + "grad_norm": 4.93811170279106, + "learning_rate": 7.625788401264146e-05, + "loss": 5.9622, + "step": 728 + }, + { + "epoch": 1.1267387944358578, + "grad_norm": 8.735095953309674, + "learning_rate": 7.623963197801263e-05, + "loss": 6.022, + "step": 729 + }, + { + "epoch": 1.1282843894899537, + "grad_norm": 6.554505113131336, + "learning_rate": 7.622133773496616e-05, + "loss": 5.9151, + "step": 730 + }, + { + "epoch": 1.1298299845440494, + "grad_norm": 6.807111122961733, + "learning_rate": 7.620300130480944e-05, + "loss": 5.8511, + "step": 731 + }, + { + "epoch": 1.1313755795981453, + "grad_norm": 4.556049123778991, + "learning_rate": 7.618462270889893e-05, + "loss": 5.7735, + "step": 732 + }, + { + "epoch": 1.1329211746522412, + "grad_norm": 8.287766607976652, + "learning_rate": 7.616620196864027e-05, + "loss": 5.7623, + "step": 733 + }, + { + "epoch": 1.1344667697063369, + "grad_norm": 7.100423364114964, + "learning_rate": 7.614773910548816e-05, + "loss": 5.6987, + "step": 734 + }, + { + "epoch": 1.1360123647604328, + "grad_norm": 4.627903031062972, + "learning_rate": 7.612923414094634e-05, + "loss": 5.6177, + "step": 735 + }, + { + "epoch": 1.1375579598145287, + "grad_norm": 4.881674498195489, + "learning_rate": 7.61106870965676e-05, + "loss": 5.5916, + "step": 736 + }, + { + "epoch": 1.1391035548686244, + "grad_norm": 3.6298034646753483, + "learning_rate": 7.609209799395377e-05, + "loss": 5.5201, + "step": 737 + }, + { + "epoch": 1.1406491499227203, + "grad_norm": 5.3843072583896845, + "learning_rate": 7.607346685475561e-05, + "loss": 5.4628, + "step": 738 + }, + { + "epoch": 1.1421947449768162, + "grad_norm": 3.6685012575901506, + "learning_rate": 7.605479370067288e-05, + "loss": 5.4009, + "step": 739 + }, + { + "epoch": 1.1437403400309119, + "grad_norm": 4.115992637127712, + "learning_rate": 7.603607855345427e-05, + "loss": 5.3897, + "step": 740 + }, + { + "epoch": 1.1452859350850078, + "grad_norm": 4.042453089971003, + "learning_rate": 7.601732143489735e-05, + "loss": 5.325, + "step": 741 + }, + { + "epoch": 1.1468315301391034, + "grad_norm": 3.16771761274949, + "learning_rate": 7.599852236684863e-05, + "loss": 5.2761, + "step": 742 + }, + { + "epoch": 1.1483771251931993, + "grad_norm": 5.038776248782134, + "learning_rate": 7.597968137120342e-05, + "loss": 5.2472, + "step": 743 + }, + { + "epoch": 1.1499227202472952, + "grad_norm": 3.0285414515015887, + "learning_rate": 7.59607984699059e-05, + "loss": 5.1769, + "step": 744 + }, + { + "epoch": 1.1514683153013912, + "grad_norm": 8.593385916790467, + "learning_rate": 7.594187368494904e-05, + "loss": 5.2736, + "step": 745 + }, + { + "epoch": 1.1530139103554868, + "grad_norm": 4.918520985824609, + "learning_rate": 7.592290703837462e-05, + "loss": 5.1426, + "step": 746 + }, + { + "epoch": 1.1545595054095827, + "grad_norm": 6.804430302655261, + "learning_rate": 7.590389855227312e-05, + "loss": 5.1824, + "step": 747 + }, + { + "epoch": 1.1561051004636784, + "grad_norm": 4.876833368982366, + "learning_rate": 7.58848482487838e-05, + "loss": 5.1545, + "step": 748 + }, + { + "epoch": 1.1576506955177743, + "grad_norm": 6.15456205410321, + "learning_rate": 7.586575615009464e-05, + "loss": 5.0909, + "step": 749 + }, + { + "epoch": 1.1591962905718702, + "grad_norm": 6.17166000852694, + "learning_rate": 7.584662227844223e-05, + "loss": 5.0788, + "step": 750 + }, + { + "epoch": 1.160741885625966, + "grad_norm": 3.8399589257173505, + "learning_rate": 7.582744665611187e-05, + "loss": 5.0191, + "step": 751 + }, + { + "epoch": 1.1622874806800618, + "grad_norm": 6.150141184578046, + "learning_rate": 7.580822930543747e-05, + "loss": 5.1103, + "step": 752 + }, + { + "epoch": 1.1638330757341577, + "grad_norm": 3.6360701998799088, + "learning_rate": 7.578897024880154e-05, + "loss": 4.9713, + "step": 753 + }, + { + "epoch": 1.1653786707882534, + "grad_norm": 4.362230382599794, + "learning_rate": 7.576966950863515e-05, + "loss": 4.9301, + "step": 754 + }, + { + "epoch": 1.1669242658423493, + "grad_norm": 4.160900154324657, + "learning_rate": 7.575032710741796e-05, + "loss": 4.935, + "step": 755 + }, + { + "epoch": 1.1684698608964452, + "grad_norm": 4.542412791649708, + "learning_rate": 7.573094306767812e-05, + "loss": 4.9442, + "step": 756 + }, + { + "epoch": 1.170015455950541, + "grad_norm": 4.245248645752507, + "learning_rate": 7.571151741199227e-05, + "loss": 4.9345, + "step": 757 + }, + { + "epoch": 1.1715610510046368, + "grad_norm": 2.9414058287299834, + "learning_rate": 7.569205016298556e-05, + "loss": 4.8409, + "step": 758 + }, + { + "epoch": 1.1731066460587325, + "grad_norm": 4.156394586232867, + "learning_rate": 7.567254134333155e-05, + "loss": 4.8069, + "step": 759 + }, + { + "epoch": 1.1746522411128284, + "grad_norm": 5.035582526248614, + "learning_rate": 7.565299097575221e-05, + "loss": 4.8857, + "step": 760 + }, + { + "epoch": 1.1761978361669243, + "grad_norm": 5.045628338913676, + "learning_rate": 7.563339908301792e-05, + "loss": 4.8286, + "step": 761 + }, + { + "epoch": 1.1777434312210202, + "grad_norm": 3.988007515418781, + "learning_rate": 7.561376568794745e-05, + "loss": 4.7804, + "step": 762 + }, + { + "epoch": 1.179289026275116, + "grad_norm": 4.186207353985047, + "learning_rate": 7.559409081340784e-05, + "loss": 4.7789, + "step": 763 + }, + { + "epoch": 1.1808346213292118, + "grad_norm": 3.9396849054608793, + "learning_rate": 7.557437448231451e-05, + "loss": 4.6864, + "step": 764 + }, + { + "epoch": 1.1823802163833075, + "grad_norm": 4.537203137435761, + "learning_rate": 7.555461671763112e-05, + "loss": 4.7164, + "step": 765 + }, + { + "epoch": 1.1839258114374034, + "grad_norm": 4.65880138729529, + "learning_rate": 7.55348175423696e-05, + "loss": 4.6988, + "step": 766 + }, + { + "epoch": 1.1854714064914993, + "grad_norm": 3.4087996084696996, + "learning_rate": 7.551497697959013e-05, + "loss": 4.6225, + "step": 767 + }, + { + "epoch": 1.187017001545595, + "grad_norm": 3.5488113498187057, + "learning_rate": 7.549509505240107e-05, + "loss": 4.6265, + "step": 768 + }, + { + "epoch": 1.1885625965996909, + "grad_norm": 5.495822131358979, + "learning_rate": 7.547517178395896e-05, + "loss": 4.6234, + "step": 769 + }, + { + "epoch": 1.1901081916537868, + "grad_norm": 4.86362818242385, + "learning_rate": 7.545520719746851e-05, + "loss": 4.6488, + "step": 770 + }, + { + "epoch": 1.1916537867078825, + "grad_norm": 2.600222845522659, + "learning_rate": 7.543520131618251e-05, + "loss": 4.5529, + "step": 771 + }, + { + "epoch": 1.1931993817619784, + "grad_norm": 5.555047812446914, + "learning_rate": 7.541515416340192e-05, + "loss": 4.615, + "step": 772 + }, + { + "epoch": 1.1947449768160743, + "grad_norm": 4.40056853472803, + "learning_rate": 7.539506576247568e-05, + "loss": 4.5056, + "step": 773 + }, + { + "epoch": 1.19629057187017, + "grad_norm": 4.982724730651727, + "learning_rate": 7.537493613680085e-05, + "loss": 4.4974, + "step": 774 + }, + { + "epoch": 1.1978361669242659, + "grad_norm": 3.321912005941974, + "learning_rate": 7.535476530982244e-05, + "loss": 4.5238, + "step": 775 + }, + { + "epoch": 1.1993817619783615, + "grad_norm": 4.643808857450901, + "learning_rate": 7.533455330503351e-05, + "loss": 4.5084, + "step": 776 + }, + { + "epoch": 1.2009273570324575, + "grad_norm": 3.0699451748744533, + "learning_rate": 7.531430014597503e-05, + "loss": 4.4445, + "step": 777 + }, + { + "epoch": 1.2024729520865534, + "grad_norm": 4.55358876465052, + "learning_rate": 7.529400585623592e-05, + "loss": 4.4033, + "step": 778 + }, + { + "epoch": 1.2040185471406493, + "grad_norm": 4.112369297404642, + "learning_rate": 7.527367045945301e-05, + "loss": 4.4221, + "step": 779 + }, + { + "epoch": 1.205564142194745, + "grad_norm": 4.311022526229789, + "learning_rate": 7.5253293979311e-05, + "loss": 4.3886, + "step": 780 + }, + { + "epoch": 1.2071097372488409, + "grad_norm": 3.456145921710313, + "learning_rate": 7.523287643954243e-05, + "loss": 4.3614, + "step": 781 + }, + { + "epoch": 1.2086553323029365, + "grad_norm": 3.576154707012333, + "learning_rate": 7.52124178639277e-05, + "loss": 4.377, + "step": 782 + }, + { + "epoch": 1.2102009273570324, + "grad_norm": 4.97025430932979, + "learning_rate": 7.519191827629497e-05, + "loss": 4.3643, + "step": 783 + }, + { + "epoch": 1.2117465224111283, + "grad_norm": 3.9979129091531562, + "learning_rate": 7.517137770052017e-05, + "loss": 4.3257, + "step": 784 + }, + { + "epoch": 1.213292117465224, + "grad_norm": 3.178227194464303, + "learning_rate": 7.515079616052699e-05, + "loss": 4.2892, + "step": 785 + }, + { + "epoch": 1.21483771251932, + "grad_norm": 4.604727245265333, + "learning_rate": 7.513017368028681e-05, + "loss": 4.2603, + "step": 786 + }, + { + "epoch": 1.2163833075734158, + "grad_norm": 3.6145987998432996, + "learning_rate": 7.510951028381869e-05, + "loss": 4.247, + "step": 787 + }, + { + "epoch": 1.2179289026275115, + "grad_norm": 3.5981722391474147, + "learning_rate": 7.508880599518936e-05, + "loss": 4.2436, + "step": 788 + }, + { + "epoch": 1.2194744976816074, + "grad_norm": 4.986630877971159, + "learning_rate": 7.506806083851319e-05, + "loss": 4.2043, + "step": 789 + }, + { + "epoch": 1.2210200927357033, + "grad_norm": 3.7121353858408104, + "learning_rate": 7.504727483795211e-05, + "loss": 4.1745, + "step": 790 + }, + { + "epoch": 1.222565687789799, + "grad_norm": 3.649859347149282, + "learning_rate": 7.502644801771564e-05, + "loss": 4.1762, + "step": 791 + }, + { + "epoch": 1.224111282843895, + "grad_norm": 4.924771754371227, + "learning_rate": 7.500558040206086e-05, + "loss": 4.1296, + "step": 792 + }, + { + "epoch": 1.2256568778979906, + "grad_norm": 3.846939395214013, + "learning_rate": 7.498467201529236e-05, + "loss": 4.0909, + "step": 793 + }, + { + "epoch": 1.2272024729520865, + "grad_norm": 4.1106153126958045, + "learning_rate": 7.496372288176219e-05, + "loss": 4.1137, + "step": 794 + }, + { + "epoch": 1.2287480680061824, + "grad_norm": 4.808135397086257, + "learning_rate": 7.494273302586986e-05, + "loss": 4.1335, + "step": 795 + }, + { + "epoch": 1.2302936630602783, + "grad_norm": 3.6829894131694014, + "learning_rate": 7.492170247206234e-05, + "loss": 4.0574, + "step": 796 + }, + { + "epoch": 1.231839258114374, + "grad_norm": 3.519370566178741, + "learning_rate": 7.490063124483401e-05, + "loss": 4.0691, + "step": 797 + }, + { + "epoch": 1.23338485316847, + "grad_norm": 3.646513485706842, + "learning_rate": 7.487951936872655e-05, + "loss": 4.0395, + "step": 798 + }, + { + "epoch": 1.2349304482225656, + "grad_norm": 4.283259741858885, + "learning_rate": 7.485836686832904e-05, + "loss": 4.0253, + "step": 799 + }, + { + "epoch": 1.2364760432766615, + "grad_norm": 2.691183180900685, + "learning_rate": 7.483717376827791e-05, + "loss": 3.9774, + "step": 800 + }, + { + "epoch": 1.2380216383307574, + "grad_norm": 3.506101475003609, + "learning_rate": 7.481594009325679e-05, + "loss": 4.0073, + "step": 801 + }, + { + "epoch": 1.239567233384853, + "grad_norm": 3.8175643338130296, + "learning_rate": 7.479466586799662e-05, + "loss": 3.9094, + "step": 802 + }, + { + "epoch": 1.241112828438949, + "grad_norm": 4.354755409166524, + "learning_rate": 7.477335111727556e-05, + "loss": 3.9642, + "step": 803 + }, + { + "epoch": 1.242658423493045, + "grad_norm": 3.2877563917762225, + "learning_rate": 7.475199586591897e-05, + "loss": 3.9092, + "step": 804 + }, + { + "epoch": 1.2442040185471406, + "grad_norm": 4.439813730748986, + "learning_rate": 7.473060013879938e-05, + "loss": 3.9291, + "step": 805 + }, + { + "epoch": 1.2457496136012365, + "grad_norm": 4.343013410523542, + "learning_rate": 7.470916396083649e-05, + "loss": 3.8698, + "step": 806 + }, + { + "epoch": 1.2472952086553324, + "grad_norm": 3.7371468862956023, + "learning_rate": 7.468768735699703e-05, + "loss": 3.8414, + "step": 807 + }, + { + "epoch": 1.248840803709428, + "grad_norm": 3.853589896317306, + "learning_rate": 7.466617035229493e-05, + "loss": 3.8265, + "step": 808 + }, + { + "epoch": 1.250386398763524, + "grad_norm": 4.032592828377498, + "learning_rate": 7.464461297179107e-05, + "loss": 3.7549, + "step": 809 + }, + { + "epoch": 1.2519319938176197, + "grad_norm": 3.9821131119978936, + "learning_rate": 7.462301524059342e-05, + "loss": 3.7342, + "step": 810 + }, + { + "epoch": 1.2534775888717156, + "grad_norm": 4.732422166817073, + "learning_rate": 7.460137718385694e-05, + "loss": 3.7275, + "step": 811 + }, + { + "epoch": 1.2550231839258115, + "grad_norm": 4.5088682750508475, + "learning_rate": 7.457969882678353e-05, + "loss": 3.6924, + "step": 812 + }, + { + "epoch": 1.2565687789799074, + "grad_norm": 5.449522575354933, + "learning_rate": 7.455798019462204e-05, + "loss": 3.7383, + "step": 813 + }, + { + "epoch": 1.258114374034003, + "grad_norm": 4.76650173402543, + "learning_rate": 7.453622131266824e-05, + "loss": 3.7432, + "step": 814 + }, + { + "epoch": 1.259659969088099, + "grad_norm": 4.067447342968902, + "learning_rate": 7.451442220626477e-05, + "loss": 3.6055, + "step": 815 + }, + { + "epoch": 1.2612055641421946, + "grad_norm": 6.253282776401617, + "learning_rate": 7.449258290080112e-05, + "loss": 3.6417, + "step": 816 + }, + { + "epoch": 1.2627511591962906, + "grad_norm": 3.2475778126347223, + "learning_rate": 7.44707034217136e-05, + "loss": 3.5223, + "step": 817 + }, + { + "epoch": 1.2642967542503865, + "grad_norm": 6.215818869272131, + "learning_rate": 7.44487837944853e-05, + "loss": 3.5363, + "step": 818 + }, + { + "epoch": 1.2658423493044824, + "grad_norm": 4.126921705951098, + "learning_rate": 7.44268240446461e-05, + "loss": 3.412, + "step": 819 + }, + { + "epoch": 1.267387944358578, + "grad_norm": 3.612290620794877, + "learning_rate": 7.440482419777259e-05, + "loss": 3.4251, + "step": 820 + }, + { + "epoch": 1.268933539412674, + "grad_norm": 6.2150029543395044, + "learning_rate": 7.438278427948805e-05, + "loss": 3.338, + "step": 821 + }, + { + "epoch": 1.2704791344667696, + "grad_norm": 3.483848021370735, + "learning_rate": 7.436070431546245e-05, + "loss": 3.2848, + "step": 822 + }, + { + "epoch": 1.2720247295208655, + "grad_norm": 5.015251343254233, + "learning_rate": 7.433858433141242e-05, + "loss": 3.2626, + "step": 823 + }, + { + "epoch": 1.2735703245749614, + "grad_norm": 4.373403459885209, + "learning_rate": 7.431642435310113e-05, + "loss": 3.1986, + "step": 824 + }, + { + "epoch": 1.2751159196290571, + "grad_norm": 6.518864003736122, + "learning_rate": 7.429422440633842e-05, + "loss": 3.1361, + "step": 825 + }, + { + "epoch": 1.276661514683153, + "grad_norm": 6.3909451423005414, + "learning_rate": 7.427198451698062e-05, + "loss": 3.0242, + "step": 826 + }, + { + "epoch": 1.2782071097372487, + "grad_norm": 6.9704608126309, + "learning_rate": 7.42497047109306e-05, + "loss": 2.9986, + "step": 827 + }, + { + "epoch": 1.2797527047913446, + "grad_norm": 4.655208475479434, + "learning_rate": 7.422738501413774e-05, + "loss": 2.8159, + "step": 828 + }, + { + "epoch": 1.2812982998454405, + "grad_norm": 15.894144967213778, + "learning_rate": 7.420502545259785e-05, + "loss": 3.3776, + "step": 829 + }, + { + "epoch": 1.2828438948995364, + "grad_norm": 7.253150436432715, + "learning_rate": 7.418262605235319e-05, + "loss": 2.9451, + "step": 830 + }, + { + "epoch": 1.2843894899536321, + "grad_norm": 6.263577298039799, + "learning_rate": 7.41601868394924e-05, + "loss": 2.6416, + "step": 831 + }, + { + "epoch": 1.285935085007728, + "grad_norm": 8.962995030592927, + "learning_rate": 7.413770784015053e-05, + "loss": 2.4536, + "step": 832 + }, + { + "epoch": 1.2874806800618237, + "grad_norm": 16.309713333237088, + "learning_rate": 7.411518908050893e-05, + "loss": 2.0685, + "step": 833 + }, + { + "epoch": 1.2890262751159196, + "grad_norm": 25.303288554745563, + "learning_rate": 7.409263058679527e-05, + "loss": 2.7383, + "step": 834 + }, + { + "epoch": 1.2905718701700155, + "grad_norm": 120.89512942503012, + "learning_rate": 7.407003238528352e-05, + "loss": 2.4238, + "step": 835 + }, + { + "epoch": 1.2921174652241114, + "grad_norm": 180.8543756230507, + "learning_rate": 7.404739450229385e-05, + "loss": 3.3396, + "step": 836 + }, + { + "epoch": 1.293663060278207, + "grad_norm": 108.89896521164664, + "learning_rate": 7.402471696419268e-05, + "loss": 3.2652, + "step": 837 + }, + { + "epoch": 1.295208655332303, + "grad_norm": 15.288203222772317, + "learning_rate": 7.400199979739263e-05, + "loss": 2.9395, + "step": 838 + }, + { + "epoch": 1.2967542503863987, + "grad_norm": 8.53625059242779, + "learning_rate": 7.397924302835247e-05, + "loss": 2.5028, + "step": 839 + }, + { + "epoch": 1.2982998454404946, + "grad_norm": 5.39838056054518, + "learning_rate": 7.395644668357704e-05, + "loss": 2.3153, + "step": 840 + }, + { + "epoch": 1.2998454404945905, + "grad_norm": 6.489588147483505, + "learning_rate": 7.393361078961735e-05, + "loss": 1.9318, + "step": 841 + }, + { + "epoch": 1.3013910355486862, + "grad_norm": 4.212253493269106, + "learning_rate": 7.391073537307044e-05, + "loss": 1.5532, + "step": 842 + }, + { + "epoch": 1.302936630602782, + "grad_norm": 3.0261374214192625, + "learning_rate": 7.388782046057936e-05, + "loss": 1.2898, + "step": 843 + }, + { + "epoch": 1.3044822256568778, + "grad_norm": 2.9650921679720463, + "learning_rate": 7.386486607883321e-05, + "loss": 1.0822, + "step": 844 + }, + { + "epoch": 1.3060278207109737, + "grad_norm": 27.710845490194725, + "learning_rate": 7.384187225456702e-05, + "loss": 0.9661, + "step": 845 + }, + { + "epoch": 1.3075734157650696, + "grad_norm": 88.50090182305416, + "learning_rate": 7.381883901456177e-05, + "loss": 4.9543, + "step": 846 + }, + { + "epoch": 1.3091190108191655, + "grad_norm": 191.46826332868898, + "learning_rate": 7.379576638564434e-05, + "loss": 2.5144, + "step": 847 + }, + { + "epoch": 1.3106646058732612, + "grad_norm": 77.57634195299751, + "learning_rate": 7.377265439468752e-05, + "loss": 2.7009, + "step": 848 + }, + { + "epoch": 1.312210200927357, + "grad_norm": 13.33061951011333, + "learning_rate": 7.374950306860991e-05, + "loss": 1.4283, + "step": 849 + }, + { + "epoch": 1.3137557959814528, + "grad_norm": 7.027804561086941, + "learning_rate": 7.372631243437593e-05, + "loss": 1.117, + "step": 850 + }, + { + "epoch": 1.3153013910355487, + "grad_norm": 8.667066051805296, + "learning_rate": 7.370308251899581e-05, + "loss": 0.9451, + "step": 851 + }, + { + "epoch": 1.3168469860896446, + "grad_norm": 6.737402708053703, + "learning_rate": 7.367981334952546e-05, + "loss": 0.9061, + "step": 852 + }, + { + "epoch": 1.3183925811437405, + "grad_norm": 3.17675567551532, + "learning_rate": 7.365650495306662e-05, + "loss": 0.7909, + "step": 853 + }, + { + "epoch": 1.3199381761978362, + "grad_norm": 3.4109002512475457, + "learning_rate": 7.36331573567666e-05, + "loss": 0.7458, + "step": 854 + }, + { + "epoch": 1.321483771251932, + "grad_norm": 0.9923781639303959, + "learning_rate": 7.360977058781847e-05, + "loss": 0.6871, + "step": 855 + }, + { + "epoch": 1.3230293663060277, + "grad_norm": 1.7370513259576752, + "learning_rate": 7.358634467346083e-05, + "loss": 0.725, + "step": 856 + }, + { + "epoch": 1.3245749613601236, + "grad_norm": 2.246456041219754, + "learning_rate": 7.356287964097795e-05, + "loss": 0.6404, + "step": 857 + }, + { + "epoch": 1.3261205564142196, + "grad_norm": 1.1800484045818855, + "learning_rate": 7.353937551769962e-05, + "loss": 0.6224, + "step": 858 + }, + { + "epoch": 1.3276661514683152, + "grad_norm": 1.4459039419605137, + "learning_rate": 7.351583233100118e-05, + "loss": 0.6058, + "step": 859 + }, + { + "epoch": 1.3292117465224111, + "grad_norm": 0.6928273626415625, + "learning_rate": 7.349225010830342e-05, + "loss": 0.5829, + "step": 860 + }, + { + "epoch": 1.3307573415765068, + "grad_norm": 1.0641321797411551, + "learning_rate": 7.346862887707269e-05, + "loss": 0.5559, + "step": 861 + }, + { + "epoch": 1.3323029366306027, + "grad_norm": 1.0485031936258609, + "learning_rate": 7.344496866482065e-05, + "loss": 0.5363, + "step": 862 + }, + { + "epoch": 1.3338485316846986, + "grad_norm": 1.0744847853013832, + "learning_rate": 7.342126949910447e-05, + "loss": 0.5442, + "step": 863 + }, + { + "epoch": 1.3353941267387945, + "grad_norm": 0.7111534457833215, + "learning_rate": 7.339753140752664e-05, + "loss": 0.5227, + "step": 864 + }, + { + "epoch": 1.3369397217928902, + "grad_norm": 0.7942962396605778, + "learning_rate": 7.337375441773499e-05, + "loss": 0.5243, + "step": 865 + }, + { + "epoch": 1.3384853168469861, + "grad_norm": 0.60059709479708, + "learning_rate": 7.334993855742265e-05, + "loss": 0.5097, + "step": 866 + }, + { + "epoch": 1.3400309119010818, + "grad_norm": 0.7581651481181462, + "learning_rate": 7.332608385432803e-05, + "loss": 0.5181, + "step": 867 + }, + { + "epoch": 1.3415765069551777, + "grad_norm": 0.5308804552574883, + "learning_rate": 7.33021903362348e-05, + "loss": 0.5008, + "step": 868 + }, + { + "epoch": 1.3431221020092736, + "grad_norm": 0.9556922253495935, + "learning_rate": 7.327825803097179e-05, + "loss": 0.4885, + "step": 869 + }, + { + "epoch": 1.3446676970633695, + "grad_norm": 1.6647993772515672, + "learning_rate": 7.325428696641306e-05, + "loss": 0.487, + "step": 870 + }, + { + "epoch": 1.3462132921174652, + "grad_norm": 0.557824650859362, + "learning_rate": 7.323027717047777e-05, + "loss": 0.4965, + "step": 871 + }, + { + "epoch": 1.3477588871715611, + "grad_norm": 0.991469443849489, + "learning_rate": 7.320622867113023e-05, + "loss": 0.4938, + "step": 872 + }, + { + "epoch": 1.3493044822256568, + "grad_norm": 0.6046190116577677, + "learning_rate": 7.318214149637976e-05, + "loss": 0.4816, + "step": 873 + }, + { + "epoch": 1.3508500772797527, + "grad_norm": 0.7644227271098579, + "learning_rate": 7.315801567428081e-05, + "loss": 0.4792, + "step": 874 + }, + { + "epoch": 1.3523956723338486, + "grad_norm": 0.6458372967237748, + "learning_rate": 7.313385123293281e-05, + "loss": 0.482, + "step": 875 + }, + { + "epoch": 1.3539412673879443, + "grad_norm": 0.6907800581217399, + "learning_rate": 7.310964820048013e-05, + "loss": 0.4808, + "step": 876 + }, + { + "epoch": 1.3554868624420402, + "grad_norm": 0.6498021432208272, + "learning_rate": 7.308540660511214e-05, + "loss": 0.4713, + "step": 877 + }, + { + "epoch": 1.3570324574961359, + "grad_norm": 0.5333204285641276, + "learning_rate": 7.30611264750631e-05, + "loss": 0.4673, + "step": 878 + }, + { + "epoch": 1.3585780525502318, + "grad_norm": 0.5259273615462304, + "learning_rate": 7.303680783861216e-05, + "loss": 0.4622, + "step": 879 + }, + { + "epoch": 1.3601236476043277, + "grad_norm": 0.4221242373141024, + "learning_rate": 7.301245072408333e-05, + "loss": 0.467, + "step": 880 + }, + { + "epoch": 1.3616692426584236, + "grad_norm": 0.5096799188944314, + "learning_rate": 7.29880551598454e-05, + "loss": 0.4589, + "step": 881 + }, + { + "epoch": 1.3632148377125193, + "grad_norm": 0.4932511435948369, + "learning_rate": 7.296362117431197e-05, + "loss": 0.4532, + "step": 882 + }, + { + "epoch": 1.3647604327666152, + "grad_norm": 0.49911424004834976, + "learning_rate": 7.293914879594138e-05, + "loss": 0.4632, + "step": 883 + }, + { + "epoch": 1.3663060278207109, + "grad_norm": 0.4229006710362692, + "learning_rate": 7.291463805323668e-05, + "loss": 0.4608, + "step": 884 + }, + { + "epoch": 1.3678516228748068, + "grad_norm": 0.5394785447176259, + "learning_rate": 7.289008897474564e-05, + "loss": 0.4577, + "step": 885 + }, + { + "epoch": 1.3693972179289027, + "grad_norm": 0.38824651713858316, + "learning_rate": 7.286550158906062e-05, + "loss": 0.4386, + "step": 886 + }, + { + "epoch": 1.3709428129829986, + "grad_norm": 0.5566933255367444, + "learning_rate": 7.284087592481867e-05, + "loss": 0.4456, + "step": 887 + }, + { + "epoch": 1.3724884080370943, + "grad_norm": 0.3563089736478836, + "learning_rate": 7.281621201070133e-05, + "loss": 0.4452, + "step": 888 + }, + { + "epoch": 1.3740340030911902, + "grad_norm": 0.3561209994399552, + "learning_rate": 7.279150987543476e-05, + "loss": 0.4557, + "step": 889 + }, + { + "epoch": 1.3755795981452859, + "grad_norm": 0.36769309647797255, + "learning_rate": 7.276676954778964e-05, + "loss": 0.4525, + "step": 890 + }, + { + "epoch": 1.3771251931993818, + "grad_norm": 0.3431461137335865, + "learning_rate": 7.27419910565811e-05, + "loss": 0.4485, + "step": 891 + }, + { + "epoch": 1.3786707882534777, + "grad_norm": 0.3652454301040794, + "learning_rate": 7.271717443066871e-05, + "loss": 0.4506, + "step": 892 + }, + { + "epoch": 1.3802163833075733, + "grad_norm": 0.31578645288650126, + "learning_rate": 7.26923196989565e-05, + "loss": 0.4332, + "step": 893 + }, + { + "epoch": 1.3817619783616693, + "grad_norm": 0.34860086522725214, + "learning_rate": 7.266742689039284e-05, + "loss": 0.4448, + "step": 894 + }, + { + "epoch": 1.383307573415765, + "grad_norm": 0.2845785029690209, + "learning_rate": 7.264249603397046e-05, + "loss": 0.4472, + "step": 895 + }, + { + "epoch": 1.3848531684698608, + "grad_norm": 0.28915228241658936, + "learning_rate": 7.261752715872643e-05, + "loss": 0.4408, + "step": 896 + }, + { + "epoch": 1.3863987635239567, + "grad_norm": 0.271133894870028, + "learning_rate": 7.259252029374209e-05, + "loss": 0.4449, + "step": 897 + }, + { + "epoch": 1.3879443585780527, + "grad_norm": 0.2311051594218128, + "learning_rate": 7.256747546814298e-05, + "loss": 0.4373, + "step": 898 + }, + { + "epoch": 1.3894899536321483, + "grad_norm": 0.2853715894695218, + "learning_rate": 7.254239271109891e-05, + "loss": 0.4466, + "step": 899 + }, + { + "epoch": 1.3910355486862442, + "grad_norm": 0.25346013690449665, + "learning_rate": 7.251727205182384e-05, + "loss": 0.4372, + "step": 900 + }, + { + "epoch": 1.39258114374034, + "grad_norm": 0.2785666994423498, + "learning_rate": 7.249211351957591e-05, + "loss": 0.4345, + "step": 901 + }, + { + "epoch": 1.3941267387944358, + "grad_norm": 0.2785254538459984, + "learning_rate": 7.246691714365732e-05, + "loss": 0.4314, + "step": 902 + }, + { + "epoch": 1.3956723338485317, + "grad_norm": 0.36191851798747987, + "learning_rate": 7.244168295341437e-05, + "loss": 0.4327, + "step": 903 + }, + { + "epoch": 1.3972179289026276, + "grad_norm": 0.23887449436457348, + "learning_rate": 7.241641097823742e-05, + "loss": 0.4418, + "step": 904 + }, + { + "epoch": 1.3987635239567233, + "grad_norm": 0.26704044921405956, + "learning_rate": 7.239110124756081e-05, + "loss": 0.4413, + "step": 905 + }, + { + "epoch": 1.4003091190108192, + "grad_norm": 0.2464024352275652, + "learning_rate": 7.236575379086286e-05, + "loss": 0.4318, + "step": 906 + }, + { + "epoch": 1.401854714064915, + "grad_norm": 0.26600554185547204, + "learning_rate": 7.234036863766586e-05, + "loss": 0.4359, + "step": 907 + }, + { + "epoch": 1.4034003091190108, + "grad_norm": 0.27451211025062616, + "learning_rate": 7.231494581753596e-05, + "loss": 0.4356, + "step": 908 + }, + { + "epoch": 1.4049459041731067, + "grad_norm": 0.2998713303413568, + "learning_rate": 7.22894853600832e-05, + "loss": 0.4493, + "step": 909 + }, + { + "epoch": 1.4064914992272024, + "grad_norm": 0.3624005258089661, + "learning_rate": 7.226398729496148e-05, + "loss": 0.4358, + "step": 910 + }, + { + "epoch": 1.4080370942812983, + "grad_norm": 0.37741966314468167, + "learning_rate": 7.223845165186845e-05, + "loss": 0.4252, + "step": 911 + }, + { + "epoch": 1.409582689335394, + "grad_norm": 0.3757885636786183, + "learning_rate": 7.221287846054559e-05, + "loss": 0.428, + "step": 912 + }, + { + "epoch": 1.41112828438949, + "grad_norm": 0.33431463700971414, + "learning_rate": 7.218726775077803e-05, + "loss": 0.431, + "step": 913 + }, + { + "epoch": 1.4126738794435858, + "grad_norm": 0.33944920095897424, + "learning_rate": 7.216161955239471e-05, + "loss": 0.4315, + "step": 914 + }, + { + "epoch": 1.4142194744976817, + "grad_norm": 0.40767393932426144, + "learning_rate": 7.213593389526812e-05, + "loss": 0.4309, + "step": 915 + }, + { + "epoch": 1.4157650695517774, + "grad_norm": 0.4332624192257046, + "learning_rate": 7.211021080931446e-05, + "loss": 0.4239, + "step": 916 + }, + { + "epoch": 1.4173106646058733, + "grad_norm": 0.5367545117142651, + "learning_rate": 7.208445032449347e-05, + "loss": 0.4356, + "step": 917 + }, + { + "epoch": 1.418856259659969, + "grad_norm": 0.6797508269994943, + "learning_rate": 7.205865247080848e-05, + "loss": 0.4236, + "step": 918 + }, + { + "epoch": 1.4204018547140649, + "grad_norm": 0.8065158861855208, + "learning_rate": 7.203281727830634e-05, + "loss": 0.4273, + "step": 919 + }, + { + "epoch": 1.4219474497681608, + "grad_norm": 0.8535877358815611, + "learning_rate": 7.200694477707737e-05, + "loss": 0.4233, + "step": 920 + }, + { + "epoch": 1.4234930448222567, + "grad_norm": 0.7057296559667867, + "learning_rate": 7.198103499725538e-05, + "loss": 0.4367, + "step": 921 + }, + { + "epoch": 1.4250386398763524, + "grad_norm": 0.5329342219908334, + "learning_rate": 7.195508796901754e-05, + "loss": 0.4394, + "step": 922 + }, + { + "epoch": 1.4265842349304483, + "grad_norm": 0.4076659192290988, + "learning_rate": 7.192910372258447e-05, + "loss": 0.4256, + "step": 923 + }, + { + "epoch": 1.428129829984544, + "grad_norm": 0.26064222228962725, + "learning_rate": 7.19030822882201e-05, + "loss": 0.4323, + "step": 924 + }, + { + "epoch": 1.4296754250386399, + "grad_norm": 0.2585035718372981, + "learning_rate": 7.187702369623167e-05, + "loss": 0.4369, + "step": 925 + }, + { + "epoch": 1.4312210200927358, + "grad_norm": 0.403830163370915, + "learning_rate": 7.185092797696969e-05, + "loss": 0.4256, + "step": 926 + }, + { + "epoch": 1.4327666151468315, + "grad_norm": 0.515104682498424, + "learning_rate": 7.182479516082797e-05, + "loss": 0.4315, + "step": 927 + }, + { + "epoch": 1.4343122102009274, + "grad_norm": 0.5590711787713675, + "learning_rate": 7.179862527824347e-05, + "loss": 0.4314, + "step": 928 + }, + { + "epoch": 1.435857805255023, + "grad_norm": 0.5742581145612555, + "learning_rate": 7.177241835969632e-05, + "loss": 0.4365, + "step": 929 + }, + { + "epoch": 1.437403400309119, + "grad_norm": 0.5507747284900131, + "learning_rate": 7.174617443570981e-05, + "loss": 0.4252, + "step": 930 + }, + { + "epoch": 1.4389489953632149, + "grad_norm": 0.43550998310891703, + "learning_rate": 7.171989353685035e-05, + "loss": 0.4335, + "step": 931 + }, + { + "epoch": 1.4404945904173108, + "grad_norm": 0.34969190686092066, + "learning_rate": 7.169357569372738e-05, + "loss": 0.4292, + "step": 932 + }, + { + "epoch": 1.4420401854714064, + "grad_norm": 0.3121259773521068, + "learning_rate": 7.166722093699335e-05, + "loss": 0.4311, + "step": 933 + }, + { + "epoch": 1.4435857805255023, + "grad_norm": 0.3040459077132056, + "learning_rate": 7.164082929734375e-05, + "loss": 0.4149, + "step": 934 + }, + { + "epoch": 1.445131375579598, + "grad_norm": 0.44094007127066864, + "learning_rate": 7.161440080551701e-05, + "loss": 0.4321, + "step": 935 + }, + { + "epoch": 1.446676970633694, + "grad_norm": 0.5279397373158785, + "learning_rate": 7.158793549229449e-05, + "loss": 0.4201, + "step": 936 + }, + { + "epoch": 1.4482225656877898, + "grad_norm": 0.4726234844469759, + "learning_rate": 7.156143338850041e-05, + "loss": 0.4244, + "step": 937 + }, + { + "epoch": 1.4497681607418857, + "grad_norm": 0.4398190703180919, + "learning_rate": 7.153489452500186e-05, + "loss": 0.4271, + "step": 938 + }, + { + "epoch": 1.4513137557959814, + "grad_norm": 0.3944713485012089, + "learning_rate": 7.150831893270874e-05, + "loss": 0.4336, + "step": 939 + }, + { + "epoch": 1.4528593508500773, + "grad_norm": 0.2704413055996092, + "learning_rate": 7.148170664257374e-05, + "loss": 0.4249, + "step": 940 + }, + { + "epoch": 1.454404945904173, + "grad_norm": 0.267487761883538, + "learning_rate": 7.145505768559225e-05, + "loss": 0.4238, + "step": 941 + }, + { + "epoch": 1.455950540958269, + "grad_norm": 0.35855880795815, + "learning_rate": 7.142837209280246e-05, + "loss": 0.4251, + "step": 942 + }, + { + "epoch": 1.4574961360123648, + "grad_norm": 0.43446999920977725, + "learning_rate": 7.14016498952851e-05, + "loss": 0.4173, + "step": 943 + }, + { + "epoch": 1.4590417310664605, + "grad_norm": 0.46506642846282903, + "learning_rate": 7.137489112416363e-05, + "loss": 0.4343, + "step": 944 + }, + { + "epoch": 1.4605873261205564, + "grad_norm": 0.4088930431765722, + "learning_rate": 7.134809581060407e-05, + "loss": 0.4211, + "step": 945 + }, + { + "epoch": 1.4621329211746523, + "grad_norm": 0.27778175203358063, + "learning_rate": 7.132126398581504e-05, + "loss": 0.4273, + "step": 946 + }, + { + "epoch": 1.463678516228748, + "grad_norm": 0.2384600394764093, + "learning_rate": 7.129439568104762e-05, + "loss": 0.4189, + "step": 947 + }, + { + "epoch": 1.465224111282844, + "grad_norm": 0.24158481481830246, + "learning_rate": 7.126749092759542e-05, + "loss": 0.4222, + "step": 948 + }, + { + "epoch": 1.4667697063369398, + "grad_norm": 0.2451523418282528, + "learning_rate": 7.124054975679449e-05, + "loss": 0.4311, + "step": 949 + }, + { + "epoch": 1.4683153013910355, + "grad_norm": 0.32254793087115824, + "learning_rate": 7.121357220002331e-05, + "loss": 0.4172, + "step": 950 + }, + { + "epoch": 1.4698608964451314, + "grad_norm": 0.44144955558690707, + "learning_rate": 7.118655828870274e-05, + "loss": 0.4185, + "step": 951 + }, + { + "epoch": 1.471406491499227, + "grad_norm": 0.537749720995868, + "learning_rate": 7.115950805429591e-05, + "loss": 0.4321, + "step": 952 + }, + { + "epoch": 1.472952086553323, + "grad_norm": 0.5715097587899174, + "learning_rate": 7.11324215283084e-05, + "loss": 0.4237, + "step": 953 + }, + { + "epoch": 1.474497681607419, + "grad_norm": 0.6273212266327671, + "learning_rate": 7.110529874228792e-05, + "loss": 0.4319, + "step": 954 + }, + { + "epoch": 1.4760432766615148, + "grad_norm": 0.7889797687201212, + "learning_rate": 7.107813972782448e-05, + "loss": 0.4234, + "step": 955 + }, + { + "epoch": 1.4775888717156105, + "grad_norm": 0.7264618248619983, + "learning_rate": 7.105094451655029e-05, + "loss": 0.4232, + "step": 956 + }, + { + "epoch": 1.4791344667697064, + "grad_norm": 0.7089214741759883, + "learning_rate": 7.102371314013969e-05, + "loss": 0.4343, + "step": 957 + }, + { + "epoch": 1.480680061823802, + "grad_norm": 0.6617679962496519, + "learning_rate": 7.099644563030914e-05, + "loss": 0.4282, + "step": 958 + }, + { + "epoch": 1.482225656877898, + "grad_norm": 0.5594902654488881, + "learning_rate": 7.096914201881722e-05, + "loss": 0.4238, + "step": 959 + }, + { + "epoch": 1.4837712519319939, + "grad_norm": 0.4634050282986392, + "learning_rate": 7.09418023374645e-05, + "loss": 0.4209, + "step": 960 + }, + { + "epoch": 1.4853168469860896, + "grad_norm": 0.38011653898573167, + "learning_rate": 7.091442661809364e-05, + "loss": 0.4226, + "step": 961 + }, + { + "epoch": 1.4868624420401855, + "grad_norm": 0.28704041077599796, + "learning_rate": 7.088701489258921e-05, + "loss": 0.4237, + "step": 962 + }, + { + "epoch": 1.4884080370942814, + "grad_norm": 0.2830439462631011, + "learning_rate": 7.085956719287773e-05, + "loss": 0.4239, + "step": 963 + }, + { + "epoch": 1.489953632148377, + "grad_norm": 0.3733602880135298, + "learning_rate": 7.083208355092763e-05, + "loss": 0.4195, + "step": 964 + }, + { + "epoch": 1.491499227202473, + "grad_norm": 0.4886140984615729, + "learning_rate": 7.080456399874919e-05, + "loss": 0.4272, + "step": 965 + }, + { + "epoch": 1.4930448222565689, + "grad_norm": 0.5820865498538151, + "learning_rate": 7.077700856839453e-05, + "loss": 0.4206, + "step": 966 + }, + { + "epoch": 1.4945904173106646, + "grad_norm": 0.6735568679899008, + "learning_rate": 7.074941729195756e-05, + "loss": 0.4216, + "step": 967 + }, + { + "epoch": 1.4961360123647605, + "grad_norm": 0.7238533873018045, + "learning_rate": 7.072179020157394e-05, + "loss": 0.4237, + "step": 968 + }, + { + "epoch": 1.4976816074188561, + "grad_norm": 0.639253964759351, + "learning_rate": 7.069412732942097e-05, + "loss": 0.4267, + "step": 969 + }, + { + "epoch": 1.499227202472952, + "grad_norm": 0.517856341056609, + "learning_rate": 7.066642870771775e-05, + "loss": 0.4221, + "step": 970 + }, + { + "epoch": 1.500772797527048, + "grad_norm": 0.3988654450549474, + "learning_rate": 7.063869436872495e-05, + "loss": 0.4327, + "step": 971 + }, + { + "epoch": 1.5023183925811439, + "grad_norm": 0.3364989820133205, + "learning_rate": 7.06109243447448e-05, + "loss": 0.4195, + "step": 972 + }, + { + "epoch": 1.5038639876352395, + "grad_norm": 0.32790952085979724, + "learning_rate": 7.05831186681212e-05, + "loss": 0.4268, + "step": 973 + }, + { + "epoch": 1.5054095826893354, + "grad_norm": 0.3985317931788811, + "learning_rate": 7.055527737123947e-05, + "loss": 0.4199, + "step": 974 + }, + { + "epoch": 1.5069551777434311, + "grad_norm": 0.5260103397842368, + "learning_rate": 7.052740048652645e-05, + "loss": 0.4263, + "step": 975 + }, + { + "epoch": 1.508500772797527, + "grad_norm": 0.6525153290567401, + "learning_rate": 7.049948804645047e-05, + "loss": 0.4272, + "step": 976 + }, + { + "epoch": 1.510046367851623, + "grad_norm": 0.6846903618954039, + "learning_rate": 7.047154008352122e-05, + "loss": 0.4206, + "step": 977 + }, + { + "epoch": 1.5115919629057188, + "grad_norm": 0.6199992700429547, + "learning_rate": 7.044355663028981e-05, + "loss": 0.4228, + "step": 978 + }, + { + "epoch": 1.5131375579598145, + "grad_norm": 0.5578406454030075, + "learning_rate": 7.041553771934862e-05, + "loss": 0.4163, + "step": 979 + }, + { + "epoch": 1.5146831530139102, + "grad_norm": 0.48284322718920997, + "learning_rate": 7.03874833833314e-05, + "loss": 0.4325, + "step": 980 + }, + { + "epoch": 1.5162287480680061, + "grad_norm": 0.4282926517985966, + "learning_rate": 7.035939365491312e-05, + "loss": 0.4275, + "step": 981 + }, + { + "epoch": 1.517774343122102, + "grad_norm": 0.3969906684645652, + "learning_rate": 7.033126856680996e-05, + "loss": 0.4251, + "step": 982 + }, + { + "epoch": 1.519319938176198, + "grad_norm": 0.33465065506818753, + "learning_rate": 7.030310815177933e-05, + "loss": 0.4177, + "step": 983 + }, + { + "epoch": 1.5208655332302936, + "grad_norm": 0.3032412982070374, + "learning_rate": 7.027491244261972e-05, + "loss": 0.4232, + "step": 984 + }, + { + "epoch": 1.5224111282843895, + "grad_norm": 0.43752458748824424, + "learning_rate": 7.02466814721708e-05, + "loss": 0.4251, + "step": 985 + }, + { + "epoch": 1.5239567233384852, + "grad_norm": 0.4924572904287951, + "learning_rate": 7.021841527331327e-05, + "loss": 0.423, + "step": 986 + }, + { + "epoch": 1.525502318392581, + "grad_norm": 0.4526093033505471, + "learning_rate": 7.019011387896884e-05, + "loss": 0.4196, + "step": 987 + }, + { + "epoch": 1.527047913446677, + "grad_norm": 0.49978471023791293, + "learning_rate": 7.016177732210025e-05, + "loss": 0.4212, + "step": 988 + }, + { + "epoch": 1.528593508500773, + "grad_norm": 0.539499544187315, + "learning_rate": 7.013340563571119e-05, + "loss": 0.4113, + "step": 989 + }, + { + "epoch": 1.5301391035548686, + "grad_norm": 0.4868929178302569, + "learning_rate": 7.010499885284624e-05, + "loss": 0.4136, + "step": 990 + }, + { + "epoch": 1.5316846986089645, + "grad_norm": 0.4980297906843877, + "learning_rate": 7.007655700659087e-05, + "loss": 0.4299, + "step": 991 + }, + { + "epoch": 1.5332302936630602, + "grad_norm": 0.582892376266293, + "learning_rate": 7.004808013007142e-05, + "loss": 0.4242, + "step": 992 + }, + { + "epoch": 1.534775888717156, + "grad_norm": 0.5798827136342914, + "learning_rate": 7.001956825645496e-05, + "loss": 0.4194, + "step": 993 + }, + { + "epoch": 1.536321483771252, + "grad_norm": 0.5309078533844069, + "learning_rate": 6.999102141894939e-05, + "loss": 0.4234, + "step": 994 + }, + { + "epoch": 1.537867078825348, + "grad_norm": 0.5430064404970218, + "learning_rate": 6.99624396508033e-05, + "loss": 0.4254, + "step": 995 + }, + { + "epoch": 1.5394126738794436, + "grad_norm": 0.46571511147801864, + "learning_rate": 6.993382298530598e-05, + "loss": 0.4128, + "step": 996 + }, + { + "epoch": 1.5409582689335393, + "grad_norm": 0.29907402931903726, + "learning_rate": 6.990517145578733e-05, + "loss": 0.4205, + "step": 997 + }, + { + "epoch": 1.5425038639876352, + "grad_norm": 0.33136439887119, + "learning_rate": 6.987648509561788e-05, + "loss": 0.4279, + "step": 998 + }, + { + "epoch": 1.544049459041731, + "grad_norm": 0.4477695438478884, + "learning_rate": 6.984776393820876e-05, + "loss": 0.4212, + "step": 999 + }, + { + "epoch": 1.545595054095827, + "grad_norm": 0.4487614235398153, + "learning_rate": 6.981900801701156e-05, + "loss": 0.4167, + "step": 1000 + }, + { + "epoch": 1.5471406491499229, + "grad_norm": 0.36555309719909523, + "learning_rate": 6.979021736551841e-05, + "loss": 0.4261, + "step": 1001 + }, + { + "epoch": 1.5486862442040186, + "grad_norm": 0.3501723419708198, + "learning_rate": 6.976139201726189e-05, + "loss": 0.418, + "step": 1002 + }, + { + "epoch": 1.5502318392581143, + "grad_norm": 0.3651564751235074, + "learning_rate": 6.973253200581493e-05, + "loss": 0.4221, + "step": 1003 + }, + { + "epoch": 1.5517774343122102, + "grad_norm": 0.34570119420075257, + "learning_rate": 6.970363736479094e-05, + "loss": 0.414, + "step": 1004 + }, + { + "epoch": 1.553323029366306, + "grad_norm": 0.4805761331573332, + "learning_rate": 6.967470812784356e-05, + "loss": 0.4182, + "step": 1005 + }, + { + "epoch": 1.554868624420402, + "grad_norm": 0.5649790110710811, + "learning_rate": 6.964574432866679e-05, + "loss": 0.4298, + "step": 1006 + }, + { + "epoch": 1.5564142194744977, + "grad_norm": 0.6221087425963319, + "learning_rate": 6.961674600099484e-05, + "loss": 0.4216, + "step": 1007 + }, + { + "epoch": 1.5579598145285936, + "grad_norm": 0.7181645539552841, + "learning_rate": 6.958771317860218e-05, + "loss": 0.4316, + "step": 1008 + }, + { + "epoch": 1.5595054095826892, + "grad_norm": 0.8221465058092516, + "learning_rate": 6.955864589530342e-05, + "loss": 0.4168, + "step": 1009 + }, + { + "epoch": 1.5610510046367851, + "grad_norm": 0.8898827490422138, + "learning_rate": 6.952954418495332e-05, + "loss": 0.4158, + "step": 1010 + }, + { + "epoch": 1.562596599690881, + "grad_norm": 0.9006817281728228, + "learning_rate": 6.950040808144674e-05, + "loss": 0.4169, + "step": 1011 + }, + { + "epoch": 1.564142194744977, + "grad_norm": 0.7845557662937626, + "learning_rate": 6.947123761871858e-05, + "loss": 0.4235, + "step": 1012 + }, + { + "epoch": 1.5656877897990726, + "grad_norm": 0.6167610107609887, + "learning_rate": 6.944203283074381e-05, + "loss": 0.4178, + "step": 1013 + }, + { + "epoch": 1.5672333848531683, + "grad_norm": 0.45690594363128245, + "learning_rate": 6.94127937515373e-05, + "loss": 0.4302, + "step": 1014 + }, + { + "epoch": 1.5687789799072642, + "grad_norm": 0.3499403023769347, + "learning_rate": 6.938352041515393e-05, + "loss": 0.4199, + "step": 1015 + }, + { + "epoch": 1.5703245749613601, + "grad_norm": 0.3572705438871156, + "learning_rate": 6.935421285568842e-05, + "loss": 0.4173, + "step": 1016 + }, + { + "epoch": 1.571870170015456, + "grad_norm": 0.44950483927508295, + "learning_rate": 6.932487110727543e-05, + "loss": 0.4209, + "step": 1017 + }, + { + "epoch": 1.573415765069552, + "grad_norm": 3.1292477500728544, + "learning_rate": 6.929549520408934e-05, + "loss": 0.4231, + "step": 1018 + }, + { + "epoch": 1.5749613601236476, + "grad_norm": 1.8980381367708854, + "learning_rate": 6.926608518034438e-05, + "loss": 0.4377, + "step": 1019 + }, + { + "epoch": 1.5765069551777433, + "grad_norm": 2.830478244840452, + "learning_rate": 6.923664107029452e-05, + "loss": 0.4763, + "step": 1020 + }, + { + "epoch": 1.5780525502318392, + "grad_norm": 0.9592222026976776, + "learning_rate": 6.920716290823337e-05, + "loss": 0.4398, + "step": 1021 + }, + { + "epoch": 1.5795981452859351, + "grad_norm": 1.6478299306991953, + "learning_rate": 6.917765072849427e-05, + "loss": 0.4653, + "step": 1022 + }, + { + "epoch": 1.581143740340031, + "grad_norm": 0.6840248212604744, + "learning_rate": 6.914810456545015e-05, + "loss": 0.4435, + "step": 1023 + }, + { + "epoch": 1.5826893353941267, + "grad_norm": 1.1177393718394315, + "learning_rate": 6.911852445351352e-05, + "loss": 0.4419, + "step": 1024 + }, + { + "epoch": 1.5842349304482226, + "grad_norm": 0.6531693590989968, + "learning_rate": 6.908891042713643e-05, + "loss": 0.4516, + "step": 1025 + }, + { + "epoch": 1.5857805255023183, + "grad_norm": 0.7123299350546227, + "learning_rate": 6.905926252081043e-05, + "loss": 0.4379, + "step": 1026 + }, + { + "epoch": 1.5873261205564142, + "grad_norm": 0.6393235977873607, + "learning_rate": 6.902958076906655e-05, + "loss": 0.4366, + "step": 1027 + }, + { + "epoch": 1.58887171561051, + "grad_norm": 0.6699748000095743, + "learning_rate": 6.899986520647521e-05, + "loss": 0.4337, + "step": 1028 + }, + { + "epoch": 1.590417310664606, + "grad_norm": 0.5965947765071413, + "learning_rate": 6.897011586764622e-05, + "loss": 0.4321, + "step": 1029 + }, + { + "epoch": 1.5919629057187017, + "grad_norm": 0.5056237819580799, + "learning_rate": 6.894033278722874e-05, + "loss": 0.4277, + "step": 1030 + }, + { + "epoch": 1.5935085007727974, + "grad_norm": 0.5083639625187817, + "learning_rate": 6.891051599991119e-05, + "loss": 0.4234, + "step": 1031 + }, + { + "epoch": 1.5950540958268933, + "grad_norm": 0.5046816637015624, + "learning_rate": 6.888066554042134e-05, + "loss": 0.4333, + "step": 1032 + }, + { + "epoch": 1.5965996908809892, + "grad_norm": 0.40281242552896174, + "learning_rate": 6.885078144352606e-05, + "loss": 0.4271, + "step": 1033 + }, + { + "epoch": 1.598145285935085, + "grad_norm": 0.4000957468169698, + "learning_rate": 6.882086374403148e-05, + "loss": 0.434, + "step": 1034 + }, + { + "epoch": 1.599690880989181, + "grad_norm": 0.4656514897022174, + "learning_rate": 6.879091247678285e-05, + "loss": 0.4336, + "step": 1035 + }, + { + "epoch": 1.6012364760432767, + "grad_norm": 0.3851788185826526, + "learning_rate": 6.87609276766645e-05, + "loss": 0.4265, + "step": 1036 + }, + { + "epoch": 1.6027820710973724, + "grad_norm": 0.3663704499902741, + "learning_rate": 6.873090937859982e-05, + "loss": 0.4309, + "step": 1037 + }, + { + "epoch": 1.6043276661514683, + "grad_norm": 0.7346406351853385, + "learning_rate": 6.870085761755122e-05, + "loss": 0.4228, + "step": 1038 + }, + { + "epoch": 1.6058732612055642, + "grad_norm": 0.46844196197914234, + "learning_rate": 6.867077242852012e-05, + "loss": 0.425, + "step": 1039 + }, + { + "epoch": 1.60741885625966, + "grad_norm": 0.49329122070089737, + "learning_rate": 6.864065384654676e-05, + "loss": 0.4243, + "step": 1040 + }, + { + "epoch": 1.6089644513137558, + "grad_norm": 0.3696669871281922, + "learning_rate": 6.861050190671043e-05, + "loss": 0.4185, + "step": 1041 + }, + { + "epoch": 1.6105100463678517, + "grad_norm": 0.33531974942756926, + "learning_rate": 6.858031664412914e-05, + "loss": 0.4301, + "step": 1042 + }, + { + "epoch": 1.6120556414219473, + "grad_norm": 0.32366833386851807, + "learning_rate": 6.85500980939598e-05, + "loss": 0.4258, + "step": 1043 + }, + { + "epoch": 1.6136012364760433, + "grad_norm": 0.3198516867842109, + "learning_rate": 6.851984629139802e-05, + "loss": 0.4248, + "step": 1044 + }, + { + "epoch": 1.6151468315301392, + "grad_norm": 0.3683588176040148, + "learning_rate": 6.84895612716782e-05, + "loss": 0.4169, + "step": 1045 + }, + { + "epoch": 1.616692426584235, + "grad_norm": 0.3701188524554859, + "learning_rate": 6.845924307007339e-05, + "loss": 0.4221, + "step": 1046 + }, + { + "epoch": 1.6182380216383307, + "grad_norm": 0.2974961436601128, + "learning_rate": 6.84288917218953e-05, + "loss": 0.4268, + "step": 1047 + }, + { + "epoch": 1.6197836166924264, + "grad_norm": 0.28565881219366634, + "learning_rate": 6.839850726249423e-05, + "loss": 0.4229, + "step": 1048 + }, + { + "epoch": 1.6213292117465223, + "grad_norm": 0.3339494717656481, + "learning_rate": 6.83680897272591e-05, + "loss": 0.4194, + "step": 1049 + }, + { + "epoch": 1.6228748068006182, + "grad_norm": 0.37085420309921574, + "learning_rate": 6.833763915161727e-05, + "loss": 0.4234, + "step": 1050 + }, + { + "epoch": 1.6244204018547141, + "grad_norm": 0.3687746373768453, + "learning_rate": 6.830715557103465e-05, + "loss": 0.4255, + "step": 1051 + }, + { + "epoch": 1.62596599690881, + "grad_norm": 0.4853540323313802, + "learning_rate": 6.827663902101554e-05, + "loss": 0.4263, + "step": 1052 + }, + { + "epoch": 1.6275115919629057, + "grad_norm": 0.506980707599996, + "learning_rate": 6.82460895371027e-05, + "loss": 0.4244, + "step": 1053 + }, + { + "epoch": 1.6290571870170014, + "grad_norm": 0.4384308700255955, + "learning_rate": 6.82155071548772e-05, + "loss": 0.4341, + "step": 1054 + }, + { + "epoch": 1.6306027820710973, + "grad_norm": 0.29139242680715716, + "learning_rate": 6.818489190995842e-05, + "loss": 0.4151, + "step": 1055 + }, + { + "epoch": 1.6321483771251932, + "grad_norm": 0.24337771986454898, + "learning_rate": 6.815424383800405e-05, + "loss": 0.4257, + "step": 1056 + }, + { + "epoch": 1.6336939721792891, + "grad_norm": 0.30221847310026523, + "learning_rate": 6.812356297471e-05, + "loss": 0.4191, + "step": 1057 + }, + { + "epoch": 1.6352395672333848, + "grad_norm": 0.34157919884657423, + "learning_rate": 6.809284935581039e-05, + "loss": 0.4249, + "step": 1058 + }, + { + "epoch": 1.6367851622874807, + "grad_norm": 0.3627131538237721, + "learning_rate": 6.806210301707745e-05, + "loss": 0.4176, + "step": 1059 + }, + { + "epoch": 1.6383307573415764, + "grad_norm": 0.38110607158808035, + "learning_rate": 6.803132399432154e-05, + "loss": 0.4169, + "step": 1060 + }, + { + "epoch": 1.6398763523956723, + "grad_norm": 0.33843619857889295, + "learning_rate": 6.800051232339111e-05, + "loss": 0.4201, + "step": 1061 + }, + { + "epoch": 1.6414219474497682, + "grad_norm": 0.2546149999249603, + "learning_rate": 6.796966804017258e-05, + "loss": 0.413, + "step": 1062 + }, + { + "epoch": 1.6429675425038641, + "grad_norm": 0.18453053277347461, + "learning_rate": 6.79387911805904e-05, + "loss": 0.4175, + "step": 1063 + }, + { + "epoch": 1.6445131375579598, + "grad_norm": 0.24557042084259198, + "learning_rate": 6.790788178060697e-05, + "loss": 0.4165, + "step": 1064 + }, + { + "epoch": 1.6460587326120555, + "grad_norm": 0.2269674504241286, + "learning_rate": 6.787693987622254e-05, + "loss": 0.4185, + "step": 1065 + }, + { + "epoch": 1.6476043276661514, + "grad_norm": 0.19447194748578533, + "learning_rate": 6.784596550347526e-05, + "loss": 0.4366, + "step": 1066 + }, + { + "epoch": 1.6491499227202473, + "grad_norm": 0.23111756984294846, + "learning_rate": 6.78149586984411e-05, + "loss": 0.4316, + "step": 1067 + }, + { + "epoch": 1.6506955177743432, + "grad_norm": 0.2256982718893023, + "learning_rate": 6.778391949723375e-05, + "loss": 0.4127, + "step": 1068 + }, + { + "epoch": 1.652241112828439, + "grad_norm": 0.2332090184401883, + "learning_rate": 6.77528479360047e-05, + "loss": 0.4049, + "step": 1069 + }, + { + "epoch": 1.6537867078825348, + "grad_norm": 0.23913692934004846, + "learning_rate": 6.772174405094309e-05, + "loss": 0.4296, + "step": 1070 + }, + { + "epoch": 1.6553323029366305, + "grad_norm": 0.3520928886160808, + "learning_rate": 6.769060787827571e-05, + "loss": 0.4093, + "step": 1071 + }, + { + "epoch": 1.6568778979907264, + "grad_norm": 0.5369071929986187, + "learning_rate": 6.765943945426697e-05, + "loss": 0.4265, + "step": 1072 + }, + { + "epoch": 1.6584234930448223, + "grad_norm": 0.7007120576774098, + "learning_rate": 6.762823881521883e-05, + "loss": 0.4203, + "step": 1073 + }, + { + "epoch": 1.6599690880989182, + "grad_norm": 0.77923246824803, + "learning_rate": 6.759700599747077e-05, + "loss": 0.4313, + "step": 1074 + }, + { + "epoch": 1.6615146831530139, + "grad_norm": 0.740452462478401, + "learning_rate": 6.756574103739979e-05, + "loss": 0.4212, + "step": 1075 + }, + { + "epoch": 1.6630602782071098, + "grad_norm": 0.6678975079052005, + "learning_rate": 6.753444397142024e-05, + "loss": 0.4264, + "step": 1076 + }, + { + "epoch": 1.6646058732612055, + "grad_norm": 0.6013591124678156, + "learning_rate": 6.750311483598394e-05, + "loss": 0.4174, + "step": 1077 + }, + { + "epoch": 1.6661514683153014, + "grad_norm": 0.4321985899818665, + "learning_rate": 6.747175366758005e-05, + "loss": 0.4174, + "step": 1078 + }, + { + "epoch": 1.6676970633693973, + "grad_norm": 0.24934380170419806, + "learning_rate": 6.7440360502735e-05, + "loss": 0.4294, + "step": 1079 + }, + { + "epoch": 1.6692426584234932, + "grad_norm": 0.3200073164107801, + "learning_rate": 6.74089353780125e-05, + "loss": 0.4118, + "step": 1080 + }, + { + "epoch": 1.6707882534775889, + "grad_norm": 0.44053524474232914, + "learning_rate": 6.737747833001354e-05, + "loss": 0.4102, + "step": 1081 + }, + { + "epoch": 1.6723338485316845, + "grad_norm": 0.47983582823985493, + "learning_rate": 6.734598939537621e-05, + "loss": 0.4181, + "step": 1082 + }, + { + "epoch": 1.6738794435857804, + "grad_norm": 0.5191238281114948, + "learning_rate": 6.731446861077581e-05, + "loss": 0.4165, + "step": 1083 + }, + { + "epoch": 1.6754250386398764, + "grad_norm": 0.5029319945077619, + "learning_rate": 6.728291601292466e-05, + "loss": 0.4303, + "step": 1084 + }, + { + "epoch": 1.6769706336939723, + "grad_norm": 0.4309530318214376, + "learning_rate": 6.725133163857219e-05, + "loss": 0.4228, + "step": 1085 + }, + { + "epoch": 1.6785162287480682, + "grad_norm": 0.4503679207752796, + "learning_rate": 6.721971552450483e-05, + "loss": 0.4153, + "step": 1086 + }, + { + "epoch": 1.6800618238021638, + "grad_norm": 0.49171363614832303, + "learning_rate": 6.7188067707546e-05, + "loss": 0.4173, + "step": 1087 + }, + { + "epoch": 1.6816074188562595, + "grad_norm": 0.4889423713612924, + "learning_rate": 6.715638822455594e-05, + "loss": 0.4163, + "step": 1088 + }, + { + "epoch": 1.6831530139103554, + "grad_norm": 0.44497904897047286, + "learning_rate": 6.712467711243191e-05, + "loss": 0.4188, + "step": 1089 + }, + { + "epoch": 1.6846986089644513, + "grad_norm": 0.39806898970401583, + "learning_rate": 6.709293440810792e-05, + "loss": 0.4131, + "step": 1090 + }, + { + "epoch": 1.6862442040185472, + "grad_norm": 0.33195307395967655, + "learning_rate": 6.70611601485548e-05, + "loss": 0.4189, + "step": 1091 + }, + { + "epoch": 1.687789799072643, + "grad_norm": 0.8871998849464271, + "learning_rate": 6.702935437078017e-05, + "loss": 0.4168, + "step": 1092 + }, + { + "epoch": 1.6893353941267388, + "grad_norm": 0.44193443814333394, + "learning_rate": 6.699751711182827e-05, + "loss": 0.4101, + "step": 1093 + }, + { + "epoch": 1.6908809891808345, + "grad_norm": 0.4141123061157703, + "learning_rate": 6.696564840878009e-05, + "loss": 0.4262, + "step": 1094 + }, + { + "epoch": 1.6924265842349304, + "grad_norm": 0.4973649245570431, + "learning_rate": 6.693374829875322e-05, + "loss": 0.4225, + "step": 1095 + }, + { + "epoch": 1.6939721792890263, + "grad_norm": 0.6688557554900625, + "learning_rate": 6.690181681890177e-05, + "loss": 0.4114, + "step": 1096 + }, + { + "epoch": 1.6955177743431222, + "grad_norm": 0.6857166826042449, + "learning_rate": 6.686985400641651e-05, + "loss": 0.4217, + "step": 1097 + }, + { + "epoch": 1.697063369397218, + "grad_norm": 0.5830106585335436, + "learning_rate": 6.68378598985246e-05, + "loss": 0.4219, + "step": 1098 + }, + { + "epoch": 1.6986089644513136, + "grad_norm": 0.501857289501519, + "learning_rate": 6.680583453248967e-05, + "loss": 0.419, + "step": 1099 + }, + { + "epoch": 1.7001545595054095, + "grad_norm": 0.4796633764215541, + "learning_rate": 6.677377794561178e-05, + "loss": 0.423, + "step": 1100 + }, + { + "epoch": 1.7017001545595054, + "grad_norm": 0.43334114327907036, + "learning_rate": 6.674169017522734e-05, + "loss": 0.4201, + "step": 1101 + }, + { + "epoch": 1.7032457496136013, + "grad_norm": 0.3364871560607749, + "learning_rate": 6.670957125870911e-05, + "loss": 0.4278, + "step": 1102 + }, + { + "epoch": 1.7047913446676972, + "grad_norm": 0.4011619253995148, + "learning_rate": 6.667742123346603e-05, + "loss": 0.4216, + "step": 1103 + }, + { + "epoch": 1.706336939721793, + "grad_norm": 0.40592447432225237, + "learning_rate": 6.664524013694341e-05, + "loss": 0.4153, + "step": 1104 + }, + { + "epoch": 1.7078825347758886, + "grad_norm": 0.40454456250038306, + "learning_rate": 6.661302800662261e-05, + "loss": 0.424, + "step": 1105 + }, + { + "epoch": 1.7094281298299845, + "grad_norm": 0.391031986132384, + "learning_rate": 6.658078488002128e-05, + "loss": 0.416, + "step": 1106 + }, + { + "epoch": 1.7109737248840804, + "grad_norm": 0.318555248550208, + "learning_rate": 6.654851079469303e-05, + "loss": 0.4194, + "step": 1107 + }, + { + "epoch": 1.7125193199381763, + "grad_norm": 0.2833193496747006, + "learning_rate": 6.651620578822762e-05, + "loss": 0.4226, + "step": 1108 + }, + { + "epoch": 1.714064914992272, + "grad_norm": 0.4102788785301504, + "learning_rate": 6.64838698982508e-05, + "loss": 0.4209, + "step": 1109 + }, + { + "epoch": 1.7156105100463679, + "grad_norm": 0.43559268625858316, + "learning_rate": 6.645150316242426e-05, + "loss": 0.414, + "step": 1110 + }, + { + "epoch": 1.7171561051004636, + "grad_norm": 0.3400148358964731, + "learning_rate": 6.641910561844569e-05, + "loss": 0.4091, + "step": 1111 + }, + { + "epoch": 1.7187017001545595, + "grad_norm": 0.3106594308436494, + "learning_rate": 6.638667730404856e-05, + "loss": 0.4236, + "step": 1112 + }, + { + "epoch": 1.7202472952086554, + "grad_norm": 0.2581600125356283, + "learning_rate": 6.635421825700228e-05, + "loss": 0.4238, + "step": 1113 + }, + { + "epoch": 1.7217928902627513, + "grad_norm": 0.3077881881891871, + "learning_rate": 6.6321728515112e-05, + "loss": 0.4231, + "step": 1114 + }, + { + "epoch": 1.723338485316847, + "grad_norm": 0.4573839706064513, + "learning_rate": 6.62892081162186e-05, + "loss": 0.4123, + "step": 1115 + }, + { + "epoch": 1.7248840803709427, + "grad_norm": 0.42187828947492784, + "learning_rate": 6.625665709819871e-05, + "loss": 0.4278, + "step": 1116 + }, + { + "epoch": 1.7264296754250386, + "grad_norm": 0.3485554331780731, + "learning_rate": 6.622407549896462e-05, + "loss": 0.4191, + "step": 1117 + }, + { + "epoch": 1.7279752704791345, + "grad_norm": 0.237263508252921, + "learning_rate": 6.619146335646422e-05, + "loss": 0.4095, + "step": 1118 + }, + { + "epoch": 1.7295208655332304, + "grad_norm": 0.3190353533959669, + "learning_rate": 6.615882070868096e-05, + "loss": 0.4143, + "step": 1119 + }, + { + "epoch": 1.7310664605873263, + "grad_norm": 0.4023411179851799, + "learning_rate": 6.612614759363386e-05, + "loss": 0.4245, + "step": 1120 + }, + { + "epoch": 1.732612055641422, + "grad_norm": 0.296489833215759, + "learning_rate": 6.609344404937739e-05, + "loss": 0.4166, + "step": 1121 + }, + { + "epoch": 1.7341576506955176, + "grad_norm": 0.34685746967755765, + "learning_rate": 6.606071011400148e-05, + "loss": 0.4272, + "step": 1122 + }, + { + "epoch": 1.7357032457496135, + "grad_norm": 0.3189024251074944, + "learning_rate": 6.602794582563146e-05, + "loss": 0.411, + "step": 1123 + }, + { + "epoch": 1.7372488408037094, + "grad_norm": 0.34829829224625086, + "learning_rate": 6.599515122242799e-05, + "loss": 0.4158, + "step": 1124 + }, + { + "epoch": 1.7387944358578054, + "grad_norm": 0.3899462123118788, + "learning_rate": 6.596232634258706e-05, + "loss": 0.4176, + "step": 1125 + }, + { + "epoch": 1.740340030911901, + "grad_norm": 0.3589594431918154, + "learning_rate": 6.592947122433992e-05, + "loss": 0.4153, + "step": 1126 + }, + { + "epoch": 1.741885625965997, + "grad_norm": 0.32624874047244823, + "learning_rate": 6.5896585905953e-05, + "loss": 0.4257, + "step": 1127 + }, + { + "epoch": 1.7434312210200926, + "grad_norm": 0.3573946165560419, + "learning_rate": 6.586367042572797e-05, + "loss": 0.4186, + "step": 1128 + }, + { + "epoch": 1.7449768160741885, + "grad_norm": 0.33050528839284254, + "learning_rate": 6.583072482200158e-05, + "loss": 0.4205, + "step": 1129 + }, + { + "epoch": 1.7465224111282844, + "grad_norm": 0.28947624973873576, + "learning_rate": 6.579774913314571e-05, + "loss": 0.4116, + "step": 1130 + }, + { + "epoch": 1.7480680061823803, + "grad_norm": 0.36295085848997277, + "learning_rate": 6.576474339756723e-05, + "loss": 0.4151, + "step": 1131 + }, + { + "epoch": 1.749613601236476, + "grad_norm": 0.3375864893741672, + "learning_rate": 6.573170765370805e-05, + "loss": 0.4207, + "step": 1132 + }, + { + "epoch": 1.7511591962905717, + "grad_norm": 0.2574122901993273, + "learning_rate": 6.569864194004498e-05, + "loss": 0.4147, + "step": 1133 + }, + { + "epoch": 1.7527047913446676, + "grad_norm": 0.23678817072974698, + "learning_rate": 6.566554629508976e-05, + "loss": 0.425, + "step": 1134 + }, + { + "epoch": 1.7542503863987635, + "grad_norm": 0.29487274917096445, + "learning_rate": 6.563242075738902e-05, + "loss": 0.4144, + "step": 1135 + }, + { + "epoch": 1.7557959814528594, + "grad_norm": 0.37107506017243763, + "learning_rate": 6.559926536552417e-05, + "loss": 0.4229, + "step": 1136 + }, + { + "epoch": 1.7573415765069553, + "grad_norm": 0.3231852598757117, + "learning_rate": 6.556608015811142e-05, + "loss": 0.4154, + "step": 1137 + }, + { + "epoch": 1.758887171561051, + "grad_norm": 0.34463743975189143, + "learning_rate": 6.553286517380166e-05, + "loss": 0.4136, + "step": 1138 + }, + { + "epoch": 1.7604327666151467, + "grad_norm": 0.241224707262446, + "learning_rate": 6.54996204512805e-05, + "loss": 0.4152, + "step": 1139 + }, + { + "epoch": 1.7619783616692426, + "grad_norm": 0.27379722315518734, + "learning_rate": 6.546634602926821e-05, + "loss": 0.4247, + "step": 1140 + }, + { + "epoch": 1.7635239567233385, + "grad_norm": 0.3486256104244915, + "learning_rate": 6.543304194651956e-05, + "loss": 0.4129, + "step": 1141 + }, + { + "epoch": 1.7650695517774344, + "grad_norm": 0.2647142219521478, + "learning_rate": 6.539970824182398e-05, + "loss": 0.4174, + "step": 1142 + }, + { + "epoch": 1.76661514683153, + "grad_norm": 0.26381420485930107, + "learning_rate": 6.536634495400531e-05, + "loss": 0.4086, + "step": 1143 + }, + { + "epoch": 1.768160741885626, + "grad_norm": 0.3078250326202281, + "learning_rate": 6.533295212192189e-05, + "loss": 0.4258, + "step": 1144 + }, + { + "epoch": 1.7697063369397217, + "grad_norm": 0.37360398510483267, + "learning_rate": 6.529952978446645e-05, + "loss": 0.4109, + "step": 1145 + }, + { + "epoch": 1.7712519319938176, + "grad_norm": 0.44167793126435473, + "learning_rate": 6.526607798056612e-05, + "loss": 0.4211, + "step": 1146 + }, + { + "epoch": 1.7727975270479135, + "grad_norm": 0.454035260878904, + "learning_rate": 6.52325967491823e-05, + "loss": 0.4215, + "step": 1147 + }, + { + "epoch": 1.7743431221020094, + "grad_norm": 0.4246007986639795, + "learning_rate": 6.519908612931071e-05, + "loss": 0.4097, + "step": 1148 + }, + { + "epoch": 1.775888717156105, + "grad_norm": 0.5040633836956859, + "learning_rate": 6.516554615998126e-05, + "loss": 0.4154, + "step": 1149 + }, + { + "epoch": 1.7774343122102008, + "grad_norm": 0.5323469886608114, + "learning_rate": 6.513197688025808e-05, + "loss": 0.4101, + "step": 1150 + }, + { + "epoch": 1.7789799072642967, + "grad_norm": 0.540753440436067, + "learning_rate": 6.509837832923939e-05, + "loss": 0.4176, + "step": 1151 + }, + { + "epoch": 1.7805255023183926, + "grad_norm": 0.5436390782696222, + "learning_rate": 6.506475054605757e-05, + "loss": 0.4108, + "step": 1152 + }, + { + "epoch": 1.7820710973724885, + "grad_norm": 0.5167138471794511, + "learning_rate": 6.503109356987899e-05, + "loss": 0.4093, + "step": 1153 + }, + { + "epoch": 1.7836166924265844, + "grad_norm": 0.5670069583521865, + "learning_rate": 6.499740743990404e-05, + "loss": 0.4169, + "step": 1154 + }, + { + "epoch": 1.78516228748068, + "grad_norm": 0.6272309850462795, + "learning_rate": 6.496369219536706e-05, + "loss": 0.4182, + "step": 1155 + }, + { + "epoch": 1.7867078825347757, + "grad_norm": 0.5321810006856598, + "learning_rate": 6.492994787553632e-05, + "loss": 0.4153, + "step": 1156 + }, + { + "epoch": 1.7882534775888717, + "grad_norm": 0.3947246482796357, + "learning_rate": 6.489617451971392e-05, + "loss": 0.4185, + "step": 1157 + }, + { + "epoch": 1.7897990726429676, + "grad_norm": 0.2919535338412543, + "learning_rate": 6.486237216723581e-05, + "loss": 0.4234, + "step": 1158 + }, + { + "epoch": 1.7913446676970635, + "grad_norm": 0.31757558124492435, + "learning_rate": 6.482854085747171e-05, + "loss": 0.4107, + "step": 1159 + }, + { + "epoch": 1.7928902627511591, + "grad_norm": 0.41985931089669054, + "learning_rate": 6.479468062982505e-05, + "loss": 0.4116, + "step": 1160 + }, + { + "epoch": 1.794435857805255, + "grad_norm": 0.43819798291205037, + "learning_rate": 6.476079152373292e-05, + "loss": 0.4077, + "step": 1161 + }, + { + "epoch": 1.7959814528593507, + "grad_norm": 0.4394654469509475, + "learning_rate": 6.472687357866611e-05, + "loss": 0.4076, + "step": 1162 + }, + { + "epoch": 1.7975270479134466, + "grad_norm": 0.3421834960647247, + "learning_rate": 6.469292683412895e-05, + "loss": 0.4157, + "step": 1163 + }, + { + "epoch": 1.7990726429675425, + "grad_norm": 0.261891084008789, + "learning_rate": 6.465895132965931e-05, + "loss": 0.4156, + "step": 1164 + }, + { + "epoch": 1.8006182380216385, + "grad_norm": 0.2566986603550055, + "learning_rate": 6.46249471048286e-05, + "loss": 0.4144, + "step": 1165 + }, + { + "epoch": 1.8021638330757341, + "grad_norm": 0.3715925083121105, + "learning_rate": 6.459091419924162e-05, + "loss": 0.4127, + "step": 1166 + }, + { + "epoch": 1.80370942812983, + "grad_norm": 0.44997059607774637, + "learning_rate": 6.455685265253663e-05, + "loss": 0.4108, + "step": 1167 + }, + { + "epoch": 1.8052550231839257, + "grad_norm": 0.39558533003170193, + "learning_rate": 6.452276250438521e-05, + "loss": 0.4082, + "step": 1168 + }, + { + "epoch": 1.8068006182380216, + "grad_norm": 0.3205412972742549, + "learning_rate": 6.448864379449227e-05, + "loss": 0.4145, + "step": 1169 + }, + { + "epoch": 1.8083462132921175, + "grad_norm": 0.33560562911213015, + "learning_rate": 6.445449656259598e-05, + "loss": 0.4077, + "step": 1170 + }, + { + "epoch": 1.8098918083462134, + "grad_norm": 0.3408721236384103, + "learning_rate": 6.442032084846775e-05, + "loss": 0.4204, + "step": 1171 + }, + { + "epoch": 1.8114374034003091, + "grad_norm": 0.32642351132837616, + "learning_rate": 6.438611669191212e-05, + "loss": 0.4099, + "step": 1172 + }, + { + "epoch": 1.8129829984544048, + "grad_norm": 0.3008497794435818, + "learning_rate": 6.435188413276679e-05, + "loss": 0.4175, + "step": 1173 + }, + { + "epoch": 1.8145285935085007, + "grad_norm": 0.3161866790828201, + "learning_rate": 6.431762321090255e-05, + "loss": 0.426, + "step": 1174 + }, + { + "epoch": 1.8160741885625966, + "grad_norm": 0.34293499287346135, + "learning_rate": 6.42833339662232e-05, + "loss": 0.4227, + "step": 1175 + }, + { + "epoch": 1.8176197836166925, + "grad_norm": 0.32808353325790235, + "learning_rate": 6.424901643866553e-05, + "loss": 0.4163, + "step": 1176 + }, + { + "epoch": 1.8191653786707882, + "grad_norm": 0.3659363750905557, + "learning_rate": 6.421467066819927e-05, + "loss": 0.4181, + "step": 1177 + }, + { + "epoch": 1.820710973724884, + "grad_norm": 0.4687589192970821, + "learning_rate": 6.418029669482707e-05, + "loss": 0.4099, + "step": 1178 + }, + { + "epoch": 1.8222565687789798, + "grad_norm": 0.46175089574388317, + "learning_rate": 6.414589455858438e-05, + "loss": 0.4068, + "step": 1179 + }, + { + "epoch": 1.8238021638330757, + "grad_norm": 0.47340277937673997, + "learning_rate": 6.411146429953951e-05, + "loss": 0.4033, + "step": 1180 + }, + { + "epoch": 1.8253477588871716, + "grad_norm": 0.5999195250605643, + "learning_rate": 6.407700595779348e-05, + "loss": 0.4199, + "step": 1181 + }, + { + "epoch": 1.8268933539412675, + "grad_norm": 0.6650778602296572, + "learning_rate": 6.404251957348005e-05, + "loss": 0.4128, + "step": 1182 + }, + { + "epoch": 1.8284389489953632, + "grad_norm": 0.6605952338300355, + "learning_rate": 6.400800518676558e-05, + "loss": 0.4236, + "step": 1183 + }, + { + "epoch": 1.829984544049459, + "grad_norm": 0.6526342849663198, + "learning_rate": 6.397346283784915e-05, + "loss": 0.412, + "step": 1184 + }, + { + "epoch": 1.8315301391035548, + "grad_norm": 0.595465682627906, + "learning_rate": 6.393889256696231e-05, + "loss": 0.4133, + "step": 1185 + }, + { + "epoch": 1.8330757341576507, + "grad_norm": 0.4792886863134555, + "learning_rate": 6.390429441436915e-05, + "loss": 0.4013, + "step": 1186 + }, + { + "epoch": 1.8346213292117466, + "grad_norm": 0.35199181474614083, + "learning_rate": 6.38696684203663e-05, + "loss": 0.4032, + "step": 1187 + }, + { + "epoch": 1.8361669242658425, + "grad_norm": 0.2625679966109965, + "learning_rate": 6.383501462528273e-05, + "loss": 0.4114, + "step": 1188 + }, + { + "epoch": 1.8377125193199382, + "grad_norm": 0.20920678505260878, + "learning_rate": 6.380033306947985e-05, + "loss": 0.4133, + "step": 1189 + }, + { + "epoch": 1.8392581143740339, + "grad_norm": 0.25469585355967583, + "learning_rate": 6.376562379335136e-05, + "loss": 0.4138, + "step": 1190 + }, + { + "epoch": 1.8408037094281298, + "grad_norm": 0.3612273794583833, + "learning_rate": 6.373088683732329e-05, + "loss": 0.4182, + "step": 1191 + }, + { + "epoch": 1.8423493044822257, + "grad_norm": 0.39257267944161245, + "learning_rate": 6.369612224185387e-05, + "loss": 0.4093, + "step": 1192 + }, + { + "epoch": 1.8438948995363216, + "grad_norm": 0.42380522893852823, + "learning_rate": 6.366133004743354e-05, + "loss": 0.4129, + "step": 1193 + }, + { + "epoch": 1.8454404945904173, + "grad_norm": 0.373374444311683, + "learning_rate": 6.362651029458488e-05, + "loss": 0.4111, + "step": 1194 + }, + { + "epoch": 1.8469860896445132, + "grad_norm": 0.2804690905657256, + "learning_rate": 6.35916630238626e-05, + "loss": 0.4162, + "step": 1195 + }, + { + "epoch": 1.8485316846986088, + "grad_norm": 0.22179321091728624, + "learning_rate": 6.355678827585339e-05, + "loss": 0.4196, + "step": 1196 + }, + { + "epoch": 1.8500772797527048, + "grad_norm": 0.25619803053508716, + "learning_rate": 6.352188609117601e-05, + "loss": 0.4165, + "step": 1197 + }, + { + "epoch": 1.8516228748068007, + "grad_norm": 0.29265947672612824, + "learning_rate": 6.348695651048113e-05, + "loss": 0.412, + "step": 1198 + }, + { + "epoch": 1.8531684698608966, + "grad_norm": 0.3248585262131145, + "learning_rate": 6.345199957445134e-05, + "loss": 0.4044, + "step": 1199 + }, + { + "epoch": 1.8547140649149922, + "grad_norm": 0.33661124084441724, + "learning_rate": 6.341701532380111e-05, + "loss": 0.4213, + "step": 1200 + }, + { + "epoch": 1.8562596599690881, + "grad_norm": 0.37602417264309734, + "learning_rate": 6.338200379927673e-05, + "loss": 0.4123, + "step": 1201 + }, + { + "epoch": 1.8578052550231838, + "grad_norm": 0.4273071512644445, + "learning_rate": 6.33469650416562e-05, + "loss": 0.417, + "step": 1202 + }, + { + "epoch": 1.8593508500772797, + "grad_norm": 0.42079618075158465, + "learning_rate": 6.33118990917493e-05, + "loss": 0.4178, + "step": 1203 + }, + { + "epoch": 1.8608964451313756, + "grad_norm": 0.40759642036475224, + "learning_rate": 6.327680599039745e-05, + "loss": 0.4094, + "step": 1204 + }, + { + "epoch": 1.8624420401854715, + "grad_norm": 0.39745878772659676, + "learning_rate": 6.324168577847371e-05, + "loss": 0.4079, + "step": 1205 + }, + { + "epoch": 1.8639876352395672, + "grad_norm": 0.3415056835125822, + "learning_rate": 6.320653849688268e-05, + "loss": 0.4157, + "step": 1206 + }, + { + "epoch": 1.865533230293663, + "grad_norm": 0.24947747232526643, + "learning_rate": 6.317136418656054e-05, + "loss": 0.4028, + "step": 1207 + }, + { + "epoch": 1.8670788253477588, + "grad_norm": 0.3005682819226224, + "learning_rate": 6.31361628884749e-05, + "loss": 0.4221, + "step": 1208 + }, + { + "epoch": 1.8686244204018547, + "grad_norm": 0.40116056221954854, + "learning_rate": 6.310093464362485e-05, + "loss": 0.4091, + "step": 1209 + }, + { + "epoch": 1.8701700154559506, + "grad_norm": 0.4979343197776133, + "learning_rate": 6.306567949304085e-05, + "loss": 0.4143, + "step": 1210 + }, + { + "epoch": 1.8717156105100463, + "grad_norm": 0.5484171918207302, + "learning_rate": 6.303039747778464e-05, + "loss": 0.42, + "step": 1211 + }, + { + "epoch": 1.8732612055641422, + "grad_norm": 0.4410930297414067, + "learning_rate": 6.299508863894933e-05, + "loss": 0.4075, + "step": 1212 + }, + { + "epoch": 1.874806800618238, + "grad_norm": 0.4015790888559192, + "learning_rate": 6.295975301765925e-05, + "loss": 0.4144, + "step": 1213 + }, + { + "epoch": 1.8763523956723338, + "grad_norm": 0.4815391033114978, + "learning_rate": 6.29243906550699e-05, + "loss": 0.4105, + "step": 1214 + }, + { + "epoch": 1.8778979907264297, + "grad_norm": 0.5005456093567422, + "learning_rate": 6.288900159236793e-05, + "loss": 0.4103, + "step": 1215 + }, + { + "epoch": 1.8794435857805256, + "grad_norm": 0.5730406470111321, + "learning_rate": 6.285358587077112e-05, + "loss": 0.4181, + "step": 1216 + }, + { + "epoch": 1.8809891808346213, + "grad_norm": 0.5962056972704061, + "learning_rate": 6.281814353152825e-05, + "loss": 0.4178, + "step": 1217 + }, + { + "epoch": 1.8825347758887172, + "grad_norm": 0.42427410875375193, + "learning_rate": 6.278267461591915e-05, + "loss": 0.4128, + "step": 1218 + }, + { + "epoch": 1.8840803709428129, + "grad_norm": 0.3196237136374311, + "learning_rate": 6.274717916525457e-05, + "loss": 0.4141, + "step": 1219 + }, + { + "epoch": 1.8856259659969088, + "grad_norm": 0.3488472721495409, + "learning_rate": 6.271165722087621e-05, + "loss": 0.4155, + "step": 1220 + }, + { + "epoch": 1.8871715610510047, + "grad_norm": 0.46154481832376043, + "learning_rate": 6.267610882415654e-05, + "loss": 0.4129, + "step": 1221 + }, + { + "epoch": 1.8887171561051006, + "grad_norm": 0.5639974047085498, + "learning_rate": 6.264053401649892e-05, + "loss": 0.4075, + "step": 1222 + }, + { + "epoch": 1.8902627511591963, + "grad_norm": 0.5253893444511016, + "learning_rate": 6.260493283933744e-05, + "loss": 0.4147, + "step": 1223 + }, + { + "epoch": 1.891808346213292, + "grad_norm": 0.4951660291809643, + "learning_rate": 6.256930533413692e-05, + "loss": 0.4233, + "step": 1224 + }, + { + "epoch": 1.8933539412673879, + "grad_norm": 0.591961666560772, + "learning_rate": 6.253365154239282e-05, + "loss": 0.4232, + "step": 1225 + }, + { + "epoch": 1.8948995363214838, + "grad_norm": 0.5675281152594379, + "learning_rate": 6.249797150563124e-05, + "loss": 0.4129, + "step": 1226 + }, + { + "epoch": 1.8964451313755797, + "grad_norm": 0.5417143579552858, + "learning_rate": 6.246226526540881e-05, + "loss": 0.4273, + "step": 1227 + }, + { + "epoch": 1.8979907264296756, + "grad_norm": 0.5979033416426762, + "learning_rate": 6.242653286331273e-05, + "loss": 0.4311, + "step": 1228 + }, + { + "epoch": 1.8995363214837713, + "grad_norm": 0.5412056674221714, + "learning_rate": 6.239077434096063e-05, + "loss": 0.4107, + "step": 1229 + }, + { + "epoch": 1.901081916537867, + "grad_norm": 0.34544360856265593, + "learning_rate": 6.235498974000057e-05, + "loss": 0.4103, + "step": 1230 + }, + { + "epoch": 1.9026275115919629, + "grad_norm": 0.3093068702664324, + "learning_rate": 6.231917910211104e-05, + "loss": 0.4187, + "step": 1231 + }, + { + "epoch": 1.9041731066460588, + "grad_norm": 0.4606991022107261, + "learning_rate": 6.228334246900075e-05, + "loss": 0.4165, + "step": 1232 + }, + { + "epoch": 1.9057187017001547, + "grad_norm": 0.5144169510223493, + "learning_rate": 6.224747988240877e-05, + "loss": 0.4062, + "step": 1233 + }, + { + "epoch": 1.9072642967542504, + "grad_norm": 0.5291361309953686, + "learning_rate": 6.221159138410436e-05, + "loss": 0.4127, + "step": 1234 + }, + { + "epoch": 1.9088098918083463, + "grad_norm": 0.5486330265411696, + "learning_rate": 6.217567701588699e-05, + "loss": 0.41, + "step": 1235 + }, + { + "epoch": 1.910355486862442, + "grad_norm": 0.5042173545118099, + "learning_rate": 6.213973681958622e-05, + "loss": 0.4079, + "step": 1236 + }, + { + "epoch": 1.9119010819165378, + "grad_norm": 0.4301677403077317, + "learning_rate": 6.210377083706173e-05, + "loss": 0.4187, + "step": 1237 + }, + { + "epoch": 1.9134466769706338, + "grad_norm": 0.3221844288424395, + "learning_rate": 6.206777911020319e-05, + "loss": 0.417, + "step": 1238 + }, + { + "epoch": 1.9149922720247297, + "grad_norm": 0.35860210442670204, + "learning_rate": 6.203176168093033e-05, + "loss": 0.4071, + "step": 1239 + }, + { + "epoch": 1.9165378670788253, + "grad_norm": 0.455646522032718, + "learning_rate": 6.199571859119273e-05, + "loss": 0.413, + "step": 1240 + }, + { + "epoch": 1.918083462132921, + "grad_norm": 0.5114258237030155, + "learning_rate": 6.19596498829699e-05, + "loss": 0.4103, + "step": 1241 + }, + { + "epoch": 1.919629057187017, + "grad_norm": 0.5018901743826315, + "learning_rate": 6.192355559827119e-05, + "loss": 0.4138, + "step": 1242 + }, + { + "epoch": 1.9211746522411128, + "grad_norm": 0.3770510003359604, + "learning_rate": 6.188743577913573e-05, + "loss": 0.4088, + "step": 1243 + }, + { + "epoch": 1.9227202472952087, + "grad_norm": 0.3075133758449342, + "learning_rate": 6.185129046763238e-05, + "loss": 0.412, + "step": 1244 + }, + { + "epoch": 1.9242658423493046, + "grad_norm": 0.31888203899725603, + "learning_rate": 6.181511970585972e-05, + "loss": 0.4169, + "step": 1245 + }, + { + "epoch": 1.9258114374034003, + "grad_norm": 0.3680308830373179, + "learning_rate": 6.177892353594594e-05, + "loss": 0.4126, + "step": 1246 + }, + { + "epoch": 1.927357032457496, + "grad_norm": 0.3419862027786341, + "learning_rate": 6.174270200004885e-05, + "loss": 0.4106, + "step": 1247 + }, + { + "epoch": 1.928902627511592, + "grad_norm": 0.4074337933081065, + "learning_rate": 6.170645514035578e-05, + "loss": 0.4114, + "step": 1248 + }, + { + "epoch": 1.9304482225656878, + "grad_norm": 0.402449633126962, + "learning_rate": 6.167018299908358e-05, + "loss": 0.4231, + "step": 1249 + }, + { + "epoch": 1.9319938176197837, + "grad_norm": 0.45296020240879586, + "learning_rate": 6.163388561847849e-05, + "loss": 0.4144, + "step": 1250 + }, + { + "epoch": 1.9335394126738794, + "grad_norm": 0.3919980871130073, + "learning_rate": 6.159756304081625e-05, + "loss": 0.4154, + "step": 1251 + }, + { + "epoch": 1.9350850077279753, + "grad_norm": 0.3488830984928472, + "learning_rate": 6.156121530840183e-05, + "loss": 0.4104, + "step": 1252 + }, + { + "epoch": 1.936630602782071, + "grad_norm": 0.33868722953018976, + "learning_rate": 6.15248424635696e-05, + "loss": 0.4142, + "step": 1253 + }, + { + "epoch": 1.938176197836167, + "grad_norm": 0.2855567516301163, + "learning_rate": 6.14884445486831e-05, + "loss": 0.4172, + "step": 1254 + }, + { + "epoch": 1.9397217928902628, + "grad_norm": 0.2922986903670274, + "learning_rate": 6.145202160613509e-05, + "loss": 0.4105, + "step": 1255 + }, + { + "epoch": 1.9412673879443587, + "grad_norm": 0.29727295624694355, + "learning_rate": 6.141557367834752e-05, + "loss": 0.4135, + "step": 1256 + }, + { + "epoch": 1.9428129829984544, + "grad_norm": 0.33154582980383096, + "learning_rate": 6.13791008077714e-05, + "loss": 0.4168, + "step": 1257 + }, + { + "epoch": 1.94435857805255, + "grad_norm": 0.29921155830945473, + "learning_rate": 6.13426030368868e-05, + "loss": 0.401, + "step": 1258 + }, + { + "epoch": 1.945904173106646, + "grad_norm": 0.3000493708088317, + "learning_rate": 6.130608040820277e-05, + "loss": 0.421, + "step": 1259 + }, + { + "epoch": 1.947449768160742, + "grad_norm": 0.31902396993102805, + "learning_rate": 6.126953296425738e-05, + "loss": 0.408, + "step": 1260 + }, + { + "epoch": 1.9489953632148378, + "grad_norm": 0.33261146579060874, + "learning_rate": 6.123296074761753e-05, + "loss": 0.4095, + "step": 1261 + }, + { + "epoch": 1.9505409582689337, + "grad_norm": 0.36866099521520945, + "learning_rate": 6.119636380087899e-05, + "loss": 0.4178, + "step": 1262 + }, + { + "epoch": 1.9520865533230294, + "grad_norm": 0.341692553925293, + "learning_rate": 6.115974216666638e-05, + "loss": 0.4156, + "step": 1263 + }, + { + "epoch": 1.953632148377125, + "grad_norm": 0.32121154073975516, + "learning_rate": 6.1123095887633e-05, + "loss": 0.4094, + "step": 1264 + }, + { + "epoch": 1.955177743431221, + "grad_norm": 0.41034270168967524, + "learning_rate": 6.108642500646091e-05, + "loss": 0.4196, + "step": 1265 + }, + { + "epoch": 1.9567233384853169, + "grad_norm": 0.41868186295386195, + "learning_rate": 6.10497295658608e-05, + "loss": 0.407, + "step": 1266 + }, + { + "epoch": 1.9582689335394128, + "grad_norm": 0.3773719691008005, + "learning_rate": 6.1013009608571954e-05, + "loss": 0.4121, + "step": 1267 + }, + { + "epoch": 1.9598145285935085, + "grad_norm": 0.4011699259879193, + "learning_rate": 6.097626517736226e-05, + "loss": 0.412, + "step": 1268 + }, + { + "epoch": 1.9613601236476044, + "grad_norm": 0.43043830359425667, + "learning_rate": 6.0939496315028044e-05, + "loss": 0.4108, + "step": 1269 + }, + { + "epoch": 1.9629057187017, + "grad_norm": 0.4408474459144848, + "learning_rate": 6.090270306439415e-05, + "loss": 0.4158, + "step": 1270 + }, + { + "epoch": 1.964451313755796, + "grad_norm": 0.5261256456743819, + "learning_rate": 6.086588546831378e-05, + "loss": 0.4226, + "step": 1271 + }, + { + "epoch": 1.9659969088098919, + "grad_norm": 0.5996786938777356, + "learning_rate": 6.082904356966849e-05, + "loss": 0.4084, + "step": 1272 + }, + { + "epoch": 1.9675425038639878, + "grad_norm": 0.6343211777203427, + "learning_rate": 6.07921774113682e-05, + "loss": 0.4164, + "step": 1273 + }, + { + "epoch": 1.9690880989180835, + "grad_norm": 0.6506414593550145, + "learning_rate": 6.075528703635102e-05, + "loss": 0.4154, + "step": 1274 + }, + { + "epoch": 1.9706336939721791, + "grad_norm": 0.5323031068469932, + "learning_rate": 6.071837248758329e-05, + "loss": 0.4165, + "step": 1275 + }, + { + "epoch": 1.972179289026275, + "grad_norm": 0.4033522028561435, + "learning_rate": 6.06814338080595e-05, + "loss": 0.4153, + "step": 1276 + }, + { + "epoch": 1.973724884080371, + "grad_norm": 0.35960287534280405, + "learning_rate": 6.064447104080229e-05, + "loss": 0.4101, + "step": 1277 + }, + { + "epoch": 1.9752704791344669, + "grad_norm": 0.3645860122710083, + "learning_rate": 6.060748422886225e-05, + "loss": 0.411, + "step": 1278 + }, + { + "epoch": 1.9768160741885628, + "grad_norm": 0.3184631350197533, + "learning_rate": 6.05704734153181e-05, + "loss": 0.4145, + "step": 1279 + }, + { + "epoch": 1.9783616692426584, + "grad_norm": 0.26820918607939453, + "learning_rate": 6.053343864327643e-05, + "loss": 0.4105, + "step": 1280 + }, + { + "epoch": 1.9799072642967541, + "grad_norm": 0.2548959130980497, + "learning_rate": 6.0496379955871756e-05, + "loss": 0.4101, + "step": 1281 + }, + { + "epoch": 1.98145285935085, + "grad_norm": 0.3122132325464997, + "learning_rate": 6.045929739626646e-05, + "loss": 0.4194, + "step": 1282 + }, + { + "epoch": 1.982998454404946, + "grad_norm": 0.3760475329731319, + "learning_rate": 6.042219100765071e-05, + "loss": 0.408, + "step": 1283 + }, + { + "epoch": 1.9845440494590418, + "grad_norm": 0.39407874614851257, + "learning_rate": 6.0385060833242454e-05, + "loss": 0.4126, + "step": 1284 + }, + { + "epoch": 1.9860896445131375, + "grad_norm": 0.4158714884606941, + "learning_rate": 6.0347906916287325e-05, + "loss": 0.4102, + "step": 1285 + }, + { + "epoch": 1.9876352395672334, + "grad_norm": 0.39739963706387743, + "learning_rate": 6.0310729300058594e-05, + "loss": 0.4074, + "step": 1286 + }, + { + "epoch": 1.989180834621329, + "grad_norm": 0.40956072410450617, + "learning_rate": 6.0273528027857174e-05, + "loss": 0.4035, + "step": 1287 + }, + { + "epoch": 1.990726429675425, + "grad_norm": 0.4235509784543519, + "learning_rate": 6.02363031430115e-05, + "loss": 0.4063, + "step": 1288 + }, + { + "epoch": 1.992272024729521, + "grad_norm": 0.35569231358204384, + "learning_rate": 6.01990546888775e-05, + "loss": 0.4088, + "step": 1289 + }, + { + "epoch": 1.9938176197836168, + "grad_norm": 0.3277558087857754, + "learning_rate": 6.016178270883859e-05, + "loss": 0.4124, + "step": 1290 + }, + { + "epoch": 1.9953632148377125, + "grad_norm": 0.39938016792163783, + "learning_rate": 6.012448724630557e-05, + "loss": 0.4065, + "step": 1291 + }, + { + "epoch": 1.9969088098918082, + "grad_norm": 0.5074637792535728, + "learning_rate": 6.008716834471657e-05, + "loss": 0.4113, + "step": 1292 + }, + { + "epoch": 1.998454404945904, + "grad_norm": 0.5378561174080769, + "learning_rate": 6.004982604753705e-05, + "loss": 0.406, + "step": 1293 + }, + { + "epoch": 2.0, + "grad_norm": 0.5512949661008176, + "learning_rate": 6.001246039825969e-05, + "loss": 0.3932, + "step": 1294 + }, + { + "epoch": 2.001545595054096, + "grad_norm": 0.48164343529296594, + "learning_rate": 5.99750714404044e-05, + "loss": 0.3942, + "step": 1295 + }, + { + "epoch": 2.003091190108192, + "grad_norm": 0.4173541891289462, + "learning_rate": 5.993765921751822e-05, + "loss": 0.3933, + "step": 1296 + }, + { + "epoch": 2.0046367851622873, + "grad_norm": 0.44874736906998797, + "learning_rate": 5.990022377317527e-05, + "loss": 0.399, + "step": 1297 + }, + { + "epoch": 2.006182380216383, + "grad_norm": 0.48852803899547276, + "learning_rate": 5.986276515097676e-05, + "loss": 0.3895, + "step": 1298 + }, + { + "epoch": 2.007727975270479, + "grad_norm": 0.47210391121002704, + "learning_rate": 5.982528339455086e-05, + "loss": 0.3965, + "step": 1299 + }, + { + "epoch": 2.009273570324575, + "grad_norm": 0.4071856200785768, + "learning_rate": 5.978777854755268e-05, + "loss": 0.3877, + "step": 1300 + }, + { + "epoch": 2.010819165378671, + "grad_norm": 0.39801788539688204, + "learning_rate": 5.975025065366425e-05, + "loss": 0.3901, + "step": 1301 + }, + { + "epoch": 2.012364760432767, + "grad_norm": 0.44597876167965794, + "learning_rate": 5.971269975659444e-05, + "loss": 0.389, + "step": 1302 + }, + { + "epoch": 2.0139103554868623, + "grad_norm": 0.39663282709561176, + "learning_rate": 5.9675125900078876e-05, + "loss": 0.3886, + "step": 1303 + }, + { + "epoch": 2.015455950540958, + "grad_norm": 0.3271883215905907, + "learning_rate": 5.963752912787998e-05, + "loss": 0.3952, + "step": 1304 + }, + { + "epoch": 2.017001545595054, + "grad_norm": 0.3443570775972986, + "learning_rate": 5.959990948378682e-05, + "loss": 0.3916, + "step": 1305 + }, + { + "epoch": 2.01854714064915, + "grad_norm": 0.43047965620621526, + "learning_rate": 5.9562267011615106e-05, + "loss": 0.3958, + "step": 1306 + }, + { + "epoch": 2.020092735703246, + "grad_norm": 0.48135344420690257, + "learning_rate": 5.952460175520719e-05, + "loss": 0.403, + "step": 1307 + }, + { + "epoch": 2.021638330757342, + "grad_norm": 0.4731700325734638, + "learning_rate": 5.948691375843187e-05, + "loss": 0.39, + "step": 1308 + }, + { + "epoch": 2.0231839258114372, + "grad_norm": 0.3648430550938961, + "learning_rate": 5.94492030651845e-05, + "loss": 0.3952, + "step": 1309 + }, + { + "epoch": 2.024729520865533, + "grad_norm": 0.36491524034826517, + "learning_rate": 5.9411469719386865e-05, + "loss": 0.387, + "step": 1310 + }, + { + "epoch": 2.026275115919629, + "grad_norm": 0.4363293882642952, + "learning_rate": 5.937371376498707e-05, + "loss": 0.3953, + "step": 1311 + }, + { + "epoch": 2.027820710973725, + "grad_norm": 0.49219155345852056, + "learning_rate": 5.9335935245959657e-05, + "loss": 0.3904, + "step": 1312 + }, + { + "epoch": 2.029366306027821, + "grad_norm": 0.46819033388509657, + "learning_rate": 5.9298134206305344e-05, + "loss": 0.405, + "step": 1313 + }, + { + "epoch": 2.0309119010819163, + "grad_norm": 0.3731531854940534, + "learning_rate": 5.926031069005115e-05, + "loss": 0.398, + "step": 1314 + }, + { + "epoch": 2.0324574961360122, + "grad_norm": 0.34717209437762364, + "learning_rate": 5.9222464741250234e-05, + "loss": 0.3906, + "step": 1315 + }, + { + "epoch": 2.034003091190108, + "grad_norm": 0.32135536723818864, + "learning_rate": 5.9184596403981907e-05, + "loss": 0.3919, + "step": 1316 + }, + { + "epoch": 2.035548686244204, + "grad_norm": 0.3187246820508031, + "learning_rate": 5.9146705722351535e-05, + "loss": 0.387, + "step": 1317 + }, + { + "epoch": 2.0370942812983, + "grad_norm": 0.3844105515744366, + "learning_rate": 5.910879274049052e-05, + "loss": 0.3858, + "step": 1318 + }, + { + "epoch": 2.038639876352396, + "grad_norm": 0.3212345741912804, + "learning_rate": 5.9070857502556255e-05, + "loss": 0.3982, + "step": 1319 + }, + { + "epoch": 2.0401854714064913, + "grad_norm": 0.2837636853315771, + "learning_rate": 5.903290005273199e-05, + "loss": 0.3967, + "step": 1320 + }, + { + "epoch": 2.041731066460587, + "grad_norm": 0.2753993787315415, + "learning_rate": 5.899492043522695e-05, + "loss": 0.3999, + "step": 1321 + }, + { + "epoch": 2.043276661514683, + "grad_norm": 0.23306084091444113, + "learning_rate": 5.895691869427608e-05, + "loss": 0.3884, + "step": 1322 + }, + { + "epoch": 2.044822256568779, + "grad_norm": 0.2752161170160501, + "learning_rate": 5.8918894874140136e-05, + "loss": 0.3999, + "step": 1323 + }, + { + "epoch": 2.046367851622875, + "grad_norm": 0.2893023265733322, + "learning_rate": 5.88808490191056e-05, + "loss": 0.3912, + "step": 1324 + }, + { + "epoch": 2.047913446676971, + "grad_norm": 0.39089214093172914, + "learning_rate": 5.884278117348458e-05, + "loss": 0.3886, + "step": 1325 + }, + { + "epoch": 2.0494590417310663, + "grad_norm": 0.47685298138721893, + "learning_rate": 5.880469138161484e-05, + "loss": 0.4072, + "step": 1326 + }, + { + "epoch": 2.051004636785162, + "grad_norm": 0.36289657252757496, + "learning_rate": 5.876657968785969e-05, + "loss": 0.3961, + "step": 1327 + }, + { + "epoch": 2.052550231839258, + "grad_norm": 0.2614837486118298, + "learning_rate": 5.872844613660792e-05, + "loss": 0.3826, + "step": 1328 + }, + { + "epoch": 2.054095826893354, + "grad_norm": 0.23327559900974454, + "learning_rate": 5.8690290772273815e-05, + "loss": 0.395, + "step": 1329 + }, + { + "epoch": 2.05564142194745, + "grad_norm": 0.25665314076887247, + "learning_rate": 5.865211363929707e-05, + "loss": 0.3943, + "step": 1330 + }, + { + "epoch": 2.0571870170015454, + "grad_norm": 0.28734827929141427, + "learning_rate": 5.861391478214268e-05, + "loss": 0.387, + "step": 1331 + }, + { + "epoch": 2.0587326120556413, + "grad_norm": 0.2927672387638488, + "learning_rate": 5.8575694245301005e-05, + "loss": 0.3867, + "step": 1332 + }, + { + "epoch": 2.060278207109737, + "grad_norm": 0.28953425176851866, + "learning_rate": 5.853745207328762e-05, + "loss": 0.4011, + "step": 1333 + }, + { + "epoch": 2.061823802163833, + "grad_norm": 0.28360602802351886, + "learning_rate": 5.8499188310643293e-05, + "loss": 0.3866, + "step": 1334 + }, + { + "epoch": 2.063369397217929, + "grad_norm": 0.28169766603083646, + "learning_rate": 5.8460903001933996e-05, + "loss": 0.3889, + "step": 1335 + }, + { + "epoch": 2.064914992272025, + "grad_norm": 0.29187158559313897, + "learning_rate": 5.842259619175071e-05, + "loss": 0.3887, + "step": 1336 + }, + { + "epoch": 2.0664605873261204, + "grad_norm": 0.30502182980252696, + "learning_rate": 5.838426792470952e-05, + "loss": 0.3931, + "step": 1337 + }, + { + "epoch": 2.0680061823802163, + "grad_norm": 0.2797816265394829, + "learning_rate": 5.834591824545149e-05, + "loss": 0.3849, + "step": 1338 + }, + { + "epoch": 2.069551777434312, + "grad_norm": 0.23022841047299905, + "learning_rate": 5.830754719864259e-05, + "loss": 0.3943, + "step": 1339 + }, + { + "epoch": 2.071097372488408, + "grad_norm": 0.25594596278299064, + "learning_rate": 5.826915482897374e-05, + "loss": 0.3889, + "step": 1340 + }, + { + "epoch": 2.072642967542504, + "grad_norm": 0.29263312885541526, + "learning_rate": 5.8230741181160625e-05, + "loss": 0.394, + "step": 1341 + }, + { + "epoch": 2.0741885625966, + "grad_norm": 0.3132155920309119, + "learning_rate": 5.8192306299943764e-05, + "loss": 0.3876, + "step": 1342 + }, + { + "epoch": 2.0757341576506954, + "grad_norm": 0.37803460940583933, + "learning_rate": 5.815385023008839e-05, + "loss": 0.3967, + "step": 1343 + }, + { + "epoch": 2.0772797527047913, + "grad_norm": 0.39599324509331707, + "learning_rate": 5.8115373016384404e-05, + "loss": 0.3831, + "step": 1344 + }, + { + "epoch": 2.078825347758887, + "grad_norm": 0.36883394820473075, + "learning_rate": 5.807687470364635e-05, + "loss": 0.4026, + "step": 1345 + }, + { + "epoch": 2.080370942812983, + "grad_norm": 0.3163780061294479, + "learning_rate": 5.803835533671334e-05, + "loss": 0.3954, + "step": 1346 + }, + { + "epoch": 2.081916537867079, + "grad_norm": 0.2548836687691033, + "learning_rate": 5.7999814960449e-05, + "loss": 0.3931, + "step": 1347 + }, + { + "epoch": 2.0834621329211744, + "grad_norm": 0.3777335328739351, + "learning_rate": 5.7961253619741415e-05, + "loss": 0.3921, + "step": 1348 + }, + { + "epoch": 2.0850077279752703, + "grad_norm": 0.4696687530915752, + "learning_rate": 5.7922671359503145e-05, + "loss": 0.3995, + "step": 1349 + }, + { + "epoch": 2.0865533230293662, + "grad_norm": 0.4653528251375817, + "learning_rate": 5.788406822467105e-05, + "loss": 0.3991, + "step": 1350 + }, + { + "epoch": 2.088098918083462, + "grad_norm": 0.4146975919005292, + "learning_rate": 5.784544426020632e-05, + "loss": 0.3966, + "step": 1351 + }, + { + "epoch": 2.089644513137558, + "grad_norm": 0.3602980967706405, + "learning_rate": 5.780679951109442e-05, + "loss": 0.4004, + "step": 1352 + }, + { + "epoch": 2.091190108191654, + "grad_norm": 0.3812258112220021, + "learning_rate": 5.776813402234503e-05, + "loss": 0.3895, + "step": 1353 + }, + { + "epoch": 2.0927357032457494, + "grad_norm": 0.5348490661093533, + "learning_rate": 5.772944783899195e-05, + "loss": 0.3868, + "step": 1354 + }, + { + "epoch": 2.0942812982998453, + "grad_norm": 0.5341805710201996, + "learning_rate": 5.769074100609311e-05, + "loss": 0.4031, + "step": 1355 + }, + { + "epoch": 2.0958268933539412, + "grad_norm": 0.42718412388594196, + "learning_rate": 5.765201356873049e-05, + "loss": 0.3924, + "step": 1356 + }, + { + "epoch": 2.097372488408037, + "grad_norm": 0.33149597465277597, + "learning_rate": 5.761326557201006e-05, + "loss": 0.3913, + "step": 1357 + }, + { + "epoch": 2.098918083462133, + "grad_norm": 0.26099816268109227, + "learning_rate": 5.7574497061061745e-05, + "loss": 0.391, + "step": 1358 + }, + { + "epoch": 2.100463678516229, + "grad_norm": 0.32062129249672705, + "learning_rate": 5.753570808103934e-05, + "loss": 0.3899, + "step": 1359 + }, + { + "epoch": 2.1020092735703244, + "grad_norm": 0.48520151913932597, + "learning_rate": 5.7496898677120506e-05, + "loss": 0.3891, + "step": 1360 + }, + { + "epoch": 2.1035548686244203, + "grad_norm": 0.4735338135968625, + "learning_rate": 5.745806889450669e-05, + "loss": 0.3927, + "step": 1361 + }, + { + "epoch": 2.105100463678516, + "grad_norm": 0.47126817539463034, + "learning_rate": 5.7419218778423045e-05, + "loss": 0.3799, + "step": 1362 + }, + { + "epoch": 2.106646058732612, + "grad_norm": 0.4650326573097983, + "learning_rate": 5.738034837411845e-05, + "loss": 0.4049, + "step": 1363 + }, + { + "epoch": 2.108191653786708, + "grad_norm": 0.5052600261623205, + "learning_rate": 5.734145772686539e-05, + "loss": 0.3907, + "step": 1364 + }, + { + "epoch": 2.109737248840804, + "grad_norm": 0.4492378071242691, + "learning_rate": 5.7302546881959905e-05, + "loss": 0.3929, + "step": 1365 + }, + { + "epoch": 2.1112828438948994, + "grad_norm": 0.4022561190795431, + "learning_rate": 5.726361588472161e-05, + "loss": 0.3957, + "step": 1366 + }, + { + "epoch": 2.1128284389489953, + "grad_norm": 0.3232004812396111, + "learning_rate": 5.722466478049356e-05, + "loss": 0.3951, + "step": 1367 + }, + { + "epoch": 2.114374034003091, + "grad_norm": 0.29120789738606623, + "learning_rate": 5.718569361464221e-05, + "loss": 0.3973, + "step": 1368 + }, + { + "epoch": 2.115919629057187, + "grad_norm": 0.3297836661004876, + "learning_rate": 5.714670243255745e-05, + "loss": 0.4035, + "step": 1369 + }, + { + "epoch": 2.117465224111283, + "grad_norm": 0.33666164737551685, + "learning_rate": 5.710769127965238e-05, + "loss": 0.4048, + "step": 1370 + }, + { + "epoch": 2.1190108191653785, + "grad_norm": 0.32318941343611973, + "learning_rate": 5.7068660201363464e-05, + "loss": 0.3886, + "step": 1371 + }, + { + "epoch": 2.1205564142194744, + "grad_norm": 0.31605456636520374, + "learning_rate": 5.7029609243150306e-05, + "loss": 0.3986, + "step": 1372 + }, + { + "epoch": 2.1221020092735703, + "grad_norm": 0.3444017106165271, + "learning_rate": 5.699053845049568e-05, + "loss": 0.3974, + "step": 1373 + }, + { + "epoch": 2.123647604327666, + "grad_norm": 0.32076054536360127, + "learning_rate": 5.695144786890547e-05, + "loss": 0.3941, + "step": 1374 + }, + { + "epoch": 2.125193199381762, + "grad_norm": 0.3375559972286297, + "learning_rate": 5.6912337543908616e-05, + "loss": 0.4049, + "step": 1375 + }, + { + "epoch": 2.126738794435858, + "grad_norm": 0.3398886977192635, + "learning_rate": 5.687320752105702e-05, + "loss": 0.4028, + "step": 1376 + }, + { + "epoch": 2.1282843894899535, + "grad_norm": 0.28589222478831905, + "learning_rate": 5.683405784592557e-05, + "loss": 0.4076, + "step": 1377 + }, + { + "epoch": 2.1298299845440494, + "grad_norm": 0.2991766255827722, + "learning_rate": 5.679488856411201e-05, + "loss": 0.3986, + "step": 1378 + }, + { + "epoch": 2.1313755795981453, + "grad_norm": 0.2730775297870511, + "learning_rate": 5.6755699721236927e-05, + "loss": 0.3962, + "step": 1379 + }, + { + "epoch": 2.132921174652241, + "grad_norm": 0.2279078127037609, + "learning_rate": 5.67164913629437e-05, + "loss": 0.3951, + "step": 1380 + }, + { + "epoch": 2.134466769706337, + "grad_norm": 0.26190056792962724, + "learning_rate": 5.6677263534898425e-05, + "loss": 0.3918, + "step": 1381 + }, + { + "epoch": 2.1360123647604325, + "grad_norm": 0.30977615037207307, + "learning_rate": 5.663801628278988e-05, + "loss": 0.388, + "step": 1382 + }, + { + "epoch": 2.1375579598145285, + "grad_norm": 0.2902462039984759, + "learning_rate": 5.659874965232948e-05, + "loss": 0.3905, + "step": 1383 + }, + { + "epoch": 2.1391035548686244, + "grad_norm": 0.27677467210321927, + "learning_rate": 5.655946368925118e-05, + "loss": 0.3923, + "step": 1384 + }, + { + "epoch": 2.1406491499227203, + "grad_norm": 0.24262357446186741, + "learning_rate": 5.652015843931149e-05, + "loss": 0.3868, + "step": 1385 + }, + { + "epoch": 2.142194744976816, + "grad_norm": 0.21439927480680235, + "learning_rate": 5.648083394828935e-05, + "loss": 0.396, + "step": 1386 + }, + { + "epoch": 2.143740340030912, + "grad_norm": 0.3024814027973054, + "learning_rate": 5.644149026198612e-05, + "loss": 0.3937, + "step": 1387 + }, + { + "epoch": 2.1452859350850075, + "grad_norm": 0.33105463220970677, + "learning_rate": 5.640212742622553e-05, + "loss": 0.4014, + "step": 1388 + }, + { + "epoch": 2.1468315301391034, + "grad_norm": 0.2393384112444071, + "learning_rate": 5.636274548685361e-05, + "loss": 0.3896, + "step": 1389 + }, + { + "epoch": 2.1483771251931993, + "grad_norm": 0.2586734232627636, + "learning_rate": 5.63233444897386e-05, + "loss": 0.3975, + "step": 1390 + }, + { + "epoch": 2.1499227202472952, + "grad_norm": 0.2303543368209657, + "learning_rate": 5.628392448077101e-05, + "loss": 0.3936, + "step": 1391 + }, + { + "epoch": 2.151468315301391, + "grad_norm": 0.2293507554950428, + "learning_rate": 5.624448550586342e-05, + "loss": 0.396, + "step": 1392 + }, + { + "epoch": 2.153013910355487, + "grad_norm": 0.2548338874339664, + "learning_rate": 5.620502761095056e-05, + "loss": 0.397, + "step": 1393 + }, + { + "epoch": 2.1545595054095825, + "grad_norm": 0.22722623652559307, + "learning_rate": 5.616555084198915e-05, + "loss": 0.3955, + "step": 1394 + }, + { + "epoch": 2.1561051004636784, + "grad_norm": 0.1918331492110652, + "learning_rate": 5.6126055244957925e-05, + "loss": 0.3882, + "step": 1395 + }, + { + "epoch": 2.1576506955177743, + "grad_norm": 0.2166050594320556, + "learning_rate": 5.608654086585752e-05, + "loss": 0.4004, + "step": 1396 + }, + { + "epoch": 2.1591962905718702, + "grad_norm": 0.2809903221060055, + "learning_rate": 5.604700775071048e-05, + "loss": 0.4012, + "step": 1397 + }, + { + "epoch": 2.160741885625966, + "grad_norm": 0.36215355227293844, + "learning_rate": 5.600745594556113e-05, + "loss": 0.3953, + "step": 1398 + }, + { + "epoch": 2.162287480680062, + "grad_norm": 0.401216351650045, + "learning_rate": 5.596788549647562e-05, + "loss": 0.3986, + "step": 1399 + }, + { + "epoch": 2.1638330757341575, + "grad_norm": 0.36923309321816083, + "learning_rate": 5.592829644954175e-05, + "loss": 0.3896, + "step": 1400 + }, + { + "epoch": 2.1653786707882534, + "grad_norm": 0.28053308499077567, + "learning_rate": 5.588868885086902e-05, + "loss": 0.3945, + "step": 1401 + }, + { + "epoch": 2.1669242658423493, + "grad_norm": 0.2347720773966801, + "learning_rate": 5.584906274658854e-05, + "loss": 0.3896, + "step": 1402 + }, + { + "epoch": 2.1684698608964452, + "grad_norm": 0.2996575204398839, + "learning_rate": 5.580941818285295e-05, + "loss": 0.3956, + "step": 1403 + }, + { + "epoch": 2.170015455950541, + "grad_norm": 0.24214511700057684, + "learning_rate": 5.576975520583642e-05, + "loss": 0.3948, + "step": 1404 + }, + { + "epoch": 2.1715610510046366, + "grad_norm": 0.20950657680483423, + "learning_rate": 5.573007386173452e-05, + "loss": 0.3925, + "step": 1405 + }, + { + "epoch": 2.1731066460587325, + "grad_norm": 0.26792625106666884, + "learning_rate": 5.569037419676427e-05, + "loss": 0.3992, + "step": 1406 + }, + { + "epoch": 2.1746522411128284, + "grad_norm": 0.2712494902998535, + "learning_rate": 5.5650656257163985e-05, + "loss": 0.3859, + "step": 1407 + }, + { + "epoch": 2.1761978361669243, + "grad_norm": 0.2930171260012353, + "learning_rate": 5.561092008919328e-05, + "loss": 0.3958, + "step": 1408 + }, + { + "epoch": 2.17774343122102, + "grad_norm": 0.304917573510114, + "learning_rate": 5.557116573913301e-05, + "loss": 0.3848, + "step": 1409 + }, + { + "epoch": 2.179289026275116, + "grad_norm": 0.2414499333351336, + "learning_rate": 5.553139325328518e-05, + "loss": 0.4, + "step": 1410 + }, + { + "epoch": 2.1808346213292116, + "grad_norm": 0.34294961997680773, + "learning_rate": 5.5491602677972945e-05, + "loss": 0.3981, + "step": 1411 + }, + { + "epoch": 2.1823802163833075, + "grad_norm": 0.37602037518813536, + "learning_rate": 5.545179405954051e-05, + "loss": 0.3921, + "step": 1412 + }, + { + "epoch": 2.1839258114374034, + "grad_norm": 0.25752242189578534, + "learning_rate": 5.5411967444353105e-05, + "loss": 0.3996, + "step": 1413 + }, + { + "epoch": 2.1854714064914993, + "grad_norm": 0.31666571784081177, + "learning_rate": 5.537212287879693e-05, + "loss": 0.3922, + "step": 1414 + }, + { + "epoch": 2.187017001545595, + "grad_norm": 0.3736689968910478, + "learning_rate": 5.533226040927906e-05, + "loss": 0.3975, + "step": 1415 + }, + { + "epoch": 2.1885625965996907, + "grad_norm": 0.2916315544781542, + "learning_rate": 5.5292380082227476e-05, + "loss": 0.3898, + "step": 1416 + }, + { + "epoch": 2.1901081916537866, + "grad_norm": 0.23508392644454182, + "learning_rate": 5.5252481944090876e-05, + "loss": 0.3905, + "step": 1417 + }, + { + "epoch": 2.1916537867078825, + "grad_norm": 0.2089635516135591, + "learning_rate": 5.521256604133879e-05, + "loss": 0.396, + "step": 1418 + }, + { + "epoch": 2.1931993817619784, + "grad_norm": 0.283463538403399, + "learning_rate": 5.51726324204614e-05, + "loss": 0.3932, + "step": 1419 + }, + { + "epoch": 2.1947449768160743, + "grad_norm": 0.20312496600238453, + "learning_rate": 5.513268112796951e-05, + "loss": 0.3871, + "step": 1420 + }, + { + "epoch": 2.19629057187017, + "grad_norm": 0.23945133879207264, + "learning_rate": 5.509271221039452e-05, + "loss": 0.4051, + "step": 1421 + }, + { + "epoch": 2.1978361669242656, + "grad_norm": 0.26650563107564185, + "learning_rate": 5.505272571428837e-05, + "loss": 0.3929, + "step": 1422 + }, + { + "epoch": 2.1993817619783615, + "grad_norm": 0.29965664042142814, + "learning_rate": 5.501272168622345e-05, + "loss": 0.4062, + "step": 1423 + }, + { + "epoch": 2.2009273570324575, + "grad_norm": 0.2829338590370037, + "learning_rate": 5.497270017279259e-05, + "loss": 0.3996, + "step": 1424 + }, + { + "epoch": 2.2024729520865534, + "grad_norm": 0.29676225707278125, + "learning_rate": 5.493266122060897e-05, + "loss": 0.3936, + "step": 1425 + }, + { + "epoch": 2.2040185471406493, + "grad_norm": 0.35621600063358044, + "learning_rate": 5.48926048763061e-05, + "loss": 0.3995, + "step": 1426 + }, + { + "epoch": 2.205564142194745, + "grad_norm": 0.382789454234158, + "learning_rate": 5.485253118653774e-05, + "loss": 0.3828, + "step": 1427 + }, + { + "epoch": 2.2071097372488406, + "grad_norm": 0.32438131697044115, + "learning_rate": 5.481244019797782e-05, + "loss": 0.3964, + "step": 1428 + }, + { + "epoch": 2.2086553323029365, + "grad_norm": 0.2522710585082952, + "learning_rate": 5.477233195732046e-05, + "loss": 0.3998, + "step": 1429 + }, + { + "epoch": 2.2102009273570324, + "grad_norm": 0.22076782465890474, + "learning_rate": 5.473220651127988e-05, + "loss": 0.3961, + "step": 1430 + }, + { + "epoch": 2.2117465224111283, + "grad_norm": 0.24445337462545094, + "learning_rate": 5.469206390659028e-05, + "loss": 0.3788, + "step": 1431 + }, + { + "epoch": 2.2132921174652243, + "grad_norm": 0.23976966421294135, + "learning_rate": 5.46519041900059e-05, + "loss": 0.3933, + "step": 1432 + }, + { + "epoch": 2.21483771251932, + "grad_norm": 0.2508674942889716, + "learning_rate": 5.461172740830091e-05, + "loss": 0.3875, + "step": 1433 + }, + { + "epoch": 2.2163833075734156, + "grad_norm": 0.2786405104063059, + "learning_rate": 5.45715336082693e-05, + "loss": 0.3798, + "step": 1434 + }, + { + "epoch": 2.2179289026275115, + "grad_norm": 0.3579458488758703, + "learning_rate": 5.453132283672496e-05, + "loss": 0.4041, + "step": 1435 + }, + { + "epoch": 2.2194744976816074, + "grad_norm": 0.3935960503455614, + "learning_rate": 5.4491095140501486e-05, + "loss": 0.3933, + "step": 1436 + }, + { + "epoch": 2.2210200927357033, + "grad_norm": 0.3697590188908626, + "learning_rate": 5.4450850566452213e-05, + "loss": 0.3921, + "step": 1437 + }, + { + "epoch": 2.2225656877897992, + "grad_norm": 0.3605308818179636, + "learning_rate": 5.441058916145011e-05, + "loss": 0.3999, + "step": 1438 + }, + { + "epoch": 2.2241112828438947, + "grad_norm": 0.3356567615719234, + "learning_rate": 5.43703109723878e-05, + "loss": 0.3909, + "step": 1439 + }, + { + "epoch": 2.2256568778979906, + "grad_norm": 0.30531472053735503, + "learning_rate": 5.43300160461774e-05, + "loss": 0.3928, + "step": 1440 + }, + { + "epoch": 2.2272024729520865, + "grad_norm": 0.2565363158264292, + "learning_rate": 5.428970442975053e-05, + "loss": 0.3945, + "step": 1441 + }, + { + "epoch": 2.2287480680061824, + "grad_norm": 0.2380223524712181, + "learning_rate": 5.424937617005829e-05, + "loss": 0.3793, + "step": 1442 + }, + { + "epoch": 2.2302936630602783, + "grad_norm": 0.2560952406565104, + "learning_rate": 5.4209031314071116e-05, + "loss": 0.3911, + "step": 1443 + }, + { + "epoch": 2.2318392581143742, + "grad_norm": 0.24313728437682017, + "learning_rate": 5.4168669908778814e-05, + "loss": 0.4018, + "step": 1444 + }, + { + "epoch": 2.2333848531684697, + "grad_norm": 0.21664727345403825, + "learning_rate": 5.412829200119041e-05, + "loss": 0.3902, + "step": 1445 + }, + { + "epoch": 2.2349304482225656, + "grad_norm": 0.2675734440025409, + "learning_rate": 5.408789763833421e-05, + "loss": 0.3898, + "step": 1446 + }, + { + "epoch": 2.2364760432766615, + "grad_norm": 0.2923302849334493, + "learning_rate": 5.4047486867257664e-05, + "loss": 0.3893, + "step": 1447 + }, + { + "epoch": 2.2380216383307574, + "grad_norm": 0.26972423411484947, + "learning_rate": 5.4007059735027315e-05, + "loss": 0.397, + "step": 1448 + }, + { + "epoch": 2.2395672333848533, + "grad_norm": 0.304336195544412, + "learning_rate": 5.3966616288728774e-05, + "loss": 0.4051, + "step": 1449 + }, + { + "epoch": 2.2411128284389488, + "grad_norm": 0.28943902807642036, + "learning_rate": 5.3926156575466675e-05, + "loss": 0.3941, + "step": 1450 + }, + { + "epoch": 2.2426584234930447, + "grad_norm": 0.28694545302948715, + "learning_rate": 5.388568064236457e-05, + "loss": 0.3915, + "step": 1451 + }, + { + "epoch": 2.2442040185471406, + "grad_norm": 0.33675761551956246, + "learning_rate": 5.3845188536564884e-05, + "loss": 0.3978, + "step": 1452 + }, + { + "epoch": 2.2457496136012365, + "grad_norm": 0.38081008970903646, + "learning_rate": 5.3804680305228934e-05, + "loss": 0.3881, + "step": 1453 + }, + { + "epoch": 2.2472952086553324, + "grad_norm": 0.34809364873567517, + "learning_rate": 5.3764155995536796e-05, + "loss": 0.396, + "step": 1454 + }, + { + "epoch": 2.2488408037094283, + "grad_norm": 0.3633482183310246, + "learning_rate": 5.3723615654687224e-05, + "loss": 0.3879, + "step": 1455 + }, + { + "epoch": 2.250386398763524, + "grad_norm": 0.36431906421062726, + "learning_rate": 5.368305932989773e-05, + "loss": 0.3941, + "step": 1456 + }, + { + "epoch": 2.2519319938176197, + "grad_norm": 0.4049586685164246, + "learning_rate": 5.364248706840437e-05, + "loss": 0.3918, + "step": 1457 + }, + { + "epoch": 2.2534775888717156, + "grad_norm": 0.39800391551422143, + "learning_rate": 5.3601898917461826e-05, + "loss": 0.3888, + "step": 1458 + }, + { + "epoch": 2.2550231839258115, + "grad_norm": 0.31734587330195074, + "learning_rate": 5.35612949243432e-05, + "loss": 0.3917, + "step": 1459 + }, + { + "epoch": 2.2565687789799074, + "grad_norm": 0.2570755668387734, + "learning_rate": 5.352067513634011e-05, + "loss": 0.391, + "step": 1460 + }, + { + "epoch": 2.2581143740340033, + "grad_norm": 0.24640982169832737, + "learning_rate": 5.3480039600762586e-05, + "loss": 0.3946, + "step": 1461 + }, + { + "epoch": 2.2596599690880987, + "grad_norm": 0.23778783202732132, + "learning_rate": 5.343938836493894e-05, + "loss": 0.3869, + "step": 1462 + }, + { + "epoch": 2.2612055641421946, + "grad_norm": 0.2653726286543613, + "learning_rate": 5.339872147621579e-05, + "loss": 0.3963, + "step": 1463 + }, + { + "epoch": 2.2627511591962906, + "grad_norm": 0.23187671980103855, + "learning_rate": 5.335803898195803e-05, + "loss": 0.3837, + "step": 1464 + }, + { + "epoch": 2.2642967542503865, + "grad_norm": 0.2480098286152096, + "learning_rate": 5.3317340929548676e-05, + "loss": 0.3948, + "step": 1465 + }, + { + "epoch": 2.2658423493044824, + "grad_norm": 0.2768215559146363, + "learning_rate": 5.327662736638886e-05, + "loss": 0.3904, + "step": 1466 + }, + { + "epoch": 2.2673879443585783, + "grad_norm": 0.3541058579001065, + "learning_rate": 5.3235898339897846e-05, + "loss": 0.3868, + "step": 1467 + }, + { + "epoch": 2.2689335394126737, + "grad_norm": 0.3455770642766049, + "learning_rate": 5.319515389751283e-05, + "loss": 0.3956, + "step": 1468 + }, + { + "epoch": 2.2704791344667696, + "grad_norm": 0.24872912183460227, + "learning_rate": 5.3154394086688995e-05, + "loss": 0.3934, + "step": 1469 + }, + { + "epoch": 2.2720247295208655, + "grad_norm": 0.22593842204717707, + "learning_rate": 5.311361895489946e-05, + "loss": 0.4029, + "step": 1470 + }, + { + "epoch": 2.2735703245749614, + "grad_norm": 0.18875541559897221, + "learning_rate": 5.3072828549635107e-05, + "loss": 0.3868, + "step": 1471 + }, + { + "epoch": 2.2751159196290573, + "grad_norm": 0.20398086887059988, + "learning_rate": 5.303202291840471e-05, + "loss": 0.3923, + "step": 1472 + }, + { + "epoch": 2.276661514683153, + "grad_norm": 0.23263084748729548, + "learning_rate": 5.299120210873467e-05, + "loss": 0.3837, + "step": 1473 + }, + { + "epoch": 2.2782071097372487, + "grad_norm": 0.21563000272531446, + "learning_rate": 5.295036616816914e-05, + "loss": 0.3931, + "step": 1474 + }, + { + "epoch": 2.2797527047913446, + "grad_norm": 0.22171142965832003, + "learning_rate": 5.290951514426988e-05, + "loss": 0.3956, + "step": 1475 + }, + { + "epoch": 2.2812982998454405, + "grad_norm": 0.22587717378598995, + "learning_rate": 5.286864908461619e-05, + "loss": 0.3872, + "step": 1476 + }, + { + "epoch": 2.2828438948995364, + "grad_norm": 0.1971272547786867, + "learning_rate": 5.282776803680492e-05, + "loss": 0.4036, + "step": 1477 + }, + { + "epoch": 2.2843894899536323, + "grad_norm": 0.2224845402260878, + "learning_rate": 5.278687204845035e-05, + "loss": 0.3897, + "step": 1478 + }, + { + "epoch": 2.285935085007728, + "grad_norm": 0.19441003764474904, + "learning_rate": 5.2745961167184185e-05, + "loss": 0.3889, + "step": 1479 + }, + { + "epoch": 2.2874806800618237, + "grad_norm": 0.18171125267942248, + "learning_rate": 5.270503544065544e-05, + "loss": 0.402, + "step": 1480 + }, + { + "epoch": 2.2890262751159196, + "grad_norm": 0.17983694611736392, + "learning_rate": 5.266409491653046e-05, + "loss": 0.3854, + "step": 1481 + }, + { + "epoch": 2.2905718701700155, + "grad_norm": 0.20254679567986317, + "learning_rate": 5.262313964249281e-05, + "loss": 0.3988, + "step": 1482 + }, + { + "epoch": 2.2921174652241114, + "grad_norm": 0.19184410298081583, + "learning_rate": 5.258216966624321e-05, + "loss": 0.3851, + "step": 1483 + }, + { + "epoch": 2.293663060278207, + "grad_norm": 0.254767017913148, + "learning_rate": 5.254118503549954e-05, + "loss": 0.3928, + "step": 1484 + }, + { + "epoch": 2.295208655332303, + "grad_norm": 0.31469261555728983, + "learning_rate": 5.2500185797996734e-05, + "loss": 0.39, + "step": 1485 + }, + { + "epoch": 2.2967542503863987, + "grad_norm": 0.28076586148567945, + "learning_rate": 5.245917200148675e-05, + "loss": 0.394, + "step": 1486 + }, + { + "epoch": 2.2982998454404946, + "grad_norm": 0.25246618561021705, + "learning_rate": 5.241814369373845e-05, + "loss": 0.3939, + "step": 1487 + }, + { + "epoch": 2.2998454404945905, + "grad_norm": 0.22736004566702298, + "learning_rate": 5.237710092253768e-05, + "loss": 0.3891, + "step": 1488 + }, + { + "epoch": 2.3013910355486864, + "grad_norm": 0.25674536214096144, + "learning_rate": 5.233604373568707e-05, + "loss": 0.3954, + "step": 1489 + }, + { + "epoch": 2.3029366306027823, + "grad_norm": 0.24910563016368026, + "learning_rate": 5.229497218100606e-05, + "loss": 0.3961, + "step": 1490 + }, + { + "epoch": 2.3044822256568778, + "grad_norm": 0.25669220171173646, + "learning_rate": 5.225388630633083e-05, + "loss": 0.4008, + "step": 1491 + }, + { + "epoch": 2.3060278207109737, + "grad_norm": 0.21963865673874589, + "learning_rate": 5.221278615951422e-05, + "loss": 0.3926, + "step": 1492 + }, + { + "epoch": 2.3075734157650696, + "grad_norm": 0.2736157715412973, + "learning_rate": 5.217167178842573e-05, + "loss": 0.396, + "step": 1493 + }, + { + "epoch": 2.3091190108191655, + "grad_norm": 0.38381019611426875, + "learning_rate": 5.2130543240951366e-05, + "loss": 0.3904, + "step": 1494 + }, + { + "epoch": 2.3106646058732614, + "grad_norm": 0.3546302751151582, + "learning_rate": 5.208940056499371e-05, + "loss": 0.3997, + "step": 1495 + }, + { + "epoch": 2.312210200927357, + "grad_norm": 0.29651312604337876, + "learning_rate": 5.204824380847177e-05, + "loss": 0.393, + "step": 1496 + }, + { + "epoch": 2.3137557959814528, + "grad_norm": 0.25451612140716673, + "learning_rate": 5.2007073019320935e-05, + "loss": 0.3971, + "step": 1497 + }, + { + "epoch": 2.3153013910355487, + "grad_norm": 0.29775277849373727, + "learning_rate": 5.196588824549298e-05, + "loss": 0.395, + "step": 1498 + }, + { + "epoch": 2.3168469860896446, + "grad_norm": 0.30367259317781947, + "learning_rate": 5.192468953495592e-05, + "loss": 0.3982, + "step": 1499 + }, + { + "epoch": 2.3183925811437405, + "grad_norm": 0.34598867891208274, + "learning_rate": 5.188347693569407e-05, + "loss": 0.395, + "step": 1500 + }, + { + "epoch": 2.3199381761978364, + "grad_norm": 0.37348045757641857, + "learning_rate": 5.184225049570783e-05, + "loss": 0.3961, + "step": 1501 + }, + { + "epoch": 2.321483771251932, + "grad_norm": 0.3362956096767625, + "learning_rate": 5.180101026301379e-05, + "loss": 0.3966, + "step": 1502 + }, + { + "epoch": 2.3230293663060277, + "grad_norm": 0.31799218245762445, + "learning_rate": 5.17597562856446e-05, + "loss": 0.3978, + "step": 1503 + }, + { + "epoch": 2.3245749613601236, + "grad_norm": 0.3464467388056877, + "learning_rate": 5.171848861164887e-05, + "loss": 0.3948, + "step": 1504 + }, + { + "epoch": 2.3261205564142196, + "grad_norm": 0.41609624207931256, + "learning_rate": 5.167720728909121e-05, + "loss": 0.3844, + "step": 1505 + }, + { + "epoch": 2.3276661514683155, + "grad_norm": 0.4745365547147811, + "learning_rate": 5.163591236605212e-05, + "loss": 0.3882, + "step": 1506 + }, + { + "epoch": 2.329211746522411, + "grad_norm": 0.4825019829743042, + "learning_rate": 5.1594603890627934e-05, + "loss": 0.3853, + "step": 1507 + }, + { + "epoch": 2.330757341576507, + "grad_norm": 0.38029637854706505, + "learning_rate": 5.155328191093074e-05, + "loss": 0.3867, + "step": 1508 + }, + { + "epoch": 2.3323029366306027, + "grad_norm": 0.3240935750344444, + "learning_rate": 5.151194647508841e-05, + "loss": 0.3859, + "step": 1509 + }, + { + "epoch": 2.3338485316846986, + "grad_norm": 0.3248722572663255, + "learning_rate": 5.147059763124445e-05, + "loss": 0.3928, + "step": 1510 + }, + { + "epoch": 2.3353941267387945, + "grad_norm": 0.4688604609285743, + "learning_rate": 5.1429235427557994e-05, + "loss": 0.3982, + "step": 1511 + }, + { + "epoch": 2.3369397217928904, + "grad_norm": 0.42273469107871264, + "learning_rate": 5.1387859912203734e-05, + "loss": 0.3942, + "step": 1512 + }, + { + "epoch": 2.338485316846986, + "grad_norm": 0.3433548338511978, + "learning_rate": 5.134647113337187e-05, + "loss": 0.3836, + "step": 1513 + }, + { + "epoch": 2.340030911901082, + "grad_norm": 0.2851349484514058, + "learning_rate": 5.130506913926806e-05, + "loss": 0.3954, + "step": 1514 + }, + { + "epoch": 2.3415765069551777, + "grad_norm": 0.4135287612524575, + "learning_rate": 5.1263653978113315e-05, + "loss": 0.3851, + "step": 1515 + }, + { + "epoch": 2.3431221020092736, + "grad_norm": 0.5691535220332999, + "learning_rate": 5.1222225698144024e-05, + "loss": 0.3911, + "step": 1516 + }, + { + "epoch": 2.3446676970633695, + "grad_norm": 0.5566667027836483, + "learning_rate": 5.118078434761185e-05, + "loss": 0.398, + "step": 1517 + }, + { + "epoch": 2.346213292117465, + "grad_norm": 0.405028370843713, + "learning_rate": 5.113932997478364e-05, + "loss": 0.3893, + "step": 1518 + }, + { + "epoch": 2.347758887171561, + "grad_norm": 0.231212331290742, + "learning_rate": 5.1097862627941465e-05, + "loss": 0.3956, + "step": 1519 + }, + { + "epoch": 2.349304482225657, + "grad_norm": 0.2972009251282469, + "learning_rate": 5.105638235538248e-05, + "loss": 0.388, + "step": 1520 + }, + { + "epoch": 2.3508500772797527, + "grad_norm": 0.39737209705772275, + "learning_rate": 5.1014889205418875e-05, + "loss": 0.3994, + "step": 1521 + }, + { + "epoch": 2.3523956723338486, + "grad_norm": 0.3992741406688027, + "learning_rate": 5.097338322637786e-05, + "loss": 0.39, + "step": 1522 + }, + { + "epoch": 2.3539412673879445, + "grad_norm": 0.3236551759909285, + "learning_rate": 5.093186446660159e-05, + "loss": 0.3868, + "step": 1523 + }, + { + "epoch": 2.3554868624420404, + "grad_norm": 0.25157713529976583, + "learning_rate": 5.0890332974447117e-05, + "loss": 0.3962, + "step": 1524 + }, + { + "epoch": 2.357032457496136, + "grad_norm": 0.27996376356540936, + "learning_rate": 5.084878879828627e-05, + "loss": 0.3941, + "step": 1525 + }, + { + "epoch": 2.358578052550232, + "grad_norm": 0.2992424334167224, + "learning_rate": 5.080723198650571e-05, + "loss": 0.3969, + "step": 1526 + }, + { + "epoch": 2.3601236476043277, + "grad_norm": 0.30496741070379774, + "learning_rate": 5.07656625875068e-05, + "loss": 0.3882, + "step": 1527 + }, + { + "epoch": 2.3616692426584236, + "grad_norm": 0.3024436138778408, + "learning_rate": 5.0724080649705526e-05, + "loss": 0.3963, + "step": 1528 + }, + { + "epoch": 2.3632148377125195, + "grad_norm": 0.32770408845151705, + "learning_rate": 5.068248622153252e-05, + "loss": 0.3842, + "step": 1529 + }, + { + "epoch": 2.364760432766615, + "grad_norm": 0.34329092443466325, + "learning_rate": 5.0640879351432934e-05, + "loss": 0.3972, + "step": 1530 + }, + { + "epoch": 2.366306027820711, + "grad_norm": 0.28538432097026656, + "learning_rate": 5.059926008786648e-05, + "loss": 0.3848, + "step": 1531 + }, + { + "epoch": 2.3678516228748068, + "grad_norm": 0.2351912104958466, + "learning_rate": 5.055762847930719e-05, + "loss": 0.3931, + "step": 1532 + }, + { + "epoch": 2.3693972179289027, + "grad_norm": 0.2675290659579395, + "learning_rate": 5.051598457424355e-05, + "loss": 0.4034, + "step": 1533 + }, + { + "epoch": 2.3709428129829986, + "grad_norm": 0.2450247585979996, + "learning_rate": 5.047432842117836e-05, + "loss": 0.3845, + "step": 1534 + }, + { + "epoch": 2.3724884080370945, + "grad_norm": 0.24341720101474015, + "learning_rate": 5.043266006862869e-05, + "loss": 0.3928, + "step": 1535 + }, + { + "epoch": 2.37403400309119, + "grad_norm": 0.2913711114752356, + "learning_rate": 5.039097956512577e-05, + "loss": 0.3888, + "step": 1536 + }, + { + "epoch": 2.375579598145286, + "grad_norm": 0.3530909123361894, + "learning_rate": 5.034928695921506e-05, + "loss": 0.3904, + "step": 1537 + }, + { + "epoch": 2.3771251931993818, + "grad_norm": 0.3945139153800936, + "learning_rate": 5.030758229945605e-05, + "loss": 0.3854, + "step": 1538 + }, + { + "epoch": 2.3786707882534777, + "grad_norm": 0.39017734529459125, + "learning_rate": 5.0265865634422284e-05, + "loss": 0.393, + "step": 1539 + }, + { + "epoch": 2.3802163833075736, + "grad_norm": 0.305288369267672, + "learning_rate": 5.0224137012701314e-05, + "loss": 0.3962, + "step": 1540 + }, + { + "epoch": 2.381761978361669, + "grad_norm": 0.3139832140402103, + "learning_rate": 5.018239648289461e-05, + "loss": 0.4092, + "step": 1541 + }, + { + "epoch": 2.383307573415765, + "grad_norm": 0.3163870471362883, + "learning_rate": 5.014064409361749e-05, + "loss": 0.3893, + "step": 1542 + }, + { + "epoch": 2.384853168469861, + "grad_norm": 0.30771143522996813, + "learning_rate": 5.009887989349909e-05, + "loss": 0.3893, + "step": 1543 + }, + { + "epoch": 2.3863987635239567, + "grad_norm": 0.304189291631867, + "learning_rate": 5.0057103931182324e-05, + "loss": 0.3891, + "step": 1544 + }, + { + "epoch": 2.3879443585780527, + "grad_norm": 0.229975198322786, + "learning_rate": 5.0015316255323803e-05, + "loss": 0.3998, + "step": 1545 + }, + { + "epoch": 2.3894899536321486, + "grad_norm": 0.28369565359041216, + "learning_rate": 4.997351691459374e-05, + "loss": 0.4022, + "step": 1546 + }, + { + "epoch": 2.391035548686244, + "grad_norm": 0.3174888594737512, + "learning_rate": 4.993170595767598e-05, + "loss": 0.396, + "step": 1547 + }, + { + "epoch": 2.39258114374034, + "grad_norm": 0.2909324953054256, + "learning_rate": 4.988988343326788e-05, + "loss": 0.3893, + "step": 1548 + }, + { + "epoch": 2.394126738794436, + "grad_norm": 0.24681751451704773, + "learning_rate": 4.984804939008027e-05, + "loss": 0.3949, + "step": 1549 + }, + { + "epoch": 2.3956723338485317, + "grad_norm": 0.3624724434626657, + "learning_rate": 4.9806203876837386e-05, + "loss": 0.3933, + "step": 1550 + }, + { + "epoch": 2.3972179289026276, + "grad_norm": 0.248485801335781, + "learning_rate": 4.976434694227687e-05, + "loss": 0.3993, + "step": 1551 + }, + { + "epoch": 2.398763523956723, + "grad_norm": 0.2535610648412525, + "learning_rate": 4.972247863514959e-05, + "loss": 0.394, + "step": 1552 + }, + { + "epoch": 2.400309119010819, + "grad_norm": 0.24641241219834106, + "learning_rate": 4.968059900421971e-05, + "loss": 0.3857, + "step": 1553 + }, + { + "epoch": 2.401854714064915, + "grad_norm": 0.16300375810319317, + "learning_rate": 4.963870809826458e-05, + "loss": 0.386, + "step": 1554 + }, + { + "epoch": 2.403400309119011, + "grad_norm": 0.19116905649294372, + "learning_rate": 4.959680596607467e-05, + "loss": 0.3924, + "step": 1555 + }, + { + "epoch": 2.4049459041731067, + "grad_norm": 0.23101898261927714, + "learning_rate": 4.955489265645355e-05, + "loss": 0.3986, + "step": 1556 + }, + { + "epoch": 2.4064914992272026, + "grad_norm": 0.2654842304461216, + "learning_rate": 4.951296821821775e-05, + "loss": 0.3832, + "step": 1557 + }, + { + "epoch": 2.4080370942812985, + "grad_norm": 0.24896210454474568, + "learning_rate": 4.947103270019682e-05, + "loss": 0.3985, + "step": 1558 + }, + { + "epoch": 2.409582689335394, + "grad_norm": 0.22508988746256473, + "learning_rate": 4.942908615123322e-05, + "loss": 0.3877, + "step": 1559 + }, + { + "epoch": 2.41112828438949, + "grad_norm": 0.20552555820521468, + "learning_rate": 4.938712862018218e-05, + "loss": 0.3911, + "step": 1560 + }, + { + "epoch": 2.412673879443586, + "grad_norm": 0.20543558148018728, + "learning_rate": 4.93451601559118e-05, + "loss": 0.3902, + "step": 1561 + }, + { + "epoch": 2.4142194744976817, + "grad_norm": 0.21293359127356493, + "learning_rate": 4.930318080730292e-05, + "loss": 0.3934, + "step": 1562 + }, + { + "epoch": 2.4157650695517776, + "grad_norm": 0.22528425535355565, + "learning_rate": 4.926119062324898e-05, + "loss": 0.3886, + "step": 1563 + }, + { + "epoch": 2.417310664605873, + "grad_norm": 0.2740527567402077, + "learning_rate": 4.921918965265608e-05, + "loss": 0.3922, + "step": 1564 + }, + { + "epoch": 2.418856259659969, + "grad_norm": 0.222753478064478, + "learning_rate": 4.917717794444292e-05, + "loss": 0.3864, + "step": 1565 + }, + { + "epoch": 2.420401854714065, + "grad_norm": 0.2317187265658363, + "learning_rate": 4.913515554754066e-05, + "loss": 0.3954, + "step": 1566 + }, + { + "epoch": 2.421947449768161, + "grad_norm": 0.23958271070659026, + "learning_rate": 4.9093122510892916e-05, + "loss": 0.3967, + "step": 1567 + }, + { + "epoch": 2.4234930448222567, + "grad_norm": 0.24081649181744408, + "learning_rate": 4.905107888345571e-05, + "loss": 0.3885, + "step": 1568 + }, + { + "epoch": 2.4250386398763526, + "grad_norm": 0.28559014304553987, + "learning_rate": 4.900902471419739e-05, + "loss": 0.3812, + "step": 1569 + }, + { + "epoch": 2.426584234930448, + "grad_norm": 0.2706678308636684, + "learning_rate": 4.896696005209858e-05, + "loss": 0.3955, + "step": 1570 + }, + { + "epoch": 2.428129829984544, + "grad_norm": 0.269485650400652, + "learning_rate": 4.8924884946152146e-05, + "loss": 0.3817, + "step": 1571 + }, + { + "epoch": 2.42967542503864, + "grad_norm": 0.2657977637708371, + "learning_rate": 4.888279944536307e-05, + "loss": 0.3978, + "step": 1572 + }, + { + "epoch": 2.4312210200927358, + "grad_norm": 0.27799517604598306, + "learning_rate": 4.884070359874852e-05, + "loss": 0.3939, + "step": 1573 + }, + { + "epoch": 2.4327666151468317, + "grad_norm": 0.2675807845283423, + "learning_rate": 4.879859745533761e-05, + "loss": 0.3952, + "step": 1574 + }, + { + "epoch": 2.434312210200927, + "grad_norm": 0.2086648242083336, + "learning_rate": 4.875648106417153e-05, + "loss": 0.3973, + "step": 1575 + }, + { + "epoch": 2.435857805255023, + "grad_norm": 0.21399432663226547, + "learning_rate": 4.871435447430339e-05, + "loss": 0.3939, + "step": 1576 + }, + { + "epoch": 2.437403400309119, + "grad_norm": 0.25115123792120214, + "learning_rate": 4.867221773479815e-05, + "loss": 0.3817, + "step": 1577 + }, + { + "epoch": 2.438948995363215, + "grad_norm": 0.2668161649482835, + "learning_rate": 4.863007089473262e-05, + "loss": 0.389, + "step": 1578 + }, + { + "epoch": 2.4404945904173108, + "grad_norm": 0.2502869198766568, + "learning_rate": 4.858791400319534e-05, + "loss": 0.3836, + "step": 1579 + }, + { + "epoch": 2.4420401854714067, + "grad_norm": 0.19761123703119207, + "learning_rate": 4.854574710928662e-05, + "loss": 0.3918, + "step": 1580 + }, + { + "epoch": 2.443585780525502, + "grad_norm": 0.20489041079125325, + "learning_rate": 4.8503570262118336e-05, + "loss": 0.388, + "step": 1581 + }, + { + "epoch": 2.445131375579598, + "grad_norm": 0.23761986430786358, + "learning_rate": 4.8461383510814e-05, + "loss": 0.3989, + "step": 1582 + }, + { + "epoch": 2.446676970633694, + "grad_norm": 0.26799255466712524, + "learning_rate": 4.841918690450871e-05, + "loss": 0.3953, + "step": 1583 + }, + { + "epoch": 2.44822256568779, + "grad_norm": 0.22044145729388834, + "learning_rate": 4.837698049234893e-05, + "loss": 0.3918, + "step": 1584 + }, + { + "epoch": 2.4497681607418857, + "grad_norm": 0.19638779708616366, + "learning_rate": 4.833476432349264e-05, + "loss": 0.3941, + "step": 1585 + }, + { + "epoch": 2.451313755795981, + "grad_norm": 0.20539014072105105, + "learning_rate": 4.829253844710914e-05, + "loss": 0.3895, + "step": 1586 + }, + { + "epoch": 2.452859350850077, + "grad_norm": 0.2615520342957774, + "learning_rate": 4.825030291237906e-05, + "loss": 0.3857, + "step": 1587 + }, + { + "epoch": 2.454404945904173, + "grad_norm": 0.2353556811492111, + "learning_rate": 4.820805776849424e-05, + "loss": 0.3931, + "step": 1588 + }, + { + "epoch": 2.455950540958269, + "grad_norm": 0.26594630149950554, + "learning_rate": 4.816580306465775e-05, + "loss": 0.4015, + "step": 1589 + }, + { + "epoch": 2.457496136012365, + "grad_norm": 0.2847862158071832, + "learning_rate": 4.8123538850083796e-05, + "loss": 0.3939, + "step": 1590 + }, + { + "epoch": 2.4590417310664607, + "grad_norm": 0.2325110176480486, + "learning_rate": 4.8081265173997624e-05, + "loss": 0.387, + "step": 1591 + }, + { + "epoch": 2.4605873261205566, + "grad_norm": 0.2605313564717108, + "learning_rate": 4.803898208563552e-05, + "loss": 0.3938, + "step": 1592 + }, + { + "epoch": 2.462132921174652, + "grad_norm": 0.251069170835732, + "learning_rate": 4.799668963424476e-05, + "loss": 0.4004, + "step": 1593 + }, + { + "epoch": 2.463678516228748, + "grad_norm": 0.1985495939966303, + "learning_rate": 4.7954387869083486e-05, + "loss": 0.3906, + "step": 1594 + }, + { + "epoch": 2.465224111282844, + "grad_norm": 0.24250392489168981, + "learning_rate": 4.7912076839420695e-05, + "loss": 0.3962, + "step": 1595 + }, + { + "epoch": 2.46676970633694, + "grad_norm": 0.24210751742714964, + "learning_rate": 4.786975659453619e-05, + "loss": 0.3926, + "step": 1596 + }, + { + "epoch": 2.4683153013910357, + "grad_norm": 0.2516813510196297, + "learning_rate": 4.78274271837205e-05, + "loss": 0.3971, + "step": 1597 + }, + { + "epoch": 2.469860896445131, + "grad_norm": 0.231444499963526, + "learning_rate": 4.7785088656274824e-05, + "loss": 0.3922, + "step": 1598 + }, + { + "epoch": 2.471406491499227, + "grad_norm": 0.1858894539903597, + "learning_rate": 4.7742741061510975e-05, + "loss": 0.394, + "step": 1599 + }, + { + "epoch": 2.472952086553323, + "grad_norm": 0.2753734441818444, + "learning_rate": 4.770038444875136e-05, + "loss": 0.3913, + "step": 1600 + }, + { + "epoch": 2.474497681607419, + "grad_norm": 0.2530319282868681, + "learning_rate": 4.765801886732884e-05, + "loss": 0.4057, + "step": 1601 + }, + { + "epoch": 2.476043276661515, + "grad_norm": 0.24971066262343666, + "learning_rate": 4.761564436658674e-05, + "loss": 0.3892, + "step": 1602 + }, + { + "epoch": 2.4775888717156107, + "grad_norm": 0.21172394410295473, + "learning_rate": 4.75732609958788e-05, + "loss": 0.4007, + "step": 1603 + }, + { + "epoch": 2.479134466769706, + "grad_norm": 0.21123359781475406, + "learning_rate": 4.753086880456907e-05, + "loss": 0.3986, + "step": 1604 + }, + { + "epoch": 2.480680061823802, + "grad_norm": 0.21942737233497195, + "learning_rate": 4.748846784203187e-05, + "loss": 0.4015, + "step": 1605 + }, + { + "epoch": 2.482225656877898, + "grad_norm": 0.30372087235313455, + "learning_rate": 4.7446058157651715e-05, + "loss": 0.3889, + "step": 1606 + }, + { + "epoch": 2.483771251931994, + "grad_norm": 0.20412899046160612, + "learning_rate": 4.7403639800823344e-05, + "loss": 0.3865, + "step": 1607 + }, + { + "epoch": 2.48531684698609, + "grad_norm": 0.2193097389895047, + "learning_rate": 4.7361212820951525e-05, + "loss": 0.3887, + "step": 1608 + }, + { + "epoch": 2.4868624420401853, + "grad_norm": 0.2721648103907066, + "learning_rate": 4.731877726745112e-05, + "loss": 0.4021, + "step": 1609 + }, + { + "epoch": 2.488408037094281, + "grad_norm": 0.22312355301605005, + "learning_rate": 4.7276333189746944e-05, + "loss": 0.3939, + "step": 1610 + }, + { + "epoch": 2.489953632148377, + "grad_norm": 0.22679494324181823, + "learning_rate": 4.723388063727375e-05, + "loss": 0.3909, + "step": 1611 + }, + { + "epoch": 2.491499227202473, + "grad_norm": 0.22064137384238353, + "learning_rate": 4.719141965947618e-05, + "loss": 0.3897, + "step": 1612 + }, + { + "epoch": 2.493044822256569, + "grad_norm": 0.24253869301521516, + "learning_rate": 4.714895030580865e-05, + "loss": 0.3903, + "step": 1613 + }, + { + "epoch": 2.4945904173106648, + "grad_norm": 0.23241390259659883, + "learning_rate": 4.7106472625735375e-05, + "loss": 0.3929, + "step": 1614 + }, + { + "epoch": 2.4961360123647607, + "grad_norm": 0.2771522090480515, + "learning_rate": 4.7063986668730246e-05, + "loss": 0.3861, + "step": 1615 + }, + { + "epoch": 2.497681607418856, + "grad_norm": 0.3106756875403178, + "learning_rate": 4.702149248427677e-05, + "loss": 0.3895, + "step": 1616 + }, + { + "epoch": 2.499227202472952, + "grad_norm": 0.34350424156535503, + "learning_rate": 4.6978990121868096e-05, + "loss": 0.3862, + "step": 1617 + }, + { + "epoch": 2.500772797527048, + "grad_norm": 0.3223694433694756, + "learning_rate": 4.6936479631006873e-05, + "loss": 0.3907, + "step": 1618 + }, + { + "epoch": 2.502318392581144, + "grad_norm": 0.2726783290551887, + "learning_rate": 4.689396106120516e-05, + "loss": 0.3985, + "step": 1619 + }, + { + "epoch": 2.5038639876352393, + "grad_norm": 0.23225676032680767, + "learning_rate": 4.6851434461984505e-05, + "loss": 0.3953, + "step": 1620 + }, + { + "epoch": 2.5054095826893352, + "grad_norm": 0.32044583909189767, + "learning_rate": 4.680889988287578e-05, + "loss": 0.39, + "step": 1621 + }, + { + "epoch": 2.506955177743431, + "grad_norm": 0.28909449180223523, + "learning_rate": 4.676635737341913e-05, + "loss": 0.3915, + "step": 1622 + }, + { + "epoch": 2.508500772797527, + "grad_norm": 0.3373093849569923, + "learning_rate": 4.672380698316398e-05, + "loss": 0.3957, + "step": 1623 + }, + { + "epoch": 2.510046367851623, + "grad_norm": 0.2934474440801747, + "learning_rate": 4.6681248761668886e-05, + "loss": 0.3944, + "step": 1624 + }, + { + "epoch": 2.511591962905719, + "grad_norm": 0.20853548303800792, + "learning_rate": 4.6638682758501553e-05, + "loss": 0.3867, + "step": 1625 + }, + { + "epoch": 2.5131375579598147, + "grad_norm": 0.20643856276794834, + "learning_rate": 4.659610902323875e-05, + "loss": 0.3841, + "step": 1626 + }, + { + "epoch": 2.51468315301391, + "grad_norm": 0.26311445438533454, + "learning_rate": 4.655352760546624e-05, + "loss": 0.3925, + "step": 1627 + }, + { + "epoch": 2.516228748068006, + "grad_norm": 0.2668275480704203, + "learning_rate": 4.651093855477874e-05, + "loss": 0.3919, + "step": 1628 + }, + { + "epoch": 2.517774343122102, + "grad_norm": 0.2694379532597783, + "learning_rate": 4.646834192077986e-05, + "loss": 0.3899, + "step": 1629 + }, + { + "epoch": 2.519319938176198, + "grad_norm": 0.26228893960462624, + "learning_rate": 4.6425737753082024e-05, + "loss": 0.3928, + "step": 1630 + }, + { + "epoch": 2.5208655332302934, + "grad_norm": 0.1936290956124116, + "learning_rate": 4.638312610130646e-05, + "loss": 0.3941, + "step": 1631 + }, + { + "epoch": 2.5224111282843893, + "grad_norm": 0.21327089164691612, + "learning_rate": 4.6340507015083116e-05, + "loss": 0.3924, + "step": 1632 + }, + { + "epoch": 2.523956723338485, + "grad_norm": 0.24903276404302455, + "learning_rate": 4.629788054405054e-05, + "loss": 0.3955, + "step": 1633 + }, + { + "epoch": 2.525502318392581, + "grad_norm": 0.25245491821379856, + "learning_rate": 4.625524673785593e-05, + "loss": 0.3829, + "step": 1634 + }, + { + "epoch": 2.527047913446677, + "grad_norm": 0.2844301896804582, + "learning_rate": 4.621260564615505e-05, + "loss": 0.3859, + "step": 1635 + }, + { + "epoch": 2.528593508500773, + "grad_norm": 0.21833037149589407, + "learning_rate": 4.6169957318612096e-05, + "loss": 0.3898, + "step": 1636 + }, + { + "epoch": 2.530139103554869, + "grad_norm": 0.19449687548089142, + "learning_rate": 4.6127301804899717e-05, + "loss": 0.3936, + "step": 1637 + }, + { + "epoch": 2.5316846986089647, + "grad_norm": 0.18782982149635674, + "learning_rate": 4.608463915469893e-05, + "loss": 0.3954, + "step": 1638 + }, + { + "epoch": 2.53323029366306, + "grad_norm": 0.2159678201262451, + "learning_rate": 4.604196941769906e-05, + "loss": 0.3918, + "step": 1639 + }, + { + "epoch": 2.534775888717156, + "grad_norm": 0.2563281079039888, + "learning_rate": 4.5999292643597693e-05, + "loss": 0.3836, + "step": 1640 + }, + { + "epoch": 2.536321483771252, + "grad_norm": 0.3491491507041857, + "learning_rate": 4.59566088821006e-05, + "loss": 0.3939, + "step": 1641 + }, + { + "epoch": 2.537867078825348, + "grad_norm": 0.25422624748850614, + "learning_rate": 4.591391818292171e-05, + "loss": 0.3817, + "step": 1642 + }, + { + "epoch": 2.5394126738794434, + "grad_norm": 0.1737983690973426, + "learning_rate": 4.5871220595783016e-05, + "loss": 0.388, + "step": 1643 + }, + { + "epoch": 2.5409582689335393, + "grad_norm": 0.2024156139334753, + "learning_rate": 4.582851617041452e-05, + "loss": 0.3919, + "step": 1644 + }, + { + "epoch": 2.542503863987635, + "grad_norm": 0.202932388910506, + "learning_rate": 4.578580495655423e-05, + "loss": 0.3953, + "step": 1645 + }, + { + "epoch": 2.544049459041731, + "grad_norm": 0.243242339398608, + "learning_rate": 4.5743087003948024e-05, + "loss": 0.3933, + "step": 1646 + }, + { + "epoch": 2.545595054095827, + "grad_norm": 0.24488570408288615, + "learning_rate": 4.570036236234963e-05, + "loss": 0.3927, + "step": 1647 + }, + { + "epoch": 2.547140649149923, + "grad_norm": 0.2685531776215014, + "learning_rate": 4.56576310815206e-05, + "loss": 0.3909, + "step": 1648 + }, + { + "epoch": 2.548686244204019, + "grad_norm": 0.2728965982324487, + "learning_rate": 4.561489321123018e-05, + "loss": 0.3876, + "step": 1649 + }, + { + "epoch": 2.5502318392581143, + "grad_norm": 0.2253682041586724, + "learning_rate": 4.557214880125533e-05, + "loss": 0.3976, + "step": 1650 + }, + { + "epoch": 2.55177743431221, + "grad_norm": 0.20483244888828098, + "learning_rate": 4.552939790138058e-05, + "loss": 0.3928, + "step": 1651 + }, + { + "epoch": 2.553323029366306, + "grad_norm": 0.2235110600526468, + "learning_rate": 4.548664056139803e-05, + "loss": 0.3915, + "step": 1652 + }, + { + "epoch": 2.554868624420402, + "grad_norm": 0.25680000871236164, + "learning_rate": 4.544387683110733e-05, + "loss": 0.4003, + "step": 1653 + }, + { + "epoch": 2.5564142194744974, + "grad_norm": 0.3330721283056668, + "learning_rate": 4.540110676031552e-05, + "loss": 0.3862, + "step": 1654 + }, + { + "epoch": 2.5579598145285933, + "grad_norm": 0.3367498008963851, + "learning_rate": 4.5358330398837035e-05, + "loss": 0.3884, + "step": 1655 + }, + { + "epoch": 2.5595054095826892, + "grad_norm": 0.285531764031909, + "learning_rate": 4.531554779649365e-05, + "loss": 0.3902, + "step": 1656 + }, + { + "epoch": 2.561051004636785, + "grad_norm": 0.25585937876834175, + "learning_rate": 4.52727590031144e-05, + "loss": 0.397, + "step": 1657 + }, + { + "epoch": 2.562596599690881, + "grad_norm": 0.2765152500473313, + "learning_rate": 4.522996406853552e-05, + "loss": 0.3977, + "step": 1658 + }, + { + "epoch": 2.564142194744977, + "grad_norm": 0.2121356374509123, + "learning_rate": 4.5187163042600416e-05, + "loss": 0.3953, + "step": 1659 + }, + { + "epoch": 2.565687789799073, + "grad_norm": 0.23605796283812225, + "learning_rate": 4.514435597515962e-05, + "loss": 0.3942, + "step": 1660 + }, + { + "epoch": 2.5672333848531683, + "grad_norm": 0.280273284639065, + "learning_rate": 4.510154291607061e-05, + "loss": 0.3974, + "step": 1661 + }, + { + "epoch": 2.5687789799072642, + "grad_norm": 0.2862636237262634, + "learning_rate": 4.5058723915197916e-05, + "loss": 0.3992, + "step": 1662 + }, + { + "epoch": 2.57032457496136, + "grad_norm": 0.20355355298012318, + "learning_rate": 4.501589902241298e-05, + "loss": 0.3969, + "step": 1663 + }, + { + "epoch": 2.571870170015456, + "grad_norm": 0.22242162610416033, + "learning_rate": 4.497306828759407e-05, + "loss": 0.3834, + "step": 1664 + }, + { + "epoch": 2.573415765069552, + "grad_norm": 0.28927596076059225, + "learning_rate": 4.493023176062631e-05, + "loss": 0.3973, + "step": 1665 + }, + { + "epoch": 2.5749613601236474, + "grad_norm": 0.2620957900323579, + "learning_rate": 4.488738949140151e-05, + "loss": 0.3883, + "step": 1666 + }, + { + "epoch": 2.5765069551777433, + "grad_norm": 0.2039195819917358, + "learning_rate": 4.4844541529818234e-05, + "loss": 0.3883, + "step": 1667 + }, + { + "epoch": 2.578052550231839, + "grad_norm": 0.23742806986642045, + "learning_rate": 4.480168792578163e-05, + "loss": 0.3871, + "step": 1668 + }, + { + "epoch": 2.579598145285935, + "grad_norm": 0.23110172448224695, + "learning_rate": 4.475882872920341e-05, + "loss": 0.3844, + "step": 1669 + }, + { + "epoch": 2.581143740340031, + "grad_norm": 0.27447312756803394, + "learning_rate": 4.471596399000186e-05, + "loss": 0.3908, + "step": 1670 + }, + { + "epoch": 2.582689335394127, + "grad_norm": 0.23463566543332323, + "learning_rate": 4.4673093758101655e-05, + "loss": 0.3887, + "step": 1671 + }, + { + "epoch": 2.584234930448223, + "grad_norm": 0.20136632940367377, + "learning_rate": 4.463021808343389e-05, + "loss": 0.3942, + "step": 1672 + }, + { + "epoch": 2.5857805255023183, + "grad_norm": 0.18778083320382974, + "learning_rate": 4.458733701593603e-05, + "loss": 0.3932, + "step": 1673 + }, + { + "epoch": 2.587326120556414, + "grad_norm": 0.19782531967338815, + "learning_rate": 4.454445060555176e-05, + "loss": 0.3945, + "step": 1674 + }, + { + "epoch": 2.58887171561051, + "grad_norm": 0.22508762999604104, + "learning_rate": 4.450155890223104e-05, + "loss": 0.3873, + "step": 1675 + }, + { + "epoch": 2.590417310664606, + "grad_norm": 0.21168664081062777, + "learning_rate": 4.445866195592998e-05, + "loss": 0.3926, + "step": 1676 + }, + { + "epoch": 2.5919629057187015, + "grad_norm": 0.16234233057601102, + "learning_rate": 4.44157598166108e-05, + "loss": 0.3957, + "step": 1677 + }, + { + "epoch": 2.5935085007727974, + "grad_norm": 0.2006506183967543, + "learning_rate": 4.437285253424174e-05, + "loss": 0.3987, + "step": 1678 + }, + { + "epoch": 2.5950540958268933, + "grad_norm": 0.19266544117512432, + "learning_rate": 4.432994015879707e-05, + "loss": 0.3885, + "step": 1679 + }, + { + "epoch": 2.596599690880989, + "grad_norm": 0.23946286724356225, + "learning_rate": 4.428702274025694e-05, + "loss": 0.402, + "step": 1680 + }, + { + "epoch": 2.598145285935085, + "grad_norm": 0.26911538712111643, + "learning_rate": 4.424410032860747e-05, + "loss": 0.3854, + "step": 1681 + }, + { + "epoch": 2.599690880989181, + "grad_norm": 0.24576664996872866, + "learning_rate": 4.420117297384048e-05, + "loss": 0.3955, + "step": 1682 + }, + { + "epoch": 2.601236476043277, + "grad_norm": 0.20026359969497476, + "learning_rate": 4.415824072595361e-05, + "loss": 0.3966, + "step": 1683 + }, + { + "epoch": 2.6027820710973724, + "grad_norm": 0.24261706436109684, + "learning_rate": 4.411530363495021e-05, + "loss": 0.3913, + "step": 1684 + }, + { + "epoch": 2.6043276661514683, + "grad_norm": 0.5551795310711124, + "learning_rate": 4.407236175083923e-05, + "loss": 0.3955, + "step": 1685 + }, + { + "epoch": 2.605873261205564, + "grad_norm": 0.29021409934929493, + "learning_rate": 4.402941512363522e-05, + "loss": 0.3961, + "step": 1686 + }, + { + "epoch": 2.60741885625966, + "grad_norm": 0.25950368123392215, + "learning_rate": 4.3986463803358284e-05, + "loss": 0.3964, + "step": 1687 + }, + { + "epoch": 2.6089644513137555, + "grad_norm": 0.24088722690149555, + "learning_rate": 4.394350784003394e-05, + "loss": 0.3969, + "step": 1688 + }, + { + "epoch": 2.6105100463678514, + "grad_norm": 0.2324338024413383, + "learning_rate": 4.3900547283693126e-05, + "loss": 0.379, + "step": 1689 + }, + { + "epoch": 2.6120556414219473, + "grad_norm": 0.21735680121166714, + "learning_rate": 4.385758218437218e-05, + "loss": 0.4003, + "step": 1690 + }, + { + "epoch": 2.6136012364760433, + "grad_norm": 0.18670609149054954, + "learning_rate": 4.3814612592112675e-05, + "loss": 0.3933, + "step": 1691 + }, + { + "epoch": 2.615146831530139, + "grad_norm": 0.23190433118432097, + "learning_rate": 4.377163855696143e-05, + "loss": 0.3833, + "step": 1692 + }, + { + "epoch": 2.616692426584235, + "grad_norm": 0.23098885325708501, + "learning_rate": 4.3728660128970454e-05, + "loss": 0.3941, + "step": 1693 + }, + { + "epoch": 2.618238021638331, + "grad_norm": 0.18041231125937515, + "learning_rate": 4.3685677358196855e-05, + "loss": 0.3935, + "step": 1694 + }, + { + "epoch": 2.6197836166924264, + "grad_norm": 0.18907903531096312, + "learning_rate": 4.364269029470281e-05, + "loss": 0.3927, + "step": 1695 + }, + { + "epoch": 2.6213292117465223, + "grad_norm": 0.21197360112624988, + "learning_rate": 4.359969898855549e-05, + "loss": 0.3862, + "step": 1696 + }, + { + "epoch": 2.6228748068006182, + "grad_norm": 0.2415173234059811, + "learning_rate": 4.355670348982701e-05, + "loss": 0.3972, + "step": 1697 + }, + { + "epoch": 2.624420401854714, + "grad_norm": 0.2538082855558849, + "learning_rate": 4.3513703848594366e-05, + "loss": 0.3971, + "step": 1698 + }, + { + "epoch": 2.62596599690881, + "grad_norm": 0.21427677280173618, + "learning_rate": 4.347070011493939e-05, + "loss": 0.3892, + "step": 1699 + }, + { + "epoch": 2.6275115919629055, + "grad_norm": 0.27485089107298366, + "learning_rate": 4.342769233894865e-05, + "loss": 0.3873, + "step": 1700 + }, + { + "epoch": 2.6290571870170014, + "grad_norm": 0.2374833907537212, + "learning_rate": 4.338468057071347e-05, + "loss": 0.4006, + "step": 1701 + }, + { + "epoch": 2.6306027820710973, + "grad_norm": 0.22520297589709645, + "learning_rate": 4.334166486032977e-05, + "loss": 0.3925, + "step": 1702 + }, + { + "epoch": 2.6321483771251932, + "grad_norm": 0.3218468827633287, + "learning_rate": 4.329864525789811e-05, + "loss": 0.3889, + "step": 1703 + }, + { + "epoch": 2.633693972179289, + "grad_norm": 0.3019997561853618, + "learning_rate": 4.325562181352356e-05, + "loss": 0.3912, + "step": 1704 + }, + { + "epoch": 2.635239567233385, + "grad_norm": 0.2093153275570432, + "learning_rate": 4.321259457731565e-05, + "loss": 0.3863, + "step": 1705 + }, + { + "epoch": 2.636785162287481, + "grad_norm": 0.20006225048976253, + "learning_rate": 4.316956359938835e-05, + "loss": 0.3932, + "step": 1706 + }, + { + "epoch": 2.6383307573415764, + "grad_norm": 0.22304616722385706, + "learning_rate": 4.3126528929859994e-05, + "loss": 0.3925, + "step": 1707 + }, + { + "epoch": 2.6398763523956723, + "grad_norm": 0.21878151455286357, + "learning_rate": 4.3083490618853175e-05, + "loss": 0.3902, + "step": 1708 + }, + { + "epoch": 2.641421947449768, + "grad_norm": 0.22444441987583222, + "learning_rate": 4.3040448716494775e-05, + "loss": 0.3961, + "step": 1709 + }, + { + "epoch": 2.642967542503864, + "grad_norm": 0.2483241102745166, + "learning_rate": 4.2997403272915836e-05, + "loss": 0.3934, + "step": 1710 + }, + { + "epoch": 2.6445131375579596, + "grad_norm": 0.22959455078957128, + "learning_rate": 4.295435433825152e-05, + "loss": 0.3824, + "step": 1711 + }, + { + "epoch": 2.6460587326120555, + "grad_norm": 0.21699514755570257, + "learning_rate": 4.2911301962641065e-05, + "loss": 0.3939, + "step": 1712 + }, + { + "epoch": 2.6476043276661514, + "grad_norm": 0.270050550781768, + "learning_rate": 4.2868246196227715e-05, + "loss": 0.3947, + "step": 1713 + }, + { + "epoch": 2.6491499227202473, + "grad_norm": 0.2272440301136524, + "learning_rate": 4.2825187089158656e-05, + "loss": 0.3932, + "step": 1714 + }, + { + "epoch": 2.650695517774343, + "grad_norm": 0.20918266050275014, + "learning_rate": 4.2782124691584984e-05, + "loss": 0.3977, + "step": 1715 + }, + { + "epoch": 2.652241112828439, + "grad_norm": 0.19535768556289942, + "learning_rate": 4.273905905366159e-05, + "loss": 0.3919, + "step": 1716 + }, + { + "epoch": 2.653786707882535, + "grad_norm": 0.27365474283090385, + "learning_rate": 4.269599022554718e-05, + "loss": 0.3978, + "step": 1717 + }, + { + "epoch": 2.6553323029366305, + "grad_norm": 0.2862309974660244, + "learning_rate": 4.265291825740416e-05, + "loss": 0.3865, + "step": 1718 + }, + { + "epoch": 2.6568778979907264, + "grad_norm": 0.24978807145564993, + "learning_rate": 4.2609843199398594e-05, + "loss": 0.3866, + "step": 1719 + }, + { + "epoch": 2.6584234930448223, + "grad_norm": 0.18804780266739948, + "learning_rate": 4.256676510170013e-05, + "loss": 0.3896, + "step": 1720 + }, + { + "epoch": 2.659969088098918, + "grad_norm": 0.20114039643341255, + "learning_rate": 4.252368401448198e-05, + "loss": 0.3877, + "step": 1721 + }, + { + "epoch": 2.6615146831530136, + "grad_norm": 0.22568422793760368, + "learning_rate": 4.248059998792081e-05, + "loss": 0.3953, + "step": 1722 + }, + { + "epoch": 2.6630602782071096, + "grad_norm": 0.23913012062540562, + "learning_rate": 4.243751307219675e-05, + "loss": 0.3958, + "step": 1723 + }, + { + "epoch": 2.6646058732612055, + "grad_norm": 0.23926122468307615, + "learning_rate": 4.2394423317493256e-05, + "loss": 0.3941, + "step": 1724 + }, + { + "epoch": 2.6661514683153014, + "grad_norm": 0.26739200174714023, + "learning_rate": 4.2351330773997116e-05, + "loss": 0.3938, + "step": 1725 + }, + { + "epoch": 2.6676970633693973, + "grad_norm": 0.2521231378359259, + "learning_rate": 4.230823549189834e-05, + "loss": 0.3964, + "step": 1726 + }, + { + "epoch": 2.669242658423493, + "grad_norm": 0.21798757466882637, + "learning_rate": 4.226513752139016e-05, + "loss": 0.3797, + "step": 1727 + }, + { + "epoch": 2.670788253477589, + "grad_norm": 0.24296505447418829, + "learning_rate": 4.2222036912668915e-05, + "loss": 0.3993, + "step": 1728 + }, + { + "epoch": 2.6723338485316845, + "grad_norm": 0.275836142582003, + "learning_rate": 4.2178933715934027e-05, + "loss": 0.3995, + "step": 1729 + }, + { + "epoch": 2.6738794435857804, + "grad_norm": 0.23257862603468263, + "learning_rate": 4.2135827981387936e-05, + "loss": 0.3852, + "step": 1730 + }, + { + "epoch": 2.6754250386398764, + "grad_norm": 0.2509912628829087, + "learning_rate": 4.2092719759236013e-05, + "loss": 0.3926, + "step": 1731 + }, + { + "epoch": 2.6769706336939723, + "grad_norm": 0.29179187494496706, + "learning_rate": 4.204960909968656e-05, + "loss": 0.3911, + "step": 1732 + }, + { + "epoch": 2.678516228748068, + "grad_norm": 0.25691820918430236, + "learning_rate": 4.200649605295071e-05, + "loss": 0.3886, + "step": 1733 + }, + { + "epoch": 2.6800618238021636, + "grad_norm": 0.19850809916040388, + "learning_rate": 4.1963380669242335e-05, + "loss": 0.3859, + "step": 1734 + }, + { + "epoch": 2.6816074188562595, + "grad_norm": 0.2584257513915739, + "learning_rate": 4.192026299877811e-05, + "loss": 0.3816, + "step": 1735 + }, + { + "epoch": 2.6831530139103554, + "grad_norm": 0.25829112210072946, + "learning_rate": 4.187714309177729e-05, + "loss": 0.3903, + "step": 1736 + }, + { + "epoch": 2.6846986089644513, + "grad_norm": 0.22790216466337682, + "learning_rate": 4.1834020998461785e-05, + "loss": 0.3861, + "step": 1737 + }, + { + "epoch": 2.6862442040185472, + "grad_norm": 0.2914850077764589, + "learning_rate": 4.179089676905605e-05, + "loss": 0.394, + "step": 1738 + }, + { + "epoch": 2.687789799072643, + "grad_norm": 0.2671371421689647, + "learning_rate": 4.174777045378699e-05, + "loss": 0.3896, + "step": 1739 + }, + { + "epoch": 2.689335394126739, + "grad_norm": 0.19241631176557675, + "learning_rate": 4.170464210288399e-05, + "loss": 0.3955, + "step": 1740 + }, + { + "epoch": 2.6908809891808345, + "grad_norm": 0.2086286450779931, + "learning_rate": 4.1661511766578774e-05, + "loss": 0.3967, + "step": 1741 + }, + { + "epoch": 2.6924265842349304, + "grad_norm": 0.23025192323619, + "learning_rate": 4.1618379495105364e-05, + "loss": 0.3981, + "step": 1742 + }, + { + "epoch": 2.6939721792890263, + "grad_norm": 0.2413035288042119, + "learning_rate": 4.157524533870009e-05, + "loss": 0.3985, + "step": 1743 + }, + { + "epoch": 2.6955177743431222, + "grad_norm": 0.22474391007406155, + "learning_rate": 4.153210934760142e-05, + "loss": 0.3896, + "step": 1744 + }, + { + "epoch": 2.6970633693972177, + "grad_norm": 0.20021352917828603, + "learning_rate": 4.148897157204999e-05, + "loss": 0.386, + "step": 1745 + }, + { + "epoch": 2.6986089644513136, + "grad_norm": 0.26474447032524473, + "learning_rate": 4.144583206228852e-05, + "loss": 0.3998, + "step": 1746 + }, + { + "epoch": 2.7001545595054095, + "grad_norm": 0.2158672684708272, + "learning_rate": 4.1402690868561735e-05, + "loss": 0.3887, + "step": 1747 + }, + { + "epoch": 2.7017001545595054, + "grad_norm": 0.24266533972938326, + "learning_rate": 4.13595480411163e-05, + "loss": 0.3943, + "step": 1748 + }, + { + "epoch": 2.7032457496136013, + "grad_norm": 0.24274619528355063, + "learning_rate": 4.1316403630200825e-05, + "loss": 0.3789, + "step": 1749 + }, + { + "epoch": 2.704791344667697, + "grad_norm": 0.2916946178505382, + "learning_rate": 4.127325768606576e-05, + "loss": 0.3962, + "step": 1750 + }, + { + "epoch": 2.706336939721793, + "grad_norm": 0.2413892655948878, + "learning_rate": 4.12301102589633e-05, + "loss": 0.3913, + "step": 1751 + }, + { + "epoch": 2.7078825347758886, + "grad_norm": 0.1904043535480117, + "learning_rate": 4.118696139914742e-05, + "loss": 0.3911, + "step": 1752 + }, + { + "epoch": 2.7094281298299845, + "grad_norm": 0.22388949005518227, + "learning_rate": 4.114381115687371e-05, + "loss": 0.3956, + "step": 1753 + }, + { + "epoch": 2.7109737248840804, + "grad_norm": 0.17867764017934373, + "learning_rate": 4.110065958239942e-05, + "loss": 0.3878, + "step": 1754 + }, + { + "epoch": 2.7125193199381763, + "grad_norm": 0.20689302795190623, + "learning_rate": 4.105750672598333e-05, + "loss": 0.4045, + "step": 1755 + }, + { + "epoch": 2.7140649149922718, + "grad_norm": 0.23874824045849408, + "learning_rate": 4.101435263788569e-05, + "loss": 0.3894, + "step": 1756 + }, + { + "epoch": 2.7156105100463677, + "grad_norm": 0.26648071941463974, + "learning_rate": 4.097119736836823e-05, + "loss": 0.3873, + "step": 1757 + }, + { + "epoch": 2.7171561051004636, + "grad_norm": 0.21952269929540558, + "learning_rate": 4.092804096769404e-05, + "loss": 0.3863, + "step": 1758 + }, + { + "epoch": 2.7187017001545595, + "grad_norm": 0.24690525236175953, + "learning_rate": 4.088488348612748e-05, + "loss": 0.3897, + "step": 1759 + }, + { + "epoch": 2.7202472952086554, + "grad_norm": 0.17406867030472212, + "learning_rate": 4.084172497393425e-05, + "loss": 0.3921, + "step": 1760 + }, + { + "epoch": 2.7217928902627513, + "grad_norm": 0.18210134548010445, + "learning_rate": 4.0798565481381194e-05, + "loss": 0.3967, + "step": 1761 + }, + { + "epoch": 2.723338485316847, + "grad_norm": 0.21945415295395937, + "learning_rate": 4.0755405058736314e-05, + "loss": 0.3954, + "step": 1762 + }, + { + "epoch": 2.7248840803709427, + "grad_norm": 0.232237741090731, + "learning_rate": 4.07122437562687e-05, + "loss": 0.3967, + "step": 1763 + }, + { + "epoch": 2.7264296754250386, + "grad_norm": 0.2974379407362513, + "learning_rate": 4.0669081624248464e-05, + "loss": 0.401, + "step": 1764 + }, + { + "epoch": 2.7279752704791345, + "grad_norm": 0.21490090328087502, + "learning_rate": 4.0625918712946665e-05, + "loss": 0.3857, + "step": 1765 + }, + { + "epoch": 2.7295208655332304, + "grad_norm": 0.19018711214547304, + "learning_rate": 4.05827550726353e-05, + "loss": 0.4022, + "step": 1766 + }, + { + "epoch": 2.7310664605873263, + "grad_norm": 0.2161740494195666, + "learning_rate": 4.0539590753587195e-05, + "loss": 0.3945, + "step": 1767 + }, + { + "epoch": 2.7326120556414217, + "grad_norm": 0.22327255615725017, + "learning_rate": 4.049642580607599e-05, + "loss": 0.3919, + "step": 1768 + }, + { + "epoch": 2.7341576506955176, + "grad_norm": 0.2002937348555192, + "learning_rate": 4.045326028037603e-05, + "loss": 0.391, + "step": 1769 + }, + { + "epoch": 2.7357032457496135, + "grad_norm": 0.21879066736924463, + "learning_rate": 4.041009422676234e-05, + "loss": 0.3932, + "step": 1770 + }, + { + "epoch": 2.7372488408037094, + "grad_norm": 0.21984116168203174, + "learning_rate": 4.036692769551056e-05, + "loss": 0.3815, + "step": 1771 + }, + { + "epoch": 2.7387944358578054, + "grad_norm": 0.22773966477195914, + "learning_rate": 4.0323760736896904e-05, + "loss": 0.3871, + "step": 1772 + }, + { + "epoch": 2.7403400309119013, + "grad_norm": 0.17651812601554587, + "learning_rate": 4.0280593401198046e-05, + "loss": 0.3874, + "step": 1773 + }, + { + "epoch": 2.741885625965997, + "grad_norm": 0.2452970756456186, + "learning_rate": 4.0237425738691156e-05, + "loss": 0.3911, + "step": 1774 + }, + { + "epoch": 2.7434312210200926, + "grad_norm": 0.28920746138391956, + "learning_rate": 4.0194257799653724e-05, + "loss": 0.3971, + "step": 1775 + }, + { + "epoch": 2.7449768160741885, + "grad_norm": 0.2139771033872057, + "learning_rate": 4.0151089634363593e-05, + "loss": 0.3817, + "step": 1776 + }, + { + "epoch": 2.7465224111282844, + "grad_norm": 0.20423584050338486, + "learning_rate": 4.010792129309888e-05, + "loss": 0.3928, + "step": 1777 + }, + { + "epoch": 2.7480680061823803, + "grad_norm": 0.3212385675865011, + "learning_rate": 4.006475282613789e-05, + "loss": 0.3874, + "step": 1778 + }, + { + "epoch": 2.749613601236476, + "grad_norm": 0.2704711707718232, + "learning_rate": 4.0021584283759066e-05, + "loss": 0.3989, + "step": 1779 + }, + { + "epoch": 2.7511591962905717, + "grad_norm": 0.2682116360791284, + "learning_rate": 3.997841571624094e-05, + "loss": 0.3916, + "step": 1780 + }, + { + "epoch": 2.7527047913446676, + "grad_norm": 0.32279677803920165, + "learning_rate": 3.9935247173862124e-05, + "loss": 0.3866, + "step": 1781 + }, + { + "epoch": 2.7542503863987635, + "grad_norm": 0.2830715557450085, + "learning_rate": 3.9892078706901135e-05, + "loss": 0.3924, + "step": 1782 + }, + { + "epoch": 2.7557959814528594, + "grad_norm": 0.24500749618625672, + "learning_rate": 3.9848910365636406e-05, + "loss": 0.3923, + "step": 1783 + }, + { + "epoch": 2.7573415765069553, + "grad_norm": 0.264674266420782, + "learning_rate": 3.980574220034629e-05, + "loss": 0.3849, + "step": 1784 + }, + { + "epoch": 2.7588871715610512, + "grad_norm": 0.21841249256542083, + "learning_rate": 3.976257426130886e-05, + "loss": 0.3891, + "step": 1785 + }, + { + "epoch": 2.7604327666151467, + "grad_norm": 0.18420110960683264, + "learning_rate": 3.971940659880196e-05, + "loss": 0.3944, + "step": 1786 + }, + { + "epoch": 2.7619783616692426, + "grad_norm": 0.2877075785983188, + "learning_rate": 3.967623926310311e-05, + "loss": 0.3906, + "step": 1787 + }, + { + "epoch": 2.7635239567233385, + "grad_norm": 0.24290364105842152, + "learning_rate": 3.963307230448945e-05, + "loss": 0.4077, + "step": 1788 + }, + { + "epoch": 2.7650695517774344, + "grad_norm": 0.21695406242088836, + "learning_rate": 3.958990577323767e-05, + "loss": 0.3937, + "step": 1789 + }, + { + "epoch": 2.76661514683153, + "grad_norm": 0.25856486111006705, + "learning_rate": 3.954673971962398e-05, + "loss": 0.3959, + "step": 1790 + }, + { + "epoch": 2.7681607418856258, + "grad_norm": 0.2491166348704581, + "learning_rate": 3.950357419392402e-05, + "loss": 0.3879, + "step": 1791 + }, + { + "epoch": 2.7697063369397217, + "grad_norm": 0.20871545391562396, + "learning_rate": 3.9460409246412805e-05, + "loss": 0.4034, + "step": 1792 + }, + { + "epoch": 2.7712519319938176, + "grad_norm": 0.20508228510925783, + "learning_rate": 3.941724492736471e-05, + "loss": 0.4042, + "step": 1793 + }, + { + "epoch": 2.7727975270479135, + "grad_norm": 0.18720477368958455, + "learning_rate": 3.9374081287053355e-05, + "loss": 0.3893, + "step": 1794 + }, + { + "epoch": 2.7743431221020094, + "grad_norm": 0.19308946884703032, + "learning_rate": 3.9330918375751556e-05, + "loss": 0.3915, + "step": 1795 + }, + { + "epoch": 2.7758887171561053, + "grad_norm": 0.18226166599343516, + "learning_rate": 3.928775624373131e-05, + "loss": 0.3918, + "step": 1796 + }, + { + "epoch": 2.7774343122102008, + "grad_norm": 0.17465834204444608, + "learning_rate": 3.92445949412637e-05, + "loss": 0.3876, + "step": 1797 + }, + { + "epoch": 2.7789799072642967, + "grad_norm": 0.1813088489376211, + "learning_rate": 3.9201434518618826e-05, + "loss": 0.3988, + "step": 1798 + }, + { + "epoch": 2.7805255023183926, + "grad_norm": 0.18100302492637613, + "learning_rate": 3.915827502606576e-05, + "loss": 0.3802, + "step": 1799 + }, + { + "epoch": 2.7820710973724885, + "grad_norm": 0.19417838342895236, + "learning_rate": 3.911511651387253e-05, + "loss": 0.377, + "step": 1800 + }, + { + "epoch": 2.7836166924265844, + "grad_norm": 0.2573213340117975, + "learning_rate": 3.907195903230599e-05, + "loss": 0.3912, + "step": 1801 + }, + { + "epoch": 2.78516228748068, + "grad_norm": 0.284000471614278, + "learning_rate": 3.902880263163178e-05, + "loss": 0.3848, + "step": 1802 + }, + { + "epoch": 2.7867078825347757, + "grad_norm": 0.22616621719359473, + "learning_rate": 3.898564736211432e-05, + "loss": 0.4022, + "step": 1803 + }, + { + "epoch": 2.7882534775888717, + "grad_norm": 0.20526110287847993, + "learning_rate": 3.894249327401669e-05, + "loss": 0.397, + "step": 1804 + }, + { + "epoch": 2.7897990726429676, + "grad_norm": 0.17058234330188535, + "learning_rate": 3.889934041760058e-05, + "loss": 0.3926, + "step": 1805 + }, + { + "epoch": 2.7913446676970635, + "grad_norm": 0.22396294493426347, + "learning_rate": 3.88561888431263e-05, + "loss": 0.3968, + "step": 1806 + }, + { + "epoch": 2.7928902627511594, + "grad_norm": 0.18761021761388325, + "learning_rate": 3.88130386008526e-05, + "loss": 0.4027, + "step": 1807 + }, + { + "epoch": 2.7944358578052553, + "grad_norm": 0.16755062760736492, + "learning_rate": 3.87698897410367e-05, + "loss": 0.3972, + "step": 1808 + }, + { + "epoch": 2.7959814528593507, + "grad_norm": 0.24803898568626884, + "learning_rate": 3.872674231393425e-05, + "loss": 0.3876, + "step": 1809 + }, + { + "epoch": 2.7975270479134466, + "grad_norm": 0.2607238677140293, + "learning_rate": 3.868359636979919e-05, + "loss": 0.396, + "step": 1810 + }, + { + "epoch": 2.7990726429675425, + "grad_norm": 0.24051231824413313, + "learning_rate": 3.8640451958883705e-05, + "loss": 0.394, + "step": 1811 + }, + { + "epoch": 2.8006182380216385, + "grad_norm": 0.2198760157566278, + "learning_rate": 3.8597309131438285e-05, + "loss": 0.3929, + "step": 1812 + }, + { + "epoch": 2.802163833075734, + "grad_norm": 0.21247205512843312, + "learning_rate": 3.85541679377115e-05, + "loss": 0.39, + "step": 1813 + }, + { + "epoch": 2.80370942812983, + "grad_norm": 0.22974063721913382, + "learning_rate": 3.851102842795001e-05, + "loss": 0.3873, + "step": 1814 + }, + { + "epoch": 2.8052550231839257, + "grad_norm": 0.24422117496900178, + "learning_rate": 3.846789065239859e-05, + "loss": 0.3866, + "step": 1815 + }, + { + "epoch": 2.8068006182380216, + "grad_norm": 0.1879780673874634, + "learning_rate": 3.8424754661299926e-05, + "loss": 0.3915, + "step": 1816 + }, + { + "epoch": 2.8083462132921175, + "grad_norm": 0.28131521964208156, + "learning_rate": 3.8381620504894635e-05, + "loss": 0.3965, + "step": 1817 + }, + { + "epoch": 2.8098918083462134, + "grad_norm": 0.4062594396282389, + "learning_rate": 3.833848823342124e-05, + "loss": 0.3882, + "step": 1818 + }, + { + "epoch": 2.8114374034003093, + "grad_norm": 0.3737581895445919, + "learning_rate": 3.8295357897116015e-05, + "loss": 0.3889, + "step": 1819 + }, + { + "epoch": 2.812982998454405, + "grad_norm": 0.3180846973132513, + "learning_rate": 3.825222954621301e-05, + "loss": 0.4007, + "step": 1820 + }, + { + "epoch": 2.8145285935085007, + "grad_norm": 0.23924262459817525, + "learning_rate": 3.8209103230943955e-05, + "loss": 0.396, + "step": 1821 + }, + { + "epoch": 2.8160741885625966, + "grad_norm": 0.17186725441862907, + "learning_rate": 3.816597900153822e-05, + "loss": 0.3825, + "step": 1822 + }, + { + "epoch": 2.8176197836166925, + "grad_norm": 0.2019171990332553, + "learning_rate": 3.812285690822272e-05, + "loss": 0.3911, + "step": 1823 + }, + { + "epoch": 2.819165378670788, + "grad_norm": 0.17955446107446282, + "learning_rate": 3.8079737001221905e-05, + "loss": 0.3931, + "step": 1824 + }, + { + "epoch": 2.820710973724884, + "grad_norm": 0.20078132013690356, + "learning_rate": 3.803661933075768e-05, + "loss": 0.3899, + "step": 1825 + }, + { + "epoch": 2.82225656877898, + "grad_norm": 0.2281153173227666, + "learning_rate": 3.799350394704932e-05, + "loss": 0.3774, + "step": 1826 + }, + { + "epoch": 2.8238021638330757, + "grad_norm": 0.21250268042646084, + "learning_rate": 3.7950390900313455e-05, + "loss": 0.3952, + "step": 1827 + }, + { + "epoch": 2.8253477588871716, + "grad_norm": 0.17452701520242536, + "learning_rate": 3.790728024076401e-05, + "loss": 0.393, + "step": 1828 + }, + { + "epoch": 2.8268933539412675, + "grad_norm": 0.18226256585586875, + "learning_rate": 3.786417201861209e-05, + "loss": 0.3896, + "step": 1829 + }, + { + "epoch": 2.8284389489953634, + "grad_norm": 0.1699200232766844, + "learning_rate": 3.782106628406599e-05, + "loss": 0.3881, + "step": 1830 + }, + { + "epoch": 2.8299845440494593, + "grad_norm": 0.17347852322248702, + "learning_rate": 3.77779630873311e-05, + "loss": 0.3945, + "step": 1831 + }, + { + "epoch": 2.8315301391035548, + "grad_norm": 0.19679976534755758, + "learning_rate": 3.773486247860986e-05, + "loss": 0.3935, + "step": 1832 + }, + { + "epoch": 2.8330757341576507, + "grad_norm": 0.18755022647775246, + "learning_rate": 3.769176450810167e-05, + "loss": 0.3909, + "step": 1833 + }, + { + "epoch": 2.8346213292117466, + "grad_norm": 0.16961535389711938, + "learning_rate": 3.7648669226002904e-05, + "loss": 0.396, + "step": 1834 + }, + { + "epoch": 2.8361669242658425, + "grad_norm": 0.18048002267687868, + "learning_rate": 3.760557668250676e-05, + "loss": 0.3845, + "step": 1835 + }, + { + "epoch": 2.837712519319938, + "grad_norm": 0.1736727737721342, + "learning_rate": 3.756248692780325e-05, + "loss": 0.3993, + "step": 1836 + }, + { + "epoch": 2.839258114374034, + "grad_norm": 0.17581172468210995, + "learning_rate": 3.7519400012079203e-05, + "loss": 0.3921, + "step": 1837 + }, + { + "epoch": 2.8408037094281298, + "grad_norm": 0.1894157962541827, + "learning_rate": 3.747631598551804e-05, + "loss": 0.3949, + "step": 1838 + }, + { + "epoch": 2.8423493044822257, + "grad_norm": 0.1921294070326651, + "learning_rate": 3.743323489829988e-05, + "loss": 0.3856, + "step": 1839 + }, + { + "epoch": 2.8438948995363216, + "grad_norm": 0.1772283655136702, + "learning_rate": 3.739015680060141e-05, + "loss": 0.3907, + "step": 1840 + }, + { + "epoch": 2.8454404945904175, + "grad_norm": 0.22891639350098605, + "learning_rate": 3.734708174259585e-05, + "loss": 0.3873, + "step": 1841 + }, + { + "epoch": 2.8469860896445134, + "grad_norm": 0.17898775501928832, + "learning_rate": 3.730400977445282e-05, + "loss": 0.3833, + "step": 1842 + }, + { + "epoch": 2.848531684698609, + "grad_norm": 0.1838794786145881, + "learning_rate": 3.7260940946338416e-05, + "loss": 0.3836, + "step": 1843 + }, + { + "epoch": 2.8500772797527048, + "grad_norm": 0.23873368949862525, + "learning_rate": 3.7217875308415036e-05, + "loss": 0.3953, + "step": 1844 + }, + { + "epoch": 2.8516228748068007, + "grad_norm": 0.2037754497050098, + "learning_rate": 3.717481291084135e-05, + "loss": 0.3879, + "step": 1845 + }, + { + "epoch": 2.8531684698608966, + "grad_norm": 0.19247884951391822, + "learning_rate": 3.713175380377229e-05, + "loss": 0.3782, + "step": 1846 + }, + { + "epoch": 2.854714064914992, + "grad_norm": 0.3062130201916458, + "learning_rate": 3.708869803735895e-05, + "loss": 0.3905, + "step": 1847 + }, + { + "epoch": 2.856259659969088, + "grad_norm": 0.19763069985499787, + "learning_rate": 3.704564566174849e-05, + "loss": 0.3832, + "step": 1848 + }, + { + "epoch": 2.857805255023184, + "grad_norm": 0.18346313517919666, + "learning_rate": 3.700259672708417e-05, + "loss": 0.3867, + "step": 1849 + }, + { + "epoch": 2.8593508500772797, + "grad_norm": 0.1961417697891756, + "learning_rate": 3.695955128350524e-05, + "loss": 0.3923, + "step": 1850 + }, + { + "epoch": 2.8608964451313756, + "grad_norm": 0.20811812847397376, + "learning_rate": 3.691650938114684e-05, + "loss": 0.4045, + "step": 1851 + }, + { + "epoch": 2.8624420401854715, + "grad_norm": 0.19118895495750038, + "learning_rate": 3.6873471070140026e-05, + "loss": 0.3877, + "step": 1852 + }, + { + "epoch": 2.8639876352395675, + "grad_norm": 0.2446781595702205, + "learning_rate": 3.6830436400611657e-05, + "loss": 0.3968, + "step": 1853 + }, + { + "epoch": 2.865533230293663, + "grad_norm": 0.27637008537499236, + "learning_rate": 3.678740542268437e-05, + "loss": 0.4017, + "step": 1854 + }, + { + "epoch": 2.867078825347759, + "grad_norm": 0.1716813443429829, + "learning_rate": 3.674437818647645e-05, + "loss": 0.3945, + "step": 1855 + }, + { + "epoch": 2.8686244204018547, + "grad_norm": 0.2843102701296005, + "learning_rate": 3.67013547421019e-05, + "loss": 0.3903, + "step": 1856 + }, + { + "epoch": 2.8701700154559506, + "grad_norm": 0.3605604803916667, + "learning_rate": 3.665833513967024e-05, + "loss": 0.3914, + "step": 1857 + }, + { + "epoch": 2.871715610510046, + "grad_norm": 0.281028305368157, + "learning_rate": 3.6615319429286544e-05, + "loss": 0.3907, + "step": 1858 + }, + { + "epoch": 2.873261205564142, + "grad_norm": 0.3706718088832441, + "learning_rate": 3.657230766105136e-05, + "loss": 0.387, + "step": 1859 + }, + { + "epoch": 2.874806800618238, + "grad_norm": 0.3322333261169484, + "learning_rate": 3.6529299885060634e-05, + "loss": 0.3863, + "step": 1860 + }, + { + "epoch": 2.876352395672334, + "grad_norm": 0.32728465454278505, + "learning_rate": 3.648629615140565e-05, + "loss": 0.3792, + "step": 1861 + }, + { + "epoch": 2.8778979907264297, + "grad_norm": 0.4173074392701693, + "learning_rate": 3.6443296510173006e-05, + "loss": 0.3881, + "step": 1862 + }, + { + "epoch": 2.8794435857805256, + "grad_norm": 0.3376004538998364, + "learning_rate": 3.640030101144453e-05, + "loss": 0.3905, + "step": 1863 + }, + { + "epoch": 2.8809891808346215, + "grad_norm": 0.21414190118493256, + "learning_rate": 3.635730970529719e-05, + "loss": 0.3845, + "step": 1864 + }, + { + "epoch": 2.8825347758887174, + "grad_norm": 0.3558180588155539, + "learning_rate": 3.631432264180316e-05, + "loss": 0.3941, + "step": 1865 + }, + { + "epoch": 2.884080370942813, + "grad_norm": 0.39320314157434616, + "learning_rate": 3.627133987102956e-05, + "loss": 0.3904, + "step": 1866 + }, + { + "epoch": 2.885625965996909, + "grad_norm": 0.32800947779896955, + "learning_rate": 3.622836144303857e-05, + "loss": 0.3979, + "step": 1867 + }, + { + "epoch": 2.8871715610510047, + "grad_norm": 0.40344014390857996, + "learning_rate": 3.618538740788733e-05, + "loss": 0.3879, + "step": 1868 + }, + { + "epoch": 2.8887171561051006, + "grad_norm": 0.3092449027860355, + "learning_rate": 3.614241781562783e-05, + "loss": 0.4004, + "step": 1869 + }, + { + "epoch": 2.890262751159196, + "grad_norm": 0.3191402898079322, + "learning_rate": 3.609945271630687e-05, + "loss": 0.387, + "step": 1870 + }, + { + "epoch": 2.891808346213292, + "grad_norm": 0.3404716830049869, + "learning_rate": 3.6056492159966076e-05, + "loss": 0.3861, + "step": 1871 + }, + { + "epoch": 2.893353941267388, + "grad_norm": 0.23888696627268902, + "learning_rate": 3.601353619664173e-05, + "loss": 0.3842, + "step": 1872 + }, + { + "epoch": 2.894899536321484, + "grad_norm": 0.3112003266748598, + "learning_rate": 3.597058487636478e-05, + "loss": 0.3816, + "step": 1873 + }, + { + "epoch": 2.8964451313755797, + "grad_norm": 0.27628158618508, + "learning_rate": 3.592763824916078e-05, + "loss": 0.3889, + "step": 1874 + }, + { + "epoch": 2.8979907264296756, + "grad_norm": 0.21858501891722934, + "learning_rate": 3.58846963650498e-05, + "loss": 0.3929, + "step": 1875 + }, + { + "epoch": 2.8995363214837715, + "grad_norm": 0.20230449524397381, + "learning_rate": 3.584175927404641e-05, + "loss": 0.3843, + "step": 1876 + }, + { + "epoch": 2.901081916537867, + "grad_norm": 0.1996976045353686, + "learning_rate": 3.5798827026159536e-05, + "loss": 0.3846, + "step": 1877 + }, + { + "epoch": 2.902627511591963, + "grad_norm": 0.2539315574524483, + "learning_rate": 3.5755899671392546e-05, + "loss": 0.3863, + "step": 1878 + }, + { + "epoch": 2.9041731066460588, + "grad_norm": 0.27347730889148514, + "learning_rate": 3.5712977259743065e-05, + "loss": 0.3809, + "step": 1879 + }, + { + "epoch": 2.9057187017001547, + "grad_norm": 0.2670957397189431, + "learning_rate": 3.5670059841202945e-05, + "loss": 0.394, + "step": 1880 + }, + { + "epoch": 2.90726429675425, + "grad_norm": 0.32819821406555794, + "learning_rate": 3.5627147465758276e-05, + "loss": 0.3888, + "step": 1881 + }, + { + "epoch": 2.908809891808346, + "grad_norm": 0.34145982364945043, + "learning_rate": 3.558424018338922e-05, + "loss": 0.3899, + "step": 1882 + }, + { + "epoch": 2.910355486862442, + "grad_norm": 0.3096000913364954, + "learning_rate": 3.554133804407002e-05, + "loss": 0.3969, + "step": 1883 + }, + { + "epoch": 2.911901081916538, + "grad_norm": 0.34394438543226763, + "learning_rate": 3.549844109776897e-05, + "loss": 0.3908, + "step": 1884 + }, + { + "epoch": 2.9134466769706338, + "grad_norm": 0.28540515538147054, + "learning_rate": 3.5455549394448255e-05, + "loss": 0.3825, + "step": 1885 + }, + { + "epoch": 2.9149922720247297, + "grad_norm": 0.3004810072116905, + "learning_rate": 3.541266298406399e-05, + "loss": 0.3892, + "step": 1886 + }, + { + "epoch": 2.9165378670788256, + "grad_norm": 0.21089141561599284, + "learning_rate": 3.5369781916566115e-05, + "loss": 0.3884, + "step": 1887 + }, + { + "epoch": 2.918083462132921, + "grad_norm": 0.26942714783902477, + "learning_rate": 3.5326906241898365e-05, + "loss": 0.3984, + "step": 1888 + }, + { + "epoch": 2.919629057187017, + "grad_norm": 0.3442399952953209, + "learning_rate": 3.528403600999815e-05, + "loss": 0.3901, + "step": 1889 + }, + { + "epoch": 2.921174652241113, + "grad_norm": 0.3001915641726886, + "learning_rate": 3.5241171270796594e-05, + "loss": 0.404, + "step": 1890 + }, + { + "epoch": 2.9227202472952087, + "grad_norm": 0.3761387251745968, + "learning_rate": 3.519831207421839e-05, + "loss": 0.3971, + "step": 1891 + }, + { + "epoch": 2.9242658423493046, + "grad_norm": 0.29892816120859234, + "learning_rate": 3.5155458470181766e-05, + "loss": 0.4035, + "step": 1892 + }, + { + "epoch": 2.9258114374034, + "grad_norm": 0.3150415273440005, + "learning_rate": 3.51126105085985e-05, + "loss": 0.3884, + "step": 1893 + }, + { + "epoch": 2.927357032457496, + "grad_norm": 0.34568619548067764, + "learning_rate": 3.506976823937371e-05, + "loss": 0.388, + "step": 1894 + }, + { + "epoch": 2.928902627511592, + "grad_norm": 0.31895937702164295, + "learning_rate": 3.502693171240594e-05, + "loss": 0.3945, + "step": 1895 + }, + { + "epoch": 2.930448222565688, + "grad_norm": 0.2747416006092413, + "learning_rate": 3.498410097758703e-05, + "loss": 0.3884, + "step": 1896 + }, + { + "epoch": 2.9319938176197837, + "grad_norm": 0.36800883335054185, + "learning_rate": 3.4941276084802104e-05, + "loss": 0.3885, + "step": 1897 + }, + { + "epoch": 2.9335394126738796, + "grad_norm": 0.3740803469722144, + "learning_rate": 3.48984570839294e-05, + "loss": 0.3886, + "step": 1898 + }, + { + "epoch": 2.9350850077279755, + "grad_norm": 0.35763463357124636, + "learning_rate": 3.485564402484039e-05, + "loss": 0.3915, + "step": 1899 + }, + { + "epoch": 2.936630602782071, + "grad_norm": 0.25083149731119164, + "learning_rate": 3.481283695739959e-05, + "loss": 0.3965, + "step": 1900 + }, + { + "epoch": 2.938176197836167, + "grad_norm": 0.23201087393763298, + "learning_rate": 3.477003593146448e-05, + "loss": 0.3846, + "step": 1901 + }, + { + "epoch": 2.939721792890263, + "grad_norm": 0.26829347016684496, + "learning_rate": 3.4727240996885616e-05, + "loss": 0.3727, + "step": 1902 + }, + { + "epoch": 2.9412673879443587, + "grad_norm": 0.25541284401326064, + "learning_rate": 3.4684452203506366e-05, + "loss": 0.3936, + "step": 1903 + }, + { + "epoch": 2.942812982998454, + "grad_norm": 0.2881274195973795, + "learning_rate": 3.464166960116299e-05, + "loss": 0.3876, + "step": 1904 + }, + { + "epoch": 2.94435857805255, + "grad_norm": 0.22051538054770017, + "learning_rate": 3.459889323968449e-05, + "loss": 0.3869, + "step": 1905 + }, + { + "epoch": 2.945904173106646, + "grad_norm": 0.2531089166492334, + "learning_rate": 3.455612316889268e-05, + "loss": 0.3905, + "step": 1906 + }, + { + "epoch": 2.947449768160742, + "grad_norm": 0.26532487494222146, + "learning_rate": 3.4513359438601976e-05, + "loss": 0.3933, + "step": 1907 + }, + { + "epoch": 2.948995363214838, + "grad_norm": 0.24694988073087612, + "learning_rate": 3.4470602098619437e-05, + "loss": 0.3883, + "step": 1908 + }, + { + "epoch": 2.9505409582689337, + "grad_norm": 0.21275726411920523, + "learning_rate": 3.442785119874469e-05, + "loss": 0.3997, + "step": 1909 + }, + { + "epoch": 2.9520865533230296, + "grad_norm": 0.21271410672737606, + "learning_rate": 3.438510678876983e-05, + "loss": 0.3777, + "step": 1910 + }, + { + "epoch": 2.953632148377125, + "grad_norm": 0.2801489130994963, + "learning_rate": 3.434236891847941e-05, + "loss": 0.3876, + "step": 1911 + }, + { + "epoch": 2.955177743431221, + "grad_norm": 0.25117899176243486, + "learning_rate": 3.4299637637650374e-05, + "loss": 0.3878, + "step": 1912 + }, + { + "epoch": 2.956723338485317, + "grad_norm": 0.29618242491492724, + "learning_rate": 3.4256912996051996e-05, + "loss": 0.394, + "step": 1913 + }, + { + "epoch": 2.958268933539413, + "grad_norm": 0.18860005187869416, + "learning_rate": 3.421419504344578e-05, + "loss": 0.3948, + "step": 1914 + }, + { + "epoch": 2.9598145285935082, + "grad_norm": 0.21394031751835493, + "learning_rate": 3.417148382958549e-05, + "loss": 0.3974, + "step": 1915 + }, + { + "epoch": 2.961360123647604, + "grad_norm": 0.3149830463145957, + "learning_rate": 3.4128779404217004e-05, + "loss": 0.3893, + "step": 1916 + }, + { + "epoch": 2.9629057187017, + "grad_norm": 0.2104129047081311, + "learning_rate": 3.40860818170783e-05, + "loss": 0.3975, + "step": 1917 + }, + { + "epoch": 2.964451313755796, + "grad_norm": 0.22771617164783442, + "learning_rate": 3.404339111789941e-05, + "loss": 0.3776, + "step": 1918 + }, + { + "epoch": 2.965996908809892, + "grad_norm": 0.2884790016562261, + "learning_rate": 3.400070735640233e-05, + "loss": 0.3941, + "step": 1919 + }, + { + "epoch": 2.9675425038639878, + "grad_norm": 0.18913219534335085, + "learning_rate": 3.395803058230095e-05, + "loss": 0.3933, + "step": 1920 + }, + { + "epoch": 2.9690880989180837, + "grad_norm": 0.2384230589963729, + "learning_rate": 3.3915360845301085e-05, + "loss": 0.3852, + "step": 1921 + }, + { + "epoch": 2.970633693972179, + "grad_norm": 0.2633334096132983, + "learning_rate": 3.38726981951003e-05, + "loss": 0.3947, + "step": 1922 + }, + { + "epoch": 2.972179289026275, + "grad_norm": 0.20814728125813683, + "learning_rate": 3.3830042681387904e-05, + "loss": 0.39, + "step": 1923 + }, + { + "epoch": 2.973724884080371, + "grad_norm": 0.24367474670415484, + "learning_rate": 3.378739435384496e-05, + "loss": 0.3791, + "step": 1924 + }, + { + "epoch": 2.975270479134467, + "grad_norm": 0.2821750783223297, + "learning_rate": 3.374475326214408e-05, + "loss": 0.3889, + "step": 1925 + }, + { + "epoch": 2.9768160741885628, + "grad_norm": 0.27530067994928065, + "learning_rate": 3.370211945594947e-05, + "loss": 0.3902, + "step": 1926 + }, + { + "epoch": 2.978361669242658, + "grad_norm": 0.2720303840999367, + "learning_rate": 3.36594929849169e-05, + "loss": 0.3842, + "step": 1927 + }, + { + "epoch": 2.979907264296754, + "grad_norm": 0.19930232210254215, + "learning_rate": 3.3616873898693554e-05, + "loss": 0.3792, + "step": 1928 + }, + { + "epoch": 2.98145285935085, + "grad_norm": 0.36206977990301475, + "learning_rate": 3.3574262246917976e-05, + "loss": 0.4014, + "step": 1929 + }, + { + "epoch": 2.982998454404946, + "grad_norm": 0.21907688820073476, + "learning_rate": 3.3531658079220154e-05, + "loss": 0.3806, + "step": 1930 + }, + { + "epoch": 2.984544049459042, + "grad_norm": 0.23541598495245927, + "learning_rate": 3.348906144522127e-05, + "loss": 0.3932, + "step": 1931 + }, + { + "epoch": 2.9860896445131377, + "grad_norm": 0.2548801503252262, + "learning_rate": 3.344647239453379e-05, + "loss": 0.3869, + "step": 1932 + }, + { + "epoch": 2.9876352395672336, + "grad_norm": 0.22769350023934656, + "learning_rate": 3.340389097676126e-05, + "loss": 0.3839, + "step": 1933 + }, + { + "epoch": 2.989180834621329, + "grad_norm": 0.26547226925373146, + "learning_rate": 3.336131724149846e-05, + "loss": 0.3886, + "step": 1934 + }, + { + "epoch": 2.990726429675425, + "grad_norm": 0.3073755040234777, + "learning_rate": 3.3318751238331134e-05, + "loss": 0.3892, + "step": 1935 + }, + { + "epoch": 2.992272024729521, + "grad_norm": 0.23743365918197448, + "learning_rate": 3.327619301683603e-05, + "loss": 0.3868, + "step": 1936 + }, + { + "epoch": 2.993817619783617, + "grad_norm": 0.2129623169710384, + "learning_rate": 3.3233642626580874e-05, + "loss": 0.3827, + "step": 1937 + }, + { + "epoch": 2.9953632148377123, + "grad_norm": 0.19930478449122238, + "learning_rate": 3.3191100117124234e-05, + "loss": 0.3922, + "step": 1938 + }, + { + "epoch": 2.996908809891808, + "grad_norm": 0.2818998655532697, + "learning_rate": 3.31485655380155e-05, + "loss": 0.3969, + "step": 1939 + }, + { + "epoch": 2.998454404945904, + "grad_norm": 0.32138858700796147, + "learning_rate": 3.3106038938794854e-05, + "loss": 0.3918, + "step": 1940 + }, + { + "epoch": 3.0, + "grad_norm": 0.23813170146158089, + "learning_rate": 3.306352036899315e-05, + "loss": 0.3779, + "step": 1941 + }, + { + "epoch": 3.001545595054096, + "grad_norm": 0.31171465584784497, + "learning_rate": 3.30210098781319e-05, + "loss": 0.3679, + "step": 1942 + }, + { + "epoch": 3.003091190108192, + "grad_norm": 0.2869157618738416, + "learning_rate": 3.2978507515723235e-05, + "loss": 0.3756, + "step": 1943 + }, + { + "epoch": 3.0046367851622873, + "grad_norm": 0.20074930726465806, + "learning_rate": 3.2936013331269775e-05, + "loss": 0.3651, + "step": 1944 + }, + { + "epoch": 3.006182380216383, + "grad_norm": 0.2148586383545826, + "learning_rate": 3.289352737426464e-05, + "loss": 0.3655, + "step": 1945 + }, + { + "epoch": 3.007727975270479, + "grad_norm": 0.2288239820290731, + "learning_rate": 3.285104969419136e-05, + "loss": 0.3619, + "step": 1946 + }, + { + "epoch": 3.009273570324575, + "grad_norm": 0.24053953157264535, + "learning_rate": 3.280858034052385e-05, + "loss": 0.3636, + "step": 1947 + }, + { + "epoch": 3.010819165378671, + "grad_norm": 0.2697801971339192, + "learning_rate": 3.2766119362726264e-05, + "loss": 0.3689, + "step": 1948 + }, + { + "epoch": 3.012364760432767, + "grad_norm": 0.21726444358702826, + "learning_rate": 3.2723666810253076e-05, + "loss": 0.3634, + "step": 1949 + }, + { + "epoch": 3.0139103554868623, + "grad_norm": 0.294339356595773, + "learning_rate": 3.26812227325489e-05, + "loss": 0.3645, + "step": 1950 + }, + { + "epoch": 3.015455950540958, + "grad_norm": 0.2410788387087104, + "learning_rate": 3.263878717904847e-05, + "loss": 0.3757, + "step": 1951 + }, + { + "epoch": 3.017001545595054, + "grad_norm": 0.2751603563610299, + "learning_rate": 3.259636019917667e-05, + "loss": 0.3646, + "step": 1952 + }, + { + "epoch": 3.01854714064915, + "grad_norm": 0.17361571571114168, + "learning_rate": 3.255394184234829e-05, + "loss": 0.3613, + "step": 1953 + }, + { + "epoch": 3.020092735703246, + "grad_norm": 0.2241225829426769, + "learning_rate": 3.251153215796814e-05, + "loss": 0.3681, + "step": 1954 + }, + { + "epoch": 3.021638330757342, + "grad_norm": 0.2365359130191026, + "learning_rate": 3.2469131195430933e-05, + "loss": 0.3717, + "step": 1955 + }, + { + "epoch": 3.0231839258114372, + "grad_norm": 0.172259454607108, + "learning_rate": 3.2426739004121214e-05, + "loss": 0.3717, + "step": 1956 + }, + { + "epoch": 3.024729520865533, + "grad_norm": 0.27340648916561816, + "learning_rate": 3.238435563341326e-05, + "loss": 0.371, + "step": 1957 + }, + { + "epoch": 3.026275115919629, + "grad_norm": 0.2413768979772997, + "learning_rate": 3.234198113267118e-05, + "loss": 0.3661, + "step": 1958 + }, + { + "epoch": 3.027820710973725, + "grad_norm": 0.26528013813158247, + "learning_rate": 3.2299615551248655e-05, + "loss": 0.3707, + "step": 1959 + }, + { + "epoch": 3.029366306027821, + "grad_norm": 0.20935732880513103, + "learning_rate": 3.2257258938489045e-05, + "loss": 0.3633, + "step": 1960 + }, + { + "epoch": 3.0309119010819163, + "grad_norm": 0.20152247145768049, + "learning_rate": 3.221491134372518e-05, + "loss": 0.357, + "step": 1961 + }, + { + "epoch": 3.0324574961360122, + "grad_norm": 0.24715380895695596, + "learning_rate": 3.217257281627951e-05, + "loss": 0.3744, + "step": 1962 + }, + { + "epoch": 3.034003091190108, + "grad_norm": 0.17849954414560545, + "learning_rate": 3.2130243405463834e-05, + "loss": 0.3602, + "step": 1963 + }, + { + "epoch": 3.035548686244204, + "grad_norm": 0.24389877635398147, + "learning_rate": 3.208792316057931e-05, + "loss": 0.3606, + "step": 1964 + }, + { + "epoch": 3.0370942812983, + "grad_norm": 0.24691720825848798, + "learning_rate": 3.204561213091653e-05, + "loss": 0.357, + "step": 1965 + }, + { + "epoch": 3.038639876352396, + "grad_norm": 0.18331040793836595, + "learning_rate": 3.200331036575525e-05, + "loss": 0.3664, + "step": 1966 + }, + { + "epoch": 3.0401854714064913, + "grad_norm": 0.2769177026846328, + "learning_rate": 3.196101791436448e-05, + "loss": 0.3713, + "step": 1967 + }, + { + "epoch": 3.041731066460587, + "grad_norm": 0.1820364877136236, + "learning_rate": 3.191873482600239e-05, + "loss": 0.3654, + "step": 1968 + }, + { + "epoch": 3.043276661514683, + "grad_norm": 0.18851113345357065, + "learning_rate": 3.187646114991622e-05, + "loss": 0.3619, + "step": 1969 + }, + { + "epoch": 3.044822256568779, + "grad_norm": 0.17214646935687045, + "learning_rate": 3.1834196935342254e-05, + "loss": 0.368, + "step": 1970 + }, + { + "epoch": 3.046367851622875, + "grad_norm": 0.16470405943858682, + "learning_rate": 3.179194223150577e-05, + "loss": 0.373, + "step": 1971 + }, + { + "epoch": 3.047913446676971, + "grad_norm": 0.1776058167891048, + "learning_rate": 3.1749697087620954e-05, + "loss": 0.3704, + "step": 1972 + }, + { + "epoch": 3.0494590417310663, + "grad_norm": 0.18969738255695962, + "learning_rate": 3.1707461552890864e-05, + "loss": 0.3698, + "step": 1973 + }, + { + "epoch": 3.051004636785162, + "grad_norm": 0.1544017040360298, + "learning_rate": 3.166523567650737e-05, + "loss": 0.3644, + "step": 1974 + }, + { + "epoch": 3.052550231839258, + "grad_norm": 0.16631194905369284, + "learning_rate": 3.162301950765109e-05, + "loss": 0.3682, + "step": 1975 + }, + { + "epoch": 3.054095826893354, + "grad_norm": 0.1645757603835096, + "learning_rate": 3.1580813095491306e-05, + "loss": 0.3614, + "step": 1976 + }, + { + "epoch": 3.05564142194745, + "grad_norm": 0.16666207079870846, + "learning_rate": 3.1538616489186e-05, + "loss": 0.3654, + "step": 1977 + }, + { + "epoch": 3.0571870170015454, + "grad_norm": 0.23611859921736, + "learning_rate": 3.1496429737881685e-05, + "loss": 0.3596, + "step": 1978 + }, + { + "epoch": 3.0587326120556413, + "grad_norm": 0.16628564713517838, + "learning_rate": 3.145425289071339e-05, + "loss": 0.3599, + "step": 1979 + }, + { + "epoch": 3.060278207109737, + "grad_norm": 0.14586628384056954, + "learning_rate": 3.1412085996804665e-05, + "loss": 0.3668, + "step": 1980 + }, + { + "epoch": 3.061823802163833, + "grad_norm": 0.17648238084761203, + "learning_rate": 3.1369929105267397e-05, + "loss": 0.3706, + "step": 1981 + }, + { + "epoch": 3.063369397217929, + "grad_norm": 0.17339756474227566, + "learning_rate": 3.132778226520185e-05, + "loss": 0.355, + "step": 1982 + }, + { + "epoch": 3.064914992272025, + "grad_norm": 0.17418137231230973, + "learning_rate": 3.128564552569662e-05, + "loss": 0.3642, + "step": 1983 + }, + { + "epoch": 3.0664605873261204, + "grad_norm": 0.18383584794423377, + "learning_rate": 3.124351893582848e-05, + "loss": 0.3696, + "step": 1984 + }, + { + "epoch": 3.0680061823802163, + "grad_norm": 0.15943423486156244, + "learning_rate": 3.120140254466239e-05, + "loss": 0.3652, + "step": 1985 + }, + { + "epoch": 3.069551777434312, + "grad_norm": 0.13558986598558756, + "learning_rate": 3.11592964012515e-05, + "loss": 0.3744, + "step": 1986 + }, + { + "epoch": 3.071097372488408, + "grad_norm": 0.17240632669153214, + "learning_rate": 3.1117200554636944e-05, + "loss": 0.3723, + "step": 1987 + }, + { + "epoch": 3.072642967542504, + "grad_norm": 0.1298321730384647, + "learning_rate": 3.107511505384788e-05, + "loss": 0.3668, + "step": 1988 + }, + { + "epoch": 3.0741885625966, + "grad_norm": 0.1407898135550854, + "learning_rate": 3.1033039947901425e-05, + "loss": 0.3674, + "step": 1989 + }, + { + "epoch": 3.0757341576506954, + "grad_norm": 0.1451842599816864, + "learning_rate": 3.099097528580262e-05, + "loss": 0.357, + "step": 1990 + }, + { + "epoch": 3.0772797527047913, + "grad_norm": 0.15038363856026724, + "learning_rate": 3.0948921116544315e-05, + "loss": 0.3619, + "step": 1991 + }, + { + "epoch": 3.078825347758887, + "grad_norm": 0.15593874991279674, + "learning_rate": 3.090687748910709e-05, + "loss": 0.3739, + "step": 1992 + }, + { + "epoch": 3.080370942812983, + "grad_norm": 0.17760818245046628, + "learning_rate": 3.086484445245935e-05, + "loss": 0.3598, + "step": 1993 + }, + { + "epoch": 3.081916537867079, + "grad_norm": 0.16189925967772206, + "learning_rate": 3.082282205555709e-05, + "loss": 0.3683, + "step": 1994 + }, + { + "epoch": 3.0834621329211744, + "grad_norm": 0.16397882352854634, + "learning_rate": 3.078081034734393e-05, + "loss": 0.3653, + "step": 1995 + }, + { + "epoch": 3.0850077279752703, + "grad_norm": 0.16873160126894193, + "learning_rate": 3.073880937675104e-05, + "loss": 0.3707, + "step": 1996 + }, + { + "epoch": 3.0865533230293662, + "grad_norm": 0.4887966427758655, + "learning_rate": 3.0696819192697094e-05, + "loss": 0.3704, + "step": 1997 + }, + { + "epoch": 3.088098918083462, + "grad_norm": 0.16174679696377703, + "learning_rate": 3.065483984408819e-05, + "loss": 0.3672, + "step": 1998 + }, + { + "epoch": 3.089644513137558, + "grad_norm": 0.1614323899965815, + "learning_rate": 3.061287137981783e-05, + "loss": 0.3619, + "step": 1999 + }, + { + "epoch": 3.091190108191654, + "grad_norm": 0.16531952527462507, + "learning_rate": 3.05709138487668e-05, + "loss": 0.3608, + "step": 2000 + }, + { + "epoch": 3.0927357032457494, + "grad_norm": 0.1665889090936738, + "learning_rate": 3.052896729980318e-05, + "loss": 0.3628, + "step": 2001 + }, + { + "epoch": 3.0942812982998453, + "grad_norm": 0.17925722528029708, + "learning_rate": 3.0487031781782265e-05, + "loss": 0.3638, + "step": 2002 + }, + { + "epoch": 3.0958268933539412, + "grad_norm": 0.15251213414087034, + "learning_rate": 3.0445107343546474e-05, + "loss": 0.3719, + "step": 2003 + }, + { + "epoch": 3.097372488408037, + "grad_norm": 0.15268078066226246, + "learning_rate": 3.040319403392533e-05, + "loss": 0.3639, + "step": 2004 + }, + { + "epoch": 3.098918083462133, + "grad_norm": 0.1919213576284206, + "learning_rate": 3.0361291901735428e-05, + "loss": 0.3778, + "step": 2005 + }, + { + "epoch": 3.100463678516229, + "grad_norm": 0.1661459402079912, + "learning_rate": 3.0319400995780305e-05, + "loss": 0.3642, + "step": 2006 + }, + { + "epoch": 3.1020092735703244, + "grad_norm": 0.17391255717889237, + "learning_rate": 3.0277521364850426e-05, + "loss": 0.3732, + "step": 2007 + }, + { + "epoch": 3.1035548686244203, + "grad_norm": 0.15570388212781644, + "learning_rate": 3.0235653057723148e-05, + "loss": 0.3627, + "step": 2008 + }, + { + "epoch": 3.105100463678516, + "grad_norm": 0.1657068281966001, + "learning_rate": 3.0193796123162617e-05, + "loss": 0.365, + "step": 2009 + }, + { + "epoch": 3.106646058732612, + "grad_norm": 0.1644518030320822, + "learning_rate": 3.0151950609919733e-05, + "loss": 0.3595, + "step": 2010 + }, + { + "epoch": 3.108191653786708, + "grad_norm": 0.15149853097947524, + "learning_rate": 3.0110116566732132e-05, + "loss": 0.3613, + "step": 2011 + }, + { + "epoch": 3.109737248840804, + "grad_norm": 0.20400354964088088, + "learning_rate": 3.006829404232404e-05, + "loss": 0.3682, + "step": 2012 + }, + { + "epoch": 3.1112828438948994, + "grad_norm": 0.1591073786268792, + "learning_rate": 3.0026483085406283e-05, + "loss": 0.3732, + "step": 2013 + }, + { + "epoch": 3.1128284389489953, + "grad_norm": 0.1570119612980249, + "learning_rate": 2.9984683744676206e-05, + "loss": 0.3659, + "step": 2014 + }, + { + "epoch": 3.114374034003091, + "grad_norm": 0.17084269419017303, + "learning_rate": 2.994289606881769e-05, + "loss": 0.3738, + "step": 2015 + }, + { + "epoch": 3.115919629057187, + "grad_norm": 0.14029104939344528, + "learning_rate": 2.9901120106500924e-05, + "loss": 0.3612, + "step": 2016 + }, + { + "epoch": 3.117465224111283, + "grad_norm": 0.1589995205079017, + "learning_rate": 2.9859355906382517e-05, + "loss": 0.3613, + "step": 2017 + }, + { + "epoch": 3.1190108191653785, + "grad_norm": 0.16915103629839312, + "learning_rate": 2.98176035171054e-05, + "loss": 0.373, + "step": 2018 + }, + { + "epoch": 3.1205564142194744, + "grad_norm": 0.13703361557235325, + "learning_rate": 2.97758629872987e-05, + "loss": 0.359, + "step": 2019 + }, + { + "epoch": 3.1221020092735703, + "grad_norm": 0.22441100620131804, + "learning_rate": 2.9734134365577723e-05, + "loss": 0.3694, + "step": 2020 + }, + { + "epoch": 3.123647604327666, + "grad_norm": 0.1478916041558205, + "learning_rate": 2.9692417700543966e-05, + "loss": 0.3638, + "step": 2021 + }, + { + "epoch": 3.125193199381762, + "grad_norm": 0.2096432660700668, + "learning_rate": 2.9650713040784954e-05, + "loss": 0.3644, + "step": 2022 + }, + { + "epoch": 3.126738794435858, + "grad_norm": 0.16526543441751307, + "learning_rate": 2.960902043487423e-05, + "loss": 0.3618, + "step": 2023 + }, + { + "epoch": 3.1282843894899535, + "grad_norm": 0.17324589785159916, + "learning_rate": 2.9567339931371327e-05, + "loss": 0.3743, + "step": 2024 + }, + { + "epoch": 3.1298299845440494, + "grad_norm": 0.1389638279469882, + "learning_rate": 2.9525671578821647e-05, + "loss": 0.3629, + "step": 2025 + }, + { + "epoch": 3.1313755795981453, + "grad_norm": 0.18503881736460734, + "learning_rate": 2.9484015425756454e-05, + "loss": 0.374, + "step": 2026 + }, + { + "epoch": 3.132921174652241, + "grad_norm": 0.12587601505049495, + "learning_rate": 2.9442371520692825e-05, + "loss": 0.3579, + "step": 2027 + }, + { + "epoch": 3.134466769706337, + "grad_norm": 0.1654549667985164, + "learning_rate": 2.9400739912133543e-05, + "loss": 0.3711, + "step": 2028 + }, + { + "epoch": 3.1360123647604325, + "grad_norm": 0.13126514642819836, + "learning_rate": 2.9359120648567062e-05, + "loss": 0.3665, + "step": 2029 + }, + { + "epoch": 3.1375579598145285, + "grad_norm": 0.1433166542494949, + "learning_rate": 2.9317513778467498e-05, + "loss": 0.3651, + "step": 2030 + }, + { + "epoch": 3.1391035548686244, + "grad_norm": 0.14785582958072702, + "learning_rate": 2.9275919350294498e-05, + "loss": 0.3599, + "step": 2031 + }, + { + "epoch": 3.1406491499227203, + "grad_norm": 0.1627068357890892, + "learning_rate": 2.9234337412493218e-05, + "loss": 0.372, + "step": 2032 + }, + { + "epoch": 3.142194744976816, + "grad_norm": 0.15647817397173375, + "learning_rate": 2.9192768013494297e-05, + "loss": 0.3666, + "step": 2033 + }, + { + "epoch": 3.143740340030912, + "grad_norm": 0.14213938932745243, + "learning_rate": 2.9151211201713744e-05, + "loss": 0.3661, + "step": 2034 + }, + { + "epoch": 3.1452859350850075, + "grad_norm": 0.13130075114364653, + "learning_rate": 2.91096670255529e-05, + "loss": 0.3577, + "step": 2035 + }, + { + "epoch": 3.1468315301391034, + "grad_norm": 0.14255253995836262, + "learning_rate": 2.9068135533398415e-05, + "loss": 0.3575, + "step": 2036 + }, + { + "epoch": 3.1483771251931993, + "grad_norm": 0.14334987938477375, + "learning_rate": 2.902661677362215e-05, + "loss": 0.3687, + "step": 2037 + }, + { + "epoch": 3.1499227202472952, + "grad_norm": 0.1493354947333077, + "learning_rate": 2.898511079458113e-05, + "loss": 0.3758, + "step": 2038 + }, + { + "epoch": 3.151468315301391, + "grad_norm": 0.12377396402553244, + "learning_rate": 2.8943617644617536e-05, + "loss": 0.368, + "step": 2039 + }, + { + "epoch": 3.153013910355487, + "grad_norm": 0.14593946865969992, + "learning_rate": 2.890213737205854e-05, + "loss": 0.3654, + "step": 2040 + }, + { + "epoch": 3.1545595054095825, + "grad_norm": 0.13530777880989286, + "learning_rate": 2.886067002521637e-05, + "loss": 0.3715, + "step": 2041 + }, + { + "epoch": 3.1561051004636784, + "grad_norm": 0.14577397239207665, + "learning_rate": 2.8819215652388162e-05, + "loss": 0.3669, + "step": 2042 + }, + { + "epoch": 3.1576506955177743, + "grad_norm": 0.15068942998498627, + "learning_rate": 2.877777430185599e-05, + "loss": 0.3603, + "step": 2043 + }, + { + "epoch": 3.1591962905718702, + "grad_norm": 0.1403873402782044, + "learning_rate": 2.873634602188671e-05, + "loss": 0.3618, + "step": 2044 + }, + { + "epoch": 3.160741885625966, + "grad_norm": 0.14312343061212757, + "learning_rate": 2.8694930860731954e-05, + "loss": 0.3609, + "step": 2045 + }, + { + "epoch": 3.162287480680062, + "grad_norm": 0.12439940271897412, + "learning_rate": 2.8653528866628132e-05, + "loss": 0.366, + "step": 2046 + }, + { + "epoch": 3.1638330757341575, + "grad_norm": 0.15363023129281814, + "learning_rate": 2.8612140087796282e-05, + "loss": 0.3599, + "step": 2047 + }, + { + "epoch": 3.1653786707882534, + "grad_norm": 0.15970831874172828, + "learning_rate": 2.8570764572442012e-05, + "loss": 0.3708, + "step": 2048 + }, + { + "epoch": 3.1669242658423493, + "grad_norm": 0.14084145298957124, + "learning_rate": 2.852940236875556e-05, + "loss": 0.3683, + "step": 2049 + }, + { + "epoch": 3.1684698608964452, + "grad_norm": 0.17499188472712682, + "learning_rate": 2.848805352491161e-05, + "loss": 0.361, + "step": 2050 + }, + { + "epoch": 3.170015455950541, + "grad_norm": 0.14425219613022813, + "learning_rate": 2.8446718089069264e-05, + "loss": 0.37, + "step": 2051 + }, + { + "epoch": 3.1715610510046366, + "grad_norm": 0.1251859772791234, + "learning_rate": 2.840539610937208e-05, + "loss": 0.3623, + "step": 2052 + }, + { + "epoch": 3.1731066460587325, + "grad_norm": 0.12579663061900387, + "learning_rate": 2.8364087633947887e-05, + "loss": 0.3583, + "step": 2053 + }, + { + "epoch": 3.1746522411128284, + "grad_norm": 0.12251358173410634, + "learning_rate": 2.8322792710908787e-05, + "loss": 0.3623, + "step": 2054 + }, + { + "epoch": 3.1761978361669243, + "grad_norm": 0.13829595686979887, + "learning_rate": 2.828151138835114e-05, + "loss": 0.3613, + "step": 2055 + }, + { + "epoch": 3.17774343122102, + "grad_norm": 0.1381224917859969, + "learning_rate": 2.8240243714355416e-05, + "loss": 0.3658, + "step": 2056 + }, + { + "epoch": 3.179289026275116, + "grad_norm": 0.14834068382748958, + "learning_rate": 2.8198989736986212e-05, + "loss": 0.3774, + "step": 2057 + }, + { + "epoch": 3.1808346213292116, + "grad_norm": 0.12234098986699041, + "learning_rate": 2.8157749504292185e-05, + "loss": 0.3632, + "step": 2058 + }, + { + "epoch": 3.1823802163833075, + "grad_norm": 0.14951922074387244, + "learning_rate": 2.811652306430595e-05, + "loss": 0.3695, + "step": 2059 + }, + { + "epoch": 3.1839258114374034, + "grad_norm": 0.12717372719527187, + "learning_rate": 2.8075310465044076e-05, + "loss": 0.368, + "step": 2060 + }, + { + "epoch": 3.1854714064914993, + "grad_norm": 0.13522718209919893, + "learning_rate": 2.8034111754507032e-05, + "loss": 0.368, + "step": 2061 + }, + { + "epoch": 3.187017001545595, + "grad_norm": 0.12239250181348509, + "learning_rate": 2.799292698067907e-05, + "loss": 0.3636, + "step": 2062 + }, + { + "epoch": 3.1885625965996907, + "grad_norm": 0.13502088259022943, + "learning_rate": 2.795175619152824e-05, + "loss": 0.3646, + "step": 2063 + }, + { + "epoch": 3.1901081916537866, + "grad_norm": 0.12908259748447184, + "learning_rate": 2.7910599435006297e-05, + "loss": 0.3724, + "step": 2064 + }, + { + "epoch": 3.1916537867078825, + "grad_norm": 0.12814919445837764, + "learning_rate": 2.7869456759048644e-05, + "loss": 0.3683, + "step": 2065 + }, + { + "epoch": 3.1931993817619784, + "grad_norm": 0.14163890351082115, + "learning_rate": 2.7828328211574277e-05, + "loss": 0.3675, + "step": 2066 + }, + { + "epoch": 3.1947449768160743, + "grad_norm": 0.1372701812449613, + "learning_rate": 2.778721384048579e-05, + "loss": 0.3705, + "step": 2067 + }, + { + "epoch": 3.19629057187017, + "grad_norm": 0.15204661234673036, + "learning_rate": 2.7746113693669186e-05, + "loss": 0.3674, + "step": 2068 + }, + { + "epoch": 3.1978361669242656, + "grad_norm": 0.15115300068077478, + "learning_rate": 2.7705027818993958e-05, + "loss": 0.3729, + "step": 2069 + }, + { + "epoch": 3.1993817619783615, + "grad_norm": 0.14370368896428348, + "learning_rate": 2.7663956264312943e-05, + "loss": 0.367, + "step": 2070 + }, + { + "epoch": 3.2009273570324575, + "grad_norm": 0.14968474740202153, + "learning_rate": 2.7622899077462335e-05, + "loss": 0.3713, + "step": 2071 + }, + { + "epoch": 3.2024729520865534, + "grad_norm": 0.1686462967770937, + "learning_rate": 2.758185630626156e-05, + "loss": 0.3684, + "step": 2072 + }, + { + "epoch": 3.2040185471406493, + "grad_norm": 0.1536991268040143, + "learning_rate": 2.7540827998513263e-05, + "loss": 0.3653, + "step": 2073 + }, + { + "epoch": 3.205564142194745, + "grad_norm": 0.14785756813895054, + "learning_rate": 2.7499814202003272e-05, + "loss": 0.3672, + "step": 2074 + }, + { + "epoch": 3.2071097372488406, + "grad_norm": 0.16013222543831185, + "learning_rate": 2.7458814964500473e-05, + "loss": 0.3689, + "step": 2075 + }, + { + "epoch": 3.2086553323029365, + "grad_norm": 0.1534977441030092, + "learning_rate": 2.741783033375679e-05, + "loss": 0.3704, + "step": 2076 + }, + { + "epoch": 3.2102009273570324, + "grad_norm": 0.17365777916889796, + "learning_rate": 2.73768603575072e-05, + "loss": 0.3659, + "step": 2077 + }, + { + "epoch": 3.2117465224111283, + "grad_norm": 0.14489892162203102, + "learning_rate": 2.7335905083469555e-05, + "loss": 0.3658, + "step": 2078 + }, + { + "epoch": 3.2132921174652243, + "grad_norm": 0.1777579032834708, + "learning_rate": 2.7294964559344563e-05, + "loss": 0.3678, + "step": 2079 + }, + { + "epoch": 3.21483771251932, + "grad_norm": 0.17040592808564636, + "learning_rate": 2.725403883281583e-05, + "loss": 0.3642, + "step": 2080 + }, + { + "epoch": 3.2163833075734156, + "grad_norm": 0.1881164082243543, + "learning_rate": 2.721312795154966e-05, + "loss": 0.3731, + "step": 2081 + }, + { + "epoch": 3.2179289026275115, + "grad_norm": 0.18418639695158592, + "learning_rate": 2.7172231963195085e-05, + "loss": 0.3681, + "step": 2082 + }, + { + "epoch": 3.2194744976816074, + "grad_norm": 0.2082446090608705, + "learning_rate": 2.713135091538382e-05, + "loss": 0.3653, + "step": 2083 + }, + { + "epoch": 3.2210200927357033, + "grad_norm": 0.16220777622045765, + "learning_rate": 2.7090484855730136e-05, + "loss": 0.3738, + "step": 2084 + }, + { + "epoch": 3.2225656877897992, + "grad_norm": 0.15783064457328178, + "learning_rate": 2.7049633831830867e-05, + "loss": 0.3643, + "step": 2085 + }, + { + "epoch": 3.2241112828438947, + "grad_norm": 0.15693558187552553, + "learning_rate": 2.7008797891265336e-05, + "loss": 0.3704, + "step": 2086 + }, + { + "epoch": 3.2256568778979906, + "grad_norm": 0.14758748067377753, + "learning_rate": 2.6967977081595304e-05, + "loss": 0.3652, + "step": 2087 + }, + { + "epoch": 3.2272024729520865, + "grad_norm": 0.16748703002722784, + "learning_rate": 2.692717145036489e-05, + "loss": 0.3653, + "step": 2088 + }, + { + "epoch": 3.2287480680061824, + "grad_norm": 0.15125269311815615, + "learning_rate": 2.6886381045100552e-05, + "loss": 0.3656, + "step": 2089 + }, + { + "epoch": 3.2302936630602783, + "grad_norm": 0.15544533844342912, + "learning_rate": 2.6845605913311012e-05, + "loss": 0.3682, + "step": 2090 + }, + { + "epoch": 3.2318392581143742, + "grad_norm": 0.1562434423567851, + "learning_rate": 2.6804846102487182e-05, + "loss": 0.3717, + "step": 2091 + }, + { + "epoch": 3.2333848531684697, + "grad_norm": 0.14291583957621115, + "learning_rate": 2.676410166010217e-05, + "loss": 0.3767, + "step": 2092 + }, + { + "epoch": 3.2349304482225656, + "grad_norm": 0.1357194755115203, + "learning_rate": 2.6723372633611147e-05, + "loss": 0.3651, + "step": 2093 + }, + { + "epoch": 3.2364760432766615, + "grad_norm": 0.17378316610340208, + "learning_rate": 2.6682659070451338e-05, + "loss": 0.3632, + "step": 2094 + }, + { + "epoch": 3.2380216383307574, + "grad_norm": 0.12682019037929812, + "learning_rate": 2.6641961018041978e-05, + "loss": 0.3672, + "step": 2095 + }, + { + "epoch": 3.2395672333848533, + "grad_norm": 0.17406933720096976, + "learning_rate": 2.6601278523784214e-05, + "loss": 0.368, + "step": 2096 + }, + { + "epoch": 3.2411128284389488, + "grad_norm": 0.14091612975788875, + "learning_rate": 2.6560611635061083e-05, + "loss": 0.3734, + "step": 2097 + }, + { + "epoch": 3.2426584234930447, + "grad_norm": 0.16253732307206356, + "learning_rate": 2.6519960399237434e-05, + "loss": 0.3733, + "step": 2098 + }, + { + "epoch": 3.2442040185471406, + "grad_norm": 0.15080109666228644, + "learning_rate": 2.6479324863659897e-05, + "loss": 0.3675, + "step": 2099 + }, + { + "epoch": 3.2457496136012365, + "grad_norm": 0.15434815098317825, + "learning_rate": 2.6438705075656822e-05, + "loss": 0.3657, + "step": 2100 + }, + { + "epoch": 3.2472952086553324, + "grad_norm": 0.15767668819755418, + "learning_rate": 2.6398101082538187e-05, + "loss": 0.3838, + "step": 2101 + }, + { + "epoch": 3.2488408037094283, + "grad_norm": 0.13494582357863177, + "learning_rate": 2.6357512931595634e-05, + "loss": 0.3638, + "step": 2102 + }, + { + "epoch": 3.250386398763524, + "grad_norm": 0.15718928860212597, + "learning_rate": 2.631694067010228e-05, + "loss": 0.3752, + "step": 2103 + }, + { + "epoch": 3.2519319938176197, + "grad_norm": 0.1535896044874011, + "learning_rate": 2.6276384345312772e-05, + "loss": 0.3688, + "step": 2104 + }, + { + "epoch": 3.2534775888717156, + "grad_norm": 0.14398376836227514, + "learning_rate": 2.6235844004463217e-05, + "loss": 0.3624, + "step": 2105 + }, + { + "epoch": 3.2550231839258115, + "grad_norm": 0.12886124949716404, + "learning_rate": 2.619531969477108e-05, + "loss": 0.3604, + "step": 2106 + }, + { + "epoch": 3.2565687789799074, + "grad_norm": 0.20057890482614704, + "learning_rate": 2.6154811463435116e-05, + "loss": 0.3754, + "step": 2107 + }, + { + "epoch": 3.2581143740340033, + "grad_norm": 0.13665360927560344, + "learning_rate": 2.6114319357635447e-05, + "loss": 0.3701, + "step": 2108 + }, + { + "epoch": 3.2596599690880987, + "grad_norm": 0.16106687173917297, + "learning_rate": 2.6073843424533332e-05, + "loss": 0.3626, + "step": 2109 + }, + { + "epoch": 3.2612055641421946, + "grad_norm": 0.13092642496388196, + "learning_rate": 2.6033383711271222e-05, + "loss": 0.3602, + "step": 2110 + }, + { + "epoch": 3.2627511591962906, + "grad_norm": 0.15358668926841232, + "learning_rate": 2.599294026497269e-05, + "loss": 0.3643, + "step": 2111 + }, + { + "epoch": 3.2642967542503865, + "grad_norm": 0.14606690153826582, + "learning_rate": 2.5952513132742346e-05, + "loss": 0.3658, + "step": 2112 + }, + { + "epoch": 3.2658423493044824, + "grad_norm": 0.1378753572686257, + "learning_rate": 2.5912102361665793e-05, + "loss": 0.37, + "step": 2113 + }, + { + "epoch": 3.2673879443585783, + "grad_norm": 0.14687750965345156, + "learning_rate": 2.5871707998809594e-05, + "loss": 0.365, + "step": 2114 + }, + { + "epoch": 3.2689335394126737, + "grad_norm": 0.14819762397571287, + "learning_rate": 2.5831330091221203e-05, + "loss": 0.3797, + "step": 2115 + }, + { + "epoch": 3.2704791344667696, + "grad_norm": 0.12710955138828492, + "learning_rate": 2.5790968685928887e-05, + "loss": 0.3732, + "step": 2116 + }, + { + "epoch": 3.2720247295208655, + "grad_norm": 0.15343239107124537, + "learning_rate": 2.575062382994172e-05, + "loss": 0.3739, + "step": 2117 + }, + { + "epoch": 3.2735703245749614, + "grad_norm": 0.1416408721693487, + "learning_rate": 2.5710295570249485e-05, + "loss": 0.3694, + "step": 2118 + }, + { + "epoch": 3.2751159196290573, + "grad_norm": 0.13009832790542777, + "learning_rate": 2.5669983953822616e-05, + "loss": 0.3742, + "step": 2119 + }, + { + "epoch": 3.276661514683153, + "grad_norm": 0.14417898261561116, + "learning_rate": 2.5629689027612213e-05, + "loss": 0.3711, + "step": 2120 + }, + { + "epoch": 3.2782071097372487, + "grad_norm": 0.13507266005557989, + "learning_rate": 2.5589410838549902e-05, + "loss": 0.3756, + "step": 2121 + }, + { + "epoch": 3.2797527047913446, + "grad_norm": 0.1650121483203225, + "learning_rate": 2.5549149433547803e-05, + "loss": 0.3703, + "step": 2122 + }, + { + "epoch": 3.2812982998454405, + "grad_norm": 0.14953876290809184, + "learning_rate": 2.5508904859498528e-05, + "loss": 0.3604, + "step": 2123 + }, + { + "epoch": 3.2828438948995364, + "grad_norm": 0.14221545791884788, + "learning_rate": 2.5468677163275056e-05, + "loss": 0.3672, + "step": 2124 + }, + { + "epoch": 3.2843894899536323, + "grad_norm": 0.14399205076304955, + "learning_rate": 2.542846639173071e-05, + "loss": 0.3681, + "step": 2125 + }, + { + "epoch": 3.285935085007728, + "grad_norm": 0.14904816147074224, + "learning_rate": 2.538827259169911e-05, + "loss": 0.3717, + "step": 2126 + }, + { + "epoch": 3.2874806800618237, + "grad_norm": 0.14830345453722532, + "learning_rate": 2.5348095809994113e-05, + "loss": 0.3694, + "step": 2127 + }, + { + "epoch": 3.2890262751159196, + "grad_norm": 0.13985328837797856, + "learning_rate": 2.530793609340974e-05, + "loss": 0.3779, + "step": 2128 + }, + { + "epoch": 3.2905718701700155, + "grad_norm": 0.17710821167389063, + "learning_rate": 2.526779348872013e-05, + "loss": 0.366, + "step": 2129 + }, + { + "epoch": 3.2921174652241114, + "grad_norm": 0.14343141353829914, + "learning_rate": 2.522766804267955e-05, + "loss": 0.3699, + "step": 2130 + }, + { + "epoch": 3.293663060278207, + "grad_norm": 0.19980441120996223, + "learning_rate": 2.5187559802022196e-05, + "loss": 0.3693, + "step": 2131 + }, + { + "epoch": 3.295208655332303, + "grad_norm": 0.15717682802009292, + "learning_rate": 2.514746881346227e-05, + "loss": 0.3682, + "step": 2132 + }, + { + "epoch": 3.2967542503863987, + "grad_norm": 0.1886639544522163, + "learning_rate": 2.5107395123693908e-05, + "loss": 0.3641, + "step": 2133 + }, + { + "epoch": 3.2982998454404946, + "grad_norm": 0.17588396501052606, + "learning_rate": 2.5067338779391044e-05, + "loss": 0.3603, + "step": 2134 + }, + { + "epoch": 3.2998454404945905, + "grad_norm": 0.17319305501363041, + "learning_rate": 2.502729982720742e-05, + "loss": 0.3762, + "step": 2135 + }, + { + "epoch": 3.3013910355486864, + "grad_norm": 0.17074748630118594, + "learning_rate": 2.498727831377656e-05, + "loss": 0.3734, + "step": 2136 + }, + { + "epoch": 3.3029366306027823, + "grad_norm": 0.13590596256129128, + "learning_rate": 2.4947274285711657e-05, + "loss": 0.3637, + "step": 2137 + }, + { + "epoch": 3.3044822256568778, + "grad_norm": 0.18090013616044767, + "learning_rate": 2.4907287789605485e-05, + "loss": 0.3706, + "step": 2138 + }, + { + "epoch": 3.3060278207109737, + "grad_norm": 0.144784935494973, + "learning_rate": 2.4867318872030503e-05, + "loss": 0.373, + "step": 2139 + }, + { + "epoch": 3.3075734157650696, + "grad_norm": 0.1891658448891262, + "learning_rate": 2.482736757953861e-05, + "loss": 0.3631, + "step": 2140 + }, + { + "epoch": 3.3091190108191655, + "grad_norm": 0.14667963718248692, + "learning_rate": 2.478743395866121e-05, + "loss": 0.3708, + "step": 2141 + }, + { + "epoch": 3.3106646058732614, + "grad_norm": 0.21035594167766744, + "learning_rate": 2.4747518055909127e-05, + "loss": 0.377, + "step": 2142 + }, + { + "epoch": 3.312210200927357, + "grad_norm": 0.16983171943181535, + "learning_rate": 2.4707619917772547e-05, + "loss": 0.3666, + "step": 2143 + }, + { + "epoch": 3.3137557959814528, + "grad_norm": 0.20078565648335242, + "learning_rate": 2.466773959072094e-05, + "loss": 0.3706, + "step": 2144 + }, + { + "epoch": 3.3153013910355487, + "grad_norm": 0.1877400727983853, + "learning_rate": 2.4627877121203082e-05, + "loss": 0.3703, + "step": 2145 + }, + { + "epoch": 3.3168469860896446, + "grad_norm": 0.16510685650538093, + "learning_rate": 2.4588032555646905e-05, + "loss": 0.3726, + "step": 2146 + }, + { + "epoch": 3.3183925811437405, + "grad_norm": 0.1528367488160912, + "learning_rate": 2.4548205940459502e-05, + "loss": 0.365, + "step": 2147 + }, + { + "epoch": 3.3199381761978364, + "grad_norm": 0.15434010432237918, + "learning_rate": 2.450839732202707e-05, + "loss": 0.3642, + "step": 2148 + }, + { + "epoch": 3.321483771251932, + "grad_norm": 0.1552317292573876, + "learning_rate": 2.4468606746714835e-05, + "loss": 0.3656, + "step": 2149 + }, + { + "epoch": 3.3230293663060277, + "grad_norm": 0.1575056447538431, + "learning_rate": 2.4428834260867e-05, + "loss": 0.3735, + "step": 2150 + }, + { + "epoch": 3.3245749613601236, + "grad_norm": 0.1623900587467734, + "learning_rate": 2.4389079910806728e-05, + "loss": 0.3677, + "step": 2151 + }, + { + "epoch": 3.3261205564142196, + "grad_norm": 0.1321736194562474, + "learning_rate": 2.4349343742836028e-05, + "loss": 0.3663, + "step": 2152 + }, + { + "epoch": 3.3276661514683155, + "grad_norm": 0.126420491100854, + "learning_rate": 2.4309625803235744e-05, + "loss": 0.3665, + "step": 2153 + }, + { + "epoch": 3.329211746522411, + "grad_norm": 0.12932422188472273, + "learning_rate": 2.426992613826549e-05, + "loss": 0.362, + "step": 2154 + }, + { + "epoch": 3.330757341576507, + "grad_norm": 0.1373101315845367, + "learning_rate": 2.42302447941636e-05, + "loss": 0.3698, + "step": 2155 + }, + { + "epoch": 3.3323029366306027, + "grad_norm": 0.16669707516861867, + "learning_rate": 2.4190581817147063e-05, + "loss": 0.358, + "step": 2156 + }, + { + "epoch": 3.3338485316846986, + "grad_norm": 0.13589432515355804, + "learning_rate": 2.415093725341147e-05, + "loss": 0.3626, + "step": 2157 + }, + { + "epoch": 3.3353941267387945, + "grad_norm": 0.16452255693638304, + "learning_rate": 2.4111311149130993e-05, + "loss": 0.3535, + "step": 2158 + }, + { + "epoch": 3.3369397217928904, + "grad_norm": 0.14103684315072043, + "learning_rate": 2.4071703550458273e-05, + "loss": 0.374, + "step": 2159 + }, + { + "epoch": 3.338485316846986, + "grad_norm": 0.1466722795459836, + "learning_rate": 2.403211450352439e-05, + "loss": 0.3605, + "step": 2160 + }, + { + "epoch": 3.340030911901082, + "grad_norm": 0.14758435816430387, + "learning_rate": 2.3992544054438877e-05, + "loss": 0.3643, + "step": 2161 + }, + { + "epoch": 3.3415765069551777, + "grad_norm": 0.14523145992398567, + "learning_rate": 2.3952992249289542e-05, + "loss": 0.3682, + "step": 2162 + }, + { + "epoch": 3.3431221020092736, + "grad_norm": 0.14600152871059888, + "learning_rate": 2.3913459134142482e-05, + "loss": 0.3656, + "step": 2163 + }, + { + "epoch": 3.3446676970633695, + "grad_norm": 0.1205791485784816, + "learning_rate": 2.3873944755042088e-05, + "loss": 0.364, + "step": 2164 + }, + { + "epoch": 3.346213292117465, + "grad_norm": 0.1429536650084363, + "learning_rate": 2.3834449158010866e-05, + "loss": 0.3645, + "step": 2165 + }, + { + "epoch": 3.347758887171561, + "grad_norm": 0.13917271002419995, + "learning_rate": 2.3794972389049448e-05, + "loss": 0.3688, + "step": 2166 + }, + { + "epoch": 3.349304482225657, + "grad_norm": 0.1374963736951724, + "learning_rate": 2.3755514494136583e-05, + "loss": 0.3737, + "step": 2167 + }, + { + "epoch": 3.3508500772797527, + "grad_norm": 0.13850266898961983, + "learning_rate": 2.3716075519229004e-05, + "loss": 0.3593, + "step": 2168 + }, + { + "epoch": 3.3523956723338486, + "grad_norm": 0.1299321402219489, + "learning_rate": 2.3676655510261402e-05, + "loss": 0.3601, + "step": 2169 + }, + { + "epoch": 3.3539412673879445, + "grad_norm": 0.1534785499695775, + "learning_rate": 2.3637254513146406e-05, + "loss": 0.3618, + "step": 2170 + }, + { + "epoch": 3.3554868624420404, + "grad_norm": 0.13334938950825415, + "learning_rate": 2.3597872573774477e-05, + "loss": 0.3752, + "step": 2171 + }, + { + "epoch": 3.357032457496136, + "grad_norm": 0.1400025344417481, + "learning_rate": 2.3558509738013884e-05, + "loss": 0.361, + "step": 2172 + }, + { + "epoch": 3.358578052550232, + "grad_norm": 0.11592633920458482, + "learning_rate": 2.3519166051710664e-05, + "loss": 0.3666, + "step": 2173 + }, + { + "epoch": 3.3601236476043277, + "grad_norm": 0.1495106821718199, + "learning_rate": 2.3479841560688527e-05, + "loss": 0.3732, + "step": 2174 + }, + { + "epoch": 3.3616692426584236, + "grad_norm": 0.12148103208951651, + "learning_rate": 2.3440536310748825e-05, + "loss": 0.3678, + "step": 2175 + }, + { + "epoch": 3.3632148377125195, + "grad_norm": 0.1560380485682236, + "learning_rate": 2.3401250347670534e-05, + "loss": 0.3653, + "step": 2176 + }, + { + "epoch": 3.364760432766615, + "grad_norm": 0.12671564592640389, + "learning_rate": 2.3361983717210136e-05, + "loss": 0.3607, + "step": 2177 + }, + { + "epoch": 3.366306027820711, + "grad_norm": 0.15976617220407052, + "learning_rate": 2.3322736465101595e-05, + "loss": 0.3768, + "step": 2178 + }, + { + "epoch": 3.3678516228748068, + "grad_norm": 0.13554607763701795, + "learning_rate": 2.3283508637056305e-05, + "loss": 0.3661, + "step": 2179 + }, + { + "epoch": 3.3693972179289027, + "grad_norm": 0.16597557158850385, + "learning_rate": 2.324430027876308e-05, + "loss": 0.367, + "step": 2180 + }, + { + "epoch": 3.3709428129829986, + "grad_norm": 0.14013672295171226, + "learning_rate": 2.3205111435888012e-05, + "loss": 0.3676, + "step": 2181 + }, + { + "epoch": 3.3724884080370945, + "grad_norm": 0.16247932045000632, + "learning_rate": 2.3165942154074435e-05, + "loss": 0.3652, + "step": 2182 + }, + { + "epoch": 3.37403400309119, + "grad_norm": 0.127265358952448, + "learning_rate": 2.3126792478942984e-05, + "loss": 0.3609, + "step": 2183 + }, + { + "epoch": 3.375579598145286, + "grad_norm": 0.146921149877785, + "learning_rate": 2.3087662456091398e-05, + "loss": 0.3657, + "step": 2184 + }, + { + "epoch": 3.3771251931993818, + "grad_norm": 0.13121490181217182, + "learning_rate": 2.3048552131094527e-05, + "loss": 0.3603, + "step": 2185 + }, + { + "epoch": 3.3786707882534777, + "grad_norm": 0.17031366399895273, + "learning_rate": 2.3009461549504327e-05, + "loss": 0.3618, + "step": 2186 + }, + { + "epoch": 3.3802163833075736, + "grad_norm": 0.14391343264973547, + "learning_rate": 2.2970390756849707e-05, + "loss": 0.3677, + "step": 2187 + }, + { + "epoch": 3.381761978361669, + "grad_norm": 0.16743317678418054, + "learning_rate": 2.293133979863655e-05, + "loss": 0.3746, + "step": 2188 + }, + { + "epoch": 3.383307573415765, + "grad_norm": 0.4433956956052001, + "learning_rate": 2.2892308720347633e-05, + "loss": 0.3724, + "step": 2189 + }, + { + "epoch": 3.384853168469861, + "grad_norm": 0.17711364532400312, + "learning_rate": 2.2853297567442576e-05, + "loss": 0.3737, + "step": 2190 + }, + { + "epoch": 3.3863987635239567, + "grad_norm": 0.12860192982862037, + "learning_rate": 2.2814306385357784e-05, + "loss": 0.3679, + "step": 2191 + }, + { + "epoch": 3.3879443585780527, + "grad_norm": 0.18114097397886067, + "learning_rate": 2.277533521950645e-05, + "loss": 0.3689, + "step": 2192 + }, + { + "epoch": 3.3894899536321486, + "grad_norm": 0.143617498461099, + "learning_rate": 2.2736384115278397e-05, + "loss": 0.3653, + "step": 2193 + }, + { + "epoch": 3.391035548686244, + "grad_norm": 0.14934037051494214, + "learning_rate": 2.2697453118040108e-05, + "loss": 0.3705, + "step": 2194 + }, + { + "epoch": 3.39258114374034, + "grad_norm": 0.12063351961554758, + "learning_rate": 2.2658542273134635e-05, + "loss": 0.3649, + "step": 2195 + }, + { + "epoch": 3.394126738794436, + "grad_norm": 0.15418817321104553, + "learning_rate": 2.2619651625881565e-05, + "loss": 0.3695, + "step": 2196 + }, + { + "epoch": 3.3956723338485317, + "grad_norm": 0.1294006613911471, + "learning_rate": 2.258078122157696e-05, + "loss": 0.3669, + "step": 2197 + }, + { + "epoch": 3.3972179289026276, + "grad_norm": 0.14619418705114054, + "learning_rate": 2.2541931105493325e-05, + "loss": 0.372, + "step": 2198 + }, + { + "epoch": 3.398763523956723, + "grad_norm": 0.14290869803965706, + "learning_rate": 2.2503101322879504e-05, + "loss": 0.3659, + "step": 2199 + }, + { + "epoch": 3.400309119010819, + "grad_norm": 0.15248398259469637, + "learning_rate": 2.2464291918960654e-05, + "loss": 0.3582, + "step": 2200 + }, + { + "epoch": 3.401854714064915, + "grad_norm": 0.13962906567160044, + "learning_rate": 2.242550293893827e-05, + "loss": 0.3693, + "step": 2201 + }, + { + "epoch": 3.403400309119011, + "grad_norm": 0.15266168076238193, + "learning_rate": 2.2386734427989953e-05, + "loss": 0.362, + "step": 2202 + }, + { + "epoch": 3.4049459041731067, + "grad_norm": 0.13968077225056355, + "learning_rate": 2.234798643126951e-05, + "loss": 0.371, + "step": 2203 + }, + { + "epoch": 3.4064914992272026, + "grad_norm": 0.18309654478650297, + "learning_rate": 2.2309258993906893e-05, + "loss": 0.3575, + "step": 2204 + }, + { + "epoch": 3.4080370942812985, + "grad_norm": 0.15824859847422365, + "learning_rate": 2.227055216100806e-05, + "loss": 0.3744, + "step": 2205 + }, + { + "epoch": 3.409582689335394, + "grad_norm": 0.1799621641151411, + "learning_rate": 2.2231865977654986e-05, + "loss": 0.3659, + "step": 2206 + }, + { + "epoch": 3.41112828438949, + "grad_norm": 0.16819288754850809, + "learning_rate": 2.2193200488905577e-05, + "loss": 0.3668, + "step": 2207 + }, + { + "epoch": 3.412673879443586, + "grad_norm": 0.14129559366228042, + "learning_rate": 2.2154555739793704e-05, + "loss": 0.3697, + "step": 2208 + }, + { + "epoch": 3.4142194744976817, + "grad_norm": 0.16403415378386554, + "learning_rate": 2.2115931775328977e-05, + "loss": 0.363, + "step": 2209 + }, + { + "epoch": 3.4157650695517776, + "grad_norm": 0.13802924828710786, + "learning_rate": 2.207732864049686e-05, + "loss": 0.3746, + "step": 2210 + }, + { + "epoch": 3.417310664605873, + "grad_norm": 0.19682770170330607, + "learning_rate": 2.2038746380258588e-05, + "loss": 0.3564, + "step": 2211 + }, + { + "epoch": 3.418856259659969, + "grad_norm": 0.17313372028050691, + "learning_rate": 2.2000185039551023e-05, + "loss": 0.365, + "step": 2212 + }, + { + "epoch": 3.420401854714065, + "grad_norm": 0.15480950449114608, + "learning_rate": 2.1961644663286666e-05, + "loss": 0.3591, + "step": 2213 + }, + { + "epoch": 3.421947449768161, + "grad_norm": 0.14347961963031264, + "learning_rate": 2.1923125296353652e-05, + "loss": 0.3721, + "step": 2214 + }, + { + "epoch": 3.4234930448222567, + "grad_norm": 0.21703862095555077, + "learning_rate": 2.1884626983615602e-05, + "loss": 0.3807, + "step": 2215 + }, + { + "epoch": 3.4250386398763526, + "grad_norm": 0.24410686962854097, + "learning_rate": 2.184614976991162e-05, + "loss": 0.3628, + "step": 2216 + }, + { + "epoch": 3.426584234930448, + "grad_norm": 0.1363111022100698, + "learning_rate": 2.1807693700056246e-05, + "loss": 0.3581, + "step": 2217 + }, + { + "epoch": 3.428129829984544, + "grad_norm": 0.23085033749351025, + "learning_rate": 2.176925881883939e-05, + "loss": 0.3686, + "step": 2218 + }, + { + "epoch": 3.42967542503864, + "grad_norm": 0.12923862126584543, + "learning_rate": 2.173084517102627e-05, + "loss": 0.3668, + "step": 2219 + }, + { + "epoch": 3.4312210200927358, + "grad_norm": 0.1651531123248392, + "learning_rate": 2.1692452801357415e-05, + "loss": 0.3668, + "step": 2220 + }, + { + "epoch": 3.4327666151468317, + "grad_norm": 0.1424252177597652, + "learning_rate": 2.1654081754548528e-05, + "loss": 0.3668, + "step": 2221 + }, + { + "epoch": 3.434312210200927, + "grad_norm": 0.13295987830609388, + "learning_rate": 2.161573207529049e-05, + "loss": 0.3693, + "step": 2222 + }, + { + "epoch": 3.435857805255023, + "grad_norm": 0.13862922963229274, + "learning_rate": 2.1577403808249304e-05, + "loss": 0.3681, + "step": 2223 + }, + { + "epoch": 3.437403400309119, + "grad_norm": 0.13194165522460818, + "learning_rate": 2.1539096998066028e-05, + "loss": 0.3591, + "step": 2224 + }, + { + "epoch": 3.438948995363215, + "grad_norm": 0.13499690881078524, + "learning_rate": 2.1500811689356703e-05, + "loss": 0.3642, + "step": 2225 + }, + { + "epoch": 3.4404945904173108, + "grad_norm": 0.1339915648072994, + "learning_rate": 2.1462547926712393e-05, + "loss": 0.3554, + "step": 2226 + }, + { + "epoch": 3.4420401854714067, + "grad_norm": 0.13631810446849388, + "learning_rate": 2.1424305754699012e-05, + "loss": 0.3682, + "step": 2227 + }, + { + "epoch": 3.443585780525502, + "grad_norm": 0.12801192112518459, + "learning_rate": 2.138608521785732e-05, + "loss": 0.3745, + "step": 2228 + }, + { + "epoch": 3.445131375579598, + "grad_norm": 0.12447615819235312, + "learning_rate": 2.1347886360702958e-05, + "loss": 0.3668, + "step": 2229 + }, + { + "epoch": 3.446676970633694, + "grad_norm": 0.1281798142444195, + "learning_rate": 2.13097092277262e-05, + "loss": 0.3592, + "step": 2230 + }, + { + "epoch": 3.44822256568779, + "grad_norm": 0.18194444810246324, + "learning_rate": 2.1271553863392084e-05, + "loss": 0.378, + "step": 2231 + }, + { + "epoch": 3.4497681607418857, + "grad_norm": 0.16293128387063224, + "learning_rate": 2.1233420312140325e-05, + "loss": 0.3609, + "step": 2232 + }, + { + "epoch": 3.451313755795981, + "grad_norm": 0.19059816836888333, + "learning_rate": 2.1195308618385164e-05, + "loss": 0.3676, + "step": 2233 + }, + { + "epoch": 3.452859350850077, + "grad_norm": 0.23789578340396286, + "learning_rate": 2.1157218826515428e-05, + "loss": 0.3533, + "step": 2234 + }, + { + "epoch": 3.454404945904173, + "grad_norm": 0.1305828277942505, + "learning_rate": 2.1119150980894405e-05, + "loss": 0.3593, + "step": 2235 + }, + { + "epoch": 3.455950540958269, + "grad_norm": 0.15005871190242023, + "learning_rate": 2.1081105125859884e-05, + "loss": 0.3739, + "step": 2236 + }, + { + "epoch": 3.457496136012365, + "grad_norm": 0.12897613860424506, + "learning_rate": 2.1043081305723945e-05, + "loss": 0.37, + "step": 2237 + }, + { + "epoch": 3.4590417310664607, + "grad_norm": 0.13550501095396109, + "learning_rate": 2.1005079564773057e-05, + "loss": 0.3722, + "step": 2238 + }, + { + "epoch": 3.4605873261205566, + "grad_norm": 0.11760751538202946, + "learning_rate": 2.0967099947268013e-05, + "loss": 0.3663, + "step": 2239 + }, + { + "epoch": 3.462132921174652, + "grad_norm": 0.14338053026336461, + "learning_rate": 2.092914249744377e-05, + "loss": 0.3662, + "step": 2240 + }, + { + "epoch": 3.463678516228748, + "grad_norm": 0.1213786296882088, + "learning_rate": 2.089120725950948e-05, + "loss": 0.3667, + "step": 2241 + }, + { + "epoch": 3.465224111282844, + "grad_norm": 0.1389070497931083, + "learning_rate": 2.0853294277648475e-05, + "loss": 0.3622, + "step": 2242 + }, + { + "epoch": 3.46676970633694, + "grad_norm": 0.3067028646142398, + "learning_rate": 2.0815403596018103e-05, + "loss": 0.3719, + "step": 2243 + }, + { + "epoch": 3.4683153013910357, + "grad_norm": 0.13083023438257277, + "learning_rate": 2.077753525874978e-05, + "loss": 0.3748, + "step": 2244 + }, + { + "epoch": 3.469860896445131, + "grad_norm": 0.2117395750689371, + "learning_rate": 2.0739689309948865e-05, + "loss": 0.3644, + "step": 2245 + }, + { + "epoch": 3.471406491499227, + "grad_norm": 0.1514838520767274, + "learning_rate": 2.070186579369467e-05, + "loss": 0.363, + "step": 2246 + }, + { + "epoch": 3.472952086553323, + "grad_norm": 0.13751256580191215, + "learning_rate": 2.066406475404035e-05, + "loss": 0.3631, + "step": 2247 + }, + { + "epoch": 3.474497681607419, + "grad_norm": 0.14978411561899216, + "learning_rate": 2.0626286235012928e-05, + "loss": 0.3615, + "step": 2248 + }, + { + "epoch": 3.476043276661515, + "grad_norm": 0.15003822080242218, + "learning_rate": 2.0588530280613152e-05, + "loss": 0.3667, + "step": 2249 + }, + { + "epoch": 3.4775888717156107, + "grad_norm": 0.1624011793089801, + "learning_rate": 2.0550796934815507e-05, + "loss": 0.363, + "step": 2250 + }, + { + "epoch": 3.479134466769706, + "grad_norm": 0.11050380571513596, + "learning_rate": 2.051308624156815e-05, + "loss": 0.3668, + "step": 2251 + }, + { + "epoch": 3.480680061823802, + "grad_norm": 0.1561929256372977, + "learning_rate": 2.0475398244792835e-05, + "loss": 0.3669, + "step": 2252 + }, + { + "epoch": 3.482225656877898, + "grad_norm": 0.1187676222355437, + "learning_rate": 2.043773298838489e-05, + "loss": 0.3565, + "step": 2253 + }, + { + "epoch": 3.483771251931994, + "grad_norm": 0.17276171169900728, + "learning_rate": 2.0400090516213192e-05, + "loss": 0.3709, + "step": 2254 + }, + { + "epoch": 3.48531684698609, + "grad_norm": 0.14069491642052284, + "learning_rate": 2.0362470872120032e-05, + "loss": 0.3766, + "step": 2255 + }, + { + "epoch": 3.4868624420401853, + "grad_norm": 0.14822815796650152, + "learning_rate": 2.032487409992113e-05, + "loss": 0.37, + "step": 2256 + }, + { + "epoch": 3.488408037094281, + "grad_norm": 0.1413867069961767, + "learning_rate": 2.0287300243405578e-05, + "loss": 0.3644, + "step": 2257 + }, + { + "epoch": 3.489953632148377, + "grad_norm": 0.15082419920184817, + "learning_rate": 2.0249749346335764e-05, + "loss": 0.37, + "step": 2258 + }, + { + "epoch": 3.491499227202473, + "grad_norm": 0.14103776048876207, + "learning_rate": 2.0212221452447328e-05, + "loss": 0.3717, + "step": 2259 + }, + { + "epoch": 3.493044822256569, + "grad_norm": 0.13815955422925869, + "learning_rate": 2.0174716605449155e-05, + "loss": 0.3677, + "step": 2260 + }, + { + "epoch": 3.4945904173106648, + "grad_norm": 0.15131638558783417, + "learning_rate": 2.0137234849023248e-05, + "loss": 0.3662, + "step": 2261 + }, + { + "epoch": 3.4961360123647607, + "grad_norm": 0.12480028276492205, + "learning_rate": 2.009977622682474e-05, + "loss": 0.362, + "step": 2262 + }, + { + "epoch": 3.497681607418856, + "grad_norm": 0.15266972037670695, + "learning_rate": 2.0062340782481785e-05, + "loss": 0.3693, + "step": 2263 + }, + { + "epoch": 3.499227202472952, + "grad_norm": 0.12720931793188006, + "learning_rate": 2.002492855959562e-05, + "loss": 0.3665, + "step": 2264 + }, + { + "epoch": 3.500772797527048, + "grad_norm": 0.1272550131960794, + "learning_rate": 1.9987539601740332e-05, + "loss": 0.3645, + "step": 2265 + }, + { + "epoch": 3.502318392581144, + "grad_norm": 0.12189387184243682, + "learning_rate": 1.9950173952462966e-05, + "loss": 0.3663, + "step": 2266 + }, + { + "epoch": 3.5038639876352393, + "grad_norm": 0.13492177300182276, + "learning_rate": 1.9912831655283446e-05, + "loss": 0.3555, + "step": 2267 + }, + { + "epoch": 3.5054095826893352, + "grad_norm": 0.12115252260644373, + "learning_rate": 1.987551275369445e-05, + "loss": 0.3732, + "step": 2268 + }, + { + "epoch": 3.506955177743431, + "grad_norm": 0.16506182656469232, + "learning_rate": 1.983821729116141e-05, + "loss": 0.3622, + "step": 2269 + }, + { + "epoch": 3.508500772797527, + "grad_norm": 0.12253683731579625, + "learning_rate": 1.9800945311122505e-05, + "loss": 0.3672, + "step": 2270 + }, + { + "epoch": 3.510046367851623, + "grad_norm": 0.1407742820609905, + "learning_rate": 1.9763696856988516e-05, + "loss": 0.37, + "step": 2271 + }, + { + "epoch": 3.511591962905719, + "grad_norm": 0.13681226830576138, + "learning_rate": 1.972647197214284e-05, + "loss": 0.3738, + "step": 2272 + }, + { + "epoch": 3.5131375579598147, + "grad_norm": 0.12079535836295369, + "learning_rate": 1.9689270699941416e-05, + "loss": 0.3649, + "step": 2273 + }, + { + "epoch": 3.51468315301391, + "grad_norm": 0.13307052757071333, + "learning_rate": 1.965209308371269e-05, + "loss": 0.3666, + "step": 2274 + }, + { + "epoch": 3.516228748068006, + "grad_norm": 0.12190116044810183, + "learning_rate": 1.9614939166757546e-05, + "loss": 0.3661, + "step": 2275 + }, + { + "epoch": 3.517774343122102, + "grad_norm": 0.12424625343826458, + "learning_rate": 1.957780899234929e-05, + "loss": 0.3676, + "step": 2276 + }, + { + "epoch": 3.519319938176198, + "grad_norm": 0.13310891983612444, + "learning_rate": 1.954070260373355e-05, + "loss": 0.3682, + "step": 2277 + }, + { + "epoch": 3.5208655332302934, + "grad_norm": 0.12058985909010728, + "learning_rate": 1.9503620044128254e-05, + "loss": 0.3736, + "step": 2278 + }, + { + "epoch": 3.5224111282843893, + "grad_norm": 0.1328700160544732, + "learning_rate": 1.9466561356723586e-05, + "loss": 0.3666, + "step": 2279 + }, + { + "epoch": 3.523956723338485, + "grad_norm": 0.12698907820026423, + "learning_rate": 1.9429526584681916e-05, + "loss": 0.3713, + "step": 2280 + }, + { + "epoch": 3.525502318392581, + "grad_norm": 0.11932373478013461, + "learning_rate": 1.9392515771137747e-05, + "loss": 0.3698, + "step": 2281 + }, + { + "epoch": 3.527047913446677, + "grad_norm": 0.13157331967424146, + "learning_rate": 1.9355528959197726e-05, + "loss": 0.369, + "step": 2282 + }, + { + "epoch": 3.528593508500773, + "grad_norm": 0.13566800341641627, + "learning_rate": 1.9318566191940504e-05, + "loss": 0.3691, + "step": 2283 + }, + { + "epoch": 3.530139103554869, + "grad_norm": 0.13162448475838684, + "learning_rate": 1.9281627512416724e-05, + "loss": 0.3729, + "step": 2284 + }, + { + "epoch": 3.5316846986089647, + "grad_norm": 0.11949373719861449, + "learning_rate": 1.9244712963649e-05, + "loss": 0.3534, + "step": 2285 + }, + { + "epoch": 3.53323029366306, + "grad_norm": 0.11736053310407618, + "learning_rate": 1.9207822588631817e-05, + "loss": 0.3707, + "step": 2286 + }, + { + "epoch": 3.534775888717156, + "grad_norm": 0.12394050065357538, + "learning_rate": 1.9170956430331507e-05, + "loss": 0.3698, + "step": 2287 + }, + { + "epoch": 3.536321483771252, + "grad_norm": 0.10797427831830231, + "learning_rate": 1.9134114531686235e-05, + "loss": 0.3545, + "step": 2288 + }, + { + "epoch": 3.537867078825348, + "grad_norm": 0.12618049750436605, + "learning_rate": 1.909729693560586e-05, + "loss": 0.3642, + "step": 2289 + }, + { + "epoch": 3.5394126738794434, + "grad_norm": 0.12076226937002642, + "learning_rate": 1.906050368497196e-05, + "loss": 0.3717, + "step": 2290 + }, + { + "epoch": 3.5409582689335393, + "grad_norm": 0.12473790053926571, + "learning_rate": 1.9023734822637743e-05, + "loss": 0.3679, + "step": 2291 + }, + { + "epoch": 3.542503863987635, + "grad_norm": 0.11560859256389514, + "learning_rate": 1.8986990391428056e-05, + "loss": 0.3687, + "step": 2292 + }, + { + "epoch": 3.544049459041731, + "grad_norm": 0.12606767866219007, + "learning_rate": 1.8950270434139227e-05, + "loss": 0.3667, + "step": 2293 + }, + { + "epoch": 3.545595054095827, + "grad_norm": 0.12389947684873036, + "learning_rate": 1.89135749935391e-05, + "loss": 0.3673, + "step": 2294 + }, + { + "epoch": 3.547140649149923, + "grad_norm": 0.1425149378605872, + "learning_rate": 1.8876904112367014e-05, + "loss": 0.3776, + "step": 2295 + }, + { + "epoch": 3.548686244204019, + "grad_norm": 0.15493403968430905, + "learning_rate": 1.884025783333364e-05, + "loss": 0.367, + "step": 2296 + }, + { + "epoch": 3.5502318392581143, + "grad_norm": 0.11972366321998125, + "learning_rate": 1.8803636199121006e-05, + "loss": 0.3698, + "step": 2297 + }, + { + "epoch": 3.55177743431221, + "grad_norm": 0.16216599108368399, + "learning_rate": 1.8767039252382474e-05, + "loss": 0.3649, + "step": 2298 + }, + { + "epoch": 3.553323029366306, + "grad_norm": 0.13076290262596807, + "learning_rate": 1.873046703574264e-05, + "loss": 0.372, + "step": 2299 + }, + { + "epoch": 3.554868624420402, + "grad_norm": 0.1801491112133388, + "learning_rate": 1.869391959179723e-05, + "loss": 0.363, + "step": 2300 + }, + { + "epoch": 3.5564142194744974, + "grad_norm": 0.1122279839845766, + "learning_rate": 1.865739696311322e-05, + "loss": 0.3632, + "step": 2301 + }, + { + "epoch": 3.5579598145285933, + "grad_norm": 0.16441164572650369, + "learning_rate": 1.8620899192228622e-05, + "loss": 0.376, + "step": 2302 + }, + { + "epoch": 3.5595054095826892, + "grad_norm": 0.12016647909976266, + "learning_rate": 1.858442632165248e-05, + "loss": 0.363, + "step": 2303 + }, + { + "epoch": 3.561051004636785, + "grad_norm": 0.13389180792505737, + "learning_rate": 1.8547978393864912e-05, + "loss": 0.3679, + "step": 2304 + }, + { + "epoch": 3.562596599690881, + "grad_norm": 0.13337931765783254, + "learning_rate": 1.851155545131692e-05, + "loss": 0.3662, + "step": 2305 + }, + { + "epoch": 3.564142194744977, + "grad_norm": 0.12029691360682021, + "learning_rate": 1.8475157536430415e-05, + "loss": 0.3775, + "step": 2306 + }, + { + "epoch": 3.565687789799073, + "grad_norm": 0.11962427742341881, + "learning_rate": 1.8438784691598176e-05, + "loss": 0.3626, + "step": 2307 + }, + { + "epoch": 3.5672333848531683, + "grad_norm": 0.12076196646675953, + "learning_rate": 1.840243695918377e-05, + "loss": 0.3503, + "step": 2308 + }, + { + "epoch": 3.5687789799072642, + "grad_norm": 0.1139843872027316, + "learning_rate": 1.8366114381521504e-05, + "loss": 0.3536, + "step": 2309 + }, + { + "epoch": 3.57032457496136, + "grad_norm": 0.1373164349875881, + "learning_rate": 1.8329817000916437e-05, + "loss": 0.3663, + "step": 2310 + }, + { + "epoch": 3.571870170015456, + "grad_norm": 0.14416586034482695, + "learning_rate": 1.8293544859644225e-05, + "loss": 0.3702, + "step": 2311 + }, + { + "epoch": 3.573415765069552, + "grad_norm": 0.12982911810173123, + "learning_rate": 1.825729799995116e-05, + "loss": 0.3662, + "step": 2312 + }, + { + "epoch": 3.5749613601236474, + "grad_norm": 0.1391599086123799, + "learning_rate": 1.8221076464054066e-05, + "loss": 0.3743, + "step": 2313 + }, + { + "epoch": 3.5765069551777433, + "grad_norm": 0.13603704128187755, + "learning_rate": 1.8184880294140293e-05, + "loss": 0.3624, + "step": 2314 + }, + { + "epoch": 3.578052550231839, + "grad_norm": 0.123219953191658, + "learning_rate": 1.814870953236762e-05, + "loss": 0.3734, + "step": 2315 + }, + { + "epoch": 3.579598145285935, + "grad_norm": 0.1545600721091344, + "learning_rate": 1.811256422086428e-05, + "loss": 0.3752, + "step": 2316 + }, + { + "epoch": 3.581143740340031, + "grad_norm": 0.12844492582796185, + "learning_rate": 1.807644440172882e-05, + "loss": 0.3571, + "step": 2317 + }, + { + "epoch": 3.582689335394127, + "grad_norm": 0.13017833846477858, + "learning_rate": 1.804035011703011e-05, + "loss": 0.3691, + "step": 2318 + }, + { + "epoch": 3.584234930448223, + "grad_norm": 0.1444866065809688, + "learning_rate": 1.8004281408807287e-05, + "loss": 0.3695, + "step": 2319 + }, + { + "epoch": 3.5857805255023183, + "grad_norm": 0.12692361732698557, + "learning_rate": 1.796823831906969e-05, + "loss": 0.3669, + "step": 2320 + }, + { + "epoch": 3.587326120556414, + "grad_norm": 0.12298294915988707, + "learning_rate": 1.793222088979682e-05, + "loss": 0.3554, + "step": 2321 + }, + { + "epoch": 3.58887171561051, + "grad_norm": 0.14366650700441572, + "learning_rate": 1.7896229162938286e-05, + "loss": 0.3647, + "step": 2322 + }, + { + "epoch": 3.590417310664606, + "grad_norm": 0.11991611653840754, + "learning_rate": 1.7860263180413797e-05, + "loss": 0.3682, + "step": 2323 + }, + { + "epoch": 3.5919629057187015, + "grad_norm": 0.12616629641714278, + "learning_rate": 1.782432298411303e-05, + "loss": 0.353, + "step": 2324 + }, + { + "epoch": 3.5935085007727974, + "grad_norm": 0.11510384978951838, + "learning_rate": 1.778840861589564e-05, + "loss": 0.3682, + "step": 2325 + }, + { + "epoch": 3.5950540958268933, + "grad_norm": 0.1427137699243558, + "learning_rate": 1.7752520117591237e-05, + "loss": 0.3586, + "step": 2326 + }, + { + "epoch": 3.596599690880989, + "grad_norm": 0.12137664650483156, + "learning_rate": 1.771665753099927e-05, + "loss": 0.3699, + "step": 2327 + }, + { + "epoch": 3.598145285935085, + "grad_norm": 0.1593699503856505, + "learning_rate": 1.768082089788897e-05, + "loss": 0.3726, + "step": 2328 + }, + { + "epoch": 3.599690880989181, + "grad_norm": 0.1132503843017808, + "learning_rate": 1.7645010259999425e-05, + "loss": 0.3631, + "step": 2329 + }, + { + "epoch": 3.601236476043277, + "grad_norm": 0.12430097431032934, + "learning_rate": 1.7609225659039384e-05, + "loss": 0.3601, + "step": 2330 + }, + { + "epoch": 3.6027820710973724, + "grad_norm": 0.12808034790865092, + "learning_rate": 1.7573467136687276e-05, + "loss": 0.3719, + "step": 2331 + }, + { + "epoch": 3.6043276661514683, + "grad_norm": 0.11166963969343933, + "learning_rate": 1.7537734734591193e-05, + "loss": 0.369, + "step": 2332 + }, + { + "epoch": 3.605873261205564, + "grad_norm": 0.12119834161313824, + "learning_rate": 1.7502028494368777e-05, + "loss": 0.3612, + "step": 2333 + }, + { + "epoch": 3.60741885625966, + "grad_norm": 0.1338599900749187, + "learning_rate": 1.746634845760719e-05, + "loss": 0.3659, + "step": 2334 + }, + { + "epoch": 3.6089644513137555, + "grad_norm": 0.1202207743667235, + "learning_rate": 1.7430694665863096e-05, + "loss": 0.3683, + "step": 2335 + }, + { + "epoch": 3.6105100463678514, + "grad_norm": 0.1237073823219968, + "learning_rate": 1.7395067160662575e-05, + "loss": 0.3714, + "step": 2336 + }, + { + "epoch": 3.6120556414219473, + "grad_norm": 0.12007218125289, + "learning_rate": 1.7359465983501088e-05, + "loss": 0.371, + "step": 2337 + }, + { + "epoch": 3.6136012364760433, + "grad_norm": 0.12842443914781965, + "learning_rate": 1.732389117584347e-05, + "loss": 0.364, + "step": 2338 + }, + { + "epoch": 3.615146831530139, + "grad_norm": 0.10596904265581793, + "learning_rate": 1.728834277912381e-05, + "loss": 0.3602, + "step": 2339 + }, + { + "epoch": 3.616692426584235, + "grad_norm": 0.10869023415659597, + "learning_rate": 1.725282083474543e-05, + "loss": 0.3621, + "step": 2340 + }, + { + "epoch": 3.618238021638331, + "grad_norm": 0.1165075477749802, + "learning_rate": 1.7217325384080856e-05, + "loss": 0.3655, + "step": 2341 + }, + { + "epoch": 3.6197836166924264, + "grad_norm": 0.10882163379175609, + "learning_rate": 1.718185646847176e-05, + "loss": 0.3596, + "step": 2342 + }, + { + "epoch": 3.6213292117465223, + "grad_norm": 0.1178456252177068, + "learning_rate": 1.7146414129228907e-05, + "loss": 0.3657, + "step": 2343 + }, + { + "epoch": 3.6228748068006182, + "grad_norm": 0.10240211125646784, + "learning_rate": 1.7110998407632075e-05, + "loss": 0.361, + "step": 2344 + }, + { + "epoch": 3.624420401854714, + "grad_norm": 0.13934509863883984, + "learning_rate": 1.707560934493012e-05, + "loss": 0.362, + "step": 2345 + }, + { + "epoch": 3.62596599690881, + "grad_norm": 0.12235121349091291, + "learning_rate": 1.7040246982340768e-05, + "loss": 0.3724, + "step": 2346 + }, + { + "epoch": 3.6275115919629055, + "grad_norm": 0.1289959917030072, + "learning_rate": 1.7004911361050688e-05, + "loss": 0.3619, + "step": 2347 + }, + { + "epoch": 3.6290571870170014, + "grad_norm": 0.12656070349013324, + "learning_rate": 1.696960252221539e-05, + "loss": 0.3765, + "step": 2348 + }, + { + "epoch": 3.6306027820710973, + "grad_norm": 0.13164208414661552, + "learning_rate": 1.6934320506959186e-05, + "loss": 0.3569, + "step": 2349 + }, + { + "epoch": 3.6321483771251932, + "grad_norm": 0.14018696204791378, + "learning_rate": 1.6899065356375157e-05, + "loss": 0.3696, + "step": 2350 + }, + { + "epoch": 3.633693972179289, + "grad_norm": 0.13335338503516184, + "learning_rate": 1.6863837111525102e-05, + "loss": 0.3583, + "step": 2351 + }, + { + "epoch": 3.635239567233385, + "grad_norm": 0.14189327242128436, + "learning_rate": 1.6828635813439474e-05, + "loss": 0.376, + "step": 2352 + }, + { + "epoch": 3.636785162287481, + "grad_norm": 0.1424767988489207, + "learning_rate": 1.6793461503117317e-05, + "loss": 0.3672, + "step": 2353 + }, + { + "epoch": 3.6383307573415764, + "grad_norm": 0.12716271067818344, + "learning_rate": 1.6758314221526295e-05, + "loss": 0.3703, + "step": 2354 + }, + { + "epoch": 3.6398763523956723, + "grad_norm": 0.14231456486144842, + "learning_rate": 1.6723194009602565e-05, + "loss": 0.362, + "step": 2355 + }, + { + "epoch": 3.641421947449768, + "grad_norm": 0.1155046849682613, + "learning_rate": 1.6688100908250706e-05, + "loss": 0.3692, + "step": 2356 + }, + { + "epoch": 3.642967542503864, + "grad_norm": 0.11984742917495439, + "learning_rate": 1.665303495834381e-05, + "loss": 0.364, + "step": 2357 + }, + { + "epoch": 3.6445131375579596, + "grad_norm": 0.11095599152364047, + "learning_rate": 1.661799620072329e-05, + "loss": 0.3625, + "step": 2358 + }, + { + "epoch": 3.6460587326120555, + "grad_norm": 0.11742773718301099, + "learning_rate": 1.658298467619889e-05, + "loss": 0.3677, + "step": 2359 + }, + { + "epoch": 3.6476043276661514, + "grad_norm": 0.1370819526528459, + "learning_rate": 1.6548000425548673e-05, + "loss": 0.3657, + "step": 2360 + }, + { + "epoch": 3.6491499227202473, + "grad_norm": 0.1310253884603917, + "learning_rate": 1.651304348951889e-05, + "loss": 0.372, + "step": 2361 + }, + { + "epoch": 3.650695517774343, + "grad_norm": 0.1352718464038083, + "learning_rate": 1.6478113908824007e-05, + "loss": 0.3683, + "step": 2362 + }, + { + "epoch": 3.652241112828439, + "grad_norm": 0.12912938776115337, + "learning_rate": 1.644321172414662e-05, + "loss": 0.3635, + "step": 2363 + }, + { + "epoch": 3.653786707882535, + "grad_norm": 0.14573174937900993, + "learning_rate": 1.6408336976137413e-05, + "loss": 0.3612, + "step": 2364 + }, + { + "epoch": 3.6553323029366305, + "grad_norm": 0.13518683413753524, + "learning_rate": 1.6373489705415105e-05, + "loss": 0.3549, + "step": 2365 + }, + { + "epoch": 3.6568778979907264, + "grad_norm": 0.14780922237985136, + "learning_rate": 1.6338669952566458e-05, + "loss": 0.3699, + "step": 2366 + }, + { + "epoch": 3.6584234930448223, + "grad_norm": 0.13451760149116115, + "learning_rate": 1.6303877758146135e-05, + "loss": 0.357, + "step": 2367 + }, + { + "epoch": 3.659969088098918, + "grad_norm": 0.1538837938353939, + "learning_rate": 1.6269113162676718e-05, + "loss": 0.3697, + "step": 2368 + }, + { + "epoch": 3.6615146831530136, + "grad_norm": 0.14661770256363155, + "learning_rate": 1.623437620664865e-05, + "loss": 0.3579, + "step": 2369 + }, + { + "epoch": 3.6630602782071096, + "grad_norm": 0.12274379462840326, + "learning_rate": 1.6199666930520167e-05, + "loss": 0.3643, + "step": 2370 + }, + { + "epoch": 3.6646058732612055, + "grad_norm": 0.1410997977440352, + "learning_rate": 1.6164985374717287e-05, + "loss": 0.3568, + "step": 2371 + }, + { + "epoch": 3.6661514683153014, + "grad_norm": 0.10481209766859342, + "learning_rate": 1.6130331579633708e-05, + "loss": 0.3653, + "step": 2372 + }, + { + "epoch": 3.6676970633693973, + "grad_norm": 0.13728656717393564, + "learning_rate": 1.609570558563085e-05, + "loss": 0.3703, + "step": 2373 + }, + { + "epoch": 3.669242658423493, + "grad_norm": 0.34913408623266173, + "learning_rate": 1.6061107433037707e-05, + "loss": 0.3802, + "step": 2374 + }, + { + "epoch": 3.670788253477589, + "grad_norm": 0.17026625701091253, + "learning_rate": 1.6026537162150866e-05, + "loss": 0.3699, + "step": 2375 + }, + { + "epoch": 3.6723338485316845, + "grad_norm": 0.10603428596954531, + "learning_rate": 1.5991994813234427e-05, + "loss": 0.3586, + "step": 2376 + }, + { + "epoch": 3.6738794435857804, + "grad_norm": 0.14456360618655534, + "learning_rate": 1.5957480426519974e-05, + "loss": 0.3641, + "step": 2377 + }, + { + "epoch": 3.6754250386398764, + "grad_norm": 0.12460666932722352, + "learning_rate": 1.5922994042206524e-05, + "loss": 0.3663, + "step": 2378 + }, + { + "epoch": 3.6769706336939723, + "grad_norm": 0.13224902476417935, + "learning_rate": 1.58885357004605e-05, + "loss": 0.3641, + "step": 2379 + }, + { + "epoch": 3.678516228748068, + "grad_norm": 0.11835773755558501, + "learning_rate": 1.5854105441415626e-05, + "loss": 0.3725, + "step": 2380 + }, + { + "epoch": 3.6800618238021636, + "grad_norm": 0.1399784920859329, + "learning_rate": 1.581970330517294e-05, + "loss": 0.3726, + "step": 2381 + }, + { + "epoch": 3.6816074188562595, + "grad_norm": 0.10863356986620387, + "learning_rate": 1.5785329331800748e-05, + "loss": 0.3699, + "step": 2382 + }, + { + "epoch": 3.6831530139103554, + "grad_norm": 0.14914152304207584, + "learning_rate": 1.5750983561334493e-05, + "loss": 0.3649, + "step": 2383 + }, + { + "epoch": 3.6846986089644513, + "grad_norm": 0.10983839097615365, + "learning_rate": 1.571666603377681e-05, + "loss": 0.3616, + "step": 2384 + }, + { + "epoch": 3.6862442040185472, + "grad_norm": 0.11864062319572434, + "learning_rate": 1.5682376789097454e-05, + "loss": 0.3686, + "step": 2385 + }, + { + "epoch": 3.687789799072643, + "grad_norm": 0.12921683619582586, + "learning_rate": 1.5648115867233217e-05, + "loss": 0.369, + "step": 2386 + }, + { + "epoch": 3.689335394126739, + "grad_norm": 0.13351544550888958, + "learning_rate": 1.5613883308087885e-05, + "loss": 0.3577, + "step": 2387 + }, + { + "epoch": 3.6908809891808345, + "grad_norm": 0.11574422097680881, + "learning_rate": 1.557967915153226e-05, + "loss": 0.3693, + "step": 2388 + }, + { + "epoch": 3.6924265842349304, + "grad_norm": 0.11967192135535802, + "learning_rate": 1.554550343740403e-05, + "loss": 0.3713, + "step": 2389 + }, + { + "epoch": 3.6939721792890263, + "grad_norm": 0.11774543564588924, + "learning_rate": 1.5511356205507745e-05, + "loss": 0.3664, + "step": 2390 + }, + { + "epoch": 3.6955177743431222, + "grad_norm": 0.11217961944837937, + "learning_rate": 1.547723749561481e-05, + "loss": 0.3731, + "step": 2391 + }, + { + "epoch": 3.6970633693972177, + "grad_norm": 0.1128032148337808, + "learning_rate": 1.544314734746339e-05, + "loss": 0.3599, + "step": 2392 + }, + { + "epoch": 3.6986089644513136, + "grad_norm": 0.11791019137532446, + "learning_rate": 1.5409085800758382e-05, + "loss": 0.3727, + "step": 2393 + }, + { + "epoch": 3.7001545595054095, + "grad_norm": 0.11014852879861646, + "learning_rate": 1.537505289517141e-05, + "loss": 0.3631, + "step": 2394 + }, + { + "epoch": 3.7017001545595054, + "grad_norm": 0.1189598231949291, + "learning_rate": 1.534104867034069e-05, + "loss": 0.3662, + "step": 2395 + }, + { + "epoch": 3.7032457496136013, + "grad_norm": 0.1064272086248934, + "learning_rate": 1.5307073165871064e-05, + "loss": 0.3714, + "step": 2396 + }, + { + "epoch": 3.704791344667697, + "grad_norm": 0.1175254940810905, + "learning_rate": 1.5273126421333902e-05, + "loss": 0.3622, + "step": 2397 + }, + { + "epoch": 3.706336939721793, + "grad_norm": 0.11869856783550627, + "learning_rate": 1.5239208476267094e-05, + "loss": 0.3855, + "step": 2398 + }, + { + "epoch": 3.7078825347758886, + "grad_norm": 0.10585418877519753, + "learning_rate": 1.520531937017498e-05, + "loss": 0.3679, + "step": 2399 + }, + { + "epoch": 3.7094281298299845, + "grad_norm": 0.10414374016494432, + "learning_rate": 1.5171459142528298e-05, + "loss": 0.3646, + "step": 2400 + }, + { + "epoch": 3.7109737248840804, + "grad_norm": 0.1154751459568018, + "learning_rate": 1.5137627832764192e-05, + "loss": 0.3663, + "step": 2401 + }, + { + "epoch": 3.7125193199381763, + "grad_norm": 0.12972038175619321, + "learning_rate": 1.5103825480286092e-05, + "loss": 0.3634, + "step": 2402 + }, + { + "epoch": 3.7140649149922718, + "grad_norm": 0.10544766399278752, + "learning_rate": 1.50700521244637e-05, + "loss": 0.363, + "step": 2403 + }, + { + "epoch": 3.7156105100463677, + "grad_norm": 0.14105149020865543, + "learning_rate": 1.5036307804632958e-05, + "loss": 0.3599, + "step": 2404 + }, + { + "epoch": 3.7171561051004636, + "grad_norm": 0.10578826342035111, + "learning_rate": 1.5002592560095984e-05, + "loss": 0.3768, + "step": 2405 + }, + { + "epoch": 3.7187017001545595, + "grad_norm": 0.1246590706351202, + "learning_rate": 1.4968906430121015e-05, + "loss": 0.3727, + "step": 2406 + }, + { + "epoch": 3.7202472952086554, + "grad_norm": 0.10149076338372942, + "learning_rate": 1.4935249453942437e-05, + "loss": 0.3615, + "step": 2407 + }, + { + "epoch": 3.7217928902627513, + "grad_norm": 0.11756217507445449, + "learning_rate": 1.4901621670760613e-05, + "loss": 0.3607, + "step": 2408 + }, + { + "epoch": 3.723338485316847, + "grad_norm": 0.11375314071535139, + "learning_rate": 1.4868023119741927e-05, + "loss": 0.3734, + "step": 2409 + }, + { + "epoch": 3.7248840803709427, + "grad_norm": 0.1192254435232237, + "learning_rate": 1.4834453840018754e-05, + "loss": 0.3741, + "step": 2410 + }, + { + "epoch": 3.7264296754250386, + "grad_norm": 0.1350920020566422, + "learning_rate": 1.480091387068931e-05, + "loss": 0.3703, + "step": 2411 + }, + { + "epoch": 3.7279752704791345, + "grad_norm": 0.12556924033907177, + "learning_rate": 1.4767403250817704e-05, + "loss": 0.362, + "step": 2412 + }, + { + "epoch": 3.7295208655332304, + "grad_norm": 0.11766856952811641, + "learning_rate": 1.4733922019433893e-05, + "loss": 0.3663, + "step": 2413 + }, + { + "epoch": 3.7310664605873263, + "grad_norm": 0.1286354996512215, + "learning_rate": 1.4700470215533558e-05, + "loss": 0.3648, + "step": 2414 + }, + { + "epoch": 3.7326120556414217, + "grad_norm": 0.11158946005891823, + "learning_rate": 1.4667047878078115e-05, + "loss": 0.3733, + "step": 2415 + }, + { + "epoch": 3.7341576506955176, + "grad_norm": 0.11544437169649772, + "learning_rate": 1.46336550459947e-05, + "loss": 0.3605, + "step": 2416 + }, + { + "epoch": 3.7357032457496135, + "grad_norm": 0.11471496762611157, + "learning_rate": 1.4600291758176028e-05, + "loss": 0.3704, + "step": 2417 + }, + { + "epoch": 3.7372488408037094, + "grad_norm": 0.1238451434723756, + "learning_rate": 1.4566958053480442e-05, + "loss": 0.3697, + "step": 2418 + }, + { + "epoch": 3.7387944358578054, + "grad_norm": 0.13135744878532335, + "learning_rate": 1.4533653970731808e-05, + "loss": 0.3663, + "step": 2419 + }, + { + "epoch": 3.7403400309119013, + "grad_norm": 0.14567429305716154, + "learning_rate": 1.4500379548719505e-05, + "loss": 0.378, + "step": 2420 + }, + { + "epoch": 3.741885625965997, + "grad_norm": 0.11847751544960321, + "learning_rate": 1.4467134826198343e-05, + "loss": 0.3637, + "step": 2421 + }, + { + "epoch": 3.7434312210200926, + "grad_norm": 0.12265362965387298, + "learning_rate": 1.443391984188859e-05, + "loss": 0.3652, + "step": 2422 + }, + { + "epoch": 3.7449768160741885, + "grad_norm": 0.10576200048700815, + "learning_rate": 1.4400734634475835e-05, + "loss": 0.3615, + "step": 2423 + }, + { + "epoch": 3.7465224111282844, + "grad_norm": 0.09971084048285343, + "learning_rate": 1.4367579242610989e-05, + "loss": 0.3642, + "step": 2424 + }, + { + "epoch": 3.7480680061823803, + "grad_norm": 0.11389348686792579, + "learning_rate": 1.4334453704910254e-05, + "loss": 0.3694, + "step": 2425 + }, + { + "epoch": 3.749613601236476, + "grad_norm": 0.11100882415194324, + "learning_rate": 1.4301358059955045e-05, + "loss": 0.3722, + "step": 2426 + }, + { + "epoch": 3.7511591962905717, + "grad_norm": 0.10612153251344403, + "learning_rate": 1.4268292346291972e-05, + "loss": 0.3722, + "step": 2427 + }, + { + "epoch": 3.7527047913446676, + "grad_norm": 0.10280001895994094, + "learning_rate": 1.4235256602432763e-05, + "loss": 0.362, + "step": 2428 + }, + { + "epoch": 3.7542503863987635, + "grad_norm": 0.11268140539427714, + "learning_rate": 1.4202250866854285e-05, + "loss": 0.3675, + "step": 2429 + }, + { + "epoch": 3.7557959814528594, + "grad_norm": 0.10771453126577249, + "learning_rate": 1.4169275177998416e-05, + "loss": 0.3659, + "step": 2430 + }, + { + "epoch": 3.7573415765069553, + "grad_norm": 0.10835455375616225, + "learning_rate": 1.4136329574272041e-05, + "loss": 0.3684, + "step": 2431 + }, + { + "epoch": 3.7588871715610512, + "grad_norm": 0.10972159383995192, + "learning_rate": 1.4103414094047021e-05, + "loss": 0.368, + "step": 2432 + }, + { + "epoch": 3.7604327666151467, + "grad_norm": 0.10179697696083068, + "learning_rate": 1.4070528775660113e-05, + "loss": 0.3571, + "step": 2433 + }, + { + "epoch": 3.7619783616692426, + "grad_norm": 0.11923762129516918, + "learning_rate": 1.4037673657412949e-05, + "loss": 0.3663, + "step": 2434 + }, + { + "epoch": 3.7635239567233385, + "grad_norm": 0.10963970754030641, + "learning_rate": 1.4004848777572022e-05, + "loss": 0.367, + "step": 2435 + }, + { + "epoch": 3.7650695517774344, + "grad_norm": 0.1083508491663517, + "learning_rate": 1.3972054174368555e-05, + "loss": 0.3719, + "step": 2436 + }, + { + "epoch": 3.76661514683153, + "grad_norm": 0.11886956305124681, + "learning_rate": 1.3939289885998516e-05, + "loss": 0.3582, + "step": 2437 + }, + { + "epoch": 3.7681607418856258, + "grad_norm": 0.09179304143810182, + "learning_rate": 1.3906555950622624e-05, + "loss": 0.361, + "step": 2438 + }, + { + "epoch": 3.7697063369397217, + "grad_norm": 0.11000935440391167, + "learning_rate": 1.3873852406366162e-05, + "loss": 0.3558, + "step": 2439 + }, + { + "epoch": 3.7712519319938176, + "grad_norm": 0.09878863374993056, + "learning_rate": 1.3841179291319047e-05, + "loss": 0.3699, + "step": 2440 + }, + { + "epoch": 3.7727975270479135, + "grad_norm": 0.11157323343796777, + "learning_rate": 1.3808536643535794e-05, + "loss": 0.3653, + "step": 2441 + }, + { + "epoch": 3.7743431221020094, + "grad_norm": 0.10868499354404498, + "learning_rate": 1.377592450103539e-05, + "loss": 0.3681, + "step": 2442 + }, + { + "epoch": 3.7758887171561053, + "grad_norm": 0.10663509845694374, + "learning_rate": 1.3743342901801287e-05, + "loss": 0.3622, + "step": 2443 + }, + { + "epoch": 3.7774343122102008, + "grad_norm": 0.12573443101677276, + "learning_rate": 1.3710791883781407e-05, + "loss": 0.3705, + "step": 2444 + }, + { + "epoch": 3.7789799072642967, + "grad_norm": 0.11492129231588556, + "learning_rate": 1.3678271484888024e-05, + "loss": 0.365, + "step": 2445 + }, + { + "epoch": 3.7805255023183926, + "grad_norm": 0.14807717541559526, + "learning_rate": 1.3645781742997723e-05, + "loss": 0.3656, + "step": 2446 + }, + { + "epoch": 3.7820710973724885, + "grad_norm": 0.12550994463979237, + "learning_rate": 1.361332269595144e-05, + "loss": 0.3654, + "step": 2447 + }, + { + "epoch": 3.7836166924265844, + "grad_norm": 0.11062588072470007, + "learning_rate": 1.358089438155433e-05, + "loss": 0.3602, + "step": 2448 + }, + { + "epoch": 3.78516228748068, + "grad_norm": 0.12207981186966346, + "learning_rate": 1.3548496837575734e-05, + "loss": 0.37, + "step": 2449 + }, + { + "epoch": 3.7867078825347757, + "grad_norm": 0.10031037388559602, + "learning_rate": 1.3516130101749214e-05, + "loss": 0.3717, + "step": 2450 + }, + { + "epoch": 3.7882534775888717, + "grad_norm": 0.13408802490768343, + "learning_rate": 1.3483794211772389e-05, + "loss": 0.3596, + "step": 2451 + }, + { + "epoch": 3.7897990726429676, + "grad_norm": 0.1369605229020428, + "learning_rate": 1.3451489205306985e-05, + "loss": 0.3711, + "step": 2452 + }, + { + "epoch": 3.7913446676970635, + "grad_norm": 0.1416302196359165, + "learning_rate": 1.3419215119978742e-05, + "loss": 0.3625, + "step": 2453 + }, + { + "epoch": 3.7928902627511594, + "grad_norm": 0.10716684360570666, + "learning_rate": 1.338697199337739e-05, + "loss": 0.3757, + "step": 2454 + }, + { + "epoch": 3.7944358578052553, + "grad_norm": 0.15103251616508864, + "learning_rate": 1.3354759863056614e-05, + "loss": 0.3762, + "step": 2455 + }, + { + "epoch": 3.7959814528593507, + "grad_norm": 0.13185075362002102, + "learning_rate": 1.3322578766533968e-05, + "loss": 0.3619, + "step": 2456 + }, + { + "epoch": 3.7975270479134466, + "grad_norm": 0.12239289098180552, + "learning_rate": 1.3290428741290908e-05, + "loss": 0.3644, + "step": 2457 + }, + { + "epoch": 3.7990726429675425, + "grad_norm": 0.13086101877955567, + "learning_rate": 1.3258309824772656e-05, + "loss": 0.3685, + "step": 2458 + }, + { + "epoch": 3.8006182380216385, + "grad_norm": 0.12886510139841528, + "learning_rate": 1.3226222054388225e-05, + "loss": 0.3613, + "step": 2459 + }, + { + "epoch": 3.802163833075734, + "grad_norm": 0.12681154241348283, + "learning_rate": 1.3194165467510337e-05, + "loss": 0.3616, + "step": 2460 + }, + { + "epoch": 3.80370942812983, + "grad_norm": 0.1374953428671208, + "learning_rate": 1.3162140101475416e-05, + "loss": 0.363, + "step": 2461 + }, + { + "epoch": 3.8052550231839257, + "grad_norm": 0.13472113941668032, + "learning_rate": 1.3130145993583487e-05, + "loss": 0.3692, + "step": 2462 + }, + { + "epoch": 3.8068006182380216, + "grad_norm": 0.12865828223260975, + "learning_rate": 1.3098183181098225e-05, + "loss": 0.3712, + "step": 2463 + }, + { + "epoch": 3.8083462132921175, + "grad_norm": 0.1320572537937874, + "learning_rate": 1.3066251701246801e-05, + "loss": 0.3695, + "step": 2464 + }, + { + "epoch": 3.8098918083462134, + "grad_norm": 0.12834562006677214, + "learning_rate": 1.303435159121992e-05, + "loss": 0.3709, + "step": 2465 + }, + { + "epoch": 3.8114374034003093, + "grad_norm": 0.12331051533044547, + "learning_rate": 1.3002482888171745e-05, + "loss": 0.3731, + "step": 2466 + }, + { + "epoch": 3.812982998454405, + "grad_norm": 0.12987931317976803, + "learning_rate": 1.2970645629219854e-05, + "loss": 0.3782, + "step": 2467 + }, + { + "epoch": 3.8145285935085007, + "grad_norm": 0.11810538966464726, + "learning_rate": 1.2938839851445195e-05, + "loss": 0.369, + "step": 2468 + }, + { + "epoch": 3.8160741885625966, + "grad_norm": 0.12288577780247337, + "learning_rate": 1.2907065591892085e-05, + "loss": 0.3606, + "step": 2469 + }, + { + "epoch": 3.8176197836166925, + "grad_norm": 0.11085041347305545, + "learning_rate": 1.2875322887568102e-05, + "loss": 0.365, + "step": 2470 + }, + { + "epoch": 3.819165378670788, + "grad_norm": 0.11836388424213261, + "learning_rate": 1.2843611775444056e-05, + "loss": 0.3711, + "step": 2471 + }, + { + "epoch": 3.820710973724884, + "grad_norm": 0.11440744882214117, + "learning_rate": 1.2811932292454015e-05, + "loss": 0.3673, + "step": 2472 + }, + { + "epoch": 3.82225656877898, + "grad_norm": 0.10902753212932297, + "learning_rate": 1.2780284475495175e-05, + "loss": 0.3581, + "step": 2473 + }, + { + "epoch": 3.8238021638330757, + "grad_norm": 0.10797678553836423, + "learning_rate": 1.2748668361427808e-05, + "loss": 0.3622, + "step": 2474 + }, + { + "epoch": 3.8253477588871716, + "grad_norm": 0.11778675829973131, + "learning_rate": 1.271708398707535e-05, + "loss": 0.3675, + "step": 2475 + }, + { + "epoch": 3.8268933539412675, + "grad_norm": 0.10984096594392173, + "learning_rate": 1.2685531389224206e-05, + "loss": 0.3613, + "step": 2476 + }, + { + "epoch": 3.8284389489953634, + "grad_norm": 0.11855504166646279, + "learning_rate": 1.2654010604623782e-05, + "loss": 0.3656, + "step": 2477 + }, + { + "epoch": 3.8299845440494593, + "grad_norm": 0.10105851607114989, + "learning_rate": 1.2622521669986463e-05, + "loss": 0.3587, + "step": 2478 + }, + { + "epoch": 3.8315301391035548, + "grad_norm": 0.11929478675749929, + "learning_rate": 1.2591064621987501e-05, + "loss": 0.3639, + "step": 2479 + }, + { + "epoch": 3.8330757341576507, + "grad_norm": 0.10465017419082835, + "learning_rate": 1.2559639497265024e-05, + "loss": 0.3708, + "step": 2480 + }, + { + "epoch": 3.8346213292117466, + "grad_norm": 0.0967672404214158, + "learning_rate": 1.2528246332419971e-05, + "loss": 0.3564, + "step": 2481 + }, + { + "epoch": 3.8361669242658425, + "grad_norm": 0.10544790052485256, + "learning_rate": 1.2496885164016073e-05, + "loss": 0.3764, + "step": 2482 + }, + { + "epoch": 3.837712519319938, + "grad_norm": 0.09617536404393245, + "learning_rate": 1.2465556028579777e-05, + "loss": 0.3646, + "step": 2483 + }, + { + "epoch": 3.839258114374034, + "grad_norm": 0.10297967559422405, + "learning_rate": 1.243425896260022e-05, + "loss": 0.3726, + "step": 2484 + }, + { + "epoch": 3.8408037094281298, + "grad_norm": 0.10244677724678106, + "learning_rate": 1.2402994002529223e-05, + "loss": 0.3705, + "step": 2485 + }, + { + "epoch": 3.8423493044822257, + "grad_norm": 0.10131219568374498, + "learning_rate": 1.2371761184781175e-05, + "loss": 0.3639, + "step": 2486 + }, + { + "epoch": 3.8438948995363216, + "grad_norm": 0.09707008193138299, + "learning_rate": 1.234056054573304e-05, + "loss": 0.3718, + "step": 2487 + }, + { + "epoch": 3.8454404945904175, + "grad_norm": 0.09992089678593088, + "learning_rate": 1.2309392121724302e-05, + "loss": 0.3625, + "step": 2488 + }, + { + "epoch": 3.8469860896445134, + "grad_norm": 0.09690112449159632, + "learning_rate": 1.2278255949056925e-05, + "loss": 0.3583, + "step": 2489 + }, + { + "epoch": 3.848531684698609, + "grad_norm": 0.10740608520702238, + "learning_rate": 1.2247152063995302e-05, + "loss": 0.3732, + "step": 2490 + }, + { + "epoch": 3.8500772797527048, + "grad_norm": 0.10377366745909304, + "learning_rate": 1.2216080502766254e-05, + "loss": 0.3644, + "step": 2491 + }, + { + "epoch": 3.8516228748068007, + "grad_norm": 0.10178958110475811, + "learning_rate": 1.2185041301558918e-05, + "loss": 0.3592, + "step": 2492 + }, + { + "epoch": 3.8531684698608966, + "grad_norm": 0.10540236073406219, + "learning_rate": 1.2154034496524748e-05, + "loss": 0.3627, + "step": 2493 + }, + { + "epoch": 3.854714064914992, + "grad_norm": 0.11341760050262721, + "learning_rate": 1.2123060123777477e-05, + "loss": 0.3642, + "step": 2494 + }, + { + "epoch": 3.856259659969088, + "grad_norm": 0.11106701813295258, + "learning_rate": 1.2092118219393058e-05, + "loss": 0.3639, + "step": 2495 + }, + { + "epoch": 3.857805255023184, + "grad_norm": 0.10845039770754565, + "learning_rate": 1.2061208819409611e-05, + "loss": 0.3657, + "step": 2496 + }, + { + "epoch": 3.8593508500772797, + "grad_norm": 0.11101560958602312, + "learning_rate": 1.203033195982744e-05, + "loss": 0.3691, + "step": 2497 + }, + { + "epoch": 3.8608964451313756, + "grad_norm": 0.10730901242536212, + "learning_rate": 1.1999487676608914e-05, + "loss": 0.3602, + "step": 2498 + }, + { + "epoch": 3.8624420401854715, + "grad_norm": 0.09987418290859781, + "learning_rate": 1.1968676005678459e-05, + "loss": 0.365, + "step": 2499 + }, + { + "epoch": 3.8639876352395675, + "grad_norm": 0.09839768397643736, + "learning_rate": 1.1937896982922555e-05, + "loss": 0.3511, + "step": 2500 + }, + { + "epoch": 3.865533230293663, + "grad_norm": 0.1141607541008281, + "learning_rate": 1.1907150644189627e-05, + "loss": 0.3595, + "step": 2501 + }, + { + "epoch": 3.867078825347759, + "grad_norm": 0.10095235543244717, + "learning_rate": 1.1876437025289995e-05, + "loss": 0.3674, + "step": 2502 + }, + { + "epoch": 3.8686244204018547, + "grad_norm": 0.11541566839456155, + "learning_rate": 1.1845756161995956e-05, + "loss": 0.3628, + "step": 2503 + }, + { + "epoch": 3.8701700154559506, + "grad_norm": 0.1838395361908257, + "learning_rate": 1.1815108090041596e-05, + "loss": 0.3647, + "step": 2504 + }, + { + "epoch": 3.871715610510046, + "grad_norm": 0.12030001829742068, + "learning_rate": 1.178449284512281e-05, + "loss": 0.3691, + "step": 2505 + }, + { + "epoch": 3.873261205564142, + "grad_norm": 0.11368358105501487, + "learning_rate": 1.1753910462897302e-05, + "loss": 0.3609, + "step": 2506 + }, + { + "epoch": 3.874806800618238, + "grad_norm": 0.10934822052637712, + "learning_rate": 1.1723360978984456e-05, + "loss": 0.3686, + "step": 2507 + }, + { + "epoch": 3.876352395672334, + "grad_norm": 0.11761633146968824, + "learning_rate": 1.169284442896537e-05, + "loss": 0.37, + "step": 2508 + }, + { + "epoch": 3.8778979907264297, + "grad_norm": 0.105574566148553, + "learning_rate": 1.166236084838274e-05, + "loss": 0.3648, + "step": 2509 + }, + { + "epoch": 3.8794435857805256, + "grad_norm": 0.1078432391498194, + "learning_rate": 1.1631910272740915e-05, + "loss": 0.3592, + "step": 2510 + }, + { + "epoch": 3.8809891808346215, + "grad_norm": 0.11406201758434457, + "learning_rate": 1.1601492737505775e-05, + "loss": 0.3789, + "step": 2511 + }, + { + "epoch": 3.8825347758887174, + "grad_norm": 0.11427280856537803, + "learning_rate": 1.1571108278104708e-05, + "loss": 0.3713, + "step": 2512 + }, + { + "epoch": 3.884080370942813, + "grad_norm": 0.10674616626343066, + "learning_rate": 1.1540756929926618e-05, + "loss": 0.3637, + "step": 2513 + }, + { + "epoch": 3.885625965996909, + "grad_norm": 0.10230105720022321, + "learning_rate": 1.1510438728321813e-05, + "loss": 0.364, + "step": 2514 + }, + { + "epoch": 3.8871715610510047, + "grad_norm": 0.10518339393882155, + "learning_rate": 1.148015370860199e-05, + "loss": 0.3669, + "step": 2515 + }, + { + "epoch": 3.8887171561051006, + "grad_norm": 0.10001859936936507, + "learning_rate": 1.1449901906040223e-05, + "loss": 0.3634, + "step": 2516 + }, + { + "epoch": 3.890262751159196, + "grad_norm": 0.11298110745277655, + "learning_rate": 1.1419683355870874e-05, + "loss": 0.362, + "step": 2517 + }, + { + "epoch": 3.891808346213292, + "grad_norm": 0.1037426756168707, + "learning_rate": 1.1389498093289579e-05, + "loss": 0.3644, + "step": 2518 + }, + { + "epoch": 3.893353941267388, + "grad_norm": 0.10133774840519594, + "learning_rate": 1.135934615345324e-05, + "loss": 0.3732, + "step": 2519 + }, + { + "epoch": 3.894899536321484, + "grad_norm": 0.09748273087606743, + "learning_rate": 1.1329227571479904e-05, + "loss": 0.3675, + "step": 2520 + }, + { + "epoch": 3.8964451313755797, + "grad_norm": 0.10083462430348453, + "learning_rate": 1.129914238244878e-05, + "loss": 0.3517, + "step": 2521 + }, + { + "epoch": 3.8979907264296756, + "grad_norm": 0.09849897729988676, + "learning_rate": 1.126909062140019e-05, + "loss": 0.3559, + "step": 2522 + }, + { + "epoch": 3.8995363214837715, + "grad_norm": 0.10555870666575547, + "learning_rate": 1.123907232333551e-05, + "loss": 0.3594, + "step": 2523 + }, + { + "epoch": 3.901081916537867, + "grad_norm": 0.11146760888284764, + "learning_rate": 1.1209087523217153e-05, + "loss": 0.3653, + "step": 2524 + }, + { + "epoch": 3.902627511591963, + "grad_norm": 0.11997983295919977, + "learning_rate": 1.1179136255968523e-05, + "loss": 0.3664, + "step": 2525 + }, + { + "epoch": 3.9041731066460588, + "grad_norm": 0.09861075251718329, + "learning_rate": 1.114921855647395e-05, + "loss": 0.3632, + "step": 2526 + }, + { + "epoch": 3.9057187017001547, + "grad_norm": 0.11435240577883474, + "learning_rate": 1.1119334459578668e-05, + "loss": 0.3712, + "step": 2527 + }, + { + "epoch": 3.90726429675425, + "grad_norm": 0.12655769296532293, + "learning_rate": 1.1089484000088805e-05, + "loss": 0.3663, + "step": 2528 + }, + { + "epoch": 3.908809891808346, + "grad_norm": 0.09803040820276109, + "learning_rate": 1.1059667212771288e-05, + "loss": 0.3654, + "step": 2529 + }, + { + "epoch": 3.910355486862442, + "grad_norm": 0.10975258191024564, + "learning_rate": 1.1029884132353793e-05, + "loss": 0.3656, + "step": 2530 + }, + { + "epoch": 3.911901081916538, + "grad_norm": 0.11451097258362002, + "learning_rate": 1.1000134793524802e-05, + "loss": 0.3545, + "step": 2531 + }, + { + "epoch": 3.9134466769706338, + "grad_norm": 0.10739152900892397, + "learning_rate": 1.0970419230933458e-05, + "loss": 0.3596, + "step": 2532 + }, + { + "epoch": 3.9149922720247297, + "grad_norm": 0.1046841267992002, + "learning_rate": 1.0940737479189555e-05, + "loss": 0.364, + "step": 2533 + }, + { + "epoch": 3.9165378670788256, + "grad_norm": 0.09670283606500873, + "learning_rate": 1.0911089572863562e-05, + "loss": 0.3617, + "step": 2534 + }, + { + "epoch": 3.918083462132921, + "grad_norm": 0.11150108076988523, + "learning_rate": 1.0881475546486478e-05, + "loss": 0.366, + "step": 2535 + }, + { + "epoch": 3.919629057187017, + "grad_norm": 0.10239205892448629, + "learning_rate": 1.0851895434549857e-05, + "loss": 0.3798, + "step": 2536 + }, + { + "epoch": 3.921174652241113, + "grad_norm": 0.10268938062417501, + "learning_rate": 1.082234927150573e-05, + "loss": 0.3629, + "step": 2537 + }, + { + "epoch": 3.9227202472952087, + "grad_norm": 0.10235283480388616, + "learning_rate": 1.0792837091766635e-05, + "loss": 0.3709, + "step": 2538 + }, + { + "epoch": 3.9242658423493046, + "grad_norm": 0.09634602615965598, + "learning_rate": 1.0763358929705499e-05, + "loss": 0.3712, + "step": 2539 + }, + { + "epoch": 3.9258114374034, + "grad_norm": 0.09301068968552884, + "learning_rate": 1.0733914819655618e-05, + "loss": 0.3667, + "step": 2540 + }, + { + "epoch": 3.927357032457496, + "grad_norm": 0.09545970724699197, + "learning_rate": 1.0704504795910667e-05, + "loss": 0.3632, + "step": 2541 + }, + { + "epoch": 3.928902627511592, + "grad_norm": 0.101607272338685, + "learning_rate": 1.0675128892724587e-05, + "loss": 0.371, + "step": 2542 + }, + { + "epoch": 3.930448222565688, + "grad_norm": 0.09517284569366352, + "learning_rate": 1.0645787144311583e-05, + "loss": 0.3602, + "step": 2543 + }, + { + "epoch": 3.9319938176197837, + "grad_norm": 0.09360351163832138, + "learning_rate": 1.0616479584846094e-05, + "loss": 0.3579, + "step": 2544 + }, + { + "epoch": 3.9335394126738796, + "grad_norm": 0.10538953893885825, + "learning_rate": 1.0587206248462718e-05, + "loss": 0.3641, + "step": 2545 + }, + { + "epoch": 3.9350850077279755, + "grad_norm": 0.10187825493669692, + "learning_rate": 1.0557967169256206e-05, + "loss": 0.3603, + "step": 2546 + }, + { + "epoch": 3.936630602782071, + "grad_norm": 0.10448244218839056, + "learning_rate": 1.0528762381281425e-05, + "loss": 0.3655, + "step": 2547 + }, + { + "epoch": 3.938176197836167, + "grad_norm": 0.09479663821005888, + "learning_rate": 1.0499591918553281e-05, + "loss": 0.3601, + "step": 2548 + }, + { + "epoch": 3.939721792890263, + "grad_norm": 0.09783003802874986, + "learning_rate": 1.04704558150467e-05, + "loss": 0.3625, + "step": 2549 + }, + { + "epoch": 3.9412673879443587, + "grad_norm": 0.10151351630021069, + "learning_rate": 1.04413541046966e-05, + "loss": 0.3661, + "step": 2550 + }, + { + "epoch": 3.942812982998454, + "grad_norm": 0.09724859515650129, + "learning_rate": 1.0412286821397838e-05, + "loss": 0.3702, + "step": 2551 + }, + { + "epoch": 3.94435857805255, + "grad_norm": 0.10697273241386306, + "learning_rate": 1.0383253999005163e-05, + "loss": 0.3755, + "step": 2552 + }, + { + "epoch": 3.945904173106646, + "grad_norm": 0.0999074538581681, + "learning_rate": 1.0354255671333222e-05, + "loss": 0.3682, + "step": 2553 + }, + { + "epoch": 3.947449768160742, + "grad_norm": 0.10390728660398883, + "learning_rate": 1.0325291872156447e-05, + "loss": 0.358, + "step": 2554 + }, + { + "epoch": 3.948995363214838, + "grad_norm": 0.1015880940924953, + "learning_rate": 1.029636263520906e-05, + "loss": 0.3661, + "step": 2555 + }, + { + "epoch": 3.9505409582689337, + "grad_norm": 0.09913974603917472, + "learning_rate": 1.0267467994185077e-05, + "loss": 0.3671, + "step": 2556 + }, + { + "epoch": 3.9520865533230296, + "grad_norm": 0.09285619168664734, + "learning_rate": 1.0238607982738133e-05, + "loss": 0.36, + "step": 2557 + }, + { + "epoch": 3.953632148377125, + "grad_norm": 0.09932654996319276, + "learning_rate": 1.0209782634481589e-05, + "loss": 0.3652, + "step": 2558 + }, + { + "epoch": 3.955177743431221, + "grad_norm": 0.10159746896733102, + "learning_rate": 1.0180991982988444e-05, + "loss": 0.3669, + "step": 2559 + }, + { + "epoch": 3.956723338485317, + "grad_norm": 0.10027230146989236, + "learning_rate": 1.015223606179125e-05, + "loss": 0.3758, + "step": 2560 + }, + { + "epoch": 3.958268933539413, + "grad_norm": 0.09630876921012405, + "learning_rate": 1.0123514904382108e-05, + "loss": 0.3578, + "step": 2561 + }, + { + "epoch": 3.9598145285935082, + "grad_norm": 0.10232753308969458, + "learning_rate": 1.0094828544212668e-05, + "loss": 0.3707, + "step": 2562 + }, + { + "epoch": 3.961360123647604, + "grad_norm": 0.10292052209086937, + "learning_rate": 1.0066177014694025e-05, + "loss": 0.3636, + "step": 2563 + }, + { + "epoch": 3.9629057187017, + "grad_norm": 0.10487621364529856, + "learning_rate": 1.003756034919671e-05, + "loss": 0.3728, + "step": 2564 + }, + { + "epoch": 3.964451313755796, + "grad_norm": 0.09723305903976502, + "learning_rate": 1.000897858105061e-05, + "loss": 0.3655, + "step": 2565 + }, + { + "epoch": 3.965996908809892, + "grad_norm": 0.11362968585616648, + "learning_rate": 9.980431743545047e-06, + "loss": 0.3637, + "step": 2566 + }, + { + "epoch": 3.9675425038639878, + "grad_norm": 0.09909344509120414, + "learning_rate": 9.951919869928601e-06, + "loss": 0.3672, + "step": 2567 + }, + { + "epoch": 3.9690880989180837, + "grad_norm": 0.11315416634975581, + "learning_rate": 9.923442993409131e-06, + "loss": 0.3691, + "step": 2568 + }, + { + "epoch": 3.970633693972179, + "grad_norm": 0.10533606955922994, + "learning_rate": 9.895001147153774e-06, + "loss": 0.3572, + "step": 2569 + }, + { + "epoch": 3.972179289026275, + "grad_norm": 0.12170280346016124, + "learning_rate": 9.866594364288825e-06, + "loss": 0.3546, + "step": 2570 + }, + { + "epoch": 3.973724884080371, + "grad_norm": 0.10999283026501391, + "learning_rate": 9.83822267789976e-06, + "loss": 0.3682, + "step": 2571 + }, + { + "epoch": 3.975270479134467, + "grad_norm": 0.10466572095659792, + "learning_rate": 9.809886121031177e-06, + "loss": 0.3671, + "step": 2572 + }, + { + "epoch": 3.9768160741885628, + "grad_norm": 0.11362183054589346, + "learning_rate": 9.781584726686751e-06, + "loss": 0.363, + "step": 2573 + }, + { + "epoch": 3.978361669242658, + "grad_norm": 0.09498875334782204, + "learning_rate": 9.7533185278292e-06, + "loss": 0.3614, + "step": 2574 + }, + { + "epoch": 3.979907264296754, + "grad_norm": 0.11332292262691414, + "learning_rate": 9.725087557380282e-06, + "loss": 0.3699, + "step": 2575 + }, + { + "epoch": 3.98145285935085, + "grad_norm": 0.10488757950290599, + "learning_rate": 9.696891848220686e-06, + "loss": 0.3602, + "step": 2576 + }, + { + "epoch": 3.982998454404946, + "grad_norm": 0.09354094065496711, + "learning_rate": 9.668731433190048e-06, + "loss": 0.3584, + "step": 2577 + }, + { + "epoch": 3.984544049459042, + "grad_norm": 0.10307736561633109, + "learning_rate": 9.640606345086896e-06, + "loss": 0.358, + "step": 2578 + }, + { + "epoch": 3.9860896445131377, + "grad_norm": 0.0991188352213659, + "learning_rate": 9.612516616668612e-06, + "loss": 0.3636, + "step": 2579 + }, + { + "epoch": 3.9876352395672336, + "grad_norm": 0.09447328499261065, + "learning_rate": 9.58446228065138e-06, + "loss": 0.3643, + "step": 2580 + }, + { + "epoch": 3.989180834621329, + "grad_norm": 0.10263294541166457, + "learning_rate": 9.5564433697102e-06, + "loss": 0.372, + "step": 2581 + }, + { + "epoch": 3.990726429675425, + "grad_norm": 0.09626087265886038, + "learning_rate": 9.528459916478785e-06, + "loss": 0.3605, + "step": 2582 + }, + { + "epoch": 3.992272024729521, + "grad_norm": 0.10037091578716863, + "learning_rate": 9.500511953549534e-06, + "loss": 0.3724, + "step": 2583 + }, + { + "epoch": 3.993817619783617, + "grad_norm": 0.10011278941164424, + "learning_rate": 9.472599513473563e-06, + "loss": 0.3709, + "step": 2584 + }, + { + "epoch": 3.9953632148377123, + "grad_norm": 0.09600870645580256, + "learning_rate": 9.44472262876056e-06, + "loss": 0.3695, + "step": 2585 + }, + { + "epoch": 3.996908809891808, + "grad_norm": 0.09299273510515463, + "learning_rate": 9.416881331878812e-06, + "loss": 0.3692, + "step": 2586 + }, + { + "epoch": 3.998454404945904, + "grad_norm": 0.09811585389206162, + "learning_rate": 9.389075655255197e-06, + "loss": 0.3739, + "step": 2587 + }, + { + "epoch": 4.0, + "grad_norm": 0.13954531960001063, + "learning_rate": 9.361305631275068e-06, + "loss": 0.3544, + "step": 2588 + }, + { + "epoch": 4.001545595054096, + "grad_norm": 0.10940343660912923, + "learning_rate": 9.333571292282243e-06, + "loss": 0.3502, + "step": 2589 + }, + { + "epoch": 4.003091190108192, + "grad_norm": 0.11320766478293215, + "learning_rate": 9.305872670579025e-06, + "loss": 0.3412, + "step": 2590 + }, + { + "epoch": 4.004636785162288, + "grad_norm": 0.11547323433966702, + "learning_rate": 9.278209798426082e-06, + "loss": 0.3428, + "step": 2591 + }, + { + "epoch": 4.006182380216384, + "grad_norm": 0.116703241057822, + "learning_rate": 9.250582708042452e-06, + "loss": 0.3461, + "step": 2592 + }, + { + "epoch": 4.0077279752704795, + "grad_norm": 0.11477518483073378, + "learning_rate": 9.222991431605472e-06, + "loss": 0.3538, + "step": 2593 + }, + { + "epoch": 4.0092735703245745, + "grad_norm": 0.11852802709349432, + "learning_rate": 9.195436001250822e-06, + "loss": 0.3511, + "step": 2594 + }, + { + "epoch": 4.0108191653786704, + "grad_norm": 0.1091164637954775, + "learning_rate": 9.167916449072392e-06, + "loss": 0.3484, + "step": 2595 + }, + { + "epoch": 4.012364760432766, + "grad_norm": 0.11766480809475463, + "learning_rate": 9.140432807122282e-06, + "loss": 0.3516, + "step": 2596 + }, + { + "epoch": 4.013910355486862, + "grad_norm": 0.11255519633655399, + "learning_rate": 9.112985107410801e-06, + "loss": 0.3436, + "step": 2597 + }, + { + "epoch": 4.015455950540958, + "grad_norm": 0.11371740300321591, + "learning_rate": 9.085573381906366e-06, + "loss": 0.3399, + "step": 2598 + }, + { + "epoch": 4.017001545595054, + "grad_norm": 0.12427309020683548, + "learning_rate": 9.058197662535501e-06, + "loss": 0.3488, + "step": 2599 + }, + { + "epoch": 4.01854714064915, + "grad_norm": 0.10365954811937023, + "learning_rate": 9.0308579811828e-06, + "loss": 0.3464, + "step": 2600 + }, + { + "epoch": 4.020092735703246, + "grad_norm": 0.14011142718575734, + "learning_rate": 9.00355436969087e-06, + "loss": 0.3524, + "step": 2601 + }, + { + "epoch": 4.021638330757342, + "grad_norm": 0.11021612748827103, + "learning_rate": 8.976286859860313e-06, + "loss": 0.3448, + "step": 2602 + }, + { + "epoch": 4.023183925811438, + "grad_norm": 0.10846563108738036, + "learning_rate": 8.949055483449708e-06, + "loss": 0.347, + "step": 2603 + }, + { + "epoch": 4.024729520865534, + "grad_norm": 0.10992941854538306, + "learning_rate": 8.921860272175515e-06, + "loss": 0.3508, + "step": 2604 + }, + { + "epoch": 4.0262751159196295, + "grad_norm": 0.1078991784580418, + "learning_rate": 8.894701257712084e-06, + "loss": 0.3507, + "step": 2605 + }, + { + "epoch": 4.0278207109737245, + "grad_norm": 0.09653065621329056, + "learning_rate": 8.867578471691609e-06, + "loss": 0.3538, + "step": 2606 + }, + { + "epoch": 4.02936630602782, + "grad_norm": 0.10408734736322096, + "learning_rate": 8.840491945704093e-06, + "loss": 0.3561, + "step": 2607 + }, + { + "epoch": 4.030911901081916, + "grad_norm": 0.10521353689745504, + "learning_rate": 8.813441711297281e-06, + "loss": 0.3406, + "step": 2608 + }, + { + "epoch": 4.032457496136012, + "grad_norm": 0.10747064265940268, + "learning_rate": 8.786427799976694e-06, + "loss": 0.3501, + "step": 2609 + }, + { + "epoch": 4.034003091190108, + "grad_norm": 0.116818989842645, + "learning_rate": 8.759450243205515e-06, + "loss": 0.3496, + "step": 2610 + }, + { + "epoch": 4.035548686244204, + "grad_norm": 0.09669902138624194, + "learning_rate": 8.73250907240458e-06, + "loss": 0.3513, + "step": 2611 + }, + { + "epoch": 4.0370942812983, + "grad_norm": 0.10804532239497597, + "learning_rate": 8.705604318952394e-06, + "loss": 0.3482, + "step": 2612 + }, + { + "epoch": 4.038639876352396, + "grad_norm": 0.3617095473258044, + "learning_rate": 8.678736014184972e-06, + "loss": 0.3485, + "step": 2613 + }, + { + "epoch": 4.040185471406492, + "grad_norm": 0.10391160408838113, + "learning_rate": 8.651904189395922e-06, + "loss": 0.3468, + "step": 2614 + }, + { + "epoch": 4.041731066460588, + "grad_norm": 0.10674311351942735, + "learning_rate": 8.625108875836377e-06, + "loss": 0.3543, + "step": 2615 + }, + { + "epoch": 4.043276661514684, + "grad_norm": 0.10698409102885607, + "learning_rate": 8.598350104714913e-06, + "loss": 0.3481, + "step": 2616 + }, + { + "epoch": 4.044822256568779, + "grad_norm": 0.09395026714733962, + "learning_rate": 8.571627907197548e-06, + "loss": 0.3511, + "step": 2617 + }, + { + "epoch": 4.0463678516228745, + "grad_norm": 0.11108836569277418, + "learning_rate": 8.54494231440774e-06, + "loss": 0.3469, + "step": 2618 + }, + { + "epoch": 4.04791344667697, + "grad_norm": 0.10152230122551807, + "learning_rate": 8.518293357426284e-06, + "loss": 0.3448, + "step": 2619 + }, + { + "epoch": 4.049459041731066, + "grad_norm": 0.10285114022610746, + "learning_rate": 8.491681067291279e-06, + "loss": 0.3462, + "step": 2620 + }, + { + "epoch": 4.051004636785162, + "grad_norm": 0.10585982621795423, + "learning_rate": 8.465105474998152e-06, + "loss": 0.3525, + "step": 2621 + }, + { + "epoch": 4.052550231839258, + "grad_norm": 0.09619262083884662, + "learning_rate": 8.438566611499603e-06, + "loss": 0.3469, + "step": 2622 + }, + { + "epoch": 4.054095826893354, + "grad_norm": 0.10071571746712828, + "learning_rate": 8.412064507705521e-06, + "loss": 0.3526, + "step": 2623 + }, + { + "epoch": 4.05564142194745, + "grad_norm": 0.10717668758266125, + "learning_rate": 8.385599194482985e-06, + "loss": 0.3541, + "step": 2624 + }, + { + "epoch": 4.057187017001546, + "grad_norm": 0.12277739259442917, + "learning_rate": 8.359170702656252e-06, + "loss": 0.3452, + "step": 2625 + }, + { + "epoch": 4.058732612055642, + "grad_norm": 0.10635331885322058, + "learning_rate": 8.33277906300666e-06, + "loss": 0.3546, + "step": 2626 + }, + { + "epoch": 4.060278207109738, + "grad_norm": 0.10375750544283012, + "learning_rate": 8.306424306272634e-06, + "loss": 0.3512, + "step": 2627 + }, + { + "epoch": 4.061823802163833, + "grad_norm": 0.10776883013627268, + "learning_rate": 8.280106463149655e-06, + "loss": 0.358, + "step": 2628 + }, + { + "epoch": 4.063369397217929, + "grad_norm": 0.08684956610614249, + "learning_rate": 8.253825564290192e-06, + "loss": 0.3526, + "step": 2629 + }, + { + "epoch": 4.0649149922720245, + "grad_norm": 0.1215722816083166, + "learning_rate": 8.227581640303688e-06, + "loss": 0.3542, + "step": 2630 + }, + { + "epoch": 4.06646058732612, + "grad_norm": 0.09308772726213017, + "learning_rate": 8.201374721756545e-06, + "loss": 0.3467, + "step": 2631 + }, + { + "epoch": 4.068006182380216, + "grad_norm": 0.09641533812883978, + "learning_rate": 8.175204839172041e-06, + "loss": 0.3521, + "step": 2632 + }, + { + "epoch": 4.069551777434312, + "grad_norm": 0.11116876668932495, + "learning_rate": 8.14907202303032e-06, + "loss": 0.3569, + "step": 2633 + }, + { + "epoch": 4.071097372488408, + "grad_norm": 0.08431105343847627, + "learning_rate": 8.122976303768357e-06, + "loss": 0.3429, + "step": 2634 + }, + { + "epoch": 4.072642967542504, + "grad_norm": 0.11091869787119317, + "learning_rate": 8.09691771177992e-06, + "loss": 0.343, + "step": 2635 + }, + { + "epoch": 4.0741885625966, + "grad_norm": 0.09404511826178034, + "learning_rate": 8.070896277415535e-06, + "loss": 0.3557, + "step": 2636 + }, + { + "epoch": 4.075734157650696, + "grad_norm": 0.09675306207406818, + "learning_rate": 8.044912030982464e-06, + "loss": 0.3477, + "step": 2637 + }, + { + "epoch": 4.077279752704792, + "grad_norm": 0.09294999121926695, + "learning_rate": 8.018965002744634e-06, + "loss": 0.345, + "step": 2638 + }, + { + "epoch": 4.078825347758887, + "grad_norm": 0.09093174893958986, + "learning_rate": 7.993055222922632e-06, + "loss": 0.3526, + "step": 2639 + }, + { + "epoch": 4.080370942812983, + "grad_norm": 0.09348658095917288, + "learning_rate": 7.96718272169367e-06, + "loss": 0.3451, + "step": 2640 + }, + { + "epoch": 4.0819165378670785, + "grad_norm": 0.09341498315647132, + "learning_rate": 7.94134752919153e-06, + "loss": 0.3467, + "step": 2641 + }, + { + "epoch": 4.083462132921174, + "grad_norm": 0.09435245349039612, + "learning_rate": 7.915549675506536e-06, + "loss": 0.3407, + "step": 2642 + }, + { + "epoch": 4.08500772797527, + "grad_norm": 0.09125878214059678, + "learning_rate": 7.889789190685548e-06, + "loss": 0.3475, + "step": 2643 + }, + { + "epoch": 4.086553323029366, + "grad_norm": 0.096903975224843, + "learning_rate": 7.864066104731885e-06, + "loss": 0.3398, + "step": 2644 + }, + { + "epoch": 4.088098918083462, + "grad_norm": 0.0901626823503655, + "learning_rate": 7.838380447605303e-06, + "loss": 0.3465, + "step": 2645 + }, + { + "epoch": 4.089644513137558, + "grad_norm": 0.0902459868780099, + "learning_rate": 7.812732249221962e-06, + "loss": 0.3485, + "step": 2646 + }, + { + "epoch": 4.091190108191654, + "grad_norm": 0.11066707787908245, + "learning_rate": 7.787121539454431e-06, + "loss": 0.3498, + "step": 2647 + }, + { + "epoch": 4.09273570324575, + "grad_norm": 0.10240484679057471, + "learning_rate": 7.761548348131564e-06, + "loss": 0.3459, + "step": 2648 + }, + { + "epoch": 4.094281298299846, + "grad_norm": 0.09512902311653947, + "learning_rate": 7.736012705038529e-06, + "loss": 0.3568, + "step": 2649 + }, + { + "epoch": 4.095826893353942, + "grad_norm": 0.1217732544379211, + "learning_rate": 7.710514639916807e-06, + "loss": 0.3466, + "step": 2650 + }, + { + "epoch": 4.097372488408037, + "grad_norm": 0.09070024923190873, + "learning_rate": 7.685054182464054e-06, + "loss": 0.3446, + "step": 2651 + }, + { + "epoch": 4.098918083462133, + "grad_norm": 0.0840054326889453, + "learning_rate": 7.659631362334146e-06, + "loss": 0.3453, + "step": 2652 + }, + { + "epoch": 4.1004636785162285, + "grad_norm": 0.1604455292133467, + "learning_rate": 7.634246209137139e-06, + "loss": 0.3526, + "step": 2653 + }, + { + "epoch": 4.102009273570324, + "grad_norm": 0.09067855239132425, + "learning_rate": 7.6088987524392025e-06, + "loss": 0.3548, + "step": 2654 + }, + { + "epoch": 4.10355486862442, + "grad_norm": 0.08831267056428002, + "learning_rate": 7.583589021762594e-06, + "loss": 0.3539, + "step": 2655 + }, + { + "epoch": 4.105100463678516, + "grad_norm": 0.10656832461232695, + "learning_rate": 7.5583170465856415e-06, + "loss": 0.3547, + "step": 2656 + }, + { + "epoch": 4.106646058732612, + "grad_norm": 0.11272606443660689, + "learning_rate": 7.5330828563427015e-06, + "loss": 0.3437, + "step": 2657 + }, + { + "epoch": 4.108191653786708, + "grad_norm": 0.10006764500407289, + "learning_rate": 7.5078864804241e-06, + "loss": 0.3581, + "step": 2658 + }, + { + "epoch": 4.109737248840804, + "grad_norm": 0.08699335455579792, + "learning_rate": 7.482727948176163e-06, + "loss": 0.343, + "step": 2659 + }, + { + "epoch": 4.1112828438949, + "grad_norm": 0.0891116336540287, + "learning_rate": 7.457607288901103e-06, + "loss": 0.3457, + "step": 2660 + }, + { + "epoch": 4.112828438948996, + "grad_norm": 0.10163097347052266, + "learning_rate": 7.432524531857037e-06, + "loss": 0.3531, + "step": 2661 + }, + { + "epoch": 4.114374034003091, + "grad_norm": 0.09494169578201946, + "learning_rate": 7.40747970625793e-06, + "loss": 0.3603, + "step": 2662 + }, + { + "epoch": 4.115919629057187, + "grad_norm": 0.09402303727235024, + "learning_rate": 7.382472841273576e-06, + "loss": 0.3387, + "step": 2663 + }, + { + "epoch": 4.117465224111283, + "grad_norm": 0.11541775732047659, + "learning_rate": 7.35750396602954e-06, + "loss": 0.3529, + "step": 2664 + }, + { + "epoch": 4.1190108191653785, + "grad_norm": 0.0968165489644101, + "learning_rate": 7.332573109607173e-06, + "loss": 0.3537, + "step": 2665 + }, + { + "epoch": 4.120556414219474, + "grad_norm": 0.09467829102180705, + "learning_rate": 7.307680301043514e-06, + "loss": 0.3503, + "step": 2666 + }, + { + "epoch": 4.12210200927357, + "grad_norm": 0.09864251003053498, + "learning_rate": 7.282825569331296e-06, + "loss": 0.3514, + "step": 2667 + }, + { + "epoch": 4.123647604327666, + "grad_norm": 0.10469599803133331, + "learning_rate": 7.258008943418913e-06, + "loss": 0.3499, + "step": 2668 + }, + { + "epoch": 4.125193199381762, + "grad_norm": 0.08812596124563764, + "learning_rate": 7.233230452210364e-06, + "loss": 0.3521, + "step": 2669 + }, + { + "epoch": 4.126738794435858, + "grad_norm": 0.10214262863411051, + "learning_rate": 7.208490124565238e-06, + "loss": 0.346, + "step": 2670 + }, + { + "epoch": 4.128284389489954, + "grad_norm": 0.09835108239310617, + "learning_rate": 7.183787989298681e-06, + "loss": 0.3507, + "step": 2671 + }, + { + "epoch": 4.12982998454405, + "grad_norm": 0.09020994042023699, + "learning_rate": 7.159124075181347e-06, + "loss": 0.3485, + "step": 2672 + }, + { + "epoch": 4.131375579598146, + "grad_norm": 0.10060564050468568, + "learning_rate": 7.134498410939383e-06, + "loss": 0.3437, + "step": 2673 + }, + { + "epoch": 4.132921174652241, + "grad_norm": 0.1186331079330418, + "learning_rate": 7.109911025254366e-06, + "loss": 0.3568, + "step": 2674 + }, + { + "epoch": 4.134466769706337, + "grad_norm": 0.09767985012419406, + "learning_rate": 7.08536194676333e-06, + "loss": 0.3458, + "step": 2675 + }, + { + "epoch": 4.1360123647604325, + "grad_norm": 0.1028522770072495, + "learning_rate": 7.060851204058643e-06, + "loss": 0.3484, + "step": 2676 + }, + { + "epoch": 4.1375579598145285, + "grad_norm": 0.09264338114853868, + "learning_rate": 7.036378825688044e-06, + "loss": 0.3522, + "step": 2677 + }, + { + "epoch": 4.139103554868624, + "grad_norm": 0.09341069540225162, + "learning_rate": 7.011944840154612e-06, + "loss": 0.346, + "step": 2678 + }, + { + "epoch": 4.14064914992272, + "grad_norm": 0.09763292400983055, + "learning_rate": 6.987549275916681e-06, + "loss": 0.3411, + "step": 2679 + }, + { + "epoch": 4.142194744976816, + "grad_norm": 0.09191041021765126, + "learning_rate": 6.963192161387837e-06, + "loss": 0.3496, + "step": 2680 + }, + { + "epoch": 4.143740340030912, + "grad_norm": 0.09712023598507023, + "learning_rate": 6.938873524936905e-06, + "loss": 0.3461, + "step": 2681 + }, + { + "epoch": 4.145285935085008, + "grad_norm": 0.09889849220678522, + "learning_rate": 6.914593394887883e-06, + "loss": 0.3498, + "step": 2682 + }, + { + "epoch": 4.146831530139104, + "grad_norm": 0.09735515745641965, + "learning_rate": 6.890351799519885e-06, + "loss": 0.3454, + "step": 2683 + }, + { + "epoch": 4.1483771251932, + "grad_norm": 0.1000739488906435, + "learning_rate": 6.866148767067211e-06, + "loss": 0.3579, + "step": 2684 + }, + { + "epoch": 4.149922720247295, + "grad_norm": 0.09416506863271783, + "learning_rate": 6.8419843257192e-06, + "loss": 0.3531, + "step": 2685 + }, + { + "epoch": 4.151468315301391, + "grad_norm": 0.08661446077850428, + "learning_rate": 6.817858503620245e-06, + "loss": 0.3456, + "step": 2686 + }, + { + "epoch": 4.153013910355487, + "grad_norm": 0.08864727511068461, + "learning_rate": 6.79377132886979e-06, + "loss": 0.3539, + "step": 2687 + }, + { + "epoch": 4.1545595054095825, + "grad_norm": 0.09814646368049065, + "learning_rate": 6.769722829522236e-06, + "loss": 0.3495, + "step": 2688 + }, + { + "epoch": 4.156105100463678, + "grad_norm": 0.10086265879889979, + "learning_rate": 6.74571303358695e-06, + "loss": 0.3524, + "step": 2689 + }, + { + "epoch": 4.157650695517774, + "grad_norm": 0.08681264189018105, + "learning_rate": 6.721741969028213e-06, + "loss": 0.348, + "step": 2690 + }, + { + "epoch": 4.15919629057187, + "grad_norm": 0.09231883798002785, + "learning_rate": 6.6978096637652134e-06, + "loss": 0.3439, + "step": 2691 + }, + { + "epoch": 4.160741885625966, + "grad_norm": 0.09883183887807469, + "learning_rate": 6.673916145671966e-06, + "loss": 0.352, + "step": 2692 + }, + { + "epoch": 4.162287480680062, + "grad_norm": 0.0929331735681632, + "learning_rate": 6.650061442577356e-06, + "loss": 0.344, + "step": 2693 + }, + { + "epoch": 4.163833075734158, + "grad_norm": 0.08868671769328039, + "learning_rate": 6.626245582265021e-06, + "loss": 0.3426, + "step": 2694 + }, + { + "epoch": 4.165378670788254, + "grad_norm": 0.09008028624081477, + "learning_rate": 6.602468592473368e-06, + "loss": 0.3479, + "step": 2695 + }, + { + "epoch": 4.166924265842349, + "grad_norm": 0.09592653299689341, + "learning_rate": 6.578730500895539e-06, + "loss": 0.3454, + "step": 2696 + }, + { + "epoch": 4.168469860896445, + "grad_norm": 0.09195938812380217, + "learning_rate": 6.555031335179362e-06, + "loss": 0.3449, + "step": 2697 + }, + { + "epoch": 4.170015455950541, + "grad_norm": 0.0979280950028967, + "learning_rate": 6.53137112292733e-06, + "loss": 0.3526, + "step": 2698 + }, + { + "epoch": 4.171561051004637, + "grad_norm": 0.09009923736562045, + "learning_rate": 6.507749891696584e-06, + "loss": 0.3509, + "step": 2699 + }, + { + "epoch": 4.1731066460587325, + "grad_norm": 0.09900863694145302, + "learning_rate": 6.484167668998838e-06, + "loss": 0.3573, + "step": 2700 + }, + { + "epoch": 4.174652241112828, + "grad_norm": 0.09786973451180507, + "learning_rate": 6.46062448230039e-06, + "loss": 0.3495, + "step": 2701 + }, + { + "epoch": 4.176197836166924, + "grad_norm": 0.08362971408681666, + "learning_rate": 6.43712035902206e-06, + "loss": 0.3458, + "step": 2702 + }, + { + "epoch": 4.17774343122102, + "grad_norm": 0.09525291321654723, + "learning_rate": 6.413655326539179e-06, + "loss": 0.3511, + "step": 2703 + }, + { + "epoch": 4.179289026275116, + "grad_norm": 0.10093476900173766, + "learning_rate": 6.390229412181548e-06, + "loss": 0.3585, + "step": 2704 + }, + { + "epoch": 4.180834621329212, + "grad_norm": 0.08978894281932101, + "learning_rate": 6.3668426432333995e-06, + "loss": 0.3548, + "step": 2705 + }, + { + "epoch": 4.182380216383308, + "grad_norm": 0.09246920589687538, + "learning_rate": 6.34349504693339e-06, + "loss": 0.3423, + "step": 2706 + }, + { + "epoch": 4.183925811437404, + "grad_norm": 0.09130903497926467, + "learning_rate": 6.32018665047454e-06, + "loss": 0.3501, + "step": 2707 + }, + { + "epoch": 4.185471406491499, + "grad_norm": 0.08660881253621892, + "learning_rate": 6.296917481004201e-06, + "loss": 0.3468, + "step": 2708 + }, + { + "epoch": 4.187017001545595, + "grad_norm": 0.08377527635446212, + "learning_rate": 6.273687565624067e-06, + "loss": 0.3542, + "step": 2709 + }, + { + "epoch": 4.188562596599691, + "grad_norm": 0.08866327627783151, + "learning_rate": 6.250496931390104e-06, + "loss": 0.3493, + "step": 2710 + }, + { + "epoch": 4.190108191653787, + "grad_norm": 0.08945807864644795, + "learning_rate": 6.227345605312489e-06, + "loss": 0.3403, + "step": 2711 + }, + { + "epoch": 4.1916537867078825, + "grad_norm": 0.0887623289261429, + "learning_rate": 6.20423361435567e-06, + "loss": 0.3528, + "step": 2712 + }, + { + "epoch": 4.193199381761978, + "grad_norm": 0.08680136607627459, + "learning_rate": 6.181160985438253e-06, + "loss": 0.3475, + "step": 2713 + }, + { + "epoch": 4.194744976816074, + "grad_norm": 0.09505680088025242, + "learning_rate": 6.158127745432994e-06, + "loss": 0.3474, + "step": 2714 + }, + { + "epoch": 4.19629057187017, + "grad_norm": 0.0910181437160273, + "learning_rate": 6.135133921166798e-06, + "loss": 0.3462, + "step": 2715 + }, + { + "epoch": 4.197836166924266, + "grad_norm": 0.08606960293504413, + "learning_rate": 6.112179539420644e-06, + "loss": 0.3506, + "step": 2716 + }, + { + "epoch": 4.199381761978362, + "grad_norm": 0.09129921444068308, + "learning_rate": 6.089264626929572e-06, + "loss": 0.3532, + "step": 2717 + }, + { + "epoch": 4.200927357032458, + "grad_norm": 0.09510092578536367, + "learning_rate": 6.0663892103826595e-06, + "loss": 0.353, + "step": 2718 + }, + { + "epoch": 4.202472952086553, + "grad_norm": 0.09601044093874106, + "learning_rate": 6.043553316422972e-06, + "loss": 0.3507, + "step": 2719 + }, + { + "epoch": 4.204018547140649, + "grad_norm": 0.10019579809178113, + "learning_rate": 6.0207569716475456e-06, + "loss": 0.3391, + "step": 2720 + }, + { + "epoch": 4.205564142194745, + "grad_norm": 0.09149772335796025, + "learning_rate": 5.998000202607372e-06, + "loss": 0.3525, + "step": 2721 + }, + { + "epoch": 4.207109737248841, + "grad_norm": 0.09678861359711653, + "learning_rate": 5.975283035807326e-06, + "loss": 0.3431, + "step": 2722 + }, + { + "epoch": 4.2086553323029365, + "grad_norm": 0.10770337561171722, + "learning_rate": 5.952605497706172e-06, + "loss": 0.3512, + "step": 2723 + }, + { + "epoch": 4.210200927357032, + "grad_norm": 0.09254643928336213, + "learning_rate": 5.929967614716506e-06, + "loss": 0.3488, + "step": 2724 + }, + { + "epoch": 4.211746522411128, + "grad_norm": 0.09719212904806131, + "learning_rate": 5.907369413204742e-06, + "loss": 0.3486, + "step": 2725 + }, + { + "epoch": 4.213292117465224, + "grad_norm": 0.10351728017384307, + "learning_rate": 5.884810919491078e-06, + "loss": 0.3498, + "step": 2726 + }, + { + "epoch": 4.21483771251932, + "grad_norm": 0.09774772910336196, + "learning_rate": 5.8622921598494764e-06, + "loss": 0.3556, + "step": 2727 + }, + { + "epoch": 4.216383307573416, + "grad_norm": 0.08249445366236399, + "learning_rate": 5.839813160507608e-06, + "loss": 0.3495, + "step": 2728 + }, + { + "epoch": 4.217928902627512, + "grad_norm": 0.08774296533632595, + "learning_rate": 5.817373947646827e-06, + "loss": 0.3456, + "step": 2729 + }, + { + "epoch": 4.219474497681608, + "grad_norm": 0.09044142173089888, + "learning_rate": 5.794974547402166e-06, + "loss": 0.3484, + "step": 2730 + }, + { + "epoch": 4.221020092735703, + "grad_norm": 0.0925796711635186, + "learning_rate": 5.77261498586228e-06, + "loss": 0.3553, + "step": 2731 + }, + { + "epoch": 4.222565687789799, + "grad_norm": 0.08649124670615871, + "learning_rate": 5.750295289069416e-06, + "loss": 0.3545, + "step": 2732 + }, + { + "epoch": 4.224111282843895, + "grad_norm": 0.0867973445003958, + "learning_rate": 5.728015483019396e-06, + "loss": 0.3471, + "step": 2733 + }, + { + "epoch": 4.225656877897991, + "grad_norm": 0.08862548922716557, + "learning_rate": 5.7057755936615975e-06, + "loss": 0.3452, + "step": 2734 + }, + { + "epoch": 4.2272024729520865, + "grad_norm": 0.0915530648024646, + "learning_rate": 5.683575646898884e-06, + "loss": 0.3545, + "step": 2735 + }, + { + "epoch": 4.228748068006182, + "grad_norm": 0.08613890956706916, + "learning_rate": 5.6614156685875954e-06, + "loss": 0.3463, + "step": 2736 + }, + { + "epoch": 4.230293663060278, + "grad_norm": 0.08691297495344168, + "learning_rate": 5.639295684537547e-06, + "loss": 0.3445, + "step": 2737 + }, + { + "epoch": 4.231839258114374, + "grad_norm": 0.08310811067759995, + "learning_rate": 5.617215720511962e-06, + "loss": 0.3526, + "step": 2738 + }, + { + "epoch": 4.23338485316847, + "grad_norm": 0.08343953697177237, + "learning_rate": 5.595175802227419e-06, + "loss": 0.3484, + "step": 2739 + }, + { + "epoch": 4.234930448222566, + "grad_norm": 0.0861910567941286, + "learning_rate": 5.573175955353903e-06, + "loss": 0.3522, + "step": 2740 + }, + { + "epoch": 4.236476043276662, + "grad_norm": 0.0777186794836491, + "learning_rate": 5.551216205514704e-06, + "loss": 0.3505, + "step": 2741 + }, + { + "epoch": 4.238021638330757, + "grad_norm": 0.08144096840432163, + "learning_rate": 5.529296578286407e-06, + "loss": 0.3446, + "step": 2742 + }, + { + "epoch": 4.239567233384853, + "grad_norm": 0.08835441368792356, + "learning_rate": 5.507417099198886e-06, + "loss": 0.3484, + "step": 2743 + }, + { + "epoch": 4.241112828438949, + "grad_norm": 0.08126870596595956, + "learning_rate": 5.485577793735237e-06, + "loss": 0.3505, + "step": 2744 + }, + { + "epoch": 4.242658423493045, + "grad_norm": 0.08121575444343013, + "learning_rate": 5.463778687331771e-06, + "loss": 0.352, + "step": 2745 + }, + { + "epoch": 4.244204018547141, + "grad_norm": 0.08322868248497733, + "learning_rate": 5.4420198053779735e-06, + "loss": 0.3401, + "step": 2746 + }, + { + "epoch": 4.2457496136012365, + "grad_norm": 0.09184739288112194, + "learning_rate": 5.420301173216484e-06, + "loss": 0.3507, + "step": 2747 + }, + { + "epoch": 4.247295208655332, + "grad_norm": 0.084403341690368, + "learning_rate": 5.398622816143064e-06, + "loss": 0.3518, + "step": 2748 + }, + { + "epoch": 4.248840803709428, + "grad_norm": 0.08224459987007703, + "learning_rate": 5.376984759406578e-06, + "loss": 0.3516, + "step": 2749 + }, + { + "epoch": 4.250386398763524, + "grad_norm": 0.08629277157699956, + "learning_rate": 5.355387028208938e-06, + "loss": 0.3549, + "step": 2750 + }, + { + "epoch": 4.25193199381762, + "grad_norm": 0.08185089574331944, + "learning_rate": 5.333829647705085e-06, + "loss": 0.3457, + "step": 2751 + }, + { + "epoch": 4.253477588871716, + "grad_norm": 0.08190279325500587, + "learning_rate": 5.312312643002972e-06, + "loss": 0.352, + "step": 2752 + }, + { + "epoch": 4.255023183925811, + "grad_norm": 0.0892876994651363, + "learning_rate": 5.290836039163529e-06, + "loss": 0.349, + "step": 2753 + }, + { + "epoch": 4.256568778979907, + "grad_norm": 0.08447915478996387, + "learning_rate": 5.269399861200612e-06, + "loss": 0.3459, + "step": 2754 + }, + { + "epoch": 4.258114374034003, + "grad_norm": 0.07748852606445458, + "learning_rate": 5.248004134081033e-06, + "loss": 0.3524, + "step": 2755 + }, + { + "epoch": 4.259659969088099, + "grad_norm": 0.08730300384433576, + "learning_rate": 5.226648882724448e-06, + "loss": 0.3443, + "step": 2756 + }, + { + "epoch": 4.261205564142195, + "grad_norm": 0.08505194826000924, + "learning_rate": 5.205334132003392e-06, + "loss": 0.3457, + "step": 2757 + }, + { + "epoch": 4.2627511591962906, + "grad_norm": 0.08362832295747517, + "learning_rate": 5.1840599067432215e-06, + "loss": 0.3556, + "step": 2758 + }, + { + "epoch": 4.2642967542503865, + "grad_norm": 0.08149843969736577, + "learning_rate": 5.162826231722102e-06, + "loss": 0.3517, + "step": 2759 + }, + { + "epoch": 4.265842349304482, + "grad_norm": 0.08443772007896898, + "learning_rate": 5.141633131670962e-06, + "loss": 0.35, + "step": 2760 + }, + { + "epoch": 4.267387944358578, + "grad_norm": 0.07929956288700393, + "learning_rate": 5.120480631273461e-06, + "loss": 0.3424, + "step": 2761 + }, + { + "epoch": 4.268933539412674, + "grad_norm": 0.08258796502296727, + "learning_rate": 5.0993687551660075e-06, + "loss": 0.3442, + "step": 2762 + }, + { + "epoch": 4.27047913446677, + "grad_norm": 0.08508110512188652, + "learning_rate": 5.078297527937661e-06, + "loss": 0.3528, + "step": 2763 + }, + { + "epoch": 4.272024729520865, + "grad_norm": 0.08895252203795366, + "learning_rate": 5.057266974130142e-06, + "loss": 0.3432, + "step": 2764 + }, + { + "epoch": 4.273570324574961, + "grad_norm": 0.08386139591937235, + "learning_rate": 5.036277118237833e-06, + "loss": 0.3468, + "step": 2765 + }, + { + "epoch": 4.275115919629057, + "grad_norm": 0.08920822772839218, + "learning_rate": 5.015327984707656e-06, + "loss": 0.3499, + "step": 2766 + }, + { + "epoch": 4.276661514683153, + "grad_norm": 0.08772240930867956, + "learning_rate": 4.9944195979391374e-06, + "loss": 0.3429, + "step": 2767 + }, + { + "epoch": 4.278207109737249, + "grad_norm": 0.09196911526263683, + "learning_rate": 4.973551982284366e-06, + "loss": 0.3497, + "step": 2768 + }, + { + "epoch": 4.279752704791345, + "grad_norm": 0.08052787163782076, + "learning_rate": 4.952725162047904e-06, + "loss": 0.3494, + "step": 2769 + }, + { + "epoch": 4.2812982998454405, + "grad_norm": 0.09452001276413037, + "learning_rate": 4.931939161486816e-06, + "loss": 0.3607, + "step": 2770 + }, + { + "epoch": 4.282843894899536, + "grad_norm": 0.08611093316743922, + "learning_rate": 4.911194004810642e-06, + "loss": 0.3569, + "step": 2771 + }, + { + "epoch": 4.284389489953632, + "grad_norm": 0.08298827219953163, + "learning_rate": 4.890489716181317e-06, + "loss": 0.3457, + "step": 2772 + }, + { + "epoch": 4.285935085007728, + "grad_norm": 0.08747580977147676, + "learning_rate": 4.8698263197132e-06, + "loss": 0.3484, + "step": 2773 + }, + { + "epoch": 4.287480680061824, + "grad_norm": 0.08509843362345738, + "learning_rate": 4.849203839473018e-06, + "loss": 0.3511, + "step": 2774 + }, + { + "epoch": 4.289026275115919, + "grad_norm": 0.08384040845348897, + "learning_rate": 4.828622299479833e-06, + "loss": 0.3518, + "step": 2775 + }, + { + "epoch": 4.290571870170015, + "grad_norm": 0.08225738324803471, + "learning_rate": 4.808081723705029e-06, + "loss": 0.3494, + "step": 2776 + }, + { + "epoch": 4.292117465224111, + "grad_norm": 0.08929428170089522, + "learning_rate": 4.787582136072302e-06, + "loss": 0.3468, + "step": 2777 + }, + { + "epoch": 4.293663060278207, + "grad_norm": 0.07893963360893906, + "learning_rate": 4.7671235604575735e-06, + "loss": 0.3458, + "step": 2778 + }, + { + "epoch": 4.295208655332303, + "grad_norm": 0.08680221156527582, + "learning_rate": 4.746706020689016e-06, + "loss": 0.3426, + "step": 2779 + }, + { + "epoch": 4.296754250386399, + "grad_norm": 0.0799084807848644, + "learning_rate": 4.726329540547005e-06, + "loss": 0.3573, + "step": 2780 + }, + { + "epoch": 4.298299845440495, + "grad_norm": 0.08603943550132623, + "learning_rate": 4.705994143764092e-06, + "loss": 0.3516, + "step": 2781 + }, + { + "epoch": 4.2998454404945905, + "grad_norm": 0.0848692785094291, + "learning_rate": 4.685699854024983e-06, + "loss": 0.3532, + "step": 2782 + }, + { + "epoch": 4.301391035548686, + "grad_norm": 0.08413350754023934, + "learning_rate": 4.665446694966491e-06, + "loss": 0.3563, + "step": 2783 + }, + { + "epoch": 4.302936630602782, + "grad_norm": 0.08726381849055063, + "learning_rate": 4.645234690177556e-06, + "loss": 0.3511, + "step": 2784 + }, + { + "epoch": 4.304482225656878, + "grad_norm": 0.07768866576856157, + "learning_rate": 4.625063863199159e-06, + "loss": 0.3455, + "step": 2785 + }, + { + "epoch": 4.306027820710974, + "grad_norm": 0.08596411111222517, + "learning_rate": 4.604934237524324e-06, + "loss": 0.3454, + "step": 2786 + }, + { + "epoch": 4.307573415765069, + "grad_norm": 0.07657915076559339, + "learning_rate": 4.584845836598093e-06, + "loss": 0.3411, + "step": 2787 + }, + { + "epoch": 4.309119010819165, + "grad_norm": 0.08908154360320228, + "learning_rate": 4.5647986838174904e-06, + "loss": 0.3553, + "step": 2788 + }, + { + "epoch": 4.310664605873261, + "grad_norm": 0.08381492332227229, + "learning_rate": 4.544792802531501e-06, + "loss": 0.3442, + "step": 2789 + }, + { + "epoch": 4.312210200927357, + "grad_norm": 0.07893020049897356, + "learning_rate": 4.5248282160410465e-06, + "loss": 0.3505, + "step": 2790 + }, + { + "epoch": 4.313755795981453, + "grad_norm": 0.09488830046072966, + "learning_rate": 4.504904947598938e-06, + "loss": 0.3581, + "step": 2791 + }, + { + "epoch": 4.315301391035549, + "grad_norm": 0.08400346949719387, + "learning_rate": 4.485023020409874e-06, + "loss": 0.3469, + "step": 2792 + }, + { + "epoch": 4.316846986089645, + "grad_norm": 0.09019417776449458, + "learning_rate": 4.465182457630408e-06, + "loss": 0.344, + "step": 2793 + }, + { + "epoch": 4.3183925811437405, + "grad_norm": 0.08258223886785825, + "learning_rate": 4.445383282368898e-06, + "loss": 0.3455, + "step": 2794 + }, + { + "epoch": 4.319938176197836, + "grad_norm": 0.08248713459639526, + "learning_rate": 4.425625517685501e-06, + "loss": 0.3499, + "step": 2795 + }, + { + "epoch": 4.321483771251932, + "grad_norm": 0.08953310369330902, + "learning_rate": 4.40590918659217e-06, + "loss": 0.3556, + "step": 2796 + }, + { + "epoch": 4.323029366306028, + "grad_norm": 0.0844215981666415, + "learning_rate": 4.3862343120525664e-06, + "loss": 0.3414, + "step": 2797 + }, + { + "epoch": 4.324574961360124, + "grad_norm": 0.0810130805737031, + "learning_rate": 4.36660091698208e-06, + "loss": 0.3603, + "step": 2798 + }, + { + "epoch": 4.326120556414219, + "grad_norm": 0.08542137093061943, + "learning_rate": 4.347009024247801e-06, + "loss": 0.346, + "step": 2799 + }, + { + "epoch": 4.327666151468315, + "grad_norm": 0.08815077895427041, + "learning_rate": 4.327458656668464e-06, + "loss": 0.3579, + "step": 2800 + }, + { + "epoch": 4.329211746522411, + "grad_norm": 0.07818746188271519, + "learning_rate": 4.307949837014445e-06, + "loss": 0.3483, + "step": 2801 + }, + { + "epoch": 4.330757341576507, + "grad_norm": 0.0808695254180279, + "learning_rate": 4.288482588007732e-06, + "loss": 0.3522, + "step": 2802 + }, + { + "epoch": 4.332302936630603, + "grad_norm": 0.08608424581458866, + "learning_rate": 4.269056932321891e-06, + "loss": 0.3478, + "step": 2803 + }, + { + "epoch": 4.333848531684699, + "grad_norm": 0.07895846412020385, + "learning_rate": 4.249672892582042e-06, + "loss": 0.348, + "step": 2804 + }, + { + "epoch": 4.3353941267387945, + "grad_norm": 0.07942966336374079, + "learning_rate": 4.230330491364853e-06, + "loss": 0.3481, + "step": 2805 + }, + { + "epoch": 4.3369397217928904, + "grad_norm": 0.0800726247678195, + "learning_rate": 4.211029751198474e-06, + "loss": 0.3468, + "step": 2806 + }, + { + "epoch": 4.338485316846986, + "grad_norm": 0.07880525487945812, + "learning_rate": 4.19177069456254e-06, + "loss": 0.3501, + "step": 2807 + }, + { + "epoch": 4.340030911901082, + "grad_norm": 0.07822112072857317, + "learning_rate": 4.172553343888139e-06, + "loss": 0.3456, + "step": 2808 + }, + { + "epoch": 4.341576506955178, + "grad_norm": 0.07982983239806374, + "learning_rate": 4.153377721557781e-06, + "loss": 0.3499, + "step": 2809 + }, + { + "epoch": 4.343122102009273, + "grad_norm": 0.08397108661721098, + "learning_rate": 4.134243849905373e-06, + "loss": 0.3489, + "step": 2810 + }, + { + "epoch": 4.344667697063369, + "grad_norm": 0.08417824581971808, + "learning_rate": 4.115151751216195e-06, + "loss": 0.3409, + "step": 2811 + }, + { + "epoch": 4.346213292117465, + "grad_norm": 0.08356380649469387, + "learning_rate": 4.0961014477268876e-06, + "loss": 0.3527, + "step": 2812 + }, + { + "epoch": 4.347758887171561, + "grad_norm": 0.08337450375715438, + "learning_rate": 4.0770929616253954e-06, + "loss": 0.3513, + "step": 2813 + }, + { + "epoch": 4.349304482225657, + "grad_norm": 0.0794790520176344, + "learning_rate": 4.0581263150509634e-06, + "loss": 0.3441, + "step": 2814 + }, + { + "epoch": 4.350850077279753, + "grad_norm": 0.07989596049121159, + "learning_rate": 4.0392015300941075e-06, + "loss": 0.3535, + "step": 2815 + }, + { + "epoch": 4.352395672333849, + "grad_norm": 0.08258184784831214, + "learning_rate": 4.020318628796589e-06, + "loss": 0.3538, + "step": 2816 + }, + { + "epoch": 4.3539412673879445, + "grad_norm": 0.08321302309092998, + "learning_rate": 4.0014776331513786e-06, + "loss": 0.3413, + "step": 2817 + }, + { + "epoch": 4.35548686244204, + "grad_norm": 0.07972304848581438, + "learning_rate": 3.982678565102651e-06, + "loss": 0.3432, + "step": 2818 + }, + { + "epoch": 4.357032457496136, + "grad_norm": 0.07740206915020002, + "learning_rate": 3.963921446545742e-06, + "loss": 0.3469, + "step": 2819 + }, + { + "epoch": 4.358578052550232, + "grad_norm": 0.08194551709179719, + "learning_rate": 3.945206299327118e-06, + "loss": 0.3483, + "step": 2820 + }, + { + "epoch": 4.360123647604327, + "grad_norm": 0.07644868697025502, + "learning_rate": 3.9265331452444e-06, + "loss": 0.3476, + "step": 2821 + }, + { + "epoch": 4.361669242658423, + "grad_norm": 0.08526880526334188, + "learning_rate": 3.9079020060462406e-06, + "loss": 0.3516, + "step": 2822 + }, + { + "epoch": 4.363214837712519, + "grad_norm": 0.08008458630230256, + "learning_rate": 3.889312903432396e-06, + "loss": 0.352, + "step": 2823 + }, + { + "epoch": 4.364760432766615, + "grad_norm": 0.07887376778827801, + "learning_rate": 3.870765859053669e-06, + "loss": 0.3506, + "step": 2824 + }, + { + "epoch": 4.366306027820711, + "grad_norm": 0.08405949873945731, + "learning_rate": 3.85226089451185e-06, + "loss": 0.3445, + "step": 2825 + }, + { + "epoch": 4.367851622874807, + "grad_norm": 0.07764345620712967, + "learning_rate": 3.833798031359725e-06, + "loss": 0.3501, + "step": 2826 + }, + { + "epoch": 4.369397217928903, + "grad_norm": 0.08110470084261369, + "learning_rate": 3.815377291101072e-06, + "loss": 0.353, + "step": 2827 + }, + { + "epoch": 4.370942812982999, + "grad_norm": 0.0837793846570553, + "learning_rate": 3.7969986951905725e-06, + "loss": 0.3516, + "step": 2828 + }, + { + "epoch": 4.3724884080370945, + "grad_norm": 0.07721899926190336, + "learning_rate": 3.7786622650338412e-06, + "loss": 0.3487, + "step": 2829 + }, + { + "epoch": 4.37403400309119, + "grad_norm": 0.08428388240342731, + "learning_rate": 3.7603680219873816e-06, + "loss": 0.3467, + "step": 2830 + }, + { + "epoch": 4.375579598145286, + "grad_norm": 0.08230952833359399, + "learning_rate": 3.74211598735855e-06, + "loss": 0.3516, + "step": 2831 + }, + { + "epoch": 4.377125193199381, + "grad_norm": 0.08388381109891997, + "learning_rate": 3.7239061824055456e-06, + "loss": 0.3575, + "step": 2832 + }, + { + "epoch": 4.378670788253477, + "grad_norm": 0.08382982365844328, + "learning_rate": 3.7057386283374074e-06, + "loss": 0.3484, + "step": 2833 + }, + { + "epoch": 4.380216383307573, + "grad_norm": 0.07995433671336721, + "learning_rate": 3.687613346313934e-06, + "loss": 0.3388, + "step": 2834 + }, + { + "epoch": 4.381761978361669, + "grad_norm": 0.07553437652817582, + "learning_rate": 3.6695303574457008e-06, + "loss": 0.3466, + "step": 2835 + }, + { + "epoch": 4.383307573415765, + "grad_norm": 0.0832900184978018, + "learning_rate": 3.651489682794025e-06, + "loss": 0.3517, + "step": 2836 + }, + { + "epoch": 4.384853168469861, + "grad_norm": 0.08166599537889928, + "learning_rate": 3.633491343370934e-06, + "loss": 0.3519, + "step": 2837 + }, + { + "epoch": 4.386398763523957, + "grad_norm": 0.08317565714983777, + "learning_rate": 3.615535360139162e-06, + "loss": 0.3503, + "step": 2838 + }, + { + "epoch": 4.387944358578053, + "grad_norm": 0.08332970138051007, + "learning_rate": 3.5976217540120863e-06, + "loss": 0.3444, + "step": 2839 + }, + { + "epoch": 4.3894899536321486, + "grad_norm": 0.07937650477051997, + "learning_rate": 3.5797505458537597e-06, + "loss": 0.3473, + "step": 2840 + }, + { + "epoch": 4.3910355486862445, + "grad_norm": 0.084930685850289, + "learning_rate": 3.5619217564788255e-06, + "loss": 0.3528, + "step": 2841 + }, + { + "epoch": 4.39258114374034, + "grad_norm": 0.08409427647011232, + "learning_rate": 3.54413540665254e-06, + "loss": 0.3444, + "step": 2842 + }, + { + "epoch": 4.394126738794436, + "grad_norm": 0.08577284327271625, + "learning_rate": 3.526391517090719e-06, + "loss": 0.3357, + "step": 2843 + }, + { + "epoch": 4.395672333848531, + "grad_norm": 0.08333543187416191, + "learning_rate": 3.5086901084597293e-06, + "loss": 0.3453, + "step": 2844 + }, + { + "epoch": 4.397217928902627, + "grad_norm": 0.07992037276703519, + "learning_rate": 3.491031201376451e-06, + "loss": 0.3454, + "step": 2845 + }, + { + "epoch": 4.398763523956723, + "grad_norm": 0.08209319555463147, + "learning_rate": 3.473414816408287e-06, + "loss": 0.3402, + "step": 2846 + }, + { + "epoch": 4.400309119010819, + "grad_norm": 0.08490154321268921, + "learning_rate": 3.455840974073086e-06, + "loss": 0.3427, + "step": 2847 + }, + { + "epoch": 4.401854714064915, + "grad_norm": 0.07698769938158478, + "learning_rate": 3.4383096948391593e-06, + "loss": 0.3441, + "step": 2848 + }, + { + "epoch": 4.403400309119011, + "grad_norm": 0.07880916167937885, + "learning_rate": 3.4208209991252626e-06, + "loss": 0.3433, + "step": 2849 + }, + { + "epoch": 4.404945904173107, + "grad_norm": 0.08174402671114551, + "learning_rate": 3.4033749073005205e-06, + "loss": 0.3469, + "step": 2850 + }, + { + "epoch": 4.406491499227203, + "grad_norm": 0.08622213318277683, + "learning_rate": 3.3859714396844434e-06, + "loss": 0.3458, + "step": 2851 + }, + { + "epoch": 4.4080370942812985, + "grad_norm": 0.0837012341239019, + "learning_rate": 3.3686106165469323e-06, + "loss": 0.3588, + "step": 2852 + }, + { + "epoch": 4.409582689335394, + "grad_norm": 0.07944231475429726, + "learning_rate": 3.3512924581081775e-06, + "loss": 0.3477, + "step": 2853 + }, + { + "epoch": 4.41112828438949, + "grad_norm": 0.08178179824143911, + "learning_rate": 3.3340169845386926e-06, + "loss": 0.356, + "step": 2854 + }, + { + "epoch": 4.412673879443586, + "grad_norm": 0.08582291436062328, + "learning_rate": 3.3167842159592946e-06, + "loss": 0.3455, + "step": 2855 + }, + { + "epoch": 4.414219474497681, + "grad_norm": 0.08762946770722586, + "learning_rate": 3.2995941724410384e-06, + "loss": 0.3526, + "step": 2856 + }, + { + "epoch": 4.415765069551777, + "grad_norm": 0.08316740590899953, + "learning_rate": 3.2824468740052072e-06, + "loss": 0.3476, + "step": 2857 + }, + { + "epoch": 4.417310664605873, + "grad_norm": 0.08262977766404131, + "learning_rate": 3.2653423406233364e-06, + "loss": 0.3479, + "step": 2858 + }, + { + "epoch": 4.418856259659969, + "grad_norm": 0.08268620916274816, + "learning_rate": 3.2482805922171257e-06, + "loss": 0.3493, + "step": 2859 + }, + { + "epoch": 4.420401854714065, + "grad_norm": 0.08850892909930494, + "learning_rate": 3.231261648658439e-06, + "loss": 0.3508, + "step": 2860 + }, + { + "epoch": 4.421947449768161, + "grad_norm": 0.08479502639571267, + "learning_rate": 3.2142855297693145e-06, + "loss": 0.3495, + "step": 2861 + }, + { + "epoch": 4.423493044822257, + "grad_norm": 0.08113628583032982, + "learning_rate": 3.1973522553218774e-06, + "loss": 0.3584, + "step": 2862 + }, + { + "epoch": 4.425038639876353, + "grad_norm": 0.08072752020645717, + "learning_rate": 3.180461845038374e-06, + "loss": 0.3488, + "step": 2863 + }, + { + "epoch": 4.4265842349304485, + "grad_norm": 0.07667822418785249, + "learning_rate": 3.1636143185911216e-06, + "loss": 0.3478, + "step": 2864 + }, + { + "epoch": 4.428129829984544, + "grad_norm": 0.07762761353502987, + "learning_rate": 3.1468096956024863e-06, + "loss": 0.3513, + "step": 2865 + }, + { + "epoch": 4.42967542503864, + "grad_norm": 0.08132451497698748, + "learning_rate": 3.1300479956448693e-06, + "loss": 0.3463, + "step": 2866 + }, + { + "epoch": 4.431221020092735, + "grad_norm": 0.08253670758213165, + "learning_rate": 3.1133292382406678e-06, + "loss": 0.3528, + "step": 2867 + }, + { + "epoch": 4.432766615146831, + "grad_norm": 0.08377799094948238, + "learning_rate": 3.096653442862292e-06, + "loss": 0.3543, + "step": 2868 + }, + { + "epoch": 4.434312210200927, + "grad_norm": 0.08446725971646331, + "learning_rate": 3.080020628932081e-06, + "loss": 0.352, + "step": 2869 + }, + { + "epoch": 4.435857805255023, + "grad_norm": 0.08136347996807272, + "learning_rate": 3.0634308158223303e-06, + "loss": 0.3443, + "step": 2870 + }, + { + "epoch": 4.437403400309119, + "grad_norm": 0.08057277380348657, + "learning_rate": 3.04688402285525e-06, + "loss": 0.3569, + "step": 2871 + }, + { + "epoch": 4.438948995363215, + "grad_norm": 0.08140459786575811, + "learning_rate": 3.030380269302944e-06, + "loss": 0.3377, + "step": 2872 + }, + { + "epoch": 4.440494590417311, + "grad_norm": 0.08261305046186905, + "learning_rate": 3.0139195743873826e-06, + "loss": 0.349, + "step": 2873 + }, + { + "epoch": 4.442040185471407, + "grad_norm": 0.08485741910587392, + "learning_rate": 2.9975019572804042e-06, + "loss": 0.3546, + "step": 2874 + }, + { + "epoch": 4.443585780525503, + "grad_norm": 0.07873815679095189, + "learning_rate": 2.98112743710365e-06, + "loss": 0.3492, + "step": 2875 + }, + { + "epoch": 4.4451313755795985, + "grad_norm": 0.07944709390269125, + "learning_rate": 2.9647960329285806e-06, + "loss": 0.3489, + "step": 2876 + }, + { + "epoch": 4.446676970633694, + "grad_norm": 0.0812646396088516, + "learning_rate": 2.948507763776438e-06, + "loss": 0.3465, + "step": 2877 + }, + { + "epoch": 4.448222565687789, + "grad_norm": 0.07652264016970232, + "learning_rate": 2.9322626486182158e-06, + "loss": 0.3574, + "step": 2878 + }, + { + "epoch": 4.449768160741885, + "grad_norm": 0.08365648675702712, + "learning_rate": 2.916060706374655e-06, + "loss": 0.3457, + "step": 2879 + }, + { + "epoch": 4.451313755795981, + "grad_norm": 0.07451328373134032, + "learning_rate": 2.8999019559162156e-06, + "loss": 0.3497, + "step": 2880 + }, + { + "epoch": 4.452859350850077, + "grad_norm": 0.07992407614754655, + "learning_rate": 2.8837864160630435e-06, + "loss": 0.3532, + "step": 2881 + }, + { + "epoch": 4.454404945904173, + "grad_norm": 0.07905390351920381, + "learning_rate": 2.8677141055849554e-06, + "loss": 0.3463, + "step": 2882 + }, + { + "epoch": 4.455950540958269, + "grad_norm": 0.07547533981426742, + "learning_rate": 2.851685043201431e-06, + "loss": 0.3494, + "step": 2883 + }, + { + "epoch": 4.457496136012365, + "grad_norm": 0.07804805683329721, + "learning_rate": 2.83569924758158e-06, + "loss": 0.3424, + "step": 2884 + }, + { + "epoch": 4.459041731066461, + "grad_norm": 0.08078161181159066, + "learning_rate": 2.8197567373440882e-06, + "loss": 0.3435, + "step": 2885 + }, + { + "epoch": 4.460587326120557, + "grad_norm": 0.08360221957028523, + "learning_rate": 2.803857531057266e-06, + "loss": 0.3538, + "step": 2886 + }, + { + "epoch": 4.4621329211746525, + "grad_norm": 0.09720199779553902, + "learning_rate": 2.788001647238967e-06, + "loss": 0.3486, + "step": 2887 + }, + { + "epoch": 4.4636785162287484, + "grad_norm": 0.07824968323511246, + "learning_rate": 2.7721891043565795e-06, + "loss": 0.3566, + "step": 2888 + }, + { + "epoch": 4.4652241112828435, + "grad_norm": 0.07520904847589219, + "learning_rate": 2.756419920827047e-06, + "loss": 0.3447, + "step": 2889 + }, + { + "epoch": 4.466769706336939, + "grad_norm": 0.07927334007453044, + "learning_rate": 2.740694115016771e-06, + "loss": 0.3518, + "step": 2890 + }, + { + "epoch": 4.468315301391035, + "grad_norm": 0.07346839525159184, + "learning_rate": 2.7250117052416557e-06, + "loss": 0.3524, + "step": 2891 + }, + { + "epoch": 4.469860896445131, + "grad_norm": 0.08347394736389456, + "learning_rate": 2.709372709767055e-06, + "loss": 0.3476, + "step": 2892 + }, + { + "epoch": 4.471406491499227, + "grad_norm": 0.07785722815714517, + "learning_rate": 2.693777146807759e-06, + "loss": 0.3524, + "step": 2893 + }, + { + "epoch": 4.472952086553323, + "grad_norm": 0.07551340491347229, + "learning_rate": 2.6782250345279616e-06, + "loss": 0.3573, + "step": 2894 + }, + { + "epoch": 4.474497681607419, + "grad_norm": 0.07976315821083518, + "learning_rate": 2.662716391041262e-06, + "loss": 0.3475, + "step": 2895 + }, + { + "epoch": 4.476043276661515, + "grad_norm": 0.07716251071719255, + "learning_rate": 2.6472512344106394e-06, + "loss": 0.3579, + "step": 2896 + }, + { + "epoch": 4.477588871715611, + "grad_norm": 0.0829097665496094, + "learning_rate": 2.631829582648404e-06, + "loss": 0.3471, + "step": 2897 + }, + { + "epoch": 4.479134466769707, + "grad_norm": 0.08103423871659719, + "learning_rate": 2.616451453716202e-06, + "loss": 0.3559, + "step": 2898 + }, + { + "epoch": 4.4806800618238025, + "grad_norm": 0.08553687067102614, + "learning_rate": 2.6011168655249906e-06, + "loss": 0.3552, + "step": 2899 + }, + { + "epoch": 4.4822256568778975, + "grad_norm": 0.08428872138223713, + "learning_rate": 2.5858258359350207e-06, + "loss": 0.3479, + "step": 2900 + }, + { + "epoch": 4.483771251931993, + "grad_norm": 0.07736643977622588, + "learning_rate": 2.5705783827557886e-06, + "loss": 0.3443, + "step": 2901 + }, + { + "epoch": 4.485316846986089, + "grad_norm": 0.07671781625871242, + "learning_rate": 2.555374523746075e-06, + "loss": 0.3378, + "step": 2902 + }, + { + "epoch": 4.486862442040185, + "grad_norm": 0.08055811720332556, + "learning_rate": 2.5402142766138528e-06, + "loss": 0.3573, + "step": 2903 + }, + { + "epoch": 4.488408037094281, + "grad_norm": 0.0799311392726142, + "learning_rate": 2.525097659016309e-06, + "loss": 0.3501, + "step": 2904 + }, + { + "epoch": 4.489953632148377, + "grad_norm": 0.07329363549243337, + "learning_rate": 2.510024688559822e-06, + "loss": 0.3442, + "step": 2905 + }, + { + "epoch": 4.491499227202473, + "grad_norm": 0.0786719091481724, + "learning_rate": 2.4949953827999276e-06, + "loss": 0.3476, + "step": 2906 + }, + { + "epoch": 4.493044822256569, + "grad_norm": 0.08454925530472679, + "learning_rate": 2.480009759241302e-06, + "loss": 0.3542, + "step": 2907 + }, + { + "epoch": 4.494590417310665, + "grad_norm": 0.08074136205065753, + "learning_rate": 2.4650678353377622e-06, + "loss": 0.3495, + "step": 2908 + }, + { + "epoch": 4.496136012364761, + "grad_norm": 0.07288391733839879, + "learning_rate": 2.4501696284922094e-06, + "loss": 0.3407, + "step": 2909 + }, + { + "epoch": 4.497681607418857, + "grad_norm": 0.0756241829800099, + "learning_rate": 2.4353151560566256e-06, + "loss": 0.3507, + "step": 2910 + }, + { + "epoch": 4.4992272024729525, + "grad_norm": 0.07722698742509931, + "learning_rate": 2.420504435332074e-06, + "loss": 0.3567, + "step": 2911 + }, + { + "epoch": 4.500772797527048, + "grad_norm": 0.08092853920519424, + "learning_rate": 2.4057374835686575e-06, + "loss": 0.3529, + "step": 2912 + }, + { + "epoch": 4.502318392581143, + "grad_norm": 0.07564872276180502, + "learning_rate": 2.391014317965468e-06, + "loss": 0.3551, + "step": 2913 + }, + { + "epoch": 4.503863987635239, + "grad_norm": 0.0747734888669493, + "learning_rate": 2.376334955670645e-06, + "loss": 0.3502, + "step": 2914 + }, + { + "epoch": 4.505409582689335, + "grad_norm": 0.07262912762423096, + "learning_rate": 2.361699413781282e-06, + "loss": 0.3466, + "step": 2915 + }, + { + "epoch": 4.506955177743431, + "grad_norm": 0.087234776281415, + "learning_rate": 2.3471077093434412e-06, + "loss": 0.3452, + "step": 2916 + }, + { + "epoch": 4.508500772797527, + "grad_norm": 0.07682097557507851, + "learning_rate": 2.3325598593521324e-06, + "loss": 0.3456, + "step": 2917 + }, + { + "epoch": 4.510046367851623, + "grad_norm": 0.07599485976755793, + "learning_rate": 2.3180558807512864e-06, + "loss": 0.3437, + "step": 2918 + }, + { + "epoch": 4.511591962905719, + "grad_norm": 0.07910867327561365, + "learning_rate": 2.303595790433728e-06, + "loss": 0.3495, + "step": 2919 + }, + { + "epoch": 4.513137557959815, + "grad_norm": 0.07673517152884492, + "learning_rate": 2.289179605241172e-06, + "loss": 0.3449, + "step": 2920 + }, + { + "epoch": 4.514683153013911, + "grad_norm": 0.08393290485261232, + "learning_rate": 2.2748073419642e-06, + "loss": 0.3519, + "step": 2921 + }, + { + "epoch": 4.516228748068007, + "grad_norm": 0.08053111068877837, + "learning_rate": 2.260479017342232e-06, + "loss": 0.3416, + "step": 2922 + }, + { + "epoch": 4.5177743431221025, + "grad_norm": 0.07617643912713123, + "learning_rate": 2.2461946480635087e-06, + "loss": 0.3552, + "step": 2923 + }, + { + "epoch": 4.5193199381761975, + "grad_norm": 0.08085550641476205, + "learning_rate": 2.231954250765096e-06, + "loss": 0.3627, + "step": 2924 + }, + { + "epoch": 4.520865533230293, + "grad_norm": 0.07409947109933004, + "learning_rate": 2.2177578420328195e-06, + "loss": 0.349, + "step": 2925 + }, + { + "epoch": 4.522411128284389, + "grad_norm": 0.0753170008154408, + "learning_rate": 2.203605438401288e-06, + "loss": 0.342, + "step": 2926 + }, + { + "epoch": 4.523956723338485, + "grad_norm": 0.07808859057256458, + "learning_rate": 2.1894970563538466e-06, + "loss": 0.3515, + "step": 2927 + }, + { + "epoch": 4.525502318392581, + "grad_norm": 0.08034407241190529, + "learning_rate": 2.1754327123225807e-06, + "loss": 0.3475, + "step": 2928 + }, + { + "epoch": 4.527047913446677, + "grad_norm": 0.07768682765516363, + "learning_rate": 2.161412422688267e-06, + "loss": 0.3479, + "step": 2929 + }, + { + "epoch": 4.528593508500773, + "grad_norm": 0.07794096057951751, + "learning_rate": 2.1474362037804e-06, + "loss": 0.3468, + "step": 2930 + }, + { + "epoch": 4.530139103554869, + "grad_norm": 0.07296325597840433, + "learning_rate": 2.133504071877113e-06, + "loss": 0.3465, + "step": 2931 + }, + { + "epoch": 4.531684698608965, + "grad_norm": 0.07538263425855553, + "learning_rate": 2.119616043205208e-06, + "loss": 0.3457, + "step": 2932 + }, + { + "epoch": 4.533230293663061, + "grad_norm": 0.08607373910052346, + "learning_rate": 2.105772133940125e-06, + "loss": 0.3536, + "step": 2933 + }, + { + "epoch": 4.5347758887171565, + "grad_norm": 0.07759588754302951, + "learning_rate": 2.091972360205903e-06, + "loss": 0.3473, + "step": 2934 + }, + { + "epoch": 4.5363214837712516, + "grad_norm": 0.07607434840567591, + "learning_rate": 2.078216738075178e-06, + "loss": 0.3453, + "step": 2935 + }, + { + "epoch": 4.5378670788253475, + "grad_norm": 0.07154002196069166, + "learning_rate": 2.0645052835691804e-06, + "loss": 0.3433, + "step": 2936 + }, + { + "epoch": 4.539412673879443, + "grad_norm": 0.07558474458105977, + "learning_rate": 2.050838012657681e-06, + "loss": 0.3453, + "step": 2937 + }, + { + "epoch": 4.540958268933539, + "grad_norm": 0.07190207377118617, + "learning_rate": 2.037214941258987e-06, + "loss": 0.3505, + "step": 2938 + }, + { + "epoch": 4.542503863987635, + "grad_norm": 0.07655013874782704, + "learning_rate": 2.023636085239957e-06, + "loss": 0.3546, + "step": 2939 + }, + { + "epoch": 4.544049459041731, + "grad_norm": 0.08127970977854401, + "learning_rate": 2.0101014604159053e-06, + "loss": 0.3489, + "step": 2940 + }, + { + "epoch": 4.545595054095827, + "grad_norm": 0.07679116403672862, + "learning_rate": 1.996611082550657e-06, + "loss": 0.3424, + "step": 2941 + }, + { + "epoch": 4.547140649149923, + "grad_norm": 0.07395008637835969, + "learning_rate": 1.9831649673565102e-06, + "loss": 0.3481, + "step": 2942 + }, + { + "epoch": 4.548686244204019, + "grad_norm": 0.12737252523291856, + "learning_rate": 1.9697631304941913e-06, + "loss": 0.357, + "step": 2943 + }, + { + "epoch": 4.550231839258115, + "grad_norm": 0.07563447201100297, + "learning_rate": 1.956405587572863e-06, + "loss": 0.3421, + "step": 2944 + }, + { + "epoch": 4.551777434312211, + "grad_norm": 0.07965408053300205, + "learning_rate": 1.9430923541501024e-06, + "loss": 0.3471, + "step": 2945 + }, + { + "epoch": 4.553323029366306, + "grad_norm": 0.07390792975317896, + "learning_rate": 1.92982344573188e-06, + "loss": 0.3479, + "step": 2946 + }, + { + "epoch": 4.5548686244204015, + "grad_norm": 0.0766260842638226, + "learning_rate": 1.916598877772531e-06, + "loss": 0.3483, + "step": 2947 + }, + { + "epoch": 4.556414219474497, + "grad_norm": 0.07754899335989722, + "learning_rate": 1.9034186656747478e-06, + "loss": 0.3486, + "step": 2948 + }, + { + "epoch": 4.557959814528593, + "grad_norm": 0.07480599837503225, + "learning_rate": 1.8902828247895754e-06, + "loss": 0.3468, + "step": 2949 + }, + { + "epoch": 4.559505409582689, + "grad_norm": 0.07774194619712663, + "learning_rate": 1.877191370416367e-06, + "loss": 0.3494, + "step": 2950 + }, + { + "epoch": 4.561051004636785, + "grad_norm": 0.0782130516852238, + "learning_rate": 1.8641443178027784e-06, + "loss": 0.3416, + "step": 2951 + }, + { + "epoch": 4.562596599690881, + "grad_norm": 0.07633523238815748, + "learning_rate": 1.8511416821447615e-06, + "loss": 0.3521, + "step": 2952 + }, + { + "epoch": 4.564142194744977, + "grad_norm": 0.07520877479556337, + "learning_rate": 1.838183478586526e-06, + "loss": 0.3526, + "step": 2953 + }, + { + "epoch": 4.565687789799073, + "grad_norm": 0.07414563679406655, + "learning_rate": 1.8252697222205328e-06, + "loss": 0.349, + "step": 2954 + }, + { + "epoch": 4.567233384853169, + "grad_norm": 0.07843850991489067, + "learning_rate": 1.8124004280874795e-06, + "loss": 0.3524, + "step": 2955 + }, + { + "epoch": 4.568778979907265, + "grad_norm": 0.0764585912548359, + "learning_rate": 1.7995756111762697e-06, + "loss": 0.3497, + "step": 2956 + }, + { + "epoch": 4.57032457496136, + "grad_norm": 0.0736996722848919, + "learning_rate": 1.786795286424008e-06, + "loss": 0.3523, + "step": 2957 + }, + { + "epoch": 4.571870170015456, + "grad_norm": 0.0742984756485489, + "learning_rate": 1.7740594687159872e-06, + "loss": 0.3439, + "step": 2958 + }, + { + "epoch": 4.5734157650695515, + "grad_norm": 0.07351571761008804, + "learning_rate": 1.7613681728856536e-06, + "loss": 0.3435, + "step": 2959 + }, + { + "epoch": 4.574961360123647, + "grad_norm": 0.07529382981935444, + "learning_rate": 1.7487214137146003e-06, + "loss": 0.347, + "step": 2960 + }, + { + "epoch": 4.576506955177743, + "grad_norm": 0.07416963134278001, + "learning_rate": 1.7361192059325472e-06, + "loss": 0.345, + "step": 2961 + }, + { + "epoch": 4.578052550231839, + "grad_norm": 0.07422932118402725, + "learning_rate": 1.7235615642173309e-06, + "loss": 0.3455, + "step": 2962 + }, + { + "epoch": 4.579598145285935, + "grad_norm": 0.07331993504490666, + "learning_rate": 1.7110485031948698e-06, + "loss": 0.3458, + "step": 2963 + }, + { + "epoch": 4.581143740340031, + "grad_norm": 0.07421717020868522, + "learning_rate": 1.698580037439177e-06, + "loss": 0.3474, + "step": 2964 + }, + { + "epoch": 4.582689335394127, + "grad_norm": 0.07909801885727724, + "learning_rate": 1.686156181472307e-06, + "loss": 0.3585, + "step": 2965 + }, + { + "epoch": 4.584234930448223, + "grad_norm": 0.07327136305060182, + "learning_rate": 1.6737769497643696e-06, + "loss": 0.3518, + "step": 2966 + }, + { + "epoch": 4.585780525502319, + "grad_norm": 0.07486603082792806, + "learning_rate": 1.6614423567335026e-06, + "loss": 0.3508, + "step": 2967 + }, + { + "epoch": 4.587326120556414, + "grad_norm": 0.07127235342834087, + "learning_rate": 1.6491524167458318e-06, + "loss": 0.3451, + "step": 2968 + }, + { + "epoch": 4.58887171561051, + "grad_norm": 0.07671073502051187, + "learning_rate": 1.6369071441154983e-06, + "loss": 0.3404, + "step": 2969 + }, + { + "epoch": 4.590417310664606, + "grad_norm": 0.07235205912469946, + "learning_rate": 1.6247065531046135e-06, + "loss": 0.3537, + "step": 2970 + }, + { + "epoch": 4.5919629057187015, + "grad_norm": 0.07405134880832523, + "learning_rate": 1.6125506579232418e-06, + "loss": 0.3473, + "step": 2971 + }, + { + "epoch": 4.593508500772797, + "grad_norm": 0.07586263584542462, + "learning_rate": 1.600439472729396e-06, + "loss": 0.3481, + "step": 2972 + }, + { + "epoch": 4.595054095826893, + "grad_norm": 0.07437360593397314, + "learning_rate": 1.5883730116290142e-06, + "loss": 0.353, + "step": 2973 + }, + { + "epoch": 4.596599690880989, + "grad_norm": 0.07336863791455472, + "learning_rate": 1.576351288675939e-06, + "loss": 0.3537, + "step": 2974 + }, + { + "epoch": 4.598145285935085, + "grad_norm": 0.07294086828924651, + "learning_rate": 1.5643743178719216e-06, + "loss": 0.348, + "step": 2975 + }, + { + "epoch": 4.599690880989181, + "grad_norm": 0.08671386820576314, + "learning_rate": 1.552442113166559e-06, + "loss": 0.3527, + "step": 2976 + }, + { + "epoch": 4.601236476043277, + "grad_norm": 0.0735449095800472, + "learning_rate": 1.540554688457343e-06, + "loss": 0.3452, + "step": 2977 + }, + { + "epoch": 4.602782071097373, + "grad_norm": 0.07394435022116418, + "learning_rate": 1.528712057589594e-06, + "loss": 0.3598, + "step": 2978 + }, + { + "epoch": 4.604327666151468, + "grad_norm": 0.07358602608758982, + "learning_rate": 1.5169142343564569e-06, + "loss": 0.3441, + "step": 2979 + }, + { + "epoch": 4.605873261205565, + "grad_norm": 0.07224280424680504, + "learning_rate": 1.505161232498904e-06, + "loss": 0.3494, + "step": 2980 + }, + { + "epoch": 4.60741885625966, + "grad_norm": 0.07744999403223182, + "learning_rate": 1.4934530657056922e-06, + "loss": 0.3557, + "step": 2981 + }, + { + "epoch": 4.6089644513137555, + "grad_norm": 0.07386541801531468, + "learning_rate": 1.481789747613358e-06, + "loss": 0.3466, + "step": 2982 + }, + { + "epoch": 4.6105100463678514, + "grad_norm": 0.07777540929230686, + "learning_rate": 1.4701712918062084e-06, + "loss": 0.3617, + "step": 2983 + }, + { + "epoch": 4.612055641421947, + "grad_norm": 0.07614144875206677, + "learning_rate": 1.4585977118162985e-06, + "loss": 0.3602, + "step": 2984 + }, + { + "epoch": 4.613601236476043, + "grad_norm": 0.07275324191348273, + "learning_rate": 1.4470690211234062e-06, + "loss": 0.3478, + "step": 2985 + }, + { + "epoch": 4.615146831530139, + "grad_norm": 0.07180343424786612, + "learning_rate": 1.4355852331550479e-06, + "loss": 0.3454, + "step": 2986 + }, + { + "epoch": 4.616692426584235, + "grad_norm": 0.07515994489808035, + "learning_rate": 1.4241463612864225e-06, + "loss": 0.3487, + "step": 2987 + }, + { + "epoch": 4.618238021638331, + "grad_norm": 0.07784492999599076, + "learning_rate": 1.4127524188404196e-06, + "loss": 0.3488, + "step": 2988 + }, + { + "epoch": 4.619783616692427, + "grad_norm": 0.07140157858302701, + "learning_rate": 1.4014034190876057e-06, + "loss": 0.3451, + "step": 2989 + }, + { + "epoch": 4.621329211746523, + "grad_norm": 2.0047265286912013, + "learning_rate": 1.3900993752461944e-06, + "loss": 0.3687, + "step": 2990 + }, + { + "epoch": 4.622874806800619, + "grad_norm": 0.07418079401508136, + "learning_rate": 1.3788403004820362e-06, + "loss": 0.353, + "step": 2991 + }, + { + "epoch": 4.624420401854714, + "grad_norm": 0.0711941253914657, + "learning_rate": 1.3676262079086278e-06, + "loss": 0.3534, + "step": 2992 + }, + { + "epoch": 4.62596599690881, + "grad_norm": 0.07313686483599284, + "learning_rate": 1.3564571105870505e-06, + "loss": 0.3492, + "step": 2993 + }, + { + "epoch": 4.6275115919629055, + "grad_norm": 0.0758597731241708, + "learning_rate": 1.345333021525983e-06, + "loss": 0.359, + "step": 2994 + }, + { + "epoch": 4.629057187017001, + "grad_norm": 0.0719067589383518, + "learning_rate": 1.3342539536817012e-06, + "loss": 0.3435, + "step": 2995 + }, + { + "epoch": 4.630602782071097, + "grad_norm": 0.07533804168035232, + "learning_rate": 1.323219919958021e-06, + "loss": 0.3449, + "step": 2996 + }, + { + "epoch": 4.632148377125193, + "grad_norm": 0.11897090431385539, + "learning_rate": 1.3122309332063198e-06, + "loss": 0.351, + "step": 2997 + }, + { + "epoch": 4.633693972179289, + "grad_norm": 0.07333757116667064, + "learning_rate": 1.3012870062255067e-06, + "loss": 0.3542, + "step": 2998 + }, + { + "epoch": 4.635239567233385, + "grad_norm": 0.0745402534552004, + "learning_rate": 1.290388151762012e-06, + "loss": 0.3538, + "step": 2999 + }, + { + "epoch": 4.636785162287481, + "grad_norm": 0.07571100275937757, + "learning_rate": 1.2795343825097618e-06, + "loss": 0.3585, + "step": 3000 + }, + { + "epoch": 4.638330757341577, + "grad_norm": 0.07176624771608246, + "learning_rate": 1.2687257111101859e-06, + "loss": 0.3499, + "step": 3001 + }, + { + "epoch": 4.639876352395673, + "grad_norm": 0.07402891349893713, + "learning_rate": 1.2579621501521743e-06, + "loss": 0.3459, + "step": 3002 + }, + { + "epoch": 4.641421947449768, + "grad_norm": 0.07579896281869608, + "learning_rate": 1.2472437121720727e-06, + "loss": 0.3528, + "step": 3003 + }, + { + "epoch": 4.642967542503864, + "grad_norm": 0.07434597609548542, + "learning_rate": 1.2365704096536901e-06, + "loss": 0.3514, + "step": 3004 + }, + { + "epoch": 4.64451313755796, + "grad_norm": 0.07328808377517211, + "learning_rate": 1.2259422550282562e-06, + "loss": 0.3402, + "step": 3005 + }, + { + "epoch": 4.6460587326120555, + "grad_norm": 0.07193528453543194, + "learning_rate": 1.2153592606744157e-06, + "loss": 0.3497, + "step": 3006 + }, + { + "epoch": 4.647604327666151, + "grad_norm": 0.0818183336209685, + "learning_rate": 1.2048214389182156e-06, + "loss": 0.3499, + "step": 3007 + }, + { + "epoch": 4.649149922720247, + "grad_norm": 0.07299007521326033, + "learning_rate": 1.1943288020330957e-06, + "loss": 0.3507, + "step": 3008 + }, + { + "epoch": 4.650695517774343, + "grad_norm": 0.07741361279679244, + "learning_rate": 1.183881362239858e-06, + "loss": 0.3461, + "step": 3009 + }, + { + "epoch": 4.652241112828439, + "grad_norm": 0.07190872237292077, + "learning_rate": 1.1734791317066763e-06, + "loss": 0.3416, + "step": 3010 + }, + { + "epoch": 4.653786707882535, + "grad_norm": 0.07112171217981218, + "learning_rate": 1.163122122549063e-06, + "loss": 0.3527, + "step": 3011 + }, + { + "epoch": 4.655332302936631, + "grad_norm": 0.06967890467673124, + "learning_rate": 1.1528103468298535e-06, + "loss": 0.3493, + "step": 3012 + }, + { + "epoch": 4.656877897990727, + "grad_norm": 0.07155519086256591, + "learning_rate": 1.142543816559205e-06, + "loss": 0.3483, + "step": 3013 + }, + { + "epoch": 4.658423493044822, + "grad_norm": 0.07391645581094995, + "learning_rate": 1.1323225436945885e-06, + "loss": 0.3535, + "step": 3014 + }, + { + "epoch": 4.659969088098918, + "grad_norm": 0.0720187041809313, + "learning_rate": 1.122146540140747e-06, + "loss": 0.3513, + "step": 3015 + }, + { + "epoch": 4.661514683153014, + "grad_norm": 0.07501960034559438, + "learning_rate": 1.112015817749712e-06, + "loss": 0.35, + "step": 3016 + }, + { + "epoch": 4.6630602782071096, + "grad_norm": 0.07153130718507449, + "learning_rate": 1.1019303883207598e-06, + "loss": 0.3513, + "step": 3017 + }, + { + "epoch": 4.6646058732612055, + "grad_norm": 0.07017952743086443, + "learning_rate": 1.0918902636004325e-06, + "loss": 0.3485, + "step": 3018 + }, + { + "epoch": 4.666151468315301, + "grad_norm": 0.07789273950609461, + "learning_rate": 1.0818954552824822e-06, + "loss": 0.3472, + "step": 3019 + }, + { + "epoch": 4.667697063369397, + "grad_norm": 0.07399218057264262, + "learning_rate": 1.0719459750079086e-06, + "loss": 0.35, + "step": 3020 + }, + { + "epoch": 4.669242658423493, + "grad_norm": 0.07127857715673765, + "learning_rate": 1.0620418343648998e-06, + "loss": 0.3464, + "step": 3021 + }, + { + "epoch": 4.670788253477589, + "grad_norm": 0.07483385161854272, + "learning_rate": 1.0521830448888414e-06, + "loss": 0.3456, + "step": 3022 + }, + { + "epoch": 4.672333848531685, + "grad_norm": 0.07204304647984132, + "learning_rate": 1.042369618062291e-06, + "loss": 0.3525, + "step": 3023 + }, + { + "epoch": 4.673879443585781, + "grad_norm": 0.07014879739671494, + "learning_rate": 1.03260156531499e-06, + "loss": 0.349, + "step": 3024 + }, + { + "epoch": 4.675425038639876, + "grad_norm": 0.07473680544412793, + "learning_rate": 1.022878898023807e-06, + "loss": 0.3618, + "step": 3025 + }, + { + "epoch": 4.676970633693972, + "grad_norm": 0.07421624584199558, + "learning_rate": 1.0132016275127766e-06, + "loss": 0.3485, + "step": 3026 + }, + { + "epoch": 4.678516228748068, + "grad_norm": 0.07477796206290063, + "learning_rate": 1.003569765053043e-06, + "loss": 0.3584, + "step": 3027 + }, + { + "epoch": 4.680061823802164, + "grad_norm": 0.07749692593469164, + "learning_rate": 9.939833218628637e-07, + "loss": 0.344, + "step": 3028 + }, + { + "epoch": 4.6816074188562595, + "grad_norm": 0.0718024898827332, + "learning_rate": 9.84442309107614e-07, + "loss": 0.3483, + "step": 3029 + }, + { + "epoch": 4.683153013910355, + "grad_norm": 0.0706931768229241, + "learning_rate": 9.74946737899729e-07, + "loss": 0.3463, + "step": 3030 + }, + { + "epoch": 4.684698608964451, + "grad_norm": 0.07347425880468403, + "learning_rate": 9.654966192987403e-07, + "loss": 0.3515, + "step": 3031 + }, + { + "epoch": 4.686244204018547, + "grad_norm": 0.07435454364586866, + "learning_rate": 9.560919643112165e-07, + "loss": 0.3462, + "step": 3032 + }, + { + "epoch": 4.687789799072643, + "grad_norm": 0.07490827768341446, + "learning_rate": 9.467327838908092e-07, + "loss": 0.3434, + "step": 3033 + }, + { + "epoch": 4.689335394126739, + "grad_norm": 0.0724556841352236, + "learning_rate": 9.374190889381762e-07, + "loss": 0.3571, + "step": 3034 + }, + { + "epoch": 4.690880989180835, + "grad_norm": 0.07406491541893734, + "learning_rate": 9.28150890301005e-07, + "loss": 0.3582, + "step": 3035 + }, + { + "epoch": 4.69242658423493, + "grad_norm": 0.07374468202194766, + "learning_rate": 9.189281987740073e-07, + "loss": 0.3525, + "step": 3036 + }, + { + "epoch": 4.693972179289027, + "grad_norm": 0.06877988053161671, + "learning_rate": 9.097510250988795e-07, + "loss": 0.3563, + "step": 3037 + }, + { + "epoch": 4.695517774343122, + "grad_norm": 0.07102803965671498, + "learning_rate": 9.006193799642982e-07, + "loss": 0.3479, + "step": 3038 + }, + { + "epoch": 4.697063369397218, + "grad_norm": 0.07113184558693575, + "learning_rate": 8.915332740059291e-07, + "loss": 0.351, + "step": 3039 + }, + { + "epoch": 4.698608964451314, + "grad_norm": 0.0718632619148234, + "learning_rate": 8.824927178063824e-07, + "loss": 0.3499, + "step": 3040 + }, + { + "epoch": 4.7001545595054095, + "grad_norm": 0.07329925111631844, + "learning_rate": 8.734977218952223e-07, + "loss": 0.359, + "step": 3041 + }, + { + "epoch": 4.701700154559505, + "grad_norm": 0.07547314176689787, + "learning_rate": 8.645482967489571e-07, + "loss": 0.3515, + "step": 3042 + }, + { + "epoch": 4.703245749613601, + "grad_norm": 0.07306475815997367, + "learning_rate": 8.556444527910135e-07, + "loss": 0.3501, + "step": 3043 + }, + { + "epoch": 4.704791344667697, + "grad_norm": 0.07526389425229102, + "learning_rate": 8.467862003917182e-07, + "loss": 0.3526, + "step": 3044 + }, + { + "epoch": 4.706336939721793, + "grad_norm": 0.07258969878008852, + "learning_rate": 8.379735498683117e-07, + "loss": 0.3429, + "step": 3045 + }, + { + "epoch": 4.707882534775889, + "grad_norm": 0.07342674835803169, + "learning_rate": 8.292065114849168e-07, + "loss": 0.3458, + "step": 3046 + }, + { + "epoch": 4.709428129829985, + "grad_norm": 0.07146671285867212, + "learning_rate": 8.204850954525256e-07, + "loss": 0.3502, + "step": 3047 + }, + { + "epoch": 4.710973724884081, + "grad_norm": 0.07026909475371605, + "learning_rate": 8.118093119290127e-07, + "loss": 0.3413, + "step": 3048 + }, + { + "epoch": 4.712519319938176, + "grad_norm": 0.07409553775894827, + "learning_rate": 8.031791710190817e-07, + "loss": 0.3557, + "step": 3049 + }, + { + "epoch": 4.714064914992272, + "grad_norm": 0.07232324119465391, + "learning_rate": 7.945946827742878e-07, + "loss": 0.3474, + "step": 3050 + }, + { + "epoch": 4.715610510046368, + "grad_norm": 0.07059867978942962, + "learning_rate": 7.86055857193011e-07, + "loss": 0.35, + "step": 3051 + }, + { + "epoch": 4.717156105100464, + "grad_norm": 0.0749773680999401, + "learning_rate": 7.775627042204559e-07, + "loss": 0.3506, + "step": 3052 + }, + { + "epoch": 4.7187017001545595, + "grad_norm": 0.07290342990514166, + "learning_rate": 7.691152337486118e-07, + "loss": 0.3443, + "step": 3053 + }, + { + "epoch": 4.720247295208655, + "grad_norm": 0.0752869756287177, + "learning_rate": 7.607134556162932e-07, + "loss": 0.3527, + "step": 3054 + }, + { + "epoch": 4.721792890262751, + "grad_norm": 0.07255599822428015, + "learning_rate": 7.523573796090678e-07, + "loss": 0.3488, + "step": 3055 + }, + { + "epoch": 4.723338485316847, + "grad_norm": 0.07165098224485363, + "learning_rate": 7.440470154592793e-07, + "loss": 0.3407, + "step": 3056 + }, + { + "epoch": 4.724884080370943, + "grad_norm": 0.0690313954012584, + "learning_rate": 7.357823728460478e-07, + "loss": 0.3533, + "step": 3057 + }, + { + "epoch": 4.726429675425039, + "grad_norm": 0.07346441068826698, + "learning_rate": 7.275634613952287e-07, + "loss": 0.3456, + "step": 3058 + }, + { + "epoch": 4.727975270479135, + "grad_norm": 0.06992603922784864, + "learning_rate": 7.193902906794092e-07, + "loss": 0.3544, + "step": 3059 + }, + { + "epoch": 4.72952086553323, + "grad_norm": 0.07167968338403391, + "learning_rate": 7.112628702179036e-07, + "loss": 0.348, + "step": 3060 + }, + { + "epoch": 4.731066460587326, + "grad_norm": 0.07244329403694023, + "learning_rate": 7.031812094767576e-07, + "loss": 0.348, + "step": 3061 + }, + { + "epoch": 4.732612055641422, + "grad_norm": 0.0695438720429314, + "learning_rate": 6.951453178686995e-07, + "loss": 0.3497, + "step": 3062 + }, + { + "epoch": 4.734157650695518, + "grad_norm": 0.06905910992436456, + "learning_rate": 6.871552047531537e-07, + "loss": 0.3374, + "step": 3063 + }, + { + "epoch": 4.7357032457496135, + "grad_norm": 0.07038312733282957, + "learning_rate": 6.792108794362451e-07, + "loss": 0.3539, + "step": 3064 + }, + { + "epoch": 4.7372488408037094, + "grad_norm": 0.07111918247888585, + "learning_rate": 6.713123511707498e-07, + "loss": 0.3443, + "step": 3065 + }, + { + "epoch": 4.738794435857805, + "grad_norm": 0.07029419548120323, + "learning_rate": 6.634596291561002e-07, + "loss": 0.3461, + "step": 3066 + }, + { + "epoch": 4.740340030911901, + "grad_norm": 0.07236006854199122, + "learning_rate": 6.556527225383979e-07, + "loss": 0.342, + "step": 3067 + }, + { + "epoch": 4.741885625965997, + "grad_norm": 0.07028641780741396, + "learning_rate": 6.478916404103697e-07, + "loss": 0.3466, + "step": 3068 + }, + { + "epoch": 4.743431221020093, + "grad_norm": 0.06913665604329292, + "learning_rate": 6.401763918113712e-07, + "loss": 0.3472, + "step": 3069 + }, + { + "epoch": 4.744976816074189, + "grad_norm": 0.07503151033597957, + "learning_rate": 6.325069857273791e-07, + "loss": 0.3533, + "step": 3070 + }, + { + "epoch": 4.746522411128284, + "grad_norm": 0.07241753183805473, + "learning_rate": 6.248834310909768e-07, + "loss": 0.3501, + "step": 3071 + }, + { + "epoch": 4.74806800618238, + "grad_norm": 0.07151901626987163, + "learning_rate": 6.173057367813418e-07, + "loss": 0.3555, + "step": 3072 + }, + { + "epoch": 4.749613601236476, + "grad_norm": 0.0724383131896621, + "learning_rate": 6.097739116242407e-07, + "loss": 0.3548, + "step": 3073 + }, + { + "epoch": 4.751159196290572, + "grad_norm": 0.07344169003089844, + "learning_rate": 6.022879643920121e-07, + "loss": 0.3525, + "step": 3074 + }, + { + "epoch": 4.752704791344668, + "grad_norm": 0.07373091575353383, + "learning_rate": 5.948479038035615e-07, + "loss": 0.3482, + "step": 3075 + }, + { + "epoch": 4.7542503863987635, + "grad_norm": 0.07454194242432469, + "learning_rate": 5.874537385243617e-07, + "loss": 0.3453, + "step": 3076 + }, + { + "epoch": 4.755795981452859, + "grad_norm": 0.06939072118384791, + "learning_rate": 5.801054771664128e-07, + "loss": 0.3447, + "step": 3077 + }, + { + "epoch": 4.757341576506955, + "grad_norm": 0.07059689129139902, + "learning_rate": 5.728031282882596e-07, + "loss": 0.344, + "step": 3078 + }, + { + "epoch": 4.758887171561051, + "grad_norm": 0.06795840612265545, + "learning_rate": 5.655467003949743e-07, + "loss": 0.3411, + "step": 3079 + }, + { + "epoch": 4.760432766615147, + "grad_norm": 0.06866144946942429, + "learning_rate": 5.583362019381389e-07, + "loss": 0.3488, + "step": 3080 + }, + { + "epoch": 4.761978361669243, + "grad_norm": 0.07448348792534418, + "learning_rate": 5.511716413158441e-07, + "loss": 0.3507, + "step": 3081 + }, + { + "epoch": 4.763523956723338, + "grad_norm": 0.07097827849903926, + "learning_rate": 5.440530268726863e-07, + "loss": 0.3545, + "step": 3082 + }, + { + "epoch": 4.765069551777434, + "grad_norm": 0.07223978699338694, + "learning_rate": 5.369803668997265e-07, + "loss": 0.3526, + "step": 3083 + }, + { + "epoch": 4.76661514683153, + "grad_norm": 0.07319239486591154, + "learning_rate": 5.29953669634522e-07, + "loss": 0.3395, + "step": 3084 + }, + { + "epoch": 4.768160741885626, + "grad_norm": 0.07189051429090473, + "learning_rate": 5.229729432610819e-07, + "loss": 0.3452, + "step": 3085 + }, + { + "epoch": 4.769706336939722, + "grad_norm": 0.06897687535342906, + "learning_rate": 5.160381959098981e-07, + "loss": 0.3446, + "step": 3086 + }, + { + "epoch": 4.771251931993818, + "grad_norm": 0.07503284062047047, + "learning_rate": 5.091494356578741e-07, + "loss": 0.3567, + "step": 3087 + }, + { + "epoch": 4.7727975270479135, + "grad_norm": 0.07270867686038726, + "learning_rate": 5.02306670528383e-07, + "loss": 0.3618, + "step": 3088 + }, + { + "epoch": 4.774343122102009, + "grad_norm": 0.0697258178900146, + "learning_rate": 4.955099084912097e-07, + "loss": 0.3509, + "step": 3089 + }, + { + "epoch": 4.775888717156105, + "grad_norm": 0.07336882901055865, + "learning_rate": 4.887591574625727e-07, + "loss": 0.3542, + "step": 3090 + }, + { + "epoch": 4.777434312210201, + "grad_norm": 0.06974428211111179, + "learning_rate": 4.820544253050851e-07, + "loss": 0.3478, + "step": 3091 + }, + { + "epoch": 4.778979907264297, + "grad_norm": 0.07163824194929422, + "learning_rate": 4.753957198277759e-07, + "loss": 0.3421, + "step": 3092 + }, + { + "epoch": 4.780525502318392, + "grad_norm": 0.07001588261468611, + "learning_rate": 4.687830487860634e-07, + "loss": 0.3476, + "step": 3093 + }, + { + "epoch": 4.782071097372488, + "grad_norm": 0.07257955809332253, + "learning_rate": 4.622164198817336e-07, + "loss": 0.3518, + "step": 3094 + }, + { + "epoch": 4.783616692426584, + "grad_norm": 0.07419073933373933, + "learning_rate": 4.556958407629797e-07, + "loss": 0.3516, + "step": 3095 + }, + { + "epoch": 4.78516228748068, + "grad_norm": 0.07323695296781084, + "learning_rate": 4.4922131902432664e-07, + "loss": 0.3482, + "step": 3096 + }, + { + "epoch": 4.786707882534776, + "grad_norm": 0.0703451573368961, + "learning_rate": 4.4279286220667574e-07, + "loss": 0.3475, + "step": 3097 + }, + { + "epoch": 4.788253477588872, + "grad_norm": 0.07283872288973346, + "learning_rate": 4.3641047779727777e-07, + "loss": 0.3593, + "step": 3098 + }, + { + "epoch": 4.789799072642968, + "grad_norm": 0.07067693341439762, + "learning_rate": 4.300741732297065e-07, + "loss": 0.345, + "step": 3099 + }, + { + "epoch": 4.7913446676970635, + "grad_norm": 0.06936899504891707, + "learning_rate": 4.2378395588388523e-07, + "loss": 0.3543, + "step": 3100 + }, + { + "epoch": 4.792890262751159, + "grad_norm": 0.06955688998075936, + "learning_rate": 4.175398330860425e-07, + "loss": 0.3473, + "step": 3101 + }, + { + "epoch": 4.794435857805255, + "grad_norm": 0.07243694849945517, + "learning_rate": 4.113418121087387e-07, + "loss": 0.3412, + "step": 3102 + }, + { + "epoch": 4.795981452859351, + "grad_norm": 0.07157169025779929, + "learning_rate": 4.0518990017081706e-07, + "loss": 0.3448, + "step": 3103 + }, + { + "epoch": 4.797527047913446, + "grad_norm": 0.07018241550207491, + "learning_rate": 3.9908410443743493e-07, + "loss": 0.3405, + "step": 3104 + }, + { + "epoch": 4.799072642967543, + "grad_norm": 0.07326576573687611, + "learning_rate": 3.9302443202003714e-07, + "loss": 0.3535, + "step": 3105 + }, + { + "epoch": 4.800618238021638, + "grad_norm": 0.0707235481534224, + "learning_rate": 3.8701088997633805e-07, + "loss": 0.3516, + "step": 3106 + }, + { + "epoch": 4.802163833075734, + "grad_norm": 0.07237622064497551, + "learning_rate": 3.810434853103351e-07, + "loss": 0.3479, + "step": 3107 + }, + { + "epoch": 4.80370942812983, + "grad_norm": 0.06873407666523473, + "learning_rate": 3.7512222497227747e-07, + "loss": 0.3555, + "step": 3108 + }, + { + "epoch": 4.805255023183926, + "grad_norm": 1.2980980480943682, + "learning_rate": 3.692471158586752e-07, + "loss": 0.3572, + "step": 3109 + }, + { + "epoch": 4.806800618238022, + "grad_norm": 0.07110026591967687, + "learning_rate": 3.6341816481229477e-07, + "loss": 0.3569, + "step": 3110 + }, + { + "epoch": 4.8083462132921175, + "grad_norm": 0.07272273956076963, + "learning_rate": 3.57635378622132e-07, + "loss": 0.3452, + "step": 3111 + }, + { + "epoch": 4.809891808346213, + "grad_norm": 0.07070547568026381, + "learning_rate": 3.5189876402341727e-07, + "loss": 0.34, + "step": 3112 + }, + { + "epoch": 4.811437403400309, + "grad_norm": 0.06861689944938604, + "learning_rate": 3.462083276976014e-07, + "loss": 0.3391, + "step": 3113 + }, + { + "epoch": 4.812982998454405, + "grad_norm": 0.07035516615838779, + "learning_rate": 3.4056407627235613e-07, + "loss": 0.3552, + "step": 3114 + }, + { + "epoch": 4.814528593508501, + "grad_norm": 0.07093647662836161, + "learning_rate": 3.3496601632155625e-07, + "loss": 0.3496, + "step": 3115 + }, + { + "epoch": 4.816074188562597, + "grad_norm": 0.07155958462422801, + "learning_rate": 3.2941415436528403e-07, + "loss": 0.3557, + "step": 3116 + }, + { + "epoch": 4.817619783616692, + "grad_norm": 0.07242485193251162, + "learning_rate": 3.239084968698114e-07, + "loss": 0.3461, + "step": 3117 + }, + { + "epoch": 4.819165378670788, + "grad_norm": 0.07226733900010342, + "learning_rate": 3.184490502475912e-07, + "loss": 0.3526, + "step": 3118 + }, + { + "epoch": 4.820710973724884, + "grad_norm": 0.07125796713408636, + "learning_rate": 3.1303582085726593e-07, + "loss": 0.3457, + "step": 3119 + }, + { + "epoch": 4.82225656877898, + "grad_norm": 0.0725996246570851, + "learning_rate": 3.076688150036367e-07, + "loss": 0.3488, + "step": 3120 + }, + { + "epoch": 4.823802163833076, + "grad_norm": 0.07052611899488413, + "learning_rate": 3.023480389376765e-07, + "loss": 0.346, + "step": 3121 + }, + { + "epoch": 4.825347758887172, + "grad_norm": 0.06981281270196775, + "learning_rate": 2.9707349885649937e-07, + "loss": 0.3412, + "step": 3122 + }, + { + "epoch": 4.8268933539412675, + "grad_norm": 0.069439337031512, + "learning_rate": 2.9184520090339565e-07, + "loss": 0.3567, + "step": 3123 + }, + { + "epoch": 4.828438948995363, + "grad_norm": 0.06951443984608201, + "learning_rate": 2.8666315116776975e-07, + "loss": 0.3479, + "step": 3124 + }, + { + "epoch": 4.829984544049459, + "grad_norm": 0.06735297642300035, + "learning_rate": 2.8152735568517606e-07, + "loss": 0.3474, + "step": 3125 + }, + { + "epoch": 4.831530139103555, + "grad_norm": 0.06971082920706823, + "learning_rate": 2.7643782043729197e-07, + "loss": 0.346, + "step": 3126 + }, + { + "epoch": 4.833075734157651, + "grad_norm": 0.08271390144614732, + "learning_rate": 2.713945513519134e-07, + "loss": 0.3502, + "step": 3127 + }, + { + "epoch": 4.834621329211746, + "grad_norm": 0.06943164183825769, + "learning_rate": 2.6639755430295957e-07, + "loss": 0.3395, + "step": 3128 + }, + { + "epoch": 4.836166924265842, + "grad_norm": 0.06967064529885418, + "learning_rate": 2.614468351104371e-07, + "loss": 0.3435, + "step": 3129 + }, + { + "epoch": 4.837712519319938, + "grad_norm": 0.06966768895389379, + "learning_rate": 2.565423995404759e-07, + "loss": 0.3436, + "step": 3130 + }, + { + "epoch": 4.839258114374034, + "grad_norm": 0.06997659712202021, + "learning_rate": 2.5168425330528434e-07, + "loss": 0.3374, + "step": 3131 + }, + { + "epoch": 4.84080370942813, + "grad_norm": 0.07170201260888213, + "learning_rate": 2.4687240206315855e-07, + "loss": 0.3523, + "step": 3132 + }, + { + "epoch": 4.842349304482226, + "grad_norm": 0.07350875287591517, + "learning_rate": 2.4210685141848213e-07, + "loss": 0.3526, + "step": 3133 + }, + { + "epoch": 4.843894899536322, + "grad_norm": 0.07151950805751307, + "learning_rate": 2.3738760692170404e-07, + "loss": 0.3509, + "step": 3134 + }, + { + "epoch": 4.8454404945904175, + "grad_norm": 0.06977292893693249, + "learning_rate": 2.327146740693431e-07, + "loss": 0.3533, + "step": 3135 + }, + { + "epoch": 4.846986089644513, + "grad_norm": 0.07230874587314665, + "learning_rate": 2.280880583039835e-07, + "loss": 0.3491, + "step": 3136 + }, + { + "epoch": 4.848531684698609, + "grad_norm": 0.07114464201845196, + "learning_rate": 2.2350776501425253e-07, + "loss": 0.345, + "step": 3137 + }, + { + "epoch": 4.850077279752705, + "grad_norm": 0.06958736284066182, + "learning_rate": 2.1897379953483844e-07, + "loss": 0.3485, + "step": 3138 + }, + { + "epoch": 4.8516228748068, + "grad_norm": 0.07376727963501929, + "learning_rate": 2.144861671464682e-07, + "loss": 0.3479, + "step": 3139 + }, + { + "epoch": 4.853168469860896, + "grad_norm": 0.07167646732802174, + "learning_rate": 2.1004487307589861e-07, + "loss": 0.3459, + "step": 3140 + }, + { + "epoch": 4.854714064914992, + "grad_norm": 0.0678011743605068, + "learning_rate": 2.056499224959163e-07, + "loss": 0.3486, + "step": 3141 + }, + { + "epoch": 4.856259659969088, + "grad_norm": 0.07083801332503054, + "learning_rate": 2.013013205253378e-07, + "loss": 0.3544, + "step": 3142 + }, + { + "epoch": 4.857805255023184, + "grad_norm": 0.07296463515942934, + "learning_rate": 1.9699907222899162e-07, + "loss": 0.345, + "step": 3143 + }, + { + "epoch": 4.85935085007728, + "grad_norm": 0.07337285719937855, + "learning_rate": 1.9274318261772286e-07, + "loss": 0.3465, + "step": 3144 + }, + { + "epoch": 4.860896445131376, + "grad_norm": 0.06945046410104723, + "learning_rate": 1.8853365664837975e-07, + "loss": 0.3545, + "step": 3145 + }, + { + "epoch": 4.8624420401854715, + "grad_norm": 0.07003201459282013, + "learning_rate": 1.8437049922381378e-07, + "loss": 0.3463, + "step": 3146 + }, + { + "epoch": 4.8639876352395675, + "grad_norm": 0.07152799831202193, + "learning_rate": 1.8025371519286182e-07, + "loss": 0.3509, + "step": 3147 + }, + { + "epoch": 4.865533230293663, + "grad_norm": 0.06784201123029486, + "learning_rate": 1.7618330935035955e-07, + "loss": 0.3495, + "step": 3148 + }, + { + "epoch": 4.867078825347759, + "grad_norm": 0.06876345540883695, + "learning_rate": 1.7215928643712353e-07, + "loss": 0.3522, + "step": 3149 + }, + { + "epoch": 4.868624420401854, + "grad_norm": 0.06917992969160941, + "learning_rate": 1.6818165113993812e-07, + "loss": 0.344, + "step": 3150 + }, + { + "epoch": 4.87017001545595, + "grad_norm": 0.07128084847857968, + "learning_rate": 1.6425040809157745e-07, + "loss": 0.3445, + "step": 3151 + }, + { + "epoch": 4.871715610510046, + "grad_norm": 0.06912986125990882, + "learning_rate": 1.6036556187077e-07, + "loss": 0.3585, + "step": 3152 + }, + { + "epoch": 4.873261205564142, + "grad_norm": 0.0697507160019747, + "learning_rate": 1.5652711700220756e-07, + "loss": 0.3442, + "step": 3153 + }, + { + "epoch": 4.874806800618238, + "grad_norm": 0.07002166993275444, + "learning_rate": 1.5273507795654063e-07, + "loss": 0.3582, + "step": 3154 + }, + { + "epoch": 4.876352395672334, + "grad_norm": 0.07164887990027259, + "learning_rate": 1.489894491503696e-07, + "loss": 0.3534, + "step": 3155 + }, + { + "epoch": 4.87789799072643, + "grad_norm": 0.07023687429723947, + "learning_rate": 1.4529023494623596e-07, + "loss": 0.3444, + "step": 3156 + }, + { + "epoch": 4.879443585780526, + "grad_norm": 0.07036932592856734, + "learning_rate": 1.4163743965263544e-07, + "loss": 0.3313, + "step": 3157 + }, + { + "epoch": 4.8809891808346215, + "grad_norm": 0.07116839284433644, + "learning_rate": 1.380310675239871e-07, + "loss": 0.3461, + "step": 3158 + }, + { + "epoch": 4.882534775888717, + "grad_norm": 0.06867559877545402, + "learning_rate": 1.3447112276063768e-07, + "loss": 0.3429, + "step": 3159 + }, + { + "epoch": 4.884080370942813, + "grad_norm": 0.07080825908539166, + "learning_rate": 1.3095760950887493e-07, + "loss": 0.3601, + "step": 3160 + }, + { + "epoch": 4.885625965996908, + "grad_norm": 0.0672162448443747, + "learning_rate": 1.2749053186089655e-07, + "loss": 0.3477, + "step": 3161 + }, + { + "epoch": 4.887171561051004, + "grad_norm": 0.06969539242372545, + "learning_rate": 1.2406989385481904e-07, + "loss": 0.3523, + "step": 3162 + }, + { + "epoch": 4.8887171561051, + "grad_norm": 0.07364444652155135, + "learning_rate": 1.2069569947466885e-07, + "loss": 0.3497, + "step": 3163 + }, + { + "epoch": 4.890262751159196, + "grad_norm": 0.06959759375214015, + "learning_rate": 1.1736795265038237e-07, + "loss": 0.3504, + "step": 3164 + }, + { + "epoch": 4.891808346213292, + "grad_norm": 0.07028677984651312, + "learning_rate": 1.14086657257797e-07, + "loss": 0.3473, + "step": 3165 + }, + { + "epoch": 4.893353941267388, + "grad_norm": 0.0720746843915264, + "learning_rate": 1.1085181711865123e-07, + "loss": 0.3506, + "step": 3166 + }, + { + "epoch": 4.894899536321484, + "grad_norm": 0.06850477026674287, + "learning_rate": 1.0766343600057128e-07, + "loss": 0.3497, + "step": 3167 + }, + { + "epoch": 4.89644513137558, + "grad_norm": 0.07195527681480882, + "learning_rate": 1.0452151761707552e-07, + "loss": 0.3477, + "step": 3168 + }, + { + "epoch": 4.897990726429676, + "grad_norm": 0.0727530453591956, + "learning_rate": 1.0142606562756119e-07, + "loss": 0.3476, + "step": 3169 + }, + { + "epoch": 4.8995363214837715, + "grad_norm": 0.0693586325825528, + "learning_rate": 9.837708363732213e-08, + "loss": 0.3578, + "step": 3170 + }, + { + "epoch": 4.901081916537867, + "grad_norm": 0.07160392575837254, + "learning_rate": 9.537457519750881e-08, + "loss": 0.3446, + "step": 3171 + }, + { + "epoch": 4.902627511591962, + "grad_norm": 0.07199585832686192, + "learning_rate": 9.241854380515946e-08, + "loss": 0.3446, + "step": 3172 + }, + { + "epoch": 4.904173106646059, + "grad_norm": 0.07078789282421856, + "learning_rate": 8.95089929031645e-08, + "loss": 0.35, + "step": 3173 + }, + { + "epoch": 4.905718701700154, + "grad_norm": 0.06870033302291263, + "learning_rate": 8.664592588029764e-08, + "loss": 0.3422, + "step": 3174 + }, + { + "epoch": 4.90726429675425, + "grad_norm": 0.07181153777857369, + "learning_rate": 8.382934607117144e-08, + "loss": 0.3469, + "step": 3175 + }, + { + "epoch": 4.908809891808346, + "grad_norm": 0.06931529355649897, + "learning_rate": 8.105925675627735e-08, + "loss": 0.3449, + "step": 3176 + }, + { + "epoch": 4.910355486862442, + "grad_norm": 0.06889547051762213, + "learning_rate": 7.83356611619368e-08, + "loss": 0.355, + "step": 3177 + }, + { + "epoch": 4.911901081916538, + "grad_norm": 0.06997885222535463, + "learning_rate": 7.565856246033676e-08, + "loss": 0.349, + "step": 3178 + }, + { + "epoch": 4.913446676970634, + "grad_norm": 0.07031322677659328, + "learning_rate": 7.302796376950305e-08, + "loss": 0.3499, + "step": 3179 + }, + { + "epoch": 4.91499227202473, + "grad_norm": 0.07026799992091529, + "learning_rate": 7.044386815330039e-08, + "loss": 0.3482, + "step": 3180 + }, + { + "epoch": 4.916537867078826, + "grad_norm": 0.07275397852011724, + "learning_rate": 6.79062786214324e-08, + "loss": 0.3436, + "step": 3181 + }, + { + "epoch": 4.9180834621329215, + "grad_norm": 0.07110659178980795, + "learning_rate": 6.541519812943709e-08, + "loss": 0.3517, + "step": 3182 + }, + { + "epoch": 4.919629057187017, + "grad_norm": 0.06934394961030939, + "learning_rate": 6.297062957869138e-08, + "loss": 0.3549, + "step": 3183 + }, + { + "epoch": 4.921174652241113, + "grad_norm": 0.06846822812054754, + "learning_rate": 6.057257581638443e-08, + "loss": 0.335, + "step": 3184 + }, + { + "epoch": 4.922720247295208, + "grad_norm": 0.06636240898387039, + "learning_rate": 5.8221039635539823e-08, + "loss": 0.3482, + "step": 3185 + }, + { + "epoch": 4.924265842349304, + "grad_norm": 0.06912364928780351, + "learning_rate": 5.591602377499783e-08, + "loss": 0.3484, + "step": 3186 + }, + { + "epoch": 4.9258114374034, + "grad_norm": 0.07010649629366013, + "learning_rate": 5.365753091941983e-08, + "loss": 0.3538, + "step": 3187 + }, + { + "epoch": 4.927357032457496, + "grad_norm": 0.07228072658478765, + "learning_rate": 5.1445563699283884e-08, + "loss": 0.3424, + "step": 3188 + }, + { + "epoch": 4.928902627511592, + "grad_norm": 0.07074974699598284, + "learning_rate": 4.928012469086696e-08, + "loss": 0.3546, + "step": 3189 + }, + { + "epoch": 4.930448222565688, + "grad_norm": 0.06738689125844004, + "learning_rate": 4.716121641626714e-08, + "loss": 0.3419, + "step": 3190 + }, + { + "epoch": 4.931993817619784, + "grad_norm": 0.06934501885583151, + "learning_rate": 4.5088841343385866e-08, + "loss": 0.3577, + "step": 3191 + }, + { + "epoch": 4.93353941267388, + "grad_norm": 0.074852165035623, + "learning_rate": 4.3063001885923496e-08, + "loss": 0.35, + "step": 3192 + }, + { + "epoch": 4.9350850077279755, + "grad_norm": 0.06897115937842231, + "learning_rate": 4.108370040337928e-08, + "loss": 0.3401, + "step": 3193 + }, + { + "epoch": 4.936630602782071, + "grad_norm": 0.07089437912697523, + "learning_rate": 3.915093920105584e-08, + "loss": 0.3567, + "step": 3194 + }, + { + "epoch": 4.938176197836167, + "grad_norm": 0.0711008694012866, + "learning_rate": 3.7264720530045815e-08, + "loss": 0.35, + "step": 3195 + }, + { + "epoch": 4.939721792890262, + "grad_norm": 0.06897510467703488, + "learning_rate": 3.542504658723189e-08, + "loss": 0.3466, + "step": 3196 + }, + { + "epoch": 4.941267387944358, + "grad_norm": 0.07161874661079988, + "learning_rate": 3.363191951528677e-08, + "loss": 0.343, + "step": 3197 + }, + { + "epoch": 4.942812982998454, + "grad_norm": 0.06800251521260453, + "learning_rate": 3.188534140267763e-08, + "loss": 0.3518, + "step": 3198 + }, + { + "epoch": 4.94435857805255, + "grad_norm": 0.06857550082527371, + "learning_rate": 3.018531428364835e-08, + "loss": 0.3541, + "step": 3199 + }, + { + "epoch": 4.945904173106646, + "grad_norm": 0.06842598322376241, + "learning_rate": 2.8531840138219523e-08, + "loss": 0.3483, + "step": 3200 + }, + { + "epoch": 4.947449768160742, + "grad_norm": 0.0699432865457825, + "learning_rate": 2.692492089220622e-08, + "loss": 0.3455, + "step": 3201 + }, + { + "epoch": 4.948995363214838, + "grad_norm": 0.07118104377541498, + "learning_rate": 2.5364558417186878e-08, + "loss": 0.3459, + "step": 3202 + }, + { + "epoch": 4.950540958268934, + "grad_norm": 0.07014681702978057, + "learning_rate": 2.3850754530521104e-08, + "loss": 0.3394, + "step": 3203 + }, + { + "epoch": 4.95208655332303, + "grad_norm": 0.07187162690120859, + "learning_rate": 2.2383510995349634e-08, + "loss": 0.3541, + "step": 3204 + }, + { + "epoch": 4.9536321483771255, + "grad_norm": 0.0686991664958423, + "learning_rate": 2.0962829520563277e-08, + "loss": 0.3551, + "step": 3205 + }, + { + "epoch": 4.955177743431221, + "grad_norm": 0.07159243273125415, + "learning_rate": 1.958871176083843e-08, + "loss": 0.3533, + "step": 3206 + }, + { + "epoch": 4.956723338485316, + "grad_norm": 0.07336825106437933, + "learning_rate": 1.8261159316619316e-08, + "loss": 0.3494, + "step": 3207 + }, + { + "epoch": 4.958268933539412, + "grad_norm": 0.06911403449022502, + "learning_rate": 1.6980173734109095e-08, + "loss": 0.3404, + "step": 3208 + }, + { + "epoch": 4.959814528593508, + "grad_norm": 0.07107073653552722, + "learning_rate": 1.5745756505269883e-08, + "loss": 0.341, + "step": 3209 + }, + { + "epoch": 4.961360123647604, + "grad_norm": 0.06968927460508988, + "learning_rate": 1.4557909067840493e-08, + "loss": 0.3577, + "step": 3210 + }, + { + "epoch": 4.9629057187017, + "grad_norm": 0.07308536629162159, + "learning_rate": 1.3416632805305363e-08, + "loss": 0.3539, + "step": 3211 + }, + { + "epoch": 4.964451313755796, + "grad_norm": 0.07152542735395838, + "learning_rate": 1.2321929046907877e-08, + "loss": 0.3455, + "step": 3212 + }, + { + "epoch": 4.965996908809892, + "grad_norm": 0.07195876470005634, + "learning_rate": 1.1273799067663683e-08, + "loss": 0.3532, + "step": 3213 + }, + { + "epoch": 4.967542503863988, + "grad_norm": 0.06941177989015325, + "learning_rate": 1.0272244088329608e-08, + "loss": 0.3445, + "step": 3214 + }, + { + "epoch": 4.969088098918084, + "grad_norm": 0.07016913867065504, + "learning_rate": 9.317265275412546e-09, + "loss": 0.353, + "step": 3215 + }, + { + "epoch": 4.97063369397218, + "grad_norm": 0.07421619000696215, + "learning_rate": 8.408863741191653e-09, + "loss": 0.3461, + "step": 3216 + }, + { + "epoch": 4.9721792890262755, + "grad_norm": 0.06969975222182093, + "learning_rate": 7.547040543673945e-09, + "loss": 0.3423, + "step": 3217 + }, + { + "epoch": 4.9737248840803705, + "grad_norm": 0.06875874300514737, + "learning_rate": 6.731796686634262e-09, + "loss": 0.3467, + "step": 3218 + }, + { + "epoch": 4.975270479134466, + "grad_norm": 0.06726335077458415, + "learning_rate": 5.963133119588627e-09, + "loss": 0.3497, + "step": 3219 + }, + { + "epoch": 4.976816074188562, + "grad_norm": 0.06995506586532436, + "learning_rate": 5.241050737803122e-09, + "loss": 0.3489, + "step": 3220 + }, + { + "epoch": 4.978361669242658, + "grad_norm": 0.06932164730463274, + "learning_rate": 4.5655503822850114e-09, + "loss": 0.3453, + "step": 3221 + }, + { + "epoch": 4.979907264296754, + "grad_norm": 0.06964687506918937, + "learning_rate": 3.936632839796062e-09, + "loss": 0.349, + "step": 3222 + }, + { + "epoch": 4.98145285935085, + "grad_norm": 0.06731212312953543, + "learning_rate": 3.35429884283478e-09, + "loss": 0.3504, + "step": 3223 + }, + { + "epoch": 4.982998454404946, + "grad_norm": 0.066728749601903, + "learning_rate": 2.8185490696497342e-09, + "loss": 0.3427, + "step": 3224 + }, + { + "epoch": 4.984544049459042, + "grad_norm": 0.06773974916352922, + "learning_rate": 2.3293841442351138e-09, + "loss": 0.3514, + "step": 3225 + }, + { + "epoch": 4.986089644513138, + "grad_norm": 0.06744823029918116, + "learning_rate": 1.8868046363174073e-09, + "loss": 0.3477, + "step": 3226 + }, + { + "epoch": 4.987635239567234, + "grad_norm": 0.07051766584886511, + "learning_rate": 1.4908110613687243e-09, + "loss": 0.3605, + "step": 3227 + }, + { + "epoch": 4.9891808346213296, + "grad_norm": 0.06998152135885344, + "learning_rate": 1.141403880606795e-09, + "loss": 0.3528, + "step": 3228 + }, + { + "epoch": 4.990726429675425, + "grad_norm": 0.06874012390063793, + "learning_rate": 8.385835009860899e-10, + "loss": 0.3448, + "step": 3229 + }, + { + "epoch": 4.992272024729521, + "grad_norm": 0.0684893242405293, + "learning_rate": 5.823502752067001e-10, + "loss": 0.3579, + "step": 3230 + }, + { + "epoch": 4.993817619783616, + "grad_norm": 0.0666169378166932, + "learning_rate": 3.7270450170101556e-10, + "loss": 0.3434, + "step": 3231 + }, + { + "epoch": 4.995363214837712, + "grad_norm": 0.0703101634456694, + "learning_rate": 2.0964642464260666e-10, + "loss": 0.3524, + "step": 3232 + }, + { + "epoch": 4.996908809891808, + "grad_norm": 0.06758631576854585, + "learning_rate": 9.317623394622388e-11, + "loss": 0.3586, + "step": 3233 + }, + { + "epoch": 4.998454404945904, + "grad_norm": 0.0689968575427259, + "learning_rate": 2.3294065267798203e-11, + "loss": 0.3527, + "step": 3234 + }, + { + "epoch": 5.0, + "grad_norm": 0.08107722730902595, + "learning_rate": 0.0, + "loss": 0.3426, + "step": 3235 + }, + { + "epoch": 5.0, + "step": 3235, + "total_flos": 5.427082020519936e+16, + "train_loss": 0.6229784268500079, + "train_runtime": 70315.6352, + "train_samples_per_second": 23.525, + "train_steps_per_second": 0.046 + } + ], + "logging_steps": 1, + "max_steps": 3235, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 5.427082020519936e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}