{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9943851768669285, "eval_steps": 500, "global_step": 333, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008983717012914094, "grad_norm": 5.839086702019255, "learning_rate": 4.705882352941177e-06, "loss": 0.851, "step": 1 }, { "epoch": 0.017967434025828188, "grad_norm": 5.796604506548498, "learning_rate": 9.411764705882354e-06, "loss": 0.8589, "step": 2 }, { "epoch": 0.02695115103874228, "grad_norm": 4.339124227516403, "learning_rate": 1.4117647058823532e-05, "loss": 0.8175, "step": 3 }, { "epoch": 0.035934868051656375, "grad_norm": 2.117331281643237, "learning_rate": 1.8823529411764708e-05, "loss": 0.7666, "step": 4 }, { "epoch": 0.044918585064570464, "grad_norm": 6.08947531782942, "learning_rate": 2.3529411764705884e-05, "loss": 0.7997, "step": 5 }, { "epoch": 0.05390230207748456, "grad_norm": 11.437654369979397, "learning_rate": 2.8235294117647063e-05, "loss": 0.8658, "step": 6 }, { "epoch": 0.06288601909039865, "grad_norm": 6.230659224656504, "learning_rate": 3.294117647058824e-05, "loss": 0.7742, "step": 7 }, { "epoch": 0.07186973610331275, "grad_norm": 4.258899179934184, "learning_rate": 3.7647058823529415e-05, "loss": 0.7478, "step": 8 }, { "epoch": 0.08085345311622684, "grad_norm": 3.163611777765137, "learning_rate": 4.235294117647059e-05, "loss": 0.7167, "step": 9 }, { "epoch": 0.08983717012914093, "grad_norm": 2.4646084609310077, "learning_rate": 4.705882352941177e-05, "loss": 0.6839, "step": 10 }, { "epoch": 0.09882088714205503, "grad_norm": 2.1835077872968514, "learning_rate": 5.176470588235295e-05, "loss": 0.6594, "step": 11 }, { "epoch": 0.10780460415496912, "grad_norm": 1.956338905834903, "learning_rate": 5.6470588235294126e-05, "loss": 0.6447, "step": 12 }, { "epoch": 0.11678832116788321, "grad_norm": 1.2530266966782506, "learning_rate": 6.117647058823529e-05, "loss": 0.6225, "step": 13 }, { "epoch": 0.1257720381807973, "grad_norm": 2.746116145636045, "learning_rate": 6.588235294117648e-05, "loss": 0.6409, "step": 14 }, { "epoch": 0.13475575519371139, "grad_norm": 1.836377286540359, "learning_rate": 7.058823529411765e-05, "loss": 0.6265, "step": 15 }, { "epoch": 0.1437394722066255, "grad_norm": 2.81566991788529, "learning_rate": 7.529411764705883e-05, "loss": 0.6182, "step": 16 }, { "epoch": 0.1527231892195396, "grad_norm": 1.9484267685166876, "learning_rate": 8e-05, "loss": 0.6088, "step": 17 }, { "epoch": 0.16170690623245368, "grad_norm": 1.4245865160208993, "learning_rate": 8.470588235294118e-05, "loss": 0.5934, "step": 18 }, { "epoch": 0.17069062324536777, "grad_norm": 2.326255459435556, "learning_rate": 8.941176470588236e-05, "loss": 0.6025, "step": 19 }, { "epoch": 0.17967434025828186, "grad_norm": 1.5418881568160665, "learning_rate": 9.411764705882353e-05, "loss": 0.5952, "step": 20 }, { "epoch": 0.18865805727119594, "grad_norm": 1.8971846222744186, "learning_rate": 9.882352941176472e-05, "loss": 0.5812, "step": 21 }, { "epoch": 0.19764177428411006, "grad_norm": 1.200246618699023, "learning_rate": 0.0001035294117647059, "loss": 0.5796, "step": 22 }, { "epoch": 0.20662549129702415, "grad_norm": 1.541626574943513, "learning_rate": 0.00010823529411764708, "loss": 0.5735, "step": 23 }, { "epoch": 0.21560920830993824, "grad_norm": 2.8830857335375746, "learning_rate": 0.00011294117647058825, "loss": 0.5932, "step": 24 }, { "epoch": 0.22459292532285233, "grad_norm": 1.2631947275776292, "learning_rate": 0.00011764705882352943, "loss": 0.5768, "step": 25 }, { "epoch": 0.23357664233576642, "grad_norm": 3.7808866136207353, "learning_rate": 0.00012235294117647058, "loss": 0.6315, "step": 26 }, { "epoch": 0.2425603593486805, "grad_norm": 2.3059028840452847, "learning_rate": 0.00012705882352941178, "loss": 0.6138, "step": 27 }, { "epoch": 0.2515440763615946, "grad_norm": 3.060917512545427, "learning_rate": 0.00013176470588235296, "loss": 0.5971, "step": 28 }, { "epoch": 0.2605277933745087, "grad_norm": 2.2144909817302425, "learning_rate": 0.00013647058823529413, "loss": 0.6019, "step": 29 }, { "epoch": 0.26951151038742277, "grad_norm": 10.523000576466588, "learning_rate": 0.0001411764705882353, "loss": 0.6902, "step": 30 }, { "epoch": 0.2784952274003369, "grad_norm": 3733.897541065198, "learning_rate": 0.00014588235294117648, "loss": 2.1691, "step": 31 }, { "epoch": 0.287478944413251, "grad_norm": 152.85200430325233, "learning_rate": 0.00015058823529411766, "loss": 2.3024, "step": 32 }, { "epoch": 0.29646266142616506, "grad_norm": 5.795238259929523, "learning_rate": 0.00015529411764705884, "loss": 0.9312, "step": 33 }, { "epoch": 0.3054463784390792, "grad_norm": 2.5238561333157286, "learning_rate": 0.00016, "loss": 0.7014, "step": 34 }, { "epoch": 0.31443009545199324, "grad_norm": 3.227790922816581, "learning_rate": 0.0001599955841596387, "loss": 0.6914, "step": 35 }, { "epoch": 0.32341381246490736, "grad_norm": 2.590941539337386, "learning_rate": 0.00015998233712604597, "loss": 0.716, "step": 36 }, { "epoch": 0.3323975294778215, "grad_norm": 9.464462354232134, "learning_rate": 0.00015996026036164143, "loss": 1.163, "step": 37 }, { "epoch": 0.34138124649073553, "grad_norm": 9.4056400087482, "learning_rate": 0.00015992935630361176, "loss": 0.9691, "step": 38 }, { "epoch": 0.35036496350364965, "grad_norm": 3.229822785794876, "learning_rate": 0.00015988962836364164, "loss": 0.8032, "step": 39 }, { "epoch": 0.3593486805165637, "grad_norm": 2.4264007116828594, "learning_rate": 0.00015984108092753708, "loss": 0.7292, "step": 40 }, { "epoch": 0.36833239752947783, "grad_norm": 5.595688226616298, "learning_rate": 0.0001597837193547413, "loss": 0.681, "step": 41 }, { "epoch": 0.3773161145423919, "grad_norm": 1.6476213464085256, "learning_rate": 0.00015971754997774298, "loss": 0.676, "step": 42 }, { "epoch": 0.386299831555306, "grad_norm": 11.92511194308152, "learning_rate": 0.0001596425801013773, "loss": 0.6598, "step": 43 }, { "epoch": 0.3952835485682201, "grad_norm": 3.9213311829331734, "learning_rate": 0.00015955881800201933, "loss": 0.7401, "step": 44 }, { "epoch": 0.4042672655811342, "grad_norm": 2.0354638536769767, "learning_rate": 0.00015946627292667064, "loss": 0.6612, "step": 45 }, { "epoch": 0.4132509825940483, "grad_norm": 1.886963956827035, "learning_rate": 0.00015936495509193816, "loss": 0.6393, "step": 46 }, { "epoch": 0.42223469960696236, "grad_norm": 1.6161025496299593, "learning_rate": 0.00015925487568290645, "loss": 0.6135, "step": 47 }, { "epoch": 0.4312184166198765, "grad_norm": 1.3094812966970897, "learning_rate": 0.00015913604685190304, "loss": 0.5968, "step": 48 }, { "epoch": 0.4402021336327906, "grad_norm": 1.0856830435724436, "learning_rate": 0.00015900848171715654, "loss": 0.5947, "step": 49 }, { "epoch": 0.44918585064570465, "grad_norm": 3.133660662863731, "learning_rate": 0.00015887219436134874, "loss": 0.5986, "step": 50 }, { "epoch": 0.45816956765861877, "grad_norm": 1.629031015427086, "learning_rate": 0.00015872719983005986, "loss": 0.6136, "step": 51 }, { "epoch": 0.46715328467153283, "grad_norm": 1.2979437140842698, "learning_rate": 0.0001585735141301074, "loss": 0.6008, "step": 52 }, { "epoch": 0.47613700168444695, "grad_norm": 2.022435241771787, "learning_rate": 0.00015841115422777935, "loss": 0.599, "step": 53 }, { "epoch": 0.485120718697361, "grad_norm": 1.0176598957138148, "learning_rate": 0.0001582401380469609, "loss": 0.5911, "step": 54 }, { "epoch": 0.4941044357102751, "grad_norm": 1.340446281009079, "learning_rate": 0.00015806048446715594, "loss": 0.6018, "step": 55 }, { "epoch": 0.5030881527231892, "grad_norm": 1.1781671057543985, "learning_rate": 0.00015787221332140268, "loss": 0.5775, "step": 56 }, { "epoch": 0.5120718697361033, "grad_norm": 0.8746372899225068, "learning_rate": 0.00015767534539408418, "loss": 0.5764, "step": 57 }, { "epoch": 0.5210555867490174, "grad_norm": 0.8619016762831145, "learning_rate": 0.00015746990241863394, "loss": 0.5621, "step": 58 }, { "epoch": 0.5300393037619315, "grad_norm": 0.8908249631898915, "learning_rate": 0.0001572559070751366, "loss": 0.5626, "step": 59 }, { "epoch": 0.5390230207748455, "grad_norm": 0.9941730669724362, "learning_rate": 0.00015703338298782396, "loss": 0.5587, "step": 60 }, { "epoch": 0.5480067377877597, "grad_norm": 1.075188528305704, "learning_rate": 0.0001568023547224672, "loss": 0.5524, "step": 61 }, { "epoch": 0.5569904548006738, "grad_norm": 0.8331471695655838, "learning_rate": 0.00015656284778366479, "loss": 0.5383, "step": 62 }, { "epoch": 0.5659741718135879, "grad_norm": 0.7654032131505463, "learning_rate": 0.00015631488861202688, "loss": 0.538, "step": 63 }, { "epoch": 0.574957888826502, "grad_norm": 0.8860510131733896, "learning_rate": 0.00015605850458125646, "loss": 0.531, "step": 64 }, { "epoch": 0.583941605839416, "grad_norm": 0.6831724579134854, "learning_rate": 0.0001557937239951273, "loss": 0.5314, "step": 65 }, { "epoch": 0.5929253228523301, "grad_norm": 0.6778833496972871, "learning_rate": 0.00015552057608435936, "loss": 0.5291, "step": 66 }, { "epoch": 0.6019090398652442, "grad_norm": 0.7986870903265276, "learning_rate": 0.00015523909100339186, "loss": 0.531, "step": 67 }, { "epoch": 0.6108927568781584, "grad_norm": 0.7775755149998839, "learning_rate": 0.0001549492998270544, "loss": 0.5297, "step": 68 }, { "epoch": 0.6198764738910725, "grad_norm": 0.4942435514859441, "learning_rate": 0.00015465123454713622, "loss": 0.5166, "step": 69 }, { "epoch": 0.6288601909039865, "grad_norm": 0.5343768774089758, "learning_rate": 0.0001543449280688547, "loss": 0.5149, "step": 70 }, { "epoch": 0.6378439079169006, "grad_norm": 0.5061205195595769, "learning_rate": 0.00015403041420722257, "loss": 0.515, "step": 71 }, { "epoch": 0.6468276249298147, "grad_norm": 0.46967339805307057, "learning_rate": 0.00015370772768331497, "loss": 0.5026, "step": 72 }, { "epoch": 0.6558113419427288, "grad_norm": 0.4141303041477235, "learning_rate": 0.00015337690412043625, "loss": 0.5137, "step": 73 }, { "epoch": 0.664795058955643, "grad_norm": 0.45219045262379115, "learning_rate": 0.0001530379800401875, "loss": 0.5059, "step": 74 }, { "epoch": 0.673778775968557, "grad_norm": 0.3940676548155913, "learning_rate": 0.0001526909928584345, "loss": 0.5036, "step": 75 }, { "epoch": 0.6827624929814711, "grad_norm": 0.5141391546743134, "learning_rate": 0.00015233598088117733, "loss": 0.5163, "step": 76 }, { "epoch": 0.6917462099943852, "grad_norm": 0.6219020292862112, "learning_rate": 0.00015197298330032146, "loss": 0.502, "step": 77 }, { "epoch": 0.7007299270072993, "grad_norm": 0.5118275087823705, "learning_rate": 0.000151602040189351, "loss": 0.5015, "step": 78 }, { "epoch": 0.7097136440202133, "grad_norm": 0.6008412729706559, "learning_rate": 0.00015122319249890513, "loss": 0.4927, "step": 79 }, { "epoch": 0.7186973610331274, "grad_norm": 0.9550462875054769, "learning_rate": 0.00015083648205225679, "loss": 0.5102, "step": 80 }, { "epoch": 0.7276810780460415, "grad_norm": 1.4871365184883178, "learning_rate": 0.000150441951540696, "loss": 0.5042, "step": 81 }, { "epoch": 0.7366647950589557, "grad_norm": 0.4668622982117298, "learning_rate": 0.00015003964451881666, "loss": 0.481, "step": 82 }, { "epoch": 0.7456485120718698, "grad_norm": 0.9148709606250665, "learning_rate": 0.0001496296053997084, "loss": 0.5007, "step": 83 }, { "epoch": 0.7546322290847838, "grad_norm": 1.279060221546284, "learning_rate": 0.00014921187945005348, "loss": 0.4993, "step": 84 }, { "epoch": 0.7636159460976979, "grad_norm": 0.616746320687276, "learning_rate": 0.00014878651278512966, "loss": 0.4903, "step": 85 }, { "epoch": 0.772599663110612, "grad_norm": 0.9313866786660758, "learning_rate": 0.0001483535523637191, "loss": 0.4951, "step": 86 }, { "epoch": 0.7815833801235261, "grad_norm": 0.6267949258981602, "learning_rate": 0.0001479130459829244, "loss": 0.4849, "step": 87 }, { "epoch": 0.7905670971364402, "grad_norm": 0.7584425730651909, "learning_rate": 0.00014746504227289192, "loss": 0.4986, "step": 88 }, { "epoch": 0.7995508141493542, "grad_norm": 0.5681276511497075, "learning_rate": 0.00014700959069144332, "loss": 0.4785, "step": 89 }, { "epoch": 0.8085345311622684, "grad_norm": 0.5906735363635705, "learning_rate": 0.0001465467415186155, "loss": 0.4907, "step": 90 }, { "epoch": 0.8175182481751825, "grad_norm": 0.7053181757428566, "learning_rate": 0.0001460765458511099, "loss": 0.4912, "step": 91 }, { "epoch": 0.8265019651880966, "grad_norm": 0.6751343616215291, "learning_rate": 0.00014559905559665169, "loss": 0.4844, "step": 92 }, { "epoch": 0.8354856822010107, "grad_norm": 0.6487198602578013, "learning_rate": 0.00014511432346825932, "loss": 0.4801, "step": 93 }, { "epoch": 0.8444693992139247, "grad_norm": 0.323668455749265, "learning_rate": 0.0001446224029784252, "loss": 0.4793, "step": 94 }, { "epoch": 0.8534531162268388, "grad_norm": 0.5861018094486455, "learning_rate": 0.0001441233484332082, "loss": 0.4822, "step": 95 }, { "epoch": 0.862436833239753, "grad_norm": 0.4370056395937098, "learning_rate": 0.00014361721492623834, "loss": 0.4878, "step": 96 }, { "epoch": 0.8714205502526671, "grad_norm": 0.5421346987411235, "learning_rate": 0.00014310405833263486, "loss": 0.4872, "step": 97 }, { "epoch": 0.8804042672655812, "grad_norm": 0.5238036281016801, "learning_rate": 0.00014258393530283773, "loss": 0.47, "step": 98 }, { "epoch": 0.8893879842784952, "grad_norm": 0.557528609344473, "learning_rate": 0.0001420569032563536, "loss": 0.4859, "step": 99 }, { "epoch": 0.8983717012914093, "grad_norm": 0.7040485273548837, "learning_rate": 0.00014152302037541707, "loss": 0.4747, "step": 100 }, { "epoch": 0.9073554183043234, "grad_norm": 0.6427583637387018, "learning_rate": 0.00014098234559856748, "loss": 0.487, "step": 101 }, { "epoch": 0.9163391353172375, "grad_norm": 0.5447220351132417, "learning_rate": 0.00014043493861414235, "loss": 0.4721, "step": 102 }, { "epoch": 0.9253228523301515, "grad_norm": 0.5093866707759475, "learning_rate": 0.0001398808598536881, "loss": 0.4838, "step": 103 }, { "epoch": 0.9343065693430657, "grad_norm": 0.5848991205095823, "learning_rate": 0.00013932017048528858, "loss": 0.4732, "step": 104 }, { "epoch": 0.9432902863559798, "grad_norm": 0.46406379014333665, "learning_rate": 0.00013875293240681237, "loss": 0.4829, "step": 105 }, { "epoch": 0.9522740033688939, "grad_norm": 0.3855555200229203, "learning_rate": 0.00013817920823907946, "loss": 0.4699, "step": 106 }, { "epoch": 0.961257720381808, "grad_norm": 0.47252048758301546, "learning_rate": 0.0001375990613189483, "loss": 0.4709, "step": 107 }, { "epoch": 0.970241437394722, "grad_norm": 0.5128650342882652, "learning_rate": 0.00013701255569232356, "loss": 0.4625, "step": 108 }, { "epoch": 0.9792251544076361, "grad_norm": 0.40600999436688917, "learning_rate": 0.0001364197561070856, "loss": 0.4696, "step": 109 }, { "epoch": 0.9882088714205502, "grad_norm": 0.44134356238524264, "learning_rate": 0.00013582072800594284, "loss": 0.4775, "step": 110 }, { "epoch": 0.9971925884334644, "grad_norm": 0.5024734834453151, "learning_rate": 0.00013521553751920693, "loss": 0.4707, "step": 111 }, { "epoch": 1.0075800112296462, "grad_norm": 1.0362269730450384, "learning_rate": 0.00013460425145749236, "loss": 0.8319, "step": 112 }, { "epoch": 1.0165637282425604, "grad_norm": 1.7796385504660885, "learning_rate": 0.00013398693730434062, "loss": 0.4451, "step": 113 }, { "epoch": 1.0255474452554745, "grad_norm": 0.7356380006169851, "learning_rate": 0.00013336366320877062, "loss": 0.448, "step": 114 }, { "epoch": 1.0345311622683886, "grad_norm": 1.7391096880778294, "learning_rate": 0.00013273449797775503, "loss": 0.4648, "step": 115 }, { "epoch": 1.0435148792813027, "grad_norm": 0.973846535405518, "learning_rate": 0.00013209951106862437, "loss": 0.4656, "step": 116 }, { "epoch": 1.0524985962942168, "grad_norm": 1.292877541779147, "learning_rate": 0.00013145877258139918, "loss": 0.457, "step": 117 }, { "epoch": 1.0614823133071307, "grad_norm": 0.9804792453442043, "learning_rate": 0.0001308123532510513, "loss": 0.465, "step": 118 }, { "epoch": 1.0704660303200448, "grad_norm": 0.7853353924269662, "learning_rate": 0.00013016032443969493, "loss": 0.4457, "step": 119 }, { "epoch": 1.079449747332959, "grad_norm": 0.7657346345603476, "learning_rate": 0.0001295027581287087, "loss": 0.4472, "step": 120 }, { "epoch": 1.088433464345873, "grad_norm": 0.5359112242814267, "learning_rate": 0.00012883972691078892, "loss": 0.4509, "step": 121 }, { "epoch": 1.0974171813587872, "grad_norm": 0.5597649131000553, "learning_rate": 0.00012817130398193595, "loss": 0.4389, "step": 122 }, { "epoch": 1.1064008983717013, "grad_norm": 0.44706566167796036, "learning_rate": 0.00012749756313337353, "loss": 0.4552, "step": 123 }, { "epoch": 1.1153846153846154, "grad_norm": 0.43781190235972983, "learning_rate": 0.00012681857874340235, "loss": 0.431, "step": 124 }, { "epoch": 1.1243683323975295, "grad_norm": 0.3759045435946263, "learning_rate": 0.0001261344257691894, "loss": 0.4577, "step": 125 }, { "epoch": 1.1333520494104437, "grad_norm": 0.30888079178718136, "learning_rate": 0.00012544517973849247, "loss": 0.4267, "step": 126 }, { "epoch": 1.1423357664233578, "grad_norm": 0.33375312345057856, "learning_rate": 0.00012475091674132266, "loss": 0.4422, "step": 127 }, { "epoch": 1.1513194834362717, "grad_norm": 0.2801704573825818, "learning_rate": 0.00012405171342154409, "loss": 0.431, "step": 128 }, { "epoch": 1.1603032004491858, "grad_norm": 0.2923926642463168, "learning_rate": 0.00012334764696841273, "loss": 0.4439, "step": 129 }, { "epoch": 1.1692869174621, "grad_norm": 0.2515690098567012, "learning_rate": 0.0001226387951080551, "loss": 0.4351, "step": 130 }, { "epoch": 1.178270634475014, "grad_norm": 0.23889820398885311, "learning_rate": 0.0001219252360948876, "loss": 0.4355, "step": 131 }, { "epoch": 1.1872543514879281, "grad_norm": 0.25178540985599773, "learning_rate": 0.00012120704870297748, "loss": 0.4339, "step": 132 }, { "epoch": 1.1962380685008422, "grad_norm": 0.21429460359169059, "learning_rate": 0.00012048431221734653, "loss": 0.4342, "step": 133 }, { "epoch": 1.2052217855137564, "grad_norm": 0.24272912960518966, "learning_rate": 0.00011975710642521837, "loss": 0.4296, "step": 134 }, { "epoch": 1.2142055025266705, "grad_norm": 0.2302992707934062, "learning_rate": 0.0001190255116072102, "loss": 0.4305, "step": 135 }, { "epoch": 1.2231892195395846, "grad_norm": 0.19950449025995803, "learning_rate": 0.00011828960852847013, "loss": 0.4302, "step": 136 }, { "epoch": 1.2321729365524985, "grad_norm": 0.26523954308583547, "learning_rate": 0.00011754947842976111, "loss": 0.4336, "step": 137 }, { "epoch": 1.2411566535654126, "grad_norm": 0.30033806349738085, "learning_rate": 0.00011680520301849219, "loss": 0.4306, "step": 138 }, { "epoch": 1.2501403705783267, "grad_norm": 0.31461278548219257, "learning_rate": 0.00011605686445969838, "loss": 0.4303, "step": 139 }, { "epoch": 1.2591240875912408, "grad_norm": 0.3198850103038025, "learning_rate": 0.00011530454536697001, "loss": 0.429, "step": 140 }, { "epoch": 1.268107804604155, "grad_norm": 0.3187477850631323, "learning_rate": 0.00011454832879333245, "loss": 0.4311, "step": 141 }, { "epoch": 1.277091521617069, "grad_norm": 0.21115367860711695, "learning_rate": 0.00011378829822207734, "loss": 0.4281, "step": 142 }, { "epoch": 1.2860752386299832, "grad_norm": 0.24800096621587547, "learning_rate": 0.00011302453755754653, "loss": 0.4303, "step": 143 }, { "epoch": 1.2950589556428973, "grad_norm": 0.22634162707057526, "learning_rate": 0.00011225713111586919, "loss": 0.4354, "step": 144 }, { "epoch": 1.3040426726558114, "grad_norm": 0.22574772625477368, "learning_rate": 0.00011148616361565386, "loss": 0.4176, "step": 145 }, { "epoch": 1.3130263896687255, "grad_norm": 0.24036094076866749, "learning_rate": 0.00011071172016863557, "loss": 0.4299, "step": 146 }, { "epoch": 1.3220101066816397, "grad_norm": 0.28738541137546303, "learning_rate": 0.00010993388627028018, "loss": 0.4304, "step": 147 }, { "epoch": 1.3309938236945535, "grad_norm": 0.2609472165218046, "learning_rate": 0.0001091527477903457, "loss": 0.4332, "step": 148 }, { "epoch": 1.3399775407074677, "grad_norm": 0.26038582479198685, "learning_rate": 0.00010836839096340285, "loss": 0.4209, "step": 149 }, { "epoch": 1.3489612577203818, "grad_norm": 0.2316024258260121, "learning_rate": 0.00010758090237931497, "loss": 0.4283, "step": 150 }, { "epoch": 1.357944974733296, "grad_norm": 0.2098606063881437, "learning_rate": 0.0001067903689736789, "loss": 0.4263, "step": 151 }, { "epoch": 1.36692869174621, "grad_norm": 0.24683787256011072, "learning_rate": 0.00010599687801822763, "loss": 0.4141, "step": 152 }, { "epoch": 1.3759124087591241, "grad_norm": 0.24229393658697693, "learning_rate": 0.00010520051711119585, "loss": 0.4202, "step": 153 }, { "epoch": 1.3848961257720382, "grad_norm": 0.20606020348790752, "learning_rate": 0.00010440137416764941, "loss": 0.4227, "step": 154 }, { "epoch": 1.3938798427849521, "grad_norm": 0.2804421841986466, "learning_rate": 0.00010359953740977998, "loss": 0.4163, "step": 155 }, { "epoch": 1.4028635597978663, "grad_norm": 0.23544388812205802, "learning_rate": 0.00010279509535716549, "loss": 0.4141, "step": 156 }, { "epoch": 1.4118472768107804, "grad_norm": 0.19406146485494044, "learning_rate": 0.00010198813681699804, "loss": 0.4276, "step": 157 }, { "epoch": 1.4208309938236945, "grad_norm": 0.24473989108124805, "learning_rate": 0.00010117875087427998, "loss": 0.4168, "step": 158 }, { "epoch": 1.4298147108366086, "grad_norm": 0.29142953549457795, "learning_rate": 0.00010036702688198906, "loss": 0.4309, "step": 159 }, { "epoch": 1.4387984278495227, "grad_norm": 0.6467486970522354, "learning_rate": 9.955305445121455e-05, "loss": 0.4269, "step": 160 }, { "epoch": 1.4477821448624368, "grad_norm": 0.3055063735132388, "learning_rate": 9.87369234412642e-05, "loss": 0.4284, "step": 161 }, { "epoch": 1.456765861875351, "grad_norm": 0.24616493986031823, "learning_rate": 9.79187239497444e-05, "loss": 0.4333, "step": 162 }, { "epoch": 1.465749578888265, "grad_norm": 0.2384894678445937, "learning_rate": 9.709854630261359e-05, "loss": 0.427, "step": 163 }, { "epoch": 1.4747332959011792, "grad_norm": 0.24871583396920668, "learning_rate": 9.627648104421073e-05, "loss": 0.4221, "step": 164 }, { "epoch": 1.4837170129140933, "grad_norm": 0.22412182435439584, "learning_rate": 9.545261892725947e-05, "loss": 0.4247, "step": 165 }, { "epoch": 1.4927007299270074, "grad_norm": 0.2487283398786264, "learning_rate": 9.462705090284956e-05, "loss": 0.4202, "step": 166 }, { "epoch": 1.5016844469399215, "grad_norm": 0.29923719674921373, "learning_rate": 9.379986811039604e-05, "loss": 0.4271, "step": 167 }, { "epoch": 1.5106681639528357, "grad_norm": 1.3405716596708455, "learning_rate": 9.297116186757798e-05, "loss": 0.4289, "step": 168 }, { "epoch": 1.5196518809657495, "grad_norm": 0.38664890635523597, "learning_rate": 9.214102366025724e-05, "loss": 0.4186, "step": 169 }, { "epoch": 1.5286355979786637, "grad_norm": 0.4762048641691243, "learning_rate": 9.130954513237883e-05, "loss": 0.4295, "step": 170 }, { "epoch": 1.5376193149915778, "grad_norm": 0.4899933718508197, "learning_rate": 9.047681807585381e-05, "loss": 0.4238, "step": 171 }, { "epoch": 1.546603032004492, "grad_norm": 0.4626782645201007, "learning_rate": 8.964293442042585e-05, "loss": 0.427, "step": 172 }, { "epoch": 1.5555867490174058, "grad_norm": 0.47452287145204264, "learning_rate": 8.880798622352253e-05, "loss": 0.4221, "step": 173 }, { "epoch": 1.56457046603032, "grad_norm": 0.41668257298151723, "learning_rate": 8.797206566009252e-05, "loss": 0.4157, "step": 174 }, { "epoch": 1.573554183043234, "grad_norm": 0.2988059378537405, "learning_rate": 8.713526501242991e-05, "loss": 0.4287, "step": 175 }, { "epoch": 1.5825379000561481, "grad_norm": 0.449887308091825, "learning_rate": 8.629767665998652e-05, "loss": 0.428, "step": 176 }, { "epoch": 1.5915216170690623, "grad_norm": 0.4683956205571529, "learning_rate": 8.54593930691737e-05, "loss": 0.419, "step": 177 }, { "epoch": 1.6005053340819764, "grad_norm": 0.2834519006631767, "learning_rate": 8.462050678315429e-05, "loss": 0.4149, "step": 178 }, { "epoch": 1.6094890510948905, "grad_norm": 0.34501079837011184, "learning_rate": 8.37811104116263e-05, "loss": 0.4229, "step": 179 }, { "epoch": 1.6184727681078046, "grad_norm": 0.3768651775591254, "learning_rate": 8.294129662059918e-05, "loss": 0.4203, "step": 180 }, { "epoch": 1.6274564851207187, "grad_norm": 0.3767673855555118, "learning_rate": 8.210115812216378e-05, "loss": 0.4236, "step": 181 }, { "epoch": 1.6364402021336328, "grad_norm": 0.4116044954095348, "learning_rate": 8.126078766425732e-05, "loss": 0.4168, "step": 182 }, { "epoch": 1.645423919146547, "grad_norm": 0.24536748096315872, "learning_rate": 8.04202780204245e-05, "loss": 0.4112, "step": 183 }, { "epoch": 1.654407636159461, "grad_norm": 0.36986369671919106, "learning_rate": 7.957972197957554e-05, "loss": 0.4119, "step": 184 }, { "epoch": 1.6633913531723752, "grad_norm": 0.36469684298934046, "learning_rate": 7.87392123357427e-05, "loss": 0.4208, "step": 185 }, { "epoch": 1.6723750701852893, "grad_norm": 0.3062531478790991, "learning_rate": 7.789884187783625e-05, "loss": 0.4199, "step": 186 }, { "epoch": 1.6813587871982034, "grad_norm": 0.2245212911708979, "learning_rate": 7.705870337940085e-05, "loss": 0.419, "step": 187 }, { "epoch": 1.6903425042111173, "grad_norm": 0.2815117246044071, "learning_rate": 7.621888958837372e-05, "loss": 0.4156, "step": 188 }, { "epoch": 1.6993262212240314, "grad_norm": 0.25130028675485033, "learning_rate": 7.537949321684574e-05, "loss": 0.4227, "step": 189 }, { "epoch": 1.7083099382369455, "grad_norm": 0.24156587078920358, "learning_rate": 7.454060693082634e-05, "loss": 0.4128, "step": 190 }, { "epoch": 1.7172936552498597, "grad_norm": 0.24268807529448605, "learning_rate": 7.37023233400135e-05, "loss": 0.4225, "step": 191 }, { "epoch": 1.7262773722627736, "grad_norm": 0.34956262423936846, "learning_rate": 7.286473498757011e-05, "loss": 0.4224, "step": 192 }, { "epoch": 1.7352610892756877, "grad_norm": 0.2608226148563663, "learning_rate": 7.20279343399075e-05, "loss": 0.4222, "step": 193 }, { "epoch": 1.7442448062886018, "grad_norm": 0.23482325735709875, "learning_rate": 7.11920137764775e-05, "loss": 0.4131, "step": 194 }, { "epoch": 1.753228523301516, "grad_norm": 0.22733452979745924, "learning_rate": 7.035706557957416e-05, "loss": 0.4234, "step": 195 }, { "epoch": 1.76221224031443, "grad_norm": 0.31731646035712113, "learning_rate": 6.952318192414623e-05, "loss": 0.4274, "step": 196 }, { "epoch": 1.7711959573273441, "grad_norm": 0.2180643349500018, "learning_rate": 6.869045486762122e-05, "loss": 0.4138, "step": 197 }, { "epoch": 1.7801796743402583, "grad_norm": 0.24249072075661643, "learning_rate": 6.78589763397428e-05, "loss": 0.4073, "step": 198 }, { "epoch": 1.7891633913531724, "grad_norm": 0.30886511591698557, "learning_rate": 6.702883813242205e-05, "loss": 0.4113, "step": 199 }, { "epoch": 1.7981471083660865, "grad_norm": 0.20119667545730122, "learning_rate": 6.620013188960398e-05, "loss": 0.4129, "step": 200 }, { "epoch": 1.8071308253790006, "grad_norm": 0.24435752019081544, "learning_rate": 6.537294909715048e-05, "loss": 0.4163, "step": 201 }, { "epoch": 1.8161145423919147, "grad_norm": 0.2791733867754067, "learning_rate": 6.454738107274056e-05, "loss": 0.4292, "step": 202 }, { "epoch": 1.8250982594048288, "grad_norm": 0.18260363770459387, "learning_rate": 6.372351895578931e-05, "loss": 0.4134, "step": 203 }, { "epoch": 1.834081976417743, "grad_norm": 0.21731481060715926, "learning_rate": 6.290145369738644e-05, "loss": 0.4078, "step": 204 }, { "epoch": 1.843065693430657, "grad_norm": 0.21873653534247425, "learning_rate": 6.208127605025564e-05, "loss": 0.4193, "step": 205 }, { "epoch": 1.8520494104435712, "grad_norm": 0.22981600238165162, "learning_rate": 6.126307655873584e-05, "loss": 0.4076, "step": 206 }, { "epoch": 1.861033127456485, "grad_norm": 0.20566956828097804, "learning_rate": 6.0446945548785506e-05, "loss": 0.4151, "step": 207 }, { "epoch": 1.8700168444693992, "grad_norm": 0.2119512773707407, "learning_rate": 5.9632973118010964e-05, "loss": 0.4248, "step": 208 }, { "epoch": 1.8790005614823133, "grad_norm": 0.20654052551807822, "learning_rate": 5.882124912572005e-05, "loss": 0.4125, "step": 209 }, { "epoch": 1.8879842784952274, "grad_norm": 0.23846846179660958, "learning_rate": 5.801186318300195e-05, "loss": 0.4122, "step": 210 }, { "epoch": 1.8969679955081415, "grad_norm": 0.17491624726700092, "learning_rate": 5.720490464283452e-05, "loss": 0.4124, "step": 211 }, { "epoch": 1.9059517125210554, "grad_norm": 0.22200392942766323, "learning_rate": 5.6400462590220024e-05, "loss": 0.4213, "step": 212 }, { "epoch": 1.9149354295339696, "grad_norm": 0.23091340411259711, "learning_rate": 5.559862583235059e-05, "loss": 0.4032, "step": 213 }, { "epoch": 1.9239191465468837, "grad_norm": 0.1665570698719248, "learning_rate": 5.4799482888804174e-05, "loss": 0.4214, "step": 214 }, { "epoch": 1.9329028635597978, "grad_norm": 0.22812374386475856, "learning_rate": 5.400312198177239e-05, "loss": 0.4116, "step": 215 }, { "epoch": 1.941886580572712, "grad_norm": 0.19255968363783327, "learning_rate": 5.320963102632111e-05, "loss": 0.4153, "step": 216 }, { "epoch": 1.950870297585626, "grad_norm": 0.20277918206935358, "learning_rate": 5.2419097620685035e-05, "loss": 0.4047, "step": 217 }, { "epoch": 1.9598540145985401, "grad_norm": 0.23199199574198273, "learning_rate": 5.163160903659715e-05, "loss": 0.4234, "step": 218 }, { "epoch": 1.9688377316114543, "grad_norm": 0.20877288575006245, "learning_rate": 5.08472522096543e-05, "loss": 0.4143, "step": 219 }, { "epoch": 1.9778214486243684, "grad_norm": 0.1718225040486589, "learning_rate": 5.006611372971982e-05, "loss": 0.4106, "step": 220 }, { "epoch": 1.9868051656372825, "grad_norm": 0.21428383606764617, "learning_rate": 4.928827983136444e-05, "loss": 0.3995, "step": 221 }, { "epoch": 1.9957888826501966, "grad_norm": 0.18488960532491983, "learning_rate": 4.8513836384346175e-05, "loss": 0.4229, "step": 222 }, { "epoch": 2.0061763054463784, "grad_norm": 0.44509287430368083, "learning_rate": 4.774286888413083e-05, "loss": 0.72, "step": 223 }, { "epoch": 2.0151600224592925, "grad_norm": 0.349453446774281, "learning_rate": 4.697546244245351e-05, "loss": 0.3579, "step": 224 }, { "epoch": 2.0241437394722066, "grad_norm": 0.2983670407361141, "learning_rate": 4.6211701777922695e-05, "loss": 0.3791, "step": 225 }, { "epoch": 2.0331274564851207, "grad_norm": 0.35065646157559205, "learning_rate": 4.545167120666758e-05, "loss": 0.3665, "step": 226 }, { "epoch": 2.042111173498035, "grad_norm": 0.2928621163403631, "learning_rate": 4.469545463303e-05, "loss": 0.3715, "step": 227 }, { "epoch": 2.051094890510949, "grad_norm": 0.2954327190991827, "learning_rate": 4.394313554030163e-05, "loss": 0.3641, "step": 228 }, { "epoch": 2.060078607523863, "grad_norm": 0.29422712186180044, "learning_rate": 4.3194796981507835e-05, "loss": 0.3781, "step": 229 }, { "epoch": 2.069062324536777, "grad_norm": 0.3071711408148744, "learning_rate": 4.245052157023891e-05, "loss": 0.3645, "step": 230 }, { "epoch": 2.0780460415496913, "grad_norm": 0.23769683777534895, "learning_rate": 4.1710391471529894e-05, "loss": 0.3654, "step": 231 }, { "epoch": 2.0870297585626054, "grad_norm": 0.28768316660855947, "learning_rate": 4.0974488392789824e-05, "loss": 0.3681, "step": 232 }, { "epoch": 2.0960134755755195, "grad_norm": 0.22020511686667768, "learning_rate": 4.0242893574781645e-05, "loss": 0.3544, "step": 233 }, { "epoch": 2.1049971925884337, "grad_norm": 0.240285704968683, "learning_rate": 3.951568778265347e-05, "loss": 0.3685, "step": 234 }, { "epoch": 2.1139809096013478, "grad_norm": 0.2146794751583024, "learning_rate": 3.879295129702254e-05, "loss": 0.367, "step": 235 }, { "epoch": 2.1229646266142614, "grad_norm": 0.21870712204994086, "learning_rate": 3.807476390511243e-05, "loss": 0.3627, "step": 236 }, { "epoch": 2.1319483436271756, "grad_norm": 0.22541727469315342, "learning_rate": 3.7361204891944926e-05, "loss": 0.371, "step": 237 }, { "epoch": 2.1409320606400897, "grad_norm": 0.19265971553935168, "learning_rate": 3.66523530315873e-05, "loss": 0.3604, "step": 238 }, { "epoch": 2.149915777653004, "grad_norm": 0.23010447898223835, "learning_rate": 3.594828657845594e-05, "loss": 0.3624, "step": 239 }, { "epoch": 2.158899494665918, "grad_norm": 0.17138567803268676, "learning_rate": 3.524908325867735e-05, "loss": 0.3639, "step": 240 }, { "epoch": 2.167883211678832, "grad_norm": 0.1951941414970527, "learning_rate": 3.455482026150755e-05, "loss": 0.3544, "step": 241 }, { "epoch": 2.176866928691746, "grad_norm": 0.18352754967573434, "learning_rate": 3.3865574230810636e-05, "loss": 0.3773, "step": 242 }, { "epoch": 2.1858506457046603, "grad_norm": 0.21536452648283833, "learning_rate": 3.318142125659765e-05, "loss": 0.3825, "step": 243 }, { "epoch": 2.1948343627175744, "grad_norm": 0.1829109115920877, "learning_rate": 3.250243686662651e-05, "loss": 0.3579, "step": 244 }, { "epoch": 2.2038180797304885, "grad_norm": 0.18214718381067138, "learning_rate": 3.182869601806406e-05, "loss": 0.3626, "step": 245 }, { "epoch": 2.2128017967434026, "grad_norm": 0.1825125643696691, "learning_rate": 3.11602730892111e-05, "loss": 0.3573, "step": 246 }, { "epoch": 2.2217855137563167, "grad_norm": 0.18441564836737437, "learning_rate": 3.0497241871291333e-05, "loss": 0.3667, "step": 247 }, { "epoch": 2.230769230769231, "grad_norm": 0.18116751476899712, "learning_rate": 2.9839675560305088e-05, "loss": 0.3539, "step": 248 }, { "epoch": 2.239752947782145, "grad_norm": 0.18390826080033681, "learning_rate": 2.918764674894875e-05, "loss": 0.3618, "step": 249 }, { "epoch": 2.248736664795059, "grad_norm": 0.14612480392036745, "learning_rate": 2.8541227418600858e-05, "loss": 0.3624, "step": 250 }, { "epoch": 2.257720381807973, "grad_norm": 0.17070194783528053, "learning_rate": 2.7900488931375666e-05, "loss": 0.3664, "step": 251 }, { "epoch": 2.2667040988208873, "grad_norm": 0.1354347326780045, "learning_rate": 2.7265502022244993e-05, "loss": 0.372, "step": 252 }, { "epoch": 2.2756878158338014, "grad_norm": 0.14413453810840812, "learning_rate": 2.6636336791229393e-05, "loss": 0.3601, "step": 253 }, { "epoch": 2.2846715328467155, "grad_norm": 0.15479999812641926, "learning_rate": 2.6013062695659397e-05, "loss": 0.3779, "step": 254 }, { "epoch": 2.293655249859629, "grad_norm": 0.13022418657134416, "learning_rate": 2.539574854250769e-05, "loss": 0.3628, "step": 255 }, { "epoch": 2.3026389668725433, "grad_norm": 0.14668634420749707, "learning_rate": 2.4784462480793084e-05, "loss": 0.3706, "step": 256 }, { "epoch": 2.3116226838854574, "grad_norm": 0.12905147875548487, "learning_rate": 2.4179271994057184e-05, "loss": 0.3702, "step": 257 }, { "epoch": 2.3206064008983716, "grad_norm": 0.1447256804180326, "learning_rate": 2.3580243892914424e-05, "loss": 0.3573, "step": 258 }, { "epoch": 2.3295901179112857, "grad_norm": 0.11748170585015046, "learning_rate": 2.2987444307676478e-05, "loss": 0.3758, "step": 259 }, { "epoch": 2.3385738349242, "grad_norm": 0.12882383618208246, "learning_rate": 2.24009386810517e-05, "loss": 0.3511, "step": 260 }, { "epoch": 2.347557551937114, "grad_norm": 0.12865217121068986, "learning_rate": 2.1820791760920547e-05, "loss": 0.3716, "step": 261 }, { "epoch": 2.356541268950028, "grad_norm": 0.12081698628061978, "learning_rate": 2.124706759318766e-05, "loss": 0.3683, "step": 262 }, { "epoch": 2.365524985962942, "grad_norm": 0.12215193992569423, "learning_rate": 2.0679829514711436e-05, "loss": 0.3646, "step": 263 }, { "epoch": 2.3745087029758563, "grad_norm": 0.11354398681818435, "learning_rate": 2.0119140146311905e-05, "loss": 0.3686, "step": 264 }, { "epoch": 2.3834924199887704, "grad_norm": 0.12582747852018303, "learning_rate": 1.9565061385857663e-05, "loss": 0.3682, "step": 265 }, { "epoch": 2.3924761370016845, "grad_norm": 0.11257895081104091, "learning_rate": 1.9017654401432546e-05, "loss": 0.3717, "step": 266 }, { "epoch": 2.4014598540145986, "grad_norm": 0.12305919426390005, "learning_rate": 1.8476979624582937e-05, "loss": 0.3581, "step": 267 }, { "epoch": 2.4104435710275127, "grad_norm": 0.11203368697453076, "learning_rate": 1.794309674364641e-05, "loss": 0.37, "step": 268 }, { "epoch": 2.419427288040427, "grad_norm": 0.12463695189114064, "learning_rate": 1.7416064697162287e-05, "loss": 0.3736, "step": 269 }, { "epoch": 2.428411005053341, "grad_norm": 0.10917678973672461, "learning_rate": 1.6895941667365168e-05, "loss": 0.3535, "step": 270 }, { "epoch": 2.437394722066255, "grad_norm": 0.11281664768314911, "learning_rate": 1.6382785073761703e-05, "loss": 0.3571, "step": 271 }, { "epoch": 2.446378439079169, "grad_norm": 0.13666809456134899, "learning_rate": 1.587665156679185e-05, "loss": 0.3658, "step": 272 }, { "epoch": 2.4553621560920833, "grad_norm": 0.10808808623606102, "learning_rate": 1.537759702157482e-05, "loss": 0.358, "step": 273 }, { "epoch": 2.464345873104997, "grad_norm": 0.1116710945877027, "learning_rate": 1.4885676531740694e-05, "loss": 0.3758, "step": 274 }, { "epoch": 2.473329590117911, "grad_norm": 0.13267100870029716, "learning_rate": 1.4400944403348321e-05, "loss": 0.366, "step": 275 }, { "epoch": 2.482313307130825, "grad_norm": 0.11602385380874608, "learning_rate": 1.392345414889011e-05, "loss": 0.3702, "step": 276 }, { "epoch": 2.4912970241437393, "grad_norm": 0.1145440180474532, "learning_rate": 1.345325848138451e-05, "loss": 0.3511, "step": 277 }, { "epoch": 2.5002807411566534, "grad_norm": 0.11526664415049362, "learning_rate": 1.2990409308556679e-05, "loss": 0.3516, "step": 278 }, { "epoch": 2.5092644581695676, "grad_norm": 0.11013047619147111, "learning_rate": 1.2534957727108083e-05, "loss": 0.3633, "step": 279 }, { "epoch": 2.5182481751824817, "grad_norm": 0.10354704499832852, "learning_rate": 1.208695401707562e-05, "loss": 0.3733, "step": 280 }, { "epoch": 2.527231892195396, "grad_norm": 0.10236481273298736, "learning_rate": 1.1646447636280916e-05, "loss": 0.3555, "step": 281 }, { "epoch": 2.53621560920831, "grad_norm": 0.11167580752401748, "learning_rate": 1.1213487214870366e-05, "loss": 0.359, "step": 282 }, { "epoch": 2.545199326221224, "grad_norm": 0.1099233626090954, "learning_rate": 1.0788120549946535e-05, "loss": 0.3744, "step": 283 }, { "epoch": 2.554183043234138, "grad_norm": 0.11082716206350583, "learning_rate": 1.0370394600291633e-05, "loss": 0.3508, "step": 284 }, { "epoch": 2.5631667602470523, "grad_norm": 0.12103690964444253, "learning_rate": 9.960355481183357e-06, "loss": 0.3789, "step": 285 }, { "epoch": 2.5721504772599664, "grad_norm": 0.11920362715834164, "learning_rate": 9.55804845930401e-06, "loss": 0.3667, "step": 286 }, { "epoch": 2.5811341942728805, "grad_norm": 0.1044793820973793, "learning_rate": 9.163517947743218e-06, "loss": 0.368, "step": 287 }, { "epoch": 2.5901179112857946, "grad_norm": 0.1079810919458012, "learning_rate": 8.776807501094891e-06, "loss": 0.3568, "step": 288 }, { "epoch": 2.5991016282987087, "grad_norm": 0.10388558819512615, "learning_rate": 8.397959810648992e-06, "loss": 0.3564, "step": 289 }, { "epoch": 2.608085345311623, "grad_norm": 0.10325472697547591, "learning_rate": 8.027016699678581e-06, "loss": 0.3726, "step": 290 }, { "epoch": 2.6170690623245365, "grad_norm": 0.1027618868655184, "learning_rate": 7.664019118822676e-06, "loss": 0.367, "step": 291 }, { "epoch": 2.626052779337451, "grad_norm": 0.0929701185172587, "learning_rate": 7.30900714156551e-06, "loss": 0.3525, "step": 292 }, { "epoch": 2.6350364963503647, "grad_norm": 0.10035396955714077, "learning_rate": 6.962019959812516e-06, "loss": 0.3699, "step": 293 }, { "epoch": 2.6440202133632793, "grad_norm": 0.0928922534561677, "learning_rate": 6.623095879563774e-06, "loss": 0.3665, "step": 294 }, { "epoch": 2.653003930376193, "grad_norm": 0.10258049793200555, "learning_rate": 6.29227231668506e-06, "loss": 0.3583, "step": 295 }, { "epoch": 2.661987647389107, "grad_norm": 0.0944834724986228, "learning_rate": 5.9695857927774396e-06, "loss": 0.3679, "step": 296 }, { "epoch": 2.670971364402021, "grad_norm": 0.0924161403478989, "learning_rate": 5.655071931145317e-06, "loss": 0.3653, "step": 297 }, { "epoch": 2.6799550814149353, "grad_norm": 0.09826749411539404, "learning_rate": 5.3487654528638025e-06, "loss": 0.3646, "step": 298 }, { "epoch": 2.6889387984278494, "grad_norm": 0.09353683680477741, "learning_rate": 5.050700172945631e-06, "loss": 0.3577, "step": 299 }, { "epoch": 2.6979225154407636, "grad_norm": 0.08959975875667178, "learning_rate": 4.760908996608145e-06, "loss": 0.3564, "step": 300 }, { "epoch": 2.7069062324536777, "grad_norm": 0.08693868685516433, "learning_rate": 4.47942391564066e-06, "loss": 0.3654, "step": 301 }, { "epoch": 2.715889949466592, "grad_norm": 0.09893623632113224, "learning_rate": 4.206276004872729e-06, "loss": 0.372, "step": 302 }, { "epoch": 2.724873666479506, "grad_norm": 0.09139811989819925, "learning_rate": 3.941495418743557e-06, "loss": 0.3588, "step": 303 }, { "epoch": 2.73385738349242, "grad_norm": 0.09200970263158591, "learning_rate": 3.685111387973139e-06, "loss": 0.3545, "step": 304 }, { "epoch": 2.742841100505334, "grad_norm": 0.0872066817053106, "learning_rate": 3.4371522163352444e-06, "loss": 0.3634, "step": 305 }, { "epoch": 2.7518248175182483, "grad_norm": 0.08567103650346183, "learning_rate": 3.1976452775328302e-06, "loss": 0.3557, "step": 306 }, { "epoch": 2.7608085345311624, "grad_norm": 0.08300411963401971, "learning_rate": 2.9666170121760698e-06, "loss": 0.3652, "step": 307 }, { "epoch": 2.7697922515440765, "grad_norm": 0.08826810677038714, "learning_rate": 2.7440929248634216e-06, "loss": 0.3493, "step": 308 }, { "epoch": 2.7787759685569906, "grad_norm": 0.08440475179999454, "learning_rate": 2.5300975813660467e-06, "loss": 0.3799, "step": 309 }, { "epoch": 2.7877596855699043, "grad_norm": 0.08633683276388147, "learning_rate": 2.324654605915839e-06, "loss": 0.3485, "step": 310 }, { "epoch": 2.796743402582819, "grad_norm": 0.08317447366792373, "learning_rate": 2.127786678597348e-06, "loss": 0.3685, "step": 311 }, { "epoch": 2.8057271195957325, "grad_norm": 0.0873801185267239, "learning_rate": 1.9395155328440605e-06, "loss": 0.3665, "step": 312 }, { "epoch": 2.814710836608647, "grad_norm": 0.08702648151496879, "learning_rate": 1.7598619530390904e-06, "loss": 0.3597, "step": 313 }, { "epoch": 2.8236945536215607, "grad_norm": 0.08722398379689802, "learning_rate": 1.5888457722206529e-06, "loss": 0.3651, "step": 314 }, { "epoch": 2.832678270634475, "grad_norm": 0.08744721551504023, "learning_rate": 1.4264858698926022e-06, "loss": 0.3596, "step": 315 }, { "epoch": 2.841661987647389, "grad_norm": 0.0830808255878642, "learning_rate": 1.2728001699401583e-06, "loss": 0.3632, "step": 316 }, { "epoch": 2.850645704660303, "grad_norm": 0.07917152492798647, "learning_rate": 1.1278056386512514e-06, "loss": 0.3698, "step": 317 }, { "epoch": 2.859629421673217, "grad_norm": 0.07773500081958919, "learning_rate": 9.915182828434688e-07, "loss": 0.3634, "step": 318 }, { "epoch": 2.8686131386861313, "grad_norm": 0.08350629325245962, "learning_rate": 8.639531480969787e-07, "loss": 0.3589, "step": 319 }, { "epoch": 2.8775968556990454, "grad_norm": 0.08414289148049224, "learning_rate": 7.451243170935396e-07, "loss": 0.3684, "step": 320 }, { "epoch": 2.8865805727119596, "grad_norm": 0.07899075375043686, "learning_rate": 6.350449080618593e-07, "loss": 0.3617, "step": 321 }, { "epoch": 2.8955642897248737, "grad_norm": 0.07798394987626306, "learning_rate": 5.337270733293665e-07, "loss": 0.3613, "step": 322 }, { "epoch": 2.904548006737788, "grad_norm": 0.0828825070825642, "learning_rate": 4.4118199798066195e-07, "loss": 0.3654, "step": 323 }, { "epoch": 2.913531723750702, "grad_norm": 0.08209736577715379, "learning_rate": 3.5741989862271955e-07, "loss": 0.3592, "step": 324 }, { "epoch": 2.922515440763616, "grad_norm": 0.08335581990528856, "learning_rate": 2.824500222570148e-07, "loss": 0.3694, "step": 325 }, { "epoch": 2.93149915777653, "grad_norm": 0.07573943820594226, "learning_rate": 2.162806452587063e-07, "loss": 0.3545, "step": 326 }, { "epoch": 2.9404828747894443, "grad_norm": 0.0776837963138844, "learning_rate": 1.5891907246292192e-07, "loss": 0.3521, "step": 327 }, { "epoch": 2.9494665918023584, "grad_norm": 0.07757662948055022, "learning_rate": 1.103716363583729e-07, "loss": 0.3645, "step": 328 }, { "epoch": 2.958450308815272, "grad_norm": 0.0806492488802, "learning_rate": 7.064369638825064e-08, "loss": 0.36, "step": 329 }, { "epoch": 2.9674340258281866, "grad_norm": 0.07961635351123682, "learning_rate": 3.973963835858463e-08, "loss": 0.363, "step": 330 }, { "epoch": 2.9764177428411003, "grad_norm": 0.08063307866994539, "learning_rate": 1.766287395404298e-08, "loss": 0.3659, "step": 331 }, { "epoch": 2.985401459854015, "grad_norm": 0.07899457396008377, "learning_rate": 4.415840361300383e-09, "loss": 0.3593, "step": 332 }, { "epoch": 2.9943851768669285, "grad_norm": 0.08099286982566073, "learning_rate": 0.0, "loss": 0.3626, "step": 333 }, { "epoch": 2.9943851768669285, "step": 333, "total_flos": 4607052149424128.0, "train_loss": 0.4725978755736136, "train_runtime": 115789.2905, "train_samples_per_second": 2.953, "train_steps_per_second": 0.003 } ], "logging_steps": 1.0, "max_steps": 333, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4607052149424128.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }