{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.985163204747774, "eval_steps": 500, "global_step": 735, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006782534972445952, "grad_norm": 7.135720261444181, "learning_rate": 1.0810810810810812e-06, "loss": 1.1988, "step": 1 }, { "epoch": 0.013565069944891903, "grad_norm": 7.10061152713567, "learning_rate": 2.1621621621621623e-06, "loss": 1.1926, "step": 2 }, { "epoch": 0.020347604917337857, "grad_norm": 7.012644871929241, "learning_rate": 3.2432432432432437e-06, "loss": 1.1873, "step": 3 }, { "epoch": 0.027130139889783807, "grad_norm": 6.4062690637507425, "learning_rate": 4.324324324324325e-06, "loss": 1.1664, "step": 4 }, { "epoch": 0.03391267486222976, "grad_norm": 4.819533337520507, "learning_rate": 5.405405405405406e-06, "loss": 1.1316, "step": 5 }, { "epoch": 0.040695209834675714, "grad_norm": 3.0815069633752463, "learning_rate": 6.486486486486487e-06, "loss": 1.0897, "step": 6 }, { "epoch": 0.04747774480712166, "grad_norm": 4.965671310290254, "learning_rate": 7.567567567567569e-06, "loss": 1.0541, "step": 7 }, { "epoch": 0.05426027977956761, "grad_norm": 5.523597681999487, "learning_rate": 8.64864864864865e-06, "loss": 1.0812, "step": 8 }, { "epoch": 0.06104281475201356, "grad_norm": 4.7758478538296245, "learning_rate": 9.729729729729732e-06, "loss": 1.0615, "step": 9 }, { "epoch": 0.06782534972445951, "grad_norm": 4.37471758206583, "learning_rate": 1.0810810810810812e-05, "loss": 1.0034, "step": 10 }, { "epoch": 0.07460788469690546, "grad_norm": 3.415263530374433, "learning_rate": 1.1891891891891894e-05, "loss": 0.9965, "step": 11 }, { "epoch": 0.08139041966935143, "grad_norm": 2.1249671206986167, "learning_rate": 1.2972972972972975e-05, "loss": 0.9672, "step": 12 }, { "epoch": 0.08817295464179738, "grad_norm": 2.804510150071328, "learning_rate": 1.4054054054054055e-05, "loss": 0.9248, "step": 13 }, { "epoch": 0.09495548961424333, "grad_norm": 2.227059944094318, "learning_rate": 1.5135135135135138e-05, "loss": 0.9174, "step": 14 }, { "epoch": 0.10173802458668928, "grad_norm": 1.9376154207127252, "learning_rate": 1.6216216216216218e-05, "loss": 0.9111, "step": 15 }, { "epoch": 0.10852055955913523, "grad_norm": 1.5867898298795111, "learning_rate": 1.72972972972973e-05, "loss": 0.8742, "step": 16 }, { "epoch": 0.11530309453158118, "grad_norm": 1.824405438277367, "learning_rate": 1.8378378378378383e-05, "loss": 0.8663, "step": 17 }, { "epoch": 0.12208562950402713, "grad_norm": 1.2484939184792223, "learning_rate": 1.9459459459459463e-05, "loss": 0.8685, "step": 18 }, { "epoch": 0.1288681644764731, "grad_norm": 1.6950608297204033, "learning_rate": 2.054054054054054e-05, "loss": 0.8643, "step": 19 }, { "epoch": 0.13565069944891903, "grad_norm": 1.4257834623155268, "learning_rate": 2.1621621621621624e-05, "loss": 0.8589, "step": 20 }, { "epoch": 0.142433234421365, "grad_norm": 1.2788529677027458, "learning_rate": 2.2702702702702705e-05, "loss": 0.832, "step": 21 }, { "epoch": 0.14921576939381093, "grad_norm": 1.4653330199646428, "learning_rate": 2.378378378378379e-05, "loss": 0.8257, "step": 22 }, { "epoch": 0.1559983043662569, "grad_norm": 1.1706491825282224, "learning_rate": 2.4864864864864866e-05, "loss": 0.8154, "step": 23 }, { "epoch": 0.16278083933870285, "grad_norm": 1.575724350793329, "learning_rate": 2.594594594594595e-05, "loss": 0.83, "step": 24 }, { "epoch": 0.1695633743111488, "grad_norm": 1.273387962120446, "learning_rate": 2.702702702702703e-05, "loss": 0.8327, "step": 25 }, { "epoch": 0.17634590928359475, "grad_norm": 1.5867731702030288, "learning_rate": 2.810810810810811e-05, "loss": 0.8204, "step": 26 }, { "epoch": 0.1831284442560407, "grad_norm": 0.8564117760497798, "learning_rate": 2.918918918918919e-05, "loss": 0.8056, "step": 27 }, { "epoch": 0.18991097922848665, "grad_norm": 1.4715395766174262, "learning_rate": 3.0270270270270275e-05, "loss": 0.8055, "step": 28 }, { "epoch": 0.1966935142009326, "grad_norm": 0.8375306490548892, "learning_rate": 3.135135135135135e-05, "loss": 0.7937, "step": 29 }, { "epoch": 0.20347604917337855, "grad_norm": 1.620124215802785, "learning_rate": 3.2432432432432436e-05, "loss": 0.8001, "step": 30 }, { "epoch": 0.2102585841458245, "grad_norm": 1.553317483743022, "learning_rate": 3.351351351351351e-05, "loss": 0.7876, "step": 31 }, { "epoch": 0.21704111911827045, "grad_norm": 1.658909246335399, "learning_rate": 3.45945945945946e-05, "loss": 0.7891, "step": 32 }, { "epoch": 0.22382365409071642, "grad_norm": 1.1987131034054377, "learning_rate": 3.567567567567568e-05, "loss": 0.7979, "step": 33 }, { "epoch": 0.23060618906316235, "grad_norm": 2.2091757491329065, "learning_rate": 3.6756756756756765e-05, "loss": 0.7921, "step": 34 }, { "epoch": 0.23738872403560832, "grad_norm": 2.173619464441479, "learning_rate": 3.783783783783784e-05, "loss": 0.7997, "step": 35 }, { "epoch": 0.24417125900805425, "grad_norm": 1.3174649366078477, "learning_rate": 3.8918918918918926e-05, "loss": 0.768, "step": 36 }, { "epoch": 0.2509537939805002, "grad_norm": 2.560545381997849, "learning_rate": 4e-05, "loss": 0.7938, "step": 37 }, { "epoch": 0.2577363289529462, "grad_norm": 1.6460978741311298, "learning_rate": 4.108108108108108e-05, "loss": 0.7873, "step": 38 }, { "epoch": 0.2645188639253921, "grad_norm": 2.1426591768978938, "learning_rate": 4.2162162162162164e-05, "loss": 0.781, "step": 39 }, { "epoch": 0.27130139889783805, "grad_norm": 1.8160361307625077, "learning_rate": 4.324324324324325e-05, "loss": 0.7809, "step": 40 }, { "epoch": 0.278083933870284, "grad_norm": 1.6910805765281172, "learning_rate": 4.4324324324324325e-05, "loss": 0.7744, "step": 41 }, { "epoch": 0.28486646884273, "grad_norm": 1.9801882908067985, "learning_rate": 4.540540540540541e-05, "loss": 0.7732, "step": 42 }, { "epoch": 0.29164900381517594, "grad_norm": 1.3912495139596244, "learning_rate": 4.6486486486486486e-05, "loss": 0.778, "step": 43 }, { "epoch": 0.29843153878762185, "grad_norm": 1.9510858316270183, "learning_rate": 4.756756756756758e-05, "loss": 0.7628, "step": 44 }, { "epoch": 0.3052140737600678, "grad_norm": 1.9755444838845442, "learning_rate": 4.8648648648648654e-05, "loss": 0.77, "step": 45 }, { "epoch": 0.3119966087325138, "grad_norm": 1.7136629431095296, "learning_rate": 4.972972972972973e-05, "loss": 0.7776, "step": 46 }, { "epoch": 0.31877914370495974, "grad_norm": 1.9262891096799022, "learning_rate": 5.081081081081081e-05, "loss": 0.7591, "step": 47 }, { "epoch": 0.3255616786774057, "grad_norm": 1.8644045069770525, "learning_rate": 5.18918918918919e-05, "loss": 0.7564, "step": 48 }, { "epoch": 0.3323442136498516, "grad_norm": 1.273771704881583, "learning_rate": 5.2972972972972976e-05, "loss": 0.7615, "step": 49 }, { "epoch": 0.3391267486222976, "grad_norm": 2.4421063866094843, "learning_rate": 5.405405405405406e-05, "loss": 0.7803, "step": 50 }, { "epoch": 0.34590928359474354, "grad_norm": 2.8264810278747197, "learning_rate": 5.513513513513514e-05, "loss": 0.7614, "step": 51 }, { "epoch": 0.3526918185671895, "grad_norm": 2.2601831362375036, "learning_rate": 5.621621621621622e-05, "loss": 0.7727, "step": 52 }, { "epoch": 0.3594743535396354, "grad_norm": 3.49645659835363, "learning_rate": 5.7297297297297305e-05, "loss": 0.7589, "step": 53 }, { "epoch": 0.3662568885120814, "grad_norm": 2.8843912596082983, "learning_rate": 5.837837837837838e-05, "loss": 0.7561, "step": 54 }, { "epoch": 0.37303942348452734, "grad_norm": 2.7371626553055965, "learning_rate": 5.945945945945946e-05, "loss": 0.7468, "step": 55 }, { "epoch": 0.3798219584569733, "grad_norm": 2.4650676398708997, "learning_rate": 6.054054054054055e-05, "loss": 0.7501, "step": 56 }, { "epoch": 0.38660449342941927, "grad_norm": 2.617962450345242, "learning_rate": 6.162162162162163e-05, "loss": 0.7504, "step": 57 }, { "epoch": 0.3933870284018652, "grad_norm": 1.9045313378628999, "learning_rate": 6.27027027027027e-05, "loss": 0.7443, "step": 58 }, { "epoch": 0.40016956337431114, "grad_norm": 3.214875277973139, "learning_rate": 6.378378378378379e-05, "loss": 0.756, "step": 59 }, { "epoch": 0.4069520983467571, "grad_norm": 2.607985787054137, "learning_rate": 6.486486486486487e-05, "loss": 0.7575, "step": 60 }, { "epoch": 0.41373463331920307, "grad_norm": 2.462297061139689, "learning_rate": 6.594594594594596e-05, "loss": 0.7395, "step": 61 }, { "epoch": 0.420517168291649, "grad_norm": 1.7928802854683021, "learning_rate": 6.702702702702703e-05, "loss": 0.7443, "step": 62 }, { "epoch": 0.42729970326409494, "grad_norm": 2.8782764835293015, "learning_rate": 6.810810810810811e-05, "loss": 0.7574, "step": 63 }, { "epoch": 0.4340822382365409, "grad_norm": 2.107005343355103, "learning_rate": 6.91891891891892e-05, "loss": 0.7524, "step": 64 }, { "epoch": 0.44086477320898687, "grad_norm": 2.400586779165952, "learning_rate": 7.027027027027028e-05, "loss": 0.7539, "step": 65 }, { "epoch": 0.44764730818143283, "grad_norm": 1.8984648015553665, "learning_rate": 7.135135135135136e-05, "loss": 0.7661, "step": 66 }, { "epoch": 0.45442984315387874, "grad_norm": 2.701931984195687, "learning_rate": 7.243243243243243e-05, "loss": 0.7423, "step": 67 }, { "epoch": 0.4612123781263247, "grad_norm": 3.1216702832116807, "learning_rate": 7.351351351351353e-05, "loss": 0.7558, "step": 68 }, { "epoch": 0.46799491309877067, "grad_norm": 1.7791622339381934, "learning_rate": 7.45945945945946e-05, "loss": 0.7338, "step": 69 }, { "epoch": 0.47477744807121663, "grad_norm": 2.283223713003048, "learning_rate": 7.567567567567568e-05, "loss": 0.7512, "step": 70 }, { "epoch": 0.48155998304366254, "grad_norm": 2.518755689574504, "learning_rate": 7.675675675675675e-05, "loss": 0.7371, "step": 71 }, { "epoch": 0.4883425180161085, "grad_norm": 1.99481044416433, "learning_rate": 7.783783783783785e-05, "loss": 0.7478, "step": 72 }, { "epoch": 0.49512505298855447, "grad_norm": 2.7463149548347183, "learning_rate": 7.891891891891892e-05, "loss": 0.7365, "step": 73 }, { "epoch": 0.5019075879610004, "grad_norm": 2.138101333819343, "learning_rate": 8e-05, "loss": 0.7254, "step": 74 }, { "epoch": 0.5086901229334464, "grad_norm": 2.640544503377941, "learning_rate": 7.999954822103665e-05, "loss": 0.7417, "step": 75 }, { "epoch": 0.5154726579058924, "grad_norm": 2.2402410683129688, "learning_rate": 7.999819289435179e-05, "loss": 0.7354, "step": 76 }, { "epoch": 0.5222551928783383, "grad_norm": 2.49341837988937, "learning_rate": 7.999593405056084e-05, "loss": 0.738, "step": 77 }, { "epoch": 0.5290377278507842, "grad_norm": 2.3287753265703253, "learning_rate": 7.999277174068872e-05, "loss": 0.7247, "step": 78 }, { "epoch": 0.5358202628232301, "grad_norm": 2.2524400032207716, "learning_rate": 7.998870603616864e-05, "loss": 0.7368, "step": 79 }, { "epoch": 0.5426027977956761, "grad_norm": 2.686780411560504, "learning_rate": 7.998373702884062e-05, "loss": 0.7315, "step": 80 }, { "epoch": 0.5493853327681221, "grad_norm": 1.7256862069861159, "learning_rate": 7.997786483094931e-05, "loss": 0.7492, "step": 81 }, { "epoch": 0.556167867740568, "grad_norm": 2.6274112743201523, "learning_rate": 7.997108957514146e-05, "loss": 0.7229, "step": 82 }, { "epoch": 0.562950402713014, "grad_norm": 2.0731276463095596, "learning_rate": 7.9963411414463e-05, "loss": 0.7375, "step": 83 }, { "epoch": 0.56973293768546, "grad_norm": 2.6126411934480123, "learning_rate": 7.99548305223555e-05, "loss": 0.7284, "step": 84 }, { "epoch": 0.5765154726579059, "grad_norm": 1.4633346753767393, "learning_rate": 7.994534709265226e-05, "loss": 0.7248, "step": 85 }, { "epoch": 0.5832980076303519, "grad_norm": 3.4984880607635307, "learning_rate": 7.993496133957401e-05, "loss": 0.7517, "step": 86 }, { "epoch": 0.5900805426027977, "grad_norm": 2.4422440886664147, "learning_rate": 7.992367349772398e-05, "loss": 0.7378, "step": 87 }, { "epoch": 0.5968630775752437, "grad_norm": 1.8295430254229907, "learning_rate": 7.991148382208265e-05, "loss": 0.7241, "step": 88 }, { "epoch": 0.6036456125476897, "grad_norm": 2.9619629036982755, "learning_rate": 7.989839258800196e-05, "loss": 0.7422, "step": 89 }, { "epoch": 0.6104281475201356, "grad_norm": 1.9664968756669656, "learning_rate": 7.988440009119911e-05, "loss": 0.751, "step": 90 }, { "epoch": 0.6172106824925816, "grad_norm": 1.6194583900993504, "learning_rate": 7.986950664774992e-05, "loss": 0.7346, "step": 91 }, { "epoch": 0.6239932174650276, "grad_norm": 2.8148837821592796, "learning_rate": 7.985371259408157e-05, "loss": 0.7413, "step": 92 }, { "epoch": 0.6307757524374735, "grad_norm": 1.486533554565982, "learning_rate": 7.983701828696515e-05, "loss": 0.7296, "step": 93 }, { "epoch": 0.6375582874099195, "grad_norm": 3.229587748852655, "learning_rate": 7.98194241035075e-05, "loss": 0.7483, "step": 94 }, { "epoch": 0.6443408223823655, "grad_norm": 2.518619876391423, "learning_rate": 7.980093044114269e-05, "loss": 0.7524, "step": 95 }, { "epoch": 0.6511233573548114, "grad_norm": 2.6606747542222915, "learning_rate": 7.978153771762311e-05, "loss": 0.7531, "step": 96 }, { "epoch": 0.6579058923272573, "grad_norm": 1.7367257209193254, "learning_rate": 7.976124637101e-05, "loss": 0.736, "step": 97 }, { "epoch": 0.6646884272997032, "grad_norm": 1.8350871328487093, "learning_rate": 7.974005685966354e-05, "loss": 0.7248, "step": 98 }, { "epoch": 0.6714709622721492, "grad_norm": 1.5007282529505472, "learning_rate": 7.971796966223248e-05, "loss": 0.7498, "step": 99 }, { "epoch": 0.6782534972445952, "grad_norm": 1.4248528231943807, "learning_rate": 7.969498527764341e-05, "loss": 0.7208, "step": 100 }, { "epoch": 0.6850360322170411, "grad_norm": 2.217134309648403, "learning_rate": 7.967110422508936e-05, "loss": 0.7295, "step": 101 }, { "epoch": 0.6918185671894871, "grad_norm": 2.2429893368943277, "learning_rate": 7.964632704401823e-05, "loss": 0.7429, "step": 102 }, { "epoch": 0.698601102161933, "grad_norm": 1.317977635899577, "learning_rate": 7.962065429412046e-05, "loss": 0.7311, "step": 103 }, { "epoch": 0.705383637134379, "grad_norm": 3.1540990631312718, "learning_rate": 7.959408655531646e-05, "loss": 0.7331, "step": 104 }, { "epoch": 0.712166172106825, "grad_norm": 2.0946781410468382, "learning_rate": 7.956662442774351e-05, "loss": 0.7357, "step": 105 }, { "epoch": 0.7189487070792708, "grad_norm": 3.6514929315006834, "learning_rate": 7.953826853174218e-05, "loss": 0.724, "step": 106 }, { "epoch": 0.7257312420517168, "grad_norm": 2.743797555116305, "learning_rate": 7.950901950784236e-05, "loss": 0.7337, "step": 107 }, { "epoch": 0.7325137770241628, "grad_norm": 2.6025703631899364, "learning_rate": 7.947887801674872e-05, "loss": 0.7231, "step": 108 }, { "epoch": 0.7392963119966087, "grad_norm": 2.33134578145788, "learning_rate": 7.944784473932583e-05, "loss": 0.7268, "step": 109 }, { "epoch": 0.7460788469690547, "grad_norm": 2.369166475084867, "learning_rate": 7.941592037658279e-05, "loss": 0.7175, "step": 110 }, { "epoch": 0.7528613819415007, "grad_norm": 1.5983852946337034, "learning_rate": 7.93831056496574e-05, "loss": 0.7215, "step": 111 }, { "epoch": 0.7596439169139466, "grad_norm": 2.3446201220972696, "learning_rate": 7.934940129979979e-05, "loss": 0.7314, "step": 112 }, { "epoch": 0.7664264518863926, "grad_norm": 1.872428876067427, "learning_rate": 7.931480808835577e-05, "loss": 0.7287, "step": 113 }, { "epoch": 0.7732089868588385, "grad_norm": 1.982878649441548, "learning_rate": 7.927932679674964e-05, "loss": 0.7193, "step": 114 }, { "epoch": 0.7799915218312844, "grad_norm": 1.5101882188042894, "learning_rate": 7.924295822646643e-05, "loss": 0.7217, "step": 115 }, { "epoch": 0.7867740568037304, "grad_norm": 1.5820041393952022, "learning_rate": 7.92057031990339e-05, "loss": 0.7199, "step": 116 }, { "epoch": 0.7935565917761763, "grad_norm": 1.5630207257272397, "learning_rate": 7.91675625560039e-05, "loss": 0.7141, "step": 117 }, { "epoch": 0.8003391267486223, "grad_norm": 2.457059696195913, "learning_rate": 7.91285371589335e-05, "loss": 0.7082, "step": 118 }, { "epoch": 0.8071216617210683, "grad_norm": 1.6028005358543407, "learning_rate": 7.908862788936532e-05, "loss": 0.7149, "step": 119 }, { "epoch": 0.8139041966935142, "grad_norm": 1.8116395322587184, "learning_rate": 7.904783564880779e-05, "loss": 0.7123, "step": 120 }, { "epoch": 0.8206867316659602, "grad_norm": 2.0950388250383996, "learning_rate": 7.900616135871474e-05, "loss": 0.7354, "step": 121 }, { "epoch": 0.8274692666384061, "grad_norm": 1.8624393412855933, "learning_rate": 7.896360596046453e-05, "loss": 0.7156, "step": 122 }, { "epoch": 0.8342518016108521, "grad_norm": 1.4851216429547307, "learning_rate": 7.892017041533886e-05, "loss": 0.7114, "step": 123 }, { "epoch": 0.841034336583298, "grad_norm": 2.9385975867591765, "learning_rate": 7.887585570450098e-05, "loss": 0.7127, "step": 124 }, { "epoch": 0.8478168715557439, "grad_norm": 1.8479656443451273, "learning_rate": 7.883066282897362e-05, "loss": 0.7173, "step": 125 }, { "epoch": 0.8545994065281899, "grad_norm": 2.894143467483064, "learning_rate": 7.878459280961629e-05, "loss": 0.7183, "step": 126 }, { "epoch": 0.8613819415006359, "grad_norm": 2.167706950028001, "learning_rate": 7.873764668710228e-05, "loss": 0.7237, "step": 127 }, { "epoch": 0.8681644764730818, "grad_norm": 2.646470616726372, "learning_rate": 7.868982552189514e-05, "loss": 0.7132, "step": 128 }, { "epoch": 0.8749470114455278, "grad_norm": 2.242758826547572, "learning_rate": 7.864113039422464e-05, "loss": 0.7296, "step": 129 }, { "epoch": 0.8817295464179737, "grad_norm": 2.493893447513006, "learning_rate": 7.859156240406252e-05, "loss": 0.7185, "step": 130 }, { "epoch": 0.8885120813904197, "grad_norm": 1.721360057791857, "learning_rate": 7.854112267109756e-05, "loss": 0.7244, "step": 131 }, { "epoch": 0.8952946163628657, "grad_norm": 2.6683450004265583, "learning_rate": 7.848981233471024e-05, "loss": 0.7149, "step": 132 }, { "epoch": 0.9020771513353115, "grad_norm": 2.061224678516105, "learning_rate": 7.843763255394711e-05, "loss": 0.7118, "step": 133 }, { "epoch": 0.9088596863077575, "grad_norm": 1.9936485418092913, "learning_rate": 7.838458450749452e-05, "loss": 0.7057, "step": 134 }, { "epoch": 0.9156422212802034, "grad_norm": 1.5127328421814061, "learning_rate": 7.833066939365206e-05, "loss": 0.7043, "step": 135 }, { "epoch": 0.9224247562526494, "grad_norm": 2.040325487116571, "learning_rate": 7.827588843030543e-05, "loss": 0.7087, "step": 136 }, { "epoch": 0.9292072912250954, "grad_norm": 1.6268257121266954, "learning_rate": 7.822024285489896e-05, "loss": 0.7105, "step": 137 }, { "epoch": 0.9359898261975413, "grad_norm": 2.3150904956674263, "learning_rate": 7.81637339244077e-05, "loss": 0.7139, "step": 138 }, { "epoch": 0.9427723611699873, "grad_norm": 1.841068832751663, "learning_rate": 7.810636291530893e-05, "loss": 0.7052, "step": 139 }, { "epoch": 0.9495548961424333, "grad_norm": 1.9983783549939655, "learning_rate": 7.804813112355339e-05, "loss": 0.7115, "step": 140 }, { "epoch": 0.9563374311148792, "grad_norm": 1.9754428939529725, "learning_rate": 7.798903986453603e-05, "loss": 0.7142, "step": 141 }, { "epoch": 0.9631199660873251, "grad_norm": 1.735226023700741, "learning_rate": 7.792909047306623e-05, "loss": 0.7205, "step": 142 }, { "epoch": 0.969902501059771, "grad_norm": 2.184380091482261, "learning_rate": 7.786828430333769e-05, "loss": 0.7107, "step": 143 }, { "epoch": 0.976685036032217, "grad_norm": 1.8757020725047386, "learning_rate": 7.78066227288978e-05, "loss": 0.7198, "step": 144 }, { "epoch": 0.983467571004663, "grad_norm": 1.0079051823953802, "learning_rate": 7.77441071426167e-05, "loss": 0.714, "step": 145 }, { "epoch": 0.9902501059771089, "grad_norm": 1.5288723958450625, "learning_rate": 7.768073895665573e-05, "loss": 0.7087, "step": 146 }, { "epoch": 0.9970326409495549, "grad_norm": 2.3013059467419996, "learning_rate": 7.761651960243554e-05, "loss": 0.7366, "step": 147 }, { "epoch": 1.0038151759220009, "grad_norm": 2.269978916078414, "learning_rate": 7.755145053060378e-05, "loss": 1.1363, "step": 148 }, { "epoch": 1.0105977108944468, "grad_norm": 1.6332374526993498, "learning_rate": 7.748553321100238e-05, "loss": 0.697, "step": 149 }, { "epoch": 1.0173802458668928, "grad_norm": 2.0144072844842054, "learning_rate": 7.741876913263422e-05, "loss": 0.7169, "step": 150 }, { "epoch": 1.0241627808393388, "grad_norm": 1.9963737399863932, "learning_rate": 7.735115980362964e-05, "loss": 0.6942, "step": 151 }, { "epoch": 1.0309453158117847, "grad_norm": 2.010956402670031, "learning_rate": 7.728270675121224e-05, "loss": 0.7108, "step": 152 }, { "epoch": 1.0377278507842307, "grad_norm": 1.1256640161683924, "learning_rate": 7.721341152166448e-05, "loss": 0.706, "step": 153 }, { "epoch": 1.0445103857566767, "grad_norm": 2.7360045649430287, "learning_rate": 7.714327568029272e-05, "loss": 0.7035, "step": 154 }, { "epoch": 1.0512929207291226, "grad_norm": 1.8273909265523915, "learning_rate": 7.707230081139184e-05, "loss": 0.7064, "step": 155 }, { "epoch": 1.0580754557015684, "grad_norm": 2.155560932773446, "learning_rate": 7.700048851820946e-05, "loss": 0.6991, "step": 156 }, { "epoch": 1.0648579906740143, "grad_norm": 2.1669039610273493, "learning_rate": 7.692784042290976e-05, "loss": 0.7099, "step": 157 }, { "epoch": 1.0716405256464603, "grad_norm": 1.503612527062841, "learning_rate": 7.685435816653681e-05, "loss": 0.6973, "step": 158 }, { "epoch": 1.0784230606189062, "grad_norm": 1.937309301259521, "learning_rate": 7.678004340897747e-05, "loss": 0.7049, "step": 159 }, { "epoch": 1.0852055955913522, "grad_norm": 1.3969091003754739, "learning_rate": 7.670489782892396e-05, "loss": 0.6886, "step": 160 }, { "epoch": 1.0919881305637982, "grad_norm": 1.9964168232948958, "learning_rate": 7.662892312383592e-05, "loss": 0.6883, "step": 161 }, { "epoch": 1.0987706655362441, "grad_norm": 1.7569755198383514, "learning_rate": 7.655212100990195e-05, "loss": 0.6987, "step": 162 }, { "epoch": 1.10555320050869, "grad_norm": 1.4447704185089303, "learning_rate": 7.647449322200108e-05, "loss": 0.6906, "step": 163 }, { "epoch": 1.112335735481136, "grad_norm": 1.2497557944098203, "learning_rate": 7.639604151366339e-05, "loss": 0.7009, "step": 164 }, { "epoch": 1.119118270453582, "grad_norm": 2.3116148247793267, "learning_rate": 7.631676765703042e-05, "loss": 0.6895, "step": 165 }, { "epoch": 1.125900805426028, "grad_norm": 1.330812583383646, "learning_rate": 7.623667344281522e-05, "loss": 0.7004, "step": 166 }, { "epoch": 1.132683340398474, "grad_norm": 1.8610464265540336, "learning_rate": 7.615576068026187e-05, "loss": 0.7077, "step": 167 }, { "epoch": 1.13946587537092, "grad_norm": 1.7667897999541395, "learning_rate": 7.607403119710453e-05, "loss": 0.6913, "step": 168 }, { "epoch": 1.1462484103433659, "grad_norm": 1.208336959762333, "learning_rate": 7.599148683952628e-05, "loss": 0.6839, "step": 169 }, { "epoch": 1.1530309453158119, "grad_norm": 1.9033319484816924, "learning_rate": 7.590812947211733e-05, "loss": 0.6898, "step": 170 }, { "epoch": 1.1598134802882578, "grad_norm": 1.213830580906014, "learning_rate": 7.582396097783294e-05, "loss": 0.6866, "step": 171 }, { "epoch": 1.1665960152607038, "grad_norm": 2.309488724064315, "learning_rate": 7.573898325795086e-05, "loss": 0.697, "step": 172 }, { "epoch": 1.1733785502331497, "grad_norm": 1.3337697012970882, "learning_rate": 7.565319823202838e-05, "loss": 0.7083, "step": 173 }, { "epoch": 1.1801610852055955, "grad_norm": 1.644453886623139, "learning_rate": 7.556660783785904e-05, "loss": 0.6905, "step": 174 }, { "epoch": 1.1869436201780414, "grad_norm": 1.3203501723312958, "learning_rate": 7.547921403142874e-05, "loss": 0.7048, "step": 175 }, { "epoch": 1.1937261551504874, "grad_norm": 1.4545383798679439, "learning_rate": 7.539101878687165e-05, "loss": 0.6865, "step": 176 }, { "epoch": 1.2005086901229334, "grad_norm": 2.1879709481996987, "learning_rate": 7.53020240964256e-05, "loss": 0.6964, "step": 177 }, { "epoch": 1.2072912250953793, "grad_norm": 1.8322469465167428, "learning_rate": 7.5212231970387e-05, "loss": 0.6921, "step": 178 }, { "epoch": 1.2140737600678253, "grad_norm": 1.2841938775963582, "learning_rate": 7.512164443706555e-05, "loss": 0.6816, "step": 179 }, { "epoch": 1.2208562950402713, "grad_norm": 3.1832942767833003, "learning_rate": 7.503026354273834e-05, "loss": 0.7107, "step": 180 }, { "epoch": 1.2276388300127172, "grad_norm": 2.223671181736338, "learning_rate": 7.493809135160367e-05, "loss": 0.7068, "step": 181 }, { "epoch": 1.2344213649851632, "grad_norm": 2.5365710286141727, "learning_rate": 7.484512994573438e-05, "loss": 0.6997, "step": 182 }, { "epoch": 1.2412038999576092, "grad_norm": 1.7033409104571895, "learning_rate": 7.475138142503083e-05, "loss": 0.6973, "step": 183 }, { "epoch": 1.2479864349300551, "grad_norm": 2.666703064771071, "learning_rate": 7.465684790717354e-05, "loss": 0.6904, "step": 184 }, { "epoch": 1.254768969902501, "grad_norm": 1.6449288965298707, "learning_rate": 7.45615315275752e-05, "loss": 0.7008, "step": 185 }, { "epoch": 1.261551504874947, "grad_norm": 2.670948821208354, "learning_rate": 7.446543443933258e-05, "loss": 0.7187, "step": 186 }, { "epoch": 1.268334039847393, "grad_norm": 2.052845388192151, "learning_rate": 7.436855881317784e-05, "loss": 0.7154, "step": 187 }, { "epoch": 1.275116574819839, "grad_norm": 2.3638246999276054, "learning_rate": 7.427090683742947e-05, "loss": 0.6942, "step": 188 }, { "epoch": 1.281899109792285, "grad_norm": 2.367916777331452, "learning_rate": 7.417248071794288e-05, "loss": 0.6978, "step": 189 }, { "epoch": 1.288681644764731, "grad_norm": 1.919070978827847, "learning_rate": 7.40732826780606e-05, "loss": 0.6778, "step": 190 }, { "epoch": 1.2954641797371766, "grad_norm": 1.5807447282945328, "learning_rate": 7.397331495856199e-05, "loss": 0.6878, "step": 191 }, { "epoch": 1.3022467147096228, "grad_norm": 1.5725379589793347, "learning_rate": 7.38725798176127e-05, "loss": 0.6942, "step": 192 }, { "epoch": 1.3090292496820686, "grad_norm": 1.5854647888888829, "learning_rate": 7.37710795307136e-05, "loss": 0.6942, "step": 193 }, { "epoch": 1.3158117846545148, "grad_norm": 0.9840396505987244, "learning_rate": 7.36688163906494e-05, "loss": 0.6822, "step": 194 }, { "epoch": 1.3225943196269605, "grad_norm": 1.9293846538238884, "learning_rate": 7.356579270743689e-05, "loss": 0.707, "step": 195 }, { "epoch": 1.3293768545994065, "grad_norm": 1.297177627346237, "learning_rate": 7.346201080827272e-05, "loss": 0.6966, "step": 196 }, { "epoch": 1.3361593895718524, "grad_norm": 2.3709366791443456, "learning_rate": 7.335747303748079e-05, "loss": 0.7001, "step": 197 }, { "epoch": 1.3429419245442984, "grad_norm": 1.7251163867182135, "learning_rate": 7.325218175645942e-05, "loss": 0.694, "step": 198 }, { "epoch": 1.3497244595167444, "grad_norm": 2.088301626825016, "learning_rate": 7.314613934362788e-05, "loss": 0.6879, "step": 199 }, { "epoch": 1.3565069944891903, "grad_norm": 1.9529188107557567, "learning_rate": 7.303934819437277e-05, "loss": 0.6944, "step": 200 }, { "epoch": 1.3632895294616363, "grad_norm": 1.6962527949805855, "learning_rate": 7.293181072099377e-05, "loss": 0.6917, "step": 201 }, { "epoch": 1.3700720644340822, "grad_norm": 1.6169230360470628, "learning_rate": 7.282352935264934e-05, "loss": 0.684, "step": 202 }, { "epoch": 1.3768545994065282, "grad_norm": 1.485732940861678, "learning_rate": 7.271450653530167e-05, "loss": 0.6862, "step": 203 }, { "epoch": 1.3836371343789742, "grad_norm": 1.6704659351121771, "learning_rate": 7.260474473166154e-05, "loss": 0.6932, "step": 204 }, { "epoch": 1.3904196693514201, "grad_norm": 1.5799837137385815, "learning_rate": 7.249424642113266e-05, "loss": 0.697, "step": 205 }, { "epoch": 1.397202204323866, "grad_norm": 1.2977311485642662, "learning_rate": 7.238301409975561e-05, "loss": 0.6919, "step": 206 }, { "epoch": 1.403984739296312, "grad_norm": 1.451392232428334, "learning_rate": 7.227105028015156e-05, "loss": 0.699, "step": 207 }, { "epoch": 1.410767274268758, "grad_norm": 1.2941406989928823, "learning_rate": 7.21583574914654e-05, "loss": 0.6708, "step": 208 }, { "epoch": 1.417549809241204, "grad_norm": 1.6593482637225496, "learning_rate": 7.204493827930869e-05, "loss": 0.6871, "step": 209 }, { "epoch": 1.4243323442136497, "grad_norm": 1.528014842634458, "learning_rate": 7.193079520570217e-05, "loss": 0.6777, "step": 210 }, { "epoch": 1.431114879186096, "grad_norm": 1.2836973493312132, "learning_rate": 7.18159308490178e-05, "loss": 0.6741, "step": 211 }, { "epoch": 1.4378974141585417, "grad_norm": 1.2069873065957852, "learning_rate": 7.170034780392055e-05, "loss": 0.6852, "step": 212 }, { "epoch": 1.4446799491309876, "grad_norm": 1.6652484448685945, "learning_rate": 7.158404868130988e-05, "loss": 0.6711, "step": 213 }, { "epoch": 1.4514624841034336, "grad_norm": 1.2576599603912495, "learning_rate": 7.14670361082606e-05, "loss": 0.6849, "step": 214 }, { "epoch": 1.4582450190758796, "grad_norm": 1.5405832213240351, "learning_rate": 7.134931272796368e-05, "loss": 0.691, "step": 215 }, { "epoch": 1.4650275540483255, "grad_norm": 1.684185898740359, "learning_rate": 7.123088119966647e-05, "loss": 0.6868, "step": 216 }, { "epoch": 1.4718100890207715, "grad_norm": 1.0827904024479826, "learning_rate": 7.11117441986126e-05, "loss": 0.6878, "step": 217 }, { "epoch": 1.4785926239932174, "grad_norm": 1.5453825081747208, "learning_rate": 7.099190441598161e-05, "loss": 0.6838, "step": 218 }, { "epoch": 1.4853751589656634, "grad_norm": 1.0691158300615973, "learning_rate": 7.087136455882816e-05, "loss": 0.6738, "step": 219 }, { "epoch": 1.4921576939381094, "grad_norm": 2.077881844593604, "learning_rate": 7.07501273500208e-05, "loss": 0.6912, "step": 220 }, { "epoch": 1.4989402289105553, "grad_norm": 0.9909091986869899, "learning_rate": 7.06281955281806e-05, "loss": 0.6875, "step": 221 }, { "epoch": 1.5057227638830013, "grad_norm": 1.826572971343745, "learning_rate": 7.050557184761915e-05, "loss": 0.6855, "step": 222 }, { "epoch": 1.5125052988554473, "grad_norm": 1.279113591847188, "learning_rate": 7.038225907827639e-05, "loss": 0.6961, "step": 223 }, { "epoch": 1.5192878338278932, "grad_norm": 1.229774591769691, "learning_rate": 7.02582600056581e-05, "loss": 0.6802, "step": 224 }, { "epoch": 1.526070368800339, "grad_norm": 1.3203007597562126, "learning_rate": 7.013357743077289e-05, "loss": 0.6965, "step": 225 }, { "epoch": 1.5328529037727852, "grad_norm": 1.2862135502175833, "learning_rate": 7.000821417006898e-05, "loss": 0.6815, "step": 226 }, { "epoch": 1.539635438745231, "grad_norm": 1.0628503441369677, "learning_rate": 6.988217305537058e-05, "loss": 0.6768, "step": 227 }, { "epoch": 1.546417973717677, "grad_norm": 1.3535468021784238, "learning_rate": 6.97554569338139e-05, "loss": 0.6959, "step": 228 }, { "epoch": 1.5532005086901228, "grad_norm": 1.1587176301112725, "learning_rate": 6.962806866778284e-05, "loss": 0.6821, "step": 229 }, { "epoch": 1.559983043662569, "grad_norm": 1.6895118703555423, "learning_rate": 6.950001113484432e-05, "loss": 0.6832, "step": 230 }, { "epoch": 1.5667655786350148, "grad_norm": 1.3111534840956083, "learning_rate": 6.937128722768333e-05, "loss": 0.6747, "step": 231 }, { "epoch": 1.573548113607461, "grad_norm": 1.2243130191258123, "learning_rate": 6.924189985403753e-05, "loss": 0.69, "step": 232 }, { "epoch": 1.5803306485799067, "grad_norm": 1.4287675639720405, "learning_rate": 6.91118519366316e-05, "loss": 0.6801, "step": 233 }, { "epoch": 1.5871131835523526, "grad_norm": 0.9267438225948489, "learning_rate": 6.898114641311122e-05, "loss": 0.6845, "step": 234 }, { "epoch": 1.5938957185247986, "grad_norm": 1.3352877429277719, "learning_rate": 6.884978623597665e-05, "loss": 0.6993, "step": 235 }, { "epoch": 1.6006782534972446, "grad_norm": 1.334254924822497, "learning_rate": 6.871777437251617e-05, "loss": 0.6817, "step": 236 }, { "epoch": 1.6074607884696905, "grad_norm": 1.2977369297336303, "learning_rate": 6.858511380473887e-05, "loss": 0.6856, "step": 237 }, { "epoch": 1.6142433234421365, "grad_norm": 1.1900674515806826, "learning_rate": 6.845180752930749e-05, "loss": 0.6947, "step": 238 }, { "epoch": 1.6210258584145825, "grad_norm": 2.7317108581757057, "learning_rate": 6.831785855747054e-05, "loss": 0.6985, "step": 239 }, { "epoch": 1.6278083933870284, "grad_norm": 1.2267690972200642, "learning_rate": 6.81832699149944e-05, "loss": 0.6773, "step": 240 }, { "epoch": 1.6345909283594744, "grad_norm": 3.8760882746190255, "learning_rate": 6.804804464209498e-05, "loss": 0.6895, "step": 241 }, { "epoch": 1.6413734633319204, "grad_norm": 3.5268150192285157, "learning_rate": 6.791218579336891e-05, "loss": 0.6994, "step": 242 }, { "epoch": 1.6481559983043663, "grad_norm": 1.7590143767536321, "learning_rate": 6.777569643772472e-05, "loss": 0.683, "step": 243 }, { "epoch": 1.654938533276812, "grad_norm": 1.7753106738494688, "learning_rate": 6.763857965831337e-05, "loss": 0.6811, "step": 244 }, { "epoch": 1.6617210682492582, "grad_norm": 1.7854517795067268, "learning_rate": 6.750083855245869e-05, "loss": 0.6818, "step": 245 }, { "epoch": 1.668503603221704, "grad_norm": 1.2937361964461016, "learning_rate": 6.736247623158738e-05, "loss": 0.6825, "step": 246 }, { "epoch": 1.6752861381941502, "grad_norm": 1.2452039091600213, "learning_rate": 6.722349582115872e-05, "loss": 0.6779, "step": 247 }, { "epoch": 1.682068673166596, "grad_norm": 1.6654925410640917, "learning_rate": 6.708390046059402e-05, "loss": 0.6897, "step": 248 }, { "epoch": 1.688851208139042, "grad_norm": 1.1695940537064082, "learning_rate": 6.694369330320563e-05, "loss": 0.6894, "step": 249 }, { "epoch": 1.6956337431114878, "grad_norm": 1.790433077118171, "learning_rate": 6.680287751612576e-05, "loss": 0.6965, "step": 250 }, { "epoch": 1.702416278083934, "grad_norm": 1.3940506254821923, "learning_rate": 6.666145628023495e-05, "loss": 0.6883, "step": 251 }, { "epoch": 1.7091988130563798, "grad_norm": 1.5788853280407766, "learning_rate": 6.651943279009015e-05, "loss": 0.674, "step": 252 }, { "epoch": 1.7159813480288257, "grad_norm": 1.135684283004857, "learning_rate": 6.637681025385261e-05, "loss": 0.6944, "step": 253 }, { "epoch": 1.7227638830012717, "grad_norm": 1.4839512076861858, "learning_rate": 6.62335918932154e-05, "loss": 0.6752, "step": 254 }, { "epoch": 1.7295464179737177, "grad_norm": 1.3150087814531235, "learning_rate": 6.608978094333068e-05, "loss": 0.6826, "step": 255 }, { "epoch": 1.7363289529461636, "grad_norm": 1.0720984172971046, "learning_rate": 6.594538065273652e-05, "loss": 0.683, "step": 256 }, { "epoch": 1.7431114879186096, "grad_norm": 1.46543732407952, "learning_rate": 6.58003942832836e-05, "loss": 0.6796, "step": 257 }, { "epoch": 1.7498940228910556, "grad_norm": 0.9292343800735161, "learning_rate": 6.56548251100615e-05, "loss": 0.6686, "step": 258 }, { "epoch": 1.7566765578635015, "grad_norm": 1.5664491391186621, "learning_rate": 6.550867642132474e-05, "loss": 0.6912, "step": 259 }, { "epoch": 1.7634590928359475, "grad_norm": 0.9821764593726072, "learning_rate": 6.536195151841847e-05, "loss": 0.6702, "step": 260 }, { "epoch": 1.7702416278083932, "grad_norm": 1.4298526249936745, "learning_rate": 6.521465371570393e-05, "loss": 0.6871, "step": 261 }, { "epoch": 1.7770241627808394, "grad_norm": 1.2452682220878895, "learning_rate": 6.506678634048353e-05, "loss": 0.6677, "step": 262 }, { "epoch": 1.7838066977532852, "grad_norm": 1.4920349233263195, "learning_rate": 6.491835273292574e-05, "loss": 0.6809, "step": 263 }, { "epoch": 1.7905892327257313, "grad_norm": 1.212833170547237, "learning_rate": 6.476935624598966e-05, "loss": 0.6809, "step": 264 }, { "epoch": 1.797371767698177, "grad_norm": 1.250429726318394, "learning_rate": 6.461980024534918e-05, "loss": 0.674, "step": 265 }, { "epoch": 1.8041543026706233, "grad_norm": 1.0681520589021245, "learning_rate": 6.446968810931707e-05, "loss": 0.6767, "step": 266 }, { "epoch": 1.810936837643069, "grad_norm": 1.470191760394408, "learning_rate": 6.431902322876855e-05, "loss": 0.6737, "step": 267 }, { "epoch": 1.8177193726155152, "grad_norm": 1.1749343585137977, "learning_rate": 6.416780900706484e-05, "loss": 0.683, "step": 268 }, { "epoch": 1.824501907587961, "grad_norm": 1.1727141312870375, "learning_rate": 6.401604885997614e-05, "loss": 0.6808, "step": 269 }, { "epoch": 1.831284442560407, "grad_norm": 1.3495790960778822, "learning_rate": 6.386374621560455e-05, "loss": 0.6847, "step": 270 }, { "epoch": 1.8380669775328529, "grad_norm": 1.075708125802829, "learning_rate": 6.37109045143066e-05, "loss": 0.676, "step": 271 }, { "epoch": 1.8448495125052988, "grad_norm": 1.1555696102312405, "learning_rate": 6.355752720861559e-05, "loss": 0.6661, "step": 272 }, { "epoch": 1.8516320474777448, "grad_norm": 1.383594134484124, "learning_rate": 6.340361776316349e-05, "loss": 0.6784, "step": 273 }, { "epoch": 1.8584145824501908, "grad_norm": 1.063837234136454, "learning_rate": 6.324917965460279e-05, "loss": 0.6781, "step": 274 }, { "epoch": 1.8651971174226367, "grad_norm": 1.173609581411256, "learning_rate": 6.309421637152794e-05, "loss": 0.6682, "step": 275 }, { "epoch": 1.8719796523950827, "grad_norm": 1.0658200919149257, "learning_rate": 6.29387314143965e-05, "loss": 0.6878, "step": 276 }, { "epoch": 1.8787621873675286, "grad_norm": 1.7400399121218273, "learning_rate": 6.278272829545011e-05, "loss": 0.6706, "step": 277 }, { "epoch": 1.8855447223399746, "grad_norm": 0.7229472871914281, "learning_rate": 6.26262105386351e-05, "loss": 0.6847, "step": 278 }, { "epoch": 1.8923272573124206, "grad_norm": 1.0742091230131698, "learning_rate": 6.246918167952304e-05, "loss": 0.6756, "step": 279 }, { "epoch": 1.8991097922848663, "grad_norm": 2.3846707594542456, "learning_rate": 6.231164526523063e-05, "loss": 0.6776, "step": 280 }, { "epoch": 1.9058923272573125, "grad_norm": 1.181568933400366, "learning_rate": 6.21536048543398e-05, "loss": 0.686, "step": 281 }, { "epoch": 1.9126748622297582, "grad_norm": 3.197777220254654, "learning_rate": 6.199506401681718e-05, "loss": 0.7007, "step": 282 }, { "epoch": 1.9194573972022044, "grad_norm": 2.538862401423035, "learning_rate": 6.183602633393352e-05, "loss": 0.7031, "step": 283 }, { "epoch": 1.9262399321746502, "grad_norm": 1.8353914421208335, "learning_rate": 6.167649539818284e-05, "loss": 0.6663, "step": 284 }, { "epoch": 1.9330224671470964, "grad_norm": 1.8245299177531569, "learning_rate": 6.151647481320114e-05, "loss": 0.7011, "step": 285 }, { "epoch": 1.939805002119542, "grad_norm": 2.284062594732862, "learning_rate": 6.135596819368512e-05, "loss": 0.701, "step": 286 }, { "epoch": 1.9465875370919883, "grad_norm": 1.5939451814656143, "learning_rate": 6.119497916531053e-05, "loss": 0.6898, "step": 287 }, { "epoch": 1.953370072064434, "grad_norm": 2.382128196264195, "learning_rate": 6.103351136465014e-05, "loss": 0.689, "step": 288 }, { "epoch": 1.96015260703688, "grad_norm": 2.255809021386828, "learning_rate": 6.0871568439091743e-05, "loss": 0.689, "step": 289 }, { "epoch": 1.966935142009326, "grad_norm": 1.192524289628154, "learning_rate": 6.070915404675571e-05, "loss": 0.6709, "step": 290 }, { "epoch": 1.973717676981772, "grad_norm": 1.2397061535202811, "learning_rate": 6.0546271856412306e-05, "loss": 0.687, "step": 291 }, { "epoch": 1.9805002119542179, "grad_norm": 0.9812283662135952, "learning_rate": 6.03829255473989e-05, "loss": 0.677, "step": 292 }, { "epoch": 1.9872827469266638, "grad_norm": 1.410310313066498, "learning_rate": 6.0219118809536794e-05, "loss": 0.6882, "step": 293 }, { "epoch": 1.9940652818991098, "grad_norm": 0.7865238012992424, "learning_rate": 6.0054855343047914e-05, "loss": 0.6769, "step": 294 }, { "epoch": 2.0008478168715556, "grad_norm": 1.9901508305957583, "learning_rate": 5.989013885847117e-05, "loss": 1.0693, "step": 295 }, { "epoch": 2.0076303518440017, "grad_norm": 1.3663813100431552, "learning_rate": 5.972497307657869e-05, "loss": 0.6673, "step": 296 }, { "epoch": 2.0144128868164475, "grad_norm": 1.1381480345523374, "learning_rate": 5.955936172829179e-05, "loss": 0.6676, "step": 297 }, { "epoch": 2.0211954217888937, "grad_norm": 1.496250854193103, "learning_rate": 5.939330855459661e-05, "loss": 0.6551, "step": 298 }, { "epoch": 2.0279779567613394, "grad_norm": 0.9716458462062207, "learning_rate": 5.922681730645968e-05, "loss": 0.6723, "step": 299 }, { "epoch": 2.0347604917337856, "grad_norm": 1.501595368602735, "learning_rate": 5.905989174474319e-05, "loss": 0.6509, "step": 300 }, { "epoch": 2.0415430267062313, "grad_norm": 0.849974962385986, "learning_rate": 5.889253564011999e-05, "loss": 0.6577, "step": 301 }, { "epoch": 2.0483255616786775, "grad_norm": 1.418902591957157, "learning_rate": 5.872475277298847e-05, "loss": 0.6665, "step": 302 }, { "epoch": 2.0551080966511233, "grad_norm": 0.8446815446152741, "learning_rate": 5.855654693338711e-05, "loss": 0.6556, "step": 303 }, { "epoch": 2.0618906316235694, "grad_norm": 1.3364133345670484, "learning_rate": 5.838792192090889e-05, "loss": 0.6604, "step": 304 }, { "epoch": 2.068673166596015, "grad_norm": 0.8691215260428371, "learning_rate": 5.821888154461549e-05, "loss": 0.6593, "step": 305 }, { "epoch": 2.0754557015684614, "grad_norm": 1.2207875769882877, "learning_rate": 5.8049429622951194e-05, "loss": 0.6637, "step": 306 }, { "epoch": 2.082238236540907, "grad_norm": 0.7775512997267359, "learning_rate": 5.7879569983656694e-05, "loss": 0.6433, "step": 307 }, { "epoch": 2.0890207715133533, "grad_norm": 1.2237758983601073, "learning_rate": 5.770930646368257e-05, "loss": 0.6681, "step": 308 }, { "epoch": 2.095803306485799, "grad_norm": 0.6935518063946065, "learning_rate": 5.7538642909102654e-05, "loss": 0.6571, "step": 309 }, { "epoch": 2.1025858414582452, "grad_norm": 1.0509844665628039, "learning_rate": 5.736758317502714e-05, "loss": 0.6462, "step": 310 }, { "epoch": 2.109368376430691, "grad_norm": 0.6501336655556583, "learning_rate": 5.719613112551546e-05, "loss": 0.642, "step": 311 }, { "epoch": 2.1161509114031367, "grad_norm": 0.791518345366507, "learning_rate": 5.702429063348912e-05, "loss": 0.6523, "step": 312 }, { "epoch": 2.122933446375583, "grad_norm": 0.7216839703720191, "learning_rate": 5.685206558064407e-05, "loss": 0.6485, "step": 313 }, { "epoch": 2.1297159813480286, "grad_norm": 1.3490427513390442, "learning_rate": 5.667945985736308e-05, "loss": 0.6622, "step": 314 }, { "epoch": 2.136498516320475, "grad_norm": 1.036563799894598, "learning_rate": 5.6506477362627926e-05, "loss": 0.6598, "step": 315 }, { "epoch": 2.1432810512929206, "grad_norm": 0.8799949127384937, "learning_rate": 5.6333122003931186e-05, "loss": 0.6631, "step": 316 }, { "epoch": 2.1500635862653668, "grad_norm": 0.9027804053204243, "learning_rate": 5.615939769718809e-05, "loss": 0.6543, "step": 317 }, { "epoch": 2.1568461212378125, "grad_norm": 1.08723379469273, "learning_rate": 5.5985308366647985e-05, "loss": 0.6453, "step": 318 }, { "epoch": 2.1636286562102587, "grad_norm": 1.198788894798635, "learning_rate": 5.5810857944805744e-05, "loss": 0.6466, "step": 319 }, { "epoch": 2.1704111911827044, "grad_norm": 1.2287340016478474, "learning_rate": 5.5636050372312896e-05, "loss": 0.6776, "step": 320 }, { "epoch": 2.1771937261551506, "grad_norm": 0.676307375412045, "learning_rate": 5.546088959788862e-05, "loss": 0.6517, "step": 321 }, { "epoch": 2.1839762611275964, "grad_norm": 0.8023647253320252, "learning_rate": 5.528537957823061e-05, "loss": 0.6565, "step": 322 }, { "epoch": 2.1907587961000425, "grad_norm": 1.4391382554036933, "learning_rate": 5.510952427792559e-05, "loss": 0.6587, "step": 323 }, { "epoch": 2.1975413310724883, "grad_norm": 0.7633580340639281, "learning_rate": 5.493332766935981e-05, "loss": 0.6433, "step": 324 }, { "epoch": 2.2043238660449345, "grad_norm": 1.0303353819004128, "learning_rate": 5.4756793732629335e-05, "loss": 0.6578, "step": 325 }, { "epoch": 2.21110640101738, "grad_norm": 1.3526690779961053, "learning_rate": 5.45799264554501e-05, "loss": 0.6499, "step": 326 }, { "epoch": 2.2178889359898264, "grad_norm": 0.5228459526464206, "learning_rate": 5.440272983306789e-05, "loss": 0.6394, "step": 327 }, { "epoch": 2.224671470962272, "grad_norm": 1.4252351086818837, "learning_rate": 5.4225207868167994e-05, "loss": 0.6554, "step": 328 }, { "epoch": 2.231454005934718, "grad_norm": 0.6950731562608328, "learning_rate": 5.404736457078489e-05, "loss": 0.6482, "step": 329 }, { "epoch": 2.238236540907164, "grad_norm": 0.7644651875362564, "learning_rate": 5.38692039582116e-05, "loss": 0.6503, "step": 330 }, { "epoch": 2.24501907587961, "grad_norm": 0.9023571857971306, "learning_rate": 5.3690730054908985e-05, "loss": 0.6468, "step": 331 }, { "epoch": 2.251801610852056, "grad_norm": 0.9107802448508358, "learning_rate": 5.3511946892414775e-05, "loss": 0.6518, "step": 332 }, { "epoch": 2.2585841458245017, "grad_norm": 1.1127754837574242, "learning_rate": 5.333285850925256e-05, "loss": 0.6565, "step": 333 }, { "epoch": 2.265366680796948, "grad_norm": 0.8282253306938299, "learning_rate": 5.315346895084056e-05, "loss": 0.6564, "step": 334 }, { "epoch": 2.2721492157693937, "grad_norm": 1.14130477616649, "learning_rate": 5.297378226940019e-05, "loss": 0.6516, "step": 335 }, { "epoch": 2.27893175074184, "grad_norm": 1.0103971093744055, "learning_rate": 5.279380252386461e-05, "loss": 0.6626, "step": 336 }, { "epoch": 2.2857142857142856, "grad_norm": 0.9536394369627204, "learning_rate": 5.2613533779786945e-05, "loss": 0.6663, "step": 337 }, { "epoch": 2.2924968206867318, "grad_norm": 0.942528066117481, "learning_rate": 5.243298010924852e-05, "loss": 0.6408, "step": 338 }, { "epoch": 2.2992793556591775, "grad_norm": 1.2117876187133245, "learning_rate": 5.225214559076683e-05, "loss": 0.6435, "step": 339 }, { "epoch": 2.3060618906316237, "grad_norm": 0.8633776831955887, "learning_rate": 5.207103430920345e-05, "loss": 0.6395, "step": 340 }, { "epoch": 2.3128444256040694, "grad_norm": 0.727531307349217, "learning_rate": 5.1889650355671725e-05, "loss": 0.654, "step": 341 }, { "epoch": 2.3196269605765156, "grad_norm": 0.669365998637684, "learning_rate": 5.1707997827444394e-05, "loss": 0.6546, "step": 342 }, { "epoch": 2.3264094955489614, "grad_norm": 0.8312703348077676, "learning_rate": 5.152608082786098e-05, "loss": 0.6646, "step": 343 }, { "epoch": 2.3331920305214076, "grad_norm": 1.0187489492222126, "learning_rate": 5.1343903466235174e-05, "loss": 0.648, "step": 344 }, { "epoch": 2.3399745654938533, "grad_norm": 1.034867027111884, "learning_rate": 5.116146985776194e-05, "loss": 0.6593, "step": 345 }, { "epoch": 2.3467571004662995, "grad_norm": 1.1577447007405788, "learning_rate": 5.0978784123424617e-05, "loss": 0.6516, "step": 346 }, { "epoch": 2.3535396354387452, "grad_norm": 1.0011140671455747, "learning_rate": 5.0795850389901784e-05, "loss": 0.6638, "step": 347 }, { "epoch": 2.360322170411191, "grad_norm": 0.9415943625257757, "learning_rate": 5.061267278947408e-05, "loss": 0.6457, "step": 348 }, { "epoch": 2.367104705383637, "grad_norm": 0.6919138650974885, "learning_rate": 5.042925545993079e-05, "loss": 0.6584, "step": 349 }, { "epoch": 2.373887240356083, "grad_norm": 0.7266901542517143, "learning_rate": 5.02456025444765e-05, "loss": 0.6348, "step": 350 }, { "epoch": 2.380669775328529, "grad_norm": 0.6635241827908179, "learning_rate": 5.0061718191637394e-05, "loss": 0.6574, "step": 351 }, { "epoch": 2.387452310300975, "grad_norm": 0.5653141179051638, "learning_rate": 4.987760655516757e-05, "loss": 0.6502, "step": 352 }, { "epoch": 2.394234845273421, "grad_norm": 0.7007173012785425, "learning_rate": 4.9693271793955255e-05, "loss": 0.6468, "step": 353 }, { "epoch": 2.4010173802458668, "grad_norm": 0.5821543619857267, "learning_rate": 4.95087180719288e-05, "loss": 0.6486, "step": 354 }, { "epoch": 2.407799915218313, "grad_norm": 0.4265984966323706, "learning_rate": 4.9323949557962684e-05, "loss": 0.6477, "step": 355 }, { "epoch": 2.4145824501907587, "grad_norm": 0.6454773425594047, "learning_rate": 4.913897042578327e-05, "loss": 0.6479, "step": 356 }, { "epoch": 2.421364985163205, "grad_norm": 0.615481868155692, "learning_rate": 4.8953784853874624e-05, "loss": 0.6605, "step": 357 }, { "epoch": 2.4281475201356506, "grad_norm": 0.4923342474878763, "learning_rate": 4.8768397025383996e-05, "loss": 0.6414, "step": 358 }, { "epoch": 2.434930055108097, "grad_norm": 0.4236872576222672, "learning_rate": 4.858281112802745e-05, "loss": 0.6526, "step": 359 }, { "epoch": 2.4417125900805425, "grad_norm": 0.4376255004390753, "learning_rate": 4.83970313539952e-05, "loss": 0.6503, "step": 360 }, { "epoch": 2.4484951250529887, "grad_norm": 0.7271305316785697, "learning_rate": 4.821106189985693e-05, "loss": 0.6447, "step": 361 }, { "epoch": 2.4552776600254345, "grad_norm": 1.001429290369445, "learning_rate": 4.8024906966467e-05, "loss": 0.6566, "step": 362 }, { "epoch": 2.4620601949978806, "grad_norm": 1.168304354121221, "learning_rate": 4.783857075886956e-05, "loss": 0.6614, "step": 363 }, { "epoch": 2.4688427299703264, "grad_norm": 0.8021141929873027, "learning_rate": 4.7652057486203525e-05, "loss": 0.6446, "step": 364 }, { "epoch": 2.4756252649427726, "grad_norm": 0.7075159940006046, "learning_rate": 4.746537136160757e-05, "loss": 0.6563, "step": 365 }, { "epoch": 2.4824077999152183, "grad_norm": 0.7183111983007646, "learning_rate": 4.727851660212487e-05, "loss": 0.6502, "step": 366 }, { "epoch": 2.489190334887664, "grad_norm": 0.755055422137936, "learning_rate": 4.709149742860792e-05, "loss": 0.6424, "step": 367 }, { "epoch": 2.4959728698601102, "grad_norm": 0.7253802127810576, "learning_rate": 4.690431806562311e-05, "loss": 0.6496, "step": 368 }, { "epoch": 2.5027554048325564, "grad_norm": 0.5125324963207099, "learning_rate": 4.6716982741355386e-05, "loss": 0.6533, "step": 369 }, { "epoch": 2.509537939805002, "grad_norm": 0.5597998347086172, "learning_rate": 4.652949568751267e-05, "loss": 0.6402, "step": 370 }, { "epoch": 2.516320474777448, "grad_norm": 0.7015467668995325, "learning_rate": 4.63418611392303e-05, "loss": 0.6482, "step": 371 }, { "epoch": 2.523103009749894, "grad_norm": 0.511973221694181, "learning_rate": 4.615408333497538e-05, "loss": 0.6522, "step": 372 }, { "epoch": 2.52988554472234, "grad_norm": 0.4401280437938751, "learning_rate": 4.5966166516450985e-05, "loss": 0.6573, "step": 373 }, { "epoch": 2.536668079694786, "grad_norm": 0.614469648497891, "learning_rate": 4.577811492850039e-05, "loss": 0.6598, "step": 374 }, { "epoch": 2.5434506146672318, "grad_norm": 0.7173941689992674, "learning_rate": 4.558993281901116e-05, "loss": 0.6465, "step": 375 }, { "epoch": 2.550233149639678, "grad_norm": 0.6690409730578648, "learning_rate": 4.540162443881922e-05, "loss": 0.6488, "step": 376 }, { "epoch": 2.5570156846121237, "grad_norm": 0.5848162533642309, "learning_rate": 4.5213194041612824e-05, "loss": 0.648, "step": 377 }, { "epoch": 2.56379821958457, "grad_norm": 0.549981546670745, "learning_rate": 4.5024645883836426e-05, "loss": 0.6579, "step": 378 }, { "epoch": 2.5705807545570156, "grad_norm": 0.556140263499365, "learning_rate": 4.4835984224594586e-05, "loss": 0.6621, "step": 379 }, { "epoch": 2.577363289529462, "grad_norm": 0.5195224064491745, "learning_rate": 4.464721332555577e-05, "loss": 0.6512, "step": 380 }, { "epoch": 2.5841458245019076, "grad_norm": 0.38259318048535124, "learning_rate": 4.445833745085602e-05, "loss": 0.6492, "step": 381 }, { "epoch": 2.5909283594743533, "grad_norm": 0.3534037726943067, "learning_rate": 4.4269360867002675e-05, "loss": 0.6589, "step": 382 }, { "epoch": 2.5977108944467995, "grad_norm": 0.315508715276319, "learning_rate": 4.408028784277799e-05, "loss": 0.6493, "step": 383 }, { "epoch": 2.6044934294192457, "grad_norm": 0.3298656413250479, "learning_rate": 4.389112264914273e-05, "loss": 0.6589, "step": 384 }, { "epoch": 2.6112759643916914, "grad_norm": 0.3016047037119661, "learning_rate": 4.370186955913962e-05, "loss": 0.6561, "step": 385 }, { "epoch": 2.618058499364137, "grad_norm": 0.34845242594318765, "learning_rate": 4.351253284779692e-05, "loss": 0.6552, "step": 386 }, { "epoch": 2.6248410343365833, "grad_norm": 0.38736846664169255, "learning_rate": 4.332311679203177e-05, "loss": 0.6364, "step": 387 }, { "epoch": 2.6316235693090295, "grad_norm": 0.3951534285172612, "learning_rate": 4.313362567055367e-05, "loss": 0.6502, "step": 388 }, { "epoch": 2.6384061042814753, "grad_norm": 0.4657313097001591, "learning_rate": 4.294406376376771e-05, "loss": 0.6561, "step": 389 }, { "epoch": 2.645188639253921, "grad_norm": 0.4131171546046392, "learning_rate": 4.2754435353677985e-05, "loss": 0.6567, "step": 390 }, { "epoch": 2.651971174226367, "grad_norm": 0.5305329556333895, "learning_rate": 4.2564744723790835e-05, "loss": 0.6568, "step": 391 }, { "epoch": 2.658753709198813, "grad_norm": 0.7002749870897172, "learning_rate": 4.237499615901805e-05, "loss": 0.6552, "step": 392 }, { "epoch": 2.665536244171259, "grad_norm": 0.7839360975679, "learning_rate": 4.218519394558013e-05, "loss": 0.6333, "step": 393 }, { "epoch": 2.672318779143705, "grad_norm": 0.7621551938800204, "learning_rate": 4.199534237090943e-05, "loss": 0.6501, "step": 394 }, { "epoch": 2.679101314116151, "grad_norm": 0.6652202397867919, "learning_rate": 4.1805445723553346e-05, "loss": 0.648, "step": 395 }, { "epoch": 2.685883849088597, "grad_norm": 0.4811941000709009, "learning_rate": 4.1615508293077394e-05, "loss": 0.6544, "step": 396 }, { "epoch": 2.692666384061043, "grad_norm": 0.3387872066498026, "learning_rate": 4.142553436996834e-05, "loss": 0.6558, "step": 397 }, { "epoch": 2.6994489190334887, "grad_norm": 0.28441864482912343, "learning_rate": 4.12355282455373e-05, "loss": 0.6354, "step": 398 }, { "epoch": 2.706231454005935, "grad_norm": 0.285244776358457, "learning_rate": 4.1045494211822756e-05, "loss": 0.6493, "step": 399 }, { "epoch": 2.7130139889783806, "grad_norm": 0.3573803066995761, "learning_rate": 4.085543656149365e-05, "loss": 0.6397, "step": 400 }, { "epoch": 2.7197965239508264, "grad_norm": 0.4278142282008419, "learning_rate": 4.0665359587752394e-05, "loss": 0.6537, "step": 401 }, { "epoch": 2.7265790589232726, "grad_norm": 0.49944968397215816, "learning_rate": 4.04752675842379e-05, "loss": 0.656, "step": 402 }, { "epoch": 2.7333615938957188, "grad_norm": 0.4642912024112797, "learning_rate": 4.028516484492857e-05, "loss": 0.6424, "step": 403 }, { "epoch": 2.7401441288681645, "grad_norm": 0.40098112979297407, "learning_rate": 4.009505566404535e-05, "loss": 0.6377, "step": 404 }, { "epoch": 2.7469266638406102, "grad_norm": 0.3708879898016676, "learning_rate": 3.990494433595466e-05, "loss": 0.6518, "step": 405 }, { "epoch": 2.7537091988130564, "grad_norm": 0.40008123571174364, "learning_rate": 3.9714835155071435e-05, "loss": 0.6536, "step": 406 }, { "epoch": 2.760491733785502, "grad_norm": 0.29605943254539613, "learning_rate": 3.952473241576211e-05, "loss": 0.6583, "step": 407 }, { "epoch": 2.7672742687579484, "grad_norm": 0.3406525853896961, "learning_rate": 3.933464041224761e-05, "loss": 0.6545, "step": 408 }, { "epoch": 2.774056803730394, "grad_norm": 0.4722405867001987, "learning_rate": 3.914456343850637e-05, "loss": 0.6513, "step": 409 }, { "epoch": 2.7808393387028403, "grad_norm": 0.5263658831108335, "learning_rate": 3.895450578817727e-05, "loss": 0.6548, "step": 410 }, { "epoch": 2.787621873675286, "grad_norm": 0.5203524120767699, "learning_rate": 3.8764471754462714e-05, "loss": 0.654, "step": 411 }, { "epoch": 2.794404408647732, "grad_norm": 0.3902260359882798, "learning_rate": 3.857446563003167e-05, "loss": 0.651, "step": 412 }, { "epoch": 2.801186943620178, "grad_norm": 0.3931366323184314, "learning_rate": 3.838449170692262e-05, "loss": 0.6418, "step": 413 }, { "epoch": 2.807969478592624, "grad_norm": 0.4496897165040769, "learning_rate": 3.819455427644666e-05, "loss": 0.6566, "step": 414 }, { "epoch": 2.81475201356507, "grad_norm": 0.4843275111384109, "learning_rate": 3.800465762909057e-05, "loss": 0.6494, "step": 415 }, { "epoch": 2.821534548537516, "grad_norm": 0.4284479617853188, "learning_rate": 3.781480605441989e-05, "loss": 0.6504, "step": 416 }, { "epoch": 2.828317083509962, "grad_norm": 0.3505752717784012, "learning_rate": 3.7625003840981976e-05, "loss": 0.6515, "step": 417 }, { "epoch": 2.835099618482408, "grad_norm": 0.2739600638831855, "learning_rate": 3.743525527620918e-05, "loss": 0.6482, "step": 418 }, { "epoch": 2.8418821534548537, "grad_norm": 0.2821715826512274, "learning_rate": 3.724556464632203e-05, "loss": 0.6558, "step": 419 }, { "epoch": 2.8486646884272995, "grad_norm": 0.3001235929581994, "learning_rate": 3.7055936236232296e-05, "loss": 0.6283, "step": 420 }, { "epoch": 2.8554472233997457, "grad_norm": 0.34084577684820244, "learning_rate": 3.686637432944634e-05, "loss": 0.6524, "step": 421 }, { "epoch": 2.862229758372192, "grad_norm": 0.31354392380727714, "learning_rate": 3.6676883207968226e-05, "loss": 0.6407, "step": 422 }, { "epoch": 2.8690122933446376, "grad_norm": 0.29938503730372684, "learning_rate": 3.648746715220309e-05, "loss": 0.6431, "step": 423 }, { "epoch": 2.8757948283170833, "grad_norm": 0.292588496703024, "learning_rate": 3.6298130440860394e-05, "loss": 0.6457, "step": 424 }, { "epoch": 2.8825773632895295, "grad_norm": 0.34211158982130174, "learning_rate": 3.6108877350857296e-05, "loss": 0.6413, "step": 425 }, { "epoch": 2.8893598982619753, "grad_norm": 0.28453013315503256, "learning_rate": 3.5919712157222014e-05, "loss": 0.6493, "step": 426 }, { "epoch": 2.8961424332344214, "grad_norm": 0.25000620848294364, "learning_rate": 3.573063913299733e-05, "loss": 0.6471, "step": 427 }, { "epoch": 2.902924968206867, "grad_norm": 0.32984361513817445, "learning_rate": 3.554166254914399e-05, "loss": 0.6525, "step": 428 }, { "epoch": 2.9097075031793134, "grad_norm": 0.34282163750459993, "learning_rate": 3.535278667444423e-05, "loss": 0.6435, "step": 429 }, { "epoch": 2.916490038151759, "grad_norm": 0.3376932266610861, "learning_rate": 3.5164015775405414e-05, "loss": 0.6537, "step": 430 }, { "epoch": 2.9232725731242053, "grad_norm": 0.2736761009463742, "learning_rate": 3.4975354116163594e-05, "loss": 0.6528, "step": 431 }, { "epoch": 2.930055108096651, "grad_norm": 0.28401811879767763, "learning_rate": 3.478680595838719e-05, "loss": 0.6552, "step": 432 }, { "epoch": 2.9368376430690972, "grad_norm": 0.21207447278197997, "learning_rate": 3.4598375561180783e-05, "loss": 0.6477, "step": 433 }, { "epoch": 2.943620178041543, "grad_norm": 0.3273666184303206, "learning_rate": 3.441006718098885e-05, "loss": 0.6415, "step": 434 }, { "epoch": 2.9504027130139887, "grad_norm": 0.3922599048141651, "learning_rate": 3.422188507149962e-05, "loss": 0.66, "step": 435 }, { "epoch": 2.957185247986435, "grad_norm": 0.33853238933941526, "learning_rate": 3.403383348354902e-05, "loss": 0.6392, "step": 436 }, { "epoch": 2.963967782958881, "grad_norm": 0.2904409798403406, "learning_rate": 3.3845916665024626e-05, "loss": 0.6571, "step": 437 }, { "epoch": 2.970750317931327, "grad_norm": 0.25650378002806623, "learning_rate": 3.36581388607697e-05, "loss": 0.658, "step": 438 }, { "epoch": 2.9775328529037726, "grad_norm": 0.30516122417008495, "learning_rate": 3.347050431248735e-05, "loss": 0.6445, "step": 439 }, { "epoch": 2.9843153878762188, "grad_norm": 0.35193441095679195, "learning_rate": 3.3283017258644634e-05, "loss": 0.6514, "step": 440 }, { "epoch": 2.991097922848665, "grad_norm": 0.26810523993687996, "learning_rate": 3.30956819343769e-05, "loss": 0.6479, "step": 441 }, { "epoch": 2.9978804578211107, "grad_norm": 0.40136478388949387, "learning_rate": 3.290850257139209e-05, "loss": 0.7446, "step": 442 }, { "epoch": 3.0046629927935564, "grad_norm": 0.6219962321725231, "learning_rate": 3.2721483397875135e-05, "loss": 0.9011, "step": 443 }, { "epoch": 3.0114455277660026, "grad_norm": 0.6301524386828496, "learning_rate": 3.253462863839243e-05, "loss": 0.6098, "step": 444 }, { "epoch": 3.0182280627384483, "grad_norm": 0.5440271202067795, "learning_rate": 3.2347942513796475e-05, "loss": 0.6367, "step": 445 }, { "epoch": 3.0250105977108945, "grad_norm": 0.5386656826635444, "learning_rate": 3.216142924113046e-05, "loss": 0.6285, "step": 446 }, { "epoch": 3.0317931326833403, "grad_norm": 0.4074410048009017, "learning_rate": 3.1975093033533003e-05, "loss": 0.6332, "step": 447 }, { "epoch": 3.0385756676557865, "grad_norm": 0.25928533106144735, "learning_rate": 3.1788938100143086e-05, "loss": 0.6246, "step": 448 }, { "epoch": 3.045358202628232, "grad_norm": 0.445439070033727, "learning_rate": 3.160296864600482e-05, "loss": 0.6193, "step": 449 }, { "epoch": 3.0521407376006784, "grad_norm": 0.46469489323619156, "learning_rate": 3.1417188871972554e-05, "loss": 0.6264, "step": 450 }, { "epoch": 3.058923272573124, "grad_norm": 0.3272979337949089, "learning_rate": 3.123160297461601e-05, "loss": 0.631, "step": 451 }, { "epoch": 3.0657058075455703, "grad_norm": 0.3178424540888323, "learning_rate": 3.104621514612538e-05, "loss": 0.6344, "step": 452 }, { "epoch": 3.072488342518016, "grad_norm": 0.2754042730785633, "learning_rate": 3.086102957421672e-05, "loss": 0.6129, "step": 453 }, { "epoch": 3.0792708774904622, "grad_norm": 0.35442538475012997, "learning_rate": 3.0676050442037336e-05, "loss": 0.6269, "step": 454 }, { "epoch": 3.086053412462908, "grad_norm": 0.2797121425291907, "learning_rate": 3.0491281928071217e-05, "loss": 0.6221, "step": 455 }, { "epoch": 3.092835947435354, "grad_norm": 0.23765468818900978, "learning_rate": 3.030672820604476e-05, "loss": 0.6256, "step": 456 }, { "epoch": 3.0996184824078, "grad_norm": 0.26572260092618777, "learning_rate": 3.012239344483244e-05, "loss": 0.6177, "step": 457 }, { "epoch": 3.1064010173802457, "grad_norm": 0.33390446968478465, "learning_rate": 2.993828180836262e-05, "loss": 0.6113, "step": 458 }, { "epoch": 3.113183552352692, "grad_norm": 0.2666734723327277, "learning_rate": 2.9754397455523497e-05, "loss": 0.6297, "step": 459 }, { "epoch": 3.1199660873251376, "grad_norm": 0.29084625313932944, "learning_rate": 2.957074454006921e-05, "loss": 0.6196, "step": 460 }, { "epoch": 3.1267486222975838, "grad_norm": 0.19046989657100505, "learning_rate": 2.9387327210525942e-05, "loss": 0.6188, "step": 461 }, { "epoch": 3.1335311572700295, "grad_norm": 0.338123523854364, "learning_rate": 2.9204149610098223e-05, "loss": 0.6281, "step": 462 }, { "epoch": 3.1403136922424757, "grad_norm": 0.22618376500160342, "learning_rate": 2.902121587657539e-05, "loss": 0.6157, "step": 463 }, { "epoch": 3.1470962272149214, "grad_norm": 0.35386494541824426, "learning_rate": 2.8838530142238076e-05, "loss": 0.626, "step": 464 }, { "epoch": 3.1538787621873676, "grad_norm": 0.2840090224796156, "learning_rate": 2.865609653376484e-05, "loss": 0.6199, "step": 465 }, { "epoch": 3.1606612971598134, "grad_norm": 0.2835278774397419, "learning_rate": 2.8473919172139034e-05, "loss": 0.635, "step": 466 }, { "epoch": 3.1674438321322596, "grad_norm": 0.2932289544670752, "learning_rate": 2.8292002172555616e-05, "loss": 0.6181, "step": 467 }, { "epoch": 3.1742263671047053, "grad_norm": 0.21388735557547445, "learning_rate": 2.8110349644328275e-05, "loss": 0.6304, "step": 468 }, { "epoch": 3.1810089020771515, "grad_norm": 0.25484830423076227, "learning_rate": 2.7928965690796562e-05, "loss": 0.6351, "step": 469 }, { "epoch": 3.187791437049597, "grad_norm": 0.24843127490388153, "learning_rate": 2.7747854409233187e-05, "loss": 0.6301, "step": 470 }, { "epoch": 3.1945739720220434, "grad_norm": 0.2451683485321557, "learning_rate": 2.7567019890751493e-05, "loss": 0.6207, "step": 471 }, { "epoch": 3.201356506994489, "grad_norm": 0.3164960123899796, "learning_rate": 2.7386466220213065e-05, "loss": 0.6213, "step": 472 }, { "epoch": 3.2081390419669353, "grad_norm": 0.2697409403652239, "learning_rate": 2.7206197476135403e-05, "loss": 0.6212, "step": 473 }, { "epoch": 3.214921576939381, "grad_norm": 0.19669860705803766, "learning_rate": 2.7026217730599814e-05, "loss": 0.6284, "step": 474 }, { "epoch": 3.2217041119118273, "grad_norm": 0.23674235246224137, "learning_rate": 2.6846531049159454e-05, "loss": 0.627, "step": 475 }, { "epoch": 3.228486646884273, "grad_norm": 0.18982523269613027, "learning_rate": 2.6667141490747454e-05, "loss": 0.6236, "step": 476 }, { "epoch": 3.2352691818567187, "grad_norm": 0.20626545109123062, "learning_rate": 2.648805310758524e-05, "loss": 0.6242, "step": 477 }, { "epoch": 3.242051716829165, "grad_norm": 0.25273617130117126, "learning_rate": 2.6309269945091025e-05, "loss": 0.6135, "step": 478 }, { "epoch": 3.2488342518016107, "grad_norm": 0.19396420036714845, "learning_rate": 2.6130796041788404e-05, "loss": 0.6401, "step": 479 }, { "epoch": 3.255616786774057, "grad_norm": 0.26449583344475813, "learning_rate": 2.5952635429215117e-05, "loss": 0.6303, "step": 480 }, { "epoch": 3.2623993217465026, "grad_norm": 0.2880771386228117, "learning_rate": 2.5774792131832012e-05, "loss": 0.6213, "step": 481 }, { "epoch": 3.269181856718949, "grad_norm": 0.19301524096425413, "learning_rate": 2.559727016693212e-05, "loss": 0.6368, "step": 482 }, { "epoch": 3.2759643916913945, "grad_norm": 0.31002080783316865, "learning_rate": 2.54200735445499e-05, "loss": 0.641, "step": 483 }, { "epoch": 3.2827469266638407, "grad_norm": 0.24500950182263187, "learning_rate": 2.5243206267370685e-05, "loss": 0.6204, "step": 484 }, { "epoch": 3.2895294616362865, "grad_norm": 0.25623143489960465, "learning_rate": 2.506667233064021e-05, "loss": 0.6207, "step": 485 }, { "epoch": 3.2963119966087326, "grad_norm": 0.2918652621084026, "learning_rate": 2.489047572207442e-05, "loss": 0.6371, "step": 486 }, { "epoch": 3.3030945315811784, "grad_norm": 0.23337516028968874, "learning_rate": 2.4714620421769394e-05, "loss": 0.6285, "step": 487 }, { "epoch": 3.3098770665536246, "grad_norm": 0.23619697217842073, "learning_rate": 2.4539110402111383e-05, "loss": 0.631, "step": 488 }, { "epoch": 3.3166596015260703, "grad_norm": 0.2448512796334409, "learning_rate": 2.436394962768712e-05, "loss": 0.6276, "step": 489 }, { "epoch": 3.3234421364985165, "grad_norm": 0.1983649853769521, "learning_rate": 2.4189142055194276e-05, "loss": 0.6244, "step": 490 }, { "epoch": 3.3302246714709622, "grad_norm": 0.21190165986074258, "learning_rate": 2.4014691633352038e-05, "loss": 0.6101, "step": 491 }, { "epoch": 3.337007206443408, "grad_norm": 0.25804827992388285, "learning_rate": 2.3840602302811923e-05, "loss": 0.6135, "step": 492 }, { "epoch": 3.343789741415854, "grad_norm": 0.18856304220278483, "learning_rate": 2.3666877996068824e-05, "loss": 0.6288, "step": 493 }, { "epoch": 3.3505722763883004, "grad_norm": 0.3301373974007395, "learning_rate": 2.3493522637372087e-05, "loss": 0.6326, "step": 494 }, { "epoch": 3.357354811360746, "grad_norm": 0.30315622915732504, "learning_rate": 2.332054014263692e-05, "loss": 0.6259, "step": 495 }, { "epoch": 3.364137346333192, "grad_norm": 0.22646576230809926, "learning_rate": 2.3147934419355935e-05, "loss": 0.6245, "step": 496 }, { "epoch": 3.370919881305638, "grad_norm": 0.30590058038547674, "learning_rate": 2.2975709366510887e-05, "loss": 0.6316, "step": 497 }, { "epoch": 3.3777024162780838, "grad_norm": 0.26486089548607467, "learning_rate": 2.2803868874484536e-05, "loss": 0.6383, "step": 498 }, { "epoch": 3.38448495125053, "grad_norm": 0.209806274707419, "learning_rate": 2.2632416824972886e-05, "loss": 0.614, "step": 499 }, { "epoch": 3.3912674862229757, "grad_norm": 0.3076979703172019, "learning_rate": 2.2461357090897352e-05, "loss": 0.6317, "step": 500 }, { "epoch": 3.398050021195422, "grad_norm": 0.27851643544183735, "learning_rate": 2.2290693536317444e-05, "loss": 0.6226, "step": 501 }, { "epoch": 3.4048325561678676, "grad_norm": 0.26720054263757376, "learning_rate": 2.212043001634332e-05, "loss": 0.6306, "step": 502 }, { "epoch": 3.411615091140314, "grad_norm": 0.3264075780455984, "learning_rate": 2.1950570377048813e-05, "loss": 0.6144, "step": 503 }, { "epoch": 3.4183976261127595, "grad_norm": 0.24476904503391836, "learning_rate": 2.178111845538453e-05, "loss": 0.6317, "step": 504 }, { "epoch": 3.4251801610852057, "grad_norm": 0.3445192467182261, "learning_rate": 2.1612078079091125e-05, "loss": 0.6227, "step": 505 }, { "epoch": 3.4319626960576515, "grad_norm": 0.2248390312896755, "learning_rate": 2.144345306661291e-05, "loss": 0.6179, "step": 506 }, { "epoch": 3.4387452310300977, "grad_norm": 0.24580333907022087, "learning_rate": 2.1275247227011536e-05, "loss": 0.6218, "step": 507 }, { "epoch": 3.4455277660025434, "grad_norm": 0.2333810028522329, "learning_rate": 2.1107464359880022e-05, "loss": 0.6313, "step": 508 }, { "epoch": 3.4523103009749896, "grad_norm": 0.20038699343598237, "learning_rate": 2.0940108255256823e-05, "loss": 0.6309, "step": 509 }, { "epoch": 3.4590928359474353, "grad_norm": 0.19291801763604194, "learning_rate": 2.0773182693540325e-05, "loss": 0.6076, "step": 510 }, { "epoch": 3.465875370919881, "grad_norm": 0.21138321490252843, "learning_rate": 2.0606691445403398e-05, "loss": 0.6238, "step": 511 }, { "epoch": 3.4726579058923273, "grad_norm": 0.2081459000808386, "learning_rate": 2.0440638271708226e-05, "loss": 0.6135, "step": 512 }, { "epoch": 3.4794404408647734, "grad_norm": 0.21359018594893273, "learning_rate": 2.0275026923421315e-05, "loss": 0.6347, "step": 513 }, { "epoch": 3.486222975837219, "grad_norm": 0.262459399100271, "learning_rate": 2.0109861141528853e-05, "loss": 0.6289, "step": 514 }, { "epoch": 3.493005510809665, "grad_norm": 0.24248389999156902, "learning_rate": 1.9945144656952103e-05, "loss": 0.6128, "step": 515 }, { "epoch": 3.499788045782111, "grad_norm": 0.22920015918246003, "learning_rate": 1.9780881190463222e-05, "loss": 0.634, "step": 516 }, { "epoch": 3.506570580754557, "grad_norm": 0.2824421315404272, "learning_rate": 1.9617074452601113e-05, "loss": 0.6322, "step": 517 }, { "epoch": 3.513353115727003, "grad_norm": 0.19432350726010572, "learning_rate": 1.9453728143587697e-05, "loss": 0.6231, "step": 518 }, { "epoch": 3.520135650699449, "grad_norm": 0.2846932920722668, "learning_rate": 1.9290845953244307e-05, "loss": 0.625, "step": 519 }, { "epoch": 3.526918185671895, "grad_norm": 0.16416357169798437, "learning_rate": 1.9128431560908263e-05, "loss": 0.6191, "step": 520 }, { "epoch": 3.5337007206443407, "grad_norm": 0.2747654462569946, "learning_rate": 1.896648863534988e-05, "loss": 0.622, "step": 521 }, { "epoch": 3.540483255616787, "grad_norm": 0.15903407751400953, "learning_rate": 1.8805020834689487e-05, "loss": 0.6194, "step": 522 }, { "epoch": 3.5472657905892326, "grad_norm": 0.20365263598678185, "learning_rate": 1.864403180631489e-05, "loss": 0.6255, "step": 523 }, { "epoch": 3.554048325561679, "grad_norm": 0.16301528679144703, "learning_rate": 1.8483525186798875e-05, "loss": 0.6253, "step": 524 }, { "epoch": 3.5608308605341246, "grad_norm": 0.23098070191792286, "learning_rate": 1.832350460181717e-05, "loss": 0.62, "step": 525 }, { "epoch": 3.5676133955065703, "grad_norm": 0.16673057537934008, "learning_rate": 1.8163973666066473e-05, "loss": 0.629, "step": 526 }, { "epoch": 3.5743959304790165, "grad_norm": 0.19378211117923547, "learning_rate": 1.8004935983182835e-05, "loss": 0.6344, "step": 527 }, { "epoch": 3.5811784654514627, "grad_norm": 0.17643236928174555, "learning_rate": 1.784639514566021e-05, "loss": 0.6284, "step": 528 }, { "epoch": 3.5879610004239084, "grad_norm": 0.19805537148789543, "learning_rate": 1.7688354734769378e-05, "loss": 0.6245, "step": 529 }, { "epoch": 3.594743535396354, "grad_norm": 0.17444243321183842, "learning_rate": 1.753081832047697e-05, "loss": 0.6241, "step": 530 }, { "epoch": 3.6015260703688003, "grad_norm": 0.22629631276174578, "learning_rate": 1.7373789461364904e-05, "loss": 0.6296, "step": 531 }, { "epoch": 3.6083086053412465, "grad_norm": 0.2322655718660276, "learning_rate": 1.7217271704549913e-05, "loss": 0.6335, "step": 532 }, { "epoch": 3.6150911403136923, "grad_norm": 0.20701067571130669, "learning_rate": 1.7061268585603507e-05, "loss": 0.6307, "step": 533 }, { "epoch": 3.621873675286138, "grad_norm": 0.2206959741176191, "learning_rate": 1.6905783628472074e-05, "loss": 0.6347, "step": 534 }, { "epoch": 3.628656210258584, "grad_norm": 0.15861912044210094, "learning_rate": 1.6750820345397217e-05, "loss": 0.6265, "step": 535 }, { "epoch": 3.63543874523103, "grad_norm": 0.2001185331683702, "learning_rate": 1.659638223683653e-05, "loss": 0.6287, "step": 536 }, { "epoch": 3.642221280203476, "grad_norm": 0.1746493458635906, "learning_rate": 1.644247279138442e-05, "loss": 0.6361, "step": 537 }, { "epoch": 3.649003815175922, "grad_norm": 0.17095450468813203, "learning_rate": 1.628909548569339e-05, "loss": 0.6251, "step": 538 }, { "epoch": 3.655786350148368, "grad_norm": 0.15588769920603499, "learning_rate": 1.6136253784395455e-05, "loss": 0.629, "step": 539 }, { "epoch": 3.662568885120814, "grad_norm": 0.14759859028229927, "learning_rate": 1.5983951140023864e-05, "loss": 0.6174, "step": 540 }, { "epoch": 3.66935142009326, "grad_norm": 0.16662793387428904, "learning_rate": 1.583219099293516e-05, "loss": 0.6192, "step": 541 }, { "epoch": 3.6761339550657057, "grad_norm": 0.16325631134586066, "learning_rate": 1.568097677123146e-05, "loss": 0.6193, "step": 542 }, { "epoch": 3.682916490038152, "grad_norm": 0.1590144709587067, "learning_rate": 1.5530311890682946e-05, "loss": 0.6126, "step": 543 }, { "epoch": 3.6896990250105977, "grad_norm": 0.16971629601446409, "learning_rate": 1.5380199754650838e-05, "loss": 0.6204, "step": 544 }, { "epoch": 3.6964815599830434, "grad_norm": 0.17653934494907608, "learning_rate": 1.5230643754010355e-05, "loss": 0.634, "step": 545 }, { "epoch": 3.7032640949554896, "grad_norm": 0.1602687042337523, "learning_rate": 1.508164726707427e-05, "loss": 0.6132, "step": 546 }, { "epoch": 3.7100466299279358, "grad_norm": 0.1853250131338786, "learning_rate": 1.4933213659516485e-05, "loss": 0.6263, "step": 547 }, { "epoch": 3.7168291649003815, "grad_norm": 0.16166905381502983, "learning_rate": 1.4785346284296078e-05, "loss": 0.6158, "step": 548 }, { "epoch": 3.7236116998728273, "grad_norm": 0.19092582748564702, "learning_rate": 1.4638048481581537e-05, "loss": 0.6235, "step": 549 }, { "epoch": 3.7303942348452734, "grad_norm": 0.14148685680959427, "learning_rate": 1.4491323578675265e-05, "loss": 0.6137, "step": 550 }, { "epoch": 3.7371767698177196, "grad_norm": 0.18984864298487686, "learning_rate": 1.4345174889938514e-05, "loss": 0.6329, "step": 551 }, { "epoch": 3.7439593047901654, "grad_norm": 0.14327461316917306, "learning_rate": 1.4199605716716414e-05, "loss": 0.6272, "step": 552 }, { "epoch": 3.750741839762611, "grad_norm": 0.18666689301215467, "learning_rate": 1.4054619347263487e-05, "loss": 0.6363, "step": 553 }, { "epoch": 3.7575243747350573, "grad_norm": 0.14946115608698116, "learning_rate": 1.3910219056669335e-05, "loss": 0.6218, "step": 554 }, { "epoch": 3.764306909707503, "grad_norm": 0.17777432794193568, "learning_rate": 1.3766408106784601e-05, "loss": 0.6278, "step": 555 }, { "epoch": 3.771089444679949, "grad_norm": 0.16174854488646243, "learning_rate": 1.3623189746147398e-05, "loss": 0.6227, "step": 556 }, { "epoch": 3.777871979652395, "grad_norm": 0.15621831808343503, "learning_rate": 1.3480567209909863e-05, "loss": 0.6296, "step": 557 }, { "epoch": 3.784654514624841, "grad_norm": 0.17960424688042897, "learning_rate": 1.333854371976505e-05, "loss": 0.6175, "step": 558 }, { "epoch": 3.791437049597287, "grad_norm": 0.161165540194335, "learning_rate": 1.319712248387424e-05, "loss": 0.6259, "step": 559 }, { "epoch": 3.798219584569733, "grad_norm": 0.18195946037024635, "learning_rate": 1.3056306696794377e-05, "loss": 0.6259, "step": 560 }, { "epoch": 3.805002119542179, "grad_norm": 0.15299855099029444, "learning_rate": 1.2916099539405997e-05, "loss": 0.623, "step": 561 }, { "epoch": 3.811784654514625, "grad_norm": 0.19532294279356358, "learning_rate": 1.2776504178841291e-05, "loss": 0.629, "step": 562 }, { "epoch": 3.8185671894870707, "grad_norm": 0.1505952319707712, "learning_rate": 1.2637523768412633e-05, "loss": 0.6239, "step": 563 }, { "epoch": 3.8253497244595165, "grad_norm": 0.15785614162676143, "learning_rate": 1.249916144754133e-05, "loss": 0.6213, "step": 564 }, { "epoch": 3.8321322594319627, "grad_norm": 0.17665393013957048, "learning_rate": 1.2361420341686641e-05, "loss": 0.6205, "step": 565 }, { "epoch": 3.838914794404409, "grad_norm": 0.16420702224755918, "learning_rate": 1.22243035622753e-05, "loss": 0.6221, "step": 566 }, { "epoch": 3.8456973293768546, "grad_norm": 0.17607498450226258, "learning_rate": 1.2087814206631095e-05, "loss": 0.6212, "step": 567 }, { "epoch": 3.8524798643493003, "grad_norm": 0.19592268500165855, "learning_rate": 1.1951955357905033e-05, "loss": 0.6199, "step": 568 }, { "epoch": 3.8592623993217465, "grad_norm": 0.1402346539755977, "learning_rate": 1.1816730085005603e-05, "loss": 0.6198, "step": 569 }, { "epoch": 3.8660449342941927, "grad_norm": 0.18099131317071607, "learning_rate": 1.1682141442529473e-05, "loss": 0.6256, "step": 570 }, { "epoch": 3.8728274692666385, "grad_norm": 0.19545074620930133, "learning_rate": 1.1548192470692516e-05, "loss": 0.6157, "step": 571 }, { "epoch": 3.879610004239084, "grad_norm": 0.15514995332077897, "learning_rate": 1.1414886195261135e-05, "loss": 0.6262, "step": 572 }, { "epoch": 3.8863925392115304, "grad_norm": 0.1534732088259275, "learning_rate": 1.128222562748384e-05, "loss": 0.6218, "step": 573 }, { "epoch": 3.893175074183976, "grad_norm": 0.16016994591208966, "learning_rate": 1.1150213764023357e-05, "loss": 0.6264, "step": 574 }, { "epoch": 3.8999576091564223, "grad_norm": 0.1393505701166259, "learning_rate": 1.1018853586888794e-05, "loss": 0.6217, "step": 575 }, { "epoch": 3.906740144128868, "grad_norm": 0.13613614708020336, "learning_rate": 1.0888148063368411e-05, "loss": 0.6271, "step": 576 }, { "epoch": 3.9135226791013142, "grad_norm": 0.13989845925554537, "learning_rate": 1.075810014596248e-05, "loss": 0.6246, "step": 577 }, { "epoch": 3.92030521407376, "grad_norm": 0.14295575937313854, "learning_rate": 1.0628712772316682e-05, "loss": 0.6186, "step": 578 }, { "epoch": 3.927087749046206, "grad_norm": 0.14578310176151307, "learning_rate": 1.04999888651557e-05, "loss": 0.6237, "step": 579 }, { "epoch": 3.933870284018652, "grad_norm": 0.1418828430912816, "learning_rate": 1.0371931332217175e-05, "loss": 0.6247, "step": 580 }, { "epoch": 3.940652818991098, "grad_norm": 0.1325595548114853, "learning_rate": 1.0244543066186102e-05, "loss": 0.623, "step": 581 }, { "epoch": 3.947435353963544, "grad_norm": 0.1458164336431073, "learning_rate": 1.0117826944629425e-05, "loss": 0.6311, "step": 582 }, { "epoch": 3.9542178889359896, "grad_norm": 0.1369850394553834, "learning_rate": 9.991785829931024e-06, "loss": 0.6352, "step": 583 }, { "epoch": 3.9610004239084358, "grad_norm": 0.11861951638152013, "learning_rate": 9.866422569227133e-06, "loss": 0.616, "step": 584 }, { "epoch": 3.967782958880882, "grad_norm": 0.12198529312439654, "learning_rate": 9.74173999434192e-06, "loss": 0.6267, "step": 585 }, { "epoch": 3.9745654938533277, "grad_norm": 0.13835745132234695, "learning_rate": 9.61774092172362e-06, "loss": 0.6228, "step": 586 }, { "epoch": 3.9813480288257734, "grad_norm": 0.14469581512913757, "learning_rate": 9.494428152380868e-06, "loss": 0.6249, "step": 587 }, { "epoch": 3.9881305637982196, "grad_norm": 0.14538533970008014, "learning_rate": 9.371804471819401e-06, "loss": 0.6327, "step": 588 }, { "epoch": 3.994913098770666, "grad_norm": 0.12842111782896357, "learning_rate": 9.249872649979203e-06, "loss": 0.6303, "step": 589 }, { "epoch": 4.001695633743111, "grad_norm": 0.26900932028906, "learning_rate": 9.128635441171854e-06, "loss": 0.947, "step": 590 }, { "epoch": 4.008478168715557, "grad_norm": 0.1546620990829906, "learning_rate": 9.008095584018406e-06, "loss": 0.6138, "step": 591 }, { "epoch": 4.0152607036880035, "grad_norm": 0.15420921751444336, "learning_rate": 8.888255801387417e-06, "loss": 0.6084, "step": 592 }, { "epoch": 4.02204323866045, "grad_norm": 0.15369650428751042, "learning_rate": 8.76911880033354e-06, "loss": 0.6196, "step": 593 }, { "epoch": 4.028825773632895, "grad_norm": 0.1491321631121837, "learning_rate": 8.65068727203633e-06, "loss": 0.6193, "step": 594 }, { "epoch": 4.035608308605341, "grad_norm": 0.14296040181465852, "learning_rate": 8.53296389173941e-06, "loss": 0.5971, "step": 595 }, { "epoch": 4.042390843577787, "grad_norm": 0.17332216986843843, "learning_rate": 8.415951318690134e-06, "loss": 0.6085, "step": 596 }, { "epoch": 4.0491733785502335, "grad_norm": 0.1392839623319017, "learning_rate": 8.299652196079462e-06, "loss": 0.6001, "step": 597 }, { "epoch": 4.055955913522679, "grad_norm": 0.13152438443257536, "learning_rate": 8.184069150982217e-06, "loss": 0.6105, "step": 598 }, { "epoch": 4.062738448495125, "grad_norm": 0.1534567228015376, "learning_rate": 8.069204794297843e-06, "loss": 0.6055, "step": 599 }, { "epoch": 4.069520983467571, "grad_norm": 0.13360654544275674, "learning_rate": 7.955061720691315e-06, "loss": 0.6032, "step": 600 }, { "epoch": 4.076303518440017, "grad_norm": 0.12240252460009761, "learning_rate": 7.841642508534617e-06, "loss": 0.6123, "step": 601 }, { "epoch": 4.083086053412463, "grad_norm": 0.13532197093706563, "learning_rate": 7.728949719848464e-06, "loss": 0.6153, "step": 602 }, { "epoch": 4.089868588384909, "grad_norm": 0.14193336659405498, "learning_rate": 7.6169859002443954e-06, "loss": 0.6168, "step": 603 }, { "epoch": 4.096651123357355, "grad_norm": 0.12702143209359149, "learning_rate": 7.505753578867354e-06, "loss": 0.6151, "step": 604 }, { "epoch": 4.1034336583298, "grad_norm": 0.13369676075093645, "learning_rate": 7.395255268338459e-06, "loss": 0.6166, "step": 605 }, { "epoch": 4.1102161933022465, "grad_norm": 0.1334402641156527, "learning_rate": 7.285493464698343e-06, "loss": 0.6103, "step": 606 }, { "epoch": 4.116998728274693, "grad_norm": 0.11322677082673437, "learning_rate": 7.176470647350675e-06, "loss": 0.6213, "step": 607 }, { "epoch": 4.123781263247139, "grad_norm": 0.1301924506080975, "learning_rate": 7.068189279006237e-06, "loss": 0.6099, "step": 608 }, { "epoch": 4.130563798219584, "grad_norm": 0.13846251217535666, "learning_rate": 6.960651805627248e-06, "loss": 0.5961, "step": 609 }, { "epoch": 4.13734633319203, "grad_norm": 0.11426017628960868, "learning_rate": 6.853860656372125e-06, "loss": 0.6001, "step": 610 }, { "epoch": 4.144128868164477, "grad_norm": 0.11173851423251971, "learning_rate": 6.747818243540587e-06, "loss": 0.6068, "step": 611 }, { "epoch": 4.150911403136923, "grad_norm": 0.13703419453889504, "learning_rate": 6.642526962519218e-06, "loss": 0.6083, "step": 612 }, { "epoch": 4.157693938109368, "grad_norm": 0.12103577637533855, "learning_rate": 6.537989191727292e-06, "loss": 0.6026, "step": 613 }, { "epoch": 4.164476473081814, "grad_norm": 0.12350144156787667, "learning_rate": 6.4342072925631125e-06, "loss": 0.6019, "step": 614 }, { "epoch": 4.17125900805426, "grad_norm": 0.11640263763641474, "learning_rate": 6.331183609350602e-06, "loss": 0.5954, "step": 615 }, { "epoch": 4.178041543026707, "grad_norm": 0.10545213840827801, "learning_rate": 6.228920469286408e-06, "loss": 0.6117, "step": 616 }, { "epoch": 4.184824077999152, "grad_norm": 0.11495651066805103, "learning_rate": 6.127420182387314e-06, "loss": 0.5946, "step": 617 }, { "epoch": 4.191606612971598, "grad_norm": 0.11620572636315642, "learning_rate": 6.026685041438018e-06, "loss": 0.6034, "step": 618 }, { "epoch": 4.198389147944044, "grad_norm": 0.10866699027698076, "learning_rate": 5.926717321939417e-06, "loss": 0.6048, "step": 619 }, { "epoch": 4.2051716829164905, "grad_norm": 0.1245636539367363, "learning_rate": 5.827519282057128e-06, "loss": 0.6037, "step": 620 }, { "epoch": 4.211954217888936, "grad_norm": 0.10644196388055245, "learning_rate": 5.729093162570545e-06, "loss": 0.607, "step": 621 }, { "epoch": 4.218736752861382, "grad_norm": 0.10894266432537414, "learning_rate": 5.631441186822168e-06, "loss": 0.6135, "step": 622 }, { "epoch": 4.225519287833828, "grad_norm": 0.12336549859525484, "learning_rate": 5.534565560667426e-06, "loss": 0.5981, "step": 623 }, { "epoch": 4.232301822806273, "grad_norm": 0.10948607920606619, "learning_rate": 5.4384684724248096e-06, "loss": 0.6118, "step": 624 }, { "epoch": 4.23908435777872, "grad_norm": 0.10785742485131251, "learning_rate": 5.343152092826476e-06, "loss": 0.6184, "step": 625 }, { "epoch": 4.245866892751166, "grad_norm": 0.1346812509747976, "learning_rate": 5.2486185749691665e-06, "loss": 0.6058, "step": 626 }, { "epoch": 4.252649427723612, "grad_norm": 0.10711676011900324, "learning_rate": 5.154870054265635e-06, "loss": 0.6071, "step": 627 }, { "epoch": 4.259431962696057, "grad_norm": 0.09686690911531301, "learning_rate": 5.061908648396338e-06, "loss": 0.6019, "step": 628 }, { "epoch": 4.2662144976685035, "grad_norm": 0.11286945799474804, "learning_rate": 4.969736457261669e-06, "loss": 0.604, "step": 629 }, { "epoch": 4.27299703264095, "grad_norm": 0.12539528001284939, "learning_rate": 4.878355562934465e-06, "loss": 0.6103, "step": 630 }, { "epoch": 4.279779567613396, "grad_norm": 0.11846976963374484, "learning_rate": 4.787768029613014e-06, "loss": 0.6102, "step": 631 }, { "epoch": 4.286562102585841, "grad_norm": 0.11455949347213798, "learning_rate": 4.697975903574419e-06, "loss": 0.6044, "step": 632 }, { "epoch": 4.293344637558287, "grad_norm": 0.11423245618436771, "learning_rate": 4.608981213128347e-06, "loss": 0.6104, "step": 633 }, { "epoch": 4.3001271725307335, "grad_norm": 0.11763066458751498, "learning_rate": 4.520785968571266e-06, "loss": 0.6103, "step": 634 }, { "epoch": 4.30690970750318, "grad_norm": 0.12407775475836062, "learning_rate": 4.433392162140968e-06, "loss": 0.6127, "step": 635 }, { "epoch": 4.313692242475625, "grad_norm": 0.1177975105922243, "learning_rate": 4.3468017679716245e-06, "loss": 0.6258, "step": 636 }, { "epoch": 4.320474777448071, "grad_norm": 0.11062581278413729, "learning_rate": 4.261016742049155e-06, "loss": 0.617, "step": 637 }, { "epoch": 4.327257312420517, "grad_norm": 0.11328064594157465, "learning_rate": 4.176039022167064e-06, "loss": 0.6086, "step": 638 }, { "epoch": 4.3340398473929636, "grad_norm": 0.12286256944458376, "learning_rate": 4.0918705278826685e-06, "loss": 0.6108, "step": 639 }, { "epoch": 4.340822382365409, "grad_norm": 0.11726282791596607, "learning_rate": 4.008513160473726e-06, "loss": 0.6122, "step": 640 }, { "epoch": 4.347604917337855, "grad_norm": 0.13243402516995725, "learning_rate": 3.925968802895477e-06, "loss": 0.6059, "step": 641 }, { "epoch": 4.354387452310301, "grad_norm": 0.11732415982857679, "learning_rate": 3.8442393197381456e-06, "loss": 0.6159, "step": 642 }, { "epoch": 4.3611699872827465, "grad_norm": 0.11432499142874704, "learning_rate": 3.7633265571847787e-06, "loss": 0.6053, "step": 643 }, { "epoch": 4.367952522255193, "grad_norm": 0.12438661313712396, "learning_rate": 3.683232342969594e-06, "loss": 0.6185, "step": 644 }, { "epoch": 4.374735057227639, "grad_norm": 0.12003558611356997, "learning_rate": 3.603958486336625e-06, "loss": 0.6311, "step": 645 }, { "epoch": 4.381517592200085, "grad_norm": 0.10916503150131288, "learning_rate": 3.5255067779989173e-06, "loss": 0.6158, "step": 646 }, { "epoch": 4.38830012717253, "grad_norm": 0.11467284610857227, "learning_rate": 3.4478789900980545e-06, "loss": 0.6176, "step": 647 }, { "epoch": 4.395082662144977, "grad_norm": 0.11307569011740914, "learning_rate": 3.371076876164101e-06, "loss": 0.614, "step": 648 }, { "epoch": 4.401865197117423, "grad_norm": 0.11385086584182105, "learning_rate": 3.295102171076039e-06, "loss": 0.6098, "step": 649 }, { "epoch": 4.408647732089869, "grad_norm": 0.13858694628446455, "learning_rate": 3.2199565910225305e-06, "loss": 0.5976, "step": 650 }, { "epoch": 4.415430267062314, "grad_norm": 0.12042166145746055, "learning_rate": 3.145641833463198e-06, "loss": 0.6164, "step": 651 }, { "epoch": 4.42221280203476, "grad_norm": 0.11519596457203823, "learning_rate": 3.0721595770902435e-06, "loss": 0.6156, "step": 652 }, { "epoch": 4.428995337007207, "grad_norm": 0.09905226368345378, "learning_rate": 2.9995114817905493e-06, "loss": 0.6001, "step": 653 }, { "epoch": 4.435777871979653, "grad_norm": 0.11732702729180129, "learning_rate": 2.927699188608171e-06, "loss": 0.607, "step": 654 }, { "epoch": 4.442560406952098, "grad_norm": 0.1179451846081634, "learning_rate": 2.856724319707289e-06, "loss": 0.6169, "step": 655 }, { "epoch": 4.449342941924544, "grad_norm": 0.10442323225441018, "learning_rate": 2.7865884783355234e-06, "loss": 0.6179, "step": 656 }, { "epoch": 4.4561254768969905, "grad_norm": 0.09435389572080079, "learning_rate": 2.717293248787769e-06, "loss": 0.6201, "step": 657 }, { "epoch": 4.462908011869436, "grad_norm": 0.11295536763791658, "learning_rate": 2.648840196370368e-06, "loss": 0.6019, "step": 658 }, { "epoch": 4.469690546841882, "grad_norm": 0.11594274422462209, "learning_rate": 2.5812308673657825e-06, "loss": 0.5977, "step": 659 }, { "epoch": 4.476473081814328, "grad_norm": 0.10182885233809727, "learning_rate": 2.5144667889976316e-06, "loss": 0.6017, "step": 660 }, { "epoch": 4.483255616786774, "grad_norm": 0.10348633393081054, "learning_rate": 2.4485494693962197e-06, "loss": 0.6086, "step": 661 }, { "epoch": 4.49003815175922, "grad_norm": 0.10025894496329021, "learning_rate": 2.3834803975644772e-06, "loss": 0.6182, "step": 662 }, { "epoch": 4.496820686731666, "grad_norm": 0.10187054293463914, "learning_rate": 2.3192610433442784e-06, "loss": 0.6074, "step": 663 }, { "epoch": 4.503603221704112, "grad_norm": 0.09715916393818302, "learning_rate": 2.2558928573833016e-06, "loss": 0.6076, "step": 664 }, { "epoch": 4.510385756676558, "grad_norm": 0.09756303138919038, "learning_rate": 2.1933772711021995e-06, "loss": 0.6019, "step": 665 }, { "epoch": 4.5171682916490035, "grad_norm": 0.09525032911781027, "learning_rate": 2.1317156966623243e-06, "loss": 0.5995, "step": 666 }, { "epoch": 4.52395082662145, "grad_norm": 0.09487242097788966, "learning_rate": 2.0709095269337755e-06, "loss": 0.6023, "step": 667 }, { "epoch": 4.530733361593896, "grad_norm": 0.11476206706719008, "learning_rate": 2.0109601354639706e-06, "loss": 0.6143, "step": 668 }, { "epoch": 4.537515896566342, "grad_norm": 0.09904862194550892, "learning_rate": 1.9518688764466096e-06, "loss": 0.6143, "step": 669 }, { "epoch": 4.544298431538787, "grad_norm": 0.09278301353904947, "learning_rate": 1.8936370846910845e-06, "loss": 0.6092, "step": 670 }, { "epoch": 4.5510809665112335, "grad_norm": 0.10688691601881584, "learning_rate": 1.8362660755923079e-06, "loss": 0.6186, "step": 671 }, { "epoch": 4.55786350148368, "grad_norm": 0.09697916608330605, "learning_rate": 1.7797571451010442e-06, "loss": 0.5983, "step": 672 }, { "epoch": 4.564646036456125, "grad_norm": 0.10471706981067595, "learning_rate": 1.7241115696945821e-06, "loss": 0.6132, "step": 673 }, { "epoch": 4.571428571428571, "grad_norm": 0.09406802929392503, "learning_rate": 1.669330606347952e-06, "loss": 0.5962, "step": 674 }, { "epoch": 4.578211106401017, "grad_norm": 0.0954037644840885, "learning_rate": 1.6154154925054878e-06, "loss": 0.6141, "step": 675 }, { "epoch": 4.5849936413734635, "grad_norm": 0.10094861062769968, "learning_rate": 1.5623674460529014e-06, "loss": 0.6024, "step": 676 }, { "epoch": 4.59177617634591, "grad_norm": 0.09573172404295381, "learning_rate": 1.510187665289773e-06, "loss": 0.6086, "step": 677 }, { "epoch": 4.598558711318355, "grad_norm": 0.09995771347446304, "learning_rate": 1.458877328902455e-06, "loss": 0.6052, "step": 678 }, { "epoch": 4.605341246290801, "grad_norm": 0.09656263063599439, "learning_rate": 1.4084375959374864e-06, "loss": 0.6072, "step": 679 }, { "epoch": 4.612123781263247, "grad_norm": 0.09822686472765482, "learning_rate": 1.3588696057753725e-06, "loss": 0.6089, "step": 680 }, { "epoch": 4.618906316235693, "grad_norm": 0.09676277658643694, "learning_rate": 1.3101744781048774e-06, "loss": 0.6016, "step": 681 }, { "epoch": 4.625688851208139, "grad_norm": 0.0933744691144583, "learning_rate": 1.262353312897715e-06, "loss": 0.6075, "step": 682 }, { "epoch": 4.632471386180585, "grad_norm": 0.0996442423520143, "learning_rate": 1.2154071903837107e-06, "loss": 0.6157, "step": 683 }, { "epoch": 4.639253921153031, "grad_norm": 0.10109190980066803, "learning_rate": 1.1693371710263857e-06, "loss": 0.6049, "step": 684 }, { "epoch": 4.6460364561254766, "grad_norm": 0.08703236124988406, "learning_rate": 1.1241442954990301e-06, "loss": 0.6048, "step": 685 }, { "epoch": 4.652818991097923, "grad_norm": 0.09280509544326945, "learning_rate": 1.0798295846611562e-06, "loss": 0.5989, "step": 686 }, { "epoch": 4.659601526070369, "grad_norm": 0.09785736700280491, "learning_rate": 1.0363940395354777e-06, "loss": 0.6098, "step": 687 }, { "epoch": 4.666384061042815, "grad_norm": 0.09508110174195833, "learning_rate": 9.938386412852652e-07, "loss": 0.6149, "step": 688 }, { "epoch": 4.67316659601526, "grad_norm": 0.09065950330259104, "learning_rate": 9.52164351192213e-07, "loss": 0.6088, "step": 689 }, { "epoch": 4.679949130987707, "grad_norm": 0.09551867620930947, "learning_rate": 9.113721106346918e-07, "loss": 0.5989, "step": 690 }, { "epoch": 4.686731665960153, "grad_norm": 0.09832290856026324, "learning_rate": 8.714628410665082e-07, "loss": 0.6172, "step": 691 }, { "epoch": 4.693514200932599, "grad_norm": 0.09371892802996978, "learning_rate": 8.324374439960947e-07, "loss": 0.6067, "step": 692 }, { "epoch": 4.700296735905044, "grad_norm": 0.09014551117205812, "learning_rate": 7.942968009661123e-07, "loss": 0.6142, "step": 693 }, { "epoch": 4.7070792708774905, "grad_norm": 0.09724706571681141, "learning_rate": 7.570417735335733e-07, "loss": 0.614, "step": 694 }, { "epoch": 4.713861805849937, "grad_norm": 0.08494611369686998, "learning_rate": 7.206732032503638e-07, "loss": 0.5969, "step": 695 }, { "epoch": 4.720644340822382, "grad_norm": 0.0927927564675014, "learning_rate": 6.851919116442274e-07, "loss": 0.6122, "step": 696 }, { "epoch": 4.727426875794828, "grad_norm": 0.09273724936533222, "learning_rate": 6.505987002002245e-07, "loss": 0.6092, "step": 697 }, { "epoch": 4.734209410767274, "grad_norm": 0.08796818616185165, "learning_rate": 6.168943503426139e-07, "loss": 0.6038, "step": 698 }, { "epoch": 4.7409919457397205, "grad_norm": 0.08686636687902749, "learning_rate": 5.840796234172085e-07, "loss": 0.5943, "step": 699 }, { "epoch": 4.747774480712166, "grad_norm": 0.09236449250012803, "learning_rate": 5.521552606741765e-07, "loss": 0.6066, "step": 700 }, { "epoch": 4.754557015684612, "grad_norm": 0.0926749240950783, "learning_rate": 5.211219832512893e-07, "loss": 0.6114, "step": 701 }, { "epoch": 4.761339550657058, "grad_norm": 0.10499405025384424, "learning_rate": 4.909804921576466e-07, "loss": 0.6185, "step": 702 }, { "epoch": 4.768122085629504, "grad_norm": 0.08848832101586154, "learning_rate": 4.6173146825782224e-07, "loss": 0.6067, "step": 703 }, { "epoch": 4.77490462060195, "grad_norm": 0.09319395321366955, "learning_rate": 4.3337557225650695e-07, "loss": 0.6047, "step": 704 }, { "epoch": 4.781687155574396, "grad_norm": 0.0915059840709739, "learning_rate": 4.0591344468355666e-07, "loss": 0.6118, "step": 705 }, { "epoch": 4.788469690546842, "grad_norm": 0.10729723170818359, "learning_rate": 3.793457058795591e-07, "loss": 0.6189, "step": 706 }, { "epoch": 4.795252225519288, "grad_norm": 0.09285587112058538, "learning_rate": 3.5367295598178307e-07, "loss": 0.6085, "step": 707 }, { "epoch": 4.8020347604917335, "grad_norm": 0.08196537866623092, "learning_rate": 3.2889577491064693e-07, "loss": 0.6045, "step": 708 }, { "epoch": 4.80881729546418, "grad_norm": 0.08720837958758754, "learning_rate": 3.050147223566091e-07, "loss": 0.6122, "step": 709 }, { "epoch": 4.815599830436626, "grad_norm": 0.09158067797742139, "learning_rate": 2.8203033776752487e-07, "loss": 0.6033, "step": 710 }, { "epoch": 4.822382365409071, "grad_norm": 0.09043295389558649, "learning_rate": 2.599431403364694e-07, "loss": 0.6094, "step": 711 }, { "epoch": 4.829164900381517, "grad_norm": 0.09180388224113892, "learning_rate": 2.3875362899000054e-07, "loss": 0.604, "step": 712 }, { "epoch": 4.8359474353539635, "grad_norm": 0.08651605692378189, "learning_rate": 2.1846228237689226e-07, "loss": 0.6087, "step": 713 }, { "epoch": 4.84272997032641, "grad_norm": 0.08511470109771488, "learning_rate": 1.9906955885732104e-07, "loss": 0.6091, "step": 714 }, { "epoch": 4.849512505298856, "grad_norm": 0.08454739481192287, "learning_rate": 1.8057589649251862e-07, "loss": 0.6146, "step": 715 }, { "epoch": 4.856295040271301, "grad_norm": 0.08724704005095565, "learning_rate": 1.62981713034851e-07, "loss": 0.6042, "step": 716 }, { "epoch": 4.863077575243747, "grad_norm": 0.08396574948424247, "learning_rate": 1.462874059184305e-07, "loss": 0.6086, "step": 717 }, { "epoch": 4.869860110216194, "grad_norm": 0.09082130018441244, "learning_rate": 1.3049335225009175e-07, "loss": 0.6295, "step": 718 }, { "epoch": 4.876642645188639, "grad_norm": 0.08722605095693269, "learning_rate": 1.1559990880089189e-07, "loss": 0.6085, "step": 719 }, { "epoch": 4.883425180161085, "grad_norm": 0.0878159462681937, "learning_rate": 1.0160741199805036e-07, "loss": 0.5998, "step": 720 }, { "epoch": 4.890207715133531, "grad_norm": 0.0878736816907958, "learning_rate": 8.851617791735933e-08, "loss": 0.6145, "step": 721 }, { "epoch": 4.896990250105977, "grad_norm": 0.08847166155449576, "learning_rate": 7.632650227602511e-08, "loss": 0.604, "step": 722 }, { "epoch": 4.903772785078423, "grad_norm": 0.08650107449834574, "learning_rate": 6.503866042599338e-08, "loss": 0.6187, "step": 723 }, { "epoch": 4.910555320050869, "grad_norm": 0.09963600285093287, "learning_rate": 5.4652907347745307e-08, "loss": 0.6166, "step": 724 }, { "epoch": 4.917337855023315, "grad_norm": 0.08660963081265105, "learning_rate": 4.516947764451107e-08, "loss": 0.6156, "step": 725 }, { "epoch": 4.924120389995761, "grad_norm": 0.09162932265931903, "learning_rate": 3.6588585537002955e-08, "loss": 0.6085, "step": 726 }, { "epoch": 4.930902924968207, "grad_norm": 0.08533741638096395, "learning_rate": 2.8910424858543673e-08, "loss": 0.6026, "step": 727 }, { "epoch": 4.937685459940653, "grad_norm": 0.08431534489847897, "learning_rate": 2.213516905070101e-08, "loss": 0.6012, "step": 728 }, { "epoch": 4.944467994913099, "grad_norm": 0.08633616914365788, "learning_rate": 1.6262971159384243e-08, "loss": 0.6146, "step": 729 }, { "epoch": 4.951250529885545, "grad_norm": 0.0931259062674402, "learning_rate": 1.1293963831366939e-08, "loss": 0.6116, "step": 730 }, { "epoch": 4.9580330648579904, "grad_norm": 0.08678440176862204, "learning_rate": 7.228259311293784e-09, "loss": 0.602, "step": 731 }, { "epoch": 4.964815599830437, "grad_norm": 0.0942499520961841, "learning_rate": 4.065949439158168e-09, "loss": 0.6099, "step": 732 }, { "epoch": 4.971598134802883, "grad_norm": 0.07935089647201646, "learning_rate": 1.8071056482105166e-09, "loss": 0.615, "step": 733 }, { "epoch": 4.978380669775328, "grad_norm": 0.08485220331271215, "learning_rate": 4.517789633551317e-10, "loss": 0.6104, "step": 734 }, { "epoch": 4.985163204747774, "grad_norm": 0.08704984136629952, "learning_rate": 0.0, "loss": 0.6111, "step": 735 }, { "epoch": 4.985163204747774, "step": 735, "total_flos": 1.8913738317545603e+19, "train_loss": 0.673498132682982, "train_runtime": 162938.9767, "train_samples_per_second": 2.316, "train_steps_per_second": 0.005 } ], "logging_steps": 1.0, "max_steps": 735, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.8913738317545603e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }