{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.996456118407338, "eval_steps": 400, "global_step": 3597, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008338544923910778, "grad_norm": 0.35252463817596436, "learning_rate": 2e-08, "loss": 0.8273, "step": 1 }, { "epoch": 0.0008338544923910778, "eval_loss": 0.861491858959198, "eval_runtime": 319.044, "eval_samples_per_second": 16.872, "eval_steps_per_second": 2.815, "step": 1 }, { "epoch": 0.0016677089847821555, "grad_norm": 0.3189675807952881, "learning_rate": 4e-08, "loss": 0.8572, "step": 2 }, { "epoch": 0.0025015634771732333, "grad_norm": 0.3477325141429901, "learning_rate": 6e-08, "loss": 0.9059, "step": 3 }, { "epoch": 0.003335417969564311, "grad_norm": 0.32423585653305054, "learning_rate": 8e-08, "loss": 0.8386, "step": 4 }, { "epoch": 0.004169272461955389, "grad_norm": 0.32164493203163147, "learning_rate": 1e-07, "loss": 0.8753, "step": 5 }, { "epoch": 0.0050031269543464665, "grad_norm": 0.33231157064437866, "learning_rate": 1.2e-07, "loss": 0.8665, "step": 6 }, { "epoch": 0.005836981446737544, "grad_norm": 0.31154102087020874, "learning_rate": 1.4e-07, "loss": 0.8813, "step": 7 }, { "epoch": 0.006670835939128622, "grad_norm": 0.3176712393760681, "learning_rate": 1.6e-07, "loss": 0.8683, "step": 8 }, { "epoch": 0.0075046904315197, "grad_norm": 0.3419969975948334, "learning_rate": 1.8e-07, "loss": 0.849, "step": 9 }, { "epoch": 0.008338544923910778, "grad_norm": 0.33408281207084656, "learning_rate": 2e-07, "loss": 0.8778, "step": 10 }, { "epoch": 0.009172399416301855, "grad_norm": 0.3125899136066437, "learning_rate": 2.1999999999999998e-07, "loss": 0.8051, "step": 11 }, { "epoch": 0.010006253908692933, "grad_norm": 0.32648083567619324, "learning_rate": 2.4e-07, "loss": 0.8297, "step": 12 }, { "epoch": 0.01084010840108401, "grad_norm": 0.3143678605556488, "learning_rate": 2.6e-07, "loss": 0.8401, "step": 13 }, { "epoch": 0.011673962893475089, "grad_norm": 0.3291407525539398, "learning_rate": 2.8e-07, "loss": 0.8668, "step": 14 }, { "epoch": 0.012507817385866166, "grad_norm": 0.3211856484413147, "learning_rate": 3e-07, "loss": 0.84, "step": 15 }, { "epoch": 0.013341671878257244, "grad_norm": 0.31169289350509644, "learning_rate": 3.2e-07, "loss": 0.8659, "step": 16 }, { "epoch": 0.014175526370648322, "grad_norm": 0.3279392719268799, "learning_rate": 3.4000000000000003e-07, "loss": 0.8317, "step": 17 }, { "epoch": 0.0150093808630394, "grad_norm": 0.3028814494609833, "learning_rate": 3.6e-07, "loss": 0.8246, "step": 18 }, { "epoch": 0.015843235355430477, "grad_norm": 0.307066410779953, "learning_rate": 3.7999999999999996e-07, "loss": 0.8025, "step": 19 }, { "epoch": 0.016677089847821555, "grad_norm": 0.3293515741825104, "learning_rate": 4e-07, "loss": 0.8674, "step": 20 }, { "epoch": 0.017510944340212633, "grad_norm": 0.3215535581111908, "learning_rate": 4.1999999999999995e-07, "loss": 0.8533, "step": 21 }, { "epoch": 0.01834479883260371, "grad_norm": 0.3236245810985565, "learning_rate": 4.3999999999999997e-07, "loss": 0.8957, "step": 22 }, { "epoch": 0.01917865332499479, "grad_norm": 0.33956846594810486, "learning_rate": 4.6e-07, "loss": 0.8293, "step": 23 }, { "epoch": 0.020012507817385866, "grad_norm": 0.33766457438468933, "learning_rate": 4.8e-07, "loss": 0.8381, "step": 24 }, { "epoch": 0.020846362309776944, "grad_norm": 0.3336811065673828, "learning_rate": 5e-07, "loss": 0.8475, "step": 25 }, { "epoch": 0.02168021680216802, "grad_norm": 0.307309627532959, "learning_rate": 5.2e-07, "loss": 0.852, "step": 26 }, { "epoch": 0.0225140712945591, "grad_norm": 0.29773110151290894, "learning_rate": 5.4e-07, "loss": 0.8042, "step": 27 }, { "epoch": 0.023347925786950177, "grad_norm": 0.3210948407649994, "learning_rate": 5.6e-07, "loss": 0.9099, "step": 28 }, { "epoch": 0.024181780279341255, "grad_norm": 0.3149116337299347, "learning_rate": 5.8e-07, "loss": 0.8, "step": 29 }, { "epoch": 0.025015634771732333, "grad_norm": 0.29307302832603455, "learning_rate": 6e-07, "loss": 0.8583, "step": 30 }, { "epoch": 0.02584948926412341, "grad_norm": 0.3131772577762604, "learning_rate": 6.2e-07, "loss": 0.8432, "step": 31 }, { "epoch": 0.026683343756514488, "grad_norm": 0.28058937191963196, "learning_rate": 6.4e-07, "loss": 0.8099, "step": 32 }, { "epoch": 0.027517198248905566, "grad_norm": 0.2649301290512085, "learning_rate": 6.6e-07, "loss": 0.8715, "step": 33 }, { "epoch": 0.028351052741296644, "grad_norm": 0.2659442126750946, "learning_rate": 6.800000000000001e-07, "loss": 0.7432, "step": 34 }, { "epoch": 0.02918490723368772, "grad_norm": 0.24875527620315552, "learning_rate": 7e-07, "loss": 0.8502, "step": 35 }, { "epoch": 0.0300187617260788, "grad_norm": 0.21730971336364746, "learning_rate": 7.2e-07, "loss": 0.813, "step": 36 }, { "epoch": 0.030852616218469877, "grad_norm": 0.19625824689865112, "learning_rate": 7.4e-07, "loss": 0.7839, "step": 37 }, { "epoch": 0.031686470710860955, "grad_norm": 0.19287347793579102, "learning_rate": 7.599999999999999e-07, "loss": 0.7949, "step": 38 }, { "epoch": 0.032520325203252036, "grad_norm": 0.1588095873594284, "learning_rate": 7.799999999999999e-07, "loss": 0.7137, "step": 39 }, { "epoch": 0.03335417969564311, "grad_norm": 0.17520244419574738, "learning_rate": 8e-07, "loss": 0.8112, "step": 40 }, { "epoch": 0.03418803418803419, "grad_norm": 0.16394659876823425, "learning_rate": 8.199999999999999e-07, "loss": 0.8096, "step": 41 }, { "epoch": 0.035021888680425266, "grad_norm": 0.1522587239742279, "learning_rate": 8.399999999999999e-07, "loss": 0.708, "step": 42 }, { "epoch": 0.03585574317281635, "grad_norm": 0.16340652108192444, "learning_rate": 8.599999999999999e-07, "loss": 0.7816, "step": 43 }, { "epoch": 0.03668959766520742, "grad_norm": 0.15452717244625092, "learning_rate": 8.799999999999999e-07, "loss": 0.7946, "step": 44 }, { "epoch": 0.0375234521575985, "grad_norm": 0.1545872688293457, "learning_rate": 9e-07, "loss": 0.7908, "step": 45 }, { "epoch": 0.03835730664998958, "grad_norm": 0.13799017667770386, "learning_rate": 9.2e-07, "loss": 0.7056, "step": 46 }, { "epoch": 0.03919116114238066, "grad_norm": 0.14599646627902985, "learning_rate": 9.399999999999999e-07, "loss": 0.8098, "step": 47 }, { "epoch": 0.04002501563477173, "grad_norm": 0.13951805233955383, "learning_rate": 9.6e-07, "loss": 0.7938, "step": 48 }, { "epoch": 0.040858870127162814, "grad_norm": 0.14058919250965118, "learning_rate": 9.8e-07, "loss": 0.7806, "step": 49 }, { "epoch": 0.04169272461955389, "grad_norm": 0.14300380647182465, "learning_rate": 1e-06, "loss": 0.8192, "step": 50 }, { "epoch": 0.04252657911194497, "grad_norm": 0.13719584047794342, "learning_rate": 1.02e-06, "loss": 0.7939, "step": 51 }, { "epoch": 0.04336043360433604, "grad_norm": 0.14002980291843414, "learning_rate": 1.04e-06, "loss": 0.8163, "step": 52 }, { "epoch": 0.044194288096727125, "grad_norm": 0.13815677165985107, "learning_rate": 1.06e-06, "loss": 0.8137, "step": 53 }, { "epoch": 0.0450281425891182, "grad_norm": 0.13772878050804138, "learning_rate": 1.08e-06, "loss": 0.771, "step": 54 }, { "epoch": 0.04586199708150928, "grad_norm": 0.1414576917886734, "learning_rate": 1.1e-06, "loss": 0.7989, "step": 55 }, { "epoch": 0.046695851573900354, "grad_norm": 0.11953306943178177, "learning_rate": 1.12e-06, "loss": 0.7074, "step": 56 }, { "epoch": 0.047529706066291436, "grad_norm": 0.13051201403141022, "learning_rate": 1.1399999999999999e-06, "loss": 0.8034, "step": 57 }, { "epoch": 0.04836356055868251, "grad_norm": 0.12802985310554504, "learning_rate": 1.16e-06, "loss": 0.7581, "step": 58 }, { "epoch": 0.04919741505107359, "grad_norm": 0.13593435287475586, "learning_rate": 1.18e-06, "loss": 0.7978, "step": 59 }, { "epoch": 0.050031269543464665, "grad_norm": 0.12595658004283905, "learning_rate": 1.2e-06, "loss": 0.7997, "step": 60 }, { "epoch": 0.05086512403585575, "grad_norm": 0.1257220059633255, "learning_rate": 1.22e-06, "loss": 0.7194, "step": 61 }, { "epoch": 0.05169897852824682, "grad_norm": 0.12169249355792999, "learning_rate": 1.24e-06, "loss": 0.7412, "step": 62 }, { "epoch": 0.0525328330206379, "grad_norm": 0.12391054630279541, "learning_rate": 1.26e-06, "loss": 0.7849, "step": 63 }, { "epoch": 0.053366687513028976, "grad_norm": 0.12134575098752975, "learning_rate": 1.28e-06, "loss": 0.6859, "step": 64 }, { "epoch": 0.05420054200542006, "grad_norm": 0.11605742573738098, "learning_rate": 1.3e-06, "loss": 0.8112, "step": 65 }, { "epoch": 0.05503439649781113, "grad_norm": 0.11593819409608841, "learning_rate": 1.32e-06, "loss": 0.6993, "step": 66 }, { "epoch": 0.05586825099020221, "grad_norm": 0.12103109061717987, "learning_rate": 1.34e-06, "loss": 0.7595, "step": 67 }, { "epoch": 0.05670210548259329, "grad_norm": 0.12391602247953415, "learning_rate": 1.3600000000000001e-06, "loss": 0.7676, "step": 68 }, { "epoch": 0.05753595997498437, "grad_norm": 0.10895267128944397, "learning_rate": 1.38e-06, "loss": 0.7256, "step": 69 }, { "epoch": 0.05836981446737544, "grad_norm": 0.1129627376794815, "learning_rate": 1.4e-06, "loss": 0.7097, "step": 70 }, { "epoch": 0.059203668959766524, "grad_norm": 0.1078682616353035, "learning_rate": 1.42e-06, "loss": 0.7052, "step": 71 }, { "epoch": 0.0600375234521576, "grad_norm": 0.1140650063753128, "learning_rate": 1.44e-06, "loss": 0.8067, "step": 72 }, { "epoch": 0.06087137794454868, "grad_norm": 0.10528790205717087, "learning_rate": 1.46e-06, "loss": 0.7241, "step": 73 }, { "epoch": 0.061705232436939754, "grad_norm": 0.11815836280584335, "learning_rate": 1.48e-06, "loss": 0.7344, "step": 74 }, { "epoch": 0.06253908692933083, "grad_norm": 0.11140462756156921, "learning_rate": 1.5e-06, "loss": 0.7664, "step": 75 }, { "epoch": 0.06337294142172191, "grad_norm": 0.11665944010019302, "learning_rate": 1.5199999999999998e-06, "loss": 0.7178, "step": 76 }, { "epoch": 0.06420679591411299, "grad_norm": 0.10751160234212875, "learning_rate": 1.5399999999999999e-06, "loss": 0.7269, "step": 77 }, { "epoch": 0.06504065040650407, "grad_norm": 0.09768356382846832, "learning_rate": 1.5599999999999999e-06, "loss": 0.7449, "step": 78 }, { "epoch": 0.06587450489889514, "grad_norm": 0.09686972200870514, "learning_rate": 1.58e-06, "loss": 0.7097, "step": 79 }, { "epoch": 0.06670835939128622, "grad_norm": 0.09527275711297989, "learning_rate": 1.6e-06, "loss": 0.7301, "step": 80 }, { "epoch": 0.0675422138836773, "grad_norm": 0.0931195393204689, "learning_rate": 1.62e-06, "loss": 0.7524, "step": 81 }, { "epoch": 0.06837606837606838, "grad_norm": 0.0892946794629097, "learning_rate": 1.6399999999999998e-06, "loss": 0.7156, "step": 82 }, { "epoch": 0.06920992286845945, "grad_norm": 0.08845791220664978, "learning_rate": 1.6599999999999998e-06, "loss": 0.7349, "step": 83 }, { "epoch": 0.07004377736085053, "grad_norm": 0.0896732360124588, "learning_rate": 1.6799999999999998e-06, "loss": 0.7145, "step": 84 }, { "epoch": 0.07087763185324161, "grad_norm": 0.09997335076332092, "learning_rate": 1.6999999999999998e-06, "loss": 0.7103, "step": 85 }, { "epoch": 0.0717114863456327, "grad_norm": 0.08821084350347519, "learning_rate": 1.7199999999999998e-06, "loss": 0.725, "step": 86 }, { "epoch": 0.07254534083802376, "grad_norm": 0.08937980234622955, "learning_rate": 1.7399999999999999e-06, "loss": 0.7194, "step": 87 }, { "epoch": 0.07337919533041484, "grad_norm": 0.09307857602834702, "learning_rate": 1.7599999999999999e-06, "loss": 0.7595, "step": 88 }, { "epoch": 0.07421304982280592, "grad_norm": 0.08865829557180405, "learning_rate": 1.78e-06, "loss": 0.7464, "step": 89 }, { "epoch": 0.075046904315197, "grad_norm": 0.09038495272397995, "learning_rate": 1.8e-06, "loss": 0.6698, "step": 90 }, { "epoch": 0.07588075880758807, "grad_norm": 0.08254078030586243, "learning_rate": 1.82e-06, "loss": 0.7426, "step": 91 }, { "epoch": 0.07671461329997915, "grad_norm": 0.08807505667209625, "learning_rate": 1.84e-06, "loss": 0.6789, "step": 92 }, { "epoch": 0.07754846779237023, "grad_norm": 0.09523889422416687, "learning_rate": 1.86e-06, "loss": 0.7218, "step": 93 }, { "epoch": 0.07838232228476132, "grad_norm": 0.08782277256250381, "learning_rate": 1.8799999999999998e-06, "loss": 0.7599, "step": 94 }, { "epoch": 0.07921617677715238, "grad_norm": 0.08968065679073334, "learning_rate": 1.8999999999999998e-06, "loss": 0.7971, "step": 95 }, { "epoch": 0.08005003126954346, "grad_norm": 0.08655782043933868, "learning_rate": 1.92e-06, "loss": 0.7479, "step": 96 }, { "epoch": 0.08088388576193455, "grad_norm": 0.0854155421257019, "learning_rate": 1.94e-06, "loss": 0.7327, "step": 97 }, { "epoch": 0.08171774025432563, "grad_norm": 0.08121508359909058, "learning_rate": 1.96e-06, "loss": 0.6749, "step": 98 }, { "epoch": 0.0825515947467167, "grad_norm": 0.08429264277219772, "learning_rate": 1.98e-06, "loss": 0.6402, "step": 99 }, { "epoch": 0.08338544923910778, "grad_norm": 0.08390086144208908, "learning_rate": 2e-06, "loss": 0.7434, "step": 100 }, { "epoch": 0.08421930373149886, "grad_norm": 0.08662284910678864, "learning_rate": 1.9999995964675577e-06, "loss": 0.6681, "step": 101 }, { "epoch": 0.08505315822388994, "grad_norm": 0.08700251579284668, "learning_rate": 1.9999983858705566e-06, "loss": 0.763, "step": 102 }, { "epoch": 0.085887012716281, "grad_norm": 0.08504343777894974, "learning_rate": 1.9999963682099734e-06, "loss": 0.77, "step": 103 }, { "epoch": 0.08672086720867209, "grad_norm": 0.08222745358943939, "learning_rate": 1.999993543487437e-06, "loss": 0.7103, "step": 104 }, { "epoch": 0.08755472170106317, "grad_norm": 0.08124719560146332, "learning_rate": 1.9999899117052263e-06, "loss": 0.6913, "step": 105 }, { "epoch": 0.08838857619345425, "grad_norm": 0.09033836424350739, "learning_rate": 1.9999854728662734e-06, "loss": 0.7065, "step": 106 }, { "epoch": 0.08922243068584532, "grad_norm": 0.09045730531215668, "learning_rate": 1.99998022697416e-06, "loss": 0.7624, "step": 107 }, { "epoch": 0.0900562851782364, "grad_norm": 0.0832991823554039, "learning_rate": 1.9999741740331203e-06, "loss": 0.7399, "step": 108 }, { "epoch": 0.09089013967062748, "grad_norm": 0.07666011154651642, "learning_rate": 1.9999673140480388e-06, "loss": 0.6477, "step": 109 }, { "epoch": 0.09172399416301856, "grad_norm": 0.08195007592439651, "learning_rate": 1.9999596470244527e-06, "loss": 0.7068, "step": 110 }, { "epoch": 0.09255784865540963, "grad_norm": 0.07755164802074432, "learning_rate": 1.99995117296855e-06, "loss": 0.7021, "step": 111 }, { "epoch": 0.09339170314780071, "grad_norm": 0.0782230794429779, "learning_rate": 1.9999418918871685e-06, "loss": 0.7039, "step": 112 }, { "epoch": 0.09422555764019179, "grad_norm": 0.08311248570680618, "learning_rate": 1.9999318037877995e-06, "loss": 0.7238, "step": 113 }, { "epoch": 0.09505941213258287, "grad_norm": 0.08455490320920944, "learning_rate": 1.999920908678585e-06, "loss": 0.7184, "step": 114 }, { "epoch": 0.09589326662497394, "grad_norm": 0.08463476598262787, "learning_rate": 1.9999092065683174e-06, "loss": 0.6162, "step": 115 }, { "epoch": 0.09672712111736502, "grad_norm": 0.07902689278125763, "learning_rate": 1.999896697466442e-06, "loss": 0.6649, "step": 116 }, { "epoch": 0.0975609756097561, "grad_norm": 0.08383440971374512, "learning_rate": 1.999883381383053e-06, "loss": 0.703, "step": 117 }, { "epoch": 0.09839483010214718, "grad_norm": 0.07288578897714615, "learning_rate": 1.999869258328899e-06, "loss": 0.7286, "step": 118 }, { "epoch": 0.09922868459453825, "grad_norm": 0.08177065849304199, "learning_rate": 1.999854328315377e-06, "loss": 0.6899, "step": 119 }, { "epoch": 0.10006253908692933, "grad_norm": 0.07529062032699585, "learning_rate": 1.999838591354537e-06, "loss": 0.6781, "step": 120 }, { "epoch": 0.10089639357932041, "grad_norm": 0.08230678737163544, "learning_rate": 1.9998220474590795e-06, "loss": 0.7059, "step": 121 }, { "epoch": 0.1017302480717115, "grad_norm": 0.08000528067350388, "learning_rate": 1.9998046966423567e-06, "loss": 0.6561, "step": 122 }, { "epoch": 0.10256410256410256, "grad_norm": 0.07669328153133392, "learning_rate": 1.9997865389183717e-06, "loss": 0.6452, "step": 123 }, { "epoch": 0.10339795705649364, "grad_norm": 0.07872146368026733, "learning_rate": 1.9997675743017794e-06, "loss": 0.7337, "step": 124 }, { "epoch": 0.10423181154888472, "grad_norm": 0.07842138409614563, "learning_rate": 1.9997478028078853e-06, "loss": 0.7126, "step": 125 }, { "epoch": 0.1050656660412758, "grad_norm": 0.07890679687261581, "learning_rate": 1.9997272244526453e-06, "loss": 0.7049, "step": 126 }, { "epoch": 0.10589952053366687, "grad_norm": 0.07983887195587158, "learning_rate": 1.999705839252669e-06, "loss": 0.7578, "step": 127 }, { "epoch": 0.10673337502605795, "grad_norm": 0.07721222192049026, "learning_rate": 1.9996836472252144e-06, "loss": 0.678, "step": 128 }, { "epoch": 0.10756722951844903, "grad_norm": 0.07825150340795517, "learning_rate": 1.9996606483881927e-06, "loss": 0.6763, "step": 129 }, { "epoch": 0.10840108401084012, "grad_norm": 0.08063995838165283, "learning_rate": 1.999636842760165e-06, "loss": 0.758, "step": 130 }, { "epoch": 0.10923493850323118, "grad_norm": 0.0724928081035614, "learning_rate": 1.9996122303603443e-06, "loss": 0.7312, "step": 131 }, { "epoch": 0.11006879299562226, "grad_norm": 0.08180397748947144, "learning_rate": 1.999586811208594e-06, "loss": 0.6794, "step": 132 }, { "epoch": 0.11090264748801335, "grad_norm": 0.0856177881360054, "learning_rate": 1.99956058532543e-06, "loss": 0.6885, "step": 133 }, { "epoch": 0.11173650198040443, "grad_norm": 0.0763324424624443, "learning_rate": 1.999533552732017e-06, "loss": 0.6821, "step": 134 }, { "epoch": 0.1125703564727955, "grad_norm": 0.07964632660150528, "learning_rate": 1.9995057134501725e-06, "loss": 0.7362, "step": 135 }, { "epoch": 0.11340421096518657, "grad_norm": 0.07633720338344574, "learning_rate": 1.999477067502365e-06, "loss": 0.7396, "step": 136 }, { "epoch": 0.11423806545757766, "grad_norm": 0.07655656337738037, "learning_rate": 1.9994476149117132e-06, "loss": 0.728, "step": 137 }, { "epoch": 0.11507191994996874, "grad_norm": 0.07770884782075882, "learning_rate": 1.9994173557019877e-06, "loss": 0.6878, "step": 138 }, { "epoch": 0.1159057744423598, "grad_norm": 0.07332731038331985, "learning_rate": 1.999386289897609e-06, "loss": 0.7115, "step": 139 }, { "epoch": 0.11673962893475089, "grad_norm": 0.0784095972776413, "learning_rate": 1.9993544175236497e-06, "loss": 0.6653, "step": 140 }, { "epoch": 0.11757348342714197, "grad_norm": 0.08311771601438522, "learning_rate": 1.9993217386058326e-06, "loss": 0.6504, "step": 141 }, { "epoch": 0.11840733791953305, "grad_norm": 0.07529395818710327, "learning_rate": 1.999288253170532e-06, "loss": 0.7083, "step": 142 }, { "epoch": 0.11924119241192412, "grad_norm": 0.07608146965503693, "learning_rate": 1.999253961244773e-06, "loss": 0.6856, "step": 143 }, { "epoch": 0.1200750469043152, "grad_norm": 0.07125360518693924, "learning_rate": 1.9992188628562303e-06, "loss": 0.6095, "step": 144 }, { "epoch": 0.12090890139670628, "grad_norm": 0.07715660333633423, "learning_rate": 1.999182958033232e-06, "loss": 0.7337, "step": 145 }, { "epoch": 0.12174275588909736, "grad_norm": 0.07756403833627701, "learning_rate": 1.999146246804755e-06, "loss": 0.6929, "step": 146 }, { "epoch": 0.12257661038148843, "grad_norm": 0.07667449861764908, "learning_rate": 1.9991087292004273e-06, "loss": 0.67, "step": 147 }, { "epoch": 0.12341046487387951, "grad_norm": 0.08285968005657196, "learning_rate": 1.9990704052505284e-06, "loss": 0.7483, "step": 148 }, { "epoch": 0.12424431936627059, "grad_norm": 0.07476554065942764, "learning_rate": 1.9990312749859887e-06, "loss": 0.6846, "step": 149 }, { "epoch": 0.12507817385866166, "grad_norm": 0.0818118005990982, "learning_rate": 1.998991338438388e-06, "loss": 0.6709, "step": 150 }, { "epoch": 0.12591202835105275, "grad_norm": 0.08350253850221634, "learning_rate": 1.998950595639958e-06, "loss": 0.7379, "step": 151 }, { "epoch": 0.12674588284344382, "grad_norm": 0.07916589826345444, "learning_rate": 1.9989090466235806e-06, "loss": 0.7206, "step": 152 }, { "epoch": 0.1275797373358349, "grad_norm": 0.07860454171895981, "learning_rate": 1.998866691422789e-06, "loss": 0.6777, "step": 153 }, { "epoch": 0.12841359182822598, "grad_norm": 0.07861131429672241, "learning_rate": 1.998823530071766e-06, "loss": 0.6938, "step": 154 }, { "epoch": 0.12924744632061705, "grad_norm": 0.07819496840238571, "learning_rate": 1.9987795626053465e-06, "loss": 0.7193, "step": 155 }, { "epoch": 0.13008130081300814, "grad_norm": 0.07828415185213089, "learning_rate": 1.9987347890590144e-06, "loss": 0.6809, "step": 156 }, { "epoch": 0.1309151553053992, "grad_norm": 0.07755222916603088, "learning_rate": 1.998689209468905e-06, "loss": 0.7213, "step": 157 }, { "epoch": 0.13174900979779028, "grad_norm": 0.07672612369060516, "learning_rate": 1.998642823871804e-06, "loss": 0.7196, "step": 158 }, { "epoch": 0.13258286429018137, "grad_norm": 0.07963719964027405, "learning_rate": 1.9985956323051475e-06, "loss": 0.7465, "step": 159 }, { "epoch": 0.13341671878257244, "grad_norm": 0.0795946940779686, "learning_rate": 1.9985476348070223e-06, "loss": 0.717, "step": 160 }, { "epoch": 0.1342505732749635, "grad_norm": 0.07344063371419907, "learning_rate": 1.9984988314161656e-06, "loss": 0.6861, "step": 161 }, { "epoch": 0.1350844277673546, "grad_norm": 0.07722171396017075, "learning_rate": 1.9984492221719645e-06, "loss": 0.6825, "step": 162 }, { "epoch": 0.13591828225974567, "grad_norm": 0.0812341496348381, "learning_rate": 1.9983988071144572e-06, "loss": 0.7245, "step": 163 }, { "epoch": 0.13675213675213677, "grad_norm": 0.08472032845020294, "learning_rate": 1.998347586284332e-06, "loss": 0.7315, "step": 164 }, { "epoch": 0.13758599124452783, "grad_norm": 0.07785464823246002, "learning_rate": 1.9982955597229275e-06, "loss": 0.6906, "step": 165 }, { "epoch": 0.1384198457369189, "grad_norm": 0.07382892072200775, "learning_rate": 1.998242727472232e-06, "loss": 0.6488, "step": 166 }, { "epoch": 0.13925370022931, "grad_norm": 0.07445723563432693, "learning_rate": 1.9981890895748845e-06, "loss": 0.692, "step": 167 }, { "epoch": 0.14008755472170106, "grad_norm": 0.07889281213283539, "learning_rate": 1.998134646074175e-06, "loss": 0.6438, "step": 168 }, { "epoch": 0.14092140921409213, "grad_norm": 0.07664698362350464, "learning_rate": 1.9980793970140426e-06, "loss": 0.6993, "step": 169 }, { "epoch": 0.14175526370648323, "grad_norm": 0.07807335257530212, "learning_rate": 1.9980233424390773e-06, "loss": 0.7334, "step": 170 }, { "epoch": 0.1425891181988743, "grad_norm": 0.0683249682188034, "learning_rate": 1.9979664823945174e-06, "loss": 0.6325, "step": 171 }, { "epoch": 0.1434229726912654, "grad_norm": 0.08339341729879379, "learning_rate": 1.997908816926254e-06, "loss": 0.697, "step": 172 }, { "epoch": 0.14425682718365646, "grad_norm": 0.07781082391738892, "learning_rate": 1.997850346080827e-06, "loss": 0.6748, "step": 173 }, { "epoch": 0.14509068167604752, "grad_norm": 0.07718750089406967, "learning_rate": 1.9977910699054247e-06, "loss": 0.7002, "step": 174 }, { "epoch": 0.14592453616843862, "grad_norm": 0.07779830694198608, "learning_rate": 1.997730988447888e-06, "loss": 0.723, "step": 175 }, { "epoch": 0.14675839066082969, "grad_norm": 0.07958701252937317, "learning_rate": 1.997670101756706e-06, "loss": 0.6651, "step": 176 }, { "epoch": 0.14759224515322075, "grad_norm": 0.08142640441656113, "learning_rate": 1.997608409881019e-06, "loss": 0.6652, "step": 177 }, { "epoch": 0.14842609964561185, "grad_norm": 0.07898704707622528, "learning_rate": 1.9975459128706152e-06, "loss": 0.6842, "step": 178 }, { "epoch": 0.14925995413800291, "grad_norm": 0.07870691269636154, "learning_rate": 1.997482610775935e-06, "loss": 0.7167, "step": 179 }, { "epoch": 0.150093808630394, "grad_norm": 0.07766727358102798, "learning_rate": 1.9974185036480658e-06, "loss": 0.6145, "step": 180 }, { "epoch": 0.15092766312278508, "grad_norm": 0.07913653552532196, "learning_rate": 1.997353591538748e-06, "loss": 0.7598, "step": 181 }, { "epoch": 0.15176151761517614, "grad_norm": 0.07850881665945053, "learning_rate": 1.9972878745003684e-06, "loss": 0.7492, "step": 182 }, { "epoch": 0.15259537210756724, "grad_norm": 0.07599500566720963, "learning_rate": 1.9972213525859656e-06, "loss": 0.6802, "step": 183 }, { "epoch": 0.1534292265999583, "grad_norm": 0.08165914565324783, "learning_rate": 1.997154025849227e-06, "loss": 0.6343, "step": 184 }, { "epoch": 0.15426308109234937, "grad_norm": 0.0763566642999649, "learning_rate": 1.9970858943444893e-06, "loss": 0.7214, "step": 185 }, { "epoch": 0.15509693558474047, "grad_norm": 0.07993835210800171, "learning_rate": 1.99701695812674e-06, "loss": 0.6975, "step": 186 }, { "epoch": 0.15593079007713154, "grad_norm": 0.07139826565980911, "learning_rate": 1.996947217251614e-06, "loss": 0.6714, "step": 187 }, { "epoch": 0.15676464456952263, "grad_norm": 0.0777943804860115, "learning_rate": 1.996876671775397e-06, "loss": 0.6684, "step": 188 }, { "epoch": 0.1575984990619137, "grad_norm": 0.07822203636169434, "learning_rate": 1.9968053217550242e-06, "loss": 0.7161, "step": 189 }, { "epoch": 0.15843235355430477, "grad_norm": 0.07645302265882492, "learning_rate": 1.9967331672480794e-06, "loss": 0.647, "step": 190 }, { "epoch": 0.15926620804669586, "grad_norm": 0.07744666188955307, "learning_rate": 1.996660208312796e-06, "loss": 0.6887, "step": 191 }, { "epoch": 0.16010006253908693, "grad_norm": 0.07616739720106125, "learning_rate": 1.9965864450080565e-06, "loss": 0.7209, "step": 192 }, { "epoch": 0.160933917031478, "grad_norm": 0.0808212161064148, "learning_rate": 1.996511877393393e-06, "loss": 0.7229, "step": 193 }, { "epoch": 0.1617677715238691, "grad_norm": 0.07879424095153809, "learning_rate": 1.996436505528986e-06, "loss": 0.6753, "step": 194 }, { "epoch": 0.16260162601626016, "grad_norm": 0.07811419665813446, "learning_rate": 1.9963603294756654e-06, "loss": 0.6427, "step": 195 }, { "epoch": 0.16343548050865125, "grad_norm": 0.08347994834184647, "learning_rate": 1.996283349294911e-06, "loss": 0.6575, "step": 196 }, { "epoch": 0.16426933500104232, "grad_norm": 0.08415097743272781, "learning_rate": 1.99620556504885e-06, "loss": 0.6859, "step": 197 }, { "epoch": 0.1651031894934334, "grad_norm": 0.08112502098083496, "learning_rate": 1.9961269768002596e-06, "loss": 0.6601, "step": 198 }, { "epoch": 0.16593704398582448, "grad_norm": 0.07908611744642258, "learning_rate": 1.9960475846125656e-06, "loss": 0.6697, "step": 199 }, { "epoch": 0.16677089847821555, "grad_norm": 0.07735829800367355, "learning_rate": 1.995967388549843e-06, "loss": 0.6631, "step": 200 }, { "epoch": 0.16760475297060662, "grad_norm": 0.07564505934715271, "learning_rate": 1.9958863886768145e-06, "loss": 0.7004, "step": 201 }, { "epoch": 0.1684386074629977, "grad_norm": 0.08117896318435669, "learning_rate": 1.9958045850588527e-06, "loss": 0.62, "step": 202 }, { "epoch": 0.16927246195538878, "grad_norm": 0.0761483758687973, "learning_rate": 1.9957219777619784e-06, "loss": 0.7145, "step": 203 }, { "epoch": 0.17010631644777988, "grad_norm": 0.08249073475599289, "learning_rate": 1.995638566852861e-06, "loss": 0.703, "step": 204 }, { "epoch": 0.17094017094017094, "grad_norm": 0.08197219669818878, "learning_rate": 1.9955543523988187e-06, "loss": 0.695, "step": 205 }, { "epoch": 0.171774025432562, "grad_norm": 0.07933253049850464, "learning_rate": 1.995469334467818e-06, "loss": 0.6575, "step": 206 }, { "epoch": 0.1726078799249531, "grad_norm": 0.07888033986091614, "learning_rate": 1.9953835131284735e-06, "loss": 0.6504, "step": 207 }, { "epoch": 0.17344173441734417, "grad_norm": 0.080783411860466, "learning_rate": 1.995296888450049e-06, "loss": 0.6586, "step": 208 }, { "epoch": 0.17427558890973524, "grad_norm": 0.07863239198923111, "learning_rate": 1.995209460502456e-06, "loss": 0.7204, "step": 209 }, { "epoch": 0.17510944340212634, "grad_norm": 0.07864314317703247, "learning_rate": 1.9951212293562544e-06, "loss": 0.6867, "step": 210 }, { "epoch": 0.1759432978945174, "grad_norm": 0.07574496418237686, "learning_rate": 1.995032195082653e-06, "loss": 0.6806, "step": 211 }, { "epoch": 0.1767771523869085, "grad_norm": 0.08079942315816879, "learning_rate": 1.9949423577535078e-06, "loss": 0.6929, "step": 212 }, { "epoch": 0.17761100687929957, "grad_norm": 0.08361431211233139, "learning_rate": 1.9948517174413235e-06, "loss": 0.6948, "step": 213 }, { "epoch": 0.17844486137169063, "grad_norm": 0.07974692434072495, "learning_rate": 1.994760274219253e-06, "loss": 0.7269, "step": 214 }, { "epoch": 0.17927871586408173, "grad_norm": 0.08275672793388367, "learning_rate": 1.9946680281610957e-06, "loss": 0.6268, "step": 215 }, { "epoch": 0.1801125703564728, "grad_norm": 0.08361789584159851, "learning_rate": 1.9945749793413017e-06, "loss": 0.6145, "step": 216 }, { "epoch": 0.18094642484886386, "grad_norm": 0.0829552635550499, "learning_rate": 1.9944811278349665e-06, "loss": 0.7203, "step": 217 }, { "epoch": 0.18178027934125496, "grad_norm": 0.07961665093898773, "learning_rate": 1.994386473717835e-06, "loss": 0.6598, "step": 218 }, { "epoch": 0.18261413383364603, "grad_norm": 0.08015939593315125, "learning_rate": 1.9942910170662986e-06, "loss": 0.6465, "step": 219 }, { "epoch": 0.18344798832603712, "grad_norm": 0.08055105805397034, "learning_rate": 1.994194757957397e-06, "loss": 0.6543, "step": 220 }, { "epoch": 0.1842818428184282, "grad_norm": 0.07763723284006119, "learning_rate": 1.994097696468818e-06, "loss": 0.6984, "step": 221 }, { "epoch": 0.18511569731081925, "grad_norm": 0.07882793247699738, "learning_rate": 1.9939998326788966e-06, "loss": 0.6672, "step": 222 }, { "epoch": 0.18594955180321035, "grad_norm": 0.08098764717578888, "learning_rate": 1.993901166666615e-06, "loss": 0.65, "step": 223 }, { "epoch": 0.18678340629560142, "grad_norm": 0.07791118323802948, "learning_rate": 1.9938016985116027e-06, "loss": 0.6954, "step": 224 }, { "epoch": 0.18761726078799248, "grad_norm": 0.07814016193151474, "learning_rate": 1.993701428294137e-06, "loss": 0.6788, "step": 225 }, { "epoch": 0.18845111528038358, "grad_norm": 0.07884248346090317, "learning_rate": 1.9936003560951433e-06, "loss": 0.6577, "step": 226 }, { "epoch": 0.18928496977277465, "grad_norm": 0.07807621359825134, "learning_rate": 1.9934984819961924e-06, "loss": 0.6424, "step": 227 }, { "epoch": 0.19011882426516574, "grad_norm": 0.08191139996051788, "learning_rate": 1.9933958060795043e-06, "loss": 0.6809, "step": 228 }, { "epoch": 0.1909526787575568, "grad_norm": 0.07597807049751282, "learning_rate": 1.9932923284279445e-06, "loss": 0.7109, "step": 229 }, { "epoch": 0.19178653324994788, "grad_norm": 0.08211328089237213, "learning_rate": 1.9931880491250263e-06, "loss": 0.6916, "step": 230 }, { "epoch": 0.19262038774233897, "grad_norm": 0.07775567471981049, "learning_rate": 1.993082968254909e-06, "loss": 0.6458, "step": 231 }, { "epoch": 0.19345424223473004, "grad_norm": 0.08438362181186676, "learning_rate": 1.992977085902402e-06, "loss": 0.6946, "step": 232 }, { "epoch": 0.1942880967271211, "grad_norm": 0.08631843328475952, "learning_rate": 1.9928704021529564e-06, "loss": 0.6643, "step": 233 }, { "epoch": 0.1951219512195122, "grad_norm": 0.07766929268836975, "learning_rate": 1.992762917092675e-06, "loss": 0.6753, "step": 234 }, { "epoch": 0.19595580571190327, "grad_norm": 0.07574877887964249, "learning_rate": 1.9926546308083044e-06, "loss": 0.6346, "step": 235 }, { "epoch": 0.19678966020429436, "grad_norm": 0.07815206050872803, "learning_rate": 1.9925455433872386e-06, "loss": 0.6946, "step": 236 }, { "epoch": 0.19762351469668543, "grad_norm": 0.07457253336906433, "learning_rate": 1.9924356549175185e-06, "loss": 0.7108, "step": 237 }, { "epoch": 0.1984573691890765, "grad_norm": 0.07688765227794647, "learning_rate": 1.992324965487831e-06, "loss": 0.6764, "step": 238 }, { "epoch": 0.1992912236814676, "grad_norm": 0.08050046861171722, "learning_rate": 1.99221347518751e-06, "loss": 0.6263, "step": 239 }, { "epoch": 0.20012507817385866, "grad_norm": 0.07763144373893738, "learning_rate": 1.9921011841065347e-06, "loss": 0.6618, "step": 240 }, { "epoch": 0.20095893266624973, "grad_norm": 0.07744456082582474, "learning_rate": 1.991988092335532e-06, "loss": 0.6922, "step": 241 }, { "epoch": 0.20179278715864082, "grad_norm": 0.07596256583929062, "learning_rate": 1.9918741999657743e-06, "loss": 0.7267, "step": 242 }, { "epoch": 0.2026266416510319, "grad_norm": 0.08003415167331696, "learning_rate": 1.9917595070891793e-06, "loss": 0.6796, "step": 243 }, { "epoch": 0.203460496143423, "grad_norm": 0.08418845385313034, "learning_rate": 1.9916440137983125e-06, "loss": 0.6815, "step": 244 }, { "epoch": 0.20429435063581405, "grad_norm": 0.0794699415564537, "learning_rate": 1.991527720186384e-06, "loss": 0.6962, "step": 245 }, { "epoch": 0.20512820512820512, "grad_norm": 0.07894127070903778, "learning_rate": 1.991410626347251e-06, "loss": 0.6898, "step": 246 }, { "epoch": 0.20596205962059622, "grad_norm": 0.0819648951292038, "learning_rate": 1.9912927323754145e-06, "loss": 0.726, "step": 247 }, { "epoch": 0.20679591411298728, "grad_norm": 0.08042652159929276, "learning_rate": 1.991174038366024e-06, "loss": 0.6398, "step": 248 }, { "epoch": 0.20762976860537835, "grad_norm": 0.07848809659481049, "learning_rate": 1.9910545444148722e-06, "loss": 0.6562, "step": 249 }, { "epoch": 0.20846362309776945, "grad_norm": 0.08295590430498123, "learning_rate": 1.9909342506183987e-06, "loss": 0.6767, "step": 250 }, { "epoch": 0.2092974775901605, "grad_norm": 0.0840587317943573, "learning_rate": 1.9908131570736887e-06, "loss": 0.7274, "step": 251 }, { "epoch": 0.2101313320825516, "grad_norm": 0.08375398814678192, "learning_rate": 1.9906912638784727e-06, "loss": 0.6716, "step": 252 }, { "epoch": 0.21096518657494268, "grad_norm": 0.08229872584342957, "learning_rate": 1.990568571131126e-06, "loss": 0.7385, "step": 253 }, { "epoch": 0.21179904106733374, "grad_norm": 0.07736154645681381, "learning_rate": 1.99044507893067e-06, "loss": 0.71, "step": 254 }, { "epoch": 0.21263289555972484, "grad_norm": 0.08812221139669418, "learning_rate": 1.9903207873767704e-06, "loss": 0.7045, "step": 255 }, { "epoch": 0.2134667500521159, "grad_norm": 0.08660853654146194, "learning_rate": 1.9901956965697386e-06, "loss": 0.7218, "step": 256 }, { "epoch": 0.21430060454450697, "grad_norm": 0.08562801033258438, "learning_rate": 1.9900698066105318e-06, "loss": 0.6376, "step": 257 }, { "epoch": 0.21513445903689807, "grad_norm": 0.07764414697885513, "learning_rate": 1.9899431176007503e-06, "loss": 0.6629, "step": 258 }, { "epoch": 0.21596831352928914, "grad_norm": 0.07547292113304138, "learning_rate": 1.989815629642641e-06, "loss": 0.6887, "step": 259 }, { "epoch": 0.21680216802168023, "grad_norm": 0.08463111519813538, "learning_rate": 1.9896873428390948e-06, "loss": 0.6169, "step": 260 }, { "epoch": 0.2176360225140713, "grad_norm": 0.08064530789852142, "learning_rate": 1.9895582572936473e-06, "loss": 0.6905, "step": 261 }, { "epoch": 0.21846987700646237, "grad_norm": 0.07929642498493195, "learning_rate": 1.9894283731104796e-06, "loss": 0.6415, "step": 262 }, { "epoch": 0.21930373149885346, "grad_norm": 0.07858303934335709, "learning_rate": 1.9892976903944154e-06, "loss": 0.6955, "step": 263 }, { "epoch": 0.22013758599124453, "grad_norm": 0.07967871427536011, "learning_rate": 1.9891662092509255e-06, "loss": 0.7004, "step": 264 }, { "epoch": 0.2209714404836356, "grad_norm": 0.08009492605924606, "learning_rate": 1.989033929786123e-06, "loss": 0.6206, "step": 265 }, { "epoch": 0.2218052949760267, "grad_norm": 0.08245829492807388, "learning_rate": 1.988900852106766e-06, "loss": 0.7182, "step": 266 }, { "epoch": 0.22263914946841776, "grad_norm": 0.07844138890504837, "learning_rate": 1.9887669763202568e-06, "loss": 0.6372, "step": 267 }, { "epoch": 0.22347300396080885, "grad_norm": 0.0783441960811615, "learning_rate": 1.9886323025346417e-06, "loss": 0.6182, "step": 268 }, { "epoch": 0.22430685845319992, "grad_norm": 0.0803346335887909, "learning_rate": 1.9884968308586117e-06, "loss": 0.6436, "step": 269 }, { "epoch": 0.225140712945591, "grad_norm": 0.0813342034816742, "learning_rate": 1.9883605614015015e-06, "loss": 0.6472, "step": 270 }, { "epoch": 0.22597456743798208, "grad_norm": 0.07891476154327393, "learning_rate": 1.988223494273288e-06, "loss": 0.6625, "step": 271 }, { "epoch": 0.22680842193037315, "grad_norm": 0.08087664842605591, "learning_rate": 1.9880856295845945e-06, "loss": 0.7226, "step": 272 }, { "epoch": 0.22764227642276422, "grad_norm": 0.08254389464855194, "learning_rate": 1.9879469674466865e-06, "loss": 0.6992, "step": 273 }, { "epoch": 0.2284761309151553, "grad_norm": 0.08207474648952484, "learning_rate": 1.987807507971473e-06, "loss": 0.6827, "step": 274 }, { "epoch": 0.22930998540754638, "grad_norm": 0.08394856005907059, "learning_rate": 1.9876672512715075e-06, "loss": 0.7445, "step": 275 }, { "epoch": 0.23014383989993747, "grad_norm": 0.07934897392988205, "learning_rate": 1.9875261974599856e-06, "loss": 0.6988, "step": 276 }, { "epoch": 0.23097769439232854, "grad_norm": 0.08756977319717407, "learning_rate": 1.987384346650747e-06, "loss": 0.6588, "step": 277 }, { "epoch": 0.2318115488847196, "grad_norm": 0.08198294788599014, "learning_rate": 1.9872416989582747e-06, "loss": 0.6729, "step": 278 }, { "epoch": 0.2326454033771107, "grad_norm": 0.07671192288398743, "learning_rate": 1.9870982544976948e-06, "loss": 0.6685, "step": 279 }, { "epoch": 0.23347925786950177, "grad_norm": 0.07943634688854218, "learning_rate": 1.986954013384776e-06, "loss": 0.7352, "step": 280 }, { "epoch": 0.23431311236189284, "grad_norm": 0.0829697921872139, "learning_rate": 1.98680897573593e-06, "loss": 0.6645, "step": 281 }, { "epoch": 0.23514696685428393, "grad_norm": 0.08583538234233856, "learning_rate": 1.986663141668212e-06, "loss": 0.713, "step": 282 }, { "epoch": 0.235980821346675, "grad_norm": 0.08206543326377869, "learning_rate": 1.9865165112993192e-06, "loss": 0.7211, "step": 283 }, { "epoch": 0.2368146758390661, "grad_norm": 0.07899665087461472, "learning_rate": 1.9863690847475924e-06, "loss": 0.6757, "step": 284 }, { "epoch": 0.23764853033145716, "grad_norm": 0.0813852995634079, "learning_rate": 1.986220862132014e-06, "loss": 0.692, "step": 285 }, { "epoch": 0.23848238482384823, "grad_norm": 0.09079395979642868, "learning_rate": 1.9860718435722093e-06, "loss": 0.6756, "step": 286 }, { "epoch": 0.23931623931623933, "grad_norm": 0.07518535107374191, "learning_rate": 1.985922029188446e-06, "loss": 0.626, "step": 287 }, { "epoch": 0.2401500938086304, "grad_norm": 0.08228084444999695, "learning_rate": 1.9857714191016337e-06, "loss": 0.7394, "step": 288 }, { "epoch": 0.24098394830102146, "grad_norm": 0.08036590367555618, "learning_rate": 1.985620013433325e-06, "loss": 0.6012, "step": 289 }, { "epoch": 0.24181780279341256, "grad_norm": 0.08429574221372604, "learning_rate": 1.985467812305714e-06, "loss": 0.6881, "step": 290 }, { "epoch": 0.24265165728580362, "grad_norm": 0.07866602391004562, "learning_rate": 1.9853148158416364e-06, "loss": 0.6708, "step": 291 }, { "epoch": 0.24348551177819472, "grad_norm": 0.0796622484922409, "learning_rate": 1.9851610241645714e-06, "loss": 0.7129, "step": 292 }, { "epoch": 0.24431936627058579, "grad_norm": 0.08136799931526184, "learning_rate": 1.9850064373986374e-06, "loss": 0.6806, "step": 293 }, { "epoch": 0.24515322076297685, "grad_norm": 0.07808342576026917, "learning_rate": 1.9848510556685973e-06, "loss": 0.6217, "step": 294 }, { "epoch": 0.24598707525536795, "grad_norm": 0.07293733209371567, "learning_rate": 1.984694879099853e-06, "loss": 0.662, "step": 295 }, { "epoch": 0.24682092974775902, "grad_norm": 0.07730741053819656, "learning_rate": 1.98453790781845e-06, "loss": 0.6386, "step": 296 }, { "epoch": 0.24765478424015008, "grad_norm": 0.08601437509059906, "learning_rate": 1.9843801419510743e-06, "loss": 0.6639, "step": 297 }, { "epoch": 0.24848863873254118, "grad_norm": 0.0815718024969101, "learning_rate": 1.9842215816250525e-06, "loss": 0.6891, "step": 298 }, { "epoch": 0.24932249322493225, "grad_norm": 0.08793067187070847, "learning_rate": 1.9840622269683537e-06, "loss": 0.684, "step": 299 }, { "epoch": 0.2501563477173233, "grad_norm": 0.07922165095806122, "learning_rate": 1.9839020781095873e-06, "loss": 0.6913, "step": 300 }, { "epoch": 0.2509902022097144, "grad_norm": 0.0789838656783104, "learning_rate": 1.9837411351780035e-06, "loss": 0.6556, "step": 301 }, { "epoch": 0.2518240567021055, "grad_norm": 0.07958388328552246, "learning_rate": 1.9835793983034944e-06, "loss": 0.6301, "step": 302 }, { "epoch": 0.25265791119449654, "grad_norm": 0.08036024868488312, "learning_rate": 1.9834168676165915e-06, "loss": 0.592, "step": 303 }, { "epoch": 0.25349176568688764, "grad_norm": 0.08134333044290543, "learning_rate": 1.9832535432484677e-06, "loss": 0.7142, "step": 304 }, { "epoch": 0.25432562017927873, "grad_norm": 0.08333199471235275, "learning_rate": 1.983089425330937e-06, "loss": 0.7021, "step": 305 }, { "epoch": 0.2551594746716698, "grad_norm": 0.08486991375684738, "learning_rate": 1.982924513996452e-06, "loss": 0.6724, "step": 306 }, { "epoch": 0.25599332916406087, "grad_norm": 0.08286510407924652, "learning_rate": 1.982758809378108e-06, "loss": 0.6856, "step": 307 }, { "epoch": 0.25682718365645196, "grad_norm": 0.08175141364336014, "learning_rate": 1.982592311609639e-06, "loss": 0.6605, "step": 308 }, { "epoch": 0.257661038148843, "grad_norm": 0.08094872534275055, "learning_rate": 1.982425020825419e-06, "loss": 0.7136, "step": 309 }, { "epoch": 0.2584948926412341, "grad_norm": 0.07797659933567047, "learning_rate": 1.9822569371604632e-06, "loss": 0.6975, "step": 310 }, { "epoch": 0.2593287471336252, "grad_norm": 0.08523422479629517, "learning_rate": 1.982088060750426e-06, "loss": 0.6487, "step": 311 }, { "epoch": 0.2601626016260163, "grad_norm": 0.09064790606498718, "learning_rate": 1.981918391731601e-06, "loss": 0.7006, "step": 312 }, { "epoch": 0.2609964561184073, "grad_norm": 0.08136545121669769, "learning_rate": 1.9817479302409225e-06, "loss": 0.67, "step": 313 }, { "epoch": 0.2618303106107984, "grad_norm": 0.07847245037555695, "learning_rate": 1.9815766764159642e-06, "loss": 0.6692, "step": 314 }, { "epoch": 0.2626641651031895, "grad_norm": 0.07690361142158508, "learning_rate": 1.9814046303949387e-06, "loss": 0.5999, "step": 315 }, { "epoch": 0.26349801959558056, "grad_norm": 0.08166715502738953, "learning_rate": 1.9812317923166987e-06, "loss": 0.6666, "step": 316 }, { "epoch": 0.26433187408797165, "grad_norm": 0.08074827492237091, "learning_rate": 1.981058162320735e-06, "loss": 0.657, "step": 317 }, { "epoch": 0.26516572858036275, "grad_norm": 0.08311966806650162, "learning_rate": 1.980883740547179e-06, "loss": 0.6679, "step": 318 }, { "epoch": 0.2659995830727538, "grad_norm": 0.08272409439086914, "learning_rate": 1.9807085271368e-06, "loss": 0.6467, "step": 319 }, { "epoch": 0.2668334375651449, "grad_norm": 0.08190209418535233, "learning_rate": 1.980532522231007e-06, "loss": 0.623, "step": 320 }, { "epoch": 0.267667292057536, "grad_norm": 0.0819978415966034, "learning_rate": 1.980355725971847e-06, "loss": 0.677, "step": 321 }, { "epoch": 0.268501146549927, "grad_norm": 0.07838036119937897, "learning_rate": 1.980178138502006e-06, "loss": 0.6545, "step": 322 }, { "epoch": 0.2693350010423181, "grad_norm": 0.08094768226146698, "learning_rate": 1.979999759964809e-06, "loss": 0.6429, "step": 323 }, { "epoch": 0.2701688555347092, "grad_norm": 0.0796407088637352, "learning_rate": 1.9798205905042184e-06, "loss": 0.7114, "step": 324 }, { "epoch": 0.27100271002710025, "grad_norm": 0.07667295634746552, "learning_rate": 1.9796406302648366e-06, "loss": 0.7059, "step": 325 }, { "epoch": 0.27183656451949134, "grad_norm": 0.07456690818071365, "learning_rate": 1.9794598793919023e-06, "loss": 0.5275, "step": 326 }, { "epoch": 0.27267041901188244, "grad_norm": 0.08084844797849655, "learning_rate": 1.9792783380312935e-06, "loss": 0.6449, "step": 327 }, { "epoch": 0.27350427350427353, "grad_norm": 0.08692745119333267, "learning_rate": 1.9790960063295254e-06, "loss": 0.6831, "step": 328 }, { "epoch": 0.27433812799666457, "grad_norm": 0.0852961540222168, "learning_rate": 1.9789128844337524e-06, "loss": 0.6518, "step": 329 }, { "epoch": 0.27517198248905567, "grad_norm": 0.08140264451503754, "learning_rate": 1.978728972491765e-06, "loss": 0.606, "step": 330 }, { "epoch": 0.27600583698144676, "grad_norm": 0.08339618146419525, "learning_rate": 1.9785442706519927e-06, "loss": 0.6792, "step": 331 }, { "epoch": 0.2768396914738378, "grad_norm": 0.08711510896682739, "learning_rate": 1.9783587790635017e-06, "loss": 0.6903, "step": 332 }, { "epoch": 0.2776735459662289, "grad_norm": 0.08612138777971268, "learning_rate": 1.978172497875995e-06, "loss": 0.6998, "step": 333 }, { "epoch": 0.27850740045862, "grad_norm": 0.07922092080116272, "learning_rate": 1.9779854272398146e-06, "loss": 0.654, "step": 334 }, { "epoch": 0.27934125495101103, "grad_norm": 0.08098642528057098, "learning_rate": 1.9777975673059382e-06, "loss": 0.6706, "step": 335 }, { "epoch": 0.2801751094434021, "grad_norm": 0.0873970165848732, "learning_rate": 1.977608918225981e-06, "loss": 0.6856, "step": 336 }, { "epoch": 0.2810089639357932, "grad_norm": 0.082435242831707, "learning_rate": 1.9774194801521947e-06, "loss": 0.6375, "step": 337 }, { "epoch": 0.28184281842818426, "grad_norm": 0.08199653774499893, "learning_rate": 1.9772292532374687e-06, "loss": 0.6785, "step": 338 }, { "epoch": 0.28267667292057536, "grad_norm": 0.08817990124225616, "learning_rate": 1.9770382376353285e-06, "loss": 0.6642, "step": 339 }, { "epoch": 0.28351052741296645, "grad_norm": 0.08359472453594208, "learning_rate": 1.976846433499935e-06, "loss": 0.6555, "step": 340 }, { "epoch": 0.2843443819053575, "grad_norm": 0.08499454706907272, "learning_rate": 1.976653840986088e-06, "loss": 0.687, "step": 341 }, { "epoch": 0.2851782363977486, "grad_norm": 0.08286130428314209, "learning_rate": 1.976460460249222e-06, "loss": 0.6904, "step": 342 }, { "epoch": 0.2860120908901397, "grad_norm": 0.08244482427835464, "learning_rate": 1.976266291445406e-06, "loss": 0.6546, "step": 343 }, { "epoch": 0.2868459453825308, "grad_norm": 0.07950432598590851, "learning_rate": 1.976071334731349e-06, "loss": 0.6683, "step": 344 }, { "epoch": 0.2876797998749218, "grad_norm": 0.0854944959282875, "learning_rate": 1.975875590264393e-06, "loss": 0.6947, "step": 345 }, { "epoch": 0.2885136543673129, "grad_norm": 0.08143036812543869, "learning_rate": 1.975679058202516e-06, "loss": 0.6422, "step": 346 }, { "epoch": 0.289347508859704, "grad_norm": 0.0797310546040535, "learning_rate": 1.9754817387043325e-06, "loss": 0.6939, "step": 347 }, { "epoch": 0.29018136335209505, "grad_norm": 0.08961952477693558, "learning_rate": 1.9752836319290926e-06, "loss": 0.6848, "step": 348 }, { "epoch": 0.29101521784448614, "grad_norm": 0.08825525641441345, "learning_rate": 1.9750847380366806e-06, "loss": 0.7088, "step": 349 }, { "epoch": 0.29184907233687724, "grad_norm": 0.08601095527410507, "learning_rate": 1.9748850571876168e-06, "loss": 0.6283, "step": 350 }, { "epoch": 0.2926829268292683, "grad_norm": 0.08371152728796005, "learning_rate": 1.9746845895430568e-06, "loss": 0.6755, "step": 351 }, { "epoch": 0.29351678132165937, "grad_norm": 0.08786465227603912, "learning_rate": 1.974483335264791e-06, "loss": 0.6149, "step": 352 }, { "epoch": 0.29435063581405047, "grad_norm": 0.083786241710186, "learning_rate": 1.974281294515245e-06, "loss": 0.7104, "step": 353 }, { "epoch": 0.2951844903064415, "grad_norm": 0.0846792683005333, "learning_rate": 1.9740784674574785e-06, "loss": 0.7032, "step": 354 }, { "epoch": 0.2960183447988326, "grad_norm": 0.08477747440338135, "learning_rate": 1.973874854255186e-06, "loss": 0.6244, "step": 355 }, { "epoch": 0.2968521992912237, "grad_norm": 0.09119998663663864, "learning_rate": 1.9736704550726965e-06, "loss": 0.6373, "step": 356 }, { "epoch": 0.29768605378361473, "grad_norm": 0.08474856615066528, "learning_rate": 1.9734652700749733e-06, "loss": 0.736, "step": 357 }, { "epoch": 0.29851990827600583, "grad_norm": 0.08070901036262512, "learning_rate": 1.9732592994276145e-06, "loss": 0.6442, "step": 358 }, { "epoch": 0.2993537627683969, "grad_norm": 0.08306030929088593, "learning_rate": 1.9730525432968517e-06, "loss": 0.6863, "step": 359 }, { "epoch": 0.300187617260788, "grad_norm": 0.08554356545209885, "learning_rate": 1.97284500184955e-06, "loss": 0.6782, "step": 360 }, { "epoch": 0.30102147175317906, "grad_norm": 0.08350827544927597, "learning_rate": 1.9726366752532096e-06, "loss": 0.6656, "step": 361 }, { "epoch": 0.30185532624557015, "grad_norm": 0.08501608669757843, "learning_rate": 1.972427563675963e-06, "loss": 0.6651, "step": 362 }, { "epoch": 0.30268918073796125, "grad_norm": 0.08401606976985931, "learning_rate": 1.972217667286577e-06, "loss": 0.7153, "step": 363 }, { "epoch": 0.3035230352303523, "grad_norm": 0.08713418990373611, "learning_rate": 1.9720069862544513e-06, "loss": 0.6623, "step": 364 }, { "epoch": 0.3043568897227434, "grad_norm": 0.0871342197060585, "learning_rate": 1.9717955207496197e-06, "loss": 0.6571, "step": 365 }, { "epoch": 0.3051907442151345, "grad_norm": 0.08501847088336945, "learning_rate": 1.9715832709427477e-06, "loss": 0.6714, "step": 366 }, { "epoch": 0.3060245987075255, "grad_norm": 0.07894369214773178, "learning_rate": 1.971370237005136e-06, "loss": 0.6737, "step": 367 }, { "epoch": 0.3068584531999166, "grad_norm": 0.08392394334077835, "learning_rate": 1.9711564191087157e-06, "loss": 0.6807, "step": 368 }, { "epoch": 0.3076923076923077, "grad_norm": 0.08396563678979874, "learning_rate": 1.970941817426052e-06, "loss": 0.6766, "step": 369 }, { "epoch": 0.30852616218469875, "grad_norm": 0.07790801674127579, "learning_rate": 1.9707264321303425e-06, "loss": 0.6293, "step": 370 }, { "epoch": 0.30936001667708984, "grad_norm": 0.08699779212474823, "learning_rate": 1.970510263395417e-06, "loss": 0.681, "step": 371 }, { "epoch": 0.31019387116948094, "grad_norm": 0.08176835626363754, "learning_rate": 1.9702933113957384e-06, "loss": 0.6931, "step": 372 }, { "epoch": 0.311027725661872, "grad_norm": 0.08946527540683746, "learning_rate": 1.9700755763063997e-06, "loss": 0.6062, "step": 373 }, { "epoch": 0.3118615801542631, "grad_norm": 0.08260657638311386, "learning_rate": 1.9698570583031284e-06, "loss": 0.6466, "step": 374 }, { "epoch": 0.31269543464665417, "grad_norm": 0.08951833099126816, "learning_rate": 1.969637757562282e-06, "loss": 0.6458, "step": 375 }, { "epoch": 0.31352928913904526, "grad_norm": 0.08605215698480606, "learning_rate": 1.9694176742608507e-06, "loss": 0.6919, "step": 376 }, { "epoch": 0.3143631436314363, "grad_norm": 0.08389750868082047, "learning_rate": 1.969196808576456e-06, "loss": 0.6941, "step": 377 }, { "epoch": 0.3151969981238274, "grad_norm": 0.08429732173681259, "learning_rate": 1.968975160687351e-06, "loss": 0.622, "step": 378 }, { "epoch": 0.3160308526162185, "grad_norm": 0.08689334988594055, "learning_rate": 1.9687527307724195e-06, "loss": 0.6235, "step": 379 }, { "epoch": 0.31686470710860953, "grad_norm": 0.08494516462087631, "learning_rate": 1.9685295190111774e-06, "loss": 0.6491, "step": 380 }, { "epoch": 0.31769856160100063, "grad_norm": 0.0890749841928482, "learning_rate": 1.9683055255837708e-06, "loss": 0.6791, "step": 381 }, { "epoch": 0.3185324160933917, "grad_norm": 0.07894483208656311, "learning_rate": 1.968080750670977e-06, "loss": 0.7075, "step": 382 }, { "epoch": 0.31936627058578276, "grad_norm": 0.08575093746185303, "learning_rate": 1.9678551944542033e-06, "loss": 0.7027, "step": 383 }, { "epoch": 0.32020012507817386, "grad_norm": 0.08360203355550766, "learning_rate": 1.9676288571154895e-06, "loss": 0.5894, "step": 384 }, { "epoch": 0.32103397957056495, "grad_norm": 0.08341158181428909, "learning_rate": 1.9674017388375036e-06, "loss": 0.658, "step": 385 }, { "epoch": 0.321867834062956, "grad_norm": 0.0844089537858963, "learning_rate": 1.9671738398035446e-06, "loss": 0.6691, "step": 386 }, { "epoch": 0.3227016885553471, "grad_norm": 0.08446256816387177, "learning_rate": 1.9669451601975426e-06, "loss": 0.7211, "step": 387 }, { "epoch": 0.3235355430477382, "grad_norm": 0.08008047193288803, "learning_rate": 1.9667157002040565e-06, "loss": 0.5843, "step": 388 }, { "epoch": 0.3243693975401292, "grad_norm": 0.08821803331375122, "learning_rate": 1.9664854600082754e-06, "loss": 0.6444, "step": 389 }, { "epoch": 0.3252032520325203, "grad_norm": 0.08548033237457275, "learning_rate": 1.966254439796018e-06, "loss": 0.6643, "step": 390 }, { "epoch": 0.3260371065249114, "grad_norm": 0.08261854946613312, "learning_rate": 1.9660226397537322e-06, "loss": 0.6469, "step": 391 }, { "epoch": 0.3268709610173025, "grad_norm": 0.07960055768489838, "learning_rate": 1.965790060068497e-06, "loss": 0.6149, "step": 392 }, { "epoch": 0.32770481550969355, "grad_norm": 0.0834561213850975, "learning_rate": 1.9655567009280177e-06, "loss": 0.6798, "step": 393 }, { "epoch": 0.32853867000208464, "grad_norm": 0.08524155616760254, "learning_rate": 1.9653225625206313e-06, "loss": 0.6999, "step": 394 }, { "epoch": 0.32937252449447574, "grad_norm": 0.08503536134958267, "learning_rate": 1.965087645035302e-06, "loss": 0.6674, "step": 395 }, { "epoch": 0.3302063789868668, "grad_norm": 0.09243112802505493, "learning_rate": 1.964851948661624e-06, "loss": 0.6391, "step": 396 }, { "epoch": 0.3310402334792579, "grad_norm": 0.08015627413988113, "learning_rate": 1.96461547358982e-06, "loss": 0.611, "step": 397 }, { "epoch": 0.33187408797164897, "grad_norm": 0.08681736886501312, "learning_rate": 1.9643782200107395e-06, "loss": 0.6967, "step": 398 }, { "epoch": 0.33270794246404, "grad_norm": 0.08576110005378723, "learning_rate": 1.9641401881158623e-06, "loss": 0.6753, "step": 399 }, { "epoch": 0.3335417969564311, "grad_norm": 0.08432309329509735, "learning_rate": 1.9639013780972954e-06, "loss": 0.6312, "step": 400 }, { "epoch": 0.3335417969564311, "eval_loss": 0.6676867008209229, "eval_runtime": 321.9965, "eval_samples_per_second": 16.718, "eval_steps_per_second": 2.789, "step": 400 }, { "epoch": 0.3343756514488222, "grad_norm": 0.08781838417053223, "learning_rate": 1.9636617901477742e-06, "loss": 0.698, "step": 401 }, { "epoch": 0.33520950594121324, "grad_norm": 0.08723490685224533, "learning_rate": 1.9634214244606613e-06, "loss": 0.6642, "step": 402 }, { "epoch": 0.33604336043360433, "grad_norm": 0.08249269425868988, "learning_rate": 1.963180281229948e-06, "loss": 0.705, "step": 403 }, { "epoch": 0.3368772149259954, "grad_norm": 0.08193965256214142, "learning_rate": 1.962938360650252e-06, "loss": 0.6813, "step": 404 }, { "epoch": 0.33771106941838647, "grad_norm": 0.08125729858875275, "learning_rate": 1.962695662916819e-06, "loss": 0.6571, "step": 405 }, { "epoch": 0.33854492391077756, "grad_norm": 0.08508846163749695, "learning_rate": 1.962452188225522e-06, "loss": 0.6212, "step": 406 }, { "epoch": 0.33937877840316866, "grad_norm": 0.0881478488445282, "learning_rate": 1.962207936772861e-06, "loss": 0.6582, "step": 407 }, { "epoch": 0.34021263289555975, "grad_norm": 0.08948707580566406, "learning_rate": 1.9619629087559622e-06, "loss": 0.6924, "step": 408 }, { "epoch": 0.3410464873879508, "grad_norm": 0.08214866369962692, "learning_rate": 1.9617171043725797e-06, "loss": 0.6266, "step": 409 }, { "epoch": 0.3418803418803419, "grad_norm": 0.09406815469264984, "learning_rate": 1.961470523821093e-06, "loss": 0.6754, "step": 410 }, { "epoch": 0.342714196372733, "grad_norm": 0.08819446712732315, "learning_rate": 1.961223167300509e-06, "loss": 0.6858, "step": 411 }, { "epoch": 0.343548050865124, "grad_norm": 0.0902935117483139, "learning_rate": 1.9609750350104606e-06, "loss": 0.6654, "step": 412 }, { "epoch": 0.3443819053575151, "grad_norm": 0.0852714329957962, "learning_rate": 1.9607261271512065e-06, "loss": 0.6604, "step": 413 }, { "epoch": 0.3452157598499062, "grad_norm": 0.08480080962181091, "learning_rate": 1.9604764439236313e-06, "loss": 0.7252, "step": 414 }, { "epoch": 0.34604961434229725, "grad_norm": 0.08467400074005127, "learning_rate": 1.960225985529246e-06, "loss": 0.6182, "step": 415 }, { "epoch": 0.34688346883468835, "grad_norm": 0.09190120548009872, "learning_rate": 1.9599747521701865e-06, "loss": 0.6605, "step": 416 }, { "epoch": 0.34771732332707944, "grad_norm": 0.08994212001562119, "learning_rate": 1.9597227440492143e-06, "loss": 0.6674, "step": 417 }, { "epoch": 0.3485511778194705, "grad_norm": 0.08530712872743607, "learning_rate": 1.959469961369716e-06, "loss": 0.673, "step": 418 }, { "epoch": 0.3493850323118616, "grad_norm": 0.0814647376537323, "learning_rate": 1.9592164043357046e-06, "loss": 0.6224, "step": 419 }, { "epoch": 0.35021888680425267, "grad_norm": 0.08769946545362473, "learning_rate": 1.9589620731518164e-06, "loss": 0.6446, "step": 420 }, { "epoch": 0.3510527412966437, "grad_norm": 0.0855315700173378, "learning_rate": 1.958706968023313e-06, "loss": 0.6665, "step": 421 }, { "epoch": 0.3518865957890348, "grad_norm": 0.08764835447072983, "learning_rate": 1.958451089156082e-06, "loss": 0.6888, "step": 422 }, { "epoch": 0.3527204502814259, "grad_norm": 0.08674878627061844, "learning_rate": 1.9581944367566324e-06, "loss": 0.6894, "step": 423 }, { "epoch": 0.353554304773817, "grad_norm": 0.07654520124197006, "learning_rate": 1.9579370110321005e-06, "loss": 0.5972, "step": 424 }, { "epoch": 0.35438815926620804, "grad_norm": 0.08671235293149948, "learning_rate": 1.9576788121902454e-06, "loss": 0.6705, "step": 425 }, { "epoch": 0.35522201375859913, "grad_norm": 0.08799296617507935, "learning_rate": 1.9574198404394505e-06, "loss": 0.66, "step": 426 }, { "epoch": 0.3560558682509902, "grad_norm": 0.08908937871456146, "learning_rate": 1.9571600959887224e-06, "loss": 0.6687, "step": 427 }, { "epoch": 0.35688972274338127, "grad_norm": 0.08397683501243591, "learning_rate": 1.9568995790476915e-06, "loss": 0.5915, "step": 428 }, { "epoch": 0.35772357723577236, "grad_norm": 0.08805875480175018, "learning_rate": 1.9566382898266126e-06, "loss": 0.6706, "step": 429 }, { "epoch": 0.35855743172816346, "grad_norm": 0.08281008154153824, "learning_rate": 1.956376228536363e-06, "loss": 0.5692, "step": 430 }, { "epoch": 0.3593912862205545, "grad_norm": 0.08514556288719177, "learning_rate": 1.9561133953884427e-06, "loss": 0.5938, "step": 431 }, { "epoch": 0.3602251407129456, "grad_norm": 0.09112891554832458, "learning_rate": 1.955849790594975e-06, "loss": 0.7044, "step": 432 }, { "epoch": 0.3610589952053367, "grad_norm": 0.08578234165906906, "learning_rate": 1.9555854143687064e-06, "loss": 0.6901, "step": 433 }, { "epoch": 0.3618928496977277, "grad_norm": 0.08405105024576187, "learning_rate": 1.955320266923006e-06, "loss": 0.6307, "step": 434 }, { "epoch": 0.3627267041901188, "grad_norm": 0.085452601313591, "learning_rate": 1.9550543484718644e-06, "loss": 0.6147, "step": 435 }, { "epoch": 0.3635605586825099, "grad_norm": 0.0871538370847702, "learning_rate": 1.9547876592298955e-06, "loss": 0.6415, "step": 436 }, { "epoch": 0.36439441317490096, "grad_norm": 0.08855723589658737, "learning_rate": 1.954520199412334e-06, "loss": 0.7026, "step": 437 }, { "epoch": 0.36522826766729205, "grad_norm": 0.08221881091594696, "learning_rate": 1.9542519692350387e-06, "loss": 0.6813, "step": 438 }, { "epoch": 0.36606212215968315, "grad_norm": 0.08595414459705353, "learning_rate": 1.9539829689144876e-06, "loss": 0.6809, "step": 439 }, { "epoch": 0.36689597665207424, "grad_norm": 0.08485814183950424, "learning_rate": 1.9537131986677816e-06, "loss": 0.6352, "step": 440 }, { "epoch": 0.3677298311444653, "grad_norm": 0.08620929718017578, "learning_rate": 1.9534426587126433e-06, "loss": 0.7072, "step": 441 }, { "epoch": 0.3685636856368564, "grad_norm": 0.0919066071510315, "learning_rate": 1.9531713492674154e-06, "loss": 0.6809, "step": 442 }, { "epoch": 0.36939754012924747, "grad_norm": 0.08376951515674591, "learning_rate": 1.9528992705510628e-06, "loss": 0.6486, "step": 443 }, { "epoch": 0.3702313946216385, "grad_norm": 0.088445745408535, "learning_rate": 1.95262642278317e-06, "loss": 0.709, "step": 444 }, { "epoch": 0.3710652491140296, "grad_norm": 0.08183503895998001, "learning_rate": 1.952352806183943e-06, "loss": 0.6277, "step": 445 }, { "epoch": 0.3718991036064207, "grad_norm": 0.08638439327478409, "learning_rate": 1.9520784209742094e-06, "loss": 0.6665, "step": 446 }, { "epoch": 0.37273295809881174, "grad_norm": 0.08837208896875381, "learning_rate": 1.951803267375414e-06, "loss": 0.6103, "step": 447 }, { "epoch": 0.37356681259120283, "grad_norm": 0.08010973036289215, "learning_rate": 1.9515273456096247e-06, "loss": 0.662, "step": 448 }, { "epoch": 0.37440066708359393, "grad_norm": 0.08395659178495407, "learning_rate": 1.9512506558995283e-06, "loss": 0.687, "step": 449 }, { "epoch": 0.37523452157598497, "grad_norm": 0.08506524562835693, "learning_rate": 1.950973198468431e-06, "loss": 0.6695, "step": 450 }, { "epoch": 0.37606837606837606, "grad_norm": 0.08150207251310349, "learning_rate": 1.950694973540259e-06, "loss": 0.6258, "step": 451 }, { "epoch": 0.37690223056076716, "grad_norm": 0.08509568125009537, "learning_rate": 1.9504159813395576e-06, "loss": 0.5774, "step": 452 }, { "epoch": 0.3777360850531582, "grad_norm": 0.08139500021934509, "learning_rate": 1.950136222091492e-06, "loss": 0.6652, "step": 453 }, { "epoch": 0.3785699395455493, "grad_norm": 0.08891302347183228, "learning_rate": 1.949855696021846e-06, "loss": 0.6602, "step": 454 }, { "epoch": 0.3794037940379404, "grad_norm": 0.08729292452335358, "learning_rate": 1.949574403357022e-06, "loss": 0.6232, "step": 455 }, { "epoch": 0.3802376485303315, "grad_norm": 0.0855722650885582, "learning_rate": 1.9492923443240425e-06, "loss": 0.7249, "step": 456 }, { "epoch": 0.3810715030227225, "grad_norm": 0.09225167334079742, "learning_rate": 1.949009519150546e-06, "loss": 0.7061, "step": 457 }, { "epoch": 0.3819053575151136, "grad_norm": 0.08833880722522736, "learning_rate": 1.9487259280647918e-06, "loss": 0.6683, "step": 458 }, { "epoch": 0.3827392120075047, "grad_norm": 0.08648336678743362, "learning_rate": 1.948441571295656e-06, "loss": 0.6716, "step": 459 }, { "epoch": 0.38357306649989575, "grad_norm": 0.08678124845027924, "learning_rate": 1.9481564490726324e-06, "loss": 0.6869, "step": 460 }, { "epoch": 0.38440692099228685, "grad_norm": 0.0866696834564209, "learning_rate": 1.9478705616258344e-06, "loss": 0.6346, "step": 461 }, { "epoch": 0.38524077548467794, "grad_norm": 0.09009408950805664, "learning_rate": 1.9475839091859904e-06, "loss": 0.6507, "step": 462 }, { "epoch": 0.386074629977069, "grad_norm": 0.09393912553787231, "learning_rate": 1.9472964919844484e-06, "loss": 0.6624, "step": 463 }, { "epoch": 0.3869084844694601, "grad_norm": 0.09663153439760208, "learning_rate": 1.9470083102531722e-06, "loss": 0.6378, "step": 464 }, { "epoch": 0.3877423389618512, "grad_norm": 0.08551555126905441, "learning_rate": 1.9467193642247435e-06, "loss": 0.6717, "step": 465 }, { "epoch": 0.3885761934542422, "grad_norm": 0.08986209332942963, "learning_rate": 1.9464296541323602e-06, "loss": 0.7313, "step": 466 }, { "epoch": 0.3894100479466333, "grad_norm": 0.08851300179958344, "learning_rate": 1.9461391802098373e-06, "loss": 0.6783, "step": 467 }, { "epoch": 0.3902439024390244, "grad_norm": 0.09088604897260666, "learning_rate": 1.9458479426916068e-06, "loss": 0.5952, "step": 468 }, { "epoch": 0.39107775693141544, "grad_norm": 0.08506208658218384, "learning_rate": 1.9455559418127144e-06, "loss": 0.6912, "step": 469 }, { "epoch": 0.39191161142380654, "grad_norm": 0.087041936814785, "learning_rate": 1.945263177808826e-06, "loss": 0.6722, "step": 470 }, { "epoch": 0.39274546591619763, "grad_norm": 0.09510099142789841, "learning_rate": 1.9449696509162193e-06, "loss": 0.697, "step": 471 }, { "epoch": 0.39357932040858873, "grad_norm": 0.0864386260509491, "learning_rate": 1.944675361371791e-06, "loss": 0.6882, "step": 472 }, { "epoch": 0.39441317490097977, "grad_norm": 0.08964356780052185, "learning_rate": 1.944380309413051e-06, "loss": 0.6403, "step": 473 }, { "epoch": 0.39524702939337086, "grad_norm": 0.09786627441644669, "learning_rate": 1.9440844952781253e-06, "loss": 0.7037, "step": 474 }, { "epoch": 0.39608088388576196, "grad_norm": 0.08709276467561722, "learning_rate": 1.9437879192057556e-06, "loss": 0.6619, "step": 475 }, { "epoch": 0.396914738378153, "grad_norm": 0.08677015453577042, "learning_rate": 1.9434905814352976e-06, "loss": 0.6609, "step": 476 }, { "epoch": 0.3977485928705441, "grad_norm": 0.08886487782001495, "learning_rate": 1.9431924822067225e-06, "loss": 0.6045, "step": 477 }, { "epoch": 0.3985824473629352, "grad_norm": 0.08575446903705597, "learning_rate": 1.942893621760616e-06, "loss": 0.6886, "step": 478 }, { "epoch": 0.39941630185532623, "grad_norm": 0.08386967331171036, "learning_rate": 1.9425940003381767e-06, "loss": 0.6721, "step": 479 }, { "epoch": 0.4002501563477173, "grad_norm": 0.08702324330806732, "learning_rate": 1.9422936181812197e-06, "loss": 0.6959, "step": 480 }, { "epoch": 0.4010840108401084, "grad_norm": 0.09090246260166168, "learning_rate": 1.9419924755321727e-06, "loss": 0.7094, "step": 481 }, { "epoch": 0.40191786533249946, "grad_norm": 0.09213659167289734, "learning_rate": 1.9416905726340767e-06, "loss": 0.6446, "step": 482 }, { "epoch": 0.40275171982489055, "grad_norm": 0.09240783751010895, "learning_rate": 1.9413879097305878e-06, "loss": 0.6401, "step": 483 }, { "epoch": 0.40358557431728165, "grad_norm": 0.09058364480733871, "learning_rate": 1.941084487065974e-06, "loss": 0.7035, "step": 484 }, { "epoch": 0.4044194288096727, "grad_norm": 0.0906047448515892, "learning_rate": 1.940780304885117e-06, "loss": 0.6519, "step": 485 }, { "epoch": 0.4052532833020638, "grad_norm": 0.09101078659296036, "learning_rate": 1.940475363433512e-06, "loss": 0.6934, "step": 486 }, { "epoch": 0.4060871377944549, "grad_norm": 0.08729363977909088, "learning_rate": 1.940169662957266e-06, "loss": 0.5822, "step": 487 }, { "epoch": 0.406920992286846, "grad_norm": 0.0946895033121109, "learning_rate": 1.9398632037031003e-06, "loss": 0.6534, "step": 488 }, { "epoch": 0.407754846779237, "grad_norm": 0.09069128334522247, "learning_rate": 1.9395559859183463e-06, "loss": 0.6706, "step": 489 }, { "epoch": 0.4085887012716281, "grad_norm": 0.08318338543176651, "learning_rate": 1.9392480098509483e-06, "loss": 0.6368, "step": 490 }, { "epoch": 0.4094225557640192, "grad_norm": 0.0826449766755104, "learning_rate": 1.9389392757494645e-06, "loss": 0.667, "step": 491 }, { "epoch": 0.41025641025641024, "grad_norm": 0.09276958554983139, "learning_rate": 1.938629783863062e-06, "loss": 0.6709, "step": 492 }, { "epoch": 0.41109026474880134, "grad_norm": 0.08281219005584717, "learning_rate": 1.9383195344415215e-06, "loss": 0.584, "step": 493 }, { "epoch": 0.41192411924119243, "grad_norm": 0.08632051944732666, "learning_rate": 1.938008527735234e-06, "loss": 0.6743, "step": 494 }, { "epoch": 0.41275797373358347, "grad_norm": 0.08970591425895691, "learning_rate": 1.937696763995202e-06, "loss": 0.6611, "step": 495 }, { "epoch": 0.41359182822597457, "grad_norm": 0.0902903825044632, "learning_rate": 1.93738424347304e-06, "loss": 0.6455, "step": 496 }, { "epoch": 0.41442568271836566, "grad_norm": 0.08964331448078156, "learning_rate": 1.9370709664209715e-06, "loss": 0.6561, "step": 497 }, { "epoch": 0.4152595372107567, "grad_norm": 0.08478616923093796, "learning_rate": 1.9367569330918317e-06, "loss": 0.595, "step": 498 }, { "epoch": 0.4160933917031478, "grad_norm": 0.09221872687339783, "learning_rate": 1.9364421437390658e-06, "loss": 0.6363, "step": 499 }, { "epoch": 0.4169272461955389, "grad_norm": 0.09467694163322449, "learning_rate": 1.936126598616729e-06, "loss": 0.6696, "step": 500 }, { "epoch": 0.41776110068792993, "grad_norm": 0.09464087337255478, "learning_rate": 1.935810297979487e-06, "loss": 0.6224, "step": 501 }, { "epoch": 0.418594955180321, "grad_norm": 0.08994987607002258, "learning_rate": 1.9354932420826147e-06, "loss": 0.6383, "step": 502 }, { "epoch": 0.4194288096727121, "grad_norm": 0.0921832025051117, "learning_rate": 1.9351754311819974e-06, "loss": 0.685, "step": 503 }, { "epoch": 0.4202626641651032, "grad_norm": 0.08833765983581543, "learning_rate": 1.934856865534129e-06, "loss": 0.6233, "step": 504 }, { "epoch": 0.42109651865749426, "grad_norm": 0.08863073587417603, "learning_rate": 1.934537545396111e-06, "loss": 0.6756, "step": 505 }, { "epoch": 0.42193037314988535, "grad_norm": 0.08703204244375229, "learning_rate": 1.934217471025658e-06, "loss": 0.6372, "step": 506 }, { "epoch": 0.42276422764227645, "grad_norm": 0.09010464698076248, "learning_rate": 1.9338966426810887e-06, "loss": 0.6677, "step": 507 }, { "epoch": 0.4235980821346675, "grad_norm": 0.09014829248189926, "learning_rate": 1.9335750606213336e-06, "loss": 0.6839, "step": 508 }, { "epoch": 0.4244319366270586, "grad_norm": 0.08871164172887802, "learning_rate": 1.93325272510593e-06, "loss": 0.7012, "step": 509 }, { "epoch": 0.4252657911194497, "grad_norm": 0.08788250386714935, "learning_rate": 1.9329296363950236e-06, "loss": 0.6169, "step": 510 }, { "epoch": 0.4260996456118407, "grad_norm": 0.08845999836921692, "learning_rate": 1.9326057947493675e-06, "loss": 0.6424, "step": 511 }, { "epoch": 0.4269335001042318, "grad_norm": 0.0858997106552124, "learning_rate": 1.9322812004303238e-06, "loss": 0.6389, "step": 512 }, { "epoch": 0.4277673545966229, "grad_norm": 0.08850055187940598, "learning_rate": 1.9319558536998603e-06, "loss": 0.6975, "step": 513 }, { "epoch": 0.42860120908901395, "grad_norm": 0.09360900521278381, "learning_rate": 1.9316297548205534e-06, "loss": 0.686, "step": 514 }, { "epoch": 0.42943506358140504, "grad_norm": 0.08473866432905197, "learning_rate": 1.931302904055586e-06, "loss": 0.6517, "step": 515 }, { "epoch": 0.43026891807379614, "grad_norm": 0.08656331151723862, "learning_rate": 1.9309753016687477e-06, "loss": 0.7003, "step": 516 }, { "epoch": 0.4311027725661872, "grad_norm": 0.08407936990261078, "learning_rate": 1.9306469479244347e-06, "loss": 0.6611, "step": 517 }, { "epoch": 0.43193662705857827, "grad_norm": 0.09026416391134262, "learning_rate": 1.930317843087651e-06, "loss": 0.6749, "step": 518 }, { "epoch": 0.43277048155096937, "grad_norm": 0.10031198710203171, "learning_rate": 1.929987987424004e-06, "loss": 0.6403, "step": 519 }, { "epoch": 0.43360433604336046, "grad_norm": 0.09739411622285843, "learning_rate": 1.9296573811997086e-06, "loss": 0.6893, "step": 520 }, { "epoch": 0.4344381905357515, "grad_norm": 0.08792299032211304, "learning_rate": 1.929326024681587e-06, "loss": 0.6268, "step": 521 }, { "epoch": 0.4352720450281426, "grad_norm": 0.0874318853020668, "learning_rate": 1.9289939181370646e-06, "loss": 0.6659, "step": 522 }, { "epoch": 0.4361058995205337, "grad_norm": 0.09086845815181732, "learning_rate": 1.9286610618341724e-06, "loss": 0.6399, "step": 523 }, { "epoch": 0.43693975401292473, "grad_norm": 0.09616074711084366, "learning_rate": 1.9283274560415477e-06, "loss": 0.6325, "step": 524 }, { "epoch": 0.4377736085053158, "grad_norm": 0.08674857765436172, "learning_rate": 1.9279931010284322e-06, "loss": 0.6487, "step": 525 }, { "epoch": 0.4386074629977069, "grad_norm": 0.09142550826072693, "learning_rate": 1.927657997064671e-06, "loss": 0.6509, "step": 526 }, { "epoch": 0.43944131749009796, "grad_norm": 0.0838690921664238, "learning_rate": 1.9273221444207158e-06, "loss": 0.6087, "step": 527 }, { "epoch": 0.44027517198248906, "grad_norm": 0.0922529399394989, "learning_rate": 1.9269855433676213e-06, "loss": 0.6815, "step": 528 }, { "epoch": 0.44110902647488015, "grad_norm": 0.08945809304714203, "learning_rate": 1.926648194177046e-06, "loss": 0.623, "step": 529 }, { "epoch": 0.4419428809672712, "grad_norm": 0.09663840383291245, "learning_rate": 1.926310097121253e-06, "loss": 0.6244, "step": 530 }, { "epoch": 0.4427767354596623, "grad_norm": 0.0872710794210434, "learning_rate": 1.9259712524731083e-06, "loss": 0.6877, "step": 531 }, { "epoch": 0.4436105899520534, "grad_norm": 0.0860145092010498, "learning_rate": 1.9256316605060813e-06, "loss": 0.6458, "step": 532 }, { "epoch": 0.4444444444444444, "grad_norm": 0.08691411465406418, "learning_rate": 1.9252913214942456e-06, "loss": 0.6286, "step": 533 }, { "epoch": 0.4452782989368355, "grad_norm": 0.0899488553404808, "learning_rate": 1.9249502357122757e-06, "loss": 0.6154, "step": 534 }, { "epoch": 0.4461121534292266, "grad_norm": 0.08670518547296524, "learning_rate": 1.924608403435451e-06, "loss": 0.6189, "step": 535 }, { "epoch": 0.4469460079216177, "grad_norm": 0.08927959948778152, "learning_rate": 1.9242658249396514e-06, "loss": 0.6591, "step": 536 }, { "epoch": 0.44777986241400874, "grad_norm": 0.09092804044485092, "learning_rate": 1.9239225005013607e-06, "loss": 0.6756, "step": 537 }, { "epoch": 0.44861371690639984, "grad_norm": 0.08876843005418777, "learning_rate": 1.923578430397664e-06, "loss": 0.6221, "step": 538 }, { "epoch": 0.44944757139879093, "grad_norm": 0.09023214131593704, "learning_rate": 1.923233614906248e-06, "loss": 0.6478, "step": 539 }, { "epoch": 0.450281425891182, "grad_norm": 0.09291035681962967, "learning_rate": 1.9228880543054006e-06, "loss": 0.6568, "step": 540 }, { "epoch": 0.45111528038357307, "grad_norm": 0.09169748425483704, "learning_rate": 1.9225417488740127e-06, "loss": 0.6966, "step": 541 }, { "epoch": 0.45194913487596416, "grad_norm": 0.0960812196135521, "learning_rate": 1.922194698891574e-06, "loss": 0.6426, "step": 542 }, { "epoch": 0.4527829893683552, "grad_norm": 0.09381034225225449, "learning_rate": 1.9218469046381778e-06, "loss": 0.7138, "step": 543 }, { "epoch": 0.4536168438607463, "grad_norm": 0.09605712443590164, "learning_rate": 1.9214983663945157e-06, "loss": 0.6677, "step": 544 }, { "epoch": 0.4544506983531374, "grad_norm": 0.09246627241373062, "learning_rate": 1.921149084441881e-06, "loss": 0.5995, "step": 545 }, { "epoch": 0.45528455284552843, "grad_norm": 0.09615227580070496, "learning_rate": 1.9207990590621663e-06, "loss": 0.6543, "step": 546 }, { "epoch": 0.45611840733791953, "grad_norm": 0.09620420634746552, "learning_rate": 1.9204482905378654e-06, "loss": 0.5895, "step": 547 }, { "epoch": 0.4569522618303106, "grad_norm": 0.08878765255212784, "learning_rate": 1.920096779152071e-06, "loss": 0.599, "step": 548 }, { "epoch": 0.45778611632270166, "grad_norm": 0.09067462384700775, "learning_rate": 1.9197445251884763e-06, "loss": 0.6897, "step": 549 }, { "epoch": 0.45861997081509276, "grad_norm": 0.09180627018213272, "learning_rate": 1.9193915289313724e-06, "loss": 0.6647, "step": 550 }, { "epoch": 0.45945382530748385, "grad_norm": 0.08824802190065384, "learning_rate": 1.91903779066565e-06, "loss": 0.6376, "step": 551 }, { "epoch": 0.46028767979987495, "grad_norm": 0.08835819363594055, "learning_rate": 1.918683310676799e-06, "loss": 0.6922, "step": 552 }, { "epoch": 0.461121534292266, "grad_norm": 0.09096652269363403, "learning_rate": 1.918328089250908e-06, "loss": 0.7169, "step": 553 }, { "epoch": 0.4619553887846571, "grad_norm": 0.09415201842784882, "learning_rate": 1.917972126674664e-06, "loss": 0.659, "step": 554 }, { "epoch": 0.4627892432770482, "grad_norm": 0.08964891731739044, "learning_rate": 1.9176154232353513e-06, "loss": 0.6869, "step": 555 }, { "epoch": 0.4636230977694392, "grad_norm": 0.08688930422067642, "learning_rate": 1.917257979220853e-06, "loss": 0.627, "step": 556 }, { "epoch": 0.4644569522618303, "grad_norm": 0.09115028381347656, "learning_rate": 1.9168997949196496e-06, "loss": 0.64, "step": 557 }, { "epoch": 0.4652908067542214, "grad_norm": 0.09430664777755737, "learning_rate": 1.9165408706208184e-06, "loss": 0.6646, "step": 558 }, { "epoch": 0.46612466124661245, "grad_norm": 0.09062688797712326, "learning_rate": 1.916181206614036e-06, "loss": 0.6392, "step": 559 }, { "epoch": 0.46695851573900354, "grad_norm": 0.09099205583333969, "learning_rate": 1.9158208031895737e-06, "loss": 0.7133, "step": 560 }, { "epoch": 0.46779237023139464, "grad_norm": 0.09944413602352142, "learning_rate": 1.9154596606383002e-06, "loss": 0.6958, "step": 561 }, { "epoch": 0.4686262247237857, "grad_norm": 0.09643740206956863, "learning_rate": 1.9150977792516816e-06, "loss": 0.6813, "step": 562 }, { "epoch": 0.4694600792161768, "grad_norm": 0.09532240033149719, "learning_rate": 1.914735159321779e-06, "loss": 0.6528, "step": 563 }, { "epoch": 0.47029393370856787, "grad_norm": 0.08927876502275467, "learning_rate": 1.914371801141251e-06, "loss": 0.639, "step": 564 }, { "epoch": 0.4711277882009589, "grad_norm": 0.09561195224523544, "learning_rate": 1.914007705003351e-06, "loss": 0.6795, "step": 565 }, { "epoch": 0.47196164269335, "grad_norm": 0.0882195308804512, "learning_rate": 1.9136428712019275e-06, "loss": 0.6437, "step": 566 }, { "epoch": 0.4727954971857411, "grad_norm": 0.08940508961677551, "learning_rate": 1.913277300031426e-06, "loss": 0.6485, "step": 567 }, { "epoch": 0.4736293516781322, "grad_norm": 0.0933694839477539, "learning_rate": 1.912910991786886e-06, "loss": 0.6876, "step": 568 }, { "epoch": 0.47446320617052323, "grad_norm": 0.09212841838598251, "learning_rate": 1.9125439467639414e-06, "loss": 0.5669, "step": 569 }, { "epoch": 0.47529706066291433, "grad_norm": 0.09037458896636963, "learning_rate": 1.9121761652588214e-06, "loss": 0.6489, "step": 570 }, { "epoch": 0.4761309151553054, "grad_norm": 0.08847213536500931, "learning_rate": 1.9118076475683506e-06, "loss": 0.6456, "step": 571 }, { "epoch": 0.47696476964769646, "grad_norm": 0.08496098965406418, "learning_rate": 1.9114383939899455e-06, "loss": 0.6174, "step": 572 }, { "epoch": 0.47779862414008756, "grad_norm": 0.09430284053087234, "learning_rate": 1.911068404821618e-06, "loss": 0.7102, "step": 573 }, { "epoch": 0.47863247863247865, "grad_norm": 0.09211906790733337, "learning_rate": 1.910697680361974e-06, "loss": 0.6161, "step": 574 }, { "epoch": 0.4794663331248697, "grad_norm": 0.09092090278863907, "learning_rate": 1.910326220910211e-06, "loss": 0.6595, "step": 575 }, { "epoch": 0.4803001876172608, "grad_norm": 0.09773701429367065, "learning_rate": 1.909954026766122e-06, "loss": 0.6303, "step": 576 }, { "epoch": 0.4811340421096519, "grad_norm": 0.09343478083610535, "learning_rate": 1.909581098230091e-06, "loss": 0.6578, "step": 577 }, { "epoch": 0.4819678966020429, "grad_norm": 0.09686136245727539, "learning_rate": 1.9092074356030966e-06, "loss": 0.6504, "step": 578 }, { "epoch": 0.482801751094434, "grad_norm": 0.09207551926374435, "learning_rate": 1.9088330391867076e-06, "loss": 0.6092, "step": 579 }, { "epoch": 0.4836356055868251, "grad_norm": 0.093394935131073, "learning_rate": 1.908457909283087e-06, "loss": 0.645, "step": 580 }, { "epoch": 0.48446946007921615, "grad_norm": 0.10280878096818924, "learning_rate": 1.9080820461949886e-06, "loss": 0.7288, "step": 581 }, { "epoch": 0.48530331457160725, "grad_norm": 0.09714383631944656, "learning_rate": 1.9077054502257585e-06, "loss": 0.6213, "step": 582 }, { "epoch": 0.48613716906399834, "grad_norm": 0.09004565328359604, "learning_rate": 1.9073281216793337e-06, "loss": 0.6459, "step": 583 }, { "epoch": 0.48697102355638944, "grad_norm": 0.12440581619739532, "learning_rate": 1.906950060860243e-06, "loss": 0.6823, "step": 584 }, { "epoch": 0.4878048780487805, "grad_norm": 0.09314379096031189, "learning_rate": 1.9065712680736066e-06, "loss": 0.6475, "step": 585 }, { "epoch": 0.48863873254117157, "grad_norm": 0.09512604027986526, "learning_rate": 1.906191743625134e-06, "loss": 0.5932, "step": 586 }, { "epoch": 0.48947258703356267, "grad_norm": 0.08977750688791275, "learning_rate": 1.9058114878211265e-06, "loss": 0.6546, "step": 587 }, { "epoch": 0.4903064415259537, "grad_norm": 0.09166161715984344, "learning_rate": 1.905430500968475e-06, "loss": 0.6772, "step": 588 }, { "epoch": 0.4911402960183448, "grad_norm": 0.10014794021844864, "learning_rate": 1.9050487833746609e-06, "loss": 0.6945, "step": 589 }, { "epoch": 0.4919741505107359, "grad_norm": 0.09376100450754166, "learning_rate": 1.9046663353477548e-06, "loss": 0.6581, "step": 590 }, { "epoch": 0.49280800500312694, "grad_norm": 0.09505487233400345, "learning_rate": 1.9042831571964171e-06, "loss": 0.6785, "step": 591 }, { "epoch": 0.49364185949551803, "grad_norm": 0.09371201694011688, "learning_rate": 1.9038992492298975e-06, "loss": 0.6331, "step": 592 }, { "epoch": 0.4944757139879091, "grad_norm": 0.09442009776830673, "learning_rate": 1.9035146117580348e-06, "loss": 0.5728, "step": 593 }, { "epoch": 0.49530956848030017, "grad_norm": 0.09772226214408875, "learning_rate": 1.903129245091256e-06, "loss": 0.6635, "step": 594 }, { "epoch": 0.49614342297269126, "grad_norm": 0.09484317898750305, "learning_rate": 1.9027431495405776e-06, "loss": 0.6327, "step": 595 }, { "epoch": 0.49697727746508236, "grad_norm": 0.09658980369567871, "learning_rate": 1.9023563254176032e-06, "loss": 0.6608, "step": 596 }, { "epoch": 0.4978111319574734, "grad_norm": 0.09717069566249847, "learning_rate": 1.9019687730345251e-06, "loss": 0.6605, "step": 597 }, { "epoch": 0.4986449864498645, "grad_norm": 0.09019803255796432, "learning_rate": 1.9015804927041233e-06, "loss": 0.6867, "step": 598 }, { "epoch": 0.4994788409422556, "grad_norm": 0.09377589821815491, "learning_rate": 1.9011914847397653e-06, "loss": 0.6419, "step": 599 }, { "epoch": 0.5003126954346466, "grad_norm": 0.09151824563741684, "learning_rate": 1.9008017494554055e-06, "loss": 0.7184, "step": 600 }, { "epoch": 0.5011465499270378, "grad_norm": 0.0963352620601654, "learning_rate": 1.900411287165586e-06, "loss": 0.6897, "step": 601 }, { "epoch": 0.5019804044194288, "grad_norm": 0.08974531292915344, "learning_rate": 1.9000200981854347e-06, "loss": 0.5847, "step": 602 }, { "epoch": 0.5028142589118199, "grad_norm": 0.09220883995294571, "learning_rate": 1.8996281828306665e-06, "loss": 0.6652, "step": 603 }, { "epoch": 0.503648113404211, "grad_norm": 0.09491855651140213, "learning_rate": 1.899235541417583e-06, "loss": 0.595, "step": 604 }, { "epoch": 0.504481967896602, "grad_norm": 0.09616294503211975, "learning_rate": 1.898842174263071e-06, "loss": 0.6378, "step": 605 }, { "epoch": 0.5053158223889931, "grad_norm": 0.08942185342311859, "learning_rate": 1.8984480816846035e-06, "loss": 0.6459, "step": 606 }, { "epoch": 0.5061496768813842, "grad_norm": 0.09059803187847137, "learning_rate": 1.8980532640002388e-06, "loss": 0.6019, "step": 607 }, { "epoch": 0.5069835313737753, "grad_norm": 0.09913185238838196, "learning_rate": 1.89765772152862e-06, "loss": 0.669, "step": 608 }, { "epoch": 0.5078173858661663, "grad_norm": 0.10223888605833054, "learning_rate": 1.8972614545889756e-06, "loss": 0.6454, "step": 609 }, { "epoch": 0.5086512403585575, "grad_norm": 0.09271295368671417, "learning_rate": 1.896864463501119e-06, "loss": 0.6467, "step": 610 }, { "epoch": 0.5094850948509485, "grad_norm": 0.08862827718257904, "learning_rate": 1.896466748585448e-06, "loss": 0.6343, "step": 611 }, { "epoch": 0.5103189493433395, "grad_norm": 0.09303930401802063, "learning_rate": 1.8960683101629435e-06, "loss": 0.6389, "step": 612 }, { "epoch": 0.5111528038357307, "grad_norm": 0.09331949055194855, "learning_rate": 1.8956691485551719e-06, "loss": 0.6507, "step": 613 }, { "epoch": 0.5119866583281217, "grad_norm": 0.0993635281920433, "learning_rate": 1.8952692640842825e-06, "loss": 0.63, "step": 614 }, { "epoch": 0.5128205128205128, "grad_norm": 0.09348301589488983, "learning_rate": 1.8948686570730074e-06, "loss": 0.676, "step": 615 }, { "epoch": 0.5136543673129039, "grad_norm": 0.0950765460729599, "learning_rate": 1.894467327844663e-06, "loss": 0.7035, "step": 616 }, { "epoch": 0.514488221805295, "grad_norm": 0.08676007390022278, "learning_rate": 1.8940652767231483e-06, "loss": 0.6067, "step": 617 }, { "epoch": 0.515322076297686, "grad_norm": 0.09285329282283783, "learning_rate": 1.8936625040329438e-06, "loss": 0.6441, "step": 618 }, { "epoch": 0.5161559307900772, "grad_norm": 0.09361077845096588, "learning_rate": 1.8932590100991136e-06, "loss": 0.6767, "step": 619 }, { "epoch": 0.5169897852824682, "grad_norm": 0.09173277020454407, "learning_rate": 1.8928547952473033e-06, "loss": 0.6395, "step": 620 }, { "epoch": 0.5178236397748592, "grad_norm": 0.09198316186666489, "learning_rate": 1.892449859803741e-06, "loss": 0.7006, "step": 621 }, { "epoch": 0.5186574942672504, "grad_norm": 0.09348779171705246, "learning_rate": 1.8920442040952356e-06, "loss": 0.6641, "step": 622 }, { "epoch": 0.5194913487596414, "grad_norm": 0.08798382431268692, "learning_rate": 1.8916378284491774e-06, "loss": 0.6478, "step": 623 }, { "epoch": 0.5203252032520326, "grad_norm": 0.09442038089036942, "learning_rate": 1.8912307331935383e-06, "loss": 0.6201, "step": 624 }, { "epoch": 0.5211590577444236, "grad_norm": 0.09519796818494797, "learning_rate": 1.8908229186568705e-06, "loss": 0.5917, "step": 625 }, { "epoch": 0.5219929122368147, "grad_norm": 0.09593921154737473, "learning_rate": 1.8904143851683062e-06, "loss": 0.6859, "step": 626 }, { "epoch": 0.5228267667292058, "grad_norm": 0.09510339796543121, "learning_rate": 1.890005133057559e-06, "loss": 0.6679, "step": 627 }, { "epoch": 0.5236606212215968, "grad_norm": 0.09473798424005508, "learning_rate": 1.8895951626549222e-06, "loss": 0.6108, "step": 628 }, { "epoch": 0.5244944757139879, "grad_norm": 0.10544616729021072, "learning_rate": 1.8891844742912678e-06, "loss": 0.6007, "step": 629 }, { "epoch": 0.525328330206379, "grad_norm": 0.09652596712112427, "learning_rate": 1.8887730682980482e-06, "loss": 0.6858, "step": 630 }, { "epoch": 0.5261621846987701, "grad_norm": 0.09385030716657639, "learning_rate": 1.888360945007295e-06, "loss": 0.692, "step": 631 }, { "epoch": 0.5269960391911611, "grad_norm": 0.10086268931627274, "learning_rate": 1.887948104751618e-06, "loss": 0.7302, "step": 632 }, { "epoch": 0.5278298936835523, "grad_norm": 0.09164178371429443, "learning_rate": 1.8875345478642065e-06, "loss": 0.5793, "step": 633 }, { "epoch": 0.5286637481759433, "grad_norm": 0.08686941117048264, "learning_rate": 1.8871202746788274e-06, "loss": 0.6304, "step": 634 }, { "epoch": 0.5294976026683343, "grad_norm": 0.09579890966415405, "learning_rate": 1.8867052855298264e-06, "loss": 0.6663, "step": 635 }, { "epoch": 0.5303314571607255, "grad_norm": 0.09054259210824966, "learning_rate": 1.8862895807521264e-06, "loss": 0.6395, "step": 636 }, { "epoch": 0.5311653116531165, "grad_norm": 0.0913679376244545, "learning_rate": 1.8858731606812284e-06, "loss": 0.6481, "step": 637 }, { "epoch": 0.5319991661455076, "grad_norm": 0.09573955833911896, "learning_rate": 1.8854560256532098e-06, "loss": 0.6543, "step": 638 }, { "epoch": 0.5328330206378987, "grad_norm": 0.09667269140481949, "learning_rate": 1.8850381760047262e-06, "loss": 0.6532, "step": 639 }, { "epoch": 0.5336668751302898, "grad_norm": 0.09536033123731613, "learning_rate": 1.8846196120730093e-06, "loss": 0.6429, "step": 640 }, { "epoch": 0.5345007296226808, "grad_norm": 0.09058564901351929, "learning_rate": 1.8842003341958673e-06, "loss": 0.579, "step": 641 }, { "epoch": 0.535334584115072, "grad_norm": 0.09930071234703064, "learning_rate": 1.8837803427116844e-06, "loss": 0.5868, "step": 642 }, { "epoch": 0.536168438607463, "grad_norm": 0.0978417843580246, "learning_rate": 1.8833596379594211e-06, "loss": 0.6822, "step": 643 }, { "epoch": 0.537002293099854, "grad_norm": 0.09346287697553635, "learning_rate": 1.882938220278614e-06, "loss": 0.6116, "step": 644 }, { "epoch": 0.5378361475922452, "grad_norm": 0.09151905030012131, "learning_rate": 1.8825160900093738e-06, "loss": 0.6496, "step": 645 }, { "epoch": 0.5386700020846362, "grad_norm": 0.0970907062292099, "learning_rate": 1.8820932474923871e-06, "loss": 0.6504, "step": 646 }, { "epoch": 0.5395038565770273, "grad_norm": 0.09408881515264511, "learning_rate": 1.8816696930689159e-06, "loss": 0.6829, "step": 647 }, { "epoch": 0.5403377110694184, "grad_norm": 0.09966862946748734, "learning_rate": 1.8812454270807951e-06, "loss": 0.5898, "step": 648 }, { "epoch": 0.5411715655618095, "grad_norm": 0.0900130420923233, "learning_rate": 1.8808204498704356e-06, "loss": 0.6042, "step": 649 }, { "epoch": 0.5420054200542005, "grad_norm": 0.09694640338420868, "learning_rate": 1.8803947617808214e-06, "loss": 0.6727, "step": 650 }, { "epoch": 0.5428392745465916, "grad_norm": 0.09688873589038849, "learning_rate": 1.8799683631555103e-06, "loss": 0.6137, "step": 651 }, { "epoch": 0.5436731290389827, "grad_norm": 0.09676310420036316, "learning_rate": 1.8795412543386337e-06, "loss": 0.612, "step": 652 }, { "epoch": 0.5445069835313737, "grad_norm": 0.09581848233938217, "learning_rate": 1.8791134356748962e-06, "loss": 0.6632, "step": 653 }, { "epoch": 0.5453408380237649, "grad_norm": 0.099350206553936, "learning_rate": 1.8786849075095753e-06, "loss": 0.6652, "step": 654 }, { "epoch": 0.5461746925161559, "grad_norm": 0.09312117099761963, "learning_rate": 1.878255670188521e-06, "loss": 0.6184, "step": 655 }, { "epoch": 0.5470085470085471, "grad_norm": 0.09312383830547333, "learning_rate": 1.8778257240581554e-06, "loss": 0.6595, "step": 656 }, { "epoch": 0.5478424015009381, "grad_norm": 0.09485237300395966, "learning_rate": 1.8773950694654732e-06, "loss": 0.716, "step": 657 }, { "epoch": 0.5486762559933291, "grad_norm": 0.09585438668727875, "learning_rate": 1.8769637067580405e-06, "loss": 0.7058, "step": 658 }, { "epoch": 0.5495101104857203, "grad_norm": 0.09024535119533539, "learning_rate": 1.876531636283995e-06, "loss": 0.594, "step": 659 }, { "epoch": 0.5503439649781113, "grad_norm": 0.09285992383956909, "learning_rate": 1.8760988583920456e-06, "loss": 0.648, "step": 660 }, { "epoch": 0.5511778194705024, "grad_norm": 0.08939266204833984, "learning_rate": 1.8756653734314722e-06, "loss": 0.5866, "step": 661 }, { "epoch": 0.5520116739628935, "grad_norm": 0.10212475061416626, "learning_rate": 1.8752311817521254e-06, "loss": 0.5485, "step": 662 }, { "epoch": 0.5528455284552846, "grad_norm": 0.0938698872923851, "learning_rate": 1.8747962837044256e-06, "loss": 0.692, "step": 663 }, { "epoch": 0.5536793829476756, "grad_norm": 0.09794861823320389, "learning_rate": 1.8743606796393644e-06, "loss": 0.6449, "step": 664 }, { "epoch": 0.5545132374400668, "grad_norm": 0.09574401378631592, "learning_rate": 1.8739243699085019e-06, "loss": 0.5878, "step": 665 }, { "epoch": 0.5553470919324578, "grad_norm": 0.10169863700866699, "learning_rate": 1.8734873548639687e-06, "loss": 0.6968, "step": 666 }, { "epoch": 0.5561809464248488, "grad_norm": 0.09560302644968033, "learning_rate": 1.8730496348584642e-06, "loss": 0.6448, "step": 667 }, { "epoch": 0.55701480091724, "grad_norm": 0.10035623610019684, "learning_rate": 1.8726112102452572e-06, "loss": 0.6893, "step": 668 }, { "epoch": 0.557848655409631, "grad_norm": 0.093927301466465, "learning_rate": 1.872172081378184e-06, "loss": 0.6541, "step": 669 }, { "epoch": 0.5586825099020221, "grad_norm": 0.0979999378323555, "learning_rate": 1.8717322486116508e-06, "loss": 0.6651, "step": 670 }, { "epoch": 0.5595163643944132, "grad_norm": 0.10020069032907486, "learning_rate": 1.8712917123006312e-06, "loss": 0.6289, "step": 671 }, { "epoch": 0.5603502188868043, "grad_norm": 0.10121606290340424, "learning_rate": 1.8708504728006664e-06, "loss": 0.6547, "step": 672 }, { "epoch": 0.5611840733791953, "grad_norm": 0.09832029789686203, "learning_rate": 1.870408530467865e-06, "loss": 0.651, "step": 673 }, { "epoch": 0.5620179278715864, "grad_norm": 0.0978146642446518, "learning_rate": 1.8699658856589037e-06, "loss": 0.6453, "step": 674 }, { "epoch": 0.5628517823639775, "grad_norm": 0.10008279979228973, "learning_rate": 1.8695225387310249e-06, "loss": 0.6945, "step": 675 }, { "epoch": 0.5636856368563685, "grad_norm": 0.09760496020317078, "learning_rate": 1.8690784900420388e-06, "loss": 0.6222, "step": 676 }, { "epoch": 0.5645194913487597, "grad_norm": 0.09725689142942429, "learning_rate": 1.8686337399503215e-06, "loss": 0.6464, "step": 677 }, { "epoch": 0.5653533458411507, "grad_norm": 0.09347565472126007, "learning_rate": 1.8681882888148152e-06, "loss": 0.6282, "step": 678 }, { "epoch": 0.5661872003335418, "grad_norm": 0.09704628586769104, "learning_rate": 1.867742136995028e-06, "loss": 0.6529, "step": 679 }, { "epoch": 0.5670210548259329, "grad_norm": 0.09918136149644852, "learning_rate": 1.8672952848510328e-06, "loss": 0.739, "step": 680 }, { "epoch": 0.5678549093183239, "grad_norm": 0.09772541373968124, "learning_rate": 1.8668477327434686e-06, "loss": 0.6383, "step": 681 }, { "epoch": 0.568688763810715, "grad_norm": 0.09866555780172348, "learning_rate": 1.866399481033539e-06, "loss": 0.6721, "step": 682 }, { "epoch": 0.5695226183031061, "grad_norm": 0.09793423861265182, "learning_rate": 1.8659505300830123e-06, "loss": 0.6055, "step": 683 }, { "epoch": 0.5703564727954972, "grad_norm": 0.10061061382293701, "learning_rate": 1.865500880254221e-06, "loss": 0.6544, "step": 684 }, { "epoch": 0.5711903272878882, "grad_norm": 0.09208130091428757, "learning_rate": 1.8650505319100617e-06, "loss": 0.6014, "step": 685 }, { "epoch": 0.5720241817802794, "grad_norm": 0.0995427742600441, "learning_rate": 1.8645994854139948e-06, "loss": 0.6965, "step": 686 }, { "epoch": 0.5728580362726704, "grad_norm": 0.09244947880506516, "learning_rate": 1.8641477411300438e-06, "loss": 0.6272, "step": 687 }, { "epoch": 0.5736918907650616, "grad_norm": 0.09937774389982224, "learning_rate": 1.863695299422796e-06, "loss": 0.6648, "step": 688 }, { "epoch": 0.5745257452574526, "grad_norm": 0.09111293405294418, "learning_rate": 1.8632421606574007e-06, "loss": 0.5476, "step": 689 }, { "epoch": 0.5753595997498436, "grad_norm": 0.10068730264902115, "learning_rate": 1.8627883251995708e-06, "loss": 0.6535, "step": 690 }, { "epoch": 0.5761934542422348, "grad_norm": 0.09240923076868057, "learning_rate": 1.862333793415581e-06, "loss": 0.575, "step": 691 }, { "epoch": 0.5770273087346258, "grad_norm": 0.09765233844518661, "learning_rate": 1.8618785656722676e-06, "loss": 0.6323, "step": 692 }, { "epoch": 0.5778611632270169, "grad_norm": 0.09996625781059265, "learning_rate": 1.861422642337029e-06, "loss": 0.639, "step": 693 }, { "epoch": 0.578695017719408, "grad_norm": 0.10148533433675766, "learning_rate": 1.860966023777825e-06, "loss": 0.7196, "step": 694 }, { "epoch": 0.579528872211799, "grad_norm": 0.09370430558919907, "learning_rate": 1.8605087103631764e-06, "loss": 0.6385, "step": 695 }, { "epoch": 0.5803627267041901, "grad_norm": 0.09228570014238358, "learning_rate": 1.8600507024621648e-06, "loss": 0.6242, "step": 696 }, { "epoch": 0.5811965811965812, "grad_norm": 0.09689746052026749, "learning_rate": 1.8595920004444324e-06, "loss": 0.6135, "step": 697 }, { "epoch": 0.5820304356889723, "grad_norm": 0.09802035987377167, "learning_rate": 1.8591326046801812e-06, "loss": 0.5789, "step": 698 }, { "epoch": 0.5828642901813633, "grad_norm": 0.10172217339277267, "learning_rate": 1.8586725155401735e-06, "loss": 0.5921, "step": 699 }, { "epoch": 0.5836981446737545, "grad_norm": 0.10300835222005844, "learning_rate": 1.858211733395731e-06, "loss": 0.6647, "step": 700 }, { "epoch": 0.5845319991661455, "grad_norm": 0.09017598628997803, "learning_rate": 1.8577502586187353e-06, "loss": 0.6409, "step": 701 }, { "epoch": 0.5853658536585366, "grad_norm": 0.10058888792991638, "learning_rate": 1.8572880915816259e-06, "loss": 0.5902, "step": 702 }, { "epoch": 0.5861997081509277, "grad_norm": 0.10125189274549484, "learning_rate": 1.856825232657402e-06, "loss": 0.6392, "step": 703 }, { "epoch": 0.5870335626433187, "grad_norm": 0.0983993262052536, "learning_rate": 1.8563616822196204e-06, "loss": 0.5915, "step": 704 }, { "epoch": 0.5878674171357098, "grad_norm": 0.09559126943349838, "learning_rate": 1.8558974406423965e-06, "loss": 0.6386, "step": 705 }, { "epoch": 0.5887012716281009, "grad_norm": 0.09088745713233948, "learning_rate": 1.8554325083004034e-06, "loss": 0.6483, "step": 706 }, { "epoch": 0.589535126120492, "grad_norm": 0.09516370296478271, "learning_rate": 1.854966885568872e-06, "loss": 0.6556, "step": 707 }, { "epoch": 0.590368980612883, "grad_norm": 0.09569456428289413, "learning_rate": 1.8545005728235896e-06, "loss": 0.6394, "step": 708 }, { "epoch": 0.5912028351052742, "grad_norm": 0.0980261042714119, "learning_rate": 1.854033570440901e-06, "loss": 0.6415, "step": 709 }, { "epoch": 0.5920366895976652, "grad_norm": 0.09796703606843948, "learning_rate": 1.8535658787977075e-06, "loss": 0.6736, "step": 710 }, { "epoch": 0.5928705440900562, "grad_norm": 0.09802737832069397, "learning_rate": 1.8530974982714664e-06, "loss": 0.5802, "step": 711 }, { "epoch": 0.5937043985824474, "grad_norm": 0.09888147562742233, "learning_rate": 1.8526284292401914e-06, "loss": 0.6648, "step": 712 }, { "epoch": 0.5945382530748384, "grad_norm": 0.10190586000680923, "learning_rate": 1.8521586720824515e-06, "loss": 0.7032, "step": 713 }, { "epoch": 0.5953721075672295, "grad_norm": 0.09422854334115982, "learning_rate": 1.8516882271773712e-06, "loss": 0.6551, "step": 714 }, { "epoch": 0.5962059620596206, "grad_norm": 0.09025522321462631, "learning_rate": 1.8512170949046305e-06, "loss": 0.6231, "step": 715 }, { "epoch": 0.5970398165520117, "grad_norm": 0.102678582072258, "learning_rate": 1.850745275644463e-06, "loss": 0.6924, "step": 716 }, { "epoch": 0.5978736710444027, "grad_norm": 0.09951562434434891, "learning_rate": 1.8502727697776579e-06, "loss": 0.5763, "step": 717 }, { "epoch": 0.5987075255367938, "grad_norm": 0.10048934072256088, "learning_rate": 1.8497995776855579e-06, "loss": 0.6369, "step": 718 }, { "epoch": 0.5995413800291849, "grad_norm": 0.11139318346977234, "learning_rate": 1.84932569975006e-06, "loss": 0.6086, "step": 719 }, { "epoch": 0.600375234521576, "grad_norm": 0.10424143821001053, "learning_rate": 1.8488511363536138e-06, "loss": 0.6394, "step": 720 }, { "epoch": 0.6012090890139671, "grad_norm": 0.09213969111442566, "learning_rate": 1.8483758878792236e-06, "loss": 0.6554, "step": 721 }, { "epoch": 0.6020429435063581, "grad_norm": 0.0998111367225647, "learning_rate": 1.8478999547104452e-06, "loss": 0.6717, "step": 722 }, { "epoch": 0.6028767979987493, "grad_norm": 0.09211792796850204, "learning_rate": 1.8474233372313877e-06, "loss": 0.5694, "step": 723 }, { "epoch": 0.6037106524911403, "grad_norm": 0.10054522007703781, "learning_rate": 1.8469460358267127e-06, "loss": 0.6501, "step": 724 }, { "epoch": 0.6045445069835313, "grad_norm": 0.09979543089866638, "learning_rate": 1.8464680508816323e-06, "loss": 0.6056, "step": 725 }, { "epoch": 0.6053783614759225, "grad_norm": 0.1000729575753212, "learning_rate": 1.8459893827819126e-06, "loss": 0.6302, "step": 726 }, { "epoch": 0.6062122159683135, "grad_norm": 0.09924089163541794, "learning_rate": 1.8455100319138692e-06, "loss": 0.6432, "step": 727 }, { "epoch": 0.6070460704607046, "grad_norm": 0.0920461118221283, "learning_rate": 1.8450299986643695e-06, "loss": 0.6081, "step": 728 }, { "epoch": 0.6078799249530957, "grad_norm": 0.10430511087179184, "learning_rate": 1.8445492834208316e-06, "loss": 0.6586, "step": 729 }, { "epoch": 0.6087137794454868, "grad_norm": 0.09923563152551651, "learning_rate": 1.8440678865712236e-06, "loss": 0.6166, "step": 730 }, { "epoch": 0.6095476339378778, "grad_norm": 0.1012277603149414, "learning_rate": 1.843585808504064e-06, "loss": 0.6843, "step": 731 }, { "epoch": 0.610381488430269, "grad_norm": 0.09866297990083694, "learning_rate": 1.8431030496084214e-06, "loss": 0.6556, "step": 732 }, { "epoch": 0.61121534292266, "grad_norm": 0.09874875843524933, "learning_rate": 1.8426196102739132e-06, "loss": 0.5993, "step": 733 }, { "epoch": 0.612049197415051, "grad_norm": 0.09275837987661362, "learning_rate": 1.8421354908907067e-06, "loss": 0.6326, "step": 734 }, { "epoch": 0.6128830519074422, "grad_norm": 0.09502032399177551, "learning_rate": 1.8416506918495173e-06, "loss": 0.6077, "step": 735 }, { "epoch": 0.6137169063998332, "grad_norm": 0.09185368567705154, "learning_rate": 1.8411652135416093e-06, "loss": 0.6064, "step": 736 }, { "epoch": 0.6145507608922243, "grad_norm": 0.09815992414951324, "learning_rate": 1.8406790563587956e-06, "loss": 0.6313, "step": 737 }, { "epoch": 0.6153846153846154, "grad_norm": 0.09629546850919724, "learning_rate": 1.8401922206934361e-06, "loss": 0.6106, "step": 738 }, { "epoch": 0.6162184698770065, "grad_norm": 0.09635171294212341, "learning_rate": 1.839704706938439e-06, "loss": 0.6523, "step": 739 }, { "epoch": 0.6170523243693975, "grad_norm": 0.10179682075977325, "learning_rate": 1.8392165154872595e-06, "loss": 0.5873, "step": 740 }, { "epoch": 0.6178861788617886, "grad_norm": 0.10270754247903824, "learning_rate": 1.8387276467338996e-06, "loss": 0.673, "step": 741 }, { "epoch": 0.6187200333541797, "grad_norm": 0.09746810793876648, "learning_rate": 1.8382381010729086e-06, "loss": 0.6564, "step": 742 }, { "epoch": 0.6195538878465707, "grad_norm": 0.0962262824177742, "learning_rate": 1.8377478788993813e-06, "loss": 0.5738, "step": 743 }, { "epoch": 0.6203877423389619, "grad_norm": 0.09518938511610031, "learning_rate": 1.8372569806089586e-06, "loss": 0.6345, "step": 744 }, { "epoch": 0.6212215968313529, "grad_norm": 0.09188467264175415, "learning_rate": 1.8367654065978276e-06, "loss": 0.6112, "step": 745 }, { "epoch": 0.622055451323744, "grad_norm": 0.09746599197387695, "learning_rate": 1.8362731572627202e-06, "loss": 0.6349, "step": 746 }, { "epoch": 0.6228893058161351, "grad_norm": 0.10009641200304031, "learning_rate": 1.8357802330009136e-06, "loss": 0.6502, "step": 747 }, { "epoch": 0.6237231603085261, "grad_norm": 0.10499203950166702, "learning_rate": 1.8352866342102296e-06, "loss": 0.6812, "step": 748 }, { "epoch": 0.6245570148009172, "grad_norm": 0.09757328033447266, "learning_rate": 1.8347923612890349e-06, "loss": 0.6025, "step": 749 }, { "epoch": 0.6253908692933083, "grad_norm": 0.10616346448659897, "learning_rate": 1.8342974146362394e-06, "loss": 0.6638, "step": 750 }, { "epoch": 0.6262247237856994, "grad_norm": 0.10131774097681046, "learning_rate": 1.8338017946512968e-06, "loss": 0.5456, "step": 751 }, { "epoch": 0.6270585782780905, "grad_norm": 0.10244927555322647, "learning_rate": 1.8333055017342054e-06, "loss": 0.6751, "step": 752 }, { "epoch": 0.6278924327704816, "grad_norm": 0.10338141769170761, "learning_rate": 1.8328085362855055e-06, "loss": 0.6647, "step": 753 }, { "epoch": 0.6287262872628726, "grad_norm": 0.10026909410953522, "learning_rate": 1.8323108987062802e-06, "loss": 0.6543, "step": 754 }, { "epoch": 0.6295601417552638, "grad_norm": 0.10641954839229584, "learning_rate": 1.8318125893981553e-06, "loss": 0.6425, "step": 755 }, { "epoch": 0.6303939962476548, "grad_norm": 0.09939148277044296, "learning_rate": 1.8313136087632995e-06, "loss": 0.6531, "step": 756 }, { "epoch": 0.6312278507400458, "grad_norm": 0.10437260568141937, "learning_rate": 1.8308139572044215e-06, "loss": 0.6783, "step": 757 }, { "epoch": 0.632061705232437, "grad_norm": 0.09760341793298721, "learning_rate": 1.830313635124773e-06, "loss": 0.6734, "step": 758 }, { "epoch": 0.632895559724828, "grad_norm": 0.1017068475484848, "learning_rate": 1.8298126429281467e-06, "loss": 0.6171, "step": 759 }, { "epoch": 0.6337294142172191, "grad_norm": 0.09881062060594559, "learning_rate": 1.8293109810188754e-06, "loss": 0.6309, "step": 760 }, { "epoch": 0.6345632687096102, "grad_norm": 0.11102797091007233, "learning_rate": 1.8288086498018326e-06, "loss": 0.6361, "step": 761 }, { "epoch": 0.6353971232020013, "grad_norm": 0.09473263472318649, "learning_rate": 1.828305649682433e-06, "loss": 0.5787, "step": 762 }, { "epoch": 0.6362309776943923, "grad_norm": 0.10742990672588348, "learning_rate": 1.8278019810666293e-06, "loss": 0.6638, "step": 763 }, { "epoch": 0.6370648321867834, "grad_norm": 0.09987856447696686, "learning_rate": 1.8272976443609156e-06, "loss": 0.6054, "step": 764 }, { "epoch": 0.6378986866791745, "grad_norm": 0.09839458018541336, "learning_rate": 1.8267926399723238e-06, "loss": 0.6258, "step": 765 }, { "epoch": 0.6387325411715655, "grad_norm": 0.09883236885070801, "learning_rate": 1.8262869683084258e-06, "loss": 0.6695, "step": 766 }, { "epoch": 0.6395663956639567, "grad_norm": 0.10263945162296295, "learning_rate": 1.825780629777331e-06, "loss": 0.6032, "step": 767 }, { "epoch": 0.6404002501563477, "grad_norm": 0.09623769670724869, "learning_rate": 1.8252736247876874e-06, "loss": 0.6794, "step": 768 }, { "epoch": 0.6412341046487388, "grad_norm": 0.10218116641044617, "learning_rate": 1.8247659537486811e-06, "loss": 0.5969, "step": 769 }, { "epoch": 0.6420679591411299, "grad_norm": 0.09405802190303802, "learning_rate": 1.8242576170700352e-06, "loss": 0.6493, "step": 770 }, { "epoch": 0.642901813633521, "grad_norm": 0.09709781408309937, "learning_rate": 1.8237486151620112e-06, "loss": 0.7054, "step": 771 }, { "epoch": 0.643735668125912, "grad_norm": 0.09584866464138031, "learning_rate": 1.823238948435406e-06, "loss": 0.6299, "step": 772 }, { "epoch": 0.6445695226183031, "grad_norm": 0.09731820970773697, "learning_rate": 1.8227286173015538e-06, "loss": 0.6633, "step": 773 }, { "epoch": 0.6454033771106942, "grad_norm": 0.09664048999547958, "learning_rate": 1.822217622172325e-06, "loss": 0.6075, "step": 774 }, { "epoch": 0.6462372316030852, "grad_norm": 0.0990557000041008, "learning_rate": 1.8217059634601259e-06, "loss": 0.6755, "step": 775 }, { "epoch": 0.6470710860954764, "grad_norm": 0.10071469098329544, "learning_rate": 1.8211936415778983e-06, "loss": 0.6197, "step": 776 }, { "epoch": 0.6479049405878674, "grad_norm": 0.10386556386947632, "learning_rate": 1.820680656939119e-06, "loss": 0.6195, "step": 777 }, { "epoch": 0.6487387950802584, "grad_norm": 0.09966234862804413, "learning_rate": 1.8201670099578001e-06, "loss": 0.6654, "step": 778 }, { "epoch": 0.6495726495726496, "grad_norm": 0.09882418066263199, "learning_rate": 1.819652701048488e-06, "loss": 0.6525, "step": 779 }, { "epoch": 0.6504065040650406, "grad_norm": 0.09504549950361252, "learning_rate": 1.8191377306262633e-06, "loss": 0.6326, "step": 780 }, { "epoch": 0.6512403585574317, "grad_norm": 0.09834848344326019, "learning_rate": 1.8186220991067404e-06, "loss": 0.6417, "step": 781 }, { "epoch": 0.6520742130498228, "grad_norm": 0.09583116322755814, "learning_rate": 1.8181058069060676e-06, "loss": 0.6644, "step": 782 }, { "epoch": 0.6529080675422139, "grad_norm": 0.09287475794553757, "learning_rate": 1.8175888544409264e-06, "loss": 0.6385, "step": 783 }, { "epoch": 0.653741922034605, "grad_norm": 0.10364864766597748, "learning_rate": 1.8170712421285305e-06, "loss": 0.6996, "step": 784 }, { "epoch": 0.6545757765269961, "grad_norm": 0.09863609820604324, "learning_rate": 1.816552970386627e-06, "loss": 0.6272, "step": 785 }, { "epoch": 0.6554096310193871, "grad_norm": 0.10557069629430771, "learning_rate": 1.8160340396334949e-06, "loss": 0.6318, "step": 786 }, { "epoch": 0.6562434855117782, "grad_norm": 0.10506842285394669, "learning_rate": 1.8155144502879443e-06, "loss": 0.6557, "step": 787 }, { "epoch": 0.6570773400041693, "grad_norm": 0.09928935766220093, "learning_rate": 1.8149942027693182e-06, "loss": 0.6832, "step": 788 }, { "epoch": 0.6579111944965603, "grad_norm": 0.09821849316358566, "learning_rate": 1.8144732974974902e-06, "loss": 0.6342, "step": 789 }, { "epoch": 0.6587450489889515, "grad_norm": 0.09966670721769333, "learning_rate": 1.8139517348928637e-06, "loss": 0.6321, "step": 790 }, { "epoch": 0.6595789034813425, "grad_norm": 0.10439924150705338, "learning_rate": 1.8134295153763745e-06, "loss": 0.6552, "step": 791 }, { "epoch": 0.6604127579737336, "grad_norm": 0.09848225861787796, "learning_rate": 1.8129066393694877e-06, "loss": 0.593, "step": 792 }, { "epoch": 0.6612466124661247, "grad_norm": 0.09793344885110855, "learning_rate": 1.8123831072941974e-06, "loss": 0.6406, "step": 793 }, { "epoch": 0.6620804669585157, "grad_norm": 0.0945219025015831, "learning_rate": 1.8118589195730285e-06, "loss": 0.6118, "step": 794 }, { "epoch": 0.6629143214509068, "grad_norm": 0.10230748355388641, "learning_rate": 1.8113340766290343e-06, "loss": 0.65, "step": 795 }, { "epoch": 0.6637481759432979, "grad_norm": 0.10141695290803909, "learning_rate": 1.8108085788857973e-06, "loss": 0.6109, "step": 796 }, { "epoch": 0.664582030435689, "grad_norm": 0.11011820286512375, "learning_rate": 1.8102824267674282e-06, "loss": 0.6531, "step": 797 }, { "epoch": 0.66541588492808, "grad_norm": 0.09687966108322144, "learning_rate": 1.8097556206985658e-06, "loss": 0.649, "step": 798 }, { "epoch": 0.6662497394204712, "grad_norm": 0.09653454273939133, "learning_rate": 1.8092281611043764e-06, "loss": 0.6487, "step": 799 }, { "epoch": 0.6670835939128622, "grad_norm": 0.10373052954673767, "learning_rate": 1.808700048410555e-06, "loss": 0.6221, "step": 800 }, { "epoch": 0.6670835939128622, "eval_loss": 0.6415942907333374, "eval_runtime": 321.6008, "eval_samples_per_second": 16.738, "eval_steps_per_second": 2.792, "step": 800 }, { "epoch": 0.6679174484052532, "grad_norm": 0.10485277324914932, "learning_rate": 1.8081712830433224e-06, "loss": 0.6176, "step": 801 }, { "epoch": 0.6687513028976444, "grad_norm": 0.10105489194393158, "learning_rate": 1.8076418654294266e-06, "loss": 0.6708, "step": 802 }, { "epoch": 0.6695851573900354, "grad_norm": 0.1096423864364624, "learning_rate": 1.8071117959961416e-06, "loss": 0.601, "step": 803 }, { "epoch": 0.6704190118824265, "grad_norm": 0.10762202739715576, "learning_rate": 1.8065810751712682e-06, "loss": 0.6567, "step": 804 }, { "epoch": 0.6712528663748176, "grad_norm": 0.1107725277543068, "learning_rate": 1.8060497033831324e-06, "loss": 0.5889, "step": 805 }, { "epoch": 0.6720867208672087, "grad_norm": 0.10468114167451859, "learning_rate": 1.8055176810605857e-06, "loss": 0.5847, "step": 806 }, { "epoch": 0.6729205753595997, "grad_norm": 0.11013887077569962, "learning_rate": 1.8049850086330048e-06, "loss": 0.6008, "step": 807 }, { "epoch": 0.6737544298519909, "grad_norm": 0.1043848916888237, "learning_rate": 1.8044516865302908e-06, "loss": 0.6368, "step": 808 }, { "epoch": 0.6745882843443819, "grad_norm": 0.09962385892868042, "learning_rate": 1.8039177151828691e-06, "loss": 0.5842, "step": 809 }, { "epoch": 0.6754221388367729, "grad_norm": 0.1098175197839737, "learning_rate": 1.8033830950216897e-06, "loss": 0.6397, "step": 810 }, { "epoch": 0.6762559933291641, "grad_norm": 0.10420270264148712, "learning_rate": 1.8028478264782252e-06, "loss": 0.6131, "step": 811 }, { "epoch": 0.6770898478215551, "grad_norm": 0.10396002978086472, "learning_rate": 1.8023119099844723e-06, "loss": 0.6499, "step": 812 }, { "epoch": 0.6779237023139462, "grad_norm": 0.09836006909608841, "learning_rate": 1.8017753459729504e-06, "loss": 0.5962, "step": 813 }, { "epoch": 0.6787575568063373, "grad_norm": 0.11005699634552002, "learning_rate": 1.8012381348767014e-06, "loss": 0.6202, "step": 814 }, { "epoch": 0.6795914112987284, "grad_norm": 0.10937905311584473, "learning_rate": 1.8007002771292896e-06, "loss": 0.6053, "step": 815 }, { "epoch": 0.6804252657911195, "grad_norm": 0.10817151516675949, "learning_rate": 1.800161773164801e-06, "loss": 0.6713, "step": 816 }, { "epoch": 0.6812591202835105, "grad_norm": 0.10142417997121811, "learning_rate": 1.7996226234178434e-06, "loss": 0.6166, "step": 817 }, { "epoch": 0.6820929747759016, "grad_norm": 0.10415297746658325, "learning_rate": 1.799082828323545e-06, "loss": 0.6912, "step": 818 }, { "epoch": 0.6829268292682927, "grad_norm": 0.09985463321208954, "learning_rate": 1.7985423883175565e-06, "loss": 0.6694, "step": 819 }, { "epoch": 0.6837606837606838, "grad_norm": 0.1025400459766388, "learning_rate": 1.7980013038360475e-06, "loss": 0.6464, "step": 820 }, { "epoch": 0.6845945382530748, "grad_norm": 0.09324987977743149, "learning_rate": 1.7974595753157082e-06, "loss": 0.6251, "step": 821 }, { "epoch": 0.685428392745466, "grad_norm": 0.1065581887960434, "learning_rate": 1.796917203193749e-06, "loss": 0.6427, "step": 822 }, { "epoch": 0.686262247237857, "grad_norm": 0.1113988533616066, "learning_rate": 1.7963741879078985e-06, "loss": 0.6652, "step": 823 }, { "epoch": 0.687096101730248, "grad_norm": 0.09690185636281967, "learning_rate": 1.7958305298964066e-06, "loss": 0.6294, "step": 824 }, { "epoch": 0.6879299562226392, "grad_norm": 0.10385391116142273, "learning_rate": 1.7952862295980397e-06, "loss": 0.611, "step": 825 }, { "epoch": 0.6887638107150302, "grad_norm": 0.10155371576547623, "learning_rate": 1.794741287452084e-06, "loss": 0.6268, "step": 826 }, { "epoch": 0.6895976652074213, "grad_norm": 0.10731323808431625, "learning_rate": 1.7941957038983425e-06, "loss": 0.6294, "step": 827 }, { "epoch": 0.6904315196998124, "grad_norm": 0.0989241898059845, "learning_rate": 1.7936494793771369e-06, "loss": 0.6739, "step": 828 }, { "epoch": 0.6912653741922035, "grad_norm": 0.10223259776830673, "learning_rate": 1.7931026143293059e-06, "loss": 0.5857, "step": 829 }, { "epoch": 0.6920992286845945, "grad_norm": 0.10547658056020737, "learning_rate": 1.792555109196205e-06, "loss": 0.6864, "step": 830 }, { "epoch": 0.6929330831769857, "grad_norm": 0.1027214452624321, "learning_rate": 1.7920069644197063e-06, "loss": 0.5857, "step": 831 }, { "epoch": 0.6937669376693767, "grad_norm": 0.10017300397157669, "learning_rate": 1.7914581804421983e-06, "loss": 0.6466, "step": 832 }, { "epoch": 0.6946007921617677, "grad_norm": 0.09962292015552521, "learning_rate": 1.7909087577065851e-06, "loss": 0.5917, "step": 833 }, { "epoch": 0.6954346466541589, "grad_norm": 0.10357961803674698, "learning_rate": 1.7903586966562868e-06, "loss": 0.6382, "step": 834 }, { "epoch": 0.6962685011465499, "grad_norm": 0.1033301129937172, "learning_rate": 1.789807997735238e-06, "loss": 0.624, "step": 835 }, { "epoch": 0.697102355638941, "grad_norm": 0.09935612231492996, "learning_rate": 1.789256661387889e-06, "loss": 0.6475, "step": 836 }, { "epoch": 0.6979362101313321, "grad_norm": 0.09738706797361374, "learning_rate": 1.7887046880592033e-06, "loss": 0.652, "step": 837 }, { "epoch": 0.6987700646237232, "grad_norm": 0.10389488190412521, "learning_rate": 1.7881520781946597e-06, "loss": 0.6755, "step": 838 }, { "epoch": 0.6996039191161142, "grad_norm": 0.09787558019161224, "learning_rate": 1.7875988322402501e-06, "loss": 0.618, "step": 839 }, { "epoch": 0.7004377736085053, "grad_norm": 0.09852594137191772, "learning_rate": 1.78704495064248e-06, "loss": 0.618, "step": 840 }, { "epoch": 0.7012716281008964, "grad_norm": 0.09819158911705017, "learning_rate": 1.7864904338483675e-06, "loss": 0.5994, "step": 841 }, { "epoch": 0.7021054825932874, "grad_norm": 0.10942944139242172, "learning_rate": 1.7859352823054437e-06, "loss": 0.6405, "step": 842 }, { "epoch": 0.7029393370856786, "grad_norm": 0.10699477791786194, "learning_rate": 1.7853794964617522e-06, "loss": 0.6444, "step": 843 }, { "epoch": 0.7037731915780696, "grad_norm": 0.10151588916778564, "learning_rate": 1.784823076765848e-06, "loss": 0.6233, "step": 844 }, { "epoch": 0.7046070460704607, "grad_norm": 0.10147272795438766, "learning_rate": 1.7842660236667974e-06, "loss": 0.6606, "step": 845 }, { "epoch": 0.7054409005628518, "grad_norm": 0.1031784638762474, "learning_rate": 1.7837083376141796e-06, "loss": 0.6332, "step": 846 }, { "epoch": 0.7062747550552428, "grad_norm": 0.10318465530872345, "learning_rate": 1.7831500190580823e-06, "loss": 0.5944, "step": 847 }, { "epoch": 0.707108609547634, "grad_norm": 0.10107962787151337, "learning_rate": 1.7825910684491054e-06, "loss": 0.686, "step": 848 }, { "epoch": 0.707942464040025, "grad_norm": 0.1008550301194191, "learning_rate": 1.7820314862383584e-06, "loss": 0.6356, "step": 849 }, { "epoch": 0.7087763185324161, "grad_norm": 0.1049947738647461, "learning_rate": 1.7814712728774598e-06, "loss": 0.6436, "step": 850 }, { "epoch": 0.7096101730248072, "grad_norm": 0.10559200495481491, "learning_rate": 1.7809104288185389e-06, "loss": 0.6522, "step": 851 }, { "epoch": 0.7104440275171983, "grad_norm": 0.09807837009429932, "learning_rate": 1.7803489545142325e-06, "loss": 0.6252, "step": 852 }, { "epoch": 0.7112778820095893, "grad_norm": 0.1072913408279419, "learning_rate": 1.7797868504176874e-06, "loss": 0.6483, "step": 853 }, { "epoch": 0.7121117365019805, "grad_norm": 0.09987325221300125, "learning_rate": 1.7792241169825579e-06, "loss": 0.5955, "step": 854 }, { "epoch": 0.7129455909943715, "grad_norm": 0.10160518437623978, "learning_rate": 1.778660754663006e-06, "loss": 0.6006, "step": 855 }, { "epoch": 0.7137794454867625, "grad_norm": 0.10053478926420212, "learning_rate": 1.7780967639137025e-06, "loss": 0.6583, "step": 856 }, { "epoch": 0.7146132999791537, "grad_norm": 0.10520876199007034, "learning_rate": 1.777532145189824e-06, "loss": 0.6061, "step": 857 }, { "epoch": 0.7154471544715447, "grad_norm": 0.10797824710607529, "learning_rate": 1.776966898947054e-06, "loss": 0.6764, "step": 858 }, { "epoch": 0.7162810089639358, "grad_norm": 0.10396280139684677, "learning_rate": 1.7764010256415837e-06, "loss": 0.6302, "step": 859 }, { "epoch": 0.7171148634563269, "grad_norm": 0.10683812946081161, "learning_rate": 1.7758345257301094e-06, "loss": 0.6618, "step": 860 }, { "epoch": 0.717948717948718, "grad_norm": 0.09601815789937973, "learning_rate": 1.7752673996698326e-06, "loss": 0.6167, "step": 861 }, { "epoch": 0.718782572441109, "grad_norm": 0.1079709529876709, "learning_rate": 1.774699647918462e-06, "loss": 0.6085, "step": 862 }, { "epoch": 0.7196164269335001, "grad_norm": 0.10923435539007187, "learning_rate": 1.774131270934209e-06, "loss": 0.6492, "step": 863 }, { "epoch": 0.7204502814258912, "grad_norm": 0.10494917631149292, "learning_rate": 1.7735622691757912e-06, "loss": 0.6281, "step": 864 }, { "epoch": 0.7212841359182822, "grad_norm": 0.10305362194776535, "learning_rate": 1.7729926431024301e-06, "loss": 0.6381, "step": 865 }, { "epoch": 0.7221179904106734, "grad_norm": 0.1005750373005867, "learning_rate": 1.7724223931738505e-06, "loss": 0.667, "step": 866 }, { "epoch": 0.7229518449030644, "grad_norm": 0.10202402621507645, "learning_rate": 1.7718515198502812e-06, "loss": 0.6685, "step": 867 }, { "epoch": 0.7237856993954555, "grad_norm": 0.10256760567426682, "learning_rate": 1.7712800235924546e-06, "loss": 0.6728, "step": 868 }, { "epoch": 0.7246195538878466, "grad_norm": 0.1020650714635849, "learning_rate": 1.7707079048616046e-06, "loss": 0.6144, "step": 869 }, { "epoch": 0.7254534083802376, "grad_norm": 0.10101509839296341, "learning_rate": 1.770135164119468e-06, "loss": 0.6244, "step": 870 }, { "epoch": 0.7262872628726287, "grad_norm": 0.09737613797187805, "learning_rate": 1.769561801828284e-06, "loss": 0.6507, "step": 871 }, { "epoch": 0.7271211173650198, "grad_norm": 0.09866712987422943, "learning_rate": 1.7689878184507937e-06, "loss": 0.5875, "step": 872 }, { "epoch": 0.7279549718574109, "grad_norm": 0.0963241383433342, "learning_rate": 1.7684132144502382e-06, "loss": 0.6237, "step": 873 }, { "epoch": 0.7287888263498019, "grad_norm": 0.09706535190343857, "learning_rate": 1.7678379902903603e-06, "loss": 0.5665, "step": 874 }, { "epoch": 0.7296226808421931, "grad_norm": 0.09960032999515533, "learning_rate": 1.7672621464354034e-06, "loss": 0.6288, "step": 875 }, { "epoch": 0.7304565353345841, "grad_norm": 0.10458897799253464, "learning_rate": 1.766685683350111e-06, "loss": 0.6148, "step": 876 }, { "epoch": 0.7312903898269751, "grad_norm": 0.10443190485239029, "learning_rate": 1.766108601499726e-06, "loss": 0.6248, "step": 877 }, { "epoch": 0.7321242443193663, "grad_norm": 0.10145988315343857, "learning_rate": 1.765530901349991e-06, "loss": 0.6117, "step": 878 }, { "epoch": 0.7329580988117573, "grad_norm": 0.10847010463476181, "learning_rate": 1.7649525833671474e-06, "loss": 0.7046, "step": 879 }, { "epoch": 0.7337919533041485, "grad_norm": 0.10503443330526352, "learning_rate": 1.7643736480179352e-06, "loss": 0.6176, "step": 880 }, { "epoch": 0.7346258077965395, "grad_norm": 0.10929647833108902, "learning_rate": 1.7637940957695934e-06, "loss": 0.6434, "step": 881 }, { "epoch": 0.7354596622889306, "grad_norm": 0.10153260827064514, "learning_rate": 1.7632139270898576e-06, "loss": 0.638, "step": 882 }, { "epoch": 0.7362935167813217, "grad_norm": 0.10493889451026917, "learning_rate": 1.7626331424469615e-06, "loss": 0.6517, "step": 883 }, { "epoch": 0.7371273712737128, "grad_norm": 0.09808988124132156, "learning_rate": 1.7620517423096368e-06, "loss": 0.6179, "step": 884 }, { "epoch": 0.7379612257661038, "grad_norm": 0.10479886084794998, "learning_rate": 1.7614697271471103e-06, "loss": 0.6419, "step": 885 }, { "epoch": 0.7387950802584949, "grad_norm": 0.09941962361335754, "learning_rate": 1.7608870974291065e-06, "loss": 0.6128, "step": 886 }, { "epoch": 0.739628934750886, "grad_norm": 0.10169458389282227, "learning_rate": 1.7603038536258453e-06, "loss": 0.609, "step": 887 }, { "epoch": 0.740462789243277, "grad_norm": 0.09726240485906601, "learning_rate": 1.7597199962080423e-06, "loss": 0.6371, "step": 888 }, { "epoch": 0.7412966437356682, "grad_norm": 0.10218486189842224, "learning_rate": 1.759135525646908e-06, "loss": 0.5768, "step": 889 }, { "epoch": 0.7421304982280592, "grad_norm": 0.10321173816919327, "learning_rate": 1.7585504424141483e-06, "loss": 0.6007, "step": 890 }, { "epoch": 0.7429643527204502, "grad_norm": 0.10717494040727615, "learning_rate": 1.7579647469819631e-06, "loss": 0.6353, "step": 891 }, { "epoch": 0.7437982072128414, "grad_norm": 0.1051764264702797, "learning_rate": 1.7573784398230474e-06, "loss": 0.6098, "step": 892 }, { "epoch": 0.7446320617052324, "grad_norm": 0.10321137309074402, "learning_rate": 1.7567915214105881e-06, "loss": 0.5986, "step": 893 }, { "epoch": 0.7454659161976235, "grad_norm": 0.10979589819908142, "learning_rate": 1.7562039922182671e-06, "loss": 0.6518, "step": 894 }, { "epoch": 0.7462997706900146, "grad_norm": 0.10153292864561081, "learning_rate": 1.7556158527202585e-06, "loss": 0.6358, "step": 895 }, { "epoch": 0.7471336251824057, "grad_norm": 0.10513276606798172, "learning_rate": 1.7550271033912287e-06, "loss": 0.6085, "step": 896 }, { "epoch": 0.7479674796747967, "grad_norm": 0.10275143384933472, "learning_rate": 1.7544377447063372e-06, "loss": 0.5748, "step": 897 }, { "epoch": 0.7488013341671879, "grad_norm": 0.10438092797994614, "learning_rate": 1.7538477771412339e-06, "loss": 0.5943, "step": 898 }, { "epoch": 0.7496351886595789, "grad_norm": 0.1010277271270752, "learning_rate": 1.7532572011720617e-06, "loss": 0.6112, "step": 899 }, { "epoch": 0.7504690431519699, "grad_norm": 0.10229716449975967, "learning_rate": 1.752666017275453e-06, "loss": 0.5861, "step": 900 }, { "epoch": 0.7513028976443611, "grad_norm": 0.10199355334043503, "learning_rate": 1.7520742259285323e-06, "loss": 0.6388, "step": 901 }, { "epoch": 0.7521367521367521, "grad_norm": 0.10268741846084595, "learning_rate": 1.7514818276089128e-06, "loss": 0.6875, "step": 902 }, { "epoch": 0.7529706066291432, "grad_norm": 0.0992840975522995, "learning_rate": 1.7508888227946992e-06, "loss": 0.6756, "step": 903 }, { "epoch": 0.7538044611215343, "grad_norm": 0.09615278989076614, "learning_rate": 1.7502952119644845e-06, "loss": 0.6474, "step": 904 }, { "epoch": 0.7546383156139254, "grad_norm": 0.10315293818712234, "learning_rate": 1.749700995597351e-06, "loss": 0.6127, "step": 905 }, { "epoch": 0.7554721701063164, "grad_norm": 0.10551444441080093, "learning_rate": 1.7491061741728702e-06, "loss": 0.6357, "step": 906 }, { "epoch": 0.7563060245987075, "grad_norm": 0.10280577838420868, "learning_rate": 1.748510748171101e-06, "loss": 0.6045, "step": 907 }, { "epoch": 0.7571398790910986, "grad_norm": 0.09516436606645584, "learning_rate": 1.7479147180725912e-06, "loss": 0.6368, "step": 908 }, { "epoch": 0.7579737335834896, "grad_norm": 0.10799692571163177, "learning_rate": 1.7473180843583762e-06, "loss": 0.6669, "step": 909 }, { "epoch": 0.7588075880758808, "grad_norm": 0.1064273938536644, "learning_rate": 1.7467208475099774e-06, "loss": 0.6481, "step": 910 }, { "epoch": 0.7596414425682718, "grad_norm": 0.1054486557841301, "learning_rate": 1.7461230080094043e-06, "loss": 0.6108, "step": 911 }, { "epoch": 0.760475297060663, "grad_norm": 0.10060249269008636, "learning_rate": 1.7455245663391516e-06, "loss": 0.5604, "step": 912 }, { "epoch": 0.761309151553054, "grad_norm": 0.1015784963965416, "learning_rate": 1.7449255229822008e-06, "loss": 0.6248, "step": 913 }, { "epoch": 0.762143006045445, "grad_norm": 0.10114108771085739, "learning_rate": 1.744325878422019e-06, "loss": 0.6683, "step": 914 }, { "epoch": 0.7629768605378362, "grad_norm": 0.10236384719610214, "learning_rate": 1.7437256331425576e-06, "loss": 0.657, "step": 915 }, { "epoch": 0.7638107150302272, "grad_norm": 0.10892149806022644, "learning_rate": 1.743124787628254e-06, "loss": 0.6434, "step": 916 }, { "epoch": 0.7646445695226183, "grad_norm": 0.10639214515686035, "learning_rate": 1.7425233423640298e-06, "loss": 0.5859, "step": 917 }, { "epoch": 0.7654784240150094, "grad_norm": 0.10279032588005066, "learning_rate": 1.7419212978352898e-06, "loss": 0.6779, "step": 918 }, { "epoch": 0.7663122785074005, "grad_norm": 0.11178586632013321, "learning_rate": 1.741318654527923e-06, "loss": 0.6827, "step": 919 }, { "epoch": 0.7671461329997915, "grad_norm": 0.1018877848982811, "learning_rate": 1.7407154129283019e-06, "loss": 0.6423, "step": 920 }, { "epoch": 0.7679799874921827, "grad_norm": 0.1052117571234703, "learning_rate": 1.7401115735232817e-06, "loss": 0.6871, "step": 921 }, { "epoch": 0.7688138419845737, "grad_norm": 0.10656667500734329, "learning_rate": 1.7395071368001997e-06, "loss": 0.6384, "step": 922 }, { "epoch": 0.7696476964769647, "grad_norm": 0.09973619878292084, "learning_rate": 1.738902103246876e-06, "loss": 0.5659, "step": 923 }, { "epoch": 0.7704815509693559, "grad_norm": 0.10072916746139526, "learning_rate": 1.7382964733516112e-06, "loss": 0.6315, "step": 924 }, { "epoch": 0.7713154054617469, "grad_norm": 0.11351827532052994, "learning_rate": 1.7376902476031882e-06, "loss": 0.7005, "step": 925 }, { "epoch": 0.772149259954138, "grad_norm": 0.1084694042801857, "learning_rate": 1.7370834264908711e-06, "loss": 0.6077, "step": 926 }, { "epoch": 0.7729831144465291, "grad_norm": 0.10465063899755478, "learning_rate": 1.7364760105044033e-06, "loss": 0.5937, "step": 927 }, { "epoch": 0.7738169689389202, "grad_norm": 0.09937559068202972, "learning_rate": 1.7358680001340092e-06, "loss": 0.5854, "step": 928 }, { "epoch": 0.7746508234313112, "grad_norm": 0.11517240107059479, "learning_rate": 1.7352593958703921e-06, "loss": 0.6738, "step": 929 }, { "epoch": 0.7754846779237023, "grad_norm": 0.10875809192657471, "learning_rate": 1.734650198204736e-06, "loss": 0.6487, "step": 930 }, { "epoch": 0.7763185324160934, "grad_norm": 0.10209451615810394, "learning_rate": 1.7340404076287021e-06, "loss": 0.6277, "step": 931 }, { "epoch": 0.7771523869084844, "grad_norm": 0.11190709471702576, "learning_rate": 1.7334300246344316e-06, "loss": 0.6393, "step": 932 }, { "epoch": 0.7779862414008756, "grad_norm": 0.10622856765985489, "learning_rate": 1.7328190497145427e-06, "loss": 0.6665, "step": 933 }, { "epoch": 0.7788200958932666, "grad_norm": 0.11038866639137268, "learning_rate": 1.7322074833621318e-06, "loss": 0.6092, "step": 934 }, { "epoch": 0.7796539503856577, "grad_norm": 0.1060342788696289, "learning_rate": 1.7315953260707735e-06, "loss": 0.6427, "step": 935 }, { "epoch": 0.7804878048780488, "grad_norm": 0.10829740762710571, "learning_rate": 1.7309825783345174e-06, "loss": 0.6851, "step": 936 }, { "epoch": 0.7813216593704398, "grad_norm": 0.1001635491847992, "learning_rate": 1.7303692406478908e-06, "loss": 0.5645, "step": 937 }, { "epoch": 0.7821555138628309, "grad_norm": 0.09842494130134583, "learning_rate": 1.7297553135058978e-06, "loss": 0.613, "step": 938 }, { "epoch": 0.782989368355222, "grad_norm": 0.10335814207792282, "learning_rate": 1.7291407974040167e-06, "loss": 0.6583, "step": 939 }, { "epoch": 0.7838232228476131, "grad_norm": 0.10193807631731033, "learning_rate": 1.728525692838202e-06, "loss": 0.6242, "step": 940 }, { "epoch": 0.7846570773400041, "grad_norm": 0.1064397543668747, "learning_rate": 1.7279100003048832e-06, "loss": 0.7043, "step": 941 }, { "epoch": 0.7854909318323953, "grad_norm": 0.09790313243865967, "learning_rate": 1.7272937203009642e-06, "loss": 0.5869, "step": 942 }, { "epoch": 0.7863247863247863, "grad_norm": 0.10425697267055511, "learning_rate": 1.7266768533238225e-06, "loss": 0.6282, "step": 943 }, { "epoch": 0.7871586408171775, "grad_norm": 0.10319948196411133, "learning_rate": 1.72605939987131e-06, "loss": 0.5899, "step": 944 }, { "epoch": 0.7879924953095685, "grad_norm": 0.10117416828870773, "learning_rate": 1.725441360441752e-06, "loss": 0.5734, "step": 945 }, { "epoch": 0.7888263498019595, "grad_norm": 0.10920300334692001, "learning_rate": 1.7248227355339458e-06, "loss": 0.612, "step": 946 }, { "epoch": 0.7896602042943507, "grad_norm": 0.0993722602725029, "learning_rate": 1.7242035256471623e-06, "loss": 0.5778, "step": 947 }, { "epoch": 0.7904940587867417, "grad_norm": 0.12078487873077393, "learning_rate": 1.723583731281144e-06, "loss": 0.7115, "step": 948 }, { "epoch": 0.7913279132791328, "grad_norm": 0.10817496478557587, "learning_rate": 1.7229633529361051e-06, "loss": 0.6699, "step": 949 }, { "epoch": 0.7921617677715239, "grad_norm": 0.10841819643974304, "learning_rate": 1.7223423911127313e-06, "loss": 0.6089, "step": 950 }, { "epoch": 0.792995622263915, "grad_norm": 0.10630793124437332, "learning_rate": 1.7217208463121788e-06, "loss": 0.6615, "step": 951 }, { "epoch": 0.793829476756306, "grad_norm": 0.1000799611210823, "learning_rate": 1.721098719036075e-06, "loss": 0.6231, "step": 952 }, { "epoch": 0.7946633312486971, "grad_norm": 0.11009353399276733, "learning_rate": 1.7204760097865167e-06, "loss": 0.6197, "step": 953 }, { "epoch": 0.7954971857410882, "grad_norm": 0.10809972882270813, "learning_rate": 1.7198527190660706e-06, "loss": 0.5975, "step": 954 }, { "epoch": 0.7963310402334792, "grad_norm": 0.10163717716932297, "learning_rate": 1.719228847377773e-06, "loss": 0.7125, "step": 955 }, { "epoch": 0.7971648947258704, "grad_norm": 0.1121022030711174, "learning_rate": 1.7186043952251286e-06, "loss": 0.6821, "step": 956 }, { "epoch": 0.7979987492182614, "grad_norm": 0.11300718039274216, "learning_rate": 1.7179793631121106e-06, "loss": 0.6137, "step": 957 }, { "epoch": 0.7988326037106525, "grad_norm": 0.10873489826917648, "learning_rate": 1.717353751543161e-06, "loss": 0.5621, "step": 958 }, { "epoch": 0.7996664582030436, "grad_norm": 0.11031791567802429, "learning_rate": 1.7167275610231887e-06, "loss": 0.6363, "step": 959 }, { "epoch": 0.8005003126954346, "grad_norm": 0.10201506316661835, "learning_rate": 1.7161007920575704e-06, "loss": 0.5696, "step": 960 }, { "epoch": 0.8013341671878257, "grad_norm": 0.09877403825521469, "learning_rate": 1.7154734451521486e-06, "loss": 0.6027, "step": 961 }, { "epoch": 0.8021680216802168, "grad_norm": 0.1055438295006752, "learning_rate": 1.7148455208132334e-06, "loss": 0.6204, "step": 962 }, { "epoch": 0.8030018761726079, "grad_norm": 0.11029175668954849, "learning_rate": 1.7142170195476005e-06, "loss": 0.5826, "step": 963 }, { "epoch": 0.8038357306649989, "grad_norm": 0.1105756014585495, "learning_rate": 1.7135879418624913e-06, "loss": 0.6526, "step": 964 }, { "epoch": 0.8046695851573901, "grad_norm": 0.1054423451423645, "learning_rate": 1.712958288265612e-06, "loss": 0.6254, "step": 965 }, { "epoch": 0.8055034396497811, "grad_norm": 0.10613211989402771, "learning_rate": 1.712328059265134e-06, "loss": 0.5974, "step": 966 }, { "epoch": 0.8063372941421721, "grad_norm": 0.09556613117456436, "learning_rate": 1.7116972553696932e-06, "loss": 0.6144, "step": 967 }, { "epoch": 0.8071711486345633, "grad_norm": 0.10485559701919556, "learning_rate": 1.711065877088389e-06, "loss": 0.5998, "step": 968 }, { "epoch": 0.8080050031269543, "grad_norm": 0.10885662585496902, "learning_rate": 1.7104339249307848e-06, "loss": 0.6298, "step": 969 }, { "epoch": 0.8088388576193454, "grad_norm": 0.10140252858400345, "learning_rate": 1.709801399406907e-06, "loss": 0.5747, "step": 970 }, { "epoch": 0.8096727121117365, "grad_norm": 0.09874554723501205, "learning_rate": 1.7091683010272446e-06, "loss": 0.5676, "step": 971 }, { "epoch": 0.8105065666041276, "grad_norm": 0.10777262598276138, "learning_rate": 1.7085346303027493e-06, "loss": 0.6326, "step": 972 }, { "epoch": 0.8113404210965186, "grad_norm": 0.09974364936351776, "learning_rate": 1.7079003877448344e-06, "loss": 0.5638, "step": 973 }, { "epoch": 0.8121742755889098, "grad_norm": 0.10661293566226959, "learning_rate": 1.7072655738653745e-06, "loss": 0.6641, "step": 974 }, { "epoch": 0.8130081300813008, "grad_norm": 0.10669861733913422, "learning_rate": 1.7066301891767061e-06, "loss": 0.6478, "step": 975 }, { "epoch": 0.813841984573692, "grad_norm": 0.10368253290653229, "learning_rate": 1.7059942341916256e-06, "loss": 0.6198, "step": 976 }, { "epoch": 0.814675839066083, "grad_norm": 0.10855797678232193, "learning_rate": 1.7053577094233897e-06, "loss": 0.6453, "step": 977 }, { "epoch": 0.815509693558474, "grad_norm": 0.11007635295391083, "learning_rate": 1.7047206153857156e-06, "loss": 0.5748, "step": 978 }, { "epoch": 0.8163435480508652, "grad_norm": 0.10489246994256973, "learning_rate": 1.7040829525927796e-06, "loss": 0.579, "step": 979 }, { "epoch": 0.8171774025432562, "grad_norm": 0.11224298179149628, "learning_rate": 1.7034447215592164e-06, "loss": 0.6484, "step": 980 }, { "epoch": 0.8180112570356473, "grad_norm": 0.10526615381240845, "learning_rate": 1.7028059228001204e-06, "loss": 0.6746, "step": 981 }, { "epoch": 0.8188451115280384, "grad_norm": 0.10187442600727081, "learning_rate": 1.7021665568310435e-06, "loss": 0.6462, "step": 982 }, { "epoch": 0.8196789660204294, "grad_norm": 0.10846269875764847, "learning_rate": 1.7015266241679952e-06, "loss": 0.6329, "step": 983 }, { "epoch": 0.8205128205128205, "grad_norm": 0.110533706843853, "learning_rate": 1.7008861253274429e-06, "loss": 0.658, "step": 984 }, { "epoch": 0.8213466750052116, "grad_norm": 0.10688811540603638, "learning_rate": 1.7002450608263107e-06, "loss": 0.5996, "step": 985 }, { "epoch": 0.8221805294976027, "grad_norm": 0.10414768755435944, "learning_rate": 1.6996034311819796e-06, "loss": 0.6333, "step": 986 }, { "epoch": 0.8230143839899937, "grad_norm": 0.0999814048409462, "learning_rate": 1.698961236912286e-06, "loss": 0.6609, "step": 987 }, { "epoch": 0.8238482384823849, "grad_norm": 0.10203441977500916, "learning_rate": 1.6983184785355222e-06, "loss": 0.6095, "step": 988 }, { "epoch": 0.8246820929747759, "grad_norm": 0.10571661591529846, "learning_rate": 1.6976751565704362e-06, "loss": 0.6125, "step": 989 }, { "epoch": 0.8255159474671669, "grad_norm": 0.10715372860431671, "learning_rate": 1.6970312715362304e-06, "loss": 0.689, "step": 990 }, { "epoch": 0.8263498019595581, "grad_norm": 0.10807687044143677, "learning_rate": 1.696386823952562e-06, "loss": 0.611, "step": 991 }, { "epoch": 0.8271836564519491, "grad_norm": 0.10625305026769638, "learning_rate": 1.6957418143395418e-06, "loss": 0.6616, "step": 992 }, { "epoch": 0.8280175109443402, "grad_norm": 0.10512302070856094, "learning_rate": 1.6950962432177348e-06, "loss": 0.6147, "step": 993 }, { "epoch": 0.8288513654367313, "grad_norm": 0.10524202138185501, "learning_rate": 1.6944501111081579e-06, "loss": 0.6094, "step": 994 }, { "epoch": 0.8296852199291224, "grad_norm": 0.10839590430259705, "learning_rate": 1.6938034185322828e-06, "loss": 0.5816, "step": 995 }, { "epoch": 0.8305190744215134, "grad_norm": 0.10930492728948593, "learning_rate": 1.6931561660120312e-06, "loss": 0.6621, "step": 996 }, { "epoch": 0.8313529289139046, "grad_norm": 0.10111601650714874, "learning_rate": 1.692508354069779e-06, "loss": 0.6083, "step": 997 }, { "epoch": 0.8321867834062956, "grad_norm": 0.11010827124118805, "learning_rate": 1.691859983228352e-06, "loss": 0.6689, "step": 998 }, { "epoch": 0.8330206378986866, "grad_norm": 0.10622293502092361, "learning_rate": 1.6912110540110272e-06, "loss": 0.6142, "step": 999 }, { "epoch": 0.8338544923910778, "grad_norm": 0.10322947055101395, "learning_rate": 1.6905615669415325e-06, "loss": 0.5797, "step": 1000 }, { "epoch": 0.8346883468834688, "grad_norm": 0.11376485228538513, "learning_rate": 1.689911522544047e-06, "loss": 0.6323, "step": 1001 }, { "epoch": 0.8355222013758599, "grad_norm": 0.10336299985647202, "learning_rate": 1.6892609213431981e-06, "loss": 0.5907, "step": 1002 }, { "epoch": 0.836356055868251, "grad_norm": 0.10549134016036987, "learning_rate": 1.6886097638640631e-06, "loss": 0.6483, "step": 1003 }, { "epoch": 0.837189910360642, "grad_norm": 0.10979744791984558, "learning_rate": 1.6879580506321687e-06, "loss": 0.5829, "step": 1004 }, { "epoch": 0.8380237648530331, "grad_norm": 0.10647737979888916, "learning_rate": 1.6873057821734894e-06, "loss": 0.5714, "step": 1005 }, { "epoch": 0.8388576193454242, "grad_norm": 0.11350049823522568, "learning_rate": 1.6866529590144485e-06, "loss": 0.6756, "step": 1006 }, { "epoch": 0.8396914738378153, "grad_norm": 0.10672125220298767, "learning_rate": 1.6859995816819167e-06, "loss": 0.6459, "step": 1007 }, { "epoch": 0.8405253283302064, "grad_norm": 0.11561840772628784, "learning_rate": 1.6853456507032112e-06, "loss": 0.6912, "step": 1008 }, { "epoch": 0.8413591828225975, "grad_norm": 0.11554259806871414, "learning_rate": 1.6846911666060973e-06, "loss": 0.6032, "step": 1009 }, { "epoch": 0.8421930373149885, "grad_norm": 0.10536365956068039, "learning_rate": 1.6840361299187857e-06, "loss": 0.6182, "step": 1010 }, { "epoch": 0.8430268918073797, "grad_norm": 0.1125800758600235, "learning_rate": 1.6833805411699344e-06, "loss": 0.6519, "step": 1011 }, { "epoch": 0.8438607462997707, "grad_norm": 0.10594160109758377, "learning_rate": 1.6827244008886453e-06, "loss": 0.6407, "step": 1012 }, { "epoch": 0.8446946007921617, "grad_norm": 0.10480530560016632, "learning_rate": 1.6820677096044667e-06, "loss": 0.6398, "step": 1013 }, { "epoch": 0.8455284552845529, "grad_norm": 0.10666308552026749, "learning_rate": 1.6814104678473905e-06, "loss": 0.612, "step": 1014 }, { "epoch": 0.8463623097769439, "grad_norm": 0.10556510835886002, "learning_rate": 1.6807526761478533e-06, "loss": 0.5729, "step": 1015 }, { "epoch": 0.847196164269335, "grad_norm": 0.11039448529481888, "learning_rate": 1.6800943350367368e-06, "loss": 0.6323, "step": 1016 }, { "epoch": 0.8480300187617261, "grad_norm": 0.10304141789674759, "learning_rate": 1.6794354450453638e-06, "loss": 0.6426, "step": 1017 }, { "epoch": 0.8488638732541172, "grad_norm": 0.10969959199428558, "learning_rate": 1.678776006705502e-06, "loss": 0.558, "step": 1018 }, { "epoch": 0.8496977277465082, "grad_norm": 0.10663289576768875, "learning_rate": 1.6781160205493605e-06, "loss": 0.593, "step": 1019 }, { "epoch": 0.8505315822388994, "grad_norm": 0.10291384905576706, "learning_rate": 1.6774554871095915e-06, "loss": 0.6281, "step": 1020 }, { "epoch": 0.8513654367312904, "grad_norm": 0.10907971113920212, "learning_rate": 1.6767944069192876e-06, "loss": 0.6158, "step": 1021 }, { "epoch": 0.8521992912236814, "grad_norm": 0.11564616858959198, "learning_rate": 1.6761327805119838e-06, "loss": 0.6662, "step": 1022 }, { "epoch": 0.8530331457160726, "grad_norm": 0.10590541362762451, "learning_rate": 1.6754706084216555e-06, "loss": 0.6323, "step": 1023 }, { "epoch": 0.8538670002084636, "grad_norm": 0.10117157548666, "learning_rate": 1.6748078911827187e-06, "loss": 0.6112, "step": 1024 }, { "epoch": 0.8547008547008547, "grad_norm": 0.11531368643045425, "learning_rate": 1.674144629330029e-06, "loss": 0.6413, "step": 1025 }, { "epoch": 0.8555347091932458, "grad_norm": 0.10621378570795059, "learning_rate": 1.673480823398882e-06, "loss": 0.6394, "step": 1026 }, { "epoch": 0.8563685636856369, "grad_norm": 0.11612808704376221, "learning_rate": 1.672816473925012e-06, "loss": 0.637, "step": 1027 }, { "epoch": 0.8572024181780279, "grad_norm": 0.10893040150403976, "learning_rate": 1.672151581444592e-06, "loss": 0.5923, "step": 1028 }, { "epoch": 0.858036272670419, "grad_norm": 0.10822536796331406, "learning_rate": 1.6714861464942333e-06, "loss": 0.6766, "step": 1029 }, { "epoch": 0.8588701271628101, "grad_norm": 0.10714786499738693, "learning_rate": 1.6708201696109856e-06, "loss": 0.5649, "step": 1030 }, { "epoch": 0.8597039816552011, "grad_norm": 0.10966484993696213, "learning_rate": 1.6701536513323349e-06, "loss": 0.6219, "step": 1031 }, { "epoch": 0.8605378361475923, "grad_norm": 0.1037144884467125, "learning_rate": 1.669486592196205e-06, "loss": 0.6352, "step": 1032 }, { "epoch": 0.8613716906399833, "grad_norm": 0.11299126595258713, "learning_rate": 1.668818992740956e-06, "loss": 0.6734, "step": 1033 }, { "epoch": 0.8622055451323744, "grad_norm": 0.11029834300279617, "learning_rate": 1.6681508535053834e-06, "loss": 0.6579, "step": 1034 }, { "epoch": 0.8630393996247655, "grad_norm": 0.10519642382860184, "learning_rate": 1.6674821750287197e-06, "loss": 0.6072, "step": 1035 }, { "epoch": 0.8638732541171565, "grad_norm": 0.11065202951431274, "learning_rate": 1.6668129578506313e-06, "loss": 0.6562, "step": 1036 }, { "epoch": 0.8647071086095476, "grad_norm": 0.11024871468544006, "learning_rate": 1.66614320251122e-06, "loss": 0.6, "step": 1037 }, { "epoch": 0.8655409631019387, "grad_norm": 0.10754359513521194, "learning_rate": 1.6654729095510219e-06, "loss": 0.6028, "step": 1038 }, { "epoch": 0.8663748175943298, "grad_norm": 0.11345981806516647, "learning_rate": 1.6648020795110069e-06, "loss": 0.7018, "step": 1039 }, { "epoch": 0.8672086720867209, "grad_norm": 0.10346484929323196, "learning_rate": 1.6641307129325783e-06, "loss": 0.6866, "step": 1040 }, { "epoch": 0.868042526579112, "grad_norm": 0.10640288889408112, "learning_rate": 1.6634588103575723e-06, "loss": 0.5851, "step": 1041 }, { "epoch": 0.868876381071503, "grad_norm": 0.11161711812019348, "learning_rate": 1.6627863723282584e-06, "loss": 0.6182, "step": 1042 }, { "epoch": 0.8697102355638942, "grad_norm": 0.10999694466590881, "learning_rate": 1.662113399387337e-06, "loss": 0.5887, "step": 1043 }, { "epoch": 0.8705440900562852, "grad_norm": 0.10471642017364502, "learning_rate": 1.6614398920779418e-06, "loss": 0.6854, "step": 1044 }, { "epoch": 0.8713779445486762, "grad_norm": 0.10635983943939209, "learning_rate": 1.660765850943636e-06, "loss": 0.5945, "step": 1045 }, { "epoch": 0.8722117990410674, "grad_norm": 0.10250985622406006, "learning_rate": 1.6600912765284153e-06, "loss": 0.6283, "step": 1046 }, { "epoch": 0.8730456535334584, "grad_norm": 0.10342993587255478, "learning_rate": 1.6594161693767046e-06, "loss": 0.5878, "step": 1047 }, { "epoch": 0.8738795080258495, "grad_norm": 0.10388782620429993, "learning_rate": 1.6587405300333593e-06, "loss": 0.6197, "step": 1048 }, { "epoch": 0.8747133625182406, "grad_norm": 0.10757216066122055, "learning_rate": 1.658064359043664e-06, "loss": 0.6643, "step": 1049 }, { "epoch": 0.8755472170106317, "grad_norm": 0.10185371339321136, "learning_rate": 1.657387656953333e-06, "loss": 0.6129, "step": 1050 }, { "epoch": 0.8763810715030227, "grad_norm": 0.10625848174095154, "learning_rate": 1.6567104243085081e-06, "loss": 0.6389, "step": 1051 }, { "epoch": 0.8772149259954138, "grad_norm": 0.1068265438079834, "learning_rate": 1.6560326616557605e-06, "loss": 0.6474, "step": 1052 }, { "epoch": 0.8780487804878049, "grad_norm": 0.10008926689624786, "learning_rate": 1.6553543695420888e-06, "loss": 0.592, "step": 1053 }, { "epoch": 0.8788826349801959, "grad_norm": 0.11250849813222885, "learning_rate": 1.6546755485149181e-06, "loss": 0.6612, "step": 1054 }, { "epoch": 0.8797164894725871, "grad_norm": 0.10780669748783112, "learning_rate": 1.6539961991221017e-06, "loss": 0.5717, "step": 1055 }, { "epoch": 0.8805503439649781, "grad_norm": 0.10522449761629105, "learning_rate": 1.6533163219119181e-06, "loss": 0.6377, "step": 1056 }, { "epoch": 0.8813841984573692, "grad_norm": 0.11719299852848053, "learning_rate": 1.6526359174330727e-06, "loss": 0.6148, "step": 1057 }, { "epoch": 0.8822180529497603, "grad_norm": 0.10418464988470078, "learning_rate": 1.6519549862346959e-06, "loss": 0.547, "step": 1058 }, { "epoch": 0.8830519074421513, "grad_norm": 0.1149667277932167, "learning_rate": 1.6512735288663433e-06, "loss": 0.5509, "step": 1059 }, { "epoch": 0.8838857619345424, "grad_norm": 0.11353281885385513, "learning_rate": 1.6505915458779954e-06, "loss": 0.6679, "step": 1060 }, { "epoch": 0.8847196164269335, "grad_norm": 0.10084139555692673, "learning_rate": 1.6499090378200564e-06, "loss": 0.6315, "step": 1061 }, { "epoch": 0.8855534709193246, "grad_norm": 0.10753504931926727, "learning_rate": 1.6492260052433551e-06, "loss": 0.6441, "step": 1062 }, { "epoch": 0.8863873254117156, "grad_norm": 0.10944227129220963, "learning_rate": 1.6485424486991427e-06, "loss": 0.6378, "step": 1063 }, { "epoch": 0.8872211799041068, "grad_norm": 0.10542403161525726, "learning_rate": 1.6478583687390937e-06, "loss": 0.5715, "step": 1064 }, { "epoch": 0.8880550343964978, "grad_norm": 0.10192226618528366, "learning_rate": 1.6471737659153054e-06, "loss": 0.6035, "step": 1065 }, { "epoch": 0.8888888888888888, "grad_norm": 0.11085714399814606, "learning_rate": 1.6464886407802958e-06, "loss": 0.6705, "step": 1066 }, { "epoch": 0.88972274338128, "grad_norm": 0.10521090775728226, "learning_rate": 1.645802993887006e-06, "loss": 0.6487, "step": 1067 }, { "epoch": 0.890556597873671, "grad_norm": 0.11182847619056702, "learning_rate": 1.645116825788798e-06, "loss": 0.5908, "step": 1068 }, { "epoch": 0.8913904523660621, "grad_norm": 0.10744032263755798, "learning_rate": 1.6444301370394533e-06, "loss": 0.6978, "step": 1069 }, { "epoch": 0.8922243068584532, "grad_norm": 0.11645323783159256, "learning_rate": 1.6437429281931742e-06, "loss": 0.6552, "step": 1070 }, { "epoch": 0.8930581613508443, "grad_norm": 0.10556191205978394, "learning_rate": 1.6430551998045833e-06, "loss": 0.6485, "step": 1071 }, { "epoch": 0.8938920158432354, "grad_norm": 0.11244357377290726, "learning_rate": 1.6423669524287216e-06, "loss": 0.6352, "step": 1072 }, { "epoch": 0.8947258703356264, "grad_norm": 0.11015337705612183, "learning_rate": 1.6416781866210494e-06, "loss": 0.5928, "step": 1073 }, { "epoch": 0.8955597248280175, "grad_norm": 0.11401670426130295, "learning_rate": 1.6409889029374457e-06, "loss": 0.591, "step": 1074 }, { "epoch": 0.8963935793204086, "grad_norm": 0.10178755223751068, "learning_rate": 1.6402991019342073e-06, "loss": 0.5772, "step": 1075 }, { "epoch": 0.8972274338127997, "grad_norm": 0.10722572356462479, "learning_rate": 1.6396087841680478e-06, "loss": 0.616, "step": 1076 }, { "epoch": 0.8980612883051907, "grad_norm": 0.10824041068553925, "learning_rate": 1.6389179501960987e-06, "loss": 0.6289, "step": 1077 }, { "epoch": 0.8988951427975819, "grad_norm": 0.10898349434137344, "learning_rate": 1.638226600575908e-06, "loss": 0.6197, "step": 1078 }, { "epoch": 0.8997289972899729, "grad_norm": 0.10828059166669846, "learning_rate": 1.6375347358654397e-06, "loss": 0.6335, "step": 1079 }, { "epoch": 0.900562851782364, "grad_norm": 0.11353094130754471, "learning_rate": 1.6368423566230728e-06, "loss": 0.6306, "step": 1080 }, { "epoch": 0.9013967062747551, "grad_norm": 0.1095949038863182, "learning_rate": 1.6361494634076033e-06, "loss": 0.6579, "step": 1081 }, { "epoch": 0.9022305607671461, "grad_norm": 0.10779423266649246, "learning_rate": 1.6354560567782406e-06, "loss": 0.6268, "step": 1082 }, { "epoch": 0.9030644152595372, "grad_norm": 0.10539865493774414, "learning_rate": 1.6347621372946088e-06, "loss": 0.6318, "step": 1083 }, { "epoch": 0.9038982697519283, "grad_norm": 0.11474636197090149, "learning_rate": 1.6340677055167458e-06, "loss": 0.6843, "step": 1084 }, { "epoch": 0.9047321242443194, "grad_norm": 0.11284147948026657, "learning_rate": 1.633372762005103e-06, "loss": 0.6167, "step": 1085 }, { "epoch": 0.9055659787367104, "grad_norm": 0.10842804610729218, "learning_rate": 1.632677307320545e-06, "loss": 0.5851, "step": 1086 }, { "epoch": 0.9063998332291016, "grad_norm": 0.10843487083911896, "learning_rate": 1.6319813420243495e-06, "loss": 0.6337, "step": 1087 }, { "epoch": 0.9072336877214926, "grad_norm": 0.10792160034179688, "learning_rate": 1.6312848666782048e-06, "loss": 0.6015, "step": 1088 }, { "epoch": 0.9080675422138836, "grad_norm": 0.10741087794303894, "learning_rate": 1.6305878818442122e-06, "loss": 0.6285, "step": 1089 }, { "epoch": 0.9089013967062748, "grad_norm": 0.10371655970811844, "learning_rate": 1.6298903880848834e-06, "loss": 0.5982, "step": 1090 }, { "epoch": 0.9097352511986658, "grad_norm": 0.10612435638904572, "learning_rate": 1.6291923859631415e-06, "loss": 0.5962, "step": 1091 }, { "epoch": 0.9105691056910569, "grad_norm": 0.10705637186765671, "learning_rate": 1.6284938760423188e-06, "loss": 0.5926, "step": 1092 }, { "epoch": 0.911402960183448, "grad_norm": 0.10797090083360672, "learning_rate": 1.627794858886159e-06, "loss": 0.632, "step": 1093 }, { "epoch": 0.9122368146758391, "grad_norm": 0.1132771223783493, "learning_rate": 1.6270953350588137e-06, "loss": 0.647, "step": 1094 }, { "epoch": 0.9130706691682301, "grad_norm": 0.10793612152338028, "learning_rate": 1.626395305124844e-06, "loss": 0.6367, "step": 1095 }, { "epoch": 0.9139045236606212, "grad_norm": 0.10706419497728348, "learning_rate": 1.6256947696492196e-06, "loss": 0.6472, "step": 1096 }, { "epoch": 0.9147383781530123, "grad_norm": 0.10842972993850708, "learning_rate": 1.6249937291973184e-06, "loss": 0.6185, "step": 1097 }, { "epoch": 0.9155722326454033, "grad_norm": 0.11361444741487503, "learning_rate": 1.6242921843349252e-06, "loss": 0.617, "step": 1098 }, { "epoch": 0.9164060871377945, "grad_norm": 0.11315753310918808, "learning_rate": 1.6235901356282322e-06, "loss": 0.6478, "step": 1099 }, { "epoch": 0.9172399416301855, "grad_norm": 0.12495766580104828, "learning_rate": 1.6228875836438385e-06, "loss": 0.6252, "step": 1100 }, { "epoch": 0.9180737961225766, "grad_norm": 0.11157601326704025, "learning_rate": 1.622184528948749e-06, "loss": 0.6214, "step": 1101 }, { "epoch": 0.9189076506149677, "grad_norm": 0.10694071650505066, "learning_rate": 1.6214809721103744e-06, "loss": 0.619, "step": 1102 }, { "epoch": 0.9197415051073587, "grad_norm": 0.1112818494439125, "learning_rate": 1.6207769136965307e-06, "loss": 0.6229, "step": 1103 }, { "epoch": 0.9205753595997499, "grad_norm": 0.10422814637422562, "learning_rate": 1.6200723542754389e-06, "loss": 0.5837, "step": 1104 }, { "epoch": 0.9214092140921409, "grad_norm": 0.11240324378013611, "learning_rate": 1.619367294415724e-06, "loss": 0.7071, "step": 1105 }, { "epoch": 0.922243068584532, "grad_norm": 0.10819468647241592, "learning_rate": 1.6186617346864151e-06, "loss": 0.5728, "step": 1106 }, { "epoch": 0.9230769230769231, "grad_norm": 0.1053207591176033, "learning_rate": 1.6179556756569448e-06, "loss": 0.5915, "step": 1107 }, { "epoch": 0.9239107775693142, "grad_norm": 0.1075674444437027, "learning_rate": 1.6172491178971482e-06, "loss": 0.5935, "step": 1108 }, { "epoch": 0.9247446320617052, "grad_norm": 0.11159633845090866, "learning_rate": 1.6165420619772635e-06, "loss": 0.6493, "step": 1109 }, { "epoch": 0.9255784865540964, "grad_norm": 0.11671009659767151, "learning_rate": 1.6158345084679307e-06, "loss": 0.6661, "step": 1110 }, { "epoch": 0.9264123410464874, "grad_norm": 0.11934785544872284, "learning_rate": 1.6151264579401917e-06, "loss": 0.6335, "step": 1111 }, { "epoch": 0.9272461955388784, "grad_norm": 0.11291314661502838, "learning_rate": 1.6144179109654887e-06, "loss": 0.6413, "step": 1112 }, { "epoch": 0.9280800500312696, "grad_norm": 0.10902924090623856, "learning_rate": 1.6137088681156654e-06, "loss": 0.6713, "step": 1113 }, { "epoch": 0.9289139045236606, "grad_norm": 0.10608460754156113, "learning_rate": 1.6129993299629651e-06, "loss": 0.6024, "step": 1114 }, { "epoch": 0.9297477590160517, "grad_norm": 0.10624940693378448, "learning_rate": 1.6122892970800317e-06, "loss": 0.6128, "step": 1115 }, { "epoch": 0.9305816135084428, "grad_norm": 0.11591742187738419, "learning_rate": 1.6115787700399071e-06, "loss": 0.6242, "step": 1116 }, { "epoch": 0.9314154680008339, "grad_norm": 0.11706274002790451, "learning_rate": 1.610867749416033e-06, "loss": 0.6356, "step": 1117 }, { "epoch": 0.9322493224932249, "grad_norm": 0.10509736090898514, "learning_rate": 1.6101562357822491e-06, "loss": 0.6337, "step": 1118 }, { "epoch": 0.933083176985616, "grad_norm": 0.10801702737808228, "learning_rate": 1.6094442297127935e-06, "loss": 0.615, "step": 1119 }, { "epoch": 0.9339170314780071, "grad_norm": 0.11627451330423355, "learning_rate": 1.6087317317823007e-06, "loss": 0.6059, "step": 1120 }, { "epoch": 0.9347508859703981, "grad_norm": 0.11316727101802826, "learning_rate": 1.6080187425658033e-06, "loss": 0.6092, "step": 1121 }, { "epoch": 0.9355847404627893, "grad_norm": 0.10899489372968674, "learning_rate": 1.6073052626387296e-06, "loss": 0.5787, "step": 1122 }, { "epoch": 0.9364185949551803, "grad_norm": 0.10977214574813843, "learning_rate": 1.606591292576904e-06, "loss": 0.6022, "step": 1123 }, { "epoch": 0.9372524494475714, "grad_norm": 0.10926050692796707, "learning_rate": 1.6058768329565469e-06, "loss": 0.6626, "step": 1124 }, { "epoch": 0.9380863039399625, "grad_norm": 0.10779014229774475, "learning_rate": 1.6051618843542736e-06, "loss": 0.6381, "step": 1125 }, { "epoch": 0.9389201584323535, "grad_norm": 0.11509402096271515, "learning_rate": 1.6044464473470939e-06, "loss": 0.5817, "step": 1126 }, { "epoch": 0.9397540129247446, "grad_norm": 0.10561665147542953, "learning_rate": 1.6037305225124121e-06, "loss": 0.5634, "step": 1127 }, { "epoch": 0.9405878674171357, "grad_norm": 0.1081278994679451, "learning_rate": 1.6030141104280253e-06, "loss": 0.6437, "step": 1128 }, { "epoch": 0.9414217219095268, "grad_norm": 0.10470031946897507, "learning_rate": 1.6022972116721256e-06, "loss": 0.5853, "step": 1129 }, { "epoch": 0.9422555764019178, "grad_norm": 0.1069476306438446, "learning_rate": 1.601579826823296e-06, "loss": 0.5653, "step": 1130 }, { "epoch": 0.943089430894309, "grad_norm": 0.1095786914229393, "learning_rate": 1.600861956460513e-06, "loss": 0.6191, "step": 1131 }, { "epoch": 0.9439232853867, "grad_norm": 0.10863108187913895, "learning_rate": 1.600143601163144e-06, "loss": 0.6166, "step": 1132 }, { "epoch": 0.944757139879091, "grad_norm": 0.110601507127285, "learning_rate": 1.5994247615109498e-06, "loss": 0.5716, "step": 1133 }, { "epoch": 0.9455909943714822, "grad_norm": 0.108455128967762, "learning_rate": 1.598705438084079e-06, "loss": 0.5888, "step": 1134 }, { "epoch": 0.9464248488638732, "grad_norm": 0.11156295984983444, "learning_rate": 1.5979856314630728e-06, "loss": 0.6723, "step": 1135 }, { "epoch": 0.9472587033562644, "grad_norm": 0.11098117381334305, "learning_rate": 1.5972653422288622e-06, "loss": 0.5671, "step": 1136 }, { "epoch": 0.9480925578486554, "grad_norm": 0.11285511404275894, "learning_rate": 1.5965445709627669e-06, "loss": 0.6437, "step": 1137 }, { "epoch": 0.9489264123410465, "grad_norm": 0.10554268956184387, "learning_rate": 1.5958233182464965e-06, "loss": 0.6289, "step": 1138 }, { "epoch": 0.9497602668334376, "grad_norm": 0.10124699026346207, "learning_rate": 1.5951015846621483e-06, "loss": 0.6068, "step": 1139 }, { "epoch": 0.9505941213258287, "grad_norm": 0.11128734052181244, "learning_rate": 1.5943793707922084e-06, "loss": 0.5907, "step": 1140 }, { "epoch": 0.9514279758182197, "grad_norm": 0.11676076054573059, "learning_rate": 1.5936566772195503e-06, "loss": 0.602, "step": 1141 }, { "epoch": 0.9522618303106108, "grad_norm": 0.11664719134569168, "learning_rate": 1.5929335045274343e-06, "loss": 0.6197, "step": 1142 }, { "epoch": 0.9530956848030019, "grad_norm": 0.1060841903090477, "learning_rate": 1.5922098532995083e-06, "loss": 0.553, "step": 1143 }, { "epoch": 0.9539295392953929, "grad_norm": 0.11080587655305862, "learning_rate": 1.591485724119805e-06, "loss": 0.6357, "step": 1144 }, { "epoch": 0.9547633937877841, "grad_norm": 0.10428017377853394, "learning_rate": 1.5907611175727442e-06, "loss": 0.5693, "step": 1145 }, { "epoch": 0.9555972482801751, "grad_norm": 0.10886859893798828, "learning_rate": 1.59003603424313e-06, "loss": 0.6142, "step": 1146 }, { "epoch": 0.9564311027725662, "grad_norm": 0.1091897264122963, "learning_rate": 1.5893104747161522e-06, "loss": 0.6234, "step": 1147 }, { "epoch": 0.9572649572649573, "grad_norm": 0.11705927550792694, "learning_rate": 1.5885844395773841e-06, "loss": 0.6322, "step": 1148 }, { "epoch": 0.9580988117573483, "grad_norm": 0.11114629358053207, "learning_rate": 1.5878579294127831e-06, "loss": 0.6233, "step": 1149 }, { "epoch": 0.9589326662497394, "grad_norm": 0.11006072908639908, "learning_rate": 1.5871309448086903e-06, "loss": 0.6271, "step": 1150 }, { "epoch": 0.9597665207421305, "grad_norm": 0.10859903693199158, "learning_rate": 1.5864034863518292e-06, "loss": 0.6101, "step": 1151 }, { "epoch": 0.9606003752345216, "grad_norm": 0.11398789286613464, "learning_rate": 1.585675554629306e-06, "loss": 0.6183, "step": 1152 }, { "epoch": 0.9614342297269126, "grad_norm": 0.10914972424507141, "learning_rate": 1.5849471502286087e-06, "loss": 0.6687, "step": 1153 }, { "epoch": 0.9622680842193038, "grad_norm": 0.11407602578401566, "learning_rate": 1.5842182737376074e-06, "loss": 0.6513, "step": 1154 }, { "epoch": 0.9631019387116948, "grad_norm": 0.11537064611911774, "learning_rate": 1.5834889257445525e-06, "loss": 0.6284, "step": 1155 }, { "epoch": 0.9639357932040858, "grad_norm": 0.10912200808525085, "learning_rate": 1.582759106838075e-06, "loss": 0.634, "step": 1156 }, { "epoch": 0.964769647696477, "grad_norm": 0.11783526092767715, "learning_rate": 1.5820288176071861e-06, "loss": 0.6435, "step": 1157 }, { "epoch": 0.965603502188868, "grad_norm": 0.11490530520677567, "learning_rate": 1.5812980586412765e-06, "loss": 0.5872, "step": 1158 }, { "epoch": 0.9664373566812591, "grad_norm": 0.10283337533473969, "learning_rate": 1.5805668305301167e-06, "loss": 0.565, "step": 1159 }, { "epoch": 0.9672712111736502, "grad_norm": 0.10747389495372772, "learning_rate": 1.5798351338638548e-06, "loss": 0.5878, "step": 1160 }, { "epoch": 0.9681050656660413, "grad_norm": 0.11634895950555801, "learning_rate": 1.5791029692330172e-06, "loss": 0.6547, "step": 1161 }, { "epoch": 0.9689389201584323, "grad_norm": 0.11201413720846176, "learning_rate": 1.5783703372285086e-06, "loss": 0.6039, "step": 1162 }, { "epoch": 0.9697727746508235, "grad_norm": 0.11117483675479889, "learning_rate": 1.5776372384416105e-06, "loss": 0.6447, "step": 1163 }, { "epoch": 0.9706066291432145, "grad_norm": 0.11274691671133041, "learning_rate": 1.5769036734639815e-06, "loss": 0.5847, "step": 1164 }, { "epoch": 0.9714404836356055, "grad_norm": 0.11471108347177505, "learning_rate": 1.5761696428876556e-06, "loss": 0.6523, "step": 1165 }, { "epoch": 0.9722743381279967, "grad_norm": 0.11089842766523361, "learning_rate": 1.5754351473050434e-06, "loss": 0.5949, "step": 1166 }, { "epoch": 0.9731081926203877, "grad_norm": 0.10404518991708755, "learning_rate": 1.5747001873089306e-06, "loss": 0.5496, "step": 1167 }, { "epoch": 0.9739420471127789, "grad_norm": 0.116578109562397, "learning_rate": 1.5739647634924773e-06, "loss": 0.6103, "step": 1168 }, { "epoch": 0.9747759016051699, "grad_norm": 0.1054786667227745, "learning_rate": 1.5732288764492184e-06, "loss": 0.623, "step": 1169 }, { "epoch": 0.975609756097561, "grad_norm": 0.11369701474905014, "learning_rate": 1.5724925267730624e-06, "loss": 0.567, "step": 1170 }, { "epoch": 0.9764436105899521, "grad_norm": 0.1198234111070633, "learning_rate": 1.5717557150582917e-06, "loss": 0.6188, "step": 1171 }, { "epoch": 0.9772774650823431, "grad_norm": 0.12100456655025482, "learning_rate": 1.5710184418995604e-06, "loss": 0.641, "step": 1172 }, { "epoch": 0.9781113195747342, "grad_norm": 0.10783829540014267, "learning_rate": 1.5702807078918965e-06, "loss": 0.5889, "step": 1173 }, { "epoch": 0.9789451740671253, "grad_norm": 0.11204763501882553, "learning_rate": 1.5695425136306987e-06, "loss": 0.6271, "step": 1174 }, { "epoch": 0.9797790285595164, "grad_norm": 0.11363273113965988, "learning_rate": 1.568803859711738e-06, "loss": 0.6099, "step": 1175 }, { "epoch": 0.9806128830519074, "grad_norm": 0.10527540743350983, "learning_rate": 1.5680647467311555e-06, "loss": 0.5782, "step": 1176 }, { "epoch": 0.9814467375442986, "grad_norm": 0.11106487363576889, "learning_rate": 1.5673251752854645e-06, "loss": 0.645, "step": 1177 }, { "epoch": 0.9822805920366896, "grad_norm": 0.11001728475093842, "learning_rate": 1.5665851459715457e-06, "loss": 0.611, "step": 1178 }, { "epoch": 0.9831144465290806, "grad_norm": 0.10988467186689377, "learning_rate": 1.5658446593866517e-06, "loss": 0.6229, "step": 1179 }, { "epoch": 0.9839483010214718, "grad_norm": 0.10910095274448395, "learning_rate": 1.565103716128403e-06, "loss": 0.5964, "step": 1180 }, { "epoch": 0.9847821555138628, "grad_norm": 0.11653497070074081, "learning_rate": 1.5643623167947891e-06, "loss": 0.6451, "step": 1181 }, { "epoch": 0.9856160100062539, "grad_norm": 0.11314839869737625, "learning_rate": 1.5636204619841667e-06, "loss": 0.6419, "step": 1182 }, { "epoch": 0.986449864498645, "grad_norm": 0.10827460885047913, "learning_rate": 1.5628781522952611e-06, "loss": 0.6129, "step": 1183 }, { "epoch": 0.9872837189910361, "grad_norm": 0.1053231805562973, "learning_rate": 1.5621353883271649e-06, "loss": 0.6455, "step": 1184 }, { "epoch": 0.9881175734834271, "grad_norm": 0.11166463792324066, "learning_rate": 1.561392170679336e-06, "loss": 0.594, "step": 1185 }, { "epoch": 0.9889514279758183, "grad_norm": 0.11850026249885559, "learning_rate": 1.5606484999516e-06, "loss": 0.575, "step": 1186 }, { "epoch": 0.9897852824682093, "grad_norm": 0.11536535620689392, "learning_rate": 1.5599043767441471e-06, "loss": 0.645, "step": 1187 }, { "epoch": 0.9906191369606003, "grad_norm": 0.1191059872508049, "learning_rate": 1.5591598016575328e-06, "loss": 0.6007, "step": 1188 }, { "epoch": 0.9914529914529915, "grad_norm": 0.1131712794303894, "learning_rate": 1.558414775292678e-06, "loss": 0.5926, "step": 1189 }, { "epoch": 0.9922868459453825, "grad_norm": 0.11502361297607422, "learning_rate": 1.5576692982508663e-06, "loss": 0.6223, "step": 1190 }, { "epoch": 0.9931207004377736, "grad_norm": 0.10861647129058838, "learning_rate": 1.5569233711337474e-06, "loss": 0.5948, "step": 1191 }, { "epoch": 0.9939545549301647, "grad_norm": 0.11109884828329086, "learning_rate": 1.5561769945433323e-06, "loss": 0.5778, "step": 1192 }, { "epoch": 0.9947884094225558, "grad_norm": 0.1124456375837326, "learning_rate": 1.5554301690819951e-06, "loss": 0.6732, "step": 1193 }, { "epoch": 0.9956222639149468, "grad_norm": 0.10982295870780945, "learning_rate": 1.5546828953524727e-06, "loss": 0.6221, "step": 1194 }, { "epoch": 0.9964561184073379, "grad_norm": 0.10954145342111588, "learning_rate": 1.553935173957863e-06, "loss": 0.6045, "step": 1195 }, { "epoch": 0.997289972899729, "grad_norm": 0.10329638421535492, "learning_rate": 1.5531870055016265e-06, "loss": 0.5827, "step": 1196 }, { "epoch": 0.99812382739212, "grad_norm": 0.1147105023264885, "learning_rate": 1.5524383905875828e-06, "loss": 0.6676, "step": 1197 }, { "epoch": 0.9989576818845112, "grad_norm": 0.10970206558704376, "learning_rate": 1.5516893298199133e-06, "loss": 0.6159, "step": 1198 }, { "epoch": 0.9997915363769022, "grad_norm": 0.10848627239465714, "learning_rate": 1.5509398238031585e-06, "loss": 0.6106, "step": 1199 }, { "epoch": 1.0, "grad_norm": 0.10848627239465714, "learning_rate": 1.5501898731422187e-06, "loss": 0.1335, "step": 1200 }, { "epoch": 1.0, "eval_loss": 0.6266602873802185, "eval_runtime": 321.4177, "eval_samples_per_second": 16.748, "eval_steps_per_second": 2.794, "step": 1200 }, { "epoch": 1.0006253908692933, "grad_norm": 0.10542084276676178, "learning_rate": 1.5494394784423524e-06, "loss": 0.4569, "step": 1201 }, { "epoch": 1.0008338544923912, "grad_norm": 0.11667848378419876, "learning_rate": 1.5486886403091765e-06, "loss": 0.6476, "step": 1202 }, { "epoch": 1.001667708984782, "grad_norm": 0.12027855962514877, "learning_rate": 1.5479373593486666e-06, "loss": 0.6156, "step": 1203 }, { "epoch": 1.0025015634771732, "grad_norm": 0.115195631980896, "learning_rate": 1.547185636167155e-06, "loss": 0.603, "step": 1204 }, { "epoch": 1.0033354179695644, "grad_norm": 0.11124297231435776, "learning_rate": 1.5464334713713309e-06, "loss": 0.6328, "step": 1205 }, { "epoch": 1.0041692724619553, "grad_norm": 0.11163129657506943, "learning_rate": 1.5456808655682405e-06, "loss": 0.658, "step": 1206 }, { "epoch": 1.0050031269543465, "grad_norm": 0.10940606147050858, "learning_rate": 1.5449278193652853e-06, "loss": 0.6146, "step": 1207 }, { "epoch": 1.0058369814467376, "grad_norm": 0.10748612880706787, "learning_rate": 1.5441743333702224e-06, "loss": 0.6022, "step": 1208 }, { "epoch": 1.0066708359391285, "grad_norm": 0.11005198210477829, "learning_rate": 1.5434204081911641e-06, "loss": 0.595, "step": 1209 }, { "epoch": 1.0075046904315197, "grad_norm": 0.11207201331853867, "learning_rate": 1.5426660444365766e-06, "loss": 0.6294, "step": 1210 }, { "epoch": 1.0083385449239108, "grad_norm": 0.12536509335041046, "learning_rate": 1.5419112427152807e-06, "loss": 0.6097, "step": 1211 }, { "epoch": 1.0091723994163018, "grad_norm": 0.11734244227409363, "learning_rate": 1.54115600363645e-06, "loss": 0.6301, "step": 1212 }, { "epoch": 1.010006253908693, "grad_norm": 0.10745465755462646, "learning_rate": 1.5404003278096118e-06, "loss": 0.5472, "step": 1213 }, { "epoch": 1.010840108401084, "grad_norm": 0.10949809849262238, "learning_rate": 1.5396442158446456e-06, "loss": 0.5806, "step": 1214 }, { "epoch": 1.011673962893475, "grad_norm": 0.10704012960195541, "learning_rate": 1.5388876683517825e-06, "loss": 0.6133, "step": 1215 }, { "epoch": 1.0125078173858661, "grad_norm": 0.11208271235227585, "learning_rate": 1.5381306859416054e-06, "loss": 0.6766, "step": 1216 }, { "epoch": 1.0133416718782573, "grad_norm": 0.11088071763515472, "learning_rate": 1.5373732692250484e-06, "loss": 0.6471, "step": 1217 }, { "epoch": 1.0141755263706482, "grad_norm": 0.11159610748291016, "learning_rate": 1.536615418813396e-06, "loss": 0.5996, "step": 1218 }, { "epoch": 1.0150093808630394, "grad_norm": 0.11161819100379944, "learning_rate": 1.5358571353182824e-06, "loss": 0.6552, "step": 1219 }, { "epoch": 1.0158432353554305, "grad_norm": 0.1059797927737236, "learning_rate": 1.5350984193516917e-06, "loss": 0.663, "step": 1220 }, { "epoch": 1.0166770898478215, "grad_norm": 0.11861960589885712, "learning_rate": 1.5343392715259572e-06, "loss": 0.6603, "step": 1221 }, { "epoch": 1.0175109443402126, "grad_norm": 0.1215897724032402, "learning_rate": 1.5335796924537598e-06, "loss": 0.6286, "step": 1222 }, { "epoch": 1.0183447988326038, "grad_norm": 0.10159210860729218, "learning_rate": 1.53281968274813e-06, "loss": 0.571, "step": 1223 }, { "epoch": 1.0191786533249947, "grad_norm": 0.11154468357563019, "learning_rate": 1.5320592430224442e-06, "loss": 0.5887, "step": 1224 }, { "epoch": 1.0200125078173858, "grad_norm": 0.11331478506326675, "learning_rate": 1.5312983738904266e-06, "loss": 0.5954, "step": 1225 }, { "epoch": 1.020846362309777, "grad_norm": 0.11255330592393875, "learning_rate": 1.5305370759661481e-06, "loss": 0.5458, "step": 1226 }, { "epoch": 1.021680216802168, "grad_norm": 0.10835165530443192, "learning_rate": 1.529775349864026e-06, "loss": 0.6382, "step": 1227 }, { "epoch": 1.022514071294559, "grad_norm": 0.11464633047580719, "learning_rate": 1.5290131961988225e-06, "loss": 0.6012, "step": 1228 }, { "epoch": 1.0233479257869502, "grad_norm": 0.11286839097738266, "learning_rate": 1.528250615585644e-06, "loss": 0.6138, "step": 1229 }, { "epoch": 1.0241817802793411, "grad_norm": 0.1153862252831459, "learning_rate": 1.5274876086399441e-06, "loss": 0.6682, "step": 1230 }, { "epoch": 1.0250156347717323, "grad_norm": 0.11663507670164108, "learning_rate": 1.526724175977518e-06, "loss": 0.6318, "step": 1231 }, { "epoch": 1.0258494892641234, "grad_norm": 0.1101766973733902, "learning_rate": 1.5259603182145057e-06, "loss": 0.6107, "step": 1232 }, { "epoch": 1.0266833437565146, "grad_norm": 0.11496511846780777, "learning_rate": 1.5251960359673897e-06, "loss": 0.6275, "step": 1233 }, { "epoch": 1.0275171982489055, "grad_norm": 0.10905829817056656, "learning_rate": 1.5244313298529959e-06, "loss": 0.623, "step": 1234 }, { "epoch": 1.0283510527412967, "grad_norm": 0.10732358694076538, "learning_rate": 1.523666200488491e-06, "loss": 0.5313, "step": 1235 }, { "epoch": 1.0291849072336878, "grad_norm": 0.11587117612361908, "learning_rate": 1.5229006484913847e-06, "loss": 0.6362, "step": 1236 }, { "epoch": 1.0300187617260788, "grad_norm": 0.10961401462554932, "learning_rate": 1.5221346744795267e-06, "loss": 0.6235, "step": 1237 }, { "epoch": 1.03085261621847, "grad_norm": 0.11670937389135361, "learning_rate": 1.521368279071108e-06, "loss": 0.5856, "step": 1238 }, { "epoch": 1.031686470710861, "grad_norm": 0.11099621653556824, "learning_rate": 1.5206014628846593e-06, "loss": 0.5848, "step": 1239 }, { "epoch": 1.032520325203252, "grad_norm": 0.11106451600790024, "learning_rate": 1.5198342265390508e-06, "loss": 0.6128, "step": 1240 }, { "epoch": 1.0333541796956431, "grad_norm": 0.11923278123140335, "learning_rate": 1.5190665706534923e-06, "loss": 0.5776, "step": 1241 }, { "epoch": 1.0341880341880343, "grad_norm": 0.11535177379846573, "learning_rate": 1.5182984958475321e-06, "loss": 0.6155, "step": 1242 }, { "epoch": 1.0350218886804252, "grad_norm": 0.10793700069189072, "learning_rate": 1.5175300027410565e-06, "loss": 0.5962, "step": 1243 }, { "epoch": 1.0358557431728164, "grad_norm": 0.11252836883068085, "learning_rate": 1.5167610919542883e-06, "loss": 0.5944, "step": 1244 }, { "epoch": 1.0366895976652075, "grad_norm": 0.11895468831062317, "learning_rate": 1.5159917641077893e-06, "loss": 0.6314, "step": 1245 }, { "epoch": 1.0375234521575984, "grad_norm": 0.1106170117855072, "learning_rate": 1.515222019822457e-06, "loss": 0.6092, "step": 1246 }, { "epoch": 1.0383573066499896, "grad_norm": 0.11284390091896057, "learning_rate": 1.5144518597195242e-06, "loss": 0.5649, "step": 1247 }, { "epoch": 1.0391911611423807, "grad_norm": 0.1112569123506546, "learning_rate": 1.5136812844205608e-06, "loss": 0.6286, "step": 1248 }, { "epoch": 1.0400250156347717, "grad_norm": 0.12193046510219574, "learning_rate": 1.512910294547471e-06, "loss": 0.6077, "step": 1249 }, { "epoch": 1.0408588701271628, "grad_norm": 0.11813755333423615, "learning_rate": 1.5121388907224936e-06, "loss": 0.6649, "step": 1250 }, { "epoch": 1.041692724619554, "grad_norm": 0.11095897108316422, "learning_rate": 1.511367073568201e-06, "loss": 0.5702, "step": 1251 }, { "epoch": 1.042526579111945, "grad_norm": 0.10681331902742386, "learning_rate": 1.5105948437075008e-06, "loss": 0.5554, "step": 1252 }, { "epoch": 1.043360433604336, "grad_norm": 0.1169658675789833, "learning_rate": 1.5098222017636316e-06, "loss": 0.614, "step": 1253 }, { "epoch": 1.0441942880967272, "grad_norm": 0.11909131705760956, "learning_rate": 1.509049148360166e-06, "loss": 0.6348, "step": 1254 }, { "epoch": 1.0450281425891181, "grad_norm": 0.11616687476634979, "learning_rate": 1.5082756841210085e-06, "loss": 0.6457, "step": 1255 }, { "epoch": 1.0458619970815093, "grad_norm": 0.10854938626289368, "learning_rate": 1.5075018096703945e-06, "loss": 0.6224, "step": 1256 }, { "epoch": 1.0466958515739004, "grad_norm": 0.1108667403459549, "learning_rate": 1.506727525632891e-06, "loss": 0.6122, "step": 1257 }, { "epoch": 1.0475297060662914, "grad_norm": 0.11282949894666672, "learning_rate": 1.5059528326333952e-06, "loss": 0.6331, "step": 1258 }, { "epoch": 1.0483635605586825, "grad_norm": 0.10732194781303406, "learning_rate": 1.5051777312971354e-06, "loss": 0.6162, "step": 1259 }, { "epoch": 1.0491974150510737, "grad_norm": 0.11677113175392151, "learning_rate": 1.5044022222496682e-06, "loss": 0.6325, "step": 1260 }, { "epoch": 1.0500312695434646, "grad_norm": 0.12134989351034164, "learning_rate": 1.5036263061168796e-06, "loss": 0.6091, "step": 1261 }, { "epoch": 1.0508651240358557, "grad_norm": 0.11596612632274628, "learning_rate": 1.5028499835249842e-06, "loss": 0.6314, "step": 1262 }, { "epoch": 1.051698978528247, "grad_norm": 0.11390900611877441, "learning_rate": 1.5020732551005249e-06, "loss": 0.627, "step": 1263 }, { "epoch": 1.0525328330206378, "grad_norm": 0.11059928685426712, "learning_rate": 1.501296121470372e-06, "loss": 0.6124, "step": 1264 }, { "epoch": 1.053366687513029, "grad_norm": 0.1167077049612999, "learning_rate": 1.5005185832617228e-06, "loss": 0.6523, "step": 1265 }, { "epoch": 1.0542005420054201, "grad_norm": 0.11124616861343384, "learning_rate": 1.4997406411021007e-06, "loss": 0.5474, "step": 1266 }, { "epoch": 1.055034396497811, "grad_norm": 0.11309702694416046, "learning_rate": 1.4989622956193559e-06, "loss": 0.5877, "step": 1267 }, { "epoch": 1.0558682509902022, "grad_norm": 0.11402348428964615, "learning_rate": 1.4981835474416638e-06, "loss": 0.6443, "step": 1268 }, { "epoch": 1.0567021054825934, "grad_norm": 0.12587358057498932, "learning_rate": 1.497404397197524e-06, "loss": 0.6602, "step": 1269 }, { "epoch": 1.0575359599749843, "grad_norm": 0.10854615271091461, "learning_rate": 1.496624845515762e-06, "loss": 0.6295, "step": 1270 }, { "epoch": 1.0583698144673754, "grad_norm": 0.11264639347791672, "learning_rate": 1.4958448930255262e-06, "loss": 0.616, "step": 1271 }, { "epoch": 1.0592036689597666, "grad_norm": 0.11427608877420425, "learning_rate": 1.4950645403562896e-06, "loss": 0.5909, "step": 1272 }, { "epoch": 1.0600375234521575, "grad_norm": 0.11705349385738373, "learning_rate": 1.4942837881378462e-06, "loss": 0.6587, "step": 1273 }, { "epoch": 1.0608713779445487, "grad_norm": 0.11993349343538284, "learning_rate": 1.4935026370003148e-06, "loss": 0.607, "step": 1274 }, { "epoch": 1.0617052324369398, "grad_norm": 0.13291329145431519, "learning_rate": 1.4927210875741345e-06, "loss": 0.716, "step": 1275 }, { "epoch": 1.0625390869293307, "grad_norm": 0.11170666664838791, "learning_rate": 1.4919391404900665e-06, "loss": 0.6195, "step": 1276 }, { "epoch": 1.063372941421722, "grad_norm": 0.10838773101568222, "learning_rate": 1.4911567963791927e-06, "loss": 0.6319, "step": 1277 }, { "epoch": 1.064206795914113, "grad_norm": 0.10962139070034027, "learning_rate": 1.490374055872916e-06, "loss": 0.6101, "step": 1278 }, { "epoch": 1.065040650406504, "grad_norm": 0.11577238142490387, "learning_rate": 1.4895909196029584e-06, "loss": 0.6285, "step": 1279 }, { "epoch": 1.0658745048988951, "grad_norm": 0.11450095474720001, "learning_rate": 1.4888073882013615e-06, "loss": 0.63, "step": 1280 }, { "epoch": 1.0667083593912863, "grad_norm": 0.11657418310642242, "learning_rate": 1.4880234623004866e-06, "loss": 0.638, "step": 1281 }, { "epoch": 1.0675422138836772, "grad_norm": 0.11416659504175186, "learning_rate": 1.4872391425330122e-06, "loss": 0.6171, "step": 1282 }, { "epoch": 1.0683760683760684, "grad_norm": 0.11568942666053772, "learning_rate": 1.4864544295319355e-06, "loss": 0.617, "step": 1283 }, { "epoch": 1.0692099228684595, "grad_norm": 0.10842934995889664, "learning_rate": 1.485669323930571e-06, "loss": 0.5741, "step": 1284 }, { "epoch": 1.0700437773608504, "grad_norm": 0.11735663563013077, "learning_rate": 1.4848838263625493e-06, "loss": 0.6143, "step": 1285 }, { "epoch": 1.0708776318532416, "grad_norm": 0.11187799274921417, "learning_rate": 1.4840979374618186e-06, "loss": 0.5643, "step": 1286 }, { "epoch": 1.0717114863456327, "grad_norm": 0.11164343357086182, "learning_rate": 1.4833116578626415e-06, "loss": 0.596, "step": 1287 }, { "epoch": 1.0725453408380237, "grad_norm": 0.11518281698226929, "learning_rate": 1.4825249881995972e-06, "loss": 0.5958, "step": 1288 }, { "epoch": 1.0733791953304148, "grad_norm": 0.12009905278682709, "learning_rate": 1.481737929107579e-06, "loss": 0.6025, "step": 1289 }, { "epoch": 1.074213049822806, "grad_norm": 0.11446750909090042, "learning_rate": 1.4809504812217949e-06, "loss": 0.6771, "step": 1290 }, { "epoch": 1.075046904315197, "grad_norm": 0.11930940300226212, "learning_rate": 1.4801626451777655e-06, "loss": 0.5431, "step": 1291 }, { "epoch": 1.075880758807588, "grad_norm": 0.11522958427667618, "learning_rate": 1.4793744216113266e-06, "loss": 0.5958, "step": 1292 }, { "epoch": 1.0767146132999792, "grad_norm": 0.10718630254268646, "learning_rate": 1.4785858111586256e-06, "loss": 0.6057, "step": 1293 }, { "epoch": 1.0775484677923703, "grad_norm": 0.10954626649618149, "learning_rate": 1.477796814456122e-06, "loss": 0.5653, "step": 1294 }, { "epoch": 1.0783823222847613, "grad_norm": 0.11433891206979752, "learning_rate": 1.4770074321405876e-06, "loss": 0.6031, "step": 1295 }, { "epoch": 1.0792161767771524, "grad_norm": 0.11025510728359222, "learning_rate": 1.4762176648491049e-06, "loss": 0.6165, "step": 1296 }, { "epoch": 1.0800500312695434, "grad_norm": 0.11881809681653976, "learning_rate": 1.4754275132190678e-06, "loss": 0.6047, "step": 1297 }, { "epoch": 1.0808838857619345, "grad_norm": 0.12111616879701614, "learning_rate": 1.474636977888179e-06, "loss": 0.585, "step": 1298 }, { "epoch": 1.0817177402543257, "grad_norm": 0.1101752445101738, "learning_rate": 1.4738460594944528e-06, "loss": 0.6288, "step": 1299 }, { "epoch": 1.0825515947467168, "grad_norm": 0.11202886700630188, "learning_rate": 1.4730547586762113e-06, "loss": 0.6521, "step": 1300 }, { "epoch": 1.0833854492391077, "grad_norm": 0.1188650131225586, "learning_rate": 1.4722630760720856e-06, "loss": 0.5981, "step": 1301 }, { "epoch": 1.0842193037314989, "grad_norm": 0.11928634345531464, "learning_rate": 1.4714710123210146e-06, "loss": 0.6131, "step": 1302 }, { "epoch": 1.08505315822389, "grad_norm": 0.11217010766267776, "learning_rate": 1.470678568062246e-06, "loss": 0.6369, "step": 1303 }, { "epoch": 1.085887012716281, "grad_norm": 0.11031745374202728, "learning_rate": 1.469885743935333e-06, "loss": 0.5831, "step": 1304 }, { "epoch": 1.0867208672086721, "grad_norm": 0.11733289062976837, "learning_rate": 1.469092540580136e-06, "loss": 0.6509, "step": 1305 }, { "epoch": 1.0875547217010633, "grad_norm": 0.1268366575241089, "learning_rate": 1.4682989586368218e-06, "loss": 0.6563, "step": 1306 }, { "epoch": 1.0883885761934542, "grad_norm": 0.11494706571102142, "learning_rate": 1.4675049987458628e-06, "loss": 0.6659, "step": 1307 }, { "epoch": 1.0892224306858453, "grad_norm": 0.11928484588861465, "learning_rate": 1.4667106615480361e-06, "loss": 0.63, "step": 1308 }, { "epoch": 1.0900562851782365, "grad_norm": 0.11468034982681274, "learning_rate": 1.465915947684423e-06, "loss": 0.6123, "step": 1309 }, { "epoch": 1.0908901396706274, "grad_norm": 0.118593730032444, "learning_rate": 1.4651208577964095e-06, "loss": 0.6211, "step": 1310 }, { "epoch": 1.0917239941630186, "grad_norm": 0.11647184193134308, "learning_rate": 1.4643253925256844e-06, "loss": 0.5772, "step": 1311 }, { "epoch": 1.0925578486554097, "grad_norm": 0.11724021285772324, "learning_rate": 1.4635295525142404e-06, "loss": 0.6618, "step": 1312 }, { "epoch": 1.0933917031478007, "grad_norm": 0.10851619392633438, "learning_rate": 1.4627333384043712e-06, "loss": 0.6384, "step": 1313 }, { "epoch": 1.0942255576401918, "grad_norm": 0.11257562786340714, "learning_rate": 1.4619367508386734e-06, "loss": 0.6153, "step": 1314 }, { "epoch": 1.095059412132583, "grad_norm": 0.10816484689712524, "learning_rate": 1.4611397904600456e-06, "loss": 0.5628, "step": 1315 }, { "epoch": 1.0958932666249739, "grad_norm": 0.11537015438079834, "learning_rate": 1.4603424579116859e-06, "loss": 0.5909, "step": 1316 }, { "epoch": 1.096727121117365, "grad_norm": 0.11379721760749817, "learning_rate": 1.4595447538370932e-06, "loss": 0.6503, "step": 1317 }, { "epoch": 1.0975609756097562, "grad_norm": 0.1122441440820694, "learning_rate": 1.4587466788800673e-06, "loss": 0.6081, "step": 1318 }, { "epoch": 1.0983948301021471, "grad_norm": 0.10928847640752792, "learning_rate": 1.4579482336847059e-06, "loss": 0.6015, "step": 1319 }, { "epoch": 1.0992286845945383, "grad_norm": 0.1118772029876709, "learning_rate": 1.4571494188954056e-06, "loss": 0.6146, "step": 1320 }, { "epoch": 1.1000625390869294, "grad_norm": 0.12636762857437134, "learning_rate": 1.4563502351568624e-06, "loss": 0.6321, "step": 1321 }, { "epoch": 1.1008963935793203, "grad_norm": 0.11796654760837555, "learning_rate": 1.4555506831140695e-06, "loss": 0.6043, "step": 1322 }, { "epoch": 1.1017302480717115, "grad_norm": 0.10846695303916931, "learning_rate": 1.4547507634123175e-06, "loss": 0.5693, "step": 1323 }, { "epoch": 1.1025641025641026, "grad_norm": 0.11494018882513046, "learning_rate": 1.4539504766971927e-06, "loss": 0.6297, "step": 1324 }, { "epoch": 1.1033979570564936, "grad_norm": 0.11545376479625702, "learning_rate": 1.4531498236145786e-06, "loss": 0.6224, "step": 1325 }, { "epoch": 1.1042318115488847, "grad_norm": 0.118451327085495, "learning_rate": 1.4523488048106548e-06, "loss": 0.6033, "step": 1326 }, { "epoch": 1.1050656660412759, "grad_norm": 0.12059949338436127, "learning_rate": 1.4515474209318947e-06, "loss": 0.621, "step": 1327 }, { "epoch": 1.1058995205336668, "grad_norm": 0.11343257129192352, "learning_rate": 1.4507456726250678e-06, "loss": 0.6262, "step": 1328 }, { "epoch": 1.106733375026058, "grad_norm": 0.11061260104179382, "learning_rate": 1.4499435605372365e-06, "loss": 0.5978, "step": 1329 }, { "epoch": 1.107567229518449, "grad_norm": 0.1068677082657814, "learning_rate": 1.4491410853157575e-06, "loss": 0.5494, "step": 1330 }, { "epoch": 1.10840108401084, "grad_norm": 0.12015094608068466, "learning_rate": 1.44833824760828e-06, "loss": 0.6331, "step": 1331 }, { "epoch": 1.1092349385032312, "grad_norm": 0.12817353010177612, "learning_rate": 1.447535048062747e-06, "loss": 0.6544, "step": 1332 }, { "epoch": 1.1100687929956223, "grad_norm": 0.10849379003047943, "learning_rate": 1.4467314873273916e-06, "loss": 0.5527, "step": 1333 }, { "epoch": 1.1109026474880133, "grad_norm": 0.11941130459308624, "learning_rate": 1.4459275660507403e-06, "loss": 0.595, "step": 1334 }, { "epoch": 1.1117365019804044, "grad_norm": 0.11218126118183136, "learning_rate": 1.4451232848816087e-06, "loss": 0.6288, "step": 1335 }, { "epoch": 1.1125703564727956, "grad_norm": 0.11410936713218689, "learning_rate": 1.444318644469105e-06, "loss": 0.607, "step": 1336 }, { "epoch": 1.1134042109651865, "grad_norm": 0.11492687463760376, "learning_rate": 1.4435136454626262e-06, "loss": 0.6503, "step": 1337 }, { "epoch": 1.1142380654575776, "grad_norm": 0.11537691950798035, "learning_rate": 1.4427082885118575e-06, "loss": 0.6566, "step": 1338 }, { "epoch": 1.1150719199499688, "grad_norm": 0.1288568675518036, "learning_rate": 1.4419025742667757e-06, "loss": 0.6144, "step": 1339 }, { "epoch": 1.1159057744423597, "grad_norm": 0.12686489522457123, "learning_rate": 1.4410965033776437e-06, "loss": 0.6014, "step": 1340 }, { "epoch": 1.1167396289347509, "grad_norm": 0.11200912296772003, "learning_rate": 1.440290076495013e-06, "loss": 0.6042, "step": 1341 }, { "epoch": 1.117573483427142, "grad_norm": 0.11744599044322968, "learning_rate": 1.439483294269722e-06, "loss": 0.579, "step": 1342 }, { "epoch": 1.118407337919533, "grad_norm": 0.11925812810659409, "learning_rate": 1.4386761573528973e-06, "loss": 0.6255, "step": 1343 }, { "epoch": 1.119241192411924, "grad_norm": 0.12317737191915512, "learning_rate": 1.4378686663959505e-06, "loss": 0.5868, "step": 1344 }, { "epoch": 1.1200750469043153, "grad_norm": 0.11927246302366257, "learning_rate": 1.4370608220505788e-06, "loss": 0.5806, "step": 1345 }, { "epoch": 1.1209089013967062, "grad_norm": 0.10501258075237274, "learning_rate": 1.436252624968765e-06, "loss": 0.4934, "step": 1346 }, { "epoch": 1.1217427558890973, "grad_norm": 0.11004379391670227, "learning_rate": 1.4354440758027771e-06, "loss": 0.6332, "step": 1347 }, { "epoch": 1.1225766103814885, "grad_norm": 0.11404111236333847, "learning_rate": 1.4346351752051662e-06, "loss": 0.6288, "step": 1348 }, { "epoch": 1.1234104648738794, "grad_norm": 0.1158967912197113, "learning_rate": 1.4338259238287678e-06, "loss": 0.648, "step": 1349 }, { "epoch": 1.1242443193662706, "grad_norm": 0.11354929208755493, "learning_rate": 1.4330163223267002e-06, "loss": 0.6073, "step": 1350 }, { "epoch": 1.1250781738586617, "grad_norm": 0.11661235243082047, "learning_rate": 1.4322063713523644e-06, "loss": 0.6168, "step": 1351 }, { "epoch": 1.1259120283510526, "grad_norm": 0.11652939021587372, "learning_rate": 1.4313960715594435e-06, "loss": 0.6285, "step": 1352 }, { "epoch": 1.1267458828434438, "grad_norm": 0.10853532701730728, "learning_rate": 1.4305854236019017e-06, "loss": 0.5987, "step": 1353 }, { "epoch": 1.127579737335835, "grad_norm": 0.11252254992723465, "learning_rate": 1.429774428133985e-06, "loss": 0.5752, "step": 1354 }, { "epoch": 1.128413591828226, "grad_norm": 0.12270165979862213, "learning_rate": 1.4289630858102188e-06, "loss": 0.6592, "step": 1355 }, { "epoch": 1.129247446320617, "grad_norm": 0.11131463944911957, "learning_rate": 1.4281513972854094e-06, "loss": 0.623, "step": 1356 }, { "epoch": 1.1300813008130082, "grad_norm": 0.11877856403589249, "learning_rate": 1.4273393632146418e-06, "loss": 0.64, "step": 1357 }, { "epoch": 1.130915155305399, "grad_norm": 0.11009548604488373, "learning_rate": 1.4265269842532805e-06, "loss": 0.5545, "step": 1358 }, { "epoch": 1.1317490097977903, "grad_norm": 0.11712039262056351, "learning_rate": 1.425714261056968e-06, "loss": 0.5737, "step": 1359 }, { "epoch": 1.1325828642901814, "grad_norm": 0.12244511395692825, "learning_rate": 1.4249011942816243e-06, "loss": 0.6165, "step": 1360 }, { "epoch": 1.1334167187825726, "grad_norm": 0.1117769330739975, "learning_rate": 1.424087784583447e-06, "loss": 0.6471, "step": 1361 }, { "epoch": 1.1342505732749635, "grad_norm": 0.11851704865694046, "learning_rate": 1.4232740326189113e-06, "loss": 0.5935, "step": 1362 }, { "epoch": 1.1350844277673546, "grad_norm": 0.11291871964931488, "learning_rate": 1.4224599390447672e-06, "loss": 0.6288, "step": 1363 }, { "epoch": 1.1359182822597456, "grad_norm": 0.12204591929912567, "learning_rate": 1.421645504518041e-06, "loss": 0.5833, "step": 1364 }, { "epoch": 1.1367521367521367, "grad_norm": 0.1169772669672966, "learning_rate": 1.4208307296960343e-06, "loss": 0.5812, "step": 1365 }, { "epoch": 1.1375859912445279, "grad_norm": 0.11904870718717575, "learning_rate": 1.4200156152363236e-06, "loss": 0.6731, "step": 1366 }, { "epoch": 1.138419845736919, "grad_norm": 0.11016523838043213, "learning_rate": 1.4192001617967585e-06, "loss": 0.6453, "step": 1367 }, { "epoch": 1.13925370022931, "grad_norm": 0.12760315835475922, "learning_rate": 1.4183843700354635e-06, "loss": 0.6603, "step": 1368 }, { "epoch": 1.140087554721701, "grad_norm": 0.12259379029273987, "learning_rate": 1.417568240610835e-06, "loss": 0.667, "step": 1369 }, { "epoch": 1.140921409214092, "grad_norm": 0.116600401699543, "learning_rate": 1.416751774181543e-06, "loss": 0.5901, "step": 1370 }, { "epoch": 1.1417552637064832, "grad_norm": 0.11861781775951385, "learning_rate": 1.415934971406528e-06, "loss": 0.5813, "step": 1371 }, { "epoch": 1.1425891181988743, "grad_norm": 0.1124439537525177, "learning_rate": 1.415117832945003e-06, "loss": 0.5673, "step": 1372 }, { "epoch": 1.1434229726912655, "grad_norm": 0.11245383322238922, "learning_rate": 1.4143003594564526e-06, "loss": 0.5528, "step": 1373 }, { "epoch": 1.1442568271836564, "grad_norm": 0.12306279689073563, "learning_rate": 1.4134825516006307e-06, "loss": 0.6384, "step": 1374 }, { "epoch": 1.1450906816760476, "grad_norm": 0.11250312626361847, "learning_rate": 1.4126644100375603e-06, "loss": 0.6299, "step": 1375 }, { "epoch": 1.1459245361684387, "grad_norm": 0.12479723989963531, "learning_rate": 1.411845935427536e-06, "loss": 0.6302, "step": 1376 }, { "epoch": 1.1467583906608296, "grad_norm": 0.11576981842517853, "learning_rate": 1.4110271284311189e-06, "loss": 0.6142, "step": 1377 }, { "epoch": 1.1475922451532208, "grad_norm": 0.1186005100607872, "learning_rate": 1.41020798970914e-06, "loss": 0.6573, "step": 1378 }, { "epoch": 1.148426099645612, "grad_norm": 0.12041036039590836, "learning_rate": 1.4093885199226969e-06, "loss": 0.5843, "step": 1379 }, { "epoch": 1.1492599541380029, "grad_norm": 0.11400541663169861, "learning_rate": 1.4085687197331558e-06, "loss": 0.6108, "step": 1380 }, { "epoch": 1.150093808630394, "grad_norm": 0.11786925792694092, "learning_rate": 1.4077485898021477e-06, "loss": 0.5824, "step": 1381 }, { "epoch": 1.1509276631227852, "grad_norm": 0.12048111855983734, "learning_rate": 1.406928130791571e-06, "loss": 0.6496, "step": 1382 }, { "epoch": 1.151761517615176, "grad_norm": 0.11885000020265579, "learning_rate": 1.4061073433635898e-06, "loss": 0.6486, "step": 1383 }, { "epoch": 1.1525953721075672, "grad_norm": 0.11726827174425125, "learning_rate": 1.4052862281806318e-06, "loss": 0.6299, "step": 1384 }, { "epoch": 1.1534292265999584, "grad_norm": 0.11078763753175735, "learning_rate": 1.4044647859053913e-06, "loss": 0.608, "step": 1385 }, { "epoch": 1.1542630810923493, "grad_norm": 0.11806345731019974, "learning_rate": 1.4036430172008246e-06, "loss": 0.5859, "step": 1386 }, { "epoch": 1.1550969355847405, "grad_norm": 0.11999361962080002, "learning_rate": 1.4028209227301532e-06, "loss": 0.6445, "step": 1387 }, { "epoch": 1.1559307900771316, "grad_norm": 0.1165865957736969, "learning_rate": 1.4019985031568604e-06, "loss": 0.5671, "step": 1388 }, { "epoch": 1.1567646445695225, "grad_norm": 0.12415032833814621, "learning_rate": 1.4011757591446915e-06, "loss": 0.6278, "step": 1389 }, { "epoch": 1.1575984990619137, "grad_norm": 0.11192337423563004, "learning_rate": 1.4003526913576554e-06, "loss": 0.5964, "step": 1390 }, { "epoch": 1.1584323535543049, "grad_norm": 0.11573266237974167, "learning_rate": 1.3995293004600203e-06, "loss": 0.6284, "step": 1391 }, { "epoch": 1.1592662080466958, "grad_norm": 0.10940641909837723, "learning_rate": 1.3987055871163168e-06, "loss": 0.6426, "step": 1392 }, { "epoch": 1.160100062539087, "grad_norm": 0.11915182322263718, "learning_rate": 1.3978815519913342e-06, "loss": 0.57, "step": 1393 }, { "epoch": 1.160933917031478, "grad_norm": 0.12485363334417343, "learning_rate": 1.3970571957501226e-06, "loss": 0.5625, "step": 1394 }, { "epoch": 1.161767771523869, "grad_norm": 0.11141739040613174, "learning_rate": 1.3962325190579919e-06, "loss": 0.5876, "step": 1395 }, { "epoch": 1.1626016260162602, "grad_norm": 0.11316776275634766, "learning_rate": 1.3954075225805085e-06, "loss": 0.5888, "step": 1396 }, { "epoch": 1.1634354805086513, "grad_norm": 0.11194372922182083, "learning_rate": 1.3945822069834982e-06, "loss": 0.5907, "step": 1397 }, { "epoch": 1.1642693350010422, "grad_norm": 0.11734558641910553, "learning_rate": 1.3937565729330452e-06, "loss": 0.6379, "step": 1398 }, { "epoch": 1.1651031894934334, "grad_norm": 0.11923021823167801, "learning_rate": 1.3929306210954888e-06, "loss": 0.624, "step": 1399 }, { "epoch": 1.1659370439858245, "grad_norm": 0.11734186112880707, "learning_rate": 1.392104352137426e-06, "loss": 0.5796, "step": 1400 }, { "epoch": 1.1667708984782155, "grad_norm": 0.11634799838066101, "learning_rate": 1.3912777667257092e-06, "loss": 0.5915, "step": 1401 }, { "epoch": 1.1676047529706066, "grad_norm": 0.11459887027740479, "learning_rate": 1.3904508655274472e-06, "loss": 0.523, "step": 1402 }, { "epoch": 1.1684386074629978, "grad_norm": 0.12342048436403275, "learning_rate": 1.3896236492100025e-06, "loss": 0.5816, "step": 1403 }, { "epoch": 1.1692724619553887, "grad_norm": 0.12243616580963135, "learning_rate": 1.388796118440992e-06, "loss": 0.6229, "step": 1404 }, { "epoch": 1.1701063164477798, "grad_norm": 0.11432825773954391, "learning_rate": 1.3879682738882871e-06, "loss": 0.5549, "step": 1405 }, { "epoch": 1.170940170940171, "grad_norm": 0.12253974378108978, "learning_rate": 1.3871401162200122e-06, "loss": 0.6121, "step": 1406 }, { "epoch": 1.171774025432562, "grad_norm": 0.12205984443426132, "learning_rate": 1.3863116461045438e-06, "loss": 0.6503, "step": 1407 }, { "epoch": 1.172607879924953, "grad_norm": 0.12476064264774323, "learning_rate": 1.3854828642105116e-06, "loss": 0.6309, "step": 1408 }, { "epoch": 1.1734417344173442, "grad_norm": 0.12862101197242737, "learning_rate": 1.384653771206796e-06, "loss": 0.6598, "step": 1409 }, { "epoch": 1.1742755889097352, "grad_norm": 0.11716258525848389, "learning_rate": 1.3838243677625289e-06, "loss": 0.619, "step": 1410 }, { "epoch": 1.1751094434021263, "grad_norm": 0.12143337726593018, "learning_rate": 1.3829946545470927e-06, "loss": 0.6529, "step": 1411 }, { "epoch": 1.1759432978945175, "grad_norm": 0.1213587149977684, "learning_rate": 1.38216463223012e-06, "loss": 0.642, "step": 1412 }, { "epoch": 1.1767771523869084, "grad_norm": 0.11838190257549286, "learning_rate": 1.3813343014814925e-06, "loss": 0.6738, "step": 1413 }, { "epoch": 1.1776110068792995, "grad_norm": 0.12712059915065765, "learning_rate": 1.3805036629713406e-06, "loss": 0.6289, "step": 1414 }, { "epoch": 1.1784448613716907, "grad_norm": 0.12272265553474426, "learning_rate": 1.3796727173700443e-06, "loss": 0.6046, "step": 1415 }, { "epoch": 1.1792787158640818, "grad_norm": 0.11514297872781754, "learning_rate": 1.3788414653482296e-06, "loss": 0.5965, "step": 1416 }, { "epoch": 1.1801125703564728, "grad_norm": 0.12112387269735336, "learning_rate": 1.3780099075767719e-06, "loss": 0.6412, "step": 1417 }, { "epoch": 1.180946424848864, "grad_norm": 0.12124162167310715, "learning_rate": 1.3771780447267916e-06, "loss": 0.5654, "step": 1418 }, { "epoch": 1.1817802793412548, "grad_norm": 0.12093062698841095, "learning_rate": 1.3763458774696562e-06, "loss": 0.6209, "step": 1419 }, { "epoch": 1.182614133833646, "grad_norm": 0.12662725150585175, "learning_rate": 1.3755134064769786e-06, "loss": 0.6165, "step": 1420 }, { "epoch": 1.1834479883260371, "grad_norm": 0.12158691138029099, "learning_rate": 1.3746806324206173e-06, "loss": 0.5725, "step": 1421 }, { "epoch": 1.1842818428184283, "grad_norm": 0.11839036643505096, "learning_rate": 1.373847555972674e-06, "loss": 0.627, "step": 1422 }, { "epoch": 1.1851156973108192, "grad_norm": 0.1114383339881897, "learning_rate": 1.3730141778054962e-06, "loss": 0.6054, "step": 1423 }, { "epoch": 1.1859495518032104, "grad_norm": 0.11537707597017288, "learning_rate": 1.3721804985916743e-06, "loss": 0.625, "step": 1424 }, { "epoch": 1.1867834062956013, "grad_norm": 0.12441827356815338, "learning_rate": 1.3713465190040414e-06, "loss": 0.6006, "step": 1425 }, { "epoch": 1.1876172607879925, "grad_norm": 0.11192117631435394, "learning_rate": 1.3705122397156726e-06, "loss": 0.6492, "step": 1426 }, { "epoch": 1.1884511152803836, "grad_norm": 0.12034382671117783, "learning_rate": 1.369677661399886e-06, "loss": 0.6147, "step": 1427 }, { "epoch": 1.1892849697727748, "grad_norm": 0.12263434380292892, "learning_rate": 1.3688427847302403e-06, "loss": 0.6354, "step": 1428 }, { "epoch": 1.1901188242651657, "grad_norm": 0.12424664944410324, "learning_rate": 1.368007610380535e-06, "loss": 0.593, "step": 1429 }, { "epoch": 1.1909526787575568, "grad_norm": 0.11351408064365387, "learning_rate": 1.3671721390248099e-06, "loss": 0.5836, "step": 1430 }, { "epoch": 1.1917865332499478, "grad_norm": 0.11502416431903839, "learning_rate": 1.3663363713373453e-06, "loss": 0.5834, "step": 1431 }, { "epoch": 1.192620387742339, "grad_norm": 0.1204412430524826, "learning_rate": 1.3655003079926593e-06, "loss": 0.5815, "step": 1432 }, { "epoch": 1.19345424223473, "grad_norm": 0.12347093969583511, "learning_rate": 1.364663949665509e-06, "loss": 0.6212, "step": 1433 }, { "epoch": 1.1942880967271212, "grad_norm": 0.12269889563322067, "learning_rate": 1.3638272970308906e-06, "loss": 0.6115, "step": 1434 }, { "epoch": 1.1951219512195121, "grad_norm": 0.12062874436378479, "learning_rate": 1.3629903507640367e-06, "loss": 0.6147, "step": 1435 }, { "epoch": 1.1959558057119033, "grad_norm": 0.12827728688716888, "learning_rate": 1.362153111540417e-06, "loss": 0.6462, "step": 1436 }, { "epoch": 1.1967896602042944, "grad_norm": 0.12133876234292984, "learning_rate": 1.3613155800357384e-06, "loss": 0.6544, "step": 1437 }, { "epoch": 1.1976235146966854, "grad_norm": 0.12726116180419922, "learning_rate": 1.3604777569259427e-06, "loss": 0.6589, "step": 1438 }, { "epoch": 1.1984573691890765, "grad_norm": 0.12063872814178467, "learning_rate": 1.3596396428872078e-06, "loss": 0.6368, "step": 1439 }, { "epoch": 1.1992912236814677, "grad_norm": 0.11616161465644836, "learning_rate": 1.3588012385959462e-06, "loss": 0.5165, "step": 1440 }, { "epoch": 1.2001250781738586, "grad_norm": 0.1210310310125351, "learning_rate": 1.3579625447288043e-06, "loss": 0.5979, "step": 1441 }, { "epoch": 1.2009589326662498, "grad_norm": 0.11908716708421707, "learning_rate": 1.3571235619626624e-06, "loss": 0.647, "step": 1442 }, { "epoch": 1.201792787158641, "grad_norm": 0.11108589917421341, "learning_rate": 1.3562842909746342e-06, "loss": 0.6023, "step": 1443 }, { "epoch": 1.2026266416510318, "grad_norm": 0.12645037472248077, "learning_rate": 1.3554447324420659e-06, "loss": 0.6239, "step": 1444 }, { "epoch": 1.203460496143423, "grad_norm": 0.11542391031980515, "learning_rate": 1.3546048870425354e-06, "loss": 0.601, "step": 1445 }, { "epoch": 1.2042943506358141, "grad_norm": 0.12430214136838913, "learning_rate": 1.3537647554538532e-06, "loss": 0.6483, "step": 1446 }, { "epoch": 1.205128205128205, "grad_norm": 0.11422111093997955, "learning_rate": 1.352924338354059e-06, "loss": 0.5866, "step": 1447 }, { "epoch": 1.2059620596205962, "grad_norm": 0.11492970585823059, "learning_rate": 1.3520836364214246e-06, "loss": 0.6163, "step": 1448 }, { "epoch": 1.2067959141129874, "grad_norm": 0.11202309280633926, "learning_rate": 1.3512426503344507e-06, "loss": 0.6274, "step": 1449 }, { "epoch": 1.2076297686053783, "grad_norm": 0.11905910074710846, "learning_rate": 1.350401380771868e-06, "loss": 0.6036, "step": 1450 }, { "epoch": 1.2084636230977694, "grad_norm": 0.11703864485025406, "learning_rate": 1.3495598284126348e-06, "loss": 0.6599, "step": 1451 }, { "epoch": 1.2092974775901606, "grad_norm": 0.12923474609851837, "learning_rate": 1.3487179939359393e-06, "loss": 0.6587, "step": 1452 }, { "epoch": 1.2101313320825515, "grad_norm": 0.12412197142839432, "learning_rate": 1.3478758780211964e-06, "loss": 0.6702, "step": 1453 }, { "epoch": 1.2109651865749427, "grad_norm": 0.12546297907829285, "learning_rate": 1.3470334813480482e-06, "loss": 0.6502, "step": 1454 }, { "epoch": 1.2117990410673338, "grad_norm": 0.11630617082118988, "learning_rate": 1.3461908045963633e-06, "loss": 0.6105, "step": 1455 }, { "epoch": 1.2126328955597248, "grad_norm": 0.11860398203134537, "learning_rate": 1.3453478484462366e-06, "loss": 0.6031, "step": 1456 }, { "epoch": 1.213466750052116, "grad_norm": 0.12004498392343521, "learning_rate": 1.3445046135779885e-06, "loss": 0.5755, "step": 1457 }, { "epoch": 1.214300604544507, "grad_norm": 0.11858487129211426, "learning_rate": 1.3436611006721641e-06, "loss": 0.6452, "step": 1458 }, { "epoch": 1.215134459036898, "grad_norm": 0.12253355234861374, "learning_rate": 1.342817310409533e-06, "loss": 0.5958, "step": 1459 }, { "epoch": 1.2159683135292891, "grad_norm": 0.1240239366889, "learning_rate": 1.3419732434710893e-06, "loss": 0.6035, "step": 1460 }, { "epoch": 1.2168021680216803, "grad_norm": 0.10970788449048996, "learning_rate": 1.3411289005380493e-06, "loss": 0.5539, "step": 1461 }, { "epoch": 1.2176360225140712, "grad_norm": 0.12056974321603775, "learning_rate": 1.340284282291852e-06, "loss": 0.6337, "step": 1462 }, { "epoch": 1.2184698770064624, "grad_norm": 0.12024622410535812, "learning_rate": 1.3394393894141602e-06, "loss": 0.6247, "step": 1463 }, { "epoch": 1.2193037314988535, "grad_norm": 0.11444680392742157, "learning_rate": 1.3385942225868568e-06, "loss": 0.6319, "step": 1464 }, { "epoch": 1.2201375859912444, "grad_norm": 0.1220422014594078, "learning_rate": 1.3377487824920457e-06, "loss": 0.5785, "step": 1465 }, { "epoch": 1.2209714404836356, "grad_norm": 0.11991899460554123, "learning_rate": 1.336903069812053e-06, "loss": 0.5802, "step": 1466 }, { "epoch": 1.2218052949760267, "grad_norm": 0.12127123773097992, "learning_rate": 1.3360570852294226e-06, "loss": 0.6411, "step": 1467 }, { "epoch": 1.2226391494684177, "grad_norm": 0.11670736223459244, "learning_rate": 1.33521082942692e-06, "loss": 0.6309, "step": 1468 }, { "epoch": 1.2234730039608088, "grad_norm": 0.12089689821004868, "learning_rate": 1.3343643030875276e-06, "loss": 0.5576, "step": 1469 }, { "epoch": 1.2243068584532, "grad_norm": 0.10762432962656021, "learning_rate": 1.3335175068944477e-06, "loss": 0.6016, "step": 1470 }, { "epoch": 1.225140712945591, "grad_norm": 0.12298929691314697, "learning_rate": 1.3326704415310998e-06, "loss": 0.5854, "step": 1471 }, { "epoch": 1.225974567437982, "grad_norm": 0.12037690728902817, "learning_rate": 1.3318231076811203e-06, "loss": 0.59, "step": 1472 }, { "epoch": 1.2268084219303732, "grad_norm": 0.12309654802083969, "learning_rate": 1.3309755060283626e-06, "loss": 0.5617, "step": 1473 }, { "epoch": 1.2276422764227641, "grad_norm": 0.11756734549999237, "learning_rate": 1.3301276372568962e-06, "loss": 0.6191, "step": 1474 }, { "epoch": 1.2284761309151553, "grad_norm": 0.12193117290735245, "learning_rate": 1.3292795020510065e-06, "loss": 0.5838, "step": 1475 }, { "epoch": 1.2293099854075464, "grad_norm": 0.11809369921684265, "learning_rate": 1.3284311010951933e-06, "loss": 0.6189, "step": 1476 }, { "epoch": 1.2301438398999376, "grad_norm": 0.11490161716938019, "learning_rate": 1.3275824350741716e-06, "loss": 0.5843, "step": 1477 }, { "epoch": 1.2309776943923285, "grad_norm": 0.12500688433647156, "learning_rate": 1.3267335046728698e-06, "loss": 0.6587, "step": 1478 }, { "epoch": 1.2318115488847197, "grad_norm": 0.12077947705984116, "learning_rate": 1.3258843105764298e-06, "loss": 0.5794, "step": 1479 }, { "epoch": 1.2326454033771106, "grad_norm": 0.1252153217792511, "learning_rate": 1.3250348534702061e-06, "loss": 0.6054, "step": 1480 }, { "epoch": 1.2334792578695017, "grad_norm": 0.11973953247070312, "learning_rate": 1.3241851340397657e-06, "loss": 0.6504, "step": 1481 }, { "epoch": 1.234313112361893, "grad_norm": 0.12270364910364151, "learning_rate": 1.3233351529708875e-06, "loss": 0.5973, "step": 1482 }, { "epoch": 1.235146966854284, "grad_norm": 0.11802547425031662, "learning_rate": 1.322484910949562e-06, "loss": 0.5782, "step": 1483 }, { "epoch": 1.235980821346675, "grad_norm": 0.11984699219465256, "learning_rate": 1.3216344086619887e-06, "loss": 0.6318, "step": 1484 }, { "epoch": 1.2368146758390661, "grad_norm": 0.11720497161149979, "learning_rate": 1.3207836467945784e-06, "loss": 0.5948, "step": 1485 }, { "epoch": 1.237648530331457, "grad_norm": 0.11490039527416229, "learning_rate": 1.3199326260339514e-06, "loss": 0.6022, "step": 1486 }, { "epoch": 1.2384823848238482, "grad_norm": 0.11624352633953094, "learning_rate": 1.3190813470669361e-06, "loss": 0.5964, "step": 1487 }, { "epoch": 1.2393162393162394, "grad_norm": 0.11498794704675674, "learning_rate": 1.3182298105805704e-06, "loss": 0.614, "step": 1488 }, { "epoch": 1.2401500938086305, "grad_norm": 0.11788622289896011, "learning_rate": 1.3173780172620996e-06, "loss": 0.56, "step": 1489 }, { "epoch": 1.2409839483010214, "grad_norm": 0.11602766811847687, "learning_rate": 1.316525967798976e-06, "loss": 0.5926, "step": 1490 }, { "epoch": 1.2418178027934126, "grad_norm": 0.11411117762327194, "learning_rate": 1.3156736628788582e-06, "loss": 0.6047, "step": 1491 }, { "epoch": 1.2426516572858035, "grad_norm": 0.12349656969308853, "learning_rate": 1.3148211031896127e-06, "loss": 0.6254, "step": 1492 }, { "epoch": 1.2434855117781947, "grad_norm": 0.11581690609455109, "learning_rate": 1.3139682894193098e-06, "loss": 0.5767, "step": 1493 }, { "epoch": 1.2443193662705858, "grad_norm": 0.12168504297733307, "learning_rate": 1.3131152222562253e-06, "loss": 0.6076, "step": 1494 }, { "epoch": 1.245153220762977, "grad_norm": 0.12642395496368408, "learning_rate": 1.3122619023888402e-06, "loss": 0.5493, "step": 1495 }, { "epoch": 1.245987075255368, "grad_norm": 0.11577428877353668, "learning_rate": 1.3114083305058388e-06, "loss": 0.5811, "step": 1496 }, { "epoch": 1.246820929747759, "grad_norm": 0.12375449389219284, "learning_rate": 1.3105545072961092e-06, "loss": 0.6018, "step": 1497 }, { "epoch": 1.24765478424015, "grad_norm": 0.12210660427808762, "learning_rate": 1.309700433448742e-06, "loss": 0.5596, "step": 1498 }, { "epoch": 1.2484886387325411, "grad_norm": 0.11797598749399185, "learning_rate": 1.3088461096530304e-06, "loss": 0.6141, "step": 1499 }, { "epoch": 1.2493224932249323, "grad_norm": 0.11823847144842148, "learning_rate": 1.3079915365984688e-06, "loss": 0.6102, "step": 1500 }, { "epoch": 1.2501563477173234, "grad_norm": 0.12179305404424667, "learning_rate": 1.3071367149747534e-06, "loss": 0.6268, "step": 1501 }, { "epoch": 1.2509902022097144, "grad_norm": 0.12396464496850967, "learning_rate": 1.3062816454717803e-06, "loss": 0.5904, "step": 1502 }, { "epoch": 1.2518240567021055, "grad_norm": 0.12284725159406662, "learning_rate": 1.3054263287796462e-06, "loss": 0.63, "step": 1503 }, { "epoch": 1.2526579111944964, "grad_norm": 0.11863099783658981, "learning_rate": 1.3045707655886478e-06, "loss": 0.6018, "step": 1504 }, { "epoch": 1.2534917656868876, "grad_norm": 0.11341888457536697, "learning_rate": 1.3037149565892793e-06, "loss": 0.5945, "step": 1505 }, { "epoch": 1.2543256201792787, "grad_norm": 0.11802545189857483, "learning_rate": 1.3028589024722345e-06, "loss": 0.6707, "step": 1506 }, { "epoch": 1.2551594746716699, "grad_norm": 0.12044145166873932, "learning_rate": 1.3020026039284044e-06, "loss": 0.6184, "step": 1507 }, { "epoch": 1.2559933291640608, "grad_norm": 0.1108793094754219, "learning_rate": 1.301146061648878e-06, "loss": 0.607, "step": 1508 }, { "epoch": 1.256827183656452, "grad_norm": 0.11588302254676819, "learning_rate": 1.3002892763249395e-06, "loss": 0.6171, "step": 1509 }, { "epoch": 1.257661038148843, "grad_norm": 0.1190834641456604, "learning_rate": 1.2994322486480714e-06, "loss": 0.7073, "step": 1510 }, { "epoch": 1.258494892641234, "grad_norm": 0.11875880509614944, "learning_rate": 1.2985749793099498e-06, "loss": 0.5856, "step": 1511 }, { "epoch": 1.2593287471336252, "grad_norm": 0.1210557222366333, "learning_rate": 1.2977174690024473e-06, "loss": 0.599, "step": 1512 }, { "epoch": 1.2601626016260163, "grad_norm": 0.11437001079320908, "learning_rate": 1.2968597184176298e-06, "loss": 0.5626, "step": 1513 }, { "epoch": 1.2609964561184073, "grad_norm": 0.11603326350450516, "learning_rate": 1.296001728247758e-06, "loss": 0.6017, "step": 1514 }, { "epoch": 1.2618303106107984, "grad_norm": 0.10780469328165054, "learning_rate": 1.2951434991852855e-06, "loss": 0.524, "step": 1515 }, { "epoch": 1.2626641651031896, "grad_norm": 0.11808061599731445, "learning_rate": 1.2942850319228585e-06, "loss": 0.5907, "step": 1516 }, { "epoch": 1.2634980195955805, "grad_norm": 0.12312963604927063, "learning_rate": 1.2934263271533169e-06, "loss": 0.6118, "step": 1517 }, { "epoch": 1.2643318740879717, "grad_norm": 0.12425033748149872, "learning_rate": 1.2925673855696899e-06, "loss": 0.5697, "step": 1518 }, { "epoch": 1.2651657285803628, "grad_norm": 0.122237429022789, "learning_rate": 1.2917082078652e-06, "loss": 0.6109, "step": 1519 }, { "epoch": 1.2659995830727537, "grad_norm": 0.1070125550031662, "learning_rate": 1.2908487947332586e-06, "loss": 0.6049, "step": 1520 }, { "epoch": 1.2668334375651449, "grad_norm": 0.12275154143571854, "learning_rate": 1.2899891468674686e-06, "loss": 0.6722, "step": 1521 }, { "epoch": 1.267667292057536, "grad_norm": 0.11539646238088608, "learning_rate": 1.289129264961621e-06, "loss": 0.6129, "step": 1522 }, { "epoch": 1.268501146549927, "grad_norm": 0.12436021119356155, "learning_rate": 1.2882691497096967e-06, "loss": 0.571, "step": 1523 }, { "epoch": 1.2693350010423181, "grad_norm": 0.1118353083729744, "learning_rate": 1.2874088018058644e-06, "loss": 0.5885, "step": 1524 }, { "epoch": 1.2701688555347093, "grad_norm": 0.1265314221382141, "learning_rate": 1.2865482219444803e-06, "loss": 0.6319, "step": 1525 }, { "epoch": 1.2710027100271002, "grad_norm": 0.11715365201234818, "learning_rate": 1.2856874108200887e-06, "loss": 0.592, "step": 1526 }, { "epoch": 1.2718365645194913, "grad_norm": 0.12056981027126312, "learning_rate": 1.28482636912742e-06, "loss": 0.6248, "step": 1527 }, { "epoch": 1.2726704190118825, "grad_norm": 0.1140827089548111, "learning_rate": 1.2839650975613904e-06, "loss": 0.6356, "step": 1528 }, { "epoch": 1.2735042735042734, "grad_norm": 0.12372983992099762, "learning_rate": 1.2831035968171024e-06, "loss": 0.6428, "step": 1529 }, { "epoch": 1.2743381279966646, "grad_norm": 0.11513406038284302, "learning_rate": 1.2822418675898425e-06, "loss": 0.6471, "step": 1530 }, { "epoch": 1.2751719824890557, "grad_norm": 0.12804672122001648, "learning_rate": 1.2813799105750821e-06, "loss": 0.627, "step": 1531 }, { "epoch": 1.2760058369814469, "grad_norm": 0.12233211100101471, "learning_rate": 1.280517726468477e-06, "loss": 0.6289, "step": 1532 }, { "epoch": 1.2768396914738378, "grad_norm": 0.11950721591711044, "learning_rate": 1.2796553159658651e-06, "loss": 0.6586, "step": 1533 }, { "epoch": 1.277673545966229, "grad_norm": 0.11652793735265732, "learning_rate": 1.2787926797632681e-06, "loss": 0.5945, "step": 1534 }, { "epoch": 1.2785074004586199, "grad_norm": 0.11893709748983383, "learning_rate": 1.277929818556889e-06, "loss": 0.5765, "step": 1535 }, { "epoch": 1.279341254951011, "grad_norm": 0.12274635583162308, "learning_rate": 1.277066733043113e-06, "loss": 0.6151, "step": 1536 }, { "epoch": 1.2801751094434022, "grad_norm": 0.12275136262178421, "learning_rate": 1.2762034239185062e-06, "loss": 0.6484, "step": 1537 }, { "epoch": 1.2810089639357933, "grad_norm": 0.12041904777288437, "learning_rate": 1.2753398918798149e-06, "loss": 0.5766, "step": 1538 }, { "epoch": 1.2818428184281843, "grad_norm": 0.12134743481874466, "learning_rate": 1.2744761376239655e-06, "loss": 0.5816, "step": 1539 }, { "epoch": 1.2826766729205754, "grad_norm": 0.1321423202753067, "learning_rate": 1.2736121618480637e-06, "loss": 0.6209, "step": 1540 }, { "epoch": 1.2835105274129663, "grad_norm": 0.11608442664146423, "learning_rate": 1.2727479652493942e-06, "loss": 0.5762, "step": 1541 }, { "epoch": 1.2843443819053575, "grad_norm": 0.11591994017362595, "learning_rate": 1.2718835485254195e-06, "loss": 0.6242, "step": 1542 }, { "epoch": 1.2851782363977486, "grad_norm": 0.12969833612442017, "learning_rate": 1.2710189123737802e-06, "loss": 0.6901, "step": 1543 }, { "epoch": 1.2860120908901398, "grad_norm": 0.12043966352939606, "learning_rate": 1.2701540574922937e-06, "loss": 0.6434, "step": 1544 }, { "epoch": 1.2868459453825307, "grad_norm": 0.12212436646223068, "learning_rate": 1.2692889845789537e-06, "loss": 0.5581, "step": 1545 }, { "epoch": 1.2876797998749219, "grad_norm": 0.11876165866851807, "learning_rate": 1.2684236943319306e-06, "loss": 0.5852, "step": 1546 }, { "epoch": 1.2885136543673128, "grad_norm": 0.11841654032468796, "learning_rate": 1.2675581874495695e-06, "loss": 0.5714, "step": 1547 }, { "epoch": 1.289347508859704, "grad_norm": 0.11584402620792389, "learning_rate": 1.266692464630391e-06, "loss": 0.6423, "step": 1548 }, { "epoch": 1.290181363352095, "grad_norm": 0.1268841028213501, "learning_rate": 1.265826526573089e-06, "loss": 0.6126, "step": 1549 }, { "epoch": 1.2910152178444863, "grad_norm": 0.1187991052865982, "learning_rate": 1.2649603739765323e-06, "loss": 0.6014, "step": 1550 }, { "epoch": 1.2918490723368772, "grad_norm": 0.12219411134719849, "learning_rate": 1.2640940075397616e-06, "loss": 0.6377, "step": 1551 }, { "epoch": 1.2926829268292683, "grad_norm": 0.11915986984968185, "learning_rate": 1.2632274279619917e-06, "loss": 0.589, "step": 1552 }, { "epoch": 1.2935167813216593, "grad_norm": 0.1265590786933899, "learning_rate": 1.2623606359426075e-06, "loss": 0.6224, "step": 1553 }, { "epoch": 1.2943506358140504, "grad_norm": 0.12014889717102051, "learning_rate": 1.2614936321811671e-06, "loss": 0.6103, "step": 1554 }, { "epoch": 1.2951844903064416, "grad_norm": 0.12002978473901749, "learning_rate": 1.2606264173773986e-06, "loss": 0.6196, "step": 1555 }, { "epoch": 1.2960183447988327, "grad_norm": 0.12422852218151093, "learning_rate": 1.2597589922312007e-06, "loss": 0.6181, "step": 1556 }, { "epoch": 1.2968521992912236, "grad_norm": 0.12736350297927856, "learning_rate": 1.2588913574426419e-06, "loss": 0.5203, "step": 1557 }, { "epoch": 1.2976860537836148, "grad_norm": 0.1263066977262497, "learning_rate": 1.2580235137119593e-06, "loss": 0.6256, "step": 1558 }, { "epoch": 1.2985199082760057, "grad_norm": 0.11893312633037567, "learning_rate": 1.2571554617395596e-06, "loss": 0.6223, "step": 1559 }, { "epoch": 1.2993537627683969, "grad_norm": 0.1329938918352127, "learning_rate": 1.2562872022260167e-06, "loss": 0.6485, "step": 1560 }, { "epoch": 1.300187617260788, "grad_norm": 0.1189078688621521, "learning_rate": 1.2554187358720723e-06, "loss": 0.5941, "step": 1561 }, { "epoch": 1.3010214717531792, "grad_norm": 0.12884029746055603, "learning_rate": 1.2545500633786356e-06, "loss": 0.6785, "step": 1562 }, { "epoch": 1.30185532624557, "grad_norm": 0.12023568153381348, "learning_rate": 1.2536811854467815e-06, "loss": 0.62, "step": 1563 }, { "epoch": 1.3026891807379612, "grad_norm": 0.1196177676320076, "learning_rate": 1.2528121027777505e-06, "loss": 0.5589, "step": 1564 }, { "epoch": 1.3035230352303522, "grad_norm": 0.11578637361526489, "learning_rate": 1.2519428160729491e-06, "loss": 0.5884, "step": 1565 }, { "epoch": 1.3043568897227433, "grad_norm": 0.12561044096946716, "learning_rate": 1.2510733260339478e-06, "loss": 0.5634, "step": 1566 }, { "epoch": 1.3051907442151345, "grad_norm": 0.11461557447910309, "learning_rate": 1.2502036333624815e-06, "loss": 0.5752, "step": 1567 }, { "epoch": 1.3060245987075256, "grad_norm": 0.12718208134174347, "learning_rate": 1.249333738760449e-06, "loss": 0.6245, "step": 1568 }, { "epoch": 1.3068584531999166, "grad_norm": 0.12742236256599426, "learning_rate": 1.2484636429299111e-06, "loss": 0.6428, "step": 1569 }, { "epoch": 1.3076923076923077, "grad_norm": 0.11149119585752487, "learning_rate": 1.247593346573092e-06, "loss": 0.6066, "step": 1570 }, { "epoch": 1.3085261621846986, "grad_norm": 0.1338125318288803, "learning_rate": 1.2467228503923772e-06, "loss": 0.5512, "step": 1571 }, { "epoch": 1.3093600166770898, "grad_norm": 0.1266402006149292, "learning_rate": 1.2458521550903138e-06, "loss": 0.6163, "step": 1572 }, { "epoch": 1.310193871169481, "grad_norm": 0.12411841750144958, "learning_rate": 1.2449812613696094e-06, "loss": 0.6586, "step": 1573 }, { "epoch": 1.311027725661872, "grad_norm": 0.12587663531303406, "learning_rate": 1.2441101699331311e-06, "loss": 0.6382, "step": 1574 }, { "epoch": 1.311861580154263, "grad_norm": 0.1223389133810997, "learning_rate": 1.243238881483907e-06, "loss": 0.6087, "step": 1575 }, { "epoch": 1.3126954346466542, "grad_norm": 0.12871743738651276, "learning_rate": 1.2423673967251227e-06, "loss": 0.6235, "step": 1576 }, { "epoch": 1.3135292891390453, "grad_norm": 0.12110795080661774, "learning_rate": 1.2414957163601237e-06, "loss": 0.5741, "step": 1577 }, { "epoch": 1.3143631436314362, "grad_norm": 0.11542119085788727, "learning_rate": 1.240623841092412e-06, "loss": 0.5549, "step": 1578 }, { "epoch": 1.3151969981238274, "grad_norm": 0.11896467208862305, "learning_rate": 1.2397517716256478e-06, "loss": 0.6275, "step": 1579 }, { "epoch": 1.3160308526162185, "grad_norm": 0.12287244945764542, "learning_rate": 1.2388795086636479e-06, "loss": 0.6034, "step": 1580 }, { "epoch": 1.3168647071086095, "grad_norm": 0.13043585419654846, "learning_rate": 1.238007052910385e-06, "loss": 0.6218, "step": 1581 }, { "epoch": 1.3176985616010006, "grad_norm": 0.12255072593688965, "learning_rate": 1.2371344050699871e-06, "loss": 0.556, "step": 1582 }, { "epoch": 1.3185324160933918, "grad_norm": 0.12159581482410431, "learning_rate": 1.2362615658467377e-06, "loss": 0.6142, "step": 1583 }, { "epoch": 1.3193662705857827, "grad_norm": 0.11910532414913177, "learning_rate": 1.2353885359450756e-06, "loss": 0.6148, "step": 1584 }, { "epoch": 1.3202001250781739, "grad_norm": 0.11686382442712784, "learning_rate": 1.2345153160695917e-06, "loss": 0.5947, "step": 1585 }, { "epoch": 1.321033979570565, "grad_norm": 0.12290305644273758, "learning_rate": 1.2336419069250308e-06, "loss": 0.5934, "step": 1586 }, { "epoch": 1.321867834062956, "grad_norm": 0.12619248032569885, "learning_rate": 1.2327683092162918e-06, "loss": 0.6704, "step": 1587 }, { "epoch": 1.322701688555347, "grad_norm": 0.13024446368217468, "learning_rate": 1.231894523648424e-06, "loss": 0.5824, "step": 1588 }, { "epoch": 1.3235355430477382, "grad_norm": 0.1289645880460739, "learning_rate": 1.2310205509266292e-06, "loss": 0.656, "step": 1589 }, { "epoch": 1.3243693975401292, "grad_norm": 0.1269690990447998, "learning_rate": 1.23014639175626e-06, "loss": 0.6433, "step": 1590 }, { "epoch": 1.3252032520325203, "grad_norm": 0.12446628510951996, "learning_rate": 1.2292720468428199e-06, "loss": 0.57, "step": 1591 }, { "epoch": 1.3260371065249115, "grad_norm": 0.12336724996566772, "learning_rate": 1.2283975168919617e-06, "loss": 0.6324, "step": 1592 }, { "epoch": 1.3268709610173026, "grad_norm": 0.11296447366476059, "learning_rate": 1.2275228026094879e-06, "loss": 0.6097, "step": 1593 }, { "epoch": 1.3277048155096935, "grad_norm": 0.12164086103439331, "learning_rate": 1.2266479047013498e-06, "loss": 0.5969, "step": 1594 }, { "epoch": 1.3285386700020847, "grad_norm": 0.11765522509813309, "learning_rate": 1.2257728238736467e-06, "loss": 0.5404, "step": 1595 }, { "epoch": 1.3293725244944756, "grad_norm": 0.12349137663841248, "learning_rate": 1.224897560832625e-06, "loss": 0.6574, "step": 1596 }, { "epoch": 1.3302063789868668, "grad_norm": 0.1171487495303154, "learning_rate": 1.22402211628468e-06, "loss": 0.5834, "step": 1597 }, { "epoch": 1.331040233479258, "grad_norm": 0.11913489550352097, "learning_rate": 1.2231464909363512e-06, "loss": 0.6022, "step": 1598 }, { "epoch": 1.331874087971649, "grad_norm": 0.11552874743938446, "learning_rate": 1.2222706854943255e-06, "loss": 0.6164, "step": 1599 }, { "epoch": 1.33270794246404, "grad_norm": 0.11986060440540314, "learning_rate": 1.2213947006654345e-06, "loss": 0.6062, "step": 1600 }, { "epoch": 1.33270794246404, "eval_loss": 0.6176180839538574, "eval_runtime": 321.4896, "eval_samples_per_second": 16.744, "eval_steps_per_second": 2.793, "step": 1600 }, { "epoch": 1.3335417969564312, "grad_norm": 0.12321524322032928, "learning_rate": 1.2205185371566554e-06, "loss": 0.6164, "step": 1601 }, { "epoch": 1.334375651448822, "grad_norm": 0.12141884118318558, "learning_rate": 1.2196421956751083e-06, "loss": 0.601, "step": 1602 }, { "epoch": 1.3352095059412132, "grad_norm": 0.12699675559997559, "learning_rate": 1.2187656769280577e-06, "loss": 0.5379, "step": 1603 }, { "epoch": 1.3360433604336044, "grad_norm": 0.12791088223457336, "learning_rate": 1.2178889816229116e-06, "loss": 0.6292, "step": 1604 }, { "epoch": 1.3368772149259955, "grad_norm": 0.12027748674154282, "learning_rate": 1.2170121104672196e-06, "loss": 0.564, "step": 1605 }, { "epoch": 1.3377110694183865, "grad_norm": 0.11741901189088821, "learning_rate": 1.2161350641686737e-06, "loss": 0.5772, "step": 1606 }, { "epoch": 1.3385449239107776, "grad_norm": 0.11572231352329254, "learning_rate": 1.2152578434351072e-06, "loss": 0.5819, "step": 1607 }, { "epoch": 1.3393787784031685, "grad_norm": 0.1255660504102707, "learning_rate": 1.214380448974494e-06, "loss": 0.6341, "step": 1608 }, { "epoch": 1.3402126328955597, "grad_norm": 0.11974442005157471, "learning_rate": 1.2135028814949485e-06, "loss": 0.6233, "step": 1609 }, { "epoch": 1.3410464873879508, "grad_norm": 0.12625208497047424, "learning_rate": 1.212625141704725e-06, "loss": 0.6231, "step": 1610 }, { "epoch": 1.341880341880342, "grad_norm": 0.12386664003133774, "learning_rate": 1.2117472303122157e-06, "loss": 0.617, "step": 1611 }, { "epoch": 1.342714196372733, "grad_norm": 0.13161849975585938, "learning_rate": 1.210869148025952e-06, "loss": 0.654, "step": 1612 }, { "epoch": 1.343548050865124, "grad_norm": 0.11914031207561493, "learning_rate": 1.2099908955546043e-06, "loss": 0.6282, "step": 1613 }, { "epoch": 1.344381905357515, "grad_norm": 0.12541799247264862, "learning_rate": 1.2091124736069788e-06, "loss": 0.6133, "step": 1614 }, { "epoch": 1.3452157598499062, "grad_norm": 0.12099876254796982, "learning_rate": 1.2082338828920185e-06, "loss": 0.5661, "step": 1615 }, { "epoch": 1.3460496143422973, "grad_norm": 0.12484125792980194, "learning_rate": 1.2073551241188037e-06, "loss": 0.6238, "step": 1616 }, { "epoch": 1.3468834688346885, "grad_norm": 0.12073414772748947, "learning_rate": 1.2064761979965497e-06, "loss": 0.5491, "step": 1617 }, { "epoch": 1.3477173233270794, "grad_norm": 0.12662388384342194, "learning_rate": 1.2055971052346067e-06, "loss": 0.6235, "step": 1618 }, { "epoch": 1.3485511778194705, "grad_norm": 0.12112808972597122, "learning_rate": 1.2047178465424594e-06, "loss": 0.5493, "step": 1619 }, { "epoch": 1.3493850323118615, "grad_norm": 0.13022390007972717, "learning_rate": 1.2038384226297271e-06, "loss": 0.6531, "step": 1620 }, { "epoch": 1.3502188868042526, "grad_norm": 0.12017896771430969, "learning_rate": 1.202958834206162e-06, "loss": 0.5819, "step": 1621 }, { "epoch": 1.3510527412966438, "grad_norm": 0.12317414581775665, "learning_rate": 1.2020790819816484e-06, "loss": 0.5736, "step": 1622 }, { "epoch": 1.351886595789035, "grad_norm": 0.12808437645435333, "learning_rate": 1.2011991666662043e-06, "loss": 0.6196, "step": 1623 }, { "epoch": 1.3527204502814258, "grad_norm": 0.11925622075796127, "learning_rate": 1.2003190889699776e-06, "loss": 0.5703, "step": 1624 }, { "epoch": 1.353554304773817, "grad_norm": 0.1228049024939537, "learning_rate": 1.1994388496032485e-06, "loss": 0.6357, "step": 1625 }, { "epoch": 1.354388159266208, "grad_norm": 0.12009157985448837, "learning_rate": 1.1985584492764273e-06, "loss": 0.5559, "step": 1626 }, { "epoch": 1.355222013758599, "grad_norm": 0.1287962794303894, "learning_rate": 1.1976778887000543e-06, "loss": 0.6269, "step": 1627 }, { "epoch": 1.3560558682509902, "grad_norm": 0.11851408332586288, "learning_rate": 1.1967971685847988e-06, "loss": 0.5966, "step": 1628 }, { "epoch": 1.3568897227433814, "grad_norm": 0.12402857095003128, "learning_rate": 1.1959162896414589e-06, "loss": 0.5611, "step": 1629 }, { "epoch": 1.3577235772357723, "grad_norm": 0.12092249095439911, "learning_rate": 1.1950352525809617e-06, "loss": 0.5631, "step": 1630 }, { "epoch": 1.3585574317281635, "grad_norm": 0.12502413988113403, "learning_rate": 1.1941540581143608e-06, "loss": 0.584, "step": 1631 }, { "epoch": 1.3593912862205544, "grad_norm": 0.12327523529529572, "learning_rate": 1.1932727069528373e-06, "loss": 0.6027, "step": 1632 }, { "epoch": 1.3602251407129455, "grad_norm": 0.11953040957450867, "learning_rate": 1.1923911998076988e-06, "loss": 0.6104, "step": 1633 }, { "epoch": 1.3610589952053367, "grad_norm": 0.11636977642774582, "learning_rate": 1.1915095373903788e-06, "loss": 0.6116, "step": 1634 }, { "epoch": 1.3618928496977278, "grad_norm": 0.12088527530431747, "learning_rate": 1.1906277204124363e-06, "loss": 0.6055, "step": 1635 }, { "epoch": 1.3627267041901188, "grad_norm": 0.119356170296669, "learning_rate": 1.1897457495855544e-06, "loss": 0.606, "step": 1636 }, { "epoch": 1.36356055868251, "grad_norm": 0.12436091899871826, "learning_rate": 1.1888636256215412e-06, "loss": 0.6475, "step": 1637 }, { "epoch": 1.3643944131749008, "grad_norm": 0.12362188845872879, "learning_rate": 1.187981349232328e-06, "loss": 0.5592, "step": 1638 }, { "epoch": 1.365228267667292, "grad_norm": 0.12188667058944702, "learning_rate": 1.1870989211299684e-06, "loss": 0.6392, "step": 1639 }, { "epoch": 1.3660621221596831, "grad_norm": 0.12498065829277039, "learning_rate": 1.1862163420266398e-06, "loss": 0.5989, "step": 1640 }, { "epoch": 1.3668959766520743, "grad_norm": 0.11544735729694366, "learning_rate": 1.1853336126346406e-06, "loss": 0.6508, "step": 1641 }, { "epoch": 1.3677298311444652, "grad_norm": 0.1186600849032402, "learning_rate": 1.1844507336663906e-06, "loss": 0.5887, "step": 1642 }, { "epoch": 1.3685636856368564, "grad_norm": 0.12236911803483963, "learning_rate": 1.1835677058344308e-06, "loss": 0.5881, "step": 1643 }, { "epoch": 1.3693975401292475, "grad_norm": 0.11984533816576004, "learning_rate": 1.1826845298514214e-06, "loss": 0.5811, "step": 1644 }, { "epoch": 1.3702313946216385, "grad_norm": 0.12120435386896133, "learning_rate": 1.1818012064301432e-06, "loss": 0.6307, "step": 1645 }, { "epoch": 1.3710652491140296, "grad_norm": 0.11709383875131607, "learning_rate": 1.1809177362834955e-06, "loss": 0.6239, "step": 1646 }, { "epoch": 1.3718991036064208, "grad_norm": 0.13146185874938965, "learning_rate": 1.1800341201244952e-06, "loss": 0.5956, "step": 1647 }, { "epoch": 1.3727329580988117, "grad_norm": 0.11478230357170105, "learning_rate": 1.1791503586662793e-06, "loss": 0.5437, "step": 1648 }, { "epoch": 1.3735668125912028, "grad_norm": 0.11662175506353378, "learning_rate": 1.1782664526220992e-06, "loss": 0.6086, "step": 1649 }, { "epoch": 1.374400667083594, "grad_norm": 0.11829902231693268, "learning_rate": 1.1773824027053255e-06, "loss": 0.5965, "step": 1650 }, { "epoch": 1.375234521575985, "grad_norm": 0.12727516889572144, "learning_rate": 1.1764982096294431e-06, "loss": 0.573, "step": 1651 }, { "epoch": 1.376068376068376, "grad_norm": 0.12310794740915298, "learning_rate": 1.1756138741080536e-06, "loss": 0.6462, "step": 1652 }, { "epoch": 1.3769022305607672, "grad_norm": 0.11653224378824234, "learning_rate": 1.1747293968548734e-06, "loss": 0.5782, "step": 1653 }, { "epoch": 1.3777360850531581, "grad_norm": 0.12128982692956924, "learning_rate": 1.1738447785837322e-06, "loss": 0.6108, "step": 1654 }, { "epoch": 1.3785699395455493, "grad_norm": 0.12801648676395416, "learning_rate": 1.1729600200085752e-06, "loss": 0.6026, "step": 1655 }, { "epoch": 1.3794037940379404, "grad_norm": 0.11884930729866028, "learning_rate": 1.1720751218434596e-06, "loss": 0.5554, "step": 1656 }, { "epoch": 1.3802376485303314, "grad_norm": 0.12543028593063354, "learning_rate": 1.1711900848025554e-06, "loss": 0.6485, "step": 1657 }, { "epoch": 1.3810715030227225, "grad_norm": 0.12205296754837036, "learning_rate": 1.1703049096001452e-06, "loss": 0.585, "step": 1658 }, { "epoch": 1.3819053575151137, "grad_norm": 0.12094337493181229, "learning_rate": 1.1694195969506229e-06, "loss": 0.588, "step": 1659 }, { "epoch": 1.3827392120075048, "grad_norm": 0.12128522247076035, "learning_rate": 1.1685341475684934e-06, "loss": 0.6043, "step": 1660 }, { "epoch": 1.3835730664998958, "grad_norm": 0.12067275494337082, "learning_rate": 1.1676485621683711e-06, "loss": 0.5896, "step": 1661 }, { "epoch": 1.384406920992287, "grad_norm": 0.11607006937265396, "learning_rate": 1.1667628414649815e-06, "loss": 0.5454, "step": 1662 }, { "epoch": 1.3852407754846778, "grad_norm": 0.11472892016172409, "learning_rate": 1.1658769861731584e-06, "loss": 0.5768, "step": 1663 }, { "epoch": 1.386074629977069, "grad_norm": 0.12670141458511353, "learning_rate": 1.1649909970078447e-06, "loss": 0.6426, "step": 1664 }, { "epoch": 1.3869084844694601, "grad_norm": 0.12123548239469528, "learning_rate": 1.1641048746840912e-06, "loss": 0.6508, "step": 1665 }, { "epoch": 1.3877423389618513, "grad_norm": 0.12310612946748734, "learning_rate": 1.1632186199170555e-06, "loss": 0.6414, "step": 1666 }, { "epoch": 1.3885761934542422, "grad_norm": 0.11864950507879257, "learning_rate": 1.1623322334220037e-06, "loss": 0.5058, "step": 1667 }, { "epoch": 1.3894100479466334, "grad_norm": 0.12240031361579895, "learning_rate": 1.1614457159143064e-06, "loss": 0.5974, "step": 1668 }, { "epoch": 1.3902439024390243, "grad_norm": 0.13276803493499756, "learning_rate": 1.1605590681094408e-06, "loss": 0.5648, "step": 1669 }, { "epoch": 1.3910777569314154, "grad_norm": 0.12129524350166321, "learning_rate": 1.1596722907229898e-06, "loss": 0.6506, "step": 1670 }, { "epoch": 1.3919116114238066, "grad_norm": 0.12178556621074677, "learning_rate": 1.1587853844706396e-06, "loss": 0.5844, "step": 1671 }, { "epoch": 1.3927454659161977, "grad_norm": 0.12161284685134888, "learning_rate": 1.1578983500681816e-06, "loss": 0.5732, "step": 1672 }, { "epoch": 1.3935793204085887, "grad_norm": 0.12326084077358246, "learning_rate": 1.1570111882315098e-06, "loss": 0.5727, "step": 1673 }, { "epoch": 1.3944131749009798, "grad_norm": 0.12302473187446594, "learning_rate": 1.1561238996766216e-06, "loss": 0.5913, "step": 1674 }, { "epoch": 1.3952470293933708, "grad_norm": 0.12436919659376144, "learning_rate": 1.1552364851196166e-06, "loss": 0.5877, "step": 1675 }, { "epoch": 1.396080883885762, "grad_norm": 0.12807847559452057, "learning_rate": 1.154348945276695e-06, "loss": 0.6274, "step": 1676 }, { "epoch": 1.396914738378153, "grad_norm": 0.12182601541280746, "learning_rate": 1.1534612808641602e-06, "loss": 0.6319, "step": 1677 }, { "epoch": 1.3977485928705442, "grad_norm": 0.13187600672245026, "learning_rate": 1.1525734925984143e-06, "loss": 0.6148, "step": 1678 }, { "epoch": 1.3985824473629351, "grad_norm": 0.11592843383550644, "learning_rate": 1.1516855811959602e-06, "loss": 0.5472, "step": 1679 }, { "epoch": 1.3994163018553263, "grad_norm": 0.11742573231458664, "learning_rate": 1.1507975473734e-06, "loss": 0.5666, "step": 1680 }, { "epoch": 1.4002501563477172, "grad_norm": 0.12708871066570282, "learning_rate": 1.1499093918474347e-06, "loss": 0.6051, "step": 1681 }, { "epoch": 1.4010840108401084, "grad_norm": 0.11694274097681046, "learning_rate": 1.1490211153348634e-06, "loss": 0.5253, "step": 1682 }, { "epoch": 1.4019178653324995, "grad_norm": 0.12378162145614624, "learning_rate": 1.1481327185525827e-06, "loss": 0.6387, "step": 1683 }, { "epoch": 1.4027517198248907, "grad_norm": 0.1409406065940857, "learning_rate": 1.1472442022175867e-06, "loss": 0.5597, "step": 1684 }, { "epoch": 1.4035855743172816, "grad_norm": 0.12476393580436707, "learning_rate": 1.1463555670469657e-06, "loss": 0.5394, "step": 1685 }, { "epoch": 1.4044194288096727, "grad_norm": 0.13176341354846954, "learning_rate": 1.1454668137579057e-06, "loss": 0.6395, "step": 1686 }, { "epoch": 1.4052532833020637, "grad_norm": 0.13114789128303528, "learning_rate": 1.1445779430676883e-06, "loss": 0.6215, "step": 1687 }, { "epoch": 1.4060871377944548, "grad_norm": 0.12079852819442749, "learning_rate": 1.1436889556936904e-06, "loss": 0.5878, "step": 1688 }, { "epoch": 1.406920992286846, "grad_norm": 0.1296648234128952, "learning_rate": 1.142799852353382e-06, "loss": 0.5442, "step": 1689 }, { "epoch": 1.4077548467792371, "grad_norm": 0.1254526525735855, "learning_rate": 1.141910633764327e-06, "loss": 0.6245, "step": 1690 }, { "epoch": 1.408588701271628, "grad_norm": 0.12021210789680481, "learning_rate": 1.1410213006441827e-06, "loss": 0.6083, "step": 1691 }, { "epoch": 1.4094225557640192, "grad_norm": 0.12678121030330658, "learning_rate": 1.1401318537106987e-06, "loss": 0.6064, "step": 1692 }, { "epoch": 1.4102564102564101, "grad_norm": 0.12526234984397888, "learning_rate": 1.1392422936817165e-06, "loss": 0.6203, "step": 1693 }, { "epoch": 1.4110902647488013, "grad_norm": 0.11436313390731812, "learning_rate": 1.1383526212751686e-06, "loss": 0.5449, "step": 1694 }, { "epoch": 1.4119241192411924, "grad_norm": 0.11527229845523834, "learning_rate": 1.1374628372090781e-06, "loss": 0.5821, "step": 1695 }, { "epoch": 1.4127579737335836, "grad_norm": 0.12278933078050613, "learning_rate": 1.1365729422015593e-06, "loss": 0.6249, "step": 1696 }, { "epoch": 1.4135918282259745, "grad_norm": 0.12223772704601288, "learning_rate": 1.1356829369708144e-06, "loss": 0.5978, "step": 1697 }, { "epoch": 1.4144256827183657, "grad_norm": 0.12515763938426971, "learning_rate": 1.1347928222351354e-06, "loss": 0.5633, "step": 1698 }, { "epoch": 1.4152595372107566, "grad_norm": 0.12488391250371933, "learning_rate": 1.1339025987129031e-06, "loss": 0.6073, "step": 1699 }, { "epoch": 1.4160933917031477, "grad_norm": 0.13267357647418976, "learning_rate": 1.1330122671225855e-06, "loss": 0.5895, "step": 1700 }, { "epoch": 1.416927246195539, "grad_norm": 0.121218241751194, "learning_rate": 1.1321218281827377e-06, "loss": 0.5815, "step": 1701 }, { "epoch": 1.41776110068793, "grad_norm": 0.11679316312074661, "learning_rate": 1.1312312826120019e-06, "loss": 0.6053, "step": 1702 }, { "epoch": 1.418594955180321, "grad_norm": 0.12714731693267822, "learning_rate": 1.1303406311291064e-06, "loss": 0.5996, "step": 1703 }, { "epoch": 1.4194288096727121, "grad_norm": 0.1288938969373703, "learning_rate": 1.1294498744528645e-06, "loss": 0.622, "step": 1704 }, { "epoch": 1.4202626641651033, "grad_norm": 0.11887186765670776, "learning_rate": 1.128559013302174e-06, "loss": 0.5892, "step": 1705 }, { "epoch": 1.4210965186574942, "grad_norm": 0.12163923680782318, "learning_rate": 1.1276680483960188e-06, "loss": 0.5992, "step": 1706 }, { "epoch": 1.4219303731498854, "grad_norm": 0.1262393295764923, "learning_rate": 1.1267769804534647e-06, "loss": 0.6356, "step": 1707 }, { "epoch": 1.4227642276422765, "grad_norm": 0.12177680432796478, "learning_rate": 1.1258858101936615e-06, "loss": 0.5822, "step": 1708 }, { "epoch": 1.4235980821346674, "grad_norm": 0.1219385415315628, "learning_rate": 1.1249945383358413e-06, "loss": 0.5946, "step": 1709 }, { "epoch": 1.4244319366270586, "grad_norm": 0.12739478051662445, "learning_rate": 1.1241031655993187e-06, "loss": 0.5985, "step": 1710 }, { "epoch": 1.4252657911194497, "grad_norm": 0.12883968651294708, "learning_rate": 1.1232116927034892e-06, "loss": 0.6438, "step": 1711 }, { "epoch": 1.4260996456118407, "grad_norm": 0.1276615411043167, "learning_rate": 1.1223201203678287e-06, "loss": 0.6173, "step": 1712 }, { "epoch": 1.4269335001042318, "grad_norm": 0.11872417479753494, "learning_rate": 1.1214284493118946e-06, "loss": 0.5831, "step": 1713 }, { "epoch": 1.427767354596623, "grad_norm": 0.12630686163902283, "learning_rate": 1.1205366802553228e-06, "loss": 0.6089, "step": 1714 }, { "epoch": 1.428601209089014, "grad_norm": 0.13489751517772675, "learning_rate": 1.1196448139178295e-06, "loss": 0.6001, "step": 1715 }, { "epoch": 1.429435063581405, "grad_norm": 0.12865175306797028, "learning_rate": 1.118752851019208e-06, "loss": 0.6606, "step": 1716 }, { "epoch": 1.4302689180737962, "grad_norm": 0.12072816491127014, "learning_rate": 1.1178607922793307e-06, "loss": 0.6222, "step": 1717 }, { "epoch": 1.4311027725661871, "grad_norm": 0.11538699269294739, "learning_rate": 1.1169686384181467e-06, "loss": 0.5956, "step": 1718 }, { "epoch": 1.4319366270585783, "grad_norm": 0.1280529797077179, "learning_rate": 1.116076390155682e-06, "loss": 0.5582, "step": 1719 }, { "epoch": 1.4327704815509694, "grad_norm": 0.11776122450828552, "learning_rate": 1.1151840482120385e-06, "loss": 0.5765, "step": 1720 }, { "epoch": 1.4336043360433606, "grad_norm": 0.12921258807182312, "learning_rate": 1.1142916133073946e-06, "loss": 0.6283, "step": 1721 }, { "epoch": 1.4344381905357515, "grad_norm": 0.11978008598089218, "learning_rate": 1.113399086162003e-06, "loss": 0.5641, "step": 1722 }, { "epoch": 1.4352720450281427, "grad_norm": 0.12989699840545654, "learning_rate": 1.1125064674961913e-06, "loss": 0.5957, "step": 1723 }, { "epoch": 1.4361058995205336, "grad_norm": 0.12442397326231003, "learning_rate": 1.1116137580303602e-06, "loss": 0.5626, "step": 1724 }, { "epoch": 1.4369397540129247, "grad_norm": 0.12246950715780258, "learning_rate": 1.1107209584849844e-06, "loss": 0.6279, "step": 1725 }, { "epoch": 1.4377736085053159, "grad_norm": 0.12757597863674164, "learning_rate": 1.109828069580611e-06, "loss": 0.6326, "step": 1726 }, { "epoch": 1.438607462997707, "grad_norm": 0.12838071584701538, "learning_rate": 1.108935092037859e-06, "loss": 0.6689, "step": 1727 }, { "epoch": 1.439441317490098, "grad_norm": 0.13266290724277496, "learning_rate": 1.10804202657742e-06, "loss": 0.6087, "step": 1728 }, { "epoch": 1.440275171982489, "grad_norm": 0.1249866783618927, "learning_rate": 1.107148873920055e-06, "loss": 0.5671, "step": 1729 }, { "epoch": 1.44110902647488, "grad_norm": 0.12549594044685364, "learning_rate": 1.1062556347865966e-06, "loss": 0.6026, "step": 1730 }, { "epoch": 1.4419428809672712, "grad_norm": 0.11809445917606354, "learning_rate": 1.1053623098979464e-06, "loss": 0.6609, "step": 1731 }, { "epoch": 1.4427767354596623, "grad_norm": 0.1284630000591278, "learning_rate": 1.1044688999750757e-06, "loss": 0.6976, "step": 1732 }, { "epoch": 1.4436105899520535, "grad_norm": 0.12497500330209732, "learning_rate": 1.1035754057390245e-06, "loss": 0.6037, "step": 1733 }, { "epoch": 1.4444444444444444, "grad_norm": 0.12468233704566956, "learning_rate": 1.1026818279109002e-06, "loss": 0.5953, "step": 1734 }, { "epoch": 1.4452782989368356, "grad_norm": 0.1244860291481018, "learning_rate": 1.1017881672118785e-06, "loss": 0.5816, "step": 1735 }, { "epoch": 1.4461121534292265, "grad_norm": 0.12283075600862503, "learning_rate": 1.1008944243632012e-06, "loss": 0.5896, "step": 1736 }, { "epoch": 1.4469460079216176, "grad_norm": 0.12096507102251053, "learning_rate": 1.100000600086177e-06, "loss": 0.5848, "step": 1737 }, { "epoch": 1.4477798624140088, "grad_norm": 0.13141663372516632, "learning_rate": 1.0991066951021802e-06, "loss": 0.5944, "step": 1738 }, { "epoch": 1.4486137169064, "grad_norm": 0.11670444160699844, "learning_rate": 1.0982127101326497e-06, "loss": 0.5609, "step": 1739 }, { "epoch": 1.4494475713987909, "grad_norm": 0.1190413385629654, "learning_rate": 1.09731864589909e-06, "loss": 0.639, "step": 1740 }, { "epoch": 1.450281425891182, "grad_norm": 0.12622325122356415, "learning_rate": 1.0964245031230684e-06, "loss": 0.6146, "step": 1741 }, { "epoch": 1.451115280383573, "grad_norm": 0.12580569088459015, "learning_rate": 1.0955302825262161e-06, "loss": 0.5947, "step": 1742 }, { "epoch": 1.451949134875964, "grad_norm": 0.1237453743815422, "learning_rate": 1.0946359848302273e-06, "loss": 0.6203, "step": 1743 }, { "epoch": 1.4527829893683553, "grad_norm": 0.11954555660486221, "learning_rate": 1.0937416107568586e-06, "loss": 0.5816, "step": 1744 }, { "epoch": 1.4536168438607464, "grad_norm": 0.125918909907341, "learning_rate": 1.0928471610279278e-06, "loss": 0.5414, "step": 1745 }, { "epoch": 1.4544506983531373, "grad_norm": 0.1135488972067833, "learning_rate": 1.0919526363653133e-06, "loss": 0.5737, "step": 1746 }, { "epoch": 1.4552845528455285, "grad_norm": 0.1270725578069687, "learning_rate": 1.091058037490955e-06, "loss": 0.7054, "step": 1747 }, { "epoch": 1.4561184073379194, "grad_norm": 0.11943162232637405, "learning_rate": 1.0901633651268524e-06, "loss": 0.601, "step": 1748 }, { "epoch": 1.4569522618303106, "grad_norm": 0.12362517416477203, "learning_rate": 1.089268619995064e-06, "loss": 0.6438, "step": 1749 }, { "epoch": 1.4577861163227017, "grad_norm": 0.11327839642763138, "learning_rate": 1.0883738028177068e-06, "loss": 0.5466, "step": 1750 }, { "epoch": 1.4586199708150929, "grad_norm": 0.11802560091018677, "learning_rate": 1.0874789143169567e-06, "loss": 0.5889, "step": 1751 }, { "epoch": 1.4594538253074838, "grad_norm": 0.1234131008386612, "learning_rate": 1.0865839552150469e-06, "loss": 0.6463, "step": 1752 }, { "epoch": 1.460287679799875, "grad_norm": 0.11657127737998962, "learning_rate": 1.085688926234267e-06, "loss": 0.5814, "step": 1753 }, { "epoch": 1.4611215342922659, "grad_norm": 0.12287198007106781, "learning_rate": 1.084793828096964e-06, "loss": 0.6292, "step": 1754 }, { "epoch": 1.461955388784657, "grad_norm": 0.11891543120145798, "learning_rate": 1.0838986615255397e-06, "loss": 0.6176, "step": 1755 }, { "epoch": 1.4627892432770482, "grad_norm": 0.12395941466093063, "learning_rate": 1.0830034272424518e-06, "loss": 0.5912, "step": 1756 }, { "epoch": 1.4636230977694393, "grad_norm": 0.12994913756847382, "learning_rate": 1.0821081259702126e-06, "loss": 0.6032, "step": 1757 }, { "epoch": 1.4644569522618303, "grad_norm": 0.12760955095291138, "learning_rate": 1.0812127584313882e-06, "loss": 0.619, "step": 1758 }, { "epoch": 1.4652908067542214, "grad_norm": 0.13421405851840973, "learning_rate": 1.0803173253485982e-06, "loss": 0.5927, "step": 1759 }, { "epoch": 1.4661246612466123, "grad_norm": 0.1249236911535263, "learning_rate": 1.0794218274445153e-06, "loss": 0.5979, "step": 1760 }, { "epoch": 1.4669585157390035, "grad_norm": 0.12936027348041534, "learning_rate": 1.0785262654418645e-06, "loss": 0.5907, "step": 1761 }, { "epoch": 1.4677923702313946, "grad_norm": 0.1282528042793274, "learning_rate": 1.0776306400634225e-06, "loss": 0.6585, "step": 1762 }, { "epoch": 1.4686262247237858, "grad_norm": 0.12033521384000778, "learning_rate": 1.0767349520320165e-06, "loss": 0.5572, "step": 1763 }, { "epoch": 1.4694600792161767, "grad_norm": 0.12083300203084946, "learning_rate": 1.0758392020705255e-06, "loss": 0.6449, "step": 1764 }, { "epoch": 1.4702939337085679, "grad_norm": 0.1208571195602417, "learning_rate": 1.0749433909018776e-06, "loss": 0.6389, "step": 1765 }, { "epoch": 1.4711277882009588, "grad_norm": 0.11647844314575195, "learning_rate": 1.0740475192490509e-06, "loss": 0.6427, "step": 1766 }, { "epoch": 1.47196164269335, "grad_norm": 0.12908311188220978, "learning_rate": 1.073151587835071e-06, "loss": 0.6074, "step": 1767 }, { "epoch": 1.472795497185741, "grad_norm": 0.12080005556344986, "learning_rate": 1.0722555973830138e-06, "loss": 0.6037, "step": 1768 }, { "epoch": 1.4736293516781322, "grad_norm": 0.12868741154670715, "learning_rate": 1.0713595486160011e-06, "loss": 0.6149, "step": 1769 }, { "epoch": 1.4744632061705232, "grad_norm": 0.12102212756872177, "learning_rate": 1.0704634422572027e-06, "loss": 0.5885, "step": 1770 }, { "epoch": 1.4752970606629143, "grad_norm": 0.11719702929258347, "learning_rate": 1.069567279029834e-06, "loss": 0.5335, "step": 1771 }, { "epoch": 1.4761309151553055, "grad_norm": 0.12366301566362381, "learning_rate": 1.0686710596571578e-06, "loss": 0.5753, "step": 1772 }, { "epoch": 1.4769647696476964, "grad_norm": 0.1275297850370407, "learning_rate": 1.0677747848624803e-06, "loss": 0.6315, "step": 1773 }, { "epoch": 1.4777986241400876, "grad_norm": 0.11969911307096481, "learning_rate": 1.066878455369154e-06, "loss": 0.5857, "step": 1774 }, { "epoch": 1.4786324786324787, "grad_norm": 0.13167330622673035, "learning_rate": 1.0659820719005749e-06, "loss": 0.6526, "step": 1775 }, { "epoch": 1.4794663331248696, "grad_norm": 0.12339498847723007, "learning_rate": 1.0650856351801824e-06, "loss": 0.6355, "step": 1776 }, { "epoch": 1.4803001876172608, "grad_norm": 0.12983542680740356, "learning_rate": 1.0641891459314595e-06, "loss": 0.6069, "step": 1777 }, { "epoch": 1.481134042109652, "grad_norm": 0.1269686222076416, "learning_rate": 1.063292604877931e-06, "loss": 0.6478, "step": 1778 }, { "epoch": 1.4819678966020429, "grad_norm": 0.12114999443292618, "learning_rate": 1.0623960127431635e-06, "loss": 0.5406, "step": 1779 }, { "epoch": 1.482801751094434, "grad_norm": 0.11573608964681625, "learning_rate": 1.061499370250765e-06, "loss": 0.5789, "step": 1780 }, { "epoch": 1.4836356055868252, "grad_norm": 0.1279306858778, "learning_rate": 1.0606026781243847e-06, "loss": 0.6475, "step": 1781 }, { "epoch": 1.484469460079216, "grad_norm": 0.12379437685012817, "learning_rate": 1.0597059370877103e-06, "loss": 0.5484, "step": 1782 }, { "epoch": 1.4853033145716072, "grad_norm": 0.12157581001520157, "learning_rate": 1.0588091478644714e-06, "loss": 0.5866, "step": 1783 }, { "epoch": 1.4861371690639984, "grad_norm": 0.13414230942726135, "learning_rate": 1.0579123111784342e-06, "loss": 0.6632, "step": 1784 }, { "epoch": 1.4869710235563893, "grad_norm": 0.12635232508182526, "learning_rate": 1.057015427753404e-06, "loss": 0.6577, "step": 1785 }, { "epoch": 1.4878048780487805, "grad_norm": 0.13163447380065918, "learning_rate": 1.0561184983132244e-06, "loss": 0.6335, "step": 1786 }, { "epoch": 1.4886387325411716, "grad_norm": 0.12453741580247879, "learning_rate": 1.0552215235817754e-06, "loss": 0.6312, "step": 1787 }, { "epoch": 1.4894725870335628, "grad_norm": 0.12222108989953995, "learning_rate": 1.054324504282974e-06, "loss": 0.6371, "step": 1788 }, { "epoch": 1.4903064415259537, "grad_norm": 0.1175316870212555, "learning_rate": 1.0534274411407725e-06, "loss": 0.5479, "step": 1789 }, { "epoch": 1.4911402960183449, "grad_norm": 0.1285734921693802, "learning_rate": 1.0525303348791596e-06, "loss": 0.6339, "step": 1790 }, { "epoch": 1.4919741505107358, "grad_norm": 0.1263425052165985, "learning_rate": 1.0516331862221582e-06, "loss": 0.6376, "step": 1791 }, { "epoch": 1.492808005003127, "grad_norm": 0.13151004910469055, "learning_rate": 1.050735995893825e-06, "loss": 0.5966, "step": 1792 }, { "epoch": 1.493641859495518, "grad_norm": 0.12273474782705307, "learning_rate": 1.0498387646182511e-06, "loss": 0.6469, "step": 1793 }, { "epoch": 1.4944757139879092, "grad_norm": 0.1246555745601654, "learning_rate": 1.0489414931195603e-06, "loss": 0.5489, "step": 1794 }, { "epoch": 1.4953095684803002, "grad_norm": 0.12344285845756531, "learning_rate": 1.0480441821219096e-06, "loss": 0.5895, "step": 1795 }, { "epoch": 1.4961434229726913, "grad_norm": 0.12281202524900436, "learning_rate": 1.047146832349486e-06, "loss": 0.595, "step": 1796 }, { "epoch": 1.4969772774650822, "grad_norm": 0.13203196227550507, "learning_rate": 1.0462494445265098e-06, "loss": 0.6427, "step": 1797 }, { "epoch": 1.4978111319574734, "grad_norm": 0.12045428156852722, "learning_rate": 1.045352019377231e-06, "loss": 0.5412, "step": 1798 }, { "epoch": 1.4986449864498645, "grad_norm": 0.1290232390165329, "learning_rate": 1.0444545576259302e-06, "loss": 0.6301, "step": 1799 }, { "epoch": 1.4994788409422557, "grad_norm": 0.12524542212486267, "learning_rate": 1.0435570599969165e-06, "loss": 0.6474, "step": 1800 }, { "epoch": 1.5003126954346466, "grad_norm": 0.11619977653026581, "learning_rate": 1.0426595272145295e-06, "loss": 0.5225, "step": 1801 }, { "epoch": 1.5011465499270378, "grad_norm": 0.12204474955797195, "learning_rate": 1.041761960003136e-06, "loss": 0.5938, "step": 1802 }, { "epoch": 1.5019804044194287, "grad_norm": 0.13166320323944092, "learning_rate": 1.0408643590871312e-06, "loss": 0.6535, "step": 1803 }, { "epoch": 1.5028142589118199, "grad_norm": 0.12408415228128433, "learning_rate": 1.0399667251909368e-06, "loss": 0.5745, "step": 1804 }, { "epoch": 1.503648113404211, "grad_norm": 0.1306605190038681, "learning_rate": 1.039069059039002e-06, "loss": 0.6438, "step": 1805 }, { "epoch": 1.5044819678966022, "grad_norm": 0.11977241933345795, "learning_rate": 1.0381713613558022e-06, "loss": 0.6051, "step": 1806 }, { "epoch": 1.505315822388993, "grad_norm": 0.11851506680250168, "learning_rate": 1.0372736328658363e-06, "loss": 0.6246, "step": 1807 }, { "epoch": 1.5061496768813842, "grad_norm": 0.12838518619537354, "learning_rate": 1.0363758742936306e-06, "loss": 0.6446, "step": 1808 }, { "epoch": 1.5069835313737752, "grad_norm": 0.11811474710702896, "learning_rate": 1.0354780863637337e-06, "loss": 0.5982, "step": 1809 }, { "epoch": 1.5078173858661663, "grad_norm": 0.12509089708328247, "learning_rate": 1.0345802698007196e-06, "loss": 0.6321, "step": 1810 }, { "epoch": 1.5086512403585575, "grad_norm": 0.1284075677394867, "learning_rate": 1.0336824253291835e-06, "loss": 0.641, "step": 1811 }, { "epoch": 1.5094850948509486, "grad_norm": 0.12310284376144409, "learning_rate": 1.0327845536737453e-06, "loss": 0.6306, "step": 1812 }, { "epoch": 1.5103189493433395, "grad_norm": 0.12668739259243011, "learning_rate": 1.031886655559045e-06, "loss": 0.6587, "step": 1813 }, { "epoch": 1.5111528038357307, "grad_norm": 0.12098442763090134, "learning_rate": 1.0309887317097444e-06, "loss": 0.6573, "step": 1814 }, { "epoch": 1.5119866583281216, "grad_norm": 0.1259874552488327, "learning_rate": 1.0300907828505267e-06, "loss": 0.641, "step": 1815 }, { "epoch": 1.5128205128205128, "grad_norm": 0.13358241319656372, "learning_rate": 1.0291928097060948e-06, "loss": 0.6342, "step": 1816 }, { "epoch": 1.513654367312904, "grad_norm": 0.12032781541347504, "learning_rate": 1.0282948130011715e-06, "loss": 0.5766, "step": 1817 }, { "epoch": 1.514488221805295, "grad_norm": 0.11778460443019867, "learning_rate": 1.027396793460498e-06, "loss": 0.5576, "step": 1818 }, { "epoch": 1.515322076297686, "grad_norm": 0.12484540045261383, "learning_rate": 1.0264987518088346e-06, "loss": 0.6544, "step": 1819 }, { "epoch": 1.5161559307900772, "grad_norm": 0.12196074426174164, "learning_rate": 1.0256006887709592e-06, "loss": 0.5936, "step": 1820 }, { "epoch": 1.516989785282468, "grad_norm": 0.13073596358299255, "learning_rate": 1.024702605071667e-06, "loss": 0.6153, "step": 1821 }, { "epoch": 1.5178236397748592, "grad_norm": 0.124057836830616, "learning_rate": 1.0238045014357692e-06, "loss": 0.6284, "step": 1822 }, { "epoch": 1.5186574942672504, "grad_norm": 0.12679198384284973, "learning_rate": 1.0229063785880948e-06, "loss": 0.5893, "step": 1823 }, { "epoch": 1.5194913487596415, "grad_norm": 0.128417506814003, "learning_rate": 1.022008237253486e-06, "loss": 0.5923, "step": 1824 }, { "epoch": 1.5203252032520327, "grad_norm": 0.12480554729700089, "learning_rate": 1.0211100781568023e-06, "loss": 0.606, "step": 1825 }, { "epoch": 1.5211590577444236, "grad_norm": 0.1268758475780487, "learning_rate": 1.0202119020229154e-06, "loss": 0.6155, "step": 1826 }, { "epoch": 1.5219929122368145, "grad_norm": 0.12820713222026825, "learning_rate": 1.0193137095767123e-06, "loss": 0.5935, "step": 1827 }, { "epoch": 1.5228267667292057, "grad_norm": 0.11923402547836304, "learning_rate": 1.0184155015430926e-06, "loss": 0.6327, "step": 1828 }, { "epoch": 1.5236606212215968, "grad_norm": 0.11409208923578262, "learning_rate": 1.017517278646968e-06, "loss": 0.5759, "step": 1829 }, { "epoch": 1.524494475713988, "grad_norm": 0.12284267693758011, "learning_rate": 1.016619041613263e-06, "loss": 0.5775, "step": 1830 }, { "epoch": 1.5253283302063791, "grad_norm": 0.12808358669281006, "learning_rate": 1.015720791166913e-06, "loss": 0.5413, "step": 1831 }, { "epoch": 1.52616218469877, "grad_norm": 0.11841126531362534, "learning_rate": 1.0148225280328647e-06, "loss": 0.573, "step": 1832 }, { "epoch": 1.526996039191161, "grad_norm": 0.12383410334587097, "learning_rate": 1.0139242529360744e-06, "loss": 0.594, "step": 1833 }, { "epoch": 1.5278298936835522, "grad_norm": 0.12949687242507935, "learning_rate": 1.0130259666015084e-06, "loss": 0.5782, "step": 1834 }, { "epoch": 1.5286637481759433, "grad_norm": 0.12045048922300339, "learning_rate": 1.0121276697541426e-06, "loss": 0.6006, "step": 1835 }, { "epoch": 1.5294976026683345, "grad_norm": 0.1283283829689026, "learning_rate": 1.01122936311896e-06, "loss": 0.638, "step": 1836 }, { "epoch": 1.5303314571607256, "grad_norm": 0.12735942006111145, "learning_rate": 1.0103310474209527e-06, "loss": 0.6341, "step": 1837 }, { "epoch": 1.5311653116531165, "grad_norm": 0.12286734580993652, "learning_rate": 1.0094327233851199e-06, "loss": 0.5176, "step": 1838 }, { "epoch": 1.5319991661455075, "grad_norm": 0.12337525188922882, "learning_rate": 1.0085343917364675e-06, "loss": 0.5797, "step": 1839 }, { "epoch": 1.5328330206378986, "grad_norm": 0.12428612262010574, "learning_rate": 1.0076360532000068e-06, "loss": 0.5969, "step": 1840 }, { "epoch": 1.5336668751302898, "grad_norm": 0.12380371987819672, "learning_rate": 1.006737708500756e-06, "loss": 0.6301, "step": 1841 }, { "epoch": 1.534500729622681, "grad_norm": 0.12506607174873352, "learning_rate": 1.0058393583637376e-06, "loss": 0.6016, "step": 1842 }, { "epoch": 1.535334584115072, "grad_norm": 0.12334273755550385, "learning_rate": 1.0049410035139774e-06, "loss": 0.5499, "step": 1843 }, { "epoch": 1.536168438607463, "grad_norm": 0.12097372114658356, "learning_rate": 1.0040426446765074e-06, "loss": 0.6183, "step": 1844 }, { "epoch": 1.537002293099854, "grad_norm": 0.13287068903446198, "learning_rate": 1.0031442825763603e-06, "loss": 0.66, "step": 1845 }, { "epoch": 1.537836147592245, "grad_norm": 0.12525850534439087, "learning_rate": 1.0022459179385736e-06, "loss": 0.6031, "step": 1846 }, { "epoch": 1.5386700020846362, "grad_norm": 0.11843365430831909, "learning_rate": 1.001347551488185e-06, "loss": 0.6228, "step": 1847 }, { "epoch": 1.5395038565770274, "grad_norm": 0.12727557122707367, "learning_rate": 1.0004491839502352e-06, "loss": 0.5877, "step": 1848 }, { "epoch": 1.5403377110694185, "grad_norm": 0.12447907030582428, "learning_rate": 9.99550816049765e-07, "loss": 0.6149, "step": 1849 }, { "epoch": 1.5411715655618095, "grad_norm": 0.12982895970344543, "learning_rate": 9.986524485118152e-07, "loss": 0.6134, "step": 1850 }, { "epoch": 1.5420054200542004, "grad_norm": 0.13101699948310852, "learning_rate": 9.977540820614265e-07, "loss": 0.5873, "step": 1851 }, { "epoch": 1.5428392745465915, "grad_norm": 0.126919686794281, "learning_rate": 9.968557174236396e-07, "loss": 0.5707, "step": 1852 }, { "epoch": 1.5436731290389827, "grad_norm": 0.12582388520240784, "learning_rate": 9.95957355323493e-07, "loss": 0.5985, "step": 1853 }, { "epoch": 1.5445069835313738, "grad_norm": 0.13102799654006958, "learning_rate": 9.950589964860226e-07, "loss": 0.5939, "step": 1854 }, { "epoch": 1.545340838023765, "grad_norm": 0.13151919841766357, "learning_rate": 9.941606416362628e-07, "loss": 0.6505, "step": 1855 }, { "epoch": 1.546174692516156, "grad_norm": 0.12860901653766632, "learning_rate": 9.932622914992437e-07, "loss": 0.6129, "step": 1856 }, { "epoch": 1.547008547008547, "grad_norm": 0.1313966065645218, "learning_rate": 9.923639467999929e-07, "loss": 0.6383, "step": 1857 }, { "epoch": 1.547842401500938, "grad_norm": 0.12255886942148209, "learning_rate": 9.914656082635326e-07, "loss": 0.5915, "step": 1858 }, { "epoch": 1.5486762559933291, "grad_norm": 0.12750515341758728, "learning_rate": 9.9056727661488e-07, "loss": 0.5972, "step": 1859 }, { "epoch": 1.5495101104857203, "grad_norm": 0.12609492242336273, "learning_rate": 9.896689525790472e-07, "loss": 0.6454, "step": 1860 }, { "epoch": 1.5503439649781114, "grad_norm": 0.1195141151547432, "learning_rate": 9.887706368810402e-07, "loss": 0.5384, "step": 1861 }, { "epoch": 1.5511778194705024, "grad_norm": 0.12656457722187042, "learning_rate": 9.878723302458576e-07, "loss": 0.6791, "step": 1862 }, { "epoch": 1.5520116739628935, "grad_norm": 0.1253191977739334, "learning_rate": 9.869740333984917e-07, "loss": 0.5846, "step": 1863 }, { "epoch": 1.5528455284552845, "grad_norm": 0.12685538828372955, "learning_rate": 9.860757470639258e-07, "loss": 0.6074, "step": 1864 }, { "epoch": 1.5536793829476756, "grad_norm": 0.12786641716957092, "learning_rate": 9.851774719671355e-07, "loss": 0.6289, "step": 1865 }, { "epoch": 1.5545132374400668, "grad_norm": 0.12862898409366608, "learning_rate": 9.84279208833087e-07, "loss": 0.6372, "step": 1866 }, { "epoch": 1.555347091932458, "grad_norm": 0.1340702474117279, "learning_rate": 9.833809583867374e-07, "loss": 0.6325, "step": 1867 }, { "epoch": 1.5561809464248488, "grad_norm": 0.12695550918579102, "learning_rate": 9.824827213530321e-07, "loss": 0.631, "step": 1868 }, { "epoch": 1.55701480091724, "grad_norm": 0.12067730724811554, "learning_rate": 9.815844984569077e-07, "loss": 0.51, "step": 1869 }, { "epoch": 1.557848655409631, "grad_norm": 0.1230134516954422, "learning_rate": 9.806862904232878e-07, "loss": 0.5481, "step": 1870 }, { "epoch": 1.558682509902022, "grad_norm": 0.11921508610248566, "learning_rate": 9.797880979770843e-07, "loss": 0.5222, "step": 1871 }, { "epoch": 1.5595163643944132, "grad_norm": 0.12758903205394745, "learning_rate": 9.788899218431978e-07, "loss": 0.5841, "step": 1872 }, { "epoch": 1.5603502188868044, "grad_norm": 0.12362825870513916, "learning_rate": 9.779917627465138e-07, "loss": 0.5896, "step": 1873 }, { "epoch": 1.5611840733791953, "grad_norm": 0.11825423687696457, "learning_rate": 9.770936214119053e-07, "loss": 0.5742, "step": 1874 }, { "epoch": 1.5620179278715864, "grad_norm": 0.12133418768644333, "learning_rate": 9.761954985642307e-07, "loss": 0.5397, "step": 1875 }, { "epoch": 1.5628517823639774, "grad_norm": 0.1267905831336975, "learning_rate": 9.752973949283332e-07, "loss": 0.5854, "step": 1876 }, { "epoch": 1.5636856368563685, "grad_norm": 0.12750284373760223, "learning_rate": 9.743993112290408e-07, "loss": 0.6361, "step": 1877 }, { "epoch": 1.5645194913487597, "grad_norm": 0.12900686264038086, "learning_rate": 9.735012481911653e-07, "loss": 0.5673, "step": 1878 }, { "epoch": 1.5653533458411508, "grad_norm": 0.12558184564113617, "learning_rate": 9.726032065395022e-07, "loss": 0.6087, "step": 1879 }, { "epoch": 1.5661872003335418, "grad_norm": 0.12879346311092377, "learning_rate": 9.717051869988286e-07, "loss": 0.6159, "step": 1880 }, { "epoch": 1.567021054825933, "grad_norm": 0.12538638710975647, "learning_rate": 9.708071902939053e-07, "loss": 0.5894, "step": 1881 }, { "epoch": 1.5678549093183238, "grad_norm": 0.12777774035930634, "learning_rate": 9.699092171494735e-07, "loss": 0.6216, "step": 1882 }, { "epoch": 1.568688763810715, "grad_norm": 0.12464749068021774, "learning_rate": 9.690112682902557e-07, "loss": 0.6147, "step": 1883 }, { "epoch": 1.5695226183031061, "grad_norm": 0.12391242384910583, "learning_rate": 9.681133444409552e-07, "loss": 0.6487, "step": 1884 }, { "epoch": 1.5703564727954973, "grad_norm": 0.12629380822181702, "learning_rate": 9.672154463262546e-07, "loss": 0.5764, "step": 1885 }, { "epoch": 1.5711903272878882, "grad_norm": 0.13301175832748413, "learning_rate": 9.663175746708162e-07, "loss": 0.5439, "step": 1886 }, { "epoch": 1.5720241817802794, "grad_norm": 0.1252022087574005, "learning_rate": 9.654197301992805e-07, "loss": 0.6084, "step": 1887 }, { "epoch": 1.5728580362726703, "grad_norm": 0.12484809011220932, "learning_rate": 9.64521913636266e-07, "loss": 0.6254, "step": 1888 }, { "epoch": 1.5736918907650614, "grad_norm": 0.13050304353237152, "learning_rate": 9.636241257063695e-07, "loss": 0.5849, "step": 1889 }, { "epoch": 1.5745257452574526, "grad_norm": 0.1395888775587082, "learning_rate": 9.627263671341638e-07, "loss": 0.6367, "step": 1890 }, { "epoch": 1.5753595997498437, "grad_norm": 0.12380406260490417, "learning_rate": 9.61828638644198e-07, "loss": 0.6102, "step": 1891 }, { "epoch": 1.576193454242235, "grad_norm": 0.1276310682296753, "learning_rate": 9.609309409609978e-07, "loss": 0.6069, "step": 1892 }, { "epoch": 1.5770273087346258, "grad_norm": 0.12831759452819824, "learning_rate": 9.600332748090633e-07, "loss": 0.6047, "step": 1893 }, { "epoch": 1.5778611632270167, "grad_norm": 0.12802143394947052, "learning_rate": 9.59135640912869e-07, "loss": 0.5987, "step": 1894 }, { "epoch": 1.578695017719408, "grad_norm": 0.12463508546352386, "learning_rate": 9.582380399968642e-07, "loss": 0.5831, "step": 1895 }, { "epoch": 1.579528872211799, "grad_norm": 0.13273175060749054, "learning_rate": 9.573404727854708e-07, "loss": 0.5763, "step": 1896 }, { "epoch": 1.5803627267041902, "grad_norm": 0.1212095320224762, "learning_rate": 9.564429400030836e-07, "loss": 0.6321, "step": 1897 }, { "epoch": 1.5811965811965814, "grad_norm": 0.13839462399482727, "learning_rate": 9.555454423740702e-07, "loss": 0.565, "step": 1898 }, { "epoch": 1.5820304356889723, "grad_norm": 0.12162493169307709, "learning_rate": 9.54647980622769e-07, "loss": 0.633, "step": 1899 }, { "epoch": 1.5828642901813632, "grad_norm": 0.13095051050186157, "learning_rate": 9.5375055547349e-07, "loss": 0.6058, "step": 1900 }, { "epoch": 1.5836981446737544, "grad_norm": 0.12818747758865356, "learning_rate": 9.528531676505139e-07, "loss": 0.63, "step": 1901 }, { "epoch": 1.5845319991661455, "grad_norm": 0.12812934815883636, "learning_rate": 9.519558178780903e-07, "loss": 0.6276, "step": 1902 }, { "epoch": 1.5853658536585367, "grad_norm": 0.12593896687030792, "learning_rate": 9.510585068804393e-07, "loss": 0.6017, "step": 1903 }, { "epoch": 1.5861997081509278, "grad_norm": 0.12588874995708466, "learning_rate": 9.501612353817489e-07, "loss": 0.5889, "step": 1904 }, { "epoch": 1.5870335626433187, "grad_norm": 0.12574082612991333, "learning_rate": 9.492640041061751e-07, "loss": 0.642, "step": 1905 }, { "epoch": 1.5878674171357097, "grad_norm": 0.125298872590065, "learning_rate": 9.48366813777842e-07, "loss": 0.5936, "step": 1906 }, { "epoch": 1.5887012716281008, "grad_norm": 0.13028943538665771, "learning_rate": 9.474696651208405e-07, "loss": 0.5672, "step": 1907 }, { "epoch": 1.589535126120492, "grad_norm": 0.12832380831241608, "learning_rate": 9.465725588592276e-07, "loss": 0.5677, "step": 1908 }, { "epoch": 1.5903689806128831, "grad_norm": 0.1250237077474594, "learning_rate": 9.456754957170261e-07, "loss": 0.5445, "step": 1909 }, { "epoch": 1.5912028351052743, "grad_norm": 0.12492944300174713, "learning_rate": 9.447784764182246e-07, "loss": 0.503, "step": 1910 }, { "epoch": 1.5920366895976652, "grad_norm": 0.13353389501571655, "learning_rate": 9.438815016867757e-07, "loss": 0.5951, "step": 1911 }, { "epoch": 1.5928705440900561, "grad_norm": 0.12580597400665283, "learning_rate": 9.429845722465959e-07, "loss": 0.6326, "step": 1912 }, { "epoch": 1.5937043985824473, "grad_norm": 0.13248313963413239, "learning_rate": 9.42087688821566e-07, "loss": 0.6321, "step": 1913 }, { "epoch": 1.5945382530748384, "grad_norm": 0.12739625573158264, "learning_rate": 9.411908521355288e-07, "loss": 0.5681, "step": 1914 }, { "epoch": 1.5953721075672296, "grad_norm": 0.12767314910888672, "learning_rate": 9.402940629122893e-07, "loss": 0.6109, "step": 1915 }, { "epoch": 1.5962059620596207, "grad_norm": 0.12399329990148544, "learning_rate": 9.393973218756155e-07, "loss": 0.6677, "step": 1916 }, { "epoch": 1.5970398165520117, "grad_norm": 0.12876498699188232, "learning_rate": 9.385006297492348e-07, "loss": 0.6214, "step": 1917 }, { "epoch": 1.5978736710444026, "grad_norm": 0.13340961933135986, "learning_rate": 9.376039872568367e-07, "loss": 0.5851, "step": 1918 }, { "epoch": 1.5987075255367937, "grad_norm": 0.12147720158100128, "learning_rate": 9.367073951220692e-07, "loss": 0.6, "step": 1919 }, { "epoch": 1.599541380029185, "grad_norm": 0.12896296381950378, "learning_rate": 9.358108540685404e-07, "loss": 0.6626, "step": 1920 }, { "epoch": 1.600375234521576, "grad_norm": 0.12212686985731125, "learning_rate": 9.349143648198174e-07, "loss": 0.5925, "step": 1921 }, { "epoch": 1.6012090890139672, "grad_norm": 0.125599205493927, "learning_rate": 9.340179280994251e-07, "loss": 0.6106, "step": 1922 }, { "epoch": 1.6020429435063581, "grad_norm": 0.1354406625032425, "learning_rate": 9.331215446308459e-07, "loss": 0.6041, "step": 1923 }, { "epoch": 1.6028767979987493, "grad_norm": 0.12907730042934418, "learning_rate": 9.322252151375199e-07, "loss": 0.6636, "step": 1924 }, { "epoch": 1.6037106524911402, "grad_norm": 0.13119882345199585, "learning_rate": 9.313289403428425e-07, "loss": 0.581, "step": 1925 }, { "epoch": 1.6045445069835313, "grad_norm": 0.13200044631958008, "learning_rate": 9.304327209701661e-07, "loss": 0.6078, "step": 1926 }, { "epoch": 1.6053783614759225, "grad_norm": 0.12081551551818848, "learning_rate": 9.295365577427975e-07, "loss": 0.6248, "step": 1927 }, { "epoch": 1.6062122159683137, "grad_norm": 0.12003374099731445, "learning_rate": 9.286404513839991e-07, "loss": 0.5611, "step": 1928 }, { "epoch": 1.6070460704607046, "grad_norm": 0.12585462629795074, "learning_rate": 9.27744402616986e-07, "loss": 0.612, "step": 1929 }, { "epoch": 1.6078799249530957, "grad_norm": 0.12522676587104797, "learning_rate": 9.268484121649287e-07, "loss": 0.5522, "step": 1930 }, { "epoch": 1.6087137794454867, "grad_norm": 0.13690777122974396, "learning_rate": 9.25952480750949e-07, "loss": 0.6234, "step": 1931 }, { "epoch": 1.6095476339378778, "grad_norm": 0.12669777870178223, "learning_rate": 9.250566090981221e-07, "loss": 0.5642, "step": 1932 }, { "epoch": 1.610381488430269, "grad_norm": 0.12308058142662048, "learning_rate": 9.241607979294744e-07, "loss": 0.5617, "step": 1933 }, { "epoch": 1.61121534292266, "grad_norm": 0.12228291481733322, "learning_rate": 9.232650479679835e-07, "loss": 0.6025, "step": 1934 }, { "epoch": 1.612049197415051, "grad_norm": 0.12602399289608002, "learning_rate": 9.223693599365776e-07, "loss": 0.5623, "step": 1935 }, { "epoch": 1.6128830519074422, "grad_norm": 0.12143569439649582, "learning_rate": 9.214737345581355e-07, "loss": 0.603, "step": 1936 }, { "epoch": 1.6137169063998331, "grad_norm": 0.13084720075130463, "learning_rate": 9.205781725554848e-07, "loss": 0.6486, "step": 1937 }, { "epoch": 1.6145507608922243, "grad_norm": 0.1243109330534935, "learning_rate": 9.196826746514017e-07, "loss": 0.5865, "step": 1938 }, { "epoch": 1.6153846153846154, "grad_norm": 0.12693606317043304, "learning_rate": 9.187872415686118e-07, "loss": 0.5937, "step": 1939 }, { "epoch": 1.6162184698770066, "grad_norm": 0.13238000869750977, "learning_rate": 9.178918740297875e-07, "loss": 0.5982, "step": 1940 }, { "epoch": 1.6170523243693975, "grad_norm": 0.12275899946689606, "learning_rate": 9.169965727575482e-07, "loss": 0.6161, "step": 1941 }, { "epoch": 1.6178861788617886, "grad_norm": 0.13157354295253754, "learning_rate": 9.161013384744605e-07, "loss": 0.5869, "step": 1942 }, { "epoch": 1.6187200333541796, "grad_norm": 0.13102547824382782, "learning_rate": 9.152061719030364e-07, "loss": 0.613, "step": 1943 }, { "epoch": 1.6195538878465707, "grad_norm": 0.12562698125839233, "learning_rate": 9.143110737657329e-07, "loss": 0.5896, "step": 1944 }, { "epoch": 1.6203877423389619, "grad_norm": 0.13320913910865784, "learning_rate": 9.134160447849532e-07, "loss": 0.6378, "step": 1945 }, { "epoch": 1.621221596831353, "grad_norm": 0.12788859009742737, "learning_rate": 9.125210856830432e-07, "loss": 0.6269, "step": 1946 }, { "epoch": 1.622055451323744, "grad_norm": 0.11895091086626053, "learning_rate": 9.116261971822932e-07, "loss": 0.6095, "step": 1947 }, { "epoch": 1.622889305816135, "grad_norm": 0.12486100941896439, "learning_rate": 9.107313800049362e-07, "loss": 0.6176, "step": 1948 }, { "epoch": 1.623723160308526, "grad_norm": 0.12875628471374512, "learning_rate": 9.098366348731476e-07, "loss": 0.6406, "step": 1949 }, { "epoch": 1.6245570148009172, "grad_norm": 0.131727933883667, "learning_rate": 9.089419625090448e-07, "loss": 0.6074, "step": 1950 }, { "epoch": 1.6253908692933083, "grad_norm": 0.1188148781657219, "learning_rate": 9.080473636346868e-07, "loss": 0.4721, "step": 1951 }, { "epoch": 1.6262247237856995, "grad_norm": 0.13043268024921417, "learning_rate": 9.071528389720723e-07, "loss": 0.5997, "step": 1952 }, { "epoch": 1.6270585782780906, "grad_norm": 0.12843738496303558, "learning_rate": 9.062583892431414e-07, "loss": 0.5812, "step": 1953 }, { "epoch": 1.6278924327704816, "grad_norm": 0.13008764386177063, "learning_rate": 9.053640151697727e-07, "loss": 0.624, "step": 1954 }, { "epoch": 1.6287262872628725, "grad_norm": 0.12412194907665253, "learning_rate": 9.044697174737842e-07, "loss": 0.5941, "step": 1955 }, { "epoch": 1.6295601417552636, "grad_norm": 0.12464737147092819, "learning_rate": 9.035754968769319e-07, "loss": 0.5998, "step": 1956 }, { "epoch": 1.6303939962476548, "grad_norm": 0.1221119835972786, "learning_rate": 9.026813541009104e-07, "loss": 0.5577, "step": 1957 }, { "epoch": 1.631227850740046, "grad_norm": 0.1313888281583786, "learning_rate": 9.017872898673501e-07, "loss": 0.6428, "step": 1958 }, { "epoch": 1.632061705232437, "grad_norm": 0.13170261681079865, "learning_rate": 9.008933048978198e-07, "loss": 0.627, "step": 1959 }, { "epoch": 1.632895559724828, "grad_norm": 0.13136637210845947, "learning_rate": 8.999993999138227e-07, "loss": 0.5789, "step": 1960 }, { "epoch": 1.633729414217219, "grad_norm": 0.12489289045333862, "learning_rate": 8.991055756367986e-07, "loss": 0.6105, "step": 1961 }, { "epoch": 1.63456326870961, "grad_norm": 0.12347642332315445, "learning_rate": 8.982118327881216e-07, "loss": 0.679, "step": 1962 }, { "epoch": 1.6353971232020013, "grad_norm": 0.12213364243507385, "learning_rate": 8.973181720890998e-07, "loss": 0.6041, "step": 1963 }, { "epoch": 1.6362309776943924, "grad_norm": 0.1292726844549179, "learning_rate": 8.964245942609755e-07, "loss": 0.6276, "step": 1964 }, { "epoch": 1.6370648321867836, "grad_norm": 0.1329735517501831, "learning_rate": 8.955311000249243e-07, "loss": 0.6325, "step": 1965 }, { "epoch": 1.6378986866791745, "grad_norm": 0.12430353462696075, "learning_rate": 8.946376901020538e-07, "loss": 0.5724, "step": 1966 }, { "epoch": 1.6387325411715654, "grad_norm": 0.12788113951683044, "learning_rate": 8.937443652134036e-07, "loss": 0.6015, "step": 1967 }, { "epoch": 1.6395663956639566, "grad_norm": 0.12255391478538513, "learning_rate": 8.928511260799452e-07, "loss": 0.5728, "step": 1968 }, { "epoch": 1.6404002501563477, "grad_norm": 0.12283147871494293, "learning_rate": 8.919579734225803e-07, "loss": 0.5962, "step": 1969 }, { "epoch": 1.6412341046487389, "grad_norm": 0.12731313705444336, "learning_rate": 8.910649079621409e-07, "loss": 0.6335, "step": 1970 }, { "epoch": 1.64206795914113, "grad_norm": 0.1330873817205429, "learning_rate": 8.901719304193893e-07, "loss": 0.5865, "step": 1971 }, { "epoch": 1.642901813633521, "grad_norm": 0.12057091295719147, "learning_rate": 8.892790415150159e-07, "loss": 0.5224, "step": 1972 }, { "epoch": 1.6437356681259119, "grad_norm": 0.12532652914524078, "learning_rate": 8.883862419696397e-07, "loss": 0.657, "step": 1973 }, { "epoch": 1.644569522618303, "grad_norm": 0.12516506016254425, "learning_rate": 8.874935325038086e-07, "loss": 0.5994, "step": 1974 }, { "epoch": 1.6454033771106942, "grad_norm": 0.12904782593250275, "learning_rate": 8.866009138379966e-07, "loss": 0.6043, "step": 1975 }, { "epoch": 1.6462372316030853, "grad_norm": 0.12918344140052795, "learning_rate": 8.857083866926053e-07, "loss": 0.5468, "step": 1976 }, { "epoch": 1.6470710860954765, "grad_norm": 0.13387003540992737, "learning_rate": 8.848159517879615e-07, "loss": 0.6415, "step": 1977 }, { "epoch": 1.6479049405878674, "grad_norm": 0.12973016500473022, "learning_rate": 8.839236098443181e-07, "loss": 0.6151, "step": 1978 }, { "epoch": 1.6487387950802583, "grad_norm": 0.13009996712207794, "learning_rate": 8.830313615818534e-07, "loss": 0.5864, "step": 1979 }, { "epoch": 1.6495726495726495, "grad_norm": 0.1275186836719513, "learning_rate": 8.821392077206694e-07, "loss": 0.5637, "step": 1980 }, { "epoch": 1.6504065040650406, "grad_norm": 0.13509085774421692, "learning_rate": 8.81247148980792e-07, "loss": 0.568, "step": 1981 }, { "epoch": 1.6512403585574318, "grad_norm": 0.12902608513832092, "learning_rate": 8.803551860821705e-07, "loss": 0.6641, "step": 1982 }, { "epoch": 1.652074213049823, "grad_norm": 0.12584422528743744, "learning_rate": 8.79463319744677e-07, "loss": 0.6211, "step": 1983 }, { "epoch": 1.6529080675422139, "grad_norm": 0.12634308636188507, "learning_rate": 8.785715506881056e-07, "loss": 0.5961, "step": 1984 }, { "epoch": 1.653741922034605, "grad_norm": 0.12757942080497742, "learning_rate": 8.776798796321713e-07, "loss": 0.6216, "step": 1985 }, { "epoch": 1.654575776526996, "grad_norm": 0.1245681568980217, "learning_rate": 8.767883072965112e-07, "loss": 0.557, "step": 1986 }, { "epoch": 1.655409631019387, "grad_norm": 0.13303357362747192, "learning_rate": 8.758968344006811e-07, "loss": 0.6282, "step": 1987 }, { "epoch": 1.6562434855117782, "grad_norm": 0.135984405875206, "learning_rate": 8.750054616641584e-07, "loss": 0.6216, "step": 1988 }, { "epoch": 1.6570773400041694, "grad_norm": 0.13211168348789215, "learning_rate": 8.741141898063385e-07, "loss": 0.5596, "step": 1989 }, { "epoch": 1.6579111944965603, "grad_norm": 0.13526758551597595, "learning_rate": 8.732230195465352e-07, "loss": 0.6068, "step": 1990 }, { "epoch": 1.6587450489889515, "grad_norm": 0.1245320737361908, "learning_rate": 8.723319516039813e-07, "loss": 0.5431, "step": 1991 }, { "epoch": 1.6595789034813424, "grad_norm": 0.13527055084705353, "learning_rate": 8.714409866978259e-07, "loss": 0.6634, "step": 1992 }, { "epoch": 1.6604127579737336, "grad_norm": 0.1319219022989273, "learning_rate": 8.705501255471357e-07, "loss": 0.63, "step": 1993 }, { "epoch": 1.6612466124661247, "grad_norm": 0.13556063175201416, "learning_rate": 8.696593688708938e-07, "loss": 0.6061, "step": 1994 }, { "epoch": 1.6620804669585159, "grad_norm": 0.13379541039466858, "learning_rate": 8.68768717387998e-07, "loss": 0.5935, "step": 1995 }, { "epoch": 1.6629143214509068, "grad_norm": 0.13060715794563293, "learning_rate": 8.678781718172622e-07, "loss": 0.6509, "step": 1996 }, { "epoch": 1.663748175943298, "grad_norm": 0.13264665007591248, "learning_rate": 8.669877328774145e-07, "loss": 0.6026, "step": 1997 }, { "epoch": 1.6645820304356889, "grad_norm": 0.12658947706222534, "learning_rate": 8.66097401287097e-07, "loss": 0.5882, "step": 1998 }, { "epoch": 1.66541588492808, "grad_norm": 0.12481638789176941, "learning_rate": 8.652071777648645e-07, "loss": 0.5412, "step": 1999 }, { "epoch": 1.6662497394204712, "grad_norm": 0.13407550752162933, "learning_rate": 8.643170630291859e-07, "loss": 0.5861, "step": 2000 }, { "epoch": 1.6662497394204712, "eval_loss": 0.6118956804275513, "eval_runtime": 321.3552, "eval_samples_per_second": 16.751, "eval_steps_per_second": 2.794, "step": 2000 }, { "epoch": 1.6670835939128623, "grad_norm": 0.13629695773124695, "learning_rate": 8.63427057798441e-07, "loss": 0.6098, "step": 2001 }, { "epoch": 1.6679174484052532, "grad_norm": 0.12663327157497406, "learning_rate": 8.625371627909215e-07, "loss": 0.6582, "step": 2002 }, { "epoch": 1.6687513028976444, "grad_norm": 0.12094560265541077, "learning_rate": 8.616473787248312e-07, "loss": 0.6911, "step": 2003 }, { "epoch": 1.6695851573900353, "grad_norm": 0.13008569180965424, "learning_rate": 8.607577063182833e-07, "loss": 0.6286, "step": 2004 }, { "epoch": 1.6704190118824265, "grad_norm": 0.12234228849411011, "learning_rate": 8.598681462893011e-07, "loss": 0.5862, "step": 2005 }, { "epoch": 1.6712528663748176, "grad_norm": 0.12451769411563873, "learning_rate": 8.589786993558172e-07, "loss": 0.5655, "step": 2006 }, { "epoch": 1.6720867208672088, "grad_norm": 0.12555864453315735, "learning_rate": 8.58089366235673e-07, "loss": 0.537, "step": 2007 }, { "epoch": 1.6729205753595997, "grad_norm": 0.12302037328481674, "learning_rate": 8.572001476466182e-07, "loss": 0.6041, "step": 2008 }, { "epoch": 1.6737544298519909, "grad_norm": 0.13534630835056305, "learning_rate": 8.563110443063098e-07, "loss": 0.6523, "step": 2009 }, { "epoch": 1.6745882843443818, "grad_norm": 0.13176940381526947, "learning_rate": 8.554220569323117e-07, "loss": 0.6085, "step": 2010 }, { "epoch": 1.675422138836773, "grad_norm": 0.12548188865184784, "learning_rate": 8.545331862420943e-07, "loss": 0.6374, "step": 2011 }, { "epoch": 1.676255993329164, "grad_norm": 0.12622122466564178, "learning_rate": 8.536444329530346e-07, "loss": 0.6144, "step": 2012 }, { "epoch": 1.6770898478215552, "grad_norm": 0.12633056938648224, "learning_rate": 8.527557977824136e-07, "loss": 0.5992, "step": 2013 }, { "epoch": 1.6779237023139462, "grad_norm": 0.1307809054851532, "learning_rate": 8.518672814474174e-07, "loss": 0.6072, "step": 2014 }, { "epoch": 1.6787575568063373, "grad_norm": 0.1285247951745987, "learning_rate": 8.509788846651369e-07, "loss": 0.5737, "step": 2015 }, { "epoch": 1.6795914112987282, "grad_norm": 0.12834513187408447, "learning_rate": 8.500906081525656e-07, "loss": 0.6045, "step": 2016 }, { "epoch": 1.6804252657911194, "grad_norm": 0.12550753355026245, "learning_rate": 8.492024526265999e-07, "loss": 0.576, "step": 2017 }, { "epoch": 1.6812591202835105, "grad_norm": 0.1294783502817154, "learning_rate": 8.483144188040399e-07, "loss": 0.6527, "step": 2018 }, { "epoch": 1.6820929747759017, "grad_norm": 0.12203014642000198, "learning_rate": 8.474265074015856e-07, "loss": 0.5592, "step": 2019 }, { "epoch": 1.6829268292682928, "grad_norm": 0.12280543148517609, "learning_rate": 8.465387191358398e-07, "loss": 0.5638, "step": 2020 }, { "epoch": 1.6837606837606838, "grad_norm": 0.13221971690654755, "learning_rate": 8.456510547233048e-07, "loss": 0.636, "step": 2021 }, { "epoch": 1.6845945382530747, "grad_norm": 0.13103988766670227, "learning_rate": 8.447635148803835e-07, "loss": 0.5729, "step": 2022 }, { "epoch": 1.6854283927454659, "grad_norm": 0.13310641050338745, "learning_rate": 8.438761003233783e-07, "loss": 0.6808, "step": 2023 }, { "epoch": 1.686262247237857, "grad_norm": 0.13218644261360168, "learning_rate": 8.429888117684902e-07, "loss": 0.5781, "step": 2024 }, { "epoch": 1.6870961017302482, "grad_norm": 0.1348038613796234, "learning_rate": 8.421016499318183e-07, "loss": 0.6018, "step": 2025 }, { "epoch": 1.6879299562226393, "grad_norm": 0.12556737661361694, "learning_rate": 8.412146155293605e-07, "loss": 0.6126, "step": 2026 }, { "epoch": 1.6887638107150302, "grad_norm": 0.12600624561309814, "learning_rate": 8.403277092770104e-07, "loss": 0.6138, "step": 2027 }, { "epoch": 1.6895976652074212, "grad_norm": 0.1401616632938385, "learning_rate": 8.394409318905592e-07, "loss": 0.5923, "step": 2028 }, { "epoch": 1.6904315196998123, "grad_norm": 0.12791836261749268, "learning_rate": 8.385542840856939e-07, "loss": 0.5626, "step": 2029 }, { "epoch": 1.6912653741922035, "grad_norm": 0.12238767743110657, "learning_rate": 8.376677665779967e-07, "loss": 0.5742, "step": 2030 }, { "epoch": 1.6920992286845946, "grad_norm": 0.12230707705020905, "learning_rate": 8.367813800829442e-07, "loss": 0.6064, "step": 2031 }, { "epoch": 1.6929330831769858, "grad_norm": 0.12692691385746002, "learning_rate": 8.358951253159089e-07, "loss": 0.6256, "step": 2032 }, { "epoch": 1.6937669376693767, "grad_norm": 0.12948474287986755, "learning_rate": 8.350090029921551e-07, "loss": 0.6033, "step": 2033 }, { "epoch": 1.6946007921617676, "grad_norm": 0.12658190727233887, "learning_rate": 8.341230138268415e-07, "loss": 0.5697, "step": 2034 }, { "epoch": 1.6954346466541588, "grad_norm": 0.12975440919399261, "learning_rate": 8.332371585350185e-07, "loss": 0.6117, "step": 2035 }, { "epoch": 1.69626850114655, "grad_norm": 0.11741238087415695, "learning_rate": 8.323514378316288e-07, "loss": 0.5446, "step": 2036 }, { "epoch": 1.697102355638941, "grad_norm": 0.12554077804088593, "learning_rate": 8.314658524315068e-07, "loss": 0.5718, "step": 2037 }, { "epoch": 1.6979362101313322, "grad_norm": 0.12683430314064026, "learning_rate": 8.30580403049377e-07, "loss": 0.6252, "step": 2038 }, { "epoch": 1.6987700646237232, "grad_norm": 0.12512685358524323, "learning_rate": 8.296950903998549e-07, "loss": 0.6054, "step": 2039 }, { "epoch": 1.699603919116114, "grad_norm": 0.13208059966564178, "learning_rate": 8.288099151974446e-07, "loss": 0.6566, "step": 2040 }, { "epoch": 1.7004377736085052, "grad_norm": 0.1311858892440796, "learning_rate": 8.279248781565407e-07, "loss": 0.6073, "step": 2041 }, { "epoch": 1.7012716281008964, "grad_norm": 0.12754926085472107, "learning_rate": 8.27039979991425e-07, "loss": 0.5843, "step": 2042 }, { "epoch": 1.7021054825932875, "grad_norm": 0.12932710349559784, "learning_rate": 8.261552214162678e-07, "loss": 0.5742, "step": 2043 }, { "epoch": 1.7029393370856787, "grad_norm": 0.12164090573787689, "learning_rate": 8.252706031451269e-07, "loss": 0.6054, "step": 2044 }, { "epoch": 1.7037731915780696, "grad_norm": 0.1326964795589447, "learning_rate": 8.243861258919465e-07, "loss": 0.5884, "step": 2045 }, { "epoch": 1.7046070460704605, "grad_norm": 0.12942658364772797, "learning_rate": 8.235017903705568e-07, "loss": 0.5781, "step": 2046 }, { "epoch": 1.7054409005628517, "grad_norm": 0.12789803743362427, "learning_rate": 8.226175972946746e-07, "loss": 0.6674, "step": 2047 }, { "epoch": 1.7062747550552428, "grad_norm": 0.13167151808738708, "learning_rate": 8.217335473779007e-07, "loss": 0.6454, "step": 2048 }, { "epoch": 1.707108609547634, "grad_norm": 0.13229158520698547, "learning_rate": 8.208496413337209e-07, "loss": 0.6279, "step": 2049 }, { "epoch": 1.7079424640400251, "grad_norm": 0.12817689776420593, "learning_rate": 8.199658798755047e-07, "loss": 0.6782, "step": 2050 }, { "epoch": 1.708776318532416, "grad_norm": 0.13753139972686768, "learning_rate": 8.190822637165047e-07, "loss": 0.6016, "step": 2051 }, { "epoch": 1.7096101730248072, "grad_norm": 0.12560203671455383, "learning_rate": 8.181987935698568e-07, "loss": 0.6165, "step": 2052 }, { "epoch": 1.7104440275171982, "grad_norm": 0.13002927601337433, "learning_rate": 8.173154701485786e-07, "loss": 0.6121, "step": 2053 }, { "epoch": 1.7112778820095893, "grad_norm": 0.13231289386749268, "learning_rate": 8.164322941655692e-07, "loss": 0.6832, "step": 2054 }, { "epoch": 1.7121117365019805, "grad_norm": 0.12360356748104095, "learning_rate": 8.155492663336093e-07, "loss": 0.5825, "step": 2055 }, { "epoch": 1.7129455909943716, "grad_norm": 0.12095890194177628, "learning_rate": 8.146663873653596e-07, "loss": 0.6056, "step": 2056 }, { "epoch": 1.7137794454867625, "grad_norm": 0.1325632631778717, "learning_rate": 8.137836579733604e-07, "loss": 0.6321, "step": 2057 }, { "epoch": 1.7146132999791537, "grad_norm": 0.1254180669784546, "learning_rate": 8.129010788700317e-07, "loss": 0.5731, "step": 2058 }, { "epoch": 1.7154471544715446, "grad_norm": 0.13764861226081848, "learning_rate": 8.120186507676724e-07, "loss": 0.5629, "step": 2059 }, { "epoch": 1.7162810089639358, "grad_norm": 0.13155192136764526, "learning_rate": 8.111363743784586e-07, "loss": 0.6458, "step": 2060 }, { "epoch": 1.717114863456327, "grad_norm": 0.13540032505989075, "learning_rate": 8.102542504144454e-07, "loss": 0.585, "step": 2061 }, { "epoch": 1.717948717948718, "grad_norm": 0.1272411197423935, "learning_rate": 8.093722795875635e-07, "loss": 0.5677, "step": 2062 }, { "epoch": 1.718782572441109, "grad_norm": 0.13317130506038666, "learning_rate": 8.084904626096209e-07, "loss": 0.6681, "step": 2063 }, { "epoch": 1.7196164269335001, "grad_norm": 0.12464749813079834, "learning_rate": 8.076088001923012e-07, "loss": 0.6021, "step": 2064 }, { "epoch": 1.720450281425891, "grad_norm": 0.12794774770736694, "learning_rate": 8.06727293047163e-07, "loss": 0.6102, "step": 2065 }, { "epoch": 1.7212841359182822, "grad_norm": 0.11895739287137985, "learning_rate": 8.058459418856393e-07, "loss": 0.538, "step": 2066 }, { "epoch": 1.7221179904106734, "grad_norm": 0.12413477897644043, "learning_rate": 8.049647474190383e-07, "loss": 0.5926, "step": 2067 }, { "epoch": 1.7229518449030645, "grad_norm": 0.1245928704738617, "learning_rate": 8.04083710358541e-07, "loss": 0.5464, "step": 2068 }, { "epoch": 1.7237856993954555, "grad_norm": 0.12484681606292725, "learning_rate": 8.032028314152012e-07, "loss": 0.5695, "step": 2069 }, { "epoch": 1.7246195538878466, "grad_norm": 0.12726366519927979, "learning_rate": 8.023221112999458e-07, "loss": 0.5966, "step": 2070 }, { "epoch": 1.7254534083802375, "grad_norm": 0.12166636437177658, "learning_rate": 8.014415507235727e-07, "loss": 0.6285, "step": 2071 }, { "epoch": 1.7262872628726287, "grad_norm": 0.1254647970199585, "learning_rate": 8.005611503967515e-07, "loss": 0.6124, "step": 2072 }, { "epoch": 1.7271211173650198, "grad_norm": 0.12607796490192413, "learning_rate": 7.996809110300226e-07, "loss": 0.5675, "step": 2073 }, { "epoch": 1.727954971857411, "grad_norm": 0.13107630610466003, "learning_rate": 7.988008333337961e-07, "loss": 0.6774, "step": 2074 }, { "epoch": 1.728788826349802, "grad_norm": 0.12005145847797394, "learning_rate": 7.979209180183514e-07, "loss": 0.6335, "step": 2075 }, { "epoch": 1.729622680842193, "grad_norm": 0.11902257800102234, "learning_rate": 7.970411657938381e-07, "loss": 0.5516, "step": 2076 }, { "epoch": 1.730456535334584, "grad_norm": 0.11752620339393616, "learning_rate": 7.961615773702727e-07, "loss": 0.5586, "step": 2077 }, { "epoch": 1.7312903898269751, "grad_norm": 0.12146768718957901, "learning_rate": 7.952821534575405e-07, "loss": 0.6039, "step": 2078 }, { "epoch": 1.7321242443193663, "grad_norm": 0.1350381076335907, "learning_rate": 7.944028947653936e-07, "loss": 0.5445, "step": 2079 }, { "epoch": 1.7329580988117574, "grad_norm": 0.11824685335159302, "learning_rate": 7.935238020034503e-07, "loss": 0.5534, "step": 2080 }, { "epoch": 1.7337919533041486, "grad_norm": 0.12622307240962982, "learning_rate": 7.926448758811963e-07, "loss": 0.5945, "step": 2081 }, { "epoch": 1.7346258077965395, "grad_norm": 0.12936830520629883, "learning_rate": 7.917661171079816e-07, "loss": 0.5754, "step": 2082 }, { "epoch": 1.7354596622889304, "grad_norm": 0.12486603856086731, "learning_rate": 7.908875263930214e-07, "loss": 0.6099, "step": 2083 }, { "epoch": 1.7362935167813216, "grad_norm": 0.13117629289627075, "learning_rate": 7.900091044453956e-07, "loss": 0.6207, "step": 2084 }, { "epoch": 1.7371273712737128, "grad_norm": 0.1298520714044571, "learning_rate": 7.891308519740479e-07, "loss": 0.5873, "step": 2085 }, { "epoch": 1.737961225766104, "grad_norm": 0.12560497224330902, "learning_rate": 7.882527696877846e-07, "loss": 0.5623, "step": 2086 }, { "epoch": 1.738795080258495, "grad_norm": 0.13026194274425507, "learning_rate": 7.873748582952753e-07, "loss": 0.6185, "step": 2087 }, { "epoch": 1.739628934750886, "grad_norm": 0.13659796118736267, "learning_rate": 7.864971185050515e-07, "loss": 0.6017, "step": 2088 }, { "epoch": 1.740462789243277, "grad_norm": 0.13491465151309967, "learning_rate": 7.856195510255058e-07, "loss": 0.5749, "step": 2089 }, { "epoch": 1.741296643735668, "grad_norm": 0.12798424065113068, "learning_rate": 7.847421565648928e-07, "loss": 0.6356, "step": 2090 }, { "epoch": 1.7421304982280592, "grad_norm": 0.12282812595367432, "learning_rate": 7.838649358313262e-07, "loss": 0.5922, "step": 2091 }, { "epoch": 1.7429643527204504, "grad_norm": 0.1255982518196106, "learning_rate": 7.829878895327804e-07, "loss": 0.645, "step": 2092 }, { "epoch": 1.7437982072128415, "grad_norm": 0.1324257254600525, "learning_rate": 7.821110183770884e-07, "loss": 0.5499, "step": 2093 }, { "epoch": 1.7446320617052324, "grad_norm": 0.12864653766155243, "learning_rate": 7.812343230719422e-07, "loss": 0.5602, "step": 2094 }, { "epoch": 1.7454659161976234, "grad_norm": 0.13943764567375183, "learning_rate": 7.803578043248918e-07, "loss": 0.6665, "step": 2095 }, { "epoch": 1.7462997706900145, "grad_norm": 0.12609517574310303, "learning_rate": 7.794814628433448e-07, "loss": 0.6295, "step": 2096 }, { "epoch": 1.7471336251824057, "grad_norm": 0.12663273513317108, "learning_rate": 7.786052993345655e-07, "loss": 0.5585, "step": 2097 }, { "epoch": 1.7479674796747968, "grad_norm": 0.13500264286994934, "learning_rate": 7.777293145056745e-07, "loss": 0.6277, "step": 2098 }, { "epoch": 1.748801334167188, "grad_norm": 0.12323655933141708, "learning_rate": 7.76853509063649e-07, "loss": 0.5386, "step": 2099 }, { "epoch": 1.749635188659579, "grad_norm": 0.13133323192596436, "learning_rate": 7.759778837153204e-07, "loss": 0.5523, "step": 2100 }, { "epoch": 1.7504690431519698, "grad_norm": 0.12879957258701324, "learning_rate": 7.751024391673749e-07, "loss": 0.6167, "step": 2101 }, { "epoch": 1.751302897644361, "grad_norm": 0.13135939836502075, "learning_rate": 7.742271761263536e-07, "loss": 0.6001, "step": 2102 }, { "epoch": 1.7521367521367521, "grad_norm": 0.13290925323963165, "learning_rate": 7.733520952986505e-07, "loss": 0.5734, "step": 2103 }, { "epoch": 1.7529706066291433, "grad_norm": 0.1278628557920456, "learning_rate": 7.724771973905119e-07, "loss": 0.6017, "step": 2104 }, { "epoch": 1.7538044611215344, "grad_norm": 0.12137078493833542, "learning_rate": 7.716024831080383e-07, "loss": 0.5904, "step": 2105 }, { "epoch": 1.7546383156139254, "grad_norm": 0.12834006547927856, "learning_rate": 7.707279531571799e-07, "loss": 0.5976, "step": 2106 }, { "epoch": 1.7554721701063163, "grad_norm": 0.12812873721122742, "learning_rate": 7.698536082437399e-07, "loss": 0.577, "step": 2107 }, { "epoch": 1.7563060245987074, "grad_norm": 0.12574169039726257, "learning_rate": 7.68979449073371e-07, "loss": 0.6299, "step": 2108 }, { "epoch": 1.7571398790910986, "grad_norm": 0.14230380952358246, "learning_rate": 7.681054763515759e-07, "loss": 0.5561, "step": 2109 }, { "epoch": 1.7579737335834897, "grad_norm": 0.1346198320388794, "learning_rate": 7.672316907837084e-07, "loss": 0.6105, "step": 2110 }, { "epoch": 1.758807588075881, "grad_norm": 0.12239620089530945, "learning_rate": 7.663580930749692e-07, "loss": 0.5807, "step": 2111 }, { "epoch": 1.7596414425682718, "grad_norm": 0.13581721484661102, "learning_rate": 7.654846839304086e-07, "loss": 0.6454, "step": 2112 }, { "epoch": 1.760475297060663, "grad_norm": 0.13057507574558258, "learning_rate": 7.646114640549246e-07, "loss": 0.5808, "step": 2113 }, { "epoch": 1.761309151553054, "grad_norm": 0.1373625248670578, "learning_rate": 7.637384341532622e-07, "loss": 0.6481, "step": 2114 }, { "epoch": 1.762143006045445, "grad_norm": 0.14396755397319794, "learning_rate": 7.628655949300132e-07, "loss": 0.5838, "step": 2115 }, { "epoch": 1.7629768605378362, "grad_norm": 0.13005192577838898, "learning_rate": 7.619929470896154e-07, "loss": 0.5892, "step": 2116 }, { "epoch": 1.7638107150302273, "grad_norm": 0.1325499266386032, "learning_rate": 7.611204913363522e-07, "loss": 0.6032, "step": 2117 }, { "epoch": 1.7646445695226183, "grad_norm": 0.12955497205257416, "learning_rate": 7.602482283743523e-07, "loss": 0.5703, "step": 2118 }, { "epoch": 1.7654784240150094, "grad_norm": 0.11970066279172897, "learning_rate": 7.593761589075879e-07, "loss": 0.5528, "step": 2119 }, { "epoch": 1.7663122785074004, "grad_norm": 0.12704381346702576, "learning_rate": 7.585042836398762e-07, "loss": 0.5799, "step": 2120 }, { "epoch": 1.7671461329997915, "grad_norm": 0.1244860365986824, "learning_rate": 7.576326032748771e-07, "loss": 0.5843, "step": 2121 }, { "epoch": 1.7679799874921827, "grad_norm": 0.13251663744449615, "learning_rate": 7.567611185160931e-07, "loss": 0.5768, "step": 2122 }, { "epoch": 1.7688138419845738, "grad_norm": 0.12903398275375366, "learning_rate": 7.55889830066869e-07, "loss": 0.56, "step": 2123 }, { "epoch": 1.7696476964769647, "grad_norm": 0.13495059311389923, "learning_rate": 7.550187386303908e-07, "loss": 0.6376, "step": 2124 }, { "epoch": 1.7704815509693559, "grad_norm": 0.1355675309896469, "learning_rate": 7.541478449096861e-07, "loss": 0.6375, "step": 2125 }, { "epoch": 1.7713154054617468, "grad_norm": 0.12867653369903564, "learning_rate": 7.532771496076226e-07, "loss": 0.5449, "step": 2126 }, { "epoch": 1.772149259954138, "grad_norm": 0.13269716501235962, "learning_rate": 7.524066534269078e-07, "loss": 0.6504, "step": 2127 }, { "epoch": 1.7729831144465291, "grad_norm": 0.12922774255275726, "learning_rate": 7.515363570700889e-07, "loss": 0.5892, "step": 2128 }, { "epoch": 1.7738169689389203, "grad_norm": 0.1222226545214653, "learning_rate": 7.506662612395513e-07, "loss": 0.5652, "step": 2129 }, { "epoch": 1.7746508234313112, "grad_norm": 0.1284438967704773, "learning_rate": 7.497963666375185e-07, "loss": 0.5996, "step": 2130 }, { "epoch": 1.7754846779237023, "grad_norm": 0.1273675560951233, "learning_rate": 7.489266739660523e-07, "loss": 0.6324, "step": 2131 }, { "epoch": 1.7763185324160933, "grad_norm": 0.13269834220409393, "learning_rate": 7.480571839270512e-07, "loss": 0.6446, "step": 2132 }, { "epoch": 1.7771523869084844, "grad_norm": 0.12966395914554596, "learning_rate": 7.471878972222494e-07, "loss": 0.5247, "step": 2133 }, { "epoch": 1.7779862414008756, "grad_norm": 0.12394700199365616, "learning_rate": 7.463188145532185e-07, "loss": 0.6179, "step": 2134 }, { "epoch": 1.7788200958932667, "grad_norm": 0.13124272227287292, "learning_rate": 7.454499366213642e-07, "loss": 0.5627, "step": 2135 }, { "epoch": 1.7796539503856577, "grad_norm": 0.13032935559749603, "learning_rate": 7.445812641279274e-07, "loss": 0.5894, "step": 2136 }, { "epoch": 1.7804878048780488, "grad_norm": 0.13481009006500244, "learning_rate": 7.437127977739835e-07, "loss": 0.6706, "step": 2137 }, { "epoch": 1.7813216593704397, "grad_norm": 0.1252714842557907, "learning_rate": 7.428445382604405e-07, "loss": 0.6167, "step": 2138 }, { "epoch": 1.7821555138628309, "grad_norm": 0.13593217730522156, "learning_rate": 7.419764862880406e-07, "loss": 0.6212, "step": 2139 }, { "epoch": 1.782989368355222, "grad_norm": 0.128921777009964, "learning_rate": 7.411086425573583e-07, "loss": 0.5644, "step": 2140 }, { "epoch": 1.7838232228476132, "grad_norm": 0.1279333531856537, "learning_rate": 7.402410077687993e-07, "loss": 0.6109, "step": 2141 }, { "epoch": 1.7846570773400041, "grad_norm": 0.1240469217300415, "learning_rate": 7.393735826226013e-07, "loss": 0.6266, "step": 2142 }, { "epoch": 1.7854909318323953, "grad_norm": 0.12809887528419495, "learning_rate": 7.38506367818833e-07, "loss": 0.6296, "step": 2143 }, { "epoch": 1.7863247863247862, "grad_norm": 0.12371251732110977, "learning_rate": 7.376393640573927e-07, "loss": 0.5797, "step": 2144 }, { "epoch": 1.7871586408171773, "grad_norm": 0.13013000786304474, "learning_rate": 7.367725720380086e-07, "loss": 0.5641, "step": 2145 }, { "epoch": 1.7879924953095685, "grad_norm": 0.12911874055862427, "learning_rate": 7.359059924602385e-07, "loss": 0.6069, "step": 2146 }, { "epoch": 1.7888263498019596, "grad_norm": 0.12248188257217407, "learning_rate": 7.35039626023468e-07, "loss": 0.5962, "step": 2147 }, { "epoch": 1.7896602042943508, "grad_norm": 0.1306978166103363, "learning_rate": 7.34173473426911e-07, "loss": 0.6135, "step": 2148 }, { "epoch": 1.7904940587867417, "grad_norm": 0.11989414691925049, "learning_rate": 7.333075353696089e-07, "loss": 0.5457, "step": 2149 }, { "epoch": 1.7913279132791327, "grad_norm": 0.1320810168981552, "learning_rate": 7.324418125504304e-07, "loss": 0.616, "step": 2150 }, { "epoch": 1.7921617677715238, "grad_norm": 0.1229398101568222, "learning_rate": 7.315763056680694e-07, "loss": 0.5835, "step": 2151 }, { "epoch": 1.792995622263915, "grad_norm": 0.12493004649877548, "learning_rate": 7.307110154210463e-07, "loss": 0.5619, "step": 2152 }, { "epoch": 1.793829476756306, "grad_norm": 0.12893696129322052, "learning_rate": 7.298459425077063e-07, "loss": 0.603, "step": 2153 }, { "epoch": 1.7946633312486973, "grad_norm": 0.1281341016292572, "learning_rate": 7.289810876262199e-07, "loss": 0.56, "step": 2154 }, { "epoch": 1.7954971857410882, "grad_norm": 0.1304667741060257, "learning_rate": 7.281164514745805e-07, "loss": 0.539, "step": 2155 }, { "epoch": 1.7963310402334791, "grad_norm": 0.12886284291744232, "learning_rate": 7.272520347506058e-07, "loss": 0.6319, "step": 2156 }, { "epoch": 1.7971648947258703, "grad_norm": 0.12648077309131622, "learning_rate": 7.263878381519364e-07, "loss": 0.5886, "step": 2157 }, { "epoch": 1.7979987492182614, "grad_norm": 0.12707141041755676, "learning_rate": 7.255238623760347e-07, "loss": 0.5575, "step": 2158 }, { "epoch": 1.7988326037106526, "grad_norm": 0.13646213710308075, "learning_rate": 7.24660108120185e-07, "loss": 0.5984, "step": 2159 }, { "epoch": 1.7996664582030437, "grad_norm": 0.1271483451128006, "learning_rate": 7.237965760814939e-07, "loss": 0.5451, "step": 2160 }, { "epoch": 1.8005003126954346, "grad_norm": 0.1413080096244812, "learning_rate": 7.22933266956887e-07, "loss": 0.609, "step": 2161 }, { "epoch": 1.8013341671878256, "grad_norm": 0.1268233358860016, "learning_rate": 7.220701814431109e-07, "loss": 0.5813, "step": 2162 }, { "epoch": 1.8021680216802167, "grad_norm": 0.13267411291599274, "learning_rate": 7.212073202367319e-07, "loss": 0.6314, "step": 2163 }, { "epoch": 1.8030018761726079, "grad_norm": 0.12609542906284332, "learning_rate": 7.203446840341347e-07, "loss": 0.5809, "step": 2164 }, { "epoch": 1.803835730664999, "grad_norm": 0.12579114735126495, "learning_rate": 7.194822735315229e-07, "loss": 0.6272, "step": 2165 }, { "epoch": 1.8046695851573902, "grad_norm": 0.1351184844970703, "learning_rate": 7.186200894249177e-07, "loss": 0.5507, "step": 2166 }, { "epoch": 1.805503439649781, "grad_norm": 0.13242338597774506, "learning_rate": 7.177581324101575e-07, "loss": 0.604, "step": 2167 }, { "epoch": 1.806337294142172, "grad_norm": 0.13314031064510345, "learning_rate": 7.168964031828976e-07, "loss": 0.6186, "step": 2168 }, { "epoch": 1.8071711486345632, "grad_norm": 0.12832415103912354, "learning_rate": 7.160349024386095e-07, "loss": 0.5621, "step": 2169 }, { "epoch": 1.8080050031269543, "grad_norm": 0.13182200491428375, "learning_rate": 7.151736308725801e-07, "loss": 0.6296, "step": 2170 }, { "epoch": 1.8088388576193455, "grad_norm": 0.13076068460941315, "learning_rate": 7.143125891799111e-07, "loss": 0.5908, "step": 2171 }, { "epoch": 1.8096727121117366, "grad_norm": 0.1297273486852646, "learning_rate": 7.134517780555197e-07, "loss": 0.5424, "step": 2172 }, { "epoch": 1.8105065666041276, "grad_norm": 0.13636283576488495, "learning_rate": 7.125911981941359e-07, "loss": 0.6602, "step": 2173 }, { "epoch": 1.8113404210965185, "grad_norm": 0.13110437989234924, "learning_rate": 7.117308502903035e-07, "loss": 0.5969, "step": 2174 }, { "epoch": 1.8121742755889096, "grad_norm": 0.1326896697282791, "learning_rate": 7.108707350383791e-07, "loss": 0.6131, "step": 2175 }, { "epoch": 1.8130081300813008, "grad_norm": 0.13135559856891632, "learning_rate": 7.100108531325317e-07, "loss": 0.5867, "step": 2176 }, { "epoch": 1.813841984573692, "grad_norm": 0.12924109399318695, "learning_rate": 7.091512052667412e-07, "loss": 0.6186, "step": 2177 }, { "epoch": 1.814675839066083, "grad_norm": 0.12946565449237823, "learning_rate": 7.082917921348001e-07, "loss": 0.6381, "step": 2178 }, { "epoch": 1.815509693558474, "grad_norm": 0.12695160508155823, "learning_rate": 7.0743261443031e-07, "loss": 0.5888, "step": 2179 }, { "epoch": 1.8163435480508652, "grad_norm": 0.12655752897262573, "learning_rate": 7.065736728466832e-07, "loss": 0.6633, "step": 2180 }, { "epoch": 1.817177402543256, "grad_norm": 0.13403771817684174, "learning_rate": 7.057149680771413e-07, "loss": 0.608, "step": 2181 }, { "epoch": 1.8180112570356473, "grad_norm": 0.12965700030326843, "learning_rate": 7.048565008147146e-07, "loss": 0.6786, "step": 2182 }, { "epoch": 1.8188451115280384, "grad_norm": 0.1377706527709961, "learning_rate": 7.039982717522422e-07, "loss": 0.6328, "step": 2183 }, { "epoch": 1.8196789660204296, "grad_norm": 0.14265958964824677, "learning_rate": 7.031402815823704e-07, "loss": 0.6493, "step": 2184 }, { "epoch": 1.8205128205128205, "grad_norm": 0.13361769914627075, "learning_rate": 7.022825309975529e-07, "loss": 0.618, "step": 2185 }, { "epoch": 1.8213466750052116, "grad_norm": 0.12783940136432648, "learning_rate": 7.014250206900502e-07, "loss": 0.6005, "step": 2186 }, { "epoch": 1.8221805294976026, "grad_norm": 0.12289536744356155, "learning_rate": 7.005677513519288e-07, "loss": 0.577, "step": 2187 }, { "epoch": 1.8230143839899937, "grad_norm": 0.12831023335456848, "learning_rate": 6.997107236750604e-07, "loss": 0.5951, "step": 2188 }, { "epoch": 1.8238482384823849, "grad_norm": 0.12774531543254852, "learning_rate": 6.988539383511222e-07, "loss": 0.6544, "step": 2189 }, { "epoch": 1.824682092974776, "grad_norm": 0.13053590059280396, "learning_rate": 6.979973960715957e-07, "loss": 0.5647, "step": 2190 }, { "epoch": 1.825515947467167, "grad_norm": 0.12190835922956467, "learning_rate": 6.971410975277654e-07, "loss": 0.6, "step": 2191 }, { "epoch": 1.826349801959558, "grad_norm": 0.13813957571983337, "learning_rate": 6.962850434107206e-07, "loss": 0.6417, "step": 2192 }, { "epoch": 1.827183656451949, "grad_norm": 0.13274197280406952, "learning_rate": 6.954292344113521e-07, "loss": 0.6578, "step": 2193 }, { "epoch": 1.8280175109443402, "grad_norm": 0.12455936521291733, "learning_rate": 6.945736712203536e-07, "loss": 0.5857, "step": 2194 }, { "epoch": 1.8288513654367313, "grad_norm": 0.13513398170471191, "learning_rate": 6.937183545282198e-07, "loss": 0.5971, "step": 2195 }, { "epoch": 1.8296852199291225, "grad_norm": 0.1251787394285202, "learning_rate": 6.928632850252467e-07, "loss": 0.6098, "step": 2196 }, { "epoch": 1.8305190744215134, "grad_norm": 0.12823201715946198, "learning_rate": 6.920084634015313e-07, "loss": 0.6184, "step": 2197 }, { "epoch": 1.8313529289139046, "grad_norm": 0.12458322197198868, "learning_rate": 6.911538903469697e-07, "loss": 0.5409, "step": 2198 }, { "epoch": 1.8321867834062955, "grad_norm": 0.13494035601615906, "learning_rate": 6.902995665512581e-07, "loss": 0.5969, "step": 2199 }, { "epoch": 1.8330206378986866, "grad_norm": 0.1326414942741394, "learning_rate": 6.894454927038906e-07, "loss": 0.5327, "step": 2200 }, { "epoch": 1.8338544923910778, "grad_norm": 0.12999312579631805, "learning_rate": 6.885916694941612e-07, "loss": 0.6224, "step": 2201 }, { "epoch": 1.834688346883469, "grad_norm": 0.12325473874807358, "learning_rate": 6.8773809761116e-07, "loss": 0.5234, "step": 2202 }, { "epoch": 1.8355222013758599, "grad_norm": 0.12891866266727448, "learning_rate": 6.868847777437747e-07, "loss": 0.5782, "step": 2203 }, { "epoch": 1.836356055868251, "grad_norm": 0.1282792091369629, "learning_rate": 6.860317105806905e-07, "loss": 0.5817, "step": 2204 }, { "epoch": 1.837189910360642, "grad_norm": 0.12752637267112732, "learning_rate": 6.851788968103876e-07, "loss": 0.6053, "step": 2205 }, { "epoch": 1.838023764853033, "grad_norm": 0.1264016330242157, "learning_rate": 6.843263371211414e-07, "loss": 0.5788, "step": 2206 }, { "epoch": 1.8388576193454242, "grad_norm": 0.12685662508010864, "learning_rate": 6.83474032201024e-07, "loss": 0.5849, "step": 2207 }, { "epoch": 1.8396914738378154, "grad_norm": 0.1329505294561386, "learning_rate": 6.826219827379002e-07, "loss": 0.5886, "step": 2208 }, { "epoch": 1.8405253283302065, "grad_norm": 0.13913948833942413, "learning_rate": 6.817701894194293e-07, "loss": 0.6189, "step": 2209 }, { "epoch": 1.8413591828225975, "grad_norm": 0.12883232533931732, "learning_rate": 6.809186529330639e-07, "loss": 0.5634, "step": 2210 }, { "epoch": 1.8421930373149884, "grad_norm": 0.12113146483898163, "learning_rate": 6.800673739660488e-07, "loss": 0.597, "step": 2211 }, { "epoch": 1.8430268918073796, "grad_norm": 0.1325560361146927, "learning_rate": 6.792163532054217e-07, "loss": 0.551, "step": 2212 }, { "epoch": 1.8438607462997707, "grad_norm": 0.14157071709632874, "learning_rate": 6.783655913380115e-07, "loss": 0.563, "step": 2213 }, { "epoch": 1.8446946007921619, "grad_norm": 0.12442418932914734, "learning_rate": 6.775150890504381e-07, "loss": 0.5486, "step": 2214 }, { "epoch": 1.845528455284553, "grad_norm": 0.1348833441734314, "learning_rate": 6.766648470291123e-07, "loss": 0.5977, "step": 2215 }, { "epoch": 1.846362309776944, "grad_norm": 0.12434475868940353, "learning_rate": 6.758148659602344e-07, "loss": 0.5013, "step": 2216 }, { "epoch": 1.8471961642693349, "grad_norm": 0.12582679092884064, "learning_rate": 6.749651465297943e-07, "loss": 0.5939, "step": 2217 }, { "epoch": 1.848030018761726, "grad_norm": 0.12978383898735046, "learning_rate": 6.741156894235705e-07, "loss": 0.6158, "step": 2218 }, { "epoch": 1.8488638732541172, "grad_norm": 0.1360258311033249, "learning_rate": 6.732664953271305e-07, "loss": 0.6059, "step": 2219 }, { "epoch": 1.8496977277465083, "grad_norm": 0.13842739164829254, "learning_rate": 6.724175649258286e-07, "loss": 0.5801, "step": 2220 }, { "epoch": 1.8505315822388995, "grad_norm": 0.12137693911790848, "learning_rate": 6.715688989048065e-07, "loss": 0.5945, "step": 2221 }, { "epoch": 1.8513654367312904, "grad_norm": 0.1248222142457962, "learning_rate": 6.707204979489933e-07, "loss": 0.5963, "step": 2222 }, { "epoch": 1.8521992912236813, "grad_norm": 0.12999458611011505, "learning_rate": 6.698723627431037e-07, "loss": 0.6183, "step": 2223 }, { "epoch": 1.8530331457160725, "grad_norm": 0.13486354053020477, "learning_rate": 6.690244939716375e-07, "loss": 0.6077, "step": 2224 }, { "epoch": 1.8538670002084636, "grad_norm": 0.12799599766731262, "learning_rate": 6.681768923188798e-07, "loss": 0.6049, "step": 2225 }, { "epoch": 1.8547008547008548, "grad_norm": 0.13147741556167603, "learning_rate": 6.673295584689003e-07, "loss": 0.6467, "step": 2226 }, { "epoch": 1.855534709193246, "grad_norm": 0.13251982629299164, "learning_rate": 6.664824931055522e-07, "loss": 0.6342, "step": 2227 }, { "epoch": 1.8563685636856369, "grad_norm": 0.13429106771945953, "learning_rate": 6.656356969124723e-07, "loss": 0.6237, "step": 2228 }, { "epoch": 1.8572024181780278, "grad_norm": 0.1291757971048355, "learning_rate": 6.647891705730801e-07, "loss": 0.5665, "step": 2229 }, { "epoch": 1.858036272670419, "grad_norm": 0.13031582534313202, "learning_rate": 6.639429147705772e-07, "loss": 0.5835, "step": 2230 }, { "epoch": 1.85887012716281, "grad_norm": 0.13219709694385529, "learning_rate": 6.630969301879473e-07, "loss": 0.5901, "step": 2231 }, { "epoch": 1.8597039816552012, "grad_norm": 0.13087095320224762, "learning_rate": 6.622512175079542e-07, "loss": 0.6978, "step": 2232 }, { "epoch": 1.8605378361475924, "grad_norm": 0.1454305499792099, "learning_rate": 6.614057774131436e-07, "loss": 0.5978, "step": 2233 }, { "epoch": 1.8613716906399833, "grad_norm": 0.12761370837688446, "learning_rate": 6.605606105858401e-07, "loss": 0.5659, "step": 2234 }, { "epoch": 1.8622055451323742, "grad_norm": 0.13098569214344025, "learning_rate": 6.597157177081477e-07, "loss": 0.5773, "step": 2235 }, { "epoch": 1.8630393996247654, "grad_norm": 0.1315004825592041, "learning_rate": 6.588710994619508e-07, "loss": 0.5531, "step": 2236 }, { "epoch": 1.8638732541171565, "grad_norm": 0.1302976906299591, "learning_rate": 6.580267565289106e-07, "loss": 0.5603, "step": 2237 }, { "epoch": 1.8647071086095477, "grad_norm": 0.12400949001312256, "learning_rate": 6.571826895904666e-07, "loss": 0.5228, "step": 2238 }, { "epoch": 1.8655409631019388, "grad_norm": 0.13697467744350433, "learning_rate": 6.563388993278359e-07, "loss": 0.6525, "step": 2239 }, { "epoch": 1.8663748175943298, "grad_norm": 0.12461115419864655, "learning_rate": 6.554953864220114e-07, "loss": 0.6154, "step": 2240 }, { "epoch": 1.867208672086721, "grad_norm": 0.13096342980861664, "learning_rate": 6.546521515537635e-07, "loss": 0.612, "step": 2241 }, { "epoch": 1.8680425265791119, "grad_norm": 0.12971916794776917, "learning_rate": 6.538091954036369e-07, "loss": 0.6228, "step": 2242 }, { "epoch": 1.868876381071503, "grad_norm": 0.13440938293933868, "learning_rate": 6.529665186519519e-07, "loss": 0.6226, "step": 2243 }, { "epoch": 1.8697102355638942, "grad_norm": 0.13049529492855072, "learning_rate": 6.521241219788037e-07, "loss": 0.6174, "step": 2244 }, { "epoch": 1.8705440900562853, "grad_norm": 0.12830443680286407, "learning_rate": 6.512820060640607e-07, "loss": 0.5727, "step": 2245 }, { "epoch": 1.8713779445486762, "grad_norm": 0.1377989947795868, "learning_rate": 6.504401715873653e-07, "loss": 0.6148, "step": 2246 }, { "epoch": 1.8722117990410674, "grad_norm": 0.12611956894397736, "learning_rate": 6.495986192281324e-07, "loss": 0.5785, "step": 2247 }, { "epoch": 1.8730456535334583, "grad_norm": 0.13040220737457275, "learning_rate": 6.487573496655495e-07, "loss": 0.6567, "step": 2248 }, { "epoch": 1.8738795080258495, "grad_norm": 0.13693012297153473, "learning_rate": 6.479163635785758e-07, "loss": 0.5877, "step": 2249 }, { "epoch": 1.8747133625182406, "grad_norm": 0.13410882651805878, "learning_rate": 6.47075661645941e-07, "loss": 0.6076, "step": 2250 }, { "epoch": 1.8755472170106318, "grad_norm": 0.12973110377788544, "learning_rate": 6.462352445461467e-07, "loss": 0.6246, "step": 2251 }, { "epoch": 1.8763810715030227, "grad_norm": 0.12795211374759674, "learning_rate": 6.453951129574643e-07, "loss": 0.6205, "step": 2252 }, { "epoch": 1.8772149259954138, "grad_norm": 0.1331990361213684, "learning_rate": 6.44555267557934e-07, "loss": 0.6227, "step": 2253 }, { "epoch": 1.8780487804878048, "grad_norm": 0.13448093831539154, "learning_rate": 6.437157090253658e-07, "loss": 0.5703, "step": 2254 }, { "epoch": 1.878882634980196, "grad_norm": 0.12563006579875946, "learning_rate": 6.428764380373375e-07, "loss": 0.5943, "step": 2255 }, { "epoch": 1.879716489472587, "grad_norm": 0.13857470452785492, "learning_rate": 6.420374552711958e-07, "loss": 0.6263, "step": 2256 }, { "epoch": 1.8805503439649782, "grad_norm": 0.13378775119781494, "learning_rate": 6.411987614040539e-07, "loss": 0.6188, "step": 2257 }, { "epoch": 1.8813841984573692, "grad_norm": 0.13999195396900177, "learning_rate": 6.40360357112792e-07, "loss": 0.6461, "step": 2258 }, { "epoch": 1.8822180529497603, "grad_norm": 0.13252532482147217, "learning_rate": 6.395222430740572e-07, "loss": 0.6247, "step": 2259 }, { "epoch": 1.8830519074421512, "grad_norm": 0.13208715617656708, "learning_rate": 6.386844199642618e-07, "loss": 0.5912, "step": 2260 }, { "epoch": 1.8838857619345424, "grad_norm": 0.12673957645893097, "learning_rate": 6.378468884595829e-07, "loss": 0.529, "step": 2261 }, { "epoch": 1.8847196164269335, "grad_norm": 0.1289004683494568, "learning_rate": 6.370096492359635e-07, "loss": 0.5575, "step": 2262 }, { "epoch": 1.8855534709193247, "grad_norm": 0.1387024074792862, "learning_rate": 6.361727029691097e-07, "loss": 0.6183, "step": 2263 }, { "epoch": 1.8863873254117156, "grad_norm": 0.13052384555339813, "learning_rate": 6.353360503344909e-07, "loss": 0.6069, "step": 2264 }, { "epoch": 1.8872211799041068, "grad_norm": 0.13136428594589233, "learning_rate": 6.344996920073409e-07, "loss": 0.6139, "step": 2265 }, { "epoch": 1.8880550343964977, "grad_norm": 0.12757977843284607, "learning_rate": 6.336636286626547e-07, "loss": 0.6091, "step": 2266 }, { "epoch": 1.8888888888888888, "grad_norm": 0.13187870383262634, "learning_rate": 6.328278609751898e-07, "loss": 0.6339, "step": 2267 }, { "epoch": 1.88972274338128, "grad_norm": 0.13330577313899994, "learning_rate": 6.319923896194651e-07, "loss": 0.5872, "step": 2268 }, { "epoch": 1.8905565978736711, "grad_norm": 0.13062001764774323, "learning_rate": 6.311572152697598e-07, "loss": 0.618, "step": 2269 }, { "epoch": 1.891390452366062, "grad_norm": 0.1248871386051178, "learning_rate": 6.30322338600114e-07, "loss": 0.576, "step": 2270 }, { "epoch": 1.8922243068584532, "grad_norm": 0.1286626160144806, "learning_rate": 6.294877602843275e-07, "loss": 0.5739, "step": 2271 }, { "epoch": 1.8930581613508441, "grad_norm": 0.137839674949646, "learning_rate": 6.286534809959587e-07, "loss": 0.5684, "step": 2272 }, { "epoch": 1.8938920158432353, "grad_norm": 0.13219650089740753, "learning_rate": 6.278195014083256e-07, "loss": 0.6107, "step": 2273 }, { "epoch": 1.8947258703356264, "grad_norm": 0.13113045692443848, "learning_rate": 6.269858221945037e-07, "loss": 0.5889, "step": 2274 }, { "epoch": 1.8955597248280176, "grad_norm": 0.13249342143535614, "learning_rate": 6.261524440273262e-07, "loss": 0.5744, "step": 2275 }, { "epoch": 1.8963935793204088, "grad_norm": 0.12959535419940948, "learning_rate": 6.253193675793831e-07, "loss": 0.597, "step": 2276 }, { "epoch": 1.8972274338127997, "grad_norm": 0.12774960696697235, "learning_rate": 6.244865935230214e-07, "loss": 0.5293, "step": 2277 }, { "epoch": 1.8980612883051906, "grad_norm": 0.13584597408771515, "learning_rate": 6.23654122530344e-07, "loss": 0.6379, "step": 2278 }, { "epoch": 1.8988951427975818, "grad_norm": 0.13189005851745605, "learning_rate": 6.228219552732083e-07, "loss": 0.5781, "step": 2279 }, { "epoch": 1.899728997289973, "grad_norm": 0.12410996854305267, "learning_rate": 6.219900924232278e-07, "loss": 0.5808, "step": 2280 }, { "epoch": 1.900562851782364, "grad_norm": 0.12360142171382904, "learning_rate": 6.211585346517701e-07, "loss": 0.6029, "step": 2281 }, { "epoch": 1.9013967062747552, "grad_norm": 0.12962640821933746, "learning_rate": 6.203272826299559e-07, "loss": 0.6324, "step": 2282 }, { "epoch": 1.9022305607671461, "grad_norm": 0.1311754435300827, "learning_rate": 6.194963370286595e-07, "loss": 0.5782, "step": 2283 }, { "epoch": 1.903064415259537, "grad_norm": 0.13473135232925415, "learning_rate": 6.186656985185077e-07, "loss": 0.6378, "step": 2284 }, { "epoch": 1.9038982697519282, "grad_norm": 0.1313280314207077, "learning_rate": 6.178353677698801e-07, "loss": 0.6647, "step": 2285 }, { "epoch": 1.9047321242443194, "grad_norm": 0.13595938682556152, "learning_rate": 6.170053454529073e-07, "loss": 0.5646, "step": 2286 }, { "epoch": 1.9055659787367105, "grad_norm": 0.13668006658554077, "learning_rate": 6.16175632237471e-07, "loss": 0.6181, "step": 2287 }, { "epoch": 1.9063998332291017, "grad_norm": 0.13875062763690948, "learning_rate": 6.15346228793204e-07, "loss": 0.5855, "step": 2288 }, { "epoch": 1.9072336877214926, "grad_norm": 0.12911437451839447, "learning_rate": 6.145171357894885e-07, "loss": 0.6139, "step": 2289 }, { "epoch": 1.9080675422138835, "grad_norm": 0.13118845224380493, "learning_rate": 6.136883538954561e-07, "loss": 0.5864, "step": 2290 }, { "epoch": 1.9089013967062747, "grad_norm": 0.12767499685287476, "learning_rate": 6.128598837799879e-07, "loss": 0.6385, "step": 2291 }, { "epoch": 1.9097352511986658, "grad_norm": 0.1284511238336563, "learning_rate": 6.12031726111713e-07, "loss": 0.5868, "step": 2292 }, { "epoch": 1.910569105691057, "grad_norm": 0.13237819075584412, "learning_rate": 6.112038815590079e-07, "loss": 0.5604, "step": 2293 }, { "epoch": 1.9114029601834481, "grad_norm": 0.13073354959487915, "learning_rate": 6.103763507899977e-07, "loss": 0.621, "step": 2294 }, { "epoch": 1.912236814675839, "grad_norm": 0.1346769481897354, "learning_rate": 6.095491344725527e-07, "loss": 0.5636, "step": 2295 }, { "epoch": 1.91307066916823, "grad_norm": 0.13018903136253357, "learning_rate": 6.087222332742905e-07, "loss": 0.582, "step": 2296 }, { "epoch": 1.9139045236606211, "grad_norm": 0.13262416422367096, "learning_rate": 6.078956478625742e-07, "loss": 0.6244, "step": 2297 }, { "epoch": 1.9147383781530123, "grad_norm": 0.1358201801776886, "learning_rate": 6.070693789045113e-07, "loss": 0.6072, "step": 2298 }, { "epoch": 1.9155722326454034, "grad_norm": 0.12717005610466003, "learning_rate": 6.06243427066955e-07, "loss": 0.6003, "step": 2299 }, { "epoch": 1.9164060871377946, "grad_norm": 0.13791069388389587, "learning_rate": 6.054177930165016e-07, "loss": 0.6324, "step": 2300 }, { "epoch": 1.9172399416301855, "grad_norm": 0.12626376748085022, "learning_rate": 6.045924774194916e-07, "loss": 0.5243, "step": 2301 }, { "epoch": 1.9180737961225764, "grad_norm": 0.12443426251411438, "learning_rate": 6.037674809420083e-07, "loss": 0.5733, "step": 2302 }, { "epoch": 1.9189076506149676, "grad_norm": 0.13789716362953186, "learning_rate": 6.029428042498772e-07, "loss": 0.6085, "step": 2303 }, { "epoch": 1.9197415051073587, "grad_norm": 0.12876053154468536, "learning_rate": 6.02118448008666e-07, "loss": 0.5843, "step": 2304 }, { "epoch": 1.92057535959975, "grad_norm": 0.1320531666278839, "learning_rate": 6.012944128836835e-07, "loss": 0.6165, "step": 2305 }, { "epoch": 1.921409214092141, "grad_norm": 0.12669098377227783, "learning_rate": 6.004706995399797e-07, "loss": 0.5911, "step": 2306 }, { "epoch": 1.922243068584532, "grad_norm": 0.12740486860275269, "learning_rate": 5.996473086423449e-07, "loss": 0.5759, "step": 2307 }, { "epoch": 1.9230769230769231, "grad_norm": 0.1266666054725647, "learning_rate": 5.988242408553083e-07, "loss": 0.6185, "step": 2308 }, { "epoch": 1.923910777569314, "grad_norm": 0.14473924040794373, "learning_rate": 5.980014968431395e-07, "loss": 0.5639, "step": 2309 }, { "epoch": 1.9247446320617052, "grad_norm": 0.13156810402870178, "learning_rate": 5.971790772698466e-07, "loss": 0.5898, "step": 2310 }, { "epoch": 1.9255784865540964, "grad_norm": 0.13030414283275604, "learning_rate": 5.963569827991751e-07, "loss": 0.5725, "step": 2311 }, { "epoch": 1.9264123410464875, "grad_norm": 0.1348404586315155, "learning_rate": 5.955352140946088e-07, "loss": 0.5914, "step": 2312 }, { "epoch": 1.9272461955388784, "grad_norm": 0.12505322694778442, "learning_rate": 5.94713771819368e-07, "loss": 0.5284, "step": 2313 }, { "epoch": 1.9280800500312696, "grad_norm": 0.1333184838294983, "learning_rate": 5.938926566364105e-07, "loss": 0.6533, "step": 2314 }, { "epoch": 1.9289139045236605, "grad_norm": 0.14823047816753387, "learning_rate": 5.930718692084289e-07, "loss": 0.6039, "step": 2315 }, { "epoch": 1.9297477590160517, "grad_norm": 0.1254442185163498, "learning_rate": 5.922514101978522e-07, "loss": 0.5922, "step": 2316 }, { "epoch": 1.9305816135084428, "grad_norm": 0.13495124876499176, "learning_rate": 5.914312802668444e-07, "loss": 0.6377, "step": 2317 }, { "epoch": 1.931415468000834, "grad_norm": 0.12776747345924377, "learning_rate": 5.90611480077303e-07, "loss": 0.5803, "step": 2318 }, { "epoch": 1.932249322493225, "grad_norm": 0.12187840789556503, "learning_rate": 5.897920102908602e-07, "loss": 0.6001, "step": 2319 }, { "epoch": 1.933083176985616, "grad_norm": 0.13361410796642303, "learning_rate": 5.889728715688814e-07, "loss": 0.6101, "step": 2320 }, { "epoch": 1.933917031478007, "grad_norm": 0.13355158269405365, "learning_rate": 5.881540645724645e-07, "loss": 0.5752, "step": 2321 }, { "epoch": 1.9347508859703981, "grad_norm": 0.12796255946159363, "learning_rate": 5.8733558996244e-07, "loss": 0.5554, "step": 2322 }, { "epoch": 1.9355847404627893, "grad_norm": 0.132767915725708, "learning_rate": 5.865174483993696e-07, "loss": 0.5639, "step": 2323 }, { "epoch": 1.9364185949551804, "grad_norm": 0.1333564668893814, "learning_rate": 5.856996405435471e-07, "loss": 0.6437, "step": 2324 }, { "epoch": 1.9372524494475714, "grad_norm": 0.13327006995677948, "learning_rate": 5.848821670549967e-07, "loss": 0.6275, "step": 2325 }, { "epoch": 1.9380863039399625, "grad_norm": 0.13549090921878815, "learning_rate": 5.840650285934722e-07, "loss": 0.5768, "step": 2326 }, { "epoch": 1.9389201584323534, "grad_norm": 0.12455825507640839, "learning_rate": 5.832482258184574e-07, "loss": 0.5617, "step": 2327 }, { "epoch": 1.9397540129247446, "grad_norm": 0.1259452998638153, "learning_rate": 5.824317593891651e-07, "loss": 0.5639, "step": 2328 }, { "epoch": 1.9405878674171357, "grad_norm": 0.12992453575134277, "learning_rate": 5.816156299645362e-07, "loss": 0.6418, "step": 2329 }, { "epoch": 1.9414217219095269, "grad_norm": 0.12744390964508057, "learning_rate": 5.807998382032413e-07, "loss": 0.5801, "step": 2330 }, { "epoch": 1.9422555764019178, "grad_norm": 0.13386395573616028, "learning_rate": 5.799843847636765e-07, "loss": 0.6192, "step": 2331 }, { "epoch": 1.943089430894309, "grad_norm": 0.13299784064292908, "learning_rate": 5.791692703039656e-07, "loss": 0.5963, "step": 2332 }, { "epoch": 1.9439232853867, "grad_norm": 0.12975671887397766, "learning_rate": 5.783544954819591e-07, "loss": 0.6515, "step": 2333 }, { "epoch": 1.944757139879091, "grad_norm": 0.13418138027191162, "learning_rate": 5.775400609552331e-07, "loss": 0.5566, "step": 2334 }, { "epoch": 1.9455909943714822, "grad_norm": 0.13558447360992432, "learning_rate": 5.767259673810889e-07, "loss": 0.6067, "step": 2335 }, { "epoch": 1.9464248488638733, "grad_norm": 0.12238030880689621, "learning_rate": 5.759122154165528e-07, "loss": 0.5724, "step": 2336 }, { "epoch": 1.9472587033562645, "grad_norm": 0.14128273725509644, "learning_rate": 5.750988057183754e-07, "loss": 0.6078, "step": 2337 }, { "epoch": 1.9480925578486554, "grad_norm": 0.14713209867477417, "learning_rate": 5.742857389430319e-07, "loss": 0.6125, "step": 2338 }, { "epoch": 1.9489264123410464, "grad_norm": 0.1362263411283493, "learning_rate": 5.734730157467194e-07, "loss": 0.5711, "step": 2339 }, { "epoch": 1.9497602668334375, "grad_norm": 0.13019274175167084, "learning_rate": 5.72660636785358e-07, "loss": 0.5702, "step": 2340 }, { "epoch": 1.9505941213258287, "grad_norm": 0.14090077579021454, "learning_rate": 5.718486027145905e-07, "loss": 0.6286, "step": 2341 }, { "epoch": 1.9514279758182198, "grad_norm": 0.13111229240894318, "learning_rate": 5.710369141897813e-07, "loss": 0.6162, "step": 2342 }, { "epoch": 1.952261830310611, "grad_norm": 0.14520929753780365, "learning_rate": 5.702255718660148e-07, "loss": 0.6235, "step": 2343 }, { "epoch": 1.9530956848030019, "grad_norm": 0.12398312240839005, "learning_rate": 5.69414576398098e-07, "loss": 0.5896, "step": 2344 }, { "epoch": 1.9539295392953928, "grad_norm": 0.14264680445194244, "learning_rate": 5.686039284405563e-07, "loss": 0.5586, "step": 2345 }, { "epoch": 1.954763393787784, "grad_norm": 0.13142624497413635, "learning_rate": 5.677936286476356e-07, "loss": 0.5624, "step": 2346 }, { "epoch": 1.9555972482801751, "grad_norm": 0.13193339109420776, "learning_rate": 5.669836776732999e-07, "loss": 0.6235, "step": 2347 }, { "epoch": 1.9564311027725663, "grad_norm": 0.12483777850866318, "learning_rate": 5.661740761712324e-07, "loss": 0.6564, "step": 2348 }, { "epoch": 1.9572649572649574, "grad_norm": 0.13298740983009338, "learning_rate": 5.653648247948341e-07, "loss": 0.6364, "step": 2349 }, { "epoch": 1.9580988117573483, "grad_norm": 0.12461144477128983, "learning_rate": 5.64555924197223e-07, "loss": 0.5612, "step": 2350 }, { "epoch": 1.9589326662497393, "grad_norm": 0.12801668047904968, "learning_rate": 5.637473750312349e-07, "loss": 0.5265, "step": 2351 }, { "epoch": 1.9597665207421304, "grad_norm": 0.13269507884979248, "learning_rate": 5.62939177949421e-07, "loss": 0.5956, "step": 2352 }, { "epoch": 1.9606003752345216, "grad_norm": 0.14033639430999756, "learning_rate": 5.621313336040494e-07, "loss": 0.6257, "step": 2353 }, { "epoch": 1.9614342297269127, "grad_norm": 0.12311799824237823, "learning_rate": 5.613238426471024e-07, "loss": 0.5558, "step": 2354 }, { "epoch": 1.9622680842193039, "grad_norm": 0.12262886762619019, "learning_rate": 5.605167057302777e-07, "loss": 0.5878, "step": 2355 }, { "epoch": 1.9631019387116948, "grad_norm": 0.12075518816709518, "learning_rate": 5.597099235049874e-07, "loss": 0.5562, "step": 2356 }, { "epoch": 1.9639357932040857, "grad_norm": 0.12985512614250183, "learning_rate": 5.589034966223566e-07, "loss": 0.6132, "step": 2357 }, { "epoch": 1.9647696476964769, "grad_norm": 0.1268652081489563, "learning_rate": 5.580974257332242e-07, "loss": 0.6154, "step": 2358 }, { "epoch": 1.965603502188868, "grad_norm": 0.13775032758712769, "learning_rate": 5.572917114881422e-07, "loss": 0.6034, "step": 2359 }, { "epoch": 1.9664373566812592, "grad_norm": 0.13750946521759033, "learning_rate": 5.564863545373739e-07, "loss": 0.5833, "step": 2360 }, { "epoch": 1.9672712111736503, "grad_norm": 0.12842507660388947, "learning_rate": 5.556813555308948e-07, "loss": 0.5999, "step": 2361 }, { "epoch": 1.9681050656660413, "grad_norm": 0.13427001237869263, "learning_rate": 5.548767151183912e-07, "loss": 0.6044, "step": 2362 }, { "epoch": 1.9689389201584322, "grad_norm": 0.13650521636009216, "learning_rate": 5.540724339492602e-07, "loss": 0.6156, "step": 2363 }, { "epoch": 1.9697727746508233, "grad_norm": 0.13917742669582367, "learning_rate": 5.532685126726086e-07, "loss": 0.5875, "step": 2364 }, { "epoch": 1.9706066291432145, "grad_norm": 0.1281631588935852, "learning_rate": 5.524649519372532e-07, "loss": 0.5927, "step": 2365 }, { "epoch": 1.9714404836356056, "grad_norm": 0.13355673849582672, "learning_rate": 5.516617523917195e-07, "loss": 0.5991, "step": 2366 }, { "epoch": 1.9722743381279968, "grad_norm": 0.12803775072097778, "learning_rate": 5.508589146842423e-07, "loss": 0.6001, "step": 2367 }, { "epoch": 1.9731081926203877, "grad_norm": 0.1371046006679535, "learning_rate": 5.500564394627634e-07, "loss": 0.5983, "step": 2368 }, { "epoch": 1.9739420471127789, "grad_norm": 0.1317107230424881, "learning_rate": 5.492543273749321e-07, "loss": 0.6025, "step": 2369 }, { "epoch": 1.9747759016051698, "grad_norm": 0.13639812171459198, "learning_rate": 5.484525790681052e-07, "loss": 0.598, "step": 2370 }, { "epoch": 1.975609756097561, "grad_norm": 0.13002654910087585, "learning_rate": 5.476511951893454e-07, "loss": 0.602, "step": 2371 }, { "epoch": 1.976443610589952, "grad_norm": 0.12995946407318115, "learning_rate": 5.468501763854214e-07, "loss": 0.575, "step": 2372 }, { "epoch": 1.9772774650823433, "grad_norm": 0.13288018107414246, "learning_rate": 5.460495233028073e-07, "loss": 0.6301, "step": 2373 }, { "epoch": 1.9781113195747342, "grad_norm": 0.13422086834907532, "learning_rate": 5.452492365876827e-07, "loss": 0.6567, "step": 2374 }, { "epoch": 1.9789451740671253, "grad_norm": 0.13234974443912506, "learning_rate": 5.444493168859303e-07, "loss": 0.5892, "step": 2375 }, { "epoch": 1.9797790285595163, "grad_norm": 0.13048966228961945, "learning_rate": 5.436497648431375e-07, "loss": 0.6009, "step": 2376 }, { "epoch": 1.9806128830519074, "grad_norm": 0.124395951628685, "learning_rate": 5.428505811045947e-07, "loss": 0.5759, "step": 2377 }, { "epoch": 1.9814467375442986, "grad_norm": 0.12869031727313995, "learning_rate": 5.420517663152948e-07, "loss": 0.6303, "step": 2378 }, { "epoch": 1.9822805920366897, "grad_norm": 0.13104695081710815, "learning_rate": 5.412533211199329e-07, "loss": 0.6003, "step": 2379 }, { "epoch": 1.9831144465290806, "grad_norm": 0.13080964982509613, "learning_rate": 5.404552461629068e-07, "loss": 0.6617, "step": 2380 }, { "epoch": 1.9839483010214718, "grad_norm": 0.12720125913619995, "learning_rate": 5.39657542088314e-07, "loss": 0.6138, "step": 2381 }, { "epoch": 1.9847821555138627, "grad_norm": 0.12719741463661194, "learning_rate": 5.388602095399544e-07, "loss": 0.6034, "step": 2382 }, { "epoch": 1.9856160100062539, "grad_norm": 0.13190889358520508, "learning_rate": 5.380632491613264e-07, "loss": 0.6329, "step": 2383 }, { "epoch": 1.986449864498645, "grad_norm": 0.13070254027843475, "learning_rate": 5.372666615956289e-07, "loss": 0.5807, "step": 2384 }, { "epoch": 1.9872837189910362, "grad_norm": 0.12874957919120789, "learning_rate": 5.3647044748576e-07, "loss": 0.5942, "step": 2385 }, { "epoch": 1.988117573483427, "grad_norm": 0.13363546133041382, "learning_rate": 5.356746074743157e-07, "loss": 0.5778, "step": 2386 }, { "epoch": 1.9889514279758183, "grad_norm": 0.1323724091053009, "learning_rate": 5.348791422035903e-07, "loss": 0.6128, "step": 2387 }, { "epoch": 1.9897852824682092, "grad_norm": 0.1363963931798935, "learning_rate": 5.340840523155768e-07, "loss": 0.6101, "step": 2388 }, { "epoch": 1.9906191369606003, "grad_norm": 0.13670435547828674, "learning_rate": 5.332893384519639e-07, "loss": 0.6146, "step": 2389 }, { "epoch": 1.9914529914529915, "grad_norm": 0.12916621565818787, "learning_rate": 5.324950012541372e-07, "loss": 0.6372, "step": 2390 }, { "epoch": 1.9922868459453826, "grad_norm": 0.12862586975097656, "learning_rate": 5.317010413631782e-07, "loss": 0.5652, "step": 2391 }, { "epoch": 1.9931207004377736, "grad_norm": 0.1284669190645218, "learning_rate": 5.309074594198644e-07, "loss": 0.6468, "step": 2392 }, { "epoch": 1.9939545549301647, "grad_norm": 0.12397977709770203, "learning_rate": 5.301142560646676e-07, "loss": 0.603, "step": 2393 }, { "epoch": 1.9947884094225556, "grad_norm": 0.13170576095581055, "learning_rate": 5.293214319377542e-07, "loss": 0.63, "step": 2394 }, { "epoch": 1.9956222639149468, "grad_norm": 0.11900755017995834, "learning_rate": 5.285289876789849e-07, "loss": 0.5428, "step": 2395 }, { "epoch": 1.996456118407338, "grad_norm": 0.13693659007549286, "learning_rate": 5.277369239279143e-07, "loss": 0.6014, "step": 2396 }, { "epoch": 1.997289972899729, "grad_norm": 0.14242641627788544, "learning_rate": 5.269452413237885e-07, "loss": 0.6007, "step": 2397 }, { "epoch": 1.99812382739212, "grad_norm": 0.12357503920793533, "learning_rate": 5.261539405055471e-07, "loss": 0.5221, "step": 2398 }, { "epoch": 1.9989576818845112, "grad_norm": 0.128871351480484, "learning_rate": 5.253630221118209e-07, "loss": 0.6091, "step": 2399 }, { "epoch": 1.999791536376902, "grad_norm": 0.14917448163032532, "learning_rate": 5.245724867809325e-07, "loss": 0.6194, "step": 2400 }, { "epoch": 1.999791536376902, "eval_loss": 0.6083878874778748, "eval_runtime": 321.3809, "eval_samples_per_second": 16.75, "eval_steps_per_second": 2.794, "step": 2400 }, { "epoch": 2.0, "grad_norm": 0.14917448163032532, "learning_rate": 5.237823351508952e-07, "loss": 0.1756, "step": 2401 }, { "epoch": 2.0006253908692933, "grad_norm": 0.13851699233055115, "learning_rate": 5.229925678594123e-07, "loss": 0.4831, "step": 2402 }, { "epoch": 2.000833854492391, "grad_norm": 0.12272315472364426, "learning_rate": 5.22203185543878e-07, "loss": 0.6342, "step": 2403 }, { "epoch": 2.0016677089847823, "grad_norm": 0.13500462472438812, "learning_rate": 5.214141888413745e-07, "loss": 0.5933, "step": 2404 }, { "epoch": 2.0025015634771735, "grad_norm": 0.1287509948015213, "learning_rate": 5.206255783886735e-07, "loss": 0.5951, "step": 2405 }, { "epoch": 2.003335417969564, "grad_norm": 0.13235405087471008, "learning_rate": 5.198373548222346e-07, "loss": 0.5602, "step": 2406 }, { "epoch": 2.0041692724619553, "grad_norm": 0.13505418598651886, "learning_rate": 5.190495187782058e-07, "loss": 0.5749, "step": 2407 }, { "epoch": 2.0050031269543465, "grad_norm": 0.12087678164243698, "learning_rate": 5.18262070892421e-07, "loss": 0.5794, "step": 2408 }, { "epoch": 2.0058369814467376, "grad_norm": 0.1379173845052719, "learning_rate": 5.174750118004028e-07, "loss": 0.5605, "step": 2409 }, { "epoch": 2.0066708359391288, "grad_norm": 0.12731070816516876, "learning_rate": 5.166883421373583e-07, "loss": 0.5652, "step": 2410 }, { "epoch": 2.00750469043152, "grad_norm": 0.12088848650455475, "learning_rate": 5.159020625381814e-07, "loss": 0.5716, "step": 2411 }, { "epoch": 2.0083385449239106, "grad_norm": 0.133113831281662, "learning_rate": 5.151161736374505e-07, "loss": 0.5916, "step": 2412 }, { "epoch": 2.0091723994163018, "grad_norm": 0.13565190136432648, "learning_rate": 5.143306760694291e-07, "loss": 0.5265, "step": 2413 }, { "epoch": 2.010006253908693, "grad_norm": 0.12744170427322388, "learning_rate": 5.135455704680644e-07, "loss": 0.5823, "step": 2414 }, { "epoch": 2.010840108401084, "grad_norm": 0.12263254076242447, "learning_rate": 5.12760857466988e-07, "loss": 0.5971, "step": 2415 }, { "epoch": 2.011673962893475, "grad_norm": 0.13080917298793793, "learning_rate": 5.119765376995133e-07, "loss": 0.5702, "step": 2416 }, { "epoch": 2.0125078173858664, "grad_norm": 0.13577453792095184, "learning_rate": 5.111926117986383e-07, "loss": 0.5877, "step": 2417 }, { "epoch": 2.013341671878257, "grad_norm": 0.130285382270813, "learning_rate": 5.104090803970417e-07, "loss": 0.5499, "step": 2418 }, { "epoch": 2.0141755263706482, "grad_norm": 0.13663120567798615, "learning_rate": 5.096259441270841e-07, "loss": 0.6155, "step": 2419 }, { "epoch": 2.0150093808630394, "grad_norm": 0.1265571266412735, "learning_rate": 5.088432036208074e-07, "loss": 0.6034, "step": 2420 }, { "epoch": 2.0158432353554305, "grad_norm": 0.13786400854587555, "learning_rate": 5.080608595099338e-07, "loss": 0.5979, "step": 2421 }, { "epoch": 2.0166770898478217, "grad_norm": 0.13637498021125793, "learning_rate": 5.072789124258659e-07, "loss": 0.5949, "step": 2422 }, { "epoch": 2.017510944340213, "grad_norm": 0.13866861164569855, "learning_rate": 5.064973629996852e-07, "loss": 0.5988, "step": 2423 }, { "epoch": 2.0183447988326035, "grad_norm": 0.12290661782026291, "learning_rate": 5.057162118621539e-07, "loss": 0.5804, "step": 2424 }, { "epoch": 2.0191786533249947, "grad_norm": 0.12569722533226013, "learning_rate": 5.049354596437103e-07, "loss": 0.5788, "step": 2425 }, { "epoch": 2.020012507817386, "grad_norm": 0.1376030594110489, "learning_rate": 5.041551069744734e-07, "loss": 0.6145, "step": 2426 }, { "epoch": 2.020846362309777, "grad_norm": 0.1364547610282898, "learning_rate": 5.033751544842379e-07, "loss": 0.5738, "step": 2427 }, { "epoch": 2.021680216802168, "grad_norm": 0.1332581341266632, "learning_rate": 5.02595602802476e-07, "loss": 0.6183, "step": 2428 }, { "epoch": 2.0225140712945593, "grad_norm": 0.13177590072155, "learning_rate": 5.018164525583366e-07, "loss": 0.6066, "step": 2429 }, { "epoch": 2.02334792578695, "grad_norm": 0.13377752900123596, "learning_rate": 5.010377043806444e-07, "loss": 0.6234, "step": 2430 }, { "epoch": 2.024181780279341, "grad_norm": 0.13108399510383606, "learning_rate": 5.002593588978991e-07, "loss": 0.5926, "step": 2431 }, { "epoch": 2.0250156347717323, "grad_norm": 0.13490580022335052, "learning_rate": 4.994814167382773e-07, "loss": 0.6087, "step": 2432 }, { "epoch": 2.0258494892641234, "grad_norm": 0.13015596568584442, "learning_rate": 4.987038785296281e-07, "loss": 0.5672, "step": 2433 }, { "epoch": 2.0266833437565146, "grad_norm": 0.13041432201862335, "learning_rate": 4.979267448994751e-07, "loss": 0.6379, "step": 2434 }, { "epoch": 2.0275171982489057, "grad_norm": 0.12589594721794128, "learning_rate": 4.971500164750161e-07, "loss": 0.5521, "step": 2435 }, { "epoch": 2.0283510527412965, "grad_norm": 0.133503720164299, "learning_rate": 4.963736938831208e-07, "loss": 0.6315, "step": 2436 }, { "epoch": 2.0291849072336876, "grad_norm": 0.13001377880573273, "learning_rate": 4.955977777503318e-07, "loss": 0.6043, "step": 2437 }, { "epoch": 2.0300187617260788, "grad_norm": 0.1424768716096878, "learning_rate": 4.948222687028646e-07, "loss": 0.568, "step": 2438 }, { "epoch": 2.03085261621847, "grad_norm": 0.12374301999807358, "learning_rate": 4.940471673666043e-07, "loss": 0.6289, "step": 2439 }, { "epoch": 2.031686470710861, "grad_norm": 0.12739478051662445, "learning_rate": 4.932724743671088e-07, "loss": 0.6265, "step": 2440 }, { "epoch": 2.032520325203252, "grad_norm": 0.13795717060565948, "learning_rate": 4.924981903296054e-07, "loss": 0.614, "step": 2441 }, { "epoch": 2.033354179695643, "grad_norm": 0.1346176266670227, "learning_rate": 4.917243158789914e-07, "loss": 0.5637, "step": 2442 }, { "epoch": 2.034188034188034, "grad_norm": 0.14109820127487183, "learning_rate": 4.909508516398338e-07, "loss": 0.5753, "step": 2443 }, { "epoch": 2.035021888680425, "grad_norm": 0.1337806135416031, "learning_rate": 4.901777982363684e-07, "loss": 0.6249, "step": 2444 }, { "epoch": 2.0358557431728164, "grad_norm": 0.12704265117645264, "learning_rate": 4.89405156292499e-07, "loss": 0.5865, "step": 2445 }, { "epoch": 2.0366895976652075, "grad_norm": 0.1360265612602234, "learning_rate": 4.886329264317986e-07, "loss": 0.631, "step": 2446 }, { "epoch": 2.0375234521575987, "grad_norm": 0.13027885556221008, "learning_rate": 4.878611092775065e-07, "loss": 0.6369, "step": 2447 }, { "epoch": 2.0383573066499894, "grad_norm": 0.14479905366897583, "learning_rate": 4.87089705452529e-07, "loss": 0.547, "step": 2448 }, { "epoch": 2.0391911611423805, "grad_norm": 0.14251482486724854, "learning_rate": 4.863187155794393e-07, "loss": 0.612, "step": 2449 }, { "epoch": 2.0400250156347717, "grad_norm": 0.13191883265972137, "learning_rate": 4.85548140280476e-07, "loss": 0.5575, "step": 2450 }, { "epoch": 2.040858870127163, "grad_norm": 0.1324061006307602, "learning_rate": 4.847779801775436e-07, "loss": 0.6469, "step": 2451 }, { "epoch": 2.041692724619554, "grad_norm": 0.1338500678539276, "learning_rate": 4.840082358922108e-07, "loss": 0.5686, "step": 2452 }, { "epoch": 2.042526579111945, "grad_norm": 0.1276562511920929, "learning_rate": 4.832389080457118e-07, "loss": 0.6212, "step": 2453 }, { "epoch": 2.043360433604336, "grad_norm": 0.1317005157470703, "learning_rate": 4.824699972589435e-07, "loss": 0.5743, "step": 2454 }, { "epoch": 2.044194288096727, "grad_norm": 0.13533517718315125, "learning_rate": 4.817015041524676e-07, "loss": 0.6228, "step": 2455 }, { "epoch": 2.045028142589118, "grad_norm": 0.1293802559375763, "learning_rate": 4.809334293465074e-07, "loss": 0.6132, "step": 2456 }, { "epoch": 2.0458619970815093, "grad_norm": 0.13336439430713654, "learning_rate": 4.801657734609491e-07, "loss": 0.5953, "step": 2457 }, { "epoch": 2.0466958515739004, "grad_norm": 0.13487955927848816, "learning_rate": 4.793985371153408e-07, "loss": 0.6019, "step": 2458 }, { "epoch": 2.0475297060662916, "grad_norm": 0.1389443278312683, "learning_rate": 4.786317209288923e-07, "loss": 0.613, "step": 2459 }, { "epoch": 2.0483635605586823, "grad_norm": 0.1412551999092102, "learning_rate": 4.778653255204732e-07, "loss": 0.6387, "step": 2460 }, { "epoch": 2.0491974150510734, "grad_norm": 0.13776306807994843, "learning_rate": 4.770993515086153e-07, "loss": 0.5802, "step": 2461 }, { "epoch": 2.0500312695434646, "grad_norm": 0.13684414327144623, "learning_rate": 4.76333799511509e-07, "loss": 0.63, "step": 2462 }, { "epoch": 2.0508651240358557, "grad_norm": 0.12975727021694183, "learning_rate": 4.755686701470043e-07, "loss": 0.5792, "step": 2463 }, { "epoch": 2.051698978528247, "grad_norm": 0.13416776061058044, "learning_rate": 4.7480396403261037e-07, "loss": 0.6301, "step": 2464 }, { "epoch": 2.052532833020638, "grad_norm": 0.12499997764825821, "learning_rate": 4.7403968178549445e-07, "loss": 0.6074, "step": 2465 }, { "epoch": 2.053366687513029, "grad_norm": 0.1321221888065338, "learning_rate": 4.732758240224818e-07, "loss": 0.5766, "step": 2466 }, { "epoch": 2.05420054200542, "grad_norm": 0.12665514647960663, "learning_rate": 4.7251239136005584e-07, "loss": 0.5713, "step": 2467 }, { "epoch": 2.055034396497811, "grad_norm": 0.13948574662208557, "learning_rate": 4.717493844143555e-07, "loss": 0.631, "step": 2468 }, { "epoch": 2.055868250990202, "grad_norm": 0.13029757142066956, "learning_rate": 4.709868038011776e-07, "loss": 0.5668, "step": 2469 }, { "epoch": 2.0567021054825934, "grad_norm": 0.13254894316196442, "learning_rate": 4.702246501359738e-07, "loss": 0.5865, "step": 2470 }, { "epoch": 2.0575359599749845, "grad_norm": 0.13072209060192108, "learning_rate": 4.6946292403385166e-07, "loss": 0.5824, "step": 2471 }, { "epoch": 2.0583698144673757, "grad_norm": 0.13296759128570557, "learning_rate": 4.6870162610957354e-07, "loss": 0.6032, "step": 2472 }, { "epoch": 2.0592036689597664, "grad_norm": 0.12592405080795288, "learning_rate": 4.679407569775562e-07, "loss": 0.5407, "step": 2473 }, { "epoch": 2.0600375234521575, "grad_norm": 0.13011068105697632, "learning_rate": 4.6718031725187003e-07, "loss": 0.5957, "step": 2474 }, { "epoch": 2.0608713779445487, "grad_norm": 0.13407057523727417, "learning_rate": 4.6642030754623997e-07, "loss": 0.6039, "step": 2475 }, { "epoch": 2.06170523243694, "grad_norm": 0.13803914189338684, "learning_rate": 4.6566072847404283e-07, "loss": 0.5814, "step": 2476 }, { "epoch": 2.062539086929331, "grad_norm": 0.13236603140830994, "learning_rate": 4.649015806483083e-07, "loss": 0.555, "step": 2477 }, { "epoch": 2.063372941421722, "grad_norm": 0.1258341372013092, "learning_rate": 4.641428646817177e-07, "loss": 0.6074, "step": 2478 }, { "epoch": 2.064206795914113, "grad_norm": 0.12374375760555267, "learning_rate": 4.633845811866043e-07, "loss": 0.56, "step": 2479 }, { "epoch": 2.065040650406504, "grad_norm": 0.1353626847267151, "learning_rate": 4.626267307749518e-07, "loss": 0.5985, "step": 2480 }, { "epoch": 2.065874504898895, "grad_norm": 0.12993483245372772, "learning_rate": 4.618693140583946e-07, "loss": 0.6149, "step": 2481 }, { "epoch": 2.0667083593912863, "grad_norm": 0.14027145504951477, "learning_rate": 4.611123316482176e-07, "loss": 0.5618, "step": 2482 }, { "epoch": 2.0675422138836774, "grad_norm": 0.13931693136692047, "learning_rate": 4.6035578415535413e-07, "loss": 0.6719, "step": 2483 }, { "epoch": 2.0683760683760686, "grad_norm": 0.13234581053256989, "learning_rate": 4.5959967219038785e-07, "loss": 0.6464, "step": 2484 }, { "epoch": 2.0692099228684593, "grad_norm": 0.14038105309009552, "learning_rate": 4.5884399636354976e-07, "loss": 0.5507, "step": 2485 }, { "epoch": 2.0700437773608504, "grad_norm": 0.12869097292423248, "learning_rate": 4.580887572847193e-07, "loss": 0.6238, "step": 2486 }, { "epoch": 2.0708776318532416, "grad_norm": 0.13558898866176605, "learning_rate": 4.573339555634235e-07, "loss": 0.6271, "step": 2487 }, { "epoch": 2.0717114863456327, "grad_norm": 0.13607515394687653, "learning_rate": 4.565795918088361e-07, "loss": 0.6484, "step": 2488 }, { "epoch": 2.072545340838024, "grad_norm": 0.1314222514629364, "learning_rate": 4.558256666297773e-07, "loss": 0.6515, "step": 2489 }, { "epoch": 2.073379195330415, "grad_norm": 0.11763909459114075, "learning_rate": 4.5507218063471454e-07, "loss": 0.5347, "step": 2490 }, { "epoch": 2.0742130498228057, "grad_norm": 0.13402071595191956, "learning_rate": 4.543191344317594e-07, "loss": 0.6751, "step": 2491 }, { "epoch": 2.075046904315197, "grad_norm": 0.13556897640228271, "learning_rate": 4.535665286286691e-07, "loss": 0.6263, "step": 2492 }, { "epoch": 2.075880758807588, "grad_norm": 0.12935727834701538, "learning_rate": 4.5281436383284524e-07, "loss": 0.5629, "step": 2493 }, { "epoch": 2.076714613299979, "grad_norm": 0.13147656619548798, "learning_rate": 4.520626406513336e-07, "loss": 0.5864, "step": 2494 }, { "epoch": 2.0775484677923703, "grad_norm": 0.132756769657135, "learning_rate": 4.5131135969082344e-07, "loss": 0.5948, "step": 2495 }, { "epoch": 2.0783823222847615, "grad_norm": 0.13642394542694092, "learning_rate": 4.505605215576478e-07, "loss": 0.5956, "step": 2496 }, { "epoch": 2.079216176777152, "grad_norm": 0.12447264790534973, "learning_rate": 4.4981012685778096e-07, "loss": 0.5471, "step": 2497 }, { "epoch": 2.0800500312695434, "grad_norm": 0.1277848333120346, "learning_rate": 4.49060176196841e-07, "loss": 0.6088, "step": 2498 }, { "epoch": 2.0808838857619345, "grad_norm": 0.12668083608150482, "learning_rate": 4.4831067018008637e-07, "loss": 0.6189, "step": 2499 }, { "epoch": 2.0817177402543257, "grad_norm": 0.12748317420482635, "learning_rate": 4.4756160941241704e-07, "loss": 0.5424, "step": 2500 }, { "epoch": 2.082551594746717, "grad_norm": 0.1347886621952057, "learning_rate": 4.468129944983737e-07, "loss": 0.5691, "step": 2501 }, { "epoch": 2.083385449239108, "grad_norm": 0.13337112963199615, "learning_rate": 4.4606482604213713e-07, "loss": 0.5919, "step": 2502 }, { "epoch": 2.0842193037314987, "grad_norm": 0.12566891312599182, "learning_rate": 4.4531710464752736e-07, "loss": 0.5504, "step": 2503 }, { "epoch": 2.08505315822389, "grad_norm": 0.1357632875442505, "learning_rate": 4.445698309180049e-07, "loss": 0.638, "step": 2504 }, { "epoch": 2.085887012716281, "grad_norm": 0.13156214356422424, "learning_rate": 4.438230054566677e-07, "loss": 0.6427, "step": 2505 }, { "epoch": 2.086720867208672, "grad_norm": 0.13448919355869293, "learning_rate": 4.430766288662525e-07, "loss": 0.5826, "step": 2506 }, { "epoch": 2.0875547217010633, "grad_norm": 0.12653423845767975, "learning_rate": 4.423307017491336e-07, "loss": 0.6506, "step": 2507 }, { "epoch": 2.0883885761934544, "grad_norm": 0.1331438571214676, "learning_rate": 4.415852247073224e-07, "loss": 0.5862, "step": 2508 }, { "epoch": 2.089222430685845, "grad_norm": 0.13306111097335815, "learning_rate": 4.4084019834246744e-07, "loss": 0.5739, "step": 2509 }, { "epoch": 2.0900562851782363, "grad_norm": 0.1239815205335617, "learning_rate": 4.400956232558529e-07, "loss": 0.5777, "step": 2510 }, { "epoch": 2.0908901396706274, "grad_norm": 0.13276396691799164, "learning_rate": 4.3935150004839995e-07, "loss": 0.5389, "step": 2511 }, { "epoch": 2.0917239941630186, "grad_norm": 0.13424333930015564, "learning_rate": 4.3860782932066353e-07, "loss": 0.6052, "step": 2512 }, { "epoch": 2.0925578486554097, "grad_norm": 0.1292356699705124, "learning_rate": 4.3786461167283493e-07, "loss": 0.5725, "step": 2513 }, { "epoch": 2.093391703147801, "grad_norm": 0.13369356095790863, "learning_rate": 4.371218477047386e-07, "loss": 0.5628, "step": 2514 }, { "epoch": 2.0942255576401916, "grad_norm": 0.12970249354839325, "learning_rate": 4.363795380158334e-07, "loss": 0.5888, "step": 2515 }, { "epoch": 2.0950594121325827, "grad_norm": 0.13806034624576569, "learning_rate": 4.356376832052112e-07, "loss": 0.6003, "step": 2516 }, { "epoch": 2.095893266624974, "grad_norm": 0.12886318564414978, "learning_rate": 4.3489628387159703e-07, "loss": 0.6059, "step": 2517 }, { "epoch": 2.096727121117365, "grad_norm": 0.141400545835495, "learning_rate": 4.3415534061334805e-07, "loss": 0.598, "step": 2518 }, { "epoch": 2.097560975609756, "grad_norm": 0.13046756386756897, "learning_rate": 4.334148540284541e-07, "loss": 0.6003, "step": 2519 }, { "epoch": 2.0983948301021473, "grad_norm": 0.13141290843486786, "learning_rate": 4.3267482471453566e-07, "loss": 0.5769, "step": 2520 }, { "epoch": 2.099228684594538, "grad_norm": 0.1417616456747055, "learning_rate": 4.3193525326884426e-07, "loss": 0.6074, "step": 2521 }, { "epoch": 2.100062539086929, "grad_norm": 0.13840587437152863, "learning_rate": 4.311961402882622e-07, "loss": 0.5996, "step": 2522 }, { "epoch": 2.1008963935793203, "grad_norm": 0.13204775750637054, "learning_rate": 4.304574863693015e-07, "loss": 0.5569, "step": 2523 }, { "epoch": 2.1017302480717115, "grad_norm": 0.12753865122795105, "learning_rate": 4.297192921081039e-07, "loss": 0.6069, "step": 2524 }, { "epoch": 2.1025641025641026, "grad_norm": 0.13271555304527283, "learning_rate": 4.289815581004396e-07, "loss": 0.5977, "step": 2525 }, { "epoch": 2.103397957056494, "grad_norm": 0.12939058244228363, "learning_rate": 4.282442849417085e-07, "loss": 0.5633, "step": 2526 }, { "epoch": 2.1042318115488845, "grad_norm": 0.13303998112678528, "learning_rate": 4.2750747322693727e-07, "loss": 0.5381, "step": 2527 }, { "epoch": 2.1050656660412757, "grad_norm": 0.13386622071266174, "learning_rate": 4.267711235507815e-07, "loss": 0.6149, "step": 2528 }, { "epoch": 2.105899520533667, "grad_norm": 0.13128864765167236, "learning_rate": 4.260352365075226e-07, "loss": 0.5882, "step": 2529 }, { "epoch": 2.106733375026058, "grad_norm": 0.13530667126178741, "learning_rate": 4.252998126910695e-07, "loss": 0.6106, "step": 2530 }, { "epoch": 2.107567229518449, "grad_norm": 0.13622409105300903, "learning_rate": 4.2456485269495666e-07, "loss": 0.6211, "step": 2531 }, { "epoch": 2.1084010840108403, "grad_norm": 0.13454468548297882, "learning_rate": 4.238303571123446e-07, "loss": 0.6532, "step": 2532 }, { "epoch": 2.1092349385032314, "grad_norm": 0.13102950155735016, "learning_rate": 4.2309632653601847e-07, "loss": 0.6462, "step": 2533 }, { "epoch": 2.110068792995622, "grad_norm": 0.12532152235507965, "learning_rate": 4.2236276155838936e-07, "loss": 0.5418, "step": 2534 }, { "epoch": 2.1109026474880133, "grad_norm": 0.1457109898328781, "learning_rate": 4.2162966277149147e-07, "loss": 0.5753, "step": 2535 }, { "epoch": 2.1117365019804044, "grad_norm": 0.1336268037557602, "learning_rate": 4.2089703076698303e-07, "loss": 0.6101, "step": 2536 }, { "epoch": 2.1125703564727956, "grad_norm": 0.13546814024448395, "learning_rate": 4.2016486613614566e-07, "loss": 0.614, "step": 2537 }, { "epoch": 2.1134042109651867, "grad_norm": 0.13902918994426727, "learning_rate": 4.194331694698836e-07, "loss": 0.6348, "step": 2538 }, { "epoch": 2.114238065457578, "grad_norm": 0.13438214361667633, "learning_rate": 4.187019413587234e-07, "loss": 0.5976, "step": 2539 }, { "epoch": 2.1150719199499686, "grad_norm": 0.1312108039855957, "learning_rate": 4.1797118239281405e-07, "loss": 0.569, "step": 2540 }, { "epoch": 2.1159057744423597, "grad_norm": 0.13684919476509094, "learning_rate": 4.172408931619249e-07, "loss": 0.5902, "step": 2541 }, { "epoch": 2.116739628934751, "grad_norm": 0.14042967557907104, "learning_rate": 4.165110742554474e-07, "loss": 0.6172, "step": 2542 }, { "epoch": 2.117573483427142, "grad_norm": 0.12899069488048553, "learning_rate": 4.157817262623924e-07, "loss": 0.6071, "step": 2543 }, { "epoch": 2.118407337919533, "grad_norm": 0.13776318728923798, "learning_rate": 4.150528497713911e-07, "loss": 0.5816, "step": 2544 }, { "epoch": 2.1192411924119243, "grad_norm": 0.13189685344696045, "learning_rate": 4.143244453706941e-07, "loss": 0.6054, "step": 2545 }, { "epoch": 2.120075046904315, "grad_norm": 0.13243719935417175, "learning_rate": 4.1359651364817095e-07, "loss": 0.5598, "step": 2546 }, { "epoch": 2.120908901396706, "grad_norm": 0.1378655880689621, "learning_rate": 4.1286905519130957e-07, "loss": 0.6007, "step": 2547 }, { "epoch": 2.1217427558890973, "grad_norm": 0.13281330466270447, "learning_rate": 4.1214207058721674e-07, "loss": 0.6074, "step": 2548 }, { "epoch": 2.1225766103814885, "grad_norm": 0.1345830112695694, "learning_rate": 4.114155604226158e-07, "loss": 0.6277, "step": 2549 }, { "epoch": 2.1234104648738796, "grad_norm": 0.1268264204263687, "learning_rate": 4.106895252838477e-07, "loss": 0.5918, "step": 2550 }, { "epoch": 2.124244319366271, "grad_norm": 0.13760782778263092, "learning_rate": 4.0996396575687e-07, "loss": 0.6027, "step": 2551 }, { "epoch": 2.1250781738586615, "grad_norm": 0.12820297479629517, "learning_rate": 4.09238882427256e-07, "loss": 0.5955, "step": 2552 }, { "epoch": 2.1259120283510526, "grad_norm": 0.13412033021450043, "learning_rate": 4.085142758801953e-07, "loss": 0.6313, "step": 2553 }, { "epoch": 2.126745882843444, "grad_norm": 0.13249512016773224, "learning_rate": 4.0779014670049183e-07, "loss": 0.6021, "step": 2554 }, { "epoch": 2.127579737335835, "grad_norm": 0.1324591338634491, "learning_rate": 4.0706649547256566e-07, "loss": 0.5915, "step": 2555 }, { "epoch": 2.128413591828226, "grad_norm": 0.13488835096359253, "learning_rate": 4.063433227804496e-07, "loss": 0.5679, "step": 2556 }, { "epoch": 2.1292474463206172, "grad_norm": 0.12676960229873657, "learning_rate": 4.056206292077915e-07, "loss": 0.5684, "step": 2557 }, { "epoch": 2.130081300813008, "grad_norm": 0.12975813448429108, "learning_rate": 4.0489841533785164e-07, "loss": 0.6028, "step": 2558 }, { "epoch": 2.130915155305399, "grad_norm": 0.13794657588005066, "learning_rate": 4.041766817535036e-07, "loss": 0.5906, "step": 2559 }, { "epoch": 2.1317490097977903, "grad_norm": 0.13147057592868805, "learning_rate": 4.0345542903723316e-07, "loss": 0.5742, "step": 2560 }, { "epoch": 2.1325828642901814, "grad_norm": 0.13411742448806763, "learning_rate": 4.02734657771138e-07, "loss": 0.6111, "step": 2561 }, { "epoch": 2.1334167187825726, "grad_norm": 0.13114993274211884, "learning_rate": 4.0201436853692704e-07, "loss": 0.5857, "step": 2562 }, { "epoch": 2.1342505732749637, "grad_norm": 0.13248169422149658, "learning_rate": 4.01294561915921e-07, "loss": 0.5844, "step": 2563 }, { "epoch": 2.1350844277673544, "grad_norm": 0.14177027344703674, "learning_rate": 4.005752384890504e-07, "loss": 0.6291, "step": 2564 }, { "epoch": 2.1359182822597456, "grad_norm": 0.13556669652462006, "learning_rate": 3.998563988368556e-07, "loss": 0.5753, "step": 2565 }, { "epoch": 2.1367521367521367, "grad_norm": 0.1372106969356537, "learning_rate": 3.991380435394872e-07, "loss": 0.6244, "step": 2566 }, { "epoch": 2.137585991244528, "grad_norm": 0.12975041568279266, "learning_rate": 3.984201731767042e-07, "loss": 0.6034, "step": 2567 }, { "epoch": 2.138419845736919, "grad_norm": 0.13041721284389496, "learning_rate": 3.977027883278744e-07, "loss": 0.5607, "step": 2568 }, { "epoch": 2.13925370022931, "grad_norm": 0.1401863396167755, "learning_rate": 3.9698588957197454e-07, "loss": 0.5751, "step": 2569 }, { "epoch": 2.140087554721701, "grad_norm": 0.13547448813915253, "learning_rate": 3.962694774875879e-07, "loss": 0.6007, "step": 2570 }, { "epoch": 2.140921409214092, "grad_norm": 0.14217804372310638, "learning_rate": 3.95553552652906e-07, "loss": 0.5935, "step": 2571 }, { "epoch": 2.141755263706483, "grad_norm": 0.12824539840221405, "learning_rate": 3.9483811564572635e-07, "loss": 0.6021, "step": 2572 }, { "epoch": 2.1425891181988743, "grad_norm": 0.13882043957710266, "learning_rate": 3.9412316704345304e-07, "loss": 0.6097, "step": 2573 }, { "epoch": 2.1434229726912655, "grad_norm": 0.13072605431079865, "learning_rate": 3.934087074230961e-07, "loss": 0.6078, "step": 2574 }, { "epoch": 2.1442568271836566, "grad_norm": 0.13155391812324524, "learning_rate": 3.9269473736127065e-07, "loss": 0.571, "step": 2575 }, { "epoch": 2.1450906816760473, "grad_norm": 0.13054433465003967, "learning_rate": 3.919812574341965e-07, "loss": 0.5635, "step": 2576 }, { "epoch": 2.1459245361684385, "grad_norm": 0.12609881162643433, "learning_rate": 3.912682682176991e-07, "loss": 0.59, "step": 2577 }, { "epoch": 2.1467583906608296, "grad_norm": 0.13616463541984558, "learning_rate": 3.9055577028720656e-07, "loss": 0.5864, "step": 2578 }, { "epoch": 2.147592245153221, "grad_norm": 0.1290634572505951, "learning_rate": 3.898437642177508e-07, "loss": 0.6198, "step": 2579 }, { "epoch": 2.148426099645612, "grad_norm": 0.1329745352268219, "learning_rate": 3.8913225058396717e-07, "loss": 0.614, "step": 2580 }, { "epoch": 2.149259954138003, "grad_norm": 0.12603352963924408, "learning_rate": 3.884212299600932e-07, "loss": 0.5614, "step": 2581 }, { "epoch": 2.150093808630394, "grad_norm": 0.13042472302913666, "learning_rate": 3.8771070291996865e-07, "loss": 0.6104, "step": 2582 }, { "epoch": 2.150927663122785, "grad_norm": 0.12941470742225647, "learning_rate": 3.870006700370347e-07, "loss": 0.6026, "step": 2583 }, { "epoch": 2.151761517615176, "grad_norm": 0.13917414844036102, "learning_rate": 3.862911318843346e-07, "loss": 0.571, "step": 2584 }, { "epoch": 2.1525953721075672, "grad_norm": 0.12937042117118835, "learning_rate": 3.855820890345109e-07, "loss": 0.5706, "step": 2585 }, { "epoch": 2.1534292265999584, "grad_norm": 0.13436506688594818, "learning_rate": 3.8487354205980806e-07, "loss": 0.6037, "step": 2586 }, { "epoch": 2.1542630810923495, "grad_norm": 0.12637221813201904, "learning_rate": 3.84165491532069e-07, "loss": 0.5746, "step": 2587 }, { "epoch": 2.1550969355847407, "grad_norm": 0.14346598088741302, "learning_rate": 3.834579380227364e-07, "loss": 0.6156, "step": 2588 }, { "epoch": 2.1559307900771314, "grad_norm": 0.12553030252456665, "learning_rate": 3.827508821028519e-07, "loss": 0.5304, "step": 2589 }, { "epoch": 2.1567646445695225, "grad_norm": 0.1333545595407486, "learning_rate": 3.8204432434305545e-07, "loss": 0.5691, "step": 2590 }, { "epoch": 2.1575984990619137, "grad_norm": 0.13001546263694763, "learning_rate": 3.8133826531358484e-07, "loss": 0.5925, "step": 2591 }, { "epoch": 2.158432353554305, "grad_norm": 0.13258573412895203, "learning_rate": 3.80632705584276e-07, "loss": 0.5485, "step": 2592 }, { "epoch": 2.159266208046696, "grad_norm": 0.13168732821941376, "learning_rate": 3.7992764572456116e-07, "loss": 0.6103, "step": 2593 }, { "epoch": 2.1601000625390867, "grad_norm": 0.12910717725753784, "learning_rate": 3.7922308630346935e-07, "loss": 0.531, "step": 2594 }, { "epoch": 2.160933917031478, "grad_norm": 0.13542519509792328, "learning_rate": 3.7851902788962573e-07, "loss": 0.583, "step": 2595 }, { "epoch": 2.161767771523869, "grad_norm": 0.14225146174430847, "learning_rate": 3.778154710512512e-07, "loss": 0.6395, "step": 2596 }, { "epoch": 2.16260162601626, "grad_norm": 0.13072539865970612, "learning_rate": 3.771124163561613e-07, "loss": 0.6106, "step": 2597 }, { "epoch": 2.1634354805086513, "grad_norm": 0.12915197014808655, "learning_rate": 3.764098643717677e-07, "loss": 0.5923, "step": 2598 }, { "epoch": 2.1642693350010425, "grad_norm": 0.13672548532485962, "learning_rate": 3.757078156650745e-07, "loss": 0.6079, "step": 2599 }, { "epoch": 2.1651031894934336, "grad_norm": 0.1376611441373825, "learning_rate": 3.7500627080268134e-07, "loss": 0.5911, "step": 2600 }, { "epoch": 2.1659370439858243, "grad_norm": 0.13688907027244568, "learning_rate": 3.743052303507801e-07, "loss": 0.6018, "step": 2601 }, { "epoch": 2.1667708984782155, "grad_norm": 0.13396038115024567, "learning_rate": 3.73604694875156e-07, "loss": 0.5586, "step": 2602 }, { "epoch": 2.1676047529706066, "grad_norm": 0.13952206075191498, "learning_rate": 3.729046649411864e-07, "loss": 0.6297, "step": 2603 }, { "epoch": 2.1684386074629978, "grad_norm": 0.1293851137161255, "learning_rate": 3.7220514111384115e-07, "loss": 0.5353, "step": 2604 }, { "epoch": 2.169272461955389, "grad_norm": 0.13454820215702057, "learning_rate": 3.715061239576809e-07, "loss": 0.624, "step": 2605 }, { "epoch": 2.17010631644778, "grad_norm": 0.13024364411830902, "learning_rate": 3.708076140368585e-07, "loss": 0.5573, "step": 2606 }, { "epoch": 2.1709401709401708, "grad_norm": 0.13569104671478271, "learning_rate": 3.701096119151165e-07, "loss": 0.6147, "step": 2607 }, { "epoch": 2.171774025432562, "grad_norm": 0.13256074488162994, "learning_rate": 3.6941211815578776e-07, "loss": 0.5346, "step": 2608 }, { "epoch": 2.172607879924953, "grad_norm": 0.1333748698234558, "learning_rate": 3.6871513332179513e-07, "loss": 0.5391, "step": 2609 }, { "epoch": 2.1734417344173442, "grad_norm": 0.12810109555721283, "learning_rate": 3.680186579756507e-07, "loss": 0.6132, "step": 2610 }, { "epoch": 2.1742755889097354, "grad_norm": 0.14090800285339355, "learning_rate": 3.6732269267945506e-07, "loss": 0.585, "step": 2611 }, { "epoch": 2.1751094434021265, "grad_norm": 0.12614275515079498, "learning_rate": 3.6662723799489703e-07, "loss": 0.5533, "step": 2612 }, { "epoch": 2.1759432978945172, "grad_norm": 0.12744340300559998, "learning_rate": 3.659322944832545e-07, "loss": 0.6332, "step": 2613 }, { "epoch": 2.1767771523869084, "grad_norm": 0.1316831409931183, "learning_rate": 3.652378627053911e-07, "loss": 0.5818, "step": 2614 }, { "epoch": 2.1776110068792995, "grad_norm": 0.1312979906797409, "learning_rate": 3.645439432217593e-07, "loss": 0.5593, "step": 2615 }, { "epoch": 2.1784448613716907, "grad_norm": 0.13055771589279175, "learning_rate": 3.6385053659239653e-07, "loss": 0.5446, "step": 2616 }, { "epoch": 2.179278715864082, "grad_norm": 0.14121727645397186, "learning_rate": 3.63157643376927e-07, "loss": 0.5793, "step": 2617 }, { "epoch": 2.180112570356473, "grad_norm": 0.13053211569786072, "learning_rate": 3.6246526413456057e-07, "loss": 0.544, "step": 2618 }, { "epoch": 2.1809464248488637, "grad_norm": 0.1390431970357895, "learning_rate": 3.617733994240921e-07, "loss": 0.6337, "step": 2619 }, { "epoch": 2.181780279341255, "grad_norm": 0.12655310332775116, "learning_rate": 3.610820498039011e-07, "loss": 0.5479, "step": 2620 }, { "epoch": 2.182614133833646, "grad_norm": 0.13425566256046295, "learning_rate": 3.603912158319522e-07, "loss": 0.6182, "step": 2621 }, { "epoch": 2.183447988326037, "grad_norm": 0.13877920806407928, "learning_rate": 3.5970089806579286e-07, "loss": 0.63, "step": 2622 }, { "epoch": 2.1842818428184283, "grad_norm": 0.1324535757303238, "learning_rate": 3.590110970625543e-07, "loss": 0.5958, "step": 2623 }, { "epoch": 2.1851156973108194, "grad_norm": 0.13059507310390472, "learning_rate": 3.5832181337895074e-07, "loss": 0.5963, "step": 2624 }, { "epoch": 2.18594955180321, "grad_norm": 0.13301368057727814, "learning_rate": 3.5763304757127877e-07, "loss": 0.5639, "step": 2625 }, { "epoch": 2.1867834062956013, "grad_norm": 0.13046878576278687, "learning_rate": 3.569448001954168e-07, "loss": 0.5987, "step": 2626 }, { "epoch": 2.1876172607879925, "grad_norm": 0.13746918737888336, "learning_rate": 3.5625707180682583e-07, "loss": 0.6051, "step": 2627 }, { "epoch": 2.1884511152803836, "grad_norm": 0.12730474770069122, "learning_rate": 3.5556986296054693e-07, "loss": 0.5908, "step": 2628 }, { "epoch": 2.1892849697727748, "grad_norm": 0.1462775319814682, "learning_rate": 3.548831742112017e-07, "loss": 0.5629, "step": 2629 }, { "epoch": 2.190118824265166, "grad_norm": 0.13233694434165955, "learning_rate": 3.5419700611299353e-07, "loss": 0.6072, "step": 2630 }, { "epoch": 2.1909526787575566, "grad_norm": 0.14441508054733276, "learning_rate": 3.535113592197041e-07, "loss": 0.6227, "step": 2631 }, { "epoch": 2.1917865332499478, "grad_norm": 0.13690954446792603, "learning_rate": 3.5282623408469494e-07, "loss": 0.6052, "step": 2632 }, { "epoch": 2.192620387742339, "grad_norm": 0.1288333684206009, "learning_rate": 3.5214163126090633e-07, "loss": 0.539, "step": 2633 }, { "epoch": 2.19345424223473, "grad_norm": 0.12828801572322845, "learning_rate": 3.514575513008572e-07, "loss": 0.5885, "step": 2634 }, { "epoch": 2.194288096727121, "grad_norm": 0.13093101978302002, "learning_rate": 3.5077399475664473e-07, "loss": 0.5463, "step": 2635 }, { "epoch": 2.1951219512195124, "grad_norm": 0.13602732121944427, "learning_rate": 3.500909621799434e-07, "loss": 0.5804, "step": 2636 }, { "epoch": 2.195955805711903, "grad_norm": 0.13446517288684845, "learning_rate": 3.4940845412200457e-07, "loss": 0.5284, "step": 2637 }, { "epoch": 2.1967896602042942, "grad_norm": 0.13395337760448456, "learning_rate": 3.487264711336567e-07, "loss": 0.6, "step": 2638 }, { "epoch": 2.1976235146966854, "grad_norm": 0.12414367496967316, "learning_rate": 3.480450137653043e-07, "loss": 0.5831, "step": 2639 }, { "epoch": 2.1984573691890765, "grad_norm": 0.14278534054756165, "learning_rate": 3.473640825669275e-07, "loss": 0.5867, "step": 2640 }, { "epoch": 2.1992912236814677, "grad_norm": 0.13531115651130676, "learning_rate": 3.466836780880817e-07, "loss": 0.5919, "step": 2641 }, { "epoch": 2.200125078173859, "grad_norm": 0.12811005115509033, "learning_rate": 3.4600380087789837e-07, "loss": 0.5688, "step": 2642 }, { "epoch": 2.2009589326662495, "grad_norm": 0.1400516927242279, "learning_rate": 3.453244514850816e-07, "loss": 0.5766, "step": 2643 }, { "epoch": 2.2017927871586407, "grad_norm": 0.13619574904441833, "learning_rate": 3.446456304579112e-07, "loss": 0.6037, "step": 2644 }, { "epoch": 2.202626641651032, "grad_norm": 0.13807177543640137, "learning_rate": 3.439673383442393e-07, "loss": 0.6823, "step": 2645 }, { "epoch": 2.203460496143423, "grad_norm": 0.13773640990257263, "learning_rate": 3.4328957569149186e-07, "loss": 0.5468, "step": 2646 }, { "epoch": 2.204294350635814, "grad_norm": 0.1348293572664261, "learning_rate": 3.4261234304666717e-07, "loss": 0.5808, "step": 2647 }, { "epoch": 2.2051282051282053, "grad_norm": 0.12254226952791214, "learning_rate": 3.419356409563361e-07, "loss": 0.5757, "step": 2648 }, { "epoch": 2.205962059620596, "grad_norm": 0.1347162276506424, "learning_rate": 3.4125946996664057e-07, "loss": 0.5107, "step": 2649 }, { "epoch": 2.206795914112987, "grad_norm": 0.13183777034282684, "learning_rate": 3.405838306232953e-07, "loss": 0.5665, "step": 2650 }, { "epoch": 2.2076297686053783, "grad_norm": 0.13085879385471344, "learning_rate": 3.399087234715846e-07, "loss": 0.5817, "step": 2651 }, { "epoch": 2.2084636230977694, "grad_norm": 0.1383403092622757, "learning_rate": 3.392341490563639e-07, "loss": 0.6288, "step": 2652 }, { "epoch": 2.2092974775901606, "grad_norm": 0.14718247950077057, "learning_rate": 3.3856010792205836e-07, "loss": 0.5499, "step": 2653 }, { "epoch": 2.2101313320825517, "grad_norm": 0.14452199637889862, "learning_rate": 3.378866006126631e-07, "loss": 0.6587, "step": 2654 }, { "epoch": 2.210965186574943, "grad_norm": 0.1354537159204483, "learning_rate": 3.3721362767174166e-07, "loss": 0.5457, "step": 2655 }, { "epoch": 2.2117990410673336, "grad_norm": 0.12763375043869019, "learning_rate": 3.3654118964242763e-07, "loss": 0.5585, "step": 2656 }, { "epoch": 2.2126328955597248, "grad_norm": 0.13515634834766388, "learning_rate": 3.3586928706742193e-07, "loss": 0.6082, "step": 2657 }, { "epoch": 2.213466750052116, "grad_norm": 0.13163356482982635, "learning_rate": 3.35197920488993e-07, "loss": 0.5796, "step": 2658 }, { "epoch": 2.214300604544507, "grad_norm": 0.1281367689371109, "learning_rate": 3.3452709044897796e-07, "loss": 0.5869, "step": 2659 }, { "epoch": 2.215134459036898, "grad_norm": 0.12514398992061615, "learning_rate": 3.338567974887799e-07, "loss": 0.5714, "step": 2660 }, { "epoch": 2.215968313529289, "grad_norm": 0.13542534410953522, "learning_rate": 3.3318704214936875e-07, "loss": 0.6169, "step": 2661 }, { "epoch": 2.21680216802168, "grad_norm": 0.134818896651268, "learning_rate": 3.3251782497128046e-07, "loss": 0.619, "step": 2662 }, { "epoch": 2.217636022514071, "grad_norm": 0.14056462049484253, "learning_rate": 3.318491464946163e-07, "loss": 0.5519, "step": 2663 }, { "epoch": 2.2184698770064624, "grad_norm": 0.1297370195388794, "learning_rate": 3.31181007259044e-07, "loss": 0.5778, "step": 2664 }, { "epoch": 2.2193037314988535, "grad_norm": 0.13773800432682037, "learning_rate": 3.3051340780379486e-07, "loss": 0.6441, "step": 2665 }, { "epoch": 2.2201375859912447, "grad_norm": 0.1310642957687378, "learning_rate": 3.2984634866766514e-07, "loss": 0.6099, "step": 2666 }, { "epoch": 2.220971440483636, "grad_norm": 0.13338206708431244, "learning_rate": 3.2917983038901455e-07, "loss": 0.5836, "step": 2667 }, { "epoch": 2.2218052949760265, "grad_norm": 0.13697199523448944, "learning_rate": 3.285138535057668e-07, "loss": 0.6261, "step": 2668 }, { "epoch": 2.2226391494684177, "grad_norm": 0.13657347857952118, "learning_rate": 3.2784841855540834e-07, "loss": 0.5845, "step": 2669 }, { "epoch": 2.223473003960809, "grad_norm": 0.13969075679779053, "learning_rate": 3.271835260749881e-07, "loss": 0.6062, "step": 2670 }, { "epoch": 2.2243068584532, "grad_norm": 0.12664192914962769, "learning_rate": 3.2651917660111807e-07, "loss": 0.569, "step": 2671 }, { "epoch": 2.225140712945591, "grad_norm": 0.1309032142162323, "learning_rate": 3.258553706699706e-07, "loss": 0.5738, "step": 2672 }, { "epoch": 2.2259745674379823, "grad_norm": 0.13431347906589508, "learning_rate": 3.251921088172811e-07, "loss": 0.5937, "step": 2673 }, { "epoch": 2.226808421930373, "grad_norm": 0.13456088304519653, "learning_rate": 3.245293915783444e-07, "loss": 0.5482, "step": 2674 }, { "epoch": 2.227642276422764, "grad_norm": 0.11839820444583893, "learning_rate": 3.238672194880162e-07, "loss": 0.558, "step": 2675 }, { "epoch": 2.2284761309151553, "grad_norm": 0.14328329265117645, "learning_rate": 3.232055930807125e-07, "loss": 0.592, "step": 2676 }, { "epoch": 2.2293099854075464, "grad_norm": 0.14188343286514282, "learning_rate": 3.2254451289040885e-07, "loss": 0.6332, "step": 2677 }, { "epoch": 2.2301438398999376, "grad_norm": 0.13297003507614136, "learning_rate": 3.2188397945063926e-07, "loss": 0.6207, "step": 2678 }, { "epoch": 2.2309776943923287, "grad_norm": 0.13137394189834595, "learning_rate": 3.212239932944979e-07, "loss": 0.5588, "step": 2679 }, { "epoch": 2.2318115488847194, "grad_norm": 0.13637933135032654, "learning_rate": 3.205645549546361e-07, "loss": 0.6085, "step": 2680 }, { "epoch": 2.2326454033771106, "grad_norm": 0.127396821975708, "learning_rate": 3.1990566496326334e-07, "loss": 0.5443, "step": 2681 }, { "epoch": 2.2334792578695017, "grad_norm": 0.13576333224773407, "learning_rate": 3.1924732385214657e-07, "loss": 0.5611, "step": 2682 }, { "epoch": 2.234313112361893, "grad_norm": 0.13989579677581787, "learning_rate": 3.1858953215260987e-07, "loss": 0.6044, "step": 2683 }, { "epoch": 2.235146966854284, "grad_norm": 0.12802767753601074, "learning_rate": 3.179322903955337e-07, "loss": 0.6045, "step": 2684 }, { "epoch": 2.235980821346675, "grad_norm": 0.1337248533964157, "learning_rate": 3.172755991113546e-07, "loss": 0.5668, "step": 2685 }, { "epoch": 2.236814675839066, "grad_norm": 0.1272716522216797, "learning_rate": 3.166194588300656e-07, "loss": 0.5934, "step": 2686 }, { "epoch": 2.237648530331457, "grad_norm": 0.1317165046930313, "learning_rate": 3.159638700812138e-07, "loss": 0.6131, "step": 2687 }, { "epoch": 2.238482384823848, "grad_norm": 0.13603195548057556, "learning_rate": 3.153088333939027e-07, "loss": 0.6115, "step": 2688 }, { "epoch": 2.2393162393162394, "grad_norm": 0.13870610296726227, "learning_rate": 3.1465434929678894e-07, "loss": 0.5999, "step": 2689 }, { "epoch": 2.2401500938086305, "grad_norm": 0.14465558528900146, "learning_rate": 3.1400041831808365e-07, "loss": 0.632, "step": 2690 }, { "epoch": 2.2409839483010217, "grad_norm": 0.13975290954113007, "learning_rate": 3.133470409855515e-07, "loss": 0.5866, "step": 2691 }, { "epoch": 2.2418178027934124, "grad_norm": 0.1341933161020279, "learning_rate": 3.1269421782651064e-07, "loss": 0.6207, "step": 2692 }, { "epoch": 2.2426516572858035, "grad_norm": 0.13104739785194397, "learning_rate": 3.120419493678311e-07, "loss": 0.5444, "step": 2693 }, { "epoch": 2.2434855117781947, "grad_norm": 0.12758061289787292, "learning_rate": 3.113902361359367e-07, "loss": 0.5622, "step": 2694 }, { "epoch": 2.244319366270586, "grad_norm": 0.14147430658340454, "learning_rate": 3.107390786568019e-07, "loss": 0.615, "step": 2695 }, { "epoch": 2.245153220762977, "grad_norm": 0.13600394129753113, "learning_rate": 3.10088477455953e-07, "loss": 0.6139, "step": 2696 }, { "epoch": 2.245987075255368, "grad_norm": 0.14371150732040405, "learning_rate": 3.094384330584674e-07, "loss": 0.6492, "step": 2697 }, { "epoch": 2.246820929747759, "grad_norm": 0.14032989740371704, "learning_rate": 3.087889459889732e-07, "loss": 0.5623, "step": 2698 }, { "epoch": 2.24765478424015, "grad_norm": 0.1351948231458664, "learning_rate": 3.0814001677164814e-07, "loss": 0.6272, "step": 2699 }, { "epoch": 2.248488638732541, "grad_norm": 0.12811696529388428, "learning_rate": 3.0749164593022105e-07, "loss": 0.5352, "step": 2700 }, { "epoch": 2.2493224932249323, "grad_norm": 0.12757426500320435, "learning_rate": 3.068438339879683e-07, "loss": 0.5254, "step": 2701 }, { "epoch": 2.2501563477173234, "grad_norm": 0.14932145178318024, "learning_rate": 3.061965814677172e-07, "loss": 0.5696, "step": 2702 }, { "epoch": 2.2509902022097146, "grad_norm": 0.12584970891475677, "learning_rate": 3.0554988889184183e-07, "loss": 0.5436, "step": 2703 }, { "epoch": 2.2518240567021053, "grad_norm": 0.1408921331167221, "learning_rate": 3.049037567822654e-07, "loss": 0.5891, "step": 2704 }, { "epoch": 2.2526579111944964, "grad_norm": 0.13569395244121552, "learning_rate": 3.0425818566045825e-07, "loss": 0.5709, "step": 2705 }, { "epoch": 2.2534917656868876, "grad_norm": 0.14067214727401733, "learning_rate": 3.0361317604743806e-07, "loss": 0.6298, "step": 2706 }, { "epoch": 2.2543256201792787, "grad_norm": 0.13769875466823578, "learning_rate": 3.029687284637694e-07, "loss": 0.6287, "step": 2707 }, { "epoch": 2.25515947467167, "grad_norm": 0.12546081840991974, "learning_rate": 3.023248434295638e-07, "loss": 0.5613, "step": 2708 }, { "epoch": 2.255993329164061, "grad_norm": 0.12654441595077515, "learning_rate": 3.0168152146447777e-07, "loss": 0.5902, "step": 2709 }, { "epoch": 2.256827183656452, "grad_norm": 0.13255999982357025, "learning_rate": 3.0103876308771414e-07, "loss": 0.6342, "step": 2710 }, { "epoch": 2.257661038148843, "grad_norm": 0.13959909975528717, "learning_rate": 3.0039656881802054e-07, "loss": 0.5497, "step": 2711 }, { "epoch": 2.258494892641234, "grad_norm": 0.13817466795444489, "learning_rate": 2.9975493917368934e-07, "loss": 0.5663, "step": 2712 }, { "epoch": 2.259328747133625, "grad_norm": 0.1371651142835617, "learning_rate": 2.9911387467255735e-07, "loss": 0.6335, "step": 2713 }, { "epoch": 2.2601626016260163, "grad_norm": 0.13358014822006226, "learning_rate": 2.9847337583200495e-07, "loss": 0.6309, "step": 2714 }, { "epoch": 2.2609964561184075, "grad_norm": 0.134216770529747, "learning_rate": 2.9783344316895677e-07, "loss": 0.5744, "step": 2715 }, { "epoch": 2.261830310610798, "grad_norm": 0.13467298448085785, "learning_rate": 2.9719407719987944e-07, "loss": 0.5931, "step": 2716 }, { "epoch": 2.2626641651031894, "grad_norm": 0.13385699689388275, "learning_rate": 2.965552784407834e-07, "loss": 0.5943, "step": 2717 }, { "epoch": 2.2634980195955805, "grad_norm": 0.14070317149162292, "learning_rate": 2.9591704740722035e-07, "loss": 0.6106, "step": 2718 }, { "epoch": 2.2643318740879717, "grad_norm": 0.13239797949790955, "learning_rate": 2.9527938461428425e-07, "loss": 0.6471, "step": 2719 }, { "epoch": 2.265165728580363, "grad_norm": 0.14040181040763855, "learning_rate": 2.9464229057661026e-07, "loss": 0.591, "step": 2720 }, { "epoch": 2.265999583072754, "grad_norm": 0.131845161318779, "learning_rate": 2.9400576580837464e-07, "loss": 0.531, "step": 2721 }, { "epoch": 2.266833437565145, "grad_norm": 0.1324426680803299, "learning_rate": 2.9336981082329383e-07, "loss": 0.5838, "step": 2722 }, { "epoch": 2.267667292057536, "grad_norm": 0.12901893258094788, "learning_rate": 2.927344261346254e-07, "loss": 0.5819, "step": 2723 }, { "epoch": 2.268501146549927, "grad_norm": 0.13579130172729492, "learning_rate": 2.9209961225516576e-07, "loss": 0.6084, "step": 2724 }, { "epoch": 2.269335001042318, "grad_norm": 0.13500390946865082, "learning_rate": 2.9146536969725077e-07, "loss": 0.5656, "step": 2725 }, { "epoch": 2.2701688555347093, "grad_norm": 0.12529639899730682, "learning_rate": 2.908316989727555e-07, "loss": 0.5497, "step": 2726 }, { "epoch": 2.2710027100271004, "grad_norm": 0.14078211784362793, "learning_rate": 2.901986005930933e-07, "loss": 0.6463, "step": 2727 }, { "epoch": 2.271836564519491, "grad_norm": 0.13957035541534424, "learning_rate": 2.8956607506921516e-07, "loss": 0.6332, "step": 2728 }, { "epoch": 2.2726704190118823, "grad_norm": 0.13345807790756226, "learning_rate": 2.889341229116111e-07, "loss": 0.5807, "step": 2729 }, { "epoch": 2.2735042735042734, "grad_norm": 0.13118784129619598, "learning_rate": 2.8830274463030703e-07, "loss": 0.5783, "step": 2730 }, { "epoch": 2.2743381279966646, "grad_norm": 0.142796590924263, "learning_rate": 2.876719407348659e-07, "loss": 0.6766, "step": 2731 }, { "epoch": 2.2751719824890557, "grad_norm": 0.1316111981868744, "learning_rate": 2.8704171173438796e-07, "loss": 0.5867, "step": 2732 }, { "epoch": 2.276005836981447, "grad_norm": 0.14183148741722107, "learning_rate": 2.864120581375088e-07, "loss": 0.5996, "step": 2733 }, { "epoch": 2.276839691473838, "grad_norm": 0.1354745328426361, "learning_rate": 2.8578298045239947e-07, "loss": 0.5927, "step": 2734 }, { "epoch": 2.2776735459662287, "grad_norm": 0.14374493062496185, "learning_rate": 2.851544791867666e-07, "loss": 0.6082, "step": 2735 }, { "epoch": 2.27850740045862, "grad_norm": 0.12484464049339294, "learning_rate": 2.8452655484785125e-07, "loss": 0.6088, "step": 2736 }, { "epoch": 2.279341254951011, "grad_norm": 0.14090214669704437, "learning_rate": 2.838992079424296e-07, "loss": 0.6311, "step": 2737 }, { "epoch": 2.280175109443402, "grad_norm": 0.14000464975833893, "learning_rate": 2.8327243897681107e-07, "loss": 0.6066, "step": 2738 }, { "epoch": 2.2810089639357933, "grad_norm": 0.12474175542593002, "learning_rate": 2.826462484568389e-07, "loss": 0.5723, "step": 2739 }, { "epoch": 2.281842818428184, "grad_norm": 0.1444791853427887, "learning_rate": 2.8202063688788933e-07, "loss": 0.6554, "step": 2740 }, { "epoch": 2.282676672920575, "grad_norm": 0.1323375105857849, "learning_rate": 2.813956047748717e-07, "loss": 0.5797, "step": 2741 }, { "epoch": 2.2835105274129663, "grad_norm": 0.13255009055137634, "learning_rate": 2.807711526222274e-07, "loss": 0.5765, "step": 2742 }, { "epoch": 2.2843443819053575, "grad_norm": 0.13164877891540527, "learning_rate": 2.801472809339294e-07, "loss": 0.5517, "step": 2743 }, { "epoch": 2.2851782363977486, "grad_norm": 0.12710009515285492, "learning_rate": 2.795239902134834e-07, "loss": 0.5657, "step": 2744 }, { "epoch": 2.28601209089014, "grad_norm": 0.13062423467636108, "learning_rate": 2.7890128096392474e-07, "loss": 0.5787, "step": 2745 }, { "epoch": 2.286845945382531, "grad_norm": 0.1336994171142578, "learning_rate": 2.7827915368782085e-07, "loss": 0.6224, "step": 2746 }, { "epoch": 2.2876797998749216, "grad_norm": 0.13501402735710144, "learning_rate": 2.776576088872685e-07, "loss": 0.5671, "step": 2747 }, { "epoch": 2.288513654367313, "grad_norm": 0.13158062100410461, "learning_rate": 2.770366470638947e-07, "loss": 0.5932, "step": 2748 }, { "epoch": 2.289347508859704, "grad_norm": 0.1350419968366623, "learning_rate": 2.7641626871885595e-07, "loss": 0.5793, "step": 2749 }, { "epoch": 2.290181363352095, "grad_norm": 0.13752858340740204, "learning_rate": 2.757964743528377e-07, "loss": 0.6149, "step": 2750 }, { "epoch": 2.2910152178444863, "grad_norm": 0.13495661318302155, "learning_rate": 2.75177264466054e-07, "loss": 0.6124, "step": 2751 }, { "epoch": 2.2918490723368774, "grad_norm": 0.14049872756004333, "learning_rate": 2.745586395582481e-07, "loss": 0.5992, "step": 2752 }, { "epoch": 2.292682926829268, "grad_norm": 0.13410940766334534, "learning_rate": 2.739406001286899e-07, "loss": 0.6323, "step": 2753 }, { "epoch": 2.2935167813216593, "grad_norm": 0.1433565318584442, "learning_rate": 2.733231466761776e-07, "loss": 0.5732, "step": 2754 }, { "epoch": 2.2943506358140504, "grad_norm": 0.12989090383052826, "learning_rate": 2.7270627969903603e-07, "loss": 0.5892, "step": 2755 }, { "epoch": 2.2951844903064416, "grad_norm": 0.13756996393203735, "learning_rate": 2.7208999969511695e-07, "loss": 0.5966, "step": 2756 }, { "epoch": 2.2960183447988327, "grad_norm": 0.12973067164421082, "learning_rate": 2.714743071617979e-07, "loss": 0.5318, "step": 2757 }, { "epoch": 2.296852199291224, "grad_norm": 0.13855144381523132, "learning_rate": 2.708592025959833e-07, "loss": 0.5771, "step": 2758 }, { "epoch": 2.2976860537836146, "grad_norm": 0.12780117988586426, "learning_rate": 2.7024468649410224e-07, "loss": 0.594, "step": 2759 }, { "epoch": 2.2985199082760057, "grad_norm": 0.13220033049583435, "learning_rate": 2.6963075935210875e-07, "loss": 0.6436, "step": 2760 }, { "epoch": 2.299353762768397, "grad_norm": 0.13308489322662354, "learning_rate": 2.690174216654826e-07, "loss": 0.5942, "step": 2761 }, { "epoch": 2.300187617260788, "grad_norm": 0.13608810305595398, "learning_rate": 2.684046739292266e-07, "loss": 0.5787, "step": 2762 }, { "epoch": 2.301021471753179, "grad_norm": 0.129466250538826, "learning_rate": 2.6779251663786797e-07, "loss": 0.5585, "step": 2763 }, { "epoch": 2.3018553262455703, "grad_norm": 0.13196049630641937, "learning_rate": 2.671809502854574e-07, "loss": 0.5404, "step": 2764 }, { "epoch": 2.3026891807379615, "grad_norm": 0.13600628077983856, "learning_rate": 2.6656997536556836e-07, "loss": 0.5732, "step": 2765 }, { "epoch": 2.303523035230352, "grad_norm": 0.1334197223186493, "learning_rate": 2.6595959237129783e-07, "loss": 0.5765, "step": 2766 }, { "epoch": 2.3043568897227433, "grad_norm": 0.1256718635559082, "learning_rate": 2.6534980179526414e-07, "loss": 0.5599, "step": 2767 }, { "epoch": 2.3051907442151345, "grad_norm": 0.14242585003376007, "learning_rate": 2.6474060412960784e-07, "loss": 0.5698, "step": 2768 }, { "epoch": 2.3060245987075256, "grad_norm": 0.12964016199111938, "learning_rate": 2.641319998659911e-07, "loss": 0.5852, "step": 2769 }, { "epoch": 2.306858453199917, "grad_norm": 0.23954735696315765, "learning_rate": 2.6352398949559695e-07, "loss": 0.6053, "step": 2770 }, { "epoch": 2.3076923076923075, "grad_norm": 0.13867707550525665, "learning_rate": 2.629165735091292e-07, "loss": 0.6047, "step": 2771 }, { "epoch": 2.3085261621846986, "grad_norm": 0.1344437450170517, "learning_rate": 2.623097523968116e-07, "loss": 0.5431, "step": 2772 }, { "epoch": 2.30936001667709, "grad_norm": 0.13694731891155243, "learning_rate": 2.61703526648389e-07, "loss": 0.5832, "step": 2773 }, { "epoch": 2.310193871169481, "grad_norm": 0.12899625301361084, "learning_rate": 2.610978967531241e-07, "loss": 0.5934, "step": 2774 }, { "epoch": 2.311027725661872, "grad_norm": 0.13291893899440765, "learning_rate": 2.604928631998001e-07, "loss": 0.6081, "step": 2775 }, { "epoch": 2.3118615801542632, "grad_norm": 0.13450887799263, "learning_rate": 2.598884264767183e-07, "loss": 0.59, "step": 2776 }, { "epoch": 2.3126954346466544, "grad_norm": 0.14478500187397003, "learning_rate": 2.592845870716981e-07, "loss": 0.5468, "step": 2777 }, { "epoch": 2.313529289139045, "grad_norm": 0.14345842599868774, "learning_rate": 2.586813454720771e-07, "loss": 0.5909, "step": 2778 }, { "epoch": 2.3143631436314362, "grad_norm": 0.1280716359615326, "learning_rate": 2.5807870216471053e-07, "loss": 0.5806, "step": 2779 }, { "epoch": 2.3151969981238274, "grad_norm": 0.13994114100933075, "learning_rate": 2.574766576359702e-07, "loss": 0.5587, "step": 2780 }, { "epoch": 2.3160308526162185, "grad_norm": 0.12269765138626099, "learning_rate": 2.568752123717458e-07, "loss": 0.5338, "step": 2781 }, { "epoch": 2.3168647071086097, "grad_norm": 0.13782238960266113, "learning_rate": 2.562743668574424e-07, "loss": 0.6204, "step": 2782 }, { "epoch": 2.3176985616010004, "grad_norm": 0.12796927988529205, "learning_rate": 2.556741215779813e-07, "loss": 0.5655, "step": 2783 }, { "epoch": 2.3185324160933916, "grad_norm": 0.14152126014232635, "learning_rate": 2.550744770177993e-07, "loss": 0.5953, "step": 2784 }, { "epoch": 2.3193662705857827, "grad_norm": 0.13737402856349945, "learning_rate": 2.544754336608486e-07, "loss": 0.6098, "step": 2785 }, { "epoch": 2.320200125078174, "grad_norm": 0.13175059854984283, "learning_rate": 2.5387699199059565e-07, "loss": 0.606, "step": 2786 }, { "epoch": 2.321033979570565, "grad_norm": 0.13624677062034607, "learning_rate": 2.5327915249002243e-07, "loss": 0.5801, "step": 2787 }, { "epoch": 2.321867834062956, "grad_norm": 0.12359346449375153, "learning_rate": 2.5268191564162377e-07, "loss": 0.5886, "step": 2788 }, { "epoch": 2.3227016885553473, "grad_norm": 0.13575242459774017, "learning_rate": 2.520852819274083e-07, "loss": 0.5796, "step": 2789 }, { "epoch": 2.323535543047738, "grad_norm": 0.13357725739479065, "learning_rate": 2.514892518288988e-07, "loss": 0.6109, "step": 2790 }, { "epoch": 2.324369397540129, "grad_norm": 0.13602130115032196, "learning_rate": 2.5089382582712995e-07, "loss": 0.6003, "step": 2791 }, { "epoch": 2.3252032520325203, "grad_norm": 0.13068358600139618, "learning_rate": 2.502990044026491e-07, "loss": 0.6246, "step": 2792 }, { "epoch": 2.3260371065249115, "grad_norm": 0.14064845442771912, "learning_rate": 2.4970478803551564e-07, "loss": 0.556, "step": 2793 }, { "epoch": 2.3268709610173026, "grad_norm": 0.13104741275310516, "learning_rate": 2.4911117720530053e-07, "loss": 0.5898, "step": 2794 }, { "epoch": 2.3277048155096933, "grad_norm": 0.13186362385749817, "learning_rate": 2.4851817239108686e-07, "loss": 0.5548, "step": 2795 }, { "epoch": 2.3285386700020845, "grad_norm": 0.12773117423057556, "learning_rate": 2.4792577407146776e-07, "loss": 0.6292, "step": 2796 }, { "epoch": 2.3293725244944756, "grad_norm": 0.1371065080165863, "learning_rate": 2.4733398272454687e-07, "loss": 0.5916, "step": 2797 }, { "epoch": 2.3302063789868668, "grad_norm": 0.13803279399871826, "learning_rate": 2.467427988279385e-07, "loss": 0.5778, "step": 2798 }, { "epoch": 2.331040233479258, "grad_norm": 0.13700875639915466, "learning_rate": 2.4615222285876613e-07, "loss": 0.6062, "step": 2799 }, { "epoch": 2.331874087971649, "grad_norm": 0.13106635212898254, "learning_rate": 2.4556225529366316e-07, "loss": 0.5953, "step": 2800 }, { "epoch": 2.331874087971649, "eval_loss": 0.6067690253257751, "eval_runtime": 321.3203, "eval_samples_per_second": 16.753, "eval_steps_per_second": 2.795, "step": 2800 }, { "epoch": 2.3327079424640402, "grad_norm": 0.1305270940065384, "learning_rate": 2.449728966087712e-07, "loss": 0.6015, "step": 2801 }, { "epoch": 2.333541796956431, "grad_norm": 0.14800278842449188, "learning_rate": 2.4438414727974155e-07, "loss": 0.6022, "step": 2802 }, { "epoch": 2.334375651448822, "grad_norm": 0.14003446698188782, "learning_rate": 2.437960077817326e-07, "loss": 0.6113, "step": 2803 }, { "epoch": 2.3352095059412132, "grad_norm": 0.13879652321338654, "learning_rate": 2.4320847858941164e-07, "loss": 0.5794, "step": 2804 }, { "epoch": 2.3360433604336044, "grad_norm": 0.13818885385990143, "learning_rate": 2.4262156017695257e-07, "loss": 0.5859, "step": 2805 }, { "epoch": 2.3368772149259955, "grad_norm": 0.1325407326221466, "learning_rate": 2.420352530180367e-07, "loss": 0.5484, "step": 2806 }, { "epoch": 2.3377110694183862, "grad_norm": 0.13550327718257904, "learning_rate": 2.4144955758585184e-07, "loss": 0.6088, "step": 2807 }, { "epoch": 2.3385449239107774, "grad_norm": 0.1369825005531311, "learning_rate": 2.408644743530922e-07, "loss": 0.59, "step": 2808 }, { "epoch": 2.3393787784031685, "grad_norm": 0.13526295125484467, "learning_rate": 2.402800037919578e-07, "loss": 0.6223, "step": 2809 }, { "epoch": 2.3402126328955597, "grad_norm": 0.12998482584953308, "learning_rate": 2.396961463741547e-07, "loss": 0.5972, "step": 2810 }, { "epoch": 2.341046487387951, "grad_norm": 0.14300423860549927, "learning_rate": 2.3911290257089345e-07, "loss": 0.621, "step": 2811 }, { "epoch": 2.341880341880342, "grad_norm": 0.14357508718967438, "learning_rate": 2.3853027285288964e-07, "loss": 0.605, "step": 2812 }, { "epoch": 2.342714196372733, "grad_norm": 0.1260567307472229, "learning_rate": 2.379482576903633e-07, "loss": 0.5874, "step": 2813 }, { "epoch": 2.343548050865124, "grad_norm": 0.13730305433273315, "learning_rate": 2.3736685755303853e-07, "loss": 0.631, "step": 2814 }, { "epoch": 2.344381905357515, "grad_norm": 0.13938766717910767, "learning_rate": 2.3678607291014242e-07, "loss": 0.5714, "step": 2815 }, { "epoch": 2.345215759849906, "grad_norm": 0.12665730714797974, "learning_rate": 2.3620590423040666e-07, "loss": 0.6359, "step": 2816 }, { "epoch": 2.3460496143422973, "grad_norm": 0.13339179754257202, "learning_rate": 2.3562635198206471e-07, "loss": 0.6047, "step": 2817 }, { "epoch": 2.3468834688346885, "grad_norm": 0.13408726453781128, "learning_rate": 2.3504741663285244e-07, "loss": 0.5561, "step": 2818 }, { "epoch": 2.3477173233270796, "grad_norm": 0.13089053332805634, "learning_rate": 2.3446909865000887e-07, "loss": 0.5648, "step": 2819 }, { "epoch": 2.3485511778194703, "grad_norm": 0.13494056463241577, "learning_rate": 2.3389139850027383e-07, "loss": 0.6143, "step": 2820 }, { "epoch": 2.3493850323118615, "grad_norm": 0.13700903952121735, "learning_rate": 2.3331431664988886e-07, "loss": 0.64, "step": 2821 }, { "epoch": 2.3502188868042526, "grad_norm": 0.1396048367023468, "learning_rate": 2.3273785356459653e-07, "loss": 0.6574, "step": 2822 }, { "epoch": 2.3510527412966438, "grad_norm": 0.1414024829864502, "learning_rate": 2.321620097096395e-07, "loss": 0.5843, "step": 2823 }, { "epoch": 2.351886595789035, "grad_norm": 0.13183845579624176, "learning_rate": 2.3158678554976185e-07, "loss": 0.5201, "step": 2824 }, { "epoch": 2.352720450281426, "grad_norm": 0.13967518508434296, "learning_rate": 2.3101218154920632e-07, "loss": 0.5453, "step": 2825 }, { "epoch": 2.3535543047738168, "grad_norm": 0.1291603147983551, "learning_rate": 2.3043819817171583e-07, "loss": 0.5646, "step": 2826 }, { "epoch": 2.354388159266208, "grad_norm": 0.13071309030056, "learning_rate": 2.2986483588053218e-07, "loss": 0.5815, "step": 2827 }, { "epoch": 2.355222013758599, "grad_norm": 0.13536536693572998, "learning_rate": 2.2929209513839575e-07, "loss": 0.545, "step": 2828 }, { "epoch": 2.3560558682509902, "grad_norm": 0.13090923428535461, "learning_rate": 2.287199764075457e-07, "loss": 0.6102, "step": 2829 }, { "epoch": 2.3568897227433814, "grad_norm": 0.1410083770751953, "learning_rate": 2.2814848014971856e-07, "loss": 0.6419, "step": 2830 }, { "epoch": 2.3577235772357725, "grad_norm": 0.146549254655838, "learning_rate": 2.2757760682614946e-07, "loss": 0.6209, "step": 2831 }, { "epoch": 2.3585574317281637, "grad_norm": 0.14046458899974823, "learning_rate": 2.2700735689757001e-07, "loss": 0.566, "step": 2832 }, { "epoch": 2.3593912862205544, "grad_norm": 0.14285323023796082, "learning_rate": 2.264377308242086e-07, "loss": 0.5977, "step": 2833 }, { "epoch": 2.3602251407129455, "grad_norm": 0.1337665170431137, "learning_rate": 2.2586872906579092e-07, "loss": 0.5739, "step": 2834 }, { "epoch": 2.3610589952053367, "grad_norm": 0.13819743692874908, "learning_rate": 2.253003520815382e-07, "loss": 0.5627, "step": 2835 }, { "epoch": 2.361892849697728, "grad_norm": 0.13330934941768646, "learning_rate": 2.2473260033016728e-07, "loss": 0.6229, "step": 2836 }, { "epoch": 2.362726704190119, "grad_norm": 0.13035665452480316, "learning_rate": 2.2416547426989084e-07, "loss": 0.5695, "step": 2837 }, { "epoch": 2.3635605586825097, "grad_norm": 0.12953516840934753, "learning_rate": 2.2359897435841612e-07, "loss": 0.5101, "step": 2838 }, { "epoch": 2.364394413174901, "grad_norm": 0.1377396434545517, "learning_rate": 2.230331010529458e-07, "loss": 0.585, "step": 2839 }, { "epoch": 2.365228267667292, "grad_norm": 0.13672246038913727, "learning_rate": 2.2246785481017617e-07, "loss": 0.629, "step": 2840 }, { "epoch": 2.366062122159683, "grad_norm": 0.13422022759914398, "learning_rate": 2.2190323608629758e-07, "loss": 0.5424, "step": 2841 }, { "epoch": 2.3668959766520743, "grad_norm": 0.13355980813503265, "learning_rate": 2.213392453369939e-07, "loss": 0.5672, "step": 2842 }, { "epoch": 2.3677298311444654, "grad_norm": 0.13359272480010986, "learning_rate": 2.2077588301744233e-07, "loss": 0.6152, "step": 2843 }, { "epoch": 2.3685636856368566, "grad_norm": 0.1357693374156952, "learning_rate": 2.2021314958231286e-07, "loss": 0.6062, "step": 2844 }, { "epoch": 2.3693975401292473, "grad_norm": 0.13334722816944122, "learning_rate": 2.196510454857675e-07, "loss": 0.6019, "step": 2845 }, { "epoch": 2.3702313946216385, "grad_norm": 0.1295909434556961, "learning_rate": 2.1908957118146131e-07, "loss": 0.5827, "step": 2846 }, { "epoch": 2.3710652491140296, "grad_norm": 0.13570016622543335, "learning_rate": 2.1852872712254e-07, "loss": 0.6128, "step": 2847 }, { "epoch": 2.3718991036064208, "grad_norm": 0.12960213422775269, "learning_rate": 2.1796851376164162e-07, "loss": 0.6107, "step": 2848 }, { "epoch": 2.372732958098812, "grad_norm": 0.13513532280921936, "learning_rate": 2.1740893155089445e-07, "loss": 0.5886, "step": 2849 }, { "epoch": 2.3735668125912026, "grad_norm": 0.14333775639533997, "learning_rate": 2.1684998094191765e-07, "loss": 0.6249, "step": 2850 }, { "epoch": 2.3744006670835938, "grad_norm": 0.1326669454574585, "learning_rate": 2.1629166238582052e-07, "loss": 0.5783, "step": 2851 }, { "epoch": 2.375234521575985, "grad_norm": 0.1371835172176361, "learning_rate": 2.1573397633320223e-07, "loss": 0.5928, "step": 2852 }, { "epoch": 2.376068376068376, "grad_norm": 0.1275784820318222, "learning_rate": 2.1517692323415205e-07, "loss": 0.5378, "step": 2853 }, { "epoch": 2.376902230560767, "grad_norm": 0.13794054090976715, "learning_rate": 2.146205035382478e-07, "loss": 0.5634, "step": 2854 }, { "epoch": 2.3777360850531584, "grad_norm": 0.1349364072084427, "learning_rate": 2.1406471769455615e-07, "loss": 0.5904, "step": 2855 }, { "epoch": 2.3785699395455495, "grad_norm": 0.1322142630815506, "learning_rate": 2.135095661516325e-07, "loss": 0.6332, "step": 2856 }, { "epoch": 2.3794037940379402, "grad_norm": 0.13864091038703918, "learning_rate": 2.1295504935752007e-07, "loss": 0.5943, "step": 2857 }, { "epoch": 2.3802376485303314, "grad_norm": 0.13323818147182465, "learning_rate": 2.1240116775974982e-07, "loss": 0.5768, "step": 2858 }, { "epoch": 2.3810715030227225, "grad_norm": 0.13591532409191132, "learning_rate": 2.1184792180534006e-07, "loss": 0.5839, "step": 2859 }, { "epoch": 2.3819053575151137, "grad_norm": 0.13431932032108307, "learning_rate": 2.1129531194079653e-07, "loss": 0.5859, "step": 2860 }, { "epoch": 2.382739212007505, "grad_norm": 0.12909963726997375, "learning_rate": 2.10743338612111e-07, "loss": 0.5506, "step": 2861 }, { "epoch": 2.3835730664998955, "grad_norm": 0.12649255990982056, "learning_rate": 2.1019200226476163e-07, "loss": 0.5855, "step": 2862 }, { "epoch": 2.3844069209922867, "grad_norm": 0.13748176395893097, "learning_rate": 2.096413033437131e-07, "loss": 0.5634, "step": 2863 }, { "epoch": 2.385240775484678, "grad_norm": 0.14034835994243622, "learning_rate": 2.090912422934148e-07, "loss": 0.6192, "step": 2864 }, { "epoch": 2.386074629977069, "grad_norm": 0.132245734333992, "learning_rate": 2.0854181955780182e-07, "loss": 0.5679, "step": 2865 }, { "epoch": 2.38690848446946, "grad_norm": 0.1347007155418396, "learning_rate": 2.079930355802939e-07, "loss": 0.5688, "step": 2866 }, { "epoch": 2.3877423389618513, "grad_norm": 0.13528944551944733, "learning_rate": 2.0744489080379503e-07, "loss": 0.6181, "step": 2867 }, { "epoch": 2.3885761934542424, "grad_norm": 0.13170389831066132, "learning_rate": 2.0689738567069416e-07, "loss": 0.5834, "step": 2868 }, { "epoch": 2.389410047946633, "grad_norm": 0.1397821456193924, "learning_rate": 2.063505206228632e-07, "loss": 0.5607, "step": 2869 }, { "epoch": 2.3902439024390243, "grad_norm": 0.13174912333488464, "learning_rate": 2.058042961016576e-07, "loss": 0.593, "step": 2870 }, { "epoch": 2.3910777569314154, "grad_norm": 0.1326981484889984, "learning_rate": 2.0525871254791627e-07, "loss": 0.637, "step": 2871 }, { "epoch": 2.3919116114238066, "grad_norm": 0.1353401243686676, "learning_rate": 2.0471377040196036e-07, "loss": 0.5314, "step": 2872 }, { "epoch": 2.3927454659161977, "grad_norm": 0.12800894677639008, "learning_rate": 2.0416947010359352e-07, "loss": 0.5818, "step": 2873 }, { "epoch": 2.393579320408589, "grad_norm": 0.13806089758872986, "learning_rate": 2.0362581209210129e-07, "loss": 0.6262, "step": 2874 }, { "epoch": 2.3944131749009796, "grad_norm": 0.13309136033058167, "learning_rate": 2.030827968062513e-07, "loss": 0.5718, "step": 2875 }, { "epoch": 2.3952470293933708, "grad_norm": 0.1417752504348755, "learning_rate": 2.0254042468429168e-07, "loss": 0.6006, "step": 2876 }, { "epoch": 2.396080883885762, "grad_norm": 0.1375095248222351, "learning_rate": 2.0199869616395238e-07, "loss": 0.6077, "step": 2877 }, { "epoch": 2.396914738378153, "grad_norm": 0.14154213666915894, "learning_rate": 2.0145761168244335e-07, "loss": 0.6331, "step": 2878 }, { "epoch": 2.397748592870544, "grad_norm": 0.135207399725914, "learning_rate": 2.009171716764547e-07, "loss": 0.6044, "step": 2879 }, { "epoch": 2.3985824473629354, "grad_norm": 0.1391584277153015, "learning_rate": 2.003773765821568e-07, "loss": 0.5592, "step": 2880 }, { "epoch": 2.399416301855326, "grad_norm": 0.13207468390464783, "learning_rate": 1.9983822683519912e-07, "loss": 0.6219, "step": 2881 }, { "epoch": 2.400250156347717, "grad_norm": 0.1367032378911972, "learning_rate": 1.9929972287071029e-07, "loss": 0.5719, "step": 2882 }, { "epoch": 2.4010840108401084, "grad_norm": 0.13686658442020416, "learning_rate": 1.987618651232985e-07, "loss": 0.5947, "step": 2883 }, { "epoch": 2.4019178653324995, "grad_norm": 0.12885454297065735, "learning_rate": 1.9822465402704958e-07, "loss": 0.5347, "step": 2884 }, { "epoch": 2.4027517198248907, "grad_norm": 0.13540036976337433, "learning_rate": 1.9768809001552767e-07, "loss": 0.5913, "step": 2885 }, { "epoch": 2.403585574317282, "grad_norm": 0.1323569118976593, "learning_rate": 1.9715217352177493e-07, "loss": 0.5551, "step": 2886 }, { "epoch": 2.4044194288096725, "grad_norm": 0.1385812908411026, "learning_rate": 1.9661690497831052e-07, "loss": 0.5905, "step": 2887 }, { "epoch": 2.4052532833020637, "grad_norm": 0.1258879005908966, "learning_rate": 1.9608228481713074e-07, "loss": 0.582, "step": 2888 }, { "epoch": 2.406087137794455, "grad_norm": 0.150419220328331, "learning_rate": 1.9554831346970923e-07, "loss": 0.6374, "step": 2889 }, { "epoch": 2.406920992286846, "grad_norm": 0.14132483303546906, "learning_rate": 1.9501499136699529e-07, "loss": 0.5928, "step": 2890 }, { "epoch": 2.407754846779237, "grad_norm": 0.13279718160629272, "learning_rate": 1.9448231893941414e-07, "loss": 0.5921, "step": 2891 }, { "epoch": 2.4085887012716283, "grad_norm": 0.13128960132598877, "learning_rate": 1.9395029661686757e-07, "loss": 0.6427, "step": 2892 }, { "epoch": 2.409422555764019, "grad_norm": 0.13115164637565613, "learning_rate": 1.934189248287319e-07, "loss": 0.5849, "step": 2893 }, { "epoch": 2.41025641025641, "grad_norm": 0.13133604824543, "learning_rate": 1.9288820400385852e-07, "loss": 0.5836, "step": 2894 }, { "epoch": 2.4110902647488013, "grad_norm": 0.13140904903411865, "learning_rate": 1.9235813457057359e-07, "loss": 0.5855, "step": 2895 }, { "epoch": 2.4119241192411924, "grad_norm": 0.12903335690498352, "learning_rate": 1.9182871695667746e-07, "loss": 0.6125, "step": 2896 }, { "epoch": 2.4127579737335836, "grad_norm": 0.1353691667318344, "learning_rate": 1.9129995158944478e-07, "loss": 0.6207, "step": 2897 }, { "epoch": 2.4135918282259747, "grad_norm": 0.13868772983551025, "learning_rate": 1.9077183889562331e-07, "loss": 0.5823, "step": 2898 }, { "epoch": 2.414425682718366, "grad_norm": 0.13506346940994263, "learning_rate": 1.9024437930143432e-07, "loss": 0.5523, "step": 2899 }, { "epoch": 2.4152595372107566, "grad_norm": 0.1333262026309967, "learning_rate": 1.8971757323257197e-07, "loss": 0.6106, "step": 2900 }, { "epoch": 2.4160933917031477, "grad_norm": 0.1286570131778717, "learning_rate": 1.8919142111420284e-07, "loss": 0.567, "step": 2901 }, { "epoch": 2.416927246195539, "grad_norm": 0.1309124082326889, "learning_rate": 1.8866592337096577e-07, "loss": 0.5119, "step": 2902 }, { "epoch": 2.41776110068793, "grad_norm": 0.13527123630046844, "learning_rate": 1.881410804269714e-07, "loss": 0.5983, "step": 2903 }, { "epoch": 2.418594955180321, "grad_norm": 0.13855566084384918, "learning_rate": 1.876168927058026e-07, "loss": 0.5794, "step": 2904 }, { "epoch": 2.419428809672712, "grad_norm": 0.13321010768413544, "learning_rate": 1.8709336063051216e-07, "loss": 0.5673, "step": 2905 }, { "epoch": 2.420262664165103, "grad_norm": 0.13595032691955566, "learning_rate": 1.8657048462362511e-07, "loss": 0.6538, "step": 2906 }, { "epoch": 2.421096518657494, "grad_norm": 0.13007234036922455, "learning_rate": 1.8604826510713612e-07, "loss": 0.5694, "step": 2907 }, { "epoch": 2.4219303731498854, "grad_norm": 0.1420540064573288, "learning_rate": 1.8552670250251002e-07, "loss": 0.5982, "step": 2908 }, { "epoch": 2.4227642276422765, "grad_norm": 0.1407056301832199, "learning_rate": 1.8500579723068176e-07, "loss": 0.5575, "step": 2909 }, { "epoch": 2.4235980821346677, "grad_norm": 0.14242278039455414, "learning_rate": 1.8448554971205577e-07, "loss": 0.5821, "step": 2910 }, { "epoch": 2.424431936627059, "grad_norm": 0.132111057639122, "learning_rate": 1.8396596036650513e-07, "loss": 0.5865, "step": 2911 }, { "epoch": 2.4252657911194495, "grad_norm": 0.13183192908763885, "learning_rate": 1.834470296133729e-07, "loss": 0.5977, "step": 2912 }, { "epoch": 2.4260996456118407, "grad_norm": 0.13186009228229523, "learning_rate": 1.8292875787146945e-07, "loss": 0.5865, "step": 2913 }, { "epoch": 2.426933500104232, "grad_norm": 0.14010071754455566, "learning_rate": 1.8241114555907367e-07, "loss": 0.6115, "step": 2914 }, { "epoch": 2.427767354596623, "grad_norm": 0.12944266200065613, "learning_rate": 1.818941930939324e-07, "loss": 0.5411, "step": 2915 }, { "epoch": 2.428601209089014, "grad_norm": 0.1330813467502594, "learning_rate": 1.8137790089325967e-07, "loss": 0.5943, "step": 2916 }, { "epoch": 2.429435063581405, "grad_norm": 0.1295856535434723, "learning_rate": 1.808622693737367e-07, "loss": 0.6727, "step": 2917 }, { "epoch": 2.430268918073796, "grad_norm": 0.1257857233285904, "learning_rate": 1.8034729895151202e-07, "loss": 0.5727, "step": 2918 }, { "epoch": 2.431102772566187, "grad_norm": 0.13976608216762543, "learning_rate": 1.7983299004219999e-07, "loss": 0.6404, "step": 2919 }, { "epoch": 2.4319366270585783, "grad_norm": 0.12914389371871948, "learning_rate": 1.793193430608808e-07, "loss": 0.5684, "step": 2920 }, { "epoch": 2.4327704815509694, "grad_norm": 0.13092781603336334, "learning_rate": 1.7880635842210169e-07, "loss": 0.6637, "step": 2921 }, { "epoch": 2.4336043360433606, "grad_norm": 0.13849668204784393, "learning_rate": 1.7829403653987407e-07, "loss": 0.5713, "step": 2922 }, { "epoch": 2.4344381905357517, "grad_norm": 0.13813416659832, "learning_rate": 1.7778237782767503e-07, "loss": 0.594, "step": 2923 }, { "epoch": 2.4352720450281424, "grad_norm": 0.12659841775894165, "learning_rate": 1.7727138269844633e-07, "loss": 0.5589, "step": 2924 }, { "epoch": 2.4361058995205336, "grad_norm": 0.14138655364513397, "learning_rate": 1.7676105156459397e-07, "loss": 0.6563, "step": 2925 }, { "epoch": 2.4369397540129247, "grad_norm": 0.12977050244808197, "learning_rate": 1.7625138483798872e-07, "loss": 0.6377, "step": 2926 }, { "epoch": 2.437773608505316, "grad_norm": 0.13506102561950684, "learning_rate": 1.7574238292996456e-07, "loss": 0.6006, "step": 2927 }, { "epoch": 2.438607462997707, "grad_norm": 0.13082827627658844, "learning_rate": 1.752340462513191e-07, "loss": 0.5117, "step": 2928 }, { "epoch": 2.4394413174900977, "grad_norm": 0.16338416934013367, "learning_rate": 1.7472637521231281e-07, "loss": 0.5716, "step": 2929 }, { "epoch": 2.440275171982489, "grad_norm": 0.13927298784255981, "learning_rate": 1.7421937022266918e-07, "loss": 0.5886, "step": 2930 }, { "epoch": 2.44110902647488, "grad_norm": 0.134609654545784, "learning_rate": 1.7371303169157437e-07, "loss": 0.5797, "step": 2931 }, { "epoch": 2.441942880967271, "grad_norm": 0.13286033272743225, "learning_rate": 1.73207360027676e-07, "loss": 0.569, "step": 2932 }, { "epoch": 2.4427767354596623, "grad_norm": 0.13371062278747559, "learning_rate": 1.7270235563908442e-07, "loss": 0.6039, "step": 2933 }, { "epoch": 2.4436105899520535, "grad_norm": 0.13399933278560638, "learning_rate": 1.721980189333707e-07, "loss": 0.5456, "step": 2934 }, { "epoch": 2.4444444444444446, "grad_norm": 0.13598956167697906, "learning_rate": 1.7169435031756708e-07, "loss": 0.572, "step": 2935 }, { "epoch": 2.4452782989368353, "grad_norm": 0.13800416886806488, "learning_rate": 1.711913501981672e-07, "loss": 0.5252, "step": 2936 }, { "epoch": 2.4461121534292265, "grad_norm": 0.1314326971769333, "learning_rate": 1.7068901898112476e-07, "loss": 0.6203, "step": 2937 }, { "epoch": 2.4469460079216176, "grad_norm": 0.134601429104805, "learning_rate": 1.701873570718534e-07, "loss": 0.5573, "step": 2938 }, { "epoch": 2.447779862414009, "grad_norm": 0.14564085006713867, "learning_rate": 1.6968636487522702e-07, "loss": 0.6396, "step": 2939 }, { "epoch": 2.4486137169064, "grad_norm": 0.13601054251194, "learning_rate": 1.6918604279557846e-07, "loss": 0.6431, "step": 2940 }, { "epoch": 2.449447571398791, "grad_norm": 0.14048826694488525, "learning_rate": 1.686863912367006e-07, "loss": 0.521, "step": 2941 }, { "epoch": 2.450281425891182, "grad_norm": 0.13788892328739166, "learning_rate": 1.681874106018445e-07, "loss": 0.5965, "step": 2942 }, { "epoch": 2.451115280383573, "grad_norm": 0.14097999036312103, "learning_rate": 1.6768910129371983e-07, "loss": 0.6377, "step": 2943 }, { "epoch": 2.451949134875964, "grad_norm": 0.132716566324234, "learning_rate": 1.6719146371449467e-07, "loss": 0.548, "step": 2944 }, { "epoch": 2.4527829893683553, "grad_norm": 0.12886781990528107, "learning_rate": 1.6669449826579463e-07, "loss": 0.5821, "step": 2945 }, { "epoch": 2.4536168438607464, "grad_norm": 0.1308266818523407, "learning_rate": 1.66198205348703e-07, "loss": 0.574, "step": 2946 }, { "epoch": 2.4544506983531376, "grad_norm": 0.13480325043201447, "learning_rate": 1.657025853637608e-07, "loss": 0.55, "step": 2947 }, { "epoch": 2.4552845528455283, "grad_norm": 0.12624162435531616, "learning_rate": 1.6520763871096522e-07, "loss": 0.5327, "step": 2948 }, { "epoch": 2.4561184073379194, "grad_norm": 0.1327810138463974, "learning_rate": 1.6471336578977013e-07, "loss": 0.622, "step": 2949 }, { "epoch": 2.4569522618303106, "grad_norm": 0.12912337481975555, "learning_rate": 1.6421976699908634e-07, "loss": 0.605, "step": 2950 }, { "epoch": 2.4577861163227017, "grad_norm": 0.14189478754997253, "learning_rate": 1.6372684273727987e-07, "loss": 0.6006, "step": 2951 }, { "epoch": 2.458619970815093, "grad_norm": 0.1329701840877533, "learning_rate": 1.6323459340217249e-07, "loss": 0.6053, "step": 2952 }, { "epoch": 2.459453825307484, "grad_norm": 0.1286649852991104, "learning_rate": 1.627430193910414e-07, "loss": 0.5873, "step": 2953 }, { "epoch": 2.460287679799875, "grad_norm": 0.135244220495224, "learning_rate": 1.622521211006186e-07, "loss": 0.6211, "step": 2954 }, { "epoch": 2.461121534292266, "grad_norm": 0.12673702836036682, "learning_rate": 1.6176189892709124e-07, "loss": 0.6082, "step": 2955 }, { "epoch": 2.461955388784657, "grad_norm": 0.13566385209560394, "learning_rate": 1.612723532661001e-07, "loss": 0.6062, "step": 2956 }, { "epoch": 2.462789243277048, "grad_norm": 0.13710834085941315, "learning_rate": 1.607834845127405e-07, "loss": 0.6035, "step": 2957 }, { "epoch": 2.4636230977694393, "grad_norm": 0.14074556529521942, "learning_rate": 1.6029529306156098e-07, "loss": 0.6363, "step": 2958 }, { "epoch": 2.4644569522618305, "grad_norm": 0.13592229783535004, "learning_rate": 1.5980777930656398e-07, "loss": 0.6648, "step": 2959 }, { "epoch": 2.465290806754221, "grad_norm": 0.13886703550815582, "learning_rate": 1.5932094364120453e-07, "loss": 0.6431, "step": 2960 }, { "epoch": 2.4661246612466123, "grad_norm": 0.12988947331905365, "learning_rate": 1.5883478645839043e-07, "loss": 0.6615, "step": 2961 }, { "epoch": 2.4669585157390035, "grad_norm": 0.132481649518013, "learning_rate": 1.5834930815048265e-07, "loss": 0.6239, "step": 2962 }, { "epoch": 2.4677923702313946, "grad_norm": 0.13364684581756592, "learning_rate": 1.5786450910929328e-07, "loss": 0.5614, "step": 2963 }, { "epoch": 2.468626224723786, "grad_norm": 0.13308215141296387, "learning_rate": 1.5738038972608647e-07, "loss": 0.6099, "step": 2964 }, { "epoch": 2.469460079216177, "grad_norm": 0.14067746698856354, "learning_rate": 1.5689695039157847e-07, "loss": 0.6513, "step": 2965 }, { "epoch": 2.470293933708568, "grad_norm": 0.13405707478523254, "learning_rate": 1.5641419149593582e-07, "loss": 0.6023, "step": 2966 }, { "epoch": 2.471127788200959, "grad_norm": 0.13046559691429138, "learning_rate": 1.5593211342877643e-07, "loss": 0.6202, "step": 2967 }, { "epoch": 2.47196164269335, "grad_norm": 0.13105064630508423, "learning_rate": 1.5545071657916842e-07, "loss": 0.5896, "step": 2968 }, { "epoch": 2.472795497185741, "grad_norm": 0.13460539281368256, "learning_rate": 1.549700013356302e-07, "loss": 0.5834, "step": 2969 }, { "epoch": 2.4736293516781322, "grad_norm": 0.13069799542427063, "learning_rate": 1.5448996808613056e-07, "loss": 0.5492, "step": 2970 }, { "epoch": 2.4744632061705234, "grad_norm": 0.13145001232624054, "learning_rate": 1.5401061721808727e-07, "loss": 0.4951, "step": 2971 }, { "epoch": 2.475297060662914, "grad_norm": 0.13705606758594513, "learning_rate": 1.5353194911836754e-07, "loss": 0.5947, "step": 2972 }, { "epoch": 2.4761309151553053, "grad_norm": 0.13895384967327118, "learning_rate": 1.5305396417328753e-07, "loss": 0.5912, "step": 2973 }, { "epoch": 2.4769647696476964, "grad_norm": 0.13812367618083954, "learning_rate": 1.5257666276861225e-07, "loss": 0.6273, "step": 2974 }, { "epoch": 2.4777986241400876, "grad_norm": 0.1422661542892456, "learning_rate": 1.5210004528955466e-07, "loss": 0.547, "step": 2975 }, { "epoch": 2.4786324786324787, "grad_norm": 0.12845875322818756, "learning_rate": 1.516241121207763e-07, "loss": 0.5784, "step": 2976 }, { "epoch": 2.47946633312487, "grad_norm": 0.15125256776809692, "learning_rate": 1.5114886364638612e-07, "loss": 0.6543, "step": 2977 }, { "epoch": 2.480300187617261, "grad_norm": 0.13215088844299316, "learning_rate": 1.5067430024993999e-07, "loss": 0.5972, "step": 2978 }, { "epoch": 2.4811340421096517, "grad_norm": 0.14214058220386505, "learning_rate": 1.5020042231444195e-07, "loss": 0.6396, "step": 2979 }, { "epoch": 2.481967896602043, "grad_norm": 0.1361965388059616, "learning_rate": 1.4972723022234211e-07, "loss": 0.5897, "step": 2980 }, { "epoch": 2.482801751094434, "grad_norm": 0.13309814035892487, "learning_rate": 1.49254724355537e-07, "loss": 0.5765, "step": 2981 }, { "epoch": 2.483635605586825, "grad_norm": 0.13403445482254028, "learning_rate": 1.4878290509536963e-07, "loss": 0.6108, "step": 2982 }, { "epoch": 2.4844694600792163, "grad_norm": 0.14112424850463867, "learning_rate": 1.4831177282262842e-07, "loss": 0.5783, "step": 2983 }, { "epoch": 2.485303314571607, "grad_norm": 0.14206255972385406, "learning_rate": 1.4784132791754834e-07, "loss": 0.6168, "step": 2984 }, { "epoch": 2.486137169063998, "grad_norm": 0.13668708503246307, "learning_rate": 1.4737157075980844e-07, "loss": 0.5941, "step": 2985 }, { "epoch": 2.4869710235563893, "grad_norm": 0.13575045764446259, "learning_rate": 1.4690250172853348e-07, "loss": 0.544, "step": 2986 }, { "epoch": 2.4878048780487805, "grad_norm": 0.14906087517738342, "learning_rate": 1.464341212022926e-07, "loss": 0.6252, "step": 2987 }, { "epoch": 2.4886387325411716, "grad_norm": 0.12553520500659943, "learning_rate": 1.459664295590991e-07, "loss": 0.5561, "step": 2988 }, { "epoch": 2.489472587033563, "grad_norm": 0.13482603430747986, "learning_rate": 1.454994271764105e-07, "loss": 0.6109, "step": 2989 }, { "epoch": 2.490306441525954, "grad_norm": 0.14026843011379242, "learning_rate": 1.4503311443112798e-07, "loss": 0.6212, "step": 2990 }, { "epoch": 2.4911402960183446, "grad_norm": 0.13011108338832855, "learning_rate": 1.4456749169959647e-07, "loss": 0.5732, "step": 2991 }, { "epoch": 2.491974150510736, "grad_norm": 0.13201136887073517, "learning_rate": 1.441025593576035e-07, "loss": 0.5915, "step": 2992 }, { "epoch": 2.492808005003127, "grad_norm": 0.13688592612743378, "learning_rate": 1.436383177803796e-07, "loss": 0.598, "step": 2993 }, { "epoch": 2.493641859495518, "grad_norm": 0.13714109361171722, "learning_rate": 1.4317476734259802e-07, "loss": 0.622, "step": 2994 }, { "epoch": 2.4944757139879092, "grad_norm": 0.13858246803283691, "learning_rate": 1.42711908418374e-07, "loss": 0.5177, "step": 2995 }, { "epoch": 2.4953095684803, "grad_norm": 0.13396722078323364, "learning_rate": 1.4224974138126468e-07, "loss": 0.6211, "step": 2996 }, { "epoch": 2.496143422972691, "grad_norm": 0.13308130204677582, "learning_rate": 1.417882666042689e-07, "loss": 0.5512, "step": 2997 }, { "epoch": 2.4969772774650822, "grad_norm": 0.13575038313865662, "learning_rate": 1.4132748445982633e-07, "loss": 0.5588, "step": 2998 }, { "epoch": 2.4978111319574734, "grad_norm": 0.1262865513563156, "learning_rate": 1.4086739531981883e-07, "loss": 0.5405, "step": 2999 }, { "epoch": 2.4986449864498645, "grad_norm": 0.1360829621553421, "learning_rate": 1.404079995555677e-07, "loss": 0.6311, "step": 3000 }, { "epoch": 2.4994788409422557, "grad_norm": 0.11881641298532486, "learning_rate": 1.3994929753783513e-07, "loss": 0.5152, "step": 3001 }, { "epoch": 2.500312695434647, "grad_norm": 0.1360657513141632, "learning_rate": 1.3949128963682354e-07, "loss": 0.6054, "step": 3002 }, { "epoch": 2.5011465499270376, "grad_norm": 0.14240588247776031, "learning_rate": 1.3903397622217507e-07, "loss": 0.628, "step": 3003 }, { "epoch": 2.5019804044194287, "grad_norm": 0.1373392790555954, "learning_rate": 1.3857735766297108e-07, "loss": 0.5672, "step": 3004 }, { "epoch": 2.50281425891182, "grad_norm": 0.13722510635852814, "learning_rate": 1.3812143432773238e-07, "loss": 0.6085, "step": 3005 }, { "epoch": 2.503648113404211, "grad_norm": 0.12611958384513855, "learning_rate": 1.3766620658441897e-07, "loss": 0.6001, "step": 3006 }, { "epoch": 2.504481967896602, "grad_norm": 0.13417090475559235, "learning_rate": 1.3721167480042883e-07, "loss": 0.6345, "step": 3007 }, { "epoch": 2.505315822388993, "grad_norm": 0.13315609097480774, "learning_rate": 1.36757839342599e-07, "loss": 0.5633, "step": 3008 }, { "epoch": 2.5061496768813845, "grad_norm": 0.13062290847301483, "learning_rate": 1.36304700577204e-07, "loss": 0.6114, "step": 3009 }, { "epoch": 2.506983531373775, "grad_norm": 0.14047642052173615, "learning_rate": 1.358522588699561e-07, "loss": 0.593, "step": 3010 }, { "epoch": 2.5078173858661663, "grad_norm": 0.1368640661239624, "learning_rate": 1.354005145860052e-07, "loss": 0.6204, "step": 3011 }, { "epoch": 2.5086512403585575, "grad_norm": 0.13820619881153107, "learning_rate": 1.3494946808993801e-07, "loss": 0.6287, "step": 3012 }, { "epoch": 2.5094850948509486, "grad_norm": 0.1320614367723465, "learning_rate": 1.3449911974577876e-07, "loss": 0.5522, "step": 3013 }, { "epoch": 2.5103189493433398, "grad_norm": 0.1310126632452011, "learning_rate": 1.340494699169875e-07, "loss": 0.6068, "step": 3014 }, { "epoch": 2.5111528038357305, "grad_norm": 0.13907015323638916, "learning_rate": 1.3360051896646084e-07, "loss": 0.5757, "step": 3015 }, { "epoch": 2.5119866583281216, "grad_norm": 0.1383555829524994, "learning_rate": 1.3315226725653138e-07, "loss": 0.5866, "step": 3016 }, { "epoch": 2.5128205128205128, "grad_norm": 0.13155722618103027, "learning_rate": 1.3270471514896741e-07, "loss": 0.5508, "step": 3017 }, { "epoch": 2.513654367312904, "grad_norm": 0.13946841657161713, "learning_rate": 1.3225786300497232e-07, "loss": 0.6024, "step": 3018 }, { "epoch": 2.514488221805295, "grad_norm": 0.1311018317937851, "learning_rate": 1.3181171118518464e-07, "loss": 0.6378, "step": 3019 }, { "epoch": 2.515322076297686, "grad_norm": 0.13334567844867706, "learning_rate": 1.3136626004967833e-07, "loss": 0.5973, "step": 3020 }, { "epoch": 2.5161559307900774, "grad_norm": 0.1304926574230194, "learning_rate": 1.3092150995796113e-07, "loss": 0.5582, "step": 3021 }, { "epoch": 2.516989785282468, "grad_norm": 0.1310420036315918, "learning_rate": 1.3047746126897496e-07, "loss": 0.6132, "step": 3022 }, { "epoch": 2.5178236397748592, "grad_norm": 0.1336178183555603, "learning_rate": 1.3003411434109646e-07, "loss": 0.5149, "step": 3023 }, { "epoch": 2.5186574942672504, "grad_norm": 0.1330386996269226, "learning_rate": 1.2959146953213497e-07, "loss": 0.6625, "step": 3024 }, { "epoch": 2.5194913487596415, "grad_norm": 0.14059826731681824, "learning_rate": 1.2914952719933369e-07, "loss": 0.5979, "step": 3025 }, { "epoch": 2.5203252032520327, "grad_norm": 0.1363728940486908, "learning_rate": 1.287082876993687e-07, "loss": 0.6561, "step": 3026 }, { "epoch": 2.5211590577444234, "grad_norm": 0.13385115563869476, "learning_rate": 1.282677513883489e-07, "loss": 0.5837, "step": 3027 }, { "epoch": 2.5219929122368145, "grad_norm": 0.13265737891197205, "learning_rate": 1.2782791862181575e-07, "loss": 0.6232, "step": 3028 }, { "epoch": 2.5228267667292057, "grad_norm": 0.14071135222911835, "learning_rate": 1.2738878975474286e-07, "loss": 0.5758, "step": 3029 }, { "epoch": 2.523660621221597, "grad_norm": 0.14203928411006927, "learning_rate": 1.2695036514153567e-07, "loss": 0.6038, "step": 3030 }, { "epoch": 2.524494475713988, "grad_norm": 0.13652530312538147, "learning_rate": 1.2651264513603132e-07, "loss": 0.5693, "step": 3031 }, { "epoch": 2.525328330206379, "grad_norm": 0.14524352550506592, "learning_rate": 1.2607563009149825e-07, "loss": 0.5795, "step": 3032 }, { "epoch": 2.5261621846987703, "grad_norm": 0.13484011590480804, "learning_rate": 1.2563932036063584e-07, "loss": 0.5782, "step": 3033 }, { "epoch": 2.526996039191161, "grad_norm": 0.14854557812213898, "learning_rate": 1.2520371629557425e-07, "loss": 0.6024, "step": 3034 }, { "epoch": 2.527829893683552, "grad_norm": 0.13873372972011566, "learning_rate": 1.2476881824787466e-07, "loss": 0.5507, "step": 3035 }, { "epoch": 2.5286637481759433, "grad_norm": 0.1351420283317566, "learning_rate": 1.2433462656852777e-07, "loss": 0.6358, "step": 3036 }, { "epoch": 2.5294976026683345, "grad_norm": 0.13378125429153442, "learning_rate": 1.2390114160795417e-07, "loss": 0.5873, "step": 3037 }, { "epoch": 2.5303314571607256, "grad_norm": 0.1382291316986084, "learning_rate": 1.2346836371600478e-07, "loss": 0.6191, "step": 3038 }, { "epoch": 2.5311653116531163, "grad_norm": 0.1428293138742447, "learning_rate": 1.230362932419594e-07, "loss": 0.567, "step": 3039 }, { "epoch": 2.5319991661455075, "grad_norm": 0.13339033722877502, "learning_rate": 1.226049305345267e-07, "loss": 0.523, "step": 3040 }, { "epoch": 2.5328330206378986, "grad_norm": 0.1312277466058731, "learning_rate": 1.2217427594184459e-07, "loss": 0.5382, "step": 3041 }, { "epoch": 2.5336668751302898, "grad_norm": 0.13031668961048126, "learning_rate": 1.2174432981147886e-07, "loss": 0.6027, "step": 3042 }, { "epoch": 2.534500729622681, "grad_norm": 0.14053155481815338, "learning_rate": 1.213150924904245e-07, "loss": 0.6168, "step": 3043 }, { "epoch": 2.535334584115072, "grad_norm": 0.14308233559131622, "learning_rate": 1.2088656432510358e-07, "loss": 0.6001, "step": 3044 }, { "epoch": 2.536168438607463, "grad_norm": 0.13693615794181824, "learning_rate": 1.2045874566136617e-07, "loss": 0.5843, "step": 3045 }, { "epoch": 2.537002293099854, "grad_norm": 0.1373409479856491, "learning_rate": 1.2003163684448968e-07, "loss": 0.6172, "step": 3046 }, { "epoch": 2.537836147592245, "grad_norm": 0.1251070201396942, "learning_rate": 1.1960523821917868e-07, "loss": 0.5047, "step": 3047 }, { "epoch": 2.5386700020846362, "grad_norm": 0.1326339989900589, "learning_rate": 1.1917955012956427e-07, "loss": 0.555, "step": 3048 }, { "epoch": 2.5395038565770274, "grad_norm": 0.12611235678195953, "learning_rate": 1.1875457291920476e-07, "loss": 0.571, "step": 3049 }, { "epoch": 2.5403377110694185, "grad_norm": 0.13283178210258484, "learning_rate": 1.1833030693108425e-07, "loss": 0.566, "step": 3050 }, { "epoch": 2.5411715655618092, "grad_norm": 0.13543598353862762, "learning_rate": 1.1790675250761262e-07, "loss": 0.628, "step": 3051 }, { "epoch": 2.5420054200542004, "grad_norm": 0.13329797983169556, "learning_rate": 1.1748390999062607e-07, "loss": 0.6256, "step": 3052 }, { "epoch": 2.5428392745465915, "grad_norm": 0.13540299236774445, "learning_rate": 1.1706177972138598e-07, "loss": 0.5603, "step": 3053 }, { "epoch": 2.5436731290389827, "grad_norm": 0.13842473924160004, "learning_rate": 1.166403620405787e-07, "loss": 0.592, "step": 3054 }, { "epoch": 2.544506983531374, "grad_norm": 0.1399260014295578, "learning_rate": 1.1621965728831562e-07, "loss": 0.6008, "step": 3055 }, { "epoch": 2.545340838023765, "grad_norm": 0.13631440699100494, "learning_rate": 1.157996658041327e-07, "loss": 0.5659, "step": 3056 }, { "epoch": 2.546174692516156, "grad_norm": 0.16495107114315033, "learning_rate": 1.1538038792699068e-07, "loss": 0.6093, "step": 3057 }, { "epoch": 2.547008547008547, "grad_norm": 0.14227885007858276, "learning_rate": 1.1496182399527365e-07, "loss": 0.6076, "step": 3058 }, { "epoch": 2.547842401500938, "grad_norm": 0.1316503882408142, "learning_rate": 1.145439743467902e-07, "loss": 0.5826, "step": 3059 }, { "epoch": 2.548676255993329, "grad_norm": 0.13284341990947723, "learning_rate": 1.1412683931877176e-07, "loss": 0.6255, "step": 3060 }, { "epoch": 2.5495101104857203, "grad_norm": 0.1408301591873169, "learning_rate": 1.1371041924787362e-07, "loss": 0.6103, "step": 3061 }, { "epoch": 2.5503439649781114, "grad_norm": 0.1347634494304657, "learning_rate": 1.1329471447017358e-07, "loss": 0.5807, "step": 3062 }, { "epoch": 2.551177819470502, "grad_norm": 0.13598300516605377, "learning_rate": 1.1287972532117229e-07, "loss": 0.6578, "step": 3063 }, { "epoch": 2.5520116739628937, "grad_norm": 0.1328735053539276, "learning_rate": 1.1246545213579339e-07, "loss": 0.549, "step": 3064 }, { "epoch": 2.5528455284552845, "grad_norm": 0.1336262822151184, "learning_rate": 1.1205189524838188e-07, "loss": 0.5827, "step": 3065 }, { "epoch": 2.5536793829476756, "grad_norm": 0.14055649936199188, "learning_rate": 1.1163905499270487e-07, "loss": 0.6581, "step": 3066 }, { "epoch": 2.5545132374400668, "grad_norm": 0.13607151806354523, "learning_rate": 1.1122693170195163e-07, "loss": 0.6446, "step": 3067 }, { "epoch": 2.555347091932458, "grad_norm": 0.13129949569702148, "learning_rate": 1.108155257087322e-07, "loss": 0.5594, "step": 3068 }, { "epoch": 2.556180946424849, "grad_norm": 0.13824865221977234, "learning_rate": 1.1040483734507788e-07, "loss": 0.5606, "step": 3069 }, { "epoch": 2.5570148009172398, "grad_norm": 0.13617265224456787, "learning_rate": 1.0999486694244086e-07, "loss": 0.6431, "step": 3070 }, { "epoch": 2.557848655409631, "grad_norm": 0.13965849578380585, "learning_rate": 1.0958561483169359e-07, "loss": 0.6381, "step": 3071 }, { "epoch": 2.558682509902022, "grad_norm": 0.13711613416671753, "learning_rate": 1.0917708134312963e-07, "loss": 0.6311, "step": 3072 }, { "epoch": 2.559516364394413, "grad_norm": 0.128121018409729, "learning_rate": 1.0876926680646159e-07, "loss": 0.5069, "step": 3073 }, { "epoch": 2.5603502188868044, "grad_norm": 0.139384925365448, "learning_rate": 1.0836217155082251e-07, "loss": 0.587, "step": 3074 }, { "epoch": 2.561184073379195, "grad_norm": 0.1332816481590271, "learning_rate": 1.0795579590476445e-07, "loss": 0.6268, "step": 3075 }, { "epoch": 2.5620179278715867, "grad_norm": 0.14762350916862488, "learning_rate": 1.0755014019625896e-07, "loss": 0.6432, "step": 3076 }, { "epoch": 2.5628517823639774, "grad_norm": 0.13618019223213196, "learning_rate": 1.071452047526965e-07, "loss": 0.5556, "step": 3077 }, { "epoch": 2.5636856368563685, "grad_norm": 0.13517871499061584, "learning_rate": 1.0674098990088654e-07, "loss": 0.5656, "step": 3078 }, { "epoch": 2.5645194913487597, "grad_norm": 0.12663470208644867, "learning_rate": 1.0633749596705643e-07, "loss": 0.5957, "step": 3079 }, { "epoch": 2.565353345841151, "grad_norm": 0.1300160437822342, "learning_rate": 1.0593472327685182e-07, "loss": 0.5604, "step": 3080 }, { "epoch": 2.566187200333542, "grad_norm": 0.13851061463356018, "learning_rate": 1.0553267215533679e-07, "loss": 0.552, "step": 3081 }, { "epoch": 2.5670210548259327, "grad_norm": 0.12825097143650055, "learning_rate": 1.0513134292699243e-07, "loss": 0.5441, "step": 3082 }, { "epoch": 2.567854909318324, "grad_norm": 0.13497650623321533, "learning_rate": 1.0473073591571757e-07, "loss": 0.589, "step": 3083 }, { "epoch": 2.568688763810715, "grad_norm": 0.14014704525470734, "learning_rate": 1.04330851444828e-07, "loss": 0.5904, "step": 3084 }, { "epoch": 2.569522618303106, "grad_norm": 0.13393735885620117, "learning_rate": 1.0393168983705625e-07, "loss": 0.5943, "step": 3085 }, { "epoch": 2.5703564727954973, "grad_norm": 0.13364675641059875, "learning_rate": 1.0353325141455205e-07, "loss": 0.6188, "step": 3086 }, { "epoch": 2.571190327287888, "grad_norm": 0.14985555410385132, "learning_rate": 1.0313553649888074e-07, "loss": 0.6061, "step": 3087 }, { "epoch": 2.5720241817802796, "grad_norm": 0.13277341425418854, "learning_rate": 1.0273854541102423e-07, "loss": 0.6115, "step": 3088 }, { "epoch": 2.5728580362726703, "grad_norm": 0.13757126033306122, "learning_rate": 1.023422784713801e-07, "loss": 0.5883, "step": 3089 }, { "epoch": 2.5736918907650614, "grad_norm": 0.13987593352794647, "learning_rate": 1.0194673599976134e-07, "loss": 0.6483, "step": 3090 }, { "epoch": 2.5745257452574526, "grad_norm": 0.13018371164798737, "learning_rate": 1.0155191831539644e-07, "loss": 0.5913, "step": 3091 }, { "epoch": 2.5753595997498437, "grad_norm": 0.13608433306217194, "learning_rate": 1.0115782573692866e-07, "loss": 0.5576, "step": 3092 }, { "epoch": 2.576193454242235, "grad_norm": 0.1433640569448471, "learning_rate": 1.0076445858241678e-07, "loss": 0.6182, "step": 3093 }, { "epoch": 2.5770273087346256, "grad_norm": 0.14163893461227417, "learning_rate": 1.003718171693333e-07, "loss": 0.5578, "step": 3094 }, { "epoch": 2.5778611632270167, "grad_norm": 0.14300225675106049, "learning_rate": 9.997990181456529e-08, "loss": 0.5916, "step": 3095 }, { "epoch": 2.578695017719408, "grad_norm": 0.1443438082933426, "learning_rate": 9.958871283441406e-08, "loss": 0.611, "step": 3096 }, { "epoch": 2.579528872211799, "grad_norm": 0.13358335196971893, "learning_rate": 9.919825054459441e-08, "loss": 0.5854, "step": 3097 }, { "epoch": 2.58036272670419, "grad_norm": 0.1267537772655487, "learning_rate": 9.880851526023471e-08, "loss": 0.6017, "step": 3098 }, { "epoch": 2.5811965811965814, "grad_norm": 0.13596156239509583, "learning_rate": 9.841950729587667e-08, "loss": 0.5797, "step": 3099 }, { "epoch": 2.5820304356889725, "grad_norm": 0.12967713177204132, "learning_rate": 9.803122696547473e-08, "loss": 0.5831, "step": 3100 }, { "epoch": 2.582864290181363, "grad_norm": 0.13567784428596497, "learning_rate": 9.764367458239675e-08, "loss": 0.6173, "step": 3101 }, { "epoch": 2.5836981446737544, "grad_norm": 0.13866503536701202, "learning_rate": 9.725685045942234e-08, "loss": 0.6123, "step": 3102 }, { "epoch": 2.5845319991661455, "grad_norm": 0.13681918382644653, "learning_rate": 9.687075490874375e-08, "loss": 0.6131, "step": 3103 }, { "epoch": 2.5853658536585367, "grad_norm": 0.1418786495923996, "learning_rate": 9.648538824196518e-08, "loss": 0.5703, "step": 3104 }, { "epoch": 2.586199708150928, "grad_norm": 0.13041797280311584, "learning_rate": 9.610075077010239e-08, "loss": 0.5712, "step": 3105 }, { "epoch": 2.5870335626433185, "grad_norm": 0.13490214943885803, "learning_rate": 9.571684280358272e-08, "loss": 0.525, "step": 3106 }, { "epoch": 2.5878674171357097, "grad_norm": 0.13328233361244202, "learning_rate": 9.533366465224513e-08, "loss": 0.6147, "step": 3107 }, { "epoch": 2.588701271628101, "grad_norm": 0.13110201060771942, "learning_rate": 9.495121662533911e-08, "loss": 0.5273, "step": 3108 }, { "epoch": 2.589535126120492, "grad_norm": 0.13410116732120514, "learning_rate": 9.456949903152478e-08, "loss": 0.6143, "step": 3109 }, { "epoch": 2.590368980612883, "grad_norm": 0.1353457123041153, "learning_rate": 9.418851217887336e-08, "loss": 0.6355, "step": 3110 }, { "epoch": 2.5912028351052743, "grad_norm": 0.12921103835105896, "learning_rate": 9.38082563748659e-08, "loss": 0.6383, "step": 3111 }, { "epoch": 2.5920366895976654, "grad_norm": 0.14108790457248688, "learning_rate": 9.342873192639345e-08, "loss": 0.561, "step": 3112 }, { "epoch": 2.592870544090056, "grad_norm": 0.13146281242370605, "learning_rate": 9.304993913975679e-08, "loss": 0.5661, "step": 3113 }, { "epoch": 2.5937043985824473, "grad_norm": 0.14770254492759705, "learning_rate": 9.267187832066614e-08, "loss": 0.5737, "step": 3114 }, { "epoch": 2.5945382530748384, "grad_norm": 0.13602465391159058, "learning_rate": 9.229454977424156e-08, "loss": 0.546, "step": 3115 }, { "epoch": 2.5953721075672296, "grad_norm": 0.1373216062784195, "learning_rate": 9.191795380501133e-08, "loss": 0.6068, "step": 3116 }, { "epoch": 2.5962059620596207, "grad_norm": 0.13257469236850739, "learning_rate": 9.154209071691288e-08, "loss": 0.5375, "step": 3117 }, { "epoch": 2.5970398165520114, "grad_norm": 0.13821159303188324, "learning_rate": 9.116696081329222e-08, "loss": 0.6067, "step": 3118 }, { "epoch": 2.5978736710444026, "grad_norm": 0.13279394805431366, "learning_rate": 9.079256439690352e-08, "loss": 0.5351, "step": 3119 }, { "epoch": 2.5987075255367937, "grad_norm": 0.13164344429969788, "learning_rate": 9.041890176990874e-08, "loss": 0.5398, "step": 3120 }, { "epoch": 2.599541380029185, "grad_norm": 0.145656019449234, "learning_rate": 9.004597323387797e-08, "loss": 0.5793, "step": 3121 }, { "epoch": 2.600375234521576, "grad_norm": 0.13412456214427948, "learning_rate": 8.967377908978902e-08, "loss": 0.6407, "step": 3122 }, { "epoch": 2.601209089013967, "grad_norm": 0.13084304332733154, "learning_rate": 8.930231963802637e-08, "loss": 0.5599, "step": 3123 }, { "epoch": 2.6020429435063583, "grad_norm": 0.1360272616147995, "learning_rate": 8.893159517838189e-08, "loss": 0.5474, "step": 3124 }, { "epoch": 2.602876797998749, "grad_norm": 0.1339552402496338, "learning_rate": 8.856160601005458e-08, "loss": 0.648, "step": 3125 }, { "epoch": 2.60371065249114, "grad_norm": 0.14748166501522064, "learning_rate": 8.819235243164957e-08, "loss": 0.5998, "step": 3126 }, { "epoch": 2.6045445069835313, "grad_norm": 0.13691233098506927, "learning_rate": 8.782383474117838e-08, "loss": 0.6163, "step": 3127 }, { "epoch": 2.6053783614759225, "grad_norm": 0.1279921531677246, "learning_rate": 8.745605323605876e-08, "loss": 0.6077, "step": 3128 }, { "epoch": 2.6062122159683137, "grad_norm": 0.14039035141468048, "learning_rate": 8.708900821311404e-08, "loss": 0.6349, "step": 3129 }, { "epoch": 2.6070460704607044, "grad_norm": 0.13836809992790222, "learning_rate": 8.672269996857373e-08, "loss": 0.5648, "step": 3130 }, { "epoch": 2.607879924953096, "grad_norm": 0.13423679769039154, "learning_rate": 8.635712879807222e-08, "loss": 0.5125, "step": 3131 }, { "epoch": 2.6087137794454867, "grad_norm": 0.13100112974643707, "learning_rate": 8.599229499664906e-08, "loss": 0.6376, "step": 3132 }, { "epoch": 2.609547633937878, "grad_norm": 0.14407749474048615, "learning_rate": 8.562819885874884e-08, "loss": 0.5973, "step": 3133 }, { "epoch": 2.610381488430269, "grad_norm": 0.13678571581840515, "learning_rate": 8.526484067822082e-08, "loss": 0.5967, "step": 3134 }, { "epoch": 2.61121534292266, "grad_norm": 0.14255864918231964, "learning_rate": 8.490222074831844e-08, "loss": 0.6606, "step": 3135 }, { "epoch": 2.6120491974150513, "grad_norm": 0.13524065911769867, "learning_rate": 8.45403393616998e-08, "loss": 0.5667, "step": 3136 }, { "epoch": 2.612883051907442, "grad_norm": 0.1418743133544922, "learning_rate": 8.417919681042651e-08, "loss": 0.5524, "step": 3137 }, { "epoch": 2.613716906399833, "grad_norm": 0.13822202384471893, "learning_rate": 8.38187933859641e-08, "loss": 0.6199, "step": 3138 }, { "epoch": 2.6145507608922243, "grad_norm": 0.14503109455108643, "learning_rate": 8.34591293791812e-08, "loss": 0.5912, "step": 3139 }, { "epoch": 2.6153846153846154, "grad_norm": 0.14785243570804596, "learning_rate": 8.310020508035054e-08, "loss": 0.6044, "step": 3140 }, { "epoch": 2.6162184698770066, "grad_norm": 0.13747577369213104, "learning_rate": 8.274202077914705e-08, "loss": 0.6059, "step": 3141 }, { "epoch": 2.6170523243693973, "grad_norm": 0.1328943818807602, "learning_rate": 8.238457676464871e-08, "loss": 0.5666, "step": 3142 }, { "epoch": 2.617886178861789, "grad_norm": 0.13470874726772308, "learning_rate": 8.202787332533589e-08, "loss": 0.6666, "step": 3143 }, { "epoch": 2.6187200333541796, "grad_norm": 0.141288623213768, "learning_rate": 8.167191074909164e-08, "loss": 0.6507, "step": 3144 }, { "epoch": 2.6195538878465707, "grad_norm": 0.1307288110256195, "learning_rate": 8.13166893232008e-08, "loss": 0.619, "step": 3145 }, { "epoch": 2.620387742338962, "grad_norm": 0.14363805949687958, "learning_rate": 8.096220933435005e-08, "loss": 0.596, "step": 3146 }, { "epoch": 2.621221596831353, "grad_norm": 0.13206861913204193, "learning_rate": 8.060847106862778e-08, "loss": 0.5791, "step": 3147 }, { "epoch": 2.622055451323744, "grad_norm": 0.13542182743549347, "learning_rate": 8.025547481152372e-08, "loss": 0.5811, "step": 3148 }, { "epoch": 2.622889305816135, "grad_norm": 0.1321006715297699, "learning_rate": 7.990322084792866e-08, "loss": 0.6358, "step": 3149 }, { "epoch": 2.623723160308526, "grad_norm": 0.13542653620243073, "learning_rate": 7.955170946213441e-08, "loss": 0.6319, "step": 3150 }, { "epoch": 2.624557014800917, "grad_norm": 0.1310838907957077, "learning_rate": 7.920094093783369e-08, "loss": 0.5897, "step": 3151 }, { "epoch": 2.6253908692933083, "grad_norm": 0.13916926085948944, "learning_rate": 7.885091555811929e-08, "loss": 0.5275, "step": 3152 }, { "epoch": 2.6262247237856995, "grad_norm": 0.13519278168678284, "learning_rate": 7.850163360548423e-08, "loss": 0.623, "step": 3153 }, { "epoch": 2.6270585782780906, "grad_norm": 0.14452557265758514, "learning_rate": 7.815309536182213e-08, "loss": 0.579, "step": 3154 }, { "epoch": 2.627892432770482, "grad_norm": 0.13447542488574982, "learning_rate": 7.780530110842565e-08, "loss": 0.658, "step": 3155 }, { "epoch": 2.6287262872628725, "grad_norm": 0.13196657598018646, "learning_rate": 7.745825112598747e-08, "loss": 0.612, "step": 3156 }, { "epoch": 2.6295601417552636, "grad_norm": 0.14191506803035736, "learning_rate": 7.711194569459933e-08, "loss": 0.5857, "step": 3157 }, { "epoch": 2.630393996247655, "grad_norm": 0.13351404666900635, "learning_rate": 7.67663850937521e-08, "loss": 0.5922, "step": 3158 }, { "epoch": 2.631227850740046, "grad_norm": 0.1378561109304428, "learning_rate": 7.642156960233592e-08, "loss": 0.563, "step": 3159 }, { "epoch": 2.632061705232437, "grad_norm": 0.14186091721057892, "learning_rate": 7.607749949863907e-08, "loss": 0.5939, "step": 3160 }, { "epoch": 2.632895559724828, "grad_norm": 0.13343293964862823, "learning_rate": 7.573417506034851e-08, "loss": 0.591, "step": 3161 }, { "epoch": 2.633729414217219, "grad_norm": 0.13701584935188293, "learning_rate": 7.539159656454919e-08, "loss": 0.5781, "step": 3162 }, { "epoch": 2.63456326870961, "grad_norm": 0.13953688740730286, "learning_rate": 7.504976428772436e-08, "loss": 0.6192, "step": 3163 }, { "epoch": 2.6353971232020013, "grad_norm": 0.14170017838478088, "learning_rate": 7.470867850575458e-08, "loss": 0.5722, "step": 3164 }, { "epoch": 2.6362309776943924, "grad_norm": 0.12384858727455139, "learning_rate": 7.436833949391852e-08, "loss": 0.5508, "step": 3165 }, { "epoch": 2.6370648321867836, "grad_norm": 0.1331976056098938, "learning_rate": 7.40287475268918e-08, "loss": 0.6255, "step": 3166 }, { "epoch": 2.6378986866791747, "grad_norm": 0.131424680352211, "learning_rate": 7.36899028787471e-08, "loss": 0.5759, "step": 3167 }, { "epoch": 2.6387325411715654, "grad_norm": 0.1211531013250351, "learning_rate": 7.335180582295387e-08, "loss": 0.5233, "step": 3168 }, { "epoch": 2.6395663956639566, "grad_norm": 0.1320580244064331, "learning_rate": 7.301445663237859e-08, "loss": 0.5967, "step": 3169 }, { "epoch": 2.6404002501563477, "grad_norm": 0.14533844590187073, "learning_rate": 7.267785557928397e-08, "loss": 0.5778, "step": 3170 }, { "epoch": 2.641234104648739, "grad_norm": 0.13105171918869019, "learning_rate": 7.234200293532889e-08, "loss": 0.6468, "step": 3171 }, { "epoch": 2.64206795914113, "grad_norm": 0.1390206217765808, "learning_rate": 7.200689897156787e-08, "loss": 0.5533, "step": 3172 }, { "epoch": 2.6429018136335207, "grad_norm": 0.13692353665828705, "learning_rate": 7.167254395845201e-08, "loss": 0.6265, "step": 3173 }, { "epoch": 2.643735668125912, "grad_norm": 0.14241954684257507, "learning_rate": 7.133893816582737e-08, "loss": 0.6269, "step": 3174 }, { "epoch": 2.644569522618303, "grad_norm": 0.13459300994873047, "learning_rate": 7.100608186293544e-08, "loss": 0.5674, "step": 3175 }, { "epoch": 2.645403377110694, "grad_norm": 0.13194911181926727, "learning_rate": 7.067397531841279e-08, "loss": 0.5999, "step": 3176 }, { "epoch": 2.6462372316030853, "grad_norm": 0.12783268094062805, "learning_rate": 7.034261880029113e-08, "loss": 0.6032, "step": 3177 }, { "epoch": 2.6470710860954765, "grad_norm": 0.1423323005437851, "learning_rate": 7.001201257599643e-08, "loss": 0.6249, "step": 3178 }, { "epoch": 2.6479049405878676, "grad_norm": 0.14864739775657654, "learning_rate": 6.968215691234935e-08, "loss": 0.5589, "step": 3179 }, { "epoch": 2.6487387950802583, "grad_norm": 0.131490558385849, "learning_rate": 6.935305207556508e-08, "loss": 0.5947, "step": 3180 }, { "epoch": 2.6495726495726495, "grad_norm": 0.1352132111787796, "learning_rate": 6.902469833125235e-08, "loss": 0.5823, "step": 3181 }, { "epoch": 2.6504065040650406, "grad_norm": 0.1323070228099823, "learning_rate": 6.869709594441398e-08, "loss": 0.5924, "step": 3182 }, { "epoch": 2.651240358557432, "grad_norm": 0.13742870092391968, "learning_rate": 6.837024517944656e-08, "loss": 0.6295, "step": 3183 }, { "epoch": 2.652074213049823, "grad_norm": 0.1480836272239685, "learning_rate": 6.804414630013966e-08, "loss": 0.6012, "step": 3184 }, { "epoch": 2.6529080675422136, "grad_norm": 0.14015209674835205, "learning_rate": 6.77187995696763e-08, "loss": 0.5595, "step": 3185 }, { "epoch": 2.6537419220346052, "grad_norm": 0.1540982872247696, "learning_rate": 6.739420525063233e-08, "loss": 0.6032, "step": 3186 }, { "epoch": 2.654575776526996, "grad_norm": 0.14946775138378143, "learning_rate": 6.707036360497631e-08, "loss": 0.6006, "step": 3187 }, { "epoch": 2.655409631019387, "grad_norm": 0.13741618394851685, "learning_rate": 6.674727489406983e-08, "loss": 0.6193, "step": 3188 }, { "epoch": 2.6562434855117782, "grad_norm": 0.1380772888660431, "learning_rate": 6.642493937866623e-08, "loss": 0.6163, "step": 3189 }, { "epoch": 2.6570773400041694, "grad_norm": 0.14575491845607758, "learning_rate": 6.610335731891115e-08, "loss": 0.5897, "step": 3190 }, { "epoch": 2.6579111944965605, "grad_norm": 0.14663982391357422, "learning_rate": 6.578252897434222e-08, "loss": 0.5779, "step": 3191 }, { "epoch": 2.6587450489889513, "grad_norm": 0.13804934918880463, "learning_rate": 6.546245460388878e-08, "loss": 0.6195, "step": 3192 }, { "epoch": 2.6595789034813424, "grad_norm": 0.13802267611026764, "learning_rate": 6.51431344658716e-08, "loss": 0.5465, "step": 3193 }, { "epoch": 2.6604127579737336, "grad_norm": 0.1352781057357788, "learning_rate": 6.482456881800246e-08, "loss": 0.6086, "step": 3194 }, { "epoch": 2.6612466124661247, "grad_norm": 0.14254027605056763, "learning_rate": 6.450675791738502e-08, "loss": 0.594, "step": 3195 }, { "epoch": 2.662080466958516, "grad_norm": 0.12728862464427948, "learning_rate": 6.418970202051299e-08, "loss": 0.6049, "step": 3196 }, { "epoch": 2.6629143214509066, "grad_norm": 0.1306411623954773, "learning_rate": 6.387340138327091e-08, "loss": 0.5897, "step": 3197 }, { "epoch": 2.663748175943298, "grad_norm": 0.15724188089370728, "learning_rate": 6.355785626093424e-08, "loss": 0.6183, "step": 3198 }, { "epoch": 2.664582030435689, "grad_norm": 0.14000767469406128, "learning_rate": 6.324306690816839e-08, "loss": 0.6041, "step": 3199 }, { "epoch": 2.66541588492808, "grad_norm": 0.1324029564857483, "learning_rate": 6.292903357902846e-08, "loss": 0.6394, "step": 3200 }, { "epoch": 2.66541588492808, "eval_loss": 0.6060388088226318, "eval_runtime": 321.4788, "eval_samples_per_second": 16.744, "eval_steps_per_second": 2.793, "step": 3200 }, { "epoch": 2.666249739420471, "grad_norm": 0.13470742106437683, "learning_rate": 6.261575652696005e-08, "loss": 0.6116, "step": 3201 }, { "epoch": 2.6670835939128623, "grad_norm": 0.13555613160133362, "learning_rate": 6.23032360047977e-08, "loss": 0.6334, "step": 3202 }, { "epoch": 2.6679174484052535, "grad_norm": 0.1306881457567215, "learning_rate": 6.199147226476608e-08, "loss": 0.6035, "step": 3203 }, { "epoch": 2.668751302897644, "grad_norm": 0.13432841002941132, "learning_rate": 6.168046555847872e-08, "loss": 0.6405, "step": 3204 }, { "epoch": 2.6695851573900353, "grad_norm": 0.13840965926647186, "learning_rate": 6.137021613693805e-08, "loss": 0.6139, "step": 3205 }, { "epoch": 2.6704190118824265, "grad_norm": 0.13548065721988678, "learning_rate": 6.106072425053565e-08, "loss": 0.5673, "step": 3206 }, { "epoch": 2.6712528663748176, "grad_norm": 0.1370995193719864, "learning_rate": 6.075199014905152e-08, "loss": 0.6277, "step": 3207 }, { "epoch": 2.6720867208672088, "grad_norm": 0.13972927629947662, "learning_rate": 6.044401408165378e-08, "loss": 0.6228, "step": 3208 }, { "epoch": 2.6729205753595995, "grad_norm": 0.13325779139995575, "learning_rate": 6.013679629689961e-08, "loss": 0.5941, "step": 3209 }, { "epoch": 2.673754429851991, "grad_norm": 0.135503888130188, "learning_rate": 5.983033704273355e-08, "loss": 0.6341, "step": 3210 }, { "epoch": 2.674588284344382, "grad_norm": 0.1303946077823639, "learning_rate": 5.952463656648776e-08, "loss": 0.6115, "step": 3211 }, { "epoch": 2.675422138836773, "grad_norm": 0.13263511657714844, "learning_rate": 5.921969511488278e-08, "loss": 0.6329, "step": 3212 }, { "epoch": 2.676255993329164, "grad_norm": 0.13432781398296356, "learning_rate": 5.8915512934026076e-08, "loss": 0.5552, "step": 3213 }, { "epoch": 2.6770898478215552, "grad_norm": 0.13039176166057587, "learning_rate": 5.8612090269412273e-08, "loss": 0.617, "step": 3214 }, { "epoch": 2.6779237023139464, "grad_norm": 0.1374451071023941, "learning_rate": 5.8309427365923147e-08, "loss": 0.6029, "step": 3215 }, { "epoch": 2.678757556806337, "grad_norm": 0.1409667581319809, "learning_rate": 5.800752446782731e-08, "loss": 0.5886, "step": 3216 }, { "epoch": 2.6795914112987282, "grad_norm": 0.13507746160030365, "learning_rate": 5.770638181878007e-08, "loss": 0.575, "step": 3217 }, { "epoch": 2.6804252657911194, "grad_norm": 0.13138867914676666, "learning_rate": 5.740599966182313e-08, "loss": 0.603, "step": 3218 }, { "epoch": 2.6812591202835105, "grad_norm": 0.13637477159500122, "learning_rate": 5.7106378239384246e-08, "loss": 0.5937, "step": 3219 }, { "epoch": 2.6820929747759017, "grad_norm": 0.13671500980854034, "learning_rate": 5.6807517793277414e-08, "loss": 0.5929, "step": 3220 }, { "epoch": 2.682926829268293, "grad_norm": 0.1326328068971634, "learning_rate": 5.6509418564702375e-08, "loss": 0.5972, "step": 3221 }, { "epoch": 2.683760683760684, "grad_norm": 0.1319316029548645, "learning_rate": 5.6212080794244444e-08, "loss": 0.5825, "step": 3222 }, { "epoch": 2.6845945382530747, "grad_norm": 0.1371231973171234, "learning_rate": 5.5915504721874674e-08, "loss": 0.5947, "step": 3223 }, { "epoch": 2.685428392745466, "grad_norm": 0.13127702474594116, "learning_rate": 5.561969058694915e-08, "loss": 0.6263, "step": 3224 }, { "epoch": 2.686262247237857, "grad_norm": 0.12458047270774841, "learning_rate": 5.532463862820902e-08, "loss": 0.5807, "step": 3225 }, { "epoch": 2.687096101730248, "grad_norm": 0.13222922384738922, "learning_rate": 5.5030349083780346e-08, "loss": 0.5531, "step": 3226 }, { "epoch": 2.6879299562226393, "grad_norm": 0.1369413137435913, "learning_rate": 5.473682219117404e-08, "loss": 0.6209, "step": 3227 }, { "epoch": 2.68876381071503, "grad_norm": 0.13719838857650757, "learning_rate": 5.444405818728526e-08, "loss": 0.6308, "step": 3228 }, { "epoch": 2.689597665207421, "grad_norm": 0.12738753855228424, "learning_rate": 5.415205730839356e-08, "loss": 0.5833, "step": 3229 }, { "epoch": 2.6904315196998123, "grad_norm": 0.14416873455047607, "learning_rate": 5.3860819790162547e-08, "loss": 0.6383, "step": 3230 }, { "epoch": 2.6912653741922035, "grad_norm": 0.12869682908058167, "learning_rate": 5.357034586763964e-08, "loss": 0.5182, "step": 3231 }, { "epoch": 2.6920992286845946, "grad_norm": 0.1447078436613083, "learning_rate": 5.3280635775256545e-08, "loss": 0.6214, "step": 3232 }, { "epoch": 2.6929330831769858, "grad_norm": 0.1304541379213333, "learning_rate": 5.299168974682788e-08, "loss": 0.584, "step": 3233 }, { "epoch": 2.693766937669377, "grad_norm": 0.13652051985263824, "learning_rate": 5.270350801555179e-08, "loss": 0.6196, "step": 3234 }, { "epoch": 2.6946007921617676, "grad_norm": 0.13038598001003265, "learning_rate": 5.241609081400966e-08, "loss": 0.589, "step": 3235 }, { "epoch": 2.6954346466541588, "grad_norm": 0.14265742897987366, "learning_rate": 5.2129438374165836e-08, "loss": 0.6241, "step": 3236 }, { "epoch": 2.69626850114655, "grad_norm": 0.13225463032722473, "learning_rate": 5.1843550927367365e-08, "loss": 0.6024, "step": 3237 }, { "epoch": 2.697102355638941, "grad_norm": 0.12445509433746338, "learning_rate": 5.155842870434413e-08, "loss": 0.5283, "step": 3238 }, { "epoch": 2.6979362101313322, "grad_norm": 0.1268155425786972, "learning_rate": 5.1274071935208276e-08, "loss": 0.545, "step": 3239 }, { "epoch": 2.698770064623723, "grad_norm": 0.13421349227428436, "learning_rate": 5.099048084945401e-08, "loss": 0.6167, "step": 3240 }, { "epoch": 2.699603919116114, "grad_norm": 0.13776518404483795, "learning_rate": 5.070765567595758e-08, "loss": 0.585, "step": 3241 }, { "epoch": 2.7004377736085052, "grad_norm": 0.1408858597278595, "learning_rate": 5.042559664297774e-08, "loss": 0.585, "step": 3242 }, { "epoch": 2.7012716281008964, "grad_norm": 0.1378757804632187, "learning_rate": 5.0144303978154054e-08, "loss": 0.5814, "step": 3243 }, { "epoch": 2.7021054825932875, "grad_norm": 0.13119475543498993, "learning_rate": 4.986377790850815e-08, "loss": 0.6186, "step": 3244 }, { "epoch": 2.7029393370856787, "grad_norm": 0.13207046687602997, "learning_rate": 4.958401866044248e-08, "loss": 0.5762, "step": 3245 }, { "epoch": 2.70377319157807, "grad_norm": 0.13125978410243988, "learning_rate": 4.930502645974122e-08, "loss": 0.608, "step": 3246 }, { "epoch": 2.7046070460704605, "grad_norm": 0.1301548033952713, "learning_rate": 4.902680153156913e-08, "loss": 0.6198, "step": 3247 }, { "epoch": 2.7054409005628517, "grad_norm": 0.13901136815547943, "learning_rate": 4.874934410047171e-08, "loss": 0.5699, "step": 3248 }, { "epoch": 2.706274755055243, "grad_norm": 0.13569803535938263, "learning_rate": 4.847265439037507e-08, "loss": 0.591, "step": 3249 }, { "epoch": 2.707108609547634, "grad_norm": 0.13454224169254303, "learning_rate": 4.81967326245859e-08, "loss": 0.5863, "step": 3250 }, { "epoch": 2.707942464040025, "grad_norm": 0.13513845205307007, "learning_rate": 4.792157902579075e-08, "loss": 0.6112, "step": 3251 }, { "epoch": 2.708776318532416, "grad_norm": 0.1458248496055603, "learning_rate": 4.764719381605642e-08, "loss": 0.6287, "step": 3252 }, { "epoch": 2.7096101730248074, "grad_norm": 0.13269999623298645, "learning_rate": 4.7373577216829885e-08, "loss": 0.623, "step": 3253 }, { "epoch": 2.710444027517198, "grad_norm": 0.131892129778862, "learning_rate": 4.710072944893739e-08, "loss": 0.5775, "step": 3254 }, { "epoch": 2.7112778820095893, "grad_norm": 0.12902632355690002, "learning_rate": 4.6828650732584466e-08, "loss": 0.5504, "step": 3255 }, { "epoch": 2.7121117365019805, "grad_norm": 0.13969209790229797, "learning_rate": 4.655734128735678e-08, "loss": 0.5942, "step": 3256 }, { "epoch": 2.7129455909943716, "grad_norm": 0.13684862852096558, "learning_rate": 4.6286801332218295e-08, "loss": 0.5699, "step": 3257 }, { "epoch": 2.7137794454867628, "grad_norm": 0.12708385288715363, "learning_rate": 4.601703108551247e-08, "loss": 0.5832, "step": 3258 }, { "epoch": 2.7146132999791535, "grad_norm": 0.14265984296798706, "learning_rate": 4.574803076496147e-08, "loss": 0.5872, "step": 3259 }, { "epoch": 2.7154471544715446, "grad_norm": 0.1383449137210846, "learning_rate": 4.547980058766565e-08, "loss": 0.5751, "step": 3260 }, { "epoch": 2.7162810089639358, "grad_norm": 0.12680986523628235, "learning_rate": 4.521234077010449e-08, "loss": 0.546, "step": 3261 }, { "epoch": 2.717114863456327, "grad_norm": 0.13447879254817963, "learning_rate": 4.4945651528135454e-08, "loss": 0.5581, "step": 3262 }, { "epoch": 2.717948717948718, "grad_norm": 0.13199739158153534, "learning_rate": 4.467973307699391e-08, "loss": 0.6055, "step": 3263 }, { "epoch": 2.7187825724411088, "grad_norm": 0.13321660459041595, "learning_rate": 4.44145856312933e-08, "loss": 0.5695, "step": 3264 }, { "epoch": 2.7196164269335004, "grad_norm": 0.1386847347021103, "learning_rate": 4.415020940502501e-08, "loss": 0.6057, "step": 3265 }, { "epoch": 2.720450281425891, "grad_norm": 0.1390652060508728, "learning_rate": 4.388660461155735e-08, "loss": 0.6325, "step": 3266 }, { "epoch": 2.721284135918282, "grad_norm": 0.1421324461698532, "learning_rate": 4.3623771463637047e-08, "loss": 0.6342, "step": 3267 }, { "epoch": 2.7221179904106734, "grad_norm": 0.1424824744462967, "learning_rate": 4.336171017338719e-08, "loss": 0.5939, "step": 3268 }, { "epoch": 2.7229518449030645, "grad_norm": 0.132848858833313, "learning_rate": 4.3100420952308366e-08, "loss": 0.5769, "step": 3269 }, { "epoch": 2.7237856993954557, "grad_norm": 0.12438632547855377, "learning_rate": 4.283990401127768e-08, "loss": 0.5994, "step": 3270 }, { "epoch": 2.7246195538878464, "grad_norm": 0.13128380477428436, "learning_rate": 4.25801595605495e-08, "loss": 0.5742, "step": 3271 }, { "epoch": 2.7254534083802375, "grad_norm": 0.13314370810985565, "learning_rate": 4.232118780975447e-08, "loss": 0.5809, "step": 3272 }, { "epoch": 2.7262872628726287, "grad_norm": 0.1332280933856964, "learning_rate": 4.2062988967899394e-08, "loss": 0.6093, "step": 3273 }, { "epoch": 2.72712111736502, "grad_norm": 0.13469170033931732, "learning_rate": 4.180556324336748e-08, "loss": 0.585, "step": 3274 }, { "epoch": 2.727954971857411, "grad_norm": 0.13007333874702454, "learning_rate": 4.1548910843918203e-08, "loss": 0.5852, "step": 3275 }, { "epoch": 2.7287888263498017, "grad_norm": 0.13419799506664276, "learning_rate": 4.129303197668666e-08, "loss": 0.5742, "step": 3276 }, { "epoch": 2.7296226808421933, "grad_norm": 0.13037191331386566, "learning_rate": 4.1037926848183525e-08, "loss": 0.5463, "step": 3277 }, { "epoch": 2.730456535334584, "grad_norm": 0.13272960484027863, "learning_rate": 4.0783595664295346e-08, "loss": 0.6169, "step": 3278 }, { "epoch": 2.731290389826975, "grad_norm": 0.12827260792255402, "learning_rate": 4.053003863028381e-08, "loss": 0.6146, "step": 3279 }, { "epoch": 2.7321242443193663, "grad_norm": 0.13763415813446045, "learning_rate": 4.02772559507859e-08, "loss": 0.5732, "step": 3280 }, { "epoch": 2.7329580988117574, "grad_norm": 0.13149921596050262, "learning_rate": 4.0025247829813646e-08, "loss": 0.6097, "step": 3281 }, { "epoch": 2.7337919533041486, "grad_norm": 0.13176915049552917, "learning_rate": 3.977401447075401e-08, "loss": 0.571, "step": 3282 }, { "epoch": 2.7346258077965393, "grad_norm": 0.13508684933185577, "learning_rate": 3.952355607636859e-08, "loss": 0.584, "step": 3283 }, { "epoch": 2.7354596622889304, "grad_norm": 0.1335073858499527, "learning_rate": 3.927387284879336e-08, "loss": 0.5121, "step": 3284 }, { "epoch": 2.7362935167813216, "grad_norm": 0.12906381487846375, "learning_rate": 3.902496498953922e-08, "loss": 0.5526, "step": 3285 }, { "epoch": 2.7371273712737128, "grad_norm": 0.14333176612854004, "learning_rate": 3.8776832699490836e-08, "loss": 0.57, "step": 3286 }, { "epoch": 2.737961225766104, "grad_norm": 0.13652856647968292, "learning_rate": 3.8529476178907004e-08, "loss": 0.5821, "step": 3287 }, { "epoch": 2.738795080258495, "grad_norm": 0.14862151443958282, "learning_rate": 3.82828956274206e-08, "loss": 0.5894, "step": 3288 }, { "epoch": 2.739628934750886, "grad_norm": 0.1358601450920105, "learning_rate": 3.8037091244037866e-08, "loss": 0.5614, "step": 3289 }, { "epoch": 2.740462789243277, "grad_norm": 0.1391635686159134, "learning_rate": 3.7792063227139126e-08, "loss": 0.602, "step": 3290 }, { "epoch": 2.741296643735668, "grad_norm": 0.1313261091709137, "learning_rate": 3.754781177447797e-08, "loss": 0.5589, "step": 3291 }, { "epoch": 2.742130498228059, "grad_norm": 0.139642596244812, "learning_rate": 3.7304337083180924e-08, "loss": 0.561, "step": 3292 }, { "epoch": 2.7429643527204504, "grad_norm": 0.15068678557872772, "learning_rate": 3.706163934974804e-08, "loss": 0.5765, "step": 3293 }, { "epoch": 2.7437982072128415, "grad_norm": 0.13592305779457092, "learning_rate": 3.6819718770052076e-08, "loss": 0.591, "step": 3294 }, { "epoch": 2.744632061705232, "grad_norm": 0.13014188408851624, "learning_rate": 3.657857553933841e-08, "loss": 0.5622, "step": 3295 }, { "epoch": 2.7454659161976234, "grad_norm": 0.13747231662273407, "learning_rate": 3.633820985222569e-08, "loss": 0.6119, "step": 3296 }, { "epoch": 2.7462997706900145, "grad_norm": 0.12931308150291443, "learning_rate": 3.6098621902704406e-08, "loss": 0.5682, "step": 3297 }, { "epoch": 2.7471336251824057, "grad_norm": 0.1329226791858673, "learning_rate": 3.585981188413767e-08, "loss": 0.5577, "step": 3298 }, { "epoch": 2.747967479674797, "grad_norm": 0.13071216642856598, "learning_rate": 3.562177998926041e-08, "loss": 0.6024, "step": 3299 }, { "epoch": 2.748801334167188, "grad_norm": 0.13766297698020935, "learning_rate": 3.538452641018008e-08, "loss": 0.5997, "step": 3300 }, { "epoch": 2.749635188659579, "grad_norm": 0.1369374394416809, "learning_rate": 3.514805133837562e-08, "loss": 0.5504, "step": 3301 }, { "epoch": 2.75046904315197, "grad_norm": 0.1379423290491104, "learning_rate": 3.49123549646978e-08, "loss": 0.6026, "step": 3302 }, { "epoch": 2.751302897644361, "grad_norm": 0.1353907734155655, "learning_rate": 3.4677437479368685e-08, "loss": 0.5925, "step": 3303 }, { "epoch": 2.752136752136752, "grad_norm": 0.12980139255523682, "learning_rate": 3.444329907198229e-08, "loss": 0.5767, "step": 3304 }, { "epoch": 2.7529706066291433, "grad_norm": 0.14043360948562622, "learning_rate": 3.420993993150323e-08, "loss": 0.5912, "step": 3305 }, { "epoch": 2.7538044611215344, "grad_norm": 0.1324455440044403, "learning_rate": 3.397736024626752e-08, "loss": 0.5764, "step": 3306 }, { "epoch": 2.754638315613925, "grad_norm": 0.1285971999168396, "learning_rate": 3.374556020398211e-08, "loss": 0.5755, "step": 3307 }, { "epoch": 2.7554721701063163, "grad_norm": 0.1404966413974762, "learning_rate": 3.351453999172471e-08, "loss": 0.6218, "step": 3308 }, { "epoch": 2.7563060245987074, "grad_norm": 0.12702247500419617, "learning_rate": 3.3284299795943495e-08, "loss": 0.6081, "step": 3309 }, { "epoch": 2.7571398790910986, "grad_norm": 0.1338385045528412, "learning_rate": 3.3054839802457156e-08, "loss": 0.5997, "step": 3310 }, { "epoch": 2.7579737335834897, "grad_norm": 0.14341551065444946, "learning_rate": 3.282616019645512e-08, "loss": 0.5597, "step": 3311 }, { "epoch": 2.758807588075881, "grad_norm": 0.1413535624742508, "learning_rate": 3.259826116249642e-08, "loss": 0.6367, "step": 3312 }, { "epoch": 2.759641442568272, "grad_norm": 0.13823261857032776, "learning_rate": 3.2371142884510375e-08, "loss": 0.6002, "step": 3313 }, { "epoch": 2.7604752970606627, "grad_norm": 0.1444302350282669, "learning_rate": 3.214480554579635e-08, "loss": 0.5833, "step": 3314 }, { "epoch": 2.761309151553054, "grad_norm": 0.13439130783081055, "learning_rate": 3.1919249329023125e-08, "loss": 0.6072, "step": 3315 }, { "epoch": 2.762143006045445, "grad_norm": 0.13830120861530304, "learning_rate": 3.169447441622919e-08, "loss": 0.5976, "step": 3316 }, { "epoch": 2.762976860537836, "grad_norm": 0.12596963346004486, "learning_rate": 3.147048098882255e-08, "loss": 0.5382, "step": 3317 }, { "epoch": 2.7638107150302273, "grad_norm": 0.12890082597732544, "learning_rate": 3.1247269227580255e-08, "loss": 0.5533, "step": 3318 }, { "epoch": 2.764644569522618, "grad_norm": 0.13346536457538605, "learning_rate": 3.1024839312648986e-08, "loss": 0.5908, "step": 3319 }, { "epoch": 2.7654784240150097, "grad_norm": 0.13776253163814545, "learning_rate": 3.080319142354393e-08, "loss": 0.61, "step": 3320 }, { "epoch": 2.7663122785074004, "grad_norm": 0.13473759591579437, "learning_rate": 3.0582325739149315e-08, "loss": 0.5677, "step": 3321 }, { "epoch": 2.7671461329997915, "grad_norm": 0.14179588854312897, "learning_rate": 3.036224243771812e-08, "loss": 0.5815, "step": 3322 }, { "epoch": 2.7679799874921827, "grad_norm": 0.13902443647384644, "learning_rate": 3.014294169687182e-08, "loss": 0.6231, "step": 3323 }, { "epoch": 2.768813841984574, "grad_norm": 0.13968312740325928, "learning_rate": 2.992442369360015e-08, "loss": 0.5573, "step": 3324 }, { "epoch": 2.769647696476965, "grad_norm": 0.14189480245113373, "learning_rate": 2.9706688604261842e-08, "loss": 0.5661, "step": 3325 }, { "epoch": 2.7704815509693557, "grad_norm": 0.13203908503055573, "learning_rate": 2.948973660458276e-08, "loss": 0.5727, "step": 3326 }, { "epoch": 2.771315405461747, "grad_norm": 0.13922102749347687, "learning_rate": 2.92735678696574e-08, "loss": 0.5758, "step": 3327 }, { "epoch": 2.772149259954138, "grad_norm": 0.13167941570281982, "learning_rate": 2.9058182573947986e-08, "loss": 0.6149, "step": 3328 }, { "epoch": 2.772983114446529, "grad_norm": 0.14113520085811615, "learning_rate": 2.884358089128436e-08, "loss": 0.5964, "step": 3329 }, { "epoch": 2.7738169689389203, "grad_norm": 0.13135458528995514, "learning_rate": 2.8629762994864194e-08, "loss": 0.6397, "step": 3330 }, { "epoch": 2.774650823431311, "grad_norm": 0.1415179967880249, "learning_rate": 2.841672905725212e-08, "loss": 0.5917, "step": 3331 }, { "epoch": 2.7754846779237026, "grad_norm": 0.12605704367160797, "learning_rate": 2.8204479250380497e-08, "loss": 0.5547, "step": 3332 }, { "epoch": 2.7763185324160933, "grad_norm": 0.14685988426208496, "learning_rate": 2.799301374554863e-08, "loss": 0.657, "step": 3333 }, { "epoch": 2.7771523869084844, "grad_norm": 0.1261330097913742, "learning_rate": 2.778233271342312e-08, "loss": 0.6016, "step": 3334 }, { "epoch": 2.7779862414008756, "grad_norm": 0.12885618209838867, "learning_rate": 2.757243632403694e-08, "loss": 0.5165, "step": 3335 }, { "epoch": 2.7788200958932667, "grad_norm": 0.130287766456604, "learning_rate": 2.7363324746790263e-08, "loss": 0.5749, "step": 3336 }, { "epoch": 2.779653950385658, "grad_norm": 0.13909180462360382, "learning_rate": 2.7154998150449638e-08, "loss": 0.5748, "step": 3337 }, { "epoch": 2.7804878048780486, "grad_norm": 0.13536633551120758, "learning_rate": 2.694745670314824e-08, "loss": 0.5769, "step": 3338 }, { "epoch": 2.7813216593704397, "grad_norm": 0.13742972910404205, "learning_rate": 2.67407005723852e-08, "loss": 0.5778, "step": 3339 }, { "epoch": 2.782155513862831, "grad_norm": 0.1485828012228012, "learning_rate": 2.653472992502648e-08, "loss": 0.6061, "step": 3340 }, { "epoch": 2.782989368355222, "grad_norm": 0.13658133149147034, "learning_rate": 2.6329544927303548e-08, "loss": 0.566, "step": 3341 }, { "epoch": 2.783823222847613, "grad_norm": 0.1493666023015976, "learning_rate": 2.612514574481417e-08, "loss": 0.6392, "step": 3342 }, { "epoch": 2.784657077340004, "grad_norm": 0.13238009810447693, "learning_rate": 2.5921532542521494e-08, "loss": 0.6062, "step": 3343 }, { "epoch": 2.7854909318323955, "grad_norm": 0.13656125962734222, "learning_rate": 2.5718705484754853e-08, "loss": 0.5821, "step": 3344 }, { "epoch": 2.786324786324786, "grad_norm": 0.1402300000190735, "learning_rate": 2.5516664735208748e-08, "loss": 0.6277, "step": 3345 }, { "epoch": 2.7871586408171773, "grad_norm": 0.13126076757907867, "learning_rate": 2.5315410456943187e-08, "loss": 0.6041, "step": 3346 }, { "epoch": 2.7879924953095685, "grad_norm": 0.13182519376277924, "learning_rate": 2.5114942812383243e-08, "loss": 0.5526, "step": 3347 }, { "epoch": 2.7888263498019596, "grad_norm": 0.1345091164112091, "learning_rate": 2.491526196331961e-08, "loss": 0.5156, "step": 3348 }, { "epoch": 2.789660204294351, "grad_norm": 0.13091179728507996, "learning_rate": 2.471636807090749e-08, "loss": 0.533, "step": 3349 }, { "epoch": 2.7904940587867415, "grad_norm": 0.1362873613834381, "learning_rate": 2.451826129566725e-08, "loss": 0.5648, "step": 3350 }, { "epoch": 2.7913279132791327, "grad_norm": 0.13758814334869385, "learning_rate": 2.4320941797483896e-08, "loss": 0.6127, "step": 3351 }, { "epoch": 2.792161767771524, "grad_norm": 0.14192192256450653, "learning_rate": 2.4124409735607033e-08, "loss": 0.6296, "step": 3352 }, { "epoch": 2.792995622263915, "grad_norm": 0.1402604877948761, "learning_rate": 2.3928665268650893e-08, "loss": 0.6048, "step": 3353 }, { "epoch": 2.793829476756306, "grad_norm": 0.13130877912044525, "learning_rate": 2.3733708554593666e-08, "loss": 0.6207, "step": 3354 }, { "epoch": 2.7946633312486973, "grad_norm": 0.12290725857019424, "learning_rate": 2.3539539750778492e-08, "loss": 0.5844, "step": 3355 }, { "epoch": 2.7954971857410884, "grad_norm": 0.1342190057039261, "learning_rate": 2.3346159013911905e-08, "loss": 0.5687, "step": 3356 }, { "epoch": 2.796331040233479, "grad_norm": 0.14114157855510712, "learning_rate": 2.3153566500064615e-08, "loss": 0.6344, "step": 3357 }, { "epoch": 2.7971648947258703, "grad_norm": 0.13066042959690094, "learning_rate": 2.2961762364671623e-08, "loss": 0.5894, "step": 3358 }, { "epoch": 2.7979987492182614, "grad_norm": 0.13218310475349426, "learning_rate": 2.277074676253121e-08, "loss": 0.5559, "step": 3359 }, { "epoch": 2.7988326037106526, "grad_norm": 0.13218623399734497, "learning_rate": 2.258051984780518e-08, "loss": 0.5501, "step": 3360 }, { "epoch": 2.7996664582030437, "grad_norm": 0.13243363797664642, "learning_rate": 2.239108177401916e-08, "loss": 0.6082, "step": 3361 }, { "epoch": 2.8005003126954344, "grad_norm": 0.133683443069458, "learning_rate": 2.2202432694061746e-08, "loss": 0.627, "step": 3362 }, { "epoch": 2.8013341671878256, "grad_norm": 0.12931255996227264, "learning_rate": 2.2014572760185258e-08, "loss": 0.5866, "step": 3363 }, { "epoch": 2.8021680216802167, "grad_norm": 0.13384999334812164, "learning_rate": 2.182750212400475e-08, "loss": 0.5792, "step": 3364 }, { "epoch": 2.803001876172608, "grad_norm": 0.14243336021900177, "learning_rate": 2.164122093649834e-08, "loss": 0.6277, "step": 3365 }, { "epoch": 2.803835730664999, "grad_norm": 0.13342176377773285, "learning_rate": 2.14557293480071e-08, "loss": 0.6342, "step": 3366 }, { "epoch": 2.80466958515739, "grad_norm": 0.15023866295814514, "learning_rate": 2.1271027508234728e-08, "loss": 0.5834, "step": 3367 }, { "epoch": 2.8055034396497813, "grad_norm": 0.12801045179367065, "learning_rate": 2.1087115566247427e-08, "loss": 0.5776, "step": 3368 }, { "epoch": 2.806337294142172, "grad_norm": 0.12522506713867188, "learning_rate": 2.090399367047435e-08, "loss": 0.5719, "step": 3369 }, { "epoch": 2.807171148634563, "grad_norm": 0.13651913404464722, "learning_rate": 2.072166196870673e-08, "loss": 0.6445, "step": 3370 }, { "epoch": 2.8080050031269543, "grad_norm": 0.13021034002304077, "learning_rate": 2.0540120608097843e-08, "loss": 0.6677, "step": 3371 }, { "epoch": 2.8088388576193455, "grad_norm": 0.13658173382282257, "learning_rate": 2.0359369735163502e-08, "loss": 0.576, "step": 3372 }, { "epoch": 2.8096727121117366, "grad_norm": 0.14392971992492676, "learning_rate": 2.017940949578134e-08, "loss": 0.6, "step": 3373 }, { "epoch": 2.8105065666041273, "grad_norm": 0.13017821311950684, "learning_rate": 2.0000240035191073e-08, "loss": 0.5618, "step": 3374 }, { "epoch": 2.8113404210965185, "grad_norm": 0.12962140142917633, "learning_rate": 1.982186149799392e-08, "loss": 0.5984, "step": 3375 }, { "epoch": 2.8121742755889096, "grad_norm": 0.1291617900133133, "learning_rate": 1.9644274028152942e-08, "loss": 0.575, "step": 3376 }, { "epoch": 2.813008130081301, "grad_norm": 0.14441603422164917, "learning_rate": 1.9467477768992935e-08, "loss": 0.6437, "step": 3377 }, { "epoch": 2.813841984573692, "grad_norm": 0.14484144747257233, "learning_rate": 1.9291472863199764e-08, "loss": 0.622, "step": 3378 }, { "epoch": 2.814675839066083, "grad_norm": 0.1424298733472824, "learning_rate": 1.9116259452820916e-08, "loss": 0.5915, "step": 3379 }, { "epoch": 2.8155096935584742, "grad_norm": 0.13980284333229065, "learning_rate": 1.894183767926494e-08, "loss": 0.5591, "step": 3380 }, { "epoch": 2.816343548050865, "grad_norm": 0.13745109736919403, "learning_rate": 1.876820768330145e-08, "loss": 0.532, "step": 3381 }, { "epoch": 2.817177402543256, "grad_norm": 0.14838650822639465, "learning_rate": 1.859536960506125e-08, "loss": 0.5929, "step": 3382 }, { "epoch": 2.8180112570356473, "grad_norm": 0.15504220128059387, "learning_rate": 1.8423323584035754e-08, "loss": 0.5978, "step": 3383 }, { "epoch": 2.8188451115280384, "grad_norm": 0.14242221415042877, "learning_rate": 1.8252069759077336e-08, "loss": 0.6339, "step": 3384 }, { "epoch": 2.8196789660204296, "grad_norm": 0.1389232724905014, "learning_rate": 1.8081608268398996e-08, "loss": 0.5449, "step": 3385 }, { "epoch": 2.8205128205128203, "grad_norm": 0.13895577192306519, "learning_rate": 1.7911939249574015e-08, "loss": 0.6285, "step": 3386 }, { "epoch": 2.821346675005212, "grad_norm": 0.1358403116464615, "learning_rate": 1.7743062839536638e-08, "loss": 0.5687, "step": 3387 }, { "epoch": 2.8221805294976026, "grad_norm": 0.12436597049236298, "learning_rate": 1.757497917458084e-08, "loss": 0.541, "step": 3388 }, { "epoch": 2.8230143839899937, "grad_norm": 0.13565631210803986, "learning_rate": 1.740768839036111e-08, "loss": 0.6191, "step": 3389 }, { "epoch": 2.823848238482385, "grad_norm": 0.12684424221515656, "learning_rate": 1.7241190621891887e-08, "loss": 0.5408, "step": 3390 }, { "epoch": 2.824682092974776, "grad_norm": 0.14209088683128357, "learning_rate": 1.707548600354769e-08, "loss": 0.5494, "step": 3391 }, { "epoch": 2.825515947467167, "grad_norm": 0.13639536499977112, "learning_rate": 1.6910574669063094e-08, "loss": 0.5999, "step": 3392 }, { "epoch": 2.826349801959558, "grad_norm": 0.1375364065170288, "learning_rate": 1.674645675153208e-08, "loss": 0.6406, "step": 3393 }, { "epoch": 2.827183656451949, "grad_norm": 0.1376044601202011, "learning_rate": 1.658313238340836e-08, "loss": 0.5619, "step": 3394 }, { "epoch": 2.82801751094434, "grad_norm": 0.1302410364151001, "learning_rate": 1.64206016965055e-08, "loss": 0.6184, "step": 3395 }, { "epoch": 2.8288513654367313, "grad_norm": 0.1312178373336792, "learning_rate": 1.625886482199623e-08, "loss": 0.5278, "step": 3396 }, { "epoch": 2.8296852199291225, "grad_norm": 0.14211532473564148, "learning_rate": 1.60979218904127e-08, "loss": 0.559, "step": 3397 }, { "epoch": 2.830519074421513, "grad_norm": 0.1271534264087677, "learning_rate": 1.593777303164623e-08, "loss": 0.5249, "step": 3398 }, { "epoch": 2.8313529289139048, "grad_norm": 0.20897027850151062, "learning_rate": 1.577841837494742e-08, "loss": 0.5185, "step": 3399 }, { "epoch": 2.8321867834062955, "grad_norm": 0.1321958750486374, "learning_rate": 1.561985804892585e-08, "loss": 0.6346, "step": 3400 }, { "epoch": 2.8330206378986866, "grad_norm": 0.13702531158924103, "learning_rate": 1.5462092181549924e-08, "loss": 0.61, "step": 3401 }, { "epoch": 2.833854492391078, "grad_norm": 0.13523337244987488, "learning_rate": 1.5305120900146905e-08, "loss": 0.638, "step": 3402 }, { "epoch": 2.834688346883469, "grad_norm": 0.13042756915092468, "learning_rate": 1.5148944331402902e-08, "loss": 0.6441, "step": 3403 }, { "epoch": 2.83552220137586, "grad_norm": 0.13690252602100372, "learning_rate": 1.499356260136253e-08, "loss": 0.6313, "step": 3404 }, { "epoch": 2.836356055868251, "grad_norm": 0.13663507997989655, "learning_rate": 1.4838975835428702e-08, "loss": 0.6309, "step": 3405 }, { "epoch": 2.837189910360642, "grad_norm": 0.12254924327135086, "learning_rate": 1.4685184158363284e-08, "loss": 0.5226, "step": 3406 }, { "epoch": 2.838023764853033, "grad_norm": 0.12567856907844543, "learning_rate": 1.4532187694285991e-08, "loss": 0.517, "step": 3407 }, { "epoch": 2.8388576193454242, "grad_norm": 0.14149676263332367, "learning_rate": 1.4379986566674939e-08, "loss": 0.5745, "step": 3408 }, { "epoch": 2.8396914738378154, "grad_norm": 0.13551267981529236, "learning_rate": 1.4228580898366317e-08, "loss": 0.5569, "step": 3409 }, { "epoch": 2.8405253283302065, "grad_norm": 0.13712963461875916, "learning_rate": 1.4077970811554152e-08, "loss": 0.6339, "step": 3410 }, { "epoch": 2.8413591828225977, "grad_norm": 0.1384069174528122, "learning_rate": 1.3928156427790771e-08, "loss": 0.5778, "step": 3411 }, { "epoch": 2.8421930373149884, "grad_norm": 0.1400592178106308, "learning_rate": 1.3779137867986012e-08, "loss": 0.5387, "step": 3412 }, { "epoch": 2.8430268918073796, "grad_norm": 0.1311006098985672, "learning_rate": 1.363091525240756e-08, "loss": 0.6006, "step": 3413 }, { "epoch": 2.8438607462997707, "grad_norm": 0.13074593245983124, "learning_rate": 1.3483488700680612e-08, "loss": 0.578, "step": 3414 }, { "epoch": 2.844694600792162, "grad_norm": 0.13855338096618652, "learning_rate": 1.3336858331787992e-08, "loss": 0.5635, "step": 3415 }, { "epoch": 2.845528455284553, "grad_norm": 0.12744846940040588, "learning_rate": 1.3191024264070039e-08, "loss": 0.5299, "step": 3416 }, { "epoch": 2.8463623097769437, "grad_norm": 0.133438840508461, "learning_rate": 1.3045986615224159e-08, "loss": 0.6169, "step": 3417 }, { "epoch": 2.847196164269335, "grad_norm": 0.1427011340856552, "learning_rate": 1.290174550230516e-08, "loss": 0.6139, "step": 3418 }, { "epoch": 2.848030018761726, "grad_norm": 0.1345617026090622, "learning_rate": 1.2758301041725039e-08, "loss": 0.63, "step": 3419 }, { "epoch": 2.848863873254117, "grad_norm": 0.16132883727550507, "learning_rate": 1.2615653349252741e-08, "loss": 0.5631, "step": 3420 }, { "epoch": 2.8496977277465083, "grad_norm": 0.13171128928661346, "learning_rate": 1.247380254001429e-08, "loss": 0.6195, "step": 3421 }, { "epoch": 2.8505315822388995, "grad_norm": 0.13379457592964172, "learning_rate": 1.2332748728492325e-08, "loss": 0.6468, "step": 3422 }, { "epoch": 2.8513654367312906, "grad_norm": 0.14426220953464508, "learning_rate": 1.2192492028526568e-08, "loss": 0.6204, "step": 3423 }, { "epoch": 2.8521992912236813, "grad_norm": 0.12451158463954926, "learning_rate": 1.2053032553313358e-08, "loss": 0.5671, "step": 3424 }, { "epoch": 2.8530331457160725, "grad_norm": 0.1390177309513092, "learning_rate": 1.1914370415405217e-08, "loss": 0.5637, "step": 3425 }, { "epoch": 2.8538670002084636, "grad_norm": 0.14198832213878632, "learning_rate": 1.1776505726711738e-08, "loss": 0.5807, "step": 3426 }, { "epoch": 2.8547008547008548, "grad_norm": 0.14354586601257324, "learning_rate": 1.1639438598498697e-08, "loss": 0.6193, "step": 3427 }, { "epoch": 2.855534709193246, "grad_norm": 0.13480442762374878, "learning_rate": 1.1503169141388048e-08, "loss": 0.6226, "step": 3428 }, { "epoch": 2.8563685636856366, "grad_norm": 0.1335025429725647, "learning_rate": 1.136769746535815e-08, "loss": 0.555, "step": 3429 }, { "epoch": 2.857202418178028, "grad_norm": 0.13503603637218475, "learning_rate": 1.1233023679743326e-08, "loss": 0.6121, "step": 3430 }, { "epoch": 2.858036272670419, "grad_norm": 0.13793711364269257, "learning_rate": 1.1099147893234184e-08, "loss": 0.6009, "step": 3431 }, { "epoch": 2.85887012716281, "grad_norm": 0.13678361475467682, "learning_rate": 1.0966070213877076e-08, "loss": 0.5543, "step": 3432 }, { "epoch": 2.8597039816552012, "grad_norm": 0.13579510152339935, "learning_rate": 1.0833790749074423e-08, "loss": 0.6516, "step": 3433 }, { "epoch": 2.8605378361475924, "grad_norm": 0.12949161231517792, "learning_rate": 1.070230960558427e-08, "loss": 0.5899, "step": 3434 }, { "epoch": 2.8613716906399835, "grad_norm": 0.1332952082157135, "learning_rate": 1.0571626889520401e-08, "loss": 0.6227, "step": 3435 }, { "epoch": 2.8622055451323742, "grad_norm": 0.13445498049259186, "learning_rate": 1.0441742706352452e-08, "loss": 0.5595, "step": 3436 }, { "epoch": 2.8630393996247654, "grad_norm": 0.1315450817346573, "learning_rate": 1.0312657160905124e-08, "loss": 0.5377, "step": 3437 }, { "epoch": 2.8638732541171565, "grad_norm": 0.13313435018062592, "learning_rate": 1.0184370357358862e-08, "loss": 0.6007, "step": 3438 }, { "epoch": 2.8647071086095477, "grad_norm": 0.1431698352098465, "learning_rate": 1.0056882399249622e-08, "loss": 0.6246, "step": 3439 }, { "epoch": 2.865540963101939, "grad_norm": 0.1385825276374817, "learning_rate": 9.930193389468322e-09, "loss": 0.6199, "step": 3440 }, { "epoch": 2.8663748175943295, "grad_norm": 0.14564557373523712, "learning_rate": 9.804303430261174e-09, "loss": 0.5946, "step": 3441 }, { "epoch": 2.867208672086721, "grad_norm": 0.13472293317317963, "learning_rate": 9.679212623229571e-09, "loss": 0.6071, "step": 3442 }, { "epoch": 2.868042526579112, "grad_norm": 0.14901019632816315, "learning_rate": 9.554921069330092e-09, "loss": 0.5897, "step": 3443 }, { "epoch": 2.868876381071503, "grad_norm": 0.13316750526428223, "learning_rate": 9.431428868873825e-09, "loss": 0.5931, "step": 3444 }, { "epoch": 2.869710235563894, "grad_norm": 0.14676709473133087, "learning_rate": 9.30873612152705e-09, "loss": 0.5984, "step": 3445 }, { "epoch": 2.8705440900562853, "grad_norm": 0.1378309726715088, "learning_rate": 9.186842926311e-09, "loss": 0.59, "step": 3446 }, { "epoch": 2.8713779445486765, "grad_norm": 0.12724530696868896, "learning_rate": 9.065749381601096e-09, "loss": 0.552, "step": 3447 }, { "epoch": 2.872211799041067, "grad_norm": 0.12778747081756592, "learning_rate": 8.945455585127825e-09, "loss": 0.6162, "step": 3448 }, { "epoch": 2.8730456535334583, "grad_norm": 0.137205109000206, "learning_rate": 8.825961633976086e-09, "loss": 0.5375, "step": 3449 }, { "epoch": 2.8738795080258495, "grad_norm": 0.1317429095506668, "learning_rate": 8.707267624585291e-09, "loss": 0.572, "step": 3450 }, { "epoch": 2.8747133625182406, "grad_norm": 0.14071056246757507, "learning_rate": 8.589373652749033e-09, "loss": 0.6069, "step": 3451 }, { "epoch": 2.8755472170106318, "grad_norm": 0.13120976090431213, "learning_rate": 8.472279813615646e-09, "loss": 0.5887, "step": 3452 }, { "epoch": 2.8763810715030225, "grad_norm": 0.13209153711795807, "learning_rate": 8.355986201687316e-09, "loss": 0.6163, "step": 3453 }, { "epoch": 2.877214925995414, "grad_norm": 0.13731753826141357, "learning_rate": 8.240492910820407e-09, "loss": 0.5864, "step": 3454 }, { "epoch": 2.8780487804878048, "grad_norm": 0.1386546641588211, "learning_rate": 8.125800034225693e-09, "loss": 0.6062, "step": 3455 }, { "epoch": 2.878882634980196, "grad_norm": 0.13179557025432587, "learning_rate": 8.011907664467798e-09, "loss": 0.6015, "step": 3456 }, { "epoch": 2.879716489472587, "grad_norm": 0.1399107426404953, "learning_rate": 7.898815893465082e-09, "loss": 0.5581, "step": 3457 }, { "epoch": 2.880550343964978, "grad_norm": 0.13265350461006165, "learning_rate": 7.78652481248998e-09, "loss": 0.5804, "step": 3458 }, { "epoch": 2.8813841984573694, "grad_norm": 0.1345401108264923, "learning_rate": 7.675034512168777e-09, "loss": 0.5722, "step": 3459 }, { "epoch": 2.88221805294976, "grad_norm": 0.13211511075496674, "learning_rate": 7.564345082481383e-09, "loss": 0.5666, "step": 3460 }, { "epoch": 2.8830519074421512, "grad_norm": 0.1365034133195877, "learning_rate": 7.454456612761228e-09, "loss": 0.6178, "step": 3461 }, { "epoch": 2.8838857619345424, "grad_norm": 0.13708631694316864, "learning_rate": 7.345369191695483e-09, "loss": 0.5751, "step": 3462 }, { "epoch": 2.8847196164269335, "grad_norm": 0.1312311738729477, "learning_rate": 7.237082907324832e-09, "loss": 0.603, "step": 3463 }, { "epoch": 2.8855534709193247, "grad_norm": 0.12686726450920105, "learning_rate": 7.129597847043256e-09, "loss": 0.5629, "step": 3464 }, { "epoch": 2.8863873254117154, "grad_norm": 0.1391042321920395, "learning_rate": 7.022914097598254e-09, "loss": 0.6266, "step": 3465 }, { "epoch": 2.887221179904107, "grad_norm": 0.14168620109558105, "learning_rate": 6.917031745090507e-09, "loss": 0.5689, "step": 3466 }, { "epoch": 2.8880550343964977, "grad_norm": 0.13832350075244904, "learning_rate": 6.811950874973993e-09, "loss": 0.6038, "step": 3467 }, { "epoch": 2.888888888888889, "grad_norm": 0.1407477855682373, "learning_rate": 6.707671572055651e-09, "loss": 0.6372, "step": 3468 }, { "epoch": 2.88972274338128, "grad_norm": 0.13231201469898224, "learning_rate": 6.6041939204957155e-09, "loss": 0.5956, "step": 3469 }, { "epoch": 2.890556597873671, "grad_norm": 0.130508154630661, "learning_rate": 6.501518003807271e-09, "loss": 0.621, "step": 3470 }, { "epoch": 2.8913904523660623, "grad_norm": 0.13559859991073608, "learning_rate": 6.3996439048567e-09, "loss": 0.5521, "step": 3471 }, { "epoch": 2.892224306858453, "grad_norm": 0.13799329102039337, "learning_rate": 6.298571705862787e-09, "loss": 0.6098, "step": 3472 }, { "epoch": 2.893058161350844, "grad_norm": 0.14324386417865753, "learning_rate": 6.198301488397395e-09, "loss": 0.6216, "step": 3473 }, { "epoch": 2.8938920158432353, "grad_norm": 0.13183069229125977, "learning_rate": 6.098833333385122e-09, "loss": 0.5785, "step": 3474 }, { "epoch": 2.8947258703356264, "grad_norm": 0.13131730258464813, "learning_rate": 6.000167321103311e-09, "loss": 0.5471, "step": 3475 }, { "epoch": 2.8955597248280176, "grad_norm": 0.1318100392818451, "learning_rate": 5.902303531181707e-09, "loss": 0.5691, "step": 3476 }, { "epoch": 2.8963935793204088, "grad_norm": 0.1341163069009781, "learning_rate": 5.805242042602798e-09, "loss": 0.5945, "step": 3477 }, { "epoch": 2.8972274338128, "grad_norm": 0.1387287676334381, "learning_rate": 5.7089829337014784e-09, "loss": 0.5524, "step": 3478 }, { "epoch": 2.8980612883051906, "grad_norm": 0.12849865853786469, "learning_rate": 5.613526282165049e-09, "loss": 0.6503, "step": 3479 }, { "epoch": 2.8988951427975818, "grad_norm": 0.12773889303207397, "learning_rate": 5.518872165033328e-09, "loss": 0.5839, "step": 3480 }, { "epoch": 2.899728997289973, "grad_norm": 0.12893427908420563, "learning_rate": 5.425020658698209e-09, "loss": 0.5371, "step": 3481 }, { "epoch": 2.900562851782364, "grad_norm": 0.13321800529956818, "learning_rate": 5.331971838904104e-09, "loss": 0.5849, "step": 3482 }, { "epoch": 2.901396706274755, "grad_norm": 0.13946333527565002, "learning_rate": 5.239725780747384e-09, "loss": 0.5579, "step": 3483 }, { "epoch": 2.902230560767146, "grad_norm": 0.13679048418998718, "learning_rate": 5.148282558676387e-09, "loss": 0.5843, "step": 3484 }, { "epoch": 2.903064415259537, "grad_norm": 0.13140372931957245, "learning_rate": 5.057642246492077e-09, "loss": 0.5889, "step": 3485 }, { "epoch": 2.903898269751928, "grad_norm": 0.13622407615184784, "learning_rate": 4.96780491734694e-09, "loss": 0.6459, "step": 3486 }, { "epoch": 2.9047321242443194, "grad_norm": 0.14550499618053436, "learning_rate": 4.8787706437454226e-09, "loss": 0.5944, "step": 3487 }, { "epoch": 2.9055659787367105, "grad_norm": 0.1305081993341446, "learning_rate": 4.790539497543933e-09, "loss": 0.5292, "step": 3488 }, { "epoch": 2.9063998332291017, "grad_norm": 0.13372571766376495, "learning_rate": 4.703111549950956e-09, "loss": 0.5904, "step": 3489 }, { "epoch": 2.907233687721493, "grad_norm": 0.13223007321357727, "learning_rate": 4.6164868715263815e-09, "loss": 0.5251, "step": 3490 }, { "epoch": 2.9080675422138835, "grad_norm": 0.1299707293510437, "learning_rate": 4.530665532181954e-09, "loss": 0.59, "step": 3491 }, { "epoch": 2.9089013967062747, "grad_norm": 0.13822266459465027, "learning_rate": 4.445647601181046e-09, "loss": 0.5531, "step": 3492 }, { "epoch": 2.909735251198666, "grad_norm": 0.13910748064517975, "learning_rate": 4.361433147138771e-09, "loss": 0.5882, "step": 3493 }, { "epoch": 2.910569105691057, "grad_norm": 0.12613445520401, "learning_rate": 4.278022238021428e-09, "loss": 0.5321, "step": 3494 }, { "epoch": 2.911402960183448, "grad_norm": 0.13864509761333466, "learning_rate": 4.1954149411471686e-09, "loss": 0.5883, "step": 3495 }, { "epoch": 2.912236814675839, "grad_norm": 0.13213272392749786, "learning_rate": 4.113611323185551e-09, "loss": 0.574, "step": 3496 }, { "epoch": 2.91307066916823, "grad_norm": 0.13092045485973358, "learning_rate": 4.032611450157208e-09, "loss": 0.5245, "step": 3497 }, { "epoch": 2.913904523660621, "grad_norm": 0.1245381310582161, "learning_rate": 3.952415387434293e-09, "loss": 0.5328, "step": 3498 }, { "epoch": 2.9147383781530123, "grad_norm": 0.13974323868751526, "learning_rate": 3.873023199740366e-09, "loss": 0.5639, "step": 3499 }, { "epoch": 2.9155722326454034, "grad_norm": 0.13678237795829773, "learning_rate": 3.794434951149949e-09, "loss": 0.5984, "step": 3500 }, { "epoch": 2.9164060871377946, "grad_norm": 0.13763700425624847, "learning_rate": 3.716650705088975e-09, "loss": 0.6297, "step": 3501 }, { "epoch": 2.9172399416301857, "grad_norm": 0.1303861439228058, "learning_rate": 3.6396705243343374e-09, "loss": 0.5556, "step": 3502 }, { "epoch": 2.9180737961225764, "grad_norm": 0.1321723759174347, "learning_rate": 3.563494471014006e-09, "loss": 0.5567, "step": 3503 }, { "epoch": 2.9189076506149676, "grad_norm": 0.13992999494075775, "learning_rate": 3.4881226066070246e-09, "loss": 0.6092, "step": 3504 }, { "epoch": 2.9197415051073587, "grad_norm": 0.13959458470344543, "learning_rate": 3.4135549919434015e-09, "loss": 0.6001, "step": 3505 }, { "epoch": 2.92057535959975, "grad_norm": 0.14178508520126343, "learning_rate": 3.3397916872039968e-09, "loss": 0.593, "step": 3506 }, { "epoch": 2.921409214092141, "grad_norm": 0.13454732298851013, "learning_rate": 3.266832751920523e-09, "loss": 0.5884, "step": 3507 }, { "epoch": 2.9222430685845318, "grad_norm": 0.12983369827270508, "learning_rate": 3.194678244975768e-09, "loss": 0.6062, "step": 3508 }, { "epoch": 2.9230769230769234, "grad_norm": 0.13077548146247864, "learning_rate": 3.123328224602817e-09, "loss": 0.6348, "step": 3509 }, { "epoch": 2.923910777569314, "grad_norm": 0.13693194091320038, "learning_rate": 3.0527827483860514e-09, "loss": 0.5853, "step": 3510 }, { "epoch": 2.924744632061705, "grad_norm": 0.1398155689239502, "learning_rate": 2.9830418732601503e-09, "loss": 0.5914, "step": 3511 }, { "epoch": 2.9255784865540964, "grad_norm": 0.1330822855234146, "learning_rate": 2.914105655510424e-09, "loss": 0.5933, "step": 3512 }, { "epoch": 2.9264123410464875, "grad_norm": 0.1414237916469574, "learning_rate": 2.845974150773034e-09, "loss": 0.6076, "step": 3513 }, { "epoch": 2.9272461955388787, "grad_norm": 0.14010551571846008, "learning_rate": 2.7786474140343298e-09, "loss": 0.6257, "step": 3514 }, { "epoch": 2.9280800500312694, "grad_norm": 0.13469921052455902, "learning_rate": 2.7121254996315126e-09, "loss": 0.5822, "step": 3515 }, { "epoch": 2.9289139045236605, "grad_norm": 0.12613841891288757, "learning_rate": 2.646408461252081e-09, "loss": 0.5375, "step": 3516 }, { "epoch": 2.9297477590160517, "grad_norm": 0.1363915503025055, "learning_rate": 2.581496351933832e-09, "loss": 0.6204, "step": 3517 }, { "epoch": 2.930581613508443, "grad_norm": 0.13360045850276947, "learning_rate": 2.5173892240650807e-09, "loss": 0.6222, "step": 3518 }, { "epoch": 2.931415468000834, "grad_norm": 0.13514232635498047, "learning_rate": 2.4540871293845524e-09, "loss": 0.5719, "step": 3519 }, { "epoch": 2.9322493224932247, "grad_norm": 0.1346537470817566, "learning_rate": 2.3915901189811572e-09, "loss": 0.5909, "step": 3520 }, { "epoch": 2.9330831769856163, "grad_norm": 0.13728389143943787, "learning_rate": 2.3298982432937708e-09, "loss": 0.6022, "step": 3521 }, { "epoch": 2.933917031478007, "grad_norm": 0.14774242043495178, "learning_rate": 2.269011552112121e-09, "loss": 0.6069, "step": 3522 }, { "epoch": 2.934750885970398, "grad_norm": 0.1371455043554306, "learning_rate": 2.208930094575345e-09, "loss": 0.5524, "step": 3523 }, { "epoch": 2.9355847404627893, "grad_norm": 0.12859797477722168, "learning_rate": 2.149653919173433e-09, "loss": 0.5689, "step": 3524 }, { "epoch": 2.9364185949551804, "grad_norm": 0.12820853292942047, "learning_rate": 2.0911830737457836e-09, "loss": 0.6056, "step": 3525 }, { "epoch": 2.9372524494475716, "grad_norm": 0.13713565468788147, "learning_rate": 2.0335176054823156e-09, "loss": 0.5488, "step": 3526 }, { "epoch": 2.9380863039399623, "grad_norm": 0.13549277186393738, "learning_rate": 1.976657560922912e-09, "loss": 0.612, "step": 3527 }, { "epoch": 2.9389201584323534, "grad_norm": 0.13806621730327606, "learning_rate": 1.920602985957087e-09, "loss": 0.581, "step": 3528 }, { "epoch": 2.9397540129247446, "grad_norm": 0.13658861815929413, "learning_rate": 1.865353925824653e-09, "loss": 0.623, "step": 3529 }, { "epoch": 2.9405878674171357, "grad_norm": 0.1355557143688202, "learning_rate": 1.810910425115164e-09, "loss": 0.5944, "step": 3530 }, { "epoch": 2.941421721909527, "grad_norm": 0.1372803896665573, "learning_rate": 1.7572725277680278e-09, "loss": 0.6322, "step": 3531 }, { "epoch": 2.9422555764019176, "grad_norm": 0.13330133259296417, "learning_rate": 1.7044402770725052e-09, "loss": 0.602, "step": 3532 }, { "epoch": 2.943089430894309, "grad_norm": 0.13340924680233002, "learning_rate": 1.652413715667822e-09, "loss": 0.5922, "step": 3533 }, { "epoch": 2.9439232853867, "grad_norm": 0.1386307179927826, "learning_rate": 1.6011928855425017e-09, "loss": 0.5757, "step": 3534 }, { "epoch": 2.944757139879091, "grad_norm": 0.1340595781803131, "learning_rate": 1.5507778280353655e-09, "loss": 0.5991, "step": 3535 }, { "epoch": 2.945590994371482, "grad_norm": 0.13373708724975586, "learning_rate": 1.5011685838343113e-09, "loss": 0.5737, "step": 3536 }, { "epoch": 2.9464248488638733, "grad_norm": 0.1402965486049652, "learning_rate": 1.4523651929775337e-09, "loss": 0.6285, "step": 3537 }, { "epoch": 2.9472587033562645, "grad_norm": 0.13696572184562683, "learning_rate": 1.4043676948524152e-09, "loss": 0.6445, "step": 3538 }, { "epoch": 2.948092557848655, "grad_norm": 0.1440492570400238, "learning_rate": 1.3571761281960804e-09, "loss": 0.6, "step": 3539 }, { "epoch": 2.9489264123410464, "grad_norm": 0.13235493004322052, "learning_rate": 1.310790531095063e-09, "loss": 0.5512, "step": 3540 }, { "epoch": 2.9497602668334375, "grad_norm": 0.13626453280448914, "learning_rate": 1.2652109409856393e-09, "loss": 0.5963, "step": 3541 }, { "epoch": 2.9505941213258287, "grad_norm": 0.1433640420436859, "learning_rate": 1.2204373946534951e-09, "loss": 0.607, "step": 3542 }, { "epoch": 2.95142797581822, "grad_norm": 0.13188263773918152, "learning_rate": 1.1764699282337254e-09, "loss": 0.6084, "step": 3543 }, { "epoch": 2.952261830310611, "grad_norm": 0.14081910252571106, "learning_rate": 1.1333085772110563e-09, "loss": 0.5635, "step": 3544 }, { "epoch": 2.953095684803002, "grad_norm": 0.13790731132030487, "learning_rate": 1.0909533764194013e-09, "loss": 0.5927, "step": 3545 }, { "epoch": 2.953929539295393, "grad_norm": 0.13086406886577606, "learning_rate": 1.0494043600423052e-09, "loss": 0.601, "step": 3546 }, { "epoch": 2.954763393787784, "grad_norm": 0.13492035865783691, "learning_rate": 1.0086615616121674e-09, "loss": 0.5422, "step": 3547 }, { "epoch": 2.955597248280175, "grad_norm": 0.13792026042938232, "learning_rate": 9.687250140114621e-10, "loss": 0.6353, "step": 3548 }, { "epoch": 2.9564311027725663, "grad_norm": 0.13961535692214966, "learning_rate": 9.295947494714073e-10, "loss": 0.623, "step": 3549 }, { "epoch": 2.9572649572649574, "grad_norm": 0.13909204304218292, "learning_rate": 8.912707995726298e-10, "loss": 0.5866, "step": 3550 }, { "epoch": 2.958098811757348, "grad_norm": 0.13544286787509918, "learning_rate": 8.53753195245055e-10, "loss": 0.6032, "step": 3551 }, { "epoch": 2.9589326662497393, "grad_norm": 0.13512249290943146, "learning_rate": 8.170419667679063e-10, "loss": 0.5788, "step": 3552 }, { "epoch": 2.9597665207421304, "grad_norm": 0.13385283946990967, "learning_rate": 7.811371437693725e-10, "loss": 0.6433, "step": 3553 }, { "epoch": 2.9606003752345216, "grad_norm": 0.14141035079956055, "learning_rate": 7.46038755227163e-10, "loss": 0.6251, "step": 3554 }, { "epoch": 2.9614342297269127, "grad_norm": 0.12954039871692657, "learning_rate": 7.117468294678408e-10, "loss": 0.5992, "step": 3555 }, { "epoch": 2.962268084219304, "grad_norm": 0.14412134885787964, "learning_rate": 6.782613941672677e-10, "loss": 0.5485, "step": 3556 }, { "epoch": 2.963101938711695, "grad_norm": 0.14092476665973663, "learning_rate": 6.455824763503815e-10, "loss": 0.5706, "step": 3557 }, { "epoch": 2.9639357932040857, "grad_norm": 0.1278911679983139, "learning_rate": 6.137101023910852e-10, "loss": 0.6091, "step": 3558 }, { "epoch": 2.964769647696477, "grad_norm": 0.13343612849712372, "learning_rate": 5.82644298012469e-10, "loss": 0.5654, "step": 3559 }, { "epoch": 2.965603502188868, "grad_norm": 0.13848499953746796, "learning_rate": 5.523850882866998e-10, "loss": 0.5886, "step": 3560 }, { "epoch": 2.966437356681259, "grad_norm": 0.1324864774942398, "learning_rate": 5.229324976350203e-10, "loss": 0.5909, "step": 3561 }, { "epoch": 2.9672712111736503, "grad_norm": 0.14334915578365326, "learning_rate": 4.942865498274162e-10, "loss": 0.5984, "step": 3562 }, { "epoch": 2.968105065666041, "grad_norm": 0.13357843458652496, "learning_rate": 4.664472679830611e-10, "loss": 0.6262, "step": 3563 }, { "epoch": 2.968938920158432, "grad_norm": 0.14388038218021393, "learning_rate": 4.3941467457009327e-10, "loss": 0.5887, "step": 3564 }, { "epoch": 2.9697727746508233, "grad_norm": 0.14118964970111847, "learning_rate": 4.1318879140561647e-10, "loss": 0.5909, "step": 3565 }, { "epoch": 2.9706066291432145, "grad_norm": 0.13682524859905243, "learning_rate": 3.877696396555885e-10, "loss": 0.5999, "step": 3566 }, { "epoch": 2.9714404836356056, "grad_norm": 0.14852957427501678, "learning_rate": 3.631572398348215e-10, "loss": 0.5844, "step": 3567 }, { "epoch": 2.972274338127997, "grad_norm": 0.1391034871339798, "learning_rate": 3.3935161180720374e-10, "loss": 0.5864, "step": 3568 }, { "epoch": 2.973108192620388, "grad_norm": 0.13956484198570251, "learning_rate": 3.1635277478547774e-10, "loss": 0.6029, "step": 3569 }, { "epoch": 2.9739420471127787, "grad_norm": 0.13877858221530914, "learning_rate": 2.941607473311292e-10, "loss": 0.6055, "step": 3570 }, { "epoch": 2.97477590160517, "grad_norm": 0.13228566944599152, "learning_rate": 2.7277554735449793e-10, "loss": 0.6097, "step": 3571 }, { "epoch": 2.975609756097561, "grad_norm": 0.14116641879081726, "learning_rate": 2.521971921148891e-10, "loss": 0.5765, "step": 3572 }, { "epoch": 2.976443610589952, "grad_norm": 0.14230525493621826, "learning_rate": 2.32425698220462e-10, "loss": 0.5518, "step": 3573 }, { "epoch": 2.9772774650823433, "grad_norm": 0.13284701108932495, "learning_rate": 2.13461081627897e-10, "loss": 0.5671, "step": 3574 }, { "epoch": 2.978111319574734, "grad_norm": 0.13724222779273987, "learning_rate": 1.9530335764306182e-10, "loss": 0.5653, "step": 3575 }, { "epoch": 2.9789451740671256, "grad_norm": 0.13540944457054138, "learning_rate": 1.7795254092034528e-10, "loss": 0.6368, "step": 3576 }, { "epoch": 2.9797790285595163, "grad_norm": 0.12903504073619843, "learning_rate": 1.6140864546299039e-10, "loss": 0.5662, "step": 3577 }, { "epoch": 2.9806128830519074, "grad_norm": 0.12735681235790253, "learning_rate": 1.4567168462298329e-10, "loss": 0.5982, "step": 3578 }, { "epoch": 2.9814467375442986, "grad_norm": 0.1325932890176773, "learning_rate": 1.307416711010534e-10, "loss": 0.6085, "step": 3579 }, { "epoch": 2.9822805920366897, "grad_norm": 0.14488919079303741, "learning_rate": 1.166186169466732e-10, "loss": 0.6152, "step": 3580 }, { "epoch": 2.983114446529081, "grad_norm": 0.1415412425994873, "learning_rate": 1.0330253355816942e-10, "loss": 0.5471, "step": 3581 }, { "epoch": 2.9839483010214716, "grad_norm": 0.13478343188762665, "learning_rate": 9.079343168227893e-11, "loss": 0.5954, "step": 3582 }, { "epoch": 2.9847821555138627, "grad_norm": 0.13751602172851562, "learning_rate": 7.90913214149258e-11, "loss": 0.6013, "step": 3583 }, { "epoch": 2.985616010006254, "grad_norm": 0.15262840688228607, "learning_rate": 6.819621220033322e-11, "loss": 0.6151, "step": 3584 }, { "epoch": 2.986449864498645, "grad_norm": 0.1271335780620575, "learning_rate": 5.810811283146755e-11, "loss": 0.5653, "step": 3585 }, { "epoch": 2.987283718991036, "grad_norm": 0.12459156662225723, "learning_rate": 4.882703145026035e-11, "loss": 0.6239, "step": 3586 }, { "epoch": 2.988117573483427, "grad_norm": 0.13420237600803375, "learning_rate": 4.03529755470533e-11, "loss": 0.6027, "step": 3587 }, { "epoch": 2.9889514279758185, "grad_norm": 0.1399519443511963, "learning_rate": 3.2685951960931266e-11, "loss": 0.5939, "step": 3588 }, { "epoch": 2.989785282468209, "grad_norm": 0.14246320724487305, "learning_rate": 2.5825966879833293e-11, "loss": 0.5824, "step": 3589 }, { "epoch": 2.9906191369606003, "grad_norm": 0.1423949897289276, "learning_rate": 1.9773025839997516e-11, "loss": 0.617, "step": 3590 }, { "epoch": 2.9914529914529915, "grad_norm": 0.13538257777690887, "learning_rate": 1.4527133726627283e-11, "loss": 0.5882, "step": 3591 }, { "epoch": 2.9922868459453826, "grad_norm": 0.1352311223745346, "learning_rate": 1.0088294773558104e-11, "loss": 0.6005, "step": 3592 }, { "epoch": 2.993120700437774, "grad_norm": 0.13805273175239563, "learning_rate": 6.456512563257633e-12, "loss": 0.6175, "step": 3593 }, { "epoch": 2.9939545549301645, "grad_norm": 0.141433984041214, "learning_rate": 3.6317900266036316e-12, "loss": 0.5934, "step": 3594 }, { "epoch": 2.9947884094225556, "grad_norm": 0.13686375319957733, "learning_rate": 1.6141294435501052e-12, "loss": 0.6121, "step": 3595 }, { "epoch": 2.995622263914947, "grad_norm": 0.14368127286434174, "learning_rate": 4.0353244235014075e-13, "loss": 0.6347, "step": 3596 }, { "epoch": 2.996456118407338, "grad_norm": 0.1274050921201706, "learning_rate": 0.0, "loss": 0.54, "step": 3597 } ], "logging_steps": 1, "max_steps": 3597, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1199, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.6359256425240723e+19, "train_batch_size": 6, "trial_name": null, "trial_params": null }