{ "best_global_step": 3689, "best_metric": 88.77438705594233, "best_model_checkpoint": "whisper-tiny-bfloat16-sada/checkpoints/checkpoint-3689", "epoch": 0.24994918354902093, "eval_steps": 3689, "global_step": 3689, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 6.775526797208483e-05, "grad_norm": 132.70848083496094, "learning_rate": 0.0, "loss": 4.4561, "step": 1 }, { "epoch": 0.00013551053594416967, "grad_norm": 133.01596069335938, "learning_rate": 1.3333333333333334e-07, "loss": 4.8642, "step": 2 }, { "epoch": 0.0002032658039162545, "grad_norm": 106.03809356689453, "learning_rate": 2.6666666666666667e-07, "loss": 4.3778, "step": 3 }, { "epoch": 0.00027102107188833934, "grad_norm": 88.75433349609375, "learning_rate": 4.0000000000000003e-07, "loss": 4.0385, "step": 4 }, { "epoch": 0.00033877633986042414, "grad_norm": 101.19609832763672, "learning_rate": 5.333333333333333e-07, "loss": 3.9092, "step": 5 }, { "epoch": 0.000406531607832509, "grad_norm": 67.02723693847656, "learning_rate": 6.666666666666667e-07, "loss": 3.7556, "step": 6 }, { "epoch": 0.0004742868758045938, "grad_norm": 67.52012634277344, "learning_rate": 8.000000000000001e-07, "loss": 3.7119, "step": 7 }, { "epoch": 0.0005420421437766787, "grad_norm": 83.27348327636719, "learning_rate": 9.333333333333334e-07, "loss": 3.6125, "step": 8 }, { "epoch": 0.0006097974117487635, "grad_norm": 84.20647430419922, "learning_rate": 1.0666666666666667e-06, "loss": 4.1354, "step": 9 }, { "epoch": 0.0006775526797208483, "grad_norm": 110.12113952636719, "learning_rate": 1.2000000000000002e-06, "loss": 3.9594, "step": 10 }, { "epoch": 0.0007453079476929331, "grad_norm": 84.61614990234375, "learning_rate": 1.3333333333333334e-06, "loss": 4.1415, "step": 11 }, { "epoch": 0.000813063215665018, "grad_norm": 119.0740966796875, "learning_rate": 1.4666666666666667e-06, "loss": 4.0689, "step": 12 }, { "epoch": 0.0008808184836371028, "grad_norm": 122.41352081298828, "learning_rate": 1.6000000000000001e-06, "loss": 4.0834, "step": 13 }, { "epoch": 0.0009485737516091876, "grad_norm": 99.06126403808594, "learning_rate": 1.7333333333333334e-06, "loss": 4.4224, "step": 14 }, { "epoch": 0.0010163290195812724, "grad_norm": 113.89531707763672, "learning_rate": 1.8666666666666669e-06, "loss": 4.0188, "step": 15 }, { "epoch": 0.0010840842875533573, "grad_norm": 70.24198150634766, "learning_rate": 2.0000000000000003e-06, "loss": 3.3951, "step": 16 }, { "epoch": 0.001151839555525442, "grad_norm": 67.99130249023438, "learning_rate": 2.1333333333333334e-06, "loss": 3.7385, "step": 17 }, { "epoch": 0.001219594823497527, "grad_norm": 54.554779052734375, "learning_rate": 2.266666666666667e-06, "loss": 2.9009, "step": 18 }, { "epoch": 0.0012873500914696117, "grad_norm": 47.79149627685547, "learning_rate": 2.4000000000000003e-06, "loss": 3.0572, "step": 19 }, { "epoch": 0.0013551053594416966, "grad_norm": 50.05776596069336, "learning_rate": 2.5333333333333334e-06, "loss": 2.8678, "step": 20 }, { "epoch": 0.0014228606274137815, "grad_norm": 55.41701889038086, "learning_rate": 2.666666666666667e-06, "loss": 3.3981, "step": 21 }, { "epoch": 0.0014906158953858662, "grad_norm": 57.1912727355957, "learning_rate": 2.8000000000000003e-06, "loss": 3.3247, "step": 22 }, { "epoch": 0.001558371163357951, "grad_norm": 39.77119445800781, "learning_rate": 2.9333333333333333e-06, "loss": 3.1466, "step": 23 }, { "epoch": 0.001626126431330036, "grad_norm": 37.258392333984375, "learning_rate": 3.066666666666667e-06, "loss": 3.0137, "step": 24 }, { "epoch": 0.0016938816993021207, "grad_norm": 41.15428924560547, "learning_rate": 3.2000000000000003e-06, "loss": 2.9908, "step": 25 }, { "epoch": 0.0017616369672742056, "grad_norm": 44.59917449951172, "learning_rate": 3.3333333333333333e-06, "loss": 3.6947, "step": 26 }, { "epoch": 0.0018293922352462903, "grad_norm": 32.7044792175293, "learning_rate": 3.466666666666667e-06, "loss": 2.9421, "step": 27 }, { "epoch": 0.0018971475032183752, "grad_norm": 28.445444107055664, "learning_rate": 3.6e-06, "loss": 2.9849, "step": 28 }, { "epoch": 0.00196490277119046, "grad_norm": 30.879247665405273, "learning_rate": 3.7333333333333337e-06, "loss": 2.9007, "step": 29 }, { "epoch": 0.002032658039162545, "grad_norm": 22.723188400268555, "learning_rate": 3.866666666666667e-06, "loss": 2.3549, "step": 30 }, { "epoch": 0.0021004133071346296, "grad_norm": 23.59457778930664, "learning_rate": 4.000000000000001e-06, "loss": 2.8564, "step": 31 }, { "epoch": 0.0021681685751067147, "grad_norm": 23.313579559326172, "learning_rate": 4.133333333333333e-06, "loss": 2.453, "step": 32 }, { "epoch": 0.0022359238430787994, "grad_norm": 23.877405166625977, "learning_rate": 4.266666666666667e-06, "loss": 2.2042, "step": 33 }, { "epoch": 0.002303679111050884, "grad_norm": 29.79376792907715, "learning_rate": 4.4e-06, "loss": 2.6679, "step": 34 }, { "epoch": 0.002371434379022969, "grad_norm": 21.018878936767578, "learning_rate": 4.533333333333334e-06, "loss": 2.3983, "step": 35 }, { "epoch": 0.002439189646995054, "grad_norm": 30.22182846069336, "learning_rate": 4.666666666666667e-06, "loss": 2.8668, "step": 36 }, { "epoch": 0.0025069449149671386, "grad_norm": 24.619905471801758, "learning_rate": 4.800000000000001e-06, "loss": 2.5986, "step": 37 }, { "epoch": 0.0025747001829392233, "grad_norm": 25.470375061035156, "learning_rate": 4.933333333333333e-06, "loss": 2.4936, "step": 38 }, { "epoch": 0.0026424554509113084, "grad_norm": 19.872745513916016, "learning_rate": 5.066666666666667e-06, "loss": 2.1633, "step": 39 }, { "epoch": 0.002710210718883393, "grad_norm": 24.880964279174805, "learning_rate": 5.2e-06, "loss": 2.6798, "step": 40 }, { "epoch": 0.002777965986855478, "grad_norm": 22.790157318115234, "learning_rate": 5.333333333333334e-06, "loss": 2.487, "step": 41 }, { "epoch": 0.002845721254827563, "grad_norm": 21.404136657714844, "learning_rate": 5.466666666666667e-06, "loss": 2.3383, "step": 42 }, { "epoch": 0.0029134765227996477, "grad_norm": 22.80590057373047, "learning_rate": 5.600000000000001e-06, "loss": 2.6347, "step": 43 }, { "epoch": 0.0029812317907717324, "grad_norm": 20.625160217285156, "learning_rate": 5.733333333333333e-06, "loss": 2.066, "step": 44 }, { "epoch": 0.0030489870587438175, "grad_norm": 21.700828552246094, "learning_rate": 5.866666666666667e-06, "loss": 2.188, "step": 45 }, { "epoch": 0.003116742326715902, "grad_norm": 18.8945255279541, "learning_rate": 6e-06, "loss": 2.1531, "step": 46 }, { "epoch": 0.003184497594687987, "grad_norm": 20.238525390625, "learning_rate": 6.133333333333334e-06, "loss": 1.9306, "step": 47 }, { "epoch": 0.003252252862660072, "grad_norm": 23.74344825744629, "learning_rate": 6.266666666666666e-06, "loss": 2.3206, "step": 48 }, { "epoch": 0.0033200081306321567, "grad_norm": 22.429964065551758, "learning_rate": 6.4000000000000006e-06, "loss": 2.05, "step": 49 }, { "epoch": 0.0033877633986042414, "grad_norm": 22.078792572021484, "learning_rate": 6.533333333333333e-06, "loss": 2.1155, "step": 50 }, { "epoch": 0.003455518666576326, "grad_norm": 19.691747665405273, "learning_rate": 6.666666666666667e-06, "loss": 1.8719, "step": 51 }, { "epoch": 0.0035232739345484113, "grad_norm": 27.412866592407227, "learning_rate": 6.800000000000001e-06, "loss": 2.0943, "step": 52 }, { "epoch": 0.003591029202520496, "grad_norm": 21.631580352783203, "learning_rate": 6.933333333333334e-06, "loss": 2.246, "step": 53 }, { "epoch": 0.0036587844704925807, "grad_norm": 16.729095458984375, "learning_rate": 7.066666666666667e-06, "loss": 1.8101, "step": 54 }, { "epoch": 0.003726539738464666, "grad_norm": 18.381364822387695, "learning_rate": 7.2e-06, "loss": 1.8668, "step": 55 }, { "epoch": 0.0037942950064367505, "grad_norm": 19.661123275756836, "learning_rate": 7.333333333333334e-06, "loss": 2.1912, "step": 56 }, { "epoch": 0.003862050274408835, "grad_norm": 13.772311210632324, "learning_rate": 7.4666666666666675e-06, "loss": 1.7408, "step": 57 }, { "epoch": 0.00392980554238092, "grad_norm": 20.559425354003906, "learning_rate": 7.6e-06, "loss": 1.9717, "step": 58 }, { "epoch": 0.003997560810353005, "grad_norm": 17.385791778564453, "learning_rate": 7.733333333333334e-06, "loss": 1.7681, "step": 59 }, { "epoch": 0.00406531607832509, "grad_norm": 18.455888748168945, "learning_rate": 7.866666666666667e-06, "loss": 1.8822, "step": 60 }, { "epoch": 0.004133071346297174, "grad_norm": 17.53643798828125, "learning_rate": 8.000000000000001e-06, "loss": 1.8158, "step": 61 }, { "epoch": 0.004200826614269259, "grad_norm": 15.640396118164062, "learning_rate": 8.133333333333332e-06, "loss": 1.8786, "step": 62 }, { "epoch": 0.004268581882241345, "grad_norm": 67.14771270751953, "learning_rate": 8.266666666666667e-06, "loss": 1.845, "step": 63 }, { "epoch": 0.004336337150213429, "grad_norm": 16.778207778930664, "learning_rate": 8.400000000000001e-06, "loss": 1.7286, "step": 64 }, { "epoch": 0.004404092418185514, "grad_norm": 20.485218048095703, "learning_rate": 8.533333333333334e-06, "loss": 1.882, "step": 65 }, { "epoch": 0.004471847686157599, "grad_norm": 15.77828598022461, "learning_rate": 8.666666666666668e-06, "loss": 1.7788, "step": 66 }, { "epoch": 0.0045396029541296835, "grad_norm": 20.225608825683594, "learning_rate": 8.8e-06, "loss": 1.9842, "step": 67 }, { "epoch": 0.004607358222101768, "grad_norm": 16.230987548828125, "learning_rate": 8.933333333333333e-06, "loss": 1.4581, "step": 68 }, { "epoch": 0.004675113490073853, "grad_norm": 13.831804275512695, "learning_rate": 9.066666666666667e-06, "loss": 1.6173, "step": 69 }, { "epoch": 0.004742868758045938, "grad_norm": 16.4691104888916, "learning_rate": 9.2e-06, "loss": 1.6055, "step": 70 }, { "epoch": 0.004810624026018023, "grad_norm": 17.788105010986328, "learning_rate": 9.333333333333334e-06, "loss": 1.7996, "step": 71 }, { "epoch": 0.004878379293990108, "grad_norm": 20.043182373046875, "learning_rate": 9.466666666666667e-06, "loss": 1.7224, "step": 72 }, { "epoch": 0.0049461345619621925, "grad_norm": 14.918837547302246, "learning_rate": 9.600000000000001e-06, "loss": 1.4448, "step": 73 }, { "epoch": 0.005013889829934277, "grad_norm": 15.70170783996582, "learning_rate": 9.733333333333334e-06, "loss": 1.7469, "step": 74 }, { "epoch": 0.005081645097906362, "grad_norm": 19.128456115722656, "learning_rate": 9.866666666666667e-06, "loss": 1.9505, "step": 75 }, { "epoch": 0.005149400365878447, "grad_norm": 34.82925796508789, "learning_rate": 1e-05, "loss": 1.6543, "step": 76 }, { "epoch": 0.005217155633850532, "grad_norm": 19.61214256286621, "learning_rate": 1.0133333333333333e-05, "loss": 1.9051, "step": 77 }, { "epoch": 0.005284910901822617, "grad_norm": 20.190847396850586, "learning_rate": 1.0266666666666668e-05, "loss": 1.6016, "step": 78 }, { "epoch": 0.005352666169794702, "grad_norm": 15.873608589172363, "learning_rate": 1.04e-05, "loss": 1.8286, "step": 79 }, { "epoch": 0.005420421437766786, "grad_norm": 17.513385772705078, "learning_rate": 1.0533333333333335e-05, "loss": 1.7122, "step": 80 }, { "epoch": 0.005488176705738871, "grad_norm": 13.794294357299805, "learning_rate": 1.0666666666666667e-05, "loss": 1.5369, "step": 81 }, { "epoch": 0.005555931973710956, "grad_norm": 16.763822555541992, "learning_rate": 1.08e-05, "loss": 1.688, "step": 82 }, { "epoch": 0.005623687241683041, "grad_norm": 15.182608604431152, "learning_rate": 1.0933333333333334e-05, "loss": 1.9076, "step": 83 }, { "epoch": 0.005691442509655126, "grad_norm": 19.11581802368164, "learning_rate": 1.1066666666666667e-05, "loss": 1.6034, "step": 84 }, { "epoch": 0.005759197777627211, "grad_norm": 18.303098678588867, "learning_rate": 1.1200000000000001e-05, "loss": 1.8248, "step": 85 }, { "epoch": 0.005826953045599295, "grad_norm": 17.912731170654297, "learning_rate": 1.1333333333333334e-05, "loss": 1.737, "step": 86 }, { "epoch": 0.00589470831357138, "grad_norm": 18.4353084564209, "learning_rate": 1.1466666666666666e-05, "loss": 1.7141, "step": 87 }, { "epoch": 0.005962463581543465, "grad_norm": 17.006074905395508, "learning_rate": 1.16e-05, "loss": 1.6568, "step": 88 }, { "epoch": 0.0060302188495155494, "grad_norm": 15.084410667419434, "learning_rate": 1.1733333333333333e-05, "loss": 1.3369, "step": 89 }, { "epoch": 0.006097974117487635, "grad_norm": 14.984724044799805, "learning_rate": 1.1866666666666668e-05, "loss": 1.777, "step": 90 }, { "epoch": 0.00616572938545972, "grad_norm": 19.447574615478516, "learning_rate": 1.2e-05, "loss": 1.4225, "step": 91 }, { "epoch": 0.006233484653431804, "grad_norm": 15.212474822998047, "learning_rate": 1.2133333333333335e-05, "loss": 1.4896, "step": 92 }, { "epoch": 0.006301239921403889, "grad_norm": 17.97924041748047, "learning_rate": 1.2266666666666667e-05, "loss": 1.5947, "step": 93 }, { "epoch": 0.006368995189375974, "grad_norm": 13.695365905761719, "learning_rate": 1.24e-05, "loss": 1.471, "step": 94 }, { "epoch": 0.0064367504573480585, "grad_norm": 15.574562072753906, "learning_rate": 1.2533333333333332e-05, "loss": 1.5803, "step": 95 }, { "epoch": 0.006504505725320144, "grad_norm": 16.71872329711914, "learning_rate": 1.2666666666666668e-05, "loss": 1.7408, "step": 96 }, { "epoch": 0.006572260993292229, "grad_norm": 14.232036590576172, "learning_rate": 1.2800000000000001e-05, "loss": 1.4414, "step": 97 }, { "epoch": 0.0066400162612643135, "grad_norm": 14.430774688720703, "learning_rate": 1.2933333333333334e-05, "loss": 1.712, "step": 98 }, { "epoch": 0.006707771529236398, "grad_norm": 15.887603759765625, "learning_rate": 1.3066666666666666e-05, "loss": 1.6097, "step": 99 }, { "epoch": 0.006775526797208483, "grad_norm": 21.155643463134766, "learning_rate": 1.32e-05, "loss": 2.0581, "step": 100 }, { "epoch": 0.0068432820651805676, "grad_norm": 17.30326271057129, "learning_rate": 1.3333333333333333e-05, "loss": 1.2823, "step": 101 }, { "epoch": 0.006911037333152652, "grad_norm": 18.184707641601562, "learning_rate": 1.3466666666666666e-05, "loss": 1.5165, "step": 102 }, { "epoch": 0.006978792601124738, "grad_norm": 23.46563148498535, "learning_rate": 1.3600000000000002e-05, "loss": 1.406, "step": 103 }, { "epoch": 0.0070465478690968225, "grad_norm": 14.586087226867676, "learning_rate": 1.3733333333333335e-05, "loss": 1.8416, "step": 104 }, { "epoch": 0.007114303137068907, "grad_norm": 14.55811595916748, "learning_rate": 1.3866666666666667e-05, "loss": 1.5466, "step": 105 }, { "epoch": 0.007182058405040992, "grad_norm": 20.244586944580078, "learning_rate": 1.4000000000000001e-05, "loss": 1.755, "step": 106 }, { "epoch": 0.007249813673013077, "grad_norm": 11.932172775268555, "learning_rate": 1.4133333333333334e-05, "loss": 1.4543, "step": 107 }, { "epoch": 0.007317568940985161, "grad_norm": 17.406888961791992, "learning_rate": 1.4266666666666667e-05, "loss": 1.4436, "step": 108 }, { "epoch": 0.007385324208957246, "grad_norm": 14.591705322265625, "learning_rate": 1.44e-05, "loss": 1.4088, "step": 109 }, { "epoch": 0.007453079476929332, "grad_norm": 18.028337478637695, "learning_rate": 1.4533333333333335e-05, "loss": 1.6675, "step": 110 }, { "epoch": 0.007520834744901416, "grad_norm": 16.37979507446289, "learning_rate": 1.4666666666666668e-05, "loss": 1.6646, "step": 111 }, { "epoch": 0.007588590012873501, "grad_norm": 19.028308868408203, "learning_rate": 1.48e-05, "loss": 1.701, "step": 112 }, { "epoch": 0.007656345280845586, "grad_norm": 17.28181266784668, "learning_rate": 1.4933333333333335e-05, "loss": 1.3984, "step": 113 }, { "epoch": 0.00772410054881767, "grad_norm": 16.78407859802246, "learning_rate": 1.5066666666666668e-05, "loss": 1.6851, "step": 114 }, { "epoch": 0.007791855816789755, "grad_norm": 15.634642601013184, "learning_rate": 1.52e-05, "loss": 1.6484, "step": 115 }, { "epoch": 0.00785961108476184, "grad_norm": 13.879626274108887, "learning_rate": 1.5333333333333334e-05, "loss": 1.4609, "step": 116 }, { "epoch": 0.007927366352733925, "grad_norm": 16.247541427612305, "learning_rate": 1.546666666666667e-05, "loss": 1.5219, "step": 117 }, { "epoch": 0.00799512162070601, "grad_norm": 20.57497215270996, "learning_rate": 1.56e-05, "loss": 1.6038, "step": 118 }, { "epoch": 0.008062876888678095, "grad_norm": 16.32591438293457, "learning_rate": 1.5733333333333334e-05, "loss": 1.5315, "step": 119 }, { "epoch": 0.00813063215665018, "grad_norm": 18.236255645751953, "learning_rate": 1.586666666666667e-05, "loss": 1.6589, "step": 120 }, { "epoch": 0.008198387424622264, "grad_norm": 20.850723266601562, "learning_rate": 1.6000000000000003e-05, "loss": 1.5556, "step": 121 }, { "epoch": 0.008266142692594349, "grad_norm": 17.931947708129883, "learning_rate": 1.6133333333333334e-05, "loss": 1.8209, "step": 122 }, { "epoch": 0.008333897960566434, "grad_norm": 20.531381607055664, "learning_rate": 1.6266666666666665e-05, "loss": 1.6421, "step": 123 }, { "epoch": 0.008401653228538518, "grad_norm": 15.6729736328125, "learning_rate": 1.6400000000000002e-05, "loss": 1.5652, "step": 124 }, { "epoch": 0.008469408496510603, "grad_norm": 18.139949798583984, "learning_rate": 1.6533333333333333e-05, "loss": 1.5578, "step": 125 }, { "epoch": 0.00853716376448269, "grad_norm": 15.77291202545166, "learning_rate": 1.6666666666666667e-05, "loss": 1.6091, "step": 126 }, { "epoch": 0.008604919032454774, "grad_norm": 18.00165557861328, "learning_rate": 1.6800000000000002e-05, "loss": 1.4725, "step": 127 }, { "epoch": 0.008672674300426859, "grad_norm": 17.33936309814453, "learning_rate": 1.6933333333333333e-05, "loss": 1.4272, "step": 128 }, { "epoch": 0.008740429568398943, "grad_norm": 16.75677490234375, "learning_rate": 1.7066666666666667e-05, "loss": 1.602, "step": 129 }, { "epoch": 0.008808184836371028, "grad_norm": 17.761140823364258, "learning_rate": 1.7199999999999998e-05, "loss": 1.8523, "step": 130 }, { "epoch": 0.008875940104343113, "grad_norm": 15.888497352600098, "learning_rate": 1.7333333333333336e-05, "loss": 1.5594, "step": 131 }, { "epoch": 0.008943695372315198, "grad_norm": 19.62826156616211, "learning_rate": 1.7466666666666667e-05, "loss": 1.6909, "step": 132 }, { "epoch": 0.009011450640287282, "grad_norm": 15.62592887878418, "learning_rate": 1.76e-05, "loss": 1.5873, "step": 133 }, { "epoch": 0.009079205908259367, "grad_norm": 17.64893913269043, "learning_rate": 1.7733333333333335e-05, "loss": 1.5681, "step": 134 }, { "epoch": 0.009146961176231452, "grad_norm": 15.14854907989502, "learning_rate": 1.7866666666666666e-05, "loss": 1.3324, "step": 135 }, { "epoch": 0.009214716444203536, "grad_norm": 18.115419387817383, "learning_rate": 1.8e-05, "loss": 1.5512, "step": 136 }, { "epoch": 0.009282471712175621, "grad_norm": 18.2511043548584, "learning_rate": 1.8133333333333335e-05, "loss": 1.6405, "step": 137 }, { "epoch": 0.009350226980147706, "grad_norm": 13.296855926513672, "learning_rate": 1.826666666666667e-05, "loss": 1.2269, "step": 138 }, { "epoch": 0.009417982248119792, "grad_norm": 17.17320442199707, "learning_rate": 1.84e-05, "loss": 1.4031, "step": 139 }, { "epoch": 0.009485737516091877, "grad_norm": 15.399650573730469, "learning_rate": 1.8533333333333334e-05, "loss": 1.4793, "step": 140 }, { "epoch": 0.009553492784063962, "grad_norm": 15.0385160446167, "learning_rate": 1.866666666666667e-05, "loss": 1.4129, "step": 141 }, { "epoch": 0.009621248052036046, "grad_norm": 14.586984634399414, "learning_rate": 1.88e-05, "loss": 1.1959, "step": 142 }, { "epoch": 0.009689003320008131, "grad_norm": 14.939247131347656, "learning_rate": 1.8933333333333334e-05, "loss": 1.6728, "step": 143 }, { "epoch": 0.009756758587980216, "grad_norm": 18.1126651763916, "learning_rate": 1.9066666666666668e-05, "loss": 1.5035, "step": 144 }, { "epoch": 0.0098245138559523, "grad_norm": 16.0823974609375, "learning_rate": 1.9200000000000003e-05, "loss": 1.544, "step": 145 }, { "epoch": 0.009892269123924385, "grad_norm": 17.45062828063965, "learning_rate": 1.9333333333333333e-05, "loss": 1.4593, "step": 146 }, { "epoch": 0.00996002439189647, "grad_norm": 21.030731201171875, "learning_rate": 1.9466666666666668e-05, "loss": 1.6544, "step": 147 }, { "epoch": 0.010027779659868554, "grad_norm": 18.31300163269043, "learning_rate": 1.9600000000000002e-05, "loss": 1.7682, "step": 148 }, { "epoch": 0.01009553492784064, "grad_norm": 13.541064262390137, "learning_rate": 1.9733333333333333e-05, "loss": 1.333, "step": 149 }, { "epoch": 0.010163290195812724, "grad_norm": 13.538237571716309, "learning_rate": 1.9866666666666667e-05, "loss": 1.3629, "step": 150 }, { "epoch": 0.010231045463784809, "grad_norm": 18.393657684326172, "learning_rate": 2e-05, "loss": 1.4467, "step": 151 }, { "epoch": 0.010298800731756893, "grad_norm": 19.18985939025879, "learning_rate": 2.0133333333333336e-05, "loss": 1.612, "step": 152 }, { "epoch": 0.01036655599972898, "grad_norm": 14.230488777160645, "learning_rate": 2.0266666666666667e-05, "loss": 1.4047, "step": 153 }, { "epoch": 0.010434311267701064, "grad_norm": 13.807863235473633, "learning_rate": 2.04e-05, "loss": 1.4977, "step": 154 }, { "epoch": 0.010502066535673149, "grad_norm": 15.149568557739258, "learning_rate": 2.0533333333333336e-05, "loss": 1.3856, "step": 155 }, { "epoch": 0.010569821803645234, "grad_norm": 16.095308303833008, "learning_rate": 2.0666666666666666e-05, "loss": 1.3687, "step": 156 }, { "epoch": 0.010637577071617318, "grad_norm": 14.864961624145508, "learning_rate": 2.08e-05, "loss": 1.5707, "step": 157 }, { "epoch": 0.010705332339589403, "grad_norm": 15.9810152053833, "learning_rate": 2.0933333333333335e-05, "loss": 1.6841, "step": 158 }, { "epoch": 0.010773087607561488, "grad_norm": 16.706584930419922, "learning_rate": 2.106666666666667e-05, "loss": 1.5017, "step": 159 }, { "epoch": 0.010840842875533573, "grad_norm": 18.726655960083008, "learning_rate": 2.12e-05, "loss": 1.7177, "step": 160 }, { "epoch": 0.010908598143505657, "grad_norm": 17.22323989868164, "learning_rate": 2.1333333333333335e-05, "loss": 1.5452, "step": 161 }, { "epoch": 0.010976353411477742, "grad_norm": 14.823019027709961, "learning_rate": 2.146666666666667e-05, "loss": 1.2856, "step": 162 }, { "epoch": 0.011044108679449827, "grad_norm": 14.756085395812988, "learning_rate": 2.16e-05, "loss": 1.4942, "step": 163 }, { "epoch": 0.011111863947421911, "grad_norm": 18.523683547973633, "learning_rate": 2.1733333333333334e-05, "loss": 1.4832, "step": 164 }, { "epoch": 0.011179619215393996, "grad_norm": 16.75520896911621, "learning_rate": 2.186666666666667e-05, "loss": 1.1847, "step": 165 }, { "epoch": 0.011247374483366082, "grad_norm": 17.045665740966797, "learning_rate": 2.2000000000000003e-05, "loss": 1.6358, "step": 166 }, { "epoch": 0.011315129751338167, "grad_norm": 17.235095977783203, "learning_rate": 2.2133333333333334e-05, "loss": 1.6984, "step": 167 }, { "epoch": 0.011382885019310252, "grad_norm": 18.120485305786133, "learning_rate": 2.2266666666666668e-05, "loss": 1.5544, "step": 168 }, { "epoch": 0.011450640287282337, "grad_norm": 15.252264976501465, "learning_rate": 2.2400000000000002e-05, "loss": 1.4386, "step": 169 }, { "epoch": 0.011518395555254421, "grad_norm": 13.91727066040039, "learning_rate": 2.2533333333333333e-05, "loss": 1.2083, "step": 170 }, { "epoch": 0.011586150823226506, "grad_norm": 16.50945472717285, "learning_rate": 2.2666666666666668e-05, "loss": 1.5878, "step": 171 }, { "epoch": 0.01165390609119859, "grad_norm": 12.260165214538574, "learning_rate": 2.2800000000000002e-05, "loss": 1.4173, "step": 172 }, { "epoch": 0.011721661359170675, "grad_norm": 17.12419891357422, "learning_rate": 2.2933333333333333e-05, "loss": 1.3117, "step": 173 }, { "epoch": 0.01178941662714276, "grad_norm": 17.508407592773438, "learning_rate": 2.3066666666666667e-05, "loss": 1.4294, "step": 174 }, { "epoch": 0.011857171895114845, "grad_norm": 17.78769302368164, "learning_rate": 2.32e-05, "loss": 1.8437, "step": 175 }, { "epoch": 0.01192492716308693, "grad_norm": 13.79132080078125, "learning_rate": 2.3333333333333336e-05, "loss": 1.3969, "step": 176 }, { "epoch": 0.011992682431059014, "grad_norm": 16.519269943237305, "learning_rate": 2.3466666666666667e-05, "loss": 1.5833, "step": 177 }, { "epoch": 0.012060437699031099, "grad_norm": 15.920942306518555, "learning_rate": 2.36e-05, "loss": 1.4783, "step": 178 }, { "epoch": 0.012128192967003185, "grad_norm": 17.00782585144043, "learning_rate": 2.3733333333333335e-05, "loss": 1.4038, "step": 179 }, { "epoch": 0.01219594823497527, "grad_norm": 18.844655990600586, "learning_rate": 2.3866666666666666e-05, "loss": 1.653, "step": 180 }, { "epoch": 0.012263703502947355, "grad_norm": 18.05002784729004, "learning_rate": 2.4e-05, "loss": 1.4137, "step": 181 }, { "epoch": 0.01233145877091944, "grad_norm": 14.200697898864746, "learning_rate": 2.4133333333333335e-05, "loss": 1.3067, "step": 182 }, { "epoch": 0.012399214038891524, "grad_norm": 16.28152084350586, "learning_rate": 2.426666666666667e-05, "loss": 1.3425, "step": 183 }, { "epoch": 0.012466969306863609, "grad_norm": 18.691898345947266, "learning_rate": 2.44e-05, "loss": 1.5995, "step": 184 }, { "epoch": 0.012534724574835693, "grad_norm": 16.62773895263672, "learning_rate": 2.4533333333333334e-05, "loss": 1.5722, "step": 185 }, { "epoch": 0.012602479842807778, "grad_norm": 19.650226593017578, "learning_rate": 2.466666666666667e-05, "loss": 1.5445, "step": 186 }, { "epoch": 0.012670235110779863, "grad_norm": 14.582402229309082, "learning_rate": 2.48e-05, "loss": 1.4822, "step": 187 }, { "epoch": 0.012737990378751948, "grad_norm": 16.5295352935791, "learning_rate": 2.4933333333333334e-05, "loss": 1.5981, "step": 188 }, { "epoch": 0.012805745646724032, "grad_norm": 17.096031188964844, "learning_rate": 2.5066666666666665e-05, "loss": 1.2527, "step": 189 }, { "epoch": 0.012873500914696117, "grad_norm": 15.424555778503418, "learning_rate": 2.5200000000000003e-05, "loss": 1.5284, "step": 190 }, { "epoch": 0.012941256182668202, "grad_norm": 14.505122184753418, "learning_rate": 2.5333333333333337e-05, "loss": 1.5726, "step": 191 }, { "epoch": 0.013009011450640288, "grad_norm": 13.430007934570312, "learning_rate": 2.5466666666666668e-05, "loss": 1.5905, "step": 192 }, { "epoch": 0.013076766718612373, "grad_norm": 12.291808128356934, "learning_rate": 2.5600000000000002e-05, "loss": 1.3122, "step": 193 }, { "epoch": 0.013144521986584458, "grad_norm": 16.053747177124023, "learning_rate": 2.5733333333333337e-05, "loss": 1.3164, "step": 194 }, { "epoch": 0.013212277254556542, "grad_norm": 17.762910842895508, "learning_rate": 2.5866666666666667e-05, "loss": 1.4727, "step": 195 }, { "epoch": 0.013280032522528627, "grad_norm": 15.695332527160645, "learning_rate": 2.6000000000000002e-05, "loss": 1.5584, "step": 196 }, { "epoch": 0.013347787790500712, "grad_norm": 14.789410591125488, "learning_rate": 2.6133333333333333e-05, "loss": 1.3253, "step": 197 }, { "epoch": 0.013415543058472796, "grad_norm": 15.248231887817383, "learning_rate": 2.6266666666666667e-05, "loss": 1.3087, "step": 198 }, { "epoch": 0.013483298326444881, "grad_norm": 18.894878387451172, "learning_rate": 2.64e-05, "loss": 1.2755, "step": 199 }, { "epoch": 0.013551053594416966, "grad_norm": 16.611616134643555, "learning_rate": 2.6533333333333332e-05, "loss": 1.4218, "step": 200 }, { "epoch": 0.01361880886238905, "grad_norm": 17.056982040405273, "learning_rate": 2.6666666666666667e-05, "loss": 1.4376, "step": 201 }, { "epoch": 0.013686564130361135, "grad_norm": 15.91457748413086, "learning_rate": 2.6800000000000004e-05, "loss": 1.4836, "step": 202 }, { "epoch": 0.01375431939833322, "grad_norm": 16.371660232543945, "learning_rate": 2.6933333333333332e-05, "loss": 1.7374, "step": 203 }, { "epoch": 0.013822074666305304, "grad_norm": 14.45882511138916, "learning_rate": 2.706666666666667e-05, "loss": 1.3324, "step": 204 }, { "epoch": 0.013889829934277391, "grad_norm": 18.779190063476562, "learning_rate": 2.7200000000000004e-05, "loss": 1.4845, "step": 205 }, { "epoch": 0.013957585202249476, "grad_norm": 19.7078914642334, "learning_rate": 2.733333333333333e-05, "loss": 1.6681, "step": 206 }, { "epoch": 0.01402534047022156, "grad_norm": 20.498266220092773, "learning_rate": 2.746666666666667e-05, "loss": 1.559, "step": 207 }, { "epoch": 0.014093095738193645, "grad_norm": 20.2978515625, "learning_rate": 2.7600000000000003e-05, "loss": 1.5821, "step": 208 }, { "epoch": 0.01416085100616573, "grad_norm": 18.85460090637207, "learning_rate": 2.7733333333333334e-05, "loss": 1.3, "step": 209 }, { "epoch": 0.014228606274137814, "grad_norm": 13.912948608398438, "learning_rate": 2.786666666666667e-05, "loss": 1.4833, "step": 210 }, { "epoch": 0.014296361542109899, "grad_norm": 19.084806442260742, "learning_rate": 2.8000000000000003e-05, "loss": 1.3379, "step": 211 }, { "epoch": 0.014364116810081984, "grad_norm": 15.123217582702637, "learning_rate": 2.8133333333333334e-05, "loss": 1.354, "step": 212 }, { "epoch": 0.014431872078054069, "grad_norm": 14.127299308776855, "learning_rate": 2.8266666666666668e-05, "loss": 1.2341, "step": 213 }, { "epoch": 0.014499627346026153, "grad_norm": 15.609139442443848, "learning_rate": 2.84e-05, "loss": 1.3474, "step": 214 }, { "epoch": 0.014567382613998238, "grad_norm": 20.588394165039062, "learning_rate": 2.8533333333333333e-05, "loss": 1.4881, "step": 215 }, { "epoch": 0.014635137881970323, "grad_norm": 12.27170467376709, "learning_rate": 2.8666666666666668e-05, "loss": 1.2255, "step": 216 }, { "epoch": 0.014702893149942407, "grad_norm": 18.14453125, "learning_rate": 2.88e-05, "loss": 1.6328, "step": 217 }, { "epoch": 0.014770648417914492, "grad_norm": 17.722230911254883, "learning_rate": 2.8933333333333333e-05, "loss": 2.0172, "step": 218 }, { "epoch": 0.014838403685886578, "grad_norm": 14.454687118530273, "learning_rate": 2.906666666666667e-05, "loss": 1.2936, "step": 219 }, { "epoch": 0.014906158953858663, "grad_norm": 16.369625091552734, "learning_rate": 2.9199999999999998e-05, "loss": 1.6289, "step": 220 }, { "epoch": 0.014973914221830748, "grad_norm": 17.64832305908203, "learning_rate": 2.9333333333333336e-05, "loss": 1.3135, "step": 221 }, { "epoch": 0.015041669489802833, "grad_norm": 14.401778221130371, "learning_rate": 2.946666666666667e-05, "loss": 1.2614, "step": 222 }, { "epoch": 0.015109424757774917, "grad_norm": 17.145814895629883, "learning_rate": 2.96e-05, "loss": 1.3473, "step": 223 }, { "epoch": 0.015177180025747002, "grad_norm": 17.273954391479492, "learning_rate": 2.9733333333333336e-05, "loss": 1.3747, "step": 224 }, { "epoch": 0.015244935293719087, "grad_norm": 17.190818786621094, "learning_rate": 2.986666666666667e-05, "loss": 1.5452, "step": 225 }, { "epoch": 0.015312690561691171, "grad_norm": 19.573585510253906, "learning_rate": 3e-05, "loss": 1.4942, "step": 226 }, { "epoch": 0.015380445829663256, "grad_norm": 15.349540710449219, "learning_rate": 3.0133333333333335e-05, "loss": 1.4277, "step": 227 }, { "epoch": 0.01544820109763534, "grad_norm": 19.786211013793945, "learning_rate": 3.0266666666666666e-05, "loss": 1.5014, "step": 228 }, { "epoch": 0.015515956365607425, "grad_norm": 17.51229476928711, "learning_rate": 3.04e-05, "loss": 1.695, "step": 229 }, { "epoch": 0.01558371163357951, "grad_norm": 19.894756317138672, "learning_rate": 3.0533333333333335e-05, "loss": 1.4994, "step": 230 }, { "epoch": 0.015651466901551597, "grad_norm": 13.575235366821289, "learning_rate": 3.066666666666667e-05, "loss": 1.4037, "step": 231 }, { "epoch": 0.01571922216952368, "grad_norm": 18.516822814941406, "learning_rate": 3.08e-05, "loss": 1.5494, "step": 232 }, { "epoch": 0.015786977437495766, "grad_norm": 17.343368530273438, "learning_rate": 3.093333333333334e-05, "loss": 1.1708, "step": 233 }, { "epoch": 0.01585473270546785, "grad_norm": 18.78635597229004, "learning_rate": 3.1066666666666665e-05, "loss": 1.6543, "step": 234 }, { "epoch": 0.015922487973439935, "grad_norm": 16.219989776611328, "learning_rate": 3.12e-05, "loss": 1.381, "step": 235 }, { "epoch": 0.01599024324141202, "grad_norm": 15.987724304199219, "learning_rate": 3.1333333333333334e-05, "loss": 1.3981, "step": 236 }, { "epoch": 0.016057998509384105, "grad_norm": 17.251537322998047, "learning_rate": 3.146666666666667e-05, "loss": 1.8181, "step": 237 }, { "epoch": 0.01612575377735619, "grad_norm": 15.804817199707031, "learning_rate": 3.16e-05, "loss": 1.4755, "step": 238 }, { "epoch": 0.016193509045328274, "grad_norm": 16.508943557739258, "learning_rate": 3.173333333333334e-05, "loss": 1.2167, "step": 239 }, { "epoch": 0.01626126431330036, "grad_norm": 17.382783889770508, "learning_rate": 3.1866666666666664e-05, "loss": 1.6186, "step": 240 }, { "epoch": 0.016329019581272444, "grad_norm": 17.938161849975586, "learning_rate": 3.2000000000000005e-05, "loss": 1.3685, "step": 241 }, { "epoch": 0.016396774849244528, "grad_norm": 21.103981018066406, "learning_rate": 3.213333333333334e-05, "loss": 1.6275, "step": 242 }, { "epoch": 0.016464530117216613, "grad_norm": 15.145224571228027, "learning_rate": 3.226666666666667e-05, "loss": 1.2058, "step": 243 }, { "epoch": 0.016532285385188698, "grad_norm": 15.368849754333496, "learning_rate": 3.24e-05, "loss": 1.3436, "step": 244 }, { "epoch": 0.016600040653160782, "grad_norm": 16.95122718811035, "learning_rate": 3.253333333333333e-05, "loss": 1.4346, "step": 245 }, { "epoch": 0.016667795921132867, "grad_norm": 14.612030029296875, "learning_rate": 3.266666666666667e-05, "loss": 1.4759, "step": 246 }, { "epoch": 0.016735551189104952, "grad_norm": 17.6989688873291, "learning_rate": 3.2800000000000004e-05, "loss": 1.3612, "step": 247 }, { "epoch": 0.016803306457077036, "grad_norm": 19.985427856445312, "learning_rate": 3.293333333333333e-05, "loss": 1.2718, "step": 248 }, { "epoch": 0.01687106172504912, "grad_norm": 12.162117958068848, "learning_rate": 3.3066666666666666e-05, "loss": 1.2133, "step": 249 }, { "epoch": 0.016938816993021206, "grad_norm": 19.053850173950195, "learning_rate": 3.32e-05, "loss": 1.1345, "step": 250 }, { "epoch": 0.01700657226099329, "grad_norm": 15.729182243347168, "learning_rate": 3.3333333333333335e-05, "loss": 1.3079, "step": 251 }, { "epoch": 0.01707432752896538, "grad_norm": 18.984880447387695, "learning_rate": 3.346666666666667e-05, "loss": 1.5124, "step": 252 }, { "epoch": 0.017142082796937463, "grad_norm": 14.444266319274902, "learning_rate": 3.3600000000000004e-05, "loss": 1.4279, "step": 253 }, { "epoch": 0.017209838064909548, "grad_norm": 17.174089431762695, "learning_rate": 3.373333333333333e-05, "loss": 1.1192, "step": 254 }, { "epoch": 0.017277593332881633, "grad_norm": 15.885781288146973, "learning_rate": 3.3866666666666665e-05, "loss": 1.3918, "step": 255 }, { "epoch": 0.017345348600853717, "grad_norm": 14.934944152832031, "learning_rate": 3.4000000000000007e-05, "loss": 1.3348, "step": 256 }, { "epoch": 0.017413103868825802, "grad_norm": 17.977643966674805, "learning_rate": 3.4133333333333334e-05, "loss": 1.392, "step": 257 }, { "epoch": 0.017480859136797887, "grad_norm": 17.527130126953125, "learning_rate": 3.426666666666667e-05, "loss": 1.2939, "step": 258 }, { "epoch": 0.01754861440476997, "grad_norm": 17.823453903198242, "learning_rate": 3.4399999999999996e-05, "loss": 1.3118, "step": 259 }, { "epoch": 0.017616369672742056, "grad_norm": 15.746111869812012, "learning_rate": 3.453333333333334e-05, "loss": 1.3865, "step": 260 }, { "epoch": 0.01768412494071414, "grad_norm": 18.798080444335938, "learning_rate": 3.466666666666667e-05, "loss": 1.399, "step": 261 }, { "epoch": 0.017751880208686226, "grad_norm": 15.139904975891113, "learning_rate": 3.48e-05, "loss": 1.5248, "step": 262 }, { "epoch": 0.01781963547665831, "grad_norm": 16.150434494018555, "learning_rate": 3.493333333333333e-05, "loss": 1.5033, "step": 263 }, { "epoch": 0.017887390744630395, "grad_norm": 19.125490188598633, "learning_rate": 3.506666666666667e-05, "loss": 1.4695, "step": 264 }, { "epoch": 0.01795514601260248, "grad_norm": 15.708895683288574, "learning_rate": 3.52e-05, "loss": 1.6189, "step": 265 }, { "epoch": 0.018022901280574564, "grad_norm": 14.185934066772461, "learning_rate": 3.5333333333333336e-05, "loss": 1.3536, "step": 266 }, { "epoch": 0.01809065654854665, "grad_norm": 15.22371768951416, "learning_rate": 3.546666666666667e-05, "loss": 1.4652, "step": 267 }, { "epoch": 0.018158411816518734, "grad_norm": 15.854277610778809, "learning_rate": 3.56e-05, "loss": 1.7211, "step": 268 }, { "epoch": 0.01822616708449082, "grad_norm": 18.41203498840332, "learning_rate": 3.573333333333333e-05, "loss": 1.54, "step": 269 }, { "epoch": 0.018293922352462903, "grad_norm": 15.323198318481445, "learning_rate": 3.586666666666667e-05, "loss": 1.3582, "step": 270 }, { "epoch": 0.018361677620434988, "grad_norm": 15.056265830993652, "learning_rate": 3.6e-05, "loss": 1.2959, "step": 271 }, { "epoch": 0.018429432888407073, "grad_norm": 18.912555694580078, "learning_rate": 3.6133333333333335e-05, "loss": 1.5964, "step": 272 }, { "epoch": 0.018497188156379157, "grad_norm": 15.093083381652832, "learning_rate": 3.626666666666667e-05, "loss": 1.3403, "step": 273 }, { "epoch": 0.018564943424351242, "grad_norm": 17.38453483581543, "learning_rate": 3.6400000000000004e-05, "loss": 1.5082, "step": 274 }, { "epoch": 0.018632698692323327, "grad_norm": 16.53204345703125, "learning_rate": 3.653333333333334e-05, "loss": 1.2373, "step": 275 }, { "epoch": 0.01870045396029541, "grad_norm": 13.423516273498535, "learning_rate": 3.6666666666666666e-05, "loss": 1.3111, "step": 276 }, { "epoch": 0.018768209228267496, "grad_norm": 14.92001724243164, "learning_rate": 3.68e-05, "loss": 1.5767, "step": 277 }, { "epoch": 0.018835964496239584, "grad_norm": 15.229578971862793, "learning_rate": 3.6933333333333334e-05, "loss": 1.19, "step": 278 }, { "epoch": 0.01890371976421167, "grad_norm": 17.159273147583008, "learning_rate": 3.706666666666667e-05, "loss": 1.335, "step": 279 }, { "epoch": 0.018971475032183754, "grad_norm": 13.7230806350708, "learning_rate": 3.72e-05, "loss": 1.2587, "step": 280 }, { "epoch": 0.01903923030015584, "grad_norm": 15.375404357910156, "learning_rate": 3.733333333333334e-05, "loss": 1.2744, "step": 281 }, { "epoch": 0.019106985568127923, "grad_norm": 21.054534912109375, "learning_rate": 3.7466666666666665e-05, "loss": 1.5581, "step": 282 }, { "epoch": 0.019174740836100008, "grad_norm": 17.954307556152344, "learning_rate": 3.76e-05, "loss": 1.6617, "step": 283 }, { "epoch": 0.019242496104072093, "grad_norm": 17.954139709472656, "learning_rate": 3.773333333333334e-05, "loss": 1.7486, "step": 284 }, { "epoch": 0.019310251372044177, "grad_norm": 16.210704803466797, "learning_rate": 3.786666666666667e-05, "loss": 1.661, "step": 285 }, { "epoch": 0.019378006640016262, "grad_norm": 18.14916229248047, "learning_rate": 3.8e-05, "loss": 1.8154, "step": 286 }, { "epoch": 0.019445761907988347, "grad_norm": 16.372896194458008, "learning_rate": 3.8133333333333336e-05, "loss": 1.5464, "step": 287 }, { "epoch": 0.01951351717596043, "grad_norm": 16.4403076171875, "learning_rate": 3.8266666666666664e-05, "loss": 1.0815, "step": 288 }, { "epoch": 0.019581272443932516, "grad_norm": 16.612815856933594, "learning_rate": 3.8400000000000005e-05, "loss": 1.4338, "step": 289 }, { "epoch": 0.0196490277119046, "grad_norm": 17.0092716217041, "learning_rate": 3.853333333333334e-05, "loss": 1.5257, "step": 290 }, { "epoch": 0.019716782979876685, "grad_norm": 13.920248031616211, "learning_rate": 3.866666666666667e-05, "loss": 1.4595, "step": 291 }, { "epoch": 0.01978453824784877, "grad_norm": 12.699677467346191, "learning_rate": 3.88e-05, "loss": 1.2602, "step": 292 }, { "epoch": 0.019852293515820855, "grad_norm": 17.31343650817871, "learning_rate": 3.8933333333333336e-05, "loss": 1.2934, "step": 293 }, { "epoch": 0.01992004878379294, "grad_norm": 16.7819881439209, "learning_rate": 3.906666666666667e-05, "loss": 1.5385, "step": 294 }, { "epoch": 0.019987804051765024, "grad_norm": 14.175074577331543, "learning_rate": 3.9200000000000004e-05, "loss": 1.1644, "step": 295 }, { "epoch": 0.02005555931973711, "grad_norm": 22.863780975341797, "learning_rate": 3.933333333333333e-05, "loss": 1.5932, "step": 296 }, { "epoch": 0.020123314587709194, "grad_norm": 15.278692245483398, "learning_rate": 3.9466666666666666e-05, "loss": 1.3997, "step": 297 }, { "epoch": 0.02019106985568128, "grad_norm": 15.878965377807617, "learning_rate": 3.960000000000001e-05, "loss": 1.5992, "step": 298 }, { "epoch": 0.020258825123653363, "grad_norm": 17.444494247436523, "learning_rate": 3.9733333333333335e-05, "loss": 1.5224, "step": 299 }, { "epoch": 0.020326580391625448, "grad_norm": 15.011860847473145, "learning_rate": 3.986666666666667e-05, "loss": 1.2609, "step": 300 }, { "epoch": 0.020394335659597532, "grad_norm": 17.057064056396484, "learning_rate": 4e-05, "loss": 1.2724, "step": 301 }, { "epoch": 0.020462090927569617, "grad_norm": 14.946621894836426, "learning_rate": 4.013333333333333e-05, "loss": 1.273, "step": 302 }, { "epoch": 0.020529846195541702, "grad_norm": 15.072589874267578, "learning_rate": 4.026666666666667e-05, "loss": 1.2029, "step": 303 }, { "epoch": 0.020597601463513786, "grad_norm": 14.304384231567383, "learning_rate": 4.0400000000000006e-05, "loss": 1.3601, "step": 304 }, { "epoch": 0.020665356731485875, "grad_norm": 16.18627166748047, "learning_rate": 4.0533333333333334e-05, "loss": 1.5844, "step": 305 }, { "epoch": 0.02073311199945796, "grad_norm": 14.144999504089355, "learning_rate": 4.066666666666667e-05, "loss": 1.1638, "step": 306 }, { "epoch": 0.020800867267430044, "grad_norm": 13.793155670166016, "learning_rate": 4.08e-05, "loss": 1.2101, "step": 307 }, { "epoch": 0.02086862253540213, "grad_norm": 19.344085693359375, "learning_rate": 4.093333333333334e-05, "loss": 1.5899, "step": 308 }, { "epoch": 0.020936377803374213, "grad_norm": 13.670002937316895, "learning_rate": 4.106666666666667e-05, "loss": 1.4271, "step": 309 }, { "epoch": 0.021004133071346298, "grad_norm": 17.72041893005371, "learning_rate": 4.12e-05, "loss": 1.5276, "step": 310 }, { "epoch": 0.021071888339318383, "grad_norm": 14.103938102722168, "learning_rate": 4.133333333333333e-05, "loss": 1.2355, "step": 311 }, { "epoch": 0.021139643607290468, "grad_norm": 16.439529418945312, "learning_rate": 4.146666666666667e-05, "loss": 1.3393, "step": 312 }, { "epoch": 0.021207398875262552, "grad_norm": 13.821382522583008, "learning_rate": 4.16e-05, "loss": 1.275, "step": 313 }, { "epoch": 0.021275154143234637, "grad_norm": 19.987564086914062, "learning_rate": 4.1733333333333336e-05, "loss": 1.415, "step": 314 }, { "epoch": 0.02134290941120672, "grad_norm": 12.660404205322266, "learning_rate": 4.186666666666667e-05, "loss": 1.2378, "step": 315 }, { "epoch": 0.021410664679178806, "grad_norm": 14.009933471679688, "learning_rate": 4.2e-05, "loss": 1.1615, "step": 316 }, { "epoch": 0.02147841994715089, "grad_norm": 16.11932945251465, "learning_rate": 4.213333333333334e-05, "loss": 1.0653, "step": 317 }, { "epoch": 0.021546175215122976, "grad_norm": 15.943892478942871, "learning_rate": 4.226666666666667e-05, "loss": 1.6208, "step": 318 }, { "epoch": 0.02161393048309506, "grad_norm": 16.01548194885254, "learning_rate": 4.24e-05, "loss": 1.4052, "step": 319 }, { "epoch": 0.021681685751067145, "grad_norm": 15.703156471252441, "learning_rate": 4.2533333333333335e-05, "loss": 1.3736, "step": 320 }, { "epoch": 0.02174944101903923, "grad_norm": 20.367021560668945, "learning_rate": 4.266666666666667e-05, "loss": 1.5758, "step": 321 }, { "epoch": 0.021817196287011315, "grad_norm": 15.678637504577637, "learning_rate": 4.2800000000000004e-05, "loss": 1.4458, "step": 322 }, { "epoch": 0.0218849515549834, "grad_norm": 15.707669258117676, "learning_rate": 4.293333333333334e-05, "loss": 1.3065, "step": 323 }, { "epoch": 0.021952706822955484, "grad_norm": 16.07583999633789, "learning_rate": 4.3066666666666665e-05, "loss": 1.5875, "step": 324 }, { "epoch": 0.02202046209092757, "grad_norm": 15.248065948486328, "learning_rate": 4.32e-05, "loss": 1.5972, "step": 325 }, { "epoch": 0.022088217358899653, "grad_norm": 17.77379608154297, "learning_rate": 4.3333333333333334e-05, "loss": 1.3385, "step": 326 }, { "epoch": 0.022155972626871738, "grad_norm": 24.381973266601562, "learning_rate": 4.346666666666667e-05, "loss": 1.6428, "step": 327 }, { "epoch": 0.022223727894843823, "grad_norm": 17.3684024810791, "learning_rate": 4.36e-05, "loss": 1.3634, "step": 328 }, { "epoch": 0.022291483162815907, "grad_norm": 20.45623207092285, "learning_rate": 4.373333333333334e-05, "loss": 1.3947, "step": 329 }, { "epoch": 0.022359238430787992, "grad_norm": 17.434114456176758, "learning_rate": 4.3866666666666665e-05, "loss": 1.4269, "step": 330 }, { "epoch": 0.02242699369876008, "grad_norm": 13.574700355529785, "learning_rate": 4.4000000000000006e-05, "loss": 1.2807, "step": 331 }, { "epoch": 0.022494748966732165, "grad_norm": 16.404966354370117, "learning_rate": 4.413333333333334e-05, "loss": 1.4549, "step": 332 }, { "epoch": 0.02256250423470425, "grad_norm": 14.14947509765625, "learning_rate": 4.426666666666667e-05, "loss": 1.4187, "step": 333 }, { "epoch": 0.022630259502676334, "grad_norm": 17.632492065429688, "learning_rate": 4.44e-05, "loss": 1.3515, "step": 334 }, { "epoch": 0.02269801477064842, "grad_norm": 18.149261474609375, "learning_rate": 4.4533333333333336e-05, "loss": 1.4875, "step": 335 }, { "epoch": 0.022765770038620504, "grad_norm": 19.864049911499023, "learning_rate": 4.466666666666667e-05, "loss": 1.5622, "step": 336 }, { "epoch": 0.02283352530659259, "grad_norm": 16.7431583404541, "learning_rate": 4.4800000000000005e-05, "loss": 1.4344, "step": 337 }, { "epoch": 0.022901280574564673, "grad_norm": 16.53998565673828, "learning_rate": 4.493333333333333e-05, "loss": 1.1745, "step": 338 }, { "epoch": 0.022969035842536758, "grad_norm": 17.786945343017578, "learning_rate": 4.5066666666666667e-05, "loss": 1.4335, "step": 339 }, { "epoch": 0.023036791110508843, "grad_norm": 17.730606079101562, "learning_rate": 4.52e-05, "loss": 1.341, "step": 340 }, { "epoch": 0.023104546378480927, "grad_norm": 15.536660194396973, "learning_rate": 4.5333333333333335e-05, "loss": 1.4935, "step": 341 }, { "epoch": 0.023172301646453012, "grad_norm": 22.405893325805664, "learning_rate": 4.546666666666667e-05, "loss": 1.5538, "step": 342 }, { "epoch": 0.023240056914425097, "grad_norm": 15.963828086853027, "learning_rate": 4.5600000000000004e-05, "loss": 1.373, "step": 343 }, { "epoch": 0.02330781218239718, "grad_norm": 13.041223526000977, "learning_rate": 4.573333333333333e-05, "loss": 1.346, "step": 344 }, { "epoch": 0.023375567450369266, "grad_norm": 14.552989959716797, "learning_rate": 4.5866666666666666e-05, "loss": 1.2723, "step": 345 }, { "epoch": 0.02344332271834135, "grad_norm": 16.034404754638672, "learning_rate": 4.600000000000001e-05, "loss": 1.3671, "step": 346 }, { "epoch": 0.023511077986313435, "grad_norm": 17.5505428314209, "learning_rate": 4.6133333333333334e-05, "loss": 1.4553, "step": 347 }, { "epoch": 0.02357883325428552, "grad_norm": 14.029112815856934, "learning_rate": 4.626666666666667e-05, "loss": 1.3639, "step": 348 }, { "epoch": 0.023646588522257605, "grad_norm": 14.633806228637695, "learning_rate": 4.64e-05, "loss": 1.61, "step": 349 }, { "epoch": 0.02371434379022969, "grad_norm": 15.407472610473633, "learning_rate": 4.653333333333334e-05, "loss": 1.2449, "step": 350 }, { "epoch": 0.023782099058201774, "grad_norm": 13.801481246948242, "learning_rate": 4.666666666666667e-05, "loss": 1.3105, "step": 351 }, { "epoch": 0.02384985432617386, "grad_norm": 17.675159454345703, "learning_rate": 4.6800000000000006e-05, "loss": 1.3142, "step": 352 }, { "epoch": 0.023917609594145944, "grad_norm": 15.383625984191895, "learning_rate": 4.6933333333333333e-05, "loss": 1.2976, "step": 353 }, { "epoch": 0.02398536486211803, "grad_norm": 18.087081909179688, "learning_rate": 4.706666666666667e-05, "loss": 1.3391, "step": 354 }, { "epoch": 0.024053120130090113, "grad_norm": 15.790519714355469, "learning_rate": 4.72e-05, "loss": 1.7103, "step": 355 }, { "epoch": 0.024120875398062198, "grad_norm": 15.931246757507324, "learning_rate": 4.7333333333333336e-05, "loss": 1.4792, "step": 356 }, { "epoch": 0.024188630666034282, "grad_norm": 18.39167022705078, "learning_rate": 4.746666666666667e-05, "loss": 1.5085, "step": 357 }, { "epoch": 0.02425638593400637, "grad_norm": 13.518149375915527, "learning_rate": 4.76e-05, "loss": 1.2711, "step": 358 }, { "epoch": 0.024324141201978455, "grad_norm": 18.02762794494629, "learning_rate": 4.773333333333333e-05, "loss": 1.4113, "step": 359 }, { "epoch": 0.02439189646995054, "grad_norm": 15.260608673095703, "learning_rate": 4.7866666666666674e-05, "loss": 1.3286, "step": 360 }, { "epoch": 0.024459651737922625, "grad_norm": 16.26275062561035, "learning_rate": 4.8e-05, "loss": 1.4436, "step": 361 }, { "epoch": 0.02452740700589471, "grad_norm": 14.213051795959473, "learning_rate": 4.8133333333333336e-05, "loss": 1.1339, "step": 362 }, { "epoch": 0.024595162273866794, "grad_norm": 16.478811264038086, "learning_rate": 4.826666666666667e-05, "loss": 1.5534, "step": 363 }, { "epoch": 0.02466291754183888, "grad_norm": 13.24108600616455, "learning_rate": 4.8400000000000004e-05, "loss": 1.4767, "step": 364 }, { "epoch": 0.024730672809810963, "grad_norm": 15.167949676513672, "learning_rate": 4.853333333333334e-05, "loss": 1.1592, "step": 365 }, { "epoch": 0.024798428077783048, "grad_norm": 18.150367736816406, "learning_rate": 4.866666666666667e-05, "loss": 1.4172, "step": 366 }, { "epoch": 0.024866183345755133, "grad_norm": 13.250176429748535, "learning_rate": 4.88e-05, "loss": 1.5022, "step": 367 }, { "epoch": 0.024933938613727218, "grad_norm": 14.780989646911621, "learning_rate": 4.8933333333333335e-05, "loss": 1.3243, "step": 368 }, { "epoch": 0.025001693881699302, "grad_norm": 13.34748649597168, "learning_rate": 4.906666666666667e-05, "loss": 1.285, "step": 369 }, { "epoch": 0.025069449149671387, "grad_norm": 14.677448272705078, "learning_rate": 4.92e-05, "loss": 1.1982, "step": 370 }, { "epoch": 0.02513720441764347, "grad_norm": 18.211145401000977, "learning_rate": 4.933333333333334e-05, "loss": 1.4018, "step": 371 }, { "epoch": 0.025204959685615556, "grad_norm": 17.68402099609375, "learning_rate": 4.9466666666666665e-05, "loss": 1.6153, "step": 372 }, { "epoch": 0.02527271495358764, "grad_norm": 14.44299030303955, "learning_rate": 4.96e-05, "loss": 1.4092, "step": 373 }, { "epoch": 0.025340470221559726, "grad_norm": 15.467754364013672, "learning_rate": 4.973333333333334e-05, "loss": 1.5647, "step": 374 }, { "epoch": 0.02540822548953181, "grad_norm": 14.037840843200684, "learning_rate": 4.986666666666667e-05, "loss": 1.2106, "step": 375 }, { "epoch": 0.025475980757503895, "grad_norm": 15.72461223602295, "learning_rate": 5e-05, "loss": 1.1832, "step": 376 }, { "epoch": 0.02554373602547598, "grad_norm": 15.03637981414795, "learning_rate": 5.013333333333333e-05, "loss": 1.5803, "step": 377 }, { "epoch": 0.025611491293448065, "grad_norm": 15.995429992675781, "learning_rate": 5.026666666666667e-05, "loss": 1.3837, "step": 378 }, { "epoch": 0.02567924656142015, "grad_norm": 17.96516227722168, "learning_rate": 5.0400000000000005e-05, "loss": 1.5114, "step": 379 }, { "epoch": 0.025747001829392234, "grad_norm": 17.960783004760742, "learning_rate": 5.053333333333333e-05, "loss": 1.4012, "step": 380 }, { "epoch": 0.02581475709736432, "grad_norm": 14.399024963378906, "learning_rate": 5.0666666666666674e-05, "loss": 1.2906, "step": 381 }, { "epoch": 0.025882512365336403, "grad_norm": 14.268051147460938, "learning_rate": 5.08e-05, "loss": 1.2354, "step": 382 }, { "epoch": 0.025950267633308488, "grad_norm": 16.070646286010742, "learning_rate": 5.0933333333333336e-05, "loss": 1.272, "step": 383 }, { "epoch": 0.026018022901280576, "grad_norm": 18.698619842529297, "learning_rate": 5.106666666666668e-05, "loss": 1.352, "step": 384 }, { "epoch": 0.02608577816925266, "grad_norm": 13.932543754577637, "learning_rate": 5.1200000000000004e-05, "loss": 1.3249, "step": 385 }, { "epoch": 0.026153533437224746, "grad_norm": 14.76308822631836, "learning_rate": 5.133333333333333e-05, "loss": 1.1807, "step": 386 }, { "epoch": 0.02622128870519683, "grad_norm": 17.602182388305664, "learning_rate": 5.146666666666667e-05, "loss": 1.4017, "step": 387 }, { "epoch": 0.026289043973168915, "grad_norm": 13.589346885681152, "learning_rate": 5.16e-05, "loss": 1.4205, "step": 388 }, { "epoch": 0.026356799241141, "grad_norm": 16.37788200378418, "learning_rate": 5.1733333333333335e-05, "loss": 1.4255, "step": 389 }, { "epoch": 0.026424554509113084, "grad_norm": 16.24977684020996, "learning_rate": 5.1866666666666676e-05, "loss": 1.2526, "step": 390 }, { "epoch": 0.02649230977708517, "grad_norm": 15.02576732635498, "learning_rate": 5.2000000000000004e-05, "loss": 1.2147, "step": 391 }, { "epoch": 0.026560065045057254, "grad_norm": 16.754850387573242, "learning_rate": 5.213333333333333e-05, "loss": 1.5725, "step": 392 }, { "epoch": 0.02662782031302934, "grad_norm": 17.120729446411133, "learning_rate": 5.2266666666666665e-05, "loss": 1.351, "step": 393 }, { "epoch": 0.026695575581001423, "grad_norm": 18.83492660522461, "learning_rate": 5.2400000000000007e-05, "loss": 1.3044, "step": 394 }, { "epoch": 0.026763330848973508, "grad_norm": 18.66061019897461, "learning_rate": 5.2533333333333334e-05, "loss": 1.4889, "step": 395 }, { "epoch": 0.026831086116945593, "grad_norm": 18.493236541748047, "learning_rate": 5.266666666666666e-05, "loss": 1.5015, "step": 396 }, { "epoch": 0.026898841384917677, "grad_norm": 19.523067474365234, "learning_rate": 5.28e-05, "loss": 1.5179, "step": 397 }, { "epoch": 0.026966596652889762, "grad_norm": 18.207366943359375, "learning_rate": 5.293333333333334e-05, "loss": 1.374, "step": 398 }, { "epoch": 0.027034351920861847, "grad_norm": 13.021437644958496, "learning_rate": 5.3066666666666665e-05, "loss": 1.0259, "step": 399 }, { "epoch": 0.02710210718883393, "grad_norm": 14.108929634094238, "learning_rate": 5.3200000000000006e-05, "loss": 1.3914, "step": 400 }, { "epoch": 0.027169862456806016, "grad_norm": 15.087890625, "learning_rate": 5.333333333333333e-05, "loss": 1.3143, "step": 401 }, { "epoch": 0.0272376177247781, "grad_norm": 16.389965057373047, "learning_rate": 5.346666666666667e-05, "loss": 1.5286, "step": 402 }, { "epoch": 0.027305372992750186, "grad_norm": 17.855867385864258, "learning_rate": 5.360000000000001e-05, "loss": 1.4221, "step": 403 }, { "epoch": 0.02737312826072227, "grad_norm": 14.016682624816895, "learning_rate": 5.3733333333333336e-05, "loss": 1.0922, "step": 404 }, { "epoch": 0.027440883528694355, "grad_norm": 16.08883285522461, "learning_rate": 5.3866666666666664e-05, "loss": 1.4932, "step": 405 }, { "epoch": 0.02750863879666644, "grad_norm": 14.883580207824707, "learning_rate": 5.4000000000000005e-05, "loss": 1.0806, "step": 406 }, { "epoch": 0.027576394064638524, "grad_norm": 13.443008422851562, "learning_rate": 5.413333333333334e-05, "loss": 1.2653, "step": 407 }, { "epoch": 0.02764414933261061, "grad_norm": 13.391244888305664, "learning_rate": 5.4266666666666667e-05, "loss": 1.2788, "step": 408 }, { "epoch": 0.027711904600582694, "grad_norm": 13.449023246765137, "learning_rate": 5.440000000000001e-05, "loss": 1.133, "step": 409 }, { "epoch": 0.027779659868554782, "grad_norm": 13.777191162109375, "learning_rate": 5.4533333333333335e-05, "loss": 1.1509, "step": 410 }, { "epoch": 0.027847415136526867, "grad_norm": 19.04452896118164, "learning_rate": 5.466666666666666e-05, "loss": 1.3306, "step": 411 }, { "epoch": 0.02791517040449895, "grad_norm": 16.583112716674805, "learning_rate": 5.4800000000000004e-05, "loss": 1.3018, "step": 412 }, { "epoch": 0.027982925672471036, "grad_norm": 16.8561954498291, "learning_rate": 5.493333333333334e-05, "loss": 1.257, "step": 413 }, { "epoch": 0.02805068094044312, "grad_norm": 15.321064949035645, "learning_rate": 5.5066666666666666e-05, "loss": 1.2498, "step": 414 }, { "epoch": 0.028118436208415205, "grad_norm": 15.825779914855957, "learning_rate": 5.520000000000001e-05, "loss": 1.2177, "step": 415 }, { "epoch": 0.02818619147638729, "grad_norm": 15.665361404418945, "learning_rate": 5.5333333333333334e-05, "loss": 1.244, "step": 416 }, { "epoch": 0.028253946744359375, "grad_norm": 15.516046524047852, "learning_rate": 5.546666666666667e-05, "loss": 1.4957, "step": 417 }, { "epoch": 0.02832170201233146, "grad_norm": 19.148746490478516, "learning_rate": 5.560000000000001e-05, "loss": 1.4346, "step": 418 }, { "epoch": 0.028389457280303544, "grad_norm": 15.215538024902344, "learning_rate": 5.573333333333334e-05, "loss": 1.081, "step": 419 }, { "epoch": 0.02845721254827563, "grad_norm": 15.612996101379395, "learning_rate": 5.5866666666666665e-05, "loss": 1.1671, "step": 420 }, { "epoch": 0.028524967816247714, "grad_norm": 13.754039764404297, "learning_rate": 5.6000000000000006e-05, "loss": 1.2679, "step": 421 }, { "epoch": 0.028592723084219798, "grad_norm": 16.350305557250977, "learning_rate": 5.613333333333334e-05, "loss": 1.3034, "step": 422 }, { "epoch": 0.028660478352191883, "grad_norm": 16.560344696044922, "learning_rate": 5.626666666666667e-05, "loss": 1.3633, "step": 423 }, { "epoch": 0.028728233620163968, "grad_norm": 20.391889572143555, "learning_rate": 5.6399999999999995e-05, "loss": 1.3533, "step": 424 }, { "epoch": 0.028795988888136052, "grad_norm": 16.13326072692871, "learning_rate": 5.6533333333333336e-05, "loss": 1.3678, "step": 425 }, { "epoch": 0.028863744156108137, "grad_norm": 21.544612884521484, "learning_rate": 5.666666666666667e-05, "loss": 1.6275, "step": 426 }, { "epoch": 0.028931499424080222, "grad_norm": 17.097408294677734, "learning_rate": 5.68e-05, "loss": 1.6119, "step": 427 }, { "epoch": 0.028999254692052306, "grad_norm": 15.699347496032715, "learning_rate": 5.693333333333334e-05, "loss": 1.2644, "step": 428 }, { "epoch": 0.02906700996002439, "grad_norm": 16.643613815307617, "learning_rate": 5.706666666666667e-05, "loss": 1.3497, "step": 429 }, { "epoch": 0.029134765227996476, "grad_norm": 20.941123962402344, "learning_rate": 5.72e-05, "loss": 1.4232, "step": 430 }, { "epoch": 0.02920252049596856, "grad_norm": 17.549013137817383, "learning_rate": 5.7333333333333336e-05, "loss": 1.3429, "step": 431 }, { "epoch": 0.029270275763940645, "grad_norm": 19.702617645263672, "learning_rate": 5.746666666666667e-05, "loss": 1.7924, "step": 432 }, { "epoch": 0.02933803103191273, "grad_norm": 16.396209716796875, "learning_rate": 5.76e-05, "loss": 1.3296, "step": 433 }, { "epoch": 0.029405786299884815, "grad_norm": 15.823278427124023, "learning_rate": 5.773333333333334e-05, "loss": 1.4228, "step": 434 }, { "epoch": 0.0294735415678569, "grad_norm": 19.61952018737793, "learning_rate": 5.7866666666666666e-05, "loss": 1.3522, "step": 435 }, { "epoch": 0.029541296835828984, "grad_norm": 14.721433639526367, "learning_rate": 5.8e-05, "loss": 1.284, "step": 436 }, { "epoch": 0.029609052103801072, "grad_norm": 16.594276428222656, "learning_rate": 5.813333333333334e-05, "loss": 1.4636, "step": 437 }, { "epoch": 0.029676807371773157, "grad_norm": 15.98005199432373, "learning_rate": 5.826666666666667e-05, "loss": 1.2825, "step": 438 }, { "epoch": 0.02974456263974524, "grad_norm": 15.501729965209961, "learning_rate": 5.8399999999999997e-05, "loss": 1.3736, "step": 439 }, { "epoch": 0.029812317907717326, "grad_norm": 18.077552795410156, "learning_rate": 5.853333333333334e-05, "loss": 1.4115, "step": 440 }, { "epoch": 0.02988007317568941, "grad_norm": 16.016721725463867, "learning_rate": 5.866666666666667e-05, "loss": 1.3177, "step": 441 }, { "epoch": 0.029947828443661496, "grad_norm": 16.39783477783203, "learning_rate": 5.88e-05, "loss": 1.4465, "step": 442 }, { "epoch": 0.03001558371163358, "grad_norm": 13.970220565795898, "learning_rate": 5.893333333333334e-05, "loss": 1.2956, "step": 443 }, { "epoch": 0.030083338979605665, "grad_norm": 16.5622615814209, "learning_rate": 5.906666666666667e-05, "loss": 1.2144, "step": 444 }, { "epoch": 0.03015109424757775, "grad_norm": 18.191911697387695, "learning_rate": 5.92e-05, "loss": 1.623, "step": 445 }, { "epoch": 0.030218849515549834, "grad_norm": 15.483613967895508, "learning_rate": 5.9333333333333343e-05, "loss": 1.3232, "step": 446 }, { "epoch": 0.03028660478352192, "grad_norm": 15.633200645446777, "learning_rate": 5.946666666666667e-05, "loss": 1.544, "step": 447 }, { "epoch": 0.030354360051494004, "grad_norm": 16.025054931640625, "learning_rate": 5.96e-05, "loss": 1.4837, "step": 448 }, { "epoch": 0.03042211531946609, "grad_norm": 15.954922676086426, "learning_rate": 5.973333333333334e-05, "loss": 1.341, "step": 449 }, { "epoch": 0.030489870587438173, "grad_norm": 14.852401733398438, "learning_rate": 5.9866666666666674e-05, "loss": 1.2346, "step": 450 }, { "epoch": 0.030557625855410258, "grad_norm": 14.887676239013672, "learning_rate": 6e-05, "loss": 1.2789, "step": 451 }, { "epoch": 0.030625381123382343, "grad_norm": 17.25469970703125, "learning_rate": 6.013333333333334e-05, "loss": 1.4446, "step": 452 }, { "epoch": 0.030693136391354427, "grad_norm": 19.356597900390625, "learning_rate": 6.026666666666667e-05, "loss": 1.3765, "step": 453 }, { "epoch": 0.030760891659326512, "grad_norm": 14.16335391998291, "learning_rate": 6.04e-05, "loss": 1.4213, "step": 454 }, { "epoch": 0.030828646927298597, "grad_norm": 15.812528610229492, "learning_rate": 6.053333333333333e-05, "loss": 1.4028, "step": 455 }, { "epoch": 0.03089640219527068, "grad_norm": 15.10727596282959, "learning_rate": 6.066666666666667e-05, "loss": 1.3841, "step": 456 }, { "epoch": 0.030964157463242766, "grad_norm": 19.91744613647461, "learning_rate": 6.08e-05, "loss": 1.4706, "step": 457 }, { "epoch": 0.03103191273121485, "grad_norm": 14.597813606262207, "learning_rate": 6.093333333333333e-05, "loss": 1.2428, "step": 458 }, { "epoch": 0.031099667999186936, "grad_norm": 14.952363014221191, "learning_rate": 6.106666666666667e-05, "loss": 1.3431, "step": 459 }, { "epoch": 0.03116742326715902, "grad_norm": 16.519468307495117, "learning_rate": 6.12e-05, "loss": 1.5399, "step": 460 }, { "epoch": 0.031235178535131105, "grad_norm": 14.55786418914795, "learning_rate": 6.133333333333334e-05, "loss": 1.3844, "step": 461 }, { "epoch": 0.03130293380310319, "grad_norm": 12.52665901184082, "learning_rate": 6.146666666666668e-05, "loss": 1.2085, "step": 462 }, { "epoch": 0.03137068907107528, "grad_norm": 17.549148559570312, "learning_rate": 6.16e-05, "loss": 1.588, "step": 463 }, { "epoch": 0.03143844433904736, "grad_norm": 14.761232376098633, "learning_rate": 6.173333333333333e-05, "loss": 1.1685, "step": 464 }, { "epoch": 0.03150619960701945, "grad_norm": 13.1014404296875, "learning_rate": 6.186666666666668e-05, "loss": 1.1718, "step": 465 }, { "epoch": 0.03157395487499153, "grad_norm": 15.998059272766113, "learning_rate": 6.2e-05, "loss": 1.3177, "step": 466 }, { "epoch": 0.03164171014296362, "grad_norm": 16.402875900268555, "learning_rate": 6.213333333333333e-05, "loss": 1.4821, "step": 467 }, { "epoch": 0.0317094654109357, "grad_norm": 12.871467590332031, "learning_rate": 6.226666666666667e-05, "loss": 1.0176, "step": 468 }, { "epoch": 0.031777220678907786, "grad_norm": 15.602563858032227, "learning_rate": 6.24e-05, "loss": 1.3418, "step": 469 }, { "epoch": 0.03184497594687987, "grad_norm": 13.5369234085083, "learning_rate": 6.253333333333333e-05, "loss": 1.4051, "step": 470 }, { "epoch": 0.031912731214851955, "grad_norm": 16.841650009155273, "learning_rate": 6.266666666666667e-05, "loss": 1.5438, "step": 471 }, { "epoch": 0.03198048648282404, "grad_norm": 14.581981658935547, "learning_rate": 6.280000000000001e-05, "loss": 1.2767, "step": 472 }, { "epoch": 0.032048241750796125, "grad_norm": 14.084460258483887, "learning_rate": 6.293333333333334e-05, "loss": 1.2224, "step": 473 }, { "epoch": 0.03211599701876821, "grad_norm": 15.54316234588623, "learning_rate": 6.306666666666668e-05, "loss": 1.6084, "step": 474 }, { "epoch": 0.032183752286740294, "grad_norm": 15.498743057250977, "learning_rate": 6.32e-05, "loss": 1.3937, "step": 475 }, { "epoch": 0.03225150755471238, "grad_norm": 17.728227615356445, "learning_rate": 6.333333333333333e-05, "loss": 1.2996, "step": 476 }, { "epoch": 0.032319262822684464, "grad_norm": 13.956122398376465, "learning_rate": 6.346666666666667e-05, "loss": 1.4472, "step": 477 }, { "epoch": 0.03238701809065655, "grad_norm": 18.578060150146484, "learning_rate": 6.36e-05, "loss": 1.4659, "step": 478 }, { "epoch": 0.03245477335862863, "grad_norm": 14.055442810058594, "learning_rate": 6.373333333333333e-05, "loss": 1.1979, "step": 479 }, { "epoch": 0.03252252862660072, "grad_norm": 15.902195930480957, "learning_rate": 6.386666666666667e-05, "loss": 1.3367, "step": 480 }, { "epoch": 0.0325902838945728, "grad_norm": 14.212138175964355, "learning_rate": 6.400000000000001e-05, "loss": 1.2716, "step": 481 }, { "epoch": 0.03265803916254489, "grad_norm": 17.711475372314453, "learning_rate": 6.413333333333334e-05, "loss": 1.3016, "step": 482 }, { "epoch": 0.03272579443051697, "grad_norm": 13.068260192871094, "learning_rate": 6.426666666666668e-05, "loss": 1.1346, "step": 483 }, { "epoch": 0.032793549698489056, "grad_norm": 17.422321319580078, "learning_rate": 6.440000000000001e-05, "loss": 1.0931, "step": 484 }, { "epoch": 0.03286130496646114, "grad_norm": 14.98038101196289, "learning_rate": 6.453333333333333e-05, "loss": 1.22, "step": 485 }, { "epoch": 0.032929060234433226, "grad_norm": 13.300479888916016, "learning_rate": 6.466666666666666e-05, "loss": 1.2647, "step": 486 }, { "epoch": 0.03299681550240531, "grad_norm": 14.611360549926758, "learning_rate": 6.48e-05, "loss": 1.2478, "step": 487 }, { "epoch": 0.033064570770377395, "grad_norm": 19.244020462036133, "learning_rate": 6.493333333333333e-05, "loss": 1.2126, "step": 488 }, { "epoch": 0.03313232603834948, "grad_norm": 16.285396575927734, "learning_rate": 6.506666666666666e-05, "loss": 1.1694, "step": 489 }, { "epoch": 0.033200081306321565, "grad_norm": 19.613059997558594, "learning_rate": 6.52e-05, "loss": 1.2118, "step": 490 }, { "epoch": 0.03326783657429365, "grad_norm": 19.19630241394043, "learning_rate": 6.533333333333334e-05, "loss": 1.3093, "step": 491 }, { "epoch": 0.033335591842265734, "grad_norm": 12.653651237487793, "learning_rate": 6.546666666666667e-05, "loss": 1.1545, "step": 492 }, { "epoch": 0.03340334711023782, "grad_norm": 13.434443473815918, "learning_rate": 6.560000000000001e-05, "loss": 1.2543, "step": 493 }, { "epoch": 0.033471102378209903, "grad_norm": 18.893598556518555, "learning_rate": 6.573333333333334e-05, "loss": 1.3632, "step": 494 }, { "epoch": 0.03353885764618199, "grad_norm": 18.339479446411133, "learning_rate": 6.586666666666666e-05, "loss": 1.4369, "step": 495 }, { "epoch": 0.03360661291415407, "grad_norm": 12.303078651428223, "learning_rate": 6.6e-05, "loss": 1.4032, "step": 496 }, { "epoch": 0.03367436818212616, "grad_norm": 17.041015625, "learning_rate": 6.613333333333333e-05, "loss": 1.2395, "step": 497 }, { "epoch": 0.03374212345009824, "grad_norm": 14.035640716552734, "learning_rate": 6.626666666666666e-05, "loss": 1.3541, "step": 498 }, { "epoch": 0.03380987871807033, "grad_norm": 16.93412208557129, "learning_rate": 6.64e-05, "loss": 1.3503, "step": 499 }, { "epoch": 0.03387763398604241, "grad_norm": 16.02039909362793, "learning_rate": 6.653333333333334e-05, "loss": 1.3226, "step": 500 }, { "epoch": 0.033945389254014496, "grad_norm": 16.061542510986328, "learning_rate": 6.666666666666667e-05, "loss": 1.433, "step": 501 }, { "epoch": 0.03401314452198658, "grad_norm": 14.810320854187012, "learning_rate": 6.680000000000001e-05, "loss": 1.3715, "step": 502 }, { "epoch": 0.03408089978995867, "grad_norm": 15.132527351379395, "learning_rate": 6.693333333333334e-05, "loss": 1.0683, "step": 503 }, { "epoch": 0.03414865505793076, "grad_norm": 13.768165588378906, "learning_rate": 6.706666666666667e-05, "loss": 1.234, "step": 504 }, { "epoch": 0.03421641032590284, "grad_norm": 16.227920532226562, "learning_rate": 6.720000000000001e-05, "loss": 1.4181, "step": 505 }, { "epoch": 0.03428416559387493, "grad_norm": 14.205772399902344, "learning_rate": 6.733333333333333e-05, "loss": 1.4342, "step": 506 }, { "epoch": 0.03435192086184701, "grad_norm": 17.451099395751953, "learning_rate": 6.746666666666666e-05, "loss": 1.3338, "step": 507 }, { "epoch": 0.034419676129819096, "grad_norm": 13.7041015625, "learning_rate": 6.76e-05, "loss": 1.3972, "step": 508 }, { "epoch": 0.03448743139779118, "grad_norm": 13.206351280212402, "learning_rate": 6.773333333333333e-05, "loss": 1.4233, "step": 509 }, { "epoch": 0.034555186665763266, "grad_norm": 11.867975234985352, "learning_rate": 6.786666666666667e-05, "loss": 1.111, "step": 510 }, { "epoch": 0.03462294193373535, "grad_norm": 19.726648330688477, "learning_rate": 6.800000000000001e-05, "loss": 1.2918, "step": 511 }, { "epoch": 0.034690697201707435, "grad_norm": 18.914016723632812, "learning_rate": 6.813333333333334e-05, "loss": 1.5998, "step": 512 }, { "epoch": 0.03475845246967952, "grad_norm": 18.065937042236328, "learning_rate": 6.826666666666667e-05, "loss": 1.4952, "step": 513 }, { "epoch": 0.034826207737651604, "grad_norm": 16.42376136779785, "learning_rate": 6.840000000000001e-05, "loss": 1.2807, "step": 514 }, { "epoch": 0.03489396300562369, "grad_norm": 15.537372589111328, "learning_rate": 6.853333333333334e-05, "loss": 1.3524, "step": 515 }, { "epoch": 0.034961718273595774, "grad_norm": 18.395238876342773, "learning_rate": 6.866666666666666e-05, "loss": 1.3022, "step": 516 }, { "epoch": 0.03502947354156786, "grad_norm": 15.80531120300293, "learning_rate": 6.879999999999999e-05, "loss": 1.4201, "step": 517 }, { "epoch": 0.03509722880953994, "grad_norm": 15.361504554748535, "learning_rate": 6.893333333333333e-05, "loss": 1.4351, "step": 518 }, { "epoch": 0.03516498407751203, "grad_norm": 12.501819610595703, "learning_rate": 6.906666666666667e-05, "loss": 1.1894, "step": 519 }, { "epoch": 0.03523273934548411, "grad_norm": 14.53589916229248, "learning_rate": 6.92e-05, "loss": 1.3435, "step": 520 }, { "epoch": 0.0353004946134562, "grad_norm": 15.057633399963379, "learning_rate": 6.933333333333334e-05, "loss": 1.5467, "step": 521 }, { "epoch": 0.03536824988142828, "grad_norm": 16.27166175842285, "learning_rate": 6.946666666666667e-05, "loss": 1.5171, "step": 522 }, { "epoch": 0.03543600514940037, "grad_norm": 18.050413131713867, "learning_rate": 6.96e-05, "loss": 1.2246, "step": 523 }, { "epoch": 0.03550376041737245, "grad_norm": 15.316112518310547, "learning_rate": 6.973333333333334e-05, "loss": 1.2551, "step": 524 }, { "epoch": 0.035571515685344536, "grad_norm": 15.416853904724121, "learning_rate": 6.986666666666667e-05, "loss": 1.1895, "step": 525 }, { "epoch": 0.03563927095331662, "grad_norm": 15.019004821777344, "learning_rate": 7e-05, "loss": 1.5957, "step": 526 }, { "epoch": 0.035707026221288705, "grad_norm": 15.33927059173584, "learning_rate": 7.013333333333333e-05, "loss": 1.3536, "step": 527 }, { "epoch": 0.03577478148926079, "grad_norm": 16.403501510620117, "learning_rate": 7.026666666666668e-05, "loss": 1.6104, "step": 528 }, { "epoch": 0.035842536757232875, "grad_norm": 13.962714195251465, "learning_rate": 7.04e-05, "loss": 1.2122, "step": 529 }, { "epoch": 0.03591029202520496, "grad_norm": 14.530926704406738, "learning_rate": 7.053333333333334e-05, "loss": 1.2879, "step": 530 }, { "epoch": 0.035978047293177044, "grad_norm": 13.99305248260498, "learning_rate": 7.066666666666667e-05, "loss": 1.2063, "step": 531 }, { "epoch": 0.03604580256114913, "grad_norm": 17.54342269897461, "learning_rate": 7.08e-05, "loss": 1.2823, "step": 532 }, { "epoch": 0.036113557829121214, "grad_norm": 11.097670555114746, "learning_rate": 7.093333333333334e-05, "loss": 1.1658, "step": 533 }, { "epoch": 0.0361813130970933, "grad_norm": 19.069889068603516, "learning_rate": 7.106666666666667e-05, "loss": 1.3368, "step": 534 }, { "epoch": 0.03624906836506538, "grad_norm": 12.675487518310547, "learning_rate": 7.12e-05, "loss": 1.1586, "step": 535 }, { "epoch": 0.03631682363303747, "grad_norm": 16.971637725830078, "learning_rate": 7.133333333333334e-05, "loss": 1.2458, "step": 536 }, { "epoch": 0.03638457890100955, "grad_norm": 14.513422012329102, "learning_rate": 7.146666666666666e-05, "loss": 1.236, "step": 537 }, { "epoch": 0.03645233416898164, "grad_norm": 14.271836280822754, "learning_rate": 7.16e-05, "loss": 1.0951, "step": 538 }, { "epoch": 0.03652008943695372, "grad_norm": 14.286022186279297, "learning_rate": 7.173333333333335e-05, "loss": 1.2411, "step": 539 }, { "epoch": 0.03658784470492581, "grad_norm": 16.81059455871582, "learning_rate": 7.186666666666667e-05, "loss": 1.2453, "step": 540 }, { "epoch": 0.03665559997289789, "grad_norm": 16.818300247192383, "learning_rate": 7.2e-05, "loss": 1.2302, "step": 541 }, { "epoch": 0.036723355240869976, "grad_norm": 14.839925765991211, "learning_rate": 7.213333333333334e-05, "loss": 1.3445, "step": 542 }, { "epoch": 0.03679111050884206, "grad_norm": 14.917344093322754, "learning_rate": 7.226666666666667e-05, "loss": 1.3658, "step": 543 }, { "epoch": 0.036858865776814145, "grad_norm": 15.310625076293945, "learning_rate": 7.24e-05, "loss": 1.2565, "step": 544 }, { "epoch": 0.03692662104478623, "grad_norm": 16.7796630859375, "learning_rate": 7.253333333333334e-05, "loss": 1.265, "step": 545 }, { "epoch": 0.036994376312758315, "grad_norm": 19.4224796295166, "learning_rate": 7.266666666666667e-05, "loss": 1.6152, "step": 546 }, { "epoch": 0.0370621315807304, "grad_norm": 15.8001127243042, "learning_rate": 7.280000000000001e-05, "loss": 1.4313, "step": 547 }, { "epoch": 0.037129886848702484, "grad_norm": 13.597784042358398, "learning_rate": 7.293333333333334e-05, "loss": 1.3812, "step": 548 }, { "epoch": 0.03719764211667457, "grad_norm": 13.139778137207031, "learning_rate": 7.306666666666668e-05, "loss": 1.1888, "step": 549 }, { "epoch": 0.037265397384646654, "grad_norm": 17.083406448364258, "learning_rate": 7.32e-05, "loss": 1.1668, "step": 550 }, { "epoch": 0.03733315265261874, "grad_norm": 14.46076774597168, "learning_rate": 7.333333333333333e-05, "loss": 1.2394, "step": 551 }, { "epoch": 0.03740090792059082, "grad_norm": 16.217782974243164, "learning_rate": 7.346666666666667e-05, "loss": 1.3068, "step": 552 }, { "epoch": 0.03746866318856291, "grad_norm": 16.06308937072754, "learning_rate": 7.36e-05, "loss": 0.9678, "step": 553 }, { "epoch": 0.03753641845653499, "grad_norm": 14.652907371520996, "learning_rate": 7.373333333333333e-05, "loss": 1.5375, "step": 554 }, { "epoch": 0.03760417372450708, "grad_norm": 13.870722770690918, "learning_rate": 7.386666666666667e-05, "loss": 1.29, "step": 555 }, { "epoch": 0.03767192899247917, "grad_norm": 14.649571418762207, "learning_rate": 7.4e-05, "loss": 1.3588, "step": 556 }, { "epoch": 0.03773968426045125, "grad_norm": 13.698415756225586, "learning_rate": 7.413333333333334e-05, "loss": 1.0801, "step": 557 }, { "epoch": 0.03780743952842334, "grad_norm": 19.412424087524414, "learning_rate": 7.426666666666668e-05, "loss": 1.4474, "step": 558 }, { "epoch": 0.03787519479639542, "grad_norm": 18.58599853515625, "learning_rate": 7.44e-05, "loss": 1.3294, "step": 559 }, { "epoch": 0.03794295006436751, "grad_norm": 15.283289909362793, "learning_rate": 7.453333333333333e-05, "loss": 1.1554, "step": 560 }, { "epoch": 0.03801070533233959, "grad_norm": 16.289731979370117, "learning_rate": 7.466666666666667e-05, "loss": 1.222, "step": 561 }, { "epoch": 0.03807846060031168, "grad_norm": 25.105520248413086, "learning_rate": 7.48e-05, "loss": 1.2213, "step": 562 }, { "epoch": 0.03814621586828376, "grad_norm": 15.428693771362305, "learning_rate": 7.493333333333333e-05, "loss": 1.2165, "step": 563 }, { "epoch": 0.038213971136255846, "grad_norm": 18.79145622253418, "learning_rate": 7.506666666666667e-05, "loss": 1.4799, "step": 564 }, { "epoch": 0.03828172640422793, "grad_norm": 12.498862266540527, "learning_rate": 7.52e-05, "loss": 1.4348, "step": 565 }, { "epoch": 0.038349481672200016, "grad_norm": 19.302845001220703, "learning_rate": 7.533333333333334e-05, "loss": 1.2656, "step": 566 }, { "epoch": 0.0384172369401721, "grad_norm": 16.28862953186035, "learning_rate": 7.546666666666668e-05, "loss": 1.4779, "step": 567 }, { "epoch": 0.038484992208144185, "grad_norm": 13.972156524658203, "learning_rate": 7.560000000000001e-05, "loss": 1.4775, "step": 568 }, { "epoch": 0.03855274747611627, "grad_norm": 15.443683624267578, "learning_rate": 7.573333333333334e-05, "loss": 1.2671, "step": 569 }, { "epoch": 0.038620502744088354, "grad_norm": 15.203113555908203, "learning_rate": 7.586666666666668e-05, "loss": 1.4209, "step": 570 }, { "epoch": 0.03868825801206044, "grad_norm": 18.516462326049805, "learning_rate": 7.6e-05, "loss": 1.3315, "step": 571 }, { "epoch": 0.038756013280032524, "grad_norm": 13.403026580810547, "learning_rate": 7.613333333333333e-05, "loss": 1.364, "step": 572 }, { "epoch": 0.03882376854800461, "grad_norm": 12.339615821838379, "learning_rate": 7.626666666666667e-05, "loss": 1.2107, "step": 573 }, { "epoch": 0.03889152381597669, "grad_norm": 15.647716522216797, "learning_rate": 7.64e-05, "loss": 1.3165, "step": 574 }, { "epoch": 0.03895927908394878, "grad_norm": 15.854050636291504, "learning_rate": 7.653333333333333e-05, "loss": 1.517, "step": 575 }, { "epoch": 0.03902703435192086, "grad_norm": 11.765929222106934, "learning_rate": 7.666666666666667e-05, "loss": 1.2676, "step": 576 }, { "epoch": 0.03909478961989295, "grad_norm": 12.515352249145508, "learning_rate": 7.680000000000001e-05, "loss": 1.4472, "step": 577 }, { "epoch": 0.03916254488786503, "grad_norm": 11.72417163848877, "learning_rate": 7.693333333333334e-05, "loss": 1.1874, "step": 578 }, { "epoch": 0.03923030015583712, "grad_norm": 15.03148365020752, "learning_rate": 7.706666666666668e-05, "loss": 1.6381, "step": 579 }, { "epoch": 0.0392980554238092, "grad_norm": 14.0188570022583, "learning_rate": 7.72e-05, "loss": 1.4351, "step": 580 }, { "epoch": 0.039365810691781286, "grad_norm": 12.343233108520508, "learning_rate": 7.733333333333333e-05, "loss": 1.2293, "step": 581 }, { "epoch": 0.03943356595975337, "grad_norm": 15.358945846557617, "learning_rate": 7.746666666666666e-05, "loss": 1.307, "step": 582 }, { "epoch": 0.039501321227725456, "grad_norm": 14.131333351135254, "learning_rate": 7.76e-05, "loss": 1.0432, "step": 583 }, { "epoch": 0.03956907649569754, "grad_norm": 16.46926498413086, "learning_rate": 7.773333333333333e-05, "loss": 1.4126, "step": 584 }, { "epoch": 0.039636831763669625, "grad_norm": 15.209906578063965, "learning_rate": 7.786666666666667e-05, "loss": 1.3242, "step": 585 }, { "epoch": 0.03970458703164171, "grad_norm": 15.657282829284668, "learning_rate": 7.800000000000001e-05, "loss": 1.4967, "step": 586 }, { "epoch": 0.039772342299613794, "grad_norm": 11.466882705688477, "learning_rate": 7.813333333333334e-05, "loss": 1.3003, "step": 587 }, { "epoch": 0.03984009756758588, "grad_norm": 19.529300689697266, "learning_rate": 7.826666666666667e-05, "loss": 1.3835, "step": 588 }, { "epoch": 0.039907852835557964, "grad_norm": 17.22064971923828, "learning_rate": 7.840000000000001e-05, "loss": 1.2655, "step": 589 }, { "epoch": 0.03997560810353005, "grad_norm": 14.788103103637695, "learning_rate": 7.853333333333334e-05, "loss": 1.1981, "step": 590 }, { "epoch": 0.04004336337150213, "grad_norm": 13.705521583557129, "learning_rate": 7.866666666666666e-05, "loss": 1.2064, "step": 591 }, { "epoch": 0.04011111863947422, "grad_norm": 14.901930809020996, "learning_rate": 7.88e-05, "loss": 1.2632, "step": 592 }, { "epoch": 0.0401788739074463, "grad_norm": 18.520828247070312, "learning_rate": 7.893333333333333e-05, "loss": 1.4316, "step": 593 }, { "epoch": 0.04024662917541839, "grad_norm": 15.101353645324707, "learning_rate": 7.906666666666667e-05, "loss": 1.3935, "step": 594 }, { "epoch": 0.04031438444339047, "grad_norm": 13.961483001708984, "learning_rate": 7.920000000000001e-05, "loss": 1.1477, "step": 595 }, { "epoch": 0.04038213971136256, "grad_norm": 16.015554428100586, "learning_rate": 7.933333333333334e-05, "loss": 1.5029, "step": 596 }, { "epoch": 0.04044989497933464, "grad_norm": 15.009637832641602, "learning_rate": 7.946666666666667e-05, "loss": 1.3715, "step": 597 }, { "epoch": 0.040517650247306726, "grad_norm": 16.295202255249023, "learning_rate": 7.960000000000001e-05, "loss": 1.5129, "step": 598 }, { "epoch": 0.04058540551527881, "grad_norm": 88.3338623046875, "learning_rate": 7.973333333333334e-05, "loss": 1.2395, "step": 599 }, { "epoch": 0.040653160783250895, "grad_norm": 16.769424438476562, "learning_rate": 7.986666666666667e-05, "loss": 1.4258, "step": 600 }, { "epoch": 0.04072091605122298, "grad_norm": 16.044578552246094, "learning_rate": 8e-05, "loss": 1.365, "step": 601 }, { "epoch": 0.040788671319195065, "grad_norm": 15.282588958740234, "learning_rate": 8.013333333333333e-05, "loss": 1.3212, "step": 602 }, { "epoch": 0.04085642658716715, "grad_norm": 15.275490760803223, "learning_rate": 8.026666666666666e-05, "loss": 1.6048, "step": 603 }, { "epoch": 0.040924181855139234, "grad_norm": 12.952759742736816, "learning_rate": 8.04e-05, "loss": 1.1982, "step": 604 }, { "epoch": 0.04099193712311132, "grad_norm": 16.79343032836914, "learning_rate": 8.053333333333334e-05, "loss": 1.281, "step": 605 }, { "epoch": 0.041059692391083404, "grad_norm": 13.70421314239502, "learning_rate": 8.066666666666667e-05, "loss": 1.2664, "step": 606 }, { "epoch": 0.04112744765905549, "grad_norm": 14.420117378234863, "learning_rate": 8.080000000000001e-05, "loss": 1.2758, "step": 607 }, { "epoch": 0.04119520292702757, "grad_norm": 16.421335220336914, "learning_rate": 8.093333333333334e-05, "loss": 1.3105, "step": 608 }, { "epoch": 0.041262958194999665, "grad_norm": 16.96218490600586, "learning_rate": 8.106666666666667e-05, "loss": 1.4665, "step": 609 }, { "epoch": 0.04133071346297175, "grad_norm": 13.73725700378418, "learning_rate": 8.120000000000001e-05, "loss": 1.2109, "step": 610 }, { "epoch": 0.041398468730943834, "grad_norm": 14.718022346496582, "learning_rate": 8.133333333333334e-05, "loss": 1.2051, "step": 611 }, { "epoch": 0.04146622399891592, "grad_norm": 14.119061470031738, "learning_rate": 8.146666666666666e-05, "loss": 1.0731, "step": 612 }, { "epoch": 0.041533979266888, "grad_norm": 17.90053367614746, "learning_rate": 8.16e-05, "loss": 1.3804, "step": 613 }, { "epoch": 0.04160173453486009, "grad_norm": 12.727055549621582, "learning_rate": 8.173333333333335e-05, "loss": 1.0926, "step": 614 }, { "epoch": 0.04166948980283217, "grad_norm": 14.953054428100586, "learning_rate": 8.186666666666667e-05, "loss": 1.1974, "step": 615 }, { "epoch": 0.04173724507080426, "grad_norm": 16.05322265625, "learning_rate": 8.2e-05, "loss": 1.2526, "step": 616 }, { "epoch": 0.04180500033877634, "grad_norm": 15.028168678283691, "learning_rate": 8.213333333333334e-05, "loss": 1.3593, "step": 617 }, { "epoch": 0.04187275560674843, "grad_norm": 13.127458572387695, "learning_rate": 8.226666666666667e-05, "loss": 1.3981, "step": 618 }, { "epoch": 0.04194051087472051, "grad_norm": 17.323017120361328, "learning_rate": 8.24e-05, "loss": 1.4622, "step": 619 }, { "epoch": 0.042008266142692596, "grad_norm": 16.106731414794922, "learning_rate": 8.253333333333334e-05, "loss": 1.257, "step": 620 }, { "epoch": 0.04207602141066468, "grad_norm": 13.039103507995605, "learning_rate": 8.266666666666667e-05, "loss": 1.2905, "step": 621 }, { "epoch": 0.042143776678636766, "grad_norm": 13.661933898925781, "learning_rate": 8.28e-05, "loss": 1.2779, "step": 622 }, { "epoch": 0.04221153194660885, "grad_norm": 17.325756072998047, "learning_rate": 8.293333333333333e-05, "loss": 1.3934, "step": 623 }, { "epoch": 0.042279287214580935, "grad_norm": 13.898777961730957, "learning_rate": 8.306666666666668e-05, "loss": 1.3735, "step": 624 }, { "epoch": 0.04234704248255302, "grad_norm": 16.787601470947266, "learning_rate": 8.32e-05, "loss": 1.5228, "step": 625 }, { "epoch": 0.042414797750525104, "grad_norm": 13.54299259185791, "learning_rate": 8.333333333333334e-05, "loss": 1.0538, "step": 626 }, { "epoch": 0.04248255301849719, "grad_norm": 14.418194770812988, "learning_rate": 8.346666666666667e-05, "loss": 1.2932, "step": 627 }, { "epoch": 0.042550308286469274, "grad_norm": 13.39255142211914, "learning_rate": 8.36e-05, "loss": 1.1989, "step": 628 }, { "epoch": 0.04261806355444136, "grad_norm": 15.445747375488281, "learning_rate": 8.373333333333334e-05, "loss": 1.3557, "step": 629 }, { "epoch": 0.04268581882241344, "grad_norm": 13.414338111877441, "learning_rate": 8.386666666666667e-05, "loss": 1.1729, "step": 630 }, { "epoch": 0.04275357409038553, "grad_norm": 16.649137496948242, "learning_rate": 8.4e-05, "loss": 1.4212, "step": 631 }, { "epoch": 0.04282132935835761, "grad_norm": 17.583528518676758, "learning_rate": 8.413333333333334e-05, "loss": 1.4838, "step": 632 }, { "epoch": 0.0428890846263297, "grad_norm": 19.32307243347168, "learning_rate": 8.426666666666668e-05, "loss": 1.5958, "step": 633 }, { "epoch": 0.04295683989430178, "grad_norm": 12.703327178955078, "learning_rate": 8.44e-05, "loss": 1.2752, "step": 634 }, { "epoch": 0.04302459516227387, "grad_norm": 15.72768497467041, "learning_rate": 8.453333333333335e-05, "loss": 1.4059, "step": 635 }, { "epoch": 0.04309235043024595, "grad_norm": 13.532344818115234, "learning_rate": 8.466666666666667e-05, "loss": 1.2252, "step": 636 }, { "epoch": 0.043160105698218036, "grad_norm": 13.305481910705566, "learning_rate": 8.48e-05, "loss": 1.5084, "step": 637 }, { "epoch": 0.04322786096619012, "grad_norm": 11.986043930053711, "learning_rate": 8.493333333333334e-05, "loss": 1.2006, "step": 638 }, { "epoch": 0.043295616234162206, "grad_norm": 13.28003978729248, "learning_rate": 8.506666666666667e-05, "loss": 1.2442, "step": 639 }, { "epoch": 0.04336337150213429, "grad_norm": 15.835545539855957, "learning_rate": 8.52e-05, "loss": 1.2466, "step": 640 }, { "epoch": 0.043431126770106375, "grad_norm": 14.887584686279297, "learning_rate": 8.533333333333334e-05, "loss": 1.3828, "step": 641 }, { "epoch": 0.04349888203807846, "grad_norm": 14.740251541137695, "learning_rate": 8.546666666666667e-05, "loss": 1.4847, "step": 642 }, { "epoch": 0.043566637306050544, "grad_norm": 14.109213829040527, "learning_rate": 8.560000000000001e-05, "loss": 1.2251, "step": 643 }, { "epoch": 0.04363439257402263, "grad_norm": 17.33670425415039, "learning_rate": 8.573333333333333e-05, "loss": 1.3817, "step": 644 }, { "epoch": 0.043702147841994714, "grad_norm": 16.92241096496582, "learning_rate": 8.586666666666668e-05, "loss": 1.458, "step": 645 }, { "epoch": 0.0437699031099668, "grad_norm": 15.369187355041504, "learning_rate": 8.6e-05, "loss": 1.28, "step": 646 }, { "epoch": 0.04383765837793888, "grad_norm": 18.748065948486328, "learning_rate": 8.613333333333333e-05, "loss": 1.6275, "step": 647 }, { "epoch": 0.04390541364591097, "grad_norm": 15.650605201721191, "learning_rate": 8.626666666666667e-05, "loss": 1.3742, "step": 648 }, { "epoch": 0.04397316891388305, "grad_norm": 16.170730590820312, "learning_rate": 8.64e-05, "loss": 1.174, "step": 649 }, { "epoch": 0.04404092418185514, "grad_norm": 12.61952018737793, "learning_rate": 8.653333333333333e-05, "loss": 1.2178, "step": 650 }, { "epoch": 0.04410867944982722, "grad_norm": 18.86512565612793, "learning_rate": 8.666666666666667e-05, "loss": 1.4993, "step": 651 }, { "epoch": 0.04417643471779931, "grad_norm": 16.412899017333984, "learning_rate": 8.680000000000001e-05, "loss": 1.2663, "step": 652 }, { "epoch": 0.04424418998577139, "grad_norm": 14.11953353881836, "learning_rate": 8.693333333333334e-05, "loss": 1.3772, "step": 653 }, { "epoch": 0.044311945253743476, "grad_norm": 17.232633590698242, "learning_rate": 8.706666666666668e-05, "loss": 1.3252, "step": 654 }, { "epoch": 0.04437970052171556, "grad_norm": 13.833864212036133, "learning_rate": 8.72e-05, "loss": 1.279, "step": 655 }, { "epoch": 0.044447455789687645, "grad_norm": 11.546829223632812, "learning_rate": 8.733333333333333e-05, "loss": 1.382, "step": 656 }, { "epoch": 0.04451521105765973, "grad_norm": 13.163644790649414, "learning_rate": 8.746666666666667e-05, "loss": 1.2253, "step": 657 }, { "epoch": 0.044582966325631815, "grad_norm": 17.037311553955078, "learning_rate": 8.76e-05, "loss": 1.2727, "step": 658 }, { "epoch": 0.0446507215936039, "grad_norm": 14.610177040100098, "learning_rate": 8.773333333333333e-05, "loss": 1.3082, "step": 659 }, { "epoch": 0.044718476861575984, "grad_norm": 16.305557250976562, "learning_rate": 8.786666666666667e-05, "loss": 1.3948, "step": 660 }, { "epoch": 0.04478623212954807, "grad_norm": 13.207799911499023, "learning_rate": 8.800000000000001e-05, "loss": 1.2764, "step": 661 }, { "epoch": 0.04485398739752016, "grad_norm": 11.451075553894043, "learning_rate": 8.813333333333334e-05, "loss": 1.0831, "step": 662 }, { "epoch": 0.044921742665492245, "grad_norm": 13.555370330810547, "learning_rate": 8.826666666666668e-05, "loss": 1.4996, "step": 663 }, { "epoch": 0.04498949793346433, "grad_norm": 13.544769287109375, "learning_rate": 8.840000000000001e-05, "loss": 1.2277, "step": 664 }, { "epoch": 0.045057253201436415, "grad_norm": 18.05879783630371, "learning_rate": 8.853333333333333e-05, "loss": 1.3964, "step": 665 }, { "epoch": 0.0451250084694085, "grad_norm": 17.309839248657227, "learning_rate": 8.866666666666668e-05, "loss": 1.6233, "step": 666 }, { "epoch": 0.045192763737380584, "grad_norm": 12.732510566711426, "learning_rate": 8.88e-05, "loss": 1.3237, "step": 667 }, { "epoch": 0.04526051900535267, "grad_norm": 13.541101455688477, "learning_rate": 8.893333333333333e-05, "loss": 1.2523, "step": 668 }, { "epoch": 0.04532827427332475, "grad_norm": 17.54905891418457, "learning_rate": 8.906666666666667e-05, "loss": 1.5824, "step": 669 }, { "epoch": 0.04539602954129684, "grad_norm": 12.52578353881836, "learning_rate": 8.92e-05, "loss": 1.2966, "step": 670 }, { "epoch": 0.04546378480926892, "grad_norm": 13.279097557067871, "learning_rate": 8.933333333333334e-05, "loss": 1.4898, "step": 671 }, { "epoch": 0.04553154007724101, "grad_norm": 15.892850875854492, "learning_rate": 8.946666666666668e-05, "loss": 1.3757, "step": 672 }, { "epoch": 0.04559929534521309, "grad_norm": 14.108098983764648, "learning_rate": 8.960000000000001e-05, "loss": 1.1758, "step": 673 }, { "epoch": 0.04566705061318518, "grad_norm": 17.15204429626465, "learning_rate": 8.973333333333334e-05, "loss": 1.3876, "step": 674 }, { "epoch": 0.04573480588115726, "grad_norm": 14.453113555908203, "learning_rate": 8.986666666666666e-05, "loss": 1.2152, "step": 675 }, { "epoch": 0.045802561149129346, "grad_norm": 17.9672794342041, "learning_rate": 9e-05, "loss": 1.2962, "step": 676 }, { "epoch": 0.04587031641710143, "grad_norm": 19.810890197753906, "learning_rate": 9.013333333333333e-05, "loss": 1.3374, "step": 677 }, { "epoch": 0.045938071685073516, "grad_norm": 16.13353729248047, "learning_rate": 9.026666666666666e-05, "loss": 1.4125, "step": 678 }, { "epoch": 0.0460058269530456, "grad_norm": 15.257608413696289, "learning_rate": 9.04e-05, "loss": 1.1931, "step": 679 }, { "epoch": 0.046073582221017685, "grad_norm": 16.88699722290039, "learning_rate": 9.053333333333334e-05, "loss": 1.362, "step": 680 }, { "epoch": 0.04614133748898977, "grad_norm": 15.46777057647705, "learning_rate": 9.066666666666667e-05, "loss": 1.3522, "step": 681 }, { "epoch": 0.046209092756961855, "grad_norm": 13.584056854248047, "learning_rate": 9.080000000000001e-05, "loss": 1.1998, "step": 682 }, { "epoch": 0.04627684802493394, "grad_norm": 14.226449966430664, "learning_rate": 9.093333333333334e-05, "loss": 1.3312, "step": 683 }, { "epoch": 0.046344603292906024, "grad_norm": 15.157097816467285, "learning_rate": 9.106666666666667e-05, "loss": 1.3422, "step": 684 }, { "epoch": 0.04641235856087811, "grad_norm": 14.748275756835938, "learning_rate": 9.120000000000001e-05, "loss": 1.2902, "step": 685 }, { "epoch": 0.04648011382885019, "grad_norm": 13.730619430541992, "learning_rate": 9.133333333333334e-05, "loss": 1.3034, "step": 686 }, { "epoch": 0.04654786909682228, "grad_norm": 12.804062843322754, "learning_rate": 9.146666666666666e-05, "loss": 1.1484, "step": 687 }, { "epoch": 0.04661562436479436, "grad_norm": 16.587923049926758, "learning_rate": 9.16e-05, "loss": 1.4875, "step": 688 }, { "epoch": 0.04668337963276645, "grad_norm": 12.228131294250488, "learning_rate": 9.173333333333333e-05, "loss": 0.917, "step": 689 }, { "epoch": 0.04675113490073853, "grad_norm": 18.680187225341797, "learning_rate": 9.186666666666667e-05, "loss": 1.497, "step": 690 }, { "epoch": 0.04681889016871062, "grad_norm": 14.80630111694336, "learning_rate": 9.200000000000001e-05, "loss": 1.5623, "step": 691 }, { "epoch": 0.0468866454366827, "grad_norm": 13.754642486572266, "learning_rate": 9.213333333333334e-05, "loss": 1.3921, "step": 692 }, { "epoch": 0.046954400704654786, "grad_norm": 14.264236450195312, "learning_rate": 9.226666666666667e-05, "loss": 1.3615, "step": 693 }, { "epoch": 0.04702215597262687, "grad_norm": 15.883113861083984, "learning_rate": 9.240000000000001e-05, "loss": 1.4528, "step": 694 }, { "epoch": 0.047089911240598956, "grad_norm": 13.528610229492188, "learning_rate": 9.253333333333334e-05, "loss": 1.0995, "step": 695 }, { "epoch": 0.04715766650857104, "grad_norm": 15.945343971252441, "learning_rate": 9.266666666666666e-05, "loss": 1.272, "step": 696 }, { "epoch": 0.047225421776543125, "grad_norm": 14.378050804138184, "learning_rate": 9.28e-05, "loss": 1.3545, "step": 697 }, { "epoch": 0.04729317704451521, "grad_norm": 13.457077980041504, "learning_rate": 9.293333333333333e-05, "loss": 1.1751, "step": 698 }, { "epoch": 0.047360932312487294, "grad_norm": 19.285078048706055, "learning_rate": 9.306666666666667e-05, "loss": 1.4383, "step": 699 }, { "epoch": 0.04742868758045938, "grad_norm": 16.683856964111328, "learning_rate": 9.320000000000002e-05, "loss": 1.2499, "step": 700 }, { "epoch": 0.047496442848431464, "grad_norm": 13.818337440490723, "learning_rate": 9.333333333333334e-05, "loss": 1.2325, "step": 701 }, { "epoch": 0.04756419811640355, "grad_norm": 12.51142406463623, "learning_rate": 9.346666666666667e-05, "loss": 1.3317, "step": 702 }, { "epoch": 0.04763195338437563, "grad_norm": 14.697171211242676, "learning_rate": 9.360000000000001e-05, "loss": 1.6055, "step": 703 }, { "epoch": 0.04769970865234772, "grad_norm": 16.942562103271484, "learning_rate": 9.373333333333334e-05, "loss": 1.4279, "step": 704 }, { "epoch": 0.0477674639203198, "grad_norm": 16.739248275756836, "learning_rate": 9.386666666666667e-05, "loss": 1.2285, "step": 705 }, { "epoch": 0.04783521918829189, "grad_norm": 14.479548454284668, "learning_rate": 9.4e-05, "loss": 1.1984, "step": 706 }, { "epoch": 0.04790297445626397, "grad_norm": 14.384824752807617, "learning_rate": 9.413333333333334e-05, "loss": 1.2128, "step": 707 }, { "epoch": 0.04797072972423606, "grad_norm": 15.41884994506836, "learning_rate": 9.426666666666666e-05, "loss": 1.1627, "step": 708 }, { "epoch": 0.04803848499220814, "grad_norm": 13.267902374267578, "learning_rate": 9.44e-05, "loss": 1.2916, "step": 709 }, { "epoch": 0.048106240260180226, "grad_norm": 14.909900665283203, "learning_rate": 9.453333333333335e-05, "loss": 1.1579, "step": 710 }, { "epoch": 0.04817399552815231, "grad_norm": 12.639839172363281, "learning_rate": 9.466666666666667e-05, "loss": 1.2053, "step": 711 }, { "epoch": 0.048241750796124395, "grad_norm": 14.957659721374512, "learning_rate": 9.48e-05, "loss": 1.4556, "step": 712 }, { "epoch": 0.04830950606409648, "grad_norm": 14.008201599121094, "learning_rate": 9.493333333333334e-05, "loss": 1.1575, "step": 713 }, { "epoch": 0.048377261332068565, "grad_norm": 13.857501983642578, "learning_rate": 9.506666666666667e-05, "loss": 1.4121, "step": 714 }, { "epoch": 0.048445016600040657, "grad_norm": 13.867570877075195, "learning_rate": 9.52e-05, "loss": 1.4908, "step": 715 }, { "epoch": 0.04851277186801274, "grad_norm": 14.3545503616333, "learning_rate": 9.533333333333334e-05, "loss": 1.2939, "step": 716 }, { "epoch": 0.048580527135984826, "grad_norm": 15.173371315002441, "learning_rate": 9.546666666666667e-05, "loss": 1.3912, "step": 717 }, { "epoch": 0.04864828240395691, "grad_norm": 13.799453735351562, "learning_rate": 9.56e-05, "loss": 1.4219, "step": 718 }, { "epoch": 0.048716037671928995, "grad_norm": 18.264965057373047, "learning_rate": 9.573333333333335e-05, "loss": 1.4967, "step": 719 }, { "epoch": 0.04878379293990108, "grad_norm": 13.809319496154785, "learning_rate": 9.586666666666667e-05, "loss": 1.3527, "step": 720 }, { "epoch": 0.048851548207873165, "grad_norm": 13.931517601013184, "learning_rate": 9.6e-05, "loss": 1.4018, "step": 721 }, { "epoch": 0.04891930347584525, "grad_norm": 11.74387264251709, "learning_rate": 9.613333333333334e-05, "loss": 1.1981, "step": 722 }, { "epoch": 0.048987058743817334, "grad_norm": 14.960589408874512, "learning_rate": 9.626666666666667e-05, "loss": 1.3373, "step": 723 }, { "epoch": 0.04905481401178942, "grad_norm": 14.170279502868652, "learning_rate": 9.64e-05, "loss": 1.2702, "step": 724 }, { "epoch": 0.049122569279761504, "grad_norm": 15.532767295837402, "learning_rate": 9.653333333333334e-05, "loss": 1.1329, "step": 725 }, { "epoch": 0.04919032454773359, "grad_norm": 12.141284942626953, "learning_rate": 9.666666666666667e-05, "loss": 1.2694, "step": 726 }, { "epoch": 0.04925807981570567, "grad_norm": 12.379233360290527, "learning_rate": 9.680000000000001e-05, "loss": 1.1545, "step": 727 }, { "epoch": 0.04932583508367776, "grad_norm": 14.45605182647705, "learning_rate": 9.693333333333335e-05, "loss": 1.4151, "step": 728 }, { "epoch": 0.04939359035164984, "grad_norm": 15.656912803649902, "learning_rate": 9.706666666666668e-05, "loss": 1.3782, "step": 729 }, { "epoch": 0.04946134561962193, "grad_norm": 13.992748260498047, "learning_rate": 9.72e-05, "loss": 1.2731, "step": 730 }, { "epoch": 0.04952910088759401, "grad_norm": 16.253517150878906, "learning_rate": 9.733333333333335e-05, "loss": 1.4503, "step": 731 }, { "epoch": 0.049596856155566096, "grad_norm": 14.88161563873291, "learning_rate": 9.746666666666667e-05, "loss": 1.3909, "step": 732 }, { "epoch": 0.04966461142353818, "grad_norm": 14.60545539855957, "learning_rate": 9.76e-05, "loss": 1.1803, "step": 733 }, { "epoch": 0.049732366691510266, "grad_norm": 14.589803695678711, "learning_rate": 9.773333333333334e-05, "loss": 1.3844, "step": 734 }, { "epoch": 0.04980012195948235, "grad_norm": 14.781076431274414, "learning_rate": 9.786666666666667e-05, "loss": 1.2706, "step": 735 }, { "epoch": 0.049867877227454435, "grad_norm": 13.717072486877441, "learning_rate": 9.8e-05, "loss": 1.2311, "step": 736 }, { "epoch": 0.04993563249542652, "grad_norm": 17.14280891418457, "learning_rate": 9.813333333333334e-05, "loss": 1.3423, "step": 737 }, { "epoch": 0.050003387763398605, "grad_norm": 14.521373748779297, "learning_rate": 9.826666666666668e-05, "loss": 1.3118, "step": 738 }, { "epoch": 0.05007114303137069, "grad_norm": 14.79515266418457, "learning_rate": 9.84e-05, "loss": 1.3622, "step": 739 }, { "epoch": 0.050138898299342774, "grad_norm": 13.63962459564209, "learning_rate": 9.853333333333333e-05, "loss": 1.3856, "step": 740 }, { "epoch": 0.05020665356731486, "grad_norm": 13.788541793823242, "learning_rate": 9.866666666666668e-05, "loss": 1.3439, "step": 741 }, { "epoch": 0.05027440883528694, "grad_norm": 14.015862464904785, "learning_rate": 9.88e-05, "loss": 1.3263, "step": 742 }, { "epoch": 0.05034216410325903, "grad_norm": 15.407167434692383, "learning_rate": 9.893333333333333e-05, "loss": 1.5115, "step": 743 }, { "epoch": 0.05040991937123111, "grad_norm": 14.542003631591797, "learning_rate": 9.906666666666667e-05, "loss": 1.3934, "step": 744 }, { "epoch": 0.0504776746392032, "grad_norm": 16.733686447143555, "learning_rate": 9.92e-05, "loss": 1.2098, "step": 745 }, { "epoch": 0.05054542990717528, "grad_norm": 14.979568481445312, "learning_rate": 9.933333333333334e-05, "loss": 1.3234, "step": 746 }, { "epoch": 0.05061318517514737, "grad_norm": 13.699801445007324, "learning_rate": 9.946666666666668e-05, "loss": 1.3173, "step": 747 }, { "epoch": 0.05068094044311945, "grad_norm": 13.262088775634766, "learning_rate": 9.960000000000001e-05, "loss": 1.3459, "step": 748 }, { "epoch": 0.050748695711091536, "grad_norm": 14.694448471069336, "learning_rate": 9.973333333333334e-05, "loss": 1.3639, "step": 749 }, { "epoch": 0.05081645097906362, "grad_norm": 13.999008178710938, "learning_rate": 9.986666666666668e-05, "loss": 1.4275, "step": 750 }, { "epoch": 0.050884206247035706, "grad_norm": 11.588037490844727, "learning_rate": 0.0001, "loss": 1.2866, "step": 751 }, { "epoch": 0.05095196151500779, "grad_norm": 16.174110412597656, "learning_rate": 9.999863098090219e-05, "loss": 1.5338, "step": 752 }, { "epoch": 0.051019716782979875, "grad_norm": 17.8669376373291, "learning_rate": 9.999726196180437e-05, "loss": 1.3247, "step": 753 }, { "epoch": 0.05108747205095196, "grad_norm": 13.849075317382812, "learning_rate": 9.999589294270656e-05, "loss": 1.1817, "step": 754 }, { "epoch": 0.051155227318924044, "grad_norm": 14.888331413269043, "learning_rate": 9.999452392360874e-05, "loss": 1.4019, "step": 755 }, { "epoch": 0.05122298258689613, "grad_norm": 18.68206024169922, "learning_rate": 9.999315490451092e-05, "loss": 1.7115, "step": 756 }, { "epoch": 0.051290737854868214, "grad_norm": 14.762079238891602, "learning_rate": 9.99917858854131e-05, "loss": 1.2553, "step": 757 }, { "epoch": 0.0513584931228403, "grad_norm": 14.649972915649414, "learning_rate": 9.99904168663153e-05, "loss": 1.0807, "step": 758 }, { "epoch": 0.05142624839081238, "grad_norm": 13.172977447509766, "learning_rate": 9.998904784721747e-05, "loss": 1.4966, "step": 759 }, { "epoch": 0.05149400365878447, "grad_norm": 14.144796371459961, "learning_rate": 9.998767882811965e-05, "loss": 1.234, "step": 760 }, { "epoch": 0.05156175892675655, "grad_norm": 17.565507888793945, "learning_rate": 9.998630980902184e-05, "loss": 1.4061, "step": 761 }, { "epoch": 0.05162951419472864, "grad_norm": 17.205589294433594, "learning_rate": 9.998494078992402e-05, "loss": 1.6626, "step": 762 }, { "epoch": 0.05169726946270072, "grad_norm": 16.14542007446289, "learning_rate": 9.998357177082621e-05, "loss": 1.3145, "step": 763 }, { "epoch": 0.05176502473067281, "grad_norm": 14.692976951599121, "learning_rate": 9.998220275172839e-05, "loss": 1.2085, "step": 764 }, { "epoch": 0.05183277999864489, "grad_norm": 12.572774887084961, "learning_rate": 9.998083373263057e-05, "loss": 1.3743, "step": 765 }, { "epoch": 0.051900535266616976, "grad_norm": 15.904753684997559, "learning_rate": 9.997946471353275e-05, "loss": 1.3863, "step": 766 }, { "epoch": 0.05196829053458906, "grad_norm": 14.471494674682617, "learning_rate": 9.997809569443493e-05, "loss": 1.3615, "step": 767 }, { "epoch": 0.05203604580256115, "grad_norm": 13.82172966003418, "learning_rate": 9.997672667533712e-05, "loss": 1.2104, "step": 768 }, { "epoch": 0.05210380107053324, "grad_norm": 14.648073196411133, "learning_rate": 9.99753576562393e-05, "loss": 1.1166, "step": 769 }, { "epoch": 0.05217155633850532, "grad_norm": 16.776750564575195, "learning_rate": 9.997398863714149e-05, "loss": 1.4242, "step": 770 }, { "epoch": 0.05223931160647741, "grad_norm": 13.999717712402344, "learning_rate": 9.997261961804367e-05, "loss": 1.2493, "step": 771 }, { "epoch": 0.05230706687444949, "grad_norm": 13.238443374633789, "learning_rate": 9.997125059894586e-05, "loss": 1.0469, "step": 772 }, { "epoch": 0.052374822142421576, "grad_norm": 12.79283332824707, "learning_rate": 9.996988157984804e-05, "loss": 1.1975, "step": 773 }, { "epoch": 0.05244257741039366, "grad_norm": 15.577791213989258, "learning_rate": 9.996851256075022e-05, "loss": 1.4972, "step": 774 }, { "epoch": 0.052510332678365745, "grad_norm": 16.215747833251953, "learning_rate": 9.996714354165241e-05, "loss": 1.4285, "step": 775 }, { "epoch": 0.05257808794633783, "grad_norm": 14.68825626373291, "learning_rate": 9.99657745225546e-05, "loss": 1.2028, "step": 776 }, { "epoch": 0.052645843214309915, "grad_norm": 11.535130500793457, "learning_rate": 9.996440550345677e-05, "loss": 1.2466, "step": 777 }, { "epoch": 0.052713598482282, "grad_norm": 14.865918159484863, "learning_rate": 9.996303648435897e-05, "loss": 1.3903, "step": 778 }, { "epoch": 0.052781353750254084, "grad_norm": 13.626388549804688, "learning_rate": 9.996166746526115e-05, "loss": 1.5446, "step": 779 }, { "epoch": 0.05284910901822617, "grad_norm": 12.957473754882812, "learning_rate": 9.996029844616333e-05, "loss": 1.1641, "step": 780 }, { "epoch": 0.052916864286198254, "grad_norm": 14.928487777709961, "learning_rate": 9.995892942706552e-05, "loss": 1.0929, "step": 781 }, { "epoch": 0.05298461955417034, "grad_norm": 13.592952728271484, "learning_rate": 9.99575604079677e-05, "loss": 1.4684, "step": 782 }, { "epoch": 0.05305237482214242, "grad_norm": 13.836453437805176, "learning_rate": 9.995619138886988e-05, "loss": 1.0128, "step": 783 }, { "epoch": 0.05312013009011451, "grad_norm": 16.654394149780273, "learning_rate": 9.995482236977206e-05, "loss": 1.2464, "step": 784 }, { "epoch": 0.05318788535808659, "grad_norm": 14.4434232711792, "learning_rate": 9.995345335067424e-05, "loss": 1.3649, "step": 785 }, { "epoch": 0.05325564062605868, "grad_norm": 14.15592098236084, "learning_rate": 9.995208433157644e-05, "loss": 1.4426, "step": 786 }, { "epoch": 0.05332339589403076, "grad_norm": 12.97901439666748, "learning_rate": 9.995071531247862e-05, "loss": 1.227, "step": 787 }, { "epoch": 0.053391151162002846, "grad_norm": 13.167902946472168, "learning_rate": 9.99493462933808e-05, "loss": 1.2432, "step": 788 }, { "epoch": 0.05345890642997493, "grad_norm": 15.92297077178955, "learning_rate": 9.994797727428298e-05, "loss": 1.3345, "step": 789 }, { "epoch": 0.053526661697947016, "grad_norm": 13.30277156829834, "learning_rate": 9.994660825518517e-05, "loss": 1.2361, "step": 790 }, { "epoch": 0.0535944169659191, "grad_norm": 15.98779582977295, "learning_rate": 9.994523923608735e-05, "loss": 1.1202, "step": 791 }, { "epoch": 0.053662172233891185, "grad_norm": 16.414382934570312, "learning_rate": 9.994387021698953e-05, "loss": 1.448, "step": 792 }, { "epoch": 0.05372992750186327, "grad_norm": 14.491677284240723, "learning_rate": 9.994250119789171e-05, "loss": 1.0862, "step": 793 }, { "epoch": 0.053797682769835355, "grad_norm": 13.155410766601562, "learning_rate": 9.99411321787939e-05, "loss": 1.312, "step": 794 }, { "epoch": 0.05386543803780744, "grad_norm": 16.11139488220215, "learning_rate": 9.993976315969609e-05, "loss": 1.2043, "step": 795 }, { "epoch": 0.053933193305779524, "grad_norm": 14.259698867797852, "learning_rate": 9.993839414059827e-05, "loss": 1.2256, "step": 796 }, { "epoch": 0.05400094857375161, "grad_norm": 15.751099586486816, "learning_rate": 9.993702512150045e-05, "loss": 1.0732, "step": 797 }, { "epoch": 0.05406870384172369, "grad_norm": 12.766170501708984, "learning_rate": 9.993565610240263e-05, "loss": 1.0362, "step": 798 }, { "epoch": 0.05413645910969578, "grad_norm": 14.61483097076416, "learning_rate": 9.993428708330481e-05, "loss": 1.3107, "step": 799 }, { "epoch": 0.05420421437766786, "grad_norm": 14.571990013122559, "learning_rate": 9.9932918064207e-05, "loss": 1.3616, "step": 800 }, { "epoch": 0.05427196964563995, "grad_norm": 14.69124984741211, "learning_rate": 9.993154904510918e-05, "loss": 1.4273, "step": 801 }, { "epoch": 0.05433972491361203, "grad_norm": 13.834383010864258, "learning_rate": 9.993018002601136e-05, "loss": 1.2886, "step": 802 }, { "epoch": 0.05440748018158412, "grad_norm": 14.178943634033203, "learning_rate": 9.992881100691355e-05, "loss": 1.44, "step": 803 }, { "epoch": 0.0544752354495562, "grad_norm": 13.185267448425293, "learning_rate": 9.992744198781574e-05, "loss": 1.5609, "step": 804 }, { "epoch": 0.054542990717528286, "grad_norm": 15.211031913757324, "learning_rate": 9.992607296871792e-05, "loss": 1.8873, "step": 805 }, { "epoch": 0.05461074598550037, "grad_norm": 11.855447769165039, "learning_rate": 9.99247039496201e-05, "loss": 1.4063, "step": 806 }, { "epoch": 0.054678501253472456, "grad_norm": 14.352787017822266, "learning_rate": 9.992333493052228e-05, "loss": 1.56, "step": 807 }, { "epoch": 0.05474625652144454, "grad_norm": 14.339908599853516, "learning_rate": 9.992196591142446e-05, "loss": 1.1985, "step": 808 }, { "epoch": 0.054814011789416625, "grad_norm": 12.696648597717285, "learning_rate": 9.992059689232665e-05, "loss": 1.1807, "step": 809 }, { "epoch": 0.05488176705738871, "grad_norm": 14.083525657653809, "learning_rate": 9.991922787322883e-05, "loss": 1.4012, "step": 810 }, { "epoch": 0.054949522325360795, "grad_norm": 14.298514366149902, "learning_rate": 9.991785885413101e-05, "loss": 1.4047, "step": 811 }, { "epoch": 0.05501727759333288, "grad_norm": 13.391324996948242, "learning_rate": 9.99164898350332e-05, "loss": 1.2292, "step": 812 }, { "epoch": 0.055085032861304964, "grad_norm": 14.0011625289917, "learning_rate": 9.991512081593539e-05, "loss": 1.3194, "step": 813 }, { "epoch": 0.05515278812927705, "grad_norm": 14.726574897766113, "learning_rate": 9.991375179683757e-05, "loss": 1.3245, "step": 814 }, { "epoch": 0.05522054339724913, "grad_norm": 14.615242004394531, "learning_rate": 9.991238277773975e-05, "loss": 1.3704, "step": 815 }, { "epoch": 0.05528829866522122, "grad_norm": 11.06546688079834, "learning_rate": 9.991101375864193e-05, "loss": 1.0841, "step": 816 }, { "epoch": 0.0553560539331933, "grad_norm": 13.768633842468262, "learning_rate": 9.990964473954411e-05, "loss": 1.2632, "step": 817 }, { "epoch": 0.05542380920116539, "grad_norm": 15.054973602294922, "learning_rate": 9.99082757204463e-05, "loss": 1.2931, "step": 818 }, { "epoch": 0.05549156446913747, "grad_norm": 15.27096176147461, "learning_rate": 9.990690670134848e-05, "loss": 1.1789, "step": 819 }, { "epoch": 0.055559319737109564, "grad_norm": 13.228281021118164, "learning_rate": 9.990553768225067e-05, "loss": 1.4458, "step": 820 }, { "epoch": 0.05562707500508165, "grad_norm": 16.016782760620117, "learning_rate": 9.990416866315286e-05, "loss": 1.5121, "step": 821 }, { "epoch": 0.05569483027305373, "grad_norm": 14.15912914276123, "learning_rate": 9.990279964405504e-05, "loss": 1.3702, "step": 822 }, { "epoch": 0.05576258554102582, "grad_norm": 14.61017894744873, "learning_rate": 9.990143062495722e-05, "loss": 1.2774, "step": 823 }, { "epoch": 0.0558303408089979, "grad_norm": 13.241140365600586, "learning_rate": 9.990006160585941e-05, "loss": 1.3857, "step": 824 }, { "epoch": 0.05589809607696999, "grad_norm": 11.815064430236816, "learning_rate": 9.98986925867616e-05, "loss": 1.1053, "step": 825 }, { "epoch": 0.05596585134494207, "grad_norm": 13.179222106933594, "learning_rate": 9.989732356766377e-05, "loss": 1.4686, "step": 826 }, { "epoch": 0.05603360661291416, "grad_norm": 11.888179779052734, "learning_rate": 9.989595454856597e-05, "loss": 1.3191, "step": 827 }, { "epoch": 0.05610136188088624, "grad_norm": 15.43813419342041, "learning_rate": 9.989458552946815e-05, "loss": 0.9732, "step": 828 }, { "epoch": 0.056169117148858326, "grad_norm": 20.0246524810791, "learning_rate": 9.989321651037033e-05, "loss": 1.5398, "step": 829 }, { "epoch": 0.05623687241683041, "grad_norm": 14.247052192687988, "learning_rate": 9.989184749127251e-05, "loss": 1.5093, "step": 830 }, { "epoch": 0.056304627684802495, "grad_norm": 15.63775634765625, "learning_rate": 9.989047847217469e-05, "loss": 1.3488, "step": 831 }, { "epoch": 0.05637238295277458, "grad_norm": 10.582650184631348, "learning_rate": 9.988910945307688e-05, "loss": 1.3888, "step": 832 }, { "epoch": 0.056440138220746665, "grad_norm": 12.344864845275879, "learning_rate": 9.988774043397906e-05, "loss": 1.2876, "step": 833 }, { "epoch": 0.05650789348871875, "grad_norm": 13.95814323425293, "learning_rate": 9.988637141488124e-05, "loss": 1.2648, "step": 834 }, { "epoch": 0.056575648756690834, "grad_norm": 9.740105628967285, "learning_rate": 9.988500239578342e-05, "loss": 0.9777, "step": 835 }, { "epoch": 0.05664340402466292, "grad_norm": 16.16588020324707, "learning_rate": 9.988363337668562e-05, "loss": 1.5859, "step": 836 }, { "epoch": 0.056711159292635004, "grad_norm": 14.394731521606445, "learning_rate": 9.98822643575878e-05, "loss": 1.3395, "step": 837 }, { "epoch": 0.05677891456060709, "grad_norm": 11.256571769714355, "learning_rate": 9.988089533848998e-05, "loss": 1.0528, "step": 838 }, { "epoch": 0.05684666982857917, "grad_norm": 17.595510482788086, "learning_rate": 9.987952631939216e-05, "loss": 1.2678, "step": 839 }, { "epoch": 0.05691442509655126, "grad_norm": 14.132645606994629, "learning_rate": 9.987815730029434e-05, "loss": 1.4129, "step": 840 }, { "epoch": 0.05698218036452334, "grad_norm": 14.438119888305664, "learning_rate": 9.987678828119653e-05, "loss": 1.4141, "step": 841 }, { "epoch": 0.05704993563249543, "grad_norm": 15.983094215393066, "learning_rate": 9.987541926209871e-05, "loss": 1.1959, "step": 842 }, { "epoch": 0.05711769090046751, "grad_norm": 17.300403594970703, "learning_rate": 9.98740502430009e-05, "loss": 1.1613, "step": 843 }, { "epoch": 0.057185446168439596, "grad_norm": 15.150660514831543, "learning_rate": 9.987268122390307e-05, "loss": 1.2253, "step": 844 }, { "epoch": 0.05725320143641168, "grad_norm": 14.234586715698242, "learning_rate": 9.987131220480525e-05, "loss": 1.2517, "step": 845 }, { "epoch": 0.057320956704383766, "grad_norm": 18.31337547302246, "learning_rate": 9.986994318570745e-05, "loss": 1.7506, "step": 846 }, { "epoch": 0.05738871197235585, "grad_norm": 14.818669319152832, "learning_rate": 9.986857416660963e-05, "loss": 1.358, "step": 847 }, { "epoch": 0.057456467240327935, "grad_norm": 14.640913009643555, "learning_rate": 9.986720514751181e-05, "loss": 1.4481, "step": 848 }, { "epoch": 0.05752422250830002, "grad_norm": 12.277986526489258, "learning_rate": 9.986583612841399e-05, "loss": 1.3428, "step": 849 }, { "epoch": 0.057591977776272105, "grad_norm": 13.817851066589355, "learning_rate": 9.986446710931618e-05, "loss": 1.3514, "step": 850 }, { "epoch": 0.05765973304424419, "grad_norm": 13.706515312194824, "learning_rate": 9.986309809021836e-05, "loss": 1.4295, "step": 851 }, { "epoch": 0.057727488312216274, "grad_norm": 16.649917602539062, "learning_rate": 9.986172907112054e-05, "loss": 1.3895, "step": 852 }, { "epoch": 0.05779524358018836, "grad_norm": 13.659167289733887, "learning_rate": 9.986036005202272e-05, "loss": 1.1545, "step": 853 }, { "epoch": 0.057862998848160443, "grad_norm": 11.264912605285645, "learning_rate": 9.98589910329249e-05, "loss": 1.0766, "step": 854 }, { "epoch": 0.05793075411613253, "grad_norm": 12.930856704711914, "learning_rate": 9.98576220138271e-05, "loss": 1.1665, "step": 855 }, { "epoch": 0.05799850938410461, "grad_norm": 15.999971389770508, "learning_rate": 9.985625299472928e-05, "loss": 1.4595, "step": 856 }, { "epoch": 0.0580662646520767, "grad_norm": 14.566671371459961, "learning_rate": 9.985488397563146e-05, "loss": 1.283, "step": 857 }, { "epoch": 0.05813401992004878, "grad_norm": 16.106964111328125, "learning_rate": 9.985351495653364e-05, "loss": 1.2842, "step": 858 }, { "epoch": 0.05820177518802087, "grad_norm": 15.47492790222168, "learning_rate": 9.985214593743583e-05, "loss": 1.3949, "step": 859 }, { "epoch": 0.05826953045599295, "grad_norm": 10.692886352539062, "learning_rate": 9.985077691833801e-05, "loss": 1.2896, "step": 860 }, { "epoch": 0.058337285723965036, "grad_norm": 14.13198184967041, "learning_rate": 9.98494078992402e-05, "loss": 1.1566, "step": 861 }, { "epoch": 0.05840504099193712, "grad_norm": 14.455452919006348, "learning_rate": 9.984803888014237e-05, "loss": 1.1949, "step": 862 }, { "epoch": 0.058472796259909206, "grad_norm": 15.020733833312988, "learning_rate": 9.984666986104456e-05, "loss": 1.5202, "step": 863 }, { "epoch": 0.05854055152788129, "grad_norm": 21.089344024658203, "learning_rate": 9.984530084194675e-05, "loss": 1.5083, "step": 864 }, { "epoch": 0.058608306795853375, "grad_norm": 14.414257049560547, "learning_rate": 9.984393182284893e-05, "loss": 1.3977, "step": 865 }, { "epoch": 0.05867606206382546, "grad_norm": 15.667798042297363, "learning_rate": 9.984256280375111e-05, "loss": 1.4044, "step": 866 }, { "epoch": 0.058743817331797545, "grad_norm": 12.503005981445312, "learning_rate": 9.98411937846533e-05, "loss": 1.3369, "step": 867 }, { "epoch": 0.05881157259976963, "grad_norm": 13.27022933959961, "learning_rate": 9.983982476555548e-05, "loss": 1.4134, "step": 868 }, { "epoch": 0.058879327867741714, "grad_norm": 16.2034969329834, "learning_rate": 9.983845574645766e-05, "loss": 1.1196, "step": 869 }, { "epoch": 0.0589470831357138, "grad_norm": 9.868896484375, "learning_rate": 9.983708672735986e-05, "loss": 1.0248, "step": 870 }, { "epoch": 0.05901483840368588, "grad_norm": 15.588685989379883, "learning_rate": 9.983571770826204e-05, "loss": 1.5865, "step": 871 }, { "epoch": 0.05908259367165797, "grad_norm": 14.688246726989746, "learning_rate": 9.983434868916422e-05, "loss": 1.2959, "step": 872 }, { "epoch": 0.05915034893963006, "grad_norm": 14.252961158752441, "learning_rate": 9.983297967006641e-05, "loss": 1.3536, "step": 873 }, { "epoch": 0.059218104207602144, "grad_norm": 11.79800033569336, "learning_rate": 9.98316106509686e-05, "loss": 1.331, "step": 874 }, { "epoch": 0.05928585947557423, "grad_norm": 11.900074005126953, "learning_rate": 9.983024163187077e-05, "loss": 1.2241, "step": 875 }, { "epoch": 0.059353614743546314, "grad_norm": 10.696773529052734, "learning_rate": 9.982887261277295e-05, "loss": 1.4041, "step": 876 }, { "epoch": 0.0594213700115184, "grad_norm": 13.532305717468262, "learning_rate": 9.982750359367513e-05, "loss": 1.3621, "step": 877 }, { "epoch": 0.05948912527949048, "grad_norm": 14.107857704162598, "learning_rate": 9.982613457457733e-05, "loss": 1.6327, "step": 878 }, { "epoch": 0.05955688054746257, "grad_norm": 11.584097862243652, "learning_rate": 9.982476555547951e-05, "loss": 0.9606, "step": 879 }, { "epoch": 0.05962463581543465, "grad_norm": 14.240161895751953, "learning_rate": 9.982339653638169e-05, "loss": 1.263, "step": 880 }, { "epoch": 0.05969239108340674, "grad_norm": 14.461871147155762, "learning_rate": 9.982202751728387e-05, "loss": 1.4201, "step": 881 }, { "epoch": 0.05976014635137882, "grad_norm": 14.072705268859863, "learning_rate": 9.982065849818606e-05, "loss": 1.5977, "step": 882 }, { "epoch": 0.05982790161935091, "grad_norm": 14.928994178771973, "learning_rate": 9.981928947908824e-05, "loss": 1.3346, "step": 883 }, { "epoch": 0.05989565688732299, "grad_norm": 14.898951530456543, "learning_rate": 9.981792045999042e-05, "loss": 1.312, "step": 884 }, { "epoch": 0.059963412155295076, "grad_norm": 13.089646339416504, "learning_rate": 9.98165514408926e-05, "loss": 1.0833, "step": 885 }, { "epoch": 0.06003116742326716, "grad_norm": 15.768043518066406, "learning_rate": 9.981518242179478e-05, "loss": 1.1246, "step": 886 }, { "epoch": 0.060098922691239245, "grad_norm": 11.8709135055542, "learning_rate": 9.981381340269698e-05, "loss": 1.1822, "step": 887 }, { "epoch": 0.06016667795921133, "grad_norm": 15.698454856872559, "learning_rate": 9.981244438359916e-05, "loss": 1.3733, "step": 888 }, { "epoch": 0.060234433227183415, "grad_norm": 14.827208518981934, "learning_rate": 9.981107536450134e-05, "loss": 1.3121, "step": 889 }, { "epoch": 0.0603021884951555, "grad_norm": 12.522045135498047, "learning_rate": 9.980970634540352e-05, "loss": 1.5584, "step": 890 }, { "epoch": 0.060369943763127584, "grad_norm": 14.417738914489746, "learning_rate": 9.980833732630571e-05, "loss": 1.2389, "step": 891 }, { "epoch": 0.06043769903109967, "grad_norm": 14.761930465698242, "learning_rate": 9.98069683072079e-05, "loss": 1.5007, "step": 892 }, { "epoch": 0.060505454299071754, "grad_norm": 15.882668495178223, "learning_rate": 9.980559928811007e-05, "loss": 1.2651, "step": 893 }, { "epoch": 0.06057320956704384, "grad_norm": 13.605412483215332, "learning_rate": 9.980423026901225e-05, "loss": 1.5186, "step": 894 }, { "epoch": 0.06064096483501592, "grad_norm": 10.654335021972656, "learning_rate": 9.980286124991443e-05, "loss": 1.1177, "step": 895 }, { "epoch": 0.06070872010298801, "grad_norm": 12.37457275390625, "learning_rate": 9.980149223081663e-05, "loss": 1.2362, "step": 896 }, { "epoch": 0.06077647537096009, "grad_norm": 12.591222763061523, "learning_rate": 9.980012321171881e-05, "loss": 1.2509, "step": 897 }, { "epoch": 0.06084423063893218, "grad_norm": 14.337310791015625, "learning_rate": 9.979875419262099e-05, "loss": 1.2987, "step": 898 }, { "epoch": 0.06091198590690426, "grad_norm": 15.496018409729004, "learning_rate": 9.979738517352317e-05, "loss": 1.5307, "step": 899 }, { "epoch": 0.06097974117487635, "grad_norm": 13.730890274047852, "learning_rate": 9.979601615442535e-05, "loss": 1.2975, "step": 900 }, { "epoch": 0.06104749644284843, "grad_norm": 12.314823150634766, "learning_rate": 9.979464713532754e-05, "loss": 1.1916, "step": 901 }, { "epoch": 0.061115251710820516, "grad_norm": 13.761808395385742, "learning_rate": 9.979327811622972e-05, "loss": 1.3037, "step": 902 }, { "epoch": 0.0611830069787926, "grad_norm": 13.308722496032715, "learning_rate": 9.97919090971319e-05, "loss": 1.0598, "step": 903 }, { "epoch": 0.061250762246764685, "grad_norm": 13.121098518371582, "learning_rate": 9.979054007803408e-05, "loss": 1.3378, "step": 904 }, { "epoch": 0.06131851751473677, "grad_norm": 16.975666046142578, "learning_rate": 9.978917105893628e-05, "loss": 1.2826, "step": 905 }, { "epoch": 0.061386272782708855, "grad_norm": 14.529984474182129, "learning_rate": 9.978780203983846e-05, "loss": 1.3467, "step": 906 }, { "epoch": 0.06145402805068094, "grad_norm": 11.081110000610352, "learning_rate": 9.978643302074064e-05, "loss": 1.1906, "step": 907 }, { "epoch": 0.061521783318653024, "grad_norm": 12.871200561523438, "learning_rate": 9.978506400164282e-05, "loss": 1.1057, "step": 908 }, { "epoch": 0.06158953858662511, "grad_norm": 13.982168197631836, "learning_rate": 9.9783694982545e-05, "loss": 1.3824, "step": 909 }, { "epoch": 0.061657293854597194, "grad_norm": 13.076074600219727, "learning_rate": 9.97823259634472e-05, "loss": 1.2985, "step": 910 }, { "epoch": 0.06172504912256928, "grad_norm": 11.015650749206543, "learning_rate": 9.978095694434937e-05, "loss": 1.2406, "step": 911 }, { "epoch": 0.06179280439054136, "grad_norm": 13.6082763671875, "learning_rate": 9.977958792525155e-05, "loss": 1.3873, "step": 912 }, { "epoch": 0.06186055965851345, "grad_norm": 15.930809020996094, "learning_rate": 9.977821890615375e-05, "loss": 1.3542, "step": 913 }, { "epoch": 0.06192831492648553, "grad_norm": 10.710271835327148, "learning_rate": 9.977684988705593e-05, "loss": 1.1571, "step": 914 }, { "epoch": 0.06199607019445762, "grad_norm": 11.110217094421387, "learning_rate": 9.977548086795811e-05, "loss": 1.3255, "step": 915 }, { "epoch": 0.0620638254624297, "grad_norm": 11.451903343200684, "learning_rate": 9.97741118488603e-05, "loss": 1.3681, "step": 916 }, { "epoch": 0.062131580730401786, "grad_norm": 10.884252548217773, "learning_rate": 9.977274282976248e-05, "loss": 1.1988, "step": 917 }, { "epoch": 0.06219933599837387, "grad_norm": 11.031237602233887, "learning_rate": 9.977137381066466e-05, "loss": 1.26, "step": 918 }, { "epoch": 0.062267091266345956, "grad_norm": 11.585648536682129, "learning_rate": 9.977000479156686e-05, "loss": 1.3099, "step": 919 }, { "epoch": 0.06233484653431804, "grad_norm": 10.867992401123047, "learning_rate": 9.976863577246904e-05, "loss": 1.2624, "step": 920 }, { "epoch": 0.062402601802290125, "grad_norm": 14.552916526794434, "learning_rate": 9.976726675337122e-05, "loss": 1.2244, "step": 921 }, { "epoch": 0.06247035707026221, "grad_norm": 12.101760864257812, "learning_rate": 9.97658977342734e-05, "loss": 1.1925, "step": 922 }, { "epoch": 0.0625381123382343, "grad_norm": 14.113842010498047, "learning_rate": 9.976452871517559e-05, "loss": 1.3395, "step": 923 }, { "epoch": 0.06260586760620639, "grad_norm": 17.214614868164062, "learning_rate": 9.976315969607777e-05, "loss": 1.3943, "step": 924 }, { "epoch": 0.06267362287417846, "grad_norm": 13.43308162689209, "learning_rate": 9.976179067697995e-05, "loss": 1.1596, "step": 925 }, { "epoch": 0.06274137814215056, "grad_norm": 13.806952476501465, "learning_rate": 9.976042165788213e-05, "loss": 1.0354, "step": 926 }, { "epoch": 0.06280913341012263, "grad_norm": 15.638693809509277, "learning_rate": 9.975905263878431e-05, "loss": 1.3807, "step": 927 }, { "epoch": 0.06287688867809473, "grad_norm": 14.337742805480957, "learning_rate": 9.975768361968651e-05, "loss": 1.537, "step": 928 }, { "epoch": 0.0629446439460668, "grad_norm": 14.540297508239746, "learning_rate": 9.975631460058869e-05, "loss": 1.2522, "step": 929 }, { "epoch": 0.0630123992140389, "grad_norm": 15.991955757141113, "learning_rate": 9.975494558149087e-05, "loss": 1.3855, "step": 930 }, { "epoch": 0.06308015448201097, "grad_norm": 13.957479476928711, "learning_rate": 9.975357656239305e-05, "loss": 1.43, "step": 931 }, { "epoch": 0.06314790974998306, "grad_norm": 16.805377960205078, "learning_rate": 9.975220754329523e-05, "loss": 1.383, "step": 932 }, { "epoch": 0.06321566501795514, "grad_norm": 12.41854476928711, "learning_rate": 9.975083852419742e-05, "loss": 1.068, "step": 933 }, { "epoch": 0.06328342028592723, "grad_norm": 15.929006576538086, "learning_rate": 9.97494695050996e-05, "loss": 1.212, "step": 934 }, { "epoch": 0.06335117555389931, "grad_norm": 13.205544471740723, "learning_rate": 9.974810048600178e-05, "loss": 1.3682, "step": 935 }, { "epoch": 0.0634189308218714, "grad_norm": 12.105626106262207, "learning_rate": 9.974673146690396e-05, "loss": 1.3821, "step": 936 }, { "epoch": 0.06348668608984348, "grad_norm": 13.776711463928223, "learning_rate": 9.974536244780616e-05, "loss": 1.1073, "step": 937 }, { "epoch": 0.06355444135781557, "grad_norm": 12.227380752563477, "learning_rate": 9.974399342870834e-05, "loss": 1.2026, "step": 938 }, { "epoch": 0.06362219662578765, "grad_norm": 12.723440170288086, "learning_rate": 9.974262440961052e-05, "loss": 1.1325, "step": 939 }, { "epoch": 0.06368995189375974, "grad_norm": 13.943262100219727, "learning_rate": 9.97412553905127e-05, "loss": 1.1878, "step": 940 }, { "epoch": 0.06375770716173182, "grad_norm": 12.644627571105957, "learning_rate": 9.973988637141488e-05, "loss": 1.3097, "step": 941 }, { "epoch": 0.06382546242970391, "grad_norm": 12.108241081237793, "learning_rate": 9.973851735231707e-05, "loss": 1.1686, "step": 942 }, { "epoch": 0.06389321769767599, "grad_norm": 14.375092506408691, "learning_rate": 9.973714833321925e-05, "loss": 1.2721, "step": 943 }, { "epoch": 0.06396097296564808, "grad_norm": 13.439800262451172, "learning_rate": 9.973577931412143e-05, "loss": 1.3898, "step": 944 }, { "epoch": 0.06402872823362016, "grad_norm": 13.717879295349121, "learning_rate": 9.973441029502361e-05, "loss": 1.3823, "step": 945 }, { "epoch": 0.06409648350159225, "grad_norm": 12.745361328125, "learning_rate": 9.973304127592581e-05, "loss": 1.2396, "step": 946 }, { "epoch": 0.06416423876956433, "grad_norm": 11.784343719482422, "learning_rate": 9.973167225682799e-05, "loss": 1.214, "step": 947 }, { "epoch": 0.06423199403753642, "grad_norm": 14.205467224121094, "learning_rate": 9.973030323773017e-05, "loss": 1.2803, "step": 948 }, { "epoch": 0.0642997493055085, "grad_norm": 13.257532119750977, "learning_rate": 9.972893421863235e-05, "loss": 1.378, "step": 949 }, { "epoch": 0.06436750457348059, "grad_norm": 15.153338432312012, "learning_rate": 9.972756519953453e-05, "loss": 1.2854, "step": 950 }, { "epoch": 0.06443525984145267, "grad_norm": 16.765771865844727, "learning_rate": 9.972619618043672e-05, "loss": 1.3016, "step": 951 }, { "epoch": 0.06450301510942476, "grad_norm": 14.636106491088867, "learning_rate": 9.97248271613389e-05, "loss": 1.3803, "step": 952 }, { "epoch": 0.06457077037739685, "grad_norm": 13.87410831451416, "learning_rate": 9.972345814224108e-05, "loss": 1.4126, "step": 953 }, { "epoch": 0.06463852564536893, "grad_norm": 14.328899383544922, "learning_rate": 9.972208912314326e-05, "loss": 1.061, "step": 954 }, { "epoch": 0.06470628091334102, "grad_norm": 12.485203742980957, "learning_rate": 9.972072010404544e-05, "loss": 1.0985, "step": 955 }, { "epoch": 0.0647740361813131, "grad_norm": 13.77907943725586, "learning_rate": 9.971935108494764e-05, "loss": 1.3534, "step": 956 }, { "epoch": 0.06484179144928519, "grad_norm": 10.579590797424316, "learning_rate": 9.971798206584982e-05, "loss": 1.166, "step": 957 }, { "epoch": 0.06490954671725727, "grad_norm": 14.690185546875, "learning_rate": 9.9716613046752e-05, "loss": 1.3666, "step": 958 }, { "epoch": 0.06497730198522936, "grad_norm": 12.904786109924316, "learning_rate": 9.97152440276542e-05, "loss": 1.14, "step": 959 }, { "epoch": 0.06504505725320144, "grad_norm": 12.126219749450684, "learning_rate": 9.971387500855637e-05, "loss": 1.4157, "step": 960 }, { "epoch": 0.06511281252117353, "grad_norm": 13.747931480407715, "learning_rate": 9.971250598945855e-05, "loss": 1.4557, "step": 961 }, { "epoch": 0.0651805677891456, "grad_norm": 13.232327461242676, "learning_rate": 9.971113697036075e-05, "loss": 1.3471, "step": 962 }, { "epoch": 0.0652483230571177, "grad_norm": 14.886791229248047, "learning_rate": 9.970976795126293e-05, "loss": 1.1866, "step": 963 }, { "epoch": 0.06531607832508977, "grad_norm": 11.747659683227539, "learning_rate": 9.970839893216511e-05, "loss": 1.0443, "step": 964 }, { "epoch": 0.06538383359306187, "grad_norm": 11.181273460388184, "learning_rate": 9.97070299130673e-05, "loss": 1.1391, "step": 965 }, { "epoch": 0.06545158886103394, "grad_norm": 11.9672269821167, "learning_rate": 9.970566089396948e-05, "loss": 1.2847, "step": 966 }, { "epoch": 0.06551934412900604, "grad_norm": 15.825364112854004, "learning_rate": 9.970429187487166e-05, "loss": 1.3301, "step": 967 }, { "epoch": 0.06558709939697811, "grad_norm": 12.26963996887207, "learning_rate": 9.970292285577384e-05, "loss": 1.2524, "step": 968 }, { "epoch": 0.0656548546649502, "grad_norm": 11.440977096557617, "learning_rate": 9.970155383667604e-05, "loss": 1.229, "step": 969 }, { "epoch": 0.06572260993292228, "grad_norm": 10.704546928405762, "learning_rate": 9.970018481757822e-05, "loss": 0.9936, "step": 970 }, { "epoch": 0.06579036520089437, "grad_norm": 13.20880126953125, "learning_rate": 9.96988157984804e-05, "loss": 1.2353, "step": 971 }, { "epoch": 0.06585812046886645, "grad_norm": 13.101622581481934, "learning_rate": 9.969744677938258e-05, "loss": 1.2559, "step": 972 }, { "epoch": 0.06592587573683854, "grad_norm": 11.725826263427734, "learning_rate": 9.969607776028476e-05, "loss": 1.2051, "step": 973 }, { "epoch": 0.06599363100481062, "grad_norm": 11.890633583068848, "learning_rate": 9.969470874118695e-05, "loss": 1.1607, "step": 974 }, { "epoch": 0.06606138627278271, "grad_norm": 11.066970825195312, "learning_rate": 9.969333972208913e-05, "loss": 1.1031, "step": 975 }, { "epoch": 0.06612914154075479, "grad_norm": 12.26187515258789, "learning_rate": 9.969197070299131e-05, "loss": 1.309, "step": 976 }, { "epoch": 0.06619689680872688, "grad_norm": 13.490363121032715, "learning_rate": 9.96906016838935e-05, "loss": 1.3062, "step": 977 }, { "epoch": 0.06626465207669896, "grad_norm": 12.306289672851562, "learning_rate": 9.968923266479567e-05, "loss": 0.9784, "step": 978 }, { "epoch": 0.06633240734467105, "grad_norm": 10.699983596801758, "learning_rate": 9.968786364569787e-05, "loss": 1.0547, "step": 979 }, { "epoch": 0.06640016261264313, "grad_norm": 12.298179626464844, "learning_rate": 9.968649462660005e-05, "loss": 1.2496, "step": 980 }, { "epoch": 0.06646791788061522, "grad_norm": 15.239167213439941, "learning_rate": 9.968512560750223e-05, "loss": 1.2545, "step": 981 }, { "epoch": 0.0665356731485873, "grad_norm": 11.61802864074707, "learning_rate": 9.968375658840441e-05, "loss": 1.1668, "step": 982 }, { "epoch": 0.06660342841655939, "grad_norm": 12.804032325744629, "learning_rate": 9.96823875693066e-05, "loss": 1.3203, "step": 983 }, { "epoch": 0.06667118368453147, "grad_norm": 14.511723518371582, "learning_rate": 9.968101855020878e-05, "loss": 1.7519, "step": 984 }, { "epoch": 0.06673893895250356, "grad_norm": 12.442008018493652, "learning_rate": 9.967964953111096e-05, "loss": 1.3758, "step": 985 }, { "epoch": 0.06680669422047564, "grad_norm": 14.486754417419434, "learning_rate": 9.967828051201314e-05, "loss": 1.1924, "step": 986 }, { "epoch": 0.06687444948844773, "grad_norm": 13.529693603515625, "learning_rate": 9.967691149291532e-05, "loss": 1.2698, "step": 987 }, { "epoch": 0.06694220475641981, "grad_norm": 12.980225563049316, "learning_rate": 9.967554247381752e-05, "loss": 1.1896, "step": 988 }, { "epoch": 0.0670099600243919, "grad_norm": 15.495257377624512, "learning_rate": 9.96741734547197e-05, "loss": 1.286, "step": 989 }, { "epoch": 0.06707771529236398, "grad_norm": 12.67573070526123, "learning_rate": 9.967280443562188e-05, "loss": 1.1751, "step": 990 }, { "epoch": 0.06714547056033607, "grad_norm": 11.141845703125, "learning_rate": 9.967143541652406e-05, "loss": 1.5109, "step": 991 }, { "epoch": 0.06721322582830815, "grad_norm": 11.975769996643066, "learning_rate": 9.967006639742625e-05, "loss": 1.1579, "step": 992 }, { "epoch": 0.06728098109628024, "grad_norm": 13.872209548950195, "learning_rate": 9.966869737832843e-05, "loss": 1.1852, "step": 993 }, { "epoch": 0.06734873636425232, "grad_norm": 11.52573299407959, "learning_rate": 9.966732835923061e-05, "loss": 1.2319, "step": 994 }, { "epoch": 0.0674164916322244, "grad_norm": 12.90494155883789, "learning_rate": 9.96659593401328e-05, "loss": 1.2415, "step": 995 }, { "epoch": 0.06748424690019648, "grad_norm": 14.210317611694336, "learning_rate": 9.966459032103497e-05, "loss": 1.2901, "step": 996 }, { "epoch": 0.06755200216816858, "grad_norm": 12.600135803222656, "learning_rate": 9.966322130193717e-05, "loss": 1.4516, "step": 997 }, { "epoch": 0.06761975743614065, "grad_norm": 14.462118148803711, "learning_rate": 9.966185228283935e-05, "loss": 1.2205, "step": 998 }, { "epoch": 0.06768751270411275, "grad_norm": 12.870843887329102, "learning_rate": 9.966048326374153e-05, "loss": 1.179, "step": 999 }, { "epoch": 0.06775526797208482, "grad_norm": 16.4424991607666, "learning_rate": 9.965911424464371e-05, "loss": 1.5714, "step": 1000 }, { "epoch": 0.06782302324005691, "grad_norm": 12.902230262756348, "learning_rate": 9.96577452255459e-05, "loss": 1.2803, "step": 1001 }, { "epoch": 0.06789077850802899, "grad_norm": 11.469466209411621, "learning_rate": 9.965637620644808e-05, "loss": 1.0771, "step": 1002 }, { "epoch": 0.06795853377600108, "grad_norm": 13.96650505065918, "learning_rate": 9.965500718735026e-05, "loss": 1.2427, "step": 1003 }, { "epoch": 0.06802628904397316, "grad_norm": 11.55516242980957, "learning_rate": 9.965363816825244e-05, "loss": 1.0396, "step": 1004 }, { "epoch": 0.06809404431194525, "grad_norm": 13.34827709197998, "learning_rate": 9.965226914915462e-05, "loss": 1.181, "step": 1005 }, { "epoch": 0.06816179957991735, "grad_norm": 11.243910789489746, "learning_rate": 9.965090013005682e-05, "loss": 1.2875, "step": 1006 }, { "epoch": 0.06822955484788942, "grad_norm": 14.152894020080566, "learning_rate": 9.9649531110959e-05, "loss": 1.3125, "step": 1007 }, { "epoch": 0.06829731011586151, "grad_norm": 13.010010719299316, "learning_rate": 9.964816209186118e-05, "loss": 1.3213, "step": 1008 }, { "epoch": 0.06836506538383359, "grad_norm": 15.990034103393555, "learning_rate": 9.964679307276337e-05, "loss": 1.1878, "step": 1009 }, { "epoch": 0.06843282065180568, "grad_norm": 12.943589210510254, "learning_rate": 9.964542405366555e-05, "loss": 1.2443, "step": 1010 }, { "epoch": 0.06850057591977776, "grad_norm": 12.108896255493164, "learning_rate": 9.964405503456775e-05, "loss": 1.1719, "step": 1011 }, { "epoch": 0.06856833118774985, "grad_norm": 12.097951889038086, "learning_rate": 9.964268601546993e-05, "loss": 1.0653, "step": 1012 }, { "epoch": 0.06863608645572193, "grad_norm": 14.222228050231934, "learning_rate": 9.964131699637211e-05, "loss": 1.3379, "step": 1013 }, { "epoch": 0.06870384172369402, "grad_norm": 12.636894226074219, "learning_rate": 9.963994797727429e-05, "loss": 1.5559, "step": 1014 }, { "epoch": 0.0687715969916661, "grad_norm": 15.458481788635254, "learning_rate": 9.963857895817648e-05, "loss": 1.2542, "step": 1015 }, { "epoch": 0.06883935225963819, "grad_norm": 11.246847152709961, "learning_rate": 9.963720993907866e-05, "loss": 1.1944, "step": 1016 }, { "epoch": 0.06890710752761027, "grad_norm": 11.699065208435059, "learning_rate": 9.963584091998084e-05, "loss": 1.0739, "step": 1017 }, { "epoch": 0.06897486279558236, "grad_norm": 12.259678840637207, "learning_rate": 9.963447190088302e-05, "loss": 1.1365, "step": 1018 }, { "epoch": 0.06904261806355444, "grad_norm": 13.594696998596191, "learning_rate": 9.96331028817852e-05, "loss": 1.1418, "step": 1019 }, { "epoch": 0.06911037333152653, "grad_norm": 12.90888786315918, "learning_rate": 9.96317338626874e-05, "loss": 1.1987, "step": 1020 }, { "epoch": 0.06917812859949861, "grad_norm": 13.04245662689209, "learning_rate": 9.963036484358958e-05, "loss": 1.4288, "step": 1021 }, { "epoch": 0.0692458838674707, "grad_norm": 12.706077575683594, "learning_rate": 9.962899582449176e-05, "loss": 1.222, "step": 1022 }, { "epoch": 0.06931363913544278, "grad_norm": 14.205679893493652, "learning_rate": 9.962762680539394e-05, "loss": 1.3305, "step": 1023 }, { "epoch": 0.06938139440341487, "grad_norm": 17.09891128540039, "learning_rate": 9.962625778629613e-05, "loss": 1.47, "step": 1024 }, { "epoch": 0.06944914967138695, "grad_norm": 14.603500366210938, "learning_rate": 9.962488876719831e-05, "loss": 1.3887, "step": 1025 }, { "epoch": 0.06951690493935904, "grad_norm": 10.820066452026367, "learning_rate": 9.962351974810049e-05, "loss": 1.1543, "step": 1026 }, { "epoch": 0.06958466020733112, "grad_norm": 10.99889850616455, "learning_rate": 9.962215072900267e-05, "loss": 1.2908, "step": 1027 }, { "epoch": 0.06965241547530321, "grad_norm": 13.470711708068848, "learning_rate": 9.962078170990485e-05, "loss": 1.1422, "step": 1028 }, { "epoch": 0.06972017074327529, "grad_norm": 11.086441040039062, "learning_rate": 9.961941269080705e-05, "loss": 0.9835, "step": 1029 }, { "epoch": 0.06978792601124738, "grad_norm": 16.768535614013672, "learning_rate": 9.961804367170923e-05, "loss": 1.404, "step": 1030 }, { "epoch": 0.06985568127921946, "grad_norm": 15.851200103759766, "learning_rate": 9.961667465261141e-05, "loss": 1.4874, "step": 1031 }, { "epoch": 0.06992343654719155, "grad_norm": 11.995482444763184, "learning_rate": 9.961530563351359e-05, "loss": 1.1497, "step": 1032 }, { "epoch": 0.06999119181516363, "grad_norm": 13.591619491577148, "learning_rate": 9.961393661441577e-05, "loss": 1.4773, "step": 1033 }, { "epoch": 0.07005894708313572, "grad_norm": 16.878938674926758, "learning_rate": 9.961256759531796e-05, "loss": 1.2522, "step": 1034 }, { "epoch": 0.0701267023511078, "grad_norm": 11.901616096496582, "learning_rate": 9.961119857622014e-05, "loss": 1.2164, "step": 1035 }, { "epoch": 0.07019445761907989, "grad_norm": 14.935117721557617, "learning_rate": 9.960982955712232e-05, "loss": 1.1251, "step": 1036 }, { "epoch": 0.07026221288705196, "grad_norm": 12.380253791809082, "learning_rate": 9.96084605380245e-05, "loss": 1.3965, "step": 1037 }, { "epoch": 0.07032996815502406, "grad_norm": 11.645035743713379, "learning_rate": 9.96070915189267e-05, "loss": 1.266, "step": 1038 }, { "epoch": 0.07039772342299613, "grad_norm": 14.525420188903809, "learning_rate": 9.960572249982888e-05, "loss": 1.3991, "step": 1039 }, { "epoch": 0.07046547869096823, "grad_norm": 14.775094985961914, "learning_rate": 9.960435348073106e-05, "loss": 1.5958, "step": 1040 }, { "epoch": 0.0705332339589403, "grad_norm": 10.2192964553833, "learning_rate": 9.960298446163324e-05, "loss": 1.1793, "step": 1041 }, { "epoch": 0.0706009892269124, "grad_norm": 13.074480056762695, "learning_rate": 9.960161544253542e-05, "loss": 1.4243, "step": 1042 }, { "epoch": 0.07066874449488447, "grad_norm": 12.679484367370605, "learning_rate": 9.960024642343761e-05, "loss": 1.3398, "step": 1043 }, { "epoch": 0.07073649976285656, "grad_norm": 9.061332702636719, "learning_rate": 9.95988774043398e-05, "loss": 1.0036, "step": 1044 }, { "epoch": 0.07080425503082864, "grad_norm": 13.423661231994629, "learning_rate": 9.959750838524197e-05, "loss": 1.2767, "step": 1045 }, { "epoch": 0.07087201029880073, "grad_norm": 13.955148696899414, "learning_rate": 9.959613936614415e-05, "loss": 1.311, "step": 1046 }, { "epoch": 0.07093976556677281, "grad_norm": 12.746015548706055, "learning_rate": 9.959477034704635e-05, "loss": 1.42, "step": 1047 }, { "epoch": 0.0710075208347449, "grad_norm": 11.409982681274414, "learning_rate": 9.959340132794853e-05, "loss": 1.3051, "step": 1048 }, { "epoch": 0.07107527610271698, "grad_norm": 11.801681518554688, "learning_rate": 9.959203230885071e-05, "loss": 1.2315, "step": 1049 }, { "epoch": 0.07114303137068907, "grad_norm": 13.041158676147461, "learning_rate": 9.959066328975289e-05, "loss": 1.3487, "step": 1050 }, { "epoch": 0.07121078663866115, "grad_norm": 13.474900245666504, "learning_rate": 9.958929427065507e-05, "loss": 1.4542, "step": 1051 }, { "epoch": 0.07127854190663324, "grad_norm": 12.335237503051758, "learning_rate": 9.958792525155726e-05, "loss": 1.1841, "step": 1052 }, { "epoch": 0.07134629717460532, "grad_norm": 14.909475326538086, "learning_rate": 9.958655623245944e-05, "loss": 1.272, "step": 1053 }, { "epoch": 0.07141405244257741, "grad_norm": 13.449742317199707, "learning_rate": 9.958518721336162e-05, "loss": 1.4881, "step": 1054 }, { "epoch": 0.07148180771054949, "grad_norm": 12.2557954788208, "learning_rate": 9.958381819426382e-05, "loss": 1.2059, "step": 1055 }, { "epoch": 0.07154956297852158, "grad_norm": 13.71298885345459, "learning_rate": 9.9582449175166e-05, "loss": 1.2627, "step": 1056 }, { "epoch": 0.07161731824649366, "grad_norm": 15.4293212890625, "learning_rate": 9.958108015606818e-05, "loss": 1.4663, "step": 1057 }, { "epoch": 0.07168507351446575, "grad_norm": 13.665759086608887, "learning_rate": 9.957971113697037e-05, "loss": 1.4634, "step": 1058 }, { "epoch": 0.07175282878243784, "grad_norm": 13.064310073852539, "learning_rate": 9.957834211787255e-05, "loss": 1.2686, "step": 1059 }, { "epoch": 0.07182058405040992, "grad_norm": 15.21746826171875, "learning_rate": 9.957697309877473e-05, "loss": 1.4861, "step": 1060 }, { "epoch": 0.07188833931838201, "grad_norm": 12.499883651733398, "learning_rate": 9.957560407967693e-05, "loss": 1.5275, "step": 1061 }, { "epoch": 0.07195609458635409, "grad_norm": 9.355907440185547, "learning_rate": 9.957423506057911e-05, "loss": 1.1509, "step": 1062 }, { "epoch": 0.07202384985432618, "grad_norm": 10.428252220153809, "learning_rate": 9.957286604148129e-05, "loss": 1.0874, "step": 1063 }, { "epoch": 0.07209160512229826, "grad_norm": 11.351346015930176, "learning_rate": 9.957149702238347e-05, "loss": 1.0887, "step": 1064 }, { "epoch": 0.07215936039027035, "grad_norm": 15.271830558776855, "learning_rate": 9.957012800328565e-05, "loss": 1.2466, "step": 1065 }, { "epoch": 0.07222711565824243, "grad_norm": 11.172418594360352, "learning_rate": 9.956875898418784e-05, "loss": 1.0958, "step": 1066 }, { "epoch": 0.07229487092621452, "grad_norm": 14.221702575683594, "learning_rate": 9.956738996509002e-05, "loss": 1.2288, "step": 1067 }, { "epoch": 0.0723626261941866, "grad_norm": 12.167356491088867, "learning_rate": 9.95660209459922e-05, "loss": 1.1159, "step": 1068 }, { "epoch": 0.07243038146215869, "grad_norm": 12.607329368591309, "learning_rate": 9.956465192689438e-05, "loss": 1.2475, "step": 1069 }, { "epoch": 0.07249813673013077, "grad_norm": 10.59451675415039, "learning_rate": 9.956328290779658e-05, "loss": 1.1493, "step": 1070 }, { "epoch": 0.07256589199810286, "grad_norm": 11.190742492675781, "learning_rate": 9.956191388869876e-05, "loss": 1.4549, "step": 1071 }, { "epoch": 0.07263364726607494, "grad_norm": 11.225564956665039, "learning_rate": 9.956054486960094e-05, "loss": 1.0741, "step": 1072 }, { "epoch": 0.07270140253404703, "grad_norm": 13.648139953613281, "learning_rate": 9.955917585050312e-05, "loss": 1.2942, "step": 1073 }, { "epoch": 0.0727691578020191, "grad_norm": 14.606941223144531, "learning_rate": 9.95578068314053e-05, "loss": 1.253, "step": 1074 }, { "epoch": 0.0728369130699912, "grad_norm": 11.043729782104492, "learning_rate": 9.955643781230749e-05, "loss": 1.2279, "step": 1075 }, { "epoch": 0.07290466833796327, "grad_norm": 12.463634490966797, "learning_rate": 9.955506879320967e-05, "loss": 1.4399, "step": 1076 }, { "epoch": 0.07297242360593537, "grad_norm": 13.74101734161377, "learning_rate": 9.955369977411185e-05, "loss": 1.4323, "step": 1077 }, { "epoch": 0.07304017887390744, "grad_norm": 10.1694974899292, "learning_rate": 9.955233075501403e-05, "loss": 1.0113, "step": 1078 }, { "epoch": 0.07310793414187954, "grad_norm": 14.407991409301758, "learning_rate": 9.955096173591623e-05, "loss": 1.7102, "step": 1079 }, { "epoch": 0.07317568940985161, "grad_norm": 13.84760570526123, "learning_rate": 9.954959271681841e-05, "loss": 1.4491, "step": 1080 }, { "epoch": 0.0732434446778237, "grad_norm": 12.220841407775879, "learning_rate": 9.954822369772059e-05, "loss": 1.1826, "step": 1081 }, { "epoch": 0.07331119994579578, "grad_norm": 11.380377769470215, "learning_rate": 9.954685467862277e-05, "loss": 1.3537, "step": 1082 }, { "epoch": 0.07337895521376787, "grad_norm": 12.916484832763672, "learning_rate": 9.954548565952495e-05, "loss": 1.3406, "step": 1083 }, { "epoch": 0.07344671048173995, "grad_norm": 14.107590675354004, "learning_rate": 9.954411664042714e-05, "loss": 1.0531, "step": 1084 }, { "epoch": 0.07351446574971204, "grad_norm": 13.498798370361328, "learning_rate": 9.954274762132932e-05, "loss": 1.1546, "step": 1085 }, { "epoch": 0.07358222101768412, "grad_norm": 13.155747413635254, "learning_rate": 9.95413786022315e-05, "loss": 1.212, "step": 1086 }, { "epoch": 0.07364997628565621, "grad_norm": 14.135687828063965, "learning_rate": 9.954000958313368e-05, "loss": 1.2001, "step": 1087 }, { "epoch": 0.07371773155362829, "grad_norm": 13.123790740966797, "learning_rate": 9.953864056403586e-05, "loss": 1.2735, "step": 1088 }, { "epoch": 0.07378548682160038, "grad_norm": 14.591660499572754, "learning_rate": 9.953727154493806e-05, "loss": 1.2977, "step": 1089 }, { "epoch": 0.07385324208957246, "grad_norm": 13.30614948272705, "learning_rate": 9.953590252584024e-05, "loss": 1.2445, "step": 1090 }, { "epoch": 0.07392099735754455, "grad_norm": 14.542524337768555, "learning_rate": 9.953453350674242e-05, "loss": 1.2327, "step": 1091 }, { "epoch": 0.07398875262551663, "grad_norm": 12.926522254943848, "learning_rate": 9.95331644876446e-05, "loss": 1.2139, "step": 1092 }, { "epoch": 0.07405650789348872, "grad_norm": 11.622479438781738, "learning_rate": 9.953179546854679e-05, "loss": 1.2887, "step": 1093 }, { "epoch": 0.0741242631614608, "grad_norm": 14.87485122680664, "learning_rate": 9.953042644944897e-05, "loss": 1.3132, "step": 1094 }, { "epoch": 0.07419201842943289, "grad_norm": 13.096991539001465, "learning_rate": 9.952905743035115e-05, "loss": 1.0783, "step": 1095 }, { "epoch": 0.07425977369740497, "grad_norm": 14.676551818847656, "learning_rate": 9.952768841125333e-05, "loss": 1.3171, "step": 1096 }, { "epoch": 0.07432752896537706, "grad_norm": 10.399755477905273, "learning_rate": 9.952631939215551e-05, "loss": 1.0747, "step": 1097 }, { "epoch": 0.07439528423334914, "grad_norm": 11.052745819091797, "learning_rate": 9.952495037305771e-05, "loss": 1.1653, "step": 1098 }, { "epoch": 0.07446303950132123, "grad_norm": 13.014352798461914, "learning_rate": 9.952358135395989e-05, "loss": 1.3727, "step": 1099 }, { "epoch": 0.07453079476929331, "grad_norm": 13.202262878417969, "learning_rate": 9.952221233486207e-05, "loss": 1.5063, "step": 1100 }, { "epoch": 0.0745985500372654, "grad_norm": 14.272111892700195, "learning_rate": 9.952084331576426e-05, "loss": 1.0183, "step": 1101 }, { "epoch": 0.07466630530523748, "grad_norm": 14.23975658416748, "learning_rate": 9.951947429666644e-05, "loss": 1.3518, "step": 1102 }, { "epoch": 0.07473406057320957, "grad_norm": 12.293742179870605, "learning_rate": 9.951810527756862e-05, "loss": 1.4021, "step": 1103 }, { "epoch": 0.07480181584118165, "grad_norm": 16.026020050048828, "learning_rate": 9.951673625847082e-05, "loss": 1.3412, "step": 1104 }, { "epoch": 0.07486957110915374, "grad_norm": 11.120819091796875, "learning_rate": 9.9515367239373e-05, "loss": 1.331, "step": 1105 }, { "epoch": 0.07493732637712582, "grad_norm": 12.363526344299316, "learning_rate": 9.951399822027518e-05, "loss": 1.4284, "step": 1106 }, { "epoch": 0.0750050816450979, "grad_norm": 14.377492904663086, "learning_rate": 9.951262920117737e-05, "loss": 1.3686, "step": 1107 }, { "epoch": 0.07507283691306998, "grad_norm": 11.366288185119629, "learning_rate": 9.951126018207955e-05, "loss": 1.2329, "step": 1108 }, { "epoch": 0.07514059218104208, "grad_norm": 10.26131820678711, "learning_rate": 9.950989116298173e-05, "loss": 1.1091, "step": 1109 }, { "epoch": 0.07520834744901415, "grad_norm": 12.64631462097168, "learning_rate": 9.950852214388391e-05, "loss": 1.1231, "step": 1110 }, { "epoch": 0.07527610271698625, "grad_norm": 13.042781829833984, "learning_rate": 9.95071531247861e-05, "loss": 1.2338, "step": 1111 }, { "epoch": 0.07534385798495834, "grad_norm": 11.577115058898926, "learning_rate": 9.950578410568829e-05, "loss": 1.2143, "step": 1112 }, { "epoch": 0.07541161325293042, "grad_norm": 13.640811920166016, "learning_rate": 9.950441508659047e-05, "loss": 1.224, "step": 1113 }, { "epoch": 0.0754793685209025, "grad_norm": 10.271018981933594, "learning_rate": 9.950304606749265e-05, "loss": 1.0459, "step": 1114 }, { "epoch": 0.07554712378887458, "grad_norm": 12.053836822509766, "learning_rate": 9.950167704839483e-05, "loss": 1.2137, "step": 1115 }, { "epoch": 0.07561487905684668, "grad_norm": 10.022509574890137, "learning_rate": 9.950030802929702e-05, "loss": 1.092, "step": 1116 }, { "epoch": 0.07568263432481875, "grad_norm": 12.28339672088623, "learning_rate": 9.94989390101992e-05, "loss": 1.1616, "step": 1117 }, { "epoch": 0.07575038959279085, "grad_norm": 14.182686805725098, "learning_rate": 9.949756999110138e-05, "loss": 1.5135, "step": 1118 }, { "epoch": 0.07581814486076292, "grad_norm": 10.668661117553711, "learning_rate": 9.949620097200356e-05, "loss": 1.2412, "step": 1119 }, { "epoch": 0.07588590012873501, "grad_norm": 14.443583488464355, "learning_rate": 9.949483195290574e-05, "loss": 1.207, "step": 1120 }, { "epoch": 0.07595365539670709, "grad_norm": 12.418794631958008, "learning_rate": 9.949346293380794e-05, "loss": 1.1907, "step": 1121 }, { "epoch": 0.07602141066467918, "grad_norm": 12.429618835449219, "learning_rate": 9.949209391471012e-05, "loss": 1.4403, "step": 1122 }, { "epoch": 0.07608916593265126, "grad_norm": 9.524622917175293, "learning_rate": 9.94907248956123e-05, "loss": 1.139, "step": 1123 }, { "epoch": 0.07615692120062335, "grad_norm": 10.974812507629395, "learning_rate": 9.948935587651448e-05, "loss": 1.4055, "step": 1124 }, { "epoch": 0.07622467646859543, "grad_norm": 16.152681350708008, "learning_rate": 9.948798685741667e-05, "loss": 1.0874, "step": 1125 }, { "epoch": 0.07629243173656752, "grad_norm": 12.023541450500488, "learning_rate": 9.948661783831885e-05, "loss": 1.2378, "step": 1126 }, { "epoch": 0.0763601870045396, "grad_norm": 11.597234725952148, "learning_rate": 9.948524881922103e-05, "loss": 1.0104, "step": 1127 }, { "epoch": 0.07642794227251169, "grad_norm": 11.374302864074707, "learning_rate": 9.948387980012321e-05, "loss": 1.1099, "step": 1128 }, { "epoch": 0.07649569754048377, "grad_norm": 14.398423194885254, "learning_rate": 9.94825107810254e-05, "loss": 1.5533, "step": 1129 }, { "epoch": 0.07656345280845586, "grad_norm": 13.1026611328125, "learning_rate": 9.948114176192759e-05, "loss": 1.3948, "step": 1130 }, { "epoch": 0.07663120807642794, "grad_norm": 12.012560844421387, "learning_rate": 9.947977274282977e-05, "loss": 1.2156, "step": 1131 }, { "epoch": 0.07669896334440003, "grad_norm": 12.900229454040527, "learning_rate": 9.947840372373195e-05, "loss": 1.2658, "step": 1132 }, { "epoch": 0.07676671861237211, "grad_norm": 15.250492095947266, "learning_rate": 9.947703470463413e-05, "loss": 1.6102, "step": 1133 }, { "epoch": 0.0768344738803442, "grad_norm": 15.08134937286377, "learning_rate": 9.947566568553632e-05, "loss": 1.1947, "step": 1134 }, { "epoch": 0.07690222914831628, "grad_norm": 13.353601455688477, "learning_rate": 9.94742966664385e-05, "loss": 1.3211, "step": 1135 }, { "epoch": 0.07696998441628837, "grad_norm": 11.310175895690918, "learning_rate": 9.947292764734068e-05, "loss": 1.2223, "step": 1136 }, { "epoch": 0.07703773968426045, "grad_norm": 11.800848960876465, "learning_rate": 9.947155862824286e-05, "loss": 1.2131, "step": 1137 }, { "epoch": 0.07710549495223254, "grad_norm": 11.748014450073242, "learning_rate": 9.947018960914504e-05, "loss": 1.0734, "step": 1138 }, { "epoch": 0.07717325022020462, "grad_norm": 12.282258033752441, "learning_rate": 9.946882059004724e-05, "loss": 1.3257, "step": 1139 }, { "epoch": 0.07724100548817671, "grad_norm": 11.93818473815918, "learning_rate": 9.946745157094942e-05, "loss": 1.1452, "step": 1140 }, { "epoch": 0.07730876075614879, "grad_norm": 13.397029876708984, "learning_rate": 9.94660825518516e-05, "loss": 1.2625, "step": 1141 }, { "epoch": 0.07737651602412088, "grad_norm": 12.135769844055176, "learning_rate": 9.946471353275378e-05, "loss": 1.3624, "step": 1142 }, { "epoch": 0.07744427129209296, "grad_norm": 11.304028511047363, "learning_rate": 9.946334451365596e-05, "loss": 1.3019, "step": 1143 }, { "epoch": 0.07751202656006505, "grad_norm": 10.978137016296387, "learning_rate": 9.946197549455815e-05, "loss": 1.3137, "step": 1144 }, { "epoch": 0.07757978182803713, "grad_norm": 10.997323989868164, "learning_rate": 9.946060647546033e-05, "loss": 1.3573, "step": 1145 }, { "epoch": 0.07764753709600922, "grad_norm": 11.883647918701172, "learning_rate": 9.945923745636251e-05, "loss": 1.2587, "step": 1146 }, { "epoch": 0.0777152923639813, "grad_norm": 10.70753288269043, "learning_rate": 9.945786843726471e-05, "loss": 1.0866, "step": 1147 }, { "epoch": 0.07778304763195339, "grad_norm": 13.318743705749512, "learning_rate": 9.945649941816689e-05, "loss": 1.3356, "step": 1148 }, { "epoch": 0.07785080289992546, "grad_norm": 10.656171798706055, "learning_rate": 9.945513039906907e-05, "loss": 1.2155, "step": 1149 }, { "epoch": 0.07791855816789756, "grad_norm": 14.73982048034668, "learning_rate": 9.945376137997126e-05, "loss": 1.2276, "step": 1150 }, { "epoch": 0.07798631343586963, "grad_norm": 12.461714744567871, "learning_rate": 9.945239236087344e-05, "loss": 1.2999, "step": 1151 }, { "epoch": 0.07805406870384173, "grad_norm": 12.277376174926758, "learning_rate": 9.945102334177562e-05, "loss": 1.1131, "step": 1152 }, { "epoch": 0.0781218239718138, "grad_norm": 12.726540565490723, "learning_rate": 9.944965432267782e-05, "loss": 1.4181, "step": 1153 }, { "epoch": 0.0781895792397859, "grad_norm": 11.086180686950684, "learning_rate": 9.944828530358e-05, "loss": 1.3852, "step": 1154 }, { "epoch": 0.07825733450775797, "grad_norm": 12.80537223815918, "learning_rate": 9.944691628448218e-05, "loss": 1.1101, "step": 1155 }, { "epoch": 0.07832508977573006, "grad_norm": 11.960269927978516, "learning_rate": 9.944554726538436e-05, "loss": 1.2186, "step": 1156 }, { "epoch": 0.07839284504370214, "grad_norm": 11.732439041137695, "learning_rate": 9.944417824628655e-05, "loss": 0.9854, "step": 1157 }, { "epoch": 0.07846060031167423, "grad_norm": 12.578715324401855, "learning_rate": 9.944280922718873e-05, "loss": 1.0792, "step": 1158 }, { "epoch": 0.07852835557964631, "grad_norm": 13.779712677001953, "learning_rate": 9.944144020809091e-05, "loss": 1.2468, "step": 1159 }, { "epoch": 0.0785961108476184, "grad_norm": 13.95693588256836, "learning_rate": 9.944007118899309e-05, "loss": 1.234, "step": 1160 }, { "epoch": 0.07866386611559048, "grad_norm": 12.056897163391113, "learning_rate": 9.943870216989527e-05, "loss": 1.3034, "step": 1161 }, { "epoch": 0.07873162138356257, "grad_norm": 11.633442878723145, "learning_rate": 9.943733315079747e-05, "loss": 1.2744, "step": 1162 }, { "epoch": 0.07879937665153465, "grad_norm": 14.062381744384766, "learning_rate": 9.943596413169965e-05, "loss": 1.2535, "step": 1163 }, { "epoch": 0.07886713191950674, "grad_norm": 11.60498332977295, "learning_rate": 9.943459511260183e-05, "loss": 1.1578, "step": 1164 }, { "epoch": 0.07893488718747883, "grad_norm": 9.667806625366211, "learning_rate": 9.943322609350401e-05, "loss": 1.0804, "step": 1165 }, { "epoch": 0.07900264245545091, "grad_norm": 12.30827808380127, "learning_rate": 9.943185707440619e-05, "loss": 1.4167, "step": 1166 }, { "epoch": 0.079070397723423, "grad_norm": 10.196819305419922, "learning_rate": 9.943048805530838e-05, "loss": 1.0547, "step": 1167 }, { "epoch": 0.07913815299139508, "grad_norm": 10.029928207397461, "learning_rate": 9.942911903621056e-05, "loss": 1.1798, "step": 1168 }, { "epoch": 0.07920590825936717, "grad_norm": 11.782978057861328, "learning_rate": 9.942775001711274e-05, "loss": 1.1454, "step": 1169 }, { "epoch": 0.07927366352733925, "grad_norm": 12.25143814086914, "learning_rate": 9.942638099801492e-05, "loss": 1.2854, "step": 1170 }, { "epoch": 0.07934141879531134, "grad_norm": 8.47904109954834, "learning_rate": 9.942501197891712e-05, "loss": 0.9381, "step": 1171 }, { "epoch": 0.07940917406328342, "grad_norm": 13.698802947998047, "learning_rate": 9.94236429598193e-05, "loss": 1.6261, "step": 1172 }, { "epoch": 0.07947692933125551, "grad_norm": 10.885397911071777, "learning_rate": 9.942227394072148e-05, "loss": 1.0645, "step": 1173 }, { "epoch": 0.07954468459922759, "grad_norm": 13.274818420410156, "learning_rate": 9.942090492162366e-05, "loss": 1.3411, "step": 1174 }, { "epoch": 0.07961243986719968, "grad_norm": 14.061238288879395, "learning_rate": 9.941953590252584e-05, "loss": 1.1916, "step": 1175 }, { "epoch": 0.07968019513517176, "grad_norm": 10.074264526367188, "learning_rate": 9.941816688342803e-05, "loss": 1.092, "step": 1176 }, { "epoch": 0.07974795040314385, "grad_norm": 14.741287231445312, "learning_rate": 9.941679786433021e-05, "loss": 1.3774, "step": 1177 }, { "epoch": 0.07981570567111593, "grad_norm": 11.308422088623047, "learning_rate": 9.941542884523239e-05, "loss": 0.9298, "step": 1178 }, { "epoch": 0.07988346093908802, "grad_norm": 14.375280380249023, "learning_rate": 9.941405982613457e-05, "loss": 1.4525, "step": 1179 }, { "epoch": 0.0799512162070601, "grad_norm": 11.724523544311523, "learning_rate": 9.941269080703677e-05, "loss": 1.3616, "step": 1180 }, { "epoch": 0.08001897147503219, "grad_norm": 12.578176498413086, "learning_rate": 9.941132178793895e-05, "loss": 1.1513, "step": 1181 }, { "epoch": 0.08008672674300427, "grad_norm": 12.100804328918457, "learning_rate": 9.940995276884113e-05, "loss": 1.0656, "step": 1182 }, { "epoch": 0.08015448201097636, "grad_norm": 11.680248260498047, "learning_rate": 9.940858374974331e-05, "loss": 1.2358, "step": 1183 }, { "epoch": 0.08022223727894844, "grad_norm": 10.066198348999023, "learning_rate": 9.940721473064549e-05, "loss": 0.9219, "step": 1184 }, { "epoch": 0.08028999254692053, "grad_norm": 10.813334465026855, "learning_rate": 9.940584571154768e-05, "loss": 1.1456, "step": 1185 }, { "epoch": 0.0803577478148926, "grad_norm": 14.004862785339355, "learning_rate": 9.940447669244986e-05, "loss": 1.2375, "step": 1186 }, { "epoch": 0.0804255030828647, "grad_norm": 11.868766784667969, "learning_rate": 9.940310767335204e-05, "loss": 1.3027, "step": 1187 }, { "epoch": 0.08049325835083677, "grad_norm": 12.48153018951416, "learning_rate": 9.940173865425422e-05, "loss": 1.1192, "step": 1188 }, { "epoch": 0.08056101361880887, "grad_norm": 12.340612411499023, "learning_rate": 9.940036963515642e-05, "loss": 1.246, "step": 1189 }, { "epoch": 0.08062876888678094, "grad_norm": 12.205392837524414, "learning_rate": 9.93990006160586e-05, "loss": 1.1233, "step": 1190 }, { "epoch": 0.08069652415475304, "grad_norm": 12.69509220123291, "learning_rate": 9.939763159696078e-05, "loss": 1.2202, "step": 1191 }, { "epoch": 0.08076427942272511, "grad_norm": 12.40784740447998, "learning_rate": 9.939626257786296e-05, "loss": 1.1345, "step": 1192 }, { "epoch": 0.0808320346906972, "grad_norm": 11.678507804870605, "learning_rate": 9.939489355876515e-05, "loss": 1.4513, "step": 1193 }, { "epoch": 0.08089978995866928, "grad_norm": 11.649873733520508, "learning_rate": 9.939352453966733e-05, "loss": 1.3827, "step": 1194 }, { "epoch": 0.08096754522664137, "grad_norm": 12.378853797912598, "learning_rate": 9.939215552056951e-05, "loss": 1.3632, "step": 1195 }, { "epoch": 0.08103530049461345, "grad_norm": 11.023188591003418, "learning_rate": 9.93907865014717e-05, "loss": 1.3892, "step": 1196 }, { "epoch": 0.08110305576258554, "grad_norm": 13.111897468566895, "learning_rate": 9.938941748237389e-05, "loss": 1.1973, "step": 1197 }, { "epoch": 0.08117081103055762, "grad_norm": 10.171613693237305, "learning_rate": 9.938804846327607e-05, "loss": 1.3862, "step": 1198 }, { "epoch": 0.08123856629852971, "grad_norm": 13.327658653259277, "learning_rate": 9.938667944417826e-05, "loss": 1.3173, "step": 1199 }, { "epoch": 0.08130632156650179, "grad_norm": 11.715154647827148, "learning_rate": 9.938531042508044e-05, "loss": 1.0839, "step": 1200 }, { "epoch": 0.08137407683447388, "grad_norm": 11.521212577819824, "learning_rate": 9.938394140598262e-05, "loss": 1.0285, "step": 1201 }, { "epoch": 0.08144183210244596, "grad_norm": 12.024236679077148, "learning_rate": 9.93825723868848e-05, "loss": 1.3025, "step": 1202 }, { "epoch": 0.08150958737041805, "grad_norm": 10.245376586914062, "learning_rate": 9.9381203367787e-05, "loss": 1.1658, "step": 1203 }, { "epoch": 0.08157734263839013, "grad_norm": 10.731759071350098, "learning_rate": 9.937983434868918e-05, "loss": 1.0565, "step": 1204 }, { "epoch": 0.08164509790636222, "grad_norm": 16.714153289794922, "learning_rate": 9.937846532959136e-05, "loss": 1.2761, "step": 1205 }, { "epoch": 0.0817128531743343, "grad_norm": 11.172699928283691, "learning_rate": 9.937709631049354e-05, "loss": 1.1887, "step": 1206 }, { "epoch": 0.08178060844230639, "grad_norm": 11.384743690490723, "learning_rate": 9.937572729139572e-05, "loss": 0.9546, "step": 1207 }, { "epoch": 0.08184836371027847, "grad_norm": 11.163822174072266, "learning_rate": 9.937435827229791e-05, "loss": 1.1498, "step": 1208 }, { "epoch": 0.08191611897825056, "grad_norm": 10.657593727111816, "learning_rate": 9.937298925320009e-05, "loss": 0.9506, "step": 1209 }, { "epoch": 0.08198387424622264, "grad_norm": 10.6862211227417, "learning_rate": 9.937162023410227e-05, "loss": 1.2308, "step": 1210 }, { "epoch": 0.08205162951419473, "grad_norm": 10.649473190307617, "learning_rate": 9.937025121500445e-05, "loss": 1.2679, "step": 1211 }, { "epoch": 0.08211938478216681, "grad_norm": 9.298782348632812, "learning_rate": 9.936888219590665e-05, "loss": 1.1327, "step": 1212 }, { "epoch": 0.0821871400501389, "grad_norm": 11.448348045349121, "learning_rate": 9.936751317680883e-05, "loss": 1.2796, "step": 1213 }, { "epoch": 0.08225489531811098, "grad_norm": 12.141517639160156, "learning_rate": 9.936614415771101e-05, "loss": 1.1128, "step": 1214 }, { "epoch": 0.08232265058608307, "grad_norm": 11.799830436706543, "learning_rate": 9.936477513861319e-05, "loss": 1.0743, "step": 1215 }, { "epoch": 0.08239040585405515, "grad_norm": 11.952958106994629, "learning_rate": 9.936340611951537e-05, "loss": 1.1659, "step": 1216 }, { "epoch": 0.08245816112202724, "grad_norm": 11.870144844055176, "learning_rate": 9.936203710041756e-05, "loss": 1.1195, "step": 1217 }, { "epoch": 0.08252591638999933, "grad_norm": 11.970368385314941, "learning_rate": 9.936066808131974e-05, "loss": 1.2139, "step": 1218 }, { "epoch": 0.08259367165797141, "grad_norm": 11.211687088012695, "learning_rate": 9.935929906222192e-05, "loss": 1.2107, "step": 1219 }, { "epoch": 0.0826614269259435, "grad_norm": 12.987563133239746, "learning_rate": 9.93579300431241e-05, "loss": 1.2804, "step": 1220 }, { "epoch": 0.08272918219391558, "grad_norm": 12.337888717651367, "learning_rate": 9.935656102402628e-05, "loss": 0.8713, "step": 1221 }, { "epoch": 0.08279693746188767, "grad_norm": 11.717150688171387, "learning_rate": 9.935519200492848e-05, "loss": 1.2684, "step": 1222 }, { "epoch": 0.08286469272985975, "grad_norm": 12.658769607543945, "learning_rate": 9.935382298583066e-05, "loss": 1.0231, "step": 1223 }, { "epoch": 0.08293244799783184, "grad_norm": 10.526476860046387, "learning_rate": 9.935245396673284e-05, "loss": 1.0605, "step": 1224 }, { "epoch": 0.08300020326580392, "grad_norm": 10.575004577636719, "learning_rate": 9.935108494763502e-05, "loss": 1.2169, "step": 1225 }, { "epoch": 0.083067958533776, "grad_norm": 10.499407768249512, "learning_rate": 9.934971592853721e-05, "loss": 1.3064, "step": 1226 }, { "epoch": 0.08313571380174808, "grad_norm": 12.25387191772461, "learning_rate": 9.934834690943939e-05, "loss": 1.1554, "step": 1227 }, { "epoch": 0.08320346906972018, "grad_norm": 12.173775672912598, "learning_rate": 9.934697789034157e-05, "loss": 1.0917, "step": 1228 }, { "epoch": 0.08327122433769225, "grad_norm": 12.424721717834473, "learning_rate": 9.934560887124375e-05, "loss": 1.3679, "step": 1229 }, { "epoch": 0.08333897960566435, "grad_norm": 11.291987419128418, "learning_rate": 9.934423985214593e-05, "loss": 1.1439, "step": 1230 }, { "epoch": 0.08340673487363642, "grad_norm": 10.908637046813965, "learning_rate": 9.934287083304813e-05, "loss": 1.0487, "step": 1231 }, { "epoch": 0.08347449014160851, "grad_norm": 14.767544746398926, "learning_rate": 9.934150181395031e-05, "loss": 1.2951, "step": 1232 }, { "epoch": 0.08354224540958059, "grad_norm": 11.959871292114258, "learning_rate": 9.934013279485249e-05, "loss": 1.1244, "step": 1233 }, { "epoch": 0.08361000067755268, "grad_norm": 11.19450569152832, "learning_rate": 9.933876377575467e-05, "loss": 1.2802, "step": 1234 }, { "epoch": 0.08367775594552476, "grad_norm": 10.71377182006836, "learning_rate": 9.933739475665686e-05, "loss": 1.0376, "step": 1235 }, { "epoch": 0.08374551121349685, "grad_norm": 12.174454689025879, "learning_rate": 9.933602573755904e-05, "loss": 1.5779, "step": 1236 }, { "epoch": 0.08381326648146893, "grad_norm": 9.863836288452148, "learning_rate": 9.933465671846122e-05, "loss": 1.0443, "step": 1237 }, { "epoch": 0.08388102174944102, "grad_norm": 11.964838027954102, "learning_rate": 9.93332876993634e-05, "loss": 1.0613, "step": 1238 }, { "epoch": 0.0839487770174131, "grad_norm": 11.49203109741211, "learning_rate": 9.93319186802656e-05, "loss": 1.162, "step": 1239 }, { "epoch": 0.08401653228538519, "grad_norm": 12.548815727233887, "learning_rate": 9.933054966116778e-05, "loss": 1.0652, "step": 1240 }, { "epoch": 0.08408428755335727, "grad_norm": 13.37637996673584, "learning_rate": 9.932918064206996e-05, "loss": 1.2281, "step": 1241 }, { "epoch": 0.08415204282132936, "grad_norm": 12.59211254119873, "learning_rate": 9.932781162297215e-05, "loss": 1.0624, "step": 1242 }, { "epoch": 0.08421979808930144, "grad_norm": 13.386221885681152, "learning_rate": 9.932644260387433e-05, "loss": 1.4393, "step": 1243 }, { "epoch": 0.08428755335727353, "grad_norm": 12.647525787353516, "learning_rate": 9.932507358477651e-05, "loss": 1.2228, "step": 1244 }, { "epoch": 0.08435530862524561, "grad_norm": 12.039474487304688, "learning_rate": 9.93237045656787e-05, "loss": 1.1354, "step": 1245 }, { "epoch": 0.0844230638932177, "grad_norm": 11.373556137084961, "learning_rate": 9.932233554658089e-05, "loss": 1.2418, "step": 1246 }, { "epoch": 0.08449081916118978, "grad_norm": 10.944781303405762, "learning_rate": 9.932096652748307e-05, "loss": 1.1515, "step": 1247 }, { "epoch": 0.08455857442916187, "grad_norm": 12.174854278564453, "learning_rate": 9.931959750838525e-05, "loss": 1.2063, "step": 1248 }, { "epoch": 0.08462632969713395, "grad_norm": 8.846879005432129, "learning_rate": 9.931822848928744e-05, "loss": 0.9639, "step": 1249 }, { "epoch": 0.08469408496510604, "grad_norm": 13.793547630310059, "learning_rate": 9.931685947018962e-05, "loss": 1.1236, "step": 1250 }, { "epoch": 0.08476184023307812, "grad_norm": 14.486831665039062, "learning_rate": 9.93154904510918e-05, "loss": 1.2047, "step": 1251 }, { "epoch": 0.08482959550105021, "grad_norm": 13.262588500976562, "learning_rate": 9.931412143199398e-05, "loss": 1.4091, "step": 1252 }, { "epoch": 0.08489735076902229, "grad_norm": 13.289068222045898, "learning_rate": 9.931275241289616e-05, "loss": 1.5065, "step": 1253 }, { "epoch": 0.08496510603699438, "grad_norm": 10.22205638885498, "learning_rate": 9.931138339379836e-05, "loss": 1.1116, "step": 1254 }, { "epoch": 0.08503286130496646, "grad_norm": 13.141668319702148, "learning_rate": 9.931001437470054e-05, "loss": 1.4006, "step": 1255 }, { "epoch": 0.08510061657293855, "grad_norm": 11.817032814025879, "learning_rate": 9.930864535560272e-05, "loss": 1.2062, "step": 1256 }, { "epoch": 0.08516837184091063, "grad_norm": 10.814498901367188, "learning_rate": 9.93072763365049e-05, "loss": 1.138, "step": 1257 }, { "epoch": 0.08523612710888272, "grad_norm": 12.598155975341797, "learning_rate": 9.930590731740709e-05, "loss": 1.1959, "step": 1258 }, { "epoch": 0.0853038823768548, "grad_norm": 11.909974098205566, "learning_rate": 9.930453829830927e-05, "loss": 1.2891, "step": 1259 }, { "epoch": 0.08537163764482689, "grad_norm": 12.663064956665039, "learning_rate": 9.930316927921145e-05, "loss": 1.3017, "step": 1260 }, { "epoch": 0.08543939291279896, "grad_norm": 10.810627937316895, "learning_rate": 9.930180026011363e-05, "loss": 1.16, "step": 1261 }, { "epoch": 0.08550714818077106, "grad_norm": 9.182926177978516, "learning_rate": 9.930043124101581e-05, "loss": 1.2625, "step": 1262 }, { "epoch": 0.08557490344874313, "grad_norm": 12.73978042602539, "learning_rate": 9.9299062221918e-05, "loss": 0.9904, "step": 1263 }, { "epoch": 0.08564265871671523, "grad_norm": 10.346587181091309, "learning_rate": 9.929769320282019e-05, "loss": 1.0167, "step": 1264 }, { "epoch": 0.0857104139846873, "grad_norm": 12.145682334899902, "learning_rate": 9.929632418372237e-05, "loss": 1.4282, "step": 1265 }, { "epoch": 0.0857781692526594, "grad_norm": 11.515445709228516, "learning_rate": 9.929495516462455e-05, "loss": 1.0527, "step": 1266 }, { "epoch": 0.08584592452063147, "grad_norm": 10.979050636291504, "learning_rate": 9.929358614552674e-05, "loss": 1.2888, "step": 1267 }, { "epoch": 0.08591367978860356, "grad_norm": 10.248215675354004, "learning_rate": 9.929221712642892e-05, "loss": 1.4316, "step": 1268 }, { "epoch": 0.08598143505657564, "grad_norm": 13.743851661682129, "learning_rate": 9.92908481073311e-05, "loss": 1.5047, "step": 1269 }, { "epoch": 0.08604919032454773, "grad_norm": 9.919225692749023, "learning_rate": 9.928947908823328e-05, "loss": 0.9602, "step": 1270 }, { "epoch": 0.08611694559251983, "grad_norm": 9.795915603637695, "learning_rate": 9.928811006913546e-05, "loss": 1.3736, "step": 1271 }, { "epoch": 0.0861847008604919, "grad_norm": 13.108200073242188, "learning_rate": 9.928674105003766e-05, "loss": 1.2591, "step": 1272 }, { "epoch": 0.086252456128464, "grad_norm": 13.108073234558105, "learning_rate": 9.928537203093984e-05, "loss": 1.102, "step": 1273 }, { "epoch": 0.08632021139643607, "grad_norm": 15.177817344665527, "learning_rate": 9.928400301184202e-05, "loss": 1.3164, "step": 1274 }, { "epoch": 0.08638796666440816, "grad_norm": 13.048440933227539, "learning_rate": 9.92826339927442e-05, "loss": 1.3029, "step": 1275 }, { "epoch": 0.08645572193238024, "grad_norm": 10.982895851135254, "learning_rate": 9.928126497364638e-05, "loss": 1.1829, "step": 1276 }, { "epoch": 0.08652347720035233, "grad_norm": 10.901629447937012, "learning_rate": 9.927989595454857e-05, "loss": 1.2116, "step": 1277 }, { "epoch": 0.08659123246832441, "grad_norm": 12.924722671508789, "learning_rate": 9.927852693545075e-05, "loss": 1.6292, "step": 1278 }, { "epoch": 0.0866589877362965, "grad_norm": 12.572770118713379, "learning_rate": 9.927715791635293e-05, "loss": 1.1012, "step": 1279 }, { "epoch": 0.08672674300426858, "grad_norm": 12.803020477294922, "learning_rate": 9.927578889725511e-05, "loss": 1.276, "step": 1280 }, { "epoch": 0.08679449827224067, "grad_norm": 10.92810344696045, "learning_rate": 9.92744198781573e-05, "loss": 1.1876, "step": 1281 }, { "epoch": 0.08686225354021275, "grad_norm": 12.253180503845215, "learning_rate": 9.927305085905949e-05, "loss": 1.1807, "step": 1282 }, { "epoch": 0.08693000880818484, "grad_norm": 11.163126945495605, "learning_rate": 9.927168183996167e-05, "loss": 1.1671, "step": 1283 }, { "epoch": 0.08699776407615692, "grad_norm": 10.726607322692871, "learning_rate": 9.927031282086385e-05, "loss": 1.185, "step": 1284 }, { "epoch": 0.08706551934412901, "grad_norm": 13.265491485595703, "learning_rate": 9.926894380176603e-05, "loss": 1.2916, "step": 1285 }, { "epoch": 0.08713327461210109, "grad_norm": 14.559592247009277, "learning_rate": 9.926757478266822e-05, "loss": 1.288, "step": 1286 }, { "epoch": 0.08720102988007318, "grad_norm": 14.816813468933105, "learning_rate": 9.92662057635704e-05, "loss": 1.4397, "step": 1287 }, { "epoch": 0.08726878514804526, "grad_norm": 11.811420440673828, "learning_rate": 9.926483674447258e-05, "loss": 1.348, "step": 1288 }, { "epoch": 0.08733654041601735, "grad_norm": 10.920133590698242, "learning_rate": 9.926346772537478e-05, "loss": 1.4169, "step": 1289 }, { "epoch": 0.08740429568398943, "grad_norm": 11.690089225769043, "learning_rate": 9.926209870627696e-05, "loss": 1.1199, "step": 1290 }, { "epoch": 0.08747205095196152, "grad_norm": 9.411031723022461, "learning_rate": 9.926072968717914e-05, "loss": 1.031, "step": 1291 }, { "epoch": 0.0875398062199336, "grad_norm": 12.174457550048828, "learning_rate": 9.925936066808133e-05, "loss": 1.0622, "step": 1292 }, { "epoch": 0.08760756148790569, "grad_norm": 10.346089363098145, "learning_rate": 9.925799164898351e-05, "loss": 1.2777, "step": 1293 }, { "epoch": 0.08767531675587777, "grad_norm": 12.534863471984863, "learning_rate": 9.925662262988569e-05, "loss": 1.0093, "step": 1294 }, { "epoch": 0.08774307202384986, "grad_norm": 12.050302505493164, "learning_rate": 9.925525361078789e-05, "loss": 1.0591, "step": 1295 }, { "epoch": 0.08781082729182194, "grad_norm": 11.556166648864746, "learning_rate": 9.925388459169007e-05, "loss": 1.1452, "step": 1296 }, { "epoch": 0.08787858255979403, "grad_norm": 9.693270683288574, "learning_rate": 9.925251557259225e-05, "loss": 1.0332, "step": 1297 }, { "epoch": 0.0879463378277661, "grad_norm": 12.646526336669922, "learning_rate": 9.925114655349443e-05, "loss": 1.1894, "step": 1298 }, { "epoch": 0.0880140930957382, "grad_norm": 10.676809310913086, "learning_rate": 9.924977753439661e-05, "loss": 1.1727, "step": 1299 }, { "epoch": 0.08808184836371027, "grad_norm": 11.182327270507812, "learning_rate": 9.92484085152988e-05, "loss": 1.0145, "step": 1300 }, { "epoch": 0.08814960363168237, "grad_norm": 11.55026626586914, "learning_rate": 9.924703949620098e-05, "loss": 1.2187, "step": 1301 }, { "epoch": 0.08821735889965444, "grad_norm": 11.502679824829102, "learning_rate": 9.924567047710316e-05, "loss": 1.3116, "step": 1302 }, { "epoch": 0.08828511416762654, "grad_norm": 9.676247596740723, "learning_rate": 9.924430145800534e-05, "loss": 1.2869, "step": 1303 }, { "epoch": 0.08835286943559861, "grad_norm": 15.123950004577637, "learning_rate": 9.924293243890754e-05, "loss": 1.0225, "step": 1304 }, { "epoch": 0.0884206247035707, "grad_norm": 14.030994415283203, "learning_rate": 9.924156341980972e-05, "loss": 0.9829, "step": 1305 }, { "epoch": 0.08848837997154278, "grad_norm": 10.00402545928955, "learning_rate": 9.92401944007119e-05, "loss": 1.0946, "step": 1306 }, { "epoch": 0.08855613523951487, "grad_norm": 9.077853202819824, "learning_rate": 9.923882538161408e-05, "loss": 0.852, "step": 1307 }, { "epoch": 0.08862389050748695, "grad_norm": 12.777885437011719, "learning_rate": 9.923745636251626e-05, "loss": 1.0513, "step": 1308 }, { "epoch": 0.08869164577545904, "grad_norm": 10.686469078063965, "learning_rate": 9.923608734341845e-05, "loss": 1.0873, "step": 1309 }, { "epoch": 0.08875940104343112, "grad_norm": 11.51689338684082, "learning_rate": 9.923471832432063e-05, "loss": 1.0037, "step": 1310 }, { "epoch": 0.08882715631140321, "grad_norm": 13.259784698486328, "learning_rate": 9.923334930522281e-05, "loss": 1.2004, "step": 1311 }, { "epoch": 0.08889491157937529, "grad_norm": 10.63463306427002, "learning_rate": 9.923198028612499e-05, "loss": 1.2067, "step": 1312 }, { "epoch": 0.08896266684734738, "grad_norm": 10.838210105895996, "learning_rate": 9.923061126702719e-05, "loss": 1.2939, "step": 1313 }, { "epoch": 0.08903042211531946, "grad_norm": 12.058418273925781, "learning_rate": 9.922924224792937e-05, "loss": 1.3091, "step": 1314 }, { "epoch": 0.08909817738329155, "grad_norm": 10.619451522827148, "learning_rate": 9.922787322883155e-05, "loss": 1.1635, "step": 1315 }, { "epoch": 0.08916593265126363, "grad_norm": 9.37607192993164, "learning_rate": 9.922650420973373e-05, "loss": 1.2206, "step": 1316 }, { "epoch": 0.08923368791923572, "grad_norm": 12.119776725769043, "learning_rate": 9.922513519063591e-05, "loss": 1.2159, "step": 1317 }, { "epoch": 0.0893014431872078, "grad_norm": 14.013461112976074, "learning_rate": 9.92237661715381e-05, "loss": 1.1445, "step": 1318 }, { "epoch": 0.08936919845517989, "grad_norm": 11.560707092285156, "learning_rate": 9.922239715244028e-05, "loss": 1.3617, "step": 1319 }, { "epoch": 0.08943695372315197, "grad_norm": 11.817791938781738, "learning_rate": 9.922102813334246e-05, "loss": 1.1931, "step": 1320 }, { "epoch": 0.08950470899112406, "grad_norm": 11.544127464294434, "learning_rate": 9.921965911424464e-05, "loss": 1.0766, "step": 1321 }, { "epoch": 0.08957246425909614, "grad_norm": 10.67740249633789, "learning_rate": 9.921829009514684e-05, "loss": 1.2103, "step": 1322 }, { "epoch": 0.08964021952706823, "grad_norm": 11.774645805358887, "learning_rate": 9.921692107604902e-05, "loss": 1.1921, "step": 1323 }, { "epoch": 0.08970797479504032, "grad_norm": 12.594759941101074, "learning_rate": 9.92155520569512e-05, "loss": 0.906, "step": 1324 }, { "epoch": 0.0897757300630124, "grad_norm": 10.988224983215332, "learning_rate": 9.921418303785338e-05, "loss": 1.4195, "step": 1325 }, { "epoch": 0.08984348533098449, "grad_norm": 12.22718620300293, "learning_rate": 9.921281401875556e-05, "loss": 1.2461, "step": 1326 }, { "epoch": 0.08991124059895657, "grad_norm": 10.964727401733398, "learning_rate": 9.921144499965775e-05, "loss": 1.1254, "step": 1327 }, { "epoch": 0.08997899586692866, "grad_norm": 11.327523231506348, "learning_rate": 9.921007598055993e-05, "loss": 1.145, "step": 1328 }, { "epoch": 0.09004675113490074, "grad_norm": 9.870691299438477, "learning_rate": 9.920870696146211e-05, "loss": 1.2408, "step": 1329 }, { "epoch": 0.09011450640287283, "grad_norm": 11.02373218536377, "learning_rate": 9.920733794236429e-05, "loss": 1.1084, "step": 1330 }, { "epoch": 0.09018226167084491, "grad_norm": 12.300410270690918, "learning_rate": 9.920596892326647e-05, "loss": 1.1542, "step": 1331 }, { "epoch": 0.090250016938817, "grad_norm": 9.832919120788574, "learning_rate": 9.920459990416867e-05, "loss": 1.1195, "step": 1332 }, { "epoch": 0.09031777220678908, "grad_norm": 10.432522773742676, "learning_rate": 9.920323088507085e-05, "loss": 1.1825, "step": 1333 }, { "epoch": 0.09038552747476117, "grad_norm": 11.878792762756348, "learning_rate": 9.920186186597303e-05, "loss": 0.9647, "step": 1334 }, { "epoch": 0.09045328274273325, "grad_norm": 11.866320610046387, "learning_rate": 9.920049284687522e-05, "loss": 1.3598, "step": 1335 }, { "epoch": 0.09052103801070534, "grad_norm": 14.11543083190918, "learning_rate": 9.91991238277774e-05, "loss": 1.4172, "step": 1336 }, { "epoch": 0.09058879327867742, "grad_norm": 13.841622352600098, "learning_rate": 9.919775480867958e-05, "loss": 1.4119, "step": 1337 }, { "epoch": 0.0906565485466495, "grad_norm": 11.077167510986328, "learning_rate": 9.919638578958178e-05, "loss": 0.984, "step": 1338 }, { "epoch": 0.09072430381462158, "grad_norm": 10.904266357421875, "learning_rate": 9.919501677048396e-05, "loss": 1.0439, "step": 1339 }, { "epoch": 0.09079205908259368, "grad_norm": 11.623948097229004, "learning_rate": 9.919364775138614e-05, "loss": 1.0318, "step": 1340 }, { "epoch": 0.09085981435056575, "grad_norm": 10.893725395202637, "learning_rate": 9.919227873228833e-05, "loss": 1.2626, "step": 1341 }, { "epoch": 0.09092756961853785, "grad_norm": 10.064491271972656, "learning_rate": 9.919090971319051e-05, "loss": 1.1483, "step": 1342 }, { "epoch": 0.09099532488650992, "grad_norm": 9.854101181030273, "learning_rate": 9.918954069409269e-05, "loss": 1.1595, "step": 1343 }, { "epoch": 0.09106308015448202, "grad_norm": 11.682498931884766, "learning_rate": 9.918817167499487e-05, "loss": 1.2887, "step": 1344 }, { "epoch": 0.09113083542245409, "grad_norm": 10.484097480773926, "learning_rate": 9.918680265589707e-05, "loss": 1.2, "step": 1345 }, { "epoch": 0.09119859069042618, "grad_norm": 12.332358360290527, "learning_rate": 9.918543363679925e-05, "loss": 1.262, "step": 1346 }, { "epoch": 0.09126634595839826, "grad_norm": 13.706925392150879, "learning_rate": 9.918406461770143e-05, "loss": 1.2151, "step": 1347 }, { "epoch": 0.09133410122637035, "grad_norm": 13.918478965759277, "learning_rate": 9.91826955986036e-05, "loss": 1.4653, "step": 1348 }, { "epoch": 0.09140185649434243, "grad_norm": 11.970015525817871, "learning_rate": 9.918132657950579e-05, "loss": 1.2452, "step": 1349 }, { "epoch": 0.09146961176231452, "grad_norm": 9.698074340820312, "learning_rate": 9.917995756040798e-05, "loss": 0.9142, "step": 1350 }, { "epoch": 0.0915373670302866, "grad_norm": 9.225728988647461, "learning_rate": 9.917858854131016e-05, "loss": 1.0293, "step": 1351 }, { "epoch": 0.09160512229825869, "grad_norm": 10.77661418914795, "learning_rate": 9.917721952221234e-05, "loss": 1.1865, "step": 1352 }, { "epoch": 0.09167287756623077, "grad_norm": 10.45409870147705, "learning_rate": 9.917585050311452e-05, "loss": 1.1553, "step": 1353 }, { "epoch": 0.09174063283420286, "grad_norm": 10.44918441772461, "learning_rate": 9.91744814840167e-05, "loss": 1.1621, "step": 1354 }, { "epoch": 0.09180838810217494, "grad_norm": 14.769590377807617, "learning_rate": 9.91731124649189e-05, "loss": 1.3502, "step": 1355 }, { "epoch": 0.09187614337014703, "grad_norm": 12.733844757080078, "learning_rate": 9.917174344582108e-05, "loss": 1.1973, "step": 1356 }, { "epoch": 0.09194389863811911, "grad_norm": 13.365818977355957, "learning_rate": 9.917037442672326e-05, "loss": 1.5412, "step": 1357 }, { "epoch": 0.0920116539060912, "grad_norm": 11.163050651550293, "learning_rate": 9.916900540762544e-05, "loss": 1.3142, "step": 1358 }, { "epoch": 0.09207940917406328, "grad_norm": 11.420190811157227, "learning_rate": 9.916763638852763e-05, "loss": 1.2257, "step": 1359 }, { "epoch": 0.09214716444203537, "grad_norm": 9.66398811340332, "learning_rate": 9.916626736942981e-05, "loss": 1.0584, "step": 1360 }, { "epoch": 0.09221491971000745, "grad_norm": 13.643363952636719, "learning_rate": 9.916489835033199e-05, "loss": 1.3783, "step": 1361 }, { "epoch": 0.09228267497797954, "grad_norm": 11.658889770507812, "learning_rate": 9.916352933123417e-05, "loss": 1.2126, "step": 1362 }, { "epoch": 0.09235043024595162, "grad_norm": 11.2728271484375, "learning_rate": 9.916216031213635e-05, "loss": 1.2117, "step": 1363 }, { "epoch": 0.09241818551392371, "grad_norm": 13.576864242553711, "learning_rate": 9.916079129303855e-05, "loss": 1.2922, "step": 1364 }, { "epoch": 0.09248594078189579, "grad_norm": 10.950700759887695, "learning_rate": 9.915942227394073e-05, "loss": 1.1394, "step": 1365 }, { "epoch": 0.09255369604986788, "grad_norm": 11.638351440429688, "learning_rate": 9.91580532548429e-05, "loss": 1.3927, "step": 1366 }, { "epoch": 0.09262145131783996, "grad_norm": 12.355545043945312, "learning_rate": 9.915668423574509e-05, "loss": 1.1861, "step": 1367 }, { "epoch": 0.09268920658581205, "grad_norm": 11.543237686157227, "learning_rate": 9.915531521664728e-05, "loss": 0.9868, "step": 1368 }, { "epoch": 0.09275696185378413, "grad_norm": 11.684252738952637, "learning_rate": 9.915394619754946e-05, "loss": 1.3332, "step": 1369 }, { "epoch": 0.09282471712175622, "grad_norm": 10.775650024414062, "learning_rate": 9.915257717845164e-05, "loss": 1.1164, "step": 1370 }, { "epoch": 0.0928924723897283, "grad_norm": 11.649751663208008, "learning_rate": 9.915120815935382e-05, "loss": 1.0369, "step": 1371 }, { "epoch": 0.09296022765770039, "grad_norm": 9.741403579711914, "learning_rate": 9.9149839140256e-05, "loss": 1.0509, "step": 1372 }, { "epoch": 0.09302798292567246, "grad_norm": 13.804118156433105, "learning_rate": 9.91484701211582e-05, "loss": 1.4318, "step": 1373 }, { "epoch": 0.09309573819364456, "grad_norm": 10.939459800720215, "learning_rate": 9.914710110206038e-05, "loss": 1.218, "step": 1374 }, { "epoch": 0.09316349346161663, "grad_norm": 14.076252937316895, "learning_rate": 9.914573208296256e-05, "loss": 1.4699, "step": 1375 }, { "epoch": 0.09323124872958873, "grad_norm": 12.90072250366211, "learning_rate": 9.914436306386474e-05, "loss": 1.3418, "step": 1376 }, { "epoch": 0.09329900399756082, "grad_norm": 12.7711820602417, "learning_rate": 9.914299404476693e-05, "loss": 0.9866, "step": 1377 }, { "epoch": 0.0933667592655329, "grad_norm": 9.205671310424805, "learning_rate": 9.914162502566911e-05, "loss": 1.1345, "step": 1378 }, { "epoch": 0.09343451453350499, "grad_norm": 11.027194023132324, "learning_rate": 9.914025600657129e-05, "loss": 1.2637, "step": 1379 }, { "epoch": 0.09350226980147706, "grad_norm": 12.861044883728027, "learning_rate": 9.913888698747347e-05, "loss": 1.2348, "step": 1380 }, { "epoch": 0.09357002506944916, "grad_norm": 11.266969680786133, "learning_rate": 9.913751796837567e-05, "loss": 1.0773, "step": 1381 }, { "epoch": 0.09363778033742123, "grad_norm": 13.137110710144043, "learning_rate": 9.913614894927785e-05, "loss": 1.0537, "step": 1382 }, { "epoch": 0.09370553560539333, "grad_norm": 11.343362808227539, "learning_rate": 9.913477993018003e-05, "loss": 1.335, "step": 1383 }, { "epoch": 0.0937732908733654, "grad_norm": 11.472663879394531, "learning_rate": 9.913341091108222e-05, "loss": 1.1362, "step": 1384 }, { "epoch": 0.0938410461413375, "grad_norm": 8.441573143005371, "learning_rate": 9.91320418919844e-05, "loss": 1.1646, "step": 1385 }, { "epoch": 0.09390880140930957, "grad_norm": 12.570130348205566, "learning_rate": 9.913067287288658e-05, "loss": 1.3535, "step": 1386 }, { "epoch": 0.09397655667728166, "grad_norm": 11.671664237976074, "learning_rate": 9.912930385378877e-05, "loss": 1.1707, "step": 1387 }, { "epoch": 0.09404431194525374, "grad_norm": 12.638328552246094, "learning_rate": 9.912793483469096e-05, "loss": 1.057, "step": 1388 }, { "epoch": 0.09411206721322583, "grad_norm": 10.506028175354004, "learning_rate": 9.912656581559314e-05, "loss": 0.9279, "step": 1389 }, { "epoch": 0.09417982248119791, "grad_norm": 11.536858558654785, "learning_rate": 9.912519679649532e-05, "loss": 1.3351, "step": 1390 }, { "epoch": 0.09424757774917, "grad_norm": 12.692436218261719, "learning_rate": 9.912382777739751e-05, "loss": 1.164, "step": 1391 }, { "epoch": 0.09431533301714208, "grad_norm": 12.088066101074219, "learning_rate": 9.912245875829969e-05, "loss": 1.2971, "step": 1392 }, { "epoch": 0.09438308828511417, "grad_norm": 12.133123397827148, "learning_rate": 9.912108973920187e-05, "loss": 1.2742, "step": 1393 }, { "epoch": 0.09445084355308625, "grad_norm": 10.168001174926758, "learning_rate": 9.911972072010405e-05, "loss": 1.0964, "step": 1394 }, { "epoch": 0.09451859882105834, "grad_norm": 10.561311721801758, "learning_rate": 9.911835170100623e-05, "loss": 1.1828, "step": 1395 }, { "epoch": 0.09458635408903042, "grad_norm": 11.497330665588379, "learning_rate": 9.911698268190843e-05, "loss": 1.4699, "step": 1396 }, { "epoch": 0.09465410935700251, "grad_norm": 12.190573692321777, "learning_rate": 9.91156136628106e-05, "loss": 1.1601, "step": 1397 }, { "epoch": 0.09472186462497459, "grad_norm": 10.633028030395508, "learning_rate": 9.911424464371279e-05, "loss": 1.333, "step": 1398 }, { "epoch": 0.09478961989294668, "grad_norm": 12.262279510498047, "learning_rate": 9.911287562461497e-05, "loss": 1.2214, "step": 1399 }, { "epoch": 0.09485737516091876, "grad_norm": 11.506840705871582, "learning_rate": 9.911150660551716e-05, "loss": 1.1172, "step": 1400 }, { "epoch": 0.09492513042889085, "grad_norm": 11.453936576843262, "learning_rate": 9.911013758641934e-05, "loss": 1.2205, "step": 1401 }, { "epoch": 0.09499288569686293, "grad_norm": 9.980772972106934, "learning_rate": 9.910876856732152e-05, "loss": 1.1004, "step": 1402 }, { "epoch": 0.09506064096483502, "grad_norm": 11.775416374206543, "learning_rate": 9.91073995482237e-05, "loss": 1.1587, "step": 1403 }, { "epoch": 0.0951283962328071, "grad_norm": 8.840147972106934, "learning_rate": 9.910603052912588e-05, "loss": 0.8702, "step": 1404 }, { "epoch": 0.09519615150077919, "grad_norm": 10.938506126403809, "learning_rate": 9.910466151002808e-05, "loss": 1.085, "step": 1405 }, { "epoch": 0.09526390676875127, "grad_norm": 11.733402252197266, "learning_rate": 9.910329249093026e-05, "loss": 1.1202, "step": 1406 }, { "epoch": 0.09533166203672336, "grad_norm": 11.616521835327148, "learning_rate": 9.910192347183244e-05, "loss": 1.3483, "step": 1407 }, { "epoch": 0.09539941730469544, "grad_norm": 12.477338790893555, "learning_rate": 9.910055445273462e-05, "loss": 1.4798, "step": 1408 }, { "epoch": 0.09546717257266753, "grad_norm": 11.233193397521973, "learning_rate": 9.90991854336368e-05, "loss": 1.0304, "step": 1409 }, { "epoch": 0.0955349278406396, "grad_norm": 12.586124420166016, "learning_rate": 9.909781641453899e-05, "loss": 1.2216, "step": 1410 }, { "epoch": 0.0956026831086117, "grad_norm": 12.974738121032715, "learning_rate": 9.909644739544117e-05, "loss": 1.1495, "step": 1411 }, { "epoch": 0.09567043837658377, "grad_norm": 9.613628387451172, "learning_rate": 9.909507837634335e-05, "loss": 0.8326, "step": 1412 }, { "epoch": 0.09573819364455587, "grad_norm": 10.644312858581543, "learning_rate": 9.909370935724553e-05, "loss": 1.0182, "step": 1413 }, { "epoch": 0.09580594891252794, "grad_norm": 11.155874252319336, "learning_rate": 9.909234033814773e-05, "loss": 1.2887, "step": 1414 }, { "epoch": 0.09587370418050004, "grad_norm": 12.068909645080566, "learning_rate": 9.90909713190499e-05, "loss": 1.1697, "step": 1415 }, { "epoch": 0.09594145944847211, "grad_norm": 10.66831111907959, "learning_rate": 9.908960229995209e-05, "loss": 1.1275, "step": 1416 }, { "epoch": 0.0960092147164442, "grad_norm": 11.80036449432373, "learning_rate": 9.908823328085427e-05, "loss": 1.38, "step": 1417 }, { "epoch": 0.09607696998441628, "grad_norm": 11.677534103393555, "learning_rate": 9.908686426175645e-05, "loss": 1.075, "step": 1418 }, { "epoch": 0.09614472525238837, "grad_norm": 10.54027271270752, "learning_rate": 9.908549524265864e-05, "loss": 0.9617, "step": 1419 }, { "epoch": 0.09621248052036045, "grad_norm": 9.70718002319336, "learning_rate": 9.908412622356082e-05, "loss": 1.0395, "step": 1420 }, { "epoch": 0.09628023578833254, "grad_norm": 10.439559936523438, "learning_rate": 9.9082757204463e-05, "loss": 1.4112, "step": 1421 }, { "epoch": 0.09634799105630462, "grad_norm": 9.328675270080566, "learning_rate": 9.908138818536518e-05, "loss": 1.0481, "step": 1422 }, { "epoch": 0.09641574632427671, "grad_norm": 12.834508895874023, "learning_rate": 9.908001916626738e-05, "loss": 1.0863, "step": 1423 }, { "epoch": 0.09648350159224879, "grad_norm": 11.885201454162598, "learning_rate": 9.907865014716956e-05, "loss": 1.3509, "step": 1424 }, { "epoch": 0.09655125686022088, "grad_norm": 11.299174308776855, "learning_rate": 9.907728112807174e-05, "loss": 1.1328, "step": 1425 }, { "epoch": 0.09661901212819296, "grad_norm": 13.024226188659668, "learning_rate": 9.907591210897392e-05, "loss": 1.1495, "step": 1426 }, { "epoch": 0.09668676739616505, "grad_norm": 13.418682098388672, "learning_rate": 9.907454308987611e-05, "loss": 1.3827, "step": 1427 }, { "epoch": 0.09675452266413713, "grad_norm": 11.28375244140625, "learning_rate": 9.907317407077829e-05, "loss": 1.3658, "step": 1428 }, { "epoch": 0.09682227793210922, "grad_norm": 9.711199760437012, "learning_rate": 9.907180505168047e-05, "loss": 0.8743, "step": 1429 }, { "epoch": 0.09689003320008131, "grad_norm": 12.292948722839355, "learning_rate": 9.907043603258267e-05, "loss": 1.203, "step": 1430 }, { "epoch": 0.09695778846805339, "grad_norm": 13.195072174072266, "learning_rate": 9.906906701348485e-05, "loss": 1.0403, "step": 1431 }, { "epoch": 0.09702554373602548, "grad_norm": 11.45721435546875, "learning_rate": 9.906769799438703e-05, "loss": 1.2075, "step": 1432 }, { "epoch": 0.09709329900399756, "grad_norm": 10.477989196777344, "learning_rate": 9.906632897528922e-05, "loss": 1.1646, "step": 1433 }, { "epoch": 0.09716105427196965, "grad_norm": 12.572269439697266, "learning_rate": 9.90649599561914e-05, "loss": 1.3944, "step": 1434 }, { "epoch": 0.09722880953994173, "grad_norm": 9.37205982208252, "learning_rate": 9.906359093709358e-05, "loss": 0.9902, "step": 1435 }, { "epoch": 0.09729656480791382, "grad_norm": 11.590779304504395, "learning_rate": 9.906222191799576e-05, "loss": 1.2658, "step": 1436 }, { "epoch": 0.0973643200758859, "grad_norm": 10.35207748413086, "learning_rate": 9.906085289889795e-05, "loss": 0.7905, "step": 1437 }, { "epoch": 0.09743207534385799, "grad_norm": 9.993937492370605, "learning_rate": 9.905948387980013e-05, "loss": 1.1522, "step": 1438 }, { "epoch": 0.09749983061183007, "grad_norm": 9.865569114685059, "learning_rate": 9.905811486070232e-05, "loss": 1.2536, "step": 1439 }, { "epoch": 0.09756758587980216, "grad_norm": 12.836588859558105, "learning_rate": 9.90567458416045e-05, "loss": 1.2216, "step": 1440 }, { "epoch": 0.09763534114777424, "grad_norm": 10.062298774719238, "learning_rate": 9.905537682250668e-05, "loss": 1.1576, "step": 1441 }, { "epoch": 0.09770309641574633, "grad_norm": 10.897071838378906, "learning_rate": 9.905400780340887e-05, "loss": 1.1315, "step": 1442 }, { "epoch": 0.09777085168371841, "grad_norm": 10.366122245788574, "learning_rate": 9.905263878431105e-05, "loss": 1.274, "step": 1443 }, { "epoch": 0.0978386069516905, "grad_norm": 11.632966995239258, "learning_rate": 9.905126976521323e-05, "loss": 1.1427, "step": 1444 }, { "epoch": 0.09790636221966258, "grad_norm": 10.537737846374512, "learning_rate": 9.904990074611541e-05, "loss": 1.0913, "step": 1445 }, { "epoch": 0.09797411748763467, "grad_norm": 9.52363109588623, "learning_rate": 9.90485317270176e-05, "loss": 0.8677, "step": 1446 }, { "epoch": 0.09804187275560675, "grad_norm": 11.511491775512695, "learning_rate": 9.904716270791979e-05, "loss": 1.0381, "step": 1447 }, { "epoch": 0.09810962802357884, "grad_norm": 12.085793495178223, "learning_rate": 9.904579368882197e-05, "loss": 1.1415, "step": 1448 }, { "epoch": 0.09817738329155092, "grad_norm": 8.665430068969727, "learning_rate": 9.904442466972415e-05, "loss": 1.0367, "step": 1449 }, { "epoch": 0.09824513855952301, "grad_norm": 10.900618553161621, "learning_rate": 9.904305565062633e-05, "loss": 0.9835, "step": 1450 }, { "epoch": 0.09831289382749508, "grad_norm": 10.3113431930542, "learning_rate": 9.904168663152852e-05, "loss": 0.924, "step": 1451 }, { "epoch": 0.09838064909546718, "grad_norm": 10.001591682434082, "learning_rate": 9.90403176124307e-05, "loss": 1.0835, "step": 1452 }, { "epoch": 0.09844840436343925, "grad_norm": 11.333273887634277, "learning_rate": 9.903894859333288e-05, "loss": 1.072, "step": 1453 }, { "epoch": 0.09851615963141135, "grad_norm": 10.107904434204102, "learning_rate": 9.903757957423506e-05, "loss": 1.0848, "step": 1454 }, { "epoch": 0.09858391489938342, "grad_norm": 12.578730583190918, "learning_rate": 9.903621055513725e-05, "loss": 1.2735, "step": 1455 }, { "epoch": 0.09865167016735552, "grad_norm": 10.453478813171387, "learning_rate": 9.903484153603944e-05, "loss": 1.3141, "step": 1456 }, { "epoch": 0.09871942543532759, "grad_norm": 10.383566856384277, "learning_rate": 9.903347251694162e-05, "loss": 1.0992, "step": 1457 }, { "epoch": 0.09878718070329968, "grad_norm": 9.612902641296387, "learning_rate": 9.90321034978438e-05, "loss": 1.3103, "step": 1458 }, { "epoch": 0.09885493597127176, "grad_norm": 12.111359596252441, "learning_rate": 9.903073447874598e-05, "loss": 1.3102, "step": 1459 }, { "epoch": 0.09892269123924385, "grad_norm": 9.987195014953613, "learning_rate": 9.902936545964817e-05, "loss": 1.1961, "step": 1460 }, { "epoch": 0.09899044650721593, "grad_norm": 10.900408744812012, "learning_rate": 9.902799644055035e-05, "loss": 1.0208, "step": 1461 }, { "epoch": 0.09905820177518802, "grad_norm": 9.94915771484375, "learning_rate": 9.902662742145253e-05, "loss": 1.3347, "step": 1462 }, { "epoch": 0.0991259570431601, "grad_norm": 13.393661499023438, "learning_rate": 9.902525840235471e-05, "loss": 1.2903, "step": 1463 }, { "epoch": 0.09919371231113219, "grad_norm": 10.122967720031738, "learning_rate": 9.902388938325689e-05, "loss": 1.0229, "step": 1464 }, { "epoch": 0.09926146757910427, "grad_norm": 10.775031089782715, "learning_rate": 9.902252036415909e-05, "loss": 1.1669, "step": 1465 }, { "epoch": 0.09932922284707636, "grad_norm": 9.733497619628906, "learning_rate": 9.902115134506127e-05, "loss": 1.0903, "step": 1466 }, { "epoch": 0.09939697811504844, "grad_norm": 9.230277061462402, "learning_rate": 9.901978232596345e-05, "loss": 1.1738, "step": 1467 }, { "epoch": 0.09946473338302053, "grad_norm": 10.822884559631348, "learning_rate": 9.901841330686563e-05, "loss": 1.3392, "step": 1468 }, { "epoch": 0.09953248865099261, "grad_norm": 10.64195442199707, "learning_rate": 9.901704428776782e-05, "loss": 1.0823, "step": 1469 }, { "epoch": 0.0996002439189647, "grad_norm": 13.73645305633545, "learning_rate": 9.901567526867e-05, "loss": 1.1128, "step": 1470 }, { "epoch": 0.09966799918693678, "grad_norm": 11.361958503723145, "learning_rate": 9.901430624957218e-05, "loss": 1.083, "step": 1471 }, { "epoch": 0.09973575445490887, "grad_norm": 10.839045524597168, "learning_rate": 9.901293723047436e-05, "loss": 1.0288, "step": 1472 }, { "epoch": 0.09980350972288095, "grad_norm": 10.41995906829834, "learning_rate": 9.901156821137656e-05, "loss": 1.0631, "step": 1473 }, { "epoch": 0.09987126499085304, "grad_norm": 11.87709903717041, "learning_rate": 9.901019919227874e-05, "loss": 1.0715, "step": 1474 }, { "epoch": 0.09993902025882512, "grad_norm": 10.46670913696289, "learning_rate": 9.900883017318092e-05, "loss": 1.1684, "step": 1475 }, { "epoch": 0.10000677552679721, "grad_norm": 12.163457870483398, "learning_rate": 9.900746115408311e-05, "loss": 1.4416, "step": 1476 }, { "epoch": 0.10007453079476929, "grad_norm": 13.417581558227539, "learning_rate": 9.900609213498529e-05, "loss": 1.1876, "step": 1477 }, { "epoch": 0.10014228606274138, "grad_norm": 11.35722541809082, "learning_rate": 9.900472311588747e-05, "loss": 1.1389, "step": 1478 }, { "epoch": 0.10021004133071346, "grad_norm": 10.042820930480957, "learning_rate": 9.900335409678966e-05, "loss": 1.2156, "step": 1479 }, { "epoch": 0.10027779659868555, "grad_norm": 10.823782920837402, "learning_rate": 9.900198507769184e-05, "loss": 1.1915, "step": 1480 }, { "epoch": 0.10034555186665763, "grad_norm": 13.6808443069458, "learning_rate": 9.900061605859403e-05, "loss": 1.1724, "step": 1481 }, { "epoch": 0.10041330713462972, "grad_norm": 11.16846752166748, "learning_rate": 9.89992470394962e-05, "loss": 1.2093, "step": 1482 }, { "epoch": 0.10048106240260181, "grad_norm": 10.391450881958008, "learning_rate": 9.89978780203984e-05, "loss": 1.142, "step": 1483 }, { "epoch": 0.10054881767057389, "grad_norm": 9.324288368225098, "learning_rate": 9.899650900130058e-05, "loss": 1.0579, "step": 1484 }, { "epoch": 0.10061657293854598, "grad_norm": 12.601625442504883, "learning_rate": 9.899513998220276e-05, "loss": 1.2643, "step": 1485 }, { "epoch": 0.10068432820651806, "grad_norm": 14.270779609680176, "learning_rate": 9.899377096310494e-05, "loss": 1.2903, "step": 1486 }, { "epoch": 0.10075208347449015, "grad_norm": 11.521232604980469, "learning_rate": 9.899240194400712e-05, "loss": 1.2385, "step": 1487 }, { "epoch": 0.10081983874246223, "grad_norm": 10.76693344116211, "learning_rate": 9.899103292490931e-05, "loss": 1.1745, "step": 1488 }, { "epoch": 0.10088759401043432, "grad_norm": 9.091184616088867, "learning_rate": 9.89896639058115e-05, "loss": 0.9571, "step": 1489 }, { "epoch": 0.1009553492784064, "grad_norm": 11.930106163024902, "learning_rate": 9.898829488671368e-05, "loss": 1.1226, "step": 1490 }, { "epoch": 0.10102310454637849, "grad_norm": 10.90937614440918, "learning_rate": 9.898692586761586e-05, "loss": 1.0776, "step": 1491 }, { "epoch": 0.10109085981435056, "grad_norm": 10.618545532226562, "learning_rate": 9.898555684851805e-05, "loss": 1.1251, "step": 1492 }, { "epoch": 0.10115861508232266, "grad_norm": 10.228861808776855, "learning_rate": 9.898418782942023e-05, "loss": 1.1987, "step": 1493 }, { "epoch": 0.10122637035029473, "grad_norm": 8.807862281799316, "learning_rate": 9.898281881032241e-05, "loss": 0.9339, "step": 1494 }, { "epoch": 0.10129412561826683, "grad_norm": 11.24593448638916, "learning_rate": 9.898144979122459e-05, "loss": 1.1495, "step": 1495 }, { "epoch": 0.1013618808862389, "grad_norm": 11.192438125610352, "learning_rate": 9.898008077212677e-05, "loss": 1.1361, "step": 1496 }, { "epoch": 0.101429636154211, "grad_norm": 10.440075874328613, "learning_rate": 9.897871175302896e-05, "loss": 1.2273, "step": 1497 }, { "epoch": 0.10149739142218307, "grad_norm": 11.103675842285156, "learning_rate": 9.897734273393115e-05, "loss": 1.1308, "step": 1498 }, { "epoch": 0.10156514669015516, "grad_norm": 9.78297233581543, "learning_rate": 9.897597371483333e-05, "loss": 1.0907, "step": 1499 }, { "epoch": 0.10163290195812724, "grad_norm": 10.98086166381836, "learning_rate": 9.89746046957355e-05, "loss": 0.9743, "step": 1500 }, { "epoch": 0.10170065722609933, "grad_norm": 9.268783569335938, "learning_rate": 9.89732356766377e-05, "loss": 0.8917, "step": 1501 }, { "epoch": 0.10176841249407141, "grad_norm": 12.674605369567871, "learning_rate": 9.897186665753988e-05, "loss": 1.2247, "step": 1502 }, { "epoch": 0.1018361677620435, "grad_norm": 10.987565040588379, "learning_rate": 9.897049763844206e-05, "loss": 1.0688, "step": 1503 }, { "epoch": 0.10190392303001558, "grad_norm": 16.014053344726562, "learning_rate": 9.896912861934424e-05, "loss": 0.8334, "step": 1504 }, { "epoch": 0.10197167829798767, "grad_norm": 11.119991302490234, "learning_rate": 9.896775960024642e-05, "loss": 1.3176, "step": 1505 }, { "epoch": 0.10203943356595975, "grad_norm": 13.23279094696045, "learning_rate": 9.896639058114861e-05, "loss": 1.1625, "step": 1506 }, { "epoch": 0.10210718883393184, "grad_norm": 9.3678560256958, "learning_rate": 9.89650215620508e-05, "loss": 1.2774, "step": 1507 }, { "epoch": 0.10217494410190392, "grad_norm": 10.829100608825684, "learning_rate": 9.896365254295298e-05, "loss": 1.2423, "step": 1508 }, { "epoch": 0.10224269936987601, "grad_norm": 12.12694263458252, "learning_rate": 9.896228352385516e-05, "loss": 1.2276, "step": 1509 }, { "epoch": 0.10231045463784809, "grad_norm": 11.626548767089844, "learning_rate": 9.896091450475735e-05, "loss": 1.0871, "step": 1510 }, { "epoch": 0.10237820990582018, "grad_norm": 11.388608932495117, "learning_rate": 9.895954548565953e-05, "loss": 1.5582, "step": 1511 }, { "epoch": 0.10244596517379226, "grad_norm": 9.463730812072754, "learning_rate": 9.895817646656171e-05, "loss": 1.1066, "step": 1512 }, { "epoch": 0.10251372044176435, "grad_norm": 10.291573524475098, "learning_rate": 9.895680744746389e-05, "loss": 1.4026, "step": 1513 }, { "epoch": 0.10258147570973643, "grad_norm": 9.778963088989258, "learning_rate": 9.895543842836607e-05, "loss": 1.1109, "step": 1514 }, { "epoch": 0.10264923097770852, "grad_norm": 9.685966491699219, "learning_rate": 9.895406940926827e-05, "loss": 1.1633, "step": 1515 }, { "epoch": 0.1027169862456806, "grad_norm": 10.76310920715332, "learning_rate": 9.895270039017045e-05, "loss": 1.0813, "step": 1516 }, { "epoch": 0.10278474151365269, "grad_norm": 9.795347213745117, "learning_rate": 9.895133137107263e-05, "loss": 1.2079, "step": 1517 }, { "epoch": 0.10285249678162477, "grad_norm": 9.980990409851074, "learning_rate": 9.89499623519748e-05, "loss": 1.0701, "step": 1518 }, { "epoch": 0.10292025204959686, "grad_norm": 9.682209014892578, "learning_rate": 9.894859333287699e-05, "loss": 1.4133, "step": 1519 }, { "epoch": 0.10298800731756894, "grad_norm": 10.632065773010254, "learning_rate": 9.894722431377918e-05, "loss": 1.0631, "step": 1520 }, { "epoch": 0.10305576258554103, "grad_norm": 10.099474906921387, "learning_rate": 9.894585529468136e-05, "loss": 1.2015, "step": 1521 }, { "epoch": 0.1031235178535131, "grad_norm": 8.289199829101562, "learning_rate": 9.894448627558354e-05, "loss": 1.3009, "step": 1522 }, { "epoch": 0.1031912731214852, "grad_norm": 9.403796195983887, "learning_rate": 9.894311725648573e-05, "loss": 1.113, "step": 1523 }, { "epoch": 0.10325902838945727, "grad_norm": 12.6613130569458, "learning_rate": 9.894174823738792e-05, "loss": 0.9394, "step": 1524 }, { "epoch": 0.10332678365742937, "grad_norm": 9.85255241394043, "learning_rate": 9.894037921829011e-05, "loss": 1.1007, "step": 1525 }, { "epoch": 0.10339453892540144, "grad_norm": 11.918173789978027, "learning_rate": 9.893901019919229e-05, "loss": 1.1817, "step": 1526 }, { "epoch": 0.10346229419337354, "grad_norm": 9.994447708129883, "learning_rate": 9.893764118009447e-05, "loss": 1.2522, "step": 1527 }, { "epoch": 0.10353004946134561, "grad_norm": 9.879289627075195, "learning_rate": 9.893627216099665e-05, "loss": 1.152, "step": 1528 }, { "epoch": 0.1035978047293177, "grad_norm": 10.103482246398926, "learning_rate": 9.893490314189884e-05, "loss": 1.2227, "step": 1529 }, { "epoch": 0.10366555999728978, "grad_norm": 11.173476219177246, "learning_rate": 9.893353412280102e-05, "loss": 1.2188, "step": 1530 }, { "epoch": 0.10373331526526187, "grad_norm": 11.540877342224121, "learning_rate": 9.89321651037032e-05, "loss": 1.1669, "step": 1531 }, { "epoch": 0.10380107053323395, "grad_norm": 10.706154823303223, "learning_rate": 9.893079608460539e-05, "loss": 1.2771, "step": 1532 }, { "epoch": 0.10386882580120604, "grad_norm": 11.781739234924316, "learning_rate": 9.892942706550758e-05, "loss": 1.0501, "step": 1533 }, { "epoch": 0.10393658106917812, "grad_norm": 10.482099533081055, "learning_rate": 9.892805804640976e-05, "loss": 1.2327, "step": 1534 }, { "epoch": 0.10400433633715021, "grad_norm": 12.450867652893066, "learning_rate": 9.892668902731194e-05, "loss": 1.3575, "step": 1535 }, { "epoch": 0.1040720916051223, "grad_norm": 10.236811637878418, "learning_rate": 9.892532000821412e-05, "loss": 1.2559, "step": 1536 }, { "epoch": 0.10413984687309438, "grad_norm": 9.311124801635742, "learning_rate": 9.89239509891163e-05, "loss": 1.1143, "step": 1537 }, { "epoch": 0.10420760214106647, "grad_norm": 9.182706832885742, "learning_rate": 9.89225819700185e-05, "loss": 0.8875, "step": 1538 }, { "epoch": 0.10427535740903855, "grad_norm": 12.762700080871582, "learning_rate": 9.892121295092067e-05, "loss": 1.0552, "step": 1539 }, { "epoch": 0.10434311267701064, "grad_norm": 12.222203254699707, "learning_rate": 9.891984393182285e-05, "loss": 1.3185, "step": 1540 }, { "epoch": 0.10441086794498272, "grad_norm": 11.45807933807373, "learning_rate": 9.891847491272504e-05, "loss": 1.2952, "step": 1541 }, { "epoch": 0.10447862321295481, "grad_norm": 10.445068359375, "learning_rate": 9.891710589362722e-05, "loss": 1.1637, "step": 1542 }, { "epoch": 0.10454637848092689, "grad_norm": 11.758063316345215, "learning_rate": 9.891573687452941e-05, "loss": 1.3202, "step": 1543 }, { "epoch": 0.10461413374889898, "grad_norm": 10.176533699035645, "learning_rate": 9.891436785543159e-05, "loss": 1.1762, "step": 1544 }, { "epoch": 0.10468188901687106, "grad_norm": 10.133155822753906, "learning_rate": 9.891299883633377e-05, "loss": 1.2226, "step": 1545 }, { "epoch": 0.10474964428484315, "grad_norm": 9.883895874023438, "learning_rate": 9.891162981723595e-05, "loss": 0.9687, "step": 1546 }, { "epoch": 0.10481739955281523, "grad_norm": 13.175050735473633, "learning_rate": 9.891026079813814e-05, "loss": 1.2075, "step": 1547 }, { "epoch": 0.10488515482078732, "grad_norm": 11.183597564697266, "learning_rate": 9.890889177904032e-05, "loss": 1.0917, "step": 1548 }, { "epoch": 0.1049529100887594, "grad_norm": 10.135035514831543, "learning_rate": 9.89075227599425e-05, "loss": 1.0087, "step": 1549 }, { "epoch": 0.10502066535673149, "grad_norm": 16.02760887145996, "learning_rate": 9.890615374084469e-05, "loss": 1.2015, "step": 1550 }, { "epoch": 0.10508842062470357, "grad_norm": 11.255363464355469, "learning_rate": 9.890478472174687e-05, "loss": 1.1106, "step": 1551 }, { "epoch": 0.10515617589267566, "grad_norm": 10.740998268127441, "learning_rate": 9.890341570264906e-05, "loss": 1.1835, "step": 1552 }, { "epoch": 0.10522393116064774, "grad_norm": 11.84919548034668, "learning_rate": 9.890204668355124e-05, "loss": 1.1655, "step": 1553 }, { "epoch": 0.10529168642861983, "grad_norm": 12.041108131408691, "learning_rate": 9.890067766445342e-05, "loss": 1.2544, "step": 1554 }, { "epoch": 0.10535944169659191, "grad_norm": 8.966646194458008, "learning_rate": 9.88993086453556e-05, "loss": 1.098, "step": 1555 }, { "epoch": 0.105427196964564, "grad_norm": 11.838338851928711, "learning_rate": 9.88979396262578e-05, "loss": 1.1774, "step": 1556 }, { "epoch": 0.10549495223253608, "grad_norm": 11.186326026916504, "learning_rate": 9.889657060715997e-05, "loss": 1.1778, "step": 1557 }, { "epoch": 0.10556270750050817, "grad_norm": 9.448702812194824, "learning_rate": 9.889520158806216e-05, "loss": 1.0171, "step": 1558 }, { "epoch": 0.10563046276848025, "grad_norm": 10.931096076965332, "learning_rate": 9.889383256896434e-05, "loss": 1.1797, "step": 1559 }, { "epoch": 0.10569821803645234, "grad_norm": 10.293981552124023, "learning_rate": 9.889246354986652e-05, "loss": 1.0723, "step": 1560 }, { "epoch": 0.10576597330442442, "grad_norm": 10.052331924438477, "learning_rate": 9.889109453076871e-05, "loss": 0.831, "step": 1561 }, { "epoch": 0.10583372857239651, "grad_norm": 10.593210220336914, "learning_rate": 9.888972551167089e-05, "loss": 1.3415, "step": 1562 }, { "epoch": 0.10590148384036858, "grad_norm": 11.580954551696777, "learning_rate": 9.888835649257307e-05, "loss": 1.1522, "step": 1563 }, { "epoch": 0.10596923910834068, "grad_norm": 11.495551109313965, "learning_rate": 9.888698747347525e-05, "loss": 1.2515, "step": 1564 }, { "epoch": 0.10603699437631275, "grad_norm": 10.543874740600586, "learning_rate": 9.888561845437743e-05, "loss": 1.1579, "step": 1565 }, { "epoch": 0.10610474964428485, "grad_norm": 10.588164329528809, "learning_rate": 9.888424943527963e-05, "loss": 1.3825, "step": 1566 }, { "epoch": 0.10617250491225692, "grad_norm": 13.591666221618652, "learning_rate": 9.88828804161818e-05, "loss": 1.0941, "step": 1567 }, { "epoch": 0.10624026018022902, "grad_norm": 10.866951942443848, "learning_rate": 9.888151139708399e-05, "loss": 1.478, "step": 1568 }, { "epoch": 0.10630801544820109, "grad_norm": 9.627554893493652, "learning_rate": 9.888014237798618e-05, "loss": 1.004, "step": 1569 }, { "epoch": 0.10637577071617318, "grad_norm": 10.871118545532227, "learning_rate": 9.887877335888836e-05, "loss": 1.1997, "step": 1570 }, { "epoch": 0.10644352598414526, "grad_norm": 10.74503231048584, "learning_rate": 9.887740433979054e-05, "loss": 1.005, "step": 1571 }, { "epoch": 0.10651128125211735, "grad_norm": 11.119452476501465, "learning_rate": 9.887603532069273e-05, "loss": 1.1168, "step": 1572 }, { "epoch": 0.10657903652008943, "grad_norm": 10.601544380187988, "learning_rate": 9.887466630159491e-05, "loss": 1.0256, "step": 1573 }, { "epoch": 0.10664679178806152, "grad_norm": 10.329113006591797, "learning_rate": 9.88732972824971e-05, "loss": 1.3432, "step": 1574 }, { "epoch": 0.1067145470560336, "grad_norm": 9.973999977111816, "learning_rate": 9.887192826339929e-05, "loss": 0.9961, "step": 1575 }, { "epoch": 0.10678230232400569, "grad_norm": 10.981974601745605, "learning_rate": 9.887055924430147e-05, "loss": 1.3936, "step": 1576 }, { "epoch": 0.10685005759197777, "grad_norm": 10.953417778015137, "learning_rate": 9.886919022520365e-05, "loss": 0.9669, "step": 1577 }, { "epoch": 0.10691781285994986, "grad_norm": 10.656323432922363, "learning_rate": 9.886782120610583e-05, "loss": 1.1393, "step": 1578 }, { "epoch": 0.10698556812792194, "grad_norm": 9.434617042541504, "learning_rate": 9.886645218700802e-05, "loss": 1.1813, "step": 1579 }, { "epoch": 0.10705332339589403, "grad_norm": 12.334831237792969, "learning_rate": 9.88650831679102e-05, "loss": 1.453, "step": 1580 }, { "epoch": 0.10712107866386611, "grad_norm": 9.892403602600098, "learning_rate": 9.886371414881238e-05, "loss": 1.1373, "step": 1581 }, { "epoch": 0.1071888339318382, "grad_norm": 9.248678207397461, "learning_rate": 9.886234512971456e-05, "loss": 1.0734, "step": 1582 }, { "epoch": 0.10725658919981028, "grad_norm": 10.317010879516602, "learning_rate": 9.886097611061675e-05, "loss": 1.1252, "step": 1583 }, { "epoch": 0.10732434446778237, "grad_norm": 9.586435317993164, "learning_rate": 9.885960709151894e-05, "loss": 1.1705, "step": 1584 }, { "epoch": 0.10739209973575445, "grad_norm": 8.883166313171387, "learning_rate": 9.885823807242112e-05, "loss": 1.3164, "step": 1585 }, { "epoch": 0.10745985500372654, "grad_norm": 12.870014190673828, "learning_rate": 9.88568690533233e-05, "loss": 1.2187, "step": 1586 }, { "epoch": 0.10752761027169863, "grad_norm": 10.858057975769043, "learning_rate": 9.885550003422548e-05, "loss": 1.2334, "step": 1587 }, { "epoch": 0.10759536553967071, "grad_norm": 9.855050086975098, "learning_rate": 9.885413101512767e-05, "loss": 1.068, "step": 1588 }, { "epoch": 0.1076631208076428, "grad_norm": 12.256099700927734, "learning_rate": 9.885276199602985e-05, "loss": 1.2481, "step": 1589 }, { "epoch": 0.10773087607561488, "grad_norm": 9.756118774414062, "learning_rate": 9.885139297693203e-05, "loss": 1.2365, "step": 1590 }, { "epoch": 0.10779863134358697, "grad_norm": 11.727996826171875, "learning_rate": 9.885002395783421e-05, "loss": 1.3839, "step": 1591 }, { "epoch": 0.10786638661155905, "grad_norm": 11.58463191986084, "learning_rate": 9.88486549387364e-05, "loss": 0.937, "step": 1592 }, { "epoch": 0.10793414187953114, "grad_norm": 12.663015365600586, "learning_rate": 9.884728591963859e-05, "loss": 1.3453, "step": 1593 }, { "epoch": 0.10800189714750322, "grad_norm": 10.950531005859375, "learning_rate": 9.884591690054077e-05, "loss": 1.1345, "step": 1594 }, { "epoch": 0.10806965241547531, "grad_norm": 10.361424446105957, "learning_rate": 9.884454788144295e-05, "loss": 1.2576, "step": 1595 }, { "epoch": 0.10813740768344739, "grad_norm": 11.155044555664062, "learning_rate": 9.884317886234513e-05, "loss": 1.1253, "step": 1596 }, { "epoch": 0.10820516295141948, "grad_norm": 12.457600593566895, "learning_rate": 9.884180984324731e-05, "loss": 1.0959, "step": 1597 }, { "epoch": 0.10827291821939156, "grad_norm": 13.494983673095703, "learning_rate": 9.88404408241495e-05, "loss": 1.1671, "step": 1598 }, { "epoch": 0.10834067348736365, "grad_norm": 11.870648384094238, "learning_rate": 9.883907180505168e-05, "loss": 1.2121, "step": 1599 }, { "epoch": 0.10840842875533573, "grad_norm": 10.327920913696289, "learning_rate": 9.883770278595387e-05, "loss": 1.1319, "step": 1600 }, { "epoch": 0.10847618402330782, "grad_norm": 9.578879356384277, "learning_rate": 9.883633376685605e-05, "loss": 0.9064, "step": 1601 }, { "epoch": 0.1085439392912799, "grad_norm": 9.598779678344727, "learning_rate": 9.883496474775824e-05, "loss": 1.1119, "step": 1602 }, { "epoch": 0.10861169455925199, "grad_norm": 13.522628784179688, "learning_rate": 9.883359572866042e-05, "loss": 1.2549, "step": 1603 }, { "epoch": 0.10867944982722406, "grad_norm": 8.123127937316895, "learning_rate": 9.88322267095626e-05, "loss": 0.8778, "step": 1604 }, { "epoch": 0.10874720509519616, "grad_norm": 13.100455284118652, "learning_rate": 9.883085769046478e-05, "loss": 1.2023, "step": 1605 }, { "epoch": 0.10881496036316823, "grad_norm": 9.897802352905273, "learning_rate": 9.882948867136696e-05, "loss": 1.2094, "step": 1606 }, { "epoch": 0.10888271563114033, "grad_norm": 8.55170726776123, "learning_rate": 9.882811965226915e-05, "loss": 1.0407, "step": 1607 }, { "epoch": 0.1089504708991124, "grad_norm": 10.867438316345215, "learning_rate": 9.882675063317133e-05, "loss": 1.3021, "step": 1608 }, { "epoch": 0.1090182261670845, "grad_norm": 11.062238693237305, "learning_rate": 9.882538161407352e-05, "loss": 1.2627, "step": 1609 }, { "epoch": 0.10908598143505657, "grad_norm": 12.710458755493164, "learning_rate": 9.88240125949757e-05, "loss": 1.6196, "step": 1610 }, { "epoch": 0.10915373670302866, "grad_norm": 10.986543655395508, "learning_rate": 9.882264357587789e-05, "loss": 1.2021, "step": 1611 }, { "epoch": 0.10922149197100074, "grad_norm": 12.792850494384766, "learning_rate": 9.882127455678007e-05, "loss": 1.1246, "step": 1612 }, { "epoch": 0.10928924723897283, "grad_norm": 12.120078086853027, "learning_rate": 9.881990553768225e-05, "loss": 1.2522, "step": 1613 }, { "epoch": 0.10935700250694491, "grad_norm": 10.562193870544434, "learning_rate": 9.881853651858443e-05, "loss": 1.1791, "step": 1614 }, { "epoch": 0.109424757774917, "grad_norm": 13.017698287963867, "learning_rate": 9.881716749948662e-05, "loss": 1.0733, "step": 1615 }, { "epoch": 0.10949251304288908, "grad_norm": 10.904980659484863, "learning_rate": 9.88157984803888e-05, "loss": 1.0166, "step": 1616 }, { "epoch": 0.10956026831086117, "grad_norm": 12.332829475402832, "learning_rate": 9.881442946129099e-05, "loss": 1.0688, "step": 1617 }, { "epoch": 0.10962802357883325, "grad_norm": 12.238106727600098, "learning_rate": 9.881306044219318e-05, "loss": 0.9869, "step": 1618 }, { "epoch": 0.10969577884680534, "grad_norm": 12.75059986114502, "learning_rate": 9.881169142309536e-05, "loss": 1.5365, "step": 1619 }, { "epoch": 0.10976353411477742, "grad_norm": 11.019241333007812, "learning_rate": 9.881032240399754e-05, "loss": 1.2332, "step": 1620 }, { "epoch": 0.10983128938274951, "grad_norm": 11.867498397827148, "learning_rate": 9.880895338489973e-05, "loss": 1.1955, "step": 1621 }, { "epoch": 0.10989904465072159, "grad_norm": 9.685249328613281, "learning_rate": 9.880758436580191e-05, "loss": 1.1972, "step": 1622 }, { "epoch": 0.10996679991869368, "grad_norm": 11.80540657043457, "learning_rate": 9.88062153467041e-05, "loss": 0.9501, "step": 1623 }, { "epoch": 0.11003455518666576, "grad_norm": 9.719080924987793, "learning_rate": 9.880484632760627e-05, "loss": 0.9904, "step": 1624 }, { "epoch": 0.11010231045463785, "grad_norm": 10.906637191772461, "learning_rate": 9.880347730850847e-05, "loss": 1.2807, "step": 1625 }, { "epoch": 0.11017006572260993, "grad_norm": 10.40942668914795, "learning_rate": 9.880210828941065e-05, "loss": 1.2772, "step": 1626 }, { "epoch": 0.11023782099058202, "grad_norm": 10.265239715576172, "learning_rate": 9.880073927031283e-05, "loss": 1.1583, "step": 1627 }, { "epoch": 0.1103055762585541, "grad_norm": 12.158036231994629, "learning_rate": 9.879937025121501e-05, "loss": 1.3907, "step": 1628 }, { "epoch": 0.11037333152652619, "grad_norm": 10.128400802612305, "learning_rate": 9.879800123211719e-05, "loss": 0.8838, "step": 1629 }, { "epoch": 0.11044108679449827, "grad_norm": 9.536952018737793, "learning_rate": 9.879663221301938e-05, "loss": 1.0525, "step": 1630 }, { "epoch": 0.11050884206247036, "grad_norm": 10.826987266540527, "learning_rate": 9.879526319392156e-05, "loss": 1.2005, "step": 1631 }, { "epoch": 0.11057659733044244, "grad_norm": 10.08153247833252, "learning_rate": 9.879389417482374e-05, "loss": 1.1521, "step": 1632 }, { "epoch": 0.11064435259841453, "grad_norm": 10.642118453979492, "learning_rate": 9.879252515572592e-05, "loss": 1.2316, "step": 1633 }, { "epoch": 0.1107121078663866, "grad_norm": 9.595836639404297, "learning_rate": 9.879115613662812e-05, "loss": 1.0969, "step": 1634 }, { "epoch": 0.1107798631343587, "grad_norm": 8.954519271850586, "learning_rate": 9.87897871175303e-05, "loss": 0.9422, "step": 1635 }, { "epoch": 0.11084761840233077, "grad_norm": 9.143898010253906, "learning_rate": 9.878841809843248e-05, "loss": 0.8949, "step": 1636 }, { "epoch": 0.11091537367030287, "grad_norm": 11.656847953796387, "learning_rate": 9.878704907933466e-05, "loss": 0.9739, "step": 1637 }, { "epoch": 0.11098312893827494, "grad_norm": 11.996994018554688, "learning_rate": 9.878568006023684e-05, "loss": 1.3361, "step": 1638 }, { "epoch": 0.11105088420624704, "grad_norm": 11.478599548339844, "learning_rate": 9.878431104113903e-05, "loss": 1.0726, "step": 1639 }, { "epoch": 0.11111863947421913, "grad_norm": 14.594179153442383, "learning_rate": 9.878294202204121e-05, "loss": 1.3725, "step": 1640 }, { "epoch": 0.1111863947421912, "grad_norm": 12.15703296661377, "learning_rate": 9.87815730029434e-05, "loss": 1.1996, "step": 1641 }, { "epoch": 0.1112541500101633, "grad_norm": 9.384215354919434, "learning_rate": 9.878020398384557e-05, "loss": 1.1294, "step": 1642 }, { "epoch": 0.11132190527813537, "grad_norm": 11.131610870361328, "learning_rate": 9.877883496474777e-05, "loss": 1.0526, "step": 1643 }, { "epoch": 0.11138966054610747, "grad_norm": 9.628592491149902, "learning_rate": 9.877746594564995e-05, "loss": 1.0715, "step": 1644 }, { "epoch": 0.11145741581407954, "grad_norm": 10.876824378967285, "learning_rate": 9.877609692655213e-05, "loss": 1.3337, "step": 1645 }, { "epoch": 0.11152517108205164, "grad_norm": 10.489126205444336, "learning_rate": 9.877472790745431e-05, "loss": 1.1727, "step": 1646 }, { "epoch": 0.11159292635002371, "grad_norm": 8.737699508666992, "learning_rate": 9.877335888835649e-05, "loss": 1.1252, "step": 1647 }, { "epoch": 0.1116606816179958, "grad_norm": 9.8843355178833, "learning_rate": 9.877198986925868e-05, "loss": 0.9222, "step": 1648 }, { "epoch": 0.11172843688596788, "grad_norm": 7.930290699005127, "learning_rate": 9.877062085016086e-05, "loss": 1.0095, "step": 1649 }, { "epoch": 0.11179619215393997, "grad_norm": 9.721529006958008, "learning_rate": 9.876925183106304e-05, "loss": 1.0812, "step": 1650 }, { "epoch": 0.11186394742191205, "grad_norm": 10.844324111938477, "learning_rate": 9.876788281196523e-05, "loss": 1.2245, "step": 1651 }, { "epoch": 0.11193170268988414, "grad_norm": 8.750361442565918, "learning_rate": 9.87665137928674e-05, "loss": 0.9983, "step": 1652 }, { "epoch": 0.11199945795785622, "grad_norm": 8.716521263122559, "learning_rate": 9.87651447737696e-05, "loss": 0.8612, "step": 1653 }, { "epoch": 0.11206721322582831, "grad_norm": 10.06696891784668, "learning_rate": 9.876377575467178e-05, "loss": 1.0806, "step": 1654 }, { "epoch": 0.11213496849380039, "grad_norm": 10.526103019714355, "learning_rate": 9.876240673557396e-05, "loss": 1.0799, "step": 1655 }, { "epoch": 0.11220272376177248, "grad_norm": 11.108748435974121, "learning_rate": 9.876103771647614e-05, "loss": 1.3871, "step": 1656 }, { "epoch": 0.11227047902974456, "grad_norm": 10.120033264160156, "learning_rate": 9.875966869737833e-05, "loss": 1.0399, "step": 1657 }, { "epoch": 0.11233823429771665, "grad_norm": 10.589370727539062, "learning_rate": 9.875829967828051e-05, "loss": 1.3249, "step": 1658 }, { "epoch": 0.11240598956568873, "grad_norm": 8.751884460449219, "learning_rate": 9.87569306591827e-05, "loss": 0.9404, "step": 1659 }, { "epoch": 0.11247374483366082, "grad_norm": 9.923823356628418, "learning_rate": 9.875556164008488e-05, "loss": 1.067, "step": 1660 }, { "epoch": 0.1125415001016329, "grad_norm": 9.70035171508789, "learning_rate": 9.875419262098707e-05, "loss": 1.0079, "step": 1661 }, { "epoch": 0.11260925536960499, "grad_norm": 12.337713241577148, "learning_rate": 9.875282360188925e-05, "loss": 1.0962, "step": 1662 }, { "epoch": 0.11267701063757707, "grad_norm": 10.142355918884277, "learning_rate": 9.875145458279143e-05, "loss": 1.1916, "step": 1663 }, { "epoch": 0.11274476590554916, "grad_norm": 11.82714557647705, "learning_rate": 9.875008556369362e-05, "loss": 1.4582, "step": 1664 }, { "epoch": 0.11281252117352124, "grad_norm": 10.524566650390625, "learning_rate": 9.87487165445958e-05, "loss": 1.2156, "step": 1665 }, { "epoch": 0.11288027644149333, "grad_norm": 12.232719421386719, "learning_rate": 9.874734752549798e-05, "loss": 1.2152, "step": 1666 }, { "epoch": 0.11294803170946541, "grad_norm": 10.665773391723633, "learning_rate": 9.874597850640018e-05, "loss": 1.085, "step": 1667 }, { "epoch": 0.1130157869774375, "grad_norm": 10.0956449508667, "learning_rate": 9.874460948730236e-05, "loss": 1.1142, "step": 1668 }, { "epoch": 0.11308354224540958, "grad_norm": 9.27825927734375, "learning_rate": 9.874324046820454e-05, "loss": 1.0263, "step": 1669 }, { "epoch": 0.11315129751338167, "grad_norm": 11.171673774719238, "learning_rate": 9.874187144910672e-05, "loss": 1.5049, "step": 1670 }, { "epoch": 0.11321905278135375, "grad_norm": 10.442279815673828, "learning_rate": 9.874050243000891e-05, "loss": 1.0697, "step": 1671 }, { "epoch": 0.11328680804932584, "grad_norm": 11.563700675964355, "learning_rate": 9.87391334109111e-05, "loss": 1.003, "step": 1672 }, { "epoch": 0.11335456331729792, "grad_norm": 12.447733879089355, "learning_rate": 9.873776439181327e-05, "loss": 1.2279, "step": 1673 }, { "epoch": 0.11342231858527001, "grad_norm": 9.70698070526123, "learning_rate": 9.873639537271545e-05, "loss": 1.2208, "step": 1674 }, { "epoch": 0.11349007385324208, "grad_norm": 10.89858341217041, "learning_rate": 9.873502635361763e-05, "loss": 1.1676, "step": 1675 }, { "epoch": 0.11355782912121418, "grad_norm": 11.106192588806152, "learning_rate": 9.873365733451983e-05, "loss": 1.0918, "step": 1676 }, { "epoch": 0.11362558438918625, "grad_norm": 10.186528205871582, "learning_rate": 9.873228831542201e-05, "loss": 1.1651, "step": 1677 }, { "epoch": 0.11369333965715835, "grad_norm": 10.69977855682373, "learning_rate": 9.873091929632419e-05, "loss": 1.2867, "step": 1678 }, { "epoch": 0.11376109492513042, "grad_norm": 11.934000015258789, "learning_rate": 9.872955027722637e-05, "loss": 1.2051, "step": 1679 }, { "epoch": 0.11382885019310252, "grad_norm": 9.277094841003418, "learning_rate": 9.872818125812856e-05, "loss": 1.17, "step": 1680 }, { "epoch": 0.1138966054610746, "grad_norm": 11.35097885131836, "learning_rate": 9.872681223903074e-05, "loss": 1.2433, "step": 1681 }, { "epoch": 0.11396436072904668, "grad_norm": 11.362833023071289, "learning_rate": 9.872544321993292e-05, "loss": 1.3577, "step": 1682 }, { "epoch": 0.11403211599701876, "grad_norm": 10.944365501403809, "learning_rate": 9.87240742008351e-05, "loss": 1.0731, "step": 1683 }, { "epoch": 0.11409987126499085, "grad_norm": 10.292986869812012, "learning_rate": 9.872270518173728e-05, "loss": 1.0639, "step": 1684 }, { "epoch": 0.11416762653296293, "grad_norm": 10.8629150390625, "learning_rate": 9.872133616263948e-05, "loss": 1.3503, "step": 1685 }, { "epoch": 0.11423538180093502, "grad_norm": 10.277920722961426, "learning_rate": 9.871996714354166e-05, "loss": 1.1544, "step": 1686 }, { "epoch": 0.1143031370689071, "grad_norm": 8.98864459991455, "learning_rate": 9.871859812444384e-05, "loss": 1.3638, "step": 1687 }, { "epoch": 0.11437089233687919, "grad_norm": 8.126718521118164, "learning_rate": 9.871722910534602e-05, "loss": 1.0343, "step": 1688 }, { "epoch": 0.11443864760485127, "grad_norm": 9.229771614074707, "learning_rate": 9.871586008624821e-05, "loss": 1.0137, "step": 1689 }, { "epoch": 0.11450640287282336, "grad_norm": 10.554340362548828, "learning_rate": 9.87144910671504e-05, "loss": 0.9908, "step": 1690 }, { "epoch": 0.11457415814079544, "grad_norm": 11.857934951782227, "learning_rate": 9.871312204805257e-05, "loss": 1.29, "step": 1691 }, { "epoch": 0.11464191340876753, "grad_norm": 9.334272384643555, "learning_rate": 9.871175302895475e-05, "loss": 1.0449, "step": 1692 }, { "epoch": 0.11470966867673962, "grad_norm": 9.092185974121094, "learning_rate": 9.871038400985693e-05, "loss": 1.1386, "step": 1693 }, { "epoch": 0.1147774239447117, "grad_norm": 9.379465103149414, "learning_rate": 9.870901499075913e-05, "loss": 1.0964, "step": 1694 }, { "epoch": 0.11484517921268379, "grad_norm": 10.038492202758789, "learning_rate": 9.870764597166131e-05, "loss": 0.8512, "step": 1695 }, { "epoch": 0.11491293448065587, "grad_norm": 8.2230863571167, "learning_rate": 9.870627695256349e-05, "loss": 1.0646, "step": 1696 }, { "epoch": 0.11498068974862796, "grad_norm": 9.939510345458984, "learning_rate": 9.870490793346567e-05, "loss": 1.081, "step": 1697 }, { "epoch": 0.11504844501660004, "grad_norm": 11.864813804626465, "learning_rate": 9.870353891436786e-05, "loss": 1.2447, "step": 1698 }, { "epoch": 0.11511620028457213, "grad_norm": 11.517714500427246, "learning_rate": 9.870216989527004e-05, "loss": 1.1274, "step": 1699 }, { "epoch": 0.11518395555254421, "grad_norm": 12.172054290771484, "learning_rate": 9.870080087617222e-05, "loss": 1.3278, "step": 1700 }, { "epoch": 0.1152517108205163, "grad_norm": 9.915261268615723, "learning_rate": 9.86994318570744e-05, "loss": 1.0298, "step": 1701 }, { "epoch": 0.11531946608848838, "grad_norm": 10.937396049499512, "learning_rate": 9.869806283797659e-05, "loss": 1.363, "step": 1702 }, { "epoch": 0.11538722135646047, "grad_norm": 11.586840629577637, "learning_rate": 9.869669381887878e-05, "loss": 0.9152, "step": 1703 }, { "epoch": 0.11545497662443255, "grad_norm": 10.035614967346191, "learning_rate": 9.869532479978096e-05, "loss": 1.2152, "step": 1704 }, { "epoch": 0.11552273189240464, "grad_norm": 8.775728225708008, "learning_rate": 9.869395578068314e-05, "loss": 1.0552, "step": 1705 }, { "epoch": 0.11559048716037672, "grad_norm": 10.85958480834961, "learning_rate": 9.869258676158532e-05, "loss": 1.0473, "step": 1706 }, { "epoch": 0.11565824242834881, "grad_norm": 9.311616897583008, "learning_rate": 9.869121774248751e-05, "loss": 1.1704, "step": 1707 }, { "epoch": 0.11572599769632089, "grad_norm": 9.503780364990234, "learning_rate": 9.86898487233897e-05, "loss": 1.1375, "step": 1708 }, { "epoch": 0.11579375296429298, "grad_norm": 9.02602481842041, "learning_rate": 9.868847970429187e-05, "loss": 0.9955, "step": 1709 }, { "epoch": 0.11586150823226506, "grad_norm": 11.473995208740234, "learning_rate": 9.868711068519407e-05, "loss": 0.8842, "step": 1710 }, { "epoch": 0.11592926350023715, "grad_norm": 10.124822616577148, "learning_rate": 9.868574166609625e-05, "loss": 1.0458, "step": 1711 }, { "epoch": 0.11599701876820923, "grad_norm": 11.004744529724121, "learning_rate": 9.868437264699843e-05, "loss": 1.2136, "step": 1712 }, { "epoch": 0.11606477403618132, "grad_norm": 10.497210502624512, "learning_rate": 9.868300362790062e-05, "loss": 1.3457, "step": 1713 }, { "epoch": 0.1161325293041534, "grad_norm": 10.933736801147461, "learning_rate": 9.86816346088028e-05, "loss": 1.0931, "step": 1714 }, { "epoch": 0.11620028457212549, "grad_norm": 11.045526504516602, "learning_rate": 9.868026558970498e-05, "loss": 1.2338, "step": 1715 }, { "epoch": 0.11626803984009756, "grad_norm": 10.763197898864746, "learning_rate": 9.867889657060716e-05, "loss": 1.2325, "step": 1716 }, { "epoch": 0.11633579510806966, "grad_norm": 9.53070068359375, "learning_rate": 9.867752755150936e-05, "loss": 1.1268, "step": 1717 }, { "epoch": 0.11640355037604173, "grad_norm": 10.572071075439453, "learning_rate": 9.867615853241154e-05, "loss": 1.0118, "step": 1718 }, { "epoch": 0.11647130564401383, "grad_norm": 7.633776664733887, "learning_rate": 9.867478951331372e-05, "loss": 1.081, "step": 1719 }, { "epoch": 0.1165390609119859, "grad_norm": 8.915566444396973, "learning_rate": 9.86734204942159e-05, "loss": 1.1526, "step": 1720 }, { "epoch": 0.116606816179958, "grad_norm": 8.296719551086426, "learning_rate": 9.867205147511809e-05, "loss": 1.0571, "step": 1721 }, { "epoch": 0.11667457144793007, "grad_norm": 9.387847900390625, "learning_rate": 9.867068245602027e-05, "loss": 1.0447, "step": 1722 }, { "epoch": 0.11674232671590216, "grad_norm": 10.832168579101562, "learning_rate": 9.866931343692245e-05, "loss": 1.2601, "step": 1723 }, { "epoch": 0.11681008198387424, "grad_norm": 11.631181716918945, "learning_rate": 9.866794441782463e-05, "loss": 1.4507, "step": 1724 }, { "epoch": 0.11687783725184633, "grad_norm": 8.372645378112793, "learning_rate": 9.866657539872681e-05, "loss": 1.1191, "step": 1725 }, { "epoch": 0.11694559251981841, "grad_norm": 9.577934265136719, "learning_rate": 9.866520637962901e-05, "loss": 1.0336, "step": 1726 }, { "epoch": 0.1170133477877905, "grad_norm": 10.626317977905273, "learning_rate": 9.866383736053119e-05, "loss": 1.1595, "step": 1727 }, { "epoch": 0.11708110305576258, "grad_norm": 9.758353233337402, "learning_rate": 9.866246834143337e-05, "loss": 0.8803, "step": 1728 }, { "epoch": 0.11714885832373467, "grad_norm": 9.643457412719727, "learning_rate": 9.866109932233555e-05, "loss": 1.3213, "step": 1729 }, { "epoch": 0.11721661359170675, "grad_norm": 8.910233497619629, "learning_rate": 9.865973030323773e-05, "loss": 1.1123, "step": 1730 }, { "epoch": 0.11728436885967884, "grad_norm": 8.893741607666016, "learning_rate": 9.865836128413992e-05, "loss": 1.1059, "step": 1731 }, { "epoch": 0.11735212412765092, "grad_norm": 12.937616348266602, "learning_rate": 9.86569922650421e-05, "loss": 1.2488, "step": 1732 }, { "epoch": 0.11741987939562301, "grad_norm": 9.696173667907715, "learning_rate": 9.865562324594428e-05, "loss": 1.1076, "step": 1733 }, { "epoch": 0.11748763466359509, "grad_norm": 11.568793296813965, "learning_rate": 9.865425422684646e-05, "loss": 1.1023, "step": 1734 }, { "epoch": 0.11755538993156718, "grad_norm": 11.438271522521973, "learning_rate": 9.865288520774866e-05, "loss": 1.1014, "step": 1735 }, { "epoch": 0.11762314519953926, "grad_norm": 8.217238426208496, "learning_rate": 9.865151618865084e-05, "loss": 0.9566, "step": 1736 }, { "epoch": 0.11769090046751135, "grad_norm": 9.716591835021973, "learning_rate": 9.865014716955302e-05, "loss": 1.0478, "step": 1737 }, { "epoch": 0.11775865573548343, "grad_norm": 11.152593612670898, "learning_rate": 9.86487781504552e-05, "loss": 1.3552, "step": 1738 }, { "epoch": 0.11782641100345552, "grad_norm": 8.824272155761719, "learning_rate": 9.864740913135738e-05, "loss": 1.0463, "step": 1739 }, { "epoch": 0.1178941662714276, "grad_norm": 8.941548347473145, "learning_rate": 9.864604011225957e-05, "loss": 0.9552, "step": 1740 }, { "epoch": 0.11796192153939969, "grad_norm": 10.76177978515625, "learning_rate": 9.864467109316175e-05, "loss": 1.0076, "step": 1741 }, { "epoch": 0.11802967680737177, "grad_norm": 9.87431812286377, "learning_rate": 9.864330207406393e-05, "loss": 0.9905, "step": 1742 }, { "epoch": 0.11809743207534386, "grad_norm": 12.406320571899414, "learning_rate": 9.864193305496611e-05, "loss": 1.2839, "step": 1743 }, { "epoch": 0.11816518734331594, "grad_norm": 12.649428367614746, "learning_rate": 9.864056403586831e-05, "loss": 1.0806, "step": 1744 }, { "epoch": 0.11823294261128803, "grad_norm": 10.888890266418457, "learning_rate": 9.863919501677049e-05, "loss": 0.8805, "step": 1745 }, { "epoch": 0.11830069787926012, "grad_norm": 10.59262752532959, "learning_rate": 9.863782599767267e-05, "loss": 0.9977, "step": 1746 }, { "epoch": 0.1183684531472322, "grad_norm": 11.14206314086914, "learning_rate": 9.863645697857485e-05, "loss": 1.4569, "step": 1747 }, { "epoch": 0.11843620841520429, "grad_norm": 12.148449897766113, "learning_rate": 9.863508795947703e-05, "loss": 1.5222, "step": 1748 }, { "epoch": 0.11850396368317637, "grad_norm": 11.814643859863281, "learning_rate": 9.863371894037922e-05, "loss": 1.3888, "step": 1749 }, { "epoch": 0.11857171895114846, "grad_norm": 12.437151908874512, "learning_rate": 9.86323499212814e-05, "loss": 1.024, "step": 1750 }, { "epoch": 0.11863947421912054, "grad_norm": 9.314751625061035, "learning_rate": 9.863098090218358e-05, "loss": 1.064, "step": 1751 }, { "epoch": 0.11870722948709263, "grad_norm": 12.978782653808594, "learning_rate": 9.862961188308576e-05, "loss": 1.4048, "step": 1752 }, { "epoch": 0.1187749847550647, "grad_norm": 9.825428009033203, "learning_rate": 9.862824286398796e-05, "loss": 0.9631, "step": 1753 }, { "epoch": 0.1188427400230368, "grad_norm": 9.127273559570312, "learning_rate": 9.862687384489014e-05, "loss": 1.0364, "step": 1754 }, { "epoch": 0.11891049529100887, "grad_norm": 10.290020942687988, "learning_rate": 9.862550482579232e-05, "loss": 1.329, "step": 1755 }, { "epoch": 0.11897825055898097, "grad_norm": 9.023946762084961, "learning_rate": 9.862413580669451e-05, "loss": 1.186, "step": 1756 }, { "epoch": 0.11904600582695304, "grad_norm": 11.889911651611328, "learning_rate": 9.86227667875967e-05, "loss": 1.0338, "step": 1757 }, { "epoch": 0.11911376109492514, "grad_norm": 8.938654899597168, "learning_rate": 9.862139776849887e-05, "loss": 1.1933, "step": 1758 }, { "epoch": 0.11918151636289721, "grad_norm": 9.240127563476562, "learning_rate": 9.862002874940107e-05, "loss": 1.2336, "step": 1759 }, { "epoch": 0.1192492716308693, "grad_norm": 10.340953826904297, "learning_rate": 9.861865973030325e-05, "loss": 1.1097, "step": 1760 }, { "epoch": 0.11931702689884138, "grad_norm": 10.807507514953613, "learning_rate": 9.861729071120543e-05, "loss": 0.9127, "step": 1761 }, { "epoch": 0.11938478216681347, "grad_norm": 9.670829772949219, "learning_rate": 9.861592169210761e-05, "loss": 1.2296, "step": 1762 }, { "epoch": 0.11945253743478555, "grad_norm": 10.85981273651123, "learning_rate": 9.86145526730098e-05, "loss": 1.1341, "step": 1763 }, { "epoch": 0.11952029270275764, "grad_norm": 9.198482513427734, "learning_rate": 9.861318365391198e-05, "loss": 1.4397, "step": 1764 }, { "epoch": 0.11958804797072972, "grad_norm": 11.130392074584961, "learning_rate": 9.861181463481416e-05, "loss": 1.381, "step": 1765 }, { "epoch": 0.11965580323870181, "grad_norm": 9.568144798278809, "learning_rate": 9.861044561571634e-05, "loss": 1.0284, "step": 1766 }, { "epoch": 0.11972355850667389, "grad_norm": 9.049298286437988, "learning_rate": 9.860907659661854e-05, "loss": 1.0714, "step": 1767 }, { "epoch": 0.11979131377464598, "grad_norm": 8.948347091674805, "learning_rate": 9.860770757752072e-05, "loss": 1.0248, "step": 1768 }, { "epoch": 0.11985906904261806, "grad_norm": 10.836170196533203, "learning_rate": 9.86063385584229e-05, "loss": 1.0551, "step": 1769 }, { "epoch": 0.11992682431059015, "grad_norm": 11.044917106628418, "learning_rate": 9.860496953932508e-05, "loss": 1.1374, "step": 1770 }, { "epoch": 0.11999457957856223, "grad_norm": 10.975854873657227, "learning_rate": 9.860360052022726e-05, "loss": 1.2018, "step": 1771 }, { "epoch": 0.12006233484653432, "grad_norm": 9.752070426940918, "learning_rate": 9.860223150112945e-05, "loss": 1.0843, "step": 1772 }, { "epoch": 0.1201300901145064, "grad_norm": 9.662758827209473, "learning_rate": 9.860086248203163e-05, "loss": 0.8866, "step": 1773 }, { "epoch": 0.12019784538247849, "grad_norm": 11.268694877624512, "learning_rate": 9.859949346293381e-05, "loss": 1.2653, "step": 1774 }, { "epoch": 0.12026560065045057, "grad_norm": 9.201732635498047, "learning_rate": 9.8598124443836e-05, "loss": 1.1466, "step": 1775 }, { "epoch": 0.12033335591842266, "grad_norm": 12.577048301696777, "learning_rate": 9.859675542473819e-05, "loss": 1.2606, "step": 1776 }, { "epoch": 0.12040111118639474, "grad_norm": 10.051258087158203, "learning_rate": 9.859538640564037e-05, "loss": 1.1533, "step": 1777 }, { "epoch": 0.12046886645436683, "grad_norm": 9.358968734741211, "learning_rate": 9.859401738654255e-05, "loss": 1.0501, "step": 1778 }, { "epoch": 0.12053662172233891, "grad_norm": 11.277347564697266, "learning_rate": 9.859264836744473e-05, "loss": 0.9804, "step": 1779 }, { "epoch": 0.120604376990311, "grad_norm": 12.41299057006836, "learning_rate": 9.859127934834691e-05, "loss": 1.427, "step": 1780 }, { "epoch": 0.12067213225828308, "grad_norm": 10.081144332885742, "learning_rate": 9.85899103292491e-05, "loss": 1.1321, "step": 1781 }, { "epoch": 0.12073988752625517, "grad_norm": 11.469686508178711, "learning_rate": 9.858854131015128e-05, "loss": 1.1972, "step": 1782 }, { "epoch": 0.12080764279422725, "grad_norm": 8.608443260192871, "learning_rate": 9.858717229105346e-05, "loss": 0.9914, "step": 1783 }, { "epoch": 0.12087539806219934, "grad_norm": 11.673405647277832, "learning_rate": 9.858580327195564e-05, "loss": 1.3406, "step": 1784 }, { "epoch": 0.12094315333017142, "grad_norm": 9.533596992492676, "learning_rate": 9.858443425285782e-05, "loss": 1.2119, "step": 1785 }, { "epoch": 0.12101090859814351, "grad_norm": 7.792461395263672, "learning_rate": 9.858306523376002e-05, "loss": 1.167, "step": 1786 }, { "epoch": 0.12107866386611559, "grad_norm": 10.209978103637695, "learning_rate": 9.85816962146622e-05, "loss": 1.164, "step": 1787 }, { "epoch": 0.12114641913408768, "grad_norm": 11.591109275817871, "learning_rate": 9.858032719556438e-05, "loss": 1.2598, "step": 1788 }, { "epoch": 0.12121417440205975, "grad_norm": 10.562797546386719, "learning_rate": 9.857895817646656e-05, "loss": 1.3071, "step": 1789 }, { "epoch": 0.12128192967003185, "grad_norm": 9.419530868530273, "learning_rate": 9.857758915736875e-05, "loss": 0.9951, "step": 1790 }, { "epoch": 0.12134968493800392, "grad_norm": 9.525064468383789, "learning_rate": 9.857622013827093e-05, "loss": 1.1833, "step": 1791 }, { "epoch": 0.12141744020597602, "grad_norm": 9.140012741088867, "learning_rate": 9.857485111917311e-05, "loss": 1.0573, "step": 1792 }, { "epoch": 0.1214851954739481, "grad_norm": 10.88685417175293, "learning_rate": 9.85734821000753e-05, "loss": 1.1669, "step": 1793 }, { "epoch": 0.12155295074192018, "grad_norm": 8.63829231262207, "learning_rate": 9.857211308097747e-05, "loss": 1.1733, "step": 1794 }, { "epoch": 0.12162070600989226, "grad_norm": 9.90237808227539, "learning_rate": 9.857074406187967e-05, "loss": 0.9175, "step": 1795 }, { "epoch": 0.12168846127786435, "grad_norm": 10.068902969360352, "learning_rate": 9.856937504278185e-05, "loss": 0.9422, "step": 1796 }, { "epoch": 0.12175621654583643, "grad_norm": 7.482577800750732, "learning_rate": 9.856800602368403e-05, "loss": 1.0359, "step": 1797 }, { "epoch": 0.12182397181380852, "grad_norm": 10.847206115722656, "learning_rate": 9.856663700458621e-05, "loss": 1.2258, "step": 1798 }, { "epoch": 0.12189172708178062, "grad_norm": 11.122177124023438, "learning_rate": 9.85652679854884e-05, "loss": 1.1378, "step": 1799 }, { "epoch": 0.1219594823497527, "grad_norm": 9.994139671325684, "learning_rate": 9.856389896639058e-05, "loss": 1.2663, "step": 1800 }, { "epoch": 0.12202723761772478, "grad_norm": 10.687590599060059, "learning_rate": 9.856252994729276e-05, "loss": 1.1777, "step": 1801 }, { "epoch": 0.12209499288569686, "grad_norm": 8.403971672058105, "learning_rate": 9.856116092819494e-05, "loss": 0.9812, "step": 1802 }, { "epoch": 0.12216274815366895, "grad_norm": 9.77443790435791, "learning_rate": 9.855979190909714e-05, "loss": 1.0615, "step": 1803 }, { "epoch": 0.12223050342164103, "grad_norm": 10.772642135620117, "learning_rate": 9.855842288999932e-05, "loss": 1.1835, "step": 1804 }, { "epoch": 0.12229825868961312, "grad_norm": 10.509920120239258, "learning_rate": 9.85570538709015e-05, "loss": 1.0361, "step": 1805 }, { "epoch": 0.1223660139575852, "grad_norm": 11.14194393157959, "learning_rate": 9.855568485180369e-05, "loss": 0.9947, "step": 1806 }, { "epoch": 0.12243376922555729, "grad_norm": 11.018975257873535, "learning_rate": 9.855431583270587e-05, "loss": 1.2036, "step": 1807 }, { "epoch": 0.12250152449352937, "grad_norm": 12.099370002746582, "learning_rate": 9.855294681360805e-05, "loss": 1.2977, "step": 1808 }, { "epoch": 0.12256927976150146, "grad_norm": 8.841024398803711, "learning_rate": 9.855157779451025e-05, "loss": 1.2093, "step": 1809 }, { "epoch": 0.12263703502947354, "grad_norm": 9.149311065673828, "learning_rate": 9.855020877541243e-05, "loss": 1.0088, "step": 1810 }, { "epoch": 0.12270479029744563, "grad_norm": 10.53498649597168, "learning_rate": 9.854883975631461e-05, "loss": 0.9527, "step": 1811 }, { "epoch": 0.12277254556541771, "grad_norm": 9.230859756469727, "learning_rate": 9.854747073721679e-05, "loss": 0.9584, "step": 1812 }, { "epoch": 0.1228403008333898, "grad_norm": 11.702610969543457, "learning_rate": 9.854610171811898e-05, "loss": 1.2369, "step": 1813 }, { "epoch": 0.12290805610136188, "grad_norm": 9.912004470825195, "learning_rate": 9.854473269902116e-05, "loss": 1.2571, "step": 1814 }, { "epoch": 0.12297581136933397, "grad_norm": 10.071510314941406, "learning_rate": 9.854336367992334e-05, "loss": 1.1666, "step": 1815 }, { "epoch": 0.12304356663730605, "grad_norm": 10.642035484313965, "learning_rate": 9.854199466082552e-05, "loss": 1.0309, "step": 1816 }, { "epoch": 0.12311132190527814, "grad_norm": 11.135947227478027, "learning_rate": 9.85406256417277e-05, "loss": 1.1276, "step": 1817 }, { "epoch": 0.12317907717325022, "grad_norm": 11.575469017028809, "learning_rate": 9.85392566226299e-05, "loss": 1.1507, "step": 1818 }, { "epoch": 0.12324683244122231, "grad_norm": 10.470771789550781, "learning_rate": 9.853788760353208e-05, "loss": 1.3777, "step": 1819 }, { "epoch": 0.12331458770919439, "grad_norm": 9.77783489227295, "learning_rate": 9.853651858443426e-05, "loss": 1.0492, "step": 1820 }, { "epoch": 0.12338234297716648, "grad_norm": 10.132309913635254, "learning_rate": 9.853514956533644e-05, "loss": 1.1488, "step": 1821 }, { "epoch": 0.12345009824513856, "grad_norm": 11.159482955932617, "learning_rate": 9.853378054623863e-05, "loss": 1.2571, "step": 1822 }, { "epoch": 0.12351785351311065, "grad_norm": 11.599156379699707, "learning_rate": 9.853241152714081e-05, "loss": 1.1734, "step": 1823 }, { "epoch": 0.12358560878108273, "grad_norm": 10.278475761413574, "learning_rate": 9.853104250804299e-05, "loss": 1.0637, "step": 1824 }, { "epoch": 0.12365336404905482, "grad_norm": 12.126015663146973, "learning_rate": 9.852967348894517e-05, "loss": 1.0546, "step": 1825 }, { "epoch": 0.1237211193170269, "grad_norm": 12.094749450683594, "learning_rate": 9.852830446984735e-05, "loss": 1.374, "step": 1826 }, { "epoch": 0.12378887458499899, "grad_norm": 7.607821941375732, "learning_rate": 9.852693545074955e-05, "loss": 0.827, "step": 1827 }, { "epoch": 0.12385662985297106, "grad_norm": 12.086756706237793, "learning_rate": 9.852556643165173e-05, "loss": 1.1219, "step": 1828 }, { "epoch": 0.12392438512094316, "grad_norm": 10.215773582458496, "learning_rate": 9.852419741255391e-05, "loss": 1.0148, "step": 1829 }, { "epoch": 0.12399214038891523, "grad_norm": 10.438709259033203, "learning_rate": 9.852282839345609e-05, "loss": 1.0649, "step": 1830 }, { "epoch": 0.12405989565688733, "grad_norm": 10.564906120300293, "learning_rate": 9.852145937435828e-05, "loss": 1.4346, "step": 1831 }, { "epoch": 0.1241276509248594, "grad_norm": 9.83198356628418, "learning_rate": 9.852009035526046e-05, "loss": 0.847, "step": 1832 }, { "epoch": 0.1241954061928315, "grad_norm": 10.990026473999023, "learning_rate": 9.851872133616264e-05, "loss": 1.2358, "step": 1833 }, { "epoch": 0.12426316146080357, "grad_norm": 10.108382225036621, "learning_rate": 9.851735231706482e-05, "loss": 0.9943, "step": 1834 }, { "epoch": 0.12433091672877566, "grad_norm": 10.23820686340332, "learning_rate": 9.8515983297967e-05, "loss": 1.2464, "step": 1835 }, { "epoch": 0.12439867199674774, "grad_norm": 11.552473068237305, "learning_rate": 9.85146142788692e-05, "loss": 1.2175, "step": 1836 }, { "epoch": 0.12446642726471983, "grad_norm": 8.646978378295898, "learning_rate": 9.851324525977138e-05, "loss": 0.9975, "step": 1837 }, { "epoch": 0.12453418253269191, "grad_norm": 9.344541549682617, "learning_rate": 9.851187624067356e-05, "loss": 0.9463, "step": 1838 }, { "epoch": 0.124601937800664, "grad_norm": 12.645013809204102, "learning_rate": 9.851050722157574e-05, "loss": 1.0396, "step": 1839 }, { "epoch": 0.12466969306863608, "grad_norm": 10.093807220458984, "learning_rate": 9.850913820247792e-05, "loss": 1.2554, "step": 1840 }, { "epoch": 0.12473744833660817, "grad_norm": 8.792567253112793, "learning_rate": 9.850776918338011e-05, "loss": 0.9852, "step": 1841 }, { "epoch": 0.12480520360458025, "grad_norm": 10.448987007141113, "learning_rate": 9.85064001642823e-05, "loss": 1.168, "step": 1842 }, { "epoch": 0.12487295887255234, "grad_norm": 9.953516960144043, "learning_rate": 9.850503114518447e-05, "loss": 1.0116, "step": 1843 }, { "epoch": 0.12494071414052442, "grad_norm": 10.187036514282227, "learning_rate": 9.850366212608665e-05, "loss": 1.203, "step": 1844 }, { "epoch": 0.1250084694084965, "grad_norm": 8.890667915344238, "learning_rate": 9.850229310698885e-05, "loss": 1.0104, "step": 1845 }, { "epoch": 0.1250762246764686, "grad_norm": 11.916625022888184, "learning_rate": 9.850092408789103e-05, "loss": 1.0254, "step": 1846 }, { "epoch": 0.12514397994444068, "grad_norm": 10.934864044189453, "learning_rate": 9.849955506879321e-05, "loss": 0.975, "step": 1847 }, { "epoch": 0.12521173521241277, "grad_norm": 11.214954376220703, "learning_rate": 9.849818604969539e-05, "loss": 1.45, "step": 1848 }, { "epoch": 0.12527949048038484, "grad_norm": 8.801512718200684, "learning_rate": 9.849681703059758e-05, "loss": 1.0899, "step": 1849 }, { "epoch": 0.12534724574835693, "grad_norm": 10.340089797973633, "learning_rate": 9.849544801149976e-05, "loss": 1.1704, "step": 1850 }, { "epoch": 0.12541500101632902, "grad_norm": 9.204201698303223, "learning_rate": 9.849407899240194e-05, "loss": 1.0569, "step": 1851 }, { "epoch": 0.1254827562843011, "grad_norm": 9.43604564666748, "learning_rate": 9.849270997330414e-05, "loss": 1.0721, "step": 1852 }, { "epoch": 0.1255505115522732, "grad_norm": 10.110416412353516, "learning_rate": 9.849134095420632e-05, "loss": 1.1265, "step": 1853 }, { "epoch": 0.12561826682024527, "grad_norm": 10.630755424499512, "learning_rate": 9.84899719351085e-05, "loss": 1.1103, "step": 1854 }, { "epoch": 0.12568602208821736, "grad_norm": 10.888835906982422, "learning_rate": 9.848860291601069e-05, "loss": 1.347, "step": 1855 }, { "epoch": 0.12575377735618945, "grad_norm": 8.84623908996582, "learning_rate": 9.848723389691287e-05, "loss": 0.8531, "step": 1856 }, { "epoch": 0.12582153262416154, "grad_norm": 10.99928092956543, "learning_rate": 9.848586487781505e-05, "loss": 1.1747, "step": 1857 }, { "epoch": 0.1258892878921336, "grad_norm": 8.556151390075684, "learning_rate": 9.848449585871723e-05, "loss": 1.0207, "step": 1858 }, { "epoch": 0.1259570431601057, "grad_norm": 10.207547187805176, "learning_rate": 9.848312683961943e-05, "loss": 1.4782, "step": 1859 }, { "epoch": 0.1260247984280778, "grad_norm": 9.39505386352539, "learning_rate": 9.848175782052161e-05, "loss": 1.1415, "step": 1860 }, { "epoch": 0.12609255369604988, "grad_norm": 10.647768020629883, "learning_rate": 9.848038880142379e-05, "loss": 1.1439, "step": 1861 }, { "epoch": 0.12616030896402194, "grad_norm": 11.517707824707031, "learning_rate": 9.847901978232597e-05, "loss": 1.3181, "step": 1862 }, { "epoch": 0.12622806423199404, "grad_norm": 10.916093826293945, "learning_rate": 9.847765076322815e-05, "loss": 1.0708, "step": 1863 }, { "epoch": 0.12629581949996613, "grad_norm": 9.893363952636719, "learning_rate": 9.847628174413034e-05, "loss": 1.1932, "step": 1864 }, { "epoch": 0.12636357476793822, "grad_norm": 12.349174499511719, "learning_rate": 9.847491272503252e-05, "loss": 1.3161, "step": 1865 }, { "epoch": 0.12643133003591028, "grad_norm": 10.157081604003906, "learning_rate": 9.84735437059347e-05, "loss": 1.1325, "step": 1866 }, { "epoch": 0.12649908530388237, "grad_norm": 9.772073745727539, "learning_rate": 9.847217468683688e-05, "loss": 1.2739, "step": 1867 }, { "epoch": 0.12656684057185447, "grad_norm": 10.468371391296387, "learning_rate": 9.847080566773908e-05, "loss": 1.0279, "step": 1868 }, { "epoch": 0.12663459583982656, "grad_norm": 11.055033683776855, "learning_rate": 9.846943664864126e-05, "loss": 1.0584, "step": 1869 }, { "epoch": 0.12670235110779862, "grad_norm": 10.47987174987793, "learning_rate": 9.846806762954344e-05, "loss": 1.0385, "step": 1870 }, { "epoch": 0.1267701063757707, "grad_norm": 8.933801651000977, "learning_rate": 9.846669861044562e-05, "loss": 1.0259, "step": 1871 }, { "epoch": 0.1268378616437428, "grad_norm": 10.32228946685791, "learning_rate": 9.84653295913478e-05, "loss": 1.0472, "step": 1872 }, { "epoch": 0.1269056169117149, "grad_norm": 9.294051170349121, "learning_rate": 9.846396057224999e-05, "loss": 1.1638, "step": 1873 }, { "epoch": 0.12697337217968696, "grad_norm": 8.32187271118164, "learning_rate": 9.846259155315217e-05, "loss": 0.9865, "step": 1874 }, { "epoch": 0.12704112744765905, "grad_norm": 8.782264709472656, "learning_rate": 9.846122253405435e-05, "loss": 1.0868, "step": 1875 }, { "epoch": 0.12710888271563114, "grad_norm": 10.444697380065918, "learning_rate": 9.845985351495653e-05, "loss": 1.2137, "step": 1876 }, { "epoch": 0.12717663798360324, "grad_norm": 10.543922424316406, "learning_rate": 9.845848449585873e-05, "loss": 1.2337, "step": 1877 }, { "epoch": 0.1272443932515753, "grad_norm": 8.504612922668457, "learning_rate": 9.845711547676091e-05, "loss": 1.0234, "step": 1878 }, { "epoch": 0.1273121485195474, "grad_norm": 8.83178997039795, "learning_rate": 9.845574645766309e-05, "loss": 1.0611, "step": 1879 }, { "epoch": 0.12737990378751948, "grad_norm": 14.37713623046875, "learning_rate": 9.845437743856527e-05, "loss": 1.1704, "step": 1880 }, { "epoch": 0.12744765905549157, "grad_norm": 12.43575668334961, "learning_rate": 9.845300841946745e-05, "loss": 0.848, "step": 1881 }, { "epoch": 0.12751541432346364, "grad_norm": 11.625554084777832, "learning_rate": 9.845163940036964e-05, "loss": 1.2395, "step": 1882 }, { "epoch": 0.12758316959143573, "grad_norm": 7.8962082862854, "learning_rate": 9.845027038127182e-05, "loss": 0.9702, "step": 1883 }, { "epoch": 0.12765092485940782, "grad_norm": 11.689674377441406, "learning_rate": 9.8448901362174e-05, "loss": 1.4885, "step": 1884 }, { "epoch": 0.1277186801273799, "grad_norm": 11.037276268005371, "learning_rate": 9.844753234307618e-05, "loss": 1.1751, "step": 1885 }, { "epoch": 0.12778643539535198, "grad_norm": 11.115680694580078, "learning_rate": 9.844616332397836e-05, "loss": 1.2912, "step": 1886 }, { "epoch": 0.12785419066332407, "grad_norm": 10.553492546081543, "learning_rate": 9.844479430488056e-05, "loss": 1.1001, "step": 1887 }, { "epoch": 0.12792194593129616, "grad_norm": 9.92845630645752, "learning_rate": 9.844342528578274e-05, "loss": 1.1697, "step": 1888 }, { "epoch": 0.12798970119926825, "grad_norm": 9.33232307434082, "learning_rate": 9.844205626668492e-05, "loss": 1.0931, "step": 1889 }, { "epoch": 0.12805745646724032, "grad_norm": 12.35257339477539, "learning_rate": 9.84406872475871e-05, "loss": 0.9819, "step": 1890 }, { "epoch": 0.1281252117352124, "grad_norm": 9.913105010986328, "learning_rate": 9.843931822848929e-05, "loss": 1.2515, "step": 1891 }, { "epoch": 0.1281929670031845, "grad_norm": 11.218729019165039, "learning_rate": 9.843794920939147e-05, "loss": 1.0073, "step": 1892 }, { "epoch": 0.1282607222711566, "grad_norm": 9.97374153137207, "learning_rate": 9.843658019029365e-05, "loss": 1.0527, "step": 1893 }, { "epoch": 0.12832847753912865, "grad_norm": 8.837554931640625, "learning_rate": 9.843521117119583e-05, "loss": 1.0175, "step": 1894 }, { "epoch": 0.12839623280710075, "grad_norm": 9.207158088684082, "learning_rate": 9.843384215209803e-05, "loss": 1.0729, "step": 1895 }, { "epoch": 0.12846398807507284, "grad_norm": 10.960060119628906, "learning_rate": 9.843247313300021e-05, "loss": 1.0485, "step": 1896 }, { "epoch": 0.12853174334304493, "grad_norm": 9.724952697753906, "learning_rate": 9.843110411390239e-05, "loss": 1.2624, "step": 1897 }, { "epoch": 0.128599498611017, "grad_norm": 9.02813720703125, "learning_rate": 9.842973509480458e-05, "loss": 1.011, "step": 1898 }, { "epoch": 0.12866725387898909, "grad_norm": 8.507847785949707, "learning_rate": 9.842836607570676e-05, "loss": 0.9155, "step": 1899 }, { "epoch": 0.12873500914696118, "grad_norm": 10.609807968139648, "learning_rate": 9.842699705660894e-05, "loss": 1.1984, "step": 1900 }, { "epoch": 0.12880276441493327, "grad_norm": 10.144070625305176, "learning_rate": 9.842562803751114e-05, "loss": 1.2331, "step": 1901 }, { "epoch": 0.12887051968290533, "grad_norm": 11.719651222229004, "learning_rate": 9.842425901841332e-05, "loss": 1.2086, "step": 1902 }, { "epoch": 0.12893827495087742, "grad_norm": 10.689997673034668, "learning_rate": 9.84228899993155e-05, "loss": 1.3769, "step": 1903 }, { "epoch": 0.12900603021884952, "grad_norm": 10.7424955368042, "learning_rate": 9.842152098021768e-05, "loss": 1.1844, "step": 1904 }, { "epoch": 0.1290737854868216, "grad_norm": 11.207498550415039, "learning_rate": 9.842015196111987e-05, "loss": 1.2151, "step": 1905 }, { "epoch": 0.1291415407547937, "grad_norm": 10.689212799072266, "learning_rate": 9.841878294202205e-05, "loss": 1.1182, "step": 1906 }, { "epoch": 0.12920929602276576, "grad_norm": 9.708647727966309, "learning_rate": 9.841741392292423e-05, "loss": 1.2874, "step": 1907 }, { "epoch": 0.12927705129073785, "grad_norm": 9.809847831726074, "learning_rate": 9.841604490382641e-05, "loss": 1.1716, "step": 1908 }, { "epoch": 0.12934480655870995, "grad_norm": 8.170798301696777, "learning_rate": 9.84146758847286e-05, "loss": 1.1342, "step": 1909 }, { "epoch": 0.12941256182668204, "grad_norm": 9.12773609161377, "learning_rate": 9.841330686563079e-05, "loss": 1.1506, "step": 1910 }, { "epoch": 0.1294803170946541, "grad_norm": 13.67032241821289, "learning_rate": 9.841193784653297e-05, "loss": 1.2065, "step": 1911 }, { "epoch": 0.1295480723626262, "grad_norm": 9.35611629486084, "learning_rate": 9.841056882743515e-05, "loss": 1.281, "step": 1912 }, { "epoch": 0.12961582763059828, "grad_norm": 9.044548988342285, "learning_rate": 9.840919980833733e-05, "loss": 1.0285, "step": 1913 }, { "epoch": 0.12968358289857038, "grad_norm": 9.955796241760254, "learning_rate": 9.840783078923952e-05, "loss": 1.1742, "step": 1914 }, { "epoch": 0.12975133816654244, "grad_norm": 10.456665992736816, "learning_rate": 9.84064617701417e-05, "loss": 1.3452, "step": 1915 }, { "epoch": 0.12981909343451453, "grad_norm": 10.861869812011719, "learning_rate": 9.840509275104388e-05, "loss": 1.0397, "step": 1916 }, { "epoch": 0.12988684870248662, "grad_norm": 8.546277046203613, "learning_rate": 9.840372373194606e-05, "loss": 0.8762, "step": 1917 }, { "epoch": 0.12995460397045872, "grad_norm": 9.103103637695312, "learning_rate": 9.840235471284824e-05, "loss": 1.0998, "step": 1918 }, { "epoch": 0.13002235923843078, "grad_norm": 9.794631958007812, "learning_rate": 9.840098569375044e-05, "loss": 1.0644, "step": 1919 }, { "epoch": 0.13009011450640287, "grad_norm": 8.702750205993652, "learning_rate": 9.839961667465262e-05, "loss": 1.108, "step": 1920 }, { "epoch": 0.13015786977437496, "grad_norm": 11.108800888061523, "learning_rate": 9.83982476555548e-05, "loss": 1.375, "step": 1921 }, { "epoch": 0.13022562504234705, "grad_norm": 11.253337860107422, "learning_rate": 9.839687863645698e-05, "loss": 1.0246, "step": 1922 }, { "epoch": 0.13029338031031912, "grad_norm": 9.817541122436523, "learning_rate": 9.839550961735917e-05, "loss": 1.0222, "step": 1923 }, { "epoch": 0.1303611355782912, "grad_norm": 9.378199577331543, "learning_rate": 9.839414059826135e-05, "loss": 1.0387, "step": 1924 }, { "epoch": 0.1304288908462633, "grad_norm": 8.789336204528809, "learning_rate": 9.839277157916353e-05, "loss": 1.0635, "step": 1925 }, { "epoch": 0.1304966461142354, "grad_norm": 11.900403022766113, "learning_rate": 9.839140256006571e-05, "loss": 1.1224, "step": 1926 }, { "epoch": 0.13056440138220746, "grad_norm": 10.007912635803223, "learning_rate": 9.83900335409679e-05, "loss": 0.8988, "step": 1927 }, { "epoch": 0.13063215665017955, "grad_norm": 10.157328605651855, "learning_rate": 9.838866452187009e-05, "loss": 0.952, "step": 1928 }, { "epoch": 0.13069991191815164, "grad_norm": 8.763729095458984, "learning_rate": 9.838729550277227e-05, "loss": 0.9304, "step": 1929 }, { "epoch": 0.13076766718612373, "grad_norm": 10.75432300567627, "learning_rate": 9.838592648367445e-05, "loss": 1.0438, "step": 1930 }, { "epoch": 0.1308354224540958, "grad_norm": 8.840702056884766, "learning_rate": 9.838455746457663e-05, "loss": 0.9164, "step": 1931 }, { "epoch": 0.1309031777220679, "grad_norm": 9.526811599731445, "learning_rate": 9.838318844547882e-05, "loss": 1.2181, "step": 1932 }, { "epoch": 0.13097093299003998, "grad_norm": 12.827199935913086, "learning_rate": 9.8381819426381e-05, "loss": 1.223, "step": 1933 }, { "epoch": 0.13103868825801207, "grad_norm": 9.47105884552002, "learning_rate": 9.838045040728318e-05, "loss": 1.1014, "step": 1934 }, { "epoch": 0.13110644352598413, "grad_norm": 9.044878959655762, "learning_rate": 9.837908138818536e-05, "loss": 1.1451, "step": 1935 }, { "epoch": 0.13117419879395623, "grad_norm": 9.24599838256836, "learning_rate": 9.837771236908754e-05, "loss": 1.1144, "step": 1936 }, { "epoch": 0.13124195406192832, "grad_norm": 12.543252944946289, "learning_rate": 9.837634334998974e-05, "loss": 1.1233, "step": 1937 }, { "epoch": 0.1313097093299004, "grad_norm": 10.022245407104492, "learning_rate": 9.837497433089192e-05, "loss": 1.3569, "step": 1938 }, { "epoch": 0.13137746459787247, "grad_norm": 11.967863082885742, "learning_rate": 9.83736053117941e-05, "loss": 1.2086, "step": 1939 }, { "epoch": 0.13144521986584456, "grad_norm": 9.64406967163086, "learning_rate": 9.837223629269628e-05, "loss": 1.1647, "step": 1940 }, { "epoch": 0.13151297513381666, "grad_norm": 8.538762092590332, "learning_rate": 9.837086727359847e-05, "loss": 1.012, "step": 1941 }, { "epoch": 0.13158073040178875, "grad_norm": 10.889129638671875, "learning_rate": 9.836949825450065e-05, "loss": 1.1559, "step": 1942 }, { "epoch": 0.1316484856697608, "grad_norm": 8.407093048095703, "learning_rate": 9.836812923540283e-05, "loss": 0.9624, "step": 1943 }, { "epoch": 0.1317162409377329, "grad_norm": 9.175569534301758, "learning_rate": 9.836676021630503e-05, "loss": 0.9814, "step": 1944 }, { "epoch": 0.131783996205705, "grad_norm": 11.240396499633789, "learning_rate": 9.836539119720721e-05, "loss": 1.3059, "step": 1945 }, { "epoch": 0.1318517514736771, "grad_norm": 8.791098594665527, "learning_rate": 9.836402217810939e-05, "loss": 0.9891, "step": 1946 }, { "epoch": 0.13191950674164915, "grad_norm": 7.401971340179443, "learning_rate": 9.836265315901158e-05, "loss": 0.9706, "step": 1947 }, { "epoch": 0.13198726200962124, "grad_norm": 13.127768516540527, "learning_rate": 9.836128413991376e-05, "loss": 0.8924, "step": 1948 }, { "epoch": 0.13205501727759333, "grad_norm": 9.805618286132812, "learning_rate": 9.835991512081594e-05, "loss": 0.9969, "step": 1949 }, { "epoch": 0.13212277254556543, "grad_norm": 10.500420570373535, "learning_rate": 9.835854610171812e-05, "loss": 1.1045, "step": 1950 }, { "epoch": 0.1321905278135375, "grad_norm": 10.043769836425781, "learning_rate": 9.835717708262032e-05, "loss": 1.2475, "step": 1951 }, { "epoch": 0.13225828308150958, "grad_norm": 10.6277437210083, "learning_rate": 9.83558080635225e-05, "loss": 0.9285, "step": 1952 }, { "epoch": 0.13232603834948167, "grad_norm": 9.011415481567383, "learning_rate": 9.835443904442468e-05, "loss": 1.1895, "step": 1953 }, { "epoch": 0.13239379361745376, "grad_norm": 11.031267166137695, "learning_rate": 9.835307002532686e-05, "loss": 1.1781, "step": 1954 }, { "epoch": 0.13246154888542583, "grad_norm": 9.960331916809082, "learning_rate": 9.835170100622905e-05, "loss": 1.0972, "step": 1955 }, { "epoch": 0.13252930415339792, "grad_norm": 9.549619674682617, "learning_rate": 9.835033198713123e-05, "loss": 1.0918, "step": 1956 }, { "epoch": 0.13259705942137, "grad_norm": 9.780478477478027, "learning_rate": 9.834896296803341e-05, "loss": 1.1491, "step": 1957 }, { "epoch": 0.1326648146893421, "grad_norm": 8.948554992675781, "learning_rate": 9.834759394893559e-05, "loss": 1.0969, "step": 1958 }, { "epoch": 0.1327325699573142, "grad_norm": 9.165532112121582, "learning_rate": 9.834622492983777e-05, "loss": 0.9294, "step": 1959 }, { "epoch": 0.13280032522528626, "grad_norm": 8.738619804382324, "learning_rate": 9.834485591073997e-05, "loss": 1.3074, "step": 1960 }, { "epoch": 0.13286808049325835, "grad_norm": 9.442314147949219, "learning_rate": 9.834348689164215e-05, "loss": 1.1238, "step": 1961 }, { "epoch": 0.13293583576123044, "grad_norm": 12.411934852600098, "learning_rate": 9.834211787254433e-05, "loss": 1.1405, "step": 1962 }, { "epoch": 0.13300359102920253, "grad_norm": 9.911120414733887, "learning_rate": 9.834074885344651e-05, "loss": 1.2197, "step": 1963 }, { "epoch": 0.1330713462971746, "grad_norm": 9.619095802307129, "learning_rate": 9.83393798343487e-05, "loss": 1.0526, "step": 1964 }, { "epoch": 0.1331391015651467, "grad_norm": 10.163374900817871, "learning_rate": 9.833801081525088e-05, "loss": 0.8374, "step": 1965 }, { "epoch": 0.13320685683311878, "grad_norm": 9.342517852783203, "learning_rate": 9.833664179615306e-05, "loss": 1.1908, "step": 1966 }, { "epoch": 0.13327461210109087, "grad_norm": 10.419418334960938, "learning_rate": 9.833527277705524e-05, "loss": 1.1175, "step": 1967 }, { "epoch": 0.13334236736906294, "grad_norm": 9.5196533203125, "learning_rate": 9.833390375795742e-05, "loss": 1.2347, "step": 1968 }, { "epoch": 0.13341012263703503, "grad_norm": 9.242755889892578, "learning_rate": 9.833253473885962e-05, "loss": 1.2401, "step": 1969 }, { "epoch": 0.13347787790500712, "grad_norm": 10.243762969970703, "learning_rate": 9.83311657197618e-05, "loss": 1.3492, "step": 1970 }, { "epoch": 0.1335456331729792, "grad_norm": 9.169745445251465, "learning_rate": 9.832979670066398e-05, "loss": 0.9533, "step": 1971 }, { "epoch": 0.13361338844095128, "grad_norm": 10.292695999145508, "learning_rate": 9.832842768156616e-05, "loss": 1.2216, "step": 1972 }, { "epoch": 0.13368114370892337, "grad_norm": 9.25019645690918, "learning_rate": 9.832705866246834e-05, "loss": 1.1613, "step": 1973 }, { "epoch": 0.13374889897689546, "grad_norm": 8.518020629882812, "learning_rate": 9.832568964337053e-05, "loss": 1.0767, "step": 1974 }, { "epoch": 0.13381665424486755, "grad_norm": 11.824376106262207, "learning_rate": 9.832432062427271e-05, "loss": 1.1884, "step": 1975 }, { "epoch": 0.13388440951283961, "grad_norm": 10.544015884399414, "learning_rate": 9.832295160517489e-05, "loss": 0.9675, "step": 1976 }, { "epoch": 0.1339521647808117, "grad_norm": 9.495721817016602, "learning_rate": 9.832158258607707e-05, "loss": 1.0586, "step": 1977 }, { "epoch": 0.1340199200487838, "grad_norm": 10.378434181213379, "learning_rate": 9.832021356697927e-05, "loss": 1.176, "step": 1978 }, { "epoch": 0.1340876753167559, "grad_norm": 10.026887893676758, "learning_rate": 9.831884454788145e-05, "loss": 1.035, "step": 1979 }, { "epoch": 0.13415543058472795, "grad_norm": 8.878249168395996, "learning_rate": 9.831747552878363e-05, "loss": 0.8441, "step": 1980 }, { "epoch": 0.13422318585270004, "grad_norm": 9.637616157531738, "learning_rate": 9.831610650968581e-05, "loss": 1.1435, "step": 1981 }, { "epoch": 0.13429094112067214, "grad_norm": 10.153711318969727, "learning_rate": 9.831473749058799e-05, "loss": 1.0833, "step": 1982 }, { "epoch": 0.13435869638864423, "grad_norm": 8.48596477508545, "learning_rate": 9.831336847149018e-05, "loss": 0.7752, "step": 1983 }, { "epoch": 0.1344264516566163, "grad_norm": 8.454994201660156, "learning_rate": 9.831199945239236e-05, "loss": 0.8657, "step": 1984 }, { "epoch": 0.13449420692458838, "grad_norm": 10.607659339904785, "learning_rate": 9.831063043329454e-05, "loss": 1.1836, "step": 1985 }, { "epoch": 0.13456196219256047, "grad_norm": 8.560189247131348, "learning_rate": 9.830926141419672e-05, "loss": 0.7877, "step": 1986 }, { "epoch": 0.13462971746053257, "grad_norm": 10.687662124633789, "learning_rate": 9.830789239509892e-05, "loss": 1.1586, "step": 1987 }, { "epoch": 0.13469747272850463, "grad_norm": 9.725050926208496, "learning_rate": 9.83065233760011e-05, "loss": 1.214, "step": 1988 }, { "epoch": 0.13476522799647672, "grad_norm": 9.808280944824219, "learning_rate": 9.830515435690328e-05, "loss": 1.0532, "step": 1989 }, { "epoch": 0.1348329832644488, "grad_norm": 8.838544845581055, "learning_rate": 9.830378533780547e-05, "loss": 1.0872, "step": 1990 }, { "epoch": 0.1349007385324209, "grad_norm": 12.043976783752441, "learning_rate": 9.830241631870765e-05, "loss": 1.2248, "step": 1991 }, { "epoch": 0.13496849380039297, "grad_norm": 10.044602394104004, "learning_rate": 9.830104729960983e-05, "loss": 0.9746, "step": 1992 }, { "epoch": 0.13503624906836506, "grad_norm": 10.861641883850098, "learning_rate": 9.829967828051203e-05, "loss": 0.9868, "step": 1993 }, { "epoch": 0.13510400433633715, "grad_norm": 9.005234718322754, "learning_rate": 9.82983092614142e-05, "loss": 1.0619, "step": 1994 }, { "epoch": 0.13517175960430924, "grad_norm": 9.902874946594238, "learning_rate": 9.829694024231639e-05, "loss": 1.3176, "step": 1995 }, { "epoch": 0.1352395148722813, "grad_norm": 12.133747100830078, "learning_rate": 9.829557122321857e-05, "loss": 1.3634, "step": 1996 }, { "epoch": 0.1353072701402534, "grad_norm": 9.571062088012695, "learning_rate": 9.829420220412076e-05, "loss": 1.139, "step": 1997 }, { "epoch": 0.1353750254082255, "grad_norm": 8.269545555114746, "learning_rate": 9.829283318502294e-05, "loss": 1.1422, "step": 1998 }, { "epoch": 0.13544278067619758, "grad_norm": 9.963309288024902, "learning_rate": 9.829146416592512e-05, "loss": 1.0612, "step": 1999 }, { "epoch": 0.13551053594416965, "grad_norm": 7.434775352478027, "learning_rate": 9.82900951468273e-05, "loss": 1.0418, "step": 2000 }, { "epoch": 0.13557829121214174, "grad_norm": 8.89494800567627, "learning_rate": 9.82887261277295e-05, "loss": 1.0677, "step": 2001 }, { "epoch": 0.13564604648011383, "grad_norm": 9.349754333496094, "learning_rate": 9.828735710863168e-05, "loss": 0.8932, "step": 2002 }, { "epoch": 0.13571380174808592, "grad_norm": 7.574460506439209, "learning_rate": 9.828598808953386e-05, "loss": 0.7652, "step": 2003 }, { "epoch": 0.13578155701605799, "grad_norm": 10.191296577453613, "learning_rate": 9.828461907043604e-05, "loss": 0.9881, "step": 2004 }, { "epoch": 0.13584931228403008, "grad_norm": 11.235671043395996, "learning_rate": 9.828325005133822e-05, "loss": 1.1942, "step": 2005 }, { "epoch": 0.13591706755200217, "grad_norm": 9.97700023651123, "learning_rate": 9.828188103224041e-05, "loss": 1.154, "step": 2006 }, { "epoch": 0.13598482281997426, "grad_norm": 8.283380508422852, "learning_rate": 9.828051201314259e-05, "loss": 0.8207, "step": 2007 }, { "epoch": 0.13605257808794632, "grad_norm": 9.147144317626953, "learning_rate": 9.827914299404477e-05, "loss": 1.0751, "step": 2008 }, { "epoch": 0.13612033335591842, "grad_norm": 9.99758529663086, "learning_rate": 9.827777397494695e-05, "loss": 1.1166, "step": 2009 }, { "epoch": 0.1361880886238905, "grad_norm": 8.340705871582031, "learning_rate": 9.827640495584915e-05, "loss": 1.0311, "step": 2010 }, { "epoch": 0.1362558438918626, "grad_norm": 10.536659240722656, "learning_rate": 9.827503593675133e-05, "loss": 1.1704, "step": 2011 }, { "epoch": 0.1363235991598347, "grad_norm": 9.01259994506836, "learning_rate": 9.827366691765351e-05, "loss": 1.0531, "step": 2012 }, { "epoch": 0.13639135442780675, "grad_norm": 10.476259231567383, "learning_rate": 9.827229789855569e-05, "loss": 1.053, "step": 2013 }, { "epoch": 0.13645910969577885, "grad_norm": 8.496943473815918, "learning_rate": 9.827092887945787e-05, "loss": 0.836, "step": 2014 }, { "epoch": 0.13652686496375094, "grad_norm": 11.638651847839355, "learning_rate": 9.826955986036006e-05, "loss": 1.0886, "step": 2015 }, { "epoch": 0.13659462023172303, "grad_norm": 7.483241558074951, "learning_rate": 9.826819084126224e-05, "loss": 1.0406, "step": 2016 }, { "epoch": 0.1366623754996951, "grad_norm": 8.648175239562988, "learning_rate": 9.826682182216442e-05, "loss": 1.0442, "step": 2017 }, { "epoch": 0.13673013076766719, "grad_norm": 8.593596458435059, "learning_rate": 9.82654528030666e-05, "loss": 1.1184, "step": 2018 }, { "epoch": 0.13679788603563928, "grad_norm": 8.563817977905273, "learning_rate": 9.826408378396878e-05, "loss": 0.8645, "step": 2019 }, { "epoch": 0.13686564130361137, "grad_norm": 7.812311172485352, "learning_rate": 9.826271476487098e-05, "loss": 1.0605, "step": 2020 }, { "epoch": 0.13693339657158343, "grad_norm": 10.97168254852295, "learning_rate": 9.826134574577316e-05, "loss": 1.1474, "step": 2021 }, { "epoch": 0.13700115183955552, "grad_norm": 7.749011993408203, "learning_rate": 9.825997672667534e-05, "loss": 0.9836, "step": 2022 }, { "epoch": 0.13706890710752762, "grad_norm": 11.088539123535156, "learning_rate": 9.825860770757752e-05, "loss": 1.1456, "step": 2023 }, { "epoch": 0.1371366623754997, "grad_norm": 10.960288047790527, "learning_rate": 9.825723868847971e-05, "loss": 1.4315, "step": 2024 }, { "epoch": 0.13720441764347177, "grad_norm": 10.804458618164062, "learning_rate": 9.825586966938189e-05, "loss": 1.0388, "step": 2025 }, { "epoch": 0.13727217291144386, "grad_norm": 7.906947612762451, "learning_rate": 9.825450065028407e-05, "loss": 1.1306, "step": 2026 }, { "epoch": 0.13733992817941595, "grad_norm": 10.404504776000977, "learning_rate": 9.825313163118625e-05, "loss": 1.0907, "step": 2027 }, { "epoch": 0.13740768344738805, "grad_norm": 9.684488296508789, "learning_rate": 9.825176261208843e-05, "loss": 1.1007, "step": 2028 }, { "epoch": 0.1374754387153601, "grad_norm": 8.345703125, "learning_rate": 9.825039359299063e-05, "loss": 0.9342, "step": 2029 }, { "epoch": 0.1375431939833322, "grad_norm": 8.513103485107422, "learning_rate": 9.824902457389281e-05, "loss": 1.1139, "step": 2030 }, { "epoch": 0.1376109492513043, "grad_norm": 8.675403594970703, "learning_rate": 9.824765555479499e-05, "loss": 0.8499, "step": 2031 }, { "epoch": 0.13767870451927638, "grad_norm": 10.550504684448242, "learning_rate": 9.824628653569717e-05, "loss": 1.2563, "step": 2032 }, { "epoch": 0.13774645978724845, "grad_norm": 7.775172710418701, "learning_rate": 9.824491751659936e-05, "loss": 1.0394, "step": 2033 }, { "epoch": 0.13781421505522054, "grad_norm": 10.057134628295898, "learning_rate": 9.824354849750154e-05, "loss": 1.337, "step": 2034 }, { "epoch": 0.13788197032319263, "grad_norm": 10.658480644226074, "learning_rate": 9.824217947840372e-05, "loss": 1.1675, "step": 2035 }, { "epoch": 0.13794972559116472, "grad_norm": 10.499738693237305, "learning_rate": 9.82408104593059e-05, "loss": 1.1397, "step": 2036 }, { "epoch": 0.1380174808591368, "grad_norm": 12.853378295898438, "learning_rate": 9.82394414402081e-05, "loss": 1.2373, "step": 2037 }, { "epoch": 0.13808523612710888, "grad_norm": 8.66174602508545, "learning_rate": 9.823807242111028e-05, "loss": 0.9863, "step": 2038 }, { "epoch": 0.13815299139508097, "grad_norm": 10.246259689331055, "learning_rate": 9.823670340201246e-05, "loss": 1.0892, "step": 2039 }, { "epoch": 0.13822074666305306, "grad_norm": 11.880420684814453, "learning_rate": 9.823533438291465e-05, "loss": 1.069, "step": 2040 }, { "epoch": 0.13828850193102513, "grad_norm": 9.475419998168945, "learning_rate": 9.823396536381683e-05, "loss": 0.9276, "step": 2041 }, { "epoch": 0.13835625719899722, "grad_norm": 9.827219009399414, "learning_rate": 9.823259634471903e-05, "loss": 1.2536, "step": 2042 }, { "epoch": 0.1384240124669693, "grad_norm": 8.558990478515625, "learning_rate": 9.82312273256212e-05, "loss": 1.2779, "step": 2043 }, { "epoch": 0.1384917677349414, "grad_norm": 12.038803100585938, "learning_rate": 9.822985830652339e-05, "loss": 0.9019, "step": 2044 }, { "epoch": 0.13855952300291347, "grad_norm": 10.764846801757812, "learning_rate": 9.822848928742557e-05, "loss": 1.1803, "step": 2045 }, { "epoch": 0.13862727827088556, "grad_norm": 10.788616180419922, "learning_rate": 9.822712026832775e-05, "loss": 1.3469, "step": 2046 }, { "epoch": 0.13869503353885765, "grad_norm": 8.718049049377441, "learning_rate": 9.822575124922994e-05, "loss": 0.9135, "step": 2047 }, { "epoch": 0.13876278880682974, "grad_norm": 12.245726585388184, "learning_rate": 9.822438223013212e-05, "loss": 1.3114, "step": 2048 }, { "epoch": 0.1388305440748018, "grad_norm": 10.55252456665039, "learning_rate": 9.82230132110343e-05, "loss": 1.1303, "step": 2049 }, { "epoch": 0.1388982993427739, "grad_norm": 8.177289962768555, "learning_rate": 9.822164419193648e-05, "loss": 0.9796, "step": 2050 }, { "epoch": 0.138966054610746, "grad_norm": 10.708680152893066, "learning_rate": 9.822027517283866e-05, "loss": 1.1546, "step": 2051 }, { "epoch": 0.13903380987871808, "grad_norm": 8.631631851196289, "learning_rate": 9.821890615374086e-05, "loss": 1.0055, "step": 2052 }, { "epoch": 0.13910156514669014, "grad_norm": 8.778770446777344, "learning_rate": 9.821753713464304e-05, "loss": 0.9626, "step": 2053 }, { "epoch": 0.13916932041466223, "grad_norm": 10.004143714904785, "learning_rate": 9.821616811554522e-05, "loss": 1.1411, "step": 2054 }, { "epoch": 0.13923707568263433, "grad_norm": 9.47324275970459, "learning_rate": 9.82147990964474e-05, "loss": 1.0081, "step": 2055 }, { "epoch": 0.13930483095060642, "grad_norm": 10.014934539794922, "learning_rate": 9.821343007734959e-05, "loss": 1.3015, "step": 2056 }, { "epoch": 0.13937258621857848, "grad_norm": 10.589959144592285, "learning_rate": 9.821206105825177e-05, "loss": 1.128, "step": 2057 }, { "epoch": 0.13944034148655057, "grad_norm": 9.273834228515625, "learning_rate": 9.821069203915395e-05, "loss": 0.7743, "step": 2058 }, { "epoch": 0.13950809675452266, "grad_norm": 10.72019100189209, "learning_rate": 9.820932302005613e-05, "loss": 0.9524, "step": 2059 }, { "epoch": 0.13957585202249476, "grad_norm": 11.212404251098633, "learning_rate": 9.820795400095831e-05, "loss": 1.2044, "step": 2060 }, { "epoch": 0.13964360729046682, "grad_norm": 9.800296783447266, "learning_rate": 9.82065849818605e-05, "loss": 1.0819, "step": 2061 }, { "epoch": 0.1397113625584389, "grad_norm": 8.668676376342773, "learning_rate": 9.820521596276269e-05, "loss": 1.0778, "step": 2062 }, { "epoch": 0.139779117826411, "grad_norm": 10.854613304138184, "learning_rate": 9.820384694366487e-05, "loss": 1.1181, "step": 2063 }, { "epoch": 0.1398468730943831, "grad_norm": 12.019871711730957, "learning_rate": 9.820247792456705e-05, "loss": 1.2507, "step": 2064 }, { "epoch": 0.1399146283623552, "grad_norm": 8.835234642028809, "learning_rate": 9.820110890546924e-05, "loss": 1.1338, "step": 2065 }, { "epoch": 0.13998238363032725, "grad_norm": 8.57636547088623, "learning_rate": 9.819973988637142e-05, "loss": 0.957, "step": 2066 }, { "epoch": 0.14005013889829934, "grad_norm": 9.080939292907715, "learning_rate": 9.81983708672736e-05, "loss": 1.0035, "step": 2067 }, { "epoch": 0.14011789416627143, "grad_norm": 9.616862297058105, "learning_rate": 9.819700184817578e-05, "loss": 0.9045, "step": 2068 }, { "epoch": 0.14018564943424353, "grad_norm": 10.04652214050293, "learning_rate": 9.819563282907796e-05, "loss": 1.1202, "step": 2069 }, { "epoch": 0.1402534047022156, "grad_norm": 9.132543563842773, "learning_rate": 9.819426380998016e-05, "loss": 0.9918, "step": 2070 }, { "epoch": 0.14032115997018768, "grad_norm": 12.023603439331055, "learning_rate": 9.819289479088234e-05, "loss": 1.1271, "step": 2071 }, { "epoch": 0.14038891523815977, "grad_norm": 9.167064666748047, "learning_rate": 9.819152577178452e-05, "loss": 0.9136, "step": 2072 }, { "epoch": 0.14045667050613186, "grad_norm": 11.954336166381836, "learning_rate": 9.81901567526867e-05, "loss": 1.0073, "step": 2073 }, { "epoch": 0.14052442577410393, "grad_norm": 9.166129112243652, "learning_rate": 9.818878773358888e-05, "loss": 1.1101, "step": 2074 }, { "epoch": 0.14059218104207602, "grad_norm": 8.53992748260498, "learning_rate": 9.818741871449107e-05, "loss": 0.9556, "step": 2075 }, { "epoch": 0.1406599363100481, "grad_norm": 10.769463539123535, "learning_rate": 9.818604969539325e-05, "loss": 1.3009, "step": 2076 }, { "epoch": 0.1407276915780202, "grad_norm": 10.938043594360352, "learning_rate": 9.818468067629543e-05, "loss": 1.4299, "step": 2077 }, { "epoch": 0.14079544684599227, "grad_norm": 10.8585844039917, "learning_rate": 9.818331165719761e-05, "loss": 1.0804, "step": 2078 }, { "epoch": 0.14086320211396436, "grad_norm": 8.763557434082031, "learning_rate": 9.81819426380998e-05, "loss": 0.8321, "step": 2079 }, { "epoch": 0.14093095738193645, "grad_norm": 11.917708396911621, "learning_rate": 9.818057361900199e-05, "loss": 1.4293, "step": 2080 }, { "epoch": 0.14099871264990854, "grad_norm": 10.189987182617188, "learning_rate": 9.817920459990417e-05, "loss": 1.139, "step": 2081 }, { "epoch": 0.1410664679178806, "grad_norm": 11.09156608581543, "learning_rate": 9.817783558080635e-05, "loss": 0.9866, "step": 2082 }, { "epoch": 0.1411342231858527, "grad_norm": 9.394566535949707, "learning_rate": 9.817646656170854e-05, "loss": 0.9401, "step": 2083 }, { "epoch": 0.1412019784538248, "grad_norm": 8.232394218444824, "learning_rate": 9.817509754261072e-05, "loss": 0.9193, "step": 2084 }, { "epoch": 0.14126973372179688, "grad_norm": 9.4952392578125, "learning_rate": 9.81737285235129e-05, "loss": 1.3217, "step": 2085 }, { "epoch": 0.14133748898976894, "grad_norm": 10.110014915466309, "learning_rate": 9.81723595044151e-05, "loss": 1.0413, "step": 2086 }, { "epoch": 0.14140524425774104, "grad_norm": 10.850225448608398, "learning_rate": 9.817099048531728e-05, "loss": 1.3722, "step": 2087 }, { "epoch": 0.14147299952571313, "grad_norm": 10.924457550048828, "learning_rate": 9.816962146621946e-05, "loss": 1.4004, "step": 2088 }, { "epoch": 0.14154075479368522, "grad_norm": 10.022381782531738, "learning_rate": 9.816825244712165e-05, "loss": 1.1989, "step": 2089 }, { "epoch": 0.14160851006165728, "grad_norm": 11.537693977355957, "learning_rate": 9.816688342802383e-05, "loss": 1.1922, "step": 2090 }, { "epoch": 0.14167626532962938, "grad_norm": 8.525372505187988, "learning_rate": 9.816551440892601e-05, "loss": 0.9291, "step": 2091 }, { "epoch": 0.14174402059760147, "grad_norm": 8.972722053527832, "learning_rate": 9.816414538982819e-05, "loss": 1.0917, "step": 2092 }, { "epoch": 0.14181177586557356, "grad_norm": 8.386235237121582, "learning_rate": 9.816277637073039e-05, "loss": 1.0482, "step": 2093 }, { "epoch": 0.14187953113354562, "grad_norm": 9.620473861694336, "learning_rate": 9.816140735163257e-05, "loss": 1.2523, "step": 2094 }, { "epoch": 0.1419472864015177, "grad_norm": 9.236804962158203, "learning_rate": 9.816003833253475e-05, "loss": 0.9474, "step": 2095 }, { "epoch": 0.1420150416694898, "grad_norm": 9.06925106048584, "learning_rate": 9.815866931343693e-05, "loss": 0.8921, "step": 2096 }, { "epoch": 0.1420827969374619, "grad_norm": 10.127729415893555, "learning_rate": 9.815730029433912e-05, "loss": 1.1693, "step": 2097 }, { "epoch": 0.14215055220543396, "grad_norm": 10.703007698059082, "learning_rate": 9.81559312752413e-05, "loss": 1.0652, "step": 2098 }, { "epoch": 0.14221830747340605, "grad_norm": 10.004093170166016, "learning_rate": 9.815456225614348e-05, "loss": 1.0158, "step": 2099 }, { "epoch": 0.14228606274137814, "grad_norm": 8.900236129760742, "learning_rate": 9.815319323704566e-05, "loss": 0.7327, "step": 2100 }, { "epoch": 0.14235381800935024, "grad_norm": 13.102290153503418, "learning_rate": 9.815182421794784e-05, "loss": 1.2837, "step": 2101 }, { "epoch": 0.1424215732773223, "grad_norm": 9.32657527923584, "learning_rate": 9.815045519885004e-05, "loss": 1.0812, "step": 2102 }, { "epoch": 0.1424893285452944, "grad_norm": 11.42785930633545, "learning_rate": 9.814908617975222e-05, "loss": 1.4467, "step": 2103 }, { "epoch": 0.14255708381326648, "grad_norm": 9.878450393676758, "learning_rate": 9.81477171606544e-05, "loss": 1.0774, "step": 2104 }, { "epoch": 0.14262483908123857, "grad_norm": 9.329227447509766, "learning_rate": 9.814634814155658e-05, "loss": 1.2536, "step": 2105 }, { "epoch": 0.14269259434921064, "grad_norm": 9.071858406066895, "learning_rate": 9.814497912245876e-05, "loss": 1.148, "step": 2106 }, { "epoch": 0.14276034961718273, "grad_norm": 11.595810890197754, "learning_rate": 9.814361010336095e-05, "loss": 1.1992, "step": 2107 }, { "epoch": 0.14282810488515482, "grad_norm": 10.041107177734375, "learning_rate": 9.814224108426313e-05, "loss": 1.0447, "step": 2108 }, { "epoch": 0.1428958601531269, "grad_norm": 9.913261413574219, "learning_rate": 9.814087206516531e-05, "loss": 1.1629, "step": 2109 }, { "epoch": 0.14296361542109898, "grad_norm": 10.983177185058594, "learning_rate": 9.813950304606749e-05, "loss": 1.0288, "step": 2110 }, { "epoch": 0.14303137068907107, "grad_norm": 10.479610443115234, "learning_rate": 9.813813402696969e-05, "loss": 1.5308, "step": 2111 }, { "epoch": 0.14309912595704316, "grad_norm": 8.70734691619873, "learning_rate": 9.813676500787187e-05, "loss": 0.9987, "step": 2112 }, { "epoch": 0.14316688122501525, "grad_norm": 8.242984771728516, "learning_rate": 9.813539598877405e-05, "loss": 1.0051, "step": 2113 }, { "epoch": 0.14323463649298732, "grad_norm": 8.627467155456543, "learning_rate": 9.813402696967623e-05, "loss": 1.1134, "step": 2114 }, { "epoch": 0.1433023917609594, "grad_norm": 9.970986366271973, "learning_rate": 9.813265795057841e-05, "loss": 1.148, "step": 2115 }, { "epoch": 0.1433701470289315, "grad_norm": 8.719794273376465, "learning_rate": 9.81312889314806e-05, "loss": 1.1013, "step": 2116 }, { "epoch": 0.1434379022969036, "grad_norm": 9.455860137939453, "learning_rate": 9.812991991238278e-05, "loss": 1.2333, "step": 2117 }, { "epoch": 0.14350565756487568, "grad_norm": 8.851629257202148, "learning_rate": 9.812855089328496e-05, "loss": 1.0985, "step": 2118 }, { "epoch": 0.14357341283284775, "grad_norm": 11.843599319458008, "learning_rate": 9.812718187418714e-05, "loss": 1.2292, "step": 2119 }, { "epoch": 0.14364116810081984, "grad_norm": 9.550506591796875, "learning_rate": 9.812581285508934e-05, "loss": 0.9466, "step": 2120 }, { "epoch": 0.14370892336879193, "grad_norm": 9.234643936157227, "learning_rate": 9.812444383599152e-05, "loss": 1.1968, "step": 2121 }, { "epoch": 0.14377667863676402, "grad_norm": 9.365439414978027, "learning_rate": 9.81230748168937e-05, "loss": 1.1643, "step": 2122 }, { "epoch": 0.14384443390473609, "grad_norm": 9.476024627685547, "learning_rate": 9.812170579779588e-05, "loss": 1.0541, "step": 2123 }, { "epoch": 0.14391218917270818, "grad_norm": 9.53847885131836, "learning_rate": 9.812033677869806e-05, "loss": 1.0808, "step": 2124 }, { "epoch": 0.14397994444068027, "grad_norm": 8.918699264526367, "learning_rate": 9.811896775960025e-05, "loss": 1.1055, "step": 2125 }, { "epoch": 0.14404769970865236, "grad_norm": 12.126187324523926, "learning_rate": 9.811759874050243e-05, "loss": 1.6079, "step": 2126 }, { "epoch": 0.14411545497662442, "grad_norm": 10.922599792480469, "learning_rate": 9.811622972140461e-05, "loss": 1.2569, "step": 2127 }, { "epoch": 0.14418321024459652, "grad_norm": 8.582880020141602, "learning_rate": 9.811486070230679e-05, "loss": 0.9562, "step": 2128 }, { "epoch": 0.1442509655125686, "grad_norm": 8.926568031311035, "learning_rate": 9.811349168320899e-05, "loss": 1.0108, "step": 2129 }, { "epoch": 0.1443187207805407, "grad_norm": 13.906332015991211, "learning_rate": 9.811212266411117e-05, "loss": 1.1851, "step": 2130 }, { "epoch": 0.14438647604851276, "grad_norm": 10.374212265014648, "learning_rate": 9.811075364501335e-05, "loss": 1.2661, "step": 2131 }, { "epoch": 0.14445423131648485, "grad_norm": 10.580810546875, "learning_rate": 9.810938462591554e-05, "loss": 0.8967, "step": 2132 }, { "epoch": 0.14452198658445695, "grad_norm": 8.91865062713623, "learning_rate": 9.810801560681772e-05, "loss": 1.032, "step": 2133 }, { "epoch": 0.14458974185242904, "grad_norm": 8.16041374206543, "learning_rate": 9.81066465877199e-05, "loss": 0.8622, "step": 2134 }, { "epoch": 0.1446574971204011, "grad_norm": 8.575905799865723, "learning_rate": 9.81052775686221e-05, "loss": 0.9956, "step": 2135 }, { "epoch": 0.1447252523883732, "grad_norm": 9.620558738708496, "learning_rate": 9.810390854952428e-05, "loss": 1.1442, "step": 2136 }, { "epoch": 0.14479300765634529, "grad_norm": 10.390005111694336, "learning_rate": 9.810253953042646e-05, "loss": 1.0518, "step": 2137 }, { "epoch": 0.14486076292431738, "grad_norm": 10.714217185974121, "learning_rate": 9.810117051132864e-05, "loss": 1.3215, "step": 2138 }, { "epoch": 0.14492851819228944, "grad_norm": 9.390836715698242, "learning_rate": 9.809980149223083e-05, "loss": 1.0936, "step": 2139 }, { "epoch": 0.14499627346026153, "grad_norm": 10.848445892333984, "learning_rate": 9.809843247313301e-05, "loss": 1.1946, "step": 2140 }, { "epoch": 0.14506402872823362, "grad_norm": 11.041672706604004, "learning_rate": 9.809706345403519e-05, "loss": 1.0422, "step": 2141 }, { "epoch": 0.14513178399620572, "grad_norm": 10.1151762008667, "learning_rate": 9.809569443493737e-05, "loss": 1.2199, "step": 2142 }, { "epoch": 0.14519953926417778, "grad_norm": 8.686629295349121, "learning_rate": 9.809432541583957e-05, "loss": 0.9861, "step": 2143 }, { "epoch": 0.14526729453214987, "grad_norm": 10.579313278198242, "learning_rate": 9.809295639674175e-05, "loss": 0.9372, "step": 2144 }, { "epoch": 0.14533504980012196, "grad_norm": 8.916631698608398, "learning_rate": 9.809158737764393e-05, "loss": 0.8024, "step": 2145 }, { "epoch": 0.14540280506809405, "grad_norm": 9.29333209991455, "learning_rate": 9.80902183585461e-05, "loss": 1.0785, "step": 2146 }, { "epoch": 0.14547056033606612, "grad_norm": 10.584277153015137, "learning_rate": 9.808884933944829e-05, "loss": 0.9167, "step": 2147 }, { "epoch": 0.1455383156040382, "grad_norm": 10.68551254272461, "learning_rate": 9.808748032035048e-05, "loss": 1.0019, "step": 2148 }, { "epoch": 0.1456060708720103, "grad_norm": 9.057500839233398, "learning_rate": 9.808611130125266e-05, "loss": 1.194, "step": 2149 }, { "epoch": 0.1456738261399824, "grad_norm": 8.642207145690918, "learning_rate": 9.808474228215484e-05, "loss": 1.0849, "step": 2150 }, { "epoch": 0.14574158140795446, "grad_norm": 9.460419654846191, "learning_rate": 9.808337326305702e-05, "loss": 1.1747, "step": 2151 }, { "epoch": 0.14580933667592655, "grad_norm": 9.003097534179688, "learning_rate": 9.80820042439592e-05, "loss": 0.8967, "step": 2152 }, { "epoch": 0.14587709194389864, "grad_norm": 9.85009765625, "learning_rate": 9.80806352248614e-05, "loss": 1.1698, "step": 2153 }, { "epoch": 0.14594484721187073, "grad_norm": 10.233894348144531, "learning_rate": 9.807926620576358e-05, "loss": 1.0496, "step": 2154 }, { "epoch": 0.1460126024798428, "grad_norm": 9.660355567932129, "learning_rate": 9.807789718666576e-05, "loss": 0.9616, "step": 2155 }, { "epoch": 0.1460803577478149, "grad_norm": 9.46469497680664, "learning_rate": 9.807652816756794e-05, "loss": 1.0005, "step": 2156 }, { "epoch": 0.14614811301578698, "grad_norm": 8.004712104797363, "learning_rate": 9.807515914847013e-05, "loss": 0.8455, "step": 2157 }, { "epoch": 0.14621586828375907, "grad_norm": 10.039002418518066, "learning_rate": 9.807379012937231e-05, "loss": 1.097, "step": 2158 }, { "epoch": 0.14628362355173113, "grad_norm": 11.16292667388916, "learning_rate": 9.807242111027449e-05, "loss": 1.4332, "step": 2159 }, { "epoch": 0.14635137881970323, "grad_norm": 9.34833812713623, "learning_rate": 9.807105209117667e-05, "loss": 1.0438, "step": 2160 }, { "epoch": 0.14641913408767532, "grad_norm": 10.256768226623535, "learning_rate": 9.806968307207885e-05, "loss": 1.2923, "step": 2161 }, { "epoch": 0.1464868893556474, "grad_norm": 8.890941619873047, "learning_rate": 9.806831405298105e-05, "loss": 0.8419, "step": 2162 }, { "epoch": 0.14655464462361947, "grad_norm": 9.340752601623535, "learning_rate": 9.806694503388323e-05, "loss": 0.9504, "step": 2163 }, { "epoch": 0.14662239989159156, "grad_norm": 10.89192008972168, "learning_rate": 9.80655760147854e-05, "loss": 1.012, "step": 2164 }, { "epoch": 0.14669015515956366, "grad_norm": 8.626432418823242, "learning_rate": 9.806420699568759e-05, "loss": 0.8943, "step": 2165 }, { "epoch": 0.14675791042753575, "grad_norm": 9.465259552001953, "learning_rate": 9.806283797658978e-05, "loss": 1.1617, "step": 2166 }, { "epoch": 0.1468256656955078, "grad_norm": 11.695369720458984, "learning_rate": 9.806146895749196e-05, "loss": 1.2719, "step": 2167 }, { "epoch": 0.1468934209634799, "grad_norm": 9.416814804077148, "learning_rate": 9.806009993839414e-05, "loss": 1.3615, "step": 2168 }, { "epoch": 0.146961176231452, "grad_norm": 9.6004638671875, "learning_rate": 9.805873091929632e-05, "loss": 1.2571, "step": 2169 }, { "epoch": 0.1470289314994241, "grad_norm": 9.187546730041504, "learning_rate": 9.80573619001985e-05, "loss": 0.962, "step": 2170 }, { "epoch": 0.14709668676739618, "grad_norm": 10.691286087036133, "learning_rate": 9.80559928811007e-05, "loss": 1.0619, "step": 2171 }, { "epoch": 0.14716444203536824, "grad_norm": 9.17457103729248, "learning_rate": 9.805462386200288e-05, "loss": 1.1507, "step": 2172 }, { "epoch": 0.14723219730334033, "grad_norm": 10.739082336425781, "learning_rate": 9.805325484290506e-05, "loss": 1.2929, "step": 2173 }, { "epoch": 0.14729995257131243, "grad_norm": 9.091232299804688, "learning_rate": 9.805188582380724e-05, "loss": 0.9422, "step": 2174 }, { "epoch": 0.14736770783928452, "grad_norm": 8.231295585632324, "learning_rate": 9.805051680470943e-05, "loss": 0.9695, "step": 2175 }, { "epoch": 0.14743546310725658, "grad_norm": 9.622401237487793, "learning_rate": 9.804914778561161e-05, "loss": 0.982, "step": 2176 }, { "epoch": 0.14750321837522867, "grad_norm": 9.49541187286377, "learning_rate": 9.804777876651379e-05, "loss": 0.9005, "step": 2177 }, { "epoch": 0.14757097364320076, "grad_norm": 10.584654808044434, "learning_rate": 9.804640974741599e-05, "loss": 0.9427, "step": 2178 }, { "epoch": 0.14763872891117286, "grad_norm": 9.132317543029785, "learning_rate": 9.804504072831817e-05, "loss": 1.0192, "step": 2179 }, { "epoch": 0.14770648417914492, "grad_norm": 8.598082542419434, "learning_rate": 9.804367170922035e-05, "loss": 0.9842, "step": 2180 }, { "epoch": 0.147774239447117, "grad_norm": 8.941360473632812, "learning_rate": 9.804230269012254e-05, "loss": 0.9907, "step": 2181 }, { "epoch": 0.1478419947150891, "grad_norm": 8.119913101196289, "learning_rate": 9.804093367102472e-05, "loss": 1.1259, "step": 2182 }, { "epoch": 0.1479097499830612, "grad_norm": 9.505135536193848, "learning_rate": 9.80395646519269e-05, "loss": 1.0509, "step": 2183 }, { "epoch": 0.14797750525103326, "grad_norm": 8.420906066894531, "learning_rate": 9.803819563282908e-05, "loss": 1.1827, "step": 2184 }, { "epoch": 0.14804526051900535, "grad_norm": 9.860353469848633, "learning_rate": 9.803682661373128e-05, "loss": 1.0286, "step": 2185 }, { "epoch": 0.14811301578697744, "grad_norm": 8.259954452514648, "learning_rate": 9.803545759463346e-05, "loss": 1.0094, "step": 2186 }, { "epoch": 0.14818077105494953, "grad_norm": 10.46882438659668, "learning_rate": 9.803408857553564e-05, "loss": 1.1817, "step": 2187 }, { "epoch": 0.1482485263229216, "grad_norm": 9.315580368041992, "learning_rate": 9.803271955643782e-05, "loss": 1.2816, "step": 2188 }, { "epoch": 0.1483162815908937, "grad_norm": 10.408548355102539, "learning_rate": 9.803135053734001e-05, "loss": 1.0355, "step": 2189 }, { "epoch": 0.14838403685886578, "grad_norm": 9.682003021240234, "learning_rate": 9.802998151824219e-05, "loss": 1.0413, "step": 2190 }, { "epoch": 0.14845179212683787, "grad_norm": 10.041797637939453, "learning_rate": 9.802861249914437e-05, "loss": 0.8388, "step": 2191 }, { "epoch": 0.14851954739480994, "grad_norm": 8.367657661437988, "learning_rate": 9.802724348004655e-05, "loss": 1.0724, "step": 2192 }, { "epoch": 0.14858730266278203, "grad_norm": 9.9558744430542, "learning_rate": 9.802587446094873e-05, "loss": 1.3534, "step": 2193 }, { "epoch": 0.14865505793075412, "grad_norm": 9.244332313537598, "learning_rate": 9.802450544185093e-05, "loss": 1.366, "step": 2194 }, { "epoch": 0.1487228131987262, "grad_norm": 9.560718536376953, "learning_rate": 9.80231364227531e-05, "loss": 1.0556, "step": 2195 }, { "epoch": 0.14879056846669828, "grad_norm": 8.724915504455566, "learning_rate": 9.802176740365529e-05, "loss": 1.316, "step": 2196 }, { "epoch": 0.14885832373467037, "grad_norm": 9.468677520751953, "learning_rate": 9.802039838455747e-05, "loss": 1.1289, "step": 2197 }, { "epoch": 0.14892607900264246, "grad_norm": 8.56347942352295, "learning_rate": 9.801902936545966e-05, "loss": 1.0402, "step": 2198 }, { "epoch": 0.14899383427061455, "grad_norm": 9.328559875488281, "learning_rate": 9.801766034636184e-05, "loss": 1.186, "step": 2199 }, { "epoch": 0.14906158953858661, "grad_norm": 10.20579719543457, "learning_rate": 9.801629132726402e-05, "loss": 1.0896, "step": 2200 }, { "epoch": 0.1491293448065587, "grad_norm": 11.614208221435547, "learning_rate": 9.80149223081662e-05, "loss": 1.2742, "step": 2201 }, { "epoch": 0.1491971000745308, "grad_norm": 8.291358947753906, "learning_rate": 9.801355328906838e-05, "loss": 1.0299, "step": 2202 }, { "epoch": 0.1492648553425029, "grad_norm": 8.734344482421875, "learning_rate": 9.801218426997058e-05, "loss": 1.0922, "step": 2203 }, { "epoch": 0.14933261061047495, "grad_norm": 8.498616218566895, "learning_rate": 9.801081525087276e-05, "loss": 0.8436, "step": 2204 }, { "epoch": 0.14940036587844704, "grad_norm": 7.4501447677612305, "learning_rate": 9.800944623177494e-05, "loss": 0.8629, "step": 2205 }, { "epoch": 0.14946812114641914, "grad_norm": 11.652278900146484, "learning_rate": 9.800807721267712e-05, "loss": 1.1089, "step": 2206 }, { "epoch": 0.14953587641439123, "grad_norm": 11.043471336364746, "learning_rate": 9.80067081935793e-05, "loss": 1.2758, "step": 2207 }, { "epoch": 0.1496036316823633, "grad_norm": 9.376245498657227, "learning_rate": 9.800533917448149e-05, "loss": 1.0178, "step": 2208 }, { "epoch": 0.14967138695033538, "grad_norm": 8.678750038146973, "learning_rate": 9.800397015538367e-05, "loss": 1.0276, "step": 2209 }, { "epoch": 0.14973914221830747, "grad_norm": 10.233052253723145, "learning_rate": 9.800260113628585e-05, "loss": 1.0468, "step": 2210 }, { "epoch": 0.14980689748627957, "grad_norm": 10.711477279663086, "learning_rate": 9.800123211718803e-05, "loss": 1.3919, "step": 2211 }, { "epoch": 0.14987465275425163, "grad_norm": 10.130411148071289, "learning_rate": 9.799986309809023e-05, "loss": 1.1085, "step": 2212 }, { "epoch": 0.14994240802222372, "grad_norm": 9.64091682434082, "learning_rate": 9.79984940789924e-05, "loss": 1.0661, "step": 2213 }, { "epoch": 0.1500101632901958, "grad_norm": 9.80176830291748, "learning_rate": 9.799712505989459e-05, "loss": 1.3426, "step": 2214 }, { "epoch": 0.1500779185581679, "grad_norm": 9.34835147857666, "learning_rate": 9.799575604079677e-05, "loss": 1.0281, "step": 2215 }, { "epoch": 0.15014567382613997, "grad_norm": 8.638712882995605, "learning_rate": 9.799438702169895e-05, "loss": 0.9037, "step": 2216 }, { "epoch": 0.15021342909411206, "grad_norm": 10.499733924865723, "learning_rate": 9.799301800260114e-05, "loss": 1.1152, "step": 2217 }, { "epoch": 0.15028118436208415, "grad_norm": 9.093478202819824, "learning_rate": 9.799164898350332e-05, "loss": 1.2026, "step": 2218 }, { "epoch": 0.15034893963005624, "grad_norm": 13.181863784790039, "learning_rate": 9.79902799644055e-05, "loss": 1.3256, "step": 2219 }, { "epoch": 0.1504166948980283, "grad_norm": 10.221563339233398, "learning_rate": 9.798891094530768e-05, "loss": 1.1867, "step": 2220 }, { "epoch": 0.1504844501660004, "grad_norm": 9.512944221496582, "learning_rate": 9.798754192620988e-05, "loss": 1.1145, "step": 2221 }, { "epoch": 0.1505522054339725, "grad_norm": 8.991315841674805, "learning_rate": 9.798617290711206e-05, "loss": 1.0289, "step": 2222 }, { "epoch": 0.15061996070194458, "grad_norm": 9.003118515014648, "learning_rate": 9.798480388801424e-05, "loss": 0.9618, "step": 2223 }, { "epoch": 0.15068771596991667, "grad_norm": 11.337166786193848, "learning_rate": 9.798343486891643e-05, "loss": 1.3864, "step": 2224 }, { "epoch": 0.15075547123788874, "grad_norm": 9.193288803100586, "learning_rate": 9.798206584981861e-05, "loss": 1.1314, "step": 2225 }, { "epoch": 0.15082322650586083, "grad_norm": 10.137048721313477, "learning_rate": 9.798069683072079e-05, "loss": 1.3325, "step": 2226 }, { "epoch": 0.15089098177383292, "grad_norm": 8.248672485351562, "learning_rate": 9.797932781162299e-05, "loss": 1.0688, "step": 2227 }, { "epoch": 0.150958737041805, "grad_norm": 10.51007080078125, "learning_rate": 9.797795879252517e-05, "loss": 1.2191, "step": 2228 }, { "epoch": 0.15102649230977708, "grad_norm": 14.15853214263916, "learning_rate": 9.797658977342735e-05, "loss": 1.0961, "step": 2229 }, { "epoch": 0.15109424757774917, "grad_norm": 8.42485523223877, "learning_rate": 9.797522075432954e-05, "loss": 1.0314, "step": 2230 }, { "epoch": 0.15116200284572126, "grad_norm": 9.3310546875, "learning_rate": 9.797385173523172e-05, "loss": 1.2372, "step": 2231 }, { "epoch": 0.15122975811369335, "grad_norm": 9.323569297790527, "learning_rate": 9.79724827161339e-05, "loss": 1.1823, "step": 2232 }, { "epoch": 0.15129751338166542, "grad_norm": 9.719592094421387, "learning_rate": 9.797111369703608e-05, "loss": 1.0689, "step": 2233 }, { "epoch": 0.1513652686496375, "grad_norm": 9.37340259552002, "learning_rate": 9.796974467793826e-05, "loss": 1.0703, "step": 2234 }, { "epoch": 0.1514330239176096, "grad_norm": 9.803778648376465, "learning_rate": 9.796837565884045e-05, "loss": 1.0422, "step": 2235 }, { "epoch": 0.1515007791855817, "grad_norm": 8.317301750183105, "learning_rate": 9.796700663974264e-05, "loss": 0.9829, "step": 2236 }, { "epoch": 0.15156853445355375, "grad_norm": 8.473258972167969, "learning_rate": 9.796563762064482e-05, "loss": 1.064, "step": 2237 }, { "epoch": 0.15163628972152585, "grad_norm": 8.495006561279297, "learning_rate": 9.7964268601547e-05, "loss": 0.8646, "step": 2238 }, { "epoch": 0.15170404498949794, "grad_norm": 9.140898704528809, "learning_rate": 9.796289958244918e-05, "loss": 1.068, "step": 2239 }, { "epoch": 0.15177180025747003, "grad_norm": 10.264453887939453, "learning_rate": 9.796153056335137e-05, "loss": 0.9681, "step": 2240 }, { "epoch": 0.1518395555254421, "grad_norm": 10.407776832580566, "learning_rate": 9.796016154425355e-05, "loss": 1.2538, "step": 2241 }, { "epoch": 0.15190731079341419, "grad_norm": 10.143677711486816, "learning_rate": 9.795879252515573e-05, "loss": 1.138, "step": 2242 }, { "epoch": 0.15197506606138628, "grad_norm": 12.528799057006836, "learning_rate": 9.795742350605791e-05, "loss": 1.2654, "step": 2243 }, { "epoch": 0.15204282132935837, "grad_norm": 10.635498046875, "learning_rate": 9.79560544869601e-05, "loss": 1.2116, "step": 2244 }, { "epoch": 0.15211057659733043, "grad_norm": 10.71164608001709, "learning_rate": 9.795468546786229e-05, "loss": 1.0262, "step": 2245 }, { "epoch": 0.15217833186530252, "grad_norm": 9.121109962463379, "learning_rate": 9.795331644876447e-05, "loss": 1.1657, "step": 2246 }, { "epoch": 0.15224608713327462, "grad_norm": 8.174636840820312, "learning_rate": 9.795194742966665e-05, "loss": 1.0027, "step": 2247 }, { "epoch": 0.1523138424012467, "grad_norm": 8.763494491577148, "learning_rate": 9.795057841056883e-05, "loss": 1.1103, "step": 2248 }, { "epoch": 0.15238159766921877, "grad_norm": 8.028278350830078, "learning_rate": 9.794920939147102e-05, "loss": 1.0205, "step": 2249 }, { "epoch": 0.15244935293719086, "grad_norm": 8.96112060546875, "learning_rate": 9.79478403723732e-05, "loss": 1.2949, "step": 2250 }, { "epoch": 0.15251710820516295, "grad_norm": 8.67423152923584, "learning_rate": 9.794647135327538e-05, "loss": 1.0602, "step": 2251 }, { "epoch": 0.15258486347313505, "grad_norm": 9.45671272277832, "learning_rate": 9.794510233417756e-05, "loss": 1.0362, "step": 2252 }, { "epoch": 0.1526526187411071, "grad_norm": 10.24669075012207, "learning_rate": 9.794373331507976e-05, "loss": 0.9918, "step": 2253 }, { "epoch": 0.1527203740090792, "grad_norm": 9.014079093933105, "learning_rate": 9.794236429598194e-05, "loss": 1.1636, "step": 2254 }, { "epoch": 0.1527881292770513, "grad_norm": 6.800943851470947, "learning_rate": 9.794099527688412e-05, "loss": 0.9089, "step": 2255 }, { "epoch": 0.15285588454502338, "grad_norm": 9.90794849395752, "learning_rate": 9.79396262577863e-05, "loss": 0.7669, "step": 2256 }, { "epoch": 0.15292363981299545, "grad_norm": 9.870927810668945, "learning_rate": 9.793825723868848e-05, "loss": 1.1243, "step": 2257 }, { "epoch": 0.15299139508096754, "grad_norm": 9.707404136657715, "learning_rate": 9.793688821959067e-05, "loss": 1.1558, "step": 2258 }, { "epoch": 0.15305915034893963, "grad_norm": 8.362896919250488, "learning_rate": 9.793551920049285e-05, "loss": 0.8601, "step": 2259 }, { "epoch": 0.15312690561691172, "grad_norm": 9.536920547485352, "learning_rate": 9.793415018139503e-05, "loss": 1.2741, "step": 2260 }, { "epoch": 0.1531946608848838, "grad_norm": 11.108535766601562, "learning_rate": 9.793278116229721e-05, "loss": 1.3118, "step": 2261 }, { "epoch": 0.15326241615285588, "grad_norm": 7.281479358673096, "learning_rate": 9.793141214319939e-05, "loss": 0.9165, "step": 2262 }, { "epoch": 0.15333017142082797, "grad_norm": 9.166728973388672, "learning_rate": 9.793004312410159e-05, "loss": 1.0845, "step": 2263 }, { "epoch": 0.15339792668880006, "grad_norm": 11.539854049682617, "learning_rate": 9.792867410500377e-05, "loss": 1.1129, "step": 2264 }, { "epoch": 0.15346568195677213, "grad_norm": 8.588869094848633, "learning_rate": 9.792730508590595e-05, "loss": 0.9698, "step": 2265 }, { "epoch": 0.15353343722474422, "grad_norm": 8.270078659057617, "learning_rate": 9.792593606680813e-05, "loss": 1.2306, "step": 2266 }, { "epoch": 0.1536011924927163, "grad_norm": 7.908688545227051, "learning_rate": 9.792456704771032e-05, "loss": 1.283, "step": 2267 }, { "epoch": 0.1536689477606884, "grad_norm": 10.376410484313965, "learning_rate": 9.79231980286125e-05, "loss": 1.0246, "step": 2268 }, { "epoch": 0.15373670302866047, "grad_norm": 9.517715454101562, "learning_rate": 9.792182900951468e-05, "loss": 0.9975, "step": 2269 }, { "epoch": 0.15380445829663256, "grad_norm": 8.789438247680664, "learning_rate": 9.792045999041688e-05, "loss": 1.0886, "step": 2270 }, { "epoch": 0.15387221356460465, "grad_norm": 9.649114608764648, "learning_rate": 9.791909097131906e-05, "loss": 1.1497, "step": 2271 }, { "epoch": 0.15393996883257674, "grad_norm": 8.533876419067383, "learning_rate": 9.791772195222124e-05, "loss": 0.8701, "step": 2272 }, { "epoch": 0.1540077241005488, "grad_norm": 10.64561653137207, "learning_rate": 9.791635293312343e-05, "loss": 1.2755, "step": 2273 }, { "epoch": 0.1540754793685209, "grad_norm": 9.510658264160156, "learning_rate": 9.791498391402561e-05, "loss": 0.8772, "step": 2274 }, { "epoch": 0.154143234636493, "grad_norm": 9.849981307983398, "learning_rate": 9.791361489492779e-05, "loss": 1.1689, "step": 2275 }, { "epoch": 0.15421098990446508, "grad_norm": 8.152152061462402, "learning_rate": 9.791224587582998e-05, "loss": 0.8148, "step": 2276 }, { "epoch": 0.15427874517243717, "grad_norm": 8.653456687927246, "learning_rate": 9.791087685673216e-05, "loss": 1.0061, "step": 2277 }, { "epoch": 0.15434650044040923, "grad_norm": 6.928426742553711, "learning_rate": 9.790950783763435e-05, "loss": 0.9192, "step": 2278 }, { "epoch": 0.15441425570838133, "grad_norm": 9.606708526611328, "learning_rate": 9.790813881853653e-05, "loss": 0.9956, "step": 2279 }, { "epoch": 0.15448201097635342, "grad_norm": 9.42773151397705, "learning_rate": 9.79067697994387e-05, "loss": 0.9302, "step": 2280 }, { "epoch": 0.1545497662443255, "grad_norm": 8.248319625854492, "learning_rate": 9.79054007803409e-05, "loss": 1.2318, "step": 2281 }, { "epoch": 0.15461752151229757, "grad_norm": 9.706393241882324, "learning_rate": 9.790403176124308e-05, "loss": 1.1489, "step": 2282 }, { "epoch": 0.15468527678026966, "grad_norm": 8.88716983795166, "learning_rate": 9.790266274214526e-05, "loss": 0.8686, "step": 2283 }, { "epoch": 0.15475303204824176, "grad_norm": 8.596991539001465, "learning_rate": 9.790129372304744e-05, "loss": 1.0623, "step": 2284 }, { "epoch": 0.15482078731621385, "grad_norm": 9.262425422668457, "learning_rate": 9.789992470394962e-05, "loss": 1.2067, "step": 2285 }, { "epoch": 0.1548885425841859, "grad_norm": 7.772172927856445, "learning_rate": 9.789855568485181e-05, "loss": 0.9976, "step": 2286 }, { "epoch": 0.154956297852158, "grad_norm": 9.320001602172852, "learning_rate": 9.7897186665754e-05, "loss": 1.1377, "step": 2287 }, { "epoch": 0.1550240531201301, "grad_norm": 11.02434253692627, "learning_rate": 9.789581764665618e-05, "loss": 1.1365, "step": 2288 }, { "epoch": 0.1550918083881022, "grad_norm": 9.90654182434082, "learning_rate": 9.789444862755836e-05, "loss": 1.208, "step": 2289 }, { "epoch": 0.15515956365607425, "grad_norm": 9.591778755187988, "learning_rate": 9.789307960846055e-05, "loss": 1.2426, "step": 2290 }, { "epoch": 0.15522731892404634, "grad_norm": 9.221457481384277, "learning_rate": 9.789171058936273e-05, "loss": 1.1449, "step": 2291 }, { "epoch": 0.15529507419201843, "grad_norm": 7.294323444366455, "learning_rate": 9.789034157026491e-05, "loss": 0.8846, "step": 2292 }, { "epoch": 0.15536282945999053, "grad_norm": 10.6463041305542, "learning_rate": 9.788897255116709e-05, "loss": 1.1397, "step": 2293 }, { "epoch": 0.1554305847279626, "grad_norm": 9.412870407104492, "learning_rate": 9.788760353206927e-05, "loss": 1.1857, "step": 2294 }, { "epoch": 0.15549833999593468, "grad_norm": 9.163009643554688, "learning_rate": 9.788623451297147e-05, "loss": 0.8907, "step": 2295 }, { "epoch": 0.15556609526390677, "grad_norm": 8.157798767089844, "learning_rate": 9.788486549387365e-05, "loss": 0.893, "step": 2296 }, { "epoch": 0.15563385053187886, "grad_norm": 9.155460357666016, "learning_rate": 9.788349647477583e-05, "loss": 1.0361, "step": 2297 }, { "epoch": 0.15570160579985093, "grad_norm": 10.280989646911621, "learning_rate": 9.7882127455678e-05, "loss": 1.2953, "step": 2298 }, { "epoch": 0.15576936106782302, "grad_norm": 9.654706954956055, "learning_rate": 9.78807584365802e-05, "loss": 0.9012, "step": 2299 }, { "epoch": 0.1558371163357951, "grad_norm": 12.425939559936523, "learning_rate": 9.787938941748238e-05, "loss": 1.4103, "step": 2300 }, { "epoch": 0.1559048716037672, "grad_norm": 9.888884544372559, "learning_rate": 9.787802039838456e-05, "loss": 1.1599, "step": 2301 }, { "epoch": 0.15597262687173927, "grad_norm": 10.30229377746582, "learning_rate": 9.787665137928674e-05, "loss": 1.0505, "step": 2302 }, { "epoch": 0.15604038213971136, "grad_norm": 11.208477973937988, "learning_rate": 9.787528236018892e-05, "loss": 1.009, "step": 2303 }, { "epoch": 0.15610813740768345, "grad_norm": 9.264431953430176, "learning_rate": 9.787391334109112e-05, "loss": 1.1453, "step": 2304 }, { "epoch": 0.15617589267565554, "grad_norm": 9.0980224609375, "learning_rate": 9.78725443219933e-05, "loss": 1.0824, "step": 2305 }, { "epoch": 0.1562436479436276, "grad_norm": 8.346585273742676, "learning_rate": 9.787117530289548e-05, "loss": 1.0632, "step": 2306 }, { "epoch": 0.1563114032115997, "grad_norm": 10.607507705688477, "learning_rate": 9.786980628379766e-05, "loss": 1.0102, "step": 2307 }, { "epoch": 0.1563791584795718, "grad_norm": 8.189208984375, "learning_rate": 9.786843726469985e-05, "loss": 1.0872, "step": 2308 }, { "epoch": 0.15644691374754388, "grad_norm": 9.84202766418457, "learning_rate": 9.786706824560203e-05, "loss": 0.9871, "step": 2309 }, { "epoch": 0.15651466901551594, "grad_norm": 11.937589645385742, "learning_rate": 9.786569922650421e-05, "loss": 1.2255, "step": 2310 }, { "epoch": 0.15658242428348804, "grad_norm": 9.855016708374023, "learning_rate": 9.786433020740639e-05, "loss": 1.0097, "step": 2311 }, { "epoch": 0.15665017955146013, "grad_norm": 8.661060333251953, "learning_rate": 9.786296118830857e-05, "loss": 0.8732, "step": 2312 }, { "epoch": 0.15671793481943222, "grad_norm": 10.57170295715332, "learning_rate": 9.786159216921077e-05, "loss": 0.862, "step": 2313 }, { "epoch": 0.15678569008740428, "grad_norm": 7.759045600891113, "learning_rate": 9.786022315011295e-05, "loss": 0.77, "step": 2314 }, { "epoch": 0.15685344535537638, "grad_norm": 10.758045196533203, "learning_rate": 9.785885413101513e-05, "loss": 0.8919, "step": 2315 }, { "epoch": 0.15692120062334847, "grad_norm": 8.521660804748535, "learning_rate": 9.78574851119173e-05, "loss": 1.131, "step": 2316 }, { "epoch": 0.15698895589132056, "grad_norm": 8.72917652130127, "learning_rate": 9.78561160928195e-05, "loss": 0.8359, "step": 2317 }, { "epoch": 0.15705671115929262, "grad_norm": 11.679365158081055, "learning_rate": 9.785474707372168e-05, "loss": 1.0078, "step": 2318 }, { "epoch": 0.15712446642726471, "grad_norm": 11.50632381439209, "learning_rate": 9.785337805462386e-05, "loss": 1.208, "step": 2319 }, { "epoch": 0.1571922216952368, "grad_norm": 9.6107759475708, "learning_rate": 9.785200903552605e-05, "loss": 1.0967, "step": 2320 }, { "epoch": 0.1572599769632089, "grad_norm": 8.629117012023926, "learning_rate": 9.785064001642824e-05, "loss": 1.0594, "step": 2321 }, { "epoch": 0.15732773223118096, "grad_norm": 11.136920928955078, "learning_rate": 9.784927099733042e-05, "loss": 1.2874, "step": 2322 }, { "epoch": 0.15739548749915305, "grad_norm": 11.097023963928223, "learning_rate": 9.784790197823261e-05, "loss": 1.1598, "step": 2323 }, { "epoch": 0.15746324276712514, "grad_norm": 11.117433547973633, "learning_rate": 9.784653295913479e-05, "loss": 1.0601, "step": 2324 }, { "epoch": 0.15753099803509724, "grad_norm": 10.152684211730957, "learning_rate": 9.784516394003697e-05, "loss": 1.2763, "step": 2325 }, { "epoch": 0.1575987533030693, "grad_norm": 8.37531852722168, "learning_rate": 9.784379492093915e-05, "loss": 1.1055, "step": 2326 }, { "epoch": 0.1576665085710414, "grad_norm": 10.463939666748047, "learning_rate": 9.784242590184134e-05, "loss": 1.3088, "step": 2327 }, { "epoch": 0.15773426383901348, "grad_norm": 8.74315357208252, "learning_rate": 9.784105688274352e-05, "loss": 1.1165, "step": 2328 }, { "epoch": 0.15780201910698557, "grad_norm": 8.691280364990234, "learning_rate": 9.78396878636457e-05, "loss": 1.1019, "step": 2329 }, { "epoch": 0.15786977437495767, "grad_norm": 10.424938201904297, "learning_rate": 9.783831884454789e-05, "loss": 1.1957, "step": 2330 }, { "epoch": 0.15793752964292973, "grad_norm": 6.867722034454346, "learning_rate": 9.783694982545008e-05, "loss": 1.0476, "step": 2331 }, { "epoch": 0.15800528491090182, "grad_norm": 9.434804916381836, "learning_rate": 9.783558080635226e-05, "loss": 1.0247, "step": 2332 }, { "epoch": 0.1580730401788739, "grad_norm": 7.771170616149902, "learning_rate": 9.783421178725444e-05, "loss": 1.2582, "step": 2333 }, { "epoch": 0.158140795446846, "grad_norm": 8.366608619689941, "learning_rate": 9.783284276815662e-05, "loss": 1.1078, "step": 2334 }, { "epoch": 0.15820855071481807, "grad_norm": 8.85851764678955, "learning_rate": 9.78314737490588e-05, "loss": 1.2405, "step": 2335 }, { "epoch": 0.15827630598279016, "grad_norm": 9.134325981140137, "learning_rate": 9.7830104729961e-05, "loss": 1.2277, "step": 2336 }, { "epoch": 0.15834406125076225, "grad_norm": 9.150130271911621, "learning_rate": 9.782873571086317e-05, "loss": 1.1355, "step": 2337 }, { "epoch": 0.15841181651873434, "grad_norm": 8.687469482421875, "learning_rate": 9.782736669176536e-05, "loss": 1.0454, "step": 2338 }, { "epoch": 0.1584795717867064, "grad_norm": 10.071285247802734, "learning_rate": 9.782599767266754e-05, "loss": 1.0041, "step": 2339 }, { "epoch": 0.1585473270546785, "grad_norm": 8.373369216918945, "learning_rate": 9.782462865356972e-05, "loss": 0.8378, "step": 2340 }, { "epoch": 0.1586150823226506, "grad_norm": 9.427014350891113, "learning_rate": 9.782325963447191e-05, "loss": 1.1413, "step": 2341 }, { "epoch": 0.15868283759062268, "grad_norm": 8.38814640045166, "learning_rate": 9.782189061537409e-05, "loss": 1.0696, "step": 2342 }, { "epoch": 0.15875059285859475, "grad_norm": 8.518644332885742, "learning_rate": 9.782052159627627e-05, "loss": 0.9814, "step": 2343 }, { "epoch": 0.15881834812656684, "grad_norm": 9.926252365112305, "learning_rate": 9.781915257717845e-05, "loss": 1.1971, "step": 2344 }, { "epoch": 0.15888610339453893, "grad_norm": 7.946019649505615, "learning_rate": 9.781778355808064e-05, "loss": 1.1902, "step": 2345 }, { "epoch": 0.15895385866251102, "grad_norm": 8.686945915222168, "learning_rate": 9.781641453898283e-05, "loss": 0.9939, "step": 2346 }, { "epoch": 0.15902161393048309, "grad_norm": 8.225680351257324, "learning_rate": 9.7815045519885e-05, "loss": 1.3665, "step": 2347 }, { "epoch": 0.15908936919845518, "grad_norm": 10.381987571716309, "learning_rate": 9.781367650078719e-05, "loss": 1.09, "step": 2348 }, { "epoch": 0.15915712446642727, "grad_norm": 8.57552719116211, "learning_rate": 9.781230748168937e-05, "loss": 0.8261, "step": 2349 }, { "epoch": 0.15922487973439936, "grad_norm": 9.105220794677734, "learning_rate": 9.781093846259156e-05, "loss": 1.102, "step": 2350 }, { "epoch": 0.15929263500237142, "grad_norm": 10.08092212677002, "learning_rate": 9.780956944349374e-05, "loss": 0.943, "step": 2351 }, { "epoch": 0.15936039027034352, "grad_norm": 10.259852409362793, "learning_rate": 9.780820042439592e-05, "loss": 0.8822, "step": 2352 }, { "epoch": 0.1594281455383156, "grad_norm": 8.31139087677002, "learning_rate": 9.78068314052981e-05, "loss": 1.2167, "step": 2353 }, { "epoch": 0.1594959008062877, "grad_norm": 7.529703617095947, "learning_rate": 9.78054623862003e-05, "loss": 0.8913, "step": 2354 }, { "epoch": 0.15956365607425976, "grad_norm": 8.792675971984863, "learning_rate": 9.780409336710248e-05, "loss": 0.966, "step": 2355 }, { "epoch": 0.15963141134223185, "grad_norm": 9.329866409301758, "learning_rate": 9.780272434800466e-05, "loss": 1.1659, "step": 2356 }, { "epoch": 0.15969916661020395, "grad_norm": 12.14089298248291, "learning_rate": 9.780135532890684e-05, "loss": 1.2019, "step": 2357 }, { "epoch": 0.15976692187817604, "grad_norm": 9.12912654876709, "learning_rate": 9.779998630980902e-05, "loss": 1.1615, "step": 2358 }, { "epoch": 0.1598346771461481, "grad_norm": 9.554464340209961, "learning_rate": 9.779861729071121e-05, "loss": 1.1695, "step": 2359 }, { "epoch": 0.1599024324141202, "grad_norm": 9.317673683166504, "learning_rate": 9.779724827161339e-05, "loss": 0.9606, "step": 2360 }, { "epoch": 0.15997018768209229, "grad_norm": 8.80395793914795, "learning_rate": 9.779587925251557e-05, "loss": 0.9693, "step": 2361 }, { "epoch": 0.16003794295006438, "grad_norm": 11.990642547607422, "learning_rate": 9.779451023341775e-05, "loss": 1.2901, "step": 2362 }, { "epoch": 0.16010569821803644, "grad_norm": 8.188547134399414, "learning_rate": 9.779314121431995e-05, "loss": 0.979, "step": 2363 }, { "epoch": 0.16017345348600853, "grad_norm": 8.316620826721191, "learning_rate": 9.779177219522213e-05, "loss": 0.8601, "step": 2364 }, { "epoch": 0.16024120875398062, "grad_norm": 7.58405876159668, "learning_rate": 9.77904031761243e-05, "loss": 1.1812, "step": 2365 }, { "epoch": 0.16030896402195272, "grad_norm": 7.725598335266113, "learning_rate": 9.77890341570265e-05, "loss": 0.9335, "step": 2366 }, { "epoch": 0.16037671928992478, "grad_norm": 8.6231107711792, "learning_rate": 9.778766513792868e-05, "loss": 1.0162, "step": 2367 }, { "epoch": 0.16044447455789687, "grad_norm": 9.762526512145996, "learning_rate": 9.778629611883086e-05, "loss": 1.3186, "step": 2368 }, { "epoch": 0.16051222982586896, "grad_norm": 11.384220123291016, "learning_rate": 9.778492709973305e-05, "loss": 1.2397, "step": 2369 }, { "epoch": 0.16057998509384105, "grad_norm": 8.841899871826172, "learning_rate": 9.778355808063523e-05, "loss": 0.8708, "step": 2370 }, { "epoch": 0.16064774036181312, "grad_norm": 7.778527736663818, "learning_rate": 9.778218906153741e-05, "loss": 0.9968, "step": 2371 }, { "epoch": 0.1607154956297852, "grad_norm": 8.559181213378906, "learning_rate": 9.77808200424396e-05, "loss": 0.9759, "step": 2372 }, { "epoch": 0.1607832508977573, "grad_norm": 10.273273468017578, "learning_rate": 9.777945102334179e-05, "loss": 0.9485, "step": 2373 }, { "epoch": 0.1608510061657294, "grad_norm": 7.946044445037842, "learning_rate": 9.777808200424397e-05, "loss": 1.105, "step": 2374 }, { "epoch": 0.16091876143370146, "grad_norm": 9.917662620544434, "learning_rate": 9.777671298514615e-05, "loss": 1.0272, "step": 2375 }, { "epoch": 0.16098651670167355, "grad_norm": 10.438239097595215, "learning_rate": 9.777534396604833e-05, "loss": 1.0197, "step": 2376 }, { "epoch": 0.16105427196964564, "grad_norm": 8.799901962280273, "learning_rate": 9.777397494695052e-05, "loss": 1.1401, "step": 2377 }, { "epoch": 0.16112202723761773, "grad_norm": 8.569243431091309, "learning_rate": 9.77726059278527e-05, "loss": 1.2019, "step": 2378 }, { "epoch": 0.1611897825055898, "grad_norm": 10.793002128601074, "learning_rate": 9.777123690875488e-05, "loss": 1.0932, "step": 2379 }, { "epoch": 0.1612575377735619, "grad_norm": 7.825850963592529, "learning_rate": 9.776986788965707e-05, "loss": 1.2226, "step": 2380 }, { "epoch": 0.16132529304153398, "grad_norm": 8.780813217163086, "learning_rate": 9.776849887055925e-05, "loss": 0.7939, "step": 2381 }, { "epoch": 0.16139304830950607, "grad_norm": 8.927638053894043, "learning_rate": 9.776712985146144e-05, "loss": 1.0551, "step": 2382 }, { "epoch": 0.16146080357747816, "grad_norm": 9.08043384552002, "learning_rate": 9.776576083236362e-05, "loss": 1.0037, "step": 2383 }, { "epoch": 0.16152855884545023, "grad_norm": 9.362268447875977, "learning_rate": 9.77643918132658e-05, "loss": 1.332, "step": 2384 }, { "epoch": 0.16159631411342232, "grad_norm": 10.533197402954102, "learning_rate": 9.776302279416798e-05, "loss": 1.0966, "step": 2385 }, { "epoch": 0.1616640693813944, "grad_norm": 9.579266548156738, "learning_rate": 9.776165377507017e-05, "loss": 0.9325, "step": 2386 }, { "epoch": 0.1617318246493665, "grad_norm": 10.071297645568848, "learning_rate": 9.776028475597235e-05, "loss": 0.9496, "step": 2387 }, { "epoch": 0.16179957991733857, "grad_norm": 9.243900299072266, "learning_rate": 9.775891573687453e-05, "loss": 1.13, "step": 2388 }, { "epoch": 0.16186733518531066, "grad_norm": 9.327018737792969, "learning_rate": 9.775754671777672e-05, "loss": 1.2587, "step": 2389 }, { "epoch": 0.16193509045328275, "grad_norm": 7.614950180053711, "learning_rate": 9.77561776986789e-05, "loss": 1.0848, "step": 2390 }, { "epoch": 0.16200284572125484, "grad_norm": 9.86501407623291, "learning_rate": 9.775480867958109e-05, "loss": 1.2504, "step": 2391 }, { "epoch": 0.1620706009892269, "grad_norm": 9.08303451538086, "learning_rate": 9.775343966048327e-05, "loss": 1.4359, "step": 2392 }, { "epoch": 0.162138356257199, "grad_norm": 8.417489051818848, "learning_rate": 9.775207064138545e-05, "loss": 1.244, "step": 2393 }, { "epoch": 0.1622061115251711, "grad_norm": 8.35366439819336, "learning_rate": 9.775070162228763e-05, "loss": 1.0706, "step": 2394 }, { "epoch": 0.16227386679314318, "grad_norm": 9.732915878295898, "learning_rate": 9.774933260318981e-05, "loss": 1.4237, "step": 2395 }, { "epoch": 0.16234162206111524, "grad_norm": 9.131365776062012, "learning_rate": 9.7747963584092e-05, "loss": 0.9207, "step": 2396 }, { "epoch": 0.16240937732908733, "grad_norm": 8.93538761138916, "learning_rate": 9.774659456499419e-05, "loss": 1.0054, "step": 2397 }, { "epoch": 0.16247713259705943, "grad_norm": 8.939055442810059, "learning_rate": 9.774522554589637e-05, "loss": 0.9782, "step": 2398 }, { "epoch": 0.16254488786503152, "grad_norm": 9.251758575439453, "learning_rate": 9.774385652679855e-05, "loss": 1.0968, "step": 2399 }, { "epoch": 0.16261264313300358, "grad_norm": 9.240782737731934, "learning_rate": 9.774248750770074e-05, "loss": 0.7982, "step": 2400 }, { "epoch": 0.16268039840097567, "grad_norm": 8.697726249694824, "learning_rate": 9.774111848860292e-05, "loss": 1.0545, "step": 2401 }, { "epoch": 0.16274815366894776, "grad_norm": 7.9362053871154785, "learning_rate": 9.77397494695051e-05, "loss": 0.8364, "step": 2402 }, { "epoch": 0.16281590893691986, "grad_norm": 11.944025993347168, "learning_rate": 9.773838045040728e-05, "loss": 1.2313, "step": 2403 }, { "epoch": 0.16288366420489192, "grad_norm": 9.496225357055664, "learning_rate": 9.773701143130946e-05, "loss": 1.1363, "step": 2404 }, { "epoch": 0.162951419472864, "grad_norm": 12.96069622039795, "learning_rate": 9.773564241221165e-05, "loss": 1.1205, "step": 2405 }, { "epoch": 0.1630191747408361, "grad_norm": 8.915671348571777, "learning_rate": 9.773427339311384e-05, "loss": 1.1149, "step": 2406 }, { "epoch": 0.1630869300088082, "grad_norm": 10.23763370513916, "learning_rate": 9.773290437401602e-05, "loss": 1.1867, "step": 2407 }, { "epoch": 0.16315468527678026, "grad_norm": 9.117730140686035, "learning_rate": 9.77315353549182e-05, "loss": 1.1792, "step": 2408 }, { "epoch": 0.16322244054475235, "grad_norm": 9.380385398864746, "learning_rate": 9.773016633582039e-05, "loss": 1.1621, "step": 2409 }, { "epoch": 0.16329019581272444, "grad_norm": 8.56508731842041, "learning_rate": 9.772879731672257e-05, "loss": 0.8372, "step": 2410 }, { "epoch": 0.16335795108069653, "grad_norm": 11.707832336425781, "learning_rate": 9.772742829762475e-05, "loss": 1.3191, "step": 2411 }, { "epoch": 0.1634257063486686, "grad_norm": 7.720577716827393, "learning_rate": 9.772605927852694e-05, "loss": 0.9617, "step": 2412 }, { "epoch": 0.1634934616166407, "grad_norm": 8.586542129516602, "learning_rate": 9.772469025942912e-05, "loss": 1.0978, "step": 2413 }, { "epoch": 0.16356121688461278, "grad_norm": 9.021394729614258, "learning_rate": 9.77233212403313e-05, "loss": 0.9358, "step": 2414 }, { "epoch": 0.16362897215258487, "grad_norm": 9.078686714172363, "learning_rate": 9.77219522212335e-05, "loss": 1.3338, "step": 2415 }, { "epoch": 0.16369672742055694, "grad_norm": 9.810312271118164, "learning_rate": 9.772058320213568e-05, "loss": 1.2472, "step": 2416 }, { "epoch": 0.16376448268852903, "grad_norm": 8.933609962463379, "learning_rate": 9.771921418303786e-05, "loss": 0.9115, "step": 2417 }, { "epoch": 0.16383223795650112, "grad_norm": 7.044286251068115, "learning_rate": 9.771784516394005e-05, "loss": 0.7983, "step": 2418 }, { "epoch": 0.1638999932244732, "grad_norm": 11.711495399475098, "learning_rate": 9.771647614484223e-05, "loss": 1.096, "step": 2419 }, { "epoch": 0.16396774849244528, "grad_norm": 9.31049633026123, "learning_rate": 9.771510712574441e-05, "loss": 1.2711, "step": 2420 }, { "epoch": 0.16403550376041737, "grad_norm": 8.10503101348877, "learning_rate": 9.77137381066466e-05, "loss": 0.9095, "step": 2421 }, { "epoch": 0.16410325902838946, "grad_norm": 7.915055751800537, "learning_rate": 9.771236908754877e-05, "loss": 1.0161, "step": 2422 }, { "epoch": 0.16417101429636155, "grad_norm": 8.185515403747559, "learning_rate": 9.771100006845097e-05, "loss": 1.1109, "step": 2423 }, { "epoch": 0.16423876956433361, "grad_norm": 9.960200309753418, "learning_rate": 9.770963104935315e-05, "loss": 0.9757, "step": 2424 }, { "epoch": 0.1643065248323057, "grad_norm": 9.646814346313477, "learning_rate": 9.770826203025533e-05, "loss": 1.0835, "step": 2425 }, { "epoch": 0.1643742801002778, "grad_norm": 9.701393127441406, "learning_rate": 9.770689301115751e-05, "loss": 1.0717, "step": 2426 }, { "epoch": 0.1644420353682499, "grad_norm": 7.887824058532715, "learning_rate": 9.770552399205969e-05, "loss": 1.0937, "step": 2427 }, { "epoch": 0.16450979063622195, "grad_norm": 10.94339370727539, "learning_rate": 9.770415497296188e-05, "loss": 1.2766, "step": 2428 }, { "epoch": 0.16457754590419404, "grad_norm": 10.051490783691406, "learning_rate": 9.770278595386406e-05, "loss": 1.248, "step": 2429 }, { "epoch": 0.16464530117216614, "grad_norm": 8.380006790161133, "learning_rate": 9.770141693476624e-05, "loss": 1.1657, "step": 2430 }, { "epoch": 0.16471305644013823, "grad_norm": 8.077753067016602, "learning_rate": 9.770004791566843e-05, "loss": 0.9511, "step": 2431 }, { "epoch": 0.1647808117081103, "grad_norm": 8.744999885559082, "learning_rate": 9.769867889657062e-05, "loss": 1.0736, "step": 2432 }, { "epoch": 0.16484856697608238, "grad_norm": 8.203909873962402, "learning_rate": 9.76973098774728e-05, "loss": 1.208, "step": 2433 }, { "epoch": 0.16491632224405448, "grad_norm": 9.462398529052734, "learning_rate": 9.769594085837498e-05, "loss": 1.0011, "step": 2434 }, { "epoch": 0.16498407751202657, "grad_norm": 11.190359115600586, "learning_rate": 9.769457183927716e-05, "loss": 1.139, "step": 2435 }, { "epoch": 0.16505183277999866, "grad_norm": 10.454118728637695, "learning_rate": 9.769320282017934e-05, "loss": 1.0799, "step": 2436 }, { "epoch": 0.16511958804797072, "grad_norm": 14.411054611206055, "learning_rate": 9.769183380108153e-05, "loss": 1.0369, "step": 2437 }, { "epoch": 0.16518734331594281, "grad_norm": 11.42679214477539, "learning_rate": 9.769046478198371e-05, "loss": 1.2486, "step": 2438 }, { "epoch": 0.1652550985839149, "grad_norm": 10.520325660705566, "learning_rate": 9.76890957628859e-05, "loss": 1.2355, "step": 2439 }, { "epoch": 0.165322853851887, "grad_norm": 7.958998680114746, "learning_rate": 9.768772674378808e-05, "loss": 1.0822, "step": 2440 }, { "epoch": 0.16539060911985906, "grad_norm": 8.649806022644043, "learning_rate": 9.768635772469027e-05, "loss": 0.9439, "step": 2441 }, { "epoch": 0.16545836438783115, "grad_norm": 8.539712905883789, "learning_rate": 9.768498870559245e-05, "loss": 0.9541, "step": 2442 }, { "epoch": 0.16552611965580324, "grad_norm": 11.40131950378418, "learning_rate": 9.768361968649463e-05, "loss": 1.0568, "step": 2443 }, { "epoch": 0.16559387492377534, "grad_norm": 8.333579063415527, "learning_rate": 9.768225066739681e-05, "loss": 0.924, "step": 2444 }, { "epoch": 0.1656616301917474, "grad_norm": 9.02564525604248, "learning_rate": 9.768088164829899e-05, "loss": 1.1901, "step": 2445 }, { "epoch": 0.1657293854597195, "grad_norm": 9.721306800842285, "learning_rate": 9.767951262920118e-05, "loss": 1.2313, "step": 2446 }, { "epoch": 0.16579714072769158, "grad_norm": 9.876781463623047, "learning_rate": 9.767814361010336e-05, "loss": 1.0165, "step": 2447 }, { "epoch": 0.16586489599566367, "grad_norm": 11.69865894317627, "learning_rate": 9.767677459100555e-05, "loss": 1.1301, "step": 2448 }, { "epoch": 0.16593265126363574, "grad_norm": 11.38391399383545, "learning_rate": 9.767540557190773e-05, "loss": 1.1522, "step": 2449 }, { "epoch": 0.16600040653160783, "grad_norm": 9.18020248413086, "learning_rate": 9.76740365528099e-05, "loss": 1.0515, "step": 2450 }, { "epoch": 0.16606816179957992, "grad_norm": 9.30802059173584, "learning_rate": 9.76726675337121e-05, "loss": 1.0789, "step": 2451 }, { "epoch": 0.166135917067552, "grad_norm": 9.58259391784668, "learning_rate": 9.767129851461428e-05, "loss": 1.1299, "step": 2452 }, { "epoch": 0.16620367233552408, "grad_norm": 9.137594223022461, "learning_rate": 9.766992949551646e-05, "loss": 1.0854, "step": 2453 }, { "epoch": 0.16627142760349617, "grad_norm": 12.410299301147461, "learning_rate": 9.766856047641864e-05, "loss": 1.1717, "step": 2454 }, { "epoch": 0.16633918287146826, "grad_norm": 9.016322135925293, "learning_rate": 9.766719145732083e-05, "loss": 1.1028, "step": 2455 }, { "epoch": 0.16640693813944035, "grad_norm": 10.166184425354004, "learning_rate": 9.766582243822301e-05, "loss": 1.0626, "step": 2456 }, { "epoch": 0.16647469340741242, "grad_norm": 9.030965805053711, "learning_rate": 9.76644534191252e-05, "loss": 1.1761, "step": 2457 }, { "epoch": 0.1665424486753845, "grad_norm": 9.294576644897461, "learning_rate": 9.766308440002739e-05, "loss": 0.943, "step": 2458 }, { "epoch": 0.1666102039433566, "grad_norm": 6.732856273651123, "learning_rate": 9.766171538092957e-05, "loss": 0.7342, "step": 2459 }, { "epoch": 0.1666779592113287, "grad_norm": 8.178942680358887, "learning_rate": 9.766034636183175e-05, "loss": 0.7889, "step": 2460 }, { "epoch": 0.16674571447930076, "grad_norm": 7.771929740905762, "learning_rate": 9.765897734273394e-05, "loss": 1.1129, "step": 2461 }, { "epoch": 0.16681346974727285, "grad_norm": 8.695874214172363, "learning_rate": 9.765760832363612e-05, "loss": 1.1968, "step": 2462 }, { "epoch": 0.16688122501524494, "grad_norm": 10.18800163269043, "learning_rate": 9.76562393045383e-05, "loss": 1.0835, "step": 2463 }, { "epoch": 0.16694898028321703, "grad_norm": 9.310625076293945, "learning_rate": 9.76548702854405e-05, "loss": 1.2042, "step": 2464 }, { "epoch": 0.1670167355511891, "grad_norm": 11.684195518493652, "learning_rate": 9.765350126634268e-05, "loss": 1.2518, "step": 2465 }, { "epoch": 0.16708449081916119, "grad_norm": 10.9056978225708, "learning_rate": 9.765213224724486e-05, "loss": 1.2465, "step": 2466 }, { "epoch": 0.16715224608713328, "grad_norm": 9.871830940246582, "learning_rate": 9.765076322814704e-05, "loss": 1.3192, "step": 2467 }, { "epoch": 0.16722000135510537, "grad_norm": 7.725397109985352, "learning_rate": 9.764939420904922e-05, "loss": 0.9082, "step": 2468 }, { "epoch": 0.16728775662307743, "grad_norm": 8.7415189743042, "learning_rate": 9.764802518995141e-05, "loss": 0.8569, "step": 2469 }, { "epoch": 0.16735551189104952, "grad_norm": 7.8990888595581055, "learning_rate": 9.76466561708536e-05, "loss": 1.0175, "step": 2470 }, { "epoch": 0.16742326715902162, "grad_norm": 8.688074111938477, "learning_rate": 9.764528715175577e-05, "loss": 1.0131, "step": 2471 }, { "epoch": 0.1674910224269937, "grad_norm": 7.902133941650391, "learning_rate": 9.764391813265795e-05, "loss": 0.8956, "step": 2472 }, { "epoch": 0.16755877769496577, "grad_norm": 9.179505348205566, "learning_rate": 9.764254911356013e-05, "loss": 1.175, "step": 2473 }, { "epoch": 0.16762653296293786, "grad_norm": 10.701058387756348, "learning_rate": 9.764118009446233e-05, "loss": 1.003, "step": 2474 }, { "epoch": 0.16769428823090995, "grad_norm": 8.76916217803955, "learning_rate": 9.763981107536451e-05, "loss": 1.102, "step": 2475 }, { "epoch": 0.16776204349888205, "grad_norm": 8.682199478149414, "learning_rate": 9.763844205626669e-05, "loss": 1.2042, "step": 2476 }, { "epoch": 0.1678297987668541, "grad_norm": 9.789544105529785, "learning_rate": 9.763707303716887e-05, "loss": 1.2469, "step": 2477 }, { "epoch": 0.1678975540348262, "grad_norm": 8.250391960144043, "learning_rate": 9.763570401807106e-05, "loss": 0.9742, "step": 2478 }, { "epoch": 0.1679653093027983, "grad_norm": 8.938610076904297, "learning_rate": 9.763433499897324e-05, "loss": 0.9968, "step": 2479 }, { "epoch": 0.16803306457077039, "grad_norm": 9.956622123718262, "learning_rate": 9.763296597987542e-05, "loss": 1.0464, "step": 2480 }, { "epoch": 0.16810081983874245, "grad_norm": 8.582858085632324, "learning_rate": 9.76315969607776e-05, "loss": 1.1915, "step": 2481 }, { "epoch": 0.16816857510671454, "grad_norm": 7.73312520980835, "learning_rate": 9.763022794167979e-05, "loss": 0.8506, "step": 2482 }, { "epoch": 0.16823633037468663, "grad_norm": 8.986891746520996, "learning_rate": 9.762885892258198e-05, "loss": 0.8959, "step": 2483 }, { "epoch": 0.16830408564265872, "grad_norm": 10.047099113464355, "learning_rate": 9.762748990348416e-05, "loss": 1.24, "step": 2484 }, { "epoch": 0.1683718409106308, "grad_norm": 8.517911911010742, "learning_rate": 9.762612088438634e-05, "loss": 0.8171, "step": 2485 }, { "epoch": 0.16843959617860288, "grad_norm": 9.586174964904785, "learning_rate": 9.762475186528852e-05, "loss": 1.1603, "step": 2486 }, { "epoch": 0.16850735144657497, "grad_norm": 9.85086441040039, "learning_rate": 9.762338284619071e-05, "loss": 0.9157, "step": 2487 }, { "epoch": 0.16857510671454706, "grad_norm": 12.023639678955078, "learning_rate": 9.76220138270929e-05, "loss": 1.2032, "step": 2488 }, { "epoch": 0.16864286198251915, "grad_norm": 8.932641983032227, "learning_rate": 9.762064480799507e-05, "loss": 1.0678, "step": 2489 }, { "epoch": 0.16871061725049122, "grad_norm": 10.568282127380371, "learning_rate": 9.761927578889725e-05, "loss": 1.3213, "step": 2490 }, { "epoch": 0.1687783725184633, "grad_norm": 9.116564750671387, "learning_rate": 9.761790676979944e-05, "loss": 1.0525, "step": 2491 }, { "epoch": 0.1688461277864354, "grad_norm": 8.192644119262695, "learning_rate": 9.761653775070163e-05, "loss": 0.9554, "step": 2492 }, { "epoch": 0.1689138830544075, "grad_norm": 9.146562576293945, "learning_rate": 9.761516873160381e-05, "loss": 0.8852, "step": 2493 }, { "epoch": 0.16898163832237956, "grad_norm": 8.82610034942627, "learning_rate": 9.761379971250599e-05, "loss": 1.0521, "step": 2494 }, { "epoch": 0.16904939359035165, "grad_norm": 9.051412582397461, "learning_rate": 9.761243069340817e-05, "loss": 0.913, "step": 2495 }, { "epoch": 0.16911714885832374, "grad_norm": 8.97696304321289, "learning_rate": 9.761106167431036e-05, "loss": 0.887, "step": 2496 }, { "epoch": 0.16918490412629583, "grad_norm": 10.083110809326172, "learning_rate": 9.760969265521254e-05, "loss": 1.3521, "step": 2497 }, { "epoch": 0.1692526593942679, "grad_norm": 7.585256099700928, "learning_rate": 9.760832363611472e-05, "loss": 0.9001, "step": 2498 }, { "epoch": 0.16932041466224, "grad_norm": 10.301995277404785, "learning_rate": 9.76069546170169e-05, "loss": 1.0915, "step": 2499 }, { "epoch": 0.16938816993021208, "grad_norm": 8.517580032348633, "learning_rate": 9.760558559791909e-05, "loss": 1.0097, "step": 2500 }, { "epoch": 0.16945592519818417, "grad_norm": 8.20002555847168, "learning_rate": 9.760421657882128e-05, "loss": 1.2988, "step": 2501 }, { "epoch": 0.16952368046615623, "grad_norm": 9.705124855041504, "learning_rate": 9.760284755972346e-05, "loss": 1.2661, "step": 2502 }, { "epoch": 0.16959143573412833, "grad_norm": 9.167060852050781, "learning_rate": 9.760147854062564e-05, "loss": 1.0556, "step": 2503 }, { "epoch": 0.16965919100210042, "grad_norm": 8.415916442871094, "learning_rate": 9.760010952152783e-05, "loss": 0.7964, "step": 2504 }, { "epoch": 0.1697269462700725, "grad_norm": 7.626298427581787, "learning_rate": 9.759874050243001e-05, "loss": 0.9153, "step": 2505 }, { "epoch": 0.16979470153804457, "grad_norm": 10.595001220703125, "learning_rate": 9.75973714833322e-05, "loss": 0.9911, "step": 2506 }, { "epoch": 0.16986245680601667, "grad_norm": 7.564423084259033, "learning_rate": 9.759600246423439e-05, "loss": 0.9843, "step": 2507 }, { "epoch": 0.16993021207398876, "grad_norm": 7.644829273223877, "learning_rate": 9.759463344513657e-05, "loss": 0.9116, "step": 2508 }, { "epoch": 0.16999796734196085, "grad_norm": 7.54351282119751, "learning_rate": 9.759326442603875e-05, "loss": 1.1528, "step": 2509 }, { "epoch": 0.1700657226099329, "grad_norm": 9.259818077087402, "learning_rate": 9.759189540694094e-05, "loss": 1.0867, "step": 2510 }, { "epoch": 0.170133477877905, "grad_norm": 8.022993087768555, "learning_rate": 9.759052638784312e-05, "loss": 0.7524, "step": 2511 }, { "epoch": 0.1702012331458771, "grad_norm": 8.264616966247559, "learning_rate": 9.75891573687453e-05, "loss": 0.9906, "step": 2512 }, { "epoch": 0.1702689884138492, "grad_norm": 9.606420516967773, "learning_rate": 9.758778834964748e-05, "loss": 0.9287, "step": 2513 }, { "epoch": 0.17033674368182125, "grad_norm": 8.019355773925781, "learning_rate": 9.758641933054966e-05, "loss": 0.879, "step": 2514 }, { "epoch": 0.17040449894979334, "grad_norm": 9.819777488708496, "learning_rate": 9.758505031145186e-05, "loss": 1.3677, "step": 2515 }, { "epoch": 0.17047225421776543, "grad_norm": 9.9795560836792, "learning_rate": 9.758368129235404e-05, "loss": 1.1614, "step": 2516 }, { "epoch": 0.17054000948573753, "grad_norm": 9.51271915435791, "learning_rate": 9.758231227325622e-05, "loss": 1.0423, "step": 2517 }, { "epoch": 0.1706077647537096, "grad_norm": 10.511359214782715, "learning_rate": 9.75809432541584e-05, "loss": 0.9492, "step": 2518 }, { "epoch": 0.17067552002168168, "grad_norm": 9.61755657196045, "learning_rate": 9.757957423506059e-05, "loss": 1.1481, "step": 2519 }, { "epoch": 0.17074327528965377, "grad_norm": 11.246142387390137, "learning_rate": 9.757820521596277e-05, "loss": 1.3981, "step": 2520 }, { "epoch": 0.17081103055762586, "grad_norm": 9.273181915283203, "learning_rate": 9.757683619686495e-05, "loss": 0.9773, "step": 2521 }, { "epoch": 0.17087878582559793, "grad_norm": 11.215822219848633, "learning_rate": 9.757546717776713e-05, "loss": 1.2303, "step": 2522 }, { "epoch": 0.17094654109357002, "grad_norm": 10.139853477478027, "learning_rate": 9.757409815866931e-05, "loss": 1.0807, "step": 2523 }, { "epoch": 0.1710142963615421, "grad_norm": 13.275606155395508, "learning_rate": 9.757272913957151e-05, "loss": 1.3326, "step": 2524 }, { "epoch": 0.1710820516295142, "grad_norm": 8.32109260559082, "learning_rate": 9.757136012047369e-05, "loss": 0.9428, "step": 2525 }, { "epoch": 0.17114980689748627, "grad_norm": 9.275816917419434, "learning_rate": 9.756999110137587e-05, "loss": 1.1531, "step": 2526 }, { "epoch": 0.17121756216545836, "grad_norm": 10.61928939819336, "learning_rate": 9.756862208227805e-05, "loss": 0.9422, "step": 2527 }, { "epoch": 0.17128531743343045, "grad_norm": 7.793631553649902, "learning_rate": 9.756725306318023e-05, "loss": 1.216, "step": 2528 }, { "epoch": 0.17135307270140254, "grad_norm": 7.453477382659912, "learning_rate": 9.756588404408242e-05, "loss": 0.7787, "step": 2529 }, { "epoch": 0.1714208279693746, "grad_norm": 10.39784049987793, "learning_rate": 9.75645150249846e-05, "loss": 1.3171, "step": 2530 }, { "epoch": 0.1714885832373467, "grad_norm": 8.57040786743164, "learning_rate": 9.756314600588678e-05, "loss": 0.8319, "step": 2531 }, { "epoch": 0.1715563385053188, "grad_norm": 8.696785926818848, "learning_rate": 9.756177698678896e-05, "loss": 1.0078, "step": 2532 }, { "epoch": 0.17162409377329088, "grad_norm": 8.212169647216797, "learning_rate": 9.756040796769116e-05, "loss": 0.7851, "step": 2533 }, { "epoch": 0.17169184904126294, "grad_norm": 10.94201946258545, "learning_rate": 9.755903894859334e-05, "loss": 1.0065, "step": 2534 }, { "epoch": 0.17175960430923504, "grad_norm": 12.041540145874023, "learning_rate": 9.755766992949552e-05, "loss": 1.2938, "step": 2535 }, { "epoch": 0.17182735957720713, "grad_norm": 9.289467811584473, "learning_rate": 9.75563009103977e-05, "loss": 1.1521, "step": 2536 }, { "epoch": 0.17189511484517922, "grad_norm": 9.231005668640137, "learning_rate": 9.755493189129988e-05, "loss": 0.9, "step": 2537 }, { "epoch": 0.17196287011315128, "grad_norm": 8.934699058532715, "learning_rate": 9.755356287220207e-05, "loss": 1.2817, "step": 2538 }, { "epoch": 0.17203062538112338, "grad_norm": 10.096917152404785, "learning_rate": 9.755219385310425e-05, "loss": 1.1587, "step": 2539 }, { "epoch": 0.17209838064909547, "grad_norm": 8.864645004272461, "learning_rate": 9.755082483400643e-05, "loss": 1.1781, "step": 2540 }, { "epoch": 0.17216613591706756, "grad_norm": 10.964715957641602, "learning_rate": 9.754945581490861e-05, "loss": 1.2729, "step": 2541 }, { "epoch": 0.17223389118503965, "grad_norm": 10.845703125, "learning_rate": 9.754808679581081e-05, "loss": 1.4475, "step": 2542 }, { "epoch": 0.17230164645301171, "grad_norm": 9.800530433654785, "learning_rate": 9.754671777671299e-05, "loss": 1.0296, "step": 2543 }, { "epoch": 0.1723694017209838, "grad_norm": 8.789834022521973, "learning_rate": 9.754534875761517e-05, "loss": 0.9573, "step": 2544 }, { "epoch": 0.1724371569889559, "grad_norm": 7.731616973876953, "learning_rate": 9.754397973851735e-05, "loss": 0.9875, "step": 2545 }, { "epoch": 0.172504912256928, "grad_norm": 8.942553520202637, "learning_rate": 9.754261071941953e-05, "loss": 1.1897, "step": 2546 }, { "epoch": 0.17257266752490005, "grad_norm": 7.468856334686279, "learning_rate": 9.754124170032172e-05, "loss": 0.8869, "step": 2547 }, { "epoch": 0.17264042279287214, "grad_norm": 8.798864364624023, "learning_rate": 9.75398726812239e-05, "loss": 1.0034, "step": 2548 }, { "epoch": 0.17270817806084424, "grad_norm": 8.579094886779785, "learning_rate": 9.753850366212608e-05, "loss": 1.0211, "step": 2549 }, { "epoch": 0.17277593332881633, "grad_norm": 10.90807819366455, "learning_rate": 9.753713464302827e-05, "loss": 1.3639, "step": 2550 }, { "epoch": 0.1728436885967884, "grad_norm": 8.337204933166504, "learning_rate": 9.753576562393046e-05, "loss": 1.323, "step": 2551 }, { "epoch": 0.17291144386476048, "grad_norm": 8.731851577758789, "learning_rate": 9.753439660483264e-05, "loss": 0.8805, "step": 2552 }, { "epoch": 0.17297919913273258, "grad_norm": 9.541427612304688, "learning_rate": 9.753302758573482e-05, "loss": 1.0368, "step": 2553 }, { "epoch": 0.17304695440070467, "grad_norm": 7.938154697418213, "learning_rate": 9.753165856663701e-05, "loss": 0.9128, "step": 2554 }, { "epoch": 0.17311470966867673, "grad_norm": 8.111212730407715, "learning_rate": 9.75302895475392e-05, "loss": 0.9857, "step": 2555 }, { "epoch": 0.17318246493664882, "grad_norm": 8.583324432373047, "learning_rate": 9.752892052844139e-05, "loss": 1.0975, "step": 2556 }, { "epoch": 0.1732502202046209, "grad_norm": 8.180643081665039, "learning_rate": 9.752755150934357e-05, "loss": 1.21, "step": 2557 }, { "epoch": 0.173317975472593, "grad_norm": 9.740217208862305, "learning_rate": 9.752618249024575e-05, "loss": 1.18, "step": 2558 }, { "epoch": 0.17338573074056507, "grad_norm": 8.725831031799316, "learning_rate": 9.752481347114793e-05, "loss": 0.9816, "step": 2559 }, { "epoch": 0.17345348600853716, "grad_norm": 10.786824226379395, "learning_rate": 9.752344445205011e-05, "loss": 1.0402, "step": 2560 }, { "epoch": 0.17352124127650925, "grad_norm": 8.91720962524414, "learning_rate": 9.75220754329523e-05, "loss": 1.2357, "step": 2561 }, { "epoch": 0.17358899654448134, "grad_norm": 9.02492618560791, "learning_rate": 9.752070641385448e-05, "loss": 1.1264, "step": 2562 }, { "epoch": 0.1736567518124534, "grad_norm": 7.917794227600098, "learning_rate": 9.751933739475666e-05, "loss": 1.0937, "step": 2563 }, { "epoch": 0.1737245070804255, "grad_norm": 11.543112754821777, "learning_rate": 9.751796837565884e-05, "loss": 1.0851, "step": 2564 }, { "epoch": 0.1737922623483976, "grad_norm": 9.114936828613281, "learning_rate": 9.751659935656104e-05, "loss": 0.9565, "step": 2565 }, { "epoch": 0.17386001761636968, "grad_norm": 12.322575569152832, "learning_rate": 9.751523033746322e-05, "loss": 1.0356, "step": 2566 }, { "epoch": 0.17392777288434175, "grad_norm": 11.39137077331543, "learning_rate": 9.75138613183654e-05, "loss": 1.2141, "step": 2567 }, { "epoch": 0.17399552815231384, "grad_norm": 9.505644798278809, "learning_rate": 9.751249229926758e-05, "loss": 0.8307, "step": 2568 }, { "epoch": 0.17406328342028593, "grad_norm": 11.05932331085205, "learning_rate": 9.751112328016976e-05, "loss": 1.1915, "step": 2569 }, { "epoch": 0.17413103868825802, "grad_norm": 11.104498863220215, "learning_rate": 9.750975426107195e-05, "loss": 1.0032, "step": 2570 }, { "epoch": 0.17419879395623009, "grad_norm": 11.366668701171875, "learning_rate": 9.750838524197413e-05, "loss": 1.0857, "step": 2571 }, { "epoch": 0.17426654922420218, "grad_norm": 8.775167465209961, "learning_rate": 9.750701622287631e-05, "loss": 1.1545, "step": 2572 }, { "epoch": 0.17433430449217427, "grad_norm": 8.820073127746582, "learning_rate": 9.75056472037785e-05, "loss": 1.0528, "step": 2573 }, { "epoch": 0.17440205976014636, "grad_norm": 9.53591537475586, "learning_rate": 9.750427818468069e-05, "loss": 1.0562, "step": 2574 }, { "epoch": 0.17446981502811842, "grad_norm": 10.08950138092041, "learning_rate": 9.750290916558287e-05, "loss": 1.0456, "step": 2575 }, { "epoch": 0.17453757029609052, "grad_norm": 8.222607612609863, "learning_rate": 9.750154014648505e-05, "loss": 0.9168, "step": 2576 }, { "epoch": 0.1746053255640626, "grad_norm": 10.067093849182129, "learning_rate": 9.750017112738723e-05, "loss": 0.9952, "step": 2577 }, { "epoch": 0.1746730808320347, "grad_norm": 9.726996421813965, "learning_rate": 9.749880210828941e-05, "loss": 0.9428, "step": 2578 }, { "epoch": 0.17474083610000676, "grad_norm": 9.970647811889648, "learning_rate": 9.74974330891916e-05, "loss": 1.1838, "step": 2579 }, { "epoch": 0.17480859136797886, "grad_norm": 7.8667893409729, "learning_rate": 9.749606407009378e-05, "loss": 1.0341, "step": 2580 }, { "epoch": 0.17487634663595095, "grad_norm": 9.227079391479492, "learning_rate": 9.749469505099596e-05, "loss": 1.0515, "step": 2581 }, { "epoch": 0.17494410190392304, "grad_norm": 9.198224067687988, "learning_rate": 9.749332603189814e-05, "loss": 1.0935, "step": 2582 }, { "epoch": 0.1750118571718951, "grad_norm": 11.105298042297363, "learning_rate": 9.749195701280032e-05, "loss": 1.2115, "step": 2583 }, { "epoch": 0.1750796124398672, "grad_norm": 8.815799713134766, "learning_rate": 9.749058799370252e-05, "loss": 1.1308, "step": 2584 }, { "epoch": 0.17514736770783929, "grad_norm": 10.571581840515137, "learning_rate": 9.74892189746047e-05, "loss": 1.1982, "step": 2585 }, { "epoch": 0.17521512297581138, "grad_norm": 7.368075370788574, "learning_rate": 9.748784995550688e-05, "loss": 0.9352, "step": 2586 }, { "epoch": 0.17528287824378344, "grad_norm": 7.103427410125732, "learning_rate": 9.748648093640906e-05, "loss": 1.0578, "step": 2587 }, { "epoch": 0.17535063351175553, "grad_norm": 7.678786754608154, "learning_rate": 9.748511191731125e-05, "loss": 0.8549, "step": 2588 }, { "epoch": 0.17541838877972762, "grad_norm": 9.514643669128418, "learning_rate": 9.748374289821343e-05, "loss": 1.1835, "step": 2589 }, { "epoch": 0.17548614404769972, "grad_norm": 7.552379131317139, "learning_rate": 9.748237387911561e-05, "loss": 0.9425, "step": 2590 }, { "epoch": 0.17555389931567178, "grad_norm": 7.663018226623535, "learning_rate": 9.74810048600178e-05, "loss": 1.0948, "step": 2591 }, { "epoch": 0.17562165458364387, "grad_norm": 10.293536186218262, "learning_rate": 9.747963584091997e-05, "loss": 1.1636, "step": 2592 }, { "epoch": 0.17568940985161596, "grad_norm": 9.024083137512207, "learning_rate": 9.747826682182217e-05, "loss": 1.1157, "step": 2593 }, { "epoch": 0.17575716511958805, "grad_norm": 8.870935440063477, "learning_rate": 9.747689780272435e-05, "loss": 0.9634, "step": 2594 }, { "epoch": 0.17582492038756015, "grad_norm": 8.553323745727539, "learning_rate": 9.747552878362653e-05, "loss": 1.1515, "step": 2595 }, { "epoch": 0.1758926756555322, "grad_norm": 9.13661003112793, "learning_rate": 9.747415976452871e-05, "loss": 0.9267, "step": 2596 }, { "epoch": 0.1759604309235043, "grad_norm": 8.66097640991211, "learning_rate": 9.74727907454309e-05, "loss": 0.6839, "step": 2597 }, { "epoch": 0.1760281861914764, "grad_norm": 10.07618236541748, "learning_rate": 9.747142172633308e-05, "loss": 1.2822, "step": 2598 }, { "epoch": 0.17609594145944849, "grad_norm": 9.136283874511719, "learning_rate": 9.747005270723526e-05, "loss": 1.0053, "step": 2599 }, { "epoch": 0.17616369672742055, "grad_norm": 9.42113208770752, "learning_rate": 9.746868368813746e-05, "loss": 1.15, "step": 2600 }, { "epoch": 0.17623145199539264, "grad_norm": 8.789713859558105, "learning_rate": 9.746731466903964e-05, "loss": 0.8504, "step": 2601 }, { "epoch": 0.17629920726336473, "grad_norm": 8.704665184020996, "learning_rate": 9.746594564994182e-05, "loss": 1.1516, "step": 2602 }, { "epoch": 0.17636696253133682, "grad_norm": 8.030630111694336, "learning_rate": 9.746457663084401e-05, "loss": 1.044, "step": 2603 }, { "epoch": 0.1764347177993089, "grad_norm": 8.640777587890625, "learning_rate": 9.746320761174619e-05, "loss": 1.0013, "step": 2604 }, { "epoch": 0.17650247306728098, "grad_norm": 7.806771278381348, "learning_rate": 9.746183859264837e-05, "loss": 1.0678, "step": 2605 }, { "epoch": 0.17657022833525307, "grad_norm": 8.883776664733887, "learning_rate": 9.746046957355055e-05, "loss": 0.9277, "step": 2606 }, { "epoch": 0.17663798360322516, "grad_norm": 7.539346694946289, "learning_rate": 9.745910055445275e-05, "loss": 1.0437, "step": 2607 }, { "epoch": 0.17670573887119723, "grad_norm": 9.873644828796387, "learning_rate": 9.745773153535493e-05, "loss": 1.206, "step": 2608 }, { "epoch": 0.17677349413916932, "grad_norm": 10.031026840209961, "learning_rate": 9.745636251625711e-05, "loss": 1.1934, "step": 2609 }, { "epoch": 0.1768412494071414, "grad_norm": 9.194849014282227, "learning_rate": 9.745499349715929e-05, "loss": 1.2365, "step": 2610 }, { "epoch": 0.1769090046751135, "grad_norm": 7.951476097106934, "learning_rate": 9.745362447806148e-05, "loss": 0.8829, "step": 2611 }, { "epoch": 0.17697675994308557, "grad_norm": 8.77597427368164, "learning_rate": 9.745225545896366e-05, "loss": 1.0097, "step": 2612 }, { "epoch": 0.17704451521105766, "grad_norm": 10.356508255004883, "learning_rate": 9.745088643986584e-05, "loss": 1.2603, "step": 2613 }, { "epoch": 0.17711227047902975, "grad_norm": 9.24199390411377, "learning_rate": 9.744951742076802e-05, "loss": 1.0181, "step": 2614 }, { "epoch": 0.17718002574700184, "grad_norm": 8.244451522827148, "learning_rate": 9.74481484016702e-05, "loss": 0.9602, "step": 2615 }, { "epoch": 0.1772477810149739, "grad_norm": 10.769842147827148, "learning_rate": 9.74467793825724e-05, "loss": 1.0646, "step": 2616 }, { "epoch": 0.177315536282946, "grad_norm": 9.381787300109863, "learning_rate": 9.744541036347458e-05, "loss": 0.8985, "step": 2617 }, { "epoch": 0.1773832915509181, "grad_norm": 8.379899978637695, "learning_rate": 9.744404134437676e-05, "loss": 1.0368, "step": 2618 }, { "epoch": 0.17745104681889018, "grad_norm": 9.59954833984375, "learning_rate": 9.744267232527894e-05, "loss": 1.1447, "step": 2619 }, { "epoch": 0.17751880208686224, "grad_norm": 8.83703327178955, "learning_rate": 9.744130330618113e-05, "loss": 1.3435, "step": 2620 }, { "epoch": 0.17758655735483433, "grad_norm": 9.448990821838379, "learning_rate": 9.743993428708331e-05, "loss": 1.1837, "step": 2621 }, { "epoch": 0.17765431262280643, "grad_norm": 9.466961860656738, "learning_rate": 9.74385652679855e-05, "loss": 1.1265, "step": 2622 }, { "epoch": 0.17772206789077852, "grad_norm": 8.68281364440918, "learning_rate": 9.743719624888767e-05, "loss": 0.8804, "step": 2623 }, { "epoch": 0.17778982315875058, "grad_norm": 7.007611274719238, "learning_rate": 9.743582722978985e-05, "loss": 1.03, "step": 2624 }, { "epoch": 0.17785757842672267, "grad_norm": 8.254279136657715, "learning_rate": 9.743445821069205e-05, "loss": 1.0328, "step": 2625 }, { "epoch": 0.17792533369469477, "grad_norm": 9.134196281433105, "learning_rate": 9.743308919159423e-05, "loss": 0.9671, "step": 2626 }, { "epoch": 0.17799308896266686, "grad_norm": 7.803997039794922, "learning_rate": 9.743172017249641e-05, "loss": 0.9692, "step": 2627 }, { "epoch": 0.17806084423063892, "grad_norm": 8.37303638458252, "learning_rate": 9.743035115339859e-05, "loss": 0.8949, "step": 2628 }, { "epoch": 0.178128599498611, "grad_norm": 9.928305625915527, "learning_rate": 9.742898213430078e-05, "loss": 1.2679, "step": 2629 }, { "epoch": 0.1781963547665831, "grad_norm": 8.58604621887207, "learning_rate": 9.742761311520296e-05, "loss": 1.23, "step": 2630 }, { "epoch": 0.1782641100345552, "grad_norm": 9.47903823852539, "learning_rate": 9.742624409610514e-05, "loss": 0.8701, "step": 2631 }, { "epoch": 0.17833186530252726, "grad_norm": 8.013197898864746, "learning_rate": 9.742487507700732e-05, "loss": 0.8325, "step": 2632 }, { "epoch": 0.17839962057049935, "grad_norm": 9.541396141052246, "learning_rate": 9.74235060579095e-05, "loss": 1.0996, "step": 2633 }, { "epoch": 0.17846737583847144, "grad_norm": 7.515398979187012, "learning_rate": 9.74221370388117e-05, "loss": 1.1174, "step": 2634 }, { "epoch": 0.17853513110644353, "grad_norm": 9.562670707702637, "learning_rate": 9.742076801971388e-05, "loss": 0.9727, "step": 2635 }, { "epoch": 0.1786028863744156, "grad_norm": 10.072449684143066, "learning_rate": 9.741939900061606e-05, "loss": 1.0021, "step": 2636 }, { "epoch": 0.1786706416423877, "grad_norm": 9.634955406188965, "learning_rate": 9.741802998151824e-05, "loss": 1.3998, "step": 2637 }, { "epoch": 0.17873839691035978, "grad_norm": 8.78954792022705, "learning_rate": 9.741666096242042e-05, "loss": 0.9602, "step": 2638 }, { "epoch": 0.17880615217833187, "grad_norm": 9.597916603088379, "learning_rate": 9.741529194332261e-05, "loss": 1.0174, "step": 2639 }, { "epoch": 0.17887390744630394, "grad_norm": 9.543583869934082, "learning_rate": 9.74139229242248e-05, "loss": 1.0686, "step": 2640 }, { "epoch": 0.17894166271427603, "grad_norm": 8.247551918029785, "learning_rate": 9.741255390512697e-05, "loss": 1.1026, "step": 2641 }, { "epoch": 0.17900941798224812, "grad_norm": 8.487943649291992, "learning_rate": 9.741118488602915e-05, "loss": 0.8843, "step": 2642 }, { "epoch": 0.1790771732502202, "grad_norm": 9.061832427978516, "learning_rate": 9.740981586693135e-05, "loss": 0.9708, "step": 2643 }, { "epoch": 0.17914492851819228, "grad_norm": 9.242405891418457, "learning_rate": 9.740844684783353e-05, "loss": 1.1344, "step": 2644 }, { "epoch": 0.17921268378616437, "grad_norm": 8.192344665527344, "learning_rate": 9.740707782873571e-05, "loss": 0.7587, "step": 2645 }, { "epoch": 0.17928043905413646, "grad_norm": 6.6164445877075195, "learning_rate": 9.74057088096379e-05, "loss": 0.842, "step": 2646 }, { "epoch": 0.17934819432210855, "grad_norm": 10.285326957702637, "learning_rate": 9.740433979054008e-05, "loss": 1.138, "step": 2647 }, { "epoch": 0.17941594959008064, "grad_norm": 8.55659294128418, "learning_rate": 9.740297077144226e-05, "loss": 0.7925, "step": 2648 }, { "epoch": 0.1794837048580527, "grad_norm": 8.384642601013184, "learning_rate": 9.740160175234446e-05, "loss": 0.9719, "step": 2649 }, { "epoch": 0.1795514601260248, "grad_norm": 9.303678512573242, "learning_rate": 9.740023273324664e-05, "loss": 1.0688, "step": 2650 }, { "epoch": 0.1796192153939969, "grad_norm": 9.642799377441406, "learning_rate": 9.739886371414882e-05, "loss": 0.9758, "step": 2651 }, { "epoch": 0.17968697066196898, "grad_norm": 9.616509437561035, "learning_rate": 9.739749469505101e-05, "loss": 0.9999, "step": 2652 }, { "epoch": 0.17975472592994104, "grad_norm": 10.511337280273438, "learning_rate": 9.739612567595319e-05, "loss": 1.0079, "step": 2653 }, { "epoch": 0.17982248119791314, "grad_norm": 7.480882167816162, "learning_rate": 9.739475665685537e-05, "loss": 0.8353, "step": 2654 }, { "epoch": 0.17989023646588523, "grad_norm": 10.277608871459961, "learning_rate": 9.739338763775755e-05, "loss": 0.9709, "step": 2655 }, { "epoch": 0.17995799173385732, "grad_norm": 9.135882377624512, "learning_rate": 9.739201861865973e-05, "loss": 1.3888, "step": 2656 }, { "epoch": 0.18002574700182938, "grad_norm": 8.147950172424316, "learning_rate": 9.739064959956193e-05, "loss": 1.2235, "step": 2657 }, { "epoch": 0.18009350226980148, "grad_norm": 8.583501815795898, "learning_rate": 9.738928058046411e-05, "loss": 0.9123, "step": 2658 }, { "epoch": 0.18016125753777357, "grad_norm": 8.202300071716309, "learning_rate": 9.738791156136629e-05, "loss": 1.0435, "step": 2659 }, { "epoch": 0.18022901280574566, "grad_norm": 8.121417045593262, "learning_rate": 9.738654254226847e-05, "loss": 0.9351, "step": 2660 }, { "epoch": 0.18029676807371772, "grad_norm": 10.645029067993164, "learning_rate": 9.738517352317065e-05, "loss": 1.5219, "step": 2661 }, { "epoch": 0.18036452334168981, "grad_norm": 9.766422271728516, "learning_rate": 9.738380450407284e-05, "loss": 1.0499, "step": 2662 }, { "epoch": 0.1804322786096619, "grad_norm": 8.137174606323242, "learning_rate": 9.738243548497502e-05, "loss": 1.1871, "step": 2663 }, { "epoch": 0.180500033877634, "grad_norm": 8.422198295593262, "learning_rate": 9.73810664658772e-05, "loss": 0.9524, "step": 2664 }, { "epoch": 0.18056778914560606, "grad_norm": 6.721381664276123, "learning_rate": 9.737969744677938e-05, "loss": 1.0449, "step": 2665 }, { "epoch": 0.18063554441357815, "grad_norm": 9.175302505493164, "learning_rate": 9.737832842768158e-05, "loss": 1.0508, "step": 2666 }, { "epoch": 0.18070329968155024, "grad_norm": 10.165428161621094, "learning_rate": 9.737695940858376e-05, "loss": 0.8744, "step": 2667 }, { "epoch": 0.18077105494952234, "grad_norm": 9.396173477172852, "learning_rate": 9.737559038948594e-05, "loss": 1.0304, "step": 2668 }, { "epoch": 0.1808388102174944, "grad_norm": 9.680516242980957, "learning_rate": 9.737422137038812e-05, "loss": 0.9773, "step": 2669 }, { "epoch": 0.1809065654854665, "grad_norm": 7.570343017578125, "learning_rate": 9.73728523512903e-05, "loss": 0.9159, "step": 2670 }, { "epoch": 0.18097432075343858, "grad_norm": 9.5789213180542, "learning_rate": 9.737148333219249e-05, "loss": 1.1269, "step": 2671 }, { "epoch": 0.18104207602141068, "grad_norm": 8.559327125549316, "learning_rate": 9.737011431309467e-05, "loss": 1.2318, "step": 2672 }, { "epoch": 0.18110983128938274, "grad_norm": 7.630974292755127, "learning_rate": 9.736874529399685e-05, "loss": 0.8996, "step": 2673 }, { "epoch": 0.18117758655735483, "grad_norm": 8.078895568847656, "learning_rate": 9.736737627489903e-05, "loss": 0.9942, "step": 2674 }, { "epoch": 0.18124534182532692, "grad_norm": 8.830656051635742, "learning_rate": 9.736600725580123e-05, "loss": 1.0614, "step": 2675 }, { "epoch": 0.181313097093299, "grad_norm": 7.66297721862793, "learning_rate": 9.736463823670341e-05, "loss": 1.025, "step": 2676 }, { "epoch": 0.18138085236127108, "grad_norm": 8.318365097045898, "learning_rate": 9.736326921760559e-05, "loss": 1.1286, "step": 2677 }, { "epoch": 0.18144860762924317, "grad_norm": 10.180238723754883, "learning_rate": 9.736190019850777e-05, "loss": 1.1304, "step": 2678 }, { "epoch": 0.18151636289721526, "grad_norm": 9.2420654296875, "learning_rate": 9.736053117940995e-05, "loss": 1.157, "step": 2679 }, { "epoch": 0.18158411816518735, "grad_norm": 7.984904766082764, "learning_rate": 9.735916216031214e-05, "loss": 1.1325, "step": 2680 }, { "epoch": 0.18165187343315942, "grad_norm": 11.136275291442871, "learning_rate": 9.735779314121432e-05, "loss": 1.0151, "step": 2681 }, { "epoch": 0.1817196287011315, "grad_norm": 9.954483032226562, "learning_rate": 9.73564241221165e-05, "loss": 1.0934, "step": 2682 }, { "epoch": 0.1817873839691036, "grad_norm": 8.491388320922852, "learning_rate": 9.735505510301868e-05, "loss": 1.2288, "step": 2683 }, { "epoch": 0.1818551392370757, "grad_norm": 9.355586051940918, "learning_rate": 9.735368608392088e-05, "loss": 0.7999, "step": 2684 }, { "epoch": 0.18192289450504776, "grad_norm": 8.829736709594727, "learning_rate": 9.735231706482306e-05, "loss": 0.9711, "step": 2685 }, { "epoch": 0.18199064977301985, "grad_norm": 9.667959213256836, "learning_rate": 9.735094804572524e-05, "loss": 1.1132, "step": 2686 }, { "epoch": 0.18205840504099194, "grad_norm": 6.971467971801758, "learning_rate": 9.734957902662742e-05, "loss": 1.0252, "step": 2687 }, { "epoch": 0.18212616030896403, "grad_norm": 9.69013500213623, "learning_rate": 9.73482100075296e-05, "loss": 0.9562, "step": 2688 }, { "epoch": 0.1821939155769361, "grad_norm": 9.47673511505127, "learning_rate": 9.734684098843179e-05, "loss": 0.9822, "step": 2689 }, { "epoch": 0.18226167084490819, "grad_norm": 9.708051681518555, "learning_rate": 9.734547196933397e-05, "loss": 1.0772, "step": 2690 }, { "epoch": 0.18232942611288028, "grad_norm": 10.617173194885254, "learning_rate": 9.734410295023615e-05, "loss": 1.2686, "step": 2691 }, { "epoch": 0.18239718138085237, "grad_norm": 9.52670955657959, "learning_rate": 9.734273393113835e-05, "loss": 0.9477, "step": 2692 }, { "epoch": 0.18246493664882443, "grad_norm": 10.090290069580078, "learning_rate": 9.734136491204053e-05, "loss": 1.1766, "step": 2693 }, { "epoch": 0.18253269191679652, "grad_norm": 8.709790229797363, "learning_rate": 9.733999589294271e-05, "loss": 1.1327, "step": 2694 }, { "epoch": 0.18260044718476862, "grad_norm": 12.07381534576416, "learning_rate": 9.73386268738449e-05, "loss": 1.3694, "step": 2695 }, { "epoch": 0.1826682024527407, "grad_norm": 8.329826354980469, "learning_rate": 9.733725785474708e-05, "loss": 1.0448, "step": 2696 }, { "epoch": 0.18273595772071277, "grad_norm": 7.551383972167969, "learning_rate": 9.733588883564926e-05, "loss": 1.029, "step": 2697 }, { "epoch": 0.18280371298868486, "grad_norm": 9.393105506896973, "learning_rate": 9.733451981655146e-05, "loss": 0.9729, "step": 2698 }, { "epoch": 0.18287146825665695, "grad_norm": 9.283944129943848, "learning_rate": 9.733315079745364e-05, "loss": 1.2459, "step": 2699 }, { "epoch": 0.18293922352462905, "grad_norm": 9.02878189086914, "learning_rate": 9.733178177835582e-05, "loss": 1.3235, "step": 2700 }, { "epoch": 0.18300697879260114, "grad_norm": 8.735793113708496, "learning_rate": 9.7330412759258e-05, "loss": 1.1447, "step": 2701 }, { "epoch": 0.1830747340605732, "grad_norm": 8.511741638183594, "learning_rate": 9.732904374016018e-05, "loss": 1.0148, "step": 2702 }, { "epoch": 0.1831424893285453, "grad_norm": 6.590076446533203, "learning_rate": 9.732767472106237e-05, "loss": 1.0904, "step": 2703 }, { "epoch": 0.18321024459651739, "grad_norm": 11.523942947387695, "learning_rate": 9.732630570196455e-05, "loss": 0.9349, "step": 2704 }, { "epoch": 0.18327799986448948, "grad_norm": 6.972995281219482, "learning_rate": 9.732493668286673e-05, "loss": 0.9837, "step": 2705 }, { "epoch": 0.18334575513246154, "grad_norm": 7.834453105926514, "learning_rate": 9.732356766376891e-05, "loss": 1.0374, "step": 2706 }, { "epoch": 0.18341351040043363, "grad_norm": 9.44674015045166, "learning_rate": 9.732219864467111e-05, "loss": 1.4479, "step": 2707 }, { "epoch": 0.18348126566840572, "grad_norm": 9.017583847045898, "learning_rate": 9.732082962557329e-05, "loss": 1.136, "step": 2708 }, { "epoch": 0.18354902093637782, "grad_norm": 8.40336799621582, "learning_rate": 9.731946060647547e-05, "loss": 0.9914, "step": 2709 }, { "epoch": 0.18361677620434988, "grad_norm": 10.586843490600586, "learning_rate": 9.731809158737765e-05, "loss": 1.1515, "step": 2710 }, { "epoch": 0.18368453147232197, "grad_norm": 7.611972808837891, "learning_rate": 9.731672256827983e-05, "loss": 1.0821, "step": 2711 }, { "epoch": 0.18375228674029406, "grad_norm": 9.093097686767578, "learning_rate": 9.731535354918202e-05, "loss": 1.1295, "step": 2712 }, { "epoch": 0.18382004200826615, "grad_norm": 8.013279914855957, "learning_rate": 9.73139845300842e-05, "loss": 0.839, "step": 2713 }, { "epoch": 0.18388779727623822, "grad_norm": 7.941558837890625, "learning_rate": 9.731261551098638e-05, "loss": 1.0819, "step": 2714 }, { "epoch": 0.1839555525442103, "grad_norm": 7.09537935256958, "learning_rate": 9.731124649188856e-05, "loss": 0.8606, "step": 2715 }, { "epoch": 0.1840233078121824, "grad_norm": 7.867136478424072, "learning_rate": 9.730987747279074e-05, "loss": 0.9942, "step": 2716 }, { "epoch": 0.1840910630801545, "grad_norm": 8.66762638092041, "learning_rate": 9.730850845369294e-05, "loss": 1.0792, "step": 2717 }, { "epoch": 0.18415881834812656, "grad_norm": 7.97219705581665, "learning_rate": 9.730713943459512e-05, "loss": 1.0366, "step": 2718 }, { "epoch": 0.18422657361609865, "grad_norm": 8.341264724731445, "learning_rate": 9.73057704154973e-05, "loss": 1.0833, "step": 2719 }, { "epoch": 0.18429432888407074, "grad_norm": 8.704034805297852, "learning_rate": 9.730440139639948e-05, "loss": 1.2909, "step": 2720 }, { "epoch": 0.18436208415204283, "grad_norm": 9.759525299072266, "learning_rate": 9.730303237730167e-05, "loss": 1.1749, "step": 2721 }, { "epoch": 0.1844298394200149, "grad_norm": 9.2392578125, "learning_rate": 9.730166335820385e-05, "loss": 1.4037, "step": 2722 }, { "epoch": 0.184497594687987, "grad_norm": 10.519861221313477, "learning_rate": 9.730029433910603e-05, "loss": 1.2692, "step": 2723 }, { "epoch": 0.18456534995595908, "grad_norm": 8.285715103149414, "learning_rate": 9.729892532000821e-05, "loss": 0.8014, "step": 2724 }, { "epoch": 0.18463310522393117, "grad_norm": 7.261229991912842, "learning_rate": 9.72975563009104e-05, "loss": 0.9302, "step": 2725 }, { "epoch": 0.18470086049190323, "grad_norm": 10.149426460266113, "learning_rate": 9.729618728181259e-05, "loss": 1.2855, "step": 2726 }, { "epoch": 0.18476861575987533, "grad_norm": 7.495121955871582, "learning_rate": 9.729481826271477e-05, "loss": 0.9771, "step": 2727 }, { "epoch": 0.18483637102784742, "grad_norm": 9.102607727050781, "learning_rate": 9.729344924361695e-05, "loss": 1.1974, "step": 2728 }, { "epoch": 0.1849041262958195, "grad_norm": 9.57135009765625, "learning_rate": 9.729208022451913e-05, "loss": 0.9758, "step": 2729 }, { "epoch": 0.18497188156379157, "grad_norm": 8.745903015136719, "learning_rate": 9.729071120542132e-05, "loss": 1.1719, "step": 2730 }, { "epoch": 0.18503963683176367, "grad_norm": 10.697607040405273, "learning_rate": 9.72893421863235e-05, "loss": 0.952, "step": 2731 }, { "epoch": 0.18510739209973576, "grad_norm": 9.907280921936035, "learning_rate": 9.728797316722568e-05, "loss": 0.9571, "step": 2732 }, { "epoch": 0.18517514736770785, "grad_norm": 7.285250186920166, "learning_rate": 9.728660414812786e-05, "loss": 0.9002, "step": 2733 }, { "epoch": 0.1852429026356799, "grad_norm": 8.134112358093262, "learning_rate": 9.728523512903004e-05, "loss": 0.8645, "step": 2734 }, { "epoch": 0.185310657903652, "grad_norm": 9.427742004394531, "learning_rate": 9.728386610993224e-05, "loss": 1.2333, "step": 2735 }, { "epoch": 0.1853784131716241, "grad_norm": 8.804586410522461, "learning_rate": 9.728249709083442e-05, "loss": 1.2601, "step": 2736 }, { "epoch": 0.1854461684395962, "grad_norm": 9.144674301147461, "learning_rate": 9.72811280717366e-05, "loss": 0.9503, "step": 2737 }, { "epoch": 0.18551392370756825, "grad_norm": 9.399337768554688, "learning_rate": 9.727975905263879e-05, "loss": 0.8586, "step": 2738 }, { "epoch": 0.18558167897554034, "grad_norm": 7.7377119064331055, "learning_rate": 9.727839003354097e-05, "loss": 0.7511, "step": 2739 }, { "epoch": 0.18564943424351243, "grad_norm": 9.146937370300293, "learning_rate": 9.727702101444315e-05, "loss": 1.0037, "step": 2740 }, { "epoch": 0.18571718951148453, "grad_norm": 7.722357273101807, "learning_rate": 9.727565199534535e-05, "loss": 1.077, "step": 2741 }, { "epoch": 0.1857849447794566, "grad_norm": 7.808566093444824, "learning_rate": 9.727428297624753e-05, "loss": 0.8953, "step": 2742 }, { "epoch": 0.18585270004742868, "grad_norm": 8.765763282775879, "learning_rate": 9.727291395714971e-05, "loss": 1.0691, "step": 2743 }, { "epoch": 0.18592045531540077, "grad_norm": 11.350706100463867, "learning_rate": 9.72715449380519e-05, "loss": 0.9308, "step": 2744 }, { "epoch": 0.18598821058337286, "grad_norm": 7.838237285614014, "learning_rate": 9.727017591895408e-05, "loss": 0.9452, "step": 2745 }, { "epoch": 0.18605596585134493, "grad_norm": 9.660740852355957, "learning_rate": 9.726880689985626e-05, "loss": 0.8705, "step": 2746 }, { "epoch": 0.18612372111931702, "grad_norm": 8.146308898925781, "learning_rate": 9.726743788075844e-05, "loss": 1.0362, "step": 2747 }, { "epoch": 0.1861914763872891, "grad_norm": 9.901165008544922, "learning_rate": 9.726606886166062e-05, "loss": 1.0842, "step": 2748 }, { "epoch": 0.1862592316552612, "grad_norm": 7.650402069091797, "learning_rate": 9.726469984256282e-05, "loss": 0.997, "step": 2749 }, { "epoch": 0.18632698692323327, "grad_norm": 7.760092735290527, "learning_rate": 9.7263330823465e-05, "loss": 0.9903, "step": 2750 }, { "epoch": 0.18639474219120536, "grad_norm": 9.523726463317871, "learning_rate": 9.726196180436718e-05, "loss": 0.9276, "step": 2751 }, { "epoch": 0.18646249745917745, "grad_norm": 7.759490489959717, "learning_rate": 9.726059278526936e-05, "loss": 0.9356, "step": 2752 }, { "epoch": 0.18653025272714954, "grad_norm": 10.949979782104492, "learning_rate": 9.725922376617155e-05, "loss": 1.1451, "step": 2753 }, { "epoch": 0.18659800799512163, "grad_norm": 8.312686920166016, "learning_rate": 9.725785474707373e-05, "loss": 0.973, "step": 2754 }, { "epoch": 0.1866657632630937, "grad_norm": 6.999983787536621, "learning_rate": 9.725648572797591e-05, "loss": 1.0699, "step": 2755 }, { "epoch": 0.1867335185310658, "grad_norm": 9.422745704650879, "learning_rate": 9.725511670887809e-05, "loss": 1.0697, "step": 2756 }, { "epoch": 0.18680127379903788, "grad_norm": 8.054603576660156, "learning_rate": 9.725374768978027e-05, "loss": 0.8764, "step": 2757 }, { "epoch": 0.18686902906700997, "grad_norm": 7.902176856994629, "learning_rate": 9.725237867068247e-05, "loss": 0.9893, "step": 2758 }, { "epoch": 0.18693678433498204, "grad_norm": 8.409537315368652, "learning_rate": 9.725100965158465e-05, "loss": 1.0526, "step": 2759 }, { "epoch": 0.18700453960295413, "grad_norm": 9.867463111877441, "learning_rate": 9.724964063248683e-05, "loss": 1.1218, "step": 2760 }, { "epoch": 0.18707229487092622, "grad_norm": 6.852199554443359, "learning_rate": 9.724827161338901e-05, "loss": 0.9838, "step": 2761 }, { "epoch": 0.1871400501388983, "grad_norm": 9.132448196411133, "learning_rate": 9.72469025942912e-05, "loss": 1.038, "step": 2762 }, { "epoch": 0.18720780540687038, "grad_norm": 8.699213981628418, "learning_rate": 9.724553357519338e-05, "loss": 1.0532, "step": 2763 }, { "epoch": 0.18727556067484247, "grad_norm": 8.520672798156738, "learning_rate": 9.724416455609556e-05, "loss": 1.2942, "step": 2764 }, { "epoch": 0.18734331594281456, "grad_norm": 9.690433502197266, "learning_rate": 9.724279553699774e-05, "loss": 0.9552, "step": 2765 }, { "epoch": 0.18741107121078665, "grad_norm": 7.540090560913086, "learning_rate": 9.724142651789992e-05, "loss": 1.0718, "step": 2766 }, { "epoch": 0.18747882647875871, "grad_norm": 9.072039604187012, "learning_rate": 9.724005749880212e-05, "loss": 1.0313, "step": 2767 }, { "epoch": 0.1875465817467308, "grad_norm": 10.155011177062988, "learning_rate": 9.72386884797043e-05, "loss": 1.3498, "step": 2768 }, { "epoch": 0.1876143370147029, "grad_norm": 7.816718101501465, "learning_rate": 9.723731946060648e-05, "loss": 0.9594, "step": 2769 }, { "epoch": 0.187682092282675, "grad_norm": 10.980911254882812, "learning_rate": 9.723595044150866e-05, "loss": 1.1675, "step": 2770 }, { "epoch": 0.18774984755064705, "grad_norm": 8.865739822387695, "learning_rate": 9.723458142241084e-05, "loss": 1.1796, "step": 2771 }, { "epoch": 0.18781760281861914, "grad_norm": 8.97850227355957, "learning_rate": 9.723321240331303e-05, "loss": 1.0907, "step": 2772 }, { "epoch": 0.18788535808659124, "grad_norm": 8.16921329498291, "learning_rate": 9.723184338421521e-05, "loss": 0.95, "step": 2773 }, { "epoch": 0.18795311335456333, "grad_norm": 8.766203880310059, "learning_rate": 9.723047436511739e-05, "loss": 0.9087, "step": 2774 }, { "epoch": 0.1880208686225354, "grad_norm": 7.410607814788818, "learning_rate": 9.722910534601957e-05, "loss": 0.8082, "step": 2775 }, { "epoch": 0.18808862389050748, "grad_norm": 9.640182495117188, "learning_rate": 9.722773632692177e-05, "loss": 0.99, "step": 2776 }, { "epoch": 0.18815637915847958, "grad_norm": 9.038297653198242, "learning_rate": 9.722636730782395e-05, "loss": 1.0017, "step": 2777 }, { "epoch": 0.18822413442645167, "grad_norm": 10.167421340942383, "learning_rate": 9.722499828872613e-05, "loss": 1.1588, "step": 2778 }, { "epoch": 0.18829188969442373, "grad_norm": 9.911538124084473, "learning_rate": 9.722362926962831e-05, "loss": 1.052, "step": 2779 }, { "epoch": 0.18835964496239582, "grad_norm": 8.78661060333252, "learning_rate": 9.722226025053049e-05, "loss": 0.9827, "step": 2780 }, { "epoch": 0.18842740023036791, "grad_norm": 8.58356761932373, "learning_rate": 9.722089123143268e-05, "loss": 0.9663, "step": 2781 }, { "epoch": 0.18849515549834, "grad_norm": 7.882653713226318, "learning_rate": 9.721952221233486e-05, "loss": 1.0282, "step": 2782 }, { "epoch": 0.18856291076631207, "grad_norm": 8.180728912353516, "learning_rate": 9.721815319323704e-05, "loss": 0.8983, "step": 2783 }, { "epoch": 0.18863066603428416, "grad_norm": 10.785475730895996, "learning_rate": 9.721678417413924e-05, "loss": 1.0949, "step": 2784 }, { "epoch": 0.18869842130225625, "grad_norm": 8.2493257522583, "learning_rate": 9.721541515504142e-05, "loss": 0.8891, "step": 2785 }, { "epoch": 0.18876617657022834, "grad_norm": 8.298515319824219, "learning_rate": 9.72140461359436e-05, "loss": 0.9384, "step": 2786 }, { "epoch": 0.1888339318382004, "grad_norm": 11.643486022949219, "learning_rate": 9.721267711684579e-05, "loss": 1.0915, "step": 2787 }, { "epoch": 0.1889016871061725, "grad_norm": 10.728472709655762, "learning_rate": 9.721130809774797e-05, "loss": 1.562, "step": 2788 }, { "epoch": 0.1889694423741446, "grad_norm": 7.9253435134887695, "learning_rate": 9.720993907865015e-05, "loss": 0.876, "step": 2789 }, { "epoch": 0.18903719764211668, "grad_norm": 9.075439453125, "learning_rate": 9.720857005955235e-05, "loss": 1.1712, "step": 2790 }, { "epoch": 0.18910495291008875, "grad_norm": 8.572853088378906, "learning_rate": 9.720720104045453e-05, "loss": 0.9697, "step": 2791 }, { "epoch": 0.18917270817806084, "grad_norm": 8.40988826751709, "learning_rate": 9.720583202135671e-05, "loss": 1.0904, "step": 2792 }, { "epoch": 0.18924046344603293, "grad_norm": 8.015021324157715, "learning_rate": 9.720446300225889e-05, "loss": 1.0264, "step": 2793 }, { "epoch": 0.18930821871400502, "grad_norm": 10.295256614685059, "learning_rate": 9.720309398316107e-05, "loss": 0.9608, "step": 2794 }, { "epoch": 0.18937597398197709, "grad_norm": 9.377728462219238, "learning_rate": 9.720172496406326e-05, "loss": 1.2169, "step": 2795 }, { "epoch": 0.18944372924994918, "grad_norm": 8.55238151550293, "learning_rate": 9.720035594496544e-05, "loss": 1.0951, "step": 2796 }, { "epoch": 0.18951148451792127, "grad_norm": 10.069438934326172, "learning_rate": 9.719898692586762e-05, "loss": 1.3206, "step": 2797 }, { "epoch": 0.18957923978589336, "grad_norm": 9.229057312011719, "learning_rate": 9.71976179067698e-05, "loss": 1.18, "step": 2798 }, { "epoch": 0.18964699505386542, "grad_norm": 8.911051750183105, "learning_rate": 9.7196248887672e-05, "loss": 1.1351, "step": 2799 }, { "epoch": 0.18971475032183752, "grad_norm": 9.285752296447754, "learning_rate": 9.719487986857418e-05, "loss": 1.0626, "step": 2800 }, { "epoch": 0.1897825055898096, "grad_norm": 8.793561935424805, "learning_rate": 9.719351084947636e-05, "loss": 1.033, "step": 2801 }, { "epoch": 0.1898502608577817, "grad_norm": 10.635384559631348, "learning_rate": 9.719214183037854e-05, "loss": 1.5004, "step": 2802 }, { "epoch": 0.18991801612575376, "grad_norm": 7.9627275466918945, "learning_rate": 9.719077281128072e-05, "loss": 1.0246, "step": 2803 }, { "epoch": 0.18998577139372586, "grad_norm": 8.960352897644043, "learning_rate": 9.718940379218291e-05, "loss": 1.0316, "step": 2804 }, { "epoch": 0.19005352666169795, "grad_norm": 9.522171020507812, "learning_rate": 9.718803477308509e-05, "loss": 1.0219, "step": 2805 }, { "epoch": 0.19012128192967004, "grad_norm": 8.412702560424805, "learning_rate": 9.718666575398727e-05, "loss": 1.0609, "step": 2806 }, { "epoch": 0.19018903719764213, "grad_norm": 8.777839660644531, "learning_rate": 9.718529673488945e-05, "loss": 1.0854, "step": 2807 }, { "epoch": 0.1902567924656142, "grad_norm": 8.932796478271484, "learning_rate": 9.718392771579165e-05, "loss": 1.1661, "step": 2808 }, { "epoch": 0.19032454773358629, "grad_norm": 8.514800071716309, "learning_rate": 9.718255869669383e-05, "loss": 1.0584, "step": 2809 }, { "epoch": 0.19039230300155838, "grad_norm": 7.001948356628418, "learning_rate": 9.718118967759601e-05, "loss": 0.8479, "step": 2810 }, { "epoch": 0.19046005826953047, "grad_norm": 8.977307319641113, "learning_rate": 9.717982065849819e-05, "loss": 1.2556, "step": 2811 }, { "epoch": 0.19052781353750253, "grad_norm": 8.513920783996582, "learning_rate": 9.717845163940037e-05, "loss": 1.2076, "step": 2812 }, { "epoch": 0.19059556880547462, "grad_norm": 8.556622505187988, "learning_rate": 9.717708262030256e-05, "loss": 1.0491, "step": 2813 }, { "epoch": 0.19066332407344672, "grad_norm": 9.8518648147583, "learning_rate": 9.717571360120474e-05, "loss": 1.1179, "step": 2814 }, { "epoch": 0.1907310793414188, "grad_norm": 8.887413024902344, "learning_rate": 9.717434458210692e-05, "loss": 1.0675, "step": 2815 }, { "epoch": 0.19079883460939087, "grad_norm": 7.888981342315674, "learning_rate": 9.71729755630091e-05, "loss": 1.0489, "step": 2816 }, { "epoch": 0.19086658987736296, "grad_norm": 7.692848205566406, "learning_rate": 9.71716065439113e-05, "loss": 1.0602, "step": 2817 }, { "epoch": 0.19093434514533505, "grad_norm": 8.591787338256836, "learning_rate": 9.717023752481348e-05, "loss": 0.8737, "step": 2818 }, { "epoch": 0.19100210041330715, "grad_norm": 9.622870445251465, "learning_rate": 9.716886850571566e-05, "loss": 0.9717, "step": 2819 }, { "epoch": 0.1910698556812792, "grad_norm": 7.949582576751709, "learning_rate": 9.716749948661784e-05, "loss": 0.8651, "step": 2820 }, { "epoch": 0.1911376109492513, "grad_norm": 9.46272087097168, "learning_rate": 9.716613046752002e-05, "loss": 0.9975, "step": 2821 }, { "epoch": 0.1912053662172234, "grad_norm": 8.86549186706543, "learning_rate": 9.716476144842221e-05, "loss": 1.0805, "step": 2822 }, { "epoch": 0.19127312148519549, "grad_norm": 8.039673805236816, "learning_rate": 9.716339242932439e-05, "loss": 1.121, "step": 2823 }, { "epoch": 0.19134087675316755, "grad_norm": 8.606979370117188, "learning_rate": 9.716202341022657e-05, "loss": 0.9217, "step": 2824 }, { "epoch": 0.19140863202113964, "grad_norm": 10.429420471191406, "learning_rate": 9.716065439112875e-05, "loss": 1.2266, "step": 2825 }, { "epoch": 0.19147638728911173, "grad_norm": 7.339411735534668, "learning_rate": 9.715928537203093e-05, "loss": 1.0235, "step": 2826 }, { "epoch": 0.19154414255708382, "grad_norm": 8.247300148010254, "learning_rate": 9.715791635293313e-05, "loss": 0.8993, "step": 2827 }, { "epoch": 0.1916118978250559, "grad_norm": 8.475278854370117, "learning_rate": 9.715654733383531e-05, "loss": 0.9611, "step": 2828 }, { "epoch": 0.19167965309302798, "grad_norm": 9.792519569396973, "learning_rate": 9.715517831473749e-05, "loss": 1.0831, "step": 2829 }, { "epoch": 0.19174740836100007, "grad_norm": 6.757070541381836, "learning_rate": 9.715380929563967e-05, "loss": 1.0282, "step": 2830 }, { "epoch": 0.19181516362897216, "grad_norm": 6.707785129547119, "learning_rate": 9.715244027654186e-05, "loss": 0.7813, "step": 2831 }, { "epoch": 0.19188291889694423, "grad_norm": 9.990489959716797, "learning_rate": 9.715107125744404e-05, "loss": 1.0504, "step": 2832 }, { "epoch": 0.19195067416491632, "grad_norm": 10.57358169555664, "learning_rate": 9.714970223834622e-05, "loss": 0.9355, "step": 2833 }, { "epoch": 0.1920184294328884, "grad_norm": 9.225300788879395, "learning_rate": 9.714833321924842e-05, "loss": 1.2404, "step": 2834 }, { "epoch": 0.1920861847008605, "grad_norm": 8.36042594909668, "learning_rate": 9.71469642001506e-05, "loss": 0.9012, "step": 2835 }, { "epoch": 0.19215393996883257, "grad_norm": 9.624984741210938, "learning_rate": 9.714559518105278e-05, "loss": 1.0509, "step": 2836 }, { "epoch": 0.19222169523680466, "grad_norm": 12.569930076599121, "learning_rate": 9.714422616195497e-05, "loss": 1.2663, "step": 2837 }, { "epoch": 0.19228945050477675, "grad_norm": 12.278885841369629, "learning_rate": 9.714285714285715e-05, "loss": 1.1547, "step": 2838 }, { "epoch": 0.19235720577274884, "grad_norm": 9.3023681640625, "learning_rate": 9.714148812375933e-05, "loss": 1.0873, "step": 2839 }, { "epoch": 0.1924249610407209, "grad_norm": 8.241714477539062, "learning_rate": 9.714011910466153e-05, "loss": 0.8222, "step": 2840 }, { "epoch": 0.192492716308693, "grad_norm": 10.45174503326416, "learning_rate": 9.71387500855637e-05, "loss": 1.1974, "step": 2841 }, { "epoch": 0.1925604715766651, "grad_norm": 8.096826553344727, "learning_rate": 9.713738106646589e-05, "loss": 1.1147, "step": 2842 }, { "epoch": 0.19262822684463718, "grad_norm": 9.053191184997559, "learning_rate": 9.713601204736807e-05, "loss": 0.9835, "step": 2843 }, { "epoch": 0.19269598211260924, "grad_norm": 8.050823211669922, "learning_rate": 9.713464302827025e-05, "loss": 1.0826, "step": 2844 }, { "epoch": 0.19276373738058133, "grad_norm": 10.000917434692383, "learning_rate": 9.713327400917244e-05, "loss": 1.1723, "step": 2845 }, { "epoch": 0.19283149264855343, "grad_norm": 8.992589950561523, "learning_rate": 9.713190499007462e-05, "loss": 1.0203, "step": 2846 }, { "epoch": 0.19289924791652552, "grad_norm": 10.340285301208496, "learning_rate": 9.71305359709768e-05, "loss": 1.2523, "step": 2847 }, { "epoch": 0.19296700318449758, "grad_norm": 8.250594139099121, "learning_rate": 9.712916695187898e-05, "loss": 1.0606, "step": 2848 }, { "epoch": 0.19303475845246967, "grad_norm": 7.16335916519165, "learning_rate": 9.712779793278116e-05, "loss": 0.9988, "step": 2849 }, { "epoch": 0.19310251372044177, "grad_norm": 9.089055061340332, "learning_rate": 9.712642891368336e-05, "loss": 0.9368, "step": 2850 }, { "epoch": 0.19317026898841386, "grad_norm": 7.568434238433838, "learning_rate": 9.712505989458554e-05, "loss": 0.9417, "step": 2851 }, { "epoch": 0.19323802425638592, "grad_norm": 8.470823287963867, "learning_rate": 9.712369087548772e-05, "loss": 1.0271, "step": 2852 }, { "epoch": 0.193305779524358, "grad_norm": 7.746623992919922, "learning_rate": 9.71223218563899e-05, "loss": 0.7641, "step": 2853 }, { "epoch": 0.1933735347923301, "grad_norm": 7.44852352142334, "learning_rate": 9.712095283729209e-05, "loss": 0.9984, "step": 2854 }, { "epoch": 0.1934412900603022, "grad_norm": 7.393777847290039, "learning_rate": 9.711958381819427e-05, "loss": 0.8515, "step": 2855 }, { "epoch": 0.19350904532827426, "grad_norm": 8.247236251831055, "learning_rate": 9.711821479909645e-05, "loss": 0.9867, "step": 2856 }, { "epoch": 0.19357680059624635, "grad_norm": 8.484920501708984, "learning_rate": 9.711684577999863e-05, "loss": 0.9099, "step": 2857 }, { "epoch": 0.19364455586421844, "grad_norm": 9.85857105255127, "learning_rate": 9.711547676090081e-05, "loss": 1.0134, "step": 2858 }, { "epoch": 0.19371231113219053, "grad_norm": 12.191691398620605, "learning_rate": 9.7114107741803e-05, "loss": 1.15, "step": 2859 }, { "epoch": 0.19378006640016263, "grad_norm": 9.691742897033691, "learning_rate": 9.711273872270519e-05, "loss": 1.2415, "step": 2860 }, { "epoch": 0.1938478216681347, "grad_norm": 11.395289421081543, "learning_rate": 9.711136970360737e-05, "loss": 1.3239, "step": 2861 }, { "epoch": 0.19391557693610678, "grad_norm": 9.222856521606445, "learning_rate": 9.711000068450955e-05, "loss": 0.9087, "step": 2862 }, { "epoch": 0.19398333220407887, "grad_norm": 8.062904357910156, "learning_rate": 9.710863166541174e-05, "loss": 1.093, "step": 2863 }, { "epoch": 0.19405108747205096, "grad_norm": 8.160481452941895, "learning_rate": 9.710726264631392e-05, "loss": 0.842, "step": 2864 }, { "epoch": 0.19411884274002303, "grad_norm": 8.165858268737793, "learning_rate": 9.71058936272161e-05, "loss": 0.8807, "step": 2865 }, { "epoch": 0.19418659800799512, "grad_norm": 8.120240211486816, "learning_rate": 9.710452460811828e-05, "loss": 0.9733, "step": 2866 }, { "epoch": 0.1942543532759672, "grad_norm": 11.363536834716797, "learning_rate": 9.710315558902046e-05, "loss": 1.1144, "step": 2867 }, { "epoch": 0.1943221085439393, "grad_norm": 8.776150703430176, "learning_rate": 9.710178656992266e-05, "loss": 0.7561, "step": 2868 }, { "epoch": 0.19438986381191137, "grad_norm": 9.393696784973145, "learning_rate": 9.710041755082484e-05, "loss": 0.8092, "step": 2869 }, { "epoch": 0.19445761907988346, "grad_norm": 10.820277214050293, "learning_rate": 9.709904853172702e-05, "loss": 0.9905, "step": 2870 }, { "epoch": 0.19452537434785555, "grad_norm": 8.041844367980957, "learning_rate": 9.70976795126292e-05, "loss": 0.8801, "step": 2871 }, { "epoch": 0.19459312961582764, "grad_norm": 8.05355167388916, "learning_rate": 9.709631049353138e-05, "loss": 0.9946, "step": 2872 }, { "epoch": 0.1946608848837997, "grad_norm": 9.46949577331543, "learning_rate": 9.709494147443357e-05, "loss": 1.0582, "step": 2873 }, { "epoch": 0.1947286401517718, "grad_norm": 8.796204566955566, "learning_rate": 9.709357245533575e-05, "loss": 1.1542, "step": 2874 }, { "epoch": 0.1947963954197439, "grad_norm": 10.45006275177002, "learning_rate": 9.709220343623793e-05, "loss": 1.2363, "step": 2875 }, { "epoch": 0.19486415068771598, "grad_norm": 10.490102767944336, "learning_rate": 9.709083441714011e-05, "loss": 1.1977, "step": 2876 }, { "epoch": 0.19493190595568805, "grad_norm": 9.048376083374023, "learning_rate": 9.708946539804231e-05, "loss": 0.8127, "step": 2877 }, { "epoch": 0.19499966122366014, "grad_norm": 6.523111343383789, "learning_rate": 9.708809637894449e-05, "loss": 0.8182, "step": 2878 }, { "epoch": 0.19506741649163223, "grad_norm": 7.170145511627197, "learning_rate": 9.708672735984667e-05, "loss": 0.9643, "step": 2879 }, { "epoch": 0.19513517175960432, "grad_norm": 7.9981818199157715, "learning_rate": 9.708535834074886e-05, "loss": 1.3179, "step": 2880 }, { "epoch": 0.19520292702757638, "grad_norm": 8.786405563354492, "learning_rate": 9.708398932165104e-05, "loss": 0.9217, "step": 2881 }, { "epoch": 0.19527068229554848, "grad_norm": 6.7907867431640625, "learning_rate": 9.708262030255322e-05, "loss": 1.0381, "step": 2882 }, { "epoch": 0.19533843756352057, "grad_norm": 7.1751556396484375, "learning_rate": 9.708125128345542e-05, "loss": 0.8334, "step": 2883 }, { "epoch": 0.19540619283149266, "grad_norm": 8.236610412597656, "learning_rate": 9.70798822643576e-05, "loss": 1.2058, "step": 2884 }, { "epoch": 0.19547394809946472, "grad_norm": 7.576273441314697, "learning_rate": 9.707851324525978e-05, "loss": 1.091, "step": 2885 }, { "epoch": 0.19554170336743681, "grad_norm": 8.698029518127441, "learning_rate": 9.707714422616197e-05, "loss": 1.0262, "step": 2886 }, { "epoch": 0.1956094586354089, "grad_norm": 8.583345413208008, "learning_rate": 9.707577520706415e-05, "loss": 1.0401, "step": 2887 }, { "epoch": 0.195677213903381, "grad_norm": 7.242405891418457, "learning_rate": 9.707440618796633e-05, "loss": 0.785, "step": 2888 }, { "epoch": 0.19574496917135306, "grad_norm": 8.4541654586792, "learning_rate": 9.707303716886851e-05, "loss": 1.0537, "step": 2889 }, { "epoch": 0.19581272443932515, "grad_norm": 7.838657855987549, "learning_rate": 9.707166814977069e-05, "loss": 1.0062, "step": 2890 }, { "epoch": 0.19588047970729724, "grad_norm": 6.812248229980469, "learning_rate": 9.707029913067289e-05, "loss": 0.885, "step": 2891 }, { "epoch": 0.19594823497526934, "grad_norm": 8.789966583251953, "learning_rate": 9.706893011157507e-05, "loss": 1.0366, "step": 2892 }, { "epoch": 0.1960159902432414, "grad_norm": 8.860052108764648, "learning_rate": 9.706756109247725e-05, "loss": 1.2247, "step": 2893 }, { "epoch": 0.1960837455112135, "grad_norm": 11.235320091247559, "learning_rate": 9.706619207337943e-05, "loss": 1.1216, "step": 2894 }, { "epoch": 0.19615150077918558, "grad_norm": 8.865259170532227, "learning_rate": 9.706482305428162e-05, "loss": 0.9467, "step": 2895 }, { "epoch": 0.19621925604715768, "grad_norm": 8.345112800598145, "learning_rate": 9.70634540351838e-05, "loss": 0.9849, "step": 2896 }, { "epoch": 0.19628701131512974, "grad_norm": 8.98128604888916, "learning_rate": 9.706208501608598e-05, "loss": 1.1421, "step": 2897 }, { "epoch": 0.19635476658310183, "grad_norm": 7.786384582519531, "learning_rate": 9.706071599698816e-05, "loss": 1.0926, "step": 2898 }, { "epoch": 0.19642252185107392, "grad_norm": 8.816730499267578, "learning_rate": 9.705934697789034e-05, "loss": 1.2626, "step": 2899 }, { "epoch": 0.19649027711904601, "grad_norm": 7.831095218658447, "learning_rate": 9.705797795879254e-05, "loss": 0.9355, "step": 2900 }, { "epoch": 0.19655803238701808, "grad_norm": 8.788371086120605, "learning_rate": 9.705660893969472e-05, "loss": 0.8883, "step": 2901 }, { "epoch": 0.19662578765499017, "grad_norm": 8.88425064086914, "learning_rate": 9.70552399205969e-05, "loss": 1.1169, "step": 2902 }, { "epoch": 0.19669354292296226, "grad_norm": 9.648268699645996, "learning_rate": 9.705387090149908e-05, "loss": 1.0628, "step": 2903 }, { "epoch": 0.19676129819093435, "grad_norm": 8.19676685333252, "learning_rate": 9.705250188240126e-05, "loss": 0.9395, "step": 2904 }, { "epoch": 0.19682905345890642, "grad_norm": 7.420725345611572, "learning_rate": 9.705113286330345e-05, "loss": 0.947, "step": 2905 }, { "epoch": 0.1968968087268785, "grad_norm": 7.221796989440918, "learning_rate": 9.704976384420563e-05, "loss": 0.8491, "step": 2906 }, { "epoch": 0.1969645639948506, "grad_norm": 9.932676315307617, "learning_rate": 9.704839482510781e-05, "loss": 1.117, "step": 2907 }, { "epoch": 0.1970323192628227, "grad_norm": 11.816266059875488, "learning_rate": 9.704702580600999e-05, "loss": 1.4347, "step": 2908 }, { "epoch": 0.19710007453079476, "grad_norm": 8.804407119750977, "learning_rate": 9.704565678691219e-05, "loss": 0.808, "step": 2909 }, { "epoch": 0.19716782979876685, "grad_norm": 7.740353584289551, "learning_rate": 9.704428776781437e-05, "loss": 1.0732, "step": 2910 }, { "epoch": 0.19723558506673894, "grad_norm": 12.040196418762207, "learning_rate": 9.704291874871655e-05, "loss": 0.947, "step": 2911 }, { "epoch": 0.19730334033471103, "grad_norm": 7.727171421051025, "learning_rate": 9.704154972961873e-05, "loss": 0.8211, "step": 2912 }, { "epoch": 0.19737109560268312, "grad_norm": 8.779428482055664, "learning_rate": 9.704018071052091e-05, "loss": 1.2204, "step": 2913 }, { "epoch": 0.19743885087065519, "grad_norm": 7.907576084136963, "learning_rate": 9.70388116914231e-05, "loss": 1.0594, "step": 2914 }, { "epoch": 0.19750660613862728, "grad_norm": 6.769292831420898, "learning_rate": 9.703744267232528e-05, "loss": 0.8743, "step": 2915 }, { "epoch": 0.19757436140659937, "grad_norm": 8.966355323791504, "learning_rate": 9.703607365322746e-05, "loss": 1.1607, "step": 2916 }, { "epoch": 0.19764211667457146, "grad_norm": 7.389810085296631, "learning_rate": 9.703470463412964e-05, "loss": 0.9728, "step": 2917 }, { "epoch": 0.19770987194254352, "grad_norm": 10.877386093139648, "learning_rate": 9.703333561503184e-05, "loss": 1.0623, "step": 2918 }, { "epoch": 0.19777762721051562, "grad_norm": 9.160116195678711, "learning_rate": 9.703196659593402e-05, "loss": 1.1374, "step": 2919 }, { "epoch": 0.1978453824784877, "grad_norm": 7.6289167404174805, "learning_rate": 9.70305975768362e-05, "loss": 0.9498, "step": 2920 }, { "epoch": 0.1979131377464598, "grad_norm": 8.663583755493164, "learning_rate": 9.702922855773838e-05, "loss": 1.1595, "step": 2921 }, { "epoch": 0.19798089301443186, "grad_norm": 7.595486640930176, "learning_rate": 9.702785953864056e-05, "loss": 1.0623, "step": 2922 }, { "epoch": 0.19804864828240396, "grad_norm": 8.680171966552734, "learning_rate": 9.702649051954275e-05, "loss": 1.0138, "step": 2923 }, { "epoch": 0.19811640355037605, "grad_norm": 9.998015403747559, "learning_rate": 9.702512150044493e-05, "loss": 1.2006, "step": 2924 }, { "epoch": 0.19818415881834814, "grad_norm": 7.801429271697998, "learning_rate": 9.702375248134711e-05, "loss": 0.8467, "step": 2925 }, { "epoch": 0.1982519140863202, "grad_norm": 8.570688247680664, "learning_rate": 9.70223834622493e-05, "loss": 1.1439, "step": 2926 }, { "epoch": 0.1983196693542923, "grad_norm": 7.835936069488525, "learning_rate": 9.702101444315149e-05, "loss": 0.9671, "step": 2927 }, { "epoch": 0.19838742462226439, "grad_norm": 8.415708541870117, "learning_rate": 9.701964542405367e-05, "loss": 1.0381, "step": 2928 }, { "epoch": 0.19845517989023648, "grad_norm": 10.370524406433105, "learning_rate": 9.701827640495586e-05, "loss": 1.151, "step": 2929 }, { "epoch": 0.19852293515820854, "grad_norm": 8.228797912597656, "learning_rate": 9.701690738585804e-05, "loss": 1.093, "step": 2930 }, { "epoch": 0.19859069042618063, "grad_norm": 8.070756912231445, "learning_rate": 9.701553836676022e-05, "loss": 1.0272, "step": 2931 }, { "epoch": 0.19865844569415272, "grad_norm": 9.19532585144043, "learning_rate": 9.701416934766242e-05, "loss": 1.1195, "step": 2932 }, { "epoch": 0.19872620096212482, "grad_norm": 10.692606925964355, "learning_rate": 9.70128003285646e-05, "loss": 1.3148, "step": 2933 }, { "epoch": 0.19879395623009688, "grad_norm": 10.058424949645996, "learning_rate": 9.701143130946678e-05, "loss": 1.4236, "step": 2934 }, { "epoch": 0.19886171149806897, "grad_norm": 6.197395324707031, "learning_rate": 9.701006229036896e-05, "loss": 0.8822, "step": 2935 }, { "epoch": 0.19892946676604106, "grad_norm": 7.603270053863525, "learning_rate": 9.700869327127114e-05, "loss": 0.9655, "step": 2936 }, { "epoch": 0.19899722203401315, "grad_norm": 8.192676544189453, "learning_rate": 9.700732425217333e-05, "loss": 0.9668, "step": 2937 }, { "epoch": 0.19906497730198522, "grad_norm": 7.121623992919922, "learning_rate": 9.700595523307551e-05, "loss": 1.0037, "step": 2938 }, { "epoch": 0.1991327325699573, "grad_norm": 6.8974127769470215, "learning_rate": 9.700458621397769e-05, "loss": 0.876, "step": 2939 }, { "epoch": 0.1992004878379294, "grad_norm": 7.590656757354736, "learning_rate": 9.700321719487987e-05, "loss": 1.0346, "step": 2940 }, { "epoch": 0.1992682431059015, "grad_norm": 8.530266761779785, "learning_rate": 9.700184817578207e-05, "loss": 1.0882, "step": 2941 }, { "epoch": 0.19933599837387356, "grad_norm": 8.064129829406738, "learning_rate": 9.700047915668425e-05, "loss": 0.9949, "step": 2942 }, { "epoch": 0.19940375364184565, "grad_norm": 7.23117208480835, "learning_rate": 9.699911013758643e-05, "loss": 0.7269, "step": 2943 }, { "epoch": 0.19947150890981774, "grad_norm": 7.326268196105957, "learning_rate": 9.69977411184886e-05, "loss": 0.8542, "step": 2944 }, { "epoch": 0.19953926417778983, "grad_norm": 7.708505153656006, "learning_rate": 9.699637209939079e-05, "loss": 0.8206, "step": 2945 }, { "epoch": 0.1996070194457619, "grad_norm": 10.134513854980469, "learning_rate": 9.699500308029298e-05, "loss": 1.0079, "step": 2946 }, { "epoch": 0.199674774713734, "grad_norm": 9.264663696289062, "learning_rate": 9.699363406119516e-05, "loss": 1.1702, "step": 2947 }, { "epoch": 0.19974252998170608, "grad_norm": 8.894827842712402, "learning_rate": 9.699226504209734e-05, "loss": 0.9184, "step": 2948 }, { "epoch": 0.19981028524967817, "grad_norm": 6.935434341430664, "learning_rate": 9.699089602299952e-05, "loss": 0.8554, "step": 2949 }, { "epoch": 0.19987804051765024, "grad_norm": 9.57607364654541, "learning_rate": 9.698952700390172e-05, "loss": 1.0937, "step": 2950 }, { "epoch": 0.19994579578562233, "grad_norm": 7.99752140045166, "learning_rate": 9.69881579848039e-05, "loss": 1.0797, "step": 2951 }, { "epoch": 0.20001355105359442, "grad_norm": 8.067659378051758, "learning_rate": 9.698678896570608e-05, "loss": 1.0515, "step": 2952 }, { "epoch": 0.2000813063215665, "grad_norm": 9.66697883605957, "learning_rate": 9.698541994660826e-05, "loss": 1.289, "step": 2953 }, { "epoch": 0.20014906158953857, "grad_norm": 7.3660664558410645, "learning_rate": 9.698405092751044e-05, "loss": 0.7409, "step": 2954 }, { "epoch": 0.20021681685751067, "grad_norm": 8.254073143005371, "learning_rate": 9.698268190841263e-05, "loss": 1.1095, "step": 2955 }, { "epoch": 0.20028457212548276, "grad_norm": 8.221102714538574, "learning_rate": 9.698131288931481e-05, "loss": 1.0083, "step": 2956 }, { "epoch": 0.20035232739345485, "grad_norm": 9.813411712646484, "learning_rate": 9.697994387021699e-05, "loss": 1.2888, "step": 2957 }, { "epoch": 0.2004200826614269, "grad_norm": 10.87628173828125, "learning_rate": 9.697857485111917e-05, "loss": 1.16, "step": 2958 }, { "epoch": 0.200487837929399, "grad_norm": 7.094732284545898, "learning_rate": 9.697720583202135e-05, "loss": 0.8959, "step": 2959 }, { "epoch": 0.2005555931973711, "grad_norm": 9.814677238464355, "learning_rate": 9.697583681292355e-05, "loss": 1.1011, "step": 2960 }, { "epoch": 0.2006233484653432, "grad_norm": 8.982966423034668, "learning_rate": 9.697446779382573e-05, "loss": 1.001, "step": 2961 }, { "epoch": 0.20069110373331525, "grad_norm": 8.49453067779541, "learning_rate": 9.697309877472791e-05, "loss": 0.9848, "step": 2962 }, { "epoch": 0.20075885900128734, "grad_norm": 7.239814758300781, "learning_rate": 9.697172975563009e-05, "loss": 0.8252, "step": 2963 }, { "epoch": 0.20082661426925943, "grad_norm": 10.874746322631836, "learning_rate": 9.697036073653228e-05, "loss": 1.0683, "step": 2964 }, { "epoch": 0.20089436953723153, "grad_norm": 7.679197311401367, "learning_rate": 9.696899171743446e-05, "loss": 1.2797, "step": 2965 }, { "epoch": 0.20096212480520362, "grad_norm": 10.089177131652832, "learning_rate": 9.696762269833664e-05, "loss": 1.1905, "step": 2966 }, { "epoch": 0.20102988007317568, "grad_norm": 8.182350158691406, "learning_rate": 9.696625367923882e-05, "loss": 0.9917, "step": 2967 }, { "epoch": 0.20109763534114777, "grad_norm": 7.8756256103515625, "learning_rate": 9.6964884660141e-05, "loss": 1.0368, "step": 2968 }, { "epoch": 0.20116539060911987, "grad_norm": 9.193910598754883, "learning_rate": 9.69635156410432e-05, "loss": 1.094, "step": 2969 }, { "epoch": 0.20123314587709196, "grad_norm": 8.854869842529297, "learning_rate": 9.696214662194538e-05, "loss": 1.1344, "step": 2970 }, { "epoch": 0.20130090114506402, "grad_norm": 10.419108390808105, "learning_rate": 9.696077760284756e-05, "loss": 0.9937, "step": 2971 }, { "epoch": 0.2013686564130361, "grad_norm": 9.329347610473633, "learning_rate": 9.695940858374975e-05, "loss": 0.9454, "step": 2972 }, { "epoch": 0.2014364116810082, "grad_norm": 8.23154067993164, "learning_rate": 9.695803956465193e-05, "loss": 0.8691, "step": 2973 }, { "epoch": 0.2015041669489803, "grad_norm": 8.016939163208008, "learning_rate": 9.695667054555411e-05, "loss": 0.9182, "step": 2974 }, { "epoch": 0.20157192221695236, "grad_norm": 9.717400550842285, "learning_rate": 9.69553015264563e-05, "loss": 1.1251, "step": 2975 }, { "epoch": 0.20163967748492445, "grad_norm": 10.470111846923828, "learning_rate": 9.695393250735849e-05, "loss": 1.1634, "step": 2976 }, { "epoch": 0.20170743275289654, "grad_norm": 8.540326118469238, "learning_rate": 9.695256348826067e-05, "loss": 0.9229, "step": 2977 }, { "epoch": 0.20177518802086863, "grad_norm": 6.997597694396973, "learning_rate": 9.695119446916286e-05, "loss": 0.9911, "step": 2978 }, { "epoch": 0.2018429432888407, "grad_norm": 7.23951530456543, "learning_rate": 9.694982545006504e-05, "loss": 0.8372, "step": 2979 }, { "epoch": 0.2019106985568128, "grad_norm": 9.217951774597168, "learning_rate": 9.694845643096722e-05, "loss": 0.8154, "step": 2980 }, { "epoch": 0.20197845382478488, "grad_norm": 8.128033638000488, "learning_rate": 9.69470874118694e-05, "loss": 0.9494, "step": 2981 }, { "epoch": 0.20204620909275697, "grad_norm": 8.328935623168945, "learning_rate": 9.694571839277158e-05, "loss": 1.0537, "step": 2982 }, { "epoch": 0.20211396436072904, "grad_norm": 8.877389907836914, "learning_rate": 9.694434937367378e-05, "loss": 0.8683, "step": 2983 }, { "epoch": 0.20218171962870113, "grad_norm": 7.285436630249023, "learning_rate": 9.694298035457596e-05, "loss": 0.7514, "step": 2984 }, { "epoch": 0.20224947489667322, "grad_norm": 9.209798812866211, "learning_rate": 9.694161133547814e-05, "loss": 1.0771, "step": 2985 }, { "epoch": 0.2023172301646453, "grad_norm": 8.625777244567871, "learning_rate": 9.694024231638032e-05, "loss": 1.0085, "step": 2986 }, { "epoch": 0.20238498543261738, "grad_norm": 8.50123405456543, "learning_rate": 9.693887329728251e-05, "loss": 0.969, "step": 2987 }, { "epoch": 0.20245274070058947, "grad_norm": 7.314642429351807, "learning_rate": 9.693750427818469e-05, "loss": 0.8075, "step": 2988 }, { "epoch": 0.20252049596856156, "grad_norm": 9.474241256713867, "learning_rate": 9.693613525908687e-05, "loss": 1.0989, "step": 2989 }, { "epoch": 0.20258825123653365, "grad_norm": 7.3510637283325195, "learning_rate": 9.693476623998905e-05, "loss": 0.8044, "step": 2990 }, { "epoch": 0.20265600650450571, "grad_norm": 7.203106880187988, "learning_rate": 9.693339722089123e-05, "loss": 1.0359, "step": 2991 }, { "epoch": 0.2027237617724778, "grad_norm": 6.71024227142334, "learning_rate": 9.693202820179343e-05, "loss": 0.8687, "step": 2992 }, { "epoch": 0.2027915170404499, "grad_norm": 8.327759742736816, "learning_rate": 9.69306591826956e-05, "loss": 1.0042, "step": 2993 }, { "epoch": 0.202859272308422, "grad_norm": 8.682476997375488, "learning_rate": 9.692929016359779e-05, "loss": 1.0273, "step": 2994 }, { "epoch": 0.20292702757639405, "grad_norm": 8.555792808532715, "learning_rate": 9.692792114449997e-05, "loss": 0.8712, "step": 2995 }, { "epoch": 0.20299478284436615, "grad_norm": 11.036639213562012, "learning_rate": 9.692655212540216e-05, "loss": 1.1452, "step": 2996 }, { "epoch": 0.20306253811233824, "grad_norm": 10.207952499389648, "learning_rate": 9.692518310630434e-05, "loss": 1.2267, "step": 2997 }, { "epoch": 0.20313029338031033, "grad_norm": 8.715107917785645, "learning_rate": 9.692381408720652e-05, "loss": 1.151, "step": 2998 }, { "epoch": 0.2031980486482824, "grad_norm": 8.72461986541748, "learning_rate": 9.69224450681087e-05, "loss": 0.9787, "step": 2999 }, { "epoch": 0.20326580391625448, "grad_norm": 8.717243194580078, "learning_rate": 9.692107604901088e-05, "loss": 1.0734, "step": 3000 }, { "epoch": 0.20333355918422658, "grad_norm": 7.039597511291504, "learning_rate": 9.691970702991308e-05, "loss": 0.9215, "step": 3001 }, { "epoch": 0.20340131445219867, "grad_norm": 10.568238258361816, "learning_rate": 9.691833801081526e-05, "loss": 1.2248, "step": 3002 }, { "epoch": 0.20346906972017073, "grad_norm": 9.515549659729004, "learning_rate": 9.691696899171744e-05, "loss": 1.1401, "step": 3003 }, { "epoch": 0.20353682498814282, "grad_norm": 9.650483131408691, "learning_rate": 9.691559997261962e-05, "loss": 0.9839, "step": 3004 }, { "epoch": 0.20360458025611491, "grad_norm": 6.817119598388672, "learning_rate": 9.691423095352181e-05, "loss": 0.8542, "step": 3005 }, { "epoch": 0.203672335524087, "grad_norm": 7.616591930389404, "learning_rate": 9.691286193442399e-05, "loss": 0.9647, "step": 3006 }, { "epoch": 0.20374009079205907, "grad_norm": 7.2600274085998535, "learning_rate": 9.691149291532617e-05, "loss": 0.962, "step": 3007 }, { "epoch": 0.20380784606003116, "grad_norm": 9.714008331298828, "learning_rate": 9.691012389622835e-05, "loss": 1.1567, "step": 3008 }, { "epoch": 0.20387560132800325, "grad_norm": 9.095394134521484, "learning_rate": 9.690875487713053e-05, "loss": 1.1315, "step": 3009 }, { "epoch": 0.20394335659597534, "grad_norm": 9.683954238891602, "learning_rate": 9.690738585803273e-05, "loss": 1.4039, "step": 3010 }, { "epoch": 0.2040111118639474, "grad_norm": 8.829015731811523, "learning_rate": 9.69060168389349e-05, "loss": 0.9176, "step": 3011 }, { "epoch": 0.2040788671319195, "grad_norm": 7.835269927978516, "learning_rate": 9.690464781983709e-05, "loss": 1.0077, "step": 3012 }, { "epoch": 0.2041466223998916, "grad_norm": 9.94642448425293, "learning_rate": 9.690327880073927e-05, "loss": 1.0815, "step": 3013 }, { "epoch": 0.20421437766786368, "grad_norm": 8.184757232666016, "learning_rate": 9.690190978164145e-05, "loss": 0.84, "step": 3014 }, { "epoch": 0.20428213293583575, "grad_norm": 9.060220718383789, "learning_rate": 9.690054076254364e-05, "loss": 0.907, "step": 3015 }, { "epoch": 0.20434988820380784, "grad_norm": 6.848534107208252, "learning_rate": 9.689917174344582e-05, "loss": 0.7549, "step": 3016 }, { "epoch": 0.20441764347177993, "grad_norm": 7.820966720581055, "learning_rate": 9.6897802724348e-05, "loss": 0.9899, "step": 3017 }, { "epoch": 0.20448539873975202, "grad_norm": 10.148963928222656, "learning_rate": 9.68964337052502e-05, "loss": 1.2403, "step": 3018 }, { "epoch": 0.20455315400772411, "grad_norm": 8.273184776306152, "learning_rate": 9.689506468615238e-05, "loss": 1.0002, "step": 3019 }, { "epoch": 0.20462090927569618, "grad_norm": 7.618801593780518, "learning_rate": 9.689369566705456e-05, "loss": 0.8864, "step": 3020 }, { "epoch": 0.20468866454366827, "grad_norm": 7.952611446380615, "learning_rate": 9.689232664795675e-05, "loss": 0.8002, "step": 3021 }, { "epoch": 0.20475641981164036, "grad_norm": 9.938977241516113, "learning_rate": 9.689095762885893e-05, "loss": 0.8956, "step": 3022 }, { "epoch": 0.20482417507961245, "grad_norm": 7.807236194610596, "learning_rate": 9.688958860976111e-05, "loss": 1.1777, "step": 3023 }, { "epoch": 0.20489193034758452, "grad_norm": 7.7249369621276855, "learning_rate": 9.68882195906633e-05, "loss": 0.8881, "step": 3024 }, { "epoch": 0.2049596856155566, "grad_norm": 7.747461795806885, "learning_rate": 9.688685057156549e-05, "loss": 0.9844, "step": 3025 }, { "epoch": 0.2050274408835287, "grad_norm": 10.017412185668945, "learning_rate": 9.688548155246767e-05, "loss": 1.0815, "step": 3026 }, { "epoch": 0.2050951961515008, "grad_norm": 6.54990816116333, "learning_rate": 9.688411253336985e-05, "loss": 0.9319, "step": 3027 }, { "epoch": 0.20516295141947286, "grad_norm": 7.358734130859375, "learning_rate": 9.688274351427204e-05, "loss": 0.9313, "step": 3028 }, { "epoch": 0.20523070668744495, "grad_norm": 7.608468055725098, "learning_rate": 9.688137449517422e-05, "loss": 1.0071, "step": 3029 }, { "epoch": 0.20529846195541704, "grad_norm": 7.013155937194824, "learning_rate": 9.68800054760764e-05, "loss": 0.9419, "step": 3030 }, { "epoch": 0.20536621722338913, "grad_norm": 9.907796859741211, "learning_rate": 9.687863645697858e-05, "loss": 1.2082, "step": 3031 }, { "epoch": 0.2054339724913612, "grad_norm": 9.059138298034668, "learning_rate": 9.687726743788076e-05, "loss": 1.0712, "step": 3032 }, { "epoch": 0.20550172775933329, "grad_norm": 10.519928932189941, "learning_rate": 9.687589841878296e-05, "loss": 1.1627, "step": 3033 }, { "epoch": 0.20556948302730538, "grad_norm": 8.381184577941895, "learning_rate": 9.687452939968514e-05, "loss": 1.087, "step": 3034 }, { "epoch": 0.20563723829527747, "grad_norm": 6.514460563659668, "learning_rate": 9.687316038058732e-05, "loss": 0.8657, "step": 3035 }, { "epoch": 0.20570499356324953, "grad_norm": 10.034708023071289, "learning_rate": 9.68717913614895e-05, "loss": 1.1529, "step": 3036 }, { "epoch": 0.20577274883122162, "grad_norm": 7.202263355255127, "learning_rate": 9.687042234239168e-05, "loss": 0.7923, "step": 3037 }, { "epoch": 0.20584050409919372, "grad_norm": 9.7435302734375, "learning_rate": 9.686905332329387e-05, "loss": 1.5071, "step": 3038 }, { "epoch": 0.2059082593671658, "grad_norm": 9.57016658782959, "learning_rate": 9.686768430419605e-05, "loss": 0.883, "step": 3039 }, { "epoch": 0.20597601463513787, "grad_norm": 7.3575029373168945, "learning_rate": 9.686631528509823e-05, "loss": 0.9999, "step": 3040 }, { "epoch": 0.20604376990310996, "grad_norm": 9.0224027633667, "learning_rate": 9.686494626600041e-05, "loss": 1.0121, "step": 3041 }, { "epoch": 0.20611152517108206, "grad_norm": 11.173224449157715, "learning_rate": 9.68635772469026e-05, "loss": 1.0556, "step": 3042 }, { "epoch": 0.20617928043905415, "grad_norm": 8.858287811279297, "learning_rate": 9.686220822780479e-05, "loss": 0.9377, "step": 3043 }, { "epoch": 0.2062470357070262, "grad_norm": 7.3096795082092285, "learning_rate": 9.686083920870697e-05, "loss": 0.8808, "step": 3044 }, { "epoch": 0.2063147909749983, "grad_norm": 8.700214385986328, "learning_rate": 9.685947018960915e-05, "loss": 1.2121, "step": 3045 }, { "epoch": 0.2063825462429704, "grad_norm": 10.57944107055664, "learning_rate": 9.685810117051133e-05, "loss": 1.0931, "step": 3046 }, { "epoch": 0.20645030151094249, "grad_norm": 8.765487670898438, "learning_rate": 9.685673215141352e-05, "loss": 1.0522, "step": 3047 }, { "epoch": 0.20651805677891455, "grad_norm": 7.717139720916748, "learning_rate": 9.68553631323157e-05, "loss": 0.9492, "step": 3048 }, { "epoch": 0.20658581204688664, "grad_norm": 9.301026344299316, "learning_rate": 9.685399411321788e-05, "loss": 1.0191, "step": 3049 }, { "epoch": 0.20665356731485873, "grad_norm": 10.251668930053711, "learning_rate": 9.685262509412006e-05, "loss": 0.991, "step": 3050 }, { "epoch": 0.20672132258283082, "grad_norm": 11.597551345825195, "learning_rate": 9.685125607502226e-05, "loss": 1.0169, "step": 3051 }, { "epoch": 0.2067890778508029, "grad_norm": 10.293901443481445, "learning_rate": 9.684988705592444e-05, "loss": 1.0623, "step": 3052 }, { "epoch": 0.20685683311877498, "grad_norm": 8.496854782104492, "learning_rate": 9.684851803682662e-05, "loss": 1.0466, "step": 3053 }, { "epoch": 0.20692458838674707, "grad_norm": 6.76383638381958, "learning_rate": 9.68471490177288e-05, "loss": 0.8155, "step": 3054 }, { "epoch": 0.20699234365471916, "grad_norm": 8.168519020080566, "learning_rate": 9.684577999863098e-05, "loss": 0.9663, "step": 3055 }, { "epoch": 0.20706009892269123, "grad_norm": 8.0086030960083, "learning_rate": 9.684441097953317e-05, "loss": 0.889, "step": 3056 }, { "epoch": 0.20712785419066332, "grad_norm": 8.347359657287598, "learning_rate": 9.684304196043535e-05, "loss": 0.938, "step": 3057 }, { "epoch": 0.2071956094586354, "grad_norm": 8.0283203125, "learning_rate": 9.684167294133753e-05, "loss": 0.7633, "step": 3058 }, { "epoch": 0.2072633647266075, "grad_norm": 8.509317398071289, "learning_rate": 9.684030392223971e-05, "loss": 1.1052, "step": 3059 }, { "epoch": 0.20733111999457957, "grad_norm": 9.441505432128906, "learning_rate": 9.683893490314189e-05, "loss": 1.1177, "step": 3060 }, { "epoch": 0.20739887526255166, "grad_norm": 8.131098747253418, "learning_rate": 9.683756588404409e-05, "loss": 0.9272, "step": 3061 }, { "epoch": 0.20746663053052375, "grad_norm": 8.423643112182617, "learning_rate": 9.683619686494627e-05, "loss": 0.875, "step": 3062 }, { "epoch": 0.20753438579849584, "grad_norm": 7.513223171234131, "learning_rate": 9.683482784584845e-05, "loss": 1.0659, "step": 3063 }, { "epoch": 0.2076021410664679, "grad_norm": 7.5881171226501465, "learning_rate": 9.683345882675063e-05, "loss": 0.9466, "step": 3064 }, { "epoch": 0.20766989633444, "grad_norm": 8.717775344848633, "learning_rate": 9.683208980765282e-05, "loss": 1.1019, "step": 3065 }, { "epoch": 0.2077376516024121, "grad_norm": 11.917694091796875, "learning_rate": 9.6830720788555e-05, "loss": 1.1085, "step": 3066 }, { "epoch": 0.20780540687038418, "grad_norm": 9.28741455078125, "learning_rate": 9.682935176945718e-05, "loss": 1.1458, "step": 3067 }, { "epoch": 0.20787316213835624, "grad_norm": 6.746860980987549, "learning_rate": 9.682798275035938e-05, "loss": 0.929, "step": 3068 }, { "epoch": 0.20794091740632833, "grad_norm": 8.48763370513916, "learning_rate": 9.682661373126156e-05, "loss": 1.0702, "step": 3069 }, { "epoch": 0.20800867267430043, "grad_norm": 8.871308326721191, "learning_rate": 9.682524471216374e-05, "loss": 0.9199, "step": 3070 }, { "epoch": 0.20807642794227252, "grad_norm": 8.275801658630371, "learning_rate": 9.682387569306593e-05, "loss": 1.0756, "step": 3071 }, { "epoch": 0.2081441832102446, "grad_norm": 8.985222816467285, "learning_rate": 9.682250667396811e-05, "loss": 1.2364, "step": 3072 }, { "epoch": 0.20821193847821667, "grad_norm": 7.749682426452637, "learning_rate": 9.682113765487029e-05, "loss": 0.9444, "step": 3073 }, { "epoch": 0.20827969374618877, "grad_norm": 9.111614227294922, "learning_rate": 9.681976863577248e-05, "loss": 1.338, "step": 3074 }, { "epoch": 0.20834744901416086, "grad_norm": 7.2874674797058105, "learning_rate": 9.681839961667466e-05, "loss": 0.838, "step": 3075 }, { "epoch": 0.20841520428213295, "grad_norm": 6.873099327087402, "learning_rate": 9.681703059757685e-05, "loss": 0.9194, "step": 3076 }, { "epoch": 0.208482959550105, "grad_norm": 7.564418792724609, "learning_rate": 9.681566157847903e-05, "loss": 1.1209, "step": 3077 }, { "epoch": 0.2085507148180771, "grad_norm": 9.088560104370117, "learning_rate": 9.68142925593812e-05, "loss": 0.8425, "step": 3078 }, { "epoch": 0.2086184700860492, "grad_norm": 7.182369709014893, "learning_rate": 9.68129235402834e-05, "loss": 1.0938, "step": 3079 }, { "epoch": 0.2086862253540213, "grad_norm": 8.853677749633789, "learning_rate": 9.681155452118558e-05, "loss": 1.2611, "step": 3080 }, { "epoch": 0.20875398062199335, "grad_norm": 8.56440258026123, "learning_rate": 9.681018550208776e-05, "loss": 1.1112, "step": 3081 }, { "epoch": 0.20882173588996544, "grad_norm": 8.356021881103516, "learning_rate": 9.680881648298994e-05, "loss": 1.0044, "step": 3082 }, { "epoch": 0.20888949115793753, "grad_norm": 9.083736419677734, "learning_rate": 9.680744746389213e-05, "loss": 1.2065, "step": 3083 }, { "epoch": 0.20895724642590963, "grad_norm": 7.990222454071045, "learning_rate": 9.680607844479432e-05, "loss": 1.0179, "step": 3084 }, { "epoch": 0.2090250016938817, "grad_norm": 8.381364822387695, "learning_rate": 9.68047094256965e-05, "loss": 0.8392, "step": 3085 }, { "epoch": 0.20909275696185378, "grad_norm": 9.017950057983398, "learning_rate": 9.680334040659868e-05, "loss": 0.9757, "step": 3086 }, { "epoch": 0.20916051222982587, "grad_norm": 8.525566101074219, "learning_rate": 9.680197138750086e-05, "loss": 1.1521, "step": 3087 }, { "epoch": 0.20922826749779797, "grad_norm": 8.432148933410645, "learning_rate": 9.680060236840305e-05, "loss": 0.8559, "step": 3088 }, { "epoch": 0.20929602276577003, "grad_norm": 9.985367774963379, "learning_rate": 9.679923334930523e-05, "loss": 1.2035, "step": 3089 }, { "epoch": 0.20936377803374212, "grad_norm": 9.771974563598633, "learning_rate": 9.679786433020741e-05, "loss": 1.1614, "step": 3090 }, { "epoch": 0.2094315333017142, "grad_norm": 10.646146774291992, "learning_rate": 9.679649531110959e-05, "loss": 1.2165, "step": 3091 }, { "epoch": 0.2094992885696863, "grad_norm": 8.93340015411377, "learning_rate": 9.679512629201177e-05, "loss": 0.9883, "step": 3092 }, { "epoch": 0.20956704383765837, "grad_norm": 8.498669624328613, "learning_rate": 9.679375727291397e-05, "loss": 1.0318, "step": 3093 }, { "epoch": 0.20963479910563046, "grad_norm": 7.536258220672607, "learning_rate": 9.679238825381615e-05, "loss": 0.9949, "step": 3094 }, { "epoch": 0.20970255437360255, "grad_norm": 8.314896583557129, "learning_rate": 9.679101923471833e-05, "loss": 1.0502, "step": 3095 }, { "epoch": 0.20977030964157464, "grad_norm": 7.470542907714844, "learning_rate": 9.67896502156205e-05, "loss": 0.8871, "step": 3096 }, { "epoch": 0.2098380649095467, "grad_norm": 8.951095581054688, "learning_rate": 9.67882811965227e-05, "loss": 0.9793, "step": 3097 }, { "epoch": 0.2099058201775188, "grad_norm": 7.879035472869873, "learning_rate": 9.678691217742488e-05, "loss": 1.0193, "step": 3098 }, { "epoch": 0.2099735754454909, "grad_norm": 8.890814781188965, "learning_rate": 9.678554315832706e-05, "loss": 0.9738, "step": 3099 }, { "epoch": 0.21004133071346298, "grad_norm": 8.863816261291504, "learning_rate": 9.678417413922924e-05, "loss": 1.1213, "step": 3100 }, { "epoch": 0.21010908598143505, "grad_norm": 9.59538745880127, "learning_rate": 9.678280512013142e-05, "loss": 1.0215, "step": 3101 }, { "epoch": 0.21017684124940714, "grad_norm": 8.811614990234375, "learning_rate": 9.678143610103362e-05, "loss": 0.8703, "step": 3102 }, { "epoch": 0.21024459651737923, "grad_norm": 7.274720191955566, "learning_rate": 9.67800670819358e-05, "loss": 0.898, "step": 3103 }, { "epoch": 0.21031235178535132, "grad_norm": 11.239364624023438, "learning_rate": 9.677869806283798e-05, "loss": 0.9823, "step": 3104 }, { "epoch": 0.21038010705332338, "grad_norm": 8.807086944580078, "learning_rate": 9.677732904374016e-05, "loss": 1.0074, "step": 3105 }, { "epoch": 0.21044786232129548, "grad_norm": 9.065536499023438, "learning_rate": 9.677596002464235e-05, "loss": 1.1861, "step": 3106 }, { "epoch": 0.21051561758926757, "grad_norm": 11.035104751586914, "learning_rate": 9.677459100554453e-05, "loss": 1.0067, "step": 3107 }, { "epoch": 0.21058337285723966, "grad_norm": 8.010696411132812, "learning_rate": 9.677322198644671e-05, "loss": 1.0855, "step": 3108 }, { "epoch": 0.21065112812521172, "grad_norm": 9.104195594787598, "learning_rate": 9.677185296734889e-05, "loss": 1.0497, "step": 3109 }, { "epoch": 0.21071888339318381, "grad_norm": 8.731512069702148, "learning_rate": 9.677048394825107e-05, "loss": 1.1108, "step": 3110 }, { "epoch": 0.2107866386611559, "grad_norm": 8.823514938354492, "learning_rate": 9.676911492915327e-05, "loss": 1.0271, "step": 3111 }, { "epoch": 0.210854393929128, "grad_norm": 7.446425914764404, "learning_rate": 9.676774591005545e-05, "loss": 0.9182, "step": 3112 }, { "epoch": 0.21092214919710006, "grad_norm": 10.75915241241455, "learning_rate": 9.676637689095763e-05, "loss": 0.93, "step": 3113 }, { "epoch": 0.21098990446507215, "grad_norm": 10.065240859985352, "learning_rate": 9.676500787185982e-05, "loss": 1.0508, "step": 3114 }, { "epoch": 0.21105765973304424, "grad_norm": 8.790117263793945, "learning_rate": 9.6763638852762e-05, "loss": 0.8971, "step": 3115 }, { "epoch": 0.21112541500101634, "grad_norm": 8.286596298217773, "learning_rate": 9.676226983366418e-05, "loss": 0.9403, "step": 3116 }, { "epoch": 0.2111931702689884, "grad_norm": 8.542399406433105, "learning_rate": 9.676090081456637e-05, "loss": 1.1638, "step": 3117 }, { "epoch": 0.2112609255369605, "grad_norm": 9.461727142333984, "learning_rate": 9.675953179546856e-05, "loss": 1.0859, "step": 3118 }, { "epoch": 0.21132868080493258, "grad_norm": 7.0917229652404785, "learning_rate": 9.675816277637074e-05, "loss": 0.9931, "step": 3119 }, { "epoch": 0.21139643607290468, "grad_norm": 9.192744255065918, "learning_rate": 9.675679375727293e-05, "loss": 1.2138, "step": 3120 }, { "epoch": 0.21146419134087674, "grad_norm": 7.744256973266602, "learning_rate": 9.675542473817511e-05, "loss": 1.3615, "step": 3121 }, { "epoch": 0.21153194660884883, "grad_norm": 6.915426254272461, "learning_rate": 9.675405571907729e-05, "loss": 0.89, "step": 3122 }, { "epoch": 0.21159970187682092, "grad_norm": 9.888227462768555, "learning_rate": 9.675268669997947e-05, "loss": 0.9569, "step": 3123 }, { "epoch": 0.21166745714479301, "grad_norm": 7.541590690612793, "learning_rate": 9.675131768088165e-05, "loss": 0.9638, "step": 3124 }, { "epoch": 0.2117352124127651, "grad_norm": 7.883132457733154, "learning_rate": 9.674994866178384e-05, "loss": 0.8901, "step": 3125 }, { "epoch": 0.21180296768073717, "grad_norm": 8.493675231933594, "learning_rate": 9.674857964268602e-05, "loss": 0.8894, "step": 3126 }, { "epoch": 0.21187072294870926, "grad_norm": 7.348284721374512, "learning_rate": 9.67472106235882e-05, "loss": 0.8721, "step": 3127 }, { "epoch": 0.21193847821668135, "grad_norm": 9.094710350036621, "learning_rate": 9.674584160449039e-05, "loss": 0.9255, "step": 3128 }, { "epoch": 0.21200623348465344, "grad_norm": 7.316446304321289, "learning_rate": 9.674447258539258e-05, "loss": 1.2577, "step": 3129 }, { "epoch": 0.2120739887526255, "grad_norm": 8.105271339416504, "learning_rate": 9.674310356629476e-05, "loss": 0.8157, "step": 3130 }, { "epoch": 0.2121417440205976, "grad_norm": 8.433457374572754, "learning_rate": 9.674173454719694e-05, "loss": 1.0398, "step": 3131 }, { "epoch": 0.2122094992885697, "grad_norm": 7.787237644195557, "learning_rate": 9.674036552809912e-05, "loss": 0.7095, "step": 3132 }, { "epoch": 0.21227725455654178, "grad_norm": 10.60180950164795, "learning_rate": 9.67389965090013e-05, "loss": 1.0482, "step": 3133 }, { "epoch": 0.21234500982451385, "grad_norm": 8.428773880004883, "learning_rate": 9.67376274899035e-05, "loss": 0.8677, "step": 3134 }, { "epoch": 0.21241276509248594, "grad_norm": 8.204195022583008, "learning_rate": 9.673625847080568e-05, "loss": 1.14, "step": 3135 }, { "epoch": 0.21248052036045803, "grad_norm": 6.449087619781494, "learning_rate": 9.673488945170786e-05, "loss": 1.0181, "step": 3136 }, { "epoch": 0.21254827562843012, "grad_norm": 9.588041305541992, "learning_rate": 9.673352043261004e-05, "loss": 0.9996, "step": 3137 }, { "epoch": 0.21261603089640219, "grad_norm": 9.626228332519531, "learning_rate": 9.673215141351223e-05, "loss": 1.0501, "step": 3138 }, { "epoch": 0.21268378616437428, "grad_norm": 9.036309242248535, "learning_rate": 9.673078239441441e-05, "loss": 1.093, "step": 3139 }, { "epoch": 0.21275154143234637, "grad_norm": 9.415257453918457, "learning_rate": 9.672941337531659e-05, "loss": 1.0983, "step": 3140 }, { "epoch": 0.21281929670031846, "grad_norm": 8.309000015258789, "learning_rate": 9.672804435621877e-05, "loss": 1.0803, "step": 3141 }, { "epoch": 0.21288705196829052, "grad_norm": 7.420774459838867, "learning_rate": 9.672667533712095e-05, "loss": 1.1027, "step": 3142 }, { "epoch": 0.21295480723626262, "grad_norm": 9.291664123535156, "learning_rate": 9.672530631802314e-05, "loss": 1.1617, "step": 3143 }, { "epoch": 0.2130225625042347, "grad_norm": 7.656317234039307, "learning_rate": 9.672393729892533e-05, "loss": 0.9423, "step": 3144 }, { "epoch": 0.2130903177722068, "grad_norm": 5.812994003295898, "learning_rate": 9.67225682798275e-05, "loss": 0.9187, "step": 3145 }, { "epoch": 0.21315807304017886, "grad_norm": 9.492706298828125, "learning_rate": 9.672119926072969e-05, "loss": 1.1804, "step": 3146 }, { "epoch": 0.21322582830815096, "grad_norm": 9.37720012664795, "learning_rate": 9.671983024163187e-05, "loss": 0.8885, "step": 3147 }, { "epoch": 0.21329358357612305, "grad_norm": 9.183650016784668, "learning_rate": 9.671846122253406e-05, "loss": 0.8967, "step": 3148 }, { "epoch": 0.21336133884409514, "grad_norm": 8.037003517150879, "learning_rate": 9.671709220343624e-05, "loss": 1.2605, "step": 3149 }, { "epoch": 0.2134290941120672, "grad_norm": 8.694345474243164, "learning_rate": 9.671572318433842e-05, "loss": 1.003, "step": 3150 }, { "epoch": 0.2134968493800393, "grad_norm": 6.429176330566406, "learning_rate": 9.67143541652406e-05, "loss": 1.1018, "step": 3151 }, { "epoch": 0.21356460464801139, "grad_norm": 7.964774131774902, "learning_rate": 9.67129851461428e-05, "loss": 0.9121, "step": 3152 }, { "epoch": 0.21363235991598348, "grad_norm": 7.5509033203125, "learning_rate": 9.671161612704498e-05, "loss": 0.9577, "step": 3153 }, { "epoch": 0.21370011518395554, "grad_norm": 7.968616962432861, "learning_rate": 9.671024710794716e-05, "loss": 1.1337, "step": 3154 }, { "epoch": 0.21376787045192763, "grad_norm": 7.340758800506592, "learning_rate": 9.670887808884934e-05, "loss": 0.8068, "step": 3155 }, { "epoch": 0.21383562571989972, "grad_norm": 7.890449523925781, "learning_rate": 9.670750906975152e-05, "loss": 1.0263, "step": 3156 }, { "epoch": 0.21390338098787182, "grad_norm": 10.908242225646973, "learning_rate": 9.670614005065371e-05, "loss": 1.0384, "step": 3157 }, { "epoch": 0.21397113625584388, "grad_norm": 7.62697696685791, "learning_rate": 9.670477103155589e-05, "loss": 0.918, "step": 3158 }, { "epoch": 0.21403889152381597, "grad_norm": 10.487192153930664, "learning_rate": 9.670340201245807e-05, "loss": 1.1135, "step": 3159 }, { "epoch": 0.21410664679178806, "grad_norm": 7.076718807220459, "learning_rate": 9.670203299336026e-05, "loss": 0.9286, "step": 3160 }, { "epoch": 0.21417440205976015, "grad_norm": 10.05949592590332, "learning_rate": 9.670066397426245e-05, "loss": 1.1317, "step": 3161 }, { "epoch": 0.21424215732773222, "grad_norm": 8.656492233276367, "learning_rate": 9.669929495516463e-05, "loss": 0.9864, "step": 3162 }, { "epoch": 0.2143099125957043, "grad_norm": 10.591787338256836, "learning_rate": 9.669792593606682e-05, "loss": 1.0801, "step": 3163 }, { "epoch": 0.2143776678636764, "grad_norm": 8.19133472442627, "learning_rate": 9.6696556916969e-05, "loss": 0.9377, "step": 3164 }, { "epoch": 0.2144454231316485, "grad_norm": 8.402669906616211, "learning_rate": 9.669518789787118e-05, "loss": 0.9283, "step": 3165 }, { "epoch": 0.21451317839962056, "grad_norm": 9.713315963745117, "learning_rate": 9.669381887877337e-05, "loss": 1.1865, "step": 3166 }, { "epoch": 0.21458093366759265, "grad_norm": 6.777700424194336, "learning_rate": 9.669244985967555e-05, "loss": 0.9094, "step": 3167 }, { "epoch": 0.21464868893556474, "grad_norm": 9.381082534790039, "learning_rate": 9.669108084057773e-05, "loss": 0.7461, "step": 3168 }, { "epoch": 0.21471644420353683, "grad_norm": 8.568790435791016, "learning_rate": 9.668971182147992e-05, "loss": 0.7339, "step": 3169 }, { "epoch": 0.2147841994715089, "grad_norm": 9.257226943969727, "learning_rate": 9.66883428023821e-05, "loss": 1.1892, "step": 3170 }, { "epoch": 0.214851954739481, "grad_norm": 8.544146537780762, "learning_rate": 9.668697378328429e-05, "loss": 1.0307, "step": 3171 }, { "epoch": 0.21491971000745308, "grad_norm": 9.540971755981445, "learning_rate": 9.668560476418647e-05, "loss": 0.8464, "step": 3172 }, { "epoch": 0.21498746527542517, "grad_norm": 9.180089950561523, "learning_rate": 9.668423574508865e-05, "loss": 0.951, "step": 3173 }, { "epoch": 0.21505522054339726, "grad_norm": 10.706409454345703, "learning_rate": 9.668286672599083e-05, "loss": 1.141, "step": 3174 }, { "epoch": 0.21512297581136933, "grad_norm": 9.421865463256836, "learning_rate": 9.668149770689302e-05, "loss": 1.0701, "step": 3175 }, { "epoch": 0.21519073107934142, "grad_norm": 9.120182037353516, "learning_rate": 9.66801286877952e-05, "loss": 1.1618, "step": 3176 }, { "epoch": 0.2152584863473135, "grad_norm": 7.879681587219238, "learning_rate": 9.667875966869738e-05, "loss": 0.8048, "step": 3177 }, { "epoch": 0.2153262416152856, "grad_norm": 7.233819007873535, "learning_rate": 9.667739064959957e-05, "loss": 0.9718, "step": 3178 }, { "epoch": 0.21539399688325767, "grad_norm": 7.79316520690918, "learning_rate": 9.667602163050175e-05, "loss": 1.0999, "step": 3179 }, { "epoch": 0.21546175215122976, "grad_norm": 8.873052597045898, "learning_rate": 9.667465261140394e-05, "loss": 0.9444, "step": 3180 }, { "epoch": 0.21552950741920185, "grad_norm": 7.386806488037109, "learning_rate": 9.667328359230612e-05, "loss": 0.8661, "step": 3181 }, { "epoch": 0.21559726268717394, "grad_norm": 7.682179927825928, "learning_rate": 9.66719145732083e-05, "loss": 0.7904, "step": 3182 }, { "epoch": 0.215665017955146, "grad_norm": 7.9192914962768555, "learning_rate": 9.667054555411048e-05, "loss": 1.0563, "step": 3183 }, { "epoch": 0.2157327732231181, "grad_norm": 6.972530364990234, "learning_rate": 9.666917653501267e-05, "loss": 0.7482, "step": 3184 }, { "epoch": 0.2158005284910902, "grad_norm": 9.844091415405273, "learning_rate": 9.666780751591485e-05, "loss": 1.0289, "step": 3185 }, { "epoch": 0.21586828375906228, "grad_norm": 8.724159240722656, "learning_rate": 9.666643849681704e-05, "loss": 1.1419, "step": 3186 }, { "epoch": 0.21593603902703434, "grad_norm": 8.859795570373535, "learning_rate": 9.666506947771922e-05, "loss": 0.9252, "step": 3187 }, { "epoch": 0.21600379429500643, "grad_norm": 10.741375923156738, "learning_rate": 9.66637004586214e-05, "loss": 1.2328, "step": 3188 }, { "epoch": 0.21607154956297853, "grad_norm": 8.468697547912598, "learning_rate": 9.666233143952359e-05, "loss": 1.1779, "step": 3189 }, { "epoch": 0.21613930483095062, "grad_norm": 7.220376014709473, "learning_rate": 9.666096242042577e-05, "loss": 0.9271, "step": 3190 }, { "epoch": 0.21620706009892268, "grad_norm": 8.3795804977417, "learning_rate": 9.665959340132795e-05, "loss": 1.0941, "step": 3191 }, { "epoch": 0.21627481536689477, "grad_norm": 9.743870735168457, "learning_rate": 9.665822438223013e-05, "loss": 0.9396, "step": 3192 }, { "epoch": 0.21634257063486687, "grad_norm": 8.927998542785645, "learning_rate": 9.665685536313231e-05, "loss": 1.0915, "step": 3193 }, { "epoch": 0.21641032590283896, "grad_norm": 8.214877128601074, "learning_rate": 9.66554863440345e-05, "loss": 0.8767, "step": 3194 }, { "epoch": 0.21647808117081102, "grad_norm": 9.605053901672363, "learning_rate": 9.665411732493669e-05, "loss": 1.1829, "step": 3195 }, { "epoch": 0.2165458364387831, "grad_norm": 8.430122375488281, "learning_rate": 9.665274830583887e-05, "loss": 0.8234, "step": 3196 }, { "epoch": 0.2166135917067552, "grad_norm": 8.868674278259277, "learning_rate": 9.665137928674105e-05, "loss": 1.0467, "step": 3197 }, { "epoch": 0.2166813469747273, "grad_norm": 8.451126098632812, "learning_rate": 9.665001026764324e-05, "loss": 1.134, "step": 3198 }, { "epoch": 0.21674910224269936, "grad_norm": 8.749373435974121, "learning_rate": 9.664864124854542e-05, "loss": 0.9825, "step": 3199 }, { "epoch": 0.21681685751067145, "grad_norm": 8.478997230529785, "learning_rate": 9.66472722294476e-05, "loss": 0.9823, "step": 3200 }, { "epoch": 0.21688461277864354, "grad_norm": 6.554266929626465, "learning_rate": 9.664590321034978e-05, "loss": 1.2293, "step": 3201 }, { "epoch": 0.21695236804661563, "grad_norm": 7.415618896484375, "learning_rate": 9.664453419125196e-05, "loss": 0.9767, "step": 3202 }, { "epoch": 0.2170201233145877, "grad_norm": 9.277578353881836, "learning_rate": 9.664316517215416e-05, "loss": 1.0042, "step": 3203 }, { "epoch": 0.2170878785825598, "grad_norm": 7.1379241943359375, "learning_rate": 9.664179615305634e-05, "loss": 0.924, "step": 3204 }, { "epoch": 0.21715563385053188, "grad_norm": 8.837613105773926, "learning_rate": 9.664042713395852e-05, "loss": 0.9889, "step": 3205 }, { "epoch": 0.21722338911850397, "grad_norm": 8.118345260620117, "learning_rate": 9.663905811486071e-05, "loss": 1.2376, "step": 3206 }, { "epoch": 0.21729114438647604, "grad_norm": 10.383713722229004, "learning_rate": 9.663768909576289e-05, "loss": 0.7988, "step": 3207 }, { "epoch": 0.21735889965444813, "grad_norm": 7.0938262939453125, "learning_rate": 9.663632007666507e-05, "loss": 0.8589, "step": 3208 }, { "epoch": 0.21742665492242022, "grad_norm": 7.732020854949951, "learning_rate": 9.663495105756726e-05, "loss": 1.0837, "step": 3209 }, { "epoch": 0.2174944101903923, "grad_norm": 7.718788146972656, "learning_rate": 9.663358203846944e-05, "loss": 0.8949, "step": 3210 }, { "epoch": 0.21756216545836438, "grad_norm": 9.187674522399902, "learning_rate": 9.663221301937162e-05, "loss": 0.9443, "step": 3211 }, { "epoch": 0.21762992072633647, "grad_norm": 8.596949577331543, "learning_rate": 9.663084400027382e-05, "loss": 1.0855, "step": 3212 }, { "epoch": 0.21769767599430856, "grad_norm": 7.854368209838867, "learning_rate": 9.6629474981176e-05, "loss": 0.9203, "step": 3213 }, { "epoch": 0.21776543126228065, "grad_norm": 9.96597957611084, "learning_rate": 9.662810596207818e-05, "loss": 0.836, "step": 3214 }, { "epoch": 0.21783318653025271, "grad_norm": 7.254636764526367, "learning_rate": 9.662673694298036e-05, "loss": 0.7926, "step": 3215 }, { "epoch": 0.2179009417982248, "grad_norm": 9.079703330993652, "learning_rate": 9.662536792388255e-05, "loss": 1.2072, "step": 3216 }, { "epoch": 0.2179686970661969, "grad_norm": 8.051166534423828, "learning_rate": 9.662399890478473e-05, "loss": 0.9719, "step": 3217 }, { "epoch": 0.218036452334169, "grad_norm": 9.189994812011719, "learning_rate": 9.662262988568691e-05, "loss": 0.9878, "step": 3218 }, { "epoch": 0.21810420760214105, "grad_norm": 9.806111335754395, "learning_rate": 9.66212608665891e-05, "loss": 1.1621, "step": 3219 }, { "epoch": 0.21817196287011315, "grad_norm": 7.833159923553467, "learning_rate": 9.661989184749128e-05, "loss": 0.9811, "step": 3220 }, { "epoch": 0.21823971813808524, "grad_norm": 7.537334442138672, "learning_rate": 9.661852282839347e-05, "loss": 0.9916, "step": 3221 }, { "epoch": 0.21830747340605733, "grad_norm": 9.828896522521973, "learning_rate": 9.661715380929565e-05, "loss": 1.09, "step": 3222 }, { "epoch": 0.2183752286740294, "grad_norm": 7.99267578125, "learning_rate": 9.661578479019783e-05, "loss": 1.0852, "step": 3223 }, { "epoch": 0.21844298394200148, "grad_norm": 8.085976600646973, "learning_rate": 9.661441577110001e-05, "loss": 1.1762, "step": 3224 }, { "epoch": 0.21851073920997358, "grad_norm": 6.523219585418701, "learning_rate": 9.661304675200219e-05, "loss": 1.0814, "step": 3225 }, { "epoch": 0.21857849447794567, "grad_norm": 9.796062469482422, "learning_rate": 9.661167773290438e-05, "loss": 1.1447, "step": 3226 }, { "epoch": 0.21864624974591776, "grad_norm": 8.570687294006348, "learning_rate": 9.661030871380656e-05, "loss": 1.0552, "step": 3227 }, { "epoch": 0.21871400501388982, "grad_norm": 9.70113468170166, "learning_rate": 9.660893969470874e-05, "loss": 1.0379, "step": 3228 }, { "epoch": 0.21878176028186191, "grad_norm": 8.195375442504883, "learning_rate": 9.660757067561093e-05, "loss": 0.9348, "step": 3229 }, { "epoch": 0.218849515549834, "grad_norm": 7.33494758605957, "learning_rate": 9.660620165651312e-05, "loss": 0.8977, "step": 3230 }, { "epoch": 0.2189172708178061, "grad_norm": 7.543430328369141, "learning_rate": 9.66048326374153e-05, "loss": 1.023, "step": 3231 }, { "epoch": 0.21898502608577816, "grad_norm": 8.45494556427002, "learning_rate": 9.660346361831748e-05, "loss": 1.2995, "step": 3232 }, { "epoch": 0.21905278135375025, "grad_norm": 8.197903633117676, "learning_rate": 9.660209459921966e-05, "loss": 0.9651, "step": 3233 }, { "epoch": 0.21912053662172234, "grad_norm": 8.392546653747559, "learning_rate": 9.660072558012184e-05, "loss": 0.9858, "step": 3234 }, { "epoch": 0.21918829188969444, "grad_norm": 6.499540328979492, "learning_rate": 9.659935656102403e-05, "loss": 0.9679, "step": 3235 }, { "epoch": 0.2192560471576665, "grad_norm": 7.784618377685547, "learning_rate": 9.659798754192621e-05, "loss": 0.9316, "step": 3236 }, { "epoch": 0.2193238024256386, "grad_norm": 9.014801979064941, "learning_rate": 9.65966185228284e-05, "loss": 0.9261, "step": 3237 }, { "epoch": 0.21939155769361068, "grad_norm": 8.238675117492676, "learning_rate": 9.659524950373058e-05, "loss": 1.0156, "step": 3238 }, { "epoch": 0.21945931296158278, "grad_norm": 9.447864532470703, "learning_rate": 9.659388048463277e-05, "loss": 1.2186, "step": 3239 }, { "epoch": 0.21952706822955484, "grad_norm": 6.894035816192627, "learning_rate": 9.659251146553495e-05, "loss": 0.9157, "step": 3240 }, { "epoch": 0.21959482349752693, "grad_norm": 7.24024772644043, "learning_rate": 9.659114244643713e-05, "loss": 1.1564, "step": 3241 }, { "epoch": 0.21966257876549902, "grad_norm": 9.135485649108887, "learning_rate": 9.658977342733931e-05, "loss": 1.2002, "step": 3242 }, { "epoch": 0.21973033403347111, "grad_norm": 6.119181156158447, "learning_rate": 9.658840440824149e-05, "loss": 0.788, "step": 3243 }, { "epoch": 0.21979808930144318, "grad_norm": 8.550108909606934, "learning_rate": 9.658703538914368e-05, "loss": 0.9251, "step": 3244 }, { "epoch": 0.21986584456941527, "grad_norm": 9.546792984008789, "learning_rate": 9.658566637004586e-05, "loss": 1.2231, "step": 3245 }, { "epoch": 0.21993359983738736, "grad_norm": 9.451306343078613, "learning_rate": 9.658429735094805e-05, "loss": 0.8991, "step": 3246 }, { "epoch": 0.22000135510535945, "grad_norm": 9.914873123168945, "learning_rate": 9.658292833185023e-05, "loss": 1.3102, "step": 3247 }, { "epoch": 0.22006911037333152, "grad_norm": 10.038002967834473, "learning_rate": 9.65815593127524e-05, "loss": 1.254, "step": 3248 }, { "epoch": 0.2201368656413036, "grad_norm": 7.590802192687988, "learning_rate": 9.65801902936546e-05, "loss": 1.0032, "step": 3249 }, { "epoch": 0.2202046209092757, "grad_norm": 8.632338523864746, "learning_rate": 9.657882127455678e-05, "loss": 0.8922, "step": 3250 }, { "epoch": 0.2202723761772478, "grad_norm": 6.483764171600342, "learning_rate": 9.657745225545896e-05, "loss": 0.9167, "step": 3251 }, { "epoch": 0.22034013144521986, "grad_norm": 8.28577709197998, "learning_rate": 9.657608323636115e-05, "loss": 1.0413, "step": 3252 }, { "epoch": 0.22040788671319195, "grad_norm": 9.218433380126953, "learning_rate": 9.657471421726333e-05, "loss": 1.1681, "step": 3253 }, { "epoch": 0.22047564198116404, "grad_norm": 6.653755187988281, "learning_rate": 9.657334519816552e-05, "loss": 0.9645, "step": 3254 }, { "epoch": 0.22054339724913613, "grad_norm": 8.784723281860352, "learning_rate": 9.657197617906771e-05, "loss": 0.7489, "step": 3255 }, { "epoch": 0.2206111525171082, "grad_norm": 9.041324615478516, "learning_rate": 9.657060715996989e-05, "loss": 1.0478, "step": 3256 }, { "epoch": 0.22067890778508029, "grad_norm": 7.811746597290039, "learning_rate": 9.656923814087207e-05, "loss": 0.9511, "step": 3257 }, { "epoch": 0.22074666305305238, "grad_norm": 9.811897277832031, "learning_rate": 9.656786912177426e-05, "loss": 1.4235, "step": 3258 }, { "epoch": 0.22081441832102447, "grad_norm": 7.313972473144531, "learning_rate": 9.656650010267644e-05, "loss": 1.0075, "step": 3259 }, { "epoch": 0.22088217358899653, "grad_norm": 9.360795021057129, "learning_rate": 9.656513108357862e-05, "loss": 0.9634, "step": 3260 }, { "epoch": 0.22094992885696862, "grad_norm": 9.322724342346191, "learning_rate": 9.65637620644808e-05, "loss": 1.0669, "step": 3261 }, { "epoch": 0.22101768412494072, "grad_norm": 10.269414901733398, "learning_rate": 9.6562393045383e-05, "loss": 1.1288, "step": 3262 }, { "epoch": 0.2210854393929128, "grad_norm": 7.672987937927246, "learning_rate": 9.656102402628518e-05, "loss": 1.0517, "step": 3263 }, { "epoch": 0.22115319466088487, "grad_norm": 8.725695610046387, "learning_rate": 9.655965500718736e-05, "loss": 1.0987, "step": 3264 }, { "epoch": 0.22122094992885696, "grad_norm": 8.577162742614746, "learning_rate": 9.655828598808954e-05, "loss": 0.821, "step": 3265 }, { "epoch": 0.22128870519682906, "grad_norm": 8.497530937194824, "learning_rate": 9.655691696899172e-05, "loss": 0.9171, "step": 3266 }, { "epoch": 0.22135646046480115, "grad_norm": 9.665203094482422, "learning_rate": 9.655554794989391e-05, "loss": 1.105, "step": 3267 }, { "epoch": 0.2214242157327732, "grad_norm": 9.42651081085205, "learning_rate": 9.65541789307961e-05, "loss": 0.9526, "step": 3268 }, { "epoch": 0.2214919710007453, "grad_norm": 7.461474418640137, "learning_rate": 9.655280991169827e-05, "loss": 0.9554, "step": 3269 }, { "epoch": 0.2215597262687174, "grad_norm": 10.56189250946045, "learning_rate": 9.655144089260045e-05, "loss": 1.1504, "step": 3270 }, { "epoch": 0.22162748153668949, "grad_norm": 8.245696067810059, "learning_rate": 9.655007187350265e-05, "loss": 1.0352, "step": 3271 }, { "epoch": 0.22169523680466155, "grad_norm": 9.182881355285645, "learning_rate": 9.654870285440483e-05, "loss": 1.1153, "step": 3272 }, { "epoch": 0.22176299207263364, "grad_norm": 8.101743698120117, "learning_rate": 9.654733383530701e-05, "loss": 1.1399, "step": 3273 }, { "epoch": 0.22183074734060573, "grad_norm": 9.072972297668457, "learning_rate": 9.654596481620919e-05, "loss": 1.1508, "step": 3274 }, { "epoch": 0.22189850260857782, "grad_norm": 8.480910301208496, "learning_rate": 9.654459579711137e-05, "loss": 0.9659, "step": 3275 }, { "epoch": 0.2219662578765499, "grad_norm": 7.095332622528076, "learning_rate": 9.654322677801356e-05, "loss": 0.9417, "step": 3276 }, { "epoch": 0.22203401314452198, "grad_norm": 8.183112144470215, "learning_rate": 9.654185775891574e-05, "loss": 0.9307, "step": 3277 }, { "epoch": 0.22210176841249407, "grad_norm": 8.60648250579834, "learning_rate": 9.654048873981792e-05, "loss": 1.141, "step": 3278 }, { "epoch": 0.22216952368046616, "grad_norm": 6.992738246917725, "learning_rate": 9.65391197207201e-05, "loss": 0.9429, "step": 3279 }, { "epoch": 0.22223727894843825, "grad_norm": 6.998913764953613, "learning_rate": 9.653775070162229e-05, "loss": 1.0159, "step": 3280 }, { "epoch": 0.22230503421641032, "grad_norm": 7.922050952911377, "learning_rate": 9.653638168252448e-05, "loss": 1.1349, "step": 3281 }, { "epoch": 0.2223727894843824, "grad_norm": 9.171984672546387, "learning_rate": 9.653501266342666e-05, "loss": 1.149, "step": 3282 }, { "epoch": 0.2224405447523545, "grad_norm": 9.88930606842041, "learning_rate": 9.653364364432884e-05, "loss": 1.1036, "step": 3283 }, { "epoch": 0.2225083000203266, "grad_norm": 10.978727340698242, "learning_rate": 9.653227462523102e-05, "loss": 1.0345, "step": 3284 }, { "epoch": 0.22257605528829866, "grad_norm": 8.712138175964355, "learning_rate": 9.653090560613321e-05, "loss": 1.1404, "step": 3285 }, { "epoch": 0.22264381055627075, "grad_norm": 8.79491901397705, "learning_rate": 9.65295365870354e-05, "loss": 1.4334, "step": 3286 }, { "epoch": 0.22271156582424284, "grad_norm": 9.580904960632324, "learning_rate": 9.652816756793757e-05, "loss": 1.0693, "step": 3287 }, { "epoch": 0.22277932109221493, "grad_norm": 7.335433006286621, "learning_rate": 9.652679854883976e-05, "loss": 1.0559, "step": 3288 }, { "epoch": 0.222847076360187, "grad_norm": 8.534089088439941, "learning_rate": 9.652542952974194e-05, "loss": 0.9835, "step": 3289 }, { "epoch": 0.2229148316281591, "grad_norm": 6.626898288726807, "learning_rate": 9.652406051064413e-05, "loss": 0.8056, "step": 3290 }, { "epoch": 0.22298258689613118, "grad_norm": 7.774187088012695, "learning_rate": 9.652269149154631e-05, "loss": 0.8032, "step": 3291 }, { "epoch": 0.22305034216410327, "grad_norm": 8.029982566833496, "learning_rate": 9.652132247244849e-05, "loss": 1.0536, "step": 3292 }, { "epoch": 0.22311809743207534, "grad_norm": 8.701712608337402, "learning_rate": 9.651995345335067e-05, "loss": 0.9422, "step": 3293 }, { "epoch": 0.22318585270004743, "grad_norm": 9.17507553100586, "learning_rate": 9.651858443425286e-05, "loss": 1.0094, "step": 3294 }, { "epoch": 0.22325360796801952, "grad_norm": 8.464656829833984, "learning_rate": 9.651721541515504e-05, "loss": 0.9546, "step": 3295 }, { "epoch": 0.2233213632359916, "grad_norm": 6.839638710021973, "learning_rate": 9.651584639605722e-05, "loss": 0.8296, "step": 3296 }, { "epoch": 0.22338911850396367, "grad_norm": 7.773430824279785, "learning_rate": 9.65144773769594e-05, "loss": 1.1344, "step": 3297 }, { "epoch": 0.22345687377193577, "grad_norm": 9.725523948669434, "learning_rate": 9.65131083578616e-05, "loss": 0.9814, "step": 3298 }, { "epoch": 0.22352462903990786, "grad_norm": 8.995366096496582, "learning_rate": 9.651173933876378e-05, "loss": 1.2263, "step": 3299 }, { "epoch": 0.22359238430787995, "grad_norm": 8.735838890075684, "learning_rate": 9.651037031966596e-05, "loss": 1.1707, "step": 3300 }, { "epoch": 0.223660139575852, "grad_norm": 7.177740097045898, "learning_rate": 9.650900130056815e-05, "loss": 0.9491, "step": 3301 }, { "epoch": 0.2237278948438241, "grad_norm": 8.393296241760254, "learning_rate": 9.650763228147033e-05, "loss": 0.9765, "step": 3302 }, { "epoch": 0.2237956501117962, "grad_norm": 9.773447036743164, "learning_rate": 9.650626326237251e-05, "loss": 1.0409, "step": 3303 }, { "epoch": 0.2238634053797683, "grad_norm": 6.951483249664307, "learning_rate": 9.650489424327471e-05, "loss": 0.9848, "step": 3304 }, { "epoch": 0.22393116064774035, "grad_norm": 8.579717636108398, "learning_rate": 9.650352522417689e-05, "loss": 0.9722, "step": 3305 }, { "epoch": 0.22399891591571244, "grad_norm": 7.798871040344238, "learning_rate": 9.650215620507907e-05, "loss": 0.8046, "step": 3306 }, { "epoch": 0.22406667118368453, "grad_norm": 9.504851341247559, "learning_rate": 9.650078718598125e-05, "loss": 1.1972, "step": 3307 }, { "epoch": 0.22413442645165663, "grad_norm": 8.384056091308594, "learning_rate": 9.649941816688344e-05, "loss": 1.0654, "step": 3308 }, { "epoch": 0.2242021817196287, "grad_norm": 10.469207763671875, "learning_rate": 9.649804914778562e-05, "loss": 1.0694, "step": 3309 }, { "epoch": 0.22426993698760078, "grad_norm": 8.5567045211792, "learning_rate": 9.64966801286878e-05, "loss": 1.0371, "step": 3310 }, { "epoch": 0.22433769225557287, "grad_norm": 9.427155494689941, "learning_rate": 9.649531110958998e-05, "loss": 1.0185, "step": 3311 }, { "epoch": 0.22440544752354497, "grad_norm": 8.238412857055664, "learning_rate": 9.649394209049216e-05, "loss": 0.9744, "step": 3312 }, { "epoch": 0.22447320279151703, "grad_norm": 9.36423110961914, "learning_rate": 9.649257307139436e-05, "loss": 0.9184, "step": 3313 }, { "epoch": 0.22454095805948912, "grad_norm": 7.485356330871582, "learning_rate": 9.649120405229654e-05, "loss": 0.9948, "step": 3314 }, { "epoch": 0.2246087133274612, "grad_norm": 7.247208118438721, "learning_rate": 9.648983503319872e-05, "loss": 0.8592, "step": 3315 }, { "epoch": 0.2246764685954333, "grad_norm": 9.266085624694824, "learning_rate": 9.64884660141009e-05, "loss": 1.0314, "step": 3316 }, { "epoch": 0.22474422386340537, "grad_norm": 11.147274017333984, "learning_rate": 9.64870969950031e-05, "loss": 1.0403, "step": 3317 }, { "epoch": 0.22481197913137746, "grad_norm": 10.961670875549316, "learning_rate": 9.648572797590527e-05, "loss": 1.3147, "step": 3318 }, { "epoch": 0.22487973439934955, "grad_norm": 7.002868175506592, "learning_rate": 9.648435895680745e-05, "loss": 0.9044, "step": 3319 }, { "epoch": 0.22494748966732164, "grad_norm": 7.886247634887695, "learning_rate": 9.648298993770963e-05, "loss": 0.8901, "step": 3320 }, { "epoch": 0.2250152449352937, "grad_norm": 8.933539390563965, "learning_rate": 9.648162091861181e-05, "loss": 0.8758, "step": 3321 }, { "epoch": 0.2250830002032658, "grad_norm": 7.223681449890137, "learning_rate": 9.648025189951401e-05, "loss": 1.1502, "step": 3322 }, { "epoch": 0.2251507554712379, "grad_norm": 8.808623313903809, "learning_rate": 9.647888288041619e-05, "loss": 1.0706, "step": 3323 }, { "epoch": 0.22521851073920998, "grad_norm": 9.557942390441895, "learning_rate": 9.647751386131837e-05, "loss": 0.8113, "step": 3324 }, { "epoch": 0.22528626600718205, "grad_norm": 9.855717658996582, "learning_rate": 9.647614484222055e-05, "loss": 1.251, "step": 3325 }, { "epoch": 0.22535402127515414, "grad_norm": 9.588946342468262, "learning_rate": 9.647477582312273e-05, "loss": 1.3029, "step": 3326 }, { "epoch": 0.22542177654312623, "grad_norm": 9.277341842651367, "learning_rate": 9.647340680402492e-05, "loss": 0.9708, "step": 3327 }, { "epoch": 0.22548953181109832, "grad_norm": 9.607316970825195, "learning_rate": 9.64720377849271e-05, "loss": 0.9614, "step": 3328 }, { "epoch": 0.22555728707907038, "grad_norm": 7.949220657348633, "learning_rate": 9.647066876582928e-05, "loss": 1.0066, "step": 3329 }, { "epoch": 0.22562504234704248, "grad_norm": 8.383744239807129, "learning_rate": 9.646929974673146e-05, "loss": 0.797, "step": 3330 }, { "epoch": 0.22569279761501457, "grad_norm": 7.203914642333984, "learning_rate": 9.646793072763366e-05, "loss": 0.9625, "step": 3331 }, { "epoch": 0.22576055288298666, "grad_norm": 8.95102596282959, "learning_rate": 9.646656170853584e-05, "loss": 1.0728, "step": 3332 }, { "epoch": 0.22582830815095875, "grad_norm": 8.044751167297363, "learning_rate": 9.646519268943802e-05, "loss": 1.2155, "step": 3333 }, { "epoch": 0.22589606341893081, "grad_norm": 9.164902687072754, "learning_rate": 9.64638236703402e-05, "loss": 1.2923, "step": 3334 }, { "epoch": 0.2259638186869029, "grad_norm": 6.792164325714111, "learning_rate": 9.646245465124238e-05, "loss": 0.9956, "step": 3335 }, { "epoch": 0.226031573954875, "grad_norm": 7.76467752456665, "learning_rate": 9.646108563214457e-05, "loss": 1.0057, "step": 3336 }, { "epoch": 0.2260993292228471, "grad_norm": 8.541545867919922, "learning_rate": 9.645971661304675e-05, "loss": 1.1013, "step": 3337 }, { "epoch": 0.22616708449081915, "grad_norm": 7.8545050621032715, "learning_rate": 9.645834759394893e-05, "loss": 0.8565, "step": 3338 }, { "epoch": 0.22623483975879125, "grad_norm": 9.322896003723145, "learning_rate": 9.645697857485112e-05, "loss": 1.0509, "step": 3339 }, { "epoch": 0.22630259502676334, "grad_norm": 9.540191650390625, "learning_rate": 9.645560955575331e-05, "loss": 1.1757, "step": 3340 }, { "epoch": 0.22637035029473543, "grad_norm": 8.97028923034668, "learning_rate": 9.645424053665549e-05, "loss": 1.0342, "step": 3341 }, { "epoch": 0.2264381055627075, "grad_norm": 6.444105625152588, "learning_rate": 9.645287151755767e-05, "loss": 0.8377, "step": 3342 }, { "epoch": 0.22650586083067958, "grad_norm": 9.237889289855957, "learning_rate": 9.645150249845985e-05, "loss": 1.0468, "step": 3343 }, { "epoch": 0.22657361609865168, "grad_norm": 9.567046165466309, "learning_rate": 9.645013347936203e-05, "loss": 1.3875, "step": 3344 }, { "epoch": 0.22664137136662377, "grad_norm": 8.302481651306152, "learning_rate": 9.644876446026422e-05, "loss": 1.0455, "step": 3345 }, { "epoch": 0.22670912663459583, "grad_norm": 9.124307632446289, "learning_rate": 9.64473954411664e-05, "loss": 0.9828, "step": 3346 }, { "epoch": 0.22677688190256792, "grad_norm": 7.700011730194092, "learning_rate": 9.644602642206858e-05, "loss": 0.8475, "step": 3347 }, { "epoch": 0.22684463717054001, "grad_norm": 6.1064276695251465, "learning_rate": 9.644465740297078e-05, "loss": 0.7779, "step": 3348 }, { "epoch": 0.2269123924385121, "grad_norm": 8.924312591552734, "learning_rate": 9.644328838387296e-05, "loss": 1.3825, "step": 3349 }, { "epoch": 0.22698014770648417, "grad_norm": 8.169050216674805, "learning_rate": 9.644191936477514e-05, "loss": 1.1198, "step": 3350 }, { "epoch": 0.22704790297445626, "grad_norm": 8.368382453918457, "learning_rate": 9.644055034567733e-05, "loss": 1.0729, "step": 3351 }, { "epoch": 0.22711565824242835, "grad_norm": 10.206897735595703, "learning_rate": 9.643918132657951e-05, "loss": 0.9022, "step": 3352 }, { "epoch": 0.22718341351040044, "grad_norm": 6.288288116455078, "learning_rate": 9.64378123074817e-05, "loss": 0.7868, "step": 3353 }, { "epoch": 0.2272511687783725, "grad_norm": 8.446578025817871, "learning_rate": 9.643644328838389e-05, "loss": 1.0426, "step": 3354 }, { "epoch": 0.2273189240463446, "grad_norm": 8.425249099731445, "learning_rate": 9.643507426928607e-05, "loss": 0.823, "step": 3355 }, { "epoch": 0.2273866793143167, "grad_norm": 6.610576629638672, "learning_rate": 9.643370525018825e-05, "loss": 0.9251, "step": 3356 }, { "epoch": 0.22745443458228878, "grad_norm": 8.365503311157227, "learning_rate": 9.643233623109043e-05, "loss": 1.0258, "step": 3357 }, { "epoch": 0.22752218985026085, "grad_norm": 8.654791831970215, "learning_rate": 9.643096721199261e-05, "loss": 0.8781, "step": 3358 }, { "epoch": 0.22758994511823294, "grad_norm": 6.612964153289795, "learning_rate": 9.64295981928948e-05, "loss": 0.7472, "step": 3359 }, { "epoch": 0.22765770038620503, "grad_norm": 9.4874267578125, "learning_rate": 9.642822917379698e-05, "loss": 0.994, "step": 3360 }, { "epoch": 0.22772545565417712, "grad_norm": 9.7284574508667, "learning_rate": 9.642686015469916e-05, "loss": 0.9293, "step": 3361 }, { "epoch": 0.2277932109221492, "grad_norm": 8.787792205810547, "learning_rate": 9.642549113560134e-05, "loss": 1.1368, "step": 3362 }, { "epoch": 0.22786096619012128, "grad_norm": 9.218424797058105, "learning_rate": 9.642412211650354e-05, "loss": 1.2437, "step": 3363 }, { "epoch": 0.22792872145809337, "grad_norm": 7.9437127113342285, "learning_rate": 9.642275309740572e-05, "loss": 0.925, "step": 3364 }, { "epoch": 0.22799647672606546, "grad_norm": 10.552806854248047, "learning_rate": 9.64213840783079e-05, "loss": 0.936, "step": 3365 }, { "epoch": 0.22806423199403753, "grad_norm": 6.572626113891602, "learning_rate": 9.642001505921008e-05, "loss": 0.8968, "step": 3366 }, { "epoch": 0.22813198726200962, "grad_norm": 8.689478874206543, "learning_rate": 9.641864604011226e-05, "loss": 0.8605, "step": 3367 }, { "epoch": 0.2281997425299817, "grad_norm": 10.450199127197266, "learning_rate": 9.641727702101445e-05, "loss": 1.0871, "step": 3368 }, { "epoch": 0.2282674977979538, "grad_norm": 9.260807991027832, "learning_rate": 9.641590800191663e-05, "loss": 0.8469, "step": 3369 }, { "epoch": 0.22833525306592586, "grad_norm": 13.626687049865723, "learning_rate": 9.641453898281881e-05, "loss": 1.0615, "step": 3370 }, { "epoch": 0.22840300833389796, "grad_norm": 7.419554710388184, "learning_rate": 9.6413169963721e-05, "loss": 0.9351, "step": 3371 }, { "epoch": 0.22847076360187005, "grad_norm": 9.33206558227539, "learning_rate": 9.641180094462319e-05, "loss": 1.1826, "step": 3372 }, { "epoch": 0.22853851886984214, "grad_norm": 7.148665904998779, "learning_rate": 9.641043192552537e-05, "loss": 1.1121, "step": 3373 }, { "epoch": 0.2286062741378142, "grad_norm": 6.784035682678223, "learning_rate": 9.640906290642755e-05, "loss": 0.9649, "step": 3374 }, { "epoch": 0.2286740294057863, "grad_norm": 7.357334613800049, "learning_rate": 9.640769388732973e-05, "loss": 0.9005, "step": 3375 }, { "epoch": 0.22874178467375839, "grad_norm": 9.157288551330566, "learning_rate": 9.640632486823191e-05, "loss": 0.9585, "step": 3376 }, { "epoch": 0.22880953994173048, "grad_norm": 9.068450927734375, "learning_rate": 9.64049558491341e-05, "loss": 0.9248, "step": 3377 }, { "epoch": 0.22887729520970254, "grad_norm": 9.446860313415527, "learning_rate": 9.640358683003628e-05, "loss": 1.0016, "step": 3378 }, { "epoch": 0.22894505047767463, "grad_norm": 7.63693904876709, "learning_rate": 9.640221781093846e-05, "loss": 1.0158, "step": 3379 }, { "epoch": 0.22901280574564672, "grad_norm": 7.569469451904297, "learning_rate": 9.640084879184064e-05, "loss": 0.9149, "step": 3380 }, { "epoch": 0.22908056101361882, "grad_norm": 8.837080955505371, "learning_rate": 9.639947977274282e-05, "loss": 0.9379, "step": 3381 }, { "epoch": 0.22914831628159088, "grad_norm": 9.141901969909668, "learning_rate": 9.639811075364502e-05, "loss": 0.9121, "step": 3382 }, { "epoch": 0.22921607154956297, "grad_norm": 7.68120813369751, "learning_rate": 9.63967417345472e-05, "loss": 1.0235, "step": 3383 }, { "epoch": 0.22928382681753506, "grad_norm": 8.82229995727539, "learning_rate": 9.639537271544938e-05, "loss": 1.1573, "step": 3384 }, { "epoch": 0.22935158208550716, "grad_norm": 9.490239143371582, "learning_rate": 9.639400369635156e-05, "loss": 0.8776, "step": 3385 }, { "epoch": 0.22941933735347925, "grad_norm": 10.257567405700684, "learning_rate": 9.639263467725375e-05, "loss": 1.0696, "step": 3386 }, { "epoch": 0.2294870926214513, "grad_norm": 7.793918609619141, "learning_rate": 9.639126565815593e-05, "loss": 1.0892, "step": 3387 }, { "epoch": 0.2295548478894234, "grad_norm": 8.983718872070312, "learning_rate": 9.638989663905811e-05, "loss": 0.908, "step": 3388 }, { "epoch": 0.2296226031573955, "grad_norm": 8.31851577758789, "learning_rate": 9.63885276199603e-05, "loss": 1.0168, "step": 3389 }, { "epoch": 0.22969035842536759, "grad_norm": 9.343503952026367, "learning_rate": 9.638715860086248e-05, "loss": 1.1174, "step": 3390 }, { "epoch": 0.22975811369333965, "grad_norm": 7.617075443267822, "learning_rate": 9.638578958176467e-05, "loss": 0.9076, "step": 3391 }, { "epoch": 0.22982586896131174, "grad_norm": 8.102355003356934, "learning_rate": 9.638442056266685e-05, "loss": 1.0476, "step": 3392 }, { "epoch": 0.22989362422928383, "grad_norm": 9.366684913635254, "learning_rate": 9.638305154356903e-05, "loss": 1.1892, "step": 3393 }, { "epoch": 0.22996137949725592, "grad_norm": 8.204293251037598, "learning_rate": 9.638168252447122e-05, "loss": 0.7731, "step": 3394 }, { "epoch": 0.230029134765228, "grad_norm": 9.68787670135498, "learning_rate": 9.63803135053734e-05, "loss": 1.1126, "step": 3395 }, { "epoch": 0.23009689003320008, "grad_norm": 6.6209330558776855, "learning_rate": 9.637894448627558e-05, "loss": 0.9006, "step": 3396 }, { "epoch": 0.23016464530117217, "grad_norm": 7.963149547576904, "learning_rate": 9.637757546717778e-05, "loss": 0.8275, "step": 3397 }, { "epoch": 0.23023240056914426, "grad_norm": 8.74716567993164, "learning_rate": 9.637620644807996e-05, "loss": 1.0592, "step": 3398 }, { "epoch": 0.23030015583711633, "grad_norm": 8.819621086120605, "learning_rate": 9.637483742898214e-05, "loss": 0.8799, "step": 3399 }, { "epoch": 0.23036791110508842, "grad_norm": 8.538536071777344, "learning_rate": 9.637346840988433e-05, "loss": 1.1498, "step": 3400 }, { "epoch": 0.2304356663730605, "grad_norm": 8.787203788757324, "learning_rate": 9.637209939078651e-05, "loss": 0.8149, "step": 3401 }, { "epoch": 0.2305034216410326, "grad_norm": 8.85105037689209, "learning_rate": 9.63707303716887e-05, "loss": 0.8525, "step": 3402 }, { "epoch": 0.23057117690900467, "grad_norm": 9.773297309875488, "learning_rate": 9.636936135259087e-05, "loss": 0.9448, "step": 3403 }, { "epoch": 0.23063893217697676, "grad_norm": 7.804196357727051, "learning_rate": 9.636799233349307e-05, "loss": 0.9549, "step": 3404 }, { "epoch": 0.23070668744494885, "grad_norm": 8.398775100708008, "learning_rate": 9.636662331439525e-05, "loss": 0.9111, "step": 3405 }, { "epoch": 0.23077444271292094, "grad_norm": 8.960867881774902, "learning_rate": 9.636525429529743e-05, "loss": 1.1533, "step": 3406 }, { "epoch": 0.230842197980893, "grad_norm": 7.513526439666748, "learning_rate": 9.636388527619961e-05, "loss": 1.113, "step": 3407 }, { "epoch": 0.2309099532488651, "grad_norm": 8.891728401184082, "learning_rate": 9.636251625710179e-05, "loss": 0.8727, "step": 3408 }, { "epoch": 0.2309777085168372, "grad_norm": 6.972326755523682, "learning_rate": 9.636114723800398e-05, "loss": 0.8085, "step": 3409 }, { "epoch": 0.23104546378480928, "grad_norm": 9.195303916931152, "learning_rate": 9.635977821890616e-05, "loss": 1.0276, "step": 3410 }, { "epoch": 0.23111321905278134, "grad_norm": 9.36483383178711, "learning_rate": 9.635840919980834e-05, "loss": 1.1369, "step": 3411 }, { "epoch": 0.23118097432075344, "grad_norm": 7.282619953155518, "learning_rate": 9.635704018071052e-05, "loss": 0.9214, "step": 3412 }, { "epoch": 0.23124872958872553, "grad_norm": 6.289726257324219, "learning_rate": 9.63556711616127e-05, "loss": 0.8556, "step": 3413 }, { "epoch": 0.23131648485669762, "grad_norm": 8.646320343017578, "learning_rate": 9.63543021425149e-05, "loss": 1.0833, "step": 3414 }, { "epoch": 0.23138424012466968, "grad_norm": 8.372559547424316, "learning_rate": 9.635293312341708e-05, "loss": 0.9944, "step": 3415 }, { "epoch": 0.23145199539264177, "grad_norm": 8.002387046813965, "learning_rate": 9.635156410431926e-05, "loss": 1.015, "step": 3416 }, { "epoch": 0.23151975066061387, "grad_norm": 8.4190092086792, "learning_rate": 9.635019508522144e-05, "loss": 0.9766, "step": 3417 }, { "epoch": 0.23158750592858596, "grad_norm": 8.759613037109375, "learning_rate": 9.634882606612363e-05, "loss": 1.2348, "step": 3418 }, { "epoch": 0.23165526119655802, "grad_norm": 8.439151763916016, "learning_rate": 9.634745704702581e-05, "loss": 0.8778, "step": 3419 }, { "epoch": 0.2317230164645301, "grad_norm": 8.05675220489502, "learning_rate": 9.6346088027928e-05, "loss": 0.9525, "step": 3420 }, { "epoch": 0.2317907717325022, "grad_norm": 9.196937561035156, "learning_rate": 9.634471900883017e-05, "loss": 1.0663, "step": 3421 }, { "epoch": 0.2318585270004743, "grad_norm": 7.622048854827881, "learning_rate": 9.634334998973235e-05, "loss": 0.9522, "step": 3422 }, { "epoch": 0.23192628226844636, "grad_norm": 10.235919952392578, "learning_rate": 9.634198097063455e-05, "loss": 1.1084, "step": 3423 }, { "epoch": 0.23199403753641845, "grad_norm": 7.920719623565674, "learning_rate": 9.634061195153673e-05, "loss": 0.9443, "step": 3424 }, { "epoch": 0.23206179280439054, "grad_norm": 7.440132141113281, "learning_rate": 9.633924293243891e-05, "loss": 1.0757, "step": 3425 }, { "epoch": 0.23212954807236263, "grad_norm": 7.0108771324157715, "learning_rate": 9.633787391334109e-05, "loss": 0.8043, "step": 3426 }, { "epoch": 0.2321973033403347, "grad_norm": 6.620635032653809, "learning_rate": 9.633650489424328e-05, "loss": 0.999, "step": 3427 }, { "epoch": 0.2322650586083068, "grad_norm": 7.27518367767334, "learning_rate": 9.633513587514546e-05, "loss": 0.888, "step": 3428 }, { "epoch": 0.23233281387627888, "grad_norm": 7.081945419311523, "learning_rate": 9.633376685604764e-05, "loss": 0.8837, "step": 3429 }, { "epoch": 0.23240056914425097, "grad_norm": 7.225597858428955, "learning_rate": 9.633239783694982e-05, "loss": 0.7774, "step": 3430 }, { "epoch": 0.23246832441222304, "grad_norm": 7.0490217208862305, "learning_rate": 9.6331028817852e-05, "loss": 1.0654, "step": 3431 }, { "epoch": 0.23253607968019513, "grad_norm": 8.641639709472656, "learning_rate": 9.63296597987542e-05, "loss": 1.1071, "step": 3432 }, { "epoch": 0.23260383494816722, "grad_norm": 7.442448139190674, "learning_rate": 9.632829077965638e-05, "loss": 0.9392, "step": 3433 }, { "epoch": 0.2326715902161393, "grad_norm": 6.6959452629089355, "learning_rate": 9.632692176055856e-05, "loss": 0.8334, "step": 3434 }, { "epoch": 0.23273934548411138, "grad_norm": 9.67419719696045, "learning_rate": 9.632555274146074e-05, "loss": 1.1948, "step": 3435 }, { "epoch": 0.23280710075208347, "grad_norm": 7.384359836578369, "learning_rate": 9.632418372236292e-05, "loss": 0.7992, "step": 3436 }, { "epoch": 0.23287485602005556, "grad_norm": 6.566294193267822, "learning_rate": 9.632281470326511e-05, "loss": 0.9393, "step": 3437 }, { "epoch": 0.23294261128802765, "grad_norm": 7.842831134796143, "learning_rate": 9.63214456841673e-05, "loss": 1.1372, "step": 3438 }, { "epoch": 0.23301036655599974, "grad_norm": 9.487961769104004, "learning_rate": 9.632007666506947e-05, "loss": 1.0381, "step": 3439 }, { "epoch": 0.2330781218239718, "grad_norm": 8.06712532043457, "learning_rate": 9.631870764597167e-05, "loss": 1.4106, "step": 3440 }, { "epoch": 0.2331458770919439, "grad_norm": 7.026076793670654, "learning_rate": 9.631733862687385e-05, "loss": 1.0662, "step": 3441 }, { "epoch": 0.233213632359916, "grad_norm": 10.926567077636719, "learning_rate": 9.631596960777603e-05, "loss": 1.2932, "step": 3442 }, { "epoch": 0.23328138762788808, "grad_norm": 8.380082130432129, "learning_rate": 9.631460058867822e-05, "loss": 1.0784, "step": 3443 }, { "epoch": 0.23334914289586015, "grad_norm": 8.768819808959961, "learning_rate": 9.63132315695804e-05, "loss": 1.1501, "step": 3444 }, { "epoch": 0.23341689816383224, "grad_norm": 8.467490196228027, "learning_rate": 9.631186255048258e-05, "loss": 0.7555, "step": 3445 }, { "epoch": 0.23348465343180433, "grad_norm": 6.0409770011901855, "learning_rate": 9.631049353138478e-05, "loss": 0.9099, "step": 3446 }, { "epoch": 0.23355240869977642, "grad_norm": 7.160163402557373, "learning_rate": 9.630912451228696e-05, "loss": 0.8386, "step": 3447 }, { "epoch": 0.23362016396774848, "grad_norm": 9.102558135986328, "learning_rate": 9.630775549318914e-05, "loss": 0.9259, "step": 3448 }, { "epoch": 0.23368791923572058, "grad_norm": 7.071728706359863, "learning_rate": 9.630638647409132e-05, "loss": 1.0426, "step": 3449 }, { "epoch": 0.23375567450369267, "grad_norm": 8.358567237854004, "learning_rate": 9.630501745499351e-05, "loss": 0.9806, "step": 3450 }, { "epoch": 0.23382342977166476, "grad_norm": 12.588993072509766, "learning_rate": 9.630364843589569e-05, "loss": 1.0796, "step": 3451 }, { "epoch": 0.23389118503963682, "grad_norm": 8.665871620178223, "learning_rate": 9.630227941679787e-05, "loss": 1.0224, "step": 3452 }, { "epoch": 0.23395894030760891, "grad_norm": 9.288796424865723, "learning_rate": 9.630091039770005e-05, "loss": 0.9329, "step": 3453 }, { "epoch": 0.234026695575581, "grad_norm": 7.319186210632324, "learning_rate": 9.629954137860223e-05, "loss": 1.057, "step": 3454 }, { "epoch": 0.2340944508435531, "grad_norm": 8.640669822692871, "learning_rate": 9.629817235950443e-05, "loss": 0.9756, "step": 3455 }, { "epoch": 0.23416220611152516, "grad_norm": 10.73513126373291, "learning_rate": 9.629680334040661e-05, "loss": 0.958, "step": 3456 }, { "epoch": 0.23422996137949725, "grad_norm": 7.40097713470459, "learning_rate": 9.629543432130879e-05, "loss": 0.9062, "step": 3457 }, { "epoch": 0.23429771664746935, "grad_norm": 9.506468772888184, "learning_rate": 9.629406530221097e-05, "loss": 1.0083, "step": 3458 }, { "epoch": 0.23436547191544144, "grad_norm": 7.291172504425049, "learning_rate": 9.629269628311315e-05, "loss": 0.9966, "step": 3459 }, { "epoch": 0.2344332271834135, "grad_norm": 8.962395668029785, "learning_rate": 9.629132726401534e-05, "loss": 0.8504, "step": 3460 }, { "epoch": 0.2345009824513856, "grad_norm": 10.29174518585205, "learning_rate": 9.628995824491752e-05, "loss": 1.0807, "step": 3461 }, { "epoch": 0.23456873771935768, "grad_norm": 9.461160659790039, "learning_rate": 9.62885892258197e-05, "loss": 0.996, "step": 3462 }, { "epoch": 0.23463649298732978, "grad_norm": 7.13394021987915, "learning_rate": 9.628722020672188e-05, "loss": 0.7163, "step": 3463 }, { "epoch": 0.23470424825530184, "grad_norm": 9.379724502563477, "learning_rate": 9.628585118762408e-05, "loss": 0.9571, "step": 3464 }, { "epoch": 0.23477200352327393, "grad_norm": 9.72339916229248, "learning_rate": 9.628448216852626e-05, "loss": 1.2559, "step": 3465 }, { "epoch": 0.23483975879124602, "grad_norm": 7.353830337524414, "learning_rate": 9.628311314942844e-05, "loss": 1.0563, "step": 3466 }, { "epoch": 0.23490751405921811, "grad_norm": 8.423667907714844, "learning_rate": 9.628174413033062e-05, "loss": 1.0799, "step": 3467 }, { "epoch": 0.23497526932719018, "grad_norm": 6.426609516143799, "learning_rate": 9.62803751112328e-05, "loss": 0.9417, "step": 3468 }, { "epoch": 0.23504302459516227, "grad_norm": 10.882522583007812, "learning_rate": 9.627900609213499e-05, "loss": 1.1498, "step": 3469 }, { "epoch": 0.23511077986313436, "grad_norm": 7.768298625946045, "learning_rate": 9.627763707303717e-05, "loss": 0.7389, "step": 3470 }, { "epoch": 0.23517853513110645, "grad_norm": 9.002137184143066, "learning_rate": 9.627626805393935e-05, "loss": 1.2338, "step": 3471 }, { "epoch": 0.23524629039907852, "grad_norm": 8.671159744262695, "learning_rate": 9.627489903484153e-05, "loss": 1.1114, "step": 3472 }, { "epoch": 0.2353140456670506, "grad_norm": 5.836034774780273, "learning_rate": 9.627353001574373e-05, "loss": 0.7075, "step": 3473 }, { "epoch": 0.2353818009350227, "grad_norm": 8.145447731018066, "learning_rate": 9.627216099664591e-05, "loss": 0.9504, "step": 3474 }, { "epoch": 0.2354495562029948, "grad_norm": 7.3333563804626465, "learning_rate": 9.627079197754809e-05, "loss": 0.8443, "step": 3475 }, { "epoch": 0.23551731147096686, "grad_norm": 8.191438674926758, "learning_rate": 9.626942295845027e-05, "loss": 0.8163, "step": 3476 }, { "epoch": 0.23558506673893895, "grad_norm": 10.042975425720215, "learning_rate": 9.626805393935245e-05, "loss": 1.0715, "step": 3477 }, { "epoch": 0.23565282200691104, "grad_norm": 7.854464054107666, "learning_rate": 9.626668492025464e-05, "loss": 0.9263, "step": 3478 }, { "epoch": 0.23572057727488313, "grad_norm": 9.286520004272461, "learning_rate": 9.626531590115682e-05, "loss": 1.0172, "step": 3479 }, { "epoch": 0.2357883325428552, "grad_norm": 7.321778297424316, "learning_rate": 9.6263946882059e-05, "loss": 1.0661, "step": 3480 }, { "epoch": 0.2358560878108273, "grad_norm": 7.432394981384277, "learning_rate": 9.626257786296118e-05, "loss": 1.1682, "step": 3481 }, { "epoch": 0.23592384307879938, "grad_norm": 10.230997085571289, "learning_rate": 9.626120884386338e-05, "loss": 1.0498, "step": 3482 }, { "epoch": 0.23599159834677147, "grad_norm": 8.530696868896484, "learning_rate": 9.625983982476556e-05, "loss": 1.0598, "step": 3483 }, { "epoch": 0.23605935361474353, "grad_norm": 7.35584020614624, "learning_rate": 9.625847080566774e-05, "loss": 1.0234, "step": 3484 }, { "epoch": 0.23612710888271563, "grad_norm": 8.030210494995117, "learning_rate": 9.625710178656992e-05, "loss": 1.0241, "step": 3485 }, { "epoch": 0.23619486415068772, "grad_norm": 7.508440971374512, "learning_rate": 9.625573276747211e-05, "loss": 0.9466, "step": 3486 }, { "epoch": 0.2362626194186598, "grad_norm": 7.4618401527404785, "learning_rate": 9.62543637483743e-05, "loss": 0.88, "step": 3487 }, { "epoch": 0.23633037468663187, "grad_norm": 7.738553524017334, "learning_rate": 9.625299472927647e-05, "loss": 1.0146, "step": 3488 }, { "epoch": 0.23639812995460396, "grad_norm": 6.782055854797363, "learning_rate": 9.625162571017867e-05, "loss": 0.8196, "step": 3489 }, { "epoch": 0.23646588522257606, "grad_norm": 7.495883464813232, "learning_rate": 9.625025669108085e-05, "loss": 1.0025, "step": 3490 }, { "epoch": 0.23653364049054815, "grad_norm": 7.224970817565918, "learning_rate": 9.624888767198303e-05, "loss": 1.0069, "step": 3491 }, { "epoch": 0.23660139575852024, "grad_norm": 8.686829566955566, "learning_rate": 9.624751865288522e-05, "loss": 0.8927, "step": 3492 }, { "epoch": 0.2366691510264923, "grad_norm": 10.551370620727539, "learning_rate": 9.62461496337874e-05, "loss": 1.0973, "step": 3493 }, { "epoch": 0.2367369062944644, "grad_norm": 8.38442611694336, "learning_rate": 9.624478061468958e-05, "loss": 0.8145, "step": 3494 }, { "epoch": 0.23680466156243649, "grad_norm": 7.099575996398926, "learning_rate": 9.624341159559176e-05, "loss": 0.8304, "step": 3495 }, { "epoch": 0.23687241683040858, "grad_norm": 7.905203342437744, "learning_rate": 9.624204257649396e-05, "loss": 1.1708, "step": 3496 }, { "epoch": 0.23694017209838064, "grad_norm": 8.08218765258789, "learning_rate": 9.624067355739614e-05, "loss": 0.8367, "step": 3497 }, { "epoch": 0.23700792736635273, "grad_norm": 7.766988277435303, "learning_rate": 9.623930453829832e-05, "loss": 1.1126, "step": 3498 }, { "epoch": 0.23707568263432482, "grad_norm": 8.3229398727417, "learning_rate": 9.62379355192005e-05, "loss": 0.9798, "step": 3499 }, { "epoch": 0.23714343790229692, "grad_norm": 7.113401889801025, "learning_rate": 9.623656650010268e-05, "loss": 0.7711, "step": 3500 }, { "epoch": 0.23721119317026898, "grad_norm": 7.286252975463867, "learning_rate": 9.623519748100487e-05, "loss": 0.8018, "step": 3501 }, { "epoch": 0.23727894843824107, "grad_norm": 8.724101066589355, "learning_rate": 9.623382846190705e-05, "loss": 0.9077, "step": 3502 }, { "epoch": 0.23734670370621316, "grad_norm": 9.201606750488281, "learning_rate": 9.623245944280923e-05, "loss": 1.2277, "step": 3503 }, { "epoch": 0.23741445897418526, "grad_norm": 6.827572822570801, "learning_rate": 9.623109042371141e-05, "loss": 0.9262, "step": 3504 }, { "epoch": 0.23748221424215732, "grad_norm": 7.41288423538208, "learning_rate": 9.622972140461361e-05, "loss": 0.762, "step": 3505 }, { "epoch": 0.2375499695101294, "grad_norm": 9.649503707885742, "learning_rate": 9.622835238551579e-05, "loss": 1.0371, "step": 3506 }, { "epoch": 0.2376177247781015, "grad_norm": 9.630754470825195, "learning_rate": 9.622698336641797e-05, "loss": 0.9994, "step": 3507 }, { "epoch": 0.2376854800460736, "grad_norm": 8.949817657470703, "learning_rate": 9.622561434732015e-05, "loss": 0.8555, "step": 3508 }, { "epoch": 0.23775323531404566, "grad_norm": 8.096979141235352, "learning_rate": 9.622424532822233e-05, "loss": 0.9939, "step": 3509 }, { "epoch": 0.23782099058201775, "grad_norm": 8.714512825012207, "learning_rate": 9.622287630912452e-05, "loss": 0.9533, "step": 3510 }, { "epoch": 0.23788874584998984, "grad_norm": 8.663968086242676, "learning_rate": 9.62215072900267e-05, "loss": 0.9974, "step": 3511 }, { "epoch": 0.23795650111796193, "grad_norm": 8.704265594482422, "learning_rate": 9.622013827092888e-05, "loss": 1.2524, "step": 3512 }, { "epoch": 0.238024256385934, "grad_norm": 6.585339069366455, "learning_rate": 9.621876925183106e-05, "loss": 0.9021, "step": 3513 }, { "epoch": 0.2380920116539061, "grad_norm": 7.992496013641357, "learning_rate": 9.621740023273324e-05, "loss": 0.9345, "step": 3514 }, { "epoch": 0.23815976692187818, "grad_norm": 7.076025009155273, "learning_rate": 9.621603121363544e-05, "loss": 1.0487, "step": 3515 }, { "epoch": 0.23822752218985027, "grad_norm": 8.833822250366211, "learning_rate": 9.621466219453762e-05, "loss": 0.9612, "step": 3516 }, { "epoch": 0.23829527745782234, "grad_norm": 7.8553385734558105, "learning_rate": 9.62132931754398e-05, "loss": 1.0452, "step": 3517 }, { "epoch": 0.23836303272579443, "grad_norm": 8.44243049621582, "learning_rate": 9.621192415634198e-05, "loss": 0.6874, "step": 3518 }, { "epoch": 0.23843078799376652, "grad_norm": 8.1638822555542, "learning_rate": 9.621055513724417e-05, "loss": 0.9239, "step": 3519 }, { "epoch": 0.2384985432617386, "grad_norm": 8.315443992614746, "learning_rate": 9.620918611814635e-05, "loss": 1.1256, "step": 3520 }, { "epoch": 0.23856629852971067, "grad_norm": 10.451863288879395, "learning_rate": 9.620781709904853e-05, "loss": 1.1453, "step": 3521 }, { "epoch": 0.23863405379768277, "grad_norm": 9.117147445678711, "learning_rate": 9.620644807995071e-05, "loss": 1.0751, "step": 3522 }, { "epoch": 0.23870180906565486, "grad_norm": 7.746578216552734, "learning_rate": 9.62050790608529e-05, "loss": 0.867, "step": 3523 }, { "epoch": 0.23876956433362695, "grad_norm": 8.366421699523926, "learning_rate": 9.620371004175509e-05, "loss": 1.0863, "step": 3524 }, { "epoch": 0.238837319601599, "grad_norm": 9.951937675476074, "learning_rate": 9.620234102265727e-05, "loss": 1.0431, "step": 3525 }, { "epoch": 0.2389050748695711, "grad_norm": 9.077424049377441, "learning_rate": 9.620097200355945e-05, "loss": 1.0433, "step": 3526 }, { "epoch": 0.2389728301375432, "grad_norm": 6.919139385223389, "learning_rate": 9.619960298446163e-05, "loss": 0.8655, "step": 3527 }, { "epoch": 0.2390405854055153, "grad_norm": 7.729475975036621, "learning_rate": 9.619823396536382e-05, "loss": 1.0943, "step": 3528 }, { "epoch": 0.23910834067348735, "grad_norm": 9.782391548156738, "learning_rate": 9.6196864946266e-05, "loss": 1.1362, "step": 3529 }, { "epoch": 0.23917609594145944, "grad_norm": 8.685064315795898, "learning_rate": 9.619549592716818e-05, "loss": 0.9885, "step": 3530 }, { "epoch": 0.23924385120943154, "grad_norm": 6.963294982910156, "learning_rate": 9.619412690807036e-05, "loss": 0.9061, "step": 3531 }, { "epoch": 0.23931160647740363, "grad_norm": 8.584861755371094, "learning_rate": 9.619275788897256e-05, "loss": 0.8527, "step": 3532 }, { "epoch": 0.2393793617453757, "grad_norm": 9.253739356994629, "learning_rate": 9.619138886987474e-05, "loss": 1.0643, "step": 3533 }, { "epoch": 0.23944711701334778, "grad_norm": 7.659543037414551, "learning_rate": 9.619001985077692e-05, "loss": 0.9553, "step": 3534 }, { "epoch": 0.23951487228131987, "grad_norm": 8.769670486450195, "learning_rate": 9.618865083167911e-05, "loss": 1.0576, "step": 3535 }, { "epoch": 0.23958262754929197, "grad_norm": 7.416141033172607, "learning_rate": 9.618728181258129e-05, "loss": 1.0064, "step": 3536 }, { "epoch": 0.23965038281726403, "grad_norm": 7.91802978515625, "learning_rate": 9.618591279348347e-05, "loss": 1.0095, "step": 3537 }, { "epoch": 0.23971813808523612, "grad_norm": 6.665622234344482, "learning_rate": 9.618454377438567e-05, "loss": 0.8099, "step": 3538 }, { "epoch": 0.2397858933532082, "grad_norm": 7.3240203857421875, "learning_rate": 9.618317475528785e-05, "loss": 0.9496, "step": 3539 }, { "epoch": 0.2398536486211803, "grad_norm": 7.268299102783203, "learning_rate": 9.618180573619003e-05, "loss": 1.0504, "step": 3540 }, { "epoch": 0.23992140388915237, "grad_norm": 8.710535049438477, "learning_rate": 9.618043671709221e-05, "loss": 0.9228, "step": 3541 }, { "epoch": 0.23998915915712446, "grad_norm": 8.5223970413208, "learning_rate": 9.61790676979944e-05, "loss": 1.0383, "step": 3542 }, { "epoch": 0.24005691442509655, "grad_norm": 7.635293960571289, "learning_rate": 9.617769867889658e-05, "loss": 0.99, "step": 3543 }, { "epoch": 0.24012466969306864, "grad_norm": 9.41180419921875, "learning_rate": 9.617632965979876e-05, "loss": 1.1214, "step": 3544 }, { "epoch": 0.24019242496104073, "grad_norm": 9.31615924835205, "learning_rate": 9.617496064070094e-05, "loss": 0.9369, "step": 3545 }, { "epoch": 0.2402601802290128, "grad_norm": 9.427112579345703, "learning_rate": 9.617359162160312e-05, "loss": 1.1829, "step": 3546 }, { "epoch": 0.2403279354969849, "grad_norm": 7.825446605682373, "learning_rate": 9.617222260250532e-05, "loss": 0.9459, "step": 3547 }, { "epoch": 0.24039569076495698, "grad_norm": 7.8191633224487305, "learning_rate": 9.61708535834075e-05, "loss": 1.1144, "step": 3548 }, { "epoch": 0.24046344603292907, "grad_norm": 8.224778175354004, "learning_rate": 9.616948456430968e-05, "loss": 0.8544, "step": 3549 }, { "epoch": 0.24053120130090114, "grad_norm": 9.618694305419922, "learning_rate": 9.616811554521186e-05, "loss": 1.2254, "step": 3550 }, { "epoch": 0.24059895656887323, "grad_norm": 7.786314964294434, "learning_rate": 9.616674652611405e-05, "loss": 0.8867, "step": 3551 }, { "epoch": 0.24066671183684532, "grad_norm": 9.397835731506348, "learning_rate": 9.616537750701623e-05, "loss": 1.0505, "step": 3552 }, { "epoch": 0.2407344671048174, "grad_norm": 7.701049327850342, "learning_rate": 9.616400848791841e-05, "loss": 0.897, "step": 3553 }, { "epoch": 0.24080222237278948, "grad_norm": 7.573019504547119, "learning_rate": 9.616263946882059e-05, "loss": 0.9972, "step": 3554 }, { "epoch": 0.24086997764076157, "grad_norm": 8.337100982666016, "learning_rate": 9.616127044972277e-05, "loss": 1.0835, "step": 3555 }, { "epoch": 0.24093773290873366, "grad_norm": 8.702056884765625, "learning_rate": 9.615990143062497e-05, "loss": 0.7688, "step": 3556 }, { "epoch": 0.24100548817670575, "grad_norm": 8.482346534729004, "learning_rate": 9.615853241152715e-05, "loss": 1.1972, "step": 3557 }, { "epoch": 0.24107324344467781, "grad_norm": 7.193674087524414, "learning_rate": 9.615716339242933e-05, "loss": 1.0389, "step": 3558 }, { "epoch": 0.2411409987126499, "grad_norm": 8.803317070007324, "learning_rate": 9.615579437333151e-05, "loss": 1.0429, "step": 3559 }, { "epoch": 0.241208753980622, "grad_norm": 9.055732727050781, "learning_rate": 9.61544253542337e-05, "loss": 1.1161, "step": 3560 }, { "epoch": 0.2412765092485941, "grad_norm": 10.80001449584961, "learning_rate": 9.615305633513588e-05, "loss": 0.9628, "step": 3561 }, { "epoch": 0.24134426451656615, "grad_norm": 7.707313060760498, "learning_rate": 9.615168731603806e-05, "loss": 0.8806, "step": 3562 }, { "epoch": 0.24141201978453825, "grad_norm": 8.823626518249512, "learning_rate": 9.615031829694024e-05, "loss": 0.8771, "step": 3563 }, { "epoch": 0.24147977505251034, "grad_norm": 7.984725475311279, "learning_rate": 9.614894927784242e-05, "loss": 0.9456, "step": 3564 }, { "epoch": 0.24154753032048243, "grad_norm": 7.335816860198975, "learning_rate": 9.614758025874462e-05, "loss": 1.0323, "step": 3565 }, { "epoch": 0.2416152855884545, "grad_norm": 5.959085941314697, "learning_rate": 9.61462112396468e-05, "loss": 0.8049, "step": 3566 }, { "epoch": 0.24168304085642658, "grad_norm": 8.775632858276367, "learning_rate": 9.614484222054898e-05, "loss": 0.8875, "step": 3567 }, { "epoch": 0.24175079612439868, "grad_norm": 6.589362621307373, "learning_rate": 9.614347320145116e-05, "loss": 1.0551, "step": 3568 }, { "epoch": 0.24181855139237077, "grad_norm": 7.49434232711792, "learning_rate": 9.614210418235334e-05, "loss": 0.7662, "step": 3569 }, { "epoch": 0.24188630666034283, "grad_norm": 7.759862899780273, "learning_rate": 9.614073516325553e-05, "loss": 0.7364, "step": 3570 }, { "epoch": 0.24195406192831492, "grad_norm": 7.9510273933410645, "learning_rate": 9.613936614415771e-05, "loss": 0.9277, "step": 3571 }, { "epoch": 0.24202181719628701, "grad_norm": 8.308568954467773, "learning_rate": 9.61379971250599e-05, "loss": 1.1689, "step": 3572 }, { "epoch": 0.2420895724642591, "grad_norm": 9.44938850402832, "learning_rate": 9.613662810596207e-05, "loss": 1.0937, "step": 3573 }, { "epoch": 0.24215732773223117, "grad_norm": 11.66707706451416, "learning_rate": 9.613525908686427e-05, "loss": 0.77, "step": 3574 }, { "epoch": 0.24222508300020326, "grad_norm": 9.25683879852295, "learning_rate": 9.613389006776645e-05, "loss": 0.9661, "step": 3575 }, { "epoch": 0.24229283826817535, "grad_norm": 7.289797782897949, "learning_rate": 9.613252104866863e-05, "loss": 0.8597, "step": 3576 }, { "epoch": 0.24236059353614745, "grad_norm": 6.669293403625488, "learning_rate": 9.613115202957081e-05, "loss": 0.9134, "step": 3577 }, { "epoch": 0.2424283488041195, "grad_norm": 8.785436630249023, "learning_rate": 9.612978301047299e-05, "loss": 0.7546, "step": 3578 }, { "epoch": 0.2424961040720916, "grad_norm": 7.386310577392578, "learning_rate": 9.612841399137518e-05, "loss": 1.0635, "step": 3579 }, { "epoch": 0.2425638593400637, "grad_norm": 8.23388957977295, "learning_rate": 9.612704497227736e-05, "loss": 0.8598, "step": 3580 }, { "epoch": 0.24263161460803578, "grad_norm": 8.498323440551758, "learning_rate": 9.612567595317954e-05, "loss": 1.083, "step": 3581 }, { "epoch": 0.24269936987600785, "grad_norm": 7.849715709686279, "learning_rate": 9.612430693408174e-05, "loss": 0.7673, "step": 3582 }, { "epoch": 0.24276712514397994, "grad_norm": 8.113242149353027, "learning_rate": 9.612293791498392e-05, "loss": 1.1003, "step": 3583 }, { "epoch": 0.24283488041195203, "grad_norm": 6.983048915863037, "learning_rate": 9.61215688958861e-05, "loss": 0.8668, "step": 3584 }, { "epoch": 0.24290263567992412, "grad_norm": 8.389126777648926, "learning_rate": 9.612019987678829e-05, "loss": 1.0008, "step": 3585 }, { "epoch": 0.2429703909478962, "grad_norm": 7.593414783477783, "learning_rate": 9.611883085769047e-05, "loss": 0.9829, "step": 3586 }, { "epoch": 0.24303814621586828, "grad_norm": 7.836172103881836, "learning_rate": 9.611746183859265e-05, "loss": 0.8128, "step": 3587 }, { "epoch": 0.24310590148384037, "grad_norm": 8.981040954589844, "learning_rate": 9.611609281949485e-05, "loss": 1.1249, "step": 3588 }, { "epoch": 0.24317365675181246, "grad_norm": 6.724935054779053, "learning_rate": 9.611472380039703e-05, "loss": 0.7699, "step": 3589 }, { "epoch": 0.24324141201978453, "grad_norm": 9.456436157226562, "learning_rate": 9.611335478129921e-05, "loss": 1.0085, "step": 3590 }, { "epoch": 0.24330916728775662, "grad_norm": 11.044548988342285, "learning_rate": 9.611198576220139e-05, "loss": 1.0671, "step": 3591 }, { "epoch": 0.2433769225557287, "grad_norm": 10.590521812438965, "learning_rate": 9.611061674310358e-05, "loss": 1.0051, "step": 3592 }, { "epoch": 0.2434446778237008, "grad_norm": 8.595380783081055, "learning_rate": 9.610924772400576e-05, "loss": 0.9655, "step": 3593 }, { "epoch": 0.24351243309167286, "grad_norm": 7.794788837432861, "learning_rate": 9.610787870490794e-05, "loss": 0.9551, "step": 3594 }, { "epoch": 0.24358018835964496, "grad_norm": 7.609074115753174, "learning_rate": 9.610650968581012e-05, "loss": 0.9856, "step": 3595 }, { "epoch": 0.24364794362761705, "grad_norm": 6.909607410430908, "learning_rate": 9.61051406667123e-05, "loss": 1.0195, "step": 3596 }, { "epoch": 0.24371569889558914, "grad_norm": 7.958381652832031, "learning_rate": 9.61037716476145e-05, "loss": 0.8263, "step": 3597 }, { "epoch": 0.24378345416356123, "grad_norm": 7.38173246383667, "learning_rate": 9.610240262851668e-05, "loss": 0.8551, "step": 3598 }, { "epoch": 0.2438512094315333, "grad_norm": 7.1448822021484375, "learning_rate": 9.610103360941886e-05, "loss": 0.8189, "step": 3599 }, { "epoch": 0.2439189646995054, "grad_norm": 7.44658088684082, "learning_rate": 9.609966459032104e-05, "loss": 0.8056, "step": 3600 }, { "epoch": 0.24398671996747748, "grad_norm": 7.480542182922363, "learning_rate": 9.609829557122322e-05, "loss": 0.8591, "step": 3601 }, { "epoch": 0.24405447523544957, "grad_norm": 9.895995140075684, "learning_rate": 9.609692655212541e-05, "loss": 1.1256, "step": 3602 }, { "epoch": 0.24412223050342163, "grad_norm": 7.009078025817871, "learning_rate": 9.609555753302759e-05, "loss": 0.9045, "step": 3603 }, { "epoch": 0.24418998577139372, "grad_norm": 8.714953422546387, "learning_rate": 9.609418851392977e-05, "loss": 1.1207, "step": 3604 }, { "epoch": 0.24425774103936582, "grad_norm": 7.239734172821045, "learning_rate": 9.609281949483195e-05, "loss": 0.7811, "step": 3605 }, { "epoch": 0.2443254963073379, "grad_norm": 10.486507415771484, "learning_rate": 9.609145047573415e-05, "loss": 1.2213, "step": 3606 }, { "epoch": 0.24439325157530997, "grad_norm": 12.031790733337402, "learning_rate": 9.609008145663633e-05, "loss": 1.0316, "step": 3607 }, { "epoch": 0.24446100684328206, "grad_norm": 7.607183456420898, "learning_rate": 9.608871243753851e-05, "loss": 0.7754, "step": 3608 }, { "epoch": 0.24452876211125416, "grad_norm": 9.313577651977539, "learning_rate": 9.608734341844069e-05, "loss": 0.9907, "step": 3609 }, { "epoch": 0.24459651737922625, "grad_norm": 7.672274589538574, "learning_rate": 9.608597439934287e-05, "loss": 0.8428, "step": 3610 }, { "epoch": 0.2446642726471983, "grad_norm": 8.259462356567383, "learning_rate": 9.608460538024506e-05, "loss": 0.8863, "step": 3611 }, { "epoch": 0.2447320279151704, "grad_norm": 8.50256061553955, "learning_rate": 9.608323636114724e-05, "loss": 0.7771, "step": 3612 }, { "epoch": 0.2447997831831425, "grad_norm": 6.8893818855285645, "learning_rate": 9.608186734204942e-05, "loss": 0.9791, "step": 3613 }, { "epoch": 0.24486753845111459, "grad_norm": 8.954825401306152, "learning_rate": 9.60804983229516e-05, "loss": 1.0712, "step": 3614 }, { "epoch": 0.24493529371908665, "grad_norm": 8.83995532989502, "learning_rate": 9.60791293038538e-05, "loss": 1.1982, "step": 3615 }, { "epoch": 0.24500304898705874, "grad_norm": 8.96689224243164, "learning_rate": 9.607776028475598e-05, "loss": 0.9816, "step": 3616 }, { "epoch": 0.24507080425503083, "grad_norm": 15.17086124420166, "learning_rate": 9.607639126565816e-05, "loss": 1.1416, "step": 3617 }, { "epoch": 0.24513855952300292, "grad_norm": 9.527314186096191, "learning_rate": 9.607502224656034e-05, "loss": 1.1699, "step": 3618 }, { "epoch": 0.245206314790975, "grad_norm": 7.378002166748047, "learning_rate": 9.607365322746252e-05, "loss": 1.1303, "step": 3619 }, { "epoch": 0.24527407005894708, "grad_norm": 7.204291343688965, "learning_rate": 9.607228420836471e-05, "loss": 0.8078, "step": 3620 }, { "epoch": 0.24534182532691917, "grad_norm": 8.181205749511719, "learning_rate": 9.607091518926689e-05, "loss": 0.9221, "step": 3621 }, { "epoch": 0.24540958059489126, "grad_norm": 8.479545593261719, "learning_rate": 9.606954617016907e-05, "loss": 0.8133, "step": 3622 }, { "epoch": 0.24547733586286333, "grad_norm": 7.591360092163086, "learning_rate": 9.606817715107125e-05, "loss": 1.0632, "step": 3623 }, { "epoch": 0.24554509113083542, "grad_norm": 8.558969497680664, "learning_rate": 9.606680813197343e-05, "loss": 1.0755, "step": 3624 }, { "epoch": 0.2456128463988075, "grad_norm": 8.02037525177002, "learning_rate": 9.606543911287563e-05, "loss": 1.0164, "step": 3625 }, { "epoch": 0.2456806016667796, "grad_norm": 7.49207878112793, "learning_rate": 9.606407009377781e-05, "loss": 1.1325, "step": 3626 }, { "epoch": 0.24574835693475167, "grad_norm": 7.376079559326172, "learning_rate": 9.606270107467999e-05, "loss": 0.7917, "step": 3627 }, { "epoch": 0.24581611220272376, "grad_norm": 7.331247329711914, "learning_rate": 9.606133205558218e-05, "loss": 1.2186, "step": 3628 }, { "epoch": 0.24588386747069585, "grad_norm": 7.393257141113281, "learning_rate": 9.605996303648436e-05, "loss": 0.7155, "step": 3629 }, { "epoch": 0.24595162273866794, "grad_norm": 9.15472412109375, "learning_rate": 9.605859401738654e-05, "loss": 0.9562, "step": 3630 }, { "epoch": 0.24601937800664, "grad_norm": 6.846646785736084, "learning_rate": 9.605722499828874e-05, "loss": 1.109, "step": 3631 }, { "epoch": 0.2460871332746121, "grad_norm": 8.57854175567627, "learning_rate": 9.605585597919092e-05, "loss": 0.8706, "step": 3632 }, { "epoch": 0.2461548885425842, "grad_norm": 7.089768886566162, "learning_rate": 9.60544869600931e-05, "loss": 0.8776, "step": 3633 }, { "epoch": 0.24622264381055628, "grad_norm": 7.549044132232666, "learning_rate": 9.605311794099529e-05, "loss": 0.9956, "step": 3634 }, { "epoch": 0.24629039907852834, "grad_norm": 6.839412689208984, "learning_rate": 9.605174892189747e-05, "loss": 0.8722, "step": 3635 }, { "epoch": 0.24635815434650044, "grad_norm": 7.378058910369873, "learning_rate": 9.605037990279965e-05, "loss": 0.8663, "step": 3636 }, { "epoch": 0.24642590961447253, "grad_norm": 7.893070220947266, "learning_rate": 9.604901088370183e-05, "loss": 0.9213, "step": 3637 }, { "epoch": 0.24649366488244462, "grad_norm": 7.3345232009887695, "learning_rate": 9.604764186460403e-05, "loss": 0.9827, "step": 3638 }, { "epoch": 0.24656142015041668, "grad_norm": 6.120781421661377, "learning_rate": 9.60462728455062e-05, "loss": 0.6371, "step": 3639 }, { "epoch": 0.24662917541838877, "grad_norm": 8.695615768432617, "learning_rate": 9.604490382640839e-05, "loss": 0.9769, "step": 3640 }, { "epoch": 0.24669693068636087, "grad_norm": 8.469325065612793, "learning_rate": 9.604353480731057e-05, "loss": 0.9267, "step": 3641 }, { "epoch": 0.24676468595433296, "grad_norm": 9.191173553466797, "learning_rate": 9.604216578821275e-05, "loss": 1.1818, "step": 3642 }, { "epoch": 0.24683244122230502, "grad_norm": 9.337483406066895, "learning_rate": 9.604079676911494e-05, "loss": 0.9904, "step": 3643 }, { "epoch": 0.2469001964902771, "grad_norm": 7.597773551940918, "learning_rate": 9.603942775001712e-05, "loss": 0.8928, "step": 3644 }, { "epoch": 0.2469679517582492, "grad_norm": 8.155903816223145, "learning_rate": 9.60380587309193e-05, "loss": 0.9511, "step": 3645 }, { "epoch": 0.2470357070262213, "grad_norm": 7.695154666900635, "learning_rate": 9.603668971182148e-05, "loss": 1.0724, "step": 3646 }, { "epoch": 0.24710346229419336, "grad_norm": 7.492908000946045, "learning_rate": 9.603532069272366e-05, "loss": 0.9133, "step": 3647 }, { "epoch": 0.24717121756216545, "grad_norm": 8.190613746643066, "learning_rate": 9.603395167362586e-05, "loss": 0.8972, "step": 3648 }, { "epoch": 0.24723897283013754, "grad_norm": 7.9199347496032715, "learning_rate": 9.603258265452804e-05, "loss": 1.0501, "step": 3649 }, { "epoch": 0.24730672809810963, "grad_norm": 8.386896133422852, "learning_rate": 9.603121363543022e-05, "loss": 0.998, "step": 3650 }, { "epoch": 0.24737448336608173, "grad_norm": 6.536781311035156, "learning_rate": 9.60298446163324e-05, "loss": 0.8379, "step": 3651 }, { "epoch": 0.2474422386340538, "grad_norm": 6.918766021728516, "learning_rate": 9.602847559723459e-05, "loss": 0.9397, "step": 3652 }, { "epoch": 0.24750999390202588, "grad_norm": 7.00775146484375, "learning_rate": 9.602710657813677e-05, "loss": 1.0079, "step": 3653 }, { "epoch": 0.24757774916999797, "grad_norm": 7.693192958831787, "learning_rate": 9.602573755903895e-05, "loss": 1.0657, "step": 3654 }, { "epoch": 0.24764550443797007, "grad_norm": 10.667771339416504, "learning_rate": 9.602436853994113e-05, "loss": 1.0657, "step": 3655 }, { "epoch": 0.24771325970594213, "grad_norm": 8.622758865356445, "learning_rate": 9.602299952084331e-05, "loss": 0.9792, "step": 3656 }, { "epoch": 0.24778101497391422, "grad_norm": 8.834444046020508, "learning_rate": 9.602163050174551e-05, "loss": 1.1014, "step": 3657 }, { "epoch": 0.2478487702418863, "grad_norm": 7.717538356781006, "learning_rate": 9.602026148264769e-05, "loss": 0.8781, "step": 3658 }, { "epoch": 0.2479165255098584, "grad_norm": 7.797954559326172, "learning_rate": 9.601889246354987e-05, "loss": 0.9772, "step": 3659 }, { "epoch": 0.24798428077783047, "grad_norm": 7.376112937927246, "learning_rate": 9.601752344445205e-05, "loss": 1.0468, "step": 3660 }, { "epoch": 0.24805203604580256, "grad_norm": 7.7266387939453125, "learning_rate": 9.601615442535424e-05, "loss": 1.007, "step": 3661 }, { "epoch": 0.24811979131377465, "grad_norm": 9.196928977966309, "learning_rate": 9.601478540625642e-05, "loss": 1.1327, "step": 3662 }, { "epoch": 0.24818754658174674, "grad_norm": 7.892288684844971, "learning_rate": 9.60134163871586e-05, "loss": 0.7511, "step": 3663 }, { "epoch": 0.2482553018497188, "grad_norm": 8.854056358337402, "learning_rate": 9.601204736806078e-05, "loss": 0.9546, "step": 3664 }, { "epoch": 0.2483230571176909, "grad_norm": 7.985452651977539, "learning_rate": 9.601067834896296e-05, "loss": 1.2186, "step": 3665 }, { "epoch": 0.248390812385663, "grad_norm": 8.336162567138672, "learning_rate": 9.600930932986516e-05, "loss": 0.864, "step": 3666 }, { "epoch": 0.24845856765363508, "grad_norm": 10.857757568359375, "learning_rate": 9.600794031076734e-05, "loss": 1.1948, "step": 3667 }, { "epoch": 0.24852632292160715, "grad_norm": 8.168721199035645, "learning_rate": 9.600657129166952e-05, "loss": 1.1151, "step": 3668 }, { "epoch": 0.24859407818957924, "grad_norm": 7.509332656860352, "learning_rate": 9.60052022725717e-05, "loss": 0.9215, "step": 3669 }, { "epoch": 0.24866183345755133, "grad_norm": 10.4354829788208, "learning_rate": 9.600383325347389e-05, "loss": 1.0696, "step": 3670 }, { "epoch": 0.24872958872552342, "grad_norm": 6.930381774902344, "learning_rate": 9.600246423437607e-05, "loss": 0.9229, "step": 3671 }, { "epoch": 0.24879734399349548, "grad_norm": 6.608088970184326, "learning_rate": 9.600109521527825e-05, "loss": 0.928, "step": 3672 }, { "epoch": 0.24886509926146758, "grad_norm": 7.4217095375061035, "learning_rate": 9.599972619618043e-05, "loss": 0.9608, "step": 3673 }, { "epoch": 0.24893285452943967, "grad_norm": 7.46991491317749, "learning_rate": 9.599835717708263e-05, "loss": 0.8928, "step": 3674 }, { "epoch": 0.24900060979741176, "grad_norm": 8.76001262664795, "learning_rate": 9.599698815798481e-05, "loss": 1.191, "step": 3675 }, { "epoch": 0.24906836506538382, "grad_norm": 8.395779609680176, "learning_rate": 9.599561913888699e-05, "loss": 1.1479, "step": 3676 }, { "epoch": 0.24913612033335591, "grad_norm": 8.00460147857666, "learning_rate": 9.599425011978918e-05, "loss": 1.1005, "step": 3677 }, { "epoch": 0.249203875601328, "grad_norm": 7.817287445068359, "learning_rate": 9.599288110069136e-05, "loss": 1.1568, "step": 3678 }, { "epoch": 0.2492716308693001, "grad_norm": 9.60706615447998, "learning_rate": 9.599151208159354e-05, "loss": 0.9734, "step": 3679 }, { "epoch": 0.24933938613727216, "grad_norm": 8.347691535949707, "learning_rate": 9.599014306249574e-05, "loss": 1.2208, "step": 3680 }, { "epoch": 0.24940714140524425, "grad_norm": 8.106613159179688, "learning_rate": 9.598877404339792e-05, "loss": 0.9253, "step": 3681 }, { "epoch": 0.24947489667321635, "grad_norm": 8.099063873291016, "learning_rate": 9.59874050243001e-05, "loss": 1.0516, "step": 3682 }, { "epoch": 0.24954265194118844, "grad_norm": 7.839589595794678, "learning_rate": 9.598603600520228e-05, "loss": 0.9764, "step": 3683 }, { "epoch": 0.2496104072091605, "grad_norm": 8.679840087890625, "learning_rate": 9.598466698610447e-05, "loss": 1.0407, "step": 3684 }, { "epoch": 0.2496781624771326, "grad_norm": 6.854926109313965, "learning_rate": 9.598329796700665e-05, "loss": 0.9457, "step": 3685 }, { "epoch": 0.24974591774510468, "grad_norm": 9.133661270141602, "learning_rate": 9.598192894790883e-05, "loss": 1.0391, "step": 3686 }, { "epoch": 0.24981367301307678, "grad_norm": 7.744726657867432, "learning_rate": 9.598055992881101e-05, "loss": 1.0256, "step": 3687 }, { "epoch": 0.24988142828104884, "grad_norm": 8.9452543258667, "learning_rate": 9.597919090971319e-05, "loss": 0.9649, "step": 3688 }, { "epoch": 0.24994918354902093, "grad_norm": 8.396431922912598, "learning_rate": 9.597782189061539e-05, "loss": 0.8189, "step": 3689 }, { "epoch": 0.24994918354902093, "eval_loss": 0.9518795609474182, "eval_noise_accuracy": 0.0, "eval_runtime": 1533.2421, "eval_samples_per_second": 3.352, "eval_steps_per_second": 0.21, "eval_wer": 88.77438705594233, "step": 3689 } ], "logging_steps": 1, "max_steps": 73795, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 3689, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.45310507532288e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }