{ "best_global_step": 7378, "best_metric": 75.89224292121845, "best_model_checkpoint": "whisper-tiny-bfloat16-sada/checkpoints/checkpoint-7378", "epoch": 0.49989836709804186, "eval_steps": 3689, "global_step": 7378, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 6.775526797208483e-05, "grad_norm": 132.70848083496094, "learning_rate": 0.0, "loss": 4.4561, "step": 1 }, { "epoch": 0.00013551053594416967, "grad_norm": 133.01596069335938, "learning_rate": 1.3333333333333334e-07, "loss": 4.8642, "step": 2 }, { "epoch": 0.0002032658039162545, "grad_norm": 106.03809356689453, "learning_rate": 2.6666666666666667e-07, "loss": 4.3778, "step": 3 }, { "epoch": 0.00027102107188833934, "grad_norm": 88.75433349609375, "learning_rate": 4.0000000000000003e-07, "loss": 4.0385, "step": 4 }, { "epoch": 0.00033877633986042414, "grad_norm": 101.19609832763672, "learning_rate": 5.333333333333333e-07, "loss": 3.9092, "step": 5 }, { "epoch": 0.000406531607832509, "grad_norm": 67.02723693847656, "learning_rate": 6.666666666666667e-07, "loss": 3.7556, "step": 6 }, { "epoch": 0.0004742868758045938, "grad_norm": 67.52012634277344, "learning_rate": 8.000000000000001e-07, "loss": 3.7119, "step": 7 }, { "epoch": 0.0005420421437766787, "grad_norm": 83.27348327636719, "learning_rate": 9.333333333333334e-07, "loss": 3.6125, "step": 8 }, { "epoch": 0.0006097974117487635, "grad_norm": 84.20647430419922, "learning_rate": 1.0666666666666667e-06, "loss": 4.1354, "step": 9 }, { "epoch": 0.0006775526797208483, "grad_norm": 110.12113952636719, "learning_rate": 1.2000000000000002e-06, "loss": 3.9594, "step": 10 }, { "epoch": 0.0007453079476929331, "grad_norm": 84.61614990234375, "learning_rate": 1.3333333333333334e-06, "loss": 4.1415, "step": 11 }, { "epoch": 0.000813063215665018, "grad_norm": 119.0740966796875, "learning_rate": 1.4666666666666667e-06, "loss": 4.0689, "step": 12 }, { "epoch": 0.0008808184836371028, "grad_norm": 122.41352081298828, "learning_rate": 1.6000000000000001e-06, "loss": 4.0834, "step": 13 }, { "epoch": 0.0009485737516091876, "grad_norm": 99.06126403808594, "learning_rate": 1.7333333333333334e-06, "loss": 4.4224, "step": 14 }, { "epoch": 0.0010163290195812724, "grad_norm": 113.89531707763672, "learning_rate": 1.8666666666666669e-06, "loss": 4.0188, "step": 15 }, { "epoch": 0.0010840842875533573, "grad_norm": 70.24198150634766, "learning_rate": 2.0000000000000003e-06, "loss": 3.3951, "step": 16 }, { "epoch": 0.001151839555525442, "grad_norm": 67.99130249023438, "learning_rate": 2.1333333333333334e-06, "loss": 3.7385, "step": 17 }, { "epoch": 0.001219594823497527, "grad_norm": 54.554779052734375, "learning_rate": 2.266666666666667e-06, "loss": 2.9009, "step": 18 }, { "epoch": 0.0012873500914696117, "grad_norm": 47.79149627685547, "learning_rate": 2.4000000000000003e-06, "loss": 3.0572, "step": 19 }, { "epoch": 0.0013551053594416966, "grad_norm": 50.05776596069336, "learning_rate": 2.5333333333333334e-06, "loss": 2.8678, "step": 20 }, { "epoch": 0.0014228606274137815, "grad_norm": 55.41701889038086, "learning_rate": 2.666666666666667e-06, "loss": 3.3981, "step": 21 }, { "epoch": 0.0014906158953858662, "grad_norm": 57.1912727355957, "learning_rate": 2.8000000000000003e-06, "loss": 3.3247, "step": 22 }, { "epoch": 0.001558371163357951, "grad_norm": 39.77119445800781, "learning_rate": 2.9333333333333333e-06, "loss": 3.1466, "step": 23 }, { "epoch": 0.001626126431330036, "grad_norm": 37.258392333984375, "learning_rate": 3.066666666666667e-06, "loss": 3.0137, "step": 24 }, { "epoch": 0.0016938816993021207, "grad_norm": 41.15428924560547, "learning_rate": 3.2000000000000003e-06, "loss": 2.9908, "step": 25 }, { "epoch": 0.0017616369672742056, "grad_norm": 44.59917449951172, "learning_rate": 3.3333333333333333e-06, "loss": 3.6947, "step": 26 }, { "epoch": 0.0018293922352462903, "grad_norm": 32.7044792175293, "learning_rate": 3.466666666666667e-06, "loss": 2.9421, "step": 27 }, { "epoch": 0.0018971475032183752, "grad_norm": 28.445444107055664, "learning_rate": 3.6e-06, "loss": 2.9849, "step": 28 }, { "epoch": 0.00196490277119046, "grad_norm": 30.879247665405273, "learning_rate": 3.7333333333333337e-06, "loss": 2.9007, "step": 29 }, { "epoch": 0.002032658039162545, "grad_norm": 22.723188400268555, "learning_rate": 3.866666666666667e-06, "loss": 2.3549, "step": 30 }, { "epoch": 0.0021004133071346296, "grad_norm": 23.59457778930664, "learning_rate": 4.000000000000001e-06, "loss": 2.8564, "step": 31 }, { "epoch": 0.0021681685751067147, "grad_norm": 23.313579559326172, "learning_rate": 4.133333333333333e-06, "loss": 2.453, "step": 32 }, { "epoch": 0.0022359238430787994, "grad_norm": 23.877405166625977, "learning_rate": 4.266666666666667e-06, "loss": 2.2042, "step": 33 }, { "epoch": 0.002303679111050884, "grad_norm": 29.79376792907715, "learning_rate": 4.4e-06, "loss": 2.6679, "step": 34 }, { "epoch": 0.002371434379022969, "grad_norm": 21.018878936767578, "learning_rate": 4.533333333333334e-06, "loss": 2.3983, "step": 35 }, { "epoch": 0.002439189646995054, "grad_norm": 30.22182846069336, "learning_rate": 4.666666666666667e-06, "loss": 2.8668, "step": 36 }, { "epoch": 0.0025069449149671386, "grad_norm": 24.619905471801758, "learning_rate": 4.800000000000001e-06, "loss": 2.5986, "step": 37 }, { "epoch": 0.0025747001829392233, "grad_norm": 25.470375061035156, "learning_rate": 4.933333333333333e-06, "loss": 2.4936, "step": 38 }, { "epoch": 0.0026424554509113084, "grad_norm": 19.872745513916016, "learning_rate": 5.066666666666667e-06, "loss": 2.1633, "step": 39 }, { "epoch": 0.002710210718883393, "grad_norm": 24.880964279174805, "learning_rate": 5.2e-06, "loss": 2.6798, "step": 40 }, { "epoch": 0.002777965986855478, "grad_norm": 22.790157318115234, "learning_rate": 5.333333333333334e-06, "loss": 2.487, "step": 41 }, { "epoch": 0.002845721254827563, "grad_norm": 21.404136657714844, "learning_rate": 5.466666666666667e-06, "loss": 2.3383, "step": 42 }, { "epoch": 0.0029134765227996477, "grad_norm": 22.80590057373047, "learning_rate": 5.600000000000001e-06, "loss": 2.6347, "step": 43 }, { "epoch": 0.0029812317907717324, "grad_norm": 20.625160217285156, "learning_rate": 5.733333333333333e-06, "loss": 2.066, "step": 44 }, { "epoch": 0.0030489870587438175, "grad_norm": 21.700828552246094, "learning_rate": 5.866666666666667e-06, "loss": 2.188, "step": 45 }, { "epoch": 0.003116742326715902, "grad_norm": 18.8945255279541, "learning_rate": 6e-06, "loss": 2.1531, "step": 46 }, { "epoch": 0.003184497594687987, "grad_norm": 20.238525390625, "learning_rate": 6.133333333333334e-06, "loss": 1.9306, "step": 47 }, { "epoch": 0.003252252862660072, "grad_norm": 23.74344825744629, "learning_rate": 6.266666666666666e-06, "loss": 2.3206, "step": 48 }, { "epoch": 0.0033200081306321567, "grad_norm": 22.429964065551758, "learning_rate": 6.4000000000000006e-06, "loss": 2.05, "step": 49 }, { "epoch": 0.0033877633986042414, "grad_norm": 22.078792572021484, "learning_rate": 6.533333333333333e-06, "loss": 2.1155, "step": 50 }, { "epoch": 0.003455518666576326, "grad_norm": 19.691747665405273, "learning_rate": 6.666666666666667e-06, "loss": 1.8719, "step": 51 }, { "epoch": 0.0035232739345484113, "grad_norm": 27.412866592407227, "learning_rate": 6.800000000000001e-06, "loss": 2.0943, "step": 52 }, { "epoch": 0.003591029202520496, "grad_norm": 21.631580352783203, "learning_rate": 6.933333333333334e-06, "loss": 2.246, "step": 53 }, { "epoch": 0.0036587844704925807, "grad_norm": 16.729095458984375, "learning_rate": 7.066666666666667e-06, "loss": 1.8101, "step": 54 }, { "epoch": 0.003726539738464666, "grad_norm": 18.381364822387695, "learning_rate": 7.2e-06, "loss": 1.8668, "step": 55 }, { "epoch": 0.0037942950064367505, "grad_norm": 19.661123275756836, "learning_rate": 7.333333333333334e-06, "loss": 2.1912, "step": 56 }, { "epoch": 0.003862050274408835, "grad_norm": 13.772311210632324, "learning_rate": 7.4666666666666675e-06, "loss": 1.7408, "step": 57 }, { "epoch": 0.00392980554238092, "grad_norm": 20.559425354003906, "learning_rate": 7.6e-06, "loss": 1.9717, "step": 58 }, { "epoch": 0.003997560810353005, "grad_norm": 17.385791778564453, "learning_rate": 7.733333333333334e-06, "loss": 1.7681, "step": 59 }, { "epoch": 0.00406531607832509, "grad_norm": 18.455888748168945, "learning_rate": 7.866666666666667e-06, "loss": 1.8822, "step": 60 }, { "epoch": 0.004133071346297174, "grad_norm": 17.53643798828125, "learning_rate": 8.000000000000001e-06, "loss": 1.8158, "step": 61 }, { "epoch": 0.004200826614269259, "grad_norm": 15.640396118164062, "learning_rate": 8.133333333333332e-06, "loss": 1.8786, "step": 62 }, { "epoch": 0.004268581882241345, "grad_norm": 67.14771270751953, "learning_rate": 8.266666666666667e-06, "loss": 1.845, "step": 63 }, { "epoch": 0.004336337150213429, "grad_norm": 16.778207778930664, "learning_rate": 8.400000000000001e-06, "loss": 1.7286, "step": 64 }, { "epoch": 0.004404092418185514, "grad_norm": 20.485218048095703, "learning_rate": 8.533333333333334e-06, "loss": 1.882, "step": 65 }, { "epoch": 0.004471847686157599, "grad_norm": 15.77828598022461, "learning_rate": 8.666666666666668e-06, "loss": 1.7788, "step": 66 }, { "epoch": 0.0045396029541296835, "grad_norm": 20.225608825683594, "learning_rate": 8.8e-06, "loss": 1.9842, "step": 67 }, { "epoch": 0.004607358222101768, "grad_norm": 16.230987548828125, "learning_rate": 8.933333333333333e-06, "loss": 1.4581, "step": 68 }, { "epoch": 0.004675113490073853, "grad_norm": 13.831804275512695, "learning_rate": 9.066666666666667e-06, "loss": 1.6173, "step": 69 }, { "epoch": 0.004742868758045938, "grad_norm": 16.4691104888916, "learning_rate": 9.2e-06, "loss": 1.6055, "step": 70 }, { "epoch": 0.004810624026018023, "grad_norm": 17.788105010986328, "learning_rate": 9.333333333333334e-06, "loss": 1.7996, "step": 71 }, { "epoch": 0.004878379293990108, "grad_norm": 20.043182373046875, "learning_rate": 9.466666666666667e-06, "loss": 1.7224, "step": 72 }, { "epoch": 0.0049461345619621925, "grad_norm": 14.918837547302246, "learning_rate": 9.600000000000001e-06, "loss": 1.4448, "step": 73 }, { "epoch": 0.005013889829934277, "grad_norm": 15.70170783996582, "learning_rate": 9.733333333333334e-06, "loss": 1.7469, "step": 74 }, { "epoch": 0.005081645097906362, "grad_norm": 19.128456115722656, "learning_rate": 9.866666666666667e-06, "loss": 1.9505, "step": 75 }, { "epoch": 0.005149400365878447, "grad_norm": 34.82925796508789, "learning_rate": 1e-05, "loss": 1.6543, "step": 76 }, { "epoch": 0.005217155633850532, "grad_norm": 19.61214256286621, "learning_rate": 1.0133333333333333e-05, "loss": 1.9051, "step": 77 }, { "epoch": 0.005284910901822617, "grad_norm": 20.190847396850586, "learning_rate": 1.0266666666666668e-05, "loss": 1.6016, "step": 78 }, { "epoch": 0.005352666169794702, "grad_norm": 15.873608589172363, "learning_rate": 1.04e-05, "loss": 1.8286, "step": 79 }, { "epoch": 0.005420421437766786, "grad_norm": 17.513385772705078, "learning_rate": 1.0533333333333335e-05, "loss": 1.7122, "step": 80 }, { "epoch": 0.005488176705738871, "grad_norm": 13.794294357299805, "learning_rate": 1.0666666666666667e-05, "loss": 1.5369, "step": 81 }, { "epoch": 0.005555931973710956, "grad_norm": 16.763822555541992, "learning_rate": 1.08e-05, "loss": 1.688, "step": 82 }, { "epoch": 0.005623687241683041, "grad_norm": 15.182608604431152, "learning_rate": 1.0933333333333334e-05, "loss": 1.9076, "step": 83 }, { "epoch": 0.005691442509655126, "grad_norm": 19.11581802368164, "learning_rate": 1.1066666666666667e-05, "loss": 1.6034, "step": 84 }, { "epoch": 0.005759197777627211, "grad_norm": 18.303098678588867, "learning_rate": 1.1200000000000001e-05, "loss": 1.8248, "step": 85 }, { "epoch": 0.005826953045599295, "grad_norm": 17.912731170654297, "learning_rate": 1.1333333333333334e-05, "loss": 1.737, "step": 86 }, { "epoch": 0.00589470831357138, "grad_norm": 18.4353084564209, "learning_rate": 1.1466666666666666e-05, "loss": 1.7141, "step": 87 }, { "epoch": 0.005962463581543465, "grad_norm": 17.006074905395508, "learning_rate": 1.16e-05, "loss": 1.6568, "step": 88 }, { "epoch": 0.0060302188495155494, "grad_norm": 15.084410667419434, "learning_rate": 1.1733333333333333e-05, "loss": 1.3369, "step": 89 }, { "epoch": 0.006097974117487635, "grad_norm": 14.984724044799805, "learning_rate": 1.1866666666666668e-05, "loss": 1.777, "step": 90 }, { "epoch": 0.00616572938545972, "grad_norm": 19.447574615478516, "learning_rate": 1.2e-05, "loss": 1.4225, "step": 91 }, { "epoch": 0.006233484653431804, "grad_norm": 15.212474822998047, "learning_rate": 1.2133333333333335e-05, "loss": 1.4896, "step": 92 }, { "epoch": 0.006301239921403889, "grad_norm": 17.97924041748047, "learning_rate": 1.2266666666666667e-05, "loss": 1.5947, "step": 93 }, { "epoch": 0.006368995189375974, "grad_norm": 13.695365905761719, "learning_rate": 1.24e-05, "loss": 1.471, "step": 94 }, { "epoch": 0.0064367504573480585, "grad_norm": 15.574562072753906, "learning_rate": 1.2533333333333332e-05, "loss": 1.5803, "step": 95 }, { "epoch": 0.006504505725320144, "grad_norm": 16.71872329711914, "learning_rate": 1.2666666666666668e-05, "loss": 1.7408, "step": 96 }, { "epoch": 0.006572260993292229, "grad_norm": 14.232036590576172, "learning_rate": 1.2800000000000001e-05, "loss": 1.4414, "step": 97 }, { "epoch": 0.0066400162612643135, "grad_norm": 14.430774688720703, "learning_rate": 1.2933333333333334e-05, "loss": 1.712, "step": 98 }, { "epoch": 0.006707771529236398, "grad_norm": 15.887603759765625, "learning_rate": 1.3066666666666666e-05, "loss": 1.6097, "step": 99 }, { "epoch": 0.006775526797208483, "grad_norm": 21.155643463134766, "learning_rate": 1.32e-05, "loss": 2.0581, "step": 100 }, { "epoch": 0.0068432820651805676, "grad_norm": 17.30326271057129, "learning_rate": 1.3333333333333333e-05, "loss": 1.2823, "step": 101 }, { "epoch": 0.006911037333152652, "grad_norm": 18.184707641601562, "learning_rate": 1.3466666666666666e-05, "loss": 1.5165, "step": 102 }, { "epoch": 0.006978792601124738, "grad_norm": 23.46563148498535, "learning_rate": 1.3600000000000002e-05, "loss": 1.406, "step": 103 }, { "epoch": 0.0070465478690968225, "grad_norm": 14.586087226867676, "learning_rate": 1.3733333333333335e-05, "loss": 1.8416, "step": 104 }, { "epoch": 0.007114303137068907, "grad_norm": 14.55811595916748, "learning_rate": 1.3866666666666667e-05, "loss": 1.5466, "step": 105 }, { "epoch": 0.007182058405040992, "grad_norm": 20.244586944580078, "learning_rate": 1.4000000000000001e-05, "loss": 1.755, "step": 106 }, { "epoch": 0.007249813673013077, "grad_norm": 11.932172775268555, "learning_rate": 1.4133333333333334e-05, "loss": 1.4543, "step": 107 }, { "epoch": 0.007317568940985161, "grad_norm": 17.406888961791992, "learning_rate": 1.4266666666666667e-05, "loss": 1.4436, "step": 108 }, { "epoch": 0.007385324208957246, "grad_norm": 14.591705322265625, "learning_rate": 1.44e-05, "loss": 1.4088, "step": 109 }, { "epoch": 0.007453079476929332, "grad_norm": 18.028337478637695, "learning_rate": 1.4533333333333335e-05, "loss": 1.6675, "step": 110 }, { "epoch": 0.007520834744901416, "grad_norm": 16.37979507446289, "learning_rate": 1.4666666666666668e-05, "loss": 1.6646, "step": 111 }, { "epoch": 0.007588590012873501, "grad_norm": 19.028308868408203, "learning_rate": 1.48e-05, "loss": 1.701, "step": 112 }, { "epoch": 0.007656345280845586, "grad_norm": 17.28181266784668, "learning_rate": 1.4933333333333335e-05, "loss": 1.3984, "step": 113 }, { "epoch": 0.00772410054881767, "grad_norm": 16.78407859802246, "learning_rate": 1.5066666666666668e-05, "loss": 1.6851, "step": 114 }, { "epoch": 0.007791855816789755, "grad_norm": 15.634642601013184, "learning_rate": 1.52e-05, "loss": 1.6484, "step": 115 }, { "epoch": 0.00785961108476184, "grad_norm": 13.879626274108887, "learning_rate": 1.5333333333333334e-05, "loss": 1.4609, "step": 116 }, { "epoch": 0.007927366352733925, "grad_norm": 16.247541427612305, "learning_rate": 1.546666666666667e-05, "loss": 1.5219, "step": 117 }, { "epoch": 0.00799512162070601, "grad_norm": 20.57497215270996, "learning_rate": 1.56e-05, "loss": 1.6038, "step": 118 }, { "epoch": 0.008062876888678095, "grad_norm": 16.32591438293457, "learning_rate": 1.5733333333333334e-05, "loss": 1.5315, "step": 119 }, { "epoch": 0.00813063215665018, "grad_norm": 18.236255645751953, "learning_rate": 1.586666666666667e-05, "loss": 1.6589, "step": 120 }, { "epoch": 0.008198387424622264, "grad_norm": 20.850723266601562, "learning_rate": 1.6000000000000003e-05, "loss": 1.5556, "step": 121 }, { "epoch": 0.008266142692594349, "grad_norm": 17.931947708129883, "learning_rate": 1.6133333333333334e-05, "loss": 1.8209, "step": 122 }, { "epoch": 0.008333897960566434, "grad_norm": 20.531381607055664, "learning_rate": 1.6266666666666665e-05, "loss": 1.6421, "step": 123 }, { "epoch": 0.008401653228538518, "grad_norm": 15.6729736328125, "learning_rate": 1.6400000000000002e-05, "loss": 1.5652, "step": 124 }, { "epoch": 0.008469408496510603, "grad_norm": 18.139949798583984, "learning_rate": 1.6533333333333333e-05, "loss": 1.5578, "step": 125 }, { "epoch": 0.00853716376448269, "grad_norm": 15.77291202545166, "learning_rate": 1.6666666666666667e-05, "loss": 1.6091, "step": 126 }, { "epoch": 0.008604919032454774, "grad_norm": 18.00165557861328, "learning_rate": 1.6800000000000002e-05, "loss": 1.4725, "step": 127 }, { "epoch": 0.008672674300426859, "grad_norm": 17.33936309814453, "learning_rate": 1.6933333333333333e-05, "loss": 1.4272, "step": 128 }, { "epoch": 0.008740429568398943, "grad_norm": 16.75677490234375, "learning_rate": 1.7066666666666667e-05, "loss": 1.602, "step": 129 }, { "epoch": 0.008808184836371028, "grad_norm": 17.761140823364258, "learning_rate": 1.7199999999999998e-05, "loss": 1.8523, "step": 130 }, { "epoch": 0.008875940104343113, "grad_norm": 15.888497352600098, "learning_rate": 1.7333333333333336e-05, "loss": 1.5594, "step": 131 }, { "epoch": 0.008943695372315198, "grad_norm": 19.62826156616211, "learning_rate": 1.7466666666666667e-05, "loss": 1.6909, "step": 132 }, { "epoch": 0.009011450640287282, "grad_norm": 15.62592887878418, "learning_rate": 1.76e-05, "loss": 1.5873, "step": 133 }, { "epoch": 0.009079205908259367, "grad_norm": 17.64893913269043, "learning_rate": 1.7733333333333335e-05, "loss": 1.5681, "step": 134 }, { "epoch": 0.009146961176231452, "grad_norm": 15.14854907989502, "learning_rate": 1.7866666666666666e-05, "loss": 1.3324, "step": 135 }, { "epoch": 0.009214716444203536, "grad_norm": 18.115419387817383, "learning_rate": 1.8e-05, "loss": 1.5512, "step": 136 }, { "epoch": 0.009282471712175621, "grad_norm": 18.2511043548584, "learning_rate": 1.8133333333333335e-05, "loss": 1.6405, "step": 137 }, { "epoch": 0.009350226980147706, "grad_norm": 13.296855926513672, "learning_rate": 1.826666666666667e-05, "loss": 1.2269, "step": 138 }, { "epoch": 0.009417982248119792, "grad_norm": 17.17320442199707, "learning_rate": 1.84e-05, "loss": 1.4031, "step": 139 }, { "epoch": 0.009485737516091877, "grad_norm": 15.399650573730469, "learning_rate": 1.8533333333333334e-05, "loss": 1.4793, "step": 140 }, { "epoch": 0.009553492784063962, "grad_norm": 15.0385160446167, "learning_rate": 1.866666666666667e-05, "loss": 1.4129, "step": 141 }, { "epoch": 0.009621248052036046, "grad_norm": 14.586984634399414, "learning_rate": 1.88e-05, "loss": 1.1959, "step": 142 }, { "epoch": 0.009689003320008131, "grad_norm": 14.939247131347656, "learning_rate": 1.8933333333333334e-05, "loss": 1.6728, "step": 143 }, { "epoch": 0.009756758587980216, "grad_norm": 18.1126651763916, "learning_rate": 1.9066666666666668e-05, "loss": 1.5035, "step": 144 }, { "epoch": 0.0098245138559523, "grad_norm": 16.0823974609375, "learning_rate": 1.9200000000000003e-05, "loss": 1.544, "step": 145 }, { "epoch": 0.009892269123924385, "grad_norm": 17.45062828063965, "learning_rate": 1.9333333333333333e-05, "loss": 1.4593, "step": 146 }, { "epoch": 0.00996002439189647, "grad_norm": 21.030731201171875, "learning_rate": 1.9466666666666668e-05, "loss": 1.6544, "step": 147 }, { "epoch": 0.010027779659868554, "grad_norm": 18.31300163269043, "learning_rate": 1.9600000000000002e-05, "loss": 1.7682, "step": 148 }, { "epoch": 0.01009553492784064, "grad_norm": 13.541064262390137, "learning_rate": 1.9733333333333333e-05, "loss": 1.333, "step": 149 }, { "epoch": 0.010163290195812724, "grad_norm": 13.538237571716309, "learning_rate": 1.9866666666666667e-05, "loss": 1.3629, "step": 150 }, { "epoch": 0.010231045463784809, "grad_norm": 18.393657684326172, "learning_rate": 2e-05, "loss": 1.4467, "step": 151 }, { "epoch": 0.010298800731756893, "grad_norm": 19.18985939025879, "learning_rate": 2.0133333333333336e-05, "loss": 1.612, "step": 152 }, { "epoch": 0.01036655599972898, "grad_norm": 14.230488777160645, "learning_rate": 2.0266666666666667e-05, "loss": 1.4047, "step": 153 }, { "epoch": 0.010434311267701064, "grad_norm": 13.807863235473633, "learning_rate": 2.04e-05, "loss": 1.4977, "step": 154 }, { "epoch": 0.010502066535673149, "grad_norm": 15.149568557739258, "learning_rate": 2.0533333333333336e-05, "loss": 1.3856, "step": 155 }, { "epoch": 0.010569821803645234, "grad_norm": 16.095308303833008, "learning_rate": 2.0666666666666666e-05, "loss": 1.3687, "step": 156 }, { "epoch": 0.010637577071617318, "grad_norm": 14.864961624145508, "learning_rate": 2.08e-05, "loss": 1.5707, "step": 157 }, { "epoch": 0.010705332339589403, "grad_norm": 15.9810152053833, "learning_rate": 2.0933333333333335e-05, "loss": 1.6841, "step": 158 }, { "epoch": 0.010773087607561488, "grad_norm": 16.706584930419922, "learning_rate": 2.106666666666667e-05, "loss": 1.5017, "step": 159 }, { "epoch": 0.010840842875533573, "grad_norm": 18.726655960083008, "learning_rate": 2.12e-05, "loss": 1.7177, "step": 160 }, { "epoch": 0.010908598143505657, "grad_norm": 17.22323989868164, "learning_rate": 2.1333333333333335e-05, "loss": 1.5452, "step": 161 }, { "epoch": 0.010976353411477742, "grad_norm": 14.823019027709961, "learning_rate": 2.146666666666667e-05, "loss": 1.2856, "step": 162 }, { "epoch": 0.011044108679449827, "grad_norm": 14.756085395812988, "learning_rate": 2.16e-05, "loss": 1.4942, "step": 163 }, { "epoch": 0.011111863947421911, "grad_norm": 18.523683547973633, "learning_rate": 2.1733333333333334e-05, "loss": 1.4832, "step": 164 }, { "epoch": 0.011179619215393996, "grad_norm": 16.75520896911621, "learning_rate": 2.186666666666667e-05, "loss": 1.1847, "step": 165 }, { "epoch": 0.011247374483366082, "grad_norm": 17.045665740966797, "learning_rate": 2.2000000000000003e-05, "loss": 1.6358, "step": 166 }, { "epoch": 0.011315129751338167, "grad_norm": 17.235095977783203, "learning_rate": 2.2133333333333334e-05, "loss": 1.6984, "step": 167 }, { "epoch": 0.011382885019310252, "grad_norm": 18.120485305786133, "learning_rate": 2.2266666666666668e-05, "loss": 1.5544, "step": 168 }, { "epoch": 0.011450640287282337, "grad_norm": 15.252264976501465, "learning_rate": 2.2400000000000002e-05, "loss": 1.4386, "step": 169 }, { "epoch": 0.011518395555254421, "grad_norm": 13.91727066040039, "learning_rate": 2.2533333333333333e-05, "loss": 1.2083, "step": 170 }, { "epoch": 0.011586150823226506, "grad_norm": 16.50945472717285, "learning_rate": 2.2666666666666668e-05, "loss": 1.5878, "step": 171 }, { "epoch": 0.01165390609119859, "grad_norm": 12.260165214538574, "learning_rate": 2.2800000000000002e-05, "loss": 1.4173, "step": 172 }, { "epoch": 0.011721661359170675, "grad_norm": 17.12419891357422, "learning_rate": 2.2933333333333333e-05, "loss": 1.3117, "step": 173 }, { "epoch": 0.01178941662714276, "grad_norm": 17.508407592773438, "learning_rate": 2.3066666666666667e-05, "loss": 1.4294, "step": 174 }, { "epoch": 0.011857171895114845, "grad_norm": 17.78769302368164, "learning_rate": 2.32e-05, "loss": 1.8437, "step": 175 }, { "epoch": 0.01192492716308693, "grad_norm": 13.79132080078125, "learning_rate": 2.3333333333333336e-05, "loss": 1.3969, "step": 176 }, { "epoch": 0.011992682431059014, "grad_norm": 16.519269943237305, "learning_rate": 2.3466666666666667e-05, "loss": 1.5833, "step": 177 }, { "epoch": 0.012060437699031099, "grad_norm": 15.920942306518555, "learning_rate": 2.36e-05, "loss": 1.4783, "step": 178 }, { "epoch": 0.012128192967003185, "grad_norm": 17.00782585144043, "learning_rate": 2.3733333333333335e-05, "loss": 1.4038, "step": 179 }, { "epoch": 0.01219594823497527, "grad_norm": 18.844655990600586, "learning_rate": 2.3866666666666666e-05, "loss": 1.653, "step": 180 }, { "epoch": 0.012263703502947355, "grad_norm": 18.05002784729004, "learning_rate": 2.4e-05, "loss": 1.4137, "step": 181 }, { "epoch": 0.01233145877091944, "grad_norm": 14.200697898864746, "learning_rate": 2.4133333333333335e-05, "loss": 1.3067, "step": 182 }, { "epoch": 0.012399214038891524, "grad_norm": 16.28152084350586, "learning_rate": 2.426666666666667e-05, "loss": 1.3425, "step": 183 }, { "epoch": 0.012466969306863609, "grad_norm": 18.691898345947266, "learning_rate": 2.44e-05, "loss": 1.5995, "step": 184 }, { "epoch": 0.012534724574835693, "grad_norm": 16.62773895263672, "learning_rate": 2.4533333333333334e-05, "loss": 1.5722, "step": 185 }, { "epoch": 0.012602479842807778, "grad_norm": 19.650226593017578, "learning_rate": 2.466666666666667e-05, "loss": 1.5445, "step": 186 }, { "epoch": 0.012670235110779863, "grad_norm": 14.582402229309082, "learning_rate": 2.48e-05, "loss": 1.4822, "step": 187 }, { "epoch": 0.012737990378751948, "grad_norm": 16.5295352935791, "learning_rate": 2.4933333333333334e-05, "loss": 1.5981, "step": 188 }, { "epoch": 0.012805745646724032, "grad_norm": 17.096031188964844, "learning_rate": 2.5066666666666665e-05, "loss": 1.2527, "step": 189 }, { "epoch": 0.012873500914696117, "grad_norm": 15.424555778503418, "learning_rate": 2.5200000000000003e-05, "loss": 1.5284, "step": 190 }, { "epoch": 0.012941256182668202, "grad_norm": 14.505122184753418, "learning_rate": 2.5333333333333337e-05, "loss": 1.5726, "step": 191 }, { "epoch": 0.013009011450640288, "grad_norm": 13.430007934570312, "learning_rate": 2.5466666666666668e-05, "loss": 1.5905, "step": 192 }, { "epoch": 0.013076766718612373, "grad_norm": 12.291808128356934, "learning_rate": 2.5600000000000002e-05, "loss": 1.3122, "step": 193 }, { "epoch": 0.013144521986584458, "grad_norm": 16.053747177124023, "learning_rate": 2.5733333333333337e-05, "loss": 1.3164, "step": 194 }, { "epoch": 0.013212277254556542, "grad_norm": 17.762910842895508, "learning_rate": 2.5866666666666667e-05, "loss": 1.4727, "step": 195 }, { "epoch": 0.013280032522528627, "grad_norm": 15.695332527160645, "learning_rate": 2.6000000000000002e-05, "loss": 1.5584, "step": 196 }, { "epoch": 0.013347787790500712, "grad_norm": 14.789410591125488, "learning_rate": 2.6133333333333333e-05, "loss": 1.3253, "step": 197 }, { "epoch": 0.013415543058472796, "grad_norm": 15.248231887817383, "learning_rate": 2.6266666666666667e-05, "loss": 1.3087, "step": 198 }, { "epoch": 0.013483298326444881, "grad_norm": 18.894878387451172, "learning_rate": 2.64e-05, "loss": 1.2755, "step": 199 }, { "epoch": 0.013551053594416966, "grad_norm": 16.611616134643555, "learning_rate": 2.6533333333333332e-05, "loss": 1.4218, "step": 200 }, { "epoch": 0.01361880886238905, "grad_norm": 17.056982040405273, "learning_rate": 2.6666666666666667e-05, "loss": 1.4376, "step": 201 }, { "epoch": 0.013686564130361135, "grad_norm": 15.91457748413086, "learning_rate": 2.6800000000000004e-05, "loss": 1.4836, "step": 202 }, { "epoch": 0.01375431939833322, "grad_norm": 16.371660232543945, "learning_rate": 2.6933333333333332e-05, "loss": 1.7374, "step": 203 }, { "epoch": 0.013822074666305304, "grad_norm": 14.45882511138916, "learning_rate": 2.706666666666667e-05, "loss": 1.3324, "step": 204 }, { "epoch": 0.013889829934277391, "grad_norm": 18.779190063476562, "learning_rate": 2.7200000000000004e-05, "loss": 1.4845, "step": 205 }, { "epoch": 0.013957585202249476, "grad_norm": 19.7078914642334, "learning_rate": 2.733333333333333e-05, "loss": 1.6681, "step": 206 }, { "epoch": 0.01402534047022156, "grad_norm": 20.498266220092773, "learning_rate": 2.746666666666667e-05, "loss": 1.559, "step": 207 }, { "epoch": 0.014093095738193645, "grad_norm": 20.2978515625, "learning_rate": 2.7600000000000003e-05, "loss": 1.5821, "step": 208 }, { "epoch": 0.01416085100616573, "grad_norm": 18.85460090637207, "learning_rate": 2.7733333333333334e-05, "loss": 1.3, "step": 209 }, { "epoch": 0.014228606274137814, "grad_norm": 13.912948608398438, "learning_rate": 2.786666666666667e-05, "loss": 1.4833, "step": 210 }, { "epoch": 0.014296361542109899, "grad_norm": 19.084806442260742, "learning_rate": 2.8000000000000003e-05, "loss": 1.3379, "step": 211 }, { "epoch": 0.014364116810081984, "grad_norm": 15.123217582702637, "learning_rate": 2.8133333333333334e-05, "loss": 1.354, "step": 212 }, { "epoch": 0.014431872078054069, "grad_norm": 14.127299308776855, "learning_rate": 2.8266666666666668e-05, "loss": 1.2341, "step": 213 }, { "epoch": 0.014499627346026153, "grad_norm": 15.609139442443848, "learning_rate": 2.84e-05, "loss": 1.3474, "step": 214 }, { "epoch": 0.014567382613998238, "grad_norm": 20.588394165039062, "learning_rate": 2.8533333333333333e-05, "loss": 1.4881, "step": 215 }, { "epoch": 0.014635137881970323, "grad_norm": 12.27170467376709, "learning_rate": 2.8666666666666668e-05, "loss": 1.2255, "step": 216 }, { "epoch": 0.014702893149942407, "grad_norm": 18.14453125, "learning_rate": 2.88e-05, "loss": 1.6328, "step": 217 }, { "epoch": 0.014770648417914492, "grad_norm": 17.722230911254883, "learning_rate": 2.8933333333333333e-05, "loss": 2.0172, "step": 218 }, { "epoch": 0.014838403685886578, "grad_norm": 14.454687118530273, "learning_rate": 2.906666666666667e-05, "loss": 1.2936, "step": 219 }, { "epoch": 0.014906158953858663, "grad_norm": 16.369625091552734, "learning_rate": 2.9199999999999998e-05, "loss": 1.6289, "step": 220 }, { "epoch": 0.014973914221830748, "grad_norm": 17.64832305908203, "learning_rate": 2.9333333333333336e-05, "loss": 1.3135, "step": 221 }, { "epoch": 0.015041669489802833, "grad_norm": 14.401778221130371, "learning_rate": 2.946666666666667e-05, "loss": 1.2614, "step": 222 }, { "epoch": 0.015109424757774917, "grad_norm": 17.145814895629883, "learning_rate": 2.96e-05, "loss": 1.3473, "step": 223 }, { "epoch": 0.015177180025747002, "grad_norm": 17.273954391479492, "learning_rate": 2.9733333333333336e-05, "loss": 1.3747, "step": 224 }, { "epoch": 0.015244935293719087, "grad_norm": 17.190818786621094, "learning_rate": 2.986666666666667e-05, "loss": 1.5452, "step": 225 }, { "epoch": 0.015312690561691171, "grad_norm": 19.573585510253906, "learning_rate": 3e-05, "loss": 1.4942, "step": 226 }, { "epoch": 0.015380445829663256, "grad_norm": 15.349540710449219, "learning_rate": 3.0133333333333335e-05, "loss": 1.4277, "step": 227 }, { "epoch": 0.01544820109763534, "grad_norm": 19.786211013793945, "learning_rate": 3.0266666666666666e-05, "loss": 1.5014, "step": 228 }, { "epoch": 0.015515956365607425, "grad_norm": 17.51229476928711, "learning_rate": 3.04e-05, "loss": 1.695, "step": 229 }, { "epoch": 0.01558371163357951, "grad_norm": 19.894756317138672, "learning_rate": 3.0533333333333335e-05, "loss": 1.4994, "step": 230 }, { "epoch": 0.015651466901551597, "grad_norm": 13.575235366821289, "learning_rate": 3.066666666666667e-05, "loss": 1.4037, "step": 231 }, { "epoch": 0.01571922216952368, "grad_norm": 18.516822814941406, "learning_rate": 3.08e-05, "loss": 1.5494, "step": 232 }, { "epoch": 0.015786977437495766, "grad_norm": 17.343368530273438, "learning_rate": 3.093333333333334e-05, "loss": 1.1708, "step": 233 }, { "epoch": 0.01585473270546785, "grad_norm": 18.78635597229004, "learning_rate": 3.1066666666666665e-05, "loss": 1.6543, "step": 234 }, { "epoch": 0.015922487973439935, "grad_norm": 16.219989776611328, "learning_rate": 3.12e-05, "loss": 1.381, "step": 235 }, { "epoch": 0.01599024324141202, "grad_norm": 15.987724304199219, "learning_rate": 3.1333333333333334e-05, "loss": 1.3981, "step": 236 }, { "epoch": 0.016057998509384105, "grad_norm": 17.251537322998047, "learning_rate": 3.146666666666667e-05, "loss": 1.8181, "step": 237 }, { "epoch": 0.01612575377735619, "grad_norm": 15.804817199707031, "learning_rate": 3.16e-05, "loss": 1.4755, "step": 238 }, { "epoch": 0.016193509045328274, "grad_norm": 16.508943557739258, "learning_rate": 3.173333333333334e-05, "loss": 1.2167, "step": 239 }, { "epoch": 0.01626126431330036, "grad_norm": 17.382783889770508, "learning_rate": 3.1866666666666664e-05, "loss": 1.6186, "step": 240 }, { "epoch": 0.016329019581272444, "grad_norm": 17.938161849975586, "learning_rate": 3.2000000000000005e-05, "loss": 1.3685, "step": 241 }, { "epoch": 0.016396774849244528, "grad_norm": 21.103981018066406, "learning_rate": 3.213333333333334e-05, "loss": 1.6275, "step": 242 }, { "epoch": 0.016464530117216613, "grad_norm": 15.145224571228027, "learning_rate": 3.226666666666667e-05, "loss": 1.2058, "step": 243 }, { "epoch": 0.016532285385188698, "grad_norm": 15.368849754333496, "learning_rate": 3.24e-05, "loss": 1.3436, "step": 244 }, { "epoch": 0.016600040653160782, "grad_norm": 16.95122718811035, "learning_rate": 3.253333333333333e-05, "loss": 1.4346, "step": 245 }, { "epoch": 0.016667795921132867, "grad_norm": 14.612030029296875, "learning_rate": 3.266666666666667e-05, "loss": 1.4759, "step": 246 }, { "epoch": 0.016735551189104952, "grad_norm": 17.6989688873291, "learning_rate": 3.2800000000000004e-05, "loss": 1.3612, "step": 247 }, { "epoch": 0.016803306457077036, "grad_norm": 19.985427856445312, "learning_rate": 3.293333333333333e-05, "loss": 1.2718, "step": 248 }, { "epoch": 0.01687106172504912, "grad_norm": 12.162117958068848, "learning_rate": 3.3066666666666666e-05, "loss": 1.2133, "step": 249 }, { "epoch": 0.016938816993021206, "grad_norm": 19.053850173950195, "learning_rate": 3.32e-05, "loss": 1.1345, "step": 250 }, { "epoch": 0.01700657226099329, "grad_norm": 15.729182243347168, "learning_rate": 3.3333333333333335e-05, "loss": 1.3079, "step": 251 }, { "epoch": 0.01707432752896538, "grad_norm": 18.984880447387695, "learning_rate": 3.346666666666667e-05, "loss": 1.5124, "step": 252 }, { "epoch": 0.017142082796937463, "grad_norm": 14.444266319274902, "learning_rate": 3.3600000000000004e-05, "loss": 1.4279, "step": 253 }, { "epoch": 0.017209838064909548, "grad_norm": 17.174089431762695, "learning_rate": 3.373333333333333e-05, "loss": 1.1192, "step": 254 }, { "epoch": 0.017277593332881633, "grad_norm": 15.885781288146973, "learning_rate": 3.3866666666666665e-05, "loss": 1.3918, "step": 255 }, { "epoch": 0.017345348600853717, "grad_norm": 14.934944152832031, "learning_rate": 3.4000000000000007e-05, "loss": 1.3348, "step": 256 }, { "epoch": 0.017413103868825802, "grad_norm": 17.977643966674805, "learning_rate": 3.4133333333333334e-05, "loss": 1.392, "step": 257 }, { "epoch": 0.017480859136797887, "grad_norm": 17.527130126953125, "learning_rate": 3.426666666666667e-05, "loss": 1.2939, "step": 258 }, { "epoch": 0.01754861440476997, "grad_norm": 17.823453903198242, "learning_rate": 3.4399999999999996e-05, "loss": 1.3118, "step": 259 }, { "epoch": 0.017616369672742056, "grad_norm": 15.746111869812012, "learning_rate": 3.453333333333334e-05, "loss": 1.3865, "step": 260 }, { "epoch": 0.01768412494071414, "grad_norm": 18.798080444335938, "learning_rate": 3.466666666666667e-05, "loss": 1.399, "step": 261 }, { "epoch": 0.017751880208686226, "grad_norm": 15.139904975891113, "learning_rate": 3.48e-05, "loss": 1.5248, "step": 262 }, { "epoch": 0.01781963547665831, "grad_norm": 16.150434494018555, "learning_rate": 3.493333333333333e-05, "loss": 1.5033, "step": 263 }, { "epoch": 0.017887390744630395, "grad_norm": 19.125490188598633, "learning_rate": 3.506666666666667e-05, "loss": 1.4695, "step": 264 }, { "epoch": 0.01795514601260248, "grad_norm": 15.708895683288574, "learning_rate": 3.52e-05, "loss": 1.6189, "step": 265 }, { "epoch": 0.018022901280574564, "grad_norm": 14.185934066772461, "learning_rate": 3.5333333333333336e-05, "loss": 1.3536, "step": 266 }, { "epoch": 0.01809065654854665, "grad_norm": 15.22371768951416, "learning_rate": 3.546666666666667e-05, "loss": 1.4652, "step": 267 }, { "epoch": 0.018158411816518734, "grad_norm": 15.854277610778809, "learning_rate": 3.56e-05, "loss": 1.7211, "step": 268 }, { "epoch": 0.01822616708449082, "grad_norm": 18.41203498840332, "learning_rate": 3.573333333333333e-05, "loss": 1.54, "step": 269 }, { "epoch": 0.018293922352462903, "grad_norm": 15.323198318481445, "learning_rate": 3.586666666666667e-05, "loss": 1.3582, "step": 270 }, { "epoch": 0.018361677620434988, "grad_norm": 15.056265830993652, "learning_rate": 3.6e-05, "loss": 1.2959, "step": 271 }, { "epoch": 0.018429432888407073, "grad_norm": 18.912555694580078, "learning_rate": 3.6133333333333335e-05, "loss": 1.5964, "step": 272 }, { "epoch": 0.018497188156379157, "grad_norm": 15.093083381652832, "learning_rate": 3.626666666666667e-05, "loss": 1.3403, "step": 273 }, { "epoch": 0.018564943424351242, "grad_norm": 17.38453483581543, "learning_rate": 3.6400000000000004e-05, "loss": 1.5082, "step": 274 }, { "epoch": 0.018632698692323327, "grad_norm": 16.53204345703125, "learning_rate": 3.653333333333334e-05, "loss": 1.2373, "step": 275 }, { "epoch": 0.01870045396029541, "grad_norm": 13.423516273498535, "learning_rate": 3.6666666666666666e-05, "loss": 1.3111, "step": 276 }, { "epoch": 0.018768209228267496, "grad_norm": 14.92001724243164, "learning_rate": 3.68e-05, "loss": 1.5767, "step": 277 }, { "epoch": 0.018835964496239584, "grad_norm": 15.229578971862793, "learning_rate": 3.6933333333333334e-05, "loss": 1.19, "step": 278 }, { "epoch": 0.01890371976421167, "grad_norm": 17.159273147583008, "learning_rate": 3.706666666666667e-05, "loss": 1.335, "step": 279 }, { "epoch": 0.018971475032183754, "grad_norm": 13.7230806350708, "learning_rate": 3.72e-05, "loss": 1.2587, "step": 280 }, { "epoch": 0.01903923030015584, "grad_norm": 15.375404357910156, "learning_rate": 3.733333333333334e-05, "loss": 1.2744, "step": 281 }, { "epoch": 0.019106985568127923, "grad_norm": 21.054534912109375, "learning_rate": 3.7466666666666665e-05, "loss": 1.5581, "step": 282 }, { "epoch": 0.019174740836100008, "grad_norm": 17.954307556152344, "learning_rate": 3.76e-05, "loss": 1.6617, "step": 283 }, { "epoch": 0.019242496104072093, "grad_norm": 17.954139709472656, "learning_rate": 3.773333333333334e-05, "loss": 1.7486, "step": 284 }, { "epoch": 0.019310251372044177, "grad_norm": 16.210704803466797, "learning_rate": 3.786666666666667e-05, "loss": 1.661, "step": 285 }, { "epoch": 0.019378006640016262, "grad_norm": 18.14916229248047, "learning_rate": 3.8e-05, "loss": 1.8154, "step": 286 }, { "epoch": 0.019445761907988347, "grad_norm": 16.372896194458008, "learning_rate": 3.8133333333333336e-05, "loss": 1.5464, "step": 287 }, { "epoch": 0.01951351717596043, "grad_norm": 16.4403076171875, "learning_rate": 3.8266666666666664e-05, "loss": 1.0815, "step": 288 }, { "epoch": 0.019581272443932516, "grad_norm": 16.612815856933594, "learning_rate": 3.8400000000000005e-05, "loss": 1.4338, "step": 289 }, { "epoch": 0.0196490277119046, "grad_norm": 17.0092716217041, "learning_rate": 3.853333333333334e-05, "loss": 1.5257, "step": 290 }, { "epoch": 0.019716782979876685, "grad_norm": 13.920248031616211, "learning_rate": 3.866666666666667e-05, "loss": 1.4595, "step": 291 }, { "epoch": 0.01978453824784877, "grad_norm": 12.699677467346191, "learning_rate": 3.88e-05, "loss": 1.2602, "step": 292 }, { "epoch": 0.019852293515820855, "grad_norm": 17.31343650817871, "learning_rate": 3.8933333333333336e-05, "loss": 1.2934, "step": 293 }, { "epoch": 0.01992004878379294, "grad_norm": 16.7819881439209, "learning_rate": 3.906666666666667e-05, "loss": 1.5385, "step": 294 }, { "epoch": 0.019987804051765024, "grad_norm": 14.175074577331543, "learning_rate": 3.9200000000000004e-05, "loss": 1.1644, "step": 295 }, { "epoch": 0.02005555931973711, "grad_norm": 22.863780975341797, "learning_rate": 3.933333333333333e-05, "loss": 1.5932, "step": 296 }, { "epoch": 0.020123314587709194, "grad_norm": 15.278692245483398, "learning_rate": 3.9466666666666666e-05, "loss": 1.3997, "step": 297 }, { "epoch": 0.02019106985568128, "grad_norm": 15.878965377807617, "learning_rate": 3.960000000000001e-05, "loss": 1.5992, "step": 298 }, { "epoch": 0.020258825123653363, "grad_norm": 17.444494247436523, "learning_rate": 3.9733333333333335e-05, "loss": 1.5224, "step": 299 }, { "epoch": 0.020326580391625448, "grad_norm": 15.011860847473145, "learning_rate": 3.986666666666667e-05, "loss": 1.2609, "step": 300 }, { "epoch": 0.020394335659597532, "grad_norm": 17.057064056396484, "learning_rate": 4e-05, "loss": 1.2724, "step": 301 }, { "epoch": 0.020462090927569617, "grad_norm": 14.946621894836426, "learning_rate": 4.013333333333333e-05, "loss": 1.273, "step": 302 }, { "epoch": 0.020529846195541702, "grad_norm": 15.072589874267578, "learning_rate": 4.026666666666667e-05, "loss": 1.2029, "step": 303 }, { "epoch": 0.020597601463513786, "grad_norm": 14.304384231567383, "learning_rate": 4.0400000000000006e-05, "loss": 1.3601, "step": 304 }, { "epoch": 0.020665356731485875, "grad_norm": 16.18627166748047, "learning_rate": 4.0533333333333334e-05, "loss": 1.5844, "step": 305 }, { "epoch": 0.02073311199945796, "grad_norm": 14.144999504089355, "learning_rate": 4.066666666666667e-05, "loss": 1.1638, "step": 306 }, { "epoch": 0.020800867267430044, "grad_norm": 13.793155670166016, "learning_rate": 4.08e-05, "loss": 1.2101, "step": 307 }, { "epoch": 0.02086862253540213, "grad_norm": 19.344085693359375, "learning_rate": 4.093333333333334e-05, "loss": 1.5899, "step": 308 }, { "epoch": 0.020936377803374213, "grad_norm": 13.670002937316895, "learning_rate": 4.106666666666667e-05, "loss": 1.4271, "step": 309 }, { "epoch": 0.021004133071346298, "grad_norm": 17.72041893005371, "learning_rate": 4.12e-05, "loss": 1.5276, "step": 310 }, { "epoch": 0.021071888339318383, "grad_norm": 14.103938102722168, "learning_rate": 4.133333333333333e-05, "loss": 1.2355, "step": 311 }, { "epoch": 0.021139643607290468, "grad_norm": 16.439529418945312, "learning_rate": 4.146666666666667e-05, "loss": 1.3393, "step": 312 }, { "epoch": 0.021207398875262552, "grad_norm": 13.821382522583008, "learning_rate": 4.16e-05, "loss": 1.275, "step": 313 }, { "epoch": 0.021275154143234637, "grad_norm": 19.987564086914062, "learning_rate": 4.1733333333333336e-05, "loss": 1.415, "step": 314 }, { "epoch": 0.02134290941120672, "grad_norm": 12.660404205322266, "learning_rate": 4.186666666666667e-05, "loss": 1.2378, "step": 315 }, { "epoch": 0.021410664679178806, "grad_norm": 14.009933471679688, "learning_rate": 4.2e-05, "loss": 1.1615, "step": 316 }, { "epoch": 0.02147841994715089, "grad_norm": 16.11932945251465, "learning_rate": 4.213333333333334e-05, "loss": 1.0653, "step": 317 }, { "epoch": 0.021546175215122976, "grad_norm": 15.943892478942871, "learning_rate": 4.226666666666667e-05, "loss": 1.6208, "step": 318 }, { "epoch": 0.02161393048309506, "grad_norm": 16.01548194885254, "learning_rate": 4.24e-05, "loss": 1.4052, "step": 319 }, { "epoch": 0.021681685751067145, "grad_norm": 15.703156471252441, "learning_rate": 4.2533333333333335e-05, "loss": 1.3736, "step": 320 }, { "epoch": 0.02174944101903923, "grad_norm": 20.367021560668945, "learning_rate": 4.266666666666667e-05, "loss": 1.5758, "step": 321 }, { "epoch": 0.021817196287011315, "grad_norm": 15.678637504577637, "learning_rate": 4.2800000000000004e-05, "loss": 1.4458, "step": 322 }, { "epoch": 0.0218849515549834, "grad_norm": 15.707669258117676, "learning_rate": 4.293333333333334e-05, "loss": 1.3065, "step": 323 }, { "epoch": 0.021952706822955484, "grad_norm": 16.07583999633789, "learning_rate": 4.3066666666666665e-05, "loss": 1.5875, "step": 324 }, { "epoch": 0.02202046209092757, "grad_norm": 15.248065948486328, "learning_rate": 4.32e-05, "loss": 1.5972, "step": 325 }, { "epoch": 0.022088217358899653, "grad_norm": 17.77379608154297, "learning_rate": 4.3333333333333334e-05, "loss": 1.3385, "step": 326 }, { "epoch": 0.022155972626871738, "grad_norm": 24.381973266601562, "learning_rate": 4.346666666666667e-05, "loss": 1.6428, "step": 327 }, { "epoch": 0.022223727894843823, "grad_norm": 17.3684024810791, "learning_rate": 4.36e-05, "loss": 1.3634, "step": 328 }, { "epoch": 0.022291483162815907, "grad_norm": 20.45623207092285, "learning_rate": 4.373333333333334e-05, "loss": 1.3947, "step": 329 }, { "epoch": 0.022359238430787992, "grad_norm": 17.434114456176758, "learning_rate": 4.3866666666666665e-05, "loss": 1.4269, "step": 330 }, { "epoch": 0.02242699369876008, "grad_norm": 13.574700355529785, "learning_rate": 4.4000000000000006e-05, "loss": 1.2807, "step": 331 }, { "epoch": 0.022494748966732165, "grad_norm": 16.404966354370117, "learning_rate": 4.413333333333334e-05, "loss": 1.4549, "step": 332 }, { "epoch": 0.02256250423470425, "grad_norm": 14.14947509765625, "learning_rate": 4.426666666666667e-05, "loss": 1.4187, "step": 333 }, { "epoch": 0.022630259502676334, "grad_norm": 17.632492065429688, "learning_rate": 4.44e-05, "loss": 1.3515, "step": 334 }, { "epoch": 0.02269801477064842, "grad_norm": 18.149261474609375, "learning_rate": 4.4533333333333336e-05, "loss": 1.4875, "step": 335 }, { "epoch": 0.022765770038620504, "grad_norm": 19.864049911499023, "learning_rate": 4.466666666666667e-05, "loss": 1.5622, "step": 336 }, { "epoch": 0.02283352530659259, "grad_norm": 16.7431583404541, "learning_rate": 4.4800000000000005e-05, "loss": 1.4344, "step": 337 }, { "epoch": 0.022901280574564673, "grad_norm": 16.53998565673828, "learning_rate": 4.493333333333333e-05, "loss": 1.1745, "step": 338 }, { "epoch": 0.022969035842536758, "grad_norm": 17.786945343017578, "learning_rate": 4.5066666666666667e-05, "loss": 1.4335, "step": 339 }, { "epoch": 0.023036791110508843, "grad_norm": 17.730606079101562, "learning_rate": 4.52e-05, "loss": 1.341, "step": 340 }, { "epoch": 0.023104546378480927, "grad_norm": 15.536660194396973, "learning_rate": 4.5333333333333335e-05, "loss": 1.4935, "step": 341 }, { "epoch": 0.023172301646453012, "grad_norm": 22.405893325805664, "learning_rate": 4.546666666666667e-05, "loss": 1.5538, "step": 342 }, { "epoch": 0.023240056914425097, "grad_norm": 15.963828086853027, "learning_rate": 4.5600000000000004e-05, "loss": 1.373, "step": 343 }, { "epoch": 0.02330781218239718, "grad_norm": 13.041223526000977, "learning_rate": 4.573333333333333e-05, "loss": 1.346, "step": 344 }, { "epoch": 0.023375567450369266, "grad_norm": 14.552989959716797, "learning_rate": 4.5866666666666666e-05, "loss": 1.2723, "step": 345 }, { "epoch": 0.02344332271834135, "grad_norm": 16.034404754638672, "learning_rate": 4.600000000000001e-05, "loss": 1.3671, "step": 346 }, { "epoch": 0.023511077986313435, "grad_norm": 17.5505428314209, "learning_rate": 4.6133333333333334e-05, "loss": 1.4553, "step": 347 }, { "epoch": 0.02357883325428552, "grad_norm": 14.029112815856934, "learning_rate": 4.626666666666667e-05, "loss": 1.3639, "step": 348 }, { "epoch": 0.023646588522257605, "grad_norm": 14.633806228637695, "learning_rate": 4.64e-05, "loss": 1.61, "step": 349 }, { "epoch": 0.02371434379022969, "grad_norm": 15.407472610473633, "learning_rate": 4.653333333333334e-05, "loss": 1.2449, "step": 350 }, { "epoch": 0.023782099058201774, "grad_norm": 13.801481246948242, "learning_rate": 4.666666666666667e-05, "loss": 1.3105, "step": 351 }, { "epoch": 0.02384985432617386, "grad_norm": 17.675159454345703, "learning_rate": 4.6800000000000006e-05, "loss": 1.3142, "step": 352 }, { "epoch": 0.023917609594145944, "grad_norm": 15.383625984191895, "learning_rate": 4.6933333333333333e-05, "loss": 1.2976, "step": 353 }, { "epoch": 0.02398536486211803, "grad_norm": 18.087081909179688, "learning_rate": 4.706666666666667e-05, "loss": 1.3391, "step": 354 }, { "epoch": 0.024053120130090113, "grad_norm": 15.790519714355469, "learning_rate": 4.72e-05, "loss": 1.7103, "step": 355 }, { "epoch": 0.024120875398062198, "grad_norm": 15.931246757507324, "learning_rate": 4.7333333333333336e-05, "loss": 1.4792, "step": 356 }, { "epoch": 0.024188630666034282, "grad_norm": 18.39167022705078, "learning_rate": 4.746666666666667e-05, "loss": 1.5085, "step": 357 }, { "epoch": 0.02425638593400637, "grad_norm": 13.518149375915527, "learning_rate": 4.76e-05, "loss": 1.2711, "step": 358 }, { "epoch": 0.024324141201978455, "grad_norm": 18.02762794494629, "learning_rate": 4.773333333333333e-05, "loss": 1.4113, "step": 359 }, { "epoch": 0.02439189646995054, "grad_norm": 15.260608673095703, "learning_rate": 4.7866666666666674e-05, "loss": 1.3286, "step": 360 }, { "epoch": 0.024459651737922625, "grad_norm": 16.26275062561035, "learning_rate": 4.8e-05, "loss": 1.4436, "step": 361 }, { "epoch": 0.02452740700589471, "grad_norm": 14.213051795959473, "learning_rate": 4.8133333333333336e-05, "loss": 1.1339, "step": 362 }, { "epoch": 0.024595162273866794, "grad_norm": 16.478811264038086, "learning_rate": 4.826666666666667e-05, "loss": 1.5534, "step": 363 }, { "epoch": 0.02466291754183888, "grad_norm": 13.24108600616455, "learning_rate": 4.8400000000000004e-05, "loss": 1.4767, "step": 364 }, { "epoch": 0.024730672809810963, "grad_norm": 15.167949676513672, "learning_rate": 4.853333333333334e-05, "loss": 1.1592, "step": 365 }, { "epoch": 0.024798428077783048, "grad_norm": 18.150367736816406, "learning_rate": 4.866666666666667e-05, "loss": 1.4172, "step": 366 }, { "epoch": 0.024866183345755133, "grad_norm": 13.250176429748535, "learning_rate": 4.88e-05, "loss": 1.5022, "step": 367 }, { "epoch": 0.024933938613727218, "grad_norm": 14.780989646911621, "learning_rate": 4.8933333333333335e-05, "loss": 1.3243, "step": 368 }, { "epoch": 0.025001693881699302, "grad_norm": 13.34748649597168, "learning_rate": 4.906666666666667e-05, "loss": 1.285, "step": 369 }, { "epoch": 0.025069449149671387, "grad_norm": 14.677448272705078, "learning_rate": 4.92e-05, "loss": 1.1982, "step": 370 }, { "epoch": 0.02513720441764347, "grad_norm": 18.211145401000977, "learning_rate": 4.933333333333334e-05, "loss": 1.4018, "step": 371 }, { "epoch": 0.025204959685615556, "grad_norm": 17.68402099609375, "learning_rate": 4.9466666666666665e-05, "loss": 1.6153, "step": 372 }, { "epoch": 0.02527271495358764, "grad_norm": 14.44299030303955, "learning_rate": 4.96e-05, "loss": 1.4092, "step": 373 }, { "epoch": 0.025340470221559726, "grad_norm": 15.467754364013672, "learning_rate": 4.973333333333334e-05, "loss": 1.5647, "step": 374 }, { "epoch": 0.02540822548953181, "grad_norm": 14.037840843200684, "learning_rate": 4.986666666666667e-05, "loss": 1.2106, "step": 375 }, { "epoch": 0.025475980757503895, "grad_norm": 15.72461223602295, "learning_rate": 5e-05, "loss": 1.1832, "step": 376 }, { "epoch": 0.02554373602547598, "grad_norm": 15.03637981414795, "learning_rate": 5.013333333333333e-05, "loss": 1.5803, "step": 377 }, { "epoch": 0.025611491293448065, "grad_norm": 15.995429992675781, "learning_rate": 5.026666666666667e-05, "loss": 1.3837, "step": 378 }, { "epoch": 0.02567924656142015, "grad_norm": 17.96516227722168, "learning_rate": 5.0400000000000005e-05, "loss": 1.5114, "step": 379 }, { "epoch": 0.025747001829392234, "grad_norm": 17.960783004760742, "learning_rate": 5.053333333333333e-05, "loss": 1.4012, "step": 380 }, { "epoch": 0.02581475709736432, "grad_norm": 14.399024963378906, "learning_rate": 5.0666666666666674e-05, "loss": 1.2906, "step": 381 }, { "epoch": 0.025882512365336403, "grad_norm": 14.268051147460938, "learning_rate": 5.08e-05, "loss": 1.2354, "step": 382 }, { "epoch": 0.025950267633308488, "grad_norm": 16.070646286010742, "learning_rate": 5.0933333333333336e-05, "loss": 1.272, "step": 383 }, { "epoch": 0.026018022901280576, "grad_norm": 18.698619842529297, "learning_rate": 5.106666666666668e-05, "loss": 1.352, "step": 384 }, { "epoch": 0.02608577816925266, "grad_norm": 13.932543754577637, "learning_rate": 5.1200000000000004e-05, "loss": 1.3249, "step": 385 }, { "epoch": 0.026153533437224746, "grad_norm": 14.76308822631836, "learning_rate": 5.133333333333333e-05, "loss": 1.1807, "step": 386 }, { "epoch": 0.02622128870519683, "grad_norm": 17.602182388305664, "learning_rate": 5.146666666666667e-05, "loss": 1.4017, "step": 387 }, { "epoch": 0.026289043973168915, "grad_norm": 13.589346885681152, "learning_rate": 5.16e-05, "loss": 1.4205, "step": 388 }, { "epoch": 0.026356799241141, "grad_norm": 16.37788200378418, "learning_rate": 5.1733333333333335e-05, "loss": 1.4255, "step": 389 }, { "epoch": 0.026424554509113084, "grad_norm": 16.24977684020996, "learning_rate": 5.1866666666666676e-05, "loss": 1.2526, "step": 390 }, { "epoch": 0.02649230977708517, "grad_norm": 15.02576732635498, "learning_rate": 5.2000000000000004e-05, "loss": 1.2147, "step": 391 }, { "epoch": 0.026560065045057254, "grad_norm": 16.754850387573242, "learning_rate": 5.213333333333333e-05, "loss": 1.5725, "step": 392 }, { "epoch": 0.02662782031302934, "grad_norm": 17.120729446411133, "learning_rate": 5.2266666666666665e-05, "loss": 1.351, "step": 393 }, { "epoch": 0.026695575581001423, "grad_norm": 18.83492660522461, "learning_rate": 5.2400000000000007e-05, "loss": 1.3044, "step": 394 }, { "epoch": 0.026763330848973508, "grad_norm": 18.66061019897461, "learning_rate": 5.2533333333333334e-05, "loss": 1.4889, "step": 395 }, { "epoch": 0.026831086116945593, "grad_norm": 18.493236541748047, "learning_rate": 5.266666666666666e-05, "loss": 1.5015, "step": 396 }, { "epoch": 0.026898841384917677, "grad_norm": 19.523067474365234, "learning_rate": 5.28e-05, "loss": 1.5179, "step": 397 }, { "epoch": 0.026966596652889762, "grad_norm": 18.207366943359375, "learning_rate": 5.293333333333334e-05, "loss": 1.374, "step": 398 }, { "epoch": 0.027034351920861847, "grad_norm": 13.021437644958496, "learning_rate": 5.3066666666666665e-05, "loss": 1.0259, "step": 399 }, { "epoch": 0.02710210718883393, "grad_norm": 14.108929634094238, "learning_rate": 5.3200000000000006e-05, "loss": 1.3914, "step": 400 }, { "epoch": 0.027169862456806016, "grad_norm": 15.087890625, "learning_rate": 5.333333333333333e-05, "loss": 1.3143, "step": 401 }, { "epoch": 0.0272376177247781, "grad_norm": 16.389965057373047, "learning_rate": 5.346666666666667e-05, "loss": 1.5286, "step": 402 }, { "epoch": 0.027305372992750186, "grad_norm": 17.855867385864258, "learning_rate": 5.360000000000001e-05, "loss": 1.4221, "step": 403 }, { "epoch": 0.02737312826072227, "grad_norm": 14.016682624816895, "learning_rate": 5.3733333333333336e-05, "loss": 1.0922, "step": 404 }, { "epoch": 0.027440883528694355, "grad_norm": 16.08883285522461, "learning_rate": 5.3866666666666664e-05, "loss": 1.4932, "step": 405 }, { "epoch": 0.02750863879666644, "grad_norm": 14.883580207824707, "learning_rate": 5.4000000000000005e-05, "loss": 1.0806, "step": 406 }, { "epoch": 0.027576394064638524, "grad_norm": 13.443008422851562, "learning_rate": 5.413333333333334e-05, "loss": 1.2653, "step": 407 }, { "epoch": 0.02764414933261061, "grad_norm": 13.391244888305664, "learning_rate": 5.4266666666666667e-05, "loss": 1.2788, "step": 408 }, { "epoch": 0.027711904600582694, "grad_norm": 13.449023246765137, "learning_rate": 5.440000000000001e-05, "loss": 1.133, "step": 409 }, { "epoch": 0.027779659868554782, "grad_norm": 13.777191162109375, "learning_rate": 5.4533333333333335e-05, "loss": 1.1509, "step": 410 }, { "epoch": 0.027847415136526867, "grad_norm": 19.04452896118164, "learning_rate": 5.466666666666666e-05, "loss": 1.3306, "step": 411 }, { "epoch": 0.02791517040449895, "grad_norm": 16.583112716674805, "learning_rate": 5.4800000000000004e-05, "loss": 1.3018, "step": 412 }, { "epoch": 0.027982925672471036, "grad_norm": 16.8561954498291, "learning_rate": 5.493333333333334e-05, "loss": 1.257, "step": 413 }, { "epoch": 0.02805068094044312, "grad_norm": 15.321064949035645, "learning_rate": 5.5066666666666666e-05, "loss": 1.2498, "step": 414 }, { "epoch": 0.028118436208415205, "grad_norm": 15.825779914855957, "learning_rate": 5.520000000000001e-05, "loss": 1.2177, "step": 415 }, { "epoch": 0.02818619147638729, "grad_norm": 15.665361404418945, "learning_rate": 5.5333333333333334e-05, "loss": 1.244, "step": 416 }, { "epoch": 0.028253946744359375, "grad_norm": 15.516046524047852, "learning_rate": 5.546666666666667e-05, "loss": 1.4957, "step": 417 }, { "epoch": 0.02832170201233146, "grad_norm": 19.148746490478516, "learning_rate": 5.560000000000001e-05, "loss": 1.4346, "step": 418 }, { "epoch": 0.028389457280303544, "grad_norm": 15.215538024902344, "learning_rate": 5.573333333333334e-05, "loss": 1.081, "step": 419 }, { "epoch": 0.02845721254827563, "grad_norm": 15.612996101379395, "learning_rate": 5.5866666666666665e-05, "loss": 1.1671, "step": 420 }, { "epoch": 0.028524967816247714, "grad_norm": 13.754039764404297, "learning_rate": 5.6000000000000006e-05, "loss": 1.2679, "step": 421 }, { "epoch": 0.028592723084219798, "grad_norm": 16.350305557250977, "learning_rate": 5.613333333333334e-05, "loss": 1.3034, "step": 422 }, { "epoch": 0.028660478352191883, "grad_norm": 16.560344696044922, "learning_rate": 5.626666666666667e-05, "loss": 1.3633, "step": 423 }, { "epoch": 0.028728233620163968, "grad_norm": 20.391889572143555, "learning_rate": 5.6399999999999995e-05, "loss": 1.3533, "step": 424 }, { "epoch": 0.028795988888136052, "grad_norm": 16.13326072692871, "learning_rate": 5.6533333333333336e-05, "loss": 1.3678, "step": 425 }, { "epoch": 0.028863744156108137, "grad_norm": 21.544612884521484, "learning_rate": 5.666666666666667e-05, "loss": 1.6275, "step": 426 }, { "epoch": 0.028931499424080222, "grad_norm": 17.097408294677734, "learning_rate": 5.68e-05, "loss": 1.6119, "step": 427 }, { "epoch": 0.028999254692052306, "grad_norm": 15.699347496032715, "learning_rate": 5.693333333333334e-05, "loss": 1.2644, "step": 428 }, { "epoch": 0.02906700996002439, "grad_norm": 16.643613815307617, "learning_rate": 5.706666666666667e-05, "loss": 1.3497, "step": 429 }, { "epoch": 0.029134765227996476, "grad_norm": 20.941123962402344, "learning_rate": 5.72e-05, "loss": 1.4232, "step": 430 }, { "epoch": 0.02920252049596856, "grad_norm": 17.549013137817383, "learning_rate": 5.7333333333333336e-05, "loss": 1.3429, "step": 431 }, { "epoch": 0.029270275763940645, "grad_norm": 19.702617645263672, "learning_rate": 5.746666666666667e-05, "loss": 1.7924, "step": 432 }, { "epoch": 0.02933803103191273, "grad_norm": 16.396209716796875, "learning_rate": 5.76e-05, "loss": 1.3296, "step": 433 }, { "epoch": 0.029405786299884815, "grad_norm": 15.823278427124023, "learning_rate": 5.773333333333334e-05, "loss": 1.4228, "step": 434 }, { "epoch": 0.0294735415678569, "grad_norm": 19.61952018737793, "learning_rate": 5.7866666666666666e-05, "loss": 1.3522, "step": 435 }, { "epoch": 0.029541296835828984, "grad_norm": 14.721433639526367, "learning_rate": 5.8e-05, "loss": 1.284, "step": 436 }, { "epoch": 0.029609052103801072, "grad_norm": 16.594276428222656, "learning_rate": 5.813333333333334e-05, "loss": 1.4636, "step": 437 }, { "epoch": 0.029676807371773157, "grad_norm": 15.98005199432373, "learning_rate": 5.826666666666667e-05, "loss": 1.2825, "step": 438 }, { "epoch": 0.02974456263974524, "grad_norm": 15.501729965209961, "learning_rate": 5.8399999999999997e-05, "loss": 1.3736, "step": 439 }, { "epoch": 0.029812317907717326, "grad_norm": 18.077552795410156, "learning_rate": 5.853333333333334e-05, "loss": 1.4115, "step": 440 }, { "epoch": 0.02988007317568941, "grad_norm": 16.016721725463867, "learning_rate": 5.866666666666667e-05, "loss": 1.3177, "step": 441 }, { "epoch": 0.029947828443661496, "grad_norm": 16.39783477783203, "learning_rate": 5.88e-05, "loss": 1.4465, "step": 442 }, { "epoch": 0.03001558371163358, "grad_norm": 13.970220565795898, "learning_rate": 5.893333333333334e-05, "loss": 1.2956, "step": 443 }, { "epoch": 0.030083338979605665, "grad_norm": 16.5622615814209, "learning_rate": 5.906666666666667e-05, "loss": 1.2144, "step": 444 }, { "epoch": 0.03015109424757775, "grad_norm": 18.191911697387695, "learning_rate": 5.92e-05, "loss": 1.623, "step": 445 }, { "epoch": 0.030218849515549834, "grad_norm": 15.483613967895508, "learning_rate": 5.9333333333333343e-05, "loss": 1.3232, "step": 446 }, { "epoch": 0.03028660478352192, "grad_norm": 15.633200645446777, "learning_rate": 5.946666666666667e-05, "loss": 1.544, "step": 447 }, { "epoch": 0.030354360051494004, "grad_norm": 16.025054931640625, "learning_rate": 5.96e-05, "loss": 1.4837, "step": 448 }, { "epoch": 0.03042211531946609, "grad_norm": 15.954922676086426, "learning_rate": 5.973333333333334e-05, "loss": 1.341, "step": 449 }, { "epoch": 0.030489870587438173, "grad_norm": 14.852401733398438, "learning_rate": 5.9866666666666674e-05, "loss": 1.2346, "step": 450 }, { "epoch": 0.030557625855410258, "grad_norm": 14.887676239013672, "learning_rate": 6e-05, "loss": 1.2789, "step": 451 }, { "epoch": 0.030625381123382343, "grad_norm": 17.25469970703125, "learning_rate": 6.013333333333334e-05, "loss": 1.4446, "step": 452 }, { "epoch": 0.030693136391354427, "grad_norm": 19.356597900390625, "learning_rate": 6.026666666666667e-05, "loss": 1.3765, "step": 453 }, { "epoch": 0.030760891659326512, "grad_norm": 14.16335391998291, "learning_rate": 6.04e-05, "loss": 1.4213, "step": 454 }, { "epoch": 0.030828646927298597, "grad_norm": 15.812528610229492, "learning_rate": 6.053333333333333e-05, "loss": 1.4028, "step": 455 }, { "epoch": 0.03089640219527068, "grad_norm": 15.10727596282959, "learning_rate": 6.066666666666667e-05, "loss": 1.3841, "step": 456 }, { "epoch": 0.030964157463242766, "grad_norm": 19.91744613647461, "learning_rate": 6.08e-05, "loss": 1.4706, "step": 457 }, { "epoch": 0.03103191273121485, "grad_norm": 14.597813606262207, "learning_rate": 6.093333333333333e-05, "loss": 1.2428, "step": 458 }, { "epoch": 0.031099667999186936, "grad_norm": 14.952363014221191, "learning_rate": 6.106666666666667e-05, "loss": 1.3431, "step": 459 }, { "epoch": 0.03116742326715902, "grad_norm": 16.519468307495117, "learning_rate": 6.12e-05, "loss": 1.5399, "step": 460 }, { "epoch": 0.031235178535131105, "grad_norm": 14.55786418914795, "learning_rate": 6.133333333333334e-05, "loss": 1.3844, "step": 461 }, { "epoch": 0.03130293380310319, "grad_norm": 12.52665901184082, "learning_rate": 6.146666666666668e-05, "loss": 1.2085, "step": 462 }, { "epoch": 0.03137068907107528, "grad_norm": 17.549148559570312, "learning_rate": 6.16e-05, "loss": 1.588, "step": 463 }, { "epoch": 0.03143844433904736, "grad_norm": 14.761232376098633, "learning_rate": 6.173333333333333e-05, "loss": 1.1685, "step": 464 }, { "epoch": 0.03150619960701945, "grad_norm": 13.1014404296875, "learning_rate": 6.186666666666668e-05, "loss": 1.1718, "step": 465 }, { "epoch": 0.03157395487499153, "grad_norm": 15.998059272766113, "learning_rate": 6.2e-05, "loss": 1.3177, "step": 466 }, { "epoch": 0.03164171014296362, "grad_norm": 16.402875900268555, "learning_rate": 6.213333333333333e-05, "loss": 1.4821, "step": 467 }, { "epoch": 0.0317094654109357, "grad_norm": 12.871467590332031, "learning_rate": 6.226666666666667e-05, "loss": 1.0176, "step": 468 }, { "epoch": 0.031777220678907786, "grad_norm": 15.602563858032227, "learning_rate": 6.24e-05, "loss": 1.3418, "step": 469 }, { "epoch": 0.03184497594687987, "grad_norm": 13.5369234085083, "learning_rate": 6.253333333333333e-05, "loss": 1.4051, "step": 470 }, { "epoch": 0.031912731214851955, "grad_norm": 16.841650009155273, "learning_rate": 6.266666666666667e-05, "loss": 1.5438, "step": 471 }, { "epoch": 0.03198048648282404, "grad_norm": 14.581981658935547, "learning_rate": 6.280000000000001e-05, "loss": 1.2767, "step": 472 }, { "epoch": 0.032048241750796125, "grad_norm": 14.084460258483887, "learning_rate": 6.293333333333334e-05, "loss": 1.2224, "step": 473 }, { "epoch": 0.03211599701876821, "grad_norm": 15.54316234588623, "learning_rate": 6.306666666666668e-05, "loss": 1.6084, "step": 474 }, { "epoch": 0.032183752286740294, "grad_norm": 15.498743057250977, "learning_rate": 6.32e-05, "loss": 1.3937, "step": 475 }, { "epoch": 0.03225150755471238, "grad_norm": 17.728227615356445, "learning_rate": 6.333333333333333e-05, "loss": 1.2996, "step": 476 }, { "epoch": 0.032319262822684464, "grad_norm": 13.956122398376465, "learning_rate": 6.346666666666667e-05, "loss": 1.4472, "step": 477 }, { "epoch": 0.03238701809065655, "grad_norm": 18.578060150146484, "learning_rate": 6.36e-05, "loss": 1.4659, "step": 478 }, { "epoch": 0.03245477335862863, "grad_norm": 14.055442810058594, "learning_rate": 6.373333333333333e-05, "loss": 1.1979, "step": 479 }, { "epoch": 0.03252252862660072, "grad_norm": 15.902195930480957, "learning_rate": 6.386666666666667e-05, "loss": 1.3367, "step": 480 }, { "epoch": 0.0325902838945728, "grad_norm": 14.212138175964355, "learning_rate": 6.400000000000001e-05, "loss": 1.2716, "step": 481 }, { "epoch": 0.03265803916254489, "grad_norm": 17.711475372314453, "learning_rate": 6.413333333333334e-05, "loss": 1.3016, "step": 482 }, { "epoch": 0.03272579443051697, "grad_norm": 13.068260192871094, "learning_rate": 6.426666666666668e-05, "loss": 1.1346, "step": 483 }, { "epoch": 0.032793549698489056, "grad_norm": 17.422321319580078, "learning_rate": 6.440000000000001e-05, "loss": 1.0931, "step": 484 }, { "epoch": 0.03286130496646114, "grad_norm": 14.98038101196289, "learning_rate": 6.453333333333333e-05, "loss": 1.22, "step": 485 }, { "epoch": 0.032929060234433226, "grad_norm": 13.300479888916016, "learning_rate": 6.466666666666666e-05, "loss": 1.2647, "step": 486 }, { "epoch": 0.03299681550240531, "grad_norm": 14.611360549926758, "learning_rate": 6.48e-05, "loss": 1.2478, "step": 487 }, { "epoch": 0.033064570770377395, "grad_norm": 19.244020462036133, "learning_rate": 6.493333333333333e-05, "loss": 1.2126, "step": 488 }, { "epoch": 0.03313232603834948, "grad_norm": 16.285396575927734, "learning_rate": 6.506666666666666e-05, "loss": 1.1694, "step": 489 }, { "epoch": 0.033200081306321565, "grad_norm": 19.613059997558594, "learning_rate": 6.52e-05, "loss": 1.2118, "step": 490 }, { "epoch": 0.03326783657429365, "grad_norm": 19.19630241394043, "learning_rate": 6.533333333333334e-05, "loss": 1.3093, "step": 491 }, { "epoch": 0.033335591842265734, "grad_norm": 12.653651237487793, "learning_rate": 6.546666666666667e-05, "loss": 1.1545, "step": 492 }, { "epoch": 0.03340334711023782, "grad_norm": 13.434443473815918, "learning_rate": 6.560000000000001e-05, "loss": 1.2543, "step": 493 }, { "epoch": 0.033471102378209903, "grad_norm": 18.893598556518555, "learning_rate": 6.573333333333334e-05, "loss": 1.3632, "step": 494 }, { "epoch": 0.03353885764618199, "grad_norm": 18.339479446411133, "learning_rate": 6.586666666666666e-05, "loss": 1.4369, "step": 495 }, { "epoch": 0.03360661291415407, "grad_norm": 12.303078651428223, "learning_rate": 6.6e-05, "loss": 1.4032, "step": 496 }, { "epoch": 0.03367436818212616, "grad_norm": 17.041015625, "learning_rate": 6.613333333333333e-05, "loss": 1.2395, "step": 497 }, { "epoch": 0.03374212345009824, "grad_norm": 14.035640716552734, "learning_rate": 6.626666666666666e-05, "loss": 1.3541, "step": 498 }, { "epoch": 0.03380987871807033, "grad_norm": 16.93412208557129, "learning_rate": 6.64e-05, "loss": 1.3503, "step": 499 }, { "epoch": 0.03387763398604241, "grad_norm": 16.02039909362793, "learning_rate": 6.653333333333334e-05, "loss": 1.3226, "step": 500 }, { "epoch": 0.033945389254014496, "grad_norm": 16.061542510986328, "learning_rate": 6.666666666666667e-05, "loss": 1.433, "step": 501 }, { "epoch": 0.03401314452198658, "grad_norm": 14.810320854187012, "learning_rate": 6.680000000000001e-05, "loss": 1.3715, "step": 502 }, { "epoch": 0.03408089978995867, "grad_norm": 15.132527351379395, "learning_rate": 6.693333333333334e-05, "loss": 1.0683, "step": 503 }, { "epoch": 0.03414865505793076, "grad_norm": 13.768165588378906, "learning_rate": 6.706666666666667e-05, "loss": 1.234, "step": 504 }, { "epoch": 0.03421641032590284, "grad_norm": 16.227920532226562, "learning_rate": 6.720000000000001e-05, "loss": 1.4181, "step": 505 }, { "epoch": 0.03428416559387493, "grad_norm": 14.205772399902344, "learning_rate": 6.733333333333333e-05, "loss": 1.4342, "step": 506 }, { "epoch": 0.03435192086184701, "grad_norm": 17.451099395751953, "learning_rate": 6.746666666666666e-05, "loss": 1.3338, "step": 507 }, { "epoch": 0.034419676129819096, "grad_norm": 13.7041015625, "learning_rate": 6.76e-05, "loss": 1.3972, "step": 508 }, { "epoch": 0.03448743139779118, "grad_norm": 13.206351280212402, "learning_rate": 6.773333333333333e-05, "loss": 1.4233, "step": 509 }, { "epoch": 0.034555186665763266, "grad_norm": 11.867975234985352, "learning_rate": 6.786666666666667e-05, "loss": 1.111, "step": 510 }, { "epoch": 0.03462294193373535, "grad_norm": 19.726648330688477, "learning_rate": 6.800000000000001e-05, "loss": 1.2918, "step": 511 }, { "epoch": 0.034690697201707435, "grad_norm": 18.914016723632812, "learning_rate": 6.813333333333334e-05, "loss": 1.5998, "step": 512 }, { "epoch": 0.03475845246967952, "grad_norm": 18.065937042236328, "learning_rate": 6.826666666666667e-05, "loss": 1.4952, "step": 513 }, { "epoch": 0.034826207737651604, "grad_norm": 16.42376136779785, "learning_rate": 6.840000000000001e-05, "loss": 1.2807, "step": 514 }, { "epoch": 0.03489396300562369, "grad_norm": 15.537372589111328, "learning_rate": 6.853333333333334e-05, "loss": 1.3524, "step": 515 }, { "epoch": 0.034961718273595774, "grad_norm": 18.395238876342773, "learning_rate": 6.866666666666666e-05, "loss": 1.3022, "step": 516 }, { "epoch": 0.03502947354156786, "grad_norm": 15.80531120300293, "learning_rate": 6.879999999999999e-05, "loss": 1.4201, "step": 517 }, { "epoch": 0.03509722880953994, "grad_norm": 15.361504554748535, "learning_rate": 6.893333333333333e-05, "loss": 1.4351, "step": 518 }, { "epoch": 0.03516498407751203, "grad_norm": 12.501819610595703, "learning_rate": 6.906666666666667e-05, "loss": 1.1894, "step": 519 }, { "epoch": 0.03523273934548411, "grad_norm": 14.53589916229248, "learning_rate": 6.92e-05, "loss": 1.3435, "step": 520 }, { "epoch": 0.0353004946134562, "grad_norm": 15.057633399963379, "learning_rate": 6.933333333333334e-05, "loss": 1.5467, "step": 521 }, { "epoch": 0.03536824988142828, "grad_norm": 16.27166175842285, "learning_rate": 6.946666666666667e-05, "loss": 1.5171, "step": 522 }, { "epoch": 0.03543600514940037, "grad_norm": 18.050413131713867, "learning_rate": 6.96e-05, "loss": 1.2246, "step": 523 }, { "epoch": 0.03550376041737245, "grad_norm": 15.316112518310547, "learning_rate": 6.973333333333334e-05, "loss": 1.2551, "step": 524 }, { "epoch": 0.035571515685344536, "grad_norm": 15.416853904724121, "learning_rate": 6.986666666666667e-05, "loss": 1.1895, "step": 525 }, { "epoch": 0.03563927095331662, "grad_norm": 15.019004821777344, "learning_rate": 7e-05, "loss": 1.5957, "step": 526 }, { "epoch": 0.035707026221288705, "grad_norm": 15.33927059173584, "learning_rate": 7.013333333333333e-05, "loss": 1.3536, "step": 527 }, { "epoch": 0.03577478148926079, "grad_norm": 16.403501510620117, "learning_rate": 7.026666666666668e-05, "loss": 1.6104, "step": 528 }, { "epoch": 0.035842536757232875, "grad_norm": 13.962714195251465, "learning_rate": 7.04e-05, "loss": 1.2122, "step": 529 }, { "epoch": 0.03591029202520496, "grad_norm": 14.530926704406738, "learning_rate": 7.053333333333334e-05, "loss": 1.2879, "step": 530 }, { "epoch": 0.035978047293177044, "grad_norm": 13.99305248260498, "learning_rate": 7.066666666666667e-05, "loss": 1.2063, "step": 531 }, { "epoch": 0.03604580256114913, "grad_norm": 17.54342269897461, "learning_rate": 7.08e-05, "loss": 1.2823, "step": 532 }, { "epoch": 0.036113557829121214, "grad_norm": 11.097670555114746, "learning_rate": 7.093333333333334e-05, "loss": 1.1658, "step": 533 }, { "epoch": 0.0361813130970933, "grad_norm": 19.069889068603516, "learning_rate": 7.106666666666667e-05, "loss": 1.3368, "step": 534 }, { "epoch": 0.03624906836506538, "grad_norm": 12.675487518310547, "learning_rate": 7.12e-05, "loss": 1.1586, "step": 535 }, { "epoch": 0.03631682363303747, "grad_norm": 16.971637725830078, "learning_rate": 7.133333333333334e-05, "loss": 1.2458, "step": 536 }, { "epoch": 0.03638457890100955, "grad_norm": 14.513422012329102, "learning_rate": 7.146666666666666e-05, "loss": 1.236, "step": 537 }, { "epoch": 0.03645233416898164, "grad_norm": 14.271836280822754, "learning_rate": 7.16e-05, "loss": 1.0951, "step": 538 }, { "epoch": 0.03652008943695372, "grad_norm": 14.286022186279297, "learning_rate": 7.173333333333335e-05, "loss": 1.2411, "step": 539 }, { "epoch": 0.03658784470492581, "grad_norm": 16.81059455871582, "learning_rate": 7.186666666666667e-05, "loss": 1.2453, "step": 540 }, { "epoch": 0.03665559997289789, "grad_norm": 16.818300247192383, "learning_rate": 7.2e-05, "loss": 1.2302, "step": 541 }, { "epoch": 0.036723355240869976, "grad_norm": 14.839925765991211, "learning_rate": 7.213333333333334e-05, "loss": 1.3445, "step": 542 }, { "epoch": 0.03679111050884206, "grad_norm": 14.917344093322754, "learning_rate": 7.226666666666667e-05, "loss": 1.3658, "step": 543 }, { "epoch": 0.036858865776814145, "grad_norm": 15.310625076293945, "learning_rate": 7.24e-05, "loss": 1.2565, "step": 544 }, { "epoch": 0.03692662104478623, "grad_norm": 16.7796630859375, "learning_rate": 7.253333333333334e-05, "loss": 1.265, "step": 545 }, { "epoch": 0.036994376312758315, "grad_norm": 19.4224796295166, "learning_rate": 7.266666666666667e-05, "loss": 1.6152, "step": 546 }, { "epoch": 0.0370621315807304, "grad_norm": 15.8001127243042, "learning_rate": 7.280000000000001e-05, "loss": 1.4313, "step": 547 }, { "epoch": 0.037129886848702484, "grad_norm": 13.597784042358398, "learning_rate": 7.293333333333334e-05, "loss": 1.3812, "step": 548 }, { "epoch": 0.03719764211667457, "grad_norm": 13.139778137207031, "learning_rate": 7.306666666666668e-05, "loss": 1.1888, "step": 549 }, { "epoch": 0.037265397384646654, "grad_norm": 17.083406448364258, "learning_rate": 7.32e-05, "loss": 1.1668, "step": 550 }, { "epoch": 0.03733315265261874, "grad_norm": 14.46076774597168, "learning_rate": 7.333333333333333e-05, "loss": 1.2394, "step": 551 }, { "epoch": 0.03740090792059082, "grad_norm": 16.217782974243164, "learning_rate": 7.346666666666667e-05, "loss": 1.3068, "step": 552 }, { "epoch": 0.03746866318856291, "grad_norm": 16.06308937072754, "learning_rate": 7.36e-05, "loss": 0.9678, "step": 553 }, { "epoch": 0.03753641845653499, "grad_norm": 14.652907371520996, "learning_rate": 7.373333333333333e-05, "loss": 1.5375, "step": 554 }, { "epoch": 0.03760417372450708, "grad_norm": 13.870722770690918, "learning_rate": 7.386666666666667e-05, "loss": 1.29, "step": 555 }, { "epoch": 0.03767192899247917, "grad_norm": 14.649571418762207, "learning_rate": 7.4e-05, "loss": 1.3588, "step": 556 }, { "epoch": 0.03773968426045125, "grad_norm": 13.698415756225586, "learning_rate": 7.413333333333334e-05, "loss": 1.0801, "step": 557 }, { "epoch": 0.03780743952842334, "grad_norm": 19.412424087524414, "learning_rate": 7.426666666666668e-05, "loss": 1.4474, "step": 558 }, { "epoch": 0.03787519479639542, "grad_norm": 18.58599853515625, "learning_rate": 7.44e-05, "loss": 1.3294, "step": 559 }, { "epoch": 0.03794295006436751, "grad_norm": 15.283289909362793, "learning_rate": 7.453333333333333e-05, "loss": 1.1554, "step": 560 }, { "epoch": 0.03801070533233959, "grad_norm": 16.289731979370117, "learning_rate": 7.466666666666667e-05, "loss": 1.222, "step": 561 }, { "epoch": 0.03807846060031168, "grad_norm": 25.105520248413086, "learning_rate": 7.48e-05, "loss": 1.2213, "step": 562 }, { "epoch": 0.03814621586828376, "grad_norm": 15.428693771362305, "learning_rate": 7.493333333333333e-05, "loss": 1.2165, "step": 563 }, { "epoch": 0.038213971136255846, "grad_norm": 18.79145622253418, "learning_rate": 7.506666666666667e-05, "loss": 1.4799, "step": 564 }, { "epoch": 0.03828172640422793, "grad_norm": 12.498862266540527, "learning_rate": 7.52e-05, "loss": 1.4348, "step": 565 }, { "epoch": 0.038349481672200016, "grad_norm": 19.302845001220703, "learning_rate": 7.533333333333334e-05, "loss": 1.2656, "step": 566 }, { "epoch": 0.0384172369401721, "grad_norm": 16.28862953186035, "learning_rate": 7.546666666666668e-05, "loss": 1.4779, "step": 567 }, { "epoch": 0.038484992208144185, "grad_norm": 13.972156524658203, "learning_rate": 7.560000000000001e-05, "loss": 1.4775, "step": 568 }, { "epoch": 0.03855274747611627, "grad_norm": 15.443683624267578, "learning_rate": 7.573333333333334e-05, "loss": 1.2671, "step": 569 }, { "epoch": 0.038620502744088354, "grad_norm": 15.203113555908203, "learning_rate": 7.586666666666668e-05, "loss": 1.4209, "step": 570 }, { "epoch": 0.03868825801206044, "grad_norm": 18.516462326049805, "learning_rate": 7.6e-05, "loss": 1.3315, "step": 571 }, { "epoch": 0.038756013280032524, "grad_norm": 13.403026580810547, "learning_rate": 7.613333333333333e-05, "loss": 1.364, "step": 572 }, { "epoch": 0.03882376854800461, "grad_norm": 12.339615821838379, "learning_rate": 7.626666666666667e-05, "loss": 1.2107, "step": 573 }, { "epoch": 0.03889152381597669, "grad_norm": 15.647716522216797, "learning_rate": 7.64e-05, "loss": 1.3165, "step": 574 }, { "epoch": 0.03895927908394878, "grad_norm": 15.854050636291504, "learning_rate": 7.653333333333333e-05, "loss": 1.517, "step": 575 }, { "epoch": 0.03902703435192086, "grad_norm": 11.765929222106934, "learning_rate": 7.666666666666667e-05, "loss": 1.2676, "step": 576 }, { "epoch": 0.03909478961989295, "grad_norm": 12.515352249145508, "learning_rate": 7.680000000000001e-05, "loss": 1.4472, "step": 577 }, { "epoch": 0.03916254488786503, "grad_norm": 11.72417163848877, "learning_rate": 7.693333333333334e-05, "loss": 1.1874, "step": 578 }, { "epoch": 0.03923030015583712, "grad_norm": 15.03148365020752, "learning_rate": 7.706666666666668e-05, "loss": 1.6381, "step": 579 }, { "epoch": 0.0392980554238092, "grad_norm": 14.0188570022583, "learning_rate": 7.72e-05, "loss": 1.4351, "step": 580 }, { "epoch": 0.039365810691781286, "grad_norm": 12.343233108520508, "learning_rate": 7.733333333333333e-05, "loss": 1.2293, "step": 581 }, { "epoch": 0.03943356595975337, "grad_norm": 15.358945846557617, "learning_rate": 7.746666666666666e-05, "loss": 1.307, "step": 582 }, { "epoch": 0.039501321227725456, "grad_norm": 14.131333351135254, "learning_rate": 7.76e-05, "loss": 1.0432, "step": 583 }, { "epoch": 0.03956907649569754, "grad_norm": 16.46926498413086, "learning_rate": 7.773333333333333e-05, "loss": 1.4126, "step": 584 }, { "epoch": 0.039636831763669625, "grad_norm": 15.209906578063965, "learning_rate": 7.786666666666667e-05, "loss": 1.3242, "step": 585 }, { "epoch": 0.03970458703164171, "grad_norm": 15.657282829284668, "learning_rate": 7.800000000000001e-05, "loss": 1.4967, "step": 586 }, { "epoch": 0.039772342299613794, "grad_norm": 11.466882705688477, "learning_rate": 7.813333333333334e-05, "loss": 1.3003, "step": 587 }, { "epoch": 0.03984009756758588, "grad_norm": 19.529300689697266, "learning_rate": 7.826666666666667e-05, "loss": 1.3835, "step": 588 }, { "epoch": 0.039907852835557964, "grad_norm": 17.22064971923828, "learning_rate": 7.840000000000001e-05, "loss": 1.2655, "step": 589 }, { "epoch": 0.03997560810353005, "grad_norm": 14.788103103637695, "learning_rate": 7.853333333333334e-05, "loss": 1.1981, "step": 590 }, { "epoch": 0.04004336337150213, "grad_norm": 13.705521583557129, "learning_rate": 7.866666666666666e-05, "loss": 1.2064, "step": 591 }, { "epoch": 0.04011111863947422, "grad_norm": 14.901930809020996, "learning_rate": 7.88e-05, "loss": 1.2632, "step": 592 }, { "epoch": 0.0401788739074463, "grad_norm": 18.520828247070312, "learning_rate": 7.893333333333333e-05, "loss": 1.4316, "step": 593 }, { "epoch": 0.04024662917541839, "grad_norm": 15.101353645324707, "learning_rate": 7.906666666666667e-05, "loss": 1.3935, "step": 594 }, { "epoch": 0.04031438444339047, "grad_norm": 13.961483001708984, "learning_rate": 7.920000000000001e-05, "loss": 1.1477, "step": 595 }, { "epoch": 0.04038213971136256, "grad_norm": 16.015554428100586, "learning_rate": 7.933333333333334e-05, "loss": 1.5029, "step": 596 }, { "epoch": 0.04044989497933464, "grad_norm": 15.009637832641602, "learning_rate": 7.946666666666667e-05, "loss": 1.3715, "step": 597 }, { "epoch": 0.040517650247306726, "grad_norm": 16.295202255249023, "learning_rate": 7.960000000000001e-05, "loss": 1.5129, "step": 598 }, { "epoch": 0.04058540551527881, "grad_norm": 88.3338623046875, "learning_rate": 7.973333333333334e-05, "loss": 1.2395, "step": 599 }, { "epoch": 0.040653160783250895, "grad_norm": 16.769424438476562, "learning_rate": 7.986666666666667e-05, "loss": 1.4258, "step": 600 }, { "epoch": 0.04072091605122298, "grad_norm": 16.044578552246094, "learning_rate": 8e-05, "loss": 1.365, "step": 601 }, { "epoch": 0.040788671319195065, "grad_norm": 15.282588958740234, "learning_rate": 8.013333333333333e-05, "loss": 1.3212, "step": 602 }, { "epoch": 0.04085642658716715, "grad_norm": 15.275490760803223, "learning_rate": 8.026666666666666e-05, "loss": 1.6048, "step": 603 }, { "epoch": 0.040924181855139234, "grad_norm": 12.952759742736816, "learning_rate": 8.04e-05, "loss": 1.1982, "step": 604 }, { "epoch": 0.04099193712311132, "grad_norm": 16.79343032836914, "learning_rate": 8.053333333333334e-05, "loss": 1.281, "step": 605 }, { "epoch": 0.041059692391083404, "grad_norm": 13.70421314239502, "learning_rate": 8.066666666666667e-05, "loss": 1.2664, "step": 606 }, { "epoch": 0.04112744765905549, "grad_norm": 14.420117378234863, "learning_rate": 8.080000000000001e-05, "loss": 1.2758, "step": 607 }, { "epoch": 0.04119520292702757, "grad_norm": 16.421335220336914, "learning_rate": 8.093333333333334e-05, "loss": 1.3105, "step": 608 }, { "epoch": 0.041262958194999665, "grad_norm": 16.96218490600586, "learning_rate": 8.106666666666667e-05, "loss": 1.4665, "step": 609 }, { "epoch": 0.04133071346297175, "grad_norm": 13.73725700378418, "learning_rate": 8.120000000000001e-05, "loss": 1.2109, "step": 610 }, { "epoch": 0.041398468730943834, "grad_norm": 14.718022346496582, "learning_rate": 8.133333333333334e-05, "loss": 1.2051, "step": 611 }, { "epoch": 0.04146622399891592, "grad_norm": 14.119061470031738, "learning_rate": 8.146666666666666e-05, "loss": 1.0731, "step": 612 }, { "epoch": 0.041533979266888, "grad_norm": 17.90053367614746, "learning_rate": 8.16e-05, "loss": 1.3804, "step": 613 }, { "epoch": 0.04160173453486009, "grad_norm": 12.727055549621582, "learning_rate": 8.173333333333335e-05, "loss": 1.0926, "step": 614 }, { "epoch": 0.04166948980283217, "grad_norm": 14.953054428100586, "learning_rate": 8.186666666666667e-05, "loss": 1.1974, "step": 615 }, { "epoch": 0.04173724507080426, "grad_norm": 16.05322265625, "learning_rate": 8.2e-05, "loss": 1.2526, "step": 616 }, { "epoch": 0.04180500033877634, "grad_norm": 15.028168678283691, "learning_rate": 8.213333333333334e-05, "loss": 1.3593, "step": 617 }, { "epoch": 0.04187275560674843, "grad_norm": 13.127458572387695, "learning_rate": 8.226666666666667e-05, "loss": 1.3981, "step": 618 }, { "epoch": 0.04194051087472051, "grad_norm": 17.323017120361328, "learning_rate": 8.24e-05, "loss": 1.4622, "step": 619 }, { "epoch": 0.042008266142692596, "grad_norm": 16.106731414794922, "learning_rate": 8.253333333333334e-05, "loss": 1.257, "step": 620 }, { "epoch": 0.04207602141066468, "grad_norm": 13.039103507995605, "learning_rate": 8.266666666666667e-05, "loss": 1.2905, "step": 621 }, { "epoch": 0.042143776678636766, "grad_norm": 13.661933898925781, "learning_rate": 8.28e-05, "loss": 1.2779, "step": 622 }, { "epoch": 0.04221153194660885, "grad_norm": 17.325756072998047, "learning_rate": 8.293333333333333e-05, "loss": 1.3934, "step": 623 }, { "epoch": 0.042279287214580935, "grad_norm": 13.898777961730957, "learning_rate": 8.306666666666668e-05, "loss": 1.3735, "step": 624 }, { "epoch": 0.04234704248255302, "grad_norm": 16.787601470947266, "learning_rate": 8.32e-05, "loss": 1.5228, "step": 625 }, { "epoch": 0.042414797750525104, "grad_norm": 13.54299259185791, "learning_rate": 8.333333333333334e-05, "loss": 1.0538, "step": 626 }, { "epoch": 0.04248255301849719, "grad_norm": 14.418194770812988, "learning_rate": 8.346666666666667e-05, "loss": 1.2932, "step": 627 }, { "epoch": 0.042550308286469274, "grad_norm": 13.39255142211914, "learning_rate": 8.36e-05, "loss": 1.1989, "step": 628 }, { "epoch": 0.04261806355444136, "grad_norm": 15.445747375488281, "learning_rate": 8.373333333333334e-05, "loss": 1.3557, "step": 629 }, { "epoch": 0.04268581882241344, "grad_norm": 13.414338111877441, "learning_rate": 8.386666666666667e-05, "loss": 1.1729, "step": 630 }, { "epoch": 0.04275357409038553, "grad_norm": 16.649137496948242, "learning_rate": 8.4e-05, "loss": 1.4212, "step": 631 }, { "epoch": 0.04282132935835761, "grad_norm": 17.583528518676758, "learning_rate": 8.413333333333334e-05, "loss": 1.4838, "step": 632 }, { "epoch": 0.0428890846263297, "grad_norm": 19.32307243347168, "learning_rate": 8.426666666666668e-05, "loss": 1.5958, "step": 633 }, { "epoch": 0.04295683989430178, "grad_norm": 12.703327178955078, "learning_rate": 8.44e-05, "loss": 1.2752, "step": 634 }, { "epoch": 0.04302459516227387, "grad_norm": 15.72768497467041, "learning_rate": 8.453333333333335e-05, "loss": 1.4059, "step": 635 }, { "epoch": 0.04309235043024595, "grad_norm": 13.532344818115234, "learning_rate": 8.466666666666667e-05, "loss": 1.2252, "step": 636 }, { "epoch": 0.043160105698218036, "grad_norm": 13.305481910705566, "learning_rate": 8.48e-05, "loss": 1.5084, "step": 637 }, { "epoch": 0.04322786096619012, "grad_norm": 11.986043930053711, "learning_rate": 8.493333333333334e-05, "loss": 1.2006, "step": 638 }, { "epoch": 0.043295616234162206, "grad_norm": 13.28003978729248, "learning_rate": 8.506666666666667e-05, "loss": 1.2442, "step": 639 }, { "epoch": 0.04336337150213429, "grad_norm": 15.835545539855957, "learning_rate": 8.52e-05, "loss": 1.2466, "step": 640 }, { "epoch": 0.043431126770106375, "grad_norm": 14.887584686279297, "learning_rate": 8.533333333333334e-05, "loss": 1.3828, "step": 641 }, { "epoch": 0.04349888203807846, "grad_norm": 14.740251541137695, "learning_rate": 8.546666666666667e-05, "loss": 1.4847, "step": 642 }, { "epoch": 0.043566637306050544, "grad_norm": 14.109213829040527, "learning_rate": 8.560000000000001e-05, "loss": 1.2251, "step": 643 }, { "epoch": 0.04363439257402263, "grad_norm": 17.33670425415039, "learning_rate": 8.573333333333333e-05, "loss": 1.3817, "step": 644 }, { "epoch": 0.043702147841994714, "grad_norm": 16.92241096496582, "learning_rate": 8.586666666666668e-05, "loss": 1.458, "step": 645 }, { "epoch": 0.0437699031099668, "grad_norm": 15.369187355041504, "learning_rate": 8.6e-05, "loss": 1.28, "step": 646 }, { "epoch": 0.04383765837793888, "grad_norm": 18.748065948486328, "learning_rate": 8.613333333333333e-05, "loss": 1.6275, "step": 647 }, { "epoch": 0.04390541364591097, "grad_norm": 15.650605201721191, "learning_rate": 8.626666666666667e-05, "loss": 1.3742, "step": 648 }, { "epoch": 0.04397316891388305, "grad_norm": 16.170730590820312, "learning_rate": 8.64e-05, "loss": 1.174, "step": 649 }, { "epoch": 0.04404092418185514, "grad_norm": 12.61952018737793, "learning_rate": 8.653333333333333e-05, "loss": 1.2178, "step": 650 }, { "epoch": 0.04410867944982722, "grad_norm": 18.86512565612793, "learning_rate": 8.666666666666667e-05, "loss": 1.4993, "step": 651 }, { "epoch": 0.04417643471779931, "grad_norm": 16.412899017333984, "learning_rate": 8.680000000000001e-05, "loss": 1.2663, "step": 652 }, { "epoch": 0.04424418998577139, "grad_norm": 14.11953353881836, "learning_rate": 8.693333333333334e-05, "loss": 1.3772, "step": 653 }, { "epoch": 0.044311945253743476, "grad_norm": 17.232633590698242, "learning_rate": 8.706666666666668e-05, "loss": 1.3252, "step": 654 }, { "epoch": 0.04437970052171556, "grad_norm": 13.833864212036133, "learning_rate": 8.72e-05, "loss": 1.279, "step": 655 }, { "epoch": 0.044447455789687645, "grad_norm": 11.546829223632812, "learning_rate": 8.733333333333333e-05, "loss": 1.382, "step": 656 }, { "epoch": 0.04451521105765973, "grad_norm": 13.163644790649414, "learning_rate": 8.746666666666667e-05, "loss": 1.2253, "step": 657 }, { "epoch": 0.044582966325631815, "grad_norm": 17.037311553955078, "learning_rate": 8.76e-05, "loss": 1.2727, "step": 658 }, { "epoch": 0.0446507215936039, "grad_norm": 14.610177040100098, "learning_rate": 8.773333333333333e-05, "loss": 1.3082, "step": 659 }, { "epoch": 0.044718476861575984, "grad_norm": 16.305557250976562, "learning_rate": 8.786666666666667e-05, "loss": 1.3948, "step": 660 }, { "epoch": 0.04478623212954807, "grad_norm": 13.207799911499023, "learning_rate": 8.800000000000001e-05, "loss": 1.2764, "step": 661 }, { "epoch": 0.04485398739752016, "grad_norm": 11.451075553894043, "learning_rate": 8.813333333333334e-05, "loss": 1.0831, "step": 662 }, { "epoch": 0.044921742665492245, "grad_norm": 13.555370330810547, "learning_rate": 8.826666666666668e-05, "loss": 1.4996, "step": 663 }, { "epoch": 0.04498949793346433, "grad_norm": 13.544769287109375, "learning_rate": 8.840000000000001e-05, "loss": 1.2277, "step": 664 }, { "epoch": 0.045057253201436415, "grad_norm": 18.05879783630371, "learning_rate": 8.853333333333333e-05, "loss": 1.3964, "step": 665 }, { "epoch": 0.0451250084694085, "grad_norm": 17.309839248657227, "learning_rate": 8.866666666666668e-05, "loss": 1.6233, "step": 666 }, { "epoch": 0.045192763737380584, "grad_norm": 12.732510566711426, "learning_rate": 8.88e-05, "loss": 1.3237, "step": 667 }, { "epoch": 0.04526051900535267, "grad_norm": 13.541101455688477, "learning_rate": 8.893333333333333e-05, "loss": 1.2523, "step": 668 }, { "epoch": 0.04532827427332475, "grad_norm": 17.54905891418457, "learning_rate": 8.906666666666667e-05, "loss": 1.5824, "step": 669 }, { "epoch": 0.04539602954129684, "grad_norm": 12.52578353881836, "learning_rate": 8.92e-05, "loss": 1.2966, "step": 670 }, { "epoch": 0.04546378480926892, "grad_norm": 13.279097557067871, "learning_rate": 8.933333333333334e-05, "loss": 1.4898, "step": 671 }, { "epoch": 0.04553154007724101, "grad_norm": 15.892850875854492, "learning_rate": 8.946666666666668e-05, "loss": 1.3757, "step": 672 }, { "epoch": 0.04559929534521309, "grad_norm": 14.108098983764648, "learning_rate": 8.960000000000001e-05, "loss": 1.1758, "step": 673 }, { "epoch": 0.04566705061318518, "grad_norm": 17.15204429626465, "learning_rate": 8.973333333333334e-05, "loss": 1.3876, "step": 674 }, { "epoch": 0.04573480588115726, "grad_norm": 14.453113555908203, "learning_rate": 8.986666666666666e-05, "loss": 1.2152, "step": 675 }, { "epoch": 0.045802561149129346, "grad_norm": 17.9672794342041, "learning_rate": 9e-05, "loss": 1.2962, "step": 676 }, { "epoch": 0.04587031641710143, "grad_norm": 19.810890197753906, "learning_rate": 9.013333333333333e-05, "loss": 1.3374, "step": 677 }, { "epoch": 0.045938071685073516, "grad_norm": 16.13353729248047, "learning_rate": 9.026666666666666e-05, "loss": 1.4125, "step": 678 }, { "epoch": 0.0460058269530456, "grad_norm": 15.257608413696289, "learning_rate": 9.04e-05, "loss": 1.1931, "step": 679 }, { "epoch": 0.046073582221017685, "grad_norm": 16.88699722290039, "learning_rate": 9.053333333333334e-05, "loss": 1.362, "step": 680 }, { "epoch": 0.04614133748898977, "grad_norm": 15.46777057647705, "learning_rate": 9.066666666666667e-05, "loss": 1.3522, "step": 681 }, { "epoch": 0.046209092756961855, "grad_norm": 13.584056854248047, "learning_rate": 9.080000000000001e-05, "loss": 1.1998, "step": 682 }, { "epoch": 0.04627684802493394, "grad_norm": 14.226449966430664, "learning_rate": 9.093333333333334e-05, "loss": 1.3312, "step": 683 }, { "epoch": 0.046344603292906024, "grad_norm": 15.157097816467285, "learning_rate": 9.106666666666667e-05, "loss": 1.3422, "step": 684 }, { "epoch": 0.04641235856087811, "grad_norm": 14.748275756835938, "learning_rate": 9.120000000000001e-05, "loss": 1.2902, "step": 685 }, { "epoch": 0.04648011382885019, "grad_norm": 13.730619430541992, "learning_rate": 9.133333333333334e-05, "loss": 1.3034, "step": 686 }, { "epoch": 0.04654786909682228, "grad_norm": 12.804062843322754, "learning_rate": 9.146666666666666e-05, "loss": 1.1484, "step": 687 }, { "epoch": 0.04661562436479436, "grad_norm": 16.587923049926758, "learning_rate": 9.16e-05, "loss": 1.4875, "step": 688 }, { "epoch": 0.04668337963276645, "grad_norm": 12.228131294250488, "learning_rate": 9.173333333333333e-05, "loss": 0.917, "step": 689 }, { "epoch": 0.04675113490073853, "grad_norm": 18.680187225341797, "learning_rate": 9.186666666666667e-05, "loss": 1.497, "step": 690 }, { "epoch": 0.04681889016871062, "grad_norm": 14.80630111694336, "learning_rate": 9.200000000000001e-05, "loss": 1.5623, "step": 691 }, { "epoch": 0.0468866454366827, "grad_norm": 13.754642486572266, "learning_rate": 9.213333333333334e-05, "loss": 1.3921, "step": 692 }, { "epoch": 0.046954400704654786, "grad_norm": 14.264236450195312, "learning_rate": 9.226666666666667e-05, "loss": 1.3615, "step": 693 }, { "epoch": 0.04702215597262687, "grad_norm": 15.883113861083984, "learning_rate": 9.240000000000001e-05, "loss": 1.4528, "step": 694 }, { "epoch": 0.047089911240598956, "grad_norm": 13.528610229492188, "learning_rate": 9.253333333333334e-05, "loss": 1.0995, "step": 695 }, { "epoch": 0.04715766650857104, "grad_norm": 15.945343971252441, "learning_rate": 9.266666666666666e-05, "loss": 1.272, "step": 696 }, { "epoch": 0.047225421776543125, "grad_norm": 14.378050804138184, "learning_rate": 9.28e-05, "loss": 1.3545, "step": 697 }, { "epoch": 0.04729317704451521, "grad_norm": 13.457077980041504, "learning_rate": 9.293333333333333e-05, "loss": 1.1751, "step": 698 }, { "epoch": 0.047360932312487294, "grad_norm": 19.285078048706055, "learning_rate": 9.306666666666667e-05, "loss": 1.4383, "step": 699 }, { "epoch": 0.04742868758045938, "grad_norm": 16.683856964111328, "learning_rate": 9.320000000000002e-05, "loss": 1.2499, "step": 700 }, { "epoch": 0.047496442848431464, "grad_norm": 13.818337440490723, "learning_rate": 9.333333333333334e-05, "loss": 1.2325, "step": 701 }, { "epoch": 0.04756419811640355, "grad_norm": 12.51142406463623, "learning_rate": 9.346666666666667e-05, "loss": 1.3317, "step": 702 }, { "epoch": 0.04763195338437563, "grad_norm": 14.697171211242676, "learning_rate": 9.360000000000001e-05, "loss": 1.6055, "step": 703 }, { "epoch": 0.04769970865234772, "grad_norm": 16.942562103271484, "learning_rate": 9.373333333333334e-05, "loss": 1.4279, "step": 704 }, { "epoch": 0.0477674639203198, "grad_norm": 16.739248275756836, "learning_rate": 9.386666666666667e-05, "loss": 1.2285, "step": 705 }, { "epoch": 0.04783521918829189, "grad_norm": 14.479548454284668, "learning_rate": 9.4e-05, "loss": 1.1984, "step": 706 }, { "epoch": 0.04790297445626397, "grad_norm": 14.384824752807617, "learning_rate": 9.413333333333334e-05, "loss": 1.2128, "step": 707 }, { "epoch": 0.04797072972423606, "grad_norm": 15.41884994506836, "learning_rate": 9.426666666666666e-05, "loss": 1.1627, "step": 708 }, { "epoch": 0.04803848499220814, "grad_norm": 13.267902374267578, "learning_rate": 9.44e-05, "loss": 1.2916, "step": 709 }, { "epoch": 0.048106240260180226, "grad_norm": 14.909900665283203, "learning_rate": 9.453333333333335e-05, "loss": 1.1579, "step": 710 }, { "epoch": 0.04817399552815231, "grad_norm": 12.639839172363281, "learning_rate": 9.466666666666667e-05, "loss": 1.2053, "step": 711 }, { "epoch": 0.048241750796124395, "grad_norm": 14.957659721374512, "learning_rate": 9.48e-05, "loss": 1.4556, "step": 712 }, { "epoch": 0.04830950606409648, "grad_norm": 14.008201599121094, "learning_rate": 9.493333333333334e-05, "loss": 1.1575, "step": 713 }, { "epoch": 0.048377261332068565, "grad_norm": 13.857501983642578, "learning_rate": 9.506666666666667e-05, "loss": 1.4121, "step": 714 }, { "epoch": 0.048445016600040657, "grad_norm": 13.867570877075195, "learning_rate": 9.52e-05, "loss": 1.4908, "step": 715 }, { "epoch": 0.04851277186801274, "grad_norm": 14.3545503616333, "learning_rate": 9.533333333333334e-05, "loss": 1.2939, "step": 716 }, { "epoch": 0.048580527135984826, "grad_norm": 15.173371315002441, "learning_rate": 9.546666666666667e-05, "loss": 1.3912, "step": 717 }, { "epoch": 0.04864828240395691, "grad_norm": 13.799453735351562, "learning_rate": 9.56e-05, "loss": 1.4219, "step": 718 }, { "epoch": 0.048716037671928995, "grad_norm": 18.264965057373047, "learning_rate": 9.573333333333335e-05, "loss": 1.4967, "step": 719 }, { "epoch": 0.04878379293990108, "grad_norm": 13.809319496154785, "learning_rate": 9.586666666666667e-05, "loss": 1.3527, "step": 720 }, { "epoch": 0.048851548207873165, "grad_norm": 13.931517601013184, "learning_rate": 9.6e-05, "loss": 1.4018, "step": 721 }, { "epoch": 0.04891930347584525, "grad_norm": 11.74387264251709, "learning_rate": 9.613333333333334e-05, "loss": 1.1981, "step": 722 }, { "epoch": 0.048987058743817334, "grad_norm": 14.960589408874512, "learning_rate": 9.626666666666667e-05, "loss": 1.3373, "step": 723 }, { "epoch": 0.04905481401178942, "grad_norm": 14.170279502868652, "learning_rate": 9.64e-05, "loss": 1.2702, "step": 724 }, { "epoch": 0.049122569279761504, "grad_norm": 15.532767295837402, "learning_rate": 9.653333333333334e-05, "loss": 1.1329, "step": 725 }, { "epoch": 0.04919032454773359, "grad_norm": 12.141284942626953, "learning_rate": 9.666666666666667e-05, "loss": 1.2694, "step": 726 }, { "epoch": 0.04925807981570567, "grad_norm": 12.379233360290527, "learning_rate": 9.680000000000001e-05, "loss": 1.1545, "step": 727 }, { "epoch": 0.04932583508367776, "grad_norm": 14.45605182647705, "learning_rate": 9.693333333333335e-05, "loss": 1.4151, "step": 728 }, { "epoch": 0.04939359035164984, "grad_norm": 15.656912803649902, "learning_rate": 9.706666666666668e-05, "loss": 1.3782, "step": 729 }, { "epoch": 0.04946134561962193, "grad_norm": 13.992748260498047, "learning_rate": 9.72e-05, "loss": 1.2731, "step": 730 }, { "epoch": 0.04952910088759401, "grad_norm": 16.253517150878906, "learning_rate": 9.733333333333335e-05, "loss": 1.4503, "step": 731 }, { "epoch": 0.049596856155566096, "grad_norm": 14.88161563873291, "learning_rate": 9.746666666666667e-05, "loss": 1.3909, "step": 732 }, { "epoch": 0.04966461142353818, "grad_norm": 14.60545539855957, "learning_rate": 9.76e-05, "loss": 1.1803, "step": 733 }, { "epoch": 0.049732366691510266, "grad_norm": 14.589803695678711, "learning_rate": 9.773333333333334e-05, "loss": 1.3844, "step": 734 }, { "epoch": 0.04980012195948235, "grad_norm": 14.781076431274414, "learning_rate": 9.786666666666667e-05, "loss": 1.2706, "step": 735 }, { "epoch": 0.049867877227454435, "grad_norm": 13.717072486877441, "learning_rate": 9.8e-05, "loss": 1.2311, "step": 736 }, { "epoch": 0.04993563249542652, "grad_norm": 17.14280891418457, "learning_rate": 9.813333333333334e-05, "loss": 1.3423, "step": 737 }, { "epoch": 0.050003387763398605, "grad_norm": 14.521373748779297, "learning_rate": 9.826666666666668e-05, "loss": 1.3118, "step": 738 }, { "epoch": 0.05007114303137069, "grad_norm": 14.79515266418457, "learning_rate": 9.84e-05, "loss": 1.3622, "step": 739 }, { "epoch": 0.050138898299342774, "grad_norm": 13.63962459564209, "learning_rate": 9.853333333333333e-05, "loss": 1.3856, "step": 740 }, { "epoch": 0.05020665356731486, "grad_norm": 13.788541793823242, "learning_rate": 9.866666666666668e-05, "loss": 1.3439, "step": 741 }, { "epoch": 0.05027440883528694, "grad_norm": 14.015862464904785, "learning_rate": 9.88e-05, "loss": 1.3263, "step": 742 }, { "epoch": 0.05034216410325903, "grad_norm": 15.407167434692383, "learning_rate": 9.893333333333333e-05, "loss": 1.5115, "step": 743 }, { "epoch": 0.05040991937123111, "grad_norm": 14.542003631591797, "learning_rate": 9.906666666666667e-05, "loss": 1.3934, "step": 744 }, { "epoch": 0.0504776746392032, "grad_norm": 16.733686447143555, "learning_rate": 9.92e-05, "loss": 1.2098, "step": 745 }, { "epoch": 0.05054542990717528, "grad_norm": 14.979568481445312, "learning_rate": 9.933333333333334e-05, "loss": 1.3234, "step": 746 }, { "epoch": 0.05061318517514737, "grad_norm": 13.699801445007324, "learning_rate": 9.946666666666668e-05, "loss": 1.3173, "step": 747 }, { "epoch": 0.05068094044311945, "grad_norm": 13.262088775634766, "learning_rate": 9.960000000000001e-05, "loss": 1.3459, "step": 748 }, { "epoch": 0.050748695711091536, "grad_norm": 14.694448471069336, "learning_rate": 9.973333333333334e-05, "loss": 1.3639, "step": 749 }, { "epoch": 0.05081645097906362, "grad_norm": 13.999008178710938, "learning_rate": 9.986666666666668e-05, "loss": 1.4275, "step": 750 }, { "epoch": 0.050884206247035706, "grad_norm": 11.588037490844727, "learning_rate": 0.0001, "loss": 1.2866, "step": 751 }, { "epoch": 0.05095196151500779, "grad_norm": 16.174110412597656, "learning_rate": 9.999863098090219e-05, "loss": 1.5338, "step": 752 }, { "epoch": 0.051019716782979875, "grad_norm": 17.8669376373291, "learning_rate": 9.999726196180437e-05, "loss": 1.3247, "step": 753 }, { "epoch": 0.05108747205095196, "grad_norm": 13.849075317382812, "learning_rate": 9.999589294270656e-05, "loss": 1.1817, "step": 754 }, { "epoch": 0.051155227318924044, "grad_norm": 14.888331413269043, "learning_rate": 9.999452392360874e-05, "loss": 1.4019, "step": 755 }, { "epoch": 0.05122298258689613, "grad_norm": 18.68206024169922, "learning_rate": 9.999315490451092e-05, "loss": 1.7115, "step": 756 }, { "epoch": 0.051290737854868214, "grad_norm": 14.762079238891602, "learning_rate": 9.99917858854131e-05, "loss": 1.2553, "step": 757 }, { "epoch": 0.0513584931228403, "grad_norm": 14.649972915649414, "learning_rate": 9.99904168663153e-05, "loss": 1.0807, "step": 758 }, { "epoch": 0.05142624839081238, "grad_norm": 13.172977447509766, "learning_rate": 9.998904784721747e-05, "loss": 1.4966, "step": 759 }, { "epoch": 0.05149400365878447, "grad_norm": 14.144796371459961, "learning_rate": 9.998767882811965e-05, "loss": 1.234, "step": 760 }, { "epoch": 0.05156175892675655, "grad_norm": 17.565507888793945, "learning_rate": 9.998630980902184e-05, "loss": 1.4061, "step": 761 }, { "epoch": 0.05162951419472864, "grad_norm": 17.205589294433594, "learning_rate": 9.998494078992402e-05, "loss": 1.6626, "step": 762 }, { "epoch": 0.05169726946270072, "grad_norm": 16.14542007446289, "learning_rate": 9.998357177082621e-05, "loss": 1.3145, "step": 763 }, { "epoch": 0.05176502473067281, "grad_norm": 14.692976951599121, "learning_rate": 9.998220275172839e-05, "loss": 1.2085, "step": 764 }, { "epoch": 0.05183277999864489, "grad_norm": 12.572774887084961, "learning_rate": 9.998083373263057e-05, "loss": 1.3743, "step": 765 }, { "epoch": 0.051900535266616976, "grad_norm": 15.904753684997559, "learning_rate": 9.997946471353275e-05, "loss": 1.3863, "step": 766 }, { "epoch": 0.05196829053458906, "grad_norm": 14.471494674682617, "learning_rate": 9.997809569443493e-05, "loss": 1.3615, "step": 767 }, { "epoch": 0.05203604580256115, "grad_norm": 13.82172966003418, "learning_rate": 9.997672667533712e-05, "loss": 1.2104, "step": 768 }, { "epoch": 0.05210380107053324, "grad_norm": 14.648073196411133, "learning_rate": 9.99753576562393e-05, "loss": 1.1166, "step": 769 }, { "epoch": 0.05217155633850532, "grad_norm": 16.776750564575195, "learning_rate": 9.997398863714149e-05, "loss": 1.4242, "step": 770 }, { "epoch": 0.05223931160647741, "grad_norm": 13.999717712402344, "learning_rate": 9.997261961804367e-05, "loss": 1.2493, "step": 771 }, { "epoch": 0.05230706687444949, "grad_norm": 13.238443374633789, "learning_rate": 9.997125059894586e-05, "loss": 1.0469, "step": 772 }, { "epoch": 0.052374822142421576, "grad_norm": 12.79283332824707, "learning_rate": 9.996988157984804e-05, "loss": 1.1975, "step": 773 }, { "epoch": 0.05244257741039366, "grad_norm": 15.577791213989258, "learning_rate": 9.996851256075022e-05, "loss": 1.4972, "step": 774 }, { "epoch": 0.052510332678365745, "grad_norm": 16.215747833251953, "learning_rate": 9.996714354165241e-05, "loss": 1.4285, "step": 775 }, { "epoch": 0.05257808794633783, "grad_norm": 14.68825626373291, "learning_rate": 9.99657745225546e-05, "loss": 1.2028, "step": 776 }, { "epoch": 0.052645843214309915, "grad_norm": 11.535130500793457, "learning_rate": 9.996440550345677e-05, "loss": 1.2466, "step": 777 }, { "epoch": 0.052713598482282, "grad_norm": 14.865918159484863, "learning_rate": 9.996303648435897e-05, "loss": 1.3903, "step": 778 }, { "epoch": 0.052781353750254084, "grad_norm": 13.626388549804688, "learning_rate": 9.996166746526115e-05, "loss": 1.5446, "step": 779 }, { "epoch": 0.05284910901822617, "grad_norm": 12.957473754882812, "learning_rate": 9.996029844616333e-05, "loss": 1.1641, "step": 780 }, { "epoch": 0.052916864286198254, "grad_norm": 14.928487777709961, "learning_rate": 9.995892942706552e-05, "loss": 1.0929, "step": 781 }, { "epoch": 0.05298461955417034, "grad_norm": 13.592952728271484, "learning_rate": 9.99575604079677e-05, "loss": 1.4684, "step": 782 }, { "epoch": 0.05305237482214242, "grad_norm": 13.836453437805176, "learning_rate": 9.995619138886988e-05, "loss": 1.0128, "step": 783 }, { "epoch": 0.05312013009011451, "grad_norm": 16.654394149780273, "learning_rate": 9.995482236977206e-05, "loss": 1.2464, "step": 784 }, { "epoch": 0.05318788535808659, "grad_norm": 14.4434232711792, "learning_rate": 9.995345335067424e-05, "loss": 1.3649, "step": 785 }, { "epoch": 0.05325564062605868, "grad_norm": 14.15592098236084, "learning_rate": 9.995208433157644e-05, "loss": 1.4426, "step": 786 }, { "epoch": 0.05332339589403076, "grad_norm": 12.97901439666748, "learning_rate": 9.995071531247862e-05, "loss": 1.227, "step": 787 }, { "epoch": 0.053391151162002846, "grad_norm": 13.167902946472168, "learning_rate": 9.99493462933808e-05, "loss": 1.2432, "step": 788 }, { "epoch": 0.05345890642997493, "grad_norm": 15.92297077178955, "learning_rate": 9.994797727428298e-05, "loss": 1.3345, "step": 789 }, { "epoch": 0.053526661697947016, "grad_norm": 13.30277156829834, "learning_rate": 9.994660825518517e-05, "loss": 1.2361, "step": 790 }, { "epoch": 0.0535944169659191, "grad_norm": 15.98779582977295, "learning_rate": 9.994523923608735e-05, "loss": 1.1202, "step": 791 }, { "epoch": 0.053662172233891185, "grad_norm": 16.414382934570312, "learning_rate": 9.994387021698953e-05, "loss": 1.448, "step": 792 }, { "epoch": 0.05372992750186327, "grad_norm": 14.491677284240723, "learning_rate": 9.994250119789171e-05, "loss": 1.0862, "step": 793 }, { "epoch": 0.053797682769835355, "grad_norm": 13.155410766601562, "learning_rate": 9.99411321787939e-05, "loss": 1.312, "step": 794 }, { "epoch": 0.05386543803780744, "grad_norm": 16.11139488220215, "learning_rate": 9.993976315969609e-05, "loss": 1.2043, "step": 795 }, { "epoch": 0.053933193305779524, "grad_norm": 14.259698867797852, "learning_rate": 9.993839414059827e-05, "loss": 1.2256, "step": 796 }, { "epoch": 0.05400094857375161, "grad_norm": 15.751099586486816, "learning_rate": 9.993702512150045e-05, "loss": 1.0732, "step": 797 }, { "epoch": 0.05406870384172369, "grad_norm": 12.766170501708984, "learning_rate": 9.993565610240263e-05, "loss": 1.0362, "step": 798 }, { "epoch": 0.05413645910969578, "grad_norm": 14.61483097076416, "learning_rate": 9.993428708330481e-05, "loss": 1.3107, "step": 799 }, { "epoch": 0.05420421437766786, "grad_norm": 14.571990013122559, "learning_rate": 9.9932918064207e-05, "loss": 1.3616, "step": 800 }, { "epoch": 0.05427196964563995, "grad_norm": 14.69124984741211, "learning_rate": 9.993154904510918e-05, "loss": 1.4273, "step": 801 }, { "epoch": 0.05433972491361203, "grad_norm": 13.834383010864258, "learning_rate": 9.993018002601136e-05, "loss": 1.2886, "step": 802 }, { "epoch": 0.05440748018158412, "grad_norm": 14.178943634033203, "learning_rate": 9.992881100691355e-05, "loss": 1.44, "step": 803 }, { "epoch": 0.0544752354495562, "grad_norm": 13.185267448425293, "learning_rate": 9.992744198781574e-05, "loss": 1.5609, "step": 804 }, { "epoch": 0.054542990717528286, "grad_norm": 15.211031913757324, "learning_rate": 9.992607296871792e-05, "loss": 1.8873, "step": 805 }, { "epoch": 0.05461074598550037, "grad_norm": 11.855447769165039, "learning_rate": 9.99247039496201e-05, "loss": 1.4063, "step": 806 }, { "epoch": 0.054678501253472456, "grad_norm": 14.352787017822266, "learning_rate": 9.992333493052228e-05, "loss": 1.56, "step": 807 }, { "epoch": 0.05474625652144454, "grad_norm": 14.339908599853516, "learning_rate": 9.992196591142446e-05, "loss": 1.1985, "step": 808 }, { "epoch": 0.054814011789416625, "grad_norm": 12.696648597717285, "learning_rate": 9.992059689232665e-05, "loss": 1.1807, "step": 809 }, { "epoch": 0.05488176705738871, "grad_norm": 14.083525657653809, "learning_rate": 9.991922787322883e-05, "loss": 1.4012, "step": 810 }, { "epoch": 0.054949522325360795, "grad_norm": 14.298514366149902, "learning_rate": 9.991785885413101e-05, "loss": 1.4047, "step": 811 }, { "epoch": 0.05501727759333288, "grad_norm": 13.391324996948242, "learning_rate": 9.99164898350332e-05, "loss": 1.2292, "step": 812 }, { "epoch": 0.055085032861304964, "grad_norm": 14.0011625289917, "learning_rate": 9.991512081593539e-05, "loss": 1.3194, "step": 813 }, { "epoch": 0.05515278812927705, "grad_norm": 14.726574897766113, "learning_rate": 9.991375179683757e-05, "loss": 1.3245, "step": 814 }, { "epoch": 0.05522054339724913, "grad_norm": 14.615242004394531, "learning_rate": 9.991238277773975e-05, "loss": 1.3704, "step": 815 }, { "epoch": 0.05528829866522122, "grad_norm": 11.06546688079834, "learning_rate": 9.991101375864193e-05, "loss": 1.0841, "step": 816 }, { "epoch": 0.0553560539331933, "grad_norm": 13.768633842468262, "learning_rate": 9.990964473954411e-05, "loss": 1.2632, "step": 817 }, { "epoch": 0.05542380920116539, "grad_norm": 15.054973602294922, "learning_rate": 9.99082757204463e-05, "loss": 1.2931, "step": 818 }, { "epoch": 0.05549156446913747, "grad_norm": 15.27096176147461, "learning_rate": 9.990690670134848e-05, "loss": 1.1789, "step": 819 }, { "epoch": 0.055559319737109564, "grad_norm": 13.228281021118164, "learning_rate": 9.990553768225067e-05, "loss": 1.4458, "step": 820 }, { "epoch": 0.05562707500508165, "grad_norm": 16.016782760620117, "learning_rate": 9.990416866315286e-05, "loss": 1.5121, "step": 821 }, { "epoch": 0.05569483027305373, "grad_norm": 14.15912914276123, "learning_rate": 9.990279964405504e-05, "loss": 1.3702, "step": 822 }, { "epoch": 0.05576258554102582, "grad_norm": 14.61017894744873, "learning_rate": 9.990143062495722e-05, "loss": 1.2774, "step": 823 }, { "epoch": 0.0558303408089979, "grad_norm": 13.241140365600586, "learning_rate": 9.990006160585941e-05, "loss": 1.3857, "step": 824 }, { "epoch": 0.05589809607696999, "grad_norm": 11.815064430236816, "learning_rate": 9.98986925867616e-05, "loss": 1.1053, "step": 825 }, { "epoch": 0.05596585134494207, "grad_norm": 13.179222106933594, "learning_rate": 9.989732356766377e-05, "loss": 1.4686, "step": 826 }, { "epoch": 0.05603360661291416, "grad_norm": 11.888179779052734, "learning_rate": 9.989595454856597e-05, "loss": 1.3191, "step": 827 }, { "epoch": 0.05610136188088624, "grad_norm": 15.43813419342041, "learning_rate": 9.989458552946815e-05, "loss": 0.9732, "step": 828 }, { "epoch": 0.056169117148858326, "grad_norm": 20.0246524810791, "learning_rate": 9.989321651037033e-05, "loss": 1.5398, "step": 829 }, { "epoch": 0.05623687241683041, "grad_norm": 14.247052192687988, "learning_rate": 9.989184749127251e-05, "loss": 1.5093, "step": 830 }, { "epoch": 0.056304627684802495, "grad_norm": 15.63775634765625, "learning_rate": 9.989047847217469e-05, "loss": 1.3488, "step": 831 }, { "epoch": 0.05637238295277458, "grad_norm": 10.582650184631348, "learning_rate": 9.988910945307688e-05, "loss": 1.3888, "step": 832 }, { "epoch": 0.056440138220746665, "grad_norm": 12.344864845275879, "learning_rate": 9.988774043397906e-05, "loss": 1.2876, "step": 833 }, { "epoch": 0.05650789348871875, "grad_norm": 13.95814323425293, "learning_rate": 9.988637141488124e-05, "loss": 1.2648, "step": 834 }, { "epoch": 0.056575648756690834, "grad_norm": 9.740105628967285, "learning_rate": 9.988500239578342e-05, "loss": 0.9777, "step": 835 }, { "epoch": 0.05664340402466292, "grad_norm": 16.16588020324707, "learning_rate": 9.988363337668562e-05, "loss": 1.5859, "step": 836 }, { "epoch": 0.056711159292635004, "grad_norm": 14.394731521606445, "learning_rate": 9.98822643575878e-05, "loss": 1.3395, "step": 837 }, { "epoch": 0.05677891456060709, "grad_norm": 11.256571769714355, "learning_rate": 9.988089533848998e-05, "loss": 1.0528, "step": 838 }, { "epoch": 0.05684666982857917, "grad_norm": 17.595510482788086, "learning_rate": 9.987952631939216e-05, "loss": 1.2678, "step": 839 }, { "epoch": 0.05691442509655126, "grad_norm": 14.132645606994629, "learning_rate": 9.987815730029434e-05, "loss": 1.4129, "step": 840 }, { "epoch": 0.05698218036452334, "grad_norm": 14.438119888305664, "learning_rate": 9.987678828119653e-05, "loss": 1.4141, "step": 841 }, { "epoch": 0.05704993563249543, "grad_norm": 15.983094215393066, "learning_rate": 9.987541926209871e-05, "loss": 1.1959, "step": 842 }, { "epoch": 0.05711769090046751, "grad_norm": 17.300403594970703, "learning_rate": 9.98740502430009e-05, "loss": 1.1613, "step": 843 }, { "epoch": 0.057185446168439596, "grad_norm": 15.150660514831543, "learning_rate": 9.987268122390307e-05, "loss": 1.2253, "step": 844 }, { "epoch": 0.05725320143641168, "grad_norm": 14.234586715698242, "learning_rate": 9.987131220480525e-05, "loss": 1.2517, "step": 845 }, { "epoch": 0.057320956704383766, "grad_norm": 18.31337547302246, "learning_rate": 9.986994318570745e-05, "loss": 1.7506, "step": 846 }, { "epoch": 0.05738871197235585, "grad_norm": 14.818669319152832, "learning_rate": 9.986857416660963e-05, "loss": 1.358, "step": 847 }, { "epoch": 0.057456467240327935, "grad_norm": 14.640913009643555, "learning_rate": 9.986720514751181e-05, "loss": 1.4481, "step": 848 }, { "epoch": 0.05752422250830002, "grad_norm": 12.277986526489258, "learning_rate": 9.986583612841399e-05, "loss": 1.3428, "step": 849 }, { "epoch": 0.057591977776272105, "grad_norm": 13.817851066589355, "learning_rate": 9.986446710931618e-05, "loss": 1.3514, "step": 850 }, { "epoch": 0.05765973304424419, "grad_norm": 13.706515312194824, "learning_rate": 9.986309809021836e-05, "loss": 1.4295, "step": 851 }, { "epoch": 0.057727488312216274, "grad_norm": 16.649917602539062, "learning_rate": 9.986172907112054e-05, "loss": 1.3895, "step": 852 }, { "epoch": 0.05779524358018836, "grad_norm": 13.659167289733887, "learning_rate": 9.986036005202272e-05, "loss": 1.1545, "step": 853 }, { "epoch": 0.057862998848160443, "grad_norm": 11.264912605285645, "learning_rate": 9.98589910329249e-05, "loss": 1.0766, "step": 854 }, { "epoch": 0.05793075411613253, "grad_norm": 12.930856704711914, "learning_rate": 9.98576220138271e-05, "loss": 1.1665, "step": 855 }, { "epoch": 0.05799850938410461, "grad_norm": 15.999971389770508, "learning_rate": 9.985625299472928e-05, "loss": 1.4595, "step": 856 }, { "epoch": 0.0580662646520767, "grad_norm": 14.566671371459961, "learning_rate": 9.985488397563146e-05, "loss": 1.283, "step": 857 }, { "epoch": 0.05813401992004878, "grad_norm": 16.106964111328125, "learning_rate": 9.985351495653364e-05, "loss": 1.2842, "step": 858 }, { "epoch": 0.05820177518802087, "grad_norm": 15.47492790222168, "learning_rate": 9.985214593743583e-05, "loss": 1.3949, "step": 859 }, { "epoch": 0.05826953045599295, "grad_norm": 10.692886352539062, "learning_rate": 9.985077691833801e-05, "loss": 1.2896, "step": 860 }, { "epoch": 0.058337285723965036, "grad_norm": 14.13198184967041, "learning_rate": 9.98494078992402e-05, "loss": 1.1566, "step": 861 }, { "epoch": 0.05840504099193712, "grad_norm": 14.455452919006348, "learning_rate": 9.984803888014237e-05, "loss": 1.1949, "step": 862 }, { "epoch": 0.058472796259909206, "grad_norm": 15.020733833312988, "learning_rate": 9.984666986104456e-05, "loss": 1.5202, "step": 863 }, { "epoch": 0.05854055152788129, "grad_norm": 21.089344024658203, "learning_rate": 9.984530084194675e-05, "loss": 1.5083, "step": 864 }, { "epoch": 0.058608306795853375, "grad_norm": 14.414257049560547, "learning_rate": 9.984393182284893e-05, "loss": 1.3977, "step": 865 }, { "epoch": 0.05867606206382546, "grad_norm": 15.667798042297363, "learning_rate": 9.984256280375111e-05, "loss": 1.4044, "step": 866 }, { "epoch": 0.058743817331797545, "grad_norm": 12.503005981445312, "learning_rate": 9.98411937846533e-05, "loss": 1.3369, "step": 867 }, { "epoch": 0.05881157259976963, "grad_norm": 13.27022933959961, "learning_rate": 9.983982476555548e-05, "loss": 1.4134, "step": 868 }, { "epoch": 0.058879327867741714, "grad_norm": 16.2034969329834, "learning_rate": 9.983845574645766e-05, "loss": 1.1196, "step": 869 }, { "epoch": 0.0589470831357138, "grad_norm": 9.868896484375, "learning_rate": 9.983708672735986e-05, "loss": 1.0248, "step": 870 }, { "epoch": 0.05901483840368588, "grad_norm": 15.588685989379883, "learning_rate": 9.983571770826204e-05, "loss": 1.5865, "step": 871 }, { "epoch": 0.05908259367165797, "grad_norm": 14.688246726989746, "learning_rate": 9.983434868916422e-05, "loss": 1.2959, "step": 872 }, { "epoch": 0.05915034893963006, "grad_norm": 14.252961158752441, "learning_rate": 9.983297967006641e-05, "loss": 1.3536, "step": 873 }, { "epoch": 0.059218104207602144, "grad_norm": 11.79800033569336, "learning_rate": 9.98316106509686e-05, "loss": 1.331, "step": 874 }, { "epoch": 0.05928585947557423, "grad_norm": 11.900074005126953, "learning_rate": 9.983024163187077e-05, "loss": 1.2241, "step": 875 }, { "epoch": 0.059353614743546314, "grad_norm": 10.696773529052734, "learning_rate": 9.982887261277295e-05, "loss": 1.4041, "step": 876 }, { "epoch": 0.0594213700115184, "grad_norm": 13.532305717468262, "learning_rate": 9.982750359367513e-05, "loss": 1.3621, "step": 877 }, { "epoch": 0.05948912527949048, "grad_norm": 14.107857704162598, "learning_rate": 9.982613457457733e-05, "loss": 1.6327, "step": 878 }, { "epoch": 0.05955688054746257, "grad_norm": 11.584097862243652, "learning_rate": 9.982476555547951e-05, "loss": 0.9606, "step": 879 }, { "epoch": 0.05962463581543465, "grad_norm": 14.240161895751953, "learning_rate": 9.982339653638169e-05, "loss": 1.263, "step": 880 }, { "epoch": 0.05969239108340674, "grad_norm": 14.461871147155762, "learning_rate": 9.982202751728387e-05, "loss": 1.4201, "step": 881 }, { "epoch": 0.05976014635137882, "grad_norm": 14.072705268859863, "learning_rate": 9.982065849818606e-05, "loss": 1.5977, "step": 882 }, { "epoch": 0.05982790161935091, "grad_norm": 14.928994178771973, "learning_rate": 9.981928947908824e-05, "loss": 1.3346, "step": 883 }, { "epoch": 0.05989565688732299, "grad_norm": 14.898951530456543, "learning_rate": 9.981792045999042e-05, "loss": 1.312, "step": 884 }, { "epoch": 0.059963412155295076, "grad_norm": 13.089646339416504, "learning_rate": 9.98165514408926e-05, "loss": 1.0833, "step": 885 }, { "epoch": 0.06003116742326716, "grad_norm": 15.768043518066406, "learning_rate": 9.981518242179478e-05, "loss": 1.1246, "step": 886 }, { "epoch": 0.060098922691239245, "grad_norm": 11.8709135055542, "learning_rate": 9.981381340269698e-05, "loss": 1.1822, "step": 887 }, { "epoch": 0.06016667795921133, "grad_norm": 15.698454856872559, "learning_rate": 9.981244438359916e-05, "loss": 1.3733, "step": 888 }, { "epoch": 0.060234433227183415, "grad_norm": 14.827208518981934, "learning_rate": 9.981107536450134e-05, "loss": 1.3121, "step": 889 }, { "epoch": 0.0603021884951555, "grad_norm": 12.522045135498047, "learning_rate": 9.980970634540352e-05, "loss": 1.5584, "step": 890 }, { "epoch": 0.060369943763127584, "grad_norm": 14.417738914489746, "learning_rate": 9.980833732630571e-05, "loss": 1.2389, "step": 891 }, { "epoch": 0.06043769903109967, "grad_norm": 14.761930465698242, "learning_rate": 9.98069683072079e-05, "loss": 1.5007, "step": 892 }, { "epoch": 0.060505454299071754, "grad_norm": 15.882668495178223, "learning_rate": 9.980559928811007e-05, "loss": 1.2651, "step": 893 }, { "epoch": 0.06057320956704384, "grad_norm": 13.605412483215332, "learning_rate": 9.980423026901225e-05, "loss": 1.5186, "step": 894 }, { "epoch": 0.06064096483501592, "grad_norm": 10.654335021972656, "learning_rate": 9.980286124991443e-05, "loss": 1.1177, "step": 895 }, { "epoch": 0.06070872010298801, "grad_norm": 12.37457275390625, "learning_rate": 9.980149223081663e-05, "loss": 1.2362, "step": 896 }, { "epoch": 0.06077647537096009, "grad_norm": 12.591222763061523, "learning_rate": 9.980012321171881e-05, "loss": 1.2509, "step": 897 }, { "epoch": 0.06084423063893218, "grad_norm": 14.337310791015625, "learning_rate": 9.979875419262099e-05, "loss": 1.2987, "step": 898 }, { "epoch": 0.06091198590690426, "grad_norm": 15.496018409729004, "learning_rate": 9.979738517352317e-05, "loss": 1.5307, "step": 899 }, { "epoch": 0.06097974117487635, "grad_norm": 13.730890274047852, "learning_rate": 9.979601615442535e-05, "loss": 1.2975, "step": 900 }, { "epoch": 0.06104749644284843, "grad_norm": 12.314823150634766, "learning_rate": 9.979464713532754e-05, "loss": 1.1916, "step": 901 }, { "epoch": 0.061115251710820516, "grad_norm": 13.761808395385742, "learning_rate": 9.979327811622972e-05, "loss": 1.3037, "step": 902 }, { "epoch": 0.0611830069787926, "grad_norm": 13.308722496032715, "learning_rate": 9.97919090971319e-05, "loss": 1.0598, "step": 903 }, { "epoch": 0.061250762246764685, "grad_norm": 13.121098518371582, "learning_rate": 9.979054007803408e-05, "loss": 1.3378, "step": 904 }, { "epoch": 0.06131851751473677, "grad_norm": 16.975666046142578, "learning_rate": 9.978917105893628e-05, "loss": 1.2826, "step": 905 }, { "epoch": 0.061386272782708855, "grad_norm": 14.529984474182129, "learning_rate": 9.978780203983846e-05, "loss": 1.3467, "step": 906 }, { "epoch": 0.06145402805068094, "grad_norm": 11.081110000610352, "learning_rate": 9.978643302074064e-05, "loss": 1.1906, "step": 907 }, { "epoch": 0.061521783318653024, "grad_norm": 12.871200561523438, "learning_rate": 9.978506400164282e-05, "loss": 1.1057, "step": 908 }, { "epoch": 0.06158953858662511, "grad_norm": 13.982168197631836, "learning_rate": 9.9783694982545e-05, "loss": 1.3824, "step": 909 }, { "epoch": 0.061657293854597194, "grad_norm": 13.076074600219727, "learning_rate": 9.97823259634472e-05, "loss": 1.2985, "step": 910 }, { "epoch": 0.06172504912256928, "grad_norm": 11.015650749206543, "learning_rate": 9.978095694434937e-05, "loss": 1.2406, "step": 911 }, { "epoch": 0.06179280439054136, "grad_norm": 13.6082763671875, "learning_rate": 9.977958792525155e-05, "loss": 1.3873, "step": 912 }, { "epoch": 0.06186055965851345, "grad_norm": 15.930809020996094, "learning_rate": 9.977821890615375e-05, "loss": 1.3542, "step": 913 }, { "epoch": 0.06192831492648553, "grad_norm": 10.710271835327148, "learning_rate": 9.977684988705593e-05, "loss": 1.1571, "step": 914 }, { "epoch": 0.06199607019445762, "grad_norm": 11.110217094421387, "learning_rate": 9.977548086795811e-05, "loss": 1.3255, "step": 915 }, { "epoch": 0.0620638254624297, "grad_norm": 11.451903343200684, "learning_rate": 9.97741118488603e-05, "loss": 1.3681, "step": 916 }, { "epoch": 0.062131580730401786, "grad_norm": 10.884252548217773, "learning_rate": 9.977274282976248e-05, "loss": 1.1988, "step": 917 }, { "epoch": 0.06219933599837387, "grad_norm": 11.031237602233887, "learning_rate": 9.977137381066466e-05, "loss": 1.26, "step": 918 }, { "epoch": 0.062267091266345956, "grad_norm": 11.585648536682129, "learning_rate": 9.977000479156686e-05, "loss": 1.3099, "step": 919 }, { "epoch": 0.06233484653431804, "grad_norm": 10.867992401123047, "learning_rate": 9.976863577246904e-05, "loss": 1.2624, "step": 920 }, { "epoch": 0.062402601802290125, "grad_norm": 14.552916526794434, "learning_rate": 9.976726675337122e-05, "loss": 1.2244, "step": 921 }, { "epoch": 0.06247035707026221, "grad_norm": 12.101760864257812, "learning_rate": 9.97658977342734e-05, "loss": 1.1925, "step": 922 }, { "epoch": 0.0625381123382343, "grad_norm": 14.113842010498047, "learning_rate": 9.976452871517559e-05, "loss": 1.3395, "step": 923 }, { "epoch": 0.06260586760620639, "grad_norm": 17.214614868164062, "learning_rate": 9.976315969607777e-05, "loss": 1.3943, "step": 924 }, { "epoch": 0.06267362287417846, "grad_norm": 13.43308162689209, "learning_rate": 9.976179067697995e-05, "loss": 1.1596, "step": 925 }, { "epoch": 0.06274137814215056, "grad_norm": 13.806952476501465, "learning_rate": 9.976042165788213e-05, "loss": 1.0354, "step": 926 }, { "epoch": 0.06280913341012263, "grad_norm": 15.638693809509277, "learning_rate": 9.975905263878431e-05, "loss": 1.3807, "step": 927 }, { "epoch": 0.06287688867809473, "grad_norm": 14.337742805480957, "learning_rate": 9.975768361968651e-05, "loss": 1.537, "step": 928 }, { "epoch": 0.0629446439460668, "grad_norm": 14.540297508239746, "learning_rate": 9.975631460058869e-05, "loss": 1.2522, "step": 929 }, { "epoch": 0.0630123992140389, "grad_norm": 15.991955757141113, "learning_rate": 9.975494558149087e-05, "loss": 1.3855, "step": 930 }, { "epoch": 0.06308015448201097, "grad_norm": 13.957479476928711, "learning_rate": 9.975357656239305e-05, "loss": 1.43, "step": 931 }, { "epoch": 0.06314790974998306, "grad_norm": 16.805377960205078, "learning_rate": 9.975220754329523e-05, "loss": 1.383, "step": 932 }, { "epoch": 0.06321566501795514, "grad_norm": 12.41854476928711, "learning_rate": 9.975083852419742e-05, "loss": 1.068, "step": 933 }, { "epoch": 0.06328342028592723, "grad_norm": 15.929006576538086, "learning_rate": 9.97494695050996e-05, "loss": 1.212, "step": 934 }, { "epoch": 0.06335117555389931, "grad_norm": 13.205544471740723, "learning_rate": 9.974810048600178e-05, "loss": 1.3682, "step": 935 }, { "epoch": 0.0634189308218714, "grad_norm": 12.105626106262207, "learning_rate": 9.974673146690396e-05, "loss": 1.3821, "step": 936 }, { "epoch": 0.06348668608984348, "grad_norm": 13.776711463928223, "learning_rate": 9.974536244780616e-05, "loss": 1.1073, "step": 937 }, { "epoch": 0.06355444135781557, "grad_norm": 12.227380752563477, "learning_rate": 9.974399342870834e-05, "loss": 1.2026, "step": 938 }, { "epoch": 0.06362219662578765, "grad_norm": 12.723440170288086, "learning_rate": 9.974262440961052e-05, "loss": 1.1325, "step": 939 }, { "epoch": 0.06368995189375974, "grad_norm": 13.943262100219727, "learning_rate": 9.97412553905127e-05, "loss": 1.1878, "step": 940 }, { "epoch": 0.06375770716173182, "grad_norm": 12.644627571105957, "learning_rate": 9.973988637141488e-05, "loss": 1.3097, "step": 941 }, { "epoch": 0.06382546242970391, "grad_norm": 12.108241081237793, "learning_rate": 9.973851735231707e-05, "loss": 1.1686, "step": 942 }, { "epoch": 0.06389321769767599, "grad_norm": 14.375092506408691, "learning_rate": 9.973714833321925e-05, "loss": 1.2721, "step": 943 }, { "epoch": 0.06396097296564808, "grad_norm": 13.439800262451172, "learning_rate": 9.973577931412143e-05, "loss": 1.3898, "step": 944 }, { "epoch": 0.06402872823362016, "grad_norm": 13.717879295349121, "learning_rate": 9.973441029502361e-05, "loss": 1.3823, "step": 945 }, { "epoch": 0.06409648350159225, "grad_norm": 12.745361328125, "learning_rate": 9.973304127592581e-05, "loss": 1.2396, "step": 946 }, { "epoch": 0.06416423876956433, "grad_norm": 11.784343719482422, "learning_rate": 9.973167225682799e-05, "loss": 1.214, "step": 947 }, { "epoch": 0.06423199403753642, "grad_norm": 14.205467224121094, "learning_rate": 9.973030323773017e-05, "loss": 1.2803, "step": 948 }, { "epoch": 0.0642997493055085, "grad_norm": 13.257532119750977, "learning_rate": 9.972893421863235e-05, "loss": 1.378, "step": 949 }, { "epoch": 0.06436750457348059, "grad_norm": 15.153338432312012, "learning_rate": 9.972756519953453e-05, "loss": 1.2854, "step": 950 }, { "epoch": 0.06443525984145267, "grad_norm": 16.765771865844727, "learning_rate": 9.972619618043672e-05, "loss": 1.3016, "step": 951 }, { "epoch": 0.06450301510942476, "grad_norm": 14.636106491088867, "learning_rate": 9.97248271613389e-05, "loss": 1.3803, "step": 952 }, { "epoch": 0.06457077037739685, "grad_norm": 13.87410831451416, "learning_rate": 9.972345814224108e-05, "loss": 1.4126, "step": 953 }, { "epoch": 0.06463852564536893, "grad_norm": 14.328899383544922, "learning_rate": 9.972208912314326e-05, "loss": 1.061, "step": 954 }, { "epoch": 0.06470628091334102, "grad_norm": 12.485203742980957, "learning_rate": 9.972072010404544e-05, "loss": 1.0985, "step": 955 }, { "epoch": 0.0647740361813131, "grad_norm": 13.77907943725586, "learning_rate": 9.971935108494764e-05, "loss": 1.3534, "step": 956 }, { "epoch": 0.06484179144928519, "grad_norm": 10.579590797424316, "learning_rate": 9.971798206584982e-05, "loss": 1.166, "step": 957 }, { "epoch": 0.06490954671725727, "grad_norm": 14.690185546875, "learning_rate": 9.9716613046752e-05, "loss": 1.3666, "step": 958 }, { "epoch": 0.06497730198522936, "grad_norm": 12.904786109924316, "learning_rate": 9.97152440276542e-05, "loss": 1.14, "step": 959 }, { "epoch": 0.06504505725320144, "grad_norm": 12.126219749450684, "learning_rate": 9.971387500855637e-05, "loss": 1.4157, "step": 960 }, { "epoch": 0.06511281252117353, "grad_norm": 13.747931480407715, "learning_rate": 9.971250598945855e-05, "loss": 1.4557, "step": 961 }, { "epoch": 0.0651805677891456, "grad_norm": 13.232327461242676, "learning_rate": 9.971113697036075e-05, "loss": 1.3471, "step": 962 }, { "epoch": 0.0652483230571177, "grad_norm": 14.886791229248047, "learning_rate": 9.970976795126293e-05, "loss": 1.1866, "step": 963 }, { "epoch": 0.06531607832508977, "grad_norm": 11.747659683227539, "learning_rate": 9.970839893216511e-05, "loss": 1.0443, "step": 964 }, { "epoch": 0.06538383359306187, "grad_norm": 11.181273460388184, "learning_rate": 9.97070299130673e-05, "loss": 1.1391, "step": 965 }, { "epoch": 0.06545158886103394, "grad_norm": 11.9672269821167, "learning_rate": 9.970566089396948e-05, "loss": 1.2847, "step": 966 }, { "epoch": 0.06551934412900604, "grad_norm": 15.825364112854004, "learning_rate": 9.970429187487166e-05, "loss": 1.3301, "step": 967 }, { "epoch": 0.06558709939697811, "grad_norm": 12.26963996887207, "learning_rate": 9.970292285577384e-05, "loss": 1.2524, "step": 968 }, { "epoch": 0.0656548546649502, "grad_norm": 11.440977096557617, "learning_rate": 9.970155383667604e-05, "loss": 1.229, "step": 969 }, { "epoch": 0.06572260993292228, "grad_norm": 10.704546928405762, "learning_rate": 9.970018481757822e-05, "loss": 0.9936, "step": 970 }, { "epoch": 0.06579036520089437, "grad_norm": 13.20880126953125, "learning_rate": 9.96988157984804e-05, "loss": 1.2353, "step": 971 }, { "epoch": 0.06585812046886645, "grad_norm": 13.101622581481934, "learning_rate": 9.969744677938258e-05, "loss": 1.2559, "step": 972 }, { "epoch": 0.06592587573683854, "grad_norm": 11.725826263427734, "learning_rate": 9.969607776028476e-05, "loss": 1.2051, "step": 973 }, { "epoch": 0.06599363100481062, "grad_norm": 11.890633583068848, "learning_rate": 9.969470874118695e-05, "loss": 1.1607, "step": 974 }, { "epoch": 0.06606138627278271, "grad_norm": 11.066970825195312, "learning_rate": 9.969333972208913e-05, "loss": 1.1031, "step": 975 }, { "epoch": 0.06612914154075479, "grad_norm": 12.26187515258789, "learning_rate": 9.969197070299131e-05, "loss": 1.309, "step": 976 }, { "epoch": 0.06619689680872688, "grad_norm": 13.490363121032715, "learning_rate": 9.96906016838935e-05, "loss": 1.3062, "step": 977 }, { "epoch": 0.06626465207669896, "grad_norm": 12.306289672851562, "learning_rate": 9.968923266479567e-05, "loss": 0.9784, "step": 978 }, { "epoch": 0.06633240734467105, "grad_norm": 10.699983596801758, "learning_rate": 9.968786364569787e-05, "loss": 1.0547, "step": 979 }, { "epoch": 0.06640016261264313, "grad_norm": 12.298179626464844, "learning_rate": 9.968649462660005e-05, "loss": 1.2496, "step": 980 }, { "epoch": 0.06646791788061522, "grad_norm": 15.239167213439941, "learning_rate": 9.968512560750223e-05, "loss": 1.2545, "step": 981 }, { "epoch": 0.0665356731485873, "grad_norm": 11.61802864074707, "learning_rate": 9.968375658840441e-05, "loss": 1.1668, "step": 982 }, { "epoch": 0.06660342841655939, "grad_norm": 12.804032325744629, "learning_rate": 9.96823875693066e-05, "loss": 1.3203, "step": 983 }, { "epoch": 0.06667118368453147, "grad_norm": 14.511723518371582, "learning_rate": 9.968101855020878e-05, "loss": 1.7519, "step": 984 }, { "epoch": 0.06673893895250356, "grad_norm": 12.442008018493652, "learning_rate": 9.967964953111096e-05, "loss": 1.3758, "step": 985 }, { "epoch": 0.06680669422047564, "grad_norm": 14.486754417419434, "learning_rate": 9.967828051201314e-05, "loss": 1.1924, "step": 986 }, { "epoch": 0.06687444948844773, "grad_norm": 13.529693603515625, "learning_rate": 9.967691149291532e-05, "loss": 1.2698, "step": 987 }, { "epoch": 0.06694220475641981, "grad_norm": 12.980225563049316, "learning_rate": 9.967554247381752e-05, "loss": 1.1896, "step": 988 }, { "epoch": 0.0670099600243919, "grad_norm": 15.495257377624512, "learning_rate": 9.96741734547197e-05, "loss": 1.286, "step": 989 }, { "epoch": 0.06707771529236398, "grad_norm": 12.67573070526123, "learning_rate": 9.967280443562188e-05, "loss": 1.1751, "step": 990 }, { "epoch": 0.06714547056033607, "grad_norm": 11.141845703125, "learning_rate": 9.967143541652406e-05, "loss": 1.5109, "step": 991 }, { "epoch": 0.06721322582830815, "grad_norm": 11.975769996643066, "learning_rate": 9.967006639742625e-05, "loss": 1.1579, "step": 992 }, { "epoch": 0.06728098109628024, "grad_norm": 13.872209548950195, "learning_rate": 9.966869737832843e-05, "loss": 1.1852, "step": 993 }, { "epoch": 0.06734873636425232, "grad_norm": 11.52573299407959, "learning_rate": 9.966732835923061e-05, "loss": 1.2319, "step": 994 }, { "epoch": 0.0674164916322244, "grad_norm": 12.90494155883789, "learning_rate": 9.96659593401328e-05, "loss": 1.2415, "step": 995 }, { "epoch": 0.06748424690019648, "grad_norm": 14.210317611694336, "learning_rate": 9.966459032103497e-05, "loss": 1.2901, "step": 996 }, { "epoch": 0.06755200216816858, "grad_norm": 12.600135803222656, "learning_rate": 9.966322130193717e-05, "loss": 1.4516, "step": 997 }, { "epoch": 0.06761975743614065, "grad_norm": 14.462118148803711, "learning_rate": 9.966185228283935e-05, "loss": 1.2205, "step": 998 }, { "epoch": 0.06768751270411275, "grad_norm": 12.870843887329102, "learning_rate": 9.966048326374153e-05, "loss": 1.179, "step": 999 }, { "epoch": 0.06775526797208482, "grad_norm": 16.4424991607666, "learning_rate": 9.965911424464371e-05, "loss": 1.5714, "step": 1000 }, { "epoch": 0.06782302324005691, "grad_norm": 12.902230262756348, "learning_rate": 9.96577452255459e-05, "loss": 1.2803, "step": 1001 }, { "epoch": 0.06789077850802899, "grad_norm": 11.469466209411621, "learning_rate": 9.965637620644808e-05, "loss": 1.0771, "step": 1002 }, { "epoch": 0.06795853377600108, "grad_norm": 13.96650505065918, "learning_rate": 9.965500718735026e-05, "loss": 1.2427, "step": 1003 }, { "epoch": 0.06802628904397316, "grad_norm": 11.55516242980957, "learning_rate": 9.965363816825244e-05, "loss": 1.0396, "step": 1004 }, { "epoch": 0.06809404431194525, "grad_norm": 13.34827709197998, "learning_rate": 9.965226914915462e-05, "loss": 1.181, "step": 1005 }, { "epoch": 0.06816179957991735, "grad_norm": 11.243910789489746, "learning_rate": 9.965090013005682e-05, "loss": 1.2875, "step": 1006 }, { "epoch": 0.06822955484788942, "grad_norm": 14.152894020080566, "learning_rate": 9.9649531110959e-05, "loss": 1.3125, "step": 1007 }, { "epoch": 0.06829731011586151, "grad_norm": 13.010010719299316, "learning_rate": 9.964816209186118e-05, "loss": 1.3213, "step": 1008 }, { "epoch": 0.06836506538383359, "grad_norm": 15.990034103393555, "learning_rate": 9.964679307276337e-05, "loss": 1.1878, "step": 1009 }, { "epoch": 0.06843282065180568, "grad_norm": 12.943589210510254, "learning_rate": 9.964542405366555e-05, "loss": 1.2443, "step": 1010 }, { "epoch": 0.06850057591977776, "grad_norm": 12.108896255493164, "learning_rate": 9.964405503456775e-05, "loss": 1.1719, "step": 1011 }, { "epoch": 0.06856833118774985, "grad_norm": 12.097951889038086, "learning_rate": 9.964268601546993e-05, "loss": 1.0653, "step": 1012 }, { "epoch": 0.06863608645572193, "grad_norm": 14.222228050231934, "learning_rate": 9.964131699637211e-05, "loss": 1.3379, "step": 1013 }, { "epoch": 0.06870384172369402, "grad_norm": 12.636894226074219, "learning_rate": 9.963994797727429e-05, "loss": 1.5559, "step": 1014 }, { "epoch": 0.0687715969916661, "grad_norm": 15.458481788635254, "learning_rate": 9.963857895817648e-05, "loss": 1.2542, "step": 1015 }, { "epoch": 0.06883935225963819, "grad_norm": 11.246847152709961, "learning_rate": 9.963720993907866e-05, "loss": 1.1944, "step": 1016 }, { "epoch": 0.06890710752761027, "grad_norm": 11.699065208435059, "learning_rate": 9.963584091998084e-05, "loss": 1.0739, "step": 1017 }, { "epoch": 0.06897486279558236, "grad_norm": 12.259678840637207, "learning_rate": 9.963447190088302e-05, "loss": 1.1365, "step": 1018 }, { "epoch": 0.06904261806355444, "grad_norm": 13.594696998596191, "learning_rate": 9.96331028817852e-05, "loss": 1.1418, "step": 1019 }, { "epoch": 0.06911037333152653, "grad_norm": 12.90888786315918, "learning_rate": 9.96317338626874e-05, "loss": 1.1987, "step": 1020 }, { "epoch": 0.06917812859949861, "grad_norm": 13.04245662689209, "learning_rate": 9.963036484358958e-05, "loss": 1.4288, "step": 1021 }, { "epoch": 0.0692458838674707, "grad_norm": 12.706077575683594, "learning_rate": 9.962899582449176e-05, "loss": 1.222, "step": 1022 }, { "epoch": 0.06931363913544278, "grad_norm": 14.205679893493652, "learning_rate": 9.962762680539394e-05, "loss": 1.3305, "step": 1023 }, { "epoch": 0.06938139440341487, "grad_norm": 17.09891128540039, "learning_rate": 9.962625778629613e-05, "loss": 1.47, "step": 1024 }, { "epoch": 0.06944914967138695, "grad_norm": 14.603500366210938, "learning_rate": 9.962488876719831e-05, "loss": 1.3887, "step": 1025 }, { "epoch": 0.06951690493935904, "grad_norm": 10.820066452026367, "learning_rate": 9.962351974810049e-05, "loss": 1.1543, "step": 1026 }, { "epoch": 0.06958466020733112, "grad_norm": 10.99889850616455, "learning_rate": 9.962215072900267e-05, "loss": 1.2908, "step": 1027 }, { "epoch": 0.06965241547530321, "grad_norm": 13.470711708068848, "learning_rate": 9.962078170990485e-05, "loss": 1.1422, "step": 1028 }, { "epoch": 0.06972017074327529, "grad_norm": 11.086441040039062, "learning_rate": 9.961941269080705e-05, "loss": 0.9835, "step": 1029 }, { "epoch": 0.06978792601124738, "grad_norm": 16.768535614013672, "learning_rate": 9.961804367170923e-05, "loss": 1.404, "step": 1030 }, { "epoch": 0.06985568127921946, "grad_norm": 15.851200103759766, "learning_rate": 9.961667465261141e-05, "loss": 1.4874, "step": 1031 }, { "epoch": 0.06992343654719155, "grad_norm": 11.995482444763184, "learning_rate": 9.961530563351359e-05, "loss": 1.1497, "step": 1032 }, { "epoch": 0.06999119181516363, "grad_norm": 13.591619491577148, "learning_rate": 9.961393661441577e-05, "loss": 1.4773, "step": 1033 }, { "epoch": 0.07005894708313572, "grad_norm": 16.878938674926758, "learning_rate": 9.961256759531796e-05, "loss": 1.2522, "step": 1034 }, { "epoch": 0.0701267023511078, "grad_norm": 11.901616096496582, "learning_rate": 9.961119857622014e-05, "loss": 1.2164, "step": 1035 }, { "epoch": 0.07019445761907989, "grad_norm": 14.935117721557617, "learning_rate": 9.960982955712232e-05, "loss": 1.1251, "step": 1036 }, { "epoch": 0.07026221288705196, "grad_norm": 12.380253791809082, "learning_rate": 9.96084605380245e-05, "loss": 1.3965, "step": 1037 }, { "epoch": 0.07032996815502406, "grad_norm": 11.645035743713379, "learning_rate": 9.96070915189267e-05, "loss": 1.266, "step": 1038 }, { "epoch": 0.07039772342299613, "grad_norm": 14.525420188903809, "learning_rate": 9.960572249982888e-05, "loss": 1.3991, "step": 1039 }, { "epoch": 0.07046547869096823, "grad_norm": 14.775094985961914, "learning_rate": 9.960435348073106e-05, "loss": 1.5958, "step": 1040 }, { "epoch": 0.0705332339589403, "grad_norm": 10.2192964553833, "learning_rate": 9.960298446163324e-05, "loss": 1.1793, "step": 1041 }, { "epoch": 0.0706009892269124, "grad_norm": 13.074480056762695, "learning_rate": 9.960161544253542e-05, "loss": 1.4243, "step": 1042 }, { "epoch": 0.07066874449488447, "grad_norm": 12.679484367370605, "learning_rate": 9.960024642343761e-05, "loss": 1.3398, "step": 1043 }, { "epoch": 0.07073649976285656, "grad_norm": 9.061332702636719, "learning_rate": 9.95988774043398e-05, "loss": 1.0036, "step": 1044 }, { "epoch": 0.07080425503082864, "grad_norm": 13.423661231994629, "learning_rate": 9.959750838524197e-05, "loss": 1.2767, "step": 1045 }, { "epoch": 0.07087201029880073, "grad_norm": 13.955148696899414, "learning_rate": 9.959613936614415e-05, "loss": 1.311, "step": 1046 }, { "epoch": 0.07093976556677281, "grad_norm": 12.746015548706055, "learning_rate": 9.959477034704635e-05, "loss": 1.42, "step": 1047 }, { "epoch": 0.0710075208347449, "grad_norm": 11.409982681274414, "learning_rate": 9.959340132794853e-05, "loss": 1.3051, "step": 1048 }, { "epoch": 0.07107527610271698, "grad_norm": 11.801681518554688, "learning_rate": 9.959203230885071e-05, "loss": 1.2315, "step": 1049 }, { "epoch": 0.07114303137068907, "grad_norm": 13.041158676147461, "learning_rate": 9.959066328975289e-05, "loss": 1.3487, "step": 1050 }, { "epoch": 0.07121078663866115, "grad_norm": 13.474900245666504, "learning_rate": 9.958929427065507e-05, "loss": 1.4542, "step": 1051 }, { "epoch": 0.07127854190663324, "grad_norm": 12.335237503051758, "learning_rate": 9.958792525155726e-05, "loss": 1.1841, "step": 1052 }, { "epoch": 0.07134629717460532, "grad_norm": 14.909475326538086, "learning_rate": 9.958655623245944e-05, "loss": 1.272, "step": 1053 }, { "epoch": 0.07141405244257741, "grad_norm": 13.449742317199707, "learning_rate": 9.958518721336162e-05, "loss": 1.4881, "step": 1054 }, { "epoch": 0.07148180771054949, "grad_norm": 12.2557954788208, "learning_rate": 9.958381819426382e-05, "loss": 1.2059, "step": 1055 }, { "epoch": 0.07154956297852158, "grad_norm": 13.71298885345459, "learning_rate": 9.9582449175166e-05, "loss": 1.2627, "step": 1056 }, { "epoch": 0.07161731824649366, "grad_norm": 15.4293212890625, "learning_rate": 9.958108015606818e-05, "loss": 1.4663, "step": 1057 }, { "epoch": 0.07168507351446575, "grad_norm": 13.665759086608887, "learning_rate": 9.957971113697037e-05, "loss": 1.4634, "step": 1058 }, { "epoch": 0.07175282878243784, "grad_norm": 13.064310073852539, "learning_rate": 9.957834211787255e-05, "loss": 1.2686, "step": 1059 }, { "epoch": 0.07182058405040992, "grad_norm": 15.21746826171875, "learning_rate": 9.957697309877473e-05, "loss": 1.4861, "step": 1060 }, { "epoch": 0.07188833931838201, "grad_norm": 12.499883651733398, "learning_rate": 9.957560407967693e-05, "loss": 1.5275, "step": 1061 }, { "epoch": 0.07195609458635409, "grad_norm": 9.355907440185547, "learning_rate": 9.957423506057911e-05, "loss": 1.1509, "step": 1062 }, { "epoch": 0.07202384985432618, "grad_norm": 10.428252220153809, "learning_rate": 9.957286604148129e-05, "loss": 1.0874, "step": 1063 }, { "epoch": 0.07209160512229826, "grad_norm": 11.351346015930176, "learning_rate": 9.957149702238347e-05, "loss": 1.0887, "step": 1064 }, { "epoch": 0.07215936039027035, "grad_norm": 15.271830558776855, "learning_rate": 9.957012800328565e-05, "loss": 1.2466, "step": 1065 }, { "epoch": 0.07222711565824243, "grad_norm": 11.172418594360352, "learning_rate": 9.956875898418784e-05, "loss": 1.0958, "step": 1066 }, { "epoch": 0.07229487092621452, "grad_norm": 14.221702575683594, "learning_rate": 9.956738996509002e-05, "loss": 1.2288, "step": 1067 }, { "epoch": 0.0723626261941866, "grad_norm": 12.167356491088867, "learning_rate": 9.95660209459922e-05, "loss": 1.1159, "step": 1068 }, { "epoch": 0.07243038146215869, "grad_norm": 12.607329368591309, "learning_rate": 9.956465192689438e-05, "loss": 1.2475, "step": 1069 }, { "epoch": 0.07249813673013077, "grad_norm": 10.59451675415039, "learning_rate": 9.956328290779658e-05, "loss": 1.1493, "step": 1070 }, { "epoch": 0.07256589199810286, "grad_norm": 11.190742492675781, "learning_rate": 9.956191388869876e-05, "loss": 1.4549, "step": 1071 }, { "epoch": 0.07263364726607494, "grad_norm": 11.225564956665039, "learning_rate": 9.956054486960094e-05, "loss": 1.0741, "step": 1072 }, { "epoch": 0.07270140253404703, "grad_norm": 13.648139953613281, "learning_rate": 9.955917585050312e-05, "loss": 1.2942, "step": 1073 }, { "epoch": 0.0727691578020191, "grad_norm": 14.606941223144531, "learning_rate": 9.95578068314053e-05, "loss": 1.253, "step": 1074 }, { "epoch": 0.0728369130699912, "grad_norm": 11.043729782104492, "learning_rate": 9.955643781230749e-05, "loss": 1.2279, "step": 1075 }, { "epoch": 0.07290466833796327, "grad_norm": 12.463634490966797, "learning_rate": 9.955506879320967e-05, "loss": 1.4399, "step": 1076 }, { "epoch": 0.07297242360593537, "grad_norm": 13.74101734161377, "learning_rate": 9.955369977411185e-05, "loss": 1.4323, "step": 1077 }, { "epoch": 0.07304017887390744, "grad_norm": 10.1694974899292, "learning_rate": 9.955233075501403e-05, "loss": 1.0113, "step": 1078 }, { "epoch": 0.07310793414187954, "grad_norm": 14.407991409301758, "learning_rate": 9.955096173591623e-05, "loss": 1.7102, "step": 1079 }, { "epoch": 0.07317568940985161, "grad_norm": 13.84760570526123, "learning_rate": 9.954959271681841e-05, "loss": 1.4491, "step": 1080 }, { "epoch": 0.0732434446778237, "grad_norm": 12.220841407775879, "learning_rate": 9.954822369772059e-05, "loss": 1.1826, "step": 1081 }, { "epoch": 0.07331119994579578, "grad_norm": 11.380377769470215, "learning_rate": 9.954685467862277e-05, "loss": 1.3537, "step": 1082 }, { "epoch": 0.07337895521376787, "grad_norm": 12.916484832763672, "learning_rate": 9.954548565952495e-05, "loss": 1.3406, "step": 1083 }, { "epoch": 0.07344671048173995, "grad_norm": 14.107590675354004, "learning_rate": 9.954411664042714e-05, "loss": 1.0531, "step": 1084 }, { "epoch": 0.07351446574971204, "grad_norm": 13.498798370361328, "learning_rate": 9.954274762132932e-05, "loss": 1.1546, "step": 1085 }, { "epoch": 0.07358222101768412, "grad_norm": 13.155747413635254, "learning_rate": 9.95413786022315e-05, "loss": 1.212, "step": 1086 }, { "epoch": 0.07364997628565621, "grad_norm": 14.135687828063965, "learning_rate": 9.954000958313368e-05, "loss": 1.2001, "step": 1087 }, { "epoch": 0.07371773155362829, "grad_norm": 13.123790740966797, "learning_rate": 9.953864056403586e-05, "loss": 1.2735, "step": 1088 }, { "epoch": 0.07378548682160038, "grad_norm": 14.591660499572754, "learning_rate": 9.953727154493806e-05, "loss": 1.2977, "step": 1089 }, { "epoch": 0.07385324208957246, "grad_norm": 13.30614948272705, "learning_rate": 9.953590252584024e-05, "loss": 1.2445, "step": 1090 }, { "epoch": 0.07392099735754455, "grad_norm": 14.542524337768555, "learning_rate": 9.953453350674242e-05, "loss": 1.2327, "step": 1091 }, { "epoch": 0.07398875262551663, "grad_norm": 12.926522254943848, "learning_rate": 9.95331644876446e-05, "loss": 1.2139, "step": 1092 }, { "epoch": 0.07405650789348872, "grad_norm": 11.622479438781738, "learning_rate": 9.953179546854679e-05, "loss": 1.2887, "step": 1093 }, { "epoch": 0.0741242631614608, "grad_norm": 14.87485122680664, "learning_rate": 9.953042644944897e-05, "loss": 1.3132, "step": 1094 }, { "epoch": 0.07419201842943289, "grad_norm": 13.096991539001465, "learning_rate": 9.952905743035115e-05, "loss": 1.0783, "step": 1095 }, { "epoch": 0.07425977369740497, "grad_norm": 14.676551818847656, "learning_rate": 9.952768841125333e-05, "loss": 1.3171, "step": 1096 }, { "epoch": 0.07432752896537706, "grad_norm": 10.399755477905273, "learning_rate": 9.952631939215551e-05, "loss": 1.0747, "step": 1097 }, { "epoch": 0.07439528423334914, "grad_norm": 11.052745819091797, "learning_rate": 9.952495037305771e-05, "loss": 1.1653, "step": 1098 }, { "epoch": 0.07446303950132123, "grad_norm": 13.014352798461914, "learning_rate": 9.952358135395989e-05, "loss": 1.3727, "step": 1099 }, { "epoch": 0.07453079476929331, "grad_norm": 13.202262878417969, "learning_rate": 9.952221233486207e-05, "loss": 1.5063, "step": 1100 }, { "epoch": 0.0745985500372654, "grad_norm": 14.272111892700195, "learning_rate": 9.952084331576426e-05, "loss": 1.0183, "step": 1101 }, { "epoch": 0.07466630530523748, "grad_norm": 14.23975658416748, "learning_rate": 9.951947429666644e-05, "loss": 1.3518, "step": 1102 }, { "epoch": 0.07473406057320957, "grad_norm": 12.293742179870605, "learning_rate": 9.951810527756862e-05, "loss": 1.4021, "step": 1103 }, { "epoch": 0.07480181584118165, "grad_norm": 16.026020050048828, "learning_rate": 9.951673625847082e-05, "loss": 1.3412, "step": 1104 }, { "epoch": 0.07486957110915374, "grad_norm": 11.120819091796875, "learning_rate": 9.9515367239373e-05, "loss": 1.331, "step": 1105 }, { "epoch": 0.07493732637712582, "grad_norm": 12.363526344299316, "learning_rate": 9.951399822027518e-05, "loss": 1.4284, "step": 1106 }, { "epoch": 0.0750050816450979, "grad_norm": 14.377492904663086, "learning_rate": 9.951262920117737e-05, "loss": 1.3686, "step": 1107 }, { "epoch": 0.07507283691306998, "grad_norm": 11.366288185119629, "learning_rate": 9.951126018207955e-05, "loss": 1.2329, "step": 1108 }, { "epoch": 0.07514059218104208, "grad_norm": 10.26131820678711, "learning_rate": 9.950989116298173e-05, "loss": 1.1091, "step": 1109 }, { "epoch": 0.07520834744901415, "grad_norm": 12.64631462097168, "learning_rate": 9.950852214388391e-05, "loss": 1.1231, "step": 1110 }, { "epoch": 0.07527610271698625, "grad_norm": 13.042781829833984, "learning_rate": 9.95071531247861e-05, "loss": 1.2338, "step": 1111 }, { "epoch": 0.07534385798495834, "grad_norm": 11.577115058898926, "learning_rate": 9.950578410568829e-05, "loss": 1.2143, "step": 1112 }, { "epoch": 0.07541161325293042, "grad_norm": 13.640811920166016, "learning_rate": 9.950441508659047e-05, "loss": 1.224, "step": 1113 }, { "epoch": 0.0754793685209025, "grad_norm": 10.271018981933594, "learning_rate": 9.950304606749265e-05, "loss": 1.0459, "step": 1114 }, { "epoch": 0.07554712378887458, "grad_norm": 12.053836822509766, "learning_rate": 9.950167704839483e-05, "loss": 1.2137, "step": 1115 }, { "epoch": 0.07561487905684668, "grad_norm": 10.022509574890137, "learning_rate": 9.950030802929702e-05, "loss": 1.092, "step": 1116 }, { "epoch": 0.07568263432481875, "grad_norm": 12.28339672088623, "learning_rate": 9.94989390101992e-05, "loss": 1.1616, "step": 1117 }, { "epoch": 0.07575038959279085, "grad_norm": 14.182686805725098, "learning_rate": 9.949756999110138e-05, "loss": 1.5135, "step": 1118 }, { "epoch": 0.07581814486076292, "grad_norm": 10.668661117553711, "learning_rate": 9.949620097200356e-05, "loss": 1.2412, "step": 1119 }, { "epoch": 0.07588590012873501, "grad_norm": 14.443583488464355, "learning_rate": 9.949483195290574e-05, "loss": 1.207, "step": 1120 }, { "epoch": 0.07595365539670709, "grad_norm": 12.418794631958008, "learning_rate": 9.949346293380794e-05, "loss": 1.1907, "step": 1121 }, { "epoch": 0.07602141066467918, "grad_norm": 12.429618835449219, "learning_rate": 9.949209391471012e-05, "loss": 1.4403, "step": 1122 }, { "epoch": 0.07608916593265126, "grad_norm": 9.524622917175293, "learning_rate": 9.94907248956123e-05, "loss": 1.139, "step": 1123 }, { "epoch": 0.07615692120062335, "grad_norm": 10.974812507629395, "learning_rate": 9.948935587651448e-05, "loss": 1.4055, "step": 1124 }, { "epoch": 0.07622467646859543, "grad_norm": 16.152681350708008, "learning_rate": 9.948798685741667e-05, "loss": 1.0874, "step": 1125 }, { "epoch": 0.07629243173656752, "grad_norm": 12.023541450500488, "learning_rate": 9.948661783831885e-05, "loss": 1.2378, "step": 1126 }, { "epoch": 0.0763601870045396, "grad_norm": 11.597234725952148, "learning_rate": 9.948524881922103e-05, "loss": 1.0104, "step": 1127 }, { "epoch": 0.07642794227251169, "grad_norm": 11.374302864074707, "learning_rate": 9.948387980012321e-05, "loss": 1.1099, "step": 1128 }, { "epoch": 0.07649569754048377, "grad_norm": 14.398423194885254, "learning_rate": 9.94825107810254e-05, "loss": 1.5533, "step": 1129 }, { "epoch": 0.07656345280845586, "grad_norm": 13.1026611328125, "learning_rate": 9.948114176192759e-05, "loss": 1.3948, "step": 1130 }, { "epoch": 0.07663120807642794, "grad_norm": 12.012560844421387, "learning_rate": 9.947977274282977e-05, "loss": 1.2156, "step": 1131 }, { "epoch": 0.07669896334440003, "grad_norm": 12.900229454040527, "learning_rate": 9.947840372373195e-05, "loss": 1.2658, "step": 1132 }, { "epoch": 0.07676671861237211, "grad_norm": 15.250492095947266, "learning_rate": 9.947703470463413e-05, "loss": 1.6102, "step": 1133 }, { "epoch": 0.0768344738803442, "grad_norm": 15.08134937286377, "learning_rate": 9.947566568553632e-05, "loss": 1.1947, "step": 1134 }, { "epoch": 0.07690222914831628, "grad_norm": 13.353601455688477, "learning_rate": 9.94742966664385e-05, "loss": 1.3211, "step": 1135 }, { "epoch": 0.07696998441628837, "grad_norm": 11.310175895690918, "learning_rate": 9.947292764734068e-05, "loss": 1.2223, "step": 1136 }, { "epoch": 0.07703773968426045, "grad_norm": 11.800848960876465, "learning_rate": 9.947155862824286e-05, "loss": 1.2131, "step": 1137 }, { "epoch": 0.07710549495223254, "grad_norm": 11.748014450073242, "learning_rate": 9.947018960914504e-05, "loss": 1.0734, "step": 1138 }, { "epoch": 0.07717325022020462, "grad_norm": 12.282258033752441, "learning_rate": 9.946882059004724e-05, "loss": 1.3257, "step": 1139 }, { "epoch": 0.07724100548817671, "grad_norm": 11.93818473815918, "learning_rate": 9.946745157094942e-05, "loss": 1.1452, "step": 1140 }, { "epoch": 0.07730876075614879, "grad_norm": 13.397029876708984, "learning_rate": 9.94660825518516e-05, "loss": 1.2625, "step": 1141 }, { "epoch": 0.07737651602412088, "grad_norm": 12.135769844055176, "learning_rate": 9.946471353275378e-05, "loss": 1.3624, "step": 1142 }, { "epoch": 0.07744427129209296, "grad_norm": 11.304028511047363, "learning_rate": 9.946334451365596e-05, "loss": 1.3019, "step": 1143 }, { "epoch": 0.07751202656006505, "grad_norm": 10.978137016296387, "learning_rate": 9.946197549455815e-05, "loss": 1.3137, "step": 1144 }, { "epoch": 0.07757978182803713, "grad_norm": 10.997323989868164, "learning_rate": 9.946060647546033e-05, "loss": 1.3573, "step": 1145 }, { "epoch": 0.07764753709600922, "grad_norm": 11.883647918701172, "learning_rate": 9.945923745636251e-05, "loss": 1.2587, "step": 1146 }, { "epoch": 0.0777152923639813, "grad_norm": 10.70753288269043, "learning_rate": 9.945786843726471e-05, "loss": 1.0866, "step": 1147 }, { "epoch": 0.07778304763195339, "grad_norm": 13.318743705749512, "learning_rate": 9.945649941816689e-05, "loss": 1.3356, "step": 1148 }, { "epoch": 0.07785080289992546, "grad_norm": 10.656171798706055, "learning_rate": 9.945513039906907e-05, "loss": 1.2155, "step": 1149 }, { "epoch": 0.07791855816789756, "grad_norm": 14.73982048034668, "learning_rate": 9.945376137997126e-05, "loss": 1.2276, "step": 1150 }, { "epoch": 0.07798631343586963, "grad_norm": 12.461714744567871, "learning_rate": 9.945239236087344e-05, "loss": 1.2999, "step": 1151 }, { "epoch": 0.07805406870384173, "grad_norm": 12.277376174926758, "learning_rate": 9.945102334177562e-05, "loss": 1.1131, "step": 1152 }, { "epoch": 0.0781218239718138, "grad_norm": 12.726540565490723, "learning_rate": 9.944965432267782e-05, "loss": 1.4181, "step": 1153 }, { "epoch": 0.0781895792397859, "grad_norm": 11.086180686950684, "learning_rate": 9.944828530358e-05, "loss": 1.3852, "step": 1154 }, { "epoch": 0.07825733450775797, "grad_norm": 12.80537223815918, "learning_rate": 9.944691628448218e-05, "loss": 1.1101, "step": 1155 }, { "epoch": 0.07832508977573006, "grad_norm": 11.960269927978516, "learning_rate": 9.944554726538436e-05, "loss": 1.2186, "step": 1156 }, { "epoch": 0.07839284504370214, "grad_norm": 11.732439041137695, "learning_rate": 9.944417824628655e-05, "loss": 0.9854, "step": 1157 }, { "epoch": 0.07846060031167423, "grad_norm": 12.578715324401855, "learning_rate": 9.944280922718873e-05, "loss": 1.0792, "step": 1158 }, { "epoch": 0.07852835557964631, "grad_norm": 13.779712677001953, "learning_rate": 9.944144020809091e-05, "loss": 1.2468, "step": 1159 }, { "epoch": 0.0785961108476184, "grad_norm": 13.95693588256836, "learning_rate": 9.944007118899309e-05, "loss": 1.234, "step": 1160 }, { "epoch": 0.07866386611559048, "grad_norm": 12.056897163391113, "learning_rate": 9.943870216989527e-05, "loss": 1.3034, "step": 1161 }, { "epoch": 0.07873162138356257, "grad_norm": 11.633442878723145, "learning_rate": 9.943733315079747e-05, "loss": 1.2744, "step": 1162 }, { "epoch": 0.07879937665153465, "grad_norm": 14.062381744384766, "learning_rate": 9.943596413169965e-05, "loss": 1.2535, "step": 1163 }, { "epoch": 0.07886713191950674, "grad_norm": 11.60498332977295, "learning_rate": 9.943459511260183e-05, "loss": 1.1578, "step": 1164 }, { "epoch": 0.07893488718747883, "grad_norm": 9.667806625366211, "learning_rate": 9.943322609350401e-05, "loss": 1.0804, "step": 1165 }, { "epoch": 0.07900264245545091, "grad_norm": 12.30827808380127, "learning_rate": 9.943185707440619e-05, "loss": 1.4167, "step": 1166 }, { "epoch": 0.079070397723423, "grad_norm": 10.196819305419922, "learning_rate": 9.943048805530838e-05, "loss": 1.0547, "step": 1167 }, { "epoch": 0.07913815299139508, "grad_norm": 10.029928207397461, "learning_rate": 9.942911903621056e-05, "loss": 1.1798, "step": 1168 }, { "epoch": 0.07920590825936717, "grad_norm": 11.782978057861328, "learning_rate": 9.942775001711274e-05, "loss": 1.1454, "step": 1169 }, { "epoch": 0.07927366352733925, "grad_norm": 12.25143814086914, "learning_rate": 9.942638099801492e-05, "loss": 1.2854, "step": 1170 }, { "epoch": 0.07934141879531134, "grad_norm": 8.47904109954834, "learning_rate": 9.942501197891712e-05, "loss": 0.9381, "step": 1171 }, { "epoch": 0.07940917406328342, "grad_norm": 13.698802947998047, "learning_rate": 9.94236429598193e-05, "loss": 1.6261, "step": 1172 }, { "epoch": 0.07947692933125551, "grad_norm": 10.885397911071777, "learning_rate": 9.942227394072148e-05, "loss": 1.0645, "step": 1173 }, { "epoch": 0.07954468459922759, "grad_norm": 13.274818420410156, "learning_rate": 9.942090492162366e-05, "loss": 1.3411, "step": 1174 }, { "epoch": 0.07961243986719968, "grad_norm": 14.061238288879395, "learning_rate": 9.941953590252584e-05, "loss": 1.1916, "step": 1175 }, { "epoch": 0.07968019513517176, "grad_norm": 10.074264526367188, "learning_rate": 9.941816688342803e-05, "loss": 1.092, "step": 1176 }, { "epoch": 0.07974795040314385, "grad_norm": 14.741287231445312, "learning_rate": 9.941679786433021e-05, "loss": 1.3774, "step": 1177 }, { "epoch": 0.07981570567111593, "grad_norm": 11.308422088623047, "learning_rate": 9.941542884523239e-05, "loss": 0.9298, "step": 1178 }, { "epoch": 0.07988346093908802, "grad_norm": 14.375280380249023, "learning_rate": 9.941405982613457e-05, "loss": 1.4525, "step": 1179 }, { "epoch": 0.0799512162070601, "grad_norm": 11.724523544311523, "learning_rate": 9.941269080703677e-05, "loss": 1.3616, "step": 1180 }, { "epoch": 0.08001897147503219, "grad_norm": 12.578176498413086, "learning_rate": 9.941132178793895e-05, "loss": 1.1513, "step": 1181 }, { "epoch": 0.08008672674300427, "grad_norm": 12.100804328918457, "learning_rate": 9.940995276884113e-05, "loss": 1.0656, "step": 1182 }, { "epoch": 0.08015448201097636, "grad_norm": 11.680248260498047, "learning_rate": 9.940858374974331e-05, "loss": 1.2358, "step": 1183 }, { "epoch": 0.08022223727894844, "grad_norm": 10.066198348999023, "learning_rate": 9.940721473064549e-05, "loss": 0.9219, "step": 1184 }, { "epoch": 0.08028999254692053, "grad_norm": 10.813334465026855, "learning_rate": 9.940584571154768e-05, "loss": 1.1456, "step": 1185 }, { "epoch": 0.0803577478148926, "grad_norm": 14.004862785339355, "learning_rate": 9.940447669244986e-05, "loss": 1.2375, "step": 1186 }, { "epoch": 0.0804255030828647, "grad_norm": 11.868766784667969, "learning_rate": 9.940310767335204e-05, "loss": 1.3027, "step": 1187 }, { "epoch": 0.08049325835083677, "grad_norm": 12.48153018951416, "learning_rate": 9.940173865425422e-05, "loss": 1.1192, "step": 1188 }, { "epoch": 0.08056101361880887, "grad_norm": 12.340612411499023, "learning_rate": 9.940036963515642e-05, "loss": 1.246, "step": 1189 }, { "epoch": 0.08062876888678094, "grad_norm": 12.205392837524414, "learning_rate": 9.93990006160586e-05, "loss": 1.1233, "step": 1190 }, { "epoch": 0.08069652415475304, "grad_norm": 12.69509220123291, "learning_rate": 9.939763159696078e-05, "loss": 1.2202, "step": 1191 }, { "epoch": 0.08076427942272511, "grad_norm": 12.40784740447998, "learning_rate": 9.939626257786296e-05, "loss": 1.1345, "step": 1192 }, { "epoch": 0.0808320346906972, "grad_norm": 11.678507804870605, "learning_rate": 9.939489355876515e-05, "loss": 1.4513, "step": 1193 }, { "epoch": 0.08089978995866928, "grad_norm": 11.649873733520508, "learning_rate": 9.939352453966733e-05, "loss": 1.3827, "step": 1194 }, { "epoch": 0.08096754522664137, "grad_norm": 12.378853797912598, "learning_rate": 9.939215552056951e-05, "loss": 1.3632, "step": 1195 }, { "epoch": 0.08103530049461345, "grad_norm": 11.023188591003418, "learning_rate": 9.93907865014717e-05, "loss": 1.3892, "step": 1196 }, { "epoch": 0.08110305576258554, "grad_norm": 13.111897468566895, "learning_rate": 9.938941748237389e-05, "loss": 1.1973, "step": 1197 }, { "epoch": 0.08117081103055762, "grad_norm": 10.171613693237305, "learning_rate": 9.938804846327607e-05, "loss": 1.3862, "step": 1198 }, { "epoch": 0.08123856629852971, "grad_norm": 13.327658653259277, "learning_rate": 9.938667944417826e-05, "loss": 1.3173, "step": 1199 }, { "epoch": 0.08130632156650179, "grad_norm": 11.715154647827148, "learning_rate": 9.938531042508044e-05, "loss": 1.0839, "step": 1200 }, { "epoch": 0.08137407683447388, "grad_norm": 11.521212577819824, "learning_rate": 9.938394140598262e-05, "loss": 1.0285, "step": 1201 }, { "epoch": 0.08144183210244596, "grad_norm": 12.024236679077148, "learning_rate": 9.93825723868848e-05, "loss": 1.3025, "step": 1202 }, { "epoch": 0.08150958737041805, "grad_norm": 10.245376586914062, "learning_rate": 9.9381203367787e-05, "loss": 1.1658, "step": 1203 }, { "epoch": 0.08157734263839013, "grad_norm": 10.731759071350098, "learning_rate": 9.937983434868918e-05, "loss": 1.0565, "step": 1204 }, { "epoch": 0.08164509790636222, "grad_norm": 16.714153289794922, "learning_rate": 9.937846532959136e-05, "loss": 1.2761, "step": 1205 }, { "epoch": 0.0817128531743343, "grad_norm": 11.172699928283691, "learning_rate": 9.937709631049354e-05, "loss": 1.1887, "step": 1206 }, { "epoch": 0.08178060844230639, "grad_norm": 11.384743690490723, "learning_rate": 9.937572729139572e-05, "loss": 0.9546, "step": 1207 }, { "epoch": 0.08184836371027847, "grad_norm": 11.163822174072266, "learning_rate": 9.937435827229791e-05, "loss": 1.1498, "step": 1208 }, { "epoch": 0.08191611897825056, "grad_norm": 10.657593727111816, "learning_rate": 9.937298925320009e-05, "loss": 0.9506, "step": 1209 }, { "epoch": 0.08198387424622264, "grad_norm": 10.6862211227417, "learning_rate": 9.937162023410227e-05, "loss": 1.2308, "step": 1210 }, { "epoch": 0.08205162951419473, "grad_norm": 10.649473190307617, "learning_rate": 9.937025121500445e-05, "loss": 1.2679, "step": 1211 }, { "epoch": 0.08211938478216681, "grad_norm": 9.298782348632812, "learning_rate": 9.936888219590665e-05, "loss": 1.1327, "step": 1212 }, { "epoch": 0.0821871400501389, "grad_norm": 11.448348045349121, "learning_rate": 9.936751317680883e-05, "loss": 1.2796, "step": 1213 }, { "epoch": 0.08225489531811098, "grad_norm": 12.141517639160156, "learning_rate": 9.936614415771101e-05, "loss": 1.1128, "step": 1214 }, { "epoch": 0.08232265058608307, "grad_norm": 11.799830436706543, "learning_rate": 9.936477513861319e-05, "loss": 1.0743, "step": 1215 }, { "epoch": 0.08239040585405515, "grad_norm": 11.952958106994629, "learning_rate": 9.936340611951537e-05, "loss": 1.1659, "step": 1216 }, { "epoch": 0.08245816112202724, "grad_norm": 11.870144844055176, "learning_rate": 9.936203710041756e-05, "loss": 1.1195, "step": 1217 }, { "epoch": 0.08252591638999933, "grad_norm": 11.970368385314941, "learning_rate": 9.936066808131974e-05, "loss": 1.2139, "step": 1218 }, { "epoch": 0.08259367165797141, "grad_norm": 11.211687088012695, "learning_rate": 9.935929906222192e-05, "loss": 1.2107, "step": 1219 }, { "epoch": 0.0826614269259435, "grad_norm": 12.987563133239746, "learning_rate": 9.93579300431241e-05, "loss": 1.2804, "step": 1220 }, { "epoch": 0.08272918219391558, "grad_norm": 12.337888717651367, "learning_rate": 9.935656102402628e-05, "loss": 0.8713, "step": 1221 }, { "epoch": 0.08279693746188767, "grad_norm": 11.717150688171387, "learning_rate": 9.935519200492848e-05, "loss": 1.2684, "step": 1222 }, { "epoch": 0.08286469272985975, "grad_norm": 12.658769607543945, "learning_rate": 9.935382298583066e-05, "loss": 1.0231, "step": 1223 }, { "epoch": 0.08293244799783184, "grad_norm": 10.526476860046387, "learning_rate": 9.935245396673284e-05, "loss": 1.0605, "step": 1224 }, { "epoch": 0.08300020326580392, "grad_norm": 10.575004577636719, "learning_rate": 9.935108494763502e-05, "loss": 1.2169, "step": 1225 }, { "epoch": 0.083067958533776, "grad_norm": 10.499407768249512, "learning_rate": 9.934971592853721e-05, "loss": 1.3064, "step": 1226 }, { "epoch": 0.08313571380174808, "grad_norm": 12.25387191772461, "learning_rate": 9.934834690943939e-05, "loss": 1.1554, "step": 1227 }, { "epoch": 0.08320346906972018, "grad_norm": 12.173775672912598, "learning_rate": 9.934697789034157e-05, "loss": 1.0917, "step": 1228 }, { "epoch": 0.08327122433769225, "grad_norm": 12.424721717834473, "learning_rate": 9.934560887124375e-05, "loss": 1.3679, "step": 1229 }, { "epoch": 0.08333897960566435, "grad_norm": 11.291987419128418, "learning_rate": 9.934423985214593e-05, "loss": 1.1439, "step": 1230 }, { "epoch": 0.08340673487363642, "grad_norm": 10.908637046813965, "learning_rate": 9.934287083304813e-05, "loss": 1.0487, "step": 1231 }, { "epoch": 0.08347449014160851, "grad_norm": 14.767544746398926, "learning_rate": 9.934150181395031e-05, "loss": 1.2951, "step": 1232 }, { "epoch": 0.08354224540958059, "grad_norm": 11.959871292114258, "learning_rate": 9.934013279485249e-05, "loss": 1.1244, "step": 1233 }, { "epoch": 0.08361000067755268, "grad_norm": 11.19450569152832, "learning_rate": 9.933876377575467e-05, "loss": 1.2802, "step": 1234 }, { "epoch": 0.08367775594552476, "grad_norm": 10.71377182006836, "learning_rate": 9.933739475665686e-05, "loss": 1.0376, "step": 1235 }, { "epoch": 0.08374551121349685, "grad_norm": 12.174454689025879, "learning_rate": 9.933602573755904e-05, "loss": 1.5779, "step": 1236 }, { "epoch": 0.08381326648146893, "grad_norm": 9.863836288452148, "learning_rate": 9.933465671846122e-05, "loss": 1.0443, "step": 1237 }, { "epoch": 0.08388102174944102, "grad_norm": 11.964838027954102, "learning_rate": 9.93332876993634e-05, "loss": 1.0613, "step": 1238 }, { "epoch": 0.0839487770174131, "grad_norm": 11.49203109741211, "learning_rate": 9.93319186802656e-05, "loss": 1.162, "step": 1239 }, { "epoch": 0.08401653228538519, "grad_norm": 12.548815727233887, "learning_rate": 9.933054966116778e-05, "loss": 1.0652, "step": 1240 }, { "epoch": 0.08408428755335727, "grad_norm": 13.37637996673584, "learning_rate": 9.932918064206996e-05, "loss": 1.2281, "step": 1241 }, { "epoch": 0.08415204282132936, "grad_norm": 12.59211254119873, "learning_rate": 9.932781162297215e-05, "loss": 1.0624, "step": 1242 }, { "epoch": 0.08421979808930144, "grad_norm": 13.386221885681152, "learning_rate": 9.932644260387433e-05, "loss": 1.4393, "step": 1243 }, { "epoch": 0.08428755335727353, "grad_norm": 12.647525787353516, "learning_rate": 9.932507358477651e-05, "loss": 1.2228, "step": 1244 }, { "epoch": 0.08435530862524561, "grad_norm": 12.039474487304688, "learning_rate": 9.93237045656787e-05, "loss": 1.1354, "step": 1245 }, { "epoch": 0.0844230638932177, "grad_norm": 11.373556137084961, "learning_rate": 9.932233554658089e-05, "loss": 1.2418, "step": 1246 }, { "epoch": 0.08449081916118978, "grad_norm": 10.944781303405762, "learning_rate": 9.932096652748307e-05, "loss": 1.1515, "step": 1247 }, { "epoch": 0.08455857442916187, "grad_norm": 12.174854278564453, "learning_rate": 9.931959750838525e-05, "loss": 1.2063, "step": 1248 }, { "epoch": 0.08462632969713395, "grad_norm": 8.846879005432129, "learning_rate": 9.931822848928744e-05, "loss": 0.9639, "step": 1249 }, { "epoch": 0.08469408496510604, "grad_norm": 13.793547630310059, "learning_rate": 9.931685947018962e-05, "loss": 1.1236, "step": 1250 }, { "epoch": 0.08476184023307812, "grad_norm": 14.486831665039062, "learning_rate": 9.93154904510918e-05, "loss": 1.2047, "step": 1251 }, { "epoch": 0.08482959550105021, "grad_norm": 13.262588500976562, "learning_rate": 9.931412143199398e-05, "loss": 1.4091, "step": 1252 }, { "epoch": 0.08489735076902229, "grad_norm": 13.289068222045898, "learning_rate": 9.931275241289616e-05, "loss": 1.5065, "step": 1253 }, { "epoch": 0.08496510603699438, "grad_norm": 10.22205638885498, "learning_rate": 9.931138339379836e-05, "loss": 1.1116, "step": 1254 }, { "epoch": 0.08503286130496646, "grad_norm": 13.141668319702148, "learning_rate": 9.931001437470054e-05, "loss": 1.4006, "step": 1255 }, { "epoch": 0.08510061657293855, "grad_norm": 11.817032814025879, "learning_rate": 9.930864535560272e-05, "loss": 1.2062, "step": 1256 }, { "epoch": 0.08516837184091063, "grad_norm": 10.814498901367188, "learning_rate": 9.93072763365049e-05, "loss": 1.138, "step": 1257 }, { "epoch": 0.08523612710888272, "grad_norm": 12.598155975341797, "learning_rate": 9.930590731740709e-05, "loss": 1.1959, "step": 1258 }, { "epoch": 0.0853038823768548, "grad_norm": 11.909974098205566, "learning_rate": 9.930453829830927e-05, "loss": 1.2891, "step": 1259 }, { "epoch": 0.08537163764482689, "grad_norm": 12.663064956665039, "learning_rate": 9.930316927921145e-05, "loss": 1.3017, "step": 1260 }, { "epoch": 0.08543939291279896, "grad_norm": 10.810627937316895, "learning_rate": 9.930180026011363e-05, "loss": 1.16, "step": 1261 }, { "epoch": 0.08550714818077106, "grad_norm": 9.182926177978516, "learning_rate": 9.930043124101581e-05, "loss": 1.2625, "step": 1262 }, { "epoch": 0.08557490344874313, "grad_norm": 12.73978042602539, "learning_rate": 9.9299062221918e-05, "loss": 0.9904, "step": 1263 }, { "epoch": 0.08564265871671523, "grad_norm": 10.346587181091309, "learning_rate": 9.929769320282019e-05, "loss": 1.0167, "step": 1264 }, { "epoch": 0.0857104139846873, "grad_norm": 12.145682334899902, "learning_rate": 9.929632418372237e-05, "loss": 1.4282, "step": 1265 }, { "epoch": 0.0857781692526594, "grad_norm": 11.515445709228516, "learning_rate": 9.929495516462455e-05, "loss": 1.0527, "step": 1266 }, { "epoch": 0.08584592452063147, "grad_norm": 10.979050636291504, "learning_rate": 9.929358614552674e-05, "loss": 1.2888, "step": 1267 }, { "epoch": 0.08591367978860356, "grad_norm": 10.248215675354004, "learning_rate": 9.929221712642892e-05, "loss": 1.4316, "step": 1268 }, { "epoch": 0.08598143505657564, "grad_norm": 13.743851661682129, "learning_rate": 9.92908481073311e-05, "loss": 1.5047, "step": 1269 }, { "epoch": 0.08604919032454773, "grad_norm": 9.919225692749023, "learning_rate": 9.928947908823328e-05, "loss": 0.9602, "step": 1270 }, { "epoch": 0.08611694559251983, "grad_norm": 9.795915603637695, "learning_rate": 9.928811006913546e-05, "loss": 1.3736, "step": 1271 }, { "epoch": 0.0861847008604919, "grad_norm": 13.108200073242188, "learning_rate": 9.928674105003766e-05, "loss": 1.2591, "step": 1272 }, { "epoch": 0.086252456128464, "grad_norm": 13.108073234558105, "learning_rate": 9.928537203093984e-05, "loss": 1.102, "step": 1273 }, { "epoch": 0.08632021139643607, "grad_norm": 15.177817344665527, "learning_rate": 9.928400301184202e-05, "loss": 1.3164, "step": 1274 }, { "epoch": 0.08638796666440816, "grad_norm": 13.048440933227539, "learning_rate": 9.92826339927442e-05, "loss": 1.3029, "step": 1275 }, { "epoch": 0.08645572193238024, "grad_norm": 10.982895851135254, "learning_rate": 9.928126497364638e-05, "loss": 1.1829, "step": 1276 }, { "epoch": 0.08652347720035233, "grad_norm": 10.901629447937012, "learning_rate": 9.927989595454857e-05, "loss": 1.2116, "step": 1277 }, { "epoch": 0.08659123246832441, "grad_norm": 12.924722671508789, "learning_rate": 9.927852693545075e-05, "loss": 1.6292, "step": 1278 }, { "epoch": 0.0866589877362965, "grad_norm": 12.572770118713379, "learning_rate": 9.927715791635293e-05, "loss": 1.1012, "step": 1279 }, { "epoch": 0.08672674300426858, "grad_norm": 12.803020477294922, "learning_rate": 9.927578889725511e-05, "loss": 1.276, "step": 1280 }, { "epoch": 0.08679449827224067, "grad_norm": 10.92810344696045, "learning_rate": 9.92744198781573e-05, "loss": 1.1876, "step": 1281 }, { "epoch": 0.08686225354021275, "grad_norm": 12.253180503845215, "learning_rate": 9.927305085905949e-05, "loss": 1.1807, "step": 1282 }, { "epoch": 0.08693000880818484, "grad_norm": 11.163126945495605, "learning_rate": 9.927168183996167e-05, "loss": 1.1671, "step": 1283 }, { "epoch": 0.08699776407615692, "grad_norm": 10.726607322692871, "learning_rate": 9.927031282086385e-05, "loss": 1.185, "step": 1284 }, { "epoch": 0.08706551934412901, "grad_norm": 13.265491485595703, "learning_rate": 9.926894380176603e-05, "loss": 1.2916, "step": 1285 }, { "epoch": 0.08713327461210109, "grad_norm": 14.559592247009277, "learning_rate": 9.926757478266822e-05, "loss": 1.288, "step": 1286 }, { "epoch": 0.08720102988007318, "grad_norm": 14.816813468933105, "learning_rate": 9.92662057635704e-05, "loss": 1.4397, "step": 1287 }, { "epoch": 0.08726878514804526, "grad_norm": 11.811420440673828, "learning_rate": 9.926483674447258e-05, "loss": 1.348, "step": 1288 }, { "epoch": 0.08733654041601735, "grad_norm": 10.920133590698242, "learning_rate": 9.926346772537478e-05, "loss": 1.4169, "step": 1289 }, { "epoch": 0.08740429568398943, "grad_norm": 11.690089225769043, "learning_rate": 9.926209870627696e-05, "loss": 1.1199, "step": 1290 }, { "epoch": 0.08747205095196152, "grad_norm": 9.411031723022461, "learning_rate": 9.926072968717914e-05, "loss": 1.031, "step": 1291 }, { "epoch": 0.0875398062199336, "grad_norm": 12.174457550048828, "learning_rate": 9.925936066808133e-05, "loss": 1.0622, "step": 1292 }, { "epoch": 0.08760756148790569, "grad_norm": 10.346089363098145, "learning_rate": 9.925799164898351e-05, "loss": 1.2777, "step": 1293 }, { "epoch": 0.08767531675587777, "grad_norm": 12.534863471984863, "learning_rate": 9.925662262988569e-05, "loss": 1.0093, "step": 1294 }, { "epoch": 0.08774307202384986, "grad_norm": 12.050302505493164, "learning_rate": 9.925525361078789e-05, "loss": 1.0591, "step": 1295 }, { "epoch": 0.08781082729182194, "grad_norm": 11.556166648864746, "learning_rate": 9.925388459169007e-05, "loss": 1.1452, "step": 1296 }, { "epoch": 0.08787858255979403, "grad_norm": 9.693270683288574, "learning_rate": 9.925251557259225e-05, "loss": 1.0332, "step": 1297 }, { "epoch": 0.0879463378277661, "grad_norm": 12.646526336669922, "learning_rate": 9.925114655349443e-05, "loss": 1.1894, "step": 1298 }, { "epoch": 0.0880140930957382, "grad_norm": 10.676809310913086, "learning_rate": 9.924977753439661e-05, "loss": 1.1727, "step": 1299 }, { "epoch": 0.08808184836371027, "grad_norm": 11.182327270507812, "learning_rate": 9.92484085152988e-05, "loss": 1.0145, "step": 1300 }, { "epoch": 0.08814960363168237, "grad_norm": 11.55026626586914, "learning_rate": 9.924703949620098e-05, "loss": 1.2187, "step": 1301 }, { "epoch": 0.08821735889965444, "grad_norm": 11.502679824829102, "learning_rate": 9.924567047710316e-05, "loss": 1.3116, "step": 1302 }, { "epoch": 0.08828511416762654, "grad_norm": 9.676247596740723, "learning_rate": 9.924430145800534e-05, "loss": 1.2869, "step": 1303 }, { "epoch": 0.08835286943559861, "grad_norm": 15.123950004577637, "learning_rate": 9.924293243890754e-05, "loss": 1.0225, "step": 1304 }, { "epoch": 0.0884206247035707, "grad_norm": 14.030994415283203, "learning_rate": 9.924156341980972e-05, "loss": 0.9829, "step": 1305 }, { "epoch": 0.08848837997154278, "grad_norm": 10.00402545928955, "learning_rate": 9.92401944007119e-05, "loss": 1.0946, "step": 1306 }, { "epoch": 0.08855613523951487, "grad_norm": 9.077853202819824, "learning_rate": 9.923882538161408e-05, "loss": 0.852, "step": 1307 }, { "epoch": 0.08862389050748695, "grad_norm": 12.777885437011719, "learning_rate": 9.923745636251626e-05, "loss": 1.0513, "step": 1308 }, { "epoch": 0.08869164577545904, "grad_norm": 10.686469078063965, "learning_rate": 9.923608734341845e-05, "loss": 1.0873, "step": 1309 }, { "epoch": 0.08875940104343112, "grad_norm": 11.51689338684082, "learning_rate": 9.923471832432063e-05, "loss": 1.0037, "step": 1310 }, { "epoch": 0.08882715631140321, "grad_norm": 13.259784698486328, "learning_rate": 9.923334930522281e-05, "loss": 1.2004, "step": 1311 }, { "epoch": 0.08889491157937529, "grad_norm": 10.63463306427002, "learning_rate": 9.923198028612499e-05, "loss": 1.2067, "step": 1312 }, { "epoch": 0.08896266684734738, "grad_norm": 10.838210105895996, "learning_rate": 9.923061126702719e-05, "loss": 1.2939, "step": 1313 }, { "epoch": 0.08903042211531946, "grad_norm": 12.058418273925781, "learning_rate": 9.922924224792937e-05, "loss": 1.3091, "step": 1314 }, { "epoch": 0.08909817738329155, "grad_norm": 10.619451522827148, "learning_rate": 9.922787322883155e-05, "loss": 1.1635, "step": 1315 }, { "epoch": 0.08916593265126363, "grad_norm": 9.37607192993164, "learning_rate": 9.922650420973373e-05, "loss": 1.2206, "step": 1316 }, { "epoch": 0.08923368791923572, "grad_norm": 12.119776725769043, "learning_rate": 9.922513519063591e-05, "loss": 1.2159, "step": 1317 }, { "epoch": 0.0893014431872078, "grad_norm": 14.013461112976074, "learning_rate": 9.92237661715381e-05, "loss": 1.1445, "step": 1318 }, { "epoch": 0.08936919845517989, "grad_norm": 11.560707092285156, "learning_rate": 9.922239715244028e-05, "loss": 1.3617, "step": 1319 }, { "epoch": 0.08943695372315197, "grad_norm": 11.817791938781738, "learning_rate": 9.922102813334246e-05, "loss": 1.1931, "step": 1320 }, { "epoch": 0.08950470899112406, "grad_norm": 11.544127464294434, "learning_rate": 9.921965911424464e-05, "loss": 1.0766, "step": 1321 }, { "epoch": 0.08957246425909614, "grad_norm": 10.67740249633789, "learning_rate": 9.921829009514684e-05, "loss": 1.2103, "step": 1322 }, { "epoch": 0.08964021952706823, "grad_norm": 11.774645805358887, "learning_rate": 9.921692107604902e-05, "loss": 1.1921, "step": 1323 }, { "epoch": 0.08970797479504032, "grad_norm": 12.594759941101074, "learning_rate": 9.92155520569512e-05, "loss": 0.906, "step": 1324 }, { "epoch": 0.0897757300630124, "grad_norm": 10.988224983215332, "learning_rate": 9.921418303785338e-05, "loss": 1.4195, "step": 1325 }, { "epoch": 0.08984348533098449, "grad_norm": 12.22718620300293, "learning_rate": 9.921281401875556e-05, "loss": 1.2461, "step": 1326 }, { "epoch": 0.08991124059895657, "grad_norm": 10.964727401733398, "learning_rate": 9.921144499965775e-05, "loss": 1.1254, "step": 1327 }, { "epoch": 0.08997899586692866, "grad_norm": 11.327523231506348, "learning_rate": 9.921007598055993e-05, "loss": 1.145, "step": 1328 }, { "epoch": 0.09004675113490074, "grad_norm": 9.870691299438477, "learning_rate": 9.920870696146211e-05, "loss": 1.2408, "step": 1329 }, { "epoch": 0.09011450640287283, "grad_norm": 11.02373218536377, "learning_rate": 9.920733794236429e-05, "loss": 1.1084, "step": 1330 }, { "epoch": 0.09018226167084491, "grad_norm": 12.300410270690918, "learning_rate": 9.920596892326647e-05, "loss": 1.1542, "step": 1331 }, { "epoch": 0.090250016938817, "grad_norm": 9.832919120788574, "learning_rate": 9.920459990416867e-05, "loss": 1.1195, "step": 1332 }, { "epoch": 0.09031777220678908, "grad_norm": 10.432522773742676, "learning_rate": 9.920323088507085e-05, "loss": 1.1825, "step": 1333 }, { "epoch": 0.09038552747476117, "grad_norm": 11.878792762756348, "learning_rate": 9.920186186597303e-05, "loss": 0.9647, "step": 1334 }, { "epoch": 0.09045328274273325, "grad_norm": 11.866320610046387, "learning_rate": 9.920049284687522e-05, "loss": 1.3598, "step": 1335 }, { "epoch": 0.09052103801070534, "grad_norm": 14.11543083190918, "learning_rate": 9.91991238277774e-05, "loss": 1.4172, "step": 1336 }, { "epoch": 0.09058879327867742, "grad_norm": 13.841622352600098, "learning_rate": 9.919775480867958e-05, "loss": 1.4119, "step": 1337 }, { "epoch": 0.0906565485466495, "grad_norm": 11.077167510986328, "learning_rate": 9.919638578958178e-05, "loss": 0.984, "step": 1338 }, { "epoch": 0.09072430381462158, "grad_norm": 10.904266357421875, "learning_rate": 9.919501677048396e-05, "loss": 1.0439, "step": 1339 }, { "epoch": 0.09079205908259368, "grad_norm": 11.623948097229004, "learning_rate": 9.919364775138614e-05, "loss": 1.0318, "step": 1340 }, { "epoch": 0.09085981435056575, "grad_norm": 10.893725395202637, "learning_rate": 9.919227873228833e-05, "loss": 1.2626, "step": 1341 }, { "epoch": 0.09092756961853785, "grad_norm": 10.064491271972656, "learning_rate": 9.919090971319051e-05, "loss": 1.1483, "step": 1342 }, { "epoch": 0.09099532488650992, "grad_norm": 9.854101181030273, "learning_rate": 9.918954069409269e-05, "loss": 1.1595, "step": 1343 }, { "epoch": 0.09106308015448202, "grad_norm": 11.682498931884766, "learning_rate": 9.918817167499487e-05, "loss": 1.2887, "step": 1344 }, { "epoch": 0.09113083542245409, "grad_norm": 10.484097480773926, "learning_rate": 9.918680265589707e-05, "loss": 1.2, "step": 1345 }, { "epoch": 0.09119859069042618, "grad_norm": 12.332358360290527, "learning_rate": 9.918543363679925e-05, "loss": 1.262, "step": 1346 }, { "epoch": 0.09126634595839826, "grad_norm": 13.706925392150879, "learning_rate": 9.918406461770143e-05, "loss": 1.2151, "step": 1347 }, { "epoch": 0.09133410122637035, "grad_norm": 13.918478965759277, "learning_rate": 9.91826955986036e-05, "loss": 1.4653, "step": 1348 }, { "epoch": 0.09140185649434243, "grad_norm": 11.970015525817871, "learning_rate": 9.918132657950579e-05, "loss": 1.2452, "step": 1349 }, { "epoch": 0.09146961176231452, "grad_norm": 9.698074340820312, "learning_rate": 9.917995756040798e-05, "loss": 0.9142, "step": 1350 }, { "epoch": 0.0915373670302866, "grad_norm": 9.225728988647461, "learning_rate": 9.917858854131016e-05, "loss": 1.0293, "step": 1351 }, { "epoch": 0.09160512229825869, "grad_norm": 10.77661418914795, "learning_rate": 9.917721952221234e-05, "loss": 1.1865, "step": 1352 }, { "epoch": 0.09167287756623077, "grad_norm": 10.45409870147705, "learning_rate": 9.917585050311452e-05, "loss": 1.1553, "step": 1353 }, { "epoch": 0.09174063283420286, "grad_norm": 10.44918441772461, "learning_rate": 9.91744814840167e-05, "loss": 1.1621, "step": 1354 }, { "epoch": 0.09180838810217494, "grad_norm": 14.769590377807617, "learning_rate": 9.91731124649189e-05, "loss": 1.3502, "step": 1355 }, { "epoch": 0.09187614337014703, "grad_norm": 12.733844757080078, "learning_rate": 9.917174344582108e-05, "loss": 1.1973, "step": 1356 }, { "epoch": 0.09194389863811911, "grad_norm": 13.365818977355957, "learning_rate": 9.917037442672326e-05, "loss": 1.5412, "step": 1357 }, { "epoch": 0.0920116539060912, "grad_norm": 11.163050651550293, "learning_rate": 9.916900540762544e-05, "loss": 1.3142, "step": 1358 }, { "epoch": 0.09207940917406328, "grad_norm": 11.420190811157227, "learning_rate": 9.916763638852763e-05, "loss": 1.2257, "step": 1359 }, { "epoch": 0.09214716444203537, "grad_norm": 9.66398811340332, "learning_rate": 9.916626736942981e-05, "loss": 1.0584, "step": 1360 }, { "epoch": 0.09221491971000745, "grad_norm": 13.643363952636719, "learning_rate": 9.916489835033199e-05, "loss": 1.3783, "step": 1361 }, { "epoch": 0.09228267497797954, "grad_norm": 11.658889770507812, "learning_rate": 9.916352933123417e-05, "loss": 1.2126, "step": 1362 }, { "epoch": 0.09235043024595162, "grad_norm": 11.2728271484375, "learning_rate": 9.916216031213635e-05, "loss": 1.2117, "step": 1363 }, { "epoch": 0.09241818551392371, "grad_norm": 13.576864242553711, "learning_rate": 9.916079129303855e-05, "loss": 1.2922, "step": 1364 }, { "epoch": 0.09248594078189579, "grad_norm": 10.950700759887695, "learning_rate": 9.915942227394073e-05, "loss": 1.1394, "step": 1365 }, { "epoch": 0.09255369604986788, "grad_norm": 11.638351440429688, "learning_rate": 9.91580532548429e-05, "loss": 1.3927, "step": 1366 }, { "epoch": 0.09262145131783996, "grad_norm": 12.355545043945312, "learning_rate": 9.915668423574509e-05, "loss": 1.1861, "step": 1367 }, { "epoch": 0.09268920658581205, "grad_norm": 11.543237686157227, "learning_rate": 9.915531521664728e-05, "loss": 0.9868, "step": 1368 }, { "epoch": 0.09275696185378413, "grad_norm": 11.684252738952637, "learning_rate": 9.915394619754946e-05, "loss": 1.3332, "step": 1369 }, { "epoch": 0.09282471712175622, "grad_norm": 10.775650024414062, "learning_rate": 9.915257717845164e-05, "loss": 1.1164, "step": 1370 }, { "epoch": 0.0928924723897283, "grad_norm": 11.649751663208008, "learning_rate": 9.915120815935382e-05, "loss": 1.0369, "step": 1371 }, { "epoch": 0.09296022765770039, "grad_norm": 9.741403579711914, "learning_rate": 9.9149839140256e-05, "loss": 1.0509, "step": 1372 }, { "epoch": 0.09302798292567246, "grad_norm": 13.804118156433105, "learning_rate": 9.91484701211582e-05, "loss": 1.4318, "step": 1373 }, { "epoch": 0.09309573819364456, "grad_norm": 10.939459800720215, "learning_rate": 9.914710110206038e-05, "loss": 1.218, "step": 1374 }, { "epoch": 0.09316349346161663, "grad_norm": 14.076252937316895, "learning_rate": 9.914573208296256e-05, "loss": 1.4699, "step": 1375 }, { "epoch": 0.09323124872958873, "grad_norm": 12.90072250366211, "learning_rate": 9.914436306386474e-05, "loss": 1.3418, "step": 1376 }, { "epoch": 0.09329900399756082, "grad_norm": 12.7711820602417, "learning_rate": 9.914299404476693e-05, "loss": 0.9866, "step": 1377 }, { "epoch": 0.0933667592655329, "grad_norm": 9.205671310424805, "learning_rate": 9.914162502566911e-05, "loss": 1.1345, "step": 1378 }, { "epoch": 0.09343451453350499, "grad_norm": 11.027194023132324, "learning_rate": 9.914025600657129e-05, "loss": 1.2637, "step": 1379 }, { "epoch": 0.09350226980147706, "grad_norm": 12.861044883728027, "learning_rate": 9.913888698747347e-05, "loss": 1.2348, "step": 1380 }, { "epoch": 0.09357002506944916, "grad_norm": 11.266969680786133, "learning_rate": 9.913751796837567e-05, "loss": 1.0773, "step": 1381 }, { "epoch": 0.09363778033742123, "grad_norm": 13.137110710144043, "learning_rate": 9.913614894927785e-05, "loss": 1.0537, "step": 1382 }, { "epoch": 0.09370553560539333, "grad_norm": 11.343362808227539, "learning_rate": 9.913477993018003e-05, "loss": 1.335, "step": 1383 }, { "epoch": 0.0937732908733654, "grad_norm": 11.472663879394531, "learning_rate": 9.913341091108222e-05, "loss": 1.1362, "step": 1384 }, { "epoch": 0.0938410461413375, "grad_norm": 8.441573143005371, "learning_rate": 9.91320418919844e-05, "loss": 1.1646, "step": 1385 }, { "epoch": 0.09390880140930957, "grad_norm": 12.570130348205566, "learning_rate": 9.913067287288658e-05, "loss": 1.3535, "step": 1386 }, { "epoch": 0.09397655667728166, "grad_norm": 11.671664237976074, "learning_rate": 9.912930385378877e-05, "loss": 1.1707, "step": 1387 }, { "epoch": 0.09404431194525374, "grad_norm": 12.638328552246094, "learning_rate": 9.912793483469096e-05, "loss": 1.057, "step": 1388 }, { "epoch": 0.09411206721322583, "grad_norm": 10.506028175354004, "learning_rate": 9.912656581559314e-05, "loss": 0.9279, "step": 1389 }, { "epoch": 0.09417982248119791, "grad_norm": 11.536858558654785, "learning_rate": 9.912519679649532e-05, "loss": 1.3351, "step": 1390 }, { "epoch": 0.09424757774917, "grad_norm": 12.692436218261719, "learning_rate": 9.912382777739751e-05, "loss": 1.164, "step": 1391 }, { "epoch": 0.09431533301714208, "grad_norm": 12.088066101074219, "learning_rate": 9.912245875829969e-05, "loss": 1.2971, "step": 1392 }, { "epoch": 0.09438308828511417, "grad_norm": 12.133123397827148, "learning_rate": 9.912108973920187e-05, "loss": 1.2742, "step": 1393 }, { "epoch": 0.09445084355308625, "grad_norm": 10.168001174926758, "learning_rate": 9.911972072010405e-05, "loss": 1.0964, "step": 1394 }, { "epoch": 0.09451859882105834, "grad_norm": 10.561311721801758, "learning_rate": 9.911835170100623e-05, "loss": 1.1828, "step": 1395 }, { "epoch": 0.09458635408903042, "grad_norm": 11.497330665588379, "learning_rate": 9.911698268190843e-05, "loss": 1.4699, "step": 1396 }, { "epoch": 0.09465410935700251, "grad_norm": 12.190573692321777, "learning_rate": 9.91156136628106e-05, "loss": 1.1601, "step": 1397 }, { "epoch": 0.09472186462497459, "grad_norm": 10.633028030395508, "learning_rate": 9.911424464371279e-05, "loss": 1.333, "step": 1398 }, { "epoch": 0.09478961989294668, "grad_norm": 12.262279510498047, "learning_rate": 9.911287562461497e-05, "loss": 1.2214, "step": 1399 }, { "epoch": 0.09485737516091876, "grad_norm": 11.506840705871582, "learning_rate": 9.911150660551716e-05, "loss": 1.1172, "step": 1400 }, { "epoch": 0.09492513042889085, "grad_norm": 11.453936576843262, "learning_rate": 9.911013758641934e-05, "loss": 1.2205, "step": 1401 }, { "epoch": 0.09499288569686293, "grad_norm": 9.980772972106934, "learning_rate": 9.910876856732152e-05, "loss": 1.1004, "step": 1402 }, { "epoch": 0.09506064096483502, "grad_norm": 11.775416374206543, "learning_rate": 9.91073995482237e-05, "loss": 1.1587, "step": 1403 }, { "epoch": 0.0951283962328071, "grad_norm": 8.840147972106934, "learning_rate": 9.910603052912588e-05, "loss": 0.8702, "step": 1404 }, { "epoch": 0.09519615150077919, "grad_norm": 10.938506126403809, "learning_rate": 9.910466151002808e-05, "loss": 1.085, "step": 1405 }, { "epoch": 0.09526390676875127, "grad_norm": 11.733402252197266, "learning_rate": 9.910329249093026e-05, "loss": 1.1202, "step": 1406 }, { "epoch": 0.09533166203672336, "grad_norm": 11.616521835327148, "learning_rate": 9.910192347183244e-05, "loss": 1.3483, "step": 1407 }, { "epoch": 0.09539941730469544, "grad_norm": 12.477338790893555, "learning_rate": 9.910055445273462e-05, "loss": 1.4798, "step": 1408 }, { "epoch": 0.09546717257266753, "grad_norm": 11.233193397521973, "learning_rate": 9.90991854336368e-05, "loss": 1.0304, "step": 1409 }, { "epoch": 0.0955349278406396, "grad_norm": 12.586124420166016, "learning_rate": 9.909781641453899e-05, "loss": 1.2216, "step": 1410 }, { "epoch": 0.0956026831086117, "grad_norm": 12.974738121032715, "learning_rate": 9.909644739544117e-05, "loss": 1.1495, "step": 1411 }, { "epoch": 0.09567043837658377, "grad_norm": 9.613628387451172, "learning_rate": 9.909507837634335e-05, "loss": 0.8326, "step": 1412 }, { "epoch": 0.09573819364455587, "grad_norm": 10.644312858581543, "learning_rate": 9.909370935724553e-05, "loss": 1.0182, "step": 1413 }, { "epoch": 0.09580594891252794, "grad_norm": 11.155874252319336, "learning_rate": 9.909234033814773e-05, "loss": 1.2887, "step": 1414 }, { "epoch": 0.09587370418050004, "grad_norm": 12.068909645080566, "learning_rate": 9.90909713190499e-05, "loss": 1.1697, "step": 1415 }, { "epoch": 0.09594145944847211, "grad_norm": 10.66831111907959, "learning_rate": 9.908960229995209e-05, "loss": 1.1275, "step": 1416 }, { "epoch": 0.0960092147164442, "grad_norm": 11.80036449432373, "learning_rate": 9.908823328085427e-05, "loss": 1.38, "step": 1417 }, { "epoch": 0.09607696998441628, "grad_norm": 11.677534103393555, "learning_rate": 9.908686426175645e-05, "loss": 1.075, "step": 1418 }, { "epoch": 0.09614472525238837, "grad_norm": 10.54027271270752, "learning_rate": 9.908549524265864e-05, "loss": 0.9617, "step": 1419 }, { "epoch": 0.09621248052036045, "grad_norm": 9.70718002319336, "learning_rate": 9.908412622356082e-05, "loss": 1.0395, "step": 1420 }, { "epoch": 0.09628023578833254, "grad_norm": 10.439559936523438, "learning_rate": 9.9082757204463e-05, "loss": 1.4112, "step": 1421 }, { "epoch": 0.09634799105630462, "grad_norm": 9.328675270080566, "learning_rate": 9.908138818536518e-05, "loss": 1.0481, "step": 1422 }, { "epoch": 0.09641574632427671, "grad_norm": 12.834508895874023, "learning_rate": 9.908001916626738e-05, "loss": 1.0863, "step": 1423 }, { "epoch": 0.09648350159224879, "grad_norm": 11.885201454162598, "learning_rate": 9.907865014716956e-05, "loss": 1.3509, "step": 1424 }, { "epoch": 0.09655125686022088, "grad_norm": 11.299174308776855, "learning_rate": 9.907728112807174e-05, "loss": 1.1328, "step": 1425 }, { "epoch": 0.09661901212819296, "grad_norm": 13.024226188659668, "learning_rate": 9.907591210897392e-05, "loss": 1.1495, "step": 1426 }, { "epoch": 0.09668676739616505, "grad_norm": 13.418682098388672, "learning_rate": 9.907454308987611e-05, "loss": 1.3827, "step": 1427 }, { "epoch": 0.09675452266413713, "grad_norm": 11.28375244140625, "learning_rate": 9.907317407077829e-05, "loss": 1.3658, "step": 1428 }, { "epoch": 0.09682227793210922, "grad_norm": 9.711199760437012, "learning_rate": 9.907180505168047e-05, "loss": 0.8743, "step": 1429 }, { "epoch": 0.09689003320008131, "grad_norm": 12.292948722839355, "learning_rate": 9.907043603258267e-05, "loss": 1.203, "step": 1430 }, { "epoch": 0.09695778846805339, "grad_norm": 13.195072174072266, "learning_rate": 9.906906701348485e-05, "loss": 1.0403, "step": 1431 }, { "epoch": 0.09702554373602548, "grad_norm": 11.45721435546875, "learning_rate": 9.906769799438703e-05, "loss": 1.2075, "step": 1432 }, { "epoch": 0.09709329900399756, "grad_norm": 10.477989196777344, "learning_rate": 9.906632897528922e-05, "loss": 1.1646, "step": 1433 }, { "epoch": 0.09716105427196965, "grad_norm": 12.572269439697266, "learning_rate": 9.90649599561914e-05, "loss": 1.3944, "step": 1434 }, { "epoch": 0.09722880953994173, "grad_norm": 9.37205982208252, "learning_rate": 9.906359093709358e-05, "loss": 0.9902, "step": 1435 }, { "epoch": 0.09729656480791382, "grad_norm": 11.590779304504395, "learning_rate": 9.906222191799576e-05, "loss": 1.2658, "step": 1436 }, { "epoch": 0.0973643200758859, "grad_norm": 10.35207748413086, "learning_rate": 9.906085289889795e-05, "loss": 0.7905, "step": 1437 }, { "epoch": 0.09743207534385799, "grad_norm": 9.993937492370605, "learning_rate": 9.905948387980013e-05, "loss": 1.1522, "step": 1438 }, { "epoch": 0.09749983061183007, "grad_norm": 9.865569114685059, "learning_rate": 9.905811486070232e-05, "loss": 1.2536, "step": 1439 }, { "epoch": 0.09756758587980216, "grad_norm": 12.836588859558105, "learning_rate": 9.90567458416045e-05, "loss": 1.2216, "step": 1440 }, { "epoch": 0.09763534114777424, "grad_norm": 10.062298774719238, "learning_rate": 9.905537682250668e-05, "loss": 1.1576, "step": 1441 }, { "epoch": 0.09770309641574633, "grad_norm": 10.897071838378906, "learning_rate": 9.905400780340887e-05, "loss": 1.1315, "step": 1442 }, { "epoch": 0.09777085168371841, "grad_norm": 10.366122245788574, "learning_rate": 9.905263878431105e-05, "loss": 1.274, "step": 1443 }, { "epoch": 0.0978386069516905, "grad_norm": 11.632966995239258, "learning_rate": 9.905126976521323e-05, "loss": 1.1427, "step": 1444 }, { "epoch": 0.09790636221966258, "grad_norm": 10.537737846374512, "learning_rate": 9.904990074611541e-05, "loss": 1.0913, "step": 1445 }, { "epoch": 0.09797411748763467, "grad_norm": 9.52363109588623, "learning_rate": 9.90485317270176e-05, "loss": 0.8677, "step": 1446 }, { "epoch": 0.09804187275560675, "grad_norm": 11.511491775512695, "learning_rate": 9.904716270791979e-05, "loss": 1.0381, "step": 1447 }, { "epoch": 0.09810962802357884, "grad_norm": 12.085793495178223, "learning_rate": 9.904579368882197e-05, "loss": 1.1415, "step": 1448 }, { "epoch": 0.09817738329155092, "grad_norm": 8.665430068969727, "learning_rate": 9.904442466972415e-05, "loss": 1.0367, "step": 1449 }, { "epoch": 0.09824513855952301, "grad_norm": 10.900618553161621, "learning_rate": 9.904305565062633e-05, "loss": 0.9835, "step": 1450 }, { "epoch": 0.09831289382749508, "grad_norm": 10.3113431930542, "learning_rate": 9.904168663152852e-05, "loss": 0.924, "step": 1451 }, { "epoch": 0.09838064909546718, "grad_norm": 10.001591682434082, "learning_rate": 9.90403176124307e-05, "loss": 1.0835, "step": 1452 }, { "epoch": 0.09844840436343925, "grad_norm": 11.333273887634277, "learning_rate": 9.903894859333288e-05, "loss": 1.072, "step": 1453 }, { "epoch": 0.09851615963141135, "grad_norm": 10.107904434204102, "learning_rate": 9.903757957423506e-05, "loss": 1.0848, "step": 1454 }, { "epoch": 0.09858391489938342, "grad_norm": 12.578730583190918, "learning_rate": 9.903621055513725e-05, "loss": 1.2735, "step": 1455 }, { "epoch": 0.09865167016735552, "grad_norm": 10.453478813171387, "learning_rate": 9.903484153603944e-05, "loss": 1.3141, "step": 1456 }, { "epoch": 0.09871942543532759, "grad_norm": 10.383566856384277, "learning_rate": 9.903347251694162e-05, "loss": 1.0992, "step": 1457 }, { "epoch": 0.09878718070329968, "grad_norm": 9.612902641296387, "learning_rate": 9.90321034978438e-05, "loss": 1.3103, "step": 1458 }, { "epoch": 0.09885493597127176, "grad_norm": 12.111359596252441, "learning_rate": 9.903073447874598e-05, "loss": 1.3102, "step": 1459 }, { "epoch": 0.09892269123924385, "grad_norm": 9.987195014953613, "learning_rate": 9.902936545964817e-05, "loss": 1.1961, "step": 1460 }, { "epoch": 0.09899044650721593, "grad_norm": 10.900408744812012, "learning_rate": 9.902799644055035e-05, "loss": 1.0208, "step": 1461 }, { "epoch": 0.09905820177518802, "grad_norm": 9.94915771484375, "learning_rate": 9.902662742145253e-05, "loss": 1.3347, "step": 1462 }, { "epoch": 0.0991259570431601, "grad_norm": 13.393661499023438, "learning_rate": 9.902525840235471e-05, "loss": 1.2903, "step": 1463 }, { "epoch": 0.09919371231113219, "grad_norm": 10.122967720031738, "learning_rate": 9.902388938325689e-05, "loss": 1.0229, "step": 1464 }, { "epoch": 0.09926146757910427, "grad_norm": 10.775031089782715, "learning_rate": 9.902252036415909e-05, "loss": 1.1669, "step": 1465 }, { "epoch": 0.09932922284707636, "grad_norm": 9.733497619628906, "learning_rate": 9.902115134506127e-05, "loss": 1.0903, "step": 1466 }, { "epoch": 0.09939697811504844, "grad_norm": 9.230277061462402, "learning_rate": 9.901978232596345e-05, "loss": 1.1738, "step": 1467 }, { "epoch": 0.09946473338302053, "grad_norm": 10.822884559631348, "learning_rate": 9.901841330686563e-05, "loss": 1.3392, "step": 1468 }, { "epoch": 0.09953248865099261, "grad_norm": 10.64195442199707, "learning_rate": 9.901704428776782e-05, "loss": 1.0823, "step": 1469 }, { "epoch": 0.0996002439189647, "grad_norm": 13.73645305633545, "learning_rate": 9.901567526867e-05, "loss": 1.1128, "step": 1470 }, { "epoch": 0.09966799918693678, "grad_norm": 11.361958503723145, "learning_rate": 9.901430624957218e-05, "loss": 1.083, "step": 1471 }, { "epoch": 0.09973575445490887, "grad_norm": 10.839045524597168, "learning_rate": 9.901293723047436e-05, "loss": 1.0288, "step": 1472 }, { "epoch": 0.09980350972288095, "grad_norm": 10.41995906829834, "learning_rate": 9.901156821137656e-05, "loss": 1.0631, "step": 1473 }, { "epoch": 0.09987126499085304, "grad_norm": 11.87709903717041, "learning_rate": 9.901019919227874e-05, "loss": 1.0715, "step": 1474 }, { "epoch": 0.09993902025882512, "grad_norm": 10.46670913696289, "learning_rate": 9.900883017318092e-05, "loss": 1.1684, "step": 1475 }, { "epoch": 0.10000677552679721, "grad_norm": 12.163457870483398, "learning_rate": 9.900746115408311e-05, "loss": 1.4416, "step": 1476 }, { "epoch": 0.10007453079476929, "grad_norm": 13.417581558227539, "learning_rate": 9.900609213498529e-05, "loss": 1.1876, "step": 1477 }, { "epoch": 0.10014228606274138, "grad_norm": 11.35722541809082, "learning_rate": 9.900472311588747e-05, "loss": 1.1389, "step": 1478 }, { "epoch": 0.10021004133071346, "grad_norm": 10.042820930480957, "learning_rate": 9.900335409678966e-05, "loss": 1.2156, "step": 1479 }, { "epoch": 0.10027779659868555, "grad_norm": 10.823782920837402, "learning_rate": 9.900198507769184e-05, "loss": 1.1915, "step": 1480 }, { "epoch": 0.10034555186665763, "grad_norm": 13.6808443069458, "learning_rate": 9.900061605859403e-05, "loss": 1.1724, "step": 1481 }, { "epoch": 0.10041330713462972, "grad_norm": 11.16846752166748, "learning_rate": 9.89992470394962e-05, "loss": 1.2093, "step": 1482 }, { "epoch": 0.10048106240260181, "grad_norm": 10.391450881958008, "learning_rate": 9.89978780203984e-05, "loss": 1.142, "step": 1483 }, { "epoch": 0.10054881767057389, "grad_norm": 9.324288368225098, "learning_rate": 9.899650900130058e-05, "loss": 1.0579, "step": 1484 }, { "epoch": 0.10061657293854598, "grad_norm": 12.601625442504883, "learning_rate": 9.899513998220276e-05, "loss": 1.2643, "step": 1485 }, { "epoch": 0.10068432820651806, "grad_norm": 14.270779609680176, "learning_rate": 9.899377096310494e-05, "loss": 1.2903, "step": 1486 }, { "epoch": 0.10075208347449015, "grad_norm": 11.521232604980469, "learning_rate": 9.899240194400712e-05, "loss": 1.2385, "step": 1487 }, { "epoch": 0.10081983874246223, "grad_norm": 10.76693344116211, "learning_rate": 9.899103292490931e-05, "loss": 1.1745, "step": 1488 }, { "epoch": 0.10088759401043432, "grad_norm": 9.091184616088867, "learning_rate": 9.89896639058115e-05, "loss": 0.9571, "step": 1489 }, { "epoch": 0.1009553492784064, "grad_norm": 11.930106163024902, "learning_rate": 9.898829488671368e-05, "loss": 1.1226, "step": 1490 }, { "epoch": 0.10102310454637849, "grad_norm": 10.90937614440918, "learning_rate": 9.898692586761586e-05, "loss": 1.0776, "step": 1491 }, { "epoch": 0.10109085981435056, "grad_norm": 10.618545532226562, "learning_rate": 9.898555684851805e-05, "loss": 1.1251, "step": 1492 }, { "epoch": 0.10115861508232266, "grad_norm": 10.228861808776855, "learning_rate": 9.898418782942023e-05, "loss": 1.1987, "step": 1493 }, { "epoch": 0.10122637035029473, "grad_norm": 8.807862281799316, "learning_rate": 9.898281881032241e-05, "loss": 0.9339, "step": 1494 }, { "epoch": 0.10129412561826683, "grad_norm": 11.24593448638916, "learning_rate": 9.898144979122459e-05, "loss": 1.1495, "step": 1495 }, { "epoch": 0.1013618808862389, "grad_norm": 11.192438125610352, "learning_rate": 9.898008077212677e-05, "loss": 1.1361, "step": 1496 }, { "epoch": 0.101429636154211, "grad_norm": 10.440075874328613, "learning_rate": 9.897871175302896e-05, "loss": 1.2273, "step": 1497 }, { "epoch": 0.10149739142218307, "grad_norm": 11.103675842285156, "learning_rate": 9.897734273393115e-05, "loss": 1.1308, "step": 1498 }, { "epoch": 0.10156514669015516, "grad_norm": 9.78297233581543, "learning_rate": 9.897597371483333e-05, "loss": 1.0907, "step": 1499 }, { "epoch": 0.10163290195812724, "grad_norm": 10.98086166381836, "learning_rate": 9.89746046957355e-05, "loss": 0.9743, "step": 1500 }, { "epoch": 0.10170065722609933, "grad_norm": 9.268783569335938, "learning_rate": 9.89732356766377e-05, "loss": 0.8917, "step": 1501 }, { "epoch": 0.10176841249407141, "grad_norm": 12.674605369567871, "learning_rate": 9.897186665753988e-05, "loss": 1.2247, "step": 1502 }, { "epoch": 0.1018361677620435, "grad_norm": 10.987565040588379, "learning_rate": 9.897049763844206e-05, "loss": 1.0688, "step": 1503 }, { "epoch": 0.10190392303001558, "grad_norm": 16.014053344726562, "learning_rate": 9.896912861934424e-05, "loss": 0.8334, "step": 1504 }, { "epoch": 0.10197167829798767, "grad_norm": 11.119991302490234, "learning_rate": 9.896775960024642e-05, "loss": 1.3176, "step": 1505 }, { "epoch": 0.10203943356595975, "grad_norm": 13.23279094696045, "learning_rate": 9.896639058114861e-05, "loss": 1.1625, "step": 1506 }, { "epoch": 0.10210718883393184, "grad_norm": 9.3678560256958, "learning_rate": 9.89650215620508e-05, "loss": 1.2774, "step": 1507 }, { "epoch": 0.10217494410190392, "grad_norm": 10.829100608825684, "learning_rate": 9.896365254295298e-05, "loss": 1.2423, "step": 1508 }, { "epoch": 0.10224269936987601, "grad_norm": 12.12694263458252, "learning_rate": 9.896228352385516e-05, "loss": 1.2276, "step": 1509 }, { "epoch": 0.10231045463784809, "grad_norm": 11.626548767089844, "learning_rate": 9.896091450475735e-05, "loss": 1.0871, "step": 1510 }, { "epoch": 0.10237820990582018, "grad_norm": 11.388608932495117, "learning_rate": 9.895954548565953e-05, "loss": 1.5582, "step": 1511 }, { "epoch": 0.10244596517379226, "grad_norm": 9.463730812072754, "learning_rate": 9.895817646656171e-05, "loss": 1.1066, "step": 1512 }, { "epoch": 0.10251372044176435, "grad_norm": 10.291573524475098, "learning_rate": 9.895680744746389e-05, "loss": 1.4026, "step": 1513 }, { "epoch": 0.10258147570973643, "grad_norm": 9.778963088989258, "learning_rate": 9.895543842836607e-05, "loss": 1.1109, "step": 1514 }, { "epoch": 0.10264923097770852, "grad_norm": 9.685966491699219, "learning_rate": 9.895406940926827e-05, "loss": 1.1633, "step": 1515 }, { "epoch": 0.1027169862456806, "grad_norm": 10.76310920715332, "learning_rate": 9.895270039017045e-05, "loss": 1.0813, "step": 1516 }, { "epoch": 0.10278474151365269, "grad_norm": 9.795347213745117, "learning_rate": 9.895133137107263e-05, "loss": 1.2079, "step": 1517 }, { "epoch": 0.10285249678162477, "grad_norm": 9.980990409851074, "learning_rate": 9.89499623519748e-05, "loss": 1.0701, "step": 1518 }, { "epoch": 0.10292025204959686, "grad_norm": 9.682209014892578, "learning_rate": 9.894859333287699e-05, "loss": 1.4133, "step": 1519 }, { "epoch": 0.10298800731756894, "grad_norm": 10.632065773010254, "learning_rate": 9.894722431377918e-05, "loss": 1.0631, "step": 1520 }, { "epoch": 0.10305576258554103, "grad_norm": 10.099474906921387, "learning_rate": 9.894585529468136e-05, "loss": 1.2015, "step": 1521 }, { "epoch": 0.1031235178535131, "grad_norm": 8.289199829101562, "learning_rate": 9.894448627558354e-05, "loss": 1.3009, "step": 1522 }, { "epoch": 0.1031912731214852, "grad_norm": 9.403796195983887, "learning_rate": 9.894311725648573e-05, "loss": 1.113, "step": 1523 }, { "epoch": 0.10325902838945727, "grad_norm": 12.6613130569458, "learning_rate": 9.894174823738792e-05, "loss": 0.9394, "step": 1524 }, { "epoch": 0.10332678365742937, "grad_norm": 9.85255241394043, "learning_rate": 9.894037921829011e-05, "loss": 1.1007, "step": 1525 }, { "epoch": 0.10339453892540144, "grad_norm": 11.918173789978027, "learning_rate": 9.893901019919229e-05, "loss": 1.1817, "step": 1526 }, { "epoch": 0.10346229419337354, "grad_norm": 9.994447708129883, "learning_rate": 9.893764118009447e-05, "loss": 1.2522, "step": 1527 }, { "epoch": 0.10353004946134561, "grad_norm": 9.879289627075195, "learning_rate": 9.893627216099665e-05, "loss": 1.152, "step": 1528 }, { "epoch": 0.1035978047293177, "grad_norm": 10.103482246398926, "learning_rate": 9.893490314189884e-05, "loss": 1.2227, "step": 1529 }, { "epoch": 0.10366555999728978, "grad_norm": 11.173476219177246, "learning_rate": 9.893353412280102e-05, "loss": 1.2188, "step": 1530 }, { "epoch": 0.10373331526526187, "grad_norm": 11.540877342224121, "learning_rate": 9.89321651037032e-05, "loss": 1.1669, "step": 1531 }, { "epoch": 0.10380107053323395, "grad_norm": 10.706154823303223, "learning_rate": 9.893079608460539e-05, "loss": 1.2771, "step": 1532 }, { "epoch": 0.10386882580120604, "grad_norm": 11.781739234924316, "learning_rate": 9.892942706550758e-05, "loss": 1.0501, "step": 1533 }, { "epoch": 0.10393658106917812, "grad_norm": 10.482099533081055, "learning_rate": 9.892805804640976e-05, "loss": 1.2327, "step": 1534 }, { "epoch": 0.10400433633715021, "grad_norm": 12.450867652893066, "learning_rate": 9.892668902731194e-05, "loss": 1.3575, "step": 1535 }, { "epoch": 0.1040720916051223, "grad_norm": 10.236811637878418, "learning_rate": 9.892532000821412e-05, "loss": 1.2559, "step": 1536 }, { "epoch": 0.10413984687309438, "grad_norm": 9.311124801635742, "learning_rate": 9.89239509891163e-05, "loss": 1.1143, "step": 1537 }, { "epoch": 0.10420760214106647, "grad_norm": 9.182706832885742, "learning_rate": 9.89225819700185e-05, "loss": 0.8875, "step": 1538 }, { "epoch": 0.10427535740903855, "grad_norm": 12.762700080871582, "learning_rate": 9.892121295092067e-05, "loss": 1.0552, "step": 1539 }, { "epoch": 0.10434311267701064, "grad_norm": 12.222203254699707, "learning_rate": 9.891984393182285e-05, "loss": 1.3185, "step": 1540 }, { "epoch": 0.10441086794498272, "grad_norm": 11.45807933807373, "learning_rate": 9.891847491272504e-05, "loss": 1.2952, "step": 1541 }, { "epoch": 0.10447862321295481, "grad_norm": 10.445068359375, "learning_rate": 9.891710589362722e-05, "loss": 1.1637, "step": 1542 }, { "epoch": 0.10454637848092689, "grad_norm": 11.758063316345215, "learning_rate": 9.891573687452941e-05, "loss": 1.3202, "step": 1543 }, { "epoch": 0.10461413374889898, "grad_norm": 10.176533699035645, "learning_rate": 9.891436785543159e-05, "loss": 1.1762, "step": 1544 }, { "epoch": 0.10468188901687106, "grad_norm": 10.133155822753906, "learning_rate": 9.891299883633377e-05, "loss": 1.2226, "step": 1545 }, { "epoch": 0.10474964428484315, "grad_norm": 9.883895874023438, "learning_rate": 9.891162981723595e-05, "loss": 0.9687, "step": 1546 }, { "epoch": 0.10481739955281523, "grad_norm": 13.175050735473633, "learning_rate": 9.891026079813814e-05, "loss": 1.2075, "step": 1547 }, { "epoch": 0.10488515482078732, "grad_norm": 11.183597564697266, "learning_rate": 9.890889177904032e-05, "loss": 1.0917, "step": 1548 }, { "epoch": 0.1049529100887594, "grad_norm": 10.135035514831543, "learning_rate": 9.89075227599425e-05, "loss": 1.0087, "step": 1549 }, { "epoch": 0.10502066535673149, "grad_norm": 16.02760887145996, "learning_rate": 9.890615374084469e-05, "loss": 1.2015, "step": 1550 }, { "epoch": 0.10508842062470357, "grad_norm": 11.255363464355469, "learning_rate": 9.890478472174687e-05, "loss": 1.1106, "step": 1551 }, { "epoch": 0.10515617589267566, "grad_norm": 10.740998268127441, "learning_rate": 9.890341570264906e-05, "loss": 1.1835, "step": 1552 }, { "epoch": 0.10522393116064774, "grad_norm": 11.84919548034668, "learning_rate": 9.890204668355124e-05, "loss": 1.1655, "step": 1553 }, { "epoch": 0.10529168642861983, "grad_norm": 12.041108131408691, "learning_rate": 9.890067766445342e-05, "loss": 1.2544, "step": 1554 }, { "epoch": 0.10535944169659191, "grad_norm": 8.966646194458008, "learning_rate": 9.88993086453556e-05, "loss": 1.098, "step": 1555 }, { "epoch": 0.105427196964564, "grad_norm": 11.838338851928711, "learning_rate": 9.88979396262578e-05, "loss": 1.1774, "step": 1556 }, { "epoch": 0.10549495223253608, "grad_norm": 11.186326026916504, "learning_rate": 9.889657060715997e-05, "loss": 1.1778, "step": 1557 }, { "epoch": 0.10556270750050817, "grad_norm": 9.448702812194824, "learning_rate": 9.889520158806216e-05, "loss": 1.0171, "step": 1558 }, { "epoch": 0.10563046276848025, "grad_norm": 10.931096076965332, "learning_rate": 9.889383256896434e-05, "loss": 1.1797, "step": 1559 }, { "epoch": 0.10569821803645234, "grad_norm": 10.293981552124023, "learning_rate": 9.889246354986652e-05, "loss": 1.0723, "step": 1560 }, { "epoch": 0.10576597330442442, "grad_norm": 10.052331924438477, "learning_rate": 9.889109453076871e-05, "loss": 0.831, "step": 1561 }, { "epoch": 0.10583372857239651, "grad_norm": 10.593210220336914, "learning_rate": 9.888972551167089e-05, "loss": 1.3415, "step": 1562 }, { "epoch": 0.10590148384036858, "grad_norm": 11.580954551696777, "learning_rate": 9.888835649257307e-05, "loss": 1.1522, "step": 1563 }, { "epoch": 0.10596923910834068, "grad_norm": 11.495551109313965, "learning_rate": 9.888698747347525e-05, "loss": 1.2515, "step": 1564 }, { "epoch": 0.10603699437631275, "grad_norm": 10.543874740600586, "learning_rate": 9.888561845437743e-05, "loss": 1.1579, "step": 1565 }, { "epoch": 0.10610474964428485, "grad_norm": 10.588164329528809, "learning_rate": 9.888424943527963e-05, "loss": 1.3825, "step": 1566 }, { "epoch": 0.10617250491225692, "grad_norm": 13.591666221618652, "learning_rate": 9.88828804161818e-05, "loss": 1.0941, "step": 1567 }, { "epoch": 0.10624026018022902, "grad_norm": 10.866951942443848, "learning_rate": 9.888151139708399e-05, "loss": 1.478, "step": 1568 }, { "epoch": 0.10630801544820109, "grad_norm": 9.627554893493652, "learning_rate": 9.888014237798618e-05, "loss": 1.004, "step": 1569 }, { "epoch": 0.10637577071617318, "grad_norm": 10.871118545532227, "learning_rate": 9.887877335888836e-05, "loss": 1.1997, "step": 1570 }, { "epoch": 0.10644352598414526, "grad_norm": 10.74503231048584, "learning_rate": 9.887740433979054e-05, "loss": 1.005, "step": 1571 }, { "epoch": 0.10651128125211735, "grad_norm": 11.119452476501465, "learning_rate": 9.887603532069273e-05, "loss": 1.1168, "step": 1572 }, { "epoch": 0.10657903652008943, "grad_norm": 10.601544380187988, "learning_rate": 9.887466630159491e-05, "loss": 1.0256, "step": 1573 }, { "epoch": 0.10664679178806152, "grad_norm": 10.329113006591797, "learning_rate": 9.88732972824971e-05, "loss": 1.3432, "step": 1574 }, { "epoch": 0.1067145470560336, "grad_norm": 9.973999977111816, "learning_rate": 9.887192826339929e-05, "loss": 0.9961, "step": 1575 }, { "epoch": 0.10678230232400569, "grad_norm": 10.981974601745605, "learning_rate": 9.887055924430147e-05, "loss": 1.3936, "step": 1576 }, { "epoch": 0.10685005759197777, "grad_norm": 10.953417778015137, "learning_rate": 9.886919022520365e-05, "loss": 0.9669, "step": 1577 }, { "epoch": 0.10691781285994986, "grad_norm": 10.656323432922363, "learning_rate": 9.886782120610583e-05, "loss": 1.1393, "step": 1578 }, { "epoch": 0.10698556812792194, "grad_norm": 9.434617042541504, "learning_rate": 9.886645218700802e-05, "loss": 1.1813, "step": 1579 }, { "epoch": 0.10705332339589403, "grad_norm": 12.334831237792969, "learning_rate": 9.88650831679102e-05, "loss": 1.453, "step": 1580 }, { "epoch": 0.10712107866386611, "grad_norm": 9.892403602600098, "learning_rate": 9.886371414881238e-05, "loss": 1.1373, "step": 1581 }, { "epoch": 0.1071888339318382, "grad_norm": 9.248678207397461, "learning_rate": 9.886234512971456e-05, "loss": 1.0734, "step": 1582 }, { "epoch": 0.10725658919981028, "grad_norm": 10.317010879516602, "learning_rate": 9.886097611061675e-05, "loss": 1.1252, "step": 1583 }, { "epoch": 0.10732434446778237, "grad_norm": 9.586435317993164, "learning_rate": 9.885960709151894e-05, "loss": 1.1705, "step": 1584 }, { "epoch": 0.10739209973575445, "grad_norm": 8.883166313171387, "learning_rate": 9.885823807242112e-05, "loss": 1.3164, "step": 1585 }, { "epoch": 0.10745985500372654, "grad_norm": 12.870014190673828, "learning_rate": 9.88568690533233e-05, "loss": 1.2187, "step": 1586 }, { "epoch": 0.10752761027169863, "grad_norm": 10.858057975769043, "learning_rate": 9.885550003422548e-05, "loss": 1.2334, "step": 1587 }, { "epoch": 0.10759536553967071, "grad_norm": 9.855050086975098, "learning_rate": 9.885413101512767e-05, "loss": 1.068, "step": 1588 }, { "epoch": 0.1076631208076428, "grad_norm": 12.256099700927734, "learning_rate": 9.885276199602985e-05, "loss": 1.2481, "step": 1589 }, { "epoch": 0.10773087607561488, "grad_norm": 9.756118774414062, "learning_rate": 9.885139297693203e-05, "loss": 1.2365, "step": 1590 }, { "epoch": 0.10779863134358697, "grad_norm": 11.727996826171875, "learning_rate": 9.885002395783421e-05, "loss": 1.3839, "step": 1591 }, { "epoch": 0.10786638661155905, "grad_norm": 11.58463191986084, "learning_rate": 9.88486549387364e-05, "loss": 0.937, "step": 1592 }, { "epoch": 0.10793414187953114, "grad_norm": 12.663015365600586, "learning_rate": 9.884728591963859e-05, "loss": 1.3453, "step": 1593 }, { "epoch": 0.10800189714750322, "grad_norm": 10.950531005859375, "learning_rate": 9.884591690054077e-05, "loss": 1.1345, "step": 1594 }, { "epoch": 0.10806965241547531, "grad_norm": 10.361424446105957, "learning_rate": 9.884454788144295e-05, "loss": 1.2576, "step": 1595 }, { "epoch": 0.10813740768344739, "grad_norm": 11.155044555664062, "learning_rate": 9.884317886234513e-05, "loss": 1.1253, "step": 1596 }, { "epoch": 0.10820516295141948, "grad_norm": 12.457600593566895, "learning_rate": 9.884180984324731e-05, "loss": 1.0959, "step": 1597 }, { "epoch": 0.10827291821939156, "grad_norm": 13.494983673095703, "learning_rate": 9.88404408241495e-05, "loss": 1.1671, "step": 1598 }, { "epoch": 0.10834067348736365, "grad_norm": 11.870648384094238, "learning_rate": 9.883907180505168e-05, "loss": 1.2121, "step": 1599 }, { "epoch": 0.10840842875533573, "grad_norm": 10.327920913696289, "learning_rate": 9.883770278595387e-05, "loss": 1.1319, "step": 1600 }, { "epoch": 0.10847618402330782, "grad_norm": 9.578879356384277, "learning_rate": 9.883633376685605e-05, "loss": 0.9064, "step": 1601 }, { "epoch": 0.1085439392912799, "grad_norm": 9.598779678344727, "learning_rate": 9.883496474775824e-05, "loss": 1.1119, "step": 1602 }, { "epoch": 0.10861169455925199, "grad_norm": 13.522628784179688, "learning_rate": 9.883359572866042e-05, "loss": 1.2549, "step": 1603 }, { "epoch": 0.10867944982722406, "grad_norm": 8.123127937316895, "learning_rate": 9.88322267095626e-05, "loss": 0.8778, "step": 1604 }, { "epoch": 0.10874720509519616, "grad_norm": 13.100455284118652, "learning_rate": 9.883085769046478e-05, "loss": 1.2023, "step": 1605 }, { "epoch": 0.10881496036316823, "grad_norm": 9.897802352905273, "learning_rate": 9.882948867136696e-05, "loss": 1.2094, "step": 1606 }, { "epoch": 0.10888271563114033, "grad_norm": 8.55170726776123, "learning_rate": 9.882811965226915e-05, "loss": 1.0407, "step": 1607 }, { "epoch": 0.1089504708991124, "grad_norm": 10.867438316345215, "learning_rate": 9.882675063317133e-05, "loss": 1.3021, "step": 1608 }, { "epoch": 0.1090182261670845, "grad_norm": 11.062238693237305, "learning_rate": 9.882538161407352e-05, "loss": 1.2627, "step": 1609 }, { "epoch": 0.10908598143505657, "grad_norm": 12.710458755493164, "learning_rate": 9.88240125949757e-05, "loss": 1.6196, "step": 1610 }, { "epoch": 0.10915373670302866, "grad_norm": 10.986543655395508, "learning_rate": 9.882264357587789e-05, "loss": 1.2021, "step": 1611 }, { "epoch": 0.10922149197100074, "grad_norm": 12.792850494384766, "learning_rate": 9.882127455678007e-05, "loss": 1.1246, "step": 1612 }, { "epoch": 0.10928924723897283, "grad_norm": 12.120078086853027, "learning_rate": 9.881990553768225e-05, "loss": 1.2522, "step": 1613 }, { "epoch": 0.10935700250694491, "grad_norm": 10.562193870544434, "learning_rate": 9.881853651858443e-05, "loss": 1.1791, "step": 1614 }, { "epoch": 0.109424757774917, "grad_norm": 13.017698287963867, "learning_rate": 9.881716749948662e-05, "loss": 1.0733, "step": 1615 }, { "epoch": 0.10949251304288908, "grad_norm": 10.904980659484863, "learning_rate": 9.88157984803888e-05, "loss": 1.0166, "step": 1616 }, { "epoch": 0.10956026831086117, "grad_norm": 12.332829475402832, "learning_rate": 9.881442946129099e-05, "loss": 1.0688, "step": 1617 }, { "epoch": 0.10962802357883325, "grad_norm": 12.238106727600098, "learning_rate": 9.881306044219318e-05, "loss": 0.9869, "step": 1618 }, { "epoch": 0.10969577884680534, "grad_norm": 12.75059986114502, "learning_rate": 9.881169142309536e-05, "loss": 1.5365, "step": 1619 }, { "epoch": 0.10976353411477742, "grad_norm": 11.019241333007812, "learning_rate": 9.881032240399754e-05, "loss": 1.2332, "step": 1620 }, { "epoch": 0.10983128938274951, "grad_norm": 11.867498397827148, "learning_rate": 9.880895338489973e-05, "loss": 1.1955, "step": 1621 }, { "epoch": 0.10989904465072159, "grad_norm": 9.685249328613281, "learning_rate": 9.880758436580191e-05, "loss": 1.1972, "step": 1622 }, { "epoch": 0.10996679991869368, "grad_norm": 11.80540657043457, "learning_rate": 9.88062153467041e-05, "loss": 0.9501, "step": 1623 }, { "epoch": 0.11003455518666576, "grad_norm": 9.719080924987793, "learning_rate": 9.880484632760627e-05, "loss": 0.9904, "step": 1624 }, { "epoch": 0.11010231045463785, "grad_norm": 10.906637191772461, "learning_rate": 9.880347730850847e-05, "loss": 1.2807, "step": 1625 }, { "epoch": 0.11017006572260993, "grad_norm": 10.40942668914795, "learning_rate": 9.880210828941065e-05, "loss": 1.2772, "step": 1626 }, { "epoch": 0.11023782099058202, "grad_norm": 10.265239715576172, "learning_rate": 9.880073927031283e-05, "loss": 1.1583, "step": 1627 }, { "epoch": 0.1103055762585541, "grad_norm": 12.158036231994629, "learning_rate": 9.879937025121501e-05, "loss": 1.3907, "step": 1628 }, { "epoch": 0.11037333152652619, "grad_norm": 10.128400802612305, "learning_rate": 9.879800123211719e-05, "loss": 0.8838, "step": 1629 }, { "epoch": 0.11044108679449827, "grad_norm": 9.536952018737793, "learning_rate": 9.879663221301938e-05, "loss": 1.0525, "step": 1630 }, { "epoch": 0.11050884206247036, "grad_norm": 10.826987266540527, "learning_rate": 9.879526319392156e-05, "loss": 1.2005, "step": 1631 }, { "epoch": 0.11057659733044244, "grad_norm": 10.08153247833252, "learning_rate": 9.879389417482374e-05, "loss": 1.1521, "step": 1632 }, { "epoch": 0.11064435259841453, "grad_norm": 10.642118453979492, "learning_rate": 9.879252515572592e-05, "loss": 1.2316, "step": 1633 }, { "epoch": 0.1107121078663866, "grad_norm": 9.595836639404297, "learning_rate": 9.879115613662812e-05, "loss": 1.0969, "step": 1634 }, { "epoch": 0.1107798631343587, "grad_norm": 8.954519271850586, "learning_rate": 9.87897871175303e-05, "loss": 0.9422, "step": 1635 }, { "epoch": 0.11084761840233077, "grad_norm": 9.143898010253906, "learning_rate": 9.878841809843248e-05, "loss": 0.8949, "step": 1636 }, { "epoch": 0.11091537367030287, "grad_norm": 11.656847953796387, "learning_rate": 9.878704907933466e-05, "loss": 0.9739, "step": 1637 }, { "epoch": 0.11098312893827494, "grad_norm": 11.996994018554688, "learning_rate": 9.878568006023684e-05, "loss": 1.3361, "step": 1638 }, { "epoch": 0.11105088420624704, "grad_norm": 11.478599548339844, "learning_rate": 9.878431104113903e-05, "loss": 1.0726, "step": 1639 }, { "epoch": 0.11111863947421913, "grad_norm": 14.594179153442383, "learning_rate": 9.878294202204121e-05, "loss": 1.3725, "step": 1640 }, { "epoch": 0.1111863947421912, "grad_norm": 12.15703296661377, "learning_rate": 9.87815730029434e-05, "loss": 1.1996, "step": 1641 }, { "epoch": 0.1112541500101633, "grad_norm": 9.384215354919434, "learning_rate": 9.878020398384557e-05, "loss": 1.1294, "step": 1642 }, { "epoch": 0.11132190527813537, "grad_norm": 11.131610870361328, "learning_rate": 9.877883496474777e-05, "loss": 1.0526, "step": 1643 }, { "epoch": 0.11138966054610747, "grad_norm": 9.628592491149902, "learning_rate": 9.877746594564995e-05, "loss": 1.0715, "step": 1644 }, { "epoch": 0.11145741581407954, "grad_norm": 10.876824378967285, "learning_rate": 9.877609692655213e-05, "loss": 1.3337, "step": 1645 }, { "epoch": 0.11152517108205164, "grad_norm": 10.489126205444336, "learning_rate": 9.877472790745431e-05, "loss": 1.1727, "step": 1646 }, { "epoch": 0.11159292635002371, "grad_norm": 8.737699508666992, "learning_rate": 9.877335888835649e-05, "loss": 1.1252, "step": 1647 }, { "epoch": 0.1116606816179958, "grad_norm": 9.8843355178833, "learning_rate": 9.877198986925868e-05, "loss": 0.9222, "step": 1648 }, { "epoch": 0.11172843688596788, "grad_norm": 7.930290699005127, "learning_rate": 9.877062085016086e-05, "loss": 1.0095, "step": 1649 }, { "epoch": 0.11179619215393997, "grad_norm": 9.721529006958008, "learning_rate": 9.876925183106304e-05, "loss": 1.0812, "step": 1650 }, { "epoch": 0.11186394742191205, "grad_norm": 10.844324111938477, "learning_rate": 9.876788281196523e-05, "loss": 1.2245, "step": 1651 }, { "epoch": 0.11193170268988414, "grad_norm": 8.750361442565918, "learning_rate": 9.87665137928674e-05, "loss": 0.9983, "step": 1652 }, { "epoch": 0.11199945795785622, "grad_norm": 8.716521263122559, "learning_rate": 9.87651447737696e-05, "loss": 0.8612, "step": 1653 }, { "epoch": 0.11206721322582831, "grad_norm": 10.06696891784668, "learning_rate": 9.876377575467178e-05, "loss": 1.0806, "step": 1654 }, { "epoch": 0.11213496849380039, "grad_norm": 10.526103019714355, "learning_rate": 9.876240673557396e-05, "loss": 1.0799, "step": 1655 }, { "epoch": 0.11220272376177248, "grad_norm": 11.108748435974121, "learning_rate": 9.876103771647614e-05, "loss": 1.3871, "step": 1656 }, { "epoch": 0.11227047902974456, "grad_norm": 10.120033264160156, "learning_rate": 9.875966869737833e-05, "loss": 1.0399, "step": 1657 }, { "epoch": 0.11233823429771665, "grad_norm": 10.589370727539062, "learning_rate": 9.875829967828051e-05, "loss": 1.3249, "step": 1658 }, { "epoch": 0.11240598956568873, "grad_norm": 8.751884460449219, "learning_rate": 9.87569306591827e-05, "loss": 0.9404, "step": 1659 }, { "epoch": 0.11247374483366082, "grad_norm": 9.923823356628418, "learning_rate": 9.875556164008488e-05, "loss": 1.067, "step": 1660 }, { "epoch": 0.1125415001016329, "grad_norm": 9.70035171508789, "learning_rate": 9.875419262098707e-05, "loss": 1.0079, "step": 1661 }, { "epoch": 0.11260925536960499, "grad_norm": 12.337713241577148, "learning_rate": 9.875282360188925e-05, "loss": 1.0962, "step": 1662 }, { "epoch": 0.11267701063757707, "grad_norm": 10.142355918884277, "learning_rate": 9.875145458279143e-05, "loss": 1.1916, "step": 1663 }, { "epoch": 0.11274476590554916, "grad_norm": 11.82714557647705, "learning_rate": 9.875008556369362e-05, "loss": 1.4582, "step": 1664 }, { "epoch": 0.11281252117352124, "grad_norm": 10.524566650390625, "learning_rate": 9.87487165445958e-05, "loss": 1.2156, "step": 1665 }, { "epoch": 0.11288027644149333, "grad_norm": 12.232719421386719, "learning_rate": 9.874734752549798e-05, "loss": 1.2152, "step": 1666 }, { "epoch": 0.11294803170946541, "grad_norm": 10.665773391723633, "learning_rate": 9.874597850640018e-05, "loss": 1.085, "step": 1667 }, { "epoch": 0.1130157869774375, "grad_norm": 10.0956449508667, "learning_rate": 9.874460948730236e-05, "loss": 1.1142, "step": 1668 }, { "epoch": 0.11308354224540958, "grad_norm": 9.27825927734375, "learning_rate": 9.874324046820454e-05, "loss": 1.0263, "step": 1669 }, { "epoch": 0.11315129751338167, "grad_norm": 11.171673774719238, "learning_rate": 9.874187144910672e-05, "loss": 1.5049, "step": 1670 }, { "epoch": 0.11321905278135375, "grad_norm": 10.442279815673828, "learning_rate": 9.874050243000891e-05, "loss": 1.0697, "step": 1671 }, { "epoch": 0.11328680804932584, "grad_norm": 11.563700675964355, "learning_rate": 9.87391334109111e-05, "loss": 1.003, "step": 1672 }, { "epoch": 0.11335456331729792, "grad_norm": 12.447733879089355, "learning_rate": 9.873776439181327e-05, "loss": 1.2279, "step": 1673 }, { "epoch": 0.11342231858527001, "grad_norm": 9.70698070526123, "learning_rate": 9.873639537271545e-05, "loss": 1.2208, "step": 1674 }, { "epoch": 0.11349007385324208, "grad_norm": 10.89858341217041, "learning_rate": 9.873502635361763e-05, "loss": 1.1676, "step": 1675 }, { "epoch": 0.11355782912121418, "grad_norm": 11.106192588806152, "learning_rate": 9.873365733451983e-05, "loss": 1.0918, "step": 1676 }, { "epoch": 0.11362558438918625, "grad_norm": 10.186528205871582, "learning_rate": 9.873228831542201e-05, "loss": 1.1651, "step": 1677 }, { "epoch": 0.11369333965715835, "grad_norm": 10.69977855682373, "learning_rate": 9.873091929632419e-05, "loss": 1.2867, "step": 1678 }, { "epoch": 0.11376109492513042, "grad_norm": 11.934000015258789, "learning_rate": 9.872955027722637e-05, "loss": 1.2051, "step": 1679 }, { "epoch": 0.11382885019310252, "grad_norm": 9.277094841003418, "learning_rate": 9.872818125812856e-05, "loss": 1.17, "step": 1680 }, { "epoch": 0.1138966054610746, "grad_norm": 11.35097885131836, "learning_rate": 9.872681223903074e-05, "loss": 1.2433, "step": 1681 }, { "epoch": 0.11396436072904668, "grad_norm": 11.362833023071289, "learning_rate": 9.872544321993292e-05, "loss": 1.3577, "step": 1682 }, { "epoch": 0.11403211599701876, "grad_norm": 10.944365501403809, "learning_rate": 9.87240742008351e-05, "loss": 1.0731, "step": 1683 }, { "epoch": 0.11409987126499085, "grad_norm": 10.292986869812012, "learning_rate": 9.872270518173728e-05, "loss": 1.0639, "step": 1684 }, { "epoch": 0.11416762653296293, "grad_norm": 10.8629150390625, "learning_rate": 9.872133616263948e-05, "loss": 1.3503, "step": 1685 }, { "epoch": 0.11423538180093502, "grad_norm": 10.277920722961426, "learning_rate": 9.871996714354166e-05, "loss": 1.1544, "step": 1686 }, { "epoch": 0.1143031370689071, "grad_norm": 8.98864459991455, "learning_rate": 9.871859812444384e-05, "loss": 1.3638, "step": 1687 }, { "epoch": 0.11437089233687919, "grad_norm": 8.126718521118164, "learning_rate": 9.871722910534602e-05, "loss": 1.0343, "step": 1688 }, { "epoch": 0.11443864760485127, "grad_norm": 9.229771614074707, "learning_rate": 9.871586008624821e-05, "loss": 1.0137, "step": 1689 }, { "epoch": 0.11450640287282336, "grad_norm": 10.554340362548828, "learning_rate": 9.87144910671504e-05, "loss": 0.9908, "step": 1690 }, { "epoch": 0.11457415814079544, "grad_norm": 11.857934951782227, "learning_rate": 9.871312204805257e-05, "loss": 1.29, "step": 1691 }, { "epoch": 0.11464191340876753, "grad_norm": 9.334272384643555, "learning_rate": 9.871175302895475e-05, "loss": 1.0449, "step": 1692 }, { "epoch": 0.11470966867673962, "grad_norm": 9.092185974121094, "learning_rate": 9.871038400985693e-05, "loss": 1.1386, "step": 1693 }, { "epoch": 0.1147774239447117, "grad_norm": 9.379465103149414, "learning_rate": 9.870901499075913e-05, "loss": 1.0964, "step": 1694 }, { "epoch": 0.11484517921268379, "grad_norm": 10.038492202758789, "learning_rate": 9.870764597166131e-05, "loss": 0.8512, "step": 1695 }, { "epoch": 0.11491293448065587, "grad_norm": 8.2230863571167, "learning_rate": 9.870627695256349e-05, "loss": 1.0646, "step": 1696 }, { "epoch": 0.11498068974862796, "grad_norm": 9.939510345458984, "learning_rate": 9.870490793346567e-05, "loss": 1.081, "step": 1697 }, { "epoch": 0.11504844501660004, "grad_norm": 11.864813804626465, "learning_rate": 9.870353891436786e-05, "loss": 1.2447, "step": 1698 }, { "epoch": 0.11511620028457213, "grad_norm": 11.517714500427246, "learning_rate": 9.870216989527004e-05, "loss": 1.1274, "step": 1699 }, { "epoch": 0.11518395555254421, "grad_norm": 12.172054290771484, "learning_rate": 9.870080087617222e-05, "loss": 1.3278, "step": 1700 }, { "epoch": 0.1152517108205163, "grad_norm": 9.915261268615723, "learning_rate": 9.86994318570744e-05, "loss": 1.0298, "step": 1701 }, { "epoch": 0.11531946608848838, "grad_norm": 10.937396049499512, "learning_rate": 9.869806283797659e-05, "loss": 1.363, "step": 1702 }, { "epoch": 0.11538722135646047, "grad_norm": 11.586840629577637, "learning_rate": 9.869669381887878e-05, "loss": 0.9152, "step": 1703 }, { "epoch": 0.11545497662443255, "grad_norm": 10.035614967346191, "learning_rate": 9.869532479978096e-05, "loss": 1.2152, "step": 1704 }, { "epoch": 0.11552273189240464, "grad_norm": 8.775728225708008, "learning_rate": 9.869395578068314e-05, "loss": 1.0552, "step": 1705 }, { "epoch": 0.11559048716037672, "grad_norm": 10.85958480834961, "learning_rate": 9.869258676158532e-05, "loss": 1.0473, "step": 1706 }, { "epoch": 0.11565824242834881, "grad_norm": 9.311616897583008, "learning_rate": 9.869121774248751e-05, "loss": 1.1704, "step": 1707 }, { "epoch": 0.11572599769632089, "grad_norm": 9.503780364990234, "learning_rate": 9.86898487233897e-05, "loss": 1.1375, "step": 1708 }, { "epoch": 0.11579375296429298, "grad_norm": 9.02602481842041, "learning_rate": 9.868847970429187e-05, "loss": 0.9955, "step": 1709 }, { "epoch": 0.11586150823226506, "grad_norm": 11.473995208740234, "learning_rate": 9.868711068519407e-05, "loss": 0.8842, "step": 1710 }, { "epoch": 0.11592926350023715, "grad_norm": 10.124822616577148, "learning_rate": 9.868574166609625e-05, "loss": 1.0458, "step": 1711 }, { "epoch": 0.11599701876820923, "grad_norm": 11.004744529724121, "learning_rate": 9.868437264699843e-05, "loss": 1.2136, "step": 1712 }, { "epoch": 0.11606477403618132, "grad_norm": 10.497210502624512, "learning_rate": 9.868300362790062e-05, "loss": 1.3457, "step": 1713 }, { "epoch": 0.1161325293041534, "grad_norm": 10.933736801147461, "learning_rate": 9.86816346088028e-05, "loss": 1.0931, "step": 1714 }, { "epoch": 0.11620028457212549, "grad_norm": 11.045526504516602, "learning_rate": 9.868026558970498e-05, "loss": 1.2338, "step": 1715 }, { "epoch": 0.11626803984009756, "grad_norm": 10.763197898864746, "learning_rate": 9.867889657060716e-05, "loss": 1.2325, "step": 1716 }, { "epoch": 0.11633579510806966, "grad_norm": 9.53070068359375, "learning_rate": 9.867752755150936e-05, "loss": 1.1268, "step": 1717 }, { "epoch": 0.11640355037604173, "grad_norm": 10.572071075439453, "learning_rate": 9.867615853241154e-05, "loss": 1.0118, "step": 1718 }, { "epoch": 0.11647130564401383, "grad_norm": 7.633776664733887, "learning_rate": 9.867478951331372e-05, "loss": 1.081, "step": 1719 }, { "epoch": 0.1165390609119859, "grad_norm": 8.915566444396973, "learning_rate": 9.86734204942159e-05, "loss": 1.1526, "step": 1720 }, { "epoch": 0.116606816179958, "grad_norm": 8.296719551086426, "learning_rate": 9.867205147511809e-05, "loss": 1.0571, "step": 1721 }, { "epoch": 0.11667457144793007, "grad_norm": 9.387847900390625, "learning_rate": 9.867068245602027e-05, "loss": 1.0447, "step": 1722 }, { "epoch": 0.11674232671590216, "grad_norm": 10.832168579101562, "learning_rate": 9.866931343692245e-05, "loss": 1.2601, "step": 1723 }, { "epoch": 0.11681008198387424, "grad_norm": 11.631181716918945, "learning_rate": 9.866794441782463e-05, "loss": 1.4507, "step": 1724 }, { "epoch": 0.11687783725184633, "grad_norm": 8.372645378112793, "learning_rate": 9.866657539872681e-05, "loss": 1.1191, "step": 1725 }, { "epoch": 0.11694559251981841, "grad_norm": 9.577934265136719, "learning_rate": 9.866520637962901e-05, "loss": 1.0336, "step": 1726 }, { "epoch": 0.1170133477877905, "grad_norm": 10.626317977905273, "learning_rate": 9.866383736053119e-05, "loss": 1.1595, "step": 1727 }, { "epoch": 0.11708110305576258, "grad_norm": 9.758353233337402, "learning_rate": 9.866246834143337e-05, "loss": 0.8803, "step": 1728 }, { "epoch": 0.11714885832373467, "grad_norm": 9.643457412719727, "learning_rate": 9.866109932233555e-05, "loss": 1.3213, "step": 1729 }, { "epoch": 0.11721661359170675, "grad_norm": 8.910233497619629, "learning_rate": 9.865973030323773e-05, "loss": 1.1123, "step": 1730 }, { "epoch": 0.11728436885967884, "grad_norm": 8.893741607666016, "learning_rate": 9.865836128413992e-05, "loss": 1.1059, "step": 1731 }, { "epoch": 0.11735212412765092, "grad_norm": 12.937616348266602, "learning_rate": 9.86569922650421e-05, "loss": 1.2488, "step": 1732 }, { "epoch": 0.11741987939562301, "grad_norm": 9.696173667907715, "learning_rate": 9.865562324594428e-05, "loss": 1.1076, "step": 1733 }, { "epoch": 0.11748763466359509, "grad_norm": 11.568793296813965, "learning_rate": 9.865425422684646e-05, "loss": 1.1023, "step": 1734 }, { "epoch": 0.11755538993156718, "grad_norm": 11.438271522521973, "learning_rate": 9.865288520774866e-05, "loss": 1.1014, "step": 1735 }, { "epoch": 0.11762314519953926, "grad_norm": 8.217238426208496, "learning_rate": 9.865151618865084e-05, "loss": 0.9566, "step": 1736 }, { "epoch": 0.11769090046751135, "grad_norm": 9.716591835021973, "learning_rate": 9.865014716955302e-05, "loss": 1.0478, "step": 1737 }, { "epoch": 0.11775865573548343, "grad_norm": 11.152593612670898, "learning_rate": 9.86487781504552e-05, "loss": 1.3552, "step": 1738 }, { "epoch": 0.11782641100345552, "grad_norm": 8.824272155761719, "learning_rate": 9.864740913135738e-05, "loss": 1.0463, "step": 1739 }, { "epoch": 0.1178941662714276, "grad_norm": 8.941548347473145, "learning_rate": 9.864604011225957e-05, "loss": 0.9552, "step": 1740 }, { "epoch": 0.11796192153939969, "grad_norm": 10.76177978515625, "learning_rate": 9.864467109316175e-05, "loss": 1.0076, "step": 1741 }, { "epoch": 0.11802967680737177, "grad_norm": 9.87431812286377, "learning_rate": 9.864330207406393e-05, "loss": 0.9905, "step": 1742 }, { "epoch": 0.11809743207534386, "grad_norm": 12.406320571899414, "learning_rate": 9.864193305496611e-05, "loss": 1.2839, "step": 1743 }, { "epoch": 0.11816518734331594, "grad_norm": 12.649428367614746, "learning_rate": 9.864056403586831e-05, "loss": 1.0806, "step": 1744 }, { "epoch": 0.11823294261128803, "grad_norm": 10.888890266418457, "learning_rate": 9.863919501677049e-05, "loss": 0.8805, "step": 1745 }, { "epoch": 0.11830069787926012, "grad_norm": 10.59262752532959, "learning_rate": 9.863782599767267e-05, "loss": 0.9977, "step": 1746 }, { "epoch": 0.1183684531472322, "grad_norm": 11.14206314086914, "learning_rate": 9.863645697857485e-05, "loss": 1.4569, "step": 1747 }, { "epoch": 0.11843620841520429, "grad_norm": 12.148449897766113, "learning_rate": 9.863508795947703e-05, "loss": 1.5222, "step": 1748 }, { "epoch": 0.11850396368317637, "grad_norm": 11.814643859863281, "learning_rate": 9.863371894037922e-05, "loss": 1.3888, "step": 1749 }, { "epoch": 0.11857171895114846, "grad_norm": 12.437151908874512, "learning_rate": 9.86323499212814e-05, "loss": 1.024, "step": 1750 }, { "epoch": 0.11863947421912054, "grad_norm": 9.314751625061035, "learning_rate": 9.863098090218358e-05, "loss": 1.064, "step": 1751 }, { "epoch": 0.11870722948709263, "grad_norm": 12.978782653808594, "learning_rate": 9.862961188308576e-05, "loss": 1.4048, "step": 1752 }, { "epoch": 0.1187749847550647, "grad_norm": 9.825428009033203, "learning_rate": 9.862824286398796e-05, "loss": 0.9631, "step": 1753 }, { "epoch": 0.1188427400230368, "grad_norm": 9.127273559570312, "learning_rate": 9.862687384489014e-05, "loss": 1.0364, "step": 1754 }, { "epoch": 0.11891049529100887, "grad_norm": 10.290020942687988, "learning_rate": 9.862550482579232e-05, "loss": 1.329, "step": 1755 }, { "epoch": 0.11897825055898097, "grad_norm": 9.023946762084961, "learning_rate": 9.862413580669451e-05, "loss": 1.186, "step": 1756 }, { "epoch": 0.11904600582695304, "grad_norm": 11.889911651611328, "learning_rate": 9.86227667875967e-05, "loss": 1.0338, "step": 1757 }, { "epoch": 0.11911376109492514, "grad_norm": 8.938654899597168, "learning_rate": 9.862139776849887e-05, "loss": 1.1933, "step": 1758 }, { "epoch": 0.11918151636289721, "grad_norm": 9.240127563476562, "learning_rate": 9.862002874940107e-05, "loss": 1.2336, "step": 1759 }, { "epoch": 0.1192492716308693, "grad_norm": 10.340953826904297, "learning_rate": 9.861865973030325e-05, "loss": 1.1097, "step": 1760 }, { "epoch": 0.11931702689884138, "grad_norm": 10.807507514953613, "learning_rate": 9.861729071120543e-05, "loss": 0.9127, "step": 1761 }, { "epoch": 0.11938478216681347, "grad_norm": 9.670829772949219, "learning_rate": 9.861592169210761e-05, "loss": 1.2296, "step": 1762 }, { "epoch": 0.11945253743478555, "grad_norm": 10.85981273651123, "learning_rate": 9.86145526730098e-05, "loss": 1.1341, "step": 1763 }, { "epoch": 0.11952029270275764, "grad_norm": 9.198482513427734, "learning_rate": 9.861318365391198e-05, "loss": 1.4397, "step": 1764 }, { "epoch": 0.11958804797072972, "grad_norm": 11.130392074584961, "learning_rate": 9.861181463481416e-05, "loss": 1.381, "step": 1765 }, { "epoch": 0.11965580323870181, "grad_norm": 9.568144798278809, "learning_rate": 9.861044561571634e-05, "loss": 1.0284, "step": 1766 }, { "epoch": 0.11972355850667389, "grad_norm": 9.049298286437988, "learning_rate": 9.860907659661854e-05, "loss": 1.0714, "step": 1767 }, { "epoch": 0.11979131377464598, "grad_norm": 8.948347091674805, "learning_rate": 9.860770757752072e-05, "loss": 1.0248, "step": 1768 }, { "epoch": 0.11985906904261806, "grad_norm": 10.836170196533203, "learning_rate": 9.86063385584229e-05, "loss": 1.0551, "step": 1769 }, { "epoch": 0.11992682431059015, "grad_norm": 11.044917106628418, "learning_rate": 9.860496953932508e-05, "loss": 1.1374, "step": 1770 }, { "epoch": 0.11999457957856223, "grad_norm": 10.975854873657227, "learning_rate": 9.860360052022726e-05, "loss": 1.2018, "step": 1771 }, { "epoch": 0.12006233484653432, "grad_norm": 9.752070426940918, "learning_rate": 9.860223150112945e-05, "loss": 1.0843, "step": 1772 }, { "epoch": 0.1201300901145064, "grad_norm": 9.662758827209473, "learning_rate": 9.860086248203163e-05, "loss": 0.8866, "step": 1773 }, { "epoch": 0.12019784538247849, "grad_norm": 11.268694877624512, "learning_rate": 9.859949346293381e-05, "loss": 1.2653, "step": 1774 }, { "epoch": 0.12026560065045057, "grad_norm": 9.201732635498047, "learning_rate": 9.8598124443836e-05, "loss": 1.1466, "step": 1775 }, { "epoch": 0.12033335591842266, "grad_norm": 12.577048301696777, "learning_rate": 9.859675542473819e-05, "loss": 1.2606, "step": 1776 }, { "epoch": 0.12040111118639474, "grad_norm": 10.051258087158203, "learning_rate": 9.859538640564037e-05, "loss": 1.1533, "step": 1777 }, { "epoch": 0.12046886645436683, "grad_norm": 9.358968734741211, "learning_rate": 9.859401738654255e-05, "loss": 1.0501, "step": 1778 }, { "epoch": 0.12053662172233891, "grad_norm": 11.277347564697266, "learning_rate": 9.859264836744473e-05, "loss": 0.9804, "step": 1779 }, { "epoch": 0.120604376990311, "grad_norm": 12.41299057006836, "learning_rate": 9.859127934834691e-05, "loss": 1.427, "step": 1780 }, { "epoch": 0.12067213225828308, "grad_norm": 10.081144332885742, "learning_rate": 9.85899103292491e-05, "loss": 1.1321, "step": 1781 }, { "epoch": 0.12073988752625517, "grad_norm": 11.469686508178711, "learning_rate": 9.858854131015128e-05, "loss": 1.1972, "step": 1782 }, { "epoch": 0.12080764279422725, "grad_norm": 8.608443260192871, "learning_rate": 9.858717229105346e-05, "loss": 0.9914, "step": 1783 }, { "epoch": 0.12087539806219934, "grad_norm": 11.673405647277832, "learning_rate": 9.858580327195564e-05, "loss": 1.3406, "step": 1784 }, { "epoch": 0.12094315333017142, "grad_norm": 9.533596992492676, "learning_rate": 9.858443425285782e-05, "loss": 1.2119, "step": 1785 }, { "epoch": 0.12101090859814351, "grad_norm": 7.792461395263672, "learning_rate": 9.858306523376002e-05, "loss": 1.167, "step": 1786 }, { "epoch": 0.12107866386611559, "grad_norm": 10.209978103637695, "learning_rate": 9.85816962146622e-05, "loss": 1.164, "step": 1787 }, { "epoch": 0.12114641913408768, "grad_norm": 11.591109275817871, "learning_rate": 9.858032719556438e-05, "loss": 1.2598, "step": 1788 }, { "epoch": 0.12121417440205975, "grad_norm": 10.562797546386719, "learning_rate": 9.857895817646656e-05, "loss": 1.3071, "step": 1789 }, { "epoch": 0.12128192967003185, "grad_norm": 9.419530868530273, "learning_rate": 9.857758915736875e-05, "loss": 0.9951, "step": 1790 }, { "epoch": 0.12134968493800392, "grad_norm": 9.525064468383789, "learning_rate": 9.857622013827093e-05, "loss": 1.1833, "step": 1791 }, { "epoch": 0.12141744020597602, "grad_norm": 9.140012741088867, "learning_rate": 9.857485111917311e-05, "loss": 1.0573, "step": 1792 }, { "epoch": 0.1214851954739481, "grad_norm": 10.88685417175293, "learning_rate": 9.85734821000753e-05, "loss": 1.1669, "step": 1793 }, { "epoch": 0.12155295074192018, "grad_norm": 8.63829231262207, "learning_rate": 9.857211308097747e-05, "loss": 1.1733, "step": 1794 }, { "epoch": 0.12162070600989226, "grad_norm": 9.90237808227539, "learning_rate": 9.857074406187967e-05, "loss": 0.9175, "step": 1795 }, { "epoch": 0.12168846127786435, "grad_norm": 10.068902969360352, "learning_rate": 9.856937504278185e-05, "loss": 0.9422, "step": 1796 }, { "epoch": 0.12175621654583643, "grad_norm": 7.482577800750732, "learning_rate": 9.856800602368403e-05, "loss": 1.0359, "step": 1797 }, { "epoch": 0.12182397181380852, "grad_norm": 10.847206115722656, "learning_rate": 9.856663700458621e-05, "loss": 1.2258, "step": 1798 }, { "epoch": 0.12189172708178062, "grad_norm": 11.122177124023438, "learning_rate": 9.85652679854884e-05, "loss": 1.1378, "step": 1799 }, { "epoch": 0.1219594823497527, "grad_norm": 9.994139671325684, "learning_rate": 9.856389896639058e-05, "loss": 1.2663, "step": 1800 }, { "epoch": 0.12202723761772478, "grad_norm": 10.687590599060059, "learning_rate": 9.856252994729276e-05, "loss": 1.1777, "step": 1801 }, { "epoch": 0.12209499288569686, "grad_norm": 8.403971672058105, "learning_rate": 9.856116092819494e-05, "loss": 0.9812, "step": 1802 }, { "epoch": 0.12216274815366895, "grad_norm": 9.77443790435791, "learning_rate": 9.855979190909714e-05, "loss": 1.0615, "step": 1803 }, { "epoch": 0.12223050342164103, "grad_norm": 10.772642135620117, "learning_rate": 9.855842288999932e-05, "loss": 1.1835, "step": 1804 }, { "epoch": 0.12229825868961312, "grad_norm": 10.509920120239258, "learning_rate": 9.85570538709015e-05, "loss": 1.0361, "step": 1805 }, { "epoch": 0.1223660139575852, "grad_norm": 11.14194393157959, "learning_rate": 9.855568485180369e-05, "loss": 0.9947, "step": 1806 }, { "epoch": 0.12243376922555729, "grad_norm": 11.018975257873535, "learning_rate": 9.855431583270587e-05, "loss": 1.2036, "step": 1807 }, { "epoch": 0.12250152449352937, "grad_norm": 12.099370002746582, "learning_rate": 9.855294681360805e-05, "loss": 1.2977, "step": 1808 }, { "epoch": 0.12256927976150146, "grad_norm": 8.841024398803711, "learning_rate": 9.855157779451025e-05, "loss": 1.2093, "step": 1809 }, { "epoch": 0.12263703502947354, "grad_norm": 9.149311065673828, "learning_rate": 9.855020877541243e-05, "loss": 1.0088, "step": 1810 }, { "epoch": 0.12270479029744563, "grad_norm": 10.53498649597168, "learning_rate": 9.854883975631461e-05, "loss": 0.9527, "step": 1811 }, { "epoch": 0.12277254556541771, "grad_norm": 9.230859756469727, "learning_rate": 9.854747073721679e-05, "loss": 0.9584, "step": 1812 }, { "epoch": 0.1228403008333898, "grad_norm": 11.702610969543457, "learning_rate": 9.854610171811898e-05, "loss": 1.2369, "step": 1813 }, { "epoch": 0.12290805610136188, "grad_norm": 9.912004470825195, "learning_rate": 9.854473269902116e-05, "loss": 1.2571, "step": 1814 }, { "epoch": 0.12297581136933397, "grad_norm": 10.071510314941406, "learning_rate": 9.854336367992334e-05, "loss": 1.1666, "step": 1815 }, { "epoch": 0.12304356663730605, "grad_norm": 10.642035484313965, "learning_rate": 9.854199466082552e-05, "loss": 1.0309, "step": 1816 }, { "epoch": 0.12311132190527814, "grad_norm": 11.135947227478027, "learning_rate": 9.85406256417277e-05, "loss": 1.1276, "step": 1817 }, { "epoch": 0.12317907717325022, "grad_norm": 11.575469017028809, "learning_rate": 9.85392566226299e-05, "loss": 1.1507, "step": 1818 }, { "epoch": 0.12324683244122231, "grad_norm": 10.470771789550781, "learning_rate": 9.853788760353208e-05, "loss": 1.3777, "step": 1819 }, { "epoch": 0.12331458770919439, "grad_norm": 9.77783489227295, "learning_rate": 9.853651858443426e-05, "loss": 1.0492, "step": 1820 }, { "epoch": 0.12338234297716648, "grad_norm": 10.132309913635254, "learning_rate": 9.853514956533644e-05, "loss": 1.1488, "step": 1821 }, { "epoch": 0.12345009824513856, "grad_norm": 11.159482955932617, "learning_rate": 9.853378054623863e-05, "loss": 1.2571, "step": 1822 }, { "epoch": 0.12351785351311065, "grad_norm": 11.599156379699707, "learning_rate": 9.853241152714081e-05, "loss": 1.1734, "step": 1823 }, { "epoch": 0.12358560878108273, "grad_norm": 10.278475761413574, "learning_rate": 9.853104250804299e-05, "loss": 1.0637, "step": 1824 }, { "epoch": 0.12365336404905482, "grad_norm": 12.126015663146973, "learning_rate": 9.852967348894517e-05, "loss": 1.0546, "step": 1825 }, { "epoch": 0.1237211193170269, "grad_norm": 12.094749450683594, "learning_rate": 9.852830446984735e-05, "loss": 1.374, "step": 1826 }, { "epoch": 0.12378887458499899, "grad_norm": 7.607821941375732, "learning_rate": 9.852693545074955e-05, "loss": 0.827, "step": 1827 }, { "epoch": 0.12385662985297106, "grad_norm": 12.086756706237793, "learning_rate": 9.852556643165173e-05, "loss": 1.1219, "step": 1828 }, { "epoch": 0.12392438512094316, "grad_norm": 10.215773582458496, "learning_rate": 9.852419741255391e-05, "loss": 1.0148, "step": 1829 }, { "epoch": 0.12399214038891523, "grad_norm": 10.438709259033203, "learning_rate": 9.852282839345609e-05, "loss": 1.0649, "step": 1830 }, { "epoch": 0.12405989565688733, "grad_norm": 10.564906120300293, "learning_rate": 9.852145937435828e-05, "loss": 1.4346, "step": 1831 }, { "epoch": 0.1241276509248594, "grad_norm": 9.83198356628418, "learning_rate": 9.852009035526046e-05, "loss": 0.847, "step": 1832 }, { "epoch": 0.1241954061928315, "grad_norm": 10.990026473999023, "learning_rate": 9.851872133616264e-05, "loss": 1.2358, "step": 1833 }, { "epoch": 0.12426316146080357, "grad_norm": 10.108382225036621, "learning_rate": 9.851735231706482e-05, "loss": 0.9943, "step": 1834 }, { "epoch": 0.12433091672877566, "grad_norm": 10.23820686340332, "learning_rate": 9.8515983297967e-05, "loss": 1.2464, "step": 1835 }, { "epoch": 0.12439867199674774, "grad_norm": 11.552473068237305, "learning_rate": 9.85146142788692e-05, "loss": 1.2175, "step": 1836 }, { "epoch": 0.12446642726471983, "grad_norm": 8.646978378295898, "learning_rate": 9.851324525977138e-05, "loss": 0.9975, "step": 1837 }, { "epoch": 0.12453418253269191, "grad_norm": 9.344541549682617, "learning_rate": 9.851187624067356e-05, "loss": 0.9463, "step": 1838 }, { "epoch": 0.124601937800664, "grad_norm": 12.645013809204102, "learning_rate": 9.851050722157574e-05, "loss": 1.0396, "step": 1839 }, { "epoch": 0.12466969306863608, "grad_norm": 10.093807220458984, "learning_rate": 9.850913820247792e-05, "loss": 1.2554, "step": 1840 }, { "epoch": 0.12473744833660817, "grad_norm": 8.792567253112793, "learning_rate": 9.850776918338011e-05, "loss": 0.9852, "step": 1841 }, { "epoch": 0.12480520360458025, "grad_norm": 10.448987007141113, "learning_rate": 9.85064001642823e-05, "loss": 1.168, "step": 1842 }, { "epoch": 0.12487295887255234, "grad_norm": 9.953516960144043, "learning_rate": 9.850503114518447e-05, "loss": 1.0116, "step": 1843 }, { "epoch": 0.12494071414052442, "grad_norm": 10.187036514282227, "learning_rate": 9.850366212608665e-05, "loss": 1.203, "step": 1844 }, { "epoch": 0.1250084694084965, "grad_norm": 8.890667915344238, "learning_rate": 9.850229310698885e-05, "loss": 1.0104, "step": 1845 }, { "epoch": 0.1250762246764686, "grad_norm": 11.916625022888184, "learning_rate": 9.850092408789103e-05, "loss": 1.0254, "step": 1846 }, { "epoch": 0.12514397994444068, "grad_norm": 10.934864044189453, "learning_rate": 9.849955506879321e-05, "loss": 0.975, "step": 1847 }, { "epoch": 0.12521173521241277, "grad_norm": 11.214954376220703, "learning_rate": 9.849818604969539e-05, "loss": 1.45, "step": 1848 }, { "epoch": 0.12527949048038484, "grad_norm": 8.801512718200684, "learning_rate": 9.849681703059758e-05, "loss": 1.0899, "step": 1849 }, { "epoch": 0.12534724574835693, "grad_norm": 10.340089797973633, "learning_rate": 9.849544801149976e-05, "loss": 1.1704, "step": 1850 }, { "epoch": 0.12541500101632902, "grad_norm": 9.204201698303223, "learning_rate": 9.849407899240194e-05, "loss": 1.0569, "step": 1851 }, { "epoch": 0.1254827562843011, "grad_norm": 9.43604564666748, "learning_rate": 9.849270997330414e-05, "loss": 1.0721, "step": 1852 }, { "epoch": 0.1255505115522732, "grad_norm": 10.110416412353516, "learning_rate": 9.849134095420632e-05, "loss": 1.1265, "step": 1853 }, { "epoch": 0.12561826682024527, "grad_norm": 10.630755424499512, "learning_rate": 9.84899719351085e-05, "loss": 1.1103, "step": 1854 }, { "epoch": 0.12568602208821736, "grad_norm": 10.888835906982422, "learning_rate": 9.848860291601069e-05, "loss": 1.347, "step": 1855 }, { "epoch": 0.12575377735618945, "grad_norm": 8.84623908996582, "learning_rate": 9.848723389691287e-05, "loss": 0.8531, "step": 1856 }, { "epoch": 0.12582153262416154, "grad_norm": 10.99928092956543, "learning_rate": 9.848586487781505e-05, "loss": 1.1747, "step": 1857 }, { "epoch": 0.1258892878921336, "grad_norm": 8.556151390075684, "learning_rate": 9.848449585871723e-05, "loss": 1.0207, "step": 1858 }, { "epoch": 0.1259570431601057, "grad_norm": 10.207547187805176, "learning_rate": 9.848312683961943e-05, "loss": 1.4782, "step": 1859 }, { "epoch": 0.1260247984280778, "grad_norm": 9.39505386352539, "learning_rate": 9.848175782052161e-05, "loss": 1.1415, "step": 1860 }, { "epoch": 0.12609255369604988, "grad_norm": 10.647768020629883, "learning_rate": 9.848038880142379e-05, "loss": 1.1439, "step": 1861 }, { "epoch": 0.12616030896402194, "grad_norm": 11.517707824707031, "learning_rate": 9.847901978232597e-05, "loss": 1.3181, "step": 1862 }, { "epoch": 0.12622806423199404, "grad_norm": 10.916093826293945, "learning_rate": 9.847765076322815e-05, "loss": 1.0708, "step": 1863 }, { "epoch": 0.12629581949996613, "grad_norm": 9.893363952636719, "learning_rate": 9.847628174413034e-05, "loss": 1.1932, "step": 1864 }, { "epoch": 0.12636357476793822, "grad_norm": 12.349174499511719, "learning_rate": 9.847491272503252e-05, "loss": 1.3161, "step": 1865 }, { "epoch": 0.12643133003591028, "grad_norm": 10.157081604003906, "learning_rate": 9.84735437059347e-05, "loss": 1.1325, "step": 1866 }, { "epoch": 0.12649908530388237, "grad_norm": 9.772073745727539, "learning_rate": 9.847217468683688e-05, "loss": 1.2739, "step": 1867 }, { "epoch": 0.12656684057185447, "grad_norm": 10.468371391296387, "learning_rate": 9.847080566773908e-05, "loss": 1.0279, "step": 1868 }, { "epoch": 0.12663459583982656, "grad_norm": 11.055033683776855, "learning_rate": 9.846943664864126e-05, "loss": 1.0584, "step": 1869 }, { "epoch": 0.12670235110779862, "grad_norm": 10.47987174987793, "learning_rate": 9.846806762954344e-05, "loss": 1.0385, "step": 1870 }, { "epoch": 0.1267701063757707, "grad_norm": 8.933801651000977, "learning_rate": 9.846669861044562e-05, "loss": 1.0259, "step": 1871 }, { "epoch": 0.1268378616437428, "grad_norm": 10.32228946685791, "learning_rate": 9.84653295913478e-05, "loss": 1.0472, "step": 1872 }, { "epoch": 0.1269056169117149, "grad_norm": 9.294051170349121, "learning_rate": 9.846396057224999e-05, "loss": 1.1638, "step": 1873 }, { "epoch": 0.12697337217968696, "grad_norm": 8.32187271118164, "learning_rate": 9.846259155315217e-05, "loss": 0.9865, "step": 1874 }, { "epoch": 0.12704112744765905, "grad_norm": 8.782264709472656, "learning_rate": 9.846122253405435e-05, "loss": 1.0868, "step": 1875 }, { "epoch": 0.12710888271563114, "grad_norm": 10.444697380065918, "learning_rate": 9.845985351495653e-05, "loss": 1.2137, "step": 1876 }, { "epoch": 0.12717663798360324, "grad_norm": 10.543922424316406, "learning_rate": 9.845848449585873e-05, "loss": 1.2337, "step": 1877 }, { "epoch": 0.1272443932515753, "grad_norm": 8.504612922668457, "learning_rate": 9.845711547676091e-05, "loss": 1.0234, "step": 1878 }, { "epoch": 0.1273121485195474, "grad_norm": 8.83178997039795, "learning_rate": 9.845574645766309e-05, "loss": 1.0611, "step": 1879 }, { "epoch": 0.12737990378751948, "grad_norm": 14.37713623046875, "learning_rate": 9.845437743856527e-05, "loss": 1.1704, "step": 1880 }, { "epoch": 0.12744765905549157, "grad_norm": 12.43575668334961, "learning_rate": 9.845300841946745e-05, "loss": 0.848, "step": 1881 }, { "epoch": 0.12751541432346364, "grad_norm": 11.625554084777832, "learning_rate": 9.845163940036964e-05, "loss": 1.2395, "step": 1882 }, { "epoch": 0.12758316959143573, "grad_norm": 7.8962082862854, "learning_rate": 9.845027038127182e-05, "loss": 0.9702, "step": 1883 }, { "epoch": 0.12765092485940782, "grad_norm": 11.689674377441406, "learning_rate": 9.8448901362174e-05, "loss": 1.4885, "step": 1884 }, { "epoch": 0.1277186801273799, "grad_norm": 11.037276268005371, "learning_rate": 9.844753234307618e-05, "loss": 1.1751, "step": 1885 }, { "epoch": 0.12778643539535198, "grad_norm": 11.115680694580078, "learning_rate": 9.844616332397836e-05, "loss": 1.2912, "step": 1886 }, { "epoch": 0.12785419066332407, "grad_norm": 10.553492546081543, "learning_rate": 9.844479430488056e-05, "loss": 1.1001, "step": 1887 }, { "epoch": 0.12792194593129616, "grad_norm": 9.92845630645752, "learning_rate": 9.844342528578274e-05, "loss": 1.1697, "step": 1888 }, { "epoch": 0.12798970119926825, "grad_norm": 9.33232307434082, "learning_rate": 9.844205626668492e-05, "loss": 1.0931, "step": 1889 }, { "epoch": 0.12805745646724032, "grad_norm": 12.35257339477539, "learning_rate": 9.84406872475871e-05, "loss": 0.9819, "step": 1890 }, { "epoch": 0.1281252117352124, "grad_norm": 9.913105010986328, "learning_rate": 9.843931822848929e-05, "loss": 1.2515, "step": 1891 }, { "epoch": 0.1281929670031845, "grad_norm": 11.218729019165039, "learning_rate": 9.843794920939147e-05, "loss": 1.0073, "step": 1892 }, { "epoch": 0.1282607222711566, "grad_norm": 9.97374153137207, "learning_rate": 9.843658019029365e-05, "loss": 1.0527, "step": 1893 }, { "epoch": 0.12832847753912865, "grad_norm": 8.837554931640625, "learning_rate": 9.843521117119583e-05, "loss": 1.0175, "step": 1894 }, { "epoch": 0.12839623280710075, "grad_norm": 9.207158088684082, "learning_rate": 9.843384215209803e-05, "loss": 1.0729, "step": 1895 }, { "epoch": 0.12846398807507284, "grad_norm": 10.960060119628906, "learning_rate": 9.843247313300021e-05, "loss": 1.0485, "step": 1896 }, { "epoch": 0.12853174334304493, "grad_norm": 9.724952697753906, "learning_rate": 9.843110411390239e-05, "loss": 1.2624, "step": 1897 }, { "epoch": 0.128599498611017, "grad_norm": 9.02813720703125, "learning_rate": 9.842973509480458e-05, "loss": 1.011, "step": 1898 }, { "epoch": 0.12866725387898909, "grad_norm": 8.507847785949707, "learning_rate": 9.842836607570676e-05, "loss": 0.9155, "step": 1899 }, { "epoch": 0.12873500914696118, "grad_norm": 10.609807968139648, "learning_rate": 9.842699705660894e-05, "loss": 1.1984, "step": 1900 }, { "epoch": 0.12880276441493327, "grad_norm": 10.144070625305176, "learning_rate": 9.842562803751114e-05, "loss": 1.2331, "step": 1901 }, { "epoch": 0.12887051968290533, "grad_norm": 11.719651222229004, "learning_rate": 9.842425901841332e-05, "loss": 1.2086, "step": 1902 }, { "epoch": 0.12893827495087742, "grad_norm": 10.689997673034668, "learning_rate": 9.84228899993155e-05, "loss": 1.3769, "step": 1903 }, { "epoch": 0.12900603021884952, "grad_norm": 10.7424955368042, "learning_rate": 9.842152098021768e-05, "loss": 1.1844, "step": 1904 }, { "epoch": 0.1290737854868216, "grad_norm": 11.207498550415039, "learning_rate": 9.842015196111987e-05, "loss": 1.2151, "step": 1905 }, { "epoch": 0.1291415407547937, "grad_norm": 10.689212799072266, "learning_rate": 9.841878294202205e-05, "loss": 1.1182, "step": 1906 }, { "epoch": 0.12920929602276576, "grad_norm": 9.708647727966309, "learning_rate": 9.841741392292423e-05, "loss": 1.2874, "step": 1907 }, { "epoch": 0.12927705129073785, "grad_norm": 9.809847831726074, "learning_rate": 9.841604490382641e-05, "loss": 1.1716, "step": 1908 }, { "epoch": 0.12934480655870995, "grad_norm": 8.170798301696777, "learning_rate": 9.84146758847286e-05, "loss": 1.1342, "step": 1909 }, { "epoch": 0.12941256182668204, "grad_norm": 9.12773609161377, "learning_rate": 9.841330686563079e-05, "loss": 1.1506, "step": 1910 }, { "epoch": 0.1294803170946541, "grad_norm": 13.67032241821289, "learning_rate": 9.841193784653297e-05, "loss": 1.2065, "step": 1911 }, { "epoch": 0.1295480723626262, "grad_norm": 9.35611629486084, "learning_rate": 9.841056882743515e-05, "loss": 1.281, "step": 1912 }, { "epoch": 0.12961582763059828, "grad_norm": 9.044548988342285, "learning_rate": 9.840919980833733e-05, "loss": 1.0285, "step": 1913 }, { "epoch": 0.12968358289857038, "grad_norm": 9.955796241760254, "learning_rate": 9.840783078923952e-05, "loss": 1.1742, "step": 1914 }, { "epoch": 0.12975133816654244, "grad_norm": 10.456665992736816, "learning_rate": 9.84064617701417e-05, "loss": 1.3452, "step": 1915 }, { "epoch": 0.12981909343451453, "grad_norm": 10.861869812011719, "learning_rate": 9.840509275104388e-05, "loss": 1.0397, "step": 1916 }, { "epoch": 0.12988684870248662, "grad_norm": 8.546277046203613, "learning_rate": 9.840372373194606e-05, "loss": 0.8762, "step": 1917 }, { "epoch": 0.12995460397045872, "grad_norm": 9.103103637695312, "learning_rate": 9.840235471284824e-05, "loss": 1.0998, "step": 1918 }, { "epoch": 0.13002235923843078, "grad_norm": 9.794631958007812, "learning_rate": 9.840098569375044e-05, "loss": 1.0644, "step": 1919 }, { "epoch": 0.13009011450640287, "grad_norm": 8.702750205993652, "learning_rate": 9.839961667465262e-05, "loss": 1.108, "step": 1920 }, { "epoch": 0.13015786977437496, "grad_norm": 11.108800888061523, "learning_rate": 9.83982476555548e-05, "loss": 1.375, "step": 1921 }, { "epoch": 0.13022562504234705, "grad_norm": 11.253337860107422, "learning_rate": 9.839687863645698e-05, "loss": 1.0246, "step": 1922 }, { "epoch": 0.13029338031031912, "grad_norm": 9.817541122436523, "learning_rate": 9.839550961735917e-05, "loss": 1.0222, "step": 1923 }, { "epoch": 0.1303611355782912, "grad_norm": 9.378199577331543, "learning_rate": 9.839414059826135e-05, "loss": 1.0387, "step": 1924 }, { "epoch": 0.1304288908462633, "grad_norm": 8.789336204528809, "learning_rate": 9.839277157916353e-05, "loss": 1.0635, "step": 1925 }, { "epoch": 0.1304966461142354, "grad_norm": 11.900403022766113, "learning_rate": 9.839140256006571e-05, "loss": 1.1224, "step": 1926 }, { "epoch": 0.13056440138220746, "grad_norm": 10.007912635803223, "learning_rate": 9.83900335409679e-05, "loss": 0.8988, "step": 1927 }, { "epoch": 0.13063215665017955, "grad_norm": 10.157328605651855, "learning_rate": 9.838866452187009e-05, "loss": 0.952, "step": 1928 }, { "epoch": 0.13069991191815164, "grad_norm": 8.763729095458984, "learning_rate": 9.838729550277227e-05, "loss": 0.9304, "step": 1929 }, { "epoch": 0.13076766718612373, "grad_norm": 10.75432300567627, "learning_rate": 9.838592648367445e-05, "loss": 1.0438, "step": 1930 }, { "epoch": 0.1308354224540958, "grad_norm": 8.840702056884766, "learning_rate": 9.838455746457663e-05, "loss": 0.9164, "step": 1931 }, { "epoch": 0.1309031777220679, "grad_norm": 9.526811599731445, "learning_rate": 9.838318844547882e-05, "loss": 1.2181, "step": 1932 }, { "epoch": 0.13097093299003998, "grad_norm": 12.827199935913086, "learning_rate": 9.8381819426381e-05, "loss": 1.223, "step": 1933 }, { "epoch": 0.13103868825801207, "grad_norm": 9.47105884552002, "learning_rate": 9.838045040728318e-05, "loss": 1.1014, "step": 1934 }, { "epoch": 0.13110644352598413, "grad_norm": 9.044878959655762, "learning_rate": 9.837908138818536e-05, "loss": 1.1451, "step": 1935 }, { "epoch": 0.13117419879395623, "grad_norm": 9.24599838256836, "learning_rate": 9.837771236908754e-05, "loss": 1.1144, "step": 1936 }, { "epoch": 0.13124195406192832, "grad_norm": 12.543252944946289, "learning_rate": 9.837634334998974e-05, "loss": 1.1233, "step": 1937 }, { "epoch": 0.1313097093299004, "grad_norm": 10.022245407104492, "learning_rate": 9.837497433089192e-05, "loss": 1.3569, "step": 1938 }, { "epoch": 0.13137746459787247, "grad_norm": 11.967863082885742, "learning_rate": 9.83736053117941e-05, "loss": 1.2086, "step": 1939 }, { "epoch": 0.13144521986584456, "grad_norm": 9.64406967163086, "learning_rate": 9.837223629269628e-05, "loss": 1.1647, "step": 1940 }, { "epoch": 0.13151297513381666, "grad_norm": 8.538762092590332, "learning_rate": 9.837086727359847e-05, "loss": 1.012, "step": 1941 }, { "epoch": 0.13158073040178875, "grad_norm": 10.889129638671875, "learning_rate": 9.836949825450065e-05, "loss": 1.1559, "step": 1942 }, { "epoch": 0.1316484856697608, "grad_norm": 8.407093048095703, "learning_rate": 9.836812923540283e-05, "loss": 0.9624, "step": 1943 }, { "epoch": 0.1317162409377329, "grad_norm": 9.175569534301758, "learning_rate": 9.836676021630503e-05, "loss": 0.9814, "step": 1944 }, { "epoch": 0.131783996205705, "grad_norm": 11.240396499633789, "learning_rate": 9.836539119720721e-05, "loss": 1.3059, "step": 1945 }, { "epoch": 0.1318517514736771, "grad_norm": 8.791098594665527, "learning_rate": 9.836402217810939e-05, "loss": 0.9891, "step": 1946 }, { "epoch": 0.13191950674164915, "grad_norm": 7.401971340179443, "learning_rate": 9.836265315901158e-05, "loss": 0.9706, "step": 1947 }, { "epoch": 0.13198726200962124, "grad_norm": 13.127768516540527, "learning_rate": 9.836128413991376e-05, "loss": 0.8924, "step": 1948 }, { "epoch": 0.13205501727759333, "grad_norm": 9.805618286132812, "learning_rate": 9.835991512081594e-05, "loss": 0.9969, "step": 1949 }, { "epoch": 0.13212277254556543, "grad_norm": 10.500420570373535, "learning_rate": 9.835854610171812e-05, "loss": 1.1045, "step": 1950 }, { "epoch": 0.1321905278135375, "grad_norm": 10.043769836425781, "learning_rate": 9.835717708262032e-05, "loss": 1.2475, "step": 1951 }, { "epoch": 0.13225828308150958, "grad_norm": 10.6277437210083, "learning_rate": 9.83558080635225e-05, "loss": 0.9285, "step": 1952 }, { "epoch": 0.13232603834948167, "grad_norm": 9.011415481567383, "learning_rate": 9.835443904442468e-05, "loss": 1.1895, "step": 1953 }, { "epoch": 0.13239379361745376, "grad_norm": 11.031267166137695, "learning_rate": 9.835307002532686e-05, "loss": 1.1781, "step": 1954 }, { "epoch": 0.13246154888542583, "grad_norm": 9.960331916809082, "learning_rate": 9.835170100622905e-05, "loss": 1.0972, "step": 1955 }, { "epoch": 0.13252930415339792, "grad_norm": 9.549619674682617, "learning_rate": 9.835033198713123e-05, "loss": 1.0918, "step": 1956 }, { "epoch": 0.13259705942137, "grad_norm": 9.780478477478027, "learning_rate": 9.834896296803341e-05, "loss": 1.1491, "step": 1957 }, { "epoch": 0.1326648146893421, "grad_norm": 8.948554992675781, "learning_rate": 9.834759394893559e-05, "loss": 1.0969, "step": 1958 }, { "epoch": 0.1327325699573142, "grad_norm": 9.165532112121582, "learning_rate": 9.834622492983777e-05, "loss": 0.9294, "step": 1959 }, { "epoch": 0.13280032522528626, "grad_norm": 8.738619804382324, "learning_rate": 9.834485591073997e-05, "loss": 1.3074, "step": 1960 }, { "epoch": 0.13286808049325835, "grad_norm": 9.442314147949219, "learning_rate": 9.834348689164215e-05, "loss": 1.1238, "step": 1961 }, { "epoch": 0.13293583576123044, "grad_norm": 12.411934852600098, "learning_rate": 9.834211787254433e-05, "loss": 1.1405, "step": 1962 }, { "epoch": 0.13300359102920253, "grad_norm": 9.911120414733887, "learning_rate": 9.834074885344651e-05, "loss": 1.2197, "step": 1963 }, { "epoch": 0.1330713462971746, "grad_norm": 9.619095802307129, "learning_rate": 9.83393798343487e-05, "loss": 1.0526, "step": 1964 }, { "epoch": 0.1331391015651467, "grad_norm": 10.163374900817871, "learning_rate": 9.833801081525088e-05, "loss": 0.8374, "step": 1965 }, { "epoch": 0.13320685683311878, "grad_norm": 9.342517852783203, "learning_rate": 9.833664179615306e-05, "loss": 1.1908, "step": 1966 }, { "epoch": 0.13327461210109087, "grad_norm": 10.419418334960938, "learning_rate": 9.833527277705524e-05, "loss": 1.1175, "step": 1967 }, { "epoch": 0.13334236736906294, "grad_norm": 9.5196533203125, "learning_rate": 9.833390375795742e-05, "loss": 1.2347, "step": 1968 }, { "epoch": 0.13341012263703503, "grad_norm": 9.242755889892578, "learning_rate": 9.833253473885962e-05, "loss": 1.2401, "step": 1969 }, { "epoch": 0.13347787790500712, "grad_norm": 10.243762969970703, "learning_rate": 9.83311657197618e-05, "loss": 1.3492, "step": 1970 }, { "epoch": 0.1335456331729792, "grad_norm": 9.169745445251465, "learning_rate": 9.832979670066398e-05, "loss": 0.9533, "step": 1971 }, { "epoch": 0.13361338844095128, "grad_norm": 10.292695999145508, "learning_rate": 9.832842768156616e-05, "loss": 1.2216, "step": 1972 }, { "epoch": 0.13368114370892337, "grad_norm": 9.25019645690918, "learning_rate": 9.832705866246834e-05, "loss": 1.1613, "step": 1973 }, { "epoch": 0.13374889897689546, "grad_norm": 8.518020629882812, "learning_rate": 9.832568964337053e-05, "loss": 1.0767, "step": 1974 }, { "epoch": 0.13381665424486755, "grad_norm": 11.824376106262207, "learning_rate": 9.832432062427271e-05, "loss": 1.1884, "step": 1975 }, { "epoch": 0.13388440951283961, "grad_norm": 10.544015884399414, "learning_rate": 9.832295160517489e-05, "loss": 0.9675, "step": 1976 }, { "epoch": 0.1339521647808117, "grad_norm": 9.495721817016602, "learning_rate": 9.832158258607707e-05, "loss": 1.0586, "step": 1977 }, { "epoch": 0.1340199200487838, "grad_norm": 10.378434181213379, "learning_rate": 9.832021356697927e-05, "loss": 1.176, "step": 1978 }, { "epoch": 0.1340876753167559, "grad_norm": 10.026887893676758, "learning_rate": 9.831884454788145e-05, "loss": 1.035, "step": 1979 }, { "epoch": 0.13415543058472795, "grad_norm": 8.878249168395996, "learning_rate": 9.831747552878363e-05, "loss": 0.8441, "step": 1980 }, { "epoch": 0.13422318585270004, "grad_norm": 9.637616157531738, "learning_rate": 9.831610650968581e-05, "loss": 1.1435, "step": 1981 }, { "epoch": 0.13429094112067214, "grad_norm": 10.153711318969727, "learning_rate": 9.831473749058799e-05, "loss": 1.0833, "step": 1982 }, { "epoch": 0.13435869638864423, "grad_norm": 8.48596477508545, "learning_rate": 9.831336847149018e-05, "loss": 0.7752, "step": 1983 }, { "epoch": 0.1344264516566163, "grad_norm": 8.454994201660156, "learning_rate": 9.831199945239236e-05, "loss": 0.8657, "step": 1984 }, { "epoch": 0.13449420692458838, "grad_norm": 10.607659339904785, "learning_rate": 9.831063043329454e-05, "loss": 1.1836, "step": 1985 }, { "epoch": 0.13456196219256047, "grad_norm": 8.560189247131348, "learning_rate": 9.830926141419672e-05, "loss": 0.7877, "step": 1986 }, { "epoch": 0.13462971746053257, "grad_norm": 10.687662124633789, "learning_rate": 9.830789239509892e-05, "loss": 1.1586, "step": 1987 }, { "epoch": 0.13469747272850463, "grad_norm": 9.725050926208496, "learning_rate": 9.83065233760011e-05, "loss": 1.214, "step": 1988 }, { "epoch": 0.13476522799647672, "grad_norm": 9.808280944824219, "learning_rate": 9.830515435690328e-05, "loss": 1.0532, "step": 1989 }, { "epoch": 0.1348329832644488, "grad_norm": 8.838544845581055, "learning_rate": 9.830378533780547e-05, "loss": 1.0872, "step": 1990 }, { "epoch": 0.1349007385324209, "grad_norm": 12.043976783752441, "learning_rate": 9.830241631870765e-05, "loss": 1.2248, "step": 1991 }, { "epoch": 0.13496849380039297, "grad_norm": 10.044602394104004, "learning_rate": 9.830104729960983e-05, "loss": 0.9746, "step": 1992 }, { "epoch": 0.13503624906836506, "grad_norm": 10.861641883850098, "learning_rate": 9.829967828051203e-05, "loss": 0.9868, "step": 1993 }, { "epoch": 0.13510400433633715, "grad_norm": 9.005234718322754, "learning_rate": 9.82983092614142e-05, "loss": 1.0619, "step": 1994 }, { "epoch": 0.13517175960430924, "grad_norm": 9.902874946594238, "learning_rate": 9.829694024231639e-05, "loss": 1.3176, "step": 1995 }, { "epoch": 0.1352395148722813, "grad_norm": 12.133747100830078, "learning_rate": 9.829557122321857e-05, "loss": 1.3634, "step": 1996 }, { "epoch": 0.1353072701402534, "grad_norm": 9.571062088012695, "learning_rate": 9.829420220412076e-05, "loss": 1.139, "step": 1997 }, { "epoch": 0.1353750254082255, "grad_norm": 8.269545555114746, "learning_rate": 9.829283318502294e-05, "loss": 1.1422, "step": 1998 }, { "epoch": 0.13544278067619758, "grad_norm": 9.963309288024902, "learning_rate": 9.829146416592512e-05, "loss": 1.0612, "step": 1999 }, { "epoch": 0.13551053594416965, "grad_norm": 7.434775352478027, "learning_rate": 9.82900951468273e-05, "loss": 1.0418, "step": 2000 }, { "epoch": 0.13557829121214174, "grad_norm": 8.89494800567627, "learning_rate": 9.82887261277295e-05, "loss": 1.0677, "step": 2001 }, { "epoch": 0.13564604648011383, "grad_norm": 9.349754333496094, "learning_rate": 9.828735710863168e-05, "loss": 0.8932, "step": 2002 }, { "epoch": 0.13571380174808592, "grad_norm": 7.574460506439209, "learning_rate": 9.828598808953386e-05, "loss": 0.7652, "step": 2003 }, { "epoch": 0.13578155701605799, "grad_norm": 10.191296577453613, "learning_rate": 9.828461907043604e-05, "loss": 0.9881, "step": 2004 }, { "epoch": 0.13584931228403008, "grad_norm": 11.235671043395996, "learning_rate": 9.828325005133822e-05, "loss": 1.1942, "step": 2005 }, { "epoch": 0.13591706755200217, "grad_norm": 9.97700023651123, "learning_rate": 9.828188103224041e-05, "loss": 1.154, "step": 2006 }, { "epoch": 0.13598482281997426, "grad_norm": 8.283380508422852, "learning_rate": 9.828051201314259e-05, "loss": 0.8207, "step": 2007 }, { "epoch": 0.13605257808794632, "grad_norm": 9.147144317626953, "learning_rate": 9.827914299404477e-05, "loss": 1.0751, "step": 2008 }, { "epoch": 0.13612033335591842, "grad_norm": 9.99758529663086, "learning_rate": 9.827777397494695e-05, "loss": 1.1166, "step": 2009 }, { "epoch": 0.1361880886238905, "grad_norm": 8.340705871582031, "learning_rate": 9.827640495584915e-05, "loss": 1.0311, "step": 2010 }, { "epoch": 0.1362558438918626, "grad_norm": 10.536659240722656, "learning_rate": 9.827503593675133e-05, "loss": 1.1704, "step": 2011 }, { "epoch": 0.1363235991598347, "grad_norm": 9.01259994506836, "learning_rate": 9.827366691765351e-05, "loss": 1.0531, "step": 2012 }, { "epoch": 0.13639135442780675, "grad_norm": 10.476259231567383, "learning_rate": 9.827229789855569e-05, "loss": 1.053, "step": 2013 }, { "epoch": 0.13645910969577885, "grad_norm": 8.496943473815918, "learning_rate": 9.827092887945787e-05, "loss": 0.836, "step": 2014 }, { "epoch": 0.13652686496375094, "grad_norm": 11.638651847839355, "learning_rate": 9.826955986036006e-05, "loss": 1.0886, "step": 2015 }, { "epoch": 0.13659462023172303, "grad_norm": 7.483241558074951, "learning_rate": 9.826819084126224e-05, "loss": 1.0406, "step": 2016 }, { "epoch": 0.1366623754996951, "grad_norm": 8.648175239562988, "learning_rate": 9.826682182216442e-05, "loss": 1.0442, "step": 2017 }, { "epoch": 0.13673013076766719, "grad_norm": 8.593596458435059, "learning_rate": 9.82654528030666e-05, "loss": 1.1184, "step": 2018 }, { "epoch": 0.13679788603563928, "grad_norm": 8.563817977905273, "learning_rate": 9.826408378396878e-05, "loss": 0.8645, "step": 2019 }, { "epoch": 0.13686564130361137, "grad_norm": 7.812311172485352, "learning_rate": 9.826271476487098e-05, "loss": 1.0605, "step": 2020 }, { "epoch": 0.13693339657158343, "grad_norm": 10.97168254852295, "learning_rate": 9.826134574577316e-05, "loss": 1.1474, "step": 2021 }, { "epoch": 0.13700115183955552, "grad_norm": 7.749011993408203, "learning_rate": 9.825997672667534e-05, "loss": 0.9836, "step": 2022 }, { "epoch": 0.13706890710752762, "grad_norm": 11.088539123535156, "learning_rate": 9.825860770757752e-05, "loss": 1.1456, "step": 2023 }, { "epoch": 0.1371366623754997, "grad_norm": 10.960288047790527, "learning_rate": 9.825723868847971e-05, "loss": 1.4315, "step": 2024 }, { "epoch": 0.13720441764347177, "grad_norm": 10.804458618164062, "learning_rate": 9.825586966938189e-05, "loss": 1.0388, "step": 2025 }, { "epoch": 0.13727217291144386, "grad_norm": 7.906947612762451, "learning_rate": 9.825450065028407e-05, "loss": 1.1306, "step": 2026 }, { "epoch": 0.13733992817941595, "grad_norm": 10.404504776000977, "learning_rate": 9.825313163118625e-05, "loss": 1.0907, "step": 2027 }, { "epoch": 0.13740768344738805, "grad_norm": 9.684488296508789, "learning_rate": 9.825176261208843e-05, "loss": 1.1007, "step": 2028 }, { "epoch": 0.1374754387153601, "grad_norm": 8.345703125, "learning_rate": 9.825039359299063e-05, "loss": 0.9342, "step": 2029 }, { "epoch": 0.1375431939833322, "grad_norm": 8.513103485107422, "learning_rate": 9.824902457389281e-05, "loss": 1.1139, "step": 2030 }, { "epoch": 0.1376109492513043, "grad_norm": 8.675403594970703, "learning_rate": 9.824765555479499e-05, "loss": 0.8499, "step": 2031 }, { "epoch": 0.13767870451927638, "grad_norm": 10.550504684448242, "learning_rate": 9.824628653569717e-05, "loss": 1.2563, "step": 2032 }, { "epoch": 0.13774645978724845, "grad_norm": 7.775172710418701, "learning_rate": 9.824491751659936e-05, "loss": 1.0394, "step": 2033 }, { "epoch": 0.13781421505522054, "grad_norm": 10.057134628295898, "learning_rate": 9.824354849750154e-05, "loss": 1.337, "step": 2034 }, { "epoch": 0.13788197032319263, "grad_norm": 10.658480644226074, "learning_rate": 9.824217947840372e-05, "loss": 1.1675, "step": 2035 }, { "epoch": 0.13794972559116472, "grad_norm": 10.499738693237305, "learning_rate": 9.82408104593059e-05, "loss": 1.1397, "step": 2036 }, { "epoch": 0.1380174808591368, "grad_norm": 12.853378295898438, "learning_rate": 9.82394414402081e-05, "loss": 1.2373, "step": 2037 }, { "epoch": 0.13808523612710888, "grad_norm": 8.66174602508545, "learning_rate": 9.823807242111028e-05, "loss": 0.9863, "step": 2038 }, { "epoch": 0.13815299139508097, "grad_norm": 10.246259689331055, "learning_rate": 9.823670340201246e-05, "loss": 1.0892, "step": 2039 }, { "epoch": 0.13822074666305306, "grad_norm": 11.880420684814453, "learning_rate": 9.823533438291465e-05, "loss": 1.069, "step": 2040 }, { "epoch": 0.13828850193102513, "grad_norm": 9.475419998168945, "learning_rate": 9.823396536381683e-05, "loss": 0.9276, "step": 2041 }, { "epoch": 0.13835625719899722, "grad_norm": 9.827219009399414, "learning_rate": 9.823259634471903e-05, "loss": 1.2536, "step": 2042 }, { "epoch": 0.1384240124669693, "grad_norm": 8.558990478515625, "learning_rate": 9.82312273256212e-05, "loss": 1.2779, "step": 2043 }, { "epoch": 0.1384917677349414, "grad_norm": 12.038803100585938, "learning_rate": 9.822985830652339e-05, "loss": 0.9019, "step": 2044 }, { "epoch": 0.13855952300291347, "grad_norm": 10.764846801757812, "learning_rate": 9.822848928742557e-05, "loss": 1.1803, "step": 2045 }, { "epoch": 0.13862727827088556, "grad_norm": 10.788616180419922, "learning_rate": 9.822712026832775e-05, "loss": 1.3469, "step": 2046 }, { "epoch": 0.13869503353885765, "grad_norm": 8.718049049377441, "learning_rate": 9.822575124922994e-05, "loss": 0.9135, "step": 2047 }, { "epoch": 0.13876278880682974, "grad_norm": 12.245726585388184, "learning_rate": 9.822438223013212e-05, "loss": 1.3114, "step": 2048 }, { "epoch": 0.1388305440748018, "grad_norm": 10.55252456665039, "learning_rate": 9.82230132110343e-05, "loss": 1.1303, "step": 2049 }, { "epoch": 0.1388982993427739, "grad_norm": 8.177289962768555, "learning_rate": 9.822164419193648e-05, "loss": 0.9796, "step": 2050 }, { "epoch": 0.138966054610746, "grad_norm": 10.708680152893066, "learning_rate": 9.822027517283866e-05, "loss": 1.1546, "step": 2051 }, { "epoch": 0.13903380987871808, "grad_norm": 8.631631851196289, "learning_rate": 9.821890615374086e-05, "loss": 1.0055, "step": 2052 }, { "epoch": 0.13910156514669014, "grad_norm": 8.778770446777344, "learning_rate": 9.821753713464304e-05, "loss": 0.9626, "step": 2053 }, { "epoch": 0.13916932041466223, "grad_norm": 10.004143714904785, "learning_rate": 9.821616811554522e-05, "loss": 1.1411, "step": 2054 }, { "epoch": 0.13923707568263433, "grad_norm": 9.47324275970459, "learning_rate": 9.82147990964474e-05, "loss": 1.0081, "step": 2055 }, { "epoch": 0.13930483095060642, "grad_norm": 10.014934539794922, "learning_rate": 9.821343007734959e-05, "loss": 1.3015, "step": 2056 }, { "epoch": 0.13937258621857848, "grad_norm": 10.589959144592285, "learning_rate": 9.821206105825177e-05, "loss": 1.128, "step": 2057 }, { "epoch": 0.13944034148655057, "grad_norm": 9.273834228515625, "learning_rate": 9.821069203915395e-05, "loss": 0.7743, "step": 2058 }, { "epoch": 0.13950809675452266, "grad_norm": 10.72019100189209, "learning_rate": 9.820932302005613e-05, "loss": 0.9524, "step": 2059 }, { "epoch": 0.13957585202249476, "grad_norm": 11.212404251098633, "learning_rate": 9.820795400095831e-05, "loss": 1.2044, "step": 2060 }, { "epoch": 0.13964360729046682, "grad_norm": 9.800296783447266, "learning_rate": 9.82065849818605e-05, "loss": 1.0819, "step": 2061 }, { "epoch": 0.1397113625584389, "grad_norm": 8.668676376342773, "learning_rate": 9.820521596276269e-05, "loss": 1.0778, "step": 2062 }, { "epoch": 0.139779117826411, "grad_norm": 10.854613304138184, "learning_rate": 9.820384694366487e-05, "loss": 1.1181, "step": 2063 }, { "epoch": 0.1398468730943831, "grad_norm": 12.019871711730957, "learning_rate": 9.820247792456705e-05, "loss": 1.2507, "step": 2064 }, { "epoch": 0.1399146283623552, "grad_norm": 8.835234642028809, "learning_rate": 9.820110890546924e-05, "loss": 1.1338, "step": 2065 }, { "epoch": 0.13998238363032725, "grad_norm": 8.57636547088623, "learning_rate": 9.819973988637142e-05, "loss": 0.957, "step": 2066 }, { "epoch": 0.14005013889829934, "grad_norm": 9.080939292907715, "learning_rate": 9.81983708672736e-05, "loss": 1.0035, "step": 2067 }, { "epoch": 0.14011789416627143, "grad_norm": 9.616862297058105, "learning_rate": 9.819700184817578e-05, "loss": 0.9045, "step": 2068 }, { "epoch": 0.14018564943424353, "grad_norm": 10.04652214050293, "learning_rate": 9.819563282907796e-05, "loss": 1.1202, "step": 2069 }, { "epoch": 0.1402534047022156, "grad_norm": 9.132543563842773, "learning_rate": 9.819426380998016e-05, "loss": 0.9918, "step": 2070 }, { "epoch": 0.14032115997018768, "grad_norm": 12.023603439331055, "learning_rate": 9.819289479088234e-05, "loss": 1.1271, "step": 2071 }, { "epoch": 0.14038891523815977, "grad_norm": 9.167064666748047, "learning_rate": 9.819152577178452e-05, "loss": 0.9136, "step": 2072 }, { "epoch": 0.14045667050613186, "grad_norm": 11.954336166381836, "learning_rate": 9.81901567526867e-05, "loss": 1.0073, "step": 2073 }, { "epoch": 0.14052442577410393, "grad_norm": 9.166129112243652, "learning_rate": 9.818878773358888e-05, "loss": 1.1101, "step": 2074 }, { "epoch": 0.14059218104207602, "grad_norm": 8.53992748260498, "learning_rate": 9.818741871449107e-05, "loss": 0.9556, "step": 2075 }, { "epoch": 0.1406599363100481, "grad_norm": 10.769463539123535, "learning_rate": 9.818604969539325e-05, "loss": 1.3009, "step": 2076 }, { "epoch": 0.1407276915780202, "grad_norm": 10.938043594360352, "learning_rate": 9.818468067629543e-05, "loss": 1.4299, "step": 2077 }, { "epoch": 0.14079544684599227, "grad_norm": 10.8585844039917, "learning_rate": 9.818331165719761e-05, "loss": 1.0804, "step": 2078 }, { "epoch": 0.14086320211396436, "grad_norm": 8.763557434082031, "learning_rate": 9.81819426380998e-05, "loss": 0.8321, "step": 2079 }, { "epoch": 0.14093095738193645, "grad_norm": 11.917708396911621, "learning_rate": 9.818057361900199e-05, "loss": 1.4293, "step": 2080 }, { "epoch": 0.14099871264990854, "grad_norm": 10.189987182617188, "learning_rate": 9.817920459990417e-05, "loss": 1.139, "step": 2081 }, { "epoch": 0.1410664679178806, "grad_norm": 11.09156608581543, "learning_rate": 9.817783558080635e-05, "loss": 0.9866, "step": 2082 }, { "epoch": 0.1411342231858527, "grad_norm": 9.394566535949707, "learning_rate": 9.817646656170854e-05, "loss": 0.9401, "step": 2083 }, { "epoch": 0.1412019784538248, "grad_norm": 8.232394218444824, "learning_rate": 9.817509754261072e-05, "loss": 0.9193, "step": 2084 }, { "epoch": 0.14126973372179688, "grad_norm": 9.4952392578125, "learning_rate": 9.81737285235129e-05, "loss": 1.3217, "step": 2085 }, { "epoch": 0.14133748898976894, "grad_norm": 10.110014915466309, "learning_rate": 9.81723595044151e-05, "loss": 1.0413, "step": 2086 }, { "epoch": 0.14140524425774104, "grad_norm": 10.850225448608398, "learning_rate": 9.817099048531728e-05, "loss": 1.3722, "step": 2087 }, { "epoch": 0.14147299952571313, "grad_norm": 10.924457550048828, "learning_rate": 9.816962146621946e-05, "loss": 1.4004, "step": 2088 }, { "epoch": 0.14154075479368522, "grad_norm": 10.022381782531738, "learning_rate": 9.816825244712165e-05, "loss": 1.1989, "step": 2089 }, { "epoch": 0.14160851006165728, "grad_norm": 11.537693977355957, "learning_rate": 9.816688342802383e-05, "loss": 1.1922, "step": 2090 }, { "epoch": 0.14167626532962938, "grad_norm": 8.525372505187988, "learning_rate": 9.816551440892601e-05, "loss": 0.9291, "step": 2091 }, { "epoch": 0.14174402059760147, "grad_norm": 8.972722053527832, "learning_rate": 9.816414538982819e-05, "loss": 1.0917, "step": 2092 }, { "epoch": 0.14181177586557356, "grad_norm": 8.386235237121582, "learning_rate": 9.816277637073039e-05, "loss": 1.0482, "step": 2093 }, { "epoch": 0.14187953113354562, "grad_norm": 9.620473861694336, "learning_rate": 9.816140735163257e-05, "loss": 1.2523, "step": 2094 }, { "epoch": 0.1419472864015177, "grad_norm": 9.236804962158203, "learning_rate": 9.816003833253475e-05, "loss": 0.9474, "step": 2095 }, { "epoch": 0.1420150416694898, "grad_norm": 9.06925106048584, "learning_rate": 9.815866931343693e-05, "loss": 0.8921, "step": 2096 }, { "epoch": 0.1420827969374619, "grad_norm": 10.127729415893555, "learning_rate": 9.815730029433912e-05, "loss": 1.1693, "step": 2097 }, { "epoch": 0.14215055220543396, "grad_norm": 10.703007698059082, "learning_rate": 9.81559312752413e-05, "loss": 1.0652, "step": 2098 }, { "epoch": 0.14221830747340605, "grad_norm": 10.004093170166016, "learning_rate": 9.815456225614348e-05, "loss": 1.0158, "step": 2099 }, { "epoch": 0.14228606274137814, "grad_norm": 8.900236129760742, "learning_rate": 9.815319323704566e-05, "loss": 0.7327, "step": 2100 }, { "epoch": 0.14235381800935024, "grad_norm": 13.102290153503418, "learning_rate": 9.815182421794784e-05, "loss": 1.2837, "step": 2101 }, { "epoch": 0.1424215732773223, "grad_norm": 9.32657527923584, "learning_rate": 9.815045519885004e-05, "loss": 1.0812, "step": 2102 }, { "epoch": 0.1424893285452944, "grad_norm": 11.42785930633545, "learning_rate": 9.814908617975222e-05, "loss": 1.4467, "step": 2103 }, { "epoch": 0.14255708381326648, "grad_norm": 9.878450393676758, "learning_rate": 9.81477171606544e-05, "loss": 1.0774, "step": 2104 }, { "epoch": 0.14262483908123857, "grad_norm": 9.329227447509766, "learning_rate": 9.814634814155658e-05, "loss": 1.2536, "step": 2105 }, { "epoch": 0.14269259434921064, "grad_norm": 9.071858406066895, "learning_rate": 9.814497912245876e-05, "loss": 1.148, "step": 2106 }, { "epoch": 0.14276034961718273, "grad_norm": 11.595810890197754, "learning_rate": 9.814361010336095e-05, "loss": 1.1992, "step": 2107 }, { "epoch": 0.14282810488515482, "grad_norm": 10.041107177734375, "learning_rate": 9.814224108426313e-05, "loss": 1.0447, "step": 2108 }, { "epoch": 0.1428958601531269, "grad_norm": 9.913261413574219, "learning_rate": 9.814087206516531e-05, "loss": 1.1629, "step": 2109 }, { "epoch": 0.14296361542109898, "grad_norm": 10.983177185058594, "learning_rate": 9.813950304606749e-05, "loss": 1.0288, "step": 2110 }, { "epoch": 0.14303137068907107, "grad_norm": 10.479610443115234, "learning_rate": 9.813813402696969e-05, "loss": 1.5308, "step": 2111 }, { "epoch": 0.14309912595704316, "grad_norm": 8.70734691619873, "learning_rate": 9.813676500787187e-05, "loss": 0.9987, "step": 2112 }, { "epoch": 0.14316688122501525, "grad_norm": 8.242984771728516, "learning_rate": 9.813539598877405e-05, "loss": 1.0051, "step": 2113 }, { "epoch": 0.14323463649298732, "grad_norm": 8.627467155456543, "learning_rate": 9.813402696967623e-05, "loss": 1.1134, "step": 2114 }, { "epoch": 0.1433023917609594, "grad_norm": 9.970986366271973, "learning_rate": 9.813265795057841e-05, "loss": 1.148, "step": 2115 }, { "epoch": 0.1433701470289315, "grad_norm": 8.719794273376465, "learning_rate": 9.81312889314806e-05, "loss": 1.1013, "step": 2116 }, { "epoch": 0.1434379022969036, "grad_norm": 9.455860137939453, "learning_rate": 9.812991991238278e-05, "loss": 1.2333, "step": 2117 }, { "epoch": 0.14350565756487568, "grad_norm": 8.851629257202148, "learning_rate": 9.812855089328496e-05, "loss": 1.0985, "step": 2118 }, { "epoch": 0.14357341283284775, "grad_norm": 11.843599319458008, "learning_rate": 9.812718187418714e-05, "loss": 1.2292, "step": 2119 }, { "epoch": 0.14364116810081984, "grad_norm": 9.550506591796875, "learning_rate": 9.812581285508934e-05, "loss": 0.9466, "step": 2120 }, { "epoch": 0.14370892336879193, "grad_norm": 9.234643936157227, "learning_rate": 9.812444383599152e-05, "loss": 1.1968, "step": 2121 }, { "epoch": 0.14377667863676402, "grad_norm": 9.365439414978027, "learning_rate": 9.81230748168937e-05, "loss": 1.1643, "step": 2122 }, { "epoch": 0.14384443390473609, "grad_norm": 9.476024627685547, "learning_rate": 9.812170579779588e-05, "loss": 1.0541, "step": 2123 }, { "epoch": 0.14391218917270818, "grad_norm": 9.53847885131836, "learning_rate": 9.812033677869806e-05, "loss": 1.0808, "step": 2124 }, { "epoch": 0.14397994444068027, "grad_norm": 8.918699264526367, "learning_rate": 9.811896775960025e-05, "loss": 1.1055, "step": 2125 }, { "epoch": 0.14404769970865236, "grad_norm": 12.126187324523926, "learning_rate": 9.811759874050243e-05, "loss": 1.6079, "step": 2126 }, { "epoch": 0.14411545497662442, "grad_norm": 10.922599792480469, "learning_rate": 9.811622972140461e-05, "loss": 1.2569, "step": 2127 }, { "epoch": 0.14418321024459652, "grad_norm": 8.582880020141602, "learning_rate": 9.811486070230679e-05, "loss": 0.9562, "step": 2128 }, { "epoch": 0.1442509655125686, "grad_norm": 8.926568031311035, "learning_rate": 9.811349168320899e-05, "loss": 1.0108, "step": 2129 }, { "epoch": 0.1443187207805407, "grad_norm": 13.906332015991211, "learning_rate": 9.811212266411117e-05, "loss": 1.1851, "step": 2130 }, { "epoch": 0.14438647604851276, "grad_norm": 10.374212265014648, "learning_rate": 9.811075364501335e-05, "loss": 1.2661, "step": 2131 }, { "epoch": 0.14445423131648485, "grad_norm": 10.580810546875, "learning_rate": 9.810938462591554e-05, "loss": 0.8967, "step": 2132 }, { "epoch": 0.14452198658445695, "grad_norm": 8.91865062713623, "learning_rate": 9.810801560681772e-05, "loss": 1.032, "step": 2133 }, { "epoch": 0.14458974185242904, "grad_norm": 8.16041374206543, "learning_rate": 9.81066465877199e-05, "loss": 0.8622, "step": 2134 }, { "epoch": 0.1446574971204011, "grad_norm": 8.575905799865723, "learning_rate": 9.81052775686221e-05, "loss": 0.9956, "step": 2135 }, { "epoch": 0.1447252523883732, "grad_norm": 9.620558738708496, "learning_rate": 9.810390854952428e-05, "loss": 1.1442, "step": 2136 }, { "epoch": 0.14479300765634529, "grad_norm": 10.390005111694336, "learning_rate": 9.810253953042646e-05, "loss": 1.0518, "step": 2137 }, { "epoch": 0.14486076292431738, "grad_norm": 10.714217185974121, "learning_rate": 9.810117051132864e-05, "loss": 1.3215, "step": 2138 }, { "epoch": 0.14492851819228944, "grad_norm": 9.390836715698242, "learning_rate": 9.809980149223083e-05, "loss": 1.0936, "step": 2139 }, { "epoch": 0.14499627346026153, "grad_norm": 10.848445892333984, "learning_rate": 9.809843247313301e-05, "loss": 1.1946, "step": 2140 }, { "epoch": 0.14506402872823362, "grad_norm": 11.041672706604004, "learning_rate": 9.809706345403519e-05, "loss": 1.0422, "step": 2141 }, { "epoch": 0.14513178399620572, "grad_norm": 10.1151762008667, "learning_rate": 9.809569443493737e-05, "loss": 1.2199, "step": 2142 }, { "epoch": 0.14519953926417778, "grad_norm": 8.686629295349121, "learning_rate": 9.809432541583957e-05, "loss": 0.9861, "step": 2143 }, { "epoch": 0.14526729453214987, "grad_norm": 10.579313278198242, "learning_rate": 9.809295639674175e-05, "loss": 0.9372, "step": 2144 }, { "epoch": 0.14533504980012196, "grad_norm": 8.916631698608398, "learning_rate": 9.809158737764393e-05, "loss": 0.8024, "step": 2145 }, { "epoch": 0.14540280506809405, "grad_norm": 9.29333209991455, "learning_rate": 9.80902183585461e-05, "loss": 1.0785, "step": 2146 }, { "epoch": 0.14547056033606612, "grad_norm": 10.584277153015137, "learning_rate": 9.808884933944829e-05, "loss": 0.9167, "step": 2147 }, { "epoch": 0.1455383156040382, "grad_norm": 10.68551254272461, "learning_rate": 9.808748032035048e-05, "loss": 1.0019, "step": 2148 }, { "epoch": 0.1456060708720103, "grad_norm": 9.057500839233398, "learning_rate": 9.808611130125266e-05, "loss": 1.194, "step": 2149 }, { "epoch": 0.1456738261399824, "grad_norm": 8.642207145690918, "learning_rate": 9.808474228215484e-05, "loss": 1.0849, "step": 2150 }, { "epoch": 0.14574158140795446, "grad_norm": 9.460419654846191, "learning_rate": 9.808337326305702e-05, "loss": 1.1747, "step": 2151 }, { "epoch": 0.14580933667592655, "grad_norm": 9.003097534179688, "learning_rate": 9.80820042439592e-05, "loss": 0.8967, "step": 2152 }, { "epoch": 0.14587709194389864, "grad_norm": 9.85009765625, "learning_rate": 9.80806352248614e-05, "loss": 1.1698, "step": 2153 }, { "epoch": 0.14594484721187073, "grad_norm": 10.233894348144531, "learning_rate": 9.807926620576358e-05, "loss": 1.0496, "step": 2154 }, { "epoch": 0.1460126024798428, "grad_norm": 9.660355567932129, "learning_rate": 9.807789718666576e-05, "loss": 0.9616, "step": 2155 }, { "epoch": 0.1460803577478149, "grad_norm": 9.46469497680664, "learning_rate": 9.807652816756794e-05, "loss": 1.0005, "step": 2156 }, { "epoch": 0.14614811301578698, "grad_norm": 8.004712104797363, "learning_rate": 9.807515914847013e-05, "loss": 0.8455, "step": 2157 }, { "epoch": 0.14621586828375907, "grad_norm": 10.039002418518066, "learning_rate": 9.807379012937231e-05, "loss": 1.097, "step": 2158 }, { "epoch": 0.14628362355173113, "grad_norm": 11.16292667388916, "learning_rate": 9.807242111027449e-05, "loss": 1.4332, "step": 2159 }, { "epoch": 0.14635137881970323, "grad_norm": 9.34833812713623, "learning_rate": 9.807105209117667e-05, "loss": 1.0438, "step": 2160 }, { "epoch": 0.14641913408767532, "grad_norm": 10.256768226623535, "learning_rate": 9.806968307207885e-05, "loss": 1.2923, "step": 2161 }, { "epoch": 0.1464868893556474, "grad_norm": 8.890941619873047, "learning_rate": 9.806831405298105e-05, "loss": 0.8419, "step": 2162 }, { "epoch": 0.14655464462361947, "grad_norm": 9.340752601623535, "learning_rate": 9.806694503388323e-05, "loss": 0.9504, "step": 2163 }, { "epoch": 0.14662239989159156, "grad_norm": 10.89192008972168, "learning_rate": 9.80655760147854e-05, "loss": 1.012, "step": 2164 }, { "epoch": 0.14669015515956366, "grad_norm": 8.626432418823242, "learning_rate": 9.806420699568759e-05, "loss": 0.8943, "step": 2165 }, { "epoch": 0.14675791042753575, "grad_norm": 9.465259552001953, "learning_rate": 9.806283797658978e-05, "loss": 1.1617, "step": 2166 }, { "epoch": 0.1468256656955078, "grad_norm": 11.695369720458984, "learning_rate": 9.806146895749196e-05, "loss": 1.2719, "step": 2167 }, { "epoch": 0.1468934209634799, "grad_norm": 9.416814804077148, "learning_rate": 9.806009993839414e-05, "loss": 1.3615, "step": 2168 }, { "epoch": 0.146961176231452, "grad_norm": 9.6004638671875, "learning_rate": 9.805873091929632e-05, "loss": 1.2571, "step": 2169 }, { "epoch": 0.1470289314994241, "grad_norm": 9.187546730041504, "learning_rate": 9.80573619001985e-05, "loss": 0.962, "step": 2170 }, { "epoch": 0.14709668676739618, "grad_norm": 10.691286087036133, "learning_rate": 9.80559928811007e-05, "loss": 1.0619, "step": 2171 }, { "epoch": 0.14716444203536824, "grad_norm": 9.17457103729248, "learning_rate": 9.805462386200288e-05, "loss": 1.1507, "step": 2172 }, { "epoch": 0.14723219730334033, "grad_norm": 10.739082336425781, "learning_rate": 9.805325484290506e-05, "loss": 1.2929, "step": 2173 }, { "epoch": 0.14729995257131243, "grad_norm": 9.091232299804688, "learning_rate": 9.805188582380724e-05, "loss": 0.9422, "step": 2174 }, { "epoch": 0.14736770783928452, "grad_norm": 8.231295585632324, "learning_rate": 9.805051680470943e-05, "loss": 0.9695, "step": 2175 }, { "epoch": 0.14743546310725658, "grad_norm": 9.622401237487793, "learning_rate": 9.804914778561161e-05, "loss": 0.982, "step": 2176 }, { "epoch": 0.14750321837522867, "grad_norm": 9.49541187286377, "learning_rate": 9.804777876651379e-05, "loss": 0.9005, "step": 2177 }, { "epoch": 0.14757097364320076, "grad_norm": 10.584654808044434, "learning_rate": 9.804640974741599e-05, "loss": 0.9427, "step": 2178 }, { "epoch": 0.14763872891117286, "grad_norm": 9.132317543029785, "learning_rate": 9.804504072831817e-05, "loss": 1.0192, "step": 2179 }, { "epoch": 0.14770648417914492, "grad_norm": 8.598082542419434, "learning_rate": 9.804367170922035e-05, "loss": 0.9842, "step": 2180 }, { "epoch": 0.147774239447117, "grad_norm": 8.941360473632812, "learning_rate": 9.804230269012254e-05, "loss": 0.9907, "step": 2181 }, { "epoch": 0.1478419947150891, "grad_norm": 8.119913101196289, "learning_rate": 9.804093367102472e-05, "loss": 1.1259, "step": 2182 }, { "epoch": 0.1479097499830612, "grad_norm": 9.505135536193848, "learning_rate": 9.80395646519269e-05, "loss": 1.0509, "step": 2183 }, { "epoch": 0.14797750525103326, "grad_norm": 8.420906066894531, "learning_rate": 9.803819563282908e-05, "loss": 1.1827, "step": 2184 }, { "epoch": 0.14804526051900535, "grad_norm": 9.860353469848633, "learning_rate": 9.803682661373128e-05, "loss": 1.0286, "step": 2185 }, { "epoch": 0.14811301578697744, "grad_norm": 8.259954452514648, "learning_rate": 9.803545759463346e-05, "loss": 1.0094, "step": 2186 }, { "epoch": 0.14818077105494953, "grad_norm": 10.46882438659668, "learning_rate": 9.803408857553564e-05, "loss": 1.1817, "step": 2187 }, { "epoch": 0.1482485263229216, "grad_norm": 9.315580368041992, "learning_rate": 9.803271955643782e-05, "loss": 1.2816, "step": 2188 }, { "epoch": 0.1483162815908937, "grad_norm": 10.408548355102539, "learning_rate": 9.803135053734001e-05, "loss": 1.0355, "step": 2189 }, { "epoch": 0.14838403685886578, "grad_norm": 9.682003021240234, "learning_rate": 9.802998151824219e-05, "loss": 1.0413, "step": 2190 }, { "epoch": 0.14845179212683787, "grad_norm": 10.041797637939453, "learning_rate": 9.802861249914437e-05, "loss": 0.8388, "step": 2191 }, { "epoch": 0.14851954739480994, "grad_norm": 8.367657661437988, "learning_rate": 9.802724348004655e-05, "loss": 1.0724, "step": 2192 }, { "epoch": 0.14858730266278203, "grad_norm": 9.9558744430542, "learning_rate": 9.802587446094873e-05, "loss": 1.3534, "step": 2193 }, { "epoch": 0.14865505793075412, "grad_norm": 9.244332313537598, "learning_rate": 9.802450544185093e-05, "loss": 1.366, "step": 2194 }, { "epoch": 0.1487228131987262, "grad_norm": 9.560718536376953, "learning_rate": 9.80231364227531e-05, "loss": 1.0556, "step": 2195 }, { "epoch": 0.14879056846669828, "grad_norm": 8.724915504455566, "learning_rate": 9.802176740365529e-05, "loss": 1.316, "step": 2196 }, { "epoch": 0.14885832373467037, "grad_norm": 9.468677520751953, "learning_rate": 9.802039838455747e-05, "loss": 1.1289, "step": 2197 }, { "epoch": 0.14892607900264246, "grad_norm": 8.56347942352295, "learning_rate": 9.801902936545966e-05, "loss": 1.0402, "step": 2198 }, { "epoch": 0.14899383427061455, "grad_norm": 9.328559875488281, "learning_rate": 9.801766034636184e-05, "loss": 1.186, "step": 2199 }, { "epoch": 0.14906158953858661, "grad_norm": 10.20579719543457, "learning_rate": 9.801629132726402e-05, "loss": 1.0896, "step": 2200 }, { "epoch": 0.1491293448065587, "grad_norm": 11.614208221435547, "learning_rate": 9.80149223081662e-05, "loss": 1.2742, "step": 2201 }, { "epoch": 0.1491971000745308, "grad_norm": 8.291358947753906, "learning_rate": 9.801355328906838e-05, "loss": 1.0299, "step": 2202 }, { "epoch": 0.1492648553425029, "grad_norm": 8.734344482421875, "learning_rate": 9.801218426997058e-05, "loss": 1.0922, "step": 2203 }, { "epoch": 0.14933261061047495, "grad_norm": 8.498616218566895, "learning_rate": 9.801081525087276e-05, "loss": 0.8436, "step": 2204 }, { "epoch": 0.14940036587844704, "grad_norm": 7.4501447677612305, "learning_rate": 9.800944623177494e-05, "loss": 0.8629, "step": 2205 }, { "epoch": 0.14946812114641914, "grad_norm": 11.652278900146484, "learning_rate": 9.800807721267712e-05, "loss": 1.1089, "step": 2206 }, { "epoch": 0.14953587641439123, "grad_norm": 11.043471336364746, "learning_rate": 9.80067081935793e-05, "loss": 1.2758, "step": 2207 }, { "epoch": 0.1496036316823633, "grad_norm": 9.376245498657227, "learning_rate": 9.800533917448149e-05, "loss": 1.0178, "step": 2208 }, { "epoch": 0.14967138695033538, "grad_norm": 8.678750038146973, "learning_rate": 9.800397015538367e-05, "loss": 1.0276, "step": 2209 }, { "epoch": 0.14973914221830747, "grad_norm": 10.233052253723145, "learning_rate": 9.800260113628585e-05, "loss": 1.0468, "step": 2210 }, { "epoch": 0.14980689748627957, "grad_norm": 10.711477279663086, "learning_rate": 9.800123211718803e-05, "loss": 1.3919, "step": 2211 }, { "epoch": 0.14987465275425163, "grad_norm": 10.130411148071289, "learning_rate": 9.799986309809023e-05, "loss": 1.1085, "step": 2212 }, { "epoch": 0.14994240802222372, "grad_norm": 9.64091682434082, "learning_rate": 9.79984940789924e-05, "loss": 1.0661, "step": 2213 }, { "epoch": 0.1500101632901958, "grad_norm": 9.80176830291748, "learning_rate": 9.799712505989459e-05, "loss": 1.3426, "step": 2214 }, { "epoch": 0.1500779185581679, "grad_norm": 9.34835147857666, "learning_rate": 9.799575604079677e-05, "loss": 1.0281, "step": 2215 }, { "epoch": 0.15014567382613997, "grad_norm": 8.638712882995605, "learning_rate": 9.799438702169895e-05, "loss": 0.9037, "step": 2216 }, { "epoch": 0.15021342909411206, "grad_norm": 10.499733924865723, "learning_rate": 9.799301800260114e-05, "loss": 1.1152, "step": 2217 }, { "epoch": 0.15028118436208415, "grad_norm": 9.093478202819824, "learning_rate": 9.799164898350332e-05, "loss": 1.2026, "step": 2218 }, { "epoch": 0.15034893963005624, "grad_norm": 13.181863784790039, "learning_rate": 9.79902799644055e-05, "loss": 1.3256, "step": 2219 }, { "epoch": 0.1504166948980283, "grad_norm": 10.221563339233398, "learning_rate": 9.798891094530768e-05, "loss": 1.1867, "step": 2220 }, { "epoch": 0.1504844501660004, "grad_norm": 9.512944221496582, "learning_rate": 9.798754192620988e-05, "loss": 1.1145, "step": 2221 }, { "epoch": 0.1505522054339725, "grad_norm": 8.991315841674805, "learning_rate": 9.798617290711206e-05, "loss": 1.0289, "step": 2222 }, { "epoch": 0.15061996070194458, "grad_norm": 9.003118515014648, "learning_rate": 9.798480388801424e-05, "loss": 0.9618, "step": 2223 }, { "epoch": 0.15068771596991667, "grad_norm": 11.337166786193848, "learning_rate": 9.798343486891643e-05, "loss": 1.3864, "step": 2224 }, { "epoch": 0.15075547123788874, "grad_norm": 9.193288803100586, "learning_rate": 9.798206584981861e-05, "loss": 1.1314, "step": 2225 }, { "epoch": 0.15082322650586083, "grad_norm": 10.137048721313477, "learning_rate": 9.798069683072079e-05, "loss": 1.3325, "step": 2226 }, { "epoch": 0.15089098177383292, "grad_norm": 8.248672485351562, "learning_rate": 9.797932781162299e-05, "loss": 1.0688, "step": 2227 }, { "epoch": 0.150958737041805, "grad_norm": 10.51007080078125, "learning_rate": 9.797795879252517e-05, "loss": 1.2191, "step": 2228 }, { "epoch": 0.15102649230977708, "grad_norm": 14.15853214263916, "learning_rate": 9.797658977342735e-05, "loss": 1.0961, "step": 2229 }, { "epoch": 0.15109424757774917, "grad_norm": 8.42485523223877, "learning_rate": 9.797522075432954e-05, "loss": 1.0314, "step": 2230 }, { "epoch": 0.15116200284572126, "grad_norm": 9.3310546875, "learning_rate": 9.797385173523172e-05, "loss": 1.2372, "step": 2231 }, { "epoch": 0.15122975811369335, "grad_norm": 9.323569297790527, "learning_rate": 9.79724827161339e-05, "loss": 1.1823, "step": 2232 }, { "epoch": 0.15129751338166542, "grad_norm": 9.719592094421387, "learning_rate": 9.797111369703608e-05, "loss": 1.0689, "step": 2233 }, { "epoch": 0.1513652686496375, "grad_norm": 9.37340259552002, "learning_rate": 9.796974467793826e-05, "loss": 1.0703, "step": 2234 }, { "epoch": 0.1514330239176096, "grad_norm": 9.803778648376465, "learning_rate": 9.796837565884045e-05, "loss": 1.0422, "step": 2235 }, { "epoch": 0.1515007791855817, "grad_norm": 8.317301750183105, "learning_rate": 9.796700663974264e-05, "loss": 0.9829, "step": 2236 }, { "epoch": 0.15156853445355375, "grad_norm": 8.473258972167969, "learning_rate": 9.796563762064482e-05, "loss": 1.064, "step": 2237 }, { "epoch": 0.15163628972152585, "grad_norm": 8.495006561279297, "learning_rate": 9.7964268601547e-05, "loss": 0.8646, "step": 2238 }, { "epoch": 0.15170404498949794, "grad_norm": 9.140898704528809, "learning_rate": 9.796289958244918e-05, "loss": 1.068, "step": 2239 }, { "epoch": 0.15177180025747003, "grad_norm": 10.264453887939453, "learning_rate": 9.796153056335137e-05, "loss": 0.9681, "step": 2240 }, { "epoch": 0.1518395555254421, "grad_norm": 10.407776832580566, "learning_rate": 9.796016154425355e-05, "loss": 1.2538, "step": 2241 }, { "epoch": 0.15190731079341419, "grad_norm": 10.143677711486816, "learning_rate": 9.795879252515573e-05, "loss": 1.138, "step": 2242 }, { "epoch": 0.15197506606138628, "grad_norm": 12.528799057006836, "learning_rate": 9.795742350605791e-05, "loss": 1.2654, "step": 2243 }, { "epoch": 0.15204282132935837, "grad_norm": 10.635498046875, "learning_rate": 9.79560544869601e-05, "loss": 1.2116, "step": 2244 }, { "epoch": 0.15211057659733043, "grad_norm": 10.71164608001709, "learning_rate": 9.795468546786229e-05, "loss": 1.0262, "step": 2245 }, { "epoch": 0.15217833186530252, "grad_norm": 9.121109962463379, "learning_rate": 9.795331644876447e-05, "loss": 1.1657, "step": 2246 }, { "epoch": 0.15224608713327462, "grad_norm": 8.174636840820312, "learning_rate": 9.795194742966665e-05, "loss": 1.0027, "step": 2247 }, { "epoch": 0.1523138424012467, "grad_norm": 8.763494491577148, "learning_rate": 9.795057841056883e-05, "loss": 1.1103, "step": 2248 }, { "epoch": 0.15238159766921877, "grad_norm": 8.028278350830078, "learning_rate": 9.794920939147102e-05, "loss": 1.0205, "step": 2249 }, { "epoch": 0.15244935293719086, "grad_norm": 8.96112060546875, "learning_rate": 9.79478403723732e-05, "loss": 1.2949, "step": 2250 }, { "epoch": 0.15251710820516295, "grad_norm": 8.67423152923584, "learning_rate": 9.794647135327538e-05, "loss": 1.0602, "step": 2251 }, { "epoch": 0.15258486347313505, "grad_norm": 9.45671272277832, "learning_rate": 9.794510233417756e-05, "loss": 1.0362, "step": 2252 }, { "epoch": 0.1526526187411071, "grad_norm": 10.24669075012207, "learning_rate": 9.794373331507976e-05, "loss": 0.9918, "step": 2253 }, { "epoch": 0.1527203740090792, "grad_norm": 9.014079093933105, "learning_rate": 9.794236429598194e-05, "loss": 1.1636, "step": 2254 }, { "epoch": 0.1527881292770513, "grad_norm": 6.800943851470947, "learning_rate": 9.794099527688412e-05, "loss": 0.9089, "step": 2255 }, { "epoch": 0.15285588454502338, "grad_norm": 9.90794849395752, "learning_rate": 9.79396262577863e-05, "loss": 0.7669, "step": 2256 }, { "epoch": 0.15292363981299545, "grad_norm": 9.870927810668945, "learning_rate": 9.793825723868848e-05, "loss": 1.1243, "step": 2257 }, { "epoch": 0.15299139508096754, "grad_norm": 9.707404136657715, "learning_rate": 9.793688821959067e-05, "loss": 1.1558, "step": 2258 }, { "epoch": 0.15305915034893963, "grad_norm": 8.362896919250488, "learning_rate": 9.793551920049285e-05, "loss": 0.8601, "step": 2259 }, { "epoch": 0.15312690561691172, "grad_norm": 9.536920547485352, "learning_rate": 9.793415018139503e-05, "loss": 1.2741, "step": 2260 }, { "epoch": 0.1531946608848838, "grad_norm": 11.108535766601562, "learning_rate": 9.793278116229721e-05, "loss": 1.3118, "step": 2261 }, { "epoch": 0.15326241615285588, "grad_norm": 7.281479358673096, "learning_rate": 9.793141214319939e-05, "loss": 0.9165, "step": 2262 }, { "epoch": 0.15333017142082797, "grad_norm": 9.166728973388672, "learning_rate": 9.793004312410159e-05, "loss": 1.0845, "step": 2263 }, { "epoch": 0.15339792668880006, "grad_norm": 11.539854049682617, "learning_rate": 9.792867410500377e-05, "loss": 1.1129, "step": 2264 }, { "epoch": 0.15346568195677213, "grad_norm": 8.588869094848633, "learning_rate": 9.792730508590595e-05, "loss": 0.9698, "step": 2265 }, { "epoch": 0.15353343722474422, "grad_norm": 8.270078659057617, "learning_rate": 9.792593606680813e-05, "loss": 1.2306, "step": 2266 }, { "epoch": 0.1536011924927163, "grad_norm": 7.908688545227051, "learning_rate": 9.792456704771032e-05, "loss": 1.283, "step": 2267 }, { "epoch": 0.1536689477606884, "grad_norm": 10.376410484313965, "learning_rate": 9.79231980286125e-05, "loss": 1.0246, "step": 2268 }, { "epoch": 0.15373670302866047, "grad_norm": 9.517715454101562, "learning_rate": 9.792182900951468e-05, "loss": 0.9975, "step": 2269 }, { "epoch": 0.15380445829663256, "grad_norm": 8.789438247680664, "learning_rate": 9.792045999041688e-05, "loss": 1.0886, "step": 2270 }, { "epoch": 0.15387221356460465, "grad_norm": 9.649114608764648, "learning_rate": 9.791909097131906e-05, "loss": 1.1497, "step": 2271 }, { "epoch": 0.15393996883257674, "grad_norm": 8.533876419067383, "learning_rate": 9.791772195222124e-05, "loss": 0.8701, "step": 2272 }, { "epoch": 0.1540077241005488, "grad_norm": 10.64561653137207, "learning_rate": 9.791635293312343e-05, "loss": 1.2755, "step": 2273 }, { "epoch": 0.1540754793685209, "grad_norm": 9.510658264160156, "learning_rate": 9.791498391402561e-05, "loss": 0.8772, "step": 2274 }, { "epoch": 0.154143234636493, "grad_norm": 9.849981307983398, "learning_rate": 9.791361489492779e-05, "loss": 1.1689, "step": 2275 }, { "epoch": 0.15421098990446508, "grad_norm": 8.152152061462402, "learning_rate": 9.791224587582998e-05, "loss": 0.8148, "step": 2276 }, { "epoch": 0.15427874517243717, "grad_norm": 8.653456687927246, "learning_rate": 9.791087685673216e-05, "loss": 1.0061, "step": 2277 }, { "epoch": 0.15434650044040923, "grad_norm": 6.928426742553711, "learning_rate": 9.790950783763435e-05, "loss": 0.9192, "step": 2278 }, { "epoch": 0.15441425570838133, "grad_norm": 9.606708526611328, "learning_rate": 9.790813881853653e-05, "loss": 0.9956, "step": 2279 }, { "epoch": 0.15448201097635342, "grad_norm": 9.42773151397705, "learning_rate": 9.79067697994387e-05, "loss": 0.9302, "step": 2280 }, { "epoch": 0.1545497662443255, "grad_norm": 8.248319625854492, "learning_rate": 9.79054007803409e-05, "loss": 1.2318, "step": 2281 }, { "epoch": 0.15461752151229757, "grad_norm": 9.706393241882324, "learning_rate": 9.790403176124308e-05, "loss": 1.1489, "step": 2282 }, { "epoch": 0.15468527678026966, "grad_norm": 8.88716983795166, "learning_rate": 9.790266274214526e-05, "loss": 0.8686, "step": 2283 }, { "epoch": 0.15475303204824176, "grad_norm": 8.596991539001465, "learning_rate": 9.790129372304744e-05, "loss": 1.0623, "step": 2284 }, { "epoch": 0.15482078731621385, "grad_norm": 9.262425422668457, "learning_rate": 9.789992470394962e-05, "loss": 1.2067, "step": 2285 }, { "epoch": 0.1548885425841859, "grad_norm": 7.772172927856445, "learning_rate": 9.789855568485181e-05, "loss": 0.9976, "step": 2286 }, { "epoch": 0.154956297852158, "grad_norm": 9.320001602172852, "learning_rate": 9.7897186665754e-05, "loss": 1.1377, "step": 2287 }, { "epoch": 0.1550240531201301, "grad_norm": 11.02434253692627, "learning_rate": 9.789581764665618e-05, "loss": 1.1365, "step": 2288 }, { "epoch": 0.1550918083881022, "grad_norm": 9.90654182434082, "learning_rate": 9.789444862755836e-05, "loss": 1.208, "step": 2289 }, { "epoch": 0.15515956365607425, "grad_norm": 9.591778755187988, "learning_rate": 9.789307960846055e-05, "loss": 1.2426, "step": 2290 }, { "epoch": 0.15522731892404634, "grad_norm": 9.221457481384277, "learning_rate": 9.789171058936273e-05, "loss": 1.1449, "step": 2291 }, { "epoch": 0.15529507419201843, "grad_norm": 7.294323444366455, "learning_rate": 9.789034157026491e-05, "loss": 0.8846, "step": 2292 }, { "epoch": 0.15536282945999053, "grad_norm": 10.6463041305542, "learning_rate": 9.788897255116709e-05, "loss": 1.1397, "step": 2293 }, { "epoch": 0.1554305847279626, "grad_norm": 9.412870407104492, "learning_rate": 9.788760353206927e-05, "loss": 1.1857, "step": 2294 }, { "epoch": 0.15549833999593468, "grad_norm": 9.163009643554688, "learning_rate": 9.788623451297147e-05, "loss": 0.8907, "step": 2295 }, { "epoch": 0.15556609526390677, "grad_norm": 8.157798767089844, "learning_rate": 9.788486549387365e-05, "loss": 0.893, "step": 2296 }, { "epoch": 0.15563385053187886, "grad_norm": 9.155460357666016, "learning_rate": 9.788349647477583e-05, "loss": 1.0361, "step": 2297 }, { "epoch": 0.15570160579985093, "grad_norm": 10.280989646911621, "learning_rate": 9.7882127455678e-05, "loss": 1.2953, "step": 2298 }, { "epoch": 0.15576936106782302, "grad_norm": 9.654706954956055, "learning_rate": 9.78807584365802e-05, "loss": 0.9012, "step": 2299 }, { "epoch": 0.1558371163357951, "grad_norm": 12.425939559936523, "learning_rate": 9.787938941748238e-05, "loss": 1.4103, "step": 2300 }, { "epoch": 0.1559048716037672, "grad_norm": 9.888884544372559, "learning_rate": 9.787802039838456e-05, "loss": 1.1599, "step": 2301 }, { "epoch": 0.15597262687173927, "grad_norm": 10.30229377746582, "learning_rate": 9.787665137928674e-05, "loss": 1.0505, "step": 2302 }, { "epoch": 0.15604038213971136, "grad_norm": 11.208477973937988, "learning_rate": 9.787528236018892e-05, "loss": 1.009, "step": 2303 }, { "epoch": 0.15610813740768345, "grad_norm": 9.264431953430176, "learning_rate": 9.787391334109112e-05, "loss": 1.1453, "step": 2304 }, { "epoch": 0.15617589267565554, "grad_norm": 9.0980224609375, "learning_rate": 9.78725443219933e-05, "loss": 1.0824, "step": 2305 }, { "epoch": 0.1562436479436276, "grad_norm": 8.346585273742676, "learning_rate": 9.787117530289548e-05, "loss": 1.0632, "step": 2306 }, { "epoch": 0.1563114032115997, "grad_norm": 10.607507705688477, "learning_rate": 9.786980628379766e-05, "loss": 1.0102, "step": 2307 }, { "epoch": 0.1563791584795718, "grad_norm": 8.189208984375, "learning_rate": 9.786843726469985e-05, "loss": 1.0872, "step": 2308 }, { "epoch": 0.15644691374754388, "grad_norm": 9.84202766418457, "learning_rate": 9.786706824560203e-05, "loss": 0.9871, "step": 2309 }, { "epoch": 0.15651466901551594, "grad_norm": 11.937589645385742, "learning_rate": 9.786569922650421e-05, "loss": 1.2255, "step": 2310 }, { "epoch": 0.15658242428348804, "grad_norm": 9.855016708374023, "learning_rate": 9.786433020740639e-05, "loss": 1.0097, "step": 2311 }, { "epoch": 0.15665017955146013, "grad_norm": 8.661060333251953, "learning_rate": 9.786296118830857e-05, "loss": 0.8732, "step": 2312 }, { "epoch": 0.15671793481943222, "grad_norm": 10.57170295715332, "learning_rate": 9.786159216921077e-05, "loss": 0.862, "step": 2313 }, { "epoch": 0.15678569008740428, "grad_norm": 7.759045600891113, "learning_rate": 9.786022315011295e-05, "loss": 0.77, "step": 2314 }, { "epoch": 0.15685344535537638, "grad_norm": 10.758045196533203, "learning_rate": 9.785885413101513e-05, "loss": 0.8919, "step": 2315 }, { "epoch": 0.15692120062334847, "grad_norm": 8.521660804748535, "learning_rate": 9.78574851119173e-05, "loss": 1.131, "step": 2316 }, { "epoch": 0.15698895589132056, "grad_norm": 8.72917652130127, "learning_rate": 9.78561160928195e-05, "loss": 0.8359, "step": 2317 }, { "epoch": 0.15705671115929262, "grad_norm": 11.679365158081055, "learning_rate": 9.785474707372168e-05, "loss": 1.0078, "step": 2318 }, { "epoch": 0.15712446642726471, "grad_norm": 11.50632381439209, "learning_rate": 9.785337805462386e-05, "loss": 1.208, "step": 2319 }, { "epoch": 0.1571922216952368, "grad_norm": 9.6107759475708, "learning_rate": 9.785200903552605e-05, "loss": 1.0967, "step": 2320 }, { "epoch": 0.1572599769632089, "grad_norm": 8.629117012023926, "learning_rate": 9.785064001642824e-05, "loss": 1.0594, "step": 2321 }, { "epoch": 0.15732773223118096, "grad_norm": 11.136920928955078, "learning_rate": 9.784927099733042e-05, "loss": 1.2874, "step": 2322 }, { "epoch": 0.15739548749915305, "grad_norm": 11.097023963928223, "learning_rate": 9.784790197823261e-05, "loss": 1.1598, "step": 2323 }, { "epoch": 0.15746324276712514, "grad_norm": 11.117433547973633, "learning_rate": 9.784653295913479e-05, "loss": 1.0601, "step": 2324 }, { "epoch": 0.15753099803509724, "grad_norm": 10.152684211730957, "learning_rate": 9.784516394003697e-05, "loss": 1.2763, "step": 2325 }, { "epoch": 0.1575987533030693, "grad_norm": 8.37531852722168, "learning_rate": 9.784379492093915e-05, "loss": 1.1055, "step": 2326 }, { "epoch": 0.1576665085710414, "grad_norm": 10.463939666748047, "learning_rate": 9.784242590184134e-05, "loss": 1.3088, "step": 2327 }, { "epoch": 0.15773426383901348, "grad_norm": 8.74315357208252, "learning_rate": 9.784105688274352e-05, "loss": 1.1165, "step": 2328 }, { "epoch": 0.15780201910698557, "grad_norm": 8.691280364990234, "learning_rate": 9.78396878636457e-05, "loss": 1.1019, "step": 2329 }, { "epoch": 0.15786977437495767, "grad_norm": 10.424938201904297, "learning_rate": 9.783831884454789e-05, "loss": 1.1957, "step": 2330 }, { "epoch": 0.15793752964292973, "grad_norm": 6.867722034454346, "learning_rate": 9.783694982545008e-05, "loss": 1.0476, "step": 2331 }, { "epoch": 0.15800528491090182, "grad_norm": 9.434804916381836, "learning_rate": 9.783558080635226e-05, "loss": 1.0247, "step": 2332 }, { "epoch": 0.1580730401788739, "grad_norm": 7.771170616149902, "learning_rate": 9.783421178725444e-05, "loss": 1.2582, "step": 2333 }, { "epoch": 0.158140795446846, "grad_norm": 8.366608619689941, "learning_rate": 9.783284276815662e-05, "loss": 1.1078, "step": 2334 }, { "epoch": 0.15820855071481807, "grad_norm": 8.85851764678955, "learning_rate": 9.78314737490588e-05, "loss": 1.2405, "step": 2335 }, { "epoch": 0.15827630598279016, "grad_norm": 9.134325981140137, "learning_rate": 9.7830104729961e-05, "loss": 1.2277, "step": 2336 }, { "epoch": 0.15834406125076225, "grad_norm": 9.150130271911621, "learning_rate": 9.782873571086317e-05, "loss": 1.1355, "step": 2337 }, { "epoch": 0.15841181651873434, "grad_norm": 8.687469482421875, "learning_rate": 9.782736669176536e-05, "loss": 1.0454, "step": 2338 }, { "epoch": 0.1584795717867064, "grad_norm": 10.071285247802734, "learning_rate": 9.782599767266754e-05, "loss": 1.0041, "step": 2339 }, { "epoch": 0.1585473270546785, "grad_norm": 8.373369216918945, "learning_rate": 9.782462865356972e-05, "loss": 0.8378, "step": 2340 }, { "epoch": 0.1586150823226506, "grad_norm": 9.427014350891113, "learning_rate": 9.782325963447191e-05, "loss": 1.1413, "step": 2341 }, { "epoch": 0.15868283759062268, "grad_norm": 8.38814640045166, "learning_rate": 9.782189061537409e-05, "loss": 1.0696, "step": 2342 }, { "epoch": 0.15875059285859475, "grad_norm": 8.518644332885742, "learning_rate": 9.782052159627627e-05, "loss": 0.9814, "step": 2343 }, { "epoch": 0.15881834812656684, "grad_norm": 9.926252365112305, "learning_rate": 9.781915257717845e-05, "loss": 1.1971, "step": 2344 }, { "epoch": 0.15888610339453893, "grad_norm": 7.946019649505615, "learning_rate": 9.781778355808064e-05, "loss": 1.1902, "step": 2345 }, { "epoch": 0.15895385866251102, "grad_norm": 8.686945915222168, "learning_rate": 9.781641453898283e-05, "loss": 0.9939, "step": 2346 }, { "epoch": 0.15902161393048309, "grad_norm": 8.225680351257324, "learning_rate": 9.7815045519885e-05, "loss": 1.3665, "step": 2347 }, { "epoch": 0.15908936919845518, "grad_norm": 10.381987571716309, "learning_rate": 9.781367650078719e-05, "loss": 1.09, "step": 2348 }, { "epoch": 0.15915712446642727, "grad_norm": 8.57552719116211, "learning_rate": 9.781230748168937e-05, "loss": 0.8261, "step": 2349 }, { "epoch": 0.15922487973439936, "grad_norm": 9.105220794677734, "learning_rate": 9.781093846259156e-05, "loss": 1.102, "step": 2350 }, { "epoch": 0.15929263500237142, "grad_norm": 10.08092212677002, "learning_rate": 9.780956944349374e-05, "loss": 0.943, "step": 2351 }, { "epoch": 0.15936039027034352, "grad_norm": 10.259852409362793, "learning_rate": 9.780820042439592e-05, "loss": 0.8822, "step": 2352 }, { "epoch": 0.1594281455383156, "grad_norm": 8.31139087677002, "learning_rate": 9.78068314052981e-05, "loss": 1.2167, "step": 2353 }, { "epoch": 0.1594959008062877, "grad_norm": 7.529703617095947, "learning_rate": 9.78054623862003e-05, "loss": 0.8913, "step": 2354 }, { "epoch": 0.15956365607425976, "grad_norm": 8.792675971984863, "learning_rate": 9.780409336710248e-05, "loss": 0.966, "step": 2355 }, { "epoch": 0.15963141134223185, "grad_norm": 9.329866409301758, "learning_rate": 9.780272434800466e-05, "loss": 1.1659, "step": 2356 }, { "epoch": 0.15969916661020395, "grad_norm": 12.14089298248291, "learning_rate": 9.780135532890684e-05, "loss": 1.2019, "step": 2357 }, { "epoch": 0.15976692187817604, "grad_norm": 9.12912654876709, "learning_rate": 9.779998630980902e-05, "loss": 1.1615, "step": 2358 }, { "epoch": 0.1598346771461481, "grad_norm": 9.554464340209961, "learning_rate": 9.779861729071121e-05, "loss": 1.1695, "step": 2359 }, { "epoch": 0.1599024324141202, "grad_norm": 9.317673683166504, "learning_rate": 9.779724827161339e-05, "loss": 0.9606, "step": 2360 }, { "epoch": 0.15997018768209229, "grad_norm": 8.80395793914795, "learning_rate": 9.779587925251557e-05, "loss": 0.9693, "step": 2361 }, { "epoch": 0.16003794295006438, "grad_norm": 11.990642547607422, "learning_rate": 9.779451023341775e-05, "loss": 1.2901, "step": 2362 }, { "epoch": 0.16010569821803644, "grad_norm": 8.188547134399414, "learning_rate": 9.779314121431995e-05, "loss": 0.979, "step": 2363 }, { "epoch": 0.16017345348600853, "grad_norm": 8.316620826721191, "learning_rate": 9.779177219522213e-05, "loss": 0.8601, "step": 2364 }, { "epoch": 0.16024120875398062, "grad_norm": 7.58405876159668, "learning_rate": 9.77904031761243e-05, "loss": 1.1812, "step": 2365 }, { "epoch": 0.16030896402195272, "grad_norm": 7.725598335266113, "learning_rate": 9.77890341570265e-05, "loss": 0.9335, "step": 2366 }, { "epoch": 0.16037671928992478, "grad_norm": 8.6231107711792, "learning_rate": 9.778766513792868e-05, "loss": 1.0162, "step": 2367 }, { "epoch": 0.16044447455789687, "grad_norm": 9.762526512145996, "learning_rate": 9.778629611883086e-05, "loss": 1.3186, "step": 2368 }, { "epoch": 0.16051222982586896, "grad_norm": 11.384220123291016, "learning_rate": 9.778492709973305e-05, "loss": 1.2397, "step": 2369 }, { "epoch": 0.16057998509384105, "grad_norm": 8.841899871826172, "learning_rate": 9.778355808063523e-05, "loss": 0.8708, "step": 2370 }, { "epoch": 0.16064774036181312, "grad_norm": 7.778527736663818, "learning_rate": 9.778218906153741e-05, "loss": 0.9968, "step": 2371 }, { "epoch": 0.1607154956297852, "grad_norm": 8.559181213378906, "learning_rate": 9.77808200424396e-05, "loss": 0.9759, "step": 2372 }, { "epoch": 0.1607832508977573, "grad_norm": 10.273273468017578, "learning_rate": 9.777945102334179e-05, "loss": 0.9485, "step": 2373 }, { "epoch": 0.1608510061657294, "grad_norm": 7.946044445037842, "learning_rate": 9.777808200424397e-05, "loss": 1.105, "step": 2374 }, { "epoch": 0.16091876143370146, "grad_norm": 9.917662620544434, "learning_rate": 9.777671298514615e-05, "loss": 1.0272, "step": 2375 }, { "epoch": 0.16098651670167355, "grad_norm": 10.438239097595215, "learning_rate": 9.777534396604833e-05, "loss": 1.0197, "step": 2376 }, { "epoch": 0.16105427196964564, "grad_norm": 8.799901962280273, "learning_rate": 9.777397494695052e-05, "loss": 1.1401, "step": 2377 }, { "epoch": 0.16112202723761773, "grad_norm": 8.569243431091309, "learning_rate": 9.77726059278527e-05, "loss": 1.2019, "step": 2378 }, { "epoch": 0.1611897825055898, "grad_norm": 10.793002128601074, "learning_rate": 9.777123690875488e-05, "loss": 1.0932, "step": 2379 }, { "epoch": 0.1612575377735619, "grad_norm": 7.825850963592529, "learning_rate": 9.776986788965707e-05, "loss": 1.2226, "step": 2380 }, { "epoch": 0.16132529304153398, "grad_norm": 8.780813217163086, "learning_rate": 9.776849887055925e-05, "loss": 0.7939, "step": 2381 }, { "epoch": 0.16139304830950607, "grad_norm": 8.927638053894043, "learning_rate": 9.776712985146144e-05, "loss": 1.0551, "step": 2382 }, { "epoch": 0.16146080357747816, "grad_norm": 9.08043384552002, "learning_rate": 9.776576083236362e-05, "loss": 1.0037, "step": 2383 }, { "epoch": 0.16152855884545023, "grad_norm": 9.362268447875977, "learning_rate": 9.77643918132658e-05, "loss": 1.332, "step": 2384 }, { "epoch": 0.16159631411342232, "grad_norm": 10.533197402954102, "learning_rate": 9.776302279416798e-05, "loss": 1.0966, "step": 2385 }, { "epoch": 0.1616640693813944, "grad_norm": 9.579266548156738, "learning_rate": 9.776165377507017e-05, "loss": 0.9325, "step": 2386 }, { "epoch": 0.1617318246493665, "grad_norm": 10.071297645568848, "learning_rate": 9.776028475597235e-05, "loss": 0.9496, "step": 2387 }, { "epoch": 0.16179957991733857, "grad_norm": 9.243900299072266, "learning_rate": 9.775891573687453e-05, "loss": 1.13, "step": 2388 }, { "epoch": 0.16186733518531066, "grad_norm": 9.327018737792969, "learning_rate": 9.775754671777672e-05, "loss": 1.2587, "step": 2389 }, { "epoch": 0.16193509045328275, "grad_norm": 7.614950180053711, "learning_rate": 9.77561776986789e-05, "loss": 1.0848, "step": 2390 }, { "epoch": 0.16200284572125484, "grad_norm": 9.86501407623291, "learning_rate": 9.775480867958109e-05, "loss": 1.2504, "step": 2391 }, { "epoch": 0.1620706009892269, "grad_norm": 9.08303451538086, "learning_rate": 9.775343966048327e-05, "loss": 1.4359, "step": 2392 }, { "epoch": 0.162138356257199, "grad_norm": 8.417489051818848, "learning_rate": 9.775207064138545e-05, "loss": 1.244, "step": 2393 }, { "epoch": 0.1622061115251711, "grad_norm": 8.35366439819336, "learning_rate": 9.775070162228763e-05, "loss": 1.0706, "step": 2394 }, { "epoch": 0.16227386679314318, "grad_norm": 9.732915878295898, "learning_rate": 9.774933260318981e-05, "loss": 1.4237, "step": 2395 }, { "epoch": 0.16234162206111524, "grad_norm": 9.131365776062012, "learning_rate": 9.7747963584092e-05, "loss": 0.9207, "step": 2396 }, { "epoch": 0.16240937732908733, "grad_norm": 8.93538761138916, "learning_rate": 9.774659456499419e-05, "loss": 1.0054, "step": 2397 }, { "epoch": 0.16247713259705943, "grad_norm": 8.939055442810059, "learning_rate": 9.774522554589637e-05, "loss": 0.9782, "step": 2398 }, { "epoch": 0.16254488786503152, "grad_norm": 9.251758575439453, "learning_rate": 9.774385652679855e-05, "loss": 1.0968, "step": 2399 }, { "epoch": 0.16261264313300358, "grad_norm": 9.240782737731934, "learning_rate": 9.774248750770074e-05, "loss": 0.7982, "step": 2400 }, { "epoch": 0.16268039840097567, "grad_norm": 8.697726249694824, "learning_rate": 9.774111848860292e-05, "loss": 1.0545, "step": 2401 }, { "epoch": 0.16274815366894776, "grad_norm": 7.9362053871154785, "learning_rate": 9.77397494695051e-05, "loss": 0.8364, "step": 2402 }, { "epoch": 0.16281590893691986, "grad_norm": 11.944025993347168, "learning_rate": 9.773838045040728e-05, "loss": 1.2313, "step": 2403 }, { "epoch": 0.16288366420489192, "grad_norm": 9.496225357055664, "learning_rate": 9.773701143130946e-05, "loss": 1.1363, "step": 2404 }, { "epoch": 0.162951419472864, "grad_norm": 12.96069622039795, "learning_rate": 9.773564241221165e-05, "loss": 1.1205, "step": 2405 }, { "epoch": 0.1630191747408361, "grad_norm": 8.915671348571777, "learning_rate": 9.773427339311384e-05, "loss": 1.1149, "step": 2406 }, { "epoch": 0.1630869300088082, "grad_norm": 10.23763370513916, "learning_rate": 9.773290437401602e-05, "loss": 1.1867, "step": 2407 }, { "epoch": 0.16315468527678026, "grad_norm": 9.117730140686035, "learning_rate": 9.77315353549182e-05, "loss": 1.1792, "step": 2408 }, { "epoch": 0.16322244054475235, "grad_norm": 9.380385398864746, "learning_rate": 9.773016633582039e-05, "loss": 1.1621, "step": 2409 }, { "epoch": 0.16329019581272444, "grad_norm": 8.56508731842041, "learning_rate": 9.772879731672257e-05, "loss": 0.8372, "step": 2410 }, { "epoch": 0.16335795108069653, "grad_norm": 11.707832336425781, "learning_rate": 9.772742829762475e-05, "loss": 1.3191, "step": 2411 }, { "epoch": 0.1634257063486686, "grad_norm": 7.720577716827393, "learning_rate": 9.772605927852694e-05, "loss": 0.9617, "step": 2412 }, { "epoch": 0.1634934616166407, "grad_norm": 8.586542129516602, "learning_rate": 9.772469025942912e-05, "loss": 1.0978, "step": 2413 }, { "epoch": 0.16356121688461278, "grad_norm": 9.021394729614258, "learning_rate": 9.77233212403313e-05, "loss": 0.9358, "step": 2414 }, { "epoch": 0.16362897215258487, "grad_norm": 9.078686714172363, "learning_rate": 9.77219522212335e-05, "loss": 1.3338, "step": 2415 }, { "epoch": 0.16369672742055694, "grad_norm": 9.810312271118164, "learning_rate": 9.772058320213568e-05, "loss": 1.2472, "step": 2416 }, { "epoch": 0.16376448268852903, "grad_norm": 8.933609962463379, "learning_rate": 9.771921418303786e-05, "loss": 0.9115, "step": 2417 }, { "epoch": 0.16383223795650112, "grad_norm": 7.044286251068115, "learning_rate": 9.771784516394005e-05, "loss": 0.7983, "step": 2418 }, { "epoch": 0.1638999932244732, "grad_norm": 11.711495399475098, "learning_rate": 9.771647614484223e-05, "loss": 1.096, "step": 2419 }, { "epoch": 0.16396774849244528, "grad_norm": 9.31049633026123, "learning_rate": 9.771510712574441e-05, "loss": 1.2711, "step": 2420 }, { "epoch": 0.16403550376041737, "grad_norm": 8.10503101348877, "learning_rate": 9.77137381066466e-05, "loss": 0.9095, "step": 2421 }, { "epoch": 0.16410325902838946, "grad_norm": 7.915055751800537, "learning_rate": 9.771236908754877e-05, "loss": 1.0161, "step": 2422 }, { "epoch": 0.16417101429636155, "grad_norm": 8.185515403747559, "learning_rate": 9.771100006845097e-05, "loss": 1.1109, "step": 2423 }, { "epoch": 0.16423876956433361, "grad_norm": 9.960200309753418, "learning_rate": 9.770963104935315e-05, "loss": 0.9757, "step": 2424 }, { "epoch": 0.1643065248323057, "grad_norm": 9.646814346313477, "learning_rate": 9.770826203025533e-05, "loss": 1.0835, "step": 2425 }, { "epoch": 0.1643742801002778, "grad_norm": 9.701393127441406, "learning_rate": 9.770689301115751e-05, "loss": 1.0717, "step": 2426 }, { "epoch": 0.1644420353682499, "grad_norm": 7.887824058532715, "learning_rate": 9.770552399205969e-05, "loss": 1.0937, "step": 2427 }, { "epoch": 0.16450979063622195, "grad_norm": 10.94339370727539, "learning_rate": 9.770415497296188e-05, "loss": 1.2766, "step": 2428 }, { "epoch": 0.16457754590419404, "grad_norm": 10.051490783691406, "learning_rate": 9.770278595386406e-05, "loss": 1.248, "step": 2429 }, { "epoch": 0.16464530117216614, "grad_norm": 8.380006790161133, "learning_rate": 9.770141693476624e-05, "loss": 1.1657, "step": 2430 }, { "epoch": 0.16471305644013823, "grad_norm": 8.077753067016602, "learning_rate": 9.770004791566843e-05, "loss": 0.9511, "step": 2431 }, { "epoch": 0.1647808117081103, "grad_norm": 8.744999885559082, "learning_rate": 9.769867889657062e-05, "loss": 1.0736, "step": 2432 }, { "epoch": 0.16484856697608238, "grad_norm": 8.203909873962402, "learning_rate": 9.76973098774728e-05, "loss": 1.208, "step": 2433 }, { "epoch": 0.16491632224405448, "grad_norm": 9.462398529052734, "learning_rate": 9.769594085837498e-05, "loss": 1.0011, "step": 2434 }, { "epoch": 0.16498407751202657, "grad_norm": 11.190359115600586, "learning_rate": 9.769457183927716e-05, "loss": 1.139, "step": 2435 }, { "epoch": 0.16505183277999866, "grad_norm": 10.454118728637695, "learning_rate": 9.769320282017934e-05, "loss": 1.0799, "step": 2436 }, { "epoch": 0.16511958804797072, "grad_norm": 14.411054611206055, "learning_rate": 9.769183380108153e-05, "loss": 1.0369, "step": 2437 }, { "epoch": 0.16518734331594281, "grad_norm": 11.42679214477539, "learning_rate": 9.769046478198371e-05, "loss": 1.2486, "step": 2438 }, { "epoch": 0.1652550985839149, "grad_norm": 10.520325660705566, "learning_rate": 9.76890957628859e-05, "loss": 1.2355, "step": 2439 }, { "epoch": 0.165322853851887, "grad_norm": 7.958998680114746, "learning_rate": 9.768772674378808e-05, "loss": 1.0822, "step": 2440 }, { "epoch": 0.16539060911985906, "grad_norm": 8.649806022644043, "learning_rate": 9.768635772469027e-05, "loss": 0.9439, "step": 2441 }, { "epoch": 0.16545836438783115, "grad_norm": 8.539712905883789, "learning_rate": 9.768498870559245e-05, "loss": 0.9541, "step": 2442 }, { "epoch": 0.16552611965580324, "grad_norm": 11.40131950378418, "learning_rate": 9.768361968649463e-05, "loss": 1.0568, "step": 2443 }, { "epoch": 0.16559387492377534, "grad_norm": 8.333579063415527, "learning_rate": 9.768225066739681e-05, "loss": 0.924, "step": 2444 }, { "epoch": 0.1656616301917474, "grad_norm": 9.02564525604248, "learning_rate": 9.768088164829899e-05, "loss": 1.1901, "step": 2445 }, { "epoch": 0.1657293854597195, "grad_norm": 9.721306800842285, "learning_rate": 9.767951262920118e-05, "loss": 1.2313, "step": 2446 }, { "epoch": 0.16579714072769158, "grad_norm": 9.876781463623047, "learning_rate": 9.767814361010336e-05, "loss": 1.0165, "step": 2447 }, { "epoch": 0.16586489599566367, "grad_norm": 11.69865894317627, "learning_rate": 9.767677459100555e-05, "loss": 1.1301, "step": 2448 }, { "epoch": 0.16593265126363574, "grad_norm": 11.38391399383545, "learning_rate": 9.767540557190773e-05, "loss": 1.1522, "step": 2449 }, { "epoch": 0.16600040653160783, "grad_norm": 9.18020248413086, "learning_rate": 9.76740365528099e-05, "loss": 1.0515, "step": 2450 }, { "epoch": 0.16606816179957992, "grad_norm": 9.30802059173584, "learning_rate": 9.76726675337121e-05, "loss": 1.0789, "step": 2451 }, { "epoch": 0.166135917067552, "grad_norm": 9.58259391784668, "learning_rate": 9.767129851461428e-05, "loss": 1.1299, "step": 2452 }, { "epoch": 0.16620367233552408, "grad_norm": 9.137594223022461, "learning_rate": 9.766992949551646e-05, "loss": 1.0854, "step": 2453 }, { "epoch": 0.16627142760349617, "grad_norm": 12.410299301147461, "learning_rate": 9.766856047641864e-05, "loss": 1.1717, "step": 2454 }, { "epoch": 0.16633918287146826, "grad_norm": 9.016322135925293, "learning_rate": 9.766719145732083e-05, "loss": 1.1028, "step": 2455 }, { "epoch": 0.16640693813944035, "grad_norm": 10.166184425354004, "learning_rate": 9.766582243822301e-05, "loss": 1.0626, "step": 2456 }, { "epoch": 0.16647469340741242, "grad_norm": 9.030965805053711, "learning_rate": 9.76644534191252e-05, "loss": 1.1761, "step": 2457 }, { "epoch": 0.1665424486753845, "grad_norm": 9.294576644897461, "learning_rate": 9.766308440002739e-05, "loss": 0.943, "step": 2458 }, { "epoch": 0.1666102039433566, "grad_norm": 6.732856273651123, "learning_rate": 9.766171538092957e-05, "loss": 0.7342, "step": 2459 }, { "epoch": 0.1666779592113287, "grad_norm": 8.178942680358887, "learning_rate": 9.766034636183175e-05, "loss": 0.7889, "step": 2460 }, { "epoch": 0.16674571447930076, "grad_norm": 7.771929740905762, "learning_rate": 9.765897734273394e-05, "loss": 1.1129, "step": 2461 }, { "epoch": 0.16681346974727285, "grad_norm": 8.695874214172363, "learning_rate": 9.765760832363612e-05, "loss": 1.1968, "step": 2462 }, { "epoch": 0.16688122501524494, "grad_norm": 10.18800163269043, "learning_rate": 9.76562393045383e-05, "loss": 1.0835, "step": 2463 }, { "epoch": 0.16694898028321703, "grad_norm": 9.310625076293945, "learning_rate": 9.76548702854405e-05, "loss": 1.2042, "step": 2464 }, { "epoch": 0.1670167355511891, "grad_norm": 11.684195518493652, "learning_rate": 9.765350126634268e-05, "loss": 1.2518, "step": 2465 }, { "epoch": 0.16708449081916119, "grad_norm": 10.9056978225708, "learning_rate": 9.765213224724486e-05, "loss": 1.2465, "step": 2466 }, { "epoch": 0.16715224608713328, "grad_norm": 9.871830940246582, "learning_rate": 9.765076322814704e-05, "loss": 1.3192, "step": 2467 }, { "epoch": 0.16722000135510537, "grad_norm": 7.725397109985352, "learning_rate": 9.764939420904922e-05, "loss": 0.9082, "step": 2468 }, { "epoch": 0.16728775662307743, "grad_norm": 8.7415189743042, "learning_rate": 9.764802518995141e-05, "loss": 0.8569, "step": 2469 }, { "epoch": 0.16735551189104952, "grad_norm": 7.8990888595581055, "learning_rate": 9.76466561708536e-05, "loss": 1.0175, "step": 2470 }, { "epoch": 0.16742326715902162, "grad_norm": 8.688074111938477, "learning_rate": 9.764528715175577e-05, "loss": 1.0131, "step": 2471 }, { "epoch": 0.1674910224269937, "grad_norm": 7.902133941650391, "learning_rate": 9.764391813265795e-05, "loss": 0.8956, "step": 2472 }, { "epoch": 0.16755877769496577, "grad_norm": 9.179505348205566, "learning_rate": 9.764254911356013e-05, "loss": 1.175, "step": 2473 }, { "epoch": 0.16762653296293786, "grad_norm": 10.701058387756348, "learning_rate": 9.764118009446233e-05, "loss": 1.003, "step": 2474 }, { "epoch": 0.16769428823090995, "grad_norm": 8.76916217803955, "learning_rate": 9.763981107536451e-05, "loss": 1.102, "step": 2475 }, { "epoch": 0.16776204349888205, "grad_norm": 8.682199478149414, "learning_rate": 9.763844205626669e-05, "loss": 1.2042, "step": 2476 }, { "epoch": 0.1678297987668541, "grad_norm": 9.789544105529785, "learning_rate": 9.763707303716887e-05, "loss": 1.2469, "step": 2477 }, { "epoch": 0.1678975540348262, "grad_norm": 8.250391960144043, "learning_rate": 9.763570401807106e-05, "loss": 0.9742, "step": 2478 }, { "epoch": 0.1679653093027983, "grad_norm": 8.938610076904297, "learning_rate": 9.763433499897324e-05, "loss": 0.9968, "step": 2479 }, { "epoch": 0.16803306457077039, "grad_norm": 9.956622123718262, "learning_rate": 9.763296597987542e-05, "loss": 1.0464, "step": 2480 }, { "epoch": 0.16810081983874245, "grad_norm": 8.582858085632324, "learning_rate": 9.76315969607776e-05, "loss": 1.1915, "step": 2481 }, { "epoch": 0.16816857510671454, "grad_norm": 7.73312520980835, "learning_rate": 9.763022794167979e-05, "loss": 0.8506, "step": 2482 }, { "epoch": 0.16823633037468663, "grad_norm": 8.986891746520996, "learning_rate": 9.762885892258198e-05, "loss": 0.8959, "step": 2483 }, { "epoch": 0.16830408564265872, "grad_norm": 10.047099113464355, "learning_rate": 9.762748990348416e-05, "loss": 1.24, "step": 2484 }, { "epoch": 0.1683718409106308, "grad_norm": 8.517911911010742, "learning_rate": 9.762612088438634e-05, "loss": 0.8171, "step": 2485 }, { "epoch": 0.16843959617860288, "grad_norm": 9.586174964904785, "learning_rate": 9.762475186528852e-05, "loss": 1.1603, "step": 2486 }, { "epoch": 0.16850735144657497, "grad_norm": 9.85086441040039, "learning_rate": 9.762338284619071e-05, "loss": 0.9157, "step": 2487 }, { "epoch": 0.16857510671454706, "grad_norm": 12.023639678955078, "learning_rate": 9.76220138270929e-05, "loss": 1.2032, "step": 2488 }, { "epoch": 0.16864286198251915, "grad_norm": 8.932641983032227, "learning_rate": 9.762064480799507e-05, "loss": 1.0678, "step": 2489 }, { "epoch": 0.16871061725049122, "grad_norm": 10.568282127380371, "learning_rate": 9.761927578889725e-05, "loss": 1.3213, "step": 2490 }, { "epoch": 0.1687783725184633, "grad_norm": 9.116564750671387, "learning_rate": 9.761790676979944e-05, "loss": 1.0525, "step": 2491 }, { "epoch": 0.1688461277864354, "grad_norm": 8.192644119262695, "learning_rate": 9.761653775070163e-05, "loss": 0.9554, "step": 2492 }, { "epoch": 0.1689138830544075, "grad_norm": 9.146562576293945, "learning_rate": 9.761516873160381e-05, "loss": 0.8852, "step": 2493 }, { "epoch": 0.16898163832237956, "grad_norm": 8.82610034942627, "learning_rate": 9.761379971250599e-05, "loss": 1.0521, "step": 2494 }, { "epoch": 0.16904939359035165, "grad_norm": 9.051412582397461, "learning_rate": 9.761243069340817e-05, "loss": 0.913, "step": 2495 }, { "epoch": 0.16911714885832374, "grad_norm": 8.97696304321289, "learning_rate": 9.761106167431036e-05, "loss": 0.887, "step": 2496 }, { "epoch": 0.16918490412629583, "grad_norm": 10.083110809326172, "learning_rate": 9.760969265521254e-05, "loss": 1.3521, "step": 2497 }, { "epoch": 0.1692526593942679, "grad_norm": 7.585256099700928, "learning_rate": 9.760832363611472e-05, "loss": 0.9001, "step": 2498 }, { "epoch": 0.16932041466224, "grad_norm": 10.301995277404785, "learning_rate": 9.76069546170169e-05, "loss": 1.0915, "step": 2499 }, { "epoch": 0.16938816993021208, "grad_norm": 8.517580032348633, "learning_rate": 9.760558559791909e-05, "loss": 1.0097, "step": 2500 }, { "epoch": 0.16945592519818417, "grad_norm": 8.20002555847168, "learning_rate": 9.760421657882128e-05, "loss": 1.2988, "step": 2501 }, { "epoch": 0.16952368046615623, "grad_norm": 9.705124855041504, "learning_rate": 9.760284755972346e-05, "loss": 1.2661, "step": 2502 }, { "epoch": 0.16959143573412833, "grad_norm": 9.167060852050781, "learning_rate": 9.760147854062564e-05, "loss": 1.0556, "step": 2503 }, { "epoch": 0.16965919100210042, "grad_norm": 8.415916442871094, "learning_rate": 9.760010952152783e-05, "loss": 0.7964, "step": 2504 }, { "epoch": 0.1697269462700725, "grad_norm": 7.626298427581787, "learning_rate": 9.759874050243001e-05, "loss": 0.9153, "step": 2505 }, { "epoch": 0.16979470153804457, "grad_norm": 10.595001220703125, "learning_rate": 9.75973714833322e-05, "loss": 0.9911, "step": 2506 }, { "epoch": 0.16986245680601667, "grad_norm": 7.564423084259033, "learning_rate": 9.759600246423439e-05, "loss": 0.9843, "step": 2507 }, { "epoch": 0.16993021207398876, "grad_norm": 7.644829273223877, "learning_rate": 9.759463344513657e-05, "loss": 0.9116, "step": 2508 }, { "epoch": 0.16999796734196085, "grad_norm": 7.54351282119751, "learning_rate": 9.759326442603875e-05, "loss": 1.1528, "step": 2509 }, { "epoch": 0.1700657226099329, "grad_norm": 9.259818077087402, "learning_rate": 9.759189540694094e-05, "loss": 1.0867, "step": 2510 }, { "epoch": 0.170133477877905, "grad_norm": 8.022993087768555, "learning_rate": 9.759052638784312e-05, "loss": 0.7524, "step": 2511 }, { "epoch": 0.1702012331458771, "grad_norm": 8.264616966247559, "learning_rate": 9.75891573687453e-05, "loss": 0.9906, "step": 2512 }, { "epoch": 0.1702689884138492, "grad_norm": 9.606420516967773, "learning_rate": 9.758778834964748e-05, "loss": 0.9287, "step": 2513 }, { "epoch": 0.17033674368182125, "grad_norm": 8.019355773925781, "learning_rate": 9.758641933054966e-05, "loss": 0.879, "step": 2514 }, { "epoch": 0.17040449894979334, "grad_norm": 9.819777488708496, "learning_rate": 9.758505031145186e-05, "loss": 1.3677, "step": 2515 }, { "epoch": 0.17047225421776543, "grad_norm": 9.9795560836792, "learning_rate": 9.758368129235404e-05, "loss": 1.1614, "step": 2516 }, { "epoch": 0.17054000948573753, "grad_norm": 9.51271915435791, "learning_rate": 9.758231227325622e-05, "loss": 1.0423, "step": 2517 }, { "epoch": 0.1706077647537096, "grad_norm": 10.511359214782715, "learning_rate": 9.75809432541584e-05, "loss": 0.9492, "step": 2518 }, { "epoch": 0.17067552002168168, "grad_norm": 9.61755657196045, "learning_rate": 9.757957423506059e-05, "loss": 1.1481, "step": 2519 }, { "epoch": 0.17074327528965377, "grad_norm": 11.246142387390137, "learning_rate": 9.757820521596277e-05, "loss": 1.3981, "step": 2520 }, { "epoch": 0.17081103055762586, "grad_norm": 9.273181915283203, "learning_rate": 9.757683619686495e-05, "loss": 0.9773, "step": 2521 }, { "epoch": 0.17087878582559793, "grad_norm": 11.215822219848633, "learning_rate": 9.757546717776713e-05, "loss": 1.2303, "step": 2522 }, { "epoch": 0.17094654109357002, "grad_norm": 10.139853477478027, "learning_rate": 9.757409815866931e-05, "loss": 1.0807, "step": 2523 }, { "epoch": 0.1710142963615421, "grad_norm": 13.275606155395508, "learning_rate": 9.757272913957151e-05, "loss": 1.3326, "step": 2524 }, { "epoch": 0.1710820516295142, "grad_norm": 8.32109260559082, "learning_rate": 9.757136012047369e-05, "loss": 0.9428, "step": 2525 }, { "epoch": 0.17114980689748627, "grad_norm": 9.275816917419434, "learning_rate": 9.756999110137587e-05, "loss": 1.1531, "step": 2526 }, { "epoch": 0.17121756216545836, "grad_norm": 10.61928939819336, "learning_rate": 9.756862208227805e-05, "loss": 0.9422, "step": 2527 }, { "epoch": 0.17128531743343045, "grad_norm": 7.793631553649902, "learning_rate": 9.756725306318023e-05, "loss": 1.216, "step": 2528 }, { "epoch": 0.17135307270140254, "grad_norm": 7.453477382659912, "learning_rate": 9.756588404408242e-05, "loss": 0.7787, "step": 2529 }, { "epoch": 0.1714208279693746, "grad_norm": 10.39784049987793, "learning_rate": 9.75645150249846e-05, "loss": 1.3171, "step": 2530 }, { "epoch": 0.1714885832373467, "grad_norm": 8.57040786743164, "learning_rate": 9.756314600588678e-05, "loss": 0.8319, "step": 2531 }, { "epoch": 0.1715563385053188, "grad_norm": 8.696785926818848, "learning_rate": 9.756177698678896e-05, "loss": 1.0078, "step": 2532 }, { "epoch": 0.17162409377329088, "grad_norm": 8.212169647216797, "learning_rate": 9.756040796769116e-05, "loss": 0.7851, "step": 2533 }, { "epoch": 0.17169184904126294, "grad_norm": 10.94201946258545, "learning_rate": 9.755903894859334e-05, "loss": 1.0065, "step": 2534 }, { "epoch": 0.17175960430923504, "grad_norm": 12.041540145874023, "learning_rate": 9.755766992949552e-05, "loss": 1.2938, "step": 2535 }, { "epoch": 0.17182735957720713, "grad_norm": 9.289467811584473, "learning_rate": 9.75563009103977e-05, "loss": 1.1521, "step": 2536 }, { "epoch": 0.17189511484517922, "grad_norm": 9.231005668640137, "learning_rate": 9.755493189129988e-05, "loss": 0.9, "step": 2537 }, { "epoch": 0.17196287011315128, "grad_norm": 8.934699058532715, "learning_rate": 9.755356287220207e-05, "loss": 1.2817, "step": 2538 }, { "epoch": 0.17203062538112338, "grad_norm": 10.096917152404785, "learning_rate": 9.755219385310425e-05, "loss": 1.1587, "step": 2539 }, { "epoch": 0.17209838064909547, "grad_norm": 8.864645004272461, "learning_rate": 9.755082483400643e-05, "loss": 1.1781, "step": 2540 }, { "epoch": 0.17216613591706756, "grad_norm": 10.964715957641602, "learning_rate": 9.754945581490861e-05, "loss": 1.2729, "step": 2541 }, { "epoch": 0.17223389118503965, "grad_norm": 10.845703125, "learning_rate": 9.754808679581081e-05, "loss": 1.4475, "step": 2542 }, { "epoch": 0.17230164645301171, "grad_norm": 9.800530433654785, "learning_rate": 9.754671777671299e-05, "loss": 1.0296, "step": 2543 }, { "epoch": 0.1723694017209838, "grad_norm": 8.789834022521973, "learning_rate": 9.754534875761517e-05, "loss": 0.9573, "step": 2544 }, { "epoch": 0.1724371569889559, "grad_norm": 7.731616973876953, "learning_rate": 9.754397973851735e-05, "loss": 0.9875, "step": 2545 }, { "epoch": 0.172504912256928, "grad_norm": 8.942553520202637, "learning_rate": 9.754261071941953e-05, "loss": 1.1897, "step": 2546 }, { "epoch": 0.17257266752490005, "grad_norm": 7.468856334686279, "learning_rate": 9.754124170032172e-05, "loss": 0.8869, "step": 2547 }, { "epoch": 0.17264042279287214, "grad_norm": 8.798864364624023, "learning_rate": 9.75398726812239e-05, "loss": 1.0034, "step": 2548 }, { "epoch": 0.17270817806084424, "grad_norm": 8.579094886779785, "learning_rate": 9.753850366212608e-05, "loss": 1.0211, "step": 2549 }, { "epoch": 0.17277593332881633, "grad_norm": 10.90807819366455, "learning_rate": 9.753713464302827e-05, "loss": 1.3639, "step": 2550 }, { "epoch": 0.1728436885967884, "grad_norm": 8.337204933166504, "learning_rate": 9.753576562393046e-05, "loss": 1.323, "step": 2551 }, { "epoch": 0.17291144386476048, "grad_norm": 8.731851577758789, "learning_rate": 9.753439660483264e-05, "loss": 0.8805, "step": 2552 }, { "epoch": 0.17297919913273258, "grad_norm": 9.541427612304688, "learning_rate": 9.753302758573482e-05, "loss": 1.0368, "step": 2553 }, { "epoch": 0.17304695440070467, "grad_norm": 7.938154697418213, "learning_rate": 9.753165856663701e-05, "loss": 0.9128, "step": 2554 }, { "epoch": 0.17311470966867673, "grad_norm": 8.111212730407715, "learning_rate": 9.75302895475392e-05, "loss": 0.9857, "step": 2555 }, { "epoch": 0.17318246493664882, "grad_norm": 8.583324432373047, "learning_rate": 9.752892052844139e-05, "loss": 1.0975, "step": 2556 }, { "epoch": 0.1732502202046209, "grad_norm": 8.180643081665039, "learning_rate": 9.752755150934357e-05, "loss": 1.21, "step": 2557 }, { "epoch": 0.173317975472593, "grad_norm": 9.740217208862305, "learning_rate": 9.752618249024575e-05, "loss": 1.18, "step": 2558 }, { "epoch": 0.17338573074056507, "grad_norm": 8.725831031799316, "learning_rate": 9.752481347114793e-05, "loss": 0.9816, "step": 2559 }, { "epoch": 0.17345348600853716, "grad_norm": 10.786824226379395, "learning_rate": 9.752344445205011e-05, "loss": 1.0402, "step": 2560 }, { "epoch": 0.17352124127650925, "grad_norm": 8.91720962524414, "learning_rate": 9.75220754329523e-05, "loss": 1.2357, "step": 2561 }, { "epoch": 0.17358899654448134, "grad_norm": 9.02492618560791, "learning_rate": 9.752070641385448e-05, "loss": 1.1264, "step": 2562 }, { "epoch": 0.1736567518124534, "grad_norm": 7.917794227600098, "learning_rate": 9.751933739475666e-05, "loss": 1.0937, "step": 2563 }, { "epoch": 0.1737245070804255, "grad_norm": 11.543112754821777, "learning_rate": 9.751796837565884e-05, "loss": 1.0851, "step": 2564 }, { "epoch": 0.1737922623483976, "grad_norm": 9.114936828613281, "learning_rate": 9.751659935656104e-05, "loss": 0.9565, "step": 2565 }, { "epoch": 0.17386001761636968, "grad_norm": 12.322575569152832, "learning_rate": 9.751523033746322e-05, "loss": 1.0356, "step": 2566 }, { "epoch": 0.17392777288434175, "grad_norm": 11.39137077331543, "learning_rate": 9.75138613183654e-05, "loss": 1.2141, "step": 2567 }, { "epoch": 0.17399552815231384, "grad_norm": 9.505644798278809, "learning_rate": 9.751249229926758e-05, "loss": 0.8307, "step": 2568 }, { "epoch": 0.17406328342028593, "grad_norm": 11.05932331085205, "learning_rate": 9.751112328016976e-05, "loss": 1.1915, "step": 2569 }, { "epoch": 0.17413103868825802, "grad_norm": 11.104498863220215, "learning_rate": 9.750975426107195e-05, "loss": 1.0032, "step": 2570 }, { "epoch": 0.17419879395623009, "grad_norm": 11.366668701171875, "learning_rate": 9.750838524197413e-05, "loss": 1.0857, "step": 2571 }, { "epoch": 0.17426654922420218, "grad_norm": 8.775167465209961, "learning_rate": 9.750701622287631e-05, "loss": 1.1545, "step": 2572 }, { "epoch": 0.17433430449217427, "grad_norm": 8.820073127746582, "learning_rate": 9.75056472037785e-05, "loss": 1.0528, "step": 2573 }, { "epoch": 0.17440205976014636, "grad_norm": 9.53591537475586, "learning_rate": 9.750427818468069e-05, "loss": 1.0562, "step": 2574 }, { "epoch": 0.17446981502811842, "grad_norm": 10.08950138092041, "learning_rate": 9.750290916558287e-05, "loss": 1.0456, "step": 2575 }, { "epoch": 0.17453757029609052, "grad_norm": 8.222607612609863, "learning_rate": 9.750154014648505e-05, "loss": 0.9168, "step": 2576 }, { "epoch": 0.1746053255640626, "grad_norm": 10.067093849182129, "learning_rate": 9.750017112738723e-05, "loss": 0.9952, "step": 2577 }, { "epoch": 0.1746730808320347, "grad_norm": 9.726996421813965, "learning_rate": 9.749880210828941e-05, "loss": 0.9428, "step": 2578 }, { "epoch": 0.17474083610000676, "grad_norm": 9.970647811889648, "learning_rate": 9.74974330891916e-05, "loss": 1.1838, "step": 2579 }, { "epoch": 0.17480859136797886, "grad_norm": 7.8667893409729, "learning_rate": 9.749606407009378e-05, "loss": 1.0341, "step": 2580 }, { "epoch": 0.17487634663595095, "grad_norm": 9.227079391479492, "learning_rate": 9.749469505099596e-05, "loss": 1.0515, "step": 2581 }, { "epoch": 0.17494410190392304, "grad_norm": 9.198224067687988, "learning_rate": 9.749332603189814e-05, "loss": 1.0935, "step": 2582 }, { "epoch": 0.1750118571718951, "grad_norm": 11.105298042297363, "learning_rate": 9.749195701280032e-05, "loss": 1.2115, "step": 2583 }, { "epoch": 0.1750796124398672, "grad_norm": 8.815799713134766, "learning_rate": 9.749058799370252e-05, "loss": 1.1308, "step": 2584 }, { "epoch": 0.17514736770783929, "grad_norm": 10.571581840515137, "learning_rate": 9.74892189746047e-05, "loss": 1.1982, "step": 2585 }, { "epoch": 0.17521512297581138, "grad_norm": 7.368075370788574, "learning_rate": 9.748784995550688e-05, "loss": 0.9352, "step": 2586 }, { "epoch": 0.17528287824378344, "grad_norm": 7.103427410125732, "learning_rate": 9.748648093640906e-05, "loss": 1.0578, "step": 2587 }, { "epoch": 0.17535063351175553, "grad_norm": 7.678786754608154, "learning_rate": 9.748511191731125e-05, "loss": 0.8549, "step": 2588 }, { "epoch": 0.17541838877972762, "grad_norm": 9.514643669128418, "learning_rate": 9.748374289821343e-05, "loss": 1.1835, "step": 2589 }, { "epoch": 0.17548614404769972, "grad_norm": 7.552379131317139, "learning_rate": 9.748237387911561e-05, "loss": 0.9425, "step": 2590 }, { "epoch": 0.17555389931567178, "grad_norm": 7.663018226623535, "learning_rate": 9.74810048600178e-05, "loss": 1.0948, "step": 2591 }, { "epoch": 0.17562165458364387, "grad_norm": 10.293536186218262, "learning_rate": 9.747963584091997e-05, "loss": 1.1636, "step": 2592 }, { "epoch": 0.17568940985161596, "grad_norm": 9.024083137512207, "learning_rate": 9.747826682182217e-05, "loss": 1.1157, "step": 2593 }, { "epoch": 0.17575716511958805, "grad_norm": 8.870935440063477, "learning_rate": 9.747689780272435e-05, "loss": 0.9634, "step": 2594 }, { "epoch": 0.17582492038756015, "grad_norm": 8.553323745727539, "learning_rate": 9.747552878362653e-05, "loss": 1.1515, "step": 2595 }, { "epoch": 0.1758926756555322, "grad_norm": 9.13661003112793, "learning_rate": 9.747415976452871e-05, "loss": 0.9267, "step": 2596 }, { "epoch": 0.1759604309235043, "grad_norm": 8.66097640991211, "learning_rate": 9.74727907454309e-05, "loss": 0.6839, "step": 2597 }, { "epoch": 0.1760281861914764, "grad_norm": 10.07618236541748, "learning_rate": 9.747142172633308e-05, "loss": 1.2822, "step": 2598 }, { "epoch": 0.17609594145944849, "grad_norm": 9.136283874511719, "learning_rate": 9.747005270723526e-05, "loss": 1.0053, "step": 2599 }, { "epoch": 0.17616369672742055, "grad_norm": 9.42113208770752, "learning_rate": 9.746868368813746e-05, "loss": 1.15, "step": 2600 }, { "epoch": 0.17623145199539264, "grad_norm": 8.789713859558105, "learning_rate": 9.746731466903964e-05, "loss": 0.8504, "step": 2601 }, { "epoch": 0.17629920726336473, "grad_norm": 8.704665184020996, "learning_rate": 9.746594564994182e-05, "loss": 1.1516, "step": 2602 }, { "epoch": 0.17636696253133682, "grad_norm": 8.030630111694336, "learning_rate": 9.746457663084401e-05, "loss": 1.044, "step": 2603 }, { "epoch": 0.1764347177993089, "grad_norm": 8.640777587890625, "learning_rate": 9.746320761174619e-05, "loss": 1.0013, "step": 2604 }, { "epoch": 0.17650247306728098, "grad_norm": 7.806771278381348, "learning_rate": 9.746183859264837e-05, "loss": 1.0678, "step": 2605 }, { "epoch": 0.17657022833525307, "grad_norm": 8.883776664733887, "learning_rate": 9.746046957355055e-05, "loss": 0.9277, "step": 2606 }, { "epoch": 0.17663798360322516, "grad_norm": 7.539346694946289, "learning_rate": 9.745910055445275e-05, "loss": 1.0437, "step": 2607 }, { "epoch": 0.17670573887119723, "grad_norm": 9.873644828796387, "learning_rate": 9.745773153535493e-05, "loss": 1.206, "step": 2608 }, { "epoch": 0.17677349413916932, "grad_norm": 10.031026840209961, "learning_rate": 9.745636251625711e-05, "loss": 1.1934, "step": 2609 }, { "epoch": 0.1768412494071414, "grad_norm": 9.194849014282227, "learning_rate": 9.745499349715929e-05, "loss": 1.2365, "step": 2610 }, { "epoch": 0.1769090046751135, "grad_norm": 7.951476097106934, "learning_rate": 9.745362447806148e-05, "loss": 0.8829, "step": 2611 }, { "epoch": 0.17697675994308557, "grad_norm": 8.77597427368164, "learning_rate": 9.745225545896366e-05, "loss": 1.0097, "step": 2612 }, { "epoch": 0.17704451521105766, "grad_norm": 10.356508255004883, "learning_rate": 9.745088643986584e-05, "loss": 1.2603, "step": 2613 }, { "epoch": 0.17711227047902975, "grad_norm": 9.24199390411377, "learning_rate": 9.744951742076802e-05, "loss": 1.0181, "step": 2614 }, { "epoch": 0.17718002574700184, "grad_norm": 8.244451522827148, "learning_rate": 9.74481484016702e-05, "loss": 0.9602, "step": 2615 }, { "epoch": 0.1772477810149739, "grad_norm": 10.769842147827148, "learning_rate": 9.74467793825724e-05, "loss": 1.0646, "step": 2616 }, { "epoch": 0.177315536282946, "grad_norm": 9.381787300109863, "learning_rate": 9.744541036347458e-05, "loss": 0.8985, "step": 2617 }, { "epoch": 0.1773832915509181, "grad_norm": 8.379899978637695, "learning_rate": 9.744404134437676e-05, "loss": 1.0368, "step": 2618 }, { "epoch": 0.17745104681889018, "grad_norm": 9.59954833984375, "learning_rate": 9.744267232527894e-05, "loss": 1.1447, "step": 2619 }, { "epoch": 0.17751880208686224, "grad_norm": 8.83703327178955, "learning_rate": 9.744130330618113e-05, "loss": 1.3435, "step": 2620 }, { "epoch": 0.17758655735483433, "grad_norm": 9.448990821838379, "learning_rate": 9.743993428708331e-05, "loss": 1.1837, "step": 2621 }, { "epoch": 0.17765431262280643, "grad_norm": 9.466961860656738, "learning_rate": 9.74385652679855e-05, "loss": 1.1265, "step": 2622 }, { "epoch": 0.17772206789077852, "grad_norm": 8.68281364440918, "learning_rate": 9.743719624888767e-05, "loss": 0.8804, "step": 2623 }, { "epoch": 0.17778982315875058, "grad_norm": 7.007611274719238, "learning_rate": 9.743582722978985e-05, "loss": 1.03, "step": 2624 }, { "epoch": 0.17785757842672267, "grad_norm": 8.254279136657715, "learning_rate": 9.743445821069205e-05, "loss": 1.0328, "step": 2625 }, { "epoch": 0.17792533369469477, "grad_norm": 9.134196281433105, "learning_rate": 9.743308919159423e-05, "loss": 0.9671, "step": 2626 }, { "epoch": 0.17799308896266686, "grad_norm": 7.803997039794922, "learning_rate": 9.743172017249641e-05, "loss": 0.9692, "step": 2627 }, { "epoch": 0.17806084423063892, "grad_norm": 8.37303638458252, "learning_rate": 9.743035115339859e-05, "loss": 0.8949, "step": 2628 }, { "epoch": 0.178128599498611, "grad_norm": 9.928305625915527, "learning_rate": 9.742898213430078e-05, "loss": 1.2679, "step": 2629 }, { "epoch": 0.1781963547665831, "grad_norm": 8.58604621887207, "learning_rate": 9.742761311520296e-05, "loss": 1.23, "step": 2630 }, { "epoch": 0.1782641100345552, "grad_norm": 9.47903823852539, "learning_rate": 9.742624409610514e-05, "loss": 0.8701, "step": 2631 }, { "epoch": 0.17833186530252726, "grad_norm": 8.013197898864746, "learning_rate": 9.742487507700732e-05, "loss": 0.8325, "step": 2632 }, { "epoch": 0.17839962057049935, "grad_norm": 9.541396141052246, "learning_rate": 9.74235060579095e-05, "loss": 1.0996, "step": 2633 }, { "epoch": 0.17846737583847144, "grad_norm": 7.515398979187012, "learning_rate": 9.74221370388117e-05, "loss": 1.1174, "step": 2634 }, { "epoch": 0.17853513110644353, "grad_norm": 9.562670707702637, "learning_rate": 9.742076801971388e-05, "loss": 0.9727, "step": 2635 }, { "epoch": 0.1786028863744156, "grad_norm": 10.072449684143066, "learning_rate": 9.741939900061606e-05, "loss": 1.0021, "step": 2636 }, { "epoch": 0.1786706416423877, "grad_norm": 9.634955406188965, "learning_rate": 9.741802998151824e-05, "loss": 1.3998, "step": 2637 }, { "epoch": 0.17873839691035978, "grad_norm": 8.78954792022705, "learning_rate": 9.741666096242042e-05, "loss": 0.9602, "step": 2638 }, { "epoch": 0.17880615217833187, "grad_norm": 9.597916603088379, "learning_rate": 9.741529194332261e-05, "loss": 1.0174, "step": 2639 }, { "epoch": 0.17887390744630394, "grad_norm": 9.543583869934082, "learning_rate": 9.74139229242248e-05, "loss": 1.0686, "step": 2640 }, { "epoch": 0.17894166271427603, "grad_norm": 8.247551918029785, "learning_rate": 9.741255390512697e-05, "loss": 1.1026, "step": 2641 }, { "epoch": 0.17900941798224812, "grad_norm": 8.487943649291992, "learning_rate": 9.741118488602915e-05, "loss": 0.8843, "step": 2642 }, { "epoch": 0.1790771732502202, "grad_norm": 9.061832427978516, "learning_rate": 9.740981586693135e-05, "loss": 0.9708, "step": 2643 }, { "epoch": 0.17914492851819228, "grad_norm": 9.242405891418457, "learning_rate": 9.740844684783353e-05, "loss": 1.1344, "step": 2644 }, { "epoch": 0.17921268378616437, "grad_norm": 8.192344665527344, "learning_rate": 9.740707782873571e-05, "loss": 0.7587, "step": 2645 }, { "epoch": 0.17928043905413646, "grad_norm": 6.6164445877075195, "learning_rate": 9.74057088096379e-05, "loss": 0.842, "step": 2646 }, { "epoch": 0.17934819432210855, "grad_norm": 10.285326957702637, "learning_rate": 9.740433979054008e-05, "loss": 1.138, "step": 2647 }, { "epoch": 0.17941594959008064, "grad_norm": 8.55659294128418, "learning_rate": 9.740297077144226e-05, "loss": 0.7925, "step": 2648 }, { "epoch": 0.1794837048580527, "grad_norm": 8.384642601013184, "learning_rate": 9.740160175234446e-05, "loss": 0.9719, "step": 2649 }, { "epoch": 0.1795514601260248, "grad_norm": 9.303678512573242, "learning_rate": 9.740023273324664e-05, "loss": 1.0688, "step": 2650 }, { "epoch": 0.1796192153939969, "grad_norm": 9.642799377441406, "learning_rate": 9.739886371414882e-05, "loss": 0.9758, "step": 2651 }, { "epoch": 0.17968697066196898, "grad_norm": 9.616509437561035, "learning_rate": 9.739749469505101e-05, "loss": 0.9999, "step": 2652 }, { "epoch": 0.17975472592994104, "grad_norm": 10.511337280273438, "learning_rate": 9.739612567595319e-05, "loss": 1.0079, "step": 2653 }, { "epoch": 0.17982248119791314, "grad_norm": 7.480882167816162, "learning_rate": 9.739475665685537e-05, "loss": 0.8353, "step": 2654 }, { "epoch": 0.17989023646588523, "grad_norm": 10.277608871459961, "learning_rate": 9.739338763775755e-05, "loss": 0.9709, "step": 2655 }, { "epoch": 0.17995799173385732, "grad_norm": 9.135882377624512, "learning_rate": 9.739201861865973e-05, "loss": 1.3888, "step": 2656 }, { "epoch": 0.18002574700182938, "grad_norm": 8.147950172424316, "learning_rate": 9.739064959956193e-05, "loss": 1.2235, "step": 2657 }, { "epoch": 0.18009350226980148, "grad_norm": 8.583501815795898, "learning_rate": 9.738928058046411e-05, "loss": 0.9123, "step": 2658 }, { "epoch": 0.18016125753777357, "grad_norm": 8.202300071716309, "learning_rate": 9.738791156136629e-05, "loss": 1.0435, "step": 2659 }, { "epoch": 0.18022901280574566, "grad_norm": 8.121417045593262, "learning_rate": 9.738654254226847e-05, "loss": 0.9351, "step": 2660 }, { "epoch": 0.18029676807371772, "grad_norm": 10.645029067993164, "learning_rate": 9.738517352317065e-05, "loss": 1.5219, "step": 2661 }, { "epoch": 0.18036452334168981, "grad_norm": 9.766422271728516, "learning_rate": 9.738380450407284e-05, "loss": 1.0499, "step": 2662 }, { "epoch": 0.1804322786096619, "grad_norm": 8.137174606323242, "learning_rate": 9.738243548497502e-05, "loss": 1.1871, "step": 2663 }, { "epoch": 0.180500033877634, "grad_norm": 8.422198295593262, "learning_rate": 9.73810664658772e-05, "loss": 0.9524, "step": 2664 }, { "epoch": 0.18056778914560606, "grad_norm": 6.721381664276123, "learning_rate": 9.737969744677938e-05, "loss": 1.0449, "step": 2665 }, { "epoch": 0.18063554441357815, "grad_norm": 9.175302505493164, "learning_rate": 9.737832842768158e-05, "loss": 1.0508, "step": 2666 }, { "epoch": 0.18070329968155024, "grad_norm": 10.165428161621094, "learning_rate": 9.737695940858376e-05, "loss": 0.8744, "step": 2667 }, { "epoch": 0.18077105494952234, "grad_norm": 9.396173477172852, "learning_rate": 9.737559038948594e-05, "loss": 1.0304, "step": 2668 }, { "epoch": 0.1808388102174944, "grad_norm": 9.680516242980957, "learning_rate": 9.737422137038812e-05, "loss": 0.9773, "step": 2669 }, { "epoch": 0.1809065654854665, "grad_norm": 7.570343017578125, "learning_rate": 9.73728523512903e-05, "loss": 0.9159, "step": 2670 }, { "epoch": 0.18097432075343858, "grad_norm": 9.5789213180542, "learning_rate": 9.737148333219249e-05, "loss": 1.1269, "step": 2671 }, { "epoch": 0.18104207602141068, "grad_norm": 8.559327125549316, "learning_rate": 9.737011431309467e-05, "loss": 1.2318, "step": 2672 }, { "epoch": 0.18110983128938274, "grad_norm": 7.630974292755127, "learning_rate": 9.736874529399685e-05, "loss": 0.8996, "step": 2673 }, { "epoch": 0.18117758655735483, "grad_norm": 8.078895568847656, "learning_rate": 9.736737627489903e-05, "loss": 0.9942, "step": 2674 }, { "epoch": 0.18124534182532692, "grad_norm": 8.830656051635742, "learning_rate": 9.736600725580123e-05, "loss": 1.0614, "step": 2675 }, { "epoch": 0.181313097093299, "grad_norm": 7.66297721862793, "learning_rate": 9.736463823670341e-05, "loss": 1.025, "step": 2676 }, { "epoch": 0.18138085236127108, "grad_norm": 8.318365097045898, "learning_rate": 9.736326921760559e-05, "loss": 1.1286, "step": 2677 }, { "epoch": 0.18144860762924317, "grad_norm": 10.180238723754883, "learning_rate": 9.736190019850777e-05, "loss": 1.1304, "step": 2678 }, { "epoch": 0.18151636289721526, "grad_norm": 9.2420654296875, "learning_rate": 9.736053117940995e-05, "loss": 1.157, "step": 2679 }, { "epoch": 0.18158411816518735, "grad_norm": 7.984904766082764, "learning_rate": 9.735916216031214e-05, "loss": 1.1325, "step": 2680 }, { "epoch": 0.18165187343315942, "grad_norm": 11.136275291442871, "learning_rate": 9.735779314121432e-05, "loss": 1.0151, "step": 2681 }, { "epoch": 0.1817196287011315, "grad_norm": 9.954483032226562, "learning_rate": 9.73564241221165e-05, "loss": 1.0934, "step": 2682 }, { "epoch": 0.1817873839691036, "grad_norm": 8.491388320922852, "learning_rate": 9.735505510301868e-05, "loss": 1.2288, "step": 2683 }, { "epoch": 0.1818551392370757, "grad_norm": 9.355586051940918, "learning_rate": 9.735368608392088e-05, "loss": 0.7999, "step": 2684 }, { "epoch": 0.18192289450504776, "grad_norm": 8.829736709594727, "learning_rate": 9.735231706482306e-05, "loss": 0.9711, "step": 2685 }, { "epoch": 0.18199064977301985, "grad_norm": 9.667959213256836, "learning_rate": 9.735094804572524e-05, "loss": 1.1132, "step": 2686 }, { "epoch": 0.18205840504099194, "grad_norm": 6.971467971801758, "learning_rate": 9.734957902662742e-05, "loss": 1.0252, "step": 2687 }, { "epoch": 0.18212616030896403, "grad_norm": 9.69013500213623, "learning_rate": 9.73482100075296e-05, "loss": 0.9562, "step": 2688 }, { "epoch": 0.1821939155769361, "grad_norm": 9.47673511505127, "learning_rate": 9.734684098843179e-05, "loss": 0.9822, "step": 2689 }, { "epoch": 0.18226167084490819, "grad_norm": 9.708051681518555, "learning_rate": 9.734547196933397e-05, "loss": 1.0772, "step": 2690 }, { "epoch": 0.18232942611288028, "grad_norm": 10.617173194885254, "learning_rate": 9.734410295023615e-05, "loss": 1.2686, "step": 2691 }, { "epoch": 0.18239718138085237, "grad_norm": 9.52670955657959, "learning_rate": 9.734273393113835e-05, "loss": 0.9477, "step": 2692 }, { "epoch": 0.18246493664882443, "grad_norm": 10.090290069580078, "learning_rate": 9.734136491204053e-05, "loss": 1.1766, "step": 2693 }, { "epoch": 0.18253269191679652, "grad_norm": 8.709790229797363, "learning_rate": 9.733999589294271e-05, "loss": 1.1327, "step": 2694 }, { "epoch": 0.18260044718476862, "grad_norm": 12.07381534576416, "learning_rate": 9.73386268738449e-05, "loss": 1.3694, "step": 2695 }, { "epoch": 0.1826682024527407, "grad_norm": 8.329826354980469, "learning_rate": 9.733725785474708e-05, "loss": 1.0448, "step": 2696 }, { "epoch": 0.18273595772071277, "grad_norm": 7.551383972167969, "learning_rate": 9.733588883564926e-05, "loss": 1.029, "step": 2697 }, { "epoch": 0.18280371298868486, "grad_norm": 9.393105506896973, "learning_rate": 9.733451981655146e-05, "loss": 0.9729, "step": 2698 }, { "epoch": 0.18287146825665695, "grad_norm": 9.283944129943848, "learning_rate": 9.733315079745364e-05, "loss": 1.2459, "step": 2699 }, { "epoch": 0.18293922352462905, "grad_norm": 9.02878189086914, "learning_rate": 9.733178177835582e-05, "loss": 1.3235, "step": 2700 }, { "epoch": 0.18300697879260114, "grad_norm": 8.735793113708496, "learning_rate": 9.7330412759258e-05, "loss": 1.1447, "step": 2701 }, { "epoch": 0.1830747340605732, "grad_norm": 8.511741638183594, "learning_rate": 9.732904374016018e-05, "loss": 1.0148, "step": 2702 }, { "epoch": 0.1831424893285453, "grad_norm": 6.590076446533203, "learning_rate": 9.732767472106237e-05, "loss": 1.0904, "step": 2703 }, { "epoch": 0.18321024459651739, "grad_norm": 11.523942947387695, "learning_rate": 9.732630570196455e-05, "loss": 0.9349, "step": 2704 }, { "epoch": 0.18327799986448948, "grad_norm": 6.972995281219482, "learning_rate": 9.732493668286673e-05, "loss": 0.9837, "step": 2705 }, { "epoch": 0.18334575513246154, "grad_norm": 7.834453105926514, "learning_rate": 9.732356766376891e-05, "loss": 1.0374, "step": 2706 }, { "epoch": 0.18341351040043363, "grad_norm": 9.44674015045166, "learning_rate": 9.732219864467111e-05, "loss": 1.4479, "step": 2707 }, { "epoch": 0.18348126566840572, "grad_norm": 9.017583847045898, "learning_rate": 9.732082962557329e-05, "loss": 1.136, "step": 2708 }, { "epoch": 0.18354902093637782, "grad_norm": 8.40336799621582, "learning_rate": 9.731946060647547e-05, "loss": 0.9914, "step": 2709 }, { "epoch": 0.18361677620434988, "grad_norm": 10.586843490600586, "learning_rate": 9.731809158737765e-05, "loss": 1.1515, "step": 2710 }, { "epoch": 0.18368453147232197, "grad_norm": 7.611972808837891, "learning_rate": 9.731672256827983e-05, "loss": 1.0821, "step": 2711 }, { "epoch": 0.18375228674029406, "grad_norm": 9.093097686767578, "learning_rate": 9.731535354918202e-05, "loss": 1.1295, "step": 2712 }, { "epoch": 0.18382004200826615, "grad_norm": 8.013279914855957, "learning_rate": 9.73139845300842e-05, "loss": 0.839, "step": 2713 }, { "epoch": 0.18388779727623822, "grad_norm": 7.941558837890625, "learning_rate": 9.731261551098638e-05, "loss": 1.0819, "step": 2714 }, { "epoch": 0.1839555525442103, "grad_norm": 7.09537935256958, "learning_rate": 9.731124649188856e-05, "loss": 0.8606, "step": 2715 }, { "epoch": 0.1840233078121824, "grad_norm": 7.867136478424072, "learning_rate": 9.730987747279074e-05, "loss": 0.9942, "step": 2716 }, { "epoch": 0.1840910630801545, "grad_norm": 8.66762638092041, "learning_rate": 9.730850845369294e-05, "loss": 1.0792, "step": 2717 }, { "epoch": 0.18415881834812656, "grad_norm": 7.97219705581665, "learning_rate": 9.730713943459512e-05, "loss": 1.0366, "step": 2718 }, { "epoch": 0.18422657361609865, "grad_norm": 8.341264724731445, "learning_rate": 9.73057704154973e-05, "loss": 1.0833, "step": 2719 }, { "epoch": 0.18429432888407074, "grad_norm": 8.704034805297852, "learning_rate": 9.730440139639948e-05, "loss": 1.2909, "step": 2720 }, { "epoch": 0.18436208415204283, "grad_norm": 9.759525299072266, "learning_rate": 9.730303237730167e-05, "loss": 1.1749, "step": 2721 }, { "epoch": 0.1844298394200149, "grad_norm": 9.2392578125, "learning_rate": 9.730166335820385e-05, "loss": 1.4037, "step": 2722 }, { "epoch": 0.184497594687987, "grad_norm": 10.519861221313477, "learning_rate": 9.730029433910603e-05, "loss": 1.2692, "step": 2723 }, { "epoch": 0.18456534995595908, "grad_norm": 8.285715103149414, "learning_rate": 9.729892532000821e-05, "loss": 0.8014, "step": 2724 }, { "epoch": 0.18463310522393117, "grad_norm": 7.261229991912842, "learning_rate": 9.72975563009104e-05, "loss": 0.9302, "step": 2725 }, { "epoch": 0.18470086049190323, "grad_norm": 10.149426460266113, "learning_rate": 9.729618728181259e-05, "loss": 1.2855, "step": 2726 }, { "epoch": 0.18476861575987533, "grad_norm": 7.495121955871582, "learning_rate": 9.729481826271477e-05, "loss": 0.9771, "step": 2727 }, { "epoch": 0.18483637102784742, "grad_norm": 9.102607727050781, "learning_rate": 9.729344924361695e-05, "loss": 1.1974, "step": 2728 }, { "epoch": 0.1849041262958195, "grad_norm": 9.57135009765625, "learning_rate": 9.729208022451913e-05, "loss": 0.9758, "step": 2729 }, { "epoch": 0.18497188156379157, "grad_norm": 8.745903015136719, "learning_rate": 9.729071120542132e-05, "loss": 1.1719, "step": 2730 }, { "epoch": 0.18503963683176367, "grad_norm": 10.697607040405273, "learning_rate": 9.72893421863235e-05, "loss": 0.952, "step": 2731 }, { "epoch": 0.18510739209973576, "grad_norm": 9.907280921936035, "learning_rate": 9.728797316722568e-05, "loss": 0.9571, "step": 2732 }, { "epoch": 0.18517514736770785, "grad_norm": 7.285250186920166, "learning_rate": 9.728660414812786e-05, "loss": 0.9002, "step": 2733 }, { "epoch": 0.1852429026356799, "grad_norm": 8.134112358093262, "learning_rate": 9.728523512903004e-05, "loss": 0.8645, "step": 2734 }, { "epoch": 0.185310657903652, "grad_norm": 9.427742004394531, "learning_rate": 9.728386610993224e-05, "loss": 1.2333, "step": 2735 }, { "epoch": 0.1853784131716241, "grad_norm": 8.804586410522461, "learning_rate": 9.728249709083442e-05, "loss": 1.2601, "step": 2736 }, { "epoch": 0.1854461684395962, "grad_norm": 9.144674301147461, "learning_rate": 9.72811280717366e-05, "loss": 0.9503, "step": 2737 }, { "epoch": 0.18551392370756825, "grad_norm": 9.399337768554688, "learning_rate": 9.727975905263879e-05, "loss": 0.8586, "step": 2738 }, { "epoch": 0.18558167897554034, "grad_norm": 7.7377119064331055, "learning_rate": 9.727839003354097e-05, "loss": 0.7511, "step": 2739 }, { "epoch": 0.18564943424351243, "grad_norm": 9.146937370300293, "learning_rate": 9.727702101444315e-05, "loss": 1.0037, "step": 2740 }, { "epoch": 0.18571718951148453, "grad_norm": 7.722357273101807, "learning_rate": 9.727565199534535e-05, "loss": 1.077, "step": 2741 }, { "epoch": 0.1857849447794566, "grad_norm": 7.808566093444824, "learning_rate": 9.727428297624753e-05, "loss": 0.8953, "step": 2742 }, { "epoch": 0.18585270004742868, "grad_norm": 8.765763282775879, "learning_rate": 9.727291395714971e-05, "loss": 1.0691, "step": 2743 }, { "epoch": 0.18592045531540077, "grad_norm": 11.350706100463867, "learning_rate": 9.72715449380519e-05, "loss": 0.9308, "step": 2744 }, { "epoch": 0.18598821058337286, "grad_norm": 7.838237285614014, "learning_rate": 9.727017591895408e-05, "loss": 0.9452, "step": 2745 }, { "epoch": 0.18605596585134493, "grad_norm": 9.660740852355957, "learning_rate": 9.726880689985626e-05, "loss": 0.8705, "step": 2746 }, { "epoch": 0.18612372111931702, "grad_norm": 8.146308898925781, "learning_rate": 9.726743788075844e-05, "loss": 1.0362, "step": 2747 }, { "epoch": 0.1861914763872891, "grad_norm": 9.901165008544922, "learning_rate": 9.726606886166062e-05, "loss": 1.0842, "step": 2748 }, { "epoch": 0.1862592316552612, "grad_norm": 7.650402069091797, "learning_rate": 9.726469984256282e-05, "loss": 0.997, "step": 2749 }, { "epoch": 0.18632698692323327, "grad_norm": 7.760092735290527, "learning_rate": 9.7263330823465e-05, "loss": 0.9903, "step": 2750 }, { "epoch": 0.18639474219120536, "grad_norm": 9.523726463317871, "learning_rate": 9.726196180436718e-05, "loss": 0.9276, "step": 2751 }, { "epoch": 0.18646249745917745, "grad_norm": 7.759490489959717, "learning_rate": 9.726059278526936e-05, "loss": 0.9356, "step": 2752 }, { "epoch": 0.18653025272714954, "grad_norm": 10.949979782104492, "learning_rate": 9.725922376617155e-05, "loss": 1.1451, "step": 2753 }, { "epoch": 0.18659800799512163, "grad_norm": 8.312686920166016, "learning_rate": 9.725785474707373e-05, "loss": 0.973, "step": 2754 }, { "epoch": 0.1866657632630937, "grad_norm": 6.999983787536621, "learning_rate": 9.725648572797591e-05, "loss": 1.0699, "step": 2755 }, { "epoch": 0.1867335185310658, "grad_norm": 9.422745704650879, "learning_rate": 9.725511670887809e-05, "loss": 1.0697, "step": 2756 }, { "epoch": 0.18680127379903788, "grad_norm": 8.054603576660156, "learning_rate": 9.725374768978027e-05, "loss": 0.8764, "step": 2757 }, { "epoch": 0.18686902906700997, "grad_norm": 7.902176856994629, "learning_rate": 9.725237867068247e-05, "loss": 0.9893, "step": 2758 }, { "epoch": 0.18693678433498204, "grad_norm": 8.409537315368652, "learning_rate": 9.725100965158465e-05, "loss": 1.0526, "step": 2759 }, { "epoch": 0.18700453960295413, "grad_norm": 9.867463111877441, "learning_rate": 9.724964063248683e-05, "loss": 1.1218, "step": 2760 }, { "epoch": 0.18707229487092622, "grad_norm": 6.852199554443359, "learning_rate": 9.724827161338901e-05, "loss": 0.9838, "step": 2761 }, { "epoch": 0.1871400501388983, "grad_norm": 9.132448196411133, "learning_rate": 9.72469025942912e-05, "loss": 1.038, "step": 2762 }, { "epoch": 0.18720780540687038, "grad_norm": 8.699213981628418, "learning_rate": 9.724553357519338e-05, "loss": 1.0532, "step": 2763 }, { "epoch": 0.18727556067484247, "grad_norm": 8.520672798156738, "learning_rate": 9.724416455609556e-05, "loss": 1.2942, "step": 2764 }, { "epoch": 0.18734331594281456, "grad_norm": 9.690433502197266, "learning_rate": 9.724279553699774e-05, "loss": 0.9552, "step": 2765 }, { "epoch": 0.18741107121078665, "grad_norm": 7.540090560913086, "learning_rate": 9.724142651789992e-05, "loss": 1.0718, "step": 2766 }, { "epoch": 0.18747882647875871, "grad_norm": 9.072039604187012, "learning_rate": 9.724005749880212e-05, "loss": 1.0313, "step": 2767 }, { "epoch": 0.1875465817467308, "grad_norm": 10.155011177062988, "learning_rate": 9.72386884797043e-05, "loss": 1.3498, "step": 2768 }, { "epoch": 0.1876143370147029, "grad_norm": 7.816718101501465, "learning_rate": 9.723731946060648e-05, "loss": 0.9594, "step": 2769 }, { "epoch": 0.187682092282675, "grad_norm": 10.980911254882812, "learning_rate": 9.723595044150866e-05, "loss": 1.1675, "step": 2770 }, { "epoch": 0.18774984755064705, "grad_norm": 8.865739822387695, "learning_rate": 9.723458142241084e-05, "loss": 1.1796, "step": 2771 }, { "epoch": 0.18781760281861914, "grad_norm": 8.97850227355957, "learning_rate": 9.723321240331303e-05, "loss": 1.0907, "step": 2772 }, { "epoch": 0.18788535808659124, "grad_norm": 8.16921329498291, "learning_rate": 9.723184338421521e-05, "loss": 0.95, "step": 2773 }, { "epoch": 0.18795311335456333, "grad_norm": 8.766203880310059, "learning_rate": 9.723047436511739e-05, "loss": 0.9087, "step": 2774 }, { "epoch": 0.1880208686225354, "grad_norm": 7.410607814788818, "learning_rate": 9.722910534601957e-05, "loss": 0.8082, "step": 2775 }, { "epoch": 0.18808862389050748, "grad_norm": 9.640182495117188, "learning_rate": 9.722773632692177e-05, "loss": 0.99, "step": 2776 }, { "epoch": 0.18815637915847958, "grad_norm": 9.038297653198242, "learning_rate": 9.722636730782395e-05, "loss": 1.0017, "step": 2777 }, { "epoch": 0.18822413442645167, "grad_norm": 10.167421340942383, "learning_rate": 9.722499828872613e-05, "loss": 1.1588, "step": 2778 }, { "epoch": 0.18829188969442373, "grad_norm": 9.911538124084473, "learning_rate": 9.722362926962831e-05, "loss": 1.052, "step": 2779 }, { "epoch": 0.18835964496239582, "grad_norm": 8.78661060333252, "learning_rate": 9.722226025053049e-05, "loss": 0.9827, "step": 2780 }, { "epoch": 0.18842740023036791, "grad_norm": 8.58356761932373, "learning_rate": 9.722089123143268e-05, "loss": 0.9663, "step": 2781 }, { "epoch": 0.18849515549834, "grad_norm": 7.882653713226318, "learning_rate": 9.721952221233486e-05, "loss": 1.0282, "step": 2782 }, { "epoch": 0.18856291076631207, "grad_norm": 8.180728912353516, "learning_rate": 9.721815319323704e-05, "loss": 0.8983, "step": 2783 }, { "epoch": 0.18863066603428416, "grad_norm": 10.785475730895996, "learning_rate": 9.721678417413924e-05, "loss": 1.0949, "step": 2784 }, { "epoch": 0.18869842130225625, "grad_norm": 8.2493257522583, "learning_rate": 9.721541515504142e-05, "loss": 0.8891, "step": 2785 }, { "epoch": 0.18876617657022834, "grad_norm": 8.298515319824219, "learning_rate": 9.72140461359436e-05, "loss": 0.9384, "step": 2786 }, { "epoch": 0.1888339318382004, "grad_norm": 11.643486022949219, "learning_rate": 9.721267711684579e-05, "loss": 1.0915, "step": 2787 }, { "epoch": 0.1889016871061725, "grad_norm": 10.728472709655762, "learning_rate": 9.721130809774797e-05, "loss": 1.562, "step": 2788 }, { "epoch": 0.1889694423741446, "grad_norm": 7.9253435134887695, "learning_rate": 9.720993907865015e-05, "loss": 0.876, "step": 2789 }, { "epoch": 0.18903719764211668, "grad_norm": 9.075439453125, "learning_rate": 9.720857005955235e-05, "loss": 1.1712, "step": 2790 }, { "epoch": 0.18910495291008875, "grad_norm": 8.572853088378906, "learning_rate": 9.720720104045453e-05, "loss": 0.9697, "step": 2791 }, { "epoch": 0.18917270817806084, "grad_norm": 8.40988826751709, "learning_rate": 9.720583202135671e-05, "loss": 1.0904, "step": 2792 }, { "epoch": 0.18924046344603293, "grad_norm": 8.015021324157715, "learning_rate": 9.720446300225889e-05, "loss": 1.0264, "step": 2793 }, { "epoch": 0.18930821871400502, "grad_norm": 10.295256614685059, "learning_rate": 9.720309398316107e-05, "loss": 0.9608, "step": 2794 }, { "epoch": 0.18937597398197709, "grad_norm": 9.377728462219238, "learning_rate": 9.720172496406326e-05, "loss": 1.2169, "step": 2795 }, { "epoch": 0.18944372924994918, "grad_norm": 8.55238151550293, "learning_rate": 9.720035594496544e-05, "loss": 1.0951, "step": 2796 }, { "epoch": 0.18951148451792127, "grad_norm": 10.069438934326172, "learning_rate": 9.719898692586762e-05, "loss": 1.3206, "step": 2797 }, { "epoch": 0.18957923978589336, "grad_norm": 9.229057312011719, "learning_rate": 9.71976179067698e-05, "loss": 1.18, "step": 2798 }, { "epoch": 0.18964699505386542, "grad_norm": 8.911051750183105, "learning_rate": 9.7196248887672e-05, "loss": 1.1351, "step": 2799 }, { "epoch": 0.18971475032183752, "grad_norm": 9.285752296447754, "learning_rate": 9.719487986857418e-05, "loss": 1.0626, "step": 2800 }, { "epoch": 0.1897825055898096, "grad_norm": 8.793561935424805, "learning_rate": 9.719351084947636e-05, "loss": 1.033, "step": 2801 }, { "epoch": 0.1898502608577817, "grad_norm": 10.635384559631348, "learning_rate": 9.719214183037854e-05, "loss": 1.5004, "step": 2802 }, { "epoch": 0.18991801612575376, "grad_norm": 7.9627275466918945, "learning_rate": 9.719077281128072e-05, "loss": 1.0246, "step": 2803 }, { "epoch": 0.18998577139372586, "grad_norm": 8.960352897644043, "learning_rate": 9.718940379218291e-05, "loss": 1.0316, "step": 2804 }, { "epoch": 0.19005352666169795, "grad_norm": 9.522171020507812, "learning_rate": 9.718803477308509e-05, "loss": 1.0219, "step": 2805 }, { "epoch": 0.19012128192967004, "grad_norm": 8.412702560424805, "learning_rate": 9.718666575398727e-05, "loss": 1.0609, "step": 2806 }, { "epoch": 0.19018903719764213, "grad_norm": 8.777839660644531, "learning_rate": 9.718529673488945e-05, "loss": 1.0854, "step": 2807 }, { "epoch": 0.1902567924656142, "grad_norm": 8.932796478271484, "learning_rate": 9.718392771579165e-05, "loss": 1.1661, "step": 2808 }, { "epoch": 0.19032454773358629, "grad_norm": 8.514800071716309, "learning_rate": 9.718255869669383e-05, "loss": 1.0584, "step": 2809 }, { "epoch": 0.19039230300155838, "grad_norm": 7.001948356628418, "learning_rate": 9.718118967759601e-05, "loss": 0.8479, "step": 2810 }, { "epoch": 0.19046005826953047, "grad_norm": 8.977307319641113, "learning_rate": 9.717982065849819e-05, "loss": 1.2556, "step": 2811 }, { "epoch": 0.19052781353750253, "grad_norm": 8.513920783996582, "learning_rate": 9.717845163940037e-05, "loss": 1.2076, "step": 2812 }, { "epoch": 0.19059556880547462, "grad_norm": 8.556622505187988, "learning_rate": 9.717708262030256e-05, "loss": 1.0491, "step": 2813 }, { "epoch": 0.19066332407344672, "grad_norm": 9.8518648147583, "learning_rate": 9.717571360120474e-05, "loss": 1.1179, "step": 2814 }, { "epoch": 0.1907310793414188, "grad_norm": 8.887413024902344, "learning_rate": 9.717434458210692e-05, "loss": 1.0675, "step": 2815 }, { "epoch": 0.19079883460939087, "grad_norm": 7.888981342315674, "learning_rate": 9.71729755630091e-05, "loss": 1.0489, "step": 2816 }, { "epoch": 0.19086658987736296, "grad_norm": 7.692848205566406, "learning_rate": 9.71716065439113e-05, "loss": 1.0602, "step": 2817 }, { "epoch": 0.19093434514533505, "grad_norm": 8.591787338256836, "learning_rate": 9.717023752481348e-05, "loss": 0.8737, "step": 2818 }, { "epoch": 0.19100210041330715, "grad_norm": 9.622870445251465, "learning_rate": 9.716886850571566e-05, "loss": 0.9717, "step": 2819 }, { "epoch": 0.1910698556812792, "grad_norm": 7.949582576751709, "learning_rate": 9.716749948661784e-05, "loss": 0.8651, "step": 2820 }, { "epoch": 0.1911376109492513, "grad_norm": 9.46272087097168, "learning_rate": 9.716613046752002e-05, "loss": 0.9975, "step": 2821 }, { "epoch": 0.1912053662172234, "grad_norm": 8.86549186706543, "learning_rate": 9.716476144842221e-05, "loss": 1.0805, "step": 2822 }, { "epoch": 0.19127312148519549, "grad_norm": 8.039673805236816, "learning_rate": 9.716339242932439e-05, "loss": 1.121, "step": 2823 }, { "epoch": 0.19134087675316755, "grad_norm": 8.606979370117188, "learning_rate": 9.716202341022657e-05, "loss": 0.9217, "step": 2824 }, { "epoch": 0.19140863202113964, "grad_norm": 10.429420471191406, "learning_rate": 9.716065439112875e-05, "loss": 1.2266, "step": 2825 }, { "epoch": 0.19147638728911173, "grad_norm": 7.339411735534668, "learning_rate": 9.715928537203093e-05, "loss": 1.0235, "step": 2826 }, { "epoch": 0.19154414255708382, "grad_norm": 8.247300148010254, "learning_rate": 9.715791635293313e-05, "loss": 0.8993, "step": 2827 }, { "epoch": 0.1916118978250559, "grad_norm": 8.475278854370117, "learning_rate": 9.715654733383531e-05, "loss": 0.9611, "step": 2828 }, { "epoch": 0.19167965309302798, "grad_norm": 9.792519569396973, "learning_rate": 9.715517831473749e-05, "loss": 1.0831, "step": 2829 }, { "epoch": 0.19174740836100007, "grad_norm": 6.757070541381836, "learning_rate": 9.715380929563967e-05, "loss": 1.0282, "step": 2830 }, { "epoch": 0.19181516362897216, "grad_norm": 6.707785129547119, "learning_rate": 9.715244027654186e-05, "loss": 0.7813, "step": 2831 }, { "epoch": 0.19188291889694423, "grad_norm": 9.990489959716797, "learning_rate": 9.715107125744404e-05, "loss": 1.0504, "step": 2832 }, { "epoch": 0.19195067416491632, "grad_norm": 10.57358169555664, "learning_rate": 9.714970223834622e-05, "loss": 0.9355, "step": 2833 }, { "epoch": 0.1920184294328884, "grad_norm": 9.225300788879395, "learning_rate": 9.714833321924842e-05, "loss": 1.2404, "step": 2834 }, { "epoch": 0.1920861847008605, "grad_norm": 8.36042594909668, "learning_rate": 9.71469642001506e-05, "loss": 0.9012, "step": 2835 }, { "epoch": 0.19215393996883257, "grad_norm": 9.624984741210938, "learning_rate": 9.714559518105278e-05, "loss": 1.0509, "step": 2836 }, { "epoch": 0.19222169523680466, "grad_norm": 12.569930076599121, "learning_rate": 9.714422616195497e-05, "loss": 1.2663, "step": 2837 }, { "epoch": 0.19228945050477675, "grad_norm": 12.278885841369629, "learning_rate": 9.714285714285715e-05, "loss": 1.1547, "step": 2838 }, { "epoch": 0.19235720577274884, "grad_norm": 9.3023681640625, "learning_rate": 9.714148812375933e-05, "loss": 1.0873, "step": 2839 }, { "epoch": 0.1924249610407209, "grad_norm": 8.241714477539062, "learning_rate": 9.714011910466153e-05, "loss": 0.8222, "step": 2840 }, { "epoch": 0.192492716308693, "grad_norm": 10.45174503326416, "learning_rate": 9.71387500855637e-05, "loss": 1.1974, "step": 2841 }, { "epoch": 0.1925604715766651, "grad_norm": 8.096826553344727, "learning_rate": 9.713738106646589e-05, "loss": 1.1147, "step": 2842 }, { "epoch": 0.19262822684463718, "grad_norm": 9.053191184997559, "learning_rate": 9.713601204736807e-05, "loss": 0.9835, "step": 2843 }, { "epoch": 0.19269598211260924, "grad_norm": 8.050823211669922, "learning_rate": 9.713464302827025e-05, "loss": 1.0826, "step": 2844 }, { "epoch": 0.19276373738058133, "grad_norm": 10.000917434692383, "learning_rate": 9.713327400917244e-05, "loss": 1.1723, "step": 2845 }, { "epoch": 0.19283149264855343, "grad_norm": 8.992589950561523, "learning_rate": 9.713190499007462e-05, "loss": 1.0203, "step": 2846 }, { "epoch": 0.19289924791652552, "grad_norm": 10.340285301208496, "learning_rate": 9.71305359709768e-05, "loss": 1.2523, "step": 2847 }, { "epoch": 0.19296700318449758, "grad_norm": 8.250594139099121, "learning_rate": 9.712916695187898e-05, "loss": 1.0606, "step": 2848 }, { "epoch": 0.19303475845246967, "grad_norm": 7.16335916519165, "learning_rate": 9.712779793278116e-05, "loss": 0.9988, "step": 2849 }, { "epoch": 0.19310251372044177, "grad_norm": 9.089055061340332, "learning_rate": 9.712642891368336e-05, "loss": 0.9368, "step": 2850 }, { "epoch": 0.19317026898841386, "grad_norm": 7.568434238433838, "learning_rate": 9.712505989458554e-05, "loss": 0.9417, "step": 2851 }, { "epoch": 0.19323802425638592, "grad_norm": 8.470823287963867, "learning_rate": 9.712369087548772e-05, "loss": 1.0271, "step": 2852 }, { "epoch": 0.193305779524358, "grad_norm": 7.746623992919922, "learning_rate": 9.71223218563899e-05, "loss": 0.7641, "step": 2853 }, { "epoch": 0.1933735347923301, "grad_norm": 7.44852352142334, "learning_rate": 9.712095283729209e-05, "loss": 0.9984, "step": 2854 }, { "epoch": 0.1934412900603022, "grad_norm": 7.393777847290039, "learning_rate": 9.711958381819427e-05, "loss": 0.8515, "step": 2855 }, { "epoch": 0.19350904532827426, "grad_norm": 8.247236251831055, "learning_rate": 9.711821479909645e-05, "loss": 0.9867, "step": 2856 }, { "epoch": 0.19357680059624635, "grad_norm": 8.484920501708984, "learning_rate": 9.711684577999863e-05, "loss": 0.9099, "step": 2857 }, { "epoch": 0.19364455586421844, "grad_norm": 9.85857105255127, "learning_rate": 9.711547676090081e-05, "loss": 1.0134, "step": 2858 }, { "epoch": 0.19371231113219053, "grad_norm": 12.191691398620605, "learning_rate": 9.7114107741803e-05, "loss": 1.15, "step": 2859 }, { "epoch": 0.19378006640016263, "grad_norm": 9.691742897033691, "learning_rate": 9.711273872270519e-05, "loss": 1.2415, "step": 2860 }, { "epoch": 0.1938478216681347, "grad_norm": 11.395289421081543, "learning_rate": 9.711136970360737e-05, "loss": 1.3239, "step": 2861 }, { "epoch": 0.19391557693610678, "grad_norm": 9.222856521606445, "learning_rate": 9.711000068450955e-05, "loss": 0.9087, "step": 2862 }, { "epoch": 0.19398333220407887, "grad_norm": 8.062904357910156, "learning_rate": 9.710863166541174e-05, "loss": 1.093, "step": 2863 }, { "epoch": 0.19405108747205096, "grad_norm": 8.160481452941895, "learning_rate": 9.710726264631392e-05, "loss": 0.842, "step": 2864 }, { "epoch": 0.19411884274002303, "grad_norm": 8.165858268737793, "learning_rate": 9.71058936272161e-05, "loss": 0.8807, "step": 2865 }, { "epoch": 0.19418659800799512, "grad_norm": 8.120240211486816, "learning_rate": 9.710452460811828e-05, "loss": 0.9733, "step": 2866 }, { "epoch": 0.1942543532759672, "grad_norm": 11.363536834716797, "learning_rate": 9.710315558902046e-05, "loss": 1.1144, "step": 2867 }, { "epoch": 0.1943221085439393, "grad_norm": 8.776150703430176, "learning_rate": 9.710178656992266e-05, "loss": 0.7561, "step": 2868 }, { "epoch": 0.19438986381191137, "grad_norm": 9.393696784973145, "learning_rate": 9.710041755082484e-05, "loss": 0.8092, "step": 2869 }, { "epoch": 0.19445761907988346, "grad_norm": 10.820277214050293, "learning_rate": 9.709904853172702e-05, "loss": 0.9905, "step": 2870 }, { "epoch": 0.19452537434785555, "grad_norm": 8.041844367980957, "learning_rate": 9.70976795126292e-05, "loss": 0.8801, "step": 2871 }, { "epoch": 0.19459312961582764, "grad_norm": 8.05355167388916, "learning_rate": 9.709631049353138e-05, "loss": 0.9946, "step": 2872 }, { "epoch": 0.1946608848837997, "grad_norm": 9.46949577331543, "learning_rate": 9.709494147443357e-05, "loss": 1.0582, "step": 2873 }, { "epoch": 0.1947286401517718, "grad_norm": 8.796204566955566, "learning_rate": 9.709357245533575e-05, "loss": 1.1542, "step": 2874 }, { "epoch": 0.1947963954197439, "grad_norm": 10.45006275177002, "learning_rate": 9.709220343623793e-05, "loss": 1.2363, "step": 2875 }, { "epoch": 0.19486415068771598, "grad_norm": 10.490102767944336, "learning_rate": 9.709083441714011e-05, "loss": 1.1977, "step": 2876 }, { "epoch": 0.19493190595568805, "grad_norm": 9.048376083374023, "learning_rate": 9.708946539804231e-05, "loss": 0.8127, "step": 2877 }, { "epoch": 0.19499966122366014, "grad_norm": 6.523111343383789, "learning_rate": 9.708809637894449e-05, "loss": 0.8182, "step": 2878 }, { "epoch": 0.19506741649163223, "grad_norm": 7.170145511627197, "learning_rate": 9.708672735984667e-05, "loss": 0.9643, "step": 2879 }, { "epoch": 0.19513517175960432, "grad_norm": 7.9981818199157715, "learning_rate": 9.708535834074886e-05, "loss": 1.3179, "step": 2880 }, { "epoch": 0.19520292702757638, "grad_norm": 8.786405563354492, "learning_rate": 9.708398932165104e-05, "loss": 0.9217, "step": 2881 }, { "epoch": 0.19527068229554848, "grad_norm": 6.7907867431640625, "learning_rate": 9.708262030255322e-05, "loss": 1.0381, "step": 2882 }, { "epoch": 0.19533843756352057, "grad_norm": 7.1751556396484375, "learning_rate": 9.708125128345542e-05, "loss": 0.8334, "step": 2883 }, { "epoch": 0.19540619283149266, "grad_norm": 8.236610412597656, "learning_rate": 9.70798822643576e-05, "loss": 1.2058, "step": 2884 }, { "epoch": 0.19547394809946472, "grad_norm": 7.576273441314697, "learning_rate": 9.707851324525978e-05, "loss": 1.091, "step": 2885 }, { "epoch": 0.19554170336743681, "grad_norm": 8.698029518127441, "learning_rate": 9.707714422616197e-05, "loss": 1.0262, "step": 2886 }, { "epoch": 0.1956094586354089, "grad_norm": 8.583345413208008, "learning_rate": 9.707577520706415e-05, "loss": 1.0401, "step": 2887 }, { "epoch": 0.195677213903381, "grad_norm": 7.242405891418457, "learning_rate": 9.707440618796633e-05, "loss": 0.785, "step": 2888 }, { "epoch": 0.19574496917135306, "grad_norm": 8.4541654586792, "learning_rate": 9.707303716886851e-05, "loss": 1.0537, "step": 2889 }, { "epoch": 0.19581272443932515, "grad_norm": 7.838657855987549, "learning_rate": 9.707166814977069e-05, "loss": 1.0062, "step": 2890 }, { "epoch": 0.19588047970729724, "grad_norm": 6.812248229980469, "learning_rate": 9.707029913067289e-05, "loss": 0.885, "step": 2891 }, { "epoch": 0.19594823497526934, "grad_norm": 8.789966583251953, "learning_rate": 9.706893011157507e-05, "loss": 1.0366, "step": 2892 }, { "epoch": 0.1960159902432414, "grad_norm": 8.860052108764648, "learning_rate": 9.706756109247725e-05, "loss": 1.2247, "step": 2893 }, { "epoch": 0.1960837455112135, "grad_norm": 11.235320091247559, "learning_rate": 9.706619207337943e-05, "loss": 1.1216, "step": 2894 }, { "epoch": 0.19615150077918558, "grad_norm": 8.865259170532227, "learning_rate": 9.706482305428162e-05, "loss": 0.9467, "step": 2895 }, { "epoch": 0.19621925604715768, "grad_norm": 8.345112800598145, "learning_rate": 9.70634540351838e-05, "loss": 0.9849, "step": 2896 }, { "epoch": 0.19628701131512974, "grad_norm": 8.98128604888916, "learning_rate": 9.706208501608598e-05, "loss": 1.1421, "step": 2897 }, { "epoch": 0.19635476658310183, "grad_norm": 7.786384582519531, "learning_rate": 9.706071599698816e-05, "loss": 1.0926, "step": 2898 }, { "epoch": 0.19642252185107392, "grad_norm": 8.816730499267578, "learning_rate": 9.705934697789034e-05, "loss": 1.2626, "step": 2899 }, { "epoch": 0.19649027711904601, "grad_norm": 7.831095218658447, "learning_rate": 9.705797795879254e-05, "loss": 0.9355, "step": 2900 }, { "epoch": 0.19655803238701808, "grad_norm": 8.788371086120605, "learning_rate": 9.705660893969472e-05, "loss": 0.8883, "step": 2901 }, { "epoch": 0.19662578765499017, "grad_norm": 8.88425064086914, "learning_rate": 9.70552399205969e-05, "loss": 1.1169, "step": 2902 }, { "epoch": 0.19669354292296226, "grad_norm": 9.648268699645996, "learning_rate": 9.705387090149908e-05, "loss": 1.0628, "step": 2903 }, { "epoch": 0.19676129819093435, "grad_norm": 8.19676685333252, "learning_rate": 9.705250188240126e-05, "loss": 0.9395, "step": 2904 }, { "epoch": 0.19682905345890642, "grad_norm": 7.420725345611572, "learning_rate": 9.705113286330345e-05, "loss": 0.947, "step": 2905 }, { "epoch": 0.1968968087268785, "grad_norm": 7.221796989440918, "learning_rate": 9.704976384420563e-05, "loss": 0.8491, "step": 2906 }, { "epoch": 0.1969645639948506, "grad_norm": 9.932676315307617, "learning_rate": 9.704839482510781e-05, "loss": 1.117, "step": 2907 }, { "epoch": 0.1970323192628227, "grad_norm": 11.816266059875488, "learning_rate": 9.704702580600999e-05, "loss": 1.4347, "step": 2908 }, { "epoch": 0.19710007453079476, "grad_norm": 8.804407119750977, "learning_rate": 9.704565678691219e-05, "loss": 0.808, "step": 2909 }, { "epoch": 0.19716782979876685, "grad_norm": 7.740353584289551, "learning_rate": 9.704428776781437e-05, "loss": 1.0732, "step": 2910 }, { "epoch": 0.19723558506673894, "grad_norm": 12.040196418762207, "learning_rate": 9.704291874871655e-05, "loss": 0.947, "step": 2911 }, { "epoch": 0.19730334033471103, "grad_norm": 7.727171421051025, "learning_rate": 9.704154972961873e-05, "loss": 0.8211, "step": 2912 }, { "epoch": 0.19737109560268312, "grad_norm": 8.779428482055664, "learning_rate": 9.704018071052091e-05, "loss": 1.2204, "step": 2913 }, { "epoch": 0.19743885087065519, "grad_norm": 7.907576084136963, "learning_rate": 9.70388116914231e-05, "loss": 1.0594, "step": 2914 }, { "epoch": 0.19750660613862728, "grad_norm": 6.769292831420898, "learning_rate": 9.703744267232528e-05, "loss": 0.8743, "step": 2915 }, { "epoch": 0.19757436140659937, "grad_norm": 8.966355323791504, "learning_rate": 9.703607365322746e-05, "loss": 1.1607, "step": 2916 }, { "epoch": 0.19764211667457146, "grad_norm": 7.389810085296631, "learning_rate": 9.703470463412964e-05, "loss": 0.9728, "step": 2917 }, { "epoch": 0.19770987194254352, "grad_norm": 10.877386093139648, "learning_rate": 9.703333561503184e-05, "loss": 1.0623, "step": 2918 }, { "epoch": 0.19777762721051562, "grad_norm": 9.160116195678711, "learning_rate": 9.703196659593402e-05, "loss": 1.1374, "step": 2919 }, { "epoch": 0.1978453824784877, "grad_norm": 7.6289167404174805, "learning_rate": 9.70305975768362e-05, "loss": 0.9498, "step": 2920 }, { "epoch": 0.1979131377464598, "grad_norm": 8.663583755493164, "learning_rate": 9.702922855773838e-05, "loss": 1.1595, "step": 2921 }, { "epoch": 0.19798089301443186, "grad_norm": 7.595486640930176, "learning_rate": 9.702785953864056e-05, "loss": 1.0623, "step": 2922 }, { "epoch": 0.19804864828240396, "grad_norm": 8.680171966552734, "learning_rate": 9.702649051954275e-05, "loss": 1.0138, "step": 2923 }, { "epoch": 0.19811640355037605, "grad_norm": 9.998015403747559, "learning_rate": 9.702512150044493e-05, "loss": 1.2006, "step": 2924 }, { "epoch": 0.19818415881834814, "grad_norm": 7.801429271697998, "learning_rate": 9.702375248134711e-05, "loss": 0.8467, "step": 2925 }, { "epoch": 0.1982519140863202, "grad_norm": 8.570688247680664, "learning_rate": 9.70223834622493e-05, "loss": 1.1439, "step": 2926 }, { "epoch": 0.1983196693542923, "grad_norm": 7.835936069488525, "learning_rate": 9.702101444315149e-05, "loss": 0.9671, "step": 2927 }, { "epoch": 0.19838742462226439, "grad_norm": 8.415708541870117, "learning_rate": 9.701964542405367e-05, "loss": 1.0381, "step": 2928 }, { "epoch": 0.19845517989023648, "grad_norm": 10.370524406433105, "learning_rate": 9.701827640495586e-05, "loss": 1.151, "step": 2929 }, { "epoch": 0.19852293515820854, "grad_norm": 8.228797912597656, "learning_rate": 9.701690738585804e-05, "loss": 1.093, "step": 2930 }, { "epoch": 0.19859069042618063, "grad_norm": 8.070756912231445, "learning_rate": 9.701553836676022e-05, "loss": 1.0272, "step": 2931 }, { "epoch": 0.19865844569415272, "grad_norm": 9.19532585144043, "learning_rate": 9.701416934766242e-05, "loss": 1.1195, "step": 2932 }, { "epoch": 0.19872620096212482, "grad_norm": 10.692606925964355, "learning_rate": 9.70128003285646e-05, "loss": 1.3148, "step": 2933 }, { "epoch": 0.19879395623009688, "grad_norm": 10.058424949645996, "learning_rate": 9.701143130946678e-05, "loss": 1.4236, "step": 2934 }, { "epoch": 0.19886171149806897, "grad_norm": 6.197395324707031, "learning_rate": 9.701006229036896e-05, "loss": 0.8822, "step": 2935 }, { "epoch": 0.19892946676604106, "grad_norm": 7.603270053863525, "learning_rate": 9.700869327127114e-05, "loss": 0.9655, "step": 2936 }, { "epoch": 0.19899722203401315, "grad_norm": 8.192676544189453, "learning_rate": 9.700732425217333e-05, "loss": 0.9668, "step": 2937 }, { "epoch": 0.19906497730198522, "grad_norm": 7.121623992919922, "learning_rate": 9.700595523307551e-05, "loss": 1.0037, "step": 2938 }, { "epoch": 0.1991327325699573, "grad_norm": 6.8974127769470215, "learning_rate": 9.700458621397769e-05, "loss": 0.876, "step": 2939 }, { "epoch": 0.1992004878379294, "grad_norm": 7.590656757354736, "learning_rate": 9.700321719487987e-05, "loss": 1.0346, "step": 2940 }, { "epoch": 0.1992682431059015, "grad_norm": 8.530266761779785, "learning_rate": 9.700184817578207e-05, "loss": 1.0882, "step": 2941 }, { "epoch": 0.19933599837387356, "grad_norm": 8.064129829406738, "learning_rate": 9.700047915668425e-05, "loss": 0.9949, "step": 2942 }, { "epoch": 0.19940375364184565, "grad_norm": 7.23117208480835, "learning_rate": 9.699911013758643e-05, "loss": 0.7269, "step": 2943 }, { "epoch": 0.19947150890981774, "grad_norm": 7.326268196105957, "learning_rate": 9.69977411184886e-05, "loss": 0.8542, "step": 2944 }, { "epoch": 0.19953926417778983, "grad_norm": 7.708505153656006, "learning_rate": 9.699637209939079e-05, "loss": 0.8206, "step": 2945 }, { "epoch": 0.1996070194457619, "grad_norm": 10.134513854980469, "learning_rate": 9.699500308029298e-05, "loss": 1.0079, "step": 2946 }, { "epoch": 0.199674774713734, "grad_norm": 9.264663696289062, "learning_rate": 9.699363406119516e-05, "loss": 1.1702, "step": 2947 }, { "epoch": 0.19974252998170608, "grad_norm": 8.894827842712402, "learning_rate": 9.699226504209734e-05, "loss": 0.9184, "step": 2948 }, { "epoch": 0.19981028524967817, "grad_norm": 6.935434341430664, "learning_rate": 9.699089602299952e-05, "loss": 0.8554, "step": 2949 }, { "epoch": 0.19987804051765024, "grad_norm": 9.57607364654541, "learning_rate": 9.698952700390172e-05, "loss": 1.0937, "step": 2950 }, { "epoch": 0.19994579578562233, "grad_norm": 7.99752140045166, "learning_rate": 9.69881579848039e-05, "loss": 1.0797, "step": 2951 }, { "epoch": 0.20001355105359442, "grad_norm": 8.067659378051758, "learning_rate": 9.698678896570608e-05, "loss": 1.0515, "step": 2952 }, { "epoch": 0.2000813063215665, "grad_norm": 9.66697883605957, "learning_rate": 9.698541994660826e-05, "loss": 1.289, "step": 2953 }, { "epoch": 0.20014906158953857, "grad_norm": 7.3660664558410645, "learning_rate": 9.698405092751044e-05, "loss": 0.7409, "step": 2954 }, { "epoch": 0.20021681685751067, "grad_norm": 8.254073143005371, "learning_rate": 9.698268190841263e-05, "loss": 1.1095, "step": 2955 }, { "epoch": 0.20028457212548276, "grad_norm": 8.221102714538574, "learning_rate": 9.698131288931481e-05, "loss": 1.0083, "step": 2956 }, { "epoch": 0.20035232739345485, "grad_norm": 9.813411712646484, "learning_rate": 9.697994387021699e-05, "loss": 1.2888, "step": 2957 }, { "epoch": 0.2004200826614269, "grad_norm": 10.87628173828125, "learning_rate": 9.697857485111917e-05, "loss": 1.16, "step": 2958 }, { "epoch": 0.200487837929399, "grad_norm": 7.094732284545898, "learning_rate": 9.697720583202135e-05, "loss": 0.8959, "step": 2959 }, { "epoch": 0.2005555931973711, "grad_norm": 9.814677238464355, "learning_rate": 9.697583681292355e-05, "loss": 1.1011, "step": 2960 }, { "epoch": 0.2006233484653432, "grad_norm": 8.982966423034668, "learning_rate": 9.697446779382573e-05, "loss": 1.001, "step": 2961 }, { "epoch": 0.20069110373331525, "grad_norm": 8.49453067779541, "learning_rate": 9.697309877472791e-05, "loss": 0.9848, "step": 2962 }, { "epoch": 0.20075885900128734, "grad_norm": 7.239814758300781, "learning_rate": 9.697172975563009e-05, "loss": 0.8252, "step": 2963 }, { "epoch": 0.20082661426925943, "grad_norm": 10.874746322631836, "learning_rate": 9.697036073653228e-05, "loss": 1.0683, "step": 2964 }, { "epoch": 0.20089436953723153, "grad_norm": 7.679197311401367, "learning_rate": 9.696899171743446e-05, "loss": 1.2797, "step": 2965 }, { "epoch": 0.20096212480520362, "grad_norm": 10.089177131652832, "learning_rate": 9.696762269833664e-05, "loss": 1.1905, "step": 2966 }, { "epoch": 0.20102988007317568, "grad_norm": 8.182350158691406, "learning_rate": 9.696625367923882e-05, "loss": 0.9917, "step": 2967 }, { "epoch": 0.20109763534114777, "grad_norm": 7.8756256103515625, "learning_rate": 9.6964884660141e-05, "loss": 1.0368, "step": 2968 }, { "epoch": 0.20116539060911987, "grad_norm": 9.193910598754883, "learning_rate": 9.69635156410432e-05, "loss": 1.094, "step": 2969 }, { "epoch": 0.20123314587709196, "grad_norm": 8.854869842529297, "learning_rate": 9.696214662194538e-05, "loss": 1.1344, "step": 2970 }, { "epoch": 0.20130090114506402, "grad_norm": 10.419108390808105, "learning_rate": 9.696077760284756e-05, "loss": 0.9937, "step": 2971 }, { "epoch": 0.2013686564130361, "grad_norm": 9.329347610473633, "learning_rate": 9.695940858374975e-05, "loss": 0.9454, "step": 2972 }, { "epoch": 0.2014364116810082, "grad_norm": 8.23154067993164, "learning_rate": 9.695803956465193e-05, "loss": 0.8691, "step": 2973 }, { "epoch": 0.2015041669489803, "grad_norm": 8.016939163208008, "learning_rate": 9.695667054555411e-05, "loss": 0.9182, "step": 2974 }, { "epoch": 0.20157192221695236, "grad_norm": 9.717400550842285, "learning_rate": 9.69553015264563e-05, "loss": 1.1251, "step": 2975 }, { "epoch": 0.20163967748492445, "grad_norm": 10.470111846923828, "learning_rate": 9.695393250735849e-05, "loss": 1.1634, "step": 2976 }, { "epoch": 0.20170743275289654, "grad_norm": 8.540326118469238, "learning_rate": 9.695256348826067e-05, "loss": 0.9229, "step": 2977 }, { "epoch": 0.20177518802086863, "grad_norm": 6.997597694396973, "learning_rate": 9.695119446916286e-05, "loss": 0.9911, "step": 2978 }, { "epoch": 0.2018429432888407, "grad_norm": 7.23951530456543, "learning_rate": 9.694982545006504e-05, "loss": 0.8372, "step": 2979 }, { "epoch": 0.2019106985568128, "grad_norm": 9.217951774597168, "learning_rate": 9.694845643096722e-05, "loss": 0.8154, "step": 2980 }, { "epoch": 0.20197845382478488, "grad_norm": 8.128033638000488, "learning_rate": 9.69470874118694e-05, "loss": 0.9494, "step": 2981 }, { "epoch": 0.20204620909275697, "grad_norm": 8.328935623168945, "learning_rate": 9.694571839277158e-05, "loss": 1.0537, "step": 2982 }, { "epoch": 0.20211396436072904, "grad_norm": 8.877389907836914, "learning_rate": 9.694434937367378e-05, "loss": 0.8683, "step": 2983 }, { "epoch": 0.20218171962870113, "grad_norm": 7.285436630249023, "learning_rate": 9.694298035457596e-05, "loss": 0.7514, "step": 2984 }, { "epoch": 0.20224947489667322, "grad_norm": 9.209798812866211, "learning_rate": 9.694161133547814e-05, "loss": 1.0771, "step": 2985 }, { "epoch": 0.2023172301646453, "grad_norm": 8.625777244567871, "learning_rate": 9.694024231638032e-05, "loss": 1.0085, "step": 2986 }, { "epoch": 0.20238498543261738, "grad_norm": 8.50123405456543, "learning_rate": 9.693887329728251e-05, "loss": 0.969, "step": 2987 }, { "epoch": 0.20245274070058947, "grad_norm": 7.314642429351807, "learning_rate": 9.693750427818469e-05, "loss": 0.8075, "step": 2988 }, { "epoch": 0.20252049596856156, "grad_norm": 9.474241256713867, "learning_rate": 9.693613525908687e-05, "loss": 1.0989, "step": 2989 }, { "epoch": 0.20258825123653365, "grad_norm": 7.3510637283325195, "learning_rate": 9.693476623998905e-05, "loss": 0.8044, "step": 2990 }, { "epoch": 0.20265600650450571, "grad_norm": 7.203106880187988, "learning_rate": 9.693339722089123e-05, "loss": 1.0359, "step": 2991 }, { "epoch": 0.2027237617724778, "grad_norm": 6.71024227142334, "learning_rate": 9.693202820179343e-05, "loss": 0.8687, "step": 2992 }, { "epoch": 0.2027915170404499, "grad_norm": 8.327759742736816, "learning_rate": 9.69306591826956e-05, "loss": 1.0042, "step": 2993 }, { "epoch": 0.202859272308422, "grad_norm": 8.682476997375488, "learning_rate": 9.692929016359779e-05, "loss": 1.0273, "step": 2994 }, { "epoch": 0.20292702757639405, "grad_norm": 8.555792808532715, "learning_rate": 9.692792114449997e-05, "loss": 0.8712, "step": 2995 }, { "epoch": 0.20299478284436615, "grad_norm": 11.036639213562012, "learning_rate": 9.692655212540216e-05, "loss": 1.1452, "step": 2996 }, { "epoch": 0.20306253811233824, "grad_norm": 10.207952499389648, "learning_rate": 9.692518310630434e-05, "loss": 1.2267, "step": 2997 }, { "epoch": 0.20313029338031033, "grad_norm": 8.715107917785645, "learning_rate": 9.692381408720652e-05, "loss": 1.151, "step": 2998 }, { "epoch": 0.2031980486482824, "grad_norm": 8.72461986541748, "learning_rate": 9.69224450681087e-05, "loss": 0.9787, "step": 2999 }, { "epoch": 0.20326580391625448, "grad_norm": 8.717243194580078, "learning_rate": 9.692107604901088e-05, "loss": 1.0734, "step": 3000 }, { "epoch": 0.20333355918422658, "grad_norm": 7.039597511291504, "learning_rate": 9.691970702991308e-05, "loss": 0.9215, "step": 3001 }, { "epoch": 0.20340131445219867, "grad_norm": 10.568238258361816, "learning_rate": 9.691833801081526e-05, "loss": 1.2248, "step": 3002 }, { "epoch": 0.20346906972017073, "grad_norm": 9.515549659729004, "learning_rate": 9.691696899171744e-05, "loss": 1.1401, "step": 3003 }, { "epoch": 0.20353682498814282, "grad_norm": 9.650483131408691, "learning_rate": 9.691559997261962e-05, "loss": 0.9839, "step": 3004 }, { "epoch": 0.20360458025611491, "grad_norm": 6.817119598388672, "learning_rate": 9.691423095352181e-05, "loss": 0.8542, "step": 3005 }, { "epoch": 0.203672335524087, "grad_norm": 7.616591930389404, "learning_rate": 9.691286193442399e-05, "loss": 0.9647, "step": 3006 }, { "epoch": 0.20374009079205907, "grad_norm": 7.2600274085998535, "learning_rate": 9.691149291532617e-05, "loss": 0.962, "step": 3007 }, { "epoch": 0.20380784606003116, "grad_norm": 9.714008331298828, "learning_rate": 9.691012389622835e-05, "loss": 1.1567, "step": 3008 }, { "epoch": 0.20387560132800325, "grad_norm": 9.095394134521484, "learning_rate": 9.690875487713053e-05, "loss": 1.1315, "step": 3009 }, { "epoch": 0.20394335659597534, "grad_norm": 9.683954238891602, "learning_rate": 9.690738585803273e-05, "loss": 1.4039, "step": 3010 }, { "epoch": 0.2040111118639474, "grad_norm": 8.829015731811523, "learning_rate": 9.69060168389349e-05, "loss": 0.9176, "step": 3011 }, { "epoch": 0.2040788671319195, "grad_norm": 7.835269927978516, "learning_rate": 9.690464781983709e-05, "loss": 1.0077, "step": 3012 }, { "epoch": 0.2041466223998916, "grad_norm": 9.94642448425293, "learning_rate": 9.690327880073927e-05, "loss": 1.0815, "step": 3013 }, { "epoch": 0.20421437766786368, "grad_norm": 8.184757232666016, "learning_rate": 9.690190978164145e-05, "loss": 0.84, "step": 3014 }, { "epoch": 0.20428213293583575, "grad_norm": 9.060220718383789, "learning_rate": 9.690054076254364e-05, "loss": 0.907, "step": 3015 }, { "epoch": 0.20434988820380784, "grad_norm": 6.848534107208252, "learning_rate": 9.689917174344582e-05, "loss": 0.7549, "step": 3016 }, { "epoch": 0.20441764347177993, "grad_norm": 7.820966720581055, "learning_rate": 9.6897802724348e-05, "loss": 0.9899, "step": 3017 }, { "epoch": 0.20448539873975202, "grad_norm": 10.148963928222656, "learning_rate": 9.68964337052502e-05, "loss": 1.2403, "step": 3018 }, { "epoch": 0.20455315400772411, "grad_norm": 8.273184776306152, "learning_rate": 9.689506468615238e-05, "loss": 1.0002, "step": 3019 }, { "epoch": 0.20462090927569618, "grad_norm": 7.618801593780518, "learning_rate": 9.689369566705456e-05, "loss": 0.8864, "step": 3020 }, { "epoch": 0.20468866454366827, "grad_norm": 7.952611446380615, "learning_rate": 9.689232664795675e-05, "loss": 0.8002, "step": 3021 }, { "epoch": 0.20475641981164036, "grad_norm": 9.938977241516113, "learning_rate": 9.689095762885893e-05, "loss": 0.8956, "step": 3022 }, { "epoch": 0.20482417507961245, "grad_norm": 7.807236194610596, "learning_rate": 9.688958860976111e-05, "loss": 1.1777, "step": 3023 }, { "epoch": 0.20489193034758452, "grad_norm": 7.7249369621276855, "learning_rate": 9.68882195906633e-05, "loss": 0.8881, "step": 3024 }, { "epoch": 0.2049596856155566, "grad_norm": 7.747461795806885, "learning_rate": 9.688685057156549e-05, "loss": 0.9844, "step": 3025 }, { "epoch": 0.2050274408835287, "grad_norm": 10.017412185668945, "learning_rate": 9.688548155246767e-05, "loss": 1.0815, "step": 3026 }, { "epoch": 0.2050951961515008, "grad_norm": 6.54990816116333, "learning_rate": 9.688411253336985e-05, "loss": 0.9319, "step": 3027 }, { "epoch": 0.20516295141947286, "grad_norm": 7.358734130859375, "learning_rate": 9.688274351427204e-05, "loss": 0.9313, "step": 3028 }, { "epoch": 0.20523070668744495, "grad_norm": 7.608468055725098, "learning_rate": 9.688137449517422e-05, "loss": 1.0071, "step": 3029 }, { "epoch": 0.20529846195541704, "grad_norm": 7.013155937194824, "learning_rate": 9.68800054760764e-05, "loss": 0.9419, "step": 3030 }, { "epoch": 0.20536621722338913, "grad_norm": 9.907796859741211, "learning_rate": 9.687863645697858e-05, "loss": 1.2082, "step": 3031 }, { "epoch": 0.2054339724913612, "grad_norm": 9.059138298034668, "learning_rate": 9.687726743788076e-05, "loss": 1.0712, "step": 3032 }, { "epoch": 0.20550172775933329, "grad_norm": 10.519928932189941, "learning_rate": 9.687589841878296e-05, "loss": 1.1627, "step": 3033 }, { "epoch": 0.20556948302730538, "grad_norm": 8.381184577941895, "learning_rate": 9.687452939968514e-05, "loss": 1.087, "step": 3034 }, { "epoch": 0.20563723829527747, "grad_norm": 6.514460563659668, "learning_rate": 9.687316038058732e-05, "loss": 0.8657, "step": 3035 }, { "epoch": 0.20570499356324953, "grad_norm": 10.034708023071289, "learning_rate": 9.68717913614895e-05, "loss": 1.1529, "step": 3036 }, { "epoch": 0.20577274883122162, "grad_norm": 7.202263355255127, "learning_rate": 9.687042234239168e-05, "loss": 0.7923, "step": 3037 }, { "epoch": 0.20584050409919372, "grad_norm": 9.7435302734375, "learning_rate": 9.686905332329387e-05, "loss": 1.5071, "step": 3038 }, { "epoch": 0.2059082593671658, "grad_norm": 9.57016658782959, "learning_rate": 9.686768430419605e-05, "loss": 0.883, "step": 3039 }, { "epoch": 0.20597601463513787, "grad_norm": 7.3575029373168945, "learning_rate": 9.686631528509823e-05, "loss": 0.9999, "step": 3040 }, { "epoch": 0.20604376990310996, "grad_norm": 9.0224027633667, "learning_rate": 9.686494626600041e-05, "loss": 1.0121, "step": 3041 }, { "epoch": 0.20611152517108206, "grad_norm": 11.173224449157715, "learning_rate": 9.68635772469026e-05, "loss": 1.0556, "step": 3042 }, { "epoch": 0.20617928043905415, "grad_norm": 8.858287811279297, "learning_rate": 9.686220822780479e-05, "loss": 0.9377, "step": 3043 }, { "epoch": 0.2062470357070262, "grad_norm": 7.3096795082092285, "learning_rate": 9.686083920870697e-05, "loss": 0.8808, "step": 3044 }, { "epoch": 0.2063147909749983, "grad_norm": 8.700214385986328, "learning_rate": 9.685947018960915e-05, "loss": 1.2121, "step": 3045 }, { "epoch": 0.2063825462429704, "grad_norm": 10.57944107055664, "learning_rate": 9.685810117051133e-05, "loss": 1.0931, "step": 3046 }, { "epoch": 0.20645030151094249, "grad_norm": 8.765487670898438, "learning_rate": 9.685673215141352e-05, "loss": 1.0522, "step": 3047 }, { "epoch": 0.20651805677891455, "grad_norm": 7.717139720916748, "learning_rate": 9.68553631323157e-05, "loss": 0.9492, "step": 3048 }, { "epoch": 0.20658581204688664, "grad_norm": 9.301026344299316, "learning_rate": 9.685399411321788e-05, "loss": 1.0191, "step": 3049 }, { "epoch": 0.20665356731485873, "grad_norm": 10.251668930053711, "learning_rate": 9.685262509412006e-05, "loss": 0.991, "step": 3050 }, { "epoch": 0.20672132258283082, "grad_norm": 11.597551345825195, "learning_rate": 9.685125607502226e-05, "loss": 1.0169, "step": 3051 }, { "epoch": 0.2067890778508029, "grad_norm": 10.293901443481445, "learning_rate": 9.684988705592444e-05, "loss": 1.0623, "step": 3052 }, { "epoch": 0.20685683311877498, "grad_norm": 8.496854782104492, "learning_rate": 9.684851803682662e-05, "loss": 1.0466, "step": 3053 }, { "epoch": 0.20692458838674707, "grad_norm": 6.76383638381958, "learning_rate": 9.68471490177288e-05, "loss": 0.8155, "step": 3054 }, { "epoch": 0.20699234365471916, "grad_norm": 8.168519020080566, "learning_rate": 9.684577999863098e-05, "loss": 0.9663, "step": 3055 }, { "epoch": 0.20706009892269123, "grad_norm": 8.0086030960083, "learning_rate": 9.684441097953317e-05, "loss": 0.889, "step": 3056 }, { "epoch": 0.20712785419066332, "grad_norm": 8.347359657287598, "learning_rate": 9.684304196043535e-05, "loss": 0.938, "step": 3057 }, { "epoch": 0.2071956094586354, "grad_norm": 8.0283203125, "learning_rate": 9.684167294133753e-05, "loss": 0.7633, "step": 3058 }, { "epoch": 0.2072633647266075, "grad_norm": 8.509317398071289, "learning_rate": 9.684030392223971e-05, "loss": 1.1052, "step": 3059 }, { "epoch": 0.20733111999457957, "grad_norm": 9.441505432128906, "learning_rate": 9.683893490314189e-05, "loss": 1.1177, "step": 3060 }, { "epoch": 0.20739887526255166, "grad_norm": 8.131098747253418, "learning_rate": 9.683756588404409e-05, "loss": 0.9272, "step": 3061 }, { "epoch": 0.20746663053052375, "grad_norm": 8.423643112182617, "learning_rate": 9.683619686494627e-05, "loss": 0.875, "step": 3062 }, { "epoch": 0.20753438579849584, "grad_norm": 7.513223171234131, "learning_rate": 9.683482784584845e-05, "loss": 1.0659, "step": 3063 }, { "epoch": 0.2076021410664679, "grad_norm": 7.5881171226501465, "learning_rate": 9.683345882675063e-05, "loss": 0.9466, "step": 3064 }, { "epoch": 0.20766989633444, "grad_norm": 8.717775344848633, "learning_rate": 9.683208980765282e-05, "loss": 1.1019, "step": 3065 }, { "epoch": 0.2077376516024121, "grad_norm": 11.917694091796875, "learning_rate": 9.6830720788555e-05, "loss": 1.1085, "step": 3066 }, { "epoch": 0.20780540687038418, "grad_norm": 9.28741455078125, "learning_rate": 9.682935176945718e-05, "loss": 1.1458, "step": 3067 }, { "epoch": 0.20787316213835624, "grad_norm": 6.746860980987549, "learning_rate": 9.682798275035938e-05, "loss": 0.929, "step": 3068 }, { "epoch": 0.20794091740632833, "grad_norm": 8.48763370513916, "learning_rate": 9.682661373126156e-05, "loss": 1.0702, "step": 3069 }, { "epoch": 0.20800867267430043, "grad_norm": 8.871308326721191, "learning_rate": 9.682524471216374e-05, "loss": 0.9199, "step": 3070 }, { "epoch": 0.20807642794227252, "grad_norm": 8.275801658630371, "learning_rate": 9.682387569306593e-05, "loss": 1.0756, "step": 3071 }, { "epoch": 0.2081441832102446, "grad_norm": 8.985222816467285, "learning_rate": 9.682250667396811e-05, "loss": 1.2364, "step": 3072 }, { "epoch": 0.20821193847821667, "grad_norm": 7.749682426452637, "learning_rate": 9.682113765487029e-05, "loss": 0.9444, "step": 3073 }, { "epoch": 0.20827969374618877, "grad_norm": 9.111614227294922, "learning_rate": 9.681976863577248e-05, "loss": 1.338, "step": 3074 }, { "epoch": 0.20834744901416086, "grad_norm": 7.2874674797058105, "learning_rate": 9.681839961667466e-05, "loss": 0.838, "step": 3075 }, { "epoch": 0.20841520428213295, "grad_norm": 6.873099327087402, "learning_rate": 9.681703059757685e-05, "loss": 0.9194, "step": 3076 }, { "epoch": 0.208482959550105, "grad_norm": 7.564418792724609, "learning_rate": 9.681566157847903e-05, "loss": 1.1209, "step": 3077 }, { "epoch": 0.2085507148180771, "grad_norm": 9.088560104370117, "learning_rate": 9.68142925593812e-05, "loss": 0.8425, "step": 3078 }, { "epoch": 0.2086184700860492, "grad_norm": 7.182369709014893, "learning_rate": 9.68129235402834e-05, "loss": 1.0938, "step": 3079 }, { "epoch": 0.2086862253540213, "grad_norm": 8.853677749633789, "learning_rate": 9.681155452118558e-05, "loss": 1.2611, "step": 3080 }, { "epoch": 0.20875398062199335, "grad_norm": 8.56440258026123, "learning_rate": 9.681018550208776e-05, "loss": 1.1112, "step": 3081 }, { "epoch": 0.20882173588996544, "grad_norm": 8.356021881103516, "learning_rate": 9.680881648298994e-05, "loss": 1.0044, "step": 3082 }, { "epoch": 0.20888949115793753, "grad_norm": 9.083736419677734, "learning_rate": 9.680744746389213e-05, "loss": 1.2065, "step": 3083 }, { "epoch": 0.20895724642590963, "grad_norm": 7.990222454071045, "learning_rate": 9.680607844479432e-05, "loss": 1.0179, "step": 3084 }, { "epoch": 0.2090250016938817, "grad_norm": 8.381364822387695, "learning_rate": 9.68047094256965e-05, "loss": 0.8392, "step": 3085 }, { "epoch": 0.20909275696185378, "grad_norm": 9.017950057983398, "learning_rate": 9.680334040659868e-05, "loss": 0.9757, "step": 3086 }, { "epoch": 0.20916051222982587, "grad_norm": 8.525566101074219, "learning_rate": 9.680197138750086e-05, "loss": 1.1521, "step": 3087 }, { "epoch": 0.20922826749779797, "grad_norm": 8.432148933410645, "learning_rate": 9.680060236840305e-05, "loss": 0.8559, "step": 3088 }, { "epoch": 0.20929602276577003, "grad_norm": 9.985367774963379, "learning_rate": 9.679923334930523e-05, "loss": 1.2035, "step": 3089 }, { "epoch": 0.20936377803374212, "grad_norm": 9.771974563598633, "learning_rate": 9.679786433020741e-05, "loss": 1.1614, "step": 3090 }, { "epoch": 0.2094315333017142, "grad_norm": 10.646146774291992, "learning_rate": 9.679649531110959e-05, "loss": 1.2165, "step": 3091 }, { "epoch": 0.2094992885696863, "grad_norm": 8.93340015411377, "learning_rate": 9.679512629201177e-05, "loss": 0.9883, "step": 3092 }, { "epoch": 0.20956704383765837, "grad_norm": 8.498669624328613, "learning_rate": 9.679375727291397e-05, "loss": 1.0318, "step": 3093 }, { "epoch": 0.20963479910563046, "grad_norm": 7.536258220672607, "learning_rate": 9.679238825381615e-05, "loss": 0.9949, "step": 3094 }, { "epoch": 0.20970255437360255, "grad_norm": 8.314896583557129, "learning_rate": 9.679101923471833e-05, "loss": 1.0502, "step": 3095 }, { "epoch": 0.20977030964157464, "grad_norm": 7.470542907714844, "learning_rate": 9.67896502156205e-05, "loss": 0.8871, "step": 3096 }, { "epoch": 0.2098380649095467, "grad_norm": 8.951095581054688, "learning_rate": 9.67882811965227e-05, "loss": 0.9793, "step": 3097 }, { "epoch": 0.2099058201775188, "grad_norm": 7.879035472869873, "learning_rate": 9.678691217742488e-05, "loss": 1.0193, "step": 3098 }, { "epoch": 0.2099735754454909, "grad_norm": 8.890814781188965, "learning_rate": 9.678554315832706e-05, "loss": 0.9738, "step": 3099 }, { "epoch": 0.21004133071346298, "grad_norm": 8.863816261291504, "learning_rate": 9.678417413922924e-05, "loss": 1.1213, "step": 3100 }, { "epoch": 0.21010908598143505, "grad_norm": 9.59538745880127, "learning_rate": 9.678280512013142e-05, "loss": 1.0215, "step": 3101 }, { "epoch": 0.21017684124940714, "grad_norm": 8.811614990234375, "learning_rate": 9.678143610103362e-05, "loss": 0.8703, "step": 3102 }, { "epoch": 0.21024459651737923, "grad_norm": 7.274720191955566, "learning_rate": 9.67800670819358e-05, "loss": 0.898, "step": 3103 }, { "epoch": 0.21031235178535132, "grad_norm": 11.239364624023438, "learning_rate": 9.677869806283798e-05, "loss": 0.9823, "step": 3104 }, { "epoch": 0.21038010705332338, "grad_norm": 8.807086944580078, "learning_rate": 9.677732904374016e-05, "loss": 1.0074, "step": 3105 }, { "epoch": 0.21044786232129548, "grad_norm": 9.065536499023438, "learning_rate": 9.677596002464235e-05, "loss": 1.1861, "step": 3106 }, { "epoch": 0.21051561758926757, "grad_norm": 11.035104751586914, "learning_rate": 9.677459100554453e-05, "loss": 1.0067, "step": 3107 }, { "epoch": 0.21058337285723966, "grad_norm": 8.010696411132812, "learning_rate": 9.677322198644671e-05, "loss": 1.0855, "step": 3108 }, { "epoch": 0.21065112812521172, "grad_norm": 9.104195594787598, "learning_rate": 9.677185296734889e-05, "loss": 1.0497, "step": 3109 }, { "epoch": 0.21071888339318381, "grad_norm": 8.731512069702148, "learning_rate": 9.677048394825107e-05, "loss": 1.1108, "step": 3110 }, { "epoch": 0.2107866386611559, "grad_norm": 8.823514938354492, "learning_rate": 9.676911492915327e-05, "loss": 1.0271, "step": 3111 }, { "epoch": 0.210854393929128, "grad_norm": 7.446425914764404, "learning_rate": 9.676774591005545e-05, "loss": 0.9182, "step": 3112 }, { "epoch": 0.21092214919710006, "grad_norm": 10.75915241241455, "learning_rate": 9.676637689095763e-05, "loss": 0.93, "step": 3113 }, { "epoch": 0.21098990446507215, "grad_norm": 10.065240859985352, "learning_rate": 9.676500787185982e-05, "loss": 1.0508, "step": 3114 }, { "epoch": 0.21105765973304424, "grad_norm": 8.790117263793945, "learning_rate": 9.6763638852762e-05, "loss": 0.8971, "step": 3115 }, { "epoch": 0.21112541500101634, "grad_norm": 8.286596298217773, "learning_rate": 9.676226983366418e-05, "loss": 0.9403, "step": 3116 }, { "epoch": 0.2111931702689884, "grad_norm": 8.542399406433105, "learning_rate": 9.676090081456637e-05, "loss": 1.1638, "step": 3117 }, { "epoch": 0.2112609255369605, "grad_norm": 9.461727142333984, "learning_rate": 9.675953179546856e-05, "loss": 1.0859, "step": 3118 }, { "epoch": 0.21132868080493258, "grad_norm": 7.0917229652404785, "learning_rate": 9.675816277637074e-05, "loss": 0.9931, "step": 3119 }, { "epoch": 0.21139643607290468, "grad_norm": 9.192744255065918, "learning_rate": 9.675679375727293e-05, "loss": 1.2138, "step": 3120 }, { "epoch": 0.21146419134087674, "grad_norm": 7.744256973266602, "learning_rate": 9.675542473817511e-05, "loss": 1.3615, "step": 3121 }, { "epoch": 0.21153194660884883, "grad_norm": 6.915426254272461, "learning_rate": 9.675405571907729e-05, "loss": 0.89, "step": 3122 }, { "epoch": 0.21159970187682092, "grad_norm": 9.888227462768555, "learning_rate": 9.675268669997947e-05, "loss": 0.9569, "step": 3123 }, { "epoch": 0.21166745714479301, "grad_norm": 7.541590690612793, "learning_rate": 9.675131768088165e-05, "loss": 0.9638, "step": 3124 }, { "epoch": 0.2117352124127651, "grad_norm": 7.883132457733154, "learning_rate": 9.674994866178384e-05, "loss": 0.8901, "step": 3125 }, { "epoch": 0.21180296768073717, "grad_norm": 8.493675231933594, "learning_rate": 9.674857964268602e-05, "loss": 0.8894, "step": 3126 }, { "epoch": 0.21187072294870926, "grad_norm": 7.348284721374512, "learning_rate": 9.67472106235882e-05, "loss": 0.8721, "step": 3127 }, { "epoch": 0.21193847821668135, "grad_norm": 9.094710350036621, "learning_rate": 9.674584160449039e-05, "loss": 0.9255, "step": 3128 }, { "epoch": 0.21200623348465344, "grad_norm": 7.316446304321289, "learning_rate": 9.674447258539258e-05, "loss": 1.2577, "step": 3129 }, { "epoch": 0.2120739887526255, "grad_norm": 8.105271339416504, "learning_rate": 9.674310356629476e-05, "loss": 0.8157, "step": 3130 }, { "epoch": 0.2121417440205976, "grad_norm": 8.433457374572754, "learning_rate": 9.674173454719694e-05, "loss": 1.0398, "step": 3131 }, { "epoch": 0.2122094992885697, "grad_norm": 7.787237644195557, "learning_rate": 9.674036552809912e-05, "loss": 0.7095, "step": 3132 }, { "epoch": 0.21227725455654178, "grad_norm": 10.60180950164795, "learning_rate": 9.67389965090013e-05, "loss": 1.0482, "step": 3133 }, { "epoch": 0.21234500982451385, "grad_norm": 8.428773880004883, "learning_rate": 9.67376274899035e-05, "loss": 0.8677, "step": 3134 }, { "epoch": 0.21241276509248594, "grad_norm": 8.204195022583008, "learning_rate": 9.673625847080568e-05, "loss": 1.14, "step": 3135 }, { "epoch": 0.21248052036045803, "grad_norm": 6.449087619781494, "learning_rate": 9.673488945170786e-05, "loss": 1.0181, "step": 3136 }, { "epoch": 0.21254827562843012, "grad_norm": 9.588041305541992, "learning_rate": 9.673352043261004e-05, "loss": 0.9996, "step": 3137 }, { "epoch": 0.21261603089640219, "grad_norm": 9.626228332519531, "learning_rate": 9.673215141351223e-05, "loss": 1.0501, "step": 3138 }, { "epoch": 0.21268378616437428, "grad_norm": 9.036309242248535, "learning_rate": 9.673078239441441e-05, "loss": 1.093, "step": 3139 }, { "epoch": 0.21275154143234637, "grad_norm": 9.415257453918457, "learning_rate": 9.672941337531659e-05, "loss": 1.0983, "step": 3140 }, { "epoch": 0.21281929670031846, "grad_norm": 8.309000015258789, "learning_rate": 9.672804435621877e-05, "loss": 1.0803, "step": 3141 }, { "epoch": 0.21288705196829052, "grad_norm": 7.420774459838867, "learning_rate": 9.672667533712095e-05, "loss": 1.1027, "step": 3142 }, { "epoch": 0.21295480723626262, "grad_norm": 9.291664123535156, "learning_rate": 9.672530631802314e-05, "loss": 1.1617, "step": 3143 }, { "epoch": 0.2130225625042347, "grad_norm": 7.656317234039307, "learning_rate": 9.672393729892533e-05, "loss": 0.9423, "step": 3144 }, { "epoch": 0.2130903177722068, "grad_norm": 5.812994003295898, "learning_rate": 9.67225682798275e-05, "loss": 0.9187, "step": 3145 }, { "epoch": 0.21315807304017886, "grad_norm": 9.492706298828125, "learning_rate": 9.672119926072969e-05, "loss": 1.1804, "step": 3146 }, { "epoch": 0.21322582830815096, "grad_norm": 9.37720012664795, "learning_rate": 9.671983024163187e-05, "loss": 0.8885, "step": 3147 }, { "epoch": 0.21329358357612305, "grad_norm": 9.183650016784668, "learning_rate": 9.671846122253406e-05, "loss": 0.8967, "step": 3148 }, { "epoch": 0.21336133884409514, "grad_norm": 8.037003517150879, "learning_rate": 9.671709220343624e-05, "loss": 1.2605, "step": 3149 }, { "epoch": 0.2134290941120672, "grad_norm": 8.694345474243164, "learning_rate": 9.671572318433842e-05, "loss": 1.003, "step": 3150 }, { "epoch": 0.2134968493800393, "grad_norm": 6.429176330566406, "learning_rate": 9.67143541652406e-05, "loss": 1.1018, "step": 3151 }, { "epoch": 0.21356460464801139, "grad_norm": 7.964774131774902, "learning_rate": 9.67129851461428e-05, "loss": 0.9121, "step": 3152 }, { "epoch": 0.21363235991598348, "grad_norm": 7.5509033203125, "learning_rate": 9.671161612704498e-05, "loss": 0.9577, "step": 3153 }, { "epoch": 0.21370011518395554, "grad_norm": 7.968616962432861, "learning_rate": 9.671024710794716e-05, "loss": 1.1337, "step": 3154 }, { "epoch": 0.21376787045192763, "grad_norm": 7.340758800506592, "learning_rate": 9.670887808884934e-05, "loss": 0.8068, "step": 3155 }, { "epoch": 0.21383562571989972, "grad_norm": 7.890449523925781, "learning_rate": 9.670750906975152e-05, "loss": 1.0263, "step": 3156 }, { "epoch": 0.21390338098787182, "grad_norm": 10.908242225646973, "learning_rate": 9.670614005065371e-05, "loss": 1.0384, "step": 3157 }, { "epoch": 0.21397113625584388, "grad_norm": 7.62697696685791, "learning_rate": 9.670477103155589e-05, "loss": 0.918, "step": 3158 }, { "epoch": 0.21403889152381597, "grad_norm": 10.487192153930664, "learning_rate": 9.670340201245807e-05, "loss": 1.1135, "step": 3159 }, { "epoch": 0.21410664679178806, "grad_norm": 7.076718807220459, "learning_rate": 9.670203299336026e-05, "loss": 0.9286, "step": 3160 }, { "epoch": 0.21417440205976015, "grad_norm": 10.05949592590332, "learning_rate": 9.670066397426245e-05, "loss": 1.1317, "step": 3161 }, { "epoch": 0.21424215732773222, "grad_norm": 8.656492233276367, "learning_rate": 9.669929495516463e-05, "loss": 0.9864, "step": 3162 }, { "epoch": 0.2143099125957043, "grad_norm": 10.591787338256836, "learning_rate": 9.669792593606682e-05, "loss": 1.0801, "step": 3163 }, { "epoch": 0.2143776678636764, "grad_norm": 8.19133472442627, "learning_rate": 9.6696556916969e-05, "loss": 0.9377, "step": 3164 }, { "epoch": 0.2144454231316485, "grad_norm": 8.402669906616211, "learning_rate": 9.669518789787118e-05, "loss": 0.9283, "step": 3165 }, { "epoch": 0.21451317839962056, "grad_norm": 9.713315963745117, "learning_rate": 9.669381887877337e-05, "loss": 1.1865, "step": 3166 }, { "epoch": 0.21458093366759265, "grad_norm": 6.777700424194336, "learning_rate": 9.669244985967555e-05, "loss": 0.9094, "step": 3167 }, { "epoch": 0.21464868893556474, "grad_norm": 9.381082534790039, "learning_rate": 9.669108084057773e-05, "loss": 0.7461, "step": 3168 }, { "epoch": 0.21471644420353683, "grad_norm": 8.568790435791016, "learning_rate": 9.668971182147992e-05, "loss": 0.7339, "step": 3169 }, { "epoch": 0.2147841994715089, "grad_norm": 9.257226943969727, "learning_rate": 9.66883428023821e-05, "loss": 1.1892, "step": 3170 }, { "epoch": 0.214851954739481, "grad_norm": 8.544146537780762, "learning_rate": 9.668697378328429e-05, "loss": 1.0307, "step": 3171 }, { "epoch": 0.21491971000745308, "grad_norm": 9.540971755981445, "learning_rate": 9.668560476418647e-05, "loss": 0.8464, "step": 3172 }, { "epoch": 0.21498746527542517, "grad_norm": 9.180089950561523, "learning_rate": 9.668423574508865e-05, "loss": 0.951, "step": 3173 }, { "epoch": 0.21505522054339726, "grad_norm": 10.706409454345703, "learning_rate": 9.668286672599083e-05, "loss": 1.141, "step": 3174 }, { "epoch": 0.21512297581136933, "grad_norm": 9.421865463256836, "learning_rate": 9.668149770689302e-05, "loss": 1.0701, "step": 3175 }, { "epoch": 0.21519073107934142, "grad_norm": 9.120182037353516, "learning_rate": 9.66801286877952e-05, "loss": 1.1618, "step": 3176 }, { "epoch": 0.2152584863473135, "grad_norm": 7.879681587219238, "learning_rate": 9.667875966869738e-05, "loss": 0.8048, "step": 3177 }, { "epoch": 0.2153262416152856, "grad_norm": 7.233819007873535, "learning_rate": 9.667739064959957e-05, "loss": 0.9718, "step": 3178 }, { "epoch": 0.21539399688325767, "grad_norm": 7.79316520690918, "learning_rate": 9.667602163050175e-05, "loss": 1.0999, "step": 3179 }, { "epoch": 0.21546175215122976, "grad_norm": 8.873052597045898, "learning_rate": 9.667465261140394e-05, "loss": 0.9444, "step": 3180 }, { "epoch": 0.21552950741920185, "grad_norm": 7.386806488037109, "learning_rate": 9.667328359230612e-05, "loss": 0.8661, "step": 3181 }, { "epoch": 0.21559726268717394, "grad_norm": 7.682179927825928, "learning_rate": 9.66719145732083e-05, "loss": 0.7904, "step": 3182 }, { "epoch": 0.215665017955146, "grad_norm": 7.9192914962768555, "learning_rate": 9.667054555411048e-05, "loss": 1.0563, "step": 3183 }, { "epoch": 0.2157327732231181, "grad_norm": 6.972530364990234, "learning_rate": 9.666917653501267e-05, "loss": 0.7482, "step": 3184 }, { "epoch": 0.2158005284910902, "grad_norm": 9.844091415405273, "learning_rate": 9.666780751591485e-05, "loss": 1.0289, "step": 3185 }, { "epoch": 0.21586828375906228, "grad_norm": 8.724159240722656, "learning_rate": 9.666643849681704e-05, "loss": 1.1419, "step": 3186 }, { "epoch": 0.21593603902703434, "grad_norm": 8.859795570373535, "learning_rate": 9.666506947771922e-05, "loss": 0.9252, "step": 3187 }, { "epoch": 0.21600379429500643, "grad_norm": 10.741375923156738, "learning_rate": 9.66637004586214e-05, "loss": 1.2328, "step": 3188 }, { "epoch": 0.21607154956297853, "grad_norm": 8.468697547912598, "learning_rate": 9.666233143952359e-05, "loss": 1.1779, "step": 3189 }, { "epoch": 0.21613930483095062, "grad_norm": 7.220376014709473, "learning_rate": 9.666096242042577e-05, "loss": 0.9271, "step": 3190 }, { "epoch": 0.21620706009892268, "grad_norm": 8.3795804977417, "learning_rate": 9.665959340132795e-05, "loss": 1.0941, "step": 3191 }, { "epoch": 0.21627481536689477, "grad_norm": 9.743870735168457, "learning_rate": 9.665822438223013e-05, "loss": 0.9396, "step": 3192 }, { "epoch": 0.21634257063486687, "grad_norm": 8.927998542785645, "learning_rate": 9.665685536313231e-05, "loss": 1.0915, "step": 3193 }, { "epoch": 0.21641032590283896, "grad_norm": 8.214877128601074, "learning_rate": 9.66554863440345e-05, "loss": 0.8767, "step": 3194 }, { "epoch": 0.21647808117081102, "grad_norm": 9.605053901672363, "learning_rate": 9.665411732493669e-05, "loss": 1.1829, "step": 3195 }, { "epoch": 0.2165458364387831, "grad_norm": 8.430122375488281, "learning_rate": 9.665274830583887e-05, "loss": 0.8234, "step": 3196 }, { "epoch": 0.2166135917067552, "grad_norm": 8.868674278259277, "learning_rate": 9.665137928674105e-05, "loss": 1.0467, "step": 3197 }, { "epoch": 0.2166813469747273, "grad_norm": 8.451126098632812, "learning_rate": 9.665001026764324e-05, "loss": 1.134, "step": 3198 }, { "epoch": 0.21674910224269936, "grad_norm": 8.749373435974121, "learning_rate": 9.664864124854542e-05, "loss": 0.9825, "step": 3199 }, { "epoch": 0.21681685751067145, "grad_norm": 8.478997230529785, "learning_rate": 9.66472722294476e-05, "loss": 0.9823, "step": 3200 }, { "epoch": 0.21688461277864354, "grad_norm": 6.554266929626465, "learning_rate": 9.664590321034978e-05, "loss": 1.2293, "step": 3201 }, { "epoch": 0.21695236804661563, "grad_norm": 7.415618896484375, "learning_rate": 9.664453419125196e-05, "loss": 0.9767, "step": 3202 }, { "epoch": 0.2170201233145877, "grad_norm": 9.277578353881836, "learning_rate": 9.664316517215416e-05, "loss": 1.0042, "step": 3203 }, { "epoch": 0.2170878785825598, "grad_norm": 7.1379241943359375, "learning_rate": 9.664179615305634e-05, "loss": 0.924, "step": 3204 }, { "epoch": 0.21715563385053188, "grad_norm": 8.837613105773926, "learning_rate": 9.664042713395852e-05, "loss": 0.9889, "step": 3205 }, { "epoch": 0.21722338911850397, "grad_norm": 8.118345260620117, "learning_rate": 9.663905811486071e-05, "loss": 1.2376, "step": 3206 }, { "epoch": 0.21729114438647604, "grad_norm": 10.383713722229004, "learning_rate": 9.663768909576289e-05, "loss": 0.7988, "step": 3207 }, { "epoch": 0.21735889965444813, "grad_norm": 7.0938262939453125, "learning_rate": 9.663632007666507e-05, "loss": 0.8589, "step": 3208 }, { "epoch": 0.21742665492242022, "grad_norm": 7.732020854949951, "learning_rate": 9.663495105756726e-05, "loss": 1.0837, "step": 3209 }, { "epoch": 0.2174944101903923, "grad_norm": 7.718788146972656, "learning_rate": 9.663358203846944e-05, "loss": 0.8949, "step": 3210 }, { "epoch": 0.21756216545836438, "grad_norm": 9.187674522399902, "learning_rate": 9.663221301937162e-05, "loss": 0.9443, "step": 3211 }, { "epoch": 0.21762992072633647, "grad_norm": 8.596949577331543, "learning_rate": 9.663084400027382e-05, "loss": 1.0855, "step": 3212 }, { "epoch": 0.21769767599430856, "grad_norm": 7.854368209838867, "learning_rate": 9.6629474981176e-05, "loss": 0.9203, "step": 3213 }, { "epoch": 0.21776543126228065, "grad_norm": 9.96597957611084, "learning_rate": 9.662810596207818e-05, "loss": 0.836, "step": 3214 }, { "epoch": 0.21783318653025271, "grad_norm": 7.254636764526367, "learning_rate": 9.662673694298036e-05, "loss": 0.7926, "step": 3215 }, { "epoch": 0.2179009417982248, "grad_norm": 9.079703330993652, "learning_rate": 9.662536792388255e-05, "loss": 1.2072, "step": 3216 }, { "epoch": 0.2179686970661969, "grad_norm": 8.051166534423828, "learning_rate": 9.662399890478473e-05, "loss": 0.9719, "step": 3217 }, { "epoch": 0.218036452334169, "grad_norm": 9.189994812011719, "learning_rate": 9.662262988568691e-05, "loss": 0.9878, "step": 3218 }, { "epoch": 0.21810420760214105, "grad_norm": 9.806111335754395, "learning_rate": 9.66212608665891e-05, "loss": 1.1621, "step": 3219 }, { "epoch": 0.21817196287011315, "grad_norm": 7.833159923553467, "learning_rate": 9.661989184749128e-05, "loss": 0.9811, "step": 3220 }, { "epoch": 0.21823971813808524, "grad_norm": 7.537334442138672, "learning_rate": 9.661852282839347e-05, "loss": 0.9916, "step": 3221 }, { "epoch": 0.21830747340605733, "grad_norm": 9.828896522521973, "learning_rate": 9.661715380929565e-05, "loss": 1.09, "step": 3222 }, { "epoch": 0.2183752286740294, "grad_norm": 7.99267578125, "learning_rate": 9.661578479019783e-05, "loss": 1.0852, "step": 3223 }, { "epoch": 0.21844298394200148, "grad_norm": 8.085976600646973, "learning_rate": 9.661441577110001e-05, "loss": 1.1762, "step": 3224 }, { "epoch": 0.21851073920997358, "grad_norm": 6.523219585418701, "learning_rate": 9.661304675200219e-05, "loss": 1.0814, "step": 3225 }, { "epoch": 0.21857849447794567, "grad_norm": 9.796062469482422, "learning_rate": 9.661167773290438e-05, "loss": 1.1447, "step": 3226 }, { "epoch": 0.21864624974591776, "grad_norm": 8.570687294006348, "learning_rate": 9.661030871380656e-05, "loss": 1.0552, "step": 3227 }, { "epoch": 0.21871400501388982, "grad_norm": 9.70113468170166, "learning_rate": 9.660893969470874e-05, "loss": 1.0379, "step": 3228 }, { "epoch": 0.21878176028186191, "grad_norm": 8.195375442504883, "learning_rate": 9.660757067561093e-05, "loss": 0.9348, "step": 3229 }, { "epoch": 0.218849515549834, "grad_norm": 7.33494758605957, "learning_rate": 9.660620165651312e-05, "loss": 0.8977, "step": 3230 }, { "epoch": 0.2189172708178061, "grad_norm": 7.543430328369141, "learning_rate": 9.66048326374153e-05, "loss": 1.023, "step": 3231 }, { "epoch": 0.21898502608577816, "grad_norm": 8.45494556427002, "learning_rate": 9.660346361831748e-05, "loss": 1.2995, "step": 3232 }, { "epoch": 0.21905278135375025, "grad_norm": 8.197903633117676, "learning_rate": 9.660209459921966e-05, "loss": 0.9651, "step": 3233 }, { "epoch": 0.21912053662172234, "grad_norm": 8.392546653747559, "learning_rate": 9.660072558012184e-05, "loss": 0.9858, "step": 3234 }, { "epoch": 0.21918829188969444, "grad_norm": 6.499540328979492, "learning_rate": 9.659935656102403e-05, "loss": 0.9679, "step": 3235 }, { "epoch": 0.2192560471576665, "grad_norm": 7.784618377685547, "learning_rate": 9.659798754192621e-05, "loss": 0.9316, "step": 3236 }, { "epoch": 0.2193238024256386, "grad_norm": 9.014801979064941, "learning_rate": 9.65966185228284e-05, "loss": 0.9261, "step": 3237 }, { "epoch": 0.21939155769361068, "grad_norm": 8.238675117492676, "learning_rate": 9.659524950373058e-05, "loss": 1.0156, "step": 3238 }, { "epoch": 0.21945931296158278, "grad_norm": 9.447864532470703, "learning_rate": 9.659388048463277e-05, "loss": 1.2186, "step": 3239 }, { "epoch": 0.21952706822955484, "grad_norm": 6.894035816192627, "learning_rate": 9.659251146553495e-05, "loss": 0.9157, "step": 3240 }, { "epoch": 0.21959482349752693, "grad_norm": 7.24024772644043, "learning_rate": 9.659114244643713e-05, "loss": 1.1564, "step": 3241 }, { "epoch": 0.21966257876549902, "grad_norm": 9.135485649108887, "learning_rate": 9.658977342733931e-05, "loss": 1.2002, "step": 3242 }, { "epoch": 0.21973033403347111, "grad_norm": 6.119181156158447, "learning_rate": 9.658840440824149e-05, "loss": 0.788, "step": 3243 }, { "epoch": 0.21979808930144318, "grad_norm": 8.550108909606934, "learning_rate": 9.658703538914368e-05, "loss": 0.9251, "step": 3244 }, { "epoch": 0.21986584456941527, "grad_norm": 9.546792984008789, "learning_rate": 9.658566637004586e-05, "loss": 1.2231, "step": 3245 }, { "epoch": 0.21993359983738736, "grad_norm": 9.451306343078613, "learning_rate": 9.658429735094805e-05, "loss": 0.8991, "step": 3246 }, { "epoch": 0.22000135510535945, "grad_norm": 9.914873123168945, "learning_rate": 9.658292833185023e-05, "loss": 1.3102, "step": 3247 }, { "epoch": 0.22006911037333152, "grad_norm": 10.038002967834473, "learning_rate": 9.65815593127524e-05, "loss": 1.254, "step": 3248 }, { "epoch": 0.2201368656413036, "grad_norm": 7.590802192687988, "learning_rate": 9.65801902936546e-05, "loss": 1.0032, "step": 3249 }, { "epoch": 0.2202046209092757, "grad_norm": 8.632338523864746, "learning_rate": 9.657882127455678e-05, "loss": 0.8922, "step": 3250 }, { "epoch": 0.2202723761772478, "grad_norm": 6.483764171600342, "learning_rate": 9.657745225545896e-05, "loss": 0.9167, "step": 3251 }, { "epoch": 0.22034013144521986, "grad_norm": 8.28577709197998, "learning_rate": 9.657608323636115e-05, "loss": 1.0413, "step": 3252 }, { "epoch": 0.22040788671319195, "grad_norm": 9.218433380126953, "learning_rate": 9.657471421726333e-05, "loss": 1.1681, "step": 3253 }, { "epoch": 0.22047564198116404, "grad_norm": 6.653755187988281, "learning_rate": 9.657334519816552e-05, "loss": 0.9645, "step": 3254 }, { "epoch": 0.22054339724913613, "grad_norm": 8.784723281860352, "learning_rate": 9.657197617906771e-05, "loss": 0.7489, "step": 3255 }, { "epoch": 0.2206111525171082, "grad_norm": 9.041324615478516, "learning_rate": 9.657060715996989e-05, "loss": 1.0478, "step": 3256 }, { "epoch": 0.22067890778508029, "grad_norm": 7.811746597290039, "learning_rate": 9.656923814087207e-05, "loss": 0.9511, "step": 3257 }, { "epoch": 0.22074666305305238, "grad_norm": 9.811897277832031, "learning_rate": 9.656786912177426e-05, "loss": 1.4235, "step": 3258 }, { "epoch": 0.22081441832102447, "grad_norm": 7.313972473144531, "learning_rate": 9.656650010267644e-05, "loss": 1.0075, "step": 3259 }, { "epoch": 0.22088217358899653, "grad_norm": 9.360795021057129, "learning_rate": 9.656513108357862e-05, "loss": 0.9634, "step": 3260 }, { "epoch": 0.22094992885696862, "grad_norm": 9.322724342346191, "learning_rate": 9.65637620644808e-05, "loss": 1.0669, "step": 3261 }, { "epoch": 0.22101768412494072, "grad_norm": 10.269414901733398, "learning_rate": 9.6562393045383e-05, "loss": 1.1288, "step": 3262 }, { "epoch": 0.2210854393929128, "grad_norm": 7.672987937927246, "learning_rate": 9.656102402628518e-05, "loss": 1.0517, "step": 3263 }, { "epoch": 0.22115319466088487, "grad_norm": 8.725695610046387, "learning_rate": 9.655965500718736e-05, "loss": 1.0987, "step": 3264 }, { "epoch": 0.22122094992885696, "grad_norm": 8.577162742614746, "learning_rate": 9.655828598808954e-05, "loss": 0.821, "step": 3265 }, { "epoch": 0.22128870519682906, "grad_norm": 8.497530937194824, "learning_rate": 9.655691696899172e-05, "loss": 0.9171, "step": 3266 }, { "epoch": 0.22135646046480115, "grad_norm": 9.665203094482422, "learning_rate": 9.655554794989391e-05, "loss": 1.105, "step": 3267 }, { "epoch": 0.2214242157327732, "grad_norm": 9.42651081085205, "learning_rate": 9.65541789307961e-05, "loss": 0.9526, "step": 3268 }, { "epoch": 0.2214919710007453, "grad_norm": 7.461474418640137, "learning_rate": 9.655280991169827e-05, "loss": 0.9554, "step": 3269 }, { "epoch": 0.2215597262687174, "grad_norm": 10.56189250946045, "learning_rate": 9.655144089260045e-05, "loss": 1.1504, "step": 3270 }, { "epoch": 0.22162748153668949, "grad_norm": 8.245696067810059, "learning_rate": 9.655007187350265e-05, "loss": 1.0352, "step": 3271 }, { "epoch": 0.22169523680466155, "grad_norm": 9.182881355285645, "learning_rate": 9.654870285440483e-05, "loss": 1.1153, "step": 3272 }, { "epoch": 0.22176299207263364, "grad_norm": 8.101743698120117, "learning_rate": 9.654733383530701e-05, "loss": 1.1399, "step": 3273 }, { "epoch": 0.22183074734060573, "grad_norm": 9.072972297668457, "learning_rate": 9.654596481620919e-05, "loss": 1.1508, "step": 3274 }, { "epoch": 0.22189850260857782, "grad_norm": 8.480910301208496, "learning_rate": 9.654459579711137e-05, "loss": 0.9659, "step": 3275 }, { "epoch": 0.2219662578765499, "grad_norm": 7.095332622528076, "learning_rate": 9.654322677801356e-05, "loss": 0.9417, "step": 3276 }, { "epoch": 0.22203401314452198, "grad_norm": 8.183112144470215, "learning_rate": 9.654185775891574e-05, "loss": 0.9307, "step": 3277 }, { "epoch": 0.22210176841249407, "grad_norm": 8.60648250579834, "learning_rate": 9.654048873981792e-05, "loss": 1.141, "step": 3278 }, { "epoch": 0.22216952368046616, "grad_norm": 6.992738246917725, "learning_rate": 9.65391197207201e-05, "loss": 0.9429, "step": 3279 }, { "epoch": 0.22223727894843825, "grad_norm": 6.998913764953613, "learning_rate": 9.653775070162229e-05, "loss": 1.0159, "step": 3280 }, { "epoch": 0.22230503421641032, "grad_norm": 7.922050952911377, "learning_rate": 9.653638168252448e-05, "loss": 1.1349, "step": 3281 }, { "epoch": 0.2223727894843824, "grad_norm": 9.171984672546387, "learning_rate": 9.653501266342666e-05, "loss": 1.149, "step": 3282 }, { "epoch": 0.2224405447523545, "grad_norm": 9.88930606842041, "learning_rate": 9.653364364432884e-05, "loss": 1.1036, "step": 3283 }, { "epoch": 0.2225083000203266, "grad_norm": 10.978727340698242, "learning_rate": 9.653227462523102e-05, "loss": 1.0345, "step": 3284 }, { "epoch": 0.22257605528829866, "grad_norm": 8.712138175964355, "learning_rate": 9.653090560613321e-05, "loss": 1.1404, "step": 3285 }, { "epoch": 0.22264381055627075, "grad_norm": 8.79491901397705, "learning_rate": 9.65295365870354e-05, "loss": 1.4334, "step": 3286 }, { "epoch": 0.22271156582424284, "grad_norm": 9.580904960632324, "learning_rate": 9.652816756793757e-05, "loss": 1.0693, "step": 3287 }, { "epoch": 0.22277932109221493, "grad_norm": 7.335433006286621, "learning_rate": 9.652679854883976e-05, "loss": 1.0559, "step": 3288 }, { "epoch": 0.222847076360187, "grad_norm": 8.534089088439941, "learning_rate": 9.652542952974194e-05, "loss": 0.9835, "step": 3289 }, { "epoch": 0.2229148316281591, "grad_norm": 6.626898288726807, "learning_rate": 9.652406051064413e-05, "loss": 0.8056, "step": 3290 }, { "epoch": 0.22298258689613118, "grad_norm": 7.774187088012695, "learning_rate": 9.652269149154631e-05, "loss": 0.8032, "step": 3291 }, { "epoch": 0.22305034216410327, "grad_norm": 8.029982566833496, "learning_rate": 9.652132247244849e-05, "loss": 1.0536, "step": 3292 }, { "epoch": 0.22311809743207534, "grad_norm": 8.701712608337402, "learning_rate": 9.651995345335067e-05, "loss": 0.9422, "step": 3293 }, { "epoch": 0.22318585270004743, "grad_norm": 9.17507553100586, "learning_rate": 9.651858443425286e-05, "loss": 1.0094, "step": 3294 }, { "epoch": 0.22325360796801952, "grad_norm": 8.464656829833984, "learning_rate": 9.651721541515504e-05, "loss": 0.9546, "step": 3295 }, { "epoch": 0.2233213632359916, "grad_norm": 6.839638710021973, "learning_rate": 9.651584639605722e-05, "loss": 0.8296, "step": 3296 }, { "epoch": 0.22338911850396367, "grad_norm": 7.773430824279785, "learning_rate": 9.65144773769594e-05, "loss": 1.1344, "step": 3297 }, { "epoch": 0.22345687377193577, "grad_norm": 9.725523948669434, "learning_rate": 9.65131083578616e-05, "loss": 0.9814, "step": 3298 }, { "epoch": 0.22352462903990786, "grad_norm": 8.995366096496582, "learning_rate": 9.651173933876378e-05, "loss": 1.2263, "step": 3299 }, { "epoch": 0.22359238430787995, "grad_norm": 8.735838890075684, "learning_rate": 9.651037031966596e-05, "loss": 1.1707, "step": 3300 }, { "epoch": 0.223660139575852, "grad_norm": 7.177740097045898, "learning_rate": 9.650900130056815e-05, "loss": 0.9491, "step": 3301 }, { "epoch": 0.2237278948438241, "grad_norm": 8.393296241760254, "learning_rate": 9.650763228147033e-05, "loss": 0.9765, "step": 3302 }, { "epoch": 0.2237956501117962, "grad_norm": 9.773447036743164, "learning_rate": 9.650626326237251e-05, "loss": 1.0409, "step": 3303 }, { "epoch": 0.2238634053797683, "grad_norm": 6.951483249664307, "learning_rate": 9.650489424327471e-05, "loss": 0.9848, "step": 3304 }, { "epoch": 0.22393116064774035, "grad_norm": 8.579717636108398, "learning_rate": 9.650352522417689e-05, "loss": 0.9722, "step": 3305 }, { "epoch": 0.22399891591571244, "grad_norm": 7.798871040344238, "learning_rate": 9.650215620507907e-05, "loss": 0.8046, "step": 3306 }, { "epoch": 0.22406667118368453, "grad_norm": 9.504851341247559, "learning_rate": 9.650078718598125e-05, "loss": 1.1972, "step": 3307 }, { "epoch": 0.22413442645165663, "grad_norm": 8.384056091308594, "learning_rate": 9.649941816688344e-05, "loss": 1.0654, "step": 3308 }, { "epoch": 0.2242021817196287, "grad_norm": 10.469207763671875, "learning_rate": 9.649804914778562e-05, "loss": 1.0694, "step": 3309 }, { "epoch": 0.22426993698760078, "grad_norm": 8.5567045211792, "learning_rate": 9.64966801286878e-05, "loss": 1.0371, "step": 3310 }, { "epoch": 0.22433769225557287, "grad_norm": 9.427155494689941, "learning_rate": 9.649531110958998e-05, "loss": 1.0185, "step": 3311 }, { "epoch": 0.22440544752354497, "grad_norm": 8.238412857055664, "learning_rate": 9.649394209049216e-05, "loss": 0.9744, "step": 3312 }, { "epoch": 0.22447320279151703, "grad_norm": 9.36423110961914, "learning_rate": 9.649257307139436e-05, "loss": 0.9184, "step": 3313 }, { "epoch": 0.22454095805948912, "grad_norm": 7.485356330871582, "learning_rate": 9.649120405229654e-05, "loss": 0.9948, "step": 3314 }, { "epoch": 0.2246087133274612, "grad_norm": 7.247208118438721, "learning_rate": 9.648983503319872e-05, "loss": 0.8592, "step": 3315 }, { "epoch": 0.2246764685954333, "grad_norm": 9.266085624694824, "learning_rate": 9.64884660141009e-05, "loss": 1.0314, "step": 3316 }, { "epoch": 0.22474422386340537, "grad_norm": 11.147274017333984, "learning_rate": 9.64870969950031e-05, "loss": 1.0403, "step": 3317 }, { "epoch": 0.22481197913137746, "grad_norm": 10.961670875549316, "learning_rate": 9.648572797590527e-05, "loss": 1.3147, "step": 3318 }, { "epoch": 0.22487973439934955, "grad_norm": 7.002868175506592, "learning_rate": 9.648435895680745e-05, "loss": 0.9044, "step": 3319 }, { "epoch": 0.22494748966732164, "grad_norm": 7.886247634887695, "learning_rate": 9.648298993770963e-05, "loss": 0.8901, "step": 3320 }, { "epoch": 0.2250152449352937, "grad_norm": 8.933539390563965, "learning_rate": 9.648162091861181e-05, "loss": 0.8758, "step": 3321 }, { "epoch": 0.2250830002032658, "grad_norm": 7.223681449890137, "learning_rate": 9.648025189951401e-05, "loss": 1.1502, "step": 3322 }, { "epoch": 0.2251507554712379, "grad_norm": 8.808623313903809, "learning_rate": 9.647888288041619e-05, "loss": 1.0706, "step": 3323 }, { "epoch": 0.22521851073920998, "grad_norm": 9.557942390441895, "learning_rate": 9.647751386131837e-05, "loss": 0.8113, "step": 3324 }, { "epoch": 0.22528626600718205, "grad_norm": 9.855717658996582, "learning_rate": 9.647614484222055e-05, "loss": 1.251, "step": 3325 }, { "epoch": 0.22535402127515414, "grad_norm": 9.588946342468262, "learning_rate": 9.647477582312273e-05, "loss": 1.3029, "step": 3326 }, { "epoch": 0.22542177654312623, "grad_norm": 9.277341842651367, "learning_rate": 9.647340680402492e-05, "loss": 0.9708, "step": 3327 }, { "epoch": 0.22548953181109832, "grad_norm": 9.607316970825195, "learning_rate": 9.64720377849271e-05, "loss": 0.9614, "step": 3328 }, { "epoch": 0.22555728707907038, "grad_norm": 7.949220657348633, "learning_rate": 9.647066876582928e-05, "loss": 1.0066, "step": 3329 }, { "epoch": 0.22562504234704248, "grad_norm": 8.383744239807129, "learning_rate": 9.646929974673146e-05, "loss": 0.797, "step": 3330 }, { "epoch": 0.22569279761501457, "grad_norm": 7.203914642333984, "learning_rate": 9.646793072763366e-05, "loss": 0.9625, "step": 3331 }, { "epoch": 0.22576055288298666, "grad_norm": 8.95102596282959, "learning_rate": 9.646656170853584e-05, "loss": 1.0728, "step": 3332 }, { "epoch": 0.22582830815095875, "grad_norm": 8.044751167297363, "learning_rate": 9.646519268943802e-05, "loss": 1.2155, "step": 3333 }, { "epoch": 0.22589606341893081, "grad_norm": 9.164902687072754, "learning_rate": 9.64638236703402e-05, "loss": 1.2923, "step": 3334 }, { "epoch": 0.2259638186869029, "grad_norm": 6.792164325714111, "learning_rate": 9.646245465124238e-05, "loss": 0.9956, "step": 3335 }, { "epoch": 0.226031573954875, "grad_norm": 7.76467752456665, "learning_rate": 9.646108563214457e-05, "loss": 1.0057, "step": 3336 }, { "epoch": 0.2260993292228471, "grad_norm": 8.541545867919922, "learning_rate": 9.645971661304675e-05, "loss": 1.1013, "step": 3337 }, { "epoch": 0.22616708449081915, "grad_norm": 7.8545050621032715, "learning_rate": 9.645834759394893e-05, "loss": 0.8565, "step": 3338 }, { "epoch": 0.22623483975879125, "grad_norm": 9.322896003723145, "learning_rate": 9.645697857485112e-05, "loss": 1.0509, "step": 3339 }, { "epoch": 0.22630259502676334, "grad_norm": 9.540191650390625, "learning_rate": 9.645560955575331e-05, "loss": 1.1757, "step": 3340 }, { "epoch": 0.22637035029473543, "grad_norm": 8.97028923034668, "learning_rate": 9.645424053665549e-05, "loss": 1.0342, "step": 3341 }, { "epoch": 0.2264381055627075, "grad_norm": 6.444105625152588, "learning_rate": 9.645287151755767e-05, "loss": 0.8377, "step": 3342 }, { "epoch": 0.22650586083067958, "grad_norm": 9.237889289855957, "learning_rate": 9.645150249845985e-05, "loss": 1.0468, "step": 3343 }, { "epoch": 0.22657361609865168, "grad_norm": 9.567046165466309, "learning_rate": 9.645013347936203e-05, "loss": 1.3875, "step": 3344 }, { "epoch": 0.22664137136662377, "grad_norm": 8.302481651306152, "learning_rate": 9.644876446026422e-05, "loss": 1.0455, "step": 3345 }, { "epoch": 0.22670912663459583, "grad_norm": 9.124307632446289, "learning_rate": 9.64473954411664e-05, "loss": 0.9828, "step": 3346 }, { "epoch": 0.22677688190256792, "grad_norm": 7.700011730194092, "learning_rate": 9.644602642206858e-05, "loss": 0.8475, "step": 3347 }, { "epoch": 0.22684463717054001, "grad_norm": 6.1064276695251465, "learning_rate": 9.644465740297078e-05, "loss": 0.7779, "step": 3348 }, { "epoch": 0.2269123924385121, "grad_norm": 8.924312591552734, "learning_rate": 9.644328838387296e-05, "loss": 1.3825, "step": 3349 }, { "epoch": 0.22698014770648417, "grad_norm": 8.169050216674805, "learning_rate": 9.644191936477514e-05, "loss": 1.1198, "step": 3350 }, { "epoch": 0.22704790297445626, "grad_norm": 8.368382453918457, "learning_rate": 9.644055034567733e-05, "loss": 1.0729, "step": 3351 }, { "epoch": 0.22711565824242835, "grad_norm": 10.206897735595703, "learning_rate": 9.643918132657951e-05, "loss": 0.9022, "step": 3352 }, { "epoch": 0.22718341351040044, "grad_norm": 6.288288116455078, "learning_rate": 9.64378123074817e-05, "loss": 0.7868, "step": 3353 }, { "epoch": 0.2272511687783725, "grad_norm": 8.446578025817871, "learning_rate": 9.643644328838389e-05, "loss": 1.0426, "step": 3354 }, { "epoch": 0.2273189240463446, "grad_norm": 8.425249099731445, "learning_rate": 9.643507426928607e-05, "loss": 0.823, "step": 3355 }, { "epoch": 0.2273866793143167, "grad_norm": 6.610576629638672, "learning_rate": 9.643370525018825e-05, "loss": 0.9251, "step": 3356 }, { "epoch": 0.22745443458228878, "grad_norm": 8.365503311157227, "learning_rate": 9.643233623109043e-05, "loss": 1.0258, "step": 3357 }, { "epoch": 0.22752218985026085, "grad_norm": 8.654791831970215, "learning_rate": 9.643096721199261e-05, "loss": 0.8781, "step": 3358 }, { "epoch": 0.22758994511823294, "grad_norm": 6.612964153289795, "learning_rate": 9.64295981928948e-05, "loss": 0.7472, "step": 3359 }, { "epoch": 0.22765770038620503, "grad_norm": 9.4874267578125, "learning_rate": 9.642822917379698e-05, "loss": 0.994, "step": 3360 }, { "epoch": 0.22772545565417712, "grad_norm": 9.7284574508667, "learning_rate": 9.642686015469916e-05, "loss": 0.9293, "step": 3361 }, { "epoch": 0.2277932109221492, "grad_norm": 8.787792205810547, "learning_rate": 9.642549113560134e-05, "loss": 1.1368, "step": 3362 }, { "epoch": 0.22786096619012128, "grad_norm": 9.218424797058105, "learning_rate": 9.642412211650354e-05, "loss": 1.2437, "step": 3363 }, { "epoch": 0.22792872145809337, "grad_norm": 7.9437127113342285, "learning_rate": 9.642275309740572e-05, "loss": 0.925, "step": 3364 }, { "epoch": 0.22799647672606546, "grad_norm": 10.552806854248047, "learning_rate": 9.64213840783079e-05, "loss": 0.936, "step": 3365 }, { "epoch": 0.22806423199403753, "grad_norm": 6.572626113891602, "learning_rate": 9.642001505921008e-05, "loss": 0.8968, "step": 3366 }, { "epoch": 0.22813198726200962, "grad_norm": 8.689478874206543, "learning_rate": 9.641864604011226e-05, "loss": 0.8605, "step": 3367 }, { "epoch": 0.2281997425299817, "grad_norm": 10.450199127197266, "learning_rate": 9.641727702101445e-05, "loss": 1.0871, "step": 3368 }, { "epoch": 0.2282674977979538, "grad_norm": 9.260807991027832, "learning_rate": 9.641590800191663e-05, "loss": 0.8469, "step": 3369 }, { "epoch": 0.22833525306592586, "grad_norm": 13.626687049865723, "learning_rate": 9.641453898281881e-05, "loss": 1.0615, "step": 3370 }, { "epoch": 0.22840300833389796, "grad_norm": 7.419554710388184, "learning_rate": 9.6413169963721e-05, "loss": 0.9351, "step": 3371 }, { "epoch": 0.22847076360187005, "grad_norm": 9.33206558227539, "learning_rate": 9.641180094462319e-05, "loss": 1.1826, "step": 3372 }, { "epoch": 0.22853851886984214, "grad_norm": 7.148665904998779, "learning_rate": 9.641043192552537e-05, "loss": 1.1121, "step": 3373 }, { "epoch": 0.2286062741378142, "grad_norm": 6.784035682678223, "learning_rate": 9.640906290642755e-05, "loss": 0.9649, "step": 3374 }, { "epoch": 0.2286740294057863, "grad_norm": 7.357334613800049, "learning_rate": 9.640769388732973e-05, "loss": 0.9005, "step": 3375 }, { "epoch": 0.22874178467375839, "grad_norm": 9.157288551330566, "learning_rate": 9.640632486823191e-05, "loss": 0.9585, "step": 3376 }, { "epoch": 0.22880953994173048, "grad_norm": 9.068450927734375, "learning_rate": 9.64049558491341e-05, "loss": 0.9248, "step": 3377 }, { "epoch": 0.22887729520970254, "grad_norm": 9.446860313415527, "learning_rate": 9.640358683003628e-05, "loss": 1.0016, "step": 3378 }, { "epoch": 0.22894505047767463, "grad_norm": 7.63693904876709, "learning_rate": 9.640221781093846e-05, "loss": 1.0158, "step": 3379 }, { "epoch": 0.22901280574564672, "grad_norm": 7.569469451904297, "learning_rate": 9.640084879184064e-05, "loss": 0.9149, "step": 3380 }, { "epoch": 0.22908056101361882, "grad_norm": 8.837080955505371, "learning_rate": 9.639947977274282e-05, "loss": 0.9379, "step": 3381 }, { "epoch": 0.22914831628159088, "grad_norm": 9.141901969909668, "learning_rate": 9.639811075364502e-05, "loss": 0.9121, "step": 3382 }, { "epoch": 0.22921607154956297, "grad_norm": 7.68120813369751, "learning_rate": 9.63967417345472e-05, "loss": 1.0235, "step": 3383 }, { "epoch": 0.22928382681753506, "grad_norm": 8.82229995727539, "learning_rate": 9.639537271544938e-05, "loss": 1.1573, "step": 3384 }, { "epoch": 0.22935158208550716, "grad_norm": 9.490239143371582, "learning_rate": 9.639400369635156e-05, "loss": 0.8776, "step": 3385 }, { "epoch": 0.22941933735347925, "grad_norm": 10.257567405700684, "learning_rate": 9.639263467725375e-05, "loss": 1.0696, "step": 3386 }, { "epoch": 0.2294870926214513, "grad_norm": 7.793918609619141, "learning_rate": 9.639126565815593e-05, "loss": 1.0892, "step": 3387 }, { "epoch": 0.2295548478894234, "grad_norm": 8.983718872070312, "learning_rate": 9.638989663905811e-05, "loss": 0.908, "step": 3388 }, { "epoch": 0.2296226031573955, "grad_norm": 8.31851577758789, "learning_rate": 9.63885276199603e-05, "loss": 1.0168, "step": 3389 }, { "epoch": 0.22969035842536759, "grad_norm": 9.343503952026367, "learning_rate": 9.638715860086248e-05, "loss": 1.1174, "step": 3390 }, { "epoch": 0.22975811369333965, "grad_norm": 7.617075443267822, "learning_rate": 9.638578958176467e-05, "loss": 0.9076, "step": 3391 }, { "epoch": 0.22982586896131174, "grad_norm": 8.102355003356934, "learning_rate": 9.638442056266685e-05, "loss": 1.0476, "step": 3392 }, { "epoch": 0.22989362422928383, "grad_norm": 9.366684913635254, "learning_rate": 9.638305154356903e-05, "loss": 1.1892, "step": 3393 }, { "epoch": 0.22996137949725592, "grad_norm": 8.204293251037598, "learning_rate": 9.638168252447122e-05, "loss": 0.7731, "step": 3394 }, { "epoch": 0.230029134765228, "grad_norm": 9.68787670135498, "learning_rate": 9.63803135053734e-05, "loss": 1.1126, "step": 3395 }, { "epoch": 0.23009689003320008, "grad_norm": 6.6209330558776855, "learning_rate": 9.637894448627558e-05, "loss": 0.9006, "step": 3396 }, { "epoch": 0.23016464530117217, "grad_norm": 7.963149547576904, "learning_rate": 9.637757546717778e-05, "loss": 0.8275, "step": 3397 }, { "epoch": 0.23023240056914426, "grad_norm": 8.74716567993164, "learning_rate": 9.637620644807996e-05, "loss": 1.0592, "step": 3398 }, { "epoch": 0.23030015583711633, "grad_norm": 8.819621086120605, "learning_rate": 9.637483742898214e-05, "loss": 0.8799, "step": 3399 }, { "epoch": 0.23036791110508842, "grad_norm": 8.538536071777344, "learning_rate": 9.637346840988433e-05, "loss": 1.1498, "step": 3400 }, { "epoch": 0.2304356663730605, "grad_norm": 8.787203788757324, "learning_rate": 9.637209939078651e-05, "loss": 0.8149, "step": 3401 }, { "epoch": 0.2305034216410326, "grad_norm": 8.85105037689209, "learning_rate": 9.63707303716887e-05, "loss": 0.8525, "step": 3402 }, { "epoch": 0.23057117690900467, "grad_norm": 9.773297309875488, "learning_rate": 9.636936135259087e-05, "loss": 0.9448, "step": 3403 }, { "epoch": 0.23063893217697676, "grad_norm": 7.804196357727051, "learning_rate": 9.636799233349307e-05, "loss": 0.9549, "step": 3404 }, { "epoch": 0.23070668744494885, "grad_norm": 8.398775100708008, "learning_rate": 9.636662331439525e-05, "loss": 0.9111, "step": 3405 }, { "epoch": 0.23077444271292094, "grad_norm": 8.960867881774902, "learning_rate": 9.636525429529743e-05, "loss": 1.1533, "step": 3406 }, { "epoch": 0.230842197980893, "grad_norm": 7.513526439666748, "learning_rate": 9.636388527619961e-05, "loss": 1.113, "step": 3407 }, { "epoch": 0.2309099532488651, "grad_norm": 8.891728401184082, "learning_rate": 9.636251625710179e-05, "loss": 0.8727, "step": 3408 }, { "epoch": 0.2309777085168372, "grad_norm": 6.972326755523682, "learning_rate": 9.636114723800398e-05, "loss": 0.8085, "step": 3409 }, { "epoch": 0.23104546378480928, "grad_norm": 9.195303916931152, "learning_rate": 9.635977821890616e-05, "loss": 1.0276, "step": 3410 }, { "epoch": 0.23111321905278134, "grad_norm": 9.36483383178711, "learning_rate": 9.635840919980834e-05, "loss": 1.1369, "step": 3411 }, { "epoch": 0.23118097432075344, "grad_norm": 7.282619953155518, "learning_rate": 9.635704018071052e-05, "loss": 0.9214, "step": 3412 }, { "epoch": 0.23124872958872553, "grad_norm": 6.289726257324219, "learning_rate": 9.63556711616127e-05, "loss": 0.8556, "step": 3413 }, { "epoch": 0.23131648485669762, "grad_norm": 8.646320343017578, "learning_rate": 9.63543021425149e-05, "loss": 1.0833, "step": 3414 }, { "epoch": 0.23138424012466968, "grad_norm": 8.372559547424316, "learning_rate": 9.635293312341708e-05, "loss": 0.9944, "step": 3415 }, { "epoch": 0.23145199539264177, "grad_norm": 8.002387046813965, "learning_rate": 9.635156410431926e-05, "loss": 1.015, "step": 3416 }, { "epoch": 0.23151975066061387, "grad_norm": 8.4190092086792, "learning_rate": 9.635019508522144e-05, "loss": 0.9766, "step": 3417 }, { "epoch": 0.23158750592858596, "grad_norm": 8.759613037109375, "learning_rate": 9.634882606612363e-05, "loss": 1.2348, "step": 3418 }, { "epoch": 0.23165526119655802, "grad_norm": 8.439151763916016, "learning_rate": 9.634745704702581e-05, "loss": 0.8778, "step": 3419 }, { "epoch": 0.2317230164645301, "grad_norm": 8.05675220489502, "learning_rate": 9.6346088027928e-05, "loss": 0.9525, "step": 3420 }, { "epoch": 0.2317907717325022, "grad_norm": 9.196937561035156, "learning_rate": 9.634471900883017e-05, "loss": 1.0663, "step": 3421 }, { "epoch": 0.2318585270004743, "grad_norm": 7.622048854827881, "learning_rate": 9.634334998973235e-05, "loss": 0.9522, "step": 3422 }, { "epoch": 0.23192628226844636, "grad_norm": 10.235919952392578, "learning_rate": 9.634198097063455e-05, "loss": 1.1084, "step": 3423 }, { "epoch": 0.23199403753641845, "grad_norm": 7.920719623565674, "learning_rate": 9.634061195153673e-05, "loss": 0.9443, "step": 3424 }, { "epoch": 0.23206179280439054, "grad_norm": 7.440132141113281, "learning_rate": 9.633924293243891e-05, "loss": 1.0757, "step": 3425 }, { "epoch": 0.23212954807236263, "grad_norm": 7.0108771324157715, "learning_rate": 9.633787391334109e-05, "loss": 0.8043, "step": 3426 }, { "epoch": 0.2321973033403347, "grad_norm": 6.620635032653809, "learning_rate": 9.633650489424328e-05, "loss": 0.999, "step": 3427 }, { "epoch": 0.2322650586083068, "grad_norm": 7.27518367767334, "learning_rate": 9.633513587514546e-05, "loss": 0.888, "step": 3428 }, { "epoch": 0.23233281387627888, "grad_norm": 7.081945419311523, "learning_rate": 9.633376685604764e-05, "loss": 0.8837, "step": 3429 }, { "epoch": 0.23240056914425097, "grad_norm": 7.225597858428955, "learning_rate": 9.633239783694982e-05, "loss": 0.7774, "step": 3430 }, { "epoch": 0.23246832441222304, "grad_norm": 7.0490217208862305, "learning_rate": 9.6331028817852e-05, "loss": 1.0654, "step": 3431 }, { "epoch": 0.23253607968019513, "grad_norm": 8.641639709472656, "learning_rate": 9.63296597987542e-05, "loss": 1.1071, "step": 3432 }, { "epoch": 0.23260383494816722, "grad_norm": 7.442448139190674, "learning_rate": 9.632829077965638e-05, "loss": 0.9392, "step": 3433 }, { "epoch": 0.2326715902161393, "grad_norm": 6.6959452629089355, "learning_rate": 9.632692176055856e-05, "loss": 0.8334, "step": 3434 }, { "epoch": 0.23273934548411138, "grad_norm": 9.67419719696045, "learning_rate": 9.632555274146074e-05, "loss": 1.1948, "step": 3435 }, { "epoch": 0.23280710075208347, "grad_norm": 7.384359836578369, "learning_rate": 9.632418372236292e-05, "loss": 0.7992, "step": 3436 }, { "epoch": 0.23287485602005556, "grad_norm": 6.566294193267822, "learning_rate": 9.632281470326511e-05, "loss": 0.9393, "step": 3437 }, { "epoch": 0.23294261128802765, "grad_norm": 7.842831134796143, "learning_rate": 9.63214456841673e-05, "loss": 1.1372, "step": 3438 }, { "epoch": 0.23301036655599974, "grad_norm": 9.487961769104004, "learning_rate": 9.632007666506947e-05, "loss": 1.0381, "step": 3439 }, { "epoch": 0.2330781218239718, "grad_norm": 8.06712532043457, "learning_rate": 9.631870764597167e-05, "loss": 1.4106, "step": 3440 }, { "epoch": 0.2331458770919439, "grad_norm": 7.026076793670654, "learning_rate": 9.631733862687385e-05, "loss": 1.0662, "step": 3441 }, { "epoch": 0.233213632359916, "grad_norm": 10.926567077636719, "learning_rate": 9.631596960777603e-05, "loss": 1.2932, "step": 3442 }, { "epoch": 0.23328138762788808, "grad_norm": 8.380082130432129, "learning_rate": 9.631460058867822e-05, "loss": 1.0784, "step": 3443 }, { "epoch": 0.23334914289586015, "grad_norm": 8.768819808959961, "learning_rate": 9.63132315695804e-05, "loss": 1.1501, "step": 3444 }, { "epoch": 0.23341689816383224, "grad_norm": 8.467490196228027, "learning_rate": 9.631186255048258e-05, "loss": 0.7555, "step": 3445 }, { "epoch": 0.23348465343180433, "grad_norm": 6.0409770011901855, "learning_rate": 9.631049353138478e-05, "loss": 0.9099, "step": 3446 }, { "epoch": 0.23355240869977642, "grad_norm": 7.160163402557373, "learning_rate": 9.630912451228696e-05, "loss": 0.8386, "step": 3447 }, { "epoch": 0.23362016396774848, "grad_norm": 9.102558135986328, "learning_rate": 9.630775549318914e-05, "loss": 0.9259, "step": 3448 }, { "epoch": 0.23368791923572058, "grad_norm": 7.071728706359863, "learning_rate": 9.630638647409132e-05, "loss": 1.0426, "step": 3449 }, { "epoch": 0.23375567450369267, "grad_norm": 8.358567237854004, "learning_rate": 9.630501745499351e-05, "loss": 0.9806, "step": 3450 }, { "epoch": 0.23382342977166476, "grad_norm": 12.588993072509766, "learning_rate": 9.630364843589569e-05, "loss": 1.0796, "step": 3451 }, { "epoch": 0.23389118503963682, "grad_norm": 8.665871620178223, "learning_rate": 9.630227941679787e-05, "loss": 1.0224, "step": 3452 }, { "epoch": 0.23395894030760891, "grad_norm": 9.288796424865723, "learning_rate": 9.630091039770005e-05, "loss": 0.9329, "step": 3453 }, { "epoch": 0.234026695575581, "grad_norm": 7.319186210632324, "learning_rate": 9.629954137860223e-05, "loss": 1.057, "step": 3454 }, { "epoch": 0.2340944508435531, "grad_norm": 8.640669822692871, "learning_rate": 9.629817235950443e-05, "loss": 0.9756, "step": 3455 }, { "epoch": 0.23416220611152516, "grad_norm": 10.73513126373291, "learning_rate": 9.629680334040661e-05, "loss": 0.958, "step": 3456 }, { "epoch": 0.23422996137949725, "grad_norm": 7.40097713470459, "learning_rate": 9.629543432130879e-05, "loss": 0.9062, "step": 3457 }, { "epoch": 0.23429771664746935, "grad_norm": 9.506468772888184, "learning_rate": 9.629406530221097e-05, "loss": 1.0083, "step": 3458 }, { "epoch": 0.23436547191544144, "grad_norm": 7.291172504425049, "learning_rate": 9.629269628311315e-05, "loss": 0.9966, "step": 3459 }, { "epoch": 0.2344332271834135, "grad_norm": 8.962395668029785, "learning_rate": 9.629132726401534e-05, "loss": 0.8504, "step": 3460 }, { "epoch": 0.2345009824513856, "grad_norm": 10.29174518585205, "learning_rate": 9.628995824491752e-05, "loss": 1.0807, "step": 3461 }, { "epoch": 0.23456873771935768, "grad_norm": 9.461160659790039, "learning_rate": 9.62885892258197e-05, "loss": 0.996, "step": 3462 }, { "epoch": 0.23463649298732978, "grad_norm": 7.13394021987915, "learning_rate": 9.628722020672188e-05, "loss": 0.7163, "step": 3463 }, { "epoch": 0.23470424825530184, "grad_norm": 9.379724502563477, "learning_rate": 9.628585118762408e-05, "loss": 0.9571, "step": 3464 }, { "epoch": 0.23477200352327393, "grad_norm": 9.72339916229248, "learning_rate": 9.628448216852626e-05, "loss": 1.2559, "step": 3465 }, { "epoch": 0.23483975879124602, "grad_norm": 7.353830337524414, "learning_rate": 9.628311314942844e-05, "loss": 1.0563, "step": 3466 }, { "epoch": 0.23490751405921811, "grad_norm": 8.423667907714844, "learning_rate": 9.628174413033062e-05, "loss": 1.0799, "step": 3467 }, { "epoch": 0.23497526932719018, "grad_norm": 6.426609516143799, "learning_rate": 9.62803751112328e-05, "loss": 0.9417, "step": 3468 }, { "epoch": 0.23504302459516227, "grad_norm": 10.882522583007812, "learning_rate": 9.627900609213499e-05, "loss": 1.1498, "step": 3469 }, { "epoch": 0.23511077986313436, "grad_norm": 7.768298625946045, "learning_rate": 9.627763707303717e-05, "loss": 0.7389, "step": 3470 }, { "epoch": 0.23517853513110645, "grad_norm": 9.002137184143066, "learning_rate": 9.627626805393935e-05, "loss": 1.2338, "step": 3471 }, { "epoch": 0.23524629039907852, "grad_norm": 8.671159744262695, "learning_rate": 9.627489903484153e-05, "loss": 1.1114, "step": 3472 }, { "epoch": 0.2353140456670506, "grad_norm": 5.836034774780273, "learning_rate": 9.627353001574373e-05, "loss": 0.7075, "step": 3473 }, { "epoch": 0.2353818009350227, "grad_norm": 8.145447731018066, "learning_rate": 9.627216099664591e-05, "loss": 0.9504, "step": 3474 }, { "epoch": 0.2354495562029948, "grad_norm": 7.3333563804626465, "learning_rate": 9.627079197754809e-05, "loss": 0.8443, "step": 3475 }, { "epoch": 0.23551731147096686, "grad_norm": 8.191438674926758, "learning_rate": 9.626942295845027e-05, "loss": 0.8163, "step": 3476 }, { "epoch": 0.23558506673893895, "grad_norm": 10.042975425720215, "learning_rate": 9.626805393935245e-05, "loss": 1.0715, "step": 3477 }, { "epoch": 0.23565282200691104, "grad_norm": 7.854464054107666, "learning_rate": 9.626668492025464e-05, "loss": 0.9263, "step": 3478 }, { "epoch": 0.23572057727488313, "grad_norm": 9.286520004272461, "learning_rate": 9.626531590115682e-05, "loss": 1.0172, "step": 3479 }, { "epoch": 0.2357883325428552, "grad_norm": 7.321778297424316, "learning_rate": 9.6263946882059e-05, "loss": 1.0661, "step": 3480 }, { "epoch": 0.2358560878108273, "grad_norm": 7.432394981384277, "learning_rate": 9.626257786296118e-05, "loss": 1.1682, "step": 3481 }, { "epoch": 0.23592384307879938, "grad_norm": 10.230997085571289, "learning_rate": 9.626120884386338e-05, "loss": 1.0498, "step": 3482 }, { "epoch": 0.23599159834677147, "grad_norm": 8.530696868896484, "learning_rate": 9.625983982476556e-05, "loss": 1.0598, "step": 3483 }, { "epoch": 0.23605935361474353, "grad_norm": 7.35584020614624, "learning_rate": 9.625847080566774e-05, "loss": 1.0234, "step": 3484 }, { "epoch": 0.23612710888271563, "grad_norm": 8.030210494995117, "learning_rate": 9.625710178656992e-05, "loss": 1.0241, "step": 3485 }, { "epoch": 0.23619486415068772, "grad_norm": 7.508440971374512, "learning_rate": 9.625573276747211e-05, "loss": 0.9466, "step": 3486 }, { "epoch": 0.2362626194186598, "grad_norm": 7.4618401527404785, "learning_rate": 9.62543637483743e-05, "loss": 0.88, "step": 3487 }, { "epoch": 0.23633037468663187, "grad_norm": 7.738553524017334, "learning_rate": 9.625299472927647e-05, "loss": 1.0146, "step": 3488 }, { "epoch": 0.23639812995460396, "grad_norm": 6.782055854797363, "learning_rate": 9.625162571017867e-05, "loss": 0.8196, "step": 3489 }, { "epoch": 0.23646588522257606, "grad_norm": 7.495883464813232, "learning_rate": 9.625025669108085e-05, "loss": 1.0025, "step": 3490 }, { "epoch": 0.23653364049054815, "grad_norm": 7.224970817565918, "learning_rate": 9.624888767198303e-05, "loss": 1.0069, "step": 3491 }, { "epoch": 0.23660139575852024, "grad_norm": 8.686829566955566, "learning_rate": 9.624751865288522e-05, "loss": 0.8927, "step": 3492 }, { "epoch": 0.2366691510264923, "grad_norm": 10.551370620727539, "learning_rate": 9.62461496337874e-05, "loss": 1.0973, "step": 3493 }, { "epoch": 0.2367369062944644, "grad_norm": 8.38442611694336, "learning_rate": 9.624478061468958e-05, "loss": 0.8145, "step": 3494 }, { "epoch": 0.23680466156243649, "grad_norm": 7.099575996398926, "learning_rate": 9.624341159559176e-05, "loss": 0.8304, "step": 3495 }, { "epoch": 0.23687241683040858, "grad_norm": 7.905203342437744, "learning_rate": 9.624204257649396e-05, "loss": 1.1708, "step": 3496 }, { "epoch": 0.23694017209838064, "grad_norm": 8.08218765258789, "learning_rate": 9.624067355739614e-05, "loss": 0.8367, "step": 3497 }, { "epoch": 0.23700792736635273, "grad_norm": 7.766988277435303, "learning_rate": 9.623930453829832e-05, "loss": 1.1126, "step": 3498 }, { "epoch": 0.23707568263432482, "grad_norm": 8.3229398727417, "learning_rate": 9.62379355192005e-05, "loss": 0.9798, "step": 3499 }, { "epoch": 0.23714343790229692, "grad_norm": 7.113401889801025, "learning_rate": 9.623656650010268e-05, "loss": 0.7711, "step": 3500 }, { "epoch": 0.23721119317026898, "grad_norm": 7.286252975463867, "learning_rate": 9.623519748100487e-05, "loss": 0.8018, "step": 3501 }, { "epoch": 0.23727894843824107, "grad_norm": 8.724101066589355, "learning_rate": 9.623382846190705e-05, "loss": 0.9077, "step": 3502 }, { "epoch": 0.23734670370621316, "grad_norm": 9.201606750488281, "learning_rate": 9.623245944280923e-05, "loss": 1.2277, "step": 3503 }, { "epoch": 0.23741445897418526, "grad_norm": 6.827572822570801, "learning_rate": 9.623109042371141e-05, "loss": 0.9262, "step": 3504 }, { "epoch": 0.23748221424215732, "grad_norm": 7.41288423538208, "learning_rate": 9.622972140461361e-05, "loss": 0.762, "step": 3505 }, { "epoch": 0.2375499695101294, "grad_norm": 9.649503707885742, "learning_rate": 9.622835238551579e-05, "loss": 1.0371, "step": 3506 }, { "epoch": 0.2376177247781015, "grad_norm": 9.630754470825195, "learning_rate": 9.622698336641797e-05, "loss": 0.9994, "step": 3507 }, { "epoch": 0.2376854800460736, "grad_norm": 8.949817657470703, "learning_rate": 9.622561434732015e-05, "loss": 0.8555, "step": 3508 }, { "epoch": 0.23775323531404566, "grad_norm": 8.096979141235352, "learning_rate": 9.622424532822233e-05, "loss": 0.9939, "step": 3509 }, { "epoch": 0.23782099058201775, "grad_norm": 8.714512825012207, "learning_rate": 9.622287630912452e-05, "loss": 0.9533, "step": 3510 }, { "epoch": 0.23788874584998984, "grad_norm": 8.663968086242676, "learning_rate": 9.62215072900267e-05, "loss": 0.9974, "step": 3511 }, { "epoch": 0.23795650111796193, "grad_norm": 8.704265594482422, "learning_rate": 9.622013827092888e-05, "loss": 1.2524, "step": 3512 }, { "epoch": 0.238024256385934, "grad_norm": 6.585339069366455, "learning_rate": 9.621876925183106e-05, "loss": 0.9021, "step": 3513 }, { "epoch": 0.2380920116539061, "grad_norm": 7.992496013641357, "learning_rate": 9.621740023273324e-05, "loss": 0.9345, "step": 3514 }, { "epoch": 0.23815976692187818, "grad_norm": 7.076025009155273, "learning_rate": 9.621603121363544e-05, "loss": 1.0487, "step": 3515 }, { "epoch": 0.23822752218985027, "grad_norm": 8.833822250366211, "learning_rate": 9.621466219453762e-05, "loss": 0.9612, "step": 3516 }, { "epoch": 0.23829527745782234, "grad_norm": 7.8553385734558105, "learning_rate": 9.62132931754398e-05, "loss": 1.0452, "step": 3517 }, { "epoch": 0.23836303272579443, "grad_norm": 8.44243049621582, "learning_rate": 9.621192415634198e-05, "loss": 0.6874, "step": 3518 }, { "epoch": 0.23843078799376652, "grad_norm": 8.1638822555542, "learning_rate": 9.621055513724417e-05, "loss": 0.9239, "step": 3519 }, { "epoch": 0.2384985432617386, "grad_norm": 8.315443992614746, "learning_rate": 9.620918611814635e-05, "loss": 1.1256, "step": 3520 }, { "epoch": 0.23856629852971067, "grad_norm": 10.451863288879395, "learning_rate": 9.620781709904853e-05, "loss": 1.1453, "step": 3521 }, { "epoch": 0.23863405379768277, "grad_norm": 9.117147445678711, "learning_rate": 9.620644807995071e-05, "loss": 1.0751, "step": 3522 }, { "epoch": 0.23870180906565486, "grad_norm": 7.746578216552734, "learning_rate": 9.62050790608529e-05, "loss": 0.867, "step": 3523 }, { "epoch": 0.23876956433362695, "grad_norm": 8.366421699523926, "learning_rate": 9.620371004175509e-05, "loss": 1.0863, "step": 3524 }, { "epoch": 0.238837319601599, "grad_norm": 9.951937675476074, "learning_rate": 9.620234102265727e-05, "loss": 1.0431, "step": 3525 }, { "epoch": 0.2389050748695711, "grad_norm": 9.077424049377441, "learning_rate": 9.620097200355945e-05, "loss": 1.0433, "step": 3526 }, { "epoch": 0.2389728301375432, "grad_norm": 6.919139385223389, "learning_rate": 9.619960298446163e-05, "loss": 0.8655, "step": 3527 }, { "epoch": 0.2390405854055153, "grad_norm": 7.729475975036621, "learning_rate": 9.619823396536382e-05, "loss": 1.0943, "step": 3528 }, { "epoch": 0.23910834067348735, "grad_norm": 9.782391548156738, "learning_rate": 9.6196864946266e-05, "loss": 1.1362, "step": 3529 }, { "epoch": 0.23917609594145944, "grad_norm": 8.685064315795898, "learning_rate": 9.619549592716818e-05, "loss": 0.9885, "step": 3530 }, { "epoch": 0.23924385120943154, "grad_norm": 6.963294982910156, "learning_rate": 9.619412690807036e-05, "loss": 0.9061, "step": 3531 }, { "epoch": 0.23931160647740363, "grad_norm": 8.584861755371094, "learning_rate": 9.619275788897256e-05, "loss": 0.8527, "step": 3532 }, { "epoch": 0.2393793617453757, "grad_norm": 9.253739356994629, "learning_rate": 9.619138886987474e-05, "loss": 1.0643, "step": 3533 }, { "epoch": 0.23944711701334778, "grad_norm": 7.659543037414551, "learning_rate": 9.619001985077692e-05, "loss": 0.9553, "step": 3534 }, { "epoch": 0.23951487228131987, "grad_norm": 8.769670486450195, "learning_rate": 9.618865083167911e-05, "loss": 1.0576, "step": 3535 }, { "epoch": 0.23958262754929197, "grad_norm": 7.416141033172607, "learning_rate": 9.618728181258129e-05, "loss": 1.0064, "step": 3536 }, { "epoch": 0.23965038281726403, "grad_norm": 7.91802978515625, "learning_rate": 9.618591279348347e-05, "loss": 1.0095, "step": 3537 }, { "epoch": 0.23971813808523612, "grad_norm": 6.665622234344482, "learning_rate": 9.618454377438567e-05, "loss": 0.8099, "step": 3538 }, { "epoch": 0.2397858933532082, "grad_norm": 7.3240203857421875, "learning_rate": 9.618317475528785e-05, "loss": 0.9496, "step": 3539 }, { "epoch": 0.2398536486211803, "grad_norm": 7.268299102783203, "learning_rate": 9.618180573619003e-05, "loss": 1.0504, "step": 3540 }, { "epoch": 0.23992140388915237, "grad_norm": 8.710535049438477, "learning_rate": 9.618043671709221e-05, "loss": 0.9228, "step": 3541 }, { "epoch": 0.23998915915712446, "grad_norm": 8.5223970413208, "learning_rate": 9.61790676979944e-05, "loss": 1.0383, "step": 3542 }, { "epoch": 0.24005691442509655, "grad_norm": 7.635293960571289, "learning_rate": 9.617769867889658e-05, "loss": 0.99, "step": 3543 }, { "epoch": 0.24012466969306864, "grad_norm": 9.41180419921875, "learning_rate": 9.617632965979876e-05, "loss": 1.1214, "step": 3544 }, { "epoch": 0.24019242496104073, "grad_norm": 9.31615924835205, "learning_rate": 9.617496064070094e-05, "loss": 0.9369, "step": 3545 }, { "epoch": 0.2402601802290128, "grad_norm": 9.427112579345703, "learning_rate": 9.617359162160312e-05, "loss": 1.1829, "step": 3546 }, { "epoch": 0.2403279354969849, "grad_norm": 7.825446605682373, "learning_rate": 9.617222260250532e-05, "loss": 0.9459, "step": 3547 }, { "epoch": 0.24039569076495698, "grad_norm": 7.8191633224487305, "learning_rate": 9.61708535834075e-05, "loss": 1.1144, "step": 3548 }, { "epoch": 0.24046344603292907, "grad_norm": 8.224778175354004, "learning_rate": 9.616948456430968e-05, "loss": 0.8544, "step": 3549 }, { "epoch": 0.24053120130090114, "grad_norm": 9.618694305419922, "learning_rate": 9.616811554521186e-05, "loss": 1.2254, "step": 3550 }, { "epoch": 0.24059895656887323, "grad_norm": 7.786314964294434, "learning_rate": 9.616674652611405e-05, "loss": 0.8867, "step": 3551 }, { "epoch": 0.24066671183684532, "grad_norm": 9.397835731506348, "learning_rate": 9.616537750701623e-05, "loss": 1.0505, "step": 3552 }, { "epoch": 0.2407344671048174, "grad_norm": 7.701049327850342, "learning_rate": 9.616400848791841e-05, "loss": 0.897, "step": 3553 }, { "epoch": 0.24080222237278948, "grad_norm": 7.573019504547119, "learning_rate": 9.616263946882059e-05, "loss": 0.9972, "step": 3554 }, { "epoch": 0.24086997764076157, "grad_norm": 8.337100982666016, "learning_rate": 9.616127044972277e-05, "loss": 1.0835, "step": 3555 }, { "epoch": 0.24093773290873366, "grad_norm": 8.702056884765625, "learning_rate": 9.615990143062497e-05, "loss": 0.7688, "step": 3556 }, { "epoch": 0.24100548817670575, "grad_norm": 8.482346534729004, "learning_rate": 9.615853241152715e-05, "loss": 1.1972, "step": 3557 }, { "epoch": 0.24107324344467781, "grad_norm": 7.193674087524414, "learning_rate": 9.615716339242933e-05, "loss": 1.0389, "step": 3558 }, { "epoch": 0.2411409987126499, "grad_norm": 8.803317070007324, "learning_rate": 9.615579437333151e-05, "loss": 1.0429, "step": 3559 }, { "epoch": 0.241208753980622, "grad_norm": 9.055732727050781, "learning_rate": 9.61544253542337e-05, "loss": 1.1161, "step": 3560 }, { "epoch": 0.2412765092485941, "grad_norm": 10.80001449584961, "learning_rate": 9.615305633513588e-05, "loss": 0.9628, "step": 3561 }, { "epoch": 0.24134426451656615, "grad_norm": 7.707313060760498, "learning_rate": 9.615168731603806e-05, "loss": 0.8806, "step": 3562 }, { "epoch": 0.24141201978453825, "grad_norm": 8.823626518249512, "learning_rate": 9.615031829694024e-05, "loss": 0.8771, "step": 3563 }, { "epoch": 0.24147977505251034, "grad_norm": 7.984725475311279, "learning_rate": 9.614894927784242e-05, "loss": 0.9456, "step": 3564 }, { "epoch": 0.24154753032048243, "grad_norm": 7.335816860198975, "learning_rate": 9.614758025874462e-05, "loss": 1.0323, "step": 3565 }, { "epoch": 0.2416152855884545, "grad_norm": 5.959085941314697, "learning_rate": 9.61462112396468e-05, "loss": 0.8049, "step": 3566 }, { "epoch": 0.24168304085642658, "grad_norm": 8.775632858276367, "learning_rate": 9.614484222054898e-05, "loss": 0.8875, "step": 3567 }, { "epoch": 0.24175079612439868, "grad_norm": 6.589362621307373, "learning_rate": 9.614347320145116e-05, "loss": 1.0551, "step": 3568 }, { "epoch": 0.24181855139237077, "grad_norm": 7.49434232711792, "learning_rate": 9.614210418235334e-05, "loss": 0.7662, "step": 3569 }, { "epoch": 0.24188630666034283, "grad_norm": 7.759862899780273, "learning_rate": 9.614073516325553e-05, "loss": 0.7364, "step": 3570 }, { "epoch": 0.24195406192831492, "grad_norm": 7.9510273933410645, "learning_rate": 9.613936614415771e-05, "loss": 0.9277, "step": 3571 }, { "epoch": 0.24202181719628701, "grad_norm": 8.308568954467773, "learning_rate": 9.61379971250599e-05, "loss": 1.1689, "step": 3572 }, { "epoch": 0.2420895724642591, "grad_norm": 9.44938850402832, "learning_rate": 9.613662810596207e-05, "loss": 1.0937, "step": 3573 }, { "epoch": 0.24215732773223117, "grad_norm": 11.66707706451416, "learning_rate": 9.613525908686427e-05, "loss": 0.77, "step": 3574 }, { "epoch": 0.24222508300020326, "grad_norm": 9.25683879852295, "learning_rate": 9.613389006776645e-05, "loss": 0.9661, "step": 3575 }, { "epoch": 0.24229283826817535, "grad_norm": 7.289797782897949, "learning_rate": 9.613252104866863e-05, "loss": 0.8597, "step": 3576 }, { "epoch": 0.24236059353614745, "grad_norm": 6.669293403625488, "learning_rate": 9.613115202957081e-05, "loss": 0.9134, "step": 3577 }, { "epoch": 0.2424283488041195, "grad_norm": 8.785436630249023, "learning_rate": 9.612978301047299e-05, "loss": 0.7546, "step": 3578 }, { "epoch": 0.2424961040720916, "grad_norm": 7.386310577392578, "learning_rate": 9.612841399137518e-05, "loss": 1.0635, "step": 3579 }, { "epoch": 0.2425638593400637, "grad_norm": 8.23388957977295, "learning_rate": 9.612704497227736e-05, "loss": 0.8598, "step": 3580 }, { "epoch": 0.24263161460803578, "grad_norm": 8.498323440551758, "learning_rate": 9.612567595317954e-05, "loss": 1.083, "step": 3581 }, { "epoch": 0.24269936987600785, "grad_norm": 7.849715709686279, "learning_rate": 9.612430693408174e-05, "loss": 0.7673, "step": 3582 }, { "epoch": 0.24276712514397994, "grad_norm": 8.113242149353027, "learning_rate": 9.612293791498392e-05, "loss": 1.1003, "step": 3583 }, { "epoch": 0.24283488041195203, "grad_norm": 6.983048915863037, "learning_rate": 9.61215688958861e-05, "loss": 0.8668, "step": 3584 }, { "epoch": 0.24290263567992412, "grad_norm": 8.389126777648926, "learning_rate": 9.612019987678829e-05, "loss": 1.0008, "step": 3585 }, { "epoch": 0.2429703909478962, "grad_norm": 7.593414783477783, "learning_rate": 9.611883085769047e-05, "loss": 0.9829, "step": 3586 }, { "epoch": 0.24303814621586828, "grad_norm": 7.836172103881836, "learning_rate": 9.611746183859265e-05, "loss": 0.8128, "step": 3587 }, { "epoch": 0.24310590148384037, "grad_norm": 8.981040954589844, "learning_rate": 9.611609281949485e-05, "loss": 1.1249, "step": 3588 }, { "epoch": 0.24317365675181246, "grad_norm": 6.724935054779053, "learning_rate": 9.611472380039703e-05, "loss": 0.7699, "step": 3589 }, { "epoch": 0.24324141201978453, "grad_norm": 9.456436157226562, "learning_rate": 9.611335478129921e-05, "loss": 1.0085, "step": 3590 }, { "epoch": 0.24330916728775662, "grad_norm": 11.044548988342285, "learning_rate": 9.611198576220139e-05, "loss": 1.0671, "step": 3591 }, { "epoch": 0.2433769225557287, "grad_norm": 10.590521812438965, "learning_rate": 9.611061674310358e-05, "loss": 1.0051, "step": 3592 }, { "epoch": 0.2434446778237008, "grad_norm": 8.595380783081055, "learning_rate": 9.610924772400576e-05, "loss": 0.9655, "step": 3593 }, { "epoch": 0.24351243309167286, "grad_norm": 7.794788837432861, "learning_rate": 9.610787870490794e-05, "loss": 0.9551, "step": 3594 }, { "epoch": 0.24358018835964496, "grad_norm": 7.609074115753174, "learning_rate": 9.610650968581012e-05, "loss": 0.9856, "step": 3595 }, { "epoch": 0.24364794362761705, "grad_norm": 6.909607410430908, "learning_rate": 9.61051406667123e-05, "loss": 1.0195, "step": 3596 }, { "epoch": 0.24371569889558914, "grad_norm": 7.958381652832031, "learning_rate": 9.61037716476145e-05, "loss": 0.8263, "step": 3597 }, { "epoch": 0.24378345416356123, "grad_norm": 7.38173246383667, "learning_rate": 9.610240262851668e-05, "loss": 0.8551, "step": 3598 }, { "epoch": 0.2438512094315333, "grad_norm": 7.1448822021484375, "learning_rate": 9.610103360941886e-05, "loss": 0.8189, "step": 3599 }, { "epoch": 0.2439189646995054, "grad_norm": 7.44658088684082, "learning_rate": 9.609966459032104e-05, "loss": 0.8056, "step": 3600 }, { "epoch": 0.24398671996747748, "grad_norm": 7.480542182922363, "learning_rate": 9.609829557122322e-05, "loss": 0.8591, "step": 3601 }, { "epoch": 0.24405447523544957, "grad_norm": 9.895995140075684, "learning_rate": 9.609692655212541e-05, "loss": 1.1256, "step": 3602 }, { "epoch": 0.24412223050342163, "grad_norm": 7.009078025817871, "learning_rate": 9.609555753302759e-05, "loss": 0.9045, "step": 3603 }, { "epoch": 0.24418998577139372, "grad_norm": 8.714953422546387, "learning_rate": 9.609418851392977e-05, "loss": 1.1207, "step": 3604 }, { "epoch": 0.24425774103936582, "grad_norm": 7.239734172821045, "learning_rate": 9.609281949483195e-05, "loss": 0.7811, "step": 3605 }, { "epoch": 0.2443254963073379, "grad_norm": 10.486507415771484, "learning_rate": 9.609145047573415e-05, "loss": 1.2213, "step": 3606 }, { "epoch": 0.24439325157530997, "grad_norm": 12.031790733337402, "learning_rate": 9.609008145663633e-05, "loss": 1.0316, "step": 3607 }, { "epoch": 0.24446100684328206, "grad_norm": 7.607183456420898, "learning_rate": 9.608871243753851e-05, "loss": 0.7754, "step": 3608 }, { "epoch": 0.24452876211125416, "grad_norm": 9.313577651977539, "learning_rate": 9.608734341844069e-05, "loss": 0.9907, "step": 3609 }, { "epoch": 0.24459651737922625, "grad_norm": 7.672274589538574, "learning_rate": 9.608597439934287e-05, "loss": 0.8428, "step": 3610 }, { "epoch": 0.2446642726471983, "grad_norm": 8.259462356567383, "learning_rate": 9.608460538024506e-05, "loss": 0.8863, "step": 3611 }, { "epoch": 0.2447320279151704, "grad_norm": 8.50256061553955, "learning_rate": 9.608323636114724e-05, "loss": 0.7771, "step": 3612 }, { "epoch": 0.2447997831831425, "grad_norm": 6.8893818855285645, "learning_rate": 9.608186734204942e-05, "loss": 0.9791, "step": 3613 }, { "epoch": 0.24486753845111459, "grad_norm": 8.954825401306152, "learning_rate": 9.60804983229516e-05, "loss": 1.0712, "step": 3614 }, { "epoch": 0.24493529371908665, "grad_norm": 8.83995532989502, "learning_rate": 9.60791293038538e-05, "loss": 1.1982, "step": 3615 }, { "epoch": 0.24500304898705874, "grad_norm": 8.96689224243164, "learning_rate": 9.607776028475598e-05, "loss": 0.9816, "step": 3616 }, { "epoch": 0.24507080425503083, "grad_norm": 15.17086124420166, "learning_rate": 9.607639126565816e-05, "loss": 1.1416, "step": 3617 }, { "epoch": 0.24513855952300292, "grad_norm": 9.527314186096191, "learning_rate": 9.607502224656034e-05, "loss": 1.1699, "step": 3618 }, { "epoch": 0.245206314790975, "grad_norm": 7.378002166748047, "learning_rate": 9.607365322746252e-05, "loss": 1.1303, "step": 3619 }, { "epoch": 0.24527407005894708, "grad_norm": 7.204291343688965, "learning_rate": 9.607228420836471e-05, "loss": 0.8078, "step": 3620 }, { "epoch": 0.24534182532691917, "grad_norm": 8.181205749511719, "learning_rate": 9.607091518926689e-05, "loss": 0.9221, "step": 3621 }, { "epoch": 0.24540958059489126, "grad_norm": 8.479545593261719, "learning_rate": 9.606954617016907e-05, "loss": 0.8133, "step": 3622 }, { "epoch": 0.24547733586286333, "grad_norm": 7.591360092163086, "learning_rate": 9.606817715107125e-05, "loss": 1.0632, "step": 3623 }, { "epoch": 0.24554509113083542, "grad_norm": 8.558969497680664, "learning_rate": 9.606680813197343e-05, "loss": 1.0755, "step": 3624 }, { "epoch": 0.2456128463988075, "grad_norm": 8.02037525177002, "learning_rate": 9.606543911287563e-05, "loss": 1.0164, "step": 3625 }, { "epoch": 0.2456806016667796, "grad_norm": 7.49207878112793, "learning_rate": 9.606407009377781e-05, "loss": 1.1325, "step": 3626 }, { "epoch": 0.24574835693475167, "grad_norm": 7.376079559326172, "learning_rate": 9.606270107467999e-05, "loss": 0.7917, "step": 3627 }, { "epoch": 0.24581611220272376, "grad_norm": 7.331247329711914, "learning_rate": 9.606133205558218e-05, "loss": 1.2186, "step": 3628 }, { "epoch": 0.24588386747069585, "grad_norm": 7.393257141113281, "learning_rate": 9.605996303648436e-05, "loss": 0.7155, "step": 3629 }, { "epoch": 0.24595162273866794, "grad_norm": 9.15472412109375, "learning_rate": 9.605859401738654e-05, "loss": 0.9562, "step": 3630 }, { "epoch": 0.24601937800664, "grad_norm": 6.846646785736084, "learning_rate": 9.605722499828874e-05, "loss": 1.109, "step": 3631 }, { "epoch": 0.2460871332746121, "grad_norm": 8.57854175567627, "learning_rate": 9.605585597919092e-05, "loss": 0.8706, "step": 3632 }, { "epoch": 0.2461548885425842, "grad_norm": 7.089768886566162, "learning_rate": 9.60544869600931e-05, "loss": 0.8776, "step": 3633 }, { "epoch": 0.24622264381055628, "grad_norm": 7.549044132232666, "learning_rate": 9.605311794099529e-05, "loss": 0.9956, "step": 3634 }, { "epoch": 0.24629039907852834, "grad_norm": 6.839412689208984, "learning_rate": 9.605174892189747e-05, "loss": 0.8722, "step": 3635 }, { "epoch": 0.24635815434650044, "grad_norm": 7.378058910369873, "learning_rate": 9.605037990279965e-05, "loss": 0.8663, "step": 3636 }, { "epoch": 0.24642590961447253, "grad_norm": 7.893070220947266, "learning_rate": 9.604901088370183e-05, "loss": 0.9213, "step": 3637 }, { "epoch": 0.24649366488244462, "grad_norm": 7.3345232009887695, "learning_rate": 9.604764186460403e-05, "loss": 0.9827, "step": 3638 }, { "epoch": 0.24656142015041668, "grad_norm": 6.120781421661377, "learning_rate": 9.60462728455062e-05, "loss": 0.6371, "step": 3639 }, { "epoch": 0.24662917541838877, "grad_norm": 8.695615768432617, "learning_rate": 9.604490382640839e-05, "loss": 0.9769, "step": 3640 }, { "epoch": 0.24669693068636087, "grad_norm": 8.469325065612793, "learning_rate": 9.604353480731057e-05, "loss": 0.9267, "step": 3641 }, { "epoch": 0.24676468595433296, "grad_norm": 9.191173553466797, "learning_rate": 9.604216578821275e-05, "loss": 1.1818, "step": 3642 }, { "epoch": 0.24683244122230502, "grad_norm": 9.337483406066895, "learning_rate": 9.604079676911494e-05, "loss": 0.9904, "step": 3643 }, { "epoch": 0.2469001964902771, "grad_norm": 7.597773551940918, "learning_rate": 9.603942775001712e-05, "loss": 0.8928, "step": 3644 }, { "epoch": 0.2469679517582492, "grad_norm": 8.155903816223145, "learning_rate": 9.60380587309193e-05, "loss": 0.9511, "step": 3645 }, { "epoch": 0.2470357070262213, "grad_norm": 7.695154666900635, "learning_rate": 9.603668971182148e-05, "loss": 1.0724, "step": 3646 }, { "epoch": 0.24710346229419336, "grad_norm": 7.492908000946045, "learning_rate": 9.603532069272366e-05, "loss": 0.9133, "step": 3647 }, { "epoch": 0.24717121756216545, "grad_norm": 8.190613746643066, "learning_rate": 9.603395167362586e-05, "loss": 0.8972, "step": 3648 }, { "epoch": 0.24723897283013754, "grad_norm": 7.9199347496032715, "learning_rate": 9.603258265452804e-05, "loss": 1.0501, "step": 3649 }, { "epoch": 0.24730672809810963, "grad_norm": 8.386896133422852, "learning_rate": 9.603121363543022e-05, "loss": 0.998, "step": 3650 }, { "epoch": 0.24737448336608173, "grad_norm": 6.536781311035156, "learning_rate": 9.60298446163324e-05, "loss": 0.8379, "step": 3651 }, { "epoch": 0.2474422386340538, "grad_norm": 6.918766021728516, "learning_rate": 9.602847559723459e-05, "loss": 0.9397, "step": 3652 }, { "epoch": 0.24750999390202588, "grad_norm": 7.00775146484375, "learning_rate": 9.602710657813677e-05, "loss": 1.0079, "step": 3653 }, { "epoch": 0.24757774916999797, "grad_norm": 7.693192958831787, "learning_rate": 9.602573755903895e-05, "loss": 1.0657, "step": 3654 }, { "epoch": 0.24764550443797007, "grad_norm": 10.667771339416504, "learning_rate": 9.602436853994113e-05, "loss": 1.0657, "step": 3655 }, { "epoch": 0.24771325970594213, "grad_norm": 8.622758865356445, "learning_rate": 9.602299952084331e-05, "loss": 0.9792, "step": 3656 }, { "epoch": 0.24778101497391422, "grad_norm": 8.834444046020508, "learning_rate": 9.602163050174551e-05, "loss": 1.1014, "step": 3657 }, { "epoch": 0.2478487702418863, "grad_norm": 7.717538356781006, "learning_rate": 9.602026148264769e-05, "loss": 0.8781, "step": 3658 }, { "epoch": 0.2479165255098584, "grad_norm": 7.797954559326172, "learning_rate": 9.601889246354987e-05, "loss": 0.9772, "step": 3659 }, { "epoch": 0.24798428077783047, "grad_norm": 7.376112937927246, "learning_rate": 9.601752344445205e-05, "loss": 1.0468, "step": 3660 }, { "epoch": 0.24805203604580256, "grad_norm": 7.7266387939453125, "learning_rate": 9.601615442535424e-05, "loss": 1.007, "step": 3661 }, { "epoch": 0.24811979131377465, "grad_norm": 9.196928977966309, "learning_rate": 9.601478540625642e-05, "loss": 1.1327, "step": 3662 }, { "epoch": 0.24818754658174674, "grad_norm": 7.892288684844971, "learning_rate": 9.60134163871586e-05, "loss": 0.7511, "step": 3663 }, { "epoch": 0.2482553018497188, "grad_norm": 8.854056358337402, "learning_rate": 9.601204736806078e-05, "loss": 0.9546, "step": 3664 }, { "epoch": 0.2483230571176909, "grad_norm": 7.985452651977539, "learning_rate": 9.601067834896296e-05, "loss": 1.2186, "step": 3665 }, { "epoch": 0.248390812385663, "grad_norm": 8.336162567138672, "learning_rate": 9.600930932986516e-05, "loss": 0.864, "step": 3666 }, { "epoch": 0.24845856765363508, "grad_norm": 10.857757568359375, "learning_rate": 9.600794031076734e-05, "loss": 1.1948, "step": 3667 }, { "epoch": 0.24852632292160715, "grad_norm": 8.168721199035645, "learning_rate": 9.600657129166952e-05, "loss": 1.1151, "step": 3668 }, { "epoch": 0.24859407818957924, "grad_norm": 7.509332656860352, "learning_rate": 9.60052022725717e-05, "loss": 0.9215, "step": 3669 }, { "epoch": 0.24866183345755133, "grad_norm": 10.4354829788208, "learning_rate": 9.600383325347389e-05, "loss": 1.0696, "step": 3670 }, { "epoch": 0.24872958872552342, "grad_norm": 6.930381774902344, "learning_rate": 9.600246423437607e-05, "loss": 0.9229, "step": 3671 }, { "epoch": 0.24879734399349548, "grad_norm": 6.608088970184326, "learning_rate": 9.600109521527825e-05, "loss": 0.928, "step": 3672 }, { "epoch": 0.24886509926146758, "grad_norm": 7.4217095375061035, "learning_rate": 9.599972619618043e-05, "loss": 0.9608, "step": 3673 }, { "epoch": 0.24893285452943967, "grad_norm": 7.46991491317749, "learning_rate": 9.599835717708263e-05, "loss": 0.8928, "step": 3674 }, { "epoch": 0.24900060979741176, "grad_norm": 8.76001262664795, "learning_rate": 9.599698815798481e-05, "loss": 1.191, "step": 3675 }, { "epoch": 0.24906836506538382, "grad_norm": 8.395779609680176, "learning_rate": 9.599561913888699e-05, "loss": 1.1479, "step": 3676 }, { "epoch": 0.24913612033335591, "grad_norm": 8.00460147857666, "learning_rate": 9.599425011978918e-05, "loss": 1.1005, "step": 3677 }, { "epoch": 0.249203875601328, "grad_norm": 7.817287445068359, "learning_rate": 9.599288110069136e-05, "loss": 1.1568, "step": 3678 }, { "epoch": 0.2492716308693001, "grad_norm": 9.60706615447998, "learning_rate": 9.599151208159354e-05, "loss": 0.9734, "step": 3679 }, { "epoch": 0.24933938613727216, "grad_norm": 8.347691535949707, "learning_rate": 9.599014306249574e-05, "loss": 1.2208, "step": 3680 }, { "epoch": 0.24940714140524425, "grad_norm": 8.106613159179688, "learning_rate": 9.598877404339792e-05, "loss": 0.9253, "step": 3681 }, { "epoch": 0.24947489667321635, "grad_norm": 8.099063873291016, "learning_rate": 9.59874050243001e-05, "loss": 1.0516, "step": 3682 }, { "epoch": 0.24954265194118844, "grad_norm": 7.839589595794678, "learning_rate": 9.598603600520228e-05, "loss": 0.9764, "step": 3683 }, { "epoch": 0.2496104072091605, "grad_norm": 8.679840087890625, "learning_rate": 9.598466698610447e-05, "loss": 1.0407, "step": 3684 }, { "epoch": 0.2496781624771326, "grad_norm": 6.854926109313965, "learning_rate": 9.598329796700665e-05, "loss": 0.9457, "step": 3685 }, { "epoch": 0.24974591774510468, "grad_norm": 9.133661270141602, "learning_rate": 9.598192894790883e-05, "loss": 1.0391, "step": 3686 }, { "epoch": 0.24981367301307678, "grad_norm": 7.744726657867432, "learning_rate": 9.598055992881101e-05, "loss": 1.0256, "step": 3687 }, { "epoch": 0.24988142828104884, "grad_norm": 8.9452543258667, "learning_rate": 9.597919090971319e-05, "loss": 0.9649, "step": 3688 }, { "epoch": 0.24994918354902093, "grad_norm": 8.396431922912598, "learning_rate": 9.597782189061539e-05, "loss": 0.8189, "step": 3689 }, { "epoch": 0.24994918354902093, "eval_loss": 0.9518795609474182, "eval_noise_accuracy": 0.0, "eval_runtime": 1533.2421, "eval_samples_per_second": 3.352, "eval_steps_per_second": 0.21, "eval_wer": 88.77438705594233, "step": 3689 }, { "epoch": 0.250016938816993, "grad_norm": 7.252607822418213, "learning_rate": 9.597645287151757e-05, "loss": 1.0734, "step": 3690 }, { "epoch": 0.2500846940849651, "grad_norm": 8.768789291381836, "learning_rate": 9.597508385241975e-05, "loss": 0.9763, "step": 3691 }, { "epoch": 0.2501524493529372, "grad_norm": 6.9475321769714355, "learning_rate": 9.597371483332193e-05, "loss": 0.8121, "step": 3692 }, { "epoch": 0.25022020462090927, "grad_norm": 7.9007248878479, "learning_rate": 9.597234581422412e-05, "loss": 0.9668, "step": 3693 }, { "epoch": 0.25028795988888136, "grad_norm": 6.524989604949951, "learning_rate": 9.59709767951263e-05, "loss": 0.9151, "step": 3694 }, { "epoch": 0.25035571515685345, "grad_norm": 7.83770751953125, "learning_rate": 9.596960777602848e-05, "loss": 0.8338, "step": 3695 }, { "epoch": 0.25042347042482554, "grad_norm": 8.619128227233887, "learning_rate": 9.596823875693066e-05, "loss": 1.0394, "step": 3696 }, { "epoch": 0.25049122569279764, "grad_norm": 7.238592624664307, "learning_rate": 9.596686973783284e-05, "loss": 0.8773, "step": 3697 }, { "epoch": 0.2505589809607697, "grad_norm": 10.257181167602539, "learning_rate": 9.596550071873504e-05, "loss": 1.0137, "step": 3698 }, { "epoch": 0.25062673622874176, "grad_norm": 9.302980422973633, "learning_rate": 9.596413169963722e-05, "loss": 0.8317, "step": 3699 }, { "epoch": 0.25069449149671386, "grad_norm": 7.485314846038818, "learning_rate": 9.59627626805394e-05, "loss": 0.9272, "step": 3700 }, { "epoch": 0.25076224676468595, "grad_norm": 7.870807647705078, "learning_rate": 9.596139366144158e-05, "loss": 1.009, "step": 3701 }, { "epoch": 0.25083000203265804, "grad_norm": 8.245805740356445, "learning_rate": 9.596002464234376e-05, "loss": 0.8616, "step": 3702 }, { "epoch": 0.25089775730063013, "grad_norm": 10.324470520019531, "learning_rate": 9.595865562324595e-05, "loss": 1.1567, "step": 3703 }, { "epoch": 0.2509655125686022, "grad_norm": 7.588774681091309, "learning_rate": 9.595728660414813e-05, "loss": 1.0025, "step": 3704 }, { "epoch": 0.2510332678365743, "grad_norm": 8.32935905456543, "learning_rate": 9.595591758505031e-05, "loss": 1.1826, "step": 3705 }, { "epoch": 0.2511010231045464, "grad_norm": 7.146906852722168, "learning_rate": 9.595454856595249e-05, "loss": 0.8108, "step": 3706 }, { "epoch": 0.25116877837251844, "grad_norm": 8.91852855682373, "learning_rate": 9.595317954685469e-05, "loss": 1.1377, "step": 3707 }, { "epoch": 0.25123653364049053, "grad_norm": 7.726437568664551, "learning_rate": 9.595181052775687e-05, "loss": 0.9971, "step": 3708 }, { "epoch": 0.2513042889084626, "grad_norm": 11.102527618408203, "learning_rate": 9.595044150865905e-05, "loss": 1.1075, "step": 3709 }, { "epoch": 0.2513720441764347, "grad_norm": 6.124303340911865, "learning_rate": 9.594907248956123e-05, "loss": 0.8523, "step": 3710 }, { "epoch": 0.2514397994444068, "grad_norm": 8.56926441192627, "learning_rate": 9.594770347046341e-05, "loss": 1.0801, "step": 3711 }, { "epoch": 0.2515075547123789, "grad_norm": 6.994394779205322, "learning_rate": 9.59463344513656e-05, "loss": 0.8559, "step": 3712 }, { "epoch": 0.251575309980351, "grad_norm": 7.428825378417969, "learning_rate": 9.594496543226778e-05, "loss": 0.8861, "step": 3713 }, { "epoch": 0.2516430652483231, "grad_norm": 8.7849760055542, "learning_rate": 9.594359641316996e-05, "loss": 0.9083, "step": 3714 }, { "epoch": 0.2517108205162951, "grad_norm": 7.608119964599609, "learning_rate": 9.594222739407214e-05, "loss": 0.8529, "step": 3715 }, { "epoch": 0.2517785757842672, "grad_norm": 7.076242446899414, "learning_rate": 9.594085837497434e-05, "loss": 0.8817, "step": 3716 }, { "epoch": 0.2518463310522393, "grad_norm": 8.78627872467041, "learning_rate": 9.593948935587652e-05, "loss": 1.1752, "step": 3717 }, { "epoch": 0.2519140863202114, "grad_norm": 8.589457511901855, "learning_rate": 9.59381203367787e-05, "loss": 0.9099, "step": 3718 }, { "epoch": 0.2519818415881835, "grad_norm": 8.67271900177002, "learning_rate": 9.593675131768088e-05, "loss": 1.0304, "step": 3719 }, { "epoch": 0.2520495968561556, "grad_norm": 9.180156707763672, "learning_rate": 9.593538229858307e-05, "loss": 1.0815, "step": 3720 }, { "epoch": 0.25211735212412767, "grad_norm": 7.973734378814697, "learning_rate": 9.593401327948525e-05, "loss": 1.0816, "step": 3721 }, { "epoch": 0.25218510739209976, "grad_norm": 7.272556781768799, "learning_rate": 9.593264426038743e-05, "loss": 0.9027, "step": 3722 }, { "epoch": 0.2522528626600718, "grad_norm": 8.033550262451172, "learning_rate": 9.593127524128963e-05, "loss": 1.0772, "step": 3723 }, { "epoch": 0.2523206179280439, "grad_norm": 7.861289978027344, "learning_rate": 9.59299062221918e-05, "loss": 0.8774, "step": 3724 }, { "epoch": 0.252388373196016, "grad_norm": 7.043121337890625, "learning_rate": 9.592853720309399e-05, "loss": 1.0165, "step": 3725 }, { "epoch": 0.25245612846398807, "grad_norm": 7.830938816070557, "learning_rate": 9.592716818399618e-05, "loss": 1.0379, "step": 3726 }, { "epoch": 0.25252388373196016, "grad_norm": 8.138580322265625, "learning_rate": 9.592579916489836e-05, "loss": 0.954, "step": 3727 }, { "epoch": 0.25259163899993226, "grad_norm": 6.810534477233887, "learning_rate": 9.592443014580054e-05, "loss": 0.7475, "step": 3728 }, { "epoch": 0.25265939426790435, "grad_norm": 6.66425895690918, "learning_rate": 9.592306112670272e-05, "loss": 0.8203, "step": 3729 }, { "epoch": 0.25272714953587644, "grad_norm": 7.962128162384033, "learning_rate": 9.592169210760492e-05, "loss": 0.9923, "step": 3730 }, { "epoch": 0.2527949048038485, "grad_norm": 8.372082710266113, "learning_rate": 9.59203230885071e-05, "loss": 0.809, "step": 3731 }, { "epoch": 0.25286266007182057, "grad_norm": 8.343878746032715, "learning_rate": 9.591895406940928e-05, "loss": 0.9373, "step": 3732 }, { "epoch": 0.25293041533979266, "grad_norm": 9.935523986816406, "learning_rate": 9.591758505031146e-05, "loss": 1.0037, "step": 3733 }, { "epoch": 0.25299817060776475, "grad_norm": 8.378336906433105, "learning_rate": 9.591621603121364e-05, "loss": 0.9834, "step": 3734 }, { "epoch": 0.25306592587573684, "grad_norm": 8.581600189208984, "learning_rate": 9.591484701211583e-05, "loss": 1.1192, "step": 3735 }, { "epoch": 0.25313368114370893, "grad_norm": 8.007279396057129, "learning_rate": 9.591347799301801e-05, "loss": 1.0261, "step": 3736 }, { "epoch": 0.253201436411681, "grad_norm": 7.40525484085083, "learning_rate": 9.591210897392019e-05, "loss": 0.9273, "step": 3737 }, { "epoch": 0.2532691916796531, "grad_norm": 7.228291988372803, "learning_rate": 9.591073995482237e-05, "loss": 1.0219, "step": 3738 }, { "epoch": 0.25333694694762515, "grad_norm": 6.714911460876465, "learning_rate": 9.590937093572457e-05, "loss": 0.9134, "step": 3739 }, { "epoch": 0.25340470221559724, "grad_norm": 6.150938510894775, "learning_rate": 9.590800191662675e-05, "loss": 0.8577, "step": 3740 }, { "epoch": 0.25347245748356934, "grad_norm": 10.693168640136719, "learning_rate": 9.590663289752893e-05, "loss": 1.2867, "step": 3741 }, { "epoch": 0.2535402127515414, "grad_norm": 8.114953994750977, "learning_rate": 9.590526387843111e-05, "loss": 1.0602, "step": 3742 }, { "epoch": 0.2536079680195135, "grad_norm": 8.472567558288574, "learning_rate": 9.590389485933329e-05, "loss": 0.9992, "step": 3743 }, { "epoch": 0.2536757232874856, "grad_norm": 8.681047439575195, "learning_rate": 9.590252584023548e-05, "loss": 1.003, "step": 3744 }, { "epoch": 0.2537434785554577, "grad_norm": 8.753557205200195, "learning_rate": 9.590115682113766e-05, "loss": 0.9447, "step": 3745 }, { "epoch": 0.2538112338234298, "grad_norm": 6.80125093460083, "learning_rate": 9.589978780203984e-05, "loss": 0.8522, "step": 3746 }, { "epoch": 0.25387898909140183, "grad_norm": 8.829830169677734, "learning_rate": 9.589841878294202e-05, "loss": 0.9131, "step": 3747 }, { "epoch": 0.2539467443593739, "grad_norm": 7.068274021148682, "learning_rate": 9.589704976384422e-05, "loss": 0.8348, "step": 3748 }, { "epoch": 0.254014499627346, "grad_norm": 8.19235897064209, "learning_rate": 9.58956807447464e-05, "loss": 0.8975, "step": 3749 }, { "epoch": 0.2540822548953181, "grad_norm": 8.1896333694458, "learning_rate": 9.589431172564858e-05, "loss": 1.1285, "step": 3750 }, { "epoch": 0.2541500101632902, "grad_norm": 7.581019878387451, "learning_rate": 9.589294270655076e-05, "loss": 0.8838, "step": 3751 }, { "epoch": 0.2542177654312623, "grad_norm": 6.806415557861328, "learning_rate": 9.589157368745294e-05, "loss": 0.9144, "step": 3752 }, { "epoch": 0.2542855206992344, "grad_norm": 6.077991485595703, "learning_rate": 9.589020466835513e-05, "loss": 0.7032, "step": 3753 }, { "epoch": 0.25435327596720647, "grad_norm": 9.278702735900879, "learning_rate": 9.588883564925731e-05, "loss": 1.1806, "step": 3754 }, { "epoch": 0.2544210312351785, "grad_norm": 7.136252403259277, "learning_rate": 9.588746663015949e-05, "loss": 0.9589, "step": 3755 }, { "epoch": 0.2544887865031506, "grad_norm": 6.349377632141113, "learning_rate": 9.588609761106167e-05, "loss": 0.9972, "step": 3756 }, { "epoch": 0.2545565417711227, "grad_norm": 10.120612144470215, "learning_rate": 9.588472859196385e-05, "loss": 1.124, "step": 3757 }, { "epoch": 0.2546242970390948, "grad_norm": 6.840261459350586, "learning_rate": 9.588335957286605e-05, "loss": 0.8897, "step": 3758 }, { "epoch": 0.2546920523070669, "grad_norm": 6.069836616516113, "learning_rate": 9.588199055376823e-05, "loss": 0.7794, "step": 3759 }, { "epoch": 0.25475980757503897, "grad_norm": 6.336123943328857, "learning_rate": 9.588062153467041e-05, "loss": 0.8936, "step": 3760 }, { "epoch": 0.25482756284301106, "grad_norm": 7.337663173675537, "learning_rate": 9.587925251557259e-05, "loss": 0.9241, "step": 3761 }, { "epoch": 0.25489531811098315, "grad_norm": 9.220332145690918, "learning_rate": 9.587788349647478e-05, "loss": 1.2426, "step": 3762 }, { "epoch": 0.25496307337895524, "grad_norm": 6.4654951095581055, "learning_rate": 9.587651447737696e-05, "loss": 0.7795, "step": 3763 }, { "epoch": 0.2550308286469273, "grad_norm": 8.458954811096191, "learning_rate": 9.587514545827914e-05, "loss": 1.0493, "step": 3764 }, { "epoch": 0.25509858391489937, "grad_norm": 6.73598575592041, "learning_rate": 9.587377643918132e-05, "loss": 0.9777, "step": 3765 }, { "epoch": 0.25516633918287146, "grad_norm": 8.673493385314941, "learning_rate": 9.587240742008352e-05, "loss": 1.1108, "step": 3766 }, { "epoch": 0.25523409445084355, "grad_norm": 7.328574180603027, "learning_rate": 9.58710384009857e-05, "loss": 0.9502, "step": 3767 }, { "epoch": 0.25530184971881564, "grad_norm": 7.9878692626953125, "learning_rate": 9.586966938188788e-05, "loss": 1.0191, "step": 3768 }, { "epoch": 0.25536960498678773, "grad_norm": 7.247650623321533, "learning_rate": 9.586830036279007e-05, "loss": 0.9356, "step": 3769 }, { "epoch": 0.2554373602547598, "grad_norm": 7.29271125793457, "learning_rate": 9.586693134369225e-05, "loss": 0.9877, "step": 3770 }, { "epoch": 0.2555051155227319, "grad_norm": 7.802029132843018, "learning_rate": 9.586556232459443e-05, "loss": 0.8923, "step": 3771 }, { "epoch": 0.25557287079070395, "grad_norm": 8.129645347595215, "learning_rate": 9.586419330549663e-05, "loss": 0.9451, "step": 3772 }, { "epoch": 0.25564062605867605, "grad_norm": 7.033285140991211, "learning_rate": 9.58628242863988e-05, "loss": 0.8529, "step": 3773 }, { "epoch": 0.25570838132664814, "grad_norm": 7.490065574645996, "learning_rate": 9.586145526730099e-05, "loss": 0.8912, "step": 3774 }, { "epoch": 0.25577613659462023, "grad_norm": 8.730104446411133, "learning_rate": 9.586008624820317e-05, "loss": 1.2685, "step": 3775 }, { "epoch": 0.2558438918625923, "grad_norm": 9.958065032958984, "learning_rate": 9.585871722910536e-05, "loss": 1.1792, "step": 3776 }, { "epoch": 0.2559116471305644, "grad_norm": 6.82180118560791, "learning_rate": 9.585734821000754e-05, "loss": 0.8734, "step": 3777 }, { "epoch": 0.2559794023985365, "grad_norm": 9.444950103759766, "learning_rate": 9.585597919090972e-05, "loss": 1.0368, "step": 3778 }, { "epoch": 0.2560471576665086, "grad_norm": 9.437919616699219, "learning_rate": 9.58546101718119e-05, "loss": 1.2728, "step": 3779 }, { "epoch": 0.25611491293448063, "grad_norm": 8.925026893615723, "learning_rate": 9.585324115271408e-05, "loss": 0.944, "step": 3780 }, { "epoch": 0.2561826682024527, "grad_norm": 8.421260833740234, "learning_rate": 9.585187213361628e-05, "loss": 0.9913, "step": 3781 }, { "epoch": 0.2562504234704248, "grad_norm": 8.5851469039917, "learning_rate": 9.585050311451846e-05, "loss": 0.9885, "step": 3782 }, { "epoch": 0.2563181787383969, "grad_norm": 7.179548263549805, "learning_rate": 9.584913409542064e-05, "loss": 1.0647, "step": 3783 }, { "epoch": 0.256385934006369, "grad_norm": 8.60708999633789, "learning_rate": 9.584776507632282e-05, "loss": 1.0475, "step": 3784 }, { "epoch": 0.2564536892743411, "grad_norm": 8.462443351745605, "learning_rate": 9.584639605722501e-05, "loss": 1.1381, "step": 3785 }, { "epoch": 0.2565214445423132, "grad_norm": 7.044154644012451, "learning_rate": 9.584502703812719e-05, "loss": 0.8223, "step": 3786 }, { "epoch": 0.2565891998102853, "grad_norm": 8.842753410339355, "learning_rate": 9.584365801902937e-05, "loss": 1.1205, "step": 3787 }, { "epoch": 0.2566569550782573, "grad_norm": 7.294439792633057, "learning_rate": 9.584228899993155e-05, "loss": 0.9563, "step": 3788 }, { "epoch": 0.2567247103462294, "grad_norm": 7.842654228210449, "learning_rate": 9.584091998083373e-05, "loss": 1.2798, "step": 3789 }, { "epoch": 0.2567924656142015, "grad_norm": 5.479234218597412, "learning_rate": 9.583955096173593e-05, "loss": 0.7835, "step": 3790 }, { "epoch": 0.2568602208821736, "grad_norm": 7.670284271240234, "learning_rate": 9.58381819426381e-05, "loss": 0.8263, "step": 3791 }, { "epoch": 0.2569279761501457, "grad_norm": 8.564105987548828, "learning_rate": 9.583681292354029e-05, "loss": 1.032, "step": 3792 }, { "epoch": 0.25699573141811777, "grad_norm": 8.706098556518555, "learning_rate": 9.583544390444247e-05, "loss": 1.0166, "step": 3793 }, { "epoch": 0.25706348668608986, "grad_norm": 6.727125644683838, "learning_rate": 9.583407488534466e-05, "loss": 0.756, "step": 3794 }, { "epoch": 0.25713124195406195, "grad_norm": 8.941418647766113, "learning_rate": 9.583270586624684e-05, "loss": 1.0223, "step": 3795 }, { "epoch": 0.257198997222034, "grad_norm": 7.35167932510376, "learning_rate": 9.583133684714902e-05, "loss": 0.9117, "step": 3796 }, { "epoch": 0.2572667524900061, "grad_norm": 9.245199203491211, "learning_rate": 9.58299678280512e-05, "loss": 1.0131, "step": 3797 }, { "epoch": 0.25733450775797817, "grad_norm": 10.583107948303223, "learning_rate": 9.582859880895338e-05, "loss": 1.0452, "step": 3798 }, { "epoch": 0.25740226302595026, "grad_norm": 7.672145843505859, "learning_rate": 9.582722978985558e-05, "loss": 0.9961, "step": 3799 }, { "epoch": 0.25747001829392235, "grad_norm": 6.5851640701293945, "learning_rate": 9.582586077075776e-05, "loss": 0.9217, "step": 3800 }, { "epoch": 0.25753777356189445, "grad_norm": 9.444985389709473, "learning_rate": 9.582449175165994e-05, "loss": 1.1704, "step": 3801 }, { "epoch": 0.25760552882986654, "grad_norm": 7.136216163635254, "learning_rate": 9.582312273256212e-05, "loss": 0.9597, "step": 3802 }, { "epoch": 0.25767328409783863, "grad_norm": 9.74889087677002, "learning_rate": 9.582175371346431e-05, "loss": 0.9866, "step": 3803 }, { "epoch": 0.25774103936581066, "grad_norm": 7.692512512207031, "learning_rate": 9.582038469436649e-05, "loss": 1.0412, "step": 3804 }, { "epoch": 0.25780879463378276, "grad_norm": 7.882124423980713, "learning_rate": 9.581901567526867e-05, "loss": 1.1129, "step": 3805 }, { "epoch": 0.25787654990175485, "grad_norm": 8.117066383361816, "learning_rate": 9.581764665617085e-05, "loss": 0.968, "step": 3806 }, { "epoch": 0.25794430516972694, "grad_norm": 8.883692741394043, "learning_rate": 9.581627763707303e-05, "loss": 1.0859, "step": 3807 }, { "epoch": 0.25801206043769903, "grad_norm": 7.985278129577637, "learning_rate": 9.581490861797523e-05, "loss": 0.9511, "step": 3808 }, { "epoch": 0.2580798157056711, "grad_norm": 10.487812995910645, "learning_rate": 9.58135395988774e-05, "loss": 1.2395, "step": 3809 }, { "epoch": 0.2581475709736432, "grad_norm": 7.5507707595825195, "learning_rate": 9.581217057977959e-05, "loss": 0.9959, "step": 3810 }, { "epoch": 0.2582153262416153, "grad_norm": 9.950063705444336, "learning_rate": 9.581080156068177e-05, "loss": 0.9453, "step": 3811 }, { "epoch": 0.2582830815095874, "grad_norm": 7.301966190338135, "learning_rate": 9.580943254158396e-05, "loss": 0.8468, "step": 3812 }, { "epoch": 0.25835083677755943, "grad_norm": 6.434390544891357, "learning_rate": 9.580806352248614e-05, "loss": 1.0513, "step": 3813 }, { "epoch": 0.2584185920455315, "grad_norm": 8.878791809082031, "learning_rate": 9.580669450338832e-05, "loss": 1.1264, "step": 3814 }, { "epoch": 0.2584863473135036, "grad_norm": 8.107111930847168, "learning_rate": 9.580532548429052e-05, "loss": 1.3037, "step": 3815 }, { "epoch": 0.2585541025814757, "grad_norm": 7.54078483581543, "learning_rate": 9.58039564651927e-05, "loss": 0.917, "step": 3816 }, { "epoch": 0.2586218578494478, "grad_norm": 9.369047164916992, "learning_rate": 9.580258744609488e-05, "loss": 0.8761, "step": 3817 }, { "epoch": 0.2586896131174199, "grad_norm": 8.927732467651367, "learning_rate": 9.580121842699707e-05, "loss": 1.0591, "step": 3818 }, { "epoch": 0.258757368385392, "grad_norm": 9.261579513549805, "learning_rate": 9.579984940789925e-05, "loss": 1.0671, "step": 3819 }, { "epoch": 0.2588251236533641, "grad_norm": 7.396904468536377, "learning_rate": 9.579848038880143e-05, "loss": 0.8242, "step": 3820 }, { "epoch": 0.2588928789213361, "grad_norm": 6.665501594543457, "learning_rate": 9.579711136970361e-05, "loss": 0.7686, "step": 3821 }, { "epoch": 0.2589606341893082, "grad_norm": 7.0808281898498535, "learning_rate": 9.57957423506058e-05, "loss": 0.9757, "step": 3822 }, { "epoch": 0.2590283894572803, "grad_norm": 7.734886646270752, "learning_rate": 9.579437333150799e-05, "loss": 1.0365, "step": 3823 }, { "epoch": 0.2590961447252524, "grad_norm": 7.383622169494629, "learning_rate": 9.579300431241017e-05, "loss": 0.8458, "step": 3824 }, { "epoch": 0.2591638999932245, "grad_norm": 7.758030891418457, "learning_rate": 9.579163529331235e-05, "loss": 0.9454, "step": 3825 }, { "epoch": 0.25923165526119657, "grad_norm": 9.456972122192383, "learning_rate": 9.579026627421454e-05, "loss": 1.065, "step": 3826 }, { "epoch": 0.25929941052916866, "grad_norm": 8.046957969665527, "learning_rate": 9.578889725511672e-05, "loss": 0.7021, "step": 3827 }, { "epoch": 0.25936716579714075, "grad_norm": 9.415145874023438, "learning_rate": 9.57875282360189e-05, "loss": 1.2344, "step": 3828 }, { "epoch": 0.2594349210651128, "grad_norm": 6.525821685791016, "learning_rate": 9.578615921692108e-05, "loss": 0.8005, "step": 3829 }, { "epoch": 0.2595026763330849, "grad_norm": 6.948854923248291, "learning_rate": 9.578479019782326e-05, "loss": 0.8313, "step": 3830 }, { "epoch": 0.259570431601057, "grad_norm": 6.991540431976318, "learning_rate": 9.578342117872546e-05, "loss": 0.945, "step": 3831 }, { "epoch": 0.25963818686902906, "grad_norm": 7.0269551277160645, "learning_rate": 9.578205215962764e-05, "loss": 0.8927, "step": 3832 }, { "epoch": 0.25970594213700116, "grad_norm": 7.773914813995361, "learning_rate": 9.578068314052982e-05, "loss": 0.8469, "step": 3833 }, { "epoch": 0.25977369740497325, "grad_norm": 9.6503267288208, "learning_rate": 9.5779314121432e-05, "loss": 1.1192, "step": 3834 }, { "epoch": 0.25984145267294534, "grad_norm": 6.582554340362549, "learning_rate": 9.577794510233418e-05, "loss": 0.8996, "step": 3835 }, { "epoch": 0.25990920794091743, "grad_norm": 10.097637176513672, "learning_rate": 9.577657608323637e-05, "loss": 1.211, "step": 3836 }, { "epoch": 0.25997696320888947, "grad_norm": 8.63124942779541, "learning_rate": 9.577520706413855e-05, "loss": 1.0353, "step": 3837 }, { "epoch": 0.26004471847686156, "grad_norm": 8.282122611999512, "learning_rate": 9.577383804504073e-05, "loss": 0.9677, "step": 3838 }, { "epoch": 0.26011247374483365, "grad_norm": 8.298484802246094, "learning_rate": 9.577246902594291e-05, "loss": 0.909, "step": 3839 }, { "epoch": 0.26018022901280574, "grad_norm": 7.91752290725708, "learning_rate": 9.57711000068451e-05, "loss": 1.1106, "step": 3840 }, { "epoch": 0.26024798428077783, "grad_norm": 8.073543548583984, "learning_rate": 9.576973098774729e-05, "loss": 0.9777, "step": 3841 }, { "epoch": 0.2603157395487499, "grad_norm": 8.225390434265137, "learning_rate": 9.576836196864947e-05, "loss": 1.0302, "step": 3842 }, { "epoch": 0.260383494816722, "grad_norm": 7.086613655090332, "learning_rate": 9.576699294955165e-05, "loss": 1.0061, "step": 3843 }, { "epoch": 0.2604512500846941, "grad_norm": 6.9043965339660645, "learning_rate": 9.576562393045383e-05, "loss": 0.8707, "step": 3844 }, { "epoch": 0.26051900535266614, "grad_norm": 9.083130836486816, "learning_rate": 9.576425491135602e-05, "loss": 1.0567, "step": 3845 }, { "epoch": 0.26058676062063824, "grad_norm": 6.962080478668213, "learning_rate": 9.57628858922582e-05, "loss": 0.8592, "step": 3846 }, { "epoch": 0.2606545158886103, "grad_norm": 7.196011066436768, "learning_rate": 9.576151687316038e-05, "loss": 0.8318, "step": 3847 }, { "epoch": 0.2607222711565824, "grad_norm": 7.421074867248535, "learning_rate": 9.576014785406256e-05, "loss": 1.0162, "step": 3848 }, { "epoch": 0.2607900264245545, "grad_norm": 7.065299034118652, "learning_rate": 9.575877883496476e-05, "loss": 0.9999, "step": 3849 }, { "epoch": 0.2608577816925266, "grad_norm": 7.442328929901123, "learning_rate": 9.575740981586694e-05, "loss": 0.8118, "step": 3850 }, { "epoch": 0.2609255369604987, "grad_norm": 6.888897895812988, "learning_rate": 9.575604079676912e-05, "loss": 1.0343, "step": 3851 }, { "epoch": 0.2609932922284708, "grad_norm": 7.788427829742432, "learning_rate": 9.57546717776713e-05, "loss": 0.9931, "step": 3852 }, { "epoch": 0.2610610474964428, "grad_norm": 7.247363567352295, "learning_rate": 9.575330275857348e-05, "loss": 0.8403, "step": 3853 }, { "epoch": 0.2611288027644149, "grad_norm": 7.305066108703613, "learning_rate": 9.575193373947567e-05, "loss": 1.0222, "step": 3854 }, { "epoch": 0.261196558032387, "grad_norm": 5.8723249435424805, "learning_rate": 9.575056472037785e-05, "loss": 0.7461, "step": 3855 }, { "epoch": 0.2612643133003591, "grad_norm": 8.933609008789062, "learning_rate": 9.574919570128003e-05, "loss": 1.1537, "step": 3856 }, { "epoch": 0.2613320685683312, "grad_norm": 7.3203125, "learning_rate": 9.574782668218221e-05, "loss": 1.0277, "step": 3857 }, { "epoch": 0.2613998238363033, "grad_norm": 7.455322742462158, "learning_rate": 9.57464576630844e-05, "loss": 0.8136, "step": 3858 }, { "epoch": 0.26146757910427537, "grad_norm": 8.074299812316895, "learning_rate": 9.574508864398659e-05, "loss": 0.9597, "step": 3859 }, { "epoch": 0.26153533437224746, "grad_norm": 8.732856750488281, "learning_rate": 9.574371962488877e-05, "loss": 1.2009, "step": 3860 }, { "epoch": 0.2616030896402195, "grad_norm": 7.179652690887451, "learning_rate": 9.574235060579095e-05, "loss": 0.8365, "step": 3861 }, { "epoch": 0.2616708449081916, "grad_norm": 8.975394248962402, "learning_rate": 9.574098158669314e-05, "loss": 0.9099, "step": 3862 }, { "epoch": 0.2617386001761637, "grad_norm": 9.083860397338867, "learning_rate": 9.573961256759532e-05, "loss": 1.0307, "step": 3863 }, { "epoch": 0.2618063554441358, "grad_norm": 7.449617385864258, "learning_rate": 9.57382435484975e-05, "loss": 1.1442, "step": 3864 }, { "epoch": 0.26187411071210787, "grad_norm": 8.299210548400879, "learning_rate": 9.57368745293997e-05, "loss": 1.0119, "step": 3865 }, { "epoch": 0.26194186598007996, "grad_norm": 6.847742080688477, "learning_rate": 9.573550551030188e-05, "loss": 0.9058, "step": 3866 }, { "epoch": 0.26200962124805205, "grad_norm": 7.057496070861816, "learning_rate": 9.573413649120406e-05, "loss": 1.0022, "step": 3867 }, { "epoch": 0.26207737651602414, "grad_norm": 7.9798359870910645, "learning_rate": 9.573276747210625e-05, "loss": 0.9555, "step": 3868 }, { "epoch": 0.26214513178399623, "grad_norm": 8.119134902954102, "learning_rate": 9.573139845300843e-05, "loss": 1.1039, "step": 3869 }, { "epoch": 0.26221288705196827, "grad_norm": 6.998579502105713, "learning_rate": 9.573002943391061e-05, "loss": 1.0448, "step": 3870 }, { "epoch": 0.26228064231994036, "grad_norm": 6.59659481048584, "learning_rate": 9.572866041481279e-05, "loss": 0.8946, "step": 3871 }, { "epoch": 0.26234839758791245, "grad_norm": 8.110078811645508, "learning_rate": 9.572729139571498e-05, "loss": 0.8568, "step": 3872 }, { "epoch": 0.26241615285588454, "grad_norm": 9.192879676818848, "learning_rate": 9.572592237661717e-05, "loss": 0.9517, "step": 3873 }, { "epoch": 0.26248390812385664, "grad_norm": 7.378695964813232, "learning_rate": 9.572455335751935e-05, "loss": 1.0917, "step": 3874 }, { "epoch": 0.2625516633918287, "grad_norm": 10.016194343566895, "learning_rate": 9.572318433842153e-05, "loss": 0.9754, "step": 3875 }, { "epoch": 0.2626194186598008, "grad_norm": 7.392832279205322, "learning_rate": 9.57218153193237e-05, "loss": 0.8258, "step": 3876 }, { "epoch": 0.2626871739277729, "grad_norm": 6.59785795211792, "learning_rate": 9.57204463002259e-05, "loss": 0.9656, "step": 3877 }, { "epoch": 0.26275492919574495, "grad_norm": 7.799993991851807, "learning_rate": 9.571907728112808e-05, "loss": 0.9776, "step": 3878 }, { "epoch": 0.26282268446371704, "grad_norm": 9.275368690490723, "learning_rate": 9.571770826203026e-05, "loss": 0.8827, "step": 3879 }, { "epoch": 0.26289043973168913, "grad_norm": 9.413054466247559, "learning_rate": 9.571633924293244e-05, "loss": 1.0671, "step": 3880 }, { "epoch": 0.2629581949996612, "grad_norm": 8.446796417236328, "learning_rate": 9.571497022383464e-05, "loss": 0.8702, "step": 3881 }, { "epoch": 0.2630259502676333, "grad_norm": 12.962898254394531, "learning_rate": 9.571360120473682e-05, "loss": 0.9281, "step": 3882 }, { "epoch": 0.2630937055356054, "grad_norm": 7.7250213623046875, "learning_rate": 9.5712232185639e-05, "loss": 0.8908, "step": 3883 }, { "epoch": 0.2631614608035775, "grad_norm": 10.405988693237305, "learning_rate": 9.571086316654118e-05, "loss": 1.2047, "step": 3884 }, { "epoch": 0.2632292160715496, "grad_norm": 9.009016036987305, "learning_rate": 9.570949414744336e-05, "loss": 1.0619, "step": 3885 }, { "epoch": 0.2632969713395216, "grad_norm": 8.600632667541504, "learning_rate": 9.570812512834555e-05, "loss": 1.4186, "step": 3886 }, { "epoch": 0.2633647266074937, "grad_norm": 7.9944071769714355, "learning_rate": 9.570675610924773e-05, "loss": 1.1493, "step": 3887 }, { "epoch": 0.2634324818754658, "grad_norm": 7.7023539543151855, "learning_rate": 9.570538709014991e-05, "loss": 0.8716, "step": 3888 }, { "epoch": 0.2635002371434379, "grad_norm": 7.743750095367432, "learning_rate": 9.570401807105209e-05, "loss": 0.9091, "step": 3889 }, { "epoch": 0.26356799241141, "grad_norm": 7.080264091491699, "learning_rate": 9.570264905195427e-05, "loss": 0.8064, "step": 3890 }, { "epoch": 0.2636357476793821, "grad_norm": 10.498579025268555, "learning_rate": 9.570128003285647e-05, "loss": 1.2763, "step": 3891 }, { "epoch": 0.2637035029473542, "grad_norm": 8.932741165161133, "learning_rate": 9.569991101375865e-05, "loss": 1.0108, "step": 3892 }, { "epoch": 0.26377125821532627, "grad_norm": 7.670261859893799, "learning_rate": 9.569854199466083e-05, "loss": 0.9312, "step": 3893 }, { "epoch": 0.2638390134832983, "grad_norm": 6.713436126708984, "learning_rate": 9.5697172975563e-05, "loss": 0.9319, "step": 3894 }, { "epoch": 0.2639067687512704, "grad_norm": 6.828521728515625, "learning_rate": 9.56958039564652e-05, "loss": 0.908, "step": 3895 }, { "epoch": 0.2639745240192425, "grad_norm": 7.664526462554932, "learning_rate": 9.569443493736738e-05, "loss": 0.9204, "step": 3896 }, { "epoch": 0.2640422792872146, "grad_norm": 7.330194473266602, "learning_rate": 9.569306591826956e-05, "loss": 0.9971, "step": 3897 }, { "epoch": 0.26411003455518667, "grad_norm": 7.202576160430908, "learning_rate": 9.569169689917174e-05, "loss": 1.1183, "step": 3898 }, { "epoch": 0.26417778982315876, "grad_norm": 7.7107720375061035, "learning_rate": 9.569032788007392e-05, "loss": 1.0794, "step": 3899 }, { "epoch": 0.26424554509113085, "grad_norm": 8.634172439575195, "learning_rate": 9.568895886097612e-05, "loss": 0.9124, "step": 3900 }, { "epoch": 0.26431330035910294, "grad_norm": 7.264395236968994, "learning_rate": 9.56875898418783e-05, "loss": 1.0044, "step": 3901 }, { "epoch": 0.264381055627075, "grad_norm": 7.352424144744873, "learning_rate": 9.568622082278048e-05, "loss": 0.9213, "step": 3902 }, { "epoch": 0.26444881089504707, "grad_norm": 8.39152717590332, "learning_rate": 9.568485180368266e-05, "loss": 0.914, "step": 3903 }, { "epoch": 0.26451656616301916, "grad_norm": 6.5833611488342285, "learning_rate": 9.568348278458485e-05, "loss": 0.9906, "step": 3904 }, { "epoch": 0.26458432143099125, "grad_norm": 7.952385902404785, "learning_rate": 9.568211376548703e-05, "loss": 0.9624, "step": 3905 }, { "epoch": 0.26465207669896335, "grad_norm": 7.232090950012207, "learning_rate": 9.568074474638921e-05, "loss": 0.8852, "step": 3906 }, { "epoch": 0.26471983196693544, "grad_norm": 7.817921161651611, "learning_rate": 9.567937572729139e-05, "loss": 0.8474, "step": 3907 }, { "epoch": 0.26478758723490753, "grad_norm": 8.926132202148438, "learning_rate": 9.567800670819359e-05, "loss": 1.0302, "step": 3908 }, { "epoch": 0.2648553425028796, "grad_norm": 8.923449516296387, "learning_rate": 9.567663768909577e-05, "loss": 1.0837, "step": 3909 }, { "epoch": 0.26492309777085166, "grad_norm": 8.431096076965332, "learning_rate": 9.567526866999795e-05, "loss": 0.7992, "step": 3910 }, { "epoch": 0.26499085303882375, "grad_norm": 10.121541976928711, "learning_rate": 9.567389965090014e-05, "loss": 1.0082, "step": 3911 }, { "epoch": 0.26505860830679584, "grad_norm": 9.349747657775879, "learning_rate": 9.567253063180232e-05, "loss": 0.8257, "step": 3912 }, { "epoch": 0.26512636357476793, "grad_norm": 9.438392639160156, "learning_rate": 9.56711616127045e-05, "loss": 0.979, "step": 3913 }, { "epoch": 0.26519411884274, "grad_norm": 7.422990798950195, "learning_rate": 9.56697925936067e-05, "loss": 0.8456, "step": 3914 }, { "epoch": 0.2652618741107121, "grad_norm": 6.354233741760254, "learning_rate": 9.566842357450888e-05, "loss": 0.6639, "step": 3915 }, { "epoch": 0.2653296293786842, "grad_norm": 9.160786628723145, "learning_rate": 9.566705455541106e-05, "loss": 1.1126, "step": 3916 }, { "epoch": 0.2653973846466563, "grad_norm": 7.786096096038818, "learning_rate": 9.566568553631324e-05, "loss": 1.2703, "step": 3917 }, { "epoch": 0.2654651399146284, "grad_norm": 7.358225345611572, "learning_rate": 9.566431651721543e-05, "loss": 1.0625, "step": 3918 }, { "epoch": 0.2655328951826004, "grad_norm": 8.920319557189941, "learning_rate": 9.566294749811761e-05, "loss": 1.0037, "step": 3919 }, { "epoch": 0.2656006504505725, "grad_norm": 7.16439151763916, "learning_rate": 9.566157847901979e-05, "loss": 1.0486, "step": 3920 }, { "epoch": 0.2656684057185446, "grad_norm": 7.374850749969482, "learning_rate": 9.566020945992197e-05, "loss": 1.011, "step": 3921 }, { "epoch": 0.2657361609865167, "grad_norm": 5.965388298034668, "learning_rate": 9.565884044082415e-05, "loss": 0.9043, "step": 3922 }, { "epoch": 0.2658039162544888, "grad_norm": 7.1143879890441895, "learning_rate": 9.565747142172634e-05, "loss": 0.923, "step": 3923 }, { "epoch": 0.2658716715224609, "grad_norm": 9.05667495727539, "learning_rate": 9.565610240262853e-05, "loss": 1.1247, "step": 3924 }, { "epoch": 0.265939426790433, "grad_norm": 6.407328128814697, "learning_rate": 9.56547333835307e-05, "loss": 0.9028, "step": 3925 }, { "epoch": 0.26600718205840507, "grad_norm": 9.335012435913086, "learning_rate": 9.565336436443289e-05, "loss": 0.9761, "step": 3926 }, { "epoch": 0.2660749373263771, "grad_norm": 7.462203025817871, "learning_rate": 9.565199534533508e-05, "loss": 0.8812, "step": 3927 }, { "epoch": 0.2661426925943492, "grad_norm": 8.162378311157227, "learning_rate": 9.565062632623726e-05, "loss": 0.991, "step": 3928 }, { "epoch": 0.2662104478623213, "grad_norm": 8.835287094116211, "learning_rate": 9.564925730713944e-05, "loss": 1.3085, "step": 3929 }, { "epoch": 0.2662782031302934, "grad_norm": 9.219624519348145, "learning_rate": 9.564788828804162e-05, "loss": 0.989, "step": 3930 }, { "epoch": 0.26634595839826547, "grad_norm": 6.832587718963623, "learning_rate": 9.56465192689438e-05, "loss": 1.0186, "step": 3931 }, { "epoch": 0.26641371366623756, "grad_norm": 8.075157165527344, "learning_rate": 9.5645150249846e-05, "loss": 0.9734, "step": 3932 }, { "epoch": 0.26648146893420965, "grad_norm": 9.403346061706543, "learning_rate": 9.564378123074818e-05, "loss": 1.1875, "step": 3933 }, { "epoch": 0.26654922420218174, "grad_norm": 8.556446075439453, "learning_rate": 9.564241221165036e-05, "loss": 1.0932, "step": 3934 }, { "epoch": 0.2666169794701538, "grad_norm": 8.178442001342773, "learning_rate": 9.564104319255254e-05, "loss": 0.7654, "step": 3935 }, { "epoch": 0.2666847347381259, "grad_norm": 8.82776165008545, "learning_rate": 9.563967417345473e-05, "loss": 1.0518, "step": 3936 }, { "epoch": 0.26675249000609796, "grad_norm": 8.001256942749023, "learning_rate": 9.563830515435691e-05, "loss": 1.0109, "step": 3937 }, { "epoch": 0.26682024527407006, "grad_norm": 7.439608573913574, "learning_rate": 9.563693613525909e-05, "loss": 0.9365, "step": 3938 }, { "epoch": 0.26688800054204215, "grad_norm": 5.78077507019043, "learning_rate": 9.563556711616127e-05, "loss": 0.9356, "step": 3939 }, { "epoch": 0.26695575581001424, "grad_norm": 7.6134819984436035, "learning_rate": 9.563419809706345e-05, "loss": 1.0003, "step": 3940 }, { "epoch": 0.26702351107798633, "grad_norm": 8.467934608459473, "learning_rate": 9.563282907796565e-05, "loss": 0.8052, "step": 3941 }, { "epoch": 0.2670912663459584, "grad_norm": 8.88598346710205, "learning_rate": 9.563146005886783e-05, "loss": 1.0069, "step": 3942 }, { "epoch": 0.26715902161393046, "grad_norm": 7.627633094787598, "learning_rate": 9.563009103977e-05, "loss": 1.1463, "step": 3943 }, { "epoch": 0.26722677688190255, "grad_norm": 7.948824882507324, "learning_rate": 9.562872202067219e-05, "loss": 0.9451, "step": 3944 }, { "epoch": 0.26729453214987464, "grad_norm": 8.10439395904541, "learning_rate": 9.562735300157437e-05, "loss": 1.1375, "step": 3945 }, { "epoch": 0.26736228741784673, "grad_norm": 6.488743782043457, "learning_rate": 9.562598398247656e-05, "loss": 0.8608, "step": 3946 }, { "epoch": 0.2674300426858188, "grad_norm": 9.731819152832031, "learning_rate": 9.562461496337874e-05, "loss": 1.2055, "step": 3947 }, { "epoch": 0.2674977979537909, "grad_norm": 10.455330848693848, "learning_rate": 9.562324594428092e-05, "loss": 1.1095, "step": 3948 }, { "epoch": 0.267565553221763, "grad_norm": 6.7713212966918945, "learning_rate": 9.56218769251831e-05, "loss": 0.9047, "step": 3949 }, { "epoch": 0.2676333084897351, "grad_norm": 9.292582511901855, "learning_rate": 9.56205079060853e-05, "loss": 0.9404, "step": 3950 }, { "epoch": 0.26770106375770714, "grad_norm": 8.252067565917969, "learning_rate": 9.561913888698748e-05, "loss": 0.8826, "step": 3951 }, { "epoch": 0.26776881902567923, "grad_norm": 7.126963138580322, "learning_rate": 9.561776986788966e-05, "loss": 0.8445, "step": 3952 }, { "epoch": 0.2678365742936513, "grad_norm": 8.352923393249512, "learning_rate": 9.561640084879184e-05, "loss": 1.1081, "step": 3953 }, { "epoch": 0.2679043295616234, "grad_norm": 6.933292388916016, "learning_rate": 9.561503182969403e-05, "loss": 0.8228, "step": 3954 }, { "epoch": 0.2679720848295955, "grad_norm": 7.9115986824035645, "learning_rate": 9.561366281059621e-05, "loss": 1.0984, "step": 3955 }, { "epoch": 0.2680398400975676, "grad_norm": 6.988186359405518, "learning_rate": 9.561229379149839e-05, "loss": 0.9334, "step": 3956 }, { "epoch": 0.2681075953655397, "grad_norm": 6.6764631271362305, "learning_rate": 9.561092477240058e-05, "loss": 0.907, "step": 3957 }, { "epoch": 0.2681753506335118, "grad_norm": 9.355245590209961, "learning_rate": 9.560955575330277e-05, "loss": 0.9087, "step": 3958 }, { "epoch": 0.2682431059014838, "grad_norm": 8.177611351013184, "learning_rate": 9.560818673420495e-05, "loss": 1.0815, "step": 3959 }, { "epoch": 0.2683108611694559, "grad_norm": 9.085289001464844, "learning_rate": 9.560681771510714e-05, "loss": 1.1291, "step": 3960 }, { "epoch": 0.268378616437428, "grad_norm": 7.056759357452393, "learning_rate": 9.560544869600932e-05, "loss": 0.8623, "step": 3961 }, { "epoch": 0.2684463717054001, "grad_norm": 8.175825119018555, "learning_rate": 9.56040796769115e-05, "loss": 0.8755, "step": 3962 }, { "epoch": 0.2685141269733722, "grad_norm": 7.127376079559326, "learning_rate": 9.560271065781368e-05, "loss": 0.8877, "step": 3963 }, { "epoch": 0.26858188224134427, "grad_norm": 9.635464668273926, "learning_rate": 9.560134163871587e-05, "loss": 0.8675, "step": 3964 }, { "epoch": 0.26864963750931636, "grad_norm": 7.8138275146484375, "learning_rate": 9.559997261961805e-05, "loss": 0.9997, "step": 3965 }, { "epoch": 0.26871739277728846, "grad_norm": 7.242639064788818, "learning_rate": 9.559860360052024e-05, "loss": 0.7692, "step": 3966 }, { "epoch": 0.2687851480452605, "grad_norm": 8.106497764587402, "learning_rate": 9.559723458142242e-05, "loss": 0.9801, "step": 3967 }, { "epoch": 0.2688529033132326, "grad_norm": 7.468952178955078, "learning_rate": 9.55958655623246e-05, "loss": 0.939, "step": 3968 }, { "epoch": 0.2689206585812047, "grad_norm": 9.002805709838867, "learning_rate": 9.559449654322679e-05, "loss": 1.1655, "step": 3969 }, { "epoch": 0.26898841384917677, "grad_norm": 8.787810325622559, "learning_rate": 9.559312752412897e-05, "loss": 0.9663, "step": 3970 }, { "epoch": 0.26905616911714886, "grad_norm": 9.113668441772461, "learning_rate": 9.559175850503115e-05, "loss": 1.0041, "step": 3971 }, { "epoch": 0.26912392438512095, "grad_norm": 6.289670944213867, "learning_rate": 9.559038948593333e-05, "loss": 0.8714, "step": 3972 }, { "epoch": 0.26919167965309304, "grad_norm": 7.755144119262695, "learning_rate": 9.558902046683552e-05, "loss": 0.8606, "step": 3973 }, { "epoch": 0.26925943492106513, "grad_norm": 8.989197731018066, "learning_rate": 9.55876514477377e-05, "loss": 0.9956, "step": 3974 }, { "epoch": 0.2693271901890372, "grad_norm": 6.45689058303833, "learning_rate": 9.558628242863989e-05, "loss": 0.7416, "step": 3975 }, { "epoch": 0.26939494545700926, "grad_norm": 7.772951126098633, "learning_rate": 9.558491340954207e-05, "loss": 0.927, "step": 3976 }, { "epoch": 0.26946270072498135, "grad_norm": 7.347445487976074, "learning_rate": 9.558354439044425e-05, "loss": 0.9337, "step": 3977 }, { "epoch": 0.26953045599295344, "grad_norm": 8.20067310333252, "learning_rate": 9.558217537134644e-05, "loss": 0.8324, "step": 3978 }, { "epoch": 0.26959821126092554, "grad_norm": 10.375189781188965, "learning_rate": 9.558080635224862e-05, "loss": 1.1103, "step": 3979 }, { "epoch": 0.2696659665288976, "grad_norm": 8.187355041503906, "learning_rate": 9.55794373331508e-05, "loss": 0.8469, "step": 3980 }, { "epoch": 0.2697337217968697, "grad_norm": 6.336839199066162, "learning_rate": 9.557806831405298e-05, "loss": 1.037, "step": 3981 }, { "epoch": 0.2698014770648418, "grad_norm": 6.368093967437744, "learning_rate": 9.557669929495517e-05, "loss": 0.6851, "step": 3982 }, { "epoch": 0.2698692323328139, "grad_norm": 7.394474506378174, "learning_rate": 9.557533027585736e-05, "loss": 1.0921, "step": 3983 }, { "epoch": 0.26993698760078594, "grad_norm": 9.0152006149292, "learning_rate": 9.557396125675954e-05, "loss": 0.9765, "step": 3984 }, { "epoch": 0.27000474286875803, "grad_norm": 8.247949600219727, "learning_rate": 9.557259223766172e-05, "loss": 1.0941, "step": 3985 }, { "epoch": 0.2700724981367301, "grad_norm": 7.9166579246521, "learning_rate": 9.55712232185639e-05, "loss": 0.798, "step": 3986 }, { "epoch": 0.2701402534047022, "grad_norm": 8.501713752746582, "learning_rate": 9.556985419946609e-05, "loss": 0.923, "step": 3987 }, { "epoch": 0.2702080086726743, "grad_norm": 7.450741767883301, "learning_rate": 9.556848518036827e-05, "loss": 0.9485, "step": 3988 }, { "epoch": 0.2702757639406464, "grad_norm": 6.7950239181518555, "learning_rate": 9.556711616127045e-05, "loss": 0.9041, "step": 3989 }, { "epoch": 0.2703435192086185, "grad_norm": 8.021660804748535, "learning_rate": 9.556574714217263e-05, "loss": 0.9494, "step": 3990 }, { "epoch": 0.2704112744765906, "grad_norm": 8.049949645996094, "learning_rate": 9.556437812307482e-05, "loss": 0.9629, "step": 3991 }, { "epoch": 0.2704790297445626, "grad_norm": 9.56615161895752, "learning_rate": 9.5563009103977e-05, "loss": 0.8822, "step": 3992 }, { "epoch": 0.2705467850125347, "grad_norm": 7.12232780456543, "learning_rate": 9.556164008487919e-05, "loss": 1.1377, "step": 3993 }, { "epoch": 0.2706145402805068, "grad_norm": 7.224277496337891, "learning_rate": 9.556027106578137e-05, "loss": 1.1023, "step": 3994 }, { "epoch": 0.2706822955484789, "grad_norm": 8.0076322555542, "learning_rate": 9.555890204668355e-05, "loss": 1.0599, "step": 3995 }, { "epoch": 0.270750050816451, "grad_norm": 7.8958845138549805, "learning_rate": 9.555753302758574e-05, "loss": 1.0455, "step": 3996 }, { "epoch": 0.2708178060844231, "grad_norm": 7.802896976470947, "learning_rate": 9.555616400848792e-05, "loss": 1.1089, "step": 3997 }, { "epoch": 0.27088556135239517, "grad_norm": 8.122269630432129, "learning_rate": 9.55547949893901e-05, "loss": 1.0708, "step": 3998 }, { "epoch": 0.27095331662036726, "grad_norm": 6.3488383293151855, "learning_rate": 9.555342597029228e-05, "loss": 0.85, "step": 3999 }, { "epoch": 0.2710210718883393, "grad_norm": 10.236666679382324, "learning_rate": 9.555205695119448e-05, "loss": 0.962, "step": 4000 }, { "epoch": 0.2710888271563114, "grad_norm": 7.594062328338623, "learning_rate": 9.555068793209666e-05, "loss": 1.1641, "step": 4001 }, { "epoch": 0.2711565824242835, "grad_norm": 6.824306964874268, "learning_rate": 9.554931891299884e-05, "loss": 0.7914, "step": 4002 }, { "epoch": 0.27122433769225557, "grad_norm": 7.826432228088379, "learning_rate": 9.554794989390103e-05, "loss": 1.0062, "step": 4003 }, { "epoch": 0.27129209296022766, "grad_norm": 7.189459800720215, "learning_rate": 9.554658087480321e-05, "loss": 0.7324, "step": 4004 }, { "epoch": 0.27135984822819975, "grad_norm": 7.949024200439453, "learning_rate": 9.554521185570539e-05, "loss": 1.1033, "step": 4005 }, { "epoch": 0.27142760349617184, "grad_norm": 6.575378894805908, "learning_rate": 9.554384283660758e-05, "loss": 1.0048, "step": 4006 }, { "epoch": 0.27149535876414393, "grad_norm": 8.585273742675781, "learning_rate": 9.554247381750976e-05, "loss": 1.0338, "step": 4007 }, { "epoch": 0.27156311403211597, "grad_norm": 7.999851703643799, "learning_rate": 9.554110479841194e-05, "loss": 1.1788, "step": 4008 }, { "epoch": 0.27163086930008806, "grad_norm": 8.40134334564209, "learning_rate": 9.553973577931413e-05, "loss": 1.0986, "step": 4009 }, { "epoch": 0.27169862456806015, "grad_norm": 6.380734920501709, "learning_rate": 9.553836676021632e-05, "loss": 0.717, "step": 4010 }, { "epoch": 0.27176637983603225, "grad_norm": 7.673857688903809, "learning_rate": 9.55369977411185e-05, "loss": 1.1646, "step": 4011 }, { "epoch": 0.27183413510400434, "grad_norm": 10.897799491882324, "learning_rate": 9.553562872202068e-05, "loss": 1.2796, "step": 4012 }, { "epoch": 0.27190189037197643, "grad_norm": 9.101582527160645, "learning_rate": 9.553425970292286e-05, "loss": 0.8778, "step": 4013 }, { "epoch": 0.2719696456399485, "grad_norm": 6.685849666595459, "learning_rate": 9.553289068382505e-05, "loss": 1.2329, "step": 4014 }, { "epoch": 0.2720374009079206, "grad_norm": 6.580325603485107, "learning_rate": 9.553152166472723e-05, "loss": 0.81, "step": 4015 }, { "epoch": 0.27210515617589265, "grad_norm": 7.854914665222168, "learning_rate": 9.553015264562941e-05, "loss": 1.1152, "step": 4016 }, { "epoch": 0.27217291144386474, "grad_norm": 7.293428421020508, "learning_rate": 9.55287836265316e-05, "loss": 0.9238, "step": 4017 }, { "epoch": 0.27224066671183683, "grad_norm": 6.944539546966553, "learning_rate": 9.552741460743378e-05, "loss": 1.018, "step": 4018 }, { "epoch": 0.2723084219798089, "grad_norm": 7.550015926361084, "learning_rate": 9.552604558833597e-05, "loss": 1.0332, "step": 4019 }, { "epoch": 0.272376177247781, "grad_norm": 8.035116195678711, "learning_rate": 9.552467656923815e-05, "loss": 1.0752, "step": 4020 }, { "epoch": 0.2724439325157531, "grad_norm": 7.088611125946045, "learning_rate": 9.552330755014033e-05, "loss": 0.833, "step": 4021 }, { "epoch": 0.2725116877837252, "grad_norm": 7.615128040313721, "learning_rate": 9.552193853104251e-05, "loss": 1.0236, "step": 4022 }, { "epoch": 0.2725794430516973, "grad_norm": 7.366427421569824, "learning_rate": 9.552056951194469e-05, "loss": 0.892, "step": 4023 }, { "epoch": 0.2726471983196694, "grad_norm": 7.073375225067139, "learning_rate": 9.551920049284688e-05, "loss": 0.8082, "step": 4024 }, { "epoch": 0.2727149535876414, "grad_norm": 6.350280284881592, "learning_rate": 9.551783147374906e-05, "loss": 0.8154, "step": 4025 }, { "epoch": 0.2727827088556135, "grad_norm": 5.656667709350586, "learning_rate": 9.551646245465125e-05, "loss": 0.8558, "step": 4026 }, { "epoch": 0.2728504641235856, "grad_norm": 6.565401077270508, "learning_rate": 9.551509343555343e-05, "loss": 1.0397, "step": 4027 }, { "epoch": 0.2729182193915577, "grad_norm": 8.4253511428833, "learning_rate": 9.551372441645562e-05, "loss": 0.9872, "step": 4028 }, { "epoch": 0.2729859746595298, "grad_norm": 7.32992696762085, "learning_rate": 9.55123553973578e-05, "loss": 0.8262, "step": 4029 }, { "epoch": 0.2730537299275019, "grad_norm": 7.277110576629639, "learning_rate": 9.551098637825998e-05, "loss": 1.218, "step": 4030 }, { "epoch": 0.27312148519547397, "grad_norm": 9.594376564025879, "learning_rate": 9.550961735916216e-05, "loss": 0.9785, "step": 4031 }, { "epoch": 0.27318924046344606, "grad_norm": 9.339418411254883, "learning_rate": 9.550824834006434e-05, "loss": 0.911, "step": 4032 }, { "epoch": 0.2732569957314181, "grad_norm": 6.717375755310059, "learning_rate": 9.550687932096653e-05, "loss": 0.8084, "step": 4033 }, { "epoch": 0.2733247509993902, "grad_norm": 6.447595596313477, "learning_rate": 9.550551030186872e-05, "loss": 0.9882, "step": 4034 }, { "epoch": 0.2733925062673623, "grad_norm": 7.6800312995910645, "learning_rate": 9.55041412827709e-05, "loss": 0.7895, "step": 4035 }, { "epoch": 0.27346026153533437, "grad_norm": 7.270735263824463, "learning_rate": 9.550277226367308e-05, "loss": 0.83, "step": 4036 }, { "epoch": 0.27352801680330646, "grad_norm": 8.246411323547363, "learning_rate": 9.550140324457527e-05, "loss": 0.8406, "step": 4037 }, { "epoch": 0.27359577207127855, "grad_norm": 9.59301471710205, "learning_rate": 9.550003422547745e-05, "loss": 0.9433, "step": 4038 }, { "epoch": 0.27366352733925065, "grad_norm": 7.304765701293945, "learning_rate": 9.549866520637963e-05, "loss": 0.8551, "step": 4039 }, { "epoch": 0.27373128260722274, "grad_norm": 10.581608772277832, "learning_rate": 9.549729618728181e-05, "loss": 1.0734, "step": 4040 }, { "epoch": 0.2737990378751948, "grad_norm": 8.858924865722656, "learning_rate": 9.549592716818399e-05, "loss": 1.1424, "step": 4041 }, { "epoch": 0.27386679314316686, "grad_norm": 7.042451858520508, "learning_rate": 9.549455814908618e-05, "loss": 0.9556, "step": 4042 }, { "epoch": 0.27393454841113896, "grad_norm": 7.213229656219482, "learning_rate": 9.549318912998837e-05, "loss": 1.0575, "step": 4043 }, { "epoch": 0.27400230367911105, "grad_norm": 10.457990646362305, "learning_rate": 9.549182011089055e-05, "loss": 1.1868, "step": 4044 }, { "epoch": 0.27407005894708314, "grad_norm": 7.598734378814697, "learning_rate": 9.549045109179273e-05, "loss": 0.7216, "step": 4045 }, { "epoch": 0.27413781421505523, "grad_norm": 7.892279148101807, "learning_rate": 9.548908207269492e-05, "loss": 1.1437, "step": 4046 }, { "epoch": 0.2742055694830273, "grad_norm": 7.455031394958496, "learning_rate": 9.54877130535971e-05, "loss": 0.8402, "step": 4047 }, { "epoch": 0.2742733247509994, "grad_norm": 7.1315107345581055, "learning_rate": 9.548634403449928e-05, "loss": 0.653, "step": 4048 }, { "epoch": 0.27434108001897145, "grad_norm": 7.116184234619141, "learning_rate": 9.548497501540147e-05, "loss": 0.9, "step": 4049 }, { "epoch": 0.27440883528694354, "grad_norm": 7.833000659942627, "learning_rate": 9.548360599630365e-05, "loss": 0.9328, "step": 4050 }, { "epoch": 0.27447659055491563, "grad_norm": 7.393906116485596, "learning_rate": 9.548223697720584e-05, "loss": 1.0643, "step": 4051 }, { "epoch": 0.2745443458228877, "grad_norm": 8.286185264587402, "learning_rate": 9.548086795810803e-05, "loss": 1.0402, "step": 4052 }, { "epoch": 0.2746121010908598, "grad_norm": 7.88281774520874, "learning_rate": 9.547949893901021e-05, "loss": 1.0117, "step": 4053 }, { "epoch": 0.2746798563588319, "grad_norm": 9.544231414794922, "learning_rate": 9.547812991991239e-05, "loss": 1.1522, "step": 4054 }, { "epoch": 0.274747611626804, "grad_norm": 7.637237071990967, "learning_rate": 9.547676090081457e-05, "loss": 1.0805, "step": 4055 }, { "epoch": 0.2748153668947761, "grad_norm": 7.00446891784668, "learning_rate": 9.547539188171676e-05, "loss": 0.9941, "step": 4056 }, { "epoch": 0.27488312216274813, "grad_norm": 6.8821306228637695, "learning_rate": 9.547402286261894e-05, "loss": 0.997, "step": 4057 }, { "epoch": 0.2749508774307202, "grad_norm": 6.466810703277588, "learning_rate": 9.547265384352112e-05, "loss": 0.8872, "step": 4058 }, { "epoch": 0.2750186326986923, "grad_norm": 7.136430263519287, "learning_rate": 9.54712848244233e-05, "loss": 0.768, "step": 4059 }, { "epoch": 0.2750863879666644, "grad_norm": 7.117071151733398, "learning_rate": 9.54699158053255e-05, "loss": 0.965, "step": 4060 }, { "epoch": 0.2751541432346365, "grad_norm": 6.811083793640137, "learning_rate": 9.546854678622768e-05, "loss": 0.8448, "step": 4061 }, { "epoch": 0.2752218985026086, "grad_norm": 6.055437088012695, "learning_rate": 9.546717776712986e-05, "loss": 1.0217, "step": 4062 }, { "epoch": 0.2752896537705807, "grad_norm": 8.154548645019531, "learning_rate": 9.546580874803204e-05, "loss": 1.0642, "step": 4063 }, { "epoch": 0.27535740903855277, "grad_norm": 7.394543170928955, "learning_rate": 9.546443972893422e-05, "loss": 1.0716, "step": 4064 }, { "epoch": 0.2754251643065248, "grad_norm": 8.716939926147461, "learning_rate": 9.546307070983641e-05, "loss": 0.8457, "step": 4065 }, { "epoch": 0.2754929195744969, "grad_norm": 8.732163429260254, "learning_rate": 9.54617016907386e-05, "loss": 1.0349, "step": 4066 }, { "epoch": 0.275560674842469, "grad_norm": 8.67320442199707, "learning_rate": 9.546033267164077e-05, "loss": 1.1179, "step": 4067 }, { "epoch": 0.2756284301104411, "grad_norm": 8.010993003845215, "learning_rate": 9.545896365254296e-05, "loss": 1.0168, "step": 4068 }, { "epoch": 0.27569618537841317, "grad_norm": 6.747826099395752, "learning_rate": 9.545759463344515e-05, "loss": 0.7905, "step": 4069 }, { "epoch": 0.27576394064638526, "grad_norm": 8.352065086364746, "learning_rate": 9.545622561434733e-05, "loss": 1.1287, "step": 4070 }, { "epoch": 0.27583169591435736, "grad_norm": 8.072574615478516, "learning_rate": 9.545485659524951e-05, "loss": 1.1582, "step": 4071 }, { "epoch": 0.27589945118232945, "grad_norm": 8.851838111877441, "learning_rate": 9.545348757615169e-05, "loss": 1.0434, "step": 4072 }, { "epoch": 0.2759672064503015, "grad_norm": 9.833956718444824, "learning_rate": 9.545211855705387e-05, "loss": 1.1956, "step": 4073 }, { "epoch": 0.2760349617182736, "grad_norm": 8.043981552124023, "learning_rate": 9.545074953795606e-05, "loss": 1.0625, "step": 4074 }, { "epoch": 0.27610271698624567, "grad_norm": 7.420129776000977, "learning_rate": 9.544938051885824e-05, "loss": 0.8933, "step": 4075 }, { "epoch": 0.27617047225421776, "grad_norm": 7.062417030334473, "learning_rate": 9.544801149976042e-05, "loss": 0.9655, "step": 4076 }, { "epoch": 0.27623822752218985, "grad_norm": 8.733392715454102, "learning_rate": 9.54466424806626e-05, "loss": 0.9182, "step": 4077 }, { "epoch": 0.27630598279016194, "grad_norm": 8.931736946105957, "learning_rate": 9.544527346156479e-05, "loss": 1.0243, "step": 4078 }, { "epoch": 0.27637373805813403, "grad_norm": 7.487978935241699, "learning_rate": 9.544390444246698e-05, "loss": 0.9967, "step": 4079 }, { "epoch": 0.2764414933261061, "grad_norm": 8.259819030761719, "learning_rate": 9.544253542336916e-05, "loss": 0.942, "step": 4080 }, { "epoch": 0.2765092485940782, "grad_norm": 9.625347137451172, "learning_rate": 9.544116640427134e-05, "loss": 1.2148, "step": 4081 }, { "epoch": 0.27657700386205025, "grad_norm": 7.737034797668457, "learning_rate": 9.543979738517352e-05, "loss": 0.7905, "step": 4082 }, { "epoch": 0.27664475913002234, "grad_norm": 7.118561744689941, "learning_rate": 9.543842836607571e-05, "loss": 0.9883, "step": 4083 }, { "epoch": 0.27671251439799444, "grad_norm": 9.749618530273438, "learning_rate": 9.54370593469779e-05, "loss": 1.0249, "step": 4084 }, { "epoch": 0.2767802696659665, "grad_norm": 8.805608749389648, "learning_rate": 9.543569032788008e-05, "loss": 1.041, "step": 4085 }, { "epoch": 0.2768480249339386, "grad_norm": 9.240931510925293, "learning_rate": 9.543432130878226e-05, "loss": 1.0598, "step": 4086 }, { "epoch": 0.2769157802019107, "grad_norm": 6.621399879455566, "learning_rate": 9.543295228968444e-05, "loss": 0.8884, "step": 4087 }, { "epoch": 0.2769835354698828, "grad_norm": 6.869698524475098, "learning_rate": 9.543158327058663e-05, "loss": 0.8309, "step": 4088 }, { "epoch": 0.2770512907378549, "grad_norm": 7.135868549346924, "learning_rate": 9.543021425148881e-05, "loss": 0.9399, "step": 4089 }, { "epoch": 0.27711904600582693, "grad_norm": 7.172493934631348, "learning_rate": 9.542884523239099e-05, "loss": 0.9626, "step": 4090 }, { "epoch": 0.277186801273799, "grad_norm": 6.903214931488037, "learning_rate": 9.542747621329317e-05, "loss": 1.0047, "step": 4091 }, { "epoch": 0.2772545565417711, "grad_norm": 7.557178020477295, "learning_rate": 9.542610719419536e-05, "loss": 0.6978, "step": 4092 }, { "epoch": 0.2773223118097432, "grad_norm": 7.468019485473633, "learning_rate": 9.542473817509754e-05, "loss": 0.8109, "step": 4093 }, { "epoch": 0.2773900670777153, "grad_norm": 8.699142456054688, "learning_rate": 9.542336915599973e-05, "loss": 0.7031, "step": 4094 }, { "epoch": 0.2774578223456874, "grad_norm": 8.03862190246582, "learning_rate": 9.54220001369019e-05, "loss": 0.9824, "step": 4095 }, { "epoch": 0.2775255776136595, "grad_norm": 9.884957313537598, "learning_rate": 9.54206311178041e-05, "loss": 0.8122, "step": 4096 }, { "epoch": 0.27759333288163157, "grad_norm": 9.435370445251465, "learning_rate": 9.541926209870628e-05, "loss": 1.2642, "step": 4097 }, { "epoch": 0.2776610881496036, "grad_norm": 8.154888153076172, "learning_rate": 9.541789307960846e-05, "loss": 0.9814, "step": 4098 }, { "epoch": 0.2777288434175757, "grad_norm": 9.771589279174805, "learning_rate": 9.541652406051065e-05, "loss": 1.0959, "step": 4099 }, { "epoch": 0.2777965986855478, "grad_norm": 7.635507106781006, "learning_rate": 9.541515504141283e-05, "loss": 0.8268, "step": 4100 }, { "epoch": 0.2778643539535199, "grad_norm": 9.028327941894531, "learning_rate": 9.541378602231501e-05, "loss": 1.227, "step": 4101 }, { "epoch": 0.277932109221492, "grad_norm": 8.327515602111816, "learning_rate": 9.541241700321721e-05, "loss": 0.8644, "step": 4102 }, { "epoch": 0.27799986448946407, "grad_norm": 7.547940254211426, "learning_rate": 9.541104798411939e-05, "loss": 1.0169, "step": 4103 }, { "epoch": 0.27806761975743616, "grad_norm": 8.0435152053833, "learning_rate": 9.540967896502157e-05, "loss": 1.0883, "step": 4104 }, { "epoch": 0.27813537502540825, "grad_norm": 7.7741217613220215, "learning_rate": 9.540830994592375e-05, "loss": 0.9389, "step": 4105 }, { "epoch": 0.2782031302933803, "grad_norm": 8.059552192687988, "learning_rate": 9.540694092682594e-05, "loss": 1.0024, "step": 4106 }, { "epoch": 0.2782708855613524, "grad_norm": 9.13268756866455, "learning_rate": 9.540557190772812e-05, "loss": 1.0448, "step": 4107 }, { "epoch": 0.27833864082932447, "grad_norm": 7.901900768280029, "learning_rate": 9.54042028886303e-05, "loss": 0.8235, "step": 4108 }, { "epoch": 0.27840639609729656, "grad_norm": 8.727076530456543, "learning_rate": 9.540283386953248e-05, "loss": 1.1047, "step": 4109 }, { "epoch": 0.27847415136526865, "grad_norm": 7.1972880363464355, "learning_rate": 9.540146485043466e-05, "loss": 0.9456, "step": 4110 }, { "epoch": 0.27854190663324074, "grad_norm": 6.886523246765137, "learning_rate": 9.540009583133686e-05, "loss": 1.0041, "step": 4111 }, { "epoch": 0.27860966190121284, "grad_norm": 7.595452308654785, "learning_rate": 9.539872681223904e-05, "loss": 1.0038, "step": 4112 }, { "epoch": 0.2786774171691849, "grad_norm": 6.007086753845215, "learning_rate": 9.539735779314122e-05, "loss": 0.8042, "step": 4113 }, { "epoch": 0.27874517243715696, "grad_norm": 7.112758159637451, "learning_rate": 9.53959887740434e-05, "loss": 0.8252, "step": 4114 }, { "epoch": 0.27881292770512905, "grad_norm": 10.120092391967773, "learning_rate": 9.53946197549456e-05, "loss": 1.1589, "step": 4115 }, { "epoch": 0.27888068297310115, "grad_norm": 7.587961196899414, "learning_rate": 9.539325073584777e-05, "loss": 1.0699, "step": 4116 }, { "epoch": 0.27894843824107324, "grad_norm": 7.671876430511475, "learning_rate": 9.539188171674995e-05, "loss": 0.9699, "step": 4117 }, { "epoch": 0.27901619350904533, "grad_norm": 7.345922470092773, "learning_rate": 9.539051269765213e-05, "loss": 0.9957, "step": 4118 }, { "epoch": 0.2790839487770174, "grad_norm": 9.215903282165527, "learning_rate": 9.538914367855432e-05, "loss": 1.156, "step": 4119 }, { "epoch": 0.2791517040449895, "grad_norm": 10.056458473205566, "learning_rate": 9.538777465945651e-05, "loss": 1.1229, "step": 4120 }, { "epoch": 0.2792194593129616, "grad_norm": 7.9655938148498535, "learning_rate": 9.538640564035869e-05, "loss": 1.017, "step": 4121 }, { "epoch": 0.27928721458093364, "grad_norm": 8.49431324005127, "learning_rate": 9.538503662126087e-05, "loss": 1.0035, "step": 4122 }, { "epoch": 0.27935496984890573, "grad_norm": 8.746543884277344, "learning_rate": 9.538366760216305e-05, "loss": 1.0711, "step": 4123 }, { "epoch": 0.2794227251168778, "grad_norm": 7.75557279586792, "learning_rate": 9.538229858306524e-05, "loss": 0.6868, "step": 4124 }, { "epoch": 0.2794904803848499, "grad_norm": 7.1494622230529785, "learning_rate": 9.538092956396742e-05, "loss": 0.721, "step": 4125 }, { "epoch": 0.279558235652822, "grad_norm": 6.971895217895508, "learning_rate": 9.53795605448696e-05, "loss": 1.0669, "step": 4126 }, { "epoch": 0.2796259909207941, "grad_norm": 6.478157043457031, "learning_rate": 9.537819152577178e-05, "loss": 0.9348, "step": 4127 }, { "epoch": 0.2796937461887662, "grad_norm": 6.307050704956055, "learning_rate": 9.537682250667397e-05, "loss": 0.9622, "step": 4128 }, { "epoch": 0.2797615014567383, "grad_norm": 9.505130767822266, "learning_rate": 9.537545348757616e-05, "loss": 1.1675, "step": 4129 }, { "epoch": 0.2798292567247104, "grad_norm": 8.11099624633789, "learning_rate": 9.537408446847834e-05, "loss": 0.853, "step": 4130 }, { "epoch": 0.2798970119926824, "grad_norm": 6.637272834777832, "learning_rate": 9.537271544938052e-05, "loss": 1.0993, "step": 4131 }, { "epoch": 0.2799647672606545, "grad_norm": 7.888055801391602, "learning_rate": 9.53713464302827e-05, "loss": 0.9582, "step": 4132 }, { "epoch": 0.2800325225286266, "grad_norm": 6.289199352264404, "learning_rate": 9.536997741118488e-05, "loss": 0.9027, "step": 4133 }, { "epoch": 0.2801002777965987, "grad_norm": 7.488378047943115, "learning_rate": 9.536860839208707e-05, "loss": 1.0697, "step": 4134 }, { "epoch": 0.2801680330645708, "grad_norm": 8.341411590576172, "learning_rate": 9.536723937298925e-05, "loss": 1.0705, "step": 4135 }, { "epoch": 0.28023578833254287, "grad_norm": 7.55519437789917, "learning_rate": 9.536587035389144e-05, "loss": 0.7589, "step": 4136 }, { "epoch": 0.28030354360051496, "grad_norm": 6.104217052459717, "learning_rate": 9.536450133479362e-05, "loss": 0.9161, "step": 4137 }, { "epoch": 0.28037129886848705, "grad_norm": 6.97914457321167, "learning_rate": 9.536313231569581e-05, "loss": 1.0342, "step": 4138 }, { "epoch": 0.2804390541364591, "grad_norm": 8.791030883789062, "learning_rate": 9.536176329659799e-05, "loss": 0.863, "step": 4139 }, { "epoch": 0.2805068094044312, "grad_norm": 6.868939399719238, "learning_rate": 9.536039427750017e-05, "loss": 0.9005, "step": 4140 }, { "epoch": 0.28057456467240327, "grad_norm": 9.854182243347168, "learning_rate": 9.535902525840235e-05, "loss": 0.773, "step": 4141 }, { "epoch": 0.28064231994037536, "grad_norm": 7.64580774307251, "learning_rate": 9.535765623930454e-05, "loss": 0.9999, "step": 4142 }, { "epoch": 0.28071007520834745, "grad_norm": 6.032886028289795, "learning_rate": 9.535628722020672e-05, "loss": 0.8916, "step": 4143 }, { "epoch": 0.28077783047631955, "grad_norm": 7.0441060066223145, "learning_rate": 9.53549182011089e-05, "loss": 0.9787, "step": 4144 }, { "epoch": 0.28084558574429164, "grad_norm": 6.4428629875183105, "learning_rate": 9.53535491820111e-05, "loss": 0.6749, "step": 4145 }, { "epoch": 0.28091334101226373, "grad_norm": 8.476522445678711, "learning_rate": 9.535218016291328e-05, "loss": 0.9408, "step": 4146 }, { "epoch": 0.28098109628023576, "grad_norm": 8.720208168029785, "learning_rate": 9.535081114381546e-05, "loss": 0.8571, "step": 4147 }, { "epoch": 0.28104885154820786, "grad_norm": 7.846646785736084, "learning_rate": 9.534944212471765e-05, "loss": 0.9193, "step": 4148 }, { "epoch": 0.28111660681617995, "grad_norm": 7.848026275634766, "learning_rate": 9.534807310561983e-05, "loss": 1.1975, "step": 4149 }, { "epoch": 0.28118436208415204, "grad_norm": 9.543595314025879, "learning_rate": 9.534670408652201e-05, "loss": 0.8769, "step": 4150 }, { "epoch": 0.28125211735212413, "grad_norm": 7.252998352050781, "learning_rate": 9.53453350674242e-05, "loss": 1.029, "step": 4151 }, { "epoch": 0.2813198726200962, "grad_norm": 6.271702289581299, "learning_rate": 9.534396604832639e-05, "loss": 0.8992, "step": 4152 }, { "epoch": 0.2813876278880683, "grad_norm": 9.071548461914062, "learning_rate": 9.534259702922857e-05, "loss": 1.0518, "step": 4153 }, { "epoch": 0.2814553831560404, "grad_norm": 7.452267169952393, "learning_rate": 9.534122801013075e-05, "loss": 0.9573, "step": 4154 }, { "epoch": 0.28152313842401244, "grad_norm": 8.72459602355957, "learning_rate": 9.533985899103293e-05, "loss": 1.09, "step": 4155 }, { "epoch": 0.28159089369198453, "grad_norm": 8.032079696655273, "learning_rate": 9.533848997193511e-05, "loss": 0.9595, "step": 4156 }, { "epoch": 0.2816586489599566, "grad_norm": 7.622939109802246, "learning_rate": 9.53371209528373e-05, "loss": 0.8722, "step": 4157 }, { "epoch": 0.2817264042279287, "grad_norm": 8.330899238586426, "learning_rate": 9.533575193373948e-05, "loss": 0.985, "step": 4158 }, { "epoch": 0.2817941594959008, "grad_norm": 9.378608703613281, "learning_rate": 9.533438291464166e-05, "loss": 0.8458, "step": 4159 }, { "epoch": 0.2818619147638729, "grad_norm": 7.305957317352295, "learning_rate": 9.533301389554384e-05, "loss": 0.9653, "step": 4160 }, { "epoch": 0.281929670031845, "grad_norm": 8.557588577270508, "learning_rate": 9.533164487644604e-05, "loss": 0.8742, "step": 4161 }, { "epoch": 0.2819974252998171, "grad_norm": 8.77814769744873, "learning_rate": 9.533027585734822e-05, "loss": 1.1782, "step": 4162 }, { "epoch": 0.2820651805677891, "grad_norm": 7.212672233581543, "learning_rate": 9.53289068382504e-05, "loss": 0.6333, "step": 4163 }, { "epoch": 0.2821329358357612, "grad_norm": 10.004176139831543, "learning_rate": 9.532753781915258e-05, "loss": 0.9544, "step": 4164 }, { "epoch": 0.2822006911037333, "grad_norm": 7.577065467834473, "learning_rate": 9.532616880005476e-05, "loss": 1.0708, "step": 4165 }, { "epoch": 0.2822684463717054, "grad_norm": 7.181521415710449, "learning_rate": 9.532479978095695e-05, "loss": 0.829, "step": 4166 }, { "epoch": 0.2823362016396775, "grad_norm": 5.7575249671936035, "learning_rate": 9.532343076185913e-05, "loss": 0.9353, "step": 4167 }, { "epoch": 0.2824039569076496, "grad_norm": 7.974564075469971, "learning_rate": 9.532206174276131e-05, "loss": 0.9976, "step": 4168 }, { "epoch": 0.28247171217562167, "grad_norm": 8.870126724243164, "learning_rate": 9.53206927236635e-05, "loss": 0.9704, "step": 4169 }, { "epoch": 0.28253946744359376, "grad_norm": 6.596248149871826, "learning_rate": 9.531932370456569e-05, "loss": 0.8763, "step": 4170 }, { "epoch": 0.2826072227115658, "grad_norm": 7.725964546203613, "learning_rate": 9.531795468546787e-05, "loss": 1.4009, "step": 4171 }, { "epoch": 0.2826749779795379, "grad_norm": 6.741204261779785, "learning_rate": 9.531658566637005e-05, "loss": 0.9103, "step": 4172 }, { "epoch": 0.28274273324751, "grad_norm": 9.002605438232422, "learning_rate": 9.531521664727223e-05, "loss": 1.1446, "step": 4173 }, { "epoch": 0.2828104885154821, "grad_norm": 7.860680103302002, "learning_rate": 9.531384762817441e-05, "loss": 1.081, "step": 4174 }, { "epoch": 0.28287824378345416, "grad_norm": 7.939533233642578, "learning_rate": 9.53124786090766e-05, "loss": 0.8876, "step": 4175 }, { "epoch": 0.28294599905142626, "grad_norm": 7.104232311248779, "learning_rate": 9.531110958997878e-05, "loss": 0.8687, "step": 4176 }, { "epoch": 0.28301375431939835, "grad_norm": 9.095148086547852, "learning_rate": 9.530974057088096e-05, "loss": 0.873, "step": 4177 }, { "epoch": 0.28308150958737044, "grad_norm": 6.869518280029297, "learning_rate": 9.530837155178314e-05, "loss": 0.8372, "step": 4178 }, { "epoch": 0.2831492648553425, "grad_norm": 7.7245049476623535, "learning_rate": 9.530700253268534e-05, "loss": 1.0267, "step": 4179 }, { "epoch": 0.28321702012331457, "grad_norm": 8.298705101013184, "learning_rate": 9.530563351358752e-05, "loss": 0.8919, "step": 4180 }, { "epoch": 0.28328477539128666, "grad_norm": 6.176532745361328, "learning_rate": 9.53042644944897e-05, "loss": 0.8603, "step": 4181 }, { "epoch": 0.28335253065925875, "grad_norm": 6.592353820800781, "learning_rate": 9.530289547539188e-05, "loss": 0.9415, "step": 4182 }, { "epoch": 0.28342028592723084, "grad_norm": 7.962296962738037, "learning_rate": 9.530152645629406e-05, "loss": 0.8764, "step": 4183 }, { "epoch": 0.28348804119520293, "grad_norm": 6.484033584594727, "learning_rate": 9.530015743719625e-05, "loss": 0.9577, "step": 4184 }, { "epoch": 0.283555796463175, "grad_norm": 8.156810760498047, "learning_rate": 9.529878841809843e-05, "loss": 1.1628, "step": 4185 }, { "epoch": 0.2836235517311471, "grad_norm": 6.916367053985596, "learning_rate": 9.529741939900061e-05, "loss": 1.1518, "step": 4186 }, { "epoch": 0.2836913069991192, "grad_norm": 7.70388650894165, "learning_rate": 9.52960503799028e-05, "loss": 0.9309, "step": 4187 }, { "epoch": 0.28375906226709124, "grad_norm": 8.770346641540527, "learning_rate": 9.529468136080499e-05, "loss": 0.891, "step": 4188 }, { "epoch": 0.28382681753506334, "grad_norm": 7.100319862365723, "learning_rate": 9.529331234170717e-05, "loss": 1.0302, "step": 4189 }, { "epoch": 0.2838945728030354, "grad_norm": 7.376253128051758, "learning_rate": 9.529194332260935e-05, "loss": 0.7906, "step": 4190 }, { "epoch": 0.2839623280710075, "grad_norm": 10.125496864318848, "learning_rate": 9.529057430351154e-05, "loss": 0.9829, "step": 4191 }, { "epoch": 0.2840300833389796, "grad_norm": 7.877635955810547, "learning_rate": 9.528920528441372e-05, "loss": 0.8803, "step": 4192 }, { "epoch": 0.2840978386069517, "grad_norm": 8.096887588500977, "learning_rate": 9.52878362653159e-05, "loss": 0.9342, "step": 4193 }, { "epoch": 0.2841655938749238, "grad_norm": 7.934850215911865, "learning_rate": 9.52864672462181e-05, "loss": 1.1699, "step": 4194 }, { "epoch": 0.2842333491428959, "grad_norm": 8.237794876098633, "learning_rate": 9.528509822712028e-05, "loss": 1.1718, "step": 4195 }, { "epoch": 0.2843011044108679, "grad_norm": 7.528624057769775, "learning_rate": 9.528372920802246e-05, "loss": 0.8229, "step": 4196 }, { "epoch": 0.28436885967884, "grad_norm": 7.210242748260498, "learning_rate": 9.528236018892464e-05, "loss": 0.8875, "step": 4197 }, { "epoch": 0.2844366149468121, "grad_norm": 7.630309581756592, "learning_rate": 9.528099116982683e-05, "loss": 0.8811, "step": 4198 }, { "epoch": 0.2845043702147842, "grad_norm": 11.624275207519531, "learning_rate": 9.527962215072901e-05, "loss": 1.0702, "step": 4199 }, { "epoch": 0.2845721254827563, "grad_norm": 7.52834415435791, "learning_rate": 9.52782531316312e-05, "loss": 0.9066, "step": 4200 }, { "epoch": 0.2846398807507284, "grad_norm": 8.418506622314453, "learning_rate": 9.527688411253337e-05, "loss": 0.8147, "step": 4201 }, { "epoch": 0.28470763601870047, "grad_norm": 9.67719554901123, "learning_rate": 9.527551509343557e-05, "loss": 0.976, "step": 4202 }, { "epoch": 0.28477539128667256, "grad_norm": 7.107409954071045, "learning_rate": 9.527414607433775e-05, "loss": 0.695, "step": 4203 }, { "epoch": 0.2848431465546446, "grad_norm": 7.840113162994385, "learning_rate": 9.527277705523993e-05, "loss": 0.9629, "step": 4204 }, { "epoch": 0.2849109018226167, "grad_norm": 8.170151710510254, "learning_rate": 9.527140803614211e-05, "loss": 0.8715, "step": 4205 }, { "epoch": 0.2849786570905888, "grad_norm": 7.882331848144531, "learning_rate": 9.527003901704429e-05, "loss": 1.0364, "step": 4206 }, { "epoch": 0.2850464123585609, "grad_norm": 10.077646255493164, "learning_rate": 9.526866999794648e-05, "loss": 1.2749, "step": 4207 }, { "epoch": 0.28511416762653297, "grad_norm": 8.749690055847168, "learning_rate": 9.526730097884866e-05, "loss": 0.9815, "step": 4208 }, { "epoch": 0.28518192289450506, "grad_norm": 6.793465614318848, "learning_rate": 9.526593195975084e-05, "loss": 1.0165, "step": 4209 }, { "epoch": 0.28524967816247715, "grad_norm": 7.186471939086914, "learning_rate": 9.526456294065302e-05, "loss": 0.911, "step": 4210 }, { "epoch": 0.28531743343044924, "grad_norm": 8.116944313049316, "learning_rate": 9.52631939215552e-05, "loss": 1.051, "step": 4211 }, { "epoch": 0.2853851886984213, "grad_norm": 5.9162750244140625, "learning_rate": 9.52618249024574e-05, "loss": 0.8165, "step": 4212 }, { "epoch": 0.28545294396639337, "grad_norm": 7.20265531539917, "learning_rate": 9.526045588335958e-05, "loss": 1.0693, "step": 4213 }, { "epoch": 0.28552069923436546, "grad_norm": 7.380153179168701, "learning_rate": 9.525908686426176e-05, "loss": 0.9611, "step": 4214 }, { "epoch": 0.28558845450233755, "grad_norm": 7.211367607116699, "learning_rate": 9.525771784516394e-05, "loss": 0.7595, "step": 4215 }, { "epoch": 0.28565620977030964, "grad_norm": 6.8030104637146, "learning_rate": 9.525634882606613e-05, "loss": 0.8281, "step": 4216 }, { "epoch": 0.28572396503828174, "grad_norm": 7.774519443511963, "learning_rate": 9.525497980696831e-05, "loss": 0.963, "step": 4217 }, { "epoch": 0.2857917203062538, "grad_norm": 8.210673332214355, "learning_rate": 9.52536107878705e-05, "loss": 1.3076, "step": 4218 }, { "epoch": 0.2858594755742259, "grad_norm": 7.3878583908081055, "learning_rate": 9.525224176877267e-05, "loss": 0.9546, "step": 4219 }, { "epoch": 0.28592723084219795, "grad_norm": 10.107827186584473, "learning_rate": 9.525087274967485e-05, "loss": 1.3039, "step": 4220 }, { "epoch": 0.28599498611017005, "grad_norm": 7.974700450897217, "learning_rate": 9.524950373057705e-05, "loss": 0.9651, "step": 4221 }, { "epoch": 0.28606274137814214, "grad_norm": 8.606413841247559, "learning_rate": 9.524813471147923e-05, "loss": 0.9459, "step": 4222 }, { "epoch": 0.28613049664611423, "grad_norm": 6.7952799797058105, "learning_rate": 9.524676569238141e-05, "loss": 0.8818, "step": 4223 }, { "epoch": 0.2861982519140863, "grad_norm": 7.2026214599609375, "learning_rate": 9.524539667328359e-05, "loss": 0.8578, "step": 4224 }, { "epoch": 0.2862660071820584, "grad_norm": 9.135619163513184, "learning_rate": 9.524402765418578e-05, "loss": 1.1594, "step": 4225 }, { "epoch": 0.2863337624500305, "grad_norm": 7.704957962036133, "learning_rate": 9.524265863508796e-05, "loss": 0.9715, "step": 4226 }, { "epoch": 0.2864015177180026, "grad_norm": 6.570467948913574, "learning_rate": 9.524128961599014e-05, "loss": 0.8916, "step": 4227 }, { "epoch": 0.28646927298597463, "grad_norm": 8.359966278076172, "learning_rate": 9.523992059689232e-05, "loss": 1.0021, "step": 4228 }, { "epoch": 0.2865370282539467, "grad_norm": 7.011820316314697, "learning_rate": 9.52385515777945e-05, "loss": 0.8718, "step": 4229 }, { "epoch": 0.2866047835219188, "grad_norm": 6.853650093078613, "learning_rate": 9.52371825586967e-05, "loss": 0.8041, "step": 4230 }, { "epoch": 0.2866725387898909, "grad_norm": 6.95853328704834, "learning_rate": 9.523581353959888e-05, "loss": 0.8763, "step": 4231 }, { "epoch": 0.286740294057863, "grad_norm": 9.217144012451172, "learning_rate": 9.523444452050106e-05, "loss": 1.046, "step": 4232 }, { "epoch": 0.2868080493258351, "grad_norm": 8.338934898376465, "learning_rate": 9.523307550140324e-05, "loss": 0.9205, "step": 4233 }, { "epoch": 0.2868758045938072, "grad_norm": 9.220541954040527, "learning_rate": 9.523170648230543e-05, "loss": 1.1468, "step": 4234 }, { "epoch": 0.2869435598617793, "grad_norm": 7.641387462615967, "learning_rate": 9.523033746320761e-05, "loss": 1.118, "step": 4235 }, { "epoch": 0.28701131512975137, "grad_norm": 7.52994441986084, "learning_rate": 9.52289684441098e-05, "loss": 0.851, "step": 4236 }, { "epoch": 0.2870790703977234, "grad_norm": 8.712708473205566, "learning_rate": 9.522759942501199e-05, "loss": 1.1741, "step": 4237 }, { "epoch": 0.2871468256656955, "grad_norm": 7.8549723625183105, "learning_rate": 9.522623040591417e-05, "loss": 0.9744, "step": 4238 }, { "epoch": 0.2872145809336676, "grad_norm": 6.8177642822265625, "learning_rate": 9.522486138681635e-05, "loss": 1.0432, "step": 4239 }, { "epoch": 0.2872823362016397, "grad_norm": 10.273691177368164, "learning_rate": 9.522349236771854e-05, "loss": 1.0777, "step": 4240 }, { "epoch": 0.28735009146961177, "grad_norm": 8.261405944824219, "learning_rate": 9.522212334862072e-05, "loss": 0.9415, "step": 4241 }, { "epoch": 0.28741784673758386, "grad_norm": 10.067824363708496, "learning_rate": 9.52207543295229e-05, "loss": 1.1091, "step": 4242 }, { "epoch": 0.28748560200555595, "grad_norm": 9.639914512634277, "learning_rate": 9.521938531042508e-05, "loss": 0.9534, "step": 4243 }, { "epoch": 0.28755335727352804, "grad_norm": 6.993269920349121, "learning_rate": 9.521801629132728e-05, "loss": 1.1193, "step": 4244 }, { "epoch": 0.2876211125415001, "grad_norm": 7.779829502105713, "learning_rate": 9.521664727222946e-05, "loss": 0.8744, "step": 4245 }, { "epoch": 0.28768886780947217, "grad_norm": 7.070007801055908, "learning_rate": 9.521527825313164e-05, "loss": 0.8609, "step": 4246 }, { "epoch": 0.28775662307744426, "grad_norm": 6.091519355773926, "learning_rate": 9.521390923403382e-05, "loss": 0.8018, "step": 4247 }, { "epoch": 0.28782437834541635, "grad_norm": 6.76396369934082, "learning_rate": 9.521254021493601e-05, "loss": 0.8509, "step": 4248 }, { "epoch": 0.28789213361338845, "grad_norm": 8.478080749511719, "learning_rate": 9.521117119583819e-05, "loss": 1.0355, "step": 4249 }, { "epoch": 0.28795988888136054, "grad_norm": 7.830933094024658, "learning_rate": 9.520980217674037e-05, "loss": 0.8699, "step": 4250 }, { "epoch": 0.28802764414933263, "grad_norm": 8.354218482971191, "learning_rate": 9.520843315764255e-05, "loss": 0.9228, "step": 4251 }, { "epoch": 0.2880953994173047, "grad_norm": 8.529300689697266, "learning_rate": 9.520706413854473e-05, "loss": 1.1215, "step": 4252 }, { "epoch": 0.28816315468527676, "grad_norm": 6.831529140472412, "learning_rate": 9.520569511944693e-05, "loss": 1.0665, "step": 4253 }, { "epoch": 0.28823090995324885, "grad_norm": 7.824577808380127, "learning_rate": 9.520432610034911e-05, "loss": 0.8864, "step": 4254 }, { "epoch": 0.28829866522122094, "grad_norm": 7.795472621917725, "learning_rate": 9.520295708125129e-05, "loss": 1.0775, "step": 4255 }, { "epoch": 0.28836642048919303, "grad_norm": 7.210735321044922, "learning_rate": 9.520158806215347e-05, "loss": 1.0244, "step": 4256 }, { "epoch": 0.2884341757571651, "grad_norm": 7.224759101867676, "learning_rate": 9.520021904305566e-05, "loss": 1.0527, "step": 4257 }, { "epoch": 0.2885019310251372, "grad_norm": 7.608676910400391, "learning_rate": 9.519885002395784e-05, "loss": 0.7978, "step": 4258 }, { "epoch": 0.2885696862931093, "grad_norm": 7.426436424255371, "learning_rate": 9.519748100486002e-05, "loss": 0.859, "step": 4259 }, { "epoch": 0.2886374415610814, "grad_norm": 7.585330963134766, "learning_rate": 9.51961119857622e-05, "loss": 1.1091, "step": 4260 }, { "epoch": 0.28870519682905343, "grad_norm": 6.930294990539551, "learning_rate": 9.519474296666438e-05, "loss": 0.8862, "step": 4261 }, { "epoch": 0.2887729520970255, "grad_norm": 8.092456817626953, "learning_rate": 9.519337394756658e-05, "loss": 1.051, "step": 4262 }, { "epoch": 0.2888407073649976, "grad_norm": 7.897385597229004, "learning_rate": 9.519200492846876e-05, "loss": 1.0589, "step": 4263 }, { "epoch": 0.2889084626329697, "grad_norm": 7.344932556152344, "learning_rate": 9.519063590937094e-05, "loss": 1.0319, "step": 4264 }, { "epoch": 0.2889762179009418, "grad_norm": 8.676694869995117, "learning_rate": 9.518926689027312e-05, "loss": 0.9119, "step": 4265 }, { "epoch": 0.2890439731689139, "grad_norm": 7.291116714477539, "learning_rate": 9.51878978711753e-05, "loss": 0.8249, "step": 4266 }, { "epoch": 0.289111728436886, "grad_norm": 7.467698574066162, "learning_rate": 9.51865288520775e-05, "loss": 0.9807, "step": 4267 }, { "epoch": 0.2891794837048581, "grad_norm": 7.744437217712402, "learning_rate": 9.518515983297967e-05, "loss": 0.8792, "step": 4268 }, { "epoch": 0.2892472389728301, "grad_norm": 6.8492560386657715, "learning_rate": 9.518379081388185e-05, "loss": 0.9216, "step": 4269 }, { "epoch": 0.2893149942408022, "grad_norm": 7.247424602508545, "learning_rate": 9.518242179478403e-05, "loss": 1.0865, "step": 4270 }, { "epoch": 0.2893827495087743, "grad_norm": 8.125252723693848, "learning_rate": 9.518105277568623e-05, "loss": 1.0298, "step": 4271 }, { "epoch": 0.2894505047767464, "grad_norm": 6.578275680541992, "learning_rate": 9.517968375658841e-05, "loss": 0.9353, "step": 4272 }, { "epoch": 0.2895182600447185, "grad_norm": 8.099616050720215, "learning_rate": 9.517831473749059e-05, "loss": 1.038, "step": 4273 }, { "epoch": 0.28958601531269057, "grad_norm": 7.529900074005127, "learning_rate": 9.517694571839277e-05, "loss": 0.9753, "step": 4274 }, { "epoch": 0.28965377058066266, "grad_norm": 7.521812915802002, "learning_rate": 9.517557669929495e-05, "loss": 0.972, "step": 4275 }, { "epoch": 0.28972152584863475, "grad_norm": 7.780272483825684, "learning_rate": 9.517420768019714e-05, "loss": 0.9927, "step": 4276 }, { "epoch": 0.2897892811166068, "grad_norm": 7.982085704803467, "learning_rate": 9.517283866109932e-05, "loss": 0.9462, "step": 4277 }, { "epoch": 0.2898570363845789, "grad_norm": 7.325984477996826, "learning_rate": 9.51714696420015e-05, "loss": 1.0814, "step": 4278 }, { "epoch": 0.289924791652551, "grad_norm": 7.833248138427734, "learning_rate": 9.517010062290368e-05, "loss": 1.1736, "step": 4279 }, { "epoch": 0.28999254692052306, "grad_norm": 6.983424186706543, "learning_rate": 9.516873160380588e-05, "loss": 0.9662, "step": 4280 }, { "epoch": 0.29006030218849516, "grad_norm": 7.954293251037598, "learning_rate": 9.516736258470806e-05, "loss": 0.8449, "step": 4281 }, { "epoch": 0.29012805745646725, "grad_norm": 7.47749137878418, "learning_rate": 9.516599356561024e-05, "loss": 0.7869, "step": 4282 }, { "epoch": 0.29019581272443934, "grad_norm": 8.70733642578125, "learning_rate": 9.516462454651243e-05, "loss": 0.915, "step": 4283 }, { "epoch": 0.29026356799241143, "grad_norm": 7.003527641296387, "learning_rate": 9.516325552741461e-05, "loss": 0.8543, "step": 4284 }, { "epoch": 0.2903313232603835, "grad_norm": 8.230785369873047, "learning_rate": 9.51618865083168e-05, "loss": 1.0022, "step": 4285 }, { "epoch": 0.29039907852835556, "grad_norm": 6.20754337310791, "learning_rate": 9.516051748921899e-05, "loss": 0.8967, "step": 4286 }, { "epoch": 0.29046683379632765, "grad_norm": 7.0158538818359375, "learning_rate": 9.515914847012117e-05, "loss": 0.9162, "step": 4287 }, { "epoch": 0.29053458906429974, "grad_norm": 6.215134143829346, "learning_rate": 9.515777945102335e-05, "loss": 0.7362, "step": 4288 }, { "epoch": 0.29060234433227183, "grad_norm": 7.366815090179443, "learning_rate": 9.515641043192553e-05, "loss": 1.0174, "step": 4289 }, { "epoch": 0.2906700996002439, "grad_norm": 7.514114856719971, "learning_rate": 9.515504141282772e-05, "loss": 0.9979, "step": 4290 }, { "epoch": 0.290737854868216, "grad_norm": 6.899611949920654, "learning_rate": 9.51536723937299e-05, "loss": 0.6909, "step": 4291 }, { "epoch": 0.2908056101361881, "grad_norm": 7.254974842071533, "learning_rate": 9.515230337463208e-05, "loss": 1.1744, "step": 4292 }, { "epoch": 0.2908733654041602, "grad_norm": 7.3820719718933105, "learning_rate": 9.515093435553426e-05, "loss": 1.0514, "step": 4293 }, { "epoch": 0.29094112067213224, "grad_norm": 7.556882858276367, "learning_rate": 9.514956533643646e-05, "loss": 1.0515, "step": 4294 }, { "epoch": 0.29100887594010433, "grad_norm": 7.479793548583984, "learning_rate": 9.514819631733864e-05, "loss": 0.9634, "step": 4295 }, { "epoch": 0.2910766312080764, "grad_norm": 6.750072956085205, "learning_rate": 9.514682729824082e-05, "loss": 0.8536, "step": 4296 }, { "epoch": 0.2911443864760485, "grad_norm": 6.5008931159973145, "learning_rate": 9.5145458279143e-05, "loss": 0.8414, "step": 4297 }, { "epoch": 0.2912121417440206, "grad_norm": 5.633193016052246, "learning_rate": 9.514408926004518e-05, "loss": 0.9628, "step": 4298 }, { "epoch": 0.2912798970119927, "grad_norm": 7.052159309387207, "learning_rate": 9.514272024094737e-05, "loss": 0.9612, "step": 4299 }, { "epoch": 0.2913476522799648, "grad_norm": 6.944911003112793, "learning_rate": 9.514135122184955e-05, "loss": 0.8237, "step": 4300 }, { "epoch": 0.2914154075479369, "grad_norm": 6.553284645080566, "learning_rate": 9.513998220275173e-05, "loss": 1.103, "step": 4301 }, { "epoch": 0.2914831628159089, "grad_norm": 6.414531707763672, "learning_rate": 9.513861318365391e-05, "loss": 1.163, "step": 4302 }, { "epoch": 0.291550918083881, "grad_norm": 9.551664352416992, "learning_rate": 9.513724416455611e-05, "loss": 1.0118, "step": 4303 }, { "epoch": 0.2916186733518531, "grad_norm": 8.332045555114746, "learning_rate": 9.513587514545829e-05, "loss": 1.0938, "step": 4304 }, { "epoch": 0.2916864286198252, "grad_norm": 7.512511253356934, "learning_rate": 9.513450612636047e-05, "loss": 0.9499, "step": 4305 }, { "epoch": 0.2917541838877973, "grad_norm": 7.233335494995117, "learning_rate": 9.513313710726265e-05, "loss": 0.9298, "step": 4306 }, { "epoch": 0.29182193915576937, "grad_norm": 6.711233615875244, "learning_rate": 9.513176808816483e-05, "loss": 0.8409, "step": 4307 }, { "epoch": 0.29188969442374146, "grad_norm": 6.433267593383789, "learning_rate": 9.513039906906702e-05, "loss": 0.7654, "step": 4308 }, { "epoch": 0.29195744969171356, "grad_norm": 7.049487113952637, "learning_rate": 9.51290300499692e-05, "loss": 0.8363, "step": 4309 }, { "epoch": 0.2920252049596856, "grad_norm": 7.030927658081055, "learning_rate": 9.512766103087138e-05, "loss": 0.8536, "step": 4310 }, { "epoch": 0.2920929602276577, "grad_norm": 7.723023414611816, "learning_rate": 9.512629201177356e-05, "loss": 0.8738, "step": 4311 }, { "epoch": 0.2921607154956298, "grad_norm": 6.293034553527832, "learning_rate": 9.512492299267576e-05, "loss": 0.8247, "step": 4312 }, { "epoch": 0.29222847076360187, "grad_norm": 7.551876544952393, "learning_rate": 9.512355397357794e-05, "loss": 0.9144, "step": 4313 }, { "epoch": 0.29229622603157396, "grad_norm": 7.78217077255249, "learning_rate": 9.512218495448012e-05, "loss": 0.911, "step": 4314 }, { "epoch": 0.29236398129954605, "grad_norm": 6.316245079040527, "learning_rate": 9.51208159353823e-05, "loss": 0.7426, "step": 4315 }, { "epoch": 0.29243173656751814, "grad_norm": 6.730735778808594, "learning_rate": 9.511944691628448e-05, "loss": 0.8938, "step": 4316 }, { "epoch": 0.29249949183549023, "grad_norm": 8.128830909729004, "learning_rate": 9.511807789718667e-05, "loss": 1.129, "step": 4317 }, { "epoch": 0.29256724710346227, "grad_norm": 8.023150444030762, "learning_rate": 9.511670887808885e-05, "loss": 1.1377, "step": 4318 }, { "epoch": 0.29263500237143436, "grad_norm": 9.075041770935059, "learning_rate": 9.511533985899103e-05, "loss": 1.0358, "step": 4319 }, { "epoch": 0.29270275763940645, "grad_norm": 9.679409980773926, "learning_rate": 9.511397083989321e-05, "loss": 0.9415, "step": 4320 }, { "epoch": 0.29277051290737854, "grad_norm": 7.316888809204102, "learning_rate": 9.51126018207954e-05, "loss": 0.9422, "step": 4321 }, { "epoch": 0.29283826817535064, "grad_norm": 8.017754554748535, "learning_rate": 9.511123280169759e-05, "loss": 0.9685, "step": 4322 }, { "epoch": 0.2929060234433227, "grad_norm": 7.990202903747559, "learning_rate": 9.510986378259977e-05, "loss": 0.9028, "step": 4323 }, { "epoch": 0.2929737787112948, "grad_norm": 7.538297653198242, "learning_rate": 9.510849476350195e-05, "loss": 0.8146, "step": 4324 }, { "epoch": 0.2930415339792669, "grad_norm": 6.794483184814453, "learning_rate": 9.510712574440413e-05, "loss": 0.8859, "step": 4325 }, { "epoch": 0.29310928924723895, "grad_norm": 9.820608139038086, "learning_rate": 9.510575672530632e-05, "loss": 1.0486, "step": 4326 }, { "epoch": 0.29317704451521104, "grad_norm": 8.107736587524414, "learning_rate": 9.51043877062085e-05, "loss": 0.9978, "step": 4327 }, { "epoch": 0.29324479978318313, "grad_norm": 8.129117965698242, "learning_rate": 9.510301868711068e-05, "loss": 0.9966, "step": 4328 }, { "epoch": 0.2933125550511552, "grad_norm": 7.15314245223999, "learning_rate": 9.510164966801288e-05, "loss": 0.8078, "step": 4329 }, { "epoch": 0.2933803103191273, "grad_norm": 7.951611518859863, "learning_rate": 9.510028064891506e-05, "loss": 0.8566, "step": 4330 }, { "epoch": 0.2934480655870994, "grad_norm": 8.350584030151367, "learning_rate": 9.509891162981724e-05, "loss": 1.1757, "step": 4331 }, { "epoch": 0.2935158208550715, "grad_norm": 7.556168079376221, "learning_rate": 9.509754261071943e-05, "loss": 1.175, "step": 4332 }, { "epoch": 0.2935835761230436, "grad_norm": 8.102100372314453, "learning_rate": 9.509617359162161e-05, "loss": 0.8948, "step": 4333 }, { "epoch": 0.2936513313910156, "grad_norm": 6.762083530426025, "learning_rate": 9.509480457252379e-05, "loss": 0.9597, "step": 4334 }, { "epoch": 0.2937190866589877, "grad_norm": 7.2654948234558105, "learning_rate": 9.509343555342599e-05, "loss": 0.9352, "step": 4335 }, { "epoch": 0.2937868419269598, "grad_norm": 8.114389419555664, "learning_rate": 9.509206653432817e-05, "loss": 0.9758, "step": 4336 }, { "epoch": 0.2938545971949319, "grad_norm": 6.972601890563965, "learning_rate": 9.509069751523035e-05, "loss": 0.9322, "step": 4337 }, { "epoch": 0.293922352462904, "grad_norm": 7.261653423309326, "learning_rate": 9.508932849613253e-05, "loss": 0.8714, "step": 4338 }, { "epoch": 0.2939901077308761, "grad_norm": 8.368372917175293, "learning_rate": 9.508795947703471e-05, "loss": 1.1235, "step": 4339 }, { "epoch": 0.2940578629988482, "grad_norm": 8.295226097106934, "learning_rate": 9.50865904579369e-05, "loss": 1.0647, "step": 4340 }, { "epoch": 0.29412561826682027, "grad_norm": 8.720281600952148, "learning_rate": 9.508522143883908e-05, "loss": 1.0779, "step": 4341 }, { "epoch": 0.29419337353479236, "grad_norm": 6.721040725708008, "learning_rate": 9.508385241974126e-05, "loss": 1.0219, "step": 4342 }, { "epoch": 0.2942611288027644, "grad_norm": 6.298686504364014, "learning_rate": 9.508248340064344e-05, "loss": 0.8011, "step": 4343 }, { "epoch": 0.2943288840707365, "grad_norm": 6.915674686431885, "learning_rate": 9.508111438154562e-05, "loss": 0.9702, "step": 4344 }, { "epoch": 0.2943966393387086, "grad_norm": 6.907165050506592, "learning_rate": 9.507974536244782e-05, "loss": 0.861, "step": 4345 }, { "epoch": 0.29446439460668067, "grad_norm": 7.540262699127197, "learning_rate": 9.507837634335e-05, "loss": 0.9426, "step": 4346 }, { "epoch": 0.29453214987465276, "grad_norm": 7.145787715911865, "learning_rate": 9.507700732425218e-05, "loss": 0.911, "step": 4347 }, { "epoch": 0.29459990514262485, "grad_norm": 9.937151908874512, "learning_rate": 9.507563830515436e-05, "loss": 0.9825, "step": 4348 }, { "epoch": 0.29466766041059694, "grad_norm": 7.125603199005127, "learning_rate": 9.507426928605655e-05, "loss": 0.855, "step": 4349 }, { "epoch": 0.29473541567856903, "grad_norm": 7.014605522155762, "learning_rate": 9.507290026695873e-05, "loss": 0.9282, "step": 4350 }, { "epoch": 0.29480317094654107, "grad_norm": 8.753725051879883, "learning_rate": 9.507153124786091e-05, "loss": 1.1047, "step": 4351 }, { "epoch": 0.29487092621451316, "grad_norm": 7.921840190887451, "learning_rate": 9.50701622287631e-05, "loss": 0.7857, "step": 4352 }, { "epoch": 0.29493868148248525, "grad_norm": 6.762521266937256, "learning_rate": 9.506879320966527e-05, "loss": 0.8537, "step": 4353 }, { "epoch": 0.29500643675045735, "grad_norm": 8.441102981567383, "learning_rate": 9.506742419056747e-05, "loss": 1.1356, "step": 4354 }, { "epoch": 0.29507419201842944, "grad_norm": 6.748636245727539, "learning_rate": 9.506605517146965e-05, "loss": 1.0962, "step": 4355 }, { "epoch": 0.29514194728640153, "grad_norm": 8.168306350708008, "learning_rate": 9.506468615237183e-05, "loss": 0.9628, "step": 4356 }, { "epoch": 0.2952097025543736, "grad_norm": 6.545300483703613, "learning_rate": 9.506331713327401e-05, "loss": 1.0041, "step": 4357 }, { "epoch": 0.2952774578223457, "grad_norm": 6.958950996398926, "learning_rate": 9.50619481141762e-05, "loss": 0.9742, "step": 4358 }, { "epoch": 0.29534521309031775, "grad_norm": 6.817789554595947, "learning_rate": 9.506057909507838e-05, "loss": 0.8281, "step": 4359 }, { "epoch": 0.29541296835828984, "grad_norm": 8.255404472351074, "learning_rate": 9.505921007598056e-05, "loss": 0.9171, "step": 4360 }, { "epoch": 0.29548072362626193, "grad_norm": 7.552668571472168, "learning_rate": 9.505784105688274e-05, "loss": 0.8665, "step": 4361 }, { "epoch": 0.295548478894234, "grad_norm": 6.980686664581299, "learning_rate": 9.505647203778492e-05, "loss": 0.9461, "step": 4362 }, { "epoch": 0.2956162341622061, "grad_norm": 6.765429973602295, "learning_rate": 9.505510301868712e-05, "loss": 0.7946, "step": 4363 }, { "epoch": 0.2956839894301782, "grad_norm": 8.403508186340332, "learning_rate": 9.50537339995893e-05, "loss": 0.9015, "step": 4364 }, { "epoch": 0.2957517446981503, "grad_norm": 7.822012901306152, "learning_rate": 9.505236498049148e-05, "loss": 1.0313, "step": 4365 }, { "epoch": 0.2958194999661224, "grad_norm": 9.843314170837402, "learning_rate": 9.505099596139366e-05, "loss": 1.11, "step": 4366 }, { "epoch": 0.2958872552340944, "grad_norm": 8.155379295349121, "learning_rate": 9.504962694229584e-05, "loss": 1.1565, "step": 4367 }, { "epoch": 0.2959550105020665, "grad_norm": 8.446005821228027, "learning_rate": 9.504825792319803e-05, "loss": 1.0588, "step": 4368 }, { "epoch": 0.2960227657700386, "grad_norm": 7.244038105010986, "learning_rate": 9.504688890410021e-05, "loss": 0.8648, "step": 4369 }, { "epoch": 0.2960905210380107, "grad_norm": 8.324630737304688, "learning_rate": 9.50455198850024e-05, "loss": 0.9862, "step": 4370 }, { "epoch": 0.2961582763059828, "grad_norm": 7.308585166931152, "learning_rate": 9.504415086590457e-05, "loss": 1.082, "step": 4371 }, { "epoch": 0.2962260315739549, "grad_norm": 8.17287826538086, "learning_rate": 9.504278184680677e-05, "loss": 1.0359, "step": 4372 }, { "epoch": 0.296293786841927, "grad_norm": 6.8774943351745605, "learning_rate": 9.504141282770895e-05, "loss": 0.904, "step": 4373 }, { "epoch": 0.29636154210989907, "grad_norm": 8.641596794128418, "learning_rate": 9.504004380861113e-05, "loss": 0.9822, "step": 4374 }, { "epoch": 0.2964292973778711, "grad_norm": 7.289584159851074, "learning_rate": 9.503867478951331e-05, "loss": 1.0274, "step": 4375 }, { "epoch": 0.2964970526458432, "grad_norm": 9.820521354675293, "learning_rate": 9.50373057704155e-05, "loss": 1.0895, "step": 4376 }, { "epoch": 0.2965648079138153, "grad_norm": 8.6587553024292, "learning_rate": 9.503593675131768e-05, "loss": 0.7513, "step": 4377 }, { "epoch": 0.2966325631817874, "grad_norm": 7.444339752197266, "learning_rate": 9.503456773221986e-05, "loss": 1.0892, "step": 4378 }, { "epoch": 0.29670031844975947, "grad_norm": 6.4494948387146, "learning_rate": 9.503319871312206e-05, "loss": 0.9465, "step": 4379 }, { "epoch": 0.29676807371773156, "grad_norm": 7.744002819061279, "learning_rate": 9.503182969402424e-05, "loss": 1.0613, "step": 4380 }, { "epoch": 0.29683582898570365, "grad_norm": 8.152511596679688, "learning_rate": 9.503046067492642e-05, "loss": 1.0115, "step": 4381 }, { "epoch": 0.29690358425367575, "grad_norm": 6.076781749725342, "learning_rate": 9.502909165582861e-05, "loss": 0.8255, "step": 4382 }, { "epoch": 0.2969713395216478, "grad_norm": 6.787397861480713, "learning_rate": 9.502772263673079e-05, "loss": 1.1132, "step": 4383 }, { "epoch": 0.2970390947896199, "grad_norm": 8.592107772827148, "learning_rate": 9.502635361763297e-05, "loss": 0.8715, "step": 4384 }, { "epoch": 0.29710685005759196, "grad_norm": 7.620471954345703, "learning_rate": 9.502498459853515e-05, "loss": 0.9521, "step": 4385 }, { "epoch": 0.29717460532556406, "grad_norm": 8.026017189025879, "learning_rate": 9.502361557943735e-05, "loss": 0.8575, "step": 4386 }, { "epoch": 0.29724236059353615, "grad_norm": 8.82768726348877, "learning_rate": 9.502224656033953e-05, "loss": 1.0642, "step": 4387 }, { "epoch": 0.29731011586150824, "grad_norm": 7.179310321807861, "learning_rate": 9.502087754124171e-05, "loss": 0.9161, "step": 4388 }, { "epoch": 0.29737787112948033, "grad_norm": 7.4363861083984375, "learning_rate": 9.501950852214389e-05, "loss": 1.0825, "step": 4389 }, { "epoch": 0.2974456263974524, "grad_norm": 6.218450546264648, "learning_rate": 9.501813950304608e-05, "loss": 0.8461, "step": 4390 }, { "epoch": 0.2975133816654245, "grad_norm": 6.540037631988525, "learning_rate": 9.501677048394826e-05, "loss": 0.9099, "step": 4391 }, { "epoch": 0.29758113693339655, "grad_norm": 9.617361068725586, "learning_rate": 9.501540146485044e-05, "loss": 0.9205, "step": 4392 }, { "epoch": 0.29764889220136864, "grad_norm": 8.519254684448242, "learning_rate": 9.501403244575262e-05, "loss": 0.9774, "step": 4393 }, { "epoch": 0.29771664746934073, "grad_norm": 8.101237297058105, "learning_rate": 9.50126634266548e-05, "loss": 0.9019, "step": 4394 }, { "epoch": 0.2977844027373128, "grad_norm": 6.703762531280518, "learning_rate": 9.5011294407557e-05, "loss": 0.9486, "step": 4395 }, { "epoch": 0.2978521580052849, "grad_norm": 8.651348114013672, "learning_rate": 9.500992538845918e-05, "loss": 0.9189, "step": 4396 }, { "epoch": 0.297919913273257, "grad_norm": 7.783169269561768, "learning_rate": 9.500855636936136e-05, "loss": 0.933, "step": 4397 }, { "epoch": 0.2979876685412291, "grad_norm": 8.011998176574707, "learning_rate": 9.500718735026354e-05, "loss": 1.0642, "step": 4398 }, { "epoch": 0.2980554238092012, "grad_norm": 10.782837867736816, "learning_rate": 9.500581833116572e-05, "loss": 1.1406, "step": 4399 }, { "epoch": 0.29812317907717323, "grad_norm": 7.699460029602051, "learning_rate": 9.500444931206791e-05, "loss": 0.9627, "step": 4400 }, { "epoch": 0.2981909343451453, "grad_norm": 8.144030570983887, "learning_rate": 9.500308029297009e-05, "loss": 0.8278, "step": 4401 }, { "epoch": 0.2982586896131174, "grad_norm": 7.882628917694092, "learning_rate": 9.500171127387227e-05, "loss": 0.7924, "step": 4402 }, { "epoch": 0.2983264448810895, "grad_norm": 6.887333393096924, "learning_rate": 9.500034225477445e-05, "loss": 0.8385, "step": 4403 }, { "epoch": 0.2983942001490616, "grad_norm": 8.446837425231934, "learning_rate": 9.499897323567665e-05, "loss": 1.0352, "step": 4404 }, { "epoch": 0.2984619554170337, "grad_norm": 8.389452934265137, "learning_rate": 9.499760421657883e-05, "loss": 0.8053, "step": 4405 }, { "epoch": 0.2985297106850058, "grad_norm": 7.09201717376709, "learning_rate": 9.499623519748101e-05, "loss": 0.999, "step": 4406 }, { "epoch": 0.29859746595297787, "grad_norm": 9.670032501220703, "learning_rate": 9.499486617838319e-05, "loss": 0.8558, "step": 4407 }, { "epoch": 0.2986652212209499, "grad_norm": 9.050987243652344, "learning_rate": 9.499349715928537e-05, "loss": 1.0356, "step": 4408 }, { "epoch": 0.298732976488922, "grad_norm": 6.220524787902832, "learning_rate": 9.499212814018756e-05, "loss": 0.8278, "step": 4409 }, { "epoch": 0.2988007317568941, "grad_norm": 7.464169979095459, "learning_rate": 9.499075912108974e-05, "loss": 0.963, "step": 4410 }, { "epoch": 0.2988684870248662, "grad_norm": 8.479538917541504, "learning_rate": 9.498939010199192e-05, "loss": 1.1108, "step": 4411 }, { "epoch": 0.2989362422928383, "grad_norm": 8.23353385925293, "learning_rate": 9.49880210828941e-05, "loss": 1.0156, "step": 4412 }, { "epoch": 0.29900399756081036, "grad_norm": 7.2161335945129395, "learning_rate": 9.49866520637963e-05, "loss": 0.9964, "step": 4413 }, { "epoch": 0.29907175282878246, "grad_norm": 7.893895149230957, "learning_rate": 9.498528304469848e-05, "loss": 0.9353, "step": 4414 }, { "epoch": 0.29913950809675455, "grad_norm": 8.448126792907715, "learning_rate": 9.498391402560066e-05, "loss": 0.9488, "step": 4415 }, { "epoch": 0.2992072633647266, "grad_norm": 8.160601615905762, "learning_rate": 9.498254500650284e-05, "loss": 0.8934, "step": 4416 }, { "epoch": 0.2992750186326987, "grad_norm": 7.6815266609191895, "learning_rate": 9.498117598740502e-05, "loss": 0.8938, "step": 4417 }, { "epoch": 0.29934277390067077, "grad_norm": 6.795403480529785, "learning_rate": 9.497980696830721e-05, "loss": 0.7074, "step": 4418 }, { "epoch": 0.29941052916864286, "grad_norm": 7.904134750366211, "learning_rate": 9.497843794920939e-05, "loss": 0.8747, "step": 4419 }, { "epoch": 0.29947828443661495, "grad_norm": 7.776915073394775, "learning_rate": 9.497706893011157e-05, "loss": 1.1141, "step": 4420 }, { "epoch": 0.29954603970458704, "grad_norm": 7.425099849700928, "learning_rate": 9.497569991101375e-05, "loss": 0.8826, "step": 4421 }, { "epoch": 0.29961379497255913, "grad_norm": 10.530119895935059, "learning_rate": 9.497433089191595e-05, "loss": 1.0893, "step": 4422 }, { "epoch": 0.2996815502405312, "grad_norm": 7.118913650512695, "learning_rate": 9.497296187281813e-05, "loss": 0.9179, "step": 4423 }, { "epoch": 0.29974930550850326, "grad_norm": 6.6407060623168945, "learning_rate": 9.497159285372031e-05, "loss": 0.8763, "step": 4424 }, { "epoch": 0.29981706077647535, "grad_norm": 6.975287437438965, "learning_rate": 9.49702238346225e-05, "loss": 0.9192, "step": 4425 }, { "epoch": 0.29988481604444744, "grad_norm": 11.773819923400879, "learning_rate": 9.496885481552468e-05, "loss": 1.3173, "step": 4426 }, { "epoch": 0.29995257131241954, "grad_norm": 13.157849311828613, "learning_rate": 9.496748579642686e-05, "loss": 0.8773, "step": 4427 }, { "epoch": 0.3000203265803916, "grad_norm": 7.372555732727051, "learning_rate": 9.496611677732906e-05, "loss": 1.0028, "step": 4428 }, { "epoch": 0.3000880818483637, "grad_norm": 8.804738998413086, "learning_rate": 9.496474775823124e-05, "loss": 1.0684, "step": 4429 }, { "epoch": 0.3001558371163358, "grad_norm": 7.723897933959961, "learning_rate": 9.496337873913342e-05, "loss": 0.9113, "step": 4430 }, { "epoch": 0.3002235923843079, "grad_norm": 10.261975288391113, "learning_rate": 9.49620097200356e-05, "loss": 1.242, "step": 4431 }, { "epoch": 0.30029134765227994, "grad_norm": 7.396425247192383, "learning_rate": 9.496064070093779e-05, "loss": 1.1045, "step": 4432 }, { "epoch": 0.30035910292025203, "grad_norm": 9.235404014587402, "learning_rate": 9.495927168183997e-05, "loss": 1.1058, "step": 4433 }, { "epoch": 0.3004268581882241, "grad_norm": 6.465355396270752, "learning_rate": 9.495790266274215e-05, "loss": 0.9163, "step": 4434 }, { "epoch": 0.3004946134561962, "grad_norm": 6.401477813720703, "learning_rate": 9.495653364364433e-05, "loss": 0.8224, "step": 4435 }, { "epoch": 0.3005623687241683, "grad_norm": 6.599462985992432, "learning_rate": 9.495516462454653e-05, "loss": 0.8104, "step": 4436 }, { "epoch": 0.3006301239921404, "grad_norm": 8.553694725036621, "learning_rate": 9.495379560544871e-05, "loss": 1.1467, "step": 4437 }, { "epoch": 0.3006978792601125, "grad_norm": 6.262206554412842, "learning_rate": 9.495242658635089e-05, "loss": 0.7903, "step": 4438 }, { "epoch": 0.3007656345280846, "grad_norm": 7.284942150115967, "learning_rate": 9.495105756725307e-05, "loss": 1.0585, "step": 4439 }, { "epoch": 0.3008333897960566, "grad_norm": 8.880256652832031, "learning_rate": 9.494968854815525e-05, "loss": 1.0553, "step": 4440 }, { "epoch": 0.3009011450640287, "grad_norm": 9.358749389648438, "learning_rate": 9.494831952905744e-05, "loss": 1.1874, "step": 4441 }, { "epoch": 0.3009689003320008, "grad_norm": 6.7088470458984375, "learning_rate": 9.494695050995962e-05, "loss": 0.8641, "step": 4442 }, { "epoch": 0.3010366555999729, "grad_norm": 9.420478820800781, "learning_rate": 9.49455814908618e-05, "loss": 0.8829, "step": 4443 }, { "epoch": 0.301104410867945, "grad_norm": 8.877934455871582, "learning_rate": 9.494421247176398e-05, "loss": 1.1206, "step": 4444 }, { "epoch": 0.3011721661359171, "grad_norm": 9.361932754516602, "learning_rate": 9.494284345266618e-05, "loss": 0.7993, "step": 4445 }, { "epoch": 0.30123992140388917, "grad_norm": 7.4104790687561035, "learning_rate": 9.494147443356836e-05, "loss": 1.0655, "step": 4446 }, { "epoch": 0.30130767667186126, "grad_norm": 7.0485992431640625, "learning_rate": 9.494010541447054e-05, "loss": 0.8878, "step": 4447 }, { "epoch": 0.30137543193983335, "grad_norm": 7.162468433380127, "learning_rate": 9.493873639537272e-05, "loss": 0.9467, "step": 4448 }, { "epoch": 0.3014431872078054, "grad_norm": 9.214662551879883, "learning_rate": 9.49373673762749e-05, "loss": 0.9288, "step": 4449 }, { "epoch": 0.3015109424757775, "grad_norm": 11.189544677734375, "learning_rate": 9.493599835717709e-05, "loss": 0.9933, "step": 4450 }, { "epoch": 0.30157869774374957, "grad_norm": 9.398331642150879, "learning_rate": 9.493462933807927e-05, "loss": 0.9248, "step": 4451 }, { "epoch": 0.30164645301172166, "grad_norm": 8.26975154876709, "learning_rate": 9.493326031898145e-05, "loss": 1.0308, "step": 4452 }, { "epoch": 0.30171420827969375, "grad_norm": 9.262918472290039, "learning_rate": 9.493189129988363e-05, "loss": 0.9367, "step": 4453 }, { "epoch": 0.30178196354766584, "grad_norm": 7.079284191131592, "learning_rate": 9.493052228078581e-05, "loss": 1.0415, "step": 4454 }, { "epoch": 0.30184971881563794, "grad_norm": 9.081875801086426, "learning_rate": 9.492915326168801e-05, "loss": 1.0783, "step": 4455 }, { "epoch": 0.30191747408361, "grad_norm": 8.476323127746582, "learning_rate": 9.492778424259019e-05, "loss": 0.9743, "step": 4456 }, { "epoch": 0.30198522935158206, "grad_norm": 7.064181327819824, "learning_rate": 9.492641522349237e-05, "loss": 0.7966, "step": 4457 }, { "epoch": 0.30205298461955415, "grad_norm": 7.837399959564209, "learning_rate": 9.492504620439455e-05, "loss": 0.8757, "step": 4458 }, { "epoch": 0.30212073988752625, "grad_norm": 7.2140936851501465, "learning_rate": 9.492367718529674e-05, "loss": 1.1334, "step": 4459 }, { "epoch": 0.30218849515549834, "grad_norm": 7.836158752441406, "learning_rate": 9.492230816619892e-05, "loss": 1.1179, "step": 4460 }, { "epoch": 0.30225625042347043, "grad_norm": 8.686471939086914, "learning_rate": 9.49209391471011e-05, "loss": 0.9826, "step": 4461 }, { "epoch": 0.3023240056914425, "grad_norm": 8.442061424255371, "learning_rate": 9.491957012800328e-05, "loss": 1.0392, "step": 4462 }, { "epoch": 0.3023917609594146, "grad_norm": 7.024100303649902, "learning_rate": 9.491820110890546e-05, "loss": 0.7857, "step": 4463 }, { "epoch": 0.3024595162273867, "grad_norm": 7.306611061096191, "learning_rate": 9.491683208980766e-05, "loss": 1.0154, "step": 4464 }, { "epoch": 0.30252727149535874, "grad_norm": 7.3064470291137695, "learning_rate": 9.491546307070984e-05, "loss": 1.0311, "step": 4465 }, { "epoch": 0.30259502676333083, "grad_norm": 7.264878749847412, "learning_rate": 9.491409405161202e-05, "loss": 0.932, "step": 4466 }, { "epoch": 0.3026627820313029, "grad_norm": 7.43487024307251, "learning_rate": 9.49127250325142e-05, "loss": 0.8024, "step": 4467 }, { "epoch": 0.302730537299275, "grad_norm": 7.710512161254883, "learning_rate": 9.491135601341639e-05, "loss": 0.7989, "step": 4468 }, { "epoch": 0.3027982925672471, "grad_norm": 7.263044834136963, "learning_rate": 9.490998699431857e-05, "loss": 0.9481, "step": 4469 }, { "epoch": 0.3028660478352192, "grad_norm": 7.284390449523926, "learning_rate": 9.490861797522075e-05, "loss": 0.8008, "step": 4470 }, { "epoch": 0.3029338031031913, "grad_norm": 8.490337371826172, "learning_rate": 9.490724895612295e-05, "loss": 1.1867, "step": 4471 }, { "epoch": 0.3030015583711634, "grad_norm": 6.555050849914551, "learning_rate": 9.490587993702513e-05, "loss": 0.8615, "step": 4472 }, { "epoch": 0.3030693136391354, "grad_norm": 6.497532367706299, "learning_rate": 9.490451091792731e-05, "loss": 0.9643, "step": 4473 }, { "epoch": 0.3031370689071075, "grad_norm": 6.556066036224365, "learning_rate": 9.49031418988295e-05, "loss": 0.9757, "step": 4474 }, { "epoch": 0.3032048241750796, "grad_norm": 6.649857044219971, "learning_rate": 9.490177287973168e-05, "loss": 1.0196, "step": 4475 }, { "epoch": 0.3032725794430517, "grad_norm": 8.725894927978516, "learning_rate": 9.490040386063386e-05, "loss": 1.1662, "step": 4476 }, { "epoch": 0.3033403347110238, "grad_norm": 6.520106792449951, "learning_rate": 9.489903484153604e-05, "loss": 0.8353, "step": 4477 }, { "epoch": 0.3034080899789959, "grad_norm": 6.688689708709717, "learning_rate": 9.489766582243824e-05, "loss": 0.8101, "step": 4478 }, { "epoch": 0.30347584524696797, "grad_norm": 7.817204475402832, "learning_rate": 9.489629680334042e-05, "loss": 0.9354, "step": 4479 }, { "epoch": 0.30354360051494006, "grad_norm": 7.536436080932617, "learning_rate": 9.48949277842426e-05, "loss": 1.2358, "step": 4480 }, { "epoch": 0.3036113557829121, "grad_norm": 8.635372161865234, "learning_rate": 9.489355876514478e-05, "loss": 0.8344, "step": 4481 }, { "epoch": 0.3036791110508842, "grad_norm": 7.561103343963623, "learning_rate": 9.489218974604697e-05, "loss": 0.864, "step": 4482 }, { "epoch": 0.3037468663188563, "grad_norm": 8.284873962402344, "learning_rate": 9.489082072694915e-05, "loss": 0.8756, "step": 4483 }, { "epoch": 0.30381462158682837, "grad_norm": 8.976608276367188, "learning_rate": 9.488945170785133e-05, "loss": 1.0523, "step": 4484 }, { "epoch": 0.30388237685480046, "grad_norm": 11.891230583190918, "learning_rate": 9.488808268875351e-05, "loss": 1.0367, "step": 4485 }, { "epoch": 0.30395013212277255, "grad_norm": 8.247873306274414, "learning_rate": 9.488671366965569e-05, "loss": 0.8173, "step": 4486 }, { "epoch": 0.30401788739074465, "grad_norm": 6.74050760269165, "learning_rate": 9.488534465055789e-05, "loss": 0.8192, "step": 4487 }, { "epoch": 0.30408564265871674, "grad_norm": 7.035127639770508, "learning_rate": 9.488397563146007e-05, "loss": 0.8416, "step": 4488 }, { "epoch": 0.3041533979266888, "grad_norm": 7.281885623931885, "learning_rate": 9.488260661236225e-05, "loss": 1.072, "step": 4489 }, { "epoch": 0.30422115319466086, "grad_norm": 6.236363887786865, "learning_rate": 9.488123759326443e-05, "loss": 0.8738, "step": 4490 }, { "epoch": 0.30428890846263296, "grad_norm": 8.20182991027832, "learning_rate": 9.487986857416662e-05, "loss": 1.1965, "step": 4491 }, { "epoch": 0.30435666373060505, "grad_norm": 9.837231636047363, "learning_rate": 9.48784995550688e-05, "loss": 0.9125, "step": 4492 }, { "epoch": 0.30442441899857714, "grad_norm": 8.810354232788086, "learning_rate": 9.487713053597098e-05, "loss": 1.0623, "step": 4493 }, { "epoch": 0.30449217426654923, "grad_norm": 6.360396385192871, "learning_rate": 9.487576151687316e-05, "loss": 0.8683, "step": 4494 }, { "epoch": 0.3045599295345213, "grad_norm": 8.158722877502441, "learning_rate": 9.487439249777534e-05, "loss": 0.9562, "step": 4495 }, { "epoch": 0.3046276848024934, "grad_norm": 12.142204284667969, "learning_rate": 9.487302347867754e-05, "loss": 0.8937, "step": 4496 }, { "epoch": 0.3046954400704655, "grad_norm": 8.125468254089355, "learning_rate": 9.487165445957972e-05, "loss": 1.0587, "step": 4497 }, { "epoch": 0.30476319533843754, "grad_norm": 7.607559680938721, "learning_rate": 9.48702854404819e-05, "loss": 0.8296, "step": 4498 }, { "epoch": 0.30483095060640963, "grad_norm": 6.430202484130859, "learning_rate": 9.486891642138408e-05, "loss": 1.067, "step": 4499 }, { "epoch": 0.3048987058743817, "grad_norm": 7.2482805252075195, "learning_rate": 9.486754740228626e-05, "loss": 1.0611, "step": 4500 }, { "epoch": 0.3049664611423538, "grad_norm": 9.431131362915039, "learning_rate": 9.486617838318845e-05, "loss": 0.9665, "step": 4501 }, { "epoch": 0.3050342164103259, "grad_norm": 8.319073677062988, "learning_rate": 9.486480936409063e-05, "loss": 0.983, "step": 4502 }, { "epoch": 0.305101971678298, "grad_norm": 6.784060478210449, "learning_rate": 9.486344034499281e-05, "loss": 0.8534, "step": 4503 }, { "epoch": 0.3051697269462701, "grad_norm": 7.521092414855957, "learning_rate": 9.486207132589499e-05, "loss": 0.9167, "step": 4504 }, { "epoch": 0.3052374822142422, "grad_norm": 7.7156982421875, "learning_rate": 9.486070230679719e-05, "loss": 0.7691, "step": 4505 }, { "epoch": 0.3053052374822142, "grad_norm": 6.727222919464111, "learning_rate": 9.485933328769937e-05, "loss": 0.7156, "step": 4506 }, { "epoch": 0.3053729927501863, "grad_norm": 7.416358470916748, "learning_rate": 9.485796426860155e-05, "loss": 1.0458, "step": 4507 }, { "epoch": 0.3054407480181584, "grad_norm": 8.601702690124512, "learning_rate": 9.485659524950373e-05, "loss": 1.1664, "step": 4508 }, { "epoch": 0.3055085032861305, "grad_norm": 6.742228984832764, "learning_rate": 9.485522623040591e-05, "loss": 0.9715, "step": 4509 }, { "epoch": 0.3055762585541026, "grad_norm": 6.782869815826416, "learning_rate": 9.48538572113081e-05, "loss": 0.7644, "step": 4510 }, { "epoch": 0.3056440138220747, "grad_norm": 5.753969669342041, "learning_rate": 9.485248819221028e-05, "loss": 0.8468, "step": 4511 }, { "epoch": 0.30571176909004677, "grad_norm": 10.150800704956055, "learning_rate": 9.485111917311246e-05, "loss": 1.1521, "step": 4512 }, { "epoch": 0.30577952435801886, "grad_norm": 7.476456642150879, "learning_rate": 9.484975015401464e-05, "loss": 1.0091, "step": 4513 }, { "epoch": 0.3058472796259909, "grad_norm": 8.955781936645508, "learning_rate": 9.484838113491684e-05, "loss": 1.1837, "step": 4514 }, { "epoch": 0.305915034893963, "grad_norm": 8.779487609863281, "learning_rate": 9.484701211581902e-05, "loss": 0.9747, "step": 4515 }, { "epoch": 0.3059827901619351, "grad_norm": 9.432082176208496, "learning_rate": 9.48456430967212e-05, "loss": 0.9298, "step": 4516 }, { "epoch": 0.3060505454299072, "grad_norm": 8.186026573181152, "learning_rate": 9.484427407762339e-05, "loss": 0.9026, "step": 4517 }, { "epoch": 0.30611830069787926, "grad_norm": 7.902477264404297, "learning_rate": 9.484290505852557e-05, "loss": 1.0116, "step": 4518 }, { "epoch": 0.30618605596585136, "grad_norm": 9.117185592651367, "learning_rate": 9.484153603942775e-05, "loss": 1.0867, "step": 4519 }, { "epoch": 0.30625381123382345, "grad_norm": 8.473755836486816, "learning_rate": 9.484016702032995e-05, "loss": 0.8742, "step": 4520 }, { "epoch": 0.30632156650179554, "grad_norm": 12.328755378723145, "learning_rate": 9.483879800123213e-05, "loss": 1.2981, "step": 4521 }, { "epoch": 0.3063893217697676, "grad_norm": 7.099585056304932, "learning_rate": 9.483742898213431e-05, "loss": 0.9685, "step": 4522 }, { "epoch": 0.30645707703773967, "grad_norm": 8.027973175048828, "learning_rate": 9.48360599630365e-05, "loss": 1.0194, "step": 4523 }, { "epoch": 0.30652483230571176, "grad_norm": 8.275866508483887, "learning_rate": 9.483469094393868e-05, "loss": 1.2424, "step": 4524 }, { "epoch": 0.30659258757368385, "grad_norm": 8.546550750732422, "learning_rate": 9.483332192484086e-05, "loss": 0.9284, "step": 4525 }, { "epoch": 0.30666034284165594, "grad_norm": 5.6628522872924805, "learning_rate": 9.483195290574304e-05, "loss": 0.882, "step": 4526 }, { "epoch": 0.30672809810962803, "grad_norm": 7.189582824707031, "learning_rate": 9.483058388664522e-05, "loss": 0.8087, "step": 4527 }, { "epoch": 0.3067958533776001, "grad_norm": 6.341355323791504, "learning_rate": 9.482921486754742e-05, "loss": 0.8759, "step": 4528 }, { "epoch": 0.3068636086455722, "grad_norm": 6.838935852050781, "learning_rate": 9.48278458484496e-05, "loss": 0.6662, "step": 4529 }, { "epoch": 0.30693136391354425, "grad_norm": 5.441661834716797, "learning_rate": 9.482647682935178e-05, "loss": 0.7008, "step": 4530 }, { "epoch": 0.30699911918151634, "grad_norm": 9.242436408996582, "learning_rate": 9.482510781025396e-05, "loss": 0.8675, "step": 4531 }, { "epoch": 0.30706687444948844, "grad_norm": 7.393209934234619, "learning_rate": 9.482373879115614e-05, "loss": 0.7829, "step": 4532 }, { "epoch": 0.30713462971746053, "grad_norm": 8.613391876220703, "learning_rate": 9.482236977205833e-05, "loss": 1.0276, "step": 4533 }, { "epoch": 0.3072023849854326, "grad_norm": 8.517586708068848, "learning_rate": 9.482100075296051e-05, "loss": 0.8115, "step": 4534 }, { "epoch": 0.3072701402534047, "grad_norm": 7.575396537780762, "learning_rate": 9.481963173386269e-05, "loss": 0.8254, "step": 4535 }, { "epoch": 0.3073378955213768, "grad_norm": 8.90748405456543, "learning_rate": 9.481826271476487e-05, "loss": 0.9864, "step": 4536 }, { "epoch": 0.3074056507893489, "grad_norm": 10.321487426757812, "learning_rate": 9.481689369566707e-05, "loss": 1.0247, "step": 4537 }, { "epoch": 0.30747340605732093, "grad_norm": 8.15263843536377, "learning_rate": 9.481552467656925e-05, "loss": 0.8475, "step": 4538 }, { "epoch": 0.307541161325293, "grad_norm": 7.198974609375, "learning_rate": 9.481415565747143e-05, "loss": 1.1028, "step": 4539 }, { "epoch": 0.3076089165932651, "grad_norm": 8.286428451538086, "learning_rate": 9.481278663837361e-05, "loss": 0.9451, "step": 4540 }, { "epoch": 0.3076766718612372, "grad_norm": 7.620658874511719, "learning_rate": 9.481141761927579e-05, "loss": 0.8243, "step": 4541 }, { "epoch": 0.3077444271292093, "grad_norm": 7.8145341873168945, "learning_rate": 9.481004860017798e-05, "loss": 0.7741, "step": 4542 }, { "epoch": 0.3078121823971814, "grad_norm": 7.216073989868164, "learning_rate": 9.480867958108016e-05, "loss": 0.7973, "step": 4543 }, { "epoch": 0.3078799376651535, "grad_norm": 7.704121112823486, "learning_rate": 9.480731056198234e-05, "loss": 0.9388, "step": 4544 }, { "epoch": 0.30794769293312557, "grad_norm": 7.872618675231934, "learning_rate": 9.480594154288452e-05, "loss": 0.8628, "step": 4545 }, { "epoch": 0.3080154482010976, "grad_norm": 5.696948528289795, "learning_rate": 9.480457252378672e-05, "loss": 0.8319, "step": 4546 }, { "epoch": 0.3080832034690697, "grad_norm": 7.372182846069336, "learning_rate": 9.48032035046889e-05, "loss": 0.9198, "step": 4547 }, { "epoch": 0.3081509587370418, "grad_norm": 6.9523420333862305, "learning_rate": 9.480183448559108e-05, "loss": 1.0383, "step": 4548 }, { "epoch": 0.3082187140050139, "grad_norm": 7.862281799316406, "learning_rate": 9.480046546649326e-05, "loss": 0.9764, "step": 4549 }, { "epoch": 0.308286469272986, "grad_norm": 6.563671588897705, "learning_rate": 9.479909644739544e-05, "loss": 0.9587, "step": 4550 }, { "epoch": 0.30835422454095807, "grad_norm": 6.150197505950928, "learning_rate": 9.479772742829763e-05, "loss": 0.8265, "step": 4551 }, { "epoch": 0.30842197980893016, "grad_norm": 5.330137729644775, "learning_rate": 9.479635840919981e-05, "loss": 0.697, "step": 4552 }, { "epoch": 0.30848973507690225, "grad_norm": 7.886617183685303, "learning_rate": 9.479498939010199e-05, "loss": 0.8645, "step": 4553 }, { "epoch": 0.30855749034487434, "grad_norm": 8.9354248046875, "learning_rate": 9.479362037100417e-05, "loss": 0.9669, "step": 4554 }, { "epoch": 0.3086252456128464, "grad_norm": 6.7398457527160645, "learning_rate": 9.479225135190635e-05, "loss": 0.6935, "step": 4555 }, { "epoch": 0.30869300088081847, "grad_norm": 6.668737888336182, "learning_rate": 9.479088233280855e-05, "loss": 1.0038, "step": 4556 }, { "epoch": 0.30876075614879056, "grad_norm": 9.82985782623291, "learning_rate": 9.478951331371073e-05, "loss": 1.1031, "step": 4557 }, { "epoch": 0.30882851141676265, "grad_norm": 7.345351696014404, "learning_rate": 9.478814429461291e-05, "loss": 1.032, "step": 4558 }, { "epoch": 0.30889626668473474, "grad_norm": 8.863839149475098, "learning_rate": 9.478677527551509e-05, "loss": 1.0042, "step": 4559 }, { "epoch": 0.30896402195270684, "grad_norm": 7.280172824859619, "learning_rate": 9.478540625641728e-05, "loss": 0.977, "step": 4560 }, { "epoch": 0.3090317772206789, "grad_norm": 7.121008396148682, "learning_rate": 9.478403723731946e-05, "loss": 0.7969, "step": 4561 }, { "epoch": 0.309099532488651, "grad_norm": 8.222799301147461, "learning_rate": 9.478266821822164e-05, "loss": 1.0655, "step": 4562 }, { "epoch": 0.30916728775662305, "grad_norm": 7.74297571182251, "learning_rate": 9.478129919912384e-05, "loss": 0.7976, "step": 4563 }, { "epoch": 0.30923504302459515, "grad_norm": 8.211404800415039, "learning_rate": 9.477993018002602e-05, "loss": 0.9004, "step": 4564 }, { "epoch": 0.30930279829256724, "grad_norm": 6.897336006164551, "learning_rate": 9.47785611609282e-05, "loss": 0.9347, "step": 4565 }, { "epoch": 0.30937055356053933, "grad_norm": 6.508474826812744, "learning_rate": 9.477719214183039e-05, "loss": 0.862, "step": 4566 }, { "epoch": 0.3094383088285114, "grad_norm": 6.972619533538818, "learning_rate": 9.477582312273257e-05, "loss": 0.8142, "step": 4567 }, { "epoch": 0.3095060640964835, "grad_norm": 8.018939971923828, "learning_rate": 9.477445410363475e-05, "loss": 1.1858, "step": 4568 }, { "epoch": 0.3095738193644556, "grad_norm": 7.123754024505615, "learning_rate": 9.477308508453695e-05, "loss": 0.9347, "step": 4569 }, { "epoch": 0.3096415746324277, "grad_norm": 7.701295375823975, "learning_rate": 9.477171606543913e-05, "loss": 1.0511, "step": 4570 }, { "epoch": 0.30970932990039973, "grad_norm": 7.858259677886963, "learning_rate": 9.47703470463413e-05, "loss": 0.8332, "step": 4571 }, { "epoch": 0.3097770851683718, "grad_norm": 9.541449546813965, "learning_rate": 9.476897802724349e-05, "loss": 1.3054, "step": 4572 }, { "epoch": 0.3098448404363439, "grad_norm": 7.877323150634766, "learning_rate": 9.476760900814567e-05, "loss": 0.9087, "step": 4573 }, { "epoch": 0.309912595704316, "grad_norm": 6.667144775390625, "learning_rate": 9.476623998904786e-05, "loss": 0.9066, "step": 4574 }, { "epoch": 0.3099803509722881, "grad_norm": 6.78439998626709, "learning_rate": 9.476487096995004e-05, "loss": 0.9524, "step": 4575 }, { "epoch": 0.3100481062402602, "grad_norm": 6.174951076507568, "learning_rate": 9.476350195085222e-05, "loss": 0.9853, "step": 4576 }, { "epoch": 0.3101158615082323, "grad_norm": 7.872679233551025, "learning_rate": 9.47621329317544e-05, "loss": 1.2686, "step": 4577 }, { "epoch": 0.3101836167762044, "grad_norm": 6.736785411834717, "learning_rate": 9.47607639126566e-05, "loss": 0.8149, "step": 4578 }, { "epoch": 0.3102513720441764, "grad_norm": 7.353178977966309, "learning_rate": 9.475939489355878e-05, "loss": 0.7092, "step": 4579 }, { "epoch": 0.3103191273121485, "grad_norm": 8.10339069366455, "learning_rate": 9.475802587446096e-05, "loss": 1.1926, "step": 4580 }, { "epoch": 0.3103868825801206, "grad_norm": 7.225834369659424, "learning_rate": 9.475665685536314e-05, "loss": 1.1534, "step": 4581 }, { "epoch": 0.3104546378480927, "grad_norm": 7.356152057647705, "learning_rate": 9.475528783626532e-05, "loss": 1.1098, "step": 4582 }, { "epoch": 0.3105223931160648, "grad_norm": 10.347126960754395, "learning_rate": 9.475391881716751e-05, "loss": 1.136, "step": 4583 }, { "epoch": 0.31059014838403687, "grad_norm": 7.562332630157471, "learning_rate": 9.475254979806969e-05, "loss": 0.9211, "step": 4584 }, { "epoch": 0.31065790365200896, "grad_norm": 6.505655288696289, "learning_rate": 9.475118077897187e-05, "loss": 0.9538, "step": 4585 }, { "epoch": 0.31072565891998105, "grad_norm": 6.551809787750244, "learning_rate": 9.474981175987405e-05, "loss": 1.1335, "step": 4586 }, { "epoch": 0.3107934141879531, "grad_norm": 7.4724321365356445, "learning_rate": 9.474844274077623e-05, "loss": 1.0441, "step": 4587 }, { "epoch": 0.3108611694559252, "grad_norm": 8.463167190551758, "learning_rate": 9.474707372167843e-05, "loss": 1.0979, "step": 4588 }, { "epoch": 0.31092892472389727, "grad_norm": 6.868551731109619, "learning_rate": 9.47457047025806e-05, "loss": 0.6893, "step": 4589 }, { "epoch": 0.31099667999186936, "grad_norm": 5.831715106964111, "learning_rate": 9.474433568348279e-05, "loss": 0.9197, "step": 4590 }, { "epoch": 0.31106443525984145, "grad_norm": 7.562331676483154, "learning_rate": 9.474296666438497e-05, "loss": 0.8768, "step": 4591 }, { "epoch": 0.31113219052781355, "grad_norm": 7.750473499298096, "learning_rate": 9.474159764528716e-05, "loss": 0.8046, "step": 4592 }, { "epoch": 0.31119994579578564, "grad_norm": 7.836174964904785, "learning_rate": 9.474022862618934e-05, "loss": 0.9346, "step": 4593 }, { "epoch": 0.31126770106375773, "grad_norm": 6.159252166748047, "learning_rate": 9.473885960709152e-05, "loss": 0.8661, "step": 4594 }, { "epoch": 0.31133545633172977, "grad_norm": 7.959331512451172, "learning_rate": 9.47374905879937e-05, "loss": 0.919, "step": 4595 }, { "epoch": 0.31140321159970186, "grad_norm": 8.349802017211914, "learning_rate": 9.473612156889588e-05, "loss": 0.9762, "step": 4596 }, { "epoch": 0.31147096686767395, "grad_norm": 6.385254859924316, "learning_rate": 9.473475254979808e-05, "loss": 1.0168, "step": 4597 }, { "epoch": 0.31153872213564604, "grad_norm": 6.848907947540283, "learning_rate": 9.473338353070026e-05, "loss": 0.8235, "step": 4598 }, { "epoch": 0.31160647740361813, "grad_norm": 8.07978630065918, "learning_rate": 9.473201451160244e-05, "loss": 0.8394, "step": 4599 }, { "epoch": 0.3116742326715902, "grad_norm": 8.702666282653809, "learning_rate": 9.473064549250462e-05, "loss": 1.1859, "step": 4600 }, { "epoch": 0.3117419879395623, "grad_norm": 7.322251796722412, "learning_rate": 9.472927647340681e-05, "loss": 0.7089, "step": 4601 }, { "epoch": 0.3118097432075344, "grad_norm": 7.717091083526611, "learning_rate": 9.472790745430899e-05, "loss": 0.8609, "step": 4602 }, { "epoch": 0.3118774984755065, "grad_norm": 7.298448085784912, "learning_rate": 9.472653843521117e-05, "loss": 1.0044, "step": 4603 }, { "epoch": 0.31194525374347853, "grad_norm": 8.221037864685059, "learning_rate": 9.472516941611335e-05, "loss": 1.0403, "step": 4604 }, { "epoch": 0.3120130090114506, "grad_norm": 8.83371639251709, "learning_rate": 9.472380039701553e-05, "loss": 1.2884, "step": 4605 }, { "epoch": 0.3120807642794227, "grad_norm": 6.774711608886719, "learning_rate": 9.472243137791773e-05, "loss": 0.8077, "step": 4606 }, { "epoch": 0.3121485195473948, "grad_norm": 8.470376968383789, "learning_rate": 9.472106235881991e-05, "loss": 0.9549, "step": 4607 }, { "epoch": 0.3122162748153669, "grad_norm": 6.509616374969482, "learning_rate": 9.471969333972209e-05, "loss": 0.9595, "step": 4608 }, { "epoch": 0.312284030083339, "grad_norm": 6.885564804077148, "learning_rate": 9.471832432062427e-05, "loss": 0.9934, "step": 4609 }, { "epoch": 0.3123517853513111, "grad_norm": 12.500927925109863, "learning_rate": 9.471695530152646e-05, "loss": 0.9493, "step": 4610 }, { "epoch": 0.3124195406192832, "grad_norm": 8.701812744140625, "learning_rate": 9.471558628242864e-05, "loss": 0.9975, "step": 4611 }, { "epoch": 0.3124872958872552, "grad_norm": 8.313292503356934, "learning_rate": 9.471421726333082e-05, "loss": 1.0069, "step": 4612 }, { "epoch": 0.3125550511552273, "grad_norm": 6.841222763061523, "learning_rate": 9.471284824423302e-05, "loss": 0.8078, "step": 4613 }, { "epoch": 0.3126228064231994, "grad_norm": 9.038453102111816, "learning_rate": 9.47114792251352e-05, "loss": 1.0445, "step": 4614 }, { "epoch": 0.3126905616911715, "grad_norm": 7.772367000579834, "learning_rate": 9.471011020603738e-05, "loss": 1.0336, "step": 4615 }, { "epoch": 0.3127583169591436, "grad_norm": 6.843810558319092, "learning_rate": 9.470874118693957e-05, "loss": 0.8171, "step": 4616 }, { "epoch": 0.31282607222711567, "grad_norm": 7.012472629547119, "learning_rate": 9.470737216784175e-05, "loss": 0.8906, "step": 4617 }, { "epoch": 0.31289382749508776, "grad_norm": 7.138260364532471, "learning_rate": 9.470600314874393e-05, "loss": 0.9742, "step": 4618 }, { "epoch": 0.31296158276305985, "grad_norm": 7.92933988571167, "learning_rate": 9.470463412964611e-05, "loss": 0.8097, "step": 4619 }, { "epoch": 0.3130293380310319, "grad_norm": 8.222073554992676, "learning_rate": 9.47032651105483e-05, "loss": 0.9354, "step": 4620 }, { "epoch": 0.313097093299004, "grad_norm": 7.705626010894775, "learning_rate": 9.470189609145049e-05, "loss": 0.7776, "step": 4621 }, { "epoch": 0.3131648485669761, "grad_norm": 7.6136345863342285, "learning_rate": 9.470052707235267e-05, "loss": 1.0204, "step": 4622 }, { "epoch": 0.31323260383494816, "grad_norm": 6.365856647491455, "learning_rate": 9.469915805325485e-05, "loss": 0.8812, "step": 4623 }, { "epoch": 0.31330035910292026, "grad_norm": 6.876771926879883, "learning_rate": 9.469778903415704e-05, "loss": 0.7518, "step": 4624 }, { "epoch": 0.31336811437089235, "grad_norm": 6.329056262969971, "learning_rate": 9.469642001505922e-05, "loss": 0.8902, "step": 4625 }, { "epoch": 0.31343586963886444, "grad_norm": 6.1461310386657715, "learning_rate": 9.46950509959614e-05, "loss": 1.0617, "step": 4626 }, { "epoch": 0.31350362490683653, "grad_norm": 7.062450408935547, "learning_rate": 9.469368197686358e-05, "loss": 0.9081, "step": 4627 }, { "epoch": 0.31357138017480857, "grad_norm": 7.412983417510986, "learning_rate": 9.469231295776576e-05, "loss": 1.0845, "step": 4628 }, { "epoch": 0.31363913544278066, "grad_norm": 7.904543876647949, "learning_rate": 9.469094393866796e-05, "loss": 0.803, "step": 4629 }, { "epoch": 0.31370689071075275, "grad_norm": 6.614920616149902, "learning_rate": 9.468957491957014e-05, "loss": 0.8369, "step": 4630 }, { "epoch": 0.31377464597872484, "grad_norm": 8.245738983154297, "learning_rate": 9.468820590047232e-05, "loss": 0.9232, "step": 4631 }, { "epoch": 0.31384240124669693, "grad_norm": 7.923001766204834, "learning_rate": 9.46868368813745e-05, "loss": 0.822, "step": 4632 }, { "epoch": 0.313910156514669, "grad_norm": 8.80931282043457, "learning_rate": 9.468546786227668e-05, "loss": 0.9253, "step": 4633 }, { "epoch": 0.3139779117826411, "grad_norm": 6.997625350952148, "learning_rate": 9.468409884317887e-05, "loss": 0.8564, "step": 4634 }, { "epoch": 0.3140456670506132, "grad_norm": 7.699014186859131, "learning_rate": 9.468272982408105e-05, "loss": 1.0097, "step": 4635 }, { "epoch": 0.31411342231858524, "grad_norm": 7.379316329956055, "learning_rate": 9.468136080498323e-05, "loss": 0.8455, "step": 4636 }, { "epoch": 0.31418117758655734, "grad_norm": 7.040482521057129, "learning_rate": 9.467999178588541e-05, "loss": 0.9065, "step": 4637 }, { "epoch": 0.31424893285452943, "grad_norm": 8.588302612304688, "learning_rate": 9.46786227667876e-05, "loss": 0.8063, "step": 4638 }, { "epoch": 0.3143166881225015, "grad_norm": 8.310260772705078, "learning_rate": 9.467725374768979e-05, "loss": 0.9012, "step": 4639 }, { "epoch": 0.3143844433904736, "grad_norm": 5.794011116027832, "learning_rate": 9.467588472859197e-05, "loss": 0.7026, "step": 4640 }, { "epoch": 0.3144521986584457, "grad_norm": 8.277170181274414, "learning_rate": 9.467451570949415e-05, "loss": 0.8472, "step": 4641 }, { "epoch": 0.3145199539264178, "grad_norm": 8.856977462768555, "learning_rate": 9.467314669039633e-05, "loss": 1.1543, "step": 4642 }, { "epoch": 0.3145877091943899, "grad_norm": 7.343659400939941, "learning_rate": 9.467177767129852e-05, "loss": 0.8366, "step": 4643 }, { "epoch": 0.3146554644623619, "grad_norm": 6.619773864746094, "learning_rate": 9.46704086522007e-05, "loss": 0.9176, "step": 4644 }, { "epoch": 0.314723219730334, "grad_norm": 8.375277519226074, "learning_rate": 9.466903963310288e-05, "loss": 1.0437, "step": 4645 }, { "epoch": 0.3147909749983061, "grad_norm": 6.792671203613281, "learning_rate": 9.466767061400506e-05, "loss": 0.7271, "step": 4646 }, { "epoch": 0.3148587302662782, "grad_norm": 7.419206619262695, "learning_rate": 9.466630159490726e-05, "loss": 0.9376, "step": 4647 }, { "epoch": 0.3149264855342503, "grad_norm": 6.413398742675781, "learning_rate": 9.466493257580944e-05, "loss": 0.8988, "step": 4648 }, { "epoch": 0.3149942408022224, "grad_norm": 7.8477373123168945, "learning_rate": 9.466356355671162e-05, "loss": 1.0791, "step": 4649 }, { "epoch": 0.31506199607019447, "grad_norm": 7.070709705352783, "learning_rate": 9.46621945376138e-05, "loss": 0.7796, "step": 4650 }, { "epoch": 0.31512975133816656, "grad_norm": 7.023820400238037, "learning_rate": 9.466082551851598e-05, "loss": 0.928, "step": 4651 }, { "epoch": 0.3151975066061386, "grad_norm": 8.347658157348633, "learning_rate": 9.465945649941817e-05, "loss": 0.8768, "step": 4652 }, { "epoch": 0.3152652618741107, "grad_norm": 7.446930408477783, "learning_rate": 9.465808748032035e-05, "loss": 0.7215, "step": 4653 }, { "epoch": 0.3153330171420828, "grad_norm": 9.261749267578125, "learning_rate": 9.465671846122253e-05, "loss": 1.0612, "step": 4654 }, { "epoch": 0.3154007724100549, "grad_norm": 6.765881061553955, "learning_rate": 9.465534944212471e-05, "loss": 0.9636, "step": 4655 }, { "epoch": 0.31546852767802697, "grad_norm": 7.752283096313477, "learning_rate": 9.46539804230269e-05, "loss": 1.0348, "step": 4656 }, { "epoch": 0.31553628294599906, "grad_norm": 8.769716262817383, "learning_rate": 9.465261140392909e-05, "loss": 0.9147, "step": 4657 }, { "epoch": 0.31560403821397115, "grad_norm": 9.27787971496582, "learning_rate": 9.465124238483127e-05, "loss": 1.0287, "step": 4658 }, { "epoch": 0.31567179348194324, "grad_norm": 5.802475452423096, "learning_rate": 9.464987336573346e-05, "loss": 0.7968, "step": 4659 }, { "epoch": 0.31573954874991533, "grad_norm": 7.436465263366699, "learning_rate": 9.464850434663564e-05, "loss": 1.1087, "step": 4660 }, { "epoch": 0.31580730401788737, "grad_norm": 8.243688583374023, "learning_rate": 9.464713532753782e-05, "loss": 1.0257, "step": 4661 }, { "epoch": 0.31587505928585946, "grad_norm": 7.885406017303467, "learning_rate": 9.464576630844002e-05, "loss": 0.9944, "step": 4662 }, { "epoch": 0.31594281455383155, "grad_norm": 8.342294692993164, "learning_rate": 9.46443972893422e-05, "loss": 0.8351, "step": 4663 }, { "epoch": 0.31601056982180364, "grad_norm": 7.476551055908203, "learning_rate": 9.464302827024438e-05, "loss": 0.8484, "step": 4664 }, { "epoch": 0.31607832508977574, "grad_norm": 7.239434719085693, "learning_rate": 9.464165925114656e-05, "loss": 0.749, "step": 4665 }, { "epoch": 0.3161460803577478, "grad_norm": 8.02225399017334, "learning_rate": 9.464029023204875e-05, "loss": 1.0679, "step": 4666 }, { "epoch": 0.3162138356257199, "grad_norm": 5.708010673522949, "learning_rate": 9.463892121295093e-05, "loss": 0.6935, "step": 4667 }, { "epoch": 0.316281590893692, "grad_norm": 9.107060432434082, "learning_rate": 9.463755219385311e-05, "loss": 0.7951, "step": 4668 }, { "epoch": 0.31634934616166405, "grad_norm": 8.045939445495605, "learning_rate": 9.463618317475529e-05, "loss": 0.9943, "step": 4669 }, { "epoch": 0.31641710142963614, "grad_norm": 8.284626007080078, "learning_rate": 9.463481415565749e-05, "loss": 0.9638, "step": 4670 }, { "epoch": 0.31648485669760823, "grad_norm": 6.9800920486450195, "learning_rate": 9.463344513655967e-05, "loss": 0.9211, "step": 4671 }, { "epoch": 0.3165526119655803, "grad_norm": 7.476776599884033, "learning_rate": 9.463207611746185e-05, "loss": 0.8704, "step": 4672 }, { "epoch": 0.3166203672335524, "grad_norm": 6.303709506988525, "learning_rate": 9.463070709836403e-05, "loss": 0.9031, "step": 4673 }, { "epoch": 0.3166881225015245, "grad_norm": 8.267644882202148, "learning_rate": 9.46293380792662e-05, "loss": 1.0899, "step": 4674 }, { "epoch": 0.3167558777694966, "grad_norm": 8.25999641418457, "learning_rate": 9.46279690601684e-05, "loss": 0.9286, "step": 4675 }, { "epoch": 0.3168236330374687, "grad_norm": 8.577073097229004, "learning_rate": 9.462660004107058e-05, "loss": 0.8837, "step": 4676 }, { "epoch": 0.3168913883054407, "grad_norm": 7.183437824249268, "learning_rate": 9.462523102197276e-05, "loss": 0.9972, "step": 4677 }, { "epoch": 0.3169591435734128, "grad_norm": 7.93414831161499, "learning_rate": 9.462386200287494e-05, "loss": 0.9806, "step": 4678 }, { "epoch": 0.3170268988413849, "grad_norm": 8.037418365478516, "learning_rate": 9.462249298377714e-05, "loss": 0.8945, "step": 4679 }, { "epoch": 0.317094654109357, "grad_norm": 7.5245585441589355, "learning_rate": 9.462112396467932e-05, "loss": 0.9444, "step": 4680 }, { "epoch": 0.3171624093773291, "grad_norm": 7.428105354309082, "learning_rate": 9.46197549455815e-05, "loss": 0.9916, "step": 4681 }, { "epoch": 0.3172301646453012, "grad_norm": 7.977080345153809, "learning_rate": 9.461838592648368e-05, "loss": 1.2206, "step": 4682 }, { "epoch": 0.3172979199132733, "grad_norm": 6.285130023956299, "learning_rate": 9.461701690738586e-05, "loss": 0.8222, "step": 4683 }, { "epoch": 0.31736567518124537, "grad_norm": 6.9027018547058105, "learning_rate": 9.461564788828805e-05, "loss": 1.0265, "step": 4684 }, { "epoch": 0.3174334304492174, "grad_norm": 7.620064735412598, "learning_rate": 9.461427886919023e-05, "loss": 0.9672, "step": 4685 }, { "epoch": 0.3175011857171895, "grad_norm": 7.635505676269531, "learning_rate": 9.461290985009241e-05, "loss": 0.9329, "step": 4686 }, { "epoch": 0.3175689409851616, "grad_norm": 7.532449245452881, "learning_rate": 9.461154083099459e-05, "loss": 1.0509, "step": 4687 }, { "epoch": 0.3176366962531337, "grad_norm": 6.428747177124023, "learning_rate": 9.461017181189677e-05, "loss": 0.9151, "step": 4688 }, { "epoch": 0.31770445152110577, "grad_norm": 7.214437007904053, "learning_rate": 9.460880279279897e-05, "loss": 0.9758, "step": 4689 }, { "epoch": 0.31777220678907786, "grad_norm": 7.525144577026367, "learning_rate": 9.460743377370115e-05, "loss": 0.7723, "step": 4690 }, { "epoch": 0.31783996205704995, "grad_norm": 8.008615493774414, "learning_rate": 9.460606475460333e-05, "loss": 0.9945, "step": 4691 }, { "epoch": 0.31790771732502204, "grad_norm": 6.715225696563721, "learning_rate": 9.460469573550551e-05, "loss": 0.9805, "step": 4692 }, { "epoch": 0.3179754725929941, "grad_norm": 7.146914005279541, "learning_rate": 9.46033267164077e-05, "loss": 0.9583, "step": 4693 }, { "epoch": 0.31804322786096617, "grad_norm": 7.280580997467041, "learning_rate": 9.460195769730988e-05, "loss": 0.9772, "step": 4694 }, { "epoch": 0.31811098312893826, "grad_norm": 7.84500789642334, "learning_rate": 9.460058867821206e-05, "loss": 0.9755, "step": 4695 }, { "epoch": 0.31817873839691035, "grad_norm": 7.289769649505615, "learning_rate": 9.459921965911424e-05, "loss": 1.1441, "step": 4696 }, { "epoch": 0.31824649366488245, "grad_norm": 9.061880111694336, "learning_rate": 9.459785064001642e-05, "loss": 0.9699, "step": 4697 }, { "epoch": 0.31831424893285454, "grad_norm": 7.284442901611328, "learning_rate": 9.459648162091862e-05, "loss": 1.0023, "step": 4698 }, { "epoch": 0.31838200420082663, "grad_norm": 7.017299175262451, "learning_rate": 9.45951126018208e-05, "loss": 0.8598, "step": 4699 }, { "epoch": 0.3184497594687987, "grad_norm": 7.236936092376709, "learning_rate": 9.459374358272298e-05, "loss": 0.9888, "step": 4700 }, { "epoch": 0.31851751473677076, "grad_norm": 8.148765563964844, "learning_rate": 9.459237456362516e-05, "loss": 0.9978, "step": 4701 }, { "epoch": 0.31858527000474285, "grad_norm": 8.023640632629395, "learning_rate": 9.459100554452735e-05, "loss": 1.0941, "step": 4702 }, { "epoch": 0.31865302527271494, "grad_norm": 6.2704291343688965, "learning_rate": 9.458963652542953e-05, "loss": 0.8616, "step": 4703 }, { "epoch": 0.31872078054068703, "grad_norm": 7.873881816864014, "learning_rate": 9.458826750633171e-05, "loss": 0.9212, "step": 4704 }, { "epoch": 0.3187885358086591, "grad_norm": 7.088031768798828, "learning_rate": 9.45868984872339e-05, "loss": 0.6453, "step": 4705 }, { "epoch": 0.3188562910766312, "grad_norm": 7.319702625274658, "learning_rate": 9.458552946813609e-05, "loss": 0.8328, "step": 4706 }, { "epoch": 0.3189240463446033, "grad_norm": 6.979453086853027, "learning_rate": 9.458416044903827e-05, "loss": 0.7806, "step": 4707 }, { "epoch": 0.3189918016125754, "grad_norm": 7.868389129638672, "learning_rate": 9.458279142994046e-05, "loss": 0.8052, "step": 4708 }, { "epoch": 0.3190595568805475, "grad_norm": 7.2435431480407715, "learning_rate": 9.458142241084264e-05, "loss": 0.8922, "step": 4709 }, { "epoch": 0.3191273121485195, "grad_norm": 8.62594985961914, "learning_rate": 9.458005339174482e-05, "loss": 1.0572, "step": 4710 }, { "epoch": 0.3191950674164916, "grad_norm": 8.404034614562988, "learning_rate": 9.457868437264701e-05, "loss": 1.1592, "step": 4711 }, { "epoch": 0.3192628226844637, "grad_norm": 6.669247150421143, "learning_rate": 9.45773153535492e-05, "loss": 0.8727, "step": 4712 }, { "epoch": 0.3193305779524358, "grad_norm": 7.565460681915283, "learning_rate": 9.457594633445138e-05, "loss": 0.7803, "step": 4713 }, { "epoch": 0.3193983332204079, "grad_norm": 7.726403713226318, "learning_rate": 9.457457731535356e-05, "loss": 0.81, "step": 4714 }, { "epoch": 0.31946608848838, "grad_norm": 7.344959259033203, "learning_rate": 9.457320829625574e-05, "loss": 0.6959, "step": 4715 }, { "epoch": 0.3195338437563521, "grad_norm": 5.9748759269714355, "learning_rate": 9.457183927715793e-05, "loss": 0.8332, "step": 4716 }, { "epoch": 0.31960159902432417, "grad_norm": 7.241995334625244, "learning_rate": 9.457047025806011e-05, "loss": 0.8243, "step": 4717 }, { "epoch": 0.3196693542922962, "grad_norm": 7.812702655792236, "learning_rate": 9.456910123896229e-05, "loss": 0.8603, "step": 4718 }, { "epoch": 0.3197371095602683, "grad_norm": 9.492036819458008, "learning_rate": 9.456773221986447e-05, "loss": 1.0208, "step": 4719 }, { "epoch": 0.3198048648282404, "grad_norm": 8.922654151916504, "learning_rate": 9.456636320076665e-05, "loss": 1.1939, "step": 4720 }, { "epoch": 0.3198726200962125, "grad_norm": 6.459314346313477, "learning_rate": 9.456499418166885e-05, "loss": 1.0332, "step": 4721 }, { "epoch": 0.31994037536418457, "grad_norm": 7.104556560516357, "learning_rate": 9.456362516257103e-05, "loss": 1.0608, "step": 4722 }, { "epoch": 0.32000813063215666, "grad_norm": 7.620473861694336, "learning_rate": 9.45622561434732e-05, "loss": 0.888, "step": 4723 }, { "epoch": 0.32007588590012875, "grad_norm": 7.757092475891113, "learning_rate": 9.456088712437539e-05, "loss": 1.0355, "step": 4724 }, { "epoch": 0.32014364116810085, "grad_norm": 7.084576606750488, "learning_rate": 9.455951810527758e-05, "loss": 0.7148, "step": 4725 }, { "epoch": 0.3202113964360729, "grad_norm": 6.5413079261779785, "learning_rate": 9.455814908617976e-05, "loss": 0.8678, "step": 4726 }, { "epoch": 0.320279151704045, "grad_norm": 8.085969924926758, "learning_rate": 9.455678006708194e-05, "loss": 1.0137, "step": 4727 }, { "epoch": 0.32034690697201706, "grad_norm": 6.338340759277344, "learning_rate": 9.455541104798412e-05, "loss": 0.9573, "step": 4728 }, { "epoch": 0.32041466223998916, "grad_norm": 8.724467277526855, "learning_rate": 9.45540420288863e-05, "loss": 1.0585, "step": 4729 }, { "epoch": 0.32048241750796125, "grad_norm": 5.904287815093994, "learning_rate": 9.45526730097885e-05, "loss": 0.6882, "step": 4730 }, { "epoch": 0.32055017277593334, "grad_norm": 6.975876331329346, "learning_rate": 9.455130399069068e-05, "loss": 0.8641, "step": 4731 }, { "epoch": 0.32061792804390543, "grad_norm": 7.307252407073975, "learning_rate": 9.454993497159286e-05, "loss": 0.8961, "step": 4732 }, { "epoch": 0.3206856833118775, "grad_norm": 6.06977653503418, "learning_rate": 9.454856595249504e-05, "loss": 0.8388, "step": 4733 }, { "epoch": 0.32075343857984956, "grad_norm": 5.903672218322754, "learning_rate": 9.454719693339723e-05, "loss": 0.6962, "step": 4734 }, { "epoch": 0.32082119384782165, "grad_norm": 9.061569213867188, "learning_rate": 9.454582791429941e-05, "loss": 1.0246, "step": 4735 }, { "epoch": 0.32088894911579374, "grad_norm": 9.154926300048828, "learning_rate": 9.454445889520159e-05, "loss": 0.9336, "step": 4736 }, { "epoch": 0.32095670438376583, "grad_norm": 6.51992654800415, "learning_rate": 9.454308987610377e-05, "loss": 0.8753, "step": 4737 }, { "epoch": 0.3210244596517379, "grad_norm": 7.160184383392334, "learning_rate": 9.454172085700595e-05, "loss": 1.0036, "step": 4738 }, { "epoch": 0.32109221491971, "grad_norm": 6.895291805267334, "learning_rate": 9.454035183790815e-05, "loss": 0.8359, "step": 4739 }, { "epoch": 0.3211599701876821, "grad_norm": 7.363986015319824, "learning_rate": 9.453898281881033e-05, "loss": 1.0207, "step": 4740 }, { "epoch": 0.3212277254556542, "grad_norm": 9.040234565734863, "learning_rate": 9.45376137997125e-05, "loss": 1.0123, "step": 4741 }, { "epoch": 0.32129548072362624, "grad_norm": 7.679563045501709, "learning_rate": 9.453624478061469e-05, "loss": 1.1201, "step": 4742 }, { "epoch": 0.32136323599159833, "grad_norm": 7.640948295593262, "learning_rate": 9.453487576151687e-05, "loss": 1.1834, "step": 4743 }, { "epoch": 0.3214309912595704, "grad_norm": 6.348153591156006, "learning_rate": 9.453350674241906e-05, "loss": 0.735, "step": 4744 }, { "epoch": 0.3214987465275425, "grad_norm": 6.899050712585449, "learning_rate": 9.453213772332124e-05, "loss": 0.9054, "step": 4745 }, { "epoch": 0.3215665017955146, "grad_norm": 8.880489349365234, "learning_rate": 9.453076870422342e-05, "loss": 1.0253, "step": 4746 }, { "epoch": 0.3216342570634867, "grad_norm": 7.2754597663879395, "learning_rate": 9.45293996851256e-05, "loss": 0.75, "step": 4747 }, { "epoch": 0.3217020123314588, "grad_norm": 7.26841402053833, "learning_rate": 9.45280306660278e-05, "loss": 1.2198, "step": 4748 }, { "epoch": 0.3217697675994309, "grad_norm": 6.0724310874938965, "learning_rate": 9.452666164692998e-05, "loss": 0.7444, "step": 4749 }, { "epoch": 0.3218375228674029, "grad_norm": 7.82569694519043, "learning_rate": 9.452529262783216e-05, "loss": 0.9266, "step": 4750 }, { "epoch": 0.321905278135375, "grad_norm": 7.153421878814697, "learning_rate": 9.452392360873435e-05, "loss": 1.2223, "step": 4751 }, { "epoch": 0.3219730334033471, "grad_norm": 6.586106777191162, "learning_rate": 9.452255458963653e-05, "loss": 0.9157, "step": 4752 }, { "epoch": 0.3220407886713192, "grad_norm": 7.776567459106445, "learning_rate": 9.452118557053871e-05, "loss": 0.7396, "step": 4753 }, { "epoch": 0.3221085439392913, "grad_norm": 7.4870381355285645, "learning_rate": 9.45198165514409e-05, "loss": 1.0812, "step": 4754 }, { "epoch": 0.3221762992072634, "grad_norm": 8.112491607666016, "learning_rate": 9.451844753234309e-05, "loss": 1.2816, "step": 4755 }, { "epoch": 0.32224405447523546, "grad_norm": 8.562600135803223, "learning_rate": 9.451707851324527e-05, "loss": 0.9872, "step": 4756 }, { "epoch": 0.32231180974320756, "grad_norm": 9.139601707458496, "learning_rate": 9.451570949414746e-05, "loss": 1.019, "step": 4757 }, { "epoch": 0.3223795650111796, "grad_norm": 7.6649370193481445, "learning_rate": 9.451434047504964e-05, "loss": 1.1594, "step": 4758 }, { "epoch": 0.3224473202791517, "grad_norm": 8.193527221679688, "learning_rate": 9.451297145595182e-05, "loss": 0.8804, "step": 4759 }, { "epoch": 0.3225150755471238, "grad_norm": 5.558340072631836, "learning_rate": 9.4511602436854e-05, "loss": 0.8778, "step": 4760 }, { "epoch": 0.32258283081509587, "grad_norm": 7.296480178833008, "learning_rate": 9.451023341775618e-05, "loss": 1.0039, "step": 4761 }, { "epoch": 0.32265058608306796, "grad_norm": 8.916117668151855, "learning_rate": 9.450886439865837e-05, "loss": 0.9516, "step": 4762 }, { "epoch": 0.32271834135104005, "grad_norm": 7.691675662994385, "learning_rate": 9.450749537956056e-05, "loss": 1.1872, "step": 4763 }, { "epoch": 0.32278609661901214, "grad_norm": 6.425968647003174, "learning_rate": 9.450612636046274e-05, "loss": 0.7914, "step": 4764 }, { "epoch": 0.32285385188698423, "grad_norm": 6.476465225219727, "learning_rate": 9.450475734136492e-05, "loss": 0.9449, "step": 4765 }, { "epoch": 0.3229216071549563, "grad_norm": 7.470714569091797, "learning_rate": 9.450338832226711e-05, "loss": 0.8132, "step": 4766 }, { "epoch": 0.32298936242292836, "grad_norm": 6.947244644165039, "learning_rate": 9.450201930316929e-05, "loss": 1.0203, "step": 4767 }, { "epoch": 0.32305711769090045, "grad_norm": 6.939558982849121, "learning_rate": 9.450065028407147e-05, "loss": 0.7484, "step": 4768 }, { "epoch": 0.32312487295887254, "grad_norm": 6.562455177307129, "learning_rate": 9.449928126497365e-05, "loss": 0.9659, "step": 4769 }, { "epoch": 0.32319262822684464, "grad_norm": 7.884627819061279, "learning_rate": 9.449791224587583e-05, "loss": 0.9394, "step": 4770 }, { "epoch": 0.3232603834948167, "grad_norm": 8.175631523132324, "learning_rate": 9.449654322677802e-05, "loss": 1.3089, "step": 4771 }, { "epoch": 0.3233281387627888, "grad_norm": 7.907345294952393, "learning_rate": 9.44951742076802e-05, "loss": 0.7641, "step": 4772 }, { "epoch": 0.3233958940307609, "grad_norm": 8.59745979309082, "learning_rate": 9.449380518858239e-05, "loss": 1.1506, "step": 4773 }, { "epoch": 0.323463649298733, "grad_norm": 8.216800689697266, "learning_rate": 9.449243616948457e-05, "loss": 1.1007, "step": 4774 }, { "epoch": 0.32353140456670504, "grad_norm": 7.630847930908203, "learning_rate": 9.449106715038675e-05, "loss": 0.6435, "step": 4775 }, { "epoch": 0.32359915983467713, "grad_norm": 8.40709400177002, "learning_rate": 9.448969813128894e-05, "loss": 1.0299, "step": 4776 }, { "epoch": 0.3236669151026492, "grad_norm": 7.3637166023254395, "learning_rate": 9.448832911219112e-05, "loss": 1.2733, "step": 4777 }, { "epoch": 0.3237346703706213, "grad_norm": 7.707301139831543, "learning_rate": 9.44869600930933e-05, "loss": 0.91, "step": 4778 }, { "epoch": 0.3238024256385934, "grad_norm": 6.911712169647217, "learning_rate": 9.448559107399548e-05, "loss": 1.0425, "step": 4779 }, { "epoch": 0.3238701809065655, "grad_norm": 7.387215614318848, "learning_rate": 9.448422205489768e-05, "loss": 1.1508, "step": 4780 }, { "epoch": 0.3239379361745376, "grad_norm": 6.074267864227295, "learning_rate": 9.448285303579986e-05, "loss": 0.7861, "step": 4781 }, { "epoch": 0.3240056914425097, "grad_norm": 7.4029436111450195, "learning_rate": 9.448148401670204e-05, "loss": 1.1438, "step": 4782 }, { "epoch": 0.3240734467104817, "grad_norm": 8.514384269714355, "learning_rate": 9.448011499760422e-05, "loss": 1.1075, "step": 4783 }, { "epoch": 0.3241412019784538, "grad_norm": 8.503793716430664, "learning_rate": 9.44787459785064e-05, "loss": 0.9377, "step": 4784 }, { "epoch": 0.3242089572464259, "grad_norm": 6.770750522613525, "learning_rate": 9.447737695940859e-05, "loss": 0.8342, "step": 4785 }, { "epoch": 0.324276712514398, "grad_norm": 7.019729137420654, "learning_rate": 9.447600794031077e-05, "loss": 0.8236, "step": 4786 }, { "epoch": 0.3243444677823701, "grad_norm": 6.147645950317383, "learning_rate": 9.447463892121295e-05, "loss": 0.9493, "step": 4787 }, { "epoch": 0.3244122230503422, "grad_norm": 9.160635948181152, "learning_rate": 9.447326990211513e-05, "loss": 1.1632, "step": 4788 }, { "epoch": 0.32447997831831427, "grad_norm": 7.918509483337402, "learning_rate": 9.447190088301733e-05, "loss": 1.0154, "step": 4789 }, { "epoch": 0.32454773358628636, "grad_norm": 8.457235336303711, "learning_rate": 9.44705318639195e-05, "loss": 1.0373, "step": 4790 }, { "epoch": 0.3246154888542584, "grad_norm": 7.332309722900391, "learning_rate": 9.446916284482169e-05, "loss": 1.09, "step": 4791 }, { "epoch": 0.3246832441222305, "grad_norm": 7.7085700035095215, "learning_rate": 9.446779382572387e-05, "loss": 0.8492, "step": 4792 }, { "epoch": 0.3247509993902026, "grad_norm": 6.121610641479492, "learning_rate": 9.446642480662605e-05, "loss": 1.0399, "step": 4793 }, { "epoch": 0.32481875465817467, "grad_norm": 6.767165184020996, "learning_rate": 9.446505578752824e-05, "loss": 0.6908, "step": 4794 }, { "epoch": 0.32488650992614676, "grad_norm": 7.629088401794434, "learning_rate": 9.446368676843042e-05, "loss": 0.9356, "step": 4795 }, { "epoch": 0.32495426519411885, "grad_norm": 7.590803146362305, "learning_rate": 9.44623177493326e-05, "loss": 0.7671, "step": 4796 }, { "epoch": 0.32502202046209094, "grad_norm": 7.876105785369873, "learning_rate": 9.44609487302348e-05, "loss": 0.9654, "step": 4797 }, { "epoch": 0.32508977573006304, "grad_norm": 8.503900527954102, "learning_rate": 9.445957971113698e-05, "loss": 0.9726, "step": 4798 }, { "epoch": 0.32515753099803507, "grad_norm": 9.96375846862793, "learning_rate": 9.445821069203916e-05, "loss": 0.7587, "step": 4799 }, { "epoch": 0.32522528626600716, "grad_norm": 6.24782133102417, "learning_rate": 9.445684167294135e-05, "loss": 0.8028, "step": 4800 }, { "epoch": 0.32529304153397925, "grad_norm": 7.5481181144714355, "learning_rate": 9.445547265384353e-05, "loss": 0.8194, "step": 4801 }, { "epoch": 0.32536079680195135, "grad_norm": 8.148533821105957, "learning_rate": 9.445410363474571e-05, "loss": 1.0009, "step": 4802 }, { "epoch": 0.32542855206992344, "grad_norm": 7.8531341552734375, "learning_rate": 9.44527346156479e-05, "loss": 1.0629, "step": 4803 }, { "epoch": 0.32549630733789553, "grad_norm": 7.7668843269348145, "learning_rate": 9.445136559655008e-05, "loss": 1.1211, "step": 4804 }, { "epoch": 0.3255640626058676, "grad_norm": 9.245609283447266, "learning_rate": 9.444999657745226e-05, "loss": 1.0407, "step": 4805 }, { "epoch": 0.3256318178738397, "grad_norm": 7.387469291687012, "learning_rate": 9.444862755835445e-05, "loss": 1.128, "step": 4806 }, { "epoch": 0.32569957314181175, "grad_norm": 6.636757850646973, "learning_rate": 9.444725853925663e-05, "loss": 0.8386, "step": 4807 }, { "epoch": 0.32576732840978384, "grad_norm": 8.264979362487793, "learning_rate": 9.444588952015882e-05, "loss": 0.904, "step": 4808 }, { "epoch": 0.32583508367775593, "grad_norm": 7.77110481262207, "learning_rate": 9.4444520501061e-05, "loss": 1.0292, "step": 4809 }, { "epoch": 0.325902838945728, "grad_norm": 6.825851917266846, "learning_rate": 9.444315148196318e-05, "loss": 0.7484, "step": 4810 }, { "epoch": 0.3259705942137001, "grad_norm": 6.475069999694824, "learning_rate": 9.444178246286536e-05, "loss": 0.9085, "step": 4811 }, { "epoch": 0.3260383494816722, "grad_norm": 8.736098289489746, "learning_rate": 9.444041344376755e-05, "loss": 1.1817, "step": 4812 }, { "epoch": 0.3261061047496443, "grad_norm": 7.0992608070373535, "learning_rate": 9.443904442466973e-05, "loss": 0.8654, "step": 4813 }, { "epoch": 0.3261738600176164, "grad_norm": 6.594883441925049, "learning_rate": 9.443767540557192e-05, "loss": 1.0002, "step": 4814 }, { "epoch": 0.3262416152855885, "grad_norm": 7.857585906982422, "learning_rate": 9.44363063864741e-05, "loss": 1.1711, "step": 4815 }, { "epoch": 0.3263093705535605, "grad_norm": 6.884295463562012, "learning_rate": 9.443493736737628e-05, "loss": 0.8003, "step": 4816 }, { "epoch": 0.3263771258215326, "grad_norm": 7.344529628753662, "learning_rate": 9.443356834827847e-05, "loss": 1.0009, "step": 4817 }, { "epoch": 0.3264448810895047, "grad_norm": 6.892088413238525, "learning_rate": 9.443219932918065e-05, "loss": 0.9797, "step": 4818 }, { "epoch": 0.3265126363574768, "grad_norm": 12.833809852600098, "learning_rate": 9.443083031008283e-05, "loss": 1.1143, "step": 4819 }, { "epoch": 0.3265803916254489, "grad_norm": 7.343682765960693, "learning_rate": 9.442946129098501e-05, "loss": 0.6302, "step": 4820 }, { "epoch": 0.326648146893421, "grad_norm": 7.089320659637451, "learning_rate": 9.442809227188719e-05, "loss": 1.0129, "step": 4821 }, { "epoch": 0.32671590216139307, "grad_norm": 6.60029935836792, "learning_rate": 9.442672325278938e-05, "loss": 0.9259, "step": 4822 }, { "epoch": 0.32678365742936516, "grad_norm": 10.470613479614258, "learning_rate": 9.442535423369157e-05, "loss": 1.1124, "step": 4823 }, { "epoch": 0.3268514126973372, "grad_norm": 8.491450309753418, "learning_rate": 9.442398521459375e-05, "loss": 0.9943, "step": 4824 }, { "epoch": 0.3269191679653093, "grad_norm": 6.033353328704834, "learning_rate": 9.442261619549593e-05, "loss": 1.0003, "step": 4825 }, { "epoch": 0.3269869232332814, "grad_norm": 7.0092549324035645, "learning_rate": 9.442124717639812e-05, "loss": 0.7843, "step": 4826 }, { "epoch": 0.32705467850125347, "grad_norm": 8.52950382232666, "learning_rate": 9.44198781573003e-05, "loss": 1.1036, "step": 4827 }, { "epoch": 0.32712243376922556, "grad_norm": 7.509829044342041, "learning_rate": 9.441850913820248e-05, "loss": 0.9969, "step": 4828 }, { "epoch": 0.32719018903719765, "grad_norm": 7.235622406005859, "learning_rate": 9.441714011910466e-05, "loss": 0.9131, "step": 4829 }, { "epoch": 0.32725794430516975, "grad_norm": 8.332746505737305, "learning_rate": 9.441577110000684e-05, "loss": 1.1663, "step": 4830 }, { "epoch": 0.32732569957314184, "grad_norm": 8.321161270141602, "learning_rate": 9.441440208090904e-05, "loss": 1.0091, "step": 4831 }, { "epoch": 0.3273934548411139, "grad_norm": 7.0502753257751465, "learning_rate": 9.441303306181122e-05, "loss": 1.0273, "step": 4832 }, { "epoch": 0.32746121010908597, "grad_norm": 6.921250343322754, "learning_rate": 9.44116640427134e-05, "loss": 0.873, "step": 4833 }, { "epoch": 0.32752896537705806, "grad_norm": 7.386787414550781, "learning_rate": 9.441029502361558e-05, "loss": 0.7415, "step": 4834 }, { "epoch": 0.32759672064503015, "grad_norm": 6.814700603485107, "learning_rate": 9.440892600451777e-05, "loss": 0.8827, "step": 4835 }, { "epoch": 0.32766447591300224, "grad_norm": 7.551968574523926, "learning_rate": 9.440755698541995e-05, "loss": 0.9752, "step": 4836 }, { "epoch": 0.32773223118097433, "grad_norm": 8.299920082092285, "learning_rate": 9.440618796632213e-05, "loss": 1.0253, "step": 4837 }, { "epoch": 0.3277999864489464, "grad_norm": 7.607963562011719, "learning_rate": 9.440481894722431e-05, "loss": 1.0562, "step": 4838 }, { "epoch": 0.3278677417169185, "grad_norm": 8.098003387451172, "learning_rate": 9.440344992812649e-05, "loss": 0.9055, "step": 4839 }, { "epoch": 0.32793549698489055, "grad_norm": 6.6647257804870605, "learning_rate": 9.440208090902869e-05, "loss": 0.8188, "step": 4840 }, { "epoch": 0.32800325225286264, "grad_norm": 7.541131496429443, "learning_rate": 9.440071188993087e-05, "loss": 0.7095, "step": 4841 }, { "epoch": 0.32807100752083473, "grad_norm": 10.203042984008789, "learning_rate": 9.439934287083305e-05, "loss": 0.8256, "step": 4842 }, { "epoch": 0.3281387627888068, "grad_norm": 8.049646377563477, "learning_rate": 9.439797385173524e-05, "loss": 1.0244, "step": 4843 }, { "epoch": 0.3282065180567789, "grad_norm": 6.727071285247803, "learning_rate": 9.439660483263742e-05, "loss": 0.9681, "step": 4844 }, { "epoch": 0.328274273324751, "grad_norm": 8.106125831604004, "learning_rate": 9.43952358135396e-05, "loss": 1.0944, "step": 4845 }, { "epoch": 0.3283420285927231, "grad_norm": 8.768218994140625, "learning_rate": 9.43938667944418e-05, "loss": 0.7493, "step": 4846 }, { "epoch": 0.3284097838606952, "grad_norm": 6.333378791809082, "learning_rate": 9.439249777534397e-05, "loss": 0.9187, "step": 4847 }, { "epoch": 0.32847753912866723, "grad_norm": 7.732221603393555, "learning_rate": 9.439112875624616e-05, "loss": 1.2522, "step": 4848 }, { "epoch": 0.3285452943966393, "grad_norm": 7.67783784866333, "learning_rate": 9.438975973714835e-05, "loss": 0.8682, "step": 4849 }, { "epoch": 0.3286130496646114, "grad_norm": 7.557129859924316, "learning_rate": 9.438839071805053e-05, "loss": 0.7843, "step": 4850 }, { "epoch": 0.3286808049325835, "grad_norm": 6.559933185577393, "learning_rate": 9.438702169895271e-05, "loss": 0.902, "step": 4851 }, { "epoch": 0.3287485602005556, "grad_norm": 7.604280471801758, "learning_rate": 9.438565267985489e-05, "loss": 0.8905, "step": 4852 }, { "epoch": 0.3288163154685277, "grad_norm": 7.508764266967773, "learning_rate": 9.438428366075707e-05, "loss": 0.932, "step": 4853 }, { "epoch": 0.3288840707364998, "grad_norm": 6.650167942047119, "learning_rate": 9.438291464165926e-05, "loss": 1.0507, "step": 4854 }, { "epoch": 0.32895182600447187, "grad_norm": 8.334061622619629, "learning_rate": 9.438154562256144e-05, "loss": 0.7949, "step": 4855 }, { "epoch": 0.3290195812724439, "grad_norm": 9.498878479003906, "learning_rate": 9.438017660346362e-05, "loss": 0.8487, "step": 4856 }, { "epoch": 0.329087336540416, "grad_norm": 7.810210227966309, "learning_rate": 9.43788075843658e-05, "loss": 0.9917, "step": 4857 }, { "epoch": 0.3291550918083881, "grad_norm": 7.337824821472168, "learning_rate": 9.4377438565268e-05, "loss": 0.8683, "step": 4858 }, { "epoch": 0.3292228470763602, "grad_norm": 8.331491470336914, "learning_rate": 9.437606954617018e-05, "loss": 1.178, "step": 4859 }, { "epoch": 0.3292906023443323, "grad_norm": 6.5013227462768555, "learning_rate": 9.437470052707236e-05, "loss": 0.7093, "step": 4860 }, { "epoch": 0.32935835761230436, "grad_norm": 8.814985275268555, "learning_rate": 9.437333150797454e-05, "loss": 0.9212, "step": 4861 }, { "epoch": 0.32942611288027646, "grad_norm": 8.78661060333252, "learning_rate": 9.437196248887672e-05, "loss": 0.8827, "step": 4862 }, { "epoch": 0.32949386814824855, "grad_norm": 8.912193298339844, "learning_rate": 9.437059346977891e-05, "loss": 1.133, "step": 4863 }, { "epoch": 0.3295616234162206, "grad_norm": 7.901734352111816, "learning_rate": 9.43692244506811e-05, "loss": 0.8293, "step": 4864 }, { "epoch": 0.3296293786841927, "grad_norm": 8.270600318908691, "learning_rate": 9.436785543158328e-05, "loss": 1.1584, "step": 4865 }, { "epoch": 0.32969713395216477, "grad_norm": 6.765751838684082, "learning_rate": 9.436648641248546e-05, "loss": 0.7705, "step": 4866 }, { "epoch": 0.32976488922013686, "grad_norm": 7.272820949554443, "learning_rate": 9.436511739338765e-05, "loss": 1.2151, "step": 4867 }, { "epoch": 0.32983264448810895, "grad_norm": 6.639741897583008, "learning_rate": 9.436374837428983e-05, "loss": 0.8868, "step": 4868 }, { "epoch": 0.32990039975608104, "grad_norm": 7.6254143714904785, "learning_rate": 9.436237935519201e-05, "loss": 1.0342, "step": 4869 }, { "epoch": 0.32996815502405313, "grad_norm": 6.546008586883545, "learning_rate": 9.436101033609419e-05, "loss": 0.7223, "step": 4870 }, { "epoch": 0.3300359102920252, "grad_norm": 6.215434551239014, "learning_rate": 9.435964131699637e-05, "loss": 0.8506, "step": 4871 }, { "epoch": 0.3301036655599973, "grad_norm": 7.897797584533691, "learning_rate": 9.435827229789856e-05, "loss": 0.8699, "step": 4872 }, { "epoch": 0.33017142082796935, "grad_norm": 8.3803129196167, "learning_rate": 9.435690327880074e-05, "loss": 1.0922, "step": 4873 }, { "epoch": 0.33023917609594144, "grad_norm": 6.675380706787109, "learning_rate": 9.435553425970293e-05, "loss": 0.8401, "step": 4874 }, { "epoch": 0.33030693136391354, "grad_norm": 7.380627632141113, "learning_rate": 9.43541652406051e-05, "loss": 0.615, "step": 4875 }, { "epoch": 0.33037468663188563, "grad_norm": 8.120410919189453, "learning_rate": 9.435279622150729e-05, "loss": 1.2622, "step": 4876 }, { "epoch": 0.3304424418998577, "grad_norm": 7.170663356781006, "learning_rate": 9.435142720240948e-05, "loss": 0.893, "step": 4877 }, { "epoch": 0.3305101971678298, "grad_norm": 12.477727890014648, "learning_rate": 9.435005818331166e-05, "loss": 0.9654, "step": 4878 }, { "epoch": 0.3305779524358019, "grad_norm": 6.562829971313477, "learning_rate": 9.434868916421384e-05, "loss": 0.7042, "step": 4879 }, { "epoch": 0.330645707703774, "grad_norm": 8.406328201293945, "learning_rate": 9.434732014511602e-05, "loss": 1.0807, "step": 4880 }, { "epoch": 0.33071346297174603, "grad_norm": 9.06248950958252, "learning_rate": 9.434595112601821e-05, "loss": 0.8586, "step": 4881 }, { "epoch": 0.3307812182397181, "grad_norm": 8.743175506591797, "learning_rate": 9.43445821069204e-05, "loss": 0.9501, "step": 4882 }, { "epoch": 0.3308489735076902, "grad_norm": 7.666022777557373, "learning_rate": 9.434321308782258e-05, "loss": 0.9411, "step": 4883 }, { "epoch": 0.3309167287756623, "grad_norm": 6.989424705505371, "learning_rate": 9.434184406872476e-05, "loss": 0.9511, "step": 4884 }, { "epoch": 0.3309844840436344, "grad_norm": 6.9908905029296875, "learning_rate": 9.434047504962694e-05, "loss": 0.9769, "step": 4885 }, { "epoch": 0.3310522393116065, "grad_norm": 7.985763072967529, "learning_rate": 9.433910603052913e-05, "loss": 1.1041, "step": 4886 }, { "epoch": 0.3311199945795786, "grad_norm": 7.9841694831848145, "learning_rate": 9.433773701143131e-05, "loss": 0.8341, "step": 4887 }, { "epoch": 0.33118774984755067, "grad_norm": 6.9602952003479, "learning_rate": 9.433636799233349e-05, "loss": 0.9388, "step": 4888 }, { "epoch": 0.3312555051155227, "grad_norm": 7.2530012130737305, "learning_rate": 9.433499897323567e-05, "loss": 0.9396, "step": 4889 }, { "epoch": 0.3313232603834948, "grad_norm": 7.338861465454102, "learning_rate": 9.433362995413786e-05, "loss": 0.9168, "step": 4890 }, { "epoch": 0.3313910156514669, "grad_norm": 7.08353853225708, "learning_rate": 9.433226093504005e-05, "loss": 0.9332, "step": 4891 }, { "epoch": 0.331458770919439, "grad_norm": 7.379842758178711, "learning_rate": 9.433089191594223e-05, "loss": 0.7139, "step": 4892 }, { "epoch": 0.3315265261874111, "grad_norm": 8.599993705749512, "learning_rate": 9.432952289684442e-05, "loss": 1.0023, "step": 4893 }, { "epoch": 0.33159428145538317, "grad_norm": 7.096752643585205, "learning_rate": 9.43281538777466e-05, "loss": 0.9052, "step": 4894 }, { "epoch": 0.33166203672335526, "grad_norm": 6.28333044052124, "learning_rate": 9.432678485864878e-05, "loss": 0.8721, "step": 4895 }, { "epoch": 0.33172979199132735, "grad_norm": 8.48362922668457, "learning_rate": 9.432541583955097e-05, "loss": 0.9774, "step": 4896 }, { "epoch": 0.3317975472592994, "grad_norm": 8.214259147644043, "learning_rate": 9.432404682045315e-05, "loss": 0.999, "step": 4897 }, { "epoch": 0.3318653025272715, "grad_norm": 9.603363990783691, "learning_rate": 9.432267780135533e-05, "loss": 0.7848, "step": 4898 }, { "epoch": 0.33193305779524357, "grad_norm": 7.982606887817383, "learning_rate": 9.432130878225753e-05, "loss": 0.7167, "step": 4899 }, { "epoch": 0.33200081306321566, "grad_norm": 6.8360915184021, "learning_rate": 9.431993976315971e-05, "loss": 1.1884, "step": 4900 }, { "epoch": 0.33206856833118775, "grad_norm": 9.70743179321289, "learning_rate": 9.431857074406189e-05, "loss": 1.0198, "step": 4901 }, { "epoch": 0.33213632359915984, "grad_norm": 7.7961554527282715, "learning_rate": 9.431720172496407e-05, "loss": 0.8637, "step": 4902 }, { "epoch": 0.33220407886713194, "grad_norm": 7.809814453125, "learning_rate": 9.431583270586625e-05, "loss": 1.0737, "step": 4903 }, { "epoch": 0.332271834135104, "grad_norm": 6.5765862464904785, "learning_rate": 9.431446368676844e-05, "loss": 0.6574, "step": 4904 }, { "epoch": 0.33233958940307606, "grad_norm": 6.912911891937256, "learning_rate": 9.431309466767062e-05, "loss": 0.9647, "step": 4905 }, { "epoch": 0.33240734467104815, "grad_norm": 7.2532877922058105, "learning_rate": 9.43117256485728e-05, "loss": 1.098, "step": 4906 }, { "epoch": 0.33247509993902025, "grad_norm": 6.589138984680176, "learning_rate": 9.431035662947498e-05, "loss": 0.9948, "step": 4907 }, { "epoch": 0.33254285520699234, "grad_norm": 9.1469144821167, "learning_rate": 9.430898761037717e-05, "loss": 0.9656, "step": 4908 }, { "epoch": 0.33261061047496443, "grad_norm": 7.124277591705322, "learning_rate": 9.430761859127936e-05, "loss": 0.8704, "step": 4909 }, { "epoch": 0.3326783657429365, "grad_norm": 6.862776279449463, "learning_rate": 9.430624957218154e-05, "loss": 0.8008, "step": 4910 }, { "epoch": 0.3327461210109086, "grad_norm": 6.245189666748047, "learning_rate": 9.430488055308372e-05, "loss": 0.8529, "step": 4911 }, { "epoch": 0.3328138762788807, "grad_norm": 8.546821594238281, "learning_rate": 9.43035115339859e-05, "loss": 1.1747, "step": 4912 }, { "epoch": 0.33288163154685274, "grad_norm": 7.5279765129089355, "learning_rate": 9.43021425148881e-05, "loss": 0.8725, "step": 4913 }, { "epoch": 0.33294938681482483, "grad_norm": 7.987123966217041, "learning_rate": 9.430077349579027e-05, "loss": 0.8752, "step": 4914 }, { "epoch": 0.3330171420827969, "grad_norm": 10.1973876953125, "learning_rate": 9.429940447669245e-05, "loss": 1.0813, "step": 4915 }, { "epoch": 0.333084897350769, "grad_norm": 7.406893253326416, "learning_rate": 9.429803545759464e-05, "loss": 0.7064, "step": 4916 }, { "epoch": 0.3331526526187411, "grad_norm": 6.545569896697998, "learning_rate": 9.429666643849682e-05, "loss": 0.9005, "step": 4917 }, { "epoch": 0.3332204078867132, "grad_norm": 8.624588012695312, "learning_rate": 9.429529741939901e-05, "loss": 0.939, "step": 4918 }, { "epoch": 0.3332881631546853, "grad_norm": 6.811842441558838, "learning_rate": 9.429392840030119e-05, "loss": 0.8367, "step": 4919 }, { "epoch": 0.3333559184226574, "grad_norm": 7.256634712219238, "learning_rate": 9.429255938120337e-05, "loss": 1.0455, "step": 4920 }, { "epoch": 0.3334236736906295, "grad_norm": 6.900022983551025, "learning_rate": 9.429119036210555e-05, "loss": 0.954, "step": 4921 }, { "epoch": 0.3334914289586015, "grad_norm": 9.217741012573242, "learning_rate": 9.428982134300774e-05, "loss": 0.6392, "step": 4922 }, { "epoch": 0.3335591842265736, "grad_norm": 7.729586124420166, "learning_rate": 9.428845232390992e-05, "loss": 0.9372, "step": 4923 }, { "epoch": 0.3336269394945457, "grad_norm": 6.606268405914307, "learning_rate": 9.42870833048121e-05, "loss": 0.8127, "step": 4924 }, { "epoch": 0.3336946947625178, "grad_norm": 8.078627586364746, "learning_rate": 9.428571428571429e-05, "loss": 0.847, "step": 4925 }, { "epoch": 0.3337624500304899, "grad_norm": 8.212615966796875, "learning_rate": 9.428434526661647e-05, "loss": 1.0133, "step": 4926 }, { "epoch": 0.33383020529846197, "grad_norm": 6.68360710144043, "learning_rate": 9.428297624751866e-05, "loss": 0.9796, "step": 4927 }, { "epoch": 0.33389796056643406, "grad_norm": 7.115147590637207, "learning_rate": 9.428160722842084e-05, "loss": 0.9106, "step": 4928 }, { "epoch": 0.33396571583440615, "grad_norm": 7.178501129150391, "learning_rate": 9.428023820932302e-05, "loss": 0.7531, "step": 4929 }, { "epoch": 0.3340334711023782, "grad_norm": 7.303642749786377, "learning_rate": 9.42788691902252e-05, "loss": 0.9163, "step": 4930 }, { "epoch": 0.3341012263703503, "grad_norm": 7.532678604125977, "learning_rate": 9.427750017112738e-05, "loss": 1.001, "step": 4931 }, { "epoch": 0.33416898163832237, "grad_norm": 6.791550636291504, "learning_rate": 9.427613115202957e-05, "loss": 0.9005, "step": 4932 }, { "epoch": 0.33423673690629446, "grad_norm": 7.264156341552734, "learning_rate": 9.427476213293176e-05, "loss": 0.8076, "step": 4933 }, { "epoch": 0.33430449217426655, "grad_norm": 8.477554321289062, "learning_rate": 9.427339311383394e-05, "loss": 0.9363, "step": 4934 }, { "epoch": 0.33437224744223865, "grad_norm": 7.020787239074707, "learning_rate": 9.427202409473612e-05, "loss": 0.6933, "step": 4935 }, { "epoch": 0.33444000271021074, "grad_norm": 7.990420818328857, "learning_rate": 9.427065507563831e-05, "loss": 0.9759, "step": 4936 }, { "epoch": 0.33450775797818283, "grad_norm": 8.950580596923828, "learning_rate": 9.426928605654049e-05, "loss": 1.2876, "step": 4937 }, { "epoch": 0.33457551324615487, "grad_norm": 6.879635334014893, "learning_rate": 9.426791703744267e-05, "loss": 0.992, "step": 4938 }, { "epoch": 0.33464326851412696, "grad_norm": 7.640803813934326, "learning_rate": 9.426654801834486e-05, "loss": 1.1523, "step": 4939 }, { "epoch": 0.33471102378209905, "grad_norm": 7.156732559204102, "learning_rate": 9.426517899924704e-05, "loss": 0.6521, "step": 4940 }, { "epoch": 0.33477877905007114, "grad_norm": 7.543867588043213, "learning_rate": 9.426380998014922e-05, "loss": 0.8495, "step": 4941 }, { "epoch": 0.33484653431804323, "grad_norm": 8.554939270019531, "learning_rate": 9.426244096105142e-05, "loss": 1.0593, "step": 4942 }, { "epoch": 0.3349142895860153, "grad_norm": 7.902100563049316, "learning_rate": 9.42610719419536e-05, "loss": 1.0271, "step": 4943 }, { "epoch": 0.3349820448539874, "grad_norm": 7.945007801055908, "learning_rate": 9.425970292285578e-05, "loss": 0.7839, "step": 4944 }, { "epoch": 0.3350498001219595, "grad_norm": 7.539274215698242, "learning_rate": 9.425833390375797e-05, "loss": 0.968, "step": 4945 }, { "epoch": 0.33511755538993154, "grad_norm": 7.23228120803833, "learning_rate": 9.425696488466015e-05, "loss": 0.8319, "step": 4946 }, { "epoch": 0.33518531065790363, "grad_norm": 7.251431941986084, "learning_rate": 9.425559586556233e-05, "loss": 0.9513, "step": 4947 }, { "epoch": 0.3352530659258757, "grad_norm": 7.047804355621338, "learning_rate": 9.425422684646451e-05, "loss": 0.8677, "step": 4948 }, { "epoch": 0.3353208211938478, "grad_norm": 7.7244696617126465, "learning_rate": 9.42528578273667e-05, "loss": 0.8705, "step": 4949 }, { "epoch": 0.3353885764618199, "grad_norm": 6.078210830688477, "learning_rate": 9.425148880826889e-05, "loss": 0.9205, "step": 4950 }, { "epoch": 0.335456331729792, "grad_norm": 7.856949329376221, "learning_rate": 9.425011978917107e-05, "loss": 0.8637, "step": 4951 }, { "epoch": 0.3355240869977641, "grad_norm": 7.341653347015381, "learning_rate": 9.424875077007325e-05, "loss": 0.9582, "step": 4952 }, { "epoch": 0.3355918422657362, "grad_norm": 7.456873893737793, "learning_rate": 9.424738175097543e-05, "loss": 0.8581, "step": 4953 }, { "epoch": 0.3356595975337082, "grad_norm": 7.885173797607422, "learning_rate": 9.424601273187761e-05, "loss": 0.9551, "step": 4954 }, { "epoch": 0.3357273528016803, "grad_norm": 7.016728401184082, "learning_rate": 9.42446437127798e-05, "loss": 0.9493, "step": 4955 }, { "epoch": 0.3357951080696524, "grad_norm": 6.383670806884766, "learning_rate": 9.424327469368198e-05, "loss": 0.8178, "step": 4956 }, { "epoch": 0.3358628633376245, "grad_norm": 6.930068492889404, "learning_rate": 9.424190567458416e-05, "loss": 1.1073, "step": 4957 }, { "epoch": 0.3359306186055966, "grad_norm": 7.001153469085693, "learning_rate": 9.424053665548634e-05, "loss": 0.844, "step": 4958 }, { "epoch": 0.3359983738735687, "grad_norm": 8.316204071044922, "learning_rate": 9.423916763638854e-05, "loss": 0.9363, "step": 4959 }, { "epoch": 0.33606612914154077, "grad_norm": 7.931922912597656, "learning_rate": 9.423779861729072e-05, "loss": 1.1617, "step": 4960 }, { "epoch": 0.33613388440951286, "grad_norm": 7.447847366333008, "learning_rate": 9.42364295981929e-05, "loss": 0.9137, "step": 4961 }, { "epoch": 0.3362016396774849, "grad_norm": 6.582141876220703, "learning_rate": 9.423506057909508e-05, "loss": 1.0516, "step": 4962 }, { "epoch": 0.336269394945457, "grad_norm": 6.278825283050537, "learning_rate": 9.423369155999726e-05, "loss": 0.9489, "step": 4963 }, { "epoch": 0.3363371502134291, "grad_norm": 7.415517330169678, "learning_rate": 9.423232254089945e-05, "loss": 0.9275, "step": 4964 }, { "epoch": 0.3364049054814012, "grad_norm": 6.073827266693115, "learning_rate": 9.423095352180163e-05, "loss": 0.7514, "step": 4965 }, { "epoch": 0.33647266074937326, "grad_norm": 9.100361824035645, "learning_rate": 9.422958450270381e-05, "loss": 1.0882, "step": 4966 }, { "epoch": 0.33654041601734536, "grad_norm": 8.23529052734375, "learning_rate": 9.4228215483606e-05, "loss": 0.8323, "step": 4967 }, { "epoch": 0.33660817128531745, "grad_norm": 9.86119270324707, "learning_rate": 9.422684646450819e-05, "loss": 0.8997, "step": 4968 }, { "epoch": 0.33667592655328954, "grad_norm": 7.328428268432617, "learning_rate": 9.422547744541037e-05, "loss": 0.8558, "step": 4969 }, { "epoch": 0.3367436818212616, "grad_norm": 6.703019618988037, "learning_rate": 9.422410842631255e-05, "loss": 0.782, "step": 4970 }, { "epoch": 0.33681143708923367, "grad_norm": 7.780190944671631, "learning_rate": 9.422273940721473e-05, "loss": 1.1681, "step": 4971 }, { "epoch": 0.33687919235720576, "grad_norm": 6.972787380218506, "learning_rate": 9.422137038811691e-05, "loss": 1.0137, "step": 4972 }, { "epoch": 0.33694694762517785, "grad_norm": 6.521172523498535, "learning_rate": 9.42200013690191e-05, "loss": 0.8866, "step": 4973 }, { "epoch": 0.33701470289314994, "grad_norm": 6.414315223693848, "learning_rate": 9.421863234992128e-05, "loss": 0.7614, "step": 4974 }, { "epoch": 0.33708245816112203, "grad_norm": 8.930048942565918, "learning_rate": 9.421726333082346e-05, "loss": 1.0757, "step": 4975 }, { "epoch": 0.3371502134290941, "grad_norm": 6.710857391357422, "learning_rate": 9.421589431172565e-05, "loss": 0.8966, "step": 4976 }, { "epoch": 0.3372179686970662, "grad_norm": 10.788898468017578, "learning_rate": 9.421452529262784e-05, "loss": 0.906, "step": 4977 }, { "epoch": 0.3372857239650383, "grad_norm": 6.601465225219727, "learning_rate": 9.421315627353002e-05, "loss": 0.967, "step": 4978 }, { "epoch": 0.33735347923301034, "grad_norm": 7.0998406410217285, "learning_rate": 9.42117872544322e-05, "loss": 1.0584, "step": 4979 }, { "epoch": 0.33742123450098244, "grad_norm": 8.517425537109375, "learning_rate": 9.421041823533438e-05, "loss": 0.83, "step": 4980 }, { "epoch": 0.33748898976895453, "grad_norm": 7.631463050842285, "learning_rate": 9.420904921623656e-05, "loss": 0.7431, "step": 4981 }, { "epoch": 0.3375567450369266, "grad_norm": 7.006965637207031, "learning_rate": 9.420768019713875e-05, "loss": 1.0081, "step": 4982 }, { "epoch": 0.3376245003048987, "grad_norm": 7.085225582122803, "learning_rate": 9.420631117804093e-05, "loss": 1.0094, "step": 4983 }, { "epoch": 0.3376922555728708, "grad_norm": 8.096879959106445, "learning_rate": 9.420494215894312e-05, "loss": 1.0376, "step": 4984 }, { "epoch": 0.3377600108408429, "grad_norm": 7.823955535888672, "learning_rate": 9.420357313984531e-05, "loss": 0.975, "step": 4985 }, { "epoch": 0.337827766108815, "grad_norm": 6.340082168579102, "learning_rate": 9.420220412074749e-05, "loss": 0.6133, "step": 4986 }, { "epoch": 0.337895521376787, "grad_norm": 9.930416107177734, "learning_rate": 9.420083510164967e-05, "loss": 1.0132, "step": 4987 }, { "epoch": 0.3379632766447591, "grad_norm": 6.7059221267700195, "learning_rate": 9.419946608255186e-05, "loss": 0.7784, "step": 4988 }, { "epoch": 0.3380310319127312, "grad_norm": 7.869287967681885, "learning_rate": 9.419809706345404e-05, "loss": 0.9114, "step": 4989 }, { "epoch": 0.3380987871807033, "grad_norm": 8.061338424682617, "learning_rate": 9.419672804435622e-05, "loss": 1.0982, "step": 4990 }, { "epoch": 0.3381665424486754, "grad_norm": 6.226239204406738, "learning_rate": 9.419535902525842e-05, "loss": 0.8251, "step": 4991 }, { "epoch": 0.3382342977166475, "grad_norm": 6.236758232116699, "learning_rate": 9.41939900061606e-05, "loss": 1.0367, "step": 4992 }, { "epoch": 0.3383020529846196, "grad_norm": 8.880475044250488, "learning_rate": 9.419262098706278e-05, "loss": 1.0941, "step": 4993 }, { "epoch": 0.33836980825259166, "grad_norm": 7.7189621925354, "learning_rate": 9.419125196796496e-05, "loss": 1.2868, "step": 4994 }, { "epoch": 0.3384375635205637, "grad_norm": 8.368658065795898, "learning_rate": 9.418988294886714e-05, "loss": 1.1495, "step": 4995 }, { "epoch": 0.3385053187885358, "grad_norm": 6.596102237701416, "learning_rate": 9.418851392976933e-05, "loss": 0.9639, "step": 4996 }, { "epoch": 0.3385730740565079, "grad_norm": 6.297356605529785, "learning_rate": 9.418714491067151e-05, "loss": 0.8802, "step": 4997 }, { "epoch": 0.33864082932448, "grad_norm": 6.869907855987549, "learning_rate": 9.41857758915737e-05, "loss": 0.8247, "step": 4998 }, { "epoch": 0.33870858459245207, "grad_norm": 8.147067070007324, "learning_rate": 9.418440687247587e-05, "loss": 0.7532, "step": 4999 }, { "epoch": 0.33877633986042416, "grad_norm": 7.368498802185059, "learning_rate": 9.418303785337807e-05, "loss": 0.8697, "step": 5000 }, { "epoch": 0.33884409512839625, "grad_norm": 6.064372539520264, "learning_rate": 9.418166883428025e-05, "loss": 0.8134, "step": 5001 }, { "epoch": 0.33891185039636834, "grad_norm": 7.22601318359375, "learning_rate": 9.418029981518243e-05, "loss": 0.8224, "step": 5002 }, { "epoch": 0.3389796056643404, "grad_norm": 7.832242012023926, "learning_rate": 9.417893079608461e-05, "loss": 0.9114, "step": 5003 }, { "epoch": 0.33904736093231247, "grad_norm": 7.05675745010376, "learning_rate": 9.417756177698679e-05, "loss": 0.8475, "step": 5004 }, { "epoch": 0.33911511620028456, "grad_norm": 6.385340690612793, "learning_rate": 9.417619275788898e-05, "loss": 0.8604, "step": 5005 }, { "epoch": 0.33918287146825665, "grad_norm": 7.100057601928711, "learning_rate": 9.417482373879116e-05, "loss": 0.8418, "step": 5006 }, { "epoch": 0.33925062673622874, "grad_norm": 8.983677864074707, "learning_rate": 9.417345471969334e-05, "loss": 1.1282, "step": 5007 }, { "epoch": 0.33931838200420084, "grad_norm": 8.526215553283691, "learning_rate": 9.417208570059552e-05, "loss": 0.8759, "step": 5008 }, { "epoch": 0.3393861372721729, "grad_norm": 7.278728485107422, "learning_rate": 9.41707166814977e-05, "loss": 1.0722, "step": 5009 }, { "epoch": 0.339453892540145, "grad_norm": 9.66884994506836, "learning_rate": 9.41693476623999e-05, "loss": 0.8776, "step": 5010 }, { "epoch": 0.33952164780811706, "grad_norm": 7.001797199249268, "learning_rate": 9.416797864330208e-05, "loss": 0.7863, "step": 5011 }, { "epoch": 0.33958940307608915, "grad_norm": 7.707855224609375, "learning_rate": 9.416660962420426e-05, "loss": 0.9138, "step": 5012 }, { "epoch": 0.33965715834406124, "grad_norm": 7.721848964691162, "learning_rate": 9.416524060510644e-05, "loss": 1.1121, "step": 5013 }, { "epoch": 0.33972491361203333, "grad_norm": 8.91602611541748, "learning_rate": 9.416387158600863e-05, "loss": 0.9152, "step": 5014 }, { "epoch": 0.3397926688800054, "grad_norm": 7.332693099975586, "learning_rate": 9.416250256691081e-05, "loss": 0.9559, "step": 5015 }, { "epoch": 0.3398604241479775, "grad_norm": 7.198998928070068, "learning_rate": 9.4161133547813e-05, "loss": 0.9239, "step": 5016 }, { "epoch": 0.3399281794159496, "grad_norm": 7.566831588745117, "learning_rate": 9.415976452871517e-05, "loss": 0.9281, "step": 5017 }, { "epoch": 0.3399959346839217, "grad_norm": 6.832939624786377, "learning_rate": 9.415839550961736e-05, "loss": 0.9164, "step": 5018 }, { "epoch": 0.34006368995189373, "grad_norm": 7.087849140167236, "learning_rate": 9.415702649051955e-05, "loss": 0.9314, "step": 5019 }, { "epoch": 0.3401314452198658, "grad_norm": 7.417871952056885, "learning_rate": 9.415565747142173e-05, "loss": 0.9687, "step": 5020 }, { "epoch": 0.3401992004878379, "grad_norm": 6.9238362312316895, "learning_rate": 9.415428845232391e-05, "loss": 1.0757, "step": 5021 }, { "epoch": 0.34026695575581, "grad_norm": 7.4261698722839355, "learning_rate": 9.415291943322609e-05, "loss": 1.0636, "step": 5022 }, { "epoch": 0.3403347110237821, "grad_norm": 6.1964030265808105, "learning_rate": 9.415155041412828e-05, "loss": 0.739, "step": 5023 }, { "epoch": 0.3404024662917542, "grad_norm": 7.07076358795166, "learning_rate": 9.415018139503046e-05, "loss": 0.9435, "step": 5024 }, { "epoch": 0.3404702215597263, "grad_norm": 8.069670677185059, "learning_rate": 9.414881237593264e-05, "loss": 0.9745, "step": 5025 }, { "epoch": 0.3405379768276984, "grad_norm": 7.559169769287109, "learning_rate": 9.414744335683482e-05, "loss": 0.9329, "step": 5026 }, { "epoch": 0.34060573209567047, "grad_norm": 8.438983917236328, "learning_rate": 9.4146074337737e-05, "loss": 1.0608, "step": 5027 }, { "epoch": 0.3406734873636425, "grad_norm": 6.522724151611328, "learning_rate": 9.41447053186392e-05, "loss": 0.7912, "step": 5028 }, { "epoch": 0.3407412426316146, "grad_norm": 8.31757926940918, "learning_rate": 9.414333629954138e-05, "loss": 0.7535, "step": 5029 }, { "epoch": 0.3408089978995867, "grad_norm": 7.540639877319336, "learning_rate": 9.414196728044356e-05, "loss": 0.7239, "step": 5030 }, { "epoch": 0.3408767531675588, "grad_norm": 7.016546726226807, "learning_rate": 9.414059826134575e-05, "loss": 1.0658, "step": 5031 }, { "epoch": 0.34094450843553087, "grad_norm": 5.6253814697265625, "learning_rate": 9.413922924224793e-05, "loss": 0.8522, "step": 5032 }, { "epoch": 0.34101226370350296, "grad_norm": 7.693328380584717, "learning_rate": 9.413786022315011e-05, "loss": 1.1011, "step": 5033 }, { "epoch": 0.34108001897147505, "grad_norm": 8.77978229522705, "learning_rate": 9.413649120405231e-05, "loss": 1.0617, "step": 5034 }, { "epoch": 0.34114777423944714, "grad_norm": 6.854990005493164, "learning_rate": 9.413512218495449e-05, "loss": 1.0232, "step": 5035 }, { "epoch": 0.3412155295074192, "grad_norm": 8.380729675292969, "learning_rate": 9.413375316585667e-05, "loss": 0.7554, "step": 5036 }, { "epoch": 0.34128328477539127, "grad_norm": 6.667641639709473, "learning_rate": 9.413238414675886e-05, "loss": 0.8992, "step": 5037 }, { "epoch": 0.34135104004336336, "grad_norm": 7.5778489112854, "learning_rate": 9.413101512766104e-05, "loss": 0.9521, "step": 5038 }, { "epoch": 0.34141879531133545, "grad_norm": 10.053051948547363, "learning_rate": 9.412964610856322e-05, "loss": 0.9595, "step": 5039 }, { "epoch": 0.34148655057930755, "grad_norm": 7.460272789001465, "learning_rate": 9.41282770894654e-05, "loss": 0.683, "step": 5040 }, { "epoch": 0.34155430584727964, "grad_norm": 7.724341869354248, "learning_rate": 9.412690807036758e-05, "loss": 0.9902, "step": 5041 }, { "epoch": 0.34162206111525173, "grad_norm": 7.729767799377441, "learning_rate": 9.412553905126978e-05, "loss": 0.9249, "step": 5042 }, { "epoch": 0.3416898163832238, "grad_norm": 6.388404846191406, "learning_rate": 9.412417003217196e-05, "loss": 1.0154, "step": 5043 }, { "epoch": 0.34175757165119586, "grad_norm": 6.2206130027771, "learning_rate": 9.412280101307414e-05, "loss": 0.8169, "step": 5044 }, { "epoch": 0.34182532691916795, "grad_norm": 7.383151054382324, "learning_rate": 9.412143199397632e-05, "loss": 1.0476, "step": 5045 }, { "epoch": 0.34189308218714004, "grad_norm": 7.733860969543457, "learning_rate": 9.412006297487851e-05, "loss": 0.9861, "step": 5046 }, { "epoch": 0.34196083745511213, "grad_norm": 7.823349475860596, "learning_rate": 9.41186939557807e-05, "loss": 0.8351, "step": 5047 }, { "epoch": 0.3420285927230842, "grad_norm": 10.20052433013916, "learning_rate": 9.411732493668287e-05, "loss": 1.2573, "step": 5048 }, { "epoch": 0.3420963479910563, "grad_norm": 7.878631114959717, "learning_rate": 9.411595591758505e-05, "loss": 1.0979, "step": 5049 }, { "epoch": 0.3421641032590284, "grad_norm": 7.230095386505127, "learning_rate": 9.411458689848723e-05, "loss": 0.9929, "step": 5050 }, { "epoch": 0.3422318585270005, "grad_norm": 7.354804992675781, "learning_rate": 9.411321787938943e-05, "loss": 1.0552, "step": 5051 }, { "epoch": 0.34229961379497253, "grad_norm": 9.846439361572266, "learning_rate": 9.411184886029161e-05, "loss": 1.0616, "step": 5052 }, { "epoch": 0.3423673690629446, "grad_norm": 6.999122142791748, "learning_rate": 9.411047984119379e-05, "loss": 0.9958, "step": 5053 }, { "epoch": 0.3424351243309167, "grad_norm": 5.719442367553711, "learning_rate": 9.410911082209597e-05, "loss": 0.8239, "step": 5054 }, { "epoch": 0.3425028795988888, "grad_norm": 7.650005340576172, "learning_rate": 9.410774180299816e-05, "loss": 0.9736, "step": 5055 }, { "epoch": 0.3425706348668609, "grad_norm": 7.938320159912109, "learning_rate": 9.410637278390034e-05, "loss": 0.7809, "step": 5056 }, { "epoch": 0.342638390134833, "grad_norm": 8.319568634033203, "learning_rate": 9.410500376480252e-05, "loss": 0.9162, "step": 5057 }, { "epoch": 0.3427061454028051, "grad_norm": 8.151114463806152, "learning_rate": 9.41036347457047e-05, "loss": 1.1071, "step": 5058 }, { "epoch": 0.3427739006707772, "grad_norm": 7.333863735198975, "learning_rate": 9.410226572660688e-05, "loss": 0.8425, "step": 5059 }, { "epoch": 0.3428416559387492, "grad_norm": 6.848262310028076, "learning_rate": 9.410089670750908e-05, "loss": 0.9385, "step": 5060 }, { "epoch": 0.3429094112067213, "grad_norm": 7.205636024475098, "learning_rate": 9.409952768841126e-05, "loss": 0.8448, "step": 5061 }, { "epoch": 0.3429771664746934, "grad_norm": 7.133283615112305, "learning_rate": 9.409815866931344e-05, "loss": 0.8306, "step": 5062 }, { "epoch": 0.3430449217426655, "grad_norm": 6.693148136138916, "learning_rate": 9.409678965021562e-05, "loss": 1.0321, "step": 5063 }, { "epoch": 0.3431126770106376, "grad_norm": 7.183022975921631, "learning_rate": 9.40954206311178e-05, "loss": 1.0067, "step": 5064 }, { "epoch": 0.34318043227860967, "grad_norm": 7.537884712219238, "learning_rate": 9.409405161202e-05, "loss": 0.9719, "step": 5065 }, { "epoch": 0.34324818754658176, "grad_norm": 7.812978267669678, "learning_rate": 9.409268259292217e-05, "loss": 0.9936, "step": 5066 }, { "epoch": 0.34331594281455385, "grad_norm": 7.93212366104126, "learning_rate": 9.409131357382435e-05, "loss": 1.1685, "step": 5067 }, { "epoch": 0.3433836980825259, "grad_norm": 7.298057556152344, "learning_rate": 9.408994455472653e-05, "loss": 0.908, "step": 5068 }, { "epoch": 0.343451453350498, "grad_norm": 8.108441352844238, "learning_rate": 9.408857553562873e-05, "loss": 0.9916, "step": 5069 }, { "epoch": 0.3435192086184701, "grad_norm": 7.206883430480957, "learning_rate": 9.408720651653091e-05, "loss": 0.9431, "step": 5070 }, { "epoch": 0.34358696388644216, "grad_norm": 8.54073715209961, "learning_rate": 9.408583749743309e-05, "loss": 0.9156, "step": 5071 }, { "epoch": 0.34365471915441426, "grad_norm": 7.274294376373291, "learning_rate": 9.408446847833527e-05, "loss": 0.7436, "step": 5072 }, { "epoch": 0.34372247442238635, "grad_norm": 7.734363079071045, "learning_rate": 9.408309945923745e-05, "loss": 1.054, "step": 5073 }, { "epoch": 0.34379022969035844, "grad_norm": 7.952118873596191, "learning_rate": 9.408173044013964e-05, "loss": 0.9152, "step": 5074 }, { "epoch": 0.34385798495833053, "grad_norm": 7.481127738952637, "learning_rate": 9.408036142104182e-05, "loss": 1.0134, "step": 5075 }, { "epoch": 0.34392574022630257, "grad_norm": 8.52331256866455, "learning_rate": 9.4078992401944e-05, "loss": 0.9724, "step": 5076 }, { "epoch": 0.34399349549427466, "grad_norm": 7.454250335693359, "learning_rate": 9.40776233828462e-05, "loss": 0.9535, "step": 5077 }, { "epoch": 0.34406125076224675, "grad_norm": 7.192932605743408, "learning_rate": 9.407625436374838e-05, "loss": 0.8923, "step": 5078 }, { "epoch": 0.34412900603021884, "grad_norm": 8.29765796661377, "learning_rate": 9.407488534465056e-05, "loss": 1.0374, "step": 5079 }, { "epoch": 0.34419676129819093, "grad_norm": 7.1159820556640625, "learning_rate": 9.407351632555275e-05, "loss": 0.7876, "step": 5080 }, { "epoch": 0.344264516566163, "grad_norm": 7.157162189483643, "learning_rate": 9.407214730645493e-05, "loss": 0.8747, "step": 5081 }, { "epoch": 0.3443322718341351, "grad_norm": 6.686028003692627, "learning_rate": 9.407077828735711e-05, "loss": 0.8189, "step": 5082 }, { "epoch": 0.3444000271021072, "grad_norm": 8.957246780395508, "learning_rate": 9.406940926825931e-05, "loss": 1.0617, "step": 5083 }, { "epoch": 0.3444677823700793, "grad_norm": 7.599720001220703, "learning_rate": 9.406804024916149e-05, "loss": 1.0455, "step": 5084 }, { "epoch": 0.34453553763805134, "grad_norm": 5.757546424865723, "learning_rate": 9.406667123006367e-05, "loss": 0.7654, "step": 5085 }, { "epoch": 0.34460329290602343, "grad_norm": 6.592649936676025, "learning_rate": 9.406530221096585e-05, "loss": 0.9713, "step": 5086 }, { "epoch": 0.3446710481739955, "grad_norm": 7.119808673858643, "learning_rate": 9.406393319186803e-05, "loss": 0.7413, "step": 5087 }, { "epoch": 0.3447388034419676, "grad_norm": 6.288692951202393, "learning_rate": 9.406256417277022e-05, "loss": 0.9479, "step": 5088 }, { "epoch": 0.3448065587099397, "grad_norm": 6.267573356628418, "learning_rate": 9.40611951536724e-05, "loss": 0.9745, "step": 5089 }, { "epoch": 0.3448743139779118, "grad_norm": 7.639352798461914, "learning_rate": 9.405982613457458e-05, "loss": 0.9655, "step": 5090 }, { "epoch": 0.3449420692458839, "grad_norm": 6.32698917388916, "learning_rate": 9.405845711547676e-05, "loss": 0.789, "step": 5091 }, { "epoch": 0.345009824513856, "grad_norm": 8.769354820251465, "learning_rate": 9.405708809637896e-05, "loss": 1.1124, "step": 5092 }, { "epoch": 0.345077579781828, "grad_norm": 7.179650783538818, "learning_rate": 9.405571907728114e-05, "loss": 0.7939, "step": 5093 }, { "epoch": 0.3451453350498001, "grad_norm": 10.112159729003906, "learning_rate": 9.405435005818332e-05, "loss": 0.943, "step": 5094 }, { "epoch": 0.3452130903177722, "grad_norm": 6.925206661224365, "learning_rate": 9.40529810390855e-05, "loss": 1.0444, "step": 5095 }, { "epoch": 0.3452808455857443, "grad_norm": 7.8664398193359375, "learning_rate": 9.405161201998768e-05, "loss": 0.9643, "step": 5096 }, { "epoch": 0.3453486008537164, "grad_norm": 7.060378551483154, "learning_rate": 9.405024300088987e-05, "loss": 0.9043, "step": 5097 }, { "epoch": 0.3454163561216885, "grad_norm": 8.105093002319336, "learning_rate": 9.404887398179205e-05, "loss": 1.1106, "step": 5098 }, { "epoch": 0.34548411138966056, "grad_norm": 7.8055739402771, "learning_rate": 9.404750496269423e-05, "loss": 0.9638, "step": 5099 }, { "epoch": 0.34555186665763266, "grad_norm": 8.225363731384277, "learning_rate": 9.404613594359641e-05, "loss": 0.7878, "step": 5100 }, { "epoch": 0.3456196219256047, "grad_norm": 7.809800148010254, "learning_rate": 9.404476692449861e-05, "loss": 0.8241, "step": 5101 }, { "epoch": 0.3456873771935768, "grad_norm": 7.520929336547852, "learning_rate": 9.404339790540079e-05, "loss": 1.0689, "step": 5102 }, { "epoch": 0.3457551324615489, "grad_norm": 5.996449947357178, "learning_rate": 9.404202888630297e-05, "loss": 0.9163, "step": 5103 }, { "epoch": 0.34582288772952097, "grad_norm": 7.2040114402771, "learning_rate": 9.404065986720515e-05, "loss": 1.0615, "step": 5104 }, { "epoch": 0.34589064299749306, "grad_norm": 7.268972873687744, "learning_rate": 9.403929084810733e-05, "loss": 0.8785, "step": 5105 }, { "epoch": 0.34595839826546515, "grad_norm": 9.244166374206543, "learning_rate": 9.403792182900952e-05, "loss": 0.8714, "step": 5106 }, { "epoch": 0.34602615353343724, "grad_norm": 6.766915321350098, "learning_rate": 9.40365528099117e-05, "loss": 0.8075, "step": 5107 }, { "epoch": 0.34609390880140933, "grad_norm": 8.137444496154785, "learning_rate": 9.403518379081388e-05, "loss": 1.0687, "step": 5108 }, { "epoch": 0.34616166406938137, "grad_norm": 7.488312244415283, "learning_rate": 9.403381477171606e-05, "loss": 0.8056, "step": 5109 }, { "epoch": 0.34622941933735346, "grad_norm": 6.555777549743652, "learning_rate": 9.403244575261826e-05, "loss": 0.8761, "step": 5110 }, { "epoch": 0.34629717460532555, "grad_norm": 11.631979942321777, "learning_rate": 9.403107673352044e-05, "loss": 0.8139, "step": 5111 }, { "epoch": 0.34636492987329764, "grad_norm": 5.973453998565674, "learning_rate": 9.402970771442262e-05, "loss": 0.8816, "step": 5112 }, { "epoch": 0.34643268514126974, "grad_norm": 5.817539215087891, "learning_rate": 9.40283386953248e-05, "loss": 0.6034, "step": 5113 }, { "epoch": 0.3465004404092418, "grad_norm": 8.215511322021484, "learning_rate": 9.402696967622698e-05, "loss": 0.8876, "step": 5114 }, { "epoch": 0.3465681956772139, "grad_norm": 5.848570346832275, "learning_rate": 9.402560065712917e-05, "loss": 0.7093, "step": 5115 }, { "epoch": 0.346635950945186, "grad_norm": 7.996893405914307, "learning_rate": 9.402423163803135e-05, "loss": 0.8865, "step": 5116 }, { "epoch": 0.34670370621315805, "grad_norm": 7.348632335662842, "learning_rate": 9.402286261893353e-05, "loss": 0.8042, "step": 5117 }, { "epoch": 0.34677146148113014, "grad_norm": 6.820111274719238, "learning_rate": 9.402149359983571e-05, "loss": 0.8744, "step": 5118 }, { "epoch": 0.34683921674910223, "grad_norm": 7.354914665222168, "learning_rate": 9.40201245807379e-05, "loss": 0.892, "step": 5119 }, { "epoch": 0.3469069720170743, "grad_norm": 7.751887321472168, "learning_rate": 9.401875556164009e-05, "loss": 0.8752, "step": 5120 }, { "epoch": 0.3469747272850464, "grad_norm": 9.630827903747559, "learning_rate": 9.401738654254227e-05, "loss": 0.999, "step": 5121 }, { "epoch": 0.3470424825530185, "grad_norm": 9.546615600585938, "learning_rate": 9.401601752344445e-05, "loss": 1.341, "step": 5122 }, { "epoch": 0.3471102378209906, "grad_norm": 7.031918048858643, "learning_rate": 9.401464850434663e-05, "loss": 1.1371, "step": 5123 }, { "epoch": 0.3471779930889627, "grad_norm": 9.030802726745605, "learning_rate": 9.401327948524882e-05, "loss": 0.9883, "step": 5124 }, { "epoch": 0.3472457483569347, "grad_norm": 6.786712646484375, "learning_rate": 9.4011910466151e-05, "loss": 0.8692, "step": 5125 }, { "epoch": 0.3473135036249068, "grad_norm": 7.327836990356445, "learning_rate": 9.401054144705318e-05, "loss": 0.8157, "step": 5126 }, { "epoch": 0.3473812588928789, "grad_norm": 6.41863489151001, "learning_rate": 9.400917242795538e-05, "loss": 0.8565, "step": 5127 }, { "epoch": 0.347449014160851, "grad_norm": 5.737055778503418, "learning_rate": 9.400780340885756e-05, "loss": 0.7385, "step": 5128 }, { "epoch": 0.3475167694288231, "grad_norm": 6.321473121643066, "learning_rate": 9.400643438975974e-05, "loss": 0.6459, "step": 5129 }, { "epoch": 0.3475845246967952, "grad_norm": 9.07598876953125, "learning_rate": 9.400506537066193e-05, "loss": 1.3719, "step": 5130 }, { "epoch": 0.3476522799647673, "grad_norm": 6.3348388671875, "learning_rate": 9.400369635156411e-05, "loss": 0.903, "step": 5131 }, { "epoch": 0.34772003523273937, "grad_norm": 7.836405277252197, "learning_rate": 9.40023273324663e-05, "loss": 0.8854, "step": 5132 }, { "epoch": 0.34778779050071146, "grad_norm": 7.712037086486816, "learning_rate": 9.400095831336849e-05, "loss": 0.8756, "step": 5133 }, { "epoch": 0.3478555457686835, "grad_norm": 7.445090293884277, "learning_rate": 9.399958929427067e-05, "loss": 1.0776, "step": 5134 }, { "epoch": 0.3479233010366556, "grad_norm": 6.7496724128723145, "learning_rate": 9.399822027517285e-05, "loss": 0.9425, "step": 5135 }, { "epoch": 0.3479910563046277, "grad_norm": 7.703073501586914, "learning_rate": 9.399685125607503e-05, "loss": 0.9217, "step": 5136 }, { "epoch": 0.34805881157259977, "grad_norm": 5.972830772399902, "learning_rate": 9.399548223697721e-05, "loss": 0.5922, "step": 5137 }, { "epoch": 0.34812656684057186, "grad_norm": 7.165718078613281, "learning_rate": 9.39941132178794e-05, "loss": 0.9624, "step": 5138 }, { "epoch": 0.34819432210854395, "grad_norm": 8.589313507080078, "learning_rate": 9.399274419878158e-05, "loss": 1.0485, "step": 5139 }, { "epoch": 0.34826207737651604, "grad_norm": 6.503042697906494, "learning_rate": 9.399137517968376e-05, "loss": 0.7833, "step": 5140 }, { "epoch": 0.34832983264448814, "grad_norm": 9.673978805541992, "learning_rate": 9.399000616058594e-05, "loss": 1.2266, "step": 5141 }, { "epoch": 0.34839758791246017, "grad_norm": 6.86154842376709, "learning_rate": 9.398863714148812e-05, "loss": 0.8709, "step": 5142 }, { "epoch": 0.34846534318043226, "grad_norm": 8.111627578735352, "learning_rate": 9.398726812239032e-05, "loss": 0.9657, "step": 5143 }, { "epoch": 0.34853309844840435, "grad_norm": 7.725754737854004, "learning_rate": 9.39858991032925e-05, "loss": 0.9551, "step": 5144 }, { "epoch": 0.34860085371637645, "grad_norm": 8.093070030212402, "learning_rate": 9.398453008419468e-05, "loss": 0.9814, "step": 5145 }, { "epoch": 0.34866860898434854, "grad_norm": 8.375652313232422, "learning_rate": 9.398316106509686e-05, "loss": 0.9232, "step": 5146 }, { "epoch": 0.34873636425232063, "grad_norm": 7.012859344482422, "learning_rate": 9.398179204599905e-05, "loss": 0.8322, "step": 5147 }, { "epoch": 0.3488041195202927, "grad_norm": 8.437539100646973, "learning_rate": 9.398042302690123e-05, "loss": 0.852, "step": 5148 }, { "epoch": 0.3488718747882648, "grad_norm": 9.844721794128418, "learning_rate": 9.397905400780341e-05, "loss": 0.8601, "step": 5149 }, { "epoch": 0.34893963005623685, "grad_norm": 7.359288215637207, "learning_rate": 9.39776849887056e-05, "loss": 0.9313, "step": 5150 }, { "epoch": 0.34900738532420894, "grad_norm": 7.528818607330322, "learning_rate": 9.397631596960777e-05, "loss": 0.923, "step": 5151 }, { "epoch": 0.34907514059218103, "grad_norm": 7.8577399253845215, "learning_rate": 9.397494695050997e-05, "loss": 0.9569, "step": 5152 }, { "epoch": 0.3491428958601531, "grad_norm": 6.960932731628418, "learning_rate": 9.397357793141215e-05, "loss": 0.9025, "step": 5153 }, { "epoch": 0.3492106511281252, "grad_norm": 6.392679691314697, "learning_rate": 9.397220891231433e-05, "loss": 0.7948, "step": 5154 }, { "epoch": 0.3492784063960973, "grad_norm": 8.22850227355957, "learning_rate": 9.397083989321651e-05, "loss": 0.9932, "step": 5155 }, { "epoch": 0.3493461616640694, "grad_norm": 6.1851277351379395, "learning_rate": 9.39694708741187e-05, "loss": 0.776, "step": 5156 }, { "epoch": 0.3494139169320415, "grad_norm": 9.058121681213379, "learning_rate": 9.396810185502088e-05, "loss": 1.1036, "step": 5157 }, { "epoch": 0.3494816722000135, "grad_norm": 7.11410665512085, "learning_rate": 9.396673283592306e-05, "loss": 0.8862, "step": 5158 }, { "epoch": 0.3495494274679856, "grad_norm": 7.148082256317139, "learning_rate": 9.396536381682524e-05, "loss": 0.9796, "step": 5159 }, { "epoch": 0.3496171827359577, "grad_norm": 7.258500576019287, "learning_rate": 9.396399479772742e-05, "loss": 0.9617, "step": 5160 }, { "epoch": 0.3496849380039298, "grad_norm": 5.1490044593811035, "learning_rate": 9.396262577862962e-05, "loss": 0.6844, "step": 5161 }, { "epoch": 0.3497526932719019, "grad_norm": 6.73121452331543, "learning_rate": 9.39612567595318e-05, "loss": 1.0669, "step": 5162 }, { "epoch": 0.349820448539874, "grad_norm": 7.891530990600586, "learning_rate": 9.395988774043398e-05, "loss": 1.0439, "step": 5163 }, { "epoch": 0.3498882038078461, "grad_norm": 7.256270885467529, "learning_rate": 9.395851872133616e-05, "loss": 0.9682, "step": 5164 }, { "epoch": 0.34995595907581817, "grad_norm": 7.153442859649658, "learning_rate": 9.395714970223835e-05, "loss": 0.913, "step": 5165 }, { "epoch": 0.3500237143437902, "grad_norm": 8.722851753234863, "learning_rate": 9.395578068314053e-05, "loss": 0.8749, "step": 5166 }, { "epoch": 0.3500914696117623, "grad_norm": 6.8031005859375, "learning_rate": 9.395441166404271e-05, "loss": 0.8755, "step": 5167 }, { "epoch": 0.3501592248797344, "grad_norm": 5.449770450592041, "learning_rate": 9.39530426449449e-05, "loss": 0.8813, "step": 5168 }, { "epoch": 0.3502269801477065, "grad_norm": 7.74420690536499, "learning_rate": 9.395167362584707e-05, "loss": 0.8361, "step": 5169 }, { "epoch": 0.35029473541567857, "grad_norm": 7.8701677322387695, "learning_rate": 9.395030460674927e-05, "loss": 0.9294, "step": 5170 }, { "epoch": 0.35036249068365066, "grad_norm": 5.897029876708984, "learning_rate": 9.394893558765145e-05, "loss": 0.7739, "step": 5171 }, { "epoch": 0.35043024595162275, "grad_norm": 6.7868781089782715, "learning_rate": 9.394756656855363e-05, "loss": 0.7252, "step": 5172 }, { "epoch": 0.35049800121959485, "grad_norm": 10.289764404296875, "learning_rate": 9.394619754945582e-05, "loss": 1.0943, "step": 5173 }, { "epoch": 0.3505657564875669, "grad_norm": 7.676051616668701, "learning_rate": 9.3944828530358e-05, "loss": 0.881, "step": 5174 }, { "epoch": 0.350633511755539, "grad_norm": 7.702653408050537, "learning_rate": 9.394345951126018e-05, "loss": 1.0346, "step": 5175 }, { "epoch": 0.35070126702351107, "grad_norm": 7.525374889373779, "learning_rate": 9.394209049216238e-05, "loss": 0.9955, "step": 5176 }, { "epoch": 0.35076902229148316, "grad_norm": 7.942657470703125, "learning_rate": 9.394072147306456e-05, "loss": 0.9715, "step": 5177 }, { "epoch": 0.35083677755945525, "grad_norm": 9.328110694885254, "learning_rate": 9.393935245396674e-05, "loss": 1.0753, "step": 5178 }, { "epoch": 0.35090453282742734, "grad_norm": 8.149508476257324, "learning_rate": 9.393798343486893e-05, "loss": 0.9818, "step": 5179 }, { "epoch": 0.35097228809539943, "grad_norm": 6.882593154907227, "learning_rate": 9.393661441577111e-05, "loss": 1.105, "step": 5180 }, { "epoch": 0.3510400433633715, "grad_norm": 7.821188926696777, "learning_rate": 9.393524539667329e-05, "loss": 0.8106, "step": 5181 }, { "epoch": 0.35110779863134356, "grad_norm": 7.235447406768799, "learning_rate": 9.393387637757547e-05, "loss": 0.9304, "step": 5182 }, { "epoch": 0.35117555389931565, "grad_norm": 6.859196662902832, "learning_rate": 9.393250735847765e-05, "loss": 0.7797, "step": 5183 }, { "epoch": 0.35124330916728774, "grad_norm": 8.33879566192627, "learning_rate": 9.393113833937985e-05, "loss": 1.0075, "step": 5184 }, { "epoch": 0.35131106443525983, "grad_norm": 6.733922004699707, "learning_rate": 9.392976932028203e-05, "loss": 0.7304, "step": 5185 }, { "epoch": 0.3513788197032319, "grad_norm": 8.839714050292969, "learning_rate": 9.392840030118421e-05, "loss": 0.8852, "step": 5186 }, { "epoch": 0.351446574971204, "grad_norm": 9.954099655151367, "learning_rate": 9.392703128208639e-05, "loss": 1.1631, "step": 5187 }, { "epoch": 0.3515143302391761, "grad_norm": 8.042675971984863, "learning_rate": 9.392566226298858e-05, "loss": 0.8741, "step": 5188 }, { "epoch": 0.3515820855071482, "grad_norm": 6.661304950714111, "learning_rate": 9.392429324389076e-05, "loss": 0.8879, "step": 5189 }, { "epoch": 0.3516498407751203, "grad_norm": 6.568962097167969, "learning_rate": 9.392292422479294e-05, "loss": 0.6823, "step": 5190 }, { "epoch": 0.35171759604309233, "grad_norm": 6.836343288421631, "learning_rate": 9.392155520569512e-05, "loss": 0.8174, "step": 5191 }, { "epoch": 0.3517853513110644, "grad_norm": 7.555830478668213, "learning_rate": 9.39201861865973e-05, "loss": 0.803, "step": 5192 }, { "epoch": 0.3518531065790365, "grad_norm": 7.264036178588867, "learning_rate": 9.39188171674995e-05, "loss": 1.0786, "step": 5193 }, { "epoch": 0.3519208618470086, "grad_norm": 5.6876420974731445, "learning_rate": 9.391744814840168e-05, "loss": 0.7553, "step": 5194 }, { "epoch": 0.3519886171149807, "grad_norm": 8.890271186828613, "learning_rate": 9.391607912930386e-05, "loss": 1.0685, "step": 5195 }, { "epoch": 0.3520563723829528, "grad_norm": 6.010801315307617, "learning_rate": 9.391471011020604e-05, "loss": 0.7915, "step": 5196 }, { "epoch": 0.3521241276509249, "grad_norm": 8.496731758117676, "learning_rate": 9.391334109110822e-05, "loss": 1.2103, "step": 5197 }, { "epoch": 0.35219188291889697, "grad_norm": 8.196046829223633, "learning_rate": 9.391197207201041e-05, "loss": 1.0173, "step": 5198 }, { "epoch": 0.352259638186869, "grad_norm": 8.809300422668457, "learning_rate": 9.391060305291259e-05, "loss": 0.906, "step": 5199 }, { "epoch": 0.3523273934548411, "grad_norm": 8.437650680541992, "learning_rate": 9.390923403381477e-05, "loss": 0.8865, "step": 5200 }, { "epoch": 0.3523951487228132, "grad_norm": 8.659364700317383, "learning_rate": 9.390786501471695e-05, "loss": 0.8878, "step": 5201 }, { "epoch": 0.3524629039907853, "grad_norm": 6.658010482788086, "learning_rate": 9.390649599561915e-05, "loss": 0.9141, "step": 5202 }, { "epoch": 0.3525306592587574, "grad_norm": 6.702786445617676, "learning_rate": 9.390512697652133e-05, "loss": 0.8193, "step": 5203 }, { "epoch": 0.35259841452672946, "grad_norm": 8.0852632522583, "learning_rate": 9.390375795742351e-05, "loss": 0.9655, "step": 5204 }, { "epoch": 0.35266616979470156, "grad_norm": 6.598480224609375, "learning_rate": 9.390238893832569e-05, "loss": 0.8858, "step": 5205 }, { "epoch": 0.35273392506267365, "grad_norm": 8.510173797607422, "learning_rate": 9.390101991922787e-05, "loss": 1.0543, "step": 5206 }, { "epoch": 0.3528016803306457, "grad_norm": 7.498827934265137, "learning_rate": 9.389965090013006e-05, "loss": 1.138, "step": 5207 }, { "epoch": 0.3528694355986178, "grad_norm": 6.370652675628662, "learning_rate": 9.389828188103224e-05, "loss": 0.996, "step": 5208 }, { "epoch": 0.35293719086658987, "grad_norm": 6.830577373504639, "learning_rate": 9.389691286193442e-05, "loss": 0.7562, "step": 5209 }, { "epoch": 0.35300494613456196, "grad_norm": 7.01290225982666, "learning_rate": 9.38955438428366e-05, "loss": 0.8141, "step": 5210 }, { "epoch": 0.35307270140253405, "grad_norm": 6.957060813903809, "learning_rate": 9.38941748237388e-05, "loss": 1.1923, "step": 5211 }, { "epoch": 0.35314045667050614, "grad_norm": 5.929959297180176, "learning_rate": 9.389280580464098e-05, "loss": 0.7928, "step": 5212 }, { "epoch": 0.35320821193847823, "grad_norm": 8.654718399047852, "learning_rate": 9.389143678554316e-05, "loss": 1.0896, "step": 5213 }, { "epoch": 0.3532759672064503, "grad_norm": 5.981748580932617, "learning_rate": 9.389006776644534e-05, "loss": 0.9777, "step": 5214 }, { "epoch": 0.35334372247442236, "grad_norm": 6.251374244689941, "learning_rate": 9.388869874734752e-05, "loss": 0.8562, "step": 5215 }, { "epoch": 0.35341147774239445, "grad_norm": 7.049698352813721, "learning_rate": 9.388732972824971e-05, "loss": 0.8297, "step": 5216 }, { "epoch": 0.35347923301036654, "grad_norm": 8.04577350616455, "learning_rate": 9.38859607091519e-05, "loss": 1.2035, "step": 5217 }, { "epoch": 0.35354698827833864, "grad_norm": 6.800088405609131, "learning_rate": 9.388459169005407e-05, "loss": 0.9551, "step": 5218 }, { "epoch": 0.35361474354631073, "grad_norm": 7.368710517883301, "learning_rate": 9.388322267095627e-05, "loss": 0.8174, "step": 5219 }, { "epoch": 0.3536824988142828, "grad_norm": 7.573550701141357, "learning_rate": 9.388185365185845e-05, "loss": 1.0921, "step": 5220 }, { "epoch": 0.3537502540822549, "grad_norm": 8.282757759094238, "learning_rate": 9.388048463276063e-05, "loss": 0.9124, "step": 5221 }, { "epoch": 0.353818009350227, "grad_norm": 8.21078872680664, "learning_rate": 9.387911561366282e-05, "loss": 1.0818, "step": 5222 }, { "epoch": 0.35388576461819904, "grad_norm": 7.401734352111816, "learning_rate": 9.3877746594565e-05, "loss": 0.9906, "step": 5223 }, { "epoch": 0.35395351988617113, "grad_norm": 8.078129768371582, "learning_rate": 9.387637757546718e-05, "loss": 1.1233, "step": 5224 }, { "epoch": 0.3540212751541432, "grad_norm": 7.231998920440674, "learning_rate": 9.387500855636938e-05, "loss": 0.8412, "step": 5225 }, { "epoch": 0.3540890304221153, "grad_norm": 8.709794044494629, "learning_rate": 9.387363953727156e-05, "loss": 0.8346, "step": 5226 }, { "epoch": 0.3541567856900874, "grad_norm": 8.45758056640625, "learning_rate": 9.387227051817374e-05, "loss": 1.0201, "step": 5227 }, { "epoch": 0.3542245409580595, "grad_norm": 5.156687259674072, "learning_rate": 9.387090149907592e-05, "loss": 0.8029, "step": 5228 }, { "epoch": 0.3542922962260316, "grad_norm": 6.2013163566589355, "learning_rate": 9.38695324799781e-05, "loss": 0.8984, "step": 5229 }, { "epoch": 0.3543600514940037, "grad_norm": 7.414056301116943, "learning_rate": 9.386816346088029e-05, "loss": 0.8606, "step": 5230 }, { "epoch": 0.3544278067619757, "grad_norm": 8.167763710021973, "learning_rate": 9.386679444178247e-05, "loss": 0.973, "step": 5231 }, { "epoch": 0.3544955620299478, "grad_norm": 7.290304183959961, "learning_rate": 9.386542542268465e-05, "loss": 0.8988, "step": 5232 }, { "epoch": 0.3545633172979199, "grad_norm": 7.181061744689941, "learning_rate": 9.386405640358683e-05, "loss": 0.8689, "step": 5233 }, { "epoch": 0.354631072565892, "grad_norm": 8.87985610961914, "learning_rate": 9.386268738448903e-05, "loss": 1.0076, "step": 5234 }, { "epoch": 0.3546988278338641, "grad_norm": 6.205500602722168, "learning_rate": 9.386131836539121e-05, "loss": 0.8101, "step": 5235 }, { "epoch": 0.3547665831018362, "grad_norm": 7.425187587738037, "learning_rate": 9.385994934629339e-05, "loss": 0.9122, "step": 5236 }, { "epoch": 0.35483433836980827, "grad_norm": 9.541454315185547, "learning_rate": 9.385858032719557e-05, "loss": 0.9868, "step": 5237 }, { "epoch": 0.35490209363778036, "grad_norm": 6.853835582733154, "learning_rate": 9.385721130809775e-05, "loss": 0.9506, "step": 5238 }, { "epoch": 0.35496984890575245, "grad_norm": 9.07067584991455, "learning_rate": 9.385584228899994e-05, "loss": 0.9146, "step": 5239 }, { "epoch": 0.3550376041737245, "grad_norm": 7.5352373123168945, "learning_rate": 9.385447326990212e-05, "loss": 1.0089, "step": 5240 }, { "epoch": 0.3551053594416966, "grad_norm": 5.626684188842773, "learning_rate": 9.38531042508043e-05, "loss": 0.7057, "step": 5241 }, { "epoch": 0.35517311470966867, "grad_norm": 6.1156768798828125, "learning_rate": 9.385173523170648e-05, "loss": 0.808, "step": 5242 }, { "epoch": 0.35524086997764076, "grad_norm": 7.227247714996338, "learning_rate": 9.385036621260868e-05, "loss": 0.9477, "step": 5243 }, { "epoch": 0.35530862524561285, "grad_norm": 7.034331798553467, "learning_rate": 9.384899719351086e-05, "loss": 0.7929, "step": 5244 }, { "epoch": 0.35537638051358494, "grad_norm": 5.943993091583252, "learning_rate": 9.384762817441304e-05, "loss": 0.8886, "step": 5245 }, { "epoch": 0.35544413578155704, "grad_norm": 7.859194278717041, "learning_rate": 9.384625915531522e-05, "loss": 1.0028, "step": 5246 }, { "epoch": 0.3555118910495291, "grad_norm": 6.8557448387146, "learning_rate": 9.38448901362174e-05, "loss": 0.8946, "step": 5247 }, { "epoch": 0.35557964631750116, "grad_norm": 6.733648777008057, "learning_rate": 9.384352111711959e-05, "loss": 0.9186, "step": 5248 }, { "epoch": 0.35564740158547326, "grad_norm": 5.929673194885254, "learning_rate": 9.384215209802177e-05, "loss": 0.8132, "step": 5249 }, { "epoch": 0.35571515685344535, "grad_norm": 6.680620193481445, "learning_rate": 9.384078307892395e-05, "loss": 0.9197, "step": 5250 }, { "epoch": 0.35578291212141744, "grad_norm": 8.043455123901367, "learning_rate": 9.383941405982613e-05, "loss": 1.0841, "step": 5251 }, { "epoch": 0.35585066738938953, "grad_norm": 9.225475311279297, "learning_rate": 9.383804504072831e-05, "loss": 1.3691, "step": 5252 }, { "epoch": 0.3559184226573616, "grad_norm": 8.588724136352539, "learning_rate": 9.383667602163051e-05, "loss": 1.1282, "step": 5253 }, { "epoch": 0.3559861779253337, "grad_norm": 8.21908950805664, "learning_rate": 9.383530700253269e-05, "loss": 0.7606, "step": 5254 }, { "epoch": 0.3560539331933058, "grad_norm": 6.107995510101318, "learning_rate": 9.383393798343487e-05, "loss": 1.1818, "step": 5255 }, { "epoch": 0.35612168846127784, "grad_norm": 6.612033367156982, "learning_rate": 9.383256896433705e-05, "loss": 0.8892, "step": 5256 }, { "epoch": 0.35618944372924993, "grad_norm": 6.935641288757324, "learning_rate": 9.383119994523924e-05, "loss": 0.8565, "step": 5257 }, { "epoch": 0.356257198997222, "grad_norm": 7.328373908996582, "learning_rate": 9.382983092614142e-05, "loss": 0.9409, "step": 5258 }, { "epoch": 0.3563249542651941, "grad_norm": 7.016412734985352, "learning_rate": 9.38284619070436e-05, "loss": 0.8768, "step": 5259 }, { "epoch": 0.3563927095331662, "grad_norm": 7.122356414794922, "learning_rate": 9.382709288794578e-05, "loss": 0.9048, "step": 5260 }, { "epoch": 0.3564604648011383, "grad_norm": 7.590730667114258, "learning_rate": 9.382572386884796e-05, "loss": 0.8154, "step": 5261 }, { "epoch": 0.3565282200691104, "grad_norm": 5.5446858406066895, "learning_rate": 9.382435484975016e-05, "loss": 0.8783, "step": 5262 }, { "epoch": 0.3565959753370825, "grad_norm": 7.153842449188232, "learning_rate": 9.382298583065234e-05, "loss": 1.2412, "step": 5263 }, { "epoch": 0.3566637306050545, "grad_norm": 7.115203857421875, "learning_rate": 9.382161681155452e-05, "loss": 1.0302, "step": 5264 }, { "epoch": 0.3567314858730266, "grad_norm": 5.7724833488464355, "learning_rate": 9.382024779245671e-05, "loss": 0.5839, "step": 5265 }, { "epoch": 0.3567992411409987, "grad_norm": 7.793944358825684, "learning_rate": 9.381887877335889e-05, "loss": 1.0893, "step": 5266 }, { "epoch": 0.3568669964089708, "grad_norm": 6.561145782470703, "learning_rate": 9.381750975426107e-05, "loss": 0.911, "step": 5267 }, { "epoch": 0.3569347516769429, "grad_norm": 6.053153038024902, "learning_rate": 9.381614073516327e-05, "loss": 0.9239, "step": 5268 }, { "epoch": 0.357002506944915, "grad_norm": 7.162718772888184, "learning_rate": 9.381477171606545e-05, "loss": 0.7193, "step": 5269 }, { "epoch": 0.35707026221288707, "grad_norm": 7.634250640869141, "learning_rate": 9.381340269696763e-05, "loss": 0.8826, "step": 5270 }, { "epoch": 0.35713801748085916, "grad_norm": 6.502168655395508, "learning_rate": 9.381203367786982e-05, "loss": 0.8278, "step": 5271 }, { "epoch": 0.3572057727488312, "grad_norm": 7.339065074920654, "learning_rate": 9.3810664658772e-05, "loss": 0.8818, "step": 5272 }, { "epoch": 0.3572735280168033, "grad_norm": 8.079582214355469, "learning_rate": 9.380929563967418e-05, "loss": 0.8596, "step": 5273 }, { "epoch": 0.3573412832847754, "grad_norm": 6.368169784545898, "learning_rate": 9.380792662057636e-05, "loss": 0.7352, "step": 5274 }, { "epoch": 0.35740903855274747, "grad_norm": 7.813303470611572, "learning_rate": 9.380655760147854e-05, "loss": 0.876, "step": 5275 }, { "epoch": 0.35747679382071956, "grad_norm": 8.180051803588867, "learning_rate": 9.380518858238074e-05, "loss": 0.8893, "step": 5276 }, { "epoch": 0.35754454908869165, "grad_norm": 7.618046283721924, "learning_rate": 9.380381956328292e-05, "loss": 0.9647, "step": 5277 }, { "epoch": 0.35761230435666375, "grad_norm": 7.507559776306152, "learning_rate": 9.38024505441851e-05, "loss": 0.9524, "step": 5278 }, { "epoch": 0.35768005962463584, "grad_norm": 9.470221519470215, "learning_rate": 9.380108152508728e-05, "loss": 1.0057, "step": 5279 }, { "epoch": 0.3577478148926079, "grad_norm": 6.612621784210205, "learning_rate": 9.379971250598947e-05, "loss": 0.8489, "step": 5280 }, { "epoch": 0.35781557016057997, "grad_norm": 7.146090030670166, "learning_rate": 9.379834348689165e-05, "loss": 0.9585, "step": 5281 }, { "epoch": 0.35788332542855206, "grad_norm": 6.300119400024414, "learning_rate": 9.379697446779383e-05, "loss": 0.881, "step": 5282 }, { "epoch": 0.35795108069652415, "grad_norm": 7.285689353942871, "learning_rate": 9.379560544869601e-05, "loss": 1.0182, "step": 5283 }, { "epoch": 0.35801883596449624, "grad_norm": 8.943527221679688, "learning_rate": 9.379423642959819e-05, "loss": 0.9782, "step": 5284 }, { "epoch": 0.35808659123246833, "grad_norm": 8.551790237426758, "learning_rate": 9.379286741050039e-05, "loss": 0.8513, "step": 5285 }, { "epoch": 0.3581543465004404, "grad_norm": 8.670862197875977, "learning_rate": 9.379149839140257e-05, "loss": 0.9026, "step": 5286 }, { "epoch": 0.3582221017684125, "grad_norm": 8.31614875793457, "learning_rate": 9.379012937230475e-05, "loss": 0.7666, "step": 5287 }, { "epoch": 0.35828985703638455, "grad_norm": 6.347659111022949, "learning_rate": 9.378876035320693e-05, "loss": 0.953, "step": 5288 }, { "epoch": 0.35835761230435664, "grad_norm": 9.098658561706543, "learning_rate": 9.378739133410912e-05, "loss": 1.3235, "step": 5289 }, { "epoch": 0.35842536757232873, "grad_norm": 7.218830585479736, "learning_rate": 9.37860223150113e-05, "loss": 1.0008, "step": 5290 }, { "epoch": 0.3584931228403008, "grad_norm": 7.346166133880615, "learning_rate": 9.378465329591348e-05, "loss": 0.8441, "step": 5291 }, { "epoch": 0.3585608781082729, "grad_norm": 6.566136360168457, "learning_rate": 9.378328427681566e-05, "loss": 1.0381, "step": 5292 }, { "epoch": 0.358628633376245, "grad_norm": 7.599377632141113, "learning_rate": 9.378191525771784e-05, "loss": 0.8131, "step": 5293 }, { "epoch": 0.3586963886442171, "grad_norm": 6.4648284912109375, "learning_rate": 9.378054623862004e-05, "loss": 0.9599, "step": 5294 }, { "epoch": 0.3587641439121892, "grad_norm": 6.155094146728516, "learning_rate": 9.377917721952222e-05, "loss": 1.0911, "step": 5295 }, { "epoch": 0.3588318991801613, "grad_norm": 5.981008529663086, "learning_rate": 9.37778082004244e-05, "loss": 0.7692, "step": 5296 }, { "epoch": 0.3588996544481333, "grad_norm": 7.424664497375488, "learning_rate": 9.377643918132658e-05, "loss": 0.9248, "step": 5297 }, { "epoch": 0.3589674097161054, "grad_norm": 8.310460090637207, "learning_rate": 9.377507016222877e-05, "loss": 0.9659, "step": 5298 }, { "epoch": 0.3590351649840775, "grad_norm": 7.189448833465576, "learning_rate": 9.377370114313095e-05, "loss": 0.9844, "step": 5299 }, { "epoch": 0.3591029202520496, "grad_norm": 6.688578128814697, "learning_rate": 9.377233212403313e-05, "loss": 0.8488, "step": 5300 }, { "epoch": 0.3591706755200217, "grad_norm": 8.917186737060547, "learning_rate": 9.377096310493531e-05, "loss": 1.0579, "step": 5301 }, { "epoch": 0.3592384307879938, "grad_norm": 8.004783630371094, "learning_rate": 9.37695940858375e-05, "loss": 0.8347, "step": 5302 }, { "epoch": 0.35930618605596587, "grad_norm": 8.461618423461914, "learning_rate": 9.376822506673969e-05, "loss": 1.1304, "step": 5303 }, { "epoch": 0.35937394132393796, "grad_norm": 7.35521936416626, "learning_rate": 9.376685604764187e-05, "loss": 0.8222, "step": 5304 }, { "epoch": 0.35944169659191, "grad_norm": 7.697165012359619, "learning_rate": 9.376548702854405e-05, "loss": 0.8885, "step": 5305 }, { "epoch": 0.3595094518598821, "grad_norm": 8.730046272277832, "learning_rate": 9.376411800944623e-05, "loss": 0.9201, "step": 5306 }, { "epoch": 0.3595772071278542, "grad_norm": 7.738508701324463, "learning_rate": 9.376274899034841e-05, "loss": 0.828, "step": 5307 }, { "epoch": 0.3596449623958263, "grad_norm": 6.961246013641357, "learning_rate": 9.37613799712506e-05, "loss": 1.2298, "step": 5308 }, { "epoch": 0.35971271766379836, "grad_norm": 7.723383903503418, "learning_rate": 9.376001095215278e-05, "loss": 0.9714, "step": 5309 }, { "epoch": 0.35978047293177046, "grad_norm": 6.885270595550537, "learning_rate": 9.375864193305496e-05, "loss": 0.8352, "step": 5310 }, { "epoch": 0.35984822819974255, "grad_norm": 8.623848915100098, "learning_rate": 9.375727291395716e-05, "loss": 0.7216, "step": 5311 }, { "epoch": 0.35991598346771464, "grad_norm": 7.042871952056885, "learning_rate": 9.375590389485934e-05, "loss": 0.8098, "step": 5312 }, { "epoch": 0.3599837387356867, "grad_norm": 6.125864505767822, "learning_rate": 9.375453487576152e-05, "loss": 0.7628, "step": 5313 }, { "epoch": 0.36005149400365877, "grad_norm": 7.560012340545654, "learning_rate": 9.375316585666371e-05, "loss": 0.9114, "step": 5314 }, { "epoch": 0.36011924927163086, "grad_norm": 6.125675678253174, "learning_rate": 9.375179683756589e-05, "loss": 0.7894, "step": 5315 }, { "epoch": 0.36018700453960295, "grad_norm": 7.076975345611572, "learning_rate": 9.375042781846807e-05, "loss": 1.1084, "step": 5316 }, { "epoch": 0.36025475980757504, "grad_norm": 6.712325096130371, "learning_rate": 9.374905879937027e-05, "loss": 0.9326, "step": 5317 }, { "epoch": 0.36032251507554713, "grad_norm": 5.849967956542969, "learning_rate": 9.374768978027245e-05, "loss": 0.7884, "step": 5318 }, { "epoch": 0.3603902703435192, "grad_norm": 7.5400614738464355, "learning_rate": 9.374632076117463e-05, "loss": 0.9844, "step": 5319 }, { "epoch": 0.3604580256114913, "grad_norm": 5.720027446746826, "learning_rate": 9.374495174207681e-05, "loss": 0.8233, "step": 5320 }, { "epoch": 0.36052578087946335, "grad_norm": 6.872372627258301, "learning_rate": 9.3743582722979e-05, "loss": 0.6808, "step": 5321 }, { "epoch": 0.36059353614743545, "grad_norm": 6.98433780670166, "learning_rate": 9.374221370388118e-05, "loss": 0.642, "step": 5322 }, { "epoch": 0.36066129141540754, "grad_norm": 7.231049537658691, "learning_rate": 9.374084468478336e-05, "loss": 1.1379, "step": 5323 }, { "epoch": 0.36072904668337963, "grad_norm": 8.301416397094727, "learning_rate": 9.373947566568554e-05, "loss": 0.9184, "step": 5324 }, { "epoch": 0.3607968019513517, "grad_norm": 6.9752068519592285, "learning_rate": 9.373810664658772e-05, "loss": 0.9984, "step": 5325 }, { "epoch": 0.3608645572193238, "grad_norm": 6.826324939727783, "learning_rate": 9.373673762748992e-05, "loss": 0.8923, "step": 5326 }, { "epoch": 0.3609323124872959, "grad_norm": 6.605769634246826, "learning_rate": 9.37353686083921e-05, "loss": 0.8809, "step": 5327 }, { "epoch": 0.361000067755268, "grad_norm": 7.791666030883789, "learning_rate": 9.373399958929428e-05, "loss": 0.9199, "step": 5328 }, { "epoch": 0.36106782302324003, "grad_norm": 7.015176773071289, "learning_rate": 9.373263057019646e-05, "loss": 0.8271, "step": 5329 }, { "epoch": 0.3611355782912121, "grad_norm": 6.63329553604126, "learning_rate": 9.373126155109864e-05, "loss": 0.9052, "step": 5330 }, { "epoch": 0.3612033335591842, "grad_norm": 9.516122817993164, "learning_rate": 9.372989253200083e-05, "loss": 1.1517, "step": 5331 }, { "epoch": 0.3612710888271563, "grad_norm": 7.7722978591918945, "learning_rate": 9.372852351290301e-05, "loss": 1.0195, "step": 5332 }, { "epoch": 0.3613388440951284, "grad_norm": 7.2577338218688965, "learning_rate": 9.372715449380519e-05, "loss": 0.9093, "step": 5333 }, { "epoch": 0.3614065993631005, "grad_norm": 6.668231010437012, "learning_rate": 9.372578547470737e-05, "loss": 0.7677, "step": 5334 }, { "epoch": 0.3614743546310726, "grad_norm": 8.366500854492188, "learning_rate": 9.372441645560957e-05, "loss": 1.0869, "step": 5335 }, { "epoch": 0.3615421098990447, "grad_norm": 7.495104789733887, "learning_rate": 9.372304743651175e-05, "loss": 0.8074, "step": 5336 }, { "epoch": 0.3616098651670167, "grad_norm": 6.546140670776367, "learning_rate": 9.372167841741393e-05, "loss": 0.8792, "step": 5337 }, { "epoch": 0.3616776204349888, "grad_norm": 7.664621353149414, "learning_rate": 9.372030939831611e-05, "loss": 0.8177, "step": 5338 }, { "epoch": 0.3617453757029609, "grad_norm": 7.482497215270996, "learning_rate": 9.371894037921829e-05, "loss": 0.7624, "step": 5339 }, { "epoch": 0.361813130970933, "grad_norm": 11.148695945739746, "learning_rate": 9.371757136012048e-05, "loss": 1.0792, "step": 5340 }, { "epoch": 0.3618808862389051, "grad_norm": 9.08121109008789, "learning_rate": 9.371620234102266e-05, "loss": 1.0027, "step": 5341 }, { "epoch": 0.36194864150687717, "grad_norm": 8.424308776855469, "learning_rate": 9.371483332192484e-05, "loss": 1.2137, "step": 5342 }, { "epoch": 0.36201639677484926, "grad_norm": 8.55742359161377, "learning_rate": 9.371346430282702e-05, "loss": 0.8023, "step": 5343 }, { "epoch": 0.36208415204282135, "grad_norm": 7.197443008422852, "learning_rate": 9.371209528372922e-05, "loss": 1.0458, "step": 5344 }, { "epoch": 0.36215190731079344, "grad_norm": 8.317964553833008, "learning_rate": 9.37107262646314e-05, "loss": 0.9829, "step": 5345 }, { "epoch": 0.3622196625787655, "grad_norm": 8.819951057434082, "learning_rate": 9.370935724553358e-05, "loss": 1.0024, "step": 5346 }, { "epoch": 0.36228741784673757, "grad_norm": 6.795368671417236, "learning_rate": 9.370798822643576e-05, "loss": 1.0216, "step": 5347 }, { "epoch": 0.36235517311470966, "grad_norm": 5.370431900024414, "learning_rate": 9.370661920733794e-05, "loss": 0.8693, "step": 5348 }, { "epoch": 0.36242292838268175, "grad_norm": 7.330100059509277, "learning_rate": 9.370525018824013e-05, "loss": 1.0576, "step": 5349 }, { "epoch": 0.36249068365065384, "grad_norm": 6.56719446182251, "learning_rate": 9.370388116914231e-05, "loss": 0.8367, "step": 5350 }, { "epoch": 0.36255843891862594, "grad_norm": 7.462427616119385, "learning_rate": 9.370251215004449e-05, "loss": 0.8284, "step": 5351 }, { "epoch": 0.362626194186598, "grad_norm": 10.382340431213379, "learning_rate": 9.370114313094667e-05, "loss": 1.1184, "step": 5352 }, { "epoch": 0.3626939494545701, "grad_norm": 6.872713088989258, "learning_rate": 9.369977411184885e-05, "loss": 0.9656, "step": 5353 }, { "epoch": 0.36276170472254216, "grad_norm": 6.2565741539001465, "learning_rate": 9.369840509275105e-05, "loss": 0.9206, "step": 5354 }, { "epoch": 0.36282945999051425, "grad_norm": 8.036307334899902, "learning_rate": 9.369703607365323e-05, "loss": 1.2556, "step": 5355 }, { "epoch": 0.36289721525848634, "grad_norm": 7.401971340179443, "learning_rate": 9.369566705455541e-05, "loss": 0.9584, "step": 5356 }, { "epoch": 0.36296497052645843, "grad_norm": 7.890522003173828, "learning_rate": 9.369429803545759e-05, "loss": 1.0172, "step": 5357 }, { "epoch": 0.3630327257944305, "grad_norm": 7.4986348152160645, "learning_rate": 9.369292901635978e-05, "loss": 0.9985, "step": 5358 }, { "epoch": 0.3631004810624026, "grad_norm": 6.275259494781494, "learning_rate": 9.369155999726196e-05, "loss": 0.9051, "step": 5359 }, { "epoch": 0.3631682363303747, "grad_norm": 6.2667365074157715, "learning_rate": 9.369019097816416e-05, "loss": 0.9137, "step": 5360 }, { "epoch": 0.3632359915983468, "grad_norm": 7.5247483253479, "learning_rate": 9.368882195906634e-05, "loss": 0.9157, "step": 5361 }, { "epoch": 0.36330374686631883, "grad_norm": 8.757854461669922, "learning_rate": 9.368745293996852e-05, "loss": 1.2596, "step": 5362 }, { "epoch": 0.3633715021342909, "grad_norm": 7.257226943969727, "learning_rate": 9.368608392087071e-05, "loss": 0.8211, "step": 5363 }, { "epoch": 0.363439257402263, "grad_norm": 6.194060325622559, "learning_rate": 9.368471490177289e-05, "loss": 0.8816, "step": 5364 }, { "epoch": 0.3635070126702351, "grad_norm": 10.445760726928711, "learning_rate": 9.368334588267507e-05, "loss": 0.858, "step": 5365 }, { "epoch": 0.3635747679382072, "grad_norm": 6.156910419464111, "learning_rate": 9.368197686357725e-05, "loss": 1.0348, "step": 5366 }, { "epoch": 0.3636425232061793, "grad_norm": 6.490479946136475, "learning_rate": 9.368060784447945e-05, "loss": 0.8834, "step": 5367 }, { "epoch": 0.3637102784741514, "grad_norm": 6.535744667053223, "learning_rate": 9.367923882538163e-05, "loss": 0.6612, "step": 5368 }, { "epoch": 0.3637780337421235, "grad_norm": 8.068840980529785, "learning_rate": 9.36778698062838e-05, "loss": 1.0348, "step": 5369 }, { "epoch": 0.3638457890100955, "grad_norm": 6.666348934173584, "learning_rate": 9.367650078718599e-05, "loss": 0.8346, "step": 5370 }, { "epoch": 0.3639135442780676, "grad_norm": 7.482582092285156, "learning_rate": 9.367513176808817e-05, "loss": 0.7356, "step": 5371 }, { "epoch": 0.3639812995460397, "grad_norm": 6.928842067718506, "learning_rate": 9.367376274899036e-05, "loss": 0.7963, "step": 5372 }, { "epoch": 0.3640490548140118, "grad_norm": 6.574978351593018, "learning_rate": 9.367239372989254e-05, "loss": 0.7893, "step": 5373 }, { "epoch": 0.3641168100819839, "grad_norm": 6.267110824584961, "learning_rate": 9.367102471079472e-05, "loss": 0.6282, "step": 5374 }, { "epoch": 0.36418456534995597, "grad_norm": 6.368738651275635, "learning_rate": 9.36696556916969e-05, "loss": 0.8983, "step": 5375 }, { "epoch": 0.36425232061792806, "grad_norm": 7.211498737335205, "learning_rate": 9.36682866725991e-05, "loss": 0.9471, "step": 5376 }, { "epoch": 0.36432007588590015, "grad_norm": 6.124608516693115, "learning_rate": 9.366691765350128e-05, "loss": 0.8649, "step": 5377 }, { "epoch": 0.3643878311538722, "grad_norm": 7.325745582580566, "learning_rate": 9.366554863440346e-05, "loss": 0.9811, "step": 5378 }, { "epoch": 0.3644555864218443, "grad_norm": 5.563783645629883, "learning_rate": 9.366417961530564e-05, "loss": 0.833, "step": 5379 }, { "epoch": 0.36452334168981637, "grad_norm": 7.581454753875732, "learning_rate": 9.366281059620782e-05, "loss": 0.9078, "step": 5380 }, { "epoch": 0.36459109695778846, "grad_norm": 7.746644020080566, "learning_rate": 9.366144157711001e-05, "loss": 0.9143, "step": 5381 }, { "epoch": 0.36465885222576055, "grad_norm": 7.729243278503418, "learning_rate": 9.366007255801219e-05, "loss": 0.856, "step": 5382 }, { "epoch": 0.36472660749373265, "grad_norm": 7.337301731109619, "learning_rate": 9.365870353891437e-05, "loss": 0.9059, "step": 5383 }, { "epoch": 0.36479436276170474, "grad_norm": 7.622936248779297, "learning_rate": 9.365733451981655e-05, "loss": 1.0493, "step": 5384 }, { "epoch": 0.36486211802967683, "grad_norm": 7.159453392028809, "learning_rate": 9.365596550071873e-05, "loss": 0.824, "step": 5385 }, { "epoch": 0.36492987329764887, "grad_norm": 7.111423492431641, "learning_rate": 9.365459648162093e-05, "loss": 0.7201, "step": 5386 }, { "epoch": 0.36499762856562096, "grad_norm": 8.7478666305542, "learning_rate": 9.365322746252311e-05, "loss": 1.2497, "step": 5387 }, { "epoch": 0.36506538383359305, "grad_norm": 9.522677421569824, "learning_rate": 9.365185844342529e-05, "loss": 0.9175, "step": 5388 }, { "epoch": 0.36513313910156514, "grad_norm": 7.881449222564697, "learning_rate": 9.365048942432747e-05, "loss": 1.1123, "step": 5389 }, { "epoch": 0.36520089436953723, "grad_norm": 6.60174560546875, "learning_rate": 9.364912040522966e-05, "loss": 0.9148, "step": 5390 }, { "epoch": 0.3652686496375093, "grad_norm": 7.549520492553711, "learning_rate": 9.364775138613184e-05, "loss": 0.991, "step": 5391 }, { "epoch": 0.3653364049054814, "grad_norm": 7.287946701049805, "learning_rate": 9.364638236703402e-05, "loss": 0.984, "step": 5392 }, { "epoch": 0.3654041601734535, "grad_norm": 9.351056098937988, "learning_rate": 9.36450133479362e-05, "loss": 1.1052, "step": 5393 }, { "epoch": 0.36547191544142554, "grad_norm": 8.443605422973633, "learning_rate": 9.364364432883838e-05, "loss": 0.941, "step": 5394 }, { "epoch": 0.36553967070939763, "grad_norm": 6.416137218475342, "learning_rate": 9.364227530974058e-05, "loss": 0.9359, "step": 5395 }, { "epoch": 0.3656074259773697, "grad_norm": 7.125934600830078, "learning_rate": 9.364090629064276e-05, "loss": 1.0006, "step": 5396 }, { "epoch": 0.3656751812453418, "grad_norm": 8.161697387695312, "learning_rate": 9.363953727154494e-05, "loss": 0.8284, "step": 5397 }, { "epoch": 0.3657429365133139, "grad_norm": 6.955849647521973, "learning_rate": 9.363816825244712e-05, "loss": 0.7467, "step": 5398 }, { "epoch": 0.365810691781286, "grad_norm": 6.489741325378418, "learning_rate": 9.363679923334931e-05, "loss": 0.7611, "step": 5399 }, { "epoch": 0.3658784470492581, "grad_norm": 7.332844257354736, "learning_rate": 9.363543021425149e-05, "loss": 0.7666, "step": 5400 }, { "epoch": 0.3659462023172302, "grad_norm": 9.001418113708496, "learning_rate": 9.363406119515367e-05, "loss": 1.0219, "step": 5401 }, { "epoch": 0.3660139575852023, "grad_norm": 8.972160339355469, "learning_rate": 9.363269217605585e-05, "loss": 1.0122, "step": 5402 }, { "epoch": 0.3660817128531743, "grad_norm": 7.487551212310791, "learning_rate": 9.363132315695803e-05, "loss": 1.2551, "step": 5403 }, { "epoch": 0.3661494681211464, "grad_norm": 5.9293904304504395, "learning_rate": 9.362995413786023e-05, "loss": 0.7594, "step": 5404 }, { "epoch": 0.3662172233891185, "grad_norm": 7.69260311126709, "learning_rate": 9.362858511876241e-05, "loss": 0.9297, "step": 5405 }, { "epoch": 0.3662849786570906, "grad_norm": 6.723849296569824, "learning_rate": 9.362721609966459e-05, "loss": 1.0034, "step": 5406 }, { "epoch": 0.3663527339250627, "grad_norm": 8.467827796936035, "learning_rate": 9.362584708056678e-05, "loss": 0.9831, "step": 5407 }, { "epoch": 0.36642048919303477, "grad_norm": 6.563052654266357, "learning_rate": 9.362447806146896e-05, "loss": 0.8464, "step": 5408 }, { "epoch": 0.36648824446100686, "grad_norm": 7.124545574188232, "learning_rate": 9.362310904237114e-05, "loss": 1.1, "step": 5409 }, { "epoch": 0.36655599972897895, "grad_norm": 8.016179084777832, "learning_rate": 9.362174002327334e-05, "loss": 0.9962, "step": 5410 }, { "epoch": 0.366623754996951, "grad_norm": 5.937708854675293, "learning_rate": 9.362037100417552e-05, "loss": 0.5908, "step": 5411 }, { "epoch": 0.3666915102649231, "grad_norm": 6.945435523986816, "learning_rate": 9.36190019850777e-05, "loss": 0.7576, "step": 5412 }, { "epoch": 0.3667592655328952, "grad_norm": 6.756751537322998, "learning_rate": 9.361763296597989e-05, "loss": 0.7966, "step": 5413 }, { "epoch": 0.36682702080086727, "grad_norm": 6.434855937957764, "learning_rate": 9.361626394688207e-05, "loss": 0.8401, "step": 5414 }, { "epoch": 0.36689477606883936, "grad_norm": 6.420316696166992, "learning_rate": 9.361489492778425e-05, "loss": 0.8031, "step": 5415 }, { "epoch": 0.36696253133681145, "grad_norm": 5.286546230316162, "learning_rate": 9.361352590868643e-05, "loss": 0.6665, "step": 5416 }, { "epoch": 0.36703028660478354, "grad_norm": 8.99885368347168, "learning_rate": 9.361215688958861e-05, "loss": 0.7545, "step": 5417 }, { "epoch": 0.36709804187275563, "grad_norm": 7.546742916107178, "learning_rate": 9.36107878704908e-05, "loss": 1.0391, "step": 5418 }, { "epoch": 0.36716579714072767, "grad_norm": 6.452430248260498, "learning_rate": 9.360941885139299e-05, "loss": 0.8458, "step": 5419 }, { "epoch": 0.36723355240869976, "grad_norm": 7.125503063201904, "learning_rate": 9.360804983229517e-05, "loss": 0.8061, "step": 5420 }, { "epoch": 0.36730130767667185, "grad_norm": 7.046570301055908, "learning_rate": 9.360668081319735e-05, "loss": 1.0968, "step": 5421 }, { "epoch": 0.36736906294464394, "grad_norm": 7.032671928405762, "learning_rate": 9.360531179409954e-05, "loss": 0.9834, "step": 5422 }, { "epoch": 0.36743681821261603, "grad_norm": 8.430721282958984, "learning_rate": 9.360394277500172e-05, "loss": 0.8875, "step": 5423 }, { "epoch": 0.3675045734805881, "grad_norm": 8.67082405090332, "learning_rate": 9.36025737559039e-05, "loss": 0.9981, "step": 5424 }, { "epoch": 0.3675723287485602, "grad_norm": 6.5781707763671875, "learning_rate": 9.360120473680608e-05, "loss": 0.9702, "step": 5425 }, { "epoch": 0.3676400840165323, "grad_norm": 9.087675094604492, "learning_rate": 9.359983571770826e-05, "loss": 1.1141, "step": 5426 }, { "epoch": 0.36770783928450435, "grad_norm": 6.949502468109131, "learning_rate": 9.359846669861046e-05, "loss": 0.9167, "step": 5427 }, { "epoch": 0.36777559455247644, "grad_norm": 9.335396766662598, "learning_rate": 9.359709767951264e-05, "loss": 0.8119, "step": 5428 }, { "epoch": 0.36784334982044853, "grad_norm": 8.041236877441406, "learning_rate": 9.359572866041482e-05, "loss": 1.0112, "step": 5429 }, { "epoch": 0.3679111050884206, "grad_norm": 6.435655117034912, "learning_rate": 9.3594359641317e-05, "loss": 0.9589, "step": 5430 }, { "epoch": 0.3679788603563927, "grad_norm": 8.047952651977539, "learning_rate": 9.359299062221919e-05, "loss": 0.7143, "step": 5431 }, { "epoch": 0.3680466156243648, "grad_norm": 6.986575126647949, "learning_rate": 9.359162160312137e-05, "loss": 0.8857, "step": 5432 }, { "epoch": 0.3681143708923369, "grad_norm": 5.447303771972656, "learning_rate": 9.359025258402355e-05, "loss": 0.7229, "step": 5433 }, { "epoch": 0.368182126160309, "grad_norm": 6.337569713592529, "learning_rate": 9.358888356492573e-05, "loss": 0.8595, "step": 5434 }, { "epoch": 0.368249881428281, "grad_norm": 7.318792819976807, "learning_rate": 9.358751454582791e-05, "loss": 0.8131, "step": 5435 }, { "epoch": 0.3683176366962531, "grad_norm": 6.816128730773926, "learning_rate": 9.35861455267301e-05, "loss": 0.6972, "step": 5436 }, { "epoch": 0.3683853919642252, "grad_norm": 6.886023998260498, "learning_rate": 9.358477650763229e-05, "loss": 0.9075, "step": 5437 }, { "epoch": 0.3684531472321973, "grad_norm": 7.4018049240112305, "learning_rate": 9.358340748853447e-05, "loss": 0.7913, "step": 5438 }, { "epoch": 0.3685209025001694, "grad_norm": 6.976071834564209, "learning_rate": 9.358203846943665e-05, "loss": 0.8306, "step": 5439 }, { "epoch": 0.3685886577681415, "grad_norm": 8.005870819091797, "learning_rate": 9.358066945033883e-05, "loss": 1.0109, "step": 5440 }, { "epoch": 0.3686564130361136, "grad_norm": 8.600711822509766, "learning_rate": 9.357930043124102e-05, "loss": 1.022, "step": 5441 }, { "epoch": 0.36872416830408566, "grad_norm": 9.857340812683105, "learning_rate": 9.35779314121432e-05, "loss": 1.1266, "step": 5442 }, { "epoch": 0.3687919235720577, "grad_norm": 8.064130783081055, "learning_rate": 9.357656239304538e-05, "loss": 0.9224, "step": 5443 }, { "epoch": 0.3688596788400298, "grad_norm": 9.217019081115723, "learning_rate": 9.357519337394756e-05, "loss": 1.0041, "step": 5444 }, { "epoch": 0.3689274341080019, "grad_norm": 7.042741298675537, "learning_rate": 9.357382435484976e-05, "loss": 0.6798, "step": 5445 }, { "epoch": 0.368995189375974, "grad_norm": 6.679181098937988, "learning_rate": 9.357245533575194e-05, "loss": 0.9641, "step": 5446 }, { "epoch": 0.36906294464394607, "grad_norm": 8.608474731445312, "learning_rate": 9.357108631665412e-05, "loss": 0.8224, "step": 5447 }, { "epoch": 0.36913069991191816, "grad_norm": 8.975968360900879, "learning_rate": 9.35697172975563e-05, "loss": 0.9103, "step": 5448 }, { "epoch": 0.36919845517989025, "grad_norm": 7.249898910522461, "learning_rate": 9.356834827845848e-05, "loss": 1.0178, "step": 5449 }, { "epoch": 0.36926621044786234, "grad_norm": 6.6788787841796875, "learning_rate": 9.356697925936067e-05, "loss": 0.7786, "step": 5450 }, { "epoch": 0.36933396571583443, "grad_norm": 7.834086894989014, "learning_rate": 9.356561024026285e-05, "loss": 1.0282, "step": 5451 }, { "epoch": 0.36940172098380647, "grad_norm": 7.36384391784668, "learning_rate": 9.356424122116503e-05, "loss": 1.0448, "step": 5452 }, { "epoch": 0.36946947625177856, "grad_norm": 7.843540191650391, "learning_rate": 9.356287220206723e-05, "loss": 1.0646, "step": 5453 }, { "epoch": 0.36953723151975065, "grad_norm": 8.009191513061523, "learning_rate": 9.35615031829694e-05, "loss": 1.0279, "step": 5454 }, { "epoch": 0.36960498678772274, "grad_norm": 7.594022274017334, "learning_rate": 9.356013416387159e-05, "loss": 1.0235, "step": 5455 }, { "epoch": 0.36967274205569484, "grad_norm": 6.167095184326172, "learning_rate": 9.355876514477378e-05, "loss": 0.7107, "step": 5456 }, { "epoch": 0.36974049732366693, "grad_norm": 8.07007884979248, "learning_rate": 9.355739612567596e-05, "loss": 0.9667, "step": 5457 }, { "epoch": 0.369808252591639, "grad_norm": 7.6446099281311035, "learning_rate": 9.355602710657814e-05, "loss": 0.9758, "step": 5458 }, { "epoch": 0.3698760078596111, "grad_norm": 7.282294273376465, "learning_rate": 9.355465808748034e-05, "loss": 0.929, "step": 5459 }, { "epoch": 0.36994376312758315, "grad_norm": 9.878494262695312, "learning_rate": 9.355328906838252e-05, "loss": 0.9909, "step": 5460 }, { "epoch": 0.37001151839555524, "grad_norm": 7.331822395324707, "learning_rate": 9.35519200492847e-05, "loss": 0.9693, "step": 5461 }, { "epoch": 0.37007927366352733, "grad_norm": 7.3209404945373535, "learning_rate": 9.355055103018688e-05, "loss": 0.8587, "step": 5462 }, { "epoch": 0.3701470289314994, "grad_norm": 8.261675834655762, "learning_rate": 9.354918201108906e-05, "loss": 0.9391, "step": 5463 }, { "epoch": 0.3702147841994715, "grad_norm": 7.214691162109375, "learning_rate": 9.354781299199125e-05, "loss": 0.8805, "step": 5464 }, { "epoch": 0.3702825394674436, "grad_norm": 7.409287452697754, "learning_rate": 9.354644397289343e-05, "loss": 0.6792, "step": 5465 }, { "epoch": 0.3703502947354157, "grad_norm": 6.368542194366455, "learning_rate": 9.354507495379561e-05, "loss": 0.6464, "step": 5466 }, { "epoch": 0.3704180500033878, "grad_norm": 7.106698513031006, "learning_rate": 9.354370593469779e-05, "loss": 1.0956, "step": 5467 }, { "epoch": 0.3704858052713598, "grad_norm": 7.591500759124756, "learning_rate": 9.354233691559999e-05, "loss": 0.9643, "step": 5468 }, { "epoch": 0.3705535605393319, "grad_norm": 7.255499839782715, "learning_rate": 9.354096789650217e-05, "loss": 0.7782, "step": 5469 }, { "epoch": 0.370621315807304, "grad_norm": 8.414693832397461, "learning_rate": 9.353959887740435e-05, "loss": 0.9433, "step": 5470 }, { "epoch": 0.3706890710752761, "grad_norm": 7.294419765472412, "learning_rate": 9.353822985830653e-05, "loss": 0.9643, "step": 5471 }, { "epoch": 0.3707568263432482, "grad_norm": 7.249752998352051, "learning_rate": 9.353686083920871e-05, "loss": 0.8529, "step": 5472 }, { "epoch": 0.3708245816112203, "grad_norm": 9.345986366271973, "learning_rate": 9.35354918201109e-05, "loss": 0.9641, "step": 5473 }, { "epoch": 0.3708923368791924, "grad_norm": 7.828972816467285, "learning_rate": 9.353412280101308e-05, "loss": 1.1, "step": 5474 }, { "epoch": 0.37096009214716447, "grad_norm": 7.739612579345703, "learning_rate": 9.353275378191526e-05, "loss": 0.8228, "step": 5475 }, { "epoch": 0.3710278474151365, "grad_norm": 7.853143215179443, "learning_rate": 9.353138476281744e-05, "loss": 0.9088, "step": 5476 }, { "epoch": 0.3710956026831086, "grad_norm": 7.996474742889404, "learning_rate": 9.353001574371964e-05, "loss": 0.9291, "step": 5477 }, { "epoch": 0.3711633579510807, "grad_norm": 7.374247074127197, "learning_rate": 9.352864672462182e-05, "loss": 1.0619, "step": 5478 }, { "epoch": 0.3712311132190528, "grad_norm": 8.037210464477539, "learning_rate": 9.3527277705524e-05, "loss": 0.9426, "step": 5479 }, { "epoch": 0.37129886848702487, "grad_norm": 5.515965938568115, "learning_rate": 9.352590868642618e-05, "loss": 0.6838, "step": 5480 }, { "epoch": 0.37136662375499696, "grad_norm": 8.530059814453125, "learning_rate": 9.352453966732836e-05, "loss": 0.9221, "step": 5481 }, { "epoch": 0.37143437902296905, "grad_norm": 8.07000732421875, "learning_rate": 9.352317064823055e-05, "loss": 0.9321, "step": 5482 }, { "epoch": 0.37150213429094114, "grad_norm": 8.79816722869873, "learning_rate": 9.352180162913273e-05, "loss": 1.0875, "step": 5483 }, { "epoch": 0.3715698895589132, "grad_norm": 8.476999282836914, "learning_rate": 9.352043261003491e-05, "loss": 1.0144, "step": 5484 }, { "epoch": 0.37163764482688527, "grad_norm": 5.7845540046691895, "learning_rate": 9.351906359093709e-05, "loss": 0.6802, "step": 5485 }, { "epoch": 0.37170540009485736, "grad_norm": 6.311066627502441, "learning_rate": 9.351769457183929e-05, "loss": 0.7673, "step": 5486 }, { "epoch": 0.37177315536282945, "grad_norm": 6.929514408111572, "learning_rate": 9.351632555274147e-05, "loss": 1.1287, "step": 5487 }, { "epoch": 0.37184091063080155, "grad_norm": 7.7819132804870605, "learning_rate": 9.351495653364365e-05, "loss": 0.8828, "step": 5488 }, { "epoch": 0.37190866589877364, "grad_norm": 7.906317710876465, "learning_rate": 9.351358751454583e-05, "loss": 0.7961, "step": 5489 }, { "epoch": 0.37197642116674573, "grad_norm": 8.190624237060547, "learning_rate": 9.351221849544801e-05, "loss": 0.8106, "step": 5490 }, { "epoch": 0.3720441764347178, "grad_norm": 6.0358734130859375, "learning_rate": 9.35108494763502e-05, "loss": 1.0342, "step": 5491 }, { "epoch": 0.37211193170268986, "grad_norm": 7.096914291381836, "learning_rate": 9.350948045725238e-05, "loss": 0.6652, "step": 5492 }, { "epoch": 0.37217968697066195, "grad_norm": 7.488151550292969, "learning_rate": 9.350811143815456e-05, "loss": 0.9283, "step": 5493 }, { "epoch": 0.37224744223863404, "grad_norm": 7.084420680999756, "learning_rate": 9.350674241905674e-05, "loss": 1.0424, "step": 5494 }, { "epoch": 0.37231519750660613, "grad_norm": 7.483108997344971, "learning_rate": 9.350537339995892e-05, "loss": 1.0657, "step": 5495 }, { "epoch": 0.3723829527745782, "grad_norm": 7.312635898590088, "learning_rate": 9.350400438086112e-05, "loss": 0.7831, "step": 5496 }, { "epoch": 0.3724507080425503, "grad_norm": 6.901159763336182, "learning_rate": 9.35026353617633e-05, "loss": 0.9832, "step": 5497 }, { "epoch": 0.3725184633105224, "grad_norm": 7.787426471710205, "learning_rate": 9.350126634266548e-05, "loss": 0.9112, "step": 5498 }, { "epoch": 0.3725862185784945, "grad_norm": 7.2583746910095215, "learning_rate": 9.349989732356767e-05, "loss": 0.757, "step": 5499 }, { "epoch": 0.37265397384646654, "grad_norm": 7.068667411804199, "learning_rate": 9.349852830446985e-05, "loss": 0.8516, "step": 5500 }, { "epoch": 0.3727217291144386, "grad_norm": 8.109786987304688, "learning_rate": 9.349715928537203e-05, "loss": 0.7925, "step": 5501 }, { "epoch": 0.3727894843824107, "grad_norm": 7.539520740509033, "learning_rate": 9.349579026627423e-05, "loss": 1.283, "step": 5502 }, { "epoch": 0.3728572396503828, "grad_norm": 8.608511924743652, "learning_rate": 9.34944212471764e-05, "loss": 1.2367, "step": 5503 }, { "epoch": 0.3729249949183549, "grad_norm": 7.932346820831299, "learning_rate": 9.349305222807859e-05, "loss": 0.894, "step": 5504 }, { "epoch": 0.372992750186327, "grad_norm": 6.520984172821045, "learning_rate": 9.349168320898078e-05, "loss": 0.884, "step": 5505 }, { "epoch": 0.3730605054542991, "grad_norm": 6.876826286315918, "learning_rate": 9.349031418988296e-05, "loss": 0.9447, "step": 5506 }, { "epoch": 0.3731282607222712, "grad_norm": 8.540080070495605, "learning_rate": 9.348894517078514e-05, "loss": 1.0709, "step": 5507 }, { "epoch": 0.37319601599024327, "grad_norm": 7.760754585266113, "learning_rate": 9.348757615168732e-05, "loss": 1.0666, "step": 5508 }, { "epoch": 0.3732637712582153, "grad_norm": 6.1746134757995605, "learning_rate": 9.348620713258951e-05, "loss": 0.8367, "step": 5509 }, { "epoch": 0.3733315265261874, "grad_norm": 8.053776741027832, "learning_rate": 9.34848381134917e-05, "loss": 1.253, "step": 5510 }, { "epoch": 0.3733992817941595, "grad_norm": 10.046021461486816, "learning_rate": 9.348346909439388e-05, "loss": 0.9086, "step": 5511 }, { "epoch": 0.3734670370621316, "grad_norm": 7.247015953063965, "learning_rate": 9.348210007529606e-05, "loss": 1.3036, "step": 5512 }, { "epoch": 0.37353479233010367, "grad_norm": 6.103085517883301, "learning_rate": 9.348073105619824e-05, "loss": 0.9778, "step": 5513 }, { "epoch": 0.37360254759807576, "grad_norm": 7.330771446228027, "learning_rate": 9.347936203710043e-05, "loss": 0.8335, "step": 5514 }, { "epoch": 0.37367030286604785, "grad_norm": 7.571770668029785, "learning_rate": 9.347799301800261e-05, "loss": 0.9963, "step": 5515 }, { "epoch": 0.37373805813401995, "grad_norm": 8.018990516662598, "learning_rate": 9.347662399890479e-05, "loss": 0.7925, "step": 5516 }, { "epoch": 0.373805813401992, "grad_norm": 7.156185626983643, "learning_rate": 9.347525497980697e-05, "loss": 0.7821, "step": 5517 }, { "epoch": 0.3738735686699641, "grad_norm": 8.189929008483887, "learning_rate": 9.347388596070915e-05, "loss": 0.8997, "step": 5518 }, { "epoch": 0.37394132393793617, "grad_norm": 6.826793670654297, "learning_rate": 9.347251694161135e-05, "loss": 0.7392, "step": 5519 }, { "epoch": 0.37400907920590826, "grad_norm": 7.931643962860107, "learning_rate": 9.347114792251353e-05, "loss": 1.1607, "step": 5520 }, { "epoch": 0.37407683447388035, "grad_norm": 6.973018646240234, "learning_rate": 9.34697789034157e-05, "loss": 0.7419, "step": 5521 }, { "epoch": 0.37414458974185244, "grad_norm": 7.43549919128418, "learning_rate": 9.346840988431789e-05, "loss": 1.0035, "step": 5522 }, { "epoch": 0.37421234500982453, "grad_norm": 6.700629234313965, "learning_rate": 9.346704086522008e-05, "loss": 1.0953, "step": 5523 }, { "epoch": 0.3742801002777966, "grad_norm": 6.039200782775879, "learning_rate": 9.346567184612226e-05, "loss": 0.6512, "step": 5524 }, { "epoch": 0.37434785554576866, "grad_norm": 8.45660400390625, "learning_rate": 9.346430282702444e-05, "loss": 1.0725, "step": 5525 }, { "epoch": 0.37441561081374075, "grad_norm": 8.787965774536133, "learning_rate": 9.346293380792662e-05, "loss": 1.0101, "step": 5526 }, { "epoch": 0.37448336608171284, "grad_norm": 8.229619026184082, "learning_rate": 9.34615647888288e-05, "loss": 1.2651, "step": 5527 }, { "epoch": 0.37455112134968493, "grad_norm": 6.913321018218994, "learning_rate": 9.3460195769731e-05, "loss": 0.9196, "step": 5528 }, { "epoch": 0.374618876617657, "grad_norm": 5.944606781005859, "learning_rate": 9.345882675063318e-05, "loss": 0.857, "step": 5529 }, { "epoch": 0.3746866318856291, "grad_norm": 9.863933563232422, "learning_rate": 9.345745773153536e-05, "loss": 1.0566, "step": 5530 }, { "epoch": 0.3747543871536012, "grad_norm": 6.494686603546143, "learning_rate": 9.345608871243754e-05, "loss": 0.8839, "step": 5531 }, { "epoch": 0.3748221424215733, "grad_norm": 6.558131694793701, "learning_rate": 9.345471969333973e-05, "loss": 0.9802, "step": 5532 }, { "epoch": 0.37488989768954534, "grad_norm": 7.316765308380127, "learning_rate": 9.345335067424191e-05, "loss": 1.1635, "step": 5533 }, { "epoch": 0.37495765295751743, "grad_norm": 6.615121841430664, "learning_rate": 9.345198165514409e-05, "loss": 0.919, "step": 5534 }, { "epoch": 0.3750254082254895, "grad_norm": 8.071523666381836, "learning_rate": 9.345061263604627e-05, "loss": 1.0023, "step": 5535 }, { "epoch": 0.3750931634934616, "grad_norm": 7.563895225524902, "learning_rate": 9.344924361694845e-05, "loss": 0.963, "step": 5536 }, { "epoch": 0.3751609187614337, "grad_norm": 6.013591766357422, "learning_rate": 9.344787459785065e-05, "loss": 0.7862, "step": 5537 }, { "epoch": 0.3752286740294058, "grad_norm": 7.296882629394531, "learning_rate": 9.344650557875283e-05, "loss": 1.0106, "step": 5538 }, { "epoch": 0.3752964292973779, "grad_norm": 7.192817211151123, "learning_rate": 9.3445136559655e-05, "loss": 0.8491, "step": 5539 }, { "epoch": 0.37536418456535, "grad_norm": 7.962235927581787, "learning_rate": 9.344376754055719e-05, "loss": 1.0711, "step": 5540 }, { "epoch": 0.375431939833322, "grad_norm": 6.515812873840332, "learning_rate": 9.344239852145937e-05, "loss": 0.8779, "step": 5541 }, { "epoch": 0.3754996951012941, "grad_norm": 8.563304901123047, "learning_rate": 9.344102950236156e-05, "loss": 1.3236, "step": 5542 }, { "epoch": 0.3755674503692662, "grad_norm": 8.214926719665527, "learning_rate": 9.343966048326374e-05, "loss": 0.9231, "step": 5543 }, { "epoch": 0.3756352056372383, "grad_norm": 6.213039398193359, "learning_rate": 9.343829146416592e-05, "loss": 0.8446, "step": 5544 }, { "epoch": 0.3757029609052104, "grad_norm": 6.392467021942139, "learning_rate": 9.343692244506812e-05, "loss": 0.6742, "step": 5545 }, { "epoch": 0.3757707161731825, "grad_norm": 6.8930206298828125, "learning_rate": 9.34355534259703e-05, "loss": 0.9973, "step": 5546 }, { "epoch": 0.37583847144115456, "grad_norm": 8.074743270874023, "learning_rate": 9.343418440687248e-05, "loss": 0.8808, "step": 5547 }, { "epoch": 0.37590622670912666, "grad_norm": 8.443988800048828, "learning_rate": 9.343281538777467e-05, "loss": 1.0066, "step": 5548 }, { "epoch": 0.3759739819770987, "grad_norm": 6.530252933502197, "learning_rate": 9.343144636867685e-05, "loss": 1.0203, "step": 5549 }, { "epoch": 0.3760417372450708, "grad_norm": 6.693957328796387, "learning_rate": 9.343007734957903e-05, "loss": 0.9484, "step": 5550 }, { "epoch": 0.3761094925130429, "grad_norm": 7.103133201599121, "learning_rate": 9.342870833048122e-05, "loss": 1.0442, "step": 5551 }, { "epoch": 0.37617724778101497, "grad_norm": 6.089579105377197, "learning_rate": 9.34273393113834e-05, "loss": 0.8828, "step": 5552 }, { "epoch": 0.37624500304898706, "grad_norm": 5.47523307800293, "learning_rate": 9.342597029228559e-05, "loss": 0.8302, "step": 5553 }, { "epoch": 0.37631275831695915, "grad_norm": 7.930117130279541, "learning_rate": 9.342460127318777e-05, "loss": 0.9565, "step": 5554 }, { "epoch": 0.37638051358493124, "grad_norm": 6.38248348236084, "learning_rate": 9.342323225408996e-05, "loss": 0.8815, "step": 5555 }, { "epoch": 0.37644826885290333, "grad_norm": 6.382977485656738, "learning_rate": 9.342186323499214e-05, "loss": 0.8472, "step": 5556 }, { "epoch": 0.3765160241208754, "grad_norm": 6.029202461242676, "learning_rate": 9.342049421589432e-05, "loss": 0.8102, "step": 5557 }, { "epoch": 0.37658377938884746, "grad_norm": 9.764986991882324, "learning_rate": 9.34191251967965e-05, "loss": 0.7955, "step": 5558 }, { "epoch": 0.37665153465681955, "grad_norm": 8.568723678588867, "learning_rate": 9.341775617769868e-05, "loss": 1.0919, "step": 5559 }, { "epoch": 0.37671928992479164, "grad_norm": 13.348160743713379, "learning_rate": 9.341638715860087e-05, "loss": 1.202, "step": 5560 }, { "epoch": 0.37678704519276374, "grad_norm": 7.161466121673584, "learning_rate": 9.341501813950306e-05, "loss": 0.909, "step": 5561 }, { "epoch": 0.37685480046073583, "grad_norm": 8.446759223937988, "learning_rate": 9.341364912040524e-05, "loss": 0.8871, "step": 5562 }, { "epoch": 0.3769225557287079, "grad_norm": 7.228363037109375, "learning_rate": 9.341228010130742e-05, "loss": 0.961, "step": 5563 }, { "epoch": 0.37699031099668, "grad_norm": 7.427947998046875, "learning_rate": 9.341091108220961e-05, "loss": 1.1086, "step": 5564 }, { "epoch": 0.3770580662646521, "grad_norm": 8.64295482635498, "learning_rate": 9.340954206311179e-05, "loss": 0.9684, "step": 5565 }, { "epoch": 0.37712582153262414, "grad_norm": 10.287007331848145, "learning_rate": 9.340817304401397e-05, "loss": 1.0202, "step": 5566 }, { "epoch": 0.37719357680059623, "grad_norm": 7.803440570831299, "learning_rate": 9.340680402491615e-05, "loss": 0.8545, "step": 5567 }, { "epoch": 0.3772613320685683, "grad_norm": 7.74205207824707, "learning_rate": 9.340543500581833e-05, "loss": 1.0585, "step": 5568 }, { "epoch": 0.3773290873365404, "grad_norm": 6.5861053466796875, "learning_rate": 9.340406598672053e-05, "loss": 0.8838, "step": 5569 }, { "epoch": 0.3773968426045125, "grad_norm": 7.1810431480407715, "learning_rate": 9.34026969676227e-05, "loss": 0.8669, "step": 5570 }, { "epoch": 0.3774645978724846, "grad_norm": 7.8471808433532715, "learning_rate": 9.340132794852489e-05, "loss": 1.0412, "step": 5571 }, { "epoch": 0.3775323531404567, "grad_norm": 7.740540027618408, "learning_rate": 9.339995892942707e-05, "loss": 0.944, "step": 5572 }, { "epoch": 0.3776001084084288, "grad_norm": 7.580658912658691, "learning_rate": 9.339858991032925e-05, "loss": 0.8676, "step": 5573 }, { "epoch": 0.3776678636764008, "grad_norm": 9.143624305725098, "learning_rate": 9.339722089123144e-05, "loss": 0.9752, "step": 5574 }, { "epoch": 0.3777356189443729, "grad_norm": 8.142240524291992, "learning_rate": 9.339585187213362e-05, "loss": 0.9733, "step": 5575 }, { "epoch": 0.377803374212345, "grad_norm": 8.034200668334961, "learning_rate": 9.33944828530358e-05, "loss": 0.9867, "step": 5576 }, { "epoch": 0.3778711294803171, "grad_norm": 7.345114231109619, "learning_rate": 9.339311383393798e-05, "loss": 0.9401, "step": 5577 }, { "epoch": 0.3779388847482892, "grad_norm": 7.522977828979492, "learning_rate": 9.339174481484018e-05, "loss": 0.965, "step": 5578 }, { "epoch": 0.3780066400162613, "grad_norm": 6.02646541595459, "learning_rate": 9.339037579574236e-05, "loss": 0.9446, "step": 5579 }, { "epoch": 0.37807439528423337, "grad_norm": 6.161006450653076, "learning_rate": 9.338900677664454e-05, "loss": 0.8803, "step": 5580 }, { "epoch": 0.37814215055220546, "grad_norm": 6.318620681762695, "learning_rate": 9.338763775754672e-05, "loss": 0.646, "step": 5581 }, { "epoch": 0.3782099058201775, "grad_norm": 5.793397903442383, "learning_rate": 9.33862687384489e-05, "loss": 0.7483, "step": 5582 }, { "epoch": 0.3782776610881496, "grad_norm": 7.235496997833252, "learning_rate": 9.338489971935109e-05, "loss": 0.9768, "step": 5583 }, { "epoch": 0.3783454163561217, "grad_norm": 6.682423114776611, "learning_rate": 9.338353070025327e-05, "loss": 0.8466, "step": 5584 }, { "epoch": 0.37841317162409377, "grad_norm": 9.257969856262207, "learning_rate": 9.338216168115545e-05, "loss": 1.1546, "step": 5585 }, { "epoch": 0.37848092689206586, "grad_norm": 7.223050594329834, "learning_rate": 9.338079266205763e-05, "loss": 0.7594, "step": 5586 }, { "epoch": 0.37854868216003795, "grad_norm": 7.735831260681152, "learning_rate": 9.337942364295983e-05, "loss": 1.0392, "step": 5587 }, { "epoch": 0.37861643742801004, "grad_norm": 6.439917087554932, "learning_rate": 9.3378054623862e-05, "loss": 1.0901, "step": 5588 }, { "epoch": 0.37868419269598214, "grad_norm": 7.018877983093262, "learning_rate": 9.337668560476419e-05, "loss": 0.9918, "step": 5589 }, { "epoch": 0.37875194796395417, "grad_norm": 5.858334541320801, "learning_rate": 9.337531658566637e-05, "loss": 0.6788, "step": 5590 }, { "epoch": 0.37881970323192626, "grad_norm": 8.445877075195312, "learning_rate": 9.337394756656856e-05, "loss": 0.9777, "step": 5591 }, { "epoch": 0.37888745849989836, "grad_norm": 6.7996439933776855, "learning_rate": 9.337257854747074e-05, "loss": 1.0086, "step": 5592 }, { "epoch": 0.37895521376787045, "grad_norm": 6.983299255371094, "learning_rate": 9.337120952837292e-05, "loss": 0.9981, "step": 5593 }, { "epoch": 0.37902296903584254, "grad_norm": 7.999194622039795, "learning_rate": 9.336984050927511e-05, "loss": 0.9166, "step": 5594 }, { "epoch": 0.37909072430381463, "grad_norm": 7.698132038116455, "learning_rate": 9.33684714901773e-05, "loss": 0.8333, "step": 5595 }, { "epoch": 0.3791584795717867, "grad_norm": 5.9128217697143555, "learning_rate": 9.336710247107948e-05, "loss": 0.7822, "step": 5596 }, { "epoch": 0.3792262348397588, "grad_norm": 7.200125217437744, "learning_rate": 9.336573345198167e-05, "loss": 0.906, "step": 5597 }, { "epoch": 0.37929399010773085, "grad_norm": 5.82244348526001, "learning_rate": 9.336436443288385e-05, "loss": 0.7376, "step": 5598 }, { "epoch": 0.37936174537570294, "grad_norm": 6.775420188903809, "learning_rate": 9.336299541378603e-05, "loss": 0.6975, "step": 5599 }, { "epoch": 0.37942950064367503, "grad_norm": 6.542332649230957, "learning_rate": 9.336162639468821e-05, "loss": 0.9788, "step": 5600 }, { "epoch": 0.3794972559116471, "grad_norm": 6.677700519561768, "learning_rate": 9.33602573755904e-05, "loss": 0.79, "step": 5601 }, { "epoch": 0.3795650111796192, "grad_norm": 8.561968803405762, "learning_rate": 9.335888835649258e-05, "loss": 0.9499, "step": 5602 }, { "epoch": 0.3796327664475913, "grad_norm": 7.0320234298706055, "learning_rate": 9.335751933739477e-05, "loss": 0.865, "step": 5603 }, { "epoch": 0.3797005217155634, "grad_norm": 7.31481409072876, "learning_rate": 9.335615031829695e-05, "loss": 1.0356, "step": 5604 }, { "epoch": 0.3797682769835355, "grad_norm": 8.205849647521973, "learning_rate": 9.335478129919913e-05, "loss": 0.7836, "step": 5605 }, { "epoch": 0.3798360322515075, "grad_norm": 6.882634162902832, "learning_rate": 9.335341228010132e-05, "loss": 0.8879, "step": 5606 }, { "epoch": 0.3799037875194796, "grad_norm": 8.171550750732422, "learning_rate": 9.33520432610035e-05, "loss": 0.8849, "step": 5607 }, { "epoch": 0.3799715427874517, "grad_norm": 7.184536457061768, "learning_rate": 9.335067424190568e-05, "loss": 0.8206, "step": 5608 }, { "epoch": 0.3800392980554238, "grad_norm": 6.50775146484375, "learning_rate": 9.334930522280786e-05, "loss": 0.7802, "step": 5609 }, { "epoch": 0.3801070533233959, "grad_norm": 7.537467002868652, "learning_rate": 9.334793620371005e-05, "loss": 1.1643, "step": 5610 }, { "epoch": 0.380174808591368, "grad_norm": 6.740983486175537, "learning_rate": 9.334656718461223e-05, "loss": 0.686, "step": 5611 }, { "epoch": 0.3802425638593401, "grad_norm": 6.01100492477417, "learning_rate": 9.334519816551442e-05, "loss": 0.8931, "step": 5612 }, { "epoch": 0.38031031912731217, "grad_norm": 6.056909084320068, "learning_rate": 9.33438291464166e-05, "loss": 0.7861, "step": 5613 }, { "epoch": 0.38037807439528426, "grad_norm": 6.271320343017578, "learning_rate": 9.334246012731878e-05, "loss": 0.8277, "step": 5614 }, { "epoch": 0.3804458296632563, "grad_norm": 5.80530309677124, "learning_rate": 9.334109110822097e-05, "loss": 0.6765, "step": 5615 }, { "epoch": 0.3805135849312284, "grad_norm": 6.031884670257568, "learning_rate": 9.333972208912315e-05, "loss": 0.6482, "step": 5616 }, { "epoch": 0.3805813401992005, "grad_norm": 7.629202842712402, "learning_rate": 9.333835307002533e-05, "loss": 1.0782, "step": 5617 }, { "epoch": 0.38064909546717257, "grad_norm": 8.182783126831055, "learning_rate": 9.333698405092751e-05, "loss": 0.8355, "step": 5618 }, { "epoch": 0.38071685073514466, "grad_norm": 7.317597389221191, "learning_rate": 9.33356150318297e-05, "loss": 0.9035, "step": 5619 }, { "epoch": 0.38078460600311675, "grad_norm": 8.783252716064453, "learning_rate": 9.333424601273189e-05, "loss": 1.1105, "step": 5620 }, { "epoch": 0.38085236127108885, "grad_norm": 6.5179290771484375, "learning_rate": 9.333287699363407e-05, "loss": 0.7859, "step": 5621 }, { "epoch": 0.38092011653906094, "grad_norm": 7.544989585876465, "learning_rate": 9.333150797453625e-05, "loss": 0.8601, "step": 5622 }, { "epoch": 0.380987871807033, "grad_norm": 7.307798862457275, "learning_rate": 9.333013895543843e-05, "loss": 1.071, "step": 5623 }, { "epoch": 0.38105562707500507, "grad_norm": 6.946296215057373, "learning_rate": 9.332876993634062e-05, "loss": 0.8445, "step": 5624 }, { "epoch": 0.38112338234297716, "grad_norm": 8.099409103393555, "learning_rate": 9.33274009172428e-05, "loss": 0.9821, "step": 5625 }, { "epoch": 0.38119113761094925, "grad_norm": 7.492844104766846, "learning_rate": 9.332603189814498e-05, "loss": 0.9696, "step": 5626 }, { "epoch": 0.38125889287892134, "grad_norm": 6.632567882537842, "learning_rate": 9.332466287904716e-05, "loss": 0.8853, "step": 5627 }, { "epoch": 0.38132664814689343, "grad_norm": 5.6510748863220215, "learning_rate": 9.332329385994934e-05, "loss": 0.8686, "step": 5628 }, { "epoch": 0.3813944034148655, "grad_norm": 6.428233623504639, "learning_rate": 9.332192484085154e-05, "loss": 0.8912, "step": 5629 }, { "epoch": 0.3814621586828376, "grad_norm": 8.161954879760742, "learning_rate": 9.332055582175372e-05, "loss": 0.888, "step": 5630 }, { "epoch": 0.38152991395080965, "grad_norm": 7.743470191955566, "learning_rate": 9.33191868026559e-05, "loss": 0.8759, "step": 5631 }, { "epoch": 0.38159766921878174, "grad_norm": 7.648897647857666, "learning_rate": 9.331781778355808e-05, "loss": 1.2331, "step": 5632 }, { "epoch": 0.38166542448675383, "grad_norm": 7.748523235321045, "learning_rate": 9.331644876446027e-05, "loss": 1.1732, "step": 5633 }, { "epoch": 0.3817331797547259, "grad_norm": 7.337912082672119, "learning_rate": 9.331507974536245e-05, "loss": 0.8043, "step": 5634 }, { "epoch": 0.381800935022698, "grad_norm": 7.525491714477539, "learning_rate": 9.331371072626463e-05, "loss": 0.8762, "step": 5635 }, { "epoch": 0.3818686902906701, "grad_norm": 7.477416515350342, "learning_rate": 9.331234170716681e-05, "loss": 0.9053, "step": 5636 }, { "epoch": 0.3819364455586422, "grad_norm": 6.855381965637207, "learning_rate": 9.331097268806899e-05, "loss": 0.9124, "step": 5637 }, { "epoch": 0.3820042008266143, "grad_norm": 7.663267135620117, "learning_rate": 9.330960366897119e-05, "loss": 0.8139, "step": 5638 }, { "epoch": 0.38207195609458633, "grad_norm": 6.08014440536499, "learning_rate": 9.330823464987337e-05, "loss": 0.7084, "step": 5639 }, { "epoch": 0.3821397113625584, "grad_norm": 7.804579734802246, "learning_rate": 9.330686563077555e-05, "loss": 1.206, "step": 5640 }, { "epoch": 0.3822074666305305, "grad_norm": 7.1148552894592285, "learning_rate": 9.330549661167774e-05, "loss": 1.1221, "step": 5641 }, { "epoch": 0.3822752218985026, "grad_norm": 6.727654457092285, "learning_rate": 9.330412759257992e-05, "loss": 1.025, "step": 5642 }, { "epoch": 0.3823429771664747, "grad_norm": 7.059203147888184, "learning_rate": 9.33027585734821e-05, "loss": 1.0819, "step": 5643 }, { "epoch": 0.3824107324344468, "grad_norm": 8.003962516784668, "learning_rate": 9.33013895543843e-05, "loss": 0.95, "step": 5644 }, { "epoch": 0.3824784877024189, "grad_norm": 7.414921283721924, "learning_rate": 9.330002053528647e-05, "loss": 0.9279, "step": 5645 }, { "epoch": 0.38254624297039097, "grad_norm": 5.462070465087891, "learning_rate": 9.329865151618866e-05, "loss": 0.8003, "step": 5646 }, { "epoch": 0.382613998238363, "grad_norm": 7.23457145690918, "learning_rate": 9.329728249709085e-05, "loss": 0.8939, "step": 5647 }, { "epoch": 0.3826817535063351, "grad_norm": 6.371642112731934, "learning_rate": 9.329591347799303e-05, "loss": 0.7125, "step": 5648 }, { "epoch": 0.3827495087743072, "grad_norm": 6.855792999267578, "learning_rate": 9.329454445889521e-05, "loss": 0.745, "step": 5649 }, { "epoch": 0.3828172640422793, "grad_norm": 6.945821762084961, "learning_rate": 9.329317543979739e-05, "loss": 1.1931, "step": 5650 }, { "epoch": 0.3828850193102514, "grad_norm": 7.453825950622559, "learning_rate": 9.329180642069957e-05, "loss": 0.9513, "step": 5651 }, { "epoch": 0.38295277457822346, "grad_norm": 7.040378093719482, "learning_rate": 9.329043740160176e-05, "loss": 0.8896, "step": 5652 }, { "epoch": 0.38302052984619556, "grad_norm": 8.740388870239258, "learning_rate": 9.328906838250394e-05, "loss": 0.771, "step": 5653 }, { "epoch": 0.38308828511416765, "grad_norm": 6.791045665740967, "learning_rate": 9.328769936340613e-05, "loss": 0.8835, "step": 5654 }, { "epoch": 0.3831560403821397, "grad_norm": 6.560173511505127, "learning_rate": 9.32863303443083e-05, "loss": 0.8516, "step": 5655 }, { "epoch": 0.3832237956501118, "grad_norm": 6.763086795806885, "learning_rate": 9.32849613252105e-05, "loss": 0.7976, "step": 5656 }, { "epoch": 0.38329155091808387, "grad_norm": 5.9037370681762695, "learning_rate": 9.328359230611268e-05, "loss": 0.943, "step": 5657 }, { "epoch": 0.38335930618605596, "grad_norm": 4.944889545440674, "learning_rate": 9.328222328701486e-05, "loss": 0.7913, "step": 5658 }, { "epoch": 0.38342706145402805, "grad_norm": 8.701375007629395, "learning_rate": 9.328085426791704e-05, "loss": 0.8, "step": 5659 }, { "epoch": 0.38349481672200014, "grad_norm": 7.1927289962768555, "learning_rate": 9.327948524881922e-05, "loss": 0.8888, "step": 5660 }, { "epoch": 0.38356257198997223, "grad_norm": 8.781030654907227, "learning_rate": 9.327811622972141e-05, "loss": 0.951, "step": 5661 }, { "epoch": 0.3836303272579443, "grad_norm": 7.27484655380249, "learning_rate": 9.32767472106236e-05, "loss": 0.8034, "step": 5662 }, { "epoch": 0.3836980825259164, "grad_norm": 7.739286422729492, "learning_rate": 9.327537819152578e-05, "loss": 1.0769, "step": 5663 }, { "epoch": 0.38376583779388845, "grad_norm": 7.0264973640441895, "learning_rate": 9.327400917242796e-05, "loss": 1.0652, "step": 5664 }, { "epoch": 0.38383359306186055, "grad_norm": 6.887399673461914, "learning_rate": 9.327264015333015e-05, "loss": 0.7986, "step": 5665 }, { "epoch": 0.38390134832983264, "grad_norm": 9.081385612487793, "learning_rate": 9.327127113423233e-05, "loss": 1.2249, "step": 5666 }, { "epoch": 0.38396910359780473, "grad_norm": 7.966154098510742, "learning_rate": 9.326990211513451e-05, "loss": 0.8924, "step": 5667 }, { "epoch": 0.3840368588657768, "grad_norm": 7.321074962615967, "learning_rate": 9.326853309603669e-05, "loss": 1.0022, "step": 5668 }, { "epoch": 0.3841046141337489, "grad_norm": 7.060865879058838, "learning_rate": 9.326716407693887e-05, "loss": 0.8849, "step": 5669 }, { "epoch": 0.384172369401721, "grad_norm": 7.009364604949951, "learning_rate": 9.326579505784106e-05, "loss": 0.9032, "step": 5670 }, { "epoch": 0.3842401246696931, "grad_norm": 8.417223930358887, "learning_rate": 9.326442603874325e-05, "loss": 1.1162, "step": 5671 }, { "epoch": 0.38430787993766513, "grad_norm": 6.774781227111816, "learning_rate": 9.326305701964543e-05, "loss": 0.773, "step": 5672 }, { "epoch": 0.3843756352056372, "grad_norm": 6.5464911460876465, "learning_rate": 9.32616880005476e-05, "loss": 1.0374, "step": 5673 }, { "epoch": 0.3844433904736093, "grad_norm": 6.618286609649658, "learning_rate": 9.326031898144979e-05, "loss": 0.9538, "step": 5674 }, { "epoch": 0.3845111457415814, "grad_norm": 7.598459720611572, "learning_rate": 9.325894996235198e-05, "loss": 0.7811, "step": 5675 }, { "epoch": 0.3845789010095535, "grad_norm": 6.456278324127197, "learning_rate": 9.325758094325416e-05, "loss": 1.055, "step": 5676 }, { "epoch": 0.3846466562775256, "grad_norm": 7.940021991729736, "learning_rate": 9.325621192415634e-05, "loss": 1.1605, "step": 5677 }, { "epoch": 0.3847144115454977, "grad_norm": 8.113789558410645, "learning_rate": 9.325484290505852e-05, "loss": 1.0405, "step": 5678 }, { "epoch": 0.3847821668134698, "grad_norm": 7.918695449829102, "learning_rate": 9.325347388596071e-05, "loss": 0.8019, "step": 5679 }, { "epoch": 0.3848499220814418, "grad_norm": 8.129600524902344, "learning_rate": 9.32521048668629e-05, "loss": 0.9581, "step": 5680 }, { "epoch": 0.3849176773494139, "grad_norm": 6.6353230476379395, "learning_rate": 9.325073584776508e-05, "loss": 0.8932, "step": 5681 }, { "epoch": 0.384985432617386, "grad_norm": 6.153687477111816, "learning_rate": 9.324936682866726e-05, "loss": 0.7646, "step": 5682 }, { "epoch": 0.3850531878853581, "grad_norm": 6.837560653686523, "learning_rate": 9.324799780956944e-05, "loss": 0.8998, "step": 5683 }, { "epoch": 0.3851209431533302, "grad_norm": 8.48747444152832, "learning_rate": 9.324662879047163e-05, "loss": 0.7761, "step": 5684 }, { "epoch": 0.38518869842130227, "grad_norm": 10.725845336914062, "learning_rate": 9.324525977137381e-05, "loss": 1.1755, "step": 5685 }, { "epoch": 0.38525645368927436, "grad_norm": 5.731265544891357, "learning_rate": 9.324389075227599e-05, "loss": 0.8974, "step": 5686 }, { "epoch": 0.38532420895724645, "grad_norm": 8.479190826416016, "learning_rate": 9.324252173317818e-05, "loss": 0.9307, "step": 5687 }, { "epoch": 0.3853919642252185, "grad_norm": 6.539626598358154, "learning_rate": 9.324115271408037e-05, "loss": 0.9463, "step": 5688 }, { "epoch": 0.3854597194931906, "grad_norm": 7.225162029266357, "learning_rate": 9.323978369498255e-05, "loss": 0.8206, "step": 5689 }, { "epoch": 0.38552747476116267, "grad_norm": 8.779760360717773, "learning_rate": 9.323841467588474e-05, "loss": 1.1208, "step": 5690 }, { "epoch": 0.38559523002913476, "grad_norm": 7.580684185028076, "learning_rate": 9.323704565678692e-05, "loss": 0.7993, "step": 5691 }, { "epoch": 0.38566298529710685, "grad_norm": 8.699392318725586, "learning_rate": 9.32356766376891e-05, "loss": 1.073, "step": 5692 }, { "epoch": 0.38573074056507894, "grad_norm": 7.190006256103516, "learning_rate": 9.32343076185913e-05, "loss": 1.0204, "step": 5693 }, { "epoch": 0.38579849583305104, "grad_norm": 9.289970397949219, "learning_rate": 9.323293859949347e-05, "loss": 0.7091, "step": 5694 }, { "epoch": 0.3858662511010231, "grad_norm": 9.533041000366211, "learning_rate": 9.323156958039565e-05, "loss": 1.103, "step": 5695 }, { "epoch": 0.38593400636899516, "grad_norm": 7.375948905944824, "learning_rate": 9.323020056129783e-05, "loss": 0.9815, "step": 5696 }, { "epoch": 0.38600176163696726, "grad_norm": 8.580230712890625, "learning_rate": 9.322883154220003e-05, "loss": 1.1056, "step": 5697 }, { "epoch": 0.38606951690493935, "grad_norm": 8.332324981689453, "learning_rate": 9.322746252310221e-05, "loss": 0.8809, "step": 5698 }, { "epoch": 0.38613727217291144, "grad_norm": 6.298001289367676, "learning_rate": 9.322609350400439e-05, "loss": 0.9255, "step": 5699 }, { "epoch": 0.38620502744088353, "grad_norm": 7.031383514404297, "learning_rate": 9.322472448490657e-05, "loss": 0.9196, "step": 5700 }, { "epoch": 0.3862727827088556, "grad_norm": 6.791995048522949, "learning_rate": 9.322335546580875e-05, "loss": 0.8184, "step": 5701 }, { "epoch": 0.3863405379768277, "grad_norm": 7.81817102432251, "learning_rate": 9.322198644671094e-05, "loss": 0.9331, "step": 5702 }, { "epoch": 0.3864082932447998, "grad_norm": 7.933851718902588, "learning_rate": 9.322061742761312e-05, "loss": 0.872, "step": 5703 }, { "epoch": 0.38647604851277184, "grad_norm": 7.805744171142578, "learning_rate": 9.32192484085153e-05, "loss": 0.8378, "step": 5704 }, { "epoch": 0.38654380378074393, "grad_norm": 9.205484390258789, "learning_rate": 9.321787938941749e-05, "loss": 0.9853, "step": 5705 }, { "epoch": 0.386611559048716, "grad_norm": 7.313584804534912, "learning_rate": 9.321651037031967e-05, "loss": 0.7576, "step": 5706 }, { "epoch": 0.3866793143166881, "grad_norm": 10.125847816467285, "learning_rate": 9.321514135122186e-05, "loss": 0.8497, "step": 5707 }, { "epoch": 0.3867470695846602, "grad_norm": 7.392860412597656, "learning_rate": 9.321377233212404e-05, "loss": 0.9151, "step": 5708 }, { "epoch": 0.3868148248526323, "grad_norm": 8.305766105651855, "learning_rate": 9.321240331302622e-05, "loss": 1.1604, "step": 5709 }, { "epoch": 0.3868825801206044, "grad_norm": 5.897716522216797, "learning_rate": 9.32110342939284e-05, "loss": 0.6471, "step": 5710 }, { "epoch": 0.3869503353885765, "grad_norm": 6.7423200607299805, "learning_rate": 9.32096652748306e-05, "loss": 0.7968, "step": 5711 }, { "epoch": 0.3870180906565485, "grad_norm": 8.511382102966309, "learning_rate": 9.320829625573277e-05, "loss": 1.1875, "step": 5712 }, { "epoch": 0.3870858459245206, "grad_norm": 8.105764389038086, "learning_rate": 9.320692723663495e-05, "loss": 1.0717, "step": 5713 }, { "epoch": 0.3871536011924927, "grad_norm": 6.555315971374512, "learning_rate": 9.320555821753714e-05, "loss": 0.7491, "step": 5714 }, { "epoch": 0.3872213564604648, "grad_norm": 7.481291770935059, "learning_rate": 9.320418919843932e-05, "loss": 1.0057, "step": 5715 }, { "epoch": 0.3872891117284369, "grad_norm": 8.126254081726074, "learning_rate": 9.320282017934151e-05, "loss": 0.94, "step": 5716 }, { "epoch": 0.387356866996409, "grad_norm": 5.760578632354736, "learning_rate": 9.320145116024369e-05, "loss": 0.8861, "step": 5717 }, { "epoch": 0.38742462226438107, "grad_norm": 7.664496421813965, "learning_rate": 9.320008214114587e-05, "loss": 0.8005, "step": 5718 }, { "epoch": 0.38749237753235316, "grad_norm": 8.473989486694336, "learning_rate": 9.319871312204805e-05, "loss": 0.8196, "step": 5719 }, { "epoch": 0.38756013280032525, "grad_norm": 8.553350448608398, "learning_rate": 9.319734410295024e-05, "loss": 0.8356, "step": 5720 }, { "epoch": 0.3876278880682973, "grad_norm": 7.713801860809326, "learning_rate": 9.319597508385242e-05, "loss": 1.0326, "step": 5721 }, { "epoch": 0.3876956433362694, "grad_norm": 7.013209819793701, "learning_rate": 9.31946060647546e-05, "loss": 1.1149, "step": 5722 }, { "epoch": 0.38776339860424147, "grad_norm": 8.171217918395996, "learning_rate": 9.319323704565679e-05, "loss": 1.141, "step": 5723 }, { "epoch": 0.38783115387221356, "grad_norm": 8.459320068359375, "learning_rate": 9.319186802655897e-05, "loss": 0.8859, "step": 5724 }, { "epoch": 0.38789890914018565, "grad_norm": 6.681031227111816, "learning_rate": 9.319049900746116e-05, "loss": 0.8746, "step": 5725 }, { "epoch": 0.38796666440815775, "grad_norm": 6.796359539031982, "learning_rate": 9.318912998836334e-05, "loss": 1.0891, "step": 5726 }, { "epoch": 0.38803441967612984, "grad_norm": 6.819052696228027, "learning_rate": 9.318776096926552e-05, "loss": 0.8087, "step": 5727 }, { "epoch": 0.38810217494410193, "grad_norm": 6.746981620788574, "learning_rate": 9.31863919501677e-05, "loss": 1.1591, "step": 5728 }, { "epoch": 0.38816993021207397, "grad_norm": 7.767449855804443, "learning_rate": 9.318502293106988e-05, "loss": 1.1012, "step": 5729 }, { "epoch": 0.38823768548004606, "grad_norm": 7.967437744140625, "learning_rate": 9.318365391197207e-05, "loss": 1.0262, "step": 5730 }, { "epoch": 0.38830544074801815, "grad_norm": 6.053138256072998, "learning_rate": 9.318228489287426e-05, "loss": 0.6821, "step": 5731 }, { "epoch": 0.38837319601599024, "grad_norm": 7.298278331756592, "learning_rate": 9.318091587377644e-05, "loss": 1.0869, "step": 5732 }, { "epoch": 0.38844095128396233, "grad_norm": 6.621678352355957, "learning_rate": 9.317954685467863e-05, "loss": 0.9388, "step": 5733 }, { "epoch": 0.3885087065519344, "grad_norm": 7.588876247406006, "learning_rate": 9.317817783558081e-05, "loss": 0.8775, "step": 5734 }, { "epoch": 0.3885764618199065, "grad_norm": 6.0856242179870605, "learning_rate": 9.317680881648299e-05, "loss": 0.8057, "step": 5735 }, { "epoch": 0.3886442170878786, "grad_norm": 6.144415378570557, "learning_rate": 9.317543979738518e-05, "loss": 0.9205, "step": 5736 }, { "epoch": 0.38871197235585064, "grad_norm": 8.25857925415039, "learning_rate": 9.317407077828736e-05, "loss": 0.9608, "step": 5737 }, { "epoch": 0.38877972762382274, "grad_norm": 9.107572555541992, "learning_rate": 9.317270175918954e-05, "loss": 0.7823, "step": 5738 }, { "epoch": 0.3888474828917948, "grad_norm": 6.879073619842529, "learning_rate": 9.317133274009174e-05, "loss": 0.9137, "step": 5739 }, { "epoch": 0.3889152381597669, "grad_norm": 6.114928245544434, "learning_rate": 9.316996372099392e-05, "loss": 0.9713, "step": 5740 }, { "epoch": 0.388982993427739, "grad_norm": 6.454460620880127, "learning_rate": 9.31685947018961e-05, "loss": 0.9224, "step": 5741 }, { "epoch": 0.3890507486957111, "grad_norm": 7.231433868408203, "learning_rate": 9.316722568279828e-05, "loss": 0.9087, "step": 5742 }, { "epoch": 0.3891185039636832, "grad_norm": 7.442675590515137, "learning_rate": 9.316585666370047e-05, "loss": 0.9038, "step": 5743 }, { "epoch": 0.3891862592316553, "grad_norm": 6.472715854644775, "learning_rate": 9.316448764460265e-05, "loss": 0.7429, "step": 5744 }, { "epoch": 0.3892540144996273, "grad_norm": 7.20242977142334, "learning_rate": 9.316311862550483e-05, "loss": 0.9592, "step": 5745 }, { "epoch": 0.3893217697675994, "grad_norm": 5.8578715324401855, "learning_rate": 9.316174960640701e-05, "loss": 0.901, "step": 5746 }, { "epoch": 0.3893895250355715, "grad_norm": 5.1851983070373535, "learning_rate": 9.31603805873092e-05, "loss": 0.6872, "step": 5747 }, { "epoch": 0.3894572803035436, "grad_norm": 5.5100531578063965, "learning_rate": 9.315901156821139e-05, "loss": 0.8329, "step": 5748 }, { "epoch": 0.3895250355715157, "grad_norm": 7.71047830581665, "learning_rate": 9.315764254911357e-05, "loss": 0.8726, "step": 5749 }, { "epoch": 0.3895927908394878, "grad_norm": 5.928194522857666, "learning_rate": 9.315627353001575e-05, "loss": 0.6503, "step": 5750 }, { "epoch": 0.38966054610745987, "grad_norm": 6.636775016784668, "learning_rate": 9.315490451091793e-05, "loss": 0.8905, "step": 5751 }, { "epoch": 0.38972830137543196, "grad_norm": 7.5196027755737305, "learning_rate": 9.315353549182012e-05, "loss": 0.9333, "step": 5752 }, { "epoch": 0.389796056643404, "grad_norm": 7.1420087814331055, "learning_rate": 9.31521664727223e-05, "loss": 0.777, "step": 5753 }, { "epoch": 0.3898638119113761, "grad_norm": 6.713923931121826, "learning_rate": 9.315079745362448e-05, "loss": 0.7667, "step": 5754 }, { "epoch": 0.3899315671793482, "grad_norm": 6.81540584564209, "learning_rate": 9.314942843452666e-05, "loss": 0.8838, "step": 5755 }, { "epoch": 0.3899993224473203, "grad_norm": 7.069910049438477, "learning_rate": 9.314805941542885e-05, "loss": 0.9638, "step": 5756 }, { "epoch": 0.39006707771529237, "grad_norm": 6.350069999694824, "learning_rate": 9.314669039633104e-05, "loss": 0.8163, "step": 5757 }, { "epoch": 0.39013483298326446, "grad_norm": 5.552088737487793, "learning_rate": 9.314532137723322e-05, "loss": 0.7061, "step": 5758 }, { "epoch": 0.39020258825123655, "grad_norm": 7.8301167488098145, "learning_rate": 9.31439523581354e-05, "loss": 1.0415, "step": 5759 }, { "epoch": 0.39027034351920864, "grad_norm": 7.079575538635254, "learning_rate": 9.314258333903758e-05, "loss": 0.7247, "step": 5760 }, { "epoch": 0.3903380987871807, "grad_norm": 7.282689571380615, "learning_rate": 9.314121431993976e-05, "loss": 0.8361, "step": 5761 }, { "epoch": 0.39040585405515277, "grad_norm": 7.18900203704834, "learning_rate": 9.313984530084195e-05, "loss": 1.0206, "step": 5762 }, { "epoch": 0.39047360932312486, "grad_norm": 5.926472187042236, "learning_rate": 9.313847628174413e-05, "loss": 0.8551, "step": 5763 }, { "epoch": 0.39054136459109695, "grad_norm": 8.176214218139648, "learning_rate": 9.313710726264631e-05, "loss": 0.8106, "step": 5764 }, { "epoch": 0.39060911985906904, "grad_norm": 7.093206882476807, "learning_rate": 9.31357382435485e-05, "loss": 0.9986, "step": 5765 }, { "epoch": 0.39067687512704113, "grad_norm": 8.531807899475098, "learning_rate": 9.313436922445069e-05, "loss": 0.9368, "step": 5766 }, { "epoch": 0.3907446303950132, "grad_norm": 6.479072570800781, "learning_rate": 9.313300020535287e-05, "loss": 0.858, "step": 5767 }, { "epoch": 0.3908123856629853, "grad_norm": 6.749851226806641, "learning_rate": 9.313163118625505e-05, "loss": 0.638, "step": 5768 }, { "epoch": 0.3908801409309574, "grad_norm": 8.79270076751709, "learning_rate": 9.313026216715723e-05, "loss": 0.7734, "step": 5769 }, { "epoch": 0.39094789619892945, "grad_norm": 6.972496509552002, "learning_rate": 9.312889314805941e-05, "loss": 0.8557, "step": 5770 }, { "epoch": 0.39101565146690154, "grad_norm": 6.48140811920166, "learning_rate": 9.31275241289616e-05, "loss": 0.8183, "step": 5771 }, { "epoch": 0.39108340673487363, "grad_norm": 6.5230536460876465, "learning_rate": 9.312615510986378e-05, "loss": 0.7863, "step": 5772 }, { "epoch": 0.3911511620028457, "grad_norm": 6.982974052429199, "learning_rate": 9.312478609076597e-05, "loss": 0.9604, "step": 5773 }, { "epoch": 0.3912189172708178, "grad_norm": 7.524245262145996, "learning_rate": 9.312341707166815e-05, "loss": 0.8145, "step": 5774 }, { "epoch": 0.3912866725387899, "grad_norm": 7.770455837249756, "learning_rate": 9.312204805257034e-05, "loss": 0.858, "step": 5775 }, { "epoch": 0.391354427806762, "grad_norm": 6.145147800445557, "learning_rate": 9.312067903347252e-05, "loss": 0.6833, "step": 5776 }, { "epoch": 0.3914221830747341, "grad_norm": 7.118679523468018, "learning_rate": 9.31193100143747e-05, "loss": 0.9945, "step": 5777 }, { "epoch": 0.3914899383427061, "grad_norm": 7.793931007385254, "learning_rate": 9.311794099527688e-05, "loss": 1.0227, "step": 5778 }, { "epoch": 0.3915576936106782, "grad_norm": 6.189599514007568, "learning_rate": 9.311657197617907e-05, "loss": 1.0355, "step": 5779 }, { "epoch": 0.3916254488786503, "grad_norm": 8.16638469696045, "learning_rate": 9.311520295708125e-05, "loss": 0.998, "step": 5780 }, { "epoch": 0.3916932041466224, "grad_norm": 7.059429168701172, "learning_rate": 9.311383393798343e-05, "loss": 0.8641, "step": 5781 }, { "epoch": 0.3917609594145945, "grad_norm": 9.555390357971191, "learning_rate": 9.311246491888563e-05, "loss": 0.9243, "step": 5782 }, { "epoch": 0.3918287146825666, "grad_norm": 7.566288948059082, "learning_rate": 9.311109589978781e-05, "loss": 0.8025, "step": 5783 }, { "epoch": 0.3918964699505387, "grad_norm": 8.842116355895996, "learning_rate": 9.310972688068999e-05, "loss": 0.8875, "step": 5784 }, { "epoch": 0.39196422521851076, "grad_norm": 7.426934719085693, "learning_rate": 9.310835786159218e-05, "loss": 0.8116, "step": 5785 }, { "epoch": 0.3920319804864828, "grad_norm": 7.222875595092773, "learning_rate": 9.310698884249436e-05, "loss": 0.8286, "step": 5786 }, { "epoch": 0.3920997357544549, "grad_norm": 8.364874839782715, "learning_rate": 9.310561982339654e-05, "loss": 1.1628, "step": 5787 }, { "epoch": 0.392167491022427, "grad_norm": 6.4203104972839355, "learning_rate": 9.310425080429872e-05, "loss": 0.8501, "step": 5788 }, { "epoch": 0.3922352462903991, "grad_norm": 9.124777793884277, "learning_rate": 9.310288178520092e-05, "loss": 1.1962, "step": 5789 }, { "epoch": 0.39230300155837117, "grad_norm": 7.944338321685791, "learning_rate": 9.31015127661031e-05, "loss": 1.3275, "step": 5790 }, { "epoch": 0.39237075682634326, "grad_norm": 7.256115913391113, "learning_rate": 9.310014374700528e-05, "loss": 0.8594, "step": 5791 }, { "epoch": 0.39243851209431535, "grad_norm": 6.452229976654053, "learning_rate": 9.309877472790746e-05, "loss": 0.7744, "step": 5792 }, { "epoch": 0.39250626736228744, "grad_norm": 8.207562446594238, "learning_rate": 9.309740570880964e-05, "loss": 1.0615, "step": 5793 }, { "epoch": 0.3925740226302595, "grad_norm": 6.627633571624756, "learning_rate": 9.309603668971183e-05, "loss": 1.1951, "step": 5794 }, { "epoch": 0.39264177789823157, "grad_norm": 6.888060092926025, "learning_rate": 9.309466767061401e-05, "loss": 0.7771, "step": 5795 }, { "epoch": 0.39270953316620366, "grad_norm": 5.327685356140137, "learning_rate": 9.30932986515162e-05, "loss": 0.6568, "step": 5796 }, { "epoch": 0.39277728843417575, "grad_norm": 7.033134460449219, "learning_rate": 9.309192963241837e-05, "loss": 0.8601, "step": 5797 }, { "epoch": 0.39284504370214784, "grad_norm": 6.694526195526123, "learning_rate": 9.309056061332057e-05, "loss": 0.9084, "step": 5798 }, { "epoch": 0.39291279897011994, "grad_norm": 6.9023027420043945, "learning_rate": 9.308919159422275e-05, "loss": 0.8983, "step": 5799 }, { "epoch": 0.39298055423809203, "grad_norm": 9.527570724487305, "learning_rate": 9.308782257512493e-05, "loss": 0.9773, "step": 5800 }, { "epoch": 0.3930483095060641, "grad_norm": 7.497427940368652, "learning_rate": 9.308645355602711e-05, "loss": 0.9713, "step": 5801 }, { "epoch": 0.39311606477403616, "grad_norm": 8.860198020935059, "learning_rate": 9.308508453692929e-05, "loss": 0.8866, "step": 5802 }, { "epoch": 0.39318382004200825, "grad_norm": 7.34425687789917, "learning_rate": 9.308371551783148e-05, "loss": 0.9442, "step": 5803 }, { "epoch": 0.39325157530998034, "grad_norm": 10.30170726776123, "learning_rate": 9.308234649873366e-05, "loss": 0.8957, "step": 5804 }, { "epoch": 0.39331933057795243, "grad_norm": 8.141131401062012, "learning_rate": 9.308097747963584e-05, "loss": 1.1858, "step": 5805 }, { "epoch": 0.3933870858459245, "grad_norm": 7.154587268829346, "learning_rate": 9.307960846053802e-05, "loss": 0.9056, "step": 5806 }, { "epoch": 0.3934548411138966, "grad_norm": 6.086984634399414, "learning_rate": 9.30782394414402e-05, "loss": 0.8943, "step": 5807 }, { "epoch": 0.3935225963818687, "grad_norm": 6.48331880569458, "learning_rate": 9.30768704223424e-05, "loss": 0.8631, "step": 5808 }, { "epoch": 0.3935903516498408, "grad_norm": 9.434244155883789, "learning_rate": 9.307550140324458e-05, "loss": 1.0951, "step": 5809 }, { "epoch": 0.39365810691781283, "grad_norm": 7.943146228790283, "learning_rate": 9.307413238414676e-05, "loss": 0.9073, "step": 5810 }, { "epoch": 0.3937258621857849, "grad_norm": 6.200165271759033, "learning_rate": 9.307276336504894e-05, "loss": 0.8406, "step": 5811 }, { "epoch": 0.393793617453757, "grad_norm": 6.221752643585205, "learning_rate": 9.307139434595113e-05, "loss": 0.9181, "step": 5812 }, { "epoch": 0.3938613727217291, "grad_norm": 6.645714282989502, "learning_rate": 9.307002532685331e-05, "loss": 0.8508, "step": 5813 }, { "epoch": 0.3939291279897012, "grad_norm": 7.978506088256836, "learning_rate": 9.30686563077555e-05, "loss": 1.1544, "step": 5814 }, { "epoch": 0.3939968832576733, "grad_norm": 6.967216968536377, "learning_rate": 9.306728728865767e-05, "loss": 0.7956, "step": 5815 }, { "epoch": 0.3940646385256454, "grad_norm": 9.8863525390625, "learning_rate": 9.306591826955986e-05, "loss": 0.9111, "step": 5816 }, { "epoch": 0.3941323937936175, "grad_norm": 7.202954292297363, "learning_rate": 9.306454925046205e-05, "loss": 0.9267, "step": 5817 }, { "epoch": 0.3942001490615895, "grad_norm": 6.386566162109375, "learning_rate": 9.306318023136423e-05, "loss": 0.9494, "step": 5818 }, { "epoch": 0.3942679043295616, "grad_norm": 6.505050182342529, "learning_rate": 9.306181121226641e-05, "loss": 0.904, "step": 5819 }, { "epoch": 0.3943356595975337, "grad_norm": 7.091371536254883, "learning_rate": 9.306044219316859e-05, "loss": 0.7305, "step": 5820 }, { "epoch": 0.3944034148655058, "grad_norm": 7.322198390960693, "learning_rate": 9.305907317407078e-05, "loss": 0.8101, "step": 5821 }, { "epoch": 0.3944711701334779, "grad_norm": 6.345345497131348, "learning_rate": 9.305770415497296e-05, "loss": 0.7821, "step": 5822 }, { "epoch": 0.39453892540144997, "grad_norm": 6.418498992919922, "learning_rate": 9.305633513587514e-05, "loss": 0.8844, "step": 5823 }, { "epoch": 0.39460668066942206, "grad_norm": 7.375457763671875, "learning_rate": 9.305496611677733e-05, "loss": 1.0099, "step": 5824 }, { "epoch": 0.39467443593739415, "grad_norm": 7.758962154388428, "learning_rate": 9.305359709767952e-05, "loss": 1.1459, "step": 5825 }, { "epoch": 0.39474219120536624, "grad_norm": 9.552523612976074, "learning_rate": 9.30522280785817e-05, "loss": 1.0285, "step": 5826 }, { "epoch": 0.3948099464733383, "grad_norm": 7.053111553192139, "learning_rate": 9.305085905948388e-05, "loss": 0.8006, "step": 5827 }, { "epoch": 0.39487770174131037, "grad_norm": 7.239533424377441, "learning_rate": 9.304949004038607e-05, "loss": 0.9594, "step": 5828 }, { "epoch": 0.39494545700928246, "grad_norm": 5.681763172149658, "learning_rate": 9.304812102128825e-05, "loss": 0.7517, "step": 5829 }, { "epoch": 0.39501321227725456, "grad_norm": 7.394958019256592, "learning_rate": 9.304675200219043e-05, "loss": 0.9656, "step": 5830 }, { "epoch": 0.39508096754522665, "grad_norm": 5.894937515258789, "learning_rate": 9.304538298309263e-05, "loss": 0.7568, "step": 5831 }, { "epoch": 0.39514872281319874, "grad_norm": 7.866422653198242, "learning_rate": 9.304401396399481e-05, "loss": 1.1937, "step": 5832 }, { "epoch": 0.39521647808117083, "grad_norm": 6.807196140289307, "learning_rate": 9.304264494489699e-05, "loss": 0.7785, "step": 5833 }, { "epoch": 0.3952842333491429, "grad_norm": 7.117645740509033, "learning_rate": 9.304127592579917e-05, "loss": 0.8462, "step": 5834 }, { "epoch": 0.39535198861711496, "grad_norm": 5.873225212097168, "learning_rate": 9.303990690670136e-05, "loss": 0.8372, "step": 5835 }, { "epoch": 0.39541974388508705, "grad_norm": 6.9400410652160645, "learning_rate": 9.303853788760354e-05, "loss": 0.8154, "step": 5836 }, { "epoch": 0.39548749915305914, "grad_norm": 5.7771100997924805, "learning_rate": 9.303716886850572e-05, "loss": 0.7411, "step": 5837 }, { "epoch": 0.39555525442103123, "grad_norm": 6.784138202667236, "learning_rate": 9.30357998494079e-05, "loss": 1.062, "step": 5838 }, { "epoch": 0.3956230096890033, "grad_norm": 7.726989269256592, "learning_rate": 9.303443083031008e-05, "loss": 0.9726, "step": 5839 }, { "epoch": 0.3956907649569754, "grad_norm": 6.342170715332031, "learning_rate": 9.303306181121228e-05, "loss": 0.8388, "step": 5840 }, { "epoch": 0.3957585202249475, "grad_norm": 8.047952651977539, "learning_rate": 9.303169279211446e-05, "loss": 1.0734, "step": 5841 }, { "epoch": 0.3958262754929196, "grad_norm": 7.073848724365234, "learning_rate": 9.303032377301664e-05, "loss": 0.6968, "step": 5842 }, { "epoch": 0.39589403076089164, "grad_norm": 7.159115791320801, "learning_rate": 9.302895475391882e-05, "loss": 0.7991, "step": 5843 }, { "epoch": 0.3959617860288637, "grad_norm": 6.908319473266602, "learning_rate": 9.302758573482101e-05, "loss": 0.9283, "step": 5844 }, { "epoch": 0.3960295412968358, "grad_norm": 6.5849690437316895, "learning_rate": 9.30262167157232e-05, "loss": 0.8258, "step": 5845 }, { "epoch": 0.3960972965648079, "grad_norm": 8.763665199279785, "learning_rate": 9.302484769662537e-05, "loss": 0.7734, "step": 5846 }, { "epoch": 0.39616505183278, "grad_norm": 7.134938716888428, "learning_rate": 9.302347867752755e-05, "loss": 0.8084, "step": 5847 }, { "epoch": 0.3962328071007521, "grad_norm": 6.418683052062988, "learning_rate": 9.302210965842973e-05, "loss": 0.8225, "step": 5848 }, { "epoch": 0.3963005623687242, "grad_norm": 7.124704837799072, "learning_rate": 9.302074063933193e-05, "loss": 0.6787, "step": 5849 }, { "epoch": 0.3963683176366963, "grad_norm": 6.653177261352539, "learning_rate": 9.301937162023411e-05, "loss": 0.8026, "step": 5850 }, { "epoch": 0.3964360729046683, "grad_norm": 7.981827735900879, "learning_rate": 9.301800260113629e-05, "loss": 0.9281, "step": 5851 }, { "epoch": 0.3965038281726404, "grad_norm": 5.823386192321777, "learning_rate": 9.301663358203847e-05, "loss": 0.8158, "step": 5852 }, { "epoch": 0.3965715834406125, "grad_norm": 8.96346664428711, "learning_rate": 9.301526456294066e-05, "loss": 1.0875, "step": 5853 }, { "epoch": 0.3966393387085846, "grad_norm": 6.114203929901123, "learning_rate": 9.301389554384284e-05, "loss": 0.7896, "step": 5854 }, { "epoch": 0.3967070939765567, "grad_norm": 6.386680603027344, "learning_rate": 9.301252652474502e-05, "loss": 1.0199, "step": 5855 }, { "epoch": 0.39677484924452877, "grad_norm": 8.502519607543945, "learning_rate": 9.30111575056472e-05, "loss": 1.0127, "step": 5856 }, { "epoch": 0.39684260451250086, "grad_norm": 6.875016689300537, "learning_rate": 9.300978848654938e-05, "loss": 0.8263, "step": 5857 }, { "epoch": 0.39691035978047295, "grad_norm": 8.344440460205078, "learning_rate": 9.300841946745158e-05, "loss": 1.0351, "step": 5858 }, { "epoch": 0.396978115048445, "grad_norm": 6.42828893661499, "learning_rate": 9.300705044835376e-05, "loss": 0.8503, "step": 5859 }, { "epoch": 0.3970458703164171, "grad_norm": 6.0403547286987305, "learning_rate": 9.300568142925594e-05, "loss": 0.9078, "step": 5860 }, { "epoch": 0.3971136255843892, "grad_norm": 5.010101795196533, "learning_rate": 9.300431241015812e-05, "loss": 0.8215, "step": 5861 }, { "epoch": 0.39718138085236127, "grad_norm": 6.132750988006592, "learning_rate": 9.30029433910603e-05, "loss": 0.7887, "step": 5862 }, { "epoch": 0.39724913612033336, "grad_norm": 6.075906753540039, "learning_rate": 9.30015743719625e-05, "loss": 0.822, "step": 5863 }, { "epoch": 0.39731689138830545, "grad_norm": 8.225122451782227, "learning_rate": 9.300020535286467e-05, "loss": 0.7771, "step": 5864 }, { "epoch": 0.39738464665627754, "grad_norm": 6.863472938537598, "learning_rate": 9.299883633376685e-05, "loss": 0.9185, "step": 5865 }, { "epoch": 0.39745240192424963, "grad_norm": 7.280022144317627, "learning_rate": 9.299746731466903e-05, "loss": 0.8206, "step": 5866 }, { "epoch": 0.39752015719222167, "grad_norm": 5.646960258483887, "learning_rate": 9.299609829557123e-05, "loss": 0.7897, "step": 5867 }, { "epoch": 0.39758791246019376, "grad_norm": 8.248296737670898, "learning_rate": 9.299472927647341e-05, "loss": 0.802, "step": 5868 }, { "epoch": 0.39765566772816585, "grad_norm": 6.389570236206055, "learning_rate": 9.299336025737559e-05, "loss": 0.7744, "step": 5869 }, { "epoch": 0.39772342299613794, "grad_norm": 6.452336311340332, "learning_rate": 9.299199123827777e-05, "loss": 0.9003, "step": 5870 }, { "epoch": 0.39779117826411003, "grad_norm": 6.798274040222168, "learning_rate": 9.299062221917995e-05, "loss": 0.9585, "step": 5871 }, { "epoch": 0.3978589335320821, "grad_norm": 6.267078876495361, "learning_rate": 9.298925320008214e-05, "loss": 0.7971, "step": 5872 }, { "epoch": 0.3979266888000542, "grad_norm": 7.0529069900512695, "learning_rate": 9.298788418098432e-05, "loss": 0.9384, "step": 5873 }, { "epoch": 0.3979944440680263, "grad_norm": 6.686244964599609, "learning_rate": 9.298651516188652e-05, "loss": 0.8589, "step": 5874 }, { "epoch": 0.3980621993359984, "grad_norm": 7.230234622955322, "learning_rate": 9.29851461427887e-05, "loss": 0.8638, "step": 5875 }, { "epoch": 0.39812995460397044, "grad_norm": 8.14789867401123, "learning_rate": 9.298377712369088e-05, "loss": 0.7231, "step": 5876 }, { "epoch": 0.39819770987194253, "grad_norm": 6.5517497062683105, "learning_rate": 9.298240810459307e-05, "loss": 1.0435, "step": 5877 }, { "epoch": 0.3982654651399146, "grad_norm": 7.095836639404297, "learning_rate": 9.298103908549525e-05, "loss": 0.7551, "step": 5878 }, { "epoch": 0.3983332204078867, "grad_norm": 6.932202339172363, "learning_rate": 9.297967006639743e-05, "loss": 0.9075, "step": 5879 }, { "epoch": 0.3984009756758588, "grad_norm": 8.045002937316895, "learning_rate": 9.297830104729961e-05, "loss": 0.9787, "step": 5880 }, { "epoch": 0.3984687309438309, "grad_norm": 7.1430511474609375, "learning_rate": 9.297693202820181e-05, "loss": 0.8549, "step": 5881 }, { "epoch": 0.398536486211803, "grad_norm": 5.857006549835205, "learning_rate": 9.297556300910399e-05, "loss": 0.7573, "step": 5882 }, { "epoch": 0.3986042414797751, "grad_norm": 6.143594741821289, "learning_rate": 9.297419399000617e-05, "loss": 0.747, "step": 5883 }, { "epoch": 0.3986719967477471, "grad_norm": 7.4967169761657715, "learning_rate": 9.297282497090835e-05, "loss": 1.0366, "step": 5884 }, { "epoch": 0.3987397520157192, "grad_norm": 6.6401166915893555, "learning_rate": 9.297145595181054e-05, "loss": 0.9889, "step": 5885 }, { "epoch": 0.3988075072836913, "grad_norm": 5.577928066253662, "learning_rate": 9.297008693271272e-05, "loss": 0.6917, "step": 5886 }, { "epoch": 0.3988752625516634, "grad_norm": 5.9933061599731445, "learning_rate": 9.29687179136149e-05, "loss": 0.9545, "step": 5887 }, { "epoch": 0.3989430178196355, "grad_norm": 6.353165149688721, "learning_rate": 9.296734889451708e-05, "loss": 0.7512, "step": 5888 }, { "epoch": 0.3990107730876076, "grad_norm": 6.181026458740234, "learning_rate": 9.296597987541926e-05, "loss": 0.767, "step": 5889 }, { "epoch": 0.39907852835557966, "grad_norm": 6.579110622406006, "learning_rate": 9.296461085632146e-05, "loss": 1.0604, "step": 5890 }, { "epoch": 0.39914628362355176, "grad_norm": 6.501206398010254, "learning_rate": 9.296324183722364e-05, "loss": 0.9821, "step": 5891 }, { "epoch": 0.3992140388915238, "grad_norm": 8.253642082214355, "learning_rate": 9.296187281812582e-05, "loss": 0.8827, "step": 5892 }, { "epoch": 0.3992817941594959, "grad_norm": 7.590344429016113, "learning_rate": 9.2960503799028e-05, "loss": 0.9614, "step": 5893 }, { "epoch": 0.399349549427468, "grad_norm": 6.755953311920166, "learning_rate": 9.295913477993018e-05, "loss": 0.8298, "step": 5894 }, { "epoch": 0.39941730469544007, "grad_norm": 6.040559768676758, "learning_rate": 9.295776576083237e-05, "loss": 0.7182, "step": 5895 }, { "epoch": 0.39948505996341216, "grad_norm": 6.301966667175293, "learning_rate": 9.295639674173455e-05, "loss": 0.8545, "step": 5896 }, { "epoch": 0.39955281523138425, "grad_norm": 5.77929162979126, "learning_rate": 9.295502772263673e-05, "loss": 0.9398, "step": 5897 }, { "epoch": 0.39962057049935634, "grad_norm": 6.633763313293457, "learning_rate": 9.295365870353891e-05, "loss": 0.8914, "step": 5898 }, { "epoch": 0.39968832576732843, "grad_norm": 7.260994911193848, "learning_rate": 9.295228968444111e-05, "loss": 0.7725, "step": 5899 }, { "epoch": 0.39975608103530047, "grad_norm": 8.820511817932129, "learning_rate": 9.295092066534329e-05, "loss": 0.7847, "step": 5900 }, { "epoch": 0.39982383630327256, "grad_norm": 7.335788726806641, "learning_rate": 9.294955164624547e-05, "loss": 0.683, "step": 5901 }, { "epoch": 0.39989159157124465, "grad_norm": 7.959702491760254, "learning_rate": 9.294818262714765e-05, "loss": 0.7598, "step": 5902 }, { "epoch": 0.39995934683921675, "grad_norm": 9.680485725402832, "learning_rate": 9.294681360804983e-05, "loss": 1.0359, "step": 5903 }, { "epoch": 0.40002710210718884, "grad_norm": 6.917464733123779, "learning_rate": 9.294544458895202e-05, "loss": 1.0155, "step": 5904 }, { "epoch": 0.40009485737516093, "grad_norm": 6.221781253814697, "learning_rate": 9.29440755698542e-05, "loss": 0.9645, "step": 5905 }, { "epoch": 0.400162612643133, "grad_norm": 6.778574466705322, "learning_rate": 9.294270655075638e-05, "loss": 0.794, "step": 5906 }, { "epoch": 0.4002303679111051, "grad_norm": 7.8015522956848145, "learning_rate": 9.294133753165856e-05, "loss": 0.9472, "step": 5907 }, { "epoch": 0.40029812317907715, "grad_norm": 5.018773555755615, "learning_rate": 9.293996851256076e-05, "loss": 0.8961, "step": 5908 }, { "epoch": 0.40036587844704924, "grad_norm": 6.184563636779785, "learning_rate": 9.293859949346294e-05, "loss": 0.7988, "step": 5909 }, { "epoch": 0.40043363371502133, "grad_norm": 6.593270778656006, "learning_rate": 9.293723047436512e-05, "loss": 0.8248, "step": 5910 }, { "epoch": 0.4005013889829934, "grad_norm": 7.14009952545166, "learning_rate": 9.29358614552673e-05, "loss": 0.8589, "step": 5911 }, { "epoch": 0.4005691442509655, "grad_norm": 8.044157981872559, "learning_rate": 9.293449243616948e-05, "loss": 0.9397, "step": 5912 }, { "epoch": 0.4006368995189376, "grad_norm": 7.3374247550964355, "learning_rate": 9.293312341707167e-05, "loss": 0.8615, "step": 5913 }, { "epoch": 0.4007046547869097, "grad_norm": 6.510500907897949, "learning_rate": 9.293175439797385e-05, "loss": 0.7151, "step": 5914 }, { "epoch": 0.4007724100548818, "grad_norm": 15.020646095275879, "learning_rate": 9.293038537887603e-05, "loss": 0.9349, "step": 5915 }, { "epoch": 0.4008401653228538, "grad_norm": 6.779942989349365, "learning_rate": 9.292901635977821e-05, "loss": 0.8018, "step": 5916 }, { "epoch": 0.4009079205908259, "grad_norm": 6.3740339279174805, "learning_rate": 9.29276473406804e-05, "loss": 0.7691, "step": 5917 }, { "epoch": 0.400975675858798, "grad_norm": 6.902404308319092, "learning_rate": 9.292627832158259e-05, "loss": 0.7431, "step": 5918 }, { "epoch": 0.4010434311267701, "grad_norm": 7.324024200439453, "learning_rate": 9.292490930248477e-05, "loss": 0.9314, "step": 5919 }, { "epoch": 0.4011111863947422, "grad_norm": 8.297179222106934, "learning_rate": 9.292354028338695e-05, "loss": 1.0072, "step": 5920 }, { "epoch": 0.4011789416627143, "grad_norm": 6.1418914794921875, "learning_rate": 9.292217126428914e-05, "loss": 0.948, "step": 5921 }, { "epoch": 0.4012466969306864, "grad_norm": 8.698518753051758, "learning_rate": 9.292080224519132e-05, "loss": 0.8509, "step": 5922 }, { "epoch": 0.40131445219865847, "grad_norm": 6.628043174743652, "learning_rate": 9.29194332260935e-05, "loss": 0.9388, "step": 5923 }, { "epoch": 0.4013822074666305, "grad_norm": 8.38621711730957, "learning_rate": 9.29180642069957e-05, "loss": 1.0795, "step": 5924 }, { "epoch": 0.4014499627346026, "grad_norm": 7.546327114105225, "learning_rate": 9.291669518789788e-05, "loss": 0.865, "step": 5925 }, { "epoch": 0.4015177180025747, "grad_norm": 7.399687767028809, "learning_rate": 9.291532616880006e-05, "loss": 0.8276, "step": 5926 }, { "epoch": 0.4015854732705468, "grad_norm": 6.921968460083008, "learning_rate": 9.291395714970225e-05, "loss": 0.9275, "step": 5927 }, { "epoch": 0.40165322853851887, "grad_norm": 7.856932640075684, "learning_rate": 9.291258813060443e-05, "loss": 1.1376, "step": 5928 }, { "epoch": 0.40172098380649096, "grad_norm": 7.86414098739624, "learning_rate": 9.291121911150661e-05, "loss": 1.0225, "step": 5929 }, { "epoch": 0.40178873907446305, "grad_norm": 7.028629302978516, "learning_rate": 9.29098500924088e-05, "loss": 1.0157, "step": 5930 }, { "epoch": 0.40185649434243514, "grad_norm": 5.826450347900391, "learning_rate": 9.290848107331099e-05, "loss": 0.8605, "step": 5931 }, { "epoch": 0.40192424961040724, "grad_norm": 6.072175025939941, "learning_rate": 9.290711205421317e-05, "loss": 0.6261, "step": 5932 }, { "epoch": 0.40199200487837927, "grad_norm": 6.829746246337891, "learning_rate": 9.290574303511535e-05, "loss": 0.849, "step": 5933 }, { "epoch": 0.40205976014635136, "grad_norm": 6.809370994567871, "learning_rate": 9.290437401601753e-05, "loss": 0.8659, "step": 5934 }, { "epoch": 0.40212751541432346, "grad_norm": 7.195353984832764, "learning_rate": 9.290300499691971e-05, "loss": 0.9559, "step": 5935 }, { "epoch": 0.40219527068229555, "grad_norm": 8.443836212158203, "learning_rate": 9.29016359778219e-05, "loss": 0.7687, "step": 5936 }, { "epoch": 0.40226302595026764, "grad_norm": 5.932136058807373, "learning_rate": 9.290026695872408e-05, "loss": 0.7125, "step": 5937 }, { "epoch": 0.40233078121823973, "grad_norm": 6.654352188110352, "learning_rate": 9.289889793962626e-05, "loss": 0.9933, "step": 5938 }, { "epoch": 0.4023985364862118, "grad_norm": 7.100243091583252, "learning_rate": 9.289752892052844e-05, "loss": 0.9271, "step": 5939 }, { "epoch": 0.4024662917541839, "grad_norm": 7.78497838973999, "learning_rate": 9.289615990143062e-05, "loss": 0.9358, "step": 5940 }, { "epoch": 0.40253404702215595, "grad_norm": 6.294493198394775, "learning_rate": 9.289479088233282e-05, "loss": 0.8308, "step": 5941 }, { "epoch": 0.40260180229012804, "grad_norm": 6.8807244300842285, "learning_rate": 9.2893421863235e-05, "loss": 0.7728, "step": 5942 }, { "epoch": 0.40266955755810013, "grad_norm": 7.030758857727051, "learning_rate": 9.289205284413718e-05, "loss": 0.8725, "step": 5943 }, { "epoch": 0.4027373128260722, "grad_norm": 6.787613868713379, "learning_rate": 9.289068382503936e-05, "loss": 0.8725, "step": 5944 }, { "epoch": 0.4028050680940443, "grad_norm": 6.4748148918151855, "learning_rate": 9.288931480594155e-05, "loss": 0.9305, "step": 5945 }, { "epoch": 0.4028728233620164, "grad_norm": 5.930908203125, "learning_rate": 9.288794578684373e-05, "loss": 0.814, "step": 5946 }, { "epoch": 0.4029405786299885, "grad_norm": 6.687366008758545, "learning_rate": 9.288657676774591e-05, "loss": 0.7893, "step": 5947 }, { "epoch": 0.4030083338979606, "grad_norm": 9.194374084472656, "learning_rate": 9.28852077486481e-05, "loss": 1.1462, "step": 5948 }, { "epoch": 0.4030760891659326, "grad_norm": 8.393781661987305, "learning_rate": 9.288383872955027e-05, "loss": 0.9375, "step": 5949 }, { "epoch": 0.4031438444339047, "grad_norm": 7.802282810211182, "learning_rate": 9.288246971045247e-05, "loss": 1.1499, "step": 5950 }, { "epoch": 0.4032115997018768, "grad_norm": 7.085485458374023, "learning_rate": 9.288110069135465e-05, "loss": 0.9834, "step": 5951 }, { "epoch": 0.4032793549698489, "grad_norm": 7.905081272125244, "learning_rate": 9.287973167225683e-05, "loss": 1.11, "step": 5952 }, { "epoch": 0.403347110237821, "grad_norm": 5.531884670257568, "learning_rate": 9.287836265315901e-05, "loss": 0.8386, "step": 5953 }, { "epoch": 0.4034148655057931, "grad_norm": 5.959394931793213, "learning_rate": 9.28769936340612e-05, "loss": 0.858, "step": 5954 }, { "epoch": 0.4034826207737652, "grad_norm": 7.020748138427734, "learning_rate": 9.287562461496338e-05, "loss": 0.8708, "step": 5955 }, { "epoch": 0.40355037604173727, "grad_norm": 9.01052474975586, "learning_rate": 9.287425559586556e-05, "loss": 1.0213, "step": 5956 }, { "epoch": 0.4036181313097093, "grad_norm": 7.091532230377197, "learning_rate": 9.287288657676774e-05, "loss": 1.0133, "step": 5957 }, { "epoch": 0.4036858865776814, "grad_norm": 5.7289276123046875, "learning_rate": 9.287151755766992e-05, "loss": 0.6517, "step": 5958 }, { "epoch": 0.4037536418456535, "grad_norm": 5.692935466766357, "learning_rate": 9.287014853857212e-05, "loss": 0.865, "step": 5959 }, { "epoch": 0.4038213971136256, "grad_norm": 7.469212532043457, "learning_rate": 9.28687795194743e-05, "loss": 0.8707, "step": 5960 }, { "epoch": 0.40388915238159767, "grad_norm": 6.172707557678223, "learning_rate": 9.286741050037648e-05, "loss": 0.8217, "step": 5961 }, { "epoch": 0.40395690764956976, "grad_norm": 8.741066932678223, "learning_rate": 9.286604148127866e-05, "loss": 0.8087, "step": 5962 }, { "epoch": 0.40402466291754185, "grad_norm": 6.240641117095947, "learning_rate": 9.286467246218085e-05, "loss": 1.1012, "step": 5963 }, { "epoch": 0.40409241818551395, "grad_norm": 6.801406383514404, "learning_rate": 9.286330344308303e-05, "loss": 0.9335, "step": 5964 }, { "epoch": 0.404160173453486, "grad_norm": 6.114485263824463, "learning_rate": 9.286193442398521e-05, "loss": 0.8987, "step": 5965 }, { "epoch": 0.4042279287214581, "grad_norm": 9.715784072875977, "learning_rate": 9.28605654048874e-05, "loss": 1.0337, "step": 5966 }, { "epoch": 0.40429568398943017, "grad_norm": 10.362801551818848, "learning_rate": 9.285919638578959e-05, "loss": 0.7685, "step": 5967 }, { "epoch": 0.40436343925740226, "grad_norm": 7.567534446716309, "learning_rate": 9.285782736669177e-05, "loss": 0.8397, "step": 5968 }, { "epoch": 0.40443119452537435, "grad_norm": 7.282614231109619, "learning_rate": 9.285645834759395e-05, "loss": 1.0515, "step": 5969 }, { "epoch": 0.40449894979334644, "grad_norm": 6.683037757873535, "learning_rate": 9.285508932849614e-05, "loss": 1.1006, "step": 5970 }, { "epoch": 0.40456670506131853, "grad_norm": 8.437498092651367, "learning_rate": 9.285372030939832e-05, "loss": 1.0656, "step": 5971 }, { "epoch": 0.4046344603292906, "grad_norm": 6.153039932250977, "learning_rate": 9.28523512903005e-05, "loss": 0.8285, "step": 5972 }, { "epoch": 0.40470221559726266, "grad_norm": 6.3334221839904785, "learning_rate": 9.28509822712027e-05, "loss": 0.8968, "step": 5973 }, { "epoch": 0.40476997086523475, "grad_norm": 9.119121551513672, "learning_rate": 9.284961325210488e-05, "loss": 0.8814, "step": 5974 }, { "epoch": 0.40483772613320684, "grad_norm": 7.621852397918701, "learning_rate": 9.284824423300706e-05, "loss": 0.9297, "step": 5975 }, { "epoch": 0.40490548140117893, "grad_norm": 7.415964126586914, "learning_rate": 9.284687521390924e-05, "loss": 0.8597, "step": 5976 }, { "epoch": 0.404973236669151, "grad_norm": 6.424054145812988, "learning_rate": 9.284550619481143e-05, "loss": 0.6792, "step": 5977 }, { "epoch": 0.4050409919371231, "grad_norm": 7.875925540924072, "learning_rate": 9.284413717571361e-05, "loss": 0.7931, "step": 5978 }, { "epoch": 0.4051087472050952, "grad_norm": 5.702389240264893, "learning_rate": 9.284276815661579e-05, "loss": 0.7124, "step": 5979 }, { "epoch": 0.4051765024730673, "grad_norm": 6.58071231842041, "learning_rate": 9.284139913751797e-05, "loss": 0.8965, "step": 5980 }, { "epoch": 0.4052442577410394, "grad_norm": 8.180785179138184, "learning_rate": 9.284003011842015e-05, "loss": 1.0545, "step": 5981 }, { "epoch": 0.40531201300901143, "grad_norm": 7.393392562866211, "learning_rate": 9.283866109932235e-05, "loss": 0.8044, "step": 5982 }, { "epoch": 0.4053797682769835, "grad_norm": 6.009011745452881, "learning_rate": 9.283729208022453e-05, "loss": 0.7688, "step": 5983 }, { "epoch": 0.4054475235449556, "grad_norm": 6.8165388107299805, "learning_rate": 9.283592306112671e-05, "loss": 0.7743, "step": 5984 }, { "epoch": 0.4055152788129277, "grad_norm": 7.851406574249268, "learning_rate": 9.283455404202889e-05, "loss": 0.9081, "step": 5985 }, { "epoch": 0.4055830340808998, "grad_norm": 7.138257026672363, "learning_rate": 9.283318502293108e-05, "loss": 0.9843, "step": 5986 }, { "epoch": 0.4056507893488719, "grad_norm": 7.70706033706665, "learning_rate": 9.283181600383326e-05, "loss": 0.9158, "step": 5987 }, { "epoch": 0.405718544616844, "grad_norm": 7.948725700378418, "learning_rate": 9.283044698473544e-05, "loss": 0.8694, "step": 5988 }, { "epoch": 0.40578629988481607, "grad_norm": 6.4174628257751465, "learning_rate": 9.282907796563762e-05, "loss": 0.8547, "step": 5989 }, { "epoch": 0.4058540551527881, "grad_norm": 7.62558126449585, "learning_rate": 9.28277089465398e-05, "loss": 0.7978, "step": 5990 }, { "epoch": 0.4059218104207602, "grad_norm": 7.624577522277832, "learning_rate": 9.2826339927442e-05, "loss": 1.0346, "step": 5991 }, { "epoch": 0.4059895656887323, "grad_norm": 7.392852783203125, "learning_rate": 9.282497090834418e-05, "loss": 0.8833, "step": 5992 }, { "epoch": 0.4060573209567044, "grad_norm": 7.652538776397705, "learning_rate": 9.282360188924636e-05, "loss": 1.0438, "step": 5993 }, { "epoch": 0.4061250762246765, "grad_norm": 7.050436973571777, "learning_rate": 9.282223287014854e-05, "loss": 0.8291, "step": 5994 }, { "epoch": 0.40619283149264857, "grad_norm": 7.119441032409668, "learning_rate": 9.282086385105072e-05, "loss": 0.6427, "step": 5995 }, { "epoch": 0.40626058676062066, "grad_norm": 6.1112565994262695, "learning_rate": 9.281949483195291e-05, "loss": 0.7737, "step": 5996 }, { "epoch": 0.40632834202859275, "grad_norm": 6.173165321350098, "learning_rate": 9.28181258128551e-05, "loss": 0.8099, "step": 5997 }, { "epoch": 0.4063960972965648, "grad_norm": 6.372697830200195, "learning_rate": 9.281675679375727e-05, "loss": 0.7902, "step": 5998 }, { "epoch": 0.4064638525645369, "grad_norm": 7.213540077209473, "learning_rate": 9.281538777465945e-05, "loss": 0.7273, "step": 5999 }, { "epoch": 0.40653160783250897, "grad_norm": 8.106986999511719, "learning_rate": 9.281401875556165e-05, "loss": 0.7516, "step": 6000 }, { "epoch": 0.40659936310048106, "grad_norm": 8.372703552246094, "learning_rate": 9.281264973646383e-05, "loss": 1.1737, "step": 6001 }, { "epoch": 0.40666711836845315, "grad_norm": 9.540267944335938, "learning_rate": 9.281128071736601e-05, "loss": 1.068, "step": 6002 }, { "epoch": 0.40673487363642524, "grad_norm": 6.878968715667725, "learning_rate": 9.280991169826819e-05, "loss": 0.9329, "step": 6003 }, { "epoch": 0.40680262890439733, "grad_norm": 6.826279163360596, "learning_rate": 9.280854267917037e-05, "loss": 0.8267, "step": 6004 }, { "epoch": 0.4068703841723694, "grad_norm": 7.155866622924805, "learning_rate": 9.280717366007256e-05, "loss": 0.9133, "step": 6005 }, { "epoch": 0.40693813944034146, "grad_norm": 6.883568286895752, "learning_rate": 9.280580464097474e-05, "loss": 0.709, "step": 6006 }, { "epoch": 0.40700589470831355, "grad_norm": 6.944139003753662, "learning_rate": 9.280443562187692e-05, "loss": 0.8137, "step": 6007 }, { "epoch": 0.40707364997628565, "grad_norm": 5.899077892303467, "learning_rate": 9.28030666027791e-05, "loss": 0.733, "step": 6008 }, { "epoch": 0.40714140524425774, "grad_norm": 7.213099956512451, "learning_rate": 9.28016975836813e-05, "loss": 1.1728, "step": 6009 }, { "epoch": 0.40720916051222983, "grad_norm": 7.830915927886963, "learning_rate": 9.280032856458348e-05, "loss": 0.7465, "step": 6010 }, { "epoch": 0.4072769157802019, "grad_norm": 7.609717845916748, "learning_rate": 9.279895954548566e-05, "loss": 1.2036, "step": 6011 }, { "epoch": 0.407344671048174, "grad_norm": 8.978927612304688, "learning_rate": 9.279759052638784e-05, "loss": 0.8903, "step": 6012 }, { "epoch": 0.4074124263161461, "grad_norm": 8.331847190856934, "learning_rate": 9.279622150729003e-05, "loss": 1.0853, "step": 6013 }, { "epoch": 0.40748018158411814, "grad_norm": 6.639584541320801, "learning_rate": 9.279485248819221e-05, "loss": 0.9113, "step": 6014 }, { "epoch": 0.40754793685209023, "grad_norm": 6.87017822265625, "learning_rate": 9.27934834690944e-05, "loss": 0.748, "step": 6015 }, { "epoch": 0.4076156921200623, "grad_norm": 6.675489902496338, "learning_rate": 9.279211444999659e-05, "loss": 0.8002, "step": 6016 }, { "epoch": 0.4076834473880344, "grad_norm": 7.666563034057617, "learning_rate": 9.279074543089877e-05, "loss": 0.9973, "step": 6017 }, { "epoch": 0.4077512026560065, "grad_norm": 7.878670692443848, "learning_rate": 9.278937641180095e-05, "loss": 1.0647, "step": 6018 }, { "epoch": 0.4078189579239786, "grad_norm": 6.263443946838379, "learning_rate": 9.278800739270314e-05, "loss": 0.8827, "step": 6019 }, { "epoch": 0.4078867131919507, "grad_norm": 9.69717788696289, "learning_rate": 9.278663837360532e-05, "loss": 0.728, "step": 6020 }, { "epoch": 0.4079544684599228, "grad_norm": 8.236658096313477, "learning_rate": 9.27852693545075e-05, "loss": 0.78, "step": 6021 }, { "epoch": 0.4080222237278948, "grad_norm": 6.9468512535095215, "learning_rate": 9.278390033540968e-05, "loss": 0.7674, "step": 6022 }, { "epoch": 0.4080899789958669, "grad_norm": 8.783222198486328, "learning_rate": 9.278253131631188e-05, "loss": 0.9062, "step": 6023 }, { "epoch": 0.408157734263839, "grad_norm": 6.449438571929932, "learning_rate": 9.278116229721406e-05, "loss": 0.9683, "step": 6024 }, { "epoch": 0.4082254895318111, "grad_norm": 6.185464382171631, "learning_rate": 9.277979327811624e-05, "loss": 0.8899, "step": 6025 }, { "epoch": 0.4082932447997832, "grad_norm": 6.422550678253174, "learning_rate": 9.277842425901842e-05, "loss": 0.8765, "step": 6026 }, { "epoch": 0.4083610000677553, "grad_norm": 6.8761372566223145, "learning_rate": 9.27770552399206e-05, "loss": 0.7326, "step": 6027 }, { "epoch": 0.40842875533572737, "grad_norm": 5.4224162101745605, "learning_rate": 9.277568622082279e-05, "loss": 0.7882, "step": 6028 }, { "epoch": 0.40849651060369946, "grad_norm": 6.6288275718688965, "learning_rate": 9.277431720172497e-05, "loss": 0.936, "step": 6029 }, { "epoch": 0.4085642658716715, "grad_norm": 7.305326461791992, "learning_rate": 9.277294818262715e-05, "loss": 0.8254, "step": 6030 }, { "epoch": 0.4086320211396436, "grad_norm": 6.697827339172363, "learning_rate": 9.277157916352933e-05, "loss": 0.9365, "step": 6031 }, { "epoch": 0.4086997764076157, "grad_norm": 8.074564933776855, "learning_rate": 9.277021014443153e-05, "loss": 1.0454, "step": 6032 }, { "epoch": 0.40876753167558777, "grad_norm": 7.539030075073242, "learning_rate": 9.276884112533371e-05, "loss": 0.9275, "step": 6033 }, { "epoch": 0.40883528694355986, "grad_norm": 7.1338582038879395, "learning_rate": 9.276747210623589e-05, "loss": 0.8395, "step": 6034 }, { "epoch": 0.40890304221153195, "grad_norm": 8.182660102844238, "learning_rate": 9.276610308713807e-05, "loss": 1.1907, "step": 6035 }, { "epoch": 0.40897079747950404, "grad_norm": 6.1452813148498535, "learning_rate": 9.276473406804025e-05, "loss": 0.6271, "step": 6036 }, { "epoch": 0.40903855274747614, "grad_norm": 7.247335433959961, "learning_rate": 9.276336504894244e-05, "loss": 0.8665, "step": 6037 }, { "epoch": 0.40910630801544823, "grad_norm": 8.27696418762207, "learning_rate": 9.276199602984462e-05, "loss": 0.968, "step": 6038 }, { "epoch": 0.40917406328342026, "grad_norm": 6.888766288757324, "learning_rate": 9.27606270107468e-05, "loss": 0.8212, "step": 6039 }, { "epoch": 0.40924181855139236, "grad_norm": 6.885414123535156, "learning_rate": 9.275925799164898e-05, "loss": 0.8978, "step": 6040 }, { "epoch": 0.40930957381936445, "grad_norm": 7.307199478149414, "learning_rate": 9.275788897255118e-05, "loss": 0.9602, "step": 6041 }, { "epoch": 0.40937732908733654, "grad_norm": 6.921801567077637, "learning_rate": 9.275651995345336e-05, "loss": 0.863, "step": 6042 }, { "epoch": 0.40944508435530863, "grad_norm": 6.910282611846924, "learning_rate": 9.275515093435554e-05, "loss": 0.8481, "step": 6043 }, { "epoch": 0.4095128396232807, "grad_norm": 7.98552131652832, "learning_rate": 9.275378191525772e-05, "loss": 0.9096, "step": 6044 }, { "epoch": 0.4095805948912528, "grad_norm": 9.74826431274414, "learning_rate": 9.27524128961599e-05, "loss": 0.867, "step": 6045 }, { "epoch": 0.4096483501592249, "grad_norm": 10.083016395568848, "learning_rate": 9.275104387706209e-05, "loss": 0.9664, "step": 6046 }, { "epoch": 0.40971610542719694, "grad_norm": 8.351798057556152, "learning_rate": 9.274967485796427e-05, "loss": 0.8164, "step": 6047 }, { "epoch": 0.40978386069516903, "grad_norm": 7.7515106201171875, "learning_rate": 9.274830583886645e-05, "loss": 0.9784, "step": 6048 }, { "epoch": 0.4098516159631411, "grad_norm": 6.339774131774902, "learning_rate": 9.274693681976863e-05, "loss": 0.7614, "step": 6049 }, { "epoch": 0.4099193712311132, "grad_norm": 8.455992698669434, "learning_rate": 9.274556780067081e-05, "loss": 0.989, "step": 6050 }, { "epoch": 0.4099871264990853, "grad_norm": 9.367591857910156, "learning_rate": 9.274419878157301e-05, "loss": 1.0155, "step": 6051 }, { "epoch": 0.4100548817670574, "grad_norm": 8.611092567443848, "learning_rate": 9.274282976247519e-05, "loss": 0.9541, "step": 6052 }, { "epoch": 0.4101226370350295, "grad_norm": 8.239481925964355, "learning_rate": 9.274146074337737e-05, "loss": 0.9511, "step": 6053 }, { "epoch": 0.4101903923030016, "grad_norm": 7.31620979309082, "learning_rate": 9.274009172427955e-05, "loss": 1.0126, "step": 6054 }, { "epoch": 0.4102581475709736, "grad_norm": 6.814750671386719, "learning_rate": 9.273872270518174e-05, "loss": 0.9463, "step": 6055 }, { "epoch": 0.4103259028389457, "grad_norm": 7.000329494476318, "learning_rate": 9.273735368608392e-05, "loss": 1.0664, "step": 6056 }, { "epoch": 0.4103936581069178, "grad_norm": 6.309933662414551, "learning_rate": 9.27359846669861e-05, "loss": 0.8605, "step": 6057 }, { "epoch": 0.4104614133748899, "grad_norm": 8.13158893585205, "learning_rate": 9.273461564788828e-05, "loss": 0.9158, "step": 6058 }, { "epoch": 0.410529168642862, "grad_norm": 8.334741592407227, "learning_rate": 9.273324662879048e-05, "loss": 0.8699, "step": 6059 }, { "epoch": 0.4105969239108341, "grad_norm": 8.019463539123535, "learning_rate": 9.273187760969266e-05, "loss": 1.0366, "step": 6060 }, { "epoch": 0.41066467917880617, "grad_norm": 6.104535102844238, "learning_rate": 9.273050859059484e-05, "loss": 0.8273, "step": 6061 }, { "epoch": 0.41073243444677826, "grad_norm": 8.345130920410156, "learning_rate": 9.272913957149703e-05, "loss": 1.3653, "step": 6062 }, { "epoch": 0.4108001897147503, "grad_norm": 8.028311729431152, "learning_rate": 9.272777055239921e-05, "loss": 1.096, "step": 6063 }, { "epoch": 0.4108679449827224, "grad_norm": 7.079861164093018, "learning_rate": 9.272640153330139e-05, "loss": 1.0157, "step": 6064 }, { "epoch": 0.4109357002506945, "grad_norm": 8.497052192687988, "learning_rate": 9.272503251420359e-05, "loss": 0.9972, "step": 6065 }, { "epoch": 0.41100345551866657, "grad_norm": 8.042581558227539, "learning_rate": 9.272366349510577e-05, "loss": 0.908, "step": 6066 }, { "epoch": 0.41107121078663866, "grad_norm": 6.567187786102295, "learning_rate": 9.272229447600795e-05, "loss": 0.7992, "step": 6067 }, { "epoch": 0.41113896605461075, "grad_norm": 5.78397274017334, "learning_rate": 9.272092545691013e-05, "loss": 0.773, "step": 6068 }, { "epoch": 0.41120672132258285, "grad_norm": 8.116515159606934, "learning_rate": 9.271955643781232e-05, "loss": 0.8101, "step": 6069 }, { "epoch": 0.41127447659055494, "grad_norm": 7.628951549530029, "learning_rate": 9.27181874187145e-05, "loss": 0.9282, "step": 6070 }, { "epoch": 0.411342231858527, "grad_norm": 7.830180644989014, "learning_rate": 9.271681839961668e-05, "loss": 0.8681, "step": 6071 }, { "epoch": 0.41140998712649907, "grad_norm": 7.690285682678223, "learning_rate": 9.271544938051886e-05, "loss": 0.9064, "step": 6072 }, { "epoch": 0.41147774239447116, "grad_norm": 7.585133075714111, "learning_rate": 9.271408036142106e-05, "loss": 0.8305, "step": 6073 }, { "epoch": 0.41154549766244325, "grad_norm": 6.18320894241333, "learning_rate": 9.271271134232324e-05, "loss": 1.0395, "step": 6074 }, { "epoch": 0.41161325293041534, "grad_norm": 8.301353454589844, "learning_rate": 9.271134232322542e-05, "loss": 1.1528, "step": 6075 }, { "epoch": 0.41168100819838743, "grad_norm": 6.261874198913574, "learning_rate": 9.27099733041276e-05, "loss": 0.7802, "step": 6076 }, { "epoch": 0.4117487634663595, "grad_norm": 7.161525726318359, "learning_rate": 9.270860428502978e-05, "loss": 1.1292, "step": 6077 }, { "epoch": 0.4118165187343316, "grad_norm": 6.522838115692139, "learning_rate": 9.270723526593197e-05, "loss": 0.8013, "step": 6078 }, { "epoch": 0.41188427400230365, "grad_norm": 6.919328212738037, "learning_rate": 9.270586624683415e-05, "loss": 0.8172, "step": 6079 }, { "epoch": 0.41195202927027574, "grad_norm": 6.236283302307129, "learning_rate": 9.270449722773633e-05, "loss": 0.9677, "step": 6080 }, { "epoch": 0.41201978453824784, "grad_norm": 5.593216896057129, "learning_rate": 9.270312820863851e-05, "loss": 0.922, "step": 6081 }, { "epoch": 0.4120875398062199, "grad_norm": 8.651509284973145, "learning_rate": 9.27017591895407e-05, "loss": 0.9209, "step": 6082 }, { "epoch": 0.412155295074192, "grad_norm": 6.787774085998535, "learning_rate": 9.270039017044289e-05, "loss": 1.0991, "step": 6083 }, { "epoch": 0.4122230503421641, "grad_norm": 6.774304389953613, "learning_rate": 9.269902115134507e-05, "loss": 0.7265, "step": 6084 }, { "epoch": 0.4122908056101362, "grad_norm": 7.097721099853516, "learning_rate": 9.269765213224725e-05, "loss": 1.0353, "step": 6085 }, { "epoch": 0.4123585608781083, "grad_norm": 8.65166187286377, "learning_rate": 9.269628311314943e-05, "loss": 1.0157, "step": 6086 }, { "epoch": 0.4124263161460804, "grad_norm": 5.778721809387207, "learning_rate": 9.269491409405162e-05, "loss": 0.8381, "step": 6087 }, { "epoch": 0.4124940714140524, "grad_norm": 6.297547340393066, "learning_rate": 9.26935450749538e-05, "loss": 0.9421, "step": 6088 }, { "epoch": 0.4125618266820245, "grad_norm": 7.8641462326049805, "learning_rate": 9.269217605585598e-05, "loss": 0.9987, "step": 6089 }, { "epoch": 0.4126295819499966, "grad_norm": 7.222883701324463, "learning_rate": 9.269080703675816e-05, "loss": 0.7314, "step": 6090 }, { "epoch": 0.4126973372179687, "grad_norm": 7.291220188140869, "learning_rate": 9.268943801766034e-05, "loss": 0.9199, "step": 6091 }, { "epoch": 0.4127650924859408, "grad_norm": 7.00157356262207, "learning_rate": 9.268806899856254e-05, "loss": 0.8965, "step": 6092 }, { "epoch": 0.4128328477539129, "grad_norm": 8.229329109191895, "learning_rate": 9.268669997946472e-05, "loss": 1.0392, "step": 6093 }, { "epoch": 0.41290060302188497, "grad_norm": 6.9192633628845215, "learning_rate": 9.26853309603669e-05, "loss": 0.8987, "step": 6094 }, { "epoch": 0.41296835828985706, "grad_norm": 6.3700761795043945, "learning_rate": 9.268396194126908e-05, "loss": 0.7713, "step": 6095 }, { "epoch": 0.4130361135578291, "grad_norm": 5.654745101928711, "learning_rate": 9.268259292217127e-05, "loss": 0.6733, "step": 6096 }, { "epoch": 0.4131038688258012, "grad_norm": 6.898359298706055, "learning_rate": 9.268122390307345e-05, "loss": 0.9611, "step": 6097 }, { "epoch": 0.4131716240937733, "grad_norm": 6.53093147277832, "learning_rate": 9.267985488397563e-05, "loss": 0.795, "step": 6098 }, { "epoch": 0.4132393793617454, "grad_norm": 8.565315246582031, "learning_rate": 9.267848586487781e-05, "loss": 0.7333, "step": 6099 }, { "epoch": 0.41330713462971747, "grad_norm": 7.8006744384765625, "learning_rate": 9.267711684578e-05, "loss": 0.9579, "step": 6100 }, { "epoch": 0.41337488989768956, "grad_norm": 7.270709037780762, "learning_rate": 9.267574782668219e-05, "loss": 0.9105, "step": 6101 }, { "epoch": 0.41344264516566165, "grad_norm": 7.98935079574585, "learning_rate": 9.267437880758437e-05, "loss": 0.9697, "step": 6102 }, { "epoch": 0.41351040043363374, "grad_norm": 6.606309413909912, "learning_rate": 9.267300978848655e-05, "loss": 0.8649, "step": 6103 }, { "epoch": 0.4135781557016058, "grad_norm": 7.456545352935791, "learning_rate": 9.267164076938873e-05, "loss": 1.0884, "step": 6104 }, { "epoch": 0.41364591096957787, "grad_norm": 6.820968151092529, "learning_rate": 9.267027175029092e-05, "loss": 0.6651, "step": 6105 }, { "epoch": 0.41371366623754996, "grad_norm": 7.569576740264893, "learning_rate": 9.26689027311931e-05, "loss": 1.024, "step": 6106 }, { "epoch": 0.41378142150552205, "grad_norm": 7.982132911682129, "learning_rate": 9.266753371209528e-05, "loss": 0.9111, "step": 6107 }, { "epoch": 0.41384917677349414, "grad_norm": 7.610587120056152, "learning_rate": 9.266616469299748e-05, "loss": 0.9005, "step": 6108 }, { "epoch": 0.41391693204146623, "grad_norm": 6.773017883300781, "learning_rate": 9.266479567389966e-05, "loss": 0.6696, "step": 6109 }, { "epoch": 0.4139846873094383, "grad_norm": 7.702723979949951, "learning_rate": 9.266342665480184e-05, "loss": 0.7951, "step": 6110 }, { "epoch": 0.4140524425774104, "grad_norm": 7.299111843109131, "learning_rate": 9.266205763570403e-05, "loss": 1.0422, "step": 6111 }, { "epoch": 0.41412019784538245, "grad_norm": 5.148745536804199, "learning_rate": 9.266068861660621e-05, "loss": 0.6821, "step": 6112 }, { "epoch": 0.41418795311335455, "grad_norm": 6.768344402313232, "learning_rate": 9.265931959750839e-05, "loss": 0.8119, "step": 6113 }, { "epoch": 0.41425570838132664, "grad_norm": 6.882130146026611, "learning_rate": 9.265795057841057e-05, "loss": 0.9931, "step": 6114 }, { "epoch": 0.41432346364929873, "grad_norm": 7.292698383331299, "learning_rate": 9.265658155931277e-05, "loss": 1.0415, "step": 6115 }, { "epoch": 0.4143912189172708, "grad_norm": 7.241159915924072, "learning_rate": 9.265521254021495e-05, "loss": 0.9122, "step": 6116 }, { "epoch": 0.4144589741852429, "grad_norm": 6.7898736000061035, "learning_rate": 9.265384352111713e-05, "loss": 0.9399, "step": 6117 }, { "epoch": 0.414526729453215, "grad_norm": 6.505312919616699, "learning_rate": 9.265247450201931e-05, "loss": 0.9463, "step": 6118 }, { "epoch": 0.4145944847211871, "grad_norm": 5.761348724365234, "learning_rate": 9.26511054829215e-05, "loss": 0.8365, "step": 6119 }, { "epoch": 0.41466223998915913, "grad_norm": 6.377706527709961, "learning_rate": 9.264973646382368e-05, "loss": 0.7451, "step": 6120 }, { "epoch": 0.4147299952571312, "grad_norm": 8.465597152709961, "learning_rate": 9.264836744472586e-05, "loss": 0.8582, "step": 6121 }, { "epoch": 0.4147977505251033, "grad_norm": 6.3482770919799805, "learning_rate": 9.264699842562804e-05, "loss": 0.8291, "step": 6122 }, { "epoch": 0.4148655057930754, "grad_norm": 7.984723091125488, "learning_rate": 9.264562940653022e-05, "loss": 0.9641, "step": 6123 }, { "epoch": 0.4149332610610475, "grad_norm": 6.695097923278809, "learning_rate": 9.264426038743242e-05, "loss": 0.7661, "step": 6124 }, { "epoch": 0.4150010163290196, "grad_norm": 6.403726100921631, "learning_rate": 9.26428913683346e-05, "loss": 0.6342, "step": 6125 }, { "epoch": 0.4150687715969917, "grad_norm": 5.906423091888428, "learning_rate": 9.264152234923678e-05, "loss": 0.9588, "step": 6126 }, { "epoch": 0.4151365268649638, "grad_norm": 6.482539653778076, "learning_rate": 9.264015333013896e-05, "loss": 1.0244, "step": 6127 }, { "epoch": 0.4152042821329358, "grad_norm": 7.783926486968994, "learning_rate": 9.263878431104114e-05, "loss": 0.8611, "step": 6128 }, { "epoch": 0.4152720374009079, "grad_norm": 8.377721786499023, "learning_rate": 9.263741529194333e-05, "loss": 0.9757, "step": 6129 }, { "epoch": 0.41533979266888, "grad_norm": 8.719101905822754, "learning_rate": 9.263604627284551e-05, "loss": 1.0766, "step": 6130 }, { "epoch": 0.4154075479368521, "grad_norm": 6.0139641761779785, "learning_rate": 9.263467725374769e-05, "loss": 0.8036, "step": 6131 }, { "epoch": 0.4154753032048242, "grad_norm": 8.683677673339844, "learning_rate": 9.263330823464987e-05, "loss": 0.9476, "step": 6132 }, { "epoch": 0.41554305847279627, "grad_norm": 7.566380977630615, "learning_rate": 9.263193921555207e-05, "loss": 1.0983, "step": 6133 }, { "epoch": 0.41561081374076836, "grad_norm": 6.716690540313721, "learning_rate": 9.263057019645425e-05, "loss": 0.8803, "step": 6134 }, { "epoch": 0.41567856900874045, "grad_norm": 6.123441696166992, "learning_rate": 9.262920117735643e-05, "loss": 0.7158, "step": 6135 }, { "epoch": 0.4157463242767125, "grad_norm": 8.015522956848145, "learning_rate": 9.262783215825861e-05, "loss": 1.0269, "step": 6136 }, { "epoch": 0.4158140795446846, "grad_norm": 5.564431667327881, "learning_rate": 9.262646313916079e-05, "loss": 0.8309, "step": 6137 }, { "epoch": 0.41588183481265667, "grad_norm": 7.946048736572266, "learning_rate": 9.262509412006298e-05, "loss": 0.9334, "step": 6138 }, { "epoch": 0.41594959008062876, "grad_norm": 6.741854667663574, "learning_rate": 9.262372510096516e-05, "loss": 0.6461, "step": 6139 }, { "epoch": 0.41601734534860085, "grad_norm": 8.229652404785156, "learning_rate": 9.262235608186734e-05, "loss": 0.9046, "step": 6140 }, { "epoch": 0.41608510061657294, "grad_norm": 5.590304851531982, "learning_rate": 9.262098706276952e-05, "loss": 0.6979, "step": 6141 }, { "epoch": 0.41615285588454504, "grad_norm": 6.79884672164917, "learning_rate": 9.261961804367172e-05, "loss": 0.853, "step": 6142 }, { "epoch": 0.41622061115251713, "grad_norm": 7.175544738769531, "learning_rate": 9.26182490245739e-05, "loss": 0.9265, "step": 6143 }, { "epoch": 0.4162883664204892, "grad_norm": 8.839093208312988, "learning_rate": 9.261688000547608e-05, "loss": 0.9541, "step": 6144 }, { "epoch": 0.41635612168846126, "grad_norm": 6.818619728088379, "learning_rate": 9.261551098637826e-05, "loss": 0.6284, "step": 6145 }, { "epoch": 0.41642387695643335, "grad_norm": 7.31305456161499, "learning_rate": 9.261414196728044e-05, "loss": 0.797, "step": 6146 }, { "epoch": 0.41649163222440544, "grad_norm": 8.566871643066406, "learning_rate": 9.261277294818263e-05, "loss": 1.0893, "step": 6147 }, { "epoch": 0.41655938749237753, "grad_norm": 6.028139114379883, "learning_rate": 9.261140392908481e-05, "loss": 0.781, "step": 6148 }, { "epoch": 0.4166271427603496, "grad_norm": 7.383317947387695, "learning_rate": 9.261003490998699e-05, "loss": 0.5846, "step": 6149 }, { "epoch": 0.4166948980283217, "grad_norm": 6.392228126525879, "learning_rate": 9.260866589088917e-05, "loss": 0.9403, "step": 6150 }, { "epoch": 0.4167626532962938, "grad_norm": 6.1830644607543945, "learning_rate": 9.260729687179137e-05, "loss": 0.9502, "step": 6151 }, { "epoch": 0.4168304085642659, "grad_norm": 6.851447582244873, "learning_rate": 9.260592785269355e-05, "loss": 0.8042, "step": 6152 }, { "epoch": 0.41689816383223793, "grad_norm": 5.995123386383057, "learning_rate": 9.260455883359573e-05, "loss": 0.8875, "step": 6153 }, { "epoch": 0.41696591910021, "grad_norm": 6.834797382354736, "learning_rate": 9.260318981449791e-05, "loss": 0.8652, "step": 6154 }, { "epoch": 0.4170336743681821, "grad_norm": 6.0200371742248535, "learning_rate": 9.26018207954001e-05, "loss": 0.9198, "step": 6155 }, { "epoch": 0.4171014296361542, "grad_norm": 7.138192176818848, "learning_rate": 9.260045177630228e-05, "loss": 0.8284, "step": 6156 }, { "epoch": 0.4171691849041263, "grad_norm": 7.5531229972839355, "learning_rate": 9.259908275720446e-05, "loss": 0.7188, "step": 6157 }, { "epoch": 0.4172369401720984, "grad_norm": 8.347415924072266, "learning_rate": 9.259771373810666e-05, "loss": 1.1853, "step": 6158 }, { "epoch": 0.4173046954400705, "grad_norm": 8.188237190246582, "learning_rate": 9.259634471900884e-05, "loss": 0.7712, "step": 6159 }, { "epoch": 0.4173724507080426, "grad_norm": 7.238736629486084, "learning_rate": 9.259497569991102e-05, "loss": 0.7731, "step": 6160 }, { "epoch": 0.4174402059760146, "grad_norm": 8.175471305847168, "learning_rate": 9.259360668081321e-05, "loss": 1.0769, "step": 6161 }, { "epoch": 0.4175079612439867, "grad_norm": 6.97186803817749, "learning_rate": 9.259223766171539e-05, "loss": 0.7469, "step": 6162 }, { "epoch": 0.4175757165119588, "grad_norm": 5.761664390563965, "learning_rate": 9.259086864261757e-05, "loss": 0.8716, "step": 6163 }, { "epoch": 0.4176434717799309, "grad_norm": 8.567249298095703, "learning_rate": 9.258949962351975e-05, "loss": 0.8376, "step": 6164 }, { "epoch": 0.417711227047903, "grad_norm": 5.827561378479004, "learning_rate": 9.258813060442195e-05, "loss": 0.8468, "step": 6165 }, { "epoch": 0.41777898231587507, "grad_norm": 7.336645603179932, "learning_rate": 9.258676158532413e-05, "loss": 0.8529, "step": 6166 }, { "epoch": 0.41784673758384716, "grad_norm": 7.711108207702637, "learning_rate": 9.258539256622631e-05, "loss": 0.8802, "step": 6167 }, { "epoch": 0.41791449285181925, "grad_norm": 8.625036239624023, "learning_rate": 9.258402354712849e-05, "loss": 1.0263, "step": 6168 }, { "epoch": 0.4179822481197913, "grad_norm": 7.005527019500732, "learning_rate": 9.258265452803067e-05, "loss": 0.9728, "step": 6169 }, { "epoch": 0.4180500033877634, "grad_norm": 6.906123161315918, "learning_rate": 9.258128550893286e-05, "loss": 0.8958, "step": 6170 }, { "epoch": 0.41811775865573547, "grad_norm": 7.147536754608154, "learning_rate": 9.257991648983504e-05, "loss": 1.1466, "step": 6171 }, { "epoch": 0.41818551392370756, "grad_norm": 5.8256001472473145, "learning_rate": 9.257854747073722e-05, "loss": 0.7189, "step": 6172 }, { "epoch": 0.41825326919167966, "grad_norm": 5.144516944885254, "learning_rate": 9.25771784516394e-05, "loss": 0.6368, "step": 6173 }, { "epoch": 0.41832102445965175, "grad_norm": 7.161872863769531, "learning_rate": 9.25758094325416e-05, "loss": 1.0074, "step": 6174 }, { "epoch": 0.41838877972762384, "grad_norm": 9.098782539367676, "learning_rate": 9.257444041344378e-05, "loss": 0.9261, "step": 6175 }, { "epoch": 0.41845653499559593, "grad_norm": 7.566091060638428, "learning_rate": 9.257307139434596e-05, "loss": 0.8845, "step": 6176 }, { "epoch": 0.41852429026356797, "grad_norm": 8.618456840515137, "learning_rate": 9.257170237524814e-05, "loss": 0.8643, "step": 6177 }, { "epoch": 0.41859204553154006, "grad_norm": 8.870187759399414, "learning_rate": 9.257033335615032e-05, "loss": 1.1915, "step": 6178 }, { "epoch": 0.41865980079951215, "grad_norm": 6.494687557220459, "learning_rate": 9.256896433705251e-05, "loss": 0.6523, "step": 6179 }, { "epoch": 0.41872755606748424, "grad_norm": 7.417816638946533, "learning_rate": 9.256759531795469e-05, "loss": 0.9454, "step": 6180 }, { "epoch": 0.41879531133545633, "grad_norm": 11.310504913330078, "learning_rate": 9.256622629885687e-05, "loss": 1.2984, "step": 6181 }, { "epoch": 0.4188630666034284, "grad_norm": 6.3585615158081055, "learning_rate": 9.256485727975905e-05, "loss": 0.8379, "step": 6182 }, { "epoch": 0.4189308218714005, "grad_norm": 8.134123802185059, "learning_rate": 9.256348826066123e-05, "loss": 0.9962, "step": 6183 }, { "epoch": 0.4189985771393726, "grad_norm": 5.962007999420166, "learning_rate": 9.256211924156343e-05, "loss": 0.8923, "step": 6184 }, { "epoch": 0.41906633240734464, "grad_norm": 6.423482894897461, "learning_rate": 9.256075022246561e-05, "loss": 0.7991, "step": 6185 }, { "epoch": 0.41913408767531674, "grad_norm": 6.406213283538818, "learning_rate": 9.255938120336779e-05, "loss": 0.9454, "step": 6186 }, { "epoch": 0.4192018429432888, "grad_norm": 4.926299571990967, "learning_rate": 9.255801218426997e-05, "loss": 0.6442, "step": 6187 }, { "epoch": 0.4192695982112609, "grad_norm": 6.160318374633789, "learning_rate": 9.255664316517216e-05, "loss": 0.8575, "step": 6188 }, { "epoch": 0.419337353479233, "grad_norm": 9.17479133605957, "learning_rate": 9.255527414607434e-05, "loss": 1.2344, "step": 6189 }, { "epoch": 0.4194051087472051, "grad_norm": 8.95871639251709, "learning_rate": 9.255390512697652e-05, "loss": 0.9193, "step": 6190 }, { "epoch": 0.4194728640151772, "grad_norm": 5.636984825134277, "learning_rate": 9.25525361078787e-05, "loss": 0.725, "step": 6191 }, { "epoch": 0.4195406192831493, "grad_norm": 6.044233798980713, "learning_rate": 9.255116708878088e-05, "loss": 0.7526, "step": 6192 }, { "epoch": 0.4196083745511214, "grad_norm": 6.230378150939941, "learning_rate": 9.254979806968308e-05, "loss": 0.7923, "step": 6193 }, { "epoch": 0.4196761298190934, "grad_norm": 8.391528129577637, "learning_rate": 9.254842905058526e-05, "loss": 1.0368, "step": 6194 }, { "epoch": 0.4197438850870655, "grad_norm": 7.150004863739014, "learning_rate": 9.254706003148744e-05, "loss": 0.9087, "step": 6195 }, { "epoch": 0.4198116403550376, "grad_norm": 8.44096565246582, "learning_rate": 9.254569101238962e-05, "loss": 0.9511, "step": 6196 }, { "epoch": 0.4198793956230097, "grad_norm": 7.8519768714904785, "learning_rate": 9.254432199329181e-05, "loss": 1.0565, "step": 6197 }, { "epoch": 0.4199471508909818, "grad_norm": 6.873769760131836, "learning_rate": 9.254295297419399e-05, "loss": 0.9016, "step": 6198 }, { "epoch": 0.42001490615895387, "grad_norm": 6.1226396560668945, "learning_rate": 9.254158395509617e-05, "loss": 0.8439, "step": 6199 }, { "epoch": 0.42008266142692596, "grad_norm": 6.68569803237915, "learning_rate": 9.254021493599835e-05, "loss": 0.9836, "step": 6200 }, { "epoch": 0.42015041669489805, "grad_norm": 5.907567501068115, "learning_rate": 9.253884591690055e-05, "loss": 0.8047, "step": 6201 }, { "epoch": 0.4202181719628701, "grad_norm": 5.849989891052246, "learning_rate": 9.253747689780273e-05, "loss": 0.8791, "step": 6202 }, { "epoch": 0.4202859272308422, "grad_norm": 5.872654914855957, "learning_rate": 9.253610787870491e-05, "loss": 0.8368, "step": 6203 }, { "epoch": 0.4203536824988143, "grad_norm": 5.799046039581299, "learning_rate": 9.25347388596071e-05, "loss": 0.8215, "step": 6204 }, { "epoch": 0.42042143776678637, "grad_norm": 5.331169605255127, "learning_rate": 9.253336984050928e-05, "loss": 0.9533, "step": 6205 }, { "epoch": 0.42048919303475846, "grad_norm": 6.436175346374512, "learning_rate": 9.253200082141146e-05, "loss": 0.7356, "step": 6206 }, { "epoch": 0.42055694830273055, "grad_norm": 6.290867328643799, "learning_rate": 9.253063180231366e-05, "loss": 0.7771, "step": 6207 }, { "epoch": 0.42062470357070264, "grad_norm": 8.476110458374023, "learning_rate": 9.252926278321584e-05, "loss": 1.0357, "step": 6208 }, { "epoch": 0.42069245883867473, "grad_norm": 8.561487197875977, "learning_rate": 9.252789376411802e-05, "loss": 0.974, "step": 6209 }, { "epoch": 0.42076021410664677, "grad_norm": 7.24920654296875, "learning_rate": 9.25265247450202e-05, "loss": 1.2393, "step": 6210 }, { "epoch": 0.42082796937461886, "grad_norm": 7.135931015014648, "learning_rate": 9.252515572592239e-05, "loss": 1.0491, "step": 6211 }, { "epoch": 0.42089572464259095, "grad_norm": 8.867389678955078, "learning_rate": 9.252378670682457e-05, "loss": 0.836, "step": 6212 }, { "epoch": 0.42096347991056304, "grad_norm": 6.738379955291748, "learning_rate": 9.252241768772675e-05, "loss": 0.7829, "step": 6213 }, { "epoch": 0.42103123517853513, "grad_norm": 7.226996898651123, "learning_rate": 9.252104866862893e-05, "loss": 0.8716, "step": 6214 }, { "epoch": 0.4210989904465072, "grad_norm": 6.236578941345215, "learning_rate": 9.251967964953111e-05, "loss": 0.62, "step": 6215 }, { "epoch": 0.4211667457144793, "grad_norm": 7.801733016967773, "learning_rate": 9.25183106304333e-05, "loss": 1.1107, "step": 6216 }, { "epoch": 0.4212345009824514, "grad_norm": 6.062146186828613, "learning_rate": 9.251694161133549e-05, "loss": 0.8655, "step": 6217 }, { "epoch": 0.42130225625042345, "grad_norm": 5.408603668212891, "learning_rate": 9.251557259223767e-05, "loss": 0.7054, "step": 6218 }, { "epoch": 0.42137001151839554, "grad_norm": 9.531839370727539, "learning_rate": 9.251420357313985e-05, "loss": 0.9609, "step": 6219 }, { "epoch": 0.42143776678636763, "grad_norm": 6.054145336151123, "learning_rate": 9.251283455404204e-05, "loss": 0.9311, "step": 6220 }, { "epoch": 0.4215055220543397, "grad_norm": 6.372133731842041, "learning_rate": 9.251146553494422e-05, "loss": 0.8797, "step": 6221 }, { "epoch": 0.4215732773223118, "grad_norm": 6.394374847412109, "learning_rate": 9.25100965158464e-05, "loss": 0.7603, "step": 6222 }, { "epoch": 0.4216410325902839, "grad_norm": 7.188971519470215, "learning_rate": 9.250872749674858e-05, "loss": 0.5915, "step": 6223 }, { "epoch": 0.421708787858256, "grad_norm": 7.44260311126709, "learning_rate": 9.250735847765076e-05, "loss": 0.8215, "step": 6224 }, { "epoch": 0.4217765431262281, "grad_norm": 7.43394660949707, "learning_rate": 9.250598945855296e-05, "loss": 0.891, "step": 6225 }, { "epoch": 0.4218442983942001, "grad_norm": 6.9161200523376465, "learning_rate": 9.250462043945514e-05, "loss": 0.8795, "step": 6226 }, { "epoch": 0.4219120536621722, "grad_norm": 6.955477714538574, "learning_rate": 9.250325142035732e-05, "loss": 0.9781, "step": 6227 }, { "epoch": 0.4219798089301443, "grad_norm": 7.632521629333496, "learning_rate": 9.25018824012595e-05, "loss": 0.7374, "step": 6228 }, { "epoch": 0.4220475641981164, "grad_norm": 7.344869613647461, "learning_rate": 9.250051338216169e-05, "loss": 0.9506, "step": 6229 }, { "epoch": 0.4221153194660885, "grad_norm": 7.199647426605225, "learning_rate": 9.249914436306387e-05, "loss": 0.9225, "step": 6230 }, { "epoch": 0.4221830747340606, "grad_norm": 5.884180068969727, "learning_rate": 9.249777534396605e-05, "loss": 0.729, "step": 6231 }, { "epoch": 0.4222508300020327, "grad_norm": 6.82388973236084, "learning_rate": 9.249640632486823e-05, "loss": 1.0978, "step": 6232 }, { "epoch": 0.42231858527000476, "grad_norm": 8.883796691894531, "learning_rate": 9.249503730577041e-05, "loss": 1.1025, "step": 6233 }, { "epoch": 0.4223863405379768, "grad_norm": 6.1692938804626465, "learning_rate": 9.24936682866726e-05, "loss": 0.6734, "step": 6234 }, { "epoch": 0.4224540958059489, "grad_norm": 8.827327728271484, "learning_rate": 9.249229926757479e-05, "loss": 1.1097, "step": 6235 }, { "epoch": 0.422521851073921, "grad_norm": 6.649320602416992, "learning_rate": 9.249093024847697e-05, "loss": 0.9978, "step": 6236 }, { "epoch": 0.4225896063418931, "grad_norm": 8.04487419128418, "learning_rate": 9.248956122937915e-05, "loss": 0.9293, "step": 6237 }, { "epoch": 0.42265736160986517, "grad_norm": 7.031024932861328, "learning_rate": 9.248819221028133e-05, "loss": 0.7474, "step": 6238 }, { "epoch": 0.42272511687783726, "grad_norm": 4.858736038208008, "learning_rate": 9.248682319118352e-05, "loss": 0.6429, "step": 6239 }, { "epoch": 0.42279287214580935, "grad_norm": 6.075960159301758, "learning_rate": 9.24854541720857e-05, "loss": 0.7595, "step": 6240 }, { "epoch": 0.42286062741378144, "grad_norm": 11.653654098510742, "learning_rate": 9.248408515298788e-05, "loss": 0.7962, "step": 6241 }, { "epoch": 0.4229283826817535, "grad_norm": 6.255251884460449, "learning_rate": 9.248271613389006e-05, "loss": 0.9027, "step": 6242 }, { "epoch": 0.42299613794972557, "grad_norm": 6.908040523529053, "learning_rate": 9.248134711479226e-05, "loss": 0.936, "step": 6243 }, { "epoch": 0.42306389321769766, "grad_norm": 7.853124141693115, "learning_rate": 9.247997809569444e-05, "loss": 1.0966, "step": 6244 }, { "epoch": 0.42313164848566975, "grad_norm": 7.661264896392822, "learning_rate": 9.247860907659662e-05, "loss": 0.8407, "step": 6245 }, { "epoch": 0.42319940375364185, "grad_norm": 6.52161169052124, "learning_rate": 9.24772400574988e-05, "loss": 0.862, "step": 6246 }, { "epoch": 0.42326715902161394, "grad_norm": 6.456472873687744, "learning_rate": 9.247587103840099e-05, "loss": 0.713, "step": 6247 }, { "epoch": 0.42333491428958603, "grad_norm": 6.655475616455078, "learning_rate": 9.247450201930317e-05, "loss": 0.7422, "step": 6248 }, { "epoch": 0.4234026695575581, "grad_norm": 6.448397636413574, "learning_rate": 9.247313300020535e-05, "loss": 0.8645, "step": 6249 }, { "epoch": 0.4234704248255302, "grad_norm": 5.964814186096191, "learning_rate": 9.247176398110755e-05, "loss": 0.6173, "step": 6250 }, { "epoch": 0.42353818009350225, "grad_norm": 6.125659465789795, "learning_rate": 9.247039496200973e-05, "loss": 0.6951, "step": 6251 }, { "epoch": 0.42360593536147434, "grad_norm": 6.881291389465332, "learning_rate": 9.246902594291191e-05, "loss": 0.7525, "step": 6252 }, { "epoch": 0.42367369062944643, "grad_norm": 6.647464752197266, "learning_rate": 9.24676569238141e-05, "loss": 0.6757, "step": 6253 }, { "epoch": 0.4237414458974185, "grad_norm": 6.530306816101074, "learning_rate": 9.246628790471628e-05, "loss": 1.1478, "step": 6254 }, { "epoch": 0.4238092011653906, "grad_norm": 7.5514984130859375, "learning_rate": 9.246491888561846e-05, "loss": 0.868, "step": 6255 }, { "epoch": 0.4238769564333627, "grad_norm": 10.405878067016602, "learning_rate": 9.246354986652064e-05, "loss": 1.0607, "step": 6256 }, { "epoch": 0.4239447117013348, "grad_norm": 6.7608418464660645, "learning_rate": 9.246218084742284e-05, "loss": 0.9173, "step": 6257 }, { "epoch": 0.4240124669693069, "grad_norm": 7.583088397979736, "learning_rate": 9.246081182832502e-05, "loss": 0.9091, "step": 6258 }, { "epoch": 0.4240802222372789, "grad_norm": 7.009425163269043, "learning_rate": 9.24594428092272e-05, "loss": 0.9509, "step": 6259 }, { "epoch": 0.424147977505251, "grad_norm": 10.783225059509277, "learning_rate": 9.245807379012938e-05, "loss": 0.9033, "step": 6260 }, { "epoch": 0.4242157327732231, "grad_norm": 6.1244611740112305, "learning_rate": 9.245670477103156e-05, "loss": 0.7285, "step": 6261 }, { "epoch": 0.4242834880411952, "grad_norm": 6.983814716339111, "learning_rate": 9.245533575193375e-05, "loss": 0.9463, "step": 6262 }, { "epoch": 0.4243512433091673, "grad_norm": 6.019556522369385, "learning_rate": 9.245396673283593e-05, "loss": 0.8072, "step": 6263 }, { "epoch": 0.4244189985771394, "grad_norm": 7.2982001304626465, "learning_rate": 9.245259771373811e-05, "loss": 0.8905, "step": 6264 }, { "epoch": 0.4244867538451115, "grad_norm": 7.231256008148193, "learning_rate": 9.245122869464029e-05, "loss": 0.9114, "step": 6265 }, { "epoch": 0.42455450911308357, "grad_norm": 6.597518444061279, "learning_rate": 9.244985967554249e-05, "loss": 1.0431, "step": 6266 }, { "epoch": 0.4246222643810556, "grad_norm": 8.651244163513184, "learning_rate": 9.244849065644467e-05, "loss": 1.1346, "step": 6267 }, { "epoch": 0.4246900196490277, "grad_norm": 9.052881240844727, "learning_rate": 9.244712163734685e-05, "loss": 0.7779, "step": 6268 }, { "epoch": 0.4247577749169998, "grad_norm": 6.886295318603516, "learning_rate": 9.244575261824903e-05, "loss": 0.9643, "step": 6269 }, { "epoch": 0.4248255301849719, "grad_norm": 6.241147518157959, "learning_rate": 9.244438359915121e-05, "loss": 0.7539, "step": 6270 }, { "epoch": 0.42489328545294397, "grad_norm": 7.910638332366943, "learning_rate": 9.24430145800534e-05, "loss": 0.9967, "step": 6271 }, { "epoch": 0.42496104072091606, "grad_norm": 7.068695545196533, "learning_rate": 9.244164556095558e-05, "loss": 0.9531, "step": 6272 }, { "epoch": 0.42502879598888815, "grad_norm": 7.7406907081604, "learning_rate": 9.244027654185776e-05, "loss": 0.8593, "step": 6273 }, { "epoch": 0.42509655125686024, "grad_norm": 6.548360824584961, "learning_rate": 9.243890752275994e-05, "loss": 0.7806, "step": 6274 }, { "epoch": 0.4251643065248323, "grad_norm": 6.20359468460083, "learning_rate": 9.243753850366214e-05, "loss": 0.846, "step": 6275 }, { "epoch": 0.42523206179280437, "grad_norm": 6.6235270500183105, "learning_rate": 9.243616948456432e-05, "loss": 0.806, "step": 6276 }, { "epoch": 0.42529981706077646, "grad_norm": 7.216398239135742, "learning_rate": 9.24348004654665e-05, "loss": 1.0497, "step": 6277 }, { "epoch": 0.42536757232874856, "grad_norm": 6.946768283843994, "learning_rate": 9.243343144636868e-05, "loss": 0.7862, "step": 6278 }, { "epoch": 0.42543532759672065, "grad_norm": 6.441595554351807, "learning_rate": 9.243206242727086e-05, "loss": 0.8663, "step": 6279 }, { "epoch": 0.42550308286469274, "grad_norm": 8.521356582641602, "learning_rate": 9.243069340817305e-05, "loss": 1.3689, "step": 6280 }, { "epoch": 0.42557083813266483, "grad_norm": 7.311733722686768, "learning_rate": 9.242932438907523e-05, "loss": 0.8788, "step": 6281 }, { "epoch": 0.4256385934006369, "grad_norm": 7.157277584075928, "learning_rate": 9.242795536997741e-05, "loss": 0.8245, "step": 6282 }, { "epoch": 0.42570634866860896, "grad_norm": 5.332032680511475, "learning_rate": 9.242658635087959e-05, "loss": 0.7333, "step": 6283 }, { "epoch": 0.42577410393658105, "grad_norm": 6.090252876281738, "learning_rate": 9.242521733178179e-05, "loss": 0.7084, "step": 6284 }, { "epoch": 0.42584185920455314, "grad_norm": 6.971512317657471, "learning_rate": 9.242384831268397e-05, "loss": 0.7226, "step": 6285 }, { "epoch": 0.42590961447252523, "grad_norm": 5.37277364730835, "learning_rate": 9.242247929358615e-05, "loss": 0.701, "step": 6286 }, { "epoch": 0.4259773697404973, "grad_norm": 6.7667622566223145, "learning_rate": 9.242111027448833e-05, "loss": 0.9389, "step": 6287 }, { "epoch": 0.4260451250084694, "grad_norm": 6.8329596519470215, "learning_rate": 9.241974125539051e-05, "loss": 0.8574, "step": 6288 }, { "epoch": 0.4261128802764415, "grad_norm": 6.124345779418945, "learning_rate": 9.24183722362927e-05, "loss": 0.9497, "step": 6289 }, { "epoch": 0.4261806355444136, "grad_norm": 6.190674304962158, "learning_rate": 9.241700321719488e-05, "loss": 0.8933, "step": 6290 }, { "epoch": 0.42624839081238564, "grad_norm": 6.881906986236572, "learning_rate": 9.241563419809706e-05, "loss": 1.0343, "step": 6291 }, { "epoch": 0.4263161460803577, "grad_norm": 6.8810648918151855, "learning_rate": 9.241426517899924e-05, "loss": 1.1141, "step": 6292 }, { "epoch": 0.4263839013483298, "grad_norm": 7.524305820465088, "learning_rate": 9.241289615990144e-05, "loss": 0.9136, "step": 6293 }, { "epoch": 0.4264516566163019, "grad_norm": 5.732180595397949, "learning_rate": 9.241152714080362e-05, "loss": 0.8808, "step": 6294 }, { "epoch": 0.426519411884274, "grad_norm": 5.485534191131592, "learning_rate": 9.24101581217058e-05, "loss": 0.8755, "step": 6295 }, { "epoch": 0.4265871671522461, "grad_norm": 7.75483512878418, "learning_rate": 9.240878910260799e-05, "loss": 0.6602, "step": 6296 }, { "epoch": 0.4266549224202182, "grad_norm": 5.6334075927734375, "learning_rate": 9.240742008351017e-05, "loss": 0.6075, "step": 6297 }, { "epoch": 0.4267226776881903, "grad_norm": 6.316882133483887, "learning_rate": 9.240605106441235e-05, "loss": 0.8607, "step": 6298 }, { "epoch": 0.42679043295616237, "grad_norm": 7.0021257400512695, "learning_rate": 9.240468204531455e-05, "loss": 0.8648, "step": 6299 }, { "epoch": 0.4268581882241344, "grad_norm": 5.8421125411987305, "learning_rate": 9.240331302621673e-05, "loss": 0.7887, "step": 6300 }, { "epoch": 0.4269259434921065, "grad_norm": 8.877615928649902, "learning_rate": 9.24019440071189e-05, "loss": 1.3777, "step": 6301 }, { "epoch": 0.4269936987600786, "grad_norm": 7.440095901489258, "learning_rate": 9.240057498802109e-05, "loss": 1.0015, "step": 6302 }, { "epoch": 0.4270614540280507, "grad_norm": 6.460360527038574, "learning_rate": 9.239920596892328e-05, "loss": 0.6861, "step": 6303 }, { "epoch": 0.42712920929602277, "grad_norm": 9.731608390808105, "learning_rate": 9.239783694982546e-05, "loss": 0.8601, "step": 6304 }, { "epoch": 0.42719696456399486, "grad_norm": 7.891725540161133, "learning_rate": 9.239646793072764e-05, "loss": 0.7478, "step": 6305 }, { "epoch": 0.42726471983196695, "grad_norm": 6.742074966430664, "learning_rate": 9.239509891162982e-05, "loss": 0.7774, "step": 6306 }, { "epoch": 0.42733247509993905, "grad_norm": 7.515460014343262, "learning_rate": 9.239372989253202e-05, "loss": 0.9524, "step": 6307 }, { "epoch": 0.4274002303679111, "grad_norm": 6.278934478759766, "learning_rate": 9.23923608734342e-05, "loss": 0.9216, "step": 6308 }, { "epoch": 0.4274679856358832, "grad_norm": 6.528146266937256, "learning_rate": 9.239099185433638e-05, "loss": 0.9959, "step": 6309 }, { "epoch": 0.42753574090385527, "grad_norm": 5.953476905822754, "learning_rate": 9.238962283523856e-05, "loss": 1.0062, "step": 6310 }, { "epoch": 0.42760349617182736, "grad_norm": 7.488530158996582, "learning_rate": 9.238825381614074e-05, "loss": 0.8633, "step": 6311 }, { "epoch": 0.42767125143979945, "grad_norm": 6.974276542663574, "learning_rate": 9.238688479704293e-05, "loss": 1.0849, "step": 6312 }, { "epoch": 0.42773900670777154, "grad_norm": 7.560183048248291, "learning_rate": 9.238551577794511e-05, "loss": 0.9238, "step": 6313 }, { "epoch": 0.42780676197574363, "grad_norm": 6.859951019287109, "learning_rate": 9.238414675884729e-05, "loss": 0.7021, "step": 6314 }, { "epoch": 0.4278745172437157, "grad_norm": 8.095294952392578, "learning_rate": 9.238277773974947e-05, "loss": 0.9374, "step": 6315 }, { "epoch": 0.42794227251168776, "grad_norm": 6.5458221435546875, "learning_rate": 9.238140872065165e-05, "loss": 0.7089, "step": 6316 }, { "epoch": 0.42801002777965985, "grad_norm": 6.9012274742126465, "learning_rate": 9.238003970155385e-05, "loss": 1.0967, "step": 6317 }, { "epoch": 0.42807778304763194, "grad_norm": 7.991293430328369, "learning_rate": 9.237867068245603e-05, "loss": 0.8149, "step": 6318 }, { "epoch": 0.42814553831560404, "grad_norm": 6.636972904205322, "learning_rate": 9.23773016633582e-05, "loss": 0.9311, "step": 6319 }, { "epoch": 0.4282132935835761, "grad_norm": 7.938980579376221, "learning_rate": 9.237593264426039e-05, "loss": 1.021, "step": 6320 }, { "epoch": 0.4282810488515482, "grad_norm": 6.600100994110107, "learning_rate": 9.237456362516258e-05, "loss": 0.9375, "step": 6321 }, { "epoch": 0.4283488041195203, "grad_norm": 8.848981857299805, "learning_rate": 9.237319460606476e-05, "loss": 0.7755, "step": 6322 }, { "epoch": 0.4284165593874924, "grad_norm": 8.488492012023926, "learning_rate": 9.237182558696694e-05, "loss": 0.8267, "step": 6323 }, { "epoch": 0.42848431465546444, "grad_norm": 5.880466938018799, "learning_rate": 9.237045656786912e-05, "loss": 0.7597, "step": 6324 }, { "epoch": 0.42855206992343653, "grad_norm": 6.318797588348389, "learning_rate": 9.23690875487713e-05, "loss": 1.0001, "step": 6325 }, { "epoch": 0.4286198251914086, "grad_norm": 6.730221271514893, "learning_rate": 9.23677185296735e-05, "loss": 0.9811, "step": 6326 }, { "epoch": 0.4286875804593807, "grad_norm": 6.913969993591309, "learning_rate": 9.236634951057568e-05, "loss": 0.8782, "step": 6327 }, { "epoch": 0.4287553357273528, "grad_norm": 7.533535480499268, "learning_rate": 9.236498049147786e-05, "loss": 0.652, "step": 6328 }, { "epoch": 0.4288230909953249, "grad_norm": 7.687053680419922, "learning_rate": 9.236361147238004e-05, "loss": 0.9665, "step": 6329 }, { "epoch": 0.428890846263297, "grad_norm": 7.747707843780518, "learning_rate": 9.236224245328223e-05, "loss": 1.1481, "step": 6330 }, { "epoch": 0.4289586015312691, "grad_norm": 7.709572792053223, "learning_rate": 9.236087343418441e-05, "loss": 0.9191, "step": 6331 }, { "epoch": 0.4290263567992411, "grad_norm": 8.473289489746094, "learning_rate": 9.235950441508659e-05, "loss": 1.0646, "step": 6332 }, { "epoch": 0.4290941120672132, "grad_norm": 6.672016620635986, "learning_rate": 9.235813539598877e-05, "loss": 0.7865, "step": 6333 }, { "epoch": 0.4291618673351853, "grad_norm": 6.022176265716553, "learning_rate": 9.235676637689095e-05, "loss": 0.8593, "step": 6334 }, { "epoch": 0.4292296226031574, "grad_norm": 6.375059604644775, "learning_rate": 9.235539735779315e-05, "loss": 0.8754, "step": 6335 }, { "epoch": 0.4292973778711295, "grad_norm": 6.617783546447754, "learning_rate": 9.235402833869533e-05, "loss": 1.022, "step": 6336 }, { "epoch": 0.4293651331391016, "grad_norm": 5.790512561798096, "learning_rate": 9.235265931959751e-05, "loss": 0.8903, "step": 6337 }, { "epoch": 0.42943288840707367, "grad_norm": 6.528816223144531, "learning_rate": 9.235129030049969e-05, "loss": 0.8999, "step": 6338 }, { "epoch": 0.42950064367504576, "grad_norm": 7.407886505126953, "learning_rate": 9.234992128140188e-05, "loss": 0.6748, "step": 6339 }, { "epoch": 0.4295683989430178, "grad_norm": 6.933940887451172, "learning_rate": 9.234855226230406e-05, "loss": 1.0334, "step": 6340 }, { "epoch": 0.4296361542109899, "grad_norm": 6.4698686599731445, "learning_rate": 9.234718324320624e-05, "loss": 0.8778, "step": 6341 }, { "epoch": 0.429703909478962, "grad_norm": 7.047656059265137, "learning_rate": 9.234581422410844e-05, "loss": 0.9406, "step": 6342 }, { "epoch": 0.42977166474693407, "grad_norm": 6.107243061065674, "learning_rate": 9.234444520501062e-05, "loss": 0.716, "step": 6343 }, { "epoch": 0.42983942001490616, "grad_norm": 7.644023895263672, "learning_rate": 9.23430761859128e-05, "loss": 0.9632, "step": 6344 }, { "epoch": 0.42990717528287825, "grad_norm": 8.450486183166504, "learning_rate": 9.234170716681499e-05, "loss": 0.9608, "step": 6345 }, { "epoch": 0.42997493055085034, "grad_norm": 6.914335250854492, "learning_rate": 9.234033814771717e-05, "loss": 0.9473, "step": 6346 }, { "epoch": 0.43004268581882243, "grad_norm": 7.677610874176025, "learning_rate": 9.233896912861935e-05, "loss": 0.7403, "step": 6347 }, { "epoch": 0.4301104410867945, "grad_norm": 8.03879165649414, "learning_rate": 9.233760010952153e-05, "loss": 1.1334, "step": 6348 }, { "epoch": 0.43017819635476656, "grad_norm": 8.243664741516113, "learning_rate": 9.233623109042373e-05, "loss": 0.8745, "step": 6349 }, { "epoch": 0.43024595162273865, "grad_norm": 5.417998313903809, "learning_rate": 9.23348620713259e-05, "loss": 0.8243, "step": 6350 }, { "epoch": 0.43031370689071075, "grad_norm": 9.148211479187012, "learning_rate": 9.233349305222809e-05, "loss": 1.0485, "step": 6351 }, { "epoch": 0.43038146215868284, "grad_norm": 5.762056827545166, "learning_rate": 9.233212403313027e-05, "loss": 0.8528, "step": 6352 }, { "epoch": 0.43044921742665493, "grad_norm": 7.960264682769775, "learning_rate": 9.233075501403246e-05, "loss": 1.1308, "step": 6353 }, { "epoch": 0.430516972694627, "grad_norm": 5.9596710205078125, "learning_rate": 9.232938599493464e-05, "loss": 0.8385, "step": 6354 }, { "epoch": 0.4305847279625991, "grad_norm": 7.483585357666016, "learning_rate": 9.232801697583682e-05, "loss": 0.8117, "step": 6355 }, { "epoch": 0.4306524832305712, "grad_norm": 11.998135566711426, "learning_rate": 9.2326647956739e-05, "loss": 0.8077, "step": 6356 }, { "epoch": 0.43072023849854324, "grad_norm": 5.418194770812988, "learning_rate": 9.232527893764118e-05, "loss": 0.6575, "step": 6357 }, { "epoch": 0.43078799376651533, "grad_norm": 7.568809509277344, "learning_rate": 9.232390991854338e-05, "loss": 0.9036, "step": 6358 }, { "epoch": 0.4308557490344874, "grad_norm": 6.613016128540039, "learning_rate": 9.232254089944556e-05, "loss": 0.9918, "step": 6359 }, { "epoch": 0.4309235043024595, "grad_norm": 7.7656683921813965, "learning_rate": 9.232117188034774e-05, "loss": 1.1026, "step": 6360 }, { "epoch": 0.4309912595704316, "grad_norm": 6.714265823364258, "learning_rate": 9.231980286124992e-05, "loss": 0.8636, "step": 6361 }, { "epoch": 0.4310590148384037, "grad_norm": 6.468122482299805, "learning_rate": 9.231843384215211e-05, "loss": 0.7608, "step": 6362 }, { "epoch": 0.4311267701063758, "grad_norm": 7.882842540740967, "learning_rate": 9.231706482305429e-05, "loss": 0.8926, "step": 6363 }, { "epoch": 0.4311945253743479, "grad_norm": 6.002429485321045, "learning_rate": 9.231569580395647e-05, "loss": 0.8137, "step": 6364 }, { "epoch": 0.4312622806423199, "grad_norm": 6.818765163421631, "learning_rate": 9.231432678485865e-05, "loss": 0.8249, "step": 6365 }, { "epoch": 0.431330035910292, "grad_norm": 8.535319328308105, "learning_rate": 9.231295776576083e-05, "loss": 0.888, "step": 6366 }, { "epoch": 0.4313977911782641, "grad_norm": 7.823378086090088, "learning_rate": 9.231158874666303e-05, "loss": 0.8847, "step": 6367 }, { "epoch": 0.4314655464462362, "grad_norm": 7.392731189727783, "learning_rate": 9.23102197275652e-05, "loss": 0.7595, "step": 6368 }, { "epoch": 0.4315333017142083, "grad_norm": 5.637443542480469, "learning_rate": 9.230885070846739e-05, "loss": 0.6636, "step": 6369 }, { "epoch": 0.4316010569821804, "grad_norm": 8.26915454864502, "learning_rate": 9.230748168936957e-05, "loss": 0.7704, "step": 6370 }, { "epoch": 0.43166881225015247, "grad_norm": 8.045785903930664, "learning_rate": 9.230611267027175e-05, "loss": 0.8427, "step": 6371 }, { "epoch": 0.43173656751812456, "grad_norm": 8.981038093566895, "learning_rate": 9.230474365117394e-05, "loss": 0.9997, "step": 6372 }, { "epoch": 0.4318043227860966, "grad_norm": 7.137066841125488, "learning_rate": 9.230337463207612e-05, "loss": 0.7654, "step": 6373 }, { "epoch": 0.4318720780540687, "grad_norm": 6.7576165199279785, "learning_rate": 9.23020056129783e-05, "loss": 0.8766, "step": 6374 }, { "epoch": 0.4319398333220408, "grad_norm": 6.381602764129639, "learning_rate": 9.230063659388048e-05, "loss": 0.9203, "step": 6375 }, { "epoch": 0.43200758859001287, "grad_norm": 6.969717502593994, "learning_rate": 9.229926757478268e-05, "loss": 0.9507, "step": 6376 }, { "epoch": 0.43207534385798496, "grad_norm": 5.76108455657959, "learning_rate": 9.229789855568486e-05, "loss": 0.8533, "step": 6377 }, { "epoch": 0.43214309912595705, "grad_norm": 6.446774005889893, "learning_rate": 9.229652953658704e-05, "loss": 0.9094, "step": 6378 }, { "epoch": 0.43221085439392914, "grad_norm": 9.543696403503418, "learning_rate": 9.229516051748922e-05, "loss": 0.8955, "step": 6379 }, { "epoch": 0.43227860966190124, "grad_norm": 8.095965385437012, "learning_rate": 9.22937914983914e-05, "loss": 0.633, "step": 6380 }, { "epoch": 0.4323463649298733, "grad_norm": 9.561396598815918, "learning_rate": 9.229242247929359e-05, "loss": 0.9367, "step": 6381 }, { "epoch": 0.43241412019784536, "grad_norm": 7.348874092102051, "learning_rate": 9.229105346019577e-05, "loss": 0.7585, "step": 6382 }, { "epoch": 0.43248187546581746, "grad_norm": 5.866179466247559, "learning_rate": 9.228968444109795e-05, "loss": 0.8046, "step": 6383 }, { "epoch": 0.43254963073378955, "grad_norm": 7.945270538330078, "learning_rate": 9.228831542200013e-05, "loss": 0.8599, "step": 6384 }, { "epoch": 0.43261738600176164, "grad_norm": 6.365564346313477, "learning_rate": 9.228694640290233e-05, "loss": 0.8213, "step": 6385 }, { "epoch": 0.43268514126973373, "grad_norm": 7.802513599395752, "learning_rate": 9.22855773838045e-05, "loss": 0.7779, "step": 6386 }, { "epoch": 0.4327528965377058, "grad_norm": 7.740975379943848, "learning_rate": 9.228420836470669e-05, "loss": 0.9564, "step": 6387 }, { "epoch": 0.4328206518056779, "grad_norm": 8.962937355041504, "learning_rate": 9.228283934560887e-05, "loss": 0.7704, "step": 6388 }, { "epoch": 0.43288840707364995, "grad_norm": 9.266582489013672, "learning_rate": 9.228147032651106e-05, "loss": 1.055, "step": 6389 }, { "epoch": 0.43295616234162204, "grad_norm": 8.01395320892334, "learning_rate": 9.228010130741324e-05, "loss": 1.0197, "step": 6390 }, { "epoch": 0.43302391760959413, "grad_norm": 6.83071231842041, "learning_rate": 9.227873228831543e-05, "loss": 0.8489, "step": 6391 }, { "epoch": 0.4330916728775662, "grad_norm": 7.923128604888916, "learning_rate": 9.227736326921762e-05, "loss": 0.8601, "step": 6392 }, { "epoch": 0.4331594281455383, "grad_norm": 6.395363807678223, "learning_rate": 9.22759942501198e-05, "loss": 0.8167, "step": 6393 }, { "epoch": 0.4332271834135104, "grad_norm": 5.6835408210754395, "learning_rate": 9.227462523102198e-05, "loss": 0.895, "step": 6394 }, { "epoch": 0.4332949386814825, "grad_norm": 7.308006286621094, "learning_rate": 9.227325621192417e-05, "loss": 0.985, "step": 6395 }, { "epoch": 0.4333626939494546, "grad_norm": 6.754067897796631, "learning_rate": 9.227188719282635e-05, "loss": 0.9144, "step": 6396 }, { "epoch": 0.4334304492174266, "grad_norm": 7.917202949523926, "learning_rate": 9.227051817372853e-05, "loss": 0.8683, "step": 6397 }, { "epoch": 0.4334982044853987, "grad_norm": 7.502418041229248, "learning_rate": 9.226914915463071e-05, "loss": 1.0849, "step": 6398 }, { "epoch": 0.4335659597533708, "grad_norm": 6.86057186126709, "learning_rate": 9.22677801355329e-05, "loss": 0.9286, "step": 6399 }, { "epoch": 0.4336337150213429, "grad_norm": 8.595507621765137, "learning_rate": 9.226641111643509e-05, "loss": 0.9766, "step": 6400 }, { "epoch": 0.433701470289315, "grad_norm": 6.231963157653809, "learning_rate": 9.226504209733727e-05, "loss": 0.8714, "step": 6401 }, { "epoch": 0.4337692255572871, "grad_norm": 7.1258625984191895, "learning_rate": 9.226367307823945e-05, "loss": 1.1432, "step": 6402 }, { "epoch": 0.4338369808252592, "grad_norm": 6.715908050537109, "learning_rate": 9.226230405914163e-05, "loss": 0.8271, "step": 6403 }, { "epoch": 0.43390473609323127, "grad_norm": 7.551729679107666, "learning_rate": 9.226093504004382e-05, "loss": 0.8919, "step": 6404 }, { "epoch": 0.43397249136120336, "grad_norm": 7.136280059814453, "learning_rate": 9.2259566020946e-05, "loss": 1.0586, "step": 6405 }, { "epoch": 0.4340402466291754, "grad_norm": 5.612763404846191, "learning_rate": 9.225819700184818e-05, "loss": 0.7163, "step": 6406 }, { "epoch": 0.4341080018971475, "grad_norm": 7.445765972137451, "learning_rate": 9.225682798275036e-05, "loss": 1.0488, "step": 6407 }, { "epoch": 0.4341757571651196, "grad_norm": 7.239571571350098, "learning_rate": 9.225545896365255e-05, "loss": 0.9174, "step": 6408 }, { "epoch": 0.43424351243309167, "grad_norm": 7.86649751663208, "learning_rate": 9.225408994455474e-05, "loss": 0.9312, "step": 6409 }, { "epoch": 0.43431126770106376, "grad_norm": 7.049384117126465, "learning_rate": 9.225272092545692e-05, "loss": 0.8739, "step": 6410 }, { "epoch": 0.43437902296903586, "grad_norm": 6.824424743652344, "learning_rate": 9.22513519063591e-05, "loss": 0.9335, "step": 6411 }, { "epoch": 0.43444677823700795, "grad_norm": 6.616923809051514, "learning_rate": 9.224998288726128e-05, "loss": 0.7854, "step": 6412 }, { "epoch": 0.43451453350498004, "grad_norm": 6.989858150482178, "learning_rate": 9.224861386816347e-05, "loss": 0.915, "step": 6413 }, { "epoch": 0.4345822887729521, "grad_norm": 7.5784406661987305, "learning_rate": 9.224724484906565e-05, "loss": 0.9316, "step": 6414 }, { "epoch": 0.43465004404092417, "grad_norm": 5.958365440368652, "learning_rate": 9.224587582996783e-05, "loss": 0.9103, "step": 6415 }, { "epoch": 0.43471779930889626, "grad_norm": 6.811526775360107, "learning_rate": 9.224450681087001e-05, "loss": 0.8778, "step": 6416 }, { "epoch": 0.43478555457686835, "grad_norm": 6.9535722732543945, "learning_rate": 9.22431377917722e-05, "loss": 0.9407, "step": 6417 }, { "epoch": 0.43485330984484044, "grad_norm": 7.412068843841553, "learning_rate": 9.224176877267439e-05, "loss": 0.7526, "step": 6418 }, { "epoch": 0.43492106511281253, "grad_norm": 6.990227699279785, "learning_rate": 9.224039975357657e-05, "loss": 0.8432, "step": 6419 }, { "epoch": 0.4349888203807846, "grad_norm": 7.477065563201904, "learning_rate": 9.223903073447875e-05, "loss": 0.9579, "step": 6420 }, { "epoch": 0.4350565756487567, "grad_norm": 6.9343581199646, "learning_rate": 9.223766171538093e-05, "loss": 0.7967, "step": 6421 }, { "epoch": 0.43512433091672875, "grad_norm": 7.360040664672852, "learning_rate": 9.223629269628312e-05, "loss": 0.807, "step": 6422 }, { "epoch": 0.43519208618470084, "grad_norm": 6.708899974822998, "learning_rate": 9.22349236771853e-05, "loss": 0.723, "step": 6423 }, { "epoch": 0.43525984145267294, "grad_norm": 7.195553302764893, "learning_rate": 9.223355465808748e-05, "loss": 1.0543, "step": 6424 }, { "epoch": 0.435327596720645, "grad_norm": 6.833982467651367, "learning_rate": 9.223218563898966e-05, "loss": 0.7874, "step": 6425 }, { "epoch": 0.4353953519886171, "grad_norm": 7.278003692626953, "learning_rate": 9.223081661989184e-05, "loss": 1.0377, "step": 6426 }, { "epoch": 0.4354631072565892, "grad_norm": 6.5219407081604, "learning_rate": 9.222944760079404e-05, "loss": 0.8975, "step": 6427 }, { "epoch": 0.4355308625245613, "grad_norm": 15.948746681213379, "learning_rate": 9.222807858169622e-05, "loss": 1.1554, "step": 6428 }, { "epoch": 0.4355986177925334, "grad_norm": 8.12380599975586, "learning_rate": 9.22267095625984e-05, "loss": 0.8078, "step": 6429 }, { "epoch": 0.43566637306050543, "grad_norm": 6.6606340408325195, "learning_rate": 9.222534054350058e-05, "loss": 1.065, "step": 6430 }, { "epoch": 0.4357341283284775, "grad_norm": 8.079832077026367, "learning_rate": 9.222397152440277e-05, "loss": 0.9253, "step": 6431 }, { "epoch": 0.4358018835964496, "grad_norm": 6.005470275878906, "learning_rate": 9.222260250530495e-05, "loss": 0.8853, "step": 6432 }, { "epoch": 0.4358696388644217, "grad_norm": 8.552530288696289, "learning_rate": 9.222123348620713e-05, "loss": 0.9477, "step": 6433 }, { "epoch": 0.4359373941323938, "grad_norm": 6.014112949371338, "learning_rate": 9.221986446710931e-05, "loss": 0.9249, "step": 6434 }, { "epoch": 0.4360051494003659, "grad_norm": 8.349777221679688, "learning_rate": 9.22184954480115e-05, "loss": 0.9278, "step": 6435 }, { "epoch": 0.436072904668338, "grad_norm": 7.474494934082031, "learning_rate": 9.221712642891369e-05, "loss": 1.1279, "step": 6436 }, { "epoch": 0.43614065993631007, "grad_norm": 7.27971076965332, "learning_rate": 9.221575740981587e-05, "loss": 0.8476, "step": 6437 }, { "epoch": 0.4362084152042821, "grad_norm": 7.819595813751221, "learning_rate": 9.221438839071806e-05, "loss": 0.828, "step": 6438 }, { "epoch": 0.4362761704722542, "grad_norm": 8.474798202514648, "learning_rate": 9.221301937162024e-05, "loss": 1.0775, "step": 6439 }, { "epoch": 0.4363439257402263, "grad_norm": 7.175133228302002, "learning_rate": 9.221165035252242e-05, "loss": 0.8889, "step": 6440 }, { "epoch": 0.4364116810081984, "grad_norm": 5.881869316101074, "learning_rate": 9.221028133342461e-05, "loss": 0.7435, "step": 6441 }, { "epoch": 0.4364794362761705, "grad_norm": 8.366180419921875, "learning_rate": 9.22089123143268e-05, "loss": 1.1461, "step": 6442 }, { "epoch": 0.43654719154414257, "grad_norm": 7.311695098876953, "learning_rate": 9.220754329522898e-05, "loss": 0.7601, "step": 6443 }, { "epoch": 0.43661494681211466, "grad_norm": 6.886138439178467, "learning_rate": 9.220617427613116e-05, "loss": 0.9432, "step": 6444 }, { "epoch": 0.43668270208008675, "grad_norm": 10.014945983886719, "learning_rate": 9.220480525703335e-05, "loss": 0.8914, "step": 6445 }, { "epoch": 0.4367504573480588, "grad_norm": 6.025598526000977, "learning_rate": 9.220343623793553e-05, "loss": 0.8601, "step": 6446 }, { "epoch": 0.4368182126160309, "grad_norm": 6.072788715362549, "learning_rate": 9.220206721883771e-05, "loss": 0.7345, "step": 6447 }, { "epoch": 0.43688596788400297, "grad_norm": 8.272831916809082, "learning_rate": 9.220069819973989e-05, "loss": 0.7427, "step": 6448 }, { "epoch": 0.43695372315197506, "grad_norm": 6.721330642700195, "learning_rate": 9.219932918064207e-05, "loss": 0.8294, "step": 6449 }, { "epoch": 0.43702147841994715, "grad_norm": 7.108659267425537, "learning_rate": 9.219796016154426e-05, "loss": 0.7635, "step": 6450 }, { "epoch": 0.43708923368791924, "grad_norm": 11.399370193481445, "learning_rate": 9.219659114244645e-05, "loss": 1.0645, "step": 6451 }, { "epoch": 0.43715698895589133, "grad_norm": 7.976221561431885, "learning_rate": 9.219522212334863e-05, "loss": 0.9437, "step": 6452 }, { "epoch": 0.4372247442238634, "grad_norm": 8.320938110351562, "learning_rate": 9.21938531042508e-05, "loss": 0.8029, "step": 6453 }, { "epoch": 0.4372924994918355, "grad_norm": 6.146340370178223, "learning_rate": 9.2192484085153e-05, "loss": 0.7872, "step": 6454 }, { "epoch": 0.43736025475980755, "grad_norm": 6.408365726470947, "learning_rate": 9.219111506605518e-05, "loss": 0.5922, "step": 6455 }, { "epoch": 0.43742801002777965, "grad_norm": 7.011279106140137, "learning_rate": 9.218974604695736e-05, "loss": 0.9152, "step": 6456 }, { "epoch": 0.43749576529575174, "grad_norm": 6.819314002990723, "learning_rate": 9.218837702785954e-05, "loss": 1.002, "step": 6457 }, { "epoch": 0.43756352056372383, "grad_norm": 10.63784122467041, "learning_rate": 9.218700800876172e-05, "loss": 1.0873, "step": 6458 }, { "epoch": 0.4376312758316959, "grad_norm": 6.964840888977051, "learning_rate": 9.218563898966391e-05, "loss": 0.6972, "step": 6459 }, { "epoch": 0.437699031099668, "grad_norm": 8.421072959899902, "learning_rate": 9.21842699705661e-05, "loss": 1.0543, "step": 6460 }, { "epoch": 0.4377667863676401, "grad_norm": 5.883727073669434, "learning_rate": 9.218290095146828e-05, "loss": 0.8138, "step": 6461 }, { "epoch": 0.4378345416356122, "grad_norm": 8.054207801818848, "learning_rate": 9.218153193237046e-05, "loss": 0.8885, "step": 6462 }, { "epoch": 0.43790229690358423, "grad_norm": 7.546504974365234, "learning_rate": 9.218016291327265e-05, "loss": 1.2157, "step": 6463 }, { "epoch": 0.4379700521715563, "grad_norm": 10.255965232849121, "learning_rate": 9.217879389417483e-05, "loss": 1.0082, "step": 6464 }, { "epoch": 0.4380378074395284, "grad_norm": 6.057038307189941, "learning_rate": 9.217742487507701e-05, "loss": 0.8386, "step": 6465 }, { "epoch": 0.4381055627075005, "grad_norm": 8.343341827392578, "learning_rate": 9.217605585597919e-05, "loss": 0.7051, "step": 6466 }, { "epoch": 0.4381733179754726, "grad_norm": 8.163117408752441, "learning_rate": 9.217468683688137e-05, "loss": 1.1442, "step": 6467 }, { "epoch": 0.4382410732434447, "grad_norm": 6.132575035095215, "learning_rate": 9.217331781778357e-05, "loss": 0.7034, "step": 6468 }, { "epoch": 0.4383088285114168, "grad_norm": 6.2581658363342285, "learning_rate": 9.217194879868575e-05, "loss": 0.919, "step": 6469 }, { "epoch": 0.4383765837793889, "grad_norm": 6.228326797485352, "learning_rate": 9.217057977958793e-05, "loss": 0.7785, "step": 6470 }, { "epoch": 0.4384443390473609, "grad_norm": 5.782427787780762, "learning_rate": 9.21692107604901e-05, "loss": 0.9185, "step": 6471 }, { "epoch": 0.438512094315333, "grad_norm": 5.726394176483154, "learning_rate": 9.21678417413923e-05, "loss": 0.593, "step": 6472 }, { "epoch": 0.4385798495833051, "grad_norm": 7.62056303024292, "learning_rate": 9.216647272229448e-05, "loss": 0.8947, "step": 6473 }, { "epoch": 0.4386476048512772, "grad_norm": 7.467504024505615, "learning_rate": 9.216510370319666e-05, "loss": 0.8738, "step": 6474 }, { "epoch": 0.4387153601192493, "grad_norm": 6.0644989013671875, "learning_rate": 9.216373468409884e-05, "loss": 0.7715, "step": 6475 }, { "epoch": 0.43878311538722137, "grad_norm": 7.380848407745361, "learning_rate": 9.216236566500102e-05, "loss": 0.8055, "step": 6476 }, { "epoch": 0.43885087065519346, "grad_norm": 7.162757873535156, "learning_rate": 9.216099664590322e-05, "loss": 0.8826, "step": 6477 }, { "epoch": 0.43891862592316555, "grad_norm": 8.315613746643066, "learning_rate": 9.21596276268054e-05, "loss": 1.1239, "step": 6478 }, { "epoch": 0.4389863811911376, "grad_norm": 6.754839897155762, "learning_rate": 9.215825860770758e-05, "loss": 0.8331, "step": 6479 }, { "epoch": 0.4390541364591097, "grad_norm": 5.4843831062316895, "learning_rate": 9.215688958860976e-05, "loss": 0.8134, "step": 6480 }, { "epoch": 0.43912189172708177, "grad_norm": 7.7780656814575195, "learning_rate": 9.215552056951195e-05, "loss": 1.0216, "step": 6481 }, { "epoch": 0.43918964699505386, "grad_norm": 7.779257297515869, "learning_rate": 9.215415155041413e-05, "loss": 0.91, "step": 6482 }, { "epoch": 0.43925740226302595, "grad_norm": 7.926174640655518, "learning_rate": 9.215278253131631e-05, "loss": 0.9635, "step": 6483 }, { "epoch": 0.43932515753099805, "grad_norm": 5.595751762390137, "learning_rate": 9.21514135122185e-05, "loss": 0.9092, "step": 6484 }, { "epoch": 0.43939291279897014, "grad_norm": 8.04053020477295, "learning_rate": 9.215004449312069e-05, "loss": 0.6872, "step": 6485 }, { "epoch": 0.43946066806694223, "grad_norm": 12.462115287780762, "learning_rate": 9.214867547402287e-05, "loss": 1.2741, "step": 6486 }, { "epoch": 0.43952842333491426, "grad_norm": 6.239933490753174, "learning_rate": 9.214730645492506e-05, "loss": 0.7711, "step": 6487 }, { "epoch": 0.43959617860288636, "grad_norm": 7.584579944610596, "learning_rate": 9.214593743582724e-05, "loss": 0.8666, "step": 6488 }, { "epoch": 0.43966393387085845, "grad_norm": 7.129024505615234, "learning_rate": 9.214456841672942e-05, "loss": 0.8664, "step": 6489 }, { "epoch": 0.43973168913883054, "grad_norm": 7.35465145111084, "learning_rate": 9.21431993976316e-05, "loss": 1.2072, "step": 6490 }, { "epoch": 0.43979944440680263, "grad_norm": 7.911463737487793, "learning_rate": 9.21418303785338e-05, "loss": 0.8019, "step": 6491 }, { "epoch": 0.4398671996747747, "grad_norm": 7.9478349685668945, "learning_rate": 9.214046135943597e-05, "loss": 0.8945, "step": 6492 }, { "epoch": 0.4399349549427468, "grad_norm": 7.73642635345459, "learning_rate": 9.213909234033815e-05, "loss": 1.2203, "step": 6493 }, { "epoch": 0.4400027102107189, "grad_norm": 8.306556701660156, "learning_rate": 9.213772332124034e-05, "loss": 0.8548, "step": 6494 }, { "epoch": 0.44007046547869094, "grad_norm": 7.496469497680664, "learning_rate": 9.213635430214253e-05, "loss": 1.1485, "step": 6495 }, { "epoch": 0.44013822074666303, "grad_norm": 6.734534740447998, "learning_rate": 9.213498528304471e-05, "loss": 0.8758, "step": 6496 }, { "epoch": 0.4402059760146351, "grad_norm": 6.782371997833252, "learning_rate": 9.213361626394689e-05, "loss": 0.5879, "step": 6497 }, { "epoch": 0.4402737312826072, "grad_norm": 6.044846057891846, "learning_rate": 9.213224724484907e-05, "loss": 1.0055, "step": 6498 }, { "epoch": 0.4403414865505793, "grad_norm": 7.979533672332764, "learning_rate": 9.213087822575125e-05, "loss": 0.8265, "step": 6499 }, { "epoch": 0.4404092418185514, "grad_norm": 8.245573997497559, "learning_rate": 9.212950920665344e-05, "loss": 0.5549, "step": 6500 }, { "epoch": 0.4404769970865235, "grad_norm": 6.346557140350342, "learning_rate": 9.212814018755562e-05, "loss": 0.857, "step": 6501 }, { "epoch": 0.4405447523544956, "grad_norm": 6.079600811004639, "learning_rate": 9.21267711684578e-05, "loss": 0.9127, "step": 6502 }, { "epoch": 0.4406125076224676, "grad_norm": 9.402070045471191, "learning_rate": 9.212540214935999e-05, "loss": 0.8104, "step": 6503 }, { "epoch": 0.4406802628904397, "grad_norm": 8.062244415283203, "learning_rate": 9.212403313026217e-05, "loss": 1.195, "step": 6504 }, { "epoch": 0.4407480181584118, "grad_norm": 8.54050064086914, "learning_rate": 9.212266411116436e-05, "loss": 0.854, "step": 6505 }, { "epoch": 0.4408157734263839, "grad_norm": 6.040203094482422, "learning_rate": 9.212129509206654e-05, "loss": 0.7028, "step": 6506 }, { "epoch": 0.440883528694356, "grad_norm": 7.689701080322266, "learning_rate": 9.211992607296872e-05, "loss": 0.9999, "step": 6507 }, { "epoch": 0.4409512839623281, "grad_norm": 7.860703468322754, "learning_rate": 9.21185570538709e-05, "loss": 1.0413, "step": 6508 }, { "epoch": 0.44101903923030017, "grad_norm": 7.390681266784668, "learning_rate": 9.21171880347731e-05, "loss": 1.0337, "step": 6509 }, { "epoch": 0.44108679449827226, "grad_norm": 6.800081253051758, "learning_rate": 9.211581901567527e-05, "loss": 0.9613, "step": 6510 }, { "epoch": 0.44115454976624435, "grad_norm": 5.59462833404541, "learning_rate": 9.211444999657746e-05, "loss": 0.7001, "step": 6511 }, { "epoch": 0.4412223050342164, "grad_norm": 8.300586700439453, "learning_rate": 9.211308097747964e-05, "loss": 1.0087, "step": 6512 }, { "epoch": 0.4412900603021885, "grad_norm": 6.590997695922852, "learning_rate": 9.211171195838182e-05, "loss": 0.8233, "step": 6513 }, { "epoch": 0.44135781557016057, "grad_norm": 8.995779991149902, "learning_rate": 9.211034293928401e-05, "loss": 0.9385, "step": 6514 }, { "epoch": 0.44142557083813266, "grad_norm": 6.689282417297363, "learning_rate": 9.210897392018619e-05, "loss": 0.9455, "step": 6515 }, { "epoch": 0.44149332610610476, "grad_norm": 6.494836330413818, "learning_rate": 9.210760490108837e-05, "loss": 0.9009, "step": 6516 }, { "epoch": 0.44156108137407685, "grad_norm": 7.502823352813721, "learning_rate": 9.210623588199055e-05, "loss": 0.7972, "step": 6517 }, { "epoch": 0.44162883664204894, "grad_norm": 6.561639785766602, "learning_rate": 9.210486686289274e-05, "loss": 0.7988, "step": 6518 }, { "epoch": 0.44169659191002103, "grad_norm": 6.1523847579956055, "learning_rate": 9.210349784379493e-05, "loss": 0.9561, "step": 6519 }, { "epoch": 0.44176434717799307, "grad_norm": 6.792056560516357, "learning_rate": 9.21021288246971e-05, "loss": 0.7623, "step": 6520 }, { "epoch": 0.44183210244596516, "grad_norm": 6.477465629577637, "learning_rate": 9.210075980559929e-05, "loss": 0.6875, "step": 6521 }, { "epoch": 0.44189985771393725, "grad_norm": 7.468865394592285, "learning_rate": 9.209939078650147e-05, "loss": 0.8727, "step": 6522 }, { "epoch": 0.44196761298190934, "grad_norm": 5.351073265075684, "learning_rate": 9.209802176740366e-05, "loss": 0.8114, "step": 6523 }, { "epoch": 0.44203536824988143, "grad_norm": 7.701816558837891, "learning_rate": 9.209665274830584e-05, "loss": 1.1438, "step": 6524 }, { "epoch": 0.4421031235178535, "grad_norm": 6.072995662689209, "learning_rate": 9.209528372920802e-05, "loss": 0.9237, "step": 6525 }, { "epoch": 0.4421708787858256, "grad_norm": 7.1209716796875, "learning_rate": 9.20939147101102e-05, "loss": 0.8445, "step": 6526 }, { "epoch": 0.4422386340537977, "grad_norm": 7.518238067626953, "learning_rate": 9.20925456910124e-05, "loss": 1.0969, "step": 6527 }, { "epoch": 0.44230638932176974, "grad_norm": 5.822054862976074, "learning_rate": 9.209117667191458e-05, "loss": 0.843, "step": 6528 }, { "epoch": 0.44237414458974184, "grad_norm": 6.8769612312316895, "learning_rate": 9.208980765281676e-05, "loss": 0.9661, "step": 6529 }, { "epoch": 0.4424418998577139, "grad_norm": 6.266312122344971, "learning_rate": 9.208843863371895e-05, "loss": 0.7924, "step": 6530 }, { "epoch": 0.442509655125686, "grad_norm": 6.57719612121582, "learning_rate": 9.208706961462113e-05, "loss": 0.7929, "step": 6531 }, { "epoch": 0.4425774103936581, "grad_norm": 6.038028240203857, "learning_rate": 9.208570059552331e-05, "loss": 0.7683, "step": 6532 }, { "epoch": 0.4426451656616302, "grad_norm": 7.042256832122803, "learning_rate": 9.20843315764255e-05, "loss": 0.6646, "step": 6533 }, { "epoch": 0.4427129209296023, "grad_norm": 7.218042850494385, "learning_rate": 9.208296255732768e-05, "loss": 0.9532, "step": 6534 }, { "epoch": 0.4427806761975744, "grad_norm": 8.851286888122559, "learning_rate": 9.208159353822986e-05, "loss": 0.819, "step": 6535 }, { "epoch": 0.4428484314655464, "grad_norm": 5.37313985824585, "learning_rate": 9.208022451913205e-05, "loss": 0.6068, "step": 6536 }, { "epoch": 0.4429161867335185, "grad_norm": 6.643946647644043, "learning_rate": 9.207885550003424e-05, "loss": 1.1088, "step": 6537 }, { "epoch": 0.4429839420014906, "grad_norm": 7.155415058135986, "learning_rate": 9.207748648093642e-05, "loss": 0.9693, "step": 6538 }, { "epoch": 0.4430516972694627, "grad_norm": 7.293460369110107, "learning_rate": 9.20761174618386e-05, "loss": 1.0701, "step": 6539 }, { "epoch": 0.4431194525374348, "grad_norm": 6.664429664611816, "learning_rate": 9.207474844274078e-05, "loss": 0.6425, "step": 6540 }, { "epoch": 0.4431872078054069, "grad_norm": 5.979015350341797, "learning_rate": 9.207337942364297e-05, "loss": 0.6666, "step": 6541 }, { "epoch": 0.44325496307337897, "grad_norm": 5.959255695343018, "learning_rate": 9.207201040454515e-05, "loss": 0.6914, "step": 6542 }, { "epoch": 0.44332271834135106, "grad_norm": 7.694621562957764, "learning_rate": 9.207064138544733e-05, "loss": 0.8844, "step": 6543 }, { "epoch": 0.4433904736093231, "grad_norm": 7.128264904022217, "learning_rate": 9.206927236634951e-05, "loss": 0.6571, "step": 6544 }, { "epoch": 0.4434582288772952, "grad_norm": 7.676843166351318, "learning_rate": 9.20679033472517e-05, "loss": 0.8944, "step": 6545 }, { "epoch": 0.4435259841452673, "grad_norm": 8.858941078186035, "learning_rate": 9.206653432815389e-05, "loss": 0.8057, "step": 6546 }, { "epoch": 0.4435937394132394, "grad_norm": 7.0214033126831055, "learning_rate": 9.206516530905607e-05, "loss": 0.7815, "step": 6547 }, { "epoch": 0.44366149468121147, "grad_norm": 5.209787845611572, "learning_rate": 9.206379628995825e-05, "loss": 0.8764, "step": 6548 }, { "epoch": 0.44372924994918356, "grad_norm": 7.076624870300293, "learning_rate": 9.206242727086043e-05, "loss": 0.8881, "step": 6549 }, { "epoch": 0.44379700521715565, "grad_norm": 7.659173965454102, "learning_rate": 9.206105825176262e-05, "loss": 0.8945, "step": 6550 }, { "epoch": 0.44386476048512774, "grad_norm": 6.2118611335754395, "learning_rate": 9.20596892326648e-05, "loss": 0.8022, "step": 6551 }, { "epoch": 0.4439325157530998, "grad_norm": 6.649642467498779, "learning_rate": 9.205832021356698e-05, "loss": 0.9765, "step": 6552 }, { "epoch": 0.44400027102107187, "grad_norm": 6.0131683349609375, "learning_rate": 9.205695119446917e-05, "loss": 0.6928, "step": 6553 }, { "epoch": 0.44406802628904396, "grad_norm": 6.295266151428223, "learning_rate": 9.205558217537135e-05, "loss": 0.9785, "step": 6554 }, { "epoch": 0.44413578155701605, "grad_norm": 6.524127006530762, "learning_rate": 9.205421315627354e-05, "loss": 0.8065, "step": 6555 }, { "epoch": 0.44420353682498814, "grad_norm": 7.478303909301758, "learning_rate": 9.205284413717572e-05, "loss": 0.9533, "step": 6556 }, { "epoch": 0.44427129209296023, "grad_norm": 7.92426061630249, "learning_rate": 9.20514751180779e-05, "loss": 1.1578, "step": 6557 }, { "epoch": 0.4443390473609323, "grad_norm": 7.532670497894287, "learning_rate": 9.205010609898008e-05, "loss": 0.9318, "step": 6558 }, { "epoch": 0.4444068026289044, "grad_norm": 8.458464622497559, "learning_rate": 9.204873707988226e-05, "loss": 1.087, "step": 6559 }, { "epoch": 0.4444745578968765, "grad_norm": 5.044344425201416, "learning_rate": 9.204736806078445e-05, "loss": 0.7836, "step": 6560 }, { "epoch": 0.44454231316484855, "grad_norm": 6.682579517364502, "learning_rate": 9.204599904168663e-05, "loss": 0.8407, "step": 6561 }, { "epoch": 0.44461006843282064, "grad_norm": 6.948788166046143, "learning_rate": 9.204463002258882e-05, "loss": 0.72, "step": 6562 }, { "epoch": 0.44467782370079273, "grad_norm": 7.333683967590332, "learning_rate": 9.2043261003491e-05, "loss": 0.934, "step": 6563 }, { "epoch": 0.4447455789687648, "grad_norm": 7.648477077484131, "learning_rate": 9.204189198439319e-05, "loss": 0.8992, "step": 6564 }, { "epoch": 0.4448133342367369, "grad_norm": 6.203789710998535, "learning_rate": 9.204052296529537e-05, "loss": 1.011, "step": 6565 }, { "epoch": 0.444881089504709, "grad_norm": 7.739291667938232, "learning_rate": 9.203915394619755e-05, "loss": 0.8695, "step": 6566 }, { "epoch": 0.4449488447726811, "grad_norm": 6.642831325531006, "learning_rate": 9.203778492709973e-05, "loss": 0.666, "step": 6567 }, { "epoch": 0.4450166000406532, "grad_norm": 7.760629653930664, "learning_rate": 9.203641590800191e-05, "loss": 1.1022, "step": 6568 }, { "epoch": 0.4450843553086252, "grad_norm": 6.180607318878174, "learning_rate": 9.20350468889041e-05, "loss": 0.8599, "step": 6569 }, { "epoch": 0.4451521105765973, "grad_norm": 6.284255504608154, "learning_rate": 9.203367786980629e-05, "loss": 0.7226, "step": 6570 }, { "epoch": 0.4452198658445694, "grad_norm": 5.078493595123291, "learning_rate": 9.203230885070847e-05, "loss": 0.8089, "step": 6571 }, { "epoch": 0.4452876211125415, "grad_norm": 6.533141136169434, "learning_rate": 9.203093983161065e-05, "loss": 0.8426, "step": 6572 }, { "epoch": 0.4453553763805136, "grad_norm": 6.589559555053711, "learning_rate": 9.202957081251284e-05, "loss": 1.0924, "step": 6573 }, { "epoch": 0.4454231316484857, "grad_norm": 5.858087539672852, "learning_rate": 9.202820179341502e-05, "loss": 0.8007, "step": 6574 }, { "epoch": 0.4454908869164578, "grad_norm": 6.8898491859436035, "learning_rate": 9.20268327743172e-05, "loss": 0.7622, "step": 6575 }, { "epoch": 0.44555864218442987, "grad_norm": 5.419859409332275, "learning_rate": 9.20254637552194e-05, "loss": 0.711, "step": 6576 }, { "epoch": 0.4456263974524019, "grad_norm": 7.31076192855835, "learning_rate": 9.202409473612157e-05, "loss": 1.0181, "step": 6577 }, { "epoch": 0.445694152720374, "grad_norm": 8.525489807128906, "learning_rate": 9.202272571702375e-05, "loss": 1.0324, "step": 6578 }, { "epoch": 0.4457619079883461, "grad_norm": 6.309749603271484, "learning_rate": 9.202135669792595e-05, "loss": 1.1596, "step": 6579 }, { "epoch": 0.4458296632563182, "grad_norm": 5.667306423187256, "learning_rate": 9.201998767882813e-05, "loss": 0.8091, "step": 6580 }, { "epoch": 0.44589741852429027, "grad_norm": 6.930799961090088, "learning_rate": 9.201861865973031e-05, "loss": 0.915, "step": 6581 }, { "epoch": 0.44596517379226236, "grad_norm": 8.236886024475098, "learning_rate": 9.201724964063249e-05, "loss": 0.8973, "step": 6582 }, { "epoch": 0.44603292906023445, "grad_norm": 8.804180145263672, "learning_rate": 9.201588062153468e-05, "loss": 0.9985, "step": 6583 }, { "epoch": 0.44610068432820654, "grad_norm": 6.966750621795654, "learning_rate": 9.201451160243686e-05, "loss": 0.9065, "step": 6584 }, { "epoch": 0.4461684395961786, "grad_norm": 6.3808698654174805, "learning_rate": 9.201314258333904e-05, "loss": 0.8519, "step": 6585 }, { "epoch": 0.44623619486415067, "grad_norm": 6.380456447601318, "learning_rate": 9.201177356424122e-05, "loss": 0.6365, "step": 6586 }, { "epoch": 0.44630395013212276, "grad_norm": 6.893950939178467, "learning_rate": 9.201040454514342e-05, "loss": 0.8017, "step": 6587 }, { "epoch": 0.44637170540009485, "grad_norm": 6.246983051300049, "learning_rate": 9.20090355260456e-05, "loss": 0.8914, "step": 6588 }, { "epoch": 0.44643946066806695, "grad_norm": 5.012611389160156, "learning_rate": 9.200766650694778e-05, "loss": 0.7421, "step": 6589 }, { "epoch": 0.44650721593603904, "grad_norm": 6.981064319610596, "learning_rate": 9.200629748784996e-05, "loss": 0.9045, "step": 6590 }, { "epoch": 0.44657497120401113, "grad_norm": 6.434818744659424, "learning_rate": 9.200492846875214e-05, "loss": 1.0952, "step": 6591 }, { "epoch": 0.4466427264719832, "grad_norm": 7.2661027908325195, "learning_rate": 9.200355944965433e-05, "loss": 0.8943, "step": 6592 }, { "epoch": 0.44671048173995526, "grad_norm": 6.347513675689697, "learning_rate": 9.200219043055651e-05, "loss": 0.7958, "step": 6593 }, { "epoch": 0.44677823700792735, "grad_norm": 6.98016881942749, "learning_rate": 9.20008214114587e-05, "loss": 0.7197, "step": 6594 }, { "epoch": 0.44684599227589944, "grad_norm": 5.671807765960693, "learning_rate": 9.199945239236087e-05, "loss": 0.7577, "step": 6595 }, { "epoch": 0.44691374754387153, "grad_norm": 8.79588794708252, "learning_rate": 9.199808337326307e-05, "loss": 0.9273, "step": 6596 }, { "epoch": 0.4469815028118436, "grad_norm": 7.276648998260498, "learning_rate": 9.199671435416525e-05, "loss": 0.8705, "step": 6597 }, { "epoch": 0.4470492580798157, "grad_norm": 8.435094833374023, "learning_rate": 9.199534533506743e-05, "loss": 0.9946, "step": 6598 }, { "epoch": 0.4471170133477878, "grad_norm": 6.148091793060303, "learning_rate": 9.199397631596961e-05, "loss": 0.9541, "step": 6599 }, { "epoch": 0.4471847686157599, "grad_norm": 8.867514610290527, "learning_rate": 9.199260729687179e-05, "loss": 1.029, "step": 6600 }, { "epoch": 0.44725252388373193, "grad_norm": 7.480152606964111, "learning_rate": 9.199123827777398e-05, "loss": 0.8819, "step": 6601 }, { "epoch": 0.447320279151704, "grad_norm": 7.422079086303711, "learning_rate": 9.198986925867616e-05, "loss": 0.9998, "step": 6602 }, { "epoch": 0.4473880344196761, "grad_norm": 6.239773750305176, "learning_rate": 9.198850023957834e-05, "loss": 0.9129, "step": 6603 }, { "epoch": 0.4474557896876482, "grad_norm": 6.59074592590332, "learning_rate": 9.198713122048053e-05, "loss": 1.1373, "step": 6604 }, { "epoch": 0.4475235449556203, "grad_norm": 6.096022129058838, "learning_rate": 9.198576220138272e-05, "loss": 0.8916, "step": 6605 }, { "epoch": 0.4475913002235924, "grad_norm": 5.775374889373779, "learning_rate": 9.19843931822849e-05, "loss": 0.8679, "step": 6606 }, { "epoch": 0.4476590554915645, "grad_norm": 6.836724758148193, "learning_rate": 9.198302416318708e-05, "loss": 0.8205, "step": 6607 }, { "epoch": 0.4477268107595366, "grad_norm": 6.876745700836182, "learning_rate": 9.198165514408926e-05, "loss": 0.9564, "step": 6608 }, { "epoch": 0.4477945660275086, "grad_norm": 6.725889205932617, "learning_rate": 9.198028612499144e-05, "loss": 0.7783, "step": 6609 }, { "epoch": 0.4478623212954807, "grad_norm": 6.312922477722168, "learning_rate": 9.197891710589363e-05, "loss": 0.8358, "step": 6610 }, { "epoch": 0.4479300765634528, "grad_norm": 5.7454071044921875, "learning_rate": 9.197754808679581e-05, "loss": 0.7174, "step": 6611 }, { "epoch": 0.4479978318314249, "grad_norm": 9.245071411132812, "learning_rate": 9.1976179067698e-05, "loss": 0.8224, "step": 6612 }, { "epoch": 0.448065587099397, "grad_norm": 7.2646307945251465, "learning_rate": 9.197481004860018e-05, "loss": 0.7895, "step": 6613 }, { "epoch": 0.44813334236736907, "grad_norm": 6.799703598022461, "learning_rate": 9.197344102950236e-05, "loss": 0.9685, "step": 6614 }, { "epoch": 0.44820109763534116, "grad_norm": 6.408104419708252, "learning_rate": 9.197207201040455e-05, "loss": 0.9136, "step": 6615 }, { "epoch": 0.44826885290331325, "grad_norm": 7.4527363777160645, "learning_rate": 9.197070299130673e-05, "loss": 0.9364, "step": 6616 }, { "epoch": 0.44833660817128534, "grad_norm": 6.287598609924316, "learning_rate": 9.196933397220891e-05, "loss": 0.8987, "step": 6617 }, { "epoch": 0.4484043634392574, "grad_norm": 7.270476341247559, "learning_rate": 9.196796495311109e-05, "loss": 1.0113, "step": 6618 }, { "epoch": 0.4484721187072295, "grad_norm": 8.098075866699219, "learning_rate": 9.196659593401328e-05, "loss": 0.7957, "step": 6619 }, { "epoch": 0.44853987397520156, "grad_norm": 6.268966197967529, "learning_rate": 9.196522691491546e-05, "loss": 0.7363, "step": 6620 }, { "epoch": 0.44860762924317366, "grad_norm": 8.81617259979248, "learning_rate": 9.196385789581765e-05, "loss": 1.1112, "step": 6621 }, { "epoch": 0.44867538451114575, "grad_norm": 7.593179225921631, "learning_rate": 9.196248887671984e-05, "loss": 0.7988, "step": 6622 }, { "epoch": 0.44874313977911784, "grad_norm": 7.044666290283203, "learning_rate": 9.196111985762202e-05, "loss": 0.832, "step": 6623 }, { "epoch": 0.44881089504708993, "grad_norm": 5.546746730804443, "learning_rate": 9.19597508385242e-05, "loss": 0.7273, "step": 6624 }, { "epoch": 0.448878650315062, "grad_norm": 7.123478412628174, "learning_rate": 9.19583818194264e-05, "loss": 0.9006, "step": 6625 }, { "epoch": 0.44894640558303406, "grad_norm": 9.073535919189453, "learning_rate": 9.195701280032857e-05, "loss": 0.8702, "step": 6626 }, { "epoch": 0.44901416085100615, "grad_norm": 7.320106506347656, "learning_rate": 9.195564378123075e-05, "loss": 0.7704, "step": 6627 }, { "epoch": 0.44908191611897824, "grad_norm": 7.6163177490234375, "learning_rate": 9.195427476213295e-05, "loss": 1.1432, "step": 6628 }, { "epoch": 0.44914967138695033, "grad_norm": 6.271851062774658, "learning_rate": 9.195290574303513e-05, "loss": 0.9378, "step": 6629 }, { "epoch": 0.4492174266549224, "grad_norm": 6.142657279968262, "learning_rate": 9.195153672393731e-05, "loss": 0.9356, "step": 6630 }, { "epoch": 0.4492851819228945, "grad_norm": 6.669857025146484, "learning_rate": 9.195016770483949e-05, "loss": 0.5939, "step": 6631 }, { "epoch": 0.4493529371908666, "grad_norm": 7.524440288543701, "learning_rate": 9.194879868574167e-05, "loss": 0.7756, "step": 6632 }, { "epoch": 0.4494206924588387, "grad_norm": 7.47619104385376, "learning_rate": 9.194742966664386e-05, "loss": 0.7881, "step": 6633 }, { "epoch": 0.44948844772681074, "grad_norm": 5.466754913330078, "learning_rate": 9.194606064754604e-05, "loss": 0.7653, "step": 6634 }, { "epoch": 0.4495562029947828, "grad_norm": 6.149755954742432, "learning_rate": 9.194469162844822e-05, "loss": 0.8208, "step": 6635 }, { "epoch": 0.4496239582627549, "grad_norm": 6.091528415679932, "learning_rate": 9.19433226093504e-05, "loss": 1.0214, "step": 6636 }, { "epoch": 0.449691713530727, "grad_norm": 5.599562168121338, "learning_rate": 9.194195359025258e-05, "loss": 1.0719, "step": 6637 }, { "epoch": 0.4497594687986991, "grad_norm": 7.242834091186523, "learning_rate": 9.194058457115478e-05, "loss": 1.0185, "step": 6638 }, { "epoch": 0.4498272240666712, "grad_norm": 6.523613452911377, "learning_rate": 9.193921555205696e-05, "loss": 0.9699, "step": 6639 }, { "epoch": 0.4498949793346433, "grad_norm": 6.815830230712891, "learning_rate": 9.193784653295914e-05, "loss": 0.814, "step": 6640 }, { "epoch": 0.4499627346026154, "grad_norm": 5.4793524742126465, "learning_rate": 9.193647751386132e-05, "loss": 0.724, "step": 6641 }, { "epoch": 0.4500304898705874, "grad_norm": 7.303586959838867, "learning_rate": 9.193510849476351e-05, "loss": 0.9991, "step": 6642 }, { "epoch": 0.4500982451385595, "grad_norm": 7.781108856201172, "learning_rate": 9.19337394756657e-05, "loss": 0.925, "step": 6643 }, { "epoch": 0.4501660004065316, "grad_norm": 5.924014568328857, "learning_rate": 9.193237045656787e-05, "loss": 0.8656, "step": 6644 }, { "epoch": 0.4502337556745037, "grad_norm": 8.77978801727295, "learning_rate": 9.193100143747005e-05, "loss": 0.6839, "step": 6645 }, { "epoch": 0.4503015109424758, "grad_norm": 6.725009441375732, "learning_rate": 9.192963241837223e-05, "loss": 1.09, "step": 6646 }, { "epoch": 0.45036926621044787, "grad_norm": 5.697434902191162, "learning_rate": 9.192826339927443e-05, "loss": 0.6516, "step": 6647 }, { "epoch": 0.45043702147841996, "grad_norm": 7.783257007598877, "learning_rate": 9.192689438017661e-05, "loss": 1.1097, "step": 6648 }, { "epoch": 0.45050477674639205, "grad_norm": 7.0213541984558105, "learning_rate": 9.192552536107879e-05, "loss": 1.0617, "step": 6649 }, { "epoch": 0.4505725320143641, "grad_norm": 9.152633666992188, "learning_rate": 9.192415634198097e-05, "loss": 0.9626, "step": 6650 }, { "epoch": 0.4506402872823362, "grad_norm": 8.690450668334961, "learning_rate": 9.192278732288316e-05, "loss": 0.7741, "step": 6651 }, { "epoch": 0.4507080425503083, "grad_norm": 6.784976482391357, "learning_rate": 9.192141830378534e-05, "loss": 1.092, "step": 6652 }, { "epoch": 0.45077579781828037, "grad_norm": 6.294474124908447, "learning_rate": 9.192004928468752e-05, "loss": 0.9322, "step": 6653 }, { "epoch": 0.45084355308625246, "grad_norm": 5.054771423339844, "learning_rate": 9.19186802655897e-05, "loss": 0.7711, "step": 6654 }, { "epoch": 0.45091130835422455, "grad_norm": 7.640350341796875, "learning_rate": 9.191731124649189e-05, "loss": 0.8006, "step": 6655 }, { "epoch": 0.45097906362219664, "grad_norm": 6.9720001220703125, "learning_rate": 9.191594222739408e-05, "loss": 0.7912, "step": 6656 }, { "epoch": 0.45104681889016873, "grad_norm": 7.61221981048584, "learning_rate": 9.191457320829626e-05, "loss": 0.8612, "step": 6657 }, { "epoch": 0.45111457415814077, "grad_norm": 7.7177958488464355, "learning_rate": 9.191320418919844e-05, "loss": 1.0009, "step": 6658 }, { "epoch": 0.45118232942611286, "grad_norm": 6.217735290527344, "learning_rate": 9.191183517010062e-05, "loss": 0.8091, "step": 6659 }, { "epoch": 0.45125008469408495, "grad_norm": 8.2019681930542, "learning_rate": 9.191046615100281e-05, "loss": 1.055, "step": 6660 }, { "epoch": 0.45131783996205704, "grad_norm": 5.853175640106201, "learning_rate": 9.1909097131905e-05, "loss": 1.1045, "step": 6661 }, { "epoch": 0.45138559523002914, "grad_norm": 6.276338577270508, "learning_rate": 9.190772811280717e-05, "loss": 0.9436, "step": 6662 }, { "epoch": 0.4514533504980012, "grad_norm": 11.320722579956055, "learning_rate": 9.190635909370935e-05, "loss": 1.1028, "step": 6663 }, { "epoch": 0.4515211057659733, "grad_norm": 6.311827659606934, "learning_rate": 9.190499007461154e-05, "loss": 0.9232, "step": 6664 }, { "epoch": 0.4515888610339454, "grad_norm": 8.186714172363281, "learning_rate": 9.190362105551373e-05, "loss": 0.7087, "step": 6665 }, { "epoch": 0.4516566163019175, "grad_norm": 6.127712249755859, "learning_rate": 9.190225203641591e-05, "loss": 1.0268, "step": 6666 }, { "epoch": 0.45172437156988954, "grad_norm": 6.840565204620361, "learning_rate": 9.190088301731809e-05, "loss": 0.9955, "step": 6667 }, { "epoch": 0.45179212683786163, "grad_norm": 6.302945137023926, "learning_rate": 9.189951399822027e-05, "loss": 0.9892, "step": 6668 }, { "epoch": 0.4518598821058337, "grad_norm": 9.08460521697998, "learning_rate": 9.189814497912246e-05, "loss": 0.7988, "step": 6669 }, { "epoch": 0.4519276373738058, "grad_norm": 7.497137069702148, "learning_rate": 9.189677596002464e-05, "loss": 0.8358, "step": 6670 }, { "epoch": 0.4519953926417779, "grad_norm": 6.515604496002197, "learning_rate": 9.189540694092682e-05, "loss": 0.7724, "step": 6671 }, { "epoch": 0.45206314790975, "grad_norm": 6.092652320861816, "learning_rate": 9.189403792182902e-05, "loss": 0.7308, "step": 6672 }, { "epoch": 0.4521309031777221, "grad_norm": 8.810023307800293, "learning_rate": 9.18926689027312e-05, "loss": 0.7249, "step": 6673 }, { "epoch": 0.4521986584456942, "grad_norm": 8.328235626220703, "learning_rate": 9.189129988363338e-05, "loss": 0.8115, "step": 6674 }, { "epoch": 0.4522664137136662, "grad_norm": 7.251335620880127, "learning_rate": 9.188993086453557e-05, "loss": 0.9843, "step": 6675 }, { "epoch": 0.4523341689816383, "grad_norm": 8.03663444519043, "learning_rate": 9.188856184543775e-05, "loss": 0.9925, "step": 6676 }, { "epoch": 0.4524019242496104, "grad_norm": 9.7393798828125, "learning_rate": 9.188719282633993e-05, "loss": 0.7575, "step": 6677 }, { "epoch": 0.4524696795175825, "grad_norm": 6.343038558959961, "learning_rate": 9.188582380724211e-05, "loss": 1.0638, "step": 6678 }, { "epoch": 0.4525374347855546, "grad_norm": 7.530363082885742, "learning_rate": 9.188445478814431e-05, "loss": 1.0212, "step": 6679 }, { "epoch": 0.4526051900535267, "grad_norm": 8.741231918334961, "learning_rate": 9.188308576904649e-05, "loss": 1.1723, "step": 6680 }, { "epoch": 0.45267294532149877, "grad_norm": 9.057670593261719, "learning_rate": 9.188171674994867e-05, "loss": 1.0396, "step": 6681 }, { "epoch": 0.45274070058947086, "grad_norm": 6.638017177581787, "learning_rate": 9.188034773085085e-05, "loss": 0.9465, "step": 6682 }, { "epoch": 0.4528084558574429, "grad_norm": 7.643139839172363, "learning_rate": 9.187897871175304e-05, "loss": 0.9346, "step": 6683 }, { "epoch": 0.452876211125415, "grad_norm": 7.176743984222412, "learning_rate": 9.187760969265522e-05, "loss": 1.036, "step": 6684 }, { "epoch": 0.4529439663933871, "grad_norm": 7.506284713745117, "learning_rate": 9.18762406735574e-05, "loss": 0.9487, "step": 6685 }, { "epoch": 0.45301172166135917, "grad_norm": 6.865815162658691, "learning_rate": 9.187487165445958e-05, "loss": 0.8211, "step": 6686 }, { "epoch": 0.45307947692933126, "grad_norm": 6.889796733856201, "learning_rate": 9.187350263536176e-05, "loss": 0.7901, "step": 6687 }, { "epoch": 0.45314723219730335, "grad_norm": 7.016167163848877, "learning_rate": 9.187213361626396e-05, "loss": 0.7699, "step": 6688 }, { "epoch": 0.45321498746527544, "grad_norm": 8.25091552734375, "learning_rate": 9.187076459716614e-05, "loss": 0.9778, "step": 6689 }, { "epoch": 0.45328274273324753, "grad_norm": 7.092465877532959, "learning_rate": 9.186939557806832e-05, "loss": 1.2544, "step": 6690 }, { "epoch": 0.45335049800121957, "grad_norm": 6.402629375457764, "learning_rate": 9.18680265589705e-05, "loss": 0.9225, "step": 6691 }, { "epoch": 0.45341825326919166, "grad_norm": 7.404470920562744, "learning_rate": 9.186665753987268e-05, "loss": 0.7975, "step": 6692 }, { "epoch": 0.45348600853716375, "grad_norm": 5.976271629333496, "learning_rate": 9.186528852077487e-05, "loss": 0.8868, "step": 6693 }, { "epoch": 0.45355376380513585, "grad_norm": 7.843896389007568, "learning_rate": 9.186391950167705e-05, "loss": 0.8034, "step": 6694 }, { "epoch": 0.45362151907310794, "grad_norm": 7.217418193817139, "learning_rate": 9.186255048257923e-05, "loss": 0.6763, "step": 6695 }, { "epoch": 0.45368927434108003, "grad_norm": 6.091420650482178, "learning_rate": 9.186118146348141e-05, "loss": 0.6565, "step": 6696 }, { "epoch": 0.4537570296090521, "grad_norm": 8.277983665466309, "learning_rate": 9.185981244438361e-05, "loss": 0.7589, "step": 6697 }, { "epoch": 0.4538247848770242, "grad_norm": 9.001012802124023, "learning_rate": 9.185844342528579e-05, "loss": 0.8664, "step": 6698 }, { "epoch": 0.45389254014499625, "grad_norm": 6.039417266845703, "learning_rate": 9.185707440618797e-05, "loss": 0.809, "step": 6699 }, { "epoch": 0.45396029541296834, "grad_norm": 6.930225849151611, "learning_rate": 9.185570538709015e-05, "loss": 0.7567, "step": 6700 }, { "epoch": 0.45402805068094043, "grad_norm": 7.135040760040283, "learning_rate": 9.185433636799233e-05, "loss": 0.7669, "step": 6701 }, { "epoch": 0.4540958059489125, "grad_norm": 7.994284152984619, "learning_rate": 9.185296734889452e-05, "loss": 0.7671, "step": 6702 }, { "epoch": 0.4541635612168846, "grad_norm": 6.146820068359375, "learning_rate": 9.18515983297967e-05, "loss": 0.8975, "step": 6703 }, { "epoch": 0.4542313164848567, "grad_norm": 6.619115352630615, "learning_rate": 9.185022931069888e-05, "loss": 0.9972, "step": 6704 }, { "epoch": 0.4542990717528288, "grad_norm": 5.762700080871582, "learning_rate": 9.184886029160106e-05, "loss": 0.6798, "step": 6705 }, { "epoch": 0.4543668270208009, "grad_norm": 7.270383358001709, "learning_rate": 9.184749127250326e-05, "loss": 0.8676, "step": 6706 }, { "epoch": 0.4544345822887729, "grad_norm": 6.667981147766113, "learning_rate": 9.184612225340544e-05, "loss": 0.7988, "step": 6707 }, { "epoch": 0.454502337556745, "grad_norm": 5.513166904449463, "learning_rate": 9.184475323430762e-05, "loss": 0.868, "step": 6708 }, { "epoch": 0.4545700928247171, "grad_norm": 6.376199245452881, "learning_rate": 9.18433842152098e-05, "loss": 0.7859, "step": 6709 }, { "epoch": 0.4546378480926892, "grad_norm": 6.444526672363281, "learning_rate": 9.184201519611198e-05, "loss": 0.7793, "step": 6710 }, { "epoch": 0.4547056033606613, "grad_norm": 8.83858871459961, "learning_rate": 9.184064617701417e-05, "loss": 0.9903, "step": 6711 }, { "epoch": 0.4547733586286334, "grad_norm": 6.07218599319458, "learning_rate": 9.183927715791635e-05, "loss": 0.8349, "step": 6712 }, { "epoch": 0.4548411138966055, "grad_norm": 6.652568340301514, "learning_rate": 9.183790813881853e-05, "loss": 0.8982, "step": 6713 }, { "epoch": 0.45490886916457757, "grad_norm": 7.582399845123291, "learning_rate": 9.183653911972071e-05, "loss": 0.9833, "step": 6714 }, { "epoch": 0.4549766244325496, "grad_norm": 8.411558151245117, "learning_rate": 9.183517010062291e-05, "loss": 0.9433, "step": 6715 }, { "epoch": 0.4550443797005217, "grad_norm": 7.420217037200928, "learning_rate": 9.183380108152509e-05, "loss": 0.9508, "step": 6716 }, { "epoch": 0.4551121349684938, "grad_norm": 7.064800262451172, "learning_rate": 9.183243206242727e-05, "loss": 0.9414, "step": 6717 }, { "epoch": 0.4551798902364659, "grad_norm": 8.102936744689941, "learning_rate": 9.183106304332946e-05, "loss": 0.9061, "step": 6718 }, { "epoch": 0.45524764550443797, "grad_norm": 8.261469841003418, "learning_rate": 9.182969402423164e-05, "loss": 1.2411, "step": 6719 }, { "epoch": 0.45531540077241006, "grad_norm": 7.5003204345703125, "learning_rate": 9.182832500513382e-05, "loss": 0.9296, "step": 6720 }, { "epoch": 0.45538315604038215, "grad_norm": 7.358273029327393, "learning_rate": 9.182695598603602e-05, "loss": 0.8109, "step": 6721 }, { "epoch": 0.45545091130835424, "grad_norm": 7.0917768478393555, "learning_rate": 9.18255869669382e-05, "loss": 0.9517, "step": 6722 }, { "epoch": 0.45551866657632634, "grad_norm": 6.559359550476074, "learning_rate": 9.182421794784038e-05, "loss": 0.8381, "step": 6723 }, { "epoch": 0.4555864218442984, "grad_norm": 6.368768215179443, "learning_rate": 9.182284892874256e-05, "loss": 0.9104, "step": 6724 }, { "epoch": 0.45565417711227046, "grad_norm": 6.576778411865234, "learning_rate": 9.182147990964475e-05, "loss": 0.9116, "step": 6725 }, { "epoch": 0.45572193238024256, "grad_norm": 6.608447551727295, "learning_rate": 9.182011089054693e-05, "loss": 0.9371, "step": 6726 }, { "epoch": 0.45578968764821465, "grad_norm": 6.38028621673584, "learning_rate": 9.181874187144911e-05, "loss": 0.9607, "step": 6727 }, { "epoch": 0.45585744291618674, "grad_norm": 8.349663734436035, "learning_rate": 9.18173728523513e-05, "loss": 0.8739, "step": 6728 }, { "epoch": 0.45592519818415883, "grad_norm": 7.229827880859375, "learning_rate": 9.181600383325349e-05, "loss": 0.6525, "step": 6729 }, { "epoch": 0.4559929534521309, "grad_norm": 7.870299816131592, "learning_rate": 9.181463481415567e-05, "loss": 0.9938, "step": 6730 }, { "epoch": 0.456060708720103, "grad_norm": 7.6586503982543945, "learning_rate": 9.181326579505785e-05, "loss": 0.9489, "step": 6731 }, { "epoch": 0.45612846398807505, "grad_norm": 8.91909122467041, "learning_rate": 9.181189677596003e-05, "loss": 0.9642, "step": 6732 }, { "epoch": 0.45619621925604714, "grad_norm": 8.341778755187988, "learning_rate": 9.181052775686221e-05, "loss": 1.0161, "step": 6733 }, { "epoch": 0.45626397452401923, "grad_norm": 7.269313812255859, "learning_rate": 9.18091587377644e-05, "loss": 0.8535, "step": 6734 }, { "epoch": 0.4563317297919913, "grad_norm": 6.635743141174316, "learning_rate": 9.180778971866658e-05, "loss": 0.6519, "step": 6735 }, { "epoch": 0.4563994850599634, "grad_norm": 7.213603973388672, "learning_rate": 9.180642069956876e-05, "loss": 0.8314, "step": 6736 }, { "epoch": 0.4564672403279355, "grad_norm": 6.491673469543457, "learning_rate": 9.180505168047094e-05, "loss": 0.8691, "step": 6737 }, { "epoch": 0.4565349955959076, "grad_norm": 6.087094783782959, "learning_rate": 9.180368266137314e-05, "loss": 0.8732, "step": 6738 }, { "epoch": 0.4566027508638797, "grad_norm": 6.0057172775268555, "learning_rate": 9.180231364227532e-05, "loss": 0.74, "step": 6739 }, { "epoch": 0.45667050613185173, "grad_norm": 6.74835729598999, "learning_rate": 9.18009446231775e-05, "loss": 0.7628, "step": 6740 }, { "epoch": 0.4567382613998238, "grad_norm": 5.433539867401123, "learning_rate": 9.179957560407968e-05, "loss": 0.6438, "step": 6741 }, { "epoch": 0.4568060166677959, "grad_norm": 7.143089771270752, "learning_rate": 9.179820658498186e-05, "loss": 0.807, "step": 6742 }, { "epoch": 0.456873771935768, "grad_norm": 6.345632076263428, "learning_rate": 9.179683756588405e-05, "loss": 0.8652, "step": 6743 }, { "epoch": 0.4569415272037401, "grad_norm": 6.4180908203125, "learning_rate": 9.179546854678623e-05, "loss": 0.8225, "step": 6744 }, { "epoch": 0.4570092824717122, "grad_norm": 7.375923156738281, "learning_rate": 9.179409952768841e-05, "loss": 0.7922, "step": 6745 }, { "epoch": 0.4570770377396843, "grad_norm": 5.134999752044678, "learning_rate": 9.17927305085906e-05, "loss": 0.8943, "step": 6746 }, { "epoch": 0.45714479300765637, "grad_norm": 8.040843963623047, "learning_rate": 9.179136148949277e-05, "loss": 0.9971, "step": 6747 }, { "epoch": 0.4572125482756284, "grad_norm": 5.302629470825195, "learning_rate": 9.178999247039497e-05, "loss": 0.6811, "step": 6748 }, { "epoch": 0.4572803035436005, "grad_norm": 6.630457878112793, "learning_rate": 9.178862345129715e-05, "loss": 0.7929, "step": 6749 }, { "epoch": 0.4573480588115726, "grad_norm": 5.959926605224609, "learning_rate": 9.178725443219933e-05, "loss": 0.7941, "step": 6750 }, { "epoch": 0.4574158140795447, "grad_norm": 7.855923175811768, "learning_rate": 9.178588541310151e-05, "loss": 0.911, "step": 6751 }, { "epoch": 0.45748356934751677, "grad_norm": 7.775002479553223, "learning_rate": 9.17845163940037e-05, "loss": 0.8315, "step": 6752 }, { "epoch": 0.45755132461548886, "grad_norm": 7.259592056274414, "learning_rate": 9.178314737490588e-05, "loss": 0.8383, "step": 6753 }, { "epoch": 0.45761907988346096, "grad_norm": 8.843462944030762, "learning_rate": 9.178177835580806e-05, "loss": 1.1005, "step": 6754 }, { "epoch": 0.45768683515143305, "grad_norm": 6.322036266326904, "learning_rate": 9.178040933671024e-05, "loss": 0.8553, "step": 6755 }, { "epoch": 0.4577545904194051, "grad_norm": 6.644796848297119, "learning_rate": 9.177904031761242e-05, "loss": 1.0933, "step": 6756 }, { "epoch": 0.4578223456873772, "grad_norm": 7.463156700134277, "learning_rate": 9.177767129851462e-05, "loss": 0.9584, "step": 6757 }, { "epoch": 0.45789010095534927, "grad_norm": 5.70056676864624, "learning_rate": 9.17763022794168e-05, "loss": 0.6945, "step": 6758 }, { "epoch": 0.45795785622332136, "grad_norm": 6.347555637359619, "learning_rate": 9.177493326031898e-05, "loss": 0.9143, "step": 6759 }, { "epoch": 0.45802561149129345, "grad_norm": 6.00151252746582, "learning_rate": 9.177356424122116e-05, "loss": 0.8139, "step": 6760 }, { "epoch": 0.45809336675926554, "grad_norm": 8.02507495880127, "learning_rate": 9.177219522212335e-05, "loss": 0.8941, "step": 6761 }, { "epoch": 0.45816112202723763, "grad_norm": 7.0266876220703125, "learning_rate": 9.177082620302553e-05, "loss": 1.0009, "step": 6762 }, { "epoch": 0.4582288772952097, "grad_norm": 5.483405113220215, "learning_rate": 9.176945718392771e-05, "loss": 0.8113, "step": 6763 }, { "epoch": 0.45829663256318176, "grad_norm": 5.661553382873535, "learning_rate": 9.176808816482991e-05, "loss": 0.7524, "step": 6764 }, { "epoch": 0.45836438783115385, "grad_norm": 7.485569953918457, "learning_rate": 9.176671914573209e-05, "loss": 0.8712, "step": 6765 }, { "epoch": 0.45843214309912594, "grad_norm": 8.262595176696777, "learning_rate": 9.176535012663427e-05, "loss": 0.7777, "step": 6766 }, { "epoch": 0.45849989836709804, "grad_norm": 6.955083847045898, "learning_rate": 9.176398110753646e-05, "loss": 0.8515, "step": 6767 }, { "epoch": 0.4585676536350701, "grad_norm": 5.7011871337890625, "learning_rate": 9.176261208843864e-05, "loss": 0.7977, "step": 6768 }, { "epoch": 0.4586354089030422, "grad_norm": 7.344122886657715, "learning_rate": 9.176124306934082e-05, "loss": 0.877, "step": 6769 }, { "epoch": 0.4587031641710143, "grad_norm": 6.595543384552002, "learning_rate": 9.1759874050243e-05, "loss": 0.7739, "step": 6770 }, { "epoch": 0.4587709194389864, "grad_norm": 6.087022304534912, "learning_rate": 9.17585050311452e-05, "loss": 0.9127, "step": 6771 }, { "epoch": 0.4588386747069585, "grad_norm": 6.13311243057251, "learning_rate": 9.175713601204738e-05, "loss": 0.8803, "step": 6772 }, { "epoch": 0.45890642997493053, "grad_norm": 5.866177558898926, "learning_rate": 9.175576699294956e-05, "loss": 0.9859, "step": 6773 }, { "epoch": 0.4589741852429026, "grad_norm": 7.406826972961426, "learning_rate": 9.175439797385174e-05, "loss": 1.2433, "step": 6774 }, { "epoch": 0.4590419405108747, "grad_norm": 5.886981964111328, "learning_rate": 9.175302895475393e-05, "loss": 0.9994, "step": 6775 }, { "epoch": 0.4591096957788468, "grad_norm": 6.240331172943115, "learning_rate": 9.175165993565611e-05, "loss": 0.6988, "step": 6776 }, { "epoch": 0.4591774510468189, "grad_norm": 6.331019878387451, "learning_rate": 9.175029091655829e-05, "loss": 0.9625, "step": 6777 }, { "epoch": 0.459245206314791, "grad_norm": 6.082772731781006, "learning_rate": 9.174892189746047e-05, "loss": 0.6876, "step": 6778 }, { "epoch": 0.4593129615827631, "grad_norm": 9.173615455627441, "learning_rate": 9.174755287836265e-05, "loss": 0.9572, "step": 6779 }, { "epoch": 0.45938071685073517, "grad_norm": 7.658091068267822, "learning_rate": 9.174618385926485e-05, "loss": 1.0724, "step": 6780 }, { "epoch": 0.4594484721187072, "grad_norm": 5.414113521575928, "learning_rate": 9.174481484016703e-05, "loss": 0.963, "step": 6781 }, { "epoch": 0.4595162273866793, "grad_norm": 8.290900230407715, "learning_rate": 9.174344582106921e-05, "loss": 0.8012, "step": 6782 }, { "epoch": 0.4595839826546514, "grad_norm": 5.778069019317627, "learning_rate": 9.174207680197139e-05, "loss": 0.9199, "step": 6783 }, { "epoch": 0.4596517379226235, "grad_norm": 7.42002010345459, "learning_rate": 9.174070778287358e-05, "loss": 1.022, "step": 6784 }, { "epoch": 0.4597194931905956, "grad_norm": 5.714671611785889, "learning_rate": 9.173933876377576e-05, "loss": 0.8194, "step": 6785 }, { "epoch": 0.45978724845856767, "grad_norm": 6.420266628265381, "learning_rate": 9.173796974467794e-05, "loss": 0.8879, "step": 6786 }, { "epoch": 0.45985500372653976, "grad_norm": 7.046072006225586, "learning_rate": 9.173660072558012e-05, "loss": 0.7672, "step": 6787 }, { "epoch": 0.45992275899451185, "grad_norm": 8.797179222106934, "learning_rate": 9.17352317064823e-05, "loss": 0.7714, "step": 6788 }, { "epoch": 0.4599905142624839, "grad_norm": 6.534037113189697, "learning_rate": 9.17338626873845e-05, "loss": 0.8629, "step": 6789 }, { "epoch": 0.460058269530456, "grad_norm": 7.647995948791504, "learning_rate": 9.173249366828668e-05, "loss": 0.8143, "step": 6790 }, { "epoch": 0.46012602479842807, "grad_norm": 5.4971418380737305, "learning_rate": 9.173112464918886e-05, "loss": 0.7946, "step": 6791 }, { "epoch": 0.46019378006640016, "grad_norm": 6.304740905761719, "learning_rate": 9.172975563009104e-05, "loss": 0.9233, "step": 6792 }, { "epoch": 0.46026153533437225, "grad_norm": 7.571120262145996, "learning_rate": 9.172838661099323e-05, "loss": 0.9919, "step": 6793 }, { "epoch": 0.46032929060234434, "grad_norm": 6.336854457855225, "learning_rate": 9.172701759189541e-05, "loss": 0.7232, "step": 6794 }, { "epoch": 0.46039704587031643, "grad_norm": 6.926676273345947, "learning_rate": 9.17256485727976e-05, "loss": 0.758, "step": 6795 }, { "epoch": 0.4604648011382885, "grad_norm": 5.480643272399902, "learning_rate": 9.172427955369977e-05, "loss": 0.5871, "step": 6796 }, { "epoch": 0.46053255640626056, "grad_norm": 8.286799430847168, "learning_rate": 9.172291053460195e-05, "loss": 0.8574, "step": 6797 }, { "epoch": 0.46060031167423265, "grad_norm": 8.043119430541992, "learning_rate": 9.172154151550415e-05, "loss": 0.6363, "step": 6798 }, { "epoch": 0.46066806694220475, "grad_norm": 8.308424949645996, "learning_rate": 9.172017249640633e-05, "loss": 0.9203, "step": 6799 }, { "epoch": 0.46073582221017684, "grad_norm": 7.131639003753662, "learning_rate": 9.171880347730851e-05, "loss": 0.685, "step": 6800 }, { "epoch": 0.46080357747814893, "grad_norm": 8.397472381591797, "learning_rate": 9.171743445821069e-05, "loss": 0.9637, "step": 6801 }, { "epoch": 0.460871332746121, "grad_norm": 6.666365146636963, "learning_rate": 9.171606543911287e-05, "loss": 1.0735, "step": 6802 }, { "epoch": 0.4609390880140931, "grad_norm": 11.102825164794922, "learning_rate": 9.171469642001506e-05, "loss": 0.8958, "step": 6803 }, { "epoch": 0.4610068432820652, "grad_norm": 7.294391632080078, "learning_rate": 9.171332740091724e-05, "loss": 1.0354, "step": 6804 }, { "epoch": 0.46107459855003724, "grad_norm": 5.9984517097473145, "learning_rate": 9.171195838181942e-05, "loss": 0.605, "step": 6805 }, { "epoch": 0.46114235381800933, "grad_norm": 6.209224224090576, "learning_rate": 9.17105893627216e-05, "loss": 0.8715, "step": 6806 }, { "epoch": 0.4612101090859814, "grad_norm": 5.404207229614258, "learning_rate": 9.17092203436238e-05, "loss": 0.6694, "step": 6807 }, { "epoch": 0.4612778643539535, "grad_norm": 6.675217151641846, "learning_rate": 9.170785132452598e-05, "loss": 0.7288, "step": 6808 }, { "epoch": 0.4613456196219256, "grad_norm": 8.444995880126953, "learning_rate": 9.170648230542816e-05, "loss": 1.0756, "step": 6809 }, { "epoch": 0.4614133748898977, "grad_norm": 8.167703628540039, "learning_rate": 9.170511328633035e-05, "loss": 1.1605, "step": 6810 }, { "epoch": 0.4614811301578698, "grad_norm": 7.55181360244751, "learning_rate": 9.170374426723253e-05, "loss": 0.7805, "step": 6811 }, { "epoch": 0.4615488854258419, "grad_norm": 6.380015850067139, "learning_rate": 9.170237524813471e-05, "loss": 0.6584, "step": 6812 }, { "epoch": 0.4616166406938139, "grad_norm": 6.2454376220703125, "learning_rate": 9.170100622903691e-05, "loss": 0.8212, "step": 6813 }, { "epoch": 0.461684395961786, "grad_norm": 7.530882358551025, "learning_rate": 9.169963720993909e-05, "loss": 0.8926, "step": 6814 }, { "epoch": 0.4617521512297581, "grad_norm": 6.3245158195495605, "learning_rate": 9.169826819084127e-05, "loss": 0.8672, "step": 6815 }, { "epoch": 0.4618199064977302, "grad_norm": 5.822272300720215, "learning_rate": 9.169689917174346e-05, "loss": 0.7972, "step": 6816 }, { "epoch": 0.4618876617657023, "grad_norm": 8.160684585571289, "learning_rate": 9.169553015264564e-05, "loss": 1.0076, "step": 6817 }, { "epoch": 0.4619554170336744, "grad_norm": 6.745192050933838, "learning_rate": 9.169416113354782e-05, "loss": 0.8187, "step": 6818 }, { "epoch": 0.46202317230164647, "grad_norm": 8.530631065368652, "learning_rate": 9.169279211445e-05, "loss": 0.7655, "step": 6819 }, { "epoch": 0.46209092756961856, "grad_norm": 7.213817596435547, "learning_rate": 9.169142309535218e-05, "loss": 0.8737, "step": 6820 }, { "epoch": 0.4621586828375906, "grad_norm": 6.5060577392578125, "learning_rate": 9.169005407625438e-05, "loss": 0.815, "step": 6821 }, { "epoch": 0.4622264381055627, "grad_norm": 6.782070636749268, "learning_rate": 9.168868505715656e-05, "loss": 0.792, "step": 6822 }, { "epoch": 0.4622941933735348, "grad_norm": 6.130987644195557, "learning_rate": 9.168731603805874e-05, "loss": 0.6743, "step": 6823 }, { "epoch": 0.46236194864150687, "grad_norm": 8.178153991699219, "learning_rate": 9.168594701896092e-05, "loss": 1.0958, "step": 6824 }, { "epoch": 0.46242970390947896, "grad_norm": 8.360578536987305, "learning_rate": 9.16845779998631e-05, "loss": 0.9709, "step": 6825 }, { "epoch": 0.46249745917745105, "grad_norm": 7.507253170013428, "learning_rate": 9.168320898076529e-05, "loss": 0.9384, "step": 6826 }, { "epoch": 0.46256521444542315, "grad_norm": 8.285658836364746, "learning_rate": 9.168183996166747e-05, "loss": 0.8913, "step": 6827 }, { "epoch": 0.46263296971339524, "grad_norm": 7.765392780303955, "learning_rate": 9.168047094256965e-05, "loss": 0.8754, "step": 6828 }, { "epoch": 0.46270072498136733, "grad_norm": 7.6219892501831055, "learning_rate": 9.167910192347183e-05, "loss": 0.7401, "step": 6829 }, { "epoch": 0.46276848024933936, "grad_norm": 9.139801979064941, "learning_rate": 9.167773290437403e-05, "loss": 0.8068, "step": 6830 }, { "epoch": 0.46283623551731146, "grad_norm": 7.655275821685791, "learning_rate": 9.167636388527621e-05, "loss": 0.8314, "step": 6831 }, { "epoch": 0.46290399078528355, "grad_norm": 6.699294090270996, "learning_rate": 9.167499486617839e-05, "loss": 0.8665, "step": 6832 }, { "epoch": 0.46297174605325564, "grad_norm": 8.436169624328613, "learning_rate": 9.167362584708057e-05, "loss": 1.3026, "step": 6833 }, { "epoch": 0.46303950132122773, "grad_norm": 5.195062160491943, "learning_rate": 9.167225682798275e-05, "loss": 0.7468, "step": 6834 }, { "epoch": 0.4631072565891998, "grad_norm": 5.665080547332764, "learning_rate": 9.167088780888494e-05, "loss": 0.7357, "step": 6835 }, { "epoch": 0.4631750118571719, "grad_norm": 6.243655681610107, "learning_rate": 9.166951878978712e-05, "loss": 0.7705, "step": 6836 }, { "epoch": 0.463242767125144, "grad_norm": 7.49260950088501, "learning_rate": 9.16681497706893e-05, "loss": 0.8061, "step": 6837 }, { "epoch": 0.46331052239311604, "grad_norm": 7.076335430145264, "learning_rate": 9.166678075159148e-05, "loss": 0.8307, "step": 6838 }, { "epoch": 0.46337827766108813, "grad_norm": 6.332518100738525, "learning_rate": 9.166541173249368e-05, "loss": 0.7431, "step": 6839 }, { "epoch": 0.4634460329290602, "grad_norm": 6.261449813842773, "learning_rate": 9.166404271339586e-05, "loss": 0.8817, "step": 6840 }, { "epoch": 0.4635137881970323, "grad_norm": 7.467769622802734, "learning_rate": 9.166267369429804e-05, "loss": 0.9473, "step": 6841 }, { "epoch": 0.4635815434650044, "grad_norm": 5.55718469619751, "learning_rate": 9.166130467520022e-05, "loss": 0.7726, "step": 6842 }, { "epoch": 0.4636492987329765, "grad_norm": 7.012959003448486, "learning_rate": 9.16599356561024e-05, "loss": 0.9274, "step": 6843 }, { "epoch": 0.4637170540009486, "grad_norm": 6.344860076904297, "learning_rate": 9.165856663700459e-05, "loss": 1.0617, "step": 6844 }, { "epoch": 0.4637848092689207, "grad_norm": 6.024655342102051, "learning_rate": 9.165719761790677e-05, "loss": 0.8224, "step": 6845 }, { "epoch": 0.4638525645368927, "grad_norm": 6.97758674621582, "learning_rate": 9.165582859880895e-05, "loss": 0.8344, "step": 6846 }, { "epoch": 0.4639203198048648, "grad_norm": 6.874304294586182, "learning_rate": 9.165445957971113e-05, "loss": 0.8698, "step": 6847 }, { "epoch": 0.4639880750728369, "grad_norm": 8.15816593170166, "learning_rate": 9.165309056061331e-05, "loss": 1.0457, "step": 6848 }, { "epoch": 0.464055830340809, "grad_norm": 7.304631233215332, "learning_rate": 9.165172154151551e-05, "loss": 0.9027, "step": 6849 }, { "epoch": 0.4641235856087811, "grad_norm": 7.377762317657471, "learning_rate": 9.165035252241769e-05, "loss": 0.6635, "step": 6850 }, { "epoch": 0.4641913408767532, "grad_norm": 6.051864147186279, "learning_rate": 9.164898350331987e-05, "loss": 0.925, "step": 6851 }, { "epoch": 0.46425909614472527, "grad_norm": 6.560476779937744, "learning_rate": 9.164761448422205e-05, "loss": 0.7674, "step": 6852 }, { "epoch": 0.46432685141269736, "grad_norm": 6.828582763671875, "learning_rate": 9.164624546512424e-05, "loss": 0.8159, "step": 6853 }, { "epoch": 0.4643946066806694, "grad_norm": 6.463281631469727, "learning_rate": 9.164487644602642e-05, "loss": 0.6261, "step": 6854 }, { "epoch": 0.4644623619486415, "grad_norm": 6.3371734619140625, "learning_rate": 9.16435074269286e-05, "loss": 0.6489, "step": 6855 }, { "epoch": 0.4645301172166136, "grad_norm": 6.052369117736816, "learning_rate": 9.16421384078308e-05, "loss": 0.6899, "step": 6856 }, { "epoch": 0.46459787248458567, "grad_norm": 6.464377403259277, "learning_rate": 9.164076938873298e-05, "loss": 0.7603, "step": 6857 }, { "epoch": 0.46466562775255776, "grad_norm": 5.9912567138671875, "learning_rate": 9.163940036963516e-05, "loss": 0.7188, "step": 6858 }, { "epoch": 0.46473338302052986, "grad_norm": 8.469727516174316, "learning_rate": 9.163803135053735e-05, "loss": 0.9991, "step": 6859 }, { "epoch": 0.46480113828850195, "grad_norm": 6.261715888977051, "learning_rate": 9.163666233143953e-05, "loss": 0.7429, "step": 6860 }, { "epoch": 0.46486889355647404, "grad_norm": 7.389404296875, "learning_rate": 9.163529331234171e-05, "loss": 0.6294, "step": 6861 }, { "epoch": 0.4649366488244461, "grad_norm": 6.988171100616455, "learning_rate": 9.16339242932439e-05, "loss": 0.8705, "step": 6862 }, { "epoch": 0.46500440409241817, "grad_norm": 6.285641193389893, "learning_rate": 9.163255527414609e-05, "loss": 0.7026, "step": 6863 }, { "epoch": 0.46507215936039026, "grad_norm": 8.77840518951416, "learning_rate": 9.163118625504827e-05, "loss": 0.8766, "step": 6864 }, { "epoch": 0.46513991462836235, "grad_norm": 6.199909210205078, "learning_rate": 9.162981723595045e-05, "loss": 0.7304, "step": 6865 }, { "epoch": 0.46520766989633444, "grad_norm": 6.947317600250244, "learning_rate": 9.162844821685263e-05, "loss": 0.8649, "step": 6866 }, { "epoch": 0.46527542516430653, "grad_norm": 8.396434783935547, "learning_rate": 9.162707919775482e-05, "loss": 1.2161, "step": 6867 }, { "epoch": 0.4653431804322786, "grad_norm": 5.935467720031738, "learning_rate": 9.1625710178657e-05, "loss": 0.8499, "step": 6868 }, { "epoch": 0.4654109357002507, "grad_norm": 6.258296966552734, "learning_rate": 9.162434115955918e-05, "loss": 0.9541, "step": 6869 }, { "epoch": 0.46547869096822275, "grad_norm": 6.29650354385376, "learning_rate": 9.162297214046136e-05, "loss": 0.7774, "step": 6870 }, { "epoch": 0.46554644623619484, "grad_norm": 5.948836803436279, "learning_rate": 9.162160312136356e-05, "loss": 0.7579, "step": 6871 }, { "epoch": 0.46561420150416694, "grad_norm": 5.719659805297852, "learning_rate": 9.162023410226574e-05, "loss": 0.7482, "step": 6872 }, { "epoch": 0.465681956772139, "grad_norm": 6.145468235015869, "learning_rate": 9.161886508316792e-05, "loss": 0.8816, "step": 6873 }, { "epoch": 0.4657497120401111, "grad_norm": 6.33701753616333, "learning_rate": 9.16174960640701e-05, "loss": 0.7617, "step": 6874 }, { "epoch": 0.4658174673080832, "grad_norm": 6.143844127655029, "learning_rate": 9.161612704497228e-05, "loss": 0.9695, "step": 6875 }, { "epoch": 0.4658852225760553, "grad_norm": 6.316319942474365, "learning_rate": 9.161475802587447e-05, "loss": 0.8837, "step": 6876 }, { "epoch": 0.4659529778440274, "grad_norm": 6.68782377243042, "learning_rate": 9.161338900677665e-05, "loss": 0.7324, "step": 6877 }, { "epoch": 0.4660207331119995, "grad_norm": 6.216282844543457, "learning_rate": 9.161201998767883e-05, "loss": 0.7787, "step": 6878 }, { "epoch": 0.4660884883799715, "grad_norm": 6.8673176765441895, "learning_rate": 9.161065096858101e-05, "loss": 1.1118, "step": 6879 }, { "epoch": 0.4661562436479436, "grad_norm": 6.71323823928833, "learning_rate": 9.16092819494832e-05, "loss": 0.9176, "step": 6880 }, { "epoch": 0.4662239989159157, "grad_norm": 8.589679718017578, "learning_rate": 9.160791293038539e-05, "loss": 0.9758, "step": 6881 }, { "epoch": 0.4662917541838878, "grad_norm": 6.692760467529297, "learning_rate": 9.160654391128757e-05, "loss": 0.8867, "step": 6882 }, { "epoch": 0.4663595094518599, "grad_norm": 8.462479591369629, "learning_rate": 9.160517489218975e-05, "loss": 0.9738, "step": 6883 }, { "epoch": 0.466427264719832, "grad_norm": 6.068343162536621, "learning_rate": 9.160380587309193e-05, "loss": 0.6652, "step": 6884 }, { "epoch": 0.46649501998780407, "grad_norm": 6.04793643951416, "learning_rate": 9.160243685399412e-05, "loss": 0.8575, "step": 6885 }, { "epoch": 0.46656277525577616, "grad_norm": 7.8360514640808105, "learning_rate": 9.16010678348963e-05, "loss": 0.8488, "step": 6886 }, { "epoch": 0.4666305305237482, "grad_norm": 6.028532981872559, "learning_rate": 9.159969881579848e-05, "loss": 0.8841, "step": 6887 }, { "epoch": 0.4666982857917203, "grad_norm": 7.673785209655762, "learning_rate": 9.159832979670066e-05, "loss": 0.9035, "step": 6888 }, { "epoch": 0.4667660410596924, "grad_norm": 7.042590618133545, "learning_rate": 9.159696077760284e-05, "loss": 0.8812, "step": 6889 }, { "epoch": 0.4668337963276645, "grad_norm": 6.562109470367432, "learning_rate": 9.159559175850504e-05, "loss": 0.7179, "step": 6890 }, { "epoch": 0.46690155159563657, "grad_norm": 6.083657264709473, "learning_rate": 9.159422273940722e-05, "loss": 0.7322, "step": 6891 }, { "epoch": 0.46696930686360866, "grad_norm": 7.408811569213867, "learning_rate": 9.15928537203094e-05, "loss": 0.9291, "step": 6892 }, { "epoch": 0.46703706213158075, "grad_norm": 6.70425271987915, "learning_rate": 9.159148470121158e-05, "loss": 0.9796, "step": 6893 }, { "epoch": 0.46710481739955284, "grad_norm": 7.486738681793213, "learning_rate": 9.159011568211377e-05, "loss": 1.128, "step": 6894 }, { "epoch": 0.4671725726675249, "grad_norm": 6.747304916381836, "learning_rate": 9.158874666301595e-05, "loss": 0.8432, "step": 6895 }, { "epoch": 0.46724032793549697, "grad_norm": 7.494656085968018, "learning_rate": 9.158737764391813e-05, "loss": 0.9587, "step": 6896 }, { "epoch": 0.46730808320346906, "grad_norm": 5.556826114654541, "learning_rate": 9.158600862482031e-05, "loss": 0.6565, "step": 6897 }, { "epoch": 0.46737583847144115, "grad_norm": 6.550345420837402, "learning_rate": 9.15846396057225e-05, "loss": 0.7854, "step": 6898 }, { "epoch": 0.46744359373941324, "grad_norm": 7.0627899169921875, "learning_rate": 9.158327058662469e-05, "loss": 0.9204, "step": 6899 }, { "epoch": 0.46751134900738534, "grad_norm": 7.1070661544799805, "learning_rate": 9.158190156752687e-05, "loss": 0.9357, "step": 6900 }, { "epoch": 0.4675791042753574, "grad_norm": 7.313487529754639, "learning_rate": 9.158053254842905e-05, "loss": 0.7561, "step": 6901 }, { "epoch": 0.4676468595433295, "grad_norm": 6.839418888092041, "learning_rate": 9.157916352933123e-05, "loss": 0.9551, "step": 6902 }, { "epoch": 0.46771461481130155, "grad_norm": 6.881319999694824, "learning_rate": 9.157779451023342e-05, "loss": 0.8799, "step": 6903 }, { "epoch": 0.46778237007927365, "grad_norm": 7.470109462738037, "learning_rate": 9.15764254911356e-05, "loss": 0.806, "step": 6904 }, { "epoch": 0.46785012534724574, "grad_norm": 7.056912899017334, "learning_rate": 9.157505647203778e-05, "loss": 0.7905, "step": 6905 }, { "epoch": 0.46791788061521783, "grad_norm": 6.295861721038818, "learning_rate": 9.157368745293998e-05, "loss": 0.8256, "step": 6906 }, { "epoch": 0.4679856358831899, "grad_norm": 5.80570650100708, "learning_rate": 9.157231843384216e-05, "loss": 0.7747, "step": 6907 }, { "epoch": 0.468053391151162, "grad_norm": 7.0094218254089355, "learning_rate": 9.157094941474435e-05, "loss": 0.9504, "step": 6908 }, { "epoch": 0.4681211464191341, "grad_norm": 5.334716320037842, "learning_rate": 9.156958039564653e-05, "loss": 0.7608, "step": 6909 }, { "epoch": 0.4681889016871062, "grad_norm": 8.058252334594727, "learning_rate": 9.156821137654871e-05, "loss": 0.889, "step": 6910 }, { "epoch": 0.46825665695507823, "grad_norm": 7.769287586212158, "learning_rate": 9.156684235745089e-05, "loss": 1.0391, "step": 6911 }, { "epoch": 0.4683244122230503, "grad_norm": 5.96143102645874, "learning_rate": 9.156547333835307e-05, "loss": 0.8793, "step": 6912 }, { "epoch": 0.4683921674910224, "grad_norm": 5.6451005935668945, "learning_rate": 9.156410431925527e-05, "loss": 0.7804, "step": 6913 }, { "epoch": 0.4684599227589945, "grad_norm": 7.539672374725342, "learning_rate": 9.156273530015745e-05, "loss": 0.8137, "step": 6914 }, { "epoch": 0.4685276780269666, "grad_norm": 7.447227954864502, "learning_rate": 9.156136628105963e-05, "loss": 1.0692, "step": 6915 }, { "epoch": 0.4685954332949387, "grad_norm": 8.100126266479492, "learning_rate": 9.155999726196181e-05, "loss": 0.8124, "step": 6916 }, { "epoch": 0.4686631885629108, "grad_norm": 8.560744285583496, "learning_rate": 9.1558628242864e-05, "loss": 0.8659, "step": 6917 }, { "epoch": 0.4687309438308829, "grad_norm": 7.135112285614014, "learning_rate": 9.155725922376618e-05, "loss": 0.9009, "step": 6918 }, { "epoch": 0.4687986990988549, "grad_norm": 7.749111175537109, "learning_rate": 9.155589020466836e-05, "loss": 0.9209, "step": 6919 }, { "epoch": 0.468866454366827, "grad_norm": 6.616466999053955, "learning_rate": 9.155452118557054e-05, "loss": 0.9283, "step": 6920 }, { "epoch": 0.4689342096347991, "grad_norm": 6.961619853973389, "learning_rate": 9.155315216647272e-05, "loss": 0.8142, "step": 6921 }, { "epoch": 0.4690019649027712, "grad_norm": 7.225759506225586, "learning_rate": 9.155178314737492e-05, "loss": 0.7626, "step": 6922 }, { "epoch": 0.4690697201707433, "grad_norm": 6.0465922355651855, "learning_rate": 9.15504141282771e-05, "loss": 0.8281, "step": 6923 }, { "epoch": 0.46913747543871537, "grad_norm": 7.507081985473633, "learning_rate": 9.154904510917928e-05, "loss": 1.1642, "step": 6924 }, { "epoch": 0.46920523070668746, "grad_norm": 6.281520843505859, "learning_rate": 9.154767609008146e-05, "loss": 0.8694, "step": 6925 }, { "epoch": 0.46927298597465955, "grad_norm": 6.0980658531188965, "learning_rate": 9.154630707098365e-05, "loss": 0.6983, "step": 6926 }, { "epoch": 0.4693407412426316, "grad_norm": 5.842291355133057, "learning_rate": 9.154493805188583e-05, "loss": 0.8392, "step": 6927 }, { "epoch": 0.4694084965106037, "grad_norm": 7.486947536468506, "learning_rate": 9.154356903278801e-05, "loss": 0.8363, "step": 6928 }, { "epoch": 0.46947625177857577, "grad_norm": 7.279699802398682, "learning_rate": 9.154220001369019e-05, "loss": 0.9697, "step": 6929 }, { "epoch": 0.46954400704654786, "grad_norm": 6.720832347869873, "learning_rate": 9.154083099459237e-05, "loss": 1.0377, "step": 6930 }, { "epoch": 0.46961176231451995, "grad_norm": 9.792084693908691, "learning_rate": 9.153946197549457e-05, "loss": 1.1697, "step": 6931 }, { "epoch": 0.46967951758249205, "grad_norm": 6.329649925231934, "learning_rate": 9.153809295639675e-05, "loss": 0.8825, "step": 6932 }, { "epoch": 0.46974727285046414, "grad_norm": 5.9268269538879395, "learning_rate": 9.153672393729893e-05, "loss": 0.9157, "step": 6933 }, { "epoch": 0.46981502811843623, "grad_norm": 7.7385430335998535, "learning_rate": 9.153535491820111e-05, "loss": 0.7662, "step": 6934 }, { "epoch": 0.4698827833864083, "grad_norm": 6.896132946014404, "learning_rate": 9.153398589910329e-05, "loss": 1.004, "step": 6935 }, { "epoch": 0.46995053865438036, "grad_norm": 6.822011470794678, "learning_rate": 9.153261688000548e-05, "loss": 1.0163, "step": 6936 }, { "epoch": 0.47001829392235245, "grad_norm": 6.4482574462890625, "learning_rate": 9.153124786090766e-05, "loss": 0.8293, "step": 6937 }, { "epoch": 0.47008604919032454, "grad_norm": 6.855703353881836, "learning_rate": 9.152987884180984e-05, "loss": 0.9737, "step": 6938 }, { "epoch": 0.47015380445829663, "grad_norm": 6.508902549743652, "learning_rate": 9.152850982271202e-05, "loss": 1.0892, "step": 6939 }, { "epoch": 0.4702215597262687, "grad_norm": 5.452862739562988, "learning_rate": 9.152714080361422e-05, "loss": 0.803, "step": 6940 }, { "epoch": 0.4702893149942408, "grad_norm": 5.583015441894531, "learning_rate": 9.15257717845164e-05, "loss": 0.6624, "step": 6941 }, { "epoch": 0.4703570702622129, "grad_norm": 7.870943069458008, "learning_rate": 9.152440276541858e-05, "loss": 0.929, "step": 6942 }, { "epoch": 0.470424825530185, "grad_norm": 7.723261833190918, "learning_rate": 9.152303374632076e-05, "loss": 0.8467, "step": 6943 }, { "epoch": 0.47049258079815703, "grad_norm": 8.4433012008667, "learning_rate": 9.152166472722294e-05, "loss": 0.9935, "step": 6944 }, { "epoch": 0.4705603360661291, "grad_norm": 7.142673015594482, "learning_rate": 9.152029570812513e-05, "loss": 0.9821, "step": 6945 }, { "epoch": 0.4706280913341012, "grad_norm": 6.800427436828613, "learning_rate": 9.151892668902731e-05, "loss": 0.7329, "step": 6946 }, { "epoch": 0.4706958466020733, "grad_norm": 6.462594985961914, "learning_rate": 9.151755766992949e-05, "loss": 0.6495, "step": 6947 }, { "epoch": 0.4707636018700454, "grad_norm": 6.2622294425964355, "learning_rate": 9.151618865083167e-05, "loss": 1.0438, "step": 6948 }, { "epoch": 0.4708313571380175, "grad_norm": 7.76660680770874, "learning_rate": 9.151481963173387e-05, "loss": 0.8855, "step": 6949 }, { "epoch": 0.4708991124059896, "grad_norm": 4.535599231719971, "learning_rate": 9.151345061263605e-05, "loss": 0.6823, "step": 6950 }, { "epoch": 0.4709668676739617, "grad_norm": 5.862152099609375, "learning_rate": 9.151208159353823e-05, "loss": 0.7322, "step": 6951 }, { "epoch": 0.4710346229419337, "grad_norm": 6.145107746124268, "learning_rate": 9.151071257444042e-05, "loss": 0.8258, "step": 6952 }, { "epoch": 0.4711023782099058, "grad_norm": 5.889834880828857, "learning_rate": 9.15093435553426e-05, "loss": 0.8009, "step": 6953 }, { "epoch": 0.4711701334778779, "grad_norm": 8.594234466552734, "learning_rate": 9.150797453624478e-05, "loss": 0.8349, "step": 6954 }, { "epoch": 0.47123788874585, "grad_norm": 7.7569379806518555, "learning_rate": 9.150660551714698e-05, "loss": 0.6711, "step": 6955 }, { "epoch": 0.4713056440138221, "grad_norm": 7.4088544845581055, "learning_rate": 9.150523649804916e-05, "loss": 0.8936, "step": 6956 }, { "epoch": 0.47137339928179417, "grad_norm": 6.610263347625732, "learning_rate": 9.150386747895134e-05, "loss": 0.9786, "step": 6957 }, { "epoch": 0.47144115454976626, "grad_norm": 6.46956729888916, "learning_rate": 9.150249845985352e-05, "loss": 0.6781, "step": 6958 }, { "epoch": 0.47150890981773835, "grad_norm": 6.152948379516602, "learning_rate": 9.150112944075571e-05, "loss": 0.8375, "step": 6959 }, { "epoch": 0.4715766650857104, "grad_norm": 6.96013879776001, "learning_rate": 9.149976042165789e-05, "loss": 0.8629, "step": 6960 }, { "epoch": 0.4716444203536825, "grad_norm": 6.767139434814453, "learning_rate": 9.149839140256007e-05, "loss": 0.8043, "step": 6961 }, { "epoch": 0.4717121756216546, "grad_norm": 7.95654296875, "learning_rate": 9.149702238346225e-05, "loss": 1.0179, "step": 6962 }, { "epoch": 0.47177993088962666, "grad_norm": 8.155875205993652, "learning_rate": 9.149565336436445e-05, "loss": 1.0663, "step": 6963 }, { "epoch": 0.47184768615759876, "grad_norm": 7.903263092041016, "learning_rate": 9.149428434526663e-05, "loss": 1.0214, "step": 6964 }, { "epoch": 0.47191544142557085, "grad_norm": 7.510760307312012, "learning_rate": 9.149291532616881e-05, "loss": 0.9942, "step": 6965 }, { "epoch": 0.47198319669354294, "grad_norm": 5.814423084259033, "learning_rate": 9.149154630707099e-05, "loss": 0.7677, "step": 6966 }, { "epoch": 0.47205095196151503, "grad_norm": 8.29617977142334, "learning_rate": 9.149017728797317e-05, "loss": 0.8268, "step": 6967 }, { "epoch": 0.47211870722948707, "grad_norm": 7.393543720245361, "learning_rate": 9.148880826887536e-05, "loss": 0.9067, "step": 6968 }, { "epoch": 0.47218646249745916, "grad_norm": 7.611250400543213, "learning_rate": 9.148743924977754e-05, "loss": 0.8908, "step": 6969 }, { "epoch": 0.47225421776543125, "grad_norm": 9.330535888671875, "learning_rate": 9.148607023067972e-05, "loss": 0.6368, "step": 6970 }, { "epoch": 0.47232197303340334, "grad_norm": 7.130900859832764, "learning_rate": 9.14847012115819e-05, "loss": 0.8157, "step": 6971 }, { "epoch": 0.47238972830137543, "grad_norm": 6.899352550506592, "learning_rate": 9.14833321924841e-05, "loss": 0.7058, "step": 6972 }, { "epoch": 0.4724574835693475, "grad_norm": 7.850022792816162, "learning_rate": 9.148196317338628e-05, "loss": 0.9622, "step": 6973 }, { "epoch": 0.4725252388373196, "grad_norm": 8.331214904785156, "learning_rate": 9.148059415428846e-05, "loss": 0.9366, "step": 6974 }, { "epoch": 0.4725929941052917, "grad_norm": 6.377279758453369, "learning_rate": 9.147922513519064e-05, "loss": 0.8701, "step": 6975 }, { "epoch": 0.47266074937326374, "grad_norm": 6.650668621063232, "learning_rate": 9.147785611609282e-05, "loss": 0.8066, "step": 6976 }, { "epoch": 0.47272850464123584, "grad_norm": 7.406231880187988, "learning_rate": 9.147648709699501e-05, "loss": 0.8833, "step": 6977 }, { "epoch": 0.4727962599092079, "grad_norm": 6.31017541885376, "learning_rate": 9.147511807789719e-05, "loss": 0.8001, "step": 6978 }, { "epoch": 0.47286401517718, "grad_norm": 6.4131927490234375, "learning_rate": 9.147374905879937e-05, "loss": 0.77, "step": 6979 }, { "epoch": 0.4729317704451521, "grad_norm": 9.2667236328125, "learning_rate": 9.147238003970155e-05, "loss": 0.9891, "step": 6980 }, { "epoch": 0.4729995257131242, "grad_norm": 7.107274055480957, "learning_rate": 9.147101102060373e-05, "loss": 1.0087, "step": 6981 }, { "epoch": 0.4730672809810963, "grad_norm": 8.102210998535156, "learning_rate": 9.146964200150593e-05, "loss": 0.7827, "step": 6982 }, { "epoch": 0.4731350362490684, "grad_norm": 6.7619099617004395, "learning_rate": 9.146827298240811e-05, "loss": 0.7432, "step": 6983 }, { "epoch": 0.4732027915170405, "grad_norm": 7.551081657409668, "learning_rate": 9.146690396331029e-05, "loss": 0.6776, "step": 6984 }, { "epoch": 0.4732705467850125, "grad_norm": 6.923361301422119, "learning_rate": 9.146553494421247e-05, "loss": 0.962, "step": 6985 }, { "epoch": 0.4733383020529846, "grad_norm": 6.971046447753906, "learning_rate": 9.146416592511466e-05, "loss": 0.9481, "step": 6986 }, { "epoch": 0.4734060573209567, "grad_norm": 7.046440601348877, "learning_rate": 9.146279690601684e-05, "loss": 1.0244, "step": 6987 }, { "epoch": 0.4734738125889288, "grad_norm": 6.281298637390137, "learning_rate": 9.146142788691902e-05, "loss": 0.8104, "step": 6988 }, { "epoch": 0.4735415678569009, "grad_norm": 10.22514820098877, "learning_rate": 9.14600588678212e-05, "loss": 1.1109, "step": 6989 }, { "epoch": 0.47360932312487297, "grad_norm": 6.466033458709717, "learning_rate": 9.145868984872338e-05, "loss": 0.9219, "step": 6990 }, { "epoch": 0.47367707839284506, "grad_norm": 7.837368011474609, "learning_rate": 9.145732082962558e-05, "loss": 1.0033, "step": 6991 }, { "epoch": 0.47374483366081716, "grad_norm": 7.501054286956787, "learning_rate": 9.145595181052776e-05, "loss": 0.919, "step": 6992 }, { "epoch": 0.4738125889287892, "grad_norm": 6.243696212768555, "learning_rate": 9.145458279142994e-05, "loss": 0.7736, "step": 6993 }, { "epoch": 0.4738803441967613, "grad_norm": 7.519147872924805, "learning_rate": 9.145321377233212e-05, "loss": 0.8014, "step": 6994 }, { "epoch": 0.4739480994647334, "grad_norm": 5.793912887573242, "learning_rate": 9.145184475323431e-05, "loss": 0.7508, "step": 6995 }, { "epoch": 0.47401585473270547, "grad_norm": 7.050177097320557, "learning_rate": 9.145047573413649e-05, "loss": 0.8174, "step": 6996 }, { "epoch": 0.47408361000067756, "grad_norm": 5.979698657989502, "learning_rate": 9.144910671503867e-05, "loss": 0.9812, "step": 6997 }, { "epoch": 0.47415136526864965, "grad_norm": 7.254084587097168, "learning_rate": 9.144773769594087e-05, "loss": 0.8015, "step": 6998 }, { "epoch": 0.47421912053662174, "grad_norm": 8.168086051940918, "learning_rate": 9.144636867684305e-05, "loss": 0.8107, "step": 6999 }, { "epoch": 0.47428687580459383, "grad_norm": 6.693539619445801, "learning_rate": 9.144499965774523e-05, "loss": 0.7048, "step": 7000 }, { "epoch": 0.47435463107256587, "grad_norm": 7.216420650482178, "learning_rate": 9.144363063864742e-05, "loss": 0.7434, "step": 7001 }, { "epoch": 0.47442238634053796, "grad_norm": 8.014084815979004, "learning_rate": 9.14422616195496e-05, "loss": 1.1213, "step": 7002 }, { "epoch": 0.47449014160851005, "grad_norm": 6.1906938552856445, "learning_rate": 9.144089260045178e-05, "loss": 0.8188, "step": 7003 }, { "epoch": 0.47455789687648214, "grad_norm": 8.70464038848877, "learning_rate": 9.143952358135398e-05, "loss": 1.0273, "step": 7004 }, { "epoch": 0.47462565214445424, "grad_norm": 5.303441524505615, "learning_rate": 9.143815456225616e-05, "loss": 0.6782, "step": 7005 }, { "epoch": 0.4746934074124263, "grad_norm": 8.176512718200684, "learning_rate": 9.143678554315834e-05, "loss": 0.7853, "step": 7006 }, { "epoch": 0.4747611626803984, "grad_norm": 6.074409008026123, "learning_rate": 9.143541652406052e-05, "loss": 0.9088, "step": 7007 }, { "epoch": 0.4748289179483705, "grad_norm": 7.32485294342041, "learning_rate": 9.14340475049627e-05, "loss": 1.0558, "step": 7008 }, { "epoch": 0.47489667321634255, "grad_norm": 5.4144463539123535, "learning_rate": 9.143267848586489e-05, "loss": 0.9064, "step": 7009 }, { "epoch": 0.47496442848431464, "grad_norm": 7.717291355133057, "learning_rate": 9.143130946676707e-05, "loss": 0.9129, "step": 7010 }, { "epoch": 0.47503218375228673, "grad_norm": 7.902177333831787, "learning_rate": 9.142994044766925e-05, "loss": 1.1144, "step": 7011 }, { "epoch": 0.4750999390202588, "grad_norm": 6.94300651550293, "learning_rate": 9.142857142857143e-05, "loss": 0.7982, "step": 7012 }, { "epoch": 0.4751676942882309, "grad_norm": 7.08957052230835, "learning_rate": 9.142720240947361e-05, "loss": 1.009, "step": 7013 }, { "epoch": 0.475235449556203, "grad_norm": 5.148087024688721, "learning_rate": 9.14258333903758e-05, "loss": 0.7443, "step": 7014 }, { "epoch": 0.4753032048241751, "grad_norm": 6.155871868133545, "learning_rate": 9.142446437127799e-05, "loss": 0.8135, "step": 7015 }, { "epoch": 0.4753709600921472, "grad_norm": 6.569172382354736, "learning_rate": 9.142309535218017e-05, "loss": 0.7684, "step": 7016 }, { "epoch": 0.4754387153601192, "grad_norm": 9.110980033874512, "learning_rate": 9.142172633308235e-05, "loss": 1.0382, "step": 7017 }, { "epoch": 0.4755064706280913, "grad_norm": 6.317762851715088, "learning_rate": 9.142035731398454e-05, "loss": 0.7206, "step": 7018 }, { "epoch": 0.4755742258960634, "grad_norm": 7.054732799530029, "learning_rate": 9.141898829488672e-05, "loss": 0.7442, "step": 7019 }, { "epoch": 0.4756419811640355, "grad_norm": 8.904619216918945, "learning_rate": 9.14176192757889e-05, "loss": 1.1686, "step": 7020 }, { "epoch": 0.4757097364320076, "grad_norm": 7.206265926361084, "learning_rate": 9.141625025669108e-05, "loss": 0.8675, "step": 7021 }, { "epoch": 0.4757774916999797, "grad_norm": 7.7408318519592285, "learning_rate": 9.141488123759326e-05, "loss": 0.9132, "step": 7022 }, { "epoch": 0.4758452469679518, "grad_norm": 5.807632923126221, "learning_rate": 9.141351221849546e-05, "loss": 0.8356, "step": 7023 }, { "epoch": 0.47591300223592387, "grad_norm": 8.07009506225586, "learning_rate": 9.141214319939764e-05, "loss": 1.1522, "step": 7024 }, { "epoch": 0.4759807575038959, "grad_norm": 7.409401893615723, "learning_rate": 9.141077418029982e-05, "loss": 0.7972, "step": 7025 }, { "epoch": 0.476048512771868, "grad_norm": 7.14201545715332, "learning_rate": 9.1409405161202e-05, "loss": 0.9284, "step": 7026 }, { "epoch": 0.4761162680398401, "grad_norm": 6.279862880706787, "learning_rate": 9.140803614210419e-05, "loss": 0.7995, "step": 7027 }, { "epoch": 0.4761840233078122, "grad_norm": 6.493180274963379, "learning_rate": 9.140666712300637e-05, "loss": 0.8848, "step": 7028 }, { "epoch": 0.47625177857578427, "grad_norm": 8.528377532958984, "learning_rate": 9.140529810390855e-05, "loss": 0.9142, "step": 7029 }, { "epoch": 0.47631953384375636, "grad_norm": 6.636556625366211, "learning_rate": 9.140392908481073e-05, "loss": 0.6841, "step": 7030 }, { "epoch": 0.47638728911172845, "grad_norm": 7.340085983276367, "learning_rate": 9.140256006571291e-05, "loss": 1.0198, "step": 7031 }, { "epoch": 0.47645504437970054, "grad_norm": 7.510453224182129, "learning_rate": 9.14011910466151e-05, "loss": 0.8581, "step": 7032 }, { "epoch": 0.4765227996476726, "grad_norm": 5.910886287689209, "learning_rate": 9.139982202751729e-05, "loss": 0.7947, "step": 7033 }, { "epoch": 0.47659055491564467, "grad_norm": 6.257016181945801, "learning_rate": 9.139845300841947e-05, "loss": 0.8759, "step": 7034 }, { "epoch": 0.47665831018361676, "grad_norm": 7.840462684631348, "learning_rate": 9.139708398932165e-05, "loss": 1.1975, "step": 7035 }, { "epoch": 0.47672606545158885, "grad_norm": 9.107316970825195, "learning_rate": 9.139571497022383e-05, "loss": 0.8321, "step": 7036 }, { "epoch": 0.47679382071956095, "grad_norm": 5.630661964416504, "learning_rate": 9.139434595112602e-05, "loss": 0.7302, "step": 7037 }, { "epoch": 0.47686157598753304, "grad_norm": 6.390323162078857, "learning_rate": 9.13929769320282e-05, "loss": 0.862, "step": 7038 }, { "epoch": 0.47692933125550513, "grad_norm": 8.85464096069336, "learning_rate": 9.139160791293038e-05, "loss": 0.9283, "step": 7039 }, { "epoch": 0.4769970865234772, "grad_norm": 6.312126159667969, "learning_rate": 9.139023889383256e-05, "loss": 0.933, "step": 7040 }, { "epoch": 0.4770648417914493, "grad_norm": 6.027670383453369, "learning_rate": 9.138886987473476e-05, "loss": 0.8221, "step": 7041 }, { "epoch": 0.47713259705942135, "grad_norm": 8.296350479125977, "learning_rate": 9.138750085563694e-05, "loss": 0.8621, "step": 7042 }, { "epoch": 0.47720035232739344, "grad_norm": 8.337299346923828, "learning_rate": 9.138613183653912e-05, "loss": 0.9126, "step": 7043 }, { "epoch": 0.47726810759536553, "grad_norm": 7.545529842376709, "learning_rate": 9.138476281744131e-05, "loss": 1.1327, "step": 7044 }, { "epoch": 0.4773358628633376, "grad_norm": 5.35225248336792, "learning_rate": 9.138339379834349e-05, "loss": 0.7311, "step": 7045 }, { "epoch": 0.4774036181313097, "grad_norm": 7.592402458190918, "learning_rate": 9.138202477924567e-05, "loss": 0.9959, "step": 7046 }, { "epoch": 0.4774713733992818, "grad_norm": 8.010600090026855, "learning_rate": 9.138065576014787e-05, "loss": 0.7788, "step": 7047 }, { "epoch": 0.4775391286672539, "grad_norm": 7.170941352844238, "learning_rate": 9.137928674105005e-05, "loss": 0.9218, "step": 7048 }, { "epoch": 0.477606883935226, "grad_norm": 6.852916240692139, "learning_rate": 9.137791772195223e-05, "loss": 0.7614, "step": 7049 }, { "epoch": 0.477674639203198, "grad_norm": 8.727351188659668, "learning_rate": 9.137654870285442e-05, "loss": 0.8449, "step": 7050 }, { "epoch": 0.4777423944711701, "grad_norm": 6.188234329223633, "learning_rate": 9.13751796837566e-05, "loss": 0.7001, "step": 7051 }, { "epoch": 0.4778101497391422, "grad_norm": 6.4075398445129395, "learning_rate": 9.137381066465878e-05, "loss": 0.9363, "step": 7052 }, { "epoch": 0.4778779050071143, "grad_norm": 7.744530200958252, "learning_rate": 9.137244164556096e-05, "loss": 0.9253, "step": 7053 }, { "epoch": 0.4779456602750864, "grad_norm": 6.745645999908447, "learning_rate": 9.137107262646314e-05, "loss": 0.9649, "step": 7054 }, { "epoch": 0.4780134155430585, "grad_norm": 8.560202598571777, "learning_rate": 9.136970360736534e-05, "loss": 1.2453, "step": 7055 }, { "epoch": 0.4780811708110306, "grad_norm": 6.941448211669922, "learning_rate": 9.136833458826752e-05, "loss": 0.8577, "step": 7056 }, { "epoch": 0.47814892607900267, "grad_norm": 6.4724626541137695, "learning_rate": 9.13669655691697e-05, "loss": 0.9304, "step": 7057 }, { "epoch": 0.4782166813469747, "grad_norm": 6.535644054412842, "learning_rate": 9.136559655007188e-05, "loss": 0.8323, "step": 7058 }, { "epoch": 0.4782844366149468, "grad_norm": 6.4093217849731445, "learning_rate": 9.136422753097407e-05, "loss": 0.9816, "step": 7059 }, { "epoch": 0.4783521918829189, "grad_norm": 6.645406246185303, "learning_rate": 9.136285851187625e-05, "loss": 0.9507, "step": 7060 }, { "epoch": 0.478419947150891, "grad_norm": 8.383099555969238, "learning_rate": 9.136148949277843e-05, "loss": 1.0339, "step": 7061 }, { "epoch": 0.47848770241886307, "grad_norm": 6.430543422698975, "learning_rate": 9.136012047368061e-05, "loss": 0.755, "step": 7062 }, { "epoch": 0.47855545768683516, "grad_norm": 6.5899176597595215, "learning_rate": 9.135875145458279e-05, "loss": 1.0841, "step": 7063 }, { "epoch": 0.47862321295480725, "grad_norm": 9.340789794921875, "learning_rate": 9.135738243548499e-05, "loss": 0.7539, "step": 7064 }, { "epoch": 0.47869096822277934, "grad_norm": 7.300266742706299, "learning_rate": 9.135601341638717e-05, "loss": 0.9049, "step": 7065 }, { "epoch": 0.4787587234907514, "grad_norm": 6.495701313018799, "learning_rate": 9.135464439728935e-05, "loss": 0.9168, "step": 7066 }, { "epoch": 0.4788264787587235, "grad_norm": 8.14792537689209, "learning_rate": 9.135327537819153e-05, "loss": 0.7678, "step": 7067 }, { "epoch": 0.47889423402669556, "grad_norm": 5.638490200042725, "learning_rate": 9.135190635909371e-05, "loss": 0.8914, "step": 7068 }, { "epoch": 0.47896198929466766, "grad_norm": 7.043061256408691, "learning_rate": 9.13505373399959e-05, "loss": 0.9575, "step": 7069 }, { "epoch": 0.47902974456263975, "grad_norm": 7.867429256439209, "learning_rate": 9.134916832089808e-05, "loss": 0.9887, "step": 7070 }, { "epoch": 0.47909749983061184, "grad_norm": 5.737957000732422, "learning_rate": 9.134779930180026e-05, "loss": 0.8605, "step": 7071 }, { "epoch": 0.47916525509858393, "grad_norm": 9.436585426330566, "learning_rate": 9.134643028270244e-05, "loss": 1.0482, "step": 7072 }, { "epoch": 0.479233010366556, "grad_norm": 6.28998327255249, "learning_rate": 9.134506126360464e-05, "loss": 1.0235, "step": 7073 }, { "epoch": 0.47930076563452806, "grad_norm": 5.121227741241455, "learning_rate": 9.134369224450682e-05, "loss": 0.8735, "step": 7074 }, { "epoch": 0.47936852090250015, "grad_norm": 5.818413734436035, "learning_rate": 9.1342323225409e-05, "loss": 0.738, "step": 7075 }, { "epoch": 0.47943627617047224, "grad_norm": 6.716359615325928, "learning_rate": 9.134095420631118e-05, "loss": 0.8825, "step": 7076 }, { "epoch": 0.47950403143844433, "grad_norm": 5.944983959197998, "learning_rate": 9.133958518721336e-05, "loss": 0.9145, "step": 7077 }, { "epoch": 0.4795717867064164, "grad_norm": 6.65972375869751, "learning_rate": 9.133821616811555e-05, "loss": 0.7746, "step": 7078 }, { "epoch": 0.4796395419743885, "grad_norm": 7.595485210418701, "learning_rate": 9.133684714901773e-05, "loss": 0.817, "step": 7079 }, { "epoch": 0.4797072972423606, "grad_norm": 5.7008280754089355, "learning_rate": 9.133547812991991e-05, "loss": 0.7996, "step": 7080 }, { "epoch": 0.4797750525103327, "grad_norm": 6.7002668380737305, "learning_rate": 9.133410911082209e-05, "loss": 1.0676, "step": 7081 }, { "epoch": 0.47984280777830474, "grad_norm": 8.43227481842041, "learning_rate": 9.133274009172429e-05, "loss": 0.9844, "step": 7082 }, { "epoch": 0.47991056304627683, "grad_norm": 5.9677653312683105, "learning_rate": 9.133137107262647e-05, "loss": 0.9158, "step": 7083 }, { "epoch": 0.4799783183142489, "grad_norm": 7.774659156799316, "learning_rate": 9.133000205352865e-05, "loss": 0.6829, "step": 7084 }, { "epoch": 0.480046073582221, "grad_norm": 7.344854354858398, "learning_rate": 9.132863303443083e-05, "loss": 0.8579, "step": 7085 }, { "epoch": 0.4801138288501931, "grad_norm": 6.4308624267578125, "learning_rate": 9.132726401533301e-05, "loss": 0.936, "step": 7086 }, { "epoch": 0.4801815841181652, "grad_norm": 7.600352764129639, "learning_rate": 9.13258949962352e-05, "loss": 0.8561, "step": 7087 }, { "epoch": 0.4802493393861373, "grad_norm": 9.475528717041016, "learning_rate": 9.132452597713738e-05, "loss": 0.9684, "step": 7088 }, { "epoch": 0.4803170946541094, "grad_norm": 5.927639007568359, "learning_rate": 9.132315695803956e-05, "loss": 0.7472, "step": 7089 }, { "epoch": 0.48038484992208147, "grad_norm": 7.304515838623047, "learning_rate": 9.132178793894176e-05, "loss": 0.8983, "step": 7090 }, { "epoch": 0.4804526051900535, "grad_norm": 7.6303391456604, "learning_rate": 9.132041891984394e-05, "loss": 0.864, "step": 7091 }, { "epoch": 0.4805203604580256, "grad_norm": 6.243314266204834, "learning_rate": 9.131904990074612e-05, "loss": 0.8448, "step": 7092 }, { "epoch": 0.4805881157259977, "grad_norm": 6.947970867156982, "learning_rate": 9.131768088164831e-05, "loss": 0.7853, "step": 7093 }, { "epoch": 0.4806558709939698, "grad_norm": 5.687379837036133, "learning_rate": 9.131631186255049e-05, "loss": 0.838, "step": 7094 }, { "epoch": 0.48072362626194187, "grad_norm": 6.477471351623535, "learning_rate": 9.131494284345267e-05, "loss": 0.9849, "step": 7095 }, { "epoch": 0.48079138152991396, "grad_norm": 6.526174068450928, "learning_rate": 9.131357382435487e-05, "loss": 0.9269, "step": 7096 }, { "epoch": 0.48085913679788606, "grad_norm": 7.404792308807373, "learning_rate": 9.131220480525705e-05, "loss": 0.8257, "step": 7097 }, { "epoch": 0.48092689206585815, "grad_norm": 7.290988445281982, "learning_rate": 9.131083578615923e-05, "loss": 1.0275, "step": 7098 }, { "epoch": 0.4809946473338302, "grad_norm": 7.744154930114746, "learning_rate": 9.13094667670614e-05, "loss": 0.9692, "step": 7099 }, { "epoch": 0.4810624026018023, "grad_norm": 5.7665252685546875, "learning_rate": 9.130809774796359e-05, "loss": 0.6751, "step": 7100 }, { "epoch": 0.48113015786977437, "grad_norm": 6.921631813049316, "learning_rate": 9.130672872886578e-05, "loss": 0.8771, "step": 7101 }, { "epoch": 0.48119791313774646, "grad_norm": 6.097098350524902, "learning_rate": 9.130535970976796e-05, "loss": 0.737, "step": 7102 }, { "epoch": 0.48126566840571855, "grad_norm": 7.192615032196045, "learning_rate": 9.130399069067014e-05, "loss": 0.9283, "step": 7103 }, { "epoch": 0.48133342367369064, "grad_norm": 6.464249610900879, "learning_rate": 9.130262167157232e-05, "loss": 1.0467, "step": 7104 }, { "epoch": 0.48140117894166273, "grad_norm": 4.616031646728516, "learning_rate": 9.130125265247452e-05, "loss": 0.5657, "step": 7105 }, { "epoch": 0.4814689342096348, "grad_norm": 6.356307029724121, "learning_rate": 9.12998836333767e-05, "loss": 0.722, "step": 7106 }, { "epoch": 0.48153668947760686, "grad_norm": 6.398674011230469, "learning_rate": 9.129851461427888e-05, "loss": 1.066, "step": 7107 }, { "epoch": 0.48160444474557895, "grad_norm": 7.015667915344238, "learning_rate": 9.129714559518106e-05, "loss": 0.8681, "step": 7108 }, { "epoch": 0.48167220001355104, "grad_norm": 8.212510108947754, "learning_rate": 9.129577657608324e-05, "loss": 0.8158, "step": 7109 }, { "epoch": 0.48173995528152314, "grad_norm": 8.409659385681152, "learning_rate": 9.129440755698543e-05, "loss": 0.7391, "step": 7110 }, { "epoch": 0.4818077105494952, "grad_norm": 7.42771053314209, "learning_rate": 9.129303853788761e-05, "loss": 0.9876, "step": 7111 }, { "epoch": 0.4818754658174673, "grad_norm": 7.816039562225342, "learning_rate": 9.129166951878979e-05, "loss": 0.9601, "step": 7112 }, { "epoch": 0.4819432210854394, "grad_norm": 8.761859893798828, "learning_rate": 9.129030049969197e-05, "loss": 0.6283, "step": 7113 }, { "epoch": 0.4820109763534115, "grad_norm": 6.981362342834473, "learning_rate": 9.128893148059415e-05, "loss": 0.9901, "step": 7114 }, { "epoch": 0.48207873162138354, "grad_norm": 7.718024253845215, "learning_rate": 9.128756246149635e-05, "loss": 0.8318, "step": 7115 }, { "epoch": 0.48214648688935563, "grad_norm": 8.191770553588867, "learning_rate": 9.128619344239853e-05, "loss": 0.9108, "step": 7116 }, { "epoch": 0.4822142421573277, "grad_norm": 8.095576286315918, "learning_rate": 9.12848244233007e-05, "loss": 0.8404, "step": 7117 }, { "epoch": 0.4822819974252998, "grad_norm": 6.66763162612915, "learning_rate": 9.128345540420289e-05, "loss": 0.6538, "step": 7118 }, { "epoch": 0.4823497526932719, "grad_norm": 7.253853797912598, "learning_rate": 9.128208638510508e-05, "loss": 0.8893, "step": 7119 }, { "epoch": 0.482417507961244, "grad_norm": 8.162705421447754, "learning_rate": 9.128071736600726e-05, "loss": 0.9602, "step": 7120 }, { "epoch": 0.4824852632292161, "grad_norm": 7.548867225646973, "learning_rate": 9.127934834690944e-05, "loss": 0.8701, "step": 7121 }, { "epoch": 0.4825530184971882, "grad_norm": 5.984834671020508, "learning_rate": 9.127797932781162e-05, "loss": 1.0416, "step": 7122 }, { "epoch": 0.4826207737651602, "grad_norm": 8.716156005859375, "learning_rate": 9.12766103087138e-05, "loss": 1.016, "step": 7123 }, { "epoch": 0.4826885290331323, "grad_norm": 6.2188873291015625, "learning_rate": 9.1275241289616e-05, "loss": 0.7147, "step": 7124 }, { "epoch": 0.4827562843011044, "grad_norm": 10.026150703430176, "learning_rate": 9.127387227051818e-05, "loss": 0.9221, "step": 7125 }, { "epoch": 0.4828240395690765, "grad_norm": 7.320329666137695, "learning_rate": 9.127250325142036e-05, "loss": 0.8486, "step": 7126 }, { "epoch": 0.4828917948370486, "grad_norm": 6.134348392486572, "learning_rate": 9.127113423232254e-05, "loss": 0.7383, "step": 7127 }, { "epoch": 0.4829595501050207, "grad_norm": 7.2880778312683105, "learning_rate": 9.126976521322473e-05, "loss": 1.059, "step": 7128 }, { "epoch": 0.48302730537299277, "grad_norm": 6.100679874420166, "learning_rate": 9.126839619412691e-05, "loss": 0.8673, "step": 7129 }, { "epoch": 0.48309506064096486, "grad_norm": 6.899023056030273, "learning_rate": 9.126702717502909e-05, "loss": 1.1572, "step": 7130 }, { "epoch": 0.4831628159089369, "grad_norm": 6.187694549560547, "learning_rate": 9.126565815593127e-05, "loss": 0.9584, "step": 7131 }, { "epoch": 0.483230571176909, "grad_norm": 7.093903064727783, "learning_rate": 9.126428913683345e-05, "loss": 0.969, "step": 7132 }, { "epoch": 0.4832983264448811, "grad_norm": 6.003389835357666, "learning_rate": 9.126292011773565e-05, "loss": 0.7638, "step": 7133 }, { "epoch": 0.48336608171285317, "grad_norm": 5.557130336761475, "learning_rate": 9.126155109863783e-05, "loss": 0.8529, "step": 7134 }, { "epoch": 0.48343383698082526, "grad_norm": 6.029399871826172, "learning_rate": 9.126018207954001e-05, "loss": 0.6359, "step": 7135 }, { "epoch": 0.48350159224879735, "grad_norm": 6.123723030090332, "learning_rate": 9.12588130604422e-05, "loss": 0.9462, "step": 7136 }, { "epoch": 0.48356934751676944, "grad_norm": 6.213245868682861, "learning_rate": 9.125744404134438e-05, "loss": 0.8322, "step": 7137 }, { "epoch": 0.48363710278474153, "grad_norm": 7.509876251220703, "learning_rate": 9.125607502224656e-05, "loss": 1.0556, "step": 7138 }, { "epoch": 0.48370485805271357, "grad_norm": 6.355532646179199, "learning_rate": 9.125470600314876e-05, "loss": 0.9281, "step": 7139 }, { "epoch": 0.48377261332068566, "grad_norm": 7.302781105041504, "learning_rate": 9.125333698405094e-05, "loss": 0.846, "step": 7140 }, { "epoch": 0.48384036858865775, "grad_norm": 7.081716537475586, "learning_rate": 9.125196796495312e-05, "loss": 0.8571, "step": 7141 }, { "epoch": 0.48390812385662985, "grad_norm": 7.652805328369141, "learning_rate": 9.125059894585531e-05, "loss": 0.7986, "step": 7142 }, { "epoch": 0.48397587912460194, "grad_norm": 5.906263828277588, "learning_rate": 9.124922992675749e-05, "loss": 0.8058, "step": 7143 }, { "epoch": 0.48404363439257403, "grad_norm": 7.398087024688721, "learning_rate": 9.124786090765967e-05, "loss": 0.8961, "step": 7144 }, { "epoch": 0.4841113896605461, "grad_norm": 9.616337776184082, "learning_rate": 9.124649188856185e-05, "loss": 0.9403, "step": 7145 }, { "epoch": 0.4841791449285182, "grad_norm": 7.74692440032959, "learning_rate": 9.124512286946403e-05, "loss": 0.7659, "step": 7146 }, { "epoch": 0.4842469001964903, "grad_norm": 6.067378520965576, "learning_rate": 9.124375385036623e-05, "loss": 0.8094, "step": 7147 }, { "epoch": 0.48431465546446234, "grad_norm": 7.943274974822998, "learning_rate": 9.12423848312684e-05, "loss": 0.8103, "step": 7148 }, { "epoch": 0.48438241073243443, "grad_norm": 7.710971355438232, "learning_rate": 9.124101581217059e-05, "loss": 1.0818, "step": 7149 }, { "epoch": 0.4844501660004065, "grad_norm": 6.904791831970215, "learning_rate": 9.123964679307277e-05, "loss": 0.9413, "step": 7150 }, { "epoch": 0.4845179212683786, "grad_norm": 8.46650505065918, "learning_rate": 9.123827777397496e-05, "loss": 1.1661, "step": 7151 }, { "epoch": 0.4845856765363507, "grad_norm": 7.021801471710205, "learning_rate": 9.123690875487714e-05, "loss": 0.7305, "step": 7152 }, { "epoch": 0.4846534318043228, "grad_norm": 7.553677082061768, "learning_rate": 9.123553973577932e-05, "loss": 0.9806, "step": 7153 }, { "epoch": 0.4847211870722949, "grad_norm": 6.204870700836182, "learning_rate": 9.12341707166815e-05, "loss": 0.8433, "step": 7154 }, { "epoch": 0.484788942340267, "grad_norm": 7.81880521774292, "learning_rate": 9.123280169758368e-05, "loss": 1.112, "step": 7155 }, { "epoch": 0.484856697608239, "grad_norm": 7.029433727264404, "learning_rate": 9.123143267848588e-05, "loss": 0.8068, "step": 7156 }, { "epoch": 0.4849244528762111, "grad_norm": 6.795009136199951, "learning_rate": 9.123006365938806e-05, "loss": 0.766, "step": 7157 }, { "epoch": 0.4849922081441832, "grad_norm": 5.924415588378906, "learning_rate": 9.122869464029024e-05, "loss": 0.7914, "step": 7158 }, { "epoch": 0.4850599634121553, "grad_norm": 5.561947345733643, "learning_rate": 9.122732562119242e-05, "loss": 1.1486, "step": 7159 }, { "epoch": 0.4851277186801274, "grad_norm": 7.202826976776123, "learning_rate": 9.122595660209461e-05, "loss": 0.8802, "step": 7160 }, { "epoch": 0.4851954739480995, "grad_norm": 6.435755729675293, "learning_rate": 9.122458758299679e-05, "loss": 0.888, "step": 7161 }, { "epoch": 0.48526322921607157, "grad_norm": 6.197578430175781, "learning_rate": 9.122321856389897e-05, "loss": 0.9139, "step": 7162 }, { "epoch": 0.48533098448404366, "grad_norm": 6.596435070037842, "learning_rate": 9.122184954480115e-05, "loss": 0.7802, "step": 7163 }, { "epoch": 0.4853987397520157, "grad_norm": 7.858447551727295, "learning_rate": 9.122048052570333e-05, "loss": 0.8759, "step": 7164 }, { "epoch": 0.4854664950199878, "grad_norm": 6.002086162567139, "learning_rate": 9.121911150660553e-05, "loss": 0.9621, "step": 7165 }, { "epoch": 0.4855342502879599, "grad_norm": 5.917041301727295, "learning_rate": 9.12177424875077e-05, "loss": 0.8144, "step": 7166 }, { "epoch": 0.48560200555593197, "grad_norm": 6.202271461486816, "learning_rate": 9.121637346840989e-05, "loss": 0.9937, "step": 7167 }, { "epoch": 0.48566976082390406, "grad_norm": 6.635425567626953, "learning_rate": 9.121500444931207e-05, "loss": 0.8103, "step": 7168 }, { "epoch": 0.48573751609187615, "grad_norm": 7.288759708404541, "learning_rate": 9.121363543021425e-05, "loss": 0.8354, "step": 7169 }, { "epoch": 0.48580527135984825, "grad_norm": 6.015995979309082, "learning_rate": 9.121226641111644e-05, "loss": 1.0732, "step": 7170 }, { "epoch": 0.48587302662782034, "grad_norm": 6.693684101104736, "learning_rate": 9.121089739201862e-05, "loss": 0.9792, "step": 7171 }, { "epoch": 0.4859407818957924, "grad_norm": 7.2573561668396, "learning_rate": 9.12095283729208e-05, "loss": 0.8823, "step": 7172 }, { "epoch": 0.48600853716376446, "grad_norm": 6.1881585121154785, "learning_rate": 9.120815935382298e-05, "loss": 0.8397, "step": 7173 }, { "epoch": 0.48607629243173656, "grad_norm": 7.354151725769043, "learning_rate": 9.120679033472518e-05, "loss": 1.0145, "step": 7174 }, { "epoch": 0.48614404769970865, "grad_norm": 6.806859016418457, "learning_rate": 9.120542131562736e-05, "loss": 0.9289, "step": 7175 }, { "epoch": 0.48621180296768074, "grad_norm": 6.098382949829102, "learning_rate": 9.120405229652954e-05, "loss": 0.9819, "step": 7176 }, { "epoch": 0.48627955823565283, "grad_norm": 6.944835186004639, "learning_rate": 9.120268327743172e-05, "loss": 0.8004, "step": 7177 }, { "epoch": 0.4863473135036249, "grad_norm": 7.9429497718811035, "learning_rate": 9.12013142583339e-05, "loss": 0.9473, "step": 7178 }, { "epoch": 0.486415068771597, "grad_norm": 6.831770420074463, "learning_rate": 9.119994523923609e-05, "loss": 0.8866, "step": 7179 }, { "epoch": 0.48648282403956905, "grad_norm": 7.992087364196777, "learning_rate": 9.119857622013827e-05, "loss": 0.6405, "step": 7180 }, { "epoch": 0.48655057930754114, "grad_norm": 7.081966400146484, "learning_rate": 9.119720720104045e-05, "loss": 0.9166, "step": 7181 }, { "epoch": 0.48661833457551323, "grad_norm": 8.257608413696289, "learning_rate": 9.119583818194263e-05, "loss": 0.7556, "step": 7182 }, { "epoch": 0.4866860898434853, "grad_norm": 6.714028835296631, "learning_rate": 9.119446916284483e-05, "loss": 0.784, "step": 7183 }, { "epoch": 0.4867538451114574, "grad_norm": 5.71464204788208, "learning_rate": 9.1193100143747e-05, "loss": 0.815, "step": 7184 }, { "epoch": 0.4868216003794295, "grad_norm": 8.290979385375977, "learning_rate": 9.119173112464919e-05, "loss": 0.822, "step": 7185 }, { "epoch": 0.4868893556474016, "grad_norm": 6.116361141204834, "learning_rate": 9.119036210555138e-05, "loss": 0.7727, "step": 7186 }, { "epoch": 0.4869571109153737, "grad_norm": 5.137567043304443, "learning_rate": 9.118899308645356e-05, "loss": 0.4916, "step": 7187 }, { "epoch": 0.48702486618334573, "grad_norm": 6.987879753112793, "learning_rate": 9.118762406735574e-05, "loss": 1.0662, "step": 7188 }, { "epoch": 0.4870926214513178, "grad_norm": 7.223506927490234, "learning_rate": 9.118625504825794e-05, "loss": 0.7408, "step": 7189 }, { "epoch": 0.4871603767192899, "grad_norm": 7.493766784667969, "learning_rate": 9.118488602916012e-05, "loss": 0.9625, "step": 7190 }, { "epoch": 0.487228131987262, "grad_norm": 7.265352725982666, "learning_rate": 9.11835170100623e-05, "loss": 0.9228, "step": 7191 }, { "epoch": 0.4872958872552341, "grad_norm": 6.04194450378418, "learning_rate": 9.118214799096449e-05, "loss": 0.8081, "step": 7192 }, { "epoch": 0.4873636425232062, "grad_norm": 7.033185958862305, "learning_rate": 9.118077897186667e-05, "loss": 1.0709, "step": 7193 }, { "epoch": 0.4874313977911783, "grad_norm": 5.993305683135986, "learning_rate": 9.117940995276885e-05, "loss": 0.8222, "step": 7194 }, { "epoch": 0.48749915305915037, "grad_norm": 6.695589542388916, "learning_rate": 9.117804093367103e-05, "loss": 0.7981, "step": 7195 }, { "epoch": 0.48756690832712246, "grad_norm": 6.6938157081604, "learning_rate": 9.117667191457321e-05, "loss": 0.8485, "step": 7196 }, { "epoch": 0.4876346635950945, "grad_norm": 6.760074138641357, "learning_rate": 9.11753028954754e-05, "loss": 0.6706, "step": 7197 }, { "epoch": 0.4877024188630666, "grad_norm": 6.397393703460693, "learning_rate": 9.117393387637759e-05, "loss": 0.8247, "step": 7198 }, { "epoch": 0.4877701741310387, "grad_norm": 7.138194561004639, "learning_rate": 9.117256485727977e-05, "loss": 1.0124, "step": 7199 }, { "epoch": 0.4878379293990108, "grad_norm": 6.4646172523498535, "learning_rate": 9.117119583818195e-05, "loss": 1.0531, "step": 7200 }, { "epoch": 0.48790568466698286, "grad_norm": 6.58534574508667, "learning_rate": 9.116982681908413e-05, "loss": 0.962, "step": 7201 }, { "epoch": 0.48797343993495496, "grad_norm": 8.237184524536133, "learning_rate": 9.116845779998632e-05, "loss": 1.0363, "step": 7202 }, { "epoch": 0.48804119520292705, "grad_norm": 9.047645568847656, "learning_rate": 9.11670887808885e-05, "loss": 0.8925, "step": 7203 }, { "epoch": 0.48810895047089914, "grad_norm": 5.87882137298584, "learning_rate": 9.116571976179068e-05, "loss": 0.7843, "step": 7204 }, { "epoch": 0.4881767057388712, "grad_norm": 7.792137622833252, "learning_rate": 9.116435074269286e-05, "loss": 0.6999, "step": 7205 }, { "epoch": 0.48824446100684327, "grad_norm": 6.028510570526123, "learning_rate": 9.116298172359506e-05, "loss": 0.9233, "step": 7206 }, { "epoch": 0.48831221627481536, "grad_norm": 6.219117164611816, "learning_rate": 9.116161270449724e-05, "loss": 0.5807, "step": 7207 }, { "epoch": 0.48837997154278745, "grad_norm": 5.99801778793335, "learning_rate": 9.116024368539942e-05, "loss": 0.9267, "step": 7208 }, { "epoch": 0.48844772681075954, "grad_norm": 7.163285255432129, "learning_rate": 9.11588746663016e-05, "loss": 0.8695, "step": 7209 }, { "epoch": 0.48851548207873163, "grad_norm": 5.209384441375732, "learning_rate": 9.115750564720378e-05, "loss": 0.8395, "step": 7210 }, { "epoch": 0.4885832373467037, "grad_norm": 5.942741394042969, "learning_rate": 9.115613662810597e-05, "loss": 0.8593, "step": 7211 }, { "epoch": 0.4886509926146758, "grad_norm": 6.742303848266602, "learning_rate": 9.115476760900815e-05, "loss": 0.7964, "step": 7212 }, { "epoch": 0.48871874788264785, "grad_norm": 6.951894283294678, "learning_rate": 9.115339858991033e-05, "loss": 0.9285, "step": 7213 }, { "epoch": 0.48878650315061994, "grad_norm": 5.752067565917969, "learning_rate": 9.115202957081251e-05, "loss": 0.6053, "step": 7214 }, { "epoch": 0.48885425841859204, "grad_norm": 6.7097883224487305, "learning_rate": 9.11506605517147e-05, "loss": 0.82, "step": 7215 }, { "epoch": 0.4889220136865641, "grad_norm": 6.396644592285156, "learning_rate": 9.114929153261689e-05, "loss": 0.9438, "step": 7216 }, { "epoch": 0.4889897689545362, "grad_norm": 6.378931999206543, "learning_rate": 9.114792251351907e-05, "loss": 0.9635, "step": 7217 }, { "epoch": 0.4890575242225083, "grad_norm": 5.802820682525635, "learning_rate": 9.114655349442125e-05, "loss": 0.7166, "step": 7218 }, { "epoch": 0.4891252794904804, "grad_norm": 6.493535041809082, "learning_rate": 9.114518447532343e-05, "loss": 0.985, "step": 7219 }, { "epoch": 0.4891930347584525, "grad_norm": 7.591537952423096, "learning_rate": 9.114381545622562e-05, "loss": 0.8348, "step": 7220 }, { "epoch": 0.48926079002642453, "grad_norm": 6.379971027374268, "learning_rate": 9.11424464371278e-05, "loss": 0.9761, "step": 7221 }, { "epoch": 0.4893285452943966, "grad_norm": 9.187169075012207, "learning_rate": 9.114107741802998e-05, "loss": 1.1654, "step": 7222 }, { "epoch": 0.4893963005623687, "grad_norm": 6.582739353179932, "learning_rate": 9.113970839893216e-05, "loss": 0.8921, "step": 7223 }, { "epoch": 0.4894640558303408, "grad_norm": 5.6152544021606445, "learning_rate": 9.113833937983434e-05, "loss": 0.7498, "step": 7224 }, { "epoch": 0.4895318110983129, "grad_norm": 6.960738182067871, "learning_rate": 9.113697036073654e-05, "loss": 0.7443, "step": 7225 }, { "epoch": 0.489599566366285, "grad_norm": 6.515749454498291, "learning_rate": 9.113560134163872e-05, "loss": 0.903, "step": 7226 }, { "epoch": 0.4896673216342571, "grad_norm": 7.327613830566406, "learning_rate": 9.11342323225409e-05, "loss": 0.8484, "step": 7227 }, { "epoch": 0.48973507690222917, "grad_norm": 6.575616359710693, "learning_rate": 9.113286330344308e-05, "loss": 1.001, "step": 7228 }, { "epoch": 0.4898028321702012, "grad_norm": 6.429412364959717, "learning_rate": 9.113149428434527e-05, "loss": 0.8028, "step": 7229 }, { "epoch": 0.4898705874381733, "grad_norm": 6.13348388671875, "learning_rate": 9.113012526524745e-05, "loss": 0.8966, "step": 7230 }, { "epoch": 0.4899383427061454, "grad_norm": 5.246626377105713, "learning_rate": 9.112875624614963e-05, "loss": 0.7548, "step": 7231 }, { "epoch": 0.4900060979741175, "grad_norm": 8.197700500488281, "learning_rate": 9.112738722705183e-05, "loss": 0.9122, "step": 7232 }, { "epoch": 0.4900738532420896, "grad_norm": 7.987671375274658, "learning_rate": 9.1126018207954e-05, "loss": 1.0089, "step": 7233 }, { "epoch": 0.49014160851006167, "grad_norm": 6.8674211502075195, "learning_rate": 9.112464918885619e-05, "loss": 0.7244, "step": 7234 }, { "epoch": 0.49020936377803376, "grad_norm": 9.719727516174316, "learning_rate": 9.112328016975838e-05, "loss": 0.9079, "step": 7235 }, { "epoch": 0.49027711904600585, "grad_norm": 6.982578277587891, "learning_rate": 9.112191115066056e-05, "loss": 0.7876, "step": 7236 }, { "epoch": 0.4903448743139779, "grad_norm": 5.029915809631348, "learning_rate": 9.112054213156274e-05, "loss": 0.7574, "step": 7237 }, { "epoch": 0.49041262958195, "grad_norm": 8.605433464050293, "learning_rate": 9.111917311246493e-05, "loss": 0.9255, "step": 7238 }, { "epoch": 0.49048038484992207, "grad_norm": 7.910008907318115, "learning_rate": 9.111780409336711e-05, "loss": 0.9514, "step": 7239 }, { "epoch": 0.49054814011789416, "grad_norm": 6.401332855224609, "learning_rate": 9.11164350742693e-05, "loss": 0.9243, "step": 7240 }, { "epoch": 0.49061589538586625, "grad_norm": 6.722992420196533, "learning_rate": 9.111506605517148e-05, "loss": 0.7595, "step": 7241 }, { "epoch": 0.49068365065383834, "grad_norm": 6.60951566696167, "learning_rate": 9.111369703607366e-05, "loss": 0.9381, "step": 7242 }, { "epoch": 0.49075140592181044, "grad_norm": 7.443787574768066, "learning_rate": 9.111232801697585e-05, "loss": 1.0037, "step": 7243 }, { "epoch": 0.4908191611897825, "grad_norm": 7.43011999130249, "learning_rate": 9.111095899787803e-05, "loss": 0.8213, "step": 7244 }, { "epoch": 0.49088691645775456, "grad_norm": 8.865147590637207, "learning_rate": 9.110958997878021e-05, "loss": 0.9547, "step": 7245 }, { "epoch": 0.49095467172572665, "grad_norm": 5.83010721206665, "learning_rate": 9.110822095968239e-05, "loss": 0.7253, "step": 7246 }, { "epoch": 0.49102242699369875, "grad_norm": 6.1638946533203125, "learning_rate": 9.110685194058458e-05, "loss": 0.7643, "step": 7247 }, { "epoch": 0.49109018226167084, "grad_norm": 6.534294128417969, "learning_rate": 9.110548292148677e-05, "loss": 0.9102, "step": 7248 }, { "epoch": 0.49115793752964293, "grad_norm": 8.244972229003906, "learning_rate": 9.110411390238895e-05, "loss": 0.9615, "step": 7249 }, { "epoch": 0.491225692797615, "grad_norm": 7.7991461753845215, "learning_rate": 9.110274488329113e-05, "loss": 0.9697, "step": 7250 }, { "epoch": 0.4912934480655871, "grad_norm": 7.489588737487793, "learning_rate": 9.11013758641933e-05, "loss": 0.7339, "step": 7251 }, { "epoch": 0.4913612033335592, "grad_norm": 6.112767696380615, "learning_rate": 9.11000068450955e-05, "loss": 1.1147, "step": 7252 }, { "epoch": 0.4914289586015313, "grad_norm": 6.319901943206787, "learning_rate": 9.109863782599768e-05, "loss": 0.9777, "step": 7253 }, { "epoch": 0.49149671386950333, "grad_norm": 7.175682544708252, "learning_rate": 9.109726880689986e-05, "loss": 0.7403, "step": 7254 }, { "epoch": 0.4915644691374754, "grad_norm": 7.254464149475098, "learning_rate": 9.109589978780204e-05, "loss": 0.8533, "step": 7255 }, { "epoch": 0.4916322244054475, "grad_norm": 6.266726493835449, "learning_rate": 9.109453076870422e-05, "loss": 0.8648, "step": 7256 }, { "epoch": 0.4916999796734196, "grad_norm": 7.290742874145508, "learning_rate": 9.109316174960642e-05, "loss": 0.7274, "step": 7257 }, { "epoch": 0.4917677349413917, "grad_norm": 6.882185459136963, "learning_rate": 9.10917927305086e-05, "loss": 0.9184, "step": 7258 }, { "epoch": 0.4918354902093638, "grad_norm": 6.796192169189453, "learning_rate": 9.109042371141078e-05, "loss": 0.9734, "step": 7259 }, { "epoch": 0.4919032454773359, "grad_norm": 7.461274147033691, "learning_rate": 9.108905469231296e-05, "loss": 0.8931, "step": 7260 }, { "epoch": 0.491971000745308, "grad_norm": 6.522415637969971, "learning_rate": 9.108768567321515e-05, "loss": 0.7529, "step": 7261 }, { "epoch": 0.49203875601328, "grad_norm": 7.108310699462891, "learning_rate": 9.108631665411733e-05, "loss": 0.7551, "step": 7262 }, { "epoch": 0.4921065112812521, "grad_norm": 6.65360164642334, "learning_rate": 9.108494763501951e-05, "loss": 1.1178, "step": 7263 }, { "epoch": 0.4921742665492242, "grad_norm": 8.197813034057617, "learning_rate": 9.108357861592169e-05, "loss": 1.2079, "step": 7264 }, { "epoch": 0.4922420218171963, "grad_norm": 6.898741245269775, "learning_rate": 9.108220959682387e-05, "loss": 0.7838, "step": 7265 }, { "epoch": 0.4923097770851684, "grad_norm": 7.363327980041504, "learning_rate": 9.108084057772607e-05, "loss": 0.763, "step": 7266 }, { "epoch": 0.49237753235314047, "grad_norm": 7.18175745010376, "learning_rate": 9.107947155862825e-05, "loss": 0.7809, "step": 7267 }, { "epoch": 0.49244528762111256, "grad_norm": 5.595573902130127, "learning_rate": 9.107810253953043e-05, "loss": 0.8471, "step": 7268 }, { "epoch": 0.49251304288908465, "grad_norm": 6.333422660827637, "learning_rate": 9.10767335204326e-05, "loss": 0.7616, "step": 7269 }, { "epoch": 0.4925807981570567, "grad_norm": 8.157796859741211, "learning_rate": 9.10753645013348e-05, "loss": 0.8233, "step": 7270 }, { "epoch": 0.4926485534250288, "grad_norm": 6.123366832733154, "learning_rate": 9.107399548223698e-05, "loss": 0.7266, "step": 7271 }, { "epoch": 0.49271630869300087, "grad_norm": 10.229715347290039, "learning_rate": 9.107262646313916e-05, "loss": 0.9589, "step": 7272 }, { "epoch": 0.49278406396097296, "grad_norm": 5.360836505889893, "learning_rate": 9.107125744404134e-05, "loss": 0.7874, "step": 7273 }, { "epoch": 0.49285181922894505, "grad_norm": 6.273800373077393, "learning_rate": 9.106988842494352e-05, "loss": 0.9512, "step": 7274 }, { "epoch": 0.49291957449691715, "grad_norm": 7.902069568634033, "learning_rate": 9.106851940584572e-05, "loss": 0.799, "step": 7275 }, { "epoch": 0.49298732976488924, "grad_norm": 9.35932731628418, "learning_rate": 9.10671503867479e-05, "loss": 0.8973, "step": 7276 }, { "epoch": 0.49305508503286133, "grad_norm": 7.119052410125732, "learning_rate": 9.106578136765008e-05, "loss": 1.2001, "step": 7277 }, { "epoch": 0.49312284030083336, "grad_norm": 6.111217498779297, "learning_rate": 9.106441234855227e-05, "loss": 0.7685, "step": 7278 }, { "epoch": 0.49319059556880546, "grad_norm": 6.093493461608887, "learning_rate": 9.106304332945445e-05, "loss": 0.7097, "step": 7279 }, { "epoch": 0.49325835083677755, "grad_norm": 6.722117900848389, "learning_rate": 9.106167431035663e-05, "loss": 0.805, "step": 7280 }, { "epoch": 0.49332610610474964, "grad_norm": 6.489585876464844, "learning_rate": 9.106030529125882e-05, "loss": 0.9748, "step": 7281 }, { "epoch": 0.49339386137272173, "grad_norm": 6.1473236083984375, "learning_rate": 9.1058936272161e-05, "loss": 0.8299, "step": 7282 }, { "epoch": 0.4934616166406938, "grad_norm": 7.472615718841553, "learning_rate": 9.105756725306319e-05, "loss": 1.0012, "step": 7283 }, { "epoch": 0.4935293719086659, "grad_norm": 7.1405463218688965, "learning_rate": 9.105619823396538e-05, "loss": 0.9959, "step": 7284 }, { "epoch": 0.493597127176638, "grad_norm": 6.21019172668457, "learning_rate": 9.105482921486756e-05, "loss": 0.867, "step": 7285 }, { "epoch": 0.49366488244461004, "grad_norm": 6.836954593658447, "learning_rate": 9.105346019576974e-05, "loss": 0.8058, "step": 7286 }, { "epoch": 0.49373263771258213, "grad_norm": 6.4547319412231445, "learning_rate": 9.105209117667192e-05, "loss": 0.8028, "step": 7287 }, { "epoch": 0.4938003929805542, "grad_norm": 6.335334300994873, "learning_rate": 9.10507221575741e-05, "loss": 0.763, "step": 7288 }, { "epoch": 0.4938681482485263, "grad_norm": 7.21290397644043, "learning_rate": 9.10493531384763e-05, "loss": 0.9477, "step": 7289 }, { "epoch": 0.4939359035164984, "grad_norm": 7.856054782867432, "learning_rate": 9.104798411937847e-05, "loss": 1.1117, "step": 7290 }, { "epoch": 0.4940036587844705, "grad_norm": 6.308975696563721, "learning_rate": 9.104661510028066e-05, "loss": 1.0397, "step": 7291 }, { "epoch": 0.4940714140524426, "grad_norm": 6.111830711364746, "learning_rate": 9.104524608118284e-05, "loss": 0.8082, "step": 7292 }, { "epoch": 0.4941391693204147, "grad_norm": 7.741870403289795, "learning_rate": 9.104387706208503e-05, "loss": 0.804, "step": 7293 }, { "epoch": 0.4942069245883867, "grad_norm": 7.447502613067627, "learning_rate": 9.104250804298721e-05, "loss": 0.9074, "step": 7294 }, { "epoch": 0.4942746798563588, "grad_norm": 4.931535243988037, "learning_rate": 9.104113902388939e-05, "loss": 0.7519, "step": 7295 }, { "epoch": 0.4943424351243309, "grad_norm": 9.830883026123047, "learning_rate": 9.103977000479157e-05, "loss": 0.8659, "step": 7296 }, { "epoch": 0.494410190392303, "grad_norm": 6.182522296905518, "learning_rate": 9.103840098569375e-05, "loss": 0.7021, "step": 7297 }, { "epoch": 0.4944779456602751, "grad_norm": 8.73188591003418, "learning_rate": 9.103703196659594e-05, "loss": 0.8496, "step": 7298 }, { "epoch": 0.4945457009282472, "grad_norm": 7.2238640785217285, "learning_rate": 9.103566294749813e-05, "loss": 1.0778, "step": 7299 }, { "epoch": 0.49461345619621927, "grad_norm": 6.101573944091797, "learning_rate": 9.10342939284003e-05, "loss": 0.9264, "step": 7300 }, { "epoch": 0.49468121146419136, "grad_norm": 7.327548503875732, "learning_rate": 9.103292490930249e-05, "loss": 0.7973, "step": 7301 }, { "epoch": 0.49474896673216345, "grad_norm": 7.1809186935424805, "learning_rate": 9.103155589020467e-05, "loss": 0.9586, "step": 7302 }, { "epoch": 0.4948167220001355, "grad_norm": 5.164478778839111, "learning_rate": 9.103018687110686e-05, "loss": 0.6197, "step": 7303 }, { "epoch": 0.4948844772681076, "grad_norm": 6.387687683105469, "learning_rate": 9.102881785200904e-05, "loss": 0.9509, "step": 7304 }, { "epoch": 0.4949522325360797, "grad_norm": 8.202160835266113, "learning_rate": 9.102744883291122e-05, "loss": 0.8557, "step": 7305 }, { "epoch": 0.49501998780405176, "grad_norm": 6.164126873016357, "learning_rate": 9.10260798138134e-05, "loss": 0.7898, "step": 7306 }, { "epoch": 0.49508774307202386, "grad_norm": 6.448176383972168, "learning_rate": 9.10247107947156e-05, "loss": 0.9094, "step": 7307 }, { "epoch": 0.49515549833999595, "grad_norm": 6.582845687866211, "learning_rate": 9.102334177561778e-05, "loss": 0.8778, "step": 7308 }, { "epoch": 0.49522325360796804, "grad_norm": 7.6885552406311035, "learning_rate": 9.102197275651996e-05, "loss": 1.0229, "step": 7309 }, { "epoch": 0.49529100887594013, "grad_norm": 6.840595722198486, "learning_rate": 9.102060373742214e-05, "loss": 0.972, "step": 7310 }, { "epoch": 0.49535876414391217, "grad_norm": 6.486509323120117, "learning_rate": 9.101923471832432e-05, "loss": 0.8788, "step": 7311 }, { "epoch": 0.49542651941188426, "grad_norm": 9.052862167358398, "learning_rate": 9.101786569922651e-05, "loss": 0.8917, "step": 7312 }, { "epoch": 0.49549427467985635, "grad_norm": 5.528444766998291, "learning_rate": 9.101649668012869e-05, "loss": 0.6167, "step": 7313 }, { "epoch": 0.49556202994782844, "grad_norm": 6.7654032707214355, "learning_rate": 9.101512766103087e-05, "loss": 0.9848, "step": 7314 }, { "epoch": 0.49562978521580053, "grad_norm": 6.247506618499756, "learning_rate": 9.101375864193305e-05, "loss": 0.5868, "step": 7315 }, { "epoch": 0.4956975404837726, "grad_norm": 6.155362129211426, "learning_rate": 9.101238962283525e-05, "loss": 0.8344, "step": 7316 }, { "epoch": 0.4957652957517447, "grad_norm": 8.219099998474121, "learning_rate": 9.101102060373743e-05, "loss": 0.8839, "step": 7317 }, { "epoch": 0.4958330510197168, "grad_norm": 5.146651744842529, "learning_rate": 9.10096515846396e-05, "loss": 0.7872, "step": 7318 }, { "epoch": 0.49590080628768884, "grad_norm": 7.989886283874512, "learning_rate": 9.100828256554179e-05, "loss": 1.0485, "step": 7319 }, { "epoch": 0.49596856155566094, "grad_norm": 7.214312553405762, "learning_rate": 9.100691354644397e-05, "loss": 0.9304, "step": 7320 }, { "epoch": 0.49603631682363303, "grad_norm": 6.856055736541748, "learning_rate": 9.100554452734616e-05, "loss": 0.8934, "step": 7321 }, { "epoch": 0.4961040720916051, "grad_norm": 6.771759986877441, "learning_rate": 9.100417550824834e-05, "loss": 0.9349, "step": 7322 }, { "epoch": 0.4961718273595772, "grad_norm": 8.19810962677002, "learning_rate": 9.100280648915052e-05, "loss": 1.0005, "step": 7323 }, { "epoch": 0.4962395826275493, "grad_norm": 6.946470737457275, "learning_rate": 9.100143747005271e-05, "loss": 0.7505, "step": 7324 }, { "epoch": 0.4963073378955214, "grad_norm": 6.6737518310546875, "learning_rate": 9.10000684509549e-05, "loss": 0.8722, "step": 7325 }, { "epoch": 0.4963750931634935, "grad_norm": 5.637021064758301, "learning_rate": 9.099869943185708e-05, "loss": 0.9587, "step": 7326 }, { "epoch": 0.4964428484314655, "grad_norm": 5.846226692199707, "learning_rate": 9.099733041275927e-05, "loss": 0.7726, "step": 7327 }, { "epoch": 0.4965106036994376, "grad_norm": 7.612580299377441, "learning_rate": 9.099596139366145e-05, "loss": 0.7203, "step": 7328 }, { "epoch": 0.4965783589674097, "grad_norm": 5.490561485290527, "learning_rate": 9.099459237456363e-05, "loss": 0.9457, "step": 7329 }, { "epoch": 0.4966461142353818, "grad_norm": 5.573283672332764, "learning_rate": 9.099322335546582e-05, "loss": 0.8116, "step": 7330 }, { "epoch": 0.4967138695033539, "grad_norm": 7.063168048858643, "learning_rate": 9.0991854336368e-05, "loss": 0.906, "step": 7331 }, { "epoch": 0.496781624771326, "grad_norm": 7.066551685333252, "learning_rate": 9.099048531727018e-05, "loss": 0.8515, "step": 7332 }, { "epoch": 0.49684938003929807, "grad_norm": 6.462795257568359, "learning_rate": 9.098911629817237e-05, "loss": 0.6325, "step": 7333 }, { "epoch": 0.49691713530727016, "grad_norm": 6.59752893447876, "learning_rate": 9.098774727907455e-05, "loss": 0.7356, "step": 7334 }, { "epoch": 0.4969848905752422, "grad_norm": 5.600460052490234, "learning_rate": 9.098637825997674e-05, "loss": 0.6307, "step": 7335 }, { "epoch": 0.4970526458432143, "grad_norm": 6.0045270919799805, "learning_rate": 9.098500924087892e-05, "loss": 1.0137, "step": 7336 }, { "epoch": 0.4971204011111864, "grad_norm": 7.3824143409729, "learning_rate": 9.09836402217811e-05, "loss": 0.8982, "step": 7337 }, { "epoch": 0.4971881563791585, "grad_norm": 6.573738098144531, "learning_rate": 9.098227120268328e-05, "loss": 0.7594, "step": 7338 }, { "epoch": 0.49725591164713057, "grad_norm": 8.101619720458984, "learning_rate": 9.098090218358547e-05, "loss": 1.1994, "step": 7339 }, { "epoch": 0.49732366691510266, "grad_norm": 5.76462459564209, "learning_rate": 9.097953316448765e-05, "loss": 0.7082, "step": 7340 }, { "epoch": 0.49739142218307475, "grad_norm": 6.358243465423584, "learning_rate": 9.097816414538983e-05, "loss": 0.8459, "step": 7341 }, { "epoch": 0.49745917745104684, "grad_norm": 7.145965576171875, "learning_rate": 9.097679512629202e-05, "loss": 0.9953, "step": 7342 }, { "epoch": 0.4975269327190189, "grad_norm": 8.23405933380127, "learning_rate": 9.09754261071942e-05, "loss": 1.1466, "step": 7343 }, { "epoch": 0.49759468798699097, "grad_norm": 6.5327982902526855, "learning_rate": 9.097405708809639e-05, "loss": 1.0226, "step": 7344 }, { "epoch": 0.49766244325496306, "grad_norm": 7.101400375366211, "learning_rate": 9.097268806899857e-05, "loss": 1.0476, "step": 7345 }, { "epoch": 0.49773019852293515, "grad_norm": 7.278493404388428, "learning_rate": 9.097131904990075e-05, "loss": 0.9499, "step": 7346 }, { "epoch": 0.49779795379090724, "grad_norm": 6.826780796051025, "learning_rate": 9.096995003080293e-05, "loss": 0.8371, "step": 7347 }, { "epoch": 0.49786570905887934, "grad_norm": 6.522684097290039, "learning_rate": 9.096858101170512e-05, "loss": 0.9318, "step": 7348 }, { "epoch": 0.4979334643268514, "grad_norm": 6.954566478729248, "learning_rate": 9.09672119926073e-05, "loss": 0.9174, "step": 7349 }, { "epoch": 0.4980012195948235, "grad_norm": 6.442493438720703, "learning_rate": 9.096584297350949e-05, "loss": 0.8619, "step": 7350 }, { "epoch": 0.49806897486279555, "grad_norm": 7.186161041259766, "learning_rate": 9.096447395441167e-05, "loss": 0.661, "step": 7351 }, { "epoch": 0.49813673013076765, "grad_norm": 6.716146945953369, "learning_rate": 9.096310493531385e-05, "loss": 0.9148, "step": 7352 }, { "epoch": 0.49820448539873974, "grad_norm": 5.643620014190674, "learning_rate": 9.096173591621604e-05, "loss": 0.8393, "step": 7353 }, { "epoch": 0.49827224066671183, "grad_norm": 7.107893466949463, "learning_rate": 9.096036689711822e-05, "loss": 0.972, "step": 7354 }, { "epoch": 0.4983399959346839, "grad_norm": 6.967519283294678, "learning_rate": 9.09589978780204e-05, "loss": 0.7033, "step": 7355 }, { "epoch": 0.498407751202656, "grad_norm": 7.291131496429443, "learning_rate": 9.095762885892258e-05, "loss": 0.8853, "step": 7356 }, { "epoch": 0.4984755064706281, "grad_norm": 7.630476474761963, "learning_rate": 9.095625983982476e-05, "loss": 0.8116, "step": 7357 }, { "epoch": 0.4985432617386002, "grad_norm": 6.2367167472839355, "learning_rate": 9.095489082072695e-05, "loss": 0.8344, "step": 7358 }, { "epoch": 0.4986110170065723, "grad_norm": 9.436936378479004, "learning_rate": 9.095352180162914e-05, "loss": 1.043, "step": 7359 }, { "epoch": 0.4986787722745443, "grad_norm": 5.330153942108154, "learning_rate": 9.095215278253132e-05, "loss": 0.7025, "step": 7360 }, { "epoch": 0.4987465275425164, "grad_norm": 5.117184162139893, "learning_rate": 9.09507837634335e-05, "loss": 0.8207, "step": 7361 }, { "epoch": 0.4988142828104885, "grad_norm": 7.965060234069824, "learning_rate": 9.094941474433569e-05, "loss": 0.9344, "step": 7362 }, { "epoch": 0.4988820380784606, "grad_norm": 6.985347747802734, "learning_rate": 9.094804572523787e-05, "loss": 0.8708, "step": 7363 }, { "epoch": 0.4989497933464327, "grad_norm": 8.380836486816406, "learning_rate": 9.094667670614005e-05, "loss": 0.8901, "step": 7364 }, { "epoch": 0.4990175486144048, "grad_norm": 6.336101055145264, "learning_rate": 9.094530768704223e-05, "loss": 0.9709, "step": 7365 }, { "epoch": 0.4990853038823769, "grad_norm": 6.435248851776123, "learning_rate": 9.094393866794441e-05, "loss": 0.8913, "step": 7366 }, { "epoch": 0.49915305915034897, "grad_norm": 7.674434661865234, "learning_rate": 9.09425696488466e-05, "loss": 1.0581, "step": 7367 }, { "epoch": 0.499220814418321, "grad_norm": 7.2185211181640625, "learning_rate": 9.094120062974879e-05, "loss": 0.9202, "step": 7368 }, { "epoch": 0.4992885696862931, "grad_norm": 6.717129230499268, "learning_rate": 9.093983161065097e-05, "loss": 0.8664, "step": 7369 }, { "epoch": 0.4993563249542652, "grad_norm": 6.229526996612549, "learning_rate": 9.093846259155316e-05, "loss": 0.7374, "step": 7370 }, { "epoch": 0.4994240802222373, "grad_norm": 7.177096843719482, "learning_rate": 9.093709357245534e-05, "loss": 1.0426, "step": 7371 }, { "epoch": 0.49949183549020937, "grad_norm": 7.289033889770508, "learning_rate": 9.093572455335752e-05, "loss": 0.8705, "step": 7372 }, { "epoch": 0.49955959075818146, "grad_norm": 8.260116577148438, "learning_rate": 9.093435553425971e-05, "loss": 0.6928, "step": 7373 }, { "epoch": 0.49962734602615355, "grad_norm": 5.730698108673096, "learning_rate": 9.09329865151619e-05, "loss": 0.7852, "step": 7374 }, { "epoch": 0.49969510129412564, "grad_norm": 6.90589714050293, "learning_rate": 9.093161749606407e-05, "loss": 0.9258, "step": 7375 }, { "epoch": 0.4997628565620977, "grad_norm": 5.358353137969971, "learning_rate": 9.093024847696627e-05, "loss": 0.7327, "step": 7376 }, { "epoch": 0.49983061183006977, "grad_norm": 5.4898505210876465, "learning_rate": 9.092887945786845e-05, "loss": 0.7371, "step": 7377 }, { "epoch": 0.49989836709804186, "grad_norm": 7.045073986053467, "learning_rate": 9.092751043877063e-05, "loss": 0.8785, "step": 7378 }, { "epoch": 0.49989836709804186, "eval_loss": 0.8431733846664429, "eval_noise_accuracy": 0.0, "eval_runtime": 1466.1466, "eval_samples_per_second": 3.505, "eval_steps_per_second": 0.22, "eval_wer": 75.89224292121845, "step": 7378 } ], "logging_steps": 1, "max_steps": 73795, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 3689, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.90621015064576e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }