diff --git "a/checkpoint-3689/trainer_state.json" "b/checkpoint-3689/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoint-3689/trainer_state.json" @@ -0,0 +1,25867 @@ +{ + "best_global_step": 3689, + "best_metric": 88.77438705594233, + "best_model_checkpoint": "whisper-tiny-bfloat16-sada/checkpoints/checkpoint-3689", + "epoch": 0.24994918354902093, + "eval_steps": 3689, + "global_step": 3689, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 6.775526797208483e-05, + "grad_norm": 132.70848083496094, + "learning_rate": 0.0, + "loss": 4.4561, + "step": 1 + }, + { + "epoch": 0.00013551053594416967, + "grad_norm": 133.01596069335938, + "learning_rate": 1.3333333333333334e-07, + "loss": 4.8642, + "step": 2 + }, + { + "epoch": 0.0002032658039162545, + "grad_norm": 106.03809356689453, + "learning_rate": 2.6666666666666667e-07, + "loss": 4.3778, + "step": 3 + }, + { + "epoch": 0.00027102107188833934, + "grad_norm": 88.75433349609375, + "learning_rate": 4.0000000000000003e-07, + "loss": 4.0385, + "step": 4 + }, + { + "epoch": 0.00033877633986042414, + "grad_norm": 101.19609832763672, + "learning_rate": 5.333333333333333e-07, + "loss": 3.9092, + "step": 5 + }, + { + "epoch": 0.000406531607832509, + "grad_norm": 67.02723693847656, + "learning_rate": 6.666666666666667e-07, + "loss": 3.7556, + "step": 6 + }, + { + "epoch": 0.0004742868758045938, + "grad_norm": 67.52012634277344, + "learning_rate": 8.000000000000001e-07, + "loss": 3.7119, + "step": 7 + }, + { + "epoch": 0.0005420421437766787, + "grad_norm": 83.27348327636719, + "learning_rate": 9.333333333333334e-07, + "loss": 3.6125, + "step": 8 + }, + { + "epoch": 0.0006097974117487635, + "grad_norm": 84.20647430419922, + "learning_rate": 1.0666666666666667e-06, + "loss": 4.1354, + "step": 9 + }, + { + "epoch": 0.0006775526797208483, + "grad_norm": 110.12113952636719, + "learning_rate": 1.2000000000000002e-06, + "loss": 3.9594, + "step": 10 + }, + { + "epoch": 0.0007453079476929331, + "grad_norm": 84.61614990234375, + "learning_rate": 1.3333333333333334e-06, + "loss": 4.1415, + "step": 11 + }, + { + "epoch": 0.000813063215665018, + "grad_norm": 119.0740966796875, + "learning_rate": 1.4666666666666667e-06, + "loss": 4.0689, + "step": 12 + }, + { + "epoch": 0.0008808184836371028, + "grad_norm": 122.41352081298828, + "learning_rate": 1.6000000000000001e-06, + "loss": 4.0834, + "step": 13 + }, + { + "epoch": 0.0009485737516091876, + "grad_norm": 99.06126403808594, + "learning_rate": 1.7333333333333334e-06, + "loss": 4.4224, + "step": 14 + }, + { + "epoch": 0.0010163290195812724, + "grad_norm": 113.89531707763672, + "learning_rate": 1.8666666666666669e-06, + "loss": 4.0188, + "step": 15 + }, + { + "epoch": 0.0010840842875533573, + "grad_norm": 70.24198150634766, + "learning_rate": 2.0000000000000003e-06, + "loss": 3.3951, + "step": 16 + }, + { + "epoch": 0.001151839555525442, + "grad_norm": 67.99130249023438, + "learning_rate": 2.1333333333333334e-06, + "loss": 3.7385, + "step": 17 + }, + { + "epoch": 0.001219594823497527, + "grad_norm": 54.554779052734375, + "learning_rate": 2.266666666666667e-06, + "loss": 2.9009, + "step": 18 + }, + { + "epoch": 0.0012873500914696117, + "grad_norm": 47.79149627685547, + "learning_rate": 2.4000000000000003e-06, + "loss": 3.0572, + "step": 19 + }, + { + "epoch": 0.0013551053594416966, + "grad_norm": 50.05776596069336, + "learning_rate": 2.5333333333333334e-06, + "loss": 2.8678, + "step": 20 + }, + { + "epoch": 0.0014228606274137815, + "grad_norm": 55.41701889038086, + "learning_rate": 2.666666666666667e-06, + "loss": 3.3981, + "step": 21 + }, + { + "epoch": 0.0014906158953858662, + "grad_norm": 57.1912727355957, + "learning_rate": 2.8000000000000003e-06, + "loss": 3.3247, + "step": 22 + }, + { + "epoch": 0.001558371163357951, + "grad_norm": 39.77119445800781, + "learning_rate": 2.9333333333333333e-06, + "loss": 3.1466, + "step": 23 + }, + { + "epoch": 0.001626126431330036, + "grad_norm": 37.258392333984375, + "learning_rate": 3.066666666666667e-06, + "loss": 3.0137, + "step": 24 + }, + { + "epoch": 0.0016938816993021207, + "grad_norm": 41.15428924560547, + "learning_rate": 3.2000000000000003e-06, + "loss": 2.9908, + "step": 25 + }, + { + "epoch": 0.0017616369672742056, + "grad_norm": 44.59917449951172, + "learning_rate": 3.3333333333333333e-06, + "loss": 3.6947, + "step": 26 + }, + { + "epoch": 0.0018293922352462903, + "grad_norm": 32.7044792175293, + "learning_rate": 3.466666666666667e-06, + "loss": 2.9421, + "step": 27 + }, + { + "epoch": 0.0018971475032183752, + "grad_norm": 28.445444107055664, + "learning_rate": 3.6e-06, + "loss": 2.9849, + "step": 28 + }, + { + "epoch": 0.00196490277119046, + "grad_norm": 30.879247665405273, + "learning_rate": 3.7333333333333337e-06, + "loss": 2.9007, + "step": 29 + }, + { + "epoch": 0.002032658039162545, + "grad_norm": 22.723188400268555, + "learning_rate": 3.866666666666667e-06, + "loss": 2.3549, + "step": 30 + }, + { + "epoch": 0.0021004133071346296, + "grad_norm": 23.59457778930664, + "learning_rate": 4.000000000000001e-06, + "loss": 2.8564, + "step": 31 + }, + { + "epoch": 0.0021681685751067147, + "grad_norm": 23.313579559326172, + "learning_rate": 4.133333333333333e-06, + "loss": 2.453, + "step": 32 + }, + { + "epoch": 0.0022359238430787994, + "grad_norm": 23.877405166625977, + "learning_rate": 4.266666666666667e-06, + "loss": 2.2042, + "step": 33 + }, + { + "epoch": 0.002303679111050884, + "grad_norm": 29.79376792907715, + "learning_rate": 4.4e-06, + "loss": 2.6679, + "step": 34 + }, + { + "epoch": 0.002371434379022969, + "grad_norm": 21.018878936767578, + "learning_rate": 4.533333333333334e-06, + "loss": 2.3983, + "step": 35 + }, + { + "epoch": 0.002439189646995054, + "grad_norm": 30.22182846069336, + "learning_rate": 4.666666666666667e-06, + "loss": 2.8668, + "step": 36 + }, + { + "epoch": 0.0025069449149671386, + "grad_norm": 24.619905471801758, + "learning_rate": 4.800000000000001e-06, + "loss": 2.5986, + "step": 37 + }, + { + "epoch": 0.0025747001829392233, + "grad_norm": 25.470375061035156, + "learning_rate": 4.933333333333333e-06, + "loss": 2.4936, + "step": 38 + }, + { + "epoch": 0.0026424554509113084, + "grad_norm": 19.872745513916016, + "learning_rate": 5.066666666666667e-06, + "loss": 2.1633, + "step": 39 + }, + { + "epoch": 0.002710210718883393, + "grad_norm": 24.880964279174805, + "learning_rate": 5.2e-06, + "loss": 2.6798, + "step": 40 + }, + { + "epoch": 0.002777965986855478, + "grad_norm": 22.790157318115234, + "learning_rate": 5.333333333333334e-06, + "loss": 2.487, + "step": 41 + }, + { + "epoch": 0.002845721254827563, + "grad_norm": 21.404136657714844, + "learning_rate": 5.466666666666667e-06, + "loss": 2.3383, + "step": 42 + }, + { + "epoch": 0.0029134765227996477, + "grad_norm": 22.80590057373047, + "learning_rate": 5.600000000000001e-06, + "loss": 2.6347, + "step": 43 + }, + { + "epoch": 0.0029812317907717324, + "grad_norm": 20.625160217285156, + "learning_rate": 5.733333333333333e-06, + "loss": 2.066, + "step": 44 + }, + { + "epoch": 0.0030489870587438175, + "grad_norm": 21.700828552246094, + "learning_rate": 5.866666666666667e-06, + "loss": 2.188, + "step": 45 + }, + { + "epoch": 0.003116742326715902, + "grad_norm": 18.8945255279541, + "learning_rate": 6e-06, + "loss": 2.1531, + "step": 46 + }, + { + "epoch": 0.003184497594687987, + "grad_norm": 20.238525390625, + "learning_rate": 6.133333333333334e-06, + "loss": 1.9306, + "step": 47 + }, + { + "epoch": 0.003252252862660072, + "grad_norm": 23.74344825744629, + "learning_rate": 6.266666666666666e-06, + "loss": 2.3206, + "step": 48 + }, + { + "epoch": 0.0033200081306321567, + "grad_norm": 22.429964065551758, + "learning_rate": 6.4000000000000006e-06, + "loss": 2.05, + "step": 49 + }, + { + "epoch": 0.0033877633986042414, + "grad_norm": 22.078792572021484, + "learning_rate": 6.533333333333333e-06, + "loss": 2.1155, + "step": 50 + }, + { + "epoch": 0.003455518666576326, + "grad_norm": 19.691747665405273, + "learning_rate": 6.666666666666667e-06, + "loss": 1.8719, + "step": 51 + }, + { + "epoch": 0.0035232739345484113, + "grad_norm": 27.412866592407227, + "learning_rate": 6.800000000000001e-06, + "loss": 2.0943, + "step": 52 + }, + { + "epoch": 0.003591029202520496, + "grad_norm": 21.631580352783203, + "learning_rate": 6.933333333333334e-06, + "loss": 2.246, + "step": 53 + }, + { + "epoch": 0.0036587844704925807, + "grad_norm": 16.729095458984375, + "learning_rate": 7.066666666666667e-06, + "loss": 1.8101, + "step": 54 + }, + { + "epoch": 0.003726539738464666, + "grad_norm": 18.381364822387695, + "learning_rate": 7.2e-06, + "loss": 1.8668, + "step": 55 + }, + { + "epoch": 0.0037942950064367505, + "grad_norm": 19.661123275756836, + "learning_rate": 7.333333333333334e-06, + "loss": 2.1912, + "step": 56 + }, + { + "epoch": 0.003862050274408835, + "grad_norm": 13.772311210632324, + "learning_rate": 7.4666666666666675e-06, + "loss": 1.7408, + "step": 57 + }, + { + "epoch": 0.00392980554238092, + "grad_norm": 20.559425354003906, + "learning_rate": 7.6e-06, + "loss": 1.9717, + "step": 58 + }, + { + "epoch": 0.003997560810353005, + "grad_norm": 17.385791778564453, + "learning_rate": 7.733333333333334e-06, + "loss": 1.7681, + "step": 59 + }, + { + "epoch": 0.00406531607832509, + "grad_norm": 18.455888748168945, + "learning_rate": 7.866666666666667e-06, + "loss": 1.8822, + "step": 60 + }, + { + "epoch": 0.004133071346297174, + "grad_norm": 17.53643798828125, + "learning_rate": 8.000000000000001e-06, + "loss": 1.8158, + "step": 61 + }, + { + "epoch": 0.004200826614269259, + "grad_norm": 15.640396118164062, + "learning_rate": 8.133333333333332e-06, + "loss": 1.8786, + "step": 62 + }, + { + "epoch": 0.004268581882241345, + "grad_norm": 67.14771270751953, + "learning_rate": 8.266666666666667e-06, + "loss": 1.845, + "step": 63 + }, + { + "epoch": 0.004336337150213429, + "grad_norm": 16.778207778930664, + "learning_rate": 8.400000000000001e-06, + "loss": 1.7286, + "step": 64 + }, + { + "epoch": 0.004404092418185514, + "grad_norm": 20.485218048095703, + "learning_rate": 8.533333333333334e-06, + "loss": 1.882, + "step": 65 + }, + { + "epoch": 0.004471847686157599, + "grad_norm": 15.77828598022461, + "learning_rate": 8.666666666666668e-06, + "loss": 1.7788, + "step": 66 + }, + { + "epoch": 0.0045396029541296835, + "grad_norm": 20.225608825683594, + "learning_rate": 8.8e-06, + "loss": 1.9842, + "step": 67 + }, + { + "epoch": 0.004607358222101768, + "grad_norm": 16.230987548828125, + "learning_rate": 8.933333333333333e-06, + "loss": 1.4581, + "step": 68 + }, + { + "epoch": 0.004675113490073853, + "grad_norm": 13.831804275512695, + "learning_rate": 9.066666666666667e-06, + "loss": 1.6173, + "step": 69 + }, + { + "epoch": 0.004742868758045938, + "grad_norm": 16.4691104888916, + "learning_rate": 9.2e-06, + "loss": 1.6055, + "step": 70 + }, + { + "epoch": 0.004810624026018023, + "grad_norm": 17.788105010986328, + "learning_rate": 9.333333333333334e-06, + "loss": 1.7996, + "step": 71 + }, + { + "epoch": 0.004878379293990108, + "grad_norm": 20.043182373046875, + "learning_rate": 9.466666666666667e-06, + "loss": 1.7224, + "step": 72 + }, + { + "epoch": 0.0049461345619621925, + "grad_norm": 14.918837547302246, + "learning_rate": 9.600000000000001e-06, + "loss": 1.4448, + "step": 73 + }, + { + "epoch": 0.005013889829934277, + "grad_norm": 15.70170783996582, + "learning_rate": 9.733333333333334e-06, + "loss": 1.7469, + "step": 74 + }, + { + "epoch": 0.005081645097906362, + "grad_norm": 19.128456115722656, + "learning_rate": 9.866666666666667e-06, + "loss": 1.9505, + "step": 75 + }, + { + "epoch": 0.005149400365878447, + "grad_norm": 34.82925796508789, + "learning_rate": 1e-05, + "loss": 1.6543, + "step": 76 + }, + { + "epoch": 0.005217155633850532, + "grad_norm": 19.61214256286621, + "learning_rate": 1.0133333333333333e-05, + "loss": 1.9051, + "step": 77 + }, + { + "epoch": 0.005284910901822617, + "grad_norm": 20.190847396850586, + "learning_rate": 1.0266666666666668e-05, + "loss": 1.6016, + "step": 78 + }, + { + "epoch": 0.005352666169794702, + "grad_norm": 15.873608589172363, + "learning_rate": 1.04e-05, + "loss": 1.8286, + "step": 79 + }, + { + "epoch": 0.005420421437766786, + "grad_norm": 17.513385772705078, + "learning_rate": 1.0533333333333335e-05, + "loss": 1.7122, + "step": 80 + }, + { + "epoch": 0.005488176705738871, + "grad_norm": 13.794294357299805, + "learning_rate": 1.0666666666666667e-05, + "loss": 1.5369, + "step": 81 + }, + { + "epoch": 0.005555931973710956, + "grad_norm": 16.763822555541992, + "learning_rate": 1.08e-05, + "loss": 1.688, + "step": 82 + }, + { + "epoch": 0.005623687241683041, + "grad_norm": 15.182608604431152, + "learning_rate": 1.0933333333333334e-05, + "loss": 1.9076, + "step": 83 + }, + { + "epoch": 0.005691442509655126, + "grad_norm": 19.11581802368164, + "learning_rate": 1.1066666666666667e-05, + "loss": 1.6034, + "step": 84 + }, + { + "epoch": 0.005759197777627211, + "grad_norm": 18.303098678588867, + "learning_rate": 1.1200000000000001e-05, + "loss": 1.8248, + "step": 85 + }, + { + "epoch": 0.005826953045599295, + "grad_norm": 17.912731170654297, + "learning_rate": 1.1333333333333334e-05, + "loss": 1.737, + "step": 86 + }, + { + "epoch": 0.00589470831357138, + "grad_norm": 18.4353084564209, + "learning_rate": 1.1466666666666666e-05, + "loss": 1.7141, + "step": 87 + }, + { + "epoch": 0.005962463581543465, + "grad_norm": 17.006074905395508, + "learning_rate": 1.16e-05, + "loss": 1.6568, + "step": 88 + }, + { + "epoch": 0.0060302188495155494, + "grad_norm": 15.084410667419434, + "learning_rate": 1.1733333333333333e-05, + "loss": 1.3369, + "step": 89 + }, + { + "epoch": 0.006097974117487635, + "grad_norm": 14.984724044799805, + "learning_rate": 1.1866666666666668e-05, + "loss": 1.777, + "step": 90 + }, + { + "epoch": 0.00616572938545972, + "grad_norm": 19.447574615478516, + "learning_rate": 1.2e-05, + "loss": 1.4225, + "step": 91 + }, + { + "epoch": 0.006233484653431804, + "grad_norm": 15.212474822998047, + "learning_rate": 1.2133333333333335e-05, + "loss": 1.4896, + "step": 92 + }, + { + "epoch": 0.006301239921403889, + "grad_norm": 17.97924041748047, + "learning_rate": 1.2266666666666667e-05, + "loss": 1.5947, + "step": 93 + }, + { + "epoch": 0.006368995189375974, + "grad_norm": 13.695365905761719, + "learning_rate": 1.24e-05, + "loss": 1.471, + "step": 94 + }, + { + "epoch": 0.0064367504573480585, + "grad_norm": 15.574562072753906, + "learning_rate": 1.2533333333333332e-05, + "loss": 1.5803, + "step": 95 + }, + { + "epoch": 0.006504505725320144, + "grad_norm": 16.71872329711914, + "learning_rate": 1.2666666666666668e-05, + "loss": 1.7408, + "step": 96 + }, + { + "epoch": 0.006572260993292229, + "grad_norm": 14.232036590576172, + "learning_rate": 1.2800000000000001e-05, + "loss": 1.4414, + "step": 97 + }, + { + "epoch": 0.0066400162612643135, + "grad_norm": 14.430774688720703, + "learning_rate": 1.2933333333333334e-05, + "loss": 1.712, + "step": 98 + }, + { + "epoch": 0.006707771529236398, + "grad_norm": 15.887603759765625, + "learning_rate": 1.3066666666666666e-05, + "loss": 1.6097, + "step": 99 + }, + { + "epoch": 0.006775526797208483, + "grad_norm": 21.155643463134766, + "learning_rate": 1.32e-05, + "loss": 2.0581, + "step": 100 + }, + { + "epoch": 0.0068432820651805676, + "grad_norm": 17.30326271057129, + "learning_rate": 1.3333333333333333e-05, + "loss": 1.2823, + "step": 101 + }, + { + "epoch": 0.006911037333152652, + "grad_norm": 18.184707641601562, + "learning_rate": 1.3466666666666666e-05, + "loss": 1.5165, + "step": 102 + }, + { + "epoch": 0.006978792601124738, + "grad_norm": 23.46563148498535, + "learning_rate": 1.3600000000000002e-05, + "loss": 1.406, + "step": 103 + }, + { + "epoch": 0.0070465478690968225, + "grad_norm": 14.586087226867676, + "learning_rate": 1.3733333333333335e-05, + "loss": 1.8416, + "step": 104 + }, + { + "epoch": 0.007114303137068907, + "grad_norm": 14.55811595916748, + "learning_rate": 1.3866666666666667e-05, + "loss": 1.5466, + "step": 105 + }, + { + "epoch": 0.007182058405040992, + "grad_norm": 20.244586944580078, + "learning_rate": 1.4000000000000001e-05, + "loss": 1.755, + "step": 106 + }, + { + "epoch": 0.007249813673013077, + "grad_norm": 11.932172775268555, + "learning_rate": 1.4133333333333334e-05, + "loss": 1.4543, + "step": 107 + }, + { + "epoch": 0.007317568940985161, + "grad_norm": 17.406888961791992, + "learning_rate": 1.4266666666666667e-05, + "loss": 1.4436, + "step": 108 + }, + { + "epoch": 0.007385324208957246, + "grad_norm": 14.591705322265625, + "learning_rate": 1.44e-05, + "loss": 1.4088, + "step": 109 + }, + { + "epoch": 0.007453079476929332, + "grad_norm": 18.028337478637695, + "learning_rate": 1.4533333333333335e-05, + "loss": 1.6675, + "step": 110 + }, + { + "epoch": 0.007520834744901416, + "grad_norm": 16.37979507446289, + "learning_rate": 1.4666666666666668e-05, + "loss": 1.6646, + "step": 111 + }, + { + "epoch": 0.007588590012873501, + "grad_norm": 19.028308868408203, + "learning_rate": 1.48e-05, + "loss": 1.701, + "step": 112 + }, + { + "epoch": 0.007656345280845586, + "grad_norm": 17.28181266784668, + "learning_rate": 1.4933333333333335e-05, + "loss": 1.3984, + "step": 113 + }, + { + "epoch": 0.00772410054881767, + "grad_norm": 16.78407859802246, + "learning_rate": 1.5066666666666668e-05, + "loss": 1.6851, + "step": 114 + }, + { + "epoch": 0.007791855816789755, + "grad_norm": 15.634642601013184, + "learning_rate": 1.52e-05, + "loss": 1.6484, + "step": 115 + }, + { + "epoch": 0.00785961108476184, + "grad_norm": 13.879626274108887, + "learning_rate": 1.5333333333333334e-05, + "loss": 1.4609, + "step": 116 + }, + { + "epoch": 0.007927366352733925, + "grad_norm": 16.247541427612305, + "learning_rate": 1.546666666666667e-05, + "loss": 1.5219, + "step": 117 + }, + { + "epoch": 0.00799512162070601, + "grad_norm": 20.57497215270996, + "learning_rate": 1.56e-05, + "loss": 1.6038, + "step": 118 + }, + { + "epoch": 0.008062876888678095, + "grad_norm": 16.32591438293457, + "learning_rate": 1.5733333333333334e-05, + "loss": 1.5315, + "step": 119 + }, + { + "epoch": 0.00813063215665018, + "grad_norm": 18.236255645751953, + "learning_rate": 1.586666666666667e-05, + "loss": 1.6589, + "step": 120 + }, + { + "epoch": 0.008198387424622264, + "grad_norm": 20.850723266601562, + "learning_rate": 1.6000000000000003e-05, + "loss": 1.5556, + "step": 121 + }, + { + "epoch": 0.008266142692594349, + "grad_norm": 17.931947708129883, + "learning_rate": 1.6133333333333334e-05, + "loss": 1.8209, + "step": 122 + }, + { + "epoch": 0.008333897960566434, + "grad_norm": 20.531381607055664, + "learning_rate": 1.6266666666666665e-05, + "loss": 1.6421, + "step": 123 + }, + { + "epoch": 0.008401653228538518, + "grad_norm": 15.6729736328125, + "learning_rate": 1.6400000000000002e-05, + "loss": 1.5652, + "step": 124 + }, + { + "epoch": 0.008469408496510603, + "grad_norm": 18.139949798583984, + "learning_rate": 1.6533333333333333e-05, + "loss": 1.5578, + "step": 125 + }, + { + "epoch": 0.00853716376448269, + "grad_norm": 15.77291202545166, + "learning_rate": 1.6666666666666667e-05, + "loss": 1.6091, + "step": 126 + }, + { + "epoch": 0.008604919032454774, + "grad_norm": 18.00165557861328, + "learning_rate": 1.6800000000000002e-05, + "loss": 1.4725, + "step": 127 + }, + { + "epoch": 0.008672674300426859, + "grad_norm": 17.33936309814453, + "learning_rate": 1.6933333333333333e-05, + "loss": 1.4272, + "step": 128 + }, + { + "epoch": 0.008740429568398943, + "grad_norm": 16.75677490234375, + "learning_rate": 1.7066666666666667e-05, + "loss": 1.602, + "step": 129 + }, + { + "epoch": 0.008808184836371028, + "grad_norm": 17.761140823364258, + "learning_rate": 1.7199999999999998e-05, + "loss": 1.8523, + "step": 130 + }, + { + "epoch": 0.008875940104343113, + "grad_norm": 15.888497352600098, + "learning_rate": 1.7333333333333336e-05, + "loss": 1.5594, + "step": 131 + }, + { + "epoch": 0.008943695372315198, + "grad_norm": 19.62826156616211, + "learning_rate": 1.7466666666666667e-05, + "loss": 1.6909, + "step": 132 + }, + { + "epoch": 0.009011450640287282, + "grad_norm": 15.62592887878418, + "learning_rate": 1.76e-05, + "loss": 1.5873, + "step": 133 + }, + { + "epoch": 0.009079205908259367, + "grad_norm": 17.64893913269043, + "learning_rate": 1.7733333333333335e-05, + "loss": 1.5681, + "step": 134 + }, + { + "epoch": 0.009146961176231452, + "grad_norm": 15.14854907989502, + "learning_rate": 1.7866666666666666e-05, + "loss": 1.3324, + "step": 135 + }, + { + "epoch": 0.009214716444203536, + "grad_norm": 18.115419387817383, + "learning_rate": 1.8e-05, + "loss": 1.5512, + "step": 136 + }, + { + "epoch": 0.009282471712175621, + "grad_norm": 18.2511043548584, + "learning_rate": 1.8133333333333335e-05, + "loss": 1.6405, + "step": 137 + }, + { + "epoch": 0.009350226980147706, + "grad_norm": 13.296855926513672, + "learning_rate": 1.826666666666667e-05, + "loss": 1.2269, + "step": 138 + }, + { + "epoch": 0.009417982248119792, + "grad_norm": 17.17320442199707, + "learning_rate": 1.84e-05, + "loss": 1.4031, + "step": 139 + }, + { + "epoch": 0.009485737516091877, + "grad_norm": 15.399650573730469, + "learning_rate": 1.8533333333333334e-05, + "loss": 1.4793, + "step": 140 + }, + { + "epoch": 0.009553492784063962, + "grad_norm": 15.0385160446167, + "learning_rate": 1.866666666666667e-05, + "loss": 1.4129, + "step": 141 + }, + { + "epoch": 0.009621248052036046, + "grad_norm": 14.586984634399414, + "learning_rate": 1.88e-05, + "loss": 1.1959, + "step": 142 + }, + { + "epoch": 0.009689003320008131, + "grad_norm": 14.939247131347656, + "learning_rate": 1.8933333333333334e-05, + "loss": 1.6728, + "step": 143 + }, + { + "epoch": 0.009756758587980216, + "grad_norm": 18.1126651763916, + "learning_rate": 1.9066666666666668e-05, + "loss": 1.5035, + "step": 144 + }, + { + "epoch": 0.0098245138559523, + "grad_norm": 16.0823974609375, + "learning_rate": 1.9200000000000003e-05, + "loss": 1.544, + "step": 145 + }, + { + "epoch": 0.009892269123924385, + "grad_norm": 17.45062828063965, + "learning_rate": 1.9333333333333333e-05, + "loss": 1.4593, + "step": 146 + }, + { + "epoch": 0.00996002439189647, + "grad_norm": 21.030731201171875, + "learning_rate": 1.9466666666666668e-05, + "loss": 1.6544, + "step": 147 + }, + { + "epoch": 0.010027779659868554, + "grad_norm": 18.31300163269043, + "learning_rate": 1.9600000000000002e-05, + "loss": 1.7682, + "step": 148 + }, + { + "epoch": 0.01009553492784064, + "grad_norm": 13.541064262390137, + "learning_rate": 1.9733333333333333e-05, + "loss": 1.333, + "step": 149 + }, + { + "epoch": 0.010163290195812724, + "grad_norm": 13.538237571716309, + "learning_rate": 1.9866666666666667e-05, + "loss": 1.3629, + "step": 150 + }, + { + "epoch": 0.010231045463784809, + "grad_norm": 18.393657684326172, + "learning_rate": 2e-05, + "loss": 1.4467, + "step": 151 + }, + { + "epoch": 0.010298800731756893, + "grad_norm": 19.18985939025879, + "learning_rate": 2.0133333333333336e-05, + "loss": 1.612, + "step": 152 + }, + { + "epoch": 0.01036655599972898, + "grad_norm": 14.230488777160645, + "learning_rate": 2.0266666666666667e-05, + "loss": 1.4047, + "step": 153 + }, + { + "epoch": 0.010434311267701064, + "grad_norm": 13.807863235473633, + "learning_rate": 2.04e-05, + "loss": 1.4977, + "step": 154 + }, + { + "epoch": 0.010502066535673149, + "grad_norm": 15.149568557739258, + "learning_rate": 2.0533333333333336e-05, + "loss": 1.3856, + "step": 155 + }, + { + "epoch": 0.010569821803645234, + "grad_norm": 16.095308303833008, + "learning_rate": 2.0666666666666666e-05, + "loss": 1.3687, + "step": 156 + }, + { + "epoch": 0.010637577071617318, + "grad_norm": 14.864961624145508, + "learning_rate": 2.08e-05, + "loss": 1.5707, + "step": 157 + }, + { + "epoch": 0.010705332339589403, + "grad_norm": 15.9810152053833, + "learning_rate": 2.0933333333333335e-05, + "loss": 1.6841, + "step": 158 + }, + { + "epoch": 0.010773087607561488, + "grad_norm": 16.706584930419922, + "learning_rate": 2.106666666666667e-05, + "loss": 1.5017, + "step": 159 + }, + { + "epoch": 0.010840842875533573, + "grad_norm": 18.726655960083008, + "learning_rate": 2.12e-05, + "loss": 1.7177, + "step": 160 + }, + { + "epoch": 0.010908598143505657, + "grad_norm": 17.22323989868164, + "learning_rate": 2.1333333333333335e-05, + "loss": 1.5452, + "step": 161 + }, + { + "epoch": 0.010976353411477742, + "grad_norm": 14.823019027709961, + "learning_rate": 2.146666666666667e-05, + "loss": 1.2856, + "step": 162 + }, + { + "epoch": 0.011044108679449827, + "grad_norm": 14.756085395812988, + "learning_rate": 2.16e-05, + "loss": 1.4942, + "step": 163 + }, + { + "epoch": 0.011111863947421911, + "grad_norm": 18.523683547973633, + "learning_rate": 2.1733333333333334e-05, + "loss": 1.4832, + "step": 164 + }, + { + "epoch": 0.011179619215393996, + "grad_norm": 16.75520896911621, + "learning_rate": 2.186666666666667e-05, + "loss": 1.1847, + "step": 165 + }, + { + "epoch": 0.011247374483366082, + "grad_norm": 17.045665740966797, + "learning_rate": 2.2000000000000003e-05, + "loss": 1.6358, + "step": 166 + }, + { + "epoch": 0.011315129751338167, + "grad_norm": 17.235095977783203, + "learning_rate": 2.2133333333333334e-05, + "loss": 1.6984, + "step": 167 + }, + { + "epoch": 0.011382885019310252, + "grad_norm": 18.120485305786133, + "learning_rate": 2.2266666666666668e-05, + "loss": 1.5544, + "step": 168 + }, + { + "epoch": 0.011450640287282337, + "grad_norm": 15.252264976501465, + "learning_rate": 2.2400000000000002e-05, + "loss": 1.4386, + "step": 169 + }, + { + "epoch": 0.011518395555254421, + "grad_norm": 13.91727066040039, + "learning_rate": 2.2533333333333333e-05, + "loss": 1.2083, + "step": 170 + }, + { + "epoch": 0.011586150823226506, + "grad_norm": 16.50945472717285, + "learning_rate": 2.2666666666666668e-05, + "loss": 1.5878, + "step": 171 + }, + { + "epoch": 0.01165390609119859, + "grad_norm": 12.260165214538574, + "learning_rate": 2.2800000000000002e-05, + "loss": 1.4173, + "step": 172 + }, + { + "epoch": 0.011721661359170675, + "grad_norm": 17.12419891357422, + "learning_rate": 2.2933333333333333e-05, + "loss": 1.3117, + "step": 173 + }, + { + "epoch": 0.01178941662714276, + "grad_norm": 17.508407592773438, + "learning_rate": 2.3066666666666667e-05, + "loss": 1.4294, + "step": 174 + }, + { + "epoch": 0.011857171895114845, + "grad_norm": 17.78769302368164, + "learning_rate": 2.32e-05, + "loss": 1.8437, + "step": 175 + }, + { + "epoch": 0.01192492716308693, + "grad_norm": 13.79132080078125, + "learning_rate": 2.3333333333333336e-05, + "loss": 1.3969, + "step": 176 + }, + { + "epoch": 0.011992682431059014, + "grad_norm": 16.519269943237305, + "learning_rate": 2.3466666666666667e-05, + "loss": 1.5833, + "step": 177 + }, + { + "epoch": 0.012060437699031099, + "grad_norm": 15.920942306518555, + "learning_rate": 2.36e-05, + "loss": 1.4783, + "step": 178 + }, + { + "epoch": 0.012128192967003185, + "grad_norm": 17.00782585144043, + "learning_rate": 2.3733333333333335e-05, + "loss": 1.4038, + "step": 179 + }, + { + "epoch": 0.01219594823497527, + "grad_norm": 18.844655990600586, + "learning_rate": 2.3866666666666666e-05, + "loss": 1.653, + "step": 180 + }, + { + "epoch": 0.012263703502947355, + "grad_norm": 18.05002784729004, + "learning_rate": 2.4e-05, + "loss": 1.4137, + "step": 181 + }, + { + "epoch": 0.01233145877091944, + "grad_norm": 14.200697898864746, + "learning_rate": 2.4133333333333335e-05, + "loss": 1.3067, + "step": 182 + }, + { + "epoch": 0.012399214038891524, + "grad_norm": 16.28152084350586, + "learning_rate": 2.426666666666667e-05, + "loss": 1.3425, + "step": 183 + }, + { + "epoch": 0.012466969306863609, + "grad_norm": 18.691898345947266, + "learning_rate": 2.44e-05, + "loss": 1.5995, + "step": 184 + }, + { + "epoch": 0.012534724574835693, + "grad_norm": 16.62773895263672, + "learning_rate": 2.4533333333333334e-05, + "loss": 1.5722, + "step": 185 + }, + { + "epoch": 0.012602479842807778, + "grad_norm": 19.650226593017578, + "learning_rate": 2.466666666666667e-05, + "loss": 1.5445, + "step": 186 + }, + { + "epoch": 0.012670235110779863, + "grad_norm": 14.582402229309082, + "learning_rate": 2.48e-05, + "loss": 1.4822, + "step": 187 + }, + { + "epoch": 0.012737990378751948, + "grad_norm": 16.5295352935791, + "learning_rate": 2.4933333333333334e-05, + "loss": 1.5981, + "step": 188 + }, + { + "epoch": 0.012805745646724032, + "grad_norm": 17.096031188964844, + "learning_rate": 2.5066666666666665e-05, + "loss": 1.2527, + "step": 189 + }, + { + "epoch": 0.012873500914696117, + "grad_norm": 15.424555778503418, + "learning_rate": 2.5200000000000003e-05, + "loss": 1.5284, + "step": 190 + }, + { + "epoch": 0.012941256182668202, + "grad_norm": 14.505122184753418, + "learning_rate": 2.5333333333333337e-05, + "loss": 1.5726, + "step": 191 + }, + { + "epoch": 0.013009011450640288, + "grad_norm": 13.430007934570312, + "learning_rate": 2.5466666666666668e-05, + "loss": 1.5905, + "step": 192 + }, + { + "epoch": 0.013076766718612373, + "grad_norm": 12.291808128356934, + "learning_rate": 2.5600000000000002e-05, + "loss": 1.3122, + "step": 193 + }, + { + "epoch": 0.013144521986584458, + "grad_norm": 16.053747177124023, + "learning_rate": 2.5733333333333337e-05, + "loss": 1.3164, + "step": 194 + }, + { + "epoch": 0.013212277254556542, + "grad_norm": 17.762910842895508, + "learning_rate": 2.5866666666666667e-05, + "loss": 1.4727, + "step": 195 + }, + { + "epoch": 0.013280032522528627, + "grad_norm": 15.695332527160645, + "learning_rate": 2.6000000000000002e-05, + "loss": 1.5584, + "step": 196 + }, + { + "epoch": 0.013347787790500712, + "grad_norm": 14.789410591125488, + "learning_rate": 2.6133333333333333e-05, + "loss": 1.3253, + "step": 197 + }, + { + "epoch": 0.013415543058472796, + "grad_norm": 15.248231887817383, + "learning_rate": 2.6266666666666667e-05, + "loss": 1.3087, + "step": 198 + }, + { + "epoch": 0.013483298326444881, + "grad_norm": 18.894878387451172, + "learning_rate": 2.64e-05, + "loss": 1.2755, + "step": 199 + }, + { + "epoch": 0.013551053594416966, + "grad_norm": 16.611616134643555, + "learning_rate": 2.6533333333333332e-05, + "loss": 1.4218, + "step": 200 + }, + { + "epoch": 0.01361880886238905, + "grad_norm": 17.056982040405273, + "learning_rate": 2.6666666666666667e-05, + "loss": 1.4376, + "step": 201 + }, + { + "epoch": 0.013686564130361135, + "grad_norm": 15.91457748413086, + "learning_rate": 2.6800000000000004e-05, + "loss": 1.4836, + "step": 202 + }, + { + "epoch": 0.01375431939833322, + "grad_norm": 16.371660232543945, + "learning_rate": 2.6933333333333332e-05, + "loss": 1.7374, + "step": 203 + }, + { + "epoch": 0.013822074666305304, + "grad_norm": 14.45882511138916, + "learning_rate": 2.706666666666667e-05, + "loss": 1.3324, + "step": 204 + }, + { + "epoch": 0.013889829934277391, + "grad_norm": 18.779190063476562, + "learning_rate": 2.7200000000000004e-05, + "loss": 1.4845, + "step": 205 + }, + { + "epoch": 0.013957585202249476, + "grad_norm": 19.7078914642334, + "learning_rate": 2.733333333333333e-05, + "loss": 1.6681, + "step": 206 + }, + { + "epoch": 0.01402534047022156, + "grad_norm": 20.498266220092773, + "learning_rate": 2.746666666666667e-05, + "loss": 1.559, + "step": 207 + }, + { + "epoch": 0.014093095738193645, + "grad_norm": 20.2978515625, + "learning_rate": 2.7600000000000003e-05, + "loss": 1.5821, + "step": 208 + }, + { + "epoch": 0.01416085100616573, + "grad_norm": 18.85460090637207, + "learning_rate": 2.7733333333333334e-05, + "loss": 1.3, + "step": 209 + }, + { + "epoch": 0.014228606274137814, + "grad_norm": 13.912948608398438, + "learning_rate": 2.786666666666667e-05, + "loss": 1.4833, + "step": 210 + }, + { + "epoch": 0.014296361542109899, + "grad_norm": 19.084806442260742, + "learning_rate": 2.8000000000000003e-05, + "loss": 1.3379, + "step": 211 + }, + { + "epoch": 0.014364116810081984, + "grad_norm": 15.123217582702637, + "learning_rate": 2.8133333333333334e-05, + "loss": 1.354, + "step": 212 + }, + { + "epoch": 0.014431872078054069, + "grad_norm": 14.127299308776855, + "learning_rate": 2.8266666666666668e-05, + "loss": 1.2341, + "step": 213 + }, + { + "epoch": 0.014499627346026153, + "grad_norm": 15.609139442443848, + "learning_rate": 2.84e-05, + "loss": 1.3474, + "step": 214 + }, + { + "epoch": 0.014567382613998238, + "grad_norm": 20.588394165039062, + "learning_rate": 2.8533333333333333e-05, + "loss": 1.4881, + "step": 215 + }, + { + "epoch": 0.014635137881970323, + "grad_norm": 12.27170467376709, + "learning_rate": 2.8666666666666668e-05, + "loss": 1.2255, + "step": 216 + }, + { + "epoch": 0.014702893149942407, + "grad_norm": 18.14453125, + "learning_rate": 2.88e-05, + "loss": 1.6328, + "step": 217 + }, + { + "epoch": 0.014770648417914492, + "grad_norm": 17.722230911254883, + "learning_rate": 2.8933333333333333e-05, + "loss": 2.0172, + "step": 218 + }, + { + "epoch": 0.014838403685886578, + "grad_norm": 14.454687118530273, + "learning_rate": 2.906666666666667e-05, + "loss": 1.2936, + "step": 219 + }, + { + "epoch": 0.014906158953858663, + "grad_norm": 16.369625091552734, + "learning_rate": 2.9199999999999998e-05, + "loss": 1.6289, + "step": 220 + }, + { + "epoch": 0.014973914221830748, + "grad_norm": 17.64832305908203, + "learning_rate": 2.9333333333333336e-05, + "loss": 1.3135, + "step": 221 + }, + { + "epoch": 0.015041669489802833, + "grad_norm": 14.401778221130371, + "learning_rate": 2.946666666666667e-05, + "loss": 1.2614, + "step": 222 + }, + { + "epoch": 0.015109424757774917, + "grad_norm": 17.145814895629883, + "learning_rate": 2.96e-05, + "loss": 1.3473, + "step": 223 + }, + { + "epoch": 0.015177180025747002, + "grad_norm": 17.273954391479492, + "learning_rate": 2.9733333333333336e-05, + "loss": 1.3747, + "step": 224 + }, + { + "epoch": 0.015244935293719087, + "grad_norm": 17.190818786621094, + "learning_rate": 2.986666666666667e-05, + "loss": 1.5452, + "step": 225 + }, + { + "epoch": 0.015312690561691171, + "grad_norm": 19.573585510253906, + "learning_rate": 3e-05, + "loss": 1.4942, + "step": 226 + }, + { + "epoch": 0.015380445829663256, + "grad_norm": 15.349540710449219, + "learning_rate": 3.0133333333333335e-05, + "loss": 1.4277, + "step": 227 + }, + { + "epoch": 0.01544820109763534, + "grad_norm": 19.786211013793945, + "learning_rate": 3.0266666666666666e-05, + "loss": 1.5014, + "step": 228 + }, + { + "epoch": 0.015515956365607425, + "grad_norm": 17.51229476928711, + "learning_rate": 3.04e-05, + "loss": 1.695, + "step": 229 + }, + { + "epoch": 0.01558371163357951, + "grad_norm": 19.894756317138672, + "learning_rate": 3.0533333333333335e-05, + "loss": 1.4994, + "step": 230 + }, + { + "epoch": 0.015651466901551597, + "grad_norm": 13.575235366821289, + "learning_rate": 3.066666666666667e-05, + "loss": 1.4037, + "step": 231 + }, + { + "epoch": 0.01571922216952368, + "grad_norm": 18.516822814941406, + "learning_rate": 3.08e-05, + "loss": 1.5494, + "step": 232 + }, + { + "epoch": 0.015786977437495766, + "grad_norm": 17.343368530273438, + "learning_rate": 3.093333333333334e-05, + "loss": 1.1708, + "step": 233 + }, + { + "epoch": 0.01585473270546785, + "grad_norm": 18.78635597229004, + "learning_rate": 3.1066666666666665e-05, + "loss": 1.6543, + "step": 234 + }, + { + "epoch": 0.015922487973439935, + "grad_norm": 16.219989776611328, + "learning_rate": 3.12e-05, + "loss": 1.381, + "step": 235 + }, + { + "epoch": 0.01599024324141202, + "grad_norm": 15.987724304199219, + "learning_rate": 3.1333333333333334e-05, + "loss": 1.3981, + "step": 236 + }, + { + "epoch": 0.016057998509384105, + "grad_norm": 17.251537322998047, + "learning_rate": 3.146666666666667e-05, + "loss": 1.8181, + "step": 237 + }, + { + "epoch": 0.01612575377735619, + "grad_norm": 15.804817199707031, + "learning_rate": 3.16e-05, + "loss": 1.4755, + "step": 238 + }, + { + "epoch": 0.016193509045328274, + "grad_norm": 16.508943557739258, + "learning_rate": 3.173333333333334e-05, + "loss": 1.2167, + "step": 239 + }, + { + "epoch": 0.01626126431330036, + "grad_norm": 17.382783889770508, + "learning_rate": 3.1866666666666664e-05, + "loss": 1.6186, + "step": 240 + }, + { + "epoch": 0.016329019581272444, + "grad_norm": 17.938161849975586, + "learning_rate": 3.2000000000000005e-05, + "loss": 1.3685, + "step": 241 + }, + { + "epoch": 0.016396774849244528, + "grad_norm": 21.103981018066406, + "learning_rate": 3.213333333333334e-05, + "loss": 1.6275, + "step": 242 + }, + { + "epoch": 0.016464530117216613, + "grad_norm": 15.145224571228027, + "learning_rate": 3.226666666666667e-05, + "loss": 1.2058, + "step": 243 + }, + { + "epoch": 0.016532285385188698, + "grad_norm": 15.368849754333496, + "learning_rate": 3.24e-05, + "loss": 1.3436, + "step": 244 + }, + { + "epoch": 0.016600040653160782, + "grad_norm": 16.95122718811035, + "learning_rate": 3.253333333333333e-05, + "loss": 1.4346, + "step": 245 + }, + { + "epoch": 0.016667795921132867, + "grad_norm": 14.612030029296875, + "learning_rate": 3.266666666666667e-05, + "loss": 1.4759, + "step": 246 + }, + { + "epoch": 0.016735551189104952, + "grad_norm": 17.6989688873291, + "learning_rate": 3.2800000000000004e-05, + "loss": 1.3612, + "step": 247 + }, + { + "epoch": 0.016803306457077036, + "grad_norm": 19.985427856445312, + "learning_rate": 3.293333333333333e-05, + "loss": 1.2718, + "step": 248 + }, + { + "epoch": 0.01687106172504912, + "grad_norm": 12.162117958068848, + "learning_rate": 3.3066666666666666e-05, + "loss": 1.2133, + "step": 249 + }, + { + "epoch": 0.016938816993021206, + "grad_norm": 19.053850173950195, + "learning_rate": 3.32e-05, + "loss": 1.1345, + "step": 250 + }, + { + "epoch": 0.01700657226099329, + "grad_norm": 15.729182243347168, + "learning_rate": 3.3333333333333335e-05, + "loss": 1.3079, + "step": 251 + }, + { + "epoch": 0.01707432752896538, + "grad_norm": 18.984880447387695, + "learning_rate": 3.346666666666667e-05, + "loss": 1.5124, + "step": 252 + }, + { + "epoch": 0.017142082796937463, + "grad_norm": 14.444266319274902, + "learning_rate": 3.3600000000000004e-05, + "loss": 1.4279, + "step": 253 + }, + { + "epoch": 0.017209838064909548, + "grad_norm": 17.174089431762695, + "learning_rate": 3.373333333333333e-05, + "loss": 1.1192, + "step": 254 + }, + { + "epoch": 0.017277593332881633, + "grad_norm": 15.885781288146973, + "learning_rate": 3.3866666666666665e-05, + "loss": 1.3918, + "step": 255 + }, + { + "epoch": 0.017345348600853717, + "grad_norm": 14.934944152832031, + "learning_rate": 3.4000000000000007e-05, + "loss": 1.3348, + "step": 256 + }, + { + "epoch": 0.017413103868825802, + "grad_norm": 17.977643966674805, + "learning_rate": 3.4133333333333334e-05, + "loss": 1.392, + "step": 257 + }, + { + "epoch": 0.017480859136797887, + "grad_norm": 17.527130126953125, + "learning_rate": 3.426666666666667e-05, + "loss": 1.2939, + "step": 258 + }, + { + "epoch": 0.01754861440476997, + "grad_norm": 17.823453903198242, + "learning_rate": 3.4399999999999996e-05, + "loss": 1.3118, + "step": 259 + }, + { + "epoch": 0.017616369672742056, + "grad_norm": 15.746111869812012, + "learning_rate": 3.453333333333334e-05, + "loss": 1.3865, + "step": 260 + }, + { + "epoch": 0.01768412494071414, + "grad_norm": 18.798080444335938, + "learning_rate": 3.466666666666667e-05, + "loss": 1.399, + "step": 261 + }, + { + "epoch": 0.017751880208686226, + "grad_norm": 15.139904975891113, + "learning_rate": 3.48e-05, + "loss": 1.5248, + "step": 262 + }, + { + "epoch": 0.01781963547665831, + "grad_norm": 16.150434494018555, + "learning_rate": 3.493333333333333e-05, + "loss": 1.5033, + "step": 263 + }, + { + "epoch": 0.017887390744630395, + "grad_norm": 19.125490188598633, + "learning_rate": 3.506666666666667e-05, + "loss": 1.4695, + "step": 264 + }, + { + "epoch": 0.01795514601260248, + "grad_norm": 15.708895683288574, + "learning_rate": 3.52e-05, + "loss": 1.6189, + "step": 265 + }, + { + "epoch": 0.018022901280574564, + "grad_norm": 14.185934066772461, + "learning_rate": 3.5333333333333336e-05, + "loss": 1.3536, + "step": 266 + }, + { + "epoch": 0.01809065654854665, + "grad_norm": 15.22371768951416, + "learning_rate": 3.546666666666667e-05, + "loss": 1.4652, + "step": 267 + }, + { + "epoch": 0.018158411816518734, + "grad_norm": 15.854277610778809, + "learning_rate": 3.56e-05, + "loss": 1.7211, + "step": 268 + }, + { + "epoch": 0.01822616708449082, + "grad_norm": 18.41203498840332, + "learning_rate": 3.573333333333333e-05, + "loss": 1.54, + "step": 269 + }, + { + "epoch": 0.018293922352462903, + "grad_norm": 15.323198318481445, + "learning_rate": 3.586666666666667e-05, + "loss": 1.3582, + "step": 270 + }, + { + "epoch": 0.018361677620434988, + "grad_norm": 15.056265830993652, + "learning_rate": 3.6e-05, + "loss": 1.2959, + "step": 271 + }, + { + "epoch": 0.018429432888407073, + "grad_norm": 18.912555694580078, + "learning_rate": 3.6133333333333335e-05, + "loss": 1.5964, + "step": 272 + }, + { + "epoch": 0.018497188156379157, + "grad_norm": 15.093083381652832, + "learning_rate": 3.626666666666667e-05, + "loss": 1.3403, + "step": 273 + }, + { + "epoch": 0.018564943424351242, + "grad_norm": 17.38453483581543, + "learning_rate": 3.6400000000000004e-05, + "loss": 1.5082, + "step": 274 + }, + { + "epoch": 0.018632698692323327, + "grad_norm": 16.53204345703125, + "learning_rate": 3.653333333333334e-05, + "loss": 1.2373, + "step": 275 + }, + { + "epoch": 0.01870045396029541, + "grad_norm": 13.423516273498535, + "learning_rate": 3.6666666666666666e-05, + "loss": 1.3111, + "step": 276 + }, + { + "epoch": 0.018768209228267496, + "grad_norm": 14.92001724243164, + "learning_rate": 3.68e-05, + "loss": 1.5767, + "step": 277 + }, + { + "epoch": 0.018835964496239584, + "grad_norm": 15.229578971862793, + "learning_rate": 3.6933333333333334e-05, + "loss": 1.19, + "step": 278 + }, + { + "epoch": 0.01890371976421167, + "grad_norm": 17.159273147583008, + "learning_rate": 3.706666666666667e-05, + "loss": 1.335, + "step": 279 + }, + { + "epoch": 0.018971475032183754, + "grad_norm": 13.7230806350708, + "learning_rate": 3.72e-05, + "loss": 1.2587, + "step": 280 + }, + { + "epoch": 0.01903923030015584, + "grad_norm": 15.375404357910156, + "learning_rate": 3.733333333333334e-05, + "loss": 1.2744, + "step": 281 + }, + { + "epoch": 0.019106985568127923, + "grad_norm": 21.054534912109375, + "learning_rate": 3.7466666666666665e-05, + "loss": 1.5581, + "step": 282 + }, + { + "epoch": 0.019174740836100008, + "grad_norm": 17.954307556152344, + "learning_rate": 3.76e-05, + "loss": 1.6617, + "step": 283 + }, + { + "epoch": 0.019242496104072093, + "grad_norm": 17.954139709472656, + "learning_rate": 3.773333333333334e-05, + "loss": 1.7486, + "step": 284 + }, + { + "epoch": 0.019310251372044177, + "grad_norm": 16.210704803466797, + "learning_rate": 3.786666666666667e-05, + "loss": 1.661, + "step": 285 + }, + { + "epoch": 0.019378006640016262, + "grad_norm": 18.14916229248047, + "learning_rate": 3.8e-05, + "loss": 1.8154, + "step": 286 + }, + { + "epoch": 0.019445761907988347, + "grad_norm": 16.372896194458008, + "learning_rate": 3.8133333333333336e-05, + "loss": 1.5464, + "step": 287 + }, + { + "epoch": 0.01951351717596043, + "grad_norm": 16.4403076171875, + "learning_rate": 3.8266666666666664e-05, + "loss": 1.0815, + "step": 288 + }, + { + "epoch": 0.019581272443932516, + "grad_norm": 16.612815856933594, + "learning_rate": 3.8400000000000005e-05, + "loss": 1.4338, + "step": 289 + }, + { + "epoch": 0.0196490277119046, + "grad_norm": 17.0092716217041, + "learning_rate": 3.853333333333334e-05, + "loss": 1.5257, + "step": 290 + }, + { + "epoch": 0.019716782979876685, + "grad_norm": 13.920248031616211, + "learning_rate": 3.866666666666667e-05, + "loss": 1.4595, + "step": 291 + }, + { + "epoch": 0.01978453824784877, + "grad_norm": 12.699677467346191, + "learning_rate": 3.88e-05, + "loss": 1.2602, + "step": 292 + }, + { + "epoch": 0.019852293515820855, + "grad_norm": 17.31343650817871, + "learning_rate": 3.8933333333333336e-05, + "loss": 1.2934, + "step": 293 + }, + { + "epoch": 0.01992004878379294, + "grad_norm": 16.7819881439209, + "learning_rate": 3.906666666666667e-05, + "loss": 1.5385, + "step": 294 + }, + { + "epoch": 0.019987804051765024, + "grad_norm": 14.175074577331543, + "learning_rate": 3.9200000000000004e-05, + "loss": 1.1644, + "step": 295 + }, + { + "epoch": 0.02005555931973711, + "grad_norm": 22.863780975341797, + "learning_rate": 3.933333333333333e-05, + "loss": 1.5932, + "step": 296 + }, + { + "epoch": 0.020123314587709194, + "grad_norm": 15.278692245483398, + "learning_rate": 3.9466666666666666e-05, + "loss": 1.3997, + "step": 297 + }, + { + "epoch": 0.02019106985568128, + "grad_norm": 15.878965377807617, + "learning_rate": 3.960000000000001e-05, + "loss": 1.5992, + "step": 298 + }, + { + "epoch": 0.020258825123653363, + "grad_norm": 17.444494247436523, + "learning_rate": 3.9733333333333335e-05, + "loss": 1.5224, + "step": 299 + }, + { + "epoch": 0.020326580391625448, + "grad_norm": 15.011860847473145, + "learning_rate": 3.986666666666667e-05, + "loss": 1.2609, + "step": 300 + }, + { + "epoch": 0.020394335659597532, + "grad_norm": 17.057064056396484, + "learning_rate": 4e-05, + "loss": 1.2724, + "step": 301 + }, + { + "epoch": 0.020462090927569617, + "grad_norm": 14.946621894836426, + "learning_rate": 4.013333333333333e-05, + "loss": 1.273, + "step": 302 + }, + { + "epoch": 0.020529846195541702, + "grad_norm": 15.072589874267578, + "learning_rate": 4.026666666666667e-05, + "loss": 1.2029, + "step": 303 + }, + { + "epoch": 0.020597601463513786, + "grad_norm": 14.304384231567383, + "learning_rate": 4.0400000000000006e-05, + "loss": 1.3601, + "step": 304 + }, + { + "epoch": 0.020665356731485875, + "grad_norm": 16.18627166748047, + "learning_rate": 4.0533333333333334e-05, + "loss": 1.5844, + "step": 305 + }, + { + "epoch": 0.02073311199945796, + "grad_norm": 14.144999504089355, + "learning_rate": 4.066666666666667e-05, + "loss": 1.1638, + "step": 306 + }, + { + "epoch": 0.020800867267430044, + "grad_norm": 13.793155670166016, + "learning_rate": 4.08e-05, + "loss": 1.2101, + "step": 307 + }, + { + "epoch": 0.02086862253540213, + "grad_norm": 19.344085693359375, + "learning_rate": 4.093333333333334e-05, + "loss": 1.5899, + "step": 308 + }, + { + "epoch": 0.020936377803374213, + "grad_norm": 13.670002937316895, + "learning_rate": 4.106666666666667e-05, + "loss": 1.4271, + "step": 309 + }, + { + "epoch": 0.021004133071346298, + "grad_norm": 17.72041893005371, + "learning_rate": 4.12e-05, + "loss": 1.5276, + "step": 310 + }, + { + "epoch": 0.021071888339318383, + "grad_norm": 14.103938102722168, + "learning_rate": 4.133333333333333e-05, + "loss": 1.2355, + "step": 311 + }, + { + "epoch": 0.021139643607290468, + "grad_norm": 16.439529418945312, + "learning_rate": 4.146666666666667e-05, + "loss": 1.3393, + "step": 312 + }, + { + "epoch": 0.021207398875262552, + "grad_norm": 13.821382522583008, + "learning_rate": 4.16e-05, + "loss": 1.275, + "step": 313 + }, + { + "epoch": 0.021275154143234637, + "grad_norm": 19.987564086914062, + "learning_rate": 4.1733333333333336e-05, + "loss": 1.415, + "step": 314 + }, + { + "epoch": 0.02134290941120672, + "grad_norm": 12.660404205322266, + "learning_rate": 4.186666666666667e-05, + "loss": 1.2378, + "step": 315 + }, + { + "epoch": 0.021410664679178806, + "grad_norm": 14.009933471679688, + "learning_rate": 4.2e-05, + "loss": 1.1615, + "step": 316 + }, + { + "epoch": 0.02147841994715089, + "grad_norm": 16.11932945251465, + "learning_rate": 4.213333333333334e-05, + "loss": 1.0653, + "step": 317 + }, + { + "epoch": 0.021546175215122976, + "grad_norm": 15.943892478942871, + "learning_rate": 4.226666666666667e-05, + "loss": 1.6208, + "step": 318 + }, + { + "epoch": 0.02161393048309506, + "grad_norm": 16.01548194885254, + "learning_rate": 4.24e-05, + "loss": 1.4052, + "step": 319 + }, + { + "epoch": 0.021681685751067145, + "grad_norm": 15.703156471252441, + "learning_rate": 4.2533333333333335e-05, + "loss": 1.3736, + "step": 320 + }, + { + "epoch": 0.02174944101903923, + "grad_norm": 20.367021560668945, + "learning_rate": 4.266666666666667e-05, + "loss": 1.5758, + "step": 321 + }, + { + "epoch": 0.021817196287011315, + "grad_norm": 15.678637504577637, + "learning_rate": 4.2800000000000004e-05, + "loss": 1.4458, + "step": 322 + }, + { + "epoch": 0.0218849515549834, + "grad_norm": 15.707669258117676, + "learning_rate": 4.293333333333334e-05, + "loss": 1.3065, + "step": 323 + }, + { + "epoch": 0.021952706822955484, + "grad_norm": 16.07583999633789, + "learning_rate": 4.3066666666666665e-05, + "loss": 1.5875, + "step": 324 + }, + { + "epoch": 0.02202046209092757, + "grad_norm": 15.248065948486328, + "learning_rate": 4.32e-05, + "loss": 1.5972, + "step": 325 + }, + { + "epoch": 0.022088217358899653, + "grad_norm": 17.77379608154297, + "learning_rate": 4.3333333333333334e-05, + "loss": 1.3385, + "step": 326 + }, + { + "epoch": 0.022155972626871738, + "grad_norm": 24.381973266601562, + "learning_rate": 4.346666666666667e-05, + "loss": 1.6428, + "step": 327 + }, + { + "epoch": 0.022223727894843823, + "grad_norm": 17.3684024810791, + "learning_rate": 4.36e-05, + "loss": 1.3634, + "step": 328 + }, + { + "epoch": 0.022291483162815907, + "grad_norm": 20.45623207092285, + "learning_rate": 4.373333333333334e-05, + "loss": 1.3947, + "step": 329 + }, + { + "epoch": 0.022359238430787992, + "grad_norm": 17.434114456176758, + "learning_rate": 4.3866666666666665e-05, + "loss": 1.4269, + "step": 330 + }, + { + "epoch": 0.02242699369876008, + "grad_norm": 13.574700355529785, + "learning_rate": 4.4000000000000006e-05, + "loss": 1.2807, + "step": 331 + }, + { + "epoch": 0.022494748966732165, + "grad_norm": 16.404966354370117, + "learning_rate": 4.413333333333334e-05, + "loss": 1.4549, + "step": 332 + }, + { + "epoch": 0.02256250423470425, + "grad_norm": 14.14947509765625, + "learning_rate": 4.426666666666667e-05, + "loss": 1.4187, + "step": 333 + }, + { + "epoch": 0.022630259502676334, + "grad_norm": 17.632492065429688, + "learning_rate": 4.44e-05, + "loss": 1.3515, + "step": 334 + }, + { + "epoch": 0.02269801477064842, + "grad_norm": 18.149261474609375, + "learning_rate": 4.4533333333333336e-05, + "loss": 1.4875, + "step": 335 + }, + { + "epoch": 0.022765770038620504, + "grad_norm": 19.864049911499023, + "learning_rate": 4.466666666666667e-05, + "loss": 1.5622, + "step": 336 + }, + { + "epoch": 0.02283352530659259, + "grad_norm": 16.7431583404541, + "learning_rate": 4.4800000000000005e-05, + "loss": 1.4344, + "step": 337 + }, + { + "epoch": 0.022901280574564673, + "grad_norm": 16.53998565673828, + "learning_rate": 4.493333333333333e-05, + "loss": 1.1745, + "step": 338 + }, + { + "epoch": 0.022969035842536758, + "grad_norm": 17.786945343017578, + "learning_rate": 4.5066666666666667e-05, + "loss": 1.4335, + "step": 339 + }, + { + "epoch": 0.023036791110508843, + "grad_norm": 17.730606079101562, + "learning_rate": 4.52e-05, + "loss": 1.341, + "step": 340 + }, + { + "epoch": 0.023104546378480927, + "grad_norm": 15.536660194396973, + "learning_rate": 4.5333333333333335e-05, + "loss": 1.4935, + "step": 341 + }, + { + "epoch": 0.023172301646453012, + "grad_norm": 22.405893325805664, + "learning_rate": 4.546666666666667e-05, + "loss": 1.5538, + "step": 342 + }, + { + "epoch": 0.023240056914425097, + "grad_norm": 15.963828086853027, + "learning_rate": 4.5600000000000004e-05, + "loss": 1.373, + "step": 343 + }, + { + "epoch": 0.02330781218239718, + "grad_norm": 13.041223526000977, + "learning_rate": 4.573333333333333e-05, + "loss": 1.346, + "step": 344 + }, + { + "epoch": 0.023375567450369266, + "grad_norm": 14.552989959716797, + "learning_rate": 4.5866666666666666e-05, + "loss": 1.2723, + "step": 345 + }, + { + "epoch": 0.02344332271834135, + "grad_norm": 16.034404754638672, + "learning_rate": 4.600000000000001e-05, + "loss": 1.3671, + "step": 346 + }, + { + "epoch": 0.023511077986313435, + "grad_norm": 17.5505428314209, + "learning_rate": 4.6133333333333334e-05, + "loss": 1.4553, + "step": 347 + }, + { + "epoch": 0.02357883325428552, + "grad_norm": 14.029112815856934, + "learning_rate": 4.626666666666667e-05, + "loss": 1.3639, + "step": 348 + }, + { + "epoch": 0.023646588522257605, + "grad_norm": 14.633806228637695, + "learning_rate": 4.64e-05, + "loss": 1.61, + "step": 349 + }, + { + "epoch": 0.02371434379022969, + "grad_norm": 15.407472610473633, + "learning_rate": 4.653333333333334e-05, + "loss": 1.2449, + "step": 350 + }, + { + "epoch": 0.023782099058201774, + "grad_norm": 13.801481246948242, + "learning_rate": 4.666666666666667e-05, + "loss": 1.3105, + "step": 351 + }, + { + "epoch": 0.02384985432617386, + "grad_norm": 17.675159454345703, + "learning_rate": 4.6800000000000006e-05, + "loss": 1.3142, + "step": 352 + }, + { + "epoch": 0.023917609594145944, + "grad_norm": 15.383625984191895, + "learning_rate": 4.6933333333333333e-05, + "loss": 1.2976, + "step": 353 + }, + { + "epoch": 0.02398536486211803, + "grad_norm": 18.087081909179688, + "learning_rate": 4.706666666666667e-05, + "loss": 1.3391, + "step": 354 + }, + { + "epoch": 0.024053120130090113, + "grad_norm": 15.790519714355469, + "learning_rate": 4.72e-05, + "loss": 1.7103, + "step": 355 + }, + { + "epoch": 0.024120875398062198, + "grad_norm": 15.931246757507324, + "learning_rate": 4.7333333333333336e-05, + "loss": 1.4792, + "step": 356 + }, + { + "epoch": 0.024188630666034282, + "grad_norm": 18.39167022705078, + "learning_rate": 4.746666666666667e-05, + "loss": 1.5085, + "step": 357 + }, + { + "epoch": 0.02425638593400637, + "grad_norm": 13.518149375915527, + "learning_rate": 4.76e-05, + "loss": 1.2711, + "step": 358 + }, + { + "epoch": 0.024324141201978455, + "grad_norm": 18.02762794494629, + "learning_rate": 4.773333333333333e-05, + "loss": 1.4113, + "step": 359 + }, + { + "epoch": 0.02439189646995054, + "grad_norm": 15.260608673095703, + "learning_rate": 4.7866666666666674e-05, + "loss": 1.3286, + "step": 360 + }, + { + "epoch": 0.024459651737922625, + "grad_norm": 16.26275062561035, + "learning_rate": 4.8e-05, + "loss": 1.4436, + "step": 361 + }, + { + "epoch": 0.02452740700589471, + "grad_norm": 14.213051795959473, + "learning_rate": 4.8133333333333336e-05, + "loss": 1.1339, + "step": 362 + }, + { + "epoch": 0.024595162273866794, + "grad_norm": 16.478811264038086, + "learning_rate": 4.826666666666667e-05, + "loss": 1.5534, + "step": 363 + }, + { + "epoch": 0.02466291754183888, + "grad_norm": 13.24108600616455, + "learning_rate": 4.8400000000000004e-05, + "loss": 1.4767, + "step": 364 + }, + { + "epoch": 0.024730672809810963, + "grad_norm": 15.167949676513672, + "learning_rate": 4.853333333333334e-05, + "loss": 1.1592, + "step": 365 + }, + { + "epoch": 0.024798428077783048, + "grad_norm": 18.150367736816406, + "learning_rate": 4.866666666666667e-05, + "loss": 1.4172, + "step": 366 + }, + { + "epoch": 0.024866183345755133, + "grad_norm": 13.250176429748535, + "learning_rate": 4.88e-05, + "loss": 1.5022, + "step": 367 + }, + { + "epoch": 0.024933938613727218, + "grad_norm": 14.780989646911621, + "learning_rate": 4.8933333333333335e-05, + "loss": 1.3243, + "step": 368 + }, + { + "epoch": 0.025001693881699302, + "grad_norm": 13.34748649597168, + "learning_rate": 4.906666666666667e-05, + "loss": 1.285, + "step": 369 + }, + { + "epoch": 0.025069449149671387, + "grad_norm": 14.677448272705078, + "learning_rate": 4.92e-05, + "loss": 1.1982, + "step": 370 + }, + { + "epoch": 0.02513720441764347, + "grad_norm": 18.211145401000977, + "learning_rate": 4.933333333333334e-05, + "loss": 1.4018, + "step": 371 + }, + { + "epoch": 0.025204959685615556, + "grad_norm": 17.68402099609375, + "learning_rate": 4.9466666666666665e-05, + "loss": 1.6153, + "step": 372 + }, + { + "epoch": 0.02527271495358764, + "grad_norm": 14.44299030303955, + "learning_rate": 4.96e-05, + "loss": 1.4092, + "step": 373 + }, + { + "epoch": 0.025340470221559726, + "grad_norm": 15.467754364013672, + "learning_rate": 4.973333333333334e-05, + "loss": 1.5647, + "step": 374 + }, + { + "epoch": 0.02540822548953181, + "grad_norm": 14.037840843200684, + "learning_rate": 4.986666666666667e-05, + "loss": 1.2106, + "step": 375 + }, + { + "epoch": 0.025475980757503895, + "grad_norm": 15.72461223602295, + "learning_rate": 5e-05, + "loss": 1.1832, + "step": 376 + }, + { + "epoch": 0.02554373602547598, + "grad_norm": 15.03637981414795, + "learning_rate": 5.013333333333333e-05, + "loss": 1.5803, + "step": 377 + }, + { + "epoch": 0.025611491293448065, + "grad_norm": 15.995429992675781, + "learning_rate": 5.026666666666667e-05, + "loss": 1.3837, + "step": 378 + }, + { + "epoch": 0.02567924656142015, + "grad_norm": 17.96516227722168, + "learning_rate": 5.0400000000000005e-05, + "loss": 1.5114, + "step": 379 + }, + { + "epoch": 0.025747001829392234, + "grad_norm": 17.960783004760742, + "learning_rate": 5.053333333333333e-05, + "loss": 1.4012, + "step": 380 + }, + { + "epoch": 0.02581475709736432, + "grad_norm": 14.399024963378906, + "learning_rate": 5.0666666666666674e-05, + "loss": 1.2906, + "step": 381 + }, + { + "epoch": 0.025882512365336403, + "grad_norm": 14.268051147460938, + "learning_rate": 5.08e-05, + "loss": 1.2354, + "step": 382 + }, + { + "epoch": 0.025950267633308488, + "grad_norm": 16.070646286010742, + "learning_rate": 5.0933333333333336e-05, + "loss": 1.272, + "step": 383 + }, + { + "epoch": 0.026018022901280576, + "grad_norm": 18.698619842529297, + "learning_rate": 5.106666666666668e-05, + "loss": 1.352, + "step": 384 + }, + { + "epoch": 0.02608577816925266, + "grad_norm": 13.932543754577637, + "learning_rate": 5.1200000000000004e-05, + "loss": 1.3249, + "step": 385 + }, + { + "epoch": 0.026153533437224746, + "grad_norm": 14.76308822631836, + "learning_rate": 5.133333333333333e-05, + "loss": 1.1807, + "step": 386 + }, + { + "epoch": 0.02622128870519683, + "grad_norm": 17.602182388305664, + "learning_rate": 5.146666666666667e-05, + "loss": 1.4017, + "step": 387 + }, + { + "epoch": 0.026289043973168915, + "grad_norm": 13.589346885681152, + "learning_rate": 5.16e-05, + "loss": 1.4205, + "step": 388 + }, + { + "epoch": 0.026356799241141, + "grad_norm": 16.37788200378418, + "learning_rate": 5.1733333333333335e-05, + "loss": 1.4255, + "step": 389 + }, + { + "epoch": 0.026424554509113084, + "grad_norm": 16.24977684020996, + "learning_rate": 5.1866666666666676e-05, + "loss": 1.2526, + "step": 390 + }, + { + "epoch": 0.02649230977708517, + "grad_norm": 15.02576732635498, + "learning_rate": 5.2000000000000004e-05, + "loss": 1.2147, + "step": 391 + }, + { + "epoch": 0.026560065045057254, + "grad_norm": 16.754850387573242, + "learning_rate": 5.213333333333333e-05, + "loss": 1.5725, + "step": 392 + }, + { + "epoch": 0.02662782031302934, + "grad_norm": 17.120729446411133, + "learning_rate": 5.2266666666666665e-05, + "loss": 1.351, + "step": 393 + }, + { + "epoch": 0.026695575581001423, + "grad_norm": 18.83492660522461, + "learning_rate": 5.2400000000000007e-05, + "loss": 1.3044, + "step": 394 + }, + { + "epoch": 0.026763330848973508, + "grad_norm": 18.66061019897461, + "learning_rate": 5.2533333333333334e-05, + "loss": 1.4889, + "step": 395 + }, + { + "epoch": 0.026831086116945593, + "grad_norm": 18.493236541748047, + "learning_rate": 5.266666666666666e-05, + "loss": 1.5015, + "step": 396 + }, + { + "epoch": 0.026898841384917677, + "grad_norm": 19.523067474365234, + "learning_rate": 5.28e-05, + "loss": 1.5179, + "step": 397 + }, + { + "epoch": 0.026966596652889762, + "grad_norm": 18.207366943359375, + "learning_rate": 5.293333333333334e-05, + "loss": 1.374, + "step": 398 + }, + { + "epoch": 0.027034351920861847, + "grad_norm": 13.021437644958496, + "learning_rate": 5.3066666666666665e-05, + "loss": 1.0259, + "step": 399 + }, + { + "epoch": 0.02710210718883393, + "grad_norm": 14.108929634094238, + "learning_rate": 5.3200000000000006e-05, + "loss": 1.3914, + "step": 400 + }, + { + "epoch": 0.027169862456806016, + "grad_norm": 15.087890625, + "learning_rate": 5.333333333333333e-05, + "loss": 1.3143, + "step": 401 + }, + { + "epoch": 0.0272376177247781, + "grad_norm": 16.389965057373047, + "learning_rate": 5.346666666666667e-05, + "loss": 1.5286, + "step": 402 + }, + { + "epoch": 0.027305372992750186, + "grad_norm": 17.855867385864258, + "learning_rate": 5.360000000000001e-05, + "loss": 1.4221, + "step": 403 + }, + { + "epoch": 0.02737312826072227, + "grad_norm": 14.016682624816895, + "learning_rate": 5.3733333333333336e-05, + "loss": 1.0922, + "step": 404 + }, + { + "epoch": 0.027440883528694355, + "grad_norm": 16.08883285522461, + "learning_rate": 5.3866666666666664e-05, + "loss": 1.4932, + "step": 405 + }, + { + "epoch": 0.02750863879666644, + "grad_norm": 14.883580207824707, + "learning_rate": 5.4000000000000005e-05, + "loss": 1.0806, + "step": 406 + }, + { + "epoch": 0.027576394064638524, + "grad_norm": 13.443008422851562, + "learning_rate": 5.413333333333334e-05, + "loss": 1.2653, + "step": 407 + }, + { + "epoch": 0.02764414933261061, + "grad_norm": 13.391244888305664, + "learning_rate": 5.4266666666666667e-05, + "loss": 1.2788, + "step": 408 + }, + { + "epoch": 0.027711904600582694, + "grad_norm": 13.449023246765137, + "learning_rate": 5.440000000000001e-05, + "loss": 1.133, + "step": 409 + }, + { + "epoch": 0.027779659868554782, + "grad_norm": 13.777191162109375, + "learning_rate": 5.4533333333333335e-05, + "loss": 1.1509, + "step": 410 + }, + { + "epoch": 0.027847415136526867, + "grad_norm": 19.04452896118164, + "learning_rate": 5.466666666666666e-05, + "loss": 1.3306, + "step": 411 + }, + { + "epoch": 0.02791517040449895, + "grad_norm": 16.583112716674805, + "learning_rate": 5.4800000000000004e-05, + "loss": 1.3018, + "step": 412 + }, + { + "epoch": 0.027982925672471036, + "grad_norm": 16.8561954498291, + "learning_rate": 5.493333333333334e-05, + "loss": 1.257, + "step": 413 + }, + { + "epoch": 0.02805068094044312, + "grad_norm": 15.321064949035645, + "learning_rate": 5.5066666666666666e-05, + "loss": 1.2498, + "step": 414 + }, + { + "epoch": 0.028118436208415205, + "grad_norm": 15.825779914855957, + "learning_rate": 5.520000000000001e-05, + "loss": 1.2177, + "step": 415 + }, + { + "epoch": 0.02818619147638729, + "grad_norm": 15.665361404418945, + "learning_rate": 5.5333333333333334e-05, + "loss": 1.244, + "step": 416 + }, + { + "epoch": 0.028253946744359375, + "grad_norm": 15.516046524047852, + "learning_rate": 5.546666666666667e-05, + "loss": 1.4957, + "step": 417 + }, + { + "epoch": 0.02832170201233146, + "grad_norm": 19.148746490478516, + "learning_rate": 5.560000000000001e-05, + "loss": 1.4346, + "step": 418 + }, + { + "epoch": 0.028389457280303544, + "grad_norm": 15.215538024902344, + "learning_rate": 5.573333333333334e-05, + "loss": 1.081, + "step": 419 + }, + { + "epoch": 0.02845721254827563, + "grad_norm": 15.612996101379395, + "learning_rate": 5.5866666666666665e-05, + "loss": 1.1671, + "step": 420 + }, + { + "epoch": 0.028524967816247714, + "grad_norm": 13.754039764404297, + "learning_rate": 5.6000000000000006e-05, + "loss": 1.2679, + "step": 421 + }, + { + "epoch": 0.028592723084219798, + "grad_norm": 16.350305557250977, + "learning_rate": 5.613333333333334e-05, + "loss": 1.3034, + "step": 422 + }, + { + "epoch": 0.028660478352191883, + "grad_norm": 16.560344696044922, + "learning_rate": 5.626666666666667e-05, + "loss": 1.3633, + "step": 423 + }, + { + "epoch": 0.028728233620163968, + "grad_norm": 20.391889572143555, + "learning_rate": 5.6399999999999995e-05, + "loss": 1.3533, + "step": 424 + }, + { + "epoch": 0.028795988888136052, + "grad_norm": 16.13326072692871, + "learning_rate": 5.6533333333333336e-05, + "loss": 1.3678, + "step": 425 + }, + { + "epoch": 0.028863744156108137, + "grad_norm": 21.544612884521484, + "learning_rate": 5.666666666666667e-05, + "loss": 1.6275, + "step": 426 + }, + { + "epoch": 0.028931499424080222, + "grad_norm": 17.097408294677734, + "learning_rate": 5.68e-05, + "loss": 1.6119, + "step": 427 + }, + { + "epoch": 0.028999254692052306, + "grad_norm": 15.699347496032715, + "learning_rate": 5.693333333333334e-05, + "loss": 1.2644, + "step": 428 + }, + { + "epoch": 0.02906700996002439, + "grad_norm": 16.643613815307617, + "learning_rate": 5.706666666666667e-05, + "loss": 1.3497, + "step": 429 + }, + { + "epoch": 0.029134765227996476, + "grad_norm": 20.941123962402344, + "learning_rate": 5.72e-05, + "loss": 1.4232, + "step": 430 + }, + { + "epoch": 0.02920252049596856, + "grad_norm": 17.549013137817383, + "learning_rate": 5.7333333333333336e-05, + "loss": 1.3429, + "step": 431 + }, + { + "epoch": 0.029270275763940645, + "grad_norm": 19.702617645263672, + "learning_rate": 5.746666666666667e-05, + "loss": 1.7924, + "step": 432 + }, + { + "epoch": 0.02933803103191273, + "grad_norm": 16.396209716796875, + "learning_rate": 5.76e-05, + "loss": 1.3296, + "step": 433 + }, + { + "epoch": 0.029405786299884815, + "grad_norm": 15.823278427124023, + "learning_rate": 5.773333333333334e-05, + "loss": 1.4228, + "step": 434 + }, + { + "epoch": 0.0294735415678569, + "grad_norm": 19.61952018737793, + "learning_rate": 5.7866666666666666e-05, + "loss": 1.3522, + "step": 435 + }, + { + "epoch": 0.029541296835828984, + "grad_norm": 14.721433639526367, + "learning_rate": 5.8e-05, + "loss": 1.284, + "step": 436 + }, + { + "epoch": 0.029609052103801072, + "grad_norm": 16.594276428222656, + "learning_rate": 5.813333333333334e-05, + "loss": 1.4636, + "step": 437 + }, + { + "epoch": 0.029676807371773157, + "grad_norm": 15.98005199432373, + "learning_rate": 5.826666666666667e-05, + "loss": 1.2825, + "step": 438 + }, + { + "epoch": 0.02974456263974524, + "grad_norm": 15.501729965209961, + "learning_rate": 5.8399999999999997e-05, + "loss": 1.3736, + "step": 439 + }, + { + "epoch": 0.029812317907717326, + "grad_norm": 18.077552795410156, + "learning_rate": 5.853333333333334e-05, + "loss": 1.4115, + "step": 440 + }, + { + "epoch": 0.02988007317568941, + "grad_norm": 16.016721725463867, + "learning_rate": 5.866666666666667e-05, + "loss": 1.3177, + "step": 441 + }, + { + "epoch": 0.029947828443661496, + "grad_norm": 16.39783477783203, + "learning_rate": 5.88e-05, + "loss": 1.4465, + "step": 442 + }, + { + "epoch": 0.03001558371163358, + "grad_norm": 13.970220565795898, + "learning_rate": 5.893333333333334e-05, + "loss": 1.2956, + "step": 443 + }, + { + "epoch": 0.030083338979605665, + "grad_norm": 16.5622615814209, + "learning_rate": 5.906666666666667e-05, + "loss": 1.2144, + "step": 444 + }, + { + "epoch": 0.03015109424757775, + "grad_norm": 18.191911697387695, + "learning_rate": 5.92e-05, + "loss": 1.623, + "step": 445 + }, + { + "epoch": 0.030218849515549834, + "grad_norm": 15.483613967895508, + "learning_rate": 5.9333333333333343e-05, + "loss": 1.3232, + "step": 446 + }, + { + "epoch": 0.03028660478352192, + "grad_norm": 15.633200645446777, + "learning_rate": 5.946666666666667e-05, + "loss": 1.544, + "step": 447 + }, + { + "epoch": 0.030354360051494004, + "grad_norm": 16.025054931640625, + "learning_rate": 5.96e-05, + "loss": 1.4837, + "step": 448 + }, + { + "epoch": 0.03042211531946609, + "grad_norm": 15.954922676086426, + "learning_rate": 5.973333333333334e-05, + "loss": 1.341, + "step": 449 + }, + { + "epoch": 0.030489870587438173, + "grad_norm": 14.852401733398438, + "learning_rate": 5.9866666666666674e-05, + "loss": 1.2346, + "step": 450 + }, + { + "epoch": 0.030557625855410258, + "grad_norm": 14.887676239013672, + "learning_rate": 6e-05, + "loss": 1.2789, + "step": 451 + }, + { + "epoch": 0.030625381123382343, + "grad_norm": 17.25469970703125, + "learning_rate": 6.013333333333334e-05, + "loss": 1.4446, + "step": 452 + }, + { + "epoch": 0.030693136391354427, + "grad_norm": 19.356597900390625, + "learning_rate": 6.026666666666667e-05, + "loss": 1.3765, + "step": 453 + }, + { + "epoch": 0.030760891659326512, + "grad_norm": 14.16335391998291, + "learning_rate": 6.04e-05, + "loss": 1.4213, + "step": 454 + }, + { + "epoch": 0.030828646927298597, + "grad_norm": 15.812528610229492, + "learning_rate": 6.053333333333333e-05, + "loss": 1.4028, + "step": 455 + }, + { + "epoch": 0.03089640219527068, + "grad_norm": 15.10727596282959, + "learning_rate": 6.066666666666667e-05, + "loss": 1.3841, + "step": 456 + }, + { + "epoch": 0.030964157463242766, + "grad_norm": 19.91744613647461, + "learning_rate": 6.08e-05, + "loss": 1.4706, + "step": 457 + }, + { + "epoch": 0.03103191273121485, + "grad_norm": 14.597813606262207, + "learning_rate": 6.093333333333333e-05, + "loss": 1.2428, + "step": 458 + }, + { + "epoch": 0.031099667999186936, + "grad_norm": 14.952363014221191, + "learning_rate": 6.106666666666667e-05, + "loss": 1.3431, + "step": 459 + }, + { + "epoch": 0.03116742326715902, + "grad_norm": 16.519468307495117, + "learning_rate": 6.12e-05, + "loss": 1.5399, + "step": 460 + }, + { + "epoch": 0.031235178535131105, + "grad_norm": 14.55786418914795, + "learning_rate": 6.133333333333334e-05, + "loss": 1.3844, + "step": 461 + }, + { + "epoch": 0.03130293380310319, + "grad_norm": 12.52665901184082, + "learning_rate": 6.146666666666668e-05, + "loss": 1.2085, + "step": 462 + }, + { + "epoch": 0.03137068907107528, + "grad_norm": 17.549148559570312, + "learning_rate": 6.16e-05, + "loss": 1.588, + "step": 463 + }, + { + "epoch": 0.03143844433904736, + "grad_norm": 14.761232376098633, + "learning_rate": 6.173333333333333e-05, + "loss": 1.1685, + "step": 464 + }, + { + "epoch": 0.03150619960701945, + "grad_norm": 13.1014404296875, + "learning_rate": 6.186666666666668e-05, + "loss": 1.1718, + "step": 465 + }, + { + "epoch": 0.03157395487499153, + "grad_norm": 15.998059272766113, + "learning_rate": 6.2e-05, + "loss": 1.3177, + "step": 466 + }, + { + "epoch": 0.03164171014296362, + "grad_norm": 16.402875900268555, + "learning_rate": 6.213333333333333e-05, + "loss": 1.4821, + "step": 467 + }, + { + "epoch": 0.0317094654109357, + "grad_norm": 12.871467590332031, + "learning_rate": 6.226666666666667e-05, + "loss": 1.0176, + "step": 468 + }, + { + "epoch": 0.031777220678907786, + "grad_norm": 15.602563858032227, + "learning_rate": 6.24e-05, + "loss": 1.3418, + "step": 469 + }, + { + "epoch": 0.03184497594687987, + "grad_norm": 13.5369234085083, + "learning_rate": 6.253333333333333e-05, + "loss": 1.4051, + "step": 470 + }, + { + "epoch": 0.031912731214851955, + "grad_norm": 16.841650009155273, + "learning_rate": 6.266666666666667e-05, + "loss": 1.5438, + "step": 471 + }, + { + "epoch": 0.03198048648282404, + "grad_norm": 14.581981658935547, + "learning_rate": 6.280000000000001e-05, + "loss": 1.2767, + "step": 472 + }, + { + "epoch": 0.032048241750796125, + "grad_norm": 14.084460258483887, + "learning_rate": 6.293333333333334e-05, + "loss": 1.2224, + "step": 473 + }, + { + "epoch": 0.03211599701876821, + "grad_norm": 15.54316234588623, + "learning_rate": 6.306666666666668e-05, + "loss": 1.6084, + "step": 474 + }, + { + "epoch": 0.032183752286740294, + "grad_norm": 15.498743057250977, + "learning_rate": 6.32e-05, + "loss": 1.3937, + "step": 475 + }, + { + "epoch": 0.03225150755471238, + "grad_norm": 17.728227615356445, + "learning_rate": 6.333333333333333e-05, + "loss": 1.2996, + "step": 476 + }, + { + "epoch": 0.032319262822684464, + "grad_norm": 13.956122398376465, + "learning_rate": 6.346666666666667e-05, + "loss": 1.4472, + "step": 477 + }, + { + "epoch": 0.03238701809065655, + "grad_norm": 18.578060150146484, + "learning_rate": 6.36e-05, + "loss": 1.4659, + "step": 478 + }, + { + "epoch": 0.03245477335862863, + "grad_norm": 14.055442810058594, + "learning_rate": 6.373333333333333e-05, + "loss": 1.1979, + "step": 479 + }, + { + "epoch": 0.03252252862660072, + "grad_norm": 15.902195930480957, + "learning_rate": 6.386666666666667e-05, + "loss": 1.3367, + "step": 480 + }, + { + "epoch": 0.0325902838945728, + "grad_norm": 14.212138175964355, + "learning_rate": 6.400000000000001e-05, + "loss": 1.2716, + "step": 481 + }, + { + "epoch": 0.03265803916254489, + "grad_norm": 17.711475372314453, + "learning_rate": 6.413333333333334e-05, + "loss": 1.3016, + "step": 482 + }, + { + "epoch": 0.03272579443051697, + "grad_norm": 13.068260192871094, + "learning_rate": 6.426666666666668e-05, + "loss": 1.1346, + "step": 483 + }, + { + "epoch": 0.032793549698489056, + "grad_norm": 17.422321319580078, + "learning_rate": 6.440000000000001e-05, + "loss": 1.0931, + "step": 484 + }, + { + "epoch": 0.03286130496646114, + "grad_norm": 14.98038101196289, + "learning_rate": 6.453333333333333e-05, + "loss": 1.22, + "step": 485 + }, + { + "epoch": 0.032929060234433226, + "grad_norm": 13.300479888916016, + "learning_rate": 6.466666666666666e-05, + "loss": 1.2647, + "step": 486 + }, + { + "epoch": 0.03299681550240531, + "grad_norm": 14.611360549926758, + "learning_rate": 6.48e-05, + "loss": 1.2478, + "step": 487 + }, + { + "epoch": 0.033064570770377395, + "grad_norm": 19.244020462036133, + "learning_rate": 6.493333333333333e-05, + "loss": 1.2126, + "step": 488 + }, + { + "epoch": 0.03313232603834948, + "grad_norm": 16.285396575927734, + "learning_rate": 6.506666666666666e-05, + "loss": 1.1694, + "step": 489 + }, + { + "epoch": 0.033200081306321565, + "grad_norm": 19.613059997558594, + "learning_rate": 6.52e-05, + "loss": 1.2118, + "step": 490 + }, + { + "epoch": 0.03326783657429365, + "grad_norm": 19.19630241394043, + "learning_rate": 6.533333333333334e-05, + "loss": 1.3093, + "step": 491 + }, + { + "epoch": 0.033335591842265734, + "grad_norm": 12.653651237487793, + "learning_rate": 6.546666666666667e-05, + "loss": 1.1545, + "step": 492 + }, + { + "epoch": 0.03340334711023782, + "grad_norm": 13.434443473815918, + "learning_rate": 6.560000000000001e-05, + "loss": 1.2543, + "step": 493 + }, + { + "epoch": 0.033471102378209903, + "grad_norm": 18.893598556518555, + "learning_rate": 6.573333333333334e-05, + "loss": 1.3632, + "step": 494 + }, + { + "epoch": 0.03353885764618199, + "grad_norm": 18.339479446411133, + "learning_rate": 6.586666666666666e-05, + "loss": 1.4369, + "step": 495 + }, + { + "epoch": 0.03360661291415407, + "grad_norm": 12.303078651428223, + "learning_rate": 6.6e-05, + "loss": 1.4032, + "step": 496 + }, + { + "epoch": 0.03367436818212616, + "grad_norm": 17.041015625, + "learning_rate": 6.613333333333333e-05, + "loss": 1.2395, + "step": 497 + }, + { + "epoch": 0.03374212345009824, + "grad_norm": 14.035640716552734, + "learning_rate": 6.626666666666666e-05, + "loss": 1.3541, + "step": 498 + }, + { + "epoch": 0.03380987871807033, + "grad_norm": 16.93412208557129, + "learning_rate": 6.64e-05, + "loss": 1.3503, + "step": 499 + }, + { + "epoch": 0.03387763398604241, + "grad_norm": 16.02039909362793, + "learning_rate": 6.653333333333334e-05, + "loss": 1.3226, + "step": 500 + }, + { + "epoch": 0.033945389254014496, + "grad_norm": 16.061542510986328, + "learning_rate": 6.666666666666667e-05, + "loss": 1.433, + "step": 501 + }, + { + "epoch": 0.03401314452198658, + "grad_norm": 14.810320854187012, + "learning_rate": 6.680000000000001e-05, + "loss": 1.3715, + "step": 502 + }, + { + "epoch": 0.03408089978995867, + "grad_norm": 15.132527351379395, + "learning_rate": 6.693333333333334e-05, + "loss": 1.0683, + "step": 503 + }, + { + "epoch": 0.03414865505793076, + "grad_norm": 13.768165588378906, + "learning_rate": 6.706666666666667e-05, + "loss": 1.234, + "step": 504 + }, + { + "epoch": 0.03421641032590284, + "grad_norm": 16.227920532226562, + "learning_rate": 6.720000000000001e-05, + "loss": 1.4181, + "step": 505 + }, + { + "epoch": 0.03428416559387493, + "grad_norm": 14.205772399902344, + "learning_rate": 6.733333333333333e-05, + "loss": 1.4342, + "step": 506 + }, + { + "epoch": 0.03435192086184701, + "grad_norm": 17.451099395751953, + "learning_rate": 6.746666666666666e-05, + "loss": 1.3338, + "step": 507 + }, + { + "epoch": 0.034419676129819096, + "grad_norm": 13.7041015625, + "learning_rate": 6.76e-05, + "loss": 1.3972, + "step": 508 + }, + { + "epoch": 0.03448743139779118, + "grad_norm": 13.206351280212402, + "learning_rate": 6.773333333333333e-05, + "loss": 1.4233, + "step": 509 + }, + { + "epoch": 0.034555186665763266, + "grad_norm": 11.867975234985352, + "learning_rate": 6.786666666666667e-05, + "loss": 1.111, + "step": 510 + }, + { + "epoch": 0.03462294193373535, + "grad_norm": 19.726648330688477, + "learning_rate": 6.800000000000001e-05, + "loss": 1.2918, + "step": 511 + }, + { + "epoch": 0.034690697201707435, + "grad_norm": 18.914016723632812, + "learning_rate": 6.813333333333334e-05, + "loss": 1.5998, + "step": 512 + }, + { + "epoch": 0.03475845246967952, + "grad_norm": 18.065937042236328, + "learning_rate": 6.826666666666667e-05, + "loss": 1.4952, + "step": 513 + }, + { + "epoch": 0.034826207737651604, + "grad_norm": 16.42376136779785, + "learning_rate": 6.840000000000001e-05, + "loss": 1.2807, + "step": 514 + }, + { + "epoch": 0.03489396300562369, + "grad_norm": 15.537372589111328, + "learning_rate": 6.853333333333334e-05, + "loss": 1.3524, + "step": 515 + }, + { + "epoch": 0.034961718273595774, + "grad_norm": 18.395238876342773, + "learning_rate": 6.866666666666666e-05, + "loss": 1.3022, + "step": 516 + }, + { + "epoch": 0.03502947354156786, + "grad_norm": 15.80531120300293, + "learning_rate": 6.879999999999999e-05, + "loss": 1.4201, + "step": 517 + }, + { + "epoch": 0.03509722880953994, + "grad_norm": 15.361504554748535, + "learning_rate": 6.893333333333333e-05, + "loss": 1.4351, + "step": 518 + }, + { + "epoch": 0.03516498407751203, + "grad_norm": 12.501819610595703, + "learning_rate": 6.906666666666667e-05, + "loss": 1.1894, + "step": 519 + }, + { + "epoch": 0.03523273934548411, + "grad_norm": 14.53589916229248, + "learning_rate": 6.92e-05, + "loss": 1.3435, + "step": 520 + }, + { + "epoch": 0.0353004946134562, + "grad_norm": 15.057633399963379, + "learning_rate": 6.933333333333334e-05, + "loss": 1.5467, + "step": 521 + }, + { + "epoch": 0.03536824988142828, + "grad_norm": 16.27166175842285, + "learning_rate": 6.946666666666667e-05, + "loss": 1.5171, + "step": 522 + }, + { + "epoch": 0.03543600514940037, + "grad_norm": 18.050413131713867, + "learning_rate": 6.96e-05, + "loss": 1.2246, + "step": 523 + }, + { + "epoch": 0.03550376041737245, + "grad_norm": 15.316112518310547, + "learning_rate": 6.973333333333334e-05, + "loss": 1.2551, + "step": 524 + }, + { + "epoch": 0.035571515685344536, + "grad_norm": 15.416853904724121, + "learning_rate": 6.986666666666667e-05, + "loss": 1.1895, + "step": 525 + }, + { + "epoch": 0.03563927095331662, + "grad_norm": 15.019004821777344, + "learning_rate": 7e-05, + "loss": 1.5957, + "step": 526 + }, + { + "epoch": 0.035707026221288705, + "grad_norm": 15.33927059173584, + "learning_rate": 7.013333333333333e-05, + "loss": 1.3536, + "step": 527 + }, + { + "epoch": 0.03577478148926079, + "grad_norm": 16.403501510620117, + "learning_rate": 7.026666666666668e-05, + "loss": 1.6104, + "step": 528 + }, + { + "epoch": 0.035842536757232875, + "grad_norm": 13.962714195251465, + "learning_rate": 7.04e-05, + "loss": 1.2122, + "step": 529 + }, + { + "epoch": 0.03591029202520496, + "grad_norm": 14.530926704406738, + "learning_rate": 7.053333333333334e-05, + "loss": 1.2879, + "step": 530 + }, + { + "epoch": 0.035978047293177044, + "grad_norm": 13.99305248260498, + "learning_rate": 7.066666666666667e-05, + "loss": 1.2063, + "step": 531 + }, + { + "epoch": 0.03604580256114913, + "grad_norm": 17.54342269897461, + "learning_rate": 7.08e-05, + "loss": 1.2823, + "step": 532 + }, + { + "epoch": 0.036113557829121214, + "grad_norm": 11.097670555114746, + "learning_rate": 7.093333333333334e-05, + "loss": 1.1658, + "step": 533 + }, + { + "epoch": 0.0361813130970933, + "grad_norm": 19.069889068603516, + "learning_rate": 7.106666666666667e-05, + "loss": 1.3368, + "step": 534 + }, + { + "epoch": 0.03624906836506538, + "grad_norm": 12.675487518310547, + "learning_rate": 7.12e-05, + "loss": 1.1586, + "step": 535 + }, + { + "epoch": 0.03631682363303747, + "grad_norm": 16.971637725830078, + "learning_rate": 7.133333333333334e-05, + "loss": 1.2458, + "step": 536 + }, + { + "epoch": 0.03638457890100955, + "grad_norm": 14.513422012329102, + "learning_rate": 7.146666666666666e-05, + "loss": 1.236, + "step": 537 + }, + { + "epoch": 0.03645233416898164, + "grad_norm": 14.271836280822754, + "learning_rate": 7.16e-05, + "loss": 1.0951, + "step": 538 + }, + { + "epoch": 0.03652008943695372, + "grad_norm": 14.286022186279297, + "learning_rate": 7.173333333333335e-05, + "loss": 1.2411, + "step": 539 + }, + { + "epoch": 0.03658784470492581, + "grad_norm": 16.81059455871582, + "learning_rate": 7.186666666666667e-05, + "loss": 1.2453, + "step": 540 + }, + { + "epoch": 0.03665559997289789, + "grad_norm": 16.818300247192383, + "learning_rate": 7.2e-05, + "loss": 1.2302, + "step": 541 + }, + { + "epoch": 0.036723355240869976, + "grad_norm": 14.839925765991211, + "learning_rate": 7.213333333333334e-05, + "loss": 1.3445, + "step": 542 + }, + { + "epoch": 0.03679111050884206, + "grad_norm": 14.917344093322754, + "learning_rate": 7.226666666666667e-05, + "loss": 1.3658, + "step": 543 + }, + { + "epoch": 0.036858865776814145, + "grad_norm": 15.310625076293945, + "learning_rate": 7.24e-05, + "loss": 1.2565, + "step": 544 + }, + { + "epoch": 0.03692662104478623, + "grad_norm": 16.7796630859375, + "learning_rate": 7.253333333333334e-05, + "loss": 1.265, + "step": 545 + }, + { + "epoch": 0.036994376312758315, + "grad_norm": 19.4224796295166, + "learning_rate": 7.266666666666667e-05, + "loss": 1.6152, + "step": 546 + }, + { + "epoch": 0.0370621315807304, + "grad_norm": 15.8001127243042, + "learning_rate": 7.280000000000001e-05, + "loss": 1.4313, + "step": 547 + }, + { + "epoch": 0.037129886848702484, + "grad_norm": 13.597784042358398, + "learning_rate": 7.293333333333334e-05, + "loss": 1.3812, + "step": 548 + }, + { + "epoch": 0.03719764211667457, + "grad_norm": 13.139778137207031, + "learning_rate": 7.306666666666668e-05, + "loss": 1.1888, + "step": 549 + }, + { + "epoch": 0.037265397384646654, + "grad_norm": 17.083406448364258, + "learning_rate": 7.32e-05, + "loss": 1.1668, + "step": 550 + }, + { + "epoch": 0.03733315265261874, + "grad_norm": 14.46076774597168, + "learning_rate": 7.333333333333333e-05, + "loss": 1.2394, + "step": 551 + }, + { + "epoch": 0.03740090792059082, + "grad_norm": 16.217782974243164, + "learning_rate": 7.346666666666667e-05, + "loss": 1.3068, + "step": 552 + }, + { + "epoch": 0.03746866318856291, + "grad_norm": 16.06308937072754, + "learning_rate": 7.36e-05, + "loss": 0.9678, + "step": 553 + }, + { + "epoch": 0.03753641845653499, + "grad_norm": 14.652907371520996, + "learning_rate": 7.373333333333333e-05, + "loss": 1.5375, + "step": 554 + }, + { + "epoch": 0.03760417372450708, + "grad_norm": 13.870722770690918, + "learning_rate": 7.386666666666667e-05, + "loss": 1.29, + "step": 555 + }, + { + "epoch": 0.03767192899247917, + "grad_norm": 14.649571418762207, + "learning_rate": 7.4e-05, + "loss": 1.3588, + "step": 556 + }, + { + "epoch": 0.03773968426045125, + "grad_norm": 13.698415756225586, + "learning_rate": 7.413333333333334e-05, + "loss": 1.0801, + "step": 557 + }, + { + "epoch": 0.03780743952842334, + "grad_norm": 19.412424087524414, + "learning_rate": 7.426666666666668e-05, + "loss": 1.4474, + "step": 558 + }, + { + "epoch": 0.03787519479639542, + "grad_norm": 18.58599853515625, + "learning_rate": 7.44e-05, + "loss": 1.3294, + "step": 559 + }, + { + "epoch": 0.03794295006436751, + "grad_norm": 15.283289909362793, + "learning_rate": 7.453333333333333e-05, + "loss": 1.1554, + "step": 560 + }, + { + "epoch": 0.03801070533233959, + "grad_norm": 16.289731979370117, + "learning_rate": 7.466666666666667e-05, + "loss": 1.222, + "step": 561 + }, + { + "epoch": 0.03807846060031168, + "grad_norm": 25.105520248413086, + "learning_rate": 7.48e-05, + "loss": 1.2213, + "step": 562 + }, + { + "epoch": 0.03814621586828376, + "grad_norm": 15.428693771362305, + "learning_rate": 7.493333333333333e-05, + "loss": 1.2165, + "step": 563 + }, + { + "epoch": 0.038213971136255846, + "grad_norm": 18.79145622253418, + "learning_rate": 7.506666666666667e-05, + "loss": 1.4799, + "step": 564 + }, + { + "epoch": 0.03828172640422793, + "grad_norm": 12.498862266540527, + "learning_rate": 7.52e-05, + "loss": 1.4348, + "step": 565 + }, + { + "epoch": 0.038349481672200016, + "grad_norm": 19.302845001220703, + "learning_rate": 7.533333333333334e-05, + "loss": 1.2656, + "step": 566 + }, + { + "epoch": 0.0384172369401721, + "grad_norm": 16.28862953186035, + "learning_rate": 7.546666666666668e-05, + "loss": 1.4779, + "step": 567 + }, + { + "epoch": 0.038484992208144185, + "grad_norm": 13.972156524658203, + "learning_rate": 7.560000000000001e-05, + "loss": 1.4775, + "step": 568 + }, + { + "epoch": 0.03855274747611627, + "grad_norm": 15.443683624267578, + "learning_rate": 7.573333333333334e-05, + "loss": 1.2671, + "step": 569 + }, + { + "epoch": 0.038620502744088354, + "grad_norm": 15.203113555908203, + "learning_rate": 7.586666666666668e-05, + "loss": 1.4209, + "step": 570 + }, + { + "epoch": 0.03868825801206044, + "grad_norm": 18.516462326049805, + "learning_rate": 7.6e-05, + "loss": 1.3315, + "step": 571 + }, + { + "epoch": 0.038756013280032524, + "grad_norm": 13.403026580810547, + "learning_rate": 7.613333333333333e-05, + "loss": 1.364, + "step": 572 + }, + { + "epoch": 0.03882376854800461, + "grad_norm": 12.339615821838379, + "learning_rate": 7.626666666666667e-05, + "loss": 1.2107, + "step": 573 + }, + { + "epoch": 0.03889152381597669, + "grad_norm": 15.647716522216797, + "learning_rate": 7.64e-05, + "loss": 1.3165, + "step": 574 + }, + { + "epoch": 0.03895927908394878, + "grad_norm": 15.854050636291504, + "learning_rate": 7.653333333333333e-05, + "loss": 1.517, + "step": 575 + }, + { + "epoch": 0.03902703435192086, + "grad_norm": 11.765929222106934, + "learning_rate": 7.666666666666667e-05, + "loss": 1.2676, + "step": 576 + }, + { + "epoch": 0.03909478961989295, + "grad_norm": 12.515352249145508, + "learning_rate": 7.680000000000001e-05, + "loss": 1.4472, + "step": 577 + }, + { + "epoch": 0.03916254488786503, + "grad_norm": 11.72417163848877, + "learning_rate": 7.693333333333334e-05, + "loss": 1.1874, + "step": 578 + }, + { + "epoch": 0.03923030015583712, + "grad_norm": 15.03148365020752, + "learning_rate": 7.706666666666668e-05, + "loss": 1.6381, + "step": 579 + }, + { + "epoch": 0.0392980554238092, + "grad_norm": 14.0188570022583, + "learning_rate": 7.72e-05, + "loss": 1.4351, + "step": 580 + }, + { + "epoch": 0.039365810691781286, + "grad_norm": 12.343233108520508, + "learning_rate": 7.733333333333333e-05, + "loss": 1.2293, + "step": 581 + }, + { + "epoch": 0.03943356595975337, + "grad_norm": 15.358945846557617, + "learning_rate": 7.746666666666666e-05, + "loss": 1.307, + "step": 582 + }, + { + "epoch": 0.039501321227725456, + "grad_norm": 14.131333351135254, + "learning_rate": 7.76e-05, + "loss": 1.0432, + "step": 583 + }, + { + "epoch": 0.03956907649569754, + "grad_norm": 16.46926498413086, + "learning_rate": 7.773333333333333e-05, + "loss": 1.4126, + "step": 584 + }, + { + "epoch": 0.039636831763669625, + "grad_norm": 15.209906578063965, + "learning_rate": 7.786666666666667e-05, + "loss": 1.3242, + "step": 585 + }, + { + "epoch": 0.03970458703164171, + "grad_norm": 15.657282829284668, + "learning_rate": 7.800000000000001e-05, + "loss": 1.4967, + "step": 586 + }, + { + "epoch": 0.039772342299613794, + "grad_norm": 11.466882705688477, + "learning_rate": 7.813333333333334e-05, + "loss": 1.3003, + "step": 587 + }, + { + "epoch": 0.03984009756758588, + "grad_norm": 19.529300689697266, + "learning_rate": 7.826666666666667e-05, + "loss": 1.3835, + "step": 588 + }, + { + "epoch": 0.039907852835557964, + "grad_norm": 17.22064971923828, + "learning_rate": 7.840000000000001e-05, + "loss": 1.2655, + "step": 589 + }, + { + "epoch": 0.03997560810353005, + "grad_norm": 14.788103103637695, + "learning_rate": 7.853333333333334e-05, + "loss": 1.1981, + "step": 590 + }, + { + "epoch": 0.04004336337150213, + "grad_norm": 13.705521583557129, + "learning_rate": 7.866666666666666e-05, + "loss": 1.2064, + "step": 591 + }, + { + "epoch": 0.04011111863947422, + "grad_norm": 14.901930809020996, + "learning_rate": 7.88e-05, + "loss": 1.2632, + "step": 592 + }, + { + "epoch": 0.0401788739074463, + "grad_norm": 18.520828247070312, + "learning_rate": 7.893333333333333e-05, + "loss": 1.4316, + "step": 593 + }, + { + "epoch": 0.04024662917541839, + "grad_norm": 15.101353645324707, + "learning_rate": 7.906666666666667e-05, + "loss": 1.3935, + "step": 594 + }, + { + "epoch": 0.04031438444339047, + "grad_norm": 13.961483001708984, + "learning_rate": 7.920000000000001e-05, + "loss": 1.1477, + "step": 595 + }, + { + "epoch": 0.04038213971136256, + "grad_norm": 16.015554428100586, + "learning_rate": 7.933333333333334e-05, + "loss": 1.5029, + "step": 596 + }, + { + "epoch": 0.04044989497933464, + "grad_norm": 15.009637832641602, + "learning_rate": 7.946666666666667e-05, + "loss": 1.3715, + "step": 597 + }, + { + "epoch": 0.040517650247306726, + "grad_norm": 16.295202255249023, + "learning_rate": 7.960000000000001e-05, + "loss": 1.5129, + "step": 598 + }, + { + "epoch": 0.04058540551527881, + "grad_norm": 88.3338623046875, + "learning_rate": 7.973333333333334e-05, + "loss": 1.2395, + "step": 599 + }, + { + "epoch": 0.040653160783250895, + "grad_norm": 16.769424438476562, + "learning_rate": 7.986666666666667e-05, + "loss": 1.4258, + "step": 600 + }, + { + "epoch": 0.04072091605122298, + "grad_norm": 16.044578552246094, + "learning_rate": 8e-05, + "loss": 1.365, + "step": 601 + }, + { + "epoch": 0.040788671319195065, + "grad_norm": 15.282588958740234, + "learning_rate": 8.013333333333333e-05, + "loss": 1.3212, + "step": 602 + }, + { + "epoch": 0.04085642658716715, + "grad_norm": 15.275490760803223, + "learning_rate": 8.026666666666666e-05, + "loss": 1.6048, + "step": 603 + }, + { + "epoch": 0.040924181855139234, + "grad_norm": 12.952759742736816, + "learning_rate": 8.04e-05, + "loss": 1.1982, + "step": 604 + }, + { + "epoch": 0.04099193712311132, + "grad_norm": 16.79343032836914, + "learning_rate": 8.053333333333334e-05, + "loss": 1.281, + "step": 605 + }, + { + "epoch": 0.041059692391083404, + "grad_norm": 13.70421314239502, + "learning_rate": 8.066666666666667e-05, + "loss": 1.2664, + "step": 606 + }, + { + "epoch": 0.04112744765905549, + "grad_norm": 14.420117378234863, + "learning_rate": 8.080000000000001e-05, + "loss": 1.2758, + "step": 607 + }, + { + "epoch": 0.04119520292702757, + "grad_norm": 16.421335220336914, + "learning_rate": 8.093333333333334e-05, + "loss": 1.3105, + "step": 608 + }, + { + "epoch": 0.041262958194999665, + "grad_norm": 16.96218490600586, + "learning_rate": 8.106666666666667e-05, + "loss": 1.4665, + "step": 609 + }, + { + "epoch": 0.04133071346297175, + "grad_norm": 13.73725700378418, + "learning_rate": 8.120000000000001e-05, + "loss": 1.2109, + "step": 610 + }, + { + "epoch": 0.041398468730943834, + "grad_norm": 14.718022346496582, + "learning_rate": 8.133333333333334e-05, + "loss": 1.2051, + "step": 611 + }, + { + "epoch": 0.04146622399891592, + "grad_norm": 14.119061470031738, + "learning_rate": 8.146666666666666e-05, + "loss": 1.0731, + "step": 612 + }, + { + "epoch": 0.041533979266888, + "grad_norm": 17.90053367614746, + "learning_rate": 8.16e-05, + "loss": 1.3804, + "step": 613 + }, + { + "epoch": 0.04160173453486009, + "grad_norm": 12.727055549621582, + "learning_rate": 8.173333333333335e-05, + "loss": 1.0926, + "step": 614 + }, + { + "epoch": 0.04166948980283217, + "grad_norm": 14.953054428100586, + "learning_rate": 8.186666666666667e-05, + "loss": 1.1974, + "step": 615 + }, + { + "epoch": 0.04173724507080426, + "grad_norm": 16.05322265625, + "learning_rate": 8.2e-05, + "loss": 1.2526, + "step": 616 + }, + { + "epoch": 0.04180500033877634, + "grad_norm": 15.028168678283691, + "learning_rate": 8.213333333333334e-05, + "loss": 1.3593, + "step": 617 + }, + { + "epoch": 0.04187275560674843, + "grad_norm": 13.127458572387695, + "learning_rate": 8.226666666666667e-05, + "loss": 1.3981, + "step": 618 + }, + { + "epoch": 0.04194051087472051, + "grad_norm": 17.323017120361328, + "learning_rate": 8.24e-05, + "loss": 1.4622, + "step": 619 + }, + { + "epoch": 0.042008266142692596, + "grad_norm": 16.106731414794922, + "learning_rate": 8.253333333333334e-05, + "loss": 1.257, + "step": 620 + }, + { + "epoch": 0.04207602141066468, + "grad_norm": 13.039103507995605, + "learning_rate": 8.266666666666667e-05, + "loss": 1.2905, + "step": 621 + }, + { + "epoch": 0.042143776678636766, + "grad_norm": 13.661933898925781, + "learning_rate": 8.28e-05, + "loss": 1.2779, + "step": 622 + }, + { + "epoch": 0.04221153194660885, + "grad_norm": 17.325756072998047, + "learning_rate": 8.293333333333333e-05, + "loss": 1.3934, + "step": 623 + }, + { + "epoch": 0.042279287214580935, + "grad_norm": 13.898777961730957, + "learning_rate": 8.306666666666668e-05, + "loss": 1.3735, + "step": 624 + }, + { + "epoch": 0.04234704248255302, + "grad_norm": 16.787601470947266, + "learning_rate": 8.32e-05, + "loss": 1.5228, + "step": 625 + }, + { + "epoch": 0.042414797750525104, + "grad_norm": 13.54299259185791, + "learning_rate": 8.333333333333334e-05, + "loss": 1.0538, + "step": 626 + }, + { + "epoch": 0.04248255301849719, + "grad_norm": 14.418194770812988, + "learning_rate": 8.346666666666667e-05, + "loss": 1.2932, + "step": 627 + }, + { + "epoch": 0.042550308286469274, + "grad_norm": 13.39255142211914, + "learning_rate": 8.36e-05, + "loss": 1.1989, + "step": 628 + }, + { + "epoch": 0.04261806355444136, + "grad_norm": 15.445747375488281, + "learning_rate": 8.373333333333334e-05, + "loss": 1.3557, + "step": 629 + }, + { + "epoch": 0.04268581882241344, + "grad_norm": 13.414338111877441, + "learning_rate": 8.386666666666667e-05, + "loss": 1.1729, + "step": 630 + }, + { + "epoch": 0.04275357409038553, + "grad_norm": 16.649137496948242, + "learning_rate": 8.4e-05, + "loss": 1.4212, + "step": 631 + }, + { + "epoch": 0.04282132935835761, + "grad_norm": 17.583528518676758, + "learning_rate": 8.413333333333334e-05, + "loss": 1.4838, + "step": 632 + }, + { + "epoch": 0.0428890846263297, + "grad_norm": 19.32307243347168, + "learning_rate": 8.426666666666668e-05, + "loss": 1.5958, + "step": 633 + }, + { + "epoch": 0.04295683989430178, + "grad_norm": 12.703327178955078, + "learning_rate": 8.44e-05, + "loss": 1.2752, + "step": 634 + }, + { + "epoch": 0.04302459516227387, + "grad_norm": 15.72768497467041, + "learning_rate": 8.453333333333335e-05, + "loss": 1.4059, + "step": 635 + }, + { + "epoch": 0.04309235043024595, + "grad_norm": 13.532344818115234, + "learning_rate": 8.466666666666667e-05, + "loss": 1.2252, + "step": 636 + }, + { + "epoch": 0.043160105698218036, + "grad_norm": 13.305481910705566, + "learning_rate": 8.48e-05, + "loss": 1.5084, + "step": 637 + }, + { + "epoch": 0.04322786096619012, + "grad_norm": 11.986043930053711, + "learning_rate": 8.493333333333334e-05, + "loss": 1.2006, + "step": 638 + }, + { + "epoch": 0.043295616234162206, + "grad_norm": 13.28003978729248, + "learning_rate": 8.506666666666667e-05, + "loss": 1.2442, + "step": 639 + }, + { + "epoch": 0.04336337150213429, + "grad_norm": 15.835545539855957, + "learning_rate": 8.52e-05, + "loss": 1.2466, + "step": 640 + }, + { + "epoch": 0.043431126770106375, + "grad_norm": 14.887584686279297, + "learning_rate": 8.533333333333334e-05, + "loss": 1.3828, + "step": 641 + }, + { + "epoch": 0.04349888203807846, + "grad_norm": 14.740251541137695, + "learning_rate": 8.546666666666667e-05, + "loss": 1.4847, + "step": 642 + }, + { + "epoch": 0.043566637306050544, + "grad_norm": 14.109213829040527, + "learning_rate": 8.560000000000001e-05, + "loss": 1.2251, + "step": 643 + }, + { + "epoch": 0.04363439257402263, + "grad_norm": 17.33670425415039, + "learning_rate": 8.573333333333333e-05, + "loss": 1.3817, + "step": 644 + }, + { + "epoch": 0.043702147841994714, + "grad_norm": 16.92241096496582, + "learning_rate": 8.586666666666668e-05, + "loss": 1.458, + "step": 645 + }, + { + "epoch": 0.0437699031099668, + "grad_norm": 15.369187355041504, + "learning_rate": 8.6e-05, + "loss": 1.28, + "step": 646 + }, + { + "epoch": 0.04383765837793888, + "grad_norm": 18.748065948486328, + "learning_rate": 8.613333333333333e-05, + "loss": 1.6275, + "step": 647 + }, + { + "epoch": 0.04390541364591097, + "grad_norm": 15.650605201721191, + "learning_rate": 8.626666666666667e-05, + "loss": 1.3742, + "step": 648 + }, + { + "epoch": 0.04397316891388305, + "grad_norm": 16.170730590820312, + "learning_rate": 8.64e-05, + "loss": 1.174, + "step": 649 + }, + { + "epoch": 0.04404092418185514, + "grad_norm": 12.61952018737793, + "learning_rate": 8.653333333333333e-05, + "loss": 1.2178, + "step": 650 + }, + { + "epoch": 0.04410867944982722, + "grad_norm": 18.86512565612793, + "learning_rate": 8.666666666666667e-05, + "loss": 1.4993, + "step": 651 + }, + { + "epoch": 0.04417643471779931, + "grad_norm": 16.412899017333984, + "learning_rate": 8.680000000000001e-05, + "loss": 1.2663, + "step": 652 + }, + { + "epoch": 0.04424418998577139, + "grad_norm": 14.11953353881836, + "learning_rate": 8.693333333333334e-05, + "loss": 1.3772, + "step": 653 + }, + { + "epoch": 0.044311945253743476, + "grad_norm": 17.232633590698242, + "learning_rate": 8.706666666666668e-05, + "loss": 1.3252, + "step": 654 + }, + { + "epoch": 0.04437970052171556, + "grad_norm": 13.833864212036133, + "learning_rate": 8.72e-05, + "loss": 1.279, + "step": 655 + }, + { + "epoch": 0.044447455789687645, + "grad_norm": 11.546829223632812, + "learning_rate": 8.733333333333333e-05, + "loss": 1.382, + "step": 656 + }, + { + "epoch": 0.04451521105765973, + "grad_norm": 13.163644790649414, + "learning_rate": 8.746666666666667e-05, + "loss": 1.2253, + "step": 657 + }, + { + "epoch": 0.044582966325631815, + "grad_norm": 17.037311553955078, + "learning_rate": 8.76e-05, + "loss": 1.2727, + "step": 658 + }, + { + "epoch": 0.0446507215936039, + "grad_norm": 14.610177040100098, + "learning_rate": 8.773333333333333e-05, + "loss": 1.3082, + "step": 659 + }, + { + "epoch": 0.044718476861575984, + "grad_norm": 16.305557250976562, + "learning_rate": 8.786666666666667e-05, + "loss": 1.3948, + "step": 660 + }, + { + "epoch": 0.04478623212954807, + "grad_norm": 13.207799911499023, + "learning_rate": 8.800000000000001e-05, + "loss": 1.2764, + "step": 661 + }, + { + "epoch": 0.04485398739752016, + "grad_norm": 11.451075553894043, + "learning_rate": 8.813333333333334e-05, + "loss": 1.0831, + "step": 662 + }, + { + "epoch": 0.044921742665492245, + "grad_norm": 13.555370330810547, + "learning_rate": 8.826666666666668e-05, + "loss": 1.4996, + "step": 663 + }, + { + "epoch": 0.04498949793346433, + "grad_norm": 13.544769287109375, + "learning_rate": 8.840000000000001e-05, + "loss": 1.2277, + "step": 664 + }, + { + "epoch": 0.045057253201436415, + "grad_norm": 18.05879783630371, + "learning_rate": 8.853333333333333e-05, + "loss": 1.3964, + "step": 665 + }, + { + "epoch": 0.0451250084694085, + "grad_norm": 17.309839248657227, + "learning_rate": 8.866666666666668e-05, + "loss": 1.6233, + "step": 666 + }, + { + "epoch": 0.045192763737380584, + "grad_norm": 12.732510566711426, + "learning_rate": 8.88e-05, + "loss": 1.3237, + "step": 667 + }, + { + "epoch": 0.04526051900535267, + "grad_norm": 13.541101455688477, + "learning_rate": 8.893333333333333e-05, + "loss": 1.2523, + "step": 668 + }, + { + "epoch": 0.04532827427332475, + "grad_norm": 17.54905891418457, + "learning_rate": 8.906666666666667e-05, + "loss": 1.5824, + "step": 669 + }, + { + "epoch": 0.04539602954129684, + "grad_norm": 12.52578353881836, + "learning_rate": 8.92e-05, + "loss": 1.2966, + "step": 670 + }, + { + "epoch": 0.04546378480926892, + "grad_norm": 13.279097557067871, + "learning_rate": 8.933333333333334e-05, + "loss": 1.4898, + "step": 671 + }, + { + "epoch": 0.04553154007724101, + "grad_norm": 15.892850875854492, + "learning_rate": 8.946666666666668e-05, + "loss": 1.3757, + "step": 672 + }, + { + "epoch": 0.04559929534521309, + "grad_norm": 14.108098983764648, + "learning_rate": 8.960000000000001e-05, + "loss": 1.1758, + "step": 673 + }, + { + "epoch": 0.04566705061318518, + "grad_norm": 17.15204429626465, + "learning_rate": 8.973333333333334e-05, + "loss": 1.3876, + "step": 674 + }, + { + "epoch": 0.04573480588115726, + "grad_norm": 14.453113555908203, + "learning_rate": 8.986666666666666e-05, + "loss": 1.2152, + "step": 675 + }, + { + "epoch": 0.045802561149129346, + "grad_norm": 17.9672794342041, + "learning_rate": 9e-05, + "loss": 1.2962, + "step": 676 + }, + { + "epoch": 0.04587031641710143, + "grad_norm": 19.810890197753906, + "learning_rate": 9.013333333333333e-05, + "loss": 1.3374, + "step": 677 + }, + { + "epoch": 0.045938071685073516, + "grad_norm": 16.13353729248047, + "learning_rate": 9.026666666666666e-05, + "loss": 1.4125, + "step": 678 + }, + { + "epoch": 0.0460058269530456, + "grad_norm": 15.257608413696289, + "learning_rate": 9.04e-05, + "loss": 1.1931, + "step": 679 + }, + { + "epoch": 0.046073582221017685, + "grad_norm": 16.88699722290039, + "learning_rate": 9.053333333333334e-05, + "loss": 1.362, + "step": 680 + }, + { + "epoch": 0.04614133748898977, + "grad_norm": 15.46777057647705, + "learning_rate": 9.066666666666667e-05, + "loss": 1.3522, + "step": 681 + }, + { + "epoch": 0.046209092756961855, + "grad_norm": 13.584056854248047, + "learning_rate": 9.080000000000001e-05, + "loss": 1.1998, + "step": 682 + }, + { + "epoch": 0.04627684802493394, + "grad_norm": 14.226449966430664, + "learning_rate": 9.093333333333334e-05, + "loss": 1.3312, + "step": 683 + }, + { + "epoch": 0.046344603292906024, + "grad_norm": 15.157097816467285, + "learning_rate": 9.106666666666667e-05, + "loss": 1.3422, + "step": 684 + }, + { + "epoch": 0.04641235856087811, + "grad_norm": 14.748275756835938, + "learning_rate": 9.120000000000001e-05, + "loss": 1.2902, + "step": 685 + }, + { + "epoch": 0.04648011382885019, + "grad_norm": 13.730619430541992, + "learning_rate": 9.133333333333334e-05, + "loss": 1.3034, + "step": 686 + }, + { + "epoch": 0.04654786909682228, + "grad_norm": 12.804062843322754, + "learning_rate": 9.146666666666666e-05, + "loss": 1.1484, + "step": 687 + }, + { + "epoch": 0.04661562436479436, + "grad_norm": 16.587923049926758, + "learning_rate": 9.16e-05, + "loss": 1.4875, + "step": 688 + }, + { + "epoch": 0.04668337963276645, + "grad_norm": 12.228131294250488, + "learning_rate": 9.173333333333333e-05, + "loss": 0.917, + "step": 689 + }, + { + "epoch": 0.04675113490073853, + "grad_norm": 18.680187225341797, + "learning_rate": 9.186666666666667e-05, + "loss": 1.497, + "step": 690 + }, + { + "epoch": 0.04681889016871062, + "grad_norm": 14.80630111694336, + "learning_rate": 9.200000000000001e-05, + "loss": 1.5623, + "step": 691 + }, + { + "epoch": 0.0468866454366827, + "grad_norm": 13.754642486572266, + "learning_rate": 9.213333333333334e-05, + "loss": 1.3921, + "step": 692 + }, + { + "epoch": 0.046954400704654786, + "grad_norm": 14.264236450195312, + "learning_rate": 9.226666666666667e-05, + "loss": 1.3615, + "step": 693 + }, + { + "epoch": 0.04702215597262687, + "grad_norm": 15.883113861083984, + "learning_rate": 9.240000000000001e-05, + "loss": 1.4528, + "step": 694 + }, + { + "epoch": 0.047089911240598956, + "grad_norm": 13.528610229492188, + "learning_rate": 9.253333333333334e-05, + "loss": 1.0995, + "step": 695 + }, + { + "epoch": 0.04715766650857104, + "grad_norm": 15.945343971252441, + "learning_rate": 9.266666666666666e-05, + "loss": 1.272, + "step": 696 + }, + { + "epoch": 0.047225421776543125, + "grad_norm": 14.378050804138184, + "learning_rate": 9.28e-05, + "loss": 1.3545, + "step": 697 + }, + { + "epoch": 0.04729317704451521, + "grad_norm": 13.457077980041504, + "learning_rate": 9.293333333333333e-05, + "loss": 1.1751, + "step": 698 + }, + { + "epoch": 0.047360932312487294, + "grad_norm": 19.285078048706055, + "learning_rate": 9.306666666666667e-05, + "loss": 1.4383, + "step": 699 + }, + { + "epoch": 0.04742868758045938, + "grad_norm": 16.683856964111328, + "learning_rate": 9.320000000000002e-05, + "loss": 1.2499, + "step": 700 + }, + { + "epoch": 0.047496442848431464, + "grad_norm": 13.818337440490723, + "learning_rate": 9.333333333333334e-05, + "loss": 1.2325, + "step": 701 + }, + { + "epoch": 0.04756419811640355, + "grad_norm": 12.51142406463623, + "learning_rate": 9.346666666666667e-05, + "loss": 1.3317, + "step": 702 + }, + { + "epoch": 0.04763195338437563, + "grad_norm": 14.697171211242676, + "learning_rate": 9.360000000000001e-05, + "loss": 1.6055, + "step": 703 + }, + { + "epoch": 0.04769970865234772, + "grad_norm": 16.942562103271484, + "learning_rate": 9.373333333333334e-05, + "loss": 1.4279, + "step": 704 + }, + { + "epoch": 0.0477674639203198, + "grad_norm": 16.739248275756836, + "learning_rate": 9.386666666666667e-05, + "loss": 1.2285, + "step": 705 + }, + { + "epoch": 0.04783521918829189, + "grad_norm": 14.479548454284668, + "learning_rate": 9.4e-05, + "loss": 1.1984, + "step": 706 + }, + { + "epoch": 0.04790297445626397, + "grad_norm": 14.384824752807617, + "learning_rate": 9.413333333333334e-05, + "loss": 1.2128, + "step": 707 + }, + { + "epoch": 0.04797072972423606, + "grad_norm": 15.41884994506836, + "learning_rate": 9.426666666666666e-05, + "loss": 1.1627, + "step": 708 + }, + { + "epoch": 0.04803848499220814, + "grad_norm": 13.267902374267578, + "learning_rate": 9.44e-05, + "loss": 1.2916, + "step": 709 + }, + { + "epoch": 0.048106240260180226, + "grad_norm": 14.909900665283203, + "learning_rate": 9.453333333333335e-05, + "loss": 1.1579, + "step": 710 + }, + { + "epoch": 0.04817399552815231, + "grad_norm": 12.639839172363281, + "learning_rate": 9.466666666666667e-05, + "loss": 1.2053, + "step": 711 + }, + { + "epoch": 0.048241750796124395, + "grad_norm": 14.957659721374512, + "learning_rate": 9.48e-05, + "loss": 1.4556, + "step": 712 + }, + { + "epoch": 0.04830950606409648, + "grad_norm": 14.008201599121094, + "learning_rate": 9.493333333333334e-05, + "loss": 1.1575, + "step": 713 + }, + { + "epoch": 0.048377261332068565, + "grad_norm": 13.857501983642578, + "learning_rate": 9.506666666666667e-05, + "loss": 1.4121, + "step": 714 + }, + { + "epoch": 0.048445016600040657, + "grad_norm": 13.867570877075195, + "learning_rate": 9.52e-05, + "loss": 1.4908, + "step": 715 + }, + { + "epoch": 0.04851277186801274, + "grad_norm": 14.3545503616333, + "learning_rate": 9.533333333333334e-05, + "loss": 1.2939, + "step": 716 + }, + { + "epoch": 0.048580527135984826, + "grad_norm": 15.173371315002441, + "learning_rate": 9.546666666666667e-05, + "loss": 1.3912, + "step": 717 + }, + { + "epoch": 0.04864828240395691, + "grad_norm": 13.799453735351562, + "learning_rate": 9.56e-05, + "loss": 1.4219, + "step": 718 + }, + { + "epoch": 0.048716037671928995, + "grad_norm": 18.264965057373047, + "learning_rate": 9.573333333333335e-05, + "loss": 1.4967, + "step": 719 + }, + { + "epoch": 0.04878379293990108, + "grad_norm": 13.809319496154785, + "learning_rate": 9.586666666666667e-05, + "loss": 1.3527, + "step": 720 + }, + { + "epoch": 0.048851548207873165, + "grad_norm": 13.931517601013184, + "learning_rate": 9.6e-05, + "loss": 1.4018, + "step": 721 + }, + { + "epoch": 0.04891930347584525, + "grad_norm": 11.74387264251709, + "learning_rate": 9.613333333333334e-05, + "loss": 1.1981, + "step": 722 + }, + { + "epoch": 0.048987058743817334, + "grad_norm": 14.960589408874512, + "learning_rate": 9.626666666666667e-05, + "loss": 1.3373, + "step": 723 + }, + { + "epoch": 0.04905481401178942, + "grad_norm": 14.170279502868652, + "learning_rate": 9.64e-05, + "loss": 1.2702, + "step": 724 + }, + { + "epoch": 0.049122569279761504, + "grad_norm": 15.532767295837402, + "learning_rate": 9.653333333333334e-05, + "loss": 1.1329, + "step": 725 + }, + { + "epoch": 0.04919032454773359, + "grad_norm": 12.141284942626953, + "learning_rate": 9.666666666666667e-05, + "loss": 1.2694, + "step": 726 + }, + { + "epoch": 0.04925807981570567, + "grad_norm": 12.379233360290527, + "learning_rate": 9.680000000000001e-05, + "loss": 1.1545, + "step": 727 + }, + { + "epoch": 0.04932583508367776, + "grad_norm": 14.45605182647705, + "learning_rate": 9.693333333333335e-05, + "loss": 1.4151, + "step": 728 + }, + { + "epoch": 0.04939359035164984, + "grad_norm": 15.656912803649902, + "learning_rate": 9.706666666666668e-05, + "loss": 1.3782, + "step": 729 + }, + { + "epoch": 0.04946134561962193, + "grad_norm": 13.992748260498047, + "learning_rate": 9.72e-05, + "loss": 1.2731, + "step": 730 + }, + { + "epoch": 0.04952910088759401, + "grad_norm": 16.253517150878906, + "learning_rate": 9.733333333333335e-05, + "loss": 1.4503, + "step": 731 + }, + { + "epoch": 0.049596856155566096, + "grad_norm": 14.88161563873291, + "learning_rate": 9.746666666666667e-05, + "loss": 1.3909, + "step": 732 + }, + { + "epoch": 0.04966461142353818, + "grad_norm": 14.60545539855957, + "learning_rate": 9.76e-05, + "loss": 1.1803, + "step": 733 + }, + { + "epoch": 0.049732366691510266, + "grad_norm": 14.589803695678711, + "learning_rate": 9.773333333333334e-05, + "loss": 1.3844, + "step": 734 + }, + { + "epoch": 0.04980012195948235, + "grad_norm": 14.781076431274414, + "learning_rate": 9.786666666666667e-05, + "loss": 1.2706, + "step": 735 + }, + { + "epoch": 0.049867877227454435, + "grad_norm": 13.717072486877441, + "learning_rate": 9.8e-05, + "loss": 1.2311, + "step": 736 + }, + { + "epoch": 0.04993563249542652, + "grad_norm": 17.14280891418457, + "learning_rate": 9.813333333333334e-05, + "loss": 1.3423, + "step": 737 + }, + { + "epoch": 0.050003387763398605, + "grad_norm": 14.521373748779297, + "learning_rate": 9.826666666666668e-05, + "loss": 1.3118, + "step": 738 + }, + { + "epoch": 0.05007114303137069, + "grad_norm": 14.79515266418457, + "learning_rate": 9.84e-05, + "loss": 1.3622, + "step": 739 + }, + { + "epoch": 0.050138898299342774, + "grad_norm": 13.63962459564209, + "learning_rate": 9.853333333333333e-05, + "loss": 1.3856, + "step": 740 + }, + { + "epoch": 0.05020665356731486, + "grad_norm": 13.788541793823242, + "learning_rate": 9.866666666666668e-05, + "loss": 1.3439, + "step": 741 + }, + { + "epoch": 0.05027440883528694, + "grad_norm": 14.015862464904785, + "learning_rate": 9.88e-05, + "loss": 1.3263, + "step": 742 + }, + { + "epoch": 0.05034216410325903, + "grad_norm": 15.407167434692383, + "learning_rate": 9.893333333333333e-05, + "loss": 1.5115, + "step": 743 + }, + { + "epoch": 0.05040991937123111, + "grad_norm": 14.542003631591797, + "learning_rate": 9.906666666666667e-05, + "loss": 1.3934, + "step": 744 + }, + { + "epoch": 0.0504776746392032, + "grad_norm": 16.733686447143555, + "learning_rate": 9.92e-05, + "loss": 1.2098, + "step": 745 + }, + { + "epoch": 0.05054542990717528, + "grad_norm": 14.979568481445312, + "learning_rate": 9.933333333333334e-05, + "loss": 1.3234, + "step": 746 + }, + { + "epoch": 0.05061318517514737, + "grad_norm": 13.699801445007324, + "learning_rate": 9.946666666666668e-05, + "loss": 1.3173, + "step": 747 + }, + { + "epoch": 0.05068094044311945, + "grad_norm": 13.262088775634766, + "learning_rate": 9.960000000000001e-05, + "loss": 1.3459, + "step": 748 + }, + { + "epoch": 0.050748695711091536, + "grad_norm": 14.694448471069336, + "learning_rate": 9.973333333333334e-05, + "loss": 1.3639, + "step": 749 + }, + { + "epoch": 0.05081645097906362, + "grad_norm": 13.999008178710938, + "learning_rate": 9.986666666666668e-05, + "loss": 1.4275, + "step": 750 + }, + { + "epoch": 0.050884206247035706, + "grad_norm": 11.588037490844727, + "learning_rate": 0.0001, + "loss": 1.2866, + "step": 751 + }, + { + "epoch": 0.05095196151500779, + "grad_norm": 16.174110412597656, + "learning_rate": 9.999863098090219e-05, + "loss": 1.5338, + "step": 752 + }, + { + "epoch": 0.051019716782979875, + "grad_norm": 17.8669376373291, + "learning_rate": 9.999726196180437e-05, + "loss": 1.3247, + "step": 753 + }, + { + "epoch": 0.05108747205095196, + "grad_norm": 13.849075317382812, + "learning_rate": 9.999589294270656e-05, + "loss": 1.1817, + "step": 754 + }, + { + "epoch": 0.051155227318924044, + "grad_norm": 14.888331413269043, + "learning_rate": 9.999452392360874e-05, + "loss": 1.4019, + "step": 755 + }, + { + "epoch": 0.05122298258689613, + "grad_norm": 18.68206024169922, + "learning_rate": 9.999315490451092e-05, + "loss": 1.7115, + "step": 756 + }, + { + "epoch": 0.051290737854868214, + "grad_norm": 14.762079238891602, + "learning_rate": 9.99917858854131e-05, + "loss": 1.2553, + "step": 757 + }, + { + "epoch": 0.0513584931228403, + "grad_norm": 14.649972915649414, + "learning_rate": 9.99904168663153e-05, + "loss": 1.0807, + "step": 758 + }, + { + "epoch": 0.05142624839081238, + "grad_norm": 13.172977447509766, + "learning_rate": 9.998904784721747e-05, + "loss": 1.4966, + "step": 759 + }, + { + "epoch": 0.05149400365878447, + "grad_norm": 14.144796371459961, + "learning_rate": 9.998767882811965e-05, + "loss": 1.234, + "step": 760 + }, + { + "epoch": 0.05156175892675655, + "grad_norm": 17.565507888793945, + "learning_rate": 9.998630980902184e-05, + "loss": 1.4061, + "step": 761 + }, + { + "epoch": 0.05162951419472864, + "grad_norm": 17.205589294433594, + "learning_rate": 9.998494078992402e-05, + "loss": 1.6626, + "step": 762 + }, + { + "epoch": 0.05169726946270072, + "grad_norm": 16.14542007446289, + "learning_rate": 9.998357177082621e-05, + "loss": 1.3145, + "step": 763 + }, + { + "epoch": 0.05176502473067281, + "grad_norm": 14.692976951599121, + "learning_rate": 9.998220275172839e-05, + "loss": 1.2085, + "step": 764 + }, + { + "epoch": 0.05183277999864489, + "grad_norm": 12.572774887084961, + "learning_rate": 9.998083373263057e-05, + "loss": 1.3743, + "step": 765 + }, + { + "epoch": 0.051900535266616976, + "grad_norm": 15.904753684997559, + "learning_rate": 9.997946471353275e-05, + "loss": 1.3863, + "step": 766 + }, + { + "epoch": 0.05196829053458906, + "grad_norm": 14.471494674682617, + "learning_rate": 9.997809569443493e-05, + "loss": 1.3615, + "step": 767 + }, + { + "epoch": 0.05203604580256115, + "grad_norm": 13.82172966003418, + "learning_rate": 9.997672667533712e-05, + "loss": 1.2104, + "step": 768 + }, + { + "epoch": 0.05210380107053324, + "grad_norm": 14.648073196411133, + "learning_rate": 9.99753576562393e-05, + "loss": 1.1166, + "step": 769 + }, + { + "epoch": 0.05217155633850532, + "grad_norm": 16.776750564575195, + "learning_rate": 9.997398863714149e-05, + "loss": 1.4242, + "step": 770 + }, + { + "epoch": 0.05223931160647741, + "grad_norm": 13.999717712402344, + "learning_rate": 9.997261961804367e-05, + "loss": 1.2493, + "step": 771 + }, + { + "epoch": 0.05230706687444949, + "grad_norm": 13.238443374633789, + "learning_rate": 9.997125059894586e-05, + "loss": 1.0469, + "step": 772 + }, + { + "epoch": 0.052374822142421576, + "grad_norm": 12.79283332824707, + "learning_rate": 9.996988157984804e-05, + "loss": 1.1975, + "step": 773 + }, + { + "epoch": 0.05244257741039366, + "grad_norm": 15.577791213989258, + "learning_rate": 9.996851256075022e-05, + "loss": 1.4972, + "step": 774 + }, + { + "epoch": 0.052510332678365745, + "grad_norm": 16.215747833251953, + "learning_rate": 9.996714354165241e-05, + "loss": 1.4285, + "step": 775 + }, + { + "epoch": 0.05257808794633783, + "grad_norm": 14.68825626373291, + "learning_rate": 9.99657745225546e-05, + "loss": 1.2028, + "step": 776 + }, + { + "epoch": 0.052645843214309915, + "grad_norm": 11.535130500793457, + "learning_rate": 9.996440550345677e-05, + "loss": 1.2466, + "step": 777 + }, + { + "epoch": 0.052713598482282, + "grad_norm": 14.865918159484863, + "learning_rate": 9.996303648435897e-05, + "loss": 1.3903, + "step": 778 + }, + { + "epoch": 0.052781353750254084, + "grad_norm": 13.626388549804688, + "learning_rate": 9.996166746526115e-05, + "loss": 1.5446, + "step": 779 + }, + { + "epoch": 0.05284910901822617, + "grad_norm": 12.957473754882812, + "learning_rate": 9.996029844616333e-05, + "loss": 1.1641, + "step": 780 + }, + { + "epoch": 0.052916864286198254, + "grad_norm": 14.928487777709961, + "learning_rate": 9.995892942706552e-05, + "loss": 1.0929, + "step": 781 + }, + { + "epoch": 0.05298461955417034, + "grad_norm": 13.592952728271484, + "learning_rate": 9.99575604079677e-05, + "loss": 1.4684, + "step": 782 + }, + { + "epoch": 0.05305237482214242, + "grad_norm": 13.836453437805176, + "learning_rate": 9.995619138886988e-05, + "loss": 1.0128, + "step": 783 + }, + { + "epoch": 0.05312013009011451, + "grad_norm": 16.654394149780273, + "learning_rate": 9.995482236977206e-05, + "loss": 1.2464, + "step": 784 + }, + { + "epoch": 0.05318788535808659, + "grad_norm": 14.4434232711792, + "learning_rate": 9.995345335067424e-05, + "loss": 1.3649, + "step": 785 + }, + { + "epoch": 0.05325564062605868, + "grad_norm": 14.15592098236084, + "learning_rate": 9.995208433157644e-05, + "loss": 1.4426, + "step": 786 + }, + { + "epoch": 0.05332339589403076, + "grad_norm": 12.97901439666748, + "learning_rate": 9.995071531247862e-05, + "loss": 1.227, + "step": 787 + }, + { + "epoch": 0.053391151162002846, + "grad_norm": 13.167902946472168, + "learning_rate": 9.99493462933808e-05, + "loss": 1.2432, + "step": 788 + }, + { + "epoch": 0.05345890642997493, + "grad_norm": 15.92297077178955, + "learning_rate": 9.994797727428298e-05, + "loss": 1.3345, + "step": 789 + }, + { + "epoch": 0.053526661697947016, + "grad_norm": 13.30277156829834, + "learning_rate": 9.994660825518517e-05, + "loss": 1.2361, + "step": 790 + }, + { + "epoch": 0.0535944169659191, + "grad_norm": 15.98779582977295, + "learning_rate": 9.994523923608735e-05, + "loss": 1.1202, + "step": 791 + }, + { + "epoch": 0.053662172233891185, + "grad_norm": 16.414382934570312, + "learning_rate": 9.994387021698953e-05, + "loss": 1.448, + "step": 792 + }, + { + "epoch": 0.05372992750186327, + "grad_norm": 14.491677284240723, + "learning_rate": 9.994250119789171e-05, + "loss": 1.0862, + "step": 793 + }, + { + "epoch": 0.053797682769835355, + "grad_norm": 13.155410766601562, + "learning_rate": 9.99411321787939e-05, + "loss": 1.312, + "step": 794 + }, + { + "epoch": 0.05386543803780744, + "grad_norm": 16.11139488220215, + "learning_rate": 9.993976315969609e-05, + "loss": 1.2043, + "step": 795 + }, + { + "epoch": 0.053933193305779524, + "grad_norm": 14.259698867797852, + "learning_rate": 9.993839414059827e-05, + "loss": 1.2256, + "step": 796 + }, + { + "epoch": 0.05400094857375161, + "grad_norm": 15.751099586486816, + "learning_rate": 9.993702512150045e-05, + "loss": 1.0732, + "step": 797 + }, + { + "epoch": 0.05406870384172369, + "grad_norm": 12.766170501708984, + "learning_rate": 9.993565610240263e-05, + "loss": 1.0362, + "step": 798 + }, + { + "epoch": 0.05413645910969578, + "grad_norm": 14.61483097076416, + "learning_rate": 9.993428708330481e-05, + "loss": 1.3107, + "step": 799 + }, + { + "epoch": 0.05420421437766786, + "grad_norm": 14.571990013122559, + "learning_rate": 9.9932918064207e-05, + "loss": 1.3616, + "step": 800 + }, + { + "epoch": 0.05427196964563995, + "grad_norm": 14.69124984741211, + "learning_rate": 9.993154904510918e-05, + "loss": 1.4273, + "step": 801 + }, + { + "epoch": 0.05433972491361203, + "grad_norm": 13.834383010864258, + "learning_rate": 9.993018002601136e-05, + "loss": 1.2886, + "step": 802 + }, + { + "epoch": 0.05440748018158412, + "grad_norm": 14.178943634033203, + "learning_rate": 9.992881100691355e-05, + "loss": 1.44, + "step": 803 + }, + { + "epoch": 0.0544752354495562, + "grad_norm": 13.185267448425293, + "learning_rate": 9.992744198781574e-05, + "loss": 1.5609, + "step": 804 + }, + { + "epoch": 0.054542990717528286, + "grad_norm": 15.211031913757324, + "learning_rate": 9.992607296871792e-05, + "loss": 1.8873, + "step": 805 + }, + { + "epoch": 0.05461074598550037, + "grad_norm": 11.855447769165039, + "learning_rate": 9.99247039496201e-05, + "loss": 1.4063, + "step": 806 + }, + { + "epoch": 0.054678501253472456, + "grad_norm": 14.352787017822266, + "learning_rate": 9.992333493052228e-05, + "loss": 1.56, + "step": 807 + }, + { + "epoch": 0.05474625652144454, + "grad_norm": 14.339908599853516, + "learning_rate": 9.992196591142446e-05, + "loss": 1.1985, + "step": 808 + }, + { + "epoch": 0.054814011789416625, + "grad_norm": 12.696648597717285, + "learning_rate": 9.992059689232665e-05, + "loss": 1.1807, + "step": 809 + }, + { + "epoch": 0.05488176705738871, + "grad_norm": 14.083525657653809, + "learning_rate": 9.991922787322883e-05, + "loss": 1.4012, + "step": 810 + }, + { + "epoch": 0.054949522325360795, + "grad_norm": 14.298514366149902, + "learning_rate": 9.991785885413101e-05, + "loss": 1.4047, + "step": 811 + }, + { + "epoch": 0.05501727759333288, + "grad_norm": 13.391324996948242, + "learning_rate": 9.99164898350332e-05, + "loss": 1.2292, + "step": 812 + }, + { + "epoch": 0.055085032861304964, + "grad_norm": 14.0011625289917, + "learning_rate": 9.991512081593539e-05, + "loss": 1.3194, + "step": 813 + }, + { + "epoch": 0.05515278812927705, + "grad_norm": 14.726574897766113, + "learning_rate": 9.991375179683757e-05, + "loss": 1.3245, + "step": 814 + }, + { + "epoch": 0.05522054339724913, + "grad_norm": 14.615242004394531, + "learning_rate": 9.991238277773975e-05, + "loss": 1.3704, + "step": 815 + }, + { + "epoch": 0.05528829866522122, + "grad_norm": 11.06546688079834, + "learning_rate": 9.991101375864193e-05, + "loss": 1.0841, + "step": 816 + }, + { + "epoch": 0.0553560539331933, + "grad_norm": 13.768633842468262, + "learning_rate": 9.990964473954411e-05, + "loss": 1.2632, + "step": 817 + }, + { + "epoch": 0.05542380920116539, + "grad_norm": 15.054973602294922, + "learning_rate": 9.99082757204463e-05, + "loss": 1.2931, + "step": 818 + }, + { + "epoch": 0.05549156446913747, + "grad_norm": 15.27096176147461, + "learning_rate": 9.990690670134848e-05, + "loss": 1.1789, + "step": 819 + }, + { + "epoch": 0.055559319737109564, + "grad_norm": 13.228281021118164, + "learning_rate": 9.990553768225067e-05, + "loss": 1.4458, + "step": 820 + }, + { + "epoch": 0.05562707500508165, + "grad_norm": 16.016782760620117, + "learning_rate": 9.990416866315286e-05, + "loss": 1.5121, + "step": 821 + }, + { + "epoch": 0.05569483027305373, + "grad_norm": 14.15912914276123, + "learning_rate": 9.990279964405504e-05, + "loss": 1.3702, + "step": 822 + }, + { + "epoch": 0.05576258554102582, + "grad_norm": 14.61017894744873, + "learning_rate": 9.990143062495722e-05, + "loss": 1.2774, + "step": 823 + }, + { + "epoch": 0.0558303408089979, + "grad_norm": 13.241140365600586, + "learning_rate": 9.990006160585941e-05, + "loss": 1.3857, + "step": 824 + }, + { + "epoch": 0.05589809607696999, + "grad_norm": 11.815064430236816, + "learning_rate": 9.98986925867616e-05, + "loss": 1.1053, + "step": 825 + }, + { + "epoch": 0.05596585134494207, + "grad_norm": 13.179222106933594, + "learning_rate": 9.989732356766377e-05, + "loss": 1.4686, + "step": 826 + }, + { + "epoch": 0.05603360661291416, + "grad_norm": 11.888179779052734, + "learning_rate": 9.989595454856597e-05, + "loss": 1.3191, + "step": 827 + }, + { + "epoch": 0.05610136188088624, + "grad_norm": 15.43813419342041, + "learning_rate": 9.989458552946815e-05, + "loss": 0.9732, + "step": 828 + }, + { + "epoch": 0.056169117148858326, + "grad_norm": 20.0246524810791, + "learning_rate": 9.989321651037033e-05, + "loss": 1.5398, + "step": 829 + }, + { + "epoch": 0.05623687241683041, + "grad_norm": 14.247052192687988, + "learning_rate": 9.989184749127251e-05, + "loss": 1.5093, + "step": 830 + }, + { + "epoch": 0.056304627684802495, + "grad_norm": 15.63775634765625, + "learning_rate": 9.989047847217469e-05, + "loss": 1.3488, + "step": 831 + }, + { + "epoch": 0.05637238295277458, + "grad_norm": 10.582650184631348, + "learning_rate": 9.988910945307688e-05, + "loss": 1.3888, + "step": 832 + }, + { + "epoch": 0.056440138220746665, + "grad_norm": 12.344864845275879, + "learning_rate": 9.988774043397906e-05, + "loss": 1.2876, + "step": 833 + }, + { + "epoch": 0.05650789348871875, + "grad_norm": 13.95814323425293, + "learning_rate": 9.988637141488124e-05, + "loss": 1.2648, + "step": 834 + }, + { + "epoch": 0.056575648756690834, + "grad_norm": 9.740105628967285, + "learning_rate": 9.988500239578342e-05, + "loss": 0.9777, + "step": 835 + }, + { + "epoch": 0.05664340402466292, + "grad_norm": 16.16588020324707, + "learning_rate": 9.988363337668562e-05, + "loss": 1.5859, + "step": 836 + }, + { + "epoch": 0.056711159292635004, + "grad_norm": 14.394731521606445, + "learning_rate": 9.98822643575878e-05, + "loss": 1.3395, + "step": 837 + }, + { + "epoch": 0.05677891456060709, + "grad_norm": 11.256571769714355, + "learning_rate": 9.988089533848998e-05, + "loss": 1.0528, + "step": 838 + }, + { + "epoch": 0.05684666982857917, + "grad_norm": 17.595510482788086, + "learning_rate": 9.987952631939216e-05, + "loss": 1.2678, + "step": 839 + }, + { + "epoch": 0.05691442509655126, + "grad_norm": 14.132645606994629, + "learning_rate": 9.987815730029434e-05, + "loss": 1.4129, + "step": 840 + }, + { + "epoch": 0.05698218036452334, + "grad_norm": 14.438119888305664, + "learning_rate": 9.987678828119653e-05, + "loss": 1.4141, + "step": 841 + }, + { + "epoch": 0.05704993563249543, + "grad_norm": 15.983094215393066, + "learning_rate": 9.987541926209871e-05, + "loss": 1.1959, + "step": 842 + }, + { + "epoch": 0.05711769090046751, + "grad_norm": 17.300403594970703, + "learning_rate": 9.98740502430009e-05, + "loss": 1.1613, + "step": 843 + }, + { + "epoch": 0.057185446168439596, + "grad_norm": 15.150660514831543, + "learning_rate": 9.987268122390307e-05, + "loss": 1.2253, + "step": 844 + }, + { + "epoch": 0.05725320143641168, + "grad_norm": 14.234586715698242, + "learning_rate": 9.987131220480525e-05, + "loss": 1.2517, + "step": 845 + }, + { + "epoch": 0.057320956704383766, + "grad_norm": 18.31337547302246, + "learning_rate": 9.986994318570745e-05, + "loss": 1.7506, + "step": 846 + }, + { + "epoch": 0.05738871197235585, + "grad_norm": 14.818669319152832, + "learning_rate": 9.986857416660963e-05, + "loss": 1.358, + "step": 847 + }, + { + "epoch": 0.057456467240327935, + "grad_norm": 14.640913009643555, + "learning_rate": 9.986720514751181e-05, + "loss": 1.4481, + "step": 848 + }, + { + "epoch": 0.05752422250830002, + "grad_norm": 12.277986526489258, + "learning_rate": 9.986583612841399e-05, + "loss": 1.3428, + "step": 849 + }, + { + "epoch": 0.057591977776272105, + "grad_norm": 13.817851066589355, + "learning_rate": 9.986446710931618e-05, + "loss": 1.3514, + "step": 850 + }, + { + "epoch": 0.05765973304424419, + "grad_norm": 13.706515312194824, + "learning_rate": 9.986309809021836e-05, + "loss": 1.4295, + "step": 851 + }, + { + "epoch": 0.057727488312216274, + "grad_norm": 16.649917602539062, + "learning_rate": 9.986172907112054e-05, + "loss": 1.3895, + "step": 852 + }, + { + "epoch": 0.05779524358018836, + "grad_norm": 13.659167289733887, + "learning_rate": 9.986036005202272e-05, + "loss": 1.1545, + "step": 853 + }, + { + "epoch": 0.057862998848160443, + "grad_norm": 11.264912605285645, + "learning_rate": 9.98589910329249e-05, + "loss": 1.0766, + "step": 854 + }, + { + "epoch": 0.05793075411613253, + "grad_norm": 12.930856704711914, + "learning_rate": 9.98576220138271e-05, + "loss": 1.1665, + "step": 855 + }, + { + "epoch": 0.05799850938410461, + "grad_norm": 15.999971389770508, + "learning_rate": 9.985625299472928e-05, + "loss": 1.4595, + "step": 856 + }, + { + "epoch": 0.0580662646520767, + "grad_norm": 14.566671371459961, + "learning_rate": 9.985488397563146e-05, + "loss": 1.283, + "step": 857 + }, + { + "epoch": 0.05813401992004878, + "grad_norm": 16.106964111328125, + "learning_rate": 9.985351495653364e-05, + "loss": 1.2842, + "step": 858 + }, + { + "epoch": 0.05820177518802087, + "grad_norm": 15.47492790222168, + "learning_rate": 9.985214593743583e-05, + "loss": 1.3949, + "step": 859 + }, + { + "epoch": 0.05826953045599295, + "grad_norm": 10.692886352539062, + "learning_rate": 9.985077691833801e-05, + "loss": 1.2896, + "step": 860 + }, + { + "epoch": 0.058337285723965036, + "grad_norm": 14.13198184967041, + "learning_rate": 9.98494078992402e-05, + "loss": 1.1566, + "step": 861 + }, + { + "epoch": 0.05840504099193712, + "grad_norm": 14.455452919006348, + "learning_rate": 9.984803888014237e-05, + "loss": 1.1949, + "step": 862 + }, + { + "epoch": 0.058472796259909206, + "grad_norm": 15.020733833312988, + "learning_rate": 9.984666986104456e-05, + "loss": 1.5202, + "step": 863 + }, + { + "epoch": 0.05854055152788129, + "grad_norm": 21.089344024658203, + "learning_rate": 9.984530084194675e-05, + "loss": 1.5083, + "step": 864 + }, + { + "epoch": 0.058608306795853375, + "grad_norm": 14.414257049560547, + "learning_rate": 9.984393182284893e-05, + "loss": 1.3977, + "step": 865 + }, + { + "epoch": 0.05867606206382546, + "grad_norm": 15.667798042297363, + "learning_rate": 9.984256280375111e-05, + "loss": 1.4044, + "step": 866 + }, + { + "epoch": 0.058743817331797545, + "grad_norm": 12.503005981445312, + "learning_rate": 9.98411937846533e-05, + "loss": 1.3369, + "step": 867 + }, + { + "epoch": 0.05881157259976963, + "grad_norm": 13.27022933959961, + "learning_rate": 9.983982476555548e-05, + "loss": 1.4134, + "step": 868 + }, + { + "epoch": 0.058879327867741714, + "grad_norm": 16.2034969329834, + "learning_rate": 9.983845574645766e-05, + "loss": 1.1196, + "step": 869 + }, + { + "epoch": 0.0589470831357138, + "grad_norm": 9.868896484375, + "learning_rate": 9.983708672735986e-05, + "loss": 1.0248, + "step": 870 + }, + { + "epoch": 0.05901483840368588, + "grad_norm": 15.588685989379883, + "learning_rate": 9.983571770826204e-05, + "loss": 1.5865, + "step": 871 + }, + { + "epoch": 0.05908259367165797, + "grad_norm": 14.688246726989746, + "learning_rate": 9.983434868916422e-05, + "loss": 1.2959, + "step": 872 + }, + { + "epoch": 0.05915034893963006, + "grad_norm": 14.252961158752441, + "learning_rate": 9.983297967006641e-05, + "loss": 1.3536, + "step": 873 + }, + { + "epoch": 0.059218104207602144, + "grad_norm": 11.79800033569336, + "learning_rate": 9.98316106509686e-05, + "loss": 1.331, + "step": 874 + }, + { + "epoch": 0.05928585947557423, + "grad_norm": 11.900074005126953, + "learning_rate": 9.983024163187077e-05, + "loss": 1.2241, + "step": 875 + }, + { + "epoch": 0.059353614743546314, + "grad_norm": 10.696773529052734, + "learning_rate": 9.982887261277295e-05, + "loss": 1.4041, + "step": 876 + }, + { + "epoch": 0.0594213700115184, + "grad_norm": 13.532305717468262, + "learning_rate": 9.982750359367513e-05, + "loss": 1.3621, + "step": 877 + }, + { + "epoch": 0.05948912527949048, + "grad_norm": 14.107857704162598, + "learning_rate": 9.982613457457733e-05, + "loss": 1.6327, + "step": 878 + }, + { + "epoch": 0.05955688054746257, + "grad_norm": 11.584097862243652, + "learning_rate": 9.982476555547951e-05, + "loss": 0.9606, + "step": 879 + }, + { + "epoch": 0.05962463581543465, + "grad_norm": 14.240161895751953, + "learning_rate": 9.982339653638169e-05, + "loss": 1.263, + "step": 880 + }, + { + "epoch": 0.05969239108340674, + "grad_norm": 14.461871147155762, + "learning_rate": 9.982202751728387e-05, + "loss": 1.4201, + "step": 881 + }, + { + "epoch": 0.05976014635137882, + "grad_norm": 14.072705268859863, + "learning_rate": 9.982065849818606e-05, + "loss": 1.5977, + "step": 882 + }, + { + "epoch": 0.05982790161935091, + "grad_norm": 14.928994178771973, + "learning_rate": 9.981928947908824e-05, + "loss": 1.3346, + "step": 883 + }, + { + "epoch": 0.05989565688732299, + "grad_norm": 14.898951530456543, + "learning_rate": 9.981792045999042e-05, + "loss": 1.312, + "step": 884 + }, + { + "epoch": 0.059963412155295076, + "grad_norm": 13.089646339416504, + "learning_rate": 9.98165514408926e-05, + "loss": 1.0833, + "step": 885 + }, + { + "epoch": 0.06003116742326716, + "grad_norm": 15.768043518066406, + "learning_rate": 9.981518242179478e-05, + "loss": 1.1246, + "step": 886 + }, + { + "epoch": 0.060098922691239245, + "grad_norm": 11.8709135055542, + "learning_rate": 9.981381340269698e-05, + "loss": 1.1822, + "step": 887 + }, + { + "epoch": 0.06016667795921133, + "grad_norm": 15.698454856872559, + "learning_rate": 9.981244438359916e-05, + "loss": 1.3733, + "step": 888 + }, + { + "epoch": 0.060234433227183415, + "grad_norm": 14.827208518981934, + "learning_rate": 9.981107536450134e-05, + "loss": 1.3121, + "step": 889 + }, + { + "epoch": 0.0603021884951555, + "grad_norm": 12.522045135498047, + "learning_rate": 9.980970634540352e-05, + "loss": 1.5584, + "step": 890 + }, + { + "epoch": 0.060369943763127584, + "grad_norm": 14.417738914489746, + "learning_rate": 9.980833732630571e-05, + "loss": 1.2389, + "step": 891 + }, + { + "epoch": 0.06043769903109967, + "grad_norm": 14.761930465698242, + "learning_rate": 9.98069683072079e-05, + "loss": 1.5007, + "step": 892 + }, + { + "epoch": 0.060505454299071754, + "grad_norm": 15.882668495178223, + "learning_rate": 9.980559928811007e-05, + "loss": 1.2651, + "step": 893 + }, + { + "epoch": 0.06057320956704384, + "grad_norm": 13.605412483215332, + "learning_rate": 9.980423026901225e-05, + "loss": 1.5186, + "step": 894 + }, + { + "epoch": 0.06064096483501592, + "grad_norm": 10.654335021972656, + "learning_rate": 9.980286124991443e-05, + "loss": 1.1177, + "step": 895 + }, + { + "epoch": 0.06070872010298801, + "grad_norm": 12.37457275390625, + "learning_rate": 9.980149223081663e-05, + "loss": 1.2362, + "step": 896 + }, + { + "epoch": 0.06077647537096009, + "grad_norm": 12.591222763061523, + "learning_rate": 9.980012321171881e-05, + "loss": 1.2509, + "step": 897 + }, + { + "epoch": 0.06084423063893218, + "grad_norm": 14.337310791015625, + "learning_rate": 9.979875419262099e-05, + "loss": 1.2987, + "step": 898 + }, + { + "epoch": 0.06091198590690426, + "grad_norm": 15.496018409729004, + "learning_rate": 9.979738517352317e-05, + "loss": 1.5307, + "step": 899 + }, + { + "epoch": 0.06097974117487635, + "grad_norm": 13.730890274047852, + "learning_rate": 9.979601615442535e-05, + "loss": 1.2975, + "step": 900 + }, + { + "epoch": 0.06104749644284843, + "grad_norm": 12.314823150634766, + "learning_rate": 9.979464713532754e-05, + "loss": 1.1916, + "step": 901 + }, + { + "epoch": 0.061115251710820516, + "grad_norm": 13.761808395385742, + "learning_rate": 9.979327811622972e-05, + "loss": 1.3037, + "step": 902 + }, + { + "epoch": 0.0611830069787926, + "grad_norm": 13.308722496032715, + "learning_rate": 9.97919090971319e-05, + "loss": 1.0598, + "step": 903 + }, + { + "epoch": 0.061250762246764685, + "grad_norm": 13.121098518371582, + "learning_rate": 9.979054007803408e-05, + "loss": 1.3378, + "step": 904 + }, + { + "epoch": 0.06131851751473677, + "grad_norm": 16.975666046142578, + "learning_rate": 9.978917105893628e-05, + "loss": 1.2826, + "step": 905 + }, + { + "epoch": 0.061386272782708855, + "grad_norm": 14.529984474182129, + "learning_rate": 9.978780203983846e-05, + "loss": 1.3467, + "step": 906 + }, + { + "epoch": 0.06145402805068094, + "grad_norm": 11.081110000610352, + "learning_rate": 9.978643302074064e-05, + "loss": 1.1906, + "step": 907 + }, + { + "epoch": 0.061521783318653024, + "grad_norm": 12.871200561523438, + "learning_rate": 9.978506400164282e-05, + "loss": 1.1057, + "step": 908 + }, + { + "epoch": 0.06158953858662511, + "grad_norm": 13.982168197631836, + "learning_rate": 9.9783694982545e-05, + "loss": 1.3824, + "step": 909 + }, + { + "epoch": 0.061657293854597194, + "grad_norm": 13.076074600219727, + "learning_rate": 9.97823259634472e-05, + "loss": 1.2985, + "step": 910 + }, + { + "epoch": 0.06172504912256928, + "grad_norm": 11.015650749206543, + "learning_rate": 9.978095694434937e-05, + "loss": 1.2406, + "step": 911 + }, + { + "epoch": 0.06179280439054136, + "grad_norm": 13.6082763671875, + "learning_rate": 9.977958792525155e-05, + "loss": 1.3873, + "step": 912 + }, + { + "epoch": 0.06186055965851345, + "grad_norm": 15.930809020996094, + "learning_rate": 9.977821890615375e-05, + "loss": 1.3542, + "step": 913 + }, + { + "epoch": 0.06192831492648553, + "grad_norm": 10.710271835327148, + "learning_rate": 9.977684988705593e-05, + "loss": 1.1571, + "step": 914 + }, + { + "epoch": 0.06199607019445762, + "grad_norm": 11.110217094421387, + "learning_rate": 9.977548086795811e-05, + "loss": 1.3255, + "step": 915 + }, + { + "epoch": 0.0620638254624297, + "grad_norm": 11.451903343200684, + "learning_rate": 9.97741118488603e-05, + "loss": 1.3681, + "step": 916 + }, + { + "epoch": 0.062131580730401786, + "grad_norm": 10.884252548217773, + "learning_rate": 9.977274282976248e-05, + "loss": 1.1988, + "step": 917 + }, + { + "epoch": 0.06219933599837387, + "grad_norm": 11.031237602233887, + "learning_rate": 9.977137381066466e-05, + "loss": 1.26, + "step": 918 + }, + { + "epoch": 0.062267091266345956, + "grad_norm": 11.585648536682129, + "learning_rate": 9.977000479156686e-05, + "loss": 1.3099, + "step": 919 + }, + { + "epoch": 0.06233484653431804, + "grad_norm": 10.867992401123047, + "learning_rate": 9.976863577246904e-05, + "loss": 1.2624, + "step": 920 + }, + { + "epoch": 0.062402601802290125, + "grad_norm": 14.552916526794434, + "learning_rate": 9.976726675337122e-05, + "loss": 1.2244, + "step": 921 + }, + { + "epoch": 0.06247035707026221, + "grad_norm": 12.101760864257812, + "learning_rate": 9.97658977342734e-05, + "loss": 1.1925, + "step": 922 + }, + { + "epoch": 0.0625381123382343, + "grad_norm": 14.113842010498047, + "learning_rate": 9.976452871517559e-05, + "loss": 1.3395, + "step": 923 + }, + { + "epoch": 0.06260586760620639, + "grad_norm": 17.214614868164062, + "learning_rate": 9.976315969607777e-05, + "loss": 1.3943, + "step": 924 + }, + { + "epoch": 0.06267362287417846, + "grad_norm": 13.43308162689209, + "learning_rate": 9.976179067697995e-05, + "loss": 1.1596, + "step": 925 + }, + { + "epoch": 0.06274137814215056, + "grad_norm": 13.806952476501465, + "learning_rate": 9.976042165788213e-05, + "loss": 1.0354, + "step": 926 + }, + { + "epoch": 0.06280913341012263, + "grad_norm": 15.638693809509277, + "learning_rate": 9.975905263878431e-05, + "loss": 1.3807, + "step": 927 + }, + { + "epoch": 0.06287688867809473, + "grad_norm": 14.337742805480957, + "learning_rate": 9.975768361968651e-05, + "loss": 1.537, + "step": 928 + }, + { + "epoch": 0.0629446439460668, + "grad_norm": 14.540297508239746, + "learning_rate": 9.975631460058869e-05, + "loss": 1.2522, + "step": 929 + }, + { + "epoch": 0.0630123992140389, + "grad_norm": 15.991955757141113, + "learning_rate": 9.975494558149087e-05, + "loss": 1.3855, + "step": 930 + }, + { + "epoch": 0.06308015448201097, + "grad_norm": 13.957479476928711, + "learning_rate": 9.975357656239305e-05, + "loss": 1.43, + "step": 931 + }, + { + "epoch": 0.06314790974998306, + "grad_norm": 16.805377960205078, + "learning_rate": 9.975220754329523e-05, + "loss": 1.383, + "step": 932 + }, + { + "epoch": 0.06321566501795514, + "grad_norm": 12.41854476928711, + "learning_rate": 9.975083852419742e-05, + "loss": 1.068, + "step": 933 + }, + { + "epoch": 0.06328342028592723, + "grad_norm": 15.929006576538086, + "learning_rate": 9.97494695050996e-05, + "loss": 1.212, + "step": 934 + }, + { + "epoch": 0.06335117555389931, + "grad_norm": 13.205544471740723, + "learning_rate": 9.974810048600178e-05, + "loss": 1.3682, + "step": 935 + }, + { + "epoch": 0.0634189308218714, + "grad_norm": 12.105626106262207, + "learning_rate": 9.974673146690396e-05, + "loss": 1.3821, + "step": 936 + }, + { + "epoch": 0.06348668608984348, + "grad_norm": 13.776711463928223, + "learning_rate": 9.974536244780616e-05, + "loss": 1.1073, + "step": 937 + }, + { + "epoch": 0.06355444135781557, + "grad_norm": 12.227380752563477, + "learning_rate": 9.974399342870834e-05, + "loss": 1.2026, + "step": 938 + }, + { + "epoch": 0.06362219662578765, + "grad_norm": 12.723440170288086, + "learning_rate": 9.974262440961052e-05, + "loss": 1.1325, + "step": 939 + }, + { + "epoch": 0.06368995189375974, + "grad_norm": 13.943262100219727, + "learning_rate": 9.97412553905127e-05, + "loss": 1.1878, + "step": 940 + }, + { + "epoch": 0.06375770716173182, + "grad_norm": 12.644627571105957, + "learning_rate": 9.973988637141488e-05, + "loss": 1.3097, + "step": 941 + }, + { + "epoch": 0.06382546242970391, + "grad_norm": 12.108241081237793, + "learning_rate": 9.973851735231707e-05, + "loss": 1.1686, + "step": 942 + }, + { + "epoch": 0.06389321769767599, + "grad_norm": 14.375092506408691, + "learning_rate": 9.973714833321925e-05, + "loss": 1.2721, + "step": 943 + }, + { + "epoch": 0.06396097296564808, + "grad_norm": 13.439800262451172, + "learning_rate": 9.973577931412143e-05, + "loss": 1.3898, + "step": 944 + }, + { + "epoch": 0.06402872823362016, + "grad_norm": 13.717879295349121, + "learning_rate": 9.973441029502361e-05, + "loss": 1.3823, + "step": 945 + }, + { + "epoch": 0.06409648350159225, + "grad_norm": 12.745361328125, + "learning_rate": 9.973304127592581e-05, + "loss": 1.2396, + "step": 946 + }, + { + "epoch": 0.06416423876956433, + "grad_norm": 11.784343719482422, + "learning_rate": 9.973167225682799e-05, + "loss": 1.214, + "step": 947 + }, + { + "epoch": 0.06423199403753642, + "grad_norm": 14.205467224121094, + "learning_rate": 9.973030323773017e-05, + "loss": 1.2803, + "step": 948 + }, + { + "epoch": 0.0642997493055085, + "grad_norm": 13.257532119750977, + "learning_rate": 9.972893421863235e-05, + "loss": 1.378, + "step": 949 + }, + { + "epoch": 0.06436750457348059, + "grad_norm": 15.153338432312012, + "learning_rate": 9.972756519953453e-05, + "loss": 1.2854, + "step": 950 + }, + { + "epoch": 0.06443525984145267, + "grad_norm": 16.765771865844727, + "learning_rate": 9.972619618043672e-05, + "loss": 1.3016, + "step": 951 + }, + { + "epoch": 0.06450301510942476, + "grad_norm": 14.636106491088867, + "learning_rate": 9.97248271613389e-05, + "loss": 1.3803, + "step": 952 + }, + { + "epoch": 0.06457077037739685, + "grad_norm": 13.87410831451416, + "learning_rate": 9.972345814224108e-05, + "loss": 1.4126, + "step": 953 + }, + { + "epoch": 0.06463852564536893, + "grad_norm": 14.328899383544922, + "learning_rate": 9.972208912314326e-05, + "loss": 1.061, + "step": 954 + }, + { + "epoch": 0.06470628091334102, + "grad_norm": 12.485203742980957, + "learning_rate": 9.972072010404544e-05, + "loss": 1.0985, + "step": 955 + }, + { + "epoch": 0.0647740361813131, + "grad_norm": 13.77907943725586, + "learning_rate": 9.971935108494764e-05, + "loss": 1.3534, + "step": 956 + }, + { + "epoch": 0.06484179144928519, + "grad_norm": 10.579590797424316, + "learning_rate": 9.971798206584982e-05, + "loss": 1.166, + "step": 957 + }, + { + "epoch": 0.06490954671725727, + "grad_norm": 14.690185546875, + "learning_rate": 9.9716613046752e-05, + "loss": 1.3666, + "step": 958 + }, + { + "epoch": 0.06497730198522936, + "grad_norm": 12.904786109924316, + "learning_rate": 9.97152440276542e-05, + "loss": 1.14, + "step": 959 + }, + { + "epoch": 0.06504505725320144, + "grad_norm": 12.126219749450684, + "learning_rate": 9.971387500855637e-05, + "loss": 1.4157, + "step": 960 + }, + { + "epoch": 0.06511281252117353, + "grad_norm": 13.747931480407715, + "learning_rate": 9.971250598945855e-05, + "loss": 1.4557, + "step": 961 + }, + { + "epoch": 0.0651805677891456, + "grad_norm": 13.232327461242676, + "learning_rate": 9.971113697036075e-05, + "loss": 1.3471, + "step": 962 + }, + { + "epoch": 0.0652483230571177, + "grad_norm": 14.886791229248047, + "learning_rate": 9.970976795126293e-05, + "loss": 1.1866, + "step": 963 + }, + { + "epoch": 0.06531607832508977, + "grad_norm": 11.747659683227539, + "learning_rate": 9.970839893216511e-05, + "loss": 1.0443, + "step": 964 + }, + { + "epoch": 0.06538383359306187, + "grad_norm": 11.181273460388184, + "learning_rate": 9.97070299130673e-05, + "loss": 1.1391, + "step": 965 + }, + { + "epoch": 0.06545158886103394, + "grad_norm": 11.9672269821167, + "learning_rate": 9.970566089396948e-05, + "loss": 1.2847, + "step": 966 + }, + { + "epoch": 0.06551934412900604, + "grad_norm": 15.825364112854004, + "learning_rate": 9.970429187487166e-05, + "loss": 1.3301, + "step": 967 + }, + { + "epoch": 0.06558709939697811, + "grad_norm": 12.26963996887207, + "learning_rate": 9.970292285577384e-05, + "loss": 1.2524, + "step": 968 + }, + { + "epoch": 0.0656548546649502, + "grad_norm": 11.440977096557617, + "learning_rate": 9.970155383667604e-05, + "loss": 1.229, + "step": 969 + }, + { + "epoch": 0.06572260993292228, + "grad_norm": 10.704546928405762, + "learning_rate": 9.970018481757822e-05, + "loss": 0.9936, + "step": 970 + }, + { + "epoch": 0.06579036520089437, + "grad_norm": 13.20880126953125, + "learning_rate": 9.96988157984804e-05, + "loss": 1.2353, + "step": 971 + }, + { + "epoch": 0.06585812046886645, + "grad_norm": 13.101622581481934, + "learning_rate": 9.969744677938258e-05, + "loss": 1.2559, + "step": 972 + }, + { + "epoch": 0.06592587573683854, + "grad_norm": 11.725826263427734, + "learning_rate": 9.969607776028476e-05, + "loss": 1.2051, + "step": 973 + }, + { + "epoch": 0.06599363100481062, + "grad_norm": 11.890633583068848, + "learning_rate": 9.969470874118695e-05, + "loss": 1.1607, + "step": 974 + }, + { + "epoch": 0.06606138627278271, + "grad_norm": 11.066970825195312, + "learning_rate": 9.969333972208913e-05, + "loss": 1.1031, + "step": 975 + }, + { + "epoch": 0.06612914154075479, + "grad_norm": 12.26187515258789, + "learning_rate": 9.969197070299131e-05, + "loss": 1.309, + "step": 976 + }, + { + "epoch": 0.06619689680872688, + "grad_norm": 13.490363121032715, + "learning_rate": 9.96906016838935e-05, + "loss": 1.3062, + "step": 977 + }, + { + "epoch": 0.06626465207669896, + "grad_norm": 12.306289672851562, + "learning_rate": 9.968923266479567e-05, + "loss": 0.9784, + "step": 978 + }, + { + "epoch": 0.06633240734467105, + "grad_norm": 10.699983596801758, + "learning_rate": 9.968786364569787e-05, + "loss": 1.0547, + "step": 979 + }, + { + "epoch": 0.06640016261264313, + "grad_norm": 12.298179626464844, + "learning_rate": 9.968649462660005e-05, + "loss": 1.2496, + "step": 980 + }, + { + "epoch": 0.06646791788061522, + "grad_norm": 15.239167213439941, + "learning_rate": 9.968512560750223e-05, + "loss": 1.2545, + "step": 981 + }, + { + "epoch": 0.0665356731485873, + "grad_norm": 11.61802864074707, + "learning_rate": 9.968375658840441e-05, + "loss": 1.1668, + "step": 982 + }, + { + "epoch": 0.06660342841655939, + "grad_norm": 12.804032325744629, + "learning_rate": 9.96823875693066e-05, + "loss": 1.3203, + "step": 983 + }, + { + "epoch": 0.06667118368453147, + "grad_norm": 14.511723518371582, + "learning_rate": 9.968101855020878e-05, + "loss": 1.7519, + "step": 984 + }, + { + "epoch": 0.06673893895250356, + "grad_norm": 12.442008018493652, + "learning_rate": 9.967964953111096e-05, + "loss": 1.3758, + "step": 985 + }, + { + "epoch": 0.06680669422047564, + "grad_norm": 14.486754417419434, + "learning_rate": 9.967828051201314e-05, + "loss": 1.1924, + "step": 986 + }, + { + "epoch": 0.06687444948844773, + "grad_norm": 13.529693603515625, + "learning_rate": 9.967691149291532e-05, + "loss": 1.2698, + "step": 987 + }, + { + "epoch": 0.06694220475641981, + "grad_norm": 12.980225563049316, + "learning_rate": 9.967554247381752e-05, + "loss": 1.1896, + "step": 988 + }, + { + "epoch": 0.0670099600243919, + "grad_norm": 15.495257377624512, + "learning_rate": 9.96741734547197e-05, + "loss": 1.286, + "step": 989 + }, + { + "epoch": 0.06707771529236398, + "grad_norm": 12.67573070526123, + "learning_rate": 9.967280443562188e-05, + "loss": 1.1751, + "step": 990 + }, + { + "epoch": 0.06714547056033607, + "grad_norm": 11.141845703125, + "learning_rate": 9.967143541652406e-05, + "loss": 1.5109, + "step": 991 + }, + { + "epoch": 0.06721322582830815, + "grad_norm": 11.975769996643066, + "learning_rate": 9.967006639742625e-05, + "loss": 1.1579, + "step": 992 + }, + { + "epoch": 0.06728098109628024, + "grad_norm": 13.872209548950195, + "learning_rate": 9.966869737832843e-05, + "loss": 1.1852, + "step": 993 + }, + { + "epoch": 0.06734873636425232, + "grad_norm": 11.52573299407959, + "learning_rate": 9.966732835923061e-05, + "loss": 1.2319, + "step": 994 + }, + { + "epoch": 0.0674164916322244, + "grad_norm": 12.90494155883789, + "learning_rate": 9.96659593401328e-05, + "loss": 1.2415, + "step": 995 + }, + { + "epoch": 0.06748424690019648, + "grad_norm": 14.210317611694336, + "learning_rate": 9.966459032103497e-05, + "loss": 1.2901, + "step": 996 + }, + { + "epoch": 0.06755200216816858, + "grad_norm": 12.600135803222656, + "learning_rate": 9.966322130193717e-05, + "loss": 1.4516, + "step": 997 + }, + { + "epoch": 0.06761975743614065, + "grad_norm": 14.462118148803711, + "learning_rate": 9.966185228283935e-05, + "loss": 1.2205, + "step": 998 + }, + { + "epoch": 0.06768751270411275, + "grad_norm": 12.870843887329102, + "learning_rate": 9.966048326374153e-05, + "loss": 1.179, + "step": 999 + }, + { + "epoch": 0.06775526797208482, + "grad_norm": 16.4424991607666, + "learning_rate": 9.965911424464371e-05, + "loss": 1.5714, + "step": 1000 + }, + { + "epoch": 0.06782302324005691, + "grad_norm": 12.902230262756348, + "learning_rate": 9.96577452255459e-05, + "loss": 1.2803, + "step": 1001 + }, + { + "epoch": 0.06789077850802899, + "grad_norm": 11.469466209411621, + "learning_rate": 9.965637620644808e-05, + "loss": 1.0771, + "step": 1002 + }, + { + "epoch": 0.06795853377600108, + "grad_norm": 13.96650505065918, + "learning_rate": 9.965500718735026e-05, + "loss": 1.2427, + "step": 1003 + }, + { + "epoch": 0.06802628904397316, + "grad_norm": 11.55516242980957, + "learning_rate": 9.965363816825244e-05, + "loss": 1.0396, + "step": 1004 + }, + { + "epoch": 0.06809404431194525, + "grad_norm": 13.34827709197998, + "learning_rate": 9.965226914915462e-05, + "loss": 1.181, + "step": 1005 + }, + { + "epoch": 0.06816179957991735, + "grad_norm": 11.243910789489746, + "learning_rate": 9.965090013005682e-05, + "loss": 1.2875, + "step": 1006 + }, + { + "epoch": 0.06822955484788942, + "grad_norm": 14.152894020080566, + "learning_rate": 9.9649531110959e-05, + "loss": 1.3125, + "step": 1007 + }, + { + "epoch": 0.06829731011586151, + "grad_norm": 13.010010719299316, + "learning_rate": 9.964816209186118e-05, + "loss": 1.3213, + "step": 1008 + }, + { + "epoch": 0.06836506538383359, + "grad_norm": 15.990034103393555, + "learning_rate": 9.964679307276337e-05, + "loss": 1.1878, + "step": 1009 + }, + { + "epoch": 0.06843282065180568, + "grad_norm": 12.943589210510254, + "learning_rate": 9.964542405366555e-05, + "loss": 1.2443, + "step": 1010 + }, + { + "epoch": 0.06850057591977776, + "grad_norm": 12.108896255493164, + "learning_rate": 9.964405503456775e-05, + "loss": 1.1719, + "step": 1011 + }, + { + "epoch": 0.06856833118774985, + "grad_norm": 12.097951889038086, + "learning_rate": 9.964268601546993e-05, + "loss": 1.0653, + "step": 1012 + }, + { + "epoch": 0.06863608645572193, + "grad_norm": 14.222228050231934, + "learning_rate": 9.964131699637211e-05, + "loss": 1.3379, + "step": 1013 + }, + { + "epoch": 0.06870384172369402, + "grad_norm": 12.636894226074219, + "learning_rate": 9.963994797727429e-05, + "loss": 1.5559, + "step": 1014 + }, + { + "epoch": 0.0687715969916661, + "grad_norm": 15.458481788635254, + "learning_rate": 9.963857895817648e-05, + "loss": 1.2542, + "step": 1015 + }, + { + "epoch": 0.06883935225963819, + "grad_norm": 11.246847152709961, + "learning_rate": 9.963720993907866e-05, + "loss": 1.1944, + "step": 1016 + }, + { + "epoch": 0.06890710752761027, + "grad_norm": 11.699065208435059, + "learning_rate": 9.963584091998084e-05, + "loss": 1.0739, + "step": 1017 + }, + { + "epoch": 0.06897486279558236, + "grad_norm": 12.259678840637207, + "learning_rate": 9.963447190088302e-05, + "loss": 1.1365, + "step": 1018 + }, + { + "epoch": 0.06904261806355444, + "grad_norm": 13.594696998596191, + "learning_rate": 9.96331028817852e-05, + "loss": 1.1418, + "step": 1019 + }, + { + "epoch": 0.06911037333152653, + "grad_norm": 12.90888786315918, + "learning_rate": 9.96317338626874e-05, + "loss": 1.1987, + "step": 1020 + }, + { + "epoch": 0.06917812859949861, + "grad_norm": 13.04245662689209, + "learning_rate": 9.963036484358958e-05, + "loss": 1.4288, + "step": 1021 + }, + { + "epoch": 0.0692458838674707, + "grad_norm": 12.706077575683594, + "learning_rate": 9.962899582449176e-05, + "loss": 1.222, + "step": 1022 + }, + { + "epoch": 0.06931363913544278, + "grad_norm": 14.205679893493652, + "learning_rate": 9.962762680539394e-05, + "loss": 1.3305, + "step": 1023 + }, + { + "epoch": 0.06938139440341487, + "grad_norm": 17.09891128540039, + "learning_rate": 9.962625778629613e-05, + "loss": 1.47, + "step": 1024 + }, + { + "epoch": 0.06944914967138695, + "grad_norm": 14.603500366210938, + "learning_rate": 9.962488876719831e-05, + "loss": 1.3887, + "step": 1025 + }, + { + "epoch": 0.06951690493935904, + "grad_norm": 10.820066452026367, + "learning_rate": 9.962351974810049e-05, + "loss": 1.1543, + "step": 1026 + }, + { + "epoch": 0.06958466020733112, + "grad_norm": 10.99889850616455, + "learning_rate": 9.962215072900267e-05, + "loss": 1.2908, + "step": 1027 + }, + { + "epoch": 0.06965241547530321, + "grad_norm": 13.470711708068848, + "learning_rate": 9.962078170990485e-05, + "loss": 1.1422, + "step": 1028 + }, + { + "epoch": 0.06972017074327529, + "grad_norm": 11.086441040039062, + "learning_rate": 9.961941269080705e-05, + "loss": 0.9835, + "step": 1029 + }, + { + "epoch": 0.06978792601124738, + "grad_norm": 16.768535614013672, + "learning_rate": 9.961804367170923e-05, + "loss": 1.404, + "step": 1030 + }, + { + "epoch": 0.06985568127921946, + "grad_norm": 15.851200103759766, + "learning_rate": 9.961667465261141e-05, + "loss": 1.4874, + "step": 1031 + }, + { + "epoch": 0.06992343654719155, + "grad_norm": 11.995482444763184, + "learning_rate": 9.961530563351359e-05, + "loss": 1.1497, + "step": 1032 + }, + { + "epoch": 0.06999119181516363, + "grad_norm": 13.591619491577148, + "learning_rate": 9.961393661441577e-05, + "loss": 1.4773, + "step": 1033 + }, + { + "epoch": 0.07005894708313572, + "grad_norm": 16.878938674926758, + "learning_rate": 9.961256759531796e-05, + "loss": 1.2522, + "step": 1034 + }, + { + "epoch": 0.0701267023511078, + "grad_norm": 11.901616096496582, + "learning_rate": 9.961119857622014e-05, + "loss": 1.2164, + "step": 1035 + }, + { + "epoch": 0.07019445761907989, + "grad_norm": 14.935117721557617, + "learning_rate": 9.960982955712232e-05, + "loss": 1.1251, + "step": 1036 + }, + { + "epoch": 0.07026221288705196, + "grad_norm": 12.380253791809082, + "learning_rate": 9.96084605380245e-05, + "loss": 1.3965, + "step": 1037 + }, + { + "epoch": 0.07032996815502406, + "grad_norm": 11.645035743713379, + "learning_rate": 9.96070915189267e-05, + "loss": 1.266, + "step": 1038 + }, + { + "epoch": 0.07039772342299613, + "grad_norm": 14.525420188903809, + "learning_rate": 9.960572249982888e-05, + "loss": 1.3991, + "step": 1039 + }, + { + "epoch": 0.07046547869096823, + "grad_norm": 14.775094985961914, + "learning_rate": 9.960435348073106e-05, + "loss": 1.5958, + "step": 1040 + }, + { + "epoch": 0.0705332339589403, + "grad_norm": 10.2192964553833, + "learning_rate": 9.960298446163324e-05, + "loss": 1.1793, + "step": 1041 + }, + { + "epoch": 0.0706009892269124, + "grad_norm": 13.074480056762695, + "learning_rate": 9.960161544253542e-05, + "loss": 1.4243, + "step": 1042 + }, + { + "epoch": 0.07066874449488447, + "grad_norm": 12.679484367370605, + "learning_rate": 9.960024642343761e-05, + "loss": 1.3398, + "step": 1043 + }, + { + "epoch": 0.07073649976285656, + "grad_norm": 9.061332702636719, + "learning_rate": 9.95988774043398e-05, + "loss": 1.0036, + "step": 1044 + }, + { + "epoch": 0.07080425503082864, + "grad_norm": 13.423661231994629, + "learning_rate": 9.959750838524197e-05, + "loss": 1.2767, + "step": 1045 + }, + { + "epoch": 0.07087201029880073, + "grad_norm": 13.955148696899414, + "learning_rate": 9.959613936614415e-05, + "loss": 1.311, + "step": 1046 + }, + { + "epoch": 0.07093976556677281, + "grad_norm": 12.746015548706055, + "learning_rate": 9.959477034704635e-05, + "loss": 1.42, + "step": 1047 + }, + { + "epoch": 0.0710075208347449, + "grad_norm": 11.409982681274414, + "learning_rate": 9.959340132794853e-05, + "loss": 1.3051, + "step": 1048 + }, + { + "epoch": 0.07107527610271698, + "grad_norm": 11.801681518554688, + "learning_rate": 9.959203230885071e-05, + "loss": 1.2315, + "step": 1049 + }, + { + "epoch": 0.07114303137068907, + "grad_norm": 13.041158676147461, + "learning_rate": 9.959066328975289e-05, + "loss": 1.3487, + "step": 1050 + }, + { + "epoch": 0.07121078663866115, + "grad_norm": 13.474900245666504, + "learning_rate": 9.958929427065507e-05, + "loss": 1.4542, + "step": 1051 + }, + { + "epoch": 0.07127854190663324, + "grad_norm": 12.335237503051758, + "learning_rate": 9.958792525155726e-05, + "loss": 1.1841, + "step": 1052 + }, + { + "epoch": 0.07134629717460532, + "grad_norm": 14.909475326538086, + "learning_rate": 9.958655623245944e-05, + "loss": 1.272, + "step": 1053 + }, + { + "epoch": 0.07141405244257741, + "grad_norm": 13.449742317199707, + "learning_rate": 9.958518721336162e-05, + "loss": 1.4881, + "step": 1054 + }, + { + "epoch": 0.07148180771054949, + "grad_norm": 12.2557954788208, + "learning_rate": 9.958381819426382e-05, + "loss": 1.2059, + "step": 1055 + }, + { + "epoch": 0.07154956297852158, + "grad_norm": 13.71298885345459, + "learning_rate": 9.9582449175166e-05, + "loss": 1.2627, + "step": 1056 + }, + { + "epoch": 0.07161731824649366, + "grad_norm": 15.4293212890625, + "learning_rate": 9.958108015606818e-05, + "loss": 1.4663, + "step": 1057 + }, + { + "epoch": 0.07168507351446575, + "grad_norm": 13.665759086608887, + "learning_rate": 9.957971113697037e-05, + "loss": 1.4634, + "step": 1058 + }, + { + "epoch": 0.07175282878243784, + "grad_norm": 13.064310073852539, + "learning_rate": 9.957834211787255e-05, + "loss": 1.2686, + "step": 1059 + }, + { + "epoch": 0.07182058405040992, + "grad_norm": 15.21746826171875, + "learning_rate": 9.957697309877473e-05, + "loss": 1.4861, + "step": 1060 + }, + { + "epoch": 0.07188833931838201, + "grad_norm": 12.499883651733398, + "learning_rate": 9.957560407967693e-05, + "loss": 1.5275, + "step": 1061 + }, + { + "epoch": 0.07195609458635409, + "grad_norm": 9.355907440185547, + "learning_rate": 9.957423506057911e-05, + "loss": 1.1509, + "step": 1062 + }, + { + "epoch": 0.07202384985432618, + "grad_norm": 10.428252220153809, + "learning_rate": 9.957286604148129e-05, + "loss": 1.0874, + "step": 1063 + }, + { + "epoch": 0.07209160512229826, + "grad_norm": 11.351346015930176, + "learning_rate": 9.957149702238347e-05, + "loss": 1.0887, + "step": 1064 + }, + { + "epoch": 0.07215936039027035, + "grad_norm": 15.271830558776855, + "learning_rate": 9.957012800328565e-05, + "loss": 1.2466, + "step": 1065 + }, + { + "epoch": 0.07222711565824243, + "grad_norm": 11.172418594360352, + "learning_rate": 9.956875898418784e-05, + "loss": 1.0958, + "step": 1066 + }, + { + "epoch": 0.07229487092621452, + "grad_norm": 14.221702575683594, + "learning_rate": 9.956738996509002e-05, + "loss": 1.2288, + "step": 1067 + }, + { + "epoch": 0.0723626261941866, + "grad_norm": 12.167356491088867, + "learning_rate": 9.95660209459922e-05, + "loss": 1.1159, + "step": 1068 + }, + { + "epoch": 0.07243038146215869, + "grad_norm": 12.607329368591309, + "learning_rate": 9.956465192689438e-05, + "loss": 1.2475, + "step": 1069 + }, + { + "epoch": 0.07249813673013077, + "grad_norm": 10.59451675415039, + "learning_rate": 9.956328290779658e-05, + "loss": 1.1493, + "step": 1070 + }, + { + "epoch": 0.07256589199810286, + "grad_norm": 11.190742492675781, + "learning_rate": 9.956191388869876e-05, + "loss": 1.4549, + "step": 1071 + }, + { + "epoch": 0.07263364726607494, + "grad_norm": 11.225564956665039, + "learning_rate": 9.956054486960094e-05, + "loss": 1.0741, + "step": 1072 + }, + { + "epoch": 0.07270140253404703, + "grad_norm": 13.648139953613281, + "learning_rate": 9.955917585050312e-05, + "loss": 1.2942, + "step": 1073 + }, + { + "epoch": 0.0727691578020191, + "grad_norm": 14.606941223144531, + "learning_rate": 9.95578068314053e-05, + "loss": 1.253, + "step": 1074 + }, + { + "epoch": 0.0728369130699912, + "grad_norm": 11.043729782104492, + "learning_rate": 9.955643781230749e-05, + "loss": 1.2279, + "step": 1075 + }, + { + "epoch": 0.07290466833796327, + "grad_norm": 12.463634490966797, + "learning_rate": 9.955506879320967e-05, + "loss": 1.4399, + "step": 1076 + }, + { + "epoch": 0.07297242360593537, + "grad_norm": 13.74101734161377, + "learning_rate": 9.955369977411185e-05, + "loss": 1.4323, + "step": 1077 + }, + { + "epoch": 0.07304017887390744, + "grad_norm": 10.1694974899292, + "learning_rate": 9.955233075501403e-05, + "loss": 1.0113, + "step": 1078 + }, + { + "epoch": 0.07310793414187954, + "grad_norm": 14.407991409301758, + "learning_rate": 9.955096173591623e-05, + "loss": 1.7102, + "step": 1079 + }, + { + "epoch": 0.07317568940985161, + "grad_norm": 13.84760570526123, + "learning_rate": 9.954959271681841e-05, + "loss": 1.4491, + "step": 1080 + }, + { + "epoch": 0.0732434446778237, + "grad_norm": 12.220841407775879, + "learning_rate": 9.954822369772059e-05, + "loss": 1.1826, + "step": 1081 + }, + { + "epoch": 0.07331119994579578, + "grad_norm": 11.380377769470215, + "learning_rate": 9.954685467862277e-05, + "loss": 1.3537, + "step": 1082 + }, + { + "epoch": 0.07337895521376787, + "grad_norm": 12.916484832763672, + "learning_rate": 9.954548565952495e-05, + "loss": 1.3406, + "step": 1083 + }, + { + "epoch": 0.07344671048173995, + "grad_norm": 14.107590675354004, + "learning_rate": 9.954411664042714e-05, + "loss": 1.0531, + "step": 1084 + }, + { + "epoch": 0.07351446574971204, + "grad_norm": 13.498798370361328, + "learning_rate": 9.954274762132932e-05, + "loss": 1.1546, + "step": 1085 + }, + { + "epoch": 0.07358222101768412, + "grad_norm": 13.155747413635254, + "learning_rate": 9.95413786022315e-05, + "loss": 1.212, + "step": 1086 + }, + { + "epoch": 0.07364997628565621, + "grad_norm": 14.135687828063965, + "learning_rate": 9.954000958313368e-05, + "loss": 1.2001, + "step": 1087 + }, + { + "epoch": 0.07371773155362829, + "grad_norm": 13.123790740966797, + "learning_rate": 9.953864056403586e-05, + "loss": 1.2735, + "step": 1088 + }, + { + "epoch": 0.07378548682160038, + "grad_norm": 14.591660499572754, + "learning_rate": 9.953727154493806e-05, + "loss": 1.2977, + "step": 1089 + }, + { + "epoch": 0.07385324208957246, + "grad_norm": 13.30614948272705, + "learning_rate": 9.953590252584024e-05, + "loss": 1.2445, + "step": 1090 + }, + { + "epoch": 0.07392099735754455, + "grad_norm": 14.542524337768555, + "learning_rate": 9.953453350674242e-05, + "loss": 1.2327, + "step": 1091 + }, + { + "epoch": 0.07398875262551663, + "grad_norm": 12.926522254943848, + "learning_rate": 9.95331644876446e-05, + "loss": 1.2139, + "step": 1092 + }, + { + "epoch": 0.07405650789348872, + "grad_norm": 11.622479438781738, + "learning_rate": 9.953179546854679e-05, + "loss": 1.2887, + "step": 1093 + }, + { + "epoch": 0.0741242631614608, + "grad_norm": 14.87485122680664, + "learning_rate": 9.953042644944897e-05, + "loss": 1.3132, + "step": 1094 + }, + { + "epoch": 0.07419201842943289, + "grad_norm": 13.096991539001465, + "learning_rate": 9.952905743035115e-05, + "loss": 1.0783, + "step": 1095 + }, + { + "epoch": 0.07425977369740497, + "grad_norm": 14.676551818847656, + "learning_rate": 9.952768841125333e-05, + "loss": 1.3171, + "step": 1096 + }, + { + "epoch": 0.07432752896537706, + "grad_norm": 10.399755477905273, + "learning_rate": 9.952631939215551e-05, + "loss": 1.0747, + "step": 1097 + }, + { + "epoch": 0.07439528423334914, + "grad_norm": 11.052745819091797, + "learning_rate": 9.952495037305771e-05, + "loss": 1.1653, + "step": 1098 + }, + { + "epoch": 0.07446303950132123, + "grad_norm": 13.014352798461914, + "learning_rate": 9.952358135395989e-05, + "loss": 1.3727, + "step": 1099 + }, + { + "epoch": 0.07453079476929331, + "grad_norm": 13.202262878417969, + "learning_rate": 9.952221233486207e-05, + "loss": 1.5063, + "step": 1100 + }, + { + "epoch": 0.0745985500372654, + "grad_norm": 14.272111892700195, + "learning_rate": 9.952084331576426e-05, + "loss": 1.0183, + "step": 1101 + }, + { + "epoch": 0.07466630530523748, + "grad_norm": 14.23975658416748, + "learning_rate": 9.951947429666644e-05, + "loss": 1.3518, + "step": 1102 + }, + { + "epoch": 0.07473406057320957, + "grad_norm": 12.293742179870605, + "learning_rate": 9.951810527756862e-05, + "loss": 1.4021, + "step": 1103 + }, + { + "epoch": 0.07480181584118165, + "grad_norm": 16.026020050048828, + "learning_rate": 9.951673625847082e-05, + "loss": 1.3412, + "step": 1104 + }, + { + "epoch": 0.07486957110915374, + "grad_norm": 11.120819091796875, + "learning_rate": 9.9515367239373e-05, + "loss": 1.331, + "step": 1105 + }, + { + "epoch": 0.07493732637712582, + "grad_norm": 12.363526344299316, + "learning_rate": 9.951399822027518e-05, + "loss": 1.4284, + "step": 1106 + }, + { + "epoch": 0.0750050816450979, + "grad_norm": 14.377492904663086, + "learning_rate": 9.951262920117737e-05, + "loss": 1.3686, + "step": 1107 + }, + { + "epoch": 0.07507283691306998, + "grad_norm": 11.366288185119629, + "learning_rate": 9.951126018207955e-05, + "loss": 1.2329, + "step": 1108 + }, + { + "epoch": 0.07514059218104208, + "grad_norm": 10.26131820678711, + "learning_rate": 9.950989116298173e-05, + "loss": 1.1091, + "step": 1109 + }, + { + "epoch": 0.07520834744901415, + "grad_norm": 12.64631462097168, + "learning_rate": 9.950852214388391e-05, + "loss": 1.1231, + "step": 1110 + }, + { + "epoch": 0.07527610271698625, + "grad_norm": 13.042781829833984, + "learning_rate": 9.95071531247861e-05, + "loss": 1.2338, + "step": 1111 + }, + { + "epoch": 0.07534385798495834, + "grad_norm": 11.577115058898926, + "learning_rate": 9.950578410568829e-05, + "loss": 1.2143, + "step": 1112 + }, + { + "epoch": 0.07541161325293042, + "grad_norm": 13.640811920166016, + "learning_rate": 9.950441508659047e-05, + "loss": 1.224, + "step": 1113 + }, + { + "epoch": 0.0754793685209025, + "grad_norm": 10.271018981933594, + "learning_rate": 9.950304606749265e-05, + "loss": 1.0459, + "step": 1114 + }, + { + "epoch": 0.07554712378887458, + "grad_norm": 12.053836822509766, + "learning_rate": 9.950167704839483e-05, + "loss": 1.2137, + "step": 1115 + }, + { + "epoch": 0.07561487905684668, + "grad_norm": 10.022509574890137, + "learning_rate": 9.950030802929702e-05, + "loss": 1.092, + "step": 1116 + }, + { + "epoch": 0.07568263432481875, + "grad_norm": 12.28339672088623, + "learning_rate": 9.94989390101992e-05, + "loss": 1.1616, + "step": 1117 + }, + { + "epoch": 0.07575038959279085, + "grad_norm": 14.182686805725098, + "learning_rate": 9.949756999110138e-05, + "loss": 1.5135, + "step": 1118 + }, + { + "epoch": 0.07581814486076292, + "grad_norm": 10.668661117553711, + "learning_rate": 9.949620097200356e-05, + "loss": 1.2412, + "step": 1119 + }, + { + "epoch": 0.07588590012873501, + "grad_norm": 14.443583488464355, + "learning_rate": 9.949483195290574e-05, + "loss": 1.207, + "step": 1120 + }, + { + "epoch": 0.07595365539670709, + "grad_norm": 12.418794631958008, + "learning_rate": 9.949346293380794e-05, + "loss": 1.1907, + "step": 1121 + }, + { + "epoch": 0.07602141066467918, + "grad_norm": 12.429618835449219, + "learning_rate": 9.949209391471012e-05, + "loss": 1.4403, + "step": 1122 + }, + { + "epoch": 0.07608916593265126, + "grad_norm": 9.524622917175293, + "learning_rate": 9.94907248956123e-05, + "loss": 1.139, + "step": 1123 + }, + { + "epoch": 0.07615692120062335, + "grad_norm": 10.974812507629395, + "learning_rate": 9.948935587651448e-05, + "loss": 1.4055, + "step": 1124 + }, + { + "epoch": 0.07622467646859543, + "grad_norm": 16.152681350708008, + "learning_rate": 9.948798685741667e-05, + "loss": 1.0874, + "step": 1125 + }, + { + "epoch": 0.07629243173656752, + "grad_norm": 12.023541450500488, + "learning_rate": 9.948661783831885e-05, + "loss": 1.2378, + "step": 1126 + }, + { + "epoch": 0.0763601870045396, + "grad_norm": 11.597234725952148, + "learning_rate": 9.948524881922103e-05, + "loss": 1.0104, + "step": 1127 + }, + { + "epoch": 0.07642794227251169, + "grad_norm": 11.374302864074707, + "learning_rate": 9.948387980012321e-05, + "loss": 1.1099, + "step": 1128 + }, + { + "epoch": 0.07649569754048377, + "grad_norm": 14.398423194885254, + "learning_rate": 9.94825107810254e-05, + "loss": 1.5533, + "step": 1129 + }, + { + "epoch": 0.07656345280845586, + "grad_norm": 13.1026611328125, + "learning_rate": 9.948114176192759e-05, + "loss": 1.3948, + "step": 1130 + }, + { + "epoch": 0.07663120807642794, + "grad_norm": 12.012560844421387, + "learning_rate": 9.947977274282977e-05, + "loss": 1.2156, + "step": 1131 + }, + { + "epoch": 0.07669896334440003, + "grad_norm": 12.900229454040527, + "learning_rate": 9.947840372373195e-05, + "loss": 1.2658, + "step": 1132 + }, + { + "epoch": 0.07676671861237211, + "grad_norm": 15.250492095947266, + "learning_rate": 9.947703470463413e-05, + "loss": 1.6102, + "step": 1133 + }, + { + "epoch": 0.0768344738803442, + "grad_norm": 15.08134937286377, + "learning_rate": 9.947566568553632e-05, + "loss": 1.1947, + "step": 1134 + }, + { + "epoch": 0.07690222914831628, + "grad_norm": 13.353601455688477, + "learning_rate": 9.94742966664385e-05, + "loss": 1.3211, + "step": 1135 + }, + { + "epoch": 0.07696998441628837, + "grad_norm": 11.310175895690918, + "learning_rate": 9.947292764734068e-05, + "loss": 1.2223, + "step": 1136 + }, + { + "epoch": 0.07703773968426045, + "grad_norm": 11.800848960876465, + "learning_rate": 9.947155862824286e-05, + "loss": 1.2131, + "step": 1137 + }, + { + "epoch": 0.07710549495223254, + "grad_norm": 11.748014450073242, + "learning_rate": 9.947018960914504e-05, + "loss": 1.0734, + "step": 1138 + }, + { + "epoch": 0.07717325022020462, + "grad_norm": 12.282258033752441, + "learning_rate": 9.946882059004724e-05, + "loss": 1.3257, + "step": 1139 + }, + { + "epoch": 0.07724100548817671, + "grad_norm": 11.93818473815918, + "learning_rate": 9.946745157094942e-05, + "loss": 1.1452, + "step": 1140 + }, + { + "epoch": 0.07730876075614879, + "grad_norm": 13.397029876708984, + "learning_rate": 9.94660825518516e-05, + "loss": 1.2625, + "step": 1141 + }, + { + "epoch": 0.07737651602412088, + "grad_norm": 12.135769844055176, + "learning_rate": 9.946471353275378e-05, + "loss": 1.3624, + "step": 1142 + }, + { + "epoch": 0.07744427129209296, + "grad_norm": 11.304028511047363, + "learning_rate": 9.946334451365596e-05, + "loss": 1.3019, + "step": 1143 + }, + { + "epoch": 0.07751202656006505, + "grad_norm": 10.978137016296387, + "learning_rate": 9.946197549455815e-05, + "loss": 1.3137, + "step": 1144 + }, + { + "epoch": 0.07757978182803713, + "grad_norm": 10.997323989868164, + "learning_rate": 9.946060647546033e-05, + "loss": 1.3573, + "step": 1145 + }, + { + "epoch": 0.07764753709600922, + "grad_norm": 11.883647918701172, + "learning_rate": 9.945923745636251e-05, + "loss": 1.2587, + "step": 1146 + }, + { + "epoch": 0.0777152923639813, + "grad_norm": 10.70753288269043, + "learning_rate": 9.945786843726471e-05, + "loss": 1.0866, + "step": 1147 + }, + { + "epoch": 0.07778304763195339, + "grad_norm": 13.318743705749512, + "learning_rate": 9.945649941816689e-05, + "loss": 1.3356, + "step": 1148 + }, + { + "epoch": 0.07785080289992546, + "grad_norm": 10.656171798706055, + "learning_rate": 9.945513039906907e-05, + "loss": 1.2155, + "step": 1149 + }, + { + "epoch": 0.07791855816789756, + "grad_norm": 14.73982048034668, + "learning_rate": 9.945376137997126e-05, + "loss": 1.2276, + "step": 1150 + }, + { + "epoch": 0.07798631343586963, + "grad_norm": 12.461714744567871, + "learning_rate": 9.945239236087344e-05, + "loss": 1.2999, + "step": 1151 + }, + { + "epoch": 0.07805406870384173, + "grad_norm": 12.277376174926758, + "learning_rate": 9.945102334177562e-05, + "loss": 1.1131, + "step": 1152 + }, + { + "epoch": 0.0781218239718138, + "grad_norm": 12.726540565490723, + "learning_rate": 9.944965432267782e-05, + "loss": 1.4181, + "step": 1153 + }, + { + "epoch": 0.0781895792397859, + "grad_norm": 11.086180686950684, + "learning_rate": 9.944828530358e-05, + "loss": 1.3852, + "step": 1154 + }, + { + "epoch": 0.07825733450775797, + "grad_norm": 12.80537223815918, + "learning_rate": 9.944691628448218e-05, + "loss": 1.1101, + "step": 1155 + }, + { + "epoch": 0.07832508977573006, + "grad_norm": 11.960269927978516, + "learning_rate": 9.944554726538436e-05, + "loss": 1.2186, + "step": 1156 + }, + { + "epoch": 0.07839284504370214, + "grad_norm": 11.732439041137695, + "learning_rate": 9.944417824628655e-05, + "loss": 0.9854, + "step": 1157 + }, + { + "epoch": 0.07846060031167423, + "grad_norm": 12.578715324401855, + "learning_rate": 9.944280922718873e-05, + "loss": 1.0792, + "step": 1158 + }, + { + "epoch": 0.07852835557964631, + "grad_norm": 13.779712677001953, + "learning_rate": 9.944144020809091e-05, + "loss": 1.2468, + "step": 1159 + }, + { + "epoch": 0.0785961108476184, + "grad_norm": 13.95693588256836, + "learning_rate": 9.944007118899309e-05, + "loss": 1.234, + "step": 1160 + }, + { + "epoch": 0.07866386611559048, + "grad_norm": 12.056897163391113, + "learning_rate": 9.943870216989527e-05, + "loss": 1.3034, + "step": 1161 + }, + { + "epoch": 0.07873162138356257, + "grad_norm": 11.633442878723145, + "learning_rate": 9.943733315079747e-05, + "loss": 1.2744, + "step": 1162 + }, + { + "epoch": 0.07879937665153465, + "grad_norm": 14.062381744384766, + "learning_rate": 9.943596413169965e-05, + "loss": 1.2535, + "step": 1163 + }, + { + "epoch": 0.07886713191950674, + "grad_norm": 11.60498332977295, + "learning_rate": 9.943459511260183e-05, + "loss": 1.1578, + "step": 1164 + }, + { + "epoch": 0.07893488718747883, + "grad_norm": 9.667806625366211, + "learning_rate": 9.943322609350401e-05, + "loss": 1.0804, + "step": 1165 + }, + { + "epoch": 0.07900264245545091, + "grad_norm": 12.30827808380127, + "learning_rate": 9.943185707440619e-05, + "loss": 1.4167, + "step": 1166 + }, + { + "epoch": 0.079070397723423, + "grad_norm": 10.196819305419922, + "learning_rate": 9.943048805530838e-05, + "loss": 1.0547, + "step": 1167 + }, + { + "epoch": 0.07913815299139508, + "grad_norm": 10.029928207397461, + "learning_rate": 9.942911903621056e-05, + "loss": 1.1798, + "step": 1168 + }, + { + "epoch": 0.07920590825936717, + "grad_norm": 11.782978057861328, + "learning_rate": 9.942775001711274e-05, + "loss": 1.1454, + "step": 1169 + }, + { + "epoch": 0.07927366352733925, + "grad_norm": 12.25143814086914, + "learning_rate": 9.942638099801492e-05, + "loss": 1.2854, + "step": 1170 + }, + { + "epoch": 0.07934141879531134, + "grad_norm": 8.47904109954834, + "learning_rate": 9.942501197891712e-05, + "loss": 0.9381, + "step": 1171 + }, + { + "epoch": 0.07940917406328342, + "grad_norm": 13.698802947998047, + "learning_rate": 9.94236429598193e-05, + "loss": 1.6261, + "step": 1172 + }, + { + "epoch": 0.07947692933125551, + "grad_norm": 10.885397911071777, + "learning_rate": 9.942227394072148e-05, + "loss": 1.0645, + "step": 1173 + }, + { + "epoch": 0.07954468459922759, + "grad_norm": 13.274818420410156, + "learning_rate": 9.942090492162366e-05, + "loss": 1.3411, + "step": 1174 + }, + { + "epoch": 0.07961243986719968, + "grad_norm": 14.061238288879395, + "learning_rate": 9.941953590252584e-05, + "loss": 1.1916, + "step": 1175 + }, + { + "epoch": 0.07968019513517176, + "grad_norm": 10.074264526367188, + "learning_rate": 9.941816688342803e-05, + "loss": 1.092, + "step": 1176 + }, + { + "epoch": 0.07974795040314385, + "grad_norm": 14.741287231445312, + "learning_rate": 9.941679786433021e-05, + "loss": 1.3774, + "step": 1177 + }, + { + "epoch": 0.07981570567111593, + "grad_norm": 11.308422088623047, + "learning_rate": 9.941542884523239e-05, + "loss": 0.9298, + "step": 1178 + }, + { + "epoch": 0.07988346093908802, + "grad_norm": 14.375280380249023, + "learning_rate": 9.941405982613457e-05, + "loss": 1.4525, + "step": 1179 + }, + { + "epoch": 0.0799512162070601, + "grad_norm": 11.724523544311523, + "learning_rate": 9.941269080703677e-05, + "loss": 1.3616, + "step": 1180 + }, + { + "epoch": 0.08001897147503219, + "grad_norm": 12.578176498413086, + "learning_rate": 9.941132178793895e-05, + "loss": 1.1513, + "step": 1181 + }, + { + "epoch": 0.08008672674300427, + "grad_norm": 12.100804328918457, + "learning_rate": 9.940995276884113e-05, + "loss": 1.0656, + "step": 1182 + }, + { + "epoch": 0.08015448201097636, + "grad_norm": 11.680248260498047, + "learning_rate": 9.940858374974331e-05, + "loss": 1.2358, + "step": 1183 + }, + { + "epoch": 0.08022223727894844, + "grad_norm": 10.066198348999023, + "learning_rate": 9.940721473064549e-05, + "loss": 0.9219, + "step": 1184 + }, + { + "epoch": 0.08028999254692053, + "grad_norm": 10.813334465026855, + "learning_rate": 9.940584571154768e-05, + "loss": 1.1456, + "step": 1185 + }, + { + "epoch": 0.0803577478148926, + "grad_norm": 14.004862785339355, + "learning_rate": 9.940447669244986e-05, + "loss": 1.2375, + "step": 1186 + }, + { + "epoch": 0.0804255030828647, + "grad_norm": 11.868766784667969, + "learning_rate": 9.940310767335204e-05, + "loss": 1.3027, + "step": 1187 + }, + { + "epoch": 0.08049325835083677, + "grad_norm": 12.48153018951416, + "learning_rate": 9.940173865425422e-05, + "loss": 1.1192, + "step": 1188 + }, + { + "epoch": 0.08056101361880887, + "grad_norm": 12.340612411499023, + "learning_rate": 9.940036963515642e-05, + "loss": 1.246, + "step": 1189 + }, + { + "epoch": 0.08062876888678094, + "grad_norm": 12.205392837524414, + "learning_rate": 9.93990006160586e-05, + "loss": 1.1233, + "step": 1190 + }, + { + "epoch": 0.08069652415475304, + "grad_norm": 12.69509220123291, + "learning_rate": 9.939763159696078e-05, + "loss": 1.2202, + "step": 1191 + }, + { + "epoch": 0.08076427942272511, + "grad_norm": 12.40784740447998, + "learning_rate": 9.939626257786296e-05, + "loss": 1.1345, + "step": 1192 + }, + { + "epoch": 0.0808320346906972, + "grad_norm": 11.678507804870605, + "learning_rate": 9.939489355876515e-05, + "loss": 1.4513, + "step": 1193 + }, + { + "epoch": 0.08089978995866928, + "grad_norm": 11.649873733520508, + "learning_rate": 9.939352453966733e-05, + "loss": 1.3827, + "step": 1194 + }, + { + "epoch": 0.08096754522664137, + "grad_norm": 12.378853797912598, + "learning_rate": 9.939215552056951e-05, + "loss": 1.3632, + "step": 1195 + }, + { + "epoch": 0.08103530049461345, + "grad_norm": 11.023188591003418, + "learning_rate": 9.93907865014717e-05, + "loss": 1.3892, + "step": 1196 + }, + { + "epoch": 0.08110305576258554, + "grad_norm": 13.111897468566895, + "learning_rate": 9.938941748237389e-05, + "loss": 1.1973, + "step": 1197 + }, + { + "epoch": 0.08117081103055762, + "grad_norm": 10.171613693237305, + "learning_rate": 9.938804846327607e-05, + "loss": 1.3862, + "step": 1198 + }, + { + "epoch": 0.08123856629852971, + "grad_norm": 13.327658653259277, + "learning_rate": 9.938667944417826e-05, + "loss": 1.3173, + "step": 1199 + }, + { + "epoch": 0.08130632156650179, + "grad_norm": 11.715154647827148, + "learning_rate": 9.938531042508044e-05, + "loss": 1.0839, + "step": 1200 + }, + { + "epoch": 0.08137407683447388, + "grad_norm": 11.521212577819824, + "learning_rate": 9.938394140598262e-05, + "loss": 1.0285, + "step": 1201 + }, + { + "epoch": 0.08144183210244596, + "grad_norm": 12.024236679077148, + "learning_rate": 9.93825723868848e-05, + "loss": 1.3025, + "step": 1202 + }, + { + "epoch": 0.08150958737041805, + "grad_norm": 10.245376586914062, + "learning_rate": 9.9381203367787e-05, + "loss": 1.1658, + "step": 1203 + }, + { + "epoch": 0.08157734263839013, + "grad_norm": 10.731759071350098, + "learning_rate": 9.937983434868918e-05, + "loss": 1.0565, + "step": 1204 + }, + { + "epoch": 0.08164509790636222, + "grad_norm": 16.714153289794922, + "learning_rate": 9.937846532959136e-05, + "loss": 1.2761, + "step": 1205 + }, + { + "epoch": 0.0817128531743343, + "grad_norm": 11.172699928283691, + "learning_rate": 9.937709631049354e-05, + "loss": 1.1887, + "step": 1206 + }, + { + "epoch": 0.08178060844230639, + "grad_norm": 11.384743690490723, + "learning_rate": 9.937572729139572e-05, + "loss": 0.9546, + "step": 1207 + }, + { + "epoch": 0.08184836371027847, + "grad_norm": 11.163822174072266, + "learning_rate": 9.937435827229791e-05, + "loss": 1.1498, + "step": 1208 + }, + { + "epoch": 0.08191611897825056, + "grad_norm": 10.657593727111816, + "learning_rate": 9.937298925320009e-05, + "loss": 0.9506, + "step": 1209 + }, + { + "epoch": 0.08198387424622264, + "grad_norm": 10.6862211227417, + "learning_rate": 9.937162023410227e-05, + "loss": 1.2308, + "step": 1210 + }, + { + "epoch": 0.08205162951419473, + "grad_norm": 10.649473190307617, + "learning_rate": 9.937025121500445e-05, + "loss": 1.2679, + "step": 1211 + }, + { + "epoch": 0.08211938478216681, + "grad_norm": 9.298782348632812, + "learning_rate": 9.936888219590665e-05, + "loss": 1.1327, + "step": 1212 + }, + { + "epoch": 0.0821871400501389, + "grad_norm": 11.448348045349121, + "learning_rate": 9.936751317680883e-05, + "loss": 1.2796, + "step": 1213 + }, + { + "epoch": 0.08225489531811098, + "grad_norm": 12.141517639160156, + "learning_rate": 9.936614415771101e-05, + "loss": 1.1128, + "step": 1214 + }, + { + "epoch": 0.08232265058608307, + "grad_norm": 11.799830436706543, + "learning_rate": 9.936477513861319e-05, + "loss": 1.0743, + "step": 1215 + }, + { + "epoch": 0.08239040585405515, + "grad_norm": 11.952958106994629, + "learning_rate": 9.936340611951537e-05, + "loss": 1.1659, + "step": 1216 + }, + { + "epoch": 0.08245816112202724, + "grad_norm": 11.870144844055176, + "learning_rate": 9.936203710041756e-05, + "loss": 1.1195, + "step": 1217 + }, + { + "epoch": 0.08252591638999933, + "grad_norm": 11.970368385314941, + "learning_rate": 9.936066808131974e-05, + "loss": 1.2139, + "step": 1218 + }, + { + "epoch": 0.08259367165797141, + "grad_norm": 11.211687088012695, + "learning_rate": 9.935929906222192e-05, + "loss": 1.2107, + "step": 1219 + }, + { + "epoch": 0.0826614269259435, + "grad_norm": 12.987563133239746, + "learning_rate": 9.93579300431241e-05, + "loss": 1.2804, + "step": 1220 + }, + { + "epoch": 0.08272918219391558, + "grad_norm": 12.337888717651367, + "learning_rate": 9.935656102402628e-05, + "loss": 0.8713, + "step": 1221 + }, + { + "epoch": 0.08279693746188767, + "grad_norm": 11.717150688171387, + "learning_rate": 9.935519200492848e-05, + "loss": 1.2684, + "step": 1222 + }, + { + "epoch": 0.08286469272985975, + "grad_norm": 12.658769607543945, + "learning_rate": 9.935382298583066e-05, + "loss": 1.0231, + "step": 1223 + }, + { + "epoch": 0.08293244799783184, + "grad_norm": 10.526476860046387, + "learning_rate": 9.935245396673284e-05, + "loss": 1.0605, + "step": 1224 + }, + { + "epoch": 0.08300020326580392, + "grad_norm": 10.575004577636719, + "learning_rate": 9.935108494763502e-05, + "loss": 1.2169, + "step": 1225 + }, + { + "epoch": 0.083067958533776, + "grad_norm": 10.499407768249512, + "learning_rate": 9.934971592853721e-05, + "loss": 1.3064, + "step": 1226 + }, + { + "epoch": 0.08313571380174808, + "grad_norm": 12.25387191772461, + "learning_rate": 9.934834690943939e-05, + "loss": 1.1554, + "step": 1227 + }, + { + "epoch": 0.08320346906972018, + "grad_norm": 12.173775672912598, + "learning_rate": 9.934697789034157e-05, + "loss": 1.0917, + "step": 1228 + }, + { + "epoch": 0.08327122433769225, + "grad_norm": 12.424721717834473, + "learning_rate": 9.934560887124375e-05, + "loss": 1.3679, + "step": 1229 + }, + { + "epoch": 0.08333897960566435, + "grad_norm": 11.291987419128418, + "learning_rate": 9.934423985214593e-05, + "loss": 1.1439, + "step": 1230 + }, + { + "epoch": 0.08340673487363642, + "grad_norm": 10.908637046813965, + "learning_rate": 9.934287083304813e-05, + "loss": 1.0487, + "step": 1231 + }, + { + "epoch": 0.08347449014160851, + "grad_norm": 14.767544746398926, + "learning_rate": 9.934150181395031e-05, + "loss": 1.2951, + "step": 1232 + }, + { + "epoch": 0.08354224540958059, + "grad_norm": 11.959871292114258, + "learning_rate": 9.934013279485249e-05, + "loss": 1.1244, + "step": 1233 + }, + { + "epoch": 0.08361000067755268, + "grad_norm": 11.19450569152832, + "learning_rate": 9.933876377575467e-05, + "loss": 1.2802, + "step": 1234 + }, + { + "epoch": 0.08367775594552476, + "grad_norm": 10.71377182006836, + "learning_rate": 9.933739475665686e-05, + "loss": 1.0376, + "step": 1235 + }, + { + "epoch": 0.08374551121349685, + "grad_norm": 12.174454689025879, + "learning_rate": 9.933602573755904e-05, + "loss": 1.5779, + "step": 1236 + }, + { + "epoch": 0.08381326648146893, + "grad_norm": 9.863836288452148, + "learning_rate": 9.933465671846122e-05, + "loss": 1.0443, + "step": 1237 + }, + { + "epoch": 0.08388102174944102, + "grad_norm": 11.964838027954102, + "learning_rate": 9.93332876993634e-05, + "loss": 1.0613, + "step": 1238 + }, + { + "epoch": 0.0839487770174131, + "grad_norm": 11.49203109741211, + "learning_rate": 9.93319186802656e-05, + "loss": 1.162, + "step": 1239 + }, + { + "epoch": 0.08401653228538519, + "grad_norm": 12.548815727233887, + "learning_rate": 9.933054966116778e-05, + "loss": 1.0652, + "step": 1240 + }, + { + "epoch": 0.08408428755335727, + "grad_norm": 13.37637996673584, + "learning_rate": 9.932918064206996e-05, + "loss": 1.2281, + "step": 1241 + }, + { + "epoch": 0.08415204282132936, + "grad_norm": 12.59211254119873, + "learning_rate": 9.932781162297215e-05, + "loss": 1.0624, + "step": 1242 + }, + { + "epoch": 0.08421979808930144, + "grad_norm": 13.386221885681152, + "learning_rate": 9.932644260387433e-05, + "loss": 1.4393, + "step": 1243 + }, + { + "epoch": 0.08428755335727353, + "grad_norm": 12.647525787353516, + "learning_rate": 9.932507358477651e-05, + "loss": 1.2228, + "step": 1244 + }, + { + "epoch": 0.08435530862524561, + "grad_norm": 12.039474487304688, + "learning_rate": 9.93237045656787e-05, + "loss": 1.1354, + "step": 1245 + }, + { + "epoch": 0.0844230638932177, + "grad_norm": 11.373556137084961, + "learning_rate": 9.932233554658089e-05, + "loss": 1.2418, + "step": 1246 + }, + { + "epoch": 0.08449081916118978, + "grad_norm": 10.944781303405762, + "learning_rate": 9.932096652748307e-05, + "loss": 1.1515, + "step": 1247 + }, + { + "epoch": 0.08455857442916187, + "grad_norm": 12.174854278564453, + "learning_rate": 9.931959750838525e-05, + "loss": 1.2063, + "step": 1248 + }, + { + "epoch": 0.08462632969713395, + "grad_norm": 8.846879005432129, + "learning_rate": 9.931822848928744e-05, + "loss": 0.9639, + "step": 1249 + }, + { + "epoch": 0.08469408496510604, + "grad_norm": 13.793547630310059, + "learning_rate": 9.931685947018962e-05, + "loss": 1.1236, + "step": 1250 + }, + { + "epoch": 0.08476184023307812, + "grad_norm": 14.486831665039062, + "learning_rate": 9.93154904510918e-05, + "loss": 1.2047, + "step": 1251 + }, + { + "epoch": 0.08482959550105021, + "grad_norm": 13.262588500976562, + "learning_rate": 9.931412143199398e-05, + "loss": 1.4091, + "step": 1252 + }, + { + "epoch": 0.08489735076902229, + "grad_norm": 13.289068222045898, + "learning_rate": 9.931275241289616e-05, + "loss": 1.5065, + "step": 1253 + }, + { + "epoch": 0.08496510603699438, + "grad_norm": 10.22205638885498, + "learning_rate": 9.931138339379836e-05, + "loss": 1.1116, + "step": 1254 + }, + { + "epoch": 0.08503286130496646, + "grad_norm": 13.141668319702148, + "learning_rate": 9.931001437470054e-05, + "loss": 1.4006, + "step": 1255 + }, + { + "epoch": 0.08510061657293855, + "grad_norm": 11.817032814025879, + "learning_rate": 9.930864535560272e-05, + "loss": 1.2062, + "step": 1256 + }, + { + "epoch": 0.08516837184091063, + "grad_norm": 10.814498901367188, + "learning_rate": 9.93072763365049e-05, + "loss": 1.138, + "step": 1257 + }, + { + "epoch": 0.08523612710888272, + "grad_norm": 12.598155975341797, + "learning_rate": 9.930590731740709e-05, + "loss": 1.1959, + "step": 1258 + }, + { + "epoch": 0.0853038823768548, + "grad_norm": 11.909974098205566, + "learning_rate": 9.930453829830927e-05, + "loss": 1.2891, + "step": 1259 + }, + { + "epoch": 0.08537163764482689, + "grad_norm": 12.663064956665039, + "learning_rate": 9.930316927921145e-05, + "loss": 1.3017, + "step": 1260 + }, + { + "epoch": 0.08543939291279896, + "grad_norm": 10.810627937316895, + "learning_rate": 9.930180026011363e-05, + "loss": 1.16, + "step": 1261 + }, + { + "epoch": 0.08550714818077106, + "grad_norm": 9.182926177978516, + "learning_rate": 9.930043124101581e-05, + "loss": 1.2625, + "step": 1262 + }, + { + "epoch": 0.08557490344874313, + "grad_norm": 12.73978042602539, + "learning_rate": 9.9299062221918e-05, + "loss": 0.9904, + "step": 1263 + }, + { + "epoch": 0.08564265871671523, + "grad_norm": 10.346587181091309, + "learning_rate": 9.929769320282019e-05, + "loss": 1.0167, + "step": 1264 + }, + { + "epoch": 0.0857104139846873, + "grad_norm": 12.145682334899902, + "learning_rate": 9.929632418372237e-05, + "loss": 1.4282, + "step": 1265 + }, + { + "epoch": 0.0857781692526594, + "grad_norm": 11.515445709228516, + "learning_rate": 9.929495516462455e-05, + "loss": 1.0527, + "step": 1266 + }, + { + "epoch": 0.08584592452063147, + "grad_norm": 10.979050636291504, + "learning_rate": 9.929358614552674e-05, + "loss": 1.2888, + "step": 1267 + }, + { + "epoch": 0.08591367978860356, + "grad_norm": 10.248215675354004, + "learning_rate": 9.929221712642892e-05, + "loss": 1.4316, + "step": 1268 + }, + { + "epoch": 0.08598143505657564, + "grad_norm": 13.743851661682129, + "learning_rate": 9.92908481073311e-05, + "loss": 1.5047, + "step": 1269 + }, + { + "epoch": 0.08604919032454773, + "grad_norm": 9.919225692749023, + "learning_rate": 9.928947908823328e-05, + "loss": 0.9602, + "step": 1270 + }, + { + "epoch": 0.08611694559251983, + "grad_norm": 9.795915603637695, + "learning_rate": 9.928811006913546e-05, + "loss": 1.3736, + "step": 1271 + }, + { + "epoch": 0.0861847008604919, + "grad_norm": 13.108200073242188, + "learning_rate": 9.928674105003766e-05, + "loss": 1.2591, + "step": 1272 + }, + { + "epoch": 0.086252456128464, + "grad_norm": 13.108073234558105, + "learning_rate": 9.928537203093984e-05, + "loss": 1.102, + "step": 1273 + }, + { + "epoch": 0.08632021139643607, + "grad_norm": 15.177817344665527, + "learning_rate": 9.928400301184202e-05, + "loss": 1.3164, + "step": 1274 + }, + { + "epoch": 0.08638796666440816, + "grad_norm": 13.048440933227539, + "learning_rate": 9.92826339927442e-05, + "loss": 1.3029, + "step": 1275 + }, + { + "epoch": 0.08645572193238024, + "grad_norm": 10.982895851135254, + "learning_rate": 9.928126497364638e-05, + "loss": 1.1829, + "step": 1276 + }, + { + "epoch": 0.08652347720035233, + "grad_norm": 10.901629447937012, + "learning_rate": 9.927989595454857e-05, + "loss": 1.2116, + "step": 1277 + }, + { + "epoch": 0.08659123246832441, + "grad_norm": 12.924722671508789, + "learning_rate": 9.927852693545075e-05, + "loss": 1.6292, + "step": 1278 + }, + { + "epoch": 0.0866589877362965, + "grad_norm": 12.572770118713379, + "learning_rate": 9.927715791635293e-05, + "loss": 1.1012, + "step": 1279 + }, + { + "epoch": 0.08672674300426858, + "grad_norm": 12.803020477294922, + "learning_rate": 9.927578889725511e-05, + "loss": 1.276, + "step": 1280 + }, + { + "epoch": 0.08679449827224067, + "grad_norm": 10.92810344696045, + "learning_rate": 9.92744198781573e-05, + "loss": 1.1876, + "step": 1281 + }, + { + "epoch": 0.08686225354021275, + "grad_norm": 12.253180503845215, + "learning_rate": 9.927305085905949e-05, + "loss": 1.1807, + "step": 1282 + }, + { + "epoch": 0.08693000880818484, + "grad_norm": 11.163126945495605, + "learning_rate": 9.927168183996167e-05, + "loss": 1.1671, + "step": 1283 + }, + { + "epoch": 0.08699776407615692, + "grad_norm": 10.726607322692871, + "learning_rate": 9.927031282086385e-05, + "loss": 1.185, + "step": 1284 + }, + { + "epoch": 0.08706551934412901, + "grad_norm": 13.265491485595703, + "learning_rate": 9.926894380176603e-05, + "loss": 1.2916, + "step": 1285 + }, + { + "epoch": 0.08713327461210109, + "grad_norm": 14.559592247009277, + "learning_rate": 9.926757478266822e-05, + "loss": 1.288, + "step": 1286 + }, + { + "epoch": 0.08720102988007318, + "grad_norm": 14.816813468933105, + "learning_rate": 9.92662057635704e-05, + "loss": 1.4397, + "step": 1287 + }, + { + "epoch": 0.08726878514804526, + "grad_norm": 11.811420440673828, + "learning_rate": 9.926483674447258e-05, + "loss": 1.348, + "step": 1288 + }, + { + "epoch": 0.08733654041601735, + "grad_norm": 10.920133590698242, + "learning_rate": 9.926346772537478e-05, + "loss": 1.4169, + "step": 1289 + }, + { + "epoch": 0.08740429568398943, + "grad_norm": 11.690089225769043, + "learning_rate": 9.926209870627696e-05, + "loss": 1.1199, + "step": 1290 + }, + { + "epoch": 0.08747205095196152, + "grad_norm": 9.411031723022461, + "learning_rate": 9.926072968717914e-05, + "loss": 1.031, + "step": 1291 + }, + { + "epoch": 0.0875398062199336, + "grad_norm": 12.174457550048828, + "learning_rate": 9.925936066808133e-05, + "loss": 1.0622, + "step": 1292 + }, + { + "epoch": 0.08760756148790569, + "grad_norm": 10.346089363098145, + "learning_rate": 9.925799164898351e-05, + "loss": 1.2777, + "step": 1293 + }, + { + "epoch": 0.08767531675587777, + "grad_norm": 12.534863471984863, + "learning_rate": 9.925662262988569e-05, + "loss": 1.0093, + "step": 1294 + }, + { + "epoch": 0.08774307202384986, + "grad_norm": 12.050302505493164, + "learning_rate": 9.925525361078789e-05, + "loss": 1.0591, + "step": 1295 + }, + { + "epoch": 0.08781082729182194, + "grad_norm": 11.556166648864746, + "learning_rate": 9.925388459169007e-05, + "loss": 1.1452, + "step": 1296 + }, + { + "epoch": 0.08787858255979403, + "grad_norm": 9.693270683288574, + "learning_rate": 9.925251557259225e-05, + "loss": 1.0332, + "step": 1297 + }, + { + "epoch": 0.0879463378277661, + "grad_norm": 12.646526336669922, + "learning_rate": 9.925114655349443e-05, + "loss": 1.1894, + "step": 1298 + }, + { + "epoch": 0.0880140930957382, + "grad_norm": 10.676809310913086, + "learning_rate": 9.924977753439661e-05, + "loss": 1.1727, + "step": 1299 + }, + { + "epoch": 0.08808184836371027, + "grad_norm": 11.182327270507812, + "learning_rate": 9.92484085152988e-05, + "loss": 1.0145, + "step": 1300 + }, + { + "epoch": 0.08814960363168237, + "grad_norm": 11.55026626586914, + "learning_rate": 9.924703949620098e-05, + "loss": 1.2187, + "step": 1301 + }, + { + "epoch": 0.08821735889965444, + "grad_norm": 11.502679824829102, + "learning_rate": 9.924567047710316e-05, + "loss": 1.3116, + "step": 1302 + }, + { + "epoch": 0.08828511416762654, + "grad_norm": 9.676247596740723, + "learning_rate": 9.924430145800534e-05, + "loss": 1.2869, + "step": 1303 + }, + { + "epoch": 0.08835286943559861, + "grad_norm": 15.123950004577637, + "learning_rate": 9.924293243890754e-05, + "loss": 1.0225, + "step": 1304 + }, + { + "epoch": 0.0884206247035707, + "grad_norm": 14.030994415283203, + "learning_rate": 9.924156341980972e-05, + "loss": 0.9829, + "step": 1305 + }, + { + "epoch": 0.08848837997154278, + "grad_norm": 10.00402545928955, + "learning_rate": 9.92401944007119e-05, + "loss": 1.0946, + "step": 1306 + }, + { + "epoch": 0.08855613523951487, + "grad_norm": 9.077853202819824, + "learning_rate": 9.923882538161408e-05, + "loss": 0.852, + "step": 1307 + }, + { + "epoch": 0.08862389050748695, + "grad_norm": 12.777885437011719, + "learning_rate": 9.923745636251626e-05, + "loss": 1.0513, + "step": 1308 + }, + { + "epoch": 0.08869164577545904, + "grad_norm": 10.686469078063965, + "learning_rate": 9.923608734341845e-05, + "loss": 1.0873, + "step": 1309 + }, + { + "epoch": 0.08875940104343112, + "grad_norm": 11.51689338684082, + "learning_rate": 9.923471832432063e-05, + "loss": 1.0037, + "step": 1310 + }, + { + "epoch": 0.08882715631140321, + "grad_norm": 13.259784698486328, + "learning_rate": 9.923334930522281e-05, + "loss": 1.2004, + "step": 1311 + }, + { + "epoch": 0.08889491157937529, + "grad_norm": 10.63463306427002, + "learning_rate": 9.923198028612499e-05, + "loss": 1.2067, + "step": 1312 + }, + { + "epoch": 0.08896266684734738, + "grad_norm": 10.838210105895996, + "learning_rate": 9.923061126702719e-05, + "loss": 1.2939, + "step": 1313 + }, + { + "epoch": 0.08903042211531946, + "grad_norm": 12.058418273925781, + "learning_rate": 9.922924224792937e-05, + "loss": 1.3091, + "step": 1314 + }, + { + "epoch": 0.08909817738329155, + "grad_norm": 10.619451522827148, + "learning_rate": 9.922787322883155e-05, + "loss": 1.1635, + "step": 1315 + }, + { + "epoch": 0.08916593265126363, + "grad_norm": 9.37607192993164, + "learning_rate": 9.922650420973373e-05, + "loss": 1.2206, + "step": 1316 + }, + { + "epoch": 0.08923368791923572, + "grad_norm": 12.119776725769043, + "learning_rate": 9.922513519063591e-05, + "loss": 1.2159, + "step": 1317 + }, + { + "epoch": 0.0893014431872078, + "grad_norm": 14.013461112976074, + "learning_rate": 9.92237661715381e-05, + "loss": 1.1445, + "step": 1318 + }, + { + "epoch": 0.08936919845517989, + "grad_norm": 11.560707092285156, + "learning_rate": 9.922239715244028e-05, + "loss": 1.3617, + "step": 1319 + }, + { + "epoch": 0.08943695372315197, + "grad_norm": 11.817791938781738, + "learning_rate": 9.922102813334246e-05, + "loss": 1.1931, + "step": 1320 + }, + { + "epoch": 0.08950470899112406, + "grad_norm": 11.544127464294434, + "learning_rate": 9.921965911424464e-05, + "loss": 1.0766, + "step": 1321 + }, + { + "epoch": 0.08957246425909614, + "grad_norm": 10.67740249633789, + "learning_rate": 9.921829009514684e-05, + "loss": 1.2103, + "step": 1322 + }, + { + "epoch": 0.08964021952706823, + "grad_norm": 11.774645805358887, + "learning_rate": 9.921692107604902e-05, + "loss": 1.1921, + "step": 1323 + }, + { + "epoch": 0.08970797479504032, + "grad_norm": 12.594759941101074, + "learning_rate": 9.92155520569512e-05, + "loss": 0.906, + "step": 1324 + }, + { + "epoch": 0.0897757300630124, + "grad_norm": 10.988224983215332, + "learning_rate": 9.921418303785338e-05, + "loss": 1.4195, + "step": 1325 + }, + { + "epoch": 0.08984348533098449, + "grad_norm": 12.22718620300293, + "learning_rate": 9.921281401875556e-05, + "loss": 1.2461, + "step": 1326 + }, + { + "epoch": 0.08991124059895657, + "grad_norm": 10.964727401733398, + "learning_rate": 9.921144499965775e-05, + "loss": 1.1254, + "step": 1327 + }, + { + "epoch": 0.08997899586692866, + "grad_norm": 11.327523231506348, + "learning_rate": 9.921007598055993e-05, + "loss": 1.145, + "step": 1328 + }, + { + "epoch": 0.09004675113490074, + "grad_norm": 9.870691299438477, + "learning_rate": 9.920870696146211e-05, + "loss": 1.2408, + "step": 1329 + }, + { + "epoch": 0.09011450640287283, + "grad_norm": 11.02373218536377, + "learning_rate": 9.920733794236429e-05, + "loss": 1.1084, + "step": 1330 + }, + { + "epoch": 0.09018226167084491, + "grad_norm": 12.300410270690918, + "learning_rate": 9.920596892326647e-05, + "loss": 1.1542, + "step": 1331 + }, + { + "epoch": 0.090250016938817, + "grad_norm": 9.832919120788574, + "learning_rate": 9.920459990416867e-05, + "loss": 1.1195, + "step": 1332 + }, + { + "epoch": 0.09031777220678908, + "grad_norm": 10.432522773742676, + "learning_rate": 9.920323088507085e-05, + "loss": 1.1825, + "step": 1333 + }, + { + "epoch": 0.09038552747476117, + "grad_norm": 11.878792762756348, + "learning_rate": 9.920186186597303e-05, + "loss": 0.9647, + "step": 1334 + }, + { + "epoch": 0.09045328274273325, + "grad_norm": 11.866320610046387, + "learning_rate": 9.920049284687522e-05, + "loss": 1.3598, + "step": 1335 + }, + { + "epoch": 0.09052103801070534, + "grad_norm": 14.11543083190918, + "learning_rate": 9.91991238277774e-05, + "loss": 1.4172, + "step": 1336 + }, + { + "epoch": 0.09058879327867742, + "grad_norm": 13.841622352600098, + "learning_rate": 9.919775480867958e-05, + "loss": 1.4119, + "step": 1337 + }, + { + "epoch": 0.0906565485466495, + "grad_norm": 11.077167510986328, + "learning_rate": 9.919638578958178e-05, + "loss": 0.984, + "step": 1338 + }, + { + "epoch": 0.09072430381462158, + "grad_norm": 10.904266357421875, + "learning_rate": 9.919501677048396e-05, + "loss": 1.0439, + "step": 1339 + }, + { + "epoch": 0.09079205908259368, + "grad_norm": 11.623948097229004, + "learning_rate": 9.919364775138614e-05, + "loss": 1.0318, + "step": 1340 + }, + { + "epoch": 0.09085981435056575, + "grad_norm": 10.893725395202637, + "learning_rate": 9.919227873228833e-05, + "loss": 1.2626, + "step": 1341 + }, + { + "epoch": 0.09092756961853785, + "grad_norm": 10.064491271972656, + "learning_rate": 9.919090971319051e-05, + "loss": 1.1483, + "step": 1342 + }, + { + "epoch": 0.09099532488650992, + "grad_norm": 9.854101181030273, + "learning_rate": 9.918954069409269e-05, + "loss": 1.1595, + "step": 1343 + }, + { + "epoch": 0.09106308015448202, + "grad_norm": 11.682498931884766, + "learning_rate": 9.918817167499487e-05, + "loss": 1.2887, + "step": 1344 + }, + { + "epoch": 0.09113083542245409, + "grad_norm": 10.484097480773926, + "learning_rate": 9.918680265589707e-05, + "loss": 1.2, + "step": 1345 + }, + { + "epoch": 0.09119859069042618, + "grad_norm": 12.332358360290527, + "learning_rate": 9.918543363679925e-05, + "loss": 1.262, + "step": 1346 + }, + { + "epoch": 0.09126634595839826, + "grad_norm": 13.706925392150879, + "learning_rate": 9.918406461770143e-05, + "loss": 1.2151, + "step": 1347 + }, + { + "epoch": 0.09133410122637035, + "grad_norm": 13.918478965759277, + "learning_rate": 9.91826955986036e-05, + "loss": 1.4653, + "step": 1348 + }, + { + "epoch": 0.09140185649434243, + "grad_norm": 11.970015525817871, + "learning_rate": 9.918132657950579e-05, + "loss": 1.2452, + "step": 1349 + }, + { + "epoch": 0.09146961176231452, + "grad_norm": 9.698074340820312, + "learning_rate": 9.917995756040798e-05, + "loss": 0.9142, + "step": 1350 + }, + { + "epoch": 0.0915373670302866, + "grad_norm": 9.225728988647461, + "learning_rate": 9.917858854131016e-05, + "loss": 1.0293, + "step": 1351 + }, + { + "epoch": 0.09160512229825869, + "grad_norm": 10.77661418914795, + "learning_rate": 9.917721952221234e-05, + "loss": 1.1865, + "step": 1352 + }, + { + "epoch": 0.09167287756623077, + "grad_norm": 10.45409870147705, + "learning_rate": 9.917585050311452e-05, + "loss": 1.1553, + "step": 1353 + }, + { + "epoch": 0.09174063283420286, + "grad_norm": 10.44918441772461, + "learning_rate": 9.91744814840167e-05, + "loss": 1.1621, + "step": 1354 + }, + { + "epoch": 0.09180838810217494, + "grad_norm": 14.769590377807617, + "learning_rate": 9.91731124649189e-05, + "loss": 1.3502, + "step": 1355 + }, + { + "epoch": 0.09187614337014703, + "grad_norm": 12.733844757080078, + "learning_rate": 9.917174344582108e-05, + "loss": 1.1973, + "step": 1356 + }, + { + "epoch": 0.09194389863811911, + "grad_norm": 13.365818977355957, + "learning_rate": 9.917037442672326e-05, + "loss": 1.5412, + "step": 1357 + }, + { + "epoch": 0.0920116539060912, + "grad_norm": 11.163050651550293, + "learning_rate": 9.916900540762544e-05, + "loss": 1.3142, + "step": 1358 + }, + { + "epoch": 0.09207940917406328, + "grad_norm": 11.420190811157227, + "learning_rate": 9.916763638852763e-05, + "loss": 1.2257, + "step": 1359 + }, + { + "epoch": 0.09214716444203537, + "grad_norm": 9.66398811340332, + "learning_rate": 9.916626736942981e-05, + "loss": 1.0584, + "step": 1360 + }, + { + "epoch": 0.09221491971000745, + "grad_norm": 13.643363952636719, + "learning_rate": 9.916489835033199e-05, + "loss": 1.3783, + "step": 1361 + }, + { + "epoch": 0.09228267497797954, + "grad_norm": 11.658889770507812, + "learning_rate": 9.916352933123417e-05, + "loss": 1.2126, + "step": 1362 + }, + { + "epoch": 0.09235043024595162, + "grad_norm": 11.2728271484375, + "learning_rate": 9.916216031213635e-05, + "loss": 1.2117, + "step": 1363 + }, + { + "epoch": 0.09241818551392371, + "grad_norm": 13.576864242553711, + "learning_rate": 9.916079129303855e-05, + "loss": 1.2922, + "step": 1364 + }, + { + "epoch": 0.09248594078189579, + "grad_norm": 10.950700759887695, + "learning_rate": 9.915942227394073e-05, + "loss": 1.1394, + "step": 1365 + }, + { + "epoch": 0.09255369604986788, + "grad_norm": 11.638351440429688, + "learning_rate": 9.91580532548429e-05, + "loss": 1.3927, + "step": 1366 + }, + { + "epoch": 0.09262145131783996, + "grad_norm": 12.355545043945312, + "learning_rate": 9.915668423574509e-05, + "loss": 1.1861, + "step": 1367 + }, + { + "epoch": 0.09268920658581205, + "grad_norm": 11.543237686157227, + "learning_rate": 9.915531521664728e-05, + "loss": 0.9868, + "step": 1368 + }, + { + "epoch": 0.09275696185378413, + "grad_norm": 11.684252738952637, + "learning_rate": 9.915394619754946e-05, + "loss": 1.3332, + "step": 1369 + }, + { + "epoch": 0.09282471712175622, + "grad_norm": 10.775650024414062, + "learning_rate": 9.915257717845164e-05, + "loss": 1.1164, + "step": 1370 + }, + { + "epoch": 0.0928924723897283, + "grad_norm": 11.649751663208008, + "learning_rate": 9.915120815935382e-05, + "loss": 1.0369, + "step": 1371 + }, + { + "epoch": 0.09296022765770039, + "grad_norm": 9.741403579711914, + "learning_rate": 9.9149839140256e-05, + "loss": 1.0509, + "step": 1372 + }, + { + "epoch": 0.09302798292567246, + "grad_norm": 13.804118156433105, + "learning_rate": 9.91484701211582e-05, + "loss": 1.4318, + "step": 1373 + }, + { + "epoch": 0.09309573819364456, + "grad_norm": 10.939459800720215, + "learning_rate": 9.914710110206038e-05, + "loss": 1.218, + "step": 1374 + }, + { + "epoch": 0.09316349346161663, + "grad_norm": 14.076252937316895, + "learning_rate": 9.914573208296256e-05, + "loss": 1.4699, + "step": 1375 + }, + { + "epoch": 0.09323124872958873, + "grad_norm": 12.90072250366211, + "learning_rate": 9.914436306386474e-05, + "loss": 1.3418, + "step": 1376 + }, + { + "epoch": 0.09329900399756082, + "grad_norm": 12.7711820602417, + "learning_rate": 9.914299404476693e-05, + "loss": 0.9866, + "step": 1377 + }, + { + "epoch": 0.0933667592655329, + "grad_norm": 9.205671310424805, + "learning_rate": 9.914162502566911e-05, + "loss": 1.1345, + "step": 1378 + }, + { + "epoch": 0.09343451453350499, + "grad_norm": 11.027194023132324, + "learning_rate": 9.914025600657129e-05, + "loss": 1.2637, + "step": 1379 + }, + { + "epoch": 0.09350226980147706, + "grad_norm": 12.861044883728027, + "learning_rate": 9.913888698747347e-05, + "loss": 1.2348, + "step": 1380 + }, + { + "epoch": 0.09357002506944916, + "grad_norm": 11.266969680786133, + "learning_rate": 9.913751796837567e-05, + "loss": 1.0773, + "step": 1381 + }, + { + "epoch": 0.09363778033742123, + "grad_norm": 13.137110710144043, + "learning_rate": 9.913614894927785e-05, + "loss": 1.0537, + "step": 1382 + }, + { + "epoch": 0.09370553560539333, + "grad_norm": 11.343362808227539, + "learning_rate": 9.913477993018003e-05, + "loss": 1.335, + "step": 1383 + }, + { + "epoch": 0.0937732908733654, + "grad_norm": 11.472663879394531, + "learning_rate": 9.913341091108222e-05, + "loss": 1.1362, + "step": 1384 + }, + { + "epoch": 0.0938410461413375, + "grad_norm": 8.441573143005371, + "learning_rate": 9.91320418919844e-05, + "loss": 1.1646, + "step": 1385 + }, + { + "epoch": 0.09390880140930957, + "grad_norm": 12.570130348205566, + "learning_rate": 9.913067287288658e-05, + "loss": 1.3535, + "step": 1386 + }, + { + "epoch": 0.09397655667728166, + "grad_norm": 11.671664237976074, + "learning_rate": 9.912930385378877e-05, + "loss": 1.1707, + "step": 1387 + }, + { + "epoch": 0.09404431194525374, + "grad_norm": 12.638328552246094, + "learning_rate": 9.912793483469096e-05, + "loss": 1.057, + "step": 1388 + }, + { + "epoch": 0.09411206721322583, + "grad_norm": 10.506028175354004, + "learning_rate": 9.912656581559314e-05, + "loss": 0.9279, + "step": 1389 + }, + { + "epoch": 0.09417982248119791, + "grad_norm": 11.536858558654785, + "learning_rate": 9.912519679649532e-05, + "loss": 1.3351, + "step": 1390 + }, + { + "epoch": 0.09424757774917, + "grad_norm": 12.692436218261719, + "learning_rate": 9.912382777739751e-05, + "loss": 1.164, + "step": 1391 + }, + { + "epoch": 0.09431533301714208, + "grad_norm": 12.088066101074219, + "learning_rate": 9.912245875829969e-05, + "loss": 1.2971, + "step": 1392 + }, + { + "epoch": 0.09438308828511417, + "grad_norm": 12.133123397827148, + "learning_rate": 9.912108973920187e-05, + "loss": 1.2742, + "step": 1393 + }, + { + "epoch": 0.09445084355308625, + "grad_norm": 10.168001174926758, + "learning_rate": 9.911972072010405e-05, + "loss": 1.0964, + "step": 1394 + }, + { + "epoch": 0.09451859882105834, + "grad_norm": 10.561311721801758, + "learning_rate": 9.911835170100623e-05, + "loss": 1.1828, + "step": 1395 + }, + { + "epoch": 0.09458635408903042, + "grad_norm": 11.497330665588379, + "learning_rate": 9.911698268190843e-05, + "loss": 1.4699, + "step": 1396 + }, + { + "epoch": 0.09465410935700251, + "grad_norm": 12.190573692321777, + "learning_rate": 9.91156136628106e-05, + "loss": 1.1601, + "step": 1397 + }, + { + "epoch": 0.09472186462497459, + "grad_norm": 10.633028030395508, + "learning_rate": 9.911424464371279e-05, + "loss": 1.333, + "step": 1398 + }, + { + "epoch": 0.09478961989294668, + "grad_norm": 12.262279510498047, + "learning_rate": 9.911287562461497e-05, + "loss": 1.2214, + "step": 1399 + }, + { + "epoch": 0.09485737516091876, + "grad_norm": 11.506840705871582, + "learning_rate": 9.911150660551716e-05, + "loss": 1.1172, + "step": 1400 + }, + { + "epoch": 0.09492513042889085, + "grad_norm": 11.453936576843262, + "learning_rate": 9.911013758641934e-05, + "loss": 1.2205, + "step": 1401 + }, + { + "epoch": 0.09499288569686293, + "grad_norm": 9.980772972106934, + "learning_rate": 9.910876856732152e-05, + "loss": 1.1004, + "step": 1402 + }, + { + "epoch": 0.09506064096483502, + "grad_norm": 11.775416374206543, + "learning_rate": 9.91073995482237e-05, + "loss": 1.1587, + "step": 1403 + }, + { + "epoch": 0.0951283962328071, + "grad_norm": 8.840147972106934, + "learning_rate": 9.910603052912588e-05, + "loss": 0.8702, + "step": 1404 + }, + { + "epoch": 0.09519615150077919, + "grad_norm": 10.938506126403809, + "learning_rate": 9.910466151002808e-05, + "loss": 1.085, + "step": 1405 + }, + { + "epoch": 0.09526390676875127, + "grad_norm": 11.733402252197266, + "learning_rate": 9.910329249093026e-05, + "loss": 1.1202, + "step": 1406 + }, + { + "epoch": 0.09533166203672336, + "grad_norm": 11.616521835327148, + "learning_rate": 9.910192347183244e-05, + "loss": 1.3483, + "step": 1407 + }, + { + "epoch": 0.09539941730469544, + "grad_norm": 12.477338790893555, + "learning_rate": 9.910055445273462e-05, + "loss": 1.4798, + "step": 1408 + }, + { + "epoch": 0.09546717257266753, + "grad_norm": 11.233193397521973, + "learning_rate": 9.90991854336368e-05, + "loss": 1.0304, + "step": 1409 + }, + { + "epoch": 0.0955349278406396, + "grad_norm": 12.586124420166016, + "learning_rate": 9.909781641453899e-05, + "loss": 1.2216, + "step": 1410 + }, + { + "epoch": 0.0956026831086117, + "grad_norm": 12.974738121032715, + "learning_rate": 9.909644739544117e-05, + "loss": 1.1495, + "step": 1411 + }, + { + "epoch": 0.09567043837658377, + "grad_norm": 9.613628387451172, + "learning_rate": 9.909507837634335e-05, + "loss": 0.8326, + "step": 1412 + }, + { + "epoch": 0.09573819364455587, + "grad_norm": 10.644312858581543, + "learning_rate": 9.909370935724553e-05, + "loss": 1.0182, + "step": 1413 + }, + { + "epoch": 0.09580594891252794, + "grad_norm": 11.155874252319336, + "learning_rate": 9.909234033814773e-05, + "loss": 1.2887, + "step": 1414 + }, + { + "epoch": 0.09587370418050004, + "grad_norm": 12.068909645080566, + "learning_rate": 9.90909713190499e-05, + "loss": 1.1697, + "step": 1415 + }, + { + "epoch": 0.09594145944847211, + "grad_norm": 10.66831111907959, + "learning_rate": 9.908960229995209e-05, + "loss": 1.1275, + "step": 1416 + }, + { + "epoch": 0.0960092147164442, + "grad_norm": 11.80036449432373, + "learning_rate": 9.908823328085427e-05, + "loss": 1.38, + "step": 1417 + }, + { + "epoch": 0.09607696998441628, + "grad_norm": 11.677534103393555, + "learning_rate": 9.908686426175645e-05, + "loss": 1.075, + "step": 1418 + }, + { + "epoch": 0.09614472525238837, + "grad_norm": 10.54027271270752, + "learning_rate": 9.908549524265864e-05, + "loss": 0.9617, + "step": 1419 + }, + { + "epoch": 0.09621248052036045, + "grad_norm": 9.70718002319336, + "learning_rate": 9.908412622356082e-05, + "loss": 1.0395, + "step": 1420 + }, + { + "epoch": 0.09628023578833254, + "grad_norm": 10.439559936523438, + "learning_rate": 9.9082757204463e-05, + "loss": 1.4112, + "step": 1421 + }, + { + "epoch": 0.09634799105630462, + "grad_norm": 9.328675270080566, + "learning_rate": 9.908138818536518e-05, + "loss": 1.0481, + "step": 1422 + }, + { + "epoch": 0.09641574632427671, + "grad_norm": 12.834508895874023, + "learning_rate": 9.908001916626738e-05, + "loss": 1.0863, + "step": 1423 + }, + { + "epoch": 0.09648350159224879, + "grad_norm": 11.885201454162598, + "learning_rate": 9.907865014716956e-05, + "loss": 1.3509, + "step": 1424 + }, + { + "epoch": 0.09655125686022088, + "grad_norm": 11.299174308776855, + "learning_rate": 9.907728112807174e-05, + "loss": 1.1328, + "step": 1425 + }, + { + "epoch": 0.09661901212819296, + "grad_norm": 13.024226188659668, + "learning_rate": 9.907591210897392e-05, + "loss": 1.1495, + "step": 1426 + }, + { + "epoch": 0.09668676739616505, + "grad_norm": 13.418682098388672, + "learning_rate": 9.907454308987611e-05, + "loss": 1.3827, + "step": 1427 + }, + { + "epoch": 0.09675452266413713, + "grad_norm": 11.28375244140625, + "learning_rate": 9.907317407077829e-05, + "loss": 1.3658, + "step": 1428 + }, + { + "epoch": 0.09682227793210922, + "grad_norm": 9.711199760437012, + "learning_rate": 9.907180505168047e-05, + "loss": 0.8743, + "step": 1429 + }, + { + "epoch": 0.09689003320008131, + "grad_norm": 12.292948722839355, + "learning_rate": 9.907043603258267e-05, + "loss": 1.203, + "step": 1430 + }, + { + "epoch": 0.09695778846805339, + "grad_norm": 13.195072174072266, + "learning_rate": 9.906906701348485e-05, + "loss": 1.0403, + "step": 1431 + }, + { + "epoch": 0.09702554373602548, + "grad_norm": 11.45721435546875, + "learning_rate": 9.906769799438703e-05, + "loss": 1.2075, + "step": 1432 + }, + { + "epoch": 0.09709329900399756, + "grad_norm": 10.477989196777344, + "learning_rate": 9.906632897528922e-05, + "loss": 1.1646, + "step": 1433 + }, + { + "epoch": 0.09716105427196965, + "grad_norm": 12.572269439697266, + "learning_rate": 9.90649599561914e-05, + "loss": 1.3944, + "step": 1434 + }, + { + "epoch": 0.09722880953994173, + "grad_norm": 9.37205982208252, + "learning_rate": 9.906359093709358e-05, + "loss": 0.9902, + "step": 1435 + }, + { + "epoch": 0.09729656480791382, + "grad_norm": 11.590779304504395, + "learning_rate": 9.906222191799576e-05, + "loss": 1.2658, + "step": 1436 + }, + { + "epoch": 0.0973643200758859, + "grad_norm": 10.35207748413086, + "learning_rate": 9.906085289889795e-05, + "loss": 0.7905, + "step": 1437 + }, + { + "epoch": 0.09743207534385799, + "grad_norm": 9.993937492370605, + "learning_rate": 9.905948387980013e-05, + "loss": 1.1522, + "step": 1438 + }, + { + "epoch": 0.09749983061183007, + "grad_norm": 9.865569114685059, + "learning_rate": 9.905811486070232e-05, + "loss": 1.2536, + "step": 1439 + }, + { + "epoch": 0.09756758587980216, + "grad_norm": 12.836588859558105, + "learning_rate": 9.90567458416045e-05, + "loss": 1.2216, + "step": 1440 + }, + { + "epoch": 0.09763534114777424, + "grad_norm": 10.062298774719238, + "learning_rate": 9.905537682250668e-05, + "loss": 1.1576, + "step": 1441 + }, + { + "epoch": 0.09770309641574633, + "grad_norm": 10.897071838378906, + "learning_rate": 9.905400780340887e-05, + "loss": 1.1315, + "step": 1442 + }, + { + "epoch": 0.09777085168371841, + "grad_norm": 10.366122245788574, + "learning_rate": 9.905263878431105e-05, + "loss": 1.274, + "step": 1443 + }, + { + "epoch": 0.0978386069516905, + "grad_norm": 11.632966995239258, + "learning_rate": 9.905126976521323e-05, + "loss": 1.1427, + "step": 1444 + }, + { + "epoch": 0.09790636221966258, + "grad_norm": 10.537737846374512, + "learning_rate": 9.904990074611541e-05, + "loss": 1.0913, + "step": 1445 + }, + { + "epoch": 0.09797411748763467, + "grad_norm": 9.52363109588623, + "learning_rate": 9.90485317270176e-05, + "loss": 0.8677, + "step": 1446 + }, + { + "epoch": 0.09804187275560675, + "grad_norm": 11.511491775512695, + "learning_rate": 9.904716270791979e-05, + "loss": 1.0381, + "step": 1447 + }, + { + "epoch": 0.09810962802357884, + "grad_norm": 12.085793495178223, + "learning_rate": 9.904579368882197e-05, + "loss": 1.1415, + "step": 1448 + }, + { + "epoch": 0.09817738329155092, + "grad_norm": 8.665430068969727, + "learning_rate": 9.904442466972415e-05, + "loss": 1.0367, + "step": 1449 + }, + { + "epoch": 0.09824513855952301, + "grad_norm": 10.900618553161621, + "learning_rate": 9.904305565062633e-05, + "loss": 0.9835, + "step": 1450 + }, + { + "epoch": 0.09831289382749508, + "grad_norm": 10.3113431930542, + "learning_rate": 9.904168663152852e-05, + "loss": 0.924, + "step": 1451 + }, + { + "epoch": 0.09838064909546718, + "grad_norm": 10.001591682434082, + "learning_rate": 9.90403176124307e-05, + "loss": 1.0835, + "step": 1452 + }, + { + "epoch": 0.09844840436343925, + "grad_norm": 11.333273887634277, + "learning_rate": 9.903894859333288e-05, + "loss": 1.072, + "step": 1453 + }, + { + "epoch": 0.09851615963141135, + "grad_norm": 10.107904434204102, + "learning_rate": 9.903757957423506e-05, + "loss": 1.0848, + "step": 1454 + }, + { + "epoch": 0.09858391489938342, + "grad_norm": 12.578730583190918, + "learning_rate": 9.903621055513725e-05, + "loss": 1.2735, + "step": 1455 + }, + { + "epoch": 0.09865167016735552, + "grad_norm": 10.453478813171387, + "learning_rate": 9.903484153603944e-05, + "loss": 1.3141, + "step": 1456 + }, + { + "epoch": 0.09871942543532759, + "grad_norm": 10.383566856384277, + "learning_rate": 9.903347251694162e-05, + "loss": 1.0992, + "step": 1457 + }, + { + "epoch": 0.09878718070329968, + "grad_norm": 9.612902641296387, + "learning_rate": 9.90321034978438e-05, + "loss": 1.3103, + "step": 1458 + }, + { + "epoch": 0.09885493597127176, + "grad_norm": 12.111359596252441, + "learning_rate": 9.903073447874598e-05, + "loss": 1.3102, + "step": 1459 + }, + { + "epoch": 0.09892269123924385, + "grad_norm": 9.987195014953613, + "learning_rate": 9.902936545964817e-05, + "loss": 1.1961, + "step": 1460 + }, + { + "epoch": 0.09899044650721593, + "grad_norm": 10.900408744812012, + "learning_rate": 9.902799644055035e-05, + "loss": 1.0208, + "step": 1461 + }, + { + "epoch": 0.09905820177518802, + "grad_norm": 9.94915771484375, + "learning_rate": 9.902662742145253e-05, + "loss": 1.3347, + "step": 1462 + }, + { + "epoch": 0.0991259570431601, + "grad_norm": 13.393661499023438, + "learning_rate": 9.902525840235471e-05, + "loss": 1.2903, + "step": 1463 + }, + { + "epoch": 0.09919371231113219, + "grad_norm": 10.122967720031738, + "learning_rate": 9.902388938325689e-05, + "loss": 1.0229, + "step": 1464 + }, + { + "epoch": 0.09926146757910427, + "grad_norm": 10.775031089782715, + "learning_rate": 9.902252036415909e-05, + "loss": 1.1669, + "step": 1465 + }, + { + "epoch": 0.09932922284707636, + "grad_norm": 9.733497619628906, + "learning_rate": 9.902115134506127e-05, + "loss": 1.0903, + "step": 1466 + }, + { + "epoch": 0.09939697811504844, + "grad_norm": 9.230277061462402, + "learning_rate": 9.901978232596345e-05, + "loss": 1.1738, + "step": 1467 + }, + { + "epoch": 0.09946473338302053, + "grad_norm": 10.822884559631348, + "learning_rate": 9.901841330686563e-05, + "loss": 1.3392, + "step": 1468 + }, + { + "epoch": 0.09953248865099261, + "grad_norm": 10.64195442199707, + "learning_rate": 9.901704428776782e-05, + "loss": 1.0823, + "step": 1469 + }, + { + "epoch": 0.0996002439189647, + "grad_norm": 13.73645305633545, + "learning_rate": 9.901567526867e-05, + "loss": 1.1128, + "step": 1470 + }, + { + "epoch": 0.09966799918693678, + "grad_norm": 11.361958503723145, + "learning_rate": 9.901430624957218e-05, + "loss": 1.083, + "step": 1471 + }, + { + "epoch": 0.09973575445490887, + "grad_norm": 10.839045524597168, + "learning_rate": 9.901293723047436e-05, + "loss": 1.0288, + "step": 1472 + }, + { + "epoch": 0.09980350972288095, + "grad_norm": 10.41995906829834, + "learning_rate": 9.901156821137656e-05, + "loss": 1.0631, + "step": 1473 + }, + { + "epoch": 0.09987126499085304, + "grad_norm": 11.87709903717041, + "learning_rate": 9.901019919227874e-05, + "loss": 1.0715, + "step": 1474 + }, + { + "epoch": 0.09993902025882512, + "grad_norm": 10.46670913696289, + "learning_rate": 9.900883017318092e-05, + "loss": 1.1684, + "step": 1475 + }, + { + "epoch": 0.10000677552679721, + "grad_norm": 12.163457870483398, + "learning_rate": 9.900746115408311e-05, + "loss": 1.4416, + "step": 1476 + }, + { + "epoch": 0.10007453079476929, + "grad_norm": 13.417581558227539, + "learning_rate": 9.900609213498529e-05, + "loss": 1.1876, + "step": 1477 + }, + { + "epoch": 0.10014228606274138, + "grad_norm": 11.35722541809082, + "learning_rate": 9.900472311588747e-05, + "loss": 1.1389, + "step": 1478 + }, + { + "epoch": 0.10021004133071346, + "grad_norm": 10.042820930480957, + "learning_rate": 9.900335409678966e-05, + "loss": 1.2156, + "step": 1479 + }, + { + "epoch": 0.10027779659868555, + "grad_norm": 10.823782920837402, + "learning_rate": 9.900198507769184e-05, + "loss": 1.1915, + "step": 1480 + }, + { + "epoch": 0.10034555186665763, + "grad_norm": 13.6808443069458, + "learning_rate": 9.900061605859403e-05, + "loss": 1.1724, + "step": 1481 + }, + { + "epoch": 0.10041330713462972, + "grad_norm": 11.16846752166748, + "learning_rate": 9.89992470394962e-05, + "loss": 1.2093, + "step": 1482 + }, + { + "epoch": 0.10048106240260181, + "grad_norm": 10.391450881958008, + "learning_rate": 9.89978780203984e-05, + "loss": 1.142, + "step": 1483 + }, + { + "epoch": 0.10054881767057389, + "grad_norm": 9.324288368225098, + "learning_rate": 9.899650900130058e-05, + "loss": 1.0579, + "step": 1484 + }, + { + "epoch": 0.10061657293854598, + "grad_norm": 12.601625442504883, + "learning_rate": 9.899513998220276e-05, + "loss": 1.2643, + "step": 1485 + }, + { + "epoch": 0.10068432820651806, + "grad_norm": 14.270779609680176, + "learning_rate": 9.899377096310494e-05, + "loss": 1.2903, + "step": 1486 + }, + { + "epoch": 0.10075208347449015, + "grad_norm": 11.521232604980469, + "learning_rate": 9.899240194400712e-05, + "loss": 1.2385, + "step": 1487 + }, + { + "epoch": 0.10081983874246223, + "grad_norm": 10.76693344116211, + "learning_rate": 9.899103292490931e-05, + "loss": 1.1745, + "step": 1488 + }, + { + "epoch": 0.10088759401043432, + "grad_norm": 9.091184616088867, + "learning_rate": 9.89896639058115e-05, + "loss": 0.9571, + "step": 1489 + }, + { + "epoch": 0.1009553492784064, + "grad_norm": 11.930106163024902, + "learning_rate": 9.898829488671368e-05, + "loss": 1.1226, + "step": 1490 + }, + { + "epoch": 0.10102310454637849, + "grad_norm": 10.90937614440918, + "learning_rate": 9.898692586761586e-05, + "loss": 1.0776, + "step": 1491 + }, + { + "epoch": 0.10109085981435056, + "grad_norm": 10.618545532226562, + "learning_rate": 9.898555684851805e-05, + "loss": 1.1251, + "step": 1492 + }, + { + "epoch": 0.10115861508232266, + "grad_norm": 10.228861808776855, + "learning_rate": 9.898418782942023e-05, + "loss": 1.1987, + "step": 1493 + }, + { + "epoch": 0.10122637035029473, + "grad_norm": 8.807862281799316, + "learning_rate": 9.898281881032241e-05, + "loss": 0.9339, + "step": 1494 + }, + { + "epoch": 0.10129412561826683, + "grad_norm": 11.24593448638916, + "learning_rate": 9.898144979122459e-05, + "loss": 1.1495, + "step": 1495 + }, + { + "epoch": 0.1013618808862389, + "grad_norm": 11.192438125610352, + "learning_rate": 9.898008077212677e-05, + "loss": 1.1361, + "step": 1496 + }, + { + "epoch": 0.101429636154211, + "grad_norm": 10.440075874328613, + "learning_rate": 9.897871175302896e-05, + "loss": 1.2273, + "step": 1497 + }, + { + "epoch": 0.10149739142218307, + "grad_norm": 11.103675842285156, + "learning_rate": 9.897734273393115e-05, + "loss": 1.1308, + "step": 1498 + }, + { + "epoch": 0.10156514669015516, + "grad_norm": 9.78297233581543, + "learning_rate": 9.897597371483333e-05, + "loss": 1.0907, + "step": 1499 + }, + { + "epoch": 0.10163290195812724, + "grad_norm": 10.98086166381836, + "learning_rate": 9.89746046957355e-05, + "loss": 0.9743, + "step": 1500 + }, + { + "epoch": 0.10170065722609933, + "grad_norm": 9.268783569335938, + "learning_rate": 9.89732356766377e-05, + "loss": 0.8917, + "step": 1501 + }, + { + "epoch": 0.10176841249407141, + "grad_norm": 12.674605369567871, + "learning_rate": 9.897186665753988e-05, + "loss": 1.2247, + "step": 1502 + }, + { + "epoch": 0.1018361677620435, + "grad_norm": 10.987565040588379, + "learning_rate": 9.897049763844206e-05, + "loss": 1.0688, + "step": 1503 + }, + { + "epoch": 0.10190392303001558, + "grad_norm": 16.014053344726562, + "learning_rate": 9.896912861934424e-05, + "loss": 0.8334, + "step": 1504 + }, + { + "epoch": 0.10197167829798767, + "grad_norm": 11.119991302490234, + "learning_rate": 9.896775960024642e-05, + "loss": 1.3176, + "step": 1505 + }, + { + "epoch": 0.10203943356595975, + "grad_norm": 13.23279094696045, + "learning_rate": 9.896639058114861e-05, + "loss": 1.1625, + "step": 1506 + }, + { + "epoch": 0.10210718883393184, + "grad_norm": 9.3678560256958, + "learning_rate": 9.89650215620508e-05, + "loss": 1.2774, + "step": 1507 + }, + { + "epoch": 0.10217494410190392, + "grad_norm": 10.829100608825684, + "learning_rate": 9.896365254295298e-05, + "loss": 1.2423, + "step": 1508 + }, + { + "epoch": 0.10224269936987601, + "grad_norm": 12.12694263458252, + "learning_rate": 9.896228352385516e-05, + "loss": 1.2276, + "step": 1509 + }, + { + "epoch": 0.10231045463784809, + "grad_norm": 11.626548767089844, + "learning_rate": 9.896091450475735e-05, + "loss": 1.0871, + "step": 1510 + }, + { + "epoch": 0.10237820990582018, + "grad_norm": 11.388608932495117, + "learning_rate": 9.895954548565953e-05, + "loss": 1.5582, + "step": 1511 + }, + { + "epoch": 0.10244596517379226, + "grad_norm": 9.463730812072754, + "learning_rate": 9.895817646656171e-05, + "loss": 1.1066, + "step": 1512 + }, + { + "epoch": 0.10251372044176435, + "grad_norm": 10.291573524475098, + "learning_rate": 9.895680744746389e-05, + "loss": 1.4026, + "step": 1513 + }, + { + "epoch": 0.10258147570973643, + "grad_norm": 9.778963088989258, + "learning_rate": 9.895543842836607e-05, + "loss": 1.1109, + "step": 1514 + }, + { + "epoch": 0.10264923097770852, + "grad_norm": 9.685966491699219, + "learning_rate": 9.895406940926827e-05, + "loss": 1.1633, + "step": 1515 + }, + { + "epoch": 0.1027169862456806, + "grad_norm": 10.76310920715332, + "learning_rate": 9.895270039017045e-05, + "loss": 1.0813, + "step": 1516 + }, + { + "epoch": 0.10278474151365269, + "grad_norm": 9.795347213745117, + "learning_rate": 9.895133137107263e-05, + "loss": 1.2079, + "step": 1517 + }, + { + "epoch": 0.10285249678162477, + "grad_norm": 9.980990409851074, + "learning_rate": 9.89499623519748e-05, + "loss": 1.0701, + "step": 1518 + }, + { + "epoch": 0.10292025204959686, + "grad_norm": 9.682209014892578, + "learning_rate": 9.894859333287699e-05, + "loss": 1.4133, + "step": 1519 + }, + { + "epoch": 0.10298800731756894, + "grad_norm": 10.632065773010254, + "learning_rate": 9.894722431377918e-05, + "loss": 1.0631, + "step": 1520 + }, + { + "epoch": 0.10305576258554103, + "grad_norm": 10.099474906921387, + "learning_rate": 9.894585529468136e-05, + "loss": 1.2015, + "step": 1521 + }, + { + "epoch": 0.1031235178535131, + "grad_norm": 8.289199829101562, + "learning_rate": 9.894448627558354e-05, + "loss": 1.3009, + "step": 1522 + }, + { + "epoch": 0.1031912731214852, + "grad_norm": 9.403796195983887, + "learning_rate": 9.894311725648573e-05, + "loss": 1.113, + "step": 1523 + }, + { + "epoch": 0.10325902838945727, + "grad_norm": 12.6613130569458, + "learning_rate": 9.894174823738792e-05, + "loss": 0.9394, + "step": 1524 + }, + { + "epoch": 0.10332678365742937, + "grad_norm": 9.85255241394043, + "learning_rate": 9.894037921829011e-05, + "loss": 1.1007, + "step": 1525 + }, + { + "epoch": 0.10339453892540144, + "grad_norm": 11.918173789978027, + "learning_rate": 9.893901019919229e-05, + "loss": 1.1817, + "step": 1526 + }, + { + "epoch": 0.10346229419337354, + "grad_norm": 9.994447708129883, + "learning_rate": 9.893764118009447e-05, + "loss": 1.2522, + "step": 1527 + }, + { + "epoch": 0.10353004946134561, + "grad_norm": 9.879289627075195, + "learning_rate": 9.893627216099665e-05, + "loss": 1.152, + "step": 1528 + }, + { + "epoch": 0.1035978047293177, + "grad_norm": 10.103482246398926, + "learning_rate": 9.893490314189884e-05, + "loss": 1.2227, + "step": 1529 + }, + { + "epoch": 0.10366555999728978, + "grad_norm": 11.173476219177246, + "learning_rate": 9.893353412280102e-05, + "loss": 1.2188, + "step": 1530 + }, + { + "epoch": 0.10373331526526187, + "grad_norm": 11.540877342224121, + "learning_rate": 9.89321651037032e-05, + "loss": 1.1669, + "step": 1531 + }, + { + "epoch": 0.10380107053323395, + "grad_norm": 10.706154823303223, + "learning_rate": 9.893079608460539e-05, + "loss": 1.2771, + "step": 1532 + }, + { + "epoch": 0.10386882580120604, + "grad_norm": 11.781739234924316, + "learning_rate": 9.892942706550758e-05, + "loss": 1.0501, + "step": 1533 + }, + { + "epoch": 0.10393658106917812, + "grad_norm": 10.482099533081055, + "learning_rate": 9.892805804640976e-05, + "loss": 1.2327, + "step": 1534 + }, + { + "epoch": 0.10400433633715021, + "grad_norm": 12.450867652893066, + "learning_rate": 9.892668902731194e-05, + "loss": 1.3575, + "step": 1535 + }, + { + "epoch": 0.1040720916051223, + "grad_norm": 10.236811637878418, + "learning_rate": 9.892532000821412e-05, + "loss": 1.2559, + "step": 1536 + }, + { + "epoch": 0.10413984687309438, + "grad_norm": 9.311124801635742, + "learning_rate": 9.89239509891163e-05, + "loss": 1.1143, + "step": 1537 + }, + { + "epoch": 0.10420760214106647, + "grad_norm": 9.182706832885742, + "learning_rate": 9.89225819700185e-05, + "loss": 0.8875, + "step": 1538 + }, + { + "epoch": 0.10427535740903855, + "grad_norm": 12.762700080871582, + "learning_rate": 9.892121295092067e-05, + "loss": 1.0552, + "step": 1539 + }, + { + "epoch": 0.10434311267701064, + "grad_norm": 12.222203254699707, + "learning_rate": 9.891984393182285e-05, + "loss": 1.3185, + "step": 1540 + }, + { + "epoch": 0.10441086794498272, + "grad_norm": 11.45807933807373, + "learning_rate": 9.891847491272504e-05, + "loss": 1.2952, + "step": 1541 + }, + { + "epoch": 0.10447862321295481, + "grad_norm": 10.445068359375, + "learning_rate": 9.891710589362722e-05, + "loss": 1.1637, + "step": 1542 + }, + { + "epoch": 0.10454637848092689, + "grad_norm": 11.758063316345215, + "learning_rate": 9.891573687452941e-05, + "loss": 1.3202, + "step": 1543 + }, + { + "epoch": 0.10461413374889898, + "grad_norm": 10.176533699035645, + "learning_rate": 9.891436785543159e-05, + "loss": 1.1762, + "step": 1544 + }, + { + "epoch": 0.10468188901687106, + "grad_norm": 10.133155822753906, + "learning_rate": 9.891299883633377e-05, + "loss": 1.2226, + "step": 1545 + }, + { + "epoch": 0.10474964428484315, + "grad_norm": 9.883895874023438, + "learning_rate": 9.891162981723595e-05, + "loss": 0.9687, + "step": 1546 + }, + { + "epoch": 0.10481739955281523, + "grad_norm": 13.175050735473633, + "learning_rate": 9.891026079813814e-05, + "loss": 1.2075, + "step": 1547 + }, + { + "epoch": 0.10488515482078732, + "grad_norm": 11.183597564697266, + "learning_rate": 9.890889177904032e-05, + "loss": 1.0917, + "step": 1548 + }, + { + "epoch": 0.1049529100887594, + "grad_norm": 10.135035514831543, + "learning_rate": 9.89075227599425e-05, + "loss": 1.0087, + "step": 1549 + }, + { + "epoch": 0.10502066535673149, + "grad_norm": 16.02760887145996, + "learning_rate": 9.890615374084469e-05, + "loss": 1.2015, + "step": 1550 + }, + { + "epoch": 0.10508842062470357, + "grad_norm": 11.255363464355469, + "learning_rate": 9.890478472174687e-05, + "loss": 1.1106, + "step": 1551 + }, + { + "epoch": 0.10515617589267566, + "grad_norm": 10.740998268127441, + "learning_rate": 9.890341570264906e-05, + "loss": 1.1835, + "step": 1552 + }, + { + "epoch": 0.10522393116064774, + "grad_norm": 11.84919548034668, + "learning_rate": 9.890204668355124e-05, + "loss": 1.1655, + "step": 1553 + }, + { + "epoch": 0.10529168642861983, + "grad_norm": 12.041108131408691, + "learning_rate": 9.890067766445342e-05, + "loss": 1.2544, + "step": 1554 + }, + { + "epoch": 0.10535944169659191, + "grad_norm": 8.966646194458008, + "learning_rate": 9.88993086453556e-05, + "loss": 1.098, + "step": 1555 + }, + { + "epoch": 0.105427196964564, + "grad_norm": 11.838338851928711, + "learning_rate": 9.88979396262578e-05, + "loss": 1.1774, + "step": 1556 + }, + { + "epoch": 0.10549495223253608, + "grad_norm": 11.186326026916504, + "learning_rate": 9.889657060715997e-05, + "loss": 1.1778, + "step": 1557 + }, + { + "epoch": 0.10556270750050817, + "grad_norm": 9.448702812194824, + "learning_rate": 9.889520158806216e-05, + "loss": 1.0171, + "step": 1558 + }, + { + "epoch": 0.10563046276848025, + "grad_norm": 10.931096076965332, + "learning_rate": 9.889383256896434e-05, + "loss": 1.1797, + "step": 1559 + }, + { + "epoch": 0.10569821803645234, + "grad_norm": 10.293981552124023, + "learning_rate": 9.889246354986652e-05, + "loss": 1.0723, + "step": 1560 + }, + { + "epoch": 0.10576597330442442, + "grad_norm": 10.052331924438477, + "learning_rate": 9.889109453076871e-05, + "loss": 0.831, + "step": 1561 + }, + { + "epoch": 0.10583372857239651, + "grad_norm": 10.593210220336914, + "learning_rate": 9.888972551167089e-05, + "loss": 1.3415, + "step": 1562 + }, + { + "epoch": 0.10590148384036858, + "grad_norm": 11.580954551696777, + "learning_rate": 9.888835649257307e-05, + "loss": 1.1522, + "step": 1563 + }, + { + "epoch": 0.10596923910834068, + "grad_norm": 11.495551109313965, + "learning_rate": 9.888698747347525e-05, + "loss": 1.2515, + "step": 1564 + }, + { + "epoch": 0.10603699437631275, + "grad_norm": 10.543874740600586, + "learning_rate": 9.888561845437743e-05, + "loss": 1.1579, + "step": 1565 + }, + { + "epoch": 0.10610474964428485, + "grad_norm": 10.588164329528809, + "learning_rate": 9.888424943527963e-05, + "loss": 1.3825, + "step": 1566 + }, + { + "epoch": 0.10617250491225692, + "grad_norm": 13.591666221618652, + "learning_rate": 9.88828804161818e-05, + "loss": 1.0941, + "step": 1567 + }, + { + "epoch": 0.10624026018022902, + "grad_norm": 10.866951942443848, + "learning_rate": 9.888151139708399e-05, + "loss": 1.478, + "step": 1568 + }, + { + "epoch": 0.10630801544820109, + "grad_norm": 9.627554893493652, + "learning_rate": 9.888014237798618e-05, + "loss": 1.004, + "step": 1569 + }, + { + "epoch": 0.10637577071617318, + "grad_norm": 10.871118545532227, + "learning_rate": 9.887877335888836e-05, + "loss": 1.1997, + "step": 1570 + }, + { + "epoch": 0.10644352598414526, + "grad_norm": 10.74503231048584, + "learning_rate": 9.887740433979054e-05, + "loss": 1.005, + "step": 1571 + }, + { + "epoch": 0.10651128125211735, + "grad_norm": 11.119452476501465, + "learning_rate": 9.887603532069273e-05, + "loss": 1.1168, + "step": 1572 + }, + { + "epoch": 0.10657903652008943, + "grad_norm": 10.601544380187988, + "learning_rate": 9.887466630159491e-05, + "loss": 1.0256, + "step": 1573 + }, + { + "epoch": 0.10664679178806152, + "grad_norm": 10.329113006591797, + "learning_rate": 9.88732972824971e-05, + "loss": 1.3432, + "step": 1574 + }, + { + "epoch": 0.1067145470560336, + "grad_norm": 9.973999977111816, + "learning_rate": 9.887192826339929e-05, + "loss": 0.9961, + "step": 1575 + }, + { + "epoch": 0.10678230232400569, + "grad_norm": 10.981974601745605, + "learning_rate": 9.887055924430147e-05, + "loss": 1.3936, + "step": 1576 + }, + { + "epoch": 0.10685005759197777, + "grad_norm": 10.953417778015137, + "learning_rate": 9.886919022520365e-05, + "loss": 0.9669, + "step": 1577 + }, + { + "epoch": 0.10691781285994986, + "grad_norm": 10.656323432922363, + "learning_rate": 9.886782120610583e-05, + "loss": 1.1393, + "step": 1578 + }, + { + "epoch": 0.10698556812792194, + "grad_norm": 9.434617042541504, + "learning_rate": 9.886645218700802e-05, + "loss": 1.1813, + "step": 1579 + }, + { + "epoch": 0.10705332339589403, + "grad_norm": 12.334831237792969, + "learning_rate": 9.88650831679102e-05, + "loss": 1.453, + "step": 1580 + }, + { + "epoch": 0.10712107866386611, + "grad_norm": 9.892403602600098, + "learning_rate": 9.886371414881238e-05, + "loss": 1.1373, + "step": 1581 + }, + { + "epoch": 0.1071888339318382, + "grad_norm": 9.248678207397461, + "learning_rate": 9.886234512971456e-05, + "loss": 1.0734, + "step": 1582 + }, + { + "epoch": 0.10725658919981028, + "grad_norm": 10.317010879516602, + "learning_rate": 9.886097611061675e-05, + "loss": 1.1252, + "step": 1583 + }, + { + "epoch": 0.10732434446778237, + "grad_norm": 9.586435317993164, + "learning_rate": 9.885960709151894e-05, + "loss": 1.1705, + "step": 1584 + }, + { + "epoch": 0.10739209973575445, + "grad_norm": 8.883166313171387, + "learning_rate": 9.885823807242112e-05, + "loss": 1.3164, + "step": 1585 + }, + { + "epoch": 0.10745985500372654, + "grad_norm": 12.870014190673828, + "learning_rate": 9.88568690533233e-05, + "loss": 1.2187, + "step": 1586 + }, + { + "epoch": 0.10752761027169863, + "grad_norm": 10.858057975769043, + "learning_rate": 9.885550003422548e-05, + "loss": 1.2334, + "step": 1587 + }, + { + "epoch": 0.10759536553967071, + "grad_norm": 9.855050086975098, + "learning_rate": 9.885413101512767e-05, + "loss": 1.068, + "step": 1588 + }, + { + "epoch": 0.1076631208076428, + "grad_norm": 12.256099700927734, + "learning_rate": 9.885276199602985e-05, + "loss": 1.2481, + "step": 1589 + }, + { + "epoch": 0.10773087607561488, + "grad_norm": 9.756118774414062, + "learning_rate": 9.885139297693203e-05, + "loss": 1.2365, + "step": 1590 + }, + { + "epoch": 0.10779863134358697, + "grad_norm": 11.727996826171875, + "learning_rate": 9.885002395783421e-05, + "loss": 1.3839, + "step": 1591 + }, + { + "epoch": 0.10786638661155905, + "grad_norm": 11.58463191986084, + "learning_rate": 9.88486549387364e-05, + "loss": 0.937, + "step": 1592 + }, + { + "epoch": 0.10793414187953114, + "grad_norm": 12.663015365600586, + "learning_rate": 9.884728591963859e-05, + "loss": 1.3453, + "step": 1593 + }, + { + "epoch": 0.10800189714750322, + "grad_norm": 10.950531005859375, + "learning_rate": 9.884591690054077e-05, + "loss": 1.1345, + "step": 1594 + }, + { + "epoch": 0.10806965241547531, + "grad_norm": 10.361424446105957, + "learning_rate": 9.884454788144295e-05, + "loss": 1.2576, + "step": 1595 + }, + { + "epoch": 0.10813740768344739, + "grad_norm": 11.155044555664062, + "learning_rate": 9.884317886234513e-05, + "loss": 1.1253, + "step": 1596 + }, + { + "epoch": 0.10820516295141948, + "grad_norm": 12.457600593566895, + "learning_rate": 9.884180984324731e-05, + "loss": 1.0959, + "step": 1597 + }, + { + "epoch": 0.10827291821939156, + "grad_norm": 13.494983673095703, + "learning_rate": 9.88404408241495e-05, + "loss": 1.1671, + "step": 1598 + }, + { + "epoch": 0.10834067348736365, + "grad_norm": 11.870648384094238, + "learning_rate": 9.883907180505168e-05, + "loss": 1.2121, + "step": 1599 + }, + { + "epoch": 0.10840842875533573, + "grad_norm": 10.327920913696289, + "learning_rate": 9.883770278595387e-05, + "loss": 1.1319, + "step": 1600 + }, + { + "epoch": 0.10847618402330782, + "grad_norm": 9.578879356384277, + "learning_rate": 9.883633376685605e-05, + "loss": 0.9064, + "step": 1601 + }, + { + "epoch": 0.1085439392912799, + "grad_norm": 9.598779678344727, + "learning_rate": 9.883496474775824e-05, + "loss": 1.1119, + "step": 1602 + }, + { + "epoch": 0.10861169455925199, + "grad_norm": 13.522628784179688, + "learning_rate": 9.883359572866042e-05, + "loss": 1.2549, + "step": 1603 + }, + { + "epoch": 0.10867944982722406, + "grad_norm": 8.123127937316895, + "learning_rate": 9.88322267095626e-05, + "loss": 0.8778, + "step": 1604 + }, + { + "epoch": 0.10874720509519616, + "grad_norm": 13.100455284118652, + "learning_rate": 9.883085769046478e-05, + "loss": 1.2023, + "step": 1605 + }, + { + "epoch": 0.10881496036316823, + "grad_norm": 9.897802352905273, + "learning_rate": 9.882948867136696e-05, + "loss": 1.2094, + "step": 1606 + }, + { + "epoch": 0.10888271563114033, + "grad_norm": 8.55170726776123, + "learning_rate": 9.882811965226915e-05, + "loss": 1.0407, + "step": 1607 + }, + { + "epoch": 0.1089504708991124, + "grad_norm": 10.867438316345215, + "learning_rate": 9.882675063317133e-05, + "loss": 1.3021, + "step": 1608 + }, + { + "epoch": 0.1090182261670845, + "grad_norm": 11.062238693237305, + "learning_rate": 9.882538161407352e-05, + "loss": 1.2627, + "step": 1609 + }, + { + "epoch": 0.10908598143505657, + "grad_norm": 12.710458755493164, + "learning_rate": 9.88240125949757e-05, + "loss": 1.6196, + "step": 1610 + }, + { + "epoch": 0.10915373670302866, + "grad_norm": 10.986543655395508, + "learning_rate": 9.882264357587789e-05, + "loss": 1.2021, + "step": 1611 + }, + { + "epoch": 0.10922149197100074, + "grad_norm": 12.792850494384766, + "learning_rate": 9.882127455678007e-05, + "loss": 1.1246, + "step": 1612 + }, + { + "epoch": 0.10928924723897283, + "grad_norm": 12.120078086853027, + "learning_rate": 9.881990553768225e-05, + "loss": 1.2522, + "step": 1613 + }, + { + "epoch": 0.10935700250694491, + "grad_norm": 10.562193870544434, + "learning_rate": 9.881853651858443e-05, + "loss": 1.1791, + "step": 1614 + }, + { + "epoch": 0.109424757774917, + "grad_norm": 13.017698287963867, + "learning_rate": 9.881716749948662e-05, + "loss": 1.0733, + "step": 1615 + }, + { + "epoch": 0.10949251304288908, + "grad_norm": 10.904980659484863, + "learning_rate": 9.88157984803888e-05, + "loss": 1.0166, + "step": 1616 + }, + { + "epoch": 0.10956026831086117, + "grad_norm": 12.332829475402832, + "learning_rate": 9.881442946129099e-05, + "loss": 1.0688, + "step": 1617 + }, + { + "epoch": 0.10962802357883325, + "grad_norm": 12.238106727600098, + "learning_rate": 9.881306044219318e-05, + "loss": 0.9869, + "step": 1618 + }, + { + "epoch": 0.10969577884680534, + "grad_norm": 12.75059986114502, + "learning_rate": 9.881169142309536e-05, + "loss": 1.5365, + "step": 1619 + }, + { + "epoch": 0.10976353411477742, + "grad_norm": 11.019241333007812, + "learning_rate": 9.881032240399754e-05, + "loss": 1.2332, + "step": 1620 + }, + { + "epoch": 0.10983128938274951, + "grad_norm": 11.867498397827148, + "learning_rate": 9.880895338489973e-05, + "loss": 1.1955, + "step": 1621 + }, + { + "epoch": 0.10989904465072159, + "grad_norm": 9.685249328613281, + "learning_rate": 9.880758436580191e-05, + "loss": 1.1972, + "step": 1622 + }, + { + "epoch": 0.10996679991869368, + "grad_norm": 11.80540657043457, + "learning_rate": 9.88062153467041e-05, + "loss": 0.9501, + "step": 1623 + }, + { + "epoch": 0.11003455518666576, + "grad_norm": 9.719080924987793, + "learning_rate": 9.880484632760627e-05, + "loss": 0.9904, + "step": 1624 + }, + { + "epoch": 0.11010231045463785, + "grad_norm": 10.906637191772461, + "learning_rate": 9.880347730850847e-05, + "loss": 1.2807, + "step": 1625 + }, + { + "epoch": 0.11017006572260993, + "grad_norm": 10.40942668914795, + "learning_rate": 9.880210828941065e-05, + "loss": 1.2772, + "step": 1626 + }, + { + "epoch": 0.11023782099058202, + "grad_norm": 10.265239715576172, + "learning_rate": 9.880073927031283e-05, + "loss": 1.1583, + "step": 1627 + }, + { + "epoch": 0.1103055762585541, + "grad_norm": 12.158036231994629, + "learning_rate": 9.879937025121501e-05, + "loss": 1.3907, + "step": 1628 + }, + { + "epoch": 0.11037333152652619, + "grad_norm": 10.128400802612305, + "learning_rate": 9.879800123211719e-05, + "loss": 0.8838, + "step": 1629 + }, + { + "epoch": 0.11044108679449827, + "grad_norm": 9.536952018737793, + "learning_rate": 9.879663221301938e-05, + "loss": 1.0525, + "step": 1630 + }, + { + "epoch": 0.11050884206247036, + "grad_norm": 10.826987266540527, + "learning_rate": 9.879526319392156e-05, + "loss": 1.2005, + "step": 1631 + }, + { + "epoch": 0.11057659733044244, + "grad_norm": 10.08153247833252, + "learning_rate": 9.879389417482374e-05, + "loss": 1.1521, + "step": 1632 + }, + { + "epoch": 0.11064435259841453, + "grad_norm": 10.642118453979492, + "learning_rate": 9.879252515572592e-05, + "loss": 1.2316, + "step": 1633 + }, + { + "epoch": 0.1107121078663866, + "grad_norm": 9.595836639404297, + "learning_rate": 9.879115613662812e-05, + "loss": 1.0969, + "step": 1634 + }, + { + "epoch": 0.1107798631343587, + "grad_norm": 8.954519271850586, + "learning_rate": 9.87897871175303e-05, + "loss": 0.9422, + "step": 1635 + }, + { + "epoch": 0.11084761840233077, + "grad_norm": 9.143898010253906, + "learning_rate": 9.878841809843248e-05, + "loss": 0.8949, + "step": 1636 + }, + { + "epoch": 0.11091537367030287, + "grad_norm": 11.656847953796387, + "learning_rate": 9.878704907933466e-05, + "loss": 0.9739, + "step": 1637 + }, + { + "epoch": 0.11098312893827494, + "grad_norm": 11.996994018554688, + "learning_rate": 9.878568006023684e-05, + "loss": 1.3361, + "step": 1638 + }, + { + "epoch": 0.11105088420624704, + "grad_norm": 11.478599548339844, + "learning_rate": 9.878431104113903e-05, + "loss": 1.0726, + "step": 1639 + }, + { + "epoch": 0.11111863947421913, + "grad_norm": 14.594179153442383, + "learning_rate": 9.878294202204121e-05, + "loss": 1.3725, + "step": 1640 + }, + { + "epoch": 0.1111863947421912, + "grad_norm": 12.15703296661377, + "learning_rate": 9.87815730029434e-05, + "loss": 1.1996, + "step": 1641 + }, + { + "epoch": 0.1112541500101633, + "grad_norm": 9.384215354919434, + "learning_rate": 9.878020398384557e-05, + "loss": 1.1294, + "step": 1642 + }, + { + "epoch": 0.11132190527813537, + "grad_norm": 11.131610870361328, + "learning_rate": 9.877883496474777e-05, + "loss": 1.0526, + "step": 1643 + }, + { + "epoch": 0.11138966054610747, + "grad_norm": 9.628592491149902, + "learning_rate": 9.877746594564995e-05, + "loss": 1.0715, + "step": 1644 + }, + { + "epoch": 0.11145741581407954, + "grad_norm": 10.876824378967285, + "learning_rate": 9.877609692655213e-05, + "loss": 1.3337, + "step": 1645 + }, + { + "epoch": 0.11152517108205164, + "grad_norm": 10.489126205444336, + "learning_rate": 9.877472790745431e-05, + "loss": 1.1727, + "step": 1646 + }, + { + "epoch": 0.11159292635002371, + "grad_norm": 8.737699508666992, + "learning_rate": 9.877335888835649e-05, + "loss": 1.1252, + "step": 1647 + }, + { + "epoch": 0.1116606816179958, + "grad_norm": 9.8843355178833, + "learning_rate": 9.877198986925868e-05, + "loss": 0.9222, + "step": 1648 + }, + { + "epoch": 0.11172843688596788, + "grad_norm": 7.930290699005127, + "learning_rate": 9.877062085016086e-05, + "loss": 1.0095, + "step": 1649 + }, + { + "epoch": 0.11179619215393997, + "grad_norm": 9.721529006958008, + "learning_rate": 9.876925183106304e-05, + "loss": 1.0812, + "step": 1650 + }, + { + "epoch": 0.11186394742191205, + "grad_norm": 10.844324111938477, + "learning_rate": 9.876788281196523e-05, + "loss": 1.2245, + "step": 1651 + }, + { + "epoch": 0.11193170268988414, + "grad_norm": 8.750361442565918, + "learning_rate": 9.87665137928674e-05, + "loss": 0.9983, + "step": 1652 + }, + { + "epoch": 0.11199945795785622, + "grad_norm": 8.716521263122559, + "learning_rate": 9.87651447737696e-05, + "loss": 0.8612, + "step": 1653 + }, + { + "epoch": 0.11206721322582831, + "grad_norm": 10.06696891784668, + "learning_rate": 9.876377575467178e-05, + "loss": 1.0806, + "step": 1654 + }, + { + "epoch": 0.11213496849380039, + "grad_norm": 10.526103019714355, + "learning_rate": 9.876240673557396e-05, + "loss": 1.0799, + "step": 1655 + }, + { + "epoch": 0.11220272376177248, + "grad_norm": 11.108748435974121, + "learning_rate": 9.876103771647614e-05, + "loss": 1.3871, + "step": 1656 + }, + { + "epoch": 0.11227047902974456, + "grad_norm": 10.120033264160156, + "learning_rate": 9.875966869737833e-05, + "loss": 1.0399, + "step": 1657 + }, + { + "epoch": 0.11233823429771665, + "grad_norm": 10.589370727539062, + "learning_rate": 9.875829967828051e-05, + "loss": 1.3249, + "step": 1658 + }, + { + "epoch": 0.11240598956568873, + "grad_norm": 8.751884460449219, + "learning_rate": 9.87569306591827e-05, + "loss": 0.9404, + "step": 1659 + }, + { + "epoch": 0.11247374483366082, + "grad_norm": 9.923823356628418, + "learning_rate": 9.875556164008488e-05, + "loss": 1.067, + "step": 1660 + }, + { + "epoch": 0.1125415001016329, + "grad_norm": 9.70035171508789, + "learning_rate": 9.875419262098707e-05, + "loss": 1.0079, + "step": 1661 + }, + { + "epoch": 0.11260925536960499, + "grad_norm": 12.337713241577148, + "learning_rate": 9.875282360188925e-05, + "loss": 1.0962, + "step": 1662 + }, + { + "epoch": 0.11267701063757707, + "grad_norm": 10.142355918884277, + "learning_rate": 9.875145458279143e-05, + "loss": 1.1916, + "step": 1663 + }, + { + "epoch": 0.11274476590554916, + "grad_norm": 11.82714557647705, + "learning_rate": 9.875008556369362e-05, + "loss": 1.4582, + "step": 1664 + }, + { + "epoch": 0.11281252117352124, + "grad_norm": 10.524566650390625, + "learning_rate": 9.87487165445958e-05, + "loss": 1.2156, + "step": 1665 + }, + { + "epoch": 0.11288027644149333, + "grad_norm": 12.232719421386719, + "learning_rate": 9.874734752549798e-05, + "loss": 1.2152, + "step": 1666 + }, + { + "epoch": 0.11294803170946541, + "grad_norm": 10.665773391723633, + "learning_rate": 9.874597850640018e-05, + "loss": 1.085, + "step": 1667 + }, + { + "epoch": 0.1130157869774375, + "grad_norm": 10.0956449508667, + "learning_rate": 9.874460948730236e-05, + "loss": 1.1142, + "step": 1668 + }, + { + "epoch": 0.11308354224540958, + "grad_norm": 9.27825927734375, + "learning_rate": 9.874324046820454e-05, + "loss": 1.0263, + "step": 1669 + }, + { + "epoch": 0.11315129751338167, + "grad_norm": 11.171673774719238, + "learning_rate": 9.874187144910672e-05, + "loss": 1.5049, + "step": 1670 + }, + { + "epoch": 0.11321905278135375, + "grad_norm": 10.442279815673828, + "learning_rate": 9.874050243000891e-05, + "loss": 1.0697, + "step": 1671 + }, + { + "epoch": 0.11328680804932584, + "grad_norm": 11.563700675964355, + "learning_rate": 9.87391334109111e-05, + "loss": 1.003, + "step": 1672 + }, + { + "epoch": 0.11335456331729792, + "grad_norm": 12.447733879089355, + "learning_rate": 9.873776439181327e-05, + "loss": 1.2279, + "step": 1673 + }, + { + "epoch": 0.11342231858527001, + "grad_norm": 9.70698070526123, + "learning_rate": 9.873639537271545e-05, + "loss": 1.2208, + "step": 1674 + }, + { + "epoch": 0.11349007385324208, + "grad_norm": 10.89858341217041, + "learning_rate": 9.873502635361763e-05, + "loss": 1.1676, + "step": 1675 + }, + { + "epoch": 0.11355782912121418, + "grad_norm": 11.106192588806152, + "learning_rate": 9.873365733451983e-05, + "loss": 1.0918, + "step": 1676 + }, + { + "epoch": 0.11362558438918625, + "grad_norm": 10.186528205871582, + "learning_rate": 9.873228831542201e-05, + "loss": 1.1651, + "step": 1677 + }, + { + "epoch": 0.11369333965715835, + "grad_norm": 10.69977855682373, + "learning_rate": 9.873091929632419e-05, + "loss": 1.2867, + "step": 1678 + }, + { + "epoch": 0.11376109492513042, + "grad_norm": 11.934000015258789, + "learning_rate": 9.872955027722637e-05, + "loss": 1.2051, + "step": 1679 + }, + { + "epoch": 0.11382885019310252, + "grad_norm": 9.277094841003418, + "learning_rate": 9.872818125812856e-05, + "loss": 1.17, + "step": 1680 + }, + { + "epoch": 0.1138966054610746, + "grad_norm": 11.35097885131836, + "learning_rate": 9.872681223903074e-05, + "loss": 1.2433, + "step": 1681 + }, + { + "epoch": 0.11396436072904668, + "grad_norm": 11.362833023071289, + "learning_rate": 9.872544321993292e-05, + "loss": 1.3577, + "step": 1682 + }, + { + "epoch": 0.11403211599701876, + "grad_norm": 10.944365501403809, + "learning_rate": 9.87240742008351e-05, + "loss": 1.0731, + "step": 1683 + }, + { + "epoch": 0.11409987126499085, + "grad_norm": 10.292986869812012, + "learning_rate": 9.872270518173728e-05, + "loss": 1.0639, + "step": 1684 + }, + { + "epoch": 0.11416762653296293, + "grad_norm": 10.8629150390625, + "learning_rate": 9.872133616263948e-05, + "loss": 1.3503, + "step": 1685 + }, + { + "epoch": 0.11423538180093502, + "grad_norm": 10.277920722961426, + "learning_rate": 9.871996714354166e-05, + "loss": 1.1544, + "step": 1686 + }, + { + "epoch": 0.1143031370689071, + "grad_norm": 8.98864459991455, + "learning_rate": 9.871859812444384e-05, + "loss": 1.3638, + "step": 1687 + }, + { + "epoch": 0.11437089233687919, + "grad_norm": 8.126718521118164, + "learning_rate": 9.871722910534602e-05, + "loss": 1.0343, + "step": 1688 + }, + { + "epoch": 0.11443864760485127, + "grad_norm": 9.229771614074707, + "learning_rate": 9.871586008624821e-05, + "loss": 1.0137, + "step": 1689 + }, + { + "epoch": 0.11450640287282336, + "grad_norm": 10.554340362548828, + "learning_rate": 9.87144910671504e-05, + "loss": 0.9908, + "step": 1690 + }, + { + "epoch": 0.11457415814079544, + "grad_norm": 11.857934951782227, + "learning_rate": 9.871312204805257e-05, + "loss": 1.29, + "step": 1691 + }, + { + "epoch": 0.11464191340876753, + "grad_norm": 9.334272384643555, + "learning_rate": 9.871175302895475e-05, + "loss": 1.0449, + "step": 1692 + }, + { + "epoch": 0.11470966867673962, + "grad_norm": 9.092185974121094, + "learning_rate": 9.871038400985693e-05, + "loss": 1.1386, + "step": 1693 + }, + { + "epoch": 0.1147774239447117, + "grad_norm": 9.379465103149414, + "learning_rate": 9.870901499075913e-05, + "loss": 1.0964, + "step": 1694 + }, + { + "epoch": 0.11484517921268379, + "grad_norm": 10.038492202758789, + "learning_rate": 9.870764597166131e-05, + "loss": 0.8512, + "step": 1695 + }, + { + "epoch": 0.11491293448065587, + "grad_norm": 8.2230863571167, + "learning_rate": 9.870627695256349e-05, + "loss": 1.0646, + "step": 1696 + }, + { + "epoch": 0.11498068974862796, + "grad_norm": 9.939510345458984, + "learning_rate": 9.870490793346567e-05, + "loss": 1.081, + "step": 1697 + }, + { + "epoch": 0.11504844501660004, + "grad_norm": 11.864813804626465, + "learning_rate": 9.870353891436786e-05, + "loss": 1.2447, + "step": 1698 + }, + { + "epoch": 0.11511620028457213, + "grad_norm": 11.517714500427246, + "learning_rate": 9.870216989527004e-05, + "loss": 1.1274, + "step": 1699 + }, + { + "epoch": 0.11518395555254421, + "grad_norm": 12.172054290771484, + "learning_rate": 9.870080087617222e-05, + "loss": 1.3278, + "step": 1700 + }, + { + "epoch": 0.1152517108205163, + "grad_norm": 9.915261268615723, + "learning_rate": 9.86994318570744e-05, + "loss": 1.0298, + "step": 1701 + }, + { + "epoch": 0.11531946608848838, + "grad_norm": 10.937396049499512, + "learning_rate": 9.869806283797659e-05, + "loss": 1.363, + "step": 1702 + }, + { + "epoch": 0.11538722135646047, + "grad_norm": 11.586840629577637, + "learning_rate": 9.869669381887878e-05, + "loss": 0.9152, + "step": 1703 + }, + { + "epoch": 0.11545497662443255, + "grad_norm": 10.035614967346191, + "learning_rate": 9.869532479978096e-05, + "loss": 1.2152, + "step": 1704 + }, + { + "epoch": 0.11552273189240464, + "grad_norm": 8.775728225708008, + "learning_rate": 9.869395578068314e-05, + "loss": 1.0552, + "step": 1705 + }, + { + "epoch": 0.11559048716037672, + "grad_norm": 10.85958480834961, + "learning_rate": 9.869258676158532e-05, + "loss": 1.0473, + "step": 1706 + }, + { + "epoch": 0.11565824242834881, + "grad_norm": 9.311616897583008, + "learning_rate": 9.869121774248751e-05, + "loss": 1.1704, + "step": 1707 + }, + { + "epoch": 0.11572599769632089, + "grad_norm": 9.503780364990234, + "learning_rate": 9.86898487233897e-05, + "loss": 1.1375, + "step": 1708 + }, + { + "epoch": 0.11579375296429298, + "grad_norm": 9.02602481842041, + "learning_rate": 9.868847970429187e-05, + "loss": 0.9955, + "step": 1709 + }, + { + "epoch": 0.11586150823226506, + "grad_norm": 11.473995208740234, + "learning_rate": 9.868711068519407e-05, + "loss": 0.8842, + "step": 1710 + }, + { + "epoch": 0.11592926350023715, + "grad_norm": 10.124822616577148, + "learning_rate": 9.868574166609625e-05, + "loss": 1.0458, + "step": 1711 + }, + { + "epoch": 0.11599701876820923, + "grad_norm": 11.004744529724121, + "learning_rate": 9.868437264699843e-05, + "loss": 1.2136, + "step": 1712 + }, + { + "epoch": 0.11606477403618132, + "grad_norm": 10.497210502624512, + "learning_rate": 9.868300362790062e-05, + "loss": 1.3457, + "step": 1713 + }, + { + "epoch": 0.1161325293041534, + "grad_norm": 10.933736801147461, + "learning_rate": 9.86816346088028e-05, + "loss": 1.0931, + "step": 1714 + }, + { + "epoch": 0.11620028457212549, + "grad_norm": 11.045526504516602, + "learning_rate": 9.868026558970498e-05, + "loss": 1.2338, + "step": 1715 + }, + { + "epoch": 0.11626803984009756, + "grad_norm": 10.763197898864746, + "learning_rate": 9.867889657060716e-05, + "loss": 1.2325, + "step": 1716 + }, + { + "epoch": 0.11633579510806966, + "grad_norm": 9.53070068359375, + "learning_rate": 9.867752755150936e-05, + "loss": 1.1268, + "step": 1717 + }, + { + "epoch": 0.11640355037604173, + "grad_norm": 10.572071075439453, + "learning_rate": 9.867615853241154e-05, + "loss": 1.0118, + "step": 1718 + }, + { + "epoch": 0.11647130564401383, + "grad_norm": 7.633776664733887, + "learning_rate": 9.867478951331372e-05, + "loss": 1.081, + "step": 1719 + }, + { + "epoch": 0.1165390609119859, + "grad_norm": 8.915566444396973, + "learning_rate": 9.86734204942159e-05, + "loss": 1.1526, + "step": 1720 + }, + { + "epoch": 0.116606816179958, + "grad_norm": 8.296719551086426, + "learning_rate": 9.867205147511809e-05, + "loss": 1.0571, + "step": 1721 + }, + { + "epoch": 0.11667457144793007, + "grad_norm": 9.387847900390625, + "learning_rate": 9.867068245602027e-05, + "loss": 1.0447, + "step": 1722 + }, + { + "epoch": 0.11674232671590216, + "grad_norm": 10.832168579101562, + "learning_rate": 9.866931343692245e-05, + "loss": 1.2601, + "step": 1723 + }, + { + "epoch": 0.11681008198387424, + "grad_norm": 11.631181716918945, + "learning_rate": 9.866794441782463e-05, + "loss": 1.4507, + "step": 1724 + }, + { + "epoch": 0.11687783725184633, + "grad_norm": 8.372645378112793, + "learning_rate": 9.866657539872681e-05, + "loss": 1.1191, + "step": 1725 + }, + { + "epoch": 0.11694559251981841, + "grad_norm": 9.577934265136719, + "learning_rate": 9.866520637962901e-05, + "loss": 1.0336, + "step": 1726 + }, + { + "epoch": 0.1170133477877905, + "grad_norm": 10.626317977905273, + "learning_rate": 9.866383736053119e-05, + "loss": 1.1595, + "step": 1727 + }, + { + "epoch": 0.11708110305576258, + "grad_norm": 9.758353233337402, + "learning_rate": 9.866246834143337e-05, + "loss": 0.8803, + "step": 1728 + }, + { + "epoch": 0.11714885832373467, + "grad_norm": 9.643457412719727, + "learning_rate": 9.866109932233555e-05, + "loss": 1.3213, + "step": 1729 + }, + { + "epoch": 0.11721661359170675, + "grad_norm": 8.910233497619629, + "learning_rate": 9.865973030323773e-05, + "loss": 1.1123, + "step": 1730 + }, + { + "epoch": 0.11728436885967884, + "grad_norm": 8.893741607666016, + "learning_rate": 9.865836128413992e-05, + "loss": 1.1059, + "step": 1731 + }, + { + "epoch": 0.11735212412765092, + "grad_norm": 12.937616348266602, + "learning_rate": 9.86569922650421e-05, + "loss": 1.2488, + "step": 1732 + }, + { + "epoch": 0.11741987939562301, + "grad_norm": 9.696173667907715, + "learning_rate": 9.865562324594428e-05, + "loss": 1.1076, + "step": 1733 + }, + { + "epoch": 0.11748763466359509, + "grad_norm": 11.568793296813965, + "learning_rate": 9.865425422684646e-05, + "loss": 1.1023, + "step": 1734 + }, + { + "epoch": 0.11755538993156718, + "grad_norm": 11.438271522521973, + "learning_rate": 9.865288520774866e-05, + "loss": 1.1014, + "step": 1735 + }, + { + "epoch": 0.11762314519953926, + "grad_norm": 8.217238426208496, + "learning_rate": 9.865151618865084e-05, + "loss": 0.9566, + "step": 1736 + }, + { + "epoch": 0.11769090046751135, + "grad_norm": 9.716591835021973, + "learning_rate": 9.865014716955302e-05, + "loss": 1.0478, + "step": 1737 + }, + { + "epoch": 0.11775865573548343, + "grad_norm": 11.152593612670898, + "learning_rate": 9.86487781504552e-05, + "loss": 1.3552, + "step": 1738 + }, + { + "epoch": 0.11782641100345552, + "grad_norm": 8.824272155761719, + "learning_rate": 9.864740913135738e-05, + "loss": 1.0463, + "step": 1739 + }, + { + "epoch": 0.1178941662714276, + "grad_norm": 8.941548347473145, + "learning_rate": 9.864604011225957e-05, + "loss": 0.9552, + "step": 1740 + }, + { + "epoch": 0.11796192153939969, + "grad_norm": 10.76177978515625, + "learning_rate": 9.864467109316175e-05, + "loss": 1.0076, + "step": 1741 + }, + { + "epoch": 0.11802967680737177, + "grad_norm": 9.87431812286377, + "learning_rate": 9.864330207406393e-05, + "loss": 0.9905, + "step": 1742 + }, + { + "epoch": 0.11809743207534386, + "grad_norm": 12.406320571899414, + "learning_rate": 9.864193305496611e-05, + "loss": 1.2839, + "step": 1743 + }, + { + "epoch": 0.11816518734331594, + "grad_norm": 12.649428367614746, + "learning_rate": 9.864056403586831e-05, + "loss": 1.0806, + "step": 1744 + }, + { + "epoch": 0.11823294261128803, + "grad_norm": 10.888890266418457, + "learning_rate": 9.863919501677049e-05, + "loss": 0.8805, + "step": 1745 + }, + { + "epoch": 0.11830069787926012, + "grad_norm": 10.59262752532959, + "learning_rate": 9.863782599767267e-05, + "loss": 0.9977, + "step": 1746 + }, + { + "epoch": 0.1183684531472322, + "grad_norm": 11.14206314086914, + "learning_rate": 9.863645697857485e-05, + "loss": 1.4569, + "step": 1747 + }, + { + "epoch": 0.11843620841520429, + "grad_norm": 12.148449897766113, + "learning_rate": 9.863508795947703e-05, + "loss": 1.5222, + "step": 1748 + }, + { + "epoch": 0.11850396368317637, + "grad_norm": 11.814643859863281, + "learning_rate": 9.863371894037922e-05, + "loss": 1.3888, + "step": 1749 + }, + { + "epoch": 0.11857171895114846, + "grad_norm": 12.437151908874512, + "learning_rate": 9.86323499212814e-05, + "loss": 1.024, + "step": 1750 + }, + { + "epoch": 0.11863947421912054, + "grad_norm": 9.314751625061035, + "learning_rate": 9.863098090218358e-05, + "loss": 1.064, + "step": 1751 + }, + { + "epoch": 0.11870722948709263, + "grad_norm": 12.978782653808594, + "learning_rate": 9.862961188308576e-05, + "loss": 1.4048, + "step": 1752 + }, + { + "epoch": 0.1187749847550647, + "grad_norm": 9.825428009033203, + "learning_rate": 9.862824286398796e-05, + "loss": 0.9631, + "step": 1753 + }, + { + "epoch": 0.1188427400230368, + "grad_norm": 9.127273559570312, + "learning_rate": 9.862687384489014e-05, + "loss": 1.0364, + "step": 1754 + }, + { + "epoch": 0.11891049529100887, + "grad_norm": 10.290020942687988, + "learning_rate": 9.862550482579232e-05, + "loss": 1.329, + "step": 1755 + }, + { + "epoch": 0.11897825055898097, + "grad_norm": 9.023946762084961, + "learning_rate": 9.862413580669451e-05, + "loss": 1.186, + "step": 1756 + }, + { + "epoch": 0.11904600582695304, + "grad_norm": 11.889911651611328, + "learning_rate": 9.86227667875967e-05, + "loss": 1.0338, + "step": 1757 + }, + { + "epoch": 0.11911376109492514, + "grad_norm": 8.938654899597168, + "learning_rate": 9.862139776849887e-05, + "loss": 1.1933, + "step": 1758 + }, + { + "epoch": 0.11918151636289721, + "grad_norm": 9.240127563476562, + "learning_rate": 9.862002874940107e-05, + "loss": 1.2336, + "step": 1759 + }, + { + "epoch": 0.1192492716308693, + "grad_norm": 10.340953826904297, + "learning_rate": 9.861865973030325e-05, + "loss": 1.1097, + "step": 1760 + }, + { + "epoch": 0.11931702689884138, + "grad_norm": 10.807507514953613, + "learning_rate": 9.861729071120543e-05, + "loss": 0.9127, + "step": 1761 + }, + { + "epoch": 0.11938478216681347, + "grad_norm": 9.670829772949219, + "learning_rate": 9.861592169210761e-05, + "loss": 1.2296, + "step": 1762 + }, + { + "epoch": 0.11945253743478555, + "grad_norm": 10.85981273651123, + "learning_rate": 9.86145526730098e-05, + "loss": 1.1341, + "step": 1763 + }, + { + "epoch": 0.11952029270275764, + "grad_norm": 9.198482513427734, + "learning_rate": 9.861318365391198e-05, + "loss": 1.4397, + "step": 1764 + }, + { + "epoch": 0.11958804797072972, + "grad_norm": 11.130392074584961, + "learning_rate": 9.861181463481416e-05, + "loss": 1.381, + "step": 1765 + }, + { + "epoch": 0.11965580323870181, + "grad_norm": 9.568144798278809, + "learning_rate": 9.861044561571634e-05, + "loss": 1.0284, + "step": 1766 + }, + { + "epoch": 0.11972355850667389, + "grad_norm": 9.049298286437988, + "learning_rate": 9.860907659661854e-05, + "loss": 1.0714, + "step": 1767 + }, + { + "epoch": 0.11979131377464598, + "grad_norm": 8.948347091674805, + "learning_rate": 9.860770757752072e-05, + "loss": 1.0248, + "step": 1768 + }, + { + "epoch": 0.11985906904261806, + "grad_norm": 10.836170196533203, + "learning_rate": 9.86063385584229e-05, + "loss": 1.0551, + "step": 1769 + }, + { + "epoch": 0.11992682431059015, + "grad_norm": 11.044917106628418, + "learning_rate": 9.860496953932508e-05, + "loss": 1.1374, + "step": 1770 + }, + { + "epoch": 0.11999457957856223, + "grad_norm": 10.975854873657227, + "learning_rate": 9.860360052022726e-05, + "loss": 1.2018, + "step": 1771 + }, + { + "epoch": 0.12006233484653432, + "grad_norm": 9.752070426940918, + "learning_rate": 9.860223150112945e-05, + "loss": 1.0843, + "step": 1772 + }, + { + "epoch": 0.1201300901145064, + "grad_norm": 9.662758827209473, + "learning_rate": 9.860086248203163e-05, + "loss": 0.8866, + "step": 1773 + }, + { + "epoch": 0.12019784538247849, + "grad_norm": 11.268694877624512, + "learning_rate": 9.859949346293381e-05, + "loss": 1.2653, + "step": 1774 + }, + { + "epoch": 0.12026560065045057, + "grad_norm": 9.201732635498047, + "learning_rate": 9.8598124443836e-05, + "loss": 1.1466, + "step": 1775 + }, + { + "epoch": 0.12033335591842266, + "grad_norm": 12.577048301696777, + "learning_rate": 9.859675542473819e-05, + "loss": 1.2606, + "step": 1776 + }, + { + "epoch": 0.12040111118639474, + "grad_norm": 10.051258087158203, + "learning_rate": 9.859538640564037e-05, + "loss": 1.1533, + "step": 1777 + }, + { + "epoch": 0.12046886645436683, + "grad_norm": 9.358968734741211, + "learning_rate": 9.859401738654255e-05, + "loss": 1.0501, + "step": 1778 + }, + { + "epoch": 0.12053662172233891, + "grad_norm": 11.277347564697266, + "learning_rate": 9.859264836744473e-05, + "loss": 0.9804, + "step": 1779 + }, + { + "epoch": 0.120604376990311, + "grad_norm": 12.41299057006836, + "learning_rate": 9.859127934834691e-05, + "loss": 1.427, + "step": 1780 + }, + { + "epoch": 0.12067213225828308, + "grad_norm": 10.081144332885742, + "learning_rate": 9.85899103292491e-05, + "loss": 1.1321, + "step": 1781 + }, + { + "epoch": 0.12073988752625517, + "grad_norm": 11.469686508178711, + "learning_rate": 9.858854131015128e-05, + "loss": 1.1972, + "step": 1782 + }, + { + "epoch": 0.12080764279422725, + "grad_norm": 8.608443260192871, + "learning_rate": 9.858717229105346e-05, + "loss": 0.9914, + "step": 1783 + }, + { + "epoch": 0.12087539806219934, + "grad_norm": 11.673405647277832, + "learning_rate": 9.858580327195564e-05, + "loss": 1.3406, + "step": 1784 + }, + { + "epoch": 0.12094315333017142, + "grad_norm": 9.533596992492676, + "learning_rate": 9.858443425285782e-05, + "loss": 1.2119, + "step": 1785 + }, + { + "epoch": 0.12101090859814351, + "grad_norm": 7.792461395263672, + "learning_rate": 9.858306523376002e-05, + "loss": 1.167, + "step": 1786 + }, + { + "epoch": 0.12107866386611559, + "grad_norm": 10.209978103637695, + "learning_rate": 9.85816962146622e-05, + "loss": 1.164, + "step": 1787 + }, + { + "epoch": 0.12114641913408768, + "grad_norm": 11.591109275817871, + "learning_rate": 9.858032719556438e-05, + "loss": 1.2598, + "step": 1788 + }, + { + "epoch": 0.12121417440205975, + "grad_norm": 10.562797546386719, + "learning_rate": 9.857895817646656e-05, + "loss": 1.3071, + "step": 1789 + }, + { + "epoch": 0.12128192967003185, + "grad_norm": 9.419530868530273, + "learning_rate": 9.857758915736875e-05, + "loss": 0.9951, + "step": 1790 + }, + { + "epoch": 0.12134968493800392, + "grad_norm": 9.525064468383789, + "learning_rate": 9.857622013827093e-05, + "loss": 1.1833, + "step": 1791 + }, + { + "epoch": 0.12141744020597602, + "grad_norm": 9.140012741088867, + "learning_rate": 9.857485111917311e-05, + "loss": 1.0573, + "step": 1792 + }, + { + "epoch": 0.1214851954739481, + "grad_norm": 10.88685417175293, + "learning_rate": 9.85734821000753e-05, + "loss": 1.1669, + "step": 1793 + }, + { + "epoch": 0.12155295074192018, + "grad_norm": 8.63829231262207, + "learning_rate": 9.857211308097747e-05, + "loss": 1.1733, + "step": 1794 + }, + { + "epoch": 0.12162070600989226, + "grad_norm": 9.90237808227539, + "learning_rate": 9.857074406187967e-05, + "loss": 0.9175, + "step": 1795 + }, + { + "epoch": 0.12168846127786435, + "grad_norm": 10.068902969360352, + "learning_rate": 9.856937504278185e-05, + "loss": 0.9422, + "step": 1796 + }, + { + "epoch": 0.12175621654583643, + "grad_norm": 7.482577800750732, + "learning_rate": 9.856800602368403e-05, + "loss": 1.0359, + "step": 1797 + }, + { + "epoch": 0.12182397181380852, + "grad_norm": 10.847206115722656, + "learning_rate": 9.856663700458621e-05, + "loss": 1.2258, + "step": 1798 + }, + { + "epoch": 0.12189172708178062, + "grad_norm": 11.122177124023438, + "learning_rate": 9.85652679854884e-05, + "loss": 1.1378, + "step": 1799 + }, + { + "epoch": 0.1219594823497527, + "grad_norm": 9.994139671325684, + "learning_rate": 9.856389896639058e-05, + "loss": 1.2663, + "step": 1800 + }, + { + "epoch": 0.12202723761772478, + "grad_norm": 10.687590599060059, + "learning_rate": 9.856252994729276e-05, + "loss": 1.1777, + "step": 1801 + }, + { + "epoch": 0.12209499288569686, + "grad_norm": 8.403971672058105, + "learning_rate": 9.856116092819494e-05, + "loss": 0.9812, + "step": 1802 + }, + { + "epoch": 0.12216274815366895, + "grad_norm": 9.77443790435791, + "learning_rate": 9.855979190909714e-05, + "loss": 1.0615, + "step": 1803 + }, + { + "epoch": 0.12223050342164103, + "grad_norm": 10.772642135620117, + "learning_rate": 9.855842288999932e-05, + "loss": 1.1835, + "step": 1804 + }, + { + "epoch": 0.12229825868961312, + "grad_norm": 10.509920120239258, + "learning_rate": 9.85570538709015e-05, + "loss": 1.0361, + "step": 1805 + }, + { + "epoch": 0.1223660139575852, + "grad_norm": 11.14194393157959, + "learning_rate": 9.855568485180369e-05, + "loss": 0.9947, + "step": 1806 + }, + { + "epoch": 0.12243376922555729, + "grad_norm": 11.018975257873535, + "learning_rate": 9.855431583270587e-05, + "loss": 1.2036, + "step": 1807 + }, + { + "epoch": 0.12250152449352937, + "grad_norm": 12.099370002746582, + "learning_rate": 9.855294681360805e-05, + "loss": 1.2977, + "step": 1808 + }, + { + "epoch": 0.12256927976150146, + "grad_norm": 8.841024398803711, + "learning_rate": 9.855157779451025e-05, + "loss": 1.2093, + "step": 1809 + }, + { + "epoch": 0.12263703502947354, + "grad_norm": 9.149311065673828, + "learning_rate": 9.855020877541243e-05, + "loss": 1.0088, + "step": 1810 + }, + { + "epoch": 0.12270479029744563, + "grad_norm": 10.53498649597168, + "learning_rate": 9.854883975631461e-05, + "loss": 0.9527, + "step": 1811 + }, + { + "epoch": 0.12277254556541771, + "grad_norm": 9.230859756469727, + "learning_rate": 9.854747073721679e-05, + "loss": 0.9584, + "step": 1812 + }, + { + "epoch": 0.1228403008333898, + "grad_norm": 11.702610969543457, + "learning_rate": 9.854610171811898e-05, + "loss": 1.2369, + "step": 1813 + }, + { + "epoch": 0.12290805610136188, + "grad_norm": 9.912004470825195, + "learning_rate": 9.854473269902116e-05, + "loss": 1.2571, + "step": 1814 + }, + { + "epoch": 0.12297581136933397, + "grad_norm": 10.071510314941406, + "learning_rate": 9.854336367992334e-05, + "loss": 1.1666, + "step": 1815 + }, + { + "epoch": 0.12304356663730605, + "grad_norm": 10.642035484313965, + "learning_rate": 9.854199466082552e-05, + "loss": 1.0309, + "step": 1816 + }, + { + "epoch": 0.12311132190527814, + "grad_norm": 11.135947227478027, + "learning_rate": 9.85406256417277e-05, + "loss": 1.1276, + "step": 1817 + }, + { + "epoch": 0.12317907717325022, + "grad_norm": 11.575469017028809, + "learning_rate": 9.85392566226299e-05, + "loss": 1.1507, + "step": 1818 + }, + { + "epoch": 0.12324683244122231, + "grad_norm": 10.470771789550781, + "learning_rate": 9.853788760353208e-05, + "loss": 1.3777, + "step": 1819 + }, + { + "epoch": 0.12331458770919439, + "grad_norm": 9.77783489227295, + "learning_rate": 9.853651858443426e-05, + "loss": 1.0492, + "step": 1820 + }, + { + "epoch": 0.12338234297716648, + "grad_norm": 10.132309913635254, + "learning_rate": 9.853514956533644e-05, + "loss": 1.1488, + "step": 1821 + }, + { + "epoch": 0.12345009824513856, + "grad_norm": 11.159482955932617, + "learning_rate": 9.853378054623863e-05, + "loss": 1.2571, + "step": 1822 + }, + { + "epoch": 0.12351785351311065, + "grad_norm": 11.599156379699707, + "learning_rate": 9.853241152714081e-05, + "loss": 1.1734, + "step": 1823 + }, + { + "epoch": 0.12358560878108273, + "grad_norm": 10.278475761413574, + "learning_rate": 9.853104250804299e-05, + "loss": 1.0637, + "step": 1824 + }, + { + "epoch": 0.12365336404905482, + "grad_norm": 12.126015663146973, + "learning_rate": 9.852967348894517e-05, + "loss": 1.0546, + "step": 1825 + }, + { + "epoch": 0.1237211193170269, + "grad_norm": 12.094749450683594, + "learning_rate": 9.852830446984735e-05, + "loss": 1.374, + "step": 1826 + }, + { + "epoch": 0.12378887458499899, + "grad_norm": 7.607821941375732, + "learning_rate": 9.852693545074955e-05, + "loss": 0.827, + "step": 1827 + }, + { + "epoch": 0.12385662985297106, + "grad_norm": 12.086756706237793, + "learning_rate": 9.852556643165173e-05, + "loss": 1.1219, + "step": 1828 + }, + { + "epoch": 0.12392438512094316, + "grad_norm": 10.215773582458496, + "learning_rate": 9.852419741255391e-05, + "loss": 1.0148, + "step": 1829 + }, + { + "epoch": 0.12399214038891523, + "grad_norm": 10.438709259033203, + "learning_rate": 9.852282839345609e-05, + "loss": 1.0649, + "step": 1830 + }, + { + "epoch": 0.12405989565688733, + "grad_norm": 10.564906120300293, + "learning_rate": 9.852145937435828e-05, + "loss": 1.4346, + "step": 1831 + }, + { + "epoch": 0.1241276509248594, + "grad_norm": 9.83198356628418, + "learning_rate": 9.852009035526046e-05, + "loss": 0.847, + "step": 1832 + }, + { + "epoch": 0.1241954061928315, + "grad_norm": 10.990026473999023, + "learning_rate": 9.851872133616264e-05, + "loss": 1.2358, + "step": 1833 + }, + { + "epoch": 0.12426316146080357, + "grad_norm": 10.108382225036621, + "learning_rate": 9.851735231706482e-05, + "loss": 0.9943, + "step": 1834 + }, + { + "epoch": 0.12433091672877566, + "grad_norm": 10.23820686340332, + "learning_rate": 9.8515983297967e-05, + "loss": 1.2464, + "step": 1835 + }, + { + "epoch": 0.12439867199674774, + "grad_norm": 11.552473068237305, + "learning_rate": 9.85146142788692e-05, + "loss": 1.2175, + "step": 1836 + }, + { + "epoch": 0.12446642726471983, + "grad_norm": 8.646978378295898, + "learning_rate": 9.851324525977138e-05, + "loss": 0.9975, + "step": 1837 + }, + { + "epoch": 0.12453418253269191, + "grad_norm": 9.344541549682617, + "learning_rate": 9.851187624067356e-05, + "loss": 0.9463, + "step": 1838 + }, + { + "epoch": 0.124601937800664, + "grad_norm": 12.645013809204102, + "learning_rate": 9.851050722157574e-05, + "loss": 1.0396, + "step": 1839 + }, + { + "epoch": 0.12466969306863608, + "grad_norm": 10.093807220458984, + "learning_rate": 9.850913820247792e-05, + "loss": 1.2554, + "step": 1840 + }, + { + "epoch": 0.12473744833660817, + "grad_norm": 8.792567253112793, + "learning_rate": 9.850776918338011e-05, + "loss": 0.9852, + "step": 1841 + }, + { + "epoch": 0.12480520360458025, + "grad_norm": 10.448987007141113, + "learning_rate": 9.85064001642823e-05, + "loss": 1.168, + "step": 1842 + }, + { + "epoch": 0.12487295887255234, + "grad_norm": 9.953516960144043, + "learning_rate": 9.850503114518447e-05, + "loss": 1.0116, + "step": 1843 + }, + { + "epoch": 0.12494071414052442, + "grad_norm": 10.187036514282227, + "learning_rate": 9.850366212608665e-05, + "loss": 1.203, + "step": 1844 + }, + { + "epoch": 0.1250084694084965, + "grad_norm": 8.890667915344238, + "learning_rate": 9.850229310698885e-05, + "loss": 1.0104, + "step": 1845 + }, + { + "epoch": 0.1250762246764686, + "grad_norm": 11.916625022888184, + "learning_rate": 9.850092408789103e-05, + "loss": 1.0254, + "step": 1846 + }, + { + "epoch": 0.12514397994444068, + "grad_norm": 10.934864044189453, + "learning_rate": 9.849955506879321e-05, + "loss": 0.975, + "step": 1847 + }, + { + "epoch": 0.12521173521241277, + "grad_norm": 11.214954376220703, + "learning_rate": 9.849818604969539e-05, + "loss": 1.45, + "step": 1848 + }, + { + "epoch": 0.12527949048038484, + "grad_norm": 8.801512718200684, + "learning_rate": 9.849681703059758e-05, + "loss": 1.0899, + "step": 1849 + }, + { + "epoch": 0.12534724574835693, + "grad_norm": 10.340089797973633, + "learning_rate": 9.849544801149976e-05, + "loss": 1.1704, + "step": 1850 + }, + { + "epoch": 0.12541500101632902, + "grad_norm": 9.204201698303223, + "learning_rate": 9.849407899240194e-05, + "loss": 1.0569, + "step": 1851 + }, + { + "epoch": 0.1254827562843011, + "grad_norm": 9.43604564666748, + "learning_rate": 9.849270997330414e-05, + "loss": 1.0721, + "step": 1852 + }, + { + "epoch": 0.1255505115522732, + "grad_norm": 10.110416412353516, + "learning_rate": 9.849134095420632e-05, + "loss": 1.1265, + "step": 1853 + }, + { + "epoch": 0.12561826682024527, + "grad_norm": 10.630755424499512, + "learning_rate": 9.84899719351085e-05, + "loss": 1.1103, + "step": 1854 + }, + { + "epoch": 0.12568602208821736, + "grad_norm": 10.888835906982422, + "learning_rate": 9.848860291601069e-05, + "loss": 1.347, + "step": 1855 + }, + { + "epoch": 0.12575377735618945, + "grad_norm": 8.84623908996582, + "learning_rate": 9.848723389691287e-05, + "loss": 0.8531, + "step": 1856 + }, + { + "epoch": 0.12582153262416154, + "grad_norm": 10.99928092956543, + "learning_rate": 9.848586487781505e-05, + "loss": 1.1747, + "step": 1857 + }, + { + "epoch": 0.1258892878921336, + "grad_norm": 8.556151390075684, + "learning_rate": 9.848449585871723e-05, + "loss": 1.0207, + "step": 1858 + }, + { + "epoch": 0.1259570431601057, + "grad_norm": 10.207547187805176, + "learning_rate": 9.848312683961943e-05, + "loss": 1.4782, + "step": 1859 + }, + { + "epoch": 0.1260247984280778, + "grad_norm": 9.39505386352539, + "learning_rate": 9.848175782052161e-05, + "loss": 1.1415, + "step": 1860 + }, + { + "epoch": 0.12609255369604988, + "grad_norm": 10.647768020629883, + "learning_rate": 9.848038880142379e-05, + "loss": 1.1439, + "step": 1861 + }, + { + "epoch": 0.12616030896402194, + "grad_norm": 11.517707824707031, + "learning_rate": 9.847901978232597e-05, + "loss": 1.3181, + "step": 1862 + }, + { + "epoch": 0.12622806423199404, + "grad_norm": 10.916093826293945, + "learning_rate": 9.847765076322815e-05, + "loss": 1.0708, + "step": 1863 + }, + { + "epoch": 0.12629581949996613, + "grad_norm": 9.893363952636719, + "learning_rate": 9.847628174413034e-05, + "loss": 1.1932, + "step": 1864 + }, + { + "epoch": 0.12636357476793822, + "grad_norm": 12.349174499511719, + "learning_rate": 9.847491272503252e-05, + "loss": 1.3161, + "step": 1865 + }, + { + "epoch": 0.12643133003591028, + "grad_norm": 10.157081604003906, + "learning_rate": 9.84735437059347e-05, + "loss": 1.1325, + "step": 1866 + }, + { + "epoch": 0.12649908530388237, + "grad_norm": 9.772073745727539, + "learning_rate": 9.847217468683688e-05, + "loss": 1.2739, + "step": 1867 + }, + { + "epoch": 0.12656684057185447, + "grad_norm": 10.468371391296387, + "learning_rate": 9.847080566773908e-05, + "loss": 1.0279, + "step": 1868 + }, + { + "epoch": 0.12663459583982656, + "grad_norm": 11.055033683776855, + "learning_rate": 9.846943664864126e-05, + "loss": 1.0584, + "step": 1869 + }, + { + "epoch": 0.12670235110779862, + "grad_norm": 10.47987174987793, + "learning_rate": 9.846806762954344e-05, + "loss": 1.0385, + "step": 1870 + }, + { + "epoch": 0.1267701063757707, + "grad_norm": 8.933801651000977, + "learning_rate": 9.846669861044562e-05, + "loss": 1.0259, + "step": 1871 + }, + { + "epoch": 0.1268378616437428, + "grad_norm": 10.32228946685791, + "learning_rate": 9.84653295913478e-05, + "loss": 1.0472, + "step": 1872 + }, + { + "epoch": 0.1269056169117149, + "grad_norm": 9.294051170349121, + "learning_rate": 9.846396057224999e-05, + "loss": 1.1638, + "step": 1873 + }, + { + "epoch": 0.12697337217968696, + "grad_norm": 8.32187271118164, + "learning_rate": 9.846259155315217e-05, + "loss": 0.9865, + "step": 1874 + }, + { + "epoch": 0.12704112744765905, + "grad_norm": 8.782264709472656, + "learning_rate": 9.846122253405435e-05, + "loss": 1.0868, + "step": 1875 + }, + { + "epoch": 0.12710888271563114, + "grad_norm": 10.444697380065918, + "learning_rate": 9.845985351495653e-05, + "loss": 1.2137, + "step": 1876 + }, + { + "epoch": 0.12717663798360324, + "grad_norm": 10.543922424316406, + "learning_rate": 9.845848449585873e-05, + "loss": 1.2337, + "step": 1877 + }, + { + "epoch": 0.1272443932515753, + "grad_norm": 8.504612922668457, + "learning_rate": 9.845711547676091e-05, + "loss": 1.0234, + "step": 1878 + }, + { + "epoch": 0.1273121485195474, + "grad_norm": 8.83178997039795, + "learning_rate": 9.845574645766309e-05, + "loss": 1.0611, + "step": 1879 + }, + { + "epoch": 0.12737990378751948, + "grad_norm": 14.37713623046875, + "learning_rate": 9.845437743856527e-05, + "loss": 1.1704, + "step": 1880 + }, + { + "epoch": 0.12744765905549157, + "grad_norm": 12.43575668334961, + "learning_rate": 9.845300841946745e-05, + "loss": 0.848, + "step": 1881 + }, + { + "epoch": 0.12751541432346364, + "grad_norm": 11.625554084777832, + "learning_rate": 9.845163940036964e-05, + "loss": 1.2395, + "step": 1882 + }, + { + "epoch": 0.12758316959143573, + "grad_norm": 7.8962082862854, + "learning_rate": 9.845027038127182e-05, + "loss": 0.9702, + "step": 1883 + }, + { + "epoch": 0.12765092485940782, + "grad_norm": 11.689674377441406, + "learning_rate": 9.8448901362174e-05, + "loss": 1.4885, + "step": 1884 + }, + { + "epoch": 0.1277186801273799, + "grad_norm": 11.037276268005371, + "learning_rate": 9.844753234307618e-05, + "loss": 1.1751, + "step": 1885 + }, + { + "epoch": 0.12778643539535198, + "grad_norm": 11.115680694580078, + "learning_rate": 9.844616332397836e-05, + "loss": 1.2912, + "step": 1886 + }, + { + "epoch": 0.12785419066332407, + "grad_norm": 10.553492546081543, + "learning_rate": 9.844479430488056e-05, + "loss": 1.1001, + "step": 1887 + }, + { + "epoch": 0.12792194593129616, + "grad_norm": 9.92845630645752, + "learning_rate": 9.844342528578274e-05, + "loss": 1.1697, + "step": 1888 + }, + { + "epoch": 0.12798970119926825, + "grad_norm": 9.33232307434082, + "learning_rate": 9.844205626668492e-05, + "loss": 1.0931, + "step": 1889 + }, + { + "epoch": 0.12805745646724032, + "grad_norm": 12.35257339477539, + "learning_rate": 9.84406872475871e-05, + "loss": 0.9819, + "step": 1890 + }, + { + "epoch": 0.1281252117352124, + "grad_norm": 9.913105010986328, + "learning_rate": 9.843931822848929e-05, + "loss": 1.2515, + "step": 1891 + }, + { + "epoch": 0.1281929670031845, + "grad_norm": 11.218729019165039, + "learning_rate": 9.843794920939147e-05, + "loss": 1.0073, + "step": 1892 + }, + { + "epoch": 0.1282607222711566, + "grad_norm": 9.97374153137207, + "learning_rate": 9.843658019029365e-05, + "loss": 1.0527, + "step": 1893 + }, + { + "epoch": 0.12832847753912865, + "grad_norm": 8.837554931640625, + "learning_rate": 9.843521117119583e-05, + "loss": 1.0175, + "step": 1894 + }, + { + "epoch": 0.12839623280710075, + "grad_norm": 9.207158088684082, + "learning_rate": 9.843384215209803e-05, + "loss": 1.0729, + "step": 1895 + }, + { + "epoch": 0.12846398807507284, + "grad_norm": 10.960060119628906, + "learning_rate": 9.843247313300021e-05, + "loss": 1.0485, + "step": 1896 + }, + { + "epoch": 0.12853174334304493, + "grad_norm": 9.724952697753906, + "learning_rate": 9.843110411390239e-05, + "loss": 1.2624, + "step": 1897 + }, + { + "epoch": 0.128599498611017, + "grad_norm": 9.02813720703125, + "learning_rate": 9.842973509480458e-05, + "loss": 1.011, + "step": 1898 + }, + { + "epoch": 0.12866725387898909, + "grad_norm": 8.507847785949707, + "learning_rate": 9.842836607570676e-05, + "loss": 0.9155, + "step": 1899 + }, + { + "epoch": 0.12873500914696118, + "grad_norm": 10.609807968139648, + "learning_rate": 9.842699705660894e-05, + "loss": 1.1984, + "step": 1900 + }, + { + "epoch": 0.12880276441493327, + "grad_norm": 10.144070625305176, + "learning_rate": 9.842562803751114e-05, + "loss": 1.2331, + "step": 1901 + }, + { + "epoch": 0.12887051968290533, + "grad_norm": 11.719651222229004, + "learning_rate": 9.842425901841332e-05, + "loss": 1.2086, + "step": 1902 + }, + { + "epoch": 0.12893827495087742, + "grad_norm": 10.689997673034668, + "learning_rate": 9.84228899993155e-05, + "loss": 1.3769, + "step": 1903 + }, + { + "epoch": 0.12900603021884952, + "grad_norm": 10.7424955368042, + "learning_rate": 9.842152098021768e-05, + "loss": 1.1844, + "step": 1904 + }, + { + "epoch": 0.1290737854868216, + "grad_norm": 11.207498550415039, + "learning_rate": 9.842015196111987e-05, + "loss": 1.2151, + "step": 1905 + }, + { + "epoch": 0.1291415407547937, + "grad_norm": 10.689212799072266, + "learning_rate": 9.841878294202205e-05, + "loss": 1.1182, + "step": 1906 + }, + { + "epoch": 0.12920929602276576, + "grad_norm": 9.708647727966309, + "learning_rate": 9.841741392292423e-05, + "loss": 1.2874, + "step": 1907 + }, + { + "epoch": 0.12927705129073785, + "grad_norm": 9.809847831726074, + "learning_rate": 9.841604490382641e-05, + "loss": 1.1716, + "step": 1908 + }, + { + "epoch": 0.12934480655870995, + "grad_norm": 8.170798301696777, + "learning_rate": 9.84146758847286e-05, + "loss": 1.1342, + "step": 1909 + }, + { + "epoch": 0.12941256182668204, + "grad_norm": 9.12773609161377, + "learning_rate": 9.841330686563079e-05, + "loss": 1.1506, + "step": 1910 + }, + { + "epoch": 0.1294803170946541, + "grad_norm": 13.67032241821289, + "learning_rate": 9.841193784653297e-05, + "loss": 1.2065, + "step": 1911 + }, + { + "epoch": 0.1295480723626262, + "grad_norm": 9.35611629486084, + "learning_rate": 9.841056882743515e-05, + "loss": 1.281, + "step": 1912 + }, + { + "epoch": 0.12961582763059828, + "grad_norm": 9.044548988342285, + "learning_rate": 9.840919980833733e-05, + "loss": 1.0285, + "step": 1913 + }, + { + "epoch": 0.12968358289857038, + "grad_norm": 9.955796241760254, + "learning_rate": 9.840783078923952e-05, + "loss": 1.1742, + "step": 1914 + }, + { + "epoch": 0.12975133816654244, + "grad_norm": 10.456665992736816, + "learning_rate": 9.84064617701417e-05, + "loss": 1.3452, + "step": 1915 + }, + { + "epoch": 0.12981909343451453, + "grad_norm": 10.861869812011719, + "learning_rate": 9.840509275104388e-05, + "loss": 1.0397, + "step": 1916 + }, + { + "epoch": 0.12988684870248662, + "grad_norm": 8.546277046203613, + "learning_rate": 9.840372373194606e-05, + "loss": 0.8762, + "step": 1917 + }, + { + "epoch": 0.12995460397045872, + "grad_norm": 9.103103637695312, + "learning_rate": 9.840235471284824e-05, + "loss": 1.0998, + "step": 1918 + }, + { + "epoch": 0.13002235923843078, + "grad_norm": 9.794631958007812, + "learning_rate": 9.840098569375044e-05, + "loss": 1.0644, + "step": 1919 + }, + { + "epoch": 0.13009011450640287, + "grad_norm": 8.702750205993652, + "learning_rate": 9.839961667465262e-05, + "loss": 1.108, + "step": 1920 + }, + { + "epoch": 0.13015786977437496, + "grad_norm": 11.108800888061523, + "learning_rate": 9.83982476555548e-05, + "loss": 1.375, + "step": 1921 + }, + { + "epoch": 0.13022562504234705, + "grad_norm": 11.253337860107422, + "learning_rate": 9.839687863645698e-05, + "loss": 1.0246, + "step": 1922 + }, + { + "epoch": 0.13029338031031912, + "grad_norm": 9.817541122436523, + "learning_rate": 9.839550961735917e-05, + "loss": 1.0222, + "step": 1923 + }, + { + "epoch": 0.1303611355782912, + "grad_norm": 9.378199577331543, + "learning_rate": 9.839414059826135e-05, + "loss": 1.0387, + "step": 1924 + }, + { + "epoch": 0.1304288908462633, + "grad_norm": 8.789336204528809, + "learning_rate": 9.839277157916353e-05, + "loss": 1.0635, + "step": 1925 + }, + { + "epoch": 0.1304966461142354, + "grad_norm": 11.900403022766113, + "learning_rate": 9.839140256006571e-05, + "loss": 1.1224, + "step": 1926 + }, + { + "epoch": 0.13056440138220746, + "grad_norm": 10.007912635803223, + "learning_rate": 9.83900335409679e-05, + "loss": 0.8988, + "step": 1927 + }, + { + "epoch": 0.13063215665017955, + "grad_norm": 10.157328605651855, + "learning_rate": 9.838866452187009e-05, + "loss": 0.952, + "step": 1928 + }, + { + "epoch": 0.13069991191815164, + "grad_norm": 8.763729095458984, + "learning_rate": 9.838729550277227e-05, + "loss": 0.9304, + "step": 1929 + }, + { + "epoch": 0.13076766718612373, + "grad_norm": 10.75432300567627, + "learning_rate": 9.838592648367445e-05, + "loss": 1.0438, + "step": 1930 + }, + { + "epoch": 0.1308354224540958, + "grad_norm": 8.840702056884766, + "learning_rate": 9.838455746457663e-05, + "loss": 0.9164, + "step": 1931 + }, + { + "epoch": 0.1309031777220679, + "grad_norm": 9.526811599731445, + "learning_rate": 9.838318844547882e-05, + "loss": 1.2181, + "step": 1932 + }, + { + "epoch": 0.13097093299003998, + "grad_norm": 12.827199935913086, + "learning_rate": 9.8381819426381e-05, + "loss": 1.223, + "step": 1933 + }, + { + "epoch": 0.13103868825801207, + "grad_norm": 9.47105884552002, + "learning_rate": 9.838045040728318e-05, + "loss": 1.1014, + "step": 1934 + }, + { + "epoch": 0.13110644352598413, + "grad_norm": 9.044878959655762, + "learning_rate": 9.837908138818536e-05, + "loss": 1.1451, + "step": 1935 + }, + { + "epoch": 0.13117419879395623, + "grad_norm": 9.24599838256836, + "learning_rate": 9.837771236908754e-05, + "loss": 1.1144, + "step": 1936 + }, + { + "epoch": 0.13124195406192832, + "grad_norm": 12.543252944946289, + "learning_rate": 9.837634334998974e-05, + "loss": 1.1233, + "step": 1937 + }, + { + "epoch": 0.1313097093299004, + "grad_norm": 10.022245407104492, + "learning_rate": 9.837497433089192e-05, + "loss": 1.3569, + "step": 1938 + }, + { + "epoch": 0.13137746459787247, + "grad_norm": 11.967863082885742, + "learning_rate": 9.83736053117941e-05, + "loss": 1.2086, + "step": 1939 + }, + { + "epoch": 0.13144521986584456, + "grad_norm": 9.64406967163086, + "learning_rate": 9.837223629269628e-05, + "loss": 1.1647, + "step": 1940 + }, + { + "epoch": 0.13151297513381666, + "grad_norm": 8.538762092590332, + "learning_rate": 9.837086727359847e-05, + "loss": 1.012, + "step": 1941 + }, + { + "epoch": 0.13158073040178875, + "grad_norm": 10.889129638671875, + "learning_rate": 9.836949825450065e-05, + "loss": 1.1559, + "step": 1942 + }, + { + "epoch": 0.1316484856697608, + "grad_norm": 8.407093048095703, + "learning_rate": 9.836812923540283e-05, + "loss": 0.9624, + "step": 1943 + }, + { + "epoch": 0.1317162409377329, + "grad_norm": 9.175569534301758, + "learning_rate": 9.836676021630503e-05, + "loss": 0.9814, + "step": 1944 + }, + { + "epoch": 0.131783996205705, + "grad_norm": 11.240396499633789, + "learning_rate": 9.836539119720721e-05, + "loss": 1.3059, + "step": 1945 + }, + { + "epoch": 0.1318517514736771, + "grad_norm": 8.791098594665527, + "learning_rate": 9.836402217810939e-05, + "loss": 0.9891, + "step": 1946 + }, + { + "epoch": 0.13191950674164915, + "grad_norm": 7.401971340179443, + "learning_rate": 9.836265315901158e-05, + "loss": 0.9706, + "step": 1947 + }, + { + "epoch": 0.13198726200962124, + "grad_norm": 13.127768516540527, + "learning_rate": 9.836128413991376e-05, + "loss": 0.8924, + "step": 1948 + }, + { + "epoch": 0.13205501727759333, + "grad_norm": 9.805618286132812, + "learning_rate": 9.835991512081594e-05, + "loss": 0.9969, + "step": 1949 + }, + { + "epoch": 0.13212277254556543, + "grad_norm": 10.500420570373535, + "learning_rate": 9.835854610171812e-05, + "loss": 1.1045, + "step": 1950 + }, + { + "epoch": 0.1321905278135375, + "grad_norm": 10.043769836425781, + "learning_rate": 9.835717708262032e-05, + "loss": 1.2475, + "step": 1951 + }, + { + "epoch": 0.13225828308150958, + "grad_norm": 10.6277437210083, + "learning_rate": 9.83558080635225e-05, + "loss": 0.9285, + "step": 1952 + }, + { + "epoch": 0.13232603834948167, + "grad_norm": 9.011415481567383, + "learning_rate": 9.835443904442468e-05, + "loss": 1.1895, + "step": 1953 + }, + { + "epoch": 0.13239379361745376, + "grad_norm": 11.031267166137695, + "learning_rate": 9.835307002532686e-05, + "loss": 1.1781, + "step": 1954 + }, + { + "epoch": 0.13246154888542583, + "grad_norm": 9.960331916809082, + "learning_rate": 9.835170100622905e-05, + "loss": 1.0972, + "step": 1955 + }, + { + "epoch": 0.13252930415339792, + "grad_norm": 9.549619674682617, + "learning_rate": 9.835033198713123e-05, + "loss": 1.0918, + "step": 1956 + }, + { + "epoch": 0.13259705942137, + "grad_norm": 9.780478477478027, + "learning_rate": 9.834896296803341e-05, + "loss": 1.1491, + "step": 1957 + }, + { + "epoch": 0.1326648146893421, + "grad_norm": 8.948554992675781, + "learning_rate": 9.834759394893559e-05, + "loss": 1.0969, + "step": 1958 + }, + { + "epoch": 0.1327325699573142, + "grad_norm": 9.165532112121582, + "learning_rate": 9.834622492983777e-05, + "loss": 0.9294, + "step": 1959 + }, + { + "epoch": 0.13280032522528626, + "grad_norm": 8.738619804382324, + "learning_rate": 9.834485591073997e-05, + "loss": 1.3074, + "step": 1960 + }, + { + "epoch": 0.13286808049325835, + "grad_norm": 9.442314147949219, + "learning_rate": 9.834348689164215e-05, + "loss": 1.1238, + "step": 1961 + }, + { + "epoch": 0.13293583576123044, + "grad_norm": 12.411934852600098, + "learning_rate": 9.834211787254433e-05, + "loss": 1.1405, + "step": 1962 + }, + { + "epoch": 0.13300359102920253, + "grad_norm": 9.911120414733887, + "learning_rate": 9.834074885344651e-05, + "loss": 1.2197, + "step": 1963 + }, + { + "epoch": 0.1330713462971746, + "grad_norm": 9.619095802307129, + "learning_rate": 9.83393798343487e-05, + "loss": 1.0526, + "step": 1964 + }, + { + "epoch": 0.1331391015651467, + "grad_norm": 10.163374900817871, + "learning_rate": 9.833801081525088e-05, + "loss": 0.8374, + "step": 1965 + }, + { + "epoch": 0.13320685683311878, + "grad_norm": 9.342517852783203, + "learning_rate": 9.833664179615306e-05, + "loss": 1.1908, + "step": 1966 + }, + { + "epoch": 0.13327461210109087, + "grad_norm": 10.419418334960938, + "learning_rate": 9.833527277705524e-05, + "loss": 1.1175, + "step": 1967 + }, + { + "epoch": 0.13334236736906294, + "grad_norm": 9.5196533203125, + "learning_rate": 9.833390375795742e-05, + "loss": 1.2347, + "step": 1968 + }, + { + "epoch": 0.13341012263703503, + "grad_norm": 9.242755889892578, + "learning_rate": 9.833253473885962e-05, + "loss": 1.2401, + "step": 1969 + }, + { + "epoch": 0.13347787790500712, + "grad_norm": 10.243762969970703, + "learning_rate": 9.83311657197618e-05, + "loss": 1.3492, + "step": 1970 + }, + { + "epoch": 0.1335456331729792, + "grad_norm": 9.169745445251465, + "learning_rate": 9.832979670066398e-05, + "loss": 0.9533, + "step": 1971 + }, + { + "epoch": 0.13361338844095128, + "grad_norm": 10.292695999145508, + "learning_rate": 9.832842768156616e-05, + "loss": 1.2216, + "step": 1972 + }, + { + "epoch": 0.13368114370892337, + "grad_norm": 9.25019645690918, + "learning_rate": 9.832705866246834e-05, + "loss": 1.1613, + "step": 1973 + }, + { + "epoch": 0.13374889897689546, + "grad_norm": 8.518020629882812, + "learning_rate": 9.832568964337053e-05, + "loss": 1.0767, + "step": 1974 + }, + { + "epoch": 0.13381665424486755, + "grad_norm": 11.824376106262207, + "learning_rate": 9.832432062427271e-05, + "loss": 1.1884, + "step": 1975 + }, + { + "epoch": 0.13388440951283961, + "grad_norm": 10.544015884399414, + "learning_rate": 9.832295160517489e-05, + "loss": 0.9675, + "step": 1976 + }, + { + "epoch": 0.1339521647808117, + "grad_norm": 9.495721817016602, + "learning_rate": 9.832158258607707e-05, + "loss": 1.0586, + "step": 1977 + }, + { + "epoch": 0.1340199200487838, + "grad_norm": 10.378434181213379, + "learning_rate": 9.832021356697927e-05, + "loss": 1.176, + "step": 1978 + }, + { + "epoch": 0.1340876753167559, + "grad_norm": 10.026887893676758, + "learning_rate": 9.831884454788145e-05, + "loss": 1.035, + "step": 1979 + }, + { + "epoch": 0.13415543058472795, + "grad_norm": 8.878249168395996, + "learning_rate": 9.831747552878363e-05, + "loss": 0.8441, + "step": 1980 + }, + { + "epoch": 0.13422318585270004, + "grad_norm": 9.637616157531738, + "learning_rate": 9.831610650968581e-05, + "loss": 1.1435, + "step": 1981 + }, + { + "epoch": 0.13429094112067214, + "grad_norm": 10.153711318969727, + "learning_rate": 9.831473749058799e-05, + "loss": 1.0833, + "step": 1982 + }, + { + "epoch": 0.13435869638864423, + "grad_norm": 8.48596477508545, + "learning_rate": 9.831336847149018e-05, + "loss": 0.7752, + "step": 1983 + }, + { + "epoch": 0.1344264516566163, + "grad_norm": 8.454994201660156, + "learning_rate": 9.831199945239236e-05, + "loss": 0.8657, + "step": 1984 + }, + { + "epoch": 0.13449420692458838, + "grad_norm": 10.607659339904785, + "learning_rate": 9.831063043329454e-05, + "loss": 1.1836, + "step": 1985 + }, + { + "epoch": 0.13456196219256047, + "grad_norm": 8.560189247131348, + "learning_rate": 9.830926141419672e-05, + "loss": 0.7877, + "step": 1986 + }, + { + "epoch": 0.13462971746053257, + "grad_norm": 10.687662124633789, + "learning_rate": 9.830789239509892e-05, + "loss": 1.1586, + "step": 1987 + }, + { + "epoch": 0.13469747272850463, + "grad_norm": 9.725050926208496, + "learning_rate": 9.83065233760011e-05, + "loss": 1.214, + "step": 1988 + }, + { + "epoch": 0.13476522799647672, + "grad_norm": 9.808280944824219, + "learning_rate": 9.830515435690328e-05, + "loss": 1.0532, + "step": 1989 + }, + { + "epoch": 0.1348329832644488, + "grad_norm": 8.838544845581055, + "learning_rate": 9.830378533780547e-05, + "loss": 1.0872, + "step": 1990 + }, + { + "epoch": 0.1349007385324209, + "grad_norm": 12.043976783752441, + "learning_rate": 9.830241631870765e-05, + "loss": 1.2248, + "step": 1991 + }, + { + "epoch": 0.13496849380039297, + "grad_norm": 10.044602394104004, + "learning_rate": 9.830104729960983e-05, + "loss": 0.9746, + "step": 1992 + }, + { + "epoch": 0.13503624906836506, + "grad_norm": 10.861641883850098, + "learning_rate": 9.829967828051203e-05, + "loss": 0.9868, + "step": 1993 + }, + { + "epoch": 0.13510400433633715, + "grad_norm": 9.005234718322754, + "learning_rate": 9.82983092614142e-05, + "loss": 1.0619, + "step": 1994 + }, + { + "epoch": 0.13517175960430924, + "grad_norm": 9.902874946594238, + "learning_rate": 9.829694024231639e-05, + "loss": 1.3176, + "step": 1995 + }, + { + "epoch": 0.1352395148722813, + "grad_norm": 12.133747100830078, + "learning_rate": 9.829557122321857e-05, + "loss": 1.3634, + "step": 1996 + }, + { + "epoch": 0.1353072701402534, + "grad_norm": 9.571062088012695, + "learning_rate": 9.829420220412076e-05, + "loss": 1.139, + "step": 1997 + }, + { + "epoch": 0.1353750254082255, + "grad_norm": 8.269545555114746, + "learning_rate": 9.829283318502294e-05, + "loss": 1.1422, + "step": 1998 + }, + { + "epoch": 0.13544278067619758, + "grad_norm": 9.963309288024902, + "learning_rate": 9.829146416592512e-05, + "loss": 1.0612, + "step": 1999 + }, + { + "epoch": 0.13551053594416965, + "grad_norm": 7.434775352478027, + "learning_rate": 9.82900951468273e-05, + "loss": 1.0418, + "step": 2000 + }, + { + "epoch": 0.13557829121214174, + "grad_norm": 8.89494800567627, + "learning_rate": 9.82887261277295e-05, + "loss": 1.0677, + "step": 2001 + }, + { + "epoch": 0.13564604648011383, + "grad_norm": 9.349754333496094, + "learning_rate": 9.828735710863168e-05, + "loss": 0.8932, + "step": 2002 + }, + { + "epoch": 0.13571380174808592, + "grad_norm": 7.574460506439209, + "learning_rate": 9.828598808953386e-05, + "loss": 0.7652, + "step": 2003 + }, + { + "epoch": 0.13578155701605799, + "grad_norm": 10.191296577453613, + "learning_rate": 9.828461907043604e-05, + "loss": 0.9881, + "step": 2004 + }, + { + "epoch": 0.13584931228403008, + "grad_norm": 11.235671043395996, + "learning_rate": 9.828325005133822e-05, + "loss": 1.1942, + "step": 2005 + }, + { + "epoch": 0.13591706755200217, + "grad_norm": 9.97700023651123, + "learning_rate": 9.828188103224041e-05, + "loss": 1.154, + "step": 2006 + }, + { + "epoch": 0.13598482281997426, + "grad_norm": 8.283380508422852, + "learning_rate": 9.828051201314259e-05, + "loss": 0.8207, + "step": 2007 + }, + { + "epoch": 0.13605257808794632, + "grad_norm": 9.147144317626953, + "learning_rate": 9.827914299404477e-05, + "loss": 1.0751, + "step": 2008 + }, + { + "epoch": 0.13612033335591842, + "grad_norm": 9.99758529663086, + "learning_rate": 9.827777397494695e-05, + "loss": 1.1166, + "step": 2009 + }, + { + "epoch": 0.1361880886238905, + "grad_norm": 8.340705871582031, + "learning_rate": 9.827640495584915e-05, + "loss": 1.0311, + "step": 2010 + }, + { + "epoch": 0.1362558438918626, + "grad_norm": 10.536659240722656, + "learning_rate": 9.827503593675133e-05, + "loss": 1.1704, + "step": 2011 + }, + { + "epoch": 0.1363235991598347, + "grad_norm": 9.01259994506836, + "learning_rate": 9.827366691765351e-05, + "loss": 1.0531, + "step": 2012 + }, + { + "epoch": 0.13639135442780675, + "grad_norm": 10.476259231567383, + "learning_rate": 9.827229789855569e-05, + "loss": 1.053, + "step": 2013 + }, + { + "epoch": 0.13645910969577885, + "grad_norm": 8.496943473815918, + "learning_rate": 9.827092887945787e-05, + "loss": 0.836, + "step": 2014 + }, + { + "epoch": 0.13652686496375094, + "grad_norm": 11.638651847839355, + "learning_rate": 9.826955986036006e-05, + "loss": 1.0886, + "step": 2015 + }, + { + "epoch": 0.13659462023172303, + "grad_norm": 7.483241558074951, + "learning_rate": 9.826819084126224e-05, + "loss": 1.0406, + "step": 2016 + }, + { + "epoch": 0.1366623754996951, + "grad_norm": 8.648175239562988, + "learning_rate": 9.826682182216442e-05, + "loss": 1.0442, + "step": 2017 + }, + { + "epoch": 0.13673013076766719, + "grad_norm": 8.593596458435059, + "learning_rate": 9.82654528030666e-05, + "loss": 1.1184, + "step": 2018 + }, + { + "epoch": 0.13679788603563928, + "grad_norm": 8.563817977905273, + "learning_rate": 9.826408378396878e-05, + "loss": 0.8645, + "step": 2019 + }, + { + "epoch": 0.13686564130361137, + "grad_norm": 7.812311172485352, + "learning_rate": 9.826271476487098e-05, + "loss": 1.0605, + "step": 2020 + }, + { + "epoch": 0.13693339657158343, + "grad_norm": 10.97168254852295, + "learning_rate": 9.826134574577316e-05, + "loss": 1.1474, + "step": 2021 + }, + { + "epoch": 0.13700115183955552, + "grad_norm": 7.749011993408203, + "learning_rate": 9.825997672667534e-05, + "loss": 0.9836, + "step": 2022 + }, + { + "epoch": 0.13706890710752762, + "grad_norm": 11.088539123535156, + "learning_rate": 9.825860770757752e-05, + "loss": 1.1456, + "step": 2023 + }, + { + "epoch": 0.1371366623754997, + "grad_norm": 10.960288047790527, + "learning_rate": 9.825723868847971e-05, + "loss": 1.4315, + "step": 2024 + }, + { + "epoch": 0.13720441764347177, + "grad_norm": 10.804458618164062, + "learning_rate": 9.825586966938189e-05, + "loss": 1.0388, + "step": 2025 + }, + { + "epoch": 0.13727217291144386, + "grad_norm": 7.906947612762451, + "learning_rate": 9.825450065028407e-05, + "loss": 1.1306, + "step": 2026 + }, + { + "epoch": 0.13733992817941595, + "grad_norm": 10.404504776000977, + "learning_rate": 9.825313163118625e-05, + "loss": 1.0907, + "step": 2027 + }, + { + "epoch": 0.13740768344738805, + "grad_norm": 9.684488296508789, + "learning_rate": 9.825176261208843e-05, + "loss": 1.1007, + "step": 2028 + }, + { + "epoch": 0.1374754387153601, + "grad_norm": 8.345703125, + "learning_rate": 9.825039359299063e-05, + "loss": 0.9342, + "step": 2029 + }, + { + "epoch": 0.1375431939833322, + "grad_norm": 8.513103485107422, + "learning_rate": 9.824902457389281e-05, + "loss": 1.1139, + "step": 2030 + }, + { + "epoch": 0.1376109492513043, + "grad_norm": 8.675403594970703, + "learning_rate": 9.824765555479499e-05, + "loss": 0.8499, + "step": 2031 + }, + { + "epoch": 0.13767870451927638, + "grad_norm": 10.550504684448242, + "learning_rate": 9.824628653569717e-05, + "loss": 1.2563, + "step": 2032 + }, + { + "epoch": 0.13774645978724845, + "grad_norm": 7.775172710418701, + "learning_rate": 9.824491751659936e-05, + "loss": 1.0394, + "step": 2033 + }, + { + "epoch": 0.13781421505522054, + "grad_norm": 10.057134628295898, + "learning_rate": 9.824354849750154e-05, + "loss": 1.337, + "step": 2034 + }, + { + "epoch": 0.13788197032319263, + "grad_norm": 10.658480644226074, + "learning_rate": 9.824217947840372e-05, + "loss": 1.1675, + "step": 2035 + }, + { + "epoch": 0.13794972559116472, + "grad_norm": 10.499738693237305, + "learning_rate": 9.82408104593059e-05, + "loss": 1.1397, + "step": 2036 + }, + { + "epoch": 0.1380174808591368, + "grad_norm": 12.853378295898438, + "learning_rate": 9.82394414402081e-05, + "loss": 1.2373, + "step": 2037 + }, + { + "epoch": 0.13808523612710888, + "grad_norm": 8.66174602508545, + "learning_rate": 9.823807242111028e-05, + "loss": 0.9863, + "step": 2038 + }, + { + "epoch": 0.13815299139508097, + "grad_norm": 10.246259689331055, + "learning_rate": 9.823670340201246e-05, + "loss": 1.0892, + "step": 2039 + }, + { + "epoch": 0.13822074666305306, + "grad_norm": 11.880420684814453, + "learning_rate": 9.823533438291465e-05, + "loss": 1.069, + "step": 2040 + }, + { + "epoch": 0.13828850193102513, + "grad_norm": 9.475419998168945, + "learning_rate": 9.823396536381683e-05, + "loss": 0.9276, + "step": 2041 + }, + { + "epoch": 0.13835625719899722, + "grad_norm": 9.827219009399414, + "learning_rate": 9.823259634471903e-05, + "loss": 1.2536, + "step": 2042 + }, + { + "epoch": 0.1384240124669693, + "grad_norm": 8.558990478515625, + "learning_rate": 9.82312273256212e-05, + "loss": 1.2779, + "step": 2043 + }, + { + "epoch": 0.1384917677349414, + "grad_norm": 12.038803100585938, + "learning_rate": 9.822985830652339e-05, + "loss": 0.9019, + "step": 2044 + }, + { + "epoch": 0.13855952300291347, + "grad_norm": 10.764846801757812, + "learning_rate": 9.822848928742557e-05, + "loss": 1.1803, + "step": 2045 + }, + { + "epoch": 0.13862727827088556, + "grad_norm": 10.788616180419922, + "learning_rate": 9.822712026832775e-05, + "loss": 1.3469, + "step": 2046 + }, + { + "epoch": 0.13869503353885765, + "grad_norm": 8.718049049377441, + "learning_rate": 9.822575124922994e-05, + "loss": 0.9135, + "step": 2047 + }, + { + "epoch": 0.13876278880682974, + "grad_norm": 12.245726585388184, + "learning_rate": 9.822438223013212e-05, + "loss": 1.3114, + "step": 2048 + }, + { + "epoch": 0.1388305440748018, + "grad_norm": 10.55252456665039, + "learning_rate": 9.82230132110343e-05, + "loss": 1.1303, + "step": 2049 + }, + { + "epoch": 0.1388982993427739, + "grad_norm": 8.177289962768555, + "learning_rate": 9.822164419193648e-05, + "loss": 0.9796, + "step": 2050 + }, + { + "epoch": 0.138966054610746, + "grad_norm": 10.708680152893066, + "learning_rate": 9.822027517283866e-05, + "loss": 1.1546, + "step": 2051 + }, + { + "epoch": 0.13903380987871808, + "grad_norm": 8.631631851196289, + "learning_rate": 9.821890615374086e-05, + "loss": 1.0055, + "step": 2052 + }, + { + "epoch": 0.13910156514669014, + "grad_norm": 8.778770446777344, + "learning_rate": 9.821753713464304e-05, + "loss": 0.9626, + "step": 2053 + }, + { + "epoch": 0.13916932041466223, + "grad_norm": 10.004143714904785, + "learning_rate": 9.821616811554522e-05, + "loss": 1.1411, + "step": 2054 + }, + { + "epoch": 0.13923707568263433, + "grad_norm": 9.47324275970459, + "learning_rate": 9.82147990964474e-05, + "loss": 1.0081, + "step": 2055 + }, + { + "epoch": 0.13930483095060642, + "grad_norm": 10.014934539794922, + "learning_rate": 9.821343007734959e-05, + "loss": 1.3015, + "step": 2056 + }, + { + "epoch": 0.13937258621857848, + "grad_norm": 10.589959144592285, + "learning_rate": 9.821206105825177e-05, + "loss": 1.128, + "step": 2057 + }, + { + "epoch": 0.13944034148655057, + "grad_norm": 9.273834228515625, + "learning_rate": 9.821069203915395e-05, + "loss": 0.7743, + "step": 2058 + }, + { + "epoch": 0.13950809675452266, + "grad_norm": 10.72019100189209, + "learning_rate": 9.820932302005613e-05, + "loss": 0.9524, + "step": 2059 + }, + { + "epoch": 0.13957585202249476, + "grad_norm": 11.212404251098633, + "learning_rate": 9.820795400095831e-05, + "loss": 1.2044, + "step": 2060 + }, + { + "epoch": 0.13964360729046682, + "grad_norm": 9.800296783447266, + "learning_rate": 9.82065849818605e-05, + "loss": 1.0819, + "step": 2061 + }, + { + "epoch": 0.1397113625584389, + "grad_norm": 8.668676376342773, + "learning_rate": 9.820521596276269e-05, + "loss": 1.0778, + "step": 2062 + }, + { + "epoch": 0.139779117826411, + "grad_norm": 10.854613304138184, + "learning_rate": 9.820384694366487e-05, + "loss": 1.1181, + "step": 2063 + }, + { + "epoch": 0.1398468730943831, + "grad_norm": 12.019871711730957, + "learning_rate": 9.820247792456705e-05, + "loss": 1.2507, + "step": 2064 + }, + { + "epoch": 0.1399146283623552, + "grad_norm": 8.835234642028809, + "learning_rate": 9.820110890546924e-05, + "loss": 1.1338, + "step": 2065 + }, + { + "epoch": 0.13998238363032725, + "grad_norm": 8.57636547088623, + "learning_rate": 9.819973988637142e-05, + "loss": 0.957, + "step": 2066 + }, + { + "epoch": 0.14005013889829934, + "grad_norm": 9.080939292907715, + "learning_rate": 9.81983708672736e-05, + "loss": 1.0035, + "step": 2067 + }, + { + "epoch": 0.14011789416627143, + "grad_norm": 9.616862297058105, + "learning_rate": 9.819700184817578e-05, + "loss": 0.9045, + "step": 2068 + }, + { + "epoch": 0.14018564943424353, + "grad_norm": 10.04652214050293, + "learning_rate": 9.819563282907796e-05, + "loss": 1.1202, + "step": 2069 + }, + { + "epoch": 0.1402534047022156, + "grad_norm": 9.132543563842773, + "learning_rate": 9.819426380998016e-05, + "loss": 0.9918, + "step": 2070 + }, + { + "epoch": 0.14032115997018768, + "grad_norm": 12.023603439331055, + "learning_rate": 9.819289479088234e-05, + "loss": 1.1271, + "step": 2071 + }, + { + "epoch": 0.14038891523815977, + "grad_norm": 9.167064666748047, + "learning_rate": 9.819152577178452e-05, + "loss": 0.9136, + "step": 2072 + }, + { + "epoch": 0.14045667050613186, + "grad_norm": 11.954336166381836, + "learning_rate": 9.81901567526867e-05, + "loss": 1.0073, + "step": 2073 + }, + { + "epoch": 0.14052442577410393, + "grad_norm": 9.166129112243652, + "learning_rate": 9.818878773358888e-05, + "loss": 1.1101, + "step": 2074 + }, + { + "epoch": 0.14059218104207602, + "grad_norm": 8.53992748260498, + "learning_rate": 9.818741871449107e-05, + "loss": 0.9556, + "step": 2075 + }, + { + "epoch": 0.1406599363100481, + "grad_norm": 10.769463539123535, + "learning_rate": 9.818604969539325e-05, + "loss": 1.3009, + "step": 2076 + }, + { + "epoch": 0.1407276915780202, + "grad_norm": 10.938043594360352, + "learning_rate": 9.818468067629543e-05, + "loss": 1.4299, + "step": 2077 + }, + { + "epoch": 0.14079544684599227, + "grad_norm": 10.8585844039917, + "learning_rate": 9.818331165719761e-05, + "loss": 1.0804, + "step": 2078 + }, + { + "epoch": 0.14086320211396436, + "grad_norm": 8.763557434082031, + "learning_rate": 9.81819426380998e-05, + "loss": 0.8321, + "step": 2079 + }, + { + "epoch": 0.14093095738193645, + "grad_norm": 11.917708396911621, + "learning_rate": 9.818057361900199e-05, + "loss": 1.4293, + "step": 2080 + }, + { + "epoch": 0.14099871264990854, + "grad_norm": 10.189987182617188, + "learning_rate": 9.817920459990417e-05, + "loss": 1.139, + "step": 2081 + }, + { + "epoch": 0.1410664679178806, + "grad_norm": 11.09156608581543, + "learning_rate": 9.817783558080635e-05, + "loss": 0.9866, + "step": 2082 + }, + { + "epoch": 0.1411342231858527, + "grad_norm": 9.394566535949707, + "learning_rate": 9.817646656170854e-05, + "loss": 0.9401, + "step": 2083 + }, + { + "epoch": 0.1412019784538248, + "grad_norm": 8.232394218444824, + "learning_rate": 9.817509754261072e-05, + "loss": 0.9193, + "step": 2084 + }, + { + "epoch": 0.14126973372179688, + "grad_norm": 9.4952392578125, + "learning_rate": 9.81737285235129e-05, + "loss": 1.3217, + "step": 2085 + }, + { + "epoch": 0.14133748898976894, + "grad_norm": 10.110014915466309, + "learning_rate": 9.81723595044151e-05, + "loss": 1.0413, + "step": 2086 + }, + { + "epoch": 0.14140524425774104, + "grad_norm": 10.850225448608398, + "learning_rate": 9.817099048531728e-05, + "loss": 1.3722, + "step": 2087 + }, + { + "epoch": 0.14147299952571313, + "grad_norm": 10.924457550048828, + "learning_rate": 9.816962146621946e-05, + "loss": 1.4004, + "step": 2088 + }, + { + "epoch": 0.14154075479368522, + "grad_norm": 10.022381782531738, + "learning_rate": 9.816825244712165e-05, + "loss": 1.1989, + "step": 2089 + }, + { + "epoch": 0.14160851006165728, + "grad_norm": 11.537693977355957, + "learning_rate": 9.816688342802383e-05, + "loss": 1.1922, + "step": 2090 + }, + { + "epoch": 0.14167626532962938, + "grad_norm": 8.525372505187988, + "learning_rate": 9.816551440892601e-05, + "loss": 0.9291, + "step": 2091 + }, + { + "epoch": 0.14174402059760147, + "grad_norm": 8.972722053527832, + "learning_rate": 9.816414538982819e-05, + "loss": 1.0917, + "step": 2092 + }, + { + "epoch": 0.14181177586557356, + "grad_norm": 8.386235237121582, + "learning_rate": 9.816277637073039e-05, + "loss": 1.0482, + "step": 2093 + }, + { + "epoch": 0.14187953113354562, + "grad_norm": 9.620473861694336, + "learning_rate": 9.816140735163257e-05, + "loss": 1.2523, + "step": 2094 + }, + { + "epoch": 0.1419472864015177, + "grad_norm": 9.236804962158203, + "learning_rate": 9.816003833253475e-05, + "loss": 0.9474, + "step": 2095 + }, + { + "epoch": 0.1420150416694898, + "grad_norm": 9.06925106048584, + "learning_rate": 9.815866931343693e-05, + "loss": 0.8921, + "step": 2096 + }, + { + "epoch": 0.1420827969374619, + "grad_norm": 10.127729415893555, + "learning_rate": 9.815730029433912e-05, + "loss": 1.1693, + "step": 2097 + }, + { + "epoch": 0.14215055220543396, + "grad_norm": 10.703007698059082, + "learning_rate": 9.81559312752413e-05, + "loss": 1.0652, + "step": 2098 + }, + { + "epoch": 0.14221830747340605, + "grad_norm": 10.004093170166016, + "learning_rate": 9.815456225614348e-05, + "loss": 1.0158, + "step": 2099 + }, + { + "epoch": 0.14228606274137814, + "grad_norm": 8.900236129760742, + "learning_rate": 9.815319323704566e-05, + "loss": 0.7327, + "step": 2100 + }, + { + "epoch": 0.14235381800935024, + "grad_norm": 13.102290153503418, + "learning_rate": 9.815182421794784e-05, + "loss": 1.2837, + "step": 2101 + }, + { + "epoch": 0.1424215732773223, + "grad_norm": 9.32657527923584, + "learning_rate": 9.815045519885004e-05, + "loss": 1.0812, + "step": 2102 + }, + { + "epoch": 0.1424893285452944, + "grad_norm": 11.42785930633545, + "learning_rate": 9.814908617975222e-05, + "loss": 1.4467, + "step": 2103 + }, + { + "epoch": 0.14255708381326648, + "grad_norm": 9.878450393676758, + "learning_rate": 9.81477171606544e-05, + "loss": 1.0774, + "step": 2104 + }, + { + "epoch": 0.14262483908123857, + "grad_norm": 9.329227447509766, + "learning_rate": 9.814634814155658e-05, + "loss": 1.2536, + "step": 2105 + }, + { + "epoch": 0.14269259434921064, + "grad_norm": 9.071858406066895, + "learning_rate": 9.814497912245876e-05, + "loss": 1.148, + "step": 2106 + }, + { + "epoch": 0.14276034961718273, + "grad_norm": 11.595810890197754, + "learning_rate": 9.814361010336095e-05, + "loss": 1.1992, + "step": 2107 + }, + { + "epoch": 0.14282810488515482, + "grad_norm": 10.041107177734375, + "learning_rate": 9.814224108426313e-05, + "loss": 1.0447, + "step": 2108 + }, + { + "epoch": 0.1428958601531269, + "grad_norm": 9.913261413574219, + "learning_rate": 9.814087206516531e-05, + "loss": 1.1629, + "step": 2109 + }, + { + "epoch": 0.14296361542109898, + "grad_norm": 10.983177185058594, + "learning_rate": 9.813950304606749e-05, + "loss": 1.0288, + "step": 2110 + }, + { + "epoch": 0.14303137068907107, + "grad_norm": 10.479610443115234, + "learning_rate": 9.813813402696969e-05, + "loss": 1.5308, + "step": 2111 + }, + { + "epoch": 0.14309912595704316, + "grad_norm": 8.70734691619873, + "learning_rate": 9.813676500787187e-05, + "loss": 0.9987, + "step": 2112 + }, + { + "epoch": 0.14316688122501525, + "grad_norm": 8.242984771728516, + "learning_rate": 9.813539598877405e-05, + "loss": 1.0051, + "step": 2113 + }, + { + "epoch": 0.14323463649298732, + "grad_norm": 8.627467155456543, + "learning_rate": 9.813402696967623e-05, + "loss": 1.1134, + "step": 2114 + }, + { + "epoch": 0.1433023917609594, + "grad_norm": 9.970986366271973, + "learning_rate": 9.813265795057841e-05, + "loss": 1.148, + "step": 2115 + }, + { + "epoch": 0.1433701470289315, + "grad_norm": 8.719794273376465, + "learning_rate": 9.81312889314806e-05, + "loss": 1.1013, + "step": 2116 + }, + { + "epoch": 0.1434379022969036, + "grad_norm": 9.455860137939453, + "learning_rate": 9.812991991238278e-05, + "loss": 1.2333, + "step": 2117 + }, + { + "epoch": 0.14350565756487568, + "grad_norm": 8.851629257202148, + "learning_rate": 9.812855089328496e-05, + "loss": 1.0985, + "step": 2118 + }, + { + "epoch": 0.14357341283284775, + "grad_norm": 11.843599319458008, + "learning_rate": 9.812718187418714e-05, + "loss": 1.2292, + "step": 2119 + }, + { + "epoch": 0.14364116810081984, + "grad_norm": 9.550506591796875, + "learning_rate": 9.812581285508934e-05, + "loss": 0.9466, + "step": 2120 + }, + { + "epoch": 0.14370892336879193, + "grad_norm": 9.234643936157227, + "learning_rate": 9.812444383599152e-05, + "loss": 1.1968, + "step": 2121 + }, + { + "epoch": 0.14377667863676402, + "grad_norm": 9.365439414978027, + "learning_rate": 9.81230748168937e-05, + "loss": 1.1643, + "step": 2122 + }, + { + "epoch": 0.14384443390473609, + "grad_norm": 9.476024627685547, + "learning_rate": 9.812170579779588e-05, + "loss": 1.0541, + "step": 2123 + }, + { + "epoch": 0.14391218917270818, + "grad_norm": 9.53847885131836, + "learning_rate": 9.812033677869806e-05, + "loss": 1.0808, + "step": 2124 + }, + { + "epoch": 0.14397994444068027, + "grad_norm": 8.918699264526367, + "learning_rate": 9.811896775960025e-05, + "loss": 1.1055, + "step": 2125 + }, + { + "epoch": 0.14404769970865236, + "grad_norm": 12.126187324523926, + "learning_rate": 9.811759874050243e-05, + "loss": 1.6079, + "step": 2126 + }, + { + "epoch": 0.14411545497662442, + "grad_norm": 10.922599792480469, + "learning_rate": 9.811622972140461e-05, + "loss": 1.2569, + "step": 2127 + }, + { + "epoch": 0.14418321024459652, + "grad_norm": 8.582880020141602, + "learning_rate": 9.811486070230679e-05, + "loss": 0.9562, + "step": 2128 + }, + { + "epoch": 0.1442509655125686, + "grad_norm": 8.926568031311035, + "learning_rate": 9.811349168320899e-05, + "loss": 1.0108, + "step": 2129 + }, + { + "epoch": 0.1443187207805407, + "grad_norm": 13.906332015991211, + "learning_rate": 9.811212266411117e-05, + "loss": 1.1851, + "step": 2130 + }, + { + "epoch": 0.14438647604851276, + "grad_norm": 10.374212265014648, + "learning_rate": 9.811075364501335e-05, + "loss": 1.2661, + "step": 2131 + }, + { + "epoch": 0.14445423131648485, + "grad_norm": 10.580810546875, + "learning_rate": 9.810938462591554e-05, + "loss": 0.8967, + "step": 2132 + }, + { + "epoch": 0.14452198658445695, + "grad_norm": 8.91865062713623, + "learning_rate": 9.810801560681772e-05, + "loss": 1.032, + "step": 2133 + }, + { + "epoch": 0.14458974185242904, + "grad_norm": 8.16041374206543, + "learning_rate": 9.81066465877199e-05, + "loss": 0.8622, + "step": 2134 + }, + { + "epoch": 0.1446574971204011, + "grad_norm": 8.575905799865723, + "learning_rate": 9.81052775686221e-05, + "loss": 0.9956, + "step": 2135 + }, + { + "epoch": 0.1447252523883732, + "grad_norm": 9.620558738708496, + "learning_rate": 9.810390854952428e-05, + "loss": 1.1442, + "step": 2136 + }, + { + "epoch": 0.14479300765634529, + "grad_norm": 10.390005111694336, + "learning_rate": 9.810253953042646e-05, + "loss": 1.0518, + "step": 2137 + }, + { + "epoch": 0.14486076292431738, + "grad_norm": 10.714217185974121, + "learning_rate": 9.810117051132864e-05, + "loss": 1.3215, + "step": 2138 + }, + { + "epoch": 0.14492851819228944, + "grad_norm": 9.390836715698242, + "learning_rate": 9.809980149223083e-05, + "loss": 1.0936, + "step": 2139 + }, + { + "epoch": 0.14499627346026153, + "grad_norm": 10.848445892333984, + "learning_rate": 9.809843247313301e-05, + "loss": 1.1946, + "step": 2140 + }, + { + "epoch": 0.14506402872823362, + "grad_norm": 11.041672706604004, + "learning_rate": 9.809706345403519e-05, + "loss": 1.0422, + "step": 2141 + }, + { + "epoch": 0.14513178399620572, + "grad_norm": 10.1151762008667, + "learning_rate": 9.809569443493737e-05, + "loss": 1.2199, + "step": 2142 + }, + { + "epoch": 0.14519953926417778, + "grad_norm": 8.686629295349121, + "learning_rate": 9.809432541583957e-05, + "loss": 0.9861, + "step": 2143 + }, + { + "epoch": 0.14526729453214987, + "grad_norm": 10.579313278198242, + "learning_rate": 9.809295639674175e-05, + "loss": 0.9372, + "step": 2144 + }, + { + "epoch": 0.14533504980012196, + "grad_norm": 8.916631698608398, + "learning_rate": 9.809158737764393e-05, + "loss": 0.8024, + "step": 2145 + }, + { + "epoch": 0.14540280506809405, + "grad_norm": 9.29333209991455, + "learning_rate": 9.80902183585461e-05, + "loss": 1.0785, + "step": 2146 + }, + { + "epoch": 0.14547056033606612, + "grad_norm": 10.584277153015137, + "learning_rate": 9.808884933944829e-05, + "loss": 0.9167, + "step": 2147 + }, + { + "epoch": 0.1455383156040382, + "grad_norm": 10.68551254272461, + "learning_rate": 9.808748032035048e-05, + "loss": 1.0019, + "step": 2148 + }, + { + "epoch": 0.1456060708720103, + "grad_norm": 9.057500839233398, + "learning_rate": 9.808611130125266e-05, + "loss": 1.194, + "step": 2149 + }, + { + "epoch": 0.1456738261399824, + "grad_norm": 8.642207145690918, + "learning_rate": 9.808474228215484e-05, + "loss": 1.0849, + "step": 2150 + }, + { + "epoch": 0.14574158140795446, + "grad_norm": 9.460419654846191, + "learning_rate": 9.808337326305702e-05, + "loss": 1.1747, + "step": 2151 + }, + { + "epoch": 0.14580933667592655, + "grad_norm": 9.003097534179688, + "learning_rate": 9.80820042439592e-05, + "loss": 0.8967, + "step": 2152 + }, + { + "epoch": 0.14587709194389864, + "grad_norm": 9.85009765625, + "learning_rate": 9.80806352248614e-05, + "loss": 1.1698, + "step": 2153 + }, + { + "epoch": 0.14594484721187073, + "grad_norm": 10.233894348144531, + "learning_rate": 9.807926620576358e-05, + "loss": 1.0496, + "step": 2154 + }, + { + "epoch": 0.1460126024798428, + "grad_norm": 9.660355567932129, + "learning_rate": 9.807789718666576e-05, + "loss": 0.9616, + "step": 2155 + }, + { + "epoch": 0.1460803577478149, + "grad_norm": 9.46469497680664, + "learning_rate": 9.807652816756794e-05, + "loss": 1.0005, + "step": 2156 + }, + { + "epoch": 0.14614811301578698, + "grad_norm": 8.004712104797363, + "learning_rate": 9.807515914847013e-05, + "loss": 0.8455, + "step": 2157 + }, + { + "epoch": 0.14621586828375907, + "grad_norm": 10.039002418518066, + "learning_rate": 9.807379012937231e-05, + "loss": 1.097, + "step": 2158 + }, + { + "epoch": 0.14628362355173113, + "grad_norm": 11.16292667388916, + "learning_rate": 9.807242111027449e-05, + "loss": 1.4332, + "step": 2159 + }, + { + "epoch": 0.14635137881970323, + "grad_norm": 9.34833812713623, + "learning_rate": 9.807105209117667e-05, + "loss": 1.0438, + "step": 2160 + }, + { + "epoch": 0.14641913408767532, + "grad_norm": 10.256768226623535, + "learning_rate": 9.806968307207885e-05, + "loss": 1.2923, + "step": 2161 + }, + { + "epoch": 0.1464868893556474, + "grad_norm": 8.890941619873047, + "learning_rate": 9.806831405298105e-05, + "loss": 0.8419, + "step": 2162 + }, + { + "epoch": 0.14655464462361947, + "grad_norm": 9.340752601623535, + "learning_rate": 9.806694503388323e-05, + "loss": 0.9504, + "step": 2163 + }, + { + "epoch": 0.14662239989159156, + "grad_norm": 10.89192008972168, + "learning_rate": 9.80655760147854e-05, + "loss": 1.012, + "step": 2164 + }, + { + "epoch": 0.14669015515956366, + "grad_norm": 8.626432418823242, + "learning_rate": 9.806420699568759e-05, + "loss": 0.8943, + "step": 2165 + }, + { + "epoch": 0.14675791042753575, + "grad_norm": 9.465259552001953, + "learning_rate": 9.806283797658978e-05, + "loss": 1.1617, + "step": 2166 + }, + { + "epoch": 0.1468256656955078, + "grad_norm": 11.695369720458984, + "learning_rate": 9.806146895749196e-05, + "loss": 1.2719, + "step": 2167 + }, + { + "epoch": 0.1468934209634799, + "grad_norm": 9.416814804077148, + "learning_rate": 9.806009993839414e-05, + "loss": 1.3615, + "step": 2168 + }, + { + "epoch": 0.146961176231452, + "grad_norm": 9.6004638671875, + "learning_rate": 9.805873091929632e-05, + "loss": 1.2571, + "step": 2169 + }, + { + "epoch": 0.1470289314994241, + "grad_norm": 9.187546730041504, + "learning_rate": 9.80573619001985e-05, + "loss": 0.962, + "step": 2170 + }, + { + "epoch": 0.14709668676739618, + "grad_norm": 10.691286087036133, + "learning_rate": 9.80559928811007e-05, + "loss": 1.0619, + "step": 2171 + }, + { + "epoch": 0.14716444203536824, + "grad_norm": 9.17457103729248, + "learning_rate": 9.805462386200288e-05, + "loss": 1.1507, + "step": 2172 + }, + { + "epoch": 0.14723219730334033, + "grad_norm": 10.739082336425781, + "learning_rate": 9.805325484290506e-05, + "loss": 1.2929, + "step": 2173 + }, + { + "epoch": 0.14729995257131243, + "grad_norm": 9.091232299804688, + "learning_rate": 9.805188582380724e-05, + "loss": 0.9422, + "step": 2174 + }, + { + "epoch": 0.14736770783928452, + "grad_norm": 8.231295585632324, + "learning_rate": 9.805051680470943e-05, + "loss": 0.9695, + "step": 2175 + }, + { + "epoch": 0.14743546310725658, + "grad_norm": 9.622401237487793, + "learning_rate": 9.804914778561161e-05, + "loss": 0.982, + "step": 2176 + }, + { + "epoch": 0.14750321837522867, + "grad_norm": 9.49541187286377, + "learning_rate": 9.804777876651379e-05, + "loss": 0.9005, + "step": 2177 + }, + { + "epoch": 0.14757097364320076, + "grad_norm": 10.584654808044434, + "learning_rate": 9.804640974741599e-05, + "loss": 0.9427, + "step": 2178 + }, + { + "epoch": 0.14763872891117286, + "grad_norm": 9.132317543029785, + "learning_rate": 9.804504072831817e-05, + "loss": 1.0192, + "step": 2179 + }, + { + "epoch": 0.14770648417914492, + "grad_norm": 8.598082542419434, + "learning_rate": 9.804367170922035e-05, + "loss": 0.9842, + "step": 2180 + }, + { + "epoch": 0.147774239447117, + "grad_norm": 8.941360473632812, + "learning_rate": 9.804230269012254e-05, + "loss": 0.9907, + "step": 2181 + }, + { + "epoch": 0.1478419947150891, + "grad_norm": 8.119913101196289, + "learning_rate": 9.804093367102472e-05, + "loss": 1.1259, + "step": 2182 + }, + { + "epoch": 0.1479097499830612, + "grad_norm": 9.505135536193848, + "learning_rate": 9.80395646519269e-05, + "loss": 1.0509, + "step": 2183 + }, + { + "epoch": 0.14797750525103326, + "grad_norm": 8.420906066894531, + "learning_rate": 9.803819563282908e-05, + "loss": 1.1827, + "step": 2184 + }, + { + "epoch": 0.14804526051900535, + "grad_norm": 9.860353469848633, + "learning_rate": 9.803682661373128e-05, + "loss": 1.0286, + "step": 2185 + }, + { + "epoch": 0.14811301578697744, + "grad_norm": 8.259954452514648, + "learning_rate": 9.803545759463346e-05, + "loss": 1.0094, + "step": 2186 + }, + { + "epoch": 0.14818077105494953, + "grad_norm": 10.46882438659668, + "learning_rate": 9.803408857553564e-05, + "loss": 1.1817, + "step": 2187 + }, + { + "epoch": 0.1482485263229216, + "grad_norm": 9.315580368041992, + "learning_rate": 9.803271955643782e-05, + "loss": 1.2816, + "step": 2188 + }, + { + "epoch": 0.1483162815908937, + "grad_norm": 10.408548355102539, + "learning_rate": 9.803135053734001e-05, + "loss": 1.0355, + "step": 2189 + }, + { + "epoch": 0.14838403685886578, + "grad_norm": 9.682003021240234, + "learning_rate": 9.802998151824219e-05, + "loss": 1.0413, + "step": 2190 + }, + { + "epoch": 0.14845179212683787, + "grad_norm": 10.041797637939453, + "learning_rate": 9.802861249914437e-05, + "loss": 0.8388, + "step": 2191 + }, + { + "epoch": 0.14851954739480994, + "grad_norm": 8.367657661437988, + "learning_rate": 9.802724348004655e-05, + "loss": 1.0724, + "step": 2192 + }, + { + "epoch": 0.14858730266278203, + "grad_norm": 9.9558744430542, + "learning_rate": 9.802587446094873e-05, + "loss": 1.3534, + "step": 2193 + }, + { + "epoch": 0.14865505793075412, + "grad_norm": 9.244332313537598, + "learning_rate": 9.802450544185093e-05, + "loss": 1.366, + "step": 2194 + }, + { + "epoch": 0.1487228131987262, + "grad_norm": 9.560718536376953, + "learning_rate": 9.80231364227531e-05, + "loss": 1.0556, + "step": 2195 + }, + { + "epoch": 0.14879056846669828, + "grad_norm": 8.724915504455566, + "learning_rate": 9.802176740365529e-05, + "loss": 1.316, + "step": 2196 + }, + { + "epoch": 0.14885832373467037, + "grad_norm": 9.468677520751953, + "learning_rate": 9.802039838455747e-05, + "loss": 1.1289, + "step": 2197 + }, + { + "epoch": 0.14892607900264246, + "grad_norm": 8.56347942352295, + "learning_rate": 9.801902936545966e-05, + "loss": 1.0402, + "step": 2198 + }, + { + "epoch": 0.14899383427061455, + "grad_norm": 9.328559875488281, + "learning_rate": 9.801766034636184e-05, + "loss": 1.186, + "step": 2199 + }, + { + "epoch": 0.14906158953858661, + "grad_norm": 10.20579719543457, + "learning_rate": 9.801629132726402e-05, + "loss": 1.0896, + "step": 2200 + }, + { + "epoch": 0.1491293448065587, + "grad_norm": 11.614208221435547, + "learning_rate": 9.80149223081662e-05, + "loss": 1.2742, + "step": 2201 + }, + { + "epoch": 0.1491971000745308, + "grad_norm": 8.291358947753906, + "learning_rate": 9.801355328906838e-05, + "loss": 1.0299, + "step": 2202 + }, + { + "epoch": 0.1492648553425029, + "grad_norm": 8.734344482421875, + "learning_rate": 9.801218426997058e-05, + "loss": 1.0922, + "step": 2203 + }, + { + "epoch": 0.14933261061047495, + "grad_norm": 8.498616218566895, + "learning_rate": 9.801081525087276e-05, + "loss": 0.8436, + "step": 2204 + }, + { + "epoch": 0.14940036587844704, + "grad_norm": 7.4501447677612305, + "learning_rate": 9.800944623177494e-05, + "loss": 0.8629, + "step": 2205 + }, + { + "epoch": 0.14946812114641914, + "grad_norm": 11.652278900146484, + "learning_rate": 9.800807721267712e-05, + "loss": 1.1089, + "step": 2206 + }, + { + "epoch": 0.14953587641439123, + "grad_norm": 11.043471336364746, + "learning_rate": 9.80067081935793e-05, + "loss": 1.2758, + "step": 2207 + }, + { + "epoch": 0.1496036316823633, + "grad_norm": 9.376245498657227, + "learning_rate": 9.800533917448149e-05, + "loss": 1.0178, + "step": 2208 + }, + { + "epoch": 0.14967138695033538, + "grad_norm": 8.678750038146973, + "learning_rate": 9.800397015538367e-05, + "loss": 1.0276, + "step": 2209 + }, + { + "epoch": 0.14973914221830747, + "grad_norm": 10.233052253723145, + "learning_rate": 9.800260113628585e-05, + "loss": 1.0468, + "step": 2210 + }, + { + "epoch": 0.14980689748627957, + "grad_norm": 10.711477279663086, + "learning_rate": 9.800123211718803e-05, + "loss": 1.3919, + "step": 2211 + }, + { + "epoch": 0.14987465275425163, + "grad_norm": 10.130411148071289, + "learning_rate": 9.799986309809023e-05, + "loss": 1.1085, + "step": 2212 + }, + { + "epoch": 0.14994240802222372, + "grad_norm": 9.64091682434082, + "learning_rate": 9.79984940789924e-05, + "loss": 1.0661, + "step": 2213 + }, + { + "epoch": 0.1500101632901958, + "grad_norm": 9.80176830291748, + "learning_rate": 9.799712505989459e-05, + "loss": 1.3426, + "step": 2214 + }, + { + "epoch": 0.1500779185581679, + "grad_norm": 9.34835147857666, + "learning_rate": 9.799575604079677e-05, + "loss": 1.0281, + "step": 2215 + }, + { + "epoch": 0.15014567382613997, + "grad_norm": 8.638712882995605, + "learning_rate": 9.799438702169895e-05, + "loss": 0.9037, + "step": 2216 + }, + { + "epoch": 0.15021342909411206, + "grad_norm": 10.499733924865723, + "learning_rate": 9.799301800260114e-05, + "loss": 1.1152, + "step": 2217 + }, + { + "epoch": 0.15028118436208415, + "grad_norm": 9.093478202819824, + "learning_rate": 9.799164898350332e-05, + "loss": 1.2026, + "step": 2218 + }, + { + "epoch": 0.15034893963005624, + "grad_norm": 13.181863784790039, + "learning_rate": 9.79902799644055e-05, + "loss": 1.3256, + "step": 2219 + }, + { + "epoch": 0.1504166948980283, + "grad_norm": 10.221563339233398, + "learning_rate": 9.798891094530768e-05, + "loss": 1.1867, + "step": 2220 + }, + { + "epoch": 0.1504844501660004, + "grad_norm": 9.512944221496582, + "learning_rate": 9.798754192620988e-05, + "loss": 1.1145, + "step": 2221 + }, + { + "epoch": 0.1505522054339725, + "grad_norm": 8.991315841674805, + "learning_rate": 9.798617290711206e-05, + "loss": 1.0289, + "step": 2222 + }, + { + "epoch": 0.15061996070194458, + "grad_norm": 9.003118515014648, + "learning_rate": 9.798480388801424e-05, + "loss": 0.9618, + "step": 2223 + }, + { + "epoch": 0.15068771596991667, + "grad_norm": 11.337166786193848, + "learning_rate": 9.798343486891643e-05, + "loss": 1.3864, + "step": 2224 + }, + { + "epoch": 0.15075547123788874, + "grad_norm": 9.193288803100586, + "learning_rate": 9.798206584981861e-05, + "loss": 1.1314, + "step": 2225 + }, + { + "epoch": 0.15082322650586083, + "grad_norm": 10.137048721313477, + "learning_rate": 9.798069683072079e-05, + "loss": 1.3325, + "step": 2226 + }, + { + "epoch": 0.15089098177383292, + "grad_norm": 8.248672485351562, + "learning_rate": 9.797932781162299e-05, + "loss": 1.0688, + "step": 2227 + }, + { + "epoch": 0.150958737041805, + "grad_norm": 10.51007080078125, + "learning_rate": 9.797795879252517e-05, + "loss": 1.2191, + "step": 2228 + }, + { + "epoch": 0.15102649230977708, + "grad_norm": 14.15853214263916, + "learning_rate": 9.797658977342735e-05, + "loss": 1.0961, + "step": 2229 + }, + { + "epoch": 0.15109424757774917, + "grad_norm": 8.42485523223877, + "learning_rate": 9.797522075432954e-05, + "loss": 1.0314, + "step": 2230 + }, + { + "epoch": 0.15116200284572126, + "grad_norm": 9.3310546875, + "learning_rate": 9.797385173523172e-05, + "loss": 1.2372, + "step": 2231 + }, + { + "epoch": 0.15122975811369335, + "grad_norm": 9.323569297790527, + "learning_rate": 9.79724827161339e-05, + "loss": 1.1823, + "step": 2232 + }, + { + "epoch": 0.15129751338166542, + "grad_norm": 9.719592094421387, + "learning_rate": 9.797111369703608e-05, + "loss": 1.0689, + "step": 2233 + }, + { + "epoch": 0.1513652686496375, + "grad_norm": 9.37340259552002, + "learning_rate": 9.796974467793826e-05, + "loss": 1.0703, + "step": 2234 + }, + { + "epoch": 0.1514330239176096, + "grad_norm": 9.803778648376465, + "learning_rate": 9.796837565884045e-05, + "loss": 1.0422, + "step": 2235 + }, + { + "epoch": 0.1515007791855817, + "grad_norm": 8.317301750183105, + "learning_rate": 9.796700663974264e-05, + "loss": 0.9829, + "step": 2236 + }, + { + "epoch": 0.15156853445355375, + "grad_norm": 8.473258972167969, + "learning_rate": 9.796563762064482e-05, + "loss": 1.064, + "step": 2237 + }, + { + "epoch": 0.15163628972152585, + "grad_norm": 8.495006561279297, + "learning_rate": 9.7964268601547e-05, + "loss": 0.8646, + "step": 2238 + }, + { + "epoch": 0.15170404498949794, + "grad_norm": 9.140898704528809, + "learning_rate": 9.796289958244918e-05, + "loss": 1.068, + "step": 2239 + }, + { + "epoch": 0.15177180025747003, + "grad_norm": 10.264453887939453, + "learning_rate": 9.796153056335137e-05, + "loss": 0.9681, + "step": 2240 + }, + { + "epoch": 0.1518395555254421, + "grad_norm": 10.407776832580566, + "learning_rate": 9.796016154425355e-05, + "loss": 1.2538, + "step": 2241 + }, + { + "epoch": 0.15190731079341419, + "grad_norm": 10.143677711486816, + "learning_rate": 9.795879252515573e-05, + "loss": 1.138, + "step": 2242 + }, + { + "epoch": 0.15197506606138628, + "grad_norm": 12.528799057006836, + "learning_rate": 9.795742350605791e-05, + "loss": 1.2654, + "step": 2243 + }, + { + "epoch": 0.15204282132935837, + "grad_norm": 10.635498046875, + "learning_rate": 9.79560544869601e-05, + "loss": 1.2116, + "step": 2244 + }, + { + "epoch": 0.15211057659733043, + "grad_norm": 10.71164608001709, + "learning_rate": 9.795468546786229e-05, + "loss": 1.0262, + "step": 2245 + }, + { + "epoch": 0.15217833186530252, + "grad_norm": 9.121109962463379, + "learning_rate": 9.795331644876447e-05, + "loss": 1.1657, + "step": 2246 + }, + { + "epoch": 0.15224608713327462, + "grad_norm": 8.174636840820312, + "learning_rate": 9.795194742966665e-05, + "loss": 1.0027, + "step": 2247 + }, + { + "epoch": 0.1523138424012467, + "grad_norm": 8.763494491577148, + "learning_rate": 9.795057841056883e-05, + "loss": 1.1103, + "step": 2248 + }, + { + "epoch": 0.15238159766921877, + "grad_norm": 8.028278350830078, + "learning_rate": 9.794920939147102e-05, + "loss": 1.0205, + "step": 2249 + }, + { + "epoch": 0.15244935293719086, + "grad_norm": 8.96112060546875, + "learning_rate": 9.79478403723732e-05, + "loss": 1.2949, + "step": 2250 + }, + { + "epoch": 0.15251710820516295, + "grad_norm": 8.67423152923584, + "learning_rate": 9.794647135327538e-05, + "loss": 1.0602, + "step": 2251 + }, + { + "epoch": 0.15258486347313505, + "grad_norm": 9.45671272277832, + "learning_rate": 9.794510233417756e-05, + "loss": 1.0362, + "step": 2252 + }, + { + "epoch": 0.1526526187411071, + "grad_norm": 10.24669075012207, + "learning_rate": 9.794373331507976e-05, + "loss": 0.9918, + "step": 2253 + }, + { + "epoch": 0.1527203740090792, + "grad_norm": 9.014079093933105, + "learning_rate": 9.794236429598194e-05, + "loss": 1.1636, + "step": 2254 + }, + { + "epoch": 0.1527881292770513, + "grad_norm": 6.800943851470947, + "learning_rate": 9.794099527688412e-05, + "loss": 0.9089, + "step": 2255 + }, + { + "epoch": 0.15285588454502338, + "grad_norm": 9.90794849395752, + "learning_rate": 9.79396262577863e-05, + "loss": 0.7669, + "step": 2256 + }, + { + "epoch": 0.15292363981299545, + "grad_norm": 9.870927810668945, + "learning_rate": 9.793825723868848e-05, + "loss": 1.1243, + "step": 2257 + }, + { + "epoch": 0.15299139508096754, + "grad_norm": 9.707404136657715, + "learning_rate": 9.793688821959067e-05, + "loss": 1.1558, + "step": 2258 + }, + { + "epoch": 0.15305915034893963, + "grad_norm": 8.362896919250488, + "learning_rate": 9.793551920049285e-05, + "loss": 0.8601, + "step": 2259 + }, + { + "epoch": 0.15312690561691172, + "grad_norm": 9.536920547485352, + "learning_rate": 9.793415018139503e-05, + "loss": 1.2741, + "step": 2260 + }, + { + "epoch": 0.1531946608848838, + "grad_norm": 11.108535766601562, + "learning_rate": 9.793278116229721e-05, + "loss": 1.3118, + "step": 2261 + }, + { + "epoch": 0.15326241615285588, + "grad_norm": 7.281479358673096, + "learning_rate": 9.793141214319939e-05, + "loss": 0.9165, + "step": 2262 + }, + { + "epoch": 0.15333017142082797, + "grad_norm": 9.166728973388672, + "learning_rate": 9.793004312410159e-05, + "loss": 1.0845, + "step": 2263 + }, + { + "epoch": 0.15339792668880006, + "grad_norm": 11.539854049682617, + "learning_rate": 9.792867410500377e-05, + "loss": 1.1129, + "step": 2264 + }, + { + "epoch": 0.15346568195677213, + "grad_norm": 8.588869094848633, + "learning_rate": 9.792730508590595e-05, + "loss": 0.9698, + "step": 2265 + }, + { + "epoch": 0.15353343722474422, + "grad_norm": 8.270078659057617, + "learning_rate": 9.792593606680813e-05, + "loss": 1.2306, + "step": 2266 + }, + { + "epoch": 0.1536011924927163, + "grad_norm": 7.908688545227051, + "learning_rate": 9.792456704771032e-05, + "loss": 1.283, + "step": 2267 + }, + { + "epoch": 0.1536689477606884, + "grad_norm": 10.376410484313965, + "learning_rate": 9.79231980286125e-05, + "loss": 1.0246, + "step": 2268 + }, + { + "epoch": 0.15373670302866047, + "grad_norm": 9.517715454101562, + "learning_rate": 9.792182900951468e-05, + "loss": 0.9975, + "step": 2269 + }, + { + "epoch": 0.15380445829663256, + "grad_norm": 8.789438247680664, + "learning_rate": 9.792045999041688e-05, + "loss": 1.0886, + "step": 2270 + }, + { + "epoch": 0.15387221356460465, + "grad_norm": 9.649114608764648, + "learning_rate": 9.791909097131906e-05, + "loss": 1.1497, + "step": 2271 + }, + { + "epoch": 0.15393996883257674, + "grad_norm": 8.533876419067383, + "learning_rate": 9.791772195222124e-05, + "loss": 0.8701, + "step": 2272 + }, + { + "epoch": 0.1540077241005488, + "grad_norm": 10.64561653137207, + "learning_rate": 9.791635293312343e-05, + "loss": 1.2755, + "step": 2273 + }, + { + "epoch": 0.1540754793685209, + "grad_norm": 9.510658264160156, + "learning_rate": 9.791498391402561e-05, + "loss": 0.8772, + "step": 2274 + }, + { + "epoch": 0.154143234636493, + "grad_norm": 9.849981307983398, + "learning_rate": 9.791361489492779e-05, + "loss": 1.1689, + "step": 2275 + }, + { + "epoch": 0.15421098990446508, + "grad_norm": 8.152152061462402, + "learning_rate": 9.791224587582998e-05, + "loss": 0.8148, + "step": 2276 + }, + { + "epoch": 0.15427874517243717, + "grad_norm": 8.653456687927246, + "learning_rate": 9.791087685673216e-05, + "loss": 1.0061, + "step": 2277 + }, + { + "epoch": 0.15434650044040923, + "grad_norm": 6.928426742553711, + "learning_rate": 9.790950783763435e-05, + "loss": 0.9192, + "step": 2278 + }, + { + "epoch": 0.15441425570838133, + "grad_norm": 9.606708526611328, + "learning_rate": 9.790813881853653e-05, + "loss": 0.9956, + "step": 2279 + }, + { + "epoch": 0.15448201097635342, + "grad_norm": 9.42773151397705, + "learning_rate": 9.79067697994387e-05, + "loss": 0.9302, + "step": 2280 + }, + { + "epoch": 0.1545497662443255, + "grad_norm": 8.248319625854492, + "learning_rate": 9.79054007803409e-05, + "loss": 1.2318, + "step": 2281 + }, + { + "epoch": 0.15461752151229757, + "grad_norm": 9.706393241882324, + "learning_rate": 9.790403176124308e-05, + "loss": 1.1489, + "step": 2282 + }, + { + "epoch": 0.15468527678026966, + "grad_norm": 8.88716983795166, + "learning_rate": 9.790266274214526e-05, + "loss": 0.8686, + "step": 2283 + }, + { + "epoch": 0.15475303204824176, + "grad_norm": 8.596991539001465, + "learning_rate": 9.790129372304744e-05, + "loss": 1.0623, + "step": 2284 + }, + { + "epoch": 0.15482078731621385, + "grad_norm": 9.262425422668457, + "learning_rate": 9.789992470394962e-05, + "loss": 1.2067, + "step": 2285 + }, + { + "epoch": 0.1548885425841859, + "grad_norm": 7.772172927856445, + "learning_rate": 9.789855568485181e-05, + "loss": 0.9976, + "step": 2286 + }, + { + "epoch": 0.154956297852158, + "grad_norm": 9.320001602172852, + "learning_rate": 9.7897186665754e-05, + "loss": 1.1377, + "step": 2287 + }, + { + "epoch": 0.1550240531201301, + "grad_norm": 11.02434253692627, + "learning_rate": 9.789581764665618e-05, + "loss": 1.1365, + "step": 2288 + }, + { + "epoch": 0.1550918083881022, + "grad_norm": 9.90654182434082, + "learning_rate": 9.789444862755836e-05, + "loss": 1.208, + "step": 2289 + }, + { + "epoch": 0.15515956365607425, + "grad_norm": 9.591778755187988, + "learning_rate": 9.789307960846055e-05, + "loss": 1.2426, + "step": 2290 + }, + { + "epoch": 0.15522731892404634, + "grad_norm": 9.221457481384277, + "learning_rate": 9.789171058936273e-05, + "loss": 1.1449, + "step": 2291 + }, + { + "epoch": 0.15529507419201843, + "grad_norm": 7.294323444366455, + "learning_rate": 9.789034157026491e-05, + "loss": 0.8846, + "step": 2292 + }, + { + "epoch": 0.15536282945999053, + "grad_norm": 10.6463041305542, + "learning_rate": 9.788897255116709e-05, + "loss": 1.1397, + "step": 2293 + }, + { + "epoch": 0.1554305847279626, + "grad_norm": 9.412870407104492, + "learning_rate": 9.788760353206927e-05, + "loss": 1.1857, + "step": 2294 + }, + { + "epoch": 0.15549833999593468, + "grad_norm": 9.163009643554688, + "learning_rate": 9.788623451297147e-05, + "loss": 0.8907, + "step": 2295 + }, + { + "epoch": 0.15556609526390677, + "grad_norm": 8.157798767089844, + "learning_rate": 9.788486549387365e-05, + "loss": 0.893, + "step": 2296 + }, + { + "epoch": 0.15563385053187886, + "grad_norm": 9.155460357666016, + "learning_rate": 9.788349647477583e-05, + "loss": 1.0361, + "step": 2297 + }, + { + "epoch": 0.15570160579985093, + "grad_norm": 10.280989646911621, + "learning_rate": 9.7882127455678e-05, + "loss": 1.2953, + "step": 2298 + }, + { + "epoch": 0.15576936106782302, + "grad_norm": 9.654706954956055, + "learning_rate": 9.78807584365802e-05, + "loss": 0.9012, + "step": 2299 + }, + { + "epoch": 0.1558371163357951, + "grad_norm": 12.425939559936523, + "learning_rate": 9.787938941748238e-05, + "loss": 1.4103, + "step": 2300 + }, + { + "epoch": 0.1559048716037672, + "grad_norm": 9.888884544372559, + "learning_rate": 9.787802039838456e-05, + "loss": 1.1599, + "step": 2301 + }, + { + "epoch": 0.15597262687173927, + "grad_norm": 10.30229377746582, + "learning_rate": 9.787665137928674e-05, + "loss": 1.0505, + "step": 2302 + }, + { + "epoch": 0.15604038213971136, + "grad_norm": 11.208477973937988, + "learning_rate": 9.787528236018892e-05, + "loss": 1.009, + "step": 2303 + }, + { + "epoch": 0.15610813740768345, + "grad_norm": 9.264431953430176, + "learning_rate": 9.787391334109112e-05, + "loss": 1.1453, + "step": 2304 + }, + { + "epoch": 0.15617589267565554, + "grad_norm": 9.0980224609375, + "learning_rate": 9.78725443219933e-05, + "loss": 1.0824, + "step": 2305 + }, + { + "epoch": 0.1562436479436276, + "grad_norm": 8.346585273742676, + "learning_rate": 9.787117530289548e-05, + "loss": 1.0632, + "step": 2306 + }, + { + "epoch": 0.1563114032115997, + "grad_norm": 10.607507705688477, + "learning_rate": 9.786980628379766e-05, + "loss": 1.0102, + "step": 2307 + }, + { + "epoch": 0.1563791584795718, + "grad_norm": 8.189208984375, + "learning_rate": 9.786843726469985e-05, + "loss": 1.0872, + "step": 2308 + }, + { + "epoch": 0.15644691374754388, + "grad_norm": 9.84202766418457, + "learning_rate": 9.786706824560203e-05, + "loss": 0.9871, + "step": 2309 + }, + { + "epoch": 0.15651466901551594, + "grad_norm": 11.937589645385742, + "learning_rate": 9.786569922650421e-05, + "loss": 1.2255, + "step": 2310 + }, + { + "epoch": 0.15658242428348804, + "grad_norm": 9.855016708374023, + "learning_rate": 9.786433020740639e-05, + "loss": 1.0097, + "step": 2311 + }, + { + "epoch": 0.15665017955146013, + "grad_norm": 8.661060333251953, + "learning_rate": 9.786296118830857e-05, + "loss": 0.8732, + "step": 2312 + }, + { + "epoch": 0.15671793481943222, + "grad_norm": 10.57170295715332, + "learning_rate": 9.786159216921077e-05, + "loss": 0.862, + "step": 2313 + }, + { + "epoch": 0.15678569008740428, + "grad_norm": 7.759045600891113, + "learning_rate": 9.786022315011295e-05, + "loss": 0.77, + "step": 2314 + }, + { + "epoch": 0.15685344535537638, + "grad_norm": 10.758045196533203, + "learning_rate": 9.785885413101513e-05, + "loss": 0.8919, + "step": 2315 + }, + { + "epoch": 0.15692120062334847, + "grad_norm": 8.521660804748535, + "learning_rate": 9.78574851119173e-05, + "loss": 1.131, + "step": 2316 + }, + { + "epoch": 0.15698895589132056, + "grad_norm": 8.72917652130127, + "learning_rate": 9.78561160928195e-05, + "loss": 0.8359, + "step": 2317 + }, + { + "epoch": 0.15705671115929262, + "grad_norm": 11.679365158081055, + "learning_rate": 9.785474707372168e-05, + "loss": 1.0078, + "step": 2318 + }, + { + "epoch": 0.15712446642726471, + "grad_norm": 11.50632381439209, + "learning_rate": 9.785337805462386e-05, + "loss": 1.208, + "step": 2319 + }, + { + "epoch": 0.1571922216952368, + "grad_norm": 9.6107759475708, + "learning_rate": 9.785200903552605e-05, + "loss": 1.0967, + "step": 2320 + }, + { + "epoch": 0.1572599769632089, + "grad_norm": 8.629117012023926, + "learning_rate": 9.785064001642824e-05, + "loss": 1.0594, + "step": 2321 + }, + { + "epoch": 0.15732773223118096, + "grad_norm": 11.136920928955078, + "learning_rate": 9.784927099733042e-05, + "loss": 1.2874, + "step": 2322 + }, + { + "epoch": 0.15739548749915305, + "grad_norm": 11.097023963928223, + "learning_rate": 9.784790197823261e-05, + "loss": 1.1598, + "step": 2323 + }, + { + "epoch": 0.15746324276712514, + "grad_norm": 11.117433547973633, + "learning_rate": 9.784653295913479e-05, + "loss": 1.0601, + "step": 2324 + }, + { + "epoch": 0.15753099803509724, + "grad_norm": 10.152684211730957, + "learning_rate": 9.784516394003697e-05, + "loss": 1.2763, + "step": 2325 + }, + { + "epoch": 0.1575987533030693, + "grad_norm": 8.37531852722168, + "learning_rate": 9.784379492093915e-05, + "loss": 1.1055, + "step": 2326 + }, + { + "epoch": 0.1576665085710414, + "grad_norm": 10.463939666748047, + "learning_rate": 9.784242590184134e-05, + "loss": 1.3088, + "step": 2327 + }, + { + "epoch": 0.15773426383901348, + "grad_norm": 8.74315357208252, + "learning_rate": 9.784105688274352e-05, + "loss": 1.1165, + "step": 2328 + }, + { + "epoch": 0.15780201910698557, + "grad_norm": 8.691280364990234, + "learning_rate": 9.78396878636457e-05, + "loss": 1.1019, + "step": 2329 + }, + { + "epoch": 0.15786977437495767, + "grad_norm": 10.424938201904297, + "learning_rate": 9.783831884454789e-05, + "loss": 1.1957, + "step": 2330 + }, + { + "epoch": 0.15793752964292973, + "grad_norm": 6.867722034454346, + "learning_rate": 9.783694982545008e-05, + "loss": 1.0476, + "step": 2331 + }, + { + "epoch": 0.15800528491090182, + "grad_norm": 9.434804916381836, + "learning_rate": 9.783558080635226e-05, + "loss": 1.0247, + "step": 2332 + }, + { + "epoch": 0.1580730401788739, + "grad_norm": 7.771170616149902, + "learning_rate": 9.783421178725444e-05, + "loss": 1.2582, + "step": 2333 + }, + { + "epoch": 0.158140795446846, + "grad_norm": 8.366608619689941, + "learning_rate": 9.783284276815662e-05, + "loss": 1.1078, + "step": 2334 + }, + { + "epoch": 0.15820855071481807, + "grad_norm": 8.85851764678955, + "learning_rate": 9.78314737490588e-05, + "loss": 1.2405, + "step": 2335 + }, + { + "epoch": 0.15827630598279016, + "grad_norm": 9.134325981140137, + "learning_rate": 9.7830104729961e-05, + "loss": 1.2277, + "step": 2336 + }, + { + "epoch": 0.15834406125076225, + "grad_norm": 9.150130271911621, + "learning_rate": 9.782873571086317e-05, + "loss": 1.1355, + "step": 2337 + }, + { + "epoch": 0.15841181651873434, + "grad_norm": 8.687469482421875, + "learning_rate": 9.782736669176536e-05, + "loss": 1.0454, + "step": 2338 + }, + { + "epoch": 0.1584795717867064, + "grad_norm": 10.071285247802734, + "learning_rate": 9.782599767266754e-05, + "loss": 1.0041, + "step": 2339 + }, + { + "epoch": 0.1585473270546785, + "grad_norm": 8.373369216918945, + "learning_rate": 9.782462865356972e-05, + "loss": 0.8378, + "step": 2340 + }, + { + "epoch": 0.1586150823226506, + "grad_norm": 9.427014350891113, + "learning_rate": 9.782325963447191e-05, + "loss": 1.1413, + "step": 2341 + }, + { + "epoch": 0.15868283759062268, + "grad_norm": 8.38814640045166, + "learning_rate": 9.782189061537409e-05, + "loss": 1.0696, + "step": 2342 + }, + { + "epoch": 0.15875059285859475, + "grad_norm": 8.518644332885742, + "learning_rate": 9.782052159627627e-05, + "loss": 0.9814, + "step": 2343 + }, + { + "epoch": 0.15881834812656684, + "grad_norm": 9.926252365112305, + "learning_rate": 9.781915257717845e-05, + "loss": 1.1971, + "step": 2344 + }, + { + "epoch": 0.15888610339453893, + "grad_norm": 7.946019649505615, + "learning_rate": 9.781778355808064e-05, + "loss": 1.1902, + "step": 2345 + }, + { + "epoch": 0.15895385866251102, + "grad_norm": 8.686945915222168, + "learning_rate": 9.781641453898283e-05, + "loss": 0.9939, + "step": 2346 + }, + { + "epoch": 0.15902161393048309, + "grad_norm": 8.225680351257324, + "learning_rate": 9.7815045519885e-05, + "loss": 1.3665, + "step": 2347 + }, + { + "epoch": 0.15908936919845518, + "grad_norm": 10.381987571716309, + "learning_rate": 9.781367650078719e-05, + "loss": 1.09, + "step": 2348 + }, + { + "epoch": 0.15915712446642727, + "grad_norm": 8.57552719116211, + "learning_rate": 9.781230748168937e-05, + "loss": 0.8261, + "step": 2349 + }, + { + "epoch": 0.15922487973439936, + "grad_norm": 9.105220794677734, + "learning_rate": 9.781093846259156e-05, + "loss": 1.102, + "step": 2350 + }, + { + "epoch": 0.15929263500237142, + "grad_norm": 10.08092212677002, + "learning_rate": 9.780956944349374e-05, + "loss": 0.943, + "step": 2351 + }, + { + "epoch": 0.15936039027034352, + "grad_norm": 10.259852409362793, + "learning_rate": 9.780820042439592e-05, + "loss": 0.8822, + "step": 2352 + }, + { + "epoch": 0.1594281455383156, + "grad_norm": 8.31139087677002, + "learning_rate": 9.78068314052981e-05, + "loss": 1.2167, + "step": 2353 + }, + { + "epoch": 0.1594959008062877, + "grad_norm": 7.529703617095947, + "learning_rate": 9.78054623862003e-05, + "loss": 0.8913, + "step": 2354 + }, + { + "epoch": 0.15956365607425976, + "grad_norm": 8.792675971984863, + "learning_rate": 9.780409336710248e-05, + "loss": 0.966, + "step": 2355 + }, + { + "epoch": 0.15963141134223185, + "grad_norm": 9.329866409301758, + "learning_rate": 9.780272434800466e-05, + "loss": 1.1659, + "step": 2356 + }, + { + "epoch": 0.15969916661020395, + "grad_norm": 12.14089298248291, + "learning_rate": 9.780135532890684e-05, + "loss": 1.2019, + "step": 2357 + }, + { + "epoch": 0.15976692187817604, + "grad_norm": 9.12912654876709, + "learning_rate": 9.779998630980902e-05, + "loss": 1.1615, + "step": 2358 + }, + { + "epoch": 0.1598346771461481, + "grad_norm": 9.554464340209961, + "learning_rate": 9.779861729071121e-05, + "loss": 1.1695, + "step": 2359 + }, + { + "epoch": 0.1599024324141202, + "grad_norm": 9.317673683166504, + "learning_rate": 9.779724827161339e-05, + "loss": 0.9606, + "step": 2360 + }, + { + "epoch": 0.15997018768209229, + "grad_norm": 8.80395793914795, + "learning_rate": 9.779587925251557e-05, + "loss": 0.9693, + "step": 2361 + }, + { + "epoch": 0.16003794295006438, + "grad_norm": 11.990642547607422, + "learning_rate": 9.779451023341775e-05, + "loss": 1.2901, + "step": 2362 + }, + { + "epoch": 0.16010569821803644, + "grad_norm": 8.188547134399414, + "learning_rate": 9.779314121431995e-05, + "loss": 0.979, + "step": 2363 + }, + { + "epoch": 0.16017345348600853, + "grad_norm": 8.316620826721191, + "learning_rate": 9.779177219522213e-05, + "loss": 0.8601, + "step": 2364 + }, + { + "epoch": 0.16024120875398062, + "grad_norm": 7.58405876159668, + "learning_rate": 9.77904031761243e-05, + "loss": 1.1812, + "step": 2365 + }, + { + "epoch": 0.16030896402195272, + "grad_norm": 7.725598335266113, + "learning_rate": 9.77890341570265e-05, + "loss": 0.9335, + "step": 2366 + }, + { + "epoch": 0.16037671928992478, + "grad_norm": 8.6231107711792, + "learning_rate": 9.778766513792868e-05, + "loss": 1.0162, + "step": 2367 + }, + { + "epoch": 0.16044447455789687, + "grad_norm": 9.762526512145996, + "learning_rate": 9.778629611883086e-05, + "loss": 1.3186, + "step": 2368 + }, + { + "epoch": 0.16051222982586896, + "grad_norm": 11.384220123291016, + "learning_rate": 9.778492709973305e-05, + "loss": 1.2397, + "step": 2369 + }, + { + "epoch": 0.16057998509384105, + "grad_norm": 8.841899871826172, + "learning_rate": 9.778355808063523e-05, + "loss": 0.8708, + "step": 2370 + }, + { + "epoch": 0.16064774036181312, + "grad_norm": 7.778527736663818, + "learning_rate": 9.778218906153741e-05, + "loss": 0.9968, + "step": 2371 + }, + { + "epoch": 0.1607154956297852, + "grad_norm": 8.559181213378906, + "learning_rate": 9.77808200424396e-05, + "loss": 0.9759, + "step": 2372 + }, + { + "epoch": 0.1607832508977573, + "grad_norm": 10.273273468017578, + "learning_rate": 9.777945102334179e-05, + "loss": 0.9485, + "step": 2373 + }, + { + "epoch": 0.1608510061657294, + "grad_norm": 7.946044445037842, + "learning_rate": 9.777808200424397e-05, + "loss": 1.105, + "step": 2374 + }, + { + "epoch": 0.16091876143370146, + "grad_norm": 9.917662620544434, + "learning_rate": 9.777671298514615e-05, + "loss": 1.0272, + "step": 2375 + }, + { + "epoch": 0.16098651670167355, + "grad_norm": 10.438239097595215, + "learning_rate": 9.777534396604833e-05, + "loss": 1.0197, + "step": 2376 + }, + { + "epoch": 0.16105427196964564, + "grad_norm": 8.799901962280273, + "learning_rate": 9.777397494695052e-05, + "loss": 1.1401, + "step": 2377 + }, + { + "epoch": 0.16112202723761773, + "grad_norm": 8.569243431091309, + "learning_rate": 9.77726059278527e-05, + "loss": 1.2019, + "step": 2378 + }, + { + "epoch": 0.1611897825055898, + "grad_norm": 10.793002128601074, + "learning_rate": 9.777123690875488e-05, + "loss": 1.0932, + "step": 2379 + }, + { + "epoch": 0.1612575377735619, + "grad_norm": 7.825850963592529, + "learning_rate": 9.776986788965707e-05, + "loss": 1.2226, + "step": 2380 + }, + { + "epoch": 0.16132529304153398, + "grad_norm": 8.780813217163086, + "learning_rate": 9.776849887055925e-05, + "loss": 0.7939, + "step": 2381 + }, + { + "epoch": 0.16139304830950607, + "grad_norm": 8.927638053894043, + "learning_rate": 9.776712985146144e-05, + "loss": 1.0551, + "step": 2382 + }, + { + "epoch": 0.16146080357747816, + "grad_norm": 9.08043384552002, + "learning_rate": 9.776576083236362e-05, + "loss": 1.0037, + "step": 2383 + }, + { + "epoch": 0.16152855884545023, + "grad_norm": 9.362268447875977, + "learning_rate": 9.77643918132658e-05, + "loss": 1.332, + "step": 2384 + }, + { + "epoch": 0.16159631411342232, + "grad_norm": 10.533197402954102, + "learning_rate": 9.776302279416798e-05, + "loss": 1.0966, + "step": 2385 + }, + { + "epoch": 0.1616640693813944, + "grad_norm": 9.579266548156738, + "learning_rate": 9.776165377507017e-05, + "loss": 0.9325, + "step": 2386 + }, + { + "epoch": 0.1617318246493665, + "grad_norm": 10.071297645568848, + "learning_rate": 9.776028475597235e-05, + "loss": 0.9496, + "step": 2387 + }, + { + "epoch": 0.16179957991733857, + "grad_norm": 9.243900299072266, + "learning_rate": 9.775891573687453e-05, + "loss": 1.13, + "step": 2388 + }, + { + "epoch": 0.16186733518531066, + "grad_norm": 9.327018737792969, + "learning_rate": 9.775754671777672e-05, + "loss": 1.2587, + "step": 2389 + }, + { + "epoch": 0.16193509045328275, + "grad_norm": 7.614950180053711, + "learning_rate": 9.77561776986789e-05, + "loss": 1.0848, + "step": 2390 + }, + { + "epoch": 0.16200284572125484, + "grad_norm": 9.86501407623291, + "learning_rate": 9.775480867958109e-05, + "loss": 1.2504, + "step": 2391 + }, + { + "epoch": 0.1620706009892269, + "grad_norm": 9.08303451538086, + "learning_rate": 9.775343966048327e-05, + "loss": 1.4359, + "step": 2392 + }, + { + "epoch": 0.162138356257199, + "grad_norm": 8.417489051818848, + "learning_rate": 9.775207064138545e-05, + "loss": 1.244, + "step": 2393 + }, + { + "epoch": 0.1622061115251711, + "grad_norm": 8.35366439819336, + "learning_rate": 9.775070162228763e-05, + "loss": 1.0706, + "step": 2394 + }, + { + "epoch": 0.16227386679314318, + "grad_norm": 9.732915878295898, + "learning_rate": 9.774933260318981e-05, + "loss": 1.4237, + "step": 2395 + }, + { + "epoch": 0.16234162206111524, + "grad_norm": 9.131365776062012, + "learning_rate": 9.7747963584092e-05, + "loss": 0.9207, + "step": 2396 + }, + { + "epoch": 0.16240937732908733, + "grad_norm": 8.93538761138916, + "learning_rate": 9.774659456499419e-05, + "loss": 1.0054, + "step": 2397 + }, + { + "epoch": 0.16247713259705943, + "grad_norm": 8.939055442810059, + "learning_rate": 9.774522554589637e-05, + "loss": 0.9782, + "step": 2398 + }, + { + "epoch": 0.16254488786503152, + "grad_norm": 9.251758575439453, + "learning_rate": 9.774385652679855e-05, + "loss": 1.0968, + "step": 2399 + }, + { + "epoch": 0.16261264313300358, + "grad_norm": 9.240782737731934, + "learning_rate": 9.774248750770074e-05, + "loss": 0.7982, + "step": 2400 + }, + { + "epoch": 0.16268039840097567, + "grad_norm": 8.697726249694824, + "learning_rate": 9.774111848860292e-05, + "loss": 1.0545, + "step": 2401 + }, + { + "epoch": 0.16274815366894776, + "grad_norm": 7.9362053871154785, + "learning_rate": 9.77397494695051e-05, + "loss": 0.8364, + "step": 2402 + }, + { + "epoch": 0.16281590893691986, + "grad_norm": 11.944025993347168, + "learning_rate": 9.773838045040728e-05, + "loss": 1.2313, + "step": 2403 + }, + { + "epoch": 0.16288366420489192, + "grad_norm": 9.496225357055664, + "learning_rate": 9.773701143130946e-05, + "loss": 1.1363, + "step": 2404 + }, + { + "epoch": 0.162951419472864, + "grad_norm": 12.96069622039795, + "learning_rate": 9.773564241221165e-05, + "loss": 1.1205, + "step": 2405 + }, + { + "epoch": 0.1630191747408361, + "grad_norm": 8.915671348571777, + "learning_rate": 9.773427339311384e-05, + "loss": 1.1149, + "step": 2406 + }, + { + "epoch": 0.1630869300088082, + "grad_norm": 10.23763370513916, + "learning_rate": 9.773290437401602e-05, + "loss": 1.1867, + "step": 2407 + }, + { + "epoch": 0.16315468527678026, + "grad_norm": 9.117730140686035, + "learning_rate": 9.77315353549182e-05, + "loss": 1.1792, + "step": 2408 + }, + { + "epoch": 0.16322244054475235, + "grad_norm": 9.380385398864746, + "learning_rate": 9.773016633582039e-05, + "loss": 1.1621, + "step": 2409 + }, + { + "epoch": 0.16329019581272444, + "grad_norm": 8.56508731842041, + "learning_rate": 9.772879731672257e-05, + "loss": 0.8372, + "step": 2410 + }, + { + "epoch": 0.16335795108069653, + "grad_norm": 11.707832336425781, + "learning_rate": 9.772742829762475e-05, + "loss": 1.3191, + "step": 2411 + }, + { + "epoch": 0.1634257063486686, + "grad_norm": 7.720577716827393, + "learning_rate": 9.772605927852694e-05, + "loss": 0.9617, + "step": 2412 + }, + { + "epoch": 0.1634934616166407, + "grad_norm": 8.586542129516602, + "learning_rate": 9.772469025942912e-05, + "loss": 1.0978, + "step": 2413 + }, + { + "epoch": 0.16356121688461278, + "grad_norm": 9.021394729614258, + "learning_rate": 9.77233212403313e-05, + "loss": 0.9358, + "step": 2414 + }, + { + "epoch": 0.16362897215258487, + "grad_norm": 9.078686714172363, + "learning_rate": 9.77219522212335e-05, + "loss": 1.3338, + "step": 2415 + }, + { + "epoch": 0.16369672742055694, + "grad_norm": 9.810312271118164, + "learning_rate": 9.772058320213568e-05, + "loss": 1.2472, + "step": 2416 + }, + { + "epoch": 0.16376448268852903, + "grad_norm": 8.933609962463379, + "learning_rate": 9.771921418303786e-05, + "loss": 0.9115, + "step": 2417 + }, + { + "epoch": 0.16383223795650112, + "grad_norm": 7.044286251068115, + "learning_rate": 9.771784516394005e-05, + "loss": 0.7983, + "step": 2418 + }, + { + "epoch": 0.1638999932244732, + "grad_norm": 11.711495399475098, + "learning_rate": 9.771647614484223e-05, + "loss": 1.096, + "step": 2419 + }, + { + "epoch": 0.16396774849244528, + "grad_norm": 9.31049633026123, + "learning_rate": 9.771510712574441e-05, + "loss": 1.2711, + "step": 2420 + }, + { + "epoch": 0.16403550376041737, + "grad_norm": 8.10503101348877, + "learning_rate": 9.77137381066466e-05, + "loss": 0.9095, + "step": 2421 + }, + { + "epoch": 0.16410325902838946, + "grad_norm": 7.915055751800537, + "learning_rate": 9.771236908754877e-05, + "loss": 1.0161, + "step": 2422 + }, + { + "epoch": 0.16417101429636155, + "grad_norm": 8.185515403747559, + "learning_rate": 9.771100006845097e-05, + "loss": 1.1109, + "step": 2423 + }, + { + "epoch": 0.16423876956433361, + "grad_norm": 9.960200309753418, + "learning_rate": 9.770963104935315e-05, + "loss": 0.9757, + "step": 2424 + }, + { + "epoch": 0.1643065248323057, + "grad_norm": 9.646814346313477, + "learning_rate": 9.770826203025533e-05, + "loss": 1.0835, + "step": 2425 + }, + { + "epoch": 0.1643742801002778, + "grad_norm": 9.701393127441406, + "learning_rate": 9.770689301115751e-05, + "loss": 1.0717, + "step": 2426 + }, + { + "epoch": 0.1644420353682499, + "grad_norm": 7.887824058532715, + "learning_rate": 9.770552399205969e-05, + "loss": 1.0937, + "step": 2427 + }, + { + "epoch": 0.16450979063622195, + "grad_norm": 10.94339370727539, + "learning_rate": 9.770415497296188e-05, + "loss": 1.2766, + "step": 2428 + }, + { + "epoch": 0.16457754590419404, + "grad_norm": 10.051490783691406, + "learning_rate": 9.770278595386406e-05, + "loss": 1.248, + "step": 2429 + }, + { + "epoch": 0.16464530117216614, + "grad_norm": 8.380006790161133, + "learning_rate": 9.770141693476624e-05, + "loss": 1.1657, + "step": 2430 + }, + { + "epoch": 0.16471305644013823, + "grad_norm": 8.077753067016602, + "learning_rate": 9.770004791566843e-05, + "loss": 0.9511, + "step": 2431 + }, + { + "epoch": 0.1647808117081103, + "grad_norm": 8.744999885559082, + "learning_rate": 9.769867889657062e-05, + "loss": 1.0736, + "step": 2432 + }, + { + "epoch": 0.16484856697608238, + "grad_norm": 8.203909873962402, + "learning_rate": 9.76973098774728e-05, + "loss": 1.208, + "step": 2433 + }, + { + "epoch": 0.16491632224405448, + "grad_norm": 9.462398529052734, + "learning_rate": 9.769594085837498e-05, + "loss": 1.0011, + "step": 2434 + }, + { + "epoch": 0.16498407751202657, + "grad_norm": 11.190359115600586, + "learning_rate": 9.769457183927716e-05, + "loss": 1.139, + "step": 2435 + }, + { + "epoch": 0.16505183277999866, + "grad_norm": 10.454118728637695, + "learning_rate": 9.769320282017934e-05, + "loss": 1.0799, + "step": 2436 + }, + { + "epoch": 0.16511958804797072, + "grad_norm": 14.411054611206055, + "learning_rate": 9.769183380108153e-05, + "loss": 1.0369, + "step": 2437 + }, + { + "epoch": 0.16518734331594281, + "grad_norm": 11.42679214477539, + "learning_rate": 9.769046478198371e-05, + "loss": 1.2486, + "step": 2438 + }, + { + "epoch": 0.1652550985839149, + "grad_norm": 10.520325660705566, + "learning_rate": 9.76890957628859e-05, + "loss": 1.2355, + "step": 2439 + }, + { + "epoch": 0.165322853851887, + "grad_norm": 7.958998680114746, + "learning_rate": 9.768772674378808e-05, + "loss": 1.0822, + "step": 2440 + }, + { + "epoch": 0.16539060911985906, + "grad_norm": 8.649806022644043, + "learning_rate": 9.768635772469027e-05, + "loss": 0.9439, + "step": 2441 + }, + { + "epoch": 0.16545836438783115, + "grad_norm": 8.539712905883789, + "learning_rate": 9.768498870559245e-05, + "loss": 0.9541, + "step": 2442 + }, + { + "epoch": 0.16552611965580324, + "grad_norm": 11.40131950378418, + "learning_rate": 9.768361968649463e-05, + "loss": 1.0568, + "step": 2443 + }, + { + "epoch": 0.16559387492377534, + "grad_norm": 8.333579063415527, + "learning_rate": 9.768225066739681e-05, + "loss": 0.924, + "step": 2444 + }, + { + "epoch": 0.1656616301917474, + "grad_norm": 9.02564525604248, + "learning_rate": 9.768088164829899e-05, + "loss": 1.1901, + "step": 2445 + }, + { + "epoch": 0.1657293854597195, + "grad_norm": 9.721306800842285, + "learning_rate": 9.767951262920118e-05, + "loss": 1.2313, + "step": 2446 + }, + { + "epoch": 0.16579714072769158, + "grad_norm": 9.876781463623047, + "learning_rate": 9.767814361010336e-05, + "loss": 1.0165, + "step": 2447 + }, + { + "epoch": 0.16586489599566367, + "grad_norm": 11.69865894317627, + "learning_rate": 9.767677459100555e-05, + "loss": 1.1301, + "step": 2448 + }, + { + "epoch": 0.16593265126363574, + "grad_norm": 11.38391399383545, + "learning_rate": 9.767540557190773e-05, + "loss": 1.1522, + "step": 2449 + }, + { + "epoch": 0.16600040653160783, + "grad_norm": 9.18020248413086, + "learning_rate": 9.76740365528099e-05, + "loss": 1.0515, + "step": 2450 + }, + { + "epoch": 0.16606816179957992, + "grad_norm": 9.30802059173584, + "learning_rate": 9.76726675337121e-05, + "loss": 1.0789, + "step": 2451 + }, + { + "epoch": 0.166135917067552, + "grad_norm": 9.58259391784668, + "learning_rate": 9.767129851461428e-05, + "loss": 1.1299, + "step": 2452 + }, + { + "epoch": 0.16620367233552408, + "grad_norm": 9.137594223022461, + "learning_rate": 9.766992949551646e-05, + "loss": 1.0854, + "step": 2453 + }, + { + "epoch": 0.16627142760349617, + "grad_norm": 12.410299301147461, + "learning_rate": 9.766856047641864e-05, + "loss": 1.1717, + "step": 2454 + }, + { + "epoch": 0.16633918287146826, + "grad_norm": 9.016322135925293, + "learning_rate": 9.766719145732083e-05, + "loss": 1.1028, + "step": 2455 + }, + { + "epoch": 0.16640693813944035, + "grad_norm": 10.166184425354004, + "learning_rate": 9.766582243822301e-05, + "loss": 1.0626, + "step": 2456 + }, + { + "epoch": 0.16647469340741242, + "grad_norm": 9.030965805053711, + "learning_rate": 9.76644534191252e-05, + "loss": 1.1761, + "step": 2457 + }, + { + "epoch": 0.1665424486753845, + "grad_norm": 9.294576644897461, + "learning_rate": 9.766308440002739e-05, + "loss": 0.943, + "step": 2458 + }, + { + "epoch": 0.1666102039433566, + "grad_norm": 6.732856273651123, + "learning_rate": 9.766171538092957e-05, + "loss": 0.7342, + "step": 2459 + }, + { + "epoch": 0.1666779592113287, + "grad_norm": 8.178942680358887, + "learning_rate": 9.766034636183175e-05, + "loss": 0.7889, + "step": 2460 + }, + { + "epoch": 0.16674571447930076, + "grad_norm": 7.771929740905762, + "learning_rate": 9.765897734273394e-05, + "loss": 1.1129, + "step": 2461 + }, + { + "epoch": 0.16681346974727285, + "grad_norm": 8.695874214172363, + "learning_rate": 9.765760832363612e-05, + "loss": 1.1968, + "step": 2462 + }, + { + "epoch": 0.16688122501524494, + "grad_norm": 10.18800163269043, + "learning_rate": 9.76562393045383e-05, + "loss": 1.0835, + "step": 2463 + }, + { + "epoch": 0.16694898028321703, + "grad_norm": 9.310625076293945, + "learning_rate": 9.76548702854405e-05, + "loss": 1.2042, + "step": 2464 + }, + { + "epoch": 0.1670167355511891, + "grad_norm": 11.684195518493652, + "learning_rate": 9.765350126634268e-05, + "loss": 1.2518, + "step": 2465 + }, + { + "epoch": 0.16708449081916119, + "grad_norm": 10.9056978225708, + "learning_rate": 9.765213224724486e-05, + "loss": 1.2465, + "step": 2466 + }, + { + "epoch": 0.16715224608713328, + "grad_norm": 9.871830940246582, + "learning_rate": 9.765076322814704e-05, + "loss": 1.3192, + "step": 2467 + }, + { + "epoch": 0.16722000135510537, + "grad_norm": 7.725397109985352, + "learning_rate": 9.764939420904922e-05, + "loss": 0.9082, + "step": 2468 + }, + { + "epoch": 0.16728775662307743, + "grad_norm": 8.7415189743042, + "learning_rate": 9.764802518995141e-05, + "loss": 0.8569, + "step": 2469 + }, + { + "epoch": 0.16735551189104952, + "grad_norm": 7.8990888595581055, + "learning_rate": 9.76466561708536e-05, + "loss": 1.0175, + "step": 2470 + }, + { + "epoch": 0.16742326715902162, + "grad_norm": 8.688074111938477, + "learning_rate": 9.764528715175577e-05, + "loss": 1.0131, + "step": 2471 + }, + { + "epoch": 0.1674910224269937, + "grad_norm": 7.902133941650391, + "learning_rate": 9.764391813265795e-05, + "loss": 0.8956, + "step": 2472 + }, + { + "epoch": 0.16755877769496577, + "grad_norm": 9.179505348205566, + "learning_rate": 9.764254911356013e-05, + "loss": 1.175, + "step": 2473 + }, + { + "epoch": 0.16762653296293786, + "grad_norm": 10.701058387756348, + "learning_rate": 9.764118009446233e-05, + "loss": 1.003, + "step": 2474 + }, + { + "epoch": 0.16769428823090995, + "grad_norm": 8.76916217803955, + "learning_rate": 9.763981107536451e-05, + "loss": 1.102, + "step": 2475 + }, + { + "epoch": 0.16776204349888205, + "grad_norm": 8.682199478149414, + "learning_rate": 9.763844205626669e-05, + "loss": 1.2042, + "step": 2476 + }, + { + "epoch": 0.1678297987668541, + "grad_norm": 9.789544105529785, + "learning_rate": 9.763707303716887e-05, + "loss": 1.2469, + "step": 2477 + }, + { + "epoch": 0.1678975540348262, + "grad_norm": 8.250391960144043, + "learning_rate": 9.763570401807106e-05, + "loss": 0.9742, + "step": 2478 + }, + { + "epoch": 0.1679653093027983, + "grad_norm": 8.938610076904297, + "learning_rate": 9.763433499897324e-05, + "loss": 0.9968, + "step": 2479 + }, + { + "epoch": 0.16803306457077039, + "grad_norm": 9.956622123718262, + "learning_rate": 9.763296597987542e-05, + "loss": 1.0464, + "step": 2480 + }, + { + "epoch": 0.16810081983874245, + "grad_norm": 8.582858085632324, + "learning_rate": 9.76315969607776e-05, + "loss": 1.1915, + "step": 2481 + }, + { + "epoch": 0.16816857510671454, + "grad_norm": 7.73312520980835, + "learning_rate": 9.763022794167979e-05, + "loss": 0.8506, + "step": 2482 + }, + { + "epoch": 0.16823633037468663, + "grad_norm": 8.986891746520996, + "learning_rate": 9.762885892258198e-05, + "loss": 0.8959, + "step": 2483 + }, + { + "epoch": 0.16830408564265872, + "grad_norm": 10.047099113464355, + "learning_rate": 9.762748990348416e-05, + "loss": 1.24, + "step": 2484 + }, + { + "epoch": 0.1683718409106308, + "grad_norm": 8.517911911010742, + "learning_rate": 9.762612088438634e-05, + "loss": 0.8171, + "step": 2485 + }, + { + "epoch": 0.16843959617860288, + "grad_norm": 9.586174964904785, + "learning_rate": 9.762475186528852e-05, + "loss": 1.1603, + "step": 2486 + }, + { + "epoch": 0.16850735144657497, + "grad_norm": 9.85086441040039, + "learning_rate": 9.762338284619071e-05, + "loss": 0.9157, + "step": 2487 + }, + { + "epoch": 0.16857510671454706, + "grad_norm": 12.023639678955078, + "learning_rate": 9.76220138270929e-05, + "loss": 1.2032, + "step": 2488 + }, + { + "epoch": 0.16864286198251915, + "grad_norm": 8.932641983032227, + "learning_rate": 9.762064480799507e-05, + "loss": 1.0678, + "step": 2489 + }, + { + "epoch": 0.16871061725049122, + "grad_norm": 10.568282127380371, + "learning_rate": 9.761927578889725e-05, + "loss": 1.3213, + "step": 2490 + }, + { + "epoch": 0.1687783725184633, + "grad_norm": 9.116564750671387, + "learning_rate": 9.761790676979944e-05, + "loss": 1.0525, + "step": 2491 + }, + { + "epoch": 0.1688461277864354, + "grad_norm": 8.192644119262695, + "learning_rate": 9.761653775070163e-05, + "loss": 0.9554, + "step": 2492 + }, + { + "epoch": 0.1689138830544075, + "grad_norm": 9.146562576293945, + "learning_rate": 9.761516873160381e-05, + "loss": 0.8852, + "step": 2493 + }, + { + "epoch": 0.16898163832237956, + "grad_norm": 8.82610034942627, + "learning_rate": 9.761379971250599e-05, + "loss": 1.0521, + "step": 2494 + }, + { + "epoch": 0.16904939359035165, + "grad_norm": 9.051412582397461, + "learning_rate": 9.761243069340817e-05, + "loss": 0.913, + "step": 2495 + }, + { + "epoch": 0.16911714885832374, + "grad_norm": 8.97696304321289, + "learning_rate": 9.761106167431036e-05, + "loss": 0.887, + "step": 2496 + }, + { + "epoch": 0.16918490412629583, + "grad_norm": 10.083110809326172, + "learning_rate": 9.760969265521254e-05, + "loss": 1.3521, + "step": 2497 + }, + { + "epoch": 0.1692526593942679, + "grad_norm": 7.585256099700928, + "learning_rate": 9.760832363611472e-05, + "loss": 0.9001, + "step": 2498 + }, + { + "epoch": 0.16932041466224, + "grad_norm": 10.301995277404785, + "learning_rate": 9.76069546170169e-05, + "loss": 1.0915, + "step": 2499 + }, + { + "epoch": 0.16938816993021208, + "grad_norm": 8.517580032348633, + "learning_rate": 9.760558559791909e-05, + "loss": 1.0097, + "step": 2500 + }, + { + "epoch": 0.16945592519818417, + "grad_norm": 8.20002555847168, + "learning_rate": 9.760421657882128e-05, + "loss": 1.2988, + "step": 2501 + }, + { + "epoch": 0.16952368046615623, + "grad_norm": 9.705124855041504, + "learning_rate": 9.760284755972346e-05, + "loss": 1.2661, + "step": 2502 + }, + { + "epoch": 0.16959143573412833, + "grad_norm": 9.167060852050781, + "learning_rate": 9.760147854062564e-05, + "loss": 1.0556, + "step": 2503 + }, + { + "epoch": 0.16965919100210042, + "grad_norm": 8.415916442871094, + "learning_rate": 9.760010952152783e-05, + "loss": 0.7964, + "step": 2504 + }, + { + "epoch": 0.1697269462700725, + "grad_norm": 7.626298427581787, + "learning_rate": 9.759874050243001e-05, + "loss": 0.9153, + "step": 2505 + }, + { + "epoch": 0.16979470153804457, + "grad_norm": 10.595001220703125, + "learning_rate": 9.75973714833322e-05, + "loss": 0.9911, + "step": 2506 + }, + { + "epoch": 0.16986245680601667, + "grad_norm": 7.564423084259033, + "learning_rate": 9.759600246423439e-05, + "loss": 0.9843, + "step": 2507 + }, + { + "epoch": 0.16993021207398876, + "grad_norm": 7.644829273223877, + "learning_rate": 9.759463344513657e-05, + "loss": 0.9116, + "step": 2508 + }, + { + "epoch": 0.16999796734196085, + "grad_norm": 7.54351282119751, + "learning_rate": 9.759326442603875e-05, + "loss": 1.1528, + "step": 2509 + }, + { + "epoch": 0.1700657226099329, + "grad_norm": 9.259818077087402, + "learning_rate": 9.759189540694094e-05, + "loss": 1.0867, + "step": 2510 + }, + { + "epoch": 0.170133477877905, + "grad_norm": 8.022993087768555, + "learning_rate": 9.759052638784312e-05, + "loss": 0.7524, + "step": 2511 + }, + { + "epoch": 0.1702012331458771, + "grad_norm": 8.264616966247559, + "learning_rate": 9.75891573687453e-05, + "loss": 0.9906, + "step": 2512 + }, + { + "epoch": 0.1702689884138492, + "grad_norm": 9.606420516967773, + "learning_rate": 9.758778834964748e-05, + "loss": 0.9287, + "step": 2513 + }, + { + "epoch": 0.17033674368182125, + "grad_norm": 8.019355773925781, + "learning_rate": 9.758641933054966e-05, + "loss": 0.879, + "step": 2514 + }, + { + "epoch": 0.17040449894979334, + "grad_norm": 9.819777488708496, + "learning_rate": 9.758505031145186e-05, + "loss": 1.3677, + "step": 2515 + }, + { + "epoch": 0.17047225421776543, + "grad_norm": 9.9795560836792, + "learning_rate": 9.758368129235404e-05, + "loss": 1.1614, + "step": 2516 + }, + { + "epoch": 0.17054000948573753, + "grad_norm": 9.51271915435791, + "learning_rate": 9.758231227325622e-05, + "loss": 1.0423, + "step": 2517 + }, + { + "epoch": 0.1706077647537096, + "grad_norm": 10.511359214782715, + "learning_rate": 9.75809432541584e-05, + "loss": 0.9492, + "step": 2518 + }, + { + "epoch": 0.17067552002168168, + "grad_norm": 9.61755657196045, + "learning_rate": 9.757957423506059e-05, + "loss": 1.1481, + "step": 2519 + }, + { + "epoch": 0.17074327528965377, + "grad_norm": 11.246142387390137, + "learning_rate": 9.757820521596277e-05, + "loss": 1.3981, + "step": 2520 + }, + { + "epoch": 0.17081103055762586, + "grad_norm": 9.273181915283203, + "learning_rate": 9.757683619686495e-05, + "loss": 0.9773, + "step": 2521 + }, + { + "epoch": 0.17087878582559793, + "grad_norm": 11.215822219848633, + "learning_rate": 9.757546717776713e-05, + "loss": 1.2303, + "step": 2522 + }, + { + "epoch": 0.17094654109357002, + "grad_norm": 10.139853477478027, + "learning_rate": 9.757409815866931e-05, + "loss": 1.0807, + "step": 2523 + }, + { + "epoch": 0.1710142963615421, + "grad_norm": 13.275606155395508, + "learning_rate": 9.757272913957151e-05, + "loss": 1.3326, + "step": 2524 + }, + { + "epoch": 0.1710820516295142, + "grad_norm": 8.32109260559082, + "learning_rate": 9.757136012047369e-05, + "loss": 0.9428, + "step": 2525 + }, + { + "epoch": 0.17114980689748627, + "grad_norm": 9.275816917419434, + "learning_rate": 9.756999110137587e-05, + "loss": 1.1531, + "step": 2526 + }, + { + "epoch": 0.17121756216545836, + "grad_norm": 10.61928939819336, + "learning_rate": 9.756862208227805e-05, + "loss": 0.9422, + "step": 2527 + }, + { + "epoch": 0.17128531743343045, + "grad_norm": 7.793631553649902, + "learning_rate": 9.756725306318023e-05, + "loss": 1.216, + "step": 2528 + }, + { + "epoch": 0.17135307270140254, + "grad_norm": 7.453477382659912, + "learning_rate": 9.756588404408242e-05, + "loss": 0.7787, + "step": 2529 + }, + { + "epoch": 0.1714208279693746, + "grad_norm": 10.39784049987793, + "learning_rate": 9.75645150249846e-05, + "loss": 1.3171, + "step": 2530 + }, + { + "epoch": 0.1714885832373467, + "grad_norm": 8.57040786743164, + "learning_rate": 9.756314600588678e-05, + "loss": 0.8319, + "step": 2531 + }, + { + "epoch": 0.1715563385053188, + "grad_norm": 8.696785926818848, + "learning_rate": 9.756177698678896e-05, + "loss": 1.0078, + "step": 2532 + }, + { + "epoch": 0.17162409377329088, + "grad_norm": 8.212169647216797, + "learning_rate": 9.756040796769116e-05, + "loss": 0.7851, + "step": 2533 + }, + { + "epoch": 0.17169184904126294, + "grad_norm": 10.94201946258545, + "learning_rate": 9.755903894859334e-05, + "loss": 1.0065, + "step": 2534 + }, + { + "epoch": 0.17175960430923504, + "grad_norm": 12.041540145874023, + "learning_rate": 9.755766992949552e-05, + "loss": 1.2938, + "step": 2535 + }, + { + "epoch": 0.17182735957720713, + "grad_norm": 9.289467811584473, + "learning_rate": 9.75563009103977e-05, + "loss": 1.1521, + "step": 2536 + }, + { + "epoch": 0.17189511484517922, + "grad_norm": 9.231005668640137, + "learning_rate": 9.755493189129988e-05, + "loss": 0.9, + "step": 2537 + }, + { + "epoch": 0.17196287011315128, + "grad_norm": 8.934699058532715, + "learning_rate": 9.755356287220207e-05, + "loss": 1.2817, + "step": 2538 + }, + { + "epoch": 0.17203062538112338, + "grad_norm": 10.096917152404785, + "learning_rate": 9.755219385310425e-05, + "loss": 1.1587, + "step": 2539 + }, + { + "epoch": 0.17209838064909547, + "grad_norm": 8.864645004272461, + "learning_rate": 9.755082483400643e-05, + "loss": 1.1781, + "step": 2540 + }, + { + "epoch": 0.17216613591706756, + "grad_norm": 10.964715957641602, + "learning_rate": 9.754945581490861e-05, + "loss": 1.2729, + "step": 2541 + }, + { + "epoch": 0.17223389118503965, + "grad_norm": 10.845703125, + "learning_rate": 9.754808679581081e-05, + "loss": 1.4475, + "step": 2542 + }, + { + "epoch": 0.17230164645301171, + "grad_norm": 9.800530433654785, + "learning_rate": 9.754671777671299e-05, + "loss": 1.0296, + "step": 2543 + }, + { + "epoch": 0.1723694017209838, + "grad_norm": 8.789834022521973, + "learning_rate": 9.754534875761517e-05, + "loss": 0.9573, + "step": 2544 + }, + { + "epoch": 0.1724371569889559, + "grad_norm": 7.731616973876953, + "learning_rate": 9.754397973851735e-05, + "loss": 0.9875, + "step": 2545 + }, + { + "epoch": 0.172504912256928, + "grad_norm": 8.942553520202637, + "learning_rate": 9.754261071941953e-05, + "loss": 1.1897, + "step": 2546 + }, + { + "epoch": 0.17257266752490005, + "grad_norm": 7.468856334686279, + "learning_rate": 9.754124170032172e-05, + "loss": 0.8869, + "step": 2547 + }, + { + "epoch": 0.17264042279287214, + "grad_norm": 8.798864364624023, + "learning_rate": 9.75398726812239e-05, + "loss": 1.0034, + "step": 2548 + }, + { + "epoch": 0.17270817806084424, + "grad_norm": 8.579094886779785, + "learning_rate": 9.753850366212608e-05, + "loss": 1.0211, + "step": 2549 + }, + { + "epoch": 0.17277593332881633, + "grad_norm": 10.90807819366455, + "learning_rate": 9.753713464302827e-05, + "loss": 1.3639, + "step": 2550 + }, + { + "epoch": 0.1728436885967884, + "grad_norm": 8.337204933166504, + "learning_rate": 9.753576562393046e-05, + "loss": 1.323, + "step": 2551 + }, + { + "epoch": 0.17291144386476048, + "grad_norm": 8.731851577758789, + "learning_rate": 9.753439660483264e-05, + "loss": 0.8805, + "step": 2552 + }, + { + "epoch": 0.17297919913273258, + "grad_norm": 9.541427612304688, + "learning_rate": 9.753302758573482e-05, + "loss": 1.0368, + "step": 2553 + }, + { + "epoch": 0.17304695440070467, + "grad_norm": 7.938154697418213, + "learning_rate": 9.753165856663701e-05, + "loss": 0.9128, + "step": 2554 + }, + { + "epoch": 0.17311470966867673, + "grad_norm": 8.111212730407715, + "learning_rate": 9.75302895475392e-05, + "loss": 0.9857, + "step": 2555 + }, + { + "epoch": 0.17318246493664882, + "grad_norm": 8.583324432373047, + "learning_rate": 9.752892052844139e-05, + "loss": 1.0975, + "step": 2556 + }, + { + "epoch": 0.1732502202046209, + "grad_norm": 8.180643081665039, + "learning_rate": 9.752755150934357e-05, + "loss": 1.21, + "step": 2557 + }, + { + "epoch": 0.173317975472593, + "grad_norm": 9.740217208862305, + "learning_rate": 9.752618249024575e-05, + "loss": 1.18, + "step": 2558 + }, + { + "epoch": 0.17338573074056507, + "grad_norm": 8.725831031799316, + "learning_rate": 9.752481347114793e-05, + "loss": 0.9816, + "step": 2559 + }, + { + "epoch": 0.17345348600853716, + "grad_norm": 10.786824226379395, + "learning_rate": 9.752344445205011e-05, + "loss": 1.0402, + "step": 2560 + }, + { + "epoch": 0.17352124127650925, + "grad_norm": 8.91720962524414, + "learning_rate": 9.75220754329523e-05, + "loss": 1.2357, + "step": 2561 + }, + { + "epoch": 0.17358899654448134, + "grad_norm": 9.02492618560791, + "learning_rate": 9.752070641385448e-05, + "loss": 1.1264, + "step": 2562 + }, + { + "epoch": 0.1736567518124534, + "grad_norm": 7.917794227600098, + "learning_rate": 9.751933739475666e-05, + "loss": 1.0937, + "step": 2563 + }, + { + "epoch": 0.1737245070804255, + "grad_norm": 11.543112754821777, + "learning_rate": 9.751796837565884e-05, + "loss": 1.0851, + "step": 2564 + }, + { + "epoch": 0.1737922623483976, + "grad_norm": 9.114936828613281, + "learning_rate": 9.751659935656104e-05, + "loss": 0.9565, + "step": 2565 + }, + { + "epoch": 0.17386001761636968, + "grad_norm": 12.322575569152832, + "learning_rate": 9.751523033746322e-05, + "loss": 1.0356, + "step": 2566 + }, + { + "epoch": 0.17392777288434175, + "grad_norm": 11.39137077331543, + "learning_rate": 9.75138613183654e-05, + "loss": 1.2141, + "step": 2567 + }, + { + "epoch": 0.17399552815231384, + "grad_norm": 9.505644798278809, + "learning_rate": 9.751249229926758e-05, + "loss": 0.8307, + "step": 2568 + }, + { + "epoch": 0.17406328342028593, + "grad_norm": 11.05932331085205, + "learning_rate": 9.751112328016976e-05, + "loss": 1.1915, + "step": 2569 + }, + { + "epoch": 0.17413103868825802, + "grad_norm": 11.104498863220215, + "learning_rate": 9.750975426107195e-05, + "loss": 1.0032, + "step": 2570 + }, + { + "epoch": 0.17419879395623009, + "grad_norm": 11.366668701171875, + "learning_rate": 9.750838524197413e-05, + "loss": 1.0857, + "step": 2571 + }, + { + "epoch": 0.17426654922420218, + "grad_norm": 8.775167465209961, + "learning_rate": 9.750701622287631e-05, + "loss": 1.1545, + "step": 2572 + }, + { + "epoch": 0.17433430449217427, + "grad_norm": 8.820073127746582, + "learning_rate": 9.75056472037785e-05, + "loss": 1.0528, + "step": 2573 + }, + { + "epoch": 0.17440205976014636, + "grad_norm": 9.53591537475586, + "learning_rate": 9.750427818468069e-05, + "loss": 1.0562, + "step": 2574 + }, + { + "epoch": 0.17446981502811842, + "grad_norm": 10.08950138092041, + "learning_rate": 9.750290916558287e-05, + "loss": 1.0456, + "step": 2575 + }, + { + "epoch": 0.17453757029609052, + "grad_norm": 8.222607612609863, + "learning_rate": 9.750154014648505e-05, + "loss": 0.9168, + "step": 2576 + }, + { + "epoch": 0.1746053255640626, + "grad_norm": 10.067093849182129, + "learning_rate": 9.750017112738723e-05, + "loss": 0.9952, + "step": 2577 + }, + { + "epoch": 0.1746730808320347, + "grad_norm": 9.726996421813965, + "learning_rate": 9.749880210828941e-05, + "loss": 0.9428, + "step": 2578 + }, + { + "epoch": 0.17474083610000676, + "grad_norm": 9.970647811889648, + "learning_rate": 9.74974330891916e-05, + "loss": 1.1838, + "step": 2579 + }, + { + "epoch": 0.17480859136797886, + "grad_norm": 7.8667893409729, + "learning_rate": 9.749606407009378e-05, + "loss": 1.0341, + "step": 2580 + }, + { + "epoch": 0.17487634663595095, + "grad_norm": 9.227079391479492, + "learning_rate": 9.749469505099596e-05, + "loss": 1.0515, + "step": 2581 + }, + { + "epoch": 0.17494410190392304, + "grad_norm": 9.198224067687988, + "learning_rate": 9.749332603189814e-05, + "loss": 1.0935, + "step": 2582 + }, + { + "epoch": 0.1750118571718951, + "grad_norm": 11.105298042297363, + "learning_rate": 9.749195701280032e-05, + "loss": 1.2115, + "step": 2583 + }, + { + "epoch": 0.1750796124398672, + "grad_norm": 8.815799713134766, + "learning_rate": 9.749058799370252e-05, + "loss": 1.1308, + "step": 2584 + }, + { + "epoch": 0.17514736770783929, + "grad_norm": 10.571581840515137, + "learning_rate": 9.74892189746047e-05, + "loss": 1.1982, + "step": 2585 + }, + { + "epoch": 0.17521512297581138, + "grad_norm": 7.368075370788574, + "learning_rate": 9.748784995550688e-05, + "loss": 0.9352, + "step": 2586 + }, + { + "epoch": 0.17528287824378344, + "grad_norm": 7.103427410125732, + "learning_rate": 9.748648093640906e-05, + "loss": 1.0578, + "step": 2587 + }, + { + "epoch": 0.17535063351175553, + "grad_norm": 7.678786754608154, + "learning_rate": 9.748511191731125e-05, + "loss": 0.8549, + "step": 2588 + }, + { + "epoch": 0.17541838877972762, + "grad_norm": 9.514643669128418, + "learning_rate": 9.748374289821343e-05, + "loss": 1.1835, + "step": 2589 + }, + { + "epoch": 0.17548614404769972, + "grad_norm": 7.552379131317139, + "learning_rate": 9.748237387911561e-05, + "loss": 0.9425, + "step": 2590 + }, + { + "epoch": 0.17555389931567178, + "grad_norm": 7.663018226623535, + "learning_rate": 9.74810048600178e-05, + "loss": 1.0948, + "step": 2591 + }, + { + "epoch": 0.17562165458364387, + "grad_norm": 10.293536186218262, + "learning_rate": 9.747963584091997e-05, + "loss": 1.1636, + "step": 2592 + }, + { + "epoch": 0.17568940985161596, + "grad_norm": 9.024083137512207, + "learning_rate": 9.747826682182217e-05, + "loss": 1.1157, + "step": 2593 + }, + { + "epoch": 0.17575716511958805, + "grad_norm": 8.870935440063477, + "learning_rate": 9.747689780272435e-05, + "loss": 0.9634, + "step": 2594 + }, + { + "epoch": 0.17582492038756015, + "grad_norm": 8.553323745727539, + "learning_rate": 9.747552878362653e-05, + "loss": 1.1515, + "step": 2595 + }, + { + "epoch": 0.1758926756555322, + "grad_norm": 9.13661003112793, + "learning_rate": 9.747415976452871e-05, + "loss": 0.9267, + "step": 2596 + }, + { + "epoch": 0.1759604309235043, + "grad_norm": 8.66097640991211, + "learning_rate": 9.74727907454309e-05, + "loss": 0.6839, + "step": 2597 + }, + { + "epoch": 0.1760281861914764, + "grad_norm": 10.07618236541748, + "learning_rate": 9.747142172633308e-05, + "loss": 1.2822, + "step": 2598 + }, + { + "epoch": 0.17609594145944849, + "grad_norm": 9.136283874511719, + "learning_rate": 9.747005270723526e-05, + "loss": 1.0053, + "step": 2599 + }, + { + "epoch": 0.17616369672742055, + "grad_norm": 9.42113208770752, + "learning_rate": 9.746868368813746e-05, + "loss": 1.15, + "step": 2600 + }, + { + "epoch": 0.17623145199539264, + "grad_norm": 8.789713859558105, + "learning_rate": 9.746731466903964e-05, + "loss": 0.8504, + "step": 2601 + }, + { + "epoch": 0.17629920726336473, + "grad_norm": 8.704665184020996, + "learning_rate": 9.746594564994182e-05, + "loss": 1.1516, + "step": 2602 + }, + { + "epoch": 0.17636696253133682, + "grad_norm": 8.030630111694336, + "learning_rate": 9.746457663084401e-05, + "loss": 1.044, + "step": 2603 + }, + { + "epoch": 0.1764347177993089, + "grad_norm": 8.640777587890625, + "learning_rate": 9.746320761174619e-05, + "loss": 1.0013, + "step": 2604 + }, + { + "epoch": 0.17650247306728098, + "grad_norm": 7.806771278381348, + "learning_rate": 9.746183859264837e-05, + "loss": 1.0678, + "step": 2605 + }, + { + "epoch": 0.17657022833525307, + "grad_norm": 8.883776664733887, + "learning_rate": 9.746046957355055e-05, + "loss": 0.9277, + "step": 2606 + }, + { + "epoch": 0.17663798360322516, + "grad_norm": 7.539346694946289, + "learning_rate": 9.745910055445275e-05, + "loss": 1.0437, + "step": 2607 + }, + { + "epoch": 0.17670573887119723, + "grad_norm": 9.873644828796387, + "learning_rate": 9.745773153535493e-05, + "loss": 1.206, + "step": 2608 + }, + { + "epoch": 0.17677349413916932, + "grad_norm": 10.031026840209961, + "learning_rate": 9.745636251625711e-05, + "loss": 1.1934, + "step": 2609 + }, + { + "epoch": 0.1768412494071414, + "grad_norm": 9.194849014282227, + "learning_rate": 9.745499349715929e-05, + "loss": 1.2365, + "step": 2610 + }, + { + "epoch": 0.1769090046751135, + "grad_norm": 7.951476097106934, + "learning_rate": 9.745362447806148e-05, + "loss": 0.8829, + "step": 2611 + }, + { + "epoch": 0.17697675994308557, + "grad_norm": 8.77597427368164, + "learning_rate": 9.745225545896366e-05, + "loss": 1.0097, + "step": 2612 + }, + { + "epoch": 0.17704451521105766, + "grad_norm": 10.356508255004883, + "learning_rate": 9.745088643986584e-05, + "loss": 1.2603, + "step": 2613 + }, + { + "epoch": 0.17711227047902975, + "grad_norm": 9.24199390411377, + "learning_rate": 9.744951742076802e-05, + "loss": 1.0181, + "step": 2614 + }, + { + "epoch": 0.17718002574700184, + "grad_norm": 8.244451522827148, + "learning_rate": 9.74481484016702e-05, + "loss": 0.9602, + "step": 2615 + }, + { + "epoch": 0.1772477810149739, + "grad_norm": 10.769842147827148, + "learning_rate": 9.74467793825724e-05, + "loss": 1.0646, + "step": 2616 + }, + { + "epoch": 0.177315536282946, + "grad_norm": 9.381787300109863, + "learning_rate": 9.744541036347458e-05, + "loss": 0.8985, + "step": 2617 + }, + { + "epoch": 0.1773832915509181, + "grad_norm": 8.379899978637695, + "learning_rate": 9.744404134437676e-05, + "loss": 1.0368, + "step": 2618 + }, + { + "epoch": 0.17745104681889018, + "grad_norm": 9.59954833984375, + "learning_rate": 9.744267232527894e-05, + "loss": 1.1447, + "step": 2619 + }, + { + "epoch": 0.17751880208686224, + "grad_norm": 8.83703327178955, + "learning_rate": 9.744130330618113e-05, + "loss": 1.3435, + "step": 2620 + }, + { + "epoch": 0.17758655735483433, + "grad_norm": 9.448990821838379, + "learning_rate": 9.743993428708331e-05, + "loss": 1.1837, + "step": 2621 + }, + { + "epoch": 0.17765431262280643, + "grad_norm": 9.466961860656738, + "learning_rate": 9.74385652679855e-05, + "loss": 1.1265, + "step": 2622 + }, + { + "epoch": 0.17772206789077852, + "grad_norm": 8.68281364440918, + "learning_rate": 9.743719624888767e-05, + "loss": 0.8804, + "step": 2623 + }, + { + "epoch": 0.17778982315875058, + "grad_norm": 7.007611274719238, + "learning_rate": 9.743582722978985e-05, + "loss": 1.03, + "step": 2624 + }, + { + "epoch": 0.17785757842672267, + "grad_norm": 8.254279136657715, + "learning_rate": 9.743445821069205e-05, + "loss": 1.0328, + "step": 2625 + }, + { + "epoch": 0.17792533369469477, + "grad_norm": 9.134196281433105, + "learning_rate": 9.743308919159423e-05, + "loss": 0.9671, + "step": 2626 + }, + { + "epoch": 0.17799308896266686, + "grad_norm": 7.803997039794922, + "learning_rate": 9.743172017249641e-05, + "loss": 0.9692, + "step": 2627 + }, + { + "epoch": 0.17806084423063892, + "grad_norm": 8.37303638458252, + "learning_rate": 9.743035115339859e-05, + "loss": 0.8949, + "step": 2628 + }, + { + "epoch": 0.178128599498611, + "grad_norm": 9.928305625915527, + "learning_rate": 9.742898213430078e-05, + "loss": 1.2679, + "step": 2629 + }, + { + "epoch": 0.1781963547665831, + "grad_norm": 8.58604621887207, + "learning_rate": 9.742761311520296e-05, + "loss": 1.23, + "step": 2630 + }, + { + "epoch": 0.1782641100345552, + "grad_norm": 9.47903823852539, + "learning_rate": 9.742624409610514e-05, + "loss": 0.8701, + "step": 2631 + }, + { + "epoch": 0.17833186530252726, + "grad_norm": 8.013197898864746, + "learning_rate": 9.742487507700732e-05, + "loss": 0.8325, + "step": 2632 + }, + { + "epoch": 0.17839962057049935, + "grad_norm": 9.541396141052246, + "learning_rate": 9.74235060579095e-05, + "loss": 1.0996, + "step": 2633 + }, + { + "epoch": 0.17846737583847144, + "grad_norm": 7.515398979187012, + "learning_rate": 9.74221370388117e-05, + "loss": 1.1174, + "step": 2634 + }, + { + "epoch": 0.17853513110644353, + "grad_norm": 9.562670707702637, + "learning_rate": 9.742076801971388e-05, + "loss": 0.9727, + "step": 2635 + }, + { + "epoch": 0.1786028863744156, + "grad_norm": 10.072449684143066, + "learning_rate": 9.741939900061606e-05, + "loss": 1.0021, + "step": 2636 + }, + { + "epoch": 0.1786706416423877, + "grad_norm": 9.634955406188965, + "learning_rate": 9.741802998151824e-05, + "loss": 1.3998, + "step": 2637 + }, + { + "epoch": 0.17873839691035978, + "grad_norm": 8.78954792022705, + "learning_rate": 9.741666096242042e-05, + "loss": 0.9602, + "step": 2638 + }, + { + "epoch": 0.17880615217833187, + "grad_norm": 9.597916603088379, + "learning_rate": 9.741529194332261e-05, + "loss": 1.0174, + "step": 2639 + }, + { + "epoch": 0.17887390744630394, + "grad_norm": 9.543583869934082, + "learning_rate": 9.74139229242248e-05, + "loss": 1.0686, + "step": 2640 + }, + { + "epoch": 0.17894166271427603, + "grad_norm": 8.247551918029785, + "learning_rate": 9.741255390512697e-05, + "loss": 1.1026, + "step": 2641 + }, + { + "epoch": 0.17900941798224812, + "grad_norm": 8.487943649291992, + "learning_rate": 9.741118488602915e-05, + "loss": 0.8843, + "step": 2642 + }, + { + "epoch": 0.1790771732502202, + "grad_norm": 9.061832427978516, + "learning_rate": 9.740981586693135e-05, + "loss": 0.9708, + "step": 2643 + }, + { + "epoch": 0.17914492851819228, + "grad_norm": 9.242405891418457, + "learning_rate": 9.740844684783353e-05, + "loss": 1.1344, + "step": 2644 + }, + { + "epoch": 0.17921268378616437, + "grad_norm": 8.192344665527344, + "learning_rate": 9.740707782873571e-05, + "loss": 0.7587, + "step": 2645 + }, + { + "epoch": 0.17928043905413646, + "grad_norm": 6.6164445877075195, + "learning_rate": 9.74057088096379e-05, + "loss": 0.842, + "step": 2646 + }, + { + "epoch": 0.17934819432210855, + "grad_norm": 10.285326957702637, + "learning_rate": 9.740433979054008e-05, + "loss": 1.138, + "step": 2647 + }, + { + "epoch": 0.17941594959008064, + "grad_norm": 8.55659294128418, + "learning_rate": 9.740297077144226e-05, + "loss": 0.7925, + "step": 2648 + }, + { + "epoch": 0.1794837048580527, + "grad_norm": 8.384642601013184, + "learning_rate": 9.740160175234446e-05, + "loss": 0.9719, + "step": 2649 + }, + { + "epoch": 0.1795514601260248, + "grad_norm": 9.303678512573242, + "learning_rate": 9.740023273324664e-05, + "loss": 1.0688, + "step": 2650 + }, + { + "epoch": 0.1796192153939969, + "grad_norm": 9.642799377441406, + "learning_rate": 9.739886371414882e-05, + "loss": 0.9758, + "step": 2651 + }, + { + "epoch": 0.17968697066196898, + "grad_norm": 9.616509437561035, + "learning_rate": 9.739749469505101e-05, + "loss": 0.9999, + "step": 2652 + }, + { + "epoch": 0.17975472592994104, + "grad_norm": 10.511337280273438, + "learning_rate": 9.739612567595319e-05, + "loss": 1.0079, + "step": 2653 + }, + { + "epoch": 0.17982248119791314, + "grad_norm": 7.480882167816162, + "learning_rate": 9.739475665685537e-05, + "loss": 0.8353, + "step": 2654 + }, + { + "epoch": 0.17989023646588523, + "grad_norm": 10.277608871459961, + "learning_rate": 9.739338763775755e-05, + "loss": 0.9709, + "step": 2655 + }, + { + "epoch": 0.17995799173385732, + "grad_norm": 9.135882377624512, + "learning_rate": 9.739201861865973e-05, + "loss": 1.3888, + "step": 2656 + }, + { + "epoch": 0.18002574700182938, + "grad_norm": 8.147950172424316, + "learning_rate": 9.739064959956193e-05, + "loss": 1.2235, + "step": 2657 + }, + { + "epoch": 0.18009350226980148, + "grad_norm": 8.583501815795898, + "learning_rate": 9.738928058046411e-05, + "loss": 0.9123, + "step": 2658 + }, + { + "epoch": 0.18016125753777357, + "grad_norm": 8.202300071716309, + "learning_rate": 9.738791156136629e-05, + "loss": 1.0435, + "step": 2659 + }, + { + "epoch": 0.18022901280574566, + "grad_norm": 8.121417045593262, + "learning_rate": 9.738654254226847e-05, + "loss": 0.9351, + "step": 2660 + }, + { + "epoch": 0.18029676807371772, + "grad_norm": 10.645029067993164, + "learning_rate": 9.738517352317065e-05, + "loss": 1.5219, + "step": 2661 + }, + { + "epoch": 0.18036452334168981, + "grad_norm": 9.766422271728516, + "learning_rate": 9.738380450407284e-05, + "loss": 1.0499, + "step": 2662 + }, + { + "epoch": 0.1804322786096619, + "grad_norm": 8.137174606323242, + "learning_rate": 9.738243548497502e-05, + "loss": 1.1871, + "step": 2663 + }, + { + "epoch": 0.180500033877634, + "grad_norm": 8.422198295593262, + "learning_rate": 9.73810664658772e-05, + "loss": 0.9524, + "step": 2664 + }, + { + "epoch": 0.18056778914560606, + "grad_norm": 6.721381664276123, + "learning_rate": 9.737969744677938e-05, + "loss": 1.0449, + "step": 2665 + }, + { + "epoch": 0.18063554441357815, + "grad_norm": 9.175302505493164, + "learning_rate": 9.737832842768158e-05, + "loss": 1.0508, + "step": 2666 + }, + { + "epoch": 0.18070329968155024, + "grad_norm": 10.165428161621094, + "learning_rate": 9.737695940858376e-05, + "loss": 0.8744, + "step": 2667 + }, + { + "epoch": 0.18077105494952234, + "grad_norm": 9.396173477172852, + "learning_rate": 9.737559038948594e-05, + "loss": 1.0304, + "step": 2668 + }, + { + "epoch": 0.1808388102174944, + "grad_norm": 9.680516242980957, + "learning_rate": 9.737422137038812e-05, + "loss": 0.9773, + "step": 2669 + }, + { + "epoch": 0.1809065654854665, + "grad_norm": 7.570343017578125, + "learning_rate": 9.73728523512903e-05, + "loss": 0.9159, + "step": 2670 + }, + { + "epoch": 0.18097432075343858, + "grad_norm": 9.5789213180542, + "learning_rate": 9.737148333219249e-05, + "loss": 1.1269, + "step": 2671 + }, + { + "epoch": 0.18104207602141068, + "grad_norm": 8.559327125549316, + "learning_rate": 9.737011431309467e-05, + "loss": 1.2318, + "step": 2672 + }, + { + "epoch": 0.18110983128938274, + "grad_norm": 7.630974292755127, + "learning_rate": 9.736874529399685e-05, + "loss": 0.8996, + "step": 2673 + }, + { + "epoch": 0.18117758655735483, + "grad_norm": 8.078895568847656, + "learning_rate": 9.736737627489903e-05, + "loss": 0.9942, + "step": 2674 + }, + { + "epoch": 0.18124534182532692, + "grad_norm": 8.830656051635742, + "learning_rate": 9.736600725580123e-05, + "loss": 1.0614, + "step": 2675 + }, + { + "epoch": 0.181313097093299, + "grad_norm": 7.66297721862793, + "learning_rate": 9.736463823670341e-05, + "loss": 1.025, + "step": 2676 + }, + { + "epoch": 0.18138085236127108, + "grad_norm": 8.318365097045898, + "learning_rate": 9.736326921760559e-05, + "loss": 1.1286, + "step": 2677 + }, + { + "epoch": 0.18144860762924317, + "grad_norm": 10.180238723754883, + "learning_rate": 9.736190019850777e-05, + "loss": 1.1304, + "step": 2678 + }, + { + "epoch": 0.18151636289721526, + "grad_norm": 9.2420654296875, + "learning_rate": 9.736053117940995e-05, + "loss": 1.157, + "step": 2679 + }, + { + "epoch": 0.18158411816518735, + "grad_norm": 7.984904766082764, + "learning_rate": 9.735916216031214e-05, + "loss": 1.1325, + "step": 2680 + }, + { + "epoch": 0.18165187343315942, + "grad_norm": 11.136275291442871, + "learning_rate": 9.735779314121432e-05, + "loss": 1.0151, + "step": 2681 + }, + { + "epoch": 0.1817196287011315, + "grad_norm": 9.954483032226562, + "learning_rate": 9.73564241221165e-05, + "loss": 1.0934, + "step": 2682 + }, + { + "epoch": 0.1817873839691036, + "grad_norm": 8.491388320922852, + "learning_rate": 9.735505510301868e-05, + "loss": 1.2288, + "step": 2683 + }, + { + "epoch": 0.1818551392370757, + "grad_norm": 9.355586051940918, + "learning_rate": 9.735368608392088e-05, + "loss": 0.7999, + "step": 2684 + }, + { + "epoch": 0.18192289450504776, + "grad_norm": 8.829736709594727, + "learning_rate": 9.735231706482306e-05, + "loss": 0.9711, + "step": 2685 + }, + { + "epoch": 0.18199064977301985, + "grad_norm": 9.667959213256836, + "learning_rate": 9.735094804572524e-05, + "loss": 1.1132, + "step": 2686 + }, + { + "epoch": 0.18205840504099194, + "grad_norm": 6.971467971801758, + "learning_rate": 9.734957902662742e-05, + "loss": 1.0252, + "step": 2687 + }, + { + "epoch": 0.18212616030896403, + "grad_norm": 9.69013500213623, + "learning_rate": 9.73482100075296e-05, + "loss": 0.9562, + "step": 2688 + }, + { + "epoch": 0.1821939155769361, + "grad_norm": 9.47673511505127, + "learning_rate": 9.734684098843179e-05, + "loss": 0.9822, + "step": 2689 + }, + { + "epoch": 0.18226167084490819, + "grad_norm": 9.708051681518555, + "learning_rate": 9.734547196933397e-05, + "loss": 1.0772, + "step": 2690 + }, + { + "epoch": 0.18232942611288028, + "grad_norm": 10.617173194885254, + "learning_rate": 9.734410295023615e-05, + "loss": 1.2686, + "step": 2691 + }, + { + "epoch": 0.18239718138085237, + "grad_norm": 9.52670955657959, + "learning_rate": 9.734273393113835e-05, + "loss": 0.9477, + "step": 2692 + }, + { + "epoch": 0.18246493664882443, + "grad_norm": 10.090290069580078, + "learning_rate": 9.734136491204053e-05, + "loss": 1.1766, + "step": 2693 + }, + { + "epoch": 0.18253269191679652, + "grad_norm": 8.709790229797363, + "learning_rate": 9.733999589294271e-05, + "loss": 1.1327, + "step": 2694 + }, + { + "epoch": 0.18260044718476862, + "grad_norm": 12.07381534576416, + "learning_rate": 9.73386268738449e-05, + "loss": 1.3694, + "step": 2695 + }, + { + "epoch": 0.1826682024527407, + "grad_norm": 8.329826354980469, + "learning_rate": 9.733725785474708e-05, + "loss": 1.0448, + "step": 2696 + }, + { + "epoch": 0.18273595772071277, + "grad_norm": 7.551383972167969, + "learning_rate": 9.733588883564926e-05, + "loss": 1.029, + "step": 2697 + }, + { + "epoch": 0.18280371298868486, + "grad_norm": 9.393105506896973, + "learning_rate": 9.733451981655146e-05, + "loss": 0.9729, + "step": 2698 + }, + { + "epoch": 0.18287146825665695, + "grad_norm": 9.283944129943848, + "learning_rate": 9.733315079745364e-05, + "loss": 1.2459, + "step": 2699 + }, + { + "epoch": 0.18293922352462905, + "grad_norm": 9.02878189086914, + "learning_rate": 9.733178177835582e-05, + "loss": 1.3235, + "step": 2700 + }, + { + "epoch": 0.18300697879260114, + "grad_norm": 8.735793113708496, + "learning_rate": 9.7330412759258e-05, + "loss": 1.1447, + "step": 2701 + }, + { + "epoch": 0.1830747340605732, + "grad_norm": 8.511741638183594, + "learning_rate": 9.732904374016018e-05, + "loss": 1.0148, + "step": 2702 + }, + { + "epoch": 0.1831424893285453, + "grad_norm": 6.590076446533203, + "learning_rate": 9.732767472106237e-05, + "loss": 1.0904, + "step": 2703 + }, + { + "epoch": 0.18321024459651739, + "grad_norm": 11.523942947387695, + "learning_rate": 9.732630570196455e-05, + "loss": 0.9349, + "step": 2704 + }, + { + "epoch": 0.18327799986448948, + "grad_norm": 6.972995281219482, + "learning_rate": 9.732493668286673e-05, + "loss": 0.9837, + "step": 2705 + }, + { + "epoch": 0.18334575513246154, + "grad_norm": 7.834453105926514, + "learning_rate": 9.732356766376891e-05, + "loss": 1.0374, + "step": 2706 + }, + { + "epoch": 0.18341351040043363, + "grad_norm": 9.44674015045166, + "learning_rate": 9.732219864467111e-05, + "loss": 1.4479, + "step": 2707 + }, + { + "epoch": 0.18348126566840572, + "grad_norm": 9.017583847045898, + "learning_rate": 9.732082962557329e-05, + "loss": 1.136, + "step": 2708 + }, + { + "epoch": 0.18354902093637782, + "grad_norm": 8.40336799621582, + "learning_rate": 9.731946060647547e-05, + "loss": 0.9914, + "step": 2709 + }, + { + "epoch": 0.18361677620434988, + "grad_norm": 10.586843490600586, + "learning_rate": 9.731809158737765e-05, + "loss": 1.1515, + "step": 2710 + }, + { + "epoch": 0.18368453147232197, + "grad_norm": 7.611972808837891, + "learning_rate": 9.731672256827983e-05, + "loss": 1.0821, + "step": 2711 + }, + { + "epoch": 0.18375228674029406, + "grad_norm": 9.093097686767578, + "learning_rate": 9.731535354918202e-05, + "loss": 1.1295, + "step": 2712 + }, + { + "epoch": 0.18382004200826615, + "grad_norm": 8.013279914855957, + "learning_rate": 9.73139845300842e-05, + "loss": 0.839, + "step": 2713 + }, + { + "epoch": 0.18388779727623822, + "grad_norm": 7.941558837890625, + "learning_rate": 9.731261551098638e-05, + "loss": 1.0819, + "step": 2714 + }, + { + "epoch": 0.1839555525442103, + "grad_norm": 7.09537935256958, + "learning_rate": 9.731124649188856e-05, + "loss": 0.8606, + "step": 2715 + }, + { + "epoch": 0.1840233078121824, + "grad_norm": 7.867136478424072, + "learning_rate": 9.730987747279074e-05, + "loss": 0.9942, + "step": 2716 + }, + { + "epoch": 0.1840910630801545, + "grad_norm": 8.66762638092041, + "learning_rate": 9.730850845369294e-05, + "loss": 1.0792, + "step": 2717 + }, + { + "epoch": 0.18415881834812656, + "grad_norm": 7.97219705581665, + "learning_rate": 9.730713943459512e-05, + "loss": 1.0366, + "step": 2718 + }, + { + "epoch": 0.18422657361609865, + "grad_norm": 8.341264724731445, + "learning_rate": 9.73057704154973e-05, + "loss": 1.0833, + "step": 2719 + }, + { + "epoch": 0.18429432888407074, + "grad_norm": 8.704034805297852, + "learning_rate": 9.730440139639948e-05, + "loss": 1.2909, + "step": 2720 + }, + { + "epoch": 0.18436208415204283, + "grad_norm": 9.759525299072266, + "learning_rate": 9.730303237730167e-05, + "loss": 1.1749, + "step": 2721 + }, + { + "epoch": 0.1844298394200149, + "grad_norm": 9.2392578125, + "learning_rate": 9.730166335820385e-05, + "loss": 1.4037, + "step": 2722 + }, + { + "epoch": 0.184497594687987, + "grad_norm": 10.519861221313477, + "learning_rate": 9.730029433910603e-05, + "loss": 1.2692, + "step": 2723 + }, + { + "epoch": 0.18456534995595908, + "grad_norm": 8.285715103149414, + "learning_rate": 9.729892532000821e-05, + "loss": 0.8014, + "step": 2724 + }, + { + "epoch": 0.18463310522393117, + "grad_norm": 7.261229991912842, + "learning_rate": 9.72975563009104e-05, + "loss": 0.9302, + "step": 2725 + }, + { + "epoch": 0.18470086049190323, + "grad_norm": 10.149426460266113, + "learning_rate": 9.729618728181259e-05, + "loss": 1.2855, + "step": 2726 + }, + { + "epoch": 0.18476861575987533, + "grad_norm": 7.495121955871582, + "learning_rate": 9.729481826271477e-05, + "loss": 0.9771, + "step": 2727 + }, + { + "epoch": 0.18483637102784742, + "grad_norm": 9.102607727050781, + "learning_rate": 9.729344924361695e-05, + "loss": 1.1974, + "step": 2728 + }, + { + "epoch": 0.1849041262958195, + "grad_norm": 9.57135009765625, + "learning_rate": 9.729208022451913e-05, + "loss": 0.9758, + "step": 2729 + }, + { + "epoch": 0.18497188156379157, + "grad_norm": 8.745903015136719, + "learning_rate": 9.729071120542132e-05, + "loss": 1.1719, + "step": 2730 + }, + { + "epoch": 0.18503963683176367, + "grad_norm": 10.697607040405273, + "learning_rate": 9.72893421863235e-05, + "loss": 0.952, + "step": 2731 + }, + { + "epoch": 0.18510739209973576, + "grad_norm": 9.907280921936035, + "learning_rate": 9.728797316722568e-05, + "loss": 0.9571, + "step": 2732 + }, + { + "epoch": 0.18517514736770785, + "grad_norm": 7.285250186920166, + "learning_rate": 9.728660414812786e-05, + "loss": 0.9002, + "step": 2733 + }, + { + "epoch": 0.1852429026356799, + "grad_norm": 8.134112358093262, + "learning_rate": 9.728523512903004e-05, + "loss": 0.8645, + "step": 2734 + }, + { + "epoch": 0.185310657903652, + "grad_norm": 9.427742004394531, + "learning_rate": 9.728386610993224e-05, + "loss": 1.2333, + "step": 2735 + }, + { + "epoch": 0.1853784131716241, + "grad_norm": 8.804586410522461, + "learning_rate": 9.728249709083442e-05, + "loss": 1.2601, + "step": 2736 + }, + { + "epoch": 0.1854461684395962, + "grad_norm": 9.144674301147461, + "learning_rate": 9.72811280717366e-05, + "loss": 0.9503, + "step": 2737 + }, + { + "epoch": 0.18551392370756825, + "grad_norm": 9.399337768554688, + "learning_rate": 9.727975905263879e-05, + "loss": 0.8586, + "step": 2738 + }, + { + "epoch": 0.18558167897554034, + "grad_norm": 7.7377119064331055, + "learning_rate": 9.727839003354097e-05, + "loss": 0.7511, + "step": 2739 + }, + { + "epoch": 0.18564943424351243, + "grad_norm": 9.146937370300293, + "learning_rate": 9.727702101444315e-05, + "loss": 1.0037, + "step": 2740 + }, + { + "epoch": 0.18571718951148453, + "grad_norm": 7.722357273101807, + "learning_rate": 9.727565199534535e-05, + "loss": 1.077, + "step": 2741 + }, + { + "epoch": 0.1857849447794566, + "grad_norm": 7.808566093444824, + "learning_rate": 9.727428297624753e-05, + "loss": 0.8953, + "step": 2742 + }, + { + "epoch": 0.18585270004742868, + "grad_norm": 8.765763282775879, + "learning_rate": 9.727291395714971e-05, + "loss": 1.0691, + "step": 2743 + }, + { + "epoch": 0.18592045531540077, + "grad_norm": 11.350706100463867, + "learning_rate": 9.72715449380519e-05, + "loss": 0.9308, + "step": 2744 + }, + { + "epoch": 0.18598821058337286, + "grad_norm": 7.838237285614014, + "learning_rate": 9.727017591895408e-05, + "loss": 0.9452, + "step": 2745 + }, + { + "epoch": 0.18605596585134493, + "grad_norm": 9.660740852355957, + "learning_rate": 9.726880689985626e-05, + "loss": 0.8705, + "step": 2746 + }, + { + "epoch": 0.18612372111931702, + "grad_norm": 8.146308898925781, + "learning_rate": 9.726743788075844e-05, + "loss": 1.0362, + "step": 2747 + }, + { + "epoch": 0.1861914763872891, + "grad_norm": 9.901165008544922, + "learning_rate": 9.726606886166062e-05, + "loss": 1.0842, + "step": 2748 + }, + { + "epoch": 0.1862592316552612, + "grad_norm": 7.650402069091797, + "learning_rate": 9.726469984256282e-05, + "loss": 0.997, + "step": 2749 + }, + { + "epoch": 0.18632698692323327, + "grad_norm": 7.760092735290527, + "learning_rate": 9.7263330823465e-05, + "loss": 0.9903, + "step": 2750 + }, + { + "epoch": 0.18639474219120536, + "grad_norm": 9.523726463317871, + "learning_rate": 9.726196180436718e-05, + "loss": 0.9276, + "step": 2751 + }, + { + "epoch": 0.18646249745917745, + "grad_norm": 7.759490489959717, + "learning_rate": 9.726059278526936e-05, + "loss": 0.9356, + "step": 2752 + }, + { + "epoch": 0.18653025272714954, + "grad_norm": 10.949979782104492, + "learning_rate": 9.725922376617155e-05, + "loss": 1.1451, + "step": 2753 + }, + { + "epoch": 0.18659800799512163, + "grad_norm": 8.312686920166016, + "learning_rate": 9.725785474707373e-05, + "loss": 0.973, + "step": 2754 + }, + { + "epoch": 0.1866657632630937, + "grad_norm": 6.999983787536621, + "learning_rate": 9.725648572797591e-05, + "loss": 1.0699, + "step": 2755 + }, + { + "epoch": 0.1867335185310658, + "grad_norm": 9.422745704650879, + "learning_rate": 9.725511670887809e-05, + "loss": 1.0697, + "step": 2756 + }, + { + "epoch": 0.18680127379903788, + "grad_norm": 8.054603576660156, + "learning_rate": 9.725374768978027e-05, + "loss": 0.8764, + "step": 2757 + }, + { + "epoch": 0.18686902906700997, + "grad_norm": 7.902176856994629, + "learning_rate": 9.725237867068247e-05, + "loss": 0.9893, + "step": 2758 + }, + { + "epoch": 0.18693678433498204, + "grad_norm": 8.409537315368652, + "learning_rate": 9.725100965158465e-05, + "loss": 1.0526, + "step": 2759 + }, + { + "epoch": 0.18700453960295413, + "grad_norm": 9.867463111877441, + "learning_rate": 9.724964063248683e-05, + "loss": 1.1218, + "step": 2760 + }, + { + "epoch": 0.18707229487092622, + "grad_norm": 6.852199554443359, + "learning_rate": 9.724827161338901e-05, + "loss": 0.9838, + "step": 2761 + }, + { + "epoch": 0.1871400501388983, + "grad_norm": 9.132448196411133, + "learning_rate": 9.72469025942912e-05, + "loss": 1.038, + "step": 2762 + }, + { + "epoch": 0.18720780540687038, + "grad_norm": 8.699213981628418, + "learning_rate": 9.724553357519338e-05, + "loss": 1.0532, + "step": 2763 + }, + { + "epoch": 0.18727556067484247, + "grad_norm": 8.520672798156738, + "learning_rate": 9.724416455609556e-05, + "loss": 1.2942, + "step": 2764 + }, + { + "epoch": 0.18734331594281456, + "grad_norm": 9.690433502197266, + "learning_rate": 9.724279553699774e-05, + "loss": 0.9552, + "step": 2765 + }, + { + "epoch": 0.18741107121078665, + "grad_norm": 7.540090560913086, + "learning_rate": 9.724142651789992e-05, + "loss": 1.0718, + "step": 2766 + }, + { + "epoch": 0.18747882647875871, + "grad_norm": 9.072039604187012, + "learning_rate": 9.724005749880212e-05, + "loss": 1.0313, + "step": 2767 + }, + { + "epoch": 0.1875465817467308, + "grad_norm": 10.155011177062988, + "learning_rate": 9.72386884797043e-05, + "loss": 1.3498, + "step": 2768 + }, + { + "epoch": 0.1876143370147029, + "grad_norm": 7.816718101501465, + "learning_rate": 9.723731946060648e-05, + "loss": 0.9594, + "step": 2769 + }, + { + "epoch": 0.187682092282675, + "grad_norm": 10.980911254882812, + "learning_rate": 9.723595044150866e-05, + "loss": 1.1675, + "step": 2770 + }, + { + "epoch": 0.18774984755064705, + "grad_norm": 8.865739822387695, + "learning_rate": 9.723458142241084e-05, + "loss": 1.1796, + "step": 2771 + }, + { + "epoch": 0.18781760281861914, + "grad_norm": 8.97850227355957, + "learning_rate": 9.723321240331303e-05, + "loss": 1.0907, + "step": 2772 + }, + { + "epoch": 0.18788535808659124, + "grad_norm": 8.16921329498291, + "learning_rate": 9.723184338421521e-05, + "loss": 0.95, + "step": 2773 + }, + { + "epoch": 0.18795311335456333, + "grad_norm": 8.766203880310059, + "learning_rate": 9.723047436511739e-05, + "loss": 0.9087, + "step": 2774 + }, + { + "epoch": 0.1880208686225354, + "grad_norm": 7.410607814788818, + "learning_rate": 9.722910534601957e-05, + "loss": 0.8082, + "step": 2775 + }, + { + "epoch": 0.18808862389050748, + "grad_norm": 9.640182495117188, + "learning_rate": 9.722773632692177e-05, + "loss": 0.99, + "step": 2776 + }, + { + "epoch": 0.18815637915847958, + "grad_norm": 9.038297653198242, + "learning_rate": 9.722636730782395e-05, + "loss": 1.0017, + "step": 2777 + }, + { + "epoch": 0.18822413442645167, + "grad_norm": 10.167421340942383, + "learning_rate": 9.722499828872613e-05, + "loss": 1.1588, + "step": 2778 + }, + { + "epoch": 0.18829188969442373, + "grad_norm": 9.911538124084473, + "learning_rate": 9.722362926962831e-05, + "loss": 1.052, + "step": 2779 + }, + { + "epoch": 0.18835964496239582, + "grad_norm": 8.78661060333252, + "learning_rate": 9.722226025053049e-05, + "loss": 0.9827, + "step": 2780 + }, + { + "epoch": 0.18842740023036791, + "grad_norm": 8.58356761932373, + "learning_rate": 9.722089123143268e-05, + "loss": 0.9663, + "step": 2781 + }, + { + "epoch": 0.18849515549834, + "grad_norm": 7.882653713226318, + "learning_rate": 9.721952221233486e-05, + "loss": 1.0282, + "step": 2782 + }, + { + "epoch": 0.18856291076631207, + "grad_norm": 8.180728912353516, + "learning_rate": 9.721815319323704e-05, + "loss": 0.8983, + "step": 2783 + }, + { + "epoch": 0.18863066603428416, + "grad_norm": 10.785475730895996, + "learning_rate": 9.721678417413924e-05, + "loss": 1.0949, + "step": 2784 + }, + { + "epoch": 0.18869842130225625, + "grad_norm": 8.2493257522583, + "learning_rate": 9.721541515504142e-05, + "loss": 0.8891, + "step": 2785 + }, + { + "epoch": 0.18876617657022834, + "grad_norm": 8.298515319824219, + "learning_rate": 9.72140461359436e-05, + "loss": 0.9384, + "step": 2786 + }, + { + "epoch": 0.1888339318382004, + "grad_norm": 11.643486022949219, + "learning_rate": 9.721267711684579e-05, + "loss": 1.0915, + "step": 2787 + }, + { + "epoch": 0.1889016871061725, + "grad_norm": 10.728472709655762, + "learning_rate": 9.721130809774797e-05, + "loss": 1.562, + "step": 2788 + }, + { + "epoch": 0.1889694423741446, + "grad_norm": 7.9253435134887695, + "learning_rate": 9.720993907865015e-05, + "loss": 0.876, + "step": 2789 + }, + { + "epoch": 0.18903719764211668, + "grad_norm": 9.075439453125, + "learning_rate": 9.720857005955235e-05, + "loss": 1.1712, + "step": 2790 + }, + { + "epoch": 0.18910495291008875, + "grad_norm": 8.572853088378906, + "learning_rate": 9.720720104045453e-05, + "loss": 0.9697, + "step": 2791 + }, + { + "epoch": 0.18917270817806084, + "grad_norm": 8.40988826751709, + "learning_rate": 9.720583202135671e-05, + "loss": 1.0904, + "step": 2792 + }, + { + "epoch": 0.18924046344603293, + "grad_norm": 8.015021324157715, + "learning_rate": 9.720446300225889e-05, + "loss": 1.0264, + "step": 2793 + }, + { + "epoch": 0.18930821871400502, + "grad_norm": 10.295256614685059, + "learning_rate": 9.720309398316107e-05, + "loss": 0.9608, + "step": 2794 + }, + { + "epoch": 0.18937597398197709, + "grad_norm": 9.377728462219238, + "learning_rate": 9.720172496406326e-05, + "loss": 1.2169, + "step": 2795 + }, + { + "epoch": 0.18944372924994918, + "grad_norm": 8.55238151550293, + "learning_rate": 9.720035594496544e-05, + "loss": 1.0951, + "step": 2796 + }, + { + "epoch": 0.18951148451792127, + "grad_norm": 10.069438934326172, + "learning_rate": 9.719898692586762e-05, + "loss": 1.3206, + "step": 2797 + }, + { + "epoch": 0.18957923978589336, + "grad_norm": 9.229057312011719, + "learning_rate": 9.71976179067698e-05, + "loss": 1.18, + "step": 2798 + }, + { + "epoch": 0.18964699505386542, + "grad_norm": 8.911051750183105, + "learning_rate": 9.7196248887672e-05, + "loss": 1.1351, + "step": 2799 + }, + { + "epoch": 0.18971475032183752, + "grad_norm": 9.285752296447754, + "learning_rate": 9.719487986857418e-05, + "loss": 1.0626, + "step": 2800 + }, + { + "epoch": 0.1897825055898096, + "grad_norm": 8.793561935424805, + "learning_rate": 9.719351084947636e-05, + "loss": 1.033, + "step": 2801 + }, + { + "epoch": 0.1898502608577817, + "grad_norm": 10.635384559631348, + "learning_rate": 9.719214183037854e-05, + "loss": 1.5004, + "step": 2802 + }, + { + "epoch": 0.18991801612575376, + "grad_norm": 7.9627275466918945, + "learning_rate": 9.719077281128072e-05, + "loss": 1.0246, + "step": 2803 + }, + { + "epoch": 0.18998577139372586, + "grad_norm": 8.960352897644043, + "learning_rate": 9.718940379218291e-05, + "loss": 1.0316, + "step": 2804 + }, + { + "epoch": 0.19005352666169795, + "grad_norm": 9.522171020507812, + "learning_rate": 9.718803477308509e-05, + "loss": 1.0219, + "step": 2805 + }, + { + "epoch": 0.19012128192967004, + "grad_norm": 8.412702560424805, + "learning_rate": 9.718666575398727e-05, + "loss": 1.0609, + "step": 2806 + }, + { + "epoch": 0.19018903719764213, + "grad_norm": 8.777839660644531, + "learning_rate": 9.718529673488945e-05, + "loss": 1.0854, + "step": 2807 + }, + { + "epoch": 0.1902567924656142, + "grad_norm": 8.932796478271484, + "learning_rate": 9.718392771579165e-05, + "loss": 1.1661, + "step": 2808 + }, + { + "epoch": 0.19032454773358629, + "grad_norm": 8.514800071716309, + "learning_rate": 9.718255869669383e-05, + "loss": 1.0584, + "step": 2809 + }, + { + "epoch": 0.19039230300155838, + "grad_norm": 7.001948356628418, + "learning_rate": 9.718118967759601e-05, + "loss": 0.8479, + "step": 2810 + }, + { + "epoch": 0.19046005826953047, + "grad_norm": 8.977307319641113, + "learning_rate": 9.717982065849819e-05, + "loss": 1.2556, + "step": 2811 + }, + { + "epoch": 0.19052781353750253, + "grad_norm": 8.513920783996582, + "learning_rate": 9.717845163940037e-05, + "loss": 1.2076, + "step": 2812 + }, + { + "epoch": 0.19059556880547462, + "grad_norm": 8.556622505187988, + "learning_rate": 9.717708262030256e-05, + "loss": 1.0491, + "step": 2813 + }, + { + "epoch": 0.19066332407344672, + "grad_norm": 9.8518648147583, + "learning_rate": 9.717571360120474e-05, + "loss": 1.1179, + "step": 2814 + }, + { + "epoch": 0.1907310793414188, + "grad_norm": 8.887413024902344, + "learning_rate": 9.717434458210692e-05, + "loss": 1.0675, + "step": 2815 + }, + { + "epoch": 0.19079883460939087, + "grad_norm": 7.888981342315674, + "learning_rate": 9.71729755630091e-05, + "loss": 1.0489, + "step": 2816 + }, + { + "epoch": 0.19086658987736296, + "grad_norm": 7.692848205566406, + "learning_rate": 9.71716065439113e-05, + "loss": 1.0602, + "step": 2817 + }, + { + "epoch": 0.19093434514533505, + "grad_norm": 8.591787338256836, + "learning_rate": 9.717023752481348e-05, + "loss": 0.8737, + "step": 2818 + }, + { + "epoch": 0.19100210041330715, + "grad_norm": 9.622870445251465, + "learning_rate": 9.716886850571566e-05, + "loss": 0.9717, + "step": 2819 + }, + { + "epoch": 0.1910698556812792, + "grad_norm": 7.949582576751709, + "learning_rate": 9.716749948661784e-05, + "loss": 0.8651, + "step": 2820 + }, + { + "epoch": 0.1911376109492513, + "grad_norm": 9.46272087097168, + "learning_rate": 9.716613046752002e-05, + "loss": 0.9975, + "step": 2821 + }, + { + "epoch": 0.1912053662172234, + "grad_norm": 8.86549186706543, + "learning_rate": 9.716476144842221e-05, + "loss": 1.0805, + "step": 2822 + }, + { + "epoch": 0.19127312148519549, + "grad_norm": 8.039673805236816, + "learning_rate": 9.716339242932439e-05, + "loss": 1.121, + "step": 2823 + }, + { + "epoch": 0.19134087675316755, + "grad_norm": 8.606979370117188, + "learning_rate": 9.716202341022657e-05, + "loss": 0.9217, + "step": 2824 + }, + { + "epoch": 0.19140863202113964, + "grad_norm": 10.429420471191406, + "learning_rate": 9.716065439112875e-05, + "loss": 1.2266, + "step": 2825 + }, + { + "epoch": 0.19147638728911173, + "grad_norm": 7.339411735534668, + "learning_rate": 9.715928537203093e-05, + "loss": 1.0235, + "step": 2826 + }, + { + "epoch": 0.19154414255708382, + "grad_norm": 8.247300148010254, + "learning_rate": 9.715791635293313e-05, + "loss": 0.8993, + "step": 2827 + }, + { + "epoch": 0.1916118978250559, + "grad_norm": 8.475278854370117, + "learning_rate": 9.715654733383531e-05, + "loss": 0.9611, + "step": 2828 + }, + { + "epoch": 0.19167965309302798, + "grad_norm": 9.792519569396973, + "learning_rate": 9.715517831473749e-05, + "loss": 1.0831, + "step": 2829 + }, + { + "epoch": 0.19174740836100007, + "grad_norm": 6.757070541381836, + "learning_rate": 9.715380929563967e-05, + "loss": 1.0282, + "step": 2830 + }, + { + "epoch": 0.19181516362897216, + "grad_norm": 6.707785129547119, + "learning_rate": 9.715244027654186e-05, + "loss": 0.7813, + "step": 2831 + }, + { + "epoch": 0.19188291889694423, + "grad_norm": 9.990489959716797, + "learning_rate": 9.715107125744404e-05, + "loss": 1.0504, + "step": 2832 + }, + { + "epoch": 0.19195067416491632, + "grad_norm": 10.57358169555664, + "learning_rate": 9.714970223834622e-05, + "loss": 0.9355, + "step": 2833 + }, + { + "epoch": 0.1920184294328884, + "grad_norm": 9.225300788879395, + "learning_rate": 9.714833321924842e-05, + "loss": 1.2404, + "step": 2834 + }, + { + "epoch": 0.1920861847008605, + "grad_norm": 8.36042594909668, + "learning_rate": 9.71469642001506e-05, + "loss": 0.9012, + "step": 2835 + }, + { + "epoch": 0.19215393996883257, + "grad_norm": 9.624984741210938, + "learning_rate": 9.714559518105278e-05, + "loss": 1.0509, + "step": 2836 + }, + { + "epoch": 0.19222169523680466, + "grad_norm": 12.569930076599121, + "learning_rate": 9.714422616195497e-05, + "loss": 1.2663, + "step": 2837 + }, + { + "epoch": 0.19228945050477675, + "grad_norm": 12.278885841369629, + "learning_rate": 9.714285714285715e-05, + "loss": 1.1547, + "step": 2838 + }, + { + "epoch": 0.19235720577274884, + "grad_norm": 9.3023681640625, + "learning_rate": 9.714148812375933e-05, + "loss": 1.0873, + "step": 2839 + }, + { + "epoch": 0.1924249610407209, + "grad_norm": 8.241714477539062, + "learning_rate": 9.714011910466153e-05, + "loss": 0.8222, + "step": 2840 + }, + { + "epoch": 0.192492716308693, + "grad_norm": 10.45174503326416, + "learning_rate": 9.71387500855637e-05, + "loss": 1.1974, + "step": 2841 + }, + { + "epoch": 0.1925604715766651, + "grad_norm": 8.096826553344727, + "learning_rate": 9.713738106646589e-05, + "loss": 1.1147, + "step": 2842 + }, + { + "epoch": 0.19262822684463718, + "grad_norm": 9.053191184997559, + "learning_rate": 9.713601204736807e-05, + "loss": 0.9835, + "step": 2843 + }, + { + "epoch": 0.19269598211260924, + "grad_norm": 8.050823211669922, + "learning_rate": 9.713464302827025e-05, + "loss": 1.0826, + "step": 2844 + }, + { + "epoch": 0.19276373738058133, + "grad_norm": 10.000917434692383, + "learning_rate": 9.713327400917244e-05, + "loss": 1.1723, + "step": 2845 + }, + { + "epoch": 0.19283149264855343, + "grad_norm": 8.992589950561523, + "learning_rate": 9.713190499007462e-05, + "loss": 1.0203, + "step": 2846 + }, + { + "epoch": 0.19289924791652552, + "grad_norm": 10.340285301208496, + "learning_rate": 9.71305359709768e-05, + "loss": 1.2523, + "step": 2847 + }, + { + "epoch": 0.19296700318449758, + "grad_norm": 8.250594139099121, + "learning_rate": 9.712916695187898e-05, + "loss": 1.0606, + "step": 2848 + }, + { + "epoch": 0.19303475845246967, + "grad_norm": 7.16335916519165, + "learning_rate": 9.712779793278116e-05, + "loss": 0.9988, + "step": 2849 + }, + { + "epoch": 0.19310251372044177, + "grad_norm": 9.089055061340332, + "learning_rate": 9.712642891368336e-05, + "loss": 0.9368, + "step": 2850 + }, + { + "epoch": 0.19317026898841386, + "grad_norm": 7.568434238433838, + "learning_rate": 9.712505989458554e-05, + "loss": 0.9417, + "step": 2851 + }, + { + "epoch": 0.19323802425638592, + "grad_norm": 8.470823287963867, + "learning_rate": 9.712369087548772e-05, + "loss": 1.0271, + "step": 2852 + }, + { + "epoch": 0.193305779524358, + "grad_norm": 7.746623992919922, + "learning_rate": 9.71223218563899e-05, + "loss": 0.7641, + "step": 2853 + }, + { + "epoch": 0.1933735347923301, + "grad_norm": 7.44852352142334, + "learning_rate": 9.712095283729209e-05, + "loss": 0.9984, + "step": 2854 + }, + { + "epoch": 0.1934412900603022, + "grad_norm": 7.393777847290039, + "learning_rate": 9.711958381819427e-05, + "loss": 0.8515, + "step": 2855 + }, + { + "epoch": 0.19350904532827426, + "grad_norm": 8.247236251831055, + "learning_rate": 9.711821479909645e-05, + "loss": 0.9867, + "step": 2856 + }, + { + "epoch": 0.19357680059624635, + "grad_norm": 8.484920501708984, + "learning_rate": 9.711684577999863e-05, + "loss": 0.9099, + "step": 2857 + }, + { + "epoch": 0.19364455586421844, + "grad_norm": 9.85857105255127, + "learning_rate": 9.711547676090081e-05, + "loss": 1.0134, + "step": 2858 + }, + { + "epoch": 0.19371231113219053, + "grad_norm": 12.191691398620605, + "learning_rate": 9.7114107741803e-05, + "loss": 1.15, + "step": 2859 + }, + { + "epoch": 0.19378006640016263, + "grad_norm": 9.691742897033691, + "learning_rate": 9.711273872270519e-05, + "loss": 1.2415, + "step": 2860 + }, + { + "epoch": 0.1938478216681347, + "grad_norm": 11.395289421081543, + "learning_rate": 9.711136970360737e-05, + "loss": 1.3239, + "step": 2861 + }, + { + "epoch": 0.19391557693610678, + "grad_norm": 9.222856521606445, + "learning_rate": 9.711000068450955e-05, + "loss": 0.9087, + "step": 2862 + }, + { + "epoch": 0.19398333220407887, + "grad_norm": 8.062904357910156, + "learning_rate": 9.710863166541174e-05, + "loss": 1.093, + "step": 2863 + }, + { + "epoch": 0.19405108747205096, + "grad_norm": 8.160481452941895, + "learning_rate": 9.710726264631392e-05, + "loss": 0.842, + "step": 2864 + }, + { + "epoch": 0.19411884274002303, + "grad_norm": 8.165858268737793, + "learning_rate": 9.71058936272161e-05, + "loss": 0.8807, + "step": 2865 + }, + { + "epoch": 0.19418659800799512, + "grad_norm": 8.120240211486816, + "learning_rate": 9.710452460811828e-05, + "loss": 0.9733, + "step": 2866 + }, + { + "epoch": 0.1942543532759672, + "grad_norm": 11.363536834716797, + "learning_rate": 9.710315558902046e-05, + "loss": 1.1144, + "step": 2867 + }, + { + "epoch": 0.1943221085439393, + "grad_norm": 8.776150703430176, + "learning_rate": 9.710178656992266e-05, + "loss": 0.7561, + "step": 2868 + }, + { + "epoch": 0.19438986381191137, + "grad_norm": 9.393696784973145, + "learning_rate": 9.710041755082484e-05, + "loss": 0.8092, + "step": 2869 + }, + { + "epoch": 0.19445761907988346, + "grad_norm": 10.820277214050293, + "learning_rate": 9.709904853172702e-05, + "loss": 0.9905, + "step": 2870 + }, + { + "epoch": 0.19452537434785555, + "grad_norm": 8.041844367980957, + "learning_rate": 9.70976795126292e-05, + "loss": 0.8801, + "step": 2871 + }, + { + "epoch": 0.19459312961582764, + "grad_norm": 8.05355167388916, + "learning_rate": 9.709631049353138e-05, + "loss": 0.9946, + "step": 2872 + }, + { + "epoch": 0.1946608848837997, + "grad_norm": 9.46949577331543, + "learning_rate": 9.709494147443357e-05, + "loss": 1.0582, + "step": 2873 + }, + { + "epoch": 0.1947286401517718, + "grad_norm": 8.796204566955566, + "learning_rate": 9.709357245533575e-05, + "loss": 1.1542, + "step": 2874 + }, + { + "epoch": 0.1947963954197439, + "grad_norm": 10.45006275177002, + "learning_rate": 9.709220343623793e-05, + "loss": 1.2363, + "step": 2875 + }, + { + "epoch": 0.19486415068771598, + "grad_norm": 10.490102767944336, + "learning_rate": 9.709083441714011e-05, + "loss": 1.1977, + "step": 2876 + }, + { + "epoch": 0.19493190595568805, + "grad_norm": 9.048376083374023, + "learning_rate": 9.708946539804231e-05, + "loss": 0.8127, + "step": 2877 + }, + { + "epoch": 0.19499966122366014, + "grad_norm": 6.523111343383789, + "learning_rate": 9.708809637894449e-05, + "loss": 0.8182, + "step": 2878 + }, + { + "epoch": 0.19506741649163223, + "grad_norm": 7.170145511627197, + "learning_rate": 9.708672735984667e-05, + "loss": 0.9643, + "step": 2879 + }, + { + "epoch": 0.19513517175960432, + "grad_norm": 7.9981818199157715, + "learning_rate": 9.708535834074886e-05, + "loss": 1.3179, + "step": 2880 + }, + { + "epoch": 0.19520292702757638, + "grad_norm": 8.786405563354492, + "learning_rate": 9.708398932165104e-05, + "loss": 0.9217, + "step": 2881 + }, + { + "epoch": 0.19527068229554848, + "grad_norm": 6.7907867431640625, + "learning_rate": 9.708262030255322e-05, + "loss": 1.0381, + "step": 2882 + }, + { + "epoch": 0.19533843756352057, + "grad_norm": 7.1751556396484375, + "learning_rate": 9.708125128345542e-05, + "loss": 0.8334, + "step": 2883 + }, + { + "epoch": 0.19540619283149266, + "grad_norm": 8.236610412597656, + "learning_rate": 9.70798822643576e-05, + "loss": 1.2058, + "step": 2884 + }, + { + "epoch": 0.19547394809946472, + "grad_norm": 7.576273441314697, + "learning_rate": 9.707851324525978e-05, + "loss": 1.091, + "step": 2885 + }, + { + "epoch": 0.19554170336743681, + "grad_norm": 8.698029518127441, + "learning_rate": 9.707714422616197e-05, + "loss": 1.0262, + "step": 2886 + }, + { + "epoch": 0.1956094586354089, + "grad_norm": 8.583345413208008, + "learning_rate": 9.707577520706415e-05, + "loss": 1.0401, + "step": 2887 + }, + { + "epoch": 0.195677213903381, + "grad_norm": 7.242405891418457, + "learning_rate": 9.707440618796633e-05, + "loss": 0.785, + "step": 2888 + }, + { + "epoch": 0.19574496917135306, + "grad_norm": 8.4541654586792, + "learning_rate": 9.707303716886851e-05, + "loss": 1.0537, + "step": 2889 + }, + { + "epoch": 0.19581272443932515, + "grad_norm": 7.838657855987549, + "learning_rate": 9.707166814977069e-05, + "loss": 1.0062, + "step": 2890 + }, + { + "epoch": 0.19588047970729724, + "grad_norm": 6.812248229980469, + "learning_rate": 9.707029913067289e-05, + "loss": 0.885, + "step": 2891 + }, + { + "epoch": 0.19594823497526934, + "grad_norm": 8.789966583251953, + "learning_rate": 9.706893011157507e-05, + "loss": 1.0366, + "step": 2892 + }, + { + "epoch": 0.1960159902432414, + "grad_norm": 8.860052108764648, + "learning_rate": 9.706756109247725e-05, + "loss": 1.2247, + "step": 2893 + }, + { + "epoch": 0.1960837455112135, + "grad_norm": 11.235320091247559, + "learning_rate": 9.706619207337943e-05, + "loss": 1.1216, + "step": 2894 + }, + { + "epoch": 0.19615150077918558, + "grad_norm": 8.865259170532227, + "learning_rate": 9.706482305428162e-05, + "loss": 0.9467, + "step": 2895 + }, + { + "epoch": 0.19621925604715768, + "grad_norm": 8.345112800598145, + "learning_rate": 9.70634540351838e-05, + "loss": 0.9849, + "step": 2896 + }, + { + "epoch": 0.19628701131512974, + "grad_norm": 8.98128604888916, + "learning_rate": 9.706208501608598e-05, + "loss": 1.1421, + "step": 2897 + }, + { + "epoch": 0.19635476658310183, + "grad_norm": 7.786384582519531, + "learning_rate": 9.706071599698816e-05, + "loss": 1.0926, + "step": 2898 + }, + { + "epoch": 0.19642252185107392, + "grad_norm": 8.816730499267578, + "learning_rate": 9.705934697789034e-05, + "loss": 1.2626, + "step": 2899 + }, + { + "epoch": 0.19649027711904601, + "grad_norm": 7.831095218658447, + "learning_rate": 9.705797795879254e-05, + "loss": 0.9355, + "step": 2900 + }, + { + "epoch": 0.19655803238701808, + "grad_norm": 8.788371086120605, + "learning_rate": 9.705660893969472e-05, + "loss": 0.8883, + "step": 2901 + }, + { + "epoch": 0.19662578765499017, + "grad_norm": 8.88425064086914, + "learning_rate": 9.70552399205969e-05, + "loss": 1.1169, + "step": 2902 + }, + { + "epoch": 0.19669354292296226, + "grad_norm": 9.648268699645996, + "learning_rate": 9.705387090149908e-05, + "loss": 1.0628, + "step": 2903 + }, + { + "epoch": 0.19676129819093435, + "grad_norm": 8.19676685333252, + "learning_rate": 9.705250188240126e-05, + "loss": 0.9395, + "step": 2904 + }, + { + "epoch": 0.19682905345890642, + "grad_norm": 7.420725345611572, + "learning_rate": 9.705113286330345e-05, + "loss": 0.947, + "step": 2905 + }, + { + "epoch": 0.1968968087268785, + "grad_norm": 7.221796989440918, + "learning_rate": 9.704976384420563e-05, + "loss": 0.8491, + "step": 2906 + }, + { + "epoch": 0.1969645639948506, + "grad_norm": 9.932676315307617, + "learning_rate": 9.704839482510781e-05, + "loss": 1.117, + "step": 2907 + }, + { + "epoch": 0.1970323192628227, + "grad_norm": 11.816266059875488, + "learning_rate": 9.704702580600999e-05, + "loss": 1.4347, + "step": 2908 + }, + { + "epoch": 0.19710007453079476, + "grad_norm": 8.804407119750977, + "learning_rate": 9.704565678691219e-05, + "loss": 0.808, + "step": 2909 + }, + { + "epoch": 0.19716782979876685, + "grad_norm": 7.740353584289551, + "learning_rate": 9.704428776781437e-05, + "loss": 1.0732, + "step": 2910 + }, + { + "epoch": 0.19723558506673894, + "grad_norm": 12.040196418762207, + "learning_rate": 9.704291874871655e-05, + "loss": 0.947, + "step": 2911 + }, + { + "epoch": 0.19730334033471103, + "grad_norm": 7.727171421051025, + "learning_rate": 9.704154972961873e-05, + "loss": 0.8211, + "step": 2912 + }, + { + "epoch": 0.19737109560268312, + "grad_norm": 8.779428482055664, + "learning_rate": 9.704018071052091e-05, + "loss": 1.2204, + "step": 2913 + }, + { + "epoch": 0.19743885087065519, + "grad_norm": 7.907576084136963, + "learning_rate": 9.70388116914231e-05, + "loss": 1.0594, + "step": 2914 + }, + { + "epoch": 0.19750660613862728, + "grad_norm": 6.769292831420898, + "learning_rate": 9.703744267232528e-05, + "loss": 0.8743, + "step": 2915 + }, + { + "epoch": 0.19757436140659937, + "grad_norm": 8.966355323791504, + "learning_rate": 9.703607365322746e-05, + "loss": 1.1607, + "step": 2916 + }, + { + "epoch": 0.19764211667457146, + "grad_norm": 7.389810085296631, + "learning_rate": 9.703470463412964e-05, + "loss": 0.9728, + "step": 2917 + }, + { + "epoch": 0.19770987194254352, + "grad_norm": 10.877386093139648, + "learning_rate": 9.703333561503184e-05, + "loss": 1.0623, + "step": 2918 + }, + { + "epoch": 0.19777762721051562, + "grad_norm": 9.160116195678711, + "learning_rate": 9.703196659593402e-05, + "loss": 1.1374, + "step": 2919 + }, + { + "epoch": 0.1978453824784877, + "grad_norm": 7.6289167404174805, + "learning_rate": 9.70305975768362e-05, + "loss": 0.9498, + "step": 2920 + }, + { + "epoch": 0.1979131377464598, + "grad_norm": 8.663583755493164, + "learning_rate": 9.702922855773838e-05, + "loss": 1.1595, + "step": 2921 + }, + { + "epoch": 0.19798089301443186, + "grad_norm": 7.595486640930176, + "learning_rate": 9.702785953864056e-05, + "loss": 1.0623, + "step": 2922 + }, + { + "epoch": 0.19804864828240396, + "grad_norm": 8.680171966552734, + "learning_rate": 9.702649051954275e-05, + "loss": 1.0138, + "step": 2923 + }, + { + "epoch": 0.19811640355037605, + "grad_norm": 9.998015403747559, + "learning_rate": 9.702512150044493e-05, + "loss": 1.2006, + "step": 2924 + }, + { + "epoch": 0.19818415881834814, + "grad_norm": 7.801429271697998, + "learning_rate": 9.702375248134711e-05, + "loss": 0.8467, + "step": 2925 + }, + { + "epoch": 0.1982519140863202, + "grad_norm": 8.570688247680664, + "learning_rate": 9.70223834622493e-05, + "loss": 1.1439, + "step": 2926 + }, + { + "epoch": 0.1983196693542923, + "grad_norm": 7.835936069488525, + "learning_rate": 9.702101444315149e-05, + "loss": 0.9671, + "step": 2927 + }, + { + "epoch": 0.19838742462226439, + "grad_norm": 8.415708541870117, + "learning_rate": 9.701964542405367e-05, + "loss": 1.0381, + "step": 2928 + }, + { + "epoch": 0.19845517989023648, + "grad_norm": 10.370524406433105, + "learning_rate": 9.701827640495586e-05, + "loss": 1.151, + "step": 2929 + }, + { + "epoch": 0.19852293515820854, + "grad_norm": 8.228797912597656, + "learning_rate": 9.701690738585804e-05, + "loss": 1.093, + "step": 2930 + }, + { + "epoch": 0.19859069042618063, + "grad_norm": 8.070756912231445, + "learning_rate": 9.701553836676022e-05, + "loss": 1.0272, + "step": 2931 + }, + { + "epoch": 0.19865844569415272, + "grad_norm": 9.19532585144043, + "learning_rate": 9.701416934766242e-05, + "loss": 1.1195, + "step": 2932 + }, + { + "epoch": 0.19872620096212482, + "grad_norm": 10.692606925964355, + "learning_rate": 9.70128003285646e-05, + "loss": 1.3148, + "step": 2933 + }, + { + "epoch": 0.19879395623009688, + "grad_norm": 10.058424949645996, + "learning_rate": 9.701143130946678e-05, + "loss": 1.4236, + "step": 2934 + }, + { + "epoch": 0.19886171149806897, + "grad_norm": 6.197395324707031, + "learning_rate": 9.701006229036896e-05, + "loss": 0.8822, + "step": 2935 + }, + { + "epoch": 0.19892946676604106, + "grad_norm": 7.603270053863525, + "learning_rate": 9.700869327127114e-05, + "loss": 0.9655, + "step": 2936 + }, + { + "epoch": 0.19899722203401315, + "grad_norm": 8.192676544189453, + "learning_rate": 9.700732425217333e-05, + "loss": 0.9668, + "step": 2937 + }, + { + "epoch": 0.19906497730198522, + "grad_norm": 7.121623992919922, + "learning_rate": 9.700595523307551e-05, + "loss": 1.0037, + "step": 2938 + }, + { + "epoch": 0.1991327325699573, + "grad_norm": 6.8974127769470215, + "learning_rate": 9.700458621397769e-05, + "loss": 0.876, + "step": 2939 + }, + { + "epoch": 0.1992004878379294, + "grad_norm": 7.590656757354736, + "learning_rate": 9.700321719487987e-05, + "loss": 1.0346, + "step": 2940 + }, + { + "epoch": 0.1992682431059015, + "grad_norm": 8.530266761779785, + "learning_rate": 9.700184817578207e-05, + "loss": 1.0882, + "step": 2941 + }, + { + "epoch": 0.19933599837387356, + "grad_norm": 8.064129829406738, + "learning_rate": 9.700047915668425e-05, + "loss": 0.9949, + "step": 2942 + }, + { + "epoch": 0.19940375364184565, + "grad_norm": 7.23117208480835, + "learning_rate": 9.699911013758643e-05, + "loss": 0.7269, + "step": 2943 + }, + { + "epoch": 0.19947150890981774, + "grad_norm": 7.326268196105957, + "learning_rate": 9.69977411184886e-05, + "loss": 0.8542, + "step": 2944 + }, + { + "epoch": 0.19953926417778983, + "grad_norm": 7.708505153656006, + "learning_rate": 9.699637209939079e-05, + "loss": 0.8206, + "step": 2945 + }, + { + "epoch": 0.1996070194457619, + "grad_norm": 10.134513854980469, + "learning_rate": 9.699500308029298e-05, + "loss": 1.0079, + "step": 2946 + }, + { + "epoch": 0.199674774713734, + "grad_norm": 9.264663696289062, + "learning_rate": 9.699363406119516e-05, + "loss": 1.1702, + "step": 2947 + }, + { + "epoch": 0.19974252998170608, + "grad_norm": 8.894827842712402, + "learning_rate": 9.699226504209734e-05, + "loss": 0.9184, + "step": 2948 + }, + { + "epoch": 0.19981028524967817, + "grad_norm": 6.935434341430664, + "learning_rate": 9.699089602299952e-05, + "loss": 0.8554, + "step": 2949 + }, + { + "epoch": 0.19987804051765024, + "grad_norm": 9.57607364654541, + "learning_rate": 9.698952700390172e-05, + "loss": 1.0937, + "step": 2950 + }, + { + "epoch": 0.19994579578562233, + "grad_norm": 7.99752140045166, + "learning_rate": 9.69881579848039e-05, + "loss": 1.0797, + "step": 2951 + }, + { + "epoch": 0.20001355105359442, + "grad_norm": 8.067659378051758, + "learning_rate": 9.698678896570608e-05, + "loss": 1.0515, + "step": 2952 + }, + { + "epoch": 0.2000813063215665, + "grad_norm": 9.66697883605957, + "learning_rate": 9.698541994660826e-05, + "loss": 1.289, + "step": 2953 + }, + { + "epoch": 0.20014906158953857, + "grad_norm": 7.3660664558410645, + "learning_rate": 9.698405092751044e-05, + "loss": 0.7409, + "step": 2954 + }, + { + "epoch": 0.20021681685751067, + "grad_norm": 8.254073143005371, + "learning_rate": 9.698268190841263e-05, + "loss": 1.1095, + "step": 2955 + }, + { + "epoch": 0.20028457212548276, + "grad_norm": 8.221102714538574, + "learning_rate": 9.698131288931481e-05, + "loss": 1.0083, + "step": 2956 + }, + { + "epoch": 0.20035232739345485, + "grad_norm": 9.813411712646484, + "learning_rate": 9.697994387021699e-05, + "loss": 1.2888, + "step": 2957 + }, + { + "epoch": 0.2004200826614269, + "grad_norm": 10.87628173828125, + "learning_rate": 9.697857485111917e-05, + "loss": 1.16, + "step": 2958 + }, + { + "epoch": 0.200487837929399, + "grad_norm": 7.094732284545898, + "learning_rate": 9.697720583202135e-05, + "loss": 0.8959, + "step": 2959 + }, + { + "epoch": 0.2005555931973711, + "grad_norm": 9.814677238464355, + "learning_rate": 9.697583681292355e-05, + "loss": 1.1011, + "step": 2960 + }, + { + "epoch": 0.2006233484653432, + "grad_norm": 8.982966423034668, + "learning_rate": 9.697446779382573e-05, + "loss": 1.001, + "step": 2961 + }, + { + "epoch": 0.20069110373331525, + "grad_norm": 8.49453067779541, + "learning_rate": 9.697309877472791e-05, + "loss": 0.9848, + "step": 2962 + }, + { + "epoch": 0.20075885900128734, + "grad_norm": 7.239814758300781, + "learning_rate": 9.697172975563009e-05, + "loss": 0.8252, + "step": 2963 + }, + { + "epoch": 0.20082661426925943, + "grad_norm": 10.874746322631836, + "learning_rate": 9.697036073653228e-05, + "loss": 1.0683, + "step": 2964 + }, + { + "epoch": 0.20089436953723153, + "grad_norm": 7.679197311401367, + "learning_rate": 9.696899171743446e-05, + "loss": 1.2797, + "step": 2965 + }, + { + "epoch": 0.20096212480520362, + "grad_norm": 10.089177131652832, + "learning_rate": 9.696762269833664e-05, + "loss": 1.1905, + "step": 2966 + }, + { + "epoch": 0.20102988007317568, + "grad_norm": 8.182350158691406, + "learning_rate": 9.696625367923882e-05, + "loss": 0.9917, + "step": 2967 + }, + { + "epoch": 0.20109763534114777, + "grad_norm": 7.8756256103515625, + "learning_rate": 9.6964884660141e-05, + "loss": 1.0368, + "step": 2968 + }, + { + "epoch": 0.20116539060911987, + "grad_norm": 9.193910598754883, + "learning_rate": 9.69635156410432e-05, + "loss": 1.094, + "step": 2969 + }, + { + "epoch": 0.20123314587709196, + "grad_norm": 8.854869842529297, + "learning_rate": 9.696214662194538e-05, + "loss": 1.1344, + "step": 2970 + }, + { + "epoch": 0.20130090114506402, + "grad_norm": 10.419108390808105, + "learning_rate": 9.696077760284756e-05, + "loss": 0.9937, + "step": 2971 + }, + { + "epoch": 0.2013686564130361, + "grad_norm": 9.329347610473633, + "learning_rate": 9.695940858374975e-05, + "loss": 0.9454, + "step": 2972 + }, + { + "epoch": 0.2014364116810082, + "grad_norm": 8.23154067993164, + "learning_rate": 9.695803956465193e-05, + "loss": 0.8691, + "step": 2973 + }, + { + "epoch": 0.2015041669489803, + "grad_norm": 8.016939163208008, + "learning_rate": 9.695667054555411e-05, + "loss": 0.9182, + "step": 2974 + }, + { + "epoch": 0.20157192221695236, + "grad_norm": 9.717400550842285, + "learning_rate": 9.69553015264563e-05, + "loss": 1.1251, + "step": 2975 + }, + { + "epoch": 0.20163967748492445, + "grad_norm": 10.470111846923828, + "learning_rate": 9.695393250735849e-05, + "loss": 1.1634, + "step": 2976 + }, + { + "epoch": 0.20170743275289654, + "grad_norm": 8.540326118469238, + "learning_rate": 9.695256348826067e-05, + "loss": 0.9229, + "step": 2977 + }, + { + "epoch": 0.20177518802086863, + "grad_norm": 6.997597694396973, + "learning_rate": 9.695119446916286e-05, + "loss": 0.9911, + "step": 2978 + }, + { + "epoch": 0.2018429432888407, + "grad_norm": 7.23951530456543, + "learning_rate": 9.694982545006504e-05, + "loss": 0.8372, + "step": 2979 + }, + { + "epoch": 0.2019106985568128, + "grad_norm": 9.217951774597168, + "learning_rate": 9.694845643096722e-05, + "loss": 0.8154, + "step": 2980 + }, + { + "epoch": 0.20197845382478488, + "grad_norm": 8.128033638000488, + "learning_rate": 9.69470874118694e-05, + "loss": 0.9494, + "step": 2981 + }, + { + "epoch": 0.20204620909275697, + "grad_norm": 8.328935623168945, + "learning_rate": 9.694571839277158e-05, + "loss": 1.0537, + "step": 2982 + }, + { + "epoch": 0.20211396436072904, + "grad_norm": 8.877389907836914, + "learning_rate": 9.694434937367378e-05, + "loss": 0.8683, + "step": 2983 + }, + { + "epoch": 0.20218171962870113, + "grad_norm": 7.285436630249023, + "learning_rate": 9.694298035457596e-05, + "loss": 0.7514, + "step": 2984 + }, + { + "epoch": 0.20224947489667322, + "grad_norm": 9.209798812866211, + "learning_rate": 9.694161133547814e-05, + "loss": 1.0771, + "step": 2985 + }, + { + "epoch": 0.2023172301646453, + "grad_norm": 8.625777244567871, + "learning_rate": 9.694024231638032e-05, + "loss": 1.0085, + "step": 2986 + }, + { + "epoch": 0.20238498543261738, + "grad_norm": 8.50123405456543, + "learning_rate": 9.693887329728251e-05, + "loss": 0.969, + "step": 2987 + }, + { + "epoch": 0.20245274070058947, + "grad_norm": 7.314642429351807, + "learning_rate": 9.693750427818469e-05, + "loss": 0.8075, + "step": 2988 + }, + { + "epoch": 0.20252049596856156, + "grad_norm": 9.474241256713867, + "learning_rate": 9.693613525908687e-05, + "loss": 1.0989, + "step": 2989 + }, + { + "epoch": 0.20258825123653365, + "grad_norm": 7.3510637283325195, + "learning_rate": 9.693476623998905e-05, + "loss": 0.8044, + "step": 2990 + }, + { + "epoch": 0.20265600650450571, + "grad_norm": 7.203106880187988, + "learning_rate": 9.693339722089123e-05, + "loss": 1.0359, + "step": 2991 + }, + { + "epoch": 0.2027237617724778, + "grad_norm": 6.71024227142334, + "learning_rate": 9.693202820179343e-05, + "loss": 0.8687, + "step": 2992 + }, + { + "epoch": 0.2027915170404499, + "grad_norm": 8.327759742736816, + "learning_rate": 9.69306591826956e-05, + "loss": 1.0042, + "step": 2993 + }, + { + "epoch": 0.202859272308422, + "grad_norm": 8.682476997375488, + "learning_rate": 9.692929016359779e-05, + "loss": 1.0273, + "step": 2994 + }, + { + "epoch": 0.20292702757639405, + "grad_norm": 8.555792808532715, + "learning_rate": 9.692792114449997e-05, + "loss": 0.8712, + "step": 2995 + }, + { + "epoch": 0.20299478284436615, + "grad_norm": 11.036639213562012, + "learning_rate": 9.692655212540216e-05, + "loss": 1.1452, + "step": 2996 + }, + { + "epoch": 0.20306253811233824, + "grad_norm": 10.207952499389648, + "learning_rate": 9.692518310630434e-05, + "loss": 1.2267, + "step": 2997 + }, + { + "epoch": 0.20313029338031033, + "grad_norm": 8.715107917785645, + "learning_rate": 9.692381408720652e-05, + "loss": 1.151, + "step": 2998 + }, + { + "epoch": 0.2031980486482824, + "grad_norm": 8.72461986541748, + "learning_rate": 9.69224450681087e-05, + "loss": 0.9787, + "step": 2999 + }, + { + "epoch": 0.20326580391625448, + "grad_norm": 8.717243194580078, + "learning_rate": 9.692107604901088e-05, + "loss": 1.0734, + "step": 3000 + }, + { + "epoch": 0.20333355918422658, + "grad_norm": 7.039597511291504, + "learning_rate": 9.691970702991308e-05, + "loss": 0.9215, + "step": 3001 + }, + { + "epoch": 0.20340131445219867, + "grad_norm": 10.568238258361816, + "learning_rate": 9.691833801081526e-05, + "loss": 1.2248, + "step": 3002 + }, + { + "epoch": 0.20346906972017073, + "grad_norm": 9.515549659729004, + "learning_rate": 9.691696899171744e-05, + "loss": 1.1401, + "step": 3003 + }, + { + "epoch": 0.20353682498814282, + "grad_norm": 9.650483131408691, + "learning_rate": 9.691559997261962e-05, + "loss": 0.9839, + "step": 3004 + }, + { + "epoch": 0.20360458025611491, + "grad_norm": 6.817119598388672, + "learning_rate": 9.691423095352181e-05, + "loss": 0.8542, + "step": 3005 + }, + { + "epoch": 0.203672335524087, + "grad_norm": 7.616591930389404, + "learning_rate": 9.691286193442399e-05, + "loss": 0.9647, + "step": 3006 + }, + { + "epoch": 0.20374009079205907, + "grad_norm": 7.2600274085998535, + "learning_rate": 9.691149291532617e-05, + "loss": 0.962, + "step": 3007 + }, + { + "epoch": 0.20380784606003116, + "grad_norm": 9.714008331298828, + "learning_rate": 9.691012389622835e-05, + "loss": 1.1567, + "step": 3008 + }, + { + "epoch": 0.20387560132800325, + "grad_norm": 9.095394134521484, + "learning_rate": 9.690875487713053e-05, + "loss": 1.1315, + "step": 3009 + }, + { + "epoch": 0.20394335659597534, + "grad_norm": 9.683954238891602, + "learning_rate": 9.690738585803273e-05, + "loss": 1.4039, + "step": 3010 + }, + { + "epoch": 0.2040111118639474, + "grad_norm": 8.829015731811523, + "learning_rate": 9.69060168389349e-05, + "loss": 0.9176, + "step": 3011 + }, + { + "epoch": 0.2040788671319195, + "grad_norm": 7.835269927978516, + "learning_rate": 9.690464781983709e-05, + "loss": 1.0077, + "step": 3012 + }, + { + "epoch": 0.2041466223998916, + "grad_norm": 9.94642448425293, + "learning_rate": 9.690327880073927e-05, + "loss": 1.0815, + "step": 3013 + }, + { + "epoch": 0.20421437766786368, + "grad_norm": 8.184757232666016, + "learning_rate": 9.690190978164145e-05, + "loss": 0.84, + "step": 3014 + }, + { + "epoch": 0.20428213293583575, + "grad_norm": 9.060220718383789, + "learning_rate": 9.690054076254364e-05, + "loss": 0.907, + "step": 3015 + }, + { + "epoch": 0.20434988820380784, + "grad_norm": 6.848534107208252, + "learning_rate": 9.689917174344582e-05, + "loss": 0.7549, + "step": 3016 + }, + { + "epoch": 0.20441764347177993, + "grad_norm": 7.820966720581055, + "learning_rate": 9.6897802724348e-05, + "loss": 0.9899, + "step": 3017 + }, + { + "epoch": 0.20448539873975202, + "grad_norm": 10.148963928222656, + "learning_rate": 9.68964337052502e-05, + "loss": 1.2403, + "step": 3018 + }, + { + "epoch": 0.20455315400772411, + "grad_norm": 8.273184776306152, + "learning_rate": 9.689506468615238e-05, + "loss": 1.0002, + "step": 3019 + }, + { + "epoch": 0.20462090927569618, + "grad_norm": 7.618801593780518, + "learning_rate": 9.689369566705456e-05, + "loss": 0.8864, + "step": 3020 + }, + { + "epoch": 0.20468866454366827, + "grad_norm": 7.952611446380615, + "learning_rate": 9.689232664795675e-05, + "loss": 0.8002, + "step": 3021 + }, + { + "epoch": 0.20475641981164036, + "grad_norm": 9.938977241516113, + "learning_rate": 9.689095762885893e-05, + "loss": 0.8956, + "step": 3022 + }, + { + "epoch": 0.20482417507961245, + "grad_norm": 7.807236194610596, + "learning_rate": 9.688958860976111e-05, + "loss": 1.1777, + "step": 3023 + }, + { + "epoch": 0.20489193034758452, + "grad_norm": 7.7249369621276855, + "learning_rate": 9.68882195906633e-05, + "loss": 0.8881, + "step": 3024 + }, + { + "epoch": 0.2049596856155566, + "grad_norm": 7.747461795806885, + "learning_rate": 9.688685057156549e-05, + "loss": 0.9844, + "step": 3025 + }, + { + "epoch": 0.2050274408835287, + "grad_norm": 10.017412185668945, + "learning_rate": 9.688548155246767e-05, + "loss": 1.0815, + "step": 3026 + }, + { + "epoch": 0.2050951961515008, + "grad_norm": 6.54990816116333, + "learning_rate": 9.688411253336985e-05, + "loss": 0.9319, + "step": 3027 + }, + { + "epoch": 0.20516295141947286, + "grad_norm": 7.358734130859375, + "learning_rate": 9.688274351427204e-05, + "loss": 0.9313, + "step": 3028 + }, + { + "epoch": 0.20523070668744495, + "grad_norm": 7.608468055725098, + "learning_rate": 9.688137449517422e-05, + "loss": 1.0071, + "step": 3029 + }, + { + "epoch": 0.20529846195541704, + "grad_norm": 7.013155937194824, + "learning_rate": 9.68800054760764e-05, + "loss": 0.9419, + "step": 3030 + }, + { + "epoch": 0.20536621722338913, + "grad_norm": 9.907796859741211, + "learning_rate": 9.687863645697858e-05, + "loss": 1.2082, + "step": 3031 + }, + { + "epoch": 0.2054339724913612, + "grad_norm": 9.059138298034668, + "learning_rate": 9.687726743788076e-05, + "loss": 1.0712, + "step": 3032 + }, + { + "epoch": 0.20550172775933329, + "grad_norm": 10.519928932189941, + "learning_rate": 9.687589841878296e-05, + "loss": 1.1627, + "step": 3033 + }, + { + "epoch": 0.20556948302730538, + "grad_norm": 8.381184577941895, + "learning_rate": 9.687452939968514e-05, + "loss": 1.087, + "step": 3034 + }, + { + "epoch": 0.20563723829527747, + "grad_norm": 6.514460563659668, + "learning_rate": 9.687316038058732e-05, + "loss": 0.8657, + "step": 3035 + }, + { + "epoch": 0.20570499356324953, + "grad_norm": 10.034708023071289, + "learning_rate": 9.68717913614895e-05, + "loss": 1.1529, + "step": 3036 + }, + { + "epoch": 0.20577274883122162, + "grad_norm": 7.202263355255127, + "learning_rate": 9.687042234239168e-05, + "loss": 0.7923, + "step": 3037 + }, + { + "epoch": 0.20584050409919372, + "grad_norm": 9.7435302734375, + "learning_rate": 9.686905332329387e-05, + "loss": 1.5071, + "step": 3038 + }, + { + "epoch": 0.2059082593671658, + "grad_norm": 9.57016658782959, + "learning_rate": 9.686768430419605e-05, + "loss": 0.883, + "step": 3039 + }, + { + "epoch": 0.20597601463513787, + "grad_norm": 7.3575029373168945, + "learning_rate": 9.686631528509823e-05, + "loss": 0.9999, + "step": 3040 + }, + { + "epoch": 0.20604376990310996, + "grad_norm": 9.0224027633667, + "learning_rate": 9.686494626600041e-05, + "loss": 1.0121, + "step": 3041 + }, + { + "epoch": 0.20611152517108206, + "grad_norm": 11.173224449157715, + "learning_rate": 9.68635772469026e-05, + "loss": 1.0556, + "step": 3042 + }, + { + "epoch": 0.20617928043905415, + "grad_norm": 8.858287811279297, + "learning_rate": 9.686220822780479e-05, + "loss": 0.9377, + "step": 3043 + }, + { + "epoch": 0.2062470357070262, + "grad_norm": 7.3096795082092285, + "learning_rate": 9.686083920870697e-05, + "loss": 0.8808, + "step": 3044 + }, + { + "epoch": 0.2063147909749983, + "grad_norm": 8.700214385986328, + "learning_rate": 9.685947018960915e-05, + "loss": 1.2121, + "step": 3045 + }, + { + "epoch": 0.2063825462429704, + "grad_norm": 10.57944107055664, + "learning_rate": 9.685810117051133e-05, + "loss": 1.0931, + "step": 3046 + }, + { + "epoch": 0.20645030151094249, + "grad_norm": 8.765487670898438, + "learning_rate": 9.685673215141352e-05, + "loss": 1.0522, + "step": 3047 + }, + { + "epoch": 0.20651805677891455, + "grad_norm": 7.717139720916748, + "learning_rate": 9.68553631323157e-05, + "loss": 0.9492, + "step": 3048 + }, + { + "epoch": 0.20658581204688664, + "grad_norm": 9.301026344299316, + "learning_rate": 9.685399411321788e-05, + "loss": 1.0191, + "step": 3049 + }, + { + "epoch": 0.20665356731485873, + "grad_norm": 10.251668930053711, + "learning_rate": 9.685262509412006e-05, + "loss": 0.991, + "step": 3050 + }, + { + "epoch": 0.20672132258283082, + "grad_norm": 11.597551345825195, + "learning_rate": 9.685125607502226e-05, + "loss": 1.0169, + "step": 3051 + }, + { + "epoch": 0.2067890778508029, + "grad_norm": 10.293901443481445, + "learning_rate": 9.684988705592444e-05, + "loss": 1.0623, + "step": 3052 + }, + { + "epoch": 0.20685683311877498, + "grad_norm": 8.496854782104492, + "learning_rate": 9.684851803682662e-05, + "loss": 1.0466, + "step": 3053 + }, + { + "epoch": 0.20692458838674707, + "grad_norm": 6.76383638381958, + "learning_rate": 9.68471490177288e-05, + "loss": 0.8155, + "step": 3054 + }, + { + "epoch": 0.20699234365471916, + "grad_norm": 8.168519020080566, + "learning_rate": 9.684577999863098e-05, + "loss": 0.9663, + "step": 3055 + }, + { + "epoch": 0.20706009892269123, + "grad_norm": 8.0086030960083, + "learning_rate": 9.684441097953317e-05, + "loss": 0.889, + "step": 3056 + }, + { + "epoch": 0.20712785419066332, + "grad_norm": 8.347359657287598, + "learning_rate": 9.684304196043535e-05, + "loss": 0.938, + "step": 3057 + }, + { + "epoch": 0.2071956094586354, + "grad_norm": 8.0283203125, + "learning_rate": 9.684167294133753e-05, + "loss": 0.7633, + "step": 3058 + }, + { + "epoch": 0.2072633647266075, + "grad_norm": 8.509317398071289, + "learning_rate": 9.684030392223971e-05, + "loss": 1.1052, + "step": 3059 + }, + { + "epoch": 0.20733111999457957, + "grad_norm": 9.441505432128906, + "learning_rate": 9.683893490314189e-05, + "loss": 1.1177, + "step": 3060 + }, + { + "epoch": 0.20739887526255166, + "grad_norm": 8.131098747253418, + "learning_rate": 9.683756588404409e-05, + "loss": 0.9272, + "step": 3061 + }, + { + "epoch": 0.20746663053052375, + "grad_norm": 8.423643112182617, + "learning_rate": 9.683619686494627e-05, + "loss": 0.875, + "step": 3062 + }, + { + "epoch": 0.20753438579849584, + "grad_norm": 7.513223171234131, + "learning_rate": 9.683482784584845e-05, + "loss": 1.0659, + "step": 3063 + }, + { + "epoch": 0.2076021410664679, + "grad_norm": 7.5881171226501465, + "learning_rate": 9.683345882675063e-05, + "loss": 0.9466, + "step": 3064 + }, + { + "epoch": 0.20766989633444, + "grad_norm": 8.717775344848633, + "learning_rate": 9.683208980765282e-05, + "loss": 1.1019, + "step": 3065 + }, + { + "epoch": 0.2077376516024121, + "grad_norm": 11.917694091796875, + "learning_rate": 9.6830720788555e-05, + "loss": 1.1085, + "step": 3066 + }, + { + "epoch": 0.20780540687038418, + "grad_norm": 9.28741455078125, + "learning_rate": 9.682935176945718e-05, + "loss": 1.1458, + "step": 3067 + }, + { + "epoch": 0.20787316213835624, + "grad_norm": 6.746860980987549, + "learning_rate": 9.682798275035938e-05, + "loss": 0.929, + "step": 3068 + }, + { + "epoch": 0.20794091740632833, + "grad_norm": 8.48763370513916, + "learning_rate": 9.682661373126156e-05, + "loss": 1.0702, + "step": 3069 + }, + { + "epoch": 0.20800867267430043, + "grad_norm": 8.871308326721191, + "learning_rate": 9.682524471216374e-05, + "loss": 0.9199, + "step": 3070 + }, + { + "epoch": 0.20807642794227252, + "grad_norm": 8.275801658630371, + "learning_rate": 9.682387569306593e-05, + "loss": 1.0756, + "step": 3071 + }, + { + "epoch": 0.2081441832102446, + "grad_norm": 8.985222816467285, + "learning_rate": 9.682250667396811e-05, + "loss": 1.2364, + "step": 3072 + }, + { + "epoch": 0.20821193847821667, + "grad_norm": 7.749682426452637, + "learning_rate": 9.682113765487029e-05, + "loss": 0.9444, + "step": 3073 + }, + { + "epoch": 0.20827969374618877, + "grad_norm": 9.111614227294922, + "learning_rate": 9.681976863577248e-05, + "loss": 1.338, + "step": 3074 + }, + { + "epoch": 0.20834744901416086, + "grad_norm": 7.2874674797058105, + "learning_rate": 9.681839961667466e-05, + "loss": 0.838, + "step": 3075 + }, + { + "epoch": 0.20841520428213295, + "grad_norm": 6.873099327087402, + "learning_rate": 9.681703059757685e-05, + "loss": 0.9194, + "step": 3076 + }, + { + "epoch": 0.208482959550105, + "grad_norm": 7.564418792724609, + "learning_rate": 9.681566157847903e-05, + "loss": 1.1209, + "step": 3077 + }, + { + "epoch": 0.2085507148180771, + "grad_norm": 9.088560104370117, + "learning_rate": 9.68142925593812e-05, + "loss": 0.8425, + "step": 3078 + }, + { + "epoch": 0.2086184700860492, + "grad_norm": 7.182369709014893, + "learning_rate": 9.68129235402834e-05, + "loss": 1.0938, + "step": 3079 + }, + { + "epoch": 0.2086862253540213, + "grad_norm": 8.853677749633789, + "learning_rate": 9.681155452118558e-05, + "loss": 1.2611, + "step": 3080 + }, + { + "epoch": 0.20875398062199335, + "grad_norm": 8.56440258026123, + "learning_rate": 9.681018550208776e-05, + "loss": 1.1112, + "step": 3081 + }, + { + "epoch": 0.20882173588996544, + "grad_norm": 8.356021881103516, + "learning_rate": 9.680881648298994e-05, + "loss": 1.0044, + "step": 3082 + }, + { + "epoch": 0.20888949115793753, + "grad_norm": 9.083736419677734, + "learning_rate": 9.680744746389213e-05, + "loss": 1.2065, + "step": 3083 + }, + { + "epoch": 0.20895724642590963, + "grad_norm": 7.990222454071045, + "learning_rate": 9.680607844479432e-05, + "loss": 1.0179, + "step": 3084 + }, + { + "epoch": 0.2090250016938817, + "grad_norm": 8.381364822387695, + "learning_rate": 9.68047094256965e-05, + "loss": 0.8392, + "step": 3085 + }, + { + "epoch": 0.20909275696185378, + "grad_norm": 9.017950057983398, + "learning_rate": 9.680334040659868e-05, + "loss": 0.9757, + "step": 3086 + }, + { + "epoch": 0.20916051222982587, + "grad_norm": 8.525566101074219, + "learning_rate": 9.680197138750086e-05, + "loss": 1.1521, + "step": 3087 + }, + { + "epoch": 0.20922826749779797, + "grad_norm": 8.432148933410645, + "learning_rate": 9.680060236840305e-05, + "loss": 0.8559, + "step": 3088 + }, + { + "epoch": 0.20929602276577003, + "grad_norm": 9.985367774963379, + "learning_rate": 9.679923334930523e-05, + "loss": 1.2035, + "step": 3089 + }, + { + "epoch": 0.20936377803374212, + "grad_norm": 9.771974563598633, + "learning_rate": 9.679786433020741e-05, + "loss": 1.1614, + "step": 3090 + }, + { + "epoch": 0.2094315333017142, + "grad_norm": 10.646146774291992, + "learning_rate": 9.679649531110959e-05, + "loss": 1.2165, + "step": 3091 + }, + { + "epoch": 0.2094992885696863, + "grad_norm": 8.93340015411377, + "learning_rate": 9.679512629201177e-05, + "loss": 0.9883, + "step": 3092 + }, + { + "epoch": 0.20956704383765837, + "grad_norm": 8.498669624328613, + "learning_rate": 9.679375727291397e-05, + "loss": 1.0318, + "step": 3093 + }, + { + "epoch": 0.20963479910563046, + "grad_norm": 7.536258220672607, + "learning_rate": 9.679238825381615e-05, + "loss": 0.9949, + "step": 3094 + }, + { + "epoch": 0.20970255437360255, + "grad_norm": 8.314896583557129, + "learning_rate": 9.679101923471833e-05, + "loss": 1.0502, + "step": 3095 + }, + { + "epoch": 0.20977030964157464, + "grad_norm": 7.470542907714844, + "learning_rate": 9.67896502156205e-05, + "loss": 0.8871, + "step": 3096 + }, + { + "epoch": 0.2098380649095467, + "grad_norm": 8.951095581054688, + "learning_rate": 9.67882811965227e-05, + "loss": 0.9793, + "step": 3097 + }, + { + "epoch": 0.2099058201775188, + "grad_norm": 7.879035472869873, + "learning_rate": 9.678691217742488e-05, + "loss": 1.0193, + "step": 3098 + }, + { + "epoch": 0.2099735754454909, + "grad_norm": 8.890814781188965, + "learning_rate": 9.678554315832706e-05, + "loss": 0.9738, + "step": 3099 + }, + { + "epoch": 0.21004133071346298, + "grad_norm": 8.863816261291504, + "learning_rate": 9.678417413922924e-05, + "loss": 1.1213, + "step": 3100 + }, + { + "epoch": 0.21010908598143505, + "grad_norm": 9.59538745880127, + "learning_rate": 9.678280512013142e-05, + "loss": 1.0215, + "step": 3101 + }, + { + "epoch": 0.21017684124940714, + "grad_norm": 8.811614990234375, + "learning_rate": 9.678143610103362e-05, + "loss": 0.8703, + "step": 3102 + }, + { + "epoch": 0.21024459651737923, + "grad_norm": 7.274720191955566, + "learning_rate": 9.67800670819358e-05, + "loss": 0.898, + "step": 3103 + }, + { + "epoch": 0.21031235178535132, + "grad_norm": 11.239364624023438, + "learning_rate": 9.677869806283798e-05, + "loss": 0.9823, + "step": 3104 + }, + { + "epoch": 0.21038010705332338, + "grad_norm": 8.807086944580078, + "learning_rate": 9.677732904374016e-05, + "loss": 1.0074, + "step": 3105 + }, + { + "epoch": 0.21044786232129548, + "grad_norm": 9.065536499023438, + "learning_rate": 9.677596002464235e-05, + "loss": 1.1861, + "step": 3106 + }, + { + "epoch": 0.21051561758926757, + "grad_norm": 11.035104751586914, + "learning_rate": 9.677459100554453e-05, + "loss": 1.0067, + "step": 3107 + }, + { + "epoch": 0.21058337285723966, + "grad_norm": 8.010696411132812, + "learning_rate": 9.677322198644671e-05, + "loss": 1.0855, + "step": 3108 + }, + { + "epoch": 0.21065112812521172, + "grad_norm": 9.104195594787598, + "learning_rate": 9.677185296734889e-05, + "loss": 1.0497, + "step": 3109 + }, + { + "epoch": 0.21071888339318381, + "grad_norm": 8.731512069702148, + "learning_rate": 9.677048394825107e-05, + "loss": 1.1108, + "step": 3110 + }, + { + "epoch": 0.2107866386611559, + "grad_norm": 8.823514938354492, + "learning_rate": 9.676911492915327e-05, + "loss": 1.0271, + "step": 3111 + }, + { + "epoch": 0.210854393929128, + "grad_norm": 7.446425914764404, + "learning_rate": 9.676774591005545e-05, + "loss": 0.9182, + "step": 3112 + }, + { + "epoch": 0.21092214919710006, + "grad_norm": 10.75915241241455, + "learning_rate": 9.676637689095763e-05, + "loss": 0.93, + "step": 3113 + }, + { + "epoch": 0.21098990446507215, + "grad_norm": 10.065240859985352, + "learning_rate": 9.676500787185982e-05, + "loss": 1.0508, + "step": 3114 + }, + { + "epoch": 0.21105765973304424, + "grad_norm": 8.790117263793945, + "learning_rate": 9.6763638852762e-05, + "loss": 0.8971, + "step": 3115 + }, + { + "epoch": 0.21112541500101634, + "grad_norm": 8.286596298217773, + "learning_rate": 9.676226983366418e-05, + "loss": 0.9403, + "step": 3116 + }, + { + "epoch": 0.2111931702689884, + "grad_norm": 8.542399406433105, + "learning_rate": 9.676090081456637e-05, + "loss": 1.1638, + "step": 3117 + }, + { + "epoch": 0.2112609255369605, + "grad_norm": 9.461727142333984, + "learning_rate": 9.675953179546856e-05, + "loss": 1.0859, + "step": 3118 + }, + { + "epoch": 0.21132868080493258, + "grad_norm": 7.0917229652404785, + "learning_rate": 9.675816277637074e-05, + "loss": 0.9931, + "step": 3119 + }, + { + "epoch": 0.21139643607290468, + "grad_norm": 9.192744255065918, + "learning_rate": 9.675679375727293e-05, + "loss": 1.2138, + "step": 3120 + }, + { + "epoch": 0.21146419134087674, + "grad_norm": 7.744256973266602, + "learning_rate": 9.675542473817511e-05, + "loss": 1.3615, + "step": 3121 + }, + { + "epoch": 0.21153194660884883, + "grad_norm": 6.915426254272461, + "learning_rate": 9.675405571907729e-05, + "loss": 0.89, + "step": 3122 + }, + { + "epoch": 0.21159970187682092, + "grad_norm": 9.888227462768555, + "learning_rate": 9.675268669997947e-05, + "loss": 0.9569, + "step": 3123 + }, + { + "epoch": 0.21166745714479301, + "grad_norm": 7.541590690612793, + "learning_rate": 9.675131768088165e-05, + "loss": 0.9638, + "step": 3124 + }, + { + "epoch": 0.2117352124127651, + "grad_norm": 7.883132457733154, + "learning_rate": 9.674994866178384e-05, + "loss": 0.8901, + "step": 3125 + }, + { + "epoch": 0.21180296768073717, + "grad_norm": 8.493675231933594, + "learning_rate": 9.674857964268602e-05, + "loss": 0.8894, + "step": 3126 + }, + { + "epoch": 0.21187072294870926, + "grad_norm": 7.348284721374512, + "learning_rate": 9.67472106235882e-05, + "loss": 0.8721, + "step": 3127 + }, + { + "epoch": 0.21193847821668135, + "grad_norm": 9.094710350036621, + "learning_rate": 9.674584160449039e-05, + "loss": 0.9255, + "step": 3128 + }, + { + "epoch": 0.21200623348465344, + "grad_norm": 7.316446304321289, + "learning_rate": 9.674447258539258e-05, + "loss": 1.2577, + "step": 3129 + }, + { + "epoch": 0.2120739887526255, + "grad_norm": 8.105271339416504, + "learning_rate": 9.674310356629476e-05, + "loss": 0.8157, + "step": 3130 + }, + { + "epoch": 0.2121417440205976, + "grad_norm": 8.433457374572754, + "learning_rate": 9.674173454719694e-05, + "loss": 1.0398, + "step": 3131 + }, + { + "epoch": 0.2122094992885697, + "grad_norm": 7.787237644195557, + "learning_rate": 9.674036552809912e-05, + "loss": 0.7095, + "step": 3132 + }, + { + "epoch": 0.21227725455654178, + "grad_norm": 10.60180950164795, + "learning_rate": 9.67389965090013e-05, + "loss": 1.0482, + "step": 3133 + }, + { + "epoch": 0.21234500982451385, + "grad_norm": 8.428773880004883, + "learning_rate": 9.67376274899035e-05, + "loss": 0.8677, + "step": 3134 + }, + { + "epoch": 0.21241276509248594, + "grad_norm": 8.204195022583008, + "learning_rate": 9.673625847080568e-05, + "loss": 1.14, + "step": 3135 + }, + { + "epoch": 0.21248052036045803, + "grad_norm": 6.449087619781494, + "learning_rate": 9.673488945170786e-05, + "loss": 1.0181, + "step": 3136 + }, + { + "epoch": 0.21254827562843012, + "grad_norm": 9.588041305541992, + "learning_rate": 9.673352043261004e-05, + "loss": 0.9996, + "step": 3137 + }, + { + "epoch": 0.21261603089640219, + "grad_norm": 9.626228332519531, + "learning_rate": 9.673215141351223e-05, + "loss": 1.0501, + "step": 3138 + }, + { + "epoch": 0.21268378616437428, + "grad_norm": 9.036309242248535, + "learning_rate": 9.673078239441441e-05, + "loss": 1.093, + "step": 3139 + }, + { + "epoch": 0.21275154143234637, + "grad_norm": 9.415257453918457, + "learning_rate": 9.672941337531659e-05, + "loss": 1.0983, + "step": 3140 + }, + { + "epoch": 0.21281929670031846, + "grad_norm": 8.309000015258789, + "learning_rate": 9.672804435621877e-05, + "loss": 1.0803, + "step": 3141 + }, + { + "epoch": 0.21288705196829052, + "grad_norm": 7.420774459838867, + "learning_rate": 9.672667533712095e-05, + "loss": 1.1027, + "step": 3142 + }, + { + "epoch": 0.21295480723626262, + "grad_norm": 9.291664123535156, + "learning_rate": 9.672530631802314e-05, + "loss": 1.1617, + "step": 3143 + }, + { + "epoch": 0.2130225625042347, + "grad_norm": 7.656317234039307, + "learning_rate": 9.672393729892533e-05, + "loss": 0.9423, + "step": 3144 + }, + { + "epoch": 0.2130903177722068, + "grad_norm": 5.812994003295898, + "learning_rate": 9.67225682798275e-05, + "loss": 0.9187, + "step": 3145 + }, + { + "epoch": 0.21315807304017886, + "grad_norm": 9.492706298828125, + "learning_rate": 9.672119926072969e-05, + "loss": 1.1804, + "step": 3146 + }, + { + "epoch": 0.21322582830815096, + "grad_norm": 9.37720012664795, + "learning_rate": 9.671983024163187e-05, + "loss": 0.8885, + "step": 3147 + }, + { + "epoch": 0.21329358357612305, + "grad_norm": 9.183650016784668, + "learning_rate": 9.671846122253406e-05, + "loss": 0.8967, + "step": 3148 + }, + { + "epoch": 0.21336133884409514, + "grad_norm": 8.037003517150879, + "learning_rate": 9.671709220343624e-05, + "loss": 1.2605, + "step": 3149 + }, + { + "epoch": 0.2134290941120672, + "grad_norm": 8.694345474243164, + "learning_rate": 9.671572318433842e-05, + "loss": 1.003, + "step": 3150 + }, + { + "epoch": 0.2134968493800393, + "grad_norm": 6.429176330566406, + "learning_rate": 9.67143541652406e-05, + "loss": 1.1018, + "step": 3151 + }, + { + "epoch": 0.21356460464801139, + "grad_norm": 7.964774131774902, + "learning_rate": 9.67129851461428e-05, + "loss": 0.9121, + "step": 3152 + }, + { + "epoch": 0.21363235991598348, + "grad_norm": 7.5509033203125, + "learning_rate": 9.671161612704498e-05, + "loss": 0.9577, + "step": 3153 + }, + { + "epoch": 0.21370011518395554, + "grad_norm": 7.968616962432861, + "learning_rate": 9.671024710794716e-05, + "loss": 1.1337, + "step": 3154 + }, + { + "epoch": 0.21376787045192763, + "grad_norm": 7.340758800506592, + "learning_rate": 9.670887808884934e-05, + "loss": 0.8068, + "step": 3155 + }, + { + "epoch": 0.21383562571989972, + "grad_norm": 7.890449523925781, + "learning_rate": 9.670750906975152e-05, + "loss": 1.0263, + "step": 3156 + }, + { + "epoch": 0.21390338098787182, + "grad_norm": 10.908242225646973, + "learning_rate": 9.670614005065371e-05, + "loss": 1.0384, + "step": 3157 + }, + { + "epoch": 0.21397113625584388, + "grad_norm": 7.62697696685791, + "learning_rate": 9.670477103155589e-05, + "loss": 0.918, + "step": 3158 + }, + { + "epoch": 0.21403889152381597, + "grad_norm": 10.487192153930664, + "learning_rate": 9.670340201245807e-05, + "loss": 1.1135, + "step": 3159 + }, + { + "epoch": 0.21410664679178806, + "grad_norm": 7.076718807220459, + "learning_rate": 9.670203299336026e-05, + "loss": 0.9286, + "step": 3160 + }, + { + "epoch": 0.21417440205976015, + "grad_norm": 10.05949592590332, + "learning_rate": 9.670066397426245e-05, + "loss": 1.1317, + "step": 3161 + }, + { + "epoch": 0.21424215732773222, + "grad_norm": 8.656492233276367, + "learning_rate": 9.669929495516463e-05, + "loss": 0.9864, + "step": 3162 + }, + { + "epoch": 0.2143099125957043, + "grad_norm": 10.591787338256836, + "learning_rate": 9.669792593606682e-05, + "loss": 1.0801, + "step": 3163 + }, + { + "epoch": 0.2143776678636764, + "grad_norm": 8.19133472442627, + "learning_rate": 9.6696556916969e-05, + "loss": 0.9377, + "step": 3164 + }, + { + "epoch": 0.2144454231316485, + "grad_norm": 8.402669906616211, + "learning_rate": 9.669518789787118e-05, + "loss": 0.9283, + "step": 3165 + }, + { + "epoch": 0.21451317839962056, + "grad_norm": 9.713315963745117, + "learning_rate": 9.669381887877337e-05, + "loss": 1.1865, + "step": 3166 + }, + { + "epoch": 0.21458093366759265, + "grad_norm": 6.777700424194336, + "learning_rate": 9.669244985967555e-05, + "loss": 0.9094, + "step": 3167 + }, + { + "epoch": 0.21464868893556474, + "grad_norm": 9.381082534790039, + "learning_rate": 9.669108084057773e-05, + "loss": 0.7461, + "step": 3168 + }, + { + "epoch": 0.21471644420353683, + "grad_norm": 8.568790435791016, + "learning_rate": 9.668971182147992e-05, + "loss": 0.7339, + "step": 3169 + }, + { + "epoch": 0.2147841994715089, + "grad_norm": 9.257226943969727, + "learning_rate": 9.66883428023821e-05, + "loss": 1.1892, + "step": 3170 + }, + { + "epoch": 0.214851954739481, + "grad_norm": 8.544146537780762, + "learning_rate": 9.668697378328429e-05, + "loss": 1.0307, + "step": 3171 + }, + { + "epoch": 0.21491971000745308, + "grad_norm": 9.540971755981445, + "learning_rate": 9.668560476418647e-05, + "loss": 0.8464, + "step": 3172 + }, + { + "epoch": 0.21498746527542517, + "grad_norm": 9.180089950561523, + "learning_rate": 9.668423574508865e-05, + "loss": 0.951, + "step": 3173 + }, + { + "epoch": 0.21505522054339726, + "grad_norm": 10.706409454345703, + "learning_rate": 9.668286672599083e-05, + "loss": 1.141, + "step": 3174 + }, + { + "epoch": 0.21512297581136933, + "grad_norm": 9.421865463256836, + "learning_rate": 9.668149770689302e-05, + "loss": 1.0701, + "step": 3175 + }, + { + "epoch": 0.21519073107934142, + "grad_norm": 9.120182037353516, + "learning_rate": 9.66801286877952e-05, + "loss": 1.1618, + "step": 3176 + }, + { + "epoch": 0.2152584863473135, + "grad_norm": 7.879681587219238, + "learning_rate": 9.667875966869738e-05, + "loss": 0.8048, + "step": 3177 + }, + { + "epoch": 0.2153262416152856, + "grad_norm": 7.233819007873535, + "learning_rate": 9.667739064959957e-05, + "loss": 0.9718, + "step": 3178 + }, + { + "epoch": 0.21539399688325767, + "grad_norm": 7.79316520690918, + "learning_rate": 9.667602163050175e-05, + "loss": 1.0999, + "step": 3179 + }, + { + "epoch": 0.21546175215122976, + "grad_norm": 8.873052597045898, + "learning_rate": 9.667465261140394e-05, + "loss": 0.9444, + "step": 3180 + }, + { + "epoch": 0.21552950741920185, + "grad_norm": 7.386806488037109, + "learning_rate": 9.667328359230612e-05, + "loss": 0.8661, + "step": 3181 + }, + { + "epoch": 0.21559726268717394, + "grad_norm": 7.682179927825928, + "learning_rate": 9.66719145732083e-05, + "loss": 0.7904, + "step": 3182 + }, + { + "epoch": 0.215665017955146, + "grad_norm": 7.9192914962768555, + "learning_rate": 9.667054555411048e-05, + "loss": 1.0563, + "step": 3183 + }, + { + "epoch": 0.2157327732231181, + "grad_norm": 6.972530364990234, + "learning_rate": 9.666917653501267e-05, + "loss": 0.7482, + "step": 3184 + }, + { + "epoch": 0.2158005284910902, + "grad_norm": 9.844091415405273, + "learning_rate": 9.666780751591485e-05, + "loss": 1.0289, + "step": 3185 + }, + { + "epoch": 0.21586828375906228, + "grad_norm": 8.724159240722656, + "learning_rate": 9.666643849681704e-05, + "loss": 1.1419, + "step": 3186 + }, + { + "epoch": 0.21593603902703434, + "grad_norm": 8.859795570373535, + "learning_rate": 9.666506947771922e-05, + "loss": 0.9252, + "step": 3187 + }, + { + "epoch": 0.21600379429500643, + "grad_norm": 10.741375923156738, + "learning_rate": 9.66637004586214e-05, + "loss": 1.2328, + "step": 3188 + }, + { + "epoch": 0.21607154956297853, + "grad_norm": 8.468697547912598, + "learning_rate": 9.666233143952359e-05, + "loss": 1.1779, + "step": 3189 + }, + { + "epoch": 0.21613930483095062, + "grad_norm": 7.220376014709473, + "learning_rate": 9.666096242042577e-05, + "loss": 0.9271, + "step": 3190 + }, + { + "epoch": 0.21620706009892268, + "grad_norm": 8.3795804977417, + "learning_rate": 9.665959340132795e-05, + "loss": 1.0941, + "step": 3191 + }, + { + "epoch": 0.21627481536689477, + "grad_norm": 9.743870735168457, + "learning_rate": 9.665822438223013e-05, + "loss": 0.9396, + "step": 3192 + }, + { + "epoch": 0.21634257063486687, + "grad_norm": 8.927998542785645, + "learning_rate": 9.665685536313231e-05, + "loss": 1.0915, + "step": 3193 + }, + { + "epoch": 0.21641032590283896, + "grad_norm": 8.214877128601074, + "learning_rate": 9.66554863440345e-05, + "loss": 0.8767, + "step": 3194 + }, + { + "epoch": 0.21647808117081102, + "grad_norm": 9.605053901672363, + "learning_rate": 9.665411732493669e-05, + "loss": 1.1829, + "step": 3195 + }, + { + "epoch": 0.2165458364387831, + "grad_norm": 8.430122375488281, + "learning_rate": 9.665274830583887e-05, + "loss": 0.8234, + "step": 3196 + }, + { + "epoch": 0.2166135917067552, + "grad_norm": 8.868674278259277, + "learning_rate": 9.665137928674105e-05, + "loss": 1.0467, + "step": 3197 + }, + { + "epoch": 0.2166813469747273, + "grad_norm": 8.451126098632812, + "learning_rate": 9.665001026764324e-05, + "loss": 1.134, + "step": 3198 + }, + { + "epoch": 0.21674910224269936, + "grad_norm": 8.749373435974121, + "learning_rate": 9.664864124854542e-05, + "loss": 0.9825, + "step": 3199 + }, + { + "epoch": 0.21681685751067145, + "grad_norm": 8.478997230529785, + "learning_rate": 9.66472722294476e-05, + "loss": 0.9823, + "step": 3200 + }, + { + "epoch": 0.21688461277864354, + "grad_norm": 6.554266929626465, + "learning_rate": 9.664590321034978e-05, + "loss": 1.2293, + "step": 3201 + }, + { + "epoch": 0.21695236804661563, + "grad_norm": 7.415618896484375, + "learning_rate": 9.664453419125196e-05, + "loss": 0.9767, + "step": 3202 + }, + { + "epoch": 0.2170201233145877, + "grad_norm": 9.277578353881836, + "learning_rate": 9.664316517215416e-05, + "loss": 1.0042, + "step": 3203 + }, + { + "epoch": 0.2170878785825598, + "grad_norm": 7.1379241943359375, + "learning_rate": 9.664179615305634e-05, + "loss": 0.924, + "step": 3204 + }, + { + "epoch": 0.21715563385053188, + "grad_norm": 8.837613105773926, + "learning_rate": 9.664042713395852e-05, + "loss": 0.9889, + "step": 3205 + }, + { + "epoch": 0.21722338911850397, + "grad_norm": 8.118345260620117, + "learning_rate": 9.663905811486071e-05, + "loss": 1.2376, + "step": 3206 + }, + { + "epoch": 0.21729114438647604, + "grad_norm": 10.383713722229004, + "learning_rate": 9.663768909576289e-05, + "loss": 0.7988, + "step": 3207 + }, + { + "epoch": 0.21735889965444813, + "grad_norm": 7.0938262939453125, + "learning_rate": 9.663632007666507e-05, + "loss": 0.8589, + "step": 3208 + }, + { + "epoch": 0.21742665492242022, + "grad_norm": 7.732020854949951, + "learning_rate": 9.663495105756726e-05, + "loss": 1.0837, + "step": 3209 + }, + { + "epoch": 0.2174944101903923, + "grad_norm": 7.718788146972656, + "learning_rate": 9.663358203846944e-05, + "loss": 0.8949, + "step": 3210 + }, + { + "epoch": 0.21756216545836438, + "grad_norm": 9.187674522399902, + "learning_rate": 9.663221301937162e-05, + "loss": 0.9443, + "step": 3211 + }, + { + "epoch": 0.21762992072633647, + "grad_norm": 8.596949577331543, + "learning_rate": 9.663084400027382e-05, + "loss": 1.0855, + "step": 3212 + }, + { + "epoch": 0.21769767599430856, + "grad_norm": 7.854368209838867, + "learning_rate": 9.6629474981176e-05, + "loss": 0.9203, + "step": 3213 + }, + { + "epoch": 0.21776543126228065, + "grad_norm": 9.96597957611084, + "learning_rate": 9.662810596207818e-05, + "loss": 0.836, + "step": 3214 + }, + { + "epoch": 0.21783318653025271, + "grad_norm": 7.254636764526367, + "learning_rate": 9.662673694298036e-05, + "loss": 0.7926, + "step": 3215 + }, + { + "epoch": 0.2179009417982248, + "grad_norm": 9.079703330993652, + "learning_rate": 9.662536792388255e-05, + "loss": 1.2072, + "step": 3216 + }, + { + "epoch": 0.2179686970661969, + "grad_norm": 8.051166534423828, + "learning_rate": 9.662399890478473e-05, + "loss": 0.9719, + "step": 3217 + }, + { + "epoch": 0.218036452334169, + "grad_norm": 9.189994812011719, + "learning_rate": 9.662262988568691e-05, + "loss": 0.9878, + "step": 3218 + }, + { + "epoch": 0.21810420760214105, + "grad_norm": 9.806111335754395, + "learning_rate": 9.66212608665891e-05, + "loss": 1.1621, + "step": 3219 + }, + { + "epoch": 0.21817196287011315, + "grad_norm": 7.833159923553467, + "learning_rate": 9.661989184749128e-05, + "loss": 0.9811, + "step": 3220 + }, + { + "epoch": 0.21823971813808524, + "grad_norm": 7.537334442138672, + "learning_rate": 9.661852282839347e-05, + "loss": 0.9916, + "step": 3221 + }, + { + "epoch": 0.21830747340605733, + "grad_norm": 9.828896522521973, + "learning_rate": 9.661715380929565e-05, + "loss": 1.09, + "step": 3222 + }, + { + "epoch": 0.2183752286740294, + "grad_norm": 7.99267578125, + "learning_rate": 9.661578479019783e-05, + "loss": 1.0852, + "step": 3223 + }, + { + "epoch": 0.21844298394200148, + "grad_norm": 8.085976600646973, + "learning_rate": 9.661441577110001e-05, + "loss": 1.1762, + "step": 3224 + }, + { + "epoch": 0.21851073920997358, + "grad_norm": 6.523219585418701, + "learning_rate": 9.661304675200219e-05, + "loss": 1.0814, + "step": 3225 + }, + { + "epoch": 0.21857849447794567, + "grad_norm": 9.796062469482422, + "learning_rate": 9.661167773290438e-05, + "loss": 1.1447, + "step": 3226 + }, + { + "epoch": 0.21864624974591776, + "grad_norm": 8.570687294006348, + "learning_rate": 9.661030871380656e-05, + "loss": 1.0552, + "step": 3227 + }, + { + "epoch": 0.21871400501388982, + "grad_norm": 9.70113468170166, + "learning_rate": 9.660893969470874e-05, + "loss": 1.0379, + "step": 3228 + }, + { + "epoch": 0.21878176028186191, + "grad_norm": 8.195375442504883, + "learning_rate": 9.660757067561093e-05, + "loss": 0.9348, + "step": 3229 + }, + { + "epoch": 0.218849515549834, + "grad_norm": 7.33494758605957, + "learning_rate": 9.660620165651312e-05, + "loss": 0.8977, + "step": 3230 + }, + { + "epoch": 0.2189172708178061, + "grad_norm": 7.543430328369141, + "learning_rate": 9.66048326374153e-05, + "loss": 1.023, + "step": 3231 + }, + { + "epoch": 0.21898502608577816, + "grad_norm": 8.45494556427002, + "learning_rate": 9.660346361831748e-05, + "loss": 1.2995, + "step": 3232 + }, + { + "epoch": 0.21905278135375025, + "grad_norm": 8.197903633117676, + "learning_rate": 9.660209459921966e-05, + "loss": 0.9651, + "step": 3233 + }, + { + "epoch": 0.21912053662172234, + "grad_norm": 8.392546653747559, + "learning_rate": 9.660072558012184e-05, + "loss": 0.9858, + "step": 3234 + }, + { + "epoch": 0.21918829188969444, + "grad_norm": 6.499540328979492, + "learning_rate": 9.659935656102403e-05, + "loss": 0.9679, + "step": 3235 + }, + { + "epoch": 0.2192560471576665, + "grad_norm": 7.784618377685547, + "learning_rate": 9.659798754192621e-05, + "loss": 0.9316, + "step": 3236 + }, + { + "epoch": 0.2193238024256386, + "grad_norm": 9.014801979064941, + "learning_rate": 9.65966185228284e-05, + "loss": 0.9261, + "step": 3237 + }, + { + "epoch": 0.21939155769361068, + "grad_norm": 8.238675117492676, + "learning_rate": 9.659524950373058e-05, + "loss": 1.0156, + "step": 3238 + }, + { + "epoch": 0.21945931296158278, + "grad_norm": 9.447864532470703, + "learning_rate": 9.659388048463277e-05, + "loss": 1.2186, + "step": 3239 + }, + { + "epoch": 0.21952706822955484, + "grad_norm": 6.894035816192627, + "learning_rate": 9.659251146553495e-05, + "loss": 0.9157, + "step": 3240 + }, + { + "epoch": 0.21959482349752693, + "grad_norm": 7.24024772644043, + "learning_rate": 9.659114244643713e-05, + "loss": 1.1564, + "step": 3241 + }, + { + "epoch": 0.21966257876549902, + "grad_norm": 9.135485649108887, + "learning_rate": 9.658977342733931e-05, + "loss": 1.2002, + "step": 3242 + }, + { + "epoch": 0.21973033403347111, + "grad_norm": 6.119181156158447, + "learning_rate": 9.658840440824149e-05, + "loss": 0.788, + "step": 3243 + }, + { + "epoch": 0.21979808930144318, + "grad_norm": 8.550108909606934, + "learning_rate": 9.658703538914368e-05, + "loss": 0.9251, + "step": 3244 + }, + { + "epoch": 0.21986584456941527, + "grad_norm": 9.546792984008789, + "learning_rate": 9.658566637004586e-05, + "loss": 1.2231, + "step": 3245 + }, + { + "epoch": 0.21993359983738736, + "grad_norm": 9.451306343078613, + "learning_rate": 9.658429735094805e-05, + "loss": 0.8991, + "step": 3246 + }, + { + "epoch": 0.22000135510535945, + "grad_norm": 9.914873123168945, + "learning_rate": 9.658292833185023e-05, + "loss": 1.3102, + "step": 3247 + }, + { + "epoch": 0.22006911037333152, + "grad_norm": 10.038002967834473, + "learning_rate": 9.65815593127524e-05, + "loss": 1.254, + "step": 3248 + }, + { + "epoch": 0.2201368656413036, + "grad_norm": 7.590802192687988, + "learning_rate": 9.65801902936546e-05, + "loss": 1.0032, + "step": 3249 + }, + { + "epoch": 0.2202046209092757, + "grad_norm": 8.632338523864746, + "learning_rate": 9.657882127455678e-05, + "loss": 0.8922, + "step": 3250 + }, + { + "epoch": 0.2202723761772478, + "grad_norm": 6.483764171600342, + "learning_rate": 9.657745225545896e-05, + "loss": 0.9167, + "step": 3251 + }, + { + "epoch": 0.22034013144521986, + "grad_norm": 8.28577709197998, + "learning_rate": 9.657608323636115e-05, + "loss": 1.0413, + "step": 3252 + }, + { + "epoch": 0.22040788671319195, + "grad_norm": 9.218433380126953, + "learning_rate": 9.657471421726333e-05, + "loss": 1.1681, + "step": 3253 + }, + { + "epoch": 0.22047564198116404, + "grad_norm": 6.653755187988281, + "learning_rate": 9.657334519816552e-05, + "loss": 0.9645, + "step": 3254 + }, + { + "epoch": 0.22054339724913613, + "grad_norm": 8.784723281860352, + "learning_rate": 9.657197617906771e-05, + "loss": 0.7489, + "step": 3255 + }, + { + "epoch": 0.2206111525171082, + "grad_norm": 9.041324615478516, + "learning_rate": 9.657060715996989e-05, + "loss": 1.0478, + "step": 3256 + }, + { + "epoch": 0.22067890778508029, + "grad_norm": 7.811746597290039, + "learning_rate": 9.656923814087207e-05, + "loss": 0.9511, + "step": 3257 + }, + { + "epoch": 0.22074666305305238, + "grad_norm": 9.811897277832031, + "learning_rate": 9.656786912177426e-05, + "loss": 1.4235, + "step": 3258 + }, + { + "epoch": 0.22081441832102447, + "grad_norm": 7.313972473144531, + "learning_rate": 9.656650010267644e-05, + "loss": 1.0075, + "step": 3259 + }, + { + "epoch": 0.22088217358899653, + "grad_norm": 9.360795021057129, + "learning_rate": 9.656513108357862e-05, + "loss": 0.9634, + "step": 3260 + }, + { + "epoch": 0.22094992885696862, + "grad_norm": 9.322724342346191, + "learning_rate": 9.65637620644808e-05, + "loss": 1.0669, + "step": 3261 + }, + { + "epoch": 0.22101768412494072, + "grad_norm": 10.269414901733398, + "learning_rate": 9.6562393045383e-05, + "loss": 1.1288, + "step": 3262 + }, + { + "epoch": 0.2210854393929128, + "grad_norm": 7.672987937927246, + "learning_rate": 9.656102402628518e-05, + "loss": 1.0517, + "step": 3263 + }, + { + "epoch": 0.22115319466088487, + "grad_norm": 8.725695610046387, + "learning_rate": 9.655965500718736e-05, + "loss": 1.0987, + "step": 3264 + }, + { + "epoch": 0.22122094992885696, + "grad_norm": 8.577162742614746, + "learning_rate": 9.655828598808954e-05, + "loss": 0.821, + "step": 3265 + }, + { + "epoch": 0.22128870519682906, + "grad_norm": 8.497530937194824, + "learning_rate": 9.655691696899172e-05, + "loss": 0.9171, + "step": 3266 + }, + { + "epoch": 0.22135646046480115, + "grad_norm": 9.665203094482422, + "learning_rate": 9.655554794989391e-05, + "loss": 1.105, + "step": 3267 + }, + { + "epoch": 0.2214242157327732, + "grad_norm": 9.42651081085205, + "learning_rate": 9.65541789307961e-05, + "loss": 0.9526, + "step": 3268 + }, + { + "epoch": 0.2214919710007453, + "grad_norm": 7.461474418640137, + "learning_rate": 9.655280991169827e-05, + "loss": 0.9554, + "step": 3269 + }, + { + "epoch": 0.2215597262687174, + "grad_norm": 10.56189250946045, + "learning_rate": 9.655144089260045e-05, + "loss": 1.1504, + "step": 3270 + }, + { + "epoch": 0.22162748153668949, + "grad_norm": 8.245696067810059, + "learning_rate": 9.655007187350265e-05, + "loss": 1.0352, + "step": 3271 + }, + { + "epoch": 0.22169523680466155, + "grad_norm": 9.182881355285645, + "learning_rate": 9.654870285440483e-05, + "loss": 1.1153, + "step": 3272 + }, + { + "epoch": 0.22176299207263364, + "grad_norm": 8.101743698120117, + "learning_rate": 9.654733383530701e-05, + "loss": 1.1399, + "step": 3273 + }, + { + "epoch": 0.22183074734060573, + "grad_norm": 9.072972297668457, + "learning_rate": 9.654596481620919e-05, + "loss": 1.1508, + "step": 3274 + }, + { + "epoch": 0.22189850260857782, + "grad_norm": 8.480910301208496, + "learning_rate": 9.654459579711137e-05, + "loss": 0.9659, + "step": 3275 + }, + { + "epoch": 0.2219662578765499, + "grad_norm": 7.095332622528076, + "learning_rate": 9.654322677801356e-05, + "loss": 0.9417, + "step": 3276 + }, + { + "epoch": 0.22203401314452198, + "grad_norm": 8.183112144470215, + "learning_rate": 9.654185775891574e-05, + "loss": 0.9307, + "step": 3277 + }, + { + "epoch": 0.22210176841249407, + "grad_norm": 8.60648250579834, + "learning_rate": 9.654048873981792e-05, + "loss": 1.141, + "step": 3278 + }, + { + "epoch": 0.22216952368046616, + "grad_norm": 6.992738246917725, + "learning_rate": 9.65391197207201e-05, + "loss": 0.9429, + "step": 3279 + }, + { + "epoch": 0.22223727894843825, + "grad_norm": 6.998913764953613, + "learning_rate": 9.653775070162229e-05, + "loss": 1.0159, + "step": 3280 + }, + { + "epoch": 0.22230503421641032, + "grad_norm": 7.922050952911377, + "learning_rate": 9.653638168252448e-05, + "loss": 1.1349, + "step": 3281 + }, + { + "epoch": 0.2223727894843824, + "grad_norm": 9.171984672546387, + "learning_rate": 9.653501266342666e-05, + "loss": 1.149, + "step": 3282 + }, + { + "epoch": 0.2224405447523545, + "grad_norm": 9.88930606842041, + "learning_rate": 9.653364364432884e-05, + "loss": 1.1036, + "step": 3283 + }, + { + "epoch": 0.2225083000203266, + "grad_norm": 10.978727340698242, + "learning_rate": 9.653227462523102e-05, + "loss": 1.0345, + "step": 3284 + }, + { + "epoch": 0.22257605528829866, + "grad_norm": 8.712138175964355, + "learning_rate": 9.653090560613321e-05, + "loss": 1.1404, + "step": 3285 + }, + { + "epoch": 0.22264381055627075, + "grad_norm": 8.79491901397705, + "learning_rate": 9.65295365870354e-05, + "loss": 1.4334, + "step": 3286 + }, + { + "epoch": 0.22271156582424284, + "grad_norm": 9.580904960632324, + "learning_rate": 9.652816756793757e-05, + "loss": 1.0693, + "step": 3287 + }, + { + "epoch": 0.22277932109221493, + "grad_norm": 7.335433006286621, + "learning_rate": 9.652679854883976e-05, + "loss": 1.0559, + "step": 3288 + }, + { + "epoch": 0.222847076360187, + "grad_norm": 8.534089088439941, + "learning_rate": 9.652542952974194e-05, + "loss": 0.9835, + "step": 3289 + }, + { + "epoch": 0.2229148316281591, + "grad_norm": 6.626898288726807, + "learning_rate": 9.652406051064413e-05, + "loss": 0.8056, + "step": 3290 + }, + { + "epoch": 0.22298258689613118, + "grad_norm": 7.774187088012695, + "learning_rate": 9.652269149154631e-05, + "loss": 0.8032, + "step": 3291 + }, + { + "epoch": 0.22305034216410327, + "grad_norm": 8.029982566833496, + "learning_rate": 9.652132247244849e-05, + "loss": 1.0536, + "step": 3292 + }, + { + "epoch": 0.22311809743207534, + "grad_norm": 8.701712608337402, + "learning_rate": 9.651995345335067e-05, + "loss": 0.9422, + "step": 3293 + }, + { + "epoch": 0.22318585270004743, + "grad_norm": 9.17507553100586, + "learning_rate": 9.651858443425286e-05, + "loss": 1.0094, + "step": 3294 + }, + { + "epoch": 0.22325360796801952, + "grad_norm": 8.464656829833984, + "learning_rate": 9.651721541515504e-05, + "loss": 0.9546, + "step": 3295 + }, + { + "epoch": 0.2233213632359916, + "grad_norm": 6.839638710021973, + "learning_rate": 9.651584639605722e-05, + "loss": 0.8296, + "step": 3296 + }, + { + "epoch": 0.22338911850396367, + "grad_norm": 7.773430824279785, + "learning_rate": 9.65144773769594e-05, + "loss": 1.1344, + "step": 3297 + }, + { + "epoch": 0.22345687377193577, + "grad_norm": 9.725523948669434, + "learning_rate": 9.65131083578616e-05, + "loss": 0.9814, + "step": 3298 + }, + { + "epoch": 0.22352462903990786, + "grad_norm": 8.995366096496582, + "learning_rate": 9.651173933876378e-05, + "loss": 1.2263, + "step": 3299 + }, + { + "epoch": 0.22359238430787995, + "grad_norm": 8.735838890075684, + "learning_rate": 9.651037031966596e-05, + "loss": 1.1707, + "step": 3300 + }, + { + "epoch": 0.223660139575852, + "grad_norm": 7.177740097045898, + "learning_rate": 9.650900130056815e-05, + "loss": 0.9491, + "step": 3301 + }, + { + "epoch": 0.2237278948438241, + "grad_norm": 8.393296241760254, + "learning_rate": 9.650763228147033e-05, + "loss": 0.9765, + "step": 3302 + }, + { + "epoch": 0.2237956501117962, + "grad_norm": 9.773447036743164, + "learning_rate": 9.650626326237251e-05, + "loss": 1.0409, + "step": 3303 + }, + { + "epoch": 0.2238634053797683, + "grad_norm": 6.951483249664307, + "learning_rate": 9.650489424327471e-05, + "loss": 0.9848, + "step": 3304 + }, + { + "epoch": 0.22393116064774035, + "grad_norm": 8.579717636108398, + "learning_rate": 9.650352522417689e-05, + "loss": 0.9722, + "step": 3305 + }, + { + "epoch": 0.22399891591571244, + "grad_norm": 7.798871040344238, + "learning_rate": 9.650215620507907e-05, + "loss": 0.8046, + "step": 3306 + }, + { + "epoch": 0.22406667118368453, + "grad_norm": 9.504851341247559, + "learning_rate": 9.650078718598125e-05, + "loss": 1.1972, + "step": 3307 + }, + { + "epoch": 0.22413442645165663, + "grad_norm": 8.384056091308594, + "learning_rate": 9.649941816688344e-05, + "loss": 1.0654, + "step": 3308 + }, + { + "epoch": 0.2242021817196287, + "grad_norm": 10.469207763671875, + "learning_rate": 9.649804914778562e-05, + "loss": 1.0694, + "step": 3309 + }, + { + "epoch": 0.22426993698760078, + "grad_norm": 8.5567045211792, + "learning_rate": 9.64966801286878e-05, + "loss": 1.0371, + "step": 3310 + }, + { + "epoch": 0.22433769225557287, + "grad_norm": 9.427155494689941, + "learning_rate": 9.649531110958998e-05, + "loss": 1.0185, + "step": 3311 + }, + { + "epoch": 0.22440544752354497, + "grad_norm": 8.238412857055664, + "learning_rate": 9.649394209049216e-05, + "loss": 0.9744, + "step": 3312 + }, + { + "epoch": 0.22447320279151703, + "grad_norm": 9.36423110961914, + "learning_rate": 9.649257307139436e-05, + "loss": 0.9184, + "step": 3313 + }, + { + "epoch": 0.22454095805948912, + "grad_norm": 7.485356330871582, + "learning_rate": 9.649120405229654e-05, + "loss": 0.9948, + "step": 3314 + }, + { + "epoch": 0.2246087133274612, + "grad_norm": 7.247208118438721, + "learning_rate": 9.648983503319872e-05, + "loss": 0.8592, + "step": 3315 + }, + { + "epoch": 0.2246764685954333, + "grad_norm": 9.266085624694824, + "learning_rate": 9.64884660141009e-05, + "loss": 1.0314, + "step": 3316 + }, + { + "epoch": 0.22474422386340537, + "grad_norm": 11.147274017333984, + "learning_rate": 9.64870969950031e-05, + "loss": 1.0403, + "step": 3317 + }, + { + "epoch": 0.22481197913137746, + "grad_norm": 10.961670875549316, + "learning_rate": 9.648572797590527e-05, + "loss": 1.3147, + "step": 3318 + }, + { + "epoch": 0.22487973439934955, + "grad_norm": 7.002868175506592, + "learning_rate": 9.648435895680745e-05, + "loss": 0.9044, + "step": 3319 + }, + { + "epoch": 0.22494748966732164, + "grad_norm": 7.886247634887695, + "learning_rate": 9.648298993770963e-05, + "loss": 0.8901, + "step": 3320 + }, + { + "epoch": 0.2250152449352937, + "grad_norm": 8.933539390563965, + "learning_rate": 9.648162091861181e-05, + "loss": 0.8758, + "step": 3321 + }, + { + "epoch": 0.2250830002032658, + "grad_norm": 7.223681449890137, + "learning_rate": 9.648025189951401e-05, + "loss": 1.1502, + "step": 3322 + }, + { + "epoch": 0.2251507554712379, + "grad_norm": 8.808623313903809, + "learning_rate": 9.647888288041619e-05, + "loss": 1.0706, + "step": 3323 + }, + { + "epoch": 0.22521851073920998, + "grad_norm": 9.557942390441895, + "learning_rate": 9.647751386131837e-05, + "loss": 0.8113, + "step": 3324 + }, + { + "epoch": 0.22528626600718205, + "grad_norm": 9.855717658996582, + "learning_rate": 9.647614484222055e-05, + "loss": 1.251, + "step": 3325 + }, + { + "epoch": 0.22535402127515414, + "grad_norm": 9.588946342468262, + "learning_rate": 9.647477582312273e-05, + "loss": 1.3029, + "step": 3326 + }, + { + "epoch": 0.22542177654312623, + "grad_norm": 9.277341842651367, + "learning_rate": 9.647340680402492e-05, + "loss": 0.9708, + "step": 3327 + }, + { + "epoch": 0.22548953181109832, + "grad_norm": 9.607316970825195, + "learning_rate": 9.64720377849271e-05, + "loss": 0.9614, + "step": 3328 + }, + { + "epoch": 0.22555728707907038, + "grad_norm": 7.949220657348633, + "learning_rate": 9.647066876582928e-05, + "loss": 1.0066, + "step": 3329 + }, + { + "epoch": 0.22562504234704248, + "grad_norm": 8.383744239807129, + "learning_rate": 9.646929974673146e-05, + "loss": 0.797, + "step": 3330 + }, + { + "epoch": 0.22569279761501457, + "grad_norm": 7.203914642333984, + "learning_rate": 9.646793072763366e-05, + "loss": 0.9625, + "step": 3331 + }, + { + "epoch": 0.22576055288298666, + "grad_norm": 8.95102596282959, + "learning_rate": 9.646656170853584e-05, + "loss": 1.0728, + "step": 3332 + }, + { + "epoch": 0.22582830815095875, + "grad_norm": 8.044751167297363, + "learning_rate": 9.646519268943802e-05, + "loss": 1.2155, + "step": 3333 + }, + { + "epoch": 0.22589606341893081, + "grad_norm": 9.164902687072754, + "learning_rate": 9.64638236703402e-05, + "loss": 1.2923, + "step": 3334 + }, + { + "epoch": 0.2259638186869029, + "grad_norm": 6.792164325714111, + "learning_rate": 9.646245465124238e-05, + "loss": 0.9956, + "step": 3335 + }, + { + "epoch": 0.226031573954875, + "grad_norm": 7.76467752456665, + "learning_rate": 9.646108563214457e-05, + "loss": 1.0057, + "step": 3336 + }, + { + "epoch": 0.2260993292228471, + "grad_norm": 8.541545867919922, + "learning_rate": 9.645971661304675e-05, + "loss": 1.1013, + "step": 3337 + }, + { + "epoch": 0.22616708449081915, + "grad_norm": 7.8545050621032715, + "learning_rate": 9.645834759394893e-05, + "loss": 0.8565, + "step": 3338 + }, + { + "epoch": 0.22623483975879125, + "grad_norm": 9.322896003723145, + "learning_rate": 9.645697857485112e-05, + "loss": 1.0509, + "step": 3339 + }, + { + "epoch": 0.22630259502676334, + "grad_norm": 9.540191650390625, + "learning_rate": 9.645560955575331e-05, + "loss": 1.1757, + "step": 3340 + }, + { + "epoch": 0.22637035029473543, + "grad_norm": 8.97028923034668, + "learning_rate": 9.645424053665549e-05, + "loss": 1.0342, + "step": 3341 + }, + { + "epoch": 0.2264381055627075, + "grad_norm": 6.444105625152588, + "learning_rate": 9.645287151755767e-05, + "loss": 0.8377, + "step": 3342 + }, + { + "epoch": 0.22650586083067958, + "grad_norm": 9.237889289855957, + "learning_rate": 9.645150249845985e-05, + "loss": 1.0468, + "step": 3343 + }, + { + "epoch": 0.22657361609865168, + "grad_norm": 9.567046165466309, + "learning_rate": 9.645013347936203e-05, + "loss": 1.3875, + "step": 3344 + }, + { + "epoch": 0.22664137136662377, + "grad_norm": 8.302481651306152, + "learning_rate": 9.644876446026422e-05, + "loss": 1.0455, + "step": 3345 + }, + { + "epoch": 0.22670912663459583, + "grad_norm": 9.124307632446289, + "learning_rate": 9.64473954411664e-05, + "loss": 0.9828, + "step": 3346 + }, + { + "epoch": 0.22677688190256792, + "grad_norm": 7.700011730194092, + "learning_rate": 9.644602642206858e-05, + "loss": 0.8475, + "step": 3347 + }, + { + "epoch": 0.22684463717054001, + "grad_norm": 6.1064276695251465, + "learning_rate": 9.644465740297078e-05, + "loss": 0.7779, + "step": 3348 + }, + { + "epoch": 0.2269123924385121, + "grad_norm": 8.924312591552734, + "learning_rate": 9.644328838387296e-05, + "loss": 1.3825, + "step": 3349 + }, + { + "epoch": 0.22698014770648417, + "grad_norm": 8.169050216674805, + "learning_rate": 9.644191936477514e-05, + "loss": 1.1198, + "step": 3350 + }, + { + "epoch": 0.22704790297445626, + "grad_norm": 8.368382453918457, + "learning_rate": 9.644055034567733e-05, + "loss": 1.0729, + "step": 3351 + }, + { + "epoch": 0.22711565824242835, + "grad_norm": 10.206897735595703, + "learning_rate": 9.643918132657951e-05, + "loss": 0.9022, + "step": 3352 + }, + { + "epoch": 0.22718341351040044, + "grad_norm": 6.288288116455078, + "learning_rate": 9.64378123074817e-05, + "loss": 0.7868, + "step": 3353 + }, + { + "epoch": 0.2272511687783725, + "grad_norm": 8.446578025817871, + "learning_rate": 9.643644328838389e-05, + "loss": 1.0426, + "step": 3354 + }, + { + "epoch": 0.2273189240463446, + "grad_norm": 8.425249099731445, + "learning_rate": 9.643507426928607e-05, + "loss": 0.823, + "step": 3355 + }, + { + "epoch": 0.2273866793143167, + "grad_norm": 6.610576629638672, + "learning_rate": 9.643370525018825e-05, + "loss": 0.9251, + "step": 3356 + }, + { + "epoch": 0.22745443458228878, + "grad_norm": 8.365503311157227, + "learning_rate": 9.643233623109043e-05, + "loss": 1.0258, + "step": 3357 + }, + { + "epoch": 0.22752218985026085, + "grad_norm": 8.654791831970215, + "learning_rate": 9.643096721199261e-05, + "loss": 0.8781, + "step": 3358 + }, + { + "epoch": 0.22758994511823294, + "grad_norm": 6.612964153289795, + "learning_rate": 9.64295981928948e-05, + "loss": 0.7472, + "step": 3359 + }, + { + "epoch": 0.22765770038620503, + "grad_norm": 9.4874267578125, + "learning_rate": 9.642822917379698e-05, + "loss": 0.994, + "step": 3360 + }, + { + "epoch": 0.22772545565417712, + "grad_norm": 9.7284574508667, + "learning_rate": 9.642686015469916e-05, + "loss": 0.9293, + "step": 3361 + }, + { + "epoch": 0.2277932109221492, + "grad_norm": 8.787792205810547, + "learning_rate": 9.642549113560134e-05, + "loss": 1.1368, + "step": 3362 + }, + { + "epoch": 0.22786096619012128, + "grad_norm": 9.218424797058105, + "learning_rate": 9.642412211650354e-05, + "loss": 1.2437, + "step": 3363 + }, + { + "epoch": 0.22792872145809337, + "grad_norm": 7.9437127113342285, + "learning_rate": 9.642275309740572e-05, + "loss": 0.925, + "step": 3364 + }, + { + "epoch": 0.22799647672606546, + "grad_norm": 10.552806854248047, + "learning_rate": 9.64213840783079e-05, + "loss": 0.936, + "step": 3365 + }, + { + "epoch": 0.22806423199403753, + "grad_norm": 6.572626113891602, + "learning_rate": 9.642001505921008e-05, + "loss": 0.8968, + "step": 3366 + }, + { + "epoch": 0.22813198726200962, + "grad_norm": 8.689478874206543, + "learning_rate": 9.641864604011226e-05, + "loss": 0.8605, + "step": 3367 + }, + { + "epoch": 0.2281997425299817, + "grad_norm": 10.450199127197266, + "learning_rate": 9.641727702101445e-05, + "loss": 1.0871, + "step": 3368 + }, + { + "epoch": 0.2282674977979538, + "grad_norm": 9.260807991027832, + "learning_rate": 9.641590800191663e-05, + "loss": 0.8469, + "step": 3369 + }, + { + "epoch": 0.22833525306592586, + "grad_norm": 13.626687049865723, + "learning_rate": 9.641453898281881e-05, + "loss": 1.0615, + "step": 3370 + }, + { + "epoch": 0.22840300833389796, + "grad_norm": 7.419554710388184, + "learning_rate": 9.6413169963721e-05, + "loss": 0.9351, + "step": 3371 + }, + { + "epoch": 0.22847076360187005, + "grad_norm": 9.33206558227539, + "learning_rate": 9.641180094462319e-05, + "loss": 1.1826, + "step": 3372 + }, + { + "epoch": 0.22853851886984214, + "grad_norm": 7.148665904998779, + "learning_rate": 9.641043192552537e-05, + "loss": 1.1121, + "step": 3373 + }, + { + "epoch": 0.2286062741378142, + "grad_norm": 6.784035682678223, + "learning_rate": 9.640906290642755e-05, + "loss": 0.9649, + "step": 3374 + }, + { + "epoch": 0.2286740294057863, + "grad_norm": 7.357334613800049, + "learning_rate": 9.640769388732973e-05, + "loss": 0.9005, + "step": 3375 + }, + { + "epoch": 0.22874178467375839, + "grad_norm": 9.157288551330566, + "learning_rate": 9.640632486823191e-05, + "loss": 0.9585, + "step": 3376 + }, + { + "epoch": 0.22880953994173048, + "grad_norm": 9.068450927734375, + "learning_rate": 9.64049558491341e-05, + "loss": 0.9248, + "step": 3377 + }, + { + "epoch": 0.22887729520970254, + "grad_norm": 9.446860313415527, + "learning_rate": 9.640358683003628e-05, + "loss": 1.0016, + "step": 3378 + }, + { + "epoch": 0.22894505047767463, + "grad_norm": 7.63693904876709, + "learning_rate": 9.640221781093846e-05, + "loss": 1.0158, + "step": 3379 + }, + { + "epoch": 0.22901280574564672, + "grad_norm": 7.569469451904297, + "learning_rate": 9.640084879184064e-05, + "loss": 0.9149, + "step": 3380 + }, + { + "epoch": 0.22908056101361882, + "grad_norm": 8.837080955505371, + "learning_rate": 9.639947977274282e-05, + "loss": 0.9379, + "step": 3381 + }, + { + "epoch": 0.22914831628159088, + "grad_norm": 9.141901969909668, + "learning_rate": 9.639811075364502e-05, + "loss": 0.9121, + "step": 3382 + }, + { + "epoch": 0.22921607154956297, + "grad_norm": 7.68120813369751, + "learning_rate": 9.63967417345472e-05, + "loss": 1.0235, + "step": 3383 + }, + { + "epoch": 0.22928382681753506, + "grad_norm": 8.82229995727539, + "learning_rate": 9.639537271544938e-05, + "loss": 1.1573, + "step": 3384 + }, + { + "epoch": 0.22935158208550716, + "grad_norm": 9.490239143371582, + "learning_rate": 9.639400369635156e-05, + "loss": 0.8776, + "step": 3385 + }, + { + "epoch": 0.22941933735347925, + "grad_norm": 10.257567405700684, + "learning_rate": 9.639263467725375e-05, + "loss": 1.0696, + "step": 3386 + }, + { + "epoch": 0.2294870926214513, + "grad_norm": 7.793918609619141, + "learning_rate": 9.639126565815593e-05, + "loss": 1.0892, + "step": 3387 + }, + { + "epoch": 0.2295548478894234, + "grad_norm": 8.983718872070312, + "learning_rate": 9.638989663905811e-05, + "loss": 0.908, + "step": 3388 + }, + { + "epoch": 0.2296226031573955, + "grad_norm": 8.31851577758789, + "learning_rate": 9.63885276199603e-05, + "loss": 1.0168, + "step": 3389 + }, + { + "epoch": 0.22969035842536759, + "grad_norm": 9.343503952026367, + "learning_rate": 9.638715860086248e-05, + "loss": 1.1174, + "step": 3390 + }, + { + "epoch": 0.22975811369333965, + "grad_norm": 7.617075443267822, + "learning_rate": 9.638578958176467e-05, + "loss": 0.9076, + "step": 3391 + }, + { + "epoch": 0.22982586896131174, + "grad_norm": 8.102355003356934, + "learning_rate": 9.638442056266685e-05, + "loss": 1.0476, + "step": 3392 + }, + { + "epoch": 0.22989362422928383, + "grad_norm": 9.366684913635254, + "learning_rate": 9.638305154356903e-05, + "loss": 1.1892, + "step": 3393 + }, + { + "epoch": 0.22996137949725592, + "grad_norm": 8.204293251037598, + "learning_rate": 9.638168252447122e-05, + "loss": 0.7731, + "step": 3394 + }, + { + "epoch": 0.230029134765228, + "grad_norm": 9.68787670135498, + "learning_rate": 9.63803135053734e-05, + "loss": 1.1126, + "step": 3395 + }, + { + "epoch": 0.23009689003320008, + "grad_norm": 6.6209330558776855, + "learning_rate": 9.637894448627558e-05, + "loss": 0.9006, + "step": 3396 + }, + { + "epoch": 0.23016464530117217, + "grad_norm": 7.963149547576904, + "learning_rate": 9.637757546717778e-05, + "loss": 0.8275, + "step": 3397 + }, + { + "epoch": 0.23023240056914426, + "grad_norm": 8.74716567993164, + "learning_rate": 9.637620644807996e-05, + "loss": 1.0592, + "step": 3398 + }, + { + "epoch": 0.23030015583711633, + "grad_norm": 8.819621086120605, + "learning_rate": 9.637483742898214e-05, + "loss": 0.8799, + "step": 3399 + }, + { + "epoch": 0.23036791110508842, + "grad_norm": 8.538536071777344, + "learning_rate": 9.637346840988433e-05, + "loss": 1.1498, + "step": 3400 + }, + { + "epoch": 0.2304356663730605, + "grad_norm": 8.787203788757324, + "learning_rate": 9.637209939078651e-05, + "loss": 0.8149, + "step": 3401 + }, + { + "epoch": 0.2305034216410326, + "grad_norm": 8.85105037689209, + "learning_rate": 9.63707303716887e-05, + "loss": 0.8525, + "step": 3402 + }, + { + "epoch": 0.23057117690900467, + "grad_norm": 9.773297309875488, + "learning_rate": 9.636936135259087e-05, + "loss": 0.9448, + "step": 3403 + }, + { + "epoch": 0.23063893217697676, + "grad_norm": 7.804196357727051, + "learning_rate": 9.636799233349307e-05, + "loss": 0.9549, + "step": 3404 + }, + { + "epoch": 0.23070668744494885, + "grad_norm": 8.398775100708008, + "learning_rate": 9.636662331439525e-05, + "loss": 0.9111, + "step": 3405 + }, + { + "epoch": 0.23077444271292094, + "grad_norm": 8.960867881774902, + "learning_rate": 9.636525429529743e-05, + "loss": 1.1533, + "step": 3406 + }, + { + "epoch": 0.230842197980893, + "grad_norm": 7.513526439666748, + "learning_rate": 9.636388527619961e-05, + "loss": 1.113, + "step": 3407 + }, + { + "epoch": 0.2309099532488651, + "grad_norm": 8.891728401184082, + "learning_rate": 9.636251625710179e-05, + "loss": 0.8727, + "step": 3408 + }, + { + "epoch": 0.2309777085168372, + "grad_norm": 6.972326755523682, + "learning_rate": 9.636114723800398e-05, + "loss": 0.8085, + "step": 3409 + }, + { + "epoch": 0.23104546378480928, + "grad_norm": 9.195303916931152, + "learning_rate": 9.635977821890616e-05, + "loss": 1.0276, + "step": 3410 + }, + { + "epoch": 0.23111321905278134, + "grad_norm": 9.36483383178711, + "learning_rate": 9.635840919980834e-05, + "loss": 1.1369, + "step": 3411 + }, + { + "epoch": 0.23118097432075344, + "grad_norm": 7.282619953155518, + "learning_rate": 9.635704018071052e-05, + "loss": 0.9214, + "step": 3412 + }, + { + "epoch": 0.23124872958872553, + "grad_norm": 6.289726257324219, + "learning_rate": 9.63556711616127e-05, + "loss": 0.8556, + "step": 3413 + }, + { + "epoch": 0.23131648485669762, + "grad_norm": 8.646320343017578, + "learning_rate": 9.63543021425149e-05, + "loss": 1.0833, + "step": 3414 + }, + { + "epoch": 0.23138424012466968, + "grad_norm": 8.372559547424316, + "learning_rate": 9.635293312341708e-05, + "loss": 0.9944, + "step": 3415 + }, + { + "epoch": 0.23145199539264177, + "grad_norm": 8.002387046813965, + "learning_rate": 9.635156410431926e-05, + "loss": 1.015, + "step": 3416 + }, + { + "epoch": 0.23151975066061387, + "grad_norm": 8.4190092086792, + "learning_rate": 9.635019508522144e-05, + "loss": 0.9766, + "step": 3417 + }, + { + "epoch": 0.23158750592858596, + "grad_norm": 8.759613037109375, + "learning_rate": 9.634882606612363e-05, + "loss": 1.2348, + "step": 3418 + }, + { + "epoch": 0.23165526119655802, + "grad_norm": 8.439151763916016, + "learning_rate": 9.634745704702581e-05, + "loss": 0.8778, + "step": 3419 + }, + { + "epoch": 0.2317230164645301, + "grad_norm": 8.05675220489502, + "learning_rate": 9.6346088027928e-05, + "loss": 0.9525, + "step": 3420 + }, + { + "epoch": 0.2317907717325022, + "grad_norm": 9.196937561035156, + "learning_rate": 9.634471900883017e-05, + "loss": 1.0663, + "step": 3421 + }, + { + "epoch": 0.2318585270004743, + "grad_norm": 7.622048854827881, + "learning_rate": 9.634334998973235e-05, + "loss": 0.9522, + "step": 3422 + }, + { + "epoch": 0.23192628226844636, + "grad_norm": 10.235919952392578, + "learning_rate": 9.634198097063455e-05, + "loss": 1.1084, + "step": 3423 + }, + { + "epoch": 0.23199403753641845, + "grad_norm": 7.920719623565674, + "learning_rate": 9.634061195153673e-05, + "loss": 0.9443, + "step": 3424 + }, + { + "epoch": 0.23206179280439054, + "grad_norm": 7.440132141113281, + "learning_rate": 9.633924293243891e-05, + "loss": 1.0757, + "step": 3425 + }, + { + "epoch": 0.23212954807236263, + "grad_norm": 7.0108771324157715, + "learning_rate": 9.633787391334109e-05, + "loss": 0.8043, + "step": 3426 + }, + { + "epoch": 0.2321973033403347, + "grad_norm": 6.620635032653809, + "learning_rate": 9.633650489424328e-05, + "loss": 0.999, + "step": 3427 + }, + { + "epoch": 0.2322650586083068, + "grad_norm": 7.27518367767334, + "learning_rate": 9.633513587514546e-05, + "loss": 0.888, + "step": 3428 + }, + { + "epoch": 0.23233281387627888, + "grad_norm": 7.081945419311523, + "learning_rate": 9.633376685604764e-05, + "loss": 0.8837, + "step": 3429 + }, + { + "epoch": 0.23240056914425097, + "grad_norm": 7.225597858428955, + "learning_rate": 9.633239783694982e-05, + "loss": 0.7774, + "step": 3430 + }, + { + "epoch": 0.23246832441222304, + "grad_norm": 7.0490217208862305, + "learning_rate": 9.6331028817852e-05, + "loss": 1.0654, + "step": 3431 + }, + { + "epoch": 0.23253607968019513, + "grad_norm": 8.641639709472656, + "learning_rate": 9.63296597987542e-05, + "loss": 1.1071, + "step": 3432 + }, + { + "epoch": 0.23260383494816722, + "grad_norm": 7.442448139190674, + "learning_rate": 9.632829077965638e-05, + "loss": 0.9392, + "step": 3433 + }, + { + "epoch": 0.2326715902161393, + "grad_norm": 6.6959452629089355, + "learning_rate": 9.632692176055856e-05, + "loss": 0.8334, + "step": 3434 + }, + { + "epoch": 0.23273934548411138, + "grad_norm": 9.67419719696045, + "learning_rate": 9.632555274146074e-05, + "loss": 1.1948, + "step": 3435 + }, + { + "epoch": 0.23280710075208347, + "grad_norm": 7.384359836578369, + "learning_rate": 9.632418372236292e-05, + "loss": 0.7992, + "step": 3436 + }, + { + "epoch": 0.23287485602005556, + "grad_norm": 6.566294193267822, + "learning_rate": 9.632281470326511e-05, + "loss": 0.9393, + "step": 3437 + }, + { + "epoch": 0.23294261128802765, + "grad_norm": 7.842831134796143, + "learning_rate": 9.63214456841673e-05, + "loss": 1.1372, + "step": 3438 + }, + { + "epoch": 0.23301036655599974, + "grad_norm": 9.487961769104004, + "learning_rate": 9.632007666506947e-05, + "loss": 1.0381, + "step": 3439 + }, + { + "epoch": 0.2330781218239718, + "grad_norm": 8.06712532043457, + "learning_rate": 9.631870764597167e-05, + "loss": 1.4106, + "step": 3440 + }, + { + "epoch": 0.2331458770919439, + "grad_norm": 7.026076793670654, + "learning_rate": 9.631733862687385e-05, + "loss": 1.0662, + "step": 3441 + }, + { + "epoch": 0.233213632359916, + "grad_norm": 10.926567077636719, + "learning_rate": 9.631596960777603e-05, + "loss": 1.2932, + "step": 3442 + }, + { + "epoch": 0.23328138762788808, + "grad_norm": 8.380082130432129, + "learning_rate": 9.631460058867822e-05, + "loss": 1.0784, + "step": 3443 + }, + { + "epoch": 0.23334914289586015, + "grad_norm": 8.768819808959961, + "learning_rate": 9.63132315695804e-05, + "loss": 1.1501, + "step": 3444 + }, + { + "epoch": 0.23341689816383224, + "grad_norm": 8.467490196228027, + "learning_rate": 9.631186255048258e-05, + "loss": 0.7555, + "step": 3445 + }, + { + "epoch": 0.23348465343180433, + "grad_norm": 6.0409770011901855, + "learning_rate": 9.631049353138478e-05, + "loss": 0.9099, + "step": 3446 + }, + { + "epoch": 0.23355240869977642, + "grad_norm": 7.160163402557373, + "learning_rate": 9.630912451228696e-05, + "loss": 0.8386, + "step": 3447 + }, + { + "epoch": 0.23362016396774848, + "grad_norm": 9.102558135986328, + "learning_rate": 9.630775549318914e-05, + "loss": 0.9259, + "step": 3448 + }, + { + "epoch": 0.23368791923572058, + "grad_norm": 7.071728706359863, + "learning_rate": 9.630638647409132e-05, + "loss": 1.0426, + "step": 3449 + }, + { + "epoch": 0.23375567450369267, + "grad_norm": 8.358567237854004, + "learning_rate": 9.630501745499351e-05, + "loss": 0.9806, + "step": 3450 + }, + { + "epoch": 0.23382342977166476, + "grad_norm": 12.588993072509766, + "learning_rate": 9.630364843589569e-05, + "loss": 1.0796, + "step": 3451 + }, + { + "epoch": 0.23389118503963682, + "grad_norm": 8.665871620178223, + "learning_rate": 9.630227941679787e-05, + "loss": 1.0224, + "step": 3452 + }, + { + "epoch": 0.23395894030760891, + "grad_norm": 9.288796424865723, + "learning_rate": 9.630091039770005e-05, + "loss": 0.9329, + "step": 3453 + }, + { + "epoch": 0.234026695575581, + "grad_norm": 7.319186210632324, + "learning_rate": 9.629954137860223e-05, + "loss": 1.057, + "step": 3454 + }, + { + "epoch": 0.2340944508435531, + "grad_norm": 8.640669822692871, + "learning_rate": 9.629817235950443e-05, + "loss": 0.9756, + "step": 3455 + }, + { + "epoch": 0.23416220611152516, + "grad_norm": 10.73513126373291, + "learning_rate": 9.629680334040661e-05, + "loss": 0.958, + "step": 3456 + }, + { + "epoch": 0.23422996137949725, + "grad_norm": 7.40097713470459, + "learning_rate": 9.629543432130879e-05, + "loss": 0.9062, + "step": 3457 + }, + { + "epoch": 0.23429771664746935, + "grad_norm": 9.506468772888184, + "learning_rate": 9.629406530221097e-05, + "loss": 1.0083, + "step": 3458 + }, + { + "epoch": 0.23436547191544144, + "grad_norm": 7.291172504425049, + "learning_rate": 9.629269628311315e-05, + "loss": 0.9966, + "step": 3459 + }, + { + "epoch": 0.2344332271834135, + "grad_norm": 8.962395668029785, + "learning_rate": 9.629132726401534e-05, + "loss": 0.8504, + "step": 3460 + }, + { + "epoch": 0.2345009824513856, + "grad_norm": 10.29174518585205, + "learning_rate": 9.628995824491752e-05, + "loss": 1.0807, + "step": 3461 + }, + { + "epoch": 0.23456873771935768, + "grad_norm": 9.461160659790039, + "learning_rate": 9.62885892258197e-05, + "loss": 0.996, + "step": 3462 + }, + { + "epoch": 0.23463649298732978, + "grad_norm": 7.13394021987915, + "learning_rate": 9.628722020672188e-05, + "loss": 0.7163, + "step": 3463 + }, + { + "epoch": 0.23470424825530184, + "grad_norm": 9.379724502563477, + "learning_rate": 9.628585118762408e-05, + "loss": 0.9571, + "step": 3464 + }, + { + "epoch": 0.23477200352327393, + "grad_norm": 9.72339916229248, + "learning_rate": 9.628448216852626e-05, + "loss": 1.2559, + "step": 3465 + }, + { + "epoch": 0.23483975879124602, + "grad_norm": 7.353830337524414, + "learning_rate": 9.628311314942844e-05, + "loss": 1.0563, + "step": 3466 + }, + { + "epoch": 0.23490751405921811, + "grad_norm": 8.423667907714844, + "learning_rate": 9.628174413033062e-05, + "loss": 1.0799, + "step": 3467 + }, + { + "epoch": 0.23497526932719018, + "grad_norm": 6.426609516143799, + "learning_rate": 9.62803751112328e-05, + "loss": 0.9417, + "step": 3468 + }, + { + "epoch": 0.23504302459516227, + "grad_norm": 10.882522583007812, + "learning_rate": 9.627900609213499e-05, + "loss": 1.1498, + "step": 3469 + }, + { + "epoch": 0.23511077986313436, + "grad_norm": 7.768298625946045, + "learning_rate": 9.627763707303717e-05, + "loss": 0.7389, + "step": 3470 + }, + { + "epoch": 0.23517853513110645, + "grad_norm": 9.002137184143066, + "learning_rate": 9.627626805393935e-05, + "loss": 1.2338, + "step": 3471 + }, + { + "epoch": 0.23524629039907852, + "grad_norm": 8.671159744262695, + "learning_rate": 9.627489903484153e-05, + "loss": 1.1114, + "step": 3472 + }, + { + "epoch": 0.2353140456670506, + "grad_norm": 5.836034774780273, + "learning_rate": 9.627353001574373e-05, + "loss": 0.7075, + "step": 3473 + }, + { + "epoch": 0.2353818009350227, + "grad_norm": 8.145447731018066, + "learning_rate": 9.627216099664591e-05, + "loss": 0.9504, + "step": 3474 + }, + { + "epoch": 0.2354495562029948, + "grad_norm": 7.3333563804626465, + "learning_rate": 9.627079197754809e-05, + "loss": 0.8443, + "step": 3475 + }, + { + "epoch": 0.23551731147096686, + "grad_norm": 8.191438674926758, + "learning_rate": 9.626942295845027e-05, + "loss": 0.8163, + "step": 3476 + }, + { + "epoch": 0.23558506673893895, + "grad_norm": 10.042975425720215, + "learning_rate": 9.626805393935245e-05, + "loss": 1.0715, + "step": 3477 + }, + { + "epoch": 0.23565282200691104, + "grad_norm": 7.854464054107666, + "learning_rate": 9.626668492025464e-05, + "loss": 0.9263, + "step": 3478 + }, + { + "epoch": 0.23572057727488313, + "grad_norm": 9.286520004272461, + "learning_rate": 9.626531590115682e-05, + "loss": 1.0172, + "step": 3479 + }, + { + "epoch": 0.2357883325428552, + "grad_norm": 7.321778297424316, + "learning_rate": 9.6263946882059e-05, + "loss": 1.0661, + "step": 3480 + }, + { + "epoch": 0.2358560878108273, + "grad_norm": 7.432394981384277, + "learning_rate": 9.626257786296118e-05, + "loss": 1.1682, + "step": 3481 + }, + { + "epoch": 0.23592384307879938, + "grad_norm": 10.230997085571289, + "learning_rate": 9.626120884386338e-05, + "loss": 1.0498, + "step": 3482 + }, + { + "epoch": 0.23599159834677147, + "grad_norm": 8.530696868896484, + "learning_rate": 9.625983982476556e-05, + "loss": 1.0598, + "step": 3483 + }, + { + "epoch": 0.23605935361474353, + "grad_norm": 7.35584020614624, + "learning_rate": 9.625847080566774e-05, + "loss": 1.0234, + "step": 3484 + }, + { + "epoch": 0.23612710888271563, + "grad_norm": 8.030210494995117, + "learning_rate": 9.625710178656992e-05, + "loss": 1.0241, + "step": 3485 + }, + { + "epoch": 0.23619486415068772, + "grad_norm": 7.508440971374512, + "learning_rate": 9.625573276747211e-05, + "loss": 0.9466, + "step": 3486 + }, + { + "epoch": 0.2362626194186598, + "grad_norm": 7.4618401527404785, + "learning_rate": 9.62543637483743e-05, + "loss": 0.88, + "step": 3487 + }, + { + "epoch": 0.23633037468663187, + "grad_norm": 7.738553524017334, + "learning_rate": 9.625299472927647e-05, + "loss": 1.0146, + "step": 3488 + }, + { + "epoch": 0.23639812995460396, + "grad_norm": 6.782055854797363, + "learning_rate": 9.625162571017867e-05, + "loss": 0.8196, + "step": 3489 + }, + { + "epoch": 0.23646588522257606, + "grad_norm": 7.495883464813232, + "learning_rate": 9.625025669108085e-05, + "loss": 1.0025, + "step": 3490 + }, + { + "epoch": 0.23653364049054815, + "grad_norm": 7.224970817565918, + "learning_rate": 9.624888767198303e-05, + "loss": 1.0069, + "step": 3491 + }, + { + "epoch": 0.23660139575852024, + "grad_norm": 8.686829566955566, + "learning_rate": 9.624751865288522e-05, + "loss": 0.8927, + "step": 3492 + }, + { + "epoch": 0.2366691510264923, + "grad_norm": 10.551370620727539, + "learning_rate": 9.62461496337874e-05, + "loss": 1.0973, + "step": 3493 + }, + { + "epoch": 0.2367369062944644, + "grad_norm": 8.38442611694336, + "learning_rate": 9.624478061468958e-05, + "loss": 0.8145, + "step": 3494 + }, + { + "epoch": 0.23680466156243649, + "grad_norm": 7.099575996398926, + "learning_rate": 9.624341159559176e-05, + "loss": 0.8304, + "step": 3495 + }, + { + "epoch": 0.23687241683040858, + "grad_norm": 7.905203342437744, + "learning_rate": 9.624204257649396e-05, + "loss": 1.1708, + "step": 3496 + }, + { + "epoch": 0.23694017209838064, + "grad_norm": 8.08218765258789, + "learning_rate": 9.624067355739614e-05, + "loss": 0.8367, + "step": 3497 + }, + { + "epoch": 0.23700792736635273, + "grad_norm": 7.766988277435303, + "learning_rate": 9.623930453829832e-05, + "loss": 1.1126, + "step": 3498 + }, + { + "epoch": 0.23707568263432482, + "grad_norm": 8.3229398727417, + "learning_rate": 9.62379355192005e-05, + "loss": 0.9798, + "step": 3499 + }, + { + "epoch": 0.23714343790229692, + "grad_norm": 7.113401889801025, + "learning_rate": 9.623656650010268e-05, + "loss": 0.7711, + "step": 3500 + }, + { + "epoch": 0.23721119317026898, + "grad_norm": 7.286252975463867, + "learning_rate": 9.623519748100487e-05, + "loss": 0.8018, + "step": 3501 + }, + { + "epoch": 0.23727894843824107, + "grad_norm": 8.724101066589355, + "learning_rate": 9.623382846190705e-05, + "loss": 0.9077, + "step": 3502 + }, + { + "epoch": 0.23734670370621316, + "grad_norm": 9.201606750488281, + "learning_rate": 9.623245944280923e-05, + "loss": 1.2277, + "step": 3503 + }, + { + "epoch": 0.23741445897418526, + "grad_norm": 6.827572822570801, + "learning_rate": 9.623109042371141e-05, + "loss": 0.9262, + "step": 3504 + }, + { + "epoch": 0.23748221424215732, + "grad_norm": 7.41288423538208, + "learning_rate": 9.622972140461361e-05, + "loss": 0.762, + "step": 3505 + }, + { + "epoch": 0.2375499695101294, + "grad_norm": 9.649503707885742, + "learning_rate": 9.622835238551579e-05, + "loss": 1.0371, + "step": 3506 + }, + { + "epoch": 0.2376177247781015, + "grad_norm": 9.630754470825195, + "learning_rate": 9.622698336641797e-05, + "loss": 0.9994, + "step": 3507 + }, + { + "epoch": 0.2376854800460736, + "grad_norm": 8.949817657470703, + "learning_rate": 9.622561434732015e-05, + "loss": 0.8555, + "step": 3508 + }, + { + "epoch": 0.23775323531404566, + "grad_norm": 8.096979141235352, + "learning_rate": 9.622424532822233e-05, + "loss": 0.9939, + "step": 3509 + }, + { + "epoch": 0.23782099058201775, + "grad_norm": 8.714512825012207, + "learning_rate": 9.622287630912452e-05, + "loss": 0.9533, + "step": 3510 + }, + { + "epoch": 0.23788874584998984, + "grad_norm": 8.663968086242676, + "learning_rate": 9.62215072900267e-05, + "loss": 0.9974, + "step": 3511 + }, + { + "epoch": 0.23795650111796193, + "grad_norm": 8.704265594482422, + "learning_rate": 9.622013827092888e-05, + "loss": 1.2524, + "step": 3512 + }, + { + "epoch": 0.238024256385934, + "grad_norm": 6.585339069366455, + "learning_rate": 9.621876925183106e-05, + "loss": 0.9021, + "step": 3513 + }, + { + "epoch": 0.2380920116539061, + "grad_norm": 7.992496013641357, + "learning_rate": 9.621740023273324e-05, + "loss": 0.9345, + "step": 3514 + }, + { + "epoch": 0.23815976692187818, + "grad_norm": 7.076025009155273, + "learning_rate": 9.621603121363544e-05, + "loss": 1.0487, + "step": 3515 + }, + { + "epoch": 0.23822752218985027, + "grad_norm": 8.833822250366211, + "learning_rate": 9.621466219453762e-05, + "loss": 0.9612, + "step": 3516 + }, + { + "epoch": 0.23829527745782234, + "grad_norm": 7.8553385734558105, + "learning_rate": 9.62132931754398e-05, + "loss": 1.0452, + "step": 3517 + }, + { + "epoch": 0.23836303272579443, + "grad_norm": 8.44243049621582, + "learning_rate": 9.621192415634198e-05, + "loss": 0.6874, + "step": 3518 + }, + { + "epoch": 0.23843078799376652, + "grad_norm": 8.1638822555542, + "learning_rate": 9.621055513724417e-05, + "loss": 0.9239, + "step": 3519 + }, + { + "epoch": 0.2384985432617386, + "grad_norm": 8.315443992614746, + "learning_rate": 9.620918611814635e-05, + "loss": 1.1256, + "step": 3520 + }, + { + "epoch": 0.23856629852971067, + "grad_norm": 10.451863288879395, + "learning_rate": 9.620781709904853e-05, + "loss": 1.1453, + "step": 3521 + }, + { + "epoch": 0.23863405379768277, + "grad_norm": 9.117147445678711, + "learning_rate": 9.620644807995071e-05, + "loss": 1.0751, + "step": 3522 + }, + { + "epoch": 0.23870180906565486, + "grad_norm": 7.746578216552734, + "learning_rate": 9.62050790608529e-05, + "loss": 0.867, + "step": 3523 + }, + { + "epoch": 0.23876956433362695, + "grad_norm": 8.366421699523926, + "learning_rate": 9.620371004175509e-05, + "loss": 1.0863, + "step": 3524 + }, + { + "epoch": 0.238837319601599, + "grad_norm": 9.951937675476074, + "learning_rate": 9.620234102265727e-05, + "loss": 1.0431, + "step": 3525 + }, + { + "epoch": 0.2389050748695711, + "grad_norm": 9.077424049377441, + "learning_rate": 9.620097200355945e-05, + "loss": 1.0433, + "step": 3526 + }, + { + "epoch": 0.2389728301375432, + "grad_norm": 6.919139385223389, + "learning_rate": 9.619960298446163e-05, + "loss": 0.8655, + "step": 3527 + }, + { + "epoch": 0.2390405854055153, + "grad_norm": 7.729475975036621, + "learning_rate": 9.619823396536382e-05, + "loss": 1.0943, + "step": 3528 + }, + { + "epoch": 0.23910834067348735, + "grad_norm": 9.782391548156738, + "learning_rate": 9.6196864946266e-05, + "loss": 1.1362, + "step": 3529 + }, + { + "epoch": 0.23917609594145944, + "grad_norm": 8.685064315795898, + "learning_rate": 9.619549592716818e-05, + "loss": 0.9885, + "step": 3530 + }, + { + "epoch": 0.23924385120943154, + "grad_norm": 6.963294982910156, + "learning_rate": 9.619412690807036e-05, + "loss": 0.9061, + "step": 3531 + }, + { + "epoch": 0.23931160647740363, + "grad_norm": 8.584861755371094, + "learning_rate": 9.619275788897256e-05, + "loss": 0.8527, + "step": 3532 + }, + { + "epoch": 0.2393793617453757, + "grad_norm": 9.253739356994629, + "learning_rate": 9.619138886987474e-05, + "loss": 1.0643, + "step": 3533 + }, + { + "epoch": 0.23944711701334778, + "grad_norm": 7.659543037414551, + "learning_rate": 9.619001985077692e-05, + "loss": 0.9553, + "step": 3534 + }, + { + "epoch": 0.23951487228131987, + "grad_norm": 8.769670486450195, + "learning_rate": 9.618865083167911e-05, + "loss": 1.0576, + "step": 3535 + }, + { + "epoch": 0.23958262754929197, + "grad_norm": 7.416141033172607, + "learning_rate": 9.618728181258129e-05, + "loss": 1.0064, + "step": 3536 + }, + { + "epoch": 0.23965038281726403, + "grad_norm": 7.91802978515625, + "learning_rate": 9.618591279348347e-05, + "loss": 1.0095, + "step": 3537 + }, + { + "epoch": 0.23971813808523612, + "grad_norm": 6.665622234344482, + "learning_rate": 9.618454377438567e-05, + "loss": 0.8099, + "step": 3538 + }, + { + "epoch": 0.2397858933532082, + "grad_norm": 7.3240203857421875, + "learning_rate": 9.618317475528785e-05, + "loss": 0.9496, + "step": 3539 + }, + { + "epoch": 0.2398536486211803, + "grad_norm": 7.268299102783203, + "learning_rate": 9.618180573619003e-05, + "loss": 1.0504, + "step": 3540 + }, + { + "epoch": 0.23992140388915237, + "grad_norm": 8.710535049438477, + "learning_rate": 9.618043671709221e-05, + "loss": 0.9228, + "step": 3541 + }, + { + "epoch": 0.23998915915712446, + "grad_norm": 8.5223970413208, + "learning_rate": 9.61790676979944e-05, + "loss": 1.0383, + "step": 3542 + }, + { + "epoch": 0.24005691442509655, + "grad_norm": 7.635293960571289, + "learning_rate": 9.617769867889658e-05, + "loss": 0.99, + "step": 3543 + }, + { + "epoch": 0.24012466969306864, + "grad_norm": 9.41180419921875, + "learning_rate": 9.617632965979876e-05, + "loss": 1.1214, + "step": 3544 + }, + { + "epoch": 0.24019242496104073, + "grad_norm": 9.31615924835205, + "learning_rate": 9.617496064070094e-05, + "loss": 0.9369, + "step": 3545 + }, + { + "epoch": 0.2402601802290128, + "grad_norm": 9.427112579345703, + "learning_rate": 9.617359162160312e-05, + "loss": 1.1829, + "step": 3546 + }, + { + "epoch": 0.2403279354969849, + "grad_norm": 7.825446605682373, + "learning_rate": 9.617222260250532e-05, + "loss": 0.9459, + "step": 3547 + }, + { + "epoch": 0.24039569076495698, + "grad_norm": 7.8191633224487305, + "learning_rate": 9.61708535834075e-05, + "loss": 1.1144, + "step": 3548 + }, + { + "epoch": 0.24046344603292907, + "grad_norm": 8.224778175354004, + "learning_rate": 9.616948456430968e-05, + "loss": 0.8544, + "step": 3549 + }, + { + "epoch": 0.24053120130090114, + "grad_norm": 9.618694305419922, + "learning_rate": 9.616811554521186e-05, + "loss": 1.2254, + "step": 3550 + }, + { + "epoch": 0.24059895656887323, + "grad_norm": 7.786314964294434, + "learning_rate": 9.616674652611405e-05, + "loss": 0.8867, + "step": 3551 + }, + { + "epoch": 0.24066671183684532, + "grad_norm": 9.397835731506348, + "learning_rate": 9.616537750701623e-05, + "loss": 1.0505, + "step": 3552 + }, + { + "epoch": 0.2407344671048174, + "grad_norm": 7.701049327850342, + "learning_rate": 9.616400848791841e-05, + "loss": 0.897, + "step": 3553 + }, + { + "epoch": 0.24080222237278948, + "grad_norm": 7.573019504547119, + "learning_rate": 9.616263946882059e-05, + "loss": 0.9972, + "step": 3554 + }, + { + "epoch": 0.24086997764076157, + "grad_norm": 8.337100982666016, + "learning_rate": 9.616127044972277e-05, + "loss": 1.0835, + "step": 3555 + }, + { + "epoch": 0.24093773290873366, + "grad_norm": 8.702056884765625, + "learning_rate": 9.615990143062497e-05, + "loss": 0.7688, + "step": 3556 + }, + { + "epoch": 0.24100548817670575, + "grad_norm": 8.482346534729004, + "learning_rate": 9.615853241152715e-05, + "loss": 1.1972, + "step": 3557 + }, + { + "epoch": 0.24107324344467781, + "grad_norm": 7.193674087524414, + "learning_rate": 9.615716339242933e-05, + "loss": 1.0389, + "step": 3558 + }, + { + "epoch": 0.2411409987126499, + "grad_norm": 8.803317070007324, + "learning_rate": 9.615579437333151e-05, + "loss": 1.0429, + "step": 3559 + }, + { + "epoch": 0.241208753980622, + "grad_norm": 9.055732727050781, + "learning_rate": 9.61544253542337e-05, + "loss": 1.1161, + "step": 3560 + }, + { + "epoch": 0.2412765092485941, + "grad_norm": 10.80001449584961, + "learning_rate": 9.615305633513588e-05, + "loss": 0.9628, + "step": 3561 + }, + { + "epoch": 0.24134426451656615, + "grad_norm": 7.707313060760498, + "learning_rate": 9.615168731603806e-05, + "loss": 0.8806, + "step": 3562 + }, + { + "epoch": 0.24141201978453825, + "grad_norm": 8.823626518249512, + "learning_rate": 9.615031829694024e-05, + "loss": 0.8771, + "step": 3563 + }, + { + "epoch": 0.24147977505251034, + "grad_norm": 7.984725475311279, + "learning_rate": 9.614894927784242e-05, + "loss": 0.9456, + "step": 3564 + }, + { + "epoch": 0.24154753032048243, + "grad_norm": 7.335816860198975, + "learning_rate": 9.614758025874462e-05, + "loss": 1.0323, + "step": 3565 + }, + { + "epoch": 0.2416152855884545, + "grad_norm": 5.959085941314697, + "learning_rate": 9.61462112396468e-05, + "loss": 0.8049, + "step": 3566 + }, + { + "epoch": 0.24168304085642658, + "grad_norm": 8.775632858276367, + "learning_rate": 9.614484222054898e-05, + "loss": 0.8875, + "step": 3567 + }, + { + "epoch": 0.24175079612439868, + "grad_norm": 6.589362621307373, + "learning_rate": 9.614347320145116e-05, + "loss": 1.0551, + "step": 3568 + }, + { + "epoch": 0.24181855139237077, + "grad_norm": 7.49434232711792, + "learning_rate": 9.614210418235334e-05, + "loss": 0.7662, + "step": 3569 + }, + { + "epoch": 0.24188630666034283, + "grad_norm": 7.759862899780273, + "learning_rate": 9.614073516325553e-05, + "loss": 0.7364, + "step": 3570 + }, + { + "epoch": 0.24195406192831492, + "grad_norm": 7.9510273933410645, + "learning_rate": 9.613936614415771e-05, + "loss": 0.9277, + "step": 3571 + }, + { + "epoch": 0.24202181719628701, + "grad_norm": 8.308568954467773, + "learning_rate": 9.61379971250599e-05, + "loss": 1.1689, + "step": 3572 + }, + { + "epoch": 0.2420895724642591, + "grad_norm": 9.44938850402832, + "learning_rate": 9.613662810596207e-05, + "loss": 1.0937, + "step": 3573 + }, + { + "epoch": 0.24215732773223117, + "grad_norm": 11.66707706451416, + "learning_rate": 9.613525908686427e-05, + "loss": 0.77, + "step": 3574 + }, + { + "epoch": 0.24222508300020326, + "grad_norm": 9.25683879852295, + "learning_rate": 9.613389006776645e-05, + "loss": 0.9661, + "step": 3575 + }, + { + "epoch": 0.24229283826817535, + "grad_norm": 7.289797782897949, + "learning_rate": 9.613252104866863e-05, + "loss": 0.8597, + "step": 3576 + }, + { + "epoch": 0.24236059353614745, + "grad_norm": 6.669293403625488, + "learning_rate": 9.613115202957081e-05, + "loss": 0.9134, + "step": 3577 + }, + { + "epoch": 0.2424283488041195, + "grad_norm": 8.785436630249023, + "learning_rate": 9.612978301047299e-05, + "loss": 0.7546, + "step": 3578 + }, + { + "epoch": 0.2424961040720916, + "grad_norm": 7.386310577392578, + "learning_rate": 9.612841399137518e-05, + "loss": 1.0635, + "step": 3579 + }, + { + "epoch": 0.2425638593400637, + "grad_norm": 8.23388957977295, + "learning_rate": 9.612704497227736e-05, + "loss": 0.8598, + "step": 3580 + }, + { + "epoch": 0.24263161460803578, + "grad_norm": 8.498323440551758, + "learning_rate": 9.612567595317954e-05, + "loss": 1.083, + "step": 3581 + }, + { + "epoch": 0.24269936987600785, + "grad_norm": 7.849715709686279, + "learning_rate": 9.612430693408174e-05, + "loss": 0.7673, + "step": 3582 + }, + { + "epoch": 0.24276712514397994, + "grad_norm": 8.113242149353027, + "learning_rate": 9.612293791498392e-05, + "loss": 1.1003, + "step": 3583 + }, + { + "epoch": 0.24283488041195203, + "grad_norm": 6.983048915863037, + "learning_rate": 9.61215688958861e-05, + "loss": 0.8668, + "step": 3584 + }, + { + "epoch": 0.24290263567992412, + "grad_norm": 8.389126777648926, + "learning_rate": 9.612019987678829e-05, + "loss": 1.0008, + "step": 3585 + }, + { + "epoch": 0.2429703909478962, + "grad_norm": 7.593414783477783, + "learning_rate": 9.611883085769047e-05, + "loss": 0.9829, + "step": 3586 + }, + { + "epoch": 0.24303814621586828, + "grad_norm": 7.836172103881836, + "learning_rate": 9.611746183859265e-05, + "loss": 0.8128, + "step": 3587 + }, + { + "epoch": 0.24310590148384037, + "grad_norm": 8.981040954589844, + "learning_rate": 9.611609281949485e-05, + "loss": 1.1249, + "step": 3588 + }, + { + "epoch": 0.24317365675181246, + "grad_norm": 6.724935054779053, + "learning_rate": 9.611472380039703e-05, + "loss": 0.7699, + "step": 3589 + }, + { + "epoch": 0.24324141201978453, + "grad_norm": 9.456436157226562, + "learning_rate": 9.611335478129921e-05, + "loss": 1.0085, + "step": 3590 + }, + { + "epoch": 0.24330916728775662, + "grad_norm": 11.044548988342285, + "learning_rate": 9.611198576220139e-05, + "loss": 1.0671, + "step": 3591 + }, + { + "epoch": 0.2433769225557287, + "grad_norm": 10.590521812438965, + "learning_rate": 9.611061674310358e-05, + "loss": 1.0051, + "step": 3592 + }, + { + "epoch": 0.2434446778237008, + "grad_norm": 8.595380783081055, + "learning_rate": 9.610924772400576e-05, + "loss": 0.9655, + "step": 3593 + }, + { + "epoch": 0.24351243309167286, + "grad_norm": 7.794788837432861, + "learning_rate": 9.610787870490794e-05, + "loss": 0.9551, + "step": 3594 + }, + { + "epoch": 0.24358018835964496, + "grad_norm": 7.609074115753174, + "learning_rate": 9.610650968581012e-05, + "loss": 0.9856, + "step": 3595 + }, + { + "epoch": 0.24364794362761705, + "grad_norm": 6.909607410430908, + "learning_rate": 9.61051406667123e-05, + "loss": 1.0195, + "step": 3596 + }, + { + "epoch": 0.24371569889558914, + "grad_norm": 7.958381652832031, + "learning_rate": 9.61037716476145e-05, + "loss": 0.8263, + "step": 3597 + }, + { + "epoch": 0.24378345416356123, + "grad_norm": 7.38173246383667, + "learning_rate": 9.610240262851668e-05, + "loss": 0.8551, + "step": 3598 + }, + { + "epoch": 0.2438512094315333, + "grad_norm": 7.1448822021484375, + "learning_rate": 9.610103360941886e-05, + "loss": 0.8189, + "step": 3599 + }, + { + "epoch": 0.2439189646995054, + "grad_norm": 7.44658088684082, + "learning_rate": 9.609966459032104e-05, + "loss": 0.8056, + "step": 3600 + }, + { + "epoch": 0.24398671996747748, + "grad_norm": 7.480542182922363, + "learning_rate": 9.609829557122322e-05, + "loss": 0.8591, + "step": 3601 + }, + { + "epoch": 0.24405447523544957, + "grad_norm": 9.895995140075684, + "learning_rate": 9.609692655212541e-05, + "loss": 1.1256, + "step": 3602 + }, + { + "epoch": 0.24412223050342163, + "grad_norm": 7.009078025817871, + "learning_rate": 9.609555753302759e-05, + "loss": 0.9045, + "step": 3603 + }, + { + "epoch": 0.24418998577139372, + "grad_norm": 8.714953422546387, + "learning_rate": 9.609418851392977e-05, + "loss": 1.1207, + "step": 3604 + }, + { + "epoch": 0.24425774103936582, + "grad_norm": 7.239734172821045, + "learning_rate": 9.609281949483195e-05, + "loss": 0.7811, + "step": 3605 + }, + { + "epoch": 0.2443254963073379, + "grad_norm": 10.486507415771484, + "learning_rate": 9.609145047573415e-05, + "loss": 1.2213, + "step": 3606 + }, + { + "epoch": 0.24439325157530997, + "grad_norm": 12.031790733337402, + "learning_rate": 9.609008145663633e-05, + "loss": 1.0316, + "step": 3607 + }, + { + "epoch": 0.24446100684328206, + "grad_norm": 7.607183456420898, + "learning_rate": 9.608871243753851e-05, + "loss": 0.7754, + "step": 3608 + }, + { + "epoch": 0.24452876211125416, + "grad_norm": 9.313577651977539, + "learning_rate": 9.608734341844069e-05, + "loss": 0.9907, + "step": 3609 + }, + { + "epoch": 0.24459651737922625, + "grad_norm": 7.672274589538574, + "learning_rate": 9.608597439934287e-05, + "loss": 0.8428, + "step": 3610 + }, + { + "epoch": 0.2446642726471983, + "grad_norm": 8.259462356567383, + "learning_rate": 9.608460538024506e-05, + "loss": 0.8863, + "step": 3611 + }, + { + "epoch": 0.2447320279151704, + "grad_norm": 8.50256061553955, + "learning_rate": 9.608323636114724e-05, + "loss": 0.7771, + "step": 3612 + }, + { + "epoch": 0.2447997831831425, + "grad_norm": 6.8893818855285645, + "learning_rate": 9.608186734204942e-05, + "loss": 0.9791, + "step": 3613 + }, + { + "epoch": 0.24486753845111459, + "grad_norm": 8.954825401306152, + "learning_rate": 9.60804983229516e-05, + "loss": 1.0712, + "step": 3614 + }, + { + "epoch": 0.24493529371908665, + "grad_norm": 8.83995532989502, + "learning_rate": 9.60791293038538e-05, + "loss": 1.1982, + "step": 3615 + }, + { + "epoch": 0.24500304898705874, + "grad_norm": 8.96689224243164, + "learning_rate": 9.607776028475598e-05, + "loss": 0.9816, + "step": 3616 + }, + { + "epoch": 0.24507080425503083, + "grad_norm": 15.17086124420166, + "learning_rate": 9.607639126565816e-05, + "loss": 1.1416, + "step": 3617 + }, + { + "epoch": 0.24513855952300292, + "grad_norm": 9.527314186096191, + "learning_rate": 9.607502224656034e-05, + "loss": 1.1699, + "step": 3618 + }, + { + "epoch": 0.245206314790975, + "grad_norm": 7.378002166748047, + "learning_rate": 9.607365322746252e-05, + "loss": 1.1303, + "step": 3619 + }, + { + "epoch": 0.24527407005894708, + "grad_norm": 7.204291343688965, + "learning_rate": 9.607228420836471e-05, + "loss": 0.8078, + "step": 3620 + }, + { + "epoch": 0.24534182532691917, + "grad_norm": 8.181205749511719, + "learning_rate": 9.607091518926689e-05, + "loss": 0.9221, + "step": 3621 + }, + { + "epoch": 0.24540958059489126, + "grad_norm": 8.479545593261719, + "learning_rate": 9.606954617016907e-05, + "loss": 0.8133, + "step": 3622 + }, + { + "epoch": 0.24547733586286333, + "grad_norm": 7.591360092163086, + "learning_rate": 9.606817715107125e-05, + "loss": 1.0632, + "step": 3623 + }, + { + "epoch": 0.24554509113083542, + "grad_norm": 8.558969497680664, + "learning_rate": 9.606680813197343e-05, + "loss": 1.0755, + "step": 3624 + }, + { + "epoch": 0.2456128463988075, + "grad_norm": 8.02037525177002, + "learning_rate": 9.606543911287563e-05, + "loss": 1.0164, + "step": 3625 + }, + { + "epoch": 0.2456806016667796, + "grad_norm": 7.49207878112793, + "learning_rate": 9.606407009377781e-05, + "loss": 1.1325, + "step": 3626 + }, + { + "epoch": 0.24574835693475167, + "grad_norm": 7.376079559326172, + "learning_rate": 9.606270107467999e-05, + "loss": 0.7917, + "step": 3627 + }, + { + "epoch": 0.24581611220272376, + "grad_norm": 7.331247329711914, + "learning_rate": 9.606133205558218e-05, + "loss": 1.2186, + "step": 3628 + }, + { + "epoch": 0.24588386747069585, + "grad_norm": 7.393257141113281, + "learning_rate": 9.605996303648436e-05, + "loss": 0.7155, + "step": 3629 + }, + { + "epoch": 0.24595162273866794, + "grad_norm": 9.15472412109375, + "learning_rate": 9.605859401738654e-05, + "loss": 0.9562, + "step": 3630 + }, + { + "epoch": 0.24601937800664, + "grad_norm": 6.846646785736084, + "learning_rate": 9.605722499828874e-05, + "loss": 1.109, + "step": 3631 + }, + { + "epoch": 0.2460871332746121, + "grad_norm": 8.57854175567627, + "learning_rate": 9.605585597919092e-05, + "loss": 0.8706, + "step": 3632 + }, + { + "epoch": 0.2461548885425842, + "grad_norm": 7.089768886566162, + "learning_rate": 9.60544869600931e-05, + "loss": 0.8776, + "step": 3633 + }, + { + "epoch": 0.24622264381055628, + "grad_norm": 7.549044132232666, + "learning_rate": 9.605311794099529e-05, + "loss": 0.9956, + "step": 3634 + }, + { + "epoch": 0.24629039907852834, + "grad_norm": 6.839412689208984, + "learning_rate": 9.605174892189747e-05, + "loss": 0.8722, + "step": 3635 + }, + { + "epoch": 0.24635815434650044, + "grad_norm": 7.378058910369873, + "learning_rate": 9.605037990279965e-05, + "loss": 0.8663, + "step": 3636 + }, + { + "epoch": 0.24642590961447253, + "grad_norm": 7.893070220947266, + "learning_rate": 9.604901088370183e-05, + "loss": 0.9213, + "step": 3637 + }, + { + "epoch": 0.24649366488244462, + "grad_norm": 7.3345232009887695, + "learning_rate": 9.604764186460403e-05, + "loss": 0.9827, + "step": 3638 + }, + { + "epoch": 0.24656142015041668, + "grad_norm": 6.120781421661377, + "learning_rate": 9.60462728455062e-05, + "loss": 0.6371, + "step": 3639 + }, + { + "epoch": 0.24662917541838877, + "grad_norm": 8.695615768432617, + "learning_rate": 9.604490382640839e-05, + "loss": 0.9769, + "step": 3640 + }, + { + "epoch": 0.24669693068636087, + "grad_norm": 8.469325065612793, + "learning_rate": 9.604353480731057e-05, + "loss": 0.9267, + "step": 3641 + }, + { + "epoch": 0.24676468595433296, + "grad_norm": 9.191173553466797, + "learning_rate": 9.604216578821275e-05, + "loss": 1.1818, + "step": 3642 + }, + { + "epoch": 0.24683244122230502, + "grad_norm": 9.337483406066895, + "learning_rate": 9.604079676911494e-05, + "loss": 0.9904, + "step": 3643 + }, + { + "epoch": 0.2469001964902771, + "grad_norm": 7.597773551940918, + "learning_rate": 9.603942775001712e-05, + "loss": 0.8928, + "step": 3644 + }, + { + "epoch": 0.2469679517582492, + "grad_norm": 8.155903816223145, + "learning_rate": 9.60380587309193e-05, + "loss": 0.9511, + "step": 3645 + }, + { + "epoch": 0.2470357070262213, + "grad_norm": 7.695154666900635, + "learning_rate": 9.603668971182148e-05, + "loss": 1.0724, + "step": 3646 + }, + { + "epoch": 0.24710346229419336, + "grad_norm": 7.492908000946045, + "learning_rate": 9.603532069272366e-05, + "loss": 0.9133, + "step": 3647 + }, + { + "epoch": 0.24717121756216545, + "grad_norm": 8.190613746643066, + "learning_rate": 9.603395167362586e-05, + "loss": 0.8972, + "step": 3648 + }, + { + "epoch": 0.24723897283013754, + "grad_norm": 7.9199347496032715, + "learning_rate": 9.603258265452804e-05, + "loss": 1.0501, + "step": 3649 + }, + { + "epoch": 0.24730672809810963, + "grad_norm": 8.386896133422852, + "learning_rate": 9.603121363543022e-05, + "loss": 0.998, + "step": 3650 + }, + { + "epoch": 0.24737448336608173, + "grad_norm": 6.536781311035156, + "learning_rate": 9.60298446163324e-05, + "loss": 0.8379, + "step": 3651 + }, + { + "epoch": 0.2474422386340538, + "grad_norm": 6.918766021728516, + "learning_rate": 9.602847559723459e-05, + "loss": 0.9397, + "step": 3652 + }, + { + "epoch": 0.24750999390202588, + "grad_norm": 7.00775146484375, + "learning_rate": 9.602710657813677e-05, + "loss": 1.0079, + "step": 3653 + }, + { + "epoch": 0.24757774916999797, + "grad_norm": 7.693192958831787, + "learning_rate": 9.602573755903895e-05, + "loss": 1.0657, + "step": 3654 + }, + { + "epoch": 0.24764550443797007, + "grad_norm": 10.667771339416504, + "learning_rate": 9.602436853994113e-05, + "loss": 1.0657, + "step": 3655 + }, + { + "epoch": 0.24771325970594213, + "grad_norm": 8.622758865356445, + "learning_rate": 9.602299952084331e-05, + "loss": 0.9792, + "step": 3656 + }, + { + "epoch": 0.24778101497391422, + "grad_norm": 8.834444046020508, + "learning_rate": 9.602163050174551e-05, + "loss": 1.1014, + "step": 3657 + }, + { + "epoch": 0.2478487702418863, + "grad_norm": 7.717538356781006, + "learning_rate": 9.602026148264769e-05, + "loss": 0.8781, + "step": 3658 + }, + { + "epoch": 0.2479165255098584, + "grad_norm": 7.797954559326172, + "learning_rate": 9.601889246354987e-05, + "loss": 0.9772, + "step": 3659 + }, + { + "epoch": 0.24798428077783047, + "grad_norm": 7.376112937927246, + "learning_rate": 9.601752344445205e-05, + "loss": 1.0468, + "step": 3660 + }, + { + "epoch": 0.24805203604580256, + "grad_norm": 7.7266387939453125, + "learning_rate": 9.601615442535424e-05, + "loss": 1.007, + "step": 3661 + }, + { + "epoch": 0.24811979131377465, + "grad_norm": 9.196928977966309, + "learning_rate": 9.601478540625642e-05, + "loss": 1.1327, + "step": 3662 + }, + { + "epoch": 0.24818754658174674, + "grad_norm": 7.892288684844971, + "learning_rate": 9.60134163871586e-05, + "loss": 0.7511, + "step": 3663 + }, + { + "epoch": 0.2482553018497188, + "grad_norm": 8.854056358337402, + "learning_rate": 9.601204736806078e-05, + "loss": 0.9546, + "step": 3664 + }, + { + "epoch": 0.2483230571176909, + "grad_norm": 7.985452651977539, + "learning_rate": 9.601067834896296e-05, + "loss": 1.2186, + "step": 3665 + }, + { + "epoch": 0.248390812385663, + "grad_norm": 8.336162567138672, + "learning_rate": 9.600930932986516e-05, + "loss": 0.864, + "step": 3666 + }, + { + "epoch": 0.24845856765363508, + "grad_norm": 10.857757568359375, + "learning_rate": 9.600794031076734e-05, + "loss": 1.1948, + "step": 3667 + }, + { + "epoch": 0.24852632292160715, + "grad_norm": 8.168721199035645, + "learning_rate": 9.600657129166952e-05, + "loss": 1.1151, + "step": 3668 + }, + { + "epoch": 0.24859407818957924, + "grad_norm": 7.509332656860352, + "learning_rate": 9.60052022725717e-05, + "loss": 0.9215, + "step": 3669 + }, + { + "epoch": 0.24866183345755133, + "grad_norm": 10.4354829788208, + "learning_rate": 9.600383325347389e-05, + "loss": 1.0696, + "step": 3670 + }, + { + "epoch": 0.24872958872552342, + "grad_norm": 6.930381774902344, + "learning_rate": 9.600246423437607e-05, + "loss": 0.9229, + "step": 3671 + }, + { + "epoch": 0.24879734399349548, + "grad_norm": 6.608088970184326, + "learning_rate": 9.600109521527825e-05, + "loss": 0.928, + "step": 3672 + }, + { + "epoch": 0.24886509926146758, + "grad_norm": 7.4217095375061035, + "learning_rate": 9.599972619618043e-05, + "loss": 0.9608, + "step": 3673 + }, + { + "epoch": 0.24893285452943967, + "grad_norm": 7.46991491317749, + "learning_rate": 9.599835717708263e-05, + "loss": 0.8928, + "step": 3674 + }, + { + "epoch": 0.24900060979741176, + "grad_norm": 8.76001262664795, + "learning_rate": 9.599698815798481e-05, + "loss": 1.191, + "step": 3675 + }, + { + "epoch": 0.24906836506538382, + "grad_norm": 8.395779609680176, + "learning_rate": 9.599561913888699e-05, + "loss": 1.1479, + "step": 3676 + }, + { + "epoch": 0.24913612033335591, + "grad_norm": 8.00460147857666, + "learning_rate": 9.599425011978918e-05, + "loss": 1.1005, + "step": 3677 + }, + { + "epoch": 0.249203875601328, + "grad_norm": 7.817287445068359, + "learning_rate": 9.599288110069136e-05, + "loss": 1.1568, + "step": 3678 + }, + { + "epoch": 0.2492716308693001, + "grad_norm": 9.60706615447998, + "learning_rate": 9.599151208159354e-05, + "loss": 0.9734, + "step": 3679 + }, + { + "epoch": 0.24933938613727216, + "grad_norm": 8.347691535949707, + "learning_rate": 9.599014306249574e-05, + "loss": 1.2208, + "step": 3680 + }, + { + "epoch": 0.24940714140524425, + "grad_norm": 8.106613159179688, + "learning_rate": 9.598877404339792e-05, + "loss": 0.9253, + "step": 3681 + }, + { + "epoch": 0.24947489667321635, + "grad_norm": 8.099063873291016, + "learning_rate": 9.59874050243001e-05, + "loss": 1.0516, + "step": 3682 + }, + { + "epoch": 0.24954265194118844, + "grad_norm": 7.839589595794678, + "learning_rate": 9.598603600520228e-05, + "loss": 0.9764, + "step": 3683 + }, + { + "epoch": 0.2496104072091605, + "grad_norm": 8.679840087890625, + "learning_rate": 9.598466698610447e-05, + "loss": 1.0407, + "step": 3684 + }, + { + "epoch": 0.2496781624771326, + "grad_norm": 6.854926109313965, + "learning_rate": 9.598329796700665e-05, + "loss": 0.9457, + "step": 3685 + }, + { + "epoch": 0.24974591774510468, + "grad_norm": 9.133661270141602, + "learning_rate": 9.598192894790883e-05, + "loss": 1.0391, + "step": 3686 + }, + { + "epoch": 0.24981367301307678, + "grad_norm": 7.744726657867432, + "learning_rate": 9.598055992881101e-05, + "loss": 1.0256, + "step": 3687 + }, + { + "epoch": 0.24988142828104884, + "grad_norm": 8.9452543258667, + "learning_rate": 9.597919090971319e-05, + "loss": 0.9649, + "step": 3688 + }, + { + "epoch": 0.24994918354902093, + "grad_norm": 8.396431922912598, + "learning_rate": 9.597782189061539e-05, + "loss": 0.8189, + "step": 3689 + }, + { + "epoch": 0.24994918354902093, + "eval_loss": 0.9518795609474182, + "eval_noise_accuracy": 0.0, + "eval_runtime": 1533.2421, + "eval_samples_per_second": 3.352, + "eval_steps_per_second": 0.21, + "eval_wer": 88.77438705594233, + "step": 3689 + } + ], + "logging_steps": 1, + "max_steps": 73795, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 3689, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.45310507532288e+18, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}