diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7033 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.5885815185403178, + "eval_steps": 500, + "global_step": 1000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0005885815185403178, + "grad_norm": 35.75, + "learning_rate": 6.666666666666667e-07, + "loss": 3.1406, + "step": 1 + }, + { + "epoch": 0.0011771630370806356, + "grad_norm": 39.0, + "learning_rate": 1.3333333333333334e-06, + "loss": 3.1719, + "step": 2 + }, + { + "epoch": 0.0017657445556209534, + "grad_norm": 59.75, + "learning_rate": 2.0000000000000003e-06, + "loss": 3.2656, + "step": 3 + }, + { + "epoch": 0.002354326074161271, + "grad_norm": 41.25, + "learning_rate": 2.666666666666667e-06, + "loss": 3.1562, + "step": 4 + }, + { + "epoch": 0.002942907592701589, + "grad_norm": 43.5, + "learning_rate": 3.3333333333333333e-06, + "loss": 3.2812, + "step": 5 + }, + { + "epoch": 0.003531489111241907, + "grad_norm": 700.0, + "learning_rate": 4.000000000000001e-06, + "loss": 3.3438, + "step": 6 + }, + { + "epoch": 0.004120070629782225, + "grad_norm": 92.5, + "learning_rate": 4.666666666666667e-06, + "loss": 3.2969, + "step": 7 + }, + { + "epoch": 0.004708652148322542, + "grad_norm": 87.0, + "learning_rate": 5.333333333333334e-06, + "loss": 3.2031, + "step": 8 + }, + { + "epoch": 0.00529723366686286, + "grad_norm": 46.5, + "learning_rate": 6e-06, + "loss": 3.1094, + "step": 9 + }, + { + "epoch": 0.005885815185403178, + "grad_norm": 27.875, + "learning_rate": 6.666666666666667e-06, + "loss": 3.0781, + "step": 10 + }, + { + "epoch": 0.006474396703943496, + "grad_norm": 53.25, + "learning_rate": 7.333333333333334e-06, + "loss": 3.0781, + "step": 11 + }, + { + "epoch": 0.007062978222483814, + "grad_norm": 48.5, + "learning_rate": 8.000000000000001e-06, + "loss": 3.25, + "step": 12 + }, + { + "epoch": 0.007651559741024131, + "grad_norm": 35.75, + "learning_rate": 8.666666666666668e-06, + "loss": 3.0469, + "step": 13 + }, + { + "epoch": 0.00824014125956445, + "grad_norm": 165.0, + "learning_rate": 9.333333333333334e-06, + "loss": 3.2969, + "step": 14 + }, + { + "epoch": 0.008828722778104767, + "grad_norm": 213.0, + "learning_rate": 1e-05, + "loss": 3.0781, + "step": 15 + }, + { + "epoch": 0.009417304296645085, + "grad_norm": 69.0, + "learning_rate": 1.0666666666666667e-05, + "loss": 3.0469, + "step": 16 + }, + { + "epoch": 0.010005885815185403, + "grad_norm": 104.5, + "learning_rate": 1.1333333333333334e-05, + "loss": 3.0469, + "step": 17 + }, + { + "epoch": 0.01059446733372572, + "grad_norm": 20.625, + "learning_rate": 1.2e-05, + "loss": 3.1094, + "step": 18 + }, + { + "epoch": 0.011183048852266038, + "grad_norm": 38.25, + "learning_rate": 1.2666666666666668e-05, + "loss": 3.1719, + "step": 19 + }, + { + "epoch": 0.011771630370806356, + "grad_norm": 65.5, + "learning_rate": 1.3333333333333333e-05, + "loss": 3.0938, + "step": 20 + }, + { + "epoch": 0.012360211889346674, + "grad_norm": 39.5, + "learning_rate": 1.4000000000000001e-05, + "loss": 3.1406, + "step": 21 + }, + { + "epoch": 0.012948793407886992, + "grad_norm": 16.5, + "learning_rate": 1.4666666666666668e-05, + "loss": 3.0625, + "step": 22 + }, + { + "epoch": 0.01353737492642731, + "grad_norm": 35.5, + "learning_rate": 1.5333333333333334e-05, + "loss": 3.125, + "step": 23 + }, + { + "epoch": 0.014125956444967627, + "grad_norm": 133.0, + "learning_rate": 1.6000000000000003e-05, + "loss": 3.0625, + "step": 24 + }, + { + "epoch": 0.014714537963507945, + "grad_norm": 31.25, + "learning_rate": 1.6666666666666667e-05, + "loss": 3.1406, + "step": 25 + }, + { + "epoch": 0.015303119482048263, + "grad_norm": 22.5, + "learning_rate": 1.7333333333333336e-05, + "loss": 3.0938, + "step": 26 + }, + { + "epoch": 0.015891701000588582, + "grad_norm": 41.25, + "learning_rate": 1.8e-05, + "loss": 3.125, + "step": 27 + }, + { + "epoch": 0.0164802825191289, + "grad_norm": 42.75, + "learning_rate": 1.866666666666667e-05, + "loss": 3.0, + "step": 28 + }, + { + "epoch": 0.017068864037669218, + "grad_norm": 90.5, + "learning_rate": 1.9333333333333333e-05, + "loss": 3.0469, + "step": 29 + }, + { + "epoch": 0.017657445556209534, + "grad_norm": 45.25, + "learning_rate": 2e-05, + "loss": 3.1406, + "step": 30 + }, + { + "epoch": 0.018246027074749854, + "grad_norm": 105.0, + "learning_rate": 2.0666666666666666e-05, + "loss": 2.9531, + "step": 31 + }, + { + "epoch": 0.01883460859329017, + "grad_norm": 29.875, + "learning_rate": 2.1333333333333335e-05, + "loss": 3.1719, + "step": 32 + }, + { + "epoch": 0.01942319011183049, + "grad_norm": 29.125, + "learning_rate": 2.2000000000000003e-05, + "loss": 3.1875, + "step": 33 + }, + { + "epoch": 0.020011771630370805, + "grad_norm": 34.0, + "learning_rate": 2.2666666666666668e-05, + "loss": 3.125, + "step": 34 + }, + { + "epoch": 0.020600353148911125, + "grad_norm": 121.0, + "learning_rate": 2.3333333333333336e-05, + "loss": 3.2812, + "step": 35 + }, + { + "epoch": 0.02118893466745144, + "grad_norm": 18.75, + "learning_rate": 2.4e-05, + "loss": 3.125, + "step": 36 + }, + { + "epoch": 0.02177751618599176, + "grad_norm": 32.0, + "learning_rate": 2.466666666666667e-05, + "loss": 2.9688, + "step": 37 + }, + { + "epoch": 0.022366097704532076, + "grad_norm": 72.0, + "learning_rate": 2.5333333333333337e-05, + "loss": 2.9531, + "step": 38 + }, + { + "epoch": 0.022954679223072396, + "grad_norm": 59.25, + "learning_rate": 2.6000000000000002e-05, + "loss": 3.125, + "step": 39 + }, + { + "epoch": 0.023543260741612712, + "grad_norm": 20.625, + "learning_rate": 2.6666666666666667e-05, + "loss": 2.9844, + "step": 40 + }, + { + "epoch": 0.02413184226015303, + "grad_norm": 73.0, + "learning_rate": 2.733333333333333e-05, + "loss": 3.0156, + "step": 41 + }, + { + "epoch": 0.024720423778693348, + "grad_norm": 11.8125, + "learning_rate": 2.8000000000000003e-05, + "loss": 2.8906, + "step": 42 + }, + { + "epoch": 0.025309005297233667, + "grad_norm": 34.5, + "learning_rate": 2.8666666666666668e-05, + "loss": 3.0312, + "step": 43 + }, + { + "epoch": 0.025897586815773983, + "grad_norm": 25.375, + "learning_rate": 2.9333333333333336e-05, + "loss": 2.9844, + "step": 44 + }, + { + "epoch": 0.026486168334314303, + "grad_norm": 17.5, + "learning_rate": 3e-05, + "loss": 3.0, + "step": 45 + }, + { + "epoch": 0.02707474985285462, + "grad_norm": 29.0, + "learning_rate": 3.066666666666667e-05, + "loss": 3.0156, + "step": 46 + }, + { + "epoch": 0.02766333137139494, + "grad_norm": 56.25, + "learning_rate": 3.1333333333333334e-05, + "loss": 2.875, + "step": 47 + }, + { + "epoch": 0.028251912889935255, + "grad_norm": 133.0, + "learning_rate": 3.2000000000000005e-05, + "loss": 3.0, + "step": 48 + }, + { + "epoch": 0.028840494408475574, + "grad_norm": 21.5, + "learning_rate": 3.266666666666667e-05, + "loss": 2.8438, + "step": 49 + }, + { + "epoch": 0.02942907592701589, + "grad_norm": 21.0, + "learning_rate": 3.3333333333333335e-05, + "loss": 2.8438, + "step": 50 + }, + { + "epoch": 0.03001765744555621, + "grad_norm": 29.125, + "learning_rate": 3.4000000000000007e-05, + "loss": 2.8125, + "step": 51 + }, + { + "epoch": 0.030606238964096526, + "grad_norm": 34.25, + "learning_rate": 3.466666666666667e-05, + "loss": 2.8906, + "step": 52 + }, + { + "epoch": 0.031194820482636845, + "grad_norm": 74.5, + "learning_rate": 3.5333333333333336e-05, + "loss": 2.7344, + "step": 53 + }, + { + "epoch": 0.031783402001177165, + "grad_norm": 19.375, + "learning_rate": 3.6e-05, + "loss": 2.7969, + "step": 54 + }, + { + "epoch": 0.032371983519717484, + "grad_norm": 146.0, + "learning_rate": 3.6666666666666666e-05, + "loss": 2.9219, + "step": 55 + }, + { + "epoch": 0.0329605650382578, + "grad_norm": 40.25, + "learning_rate": 3.733333333333334e-05, + "loss": 2.6719, + "step": 56 + }, + { + "epoch": 0.033549146556798116, + "grad_norm": 64.0, + "learning_rate": 3.8e-05, + "loss": 2.5938, + "step": 57 + }, + { + "epoch": 0.034137728075338436, + "grad_norm": 44.0, + "learning_rate": 3.866666666666667e-05, + "loss": 2.5, + "step": 58 + }, + { + "epoch": 0.034726309593878756, + "grad_norm": 30.625, + "learning_rate": 3.933333333333333e-05, + "loss": 2.6719, + "step": 59 + }, + { + "epoch": 0.03531489111241907, + "grad_norm": 59.25, + "learning_rate": 4e-05, + "loss": 2.5312, + "step": 60 + }, + { + "epoch": 0.03590347263095939, + "grad_norm": 13.9375, + "learning_rate": 4.066666666666667e-05, + "loss": 2.4688, + "step": 61 + }, + { + "epoch": 0.03649205414949971, + "grad_norm": 91.5, + "learning_rate": 4.133333333333333e-05, + "loss": 2.5625, + "step": 62 + }, + { + "epoch": 0.03708063566804003, + "grad_norm": 54.5, + "learning_rate": 4.2e-05, + "loss": 2.5312, + "step": 63 + }, + { + "epoch": 0.03766921718658034, + "grad_norm": 234.0, + "learning_rate": 4.266666666666667e-05, + "loss": 2.5, + "step": 64 + }, + { + "epoch": 0.03825779870512066, + "grad_norm": 58.75, + "learning_rate": 4.3333333333333334e-05, + "loss": 2.2969, + "step": 65 + }, + { + "epoch": 0.03884638022366098, + "grad_norm": 26.875, + "learning_rate": 4.4000000000000006e-05, + "loss": 2.1562, + "step": 66 + }, + { + "epoch": 0.0394349617422013, + "grad_norm": 30.25, + "learning_rate": 4.466666666666667e-05, + "loss": 2.25, + "step": 67 + }, + { + "epoch": 0.04002354326074161, + "grad_norm": 61.0, + "learning_rate": 4.5333333333333335e-05, + "loss": 2.3594, + "step": 68 + }, + { + "epoch": 0.04061212477928193, + "grad_norm": 36.0, + "learning_rate": 4.600000000000001e-05, + "loss": 2.1719, + "step": 69 + }, + { + "epoch": 0.04120070629782225, + "grad_norm": 56.0, + "learning_rate": 4.666666666666667e-05, + "loss": 2.1875, + "step": 70 + }, + { + "epoch": 0.04178928781636257, + "grad_norm": 34.75, + "learning_rate": 4.7333333333333336e-05, + "loss": 2.1719, + "step": 71 + }, + { + "epoch": 0.04237786933490288, + "grad_norm": 58.25, + "learning_rate": 4.8e-05, + "loss": 2.125, + "step": 72 + }, + { + "epoch": 0.0429664508534432, + "grad_norm": 67.0, + "learning_rate": 4.866666666666667e-05, + "loss": 2.4531, + "step": 73 + }, + { + "epoch": 0.04355503237198352, + "grad_norm": 21.625, + "learning_rate": 4.933333333333334e-05, + "loss": 2.1094, + "step": 74 + }, + { + "epoch": 0.04414361389052384, + "grad_norm": 43.5, + "learning_rate": 5e-05, + "loss": 2.0156, + "step": 75 + }, + { + "epoch": 0.04473219540906415, + "grad_norm": 17.125, + "learning_rate": 5.0666666666666674e-05, + "loss": 2.0156, + "step": 76 + }, + { + "epoch": 0.04532077692760447, + "grad_norm": 40.25, + "learning_rate": 5.133333333333333e-05, + "loss": 2.1094, + "step": 77 + }, + { + "epoch": 0.04590935844614479, + "grad_norm": 29.875, + "learning_rate": 5.2000000000000004e-05, + "loss": 1.9297, + "step": 78 + }, + { + "epoch": 0.04649793996468511, + "grad_norm": 45.5, + "learning_rate": 5.266666666666666e-05, + "loss": 2.0938, + "step": 79 + }, + { + "epoch": 0.047086521483225424, + "grad_norm": 73.0, + "learning_rate": 5.333333333333333e-05, + "loss": 1.9453, + "step": 80 + }, + { + "epoch": 0.047675103001765744, + "grad_norm": 28.5, + "learning_rate": 5.4000000000000005e-05, + "loss": 1.7891, + "step": 81 + }, + { + "epoch": 0.04826368452030606, + "grad_norm": 544.0, + "learning_rate": 5.466666666666666e-05, + "loss": 2.0469, + "step": 82 + }, + { + "epoch": 0.04885226603884638, + "grad_norm": 42.25, + "learning_rate": 5.5333333333333334e-05, + "loss": 1.75, + "step": 83 + }, + { + "epoch": 0.049440847557386695, + "grad_norm": 52.25, + "learning_rate": 5.6000000000000006e-05, + "loss": 1.7031, + "step": 84 + }, + { + "epoch": 0.050029429075927015, + "grad_norm": 41.5, + "learning_rate": 5.666666666666667e-05, + "loss": 1.7969, + "step": 85 + }, + { + "epoch": 0.050618010594467334, + "grad_norm": 131.0, + "learning_rate": 5.7333333333333336e-05, + "loss": 1.5234, + "step": 86 + }, + { + "epoch": 0.051206592113007654, + "grad_norm": 52.0, + "learning_rate": 5.8e-05, + "loss": 1.625, + "step": 87 + }, + { + "epoch": 0.05179517363154797, + "grad_norm": 43.75, + "learning_rate": 5.866666666666667e-05, + "loss": 1.6719, + "step": 88 + }, + { + "epoch": 0.052383755150088286, + "grad_norm": 27.375, + "learning_rate": 5.9333333333333343e-05, + "loss": 1.7188, + "step": 89 + }, + { + "epoch": 0.052972336668628606, + "grad_norm": 32.25, + "learning_rate": 6e-05, + "loss": 1.8047, + "step": 90 + }, + { + "epoch": 0.053560918187168925, + "grad_norm": 46.25, + "learning_rate": 6.066666666666667e-05, + "loss": 1.5859, + "step": 91 + }, + { + "epoch": 0.05414949970570924, + "grad_norm": 76.5, + "learning_rate": 6.133333333333334e-05, + "loss": 1.4062, + "step": 92 + }, + { + "epoch": 0.05473808122424956, + "grad_norm": 187.0, + "learning_rate": 6.2e-05, + "loss": 1.3359, + "step": 93 + }, + { + "epoch": 0.05532666274278988, + "grad_norm": 36.5, + "learning_rate": 6.266666666666667e-05, + "loss": 1.3672, + "step": 94 + }, + { + "epoch": 0.055915244261330196, + "grad_norm": 41.5, + "learning_rate": 6.333333333333333e-05, + "loss": 1.1172, + "step": 95 + }, + { + "epoch": 0.05650382577987051, + "grad_norm": 141.0, + "learning_rate": 6.400000000000001e-05, + "loss": 1.6016, + "step": 96 + }, + { + "epoch": 0.05709240729841083, + "grad_norm": 49.75, + "learning_rate": 6.466666666666666e-05, + "loss": 1.4375, + "step": 97 + }, + { + "epoch": 0.05768098881695115, + "grad_norm": 23.875, + "learning_rate": 6.533333333333334e-05, + "loss": 1.3594, + "step": 98 + }, + { + "epoch": 0.05826957033549147, + "grad_norm": 37.75, + "learning_rate": 6.6e-05, + "loss": 1.5, + "step": 99 + }, + { + "epoch": 0.05885815185403178, + "grad_norm": 32.0, + "learning_rate": 6.666666666666667e-05, + "loss": 1.3438, + "step": 100 + }, + { + "epoch": 0.0594467333725721, + "grad_norm": 37.0, + "learning_rate": 6.733333333333333e-05, + "loss": 1.2578, + "step": 101 + }, + { + "epoch": 0.06003531489111242, + "grad_norm": 66.5, + "learning_rate": 6.800000000000001e-05, + "loss": 1.3125, + "step": 102 + }, + { + "epoch": 0.06062389640965274, + "grad_norm": 52.5, + "learning_rate": 6.866666666666666e-05, + "loss": 1.3281, + "step": 103 + }, + { + "epoch": 0.06121247792819305, + "grad_norm": 122.0, + "learning_rate": 6.933333333333334e-05, + "loss": 1.1641, + "step": 104 + }, + { + "epoch": 0.06180105944673337, + "grad_norm": 13.375, + "learning_rate": 7e-05, + "loss": 1.0547, + "step": 105 + }, + { + "epoch": 0.06238964096527369, + "grad_norm": 97.5, + "learning_rate": 7.066666666666667e-05, + "loss": 1.3203, + "step": 106 + }, + { + "epoch": 0.062978222483814, + "grad_norm": 53.25, + "learning_rate": 7.133333333333334e-05, + "loss": 1.0547, + "step": 107 + }, + { + "epoch": 0.06356680400235433, + "grad_norm": 25.125, + "learning_rate": 7.2e-05, + "loss": 1.3906, + "step": 108 + }, + { + "epoch": 0.06415538552089464, + "grad_norm": 50.75, + "learning_rate": 7.266666666666667e-05, + "loss": 1.2109, + "step": 109 + }, + { + "epoch": 0.06474396703943497, + "grad_norm": 78.0, + "learning_rate": 7.333333333333333e-05, + "loss": 1.1406, + "step": 110 + }, + { + "epoch": 0.06533254855797528, + "grad_norm": 81.0, + "learning_rate": 7.4e-05, + "loss": 1.6875, + "step": 111 + }, + { + "epoch": 0.0659211300765156, + "grad_norm": 187.0, + "learning_rate": 7.466666666666667e-05, + "loss": 1.3125, + "step": 112 + }, + { + "epoch": 0.06650971159505592, + "grad_norm": 104.5, + "learning_rate": 7.533333333333334e-05, + "loss": 0.8438, + "step": 113 + }, + { + "epoch": 0.06709829311359623, + "grad_norm": 34.75, + "learning_rate": 7.6e-05, + "loss": 1.0703, + "step": 114 + }, + { + "epoch": 0.06768687463213655, + "grad_norm": 42.75, + "learning_rate": 7.666666666666667e-05, + "loss": 1.3359, + "step": 115 + }, + { + "epoch": 0.06827545615067687, + "grad_norm": 61.75, + "learning_rate": 7.733333333333333e-05, + "loss": 1.1562, + "step": 116 + }, + { + "epoch": 0.06886403766921718, + "grad_norm": 53.0, + "learning_rate": 7.800000000000001e-05, + "loss": 1.1797, + "step": 117 + }, + { + "epoch": 0.06945261918775751, + "grad_norm": 50.75, + "learning_rate": 7.866666666666666e-05, + "loss": 1.3906, + "step": 118 + }, + { + "epoch": 0.07004120070629782, + "grad_norm": 57.75, + "learning_rate": 7.933333333333334e-05, + "loss": 0.9883, + "step": 119 + }, + { + "epoch": 0.07062978222483814, + "grad_norm": 43.5, + "learning_rate": 8e-05, + "loss": 0.8828, + "step": 120 + }, + { + "epoch": 0.07121836374337846, + "grad_norm": 38.25, + "learning_rate": 8.066666666666667e-05, + "loss": 0.918, + "step": 121 + }, + { + "epoch": 0.07180694526191878, + "grad_norm": 37.75, + "learning_rate": 8.133333333333334e-05, + "loss": 1.1875, + "step": 122 + }, + { + "epoch": 0.07239552678045909, + "grad_norm": 21.125, + "learning_rate": 8.2e-05, + "loss": 0.7461, + "step": 123 + }, + { + "epoch": 0.07298410829899941, + "grad_norm": 45.0, + "learning_rate": 8.266666666666667e-05, + "loss": 0.8945, + "step": 124 + }, + { + "epoch": 0.07357268981753973, + "grad_norm": 130.0, + "learning_rate": 8.333333333333334e-05, + "loss": 1.1484, + "step": 125 + }, + { + "epoch": 0.07416127133608005, + "grad_norm": 41.25, + "learning_rate": 8.4e-05, + "loss": 1.375, + "step": 126 + }, + { + "epoch": 0.07474985285462037, + "grad_norm": 53.5, + "learning_rate": 8.466666666666667e-05, + "loss": 1.0469, + "step": 127 + }, + { + "epoch": 0.07533843437316068, + "grad_norm": 318.0, + "learning_rate": 8.533333333333334e-05, + "loss": 1.0703, + "step": 128 + }, + { + "epoch": 0.075927015891701, + "grad_norm": 30.125, + "learning_rate": 8.6e-05, + "loss": 1.0547, + "step": 129 + }, + { + "epoch": 0.07651559741024132, + "grad_norm": 170.0, + "learning_rate": 8.666666666666667e-05, + "loss": 0.8633, + "step": 130 + }, + { + "epoch": 0.07710417892878163, + "grad_norm": 26.75, + "learning_rate": 8.733333333333333e-05, + "loss": 0.6914, + "step": 131 + }, + { + "epoch": 0.07769276044732196, + "grad_norm": 64.5, + "learning_rate": 8.800000000000001e-05, + "loss": 1.0391, + "step": 132 + }, + { + "epoch": 0.07828134196586227, + "grad_norm": 67.5, + "learning_rate": 8.866666666666668e-05, + "loss": 0.9961, + "step": 133 + }, + { + "epoch": 0.0788699234844026, + "grad_norm": 62.0, + "learning_rate": 8.933333333333334e-05, + "loss": 0.9023, + "step": 134 + }, + { + "epoch": 0.07945850500294291, + "grad_norm": 40.0, + "learning_rate": 9e-05, + "loss": 0.7578, + "step": 135 + }, + { + "epoch": 0.08004708652148322, + "grad_norm": 63.5, + "learning_rate": 9.066666666666667e-05, + "loss": 1.0, + "step": 136 + }, + { + "epoch": 0.08063566804002355, + "grad_norm": 94.5, + "learning_rate": 9.133333333333334e-05, + "loss": 1.0859, + "step": 137 + }, + { + "epoch": 0.08122424955856386, + "grad_norm": 39.5, + "learning_rate": 9.200000000000001e-05, + "loss": 0.6836, + "step": 138 + }, + { + "epoch": 0.08181283107710417, + "grad_norm": 26.5, + "learning_rate": 9.266666666666666e-05, + "loss": 0.7031, + "step": 139 + }, + { + "epoch": 0.0824014125956445, + "grad_norm": 108.0, + "learning_rate": 9.333333333333334e-05, + "loss": 1.0938, + "step": 140 + }, + { + "epoch": 0.08298999411418481, + "grad_norm": 38.0, + "learning_rate": 9.4e-05, + "loss": 0.6289, + "step": 141 + }, + { + "epoch": 0.08357857563272514, + "grad_norm": 89.0, + "learning_rate": 9.466666666666667e-05, + "loss": 0.8203, + "step": 142 + }, + { + "epoch": 0.08416715715126545, + "grad_norm": 98.5, + "learning_rate": 9.533333333333334e-05, + "loss": 0.5508, + "step": 143 + }, + { + "epoch": 0.08475573866980576, + "grad_norm": 39.25, + "learning_rate": 9.6e-05, + "loss": 0.7773, + "step": 144 + }, + { + "epoch": 0.08534432018834609, + "grad_norm": 49.5, + "learning_rate": 9.666666666666667e-05, + "loss": 0.7852, + "step": 145 + }, + { + "epoch": 0.0859329017068864, + "grad_norm": 190.0, + "learning_rate": 9.733333333333335e-05, + "loss": 1.1328, + "step": 146 + }, + { + "epoch": 0.08652148322542672, + "grad_norm": 48.0, + "learning_rate": 9.8e-05, + "loss": 0.7305, + "step": 147 + }, + { + "epoch": 0.08711006474396704, + "grad_norm": 284.0, + "learning_rate": 9.866666666666668e-05, + "loss": 0.8867, + "step": 148 + }, + { + "epoch": 0.08769864626250735, + "grad_norm": 41.5, + "learning_rate": 9.933333333333334e-05, + "loss": 0.9023, + "step": 149 + }, + { + "epoch": 0.08828722778104768, + "grad_norm": 29.75, + "learning_rate": 0.0001, + "loss": 0.5391, + "step": 150 + }, + { + "epoch": 0.088875809299588, + "grad_norm": 35.25, + "learning_rate": 0.00010066666666666667, + "loss": 0.543, + "step": 151 + }, + { + "epoch": 0.0894643908181283, + "grad_norm": 117.0, + "learning_rate": 0.00010133333333333335, + "loss": 0.6094, + "step": 152 + }, + { + "epoch": 0.09005297233666863, + "grad_norm": 16.0, + "learning_rate": 0.00010200000000000001, + "loss": 0.4746, + "step": 153 + }, + { + "epoch": 0.09064155385520895, + "grad_norm": 62.25, + "learning_rate": 0.00010266666666666666, + "loss": 0.8164, + "step": 154 + }, + { + "epoch": 0.09123013537374926, + "grad_norm": 43.5, + "learning_rate": 0.00010333333333333334, + "loss": 0.5977, + "step": 155 + }, + { + "epoch": 0.09181871689228958, + "grad_norm": 16.375, + "learning_rate": 0.00010400000000000001, + "loss": 0.4199, + "step": 156 + }, + { + "epoch": 0.0924072984108299, + "grad_norm": 18.75, + "learning_rate": 0.00010466666666666667, + "loss": 0.3906, + "step": 157 + }, + { + "epoch": 0.09299587992937022, + "grad_norm": 44.25, + "learning_rate": 0.00010533333333333332, + "loss": 0.5742, + "step": 158 + }, + { + "epoch": 0.09358446144791054, + "grad_norm": 17.5, + "learning_rate": 0.00010600000000000002, + "loss": 0.4531, + "step": 159 + }, + { + "epoch": 0.09417304296645085, + "grad_norm": 35.25, + "learning_rate": 0.00010666666666666667, + "loss": 0.6992, + "step": 160 + }, + { + "epoch": 0.09476162448499117, + "grad_norm": 32.0, + "learning_rate": 0.00010733333333333333, + "loss": 0.5234, + "step": 161 + }, + { + "epoch": 0.09535020600353149, + "grad_norm": 15.5625, + "learning_rate": 0.00010800000000000001, + "loss": 0.3984, + "step": 162 + }, + { + "epoch": 0.0959387875220718, + "grad_norm": 14.875, + "learning_rate": 0.00010866666666666667, + "loss": 0.4062, + "step": 163 + }, + { + "epoch": 0.09652736904061213, + "grad_norm": 15.125, + "learning_rate": 0.00010933333333333333, + "loss": 0.4141, + "step": 164 + }, + { + "epoch": 0.09711595055915244, + "grad_norm": 33.0, + "learning_rate": 0.00011000000000000002, + "loss": 0.4551, + "step": 165 + }, + { + "epoch": 0.09770453207769277, + "grad_norm": 11.8125, + "learning_rate": 0.00011066666666666667, + "loss": 0.4023, + "step": 166 + }, + { + "epoch": 0.09829311359623308, + "grad_norm": 18.625, + "learning_rate": 0.00011133333333333333, + "loss": 0.4805, + "step": 167 + }, + { + "epoch": 0.09888169511477339, + "grad_norm": 8.5, + "learning_rate": 0.00011200000000000001, + "loss": 0.2812, + "step": 168 + }, + { + "epoch": 0.09947027663331372, + "grad_norm": 21.0, + "learning_rate": 0.00011266666666666668, + "loss": 0.4199, + "step": 169 + }, + { + "epoch": 0.10005885815185403, + "grad_norm": 29.875, + "learning_rate": 0.00011333333333333334, + "loss": 0.3711, + "step": 170 + }, + { + "epoch": 0.10064743967039436, + "grad_norm": 72.0, + "learning_rate": 0.00011399999999999999, + "loss": 0.375, + "step": 171 + }, + { + "epoch": 0.10123602118893467, + "grad_norm": 72.5, + "learning_rate": 0.00011466666666666667, + "loss": 0.4961, + "step": 172 + }, + { + "epoch": 0.10182460270747498, + "grad_norm": 19.5, + "learning_rate": 0.00011533333333333334, + "loss": 0.4766, + "step": 173 + }, + { + "epoch": 0.10241318422601531, + "grad_norm": 56.75, + "learning_rate": 0.000116, + "loss": 0.3223, + "step": 174 + }, + { + "epoch": 0.10300176574455562, + "grad_norm": 20.25, + "learning_rate": 0.00011666666666666668, + "loss": 0.2363, + "step": 175 + }, + { + "epoch": 0.10359034726309593, + "grad_norm": 26.75, + "learning_rate": 0.00011733333333333334, + "loss": 0.3164, + "step": 176 + }, + { + "epoch": 0.10417892878163626, + "grad_norm": 23.75, + "learning_rate": 0.000118, + "loss": 0.4141, + "step": 177 + }, + { + "epoch": 0.10476751030017657, + "grad_norm": 13.625, + "learning_rate": 0.00011866666666666669, + "loss": 0.3438, + "step": 178 + }, + { + "epoch": 0.1053560918187169, + "grad_norm": 20.25, + "learning_rate": 0.00011933333333333334, + "loss": 0.373, + "step": 179 + }, + { + "epoch": 0.10594467333725721, + "grad_norm": 17.875, + "learning_rate": 0.00012, + "loss": 0.4336, + "step": 180 + }, + { + "epoch": 0.10653325485579752, + "grad_norm": 8.875, + "learning_rate": 0.00012066666666666668, + "loss": 0.2002, + "step": 181 + }, + { + "epoch": 0.10712183637433785, + "grad_norm": 41.5, + "learning_rate": 0.00012133333333333335, + "loss": 0.2812, + "step": 182 + }, + { + "epoch": 0.10771041789287816, + "grad_norm": 24.375, + "learning_rate": 0.000122, + "loss": 0.2119, + "step": 183 + }, + { + "epoch": 0.10829899941141848, + "grad_norm": 8.5625, + "learning_rate": 0.00012266666666666668, + "loss": 0.1934, + "step": 184 + }, + { + "epoch": 0.1088875809299588, + "grad_norm": 540.0, + "learning_rate": 0.00012333333333333334, + "loss": 0.3203, + "step": 185 + }, + { + "epoch": 0.10947616244849911, + "grad_norm": 56.5, + "learning_rate": 0.000124, + "loss": 0.3809, + "step": 186 + }, + { + "epoch": 0.11006474396703944, + "grad_norm": 50.0, + "learning_rate": 0.00012466666666666667, + "loss": 0.207, + "step": 187 + }, + { + "epoch": 0.11065332548557975, + "grad_norm": 28.5, + "learning_rate": 0.00012533333333333334, + "loss": 0.3281, + "step": 188 + }, + { + "epoch": 0.11124190700412007, + "grad_norm": 20.375, + "learning_rate": 0.000126, + "loss": 0.373, + "step": 189 + }, + { + "epoch": 0.11183048852266039, + "grad_norm": 35.25, + "learning_rate": 0.00012666666666666666, + "loss": 0.3535, + "step": 190 + }, + { + "epoch": 0.1124190700412007, + "grad_norm": 23.5, + "learning_rate": 0.00012733333333333336, + "loss": 0.1885, + "step": 191 + }, + { + "epoch": 0.11300765155974102, + "grad_norm": 18.75, + "learning_rate": 0.00012800000000000002, + "loss": 0.4121, + "step": 192 + }, + { + "epoch": 0.11359623307828134, + "grad_norm": 96.0, + "learning_rate": 0.00012866666666666666, + "loss": 0.3574, + "step": 193 + }, + { + "epoch": 0.11418481459682166, + "grad_norm": 12.9375, + "learning_rate": 0.00012933333333333332, + "loss": 0.4023, + "step": 194 + }, + { + "epoch": 0.11477339611536198, + "grad_norm": 103.5, + "learning_rate": 0.00013000000000000002, + "loss": 0.3477, + "step": 195 + }, + { + "epoch": 0.1153619776339023, + "grad_norm": 13.3125, + "learning_rate": 0.00013066666666666668, + "loss": 0.3047, + "step": 196 + }, + { + "epoch": 0.11595055915244261, + "grad_norm": 96.5, + "learning_rate": 0.00013133333333333332, + "loss": 0.2773, + "step": 197 + }, + { + "epoch": 0.11653914067098294, + "grad_norm": 15.3125, + "learning_rate": 0.000132, + "loss": 0.2148, + "step": 198 + }, + { + "epoch": 0.11712772218952325, + "grad_norm": 24.125, + "learning_rate": 0.00013266666666666667, + "loss": 0.2461, + "step": 199 + }, + { + "epoch": 0.11771630370806356, + "grad_norm": 22.75, + "learning_rate": 0.00013333333333333334, + "loss": 0.2617, + "step": 200 + }, + { + "epoch": 0.11830488522660389, + "grad_norm": 23.625, + "learning_rate": 0.000134, + "loss": 0.1396, + "step": 201 + }, + { + "epoch": 0.1188934667451442, + "grad_norm": 7.21875, + "learning_rate": 0.00013466666666666667, + "loss": 0.1514, + "step": 202 + }, + { + "epoch": 0.11948204826368453, + "grad_norm": 37.0, + "learning_rate": 0.00013533333333333333, + "loss": 0.3223, + "step": 203 + }, + { + "epoch": 0.12007062978222484, + "grad_norm": 17.625, + "learning_rate": 0.00013600000000000003, + "loss": 0.2324, + "step": 204 + }, + { + "epoch": 0.12065921130076515, + "grad_norm": 11.875, + "learning_rate": 0.00013666666666666666, + "loss": 0.2051, + "step": 205 + }, + { + "epoch": 0.12124779281930548, + "grad_norm": 13.3125, + "learning_rate": 0.00013733333333333333, + "loss": 0.2793, + "step": 206 + }, + { + "epoch": 0.12183637433784579, + "grad_norm": 13.3125, + "learning_rate": 0.000138, + "loss": 0.2061, + "step": 207 + }, + { + "epoch": 0.1224249558563861, + "grad_norm": 16.75, + "learning_rate": 0.00013866666666666669, + "loss": 0.2148, + "step": 208 + }, + { + "epoch": 0.12301353737492643, + "grad_norm": 11.625, + "learning_rate": 0.00013933333333333335, + "loss": 0.2324, + "step": 209 + }, + { + "epoch": 0.12360211889346674, + "grad_norm": 17.5, + "learning_rate": 0.00014, + "loss": 0.1396, + "step": 210 + }, + { + "epoch": 0.12419070041200707, + "grad_norm": 4.40625, + "learning_rate": 0.00014066666666666668, + "loss": 0.106, + "step": 211 + }, + { + "epoch": 0.12477928193054738, + "grad_norm": 72.0, + "learning_rate": 0.00014133333333333334, + "loss": 0.2012, + "step": 212 + }, + { + "epoch": 0.1253678634490877, + "grad_norm": 23.25, + "learning_rate": 0.000142, + "loss": 0.0991, + "step": 213 + }, + { + "epoch": 0.125956444967628, + "grad_norm": 21.25, + "learning_rate": 0.00014266666666666667, + "loss": 0.3809, + "step": 214 + }, + { + "epoch": 0.12654502648616833, + "grad_norm": 6.0625, + "learning_rate": 0.00014333333333333334, + "loss": 0.1206, + "step": 215 + }, + { + "epoch": 0.12713360800470866, + "grad_norm": 42.0, + "learning_rate": 0.000144, + "loss": 0.1445, + "step": 216 + }, + { + "epoch": 0.12772218952324896, + "grad_norm": 17.125, + "learning_rate": 0.0001446666666666667, + "loss": 0.2344, + "step": 217 + }, + { + "epoch": 0.12831077104178928, + "grad_norm": 19.875, + "learning_rate": 0.00014533333333333333, + "loss": 0.2266, + "step": 218 + }, + { + "epoch": 0.1288993525603296, + "grad_norm": 25.0, + "learning_rate": 0.000146, + "loss": 0.1045, + "step": 219 + }, + { + "epoch": 0.12948793407886994, + "grad_norm": 14.1875, + "learning_rate": 0.00014666666666666666, + "loss": 0.1592, + "step": 220 + }, + { + "epoch": 0.13007651559741024, + "grad_norm": 8.75, + "learning_rate": 0.00014733333333333335, + "loss": 0.1982, + "step": 221 + }, + { + "epoch": 0.13066509711595056, + "grad_norm": 9.1875, + "learning_rate": 0.000148, + "loss": 0.1436, + "step": 222 + }, + { + "epoch": 0.1312536786344909, + "grad_norm": 6.8125, + "learning_rate": 0.00014866666666666666, + "loss": 0.1172, + "step": 223 + }, + { + "epoch": 0.1318422601530312, + "grad_norm": 17.25, + "learning_rate": 0.00014933333333333335, + "loss": 0.123, + "step": 224 + }, + { + "epoch": 0.13243084167157151, + "grad_norm": 12.625, + "learning_rate": 0.00015000000000000001, + "loss": 0.1055, + "step": 225 + }, + { + "epoch": 0.13301942319011184, + "grad_norm": 7.46875, + "learning_rate": 0.00015066666666666668, + "loss": 0.1641, + "step": 226 + }, + { + "epoch": 0.13360800470865214, + "grad_norm": 8.4375, + "learning_rate": 0.00015133333333333334, + "loss": 0.1069, + "step": 227 + }, + { + "epoch": 0.13419658622719247, + "grad_norm": 15.5, + "learning_rate": 0.000152, + "loss": 0.1436, + "step": 228 + }, + { + "epoch": 0.1347851677457328, + "grad_norm": 18.625, + "learning_rate": 0.00015266666666666667, + "loss": 0.1172, + "step": 229 + }, + { + "epoch": 0.1353737492642731, + "grad_norm": 8.4375, + "learning_rate": 0.00015333333333333334, + "loss": 0.1631, + "step": 230 + }, + { + "epoch": 0.13596233078281342, + "grad_norm": 8.1875, + "learning_rate": 0.000154, + "loss": 0.1118, + "step": 231 + }, + { + "epoch": 0.13655091230135374, + "grad_norm": 7.71875, + "learning_rate": 0.00015466666666666667, + "loss": 0.2559, + "step": 232 + }, + { + "epoch": 0.13713949381989404, + "grad_norm": 28.875, + "learning_rate": 0.00015533333333333333, + "loss": 0.1641, + "step": 233 + }, + { + "epoch": 0.13772807533843437, + "grad_norm": 5.78125, + "learning_rate": 0.00015600000000000002, + "loss": 0.0732, + "step": 234 + }, + { + "epoch": 0.1383166568569747, + "grad_norm": 21.5, + "learning_rate": 0.00015666666666666666, + "loss": 0.248, + "step": 235 + }, + { + "epoch": 0.13890523837551502, + "grad_norm": 9.375, + "learning_rate": 0.00015733333333333333, + "loss": 0.1504, + "step": 236 + }, + { + "epoch": 0.13949381989405532, + "grad_norm": 8.5625, + "learning_rate": 0.00015800000000000002, + "loss": 0.1226, + "step": 237 + }, + { + "epoch": 0.14008240141259565, + "grad_norm": 2.9375, + "learning_rate": 0.00015866666666666668, + "loss": 0.106, + "step": 238 + }, + { + "epoch": 0.14067098293113597, + "grad_norm": 5.53125, + "learning_rate": 0.00015933333333333332, + "loss": 0.1328, + "step": 239 + }, + { + "epoch": 0.14125956444967627, + "grad_norm": 8.3125, + "learning_rate": 0.00016, + "loss": 0.1001, + "step": 240 + }, + { + "epoch": 0.1418481459682166, + "grad_norm": 3.53125, + "learning_rate": 0.00016066666666666668, + "loss": 0.0588, + "step": 241 + }, + { + "epoch": 0.14243672748675693, + "grad_norm": 5.25, + "learning_rate": 0.00016133333333333334, + "loss": 0.1128, + "step": 242 + }, + { + "epoch": 0.14302530900529722, + "grad_norm": 14.375, + "learning_rate": 0.000162, + "loss": 0.293, + "step": 243 + }, + { + "epoch": 0.14361389052383755, + "grad_norm": 13.5, + "learning_rate": 0.00016266666666666667, + "loss": 0.0859, + "step": 244 + }, + { + "epoch": 0.14420247204237788, + "grad_norm": 21.375, + "learning_rate": 0.00016333333333333334, + "loss": 0.124, + "step": 245 + }, + { + "epoch": 0.14479105356091818, + "grad_norm": 11.625, + "learning_rate": 0.000164, + "loss": 0.1191, + "step": 246 + }, + { + "epoch": 0.1453796350794585, + "grad_norm": 5.28125, + "learning_rate": 0.00016466666666666667, + "loss": 0.0728, + "step": 247 + }, + { + "epoch": 0.14596821659799883, + "grad_norm": 10.3125, + "learning_rate": 0.00016533333333333333, + "loss": 0.1562, + "step": 248 + }, + { + "epoch": 0.14655679811653913, + "grad_norm": 2.53125, + "learning_rate": 0.000166, + "loss": 0.0752, + "step": 249 + }, + { + "epoch": 0.14714537963507945, + "grad_norm": 3.078125, + "learning_rate": 0.0001666666666666667, + "loss": 0.2041, + "step": 250 + }, + { + "epoch": 0.14773396115361978, + "grad_norm": 4.21875, + "learning_rate": 0.00016733333333333335, + "loss": 0.0796, + "step": 251 + }, + { + "epoch": 0.1483225426721601, + "grad_norm": 13.9375, + "learning_rate": 0.000168, + "loss": 0.0781, + "step": 252 + }, + { + "epoch": 0.1489111241907004, + "grad_norm": 11.5625, + "learning_rate": 0.00016866666666666668, + "loss": 0.1162, + "step": 253 + }, + { + "epoch": 0.14949970570924073, + "grad_norm": 13.0625, + "learning_rate": 0.00016933333333333335, + "loss": 0.083, + "step": 254 + }, + { + "epoch": 0.15008828722778106, + "grad_norm": 2.765625, + "learning_rate": 0.00017, + "loss": 0.053, + "step": 255 + }, + { + "epoch": 0.15067686874632136, + "grad_norm": 3.484375, + "learning_rate": 0.00017066666666666668, + "loss": 0.0806, + "step": 256 + }, + { + "epoch": 0.15126545026486168, + "grad_norm": 8.125, + "learning_rate": 0.00017133333333333334, + "loss": 0.0771, + "step": 257 + }, + { + "epoch": 0.151854031783402, + "grad_norm": 16.875, + "learning_rate": 0.000172, + "loss": 0.0752, + "step": 258 + }, + { + "epoch": 0.1524426133019423, + "grad_norm": 7.84375, + "learning_rate": 0.00017266666666666667, + "loss": 0.1406, + "step": 259 + }, + { + "epoch": 0.15303119482048264, + "grad_norm": 1.921875, + "learning_rate": 0.00017333333333333334, + "loss": 0.0603, + "step": 260 + }, + { + "epoch": 0.15361977633902296, + "grad_norm": 5.46875, + "learning_rate": 0.000174, + "loss": 0.0547, + "step": 261 + }, + { + "epoch": 0.15420835785756326, + "grad_norm": 3.921875, + "learning_rate": 0.00017466666666666667, + "loss": 0.0525, + "step": 262 + }, + { + "epoch": 0.1547969393761036, + "grad_norm": 7.53125, + "learning_rate": 0.00017533333333333336, + "loss": 0.0713, + "step": 263 + }, + { + "epoch": 0.1553855208946439, + "grad_norm": 2.90625, + "learning_rate": 0.00017600000000000002, + "loss": 0.0537, + "step": 264 + }, + { + "epoch": 0.1559741024131842, + "grad_norm": 2.0625, + "learning_rate": 0.00017666666666666666, + "loss": 0.0698, + "step": 265 + }, + { + "epoch": 0.15656268393172454, + "grad_norm": 5.875, + "learning_rate": 0.00017733333333333335, + "loss": 0.0544, + "step": 266 + }, + { + "epoch": 0.15715126545026487, + "grad_norm": 4.28125, + "learning_rate": 0.00017800000000000002, + "loss": 0.1099, + "step": 267 + }, + { + "epoch": 0.1577398469688052, + "grad_norm": 1.6171875, + "learning_rate": 0.00017866666666666668, + "loss": 0.0593, + "step": 268 + }, + { + "epoch": 0.1583284284873455, + "grad_norm": 20.625, + "learning_rate": 0.00017933333333333332, + "loss": 0.0728, + "step": 269 + }, + { + "epoch": 0.15891701000588582, + "grad_norm": 15.375, + "learning_rate": 0.00018, + "loss": 0.084, + "step": 270 + }, + { + "epoch": 0.15950559152442614, + "grad_norm": 2.09375, + "learning_rate": 0.00018066666666666668, + "loss": 0.0776, + "step": 271 + }, + { + "epoch": 0.16009417304296644, + "grad_norm": 8.0, + "learning_rate": 0.00018133333333333334, + "loss": 0.1187, + "step": 272 + }, + { + "epoch": 0.16068275456150677, + "grad_norm": 0.7421875, + "learning_rate": 0.000182, + "loss": 0.0398, + "step": 273 + }, + { + "epoch": 0.1612713360800471, + "grad_norm": 2.171875, + "learning_rate": 0.00018266666666666667, + "loss": 0.053, + "step": 274 + }, + { + "epoch": 0.1618599175985874, + "grad_norm": 11.0, + "learning_rate": 0.00018333333333333334, + "loss": 0.0723, + "step": 275 + }, + { + "epoch": 0.16244849911712772, + "grad_norm": 2.21875, + "learning_rate": 0.00018400000000000003, + "loss": 0.1396, + "step": 276 + }, + { + "epoch": 0.16303708063566805, + "grad_norm": 0.4296875, + "learning_rate": 0.00018466666666666666, + "loss": 0.0291, + "step": 277 + }, + { + "epoch": 0.16362566215420835, + "grad_norm": 38.75, + "learning_rate": 0.00018533333333333333, + "loss": 0.0547, + "step": 278 + }, + { + "epoch": 0.16421424367274867, + "grad_norm": 0.62109375, + "learning_rate": 0.00018600000000000002, + "loss": 0.0403, + "step": 279 + }, + { + "epoch": 0.164802825191289, + "grad_norm": 2.640625, + "learning_rate": 0.0001866666666666667, + "loss": 0.063, + "step": 280 + }, + { + "epoch": 0.16539140670982933, + "grad_norm": 4.21875, + "learning_rate": 0.00018733333333333335, + "loss": 0.0447, + "step": 281 + }, + { + "epoch": 0.16597998822836962, + "grad_norm": 3.09375, + "learning_rate": 0.000188, + "loss": 0.0415, + "step": 282 + }, + { + "epoch": 0.16656856974690995, + "grad_norm": 1.453125, + "learning_rate": 0.00018866666666666668, + "loss": 0.0513, + "step": 283 + }, + { + "epoch": 0.16715715126545028, + "grad_norm": 9.0625, + "learning_rate": 0.00018933333333333335, + "loss": 0.0295, + "step": 284 + }, + { + "epoch": 0.16774573278399058, + "grad_norm": 17.75, + "learning_rate": 0.00019, + "loss": 0.0747, + "step": 285 + }, + { + "epoch": 0.1683343143025309, + "grad_norm": 10.6875, + "learning_rate": 0.00019066666666666668, + "loss": 0.0664, + "step": 286 + }, + { + "epoch": 0.16892289582107123, + "grad_norm": 13.8125, + "learning_rate": 0.00019133333333333334, + "loss": 0.052, + "step": 287 + }, + { + "epoch": 0.16951147733961153, + "grad_norm": 0.94140625, + "learning_rate": 0.000192, + "loss": 0.0444, + "step": 288 + }, + { + "epoch": 0.17010005885815185, + "grad_norm": 8.5, + "learning_rate": 0.0001926666666666667, + "loss": 0.0618, + "step": 289 + }, + { + "epoch": 0.17068864037669218, + "grad_norm": 4.59375, + "learning_rate": 0.00019333333333333333, + "loss": 0.0569, + "step": 290 + }, + { + "epoch": 0.17127722189523248, + "grad_norm": 2.78125, + "learning_rate": 0.000194, + "loss": 0.0598, + "step": 291 + }, + { + "epoch": 0.1718658034137728, + "grad_norm": 15.125, + "learning_rate": 0.0001946666666666667, + "loss": 0.0986, + "step": 292 + }, + { + "epoch": 0.17245438493231313, + "grad_norm": 0.33984375, + "learning_rate": 0.00019533333333333336, + "loss": 0.026, + "step": 293 + }, + { + "epoch": 0.17304296645085343, + "grad_norm": 1.6875, + "learning_rate": 0.000196, + "loss": 0.0369, + "step": 294 + }, + { + "epoch": 0.17363154796939376, + "grad_norm": 0.7890625, + "learning_rate": 0.00019666666666666666, + "loss": 0.031, + "step": 295 + }, + { + "epoch": 0.17422012948793408, + "grad_norm": 8.5, + "learning_rate": 0.00019733333333333335, + "loss": 0.0903, + "step": 296 + }, + { + "epoch": 0.1748087110064744, + "grad_norm": 1.015625, + "learning_rate": 0.00019800000000000002, + "loss": 0.0286, + "step": 297 + }, + { + "epoch": 0.1753972925250147, + "grad_norm": 1.90625, + "learning_rate": 0.00019866666666666668, + "loss": 0.1016, + "step": 298 + }, + { + "epoch": 0.17598587404355504, + "grad_norm": 0.361328125, + "learning_rate": 0.00019933333333333334, + "loss": 0.0239, + "step": 299 + }, + { + "epoch": 0.17657445556209536, + "grad_norm": 7.9375, + "learning_rate": 0.0002, + "loss": 0.0408, + "step": 300 + }, + { + "epoch": 0.17716303708063566, + "grad_norm": 1.71875, + "learning_rate": 0.00019997559487492376, + "loss": 0.0317, + "step": 301 + }, + { + "epoch": 0.177751618599176, + "grad_norm": 0.81640625, + "learning_rate": 0.00019995118974984748, + "loss": 0.0327, + "step": 302 + }, + { + "epoch": 0.1783402001177163, + "grad_norm": 2.21875, + "learning_rate": 0.00019992678462477123, + "loss": 0.0698, + "step": 303 + }, + { + "epoch": 0.1789287816362566, + "grad_norm": 0.24609375, + "learning_rate": 0.00019990237949969495, + "loss": 0.022, + "step": 304 + }, + { + "epoch": 0.17951736315479694, + "grad_norm": 0.3125, + "learning_rate": 0.0001998779743746187, + "loss": 0.0231, + "step": 305 + }, + { + "epoch": 0.18010594467333726, + "grad_norm": 4.375, + "learning_rate": 0.00019985356924954241, + "loss": 0.0232, + "step": 306 + }, + { + "epoch": 0.18069452619187756, + "grad_norm": 3.1875, + "learning_rate": 0.00019982916412446616, + "loss": 0.0403, + "step": 307 + }, + { + "epoch": 0.1812831077104179, + "grad_norm": 2.984375, + "learning_rate": 0.00019980475899938988, + "loss": 0.085, + "step": 308 + }, + { + "epoch": 0.18187168922895822, + "grad_norm": 1.0, + "learning_rate": 0.00019978035387431363, + "loss": 0.0304, + "step": 309 + }, + { + "epoch": 0.18246027074749852, + "grad_norm": 5.09375, + "learning_rate": 0.00019975594874923735, + "loss": 0.033, + "step": 310 + }, + { + "epoch": 0.18304885226603884, + "grad_norm": 0.26953125, + "learning_rate": 0.0001997315436241611, + "loss": 0.0184, + "step": 311 + }, + { + "epoch": 0.18363743378457917, + "grad_norm": 0.47265625, + "learning_rate": 0.00019970713849908482, + "loss": 0.0287, + "step": 312 + }, + { + "epoch": 0.1842260153031195, + "grad_norm": 0.466796875, + "learning_rate": 0.00019968273337400857, + "loss": 0.0228, + "step": 313 + }, + { + "epoch": 0.1848145968216598, + "grad_norm": 1.2734375, + "learning_rate": 0.00019965832824893229, + "loss": 0.0337, + "step": 314 + }, + { + "epoch": 0.18540317834020012, + "grad_norm": 25.25, + "learning_rate": 0.00019963392312385603, + "loss": 0.1172, + "step": 315 + }, + { + "epoch": 0.18599175985874045, + "grad_norm": 2.0, + "learning_rate": 0.00019960951799877975, + "loss": 0.0398, + "step": 316 + }, + { + "epoch": 0.18658034137728075, + "grad_norm": 0.29296875, + "learning_rate": 0.0001995851128737035, + "loss": 0.0248, + "step": 317 + }, + { + "epoch": 0.18716892289582107, + "grad_norm": 0.439453125, + "learning_rate": 0.00019956070774862722, + "loss": 0.0254, + "step": 318 + }, + { + "epoch": 0.1877575044143614, + "grad_norm": 0.244140625, + "learning_rate": 0.00019953630262355097, + "loss": 0.022, + "step": 319 + }, + { + "epoch": 0.1883460859329017, + "grad_norm": 2.671875, + "learning_rate": 0.0001995118974984747, + "loss": 0.0427, + "step": 320 + }, + { + "epoch": 0.18893466745144202, + "grad_norm": 1.0546875, + "learning_rate": 0.00019948749237339844, + "loss": 0.0415, + "step": 321 + }, + { + "epoch": 0.18952324896998235, + "grad_norm": 2.0625, + "learning_rate": 0.00019946308724832216, + "loss": 0.0806, + "step": 322 + }, + { + "epoch": 0.19011183048852265, + "grad_norm": 1.6484375, + "learning_rate": 0.0001994386821232459, + "loss": 0.0405, + "step": 323 + }, + { + "epoch": 0.19070041200706297, + "grad_norm": 2.625, + "learning_rate": 0.00019941427699816963, + "loss": 0.0258, + "step": 324 + }, + { + "epoch": 0.1912889935256033, + "grad_norm": 1.21875, + "learning_rate": 0.00019938987187309337, + "loss": 0.0344, + "step": 325 + }, + { + "epoch": 0.1918775750441436, + "grad_norm": 0.6171875, + "learning_rate": 0.0001993654667480171, + "loss": 0.026, + "step": 326 + }, + { + "epoch": 0.19246615656268393, + "grad_norm": 0.52734375, + "learning_rate": 0.00019934106162294081, + "loss": 0.0264, + "step": 327 + }, + { + "epoch": 0.19305473808122425, + "grad_norm": 0.8828125, + "learning_rate": 0.00019931665649786456, + "loss": 0.0262, + "step": 328 + }, + { + "epoch": 0.19364331959976458, + "grad_norm": 0.2265625, + "learning_rate": 0.00019929225137278828, + "loss": 0.0195, + "step": 329 + }, + { + "epoch": 0.19423190111830488, + "grad_norm": 0.9375, + "learning_rate": 0.00019926784624771203, + "loss": 0.0374, + "step": 330 + }, + { + "epoch": 0.1948204826368452, + "grad_norm": 2.875, + "learning_rate": 0.00019924344112263575, + "loss": 0.0344, + "step": 331 + }, + { + "epoch": 0.19540906415538553, + "grad_norm": 0.3125, + "learning_rate": 0.0001992190359975595, + "loss": 0.0199, + "step": 332 + }, + { + "epoch": 0.19599764567392583, + "grad_norm": 0.87109375, + "learning_rate": 0.00019919463087248322, + "loss": 0.033, + "step": 333 + }, + { + "epoch": 0.19658622719246616, + "grad_norm": 0.166015625, + "learning_rate": 0.00019917022574740697, + "loss": 0.0157, + "step": 334 + }, + { + "epoch": 0.19717480871100648, + "grad_norm": 1.5859375, + "learning_rate": 0.00019914582062233069, + "loss": 0.0266, + "step": 335 + }, + { + "epoch": 0.19776339022954678, + "grad_norm": 0.150390625, + "learning_rate": 0.00019912141549725443, + "loss": 0.0153, + "step": 336 + }, + { + "epoch": 0.1983519717480871, + "grad_norm": 0.318359375, + "learning_rate": 0.00019909701037217815, + "loss": 0.0197, + "step": 337 + }, + { + "epoch": 0.19894055326662743, + "grad_norm": 4.8125, + "learning_rate": 0.0001990726052471019, + "loss": 0.0471, + "step": 338 + }, + { + "epoch": 0.19952913478516773, + "grad_norm": 1.2890625, + "learning_rate": 0.00019904820012202562, + "loss": 0.0786, + "step": 339 + }, + { + "epoch": 0.20011771630370806, + "grad_norm": 11.5, + "learning_rate": 0.00019902379499694937, + "loss": 0.0413, + "step": 340 + }, + { + "epoch": 0.2007062978222484, + "grad_norm": 1.640625, + "learning_rate": 0.0001989993898718731, + "loss": 0.0508, + "step": 341 + }, + { + "epoch": 0.2012948793407887, + "grad_norm": 1.15625, + "learning_rate": 0.00019897498474679684, + "loss": 0.0297, + "step": 342 + }, + { + "epoch": 0.201883460859329, + "grad_norm": 0.384765625, + "learning_rate": 0.00019895057962172056, + "loss": 0.0261, + "step": 343 + }, + { + "epoch": 0.20247204237786934, + "grad_norm": 0.12890625, + "learning_rate": 0.0001989261744966443, + "loss": 0.0145, + "step": 344 + }, + { + "epoch": 0.20306062389640966, + "grad_norm": 5.0, + "learning_rate": 0.00019890176937156803, + "loss": 0.0239, + "step": 345 + }, + { + "epoch": 0.20364920541494996, + "grad_norm": 0.96484375, + "learning_rate": 0.00019887736424649177, + "loss": 0.0206, + "step": 346 + }, + { + "epoch": 0.2042377869334903, + "grad_norm": 0.318359375, + "learning_rate": 0.0001988529591214155, + "loss": 0.0181, + "step": 347 + }, + { + "epoch": 0.20482636845203062, + "grad_norm": 0.671875, + "learning_rate": 0.00019882855399633924, + "loss": 0.0226, + "step": 348 + }, + { + "epoch": 0.20541494997057091, + "grad_norm": 0.365234375, + "learning_rate": 0.00019880414887126296, + "loss": 0.0175, + "step": 349 + }, + { + "epoch": 0.20600353148911124, + "grad_norm": 0.1904296875, + "learning_rate": 0.0001987797437461867, + "loss": 0.0181, + "step": 350 + }, + { + "epoch": 0.20659211300765157, + "grad_norm": 0.1279296875, + "learning_rate": 0.00019875533862111043, + "loss": 0.0135, + "step": 351 + }, + { + "epoch": 0.20718069452619187, + "grad_norm": 0.2041015625, + "learning_rate": 0.00019873093349603418, + "loss": 0.0388, + "step": 352 + }, + { + "epoch": 0.2077692760447322, + "grad_norm": 0.2041015625, + "learning_rate": 0.0001987065283709579, + "loss": 0.0172, + "step": 353 + }, + { + "epoch": 0.20835785756327252, + "grad_norm": 7.71875, + "learning_rate": 0.00019868212324588165, + "loss": 0.0327, + "step": 354 + }, + { + "epoch": 0.20894643908181282, + "grad_norm": 0.1376953125, + "learning_rate": 0.0001986577181208054, + "loss": 0.0153, + "step": 355 + }, + { + "epoch": 0.20953502060035314, + "grad_norm": 2.859375, + "learning_rate": 0.00019863331299572911, + "loss": 0.0693, + "step": 356 + }, + { + "epoch": 0.21012360211889347, + "grad_norm": 0.1376953125, + "learning_rate": 0.00019860890787065286, + "loss": 0.0157, + "step": 357 + }, + { + "epoch": 0.2107121836374338, + "grad_norm": 0.3046875, + "learning_rate": 0.00019858450274557658, + "loss": 0.022, + "step": 358 + }, + { + "epoch": 0.2113007651559741, + "grad_norm": 0.2490234375, + "learning_rate": 0.00019856009762050033, + "loss": 0.0205, + "step": 359 + }, + { + "epoch": 0.21188934667451442, + "grad_norm": 0.376953125, + "learning_rate": 0.00019853569249542405, + "loss": 0.0226, + "step": 360 + }, + { + "epoch": 0.21247792819305475, + "grad_norm": 0.2314453125, + "learning_rate": 0.0001985112873703478, + "loss": 0.0165, + "step": 361 + }, + { + "epoch": 0.21306650971159505, + "grad_norm": 0.134765625, + "learning_rate": 0.00019848688224527152, + "loss": 0.0166, + "step": 362 + }, + { + "epoch": 0.21365509123013537, + "grad_norm": 0.134765625, + "learning_rate": 0.00019846247712019527, + "loss": 0.0136, + "step": 363 + }, + { + "epoch": 0.2142436727486757, + "grad_norm": 0.71875, + "learning_rate": 0.00019843807199511899, + "loss": 0.0369, + "step": 364 + }, + { + "epoch": 0.214832254267216, + "grad_norm": 9.0, + "learning_rate": 0.00019841366687004273, + "loss": 0.0464, + "step": 365 + }, + { + "epoch": 0.21542083578575633, + "grad_norm": 0.361328125, + "learning_rate": 0.00019838926174496645, + "loss": 0.0149, + "step": 366 + }, + { + "epoch": 0.21600941730429665, + "grad_norm": 0.2890625, + "learning_rate": 0.0001983648566198902, + "loss": 0.0188, + "step": 367 + }, + { + "epoch": 0.21659799882283695, + "grad_norm": 0.1591796875, + "learning_rate": 0.00019834045149481392, + "loss": 0.0172, + "step": 368 + }, + { + "epoch": 0.21718658034137728, + "grad_norm": 0.1708984375, + "learning_rate": 0.00019831604636973767, + "loss": 0.0168, + "step": 369 + }, + { + "epoch": 0.2177751618599176, + "grad_norm": 0.1318359375, + "learning_rate": 0.0001982916412446614, + "loss": 0.015, + "step": 370 + }, + { + "epoch": 0.2183637433784579, + "grad_norm": 2.734375, + "learning_rate": 0.00019826723611958514, + "loss": 0.0378, + "step": 371 + }, + { + "epoch": 0.21895232489699823, + "grad_norm": 1.234375, + "learning_rate": 0.00019824283099450886, + "loss": 0.0237, + "step": 372 + }, + { + "epoch": 0.21954090641553856, + "grad_norm": 0.205078125, + "learning_rate": 0.0001982184258694326, + "loss": 0.0156, + "step": 373 + }, + { + "epoch": 0.22012948793407888, + "grad_norm": 0.376953125, + "learning_rate": 0.00019819402074435633, + "loss": 0.0204, + "step": 374 + }, + { + "epoch": 0.22071806945261918, + "grad_norm": 1.453125, + "learning_rate": 0.00019816961561928007, + "loss": 0.0189, + "step": 375 + }, + { + "epoch": 0.2213066509711595, + "grad_norm": 0.64453125, + "learning_rate": 0.0001981452104942038, + "loss": 0.0208, + "step": 376 + }, + { + "epoch": 0.22189523248969983, + "grad_norm": 0.25, + "learning_rate": 0.00019812080536912751, + "loss": 0.0203, + "step": 377 + }, + { + "epoch": 0.22248381400824013, + "grad_norm": 0.2431640625, + "learning_rate": 0.00019809640024405126, + "loss": 0.0175, + "step": 378 + }, + { + "epoch": 0.22307239552678046, + "grad_norm": 0.138671875, + "learning_rate": 0.00019807199511897498, + "loss": 0.0154, + "step": 379 + }, + { + "epoch": 0.22366097704532079, + "grad_norm": 2.328125, + "learning_rate": 0.00019804758999389873, + "loss": 0.0164, + "step": 380 + }, + { + "epoch": 0.22424955856386108, + "grad_norm": 0.53515625, + "learning_rate": 0.00019802318486882245, + "loss": 0.0175, + "step": 381 + }, + { + "epoch": 0.2248381400824014, + "grad_norm": 0.427734375, + "learning_rate": 0.0001979987797437462, + "loss": 0.0154, + "step": 382 + }, + { + "epoch": 0.22542672160094174, + "grad_norm": 0.1328125, + "learning_rate": 0.00019797437461866992, + "loss": 0.014, + "step": 383 + }, + { + "epoch": 0.22601530311948204, + "grad_norm": 0.166015625, + "learning_rate": 0.00019794996949359367, + "loss": 0.0153, + "step": 384 + }, + { + "epoch": 0.22660388463802236, + "grad_norm": 0.28515625, + "learning_rate": 0.0001979255643685174, + "loss": 0.0172, + "step": 385 + }, + { + "epoch": 0.2271924661565627, + "grad_norm": 0.0830078125, + "learning_rate": 0.00019790115924344113, + "loss": 0.0115, + "step": 386 + }, + { + "epoch": 0.22778104767510302, + "grad_norm": 0.11474609375, + "learning_rate": 0.00019787675411836485, + "loss": 0.0132, + "step": 387 + }, + { + "epoch": 0.22836962919364331, + "grad_norm": 1.4609375, + "learning_rate": 0.0001978523489932886, + "loss": 0.0234, + "step": 388 + }, + { + "epoch": 0.22895821071218364, + "grad_norm": 0.10498046875, + "learning_rate": 0.00019782794386821232, + "loss": 0.0125, + "step": 389 + }, + { + "epoch": 0.22954679223072397, + "grad_norm": 0.1328125, + "learning_rate": 0.00019780353874313607, + "loss": 0.0156, + "step": 390 + }, + { + "epoch": 0.23013537374926427, + "grad_norm": 1.15625, + "learning_rate": 0.0001977791336180598, + "loss": 0.0522, + "step": 391 + }, + { + "epoch": 0.2307239552678046, + "grad_norm": 0.154296875, + "learning_rate": 0.00019775472849298354, + "loss": 0.0151, + "step": 392 + }, + { + "epoch": 0.23131253678634492, + "grad_norm": 0.1591796875, + "learning_rate": 0.00019773032336790726, + "loss": 0.016, + "step": 393 + }, + { + "epoch": 0.23190111830488522, + "grad_norm": 1.1953125, + "learning_rate": 0.000197705918242831, + "loss": 0.0198, + "step": 394 + }, + { + "epoch": 0.23248969982342554, + "grad_norm": 0.1513671875, + "learning_rate": 0.00019768151311775473, + "loss": 0.0133, + "step": 395 + }, + { + "epoch": 0.23307828134196587, + "grad_norm": 0.1259765625, + "learning_rate": 0.00019765710799267847, + "loss": 0.0128, + "step": 396 + }, + { + "epoch": 0.23366686286050617, + "grad_norm": 0.251953125, + "learning_rate": 0.0001976327028676022, + "loss": 0.0173, + "step": 397 + }, + { + "epoch": 0.2342554443790465, + "grad_norm": 1.5625, + "learning_rate": 0.00019760829774252594, + "loss": 0.0593, + "step": 398 + }, + { + "epoch": 0.23484402589758682, + "grad_norm": 0.1376953125, + "learning_rate": 0.00019758389261744966, + "loss": 0.0148, + "step": 399 + }, + { + "epoch": 0.23543260741612712, + "grad_norm": 0.189453125, + "learning_rate": 0.0001975594874923734, + "loss": 0.0144, + "step": 400 + }, + { + "epoch": 0.23602118893466745, + "grad_norm": 0.10009765625, + "learning_rate": 0.00019753508236729713, + "loss": 0.0127, + "step": 401 + }, + { + "epoch": 0.23660977045320777, + "grad_norm": 0.12890625, + "learning_rate": 0.00019751067724222088, + "loss": 0.0137, + "step": 402 + }, + { + "epoch": 0.2371983519717481, + "grad_norm": 0.10888671875, + "learning_rate": 0.0001974862721171446, + "loss": 0.0126, + "step": 403 + }, + { + "epoch": 0.2377869334902884, + "grad_norm": 0.119140625, + "learning_rate": 0.00019746186699206835, + "loss": 0.0149, + "step": 404 + }, + { + "epoch": 0.23837551500882873, + "grad_norm": 0.318359375, + "learning_rate": 0.00019743746186699207, + "loss": 0.0168, + "step": 405 + }, + { + "epoch": 0.23896409652736905, + "grad_norm": 0.2265625, + "learning_rate": 0.00019741305674191581, + "loss": 0.0188, + "step": 406 + }, + { + "epoch": 0.23955267804590935, + "grad_norm": 0.11083984375, + "learning_rate": 0.00019738865161683954, + "loss": 0.0125, + "step": 407 + }, + { + "epoch": 0.24014125956444968, + "grad_norm": 0.84765625, + "learning_rate": 0.00019736424649176328, + "loss": 0.0361, + "step": 408 + }, + { + "epoch": 0.24072984108299, + "grad_norm": 0.181640625, + "learning_rate": 0.00019733984136668703, + "loss": 0.014, + "step": 409 + }, + { + "epoch": 0.2413184226015303, + "grad_norm": 0.1171875, + "learning_rate": 0.00019731543624161075, + "loss": 0.0139, + "step": 410 + }, + { + "epoch": 0.24190700412007063, + "grad_norm": 0.08544921875, + "learning_rate": 0.0001972910311165345, + "loss": 0.0115, + "step": 411 + }, + { + "epoch": 0.24249558563861096, + "grad_norm": 0.318359375, + "learning_rate": 0.00019726662599145822, + "loss": 0.0173, + "step": 412 + }, + { + "epoch": 0.24308416715715125, + "grad_norm": 0.08935546875, + "learning_rate": 0.00019724222086638197, + "loss": 0.012, + "step": 413 + }, + { + "epoch": 0.24367274867569158, + "grad_norm": 1.0859375, + "learning_rate": 0.0001972178157413057, + "loss": 0.0273, + "step": 414 + }, + { + "epoch": 0.2442613301942319, + "grad_norm": 0.08740234375, + "learning_rate": 0.00019719341061622943, + "loss": 0.0117, + "step": 415 + }, + { + "epoch": 0.2448499117127722, + "grad_norm": 0.0751953125, + "learning_rate": 0.00019716900549115315, + "loss": 0.0109, + "step": 416 + }, + { + "epoch": 0.24543849323131253, + "grad_norm": 0.2294921875, + "learning_rate": 0.0001971446003660769, + "loss": 0.0188, + "step": 417 + }, + { + "epoch": 0.24602707474985286, + "grad_norm": 0.12890625, + "learning_rate": 0.00019712019524100062, + "loss": 0.015, + "step": 418 + }, + { + "epoch": 0.24661565626839319, + "grad_norm": 0.138671875, + "learning_rate": 0.00019709579011592437, + "loss": 0.014, + "step": 419 + }, + { + "epoch": 0.24720423778693348, + "grad_norm": 0.10546875, + "learning_rate": 0.0001970713849908481, + "loss": 0.0106, + "step": 420 + }, + { + "epoch": 0.2477928193054738, + "grad_norm": 0.083984375, + "learning_rate": 0.00019704697986577184, + "loss": 0.0115, + "step": 421 + }, + { + "epoch": 0.24838140082401414, + "grad_norm": 0.314453125, + "learning_rate": 0.00019702257474069556, + "loss": 0.0153, + "step": 422 + }, + { + "epoch": 0.24896998234255444, + "grad_norm": 0.23046875, + "learning_rate": 0.0001969981696156193, + "loss": 0.0181, + "step": 423 + }, + { + "epoch": 0.24955856386109476, + "grad_norm": 0.1171875, + "learning_rate": 0.00019697376449054303, + "loss": 0.0134, + "step": 424 + }, + { + "epoch": 0.25014714537963506, + "grad_norm": 0.27734375, + "learning_rate": 0.00019694935936546677, + "loss": 0.019, + "step": 425 + }, + { + "epoch": 0.2507357268981754, + "grad_norm": 0.177734375, + "learning_rate": 0.0001969249542403905, + "loss": 0.0149, + "step": 426 + }, + { + "epoch": 0.2513243084167157, + "grad_norm": 1.15625, + "learning_rate": 0.00019690054911531422, + "loss": 0.0165, + "step": 427 + }, + { + "epoch": 0.251912889935256, + "grad_norm": 0.84375, + "learning_rate": 0.00019687614399023794, + "loss": 0.0131, + "step": 428 + }, + { + "epoch": 0.25250147145379637, + "grad_norm": 0.115234375, + "learning_rate": 0.00019685173886516168, + "loss": 0.0109, + "step": 429 + }, + { + "epoch": 0.25309005297233667, + "grad_norm": 0.67578125, + "learning_rate": 0.00019682733374008543, + "loss": 0.0229, + "step": 430 + }, + { + "epoch": 0.25367863449087696, + "grad_norm": 0.484375, + "learning_rate": 0.00019680292861500915, + "loss": 0.0172, + "step": 431 + }, + { + "epoch": 0.2542672160094173, + "grad_norm": 0.1484375, + "learning_rate": 0.0001967785234899329, + "loss": 0.0138, + "step": 432 + }, + { + "epoch": 0.2548557975279576, + "grad_norm": 0.220703125, + "learning_rate": 0.00019675411836485662, + "loss": 0.017, + "step": 433 + }, + { + "epoch": 0.2554443790464979, + "grad_norm": 0.11328125, + "learning_rate": 0.00019672971323978037, + "loss": 0.0125, + "step": 434 + }, + { + "epoch": 0.25603296056503827, + "grad_norm": 0.1474609375, + "learning_rate": 0.0001967053081147041, + "loss": 0.0131, + "step": 435 + }, + { + "epoch": 0.25662154208357857, + "grad_norm": 0.1533203125, + "learning_rate": 0.00019668090298962784, + "loss": 0.0147, + "step": 436 + }, + { + "epoch": 0.25721012360211887, + "grad_norm": 0.0966796875, + "learning_rate": 0.00019665649786455156, + "loss": 0.0122, + "step": 437 + }, + { + "epoch": 0.2577987051206592, + "grad_norm": 0.11474609375, + "learning_rate": 0.0001966320927394753, + "loss": 0.014, + "step": 438 + }, + { + "epoch": 0.2583872866391995, + "grad_norm": 0.06640625, + "learning_rate": 0.00019660768761439902, + "loss": 0.0097, + "step": 439 + }, + { + "epoch": 0.2589758681577399, + "grad_norm": 0.2470703125, + "learning_rate": 0.00019658328248932277, + "loss": 0.0133, + "step": 440 + }, + { + "epoch": 0.2595644496762802, + "grad_norm": 0.07421875, + "learning_rate": 0.0001965588773642465, + "loss": 0.0104, + "step": 441 + }, + { + "epoch": 0.26015303119482047, + "grad_norm": 0.08154296875, + "learning_rate": 0.00019653447223917024, + "loss": 0.0123, + "step": 442 + }, + { + "epoch": 0.2607416127133608, + "grad_norm": 0.1005859375, + "learning_rate": 0.00019651006711409396, + "loss": 0.0117, + "step": 443 + }, + { + "epoch": 0.2613301942319011, + "grad_norm": 0.55859375, + "learning_rate": 0.0001964856619890177, + "loss": 0.0182, + "step": 444 + }, + { + "epoch": 0.2619187757504414, + "grad_norm": 0.0712890625, + "learning_rate": 0.00019646125686394143, + "loss": 0.0111, + "step": 445 + }, + { + "epoch": 0.2625073572689818, + "grad_norm": 0.52734375, + "learning_rate": 0.00019643685173886518, + "loss": 0.0143, + "step": 446 + }, + { + "epoch": 0.2630959387875221, + "grad_norm": 0.150390625, + "learning_rate": 0.0001964124466137889, + "loss": 0.0293, + "step": 447 + }, + { + "epoch": 0.2636845203060624, + "grad_norm": 0.2333984375, + "learning_rate": 0.00019638804148871264, + "loss": 0.0152, + "step": 448 + }, + { + "epoch": 0.26427310182460273, + "grad_norm": 0.166015625, + "learning_rate": 0.00019636363636363636, + "loss": 0.0135, + "step": 449 + }, + { + "epoch": 0.26486168334314303, + "grad_norm": 0.08447265625, + "learning_rate": 0.0001963392312385601, + "loss": 0.011, + "step": 450 + }, + { + "epoch": 0.2654502648616833, + "grad_norm": 0.359375, + "learning_rate": 0.00019631482611348383, + "loss": 0.0122, + "step": 451 + }, + { + "epoch": 0.2660388463802237, + "grad_norm": 0.076171875, + "learning_rate": 0.00019629042098840758, + "loss": 0.0107, + "step": 452 + }, + { + "epoch": 0.266627427898764, + "grad_norm": 0.53125, + "learning_rate": 0.0001962660158633313, + "loss": 0.0151, + "step": 453 + }, + { + "epoch": 0.2672160094173043, + "grad_norm": 0.359375, + "learning_rate": 0.00019624161073825505, + "loss": 0.016, + "step": 454 + }, + { + "epoch": 0.26780459093584463, + "grad_norm": 0.25, + "learning_rate": 0.00019621720561317877, + "loss": 0.0182, + "step": 455 + }, + { + "epoch": 0.26839317245438493, + "grad_norm": 1.0390625, + "learning_rate": 0.00019619280048810252, + "loss": 0.0123, + "step": 456 + }, + { + "epoch": 0.26898175397292523, + "grad_norm": 0.099609375, + "learning_rate": 0.00019616839536302624, + "loss": 0.013, + "step": 457 + }, + { + "epoch": 0.2695703354914656, + "grad_norm": 1.0625, + "learning_rate": 0.00019614399023794998, + "loss": 0.0222, + "step": 458 + }, + { + "epoch": 0.2701589170100059, + "grad_norm": 0.2109375, + "learning_rate": 0.0001961195851128737, + "loss": 0.0125, + "step": 459 + }, + { + "epoch": 0.2707474985285462, + "grad_norm": 0.15625, + "learning_rate": 0.00019609517998779745, + "loss": 0.013, + "step": 460 + }, + { + "epoch": 0.27133608004708654, + "grad_norm": 0.10498046875, + "learning_rate": 0.00019607077486272117, + "loss": 0.0107, + "step": 461 + }, + { + "epoch": 0.27192466156562684, + "grad_norm": 1.0703125, + "learning_rate": 0.00019604636973764492, + "loss": 0.0153, + "step": 462 + }, + { + "epoch": 0.27251324308416713, + "grad_norm": 0.173828125, + "learning_rate": 0.00019602196461256867, + "loss": 0.0139, + "step": 463 + }, + { + "epoch": 0.2731018246027075, + "grad_norm": 1.390625, + "learning_rate": 0.0001959975594874924, + "loss": 0.019, + "step": 464 + }, + { + "epoch": 0.2736904061212478, + "grad_norm": 0.1474609375, + "learning_rate": 0.00019597315436241613, + "loss": 0.0128, + "step": 465 + }, + { + "epoch": 0.2742789876397881, + "grad_norm": 0.11962890625, + "learning_rate": 0.00019594874923733986, + "loss": 0.012, + "step": 466 + }, + { + "epoch": 0.27486756915832844, + "grad_norm": 0.248046875, + "learning_rate": 0.0001959243441122636, + "loss": 0.0151, + "step": 467 + }, + { + "epoch": 0.27545615067686874, + "grad_norm": 6.1875, + "learning_rate": 0.00019589993898718732, + "loss": 0.0189, + "step": 468 + }, + { + "epoch": 0.27604473219540904, + "grad_norm": 0.08349609375, + "learning_rate": 0.00019587553386211107, + "loss": 0.0112, + "step": 469 + }, + { + "epoch": 0.2766333137139494, + "grad_norm": 0.1181640625, + "learning_rate": 0.0001958511287370348, + "loss": 0.0143, + "step": 470 + }, + { + "epoch": 0.2772218952324897, + "grad_norm": 0.388671875, + "learning_rate": 0.00019582672361195854, + "loss": 0.0141, + "step": 471 + }, + { + "epoch": 0.27781047675103004, + "grad_norm": 0.08837890625, + "learning_rate": 0.00019580231848688226, + "loss": 0.011, + "step": 472 + }, + { + "epoch": 0.27839905826957034, + "grad_norm": 0.07373046875, + "learning_rate": 0.000195777913361806, + "loss": 0.0106, + "step": 473 + }, + { + "epoch": 0.27898763978811064, + "grad_norm": 0.2373046875, + "learning_rate": 0.00019575350823672973, + "loss": 0.0139, + "step": 474 + }, + { + "epoch": 0.279576221306651, + "grad_norm": 0.1025390625, + "learning_rate": 0.00019572910311165348, + "loss": 0.0119, + "step": 475 + }, + { + "epoch": 0.2801648028251913, + "grad_norm": 0.09814453125, + "learning_rate": 0.0001957046979865772, + "loss": 0.0122, + "step": 476 + }, + { + "epoch": 0.2807533843437316, + "grad_norm": 0.09619140625, + "learning_rate": 0.00019568029286150094, + "loss": 0.0113, + "step": 477 + }, + { + "epoch": 0.28134196586227195, + "grad_norm": 0.08203125, + "learning_rate": 0.00019565588773642464, + "loss": 0.0121, + "step": 478 + }, + { + "epoch": 0.28193054738081225, + "grad_norm": 0.59375, + "learning_rate": 0.00019563148261134838, + "loss": 0.013, + "step": 479 + }, + { + "epoch": 0.28251912889935255, + "grad_norm": 0.2734375, + "learning_rate": 0.0001956070774862721, + "loss": 0.0159, + "step": 480 + }, + { + "epoch": 0.2831077104178929, + "grad_norm": 0.1201171875, + "learning_rate": 0.00019558267236119585, + "loss": 0.0125, + "step": 481 + }, + { + "epoch": 0.2836962919364332, + "grad_norm": 0.890625, + "learning_rate": 0.00019555826723611957, + "loss": 0.0294, + "step": 482 + }, + { + "epoch": 0.2842848734549735, + "grad_norm": 0.08740234375, + "learning_rate": 0.00019553386211104332, + "loss": 0.011, + "step": 483 + }, + { + "epoch": 0.28487345497351385, + "grad_norm": 0.39453125, + "learning_rate": 0.00019550945698596707, + "loss": 0.0182, + "step": 484 + }, + { + "epoch": 0.28546203649205415, + "grad_norm": 0.52734375, + "learning_rate": 0.0001954850518608908, + "loss": 0.0165, + "step": 485 + }, + { + "epoch": 0.28605061801059445, + "grad_norm": 0.15625, + "learning_rate": 0.00019546064673581454, + "loss": 0.03, + "step": 486 + }, + { + "epoch": 0.2866391995291348, + "grad_norm": 0.1298828125, + "learning_rate": 0.00019543624161073826, + "loss": 0.0118, + "step": 487 + }, + { + "epoch": 0.2872277810476751, + "grad_norm": 0.10888671875, + "learning_rate": 0.000195411836485662, + "loss": 0.0121, + "step": 488 + }, + { + "epoch": 0.2878163625662154, + "grad_norm": 1.890625, + "learning_rate": 0.00019538743136058572, + "loss": 0.0276, + "step": 489 + }, + { + "epoch": 0.28840494408475575, + "grad_norm": 0.11962890625, + "learning_rate": 0.00019536302623550947, + "loss": 0.0146, + "step": 490 + }, + { + "epoch": 0.28899352560329605, + "grad_norm": 0.1044921875, + "learning_rate": 0.0001953386211104332, + "loss": 0.0118, + "step": 491 + }, + { + "epoch": 0.28958210712183635, + "grad_norm": 0.0888671875, + "learning_rate": 0.00019531421598535694, + "loss": 0.0112, + "step": 492 + }, + { + "epoch": 0.2901706886403767, + "grad_norm": 0.2333984375, + "learning_rate": 0.00019528981086028066, + "loss": 0.0107, + "step": 493 + }, + { + "epoch": 0.290759270158917, + "grad_norm": 0.09033203125, + "learning_rate": 0.0001952654057352044, + "loss": 0.0112, + "step": 494 + }, + { + "epoch": 0.2913478516774573, + "grad_norm": 0.08837890625, + "learning_rate": 0.00019524100061012813, + "loss": 0.0117, + "step": 495 + }, + { + "epoch": 0.29193643319599766, + "grad_norm": 0.0966796875, + "learning_rate": 0.00019521659548505188, + "loss": 0.0117, + "step": 496 + }, + { + "epoch": 0.29252501471453796, + "grad_norm": 1.34375, + "learning_rate": 0.0001951921903599756, + "loss": 0.0131, + "step": 497 + }, + { + "epoch": 0.29311359623307826, + "grad_norm": 0.171875, + "learning_rate": 0.00019516778523489934, + "loss": 0.0114, + "step": 498 + }, + { + "epoch": 0.2937021777516186, + "grad_norm": 0.1875, + "learning_rate": 0.00019514338010982306, + "loss": 0.0117, + "step": 499 + }, + { + "epoch": 0.2942907592701589, + "grad_norm": 0.08203125, + "learning_rate": 0.0001951189749847468, + "loss": 0.0114, + "step": 500 + }, + { + "epoch": 0.29487934078869926, + "grad_norm": 0.10205078125, + "learning_rate": 0.00019509456985967053, + "loss": 0.012, + "step": 501 + }, + { + "epoch": 0.29546792230723956, + "grad_norm": 0.240234375, + "learning_rate": 0.00019507016473459428, + "loss": 0.014, + "step": 502 + }, + { + "epoch": 0.29605650382577986, + "grad_norm": 0.24609375, + "learning_rate": 0.000195045759609518, + "loss": 0.0193, + "step": 503 + }, + { + "epoch": 0.2966450853443202, + "grad_norm": 0.06396484375, + "learning_rate": 0.00019502135448444175, + "loss": 0.0114, + "step": 504 + }, + { + "epoch": 0.2972336668628605, + "grad_norm": 0.08740234375, + "learning_rate": 0.00019499694935936547, + "loss": 0.0111, + "step": 505 + }, + { + "epoch": 0.2978222483814008, + "grad_norm": 0.11767578125, + "learning_rate": 0.00019497254423428922, + "loss": 0.0135, + "step": 506 + }, + { + "epoch": 0.29841082989994117, + "grad_norm": 0.1640625, + "learning_rate": 0.00019494813910921294, + "loss": 0.0125, + "step": 507 + }, + { + "epoch": 0.29899941141848146, + "grad_norm": 8.3125, + "learning_rate": 0.00019492373398413668, + "loss": 0.0181, + "step": 508 + }, + { + "epoch": 0.29958799293702176, + "grad_norm": 0.083984375, + "learning_rate": 0.0001948993288590604, + "loss": 0.011, + "step": 509 + }, + { + "epoch": 0.3001765744555621, + "grad_norm": 0.1904296875, + "learning_rate": 0.00019487492373398415, + "loss": 0.0129, + "step": 510 + }, + { + "epoch": 0.3007651559741024, + "grad_norm": 0.09619140625, + "learning_rate": 0.00019485051860890787, + "loss": 0.0108, + "step": 511 + }, + { + "epoch": 0.3013537374926427, + "grad_norm": 0.1005859375, + "learning_rate": 0.00019482611348383162, + "loss": 0.0129, + "step": 512 + }, + { + "epoch": 0.30194231901118307, + "grad_norm": 0.08447265625, + "learning_rate": 0.00019480170835875534, + "loss": 0.0109, + "step": 513 + }, + { + "epoch": 0.30253090052972337, + "grad_norm": 0.055419921875, + "learning_rate": 0.0001947773032336791, + "loss": 0.0097, + "step": 514 + }, + { + "epoch": 0.30311948204826367, + "grad_norm": 0.09716796875, + "learning_rate": 0.0001947528981086028, + "loss": 0.0133, + "step": 515 + }, + { + "epoch": 0.303708063566804, + "grad_norm": 2.546875, + "learning_rate": 0.00019472849298352656, + "loss": 0.0112, + "step": 516 + }, + { + "epoch": 0.3042966450853443, + "grad_norm": 0.1318359375, + "learning_rate": 0.0001947040878584503, + "loss": 0.0129, + "step": 517 + }, + { + "epoch": 0.3048852266038846, + "grad_norm": 0.294921875, + "learning_rate": 0.00019467968273337402, + "loss": 0.0147, + "step": 518 + }, + { + "epoch": 0.305473808122425, + "grad_norm": 0.08349609375, + "learning_rate": 0.00019465527760829777, + "loss": 0.0116, + "step": 519 + }, + { + "epoch": 0.30606238964096527, + "grad_norm": 0.08837890625, + "learning_rate": 0.0001946308724832215, + "loss": 0.0106, + "step": 520 + }, + { + "epoch": 0.30665097115950557, + "grad_norm": 0.0908203125, + "learning_rate": 0.00019460646735814524, + "loss": 0.0113, + "step": 521 + }, + { + "epoch": 0.3072395526780459, + "grad_norm": 0.400390625, + "learning_rate": 0.00019458206223306896, + "loss": 0.0187, + "step": 522 + }, + { + "epoch": 0.3078281341965862, + "grad_norm": 0.48828125, + "learning_rate": 0.0001945576571079927, + "loss": 0.0226, + "step": 523 + }, + { + "epoch": 0.3084167157151265, + "grad_norm": 0.1328125, + "learning_rate": 0.00019453325198291643, + "loss": 0.0149, + "step": 524 + }, + { + "epoch": 0.3090052972336669, + "grad_norm": 0.08203125, + "learning_rate": 0.00019450884685784018, + "loss": 0.0112, + "step": 525 + }, + { + "epoch": 0.3095938787522072, + "grad_norm": 0.054443359375, + "learning_rate": 0.0001944844417327639, + "loss": 0.0103, + "step": 526 + }, + { + "epoch": 0.3101824602707475, + "grad_norm": 0.609375, + "learning_rate": 0.00019446003660768764, + "loss": 0.0234, + "step": 527 + }, + { + "epoch": 0.3107710417892878, + "grad_norm": 0.267578125, + "learning_rate": 0.00019443563148261134, + "loss": 0.0108, + "step": 528 + }, + { + "epoch": 0.3113596233078281, + "grad_norm": 0.2216796875, + "learning_rate": 0.00019441122635753508, + "loss": 0.0138, + "step": 529 + }, + { + "epoch": 0.3119482048263684, + "grad_norm": 0.125, + "learning_rate": 0.0001943868212324588, + "loss": 0.0107, + "step": 530 + }, + { + "epoch": 0.3125367863449088, + "grad_norm": 0.1357421875, + "learning_rate": 0.00019436241610738255, + "loss": 0.0128, + "step": 531 + }, + { + "epoch": 0.3131253678634491, + "grad_norm": 0.10986328125, + "learning_rate": 0.00019433801098230627, + "loss": 0.0118, + "step": 532 + }, + { + "epoch": 0.31371394938198943, + "grad_norm": 0.09326171875, + "learning_rate": 0.00019431360585723002, + "loss": 0.0113, + "step": 533 + }, + { + "epoch": 0.31430253090052973, + "grad_norm": 0.125, + "learning_rate": 0.00019428920073215374, + "loss": 0.0124, + "step": 534 + }, + { + "epoch": 0.31489111241907003, + "grad_norm": 0.142578125, + "learning_rate": 0.0001942647956070775, + "loss": 0.0123, + "step": 535 + }, + { + "epoch": 0.3154796939376104, + "grad_norm": 0.07275390625, + "learning_rate": 0.0001942403904820012, + "loss": 0.0104, + "step": 536 + }, + { + "epoch": 0.3160682754561507, + "grad_norm": 0.10009765625, + "learning_rate": 0.00019421598535692496, + "loss": 0.012, + "step": 537 + }, + { + "epoch": 0.316656856974691, + "grad_norm": 0.279296875, + "learning_rate": 0.0001941915802318487, + "loss": 0.0141, + "step": 538 + }, + { + "epoch": 0.31724543849323134, + "grad_norm": 0.72265625, + "learning_rate": 0.00019416717510677242, + "loss": 0.0271, + "step": 539 + }, + { + "epoch": 0.31783402001177163, + "grad_norm": 0.08203125, + "learning_rate": 0.00019414276998169617, + "loss": 0.0114, + "step": 540 + }, + { + "epoch": 0.31842260153031193, + "grad_norm": 1.4765625, + "learning_rate": 0.0001941183648566199, + "loss": 0.0378, + "step": 541 + }, + { + "epoch": 0.3190111830488523, + "grad_norm": 0.2578125, + "learning_rate": 0.00019409395973154364, + "loss": 0.0176, + "step": 542 + }, + { + "epoch": 0.3195997645673926, + "grad_norm": 0.2158203125, + "learning_rate": 0.00019406955460646736, + "loss": 0.0148, + "step": 543 + }, + { + "epoch": 0.3201883460859329, + "grad_norm": 0.251953125, + "learning_rate": 0.0001940451494813911, + "loss": 0.0144, + "step": 544 + }, + { + "epoch": 0.32077692760447324, + "grad_norm": 0.11669921875, + "learning_rate": 0.00019402074435631483, + "loss": 0.0121, + "step": 545 + }, + { + "epoch": 0.32136550912301354, + "grad_norm": 0.130859375, + "learning_rate": 0.00019399633923123858, + "loss": 0.0131, + "step": 546 + }, + { + "epoch": 0.32195409064155384, + "grad_norm": 0.10107421875, + "learning_rate": 0.0001939719341061623, + "loss": 0.0115, + "step": 547 + }, + { + "epoch": 0.3225426721600942, + "grad_norm": 0.1806640625, + "learning_rate": 0.00019394752898108604, + "loss": 0.0135, + "step": 548 + }, + { + "epoch": 0.3231312536786345, + "grad_norm": 0.62890625, + "learning_rate": 0.00019392312385600976, + "loss": 0.0154, + "step": 549 + }, + { + "epoch": 0.3237198351971748, + "grad_norm": 0.08935546875, + "learning_rate": 0.0001938987187309335, + "loss": 0.0112, + "step": 550 + }, + { + "epoch": 0.32430841671571514, + "grad_norm": 0.0908203125, + "learning_rate": 0.00019387431360585723, + "loss": 0.0118, + "step": 551 + }, + { + "epoch": 0.32489699823425544, + "grad_norm": 17.375, + "learning_rate": 0.00019384990848078098, + "loss": 0.0457, + "step": 552 + }, + { + "epoch": 0.32548557975279574, + "grad_norm": 4.5625, + "learning_rate": 0.0001938255033557047, + "loss": 0.0131, + "step": 553 + }, + { + "epoch": 0.3260741612713361, + "grad_norm": 0.15234375, + "learning_rate": 0.00019380109823062845, + "loss": 0.0132, + "step": 554 + }, + { + "epoch": 0.3266627427898764, + "grad_norm": 0.220703125, + "learning_rate": 0.00019377669310555217, + "loss": 0.0148, + "step": 555 + }, + { + "epoch": 0.3272513243084167, + "grad_norm": 0.08251953125, + "learning_rate": 0.00019375228798047592, + "loss": 0.0113, + "step": 556 + }, + { + "epoch": 0.32783990582695705, + "grad_norm": 0.52734375, + "learning_rate": 0.00019372788285539964, + "loss": 0.012, + "step": 557 + }, + { + "epoch": 0.32842848734549734, + "grad_norm": 0.083984375, + "learning_rate": 0.00019370347773032338, + "loss": 0.0106, + "step": 558 + }, + { + "epoch": 0.32901706886403764, + "grad_norm": 0.2021484375, + "learning_rate": 0.0001936790726052471, + "loss": 0.0145, + "step": 559 + }, + { + "epoch": 0.329605650382578, + "grad_norm": 0.0859375, + "learning_rate": 0.00019365466748017085, + "loss": 0.0099, + "step": 560 + }, + { + "epoch": 0.3301942319011183, + "grad_norm": 0.06640625, + "learning_rate": 0.00019363026235509457, + "loss": 0.0114, + "step": 561 + }, + { + "epoch": 0.33078281341965865, + "grad_norm": 0.1982421875, + "learning_rate": 0.00019360585723001832, + "loss": 0.015, + "step": 562 + }, + { + "epoch": 0.33137139493819895, + "grad_norm": 0.0810546875, + "learning_rate": 0.00019358145210494204, + "loss": 0.0107, + "step": 563 + }, + { + "epoch": 0.33195997645673925, + "grad_norm": 0.0625, + "learning_rate": 0.0001935570469798658, + "loss": 0.0106, + "step": 564 + }, + { + "epoch": 0.3325485579752796, + "grad_norm": 0.056396484375, + "learning_rate": 0.0001935326418547895, + "loss": 0.0087, + "step": 565 + }, + { + "epoch": 0.3331371394938199, + "grad_norm": 0.08642578125, + "learning_rate": 0.00019350823672971326, + "loss": 0.0125, + "step": 566 + }, + { + "epoch": 0.3337257210123602, + "grad_norm": 0.23828125, + "learning_rate": 0.00019348383160463698, + "loss": 0.0127, + "step": 567 + }, + { + "epoch": 0.33431430253090055, + "grad_norm": 0.5390625, + "learning_rate": 0.00019345942647956072, + "loss": 0.0187, + "step": 568 + }, + { + "epoch": 0.33490288404944085, + "grad_norm": 1.015625, + "learning_rate": 0.00019343502135448445, + "loss": 0.0132, + "step": 569 + }, + { + "epoch": 0.33549146556798115, + "grad_norm": 0.10302734375, + "learning_rate": 0.0001934106162294082, + "loss": 0.0126, + "step": 570 + }, + { + "epoch": 0.3360800470865215, + "grad_norm": 0.126953125, + "learning_rate": 0.00019338621110433194, + "loss": 0.0192, + "step": 571 + }, + { + "epoch": 0.3366686286050618, + "grad_norm": 0.087890625, + "learning_rate": 0.00019336180597925566, + "loss": 0.0114, + "step": 572 + }, + { + "epoch": 0.3372572101236021, + "grad_norm": 0.16796875, + "learning_rate": 0.0001933374008541794, + "loss": 0.014, + "step": 573 + }, + { + "epoch": 0.33784579164214246, + "grad_norm": 0.076171875, + "learning_rate": 0.00019331299572910313, + "loss": 0.0106, + "step": 574 + }, + { + "epoch": 0.33843437316068276, + "grad_norm": 0.0693359375, + "learning_rate": 0.00019328859060402688, + "loss": 0.0105, + "step": 575 + }, + { + "epoch": 0.33902295467922305, + "grad_norm": 0.07080078125, + "learning_rate": 0.0001932641854789506, + "loss": 0.0108, + "step": 576 + }, + { + "epoch": 0.3396115361977634, + "grad_norm": 0.059326171875, + "learning_rate": 0.00019323978035387434, + "loss": 0.0095, + "step": 577 + }, + { + "epoch": 0.3402001177163037, + "grad_norm": 0.0615234375, + "learning_rate": 0.00019321537522879806, + "loss": 0.0102, + "step": 578 + }, + { + "epoch": 0.340788699234844, + "grad_norm": 0.0634765625, + "learning_rate": 0.00019319097010372179, + "loss": 0.0098, + "step": 579 + }, + { + "epoch": 0.34137728075338436, + "grad_norm": 3.6875, + "learning_rate": 0.0001931665649786455, + "loss": 0.0378, + "step": 580 + }, + { + "epoch": 0.34196586227192466, + "grad_norm": 0.173828125, + "learning_rate": 0.00019314215985356925, + "loss": 0.0141, + "step": 581 + }, + { + "epoch": 0.34255444379046496, + "grad_norm": 0.09033203125, + "learning_rate": 0.00019311775472849297, + "loss": 0.0107, + "step": 582 + }, + { + "epoch": 0.3431430253090053, + "grad_norm": 4.40625, + "learning_rate": 0.00019309334960341672, + "loss": 0.0598, + "step": 583 + }, + { + "epoch": 0.3437316068275456, + "grad_norm": 0.6328125, + "learning_rate": 0.00019306894447834044, + "loss": 0.032, + "step": 584 + }, + { + "epoch": 0.3443201883460859, + "grad_norm": 0.11669921875, + "learning_rate": 0.0001930445393532642, + "loss": 0.0117, + "step": 585 + }, + { + "epoch": 0.34490876986462626, + "grad_norm": 0.064453125, + "learning_rate": 0.0001930201342281879, + "loss": 0.0093, + "step": 586 + }, + { + "epoch": 0.34549735138316656, + "grad_norm": 0.07373046875, + "learning_rate": 0.00019299572910311166, + "loss": 0.0102, + "step": 587 + }, + { + "epoch": 0.34608593290170686, + "grad_norm": 0.060302734375, + "learning_rate": 0.00019297132397803538, + "loss": 0.0096, + "step": 588 + }, + { + "epoch": 0.3466745144202472, + "grad_norm": 0.73046875, + "learning_rate": 0.00019294691885295913, + "loss": 0.0176, + "step": 589 + }, + { + "epoch": 0.3472630959387875, + "grad_norm": 0.07763671875, + "learning_rate": 0.00019292251372788285, + "loss": 0.0107, + "step": 590 + }, + { + "epoch": 0.3478516774573278, + "grad_norm": 0.06396484375, + "learning_rate": 0.0001928981086028066, + "loss": 0.0096, + "step": 591 + }, + { + "epoch": 0.34844025897586817, + "grad_norm": 0.078125, + "learning_rate": 0.00019287370347773034, + "loss": 0.0264, + "step": 592 + }, + { + "epoch": 0.34902884049440847, + "grad_norm": 0.275390625, + "learning_rate": 0.00019284929835265406, + "loss": 0.0139, + "step": 593 + }, + { + "epoch": 0.3496174220129488, + "grad_norm": 0.064453125, + "learning_rate": 0.0001928248932275778, + "loss": 0.0101, + "step": 594 + }, + { + "epoch": 0.3502060035314891, + "grad_norm": 0.0634765625, + "learning_rate": 0.00019280048810250153, + "loss": 0.01, + "step": 595 + }, + { + "epoch": 0.3507945850500294, + "grad_norm": 0.0771484375, + "learning_rate": 0.00019277608297742528, + "loss": 0.0106, + "step": 596 + }, + { + "epoch": 0.35138316656856977, + "grad_norm": 0.08056640625, + "learning_rate": 0.000192751677852349, + "loss": 0.0106, + "step": 597 + }, + { + "epoch": 0.35197174808711007, + "grad_norm": 0.07177734375, + "learning_rate": 0.00019272727272727274, + "loss": 0.0109, + "step": 598 + }, + { + "epoch": 0.35256032960565037, + "grad_norm": 0.052001953125, + "learning_rate": 0.00019270286760219647, + "loss": 0.0094, + "step": 599 + }, + { + "epoch": 0.3531489111241907, + "grad_norm": 0.08203125, + "learning_rate": 0.0001926784624771202, + "loss": 0.0096, + "step": 600 + }, + { + "epoch": 0.353737492642731, + "grad_norm": 0.0810546875, + "learning_rate": 0.00019265405735204393, + "loss": 0.0115, + "step": 601 + }, + { + "epoch": 0.3543260741612713, + "grad_norm": 0.1123046875, + "learning_rate": 0.00019262965222696768, + "loss": 0.0115, + "step": 602 + }, + { + "epoch": 0.3549146556798117, + "grad_norm": 0.060302734375, + "learning_rate": 0.0001926052471018914, + "loss": 0.0095, + "step": 603 + }, + { + "epoch": 0.355503237198352, + "grad_norm": 0.10693359375, + "learning_rate": 0.00019258084197681515, + "loss": 0.0121, + "step": 604 + }, + { + "epoch": 0.35609181871689227, + "grad_norm": 0.1259765625, + "learning_rate": 0.00019255643685173887, + "loss": 0.0129, + "step": 605 + }, + { + "epoch": 0.3566804002354326, + "grad_norm": 0.062255859375, + "learning_rate": 0.00019253203172666262, + "loss": 0.0099, + "step": 606 + }, + { + "epoch": 0.3572689817539729, + "grad_norm": 0.0908203125, + "learning_rate": 0.00019250762660158634, + "loss": 0.0108, + "step": 607 + }, + { + "epoch": 0.3578575632725132, + "grad_norm": 0.061767578125, + "learning_rate": 0.00019248322147651008, + "loss": 0.0094, + "step": 608 + }, + { + "epoch": 0.3584461447910536, + "grad_norm": 0.11572265625, + "learning_rate": 0.0001924588163514338, + "loss": 0.0097, + "step": 609 + }, + { + "epoch": 0.3590347263095939, + "grad_norm": 0.083984375, + "learning_rate": 0.00019243441122635755, + "loss": 0.0121, + "step": 610 + }, + { + "epoch": 0.3596233078281342, + "grad_norm": 0.10498046875, + "learning_rate": 0.00019241000610128127, + "loss": 0.012, + "step": 611 + }, + { + "epoch": 0.36021188934667453, + "grad_norm": 0.068359375, + "learning_rate": 0.00019238560097620502, + "loss": 0.0091, + "step": 612 + }, + { + "epoch": 0.36080047086521483, + "grad_norm": 0.060791015625, + "learning_rate": 0.00019236119585112874, + "loss": 0.0087, + "step": 613 + }, + { + "epoch": 0.3613890523837551, + "grad_norm": 0.061767578125, + "learning_rate": 0.0001923367907260525, + "loss": 0.009, + "step": 614 + }, + { + "epoch": 0.3619776339022955, + "grad_norm": 0.11083984375, + "learning_rate": 0.0001923123856009762, + "loss": 0.0115, + "step": 615 + }, + { + "epoch": 0.3625662154208358, + "grad_norm": 0.06884765625, + "learning_rate": 0.00019228798047589996, + "loss": 0.0102, + "step": 616 + }, + { + "epoch": 0.3631547969393761, + "grad_norm": 0.07373046875, + "learning_rate": 0.00019226357535082368, + "loss": 0.0109, + "step": 617 + }, + { + "epoch": 0.36374337845791643, + "grad_norm": 0.0859375, + "learning_rate": 0.00019223917022574743, + "loss": 0.0108, + "step": 618 + }, + { + "epoch": 0.36433195997645673, + "grad_norm": 0.054443359375, + "learning_rate": 0.00019221476510067115, + "loss": 0.0091, + "step": 619 + }, + { + "epoch": 0.36492054149499703, + "grad_norm": 0.1806640625, + "learning_rate": 0.0001921903599755949, + "loss": 0.0134, + "step": 620 + }, + { + "epoch": 0.3655091230135374, + "grad_norm": 0.057861328125, + "learning_rate": 0.00019216595485051861, + "loss": 0.0092, + "step": 621 + }, + { + "epoch": 0.3660977045320777, + "grad_norm": 0.07080078125, + "learning_rate": 0.00019214154972544236, + "loss": 0.011, + "step": 622 + }, + { + "epoch": 0.36668628605061804, + "grad_norm": 0.07080078125, + "learning_rate": 0.00019211714460036608, + "loss": 0.0095, + "step": 623 + }, + { + "epoch": 0.36727486756915834, + "grad_norm": 0.1748046875, + "learning_rate": 0.00019209273947528983, + "loss": 0.0114, + "step": 624 + }, + { + "epoch": 0.36786344908769864, + "grad_norm": 0.56640625, + "learning_rate": 0.00019206833435021358, + "loss": 0.0288, + "step": 625 + }, + { + "epoch": 0.368452030606239, + "grad_norm": 0.208984375, + "learning_rate": 0.0001920439292251373, + "loss": 0.0131, + "step": 626 + }, + { + "epoch": 0.3690406121247793, + "grad_norm": 0.07470703125, + "learning_rate": 0.00019201952410006104, + "loss": 0.0103, + "step": 627 + }, + { + "epoch": 0.3696291936433196, + "grad_norm": 0.09716796875, + "learning_rate": 0.00019199511897498477, + "loss": 0.0109, + "step": 628 + }, + { + "epoch": 0.37021777516185994, + "grad_norm": 0.0712890625, + "learning_rate": 0.00019197071384990849, + "loss": 0.0093, + "step": 629 + }, + { + "epoch": 0.37080635668040024, + "grad_norm": 0.09912109375, + "learning_rate": 0.0001919463087248322, + "loss": 0.0117, + "step": 630 + }, + { + "epoch": 0.37139493819894054, + "grad_norm": 0.0673828125, + "learning_rate": 0.00019192190359975595, + "loss": 0.0089, + "step": 631 + }, + { + "epoch": 0.3719835197174809, + "grad_norm": 0.16796875, + "learning_rate": 0.00019189749847467967, + "loss": 0.013, + "step": 632 + }, + { + "epoch": 0.3725721012360212, + "grad_norm": 0.064453125, + "learning_rate": 0.00019187309334960342, + "loss": 0.0095, + "step": 633 + }, + { + "epoch": 0.3731606827545615, + "grad_norm": 0.099609375, + "learning_rate": 0.00019184868822452714, + "loss": 0.0101, + "step": 634 + }, + { + "epoch": 0.37374926427310184, + "grad_norm": 0.0673828125, + "learning_rate": 0.0001918242830994509, + "loss": 0.0096, + "step": 635 + }, + { + "epoch": 0.37433784579164214, + "grad_norm": 0.08935546875, + "learning_rate": 0.0001917998779743746, + "loss": 0.0106, + "step": 636 + }, + { + "epoch": 0.37492642731018244, + "grad_norm": 0.068359375, + "learning_rate": 0.00019177547284929836, + "loss": 0.0098, + "step": 637 + }, + { + "epoch": 0.3755150088287228, + "grad_norm": 0.11328125, + "learning_rate": 0.00019175106772422208, + "loss": 0.0118, + "step": 638 + }, + { + "epoch": 0.3761035903472631, + "grad_norm": 0.06689453125, + "learning_rate": 0.00019172666259914583, + "loss": 0.0096, + "step": 639 + }, + { + "epoch": 0.3766921718658034, + "grad_norm": 0.1591796875, + "learning_rate": 0.00019170225747406955, + "loss": 0.0153, + "step": 640 + }, + { + "epoch": 0.37728075338434375, + "grad_norm": 0.06884765625, + "learning_rate": 0.0001916778523489933, + "loss": 0.0093, + "step": 641 + }, + { + "epoch": 0.37786933490288405, + "grad_norm": 0.1396484375, + "learning_rate": 0.00019165344722391701, + "loss": 0.0118, + "step": 642 + }, + { + "epoch": 0.37845791642142435, + "grad_norm": 1.015625, + "learning_rate": 0.00019162904209884076, + "loss": 0.0164, + "step": 643 + }, + { + "epoch": 0.3790464979399647, + "grad_norm": 0.1552734375, + "learning_rate": 0.00019160463697376448, + "loss": 0.0121, + "step": 644 + }, + { + "epoch": 0.379635079458505, + "grad_norm": 0.2431640625, + "learning_rate": 0.00019158023184868823, + "loss": 0.0167, + "step": 645 + }, + { + "epoch": 0.3802236609770453, + "grad_norm": 0.05322265625, + "learning_rate": 0.00019155582672361198, + "loss": 0.0091, + "step": 646 + }, + { + "epoch": 0.38081224249558565, + "grad_norm": 0.061767578125, + "learning_rate": 0.0001915314215985357, + "loss": 0.0086, + "step": 647 + }, + { + "epoch": 0.38140082401412595, + "grad_norm": 0.3828125, + "learning_rate": 0.00019150701647345945, + "loss": 0.0157, + "step": 648 + }, + { + "epoch": 0.38198940553266625, + "grad_norm": 1.4296875, + "learning_rate": 0.00019148261134838317, + "loss": 0.0136, + "step": 649 + }, + { + "epoch": 0.3825779870512066, + "grad_norm": 1.40625, + "learning_rate": 0.0001914582062233069, + "loss": 0.026, + "step": 650 + }, + { + "epoch": 0.3831665685697469, + "grad_norm": 0.06396484375, + "learning_rate": 0.00019143380109823063, + "loss": 0.0096, + "step": 651 + }, + { + "epoch": 0.3837551500882872, + "grad_norm": 0.06982421875, + "learning_rate": 0.00019140939597315438, + "loss": 0.0103, + "step": 652 + }, + { + "epoch": 0.38434373160682755, + "grad_norm": 11.625, + "learning_rate": 0.0001913849908480781, + "loss": 0.0148, + "step": 653 + }, + { + "epoch": 0.38493231312536785, + "grad_norm": 0.08837890625, + "learning_rate": 0.00019136058572300185, + "loss": 0.0103, + "step": 654 + }, + { + "epoch": 0.3855208946439082, + "grad_norm": 0.083984375, + "learning_rate": 0.00019133618059792557, + "loss": 0.0115, + "step": 655 + }, + { + "epoch": 0.3861094761624485, + "grad_norm": 3.09375, + "learning_rate": 0.00019131177547284932, + "loss": 0.0154, + "step": 656 + }, + { + "epoch": 0.3866980576809888, + "grad_norm": 0.2041015625, + "learning_rate": 0.00019128737034777304, + "loss": 0.0125, + "step": 657 + }, + { + "epoch": 0.38728663919952916, + "grad_norm": 0.053955078125, + "learning_rate": 0.00019126296522269679, + "loss": 0.0089, + "step": 658 + }, + { + "epoch": 0.38787522071806946, + "grad_norm": 0.08349609375, + "learning_rate": 0.0001912385600976205, + "loss": 0.0107, + "step": 659 + }, + { + "epoch": 0.38846380223660976, + "grad_norm": 0.09521484375, + "learning_rate": 0.00019121415497254425, + "loss": 0.0104, + "step": 660 + }, + { + "epoch": 0.3890523837551501, + "grad_norm": 1.6328125, + "learning_rate": 0.00019118974984746797, + "loss": 0.0222, + "step": 661 + }, + { + "epoch": 0.3896409652736904, + "grad_norm": 0.08984375, + "learning_rate": 0.00019116534472239172, + "loss": 0.0101, + "step": 662 + }, + { + "epoch": 0.3902295467922307, + "grad_norm": 0.07421875, + "learning_rate": 0.00019114093959731544, + "loss": 0.0102, + "step": 663 + }, + { + "epoch": 0.39081812831077106, + "grad_norm": 0.087890625, + "learning_rate": 0.0001911165344722392, + "loss": 0.0109, + "step": 664 + }, + { + "epoch": 0.39140670982931136, + "grad_norm": 0.80859375, + "learning_rate": 0.0001910921293471629, + "loss": 0.0181, + "step": 665 + }, + { + "epoch": 0.39199529134785166, + "grad_norm": 0.0947265625, + "learning_rate": 0.00019106772422208666, + "loss": 0.0104, + "step": 666 + }, + { + "epoch": 0.392583872866392, + "grad_norm": 0.080078125, + "learning_rate": 0.00019104331909701038, + "loss": 0.01, + "step": 667 + }, + { + "epoch": 0.3931724543849323, + "grad_norm": 0.1396484375, + "learning_rate": 0.00019101891397193413, + "loss": 0.0177, + "step": 668 + }, + { + "epoch": 0.3937610359034726, + "grad_norm": 0.0869140625, + "learning_rate": 0.00019099450884685785, + "loss": 0.0101, + "step": 669 + }, + { + "epoch": 0.39434961742201297, + "grad_norm": 0.076171875, + "learning_rate": 0.0001909701037217816, + "loss": 0.0101, + "step": 670 + }, + { + "epoch": 0.39493819894055326, + "grad_norm": 0.13671875, + "learning_rate": 0.00019094569859670531, + "loss": 0.0128, + "step": 671 + }, + { + "epoch": 0.39552678045909356, + "grad_norm": 0.3984375, + "learning_rate": 0.00019092129347162906, + "loss": 0.0143, + "step": 672 + }, + { + "epoch": 0.3961153619776339, + "grad_norm": 0.059814453125, + "learning_rate": 0.00019089688834655278, + "loss": 0.0103, + "step": 673 + }, + { + "epoch": 0.3967039434961742, + "grad_norm": 0.08740234375, + "learning_rate": 0.00019087248322147653, + "loss": 0.0098, + "step": 674 + }, + { + "epoch": 0.3972925250147145, + "grad_norm": 0.0966796875, + "learning_rate": 0.00019084807809640025, + "loss": 0.0106, + "step": 675 + }, + { + "epoch": 0.39788110653325487, + "grad_norm": 0.072265625, + "learning_rate": 0.000190823672971324, + "loss": 0.0097, + "step": 676 + }, + { + "epoch": 0.39846968805179517, + "grad_norm": 0.0615234375, + "learning_rate": 0.00019079926784624772, + "loss": 0.0084, + "step": 677 + }, + { + "epoch": 0.39905826957033547, + "grad_norm": 0.1962890625, + "learning_rate": 0.00019077486272117147, + "loss": 0.012, + "step": 678 + }, + { + "epoch": 0.3996468510888758, + "grad_norm": 0.06201171875, + "learning_rate": 0.0001907504575960952, + "loss": 0.0095, + "step": 679 + }, + { + "epoch": 0.4002354326074161, + "grad_norm": 0.06591796875, + "learning_rate": 0.0001907260524710189, + "loss": 0.0091, + "step": 680 + }, + { + "epoch": 0.4008240141259564, + "grad_norm": 0.0625, + "learning_rate": 0.00019070164734594265, + "loss": 0.0091, + "step": 681 + }, + { + "epoch": 0.4014125956444968, + "grad_norm": 0.0908203125, + "learning_rate": 0.00019067724222086637, + "loss": 0.0108, + "step": 682 + }, + { + "epoch": 0.40200117716303707, + "grad_norm": 0.06689453125, + "learning_rate": 0.00019065283709579012, + "loss": 0.0094, + "step": 683 + }, + { + "epoch": 0.4025897586815774, + "grad_norm": 0.083984375, + "learning_rate": 0.00019062843197071384, + "loss": 0.0107, + "step": 684 + }, + { + "epoch": 0.4031783402001177, + "grad_norm": 0.11962890625, + "learning_rate": 0.0001906040268456376, + "loss": 0.012, + "step": 685 + }, + { + "epoch": 0.403766921718658, + "grad_norm": 0.07373046875, + "learning_rate": 0.0001905796217205613, + "loss": 0.0093, + "step": 686 + }, + { + "epoch": 0.4043555032371984, + "grad_norm": 0.20703125, + "learning_rate": 0.00019055521659548506, + "loss": 0.0141, + "step": 687 + }, + { + "epoch": 0.4049440847557387, + "grad_norm": 0.1943359375, + "learning_rate": 0.00019053081147040878, + "loss": 0.0154, + "step": 688 + }, + { + "epoch": 0.405532666274279, + "grad_norm": 0.05859375, + "learning_rate": 0.00019050640634533253, + "loss": 0.0095, + "step": 689 + }, + { + "epoch": 0.40612124779281933, + "grad_norm": 0.11328125, + "learning_rate": 0.00019048200122025625, + "loss": 0.0123, + "step": 690 + }, + { + "epoch": 0.4067098293113596, + "grad_norm": 0.0556640625, + "learning_rate": 0.00019045759609518, + "loss": 0.0084, + "step": 691 + }, + { + "epoch": 0.4072984108298999, + "grad_norm": 0.07421875, + "learning_rate": 0.00019043319097010371, + "loss": 0.0096, + "step": 692 + }, + { + "epoch": 0.4078869923484403, + "grad_norm": 0.05712890625, + "learning_rate": 0.00019040878584502746, + "loss": 0.0092, + "step": 693 + }, + { + "epoch": 0.4084755738669806, + "grad_norm": 0.7578125, + "learning_rate": 0.00019038438071995118, + "loss": 0.0114, + "step": 694 + }, + { + "epoch": 0.4090641553855209, + "grad_norm": 0.193359375, + "learning_rate": 0.00019035997559487493, + "loss": 0.0125, + "step": 695 + }, + { + "epoch": 0.40965273690406123, + "grad_norm": 0.11572265625, + "learning_rate": 0.00019033557046979865, + "loss": 0.0146, + "step": 696 + }, + { + "epoch": 0.41024131842260153, + "grad_norm": 0.11083984375, + "learning_rate": 0.0001903111653447224, + "loss": 0.011, + "step": 697 + }, + { + "epoch": 0.41082989994114183, + "grad_norm": 0.08251953125, + "learning_rate": 0.00019028676021964612, + "loss": 0.0099, + "step": 698 + }, + { + "epoch": 0.4114184814596822, + "grad_norm": 0.1455078125, + "learning_rate": 0.00019026235509456987, + "loss": 0.0139, + "step": 699 + }, + { + "epoch": 0.4120070629782225, + "grad_norm": 0.130859375, + "learning_rate": 0.00019023794996949361, + "loss": 0.0121, + "step": 700 + }, + { + "epoch": 0.4125956444967628, + "grad_norm": 0.310546875, + "learning_rate": 0.00019021354484441733, + "loss": 0.0122, + "step": 701 + }, + { + "epoch": 0.41318422601530314, + "grad_norm": 0.1025390625, + "learning_rate": 0.00019018913971934108, + "loss": 0.0109, + "step": 702 + }, + { + "epoch": 0.41377280753384343, + "grad_norm": 0.07470703125, + "learning_rate": 0.0001901647345942648, + "loss": 0.0099, + "step": 703 + }, + { + "epoch": 0.41436138905238373, + "grad_norm": 0.068359375, + "learning_rate": 0.00019014032946918855, + "loss": 0.0098, + "step": 704 + }, + { + "epoch": 0.4149499705709241, + "grad_norm": 0.0576171875, + "learning_rate": 0.00019011592434411227, + "loss": 0.0095, + "step": 705 + }, + { + "epoch": 0.4155385520894644, + "grad_norm": 0.1201171875, + "learning_rate": 0.00019009151921903602, + "loss": 0.0113, + "step": 706 + }, + { + "epoch": 0.4161271336080047, + "grad_norm": 0.05419921875, + "learning_rate": 0.00019006711409395974, + "loss": 0.009, + "step": 707 + }, + { + "epoch": 0.41671571512654504, + "grad_norm": 0.10302734375, + "learning_rate": 0.00019004270896888349, + "loss": 0.011, + "step": 708 + }, + { + "epoch": 0.41730429664508534, + "grad_norm": 0.056640625, + "learning_rate": 0.0001900183038438072, + "loss": 0.009, + "step": 709 + }, + { + "epoch": 0.41789287816362564, + "grad_norm": 0.1083984375, + "learning_rate": 0.00018999389871873095, + "loss": 0.0107, + "step": 710 + }, + { + "epoch": 0.418481459682166, + "grad_norm": 0.07763671875, + "learning_rate": 0.00018996949359365467, + "loss": 0.0112, + "step": 711 + }, + { + "epoch": 0.4190700412007063, + "grad_norm": 0.0556640625, + "learning_rate": 0.00018994508846857842, + "loss": 0.0087, + "step": 712 + }, + { + "epoch": 0.41965862271924664, + "grad_norm": 0.07421875, + "learning_rate": 0.00018992068334350214, + "loss": 0.0093, + "step": 713 + }, + { + "epoch": 0.42024720423778694, + "grad_norm": 0.058837890625, + "learning_rate": 0.0001898962782184259, + "loss": 0.009, + "step": 714 + }, + { + "epoch": 0.42083578575632724, + "grad_norm": 1.90625, + "learning_rate": 0.0001898718730933496, + "loss": 0.0347, + "step": 715 + }, + { + "epoch": 0.4214243672748676, + "grad_norm": 0.048095703125, + "learning_rate": 0.00018984746796827336, + "loss": 0.0087, + "step": 716 + }, + { + "epoch": 0.4220129487934079, + "grad_norm": 0.2275390625, + "learning_rate": 0.00018982306284319708, + "loss": 0.0143, + "step": 717 + }, + { + "epoch": 0.4226015303119482, + "grad_norm": 2.234375, + "learning_rate": 0.00018979865771812083, + "loss": 0.0284, + "step": 718 + }, + { + "epoch": 0.42319011183048855, + "grad_norm": 0.0576171875, + "learning_rate": 0.00018977425259304455, + "loss": 0.009, + "step": 719 + }, + { + "epoch": 0.42377869334902885, + "grad_norm": 0.049560546875, + "learning_rate": 0.0001897498474679683, + "loss": 0.0081, + "step": 720 + }, + { + "epoch": 0.42436727486756914, + "grad_norm": 0.06982421875, + "learning_rate": 0.00018972544234289201, + "loss": 0.0099, + "step": 721 + }, + { + "epoch": 0.4249558563861095, + "grad_norm": 0.06982421875, + "learning_rate": 0.00018970103721781576, + "loss": 0.0095, + "step": 722 + }, + { + "epoch": 0.4255444379046498, + "grad_norm": 0.0703125, + "learning_rate": 0.00018967663209273948, + "loss": 0.0103, + "step": 723 + }, + { + "epoch": 0.4261330194231901, + "grad_norm": 0.068359375, + "learning_rate": 0.00018965222696766323, + "loss": 0.0093, + "step": 724 + }, + { + "epoch": 0.42672160094173045, + "grad_norm": 0.07080078125, + "learning_rate": 0.00018962782184258695, + "loss": 0.0094, + "step": 725 + }, + { + "epoch": 0.42731018246027075, + "grad_norm": 0.06689453125, + "learning_rate": 0.0001896034167175107, + "loss": 0.0093, + "step": 726 + }, + { + "epoch": 0.42789876397881105, + "grad_norm": 0.2236328125, + "learning_rate": 0.00018957901159243442, + "loss": 0.0145, + "step": 727 + }, + { + "epoch": 0.4284873454973514, + "grad_norm": 0.09765625, + "learning_rate": 0.00018955460646735817, + "loss": 0.0085, + "step": 728 + }, + { + "epoch": 0.4290759270158917, + "grad_norm": 0.059326171875, + "learning_rate": 0.0001895302013422819, + "loss": 0.0092, + "step": 729 + }, + { + "epoch": 0.429664508534432, + "grad_norm": 0.1240234375, + "learning_rate": 0.0001895057962172056, + "loss": 0.0118, + "step": 730 + }, + { + "epoch": 0.43025309005297235, + "grad_norm": 0.056396484375, + "learning_rate": 0.00018948139109212935, + "loss": 0.0087, + "step": 731 + }, + { + "epoch": 0.43084167157151265, + "grad_norm": 0.171875, + "learning_rate": 0.00018945698596705308, + "loss": 0.0127, + "step": 732 + }, + { + "epoch": 0.43143025309005295, + "grad_norm": 0.2197265625, + "learning_rate": 0.00018943258084197682, + "loss": 0.0123, + "step": 733 + }, + { + "epoch": 0.4320188346085933, + "grad_norm": 0.06689453125, + "learning_rate": 0.00018940817571690054, + "loss": 0.01, + "step": 734 + }, + { + "epoch": 0.4326074161271336, + "grad_norm": 0.142578125, + "learning_rate": 0.0001893837705918243, + "loss": 0.011, + "step": 735 + }, + { + "epoch": 0.4331959976456739, + "grad_norm": 0.369140625, + "learning_rate": 0.000189359365466748, + "loss": 0.0121, + "step": 736 + }, + { + "epoch": 0.43378457916421426, + "grad_norm": 0.0478515625, + "learning_rate": 0.00018933496034167176, + "loss": 0.0085, + "step": 737 + }, + { + "epoch": 0.43437316068275456, + "grad_norm": 8.6875, + "learning_rate": 0.00018931055521659548, + "loss": 0.0413, + "step": 738 + }, + { + "epoch": 0.43496174220129485, + "grad_norm": 0.091796875, + "learning_rate": 0.00018928615009151923, + "loss": 0.0099, + "step": 739 + }, + { + "epoch": 0.4355503237198352, + "grad_norm": 0.1484375, + "learning_rate": 0.00018926174496644295, + "loss": 0.0145, + "step": 740 + }, + { + "epoch": 0.4361389052383755, + "grad_norm": 0.0654296875, + "learning_rate": 0.0001892373398413667, + "loss": 0.009, + "step": 741 + }, + { + "epoch": 0.4367274867569158, + "grad_norm": 0.058837890625, + "learning_rate": 0.00018921293471629042, + "loss": 0.0085, + "step": 742 + }, + { + "epoch": 0.43731606827545616, + "grad_norm": 0.1953125, + "learning_rate": 0.00018918852959121416, + "loss": 0.0112, + "step": 743 + }, + { + "epoch": 0.43790464979399646, + "grad_norm": 0.07470703125, + "learning_rate": 0.00018916412446613788, + "loss": 0.0101, + "step": 744 + }, + { + "epoch": 0.4384932313125368, + "grad_norm": 0.1123046875, + "learning_rate": 0.00018913971934106163, + "loss": 0.0339, + "step": 745 + }, + { + "epoch": 0.4390818128310771, + "grad_norm": 0.89453125, + "learning_rate": 0.00018911531421598535, + "loss": 0.0156, + "step": 746 + }, + { + "epoch": 0.4396703943496174, + "grad_norm": 0.103515625, + "learning_rate": 0.0001890909090909091, + "loss": 0.0125, + "step": 747 + }, + { + "epoch": 0.44025897586815776, + "grad_norm": 0.08935546875, + "learning_rate": 0.00018906650396583282, + "loss": 0.0123, + "step": 748 + }, + { + "epoch": 0.44084755738669806, + "grad_norm": 0.06884765625, + "learning_rate": 0.00018904209884075657, + "loss": 0.0098, + "step": 749 + }, + { + "epoch": 0.44143613890523836, + "grad_norm": 0.06884765625, + "learning_rate": 0.0001890176937156803, + "loss": 0.0096, + "step": 750 + }, + { + "epoch": 0.4420247204237787, + "grad_norm": 0.05712890625, + "learning_rate": 0.00018899328859060404, + "loss": 0.0085, + "step": 751 + }, + { + "epoch": 0.442613301942319, + "grad_norm": 0.1484375, + "learning_rate": 0.00018896888346552776, + "loss": 0.0121, + "step": 752 + }, + { + "epoch": 0.4432018834608593, + "grad_norm": 0.07275390625, + "learning_rate": 0.0001889444783404515, + "loss": 0.0094, + "step": 753 + }, + { + "epoch": 0.44379046497939967, + "grad_norm": 0.08251953125, + "learning_rate": 0.00018892007321537525, + "loss": 0.0103, + "step": 754 + }, + { + "epoch": 0.44437904649793997, + "grad_norm": 0.07177734375, + "learning_rate": 0.00018889566809029897, + "loss": 0.0106, + "step": 755 + }, + { + "epoch": 0.44496762801648027, + "grad_norm": 0.061279296875, + "learning_rate": 0.00018887126296522272, + "loss": 0.0095, + "step": 756 + }, + { + "epoch": 0.4455562095350206, + "grad_norm": 0.06640625, + "learning_rate": 0.00018884685784014644, + "loss": 0.0099, + "step": 757 + }, + { + "epoch": 0.4461447910535609, + "grad_norm": 0.0556640625, + "learning_rate": 0.0001888224527150702, + "loss": 0.0086, + "step": 758 + }, + { + "epoch": 0.4467333725721012, + "grad_norm": 0.05419921875, + "learning_rate": 0.0001887980475899939, + "loss": 0.0085, + "step": 759 + }, + { + "epoch": 0.44732195409064157, + "grad_norm": 0.052490234375, + "learning_rate": 0.00018877364246491765, + "loss": 0.0087, + "step": 760 + }, + { + "epoch": 0.44791053560918187, + "grad_norm": 0.09130859375, + "learning_rate": 0.00018874923733984138, + "loss": 0.0096, + "step": 761 + }, + { + "epoch": 0.44849911712772217, + "grad_norm": 0.061279296875, + "learning_rate": 0.00018872483221476512, + "loss": 0.0092, + "step": 762 + }, + { + "epoch": 0.4490876986462625, + "grad_norm": 5.4375, + "learning_rate": 0.00018870042708968884, + "loss": 0.0142, + "step": 763 + }, + { + "epoch": 0.4496762801648028, + "grad_norm": 0.059814453125, + "learning_rate": 0.0001886760219646126, + "loss": 0.0096, + "step": 764 + }, + { + "epoch": 0.4502648616833431, + "grad_norm": 0.55078125, + "learning_rate": 0.0001886516168395363, + "loss": 0.0229, + "step": 765 + }, + { + "epoch": 0.4508534432018835, + "grad_norm": 2.109375, + "learning_rate": 0.00018862721171446006, + "loss": 0.0161, + "step": 766 + }, + { + "epoch": 0.4514420247204238, + "grad_norm": 1.0, + "learning_rate": 0.00018860280658938378, + "loss": 0.0103, + "step": 767 + }, + { + "epoch": 0.45203060623896407, + "grad_norm": 0.06396484375, + "learning_rate": 0.00018857840146430753, + "loss": 0.0088, + "step": 768 + }, + { + "epoch": 0.4526191877575044, + "grad_norm": 0.053955078125, + "learning_rate": 0.00018855399633923125, + "loss": 0.0085, + "step": 769 + }, + { + "epoch": 0.4532077692760447, + "grad_norm": 0.064453125, + "learning_rate": 0.000188529591214155, + "loss": 0.0089, + "step": 770 + }, + { + "epoch": 0.453796350794585, + "grad_norm": 0.0693359375, + "learning_rate": 0.00018850518608907872, + "loss": 0.009, + "step": 771 + }, + { + "epoch": 0.4543849323131254, + "grad_norm": 0.49609375, + "learning_rate": 0.00018848078096400246, + "loss": 0.0176, + "step": 772 + }, + { + "epoch": 0.4549735138316657, + "grad_norm": 0.419921875, + "learning_rate": 0.00018845637583892618, + "loss": 0.0164, + "step": 773 + }, + { + "epoch": 0.45556209535020603, + "grad_norm": 0.06787109375, + "learning_rate": 0.00018843197071384993, + "loss": 0.0096, + "step": 774 + }, + { + "epoch": 0.45615067686874633, + "grad_norm": 0.051513671875, + "learning_rate": 0.00018840756558877365, + "loss": 0.0085, + "step": 775 + }, + { + "epoch": 0.45673925838728663, + "grad_norm": 0.14453125, + "learning_rate": 0.0001883831604636974, + "loss": 0.0134, + "step": 776 + }, + { + "epoch": 0.457327839905827, + "grad_norm": 0.1279296875, + "learning_rate": 0.00018835875533862112, + "loss": 0.0098, + "step": 777 + }, + { + "epoch": 0.4579164214243673, + "grad_norm": 0.07373046875, + "learning_rate": 0.00018833435021354487, + "loss": 0.0087, + "step": 778 + }, + { + "epoch": 0.4585050029429076, + "grad_norm": 0.138671875, + "learning_rate": 0.0001883099450884686, + "loss": 0.0111, + "step": 779 + }, + { + "epoch": 0.45909358446144793, + "grad_norm": 0.10595703125, + "learning_rate": 0.00018828553996339233, + "loss": 0.0116, + "step": 780 + }, + { + "epoch": 0.45968216597998823, + "grad_norm": 0.0693359375, + "learning_rate": 0.00018826113483831606, + "loss": 0.0094, + "step": 781 + }, + { + "epoch": 0.46027074749852853, + "grad_norm": 0.07763671875, + "learning_rate": 0.00018823672971323978, + "loss": 0.0098, + "step": 782 + }, + { + "epoch": 0.4608593290170689, + "grad_norm": 0.427734375, + "learning_rate": 0.00018821232458816352, + "loss": 0.0129, + "step": 783 + }, + { + "epoch": 0.4614479105356092, + "grad_norm": 0.06591796875, + "learning_rate": 0.00018818791946308724, + "loss": 0.009, + "step": 784 + }, + { + "epoch": 0.4620364920541495, + "grad_norm": 0.25, + "learning_rate": 0.000188163514338011, + "loss": 0.0124, + "step": 785 + }, + { + "epoch": 0.46262507357268984, + "grad_norm": 0.061279296875, + "learning_rate": 0.0001881391092129347, + "loss": 0.0092, + "step": 786 + }, + { + "epoch": 0.46321365509123014, + "grad_norm": 0.1708984375, + "learning_rate": 0.00018811470408785846, + "loss": 0.0116, + "step": 787 + }, + { + "epoch": 0.46380223660977044, + "grad_norm": 0.06884765625, + "learning_rate": 0.00018809029896278218, + "loss": 0.0102, + "step": 788 + }, + { + "epoch": 0.4643908181283108, + "grad_norm": 0.0751953125, + "learning_rate": 0.00018806589383770593, + "loss": 0.0107, + "step": 789 + }, + { + "epoch": 0.4649793996468511, + "grad_norm": 0.06494140625, + "learning_rate": 0.00018804148871262965, + "loss": 0.0096, + "step": 790 + }, + { + "epoch": 0.4655679811653914, + "grad_norm": 0.06689453125, + "learning_rate": 0.0001880170835875534, + "loss": 0.0103, + "step": 791 + }, + { + "epoch": 0.46615656268393174, + "grad_norm": 0.11865234375, + "learning_rate": 0.00018799267846247712, + "loss": 0.0099, + "step": 792 + }, + { + "epoch": 0.46674514420247204, + "grad_norm": 0.06982421875, + "learning_rate": 0.00018796827333740086, + "loss": 0.0098, + "step": 793 + }, + { + "epoch": 0.46733372572101234, + "grad_norm": 0.06787109375, + "learning_rate": 0.00018794386821232458, + "loss": 0.0086, + "step": 794 + }, + { + "epoch": 0.4679223072395527, + "grad_norm": 0.09619140625, + "learning_rate": 0.00018791946308724833, + "loss": 0.0105, + "step": 795 + }, + { + "epoch": 0.468510888758093, + "grad_norm": 2.234375, + "learning_rate": 0.00018789505796217205, + "loss": 0.0322, + "step": 796 + }, + { + "epoch": 0.4690994702766333, + "grad_norm": 0.05810546875, + "learning_rate": 0.0001878706528370958, + "loss": 0.0092, + "step": 797 + }, + { + "epoch": 0.46968805179517364, + "grad_norm": 0.0771484375, + "learning_rate": 0.00018784624771201952, + "loss": 0.0101, + "step": 798 + }, + { + "epoch": 0.47027663331371394, + "grad_norm": 0.1015625, + "learning_rate": 0.00018782184258694327, + "loss": 0.012, + "step": 799 + }, + { + "epoch": 0.47086521483225424, + "grad_norm": 0.1875, + "learning_rate": 0.000187797437461867, + "loss": 0.0137, + "step": 800 + }, + { + "epoch": 0.4714537963507946, + "grad_norm": 0.054931640625, + "learning_rate": 0.00018777303233679074, + "loss": 0.0084, + "step": 801 + }, + { + "epoch": 0.4720423778693349, + "grad_norm": 0.064453125, + "learning_rate": 0.00018774862721171446, + "loss": 0.0096, + "step": 802 + }, + { + "epoch": 0.4726309593878752, + "grad_norm": 4.6875, + "learning_rate": 0.0001877242220866382, + "loss": 0.0297, + "step": 803 + }, + { + "epoch": 0.47321954090641555, + "grad_norm": 0.06640625, + "learning_rate": 0.00018769981696156192, + "loss": 0.0095, + "step": 804 + }, + { + "epoch": 0.47380812242495585, + "grad_norm": 0.058837890625, + "learning_rate": 0.00018767541183648567, + "loss": 0.009, + "step": 805 + }, + { + "epoch": 0.4743967039434962, + "grad_norm": 0.078125, + "learning_rate": 0.0001876510067114094, + "loss": 0.0103, + "step": 806 + }, + { + "epoch": 0.4749852854620365, + "grad_norm": 0.0634765625, + "learning_rate": 0.00018762660158633314, + "loss": 0.0094, + "step": 807 + }, + { + "epoch": 0.4755738669805768, + "grad_norm": 0.05615234375, + "learning_rate": 0.0001876021964612569, + "loss": 0.0086, + "step": 808 + }, + { + "epoch": 0.47616244849911715, + "grad_norm": 0.0703125, + "learning_rate": 0.0001875777913361806, + "loss": 0.0096, + "step": 809 + }, + { + "epoch": 0.47675103001765745, + "grad_norm": 0.0625, + "learning_rate": 0.00018755338621110436, + "loss": 0.0091, + "step": 810 + }, + { + "epoch": 0.47733961153619775, + "grad_norm": 0.2314453125, + "learning_rate": 0.00018752898108602808, + "loss": 0.0112, + "step": 811 + }, + { + "epoch": 0.4779281930547381, + "grad_norm": 0.72265625, + "learning_rate": 0.00018750457596095182, + "loss": 0.0141, + "step": 812 + }, + { + "epoch": 0.4785167745732784, + "grad_norm": 0.07958984375, + "learning_rate": 0.00018748017083587554, + "loss": 0.0104, + "step": 813 + }, + { + "epoch": 0.4791053560918187, + "grad_norm": 0.05810546875, + "learning_rate": 0.0001874557657107993, + "loss": 0.01, + "step": 814 + }, + { + "epoch": 0.47969393761035906, + "grad_norm": 0.06591796875, + "learning_rate": 0.000187431360585723, + "loss": 0.0098, + "step": 815 + }, + { + "epoch": 0.48028251912889935, + "grad_norm": 0.05078125, + "learning_rate": 0.00018740695546064676, + "loss": 0.0088, + "step": 816 + }, + { + "epoch": 0.48087110064743965, + "grad_norm": 0.24609375, + "learning_rate": 0.00018738255033557048, + "loss": 0.0145, + "step": 817 + }, + { + "epoch": 0.48145968216598, + "grad_norm": 0.0732421875, + "learning_rate": 0.00018735814521049423, + "loss": 0.0101, + "step": 818 + }, + { + "epoch": 0.4820482636845203, + "grad_norm": 0.087890625, + "learning_rate": 0.00018733374008541795, + "loss": 0.0118, + "step": 819 + }, + { + "epoch": 0.4826368452030606, + "grad_norm": 0.06103515625, + "learning_rate": 0.0001873093349603417, + "loss": 0.0095, + "step": 820 + }, + { + "epoch": 0.48322542672160096, + "grad_norm": 0.11376953125, + "learning_rate": 0.00018728492983526542, + "loss": 0.0115, + "step": 821 + }, + { + "epoch": 0.48381400824014126, + "grad_norm": 0.055908203125, + "learning_rate": 0.00018726052471018916, + "loss": 0.009, + "step": 822 + }, + { + "epoch": 0.48440258975868156, + "grad_norm": 0.068359375, + "learning_rate": 0.00018723611958511288, + "loss": 0.0092, + "step": 823 + }, + { + "epoch": 0.4849911712772219, + "grad_norm": 0.06298828125, + "learning_rate": 0.00018721171446003663, + "loss": 0.0093, + "step": 824 + }, + { + "epoch": 0.4855797527957622, + "grad_norm": 0.054931640625, + "learning_rate": 0.00018718730933496035, + "loss": 0.0089, + "step": 825 + }, + { + "epoch": 0.4861683343143025, + "grad_norm": 0.052490234375, + "learning_rate": 0.0001871629042098841, + "loss": 0.0087, + "step": 826 + }, + { + "epoch": 0.48675691583284286, + "grad_norm": 0.06591796875, + "learning_rate": 0.00018713849908480782, + "loss": 0.0089, + "step": 827 + }, + { + "epoch": 0.48734549735138316, + "grad_norm": 0.1318359375, + "learning_rate": 0.00018711409395973157, + "loss": 0.0099, + "step": 828 + }, + { + "epoch": 0.48793407886992346, + "grad_norm": 0.1494140625, + "learning_rate": 0.0001870896888346553, + "loss": 0.0129, + "step": 829 + }, + { + "epoch": 0.4885226603884638, + "grad_norm": 0.173828125, + "learning_rate": 0.00018706528370957904, + "loss": 0.0118, + "step": 830 + }, + { + "epoch": 0.4891112419070041, + "grad_norm": 0.05224609375, + "learning_rate": 0.00018704087858450276, + "loss": 0.0087, + "step": 831 + }, + { + "epoch": 0.4896998234255444, + "grad_norm": 0.09716796875, + "learning_rate": 0.00018701647345942648, + "loss": 0.0101, + "step": 832 + }, + { + "epoch": 0.49028840494408477, + "grad_norm": 0.043701171875, + "learning_rate": 0.00018699206833435022, + "loss": 0.0079, + "step": 833 + }, + { + "epoch": 0.49087698646262506, + "grad_norm": 0.0498046875, + "learning_rate": 0.00018696766320927394, + "loss": 0.009, + "step": 834 + }, + { + "epoch": 0.4914655679811654, + "grad_norm": 0.048583984375, + "learning_rate": 0.0001869432580841977, + "loss": 0.0078, + "step": 835 + }, + { + "epoch": 0.4920541494997057, + "grad_norm": 0.06982421875, + "learning_rate": 0.0001869188529591214, + "loss": 0.009, + "step": 836 + }, + { + "epoch": 0.492642731018246, + "grad_norm": 0.287109375, + "learning_rate": 0.00018689444783404516, + "loss": 0.0124, + "step": 837 + }, + { + "epoch": 0.49323131253678637, + "grad_norm": 0.0478515625, + "learning_rate": 0.00018687004270896888, + "loss": 0.0082, + "step": 838 + }, + { + "epoch": 0.49381989405532667, + "grad_norm": 0.05908203125, + "learning_rate": 0.00018684563758389263, + "loss": 0.0085, + "step": 839 + }, + { + "epoch": 0.49440847557386697, + "grad_norm": 0.126953125, + "learning_rate": 0.00018682123245881635, + "loss": 0.0135, + "step": 840 + }, + { + "epoch": 0.4949970570924073, + "grad_norm": 0.0673828125, + "learning_rate": 0.0001867968273337401, + "loss": 0.0099, + "step": 841 + }, + { + "epoch": 0.4955856386109476, + "grad_norm": 0.0576171875, + "learning_rate": 0.00018677242220866382, + "loss": 0.0093, + "step": 842 + }, + { + "epoch": 0.4961742201294879, + "grad_norm": 0.0673828125, + "learning_rate": 0.00018674801708358756, + "loss": 0.0096, + "step": 843 + }, + { + "epoch": 0.4967628016480283, + "grad_norm": 0.291015625, + "learning_rate": 0.00018672361195851128, + "loss": 0.0126, + "step": 844 + }, + { + "epoch": 0.4973513831665686, + "grad_norm": 0.08056640625, + "learning_rate": 0.00018669920683343503, + "loss": 0.0087, + "step": 845 + }, + { + "epoch": 0.49793996468510887, + "grad_norm": 0.0712890625, + "learning_rate": 0.00018667480170835875, + "loss": 0.0098, + "step": 846 + }, + { + "epoch": 0.4985285462036492, + "grad_norm": 0.046630859375, + "learning_rate": 0.0001866503965832825, + "loss": 0.0084, + "step": 847 + }, + { + "epoch": 0.4991171277221895, + "grad_norm": 0.0615234375, + "learning_rate": 0.00018662599145820622, + "loss": 0.0091, + "step": 848 + }, + { + "epoch": 0.4997057092407298, + "grad_norm": 0.09033203125, + "learning_rate": 0.00018660158633312997, + "loss": 0.0094, + "step": 849 + }, + { + "epoch": 0.5002942907592701, + "grad_norm": 0.0791015625, + "learning_rate": 0.0001865771812080537, + "loss": 0.0099, + "step": 850 + }, + { + "epoch": 0.5008828722778105, + "grad_norm": 0.625, + "learning_rate": 0.00018655277608297744, + "loss": 0.0119, + "step": 851 + }, + { + "epoch": 0.5014714537963508, + "grad_norm": 0.056640625, + "learning_rate": 0.00018652837095790116, + "loss": 0.0088, + "step": 852 + }, + { + "epoch": 0.5020600353148911, + "grad_norm": 0.07421875, + "learning_rate": 0.0001865039658328249, + "loss": 0.0103, + "step": 853 + }, + { + "epoch": 0.5026486168334314, + "grad_norm": 0.054931640625, + "learning_rate": 0.00018647956070774862, + "loss": 0.0083, + "step": 854 + }, + { + "epoch": 0.5032371983519718, + "grad_norm": 0.083984375, + "learning_rate": 0.00018645515558267237, + "loss": 0.0104, + "step": 855 + }, + { + "epoch": 0.503825779870512, + "grad_norm": 0.048095703125, + "learning_rate": 0.0001864307504575961, + "loss": 0.0081, + "step": 856 + }, + { + "epoch": 0.5044143613890524, + "grad_norm": 0.07421875, + "learning_rate": 0.00018640634533251984, + "loss": 0.0096, + "step": 857 + }, + { + "epoch": 0.5050029429075927, + "grad_norm": 0.1064453125, + "learning_rate": 0.00018638194020744356, + "loss": 0.0098, + "step": 858 + }, + { + "epoch": 0.505591524426133, + "grad_norm": 0.36328125, + "learning_rate": 0.0001863575350823673, + "loss": 0.0154, + "step": 859 + }, + { + "epoch": 0.5061801059446733, + "grad_norm": 0.04833984375, + "learning_rate": 0.00018633312995729103, + "loss": 0.0086, + "step": 860 + }, + { + "epoch": 0.5067686874632137, + "grad_norm": 0.0732421875, + "learning_rate": 0.00018630872483221478, + "loss": 0.0103, + "step": 861 + }, + { + "epoch": 0.5073572689817539, + "grad_norm": 0.052734375, + "learning_rate": 0.00018628431970713852, + "loss": 0.009, + "step": 862 + }, + { + "epoch": 0.5079458505002943, + "grad_norm": 0.1533203125, + "learning_rate": 0.00018625991458206224, + "loss": 0.012, + "step": 863 + }, + { + "epoch": 0.5085344320188346, + "grad_norm": 0.0556640625, + "learning_rate": 0.000186235509456986, + "loss": 0.0084, + "step": 864 + }, + { + "epoch": 0.5091230135373749, + "grad_norm": 0.072265625, + "learning_rate": 0.0001862111043319097, + "loss": 0.0103, + "step": 865 + }, + { + "epoch": 0.5097115950559152, + "grad_norm": 0.283203125, + "learning_rate": 0.00018618669920683346, + "loss": 0.0098, + "step": 866 + }, + { + "epoch": 0.5103001765744556, + "grad_norm": 0.2353515625, + "learning_rate": 0.00018616229408175718, + "loss": 0.017, + "step": 867 + }, + { + "epoch": 0.5108887580929958, + "grad_norm": 0.058837890625, + "learning_rate": 0.00018613788895668093, + "loss": 0.0097, + "step": 868 + }, + { + "epoch": 0.5114773396115362, + "grad_norm": 0.08935546875, + "learning_rate": 0.00018611348383160465, + "loss": 0.0099, + "step": 869 + }, + { + "epoch": 0.5120659211300765, + "grad_norm": 0.068359375, + "learning_rate": 0.0001860890787065284, + "loss": 0.0089, + "step": 870 + }, + { + "epoch": 0.5126545026486168, + "grad_norm": 0.29296875, + "learning_rate": 0.00018606467358145212, + "loss": 0.0151, + "step": 871 + }, + { + "epoch": 0.5132430841671571, + "grad_norm": 0.046630859375, + "learning_rate": 0.00018604026845637586, + "loss": 0.0091, + "step": 872 + }, + { + "epoch": 0.5138316656856975, + "grad_norm": 0.08349609375, + "learning_rate": 0.00018601586333129958, + "loss": 0.011, + "step": 873 + }, + { + "epoch": 0.5144202472042377, + "grad_norm": 0.0966796875, + "learning_rate": 0.00018599145820622333, + "loss": 0.0101, + "step": 874 + }, + { + "epoch": 0.5150088287227781, + "grad_norm": 0.053466796875, + "learning_rate": 0.00018596705308114705, + "loss": 0.0083, + "step": 875 + }, + { + "epoch": 0.5155974102413184, + "grad_norm": 0.0654296875, + "learning_rate": 0.0001859426479560708, + "loss": 0.0092, + "step": 876 + }, + { + "epoch": 0.5161859917598587, + "grad_norm": 0.061279296875, + "learning_rate": 0.00018591824283099452, + "loss": 0.009, + "step": 877 + }, + { + "epoch": 0.516774573278399, + "grad_norm": 0.07763671875, + "learning_rate": 0.00018589383770591827, + "loss": 0.0283, + "step": 878 + }, + { + "epoch": 0.5173631547969394, + "grad_norm": 0.052734375, + "learning_rate": 0.000185869432580842, + "loss": 0.0086, + "step": 879 + }, + { + "epoch": 0.5179517363154797, + "grad_norm": 0.060546875, + "learning_rate": 0.00018584502745576574, + "loss": 0.0087, + "step": 880 + }, + { + "epoch": 0.51854031783402, + "grad_norm": 0.0830078125, + "learning_rate": 0.00018582062233068946, + "loss": 0.0103, + "step": 881 + }, + { + "epoch": 0.5191288993525603, + "grad_norm": 0.07470703125, + "learning_rate": 0.00018579621720561318, + "loss": 0.0103, + "step": 882 + }, + { + "epoch": 0.5197174808711007, + "grad_norm": 0.10009765625, + "learning_rate": 0.00018577181208053692, + "loss": 0.0128, + "step": 883 + }, + { + "epoch": 0.5203060623896409, + "grad_norm": 0.05322265625, + "learning_rate": 0.00018574740695546064, + "loss": 0.009, + "step": 884 + }, + { + "epoch": 0.5208946439081813, + "grad_norm": 0.06201171875, + "learning_rate": 0.0001857230018303844, + "loss": 0.0089, + "step": 885 + }, + { + "epoch": 0.5214832254267217, + "grad_norm": 0.1982421875, + "learning_rate": 0.0001856985967053081, + "loss": 0.0125, + "step": 886 + }, + { + "epoch": 0.5220718069452619, + "grad_norm": 0.11328125, + "learning_rate": 0.00018567419158023186, + "loss": 0.0114, + "step": 887 + }, + { + "epoch": 0.5226603884638023, + "grad_norm": 0.054443359375, + "learning_rate": 0.00018564978645515558, + "loss": 0.0087, + "step": 888 + }, + { + "epoch": 0.5232489699823426, + "grad_norm": 0.05859375, + "learning_rate": 0.00018562538133007933, + "loss": 0.0099, + "step": 889 + }, + { + "epoch": 0.5238375515008828, + "grad_norm": 0.0556640625, + "learning_rate": 0.00018560097620500305, + "loss": 0.0084, + "step": 890 + }, + { + "epoch": 0.5244261330194232, + "grad_norm": 0.05615234375, + "learning_rate": 0.0001855765710799268, + "loss": 0.0083, + "step": 891 + }, + { + "epoch": 0.5250147145379636, + "grad_norm": 0.0634765625, + "learning_rate": 0.00018555216595485052, + "loss": 0.009, + "step": 892 + }, + { + "epoch": 0.5256032960565038, + "grad_norm": 0.06640625, + "learning_rate": 0.00018552776082977426, + "loss": 0.0096, + "step": 893 + }, + { + "epoch": 0.5261918775750442, + "grad_norm": 0.09130859375, + "learning_rate": 0.00018550335570469799, + "loss": 0.0098, + "step": 894 + }, + { + "epoch": 0.5267804590935845, + "grad_norm": 0.062255859375, + "learning_rate": 0.00018547895057962173, + "loss": 0.0098, + "step": 895 + }, + { + "epoch": 0.5273690406121248, + "grad_norm": 0.1767578125, + "learning_rate": 0.00018545454545454545, + "loss": 0.011, + "step": 896 + }, + { + "epoch": 0.5279576221306651, + "grad_norm": 0.1796875, + "learning_rate": 0.0001854301403294692, + "loss": 0.0106, + "step": 897 + }, + { + "epoch": 0.5285462036492055, + "grad_norm": 0.07666015625, + "learning_rate": 0.00018540573520439292, + "loss": 0.0106, + "step": 898 + }, + { + "epoch": 0.5291347851677457, + "grad_norm": 0.0888671875, + "learning_rate": 0.00018538133007931667, + "loss": 0.0107, + "step": 899 + }, + { + "epoch": 0.5297233666862861, + "grad_norm": 0.055908203125, + "learning_rate": 0.0001853569249542404, + "loss": 0.0091, + "step": 900 + }, + { + "epoch": 0.5303119482048264, + "grad_norm": 0.22265625, + "learning_rate": 0.00018533251982916414, + "loss": 0.0155, + "step": 901 + }, + { + "epoch": 0.5309005297233667, + "grad_norm": 0.08447265625, + "learning_rate": 0.00018530811470408786, + "loss": 0.0111, + "step": 902 + }, + { + "epoch": 0.531489111241907, + "grad_norm": 0.0869140625, + "learning_rate": 0.0001852837095790116, + "loss": 0.0099, + "step": 903 + }, + { + "epoch": 0.5320776927604474, + "grad_norm": 0.09765625, + "learning_rate": 0.00018525930445393533, + "loss": 0.0096, + "step": 904 + }, + { + "epoch": 0.5326662742789876, + "grad_norm": 0.057373046875, + "learning_rate": 0.00018523489932885907, + "loss": 0.0088, + "step": 905 + }, + { + "epoch": 0.533254855797528, + "grad_norm": 0.058837890625, + "learning_rate": 0.0001852104942037828, + "loss": 0.009, + "step": 906 + }, + { + "epoch": 0.5338434373160683, + "grad_norm": 0.058837890625, + "learning_rate": 0.00018518608907870654, + "loss": 0.0095, + "step": 907 + }, + { + "epoch": 0.5344320188346086, + "grad_norm": 0.70703125, + "learning_rate": 0.00018516168395363026, + "loss": 0.0107, + "step": 908 + }, + { + "epoch": 0.5350206003531489, + "grad_norm": 0.1171875, + "learning_rate": 0.000185137278828554, + "loss": 0.0112, + "step": 909 + }, + { + "epoch": 0.5356091818716893, + "grad_norm": 0.11572265625, + "learning_rate": 0.00018511287370347773, + "loss": 0.0111, + "step": 910 + }, + { + "epoch": 0.5361977633902295, + "grad_norm": 0.130859375, + "learning_rate": 0.00018508846857840148, + "loss": 0.0114, + "step": 911 + }, + { + "epoch": 0.5367863449087699, + "grad_norm": 1.5, + "learning_rate": 0.0001850640634533252, + "loss": 0.0131, + "step": 912 + }, + { + "epoch": 0.5373749264273102, + "grad_norm": 0.1025390625, + "learning_rate": 0.00018503965832824894, + "loss": 0.0107, + "step": 913 + }, + { + "epoch": 0.5379635079458505, + "grad_norm": 0.0673828125, + "learning_rate": 0.00018501525320317267, + "loss": 0.0099, + "step": 914 + }, + { + "epoch": 0.5385520894643908, + "grad_norm": 0.06298828125, + "learning_rate": 0.0001849908480780964, + "loss": 0.0085, + "step": 915 + }, + { + "epoch": 0.5391406709829312, + "grad_norm": 0.0771484375, + "learning_rate": 0.00018496644295302016, + "loss": 0.0104, + "step": 916 + }, + { + "epoch": 0.5397292525014714, + "grad_norm": 0.060302734375, + "learning_rate": 0.00018494203782794388, + "loss": 0.0092, + "step": 917 + }, + { + "epoch": 0.5403178340200118, + "grad_norm": 0.07275390625, + "learning_rate": 0.00018491763270286763, + "loss": 0.0089, + "step": 918 + }, + { + "epoch": 0.5409064155385521, + "grad_norm": 0.055419921875, + "learning_rate": 0.00018489322757779135, + "loss": 0.0088, + "step": 919 + }, + { + "epoch": 0.5414949970570924, + "grad_norm": 0.0771484375, + "learning_rate": 0.0001848688224527151, + "loss": 0.0101, + "step": 920 + }, + { + "epoch": 0.5420835785756327, + "grad_norm": 0.5078125, + "learning_rate": 0.00018484441732763882, + "loss": 0.0198, + "step": 921 + }, + { + "epoch": 0.5426721600941731, + "grad_norm": 0.1572265625, + "learning_rate": 0.00018482001220256256, + "loss": 0.0093, + "step": 922 + }, + { + "epoch": 0.5432607416127133, + "grad_norm": 0.05078125, + "learning_rate": 0.00018479560707748628, + "loss": 0.0084, + "step": 923 + }, + { + "epoch": 0.5438493231312537, + "grad_norm": 0.05712890625, + "learning_rate": 0.00018477120195241003, + "loss": 0.0084, + "step": 924 + }, + { + "epoch": 0.544437904649794, + "grad_norm": 0.09423828125, + "learning_rate": 0.00018474679682733375, + "loss": 0.0109, + "step": 925 + }, + { + "epoch": 0.5450264861683343, + "grad_norm": 0.0517578125, + "learning_rate": 0.0001847223917022575, + "loss": 0.0084, + "step": 926 + }, + { + "epoch": 0.5456150676868746, + "grad_norm": 0.052978515625, + "learning_rate": 0.00018469798657718122, + "loss": 0.0083, + "step": 927 + }, + { + "epoch": 0.546203649205415, + "grad_norm": 0.08251953125, + "learning_rate": 0.00018467358145210497, + "loss": 0.0108, + "step": 928 + }, + { + "epoch": 0.5467922307239552, + "grad_norm": 0.07275390625, + "learning_rate": 0.0001846491763270287, + "loss": 0.01, + "step": 929 + }, + { + "epoch": 0.5473808122424956, + "grad_norm": 0.373046875, + "learning_rate": 0.00018462477120195244, + "loss": 0.0162, + "step": 930 + }, + { + "epoch": 0.5479693937610359, + "grad_norm": 0.078125, + "learning_rate": 0.00018460036607687616, + "loss": 0.012, + "step": 931 + }, + { + "epoch": 0.5485579752795762, + "grad_norm": 0.07177734375, + "learning_rate": 0.0001845759609517999, + "loss": 0.0087, + "step": 932 + }, + { + "epoch": 0.5491465567981165, + "grad_norm": 0.06005859375, + "learning_rate": 0.0001845515558267236, + "loss": 0.0094, + "step": 933 + }, + { + "epoch": 0.5497351383166569, + "grad_norm": 0.058349609375, + "learning_rate": 0.00018452715070164735, + "loss": 0.0087, + "step": 934 + }, + { + "epoch": 0.5503237198351971, + "grad_norm": 0.060546875, + "learning_rate": 0.00018450274557657107, + "loss": 0.0093, + "step": 935 + }, + { + "epoch": 0.5509123013537375, + "grad_norm": 0.056396484375, + "learning_rate": 0.0001844783404514948, + "loss": 0.009, + "step": 936 + }, + { + "epoch": 0.5515008828722778, + "grad_norm": 0.0439453125, + "learning_rate": 0.00018445393532641856, + "loss": 0.0081, + "step": 937 + }, + { + "epoch": 0.5520894643908181, + "grad_norm": 0.07861328125, + "learning_rate": 0.00018442953020134228, + "loss": 0.0095, + "step": 938 + }, + { + "epoch": 0.5526780459093584, + "grad_norm": 0.0927734375, + "learning_rate": 0.00018440512507626603, + "loss": 0.0101, + "step": 939 + }, + { + "epoch": 0.5532666274278988, + "grad_norm": 0.044677734375, + "learning_rate": 0.00018438071995118975, + "loss": 0.0082, + "step": 940 + }, + { + "epoch": 0.5538552089464391, + "grad_norm": 0.08984375, + "learning_rate": 0.0001843563148261135, + "loss": 0.0093, + "step": 941 + }, + { + "epoch": 0.5544437904649794, + "grad_norm": 0.0791015625, + "learning_rate": 0.00018433190970103722, + "loss": 0.0094, + "step": 942 + }, + { + "epoch": 0.5550323719835197, + "grad_norm": 0.06640625, + "learning_rate": 0.00018430750457596097, + "loss": 0.0097, + "step": 943 + }, + { + "epoch": 0.5556209535020601, + "grad_norm": 0.05615234375, + "learning_rate": 0.00018428309945088469, + "loss": 0.0085, + "step": 944 + }, + { + "epoch": 0.5562095350206003, + "grad_norm": 0.1123046875, + "learning_rate": 0.00018425869432580843, + "loss": 0.0107, + "step": 945 + }, + { + "epoch": 0.5567981165391407, + "grad_norm": 0.076171875, + "learning_rate": 0.00018423428920073215, + "loss": 0.0106, + "step": 946 + }, + { + "epoch": 0.557386698057681, + "grad_norm": 0.0537109375, + "learning_rate": 0.0001842098840756559, + "loss": 0.0086, + "step": 947 + }, + { + "epoch": 0.5579752795762213, + "grad_norm": 0.05322265625, + "learning_rate": 0.00018418547895057962, + "loss": 0.0082, + "step": 948 + }, + { + "epoch": 0.5585638610947616, + "grad_norm": 0.1396484375, + "learning_rate": 0.00018416107382550337, + "loss": 0.0107, + "step": 949 + }, + { + "epoch": 0.559152442613302, + "grad_norm": 0.07177734375, + "learning_rate": 0.0001841366687004271, + "loss": 0.0102, + "step": 950 + }, + { + "epoch": 0.5597410241318422, + "grad_norm": 0.057861328125, + "learning_rate": 0.00018411226357535084, + "loss": 0.0092, + "step": 951 + }, + { + "epoch": 0.5603296056503826, + "grad_norm": 0.06591796875, + "learning_rate": 0.00018408785845027456, + "loss": 0.0099, + "step": 952 + }, + { + "epoch": 0.5609181871689229, + "grad_norm": 0.0478515625, + "learning_rate": 0.0001840634533251983, + "loss": 0.0082, + "step": 953 + }, + { + "epoch": 0.5615067686874632, + "grad_norm": 0.07861328125, + "learning_rate": 0.00018403904820012203, + "loss": 0.0105, + "step": 954 + }, + { + "epoch": 0.5620953502060035, + "grad_norm": 0.068359375, + "learning_rate": 0.00018401464307504577, + "loss": 0.0092, + "step": 955 + }, + { + "epoch": 0.5626839317245439, + "grad_norm": 0.05029296875, + "learning_rate": 0.0001839902379499695, + "loss": 0.0084, + "step": 956 + }, + { + "epoch": 0.5632725132430841, + "grad_norm": 0.07666015625, + "learning_rate": 0.00018396583282489324, + "loss": 0.0106, + "step": 957 + }, + { + "epoch": 0.5638610947616245, + "grad_norm": 0.06591796875, + "learning_rate": 0.00018394142769981696, + "loss": 0.0103, + "step": 958 + }, + { + "epoch": 0.5644496762801648, + "grad_norm": 0.19921875, + "learning_rate": 0.0001839170225747407, + "loss": 0.0171, + "step": 959 + }, + { + "epoch": 0.5650382577987051, + "grad_norm": 0.040771484375, + "learning_rate": 0.00018389261744966443, + "loss": 0.0074, + "step": 960 + }, + { + "epoch": 0.5656268393172454, + "grad_norm": 0.055419921875, + "learning_rate": 0.00018386821232458818, + "loss": 0.0085, + "step": 961 + }, + { + "epoch": 0.5662154208357858, + "grad_norm": 0.052978515625, + "learning_rate": 0.0001838438071995119, + "loss": 0.0094, + "step": 962 + }, + { + "epoch": 0.566804002354326, + "grad_norm": 0.050048828125, + "learning_rate": 0.00018381940207443565, + "loss": 0.0082, + "step": 963 + }, + { + "epoch": 0.5673925838728664, + "grad_norm": 0.05517578125, + "learning_rate": 0.00018379499694935937, + "loss": 0.0085, + "step": 964 + }, + { + "epoch": 0.5679811653914068, + "grad_norm": 0.053466796875, + "learning_rate": 0.0001837705918242831, + "loss": 0.0087, + "step": 965 + }, + { + "epoch": 0.568569746909947, + "grad_norm": 0.08203125, + "learning_rate": 0.00018374618669920683, + "loss": 0.0091, + "step": 966 + }, + { + "epoch": 0.5691583284284873, + "grad_norm": 0.0615234375, + "learning_rate": 0.00018372178157413058, + "loss": 0.0088, + "step": 967 + }, + { + "epoch": 0.5697469099470277, + "grad_norm": 0.0693359375, + "learning_rate": 0.0001836973764490543, + "loss": 0.009, + "step": 968 + }, + { + "epoch": 0.570335491465568, + "grad_norm": 0.050537109375, + "learning_rate": 0.00018367297132397805, + "loss": 0.0083, + "step": 969 + }, + { + "epoch": 0.5709240729841083, + "grad_norm": 0.06591796875, + "learning_rate": 0.0001836485661989018, + "loss": 0.0096, + "step": 970 + }, + { + "epoch": 0.5715126545026487, + "grad_norm": 0.06640625, + "learning_rate": 0.00018362416107382552, + "loss": 0.0098, + "step": 971 + }, + { + "epoch": 0.5721012360211889, + "grad_norm": 0.11181640625, + "learning_rate": 0.00018359975594874927, + "loss": 0.0102, + "step": 972 + }, + { + "epoch": 0.5726898175397293, + "grad_norm": 0.056640625, + "learning_rate": 0.00018357535082367299, + "loss": 0.0093, + "step": 973 + }, + { + "epoch": 0.5732783990582696, + "grad_norm": 0.06298828125, + "learning_rate": 0.00018355094569859673, + "loss": 0.0092, + "step": 974 + }, + { + "epoch": 0.5738669805768098, + "grad_norm": 0.052490234375, + "learning_rate": 0.00018352654057352045, + "loss": 0.0082, + "step": 975 + }, + { + "epoch": 0.5744555620953502, + "grad_norm": 0.1826171875, + "learning_rate": 0.0001835021354484442, + "loss": 0.0129, + "step": 976 + }, + { + "epoch": 0.5750441436138906, + "grad_norm": 0.045166015625, + "learning_rate": 0.00018347773032336792, + "loss": 0.0083, + "step": 977 + }, + { + "epoch": 0.5756327251324308, + "grad_norm": 0.053955078125, + "learning_rate": 0.00018345332519829167, + "loss": 0.0092, + "step": 978 + }, + { + "epoch": 0.5762213066509712, + "grad_norm": 0.0751953125, + "learning_rate": 0.0001834289200732154, + "loss": 0.0099, + "step": 979 + }, + { + "epoch": 0.5768098881695115, + "grad_norm": 1.78125, + "learning_rate": 0.00018340451494813914, + "loss": 0.0282, + "step": 980 + }, + { + "epoch": 0.5773984696880518, + "grad_norm": 0.0478515625, + "learning_rate": 0.00018338010982306286, + "loss": 0.0089, + "step": 981 + }, + { + "epoch": 0.5779870512065921, + "grad_norm": 0.049072265625, + "learning_rate": 0.0001833557046979866, + "loss": 0.0082, + "step": 982 + }, + { + "epoch": 0.5785756327251325, + "grad_norm": 0.05615234375, + "learning_rate": 0.0001833312995729103, + "loss": 0.0081, + "step": 983 + }, + { + "epoch": 0.5791642142436727, + "grad_norm": 0.0947265625, + "learning_rate": 0.00018330689444783405, + "loss": 0.0118, + "step": 984 + }, + { + "epoch": 0.5797527957622131, + "grad_norm": 0.07177734375, + "learning_rate": 0.00018328248932275777, + "loss": 0.0089, + "step": 985 + }, + { + "epoch": 0.5803413772807534, + "grad_norm": 0.05078125, + "learning_rate": 0.00018325808419768151, + "loss": 0.0082, + "step": 986 + }, + { + "epoch": 0.5809299587992937, + "grad_norm": 0.058837890625, + "learning_rate": 0.00018323367907260523, + "loss": 0.0082, + "step": 987 + }, + { + "epoch": 0.581518540317834, + "grad_norm": 0.05859375, + "learning_rate": 0.00018320927394752898, + "loss": 0.0088, + "step": 988 + }, + { + "epoch": 0.5821071218363744, + "grad_norm": 0.0654296875, + "learning_rate": 0.0001831848688224527, + "loss": 0.0096, + "step": 989 + }, + { + "epoch": 0.5826957033549146, + "grad_norm": 0.048828125, + "learning_rate": 0.00018316046369737645, + "loss": 0.0074, + "step": 990 + }, + { + "epoch": 0.583284284873455, + "grad_norm": 0.06591796875, + "learning_rate": 0.0001831360585723002, + "loss": 0.0092, + "step": 991 + }, + { + "epoch": 0.5838728663919953, + "grad_norm": 0.0556640625, + "learning_rate": 0.00018311165344722392, + "loss": 0.0087, + "step": 992 + }, + { + "epoch": 0.5844614479105356, + "grad_norm": 0.10302734375, + "learning_rate": 0.00018308724832214767, + "loss": 0.0112, + "step": 993 + }, + { + "epoch": 0.5850500294290759, + "grad_norm": 0.055908203125, + "learning_rate": 0.00018306284319707139, + "loss": 0.0084, + "step": 994 + }, + { + "epoch": 0.5856386109476163, + "grad_norm": 0.0595703125, + "learning_rate": 0.00018303843807199513, + "loss": 0.0089, + "step": 995 + }, + { + "epoch": 0.5862271924661565, + "grad_norm": 0.053955078125, + "learning_rate": 0.00018301403294691885, + "loss": 0.0094, + "step": 996 + }, + { + "epoch": 0.5868157739846969, + "grad_norm": 0.06494140625, + "learning_rate": 0.0001829896278218426, + "loss": 0.0099, + "step": 997 + }, + { + "epoch": 0.5874043555032372, + "grad_norm": 0.044189453125, + "learning_rate": 0.00018296522269676632, + "loss": 0.0076, + "step": 998 + }, + { + "epoch": 0.5879929370217775, + "grad_norm": 0.056396484375, + "learning_rate": 0.00018294081757169007, + "loss": 0.009, + "step": 999 + }, + { + "epoch": 0.5885815185403178, + "grad_norm": 0.3515625, + "learning_rate": 0.0001829164124466138, + "loss": 0.0125, + "step": 1000 + } + ], + "logging_steps": 1, + "max_steps": 8495, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 64, + "trial_name": null, + "trial_params": null +}