| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.5885815185403178, | |
| "eval_steps": 500, | |
| "global_step": 1000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0005885815185403178, | |
| "grad_norm": 35.75, | |
| "learning_rate": 6.666666666666667e-07, | |
| "loss": 3.1406, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0011771630370806356, | |
| "grad_norm": 39.0, | |
| "learning_rate": 1.3333333333333334e-06, | |
| "loss": 3.1719, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0017657445556209534, | |
| "grad_norm": 59.75, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 3.2656, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.002354326074161271, | |
| "grad_norm": 41.25, | |
| "learning_rate": 2.666666666666667e-06, | |
| "loss": 3.1562, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.002942907592701589, | |
| "grad_norm": 43.5, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 3.2812, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.003531489111241907, | |
| "grad_norm": 700.0, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 3.3438, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.004120070629782225, | |
| "grad_norm": 92.5, | |
| "learning_rate": 4.666666666666667e-06, | |
| "loss": 3.2969, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.004708652148322542, | |
| "grad_norm": 87.0, | |
| "learning_rate": 5.333333333333334e-06, | |
| "loss": 3.2031, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.00529723366686286, | |
| "grad_norm": 46.5, | |
| "learning_rate": 6e-06, | |
| "loss": 3.1094, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.005885815185403178, | |
| "grad_norm": 27.875, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 3.0781, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.006474396703943496, | |
| "grad_norm": 53.25, | |
| "learning_rate": 7.333333333333334e-06, | |
| "loss": 3.0781, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.007062978222483814, | |
| "grad_norm": 48.5, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 3.25, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.007651559741024131, | |
| "grad_norm": 35.75, | |
| "learning_rate": 8.666666666666668e-06, | |
| "loss": 3.0469, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.00824014125956445, | |
| "grad_norm": 165.0, | |
| "learning_rate": 9.333333333333334e-06, | |
| "loss": 3.2969, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.008828722778104767, | |
| "grad_norm": 213.0, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0781, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.009417304296645085, | |
| "grad_norm": 69.0, | |
| "learning_rate": 1.0666666666666667e-05, | |
| "loss": 3.0469, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.010005885815185403, | |
| "grad_norm": 104.5, | |
| "learning_rate": 1.1333333333333334e-05, | |
| "loss": 3.0469, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.01059446733372572, | |
| "grad_norm": 20.625, | |
| "learning_rate": 1.2e-05, | |
| "loss": 3.1094, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.011183048852266038, | |
| "grad_norm": 38.25, | |
| "learning_rate": 1.2666666666666668e-05, | |
| "loss": 3.1719, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.011771630370806356, | |
| "grad_norm": 65.5, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 3.0938, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.012360211889346674, | |
| "grad_norm": 39.5, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 3.1406, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.012948793407886992, | |
| "grad_norm": 16.5, | |
| "learning_rate": 1.4666666666666668e-05, | |
| "loss": 3.0625, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.01353737492642731, | |
| "grad_norm": 35.5, | |
| "learning_rate": 1.5333333333333334e-05, | |
| "loss": 3.125, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.014125956444967627, | |
| "grad_norm": 133.0, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 3.0625, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.014714537963507945, | |
| "grad_norm": 31.25, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 3.1406, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.015303119482048263, | |
| "grad_norm": 22.5, | |
| "learning_rate": 1.7333333333333336e-05, | |
| "loss": 3.0938, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.015891701000588582, | |
| "grad_norm": 41.25, | |
| "learning_rate": 1.8e-05, | |
| "loss": 3.125, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.0164802825191289, | |
| "grad_norm": 42.75, | |
| "learning_rate": 1.866666666666667e-05, | |
| "loss": 3.0, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.017068864037669218, | |
| "grad_norm": 90.5, | |
| "learning_rate": 1.9333333333333333e-05, | |
| "loss": 3.0469, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.017657445556209534, | |
| "grad_norm": 45.25, | |
| "learning_rate": 2e-05, | |
| "loss": 3.1406, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.018246027074749854, | |
| "grad_norm": 105.0, | |
| "learning_rate": 2.0666666666666666e-05, | |
| "loss": 2.9531, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.01883460859329017, | |
| "grad_norm": 29.875, | |
| "learning_rate": 2.1333333333333335e-05, | |
| "loss": 3.1719, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.01942319011183049, | |
| "grad_norm": 29.125, | |
| "learning_rate": 2.2000000000000003e-05, | |
| "loss": 3.1875, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.020011771630370805, | |
| "grad_norm": 34.0, | |
| "learning_rate": 2.2666666666666668e-05, | |
| "loss": 3.125, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.020600353148911125, | |
| "grad_norm": 121.0, | |
| "learning_rate": 2.3333333333333336e-05, | |
| "loss": 3.2812, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.02118893466745144, | |
| "grad_norm": 18.75, | |
| "learning_rate": 2.4e-05, | |
| "loss": 3.125, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.02177751618599176, | |
| "grad_norm": 32.0, | |
| "learning_rate": 2.466666666666667e-05, | |
| "loss": 2.9688, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.022366097704532076, | |
| "grad_norm": 72.0, | |
| "learning_rate": 2.5333333333333337e-05, | |
| "loss": 2.9531, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.022954679223072396, | |
| "grad_norm": 59.25, | |
| "learning_rate": 2.6000000000000002e-05, | |
| "loss": 3.125, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.023543260741612712, | |
| "grad_norm": 20.625, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 2.9844, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02413184226015303, | |
| "grad_norm": 73.0, | |
| "learning_rate": 2.733333333333333e-05, | |
| "loss": 3.0156, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.024720423778693348, | |
| "grad_norm": 11.8125, | |
| "learning_rate": 2.8000000000000003e-05, | |
| "loss": 2.8906, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.025309005297233667, | |
| "grad_norm": 34.5, | |
| "learning_rate": 2.8666666666666668e-05, | |
| "loss": 3.0312, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.025897586815773983, | |
| "grad_norm": 25.375, | |
| "learning_rate": 2.9333333333333336e-05, | |
| "loss": 2.9844, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.026486168334314303, | |
| "grad_norm": 17.5, | |
| "learning_rate": 3e-05, | |
| "loss": 3.0, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.02707474985285462, | |
| "grad_norm": 29.0, | |
| "learning_rate": 3.066666666666667e-05, | |
| "loss": 3.0156, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.02766333137139494, | |
| "grad_norm": 56.25, | |
| "learning_rate": 3.1333333333333334e-05, | |
| "loss": 2.875, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.028251912889935255, | |
| "grad_norm": 133.0, | |
| "learning_rate": 3.2000000000000005e-05, | |
| "loss": 3.0, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.028840494408475574, | |
| "grad_norm": 21.5, | |
| "learning_rate": 3.266666666666667e-05, | |
| "loss": 2.8438, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.02942907592701589, | |
| "grad_norm": 21.0, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 2.8438, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03001765744555621, | |
| "grad_norm": 29.125, | |
| "learning_rate": 3.4000000000000007e-05, | |
| "loss": 2.8125, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.030606238964096526, | |
| "grad_norm": 34.25, | |
| "learning_rate": 3.466666666666667e-05, | |
| "loss": 2.8906, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.031194820482636845, | |
| "grad_norm": 74.5, | |
| "learning_rate": 3.5333333333333336e-05, | |
| "loss": 2.7344, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.031783402001177165, | |
| "grad_norm": 19.375, | |
| "learning_rate": 3.6e-05, | |
| "loss": 2.7969, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.032371983519717484, | |
| "grad_norm": 146.0, | |
| "learning_rate": 3.6666666666666666e-05, | |
| "loss": 2.9219, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.0329605650382578, | |
| "grad_norm": 40.25, | |
| "learning_rate": 3.733333333333334e-05, | |
| "loss": 2.6719, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.033549146556798116, | |
| "grad_norm": 64.0, | |
| "learning_rate": 3.8e-05, | |
| "loss": 2.5938, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.034137728075338436, | |
| "grad_norm": 44.0, | |
| "learning_rate": 3.866666666666667e-05, | |
| "loss": 2.5, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.034726309593878756, | |
| "grad_norm": 30.625, | |
| "learning_rate": 3.933333333333333e-05, | |
| "loss": 2.6719, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.03531489111241907, | |
| "grad_norm": 59.25, | |
| "learning_rate": 4e-05, | |
| "loss": 2.5312, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03590347263095939, | |
| "grad_norm": 13.9375, | |
| "learning_rate": 4.066666666666667e-05, | |
| "loss": 2.4688, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.03649205414949971, | |
| "grad_norm": 91.5, | |
| "learning_rate": 4.133333333333333e-05, | |
| "loss": 2.5625, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.03708063566804003, | |
| "grad_norm": 54.5, | |
| "learning_rate": 4.2e-05, | |
| "loss": 2.5312, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.03766921718658034, | |
| "grad_norm": 234.0, | |
| "learning_rate": 4.266666666666667e-05, | |
| "loss": 2.5, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.03825779870512066, | |
| "grad_norm": 58.75, | |
| "learning_rate": 4.3333333333333334e-05, | |
| "loss": 2.2969, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.03884638022366098, | |
| "grad_norm": 26.875, | |
| "learning_rate": 4.4000000000000006e-05, | |
| "loss": 2.1562, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.0394349617422013, | |
| "grad_norm": 30.25, | |
| "learning_rate": 4.466666666666667e-05, | |
| "loss": 2.25, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.04002354326074161, | |
| "grad_norm": 61.0, | |
| "learning_rate": 4.5333333333333335e-05, | |
| "loss": 2.3594, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.04061212477928193, | |
| "grad_norm": 36.0, | |
| "learning_rate": 4.600000000000001e-05, | |
| "loss": 2.1719, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.04120070629782225, | |
| "grad_norm": 56.0, | |
| "learning_rate": 4.666666666666667e-05, | |
| "loss": 2.1875, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.04178928781636257, | |
| "grad_norm": 34.75, | |
| "learning_rate": 4.7333333333333336e-05, | |
| "loss": 2.1719, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.04237786933490288, | |
| "grad_norm": 58.25, | |
| "learning_rate": 4.8e-05, | |
| "loss": 2.125, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.0429664508534432, | |
| "grad_norm": 67.0, | |
| "learning_rate": 4.866666666666667e-05, | |
| "loss": 2.4531, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.04355503237198352, | |
| "grad_norm": 21.625, | |
| "learning_rate": 4.933333333333334e-05, | |
| "loss": 2.1094, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.04414361389052384, | |
| "grad_norm": 43.5, | |
| "learning_rate": 5e-05, | |
| "loss": 2.0156, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.04473219540906415, | |
| "grad_norm": 17.125, | |
| "learning_rate": 5.0666666666666674e-05, | |
| "loss": 2.0156, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.04532077692760447, | |
| "grad_norm": 40.25, | |
| "learning_rate": 5.133333333333333e-05, | |
| "loss": 2.1094, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.04590935844614479, | |
| "grad_norm": 29.875, | |
| "learning_rate": 5.2000000000000004e-05, | |
| "loss": 1.9297, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.04649793996468511, | |
| "grad_norm": 45.5, | |
| "learning_rate": 5.266666666666666e-05, | |
| "loss": 2.0938, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.047086521483225424, | |
| "grad_norm": 73.0, | |
| "learning_rate": 5.333333333333333e-05, | |
| "loss": 1.9453, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.047675103001765744, | |
| "grad_norm": 28.5, | |
| "learning_rate": 5.4000000000000005e-05, | |
| "loss": 1.7891, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.04826368452030606, | |
| "grad_norm": 544.0, | |
| "learning_rate": 5.466666666666666e-05, | |
| "loss": 2.0469, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.04885226603884638, | |
| "grad_norm": 42.25, | |
| "learning_rate": 5.5333333333333334e-05, | |
| "loss": 1.75, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.049440847557386695, | |
| "grad_norm": 52.25, | |
| "learning_rate": 5.6000000000000006e-05, | |
| "loss": 1.7031, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.050029429075927015, | |
| "grad_norm": 41.5, | |
| "learning_rate": 5.666666666666667e-05, | |
| "loss": 1.7969, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.050618010594467334, | |
| "grad_norm": 131.0, | |
| "learning_rate": 5.7333333333333336e-05, | |
| "loss": 1.5234, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.051206592113007654, | |
| "grad_norm": 52.0, | |
| "learning_rate": 5.8e-05, | |
| "loss": 1.625, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.05179517363154797, | |
| "grad_norm": 43.75, | |
| "learning_rate": 5.866666666666667e-05, | |
| "loss": 1.6719, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.052383755150088286, | |
| "grad_norm": 27.375, | |
| "learning_rate": 5.9333333333333343e-05, | |
| "loss": 1.7188, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.052972336668628606, | |
| "grad_norm": 32.25, | |
| "learning_rate": 6e-05, | |
| "loss": 1.8047, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.053560918187168925, | |
| "grad_norm": 46.25, | |
| "learning_rate": 6.066666666666667e-05, | |
| "loss": 1.5859, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.05414949970570924, | |
| "grad_norm": 76.5, | |
| "learning_rate": 6.133333333333334e-05, | |
| "loss": 1.4062, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.05473808122424956, | |
| "grad_norm": 187.0, | |
| "learning_rate": 6.2e-05, | |
| "loss": 1.3359, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.05532666274278988, | |
| "grad_norm": 36.5, | |
| "learning_rate": 6.266666666666667e-05, | |
| "loss": 1.3672, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.055915244261330196, | |
| "grad_norm": 41.5, | |
| "learning_rate": 6.333333333333333e-05, | |
| "loss": 1.1172, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.05650382577987051, | |
| "grad_norm": 141.0, | |
| "learning_rate": 6.400000000000001e-05, | |
| "loss": 1.6016, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.05709240729841083, | |
| "grad_norm": 49.75, | |
| "learning_rate": 6.466666666666666e-05, | |
| "loss": 1.4375, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.05768098881695115, | |
| "grad_norm": 23.875, | |
| "learning_rate": 6.533333333333334e-05, | |
| "loss": 1.3594, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.05826957033549147, | |
| "grad_norm": 37.75, | |
| "learning_rate": 6.6e-05, | |
| "loss": 1.5, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.05885815185403178, | |
| "grad_norm": 32.0, | |
| "learning_rate": 6.666666666666667e-05, | |
| "loss": 1.3438, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0594467333725721, | |
| "grad_norm": 37.0, | |
| "learning_rate": 6.733333333333333e-05, | |
| "loss": 1.2578, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.06003531489111242, | |
| "grad_norm": 66.5, | |
| "learning_rate": 6.800000000000001e-05, | |
| "loss": 1.3125, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.06062389640965274, | |
| "grad_norm": 52.5, | |
| "learning_rate": 6.866666666666666e-05, | |
| "loss": 1.3281, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.06121247792819305, | |
| "grad_norm": 122.0, | |
| "learning_rate": 6.933333333333334e-05, | |
| "loss": 1.1641, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.06180105944673337, | |
| "grad_norm": 13.375, | |
| "learning_rate": 7e-05, | |
| "loss": 1.0547, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.06238964096527369, | |
| "grad_norm": 97.5, | |
| "learning_rate": 7.066666666666667e-05, | |
| "loss": 1.3203, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.062978222483814, | |
| "grad_norm": 53.25, | |
| "learning_rate": 7.133333333333334e-05, | |
| "loss": 1.0547, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.06356680400235433, | |
| "grad_norm": 25.125, | |
| "learning_rate": 7.2e-05, | |
| "loss": 1.3906, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.06415538552089464, | |
| "grad_norm": 50.75, | |
| "learning_rate": 7.266666666666667e-05, | |
| "loss": 1.2109, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.06474396703943497, | |
| "grad_norm": 78.0, | |
| "learning_rate": 7.333333333333333e-05, | |
| "loss": 1.1406, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.06533254855797528, | |
| "grad_norm": 81.0, | |
| "learning_rate": 7.4e-05, | |
| "loss": 1.6875, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.0659211300765156, | |
| "grad_norm": 187.0, | |
| "learning_rate": 7.466666666666667e-05, | |
| "loss": 1.3125, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.06650971159505592, | |
| "grad_norm": 104.5, | |
| "learning_rate": 7.533333333333334e-05, | |
| "loss": 0.8438, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.06709829311359623, | |
| "grad_norm": 34.75, | |
| "learning_rate": 7.6e-05, | |
| "loss": 1.0703, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.06768687463213655, | |
| "grad_norm": 42.75, | |
| "learning_rate": 7.666666666666667e-05, | |
| "loss": 1.3359, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.06827545615067687, | |
| "grad_norm": 61.75, | |
| "learning_rate": 7.733333333333333e-05, | |
| "loss": 1.1562, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.06886403766921718, | |
| "grad_norm": 53.0, | |
| "learning_rate": 7.800000000000001e-05, | |
| "loss": 1.1797, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.06945261918775751, | |
| "grad_norm": 50.75, | |
| "learning_rate": 7.866666666666666e-05, | |
| "loss": 1.3906, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.07004120070629782, | |
| "grad_norm": 57.75, | |
| "learning_rate": 7.933333333333334e-05, | |
| "loss": 0.9883, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.07062978222483814, | |
| "grad_norm": 43.5, | |
| "learning_rate": 8e-05, | |
| "loss": 0.8828, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.07121836374337846, | |
| "grad_norm": 38.25, | |
| "learning_rate": 8.066666666666667e-05, | |
| "loss": 0.918, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.07180694526191878, | |
| "grad_norm": 37.75, | |
| "learning_rate": 8.133333333333334e-05, | |
| "loss": 1.1875, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.07239552678045909, | |
| "grad_norm": 21.125, | |
| "learning_rate": 8.2e-05, | |
| "loss": 0.7461, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.07298410829899941, | |
| "grad_norm": 45.0, | |
| "learning_rate": 8.266666666666667e-05, | |
| "loss": 0.8945, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.07357268981753973, | |
| "grad_norm": 130.0, | |
| "learning_rate": 8.333333333333334e-05, | |
| "loss": 1.1484, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.07416127133608005, | |
| "grad_norm": 41.25, | |
| "learning_rate": 8.4e-05, | |
| "loss": 1.375, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.07474985285462037, | |
| "grad_norm": 53.5, | |
| "learning_rate": 8.466666666666667e-05, | |
| "loss": 1.0469, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.07533843437316068, | |
| "grad_norm": 318.0, | |
| "learning_rate": 8.533333333333334e-05, | |
| "loss": 1.0703, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.075927015891701, | |
| "grad_norm": 30.125, | |
| "learning_rate": 8.6e-05, | |
| "loss": 1.0547, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.07651559741024132, | |
| "grad_norm": 170.0, | |
| "learning_rate": 8.666666666666667e-05, | |
| "loss": 0.8633, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.07710417892878163, | |
| "grad_norm": 26.75, | |
| "learning_rate": 8.733333333333333e-05, | |
| "loss": 0.6914, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.07769276044732196, | |
| "grad_norm": 64.5, | |
| "learning_rate": 8.800000000000001e-05, | |
| "loss": 1.0391, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.07828134196586227, | |
| "grad_norm": 67.5, | |
| "learning_rate": 8.866666666666668e-05, | |
| "loss": 0.9961, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.0788699234844026, | |
| "grad_norm": 62.0, | |
| "learning_rate": 8.933333333333334e-05, | |
| "loss": 0.9023, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.07945850500294291, | |
| "grad_norm": 40.0, | |
| "learning_rate": 9e-05, | |
| "loss": 0.7578, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.08004708652148322, | |
| "grad_norm": 63.5, | |
| "learning_rate": 9.066666666666667e-05, | |
| "loss": 1.0, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.08063566804002355, | |
| "grad_norm": 94.5, | |
| "learning_rate": 9.133333333333334e-05, | |
| "loss": 1.0859, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.08122424955856386, | |
| "grad_norm": 39.5, | |
| "learning_rate": 9.200000000000001e-05, | |
| "loss": 0.6836, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.08181283107710417, | |
| "grad_norm": 26.5, | |
| "learning_rate": 9.266666666666666e-05, | |
| "loss": 0.7031, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.0824014125956445, | |
| "grad_norm": 108.0, | |
| "learning_rate": 9.333333333333334e-05, | |
| "loss": 1.0938, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.08298999411418481, | |
| "grad_norm": 38.0, | |
| "learning_rate": 9.4e-05, | |
| "loss": 0.6289, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.08357857563272514, | |
| "grad_norm": 89.0, | |
| "learning_rate": 9.466666666666667e-05, | |
| "loss": 0.8203, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.08416715715126545, | |
| "grad_norm": 98.5, | |
| "learning_rate": 9.533333333333334e-05, | |
| "loss": 0.5508, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.08475573866980576, | |
| "grad_norm": 39.25, | |
| "learning_rate": 9.6e-05, | |
| "loss": 0.7773, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.08534432018834609, | |
| "grad_norm": 49.5, | |
| "learning_rate": 9.666666666666667e-05, | |
| "loss": 0.7852, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.0859329017068864, | |
| "grad_norm": 190.0, | |
| "learning_rate": 9.733333333333335e-05, | |
| "loss": 1.1328, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.08652148322542672, | |
| "grad_norm": 48.0, | |
| "learning_rate": 9.8e-05, | |
| "loss": 0.7305, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.08711006474396704, | |
| "grad_norm": 284.0, | |
| "learning_rate": 9.866666666666668e-05, | |
| "loss": 0.8867, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.08769864626250735, | |
| "grad_norm": 41.5, | |
| "learning_rate": 9.933333333333334e-05, | |
| "loss": 0.9023, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.08828722778104768, | |
| "grad_norm": 29.75, | |
| "learning_rate": 0.0001, | |
| "loss": 0.5391, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.088875809299588, | |
| "grad_norm": 35.25, | |
| "learning_rate": 0.00010066666666666667, | |
| "loss": 0.543, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.0894643908181283, | |
| "grad_norm": 117.0, | |
| "learning_rate": 0.00010133333333333335, | |
| "loss": 0.6094, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.09005297233666863, | |
| "grad_norm": 16.0, | |
| "learning_rate": 0.00010200000000000001, | |
| "loss": 0.4746, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.09064155385520895, | |
| "grad_norm": 62.25, | |
| "learning_rate": 0.00010266666666666666, | |
| "loss": 0.8164, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.09123013537374926, | |
| "grad_norm": 43.5, | |
| "learning_rate": 0.00010333333333333334, | |
| "loss": 0.5977, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.09181871689228958, | |
| "grad_norm": 16.375, | |
| "learning_rate": 0.00010400000000000001, | |
| "loss": 0.4199, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.0924072984108299, | |
| "grad_norm": 18.75, | |
| "learning_rate": 0.00010466666666666667, | |
| "loss": 0.3906, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.09299587992937022, | |
| "grad_norm": 44.25, | |
| "learning_rate": 0.00010533333333333332, | |
| "loss": 0.5742, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.09358446144791054, | |
| "grad_norm": 17.5, | |
| "learning_rate": 0.00010600000000000002, | |
| "loss": 0.4531, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.09417304296645085, | |
| "grad_norm": 35.25, | |
| "learning_rate": 0.00010666666666666667, | |
| "loss": 0.6992, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.09476162448499117, | |
| "grad_norm": 32.0, | |
| "learning_rate": 0.00010733333333333333, | |
| "loss": 0.5234, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.09535020600353149, | |
| "grad_norm": 15.5625, | |
| "learning_rate": 0.00010800000000000001, | |
| "loss": 0.3984, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.0959387875220718, | |
| "grad_norm": 14.875, | |
| "learning_rate": 0.00010866666666666667, | |
| "loss": 0.4062, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.09652736904061213, | |
| "grad_norm": 15.125, | |
| "learning_rate": 0.00010933333333333333, | |
| "loss": 0.4141, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.09711595055915244, | |
| "grad_norm": 33.0, | |
| "learning_rate": 0.00011000000000000002, | |
| "loss": 0.4551, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.09770453207769277, | |
| "grad_norm": 11.8125, | |
| "learning_rate": 0.00011066666666666667, | |
| "loss": 0.4023, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.09829311359623308, | |
| "grad_norm": 18.625, | |
| "learning_rate": 0.00011133333333333333, | |
| "loss": 0.4805, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.09888169511477339, | |
| "grad_norm": 8.5, | |
| "learning_rate": 0.00011200000000000001, | |
| "loss": 0.2812, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.09947027663331372, | |
| "grad_norm": 21.0, | |
| "learning_rate": 0.00011266666666666668, | |
| "loss": 0.4199, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.10005885815185403, | |
| "grad_norm": 29.875, | |
| "learning_rate": 0.00011333333333333334, | |
| "loss": 0.3711, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.10064743967039436, | |
| "grad_norm": 72.0, | |
| "learning_rate": 0.00011399999999999999, | |
| "loss": 0.375, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.10123602118893467, | |
| "grad_norm": 72.5, | |
| "learning_rate": 0.00011466666666666667, | |
| "loss": 0.4961, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.10182460270747498, | |
| "grad_norm": 19.5, | |
| "learning_rate": 0.00011533333333333334, | |
| "loss": 0.4766, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.10241318422601531, | |
| "grad_norm": 56.75, | |
| "learning_rate": 0.000116, | |
| "loss": 0.3223, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.10300176574455562, | |
| "grad_norm": 20.25, | |
| "learning_rate": 0.00011666666666666668, | |
| "loss": 0.2363, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.10359034726309593, | |
| "grad_norm": 26.75, | |
| "learning_rate": 0.00011733333333333334, | |
| "loss": 0.3164, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.10417892878163626, | |
| "grad_norm": 23.75, | |
| "learning_rate": 0.000118, | |
| "loss": 0.4141, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.10476751030017657, | |
| "grad_norm": 13.625, | |
| "learning_rate": 0.00011866666666666669, | |
| "loss": 0.3438, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.1053560918187169, | |
| "grad_norm": 20.25, | |
| "learning_rate": 0.00011933333333333334, | |
| "loss": 0.373, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.10594467333725721, | |
| "grad_norm": 17.875, | |
| "learning_rate": 0.00012, | |
| "loss": 0.4336, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.10653325485579752, | |
| "grad_norm": 8.875, | |
| "learning_rate": 0.00012066666666666668, | |
| "loss": 0.2002, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.10712183637433785, | |
| "grad_norm": 41.5, | |
| "learning_rate": 0.00012133333333333335, | |
| "loss": 0.2812, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.10771041789287816, | |
| "grad_norm": 24.375, | |
| "learning_rate": 0.000122, | |
| "loss": 0.2119, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.10829899941141848, | |
| "grad_norm": 8.5625, | |
| "learning_rate": 0.00012266666666666668, | |
| "loss": 0.1934, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.1088875809299588, | |
| "grad_norm": 540.0, | |
| "learning_rate": 0.00012333333333333334, | |
| "loss": 0.3203, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.10947616244849911, | |
| "grad_norm": 56.5, | |
| "learning_rate": 0.000124, | |
| "loss": 0.3809, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.11006474396703944, | |
| "grad_norm": 50.0, | |
| "learning_rate": 0.00012466666666666667, | |
| "loss": 0.207, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.11065332548557975, | |
| "grad_norm": 28.5, | |
| "learning_rate": 0.00012533333333333334, | |
| "loss": 0.3281, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.11124190700412007, | |
| "grad_norm": 20.375, | |
| "learning_rate": 0.000126, | |
| "loss": 0.373, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.11183048852266039, | |
| "grad_norm": 35.25, | |
| "learning_rate": 0.00012666666666666666, | |
| "loss": 0.3535, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.1124190700412007, | |
| "grad_norm": 23.5, | |
| "learning_rate": 0.00012733333333333336, | |
| "loss": 0.1885, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.11300765155974102, | |
| "grad_norm": 18.75, | |
| "learning_rate": 0.00012800000000000002, | |
| "loss": 0.4121, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.11359623307828134, | |
| "grad_norm": 96.0, | |
| "learning_rate": 0.00012866666666666666, | |
| "loss": 0.3574, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.11418481459682166, | |
| "grad_norm": 12.9375, | |
| "learning_rate": 0.00012933333333333332, | |
| "loss": 0.4023, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.11477339611536198, | |
| "grad_norm": 103.5, | |
| "learning_rate": 0.00013000000000000002, | |
| "loss": 0.3477, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.1153619776339023, | |
| "grad_norm": 13.3125, | |
| "learning_rate": 0.00013066666666666668, | |
| "loss": 0.3047, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.11595055915244261, | |
| "grad_norm": 96.5, | |
| "learning_rate": 0.00013133333333333332, | |
| "loss": 0.2773, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.11653914067098294, | |
| "grad_norm": 15.3125, | |
| "learning_rate": 0.000132, | |
| "loss": 0.2148, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.11712772218952325, | |
| "grad_norm": 24.125, | |
| "learning_rate": 0.00013266666666666667, | |
| "loss": 0.2461, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.11771630370806356, | |
| "grad_norm": 22.75, | |
| "learning_rate": 0.00013333333333333334, | |
| "loss": 0.2617, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.11830488522660389, | |
| "grad_norm": 23.625, | |
| "learning_rate": 0.000134, | |
| "loss": 0.1396, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.1188934667451442, | |
| "grad_norm": 7.21875, | |
| "learning_rate": 0.00013466666666666667, | |
| "loss": 0.1514, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.11948204826368453, | |
| "grad_norm": 37.0, | |
| "learning_rate": 0.00013533333333333333, | |
| "loss": 0.3223, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.12007062978222484, | |
| "grad_norm": 17.625, | |
| "learning_rate": 0.00013600000000000003, | |
| "loss": 0.2324, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.12065921130076515, | |
| "grad_norm": 11.875, | |
| "learning_rate": 0.00013666666666666666, | |
| "loss": 0.2051, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.12124779281930548, | |
| "grad_norm": 13.3125, | |
| "learning_rate": 0.00013733333333333333, | |
| "loss": 0.2793, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.12183637433784579, | |
| "grad_norm": 13.3125, | |
| "learning_rate": 0.000138, | |
| "loss": 0.2061, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.1224249558563861, | |
| "grad_norm": 16.75, | |
| "learning_rate": 0.00013866666666666669, | |
| "loss": 0.2148, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.12301353737492643, | |
| "grad_norm": 11.625, | |
| "learning_rate": 0.00013933333333333335, | |
| "loss": 0.2324, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.12360211889346674, | |
| "grad_norm": 17.5, | |
| "learning_rate": 0.00014, | |
| "loss": 0.1396, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.12419070041200707, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 0.00014066666666666668, | |
| "loss": 0.106, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.12477928193054738, | |
| "grad_norm": 72.0, | |
| "learning_rate": 0.00014133333333333334, | |
| "loss": 0.2012, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.1253678634490877, | |
| "grad_norm": 23.25, | |
| "learning_rate": 0.000142, | |
| "loss": 0.0991, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.125956444967628, | |
| "grad_norm": 21.25, | |
| "learning_rate": 0.00014266666666666667, | |
| "loss": 0.3809, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.12654502648616833, | |
| "grad_norm": 6.0625, | |
| "learning_rate": 0.00014333333333333334, | |
| "loss": 0.1206, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.12713360800470866, | |
| "grad_norm": 42.0, | |
| "learning_rate": 0.000144, | |
| "loss": 0.1445, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.12772218952324896, | |
| "grad_norm": 17.125, | |
| "learning_rate": 0.0001446666666666667, | |
| "loss": 0.2344, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.12831077104178928, | |
| "grad_norm": 19.875, | |
| "learning_rate": 0.00014533333333333333, | |
| "loss": 0.2266, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.1288993525603296, | |
| "grad_norm": 25.0, | |
| "learning_rate": 0.000146, | |
| "loss": 0.1045, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.12948793407886994, | |
| "grad_norm": 14.1875, | |
| "learning_rate": 0.00014666666666666666, | |
| "loss": 0.1592, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.13007651559741024, | |
| "grad_norm": 8.75, | |
| "learning_rate": 0.00014733333333333335, | |
| "loss": 0.1982, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.13066509711595056, | |
| "grad_norm": 9.1875, | |
| "learning_rate": 0.000148, | |
| "loss": 0.1436, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.1312536786344909, | |
| "grad_norm": 6.8125, | |
| "learning_rate": 0.00014866666666666666, | |
| "loss": 0.1172, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.1318422601530312, | |
| "grad_norm": 17.25, | |
| "learning_rate": 0.00014933333333333335, | |
| "loss": 0.123, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.13243084167157151, | |
| "grad_norm": 12.625, | |
| "learning_rate": 0.00015000000000000001, | |
| "loss": 0.1055, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.13301942319011184, | |
| "grad_norm": 7.46875, | |
| "learning_rate": 0.00015066666666666668, | |
| "loss": 0.1641, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.13360800470865214, | |
| "grad_norm": 8.4375, | |
| "learning_rate": 0.00015133333333333334, | |
| "loss": 0.1069, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.13419658622719247, | |
| "grad_norm": 15.5, | |
| "learning_rate": 0.000152, | |
| "loss": 0.1436, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.1347851677457328, | |
| "grad_norm": 18.625, | |
| "learning_rate": 0.00015266666666666667, | |
| "loss": 0.1172, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.1353737492642731, | |
| "grad_norm": 8.4375, | |
| "learning_rate": 0.00015333333333333334, | |
| "loss": 0.1631, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.13596233078281342, | |
| "grad_norm": 8.1875, | |
| "learning_rate": 0.000154, | |
| "loss": 0.1118, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.13655091230135374, | |
| "grad_norm": 7.71875, | |
| "learning_rate": 0.00015466666666666667, | |
| "loss": 0.2559, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.13713949381989404, | |
| "grad_norm": 28.875, | |
| "learning_rate": 0.00015533333333333333, | |
| "loss": 0.1641, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.13772807533843437, | |
| "grad_norm": 5.78125, | |
| "learning_rate": 0.00015600000000000002, | |
| "loss": 0.0732, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.1383166568569747, | |
| "grad_norm": 21.5, | |
| "learning_rate": 0.00015666666666666666, | |
| "loss": 0.248, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.13890523837551502, | |
| "grad_norm": 9.375, | |
| "learning_rate": 0.00015733333333333333, | |
| "loss": 0.1504, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.13949381989405532, | |
| "grad_norm": 8.5625, | |
| "learning_rate": 0.00015800000000000002, | |
| "loss": 0.1226, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.14008240141259565, | |
| "grad_norm": 2.9375, | |
| "learning_rate": 0.00015866666666666668, | |
| "loss": 0.106, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.14067098293113597, | |
| "grad_norm": 5.53125, | |
| "learning_rate": 0.00015933333333333332, | |
| "loss": 0.1328, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.14125956444967627, | |
| "grad_norm": 8.3125, | |
| "learning_rate": 0.00016, | |
| "loss": 0.1001, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.1418481459682166, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 0.00016066666666666668, | |
| "loss": 0.0588, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.14243672748675693, | |
| "grad_norm": 5.25, | |
| "learning_rate": 0.00016133333333333334, | |
| "loss": 0.1128, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.14302530900529722, | |
| "grad_norm": 14.375, | |
| "learning_rate": 0.000162, | |
| "loss": 0.293, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.14361389052383755, | |
| "grad_norm": 13.5, | |
| "learning_rate": 0.00016266666666666667, | |
| "loss": 0.0859, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.14420247204237788, | |
| "grad_norm": 21.375, | |
| "learning_rate": 0.00016333333333333334, | |
| "loss": 0.124, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.14479105356091818, | |
| "grad_norm": 11.625, | |
| "learning_rate": 0.000164, | |
| "loss": 0.1191, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.1453796350794585, | |
| "grad_norm": 5.28125, | |
| "learning_rate": 0.00016466666666666667, | |
| "loss": 0.0728, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.14596821659799883, | |
| "grad_norm": 10.3125, | |
| "learning_rate": 0.00016533333333333333, | |
| "loss": 0.1562, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.14655679811653913, | |
| "grad_norm": 2.53125, | |
| "learning_rate": 0.000166, | |
| "loss": 0.0752, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.14714537963507945, | |
| "grad_norm": 3.078125, | |
| "learning_rate": 0.0001666666666666667, | |
| "loss": 0.2041, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.14773396115361978, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 0.00016733333333333335, | |
| "loss": 0.0796, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.1483225426721601, | |
| "grad_norm": 13.9375, | |
| "learning_rate": 0.000168, | |
| "loss": 0.0781, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.1489111241907004, | |
| "grad_norm": 11.5625, | |
| "learning_rate": 0.00016866666666666668, | |
| "loss": 0.1162, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.14949970570924073, | |
| "grad_norm": 13.0625, | |
| "learning_rate": 0.00016933333333333335, | |
| "loss": 0.083, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.15008828722778106, | |
| "grad_norm": 2.765625, | |
| "learning_rate": 0.00017, | |
| "loss": 0.053, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.15067686874632136, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 0.00017066666666666668, | |
| "loss": 0.0806, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.15126545026486168, | |
| "grad_norm": 8.125, | |
| "learning_rate": 0.00017133333333333334, | |
| "loss": 0.0771, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.151854031783402, | |
| "grad_norm": 16.875, | |
| "learning_rate": 0.000172, | |
| "loss": 0.0752, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.1524426133019423, | |
| "grad_norm": 7.84375, | |
| "learning_rate": 0.00017266666666666667, | |
| "loss": 0.1406, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.15303119482048264, | |
| "grad_norm": 1.921875, | |
| "learning_rate": 0.00017333333333333334, | |
| "loss": 0.0603, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.15361977633902296, | |
| "grad_norm": 5.46875, | |
| "learning_rate": 0.000174, | |
| "loss": 0.0547, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.15420835785756326, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 0.00017466666666666667, | |
| "loss": 0.0525, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.1547969393761036, | |
| "grad_norm": 7.53125, | |
| "learning_rate": 0.00017533333333333336, | |
| "loss": 0.0713, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.1553855208946439, | |
| "grad_norm": 2.90625, | |
| "learning_rate": 0.00017600000000000002, | |
| "loss": 0.0537, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.1559741024131842, | |
| "grad_norm": 2.0625, | |
| "learning_rate": 0.00017666666666666666, | |
| "loss": 0.0698, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.15656268393172454, | |
| "grad_norm": 5.875, | |
| "learning_rate": 0.00017733333333333335, | |
| "loss": 0.0544, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.15715126545026487, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 0.00017800000000000002, | |
| "loss": 0.1099, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.1577398469688052, | |
| "grad_norm": 1.6171875, | |
| "learning_rate": 0.00017866666666666668, | |
| "loss": 0.0593, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.1583284284873455, | |
| "grad_norm": 20.625, | |
| "learning_rate": 0.00017933333333333332, | |
| "loss": 0.0728, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.15891701000588582, | |
| "grad_norm": 15.375, | |
| "learning_rate": 0.00018, | |
| "loss": 0.084, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.15950559152442614, | |
| "grad_norm": 2.09375, | |
| "learning_rate": 0.00018066666666666668, | |
| "loss": 0.0776, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.16009417304296644, | |
| "grad_norm": 8.0, | |
| "learning_rate": 0.00018133333333333334, | |
| "loss": 0.1187, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.16068275456150677, | |
| "grad_norm": 0.7421875, | |
| "learning_rate": 0.000182, | |
| "loss": 0.0398, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.1612713360800471, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 0.00018266666666666667, | |
| "loss": 0.053, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.1618599175985874, | |
| "grad_norm": 11.0, | |
| "learning_rate": 0.00018333333333333334, | |
| "loss": 0.0723, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.16244849911712772, | |
| "grad_norm": 2.21875, | |
| "learning_rate": 0.00018400000000000003, | |
| "loss": 0.1396, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.16303708063566805, | |
| "grad_norm": 0.4296875, | |
| "learning_rate": 0.00018466666666666666, | |
| "loss": 0.0291, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.16362566215420835, | |
| "grad_norm": 38.75, | |
| "learning_rate": 0.00018533333333333333, | |
| "loss": 0.0547, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.16421424367274867, | |
| "grad_norm": 0.62109375, | |
| "learning_rate": 0.00018600000000000002, | |
| "loss": 0.0403, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.164802825191289, | |
| "grad_norm": 2.640625, | |
| "learning_rate": 0.0001866666666666667, | |
| "loss": 0.063, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.16539140670982933, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 0.00018733333333333335, | |
| "loss": 0.0447, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.16597998822836962, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 0.000188, | |
| "loss": 0.0415, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.16656856974690995, | |
| "grad_norm": 1.453125, | |
| "learning_rate": 0.00018866666666666668, | |
| "loss": 0.0513, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.16715715126545028, | |
| "grad_norm": 9.0625, | |
| "learning_rate": 0.00018933333333333335, | |
| "loss": 0.0295, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.16774573278399058, | |
| "grad_norm": 17.75, | |
| "learning_rate": 0.00019, | |
| "loss": 0.0747, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.1683343143025309, | |
| "grad_norm": 10.6875, | |
| "learning_rate": 0.00019066666666666668, | |
| "loss": 0.0664, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.16892289582107123, | |
| "grad_norm": 13.8125, | |
| "learning_rate": 0.00019133333333333334, | |
| "loss": 0.052, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.16951147733961153, | |
| "grad_norm": 0.94140625, | |
| "learning_rate": 0.000192, | |
| "loss": 0.0444, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.17010005885815185, | |
| "grad_norm": 8.5, | |
| "learning_rate": 0.0001926666666666667, | |
| "loss": 0.0618, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.17068864037669218, | |
| "grad_norm": 4.59375, | |
| "learning_rate": 0.00019333333333333333, | |
| "loss": 0.0569, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.17127722189523248, | |
| "grad_norm": 2.78125, | |
| "learning_rate": 0.000194, | |
| "loss": 0.0598, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.1718658034137728, | |
| "grad_norm": 15.125, | |
| "learning_rate": 0.0001946666666666667, | |
| "loss": 0.0986, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.17245438493231313, | |
| "grad_norm": 0.33984375, | |
| "learning_rate": 0.00019533333333333336, | |
| "loss": 0.026, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.17304296645085343, | |
| "grad_norm": 1.6875, | |
| "learning_rate": 0.000196, | |
| "loss": 0.0369, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.17363154796939376, | |
| "grad_norm": 0.7890625, | |
| "learning_rate": 0.00019666666666666666, | |
| "loss": 0.031, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.17422012948793408, | |
| "grad_norm": 8.5, | |
| "learning_rate": 0.00019733333333333335, | |
| "loss": 0.0903, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.1748087110064744, | |
| "grad_norm": 1.015625, | |
| "learning_rate": 0.00019800000000000002, | |
| "loss": 0.0286, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.1753972925250147, | |
| "grad_norm": 1.90625, | |
| "learning_rate": 0.00019866666666666668, | |
| "loss": 0.1016, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.17598587404355504, | |
| "grad_norm": 0.361328125, | |
| "learning_rate": 0.00019933333333333334, | |
| "loss": 0.0239, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.17657445556209536, | |
| "grad_norm": 7.9375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0408, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.17716303708063566, | |
| "grad_norm": 1.71875, | |
| "learning_rate": 0.00019997559487492376, | |
| "loss": 0.0317, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.177751618599176, | |
| "grad_norm": 0.81640625, | |
| "learning_rate": 0.00019995118974984748, | |
| "loss": 0.0327, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.1783402001177163, | |
| "grad_norm": 2.21875, | |
| "learning_rate": 0.00019992678462477123, | |
| "loss": 0.0698, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.1789287816362566, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.00019990237949969495, | |
| "loss": 0.022, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.17951736315479694, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 0.0001998779743746187, | |
| "loss": 0.0231, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.18010594467333726, | |
| "grad_norm": 4.375, | |
| "learning_rate": 0.00019985356924954241, | |
| "loss": 0.0232, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.18069452619187756, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 0.00019982916412446616, | |
| "loss": 0.0403, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.1812831077104179, | |
| "grad_norm": 2.984375, | |
| "learning_rate": 0.00019980475899938988, | |
| "loss": 0.085, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.18187168922895822, | |
| "grad_norm": 1.0, | |
| "learning_rate": 0.00019978035387431363, | |
| "loss": 0.0304, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.18246027074749852, | |
| "grad_norm": 5.09375, | |
| "learning_rate": 0.00019975594874923735, | |
| "loss": 0.033, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.18304885226603884, | |
| "grad_norm": 0.26953125, | |
| "learning_rate": 0.0001997315436241611, | |
| "loss": 0.0184, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.18363743378457917, | |
| "grad_norm": 0.47265625, | |
| "learning_rate": 0.00019970713849908482, | |
| "loss": 0.0287, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.1842260153031195, | |
| "grad_norm": 0.466796875, | |
| "learning_rate": 0.00019968273337400857, | |
| "loss": 0.0228, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.1848145968216598, | |
| "grad_norm": 1.2734375, | |
| "learning_rate": 0.00019965832824893229, | |
| "loss": 0.0337, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.18540317834020012, | |
| "grad_norm": 25.25, | |
| "learning_rate": 0.00019963392312385603, | |
| "loss": 0.1172, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.18599175985874045, | |
| "grad_norm": 2.0, | |
| "learning_rate": 0.00019960951799877975, | |
| "loss": 0.0398, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.18658034137728075, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.0001995851128737035, | |
| "loss": 0.0248, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.18716892289582107, | |
| "grad_norm": 0.439453125, | |
| "learning_rate": 0.00019956070774862722, | |
| "loss": 0.0254, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.1877575044143614, | |
| "grad_norm": 0.244140625, | |
| "learning_rate": 0.00019953630262355097, | |
| "loss": 0.022, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.1883460859329017, | |
| "grad_norm": 2.671875, | |
| "learning_rate": 0.0001995118974984747, | |
| "loss": 0.0427, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.18893466745144202, | |
| "grad_norm": 1.0546875, | |
| "learning_rate": 0.00019948749237339844, | |
| "loss": 0.0415, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.18952324896998235, | |
| "grad_norm": 2.0625, | |
| "learning_rate": 0.00019946308724832216, | |
| "loss": 0.0806, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.19011183048852265, | |
| "grad_norm": 1.6484375, | |
| "learning_rate": 0.0001994386821232459, | |
| "loss": 0.0405, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.19070041200706297, | |
| "grad_norm": 2.625, | |
| "learning_rate": 0.00019941427699816963, | |
| "loss": 0.0258, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.1912889935256033, | |
| "grad_norm": 1.21875, | |
| "learning_rate": 0.00019938987187309337, | |
| "loss": 0.0344, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.1918775750441436, | |
| "grad_norm": 0.6171875, | |
| "learning_rate": 0.0001993654667480171, | |
| "loss": 0.026, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.19246615656268393, | |
| "grad_norm": 0.52734375, | |
| "learning_rate": 0.00019934106162294081, | |
| "loss": 0.0264, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.19305473808122425, | |
| "grad_norm": 0.8828125, | |
| "learning_rate": 0.00019931665649786456, | |
| "loss": 0.0262, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.19364331959976458, | |
| "grad_norm": 0.2265625, | |
| "learning_rate": 0.00019929225137278828, | |
| "loss": 0.0195, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.19423190111830488, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 0.00019926784624771203, | |
| "loss": 0.0374, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.1948204826368452, | |
| "grad_norm": 2.875, | |
| "learning_rate": 0.00019924344112263575, | |
| "loss": 0.0344, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.19540906415538553, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 0.0001992190359975595, | |
| "loss": 0.0199, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.19599764567392583, | |
| "grad_norm": 0.87109375, | |
| "learning_rate": 0.00019919463087248322, | |
| "loss": 0.033, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.19658622719246616, | |
| "grad_norm": 0.166015625, | |
| "learning_rate": 0.00019917022574740697, | |
| "loss": 0.0157, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.19717480871100648, | |
| "grad_norm": 1.5859375, | |
| "learning_rate": 0.00019914582062233069, | |
| "loss": 0.0266, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.19776339022954678, | |
| "grad_norm": 0.150390625, | |
| "learning_rate": 0.00019912141549725443, | |
| "loss": 0.0153, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.1983519717480871, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.00019909701037217815, | |
| "loss": 0.0197, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.19894055326662743, | |
| "grad_norm": 4.8125, | |
| "learning_rate": 0.0001990726052471019, | |
| "loss": 0.0471, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.19952913478516773, | |
| "grad_norm": 1.2890625, | |
| "learning_rate": 0.00019904820012202562, | |
| "loss": 0.0786, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.20011771630370806, | |
| "grad_norm": 11.5, | |
| "learning_rate": 0.00019902379499694937, | |
| "loss": 0.0413, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.2007062978222484, | |
| "grad_norm": 1.640625, | |
| "learning_rate": 0.0001989993898718731, | |
| "loss": 0.0508, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.2012948793407887, | |
| "grad_norm": 1.15625, | |
| "learning_rate": 0.00019897498474679684, | |
| "loss": 0.0297, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.201883460859329, | |
| "grad_norm": 0.384765625, | |
| "learning_rate": 0.00019895057962172056, | |
| "loss": 0.0261, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.20247204237786934, | |
| "grad_norm": 0.12890625, | |
| "learning_rate": 0.0001989261744966443, | |
| "loss": 0.0145, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.20306062389640966, | |
| "grad_norm": 5.0, | |
| "learning_rate": 0.00019890176937156803, | |
| "loss": 0.0239, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.20364920541494996, | |
| "grad_norm": 0.96484375, | |
| "learning_rate": 0.00019887736424649177, | |
| "loss": 0.0206, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.2042377869334903, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.0001988529591214155, | |
| "loss": 0.0181, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.20482636845203062, | |
| "grad_norm": 0.671875, | |
| "learning_rate": 0.00019882855399633924, | |
| "loss": 0.0226, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.20541494997057091, | |
| "grad_norm": 0.365234375, | |
| "learning_rate": 0.00019880414887126296, | |
| "loss": 0.0175, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.20600353148911124, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 0.0001987797437461867, | |
| "loss": 0.0181, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.20659211300765157, | |
| "grad_norm": 0.1279296875, | |
| "learning_rate": 0.00019875533862111043, | |
| "loss": 0.0135, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.20718069452619187, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 0.00019873093349603418, | |
| "loss": 0.0388, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.2077692760447322, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 0.0001987065283709579, | |
| "loss": 0.0172, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.20835785756327252, | |
| "grad_norm": 7.71875, | |
| "learning_rate": 0.00019868212324588165, | |
| "loss": 0.0327, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.20894643908181282, | |
| "grad_norm": 0.1376953125, | |
| "learning_rate": 0.0001986577181208054, | |
| "loss": 0.0153, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.20953502060035314, | |
| "grad_norm": 2.859375, | |
| "learning_rate": 0.00019863331299572911, | |
| "loss": 0.0693, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.21012360211889347, | |
| "grad_norm": 0.1376953125, | |
| "learning_rate": 0.00019860890787065286, | |
| "loss": 0.0157, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.2107121836374338, | |
| "grad_norm": 0.3046875, | |
| "learning_rate": 0.00019858450274557658, | |
| "loss": 0.022, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.2113007651559741, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 0.00019856009762050033, | |
| "loss": 0.0205, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.21188934667451442, | |
| "grad_norm": 0.376953125, | |
| "learning_rate": 0.00019853569249542405, | |
| "loss": 0.0226, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.21247792819305475, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.0001985112873703478, | |
| "loss": 0.0165, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.21306650971159505, | |
| "grad_norm": 0.134765625, | |
| "learning_rate": 0.00019848688224527152, | |
| "loss": 0.0166, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.21365509123013537, | |
| "grad_norm": 0.134765625, | |
| "learning_rate": 0.00019846247712019527, | |
| "loss": 0.0136, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.2142436727486757, | |
| "grad_norm": 0.71875, | |
| "learning_rate": 0.00019843807199511899, | |
| "loss": 0.0369, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.214832254267216, | |
| "grad_norm": 9.0, | |
| "learning_rate": 0.00019841366687004273, | |
| "loss": 0.0464, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.21542083578575633, | |
| "grad_norm": 0.361328125, | |
| "learning_rate": 0.00019838926174496645, | |
| "loss": 0.0149, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.21600941730429665, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.0001983648566198902, | |
| "loss": 0.0188, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.21659799882283695, | |
| "grad_norm": 0.1591796875, | |
| "learning_rate": 0.00019834045149481392, | |
| "loss": 0.0172, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.21718658034137728, | |
| "grad_norm": 0.1708984375, | |
| "learning_rate": 0.00019831604636973767, | |
| "loss": 0.0168, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.2177751618599176, | |
| "grad_norm": 0.1318359375, | |
| "learning_rate": 0.0001982916412446614, | |
| "loss": 0.015, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.2183637433784579, | |
| "grad_norm": 2.734375, | |
| "learning_rate": 0.00019826723611958514, | |
| "loss": 0.0378, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.21895232489699823, | |
| "grad_norm": 1.234375, | |
| "learning_rate": 0.00019824283099450886, | |
| "loss": 0.0237, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.21954090641553856, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 0.0001982184258694326, | |
| "loss": 0.0156, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.22012948793407888, | |
| "grad_norm": 0.376953125, | |
| "learning_rate": 0.00019819402074435633, | |
| "loss": 0.0204, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.22071806945261918, | |
| "grad_norm": 1.453125, | |
| "learning_rate": 0.00019816961561928007, | |
| "loss": 0.0189, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.2213066509711595, | |
| "grad_norm": 0.64453125, | |
| "learning_rate": 0.0001981452104942038, | |
| "loss": 0.0208, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.22189523248969983, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.00019812080536912751, | |
| "loss": 0.0203, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.22248381400824013, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 0.00019809640024405126, | |
| "loss": 0.0175, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.22307239552678046, | |
| "grad_norm": 0.138671875, | |
| "learning_rate": 0.00019807199511897498, | |
| "loss": 0.0154, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.22366097704532079, | |
| "grad_norm": 2.328125, | |
| "learning_rate": 0.00019804758999389873, | |
| "loss": 0.0164, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.22424955856386108, | |
| "grad_norm": 0.53515625, | |
| "learning_rate": 0.00019802318486882245, | |
| "loss": 0.0175, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.2248381400824014, | |
| "grad_norm": 0.427734375, | |
| "learning_rate": 0.0001979987797437462, | |
| "loss": 0.0154, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.22542672160094174, | |
| "grad_norm": 0.1328125, | |
| "learning_rate": 0.00019797437461866992, | |
| "loss": 0.014, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.22601530311948204, | |
| "grad_norm": 0.166015625, | |
| "learning_rate": 0.00019794996949359367, | |
| "loss": 0.0153, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.22660388463802236, | |
| "grad_norm": 0.28515625, | |
| "learning_rate": 0.0001979255643685174, | |
| "loss": 0.0172, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.2271924661565627, | |
| "grad_norm": 0.0830078125, | |
| "learning_rate": 0.00019790115924344113, | |
| "loss": 0.0115, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.22778104767510302, | |
| "grad_norm": 0.11474609375, | |
| "learning_rate": 0.00019787675411836485, | |
| "loss": 0.0132, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.22836962919364331, | |
| "grad_norm": 1.4609375, | |
| "learning_rate": 0.0001978523489932886, | |
| "loss": 0.0234, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.22895821071218364, | |
| "grad_norm": 0.10498046875, | |
| "learning_rate": 0.00019782794386821232, | |
| "loss": 0.0125, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.22954679223072397, | |
| "grad_norm": 0.1328125, | |
| "learning_rate": 0.00019780353874313607, | |
| "loss": 0.0156, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.23013537374926427, | |
| "grad_norm": 1.15625, | |
| "learning_rate": 0.0001977791336180598, | |
| "loss": 0.0522, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.2307239552678046, | |
| "grad_norm": 0.154296875, | |
| "learning_rate": 0.00019775472849298354, | |
| "loss": 0.0151, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.23131253678634492, | |
| "grad_norm": 0.1591796875, | |
| "learning_rate": 0.00019773032336790726, | |
| "loss": 0.016, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.23190111830488522, | |
| "grad_norm": 1.1953125, | |
| "learning_rate": 0.000197705918242831, | |
| "loss": 0.0198, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.23248969982342554, | |
| "grad_norm": 0.1513671875, | |
| "learning_rate": 0.00019768151311775473, | |
| "loss": 0.0133, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.23307828134196587, | |
| "grad_norm": 0.1259765625, | |
| "learning_rate": 0.00019765710799267847, | |
| "loss": 0.0128, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.23366686286050617, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.0001976327028676022, | |
| "loss": 0.0173, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.2342554443790465, | |
| "grad_norm": 1.5625, | |
| "learning_rate": 0.00019760829774252594, | |
| "loss": 0.0593, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.23484402589758682, | |
| "grad_norm": 0.1376953125, | |
| "learning_rate": 0.00019758389261744966, | |
| "loss": 0.0148, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.23543260741612712, | |
| "grad_norm": 0.189453125, | |
| "learning_rate": 0.0001975594874923734, | |
| "loss": 0.0144, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.23602118893466745, | |
| "grad_norm": 0.10009765625, | |
| "learning_rate": 0.00019753508236729713, | |
| "loss": 0.0127, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.23660977045320777, | |
| "grad_norm": 0.12890625, | |
| "learning_rate": 0.00019751067724222088, | |
| "loss": 0.0137, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.2371983519717481, | |
| "grad_norm": 0.10888671875, | |
| "learning_rate": 0.0001974862721171446, | |
| "loss": 0.0126, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.2377869334902884, | |
| "grad_norm": 0.119140625, | |
| "learning_rate": 0.00019746186699206835, | |
| "loss": 0.0149, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.23837551500882873, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.00019743746186699207, | |
| "loss": 0.0168, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.23896409652736905, | |
| "grad_norm": 0.2265625, | |
| "learning_rate": 0.00019741305674191581, | |
| "loss": 0.0188, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.23955267804590935, | |
| "grad_norm": 0.11083984375, | |
| "learning_rate": 0.00019738865161683954, | |
| "loss": 0.0125, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.24014125956444968, | |
| "grad_norm": 0.84765625, | |
| "learning_rate": 0.00019736424649176328, | |
| "loss": 0.0361, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.24072984108299, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 0.00019733984136668703, | |
| "loss": 0.014, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.2413184226015303, | |
| "grad_norm": 0.1171875, | |
| "learning_rate": 0.00019731543624161075, | |
| "loss": 0.0139, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.24190700412007063, | |
| "grad_norm": 0.08544921875, | |
| "learning_rate": 0.0001972910311165345, | |
| "loss": 0.0115, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.24249558563861096, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.00019726662599145822, | |
| "loss": 0.0173, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.24308416715715125, | |
| "grad_norm": 0.08935546875, | |
| "learning_rate": 0.00019724222086638197, | |
| "loss": 0.012, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.24367274867569158, | |
| "grad_norm": 1.0859375, | |
| "learning_rate": 0.0001972178157413057, | |
| "loss": 0.0273, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.2442613301942319, | |
| "grad_norm": 0.08740234375, | |
| "learning_rate": 0.00019719341061622943, | |
| "loss": 0.0117, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.2448499117127722, | |
| "grad_norm": 0.0751953125, | |
| "learning_rate": 0.00019716900549115315, | |
| "loss": 0.0109, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.24543849323131253, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 0.0001971446003660769, | |
| "loss": 0.0188, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.24602707474985286, | |
| "grad_norm": 0.12890625, | |
| "learning_rate": 0.00019712019524100062, | |
| "loss": 0.015, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.24661565626839319, | |
| "grad_norm": 0.138671875, | |
| "learning_rate": 0.00019709579011592437, | |
| "loss": 0.014, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.24720423778693348, | |
| "grad_norm": 0.10546875, | |
| "learning_rate": 0.0001970713849908481, | |
| "loss": 0.0106, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.2477928193054738, | |
| "grad_norm": 0.083984375, | |
| "learning_rate": 0.00019704697986577184, | |
| "loss": 0.0115, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.24838140082401414, | |
| "grad_norm": 0.314453125, | |
| "learning_rate": 0.00019702257474069556, | |
| "loss": 0.0153, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.24896998234255444, | |
| "grad_norm": 0.23046875, | |
| "learning_rate": 0.0001969981696156193, | |
| "loss": 0.0181, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.24955856386109476, | |
| "grad_norm": 0.1171875, | |
| "learning_rate": 0.00019697376449054303, | |
| "loss": 0.0134, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.25014714537963506, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 0.00019694935936546677, | |
| "loss": 0.019, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.2507357268981754, | |
| "grad_norm": 0.177734375, | |
| "learning_rate": 0.0001969249542403905, | |
| "loss": 0.0149, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.2513243084167157, | |
| "grad_norm": 1.15625, | |
| "learning_rate": 0.00019690054911531422, | |
| "loss": 0.0165, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.251912889935256, | |
| "grad_norm": 0.84375, | |
| "learning_rate": 0.00019687614399023794, | |
| "loss": 0.0131, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.25250147145379637, | |
| "grad_norm": 0.115234375, | |
| "learning_rate": 0.00019685173886516168, | |
| "loss": 0.0109, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.25309005297233667, | |
| "grad_norm": 0.67578125, | |
| "learning_rate": 0.00019682733374008543, | |
| "loss": 0.0229, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.25367863449087696, | |
| "grad_norm": 0.484375, | |
| "learning_rate": 0.00019680292861500915, | |
| "loss": 0.0172, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.2542672160094173, | |
| "grad_norm": 0.1484375, | |
| "learning_rate": 0.0001967785234899329, | |
| "loss": 0.0138, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.2548557975279576, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 0.00019675411836485662, | |
| "loss": 0.017, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.2554443790464979, | |
| "grad_norm": 0.11328125, | |
| "learning_rate": 0.00019672971323978037, | |
| "loss": 0.0125, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.25603296056503827, | |
| "grad_norm": 0.1474609375, | |
| "learning_rate": 0.0001967053081147041, | |
| "loss": 0.0131, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.25662154208357857, | |
| "grad_norm": 0.1533203125, | |
| "learning_rate": 0.00019668090298962784, | |
| "loss": 0.0147, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.25721012360211887, | |
| "grad_norm": 0.0966796875, | |
| "learning_rate": 0.00019665649786455156, | |
| "loss": 0.0122, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.2577987051206592, | |
| "grad_norm": 0.11474609375, | |
| "learning_rate": 0.0001966320927394753, | |
| "loss": 0.014, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.2583872866391995, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.00019660768761439902, | |
| "loss": 0.0097, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.2589758681577399, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 0.00019658328248932277, | |
| "loss": 0.0133, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.2595644496762802, | |
| "grad_norm": 0.07421875, | |
| "learning_rate": 0.0001965588773642465, | |
| "loss": 0.0104, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.26015303119482047, | |
| "grad_norm": 0.08154296875, | |
| "learning_rate": 0.00019653447223917024, | |
| "loss": 0.0123, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.2607416127133608, | |
| "grad_norm": 0.1005859375, | |
| "learning_rate": 0.00019651006711409396, | |
| "loss": 0.0117, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.2613301942319011, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 0.0001964856619890177, | |
| "loss": 0.0182, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.2619187757504414, | |
| "grad_norm": 0.0712890625, | |
| "learning_rate": 0.00019646125686394143, | |
| "loss": 0.0111, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.2625073572689818, | |
| "grad_norm": 0.52734375, | |
| "learning_rate": 0.00019643685173886518, | |
| "loss": 0.0143, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.2630959387875221, | |
| "grad_norm": 0.150390625, | |
| "learning_rate": 0.0001964124466137889, | |
| "loss": 0.0293, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.2636845203060624, | |
| "grad_norm": 0.2333984375, | |
| "learning_rate": 0.00019638804148871264, | |
| "loss": 0.0152, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.26427310182460273, | |
| "grad_norm": 0.166015625, | |
| "learning_rate": 0.00019636363636363636, | |
| "loss": 0.0135, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.26486168334314303, | |
| "grad_norm": 0.08447265625, | |
| "learning_rate": 0.0001963392312385601, | |
| "loss": 0.011, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.2654502648616833, | |
| "grad_norm": 0.359375, | |
| "learning_rate": 0.00019631482611348383, | |
| "loss": 0.0122, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.2660388463802237, | |
| "grad_norm": 0.076171875, | |
| "learning_rate": 0.00019629042098840758, | |
| "loss": 0.0107, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.266627427898764, | |
| "grad_norm": 0.53125, | |
| "learning_rate": 0.0001962660158633313, | |
| "loss": 0.0151, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.2672160094173043, | |
| "grad_norm": 0.359375, | |
| "learning_rate": 0.00019624161073825505, | |
| "loss": 0.016, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.26780459093584463, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.00019621720561317877, | |
| "loss": 0.0182, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.26839317245438493, | |
| "grad_norm": 1.0390625, | |
| "learning_rate": 0.00019619280048810252, | |
| "loss": 0.0123, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.26898175397292523, | |
| "grad_norm": 0.099609375, | |
| "learning_rate": 0.00019616839536302624, | |
| "loss": 0.013, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.2695703354914656, | |
| "grad_norm": 1.0625, | |
| "learning_rate": 0.00019614399023794998, | |
| "loss": 0.0222, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.2701589170100059, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 0.0001961195851128737, | |
| "loss": 0.0125, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.2707474985285462, | |
| "grad_norm": 0.15625, | |
| "learning_rate": 0.00019609517998779745, | |
| "loss": 0.013, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.27133608004708654, | |
| "grad_norm": 0.10498046875, | |
| "learning_rate": 0.00019607077486272117, | |
| "loss": 0.0107, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.27192466156562684, | |
| "grad_norm": 1.0703125, | |
| "learning_rate": 0.00019604636973764492, | |
| "loss": 0.0153, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.27251324308416713, | |
| "grad_norm": 0.173828125, | |
| "learning_rate": 0.00019602196461256867, | |
| "loss": 0.0139, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.2731018246027075, | |
| "grad_norm": 1.390625, | |
| "learning_rate": 0.0001959975594874924, | |
| "loss": 0.019, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.2736904061212478, | |
| "grad_norm": 0.1474609375, | |
| "learning_rate": 0.00019597315436241613, | |
| "loss": 0.0128, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.2742789876397881, | |
| "grad_norm": 0.11962890625, | |
| "learning_rate": 0.00019594874923733986, | |
| "loss": 0.012, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.27486756915832844, | |
| "grad_norm": 0.248046875, | |
| "learning_rate": 0.0001959243441122636, | |
| "loss": 0.0151, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.27545615067686874, | |
| "grad_norm": 6.1875, | |
| "learning_rate": 0.00019589993898718732, | |
| "loss": 0.0189, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.27604473219540904, | |
| "grad_norm": 0.08349609375, | |
| "learning_rate": 0.00019587553386211107, | |
| "loss": 0.0112, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.2766333137139494, | |
| "grad_norm": 0.1181640625, | |
| "learning_rate": 0.0001958511287370348, | |
| "loss": 0.0143, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.2772218952324897, | |
| "grad_norm": 0.388671875, | |
| "learning_rate": 0.00019582672361195854, | |
| "loss": 0.0141, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.27781047675103004, | |
| "grad_norm": 0.08837890625, | |
| "learning_rate": 0.00019580231848688226, | |
| "loss": 0.011, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.27839905826957034, | |
| "grad_norm": 0.07373046875, | |
| "learning_rate": 0.000195777913361806, | |
| "loss": 0.0106, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.27898763978811064, | |
| "grad_norm": 0.2373046875, | |
| "learning_rate": 0.00019575350823672973, | |
| "loss": 0.0139, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.279576221306651, | |
| "grad_norm": 0.1025390625, | |
| "learning_rate": 0.00019572910311165348, | |
| "loss": 0.0119, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.2801648028251913, | |
| "grad_norm": 0.09814453125, | |
| "learning_rate": 0.0001957046979865772, | |
| "loss": 0.0122, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.2807533843437316, | |
| "grad_norm": 0.09619140625, | |
| "learning_rate": 0.00019568029286150094, | |
| "loss": 0.0113, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.28134196586227195, | |
| "grad_norm": 0.08203125, | |
| "learning_rate": 0.00019565588773642464, | |
| "loss": 0.0121, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.28193054738081225, | |
| "grad_norm": 0.59375, | |
| "learning_rate": 0.00019563148261134838, | |
| "loss": 0.013, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.28251912889935255, | |
| "grad_norm": 0.2734375, | |
| "learning_rate": 0.0001956070774862721, | |
| "loss": 0.0159, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.2831077104178929, | |
| "grad_norm": 0.1201171875, | |
| "learning_rate": 0.00019558267236119585, | |
| "loss": 0.0125, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.2836962919364332, | |
| "grad_norm": 0.890625, | |
| "learning_rate": 0.00019555826723611957, | |
| "loss": 0.0294, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.2842848734549735, | |
| "grad_norm": 0.08740234375, | |
| "learning_rate": 0.00019553386211104332, | |
| "loss": 0.011, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.28487345497351385, | |
| "grad_norm": 0.39453125, | |
| "learning_rate": 0.00019550945698596707, | |
| "loss": 0.0182, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.28546203649205415, | |
| "grad_norm": 0.52734375, | |
| "learning_rate": 0.0001954850518608908, | |
| "loss": 0.0165, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.28605061801059445, | |
| "grad_norm": 0.15625, | |
| "learning_rate": 0.00019546064673581454, | |
| "loss": 0.03, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.2866391995291348, | |
| "grad_norm": 0.1298828125, | |
| "learning_rate": 0.00019543624161073826, | |
| "loss": 0.0118, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.2872277810476751, | |
| "grad_norm": 0.10888671875, | |
| "learning_rate": 0.000195411836485662, | |
| "loss": 0.0121, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.2878163625662154, | |
| "grad_norm": 1.890625, | |
| "learning_rate": 0.00019538743136058572, | |
| "loss": 0.0276, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.28840494408475575, | |
| "grad_norm": 0.11962890625, | |
| "learning_rate": 0.00019536302623550947, | |
| "loss": 0.0146, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.28899352560329605, | |
| "grad_norm": 0.1044921875, | |
| "learning_rate": 0.0001953386211104332, | |
| "loss": 0.0118, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.28958210712183635, | |
| "grad_norm": 0.0888671875, | |
| "learning_rate": 0.00019531421598535694, | |
| "loss": 0.0112, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.2901706886403767, | |
| "grad_norm": 0.2333984375, | |
| "learning_rate": 0.00019528981086028066, | |
| "loss": 0.0107, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.290759270158917, | |
| "grad_norm": 0.09033203125, | |
| "learning_rate": 0.0001952654057352044, | |
| "loss": 0.0112, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.2913478516774573, | |
| "grad_norm": 0.08837890625, | |
| "learning_rate": 0.00019524100061012813, | |
| "loss": 0.0117, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.29193643319599766, | |
| "grad_norm": 0.0966796875, | |
| "learning_rate": 0.00019521659548505188, | |
| "loss": 0.0117, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.29252501471453796, | |
| "grad_norm": 1.34375, | |
| "learning_rate": 0.0001951921903599756, | |
| "loss": 0.0131, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.29311359623307826, | |
| "grad_norm": 0.171875, | |
| "learning_rate": 0.00019516778523489934, | |
| "loss": 0.0114, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.2937021777516186, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 0.00019514338010982306, | |
| "loss": 0.0117, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.2942907592701589, | |
| "grad_norm": 0.08203125, | |
| "learning_rate": 0.0001951189749847468, | |
| "loss": 0.0114, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.29487934078869926, | |
| "grad_norm": 0.10205078125, | |
| "learning_rate": 0.00019509456985967053, | |
| "loss": 0.012, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.29546792230723956, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 0.00019507016473459428, | |
| "loss": 0.014, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.29605650382577986, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.000195045759609518, | |
| "loss": 0.0193, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.2966450853443202, | |
| "grad_norm": 0.06396484375, | |
| "learning_rate": 0.00019502135448444175, | |
| "loss": 0.0114, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.2972336668628605, | |
| "grad_norm": 0.08740234375, | |
| "learning_rate": 0.00019499694935936547, | |
| "loss": 0.0111, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.2978222483814008, | |
| "grad_norm": 0.11767578125, | |
| "learning_rate": 0.00019497254423428922, | |
| "loss": 0.0135, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.29841082989994117, | |
| "grad_norm": 0.1640625, | |
| "learning_rate": 0.00019494813910921294, | |
| "loss": 0.0125, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.29899941141848146, | |
| "grad_norm": 8.3125, | |
| "learning_rate": 0.00019492373398413668, | |
| "loss": 0.0181, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.29958799293702176, | |
| "grad_norm": 0.083984375, | |
| "learning_rate": 0.0001948993288590604, | |
| "loss": 0.011, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.3001765744555621, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 0.00019487492373398415, | |
| "loss": 0.0129, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.3007651559741024, | |
| "grad_norm": 0.09619140625, | |
| "learning_rate": 0.00019485051860890787, | |
| "loss": 0.0108, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.3013537374926427, | |
| "grad_norm": 0.1005859375, | |
| "learning_rate": 0.00019482611348383162, | |
| "loss": 0.0129, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.30194231901118307, | |
| "grad_norm": 0.08447265625, | |
| "learning_rate": 0.00019480170835875534, | |
| "loss": 0.0109, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.30253090052972337, | |
| "grad_norm": 0.055419921875, | |
| "learning_rate": 0.0001947773032336791, | |
| "loss": 0.0097, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.30311948204826367, | |
| "grad_norm": 0.09716796875, | |
| "learning_rate": 0.0001947528981086028, | |
| "loss": 0.0133, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.303708063566804, | |
| "grad_norm": 2.546875, | |
| "learning_rate": 0.00019472849298352656, | |
| "loss": 0.0112, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.3042966450853443, | |
| "grad_norm": 0.1318359375, | |
| "learning_rate": 0.0001947040878584503, | |
| "loss": 0.0129, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.3048852266038846, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 0.00019467968273337402, | |
| "loss": 0.0147, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.305473808122425, | |
| "grad_norm": 0.08349609375, | |
| "learning_rate": 0.00019465527760829777, | |
| "loss": 0.0116, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.30606238964096527, | |
| "grad_norm": 0.08837890625, | |
| "learning_rate": 0.0001946308724832215, | |
| "loss": 0.0106, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.30665097115950557, | |
| "grad_norm": 0.0908203125, | |
| "learning_rate": 0.00019460646735814524, | |
| "loss": 0.0113, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.3072395526780459, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 0.00019458206223306896, | |
| "loss": 0.0187, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.3078281341965862, | |
| "grad_norm": 0.48828125, | |
| "learning_rate": 0.0001945576571079927, | |
| "loss": 0.0226, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.3084167157151265, | |
| "grad_norm": 0.1328125, | |
| "learning_rate": 0.00019453325198291643, | |
| "loss": 0.0149, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.3090052972336669, | |
| "grad_norm": 0.08203125, | |
| "learning_rate": 0.00019450884685784018, | |
| "loss": 0.0112, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.3095938787522072, | |
| "grad_norm": 0.054443359375, | |
| "learning_rate": 0.0001944844417327639, | |
| "loss": 0.0103, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.3101824602707475, | |
| "grad_norm": 0.609375, | |
| "learning_rate": 0.00019446003660768764, | |
| "loss": 0.0234, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.3107710417892878, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 0.00019443563148261134, | |
| "loss": 0.0108, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.3113596233078281, | |
| "grad_norm": 0.2216796875, | |
| "learning_rate": 0.00019441122635753508, | |
| "loss": 0.0138, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.3119482048263684, | |
| "grad_norm": 0.125, | |
| "learning_rate": 0.0001943868212324588, | |
| "loss": 0.0107, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.3125367863449088, | |
| "grad_norm": 0.1357421875, | |
| "learning_rate": 0.00019436241610738255, | |
| "loss": 0.0128, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.3131253678634491, | |
| "grad_norm": 0.10986328125, | |
| "learning_rate": 0.00019433801098230627, | |
| "loss": 0.0118, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.31371394938198943, | |
| "grad_norm": 0.09326171875, | |
| "learning_rate": 0.00019431360585723002, | |
| "loss": 0.0113, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.31430253090052973, | |
| "grad_norm": 0.125, | |
| "learning_rate": 0.00019428920073215374, | |
| "loss": 0.0124, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.31489111241907003, | |
| "grad_norm": 0.142578125, | |
| "learning_rate": 0.0001942647956070775, | |
| "loss": 0.0123, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.3154796939376104, | |
| "grad_norm": 0.07275390625, | |
| "learning_rate": 0.0001942403904820012, | |
| "loss": 0.0104, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.3160682754561507, | |
| "grad_norm": 0.10009765625, | |
| "learning_rate": 0.00019421598535692496, | |
| "loss": 0.012, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.316656856974691, | |
| "grad_norm": 0.279296875, | |
| "learning_rate": 0.0001941915802318487, | |
| "loss": 0.0141, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.31724543849323134, | |
| "grad_norm": 0.72265625, | |
| "learning_rate": 0.00019416717510677242, | |
| "loss": 0.0271, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.31783402001177163, | |
| "grad_norm": 0.08203125, | |
| "learning_rate": 0.00019414276998169617, | |
| "loss": 0.0114, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.31842260153031193, | |
| "grad_norm": 1.4765625, | |
| "learning_rate": 0.0001941183648566199, | |
| "loss": 0.0378, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.3190111830488523, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.00019409395973154364, | |
| "loss": 0.0176, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.3195997645673926, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 0.00019406955460646736, | |
| "loss": 0.0148, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.3201883460859329, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.0001940451494813911, | |
| "loss": 0.0144, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.32077692760447324, | |
| "grad_norm": 0.11669921875, | |
| "learning_rate": 0.00019402074435631483, | |
| "loss": 0.0121, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.32136550912301354, | |
| "grad_norm": 0.130859375, | |
| "learning_rate": 0.00019399633923123858, | |
| "loss": 0.0131, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.32195409064155384, | |
| "grad_norm": 0.10107421875, | |
| "learning_rate": 0.0001939719341061623, | |
| "loss": 0.0115, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.3225426721600942, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 0.00019394752898108604, | |
| "loss": 0.0135, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.3231312536786345, | |
| "grad_norm": 0.62890625, | |
| "learning_rate": 0.00019392312385600976, | |
| "loss": 0.0154, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.3237198351971748, | |
| "grad_norm": 0.08935546875, | |
| "learning_rate": 0.0001938987187309335, | |
| "loss": 0.0112, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.32430841671571514, | |
| "grad_norm": 0.0908203125, | |
| "learning_rate": 0.00019387431360585723, | |
| "loss": 0.0118, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.32489699823425544, | |
| "grad_norm": 17.375, | |
| "learning_rate": 0.00019384990848078098, | |
| "loss": 0.0457, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.32548557975279574, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 0.0001938255033557047, | |
| "loss": 0.0131, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.3260741612713361, | |
| "grad_norm": 0.15234375, | |
| "learning_rate": 0.00019380109823062845, | |
| "loss": 0.0132, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.3266627427898764, | |
| "grad_norm": 0.220703125, | |
| "learning_rate": 0.00019377669310555217, | |
| "loss": 0.0148, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.3272513243084167, | |
| "grad_norm": 0.08251953125, | |
| "learning_rate": 0.00019375228798047592, | |
| "loss": 0.0113, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.32783990582695705, | |
| "grad_norm": 0.52734375, | |
| "learning_rate": 0.00019372788285539964, | |
| "loss": 0.012, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.32842848734549734, | |
| "grad_norm": 0.083984375, | |
| "learning_rate": 0.00019370347773032338, | |
| "loss": 0.0106, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.32901706886403764, | |
| "grad_norm": 0.2021484375, | |
| "learning_rate": 0.0001936790726052471, | |
| "loss": 0.0145, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.329605650382578, | |
| "grad_norm": 0.0859375, | |
| "learning_rate": 0.00019365466748017085, | |
| "loss": 0.0099, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3301942319011183, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.00019363026235509457, | |
| "loss": 0.0114, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.33078281341965865, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 0.00019360585723001832, | |
| "loss": 0.015, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.33137139493819895, | |
| "grad_norm": 0.0810546875, | |
| "learning_rate": 0.00019358145210494204, | |
| "loss": 0.0107, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.33195997645673925, | |
| "grad_norm": 0.0625, | |
| "learning_rate": 0.0001935570469798658, | |
| "loss": 0.0106, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.3325485579752796, | |
| "grad_norm": 0.056396484375, | |
| "learning_rate": 0.0001935326418547895, | |
| "loss": 0.0087, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.3331371394938199, | |
| "grad_norm": 0.08642578125, | |
| "learning_rate": 0.00019350823672971326, | |
| "loss": 0.0125, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.3337257210123602, | |
| "grad_norm": 0.23828125, | |
| "learning_rate": 0.00019348383160463698, | |
| "loss": 0.0127, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.33431430253090055, | |
| "grad_norm": 0.5390625, | |
| "learning_rate": 0.00019345942647956072, | |
| "loss": 0.0187, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.33490288404944085, | |
| "grad_norm": 1.015625, | |
| "learning_rate": 0.00019343502135448445, | |
| "loss": 0.0132, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.33549146556798115, | |
| "grad_norm": 0.10302734375, | |
| "learning_rate": 0.0001934106162294082, | |
| "loss": 0.0126, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.3360800470865215, | |
| "grad_norm": 0.126953125, | |
| "learning_rate": 0.00019338621110433194, | |
| "loss": 0.0192, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.3366686286050618, | |
| "grad_norm": 0.087890625, | |
| "learning_rate": 0.00019336180597925566, | |
| "loss": 0.0114, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.3372572101236021, | |
| "grad_norm": 0.16796875, | |
| "learning_rate": 0.0001933374008541794, | |
| "loss": 0.014, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.33784579164214246, | |
| "grad_norm": 0.076171875, | |
| "learning_rate": 0.00019331299572910313, | |
| "loss": 0.0106, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.33843437316068276, | |
| "grad_norm": 0.0693359375, | |
| "learning_rate": 0.00019328859060402688, | |
| "loss": 0.0105, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.33902295467922305, | |
| "grad_norm": 0.07080078125, | |
| "learning_rate": 0.0001932641854789506, | |
| "loss": 0.0108, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.3396115361977634, | |
| "grad_norm": 0.059326171875, | |
| "learning_rate": 0.00019323978035387434, | |
| "loss": 0.0095, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.3402001177163037, | |
| "grad_norm": 0.0615234375, | |
| "learning_rate": 0.00019321537522879806, | |
| "loss": 0.0102, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.340788699234844, | |
| "grad_norm": 0.0634765625, | |
| "learning_rate": 0.00019319097010372179, | |
| "loss": 0.0098, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.34137728075338436, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 0.0001931665649786455, | |
| "loss": 0.0378, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.34196586227192466, | |
| "grad_norm": 0.173828125, | |
| "learning_rate": 0.00019314215985356925, | |
| "loss": 0.0141, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.34255444379046496, | |
| "grad_norm": 0.09033203125, | |
| "learning_rate": 0.00019311775472849297, | |
| "loss": 0.0107, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.3431430253090053, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 0.00019309334960341672, | |
| "loss": 0.0598, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.3437316068275456, | |
| "grad_norm": 0.6328125, | |
| "learning_rate": 0.00019306894447834044, | |
| "loss": 0.032, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.3443201883460859, | |
| "grad_norm": 0.11669921875, | |
| "learning_rate": 0.0001930445393532642, | |
| "loss": 0.0117, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.34490876986462626, | |
| "grad_norm": 0.064453125, | |
| "learning_rate": 0.0001930201342281879, | |
| "loss": 0.0093, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.34549735138316656, | |
| "grad_norm": 0.07373046875, | |
| "learning_rate": 0.00019299572910311166, | |
| "loss": 0.0102, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.34608593290170686, | |
| "grad_norm": 0.060302734375, | |
| "learning_rate": 0.00019297132397803538, | |
| "loss": 0.0096, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.3466745144202472, | |
| "grad_norm": 0.73046875, | |
| "learning_rate": 0.00019294691885295913, | |
| "loss": 0.0176, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.3472630959387875, | |
| "grad_norm": 0.07763671875, | |
| "learning_rate": 0.00019292251372788285, | |
| "loss": 0.0107, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.3478516774573278, | |
| "grad_norm": 0.06396484375, | |
| "learning_rate": 0.0001928981086028066, | |
| "loss": 0.0096, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.34844025897586817, | |
| "grad_norm": 0.078125, | |
| "learning_rate": 0.00019287370347773034, | |
| "loss": 0.0264, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.34902884049440847, | |
| "grad_norm": 0.275390625, | |
| "learning_rate": 0.00019284929835265406, | |
| "loss": 0.0139, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.3496174220129488, | |
| "grad_norm": 0.064453125, | |
| "learning_rate": 0.0001928248932275778, | |
| "loss": 0.0101, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.3502060035314891, | |
| "grad_norm": 0.0634765625, | |
| "learning_rate": 0.00019280048810250153, | |
| "loss": 0.01, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.3507945850500294, | |
| "grad_norm": 0.0771484375, | |
| "learning_rate": 0.00019277608297742528, | |
| "loss": 0.0106, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.35138316656856977, | |
| "grad_norm": 0.08056640625, | |
| "learning_rate": 0.000192751677852349, | |
| "loss": 0.0106, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.35197174808711007, | |
| "grad_norm": 0.07177734375, | |
| "learning_rate": 0.00019272727272727274, | |
| "loss": 0.0109, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.35256032960565037, | |
| "grad_norm": 0.052001953125, | |
| "learning_rate": 0.00019270286760219647, | |
| "loss": 0.0094, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.3531489111241907, | |
| "grad_norm": 0.08203125, | |
| "learning_rate": 0.0001926784624771202, | |
| "loss": 0.0096, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.353737492642731, | |
| "grad_norm": 0.0810546875, | |
| "learning_rate": 0.00019265405735204393, | |
| "loss": 0.0115, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.3543260741612713, | |
| "grad_norm": 0.1123046875, | |
| "learning_rate": 0.00019262965222696768, | |
| "loss": 0.0115, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.3549146556798117, | |
| "grad_norm": 0.060302734375, | |
| "learning_rate": 0.0001926052471018914, | |
| "loss": 0.0095, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.355503237198352, | |
| "grad_norm": 0.10693359375, | |
| "learning_rate": 0.00019258084197681515, | |
| "loss": 0.0121, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.35609181871689227, | |
| "grad_norm": 0.1259765625, | |
| "learning_rate": 0.00019255643685173887, | |
| "loss": 0.0129, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.3566804002354326, | |
| "grad_norm": 0.062255859375, | |
| "learning_rate": 0.00019253203172666262, | |
| "loss": 0.0099, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.3572689817539729, | |
| "grad_norm": 0.0908203125, | |
| "learning_rate": 0.00019250762660158634, | |
| "loss": 0.0108, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.3578575632725132, | |
| "grad_norm": 0.061767578125, | |
| "learning_rate": 0.00019248322147651008, | |
| "loss": 0.0094, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.3584461447910536, | |
| "grad_norm": 0.11572265625, | |
| "learning_rate": 0.0001924588163514338, | |
| "loss": 0.0097, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.3590347263095939, | |
| "grad_norm": 0.083984375, | |
| "learning_rate": 0.00019243441122635755, | |
| "loss": 0.0121, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.3596233078281342, | |
| "grad_norm": 0.10498046875, | |
| "learning_rate": 0.00019241000610128127, | |
| "loss": 0.012, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.36021188934667453, | |
| "grad_norm": 0.068359375, | |
| "learning_rate": 0.00019238560097620502, | |
| "loss": 0.0091, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.36080047086521483, | |
| "grad_norm": 0.060791015625, | |
| "learning_rate": 0.00019236119585112874, | |
| "loss": 0.0087, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.3613890523837551, | |
| "grad_norm": 0.061767578125, | |
| "learning_rate": 0.0001923367907260525, | |
| "loss": 0.009, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.3619776339022955, | |
| "grad_norm": 0.11083984375, | |
| "learning_rate": 0.0001923123856009762, | |
| "loss": 0.0115, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.3625662154208358, | |
| "grad_norm": 0.06884765625, | |
| "learning_rate": 0.00019228798047589996, | |
| "loss": 0.0102, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.3631547969393761, | |
| "grad_norm": 0.07373046875, | |
| "learning_rate": 0.00019226357535082368, | |
| "loss": 0.0109, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.36374337845791643, | |
| "grad_norm": 0.0859375, | |
| "learning_rate": 0.00019223917022574743, | |
| "loss": 0.0108, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.36433195997645673, | |
| "grad_norm": 0.054443359375, | |
| "learning_rate": 0.00019221476510067115, | |
| "loss": 0.0091, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.36492054149499703, | |
| "grad_norm": 0.1806640625, | |
| "learning_rate": 0.0001921903599755949, | |
| "loss": 0.0134, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.3655091230135374, | |
| "grad_norm": 0.057861328125, | |
| "learning_rate": 0.00019216595485051861, | |
| "loss": 0.0092, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.3660977045320777, | |
| "grad_norm": 0.07080078125, | |
| "learning_rate": 0.00019214154972544236, | |
| "loss": 0.011, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.36668628605061804, | |
| "grad_norm": 0.07080078125, | |
| "learning_rate": 0.00019211714460036608, | |
| "loss": 0.0095, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.36727486756915834, | |
| "grad_norm": 0.1748046875, | |
| "learning_rate": 0.00019209273947528983, | |
| "loss": 0.0114, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.36786344908769864, | |
| "grad_norm": 0.56640625, | |
| "learning_rate": 0.00019206833435021358, | |
| "loss": 0.0288, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.368452030606239, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 0.0001920439292251373, | |
| "loss": 0.0131, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.3690406121247793, | |
| "grad_norm": 0.07470703125, | |
| "learning_rate": 0.00019201952410006104, | |
| "loss": 0.0103, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.3696291936433196, | |
| "grad_norm": 0.09716796875, | |
| "learning_rate": 0.00019199511897498477, | |
| "loss": 0.0109, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.37021777516185994, | |
| "grad_norm": 0.0712890625, | |
| "learning_rate": 0.00019197071384990849, | |
| "loss": 0.0093, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.37080635668040024, | |
| "grad_norm": 0.09912109375, | |
| "learning_rate": 0.0001919463087248322, | |
| "loss": 0.0117, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.37139493819894054, | |
| "grad_norm": 0.0673828125, | |
| "learning_rate": 0.00019192190359975595, | |
| "loss": 0.0089, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.3719835197174809, | |
| "grad_norm": 0.16796875, | |
| "learning_rate": 0.00019189749847467967, | |
| "loss": 0.013, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.3725721012360212, | |
| "grad_norm": 0.064453125, | |
| "learning_rate": 0.00019187309334960342, | |
| "loss": 0.0095, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.3731606827545615, | |
| "grad_norm": 0.099609375, | |
| "learning_rate": 0.00019184868822452714, | |
| "loss": 0.0101, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.37374926427310184, | |
| "grad_norm": 0.0673828125, | |
| "learning_rate": 0.0001918242830994509, | |
| "loss": 0.0096, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.37433784579164214, | |
| "grad_norm": 0.08935546875, | |
| "learning_rate": 0.0001917998779743746, | |
| "loss": 0.0106, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.37492642731018244, | |
| "grad_norm": 0.068359375, | |
| "learning_rate": 0.00019177547284929836, | |
| "loss": 0.0098, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.3755150088287228, | |
| "grad_norm": 0.11328125, | |
| "learning_rate": 0.00019175106772422208, | |
| "loss": 0.0118, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.3761035903472631, | |
| "grad_norm": 0.06689453125, | |
| "learning_rate": 0.00019172666259914583, | |
| "loss": 0.0096, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.3766921718658034, | |
| "grad_norm": 0.1591796875, | |
| "learning_rate": 0.00019170225747406955, | |
| "loss": 0.0153, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.37728075338434375, | |
| "grad_norm": 0.06884765625, | |
| "learning_rate": 0.0001916778523489933, | |
| "loss": 0.0093, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.37786933490288405, | |
| "grad_norm": 0.1396484375, | |
| "learning_rate": 0.00019165344722391701, | |
| "loss": 0.0118, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.37845791642142435, | |
| "grad_norm": 1.015625, | |
| "learning_rate": 0.00019162904209884076, | |
| "loss": 0.0164, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.3790464979399647, | |
| "grad_norm": 0.1552734375, | |
| "learning_rate": 0.00019160463697376448, | |
| "loss": 0.0121, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.379635079458505, | |
| "grad_norm": 0.2431640625, | |
| "learning_rate": 0.00019158023184868823, | |
| "loss": 0.0167, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.3802236609770453, | |
| "grad_norm": 0.05322265625, | |
| "learning_rate": 0.00019155582672361198, | |
| "loss": 0.0091, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.38081224249558565, | |
| "grad_norm": 0.061767578125, | |
| "learning_rate": 0.0001915314215985357, | |
| "loss": 0.0086, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.38140082401412595, | |
| "grad_norm": 0.3828125, | |
| "learning_rate": 0.00019150701647345945, | |
| "loss": 0.0157, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.38198940553266625, | |
| "grad_norm": 1.4296875, | |
| "learning_rate": 0.00019148261134838317, | |
| "loss": 0.0136, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.3825779870512066, | |
| "grad_norm": 1.40625, | |
| "learning_rate": 0.0001914582062233069, | |
| "loss": 0.026, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.3831665685697469, | |
| "grad_norm": 0.06396484375, | |
| "learning_rate": 0.00019143380109823063, | |
| "loss": 0.0096, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.3837551500882872, | |
| "grad_norm": 0.06982421875, | |
| "learning_rate": 0.00019140939597315438, | |
| "loss": 0.0103, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.38434373160682755, | |
| "grad_norm": 11.625, | |
| "learning_rate": 0.0001913849908480781, | |
| "loss": 0.0148, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.38493231312536785, | |
| "grad_norm": 0.08837890625, | |
| "learning_rate": 0.00019136058572300185, | |
| "loss": 0.0103, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.3855208946439082, | |
| "grad_norm": 0.083984375, | |
| "learning_rate": 0.00019133618059792557, | |
| "loss": 0.0115, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.3861094761624485, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 0.00019131177547284932, | |
| "loss": 0.0154, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.3866980576809888, | |
| "grad_norm": 0.2041015625, | |
| "learning_rate": 0.00019128737034777304, | |
| "loss": 0.0125, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.38728663919952916, | |
| "grad_norm": 0.053955078125, | |
| "learning_rate": 0.00019126296522269679, | |
| "loss": 0.0089, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.38787522071806946, | |
| "grad_norm": 0.08349609375, | |
| "learning_rate": 0.0001912385600976205, | |
| "loss": 0.0107, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.38846380223660976, | |
| "grad_norm": 0.09521484375, | |
| "learning_rate": 0.00019121415497254425, | |
| "loss": 0.0104, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.3890523837551501, | |
| "grad_norm": 1.6328125, | |
| "learning_rate": 0.00019118974984746797, | |
| "loss": 0.0222, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.3896409652736904, | |
| "grad_norm": 0.08984375, | |
| "learning_rate": 0.00019116534472239172, | |
| "loss": 0.0101, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.3902295467922307, | |
| "grad_norm": 0.07421875, | |
| "learning_rate": 0.00019114093959731544, | |
| "loss": 0.0102, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.39081812831077106, | |
| "grad_norm": 0.087890625, | |
| "learning_rate": 0.0001911165344722392, | |
| "loss": 0.0109, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.39140670982931136, | |
| "grad_norm": 0.80859375, | |
| "learning_rate": 0.0001910921293471629, | |
| "loss": 0.0181, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.39199529134785166, | |
| "grad_norm": 0.0947265625, | |
| "learning_rate": 0.00019106772422208666, | |
| "loss": 0.0104, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.392583872866392, | |
| "grad_norm": 0.080078125, | |
| "learning_rate": 0.00019104331909701038, | |
| "loss": 0.01, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.3931724543849323, | |
| "grad_norm": 0.1396484375, | |
| "learning_rate": 0.00019101891397193413, | |
| "loss": 0.0177, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.3937610359034726, | |
| "grad_norm": 0.0869140625, | |
| "learning_rate": 0.00019099450884685785, | |
| "loss": 0.0101, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.39434961742201297, | |
| "grad_norm": 0.076171875, | |
| "learning_rate": 0.0001909701037217816, | |
| "loss": 0.0101, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.39493819894055326, | |
| "grad_norm": 0.13671875, | |
| "learning_rate": 0.00019094569859670531, | |
| "loss": 0.0128, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.39552678045909356, | |
| "grad_norm": 0.3984375, | |
| "learning_rate": 0.00019092129347162906, | |
| "loss": 0.0143, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.3961153619776339, | |
| "grad_norm": 0.059814453125, | |
| "learning_rate": 0.00019089688834655278, | |
| "loss": 0.0103, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.3967039434961742, | |
| "grad_norm": 0.08740234375, | |
| "learning_rate": 0.00019087248322147653, | |
| "loss": 0.0098, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.3972925250147145, | |
| "grad_norm": 0.0966796875, | |
| "learning_rate": 0.00019084807809640025, | |
| "loss": 0.0106, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.39788110653325487, | |
| "grad_norm": 0.072265625, | |
| "learning_rate": 0.000190823672971324, | |
| "loss": 0.0097, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.39846968805179517, | |
| "grad_norm": 0.0615234375, | |
| "learning_rate": 0.00019079926784624772, | |
| "loss": 0.0084, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.39905826957033547, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 0.00019077486272117147, | |
| "loss": 0.012, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.3996468510888758, | |
| "grad_norm": 0.06201171875, | |
| "learning_rate": 0.0001907504575960952, | |
| "loss": 0.0095, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.4002354326074161, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.0001907260524710189, | |
| "loss": 0.0091, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.4008240141259564, | |
| "grad_norm": 0.0625, | |
| "learning_rate": 0.00019070164734594265, | |
| "loss": 0.0091, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 0.4014125956444968, | |
| "grad_norm": 0.0908203125, | |
| "learning_rate": 0.00019067724222086637, | |
| "loss": 0.0108, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.40200117716303707, | |
| "grad_norm": 0.06689453125, | |
| "learning_rate": 0.00019065283709579012, | |
| "loss": 0.0094, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 0.4025897586815774, | |
| "grad_norm": 0.083984375, | |
| "learning_rate": 0.00019062843197071384, | |
| "loss": 0.0107, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.4031783402001177, | |
| "grad_norm": 0.11962890625, | |
| "learning_rate": 0.0001906040268456376, | |
| "loss": 0.012, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.403766921718658, | |
| "grad_norm": 0.07373046875, | |
| "learning_rate": 0.0001905796217205613, | |
| "loss": 0.0093, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.4043555032371984, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 0.00019055521659548506, | |
| "loss": 0.0141, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 0.4049440847557387, | |
| "grad_norm": 0.1943359375, | |
| "learning_rate": 0.00019053081147040878, | |
| "loss": 0.0154, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.405532666274279, | |
| "grad_norm": 0.05859375, | |
| "learning_rate": 0.00019050640634533253, | |
| "loss": 0.0095, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 0.40612124779281933, | |
| "grad_norm": 0.11328125, | |
| "learning_rate": 0.00019048200122025625, | |
| "loss": 0.0123, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.4067098293113596, | |
| "grad_norm": 0.0556640625, | |
| "learning_rate": 0.00019045759609518, | |
| "loss": 0.0084, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 0.4072984108298999, | |
| "grad_norm": 0.07421875, | |
| "learning_rate": 0.00019043319097010371, | |
| "loss": 0.0096, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.4078869923484403, | |
| "grad_norm": 0.05712890625, | |
| "learning_rate": 0.00019040878584502746, | |
| "loss": 0.0092, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 0.4084755738669806, | |
| "grad_norm": 0.7578125, | |
| "learning_rate": 0.00019038438071995118, | |
| "loss": 0.0114, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.4090641553855209, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 0.00019035997559487493, | |
| "loss": 0.0125, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.40965273690406123, | |
| "grad_norm": 0.11572265625, | |
| "learning_rate": 0.00019033557046979865, | |
| "loss": 0.0146, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.41024131842260153, | |
| "grad_norm": 0.11083984375, | |
| "learning_rate": 0.0001903111653447224, | |
| "loss": 0.011, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 0.41082989994114183, | |
| "grad_norm": 0.08251953125, | |
| "learning_rate": 0.00019028676021964612, | |
| "loss": 0.0099, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.4114184814596822, | |
| "grad_norm": 0.1455078125, | |
| "learning_rate": 0.00019026235509456987, | |
| "loss": 0.0139, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 0.4120070629782225, | |
| "grad_norm": 0.130859375, | |
| "learning_rate": 0.00019023794996949361, | |
| "loss": 0.0121, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.4125956444967628, | |
| "grad_norm": 0.310546875, | |
| "learning_rate": 0.00019021354484441733, | |
| "loss": 0.0122, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 0.41318422601530314, | |
| "grad_norm": 0.1025390625, | |
| "learning_rate": 0.00019018913971934108, | |
| "loss": 0.0109, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.41377280753384343, | |
| "grad_norm": 0.07470703125, | |
| "learning_rate": 0.0001901647345942648, | |
| "loss": 0.0099, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 0.41436138905238373, | |
| "grad_norm": 0.068359375, | |
| "learning_rate": 0.00019014032946918855, | |
| "loss": 0.0098, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.4149499705709241, | |
| "grad_norm": 0.0576171875, | |
| "learning_rate": 0.00019011592434411227, | |
| "loss": 0.0095, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.4155385520894644, | |
| "grad_norm": 0.1201171875, | |
| "learning_rate": 0.00019009151921903602, | |
| "loss": 0.0113, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.4161271336080047, | |
| "grad_norm": 0.05419921875, | |
| "learning_rate": 0.00019006711409395974, | |
| "loss": 0.009, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 0.41671571512654504, | |
| "grad_norm": 0.10302734375, | |
| "learning_rate": 0.00019004270896888349, | |
| "loss": 0.011, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.41730429664508534, | |
| "grad_norm": 0.056640625, | |
| "learning_rate": 0.0001900183038438072, | |
| "loss": 0.009, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 0.41789287816362564, | |
| "grad_norm": 0.1083984375, | |
| "learning_rate": 0.00018999389871873095, | |
| "loss": 0.0107, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.418481459682166, | |
| "grad_norm": 0.07763671875, | |
| "learning_rate": 0.00018996949359365467, | |
| "loss": 0.0112, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 0.4190700412007063, | |
| "grad_norm": 0.0556640625, | |
| "learning_rate": 0.00018994508846857842, | |
| "loss": 0.0087, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.41965862271924664, | |
| "grad_norm": 0.07421875, | |
| "learning_rate": 0.00018992068334350214, | |
| "loss": 0.0093, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 0.42024720423778694, | |
| "grad_norm": 0.058837890625, | |
| "learning_rate": 0.0001898962782184259, | |
| "loss": 0.009, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.42083578575632724, | |
| "grad_norm": 1.90625, | |
| "learning_rate": 0.0001898718730933496, | |
| "loss": 0.0347, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.4214243672748676, | |
| "grad_norm": 0.048095703125, | |
| "learning_rate": 0.00018984746796827336, | |
| "loss": 0.0087, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.4220129487934079, | |
| "grad_norm": 0.2275390625, | |
| "learning_rate": 0.00018982306284319708, | |
| "loss": 0.0143, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 0.4226015303119482, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 0.00018979865771812083, | |
| "loss": 0.0284, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.42319011183048855, | |
| "grad_norm": 0.0576171875, | |
| "learning_rate": 0.00018977425259304455, | |
| "loss": 0.009, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 0.42377869334902885, | |
| "grad_norm": 0.049560546875, | |
| "learning_rate": 0.0001897498474679683, | |
| "loss": 0.0081, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.42436727486756914, | |
| "grad_norm": 0.06982421875, | |
| "learning_rate": 0.00018972544234289201, | |
| "loss": 0.0099, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 0.4249558563861095, | |
| "grad_norm": 0.06982421875, | |
| "learning_rate": 0.00018970103721781576, | |
| "loss": 0.0095, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.4255444379046498, | |
| "grad_norm": 0.0703125, | |
| "learning_rate": 0.00018967663209273948, | |
| "loss": 0.0103, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 0.4261330194231901, | |
| "grad_norm": 0.068359375, | |
| "learning_rate": 0.00018965222696766323, | |
| "loss": 0.0093, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.42672160094173045, | |
| "grad_norm": 0.07080078125, | |
| "learning_rate": 0.00018962782184258695, | |
| "loss": 0.0094, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.42731018246027075, | |
| "grad_norm": 0.06689453125, | |
| "learning_rate": 0.0001896034167175107, | |
| "loss": 0.0093, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.42789876397881105, | |
| "grad_norm": 0.2236328125, | |
| "learning_rate": 0.00018957901159243442, | |
| "loss": 0.0145, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 0.4284873454973514, | |
| "grad_norm": 0.09765625, | |
| "learning_rate": 0.00018955460646735817, | |
| "loss": 0.0085, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.4290759270158917, | |
| "grad_norm": 0.059326171875, | |
| "learning_rate": 0.0001895302013422819, | |
| "loss": 0.0092, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 0.429664508534432, | |
| "grad_norm": 0.1240234375, | |
| "learning_rate": 0.0001895057962172056, | |
| "loss": 0.0118, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.43025309005297235, | |
| "grad_norm": 0.056396484375, | |
| "learning_rate": 0.00018948139109212935, | |
| "loss": 0.0087, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 0.43084167157151265, | |
| "grad_norm": 0.171875, | |
| "learning_rate": 0.00018945698596705308, | |
| "loss": 0.0127, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.43143025309005295, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 0.00018943258084197682, | |
| "loss": 0.0123, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 0.4320188346085933, | |
| "grad_norm": 0.06689453125, | |
| "learning_rate": 0.00018940817571690054, | |
| "loss": 0.01, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.4326074161271336, | |
| "grad_norm": 0.142578125, | |
| "learning_rate": 0.0001893837705918243, | |
| "loss": 0.011, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.4331959976456739, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 0.000189359365466748, | |
| "loss": 0.0121, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.43378457916421426, | |
| "grad_norm": 0.0478515625, | |
| "learning_rate": 0.00018933496034167176, | |
| "loss": 0.0085, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 0.43437316068275456, | |
| "grad_norm": 8.6875, | |
| "learning_rate": 0.00018931055521659548, | |
| "loss": 0.0413, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.43496174220129485, | |
| "grad_norm": 0.091796875, | |
| "learning_rate": 0.00018928615009151923, | |
| "loss": 0.0099, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 0.4355503237198352, | |
| "grad_norm": 0.1484375, | |
| "learning_rate": 0.00018926174496644295, | |
| "loss": 0.0145, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.4361389052383755, | |
| "grad_norm": 0.0654296875, | |
| "learning_rate": 0.0001892373398413667, | |
| "loss": 0.009, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 0.4367274867569158, | |
| "grad_norm": 0.058837890625, | |
| "learning_rate": 0.00018921293471629042, | |
| "loss": 0.0085, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.43731606827545616, | |
| "grad_norm": 0.1953125, | |
| "learning_rate": 0.00018918852959121416, | |
| "loss": 0.0112, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 0.43790464979399646, | |
| "grad_norm": 0.07470703125, | |
| "learning_rate": 0.00018916412446613788, | |
| "loss": 0.0101, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.4384932313125368, | |
| "grad_norm": 0.1123046875, | |
| "learning_rate": 0.00018913971934106163, | |
| "loss": 0.0339, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.4390818128310771, | |
| "grad_norm": 0.89453125, | |
| "learning_rate": 0.00018911531421598535, | |
| "loss": 0.0156, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.4396703943496174, | |
| "grad_norm": 0.103515625, | |
| "learning_rate": 0.0001890909090909091, | |
| "loss": 0.0125, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 0.44025897586815776, | |
| "grad_norm": 0.08935546875, | |
| "learning_rate": 0.00018906650396583282, | |
| "loss": 0.0123, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.44084755738669806, | |
| "grad_norm": 0.06884765625, | |
| "learning_rate": 0.00018904209884075657, | |
| "loss": 0.0098, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 0.44143613890523836, | |
| "grad_norm": 0.06884765625, | |
| "learning_rate": 0.0001890176937156803, | |
| "loss": 0.0096, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.4420247204237787, | |
| "grad_norm": 0.05712890625, | |
| "learning_rate": 0.00018899328859060404, | |
| "loss": 0.0085, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 0.442613301942319, | |
| "grad_norm": 0.1484375, | |
| "learning_rate": 0.00018896888346552776, | |
| "loss": 0.0121, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.4432018834608593, | |
| "grad_norm": 0.07275390625, | |
| "learning_rate": 0.0001889444783404515, | |
| "loss": 0.0094, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 0.44379046497939967, | |
| "grad_norm": 0.08251953125, | |
| "learning_rate": 0.00018892007321537525, | |
| "loss": 0.0103, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.44437904649793997, | |
| "grad_norm": 0.07177734375, | |
| "learning_rate": 0.00018889566809029897, | |
| "loss": 0.0106, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.44496762801648027, | |
| "grad_norm": 0.061279296875, | |
| "learning_rate": 0.00018887126296522272, | |
| "loss": 0.0095, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.4455562095350206, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.00018884685784014644, | |
| "loss": 0.0099, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 0.4461447910535609, | |
| "grad_norm": 0.0556640625, | |
| "learning_rate": 0.0001888224527150702, | |
| "loss": 0.0086, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.4467333725721012, | |
| "grad_norm": 0.05419921875, | |
| "learning_rate": 0.0001887980475899939, | |
| "loss": 0.0085, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 0.44732195409064157, | |
| "grad_norm": 0.052490234375, | |
| "learning_rate": 0.00018877364246491765, | |
| "loss": 0.0087, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.44791053560918187, | |
| "grad_norm": 0.09130859375, | |
| "learning_rate": 0.00018874923733984138, | |
| "loss": 0.0096, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 0.44849911712772217, | |
| "grad_norm": 0.061279296875, | |
| "learning_rate": 0.00018872483221476512, | |
| "loss": 0.0092, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.4490876986462625, | |
| "grad_norm": 5.4375, | |
| "learning_rate": 0.00018870042708968884, | |
| "loss": 0.0142, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 0.4496762801648028, | |
| "grad_norm": 0.059814453125, | |
| "learning_rate": 0.0001886760219646126, | |
| "loss": 0.0096, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.4502648616833431, | |
| "grad_norm": 0.55078125, | |
| "learning_rate": 0.0001886516168395363, | |
| "loss": 0.0229, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.4508534432018835, | |
| "grad_norm": 2.109375, | |
| "learning_rate": 0.00018862721171446006, | |
| "loss": 0.0161, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.4514420247204238, | |
| "grad_norm": 1.0, | |
| "learning_rate": 0.00018860280658938378, | |
| "loss": 0.0103, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 0.45203060623896407, | |
| "grad_norm": 0.06396484375, | |
| "learning_rate": 0.00018857840146430753, | |
| "loss": 0.0088, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.4526191877575044, | |
| "grad_norm": 0.053955078125, | |
| "learning_rate": 0.00018855399633923125, | |
| "loss": 0.0085, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 0.4532077692760447, | |
| "grad_norm": 0.064453125, | |
| "learning_rate": 0.000188529591214155, | |
| "loss": 0.0089, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.453796350794585, | |
| "grad_norm": 0.0693359375, | |
| "learning_rate": 0.00018850518608907872, | |
| "loss": 0.009, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 0.4543849323131254, | |
| "grad_norm": 0.49609375, | |
| "learning_rate": 0.00018848078096400246, | |
| "loss": 0.0176, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.4549735138316657, | |
| "grad_norm": 0.419921875, | |
| "learning_rate": 0.00018845637583892618, | |
| "loss": 0.0164, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 0.45556209535020603, | |
| "grad_norm": 0.06787109375, | |
| "learning_rate": 0.00018843197071384993, | |
| "loss": 0.0096, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.45615067686874633, | |
| "grad_norm": 0.051513671875, | |
| "learning_rate": 0.00018840756558877365, | |
| "loss": 0.0085, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.45673925838728663, | |
| "grad_norm": 0.14453125, | |
| "learning_rate": 0.0001883831604636974, | |
| "loss": 0.0134, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.457327839905827, | |
| "grad_norm": 0.1279296875, | |
| "learning_rate": 0.00018835875533862112, | |
| "loss": 0.0098, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 0.4579164214243673, | |
| "grad_norm": 0.07373046875, | |
| "learning_rate": 0.00018833435021354487, | |
| "loss": 0.0087, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.4585050029429076, | |
| "grad_norm": 0.138671875, | |
| "learning_rate": 0.0001883099450884686, | |
| "loss": 0.0111, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 0.45909358446144793, | |
| "grad_norm": 0.10595703125, | |
| "learning_rate": 0.00018828553996339233, | |
| "loss": 0.0116, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.45968216597998823, | |
| "grad_norm": 0.0693359375, | |
| "learning_rate": 0.00018826113483831606, | |
| "loss": 0.0094, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 0.46027074749852853, | |
| "grad_norm": 0.07763671875, | |
| "learning_rate": 0.00018823672971323978, | |
| "loss": 0.0098, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.4608593290170689, | |
| "grad_norm": 0.427734375, | |
| "learning_rate": 0.00018821232458816352, | |
| "loss": 0.0129, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 0.4614479105356092, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.00018818791946308724, | |
| "loss": 0.009, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.4620364920541495, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.000188163514338011, | |
| "loss": 0.0124, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.46262507357268984, | |
| "grad_norm": 0.061279296875, | |
| "learning_rate": 0.0001881391092129347, | |
| "loss": 0.0092, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.46321365509123014, | |
| "grad_norm": 0.1708984375, | |
| "learning_rate": 0.00018811470408785846, | |
| "loss": 0.0116, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 0.46380223660977044, | |
| "grad_norm": 0.06884765625, | |
| "learning_rate": 0.00018809029896278218, | |
| "loss": 0.0102, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.4643908181283108, | |
| "grad_norm": 0.0751953125, | |
| "learning_rate": 0.00018806589383770593, | |
| "loss": 0.0107, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 0.4649793996468511, | |
| "grad_norm": 0.06494140625, | |
| "learning_rate": 0.00018804148871262965, | |
| "loss": 0.0096, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.4655679811653914, | |
| "grad_norm": 0.06689453125, | |
| "learning_rate": 0.0001880170835875534, | |
| "loss": 0.0103, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 0.46615656268393174, | |
| "grad_norm": 0.11865234375, | |
| "learning_rate": 0.00018799267846247712, | |
| "loss": 0.0099, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.46674514420247204, | |
| "grad_norm": 0.06982421875, | |
| "learning_rate": 0.00018796827333740086, | |
| "loss": 0.0098, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 0.46733372572101234, | |
| "grad_norm": 0.06787109375, | |
| "learning_rate": 0.00018794386821232458, | |
| "loss": 0.0086, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.4679223072395527, | |
| "grad_norm": 0.09619140625, | |
| "learning_rate": 0.00018791946308724833, | |
| "loss": 0.0105, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.468510888758093, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 0.00018789505796217205, | |
| "loss": 0.0322, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.4690994702766333, | |
| "grad_norm": 0.05810546875, | |
| "learning_rate": 0.0001878706528370958, | |
| "loss": 0.0092, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 0.46968805179517364, | |
| "grad_norm": 0.0771484375, | |
| "learning_rate": 0.00018784624771201952, | |
| "loss": 0.0101, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 0.47027663331371394, | |
| "grad_norm": 0.1015625, | |
| "learning_rate": 0.00018782184258694327, | |
| "loss": 0.012, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 0.47086521483225424, | |
| "grad_norm": 0.1875, | |
| "learning_rate": 0.000187797437461867, | |
| "loss": 0.0137, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.4714537963507946, | |
| "grad_norm": 0.054931640625, | |
| "learning_rate": 0.00018777303233679074, | |
| "loss": 0.0084, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 0.4720423778693349, | |
| "grad_norm": 0.064453125, | |
| "learning_rate": 0.00018774862721171446, | |
| "loss": 0.0096, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 0.4726309593878752, | |
| "grad_norm": 4.6875, | |
| "learning_rate": 0.0001877242220866382, | |
| "loss": 0.0297, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 0.47321954090641555, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.00018769981696156192, | |
| "loss": 0.0095, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.47380812242495585, | |
| "grad_norm": 0.058837890625, | |
| "learning_rate": 0.00018767541183648567, | |
| "loss": 0.009, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.4743967039434962, | |
| "grad_norm": 0.078125, | |
| "learning_rate": 0.0001876510067114094, | |
| "loss": 0.0103, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 0.4749852854620365, | |
| "grad_norm": 0.0634765625, | |
| "learning_rate": 0.00018762660158633314, | |
| "loss": 0.0094, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 0.4755738669805768, | |
| "grad_norm": 0.05615234375, | |
| "learning_rate": 0.0001876021964612569, | |
| "loss": 0.0086, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.47616244849911715, | |
| "grad_norm": 0.0703125, | |
| "learning_rate": 0.0001875777913361806, | |
| "loss": 0.0096, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 0.47675103001765745, | |
| "grad_norm": 0.0625, | |
| "learning_rate": 0.00018755338621110436, | |
| "loss": 0.0091, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.47733961153619775, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.00018752898108602808, | |
| "loss": 0.0112, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 0.4779281930547381, | |
| "grad_norm": 0.72265625, | |
| "learning_rate": 0.00018750457596095182, | |
| "loss": 0.0141, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.4785167745732784, | |
| "grad_norm": 0.07958984375, | |
| "learning_rate": 0.00018748017083587554, | |
| "loss": 0.0104, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 0.4791053560918187, | |
| "grad_norm": 0.05810546875, | |
| "learning_rate": 0.0001874557657107993, | |
| "loss": 0.01, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 0.47969393761035906, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.000187431360585723, | |
| "loss": 0.0098, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.48028251912889935, | |
| "grad_norm": 0.05078125, | |
| "learning_rate": 0.00018740695546064676, | |
| "loss": 0.0088, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.48087110064743965, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.00018738255033557048, | |
| "loss": 0.0145, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 0.48145968216598, | |
| "grad_norm": 0.0732421875, | |
| "learning_rate": 0.00018735814521049423, | |
| "loss": 0.0101, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 0.4820482636845203, | |
| "grad_norm": 0.087890625, | |
| "learning_rate": 0.00018733374008541795, | |
| "loss": 0.0118, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 0.4826368452030606, | |
| "grad_norm": 0.06103515625, | |
| "learning_rate": 0.0001873093349603417, | |
| "loss": 0.0095, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.48322542672160096, | |
| "grad_norm": 0.11376953125, | |
| "learning_rate": 0.00018728492983526542, | |
| "loss": 0.0115, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 0.48381400824014126, | |
| "grad_norm": 0.055908203125, | |
| "learning_rate": 0.00018726052471018916, | |
| "loss": 0.009, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 0.48440258975868156, | |
| "grad_norm": 0.068359375, | |
| "learning_rate": 0.00018723611958511288, | |
| "loss": 0.0092, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 0.4849911712772219, | |
| "grad_norm": 0.06298828125, | |
| "learning_rate": 0.00018721171446003663, | |
| "loss": 0.0093, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.4855797527957622, | |
| "grad_norm": 0.054931640625, | |
| "learning_rate": 0.00018718730933496035, | |
| "loss": 0.0089, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.4861683343143025, | |
| "grad_norm": 0.052490234375, | |
| "learning_rate": 0.0001871629042098841, | |
| "loss": 0.0087, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 0.48675691583284286, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.00018713849908480782, | |
| "loss": 0.0089, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 0.48734549735138316, | |
| "grad_norm": 0.1318359375, | |
| "learning_rate": 0.00018711409395973157, | |
| "loss": 0.0099, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.48793407886992346, | |
| "grad_norm": 0.1494140625, | |
| "learning_rate": 0.0001870896888346553, | |
| "loss": 0.0129, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 0.4885226603884638, | |
| "grad_norm": 0.173828125, | |
| "learning_rate": 0.00018706528370957904, | |
| "loss": 0.0118, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.4891112419070041, | |
| "grad_norm": 0.05224609375, | |
| "learning_rate": 0.00018704087858450276, | |
| "loss": 0.0087, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 0.4896998234255444, | |
| "grad_norm": 0.09716796875, | |
| "learning_rate": 0.00018701647345942648, | |
| "loss": 0.0101, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.49028840494408477, | |
| "grad_norm": 0.043701171875, | |
| "learning_rate": 0.00018699206833435022, | |
| "loss": 0.0079, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 0.49087698646262506, | |
| "grad_norm": 0.0498046875, | |
| "learning_rate": 0.00018696766320927394, | |
| "loss": 0.009, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 0.4914655679811654, | |
| "grad_norm": 0.048583984375, | |
| "learning_rate": 0.0001869432580841977, | |
| "loss": 0.0078, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.4920541494997057, | |
| "grad_norm": 0.06982421875, | |
| "learning_rate": 0.0001869188529591214, | |
| "loss": 0.009, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.492642731018246, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 0.00018689444783404516, | |
| "loss": 0.0124, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 0.49323131253678637, | |
| "grad_norm": 0.0478515625, | |
| "learning_rate": 0.00018687004270896888, | |
| "loss": 0.0082, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 0.49381989405532667, | |
| "grad_norm": 0.05908203125, | |
| "learning_rate": 0.00018684563758389263, | |
| "loss": 0.0085, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 0.49440847557386697, | |
| "grad_norm": 0.126953125, | |
| "learning_rate": 0.00018682123245881635, | |
| "loss": 0.0135, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.4949970570924073, | |
| "grad_norm": 0.0673828125, | |
| "learning_rate": 0.0001867968273337401, | |
| "loss": 0.0099, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 0.4955856386109476, | |
| "grad_norm": 0.0576171875, | |
| "learning_rate": 0.00018677242220866382, | |
| "loss": 0.0093, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 0.4961742201294879, | |
| "grad_norm": 0.0673828125, | |
| "learning_rate": 0.00018674801708358756, | |
| "loss": 0.0096, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 0.4967628016480283, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.00018672361195851128, | |
| "loss": 0.0126, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.4973513831665686, | |
| "grad_norm": 0.08056640625, | |
| "learning_rate": 0.00018669920683343503, | |
| "loss": 0.0087, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.49793996468510887, | |
| "grad_norm": 0.0712890625, | |
| "learning_rate": 0.00018667480170835875, | |
| "loss": 0.0098, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 0.4985285462036492, | |
| "grad_norm": 0.046630859375, | |
| "learning_rate": 0.0001866503965832825, | |
| "loss": 0.0084, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 0.4991171277221895, | |
| "grad_norm": 0.0615234375, | |
| "learning_rate": 0.00018662599145820622, | |
| "loss": 0.0091, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.4997057092407298, | |
| "grad_norm": 0.09033203125, | |
| "learning_rate": 0.00018660158633312997, | |
| "loss": 0.0094, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 0.5002942907592701, | |
| "grad_norm": 0.0791015625, | |
| "learning_rate": 0.0001865771812080537, | |
| "loss": 0.0099, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.5008828722778105, | |
| "grad_norm": 0.625, | |
| "learning_rate": 0.00018655277608297744, | |
| "loss": 0.0119, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 0.5014714537963508, | |
| "grad_norm": 0.056640625, | |
| "learning_rate": 0.00018652837095790116, | |
| "loss": 0.0088, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.5020600353148911, | |
| "grad_norm": 0.07421875, | |
| "learning_rate": 0.0001865039658328249, | |
| "loss": 0.0103, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 0.5026486168334314, | |
| "grad_norm": 0.054931640625, | |
| "learning_rate": 0.00018647956070774862, | |
| "loss": 0.0083, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 0.5032371983519718, | |
| "grad_norm": 0.083984375, | |
| "learning_rate": 0.00018645515558267237, | |
| "loss": 0.0104, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.503825779870512, | |
| "grad_norm": 0.048095703125, | |
| "learning_rate": 0.0001864307504575961, | |
| "loss": 0.0081, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.5044143613890524, | |
| "grad_norm": 0.07421875, | |
| "learning_rate": 0.00018640634533251984, | |
| "loss": 0.0096, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 0.5050029429075927, | |
| "grad_norm": 0.1064453125, | |
| "learning_rate": 0.00018638194020744356, | |
| "loss": 0.0098, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 0.505591524426133, | |
| "grad_norm": 0.36328125, | |
| "learning_rate": 0.0001863575350823673, | |
| "loss": 0.0154, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 0.5061801059446733, | |
| "grad_norm": 0.04833984375, | |
| "learning_rate": 0.00018633312995729103, | |
| "loss": 0.0086, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.5067686874632137, | |
| "grad_norm": 0.0732421875, | |
| "learning_rate": 0.00018630872483221478, | |
| "loss": 0.0103, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 0.5073572689817539, | |
| "grad_norm": 0.052734375, | |
| "learning_rate": 0.00018628431970713852, | |
| "loss": 0.009, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 0.5079458505002943, | |
| "grad_norm": 0.1533203125, | |
| "learning_rate": 0.00018625991458206224, | |
| "loss": 0.012, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 0.5085344320188346, | |
| "grad_norm": 0.0556640625, | |
| "learning_rate": 0.000186235509456986, | |
| "loss": 0.0084, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.5091230135373749, | |
| "grad_norm": 0.072265625, | |
| "learning_rate": 0.0001862111043319097, | |
| "loss": 0.0103, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.5097115950559152, | |
| "grad_norm": 0.283203125, | |
| "learning_rate": 0.00018618669920683346, | |
| "loss": 0.0098, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 0.5103001765744556, | |
| "grad_norm": 0.2353515625, | |
| "learning_rate": 0.00018616229408175718, | |
| "loss": 0.017, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 0.5108887580929958, | |
| "grad_norm": 0.058837890625, | |
| "learning_rate": 0.00018613788895668093, | |
| "loss": 0.0097, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.5114773396115362, | |
| "grad_norm": 0.08935546875, | |
| "learning_rate": 0.00018611348383160465, | |
| "loss": 0.0099, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 0.5120659211300765, | |
| "grad_norm": 0.068359375, | |
| "learning_rate": 0.0001860890787065284, | |
| "loss": 0.0089, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5126545026486168, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.00018606467358145212, | |
| "loss": 0.0151, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 0.5132430841671571, | |
| "grad_norm": 0.046630859375, | |
| "learning_rate": 0.00018604026845637586, | |
| "loss": 0.0091, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.5138316656856975, | |
| "grad_norm": 0.08349609375, | |
| "learning_rate": 0.00018601586333129958, | |
| "loss": 0.011, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 0.5144202472042377, | |
| "grad_norm": 0.0966796875, | |
| "learning_rate": 0.00018599145820622333, | |
| "loss": 0.0101, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 0.5150088287227781, | |
| "grad_norm": 0.053466796875, | |
| "learning_rate": 0.00018596705308114705, | |
| "loss": 0.0083, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.5155974102413184, | |
| "grad_norm": 0.0654296875, | |
| "learning_rate": 0.0001859426479560708, | |
| "loss": 0.0092, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.5161859917598587, | |
| "grad_norm": 0.061279296875, | |
| "learning_rate": 0.00018591824283099452, | |
| "loss": 0.009, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 0.516774573278399, | |
| "grad_norm": 0.07763671875, | |
| "learning_rate": 0.00018589383770591827, | |
| "loss": 0.0283, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 0.5173631547969394, | |
| "grad_norm": 0.052734375, | |
| "learning_rate": 0.000185869432580842, | |
| "loss": 0.0086, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 0.5179517363154797, | |
| "grad_norm": 0.060546875, | |
| "learning_rate": 0.00018584502745576574, | |
| "loss": 0.0087, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.51854031783402, | |
| "grad_norm": 0.0830078125, | |
| "learning_rate": 0.00018582062233068946, | |
| "loss": 0.0103, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 0.5191288993525603, | |
| "grad_norm": 0.07470703125, | |
| "learning_rate": 0.00018579621720561318, | |
| "loss": 0.0103, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 0.5197174808711007, | |
| "grad_norm": 0.10009765625, | |
| "learning_rate": 0.00018577181208053692, | |
| "loss": 0.0128, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 0.5203060623896409, | |
| "grad_norm": 0.05322265625, | |
| "learning_rate": 0.00018574740695546064, | |
| "loss": 0.009, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.5208946439081813, | |
| "grad_norm": 0.06201171875, | |
| "learning_rate": 0.0001857230018303844, | |
| "loss": 0.0089, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.5214832254267217, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 0.0001856985967053081, | |
| "loss": 0.0125, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 0.5220718069452619, | |
| "grad_norm": 0.11328125, | |
| "learning_rate": 0.00018567419158023186, | |
| "loss": 0.0114, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 0.5226603884638023, | |
| "grad_norm": 0.054443359375, | |
| "learning_rate": 0.00018564978645515558, | |
| "loss": 0.0087, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.5232489699823426, | |
| "grad_norm": 0.05859375, | |
| "learning_rate": 0.00018562538133007933, | |
| "loss": 0.0099, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 0.5238375515008828, | |
| "grad_norm": 0.0556640625, | |
| "learning_rate": 0.00018560097620500305, | |
| "loss": 0.0084, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5244261330194232, | |
| "grad_norm": 0.05615234375, | |
| "learning_rate": 0.0001855765710799268, | |
| "loss": 0.0083, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 0.5250147145379636, | |
| "grad_norm": 0.0634765625, | |
| "learning_rate": 0.00018555216595485052, | |
| "loss": 0.009, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.5256032960565038, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.00018552776082977426, | |
| "loss": 0.0096, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 0.5261918775750442, | |
| "grad_norm": 0.09130859375, | |
| "learning_rate": 0.00018550335570469799, | |
| "loss": 0.0098, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 0.5267804590935845, | |
| "grad_norm": 0.062255859375, | |
| "learning_rate": 0.00018547895057962173, | |
| "loss": 0.0098, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.5273690406121248, | |
| "grad_norm": 0.1767578125, | |
| "learning_rate": 0.00018545454545454545, | |
| "loss": 0.011, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.5279576221306651, | |
| "grad_norm": 0.1796875, | |
| "learning_rate": 0.0001854301403294692, | |
| "loss": 0.0106, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 0.5285462036492055, | |
| "grad_norm": 0.07666015625, | |
| "learning_rate": 0.00018540573520439292, | |
| "loss": 0.0106, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 0.5291347851677457, | |
| "grad_norm": 0.0888671875, | |
| "learning_rate": 0.00018538133007931667, | |
| "loss": 0.0107, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 0.5297233666862861, | |
| "grad_norm": 0.055908203125, | |
| "learning_rate": 0.0001853569249542404, | |
| "loss": 0.0091, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.5303119482048264, | |
| "grad_norm": 0.22265625, | |
| "learning_rate": 0.00018533251982916414, | |
| "loss": 0.0155, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 0.5309005297233667, | |
| "grad_norm": 0.08447265625, | |
| "learning_rate": 0.00018530811470408786, | |
| "loss": 0.0111, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 0.531489111241907, | |
| "grad_norm": 0.0869140625, | |
| "learning_rate": 0.0001852837095790116, | |
| "loss": 0.0099, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 0.5320776927604474, | |
| "grad_norm": 0.09765625, | |
| "learning_rate": 0.00018525930445393533, | |
| "loss": 0.0096, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.5326662742789876, | |
| "grad_norm": 0.057373046875, | |
| "learning_rate": 0.00018523489932885907, | |
| "loss": 0.0088, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.533254855797528, | |
| "grad_norm": 0.058837890625, | |
| "learning_rate": 0.0001852104942037828, | |
| "loss": 0.009, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 0.5338434373160683, | |
| "grad_norm": 0.058837890625, | |
| "learning_rate": 0.00018518608907870654, | |
| "loss": 0.0095, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 0.5344320188346086, | |
| "grad_norm": 0.70703125, | |
| "learning_rate": 0.00018516168395363026, | |
| "loss": 0.0107, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.5350206003531489, | |
| "grad_norm": 0.1171875, | |
| "learning_rate": 0.000185137278828554, | |
| "loss": 0.0112, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 0.5356091818716893, | |
| "grad_norm": 0.11572265625, | |
| "learning_rate": 0.00018511287370347773, | |
| "loss": 0.0111, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.5361977633902295, | |
| "grad_norm": 0.130859375, | |
| "learning_rate": 0.00018508846857840148, | |
| "loss": 0.0114, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 0.5367863449087699, | |
| "grad_norm": 1.5, | |
| "learning_rate": 0.0001850640634533252, | |
| "loss": 0.0131, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.5373749264273102, | |
| "grad_norm": 0.1025390625, | |
| "learning_rate": 0.00018503965832824894, | |
| "loss": 0.0107, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 0.5379635079458505, | |
| "grad_norm": 0.0673828125, | |
| "learning_rate": 0.00018501525320317267, | |
| "loss": 0.0099, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 0.5385520894643908, | |
| "grad_norm": 0.06298828125, | |
| "learning_rate": 0.0001849908480780964, | |
| "loss": 0.0085, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.5391406709829312, | |
| "grad_norm": 0.0771484375, | |
| "learning_rate": 0.00018496644295302016, | |
| "loss": 0.0104, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.5397292525014714, | |
| "grad_norm": 0.060302734375, | |
| "learning_rate": 0.00018494203782794388, | |
| "loss": 0.0092, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 0.5403178340200118, | |
| "grad_norm": 0.07275390625, | |
| "learning_rate": 0.00018491763270286763, | |
| "loss": 0.0089, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 0.5409064155385521, | |
| "grad_norm": 0.055419921875, | |
| "learning_rate": 0.00018489322757779135, | |
| "loss": 0.0088, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 0.5414949970570924, | |
| "grad_norm": 0.0771484375, | |
| "learning_rate": 0.0001848688224527151, | |
| "loss": 0.0101, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.5420835785756327, | |
| "grad_norm": 0.5078125, | |
| "learning_rate": 0.00018484441732763882, | |
| "loss": 0.0198, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 0.5426721600941731, | |
| "grad_norm": 0.1572265625, | |
| "learning_rate": 0.00018482001220256256, | |
| "loss": 0.0093, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 0.5432607416127133, | |
| "grad_norm": 0.05078125, | |
| "learning_rate": 0.00018479560707748628, | |
| "loss": 0.0084, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 0.5438493231312537, | |
| "grad_norm": 0.05712890625, | |
| "learning_rate": 0.00018477120195241003, | |
| "loss": 0.0084, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.544437904649794, | |
| "grad_norm": 0.09423828125, | |
| "learning_rate": 0.00018474679682733375, | |
| "loss": 0.0109, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.5450264861683343, | |
| "grad_norm": 0.0517578125, | |
| "learning_rate": 0.0001847223917022575, | |
| "loss": 0.0084, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 0.5456150676868746, | |
| "grad_norm": 0.052978515625, | |
| "learning_rate": 0.00018469798657718122, | |
| "loss": 0.0083, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 0.546203649205415, | |
| "grad_norm": 0.08251953125, | |
| "learning_rate": 0.00018467358145210497, | |
| "loss": 0.0108, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.5467922307239552, | |
| "grad_norm": 0.07275390625, | |
| "learning_rate": 0.0001846491763270287, | |
| "loss": 0.01, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 0.5473808122424956, | |
| "grad_norm": 0.373046875, | |
| "learning_rate": 0.00018462477120195244, | |
| "loss": 0.0162, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.5479693937610359, | |
| "grad_norm": 0.078125, | |
| "learning_rate": 0.00018460036607687616, | |
| "loss": 0.012, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 0.5485579752795762, | |
| "grad_norm": 0.07177734375, | |
| "learning_rate": 0.0001845759609517999, | |
| "loss": 0.0087, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.5491465567981165, | |
| "grad_norm": 0.06005859375, | |
| "learning_rate": 0.0001845515558267236, | |
| "loss": 0.0094, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 0.5497351383166569, | |
| "grad_norm": 0.058349609375, | |
| "learning_rate": 0.00018452715070164735, | |
| "loss": 0.0087, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 0.5503237198351971, | |
| "grad_norm": 0.060546875, | |
| "learning_rate": 0.00018450274557657107, | |
| "loss": 0.0093, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.5509123013537375, | |
| "grad_norm": 0.056396484375, | |
| "learning_rate": 0.0001844783404514948, | |
| "loss": 0.009, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 0.5515008828722778, | |
| "grad_norm": 0.0439453125, | |
| "learning_rate": 0.00018445393532641856, | |
| "loss": 0.0081, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 0.5520894643908181, | |
| "grad_norm": 0.07861328125, | |
| "learning_rate": 0.00018442953020134228, | |
| "loss": 0.0095, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 0.5526780459093584, | |
| "grad_norm": 0.0927734375, | |
| "learning_rate": 0.00018440512507626603, | |
| "loss": 0.0101, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 0.5532666274278988, | |
| "grad_norm": 0.044677734375, | |
| "learning_rate": 0.00018438071995118975, | |
| "loss": 0.0082, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.5538552089464391, | |
| "grad_norm": 0.08984375, | |
| "learning_rate": 0.0001843563148261135, | |
| "loss": 0.0093, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 0.5544437904649794, | |
| "grad_norm": 0.0791015625, | |
| "learning_rate": 0.00018433190970103722, | |
| "loss": 0.0094, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 0.5550323719835197, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.00018430750457596097, | |
| "loss": 0.0097, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 0.5556209535020601, | |
| "grad_norm": 0.05615234375, | |
| "learning_rate": 0.00018428309945088469, | |
| "loss": 0.0085, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 0.5562095350206003, | |
| "grad_norm": 0.1123046875, | |
| "learning_rate": 0.00018425869432580843, | |
| "loss": 0.0107, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.5567981165391407, | |
| "grad_norm": 0.076171875, | |
| "learning_rate": 0.00018423428920073215, | |
| "loss": 0.0106, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 0.557386698057681, | |
| "grad_norm": 0.0537109375, | |
| "learning_rate": 0.0001842098840756559, | |
| "loss": 0.0086, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 0.5579752795762213, | |
| "grad_norm": 0.05322265625, | |
| "learning_rate": 0.00018418547895057962, | |
| "loss": 0.0082, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 0.5585638610947616, | |
| "grad_norm": 0.1396484375, | |
| "learning_rate": 0.00018416107382550337, | |
| "loss": 0.0107, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 0.559152442613302, | |
| "grad_norm": 0.07177734375, | |
| "learning_rate": 0.0001841366687004271, | |
| "loss": 0.0102, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.5597410241318422, | |
| "grad_norm": 0.057861328125, | |
| "learning_rate": 0.00018411226357535084, | |
| "loss": 0.0092, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 0.5603296056503826, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.00018408785845027456, | |
| "loss": 0.0099, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 0.5609181871689229, | |
| "grad_norm": 0.0478515625, | |
| "learning_rate": 0.0001840634533251983, | |
| "loss": 0.0082, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 0.5615067686874632, | |
| "grad_norm": 0.07861328125, | |
| "learning_rate": 0.00018403904820012203, | |
| "loss": 0.0105, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 0.5620953502060035, | |
| "grad_norm": 0.068359375, | |
| "learning_rate": 0.00018401464307504577, | |
| "loss": 0.0092, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.5626839317245439, | |
| "grad_norm": 0.05029296875, | |
| "learning_rate": 0.0001839902379499695, | |
| "loss": 0.0084, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 0.5632725132430841, | |
| "grad_norm": 0.07666015625, | |
| "learning_rate": 0.00018396583282489324, | |
| "loss": 0.0106, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 0.5638610947616245, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.00018394142769981696, | |
| "loss": 0.0103, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 0.5644496762801648, | |
| "grad_norm": 0.19921875, | |
| "learning_rate": 0.0001839170225747407, | |
| "loss": 0.0171, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 0.5650382577987051, | |
| "grad_norm": 0.040771484375, | |
| "learning_rate": 0.00018389261744966443, | |
| "loss": 0.0074, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5656268393172454, | |
| "grad_norm": 0.055419921875, | |
| "learning_rate": 0.00018386821232458818, | |
| "loss": 0.0085, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 0.5662154208357858, | |
| "grad_norm": 0.052978515625, | |
| "learning_rate": 0.0001838438071995119, | |
| "loss": 0.0094, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 0.566804002354326, | |
| "grad_norm": 0.050048828125, | |
| "learning_rate": 0.00018381940207443565, | |
| "loss": 0.0082, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 0.5673925838728664, | |
| "grad_norm": 0.05517578125, | |
| "learning_rate": 0.00018379499694935937, | |
| "loss": 0.0085, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 0.5679811653914068, | |
| "grad_norm": 0.053466796875, | |
| "learning_rate": 0.0001837705918242831, | |
| "loss": 0.0087, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.568569746909947, | |
| "grad_norm": 0.08203125, | |
| "learning_rate": 0.00018374618669920683, | |
| "loss": 0.0091, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 0.5691583284284873, | |
| "grad_norm": 0.0615234375, | |
| "learning_rate": 0.00018372178157413058, | |
| "loss": 0.0088, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 0.5697469099470277, | |
| "grad_norm": 0.0693359375, | |
| "learning_rate": 0.0001836973764490543, | |
| "loss": 0.009, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 0.570335491465568, | |
| "grad_norm": 0.050537109375, | |
| "learning_rate": 0.00018367297132397805, | |
| "loss": 0.0083, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 0.5709240729841083, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.0001836485661989018, | |
| "loss": 0.0096, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.5715126545026487, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.00018362416107382552, | |
| "loss": 0.0098, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 0.5721012360211889, | |
| "grad_norm": 0.11181640625, | |
| "learning_rate": 0.00018359975594874927, | |
| "loss": 0.0102, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 0.5726898175397293, | |
| "grad_norm": 0.056640625, | |
| "learning_rate": 0.00018357535082367299, | |
| "loss": 0.0093, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 0.5732783990582696, | |
| "grad_norm": 0.06298828125, | |
| "learning_rate": 0.00018355094569859673, | |
| "loss": 0.0092, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 0.5738669805768098, | |
| "grad_norm": 0.052490234375, | |
| "learning_rate": 0.00018352654057352045, | |
| "loss": 0.0082, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.5744555620953502, | |
| "grad_norm": 0.1826171875, | |
| "learning_rate": 0.0001835021354484442, | |
| "loss": 0.0129, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 0.5750441436138906, | |
| "grad_norm": 0.045166015625, | |
| "learning_rate": 0.00018347773032336792, | |
| "loss": 0.0083, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 0.5756327251324308, | |
| "grad_norm": 0.053955078125, | |
| "learning_rate": 0.00018345332519829167, | |
| "loss": 0.0092, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 0.5762213066509712, | |
| "grad_norm": 0.0751953125, | |
| "learning_rate": 0.0001834289200732154, | |
| "loss": 0.0099, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 0.5768098881695115, | |
| "grad_norm": 1.78125, | |
| "learning_rate": 0.00018340451494813914, | |
| "loss": 0.0282, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.5773984696880518, | |
| "grad_norm": 0.0478515625, | |
| "learning_rate": 0.00018338010982306286, | |
| "loss": 0.0089, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 0.5779870512065921, | |
| "grad_norm": 0.049072265625, | |
| "learning_rate": 0.0001833557046979866, | |
| "loss": 0.0082, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 0.5785756327251325, | |
| "grad_norm": 0.05615234375, | |
| "learning_rate": 0.0001833312995729103, | |
| "loss": 0.0081, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 0.5791642142436727, | |
| "grad_norm": 0.0947265625, | |
| "learning_rate": 0.00018330689444783405, | |
| "loss": 0.0118, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 0.5797527957622131, | |
| "grad_norm": 0.07177734375, | |
| "learning_rate": 0.00018328248932275777, | |
| "loss": 0.0089, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.5803413772807534, | |
| "grad_norm": 0.05078125, | |
| "learning_rate": 0.00018325808419768151, | |
| "loss": 0.0082, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 0.5809299587992937, | |
| "grad_norm": 0.058837890625, | |
| "learning_rate": 0.00018323367907260523, | |
| "loss": 0.0082, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 0.581518540317834, | |
| "grad_norm": 0.05859375, | |
| "learning_rate": 0.00018320927394752898, | |
| "loss": 0.0088, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 0.5821071218363744, | |
| "grad_norm": 0.0654296875, | |
| "learning_rate": 0.0001831848688224527, | |
| "loss": 0.0096, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 0.5826957033549146, | |
| "grad_norm": 0.048828125, | |
| "learning_rate": 0.00018316046369737645, | |
| "loss": 0.0074, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.583284284873455, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.0001831360585723002, | |
| "loss": 0.0092, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 0.5838728663919953, | |
| "grad_norm": 0.0556640625, | |
| "learning_rate": 0.00018311165344722392, | |
| "loss": 0.0087, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 0.5844614479105356, | |
| "grad_norm": 0.10302734375, | |
| "learning_rate": 0.00018308724832214767, | |
| "loss": 0.0112, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 0.5850500294290759, | |
| "grad_norm": 0.055908203125, | |
| "learning_rate": 0.00018306284319707139, | |
| "loss": 0.0084, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 0.5856386109476163, | |
| "grad_norm": 0.0595703125, | |
| "learning_rate": 0.00018303843807199513, | |
| "loss": 0.0089, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.5862271924661565, | |
| "grad_norm": 0.053955078125, | |
| "learning_rate": 0.00018301403294691885, | |
| "loss": 0.0094, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 0.5868157739846969, | |
| "grad_norm": 0.06494140625, | |
| "learning_rate": 0.0001829896278218426, | |
| "loss": 0.0099, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 0.5874043555032372, | |
| "grad_norm": 0.044189453125, | |
| "learning_rate": 0.00018296522269676632, | |
| "loss": 0.0076, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 0.5879929370217775, | |
| "grad_norm": 0.056396484375, | |
| "learning_rate": 0.00018294081757169007, | |
| "loss": 0.009, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 0.5885815185403178, | |
| "grad_norm": 0.3515625, | |
| "learning_rate": 0.0001829164124466138, | |
| "loss": 0.0125, | |
| "step": 1000 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 8495, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |