| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 1700, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.000588235294117647, | |
| "grad_norm": 86.25792069448507, | |
| "learning_rate": 5.882352941176471e-08, | |
| "loss": 4.144, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0029411764705882353, | |
| "grad_norm": 92.94334261133154, | |
| "learning_rate": 2.9411764705882356e-07, | |
| "loss": 4.2053, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0058823529411764705, | |
| "grad_norm": 34.41771480337444, | |
| "learning_rate": 5.882352941176471e-07, | |
| "loss": 3.9703, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.008823529411764706, | |
| "grad_norm": 20.69702610638224, | |
| "learning_rate": 8.823529411764707e-07, | |
| "loss": 3.4529, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.011764705882352941, | |
| "grad_norm": 12.865609762875932, | |
| "learning_rate": 1.1764705882352942e-06, | |
| "loss": 3.1192, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.014705882352941176, | |
| "grad_norm": 14.658994779954883, | |
| "learning_rate": 1.4705882352941177e-06, | |
| "loss": 2.9015, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.01764705882352941, | |
| "grad_norm": 19.64009310214488, | |
| "learning_rate": 1.7647058823529414e-06, | |
| "loss": 2.5883, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.020588235294117647, | |
| "grad_norm": 27.20931711935924, | |
| "learning_rate": 2.058823529411765e-06, | |
| "loss": 2.2725, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.023529411764705882, | |
| "grad_norm": 23.446531565331924, | |
| "learning_rate": 2.3529411764705885e-06, | |
| "loss": 1.8073, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.026470588235294117, | |
| "grad_norm": 13.31815173013865, | |
| "learning_rate": 2.647058823529412e-06, | |
| "loss": 1.3518, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.029411764705882353, | |
| "grad_norm": 5.34093550595349, | |
| "learning_rate": 2.9411764705882355e-06, | |
| "loss": 1.1676, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03235294117647059, | |
| "grad_norm": 3.3456911840455628, | |
| "learning_rate": 3.2352941176470594e-06, | |
| "loss": 1.147, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.03529411764705882, | |
| "grad_norm": 5.032424322254707, | |
| "learning_rate": 3.529411764705883e-06, | |
| "loss": 1.0965, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03823529411764706, | |
| "grad_norm": 4.206085507780853, | |
| "learning_rate": 3.8235294117647055e-06, | |
| "loss": 1.0636, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.041176470588235294, | |
| "grad_norm": 8.089055829866213, | |
| "learning_rate": 4.11764705882353e-06, | |
| "loss": 1.0582, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.04411764705882353, | |
| "grad_norm": 3.347364025882152, | |
| "learning_rate": 4.411764705882353e-06, | |
| "loss": 1.0539, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.047058823529411764, | |
| "grad_norm": 6.336557333099317, | |
| "learning_rate": 4.705882352941177e-06, | |
| "loss": 1.0379, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 10.978187369479212, | |
| "learning_rate": 5e-06, | |
| "loss": 1.0242, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.052941176470588235, | |
| "grad_norm": 7.818112219971164, | |
| "learning_rate": 5.294117647058824e-06, | |
| "loss": 1.0016, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.05588235294117647, | |
| "grad_norm": 8.433408497312906, | |
| "learning_rate": 5.588235294117647e-06, | |
| "loss": 0.9771, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.058823529411764705, | |
| "grad_norm": 6.212980683823607, | |
| "learning_rate": 5.882352941176471e-06, | |
| "loss": 0.9861, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.061764705882352944, | |
| "grad_norm": 6.647220375875289, | |
| "learning_rate": 6.176470588235295e-06, | |
| "loss": 0.9602, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.06470588235294118, | |
| "grad_norm": 6.527727138992631, | |
| "learning_rate": 6.470588235294119e-06, | |
| "loss": 0.9525, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.06764705882352941, | |
| "grad_norm": 7.509553679384927, | |
| "learning_rate": 6.764705882352942e-06, | |
| "loss": 0.9416, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.07058823529411765, | |
| "grad_norm": 8.103718249384503, | |
| "learning_rate": 7.058823529411766e-06, | |
| "loss": 0.9267, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.07352941176470588, | |
| "grad_norm": 8.520097429021, | |
| "learning_rate": 7.352941176470589e-06, | |
| "loss": 0.9396, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.07647058823529412, | |
| "grad_norm": 9.982219276844697, | |
| "learning_rate": 7.647058823529411e-06, | |
| "loss": 0.9259, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.07941176470588235, | |
| "grad_norm": 10.609383365594049, | |
| "learning_rate": 7.941176470588236e-06, | |
| "loss": 0.9262, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.08235294117647059, | |
| "grad_norm": 6.496638967078595, | |
| "learning_rate": 8.23529411764706e-06, | |
| "loss": 0.9169, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.08529411764705883, | |
| "grad_norm": 5.899471704114852, | |
| "learning_rate": 8.529411764705883e-06, | |
| "loss": 0.9172, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.08823529411764706, | |
| "grad_norm": 19.813714293488992, | |
| "learning_rate": 8.823529411764707e-06, | |
| "loss": 0.8899, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.09117647058823529, | |
| "grad_norm": 3.91249318240706, | |
| "learning_rate": 9.11764705882353e-06, | |
| "loss": 0.9074, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.09411764705882353, | |
| "grad_norm": 10.124669032565054, | |
| "learning_rate": 9.411764705882354e-06, | |
| "loss": 0.898, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.09705882352941177, | |
| "grad_norm": 13.840426312968154, | |
| "learning_rate": 9.705882352941177e-06, | |
| "loss": 0.9049, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 10.60421096496902, | |
| "learning_rate": 1e-05, | |
| "loss": 0.9154, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.10294117647058823, | |
| "grad_norm": 6.134422545965741, | |
| "learning_rate": 9.999736492435867e-06, | |
| "loss": 0.8836, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.10588235294117647, | |
| "grad_norm": 11.600519341121485, | |
| "learning_rate": 9.998945997517957e-06, | |
| "loss": 0.8904, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.10882352941176471, | |
| "grad_norm": 11.799379752373861, | |
| "learning_rate": 9.99762859856683e-06, | |
| "loss": 0.8941, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.11176470588235295, | |
| "grad_norm": 6.623810329677796, | |
| "learning_rate": 9.99578443444032e-06, | |
| "loss": 0.8729, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.11470588235294117, | |
| "grad_norm": 6.965926377173582, | |
| "learning_rate": 9.993413699518906e-06, | |
| "loss": 0.8733, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.11764705882352941, | |
| "grad_norm": 4.929345626084727, | |
| "learning_rate": 9.990516643685222e-06, | |
| "loss": 0.8559, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.12058823529411765, | |
| "grad_norm": 4.438658396751935, | |
| "learning_rate": 9.987093572297716e-06, | |
| "loss": 0.8693, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.12352941176470589, | |
| "grad_norm": 9.676542191697266, | |
| "learning_rate": 9.983144846158472e-06, | |
| "loss": 0.8788, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.1264705882352941, | |
| "grad_norm": 10.638416959011728, | |
| "learning_rate": 9.978670881475173e-06, | |
| "loss": 0.8517, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.12941176470588237, | |
| "grad_norm": 15.793160052452487, | |
| "learning_rate": 9.973672149817232e-06, | |
| "loss": 0.8362, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1323529411764706, | |
| "grad_norm": 14.96288140935782, | |
| "learning_rate": 9.96814917806609e-06, | |
| "loss": 0.8517, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.13529411764705881, | |
| "grad_norm": 17.94048430648359, | |
| "learning_rate": 9.96210254835968e-06, | |
| "loss": 0.8452, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.13823529411764707, | |
| "grad_norm": 10.072819918084187, | |
| "learning_rate": 9.955532898031069e-06, | |
| "loss": 0.8391, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.1411764705882353, | |
| "grad_norm": 7.684632649087268, | |
| "learning_rate": 9.948440919541277e-06, | |
| "loss": 0.8289, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.14411764705882352, | |
| "grad_norm": 17.954771803869264, | |
| "learning_rate": 9.940827360406297e-06, | |
| "loss": 0.8067, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.14705882352941177, | |
| "grad_norm": 12.023797137924742, | |
| "learning_rate": 9.932693023118299e-06, | |
| "loss": 0.8175, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 10.948930140888852, | |
| "learning_rate": 9.924038765061042e-06, | |
| "loss": 0.8226, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.15294117647058825, | |
| "grad_norm": 17.86031699503285, | |
| "learning_rate": 9.91486549841951e-06, | |
| "loss": 0.8168, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.15588235294117647, | |
| "grad_norm": 10.990033307622202, | |
| "learning_rate": 9.905174190083763e-06, | |
| "loss": 0.8047, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.1588235294117647, | |
| "grad_norm": 8.715291883213558, | |
| "learning_rate": 9.894965861547023e-06, | |
| "loss": 0.8007, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.16176470588235295, | |
| "grad_norm": 13.856071733371282, | |
| "learning_rate": 9.884241588798004e-06, | |
| "loss": 0.8017, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.16470588235294117, | |
| "grad_norm": 15.297710097108334, | |
| "learning_rate": 9.873002502207502e-06, | |
| "loss": 0.8055, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.1676470588235294, | |
| "grad_norm": 10.725176474324847, | |
| "learning_rate": 9.861249786409248e-06, | |
| "loss": 0.7972, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.17058823529411765, | |
| "grad_norm": 12.860465116141345, | |
| "learning_rate": 9.848984680175049e-06, | |
| "loss": 0.8145, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.17352941176470588, | |
| "grad_norm": 13.411817152907572, | |
| "learning_rate": 9.836208476284208e-06, | |
| "loss": 0.7992, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.17647058823529413, | |
| "grad_norm": 10.317196889908503, | |
| "learning_rate": 9.822922521387277e-06, | |
| "loss": 0.7846, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.17941176470588235, | |
| "grad_norm": 4.179704998479472, | |
| "learning_rate": 9.809128215864096e-06, | |
| "loss": 0.8025, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.18235294117647058, | |
| "grad_norm": 6.430799430688913, | |
| "learning_rate": 9.794827013676206e-06, | |
| "loss": 0.7991, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.18529411764705883, | |
| "grad_norm": 5.850743250582622, | |
| "learning_rate": 9.78002042221359e-06, | |
| "loss": 0.8157, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.18823529411764706, | |
| "grad_norm": 7.981852850027179, | |
| "learning_rate": 9.764710002135784e-06, | |
| "loss": 0.8315, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.19117647058823528, | |
| "grad_norm": 4.246718673048229, | |
| "learning_rate": 9.748897367207391e-06, | |
| "loss": 0.8126, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.19411764705882353, | |
| "grad_norm": 3.197412855424511, | |
| "learning_rate": 9.732584184127973e-06, | |
| "loss": 0.8054, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.19705882352941176, | |
| "grad_norm": 3.866629151755055, | |
| "learning_rate": 9.715772172356388e-06, | |
| "loss": 0.8106, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 3.4312289366659603, | |
| "learning_rate": 9.698463103929542e-06, | |
| "loss": 0.803, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.20294117647058824, | |
| "grad_norm": 4.537035689074838, | |
| "learning_rate": 9.68065880327562e-06, | |
| "loss": 0.8197, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.20588235294117646, | |
| "grad_norm": 4.167020748172961, | |
| "learning_rate": 9.66236114702178e-06, | |
| "loss": 0.8128, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.2088235294117647, | |
| "grad_norm": 7.084719945270455, | |
| "learning_rate": 9.643572063796352e-06, | |
| "loss": 0.7983, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.21176470588235294, | |
| "grad_norm": 4.345120253918932, | |
| "learning_rate": 9.62429353402556e-06, | |
| "loss": 0.7932, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.21470588235294116, | |
| "grad_norm": 3.8354401768943536, | |
| "learning_rate": 9.60452758972477e-06, | |
| "loss": 0.7911, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.21764705882352942, | |
| "grad_norm": 3.895028761361664, | |
| "learning_rate": 9.584276314284316e-06, | |
| "loss": 0.7903, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.22058823529411764, | |
| "grad_norm": 4.768654905615113, | |
| "learning_rate": 9.563541842249903e-06, | |
| "loss": 0.814, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.2235294117647059, | |
| "grad_norm": 4.001695620575136, | |
| "learning_rate": 9.542326359097619e-06, | |
| "loss": 0.7983, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.22647058823529412, | |
| "grad_norm": 5.464979383463447, | |
| "learning_rate": 9.520632101003579e-06, | |
| "loss": 0.8113, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.22941176470588234, | |
| "grad_norm": 4.410052168842928, | |
| "learning_rate": 9.498461354608228e-06, | |
| "loss": 0.7871, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2323529411764706, | |
| "grad_norm": 10.214130104526346, | |
| "learning_rate": 9.475816456775313e-06, | |
| "loss": 0.7673, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.23529411764705882, | |
| "grad_norm": 7.775696918112043, | |
| "learning_rate": 9.452699794345583e-06, | |
| "loss": 0.8504, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.23823529411764705, | |
| "grad_norm": 4.793572966156681, | |
| "learning_rate": 9.429113803885199e-06, | |
| "loss": 0.7983, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.2411764705882353, | |
| "grad_norm": 6.2733719119866755, | |
| "learning_rate": 9.405060971428924e-06, | |
| "loss": 0.7545, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.24411764705882352, | |
| "grad_norm": 5.665044173485091, | |
| "learning_rate": 9.380543832218069e-06, | |
| "loss": 0.7694, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.24705882352941178, | |
| "grad_norm": 6.667717620194777, | |
| "learning_rate": 9.355564970433288e-06, | |
| "loss": 0.7749, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 3.3946898220456827, | |
| "learning_rate": 9.330127018922195e-06, | |
| "loss": 0.7857, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.2529411764705882, | |
| "grad_norm": 7.472860005366773, | |
| "learning_rate": 9.30423265892184e-06, | |
| "loss": 0.7711, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.25588235294117645, | |
| "grad_norm": 4.70932174586482, | |
| "learning_rate": 9.277884619776116e-06, | |
| "loss": 0.7765, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.25882352941176473, | |
| "grad_norm": 9.57060451964989, | |
| "learning_rate": 9.251085678648072e-06, | |
| "loss": 0.7604, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.26176470588235295, | |
| "grad_norm": 4.681414682219643, | |
| "learning_rate": 9.223838660227183e-06, | |
| "loss": 0.7747, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.2647058823529412, | |
| "grad_norm": 7.367138755832612, | |
| "learning_rate": 9.196146436431635e-06, | |
| "loss": 0.7634, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.2676470588235294, | |
| "grad_norm": 5.627007459776944, | |
| "learning_rate": 9.168011926105598e-06, | |
| "loss": 0.7641, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.27058823529411763, | |
| "grad_norm": 3.5740204640033078, | |
| "learning_rate": 9.13943809471159e-06, | |
| "loss": 0.7565, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.2735294117647059, | |
| "grad_norm": 7.69019676925858, | |
| "learning_rate": 9.110427954017891e-06, | |
| "loss": 0.7733, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.27647058823529413, | |
| "grad_norm": 3.5962732922894975, | |
| "learning_rate": 9.08098456178111e-06, | |
| "loss": 0.765, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.27941176470588236, | |
| "grad_norm": 4.241563630767514, | |
| "learning_rate": 9.051111021423868e-06, | |
| "loss": 0.7754, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.2823529411764706, | |
| "grad_norm": 7.8000912530998745, | |
| "learning_rate": 9.020810481707709e-06, | |
| "loss": 0.7585, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.2852941176470588, | |
| "grad_norm": 3.186234693569658, | |
| "learning_rate": 8.990086136401199e-06, | |
| "loss": 0.7547, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.28823529411764703, | |
| "grad_norm": 7.586644694603505, | |
| "learning_rate": 8.958941223943292e-06, | |
| "loss": 0.7427, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.2911764705882353, | |
| "grad_norm": 3.258615795641642, | |
| "learning_rate": 8.927379027101994e-06, | |
| "loss": 0.7399, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.29411764705882354, | |
| "grad_norm": 3.731634633993737, | |
| "learning_rate": 8.895402872628352e-06, | |
| "loss": 0.7391, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.29705882352941176, | |
| "grad_norm": 4.803677561739918, | |
| "learning_rate": 8.863016130905795e-06, | |
| "loss": 0.7407, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 7.655699535414225, | |
| "learning_rate": 8.83022221559489e-06, | |
| "loss": 0.7506, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.3029411764705882, | |
| "grad_norm": 4.373831528386331, | |
| "learning_rate": 8.797024583273536e-06, | |
| "loss": 0.7355, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.3058823529411765, | |
| "grad_norm": 3.9385838728737266, | |
| "learning_rate": 8.763426733072624e-06, | |
| "loss": 0.735, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.3088235294117647, | |
| "grad_norm": 4.929470086337327, | |
| "learning_rate": 8.729432206307218e-06, | |
| "loss": 0.7407, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.31176470588235294, | |
| "grad_norm": 4.985252344557763, | |
| "learning_rate": 8.695044586103297e-06, | |
| "loss": 0.7368, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.31470588235294117, | |
| "grad_norm": 4.402532806236826, | |
| "learning_rate": 8.660267497020074e-06, | |
| "loss": 0.7378, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.3176470588235294, | |
| "grad_norm": 6.632318764297617, | |
| "learning_rate": 8.625104604667965e-06, | |
| "loss": 0.7182, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.3205882352941177, | |
| "grad_norm": 3.0548060644071566, | |
| "learning_rate": 8.58955961532221e-06, | |
| "loss": 0.7198, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.3235294117647059, | |
| "grad_norm": 7.1426568719242445, | |
| "learning_rate": 8.553636275532236e-06, | |
| "loss": 0.7336, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.3264705882352941, | |
| "grad_norm": 10.88738661858406, | |
| "learning_rate": 8.51733837172675e-06, | |
| "loss": 0.7368, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.32941176470588235, | |
| "grad_norm": 6.282693925795995, | |
| "learning_rate": 8.480669729814635e-06, | |
| "loss": 0.7276, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3323529411764706, | |
| "grad_norm": 9.555955539230844, | |
| "learning_rate": 8.443634214781693e-06, | |
| "loss": 0.702, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.3352941176470588, | |
| "grad_norm": 4.506254396871987, | |
| "learning_rate": 8.40623573028327e-06, | |
| "loss": 0.7129, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.3382352941176471, | |
| "grad_norm": 3.63704291073538, | |
| "learning_rate": 8.368478218232787e-06, | |
| "loss": 0.719, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.3411764705882353, | |
| "grad_norm": 2.9414631231401462, | |
| "learning_rate": 8.330365658386252e-06, | |
| "loss": 0.7119, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.34411764705882353, | |
| "grad_norm": 3.8962634035076675, | |
| "learning_rate": 8.291902067922791e-06, | |
| "loss": 0.7044, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.34705882352941175, | |
| "grad_norm": 2.996619696992914, | |
| "learning_rate": 8.25309150102121e-06, | |
| "loss": 0.7221, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 3.6918987952391653, | |
| "learning_rate": 8.213938048432697e-06, | |
| "loss": 0.7152, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.35294117647058826, | |
| "grad_norm": 5.1406824761237475, | |
| "learning_rate": 8.174445837049614e-06, | |
| "loss": 0.7122, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.3558823529411765, | |
| "grad_norm": 3.334206362728506, | |
| "learning_rate": 8.134619029470535e-06, | |
| "loss": 0.7021, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.3588235294117647, | |
| "grad_norm": 4.968392821499114, | |
| "learning_rate": 8.094461823561473e-06, | |
| "loss": 0.7053, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.36176470588235293, | |
| "grad_norm": 3.1842109451175165, | |
| "learning_rate": 8.05397845201344e-06, | |
| "loss": 0.7142, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.36470588235294116, | |
| "grad_norm": 7.122667502997686, | |
| "learning_rate": 8.013173181896283e-06, | |
| "loss": 0.7158, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.36764705882352944, | |
| "grad_norm": 3.7490095954612563, | |
| "learning_rate": 7.972050314208934e-06, | |
| "loss": 0.7073, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.37058823529411766, | |
| "grad_norm": 3.418137694246427, | |
| "learning_rate": 7.930614183426074e-06, | |
| "loss": 0.7098, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.3735294117647059, | |
| "grad_norm": 3.8866651266730092, | |
| "learning_rate": 7.888869157041257e-06, | |
| "loss": 0.6827, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.3764705882352941, | |
| "grad_norm": 3.5194202567335466, | |
| "learning_rate": 7.846819635106569e-06, | |
| "loss": 0.6891, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.37941176470588234, | |
| "grad_norm": 3.847874692639519, | |
| "learning_rate": 7.80447004976885e-06, | |
| "loss": 0.667, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.38235294117647056, | |
| "grad_norm": 5.871281951965833, | |
| "learning_rate": 7.76182486480253e-06, | |
| "loss": 0.6792, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.38529411764705884, | |
| "grad_norm": 2.9461535013380344, | |
| "learning_rate": 7.718888575139134e-06, | |
| "loss": 0.6821, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.38823529411764707, | |
| "grad_norm": 3.545429320117265, | |
| "learning_rate": 7.675665706393502e-06, | |
| "loss": 0.6837, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.3911764705882353, | |
| "grad_norm": 3.9453651020162406, | |
| "learning_rate": 7.63216081438678e-06, | |
| "loss": 0.6944, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.3941176470588235, | |
| "grad_norm": 4.157385411843414, | |
| "learning_rate": 7.588378484666214e-06, | |
| "loss": 0.6803, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.39705882352941174, | |
| "grad_norm": 2.9879188800031287, | |
| "learning_rate": 7.544323332021826e-06, | |
| "loss": 0.6716, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 5.065231322987666, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 0.6899, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.40294117647058825, | |
| "grad_norm": 5.937109080515486, | |
| "learning_rate": 7.4554131604140425e-06, | |
| "loss": 0.6858, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.40588235294117647, | |
| "grad_norm": 3.589250867729836, | |
| "learning_rate": 7.4105675128517456e-06, | |
| "loss": 0.6911, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.4088235294117647, | |
| "grad_norm": 3.459912742752029, | |
| "learning_rate": 7.365467784180051e-06, | |
| "loss": 0.6931, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.4117647058823529, | |
| "grad_norm": 7.570102416020596, | |
| "learning_rate": 7.320118728046818e-06, | |
| "loss": 0.6919, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.4147058823529412, | |
| "grad_norm": 5.7516115714170235, | |
| "learning_rate": 7.274525124379773e-06, | |
| "loss": 0.6613, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.4176470588235294, | |
| "grad_norm": 6.937823731044612, | |
| "learning_rate": 7.2286917788826926e-06, | |
| "loss": 0.6681, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.42058823529411765, | |
| "grad_norm": 2.702492838292246, | |
| "learning_rate": 7.182623522528866e-06, | |
| "loss": 0.6621, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.4235294117647059, | |
| "grad_norm": 4.510429847780033, | |
| "learning_rate": 7.136325211051905e-06, | |
| "loss": 0.6577, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.4264705882352941, | |
| "grad_norm": 3.6330047784117503, | |
| "learning_rate": 7.089801724433918e-06, | |
| "loss": 0.6682, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.4294117647058823, | |
| "grad_norm": 3.105718892497887, | |
| "learning_rate": 7.043057966391158e-06, | |
| "loss": 0.6632, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.4323529411764706, | |
| "grad_norm": 5.111900502997939, | |
| "learning_rate": 6.996098863857155e-06, | |
| "loss": 0.6486, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.43529411764705883, | |
| "grad_norm": 2.5268899583639315, | |
| "learning_rate": 6.948929366463397e-06, | |
| "loss": 0.6523, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.43823529411764706, | |
| "grad_norm": 3.2780721096181407, | |
| "learning_rate": 6.9015544460176296e-06, | |
| "loss": 0.655, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.4411764705882353, | |
| "grad_norm": 3.3381264245743596, | |
| "learning_rate": 6.8539790959798045e-06, | |
| "loss": 0.654, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.4441176470588235, | |
| "grad_norm": 4.44121574928845, | |
| "learning_rate": 6.806208330935766e-06, | |
| "loss": 0.6372, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.4470588235294118, | |
| "grad_norm": 2.685752664996779, | |
| "learning_rate": 6.758247186068684e-06, | |
| "loss": 0.6413, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.955512012069018, | |
| "learning_rate": 6.710100716628345e-06, | |
| "loss": 0.6516, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.45294117647058824, | |
| "grad_norm": 3.1611374536081454, | |
| "learning_rate": 6.6617739973982985e-06, | |
| "loss": 0.6649, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.45588235294117646, | |
| "grad_norm": 4.369165428910403, | |
| "learning_rate": 6.613272122160975e-06, | |
| "loss": 0.6622, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.4588235294117647, | |
| "grad_norm": 3.614068134809545, | |
| "learning_rate": 6.5646002031607726e-06, | |
| "loss": 0.6469, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.46176470588235297, | |
| "grad_norm": 2.902322450613806, | |
| "learning_rate": 6.515763370565218e-06, | |
| "loss": 0.631, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.4647058823529412, | |
| "grad_norm": 5.870110269988736, | |
| "learning_rate": 6.466766771924231e-06, | |
| "loss": 0.652, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.4676470588235294, | |
| "grad_norm": 3.9059862931450313, | |
| "learning_rate": 6.417615571627555e-06, | |
| "loss": 0.6318, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.47058823529411764, | |
| "grad_norm": 2.8197589495167192, | |
| "learning_rate": 6.368314950360416e-06, | |
| "loss": 0.6322, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.47352941176470587, | |
| "grad_norm": 3.418950733265166, | |
| "learning_rate": 6.318870104557459e-06, | |
| "loss": 0.6187, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.4764705882352941, | |
| "grad_norm": 4.432452311916624, | |
| "learning_rate": 6.269286245855039e-06, | |
| "loss": 0.602, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.47941176470588237, | |
| "grad_norm": 2.5812197112565887, | |
| "learning_rate": 6.219568600541886e-06, | |
| "loss": 0.6174, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.4823529411764706, | |
| "grad_norm": 3.937227710483061, | |
| "learning_rate": 6.169722409008244e-06, | |
| "loss": 0.619, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.4852941176470588, | |
| "grad_norm": 3.0605504256904443, | |
| "learning_rate": 6.119752925193516e-06, | |
| "loss": 0.6334, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.48823529411764705, | |
| "grad_norm": 4.037595617734812, | |
| "learning_rate": 6.0696654160324875e-06, | |
| "loss": 0.6279, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.49117647058823527, | |
| "grad_norm": 5.106015487464359, | |
| "learning_rate": 6.019465160900173e-06, | |
| "loss": 0.6354, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.49411764705882355, | |
| "grad_norm": 3.845547749417895, | |
| "learning_rate": 5.9691574510553505e-06, | |
| "loss": 0.6415, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.4970588235294118, | |
| "grad_norm": 3.1754711810999394, | |
| "learning_rate": 5.918747589082853e-06, | |
| "loss": 0.6243, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 3.8222639828081775, | |
| "learning_rate": 5.8682408883346535e-06, | |
| "loss": 0.6065, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.5029411764705882, | |
| "grad_norm": 3.5634408734197054, | |
| "learning_rate": 5.817642672369825e-06, | |
| "loss": 0.6008, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.5058823529411764, | |
| "grad_norm": 2.923823486420052, | |
| "learning_rate": 5.766958274393428e-06, | |
| "loss": 0.6266, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.5088235294117647, | |
| "grad_norm": 3.4223776040897373, | |
| "learning_rate": 5.716193036694359e-06, | |
| "loss": 0.6141, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.5117647058823529, | |
| "grad_norm": 3.036605144513134, | |
| "learning_rate": 5.66535231008227e-06, | |
| "loss": 0.598, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5147058823529411, | |
| "grad_norm": 2.9215046859208695, | |
| "learning_rate": 5.614441453323571e-06, | |
| "loss": 0.6074, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.5176470588235295, | |
| "grad_norm": 3.5327811692579387, | |
| "learning_rate": 5.5634658325766066e-06, | |
| "loss": 0.6052, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.5205882352941177, | |
| "grad_norm": 4.232585058220642, | |
| "learning_rate": 5.512430820826035e-06, | |
| "loss": 0.6165, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.5235294117647059, | |
| "grad_norm": 7.156188965411479, | |
| "learning_rate": 5.46134179731651e-06, | |
| "loss": 0.6012, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5264705882352941, | |
| "grad_norm": 3.0787975216808356, | |
| "learning_rate": 5.41020414698569e-06, | |
| "loss": 0.5919, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.5294117647058824, | |
| "grad_norm": 2.641117254170942, | |
| "learning_rate": 5.359023259896638e-06, | |
| "loss": 0.5921, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.5323529411764706, | |
| "grad_norm": 5.7492713146889205, | |
| "learning_rate": 5.3078045306697154e-06, | |
| "loss": 0.6009, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.5352941176470588, | |
| "grad_norm": 3.0420972441444984, | |
| "learning_rate": 5.2565533579139484e-06, | |
| "loss": 0.5803, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.538235294117647, | |
| "grad_norm": 4.94141259844524, | |
| "learning_rate": 5.205275143658018e-06, | |
| "loss": 0.5777, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.5411764705882353, | |
| "grad_norm": 2.8294470597059362, | |
| "learning_rate": 5.153975292780852e-06, | |
| "loss": 0.5853, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.5441176470588235, | |
| "grad_norm": 2.657834885061226, | |
| "learning_rate": 5.102659212441953e-06, | |
| "loss": 0.5753, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.5470588235294118, | |
| "grad_norm": 2.5027366282176264, | |
| "learning_rate": 5.05133231151145e-06, | |
| "loss": 0.5979, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 2.972259109016088, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5934, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.5529411764705883, | |
| "grad_norm": 3.7344170961635235, | |
| "learning_rate": 4.948667688488552e-06, | |
| "loss": 0.5765, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.5558823529411765, | |
| "grad_norm": 3.6094582616510187, | |
| "learning_rate": 4.8973407875580485e-06, | |
| "loss": 0.5763, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.5588235294117647, | |
| "grad_norm": 3.0365592437357183, | |
| "learning_rate": 4.846024707219149e-06, | |
| "loss": 0.5637, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.5617647058823529, | |
| "grad_norm": 2.689099476865793, | |
| "learning_rate": 4.794724856341985e-06, | |
| "loss": 0.5702, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.5647058823529412, | |
| "grad_norm": 3.1936877032969977, | |
| "learning_rate": 4.7434466420860515e-06, | |
| "loss": 0.566, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5676470588235294, | |
| "grad_norm": 3.3687980239848976, | |
| "learning_rate": 4.692195469330286e-06, | |
| "loss": 0.5855, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.5705882352941176, | |
| "grad_norm": 3.652904628883288, | |
| "learning_rate": 4.640976740103363e-06, | |
| "loss": 0.5712, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.5735294117647058, | |
| "grad_norm": 2.6471377767162116, | |
| "learning_rate": 4.589795853014313e-06, | |
| "loss": 0.5764, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.5764705882352941, | |
| "grad_norm": 3.4072798307585845, | |
| "learning_rate": 4.53865820268349e-06, | |
| "loss": 0.5584, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.5794117647058824, | |
| "grad_norm": 2.9495397115764717, | |
| "learning_rate": 4.4875691791739655e-06, | |
| "loss": 0.5603, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.5823529411764706, | |
| "grad_norm": 2.967359329887194, | |
| "learning_rate": 4.436534167423395e-06, | |
| "loss": 0.5606, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.5852941176470589, | |
| "grad_norm": 3.6733313275854824, | |
| "learning_rate": 4.3855585466764305e-06, | |
| "loss": 0.5753, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.5882352941176471, | |
| "grad_norm": 3.3017502365781692, | |
| "learning_rate": 4.334647689917734e-06, | |
| "loss": 0.562, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.5911764705882353, | |
| "grad_norm": 2.60212577629923, | |
| "learning_rate": 4.283806963305644e-06, | |
| "loss": 0.5614, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.5941176470588235, | |
| "grad_norm": 3.2117623019305164, | |
| "learning_rate": 4.233041725606573e-06, | |
| "loss": 0.5561, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.5970588235294118, | |
| "grad_norm": 2.5342533645318697, | |
| "learning_rate": 4.182357327630175e-06, | |
| "loss": 0.5456, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 3.491419616001698, | |
| "learning_rate": 4.131759111665349e-06, | |
| "loss": 0.5458, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.6029411764705882, | |
| "grad_norm": 2.5355347312480108, | |
| "learning_rate": 4.081252410917148e-06, | |
| "loss": 0.5639, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.6058823529411764, | |
| "grad_norm": 3.9415364552013177, | |
| "learning_rate": 4.03084254894465e-06, | |
| "loss": 0.5532, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.6088235294117647, | |
| "grad_norm": 4.101652297241306, | |
| "learning_rate": 3.980534839099829e-06, | |
| "loss": 0.5513, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.611764705882353, | |
| "grad_norm": 2.8200195019787944, | |
| "learning_rate": 3.930334583967514e-06, | |
| "loss": 0.543, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.6147058823529412, | |
| "grad_norm": 5.052470209996225, | |
| "learning_rate": 3.8802470748064855e-06, | |
| "loss": 0.5317, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.6176470588235294, | |
| "grad_norm": 3.0712588067545545, | |
| "learning_rate": 3.8302775909917585e-06, | |
| "loss": 0.5276, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.6205882352941177, | |
| "grad_norm": 2.8843549116949676, | |
| "learning_rate": 3.7804313994581143e-06, | |
| "loss": 0.5436, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.6235294117647059, | |
| "grad_norm": 3.71731186768865, | |
| "learning_rate": 3.730713754144961e-06, | |
| "loss": 0.542, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.6264705882352941, | |
| "grad_norm": 3.195039960104732, | |
| "learning_rate": 3.68112989544254e-06, | |
| "loss": 0.5569, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.6294117647058823, | |
| "grad_norm": 3.0760143966442426, | |
| "learning_rate": 3.6316850496395863e-06, | |
| "loss": 0.5279, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.6323529411764706, | |
| "grad_norm": 2.706150933189529, | |
| "learning_rate": 3.5823844283724464e-06, | |
| "loss": 0.5325, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.6352941176470588, | |
| "grad_norm": 2.8909978900989683, | |
| "learning_rate": 3.5332332280757706e-06, | |
| "loss": 0.5275, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.638235294117647, | |
| "grad_norm": 2.4810301000022346, | |
| "learning_rate": 3.484236629434783e-06, | |
| "loss": 0.5449, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.6411764705882353, | |
| "grad_norm": 2.6110146697831036, | |
| "learning_rate": 3.4353997968392295e-06, | |
| "loss": 0.5348, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.6441176470588236, | |
| "grad_norm": 2.6207736282199208, | |
| "learning_rate": 3.386727877839027e-06, | |
| "loss": 0.5362, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.6470588235294118, | |
| "grad_norm": 2.81407811382054, | |
| "learning_rate": 3.3382260026017027e-06, | |
| "loss": 0.5238, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 2.572013835784688, | |
| "learning_rate": 3.289899283371657e-06, | |
| "loss": 0.5251, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.6529411764705882, | |
| "grad_norm": 2.587794908048521, | |
| "learning_rate": 3.241752813931316e-06, | |
| "loss": 0.5362, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.6558823529411765, | |
| "grad_norm": 2.6632456475884423, | |
| "learning_rate": 3.1937916690642356e-06, | |
| "loss": 0.524, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.6588235294117647, | |
| "grad_norm": 3.2994948697776407, | |
| "learning_rate": 3.1460209040201967e-06, | |
| "loss": 0.5246, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.6617647058823529, | |
| "grad_norm": 2.5014355845754426, | |
| "learning_rate": 3.098445553982372e-06, | |
| "loss": 0.5201, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.6647058823529411, | |
| "grad_norm": 3.615311051852331, | |
| "learning_rate": 3.0510706335366034e-06, | |
| "loss": 0.5328, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.6676470588235294, | |
| "grad_norm": 2.8016358462604716, | |
| "learning_rate": 3.0039011361428466e-06, | |
| "loss": 0.5307, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.6705882352941176, | |
| "grad_norm": 2.5797528822198896, | |
| "learning_rate": 2.956942033608843e-06, | |
| "loss": 0.5086, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.6735294117647059, | |
| "grad_norm": 3.703697916863966, | |
| "learning_rate": 2.910198275566085e-06, | |
| "loss": 0.5119, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.6764705882352942, | |
| "grad_norm": 2.4989182080083716, | |
| "learning_rate": 2.863674788948097e-06, | |
| "loss": 0.506, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.6794117647058824, | |
| "grad_norm": 2.926008906071238, | |
| "learning_rate": 2.817376477471132e-06, | |
| "loss": 0.5061, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.6823529411764706, | |
| "grad_norm": 2.5057586009739117, | |
| "learning_rate": 2.771308221117309e-06, | |
| "loss": 0.5158, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.6852941176470588, | |
| "grad_norm": 2.556924814964562, | |
| "learning_rate": 2.725474875620228e-06, | |
| "loss": 0.5241, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.6882352941176471, | |
| "grad_norm": 2.6209251376261586, | |
| "learning_rate": 2.6798812719531843e-06, | |
| "loss": 0.5139, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.6911764705882353, | |
| "grad_norm": 3.565884155626241, | |
| "learning_rate": 2.6345322158199503e-06, | |
| "loss": 0.5196, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.6941176470588235, | |
| "grad_norm": 2.711017441488018, | |
| "learning_rate": 2.5894324871482557e-06, | |
| "loss": 0.4967, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.6970588235294117, | |
| "grad_norm": 2.4420021545344355, | |
| "learning_rate": 2.544586839585961e-06, | |
| "loss": 0.4988, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 3.318545670339647, | |
| "learning_rate": 2.5000000000000015e-06, | |
| "loss": 0.4935, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.7029411764705882, | |
| "grad_norm": 2.7341852078487987, | |
| "learning_rate": 2.4556766679781763e-06, | |
| "loss": 0.4996, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.7058823529411765, | |
| "grad_norm": 2.5448910575508994, | |
| "learning_rate": 2.411621515333788e-06, | |
| "loss": 0.5095, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.7088235294117647, | |
| "grad_norm": 2.3716682194974448, | |
| "learning_rate": 2.3678391856132203e-06, | |
| "loss": 0.4869, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.711764705882353, | |
| "grad_norm": 2.5986125491232017, | |
| "learning_rate": 2.324334293606499e-06, | |
| "loss": 0.4964, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.7147058823529412, | |
| "grad_norm": 2.6704869293015565, | |
| "learning_rate": 2.2811114248608675e-06, | |
| "loss": 0.4851, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.7176470588235294, | |
| "grad_norm": 2.434069159222151, | |
| "learning_rate": 2.238175135197471e-06, | |
| "loss": 0.5144, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.7205882352941176, | |
| "grad_norm": 2.8751039146064996, | |
| "learning_rate": 2.1955299502311523e-06, | |
| "loss": 0.4893, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.7235294117647059, | |
| "grad_norm": 2.8716211792398103, | |
| "learning_rate": 2.1531803648934333e-06, | |
| "loss": 0.4998, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.7264705882352941, | |
| "grad_norm": 2.5538647251200417, | |
| "learning_rate": 2.1111308429587446e-06, | |
| "loss": 0.4968, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.7294117647058823, | |
| "grad_norm": 2.3552149327326406, | |
| "learning_rate": 2.069385816573928e-06, | |
| "loss": 0.4751, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.7323529411764705, | |
| "grad_norm": 2.9092840839893994, | |
| "learning_rate": 2.0279496857910667e-06, | |
| "loss": 0.4892, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.7352941176470589, | |
| "grad_norm": 2.7140627954551797, | |
| "learning_rate": 1.9868268181037186e-06, | |
| "loss": 0.5007, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.7382352941176471, | |
| "grad_norm": 2.494839501975104, | |
| "learning_rate": 1.9460215479865613e-06, | |
| "loss": 0.4845, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.7411764705882353, | |
| "grad_norm": 2.3723796356846556, | |
| "learning_rate": 1.9055381764385272e-06, | |
| "loss": 0.4829, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.7441176470588236, | |
| "grad_norm": 2.3842886824345113, | |
| "learning_rate": 1.865380970529469e-06, | |
| "loss": 0.4853, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.7470588235294118, | |
| "grad_norm": 3.0004523133733523, | |
| "learning_rate": 1.8255541629503865e-06, | |
| "loss": 0.4721, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 4.17555574871206, | |
| "learning_rate": 1.7860619515673034e-06, | |
| "loss": 0.4744, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.7529411764705882, | |
| "grad_norm": 2.4034252951946233, | |
| "learning_rate": 1.746908498978791e-06, | |
| "loss": 0.4848, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.7558823529411764, | |
| "grad_norm": 2.57638925801068, | |
| "learning_rate": 1.708097932077213e-06, | |
| "loss": 0.4771, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.7588235294117647, | |
| "grad_norm": 2.648859686704048, | |
| "learning_rate": 1.6696343416137495e-06, | |
| "loss": 0.4977, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.7617647058823529, | |
| "grad_norm": 2.695298184877126, | |
| "learning_rate": 1.6315217817672142e-06, | |
| "loss": 0.4785, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.7647058823529411, | |
| "grad_norm": 2.5055321455184485, | |
| "learning_rate": 1.5937642697167288e-06, | |
| "loss": 0.4734, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.7676470588235295, | |
| "grad_norm": 2.635547159269979, | |
| "learning_rate": 1.5563657852183072e-06, | |
| "loss": 0.4824, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.7705882352941177, | |
| "grad_norm": 2.3664290913138757, | |
| "learning_rate": 1.5193302701853674e-06, | |
| "loss": 0.4622, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.7735294117647059, | |
| "grad_norm": 3.119885061171242, | |
| "learning_rate": 1.4826616282732509e-06, | |
| "loss": 0.4718, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.7764705882352941, | |
| "grad_norm": 2.490972462452176, | |
| "learning_rate": 1.4463637244677648e-06, | |
| "loss": 0.4669, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.7794117647058824, | |
| "grad_norm": 2.458731892804391, | |
| "learning_rate": 1.410440384677791e-06, | |
| "loss": 0.4617, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.7823529411764706, | |
| "grad_norm": 2.958025742420681, | |
| "learning_rate": 1.374895395332037e-06, | |
| "loss": 0.4696, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.7852941176470588, | |
| "grad_norm": 2.9915004376321104, | |
| "learning_rate": 1.339732502979928e-06, | |
| "loss": 0.4636, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.788235294117647, | |
| "grad_norm": 2.966122147337875, | |
| "learning_rate": 1.3049554138967052e-06, | |
| "loss": 0.4506, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.7911764705882353, | |
| "grad_norm": 2.37703783512199, | |
| "learning_rate": 1.2705677936927841e-06, | |
| "loss": 0.47, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.7941176470588235, | |
| "grad_norm": 2.378564585605087, | |
| "learning_rate": 1.2365732669273778e-06, | |
| "loss": 0.4494, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.7970588235294118, | |
| "grad_norm": 2.9848625505080766, | |
| "learning_rate": 1.202975416726464e-06, | |
| "loss": 0.4545, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 2.30742224730397, | |
| "learning_rate": 1.1697777844051105e-06, | |
| "loss": 0.4661, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.8029411764705883, | |
| "grad_norm": 2.6109048720465386, | |
| "learning_rate": 1.1369838690942059e-06, | |
| "loss": 0.4533, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.8058823529411765, | |
| "grad_norm": 2.9803797967996633, | |
| "learning_rate": 1.1045971273716476e-06, | |
| "loss": 0.4437, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.8088235294117647, | |
| "grad_norm": 2.7563794002126416, | |
| "learning_rate": 1.072620972898007e-06, | |
| "loss": 0.4524, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.8117647058823529, | |
| "grad_norm": 2.397027701415243, | |
| "learning_rate": 1.0410587760567104e-06, | |
| "loss": 0.4648, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.8147058823529412, | |
| "grad_norm": 2.862646785242907, | |
| "learning_rate": 1.0099138635988026e-06, | |
| "loss": 0.4592, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.8176470588235294, | |
| "grad_norm": 2.703458227745046, | |
| "learning_rate": 9.791895182922911e-07, | |
| "loss": 0.4656, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.8205882352941176, | |
| "grad_norm": 2.502457085827656, | |
| "learning_rate": 9.488889785761324e-07, | |
| "loss": 0.4487, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.8235294117647058, | |
| "grad_norm": 2.6423546848684167, | |
| "learning_rate": 9.190154382188921e-07, | |
| "loss": 0.4565, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.8264705882352941, | |
| "grad_norm": 2.532810863987098, | |
| "learning_rate": 8.895720459821089e-07, | |
| "loss": 0.4603, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.8294117647058824, | |
| "grad_norm": 2.24073062154865, | |
| "learning_rate": 8.605619052884106e-07, | |
| "loss": 0.455, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.8323529411764706, | |
| "grad_norm": 2.3347312069958477, | |
| "learning_rate": 8.31988073894403e-07, | |
| "loss": 0.4542, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.8352941176470589, | |
| "grad_norm": 2.2510986140412723, | |
| "learning_rate": 8.03853563568367e-07, | |
| "loss": 0.4422, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.8382352941176471, | |
| "grad_norm": 2.466059819140582, | |
| "learning_rate": 7.761613397728174e-07, | |
| "loss": 0.4436, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.8411764705882353, | |
| "grad_norm": 2.314552647197002, | |
| "learning_rate": 7.489143213519301e-07, | |
| "loss": 0.4526, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.8441176470588235, | |
| "grad_norm": 2.2167187458342186, | |
| "learning_rate": 7.221153802238845e-07, | |
| "loss": 0.4396, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.8470588235294118, | |
| "grad_norm": 2.4410281100307536, | |
| "learning_rate": 6.957673410781617e-07, | |
| "loss": 0.4493, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 2.391369587902256, | |
| "learning_rate": 6.698729810778065e-07, | |
| "loss": 0.4559, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.8529411764705882, | |
| "grad_norm": 2.147240414992554, | |
| "learning_rate": 6.444350295667112e-07, | |
| "loss": 0.4423, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.8558823529411764, | |
| "grad_norm": 2.425220228026074, | |
| "learning_rate": 6.194561677819327e-07, | |
| "loss": 0.454, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.8588235294117647, | |
| "grad_norm": 2.478981866538809, | |
| "learning_rate": 5.949390285710777e-07, | |
| "loss": 0.4565, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.861764705882353, | |
| "grad_norm": 2.4324955757388484, | |
| "learning_rate": 5.708861961148004e-07, | |
| "loss": 0.4485, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.8647058823529412, | |
| "grad_norm": 2.518242605502862, | |
| "learning_rate": 5.473002056544191e-07, | |
| "loss": 0.4465, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.8676470588235294, | |
| "grad_norm": 2.4153603903758767, | |
| "learning_rate": 5.241835432246888e-07, | |
| "loss": 0.4449, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.8705882352941177, | |
| "grad_norm": 2.5154960957655463, | |
| "learning_rate": 5.015386453917742e-07, | |
| "loss": 0.4416, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.8735294117647059, | |
| "grad_norm": 2.5868546835610897, | |
| "learning_rate": 4.793678989964207e-07, | |
| "loss": 0.4343, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.8764705882352941, | |
| "grad_norm": 2.483503936880007, | |
| "learning_rate": 4.576736409023813e-07, | |
| "loss": 0.4375, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.8794117647058823, | |
| "grad_norm": 2.354484991161176, | |
| "learning_rate": 4.364581577500987e-07, | |
| "loss": 0.4396, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.8823529411764706, | |
| "grad_norm": 2.32433065376712, | |
| "learning_rate": 4.15723685715686e-07, | |
| "loss": 0.4417, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.8852941176470588, | |
| "grad_norm": 2.3785384367167173, | |
| "learning_rate": 3.9547241027523164e-07, | |
| "loss": 0.4372, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.888235294117647, | |
| "grad_norm": 2.2739855313885484, | |
| "learning_rate": 3.7570646597444196e-07, | |
| "loss": 0.4382, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.8911764705882353, | |
| "grad_norm": 2.4248149464102324, | |
| "learning_rate": 3.564279362036488e-07, | |
| "loss": 0.4504, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.8941176470588236, | |
| "grad_norm": 2.5206865046215623, | |
| "learning_rate": 3.3763885297822153e-07, | |
| "loss": 0.4425, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.8970588235294118, | |
| "grad_norm": 2.321436598632796, | |
| "learning_rate": 3.1934119672438093e-07, | |
| "loss": 0.4508, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 2.2956489839218253, | |
| "learning_rate": 3.015368960704584e-07, | |
| "loss": 0.4218, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.9029411764705882, | |
| "grad_norm": 2.6380831075744045, | |
| "learning_rate": 2.842278276436128e-07, | |
| "loss": 0.4312, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.9058823529411765, | |
| "grad_norm": 2.353068804554375, | |
| "learning_rate": 2.6741581587202747e-07, | |
| "loss": 0.4572, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.9088235294117647, | |
| "grad_norm": 2.3504938166434286, | |
| "learning_rate": 2.511026327926114e-07, | |
| "loss": 0.4415, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.9117647058823529, | |
| "grad_norm": 2.4033670936853864, | |
| "learning_rate": 2.3528999786421758e-07, | |
| "loss": 0.4383, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.9147058823529411, | |
| "grad_norm": 2.5018773276624557, | |
| "learning_rate": 2.1997957778641166e-07, | |
| "loss": 0.448, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.9176470588235294, | |
| "grad_norm": 2.3367055632905753, | |
| "learning_rate": 2.0517298632379445e-07, | |
| "loss": 0.4497, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.9205882352941176, | |
| "grad_norm": 2.765125262874024, | |
| "learning_rate": 1.908717841359048e-07, | |
| "loss": 0.4507, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.9235294117647059, | |
| "grad_norm": 2.376902313703198, | |
| "learning_rate": 1.770774786127244e-07, | |
| "loss": 0.4183, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.9264705882352942, | |
| "grad_norm": 2.413442304678235, | |
| "learning_rate": 1.6379152371579277e-07, | |
| "loss": 0.4334, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.9294117647058824, | |
| "grad_norm": 2.3143531531697596, | |
| "learning_rate": 1.510153198249531e-07, | |
| "loss": 0.436, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.9323529411764706, | |
| "grad_norm": 2.34608810626235, | |
| "learning_rate": 1.3875021359075257e-07, | |
| "loss": 0.44, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.9352941176470588, | |
| "grad_norm": 2.35355433201976, | |
| "learning_rate": 1.2699749779249926e-07, | |
| "loss": 0.4314, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.9382352941176471, | |
| "grad_norm": 2.218100203321306, | |
| "learning_rate": 1.157584112019966e-07, | |
| "loss": 0.4384, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.9411764705882353, | |
| "grad_norm": 2.3877266166874676, | |
| "learning_rate": 1.0503413845297739e-07, | |
| "loss": 0.4357, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.9441176470588235, | |
| "grad_norm": 2.3187529416036936, | |
| "learning_rate": 9.482580991623747e-08, | |
| "loss": 0.4323, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.9470588235294117, | |
| "grad_norm": 2.2887581674114648, | |
| "learning_rate": 8.513450158049109e-08, | |
| "loss": 0.4448, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 2.3836470257770945, | |
| "learning_rate": 7.59612349389599e-08, | |
| "loss": 0.4269, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.9529411764705882, | |
| "grad_norm": 2.233426183677112, | |
| "learning_rate": 6.730697688170251e-08, | |
| "loss": 0.4351, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.9558823529411765, | |
| "grad_norm": 2.2576790941626284, | |
| "learning_rate": 5.917263959370312e-08, | |
| "loss": 0.4355, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.9588235294117647, | |
| "grad_norm": 2.2984294403136163, | |
| "learning_rate": 5.155908045872349e-08, | |
| "loss": 0.4406, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.961764705882353, | |
| "grad_norm": 2.3707792344113945, | |
| "learning_rate": 4.446710196893245e-08, | |
| "loss": 0.4385, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.9647058823529412, | |
| "grad_norm": 2.2246371108164262, | |
| "learning_rate": 3.7897451640321326e-08, | |
| "loss": 0.4246, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.9676470588235294, | |
| "grad_norm": 2.622284137214347, | |
| "learning_rate": 3.185082193391143e-08, | |
| "loss": 0.4279, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.9705882352941176, | |
| "grad_norm": 2.3946023382142383, | |
| "learning_rate": 2.6327850182769065e-08, | |
| "loss": 0.4573, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.9735294117647059, | |
| "grad_norm": 2.354547522725208, | |
| "learning_rate": 2.1329118524827662e-08, | |
| "loss": 0.4405, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.9764705882352941, | |
| "grad_norm": 2.3346382733247815, | |
| "learning_rate": 1.6855153841527915e-08, | |
| "loss": 0.4386, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.9794117647058823, | |
| "grad_norm": 2.3205008178624764, | |
| "learning_rate": 1.2906427702284452e-08, | |
| "loss": 0.4304, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.9823529411764705, | |
| "grad_norm": 2.4270314026839714, | |
| "learning_rate": 9.48335631477948e-09, | |
| "loss": 0.4415, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.9852941176470589, | |
| "grad_norm": 2.6174654383893214, | |
| "learning_rate": 6.586300481095098e-09, | |
| "loss": 0.4272, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.9882352941176471, | |
| "grad_norm": 2.2960002654994973, | |
| "learning_rate": 4.2155655596809455e-09, | |
| "loss": 0.4334, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.9911764705882353, | |
| "grad_norm": 2.443491425156529, | |
| "learning_rate": 2.371401433170495e-09, | |
| "loss": 0.4249, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.9941176470588236, | |
| "grad_norm": 2.2442264144591926, | |
| "learning_rate": 1.054002482043237e-09, | |
| "loss": 0.428, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.9970588235294118, | |
| "grad_norm": 2.4769821745544953, | |
| "learning_rate": 2.6350756413440203e-10, | |
| "loss": 0.4334, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 2.3764816774266806, | |
| "learning_rate": 0.0, | |
| "loss": 0.4299, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_runtime": 4.4513, | |
| "eval_samples_per_second": 2.247, | |
| "eval_steps_per_second": 0.674, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 1700, | |
| "total_flos": 117051280588800.0, | |
| "train_loss": 0.691351400122923, | |
| "train_runtime": 20676.98, | |
| "train_samples_per_second": 1.315, | |
| "train_steps_per_second": 0.082 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 1700, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 117051280588800.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |