| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9993805492463349, |
| "eval_steps": 500, |
| "global_step": 1210, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0008259343382201115, |
| "grad_norm": 23.24616747351659, |
| "learning_rate": 8.264462809917357e-08, |
| "loss": 1.4263, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.0041296716911005575, |
| "grad_norm": 24.213699866017716, |
| "learning_rate": 4.132231404958678e-07, |
| "loss": 1.4129, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.008259343382201115, |
| "grad_norm": 8.707304164936993, |
| "learning_rate": 8.264462809917356e-07, |
| "loss": 1.3289, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.012389015073301672, |
| "grad_norm": 10.918691302163335, |
| "learning_rate": 1.2396694214876035e-06, |
| "loss": 1.1721, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.01651868676440223, |
| "grad_norm": 2.87174085951705, |
| "learning_rate": 1.6528925619834712e-06, |
| "loss": 1.0219, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.020648358455502787, |
| "grad_norm": 3.071874166303048, |
| "learning_rate": 2.066115702479339e-06, |
| "loss": 0.9724, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.024778030146603344, |
| "grad_norm": 2.4133173682706213, |
| "learning_rate": 2.479338842975207e-06, |
| "loss": 0.9426, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.028907701837703904, |
| "grad_norm": 2.236585592513117, |
| "learning_rate": 2.8925619834710743e-06, |
| "loss": 0.921, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.03303737352880446, |
| "grad_norm": 2.385522667127509, |
| "learning_rate": 3.3057851239669424e-06, |
| "loss": 0.9121, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.03716704521990502, |
| "grad_norm": 2.380746269584164, |
| "learning_rate": 3.71900826446281e-06, |
| "loss": 0.9001, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.041296716911005574, |
| "grad_norm": 2.569197332815026, |
| "learning_rate": 4.132231404958678e-06, |
| "loss": 0.8867, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.045426388602106134, |
| "grad_norm": 2.3723293989848213, |
| "learning_rate": 4.5454545454545455e-06, |
| "loss": 0.8707, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.04955606029320669, |
| "grad_norm": 2.3355520193379142, |
| "learning_rate": 4.958677685950414e-06, |
| "loss": 0.8698, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.05368573198430725, |
| "grad_norm": 2.375730929200025, |
| "learning_rate": 5.371900826446281e-06, |
| "loss": 0.859, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.05781540367540781, |
| "grad_norm": 2.3481213761418838, |
| "learning_rate": 5.785123966942149e-06, |
| "loss": 0.8632, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.06194507536650836, |
| "grad_norm": 2.3161171161194254, |
| "learning_rate": 6.198347107438017e-06, |
| "loss": 0.8462, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.06607474705760892, |
| "grad_norm": 2.273903126512892, |
| "learning_rate": 6.611570247933885e-06, |
| "loss": 0.8593, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.07020441874870947, |
| "grad_norm": 2.3651082830117787, |
| "learning_rate": 7.0247933884297525e-06, |
| "loss": 0.8388, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.07433409043981004, |
| "grad_norm": 2.5780580351650584, |
| "learning_rate": 7.43801652892562e-06, |
| "loss": 0.8354, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.0784637621309106, |
| "grad_norm": 2.5881003929304165, |
| "learning_rate": 7.851239669421489e-06, |
| "loss": 0.8177, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.08259343382201115, |
| "grad_norm": 2.3662580429191262, |
| "learning_rate": 8.264462809917356e-06, |
| "loss": 0.8397, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.0867231055131117, |
| "grad_norm": 2.4549245207596173, |
| "learning_rate": 8.677685950413224e-06, |
| "loss": 0.8209, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.09085277720421227, |
| "grad_norm": 2.6986718605229454, |
| "learning_rate": 9.090909090909091e-06, |
| "loss": 0.8076, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.09498244889531282, |
| "grad_norm": 2.6597400001383136, |
| "learning_rate": 9.50413223140496e-06, |
| "loss": 0.812, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.09911212058641337, |
| "grad_norm": 2.6969489551453703, |
| "learning_rate": 9.917355371900828e-06, |
| "loss": 0.8064, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.10324179227751394, |
| "grad_norm": 2.5801588887536204, |
| "learning_rate": 9.999667111219573e-06, |
| "loss": 0.7993, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.1073714639686145, |
| "grad_norm": 2.5177823470227314, |
| "learning_rate": 9.998314826517564e-06, |
| "loss": 0.8123, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.11150113565971505, |
| "grad_norm": 2.3846369879591767, |
| "learning_rate": 9.995922621477252e-06, |
| "loss": 0.8064, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.11563080735081561, |
| "grad_norm": 2.37259551970217, |
| "learning_rate": 9.99249099380692e-06, |
| "loss": 0.7961, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.11976047904191617, |
| "grad_norm": 2.1621998588898346, |
| "learning_rate": 9.988020657471078e-06, |
| "loss": 0.7924, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.12389015073301672, |
| "grad_norm": 2.3342111944519623, |
| "learning_rate": 9.98251254254193e-06, |
| "loss": 0.7785, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.12801982242411727, |
| "grad_norm": 2.3550905722953823, |
| "learning_rate": 9.97596779500586e-06, |
| "loss": 0.7695, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.13214949411521784, |
| "grad_norm": 2.4721774789219344, |
| "learning_rate": 9.968387776525009e-06, |
| "loss": 0.7777, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.1362791658063184, |
| "grad_norm": 2.390765566890305, |
| "learning_rate": 9.959774064153977e-06, |
| "loss": 0.7865, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.14040883749741895, |
| "grad_norm": 2.392413139550164, |
| "learning_rate": 9.950128450011706e-06, |
| "loss": 0.7723, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.14453850918851952, |
| "grad_norm": 2.24515023866806, |
| "learning_rate": 9.939452940908627e-06, |
| "loss": 0.755, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.14866818087962008, |
| "grad_norm": 2.546933963480071, |
| "learning_rate": 9.927749757929125e-06, |
| "loss": 0.7538, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.15279785257072062, |
| "grad_norm": 2.5390977331021136, |
| "learning_rate": 9.915021335969452e-06, |
| "loss": 0.7436, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.1569275242618212, |
| "grad_norm": 2.435644472976208, |
| "learning_rate": 9.901270323231114e-06, |
| "loss": 0.7423, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.16105719595292176, |
| "grad_norm": 2.3962427953474794, |
| "learning_rate": 9.886499580669917e-06, |
| "loss": 0.7402, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.1651868676440223, |
| "grad_norm": 2.293192651317426, |
| "learning_rate": 9.870712181400726e-06, |
| "loss": 0.7328, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.16931653933512286, |
| "grad_norm": 2.3625939065789523, |
| "learning_rate": 9.853911410058097e-06, |
| "loss": 0.7414, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.1734462110262234, |
| "grad_norm": 2.1188762141799633, |
| "learning_rate": 9.836100762112887e-06, |
| "loss": 0.7324, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.17757588271732397, |
| "grad_norm": 2.3459257676868885, |
| "learning_rate": 9.817283943145014e-06, |
| "loss": 0.7216, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.18170555440842454, |
| "grad_norm": 2.353468899573234, |
| "learning_rate": 9.797464868072489e-06, |
| "loss": 0.7244, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.18583522609952507, |
| "grad_norm": 2.204966096546786, |
| "learning_rate": 9.776647660336905e-06, |
| "loss": 0.7193, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.18996489779062564, |
| "grad_norm": 2.310006498983248, |
| "learning_rate": 9.754836651045538e-06, |
| "loss": 0.7161, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.1940945694817262, |
| "grad_norm": 2.5104871936969277, |
| "learning_rate": 9.732036378070243e-06, |
| "loss": 0.7044, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.19822424117282675, |
| "grad_norm": 2.366571213288955, |
| "learning_rate": 9.708251585103324e-06, |
| "loss": 0.7004, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.20235391286392732, |
| "grad_norm": 2.570331809825131, |
| "learning_rate": 9.683487220670595e-06, |
| "loss": 0.7158, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.20648358455502788, |
| "grad_norm": 2.4620380694723147, |
| "learning_rate": 9.657748437101819e-06, |
| "loss": 0.7019, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.21061325624612842, |
| "grad_norm": 2.3314886422942513, |
| "learning_rate": 9.631040589458742e-06, |
| "loss": 0.6989, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.214742927937229, |
| "grad_norm": 2.2564618217324743, |
| "learning_rate": 9.603369234420944e-06, |
| "loss": 0.6947, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.21887259962832956, |
| "grad_norm": 2.2447803221031903, |
| "learning_rate": 9.574740129129767e-06, |
| "loss": 0.6756, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.2230022713194301, |
| "grad_norm": 2.590402933040932, |
| "learning_rate": 9.545159229990493e-06, |
| "loss": 0.6795, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.22713194301053066, |
| "grad_norm": 2.3017458472761634, |
| "learning_rate": 9.514632691433108e-06, |
| "loss": 0.6905, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.23126161470163123, |
| "grad_norm": 2.363184146884425, |
| "learning_rate": 9.483166864631837e-06, |
| "loss": 0.6572, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.23539128639273177, |
| "grad_norm": 2.531694092570608, |
| "learning_rate": 9.450768296183764e-06, |
| "loss": 0.6694, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.23952095808383234, |
| "grad_norm": 2.3202550588633017, |
| "learning_rate": 9.417443726746776e-06, |
| "loss": 0.6704, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.2436506297749329, |
| "grad_norm": 2.3120309970976685, |
| "learning_rate": 9.383200089637143e-06, |
| "loss": 0.6572, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.24778030146603344, |
| "grad_norm": 2.3498499961336687, |
| "learning_rate": 9.348044509387022e-06, |
| "loss": 0.6655, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.251909973157134, |
| "grad_norm": 2.5440021629079665, |
| "learning_rate": 9.311984300262151e-06, |
| "loss": 0.6612, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.25603964484823455, |
| "grad_norm": 2.358901339481344, |
| "learning_rate": 9.275026964740101e-06, |
| "loss": 0.6444, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.2601693165393351, |
| "grad_norm": 2.206937338630286, |
| "learning_rate": 9.237180191949347e-06, |
| "loss": 0.6445, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.2642989882304357, |
| "grad_norm": 2.226748345548228, |
| "learning_rate": 9.198451856069514e-06, |
| "loss": 0.6474, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.26842865992153625, |
| "grad_norm": 2.366603978699467, |
| "learning_rate": 9.158850014693123e-06, |
| "loss": 0.6361, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.2725583316126368, |
| "grad_norm": 2.2830136404441657, |
| "learning_rate": 9.118382907149164e-06, |
| "loss": 0.6395, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.27668800330373733, |
| "grad_norm": 2.4029340921426727, |
| "learning_rate": 9.077058952788888e-06, |
| "loss": 0.6379, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.2808176749948379, |
| "grad_norm": 2.4477206203283512, |
| "learning_rate": 9.034886749234112e-06, |
| "loss": 0.6272, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.28494734668593846, |
| "grad_norm": 2.3054439086007243, |
| "learning_rate": 8.991875070588449e-06, |
| "loss": 0.6279, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.28907701837703903, |
| "grad_norm": 2.2372770463589196, |
| "learning_rate": 8.948032865611823e-06, |
| "loss": 0.6276, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.2932066900681396, |
| "grad_norm": 2.2567729960946803, |
| "learning_rate": 8.90336925585864e-06, |
| "loss": 0.6107, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.29733636175924016, |
| "grad_norm": 2.2634696566360697, |
| "learning_rate": 8.857893533780015e-06, |
| "loss": 0.6222, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.3014660334503407, |
| "grad_norm": 2.2158775605445573, |
| "learning_rate": 8.811615160790428e-06, |
| "loss": 0.6099, |
| "step": 365 |
| }, |
| { |
| "epoch": 0.30559570514144124, |
| "grad_norm": 2.359375453911521, |
| "learning_rate": 8.764543765299245e-06, |
| "loss": 0.6088, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.3097253768325418, |
| "grad_norm": 2.2574003131115017, |
| "learning_rate": 8.716689140707488e-06, |
| "loss": 0.5978, |
| "step": 375 |
| }, |
| { |
| "epoch": 0.3138550485236424, |
| "grad_norm": 2.305475140301284, |
| "learning_rate": 8.668061243370273e-06, |
| "loss": 0.5955, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.31798472021474294, |
| "grad_norm": 2.512743372685998, |
| "learning_rate": 8.61867019052535e-06, |
| "loss": 0.6035, |
| "step": 385 |
| }, |
| { |
| "epoch": 0.3221143919058435, |
| "grad_norm": 2.407216931553165, |
| "learning_rate": 8.568526258188172e-06, |
| "loss": 0.5836, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.326244063596944, |
| "grad_norm": 2.2192688797123066, |
| "learning_rate": 8.517639879013918e-06, |
| "loss": 0.5888, |
| "step": 395 |
| }, |
| { |
| "epoch": 0.3303737352880446, |
| "grad_norm": 2.1358480481396267, |
| "learning_rate": 8.466021640126946e-06, |
| "loss": 0.5916, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.33450340697914516, |
| "grad_norm": 2.1925271239115034, |
| "learning_rate": 8.413682280918094e-06, |
| "loss": 0.577, |
| "step": 405 |
| }, |
| { |
| "epoch": 0.3386330786702457, |
| "grad_norm": 2.2104260296514613, |
| "learning_rate": 8.3606326908103e-06, |
| "loss": 0.5921, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.3427627503613463, |
| "grad_norm": 2.3351708112925946, |
| "learning_rate": 8.306883906993022e-06, |
| "loss": 0.5664, |
| "step": 415 |
| }, |
| { |
| "epoch": 0.3468924220524468, |
| "grad_norm": 2.129763886062084, |
| "learning_rate": 8.25244711212589e-06, |
| "loss": 0.575, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.35102209374354737, |
| "grad_norm": 2.210580574947652, |
| "learning_rate": 8.197333632012123e-06, |
| "loss": 0.5797, |
| "step": 425 |
| }, |
| { |
| "epoch": 0.35515176543464794, |
| "grad_norm": 2.3295885927377387, |
| "learning_rate": 8.141554933242135e-06, |
| "loss": 0.569, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.3592814371257485, |
| "grad_norm": 2.332203260771466, |
| "learning_rate": 8.08512262080787e-06, |
| "loss": 0.573, |
| "step": 435 |
| }, |
| { |
| "epoch": 0.36341110881684907, |
| "grad_norm": 2.13902847391916, |
| "learning_rate": 8.028048435688333e-06, |
| "loss": 0.5788, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.36754078050794964, |
| "grad_norm": 2.3592904701751145, |
| "learning_rate": 7.970344252406832e-06, |
| "loss": 0.569, |
| "step": 445 |
| }, |
| { |
| "epoch": 0.37167045219905015, |
| "grad_norm": 2.4900138472719617, |
| "learning_rate": 7.912022076560426e-06, |
| "loss": 0.5697, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.3758001238901507, |
| "grad_norm": 2.1966947004630826, |
| "learning_rate": 7.853094042322121e-06, |
| "loss": 0.5577, |
| "step": 455 |
| }, |
| { |
| "epoch": 0.3799297955812513, |
| "grad_norm": 2.4924992867127, |
| "learning_rate": 7.7935724099163e-06, |
| "loss": 0.5564, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.38405946727235185, |
| "grad_norm": 2.3168755091791775, |
| "learning_rate": 7.733469563067928e-06, |
| "loss": 0.558, |
| "step": 465 |
| }, |
| { |
| "epoch": 0.3881891389634524, |
| "grad_norm": 2.090377992450065, |
| "learning_rate": 7.67279800642607e-06, |
| "loss": 0.5472, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.392318810654553, |
| "grad_norm": 2.19832848896829, |
| "learning_rate": 7.611570362962247e-06, |
| "loss": 0.547, |
| "step": 475 |
| }, |
| { |
| "epoch": 0.3964484823456535, |
| "grad_norm": 2.139436532548822, |
| "learning_rate": 7.549799371344175e-06, |
| "loss": 0.5376, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.40057815403675406, |
| "grad_norm": 2.1648898845336304, |
| "learning_rate": 7.487497883285428e-06, |
| "loss": 0.5435, |
| "step": 485 |
| }, |
| { |
| "epoch": 0.40470782572785463, |
| "grad_norm": 2.1434615200482403, |
| "learning_rate": 7.424678860871584e-06, |
| "loss": 0.5347, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.4088374974189552, |
| "grad_norm": 2.028970353253071, |
| "learning_rate": 7.361355373863415e-06, |
| "loss": 0.5349, |
| "step": 495 |
| }, |
| { |
| "epoch": 0.41296716911005577, |
| "grad_norm": 2.1772432202945624, |
| "learning_rate": 7.297540596977663e-06, |
| "loss": 0.532, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.41709684080115633, |
| "grad_norm": 2.2430530701967832, |
| "learning_rate": 7.233247807145989e-06, |
| "loss": 0.5385, |
| "step": 505 |
| }, |
| { |
| "epoch": 0.42122651249225684, |
| "grad_norm": 2.0619754497698035, |
| "learning_rate": 7.168490380752648e-06, |
| "loss": 0.5349, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.4253561841833574, |
| "grad_norm": 2.1855473382543993, |
| "learning_rate": 7.103281790851483e-06, |
| "loss": 0.5233, |
| "step": 515 |
| }, |
| { |
| "epoch": 0.429485855874458, |
| "grad_norm": 2.2538669231669712, |
| "learning_rate": 7.037635604362786e-06, |
| "loss": 0.5283, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.43361552756555855, |
| "grad_norm": 2.2135151111234794, |
| "learning_rate": 6.971565479250659e-06, |
| "loss": 0.5167, |
| "step": 525 |
| }, |
| { |
| "epoch": 0.4377451992566591, |
| "grad_norm": 2.383590881907924, |
| "learning_rate": 6.905085161681408e-06, |
| "loss": 0.5219, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.4418748709477597, |
| "grad_norm": 2.3721647738464258, |
| "learning_rate": 6.838208483163601e-06, |
| "loss": 0.5174, |
| "step": 535 |
| }, |
| { |
| "epoch": 0.4460045426388602, |
| "grad_norm": 2.338909227744761, |
| "learning_rate": 6.770949357670358e-06, |
| "loss": 0.5273, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.45013421432996076, |
| "grad_norm": 2.1495131685935163, |
| "learning_rate": 6.703321778744495e-06, |
| "loss": 0.5186, |
| "step": 545 |
| }, |
| { |
| "epoch": 0.4542638860210613, |
| "grad_norm": 2.010655198436141, |
| "learning_rate": 6.635339816587109e-06, |
| "loss": 0.5164, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.4583935577121619, |
| "grad_norm": 2.1095925263837385, |
| "learning_rate": 6.5670176151302136e-06, |
| "loss": 0.4991, |
| "step": 555 |
| }, |
| { |
| "epoch": 0.46252322940326246, |
| "grad_norm": 2.2264319027764383, |
| "learning_rate": 6.4983693890940335e-06, |
| "loss": 0.5195, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.46665290109436297, |
| "grad_norm": 2.1349760132456246, |
| "learning_rate": 6.4294094210295725e-06, |
| "loss": 0.504, |
| "step": 565 |
| }, |
| { |
| "epoch": 0.47078257278546354, |
| "grad_norm": 2.2896775214795033, |
| "learning_rate": 6.360152058347068e-06, |
| "loss": 0.5078, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.4749122444765641, |
| "grad_norm": 2.1221923536777747, |
| "learning_rate": 6.290611710330957e-06, |
| "loss": 0.5059, |
| "step": 575 |
| }, |
| { |
| "epoch": 0.47904191616766467, |
| "grad_norm": 2.1068836258621406, |
| "learning_rate": 6.2208028451419575e-06, |
| "loss": 0.5083, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.48317158785876524, |
| "grad_norm": 2.0972952063729475, |
| "learning_rate": 6.150739986806911e-06, |
| "loss": 0.4923, |
| "step": 585 |
| }, |
| { |
| "epoch": 0.4873012595498658, |
| "grad_norm": 2.233487162462348, |
| "learning_rate": 6.0804377121969985e-06, |
| "loss": 0.4953, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.4914309312409663, |
| "grad_norm": 2.108166108174955, |
| "learning_rate": 6.009910647994956e-06, |
| "loss": 0.4918, |
| "step": 595 |
| }, |
| { |
| "epoch": 0.4955606029320669, |
| "grad_norm": 2.055777194538188, |
| "learning_rate": 5.939173467651942e-06, |
| "loss": 0.483, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.49969027462316745, |
| "grad_norm": 2.1437965393063374, |
| "learning_rate": 5.8682408883346535e-06, |
| "loss": 0.4847, |
| "step": 605 |
| }, |
| { |
| "epoch": 0.503819946314268, |
| "grad_norm": 2.019446515641709, |
| "learning_rate": 5.7971276678633625e-06, |
| "loss": 0.4935, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.5079496180053685, |
| "grad_norm": 2.1084072043678894, |
| "learning_rate": 5.725848601641492e-06, |
| "loss": 0.4736, |
| "step": 615 |
| }, |
| { |
| "epoch": 0.5120792896964691, |
| "grad_norm": 2.185992748799692, |
| "learning_rate": 5.654418519577369e-06, |
| "loss": 0.4743, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.5162089613875697, |
| "grad_norm": 2.0909394304010416, |
| "learning_rate": 5.5828522829987965e-06, |
| "loss": 0.4808, |
| "step": 625 |
| }, |
| { |
| "epoch": 0.5203386330786702, |
| "grad_norm": 2.058129072039577, |
| "learning_rate": 5.511164781561096e-06, |
| "loss": 0.4727, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.5244683047697708, |
| "grad_norm": 2.1781635883341837, |
| "learning_rate": 5.439370930149252e-06, |
| "loss": 0.4665, |
| "step": 635 |
| }, |
| { |
| "epoch": 0.5285979764608714, |
| "grad_norm": 2.132119418871688, |
| "learning_rate": 5.367485665774802e-06, |
| "loss": 0.4647, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.5327276481519719, |
| "grad_norm": 2.0923424452918953, |
| "learning_rate": 5.295523944468137e-06, |
| "loss": 0.4752, |
| "step": 645 |
| }, |
| { |
| "epoch": 0.5368573198430725, |
| "grad_norm": 2.1500817775070704, |
| "learning_rate": 5.223500738166837e-06, |
| "loss": 0.4653, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.5409869915341731, |
| "grad_norm": 2.3057702591417764, |
| "learning_rate": 5.1514310316006835e-06, |
| "loss": 0.459, |
| "step": 655 |
| }, |
| { |
| "epoch": 0.5451166632252736, |
| "grad_norm": 2.027442727638895, |
| "learning_rate": 5.07932981917404e-06, |
| "loss": 0.467, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.5492463349163742, |
| "grad_norm": 2.003781400613107, |
| "learning_rate": 5.007212101846194e-06, |
| "loss": 0.4534, |
| "step": 665 |
| }, |
| { |
| "epoch": 0.5533760066074747, |
| "grad_norm": 2.1173198648186236, |
| "learning_rate": 4.935092884010347e-06, |
| "loss": 0.4582, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.5575056782985752, |
| "grad_norm": 2.1197258530702303, |
| "learning_rate": 4.8629871703718844e-06, |
| "loss": 0.4484, |
| "step": 675 |
| }, |
| { |
| "epoch": 0.5616353499896758, |
| "grad_norm": 2.112938054969618, |
| "learning_rate": 4.7909099628265946e-06, |
| "loss": 0.4504, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.5657650216807764, |
| "grad_norm": 2.1008936035245056, |
| "learning_rate": 4.718876257339444e-06, |
| "loss": 0.4485, |
| "step": 685 |
| }, |
| { |
| "epoch": 0.5698946933718769, |
| "grad_norm": 2.925269693235509, |
| "learning_rate": 4.646901040824622e-06, |
| "loss": 0.4523, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.5740243650629775, |
| "grad_norm": 2.013226878250545, |
| "learning_rate": 4.574999288027437e-06, |
| "loss": 0.4544, |
| "step": 695 |
| }, |
| { |
| "epoch": 0.5781540367540781, |
| "grad_norm": 2.059493099572322, |
| "learning_rate": 4.503185958408767e-06, |
| "loss": 0.4467, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.5822837084451786, |
| "grad_norm": 2.0049953659179716, |
| "learning_rate": 4.431475993032673e-06, |
| "loss": 0.441, |
| "step": 705 |
| }, |
| { |
| "epoch": 0.5864133801362792, |
| "grad_norm": 2.1512212806706255, |
| "learning_rate": 4.359884311457857e-06, |
| "loss": 0.4337, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.5905430518273798, |
| "grad_norm": 2.1177615430048755, |
| "learning_rate": 4.2884258086335755e-06, |
| "loss": 0.4458, |
| "step": 715 |
| }, |
| { |
| "epoch": 0.5946727235184803, |
| "grad_norm": 2.008192166381139, |
| "learning_rate": 4.217115351800693e-06, |
| "loss": 0.4413, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.5988023952095808, |
| "grad_norm": 2.063297374873059, |
| "learning_rate": 4.145967777398481e-06, |
| "loss": 0.4344, |
| "step": 725 |
| }, |
| { |
| "epoch": 0.6029320669006814, |
| "grad_norm": 2.004714733933455, |
| "learning_rate": 4.074997887977843e-06, |
| "loss": 0.4358, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.6070617385917819, |
| "grad_norm": 2.025598614826924, |
| "learning_rate": 4.004220449121574e-06, |
| "loss": 0.4482, |
| "step": 735 |
| }, |
| { |
| "epoch": 0.6111914102828825, |
| "grad_norm": 1.938185383839704, |
| "learning_rate": 3.933650186372329e-06, |
| "loss": 0.4337, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.615321081973983, |
| "grad_norm": 2.0495762719124104, |
| "learning_rate": 3.863301782168896e-06, |
| "loss": 0.4412, |
| "step": 745 |
| }, |
| { |
| "epoch": 0.6194507536650836, |
| "grad_norm": 2.028634933360134, |
| "learning_rate": 3.7931898727914723e-06, |
| "loss": 0.4293, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.6235804253561842, |
| "grad_norm": 2.0029138326822404, |
| "learning_rate": 3.7233290453165127e-06, |
| "loss": 0.4304, |
| "step": 755 |
| }, |
| { |
| "epoch": 0.6277100970472848, |
| "grad_norm": 2.048375432552439, |
| "learning_rate": 3.6537338345818273e-06, |
| "loss": 0.4213, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.6318397687383853, |
| "grad_norm": 2.0326809490611852, |
| "learning_rate": 3.5844187201625567e-06, |
| "loss": 0.424, |
| "step": 765 |
| }, |
| { |
| "epoch": 0.6359694404294859, |
| "grad_norm": 2.019159008104501, |
| "learning_rate": 3.5153981233586277e-06, |
| "loss": 0.4228, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.6400991121205865, |
| "grad_norm": 2.09985714316485, |
| "learning_rate": 3.446686404194337e-06, |
| "loss": 0.4206, |
| "step": 775 |
| }, |
| { |
| "epoch": 0.644228783811687, |
| "grad_norm": 1.9414886074513862, |
| "learning_rate": 3.3782978584307035e-06, |
| "loss": 0.4173, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.6483584555027875, |
| "grad_norm": 2.3066100655713755, |
| "learning_rate": 3.310246714591162e-06, |
| "loss": 0.4248, |
| "step": 785 |
| }, |
| { |
| "epoch": 0.652488127193888, |
| "grad_norm": 2.106953928933072, |
| "learning_rate": 3.2425471310012645e-06, |
| "loss": 0.4159, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.6566177988849886, |
| "grad_norm": 2.197357084281967, |
| "learning_rate": 3.1752131928429787e-06, |
| "loss": 0.425, |
| "step": 795 |
| }, |
| { |
| "epoch": 0.6607474705760892, |
| "grad_norm": 1.9729477591043592, |
| "learning_rate": 3.1082589092242116e-06, |
| "loss": 0.4241, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.6648771422671897, |
| "grad_norm": 2.0442009757093875, |
| "learning_rate": 3.041698210264149e-06, |
| "loss": 0.4238, |
| "step": 805 |
| }, |
| { |
| "epoch": 0.6690068139582903, |
| "grad_norm": 2.031799210389413, |
| "learning_rate": 2.9755449441950434e-06, |
| "loss": 0.4131, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.6731364856493909, |
| "grad_norm": 2.070029206474213, |
| "learning_rate": 2.9098128744810245e-06, |
| "loss": 0.4159, |
| "step": 815 |
| }, |
| { |
| "epoch": 0.6772661573404914, |
| "grad_norm": 2.0394445933724414, |
| "learning_rate": 2.84451567695456e-06, |
| "loss": 0.42, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.681395829031592, |
| "grad_norm": 2.0377377910207657, |
| "learning_rate": 2.7796669369711294e-06, |
| "loss": 0.4095, |
| "step": 825 |
| }, |
| { |
| "epoch": 0.6855255007226926, |
| "grad_norm": 2.0626337564803965, |
| "learning_rate": 2.715280146582752e-06, |
| "loss": 0.4069, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.6896551724137931, |
| "grad_norm": 2.0040834447649454, |
| "learning_rate": 2.651368701730889e-06, |
| "loss": 0.41, |
| "step": 835 |
| }, |
| { |
| "epoch": 0.6937848441048936, |
| "grad_norm": 1.917041220777535, |
| "learning_rate": 2.5879458994593786e-06, |
| "loss": 0.4074, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.6979145157959942, |
| "grad_norm": 2.0269668889490515, |
| "learning_rate": 2.5250249351479206e-06, |
| "loss": 0.4081, |
| "step": 845 |
| }, |
| { |
| "epoch": 0.7020441874870947, |
| "grad_norm": 1.9753046952791828, |
| "learning_rate": 2.4626188997667224e-06, |
| "loss": 0.4018, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.7061738591781953, |
| "grad_norm": 2.014280374102551, |
| "learning_rate": 2.400740777152874e-06, |
| "loss": 0.4063, |
| "step": 855 |
| }, |
| { |
| "epoch": 0.7103035308692959, |
| "grad_norm": 2.026632120733398, |
| "learning_rate": 2.3394034413090015e-06, |
| "loss": 0.3955, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.7144332025603964, |
| "grad_norm": 2.088872002071019, |
| "learning_rate": 2.278619653724781e-06, |
| "loss": 0.4038, |
| "step": 865 |
| }, |
| { |
| "epoch": 0.718562874251497, |
| "grad_norm": 2.096103572784026, |
| "learning_rate": 2.218402060721845e-06, |
| "loss": 0.3921, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.7226925459425976, |
| "grad_norm": 1.9266065409655262, |
| "learning_rate": 2.1587631908226812e-06, |
| "loss": 0.4085, |
| "step": 875 |
| }, |
| { |
| "epoch": 0.7268222176336981, |
| "grad_norm": 1.9536655952796587, |
| "learning_rate": 2.09971545214401e-06, |
| "loss": 0.3965, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.7309518893247987, |
| "grad_norm": 2.061222200669705, |
| "learning_rate": 2.04127112981522e-06, |
| "loss": 0.3941, |
| "step": 885 |
| }, |
| { |
| "epoch": 0.7350815610158993, |
| "grad_norm": 2.010203957849096, |
| "learning_rate": 1.9834423834224014e-06, |
| "loss": 0.3982, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.7392112327069998, |
| "grad_norm": 1.9380026785117144, |
| "learning_rate": 1.926241244478496e-06, |
| "loss": 0.3965, |
| "step": 895 |
| }, |
| { |
| "epoch": 0.7433409043981003, |
| "grad_norm": 2.0416911208044244, |
| "learning_rate": 1.8696796139200852e-06, |
| "loss": 0.4015, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.7474705760892009, |
| "grad_norm": 2.040772561659114, |
| "learning_rate": 1.8137692596313528e-06, |
| "loss": 0.3827, |
| "step": 905 |
| }, |
| { |
| "epoch": 0.7516002477803014, |
| "grad_norm": 1.9118935892662992, |
| "learning_rate": 1.7585218139957205e-06, |
| "loss": 0.3895, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.755729919471402, |
| "grad_norm": 1.927575457582737, |
| "learning_rate": 1.7039487714756953e-06, |
| "loss": 0.3965, |
| "step": 915 |
| }, |
| { |
| "epoch": 0.7598595911625026, |
| "grad_norm": 1.9797748596618283, |
| "learning_rate": 1.6500614862213866e-06, |
| "loss": 0.3937, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.7639892628536031, |
| "grad_norm": 2.05432833461262, |
| "learning_rate": 1.596871169708235e-06, |
| "loss": 0.3891, |
| "step": 925 |
| }, |
| { |
| "epoch": 0.7681189345447037, |
| "grad_norm": 1.9501310111794177, |
| "learning_rate": 1.54438888840441e-06, |
| "loss": 0.3801, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.7722486062358043, |
| "grad_norm": 1.970421056222239, |
| "learning_rate": 1.4926255614683931e-06, |
| "loss": 0.3806, |
| "step": 935 |
| }, |
| { |
| "epoch": 0.7763782779269048, |
| "grad_norm": 2.1277167725190598, |
| "learning_rate": 1.4415919584771999e-06, |
| "loss": 0.3843, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.7805079496180054, |
| "grad_norm": 1.9784981959186678, |
| "learning_rate": 1.3912986971857168e-06, |
| "loss": 0.3828, |
| "step": 945 |
| }, |
| { |
| "epoch": 0.784637621309106, |
| "grad_norm": 2.0003697611347975, |
| "learning_rate": 1.3417562413176405e-06, |
| "loss": 0.3859, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.7887672930002065, |
| "grad_norm": 2.0785112893490485, |
| "learning_rate": 1.292974898388456e-06, |
| "loss": 0.3812, |
| "step": 955 |
| }, |
| { |
| "epoch": 0.792896964691307, |
| "grad_norm": 2.045681460940008, |
| "learning_rate": 1.2449648175609115e-06, |
| "loss": 0.3794, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.7970266363824076, |
| "grad_norm": 2.006461021136274, |
| "learning_rate": 1.1977359875334454e-06, |
| "loss": 0.3867, |
| "step": 965 |
| }, |
| { |
| "epoch": 0.8011563080735081, |
| "grad_norm": 2.0404335899757187, |
| "learning_rate": 1.1512982344619904e-06, |
| "loss": 0.3845, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.8052859797646087, |
| "grad_norm": 2.1247685897806186, |
| "learning_rate": 1.1056612199156093e-06, |
| "loss": 0.3838, |
| "step": 975 |
| }, |
| { |
| "epoch": 0.8094156514557093, |
| "grad_norm": 2.059455987291636, |
| "learning_rate": 1.0608344388663576e-06, |
| "loss": 0.3769, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.8135453231468098, |
| "grad_norm": 2.010190102670995, |
| "learning_rate": 1.01682721771382e-06, |
| "loss": 0.38, |
| "step": 985 |
| }, |
| { |
| "epoch": 0.8176749948379104, |
| "grad_norm": 2.1310463662484755, |
| "learning_rate": 9.73648712344707e-07, |
| "loss": 0.3727, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.821804666529011, |
| "grad_norm": 2.026722500514339, |
| "learning_rate": 9.313079062279429e-07, |
| "loss": 0.3775, |
| "step": 995 |
| }, |
| { |
| "epoch": 0.8259343382201115, |
| "grad_norm": 1.98494077041973, |
| "learning_rate": 8.898136085456127e-07, |
| "loss": 0.3795, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.8300640099112121, |
| "grad_norm": 2.0710248547928267, |
| "learning_rate": 8.491744523601741e-07, |
| "loss": 0.3789, |
| "step": 1005 |
| }, |
| { |
| "epoch": 0.8341936816023127, |
| "grad_norm": 2.052056544033471, |
| "learning_rate": 8.093988928183127e-07, |
| "loss": 0.376, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.8383233532934131, |
| "grad_norm": 2.0064833870293612, |
| "learning_rate": 7.70495205391818e-07, |
| "loss": 0.3728, |
| "step": 1015 |
| }, |
| { |
| "epoch": 0.8424530249845137, |
| "grad_norm": 2.032309328318966, |
| "learning_rate": 7.324714841558322e-07, |
| "loss": 0.3701, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.8465826966756143, |
| "grad_norm": 1.976437679829882, |
| "learning_rate": 6.953356401048466e-07, |
| "loss": 0.3796, |
| "step": 1025 |
| }, |
| { |
| "epoch": 0.8507123683667148, |
| "grad_norm": 2.0095046825383927, |
| "learning_rate": 6.590953995067812e-07, |
| "loss": 0.3718, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.8548420400578154, |
| "grad_norm": 2.1362491401601598, |
| "learning_rate": 6.237583022955079e-07, |
| "loss": 0.3669, |
| "step": 1035 |
| }, |
| { |
| "epoch": 0.858971711748916, |
| "grad_norm": 2.030103317522303, |
| "learning_rate": 5.89331700502136e-07, |
| "loss": 0.3674, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.8631013834400165, |
| "grad_norm": 2.0320527842019316, |
| "learning_rate": 5.558227567253832e-07, |
| "loss": 0.3679, |
| "step": 1045 |
| }, |
| { |
| "epoch": 0.8672310551311171, |
| "grad_norm": 2.1203806417251214, |
| "learning_rate": 5.232384426413772e-07, |
| "loss": 0.3697, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.8713607268222177, |
| "grad_norm": 1.9029660119177207, |
| "learning_rate": 4.915855375531642e-07, |
| "loss": 0.3708, |
| "step": 1055 |
| }, |
| { |
| "epoch": 0.8754903985133182, |
| "grad_norm": 1.9598889356044642, |
| "learning_rate": 4.608706269802471e-07, |
| "loss": 0.3641, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.8796200702044188, |
| "grad_norm": 2.015130495239755, |
| "learning_rate": 4.3110010128843747e-07, |
| "loss": 0.3659, |
| "step": 1065 |
| }, |
| { |
| "epoch": 0.8837497418955194, |
| "grad_norm": 1.9290620704097432, |
| "learning_rate": 4.022801543603194e-07, |
| "loss": 0.3648, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.8878794135866198, |
| "grad_norm": 2.131063629467497, |
| "learning_rate": 3.744167823065814e-07, |
| "loss": 0.3619, |
| "step": 1075 |
| }, |
| { |
| "epoch": 0.8920090852777204, |
| "grad_norm": 2.007895538821572, |
| "learning_rate": 3.4751578221850377e-07, |
| "loss": 0.3729, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.896138756968821, |
| "grad_norm": 2.0860597695484113, |
| "learning_rate": 3.2158275096184744e-07, |
| "loss": 0.3674, |
| "step": 1085 |
| }, |
| { |
| "epoch": 0.9002684286599215, |
| "grad_norm": 2.0200760244926044, |
| "learning_rate": 2.966230840124007e-07, |
| "loss": 0.3677, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.9043981003510221, |
| "grad_norm": 2.0288061050580053, |
| "learning_rate": 2.72641974333433e-07, |
| "loss": 0.3698, |
| "step": 1095 |
| }, |
| { |
| "epoch": 0.9085277720421226, |
| "grad_norm": 2.0237726744767, |
| "learning_rate": 2.4964441129527337e-07, |
| "loss": 0.3622, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.9126574437332232, |
| "grad_norm": 1.9537324233727318, |
| "learning_rate": 2.2763517963725169e-07, |
| "loss": 0.3678, |
| "step": 1105 |
| }, |
| { |
| "epoch": 0.9167871154243238, |
| "grad_norm": 2.0620765585117837, |
| "learning_rate": 2.0661885847221398e-07, |
| "loss": 0.3698, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.9209167871154244, |
| "grad_norm": 1.9666263090601819, |
| "learning_rate": 1.8659982033381928e-07, |
| "loss": 0.3615, |
| "step": 1115 |
| }, |
| { |
| "epoch": 0.9250464588065249, |
| "grad_norm": 2.0277290304841142, |
| "learning_rate": 1.6758223026681507e-07, |
| "loss": 0.3613, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.9291761304976255, |
| "grad_norm": 2.106138721077295, |
| "learning_rate": 1.4957004496048256e-07, |
| "loss": 0.3621, |
| "step": 1125 |
| }, |
| { |
| "epoch": 0.9333058021887259, |
| "grad_norm": 1.9942405326271708, |
| "learning_rate": 1.3256701192542853e-07, |
| "loss": 0.3659, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.9374354738798265, |
| "grad_norm": 1.9945899501208246, |
| "learning_rate": 1.1657666871390471e-07, |
| "loss": 0.3671, |
| "step": 1135 |
| }, |
| { |
| "epoch": 0.9415651455709271, |
| "grad_norm": 1.8785796566381439, |
| "learning_rate": 1.0160234218380305e-07, |
| "loss": 0.3593, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.9456948172620276, |
| "grad_norm": 2.025291462525694, |
| "learning_rate": 8.76471478064872e-08, |
| "loss": 0.3723, |
| "step": 1145 |
| }, |
| { |
| "epoch": 0.9498244889531282, |
| "grad_norm": 1.9718773659356506, |
| "learning_rate": 7.471398901860772e-08, |
| "loss": 0.3665, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.9539541606442288, |
| "grad_norm": 1.9769770127109265, |
| "learning_rate": 6.280555661802857e-08, |
| "loss": 0.3646, |
| "step": 1155 |
| }, |
| { |
| "epoch": 0.9580838323353293, |
| "grad_norm": 2.0267147785284716, |
| "learning_rate": 5.192432820399718e-08, |
| "loss": 0.3657, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.9622135040264299, |
| "grad_norm": 2.07581128095638, |
| "learning_rate": 4.207256766166845e-08, |
| "loss": 0.3565, |
| "step": 1165 |
| }, |
| { |
| "epoch": 0.9663431757175305, |
| "grad_norm": 2.0255359776120065, |
| "learning_rate": 3.3252324691093185e-08, |
| "loss": 0.3616, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.970472847408631, |
| "grad_norm": 2.0868262303852396, |
| "learning_rate": 2.546543438077087e-08, |
| "loss": 0.3683, |
| "step": 1175 |
| }, |
| { |
| "epoch": 0.9746025190997316, |
| "grad_norm": 2.0003015972698504, |
| "learning_rate": 1.8713516825851207e-08, |
| "loss": 0.371, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.9787321907908322, |
| "grad_norm": 1.9827933568645475, |
| "learning_rate": 1.2997976791065403e-08, |
| "loss": 0.3549, |
| "step": 1185 |
| }, |
| { |
| "epoch": 0.9828618624819326, |
| "grad_norm": 2.020348792034948, |
| "learning_rate": 8.32000341846162e-09, |
| "loss": 0.3645, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.9869915341730332, |
| "grad_norm": 1.9998961679855551, |
| "learning_rate": 4.6805699799967744e-09, |
| "loss": 0.3624, |
| "step": 1195 |
| }, |
| { |
| "epoch": 0.9911212058641338, |
| "grad_norm": 1.952421429252492, |
| "learning_rate": 2.0804336750429588e-09, |
| "loss": 0.3733, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.9952508775552343, |
| "grad_norm": 2.101729255399416, |
| "learning_rate": 5.201354728517905e-10, |
| "loss": 0.3664, |
| "step": 1205 |
| }, |
| { |
| "epoch": 0.9993805492463349, |
| "grad_norm": 2.1247280097277037, |
| "learning_rate": 0.0, |
| "loss": 0.3624, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.9993805492463349, |
| "eval_loss": 0.3181818723678589, |
| "eval_runtime": 1.1842, |
| "eval_samples_per_second": 2.533, |
| "eval_steps_per_second": 0.844, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.9993805492463349, |
| "step": 1210, |
| "total_flos": 253297038458880.0, |
| "train_loss": 0.5432274522860188, |
| "train_runtime": 27235.9072, |
| "train_samples_per_second": 1.422, |
| "train_steps_per_second": 0.044 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 1210, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 253297038458880.0, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|