| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 14.723716381418093, | |
| "eval_steps": 500, | |
| "global_step": 765, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.097799511002445, | |
| "grad_norm": 2.395451784133911, | |
| "learning_rate": 1.9997891995035914e-05, | |
| "loss": 0.8739, | |
| "num_input_tokens_seen": 163840, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.19559902200489, | |
| "grad_norm": 1.3395614624023438, | |
| "learning_rate": 1.999156886888064e-05, | |
| "loss": 0.7957, | |
| "num_input_tokens_seen": 327680, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.293398533007335, | |
| "grad_norm": 1.125549554824829, | |
| "learning_rate": 1.9981033287370443e-05, | |
| "loss": 0.7822, | |
| "num_input_tokens_seen": 491520, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.39119804400978, | |
| "grad_norm": 1.1936776638031006, | |
| "learning_rate": 1.9966289692316944e-05, | |
| "loss": 0.7815, | |
| "num_input_tokens_seen": 655360, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.4889975550122249, | |
| "grad_norm": 1.230684757232666, | |
| "learning_rate": 1.9947344299634464e-05, | |
| "loss": 0.7432, | |
| "num_input_tokens_seen": 819200, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.58679706601467, | |
| "grad_norm": 1.0630358457565308, | |
| "learning_rate": 1.992420509671936e-05, | |
| "loss": 0.7389, | |
| "num_input_tokens_seen": 982368, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.684596577017115, | |
| "grad_norm": 1.1900779008865356, | |
| "learning_rate": 1.9896881839082554e-05, | |
| "loss": 0.7654, | |
| "num_input_tokens_seen": 1146208, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.78239608801956, | |
| "grad_norm": 1.0942089557647705, | |
| "learning_rate": 1.9865386046236597e-05, | |
| "loss": 0.7753, | |
| "num_input_tokens_seen": 1310048, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.8801955990220048, | |
| "grad_norm": 1.0910882949829102, | |
| "learning_rate": 1.982973099683902e-05, | |
| "loss": 0.7256, | |
| "num_input_tokens_seen": 1473888, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.9779951100244498, | |
| "grad_norm": 1.0902478694915771, | |
| "learning_rate": 1.9789931723094046e-05, | |
| "loss": 0.727, | |
| "num_input_tokens_seen": 1637728, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.058679706601467, | |
| "grad_norm": 1.3777414560317993, | |
| "learning_rate": 1.9746005004415004e-05, | |
| "loss": 0.582, | |
| "num_input_tokens_seen": 1770848, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.156479217603912, | |
| "grad_norm": 1.3062232732772827, | |
| "learning_rate": 1.9697969360350098e-05, | |
| "loss": 0.5628, | |
| "num_input_tokens_seen": 1934688, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.254278728606357, | |
| "grad_norm": 1.239915132522583, | |
| "learning_rate": 1.9645845042774555e-05, | |
| "loss": 0.5561, | |
| "num_input_tokens_seen": 2098528, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.352078239608802, | |
| "grad_norm": 1.1612083911895752, | |
| "learning_rate": 1.9589654027352412e-05, | |
| "loss": 0.5778, | |
| "num_input_tokens_seen": 2262368, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.449877750611247, | |
| "grad_norm": 1.2168060541152954, | |
| "learning_rate": 1.9529420004271568e-05, | |
| "loss": 0.5207, | |
| "num_input_tokens_seen": 2426128, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.5476772616136918, | |
| "grad_norm": 1.2806190252304077, | |
| "learning_rate": 1.9465168368255946e-05, | |
| "loss": 0.5452, | |
| "num_input_tokens_seen": 2589968, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.6454767726161368, | |
| "grad_norm": 1.17153000831604, | |
| "learning_rate": 1.9396926207859085e-05, | |
| "loss": 0.5379, | |
| "num_input_tokens_seen": 2753808, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 1.7432762836185818, | |
| "grad_norm": 1.1548662185668945, | |
| "learning_rate": 1.932472229404356e-05, | |
| "loss": 0.5116, | |
| "num_input_tokens_seen": 2917648, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.8410757946210268, | |
| "grad_norm": 1.3172167539596558, | |
| "learning_rate": 1.924858706805112e-05, | |
| "loss": 0.5532, | |
| "num_input_tokens_seen": 3081488, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.9388753056234718, | |
| "grad_norm": 1.1930724382400513, | |
| "learning_rate": 1.9168552628568632e-05, | |
| "loss": 0.5659, | |
| "num_input_tokens_seen": 3245328, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.019559902200489, | |
| "grad_norm": 1.7079554796218872, | |
| "learning_rate": 1.9084652718195237e-05, | |
| "loss": 0.6028, | |
| "num_input_tokens_seen": 3378448, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 2.117359413202934, | |
| "grad_norm": 1.7393510341644287, | |
| "learning_rate": 1.8996922709216456e-05, | |
| "loss": 0.4228, | |
| "num_input_tokens_seen": 3542288, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.215158924205379, | |
| "grad_norm": 1.2822431325912476, | |
| "learning_rate": 1.8905399588691165e-05, | |
| "loss": 0.3648, | |
| "num_input_tokens_seen": 3706128, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 2.312958435207824, | |
| "grad_norm": 1.3198189735412598, | |
| "learning_rate": 1.8810121942857848e-05, | |
| "loss": 0.3822, | |
| "num_input_tokens_seen": 3869968, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.410757946210269, | |
| "grad_norm": 1.2831072807312012, | |
| "learning_rate": 1.8711129940866577e-05, | |
| "loss": 0.3907, | |
| "num_input_tokens_seen": 4033808, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 2.508557457212714, | |
| "grad_norm": 1.4272454977035522, | |
| "learning_rate": 1.860846531784368e-05, | |
| "loss": 0.3936, | |
| "num_input_tokens_seen": 4197648, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.606356968215159, | |
| "grad_norm": 1.3846728801727295, | |
| "learning_rate": 1.8502171357296144e-05, | |
| "loss": 0.376, | |
| "num_input_tokens_seen": 4361488, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 2.704156479217604, | |
| "grad_norm": 1.1824907064437866, | |
| "learning_rate": 1.839229287286327e-05, | |
| "loss": 0.3735, | |
| "num_input_tokens_seen": 4525328, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.801955990220049, | |
| "grad_norm": 1.2617613077163696, | |
| "learning_rate": 1.827887618942318e-05, | |
| "loss": 0.4114, | |
| "num_input_tokens_seen": 4689168, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 2.899755501222494, | |
| "grad_norm": 1.1593624353408813, | |
| "learning_rate": 1.816196912356222e-05, | |
| "loss": 0.3696, | |
| "num_input_tokens_seen": 4853008, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.997555012224939, | |
| "grad_norm": 1.1450086832046509, | |
| "learning_rate": 1.8041620963415418e-05, | |
| "loss": 0.3888, | |
| "num_input_tokens_seen": 5016848, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 3.078239608801956, | |
| "grad_norm": 1.4760794639587402, | |
| "learning_rate": 1.7917882447886585e-05, | |
| "loss": 0.2605, | |
| "num_input_tokens_seen": 5149968, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.176039119804401, | |
| "grad_norm": 1.520880103111267, | |
| "learning_rate": 1.7790805745256703e-05, | |
| "loss": 0.2475, | |
| "num_input_tokens_seen": 5313808, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 3.273838630806846, | |
| "grad_norm": 1.4768394231796265, | |
| "learning_rate": 1.766044443118978e-05, | |
| "loss": 0.2344, | |
| "num_input_tokens_seen": 5477648, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 3.371638141809291, | |
| "grad_norm": 1.4958367347717285, | |
| "learning_rate": 1.7526853466145248e-05, | |
| "loss": 0.2665, | |
| "num_input_tokens_seen": 5641488, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 3.469437652811736, | |
| "grad_norm": 1.4445388317108154, | |
| "learning_rate": 1.7390089172206594e-05, | |
| "loss": 0.2477, | |
| "num_input_tokens_seen": 5805328, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 3.567237163814181, | |
| "grad_norm": 1.4173372983932495, | |
| "learning_rate": 1.725020920933593e-05, | |
| "loss": 0.2679, | |
| "num_input_tokens_seen": 5969168, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 3.665036674816626, | |
| "grad_norm": 1.6111114025115967, | |
| "learning_rate": 1.710727255106447e-05, | |
| "loss": 0.234, | |
| "num_input_tokens_seen": 6133008, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 3.762836185819071, | |
| "grad_norm": 1.4121896028518677, | |
| "learning_rate": 1.696133945962927e-05, | |
| "loss": 0.2587, | |
| "num_input_tokens_seen": 6296848, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 3.860635696821516, | |
| "grad_norm": 1.3143867254257202, | |
| "learning_rate": 1.681247146056654e-05, | |
| "loss": 0.2606, | |
| "num_input_tokens_seen": 6460688, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 3.958435207823961, | |
| "grad_norm": 1.2710639238357544, | |
| "learning_rate": 1.6660731316772503e-05, | |
| "loss": 0.2543, | |
| "num_input_tokens_seen": 6624528, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 4.039119804400978, | |
| "grad_norm": 1.284611463546753, | |
| "learning_rate": 1.650618300204242e-05, | |
| "loss": 0.1856, | |
| "num_input_tokens_seen": 6757128, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 4.136919315403423, | |
| "grad_norm": 1.6113184690475464, | |
| "learning_rate": 1.634889167409923e-05, | |
| "loss": 0.1462, | |
| "num_input_tokens_seen": 6920968, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 4.234718826405868, | |
| "grad_norm": 1.230480432510376, | |
| "learning_rate": 1.6188923647122946e-05, | |
| "loss": 0.17, | |
| "num_input_tokens_seen": 7084808, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 4.332518337408313, | |
| "grad_norm": 1.388247013092041, | |
| "learning_rate": 1.6026346363792565e-05, | |
| "loss": 0.1427, | |
| "num_input_tokens_seen": 7248648, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 4.430317848410758, | |
| "grad_norm": 1.3794785737991333, | |
| "learning_rate": 1.5861228366852148e-05, | |
| "loss": 0.15, | |
| "num_input_tokens_seen": 7412488, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 4.528117359413203, | |
| "grad_norm": 1.5547810792922974, | |
| "learning_rate": 1.5693639270213138e-05, | |
| "loss": 0.1603, | |
| "num_input_tokens_seen": 7576264, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 4.625916870415648, | |
| "grad_norm": 1.49036705493927, | |
| "learning_rate": 1.552364972960506e-05, | |
| "loss": 0.1536, | |
| "num_input_tokens_seen": 7740104, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 4.723716381418093, | |
| "grad_norm": 1.2922449111938477, | |
| "learning_rate": 1.5351331412787004e-05, | |
| "loss": 0.1415, | |
| "num_input_tokens_seen": 7903944, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 4.821515892420538, | |
| "grad_norm": 1.5458725690841675, | |
| "learning_rate": 1.5176756969332428e-05, | |
| "loss": 0.159, | |
| "num_input_tokens_seen": 8067784, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 4.919315403422983, | |
| "grad_norm": 1.3060288429260254, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.1699, | |
| "num_input_tokens_seen": 8231624, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 4.3814568519592285, | |
| "learning_rate": 1.4821135025703491e-05, | |
| "loss": 0.1295, | |
| "num_input_tokens_seen": 8364744, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 5.097799511002445, | |
| "grad_norm": 1.496863842010498, | |
| "learning_rate": 1.4640237456093636e-05, | |
| "loss": 0.1019, | |
| "num_input_tokens_seen": 8528584, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 5.19559902200489, | |
| "grad_norm": 1.3016966581344604, | |
| "learning_rate": 1.4457383557765385e-05, | |
| "loss": 0.0916, | |
| "num_input_tokens_seen": 8692424, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 5.293398533007335, | |
| "grad_norm": 1.3272321224212646, | |
| "learning_rate": 1.427265042210381e-05, | |
| "loss": 0.0935, | |
| "num_input_tokens_seen": 8856264, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 5.39119804400978, | |
| "grad_norm": 1.2260195016860962, | |
| "learning_rate": 1.4086115932782316e-05, | |
| "loss": 0.0679, | |
| "num_input_tokens_seen": 9020104, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 5.488997555012225, | |
| "grad_norm": 1.1941609382629395, | |
| "learning_rate": 1.3897858732926794e-05, | |
| "loss": 0.101, | |
| "num_input_tokens_seen": 9183944, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 5.58679706601467, | |
| "grad_norm": 1.3348464965820312, | |
| "learning_rate": 1.3707958191959609e-05, | |
| "loss": 0.0802, | |
| "num_input_tokens_seen": 9347784, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 5.684596577017115, | |
| "grad_norm": 1.3610597848892212, | |
| "learning_rate": 1.3516494372137368e-05, | |
| "loss": 0.0812, | |
| "num_input_tokens_seen": 9511624, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 5.78239608801956, | |
| "grad_norm": 1.1276705265045166, | |
| "learning_rate": 1.3323547994796597e-05, | |
| "loss": 0.0812, | |
| "num_input_tokens_seen": 9675464, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 5.880195599022005, | |
| "grad_norm": 1.4052499532699585, | |
| "learning_rate": 1.3129200406321545e-05, | |
| "loss": 0.0884, | |
| "num_input_tokens_seen": 9839304, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 5.97799511002445, | |
| "grad_norm": 1.2434982061386108, | |
| "learning_rate": 1.2933533543848462e-05, | |
| "loss": 0.0906, | |
| "num_input_tokens_seen": 10003144, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 6.058679706601467, | |
| "grad_norm": 1.19767427444458, | |
| "learning_rate": 1.2736629900720832e-05, | |
| "loss": 0.0616, | |
| "num_input_tokens_seen": 10136264, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 6.156479217603912, | |
| "grad_norm": 1.2783230543136597, | |
| "learning_rate": 1.2538572491710079e-05, | |
| "loss": 0.0447, | |
| "num_input_tokens_seen": 10300104, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 6.254278728606357, | |
| "grad_norm": 1.1590359210968018, | |
| "learning_rate": 1.2339444818016488e-05, | |
| "loss": 0.0528, | |
| "num_input_tokens_seen": 10463944, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 6.352078239608802, | |
| "grad_norm": 1.2891935110092163, | |
| "learning_rate": 1.2139330832064975e-05, | |
| "loss": 0.0429, | |
| "num_input_tokens_seen": 10627784, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 6.449877750611247, | |
| "grad_norm": 1.0882583856582642, | |
| "learning_rate": 1.1938314902110701e-05, | |
| "loss": 0.0442, | |
| "num_input_tokens_seen": 10791624, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 6.547677261613692, | |
| "grad_norm": 1.1069742441177368, | |
| "learning_rate": 1.1736481776669307e-05, | |
| "loss": 0.047, | |
| "num_input_tokens_seen": 10955464, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 6.645476772616137, | |
| "grad_norm": 1.0986578464508057, | |
| "learning_rate": 1.1533916548786856e-05, | |
| "loss": 0.0437, | |
| "num_input_tokens_seen": 11119304, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 6.743276283618582, | |
| "grad_norm": 1.1824274063110352, | |
| "learning_rate": 1.133070462016454e-05, | |
| "loss": 0.0466, | |
| "num_input_tokens_seen": 11282568, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 6.841075794621027, | |
| "grad_norm": 1.197988510131836, | |
| "learning_rate": 1.1126931665153213e-05, | |
| "loss": 0.0498, | |
| "num_input_tokens_seen": 11446408, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 6.938875305623472, | |
| "grad_norm": 1.1721147298812866, | |
| "learning_rate": 1.092268359463302e-05, | |
| "loss": 0.0409, | |
| "num_input_tokens_seen": 11610248, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 7.019559902200489, | |
| "grad_norm": 0.9280807971954346, | |
| "learning_rate": 1.0718046519793276e-05, | |
| "loss": 0.0444, | |
| "num_input_tokens_seen": 11743368, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 7.117359413202934, | |
| "grad_norm": 0.8121551871299744, | |
| "learning_rate": 1.0513106715827897e-05, | |
| "loss": 0.0212, | |
| "num_input_tokens_seen": 11906632, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 7.215158924205379, | |
| "grad_norm": 0.974238932132721, | |
| "learning_rate": 1.0307950585561705e-05, | |
| "loss": 0.0275, | |
| "num_input_tokens_seen": 12070472, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 7.312958435207824, | |
| "grad_norm": 0.8143863081932068, | |
| "learning_rate": 1.01026646230229e-05, | |
| "loss": 0.0234, | |
| "num_input_tokens_seen": 12234312, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 7.410757946210269, | |
| "grad_norm": 0.9007371664047241, | |
| "learning_rate": 9.897335376977104e-06, | |
| "loss": 0.0227, | |
| "num_input_tokens_seen": 12398152, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 7.508557457212714, | |
| "grad_norm": 0.8347233533859253, | |
| "learning_rate": 9.692049414438298e-06, | |
| "loss": 0.0264, | |
| "num_input_tokens_seen": 12561992, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 7.606356968215159, | |
| "grad_norm": 1.0170607566833496, | |
| "learning_rate": 9.486893284172103e-06, | |
| "loss": 0.0251, | |
| "num_input_tokens_seen": 12725832, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 7.704156479217604, | |
| "grad_norm": 0.8994267582893372, | |
| "learning_rate": 9.281953480206725e-06, | |
| "loss": 0.0237, | |
| "num_input_tokens_seen": 12889672, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 7.801955990220049, | |
| "grad_norm": 1.0272514820098877, | |
| "learning_rate": 9.07731640536698e-06, | |
| "loss": 0.0228, | |
| "num_input_tokens_seen": 13053512, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 7.899755501222494, | |
| "grad_norm": 0.8432409167289734, | |
| "learning_rate": 8.87306833484679e-06, | |
| "loss": 0.0238, | |
| "num_input_tokens_seen": 13217352, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 7.997555012224939, | |
| "grad_norm": 0.9295158386230469, | |
| "learning_rate": 8.669295379835467e-06, | |
| "loss": 0.0222, | |
| "num_input_tokens_seen": 13381192, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 8.078239608801956, | |
| "grad_norm": 0.6758769750595093, | |
| "learning_rate": 8.466083451213145e-06, | |
| "loss": 0.0109, | |
| "num_input_tokens_seen": 13514312, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 8.1760391198044, | |
| "grad_norm": 0.5933005213737488, | |
| "learning_rate": 8.263518223330698e-06, | |
| "loss": 0.0118, | |
| "num_input_tokens_seen": 13677448, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 8.273838630806846, | |
| "grad_norm": 0.7031515836715698, | |
| "learning_rate": 8.0616850978893e-06, | |
| "loss": 0.01, | |
| "num_input_tokens_seen": 13841288, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 8.37163814180929, | |
| "grad_norm": 0.8007884621620178, | |
| "learning_rate": 7.860669167935028e-06, | |
| "loss": 0.0112, | |
| "num_input_tokens_seen": 14005128, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 8.469437652811736, | |
| "grad_norm": 0.6427900195121765, | |
| "learning_rate": 7.660555181983517e-06, | |
| "loss": 0.0133, | |
| "num_input_tokens_seen": 14168968, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 8.56723716381418, | |
| "grad_norm": 0.6638085246086121, | |
| "learning_rate": 7.461427508289922e-06, | |
| "loss": 0.0113, | |
| "num_input_tokens_seen": 14332808, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 8.665036674816626, | |
| "grad_norm": 0.6730697751045227, | |
| "learning_rate": 7.263370099279173e-06, | |
| "loss": 0.0102, | |
| "num_input_tokens_seen": 14496648, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 8.76283618581907, | |
| "grad_norm": 0.7171938419342041, | |
| "learning_rate": 7.066466456151541e-06, | |
| "loss": 0.01, | |
| "num_input_tokens_seen": 14660488, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 8.860635696821516, | |
| "grad_norm": 0.7203698754310608, | |
| "learning_rate": 6.870799593678459e-06, | |
| "loss": 0.0116, | |
| "num_input_tokens_seen": 14824328, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 8.95843520782396, | |
| "grad_norm": 0.7533681392669678, | |
| "learning_rate": 6.6764520052034054e-06, | |
| "loss": 0.0111, | |
| "num_input_tokens_seen": 14988168, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 9.039119804400977, | |
| "grad_norm": 0.22926057875156403, | |
| "learning_rate": 6.483505627862632e-06, | |
| "loss": 0.0071, | |
| "num_input_tokens_seen": 15121288, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 9.136919315403423, | |
| "grad_norm": 0.4467085301876068, | |
| "learning_rate": 6.292041808040393e-06, | |
| "loss": 0.0054, | |
| "num_input_tokens_seen": 15285128, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 9.234718826405867, | |
| "grad_norm": 0.5730965733528137, | |
| "learning_rate": 6.102141267073207e-06, | |
| "loss": 0.0059, | |
| "num_input_tokens_seen": 15448968, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 9.332518337408313, | |
| "grad_norm": 0.3801887333393097, | |
| "learning_rate": 5.913884067217686e-06, | |
| "loss": 0.0045, | |
| "num_input_tokens_seen": 15612552, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 9.430317848410757, | |
| "grad_norm": 0.4900866448879242, | |
| "learning_rate": 5.727349577896194e-06, | |
| "loss": 0.004, | |
| "num_input_tokens_seen": 15776392, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 9.528117359413203, | |
| "grad_norm": 0.3590303957462311, | |
| "learning_rate": 5.542616442234618e-06, | |
| "loss": 0.0042, | |
| "num_input_tokens_seen": 15940232, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 9.625916870415647, | |
| "grad_norm": 0.36999812722206116, | |
| "learning_rate": 5.3597625439063685e-06, | |
| "loss": 0.0049, | |
| "num_input_tokens_seen": 16104072, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 9.723716381418093, | |
| "grad_norm": 0.3589613735675812, | |
| "learning_rate": 5.178864974296511e-06, | |
| "loss": 0.0045, | |
| "num_input_tokens_seen": 16267912, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 9.821515892420537, | |
| "grad_norm": 0.30231136083602905, | |
| "learning_rate": 5.000000000000003e-06, | |
| "loss": 0.004, | |
| "num_input_tokens_seen": 16431752, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 9.919315403422983, | |
| "grad_norm": 0.34575000405311584, | |
| "learning_rate": 4.823243030667576e-06, | |
| "loss": 0.0042, | |
| "num_input_tokens_seen": 16595592, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 0.6714850068092346, | |
| "learning_rate": 4.648668587212998e-06, | |
| "loss": 0.0037, | |
| "num_input_tokens_seen": 16728712, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 10.097799511002446, | |
| "grad_norm": 0.19611337780952454, | |
| "learning_rate": 4.476350270394942e-06, | |
| "loss": 0.0021, | |
| "num_input_tokens_seen": 16892552, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 10.19559902200489, | |
| "grad_norm": 0.13641950488090515, | |
| "learning_rate": 4.306360729786867e-06, | |
| "loss": 0.002, | |
| "num_input_tokens_seen": 17056392, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 10.293398533007334, | |
| "grad_norm": 0.16702738404273987, | |
| "learning_rate": 4.138771633147856e-06, | |
| "loss": 0.0023, | |
| "num_input_tokens_seen": 17220232, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 10.39119804400978, | |
| "grad_norm": 0.2278226763010025, | |
| "learning_rate": 3.973653636207437e-06, | |
| "loss": 0.002, | |
| "num_input_tokens_seen": 17384072, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 10.488997555012224, | |
| "grad_norm": 0.1814439296722412, | |
| "learning_rate": 3.8110763528770543e-06, | |
| "loss": 0.0021, | |
| "num_input_tokens_seen": 17547912, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 10.58679706601467, | |
| "grad_norm": 0.2528248429298401, | |
| "learning_rate": 3.651108325900773e-06, | |
| "loss": 0.0025, | |
| "num_input_tokens_seen": 17711752, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 10.684596577017114, | |
| "grad_norm": 0.165152445435524, | |
| "learning_rate": 3.493816997957582e-06, | |
| "loss": 0.0021, | |
| "num_input_tokens_seen": 17875592, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 10.78239608801956, | |
| "grad_norm": 0.08041153103113174, | |
| "learning_rate": 3.339268683227499e-06, | |
| "loss": 0.0017, | |
| "num_input_tokens_seen": 18039432, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 10.880195599022004, | |
| "grad_norm": 0.24567244946956635, | |
| "learning_rate": 3.1875285394334575e-06, | |
| "loss": 0.0018, | |
| "num_input_tokens_seen": 18203272, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 10.97799511002445, | |
| "grad_norm": 0.21529506146907806, | |
| "learning_rate": 3.0386605403707347e-06, | |
| "loss": 0.0018, | |
| "num_input_tokens_seen": 18367112, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 11.058679706601467, | |
| "grad_norm": 0.05531800910830498, | |
| "learning_rate": 2.8927274489355296e-06, | |
| "loss": 0.0014, | |
| "num_input_tokens_seen": 18500232, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 11.156479217603913, | |
| "grad_norm": 0.04145563766360283, | |
| "learning_rate": 2.749790790664074e-06, | |
| "loss": 0.0013, | |
| "num_input_tokens_seen": 18662984, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 11.254278728606357, | |
| "grad_norm": 0.07673907279968262, | |
| "learning_rate": 2.6099108277934105e-06, | |
| "loss": 0.0014, | |
| "num_input_tokens_seen": 18826824, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 11.352078239608803, | |
| "grad_norm": 0.054488956928253174, | |
| "learning_rate": 2.4731465338547556e-06, | |
| "loss": 0.0013, | |
| "num_input_tokens_seen": 18990664, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 11.449877750611247, | |
| "grad_norm": 0.11207219213247299, | |
| "learning_rate": 2.339555568810221e-06, | |
| "loss": 0.0013, | |
| "num_input_tokens_seen": 19154504, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 11.547677261613693, | |
| "grad_norm": 0.062526635825634, | |
| "learning_rate": 2.209194254743295e-06, | |
| "loss": 0.0012, | |
| "num_input_tokens_seen": 19318344, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 11.645476772616137, | |
| "grad_norm": 0.05114143341779709, | |
| "learning_rate": 2.0821175521134208e-06, | |
| "loss": 0.0013, | |
| "num_input_tokens_seen": 19482184, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 11.743276283618583, | |
| "grad_norm": 0.05199455842375755, | |
| "learning_rate": 1.9583790365845823e-06, | |
| "loss": 0.0013, | |
| "num_input_tokens_seen": 19646024, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 11.841075794621027, | |
| "grad_norm": 0.057281751185655594, | |
| "learning_rate": 1.8380308764377841e-06, | |
| "loss": 0.0014, | |
| "num_input_tokens_seen": 19809864, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 11.938875305623473, | |
| "grad_norm": 0.04751597344875336, | |
| "learning_rate": 1.7211238105768213e-06, | |
| "loss": 0.0012, | |
| "num_input_tokens_seen": 19973704, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 12.01955990220049, | |
| "grad_norm": 0.04285968840122223, | |
| "learning_rate": 1.607707127136734e-06, | |
| "loss": 0.0014, | |
| "num_input_tokens_seen": 20106824, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 12.117359413202934, | |
| "grad_norm": 0.044368330389261246, | |
| "learning_rate": 1.4978286427038602e-06, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 20270664, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 12.21515892420538, | |
| "grad_norm": 0.04820827767252922, | |
| "learning_rate": 1.3915346821563235e-06, | |
| "loss": 0.0012, | |
| "num_input_tokens_seen": 20434504, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 12.312958435207824, | |
| "grad_norm": 0.034025732427835464, | |
| "learning_rate": 1.2888700591334225e-06, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 20598280, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 12.41075794621027, | |
| "grad_norm": 0.03853330388665199, | |
| "learning_rate": 1.1898780571421554e-06, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 20762120, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 12.508557457212714, | |
| "grad_norm": 0.039888255298137665, | |
| "learning_rate": 1.0946004113088381e-06, | |
| "loss": 0.0012, | |
| "num_input_tokens_seen": 20925960, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 12.60635696821516, | |
| "grad_norm": 0.04366978630423546, | |
| "learning_rate": 1.0030772907835484e-06, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 21089800, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 12.704156479217604, | |
| "grad_norm": 0.03416445106267929, | |
| "learning_rate": 9.153472818047627e-07, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 21253640, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 12.80195599022005, | |
| "grad_norm": 0.04817335307598114, | |
| "learning_rate": 8.31447371431372e-07, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 21417480, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 12.899755501222494, | |
| "grad_norm": 0.03280309960246086, | |
| "learning_rate": 7.514129319488839e-07, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 21581320, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 12.99755501222494, | |
| "grad_norm": 0.04426475614309311, | |
| "learning_rate": 6.752777059564431e-07, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 21745160, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 13.078239608801956, | |
| "grad_norm": 0.03248964622616768, | |
| "learning_rate": 6.030737921409169e-07, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 21877864, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 13.1760391198044, | |
| "grad_norm": 0.0347786545753479, | |
| "learning_rate": 5.348316317440549e-07, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 22041704, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 13.273838630806846, | |
| "grad_norm": 0.04258348420262337, | |
| "learning_rate": 4.7057999572843516e-07, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 22205544, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 13.37163814180929, | |
| "grad_norm": 0.0364152230322361, | |
| "learning_rate": 4.103459726475889e-07, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 22369384, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 13.469437652811736, | |
| "grad_norm": 0.039297617971897125, | |
| "learning_rate": 3.541549572254488e-07, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 22533224, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 13.56723716381418, | |
| "grad_norm": 0.037363260984420776, | |
| "learning_rate": 3.020306396499062e-07, | |
| "loss": 0.0009, | |
| "num_input_tokens_seen": 22697064, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 13.665036674816626, | |
| "grad_norm": 0.032989222556352615, | |
| "learning_rate": 2.539949955849985e-07, | |
| "loss": 0.0011, | |
| "num_input_tokens_seen": 22860904, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 13.76283618581907, | |
| "grad_norm": 0.03354249894618988, | |
| "learning_rate": 2.1006827690595478e-07, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 23024744, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 13.860635696821516, | |
| "grad_norm": 0.03322712704539299, | |
| "learning_rate": 1.7026900316098217e-07, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 23188584, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 13.95843520782396, | |
| "grad_norm": 0.03775469958782196, | |
| "learning_rate": 1.3461395376340502e-07, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 23352424, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 14.039119804400977, | |
| "grad_norm": 0.039018385112285614, | |
| "learning_rate": 1.0311816091744698e-07, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 23485296, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 14.136919315403423, | |
| "grad_norm": 0.03440910577774048, | |
| "learning_rate": 7.579490328064265e-08, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 23649136, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 14.234718826405867, | |
| "grad_norm": 0.03226885199546814, | |
| "learning_rate": 5.265570036553813e-08, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 23812976, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 14.332518337408313, | |
| "grad_norm": 0.03353099152445793, | |
| "learning_rate": 3.371030768305583e-08, | |
| "loss": 0.0009, | |
| "num_input_tokens_seen": 23976816, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 14.430317848410757, | |
| "grad_norm": 0.04043235257267952, | |
| "learning_rate": 1.896671262955896e-08, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 24140656, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 14.528117359413203, | |
| "grad_norm": 0.0331251285970211, | |
| "learning_rate": 8.431131119361891e-09, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 24304496, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 14.625916870415647, | |
| "grad_norm": 0.03793644160032272, | |
| "learning_rate": 2.108004964086474e-09, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 24467760, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 14.723716381418093, | |
| "grad_norm": 0.03213175758719444, | |
| "learning_rate": 0.0, | |
| "loss": 0.001, | |
| "num_input_tokens_seen": 24631600, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 14.723716381418093, | |
| "num_input_tokens_seen": 24631600, | |
| "step": 765, | |
| "total_flos": 1.936547143200768e+17, | |
| "train_loss": 0.1555636612302776, | |
| "train_runtime": 3426.6896, | |
| "train_samples_per_second": 3.576, | |
| "train_steps_per_second": 0.223 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 765, | |
| "num_input_tokens_seen": 24631600, | |
| "num_train_epochs": 15, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.936547143200768e+17, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |