| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 5859, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0008533879501621437, | |
| "grad_norm": 4.798749402792709, | |
| "learning_rate": 6.825938566552902e-07, | |
| "loss": 1.0233, | |
| "num_tokens": 495890.0, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0017067759003242873, | |
| "grad_norm": 4.0425866323185256, | |
| "learning_rate": 1.5358361774744028e-06, | |
| "loss": 1.0272, | |
| "num_tokens": 949600.0, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.002560163850486431, | |
| "grad_norm": 3.3086017576382143, | |
| "learning_rate": 2.389078498293516e-06, | |
| "loss": 1.0324, | |
| "num_tokens": 1480979.0, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.0034135518006485747, | |
| "grad_norm": 2.635692129663441, | |
| "learning_rate": 3.242320819112628e-06, | |
| "loss": 0.9332, | |
| "num_tokens": 1933925.0, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.004266939750810718, | |
| "grad_norm": 2.1861061417641277, | |
| "learning_rate": 4.095563139931741e-06, | |
| "loss": 0.8952, | |
| "num_tokens": 2403607.0, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.005120327700972862, | |
| "grad_norm": 1.7489741971256096, | |
| "learning_rate": 4.948805460750854e-06, | |
| "loss": 0.8596, | |
| "num_tokens": 2810063.0, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.005973715651135006, | |
| "grad_norm": 1.428222790090035, | |
| "learning_rate": 5.802047781569966e-06, | |
| "loss": 0.8399, | |
| "num_tokens": 3245647.0, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.006827103601297149, | |
| "grad_norm": 1.3331284271200252, | |
| "learning_rate": 6.655290102389079e-06, | |
| "loss": 0.818, | |
| "num_tokens": 3759146.0, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.007680491551459293, | |
| "grad_norm": 1.6553057262744924, | |
| "learning_rate": 7.508532423208192e-06, | |
| "loss": 0.7976, | |
| "num_tokens": 4202971.0, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.008533879501621437, | |
| "grad_norm": 1.516759211785286, | |
| "learning_rate": 8.361774744027304e-06, | |
| "loss": 0.8221, | |
| "num_tokens": 4718430.0, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.00938726745178358, | |
| "grad_norm": 1.4828872159472009, | |
| "learning_rate": 9.215017064846417e-06, | |
| "loss": 0.8361, | |
| "num_tokens": 5162362.0, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.010240655401945725, | |
| "grad_norm": 1.4796322255718408, | |
| "learning_rate": 1.006825938566553e-05, | |
| "loss": 0.8286, | |
| "num_tokens": 5644524.0, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.011094043352107869, | |
| "grad_norm": 1.3342423505815704, | |
| "learning_rate": 1.0921501706484643e-05, | |
| "loss": 0.7944, | |
| "num_tokens": 6151522.0, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.011947431302270013, | |
| "grad_norm": 1.3878247508488946, | |
| "learning_rate": 1.1774744027303754e-05, | |
| "loss": 0.8079, | |
| "num_tokens": 6674546.0, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.012800819252432157, | |
| "grad_norm": 1.4745878596040554, | |
| "learning_rate": 1.2627986348122867e-05, | |
| "loss": 0.7936, | |
| "num_tokens": 7178713.0, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.013654207202594299, | |
| "grad_norm": 1.400943190419203, | |
| "learning_rate": 1.348122866894198e-05, | |
| "loss": 0.7827, | |
| "num_tokens": 7668808.0, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.014507595152756443, | |
| "grad_norm": 1.4092175999951435, | |
| "learning_rate": 1.4334470989761092e-05, | |
| "loss": 0.7837, | |
| "num_tokens": 8131699.0, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.015360983102918587, | |
| "grad_norm": 1.2931296016661182, | |
| "learning_rate": 1.5187713310580206e-05, | |
| "loss": 0.7488, | |
| "num_tokens": 8560803.0, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.01621437105308073, | |
| "grad_norm": 1.5862120531875272, | |
| "learning_rate": 1.604095563139932e-05, | |
| "loss": 0.8045, | |
| "num_tokens": 9030225.0, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.017067759003242873, | |
| "grad_norm": 1.510461750449168, | |
| "learning_rate": 1.689419795221843e-05, | |
| "loss": 0.762, | |
| "num_tokens": 9513010.0, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.017921146953405017, | |
| "grad_norm": 1.5232874703838686, | |
| "learning_rate": 1.7747440273037545e-05, | |
| "loss": 0.7473, | |
| "num_tokens": 9959461.0, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.01877453490356716, | |
| "grad_norm": 1.5288148915863156, | |
| "learning_rate": 1.8600682593856656e-05, | |
| "loss": 0.7436, | |
| "num_tokens": 10421917.0, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.019627922853729305, | |
| "grad_norm": 1.729552367155352, | |
| "learning_rate": 1.945392491467577e-05, | |
| "loss": 0.7726, | |
| "num_tokens": 10871710.0, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.02048131080389145, | |
| "grad_norm": 1.8434967482782947, | |
| "learning_rate": 2.0307167235494882e-05, | |
| "loss": 0.7253, | |
| "num_tokens": 11368873.0, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.021334698754053593, | |
| "grad_norm": 1.8640322213390468, | |
| "learning_rate": 2.1160409556313997e-05, | |
| "loss": 0.7744, | |
| "num_tokens": 11819871.0, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.022188086704215737, | |
| "grad_norm": 1.9701041320400614, | |
| "learning_rate": 2.201365187713311e-05, | |
| "loss": 0.7393, | |
| "num_tokens": 12202952.0, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.02304147465437788, | |
| "grad_norm": 1.5669191735272756, | |
| "learning_rate": 2.286689419795222e-05, | |
| "loss": 0.7683, | |
| "num_tokens": 12733944.0, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.023894862604540025, | |
| "grad_norm": 1.6475291377422423, | |
| "learning_rate": 2.3720136518771334e-05, | |
| "loss": 0.7404, | |
| "num_tokens": 13219424.0, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.02474825055470217, | |
| "grad_norm": 1.6445326464409311, | |
| "learning_rate": 2.4573378839590446e-05, | |
| "loss": 0.7317, | |
| "num_tokens": 13674985.0, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.025601638504864313, | |
| "grad_norm": 1.4791858001610043, | |
| "learning_rate": 2.5426621160409557e-05, | |
| "loss": 0.7416, | |
| "num_tokens": 14130423.0, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.026455026455026454, | |
| "grad_norm": 1.625734240185079, | |
| "learning_rate": 2.627986348122867e-05, | |
| "loss": 0.7464, | |
| "num_tokens": 14639707.0, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.027308414405188598, | |
| "grad_norm": 1.663672051577665, | |
| "learning_rate": 2.7133105802047783e-05, | |
| "loss": 0.7506, | |
| "num_tokens": 15118063.0, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.02816180235535074, | |
| "grad_norm": 1.729758637917448, | |
| "learning_rate": 2.7986348122866894e-05, | |
| "loss": 0.762, | |
| "num_tokens": 15588155.0, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.029015190305512886, | |
| "grad_norm": 1.583147120158948, | |
| "learning_rate": 2.883959044368601e-05, | |
| "loss": 0.773, | |
| "num_tokens": 16120926.0, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.02986857825567503, | |
| "grad_norm": 1.650347974549021, | |
| "learning_rate": 2.969283276450512e-05, | |
| "loss": 0.7748, | |
| "num_tokens": 16622154.0, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.030721966205837174, | |
| "grad_norm": 1.6026594473880433, | |
| "learning_rate": 3.054607508532423e-05, | |
| "loss": 0.7031, | |
| "num_tokens": 17042190.0, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.03157535415599932, | |
| "grad_norm": 1.5107968926974098, | |
| "learning_rate": 3.139931740614335e-05, | |
| "loss": 0.7557, | |
| "num_tokens": 17488690.0, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.03242874210616146, | |
| "grad_norm": 1.5595215938461728, | |
| "learning_rate": 3.225255972696246e-05, | |
| "loss": 0.7512, | |
| "num_tokens": 17976522.0, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.033282130056323606, | |
| "grad_norm": 1.6617908648933237, | |
| "learning_rate": 3.310580204778157e-05, | |
| "loss": 0.7599, | |
| "num_tokens": 18452237.0, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.034135518006485746, | |
| "grad_norm": 1.565404710832716, | |
| "learning_rate": 3.395904436860068e-05, | |
| "loss": 0.7086, | |
| "num_tokens": 18906839.0, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.034988905956647894, | |
| "grad_norm": 1.865841075640289, | |
| "learning_rate": 3.48122866894198e-05, | |
| "loss": 0.7623, | |
| "num_tokens": 19413448.0, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.035842293906810034, | |
| "grad_norm": 1.5414831434561684, | |
| "learning_rate": 3.5665529010238906e-05, | |
| "loss": 0.745, | |
| "num_tokens": 19889730.0, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.03669568185697218, | |
| "grad_norm": 1.5168119568091742, | |
| "learning_rate": 3.6518771331058024e-05, | |
| "loss": 0.7275, | |
| "num_tokens": 20362643.0, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.03754906980713432, | |
| "grad_norm": 1.930454368475353, | |
| "learning_rate": 3.7372013651877135e-05, | |
| "loss": 0.6998, | |
| "num_tokens": 20791105.0, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.03840245775729647, | |
| "grad_norm": 1.6010073717621283, | |
| "learning_rate": 3.822525597269625e-05, | |
| "loss": 0.7435, | |
| "num_tokens": 21256777.0, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.03925584570745861, | |
| "grad_norm": 1.5494732570872156, | |
| "learning_rate": 3.907849829351536e-05, | |
| "loss": 0.7465, | |
| "num_tokens": 21704917.0, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.04010923365762076, | |
| "grad_norm": 1.6190828576003704, | |
| "learning_rate": 3.9931740614334476e-05, | |
| "loss": 0.7602, | |
| "num_tokens": 22185554.0, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.0409626216077829, | |
| "grad_norm": 1.4786172771037231, | |
| "learning_rate": 4.078498293515359e-05, | |
| "loss": 0.7384, | |
| "num_tokens": 22641755.0, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.04181600955794504, | |
| "grad_norm": 1.696545361799666, | |
| "learning_rate": 4.16382252559727e-05, | |
| "loss": 0.7297, | |
| "num_tokens": 23146999.0, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.042669397508107186, | |
| "grad_norm": 1.6836150060894002, | |
| "learning_rate": 4.249146757679181e-05, | |
| "loss": 0.7471, | |
| "num_tokens": 23595511.0, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.04352278545826933, | |
| "grad_norm": 1.3649462368231138, | |
| "learning_rate": 4.334470989761093e-05, | |
| "loss": 0.7918, | |
| "num_tokens": 24118220.0, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.044376173408431474, | |
| "grad_norm": 1.6116458751617535, | |
| "learning_rate": 4.419795221843004e-05, | |
| "loss": 0.7442, | |
| "num_tokens": 24604914.0, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.045229561358593615, | |
| "grad_norm": 2.0923322596387215, | |
| "learning_rate": 4.505119453924915e-05, | |
| "loss": 0.7513, | |
| "num_tokens": 25123670.0, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.04608294930875576, | |
| "grad_norm": 1.5206213010037783, | |
| "learning_rate": 4.590443686006826e-05, | |
| "loss": 0.7525, | |
| "num_tokens": 25614380.0, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.0469363372589179, | |
| "grad_norm": 1.4529405567523968, | |
| "learning_rate": 4.675767918088737e-05, | |
| "loss": 0.7455, | |
| "num_tokens": 26164927.0, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.04778972520908005, | |
| "grad_norm": 2.068428332035821, | |
| "learning_rate": 4.7610921501706484e-05, | |
| "loss": 0.7382, | |
| "num_tokens": 26636629.0, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.04864311315924219, | |
| "grad_norm": 1.7574119116801188, | |
| "learning_rate": 4.84641638225256e-05, | |
| "loss": 0.6937, | |
| "num_tokens": 27092899.0, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.04949650110940434, | |
| "grad_norm": 1.871558963994762, | |
| "learning_rate": 4.931740614334471e-05, | |
| "loss": 0.7411, | |
| "num_tokens": 27513214.0, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.05034988905956648, | |
| "grad_norm": 1.4982774488949393, | |
| "learning_rate": 4.999999641601773e-05, | |
| "loss": 0.7004, | |
| "num_tokens": 27982873.0, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.051203277009728626, | |
| "grad_norm": 1.4747335125750545, | |
| "learning_rate": 4.999987097675823e-05, | |
| "loss": 0.7196, | |
| "num_tokens": 28441260.0, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.05205666495989077, | |
| "grad_norm": 1.9186147707485868, | |
| "learning_rate": 4.99995663395271e-05, | |
| "loss": 0.7463, | |
| "num_tokens": 28958598.0, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.05291005291005291, | |
| "grad_norm": 1.5419614690981815, | |
| "learning_rate": 4.999908250675058e-05, | |
| "loss": 0.7287, | |
| "num_tokens": 29484984.0, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.053763440860215055, | |
| "grad_norm": 1.545284237235818, | |
| "learning_rate": 4.999841948228211e-05, | |
| "loss": 0.7488, | |
| "num_tokens": 29961915.0, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.054616828810377195, | |
| "grad_norm": 1.5650635439857692, | |
| "learning_rate": 4.999757727140229e-05, | |
| "loss": 0.7505, | |
| "num_tokens": 30417967.0, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.05547021676053934, | |
| "grad_norm": 1.5662720451827132, | |
| "learning_rate": 4.999655588081883e-05, | |
| "loss": 0.7408, | |
| "num_tokens": 30905676.0, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.05632360471070148, | |
| "grad_norm": 1.4790884805746911, | |
| "learning_rate": 4.999535531866646e-05, | |
| "loss": 0.7089, | |
| "num_tokens": 31367061.0, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.05717699266086363, | |
| "grad_norm": 1.3765457548452067, | |
| "learning_rate": 4.9993975594506975e-05, | |
| "loss": 0.7347, | |
| "num_tokens": 31862251.0, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.05803038061102577, | |
| "grad_norm": 1.3559491046855545, | |
| "learning_rate": 4.999241671932903e-05, | |
| "loss": 0.7641, | |
| "num_tokens": 32351323.0, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.05888376856118792, | |
| "grad_norm": 1.4705848919024582, | |
| "learning_rate": 4.999067870554814e-05, | |
| "loss": 0.7501, | |
| "num_tokens": 32837613.0, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.05973715651135006, | |
| "grad_norm": 1.4932750683350335, | |
| "learning_rate": 4.9988761567006536e-05, | |
| "loss": 0.7352, | |
| "num_tokens": 33256284.0, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.06059054446151221, | |
| "grad_norm": 1.5883669600294514, | |
| "learning_rate": 4.998666531897308e-05, | |
| "loss": 0.7738, | |
| "num_tokens": 33761379.0, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.06144393241167435, | |
| "grad_norm": 1.3497135990866862, | |
| "learning_rate": 4.998438997814312e-05, | |
| "loss": 0.7155, | |
| "num_tokens": 34231333.0, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.06229732036183649, | |
| "grad_norm": 1.3866569803887845, | |
| "learning_rate": 4.9981935562638395e-05, | |
| "loss": 0.7532, | |
| "num_tokens": 34703212.0, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.06315070831199864, | |
| "grad_norm": 1.329023893890578, | |
| "learning_rate": 4.997930209200684e-05, | |
| "loss": 0.7337, | |
| "num_tokens": 35161427.0, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.06400409626216078, | |
| "grad_norm": 1.3726123506316625, | |
| "learning_rate": 4.997648958722248e-05, | |
| "loss": 0.728, | |
| "num_tokens": 35595055.0, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.06485748421232292, | |
| "grad_norm": 1.3990487623850296, | |
| "learning_rate": 4.997349807068521e-05, | |
| "loss": 0.7252, | |
| "num_tokens": 36068371.0, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.06571087216248507, | |
| "grad_norm": 1.2097191821688327, | |
| "learning_rate": 4.997032756622068e-05, | |
| "loss": 0.7592, | |
| "num_tokens": 36530284.0, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.06656426011264721, | |
| "grad_norm": 1.3220720351207618, | |
| "learning_rate": 4.996697809908006e-05, | |
| "loss": 0.798, | |
| "num_tokens": 37049317.0, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.06741764806280935, | |
| "grad_norm": 1.3047177399220788, | |
| "learning_rate": 4.9963449695939824e-05, | |
| "loss": 0.7175, | |
| "num_tokens": 37518919.0, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.06827103601297149, | |
| "grad_norm": 1.3310960648595318, | |
| "learning_rate": 4.995974238490161e-05, | |
| "loss": 0.7165, | |
| "num_tokens": 37961424.0, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.06912442396313365, | |
| "grad_norm": 1.2227488039808938, | |
| "learning_rate": 4.9955856195491904e-05, | |
| "loss": 0.7512, | |
| "num_tokens": 38449337.0, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.06997781191329579, | |
| "grad_norm": 1.310648114420623, | |
| "learning_rate": 4.995179115866189e-05, | |
| "loss": 0.7444, | |
| "num_tokens": 38917825.0, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.07083119986345793, | |
| "grad_norm": 1.957817056349623, | |
| "learning_rate": 4.994754730678713e-05, | |
| "loss": 0.7321, | |
| "num_tokens": 39359207.0, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.07168458781362007, | |
| "grad_norm": 1.306911496638819, | |
| "learning_rate": 4.994312467366738e-05, | |
| "loss": 0.7609, | |
| "num_tokens": 39846738.0, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.07253797576378221, | |
| "grad_norm": 1.2794594670813513, | |
| "learning_rate": 4.9938523294526243e-05, | |
| "loss": 0.7203, | |
| "num_tokens": 40311353.0, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.07339136371394436, | |
| "grad_norm": 1.4448318443084942, | |
| "learning_rate": 4.993374320601095e-05, | |
| "loss": 0.7343, | |
| "num_tokens": 40804304.0, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.0742447516641065, | |
| "grad_norm": 1.5087684442212936, | |
| "learning_rate": 4.992878444619203e-05, | |
| "loss": 0.7456, | |
| "num_tokens": 41244260.0, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.07509813961426864, | |
| "grad_norm": 1.7043643961943444, | |
| "learning_rate": 4.992364705456304e-05, | |
| "loss": 0.7613, | |
| "num_tokens": 41826860.0, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.07595152756443078, | |
| "grad_norm": 1.4073577965500255, | |
| "learning_rate": 4.991833107204022e-05, | |
| "loss": 0.7162, | |
| "num_tokens": 42279795.0, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.07680491551459294, | |
| "grad_norm": 1.2103323507083124, | |
| "learning_rate": 4.9912836540962165e-05, | |
| "loss": 0.7176, | |
| "num_tokens": 42737662.0, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.07765830346475508, | |
| "grad_norm": 1.228293757492738, | |
| "learning_rate": 4.9907163505089535e-05, | |
| "loss": 0.7289, | |
| "num_tokens": 43200466.0, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.07851169141491722, | |
| "grad_norm": 1.2950510394136565, | |
| "learning_rate": 4.9901312009604665e-05, | |
| "loss": 0.782, | |
| "num_tokens": 43675049.0, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.07936507936507936, | |
| "grad_norm": 1.4178723133579891, | |
| "learning_rate": 4.989528210111117e-05, | |
| "loss": 0.756, | |
| "num_tokens": 44122925.0, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.08021846731524152, | |
| "grad_norm": 1.2829343514459521, | |
| "learning_rate": 4.98890738276337e-05, | |
| "loss": 0.6999, | |
| "num_tokens": 44601987.0, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.08107185526540366, | |
| "grad_norm": 1.4374504524826346, | |
| "learning_rate": 4.988268723861739e-05, | |
| "loss": 0.6932, | |
| "num_tokens": 45053630.0, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.0819252432155658, | |
| "grad_norm": 1.3512486308390388, | |
| "learning_rate": 4.9876122384927606e-05, | |
| "loss": 0.7527, | |
| "num_tokens": 45543403.0, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.08277863116572794, | |
| "grad_norm": 1.4860548613695417, | |
| "learning_rate": 4.9869379318849456e-05, | |
| "loss": 0.7689, | |
| "num_tokens": 45978851.0, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.08363201911589008, | |
| "grad_norm": 1.312504852785056, | |
| "learning_rate": 4.9862458094087435e-05, | |
| "loss": 0.7236, | |
| "num_tokens": 46425608.0, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.08448540706605223, | |
| "grad_norm": 1.2303271888413203, | |
| "learning_rate": 4.985535876576493e-05, | |
| "loss": 0.7189, | |
| "num_tokens": 46939676.0, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.08533879501621437, | |
| "grad_norm": 1.3194117498740927, | |
| "learning_rate": 4.984808139042385e-05, | |
| "loss": 0.7435, | |
| "num_tokens": 47466657.0, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.08619218296637651, | |
| "grad_norm": 1.1193590719992939, | |
| "learning_rate": 4.9840626026024094e-05, | |
| "loss": 0.7237, | |
| "num_tokens": 47942013.0, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.08704557091653865, | |
| "grad_norm": 1.4253925033083865, | |
| "learning_rate": 4.983299273194318e-05, | |
| "loss": 0.7362, | |
| "num_tokens": 48417309.0, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.08789895886670081, | |
| "grad_norm": 1.1947452604617999, | |
| "learning_rate": 4.982518156897573e-05, | |
| "loss": 0.7235, | |
| "num_tokens": 48882486.0, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.08875234681686295, | |
| "grad_norm": 1.3561034274556405, | |
| "learning_rate": 4.981719259933295e-05, | |
| "loss": 0.7308, | |
| "num_tokens": 49369486.0, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.08960573476702509, | |
| "grad_norm": 1.4421474252237962, | |
| "learning_rate": 4.980902588664219e-05, | |
| "loss": 0.7568, | |
| "num_tokens": 49823574.0, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.09045912271718723, | |
| "grad_norm": 1.4711373305504778, | |
| "learning_rate": 4.9800681495946424e-05, | |
| "loss": 0.6981, | |
| "num_tokens": 50289138.0, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.09131251066734938, | |
| "grad_norm": 1.4720189999012312, | |
| "learning_rate": 4.979215949370372e-05, | |
| "loss": 0.691, | |
| "num_tokens": 50721951.0, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.09216589861751152, | |
| "grad_norm": 1.4042362668914812, | |
| "learning_rate": 4.9783459947786706e-05, | |
| "loss": 0.7536, | |
| "num_tokens": 51193029.0, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.09301928656767366, | |
| "grad_norm": 1.3883834114302933, | |
| "learning_rate": 4.977458292748204e-05, | |
| "loss": 0.7222, | |
| "num_tokens": 51676036.0, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.0938726745178358, | |
| "grad_norm": 1.5198834286996736, | |
| "learning_rate": 4.9765528503489875e-05, | |
| "loss": 0.6972, | |
| "num_tokens": 52126399.0, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.09472606246799795, | |
| "grad_norm": 1.6818137310699364, | |
| "learning_rate": 4.975629674792326e-05, | |
| "loss": 0.746, | |
| "num_tokens": 52565059.0, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.0955794504181601, | |
| "grad_norm": 1.2941381353196966, | |
| "learning_rate": 4.974688773430759e-05, | |
| "loss": 0.7239, | |
| "num_tokens": 53010874.0, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.09643283836832224, | |
| "grad_norm": 1.115367136570152, | |
| "learning_rate": 4.973730153758e-05, | |
| "loss": 0.6817, | |
| "num_tokens": 53502621.0, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.09728622631848438, | |
| "grad_norm": 1.2236644265034253, | |
| "learning_rate": 4.972753823408882e-05, | |
| "loss": 0.7467, | |
| "num_tokens": 54012592.0, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.09813961426864652, | |
| "grad_norm": 1.3614539179632443, | |
| "learning_rate": 4.9717597901592886e-05, | |
| "loss": 0.695, | |
| "num_tokens": 54452602.0, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.09899300221880868, | |
| "grad_norm": 1.4195040928837384, | |
| "learning_rate": 4.970748061926097e-05, | |
| "loss": 0.7252, | |
| "num_tokens": 54919625.0, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.09984639016897082, | |
| "grad_norm": 1.4702094997469366, | |
| "learning_rate": 4.9697186467671194e-05, | |
| "loss": 0.6918, | |
| "num_tokens": 55377389.0, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.10069977811913296, | |
| "grad_norm": 1.1828673990749254, | |
| "learning_rate": 4.968671552881026e-05, | |
| "loss": 0.7398, | |
| "num_tokens": 55892431.0, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.1015531660692951, | |
| "grad_norm": 1.3726766976603892, | |
| "learning_rate": 4.967606788607292e-05, | |
| "loss": 0.7646, | |
| "num_tokens": 56408483.0, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.10240655401945725, | |
| "grad_norm": 1.4496405231472012, | |
| "learning_rate": 4.966524362426128e-05, | |
| "loss": 0.7378, | |
| "num_tokens": 56909889.0, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.10325994196961939, | |
| "grad_norm": 1.232834277723852, | |
| "learning_rate": 4.965424282958407e-05, | |
| "loss": 0.6929, | |
| "num_tokens": 57369675.0, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.10411332991978153, | |
| "grad_norm": 1.2396125326108323, | |
| "learning_rate": 4.964306558965604e-05, | |
| "loss": 0.676, | |
| "num_tokens": 57850087.0, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.10496671786994367, | |
| "grad_norm": 1.3528031434293055, | |
| "learning_rate": 4.963171199349718e-05, | |
| "loss": 0.7179, | |
| "num_tokens": 58331874.0, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.10582010582010581, | |
| "grad_norm": 1.3381471923806956, | |
| "learning_rate": 4.9620182131532074e-05, | |
| "loss": 0.7139, | |
| "num_tokens": 58759721.0, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.10667349377026797, | |
| "grad_norm": 1.2470605568646416, | |
| "learning_rate": 4.960847609558916e-05, | |
| "loss": 0.7068, | |
| "num_tokens": 59248661.0, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.10752688172043011, | |
| "grad_norm": 1.2526599749042204, | |
| "learning_rate": 4.959659397889998e-05, | |
| "loss": 0.6864, | |
| "num_tokens": 59676292.0, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.10838026967059225, | |
| "grad_norm": 1.2713776223670887, | |
| "learning_rate": 4.958453587609848e-05, | |
| "loss": 0.7325, | |
| "num_tokens": 60199916.0, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.10923365762075439, | |
| "grad_norm": 1.260529146049976, | |
| "learning_rate": 4.9572301883220196e-05, | |
| "loss": 0.7068, | |
| "num_tokens": 60650570.0, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.11008704557091654, | |
| "grad_norm": 1.3188706335664067, | |
| "learning_rate": 4.955989209770155e-05, | |
| "loss": 0.6759, | |
| "num_tokens": 61094637.0, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.11094043352107869, | |
| "grad_norm": 1.2746918780007785, | |
| "learning_rate": 4.954730661837904e-05, | |
| "loss": 0.7157, | |
| "num_tokens": 61630713.0, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.11179382147124083, | |
| "grad_norm": 1.3647999091533176, | |
| "learning_rate": 4.9534545545488454e-05, | |
| "loss": 0.6628, | |
| "num_tokens": 62149084.0, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.11264720942140297, | |
| "grad_norm": 1.1251187771811682, | |
| "learning_rate": 4.952160898066408e-05, | |
| "loss": 0.6955, | |
| "num_tokens": 62662896.0, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.1135005973715651, | |
| "grad_norm": 1.1296326633503708, | |
| "learning_rate": 4.950849702693789e-05, | |
| "loss": 0.696, | |
| "num_tokens": 63159787.0, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.11435398532172726, | |
| "grad_norm": 1.1820195544892163, | |
| "learning_rate": 4.949520978873874e-05, | |
| "loss": 0.7212, | |
| "num_tokens": 63644240.0, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.1152073732718894, | |
| "grad_norm": 1.0907000254670258, | |
| "learning_rate": 4.9481747371891495e-05, | |
| "loss": 0.671, | |
| "num_tokens": 64108252.0, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.11606076122205154, | |
| "grad_norm": 1.2697067765385264, | |
| "learning_rate": 4.946810988361623e-05, | |
| "loss": 0.7178, | |
| "num_tokens": 64564660.0, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.11691414917221368, | |
| "grad_norm": 1.162053282714903, | |
| "learning_rate": 4.945429743252737e-05, | |
| "loss": 0.71, | |
| "num_tokens": 65057882.0, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.11776753712237584, | |
| "grad_norm": 1.1299555887132948, | |
| "learning_rate": 4.9440310128632784e-05, | |
| "loss": 0.7068, | |
| "num_tokens": 65561717.0, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.11862092507253798, | |
| "grad_norm": 1.1824136631290918, | |
| "learning_rate": 4.942614808333296e-05, | |
| "loss": 0.7188, | |
| "num_tokens": 65991955.0, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.11947431302270012, | |
| "grad_norm": 1.1229803744745124, | |
| "learning_rate": 4.9411811409420094e-05, | |
| "loss": 0.703, | |
| "num_tokens": 66485190.0, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.12032770097286226, | |
| "grad_norm": 1.1747259463736386, | |
| "learning_rate": 4.9397300221077194e-05, | |
| "loss": 0.7358, | |
| "num_tokens": 66953512.0, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.12118108892302441, | |
| "grad_norm": 1.2548061942186421, | |
| "learning_rate": 4.9382614633877156e-05, | |
| "loss": 0.7503, | |
| "num_tokens": 67466205.0, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.12203447687318655, | |
| "grad_norm": 1.1195140238315797, | |
| "learning_rate": 4.936775476478187e-05, | |
| "loss": 0.7031, | |
| "num_tokens": 67951050.0, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.1228878648233487, | |
| "grad_norm": 1.2386873813403512, | |
| "learning_rate": 4.93527207321413e-05, | |
| "loss": 0.7079, | |
| "num_tokens": 68426882.0, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.12374125277351083, | |
| "grad_norm": 1.1091542877845337, | |
| "learning_rate": 4.933751265569247e-05, | |
| "loss": 0.7485, | |
| "num_tokens": 68930147.0, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.12459464072367298, | |
| "grad_norm": 1.0745869384498508, | |
| "learning_rate": 4.9322130656558604e-05, | |
| "loss": 0.6681, | |
| "num_tokens": 69444477.0, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.12544802867383512, | |
| "grad_norm": 1.2380251283369912, | |
| "learning_rate": 4.9306574857248065e-05, | |
| "loss": 0.7349, | |
| "num_tokens": 69913688.0, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.12630141662399727, | |
| "grad_norm": 1.3475066813612873, | |
| "learning_rate": 4.929084538165349e-05, | |
| "loss": 0.7064, | |
| "num_tokens": 70332714.0, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.12715480457415942, | |
| "grad_norm": 1.4400683759689517, | |
| "learning_rate": 4.9274942355050705e-05, | |
| "loss": 0.7243, | |
| "num_tokens": 70813408.0, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.12800819252432155, | |
| "grad_norm": 1.2444145807438363, | |
| "learning_rate": 4.9258865904097775e-05, | |
| "loss": 0.7396, | |
| "num_tokens": 71299972.0, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.1288615804744837, | |
| "grad_norm": 1.0242414181042974, | |
| "learning_rate": 4.924261615683398e-05, | |
| "loss": 0.7108, | |
| "num_tokens": 71759515.0, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.12971496842464583, | |
| "grad_norm": 1.188493882753878, | |
| "learning_rate": 4.922619324267881e-05, | |
| "loss": 0.6651, | |
| "num_tokens": 72252819.0, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.130568356374808, | |
| "grad_norm": 1.3126324479191385, | |
| "learning_rate": 4.920959729243091e-05, | |
| "loss": 0.6698, | |
| "num_tokens": 72771440.0, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.13142174432497014, | |
| "grad_norm": 1.1411508728365074, | |
| "learning_rate": 4.919282843826709e-05, | |
| "loss": 0.6685, | |
| "num_tokens": 73242194.0, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.13227513227513227, | |
| "grad_norm": 1.307936016764042, | |
| "learning_rate": 4.91758868137412e-05, | |
| "loss": 0.7338, | |
| "num_tokens": 73786207.0, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.13312852022529442, | |
| "grad_norm": 1.1613529320310767, | |
| "learning_rate": 4.9158772553783105e-05, | |
| "loss": 0.6517, | |
| "num_tokens": 74221358.0, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.13398190817545655, | |
| "grad_norm": 1.1510539792968009, | |
| "learning_rate": 4.914148579469763e-05, | |
| "loss": 0.6938, | |
| "num_tokens": 74742975.0, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.1348352961256187, | |
| "grad_norm": 1.3850463951205576, | |
| "learning_rate": 4.912402667416344e-05, | |
| "loss": 0.7233, | |
| "num_tokens": 75204120.0, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.13568868407578086, | |
| "grad_norm": 1.0345122553008024, | |
| "learning_rate": 4.910639533123193e-05, | |
| "loss": 0.7333, | |
| "num_tokens": 75684137.0, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.13654207202594298, | |
| "grad_norm": 1.1993794512248417, | |
| "learning_rate": 4.90885919063262e-05, | |
| "loss": 0.6847, | |
| "num_tokens": 76160914.0, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.13739545997610514, | |
| "grad_norm": 1.0772949797690683, | |
| "learning_rate": 4.907061654123982e-05, | |
| "loss": 0.6537, | |
| "num_tokens": 76666991.0, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.1382488479262673, | |
| "grad_norm": 1.363751747556385, | |
| "learning_rate": 4.9052469379135796e-05, | |
| "loss": 0.7245, | |
| "num_tokens": 77124379.0, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.13910223587642942, | |
| "grad_norm": 1.1901636342780861, | |
| "learning_rate": 4.903415056454539e-05, | |
| "loss": 0.7103, | |
| "num_tokens": 77613980.0, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.13995562382659157, | |
| "grad_norm": 1.1910421162849438, | |
| "learning_rate": 4.901566024336696e-05, | |
| "loss": 0.691, | |
| "num_tokens": 78170639.0, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.1408090117767537, | |
| "grad_norm": 1.129358397876882, | |
| "learning_rate": 4.899699856286484e-05, | |
| "loss": 0.7184, | |
| "num_tokens": 78667993.0, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.14166239972691586, | |
| "grad_norm": 1.184336348998445, | |
| "learning_rate": 4.8978165671668086e-05, | |
| "loss": 0.7035, | |
| "num_tokens": 79150600.0, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.142515787677078, | |
| "grad_norm": 1.2229940526291831, | |
| "learning_rate": 4.8959161719769395e-05, | |
| "loss": 0.657, | |
| "num_tokens": 79643316.0, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.14336917562724014, | |
| "grad_norm": 1.181968986229247, | |
| "learning_rate": 4.893998685852385e-05, | |
| "loss": 0.7237, | |
| "num_tokens": 80152955.0, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.1442225635774023, | |
| "grad_norm": 1.2870971641864188, | |
| "learning_rate": 4.892064124064768e-05, | |
| "loss": 0.6794, | |
| "num_tokens": 80600797.0, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.14507595152756442, | |
| "grad_norm": 1.1431233390574438, | |
| "learning_rate": 4.8901125020217165e-05, | |
| "loss": 0.6922, | |
| "num_tokens": 81108472.0, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.14592933947772657, | |
| "grad_norm": 1.1722412394244004, | |
| "learning_rate": 4.888143835266726e-05, | |
| "loss": 0.7057, | |
| "num_tokens": 81535402.0, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.14678272742788873, | |
| "grad_norm": 1.3721276457948681, | |
| "learning_rate": 4.8861581394790484e-05, | |
| "loss": 0.7128, | |
| "num_tokens": 81973764.0, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.14763611537805085, | |
| "grad_norm": 1.0428988961171062, | |
| "learning_rate": 4.884155430473557e-05, | |
| "loss": 0.7359, | |
| "num_tokens": 82469903.0, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.148489503328213, | |
| "grad_norm": 1.1967509236053502, | |
| "learning_rate": 4.882135724200628e-05, | |
| "loss": 0.7002, | |
| "num_tokens": 82980445.0, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.14934289127837516, | |
| "grad_norm": 1.2240077286805202, | |
| "learning_rate": 4.8800990367460106e-05, | |
| "loss": 0.7242, | |
| "num_tokens": 83441946.0, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.1501962792285373, | |
| "grad_norm": 1.3039003245323004, | |
| "learning_rate": 4.878045384330698e-05, | |
| "loss": 0.6752, | |
| "num_tokens": 83901702.0, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.15104966717869944, | |
| "grad_norm": 1.1382200264159172, | |
| "learning_rate": 4.875974783310799e-05, | |
| "loss": 0.6533, | |
| "num_tokens": 84319720.0, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.15190305512886157, | |
| "grad_norm": 1.1777300254187848, | |
| "learning_rate": 4.873887250177408e-05, | |
| "loss": 0.7036, | |
| "num_tokens": 84741165.0, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.15275644307902372, | |
| "grad_norm": 1.2494679499905845, | |
| "learning_rate": 4.871782801556476e-05, | |
| "loss": 0.677, | |
| "num_tokens": 85228531.0, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.15360983102918588, | |
| "grad_norm": 1.019491634265827, | |
| "learning_rate": 4.869661454208671e-05, | |
| "loss": 0.6778, | |
| "num_tokens": 85675897.0, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.154463218979348, | |
| "grad_norm": 1.2130254412309611, | |
| "learning_rate": 4.867523225029253e-05, | |
| "loss": 0.7389, | |
| "num_tokens": 86192179.0, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.15531660692951016, | |
| "grad_norm": 1.1180346575861824, | |
| "learning_rate": 4.865368131047933e-05, | |
| "loss": 0.7269, | |
| "num_tokens": 86722595.0, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.15616999487967229, | |
| "grad_norm": 1.0908623944975362, | |
| "learning_rate": 4.8631961894287436e-05, | |
| "loss": 0.7318, | |
| "num_tokens": 87175045.0, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.15702338282983444, | |
| "grad_norm": 1.1815996060870517, | |
| "learning_rate": 4.861007417469895e-05, | |
| "loss": 0.6944, | |
| "num_tokens": 87691018.0, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.1578767707799966, | |
| "grad_norm": 1.189743371344953, | |
| "learning_rate": 4.858801832603643e-05, | |
| "loss": 0.679, | |
| "num_tokens": 88141621.0, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.15873015873015872, | |
| "grad_norm": 1.1579008757807039, | |
| "learning_rate": 4.856579452396148e-05, | |
| "loss": 0.6856, | |
| "num_tokens": 88637962.0, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.15958354668032088, | |
| "grad_norm": 1.05143374186143, | |
| "learning_rate": 4.854340294547334e-05, | |
| "loss": 0.6718, | |
| "num_tokens": 89133287.0, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.16043693463048303, | |
| "grad_norm": 1.255304539583812, | |
| "learning_rate": 4.85208437689075e-05, | |
| "loss": 0.7068, | |
| "num_tokens": 89579835.0, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.16129032258064516, | |
| "grad_norm": 1.140267352892183, | |
| "learning_rate": 4.8498117173934274e-05, | |
| "loss": 0.7009, | |
| "num_tokens": 90060334.0, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.1621437105308073, | |
| "grad_norm": 1.1687127700233666, | |
| "learning_rate": 4.847522334155734e-05, | |
| "loss": 0.6597, | |
| "num_tokens": 90579891.0, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.16299709848096944, | |
| "grad_norm": 1.1682858279061235, | |
| "learning_rate": 4.845216245411234e-05, | |
| "loss": 0.7037, | |
| "num_tokens": 91038959.0, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.1638504864311316, | |
| "grad_norm": 1.0192382594480243, | |
| "learning_rate": 4.842893469526542e-05, | |
| "loss": 0.6798, | |
| "num_tokens": 91542428.0, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.16470387438129375, | |
| "grad_norm": 1.3088556304371173, | |
| "learning_rate": 4.840554025001172e-05, | |
| "loss": 0.6759, | |
| "num_tokens": 92005035.0, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.16555726233145587, | |
| "grad_norm": 1.120110093470671, | |
| "learning_rate": 4.838197930467397e-05, | |
| "loss": 0.7155, | |
| "num_tokens": 92505502.0, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.16641065028161803, | |
| "grad_norm": 1.0045477139885044, | |
| "learning_rate": 4.835825204690096e-05, | |
| "loss": 0.6813, | |
| "num_tokens": 93031983.0, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.16726403823178015, | |
| "grad_norm": 1.037997321014646, | |
| "learning_rate": 4.833435866566607e-05, | |
| "loss": 0.6795, | |
| "num_tokens": 93492833.0, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.1681174261819423, | |
| "grad_norm": 1.070891638236124, | |
| "learning_rate": 4.831029935126572e-05, | |
| "loss": 0.6678, | |
| "num_tokens": 93977963.0, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.16897081413210446, | |
| "grad_norm": 1.1011467509105966, | |
| "learning_rate": 4.828607429531795e-05, | |
| "loss": 0.7554, | |
| "num_tokens": 94474291.0, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.1698242020822666, | |
| "grad_norm": 1.2308698141252665, | |
| "learning_rate": 4.826168369076076e-05, | |
| "loss": 0.6906, | |
| "num_tokens": 94951484.0, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.17067759003242874, | |
| "grad_norm": 1.1337629675512162, | |
| "learning_rate": 4.82371277318507e-05, | |
| "loss": 0.6905, | |
| "num_tokens": 95438170.0, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.1715309779825909, | |
| "grad_norm": 1.2104358520634946, | |
| "learning_rate": 4.8212406614161244e-05, | |
| "loss": 0.6794, | |
| "num_tokens": 421288.0, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.17238436593275303, | |
| "grad_norm": 1.2353406855810571, | |
| "learning_rate": 4.818752053458126e-05, | |
| "loss": 0.7033, | |
| "num_tokens": 879235.0, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.17323775388291518, | |
| "grad_norm": 1.0518126800684762, | |
| "learning_rate": 4.816246969131342e-05, | |
| "loss": 0.6864, | |
| "num_tokens": 1334650.0, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.1740911418330773, | |
| "grad_norm": 1.1609349480835343, | |
| "learning_rate": 4.8137254283872696e-05, | |
| "loss": 0.6996, | |
| "num_tokens": 1864974.0, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.17494452978323946, | |
| "grad_norm": 1.2355298722821886, | |
| "learning_rate": 4.8111874513084656e-05, | |
| "loss": 0.7045, | |
| "num_tokens": 2399861.0, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.17579791773340162, | |
| "grad_norm": 1.0394044113377778, | |
| "learning_rate": 4.808633058108395e-05, | |
| "loss": 0.696, | |
| "num_tokens": 2920092.0, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.17665130568356374, | |
| "grad_norm": 1.1557297341582518, | |
| "learning_rate": 4.806062269131267e-05, | |
| "loss": 0.655, | |
| "num_tokens": 3390315.0, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.1775046936337259, | |
| "grad_norm": 1.0701341821683303, | |
| "learning_rate": 4.803475104851872e-05, | |
| "loss": 0.6751, | |
| "num_tokens": 3945522.0, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.17835808158388802, | |
| "grad_norm": 1.2210516256173702, | |
| "learning_rate": 4.800871585875424e-05, | |
| "loss": 0.7172, | |
| "num_tokens": 4434383.0, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.17921146953405018, | |
| "grad_norm": 1.0817386299007326, | |
| "learning_rate": 4.798251732937387e-05, | |
| "loss": 0.6945, | |
| "num_tokens": 4965050.0, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.18006485748421233, | |
| "grad_norm": 1.0999316550186282, | |
| "learning_rate": 4.795615566903318e-05, | |
| "loss": 0.6711, | |
| "num_tokens": 5453897.0, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.18091824543437446, | |
| "grad_norm": 1.083934913872008, | |
| "learning_rate": 4.792963108768698e-05, | |
| "loss": 0.6775, | |
| "num_tokens": 5919042.0, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.1817716333845366, | |
| "grad_norm": 1.1472880921199, | |
| "learning_rate": 4.7902943796587645e-05, | |
| "loss": 0.6866, | |
| "num_tokens": 6415784.0, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.18262502133469877, | |
| "grad_norm": 1.0826593628805092, | |
| "learning_rate": 4.787609400828343e-05, | |
| "loss": 0.6851, | |
| "num_tokens": 6876666.0, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.1834784092848609, | |
| "grad_norm": 1.0224289865025395, | |
| "learning_rate": 4.78490819366168e-05, | |
| "loss": 0.6485, | |
| "num_tokens": 7357276.0, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.18433179723502305, | |
| "grad_norm": 1.333695422545476, | |
| "learning_rate": 4.782190779672269e-05, | |
| "loss": 0.6996, | |
| "num_tokens": 7785367.0, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.18518518518518517, | |
| "grad_norm": 1.1213413512803674, | |
| "learning_rate": 4.779457180502682e-05, | |
| "loss": 0.7119, | |
| "num_tokens": 8254378.0, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.18603857313534733, | |
| "grad_norm": 1.4925743105514908, | |
| "learning_rate": 4.7767074179243957e-05, | |
| "loss": 0.7025, | |
| "num_tokens": 8707303.0, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.18689196108550948, | |
| "grad_norm": 1.1048073112660228, | |
| "learning_rate": 4.77394151383762e-05, | |
| "loss": 0.6978, | |
| "num_tokens": 9151397.0, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.1877453490356716, | |
| "grad_norm": 1.0780037201458406, | |
| "learning_rate": 4.771159490271121e-05, | |
| "loss": 0.6762, | |
| "num_tokens": 9606447.0, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.18859873698583376, | |
| "grad_norm": 0.9849852514233166, | |
| "learning_rate": 4.768361369382046e-05, | |
| "loss": 0.732, | |
| "num_tokens": 10046430.0, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.1894521249359959, | |
| "grad_norm": 1.1071881064820523, | |
| "learning_rate": 4.765547173455751e-05, | |
| "loss": 0.6644, | |
| "num_tokens": 10464352.0, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.19030551288615805, | |
| "grad_norm": 1.03281339435922, | |
| "learning_rate": 4.762716924905615e-05, | |
| "loss": 0.694, | |
| "num_tokens": 10939418.0, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.1911589008363202, | |
| "grad_norm": 1.0457272637108797, | |
| "learning_rate": 4.7598706462728724e-05, | |
| "loss": 0.7351, | |
| "num_tokens": 11463517.0, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.19201228878648233, | |
| "grad_norm": 1.102053251146349, | |
| "learning_rate": 4.757008360226423e-05, | |
| "loss": 0.6274, | |
| "num_tokens": 11941810.0, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.19286567673664448, | |
| "grad_norm": 0.9964910402343398, | |
| "learning_rate": 4.754130089562658e-05, | |
| "loss": 0.6472, | |
| "num_tokens": 12391040.0, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.19371906468680664, | |
| "grad_norm": 1.0404504289704928, | |
| "learning_rate": 4.751235857205277e-05, | |
| "loss": 0.694, | |
| "num_tokens": 12884964.0, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.19457245263696876, | |
| "grad_norm": 1.1174967829219296, | |
| "learning_rate": 4.748325686205103e-05, | |
| "loss": 0.64, | |
| "num_tokens": 13363123.0, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.19542584058713092, | |
| "grad_norm": 1.0038620952250452, | |
| "learning_rate": 4.7453995997399025e-05, | |
| "loss": 0.6947, | |
| "num_tokens": 13900418.0, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.19627922853729304, | |
| "grad_norm": 1.0216529119957618, | |
| "learning_rate": 4.742457621114198e-05, | |
| "loss": 0.663, | |
| "num_tokens": 14389010.0, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.1971326164874552, | |
| "grad_norm": 1.27036867063678, | |
| "learning_rate": 4.739499773759084e-05, | |
| "loss": 0.692, | |
| "num_tokens": 14884709.0, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.19798600443761735, | |
| "grad_norm": 1.0489710663459084, | |
| "learning_rate": 4.7365260812320395e-05, | |
| "loss": 0.6778, | |
| "num_tokens": 15402588.0, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.19883939238777948, | |
| "grad_norm": 1.0811664736154534, | |
| "learning_rate": 4.733536567216742e-05, | |
| "loss": 0.6461, | |
| "num_tokens": 15814538.0, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.19969278033794163, | |
| "grad_norm": 1.0711433690647847, | |
| "learning_rate": 4.7305312555228764e-05, | |
| "loss": 0.7362, | |
| "num_tokens": 16331982.0, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.20054616828810376, | |
| "grad_norm": 1.0874721305575152, | |
| "learning_rate": 4.7275101700859476e-05, | |
| "loss": 0.6543, | |
| "num_tokens": 16784047.0, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.20139955623826591, | |
| "grad_norm": 0.9190334431115768, | |
| "learning_rate": 4.724473334967087e-05, | |
| "loss": 0.6998, | |
| "num_tokens": 17275371.0, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.20225294418842807, | |
| "grad_norm": 1.176443033005418, | |
| "learning_rate": 4.721420774352866e-05, | |
| "loss": 0.705, | |
| "num_tokens": 17732176.0, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.2031063321385902, | |
| "grad_norm": 0.9879245433547331, | |
| "learning_rate": 4.7183525125550965e-05, | |
| "loss": 0.6588, | |
| "num_tokens": 18192140.0, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.20395972008875235, | |
| "grad_norm": 1.2696399236895504, | |
| "learning_rate": 4.715268574010644e-05, | |
| "loss": 0.6869, | |
| "num_tokens": 18655895.0, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.2048131080389145, | |
| "grad_norm": 1.1183962733988029, | |
| "learning_rate": 4.712168983281228e-05, | |
| "loss": 0.6926, | |
| "num_tokens": 19083334.0, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.20566649598907663, | |
| "grad_norm": 1.1432216535816588, | |
| "learning_rate": 4.709053765053228e-05, | |
| "loss": 0.6894, | |
| "num_tokens": 19593423.0, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.20651988393923879, | |
| "grad_norm": 1.1410144790578511, | |
| "learning_rate": 4.7059229441374894e-05, | |
| "loss": 0.6772, | |
| "num_tokens": 20023796.0, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.2073732718894009, | |
| "grad_norm": 1.0803016455976846, | |
| "learning_rate": 4.7027765454691204e-05, | |
| "loss": 0.6341, | |
| "num_tokens": 20497719.0, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.20822665983956307, | |
| "grad_norm": 0.8975736502997854, | |
| "learning_rate": 4.6996145941073003e-05, | |
| "loss": 0.6698, | |
| "num_tokens": 21006857.0, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.20908004778972522, | |
| "grad_norm": 1.1231707628107581, | |
| "learning_rate": 4.6964371152350735e-05, | |
| "loss": 0.6787, | |
| "num_tokens": 21477891.0, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.20993343573988735, | |
| "grad_norm": 1.026237227187416, | |
| "learning_rate": 4.693244134159153e-05, | |
| "loss": 0.6631, | |
| "num_tokens": 21955868.0, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.2107868236900495, | |
| "grad_norm": 1.1140790482607805, | |
| "learning_rate": 4.690035676309716e-05, | |
| "loss": 0.6814, | |
| "num_tokens": 22520364.0, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.21164021164021163, | |
| "grad_norm": 1.1246414543571086, | |
| "learning_rate": 4.686811767240206e-05, | |
| "loss": 0.6437, | |
| "num_tokens": 23007589.0, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.21249359959037378, | |
| "grad_norm": 1.105284668255734, | |
| "learning_rate": 4.683572432627124e-05, | |
| "loss": 0.7045, | |
| "num_tokens": 23523518.0, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.21334698754053594, | |
| "grad_norm": 1.1324649088135432, | |
| "learning_rate": 4.6803176982698244e-05, | |
| "loss": 0.6566, | |
| "num_tokens": 23910445.0, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.21420037549069806, | |
| "grad_norm": 1.1487613686834988, | |
| "learning_rate": 4.677047590090315e-05, | |
| "loss": 0.6342, | |
| "num_tokens": 24373538.0, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.21505376344086022, | |
| "grad_norm": 1.2715099668807717, | |
| "learning_rate": 4.6737621341330454e-05, | |
| "loss": 0.6977, | |
| "num_tokens": 24865513.0, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.21590715139102234, | |
| "grad_norm": 1.1514001156275855, | |
| "learning_rate": 4.6704613565647005e-05, | |
| "loss": 0.7177, | |
| "num_tokens": 25386703.0, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.2167605393411845, | |
| "grad_norm": 0.9774936460308985, | |
| "learning_rate": 4.667145283673993e-05, | |
| "loss": 0.6766, | |
| "num_tokens": 25877951.0, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.21761392729134665, | |
| "grad_norm": 1.2440248436188435, | |
| "learning_rate": 4.663813941871454e-05, | |
| "loss": 0.663, | |
| "num_tokens": 26407535.0, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.21846731524150878, | |
| "grad_norm": 1.0641868684389462, | |
| "learning_rate": 4.6604673576892216e-05, | |
| "loss": 0.6613, | |
| "num_tokens": 26873523.0, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.21932070319167093, | |
| "grad_norm": 0.9882397960743491, | |
| "learning_rate": 4.657105557780831e-05, | |
| "loss": 0.683, | |
| "num_tokens": 27345312.0, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.2201740911418331, | |
| "grad_norm": 0.96889308591211, | |
| "learning_rate": 4.653728568921001e-05, | |
| "loss": 0.6703, | |
| "num_tokens": 27818123.0, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.22102747909199522, | |
| "grad_norm": 1.03389444953015, | |
| "learning_rate": 4.650336418005423e-05, | |
| "loss": 0.6913, | |
| "num_tokens": 28283920.0, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.22188086704215737, | |
| "grad_norm": 0.992222197396554, | |
| "learning_rate": 4.6469291320505423e-05, | |
| "loss": 0.6861, | |
| "num_tokens": 28802356.0, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.2227342549923195, | |
| "grad_norm": 1.0960140838999854, | |
| "learning_rate": 4.643506738193346e-05, | |
| "loss": 0.7201, | |
| "num_tokens": 29231839.0, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.22358764294248165, | |
| "grad_norm": 1.244536050952739, | |
| "learning_rate": 4.64006926369115e-05, | |
| "loss": 0.6644, | |
| "num_tokens": 29719816.0, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.2244410308926438, | |
| "grad_norm": 1.0506600574878056, | |
| "learning_rate": 4.6366167359213744e-05, | |
| "loss": 0.6563, | |
| "num_tokens": 30171452.0, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.22529441884280593, | |
| "grad_norm": 1.0321133449697444, | |
| "learning_rate": 4.6331491823813325e-05, | |
| "loss": 0.6675, | |
| "num_tokens": 30656354.0, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.2261478067929681, | |
| "grad_norm": 0.905068136544555, | |
| "learning_rate": 4.629666630688006e-05, | |
| "loss": 0.6644, | |
| "num_tokens": 31143082.0, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.2270011947431302, | |
| "grad_norm": 0.9569551554854419, | |
| "learning_rate": 4.6261691085778315e-05, | |
| "loss": 0.6478, | |
| "num_tokens": 31664428.0, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.22785458269329237, | |
| "grad_norm": 0.9719534189212212, | |
| "learning_rate": 4.622656643906472e-05, | |
| "loss": 0.6515, | |
| "num_tokens": 32152562.0, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.22870797064345452, | |
| "grad_norm": 0.9550619554021147, | |
| "learning_rate": 4.619129264648602e-05, | |
| "loss": 0.6569, | |
| "num_tokens": 32637004.0, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.22956135859361665, | |
| "grad_norm": 1.0541566817823893, | |
| "learning_rate": 4.615586998897681e-05, | |
| "loss": 0.6805, | |
| "num_tokens": 33091371.0, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.2304147465437788, | |
| "grad_norm": 0.9447239325405086, | |
| "learning_rate": 4.6120298748657295e-05, | |
| "loss": 0.6662, | |
| "num_tokens": 33611317.0, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.23126813449394096, | |
| "grad_norm": 1.3796632522640218, | |
| "learning_rate": 4.6084579208831066e-05, | |
| "loss": 0.6974, | |
| "num_tokens": 34103804.0, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.23212152244410308, | |
| "grad_norm": 1.0266712169012189, | |
| "learning_rate": 4.604871165398282e-05, | |
| "loss": 0.6952, | |
| "num_tokens": 34532886.0, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.23297491039426524, | |
| "grad_norm": 0.9833688702297602, | |
| "learning_rate": 4.601269636977611e-05, | |
| "loss": 0.6767, | |
| "num_tokens": 35015944.0, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.23382829834442737, | |
| "grad_norm": 1.0726665743497261, | |
| "learning_rate": 4.5976533643051076e-05, | |
| "loss": 0.6559, | |
| "num_tokens": 35457611.0, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.23468168629458952, | |
| "grad_norm": 1.0076220154429152, | |
| "learning_rate": 4.594022376182212e-05, | |
| "loss": 0.6934, | |
| "num_tokens": 35915425.0, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.23553507424475167, | |
| "grad_norm": 1.0848385881255227, | |
| "learning_rate": 4.590376701527566e-05, | |
| "loss": 0.7094, | |
| "num_tokens": 36409135.0, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.2363884621949138, | |
| "grad_norm": 1.07913510618759, | |
| "learning_rate": 4.586716369376782e-05, | |
| "loss": 0.7151, | |
| "num_tokens": 36931490.0, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.23724185014507596, | |
| "grad_norm": 1.0318989559802685, | |
| "learning_rate": 4.5830414088822097e-05, | |
| "loss": 0.6608, | |
| "num_tokens": 37411796.0, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.23809523809523808, | |
| "grad_norm": 0.9603277138511171, | |
| "learning_rate": 4.579351849312703e-05, | |
| "loss": 0.6751, | |
| "num_tokens": 37874802.0, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.23894862604540024, | |
| "grad_norm": 1.1478993387134349, | |
| "learning_rate": 4.575647720053389e-05, | |
| "loss": 0.7149, | |
| "num_tokens": 38327812.0, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.2398020139955624, | |
| "grad_norm": 1.0990531300486552, | |
| "learning_rate": 4.5719290506054366e-05, | |
| "loss": 0.6453, | |
| "num_tokens": 421060.0, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.24065540194572452, | |
| "grad_norm": 0.9921453796582193, | |
| "learning_rate": 4.5681958705858155e-05, | |
| "loss": 0.6657, | |
| "num_tokens": 877105.0, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.24150878989588667, | |
| "grad_norm": 1.028795003501946, | |
| "learning_rate": 4.564448209727066e-05, | |
| "loss": 0.6523, | |
| "num_tokens": 1339698.0, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.24236217784604883, | |
| "grad_norm": 1.2134426086619026, | |
| "learning_rate": 4.5606860978770554e-05, | |
| "loss": 0.6807, | |
| "num_tokens": 1775407.0, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.24321556579621095, | |
| "grad_norm": 0.9964638495357863, | |
| "learning_rate": 4.55690956499875e-05, | |
| "loss": 0.6524, | |
| "num_tokens": 2215634.0, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.2440689537463731, | |
| "grad_norm": 1.1990942172437682, | |
| "learning_rate": 4.553118641169967e-05, | |
| "loss": 0.6732, | |
| "num_tokens": 2741804.0, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.24492234169653523, | |
| "grad_norm": 1.1330285046336883, | |
| "learning_rate": 4.5493133565831395e-05, | |
| "loss": 0.656, | |
| "num_tokens": 3266679.0, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.2457757296466974, | |
| "grad_norm": 1.0388065194387546, | |
| "learning_rate": 4.5454937415450774e-05, | |
| "loss": 0.6909, | |
| "num_tokens": 3756299.0, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.24662911759685954, | |
| "grad_norm": 0.9196704665381511, | |
| "learning_rate": 4.54165982647672e-05, | |
| "loss": 0.6585, | |
| "num_tokens": 4261608.0, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.24748250554702167, | |
| "grad_norm": 1.104949625525001, | |
| "learning_rate": 4.5378116419129035e-05, | |
| "loss": 0.6712, | |
| "num_tokens": 4687015.0, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.24833589349718382, | |
| "grad_norm": 0.9995929020146768, | |
| "learning_rate": 4.5339492185021066e-05, | |
| "loss": 0.6367, | |
| "num_tokens": 5154135.0, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.24918928144734595, | |
| "grad_norm": 1.1234154727606218, | |
| "learning_rate": 4.5300725870062153e-05, | |
| "loss": 0.6432, | |
| "num_tokens": 5646859.0, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.25004266939750813, | |
| "grad_norm": 1.1513733815654665, | |
| "learning_rate": 4.5261817783002726e-05, | |
| "loss": 0.6696, | |
| "num_tokens": 6122696.0, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.25089605734767023, | |
| "grad_norm": 1.0606906202691488, | |
| "learning_rate": 4.522276823372236e-05, | |
| "loss": 0.651, | |
| "num_tokens": 6525181.0, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.2517494452978324, | |
| "grad_norm": 0.9055146266707058, | |
| "learning_rate": 4.518357753322728e-05, | |
| "loss": 0.6346, | |
| "num_tokens": 6990704.0, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.25260283324799454, | |
| "grad_norm": 1.1044997308379876, | |
| "learning_rate": 4.5144245993647896e-05, | |
| "loss": 0.6767, | |
| "num_tokens": 7430102.0, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.2534562211981567, | |
| "grad_norm": 0.9035531176190219, | |
| "learning_rate": 4.5104773928236324e-05, | |
| "loss": 0.6372, | |
| "num_tokens": 8010997.0, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.25430960914831885, | |
| "grad_norm": 1.0509691108933237, | |
| "learning_rate": 4.506516165136388e-05, | |
| "loss": 0.6853, | |
| "num_tokens": 8514125.0, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.25516299709848095, | |
| "grad_norm": 0.9387925154248077, | |
| "learning_rate": 4.502540947851859e-05, | |
| "loss": 0.7308, | |
| "num_tokens": 9023552.0, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.2560163850486431, | |
| "grad_norm": 0.9831579527984785, | |
| "learning_rate": 4.498551772630264e-05, | |
| "loss": 0.6966, | |
| "num_tokens": 9511775.0, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.25686977299880526, | |
| "grad_norm": 1.0854175039973468, | |
| "learning_rate": 4.494548671242991e-05, | |
| "loss": 0.6905, | |
| "num_tokens": 9998350.0, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.2577231609489674, | |
| "grad_norm": 1.1084538715121595, | |
| "learning_rate": 4.490531675572341e-05, | |
| "loss": 0.6185, | |
| "num_tokens": 10442163.0, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.25857654889912957, | |
| "grad_norm": 1.0374246201385566, | |
| "learning_rate": 4.486500817611273e-05, | |
| "loss": 0.6565, | |
| "num_tokens": 10874287.0, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.25942993684929166, | |
| "grad_norm": 1.0288126747538786, | |
| "learning_rate": 4.482456129463153e-05, | |
| "loss": 0.6731, | |
| "num_tokens": 11370722.0, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.2602833247994538, | |
| "grad_norm": 1.0219221093241808, | |
| "learning_rate": 4.478397643341495e-05, | |
| "loss": 0.6421, | |
| "num_tokens": 11837926.0, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.261136712749616, | |
| "grad_norm": 0.9861358682693042, | |
| "learning_rate": 4.474325391569706e-05, | |
| "loss": 0.6584, | |
| "num_tokens": 12331519.0, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.26199010069977813, | |
| "grad_norm": 0.973697367336259, | |
| "learning_rate": 4.470239406580827e-05, | |
| "loss": 0.6459, | |
| "num_tokens": 12836674.0, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.2628434886499403, | |
| "grad_norm": 1.0550005993382137, | |
| "learning_rate": 4.466139720917277e-05, | |
| "loss": 0.6312, | |
| "num_tokens": 13241414.0, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.2636968766001024, | |
| "grad_norm": 0.9104618916882102, | |
| "learning_rate": 4.4620263672305916e-05, | |
| "loss": 0.6637, | |
| "num_tokens": 13703806.0, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.26455026455026454, | |
| "grad_norm": 1.0295282137438284, | |
| "learning_rate": 4.457899378281167e-05, | |
| "loss": 0.636, | |
| "num_tokens": 14190906.0, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.2654036525004267, | |
| "grad_norm": 0.895888271171603, | |
| "learning_rate": 4.453758786937992e-05, | |
| "loss": 0.6742, | |
| "num_tokens": 14719491.0, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.26625704045058884, | |
| "grad_norm": 0.918926913872113, | |
| "learning_rate": 4.449604626178393e-05, | |
| "loss": 0.6785, | |
| "num_tokens": 15174140.0, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.267110428400751, | |
| "grad_norm": 0.9505309720082539, | |
| "learning_rate": 4.445436929087767e-05, | |
| "loss": 0.7024, | |
| "num_tokens": 15688067.0, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.2679638163509131, | |
| "grad_norm": 0.9369802432268316, | |
| "learning_rate": 4.441255728859321e-05, | |
| "loss": 0.6402, | |
| "num_tokens": 16217563.0, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.26881720430107525, | |
| "grad_norm": 1.0494626842513919, | |
| "learning_rate": 4.437061058793806e-05, | |
| "loss": 0.6755, | |
| "num_tokens": 16704728.0, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.2696705922512374, | |
| "grad_norm": 1.0331486147107156, | |
| "learning_rate": 4.432852952299252e-05, | |
| "loss": 0.6699, | |
| "num_tokens": 17146472.0, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.27052398020139956, | |
| "grad_norm": 1.1119123215431965, | |
| "learning_rate": 4.428631442890702e-05, | |
| "loss": 0.6374, | |
| "num_tokens": 17589690.0, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.2713773681515617, | |
| "grad_norm": 1.1833713096668448, | |
| "learning_rate": 4.424396564189947e-05, | |
| "loss": 0.6965, | |
| "num_tokens": 18069122.0, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.27223075610172387, | |
| "grad_norm": 1.0755796928146804, | |
| "learning_rate": 4.420148349925252e-05, | |
| "loss": 0.6645, | |
| "num_tokens": 18595086.0, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.27308414405188597, | |
| "grad_norm": 1.1333750082764271, | |
| "learning_rate": 4.415886833931097e-05, | |
| "loss": 0.6428, | |
| "num_tokens": 19043696.0, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.2739375320020481, | |
| "grad_norm": 1.043025277361079, | |
| "learning_rate": 4.411612050147899e-05, | |
| "loss": 0.619, | |
| "num_tokens": 19518397.0, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.2747909199522103, | |
| "grad_norm": 0.9822809038945924, | |
| "learning_rate": 4.4073240326217446e-05, | |
| "loss": 0.6365, | |
| "num_tokens": 19970492.0, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.27564430790237243, | |
| "grad_norm": 0.9891307971026575, | |
| "learning_rate": 4.403022815504122e-05, | |
| "loss": 0.6459, | |
| "num_tokens": 20461858.0, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.2764976958525346, | |
| "grad_norm": 1.0566611352935433, | |
| "learning_rate": 4.398708433051645e-05, | |
| "loss": 0.6543, | |
| "num_tokens": 20947870.0, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.2773510838026967, | |
| "grad_norm": 1.0064327049389676, | |
| "learning_rate": 4.3943809196257794e-05, | |
| "loss": 0.7042, | |
| "num_tokens": 21466994.0, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.27820447175285884, | |
| "grad_norm": 0.998464431867688, | |
| "learning_rate": 4.390040309692574e-05, | |
| "loss": 0.6652, | |
| "num_tokens": 21934487.0, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.279057859703021, | |
| "grad_norm": 1.1034198382023583, | |
| "learning_rate": 4.385686637822382e-05, | |
| "loss": 0.7008, | |
| "num_tokens": 22428073.0, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.27991124765318315, | |
| "grad_norm": 1.1452495006729742, | |
| "learning_rate": 4.381319938689588e-05, | |
| "loss": 0.6417, | |
| "num_tokens": 22917591.0, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.2807646356033453, | |
| "grad_norm": 1.1134914421972593, | |
| "learning_rate": 4.376940247072331e-05, | |
| "loss": 0.6593, | |
| "num_tokens": 23362074.0, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.2816180235535074, | |
| "grad_norm": 0.9360103567579704, | |
| "learning_rate": 4.372547597852225e-05, | |
| "loss": 0.6506, | |
| "num_tokens": 23813090.0, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.28247141150366956, | |
| "grad_norm": 0.9486301910092465, | |
| "learning_rate": 4.368142026014086e-05, | |
| "loss": 0.6631, | |
| "num_tokens": 24277894.0, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.2833247994538317, | |
| "grad_norm": 0.9431082592173952, | |
| "learning_rate": 4.3637235666456506e-05, | |
| "loss": 0.6129, | |
| "num_tokens": 24749220.0, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.28417818740399386, | |
| "grad_norm": 1.1286583774589005, | |
| "learning_rate": 4.359292254937296e-05, | |
| "loss": 0.6812, | |
| "num_tokens": 25227258.0, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.285031575354156, | |
| "grad_norm": 0.9400669821950434, | |
| "learning_rate": 4.354848126181762e-05, | |
| "loss": 0.6652, | |
| "num_tokens": 25733368.0, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.2858849633043181, | |
| "grad_norm": 1.0759732617628444, | |
| "learning_rate": 4.350391215773867e-05, | |
| "loss": 0.6667, | |
| "num_tokens": 26193707.0, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.2867383512544803, | |
| "grad_norm": 1.0322942760702165, | |
| "learning_rate": 4.345921559210227e-05, | |
| "loss": 0.6621, | |
| "num_tokens": 26667344.0, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.2875917392046424, | |
| "grad_norm": 0.9563146317074364, | |
| "learning_rate": 4.341439192088976e-05, | |
| "loss": 0.638, | |
| "num_tokens": 27107554.0, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.2884451271548046, | |
| "grad_norm": 1.1135033435484845, | |
| "learning_rate": 4.336944150109478e-05, | |
| "loss": 0.6513, | |
| "num_tokens": 27590348.0, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.28929851510496674, | |
| "grad_norm": 1.1239383737366093, | |
| "learning_rate": 4.332436469072044e-05, | |
| "loss": 0.6513, | |
| "num_tokens": 28111682.0, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.29015190305512883, | |
| "grad_norm": 1.0086351161064906, | |
| "learning_rate": 4.327916184877652e-05, | |
| "loss": 0.666, | |
| "num_tokens": 28565548.0, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.291005291005291, | |
| "grad_norm": 0.9803052286266831, | |
| "learning_rate": 4.3233833335276494e-05, | |
| "loss": 0.6848, | |
| "num_tokens": 29069662.0, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.29185867895545314, | |
| "grad_norm": 1.2970656590944643, | |
| "learning_rate": 4.31883795112348e-05, | |
| "loss": 0.6702, | |
| "num_tokens": 29540008.0, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.2927120669056153, | |
| "grad_norm": 1.002197523474358, | |
| "learning_rate": 4.314280073866386e-05, | |
| "loss": 0.6677, | |
| "num_tokens": 30069043.0, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.29356545485577745, | |
| "grad_norm": 1.158257731716908, | |
| "learning_rate": 4.3097097380571256e-05, | |
| "loss": 0.6536, | |
| "num_tokens": 30608334.0, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.2944188428059396, | |
| "grad_norm": 1.1565617780482107, | |
| "learning_rate": 4.305126980095681e-05, | |
| "loss": 0.627, | |
| "num_tokens": 31077481.0, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.2952722307561017, | |
| "grad_norm": 1.0631530173310553, | |
| "learning_rate": 4.300531836480968e-05, | |
| "loss": 0.6164, | |
| "num_tokens": 31523212.0, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.29612561870626386, | |
| "grad_norm": 1.057549223421548, | |
| "learning_rate": 4.295924343810551e-05, | |
| "loss": 0.6402, | |
| "num_tokens": 32009162.0, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.296979006656426, | |
| "grad_norm": 0.9450402858117135, | |
| "learning_rate": 4.291304538780343e-05, | |
| "loss": 0.6341, | |
| "num_tokens": 32483562.0, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.29783239460658817, | |
| "grad_norm": 0.983929424012367, | |
| "learning_rate": 4.286672458184319e-05, | |
| "loss": 0.6368, | |
| "num_tokens": 33004902.0, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.2986857825567503, | |
| "grad_norm": 0.9137933823639911, | |
| "learning_rate": 4.282028138914221e-05, | |
| "loss": 0.6215, | |
| "num_tokens": 33504512.0, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.2995391705069124, | |
| "grad_norm": 0.9871617316090521, | |
| "learning_rate": 4.2773716179592666e-05, | |
| "loss": 0.6432, | |
| "num_tokens": 33951768.0, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.3003925584570746, | |
| "grad_norm": 1.160100489576663, | |
| "learning_rate": 4.27270293240585e-05, | |
| "loss": 0.6645, | |
| "num_tokens": 34457317.0, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.30124594640723673, | |
| "grad_norm": 0.8685501113426534, | |
| "learning_rate": 4.26802211943725e-05, | |
| "loss": 0.661, | |
| "num_tokens": 34964422.0, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.3020993343573989, | |
| "grad_norm": 0.9480039454372314, | |
| "learning_rate": 4.263329216333335e-05, | |
| "loss": 0.6468, | |
| "num_tokens": 35451730.0, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.30295272230756104, | |
| "grad_norm": 1.034557261853204, | |
| "learning_rate": 4.258624260470262e-05, | |
| "loss": 0.6767, | |
| "num_tokens": 35975531.0, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.30380611025772314, | |
| "grad_norm": 1.0583649410054492, | |
| "learning_rate": 4.253907289320179e-05, | |
| "loss": 0.6699, | |
| "num_tokens": 36423440.0, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.3046594982078853, | |
| "grad_norm": 0.9244887110033405, | |
| "learning_rate": 4.249178340450933e-05, | |
| "loss": 0.6536, | |
| "num_tokens": 36907135.0, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.30551288615804745, | |
| "grad_norm": 1.0956310368630402, | |
| "learning_rate": 4.244437451525764e-05, | |
| "loss": 0.6119, | |
| "num_tokens": 37349185.0, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.3063662741082096, | |
| "grad_norm": 0.8894609101942197, | |
| "learning_rate": 4.239684660303006e-05, | |
| "loss": 0.6206, | |
| "num_tokens": 37857975.0, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.30721966205837176, | |
| "grad_norm": 1.0075648632984608, | |
| "learning_rate": 4.234920004635792e-05, | |
| "loss": 0.6587, | |
| "num_tokens": 38366018.0, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.30807305000853386, | |
| "grad_norm": 0.9178640617606731, | |
| "learning_rate": 4.230143522471743e-05, | |
| "loss": 0.6422, | |
| "num_tokens": 38885944.0, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.308926437958696, | |
| "grad_norm": 0.9916540849655225, | |
| "learning_rate": 4.225355251852675e-05, | |
| "loss": 0.6407, | |
| "num_tokens": 39400675.0, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.30977982590885816, | |
| "grad_norm": 0.9767548190932287, | |
| "learning_rate": 4.2205552309142885e-05, | |
| "loss": 0.6349, | |
| "num_tokens": 39850819.0, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.3106332138590203, | |
| "grad_norm": 0.9626467489189492, | |
| "learning_rate": 4.215743497885873e-05, | |
| "loss": 0.6825, | |
| "num_tokens": 40317411.0, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.3114866018091825, | |
| "grad_norm": 0.9835758110325883, | |
| "learning_rate": 4.2109200910899916e-05, | |
| "loss": 0.6379, | |
| "num_tokens": 40821787.0, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.31233998975934457, | |
| "grad_norm": 0.8761549604162322, | |
| "learning_rate": 4.206085048942187e-05, | |
| "loss": 0.6712, | |
| "num_tokens": 41357145.0, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.3131933777095067, | |
| "grad_norm": 1.1816568369777558, | |
| "learning_rate": 4.2012384099506694e-05, | |
| "loss": 0.626, | |
| "num_tokens": 41867236.0, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.3140467656596689, | |
| "grad_norm": 1.0101399172107357, | |
| "learning_rate": 4.196380212716008e-05, | |
| "loss": 0.6177, | |
| "num_tokens": 42320349.0, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.31490015360983103, | |
| "grad_norm": 1.1136485632953617, | |
| "learning_rate": 4.191510495930831e-05, | |
| "loss": 0.6361, | |
| "num_tokens": 42827624.0, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.3157535415599932, | |
| "grad_norm": 0.9957428721397101, | |
| "learning_rate": 4.1866292983795084e-05, | |
| "loss": 0.6191, | |
| "num_tokens": 43289900.0, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.31660692951015534, | |
| "grad_norm": 0.8747769330269698, | |
| "learning_rate": 4.1817366589378526e-05, | |
| "loss": 0.612, | |
| "num_tokens": 43788954.0, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.31746031746031744, | |
| "grad_norm": 0.8584924919689573, | |
| "learning_rate": 4.1768326165727975e-05, | |
| "loss": 0.6061, | |
| "num_tokens": 44253905.0, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.3183137054104796, | |
| "grad_norm": 1.062043835867046, | |
| "learning_rate": 4.171917210342101e-05, | |
| "loss": 0.6599, | |
| "num_tokens": 44729983.0, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.31916709336064175, | |
| "grad_norm": 1.0217914781820183, | |
| "learning_rate": 4.166990479394023e-05, | |
| "loss": 0.648, | |
| "num_tokens": 45195098.0, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.3200204813108039, | |
| "grad_norm": 0.9201344018326302, | |
| "learning_rate": 4.1620524629670196e-05, | |
| "loss": 0.6671, | |
| "num_tokens": 45672842.0, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.32087386926096606, | |
| "grad_norm": 1.1397216517807938, | |
| "learning_rate": 4.157103200389428e-05, | |
| "loss": 0.6322, | |
| "num_tokens": 46151658.0, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.32172725721112816, | |
| "grad_norm": 0.9990337165224222, | |
| "learning_rate": 4.152142731079155e-05, | |
| "loss": 0.6761, | |
| "num_tokens": 46666876.0, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.3225806451612903, | |
| "grad_norm": 1.208502511536086, | |
| "learning_rate": 4.147171094543363e-05, | |
| "loss": 0.6406, | |
| "num_tokens": 47099591.0, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.32343403311145247, | |
| "grad_norm": 0.9304907451089872, | |
| "learning_rate": 4.142188330378155e-05, | |
| "loss": 0.6854, | |
| "num_tokens": 47632239.0, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.3242874210616146, | |
| "grad_norm": 1.0436402924297088, | |
| "learning_rate": 4.1371944782682536e-05, | |
| "loss": 0.6202, | |
| "num_tokens": 48095443.0, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.3251408090117768, | |
| "grad_norm": 0.9644270813331612, | |
| "learning_rate": 4.132189577986699e-05, | |
| "loss": 0.6237, | |
| "num_tokens": 48554427.0, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.3259941969619389, | |
| "grad_norm": 1.1748268554914663, | |
| "learning_rate": 4.127173669394516e-05, | |
| "loss": 0.6729, | |
| "num_tokens": 49033431.0, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.32684758491210103, | |
| "grad_norm": 1.0380245238600094, | |
| "learning_rate": 4.1221467924404075e-05, | |
| "loss": 0.6324, | |
| "num_tokens": 49483384.0, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 0.3277009728622632, | |
| "grad_norm": 0.9648072557402981, | |
| "learning_rate": 4.117108987160432e-05, | |
| "loss": 0.6481, | |
| "num_tokens": 49980147.0, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.32855436081242534, | |
| "grad_norm": 1.0206985745749346, | |
| "learning_rate": 4.112060293677687e-05, | |
| "loss": 0.6373, | |
| "num_tokens": 50417382.0, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.3294077487625875, | |
| "grad_norm": 0.8984622561385278, | |
| "learning_rate": 4.107000752201984e-05, | |
| "loss": 0.6332, | |
| "num_tokens": 50943097.0, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.3302611367127496, | |
| "grad_norm": 1.0000056966065096, | |
| "learning_rate": 4.101930403029538e-05, | |
| "loss": 0.6896, | |
| "num_tokens": 51447513.0, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.33111452466291175, | |
| "grad_norm": 1.3267825088425038, | |
| "learning_rate": 4.0968492865426367e-05, | |
| "loss": 0.6939, | |
| "num_tokens": 51924577.0, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.3319679126130739, | |
| "grad_norm": 1.0364145972348338, | |
| "learning_rate": 4.091757443209322e-05, | |
| "loss": 0.6587, | |
| "num_tokens": 52461321.0, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 0.33282130056323606, | |
| "grad_norm": 0.9473197857926919, | |
| "learning_rate": 4.0866549135830745e-05, | |
| "loss": 0.6492, | |
| "num_tokens": 52942461.0, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.3336746885133982, | |
| "grad_norm": 1.027596288413729, | |
| "learning_rate": 4.081541738302478e-05, | |
| "loss": 0.6149, | |
| "num_tokens": 53444941.0, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 0.3345280764635603, | |
| "grad_norm": 1.0567423158418994, | |
| "learning_rate": 4.076417958090906e-05, | |
| "loss": 0.7059, | |
| "num_tokens": 53933388.0, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.33538146441372246, | |
| "grad_norm": 0.9817599310427655, | |
| "learning_rate": 4.071283613756196e-05, | |
| "loss": 0.6472, | |
| "num_tokens": 54391687.0, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.3362348523638846, | |
| "grad_norm": 0.995280470621319, | |
| "learning_rate": 4.0661387461903154e-05, | |
| "loss": 0.6344, | |
| "num_tokens": 54855386.0, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.33708824031404677, | |
| "grad_norm": 0.9154397814757621, | |
| "learning_rate": 4.060983396369051e-05, | |
| "loss": 0.6384, | |
| "num_tokens": 55368447.0, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.3379416282642089, | |
| "grad_norm": 0.9448740168747737, | |
| "learning_rate": 4.055817605351669e-05, | |
| "loss": 0.6411, | |
| "num_tokens": 55814026.0, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.3387950162143711, | |
| "grad_norm": 1.094004234955502, | |
| "learning_rate": 4.050641414280597e-05, | |
| "loss": 0.614, | |
| "num_tokens": 56255121.0, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 0.3396484041645332, | |
| "grad_norm": 0.7742679468622998, | |
| "learning_rate": 4.045454864381088e-05, | |
| "loss": 0.6796, | |
| "num_tokens": 56796489.0, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.34050179211469533, | |
| "grad_norm": 0.9621627980405628, | |
| "learning_rate": 4.0402579969609024e-05, | |
| "loss": 0.6284, | |
| "num_tokens": 57279668.0, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.3413551800648575, | |
| "grad_norm": 0.9077184248917183, | |
| "learning_rate": 4.035050853409969e-05, | |
| "loss": 0.6386, | |
| "num_tokens": 57746160.0, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.34220856801501964, | |
| "grad_norm": 0.8295552770492037, | |
| "learning_rate": 4.0298334752000634e-05, | |
| "loss": 0.6142, | |
| "num_tokens": 58261166.0, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 0.3430619559651818, | |
| "grad_norm": 0.9468242453122664, | |
| "learning_rate": 4.0246059038844716e-05, | |
| "loss": 0.6364, | |
| "num_tokens": 58738443.0, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.3439153439153439, | |
| "grad_norm": 1.0397478691275794, | |
| "learning_rate": 4.019368181097663e-05, | |
| "loss": 0.6293, | |
| "num_tokens": 59205817.0, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 0.34476873186550605, | |
| "grad_norm": 1.0350500134093694, | |
| "learning_rate": 4.0141203485549564e-05, | |
| "loss": 0.6505, | |
| "num_tokens": 59653617.0, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.3456221198156682, | |
| "grad_norm": 0.9043465643179709, | |
| "learning_rate": 4.008862448052188e-05, | |
| "loss": 0.6542, | |
| "num_tokens": 60145641.0, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.34647550776583036, | |
| "grad_norm": 0.9539526130039545, | |
| "learning_rate": 4.003594521465381e-05, | |
| "loss": 0.651, | |
| "num_tokens": 60593934.0, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.3473288957159925, | |
| "grad_norm": 0.9659486932550566, | |
| "learning_rate": 3.9983166107504114e-05, | |
| "loss": 0.6528, | |
| "num_tokens": 61088175.0, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 0.3481822836661546, | |
| "grad_norm": 0.964763472219998, | |
| "learning_rate": 3.99302875794267e-05, | |
| "loss": 0.6141, | |
| "num_tokens": 61601767.0, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.34903567161631677, | |
| "grad_norm": 0.9675579189227672, | |
| "learning_rate": 3.987731005156731e-05, | |
| "loss": 0.6199, | |
| "num_tokens": 62100317.0, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 0.3498890595664789, | |
| "grad_norm": 0.9032844144807181, | |
| "learning_rate": 3.9824233945860165e-05, | |
| "loss": 0.6304, | |
| "num_tokens": 62565169.0, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.3507424475166411, | |
| "grad_norm": 0.98765077203059, | |
| "learning_rate": 3.977105968502461e-05, | |
| "loss": 0.5942, | |
| "num_tokens": 63032634.0, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.35159583546680323, | |
| "grad_norm": 1.0669138250049846, | |
| "learning_rate": 3.971778769256172e-05, | |
| "loss": 0.6312, | |
| "num_tokens": 63491002.0, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.35244922341696533, | |
| "grad_norm": 0.9971058529881554, | |
| "learning_rate": 3.966441839275095e-05, | |
| "loss": 0.6788, | |
| "num_tokens": 63993961.0, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 0.3533026113671275, | |
| "grad_norm": 0.9440611528418567, | |
| "learning_rate": 3.9610952210646746e-05, | |
| "loss": 0.6253, | |
| "num_tokens": 64465071.0, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.35415599931728964, | |
| "grad_norm": 0.9864109612249565, | |
| "learning_rate": 3.955738957207517e-05, | |
| "loss": 0.5995, | |
| "num_tokens": 64924863.0, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.3550093872674518, | |
| "grad_norm": 0.8734114010653501, | |
| "learning_rate": 3.95037309036305e-05, | |
| "loss": 0.6147, | |
| "num_tokens": 65449389.0, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.35586277521761395, | |
| "grad_norm": 0.9323878220474048, | |
| "learning_rate": 3.944997663267183e-05, | |
| "loss": 0.6499, | |
| "num_tokens": 65933344.0, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.35671616316777605, | |
| "grad_norm": 1.2029027862535284, | |
| "learning_rate": 3.939612718731968e-05, | |
| "loss": 0.6352, | |
| "num_tokens": 66383890.0, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.3575695511179382, | |
| "grad_norm": 1.062980629087297, | |
| "learning_rate": 3.934218299645256e-05, | |
| "loss": 0.6787, | |
| "num_tokens": 66852966.0, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 0.35842293906810035, | |
| "grad_norm": 1.068461942262625, | |
| "learning_rate": 3.9288144489703595e-05, | |
| "loss": 0.688, | |
| "num_tokens": 67336145.0, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.3592763270182625, | |
| "grad_norm": 1.1070164272184257, | |
| "learning_rate": 3.923401209745705e-05, | |
| "loss": 0.6366, | |
| "num_tokens": 67785161.0, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 0.36012971496842466, | |
| "grad_norm": 1.0589262427093307, | |
| "learning_rate": 3.917978625084497e-05, | |
| "loss": 0.6272, | |
| "num_tokens": 68235334.0, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.36098310291858676, | |
| "grad_norm": 0.9562551264817776, | |
| "learning_rate": 3.912546738174367e-05, | |
| "loss": 0.6758, | |
| "num_tokens": 68709817.0, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.3618364908687489, | |
| "grad_norm": 0.9656995006466078, | |
| "learning_rate": 3.907105592277035e-05, | |
| "loss": 0.5975, | |
| "num_tokens": 69137066.0, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.36268987881891107, | |
| "grad_norm": 1.194718586831029, | |
| "learning_rate": 3.901655230727964e-05, | |
| "loss": 0.622, | |
| "num_tokens": 69534328.0, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.3635432667690732, | |
| "grad_norm": 0.9072228311900601, | |
| "learning_rate": 3.896195696936012e-05, | |
| "loss": 0.6298, | |
| "num_tokens": 70009467.0, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.3643966547192354, | |
| "grad_norm": 0.979540589242795, | |
| "learning_rate": 3.890727034383092e-05, | |
| "loss": 0.6555, | |
| "num_tokens": 70495279.0, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 0.36525004266939753, | |
| "grad_norm": 1.0006563543000302, | |
| "learning_rate": 3.885249286623816e-05, | |
| "loss": 0.6323, | |
| "num_tokens": 71003407.0, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.36610343061955963, | |
| "grad_norm": 0.9867551345086583, | |
| "learning_rate": 3.879762497285162e-05, | |
| "loss": 0.6493, | |
| "num_tokens": 71504190.0, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.3669568185697218, | |
| "grad_norm": 1.0604473548798987, | |
| "learning_rate": 3.874266710066113e-05, | |
| "loss": 0.6375, | |
| "num_tokens": 71964614.0, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.36781020651988394, | |
| "grad_norm": 1.0541953933655328, | |
| "learning_rate": 3.868761968737318e-05, | |
| "loss": 0.6261, | |
| "num_tokens": 72517880.0, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 0.3686635944700461, | |
| "grad_norm": 0.9282749054885615, | |
| "learning_rate": 3.863248317140736e-05, | |
| "loss": 0.6267, | |
| "num_tokens": 72995231.0, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.36951698242020825, | |
| "grad_norm": 0.8559509162708646, | |
| "learning_rate": 3.857725799189296e-05, | |
| "loss": 0.6473, | |
| "num_tokens": 73496130.0, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 0.37037037037037035, | |
| "grad_norm": 0.9466112720742375, | |
| "learning_rate": 3.852194458866538e-05, | |
| "loss": 0.6232, | |
| "num_tokens": 73976169.0, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.3712237583205325, | |
| "grad_norm": 0.916022914049245, | |
| "learning_rate": 3.846654340226271e-05, | |
| "loss": 0.6342, | |
| "num_tokens": 74473862.0, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.37207714627069466, | |
| "grad_norm": 0.9594372023699115, | |
| "learning_rate": 3.841105487392213e-05, | |
| "loss": 0.6201, | |
| "num_tokens": 75006419.0, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.3729305342208568, | |
| "grad_norm": 1.1199332662310155, | |
| "learning_rate": 3.835547944557648e-05, | |
| "loss": 0.6321, | |
| "num_tokens": 75460453.0, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 0.37378392217101897, | |
| "grad_norm": 1.0298123419544616, | |
| "learning_rate": 3.829981755985072e-05, | |
| "loss": 0.6206, | |
| "num_tokens": 75905545.0, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.37463731012118107, | |
| "grad_norm": 1.0059371478730523, | |
| "learning_rate": 3.824406966005835e-05, | |
| "loss": 0.6316, | |
| "num_tokens": 76325746.0, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 0.3754906980713432, | |
| "grad_norm": 0.8565948465056196, | |
| "learning_rate": 3.818823619019795e-05, | |
| "loss": 0.63, | |
| "num_tokens": 76819650.0, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.3763440860215054, | |
| "grad_norm": 0.971718735543094, | |
| "learning_rate": 3.8132317594949593e-05, | |
| "loss": 0.6303, | |
| "num_tokens": 77267120.0, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.37719747397166753, | |
| "grad_norm": 0.964877825336049, | |
| "learning_rate": 3.807631431967135e-05, | |
| "loss": 0.6599, | |
| "num_tokens": 77718204.0, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.3780508619218297, | |
| "grad_norm": 0.9131715426906987, | |
| "learning_rate": 3.8020226810395706e-05, | |
| "loss": 0.6357, | |
| "num_tokens": 78201205.0, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 0.3789042498719918, | |
| "grad_norm": 0.9681729255888176, | |
| "learning_rate": 3.796405551382602e-05, | |
| "loss": 0.6362, | |
| "num_tokens": 78680163.0, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.37975763782215394, | |
| "grad_norm": 1.024261468471383, | |
| "learning_rate": 3.7907800877332945e-05, | |
| "loss": 0.6262, | |
| "num_tokens": 79147537.0, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.3806110257723161, | |
| "grad_norm": 0.8968073214866014, | |
| "learning_rate": 3.785146334895093e-05, | |
| "loss": 0.613, | |
| "num_tokens": 79600159.0, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.38146441372247825, | |
| "grad_norm": 0.948257919338274, | |
| "learning_rate": 3.779504337737456e-05, | |
| "loss": 0.6318, | |
| "num_tokens": 80042192.0, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.3823178016726404, | |
| "grad_norm": 0.9629472763837074, | |
| "learning_rate": 3.7738541411955074e-05, | |
| "loss": 0.6534, | |
| "num_tokens": 80495756.0, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.3831711896228025, | |
| "grad_norm": 0.877217572957814, | |
| "learning_rate": 3.768195790269672e-05, | |
| "loss": 0.6071, | |
| "num_tokens": 81028627.0, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 0.38402457757296465, | |
| "grad_norm": 0.8818900363189004, | |
| "learning_rate": 3.762529330025319e-05, | |
| "loss": 0.6812, | |
| "num_tokens": 81505213.0, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.3848779655231268, | |
| "grad_norm": 0.8225263179706899, | |
| "learning_rate": 3.756854805592405e-05, | |
| "loss": 0.6587, | |
| "num_tokens": 81971160.0, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 0.38573135347328896, | |
| "grad_norm": 0.9901818767360601, | |
| "learning_rate": 3.7511722621651116e-05, | |
| "loss": 0.6401, | |
| "num_tokens": 82465162.0, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.3865847414234511, | |
| "grad_norm": 1.0000087207676795, | |
| "learning_rate": 3.745481745001488e-05, | |
| "loss": 0.6431, | |
| "num_tokens": 82959331.0, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.38743812937361327, | |
| "grad_norm": 0.8801367506244648, | |
| "learning_rate": 3.7397832994230886e-05, | |
| "loss": 0.6524, | |
| "num_tokens": 83442507.0, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.38829151732377537, | |
| "grad_norm": 0.8357184360328325, | |
| "learning_rate": 3.734076970814613e-05, | |
| "loss": 0.6123, | |
| "num_tokens": 83886573.0, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.3891449052739375, | |
| "grad_norm": 0.9819493221246193, | |
| "learning_rate": 3.728362804623544e-05, | |
| "loss": 0.642, | |
| "num_tokens": 84378042.0, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.3899982932240997, | |
| "grad_norm": 1.2011907738377074, | |
| "learning_rate": 3.7226408463597885e-05, | |
| "loss": 0.6725, | |
| "num_tokens": 84863488.0, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 0.39085168117426183, | |
| "grad_norm": 0.8281172195899965, | |
| "learning_rate": 3.716911141595309e-05, | |
| "loss": 0.6141, | |
| "num_tokens": 85415205.0, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.391705069124424, | |
| "grad_norm": 0.9770574138656928, | |
| "learning_rate": 3.711173735963767e-05, | |
| "loss": 0.6286, | |
| "num_tokens": 85874153.0, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.3925584570745861, | |
| "grad_norm": 0.9169044783146749, | |
| "learning_rate": 3.705428675160154e-05, | |
| "loss": 0.6239, | |
| "num_tokens": 86342702.0, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.39341184502474824, | |
| "grad_norm": 0.8589301929082652, | |
| "learning_rate": 3.6996760049404344e-05, | |
| "loss": 0.6022, | |
| "num_tokens": 86838157.0, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 0.3942652329749104, | |
| "grad_norm": 1.0061321674465398, | |
| "learning_rate": 3.693915771121173e-05, | |
| "loss": 0.6518, | |
| "num_tokens": 87329550.0, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.39511862092507255, | |
| "grad_norm": 1.0648969069589578, | |
| "learning_rate": 3.688148019579177e-05, | |
| "loss": 0.6574, | |
| "num_tokens": 87806338.0, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 0.3959720088752347, | |
| "grad_norm": 0.9160782426371946, | |
| "learning_rate": 3.682372796251125e-05, | |
| "loss": 0.6356, | |
| "num_tokens": 88291313.0, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.3968253968253968, | |
| "grad_norm": 1.0359966560447595, | |
| "learning_rate": 3.676590147133206e-05, | |
| "loss": 0.6411, | |
| "num_tokens": 88782067.0, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.39767878477555896, | |
| "grad_norm": 1.013568304434419, | |
| "learning_rate": 3.670800118280753e-05, | |
| "loss": 0.6383, | |
| "num_tokens": 89251725.0, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.3985321727257211, | |
| "grad_norm": 0.8483728475918468, | |
| "learning_rate": 3.665002755807868e-05, | |
| "loss": 0.6106, | |
| "num_tokens": 89677245.0, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 0.39938556067588327, | |
| "grad_norm": 0.9049641479903795, | |
| "learning_rate": 3.6591981058870666e-05, | |
| "loss": 0.5812, | |
| "num_tokens": 90126369.0, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.4002389486260454, | |
| "grad_norm": 0.8707887587181531, | |
| "learning_rate": 3.653386214748902e-05, | |
| "loss": 0.6756, | |
| "num_tokens": 90629081.0, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 0.4010923365762075, | |
| "grad_norm": 0.9654943475549265, | |
| "learning_rate": 3.647567128681598e-05, | |
| "loss": 0.5985, | |
| "num_tokens": 91062791.0, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.4019457245263697, | |
| "grad_norm": 1.434236971759778, | |
| "learning_rate": 3.641740894030687e-05, | |
| "loss": 0.6202, | |
| "num_tokens": 91544042.0, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.40279911247653183, | |
| "grad_norm": 0.891703650231472, | |
| "learning_rate": 3.635907557198629e-05, | |
| "loss": 0.6246, | |
| "num_tokens": 92062591.0, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.403652500426694, | |
| "grad_norm": 0.9924195803456223, | |
| "learning_rate": 3.630067164644453e-05, | |
| "loss": 0.5949, | |
| "num_tokens": 92465083.0, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 0.40450588837685614, | |
| "grad_norm": 0.9934691293711319, | |
| "learning_rate": 3.624219762883381e-05, | |
| "loss": 0.6093, | |
| "num_tokens": 92946072.0, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.40535927632701824, | |
| "grad_norm": 0.9356359151969718, | |
| "learning_rate": 3.61836539848646e-05, | |
| "loss": 0.6722, | |
| "num_tokens": 93480913.0, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.4062126642771804, | |
| "grad_norm": 0.8475881209916387, | |
| "learning_rate": 3.6125041180801876e-05, | |
| "loss": 0.6104, | |
| "num_tokens": 93982593.0, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.40706605222734255, | |
| "grad_norm": 0.876694958055183, | |
| "learning_rate": 3.606635968346148e-05, | |
| "loss": 0.6004, | |
| "num_tokens": 94448263.0, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.4079194401775047, | |
| "grad_norm": 0.8457478713791695, | |
| "learning_rate": 3.6007609960206316e-05, | |
| "loss": 0.6372, | |
| "num_tokens": 94960061.0, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.40877282812766685, | |
| "grad_norm": 0.9818501941975467, | |
| "learning_rate": 3.5948792478942666e-05, | |
| "loss": 0.6157, | |
| "num_tokens": 95470351.0, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 0.409626216077829, | |
| "grad_norm": 0.9205272547654763, | |
| "learning_rate": 3.588990770811649e-05, | |
| "loss": 0.6256, | |
| "num_tokens": 95944505.0, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.4104796040279911, | |
| "grad_norm": 1.0164677508459452, | |
| "learning_rate": 3.583095611670965e-05, | |
| "loss": 0.6084, | |
| "num_tokens": 482939.0, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 0.41133299197815326, | |
| "grad_norm": 0.9091799199639967, | |
| "learning_rate": 3.57719381742362e-05, | |
| "loss": 0.6317, | |
| "num_tokens": 989777.0, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.4121863799283154, | |
| "grad_norm": 0.9187327421482128, | |
| "learning_rate": 3.571285435073865e-05, | |
| "loss": 0.6409, | |
| "num_tokens": 1451952.0, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.41303976787847757, | |
| "grad_norm": 0.8268322159131443, | |
| "learning_rate": 3.5653705116784174e-05, | |
| "loss": 0.6317, | |
| "num_tokens": 1967639.0, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.4138931558286397, | |
| "grad_norm": 0.8917435541473175, | |
| "learning_rate": 3.559449094346096e-05, | |
| "loss": 0.648, | |
| "num_tokens": 2432780.0, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.4147465437788018, | |
| "grad_norm": 0.823719365284321, | |
| "learning_rate": 3.5535212302374334e-05, | |
| "loss": 0.6364, | |
| "num_tokens": 2949439.0, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.415599931728964, | |
| "grad_norm": 1.0803913143852297, | |
| "learning_rate": 3.547586966564314e-05, | |
| "loss": 0.6751, | |
| "num_tokens": 3411629.0, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 0.41645331967912613, | |
| "grad_norm": 0.9749590498819914, | |
| "learning_rate": 3.5416463505895836e-05, | |
| "loss": 0.6443, | |
| "num_tokens": 3907038.0, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.4173067076292883, | |
| "grad_norm": 0.8505045179117103, | |
| "learning_rate": 3.5356994296266874e-05, | |
| "loss": 0.6228, | |
| "num_tokens": 4398651.0, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.41816009557945044, | |
| "grad_norm": 0.9912020364637751, | |
| "learning_rate": 3.5297462510392796e-05, | |
| "loss": 0.6325, | |
| "num_tokens": 4847281.0, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.41901348352961254, | |
| "grad_norm": 0.9095743217389802, | |
| "learning_rate": 3.5237868622408574e-05, | |
| "loss": 0.6658, | |
| "num_tokens": 5318809.0, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 0.4198668714797747, | |
| "grad_norm": 0.8293429203316864, | |
| "learning_rate": 3.5178213106943754e-05, | |
| "loss": 0.6082, | |
| "num_tokens": 5753540.0, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.42072025942993685, | |
| "grad_norm": 1.0446630574566, | |
| "learning_rate": 3.5118496439118734e-05, | |
| "loss": 0.6353, | |
| "num_tokens": 6226147.0, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 0.421573647380099, | |
| "grad_norm": 0.8769207386177866, | |
| "learning_rate": 3.505871909454093e-05, | |
| "loss": 0.6529, | |
| "num_tokens": 6725686.0, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.42242703533026116, | |
| "grad_norm": 0.9508154637486791, | |
| "learning_rate": 3.4998881549301025e-05, | |
| "loss": 0.6222, | |
| "num_tokens": 7265229.0, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.42328042328042326, | |
| "grad_norm": 0.9566300180169934, | |
| "learning_rate": 3.493898427996917e-05, | |
| "loss": 0.6558, | |
| "num_tokens": 7683984.0, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.4241338112305854, | |
| "grad_norm": 1.0927270546087369, | |
| "learning_rate": 3.487902776359118e-05, | |
| "loss": 0.6249, | |
| "num_tokens": 8156250.0, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 0.42498719918074757, | |
| "grad_norm": 0.9928807292515285, | |
| "learning_rate": 3.481901247768471e-05, | |
| "loss": 0.5972, | |
| "num_tokens": 8628772.0, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.4258405871309097, | |
| "grad_norm": 0.9449942264532448, | |
| "learning_rate": 3.4758938900235514e-05, | |
| "loss": 0.6578, | |
| "num_tokens": 9120921.0, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 0.4266939750810719, | |
| "grad_norm": 1.0607940692442015, | |
| "learning_rate": 3.46988075096936e-05, | |
| "loss": 0.6304, | |
| "num_tokens": 9560658.0, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.427547363031234, | |
| "grad_norm": 0.9152490107525716, | |
| "learning_rate": 3.463861878496939e-05, | |
| "loss": 0.6212, | |
| "num_tokens": 10061014.0, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.42840075098139613, | |
| "grad_norm": 1.0829777365198388, | |
| "learning_rate": 3.457837320542998e-05, | |
| "loss": 0.6402, | |
| "num_tokens": 10515660.0, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.4292541389315583, | |
| "grad_norm": 0.9597202991590822, | |
| "learning_rate": 3.451807125089525e-05, | |
| "loss": 0.6428, | |
| "num_tokens": 10976284.0, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 0.43010752688172044, | |
| "grad_norm": 1.0855106871299507, | |
| "learning_rate": 3.445771340163409e-05, | |
| "loss": 0.6245, | |
| "num_tokens": 11403774.0, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.4309609148318826, | |
| "grad_norm": 1.130452291984405, | |
| "learning_rate": 3.4397300138360565e-05, | |
| "loss": 0.6442, | |
| "num_tokens": 11874868.0, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 0.4318143027820447, | |
| "grad_norm": 0.8901582972402161, | |
| "learning_rate": 3.433683194223004e-05, | |
| "loss": 0.6598, | |
| "num_tokens": 12323802.0, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.43266769073220684, | |
| "grad_norm": 1.0163374881517988, | |
| "learning_rate": 3.427630929483543e-05, | |
| "loss": 0.6568, | |
| "num_tokens": 12821783.0, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.433521078682369, | |
| "grad_norm": 1.2078170030289623, | |
| "learning_rate": 3.42157326782033e-05, | |
| "loss": 0.6282, | |
| "num_tokens": 13282255.0, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.43437446663253115, | |
| "grad_norm": 0.9781762323975279, | |
| "learning_rate": 3.415510257479008e-05, | |
| "loss": 0.6611, | |
| "num_tokens": 13799399.0, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 0.4352278545826933, | |
| "grad_norm": 1.0308563805242064, | |
| "learning_rate": 3.409441946747813e-05, | |
| "loss": 0.6139, | |
| "num_tokens": 14301887.0, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.43608124253285546, | |
| "grad_norm": 0.9289960246789825, | |
| "learning_rate": 3.403368383957199e-05, | |
| "loss": 0.5933, | |
| "num_tokens": 14745656.0, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 0.43693463048301756, | |
| "grad_norm": 0.8577337044308939, | |
| "learning_rate": 3.39728961747945e-05, | |
| "loss": 0.5892, | |
| "num_tokens": 15257500.0, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.4377880184331797, | |
| "grad_norm": 0.9491205699096495, | |
| "learning_rate": 3.391205695728295e-05, | |
| "loss": 0.643, | |
| "num_tokens": 15815630.0, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.43864140638334187, | |
| "grad_norm": 0.819417628826071, | |
| "learning_rate": 3.3851166671585176e-05, | |
| "loss": 0.633, | |
| "num_tokens": 16325393.0, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.439494794333504, | |
| "grad_norm": 0.9084966436728551, | |
| "learning_rate": 3.3790225802655796e-05, | |
| "loss": 0.7004, | |
| "num_tokens": 16832508.0, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 0.4403481822836662, | |
| "grad_norm": 0.9355719385094893, | |
| "learning_rate": 3.3729234835852236e-05, | |
| "loss": 0.6495, | |
| "num_tokens": 17265900.0, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.4412015702338283, | |
| "grad_norm": 0.9198866510461836, | |
| "learning_rate": 3.3668194256930966e-05, | |
| "loss": 0.6203, | |
| "num_tokens": 17758587.0, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 0.44205495818399043, | |
| "grad_norm": 0.9864299064702182, | |
| "learning_rate": 3.360710455204357e-05, | |
| "loss": 0.5901, | |
| "num_tokens": 18162019.0, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.4429083461341526, | |
| "grad_norm": 0.8533788630772795, | |
| "learning_rate": 3.354596620773288e-05, | |
| "loss": 0.6233, | |
| "num_tokens": 18623614.0, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.44376173408431474, | |
| "grad_norm": 0.897565045339347, | |
| "learning_rate": 3.348477971092914e-05, | |
| "loss": 0.6353, | |
| "num_tokens": 19040608.0, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.4446151220344769, | |
| "grad_norm": 0.8517543561030608, | |
| "learning_rate": 3.3423545548946074e-05, | |
| "loss": 0.6051, | |
| "num_tokens": 19548853.0, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 0.445468509984639, | |
| "grad_norm": 0.9520584025010822, | |
| "learning_rate": 3.336226420947704e-05, | |
| "loss": 0.648, | |
| "num_tokens": 19984697.0, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.44632189793480115, | |
| "grad_norm": 0.8190289095396062, | |
| "learning_rate": 3.330093618059114e-05, | |
| "loss": 0.6475, | |
| "num_tokens": 20478274.0, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 0.4471752858849633, | |
| "grad_norm": 0.9202078769277446, | |
| "learning_rate": 3.323956195072932e-05, | |
| "loss": 0.6154, | |
| "num_tokens": 20961369.0, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.44802867383512546, | |
| "grad_norm": 0.9572401045682531, | |
| "learning_rate": 3.3178142008700494e-05, | |
| "loss": 0.6179, | |
| "num_tokens": 21399865.0, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.4488820617852876, | |
| "grad_norm": 1.0143431696499436, | |
| "learning_rate": 3.311667684367765e-05, | |
| "loss": 0.6418, | |
| "num_tokens": 21883446.0, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.4497354497354497, | |
| "grad_norm": 0.9286975643218015, | |
| "learning_rate": 3.3055166945193944e-05, | |
| "loss": 0.6044, | |
| "num_tokens": 22380147.0, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 0.45058883768561186, | |
| "grad_norm": 0.8826200601623698, | |
| "learning_rate": 3.299361280313881e-05, | |
| "loss": 0.5855, | |
| "num_tokens": 22896778.0, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.451442225635774, | |
| "grad_norm": 0.9121057769453851, | |
| "learning_rate": 3.293201490775406e-05, | |
| "loss": 0.6272, | |
| "num_tokens": 23387347.0, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 0.4522956135859362, | |
| "grad_norm": 1.021199041775351, | |
| "learning_rate": 3.2870373749629954e-05, | |
| "loss": 0.6673, | |
| "num_tokens": 23866169.0, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.45314900153609833, | |
| "grad_norm": 0.7934600813626838, | |
| "learning_rate": 3.280868981970134e-05, | |
| "loss": 0.6036, | |
| "num_tokens": 24294707.0, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.4540023894862604, | |
| "grad_norm": 1.0668312744043429, | |
| "learning_rate": 3.2746963609243716e-05, | |
| "loss": 0.6176, | |
| "num_tokens": 24786772.0, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.4548557774364226, | |
| "grad_norm": 0.8327511404831184, | |
| "learning_rate": 3.26851956098693e-05, | |
| "loss": 0.6254, | |
| "num_tokens": 25263421.0, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 0.45570916538658474, | |
| "grad_norm": 0.8948559186273163, | |
| "learning_rate": 3.2623386313523145e-05, | |
| "loss": 0.6201, | |
| "num_tokens": 25723061.0, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.4565625533367469, | |
| "grad_norm": 0.8282682651999119, | |
| "learning_rate": 3.256153621247921e-05, | |
| "loss": 0.6495, | |
| "num_tokens": 26259692.0, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 0.45741594128690904, | |
| "grad_norm": 0.9165264208822277, | |
| "learning_rate": 3.249964579933644e-05, | |
| "loss": 0.6151, | |
| "num_tokens": 26707422.0, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.4582693292370712, | |
| "grad_norm": 0.9824352883104055, | |
| "learning_rate": 3.2437715567014836e-05, | |
| "loss": 0.5971, | |
| "num_tokens": 27173821.0, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.4591227171872333, | |
| "grad_norm": 1.0074666109427157, | |
| "learning_rate": 3.237574600875154e-05, | |
| "loss": 0.615, | |
| "num_tokens": 27631439.0, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.45997610513739545, | |
| "grad_norm": 0.858111222643122, | |
| "learning_rate": 3.231373761809689e-05, | |
| "loss": 0.6305, | |
| "num_tokens": 28197458.0, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 0.4608294930875576, | |
| "grad_norm": 0.9212213548939474, | |
| "learning_rate": 3.2251690888910535e-05, | |
| "loss": 0.5823, | |
| "num_tokens": 28686498.0, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.46168288103771976, | |
| "grad_norm": 1.0348751550407185, | |
| "learning_rate": 3.218960631535742e-05, | |
| "loss": 0.622, | |
| "num_tokens": 29144972.0, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 0.4625362689878819, | |
| "grad_norm": 1.0248437192672697, | |
| "learning_rate": 3.212748439190392e-05, | |
| "loss": 0.6212, | |
| "num_tokens": 29666782.0, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.463389656938044, | |
| "grad_norm": 0.84228312481392, | |
| "learning_rate": 3.20653256133139e-05, | |
| "loss": 0.6194, | |
| "num_tokens": 30192754.0, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 0.46424304488820617, | |
| "grad_norm": 0.818769118662692, | |
| "learning_rate": 3.200313047464471e-05, | |
| "loss": 0.6678, | |
| "num_tokens": 30745942.0, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.4650964328383683, | |
| "grad_norm": 1.0823879075880105, | |
| "learning_rate": 3.194089947124333e-05, | |
| "loss": 0.5908, | |
| "num_tokens": 31152396.0, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 0.4659498207885305, | |
| "grad_norm": 0.8659299087412192, | |
| "learning_rate": 3.1878633098742344e-05, | |
| "loss": 0.6113, | |
| "num_tokens": 31689199.0, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.46680320873869263, | |
| "grad_norm": 0.8851915995943411, | |
| "learning_rate": 3.1816331853056063e-05, | |
| "loss": 0.6078, | |
| "num_tokens": 32128480.0, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 0.46765659668885473, | |
| "grad_norm": 0.7702876245179221, | |
| "learning_rate": 3.175399623037652e-05, | |
| "loss": 0.6129, | |
| "num_tokens": 32674508.0, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.4685099846390169, | |
| "grad_norm": 0.8836130988239285, | |
| "learning_rate": 3.169162672716954e-05, | |
| "loss": 0.5919, | |
| "num_tokens": 33081921.0, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 0.46936337258917904, | |
| "grad_norm": 0.8231620704033741, | |
| "learning_rate": 3.162922384017081e-05, | |
| "loss": 0.5991, | |
| "num_tokens": 33556906.0, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.4702167605393412, | |
| "grad_norm": 0.952339994710836, | |
| "learning_rate": 3.156678806638186e-05, | |
| "loss": 0.6042, | |
| "num_tokens": 34031283.0, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 0.47107014848950335, | |
| "grad_norm": 0.9622665131906707, | |
| "learning_rate": 3.150431990306617e-05, | |
| "loss": 0.6557, | |
| "num_tokens": 34577418.0, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.47192353643966545, | |
| "grad_norm": 0.8311985943600112, | |
| "learning_rate": 3.1441819847745186e-05, | |
| "loss": 0.6042, | |
| "num_tokens": 35007918.0, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 0.4727769243898276, | |
| "grad_norm": 0.947220685474545, | |
| "learning_rate": 3.137928839819434e-05, | |
| "loss": 0.638, | |
| "num_tokens": 35516081.0, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.47363031233998976, | |
| "grad_norm": 0.9708185654838827, | |
| "learning_rate": 3.131672605243911e-05, | |
| "loss": 0.6151, | |
| "num_tokens": 35972013.0, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.4744837002901519, | |
| "grad_norm": 0.9665397926545222, | |
| "learning_rate": 3.125413330875104e-05, | |
| "loss": 0.6332, | |
| "num_tokens": 36487961.0, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.47533708824031407, | |
| "grad_norm": 0.825937896752512, | |
| "learning_rate": 3.119151066564378e-05, | |
| "loss": 0.6508, | |
| "num_tokens": 36954084.0, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 0.47619047619047616, | |
| "grad_norm": 0.8814205249866639, | |
| "learning_rate": 3.1128858621869084e-05, | |
| "loss": 0.5928, | |
| "num_tokens": 37436748.0, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.4770438641406383, | |
| "grad_norm": 0.8560968867180144, | |
| "learning_rate": 3.106617767641291e-05, | |
| "loss": 0.6041, | |
| "num_tokens": 37908735.0, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 0.4778972520908005, | |
| "grad_norm": 0.8960390256733356, | |
| "learning_rate": 3.100346832849137e-05, | |
| "loss": 0.633, | |
| "num_tokens": 38388303.0, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.4787506400409626, | |
| "grad_norm": 0.8289004343409846, | |
| "learning_rate": 3.0940731077546806e-05, | |
| "loss": 0.6412, | |
| "num_tokens": 38876259.0, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 0.4796040279911248, | |
| "grad_norm": 0.9023913726126142, | |
| "learning_rate": 3.087796642324376e-05, | |
| "loss": 0.6347, | |
| "num_tokens": 39361858.0, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.48045741594128694, | |
| "grad_norm": 0.8817899043399531, | |
| "learning_rate": 3.0815174865465046e-05, | |
| "loss": 0.6335, | |
| "num_tokens": 39877714.0, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 0.48131080389144903, | |
| "grad_norm": 0.9020909342984744, | |
| "learning_rate": 3.075235690430775e-05, | |
| "loss": 0.6206, | |
| "num_tokens": 40323502.0, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.4821641918416112, | |
| "grad_norm": 0.9191982487626535, | |
| "learning_rate": 3.0689513040079235e-05, | |
| "loss": 0.6072, | |
| "num_tokens": 40800904.0, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 0.48301757979177334, | |
| "grad_norm": 0.9142822978216951, | |
| "learning_rate": 3.062664377329317e-05, | |
| "loss": 0.6488, | |
| "num_tokens": 41310389.0, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.4838709677419355, | |
| "grad_norm": 0.9574374642953535, | |
| "learning_rate": 3.0563749604665556e-05, | |
| "loss": 0.6544, | |
| "num_tokens": 41801610.0, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 0.48472435569209765, | |
| "grad_norm": 1.021826726972745, | |
| "learning_rate": 3.0500831035110677e-05, | |
| "loss": 0.6066, | |
| "num_tokens": 42263578.0, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.48557774364225975, | |
| "grad_norm": 0.8711698710249252, | |
| "learning_rate": 3.0437888565737215e-05, | |
| "loss": 0.6565, | |
| "num_tokens": 42794652.0, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 0.4864311315924219, | |
| "grad_norm": 0.9515362101152438, | |
| "learning_rate": 3.0374922697844167e-05, | |
| "loss": 0.5999, | |
| "num_tokens": 43295831.0, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.48728451954258406, | |
| "grad_norm": 0.7767880334427985, | |
| "learning_rate": 3.0311933932916874e-05, | |
| "loss": 0.6604, | |
| "num_tokens": 43762546.0, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 0.4881379074927462, | |
| "grad_norm": 0.999247983508965, | |
| "learning_rate": 3.0248922772623066e-05, | |
| "loss": 0.635, | |
| "num_tokens": 44271674.0, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.48899129544290837, | |
| "grad_norm": 0.9808245624992524, | |
| "learning_rate": 3.0185889718808813e-05, | |
| "loss": 0.6324, | |
| "num_tokens": 44767111.0, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 0.48984468339307047, | |
| "grad_norm": 0.9794159238802201, | |
| "learning_rate": 3.012283527349458e-05, | |
| "loss": 0.592, | |
| "num_tokens": 45220967.0, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.4906980713432326, | |
| "grad_norm": 0.9730204959795733, | |
| "learning_rate": 3.0059759938871194e-05, | |
| "loss": 0.6593, | |
| "num_tokens": 45692929.0, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 0.4915514592933948, | |
| "grad_norm": 0.999538493061937, | |
| "learning_rate": 2.9996664217295832e-05, | |
| "loss": 0.6325, | |
| "num_tokens": 46141558.0, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.49240484724355693, | |
| "grad_norm": 0.9461090327652082, | |
| "learning_rate": 2.9933548611288064e-05, | |
| "loss": 0.5998, | |
| "num_tokens": 46591176.0, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 0.4932582351937191, | |
| "grad_norm": 0.990050853768675, | |
| "learning_rate": 2.987041362352581e-05, | |
| "loss": 0.5853, | |
| "num_tokens": 47011064.0, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.4941116231438812, | |
| "grad_norm": 0.8903704890612268, | |
| "learning_rate": 2.9807259756841383e-05, | |
| "loss": 0.5927, | |
| "num_tokens": 47456082.0, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 0.49496501109404334, | |
| "grad_norm": 0.8895395920757792, | |
| "learning_rate": 2.974408751421743e-05, | |
| "loss": 0.6047, | |
| "num_tokens": 47977344.0, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.4958183990442055, | |
| "grad_norm": 1.0762715239860545, | |
| "learning_rate": 2.9680897398782946e-05, | |
| "loss": 0.6443, | |
| "num_tokens": 48413819.0, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 0.49667178699436765, | |
| "grad_norm": 1.0093778915297285, | |
| "learning_rate": 2.9617689913809304e-05, | |
| "loss": 0.6147, | |
| "num_tokens": 48898798.0, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.4975251749445298, | |
| "grad_norm": 0.9519508557055824, | |
| "learning_rate": 2.955446556270618e-05, | |
| "loss": 0.6226, | |
| "num_tokens": 49346202.0, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 0.4983785628946919, | |
| "grad_norm": 0.945004577120924, | |
| "learning_rate": 2.9491224849017602e-05, | |
| "loss": 0.6104, | |
| "num_tokens": 49793997.0, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.49923195084485406, | |
| "grad_norm": 0.9129913648493336, | |
| "learning_rate": 2.94279682764179e-05, | |
| "loss": 0.605, | |
| "num_tokens": 50284460.0, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.5000853387950163, | |
| "grad_norm": 0.8409386583627643, | |
| "learning_rate": 2.9364696348707726e-05, | |
| "loss": 0.6135, | |
| "num_tokens": 50836014.0, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.5009387267451784, | |
| "grad_norm": 0.990670854143676, | |
| "learning_rate": 2.930140956981002e-05, | |
| "loss": 0.5999, | |
| "num_tokens": 51250570.0, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 0.5017921146953405, | |
| "grad_norm": 0.9459574916127076, | |
| "learning_rate": 2.9238108443765988e-05, | |
| "loss": 0.5934, | |
| "num_tokens": 51745806.0, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.5026455026455027, | |
| "grad_norm": 0.8678525092738497, | |
| "learning_rate": 2.9174793474731133e-05, | |
| "loss": 0.6233, | |
| "num_tokens": 52250600.0, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 0.5034988905956648, | |
| "grad_norm": 0.9384958904208902, | |
| "learning_rate": 2.911146516697118e-05, | |
| "loss": 0.6299, | |
| "num_tokens": 52797693.0, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.504352278545827, | |
| "grad_norm": 0.9868110919179371, | |
| "learning_rate": 2.904812402485811e-05, | |
| "loss": 0.6446, | |
| "num_tokens": 53261423.0, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 0.5052056664959891, | |
| "grad_norm": 1.0091887647980669, | |
| "learning_rate": 2.8984770552866108e-05, | |
| "loss": 0.5639, | |
| "num_tokens": 53751059.0, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.5060590544461512, | |
| "grad_norm": 0.8342160700122503, | |
| "learning_rate": 2.8921405255567578e-05, | |
| "loss": 0.5679, | |
| "num_tokens": 54223040.0, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 0.5069124423963134, | |
| "grad_norm": 0.8986532584280877, | |
| "learning_rate": 2.8858028637629063e-05, | |
| "loss": 0.607, | |
| "num_tokens": 54673453.0, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.5077658303464755, | |
| "grad_norm": 0.8925992400801118, | |
| "learning_rate": 2.8794641203807314e-05, | |
| "loss": 0.6218, | |
| "num_tokens": 55202493.0, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 0.5086192182966377, | |
| "grad_norm": 0.7890714280668591, | |
| "learning_rate": 2.873124345894521e-05, | |
| "loss": 0.638, | |
| "num_tokens": 55722817.0, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.5094726062467998, | |
| "grad_norm": 0.9064838230559494, | |
| "learning_rate": 2.8667835907967748e-05, | |
| "loss": 0.6314, | |
| "num_tokens": 56227280.0, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 0.5103259941969619, | |
| "grad_norm": 0.998743125755284, | |
| "learning_rate": 2.8604419055878017e-05, | |
| "loss": 0.5891, | |
| "num_tokens": 56669120.0, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.5111793821471241, | |
| "grad_norm": 0.940747148658803, | |
| "learning_rate": 2.854099340775319e-05, | |
| "loss": 0.6301, | |
| "num_tokens": 57122362.0, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 0.5120327700972862, | |
| "grad_norm": 0.9437400600614468, | |
| "learning_rate": 2.8477559468740506e-05, | |
| "loss": 0.6555, | |
| "num_tokens": 57661431.0, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.5128861580474484, | |
| "grad_norm": 0.8394567709397782, | |
| "learning_rate": 2.8414117744053225e-05, | |
| "loss": 0.6056, | |
| "num_tokens": 58131273.0, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 0.5137395459976105, | |
| "grad_norm": 0.9489257789311477, | |
| "learning_rate": 2.8350668738966612e-05, | |
| "loss": 0.5858, | |
| "num_tokens": 58585451.0, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.5145929339477726, | |
| "grad_norm": 0.8727065090879019, | |
| "learning_rate": 2.8287212958813926e-05, | |
| "loss": 0.5646, | |
| "num_tokens": 59037920.0, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 0.5154463218979348, | |
| "grad_norm": 0.9682582828974684, | |
| "learning_rate": 2.8223750908982378e-05, | |
| "loss": 0.6208, | |
| "num_tokens": 59548752.0, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.5162997098480969, | |
| "grad_norm": 0.9339447380894269, | |
| "learning_rate": 2.8160283094909105e-05, | |
| "loss": 0.6118, | |
| "num_tokens": 60023348.0, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 0.5171530977982591, | |
| "grad_norm": 0.9091766179100768, | |
| "learning_rate": 2.8096810022077184e-05, | |
| "loss": 0.6063, | |
| "num_tokens": 60478370.0, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.5180064857484212, | |
| "grad_norm": 0.8832027277397922, | |
| "learning_rate": 2.8033332196011548e-05, | |
| "loss": 0.6184, | |
| "num_tokens": 60896095.0, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 0.5188598736985833, | |
| "grad_norm": 1.0271689932750887, | |
| "learning_rate": 2.7969850122274977e-05, | |
| "loss": 0.6155, | |
| "num_tokens": 61347251.0, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.5197132616487455, | |
| "grad_norm": 0.9743044480912783, | |
| "learning_rate": 2.7906364306464116e-05, | |
| "loss": 0.6648, | |
| "num_tokens": 61804678.0, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 0.5205666495989076, | |
| "grad_norm": 0.9109542005139061, | |
| "learning_rate": 2.784287525420538e-05, | |
| "loss": 0.6142, | |
| "num_tokens": 62331454.0, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.5214200375490698, | |
| "grad_norm": 0.8849038752588272, | |
| "learning_rate": 2.777938347115098e-05, | |
| "loss": 0.5899, | |
| "num_tokens": 62817171.0, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 0.522273425499232, | |
| "grad_norm": 0.8074591324273821, | |
| "learning_rate": 2.771588946297488e-05, | |
| "loss": 0.6066, | |
| "num_tokens": 63324176.0, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.523126813449394, | |
| "grad_norm": 0.8686242811447925, | |
| "learning_rate": 2.7652393735368747e-05, | |
| "loss": 0.61, | |
| "num_tokens": 63793104.0, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 0.5239802013995563, | |
| "grad_norm": 0.9096641571043528, | |
| "learning_rate": 2.758889679403795e-05, | |
| "loss": 0.6088, | |
| "num_tokens": 64198275.0, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.5248335893497184, | |
| "grad_norm": 0.9775542616996967, | |
| "learning_rate": 2.7525399144697534e-05, | |
| "loss": 0.6071, | |
| "num_tokens": 64704063.0, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.5256869772998806, | |
| "grad_norm": 0.94132553062766, | |
| "learning_rate": 2.746190129306816e-05, | |
| "loss": 0.605, | |
| "num_tokens": 65171476.0, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.5265403652500427, | |
| "grad_norm": 0.9605299925817627, | |
| "learning_rate": 2.739840374487214e-05, | |
| "loss": 0.6351, | |
| "num_tokens": 65672579.0, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 0.5273937532002048, | |
| "grad_norm": 0.8545901170879364, | |
| "learning_rate": 2.733490700582932e-05, | |
| "loss": 0.6194, | |
| "num_tokens": 66139328.0, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.528247141150367, | |
| "grad_norm": 0.8486195663590458, | |
| "learning_rate": 2.7271411581653145e-05, | |
| "loss": 0.5836, | |
| "num_tokens": 66594619.0, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 0.5291005291005291, | |
| "grad_norm": 0.9274528003638315, | |
| "learning_rate": 2.720791797804656e-05, | |
| "loss": 0.6368, | |
| "num_tokens": 67083533.0, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.5299539170506913, | |
| "grad_norm": 0.8969308802413379, | |
| "learning_rate": 2.7144426700698012e-05, | |
| "loss": 0.6316, | |
| "num_tokens": 67618627.0, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 0.5308073050008534, | |
| "grad_norm": 0.8416577756233091, | |
| "learning_rate": 2.708093825527745e-05, | |
| "loss": 0.6116, | |
| "num_tokens": 68115381.0, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.5316606929510155, | |
| "grad_norm": 0.8131996556741273, | |
| "learning_rate": 2.7017453147432225e-05, | |
| "loss": 0.5828, | |
| "num_tokens": 68538403.0, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 0.5325140809011777, | |
| "grad_norm": 1.0454639098669911, | |
| "learning_rate": 2.6953971882783142e-05, | |
| "loss": 0.6265, | |
| "num_tokens": 68967408.0, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.5333674688513398, | |
| "grad_norm": 0.9549895357737828, | |
| "learning_rate": 2.689049496692037e-05, | |
| "loss": 0.6101, | |
| "num_tokens": 69472052.0, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 0.534220856801502, | |
| "grad_norm": 0.8308698422547768, | |
| "learning_rate": 2.6827022905399456e-05, | |
| "loss": 0.6269, | |
| "num_tokens": 69964036.0, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.5350742447516641, | |
| "grad_norm": 0.9102679595896443, | |
| "learning_rate": 2.676355620373731e-05, | |
| "loss": 0.6088, | |
| "num_tokens": 70445245.0, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 0.5359276327018262, | |
| "grad_norm": 0.9313136439436694, | |
| "learning_rate": 2.67000953674081e-05, | |
| "loss": 0.5689, | |
| "num_tokens": 70900222.0, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.5367810206519884, | |
| "grad_norm": 0.9388407450010738, | |
| "learning_rate": 2.663664090183932e-05, | |
| "loss": 0.6126, | |
| "num_tokens": 71305878.0, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 0.5376344086021505, | |
| "grad_norm": 0.914581426947238, | |
| "learning_rate": 2.657319331240771e-05, | |
| "loss": 0.5944, | |
| "num_tokens": 71729509.0, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.5384877965523127, | |
| "grad_norm": 0.83701319276453, | |
| "learning_rate": 2.650975310443525e-05, | |
| "loss": 0.5664, | |
| "num_tokens": 72238073.0, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 0.5393411845024748, | |
| "grad_norm": 0.9171217186457761, | |
| "learning_rate": 2.644632078318513e-05, | |
| "loss": 0.5879, | |
| "num_tokens": 72677498.0, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.540194572452637, | |
| "grad_norm": 0.8098622964637798, | |
| "learning_rate": 2.6382896853857736e-05, | |
| "loss": 0.6153, | |
| "num_tokens": 73179076.0, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 0.5410479604027991, | |
| "grad_norm": 0.9000788802277151, | |
| "learning_rate": 2.63194818215866e-05, | |
| "loss": 0.6381, | |
| "num_tokens": 73703391.0, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.5419013483529612, | |
| "grad_norm": 0.9158524475261981, | |
| "learning_rate": 2.625607619143439e-05, | |
| "loss": 0.6276, | |
| "num_tokens": 74152776.0, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 0.5427547363031234, | |
| "grad_norm": 0.8909922680173544, | |
| "learning_rate": 2.619268046838893e-05, | |
| "loss": 0.5935, | |
| "num_tokens": 74582980.0, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.5436081242532855, | |
| "grad_norm": 0.9278472564551875, | |
| "learning_rate": 2.61292951573591e-05, | |
| "loss": 0.6297, | |
| "num_tokens": 75055808.0, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 0.5444615122034477, | |
| "grad_norm": 0.8845869821677456, | |
| "learning_rate": 2.606592076317087e-05, | |
| "loss": 0.6279, | |
| "num_tokens": 75509960.0, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.5453149001536098, | |
| "grad_norm": 0.7587670201893357, | |
| "learning_rate": 2.6002557790563276e-05, | |
| "loss": 0.6239, | |
| "num_tokens": 76023431.0, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 0.5461682881037719, | |
| "grad_norm": 0.9336917428778075, | |
| "learning_rate": 2.5939206744184354e-05, | |
| "loss": 0.5938, | |
| "num_tokens": 76420106.0, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.5470216760539341, | |
| "grad_norm": 0.8806170933713402, | |
| "learning_rate": 2.5875868128587177e-05, | |
| "loss": 0.6088, | |
| "num_tokens": 76908186.0, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 0.5478750640040962, | |
| "grad_norm": 1.0263401501278369, | |
| "learning_rate": 2.5812542448225836e-05, | |
| "loss": 0.6792, | |
| "num_tokens": 77391407.0, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.5487284519542585, | |
| "grad_norm": 0.8644207041767296, | |
| "learning_rate": 2.574923020745135e-05, | |
| "loss": 0.6314, | |
| "num_tokens": 77845333.0, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 0.5495818399044206, | |
| "grad_norm": 0.8712907469129566, | |
| "learning_rate": 2.5685931910507756e-05, | |
| "loss": 0.5854, | |
| "num_tokens": 78314883.0, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.5504352278545827, | |
| "grad_norm": 0.7917989133845237, | |
| "learning_rate": 2.562264806152798e-05, | |
| "loss": 0.6453, | |
| "num_tokens": 78792697.0, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.5512886158047449, | |
| "grad_norm": 1.2698451028714428, | |
| "learning_rate": 2.5559379164529916e-05, | |
| "loss": 0.6055, | |
| "num_tokens": 79268449.0, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.552142003754907, | |
| "grad_norm": 0.8698617334948361, | |
| "learning_rate": 2.5496125723412378e-05, | |
| "loss": 0.5981, | |
| "num_tokens": 79777685.0, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 0.5529953917050692, | |
| "grad_norm": 1.1367372216369092, | |
| "learning_rate": 2.5432888241951047e-05, | |
| "loss": 0.6072, | |
| "num_tokens": 80241521.0, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.5538487796552313, | |
| "grad_norm": 1.1000070125764803, | |
| "learning_rate": 2.5369667223794546e-05, | |
| "loss": 0.5812, | |
| "num_tokens": 80731872.0, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 0.5547021676053934, | |
| "grad_norm": 0.9833667568222638, | |
| "learning_rate": 2.5306463172460327e-05, | |
| "loss": 0.5954, | |
| "num_tokens": 81185238.0, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.5555555555555556, | |
| "grad_norm": 0.894923840538592, | |
| "learning_rate": 2.524327659133075e-05, | |
| "loss": 0.6241, | |
| "num_tokens": 81670790.0, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 0.5564089435057177, | |
| "grad_norm": 0.8738246405469254, | |
| "learning_rate": 2.5180107983649e-05, | |
| "loss": 0.585, | |
| "num_tokens": 82146300.0, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.5572623314558799, | |
| "grad_norm": 0.9163256478577108, | |
| "learning_rate": 2.5116957852515144e-05, | |
| "loss": 0.6172, | |
| "num_tokens": 82649931.0, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 0.558115719406042, | |
| "grad_norm": 0.9132249310168558, | |
| "learning_rate": 2.50538267008821e-05, | |
| "loss": 0.6112, | |
| "num_tokens": 83046925.0, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.5589691073562041, | |
| "grad_norm": 1.0076436822777552, | |
| "learning_rate": 2.4990715031551576e-05, | |
| "loss": 0.5812, | |
| "num_tokens": 83483265.0, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 0.5598224953063663, | |
| "grad_norm": 0.8149595056663635, | |
| "learning_rate": 2.4927623347170187e-05, | |
| "loss": 0.6116, | |
| "num_tokens": 84015002.0, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.5606758832565284, | |
| "grad_norm": 0.8659160058321907, | |
| "learning_rate": 2.4864552150225313e-05, | |
| "loss": 0.601, | |
| "num_tokens": 84530159.0, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 0.5615292712066906, | |
| "grad_norm": 1.6488676893470589, | |
| "learning_rate": 2.4801501943041207e-05, | |
| "loss": 0.6287, | |
| "num_tokens": 84976777.0, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.5623826591568527, | |
| "grad_norm": 0.903284657679167, | |
| "learning_rate": 2.473847322777494e-05, | |
| "loss": 0.5746, | |
| "num_tokens": 85475390.0, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 0.5632360471070148, | |
| "grad_norm": 0.9168478565461777, | |
| "learning_rate": 2.467546650641239e-05, | |
| "loss": 0.6107, | |
| "num_tokens": 85979990.0, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.564089435057177, | |
| "grad_norm": 0.9421153163859376, | |
| "learning_rate": 2.461248228076431e-05, | |
| "loss": 0.6217, | |
| "num_tokens": 86427182.0, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 0.5649428230073391, | |
| "grad_norm": 0.9418335674217378, | |
| "learning_rate": 2.454952105246225e-05, | |
| "loss": 0.6229, | |
| "num_tokens": 86974407.0, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.5657962109575013, | |
| "grad_norm": 0.9616597438867678, | |
| "learning_rate": 2.4486583322954615e-05, | |
| "loss": 0.5953, | |
| "num_tokens": 87406506.0, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 0.5666495989076634, | |
| "grad_norm": 0.9001557558355506, | |
| "learning_rate": 2.4423669593502674e-05, | |
| "loss": 0.6028, | |
| "num_tokens": 87814661.0, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.5675029868578255, | |
| "grad_norm": 0.9282668273832819, | |
| "learning_rate": 2.43607803651765e-05, | |
| "loss": 0.6505, | |
| "num_tokens": 88286707.0, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 0.5683563748079877, | |
| "grad_norm": 0.8061451148141976, | |
| "learning_rate": 2.429791613885109e-05, | |
| "loss": 0.5717, | |
| "num_tokens": 88814762.0, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.5692097627581498, | |
| "grad_norm": 0.8372348822992951, | |
| "learning_rate": 2.4235077415202267e-05, | |
| "loss": 0.6131, | |
| "num_tokens": 89333477.0, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 0.570063150708312, | |
| "grad_norm": 1.074881795157168, | |
| "learning_rate": 2.4172264694702766e-05, | |
| "loss": 0.5847, | |
| "num_tokens": 89794835.0, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.5709165386584741, | |
| "grad_norm": 1.0051952346241788, | |
| "learning_rate": 2.4109478477618226e-05, | |
| "loss": 0.6132, | |
| "num_tokens": 90209243.0, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 0.5717699266086362, | |
| "grad_norm": 0.9160834252199135, | |
| "learning_rate": 2.404671926400317e-05, | |
| "loss": 0.5956, | |
| "num_tokens": 90748582.0, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.5726233145587984, | |
| "grad_norm": 0.8831271838249432, | |
| "learning_rate": 2.39839875536971e-05, | |
| "loss": 0.628, | |
| "num_tokens": 91203963.0, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 0.5734767025089605, | |
| "grad_norm": 1.0215192824643375, | |
| "learning_rate": 2.3921283846320434e-05, | |
| "loss": 0.5969, | |
| "num_tokens": 91699614.0, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.5743300904591228, | |
| "grad_norm": 0.8198005107193644, | |
| "learning_rate": 2.385860864127057e-05, | |
| "loss": 0.5841, | |
| "num_tokens": 92169991.0, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 0.5751834784092849, | |
| "grad_norm": 0.8904708685605999, | |
| "learning_rate": 2.3795962437717933e-05, | |
| "loss": 0.5831, | |
| "num_tokens": 92616892.0, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.576036866359447, | |
| "grad_norm": 0.8480544139333845, | |
| "learning_rate": 2.3733345734601926e-05, | |
| "loss": 0.6098, | |
| "num_tokens": 93089897.0, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.5768902543096092, | |
| "grad_norm": 0.9175224967317364, | |
| "learning_rate": 2.3670759030627026e-05, | |
| "loss": 0.6125, | |
| "num_tokens": 93591669.0, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.5777436422597713, | |
| "grad_norm": 0.8457756162517169, | |
| "learning_rate": 2.3608202824258756e-05, | |
| "loss": 0.5802, | |
| "num_tokens": 94048412.0, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 0.5785970302099335, | |
| "grad_norm": 0.8657800668830039, | |
| "learning_rate": 2.3545677613719796e-05, | |
| "loss": 0.638, | |
| "num_tokens": 94512997.0, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.5794504181600956, | |
| "grad_norm": 0.9449477364366197, | |
| "learning_rate": 2.3483183896985905e-05, | |
| "loss": 0.6279, | |
| "num_tokens": 94989943.0, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 0.5803038061102577, | |
| "grad_norm": 0.89955796478803, | |
| "learning_rate": 2.3420722171782044e-05, | |
| "loss": 0.559, | |
| "num_tokens": 95391791.0, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.5811571940604199, | |
| "grad_norm": 0.929574266645046, | |
| "learning_rate": 2.335829293557839e-05, | |
| "loss": 0.562, | |
| "num_tokens": 95831622.0, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 0.582010582010582, | |
| "grad_norm": 0.8975741522316928, | |
| "learning_rate": 2.3295896685586327e-05, | |
| "loss": 0.5849, | |
| "num_tokens": 96232233.0, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.5828639699607442, | |
| "grad_norm": 0.9836672146613759, | |
| "learning_rate": 2.3233533918754568e-05, | |
| "loss": 0.629, | |
| "num_tokens": 96733338.0, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 0.5837173579109063, | |
| "grad_norm": 0.9309413627535794, | |
| "learning_rate": 2.3171205131765106e-05, | |
| "loss": 0.6573, | |
| "num_tokens": 97192618.0, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.5845707458610685, | |
| "grad_norm": 0.912682856813927, | |
| "learning_rate": 2.3108910821029357e-05, | |
| "loss": 0.6005, | |
| "num_tokens": 97721791.0, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 0.5854241338112306, | |
| "grad_norm": 1.0215174631608348, | |
| "learning_rate": 2.304665148268411e-05, | |
| "loss": 0.6355, | |
| "num_tokens": 98220639.0, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.5862775217613927, | |
| "grad_norm": 0.842703755218707, | |
| "learning_rate": 2.2984427612587638e-05, | |
| "loss": 0.5832, | |
| "num_tokens": 98701613.0, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 0.5871309097115549, | |
| "grad_norm": 0.9052396390768784, | |
| "learning_rate": 2.2922239706315745e-05, | |
| "loss": 0.5967, | |
| "num_tokens": 99120584.0, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.587984297661717, | |
| "grad_norm": 0.8641185188951044, | |
| "learning_rate": 2.2860088259157776e-05, | |
| "loss": 0.6187, | |
| "num_tokens": 99620033.0, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 0.5888376856118792, | |
| "grad_norm": 0.8898937101798601, | |
| "learning_rate": 2.2797973766112702e-05, | |
| "loss": 0.6015, | |
| "num_tokens": 100158385.0, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.5896910735620413, | |
| "grad_norm": 0.9808515333821658, | |
| "learning_rate": 2.2735896721885218e-05, | |
| "loss": 0.5857, | |
| "num_tokens": 100605881.0, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 0.5905444615122034, | |
| "grad_norm": 0.9118752503463046, | |
| "learning_rate": 2.2673857620881712e-05, | |
| "loss": 0.5949, | |
| "num_tokens": 101113231.0, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.5913978494623656, | |
| "grad_norm": 0.851226680431884, | |
| "learning_rate": 2.2611856957206413e-05, | |
| "loss": 0.5625, | |
| "num_tokens": 101623625.0, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 0.5922512374125277, | |
| "grad_norm": 0.9845603139612755, | |
| "learning_rate": 2.2549895224657392e-05, | |
| "loss": 0.567, | |
| "num_tokens": 102084661.0, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.5931046253626899, | |
| "grad_norm": 1.081339352649647, | |
| "learning_rate": 2.248797291672267e-05, | |
| "loss": 0.5548, | |
| "num_tokens": 102541320.0, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 0.593958013312852, | |
| "grad_norm": 1.0080529824324478, | |
| "learning_rate": 2.2426090526576288e-05, | |
| "loss": 0.6448, | |
| "num_tokens": 103056557.0, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.5948114012630141, | |
| "grad_norm": 0.932290425878926, | |
| "learning_rate": 2.2364248547074335e-05, | |
| "loss": 0.5876, | |
| "num_tokens": 103525382.0, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 0.5956647892131763, | |
| "grad_norm": 0.9350679504132159, | |
| "learning_rate": 2.2302447470751087e-05, | |
| "loss": 0.6405, | |
| "num_tokens": 104029345.0, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.5965181771633384, | |
| "grad_norm": 0.9496853296189571, | |
| "learning_rate": 2.224068778981501e-05, | |
| "loss": 0.5794, | |
| "num_tokens": 104507487.0, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 0.5973715651135006, | |
| "grad_norm": 0.9723799452134716, | |
| "learning_rate": 2.2178969996144933e-05, | |
| "loss": 0.6131, | |
| "num_tokens": 105032823.0, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.5982249530636627, | |
| "grad_norm": 0.9189931555598423, | |
| "learning_rate": 2.211729458128603e-05, | |
| "loss": 0.5845, | |
| "num_tokens": 105477154.0, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 0.5990783410138248, | |
| "grad_norm": 0.9239578729799689, | |
| "learning_rate": 2.205566203644598e-05, | |
| "loss": 0.5822, | |
| "num_tokens": 105921473.0, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.599931728963987, | |
| "grad_norm": 0.8956713854458557, | |
| "learning_rate": 2.1994072852491028e-05, | |
| "loss": 0.5684, | |
| "num_tokens": 106425660.0, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 0.6007851169141492, | |
| "grad_norm": 0.8483775695612638, | |
| "learning_rate": 2.1932527519942048e-05, | |
| "loss": 0.5502, | |
| "num_tokens": 106885151.0, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.6016385048643114, | |
| "grad_norm": 0.8076921584034902, | |
| "learning_rate": 2.1871026528970706e-05, | |
| "loss": 0.5755, | |
| "num_tokens": 107317240.0, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.6024918928144735, | |
| "grad_norm": 0.8309185583056886, | |
| "learning_rate": 2.1809570369395476e-05, | |
| "loss": 0.5729, | |
| "num_tokens": 107753016.0, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.6033452807646356, | |
| "grad_norm": 0.9182402156829599, | |
| "learning_rate": 2.1748159530677808e-05, | |
| "loss": 0.6363, | |
| "num_tokens": 108236529.0, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 0.6041986687147978, | |
| "grad_norm": 0.9215609061374522, | |
| "learning_rate": 2.1686794501918183e-05, | |
| "loss": 0.6056, | |
| "num_tokens": 108671786.0, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.6050520566649599, | |
| "grad_norm": 0.9245542050305942, | |
| "learning_rate": 2.1625475771852217e-05, | |
| "loss": 0.6313, | |
| "num_tokens": 109188437.0, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 0.6059054446151221, | |
| "grad_norm": 0.7511501718290673, | |
| "learning_rate": 2.156420382884682e-05, | |
| "loss": 0.6066, | |
| "num_tokens": 109739863.0, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.6067588325652842, | |
| "grad_norm": 0.9465637379583386, | |
| "learning_rate": 2.1502979160896243e-05, | |
| "loss": 0.5925, | |
| "num_tokens": 110107306.0, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 0.6076122205154463, | |
| "grad_norm": 0.8693114102932104, | |
| "learning_rate": 2.1441802255618227e-05, | |
| "loss": 0.602, | |
| "num_tokens": 110648438.0, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.6084656084656085, | |
| "grad_norm": 1.013343208376697, | |
| "learning_rate": 2.138067360025012e-05, | |
| "loss": 0.6372, | |
| "num_tokens": 111128095.0, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 0.6093189964157706, | |
| "grad_norm": 1.0456050971697508, | |
| "learning_rate": 2.1319593681644983e-05, | |
| "loss": 0.6223, | |
| "num_tokens": 111581630.0, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.6101723843659328, | |
| "grad_norm": 0.9777795049451665, | |
| "learning_rate": 2.125856298626772e-05, | |
| "loss": 0.6072, | |
| "num_tokens": 112053972.0, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 0.6110257723160949, | |
| "grad_norm": 0.9366594738407755, | |
| "learning_rate": 2.1197582000191195e-05, | |
| "loss": 0.5691, | |
| "num_tokens": 112477812.0, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.611879160266257, | |
| "grad_norm": 0.8631576042358379, | |
| "learning_rate": 2.1136651209092366e-05, | |
| "loss": 0.578, | |
| "num_tokens": 112954742.0, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 0.6127325482164192, | |
| "grad_norm": 0.9229121201174906, | |
| "learning_rate": 2.1075771098248435e-05, | |
| "loss": 0.5906, | |
| "num_tokens": 113477895.0, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.6135859361665813, | |
| "grad_norm": 0.8747642696537712, | |
| "learning_rate": 2.101494215253295e-05, | |
| "loss": 0.598, | |
| "num_tokens": 113996433.0, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 0.6144393241167435, | |
| "grad_norm": 0.8226734099514453, | |
| "learning_rate": 2.095416485641197e-05, | |
| "loss": 0.6226, | |
| "num_tokens": 114528571.0, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.6152927120669056, | |
| "grad_norm": 0.8561260036624176, | |
| "learning_rate": 2.0893439693940164e-05, | |
| "loss": 0.5939, | |
| "num_tokens": 114979015.0, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 0.6161461000170677, | |
| "grad_norm": 0.7185593819396288, | |
| "learning_rate": 2.083276714875704e-05, | |
| "loss": 0.6138, | |
| "num_tokens": 115545543.0, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.6169994879672299, | |
| "grad_norm": 0.8727068314836006, | |
| "learning_rate": 2.0772147704083018e-05, | |
| "loss": 0.6458, | |
| "num_tokens": 116033061.0, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 0.617852875917392, | |
| "grad_norm": 1.0114732688960872, | |
| "learning_rate": 2.071158184271558e-05, | |
| "loss": 0.5838, | |
| "num_tokens": 116495965.0, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.6187062638675542, | |
| "grad_norm": 0.8448415729964601, | |
| "learning_rate": 2.0651070047025495e-05, | |
| "loss": 0.6449, | |
| "num_tokens": 116996015.0, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 0.6195596518177163, | |
| "grad_norm": 0.9318187070912285, | |
| "learning_rate": 2.059061279895288e-05, | |
| "loss": 0.5673, | |
| "num_tokens": 117451785.0, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.6204130397678784, | |
| "grad_norm": 0.900817894542971, | |
| "learning_rate": 2.0530210580003462e-05, | |
| "loss": 0.6137, | |
| "num_tokens": 117898073.0, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 0.6212664277180406, | |
| "grad_norm": 0.9353646155780253, | |
| "learning_rate": 2.0469863871244653e-05, | |
| "loss": 0.586, | |
| "num_tokens": 118370098.0, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.6221198156682027, | |
| "grad_norm": 0.9320717548727218, | |
| "learning_rate": 2.040957315330179e-05, | |
| "loss": 0.6319, | |
| "num_tokens": 118836147.0, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 0.622973203618365, | |
| "grad_norm": 0.9657934461409301, | |
| "learning_rate": 2.0349338906354265e-05, | |
| "loss": 0.6149, | |
| "num_tokens": 119296261.0, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.623826591568527, | |
| "grad_norm": 0.8123400658647655, | |
| "learning_rate": 2.028916161013171e-05, | |
| "loss": 0.5993, | |
| "num_tokens": 119753329.0, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 0.6246799795186891, | |
| "grad_norm": 0.8641657920470441, | |
| "learning_rate": 2.0229041743910177e-05, | |
| "loss": 0.6031, | |
| "num_tokens": 120219689.0, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.6255333674688514, | |
| "grad_norm": 0.8271950377447332, | |
| "learning_rate": 2.016897978650833e-05, | |
| "loss": 0.5861, | |
| "num_tokens": 120695369.0, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 0.6263867554190135, | |
| "grad_norm": 0.8151322459613584, | |
| "learning_rate": 2.010897621628362e-05, | |
| "loss": 0.6138, | |
| "num_tokens": 121195424.0, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.6272401433691757, | |
| "grad_norm": 0.9177483702936727, | |
| "learning_rate": 2.0049031511128485e-05, | |
| "loss": 0.588, | |
| "num_tokens": 121647191.0, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.6280935313193378, | |
| "grad_norm": 0.9121002779909388, | |
| "learning_rate": 1.998914614846652e-05, | |
| "loss": 0.5563, | |
| "num_tokens": 122107208.0, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.6289469192694999, | |
| "grad_norm": 0.8173540767440151, | |
| "learning_rate": 1.9929320605248724e-05, | |
| "loss": 0.564, | |
| "num_tokens": 122573529.0, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 0.6298003072196621, | |
| "grad_norm": 1.043731044651841, | |
| "learning_rate": 1.9869555357949632e-05, | |
| "loss": 0.6045, | |
| "num_tokens": 123065555.0, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.6306536951698242, | |
| "grad_norm": 1.0739948934639034, | |
| "learning_rate": 1.980985088256358e-05, | |
| "loss": 0.5956, | |
| "num_tokens": 123541397.0, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 0.6315070831199864, | |
| "grad_norm": 0.9196803826879293, | |
| "learning_rate": 1.975020765460091e-05, | |
| "loss": 0.5922, | |
| "num_tokens": 124027333.0, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.6323604710701485, | |
| "grad_norm": 0.850161249395921, | |
| "learning_rate": 1.9690626149084123e-05, | |
| "loss": 0.6113, | |
| "num_tokens": 124511073.0, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 0.6332138590203107, | |
| "grad_norm": 0.8586604654257546, | |
| "learning_rate": 1.9631106840544172e-05, | |
| "loss": 0.5594, | |
| "num_tokens": 124983891.0, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.6340672469704728, | |
| "grad_norm": 0.8071502796590858, | |
| "learning_rate": 1.9571650203016617e-05, | |
| "loss": 0.6022, | |
| "num_tokens": 125450183.0, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 0.6349206349206349, | |
| "grad_norm": 0.8369926469871681, | |
| "learning_rate": 1.9512256710037917e-05, | |
| "loss": 0.5924, | |
| "num_tokens": 125928696.0, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.6357740228707971, | |
| "grad_norm": 0.8487805091950206, | |
| "learning_rate": 1.9452926834641617e-05, | |
| "loss": 0.5845, | |
| "num_tokens": 126461448.0, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 0.6366274108209592, | |
| "grad_norm": 0.7988342822329317, | |
| "learning_rate": 1.939366104935455e-05, | |
| "loss": 0.599, | |
| "num_tokens": 126938768.0, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.6374807987711214, | |
| "grad_norm": 0.926173531550069, | |
| "learning_rate": 1.9334459826193145e-05, | |
| "loss": 0.6254, | |
| "num_tokens": 127319352.0, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 0.6383341867212835, | |
| "grad_norm": 0.837561318846321, | |
| "learning_rate": 1.927532363665962e-05, | |
| "loss": 0.653, | |
| "num_tokens": 127838140.0, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.6391875746714456, | |
| "grad_norm": 0.8466881218592522, | |
| "learning_rate": 1.921625295173824e-05, | |
| "loss": 0.5964, | |
| "num_tokens": 128310279.0, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 0.6400409626216078, | |
| "grad_norm": 0.7976790416425076, | |
| "learning_rate": 1.9157248241891574e-05, | |
| "loss": 0.6267, | |
| "num_tokens": 128780236.0, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.6408943505717699, | |
| "grad_norm": 0.9541677508158435, | |
| "learning_rate": 1.9098309977056717e-05, | |
| "loss": 0.5881, | |
| "num_tokens": 129199502.0, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 0.6417477385219321, | |
| "grad_norm": 1.0530842262619469, | |
| "learning_rate": 1.9039438626641598e-05, | |
| "loss": 0.6341, | |
| "num_tokens": 129727094.0, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.6426011264720942, | |
| "grad_norm": 0.8739037742597267, | |
| "learning_rate": 1.8980634659521183e-05, | |
| "loss": 0.5494, | |
| "num_tokens": 130229389.0, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 0.6434545144222563, | |
| "grad_norm": 0.8423251381628554, | |
| "learning_rate": 1.89218985440338e-05, | |
| "loss": 0.5699, | |
| "num_tokens": 130679646.0, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.6443079023724185, | |
| "grad_norm": 0.9485543731171855, | |
| "learning_rate": 1.886323074797736e-05, | |
| "loss": 0.5918, | |
| "num_tokens": 131161913.0, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.8516489909370533, | |
| "learning_rate": 1.880463173860565e-05, | |
| "loss": 0.6067, | |
| "num_tokens": 131619026.0, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.6460146782727428, | |
| "grad_norm": 0.851808651800723, | |
| "learning_rate": 1.8746101982624632e-05, | |
| "loss": 0.5866, | |
| "num_tokens": 132075908.0, | |
| "step": 3785 | |
| }, | |
| { | |
| "epoch": 0.6468680662229049, | |
| "grad_norm": 0.8435590362188361, | |
| "learning_rate": 1.8687641946188673e-05, | |
| "loss": 0.5547, | |
| "num_tokens": 132528956.0, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.647721454173067, | |
| "grad_norm": 0.9957301124037699, | |
| "learning_rate": 1.8629252094896903e-05, | |
| "loss": 0.6261, | |
| "num_tokens": 132985693.0, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 0.6485748421232292, | |
| "grad_norm": 0.8507050967745332, | |
| "learning_rate": 1.8570932893789443e-05, | |
| "loss": 0.5779, | |
| "num_tokens": 133479496.0, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.6494282300733913, | |
| "grad_norm": 0.918721141542831, | |
| "learning_rate": 1.8512684807343734e-05, | |
| "loss": 0.5992, | |
| "num_tokens": 133904688.0, | |
| "step": 3805 | |
| }, | |
| { | |
| "epoch": 0.6502816180235536, | |
| "grad_norm": 0.8982922900971878, | |
| "learning_rate": 1.8454508299470846e-05, | |
| "loss": 0.6113, | |
| "num_tokens": 134399643.0, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.6511350059737157, | |
| "grad_norm": 1.0182409968166741, | |
| "learning_rate": 1.8396403833511744e-05, | |
| "loss": 0.6079, | |
| "num_tokens": 134876960.0, | |
| "step": 3815 | |
| }, | |
| { | |
| "epoch": 0.6519883939238778, | |
| "grad_norm": 0.8980802950652549, | |
| "learning_rate": 1.8338371872233646e-05, | |
| "loss": 0.5456, | |
| "num_tokens": 135314218.0, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.65284178187404, | |
| "grad_norm": 0.8141386565121531, | |
| "learning_rate": 1.828041287782628e-05, | |
| "loss": 0.596, | |
| "num_tokens": 135803609.0, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 0.6536951698242021, | |
| "grad_norm": 0.7990864081217415, | |
| "learning_rate": 1.8222527311898274e-05, | |
| "loss": 0.5935, | |
| "num_tokens": 136226687.0, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.6545485577743643, | |
| "grad_norm": 0.9144391970153878, | |
| "learning_rate": 1.8164715635473438e-05, | |
| "loss": 0.5617, | |
| "num_tokens": 136652839.0, | |
| "step": 3835 | |
| }, | |
| { | |
| "epoch": 0.6554019457245264, | |
| "grad_norm": 0.9096006554057322, | |
| "learning_rate": 1.8106978308987076e-05, | |
| "loss": 0.5756, | |
| "num_tokens": 137152722.0, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.6562553336746885, | |
| "grad_norm": 0.8047183013047816, | |
| "learning_rate": 1.8049315792282345e-05, | |
| "loss": 0.5913, | |
| "num_tokens": 137684409.0, | |
| "step": 3845 | |
| }, | |
| { | |
| "epoch": 0.6571087216248507, | |
| "grad_norm": 0.8641039631825128, | |
| "learning_rate": 1.799172854460659e-05, | |
| "loss": 0.6168, | |
| "num_tokens": 138187589.0, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.6579621095750128, | |
| "grad_norm": 0.8900431344536037, | |
| "learning_rate": 1.793421702460769e-05, | |
| "loss": 0.5878, | |
| "num_tokens": 138652113.0, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 0.658815497525175, | |
| "grad_norm": 0.9005201091983084, | |
| "learning_rate": 1.78767816903304e-05, | |
| "loss": 0.6507, | |
| "num_tokens": 139126560.0, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.6596688854753371, | |
| "grad_norm": 0.901548092005396, | |
| "learning_rate": 1.7819422999212677e-05, | |
| "loss": 0.5812, | |
| "num_tokens": 139561418.0, | |
| "step": 3865 | |
| }, | |
| { | |
| "epoch": 0.6605222734254992, | |
| "grad_norm": 0.9380903683320696, | |
| "learning_rate": 1.7762141408082096e-05, | |
| "loss": 0.6025, | |
| "num_tokens": 140037433.0, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.6613756613756614, | |
| "grad_norm": 0.953322480284107, | |
| "learning_rate": 1.7704937373152147e-05, | |
| "loss": 0.5809, | |
| "num_tokens": 140523109.0, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 0.6622290493258235, | |
| "grad_norm": 0.8874432089428697, | |
| "learning_rate": 1.7647811350018646e-05, | |
| "loss": 0.6044, | |
| "num_tokens": 141015373.0, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.6630824372759857, | |
| "grad_norm": 0.8957954095580423, | |
| "learning_rate": 1.7590763793656095e-05, | |
| "loss": 0.5643, | |
| "num_tokens": 141446951.0, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 0.6639358252261478, | |
| "grad_norm": 0.855096578289474, | |
| "learning_rate": 1.753379515841404e-05, | |
| "loss": 0.564, | |
| "num_tokens": 141886136.0, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.6647892131763099, | |
| "grad_norm": 0.8271763713417735, | |
| "learning_rate": 1.7476905898013494e-05, | |
| "loss": 0.5805, | |
| "num_tokens": 142352308.0, | |
| "step": 3895 | |
| }, | |
| { | |
| "epoch": 0.6656426011264721, | |
| "grad_norm": 0.7605289501292125, | |
| "learning_rate": 1.7420096465543262e-05, | |
| "loss": 0.5847, | |
| "num_tokens": 142858814.0, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.6664959890766342, | |
| "grad_norm": 0.9062476293435004, | |
| "learning_rate": 1.73633673134564e-05, | |
| "loss": 0.5945, | |
| "num_tokens": 143326743.0, | |
| "step": 3905 | |
| }, | |
| { | |
| "epoch": 0.6673493770267964, | |
| "grad_norm": 0.8738648590116904, | |
| "learning_rate": 1.7306718893566556e-05, | |
| "loss": 0.6153, | |
| "num_tokens": 143878702.0, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.6682027649769585, | |
| "grad_norm": 0.9899124493608162, | |
| "learning_rate": 1.725015165704441e-05, | |
| "loss": 0.598, | |
| "num_tokens": 144328462.0, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 0.6690561529271206, | |
| "grad_norm": 0.8292704517142464, | |
| "learning_rate": 1.7193666054414062e-05, | |
| "loss": 0.5899, | |
| "num_tokens": 144856028.0, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.6699095408772828, | |
| "grad_norm": 0.8783635732800318, | |
| "learning_rate": 1.7137262535549424e-05, | |
| "loss": 0.5837, | |
| "num_tokens": 145319975.0, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 0.6707629288274449, | |
| "grad_norm": 0.9691306002384159, | |
| "learning_rate": 1.7080941549670704e-05, | |
| "loss": 0.5947, | |
| "num_tokens": 145801559.0, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.6716163167776071, | |
| "grad_norm": 0.8355517780744063, | |
| "learning_rate": 1.7024703545340738e-05, | |
| "loss": 0.5661, | |
| "num_tokens": 146254549.0, | |
| "step": 3935 | |
| }, | |
| { | |
| "epoch": 0.6724697047277692, | |
| "grad_norm": 0.7902115733088367, | |
| "learning_rate": 1.6968548970461497e-05, | |
| "loss": 0.5626, | |
| "num_tokens": 146721482.0, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.6733230926779313, | |
| "grad_norm": 0.8524753464674293, | |
| "learning_rate": 1.691247827227049e-05, | |
| "loss": 0.5552, | |
| "num_tokens": 147214936.0, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 0.6741764806280935, | |
| "grad_norm": 0.8097399865281222, | |
| "learning_rate": 1.6856491897337152e-05, | |
| "loss": 0.5702, | |
| "num_tokens": 147707141.0, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.6750298685782556, | |
| "grad_norm": 0.8749505840231333, | |
| "learning_rate": 1.6800590291559395e-05, | |
| "loss": 0.5799, | |
| "num_tokens": 148175414.0, | |
| "step": 3955 | |
| }, | |
| { | |
| "epoch": 0.6758832565284179, | |
| "grad_norm": 0.9825297053765945, | |
| "learning_rate": 1.6744773900159954e-05, | |
| "loss": 0.5953, | |
| "num_tokens": 148620002.0, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.67673664447858, | |
| "grad_norm": 0.8965602948755153, | |
| "learning_rate": 1.6689043167682884e-05, | |
| "loss": 0.5825, | |
| "num_tokens": 149121519.0, | |
| "step": 3965 | |
| }, | |
| { | |
| "epoch": 0.6775900324287422, | |
| "grad_norm": 0.8623362168580441, | |
| "learning_rate": 1.663339853799005e-05, | |
| "loss": 0.5578, | |
| "num_tokens": 149555256.0, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.6784434203789043, | |
| "grad_norm": 1.0250814917423345, | |
| "learning_rate": 1.657784045425752e-05, | |
| "loss": 0.5544, | |
| "num_tokens": 150051458.0, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 0.6792968083290664, | |
| "grad_norm": 0.9016034723303509, | |
| "learning_rate": 1.6522369358972107e-05, | |
| "loss": 0.5731, | |
| "num_tokens": 150535405.0, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.6801501962792286, | |
| "grad_norm": 1.0801481065719016, | |
| "learning_rate": 1.646698569392779e-05, | |
| "loss": 0.609, | |
| "num_tokens": 150994090.0, | |
| "step": 3985 | |
| }, | |
| { | |
| "epoch": 0.6810035842293907, | |
| "grad_norm": 0.9026329662433026, | |
| "learning_rate": 1.6411689900222233e-05, | |
| "loss": 0.589, | |
| "num_tokens": 151437507.0, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.6818569721795529, | |
| "grad_norm": 0.8305286612610906, | |
| "learning_rate": 1.6356482418253264e-05, | |
| "loss": 0.6108, | |
| "num_tokens": 151897755.0, | |
| "step": 3995 | |
| }, | |
| { | |
| "epoch": 0.682710360129715, | |
| "grad_norm": 0.8759053799322227, | |
| "learning_rate": 1.630136368771534e-05, | |
| "loss": 0.5727, | |
| "num_tokens": 152392981.0, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.6835637480798771, | |
| "grad_norm": 0.8498248549941118, | |
| "learning_rate": 1.624633414759608e-05, | |
| "loss": 0.628, | |
| "num_tokens": 152945921.0, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 0.6844171360300393, | |
| "grad_norm": 0.9577467020674911, | |
| "learning_rate": 1.619139423617274e-05, | |
| "loss": 0.5955, | |
| "num_tokens": 153405452.0, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.6852705239802014, | |
| "grad_norm": 0.8677137719544256, | |
| "learning_rate": 1.6136544391008766e-05, | |
| "loss": 0.6011, | |
| "num_tokens": 153866503.0, | |
| "step": 4015 | |
| }, | |
| { | |
| "epoch": 0.6861239119303636, | |
| "grad_norm": 0.8762759941080992, | |
| "learning_rate": 1.608178504895025e-05, | |
| "loss": 0.6003, | |
| "num_tokens": 154301458.0, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.6869772998805257, | |
| "grad_norm": 0.9447545948904823, | |
| "learning_rate": 1.6027116646122497e-05, | |
| "loss": 0.5676, | |
| "num_tokens": 154747758.0, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 0.6878306878306878, | |
| "grad_norm": 0.8154066594371561, | |
| "learning_rate": 1.5972539617926547e-05, | |
| "loss": 0.5603, | |
| "num_tokens": 155226031.0, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.68868407578085, | |
| "grad_norm": 0.8283393416577676, | |
| "learning_rate": 1.5918054399035656e-05, | |
| "loss": 0.5773, | |
| "num_tokens": 155700124.0, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 0.6895374637310121, | |
| "grad_norm": 0.9745180674537747, | |
| "learning_rate": 1.5863661423391924e-05, | |
| "loss": 0.6037, | |
| "num_tokens": 156126810.0, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.6903908516811743, | |
| "grad_norm": 0.8064566786044969, | |
| "learning_rate": 1.580936112420275e-05, | |
| "loss": 0.5807, | |
| "num_tokens": 156601011.0, | |
| "step": 4045 | |
| }, | |
| { | |
| "epoch": 0.6912442396313364, | |
| "grad_norm": 0.8001959729849549, | |
| "learning_rate": 1.5755153933937433e-05, | |
| "loss": 0.6164, | |
| "num_tokens": 157095880.0, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.6920976275814985, | |
| "grad_norm": 0.942834039862139, | |
| "learning_rate": 1.5701040284323733e-05, | |
| "loss": 0.5941, | |
| "num_tokens": 157586499.0, | |
| "step": 4055 | |
| }, | |
| { | |
| "epoch": 0.6929510155316607, | |
| "grad_norm": 0.9715453174727748, | |
| "learning_rate": 1.5647020606344374e-05, | |
| "loss": 0.5795, | |
| "num_tokens": 158008275.0, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.6938044034818228, | |
| "grad_norm": 0.9026555991717418, | |
| "learning_rate": 1.5593095330233702e-05, | |
| "loss": 0.5771, | |
| "num_tokens": 158452236.0, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 0.694657791431985, | |
| "grad_norm": 1.0117298958493552, | |
| "learning_rate": 1.553926488547417e-05, | |
| "loss": 0.6012, | |
| "num_tokens": 158941412.0, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.6955111793821471, | |
| "grad_norm": 0.8584768240163265, | |
| "learning_rate": 1.5485529700792972e-05, | |
| "loss": 0.5709, | |
| "num_tokens": 159418312.0, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 0.6963645673323092, | |
| "grad_norm": 0.796633996689525, | |
| "learning_rate": 1.5431890204158623e-05, | |
| "loss": 0.5733, | |
| "num_tokens": 159876193.0, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.6972179552824714, | |
| "grad_norm": 0.8652610369119335, | |
| "learning_rate": 1.5378346822777506e-05, | |
| "loss": 0.5653, | |
| "num_tokens": 160337567.0, | |
| "step": 4085 | |
| }, | |
| { | |
| "epoch": 0.6980713432326335, | |
| "grad_norm": 0.883756733604466, | |
| "learning_rate": 1.5324899983090552e-05, | |
| "loss": 0.5973, | |
| "num_tokens": 160836081.0, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.6989247311827957, | |
| "grad_norm": 0.8346242397306459, | |
| "learning_rate": 1.5271550110769756e-05, | |
| "loss": 0.6197, | |
| "num_tokens": 161374383.0, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 0.6997781191329578, | |
| "grad_norm": 0.8332005290816415, | |
| "learning_rate": 1.5218297630714829e-05, | |
| "loss": 0.6205, | |
| "num_tokens": 161857305.0, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.7006315070831199, | |
| "grad_norm": 0.8712175844197076, | |
| "learning_rate": 1.516514296704984e-05, | |
| "loss": 0.5773, | |
| "num_tokens": 162348361.0, | |
| "step": 4105 | |
| }, | |
| { | |
| "epoch": 0.7014848950332822, | |
| "grad_norm": 0.9058422726334173, | |
| "learning_rate": 1.511208654311977e-05, | |
| "loss": 0.5879, | |
| "num_tokens": 162851803.0, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.7023382829834443, | |
| "grad_norm": 0.9044790346726684, | |
| "learning_rate": 1.5059128781487225e-05, | |
| "loss": 0.5829, | |
| "num_tokens": 163287986.0, | |
| "step": 4115 | |
| }, | |
| { | |
| "epoch": 0.7031916709336065, | |
| "grad_norm": 0.8648827055391948, | |
| "learning_rate": 1.5006270103928976e-05, | |
| "loss": 0.5817, | |
| "num_tokens": 163718273.0, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.7040450588837686, | |
| "grad_norm": 0.8076594039591006, | |
| "learning_rate": 1.4953510931432685e-05, | |
| "loss": 0.5985, | |
| "num_tokens": 164166532.0, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 0.7048984468339307, | |
| "grad_norm": 0.815303572116947, | |
| "learning_rate": 1.4900851684193512e-05, | |
| "loss": 0.6236, | |
| "num_tokens": 164596080.0, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.7057518347840929, | |
| "grad_norm": 0.8302985383503674, | |
| "learning_rate": 1.4848292781610751e-05, | |
| "loss": 0.597, | |
| "num_tokens": 165078423.0, | |
| "step": 4135 | |
| }, | |
| { | |
| "epoch": 0.706605222734255, | |
| "grad_norm": 0.8719511002078489, | |
| "learning_rate": 1.4795834642284528e-05, | |
| "loss": 0.5958, | |
| "num_tokens": 165553538.0, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.7074586106844172, | |
| "grad_norm": 0.8963125498569914, | |
| "learning_rate": 1.4743477684012438e-05, | |
| "loss": 0.5988, | |
| "num_tokens": 166052376.0, | |
| "step": 4145 | |
| }, | |
| { | |
| "epoch": 0.7083119986345793, | |
| "grad_norm": 0.9500910001284955, | |
| "learning_rate": 1.4691222323786253e-05, | |
| "loss": 0.5795, | |
| "num_tokens": 166509198.0, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.7091653865847414, | |
| "grad_norm": 0.9576655914462456, | |
| "learning_rate": 1.4639068977788542e-05, | |
| "loss": 0.6329, | |
| "num_tokens": 167013563.0, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 0.7100187745349036, | |
| "grad_norm": 0.8549981482176169, | |
| "learning_rate": 1.4587018061389426e-05, | |
| "loss": 0.63, | |
| "num_tokens": 167448213.0, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.7108721624850657, | |
| "grad_norm": 0.8226809221364069, | |
| "learning_rate": 1.453506998914323e-05, | |
| "loss": 0.5785, | |
| "num_tokens": 167971432.0, | |
| "step": 4165 | |
| }, | |
| { | |
| "epoch": 0.7117255504352279, | |
| "grad_norm": 0.8872256261929765, | |
| "learning_rate": 1.448322517478516e-05, | |
| "loss": 0.5908, | |
| "num_tokens": 168433995.0, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.71257893838539, | |
| "grad_norm": 0.8360410226445361, | |
| "learning_rate": 1.4431484031228069e-05, | |
| "loss": 0.5594, | |
| "num_tokens": 168904923.0, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 0.7134323263355521, | |
| "grad_norm": 0.8987026195932659, | |
| "learning_rate": 1.4379846970559113e-05, | |
| "loss": 0.5793, | |
| "num_tokens": 169381861.0, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.9097929846772739, | |
| "learning_rate": 1.4328314404036486e-05, | |
| "loss": 0.5907, | |
| "num_tokens": 169808194.0, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 0.7151391022358764, | |
| "grad_norm": 1.0974482399269831, | |
| "learning_rate": 1.4276886742086175e-05, | |
| "loss": 0.5923, | |
| "num_tokens": 170298702.0, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.7159924901860386, | |
| "grad_norm": 0.9360284389299645, | |
| "learning_rate": 1.4225564394298641e-05, | |
| "loss": 0.6333, | |
| "num_tokens": 170759502.0, | |
| "step": 4195 | |
| }, | |
| { | |
| "epoch": 0.7168458781362007, | |
| "grad_norm": 0.9152888746195195, | |
| "learning_rate": 1.4174347769425594e-05, | |
| "loss": 0.5702, | |
| "num_tokens": 171213451.0, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.7176992660863628, | |
| "grad_norm": 0.886847518370069, | |
| "learning_rate": 1.412323727537671e-05, | |
| "loss": 0.6102, | |
| "num_tokens": 171667948.0, | |
| "step": 4205 | |
| }, | |
| { | |
| "epoch": 0.718552654036525, | |
| "grad_norm": 0.791848385672325, | |
| "learning_rate": 1.407223331921641e-05, | |
| "loss": 0.5422, | |
| "num_tokens": 172135689.0, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.7194060419866871, | |
| "grad_norm": 0.8410700768427187, | |
| "learning_rate": 1.4021336307160612e-05, | |
| "loss": 0.5762, | |
| "num_tokens": 172566490.0, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 0.7202594299368493, | |
| "grad_norm": 0.8585840404859496, | |
| "learning_rate": 1.3970546644573467e-05, | |
| "loss": 0.5619, | |
| "num_tokens": 173084261.0, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.7211128178870114, | |
| "grad_norm": 0.9115552103460094, | |
| "learning_rate": 1.3919864735964167e-05, | |
| "loss": 0.6028, | |
| "num_tokens": 173525169.0, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 0.7219662058371735, | |
| "grad_norm": 0.8649833459661458, | |
| "learning_rate": 1.3869290984983685e-05, | |
| "loss": 0.5774, | |
| "num_tokens": 173986927.0, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.7228195937873357, | |
| "grad_norm": 0.8467791476139611, | |
| "learning_rate": 1.3818825794421619e-05, | |
| "loss": 0.5741, | |
| "num_tokens": 174485026.0, | |
| "step": 4235 | |
| }, | |
| { | |
| "epoch": 0.7236729817374978, | |
| "grad_norm": 0.8692688687345984, | |
| "learning_rate": 1.376846956620293e-05, | |
| "loss": 0.5916, | |
| "num_tokens": 174975168.0, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.72452636968766, | |
| "grad_norm": 0.7864597052391264, | |
| "learning_rate": 1.3718222701384757e-05, | |
| "loss": 0.5736, | |
| "num_tokens": 175474304.0, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 0.7253797576378221, | |
| "grad_norm": 0.8607415619129325, | |
| "learning_rate": 1.3668085600153232e-05, | |
| "loss": 0.5864, | |
| "num_tokens": 175914177.0, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.7262331455879844, | |
| "grad_norm": 0.9509781922334833, | |
| "learning_rate": 1.3618058661820277e-05, | |
| "loss": 0.5675, | |
| "num_tokens": 176354413.0, | |
| "step": 4255 | |
| }, | |
| { | |
| "epoch": 0.7270865335381465, | |
| "grad_norm": 0.8824134882346422, | |
| "learning_rate": 1.3568142284820442e-05, | |
| "loss": 0.6194, | |
| "num_tokens": 176903257.0, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.7279399214883086, | |
| "grad_norm": 0.8709398836701243, | |
| "learning_rate": 1.3518336866707723e-05, | |
| "loss": 0.5883, | |
| "num_tokens": 177361648.0, | |
| "step": 4265 | |
| }, | |
| { | |
| "epoch": 0.7287933094384708, | |
| "grad_norm": 0.880944482403639, | |
| "learning_rate": 1.3468642804152374e-05, | |
| "loss": 0.5893, | |
| "num_tokens": 177905691.0, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.7296466973886329, | |
| "grad_norm": 0.8470029010342051, | |
| "learning_rate": 1.3419060492937802e-05, | |
| "loss": 0.5644, | |
| "num_tokens": 178398025.0, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 0.7305000853387951, | |
| "grad_norm": 0.9227327888215903, | |
| "learning_rate": 1.3369590327957348e-05, | |
| "loss": 0.5975, | |
| "num_tokens": 178872312.0, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.7313534732889572, | |
| "grad_norm": 0.9366754745843368, | |
| "learning_rate": 1.3320232703211214e-05, | |
| "loss": 0.5911, | |
| "num_tokens": 179323634.0, | |
| "step": 4285 | |
| }, | |
| { | |
| "epoch": 0.7322068612391193, | |
| "grad_norm": 0.8849252887229547, | |
| "learning_rate": 1.3270988011803243e-05, | |
| "loss": 0.6062, | |
| "num_tokens": 179771488.0, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.7330602491892815, | |
| "grad_norm": 0.7892258484928227, | |
| "learning_rate": 1.3221856645937868e-05, | |
| "loss": 0.5679, | |
| "num_tokens": 180256627.0, | |
| "step": 4295 | |
| }, | |
| { | |
| "epoch": 0.7339136371394436, | |
| "grad_norm": 0.9137667870343035, | |
| "learning_rate": 1.3172838996916936e-05, | |
| "loss": 0.61, | |
| "num_tokens": 180732439.0, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.7347670250896058, | |
| "grad_norm": 0.9631346489471143, | |
| "learning_rate": 1.3123935455136599e-05, | |
| "loss": 0.6022, | |
| "num_tokens": 181226259.0, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 0.7356204130397679, | |
| "grad_norm": 0.8968709358166163, | |
| "learning_rate": 1.307514641008424e-05, | |
| "loss": 0.5965, | |
| "num_tokens": 181714583.0, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.73647380098993, | |
| "grad_norm": 0.9166240270197292, | |
| "learning_rate": 1.302647225033532e-05, | |
| "loss": 0.5887, | |
| "num_tokens": 182146534.0, | |
| "step": 4315 | |
| }, | |
| { | |
| "epoch": 0.7373271889400922, | |
| "grad_norm": 0.8160802329739986, | |
| "learning_rate": 1.2977913363550304e-05, | |
| "loss": 0.5769, | |
| "num_tokens": 182675522.0, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.7381805768902543, | |
| "grad_norm": 0.8562280243416345, | |
| "learning_rate": 1.2929470136471607e-05, | |
| "loss": 0.6258, | |
| "num_tokens": 183187657.0, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 0.7390339648404165, | |
| "grad_norm": 0.8987806086881184, | |
| "learning_rate": 1.288114295492045e-05, | |
| "loss": 0.5997, | |
| "num_tokens": 183642551.0, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.7398873527905786, | |
| "grad_norm": 0.7328617713394691, | |
| "learning_rate": 1.2832932203793848e-05, | |
| "loss": 0.5957, | |
| "num_tokens": 184184444.0, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 0.7407407407407407, | |
| "grad_norm": 0.7980388669261456, | |
| "learning_rate": 1.2784838267061491e-05, | |
| "loss": 0.5857, | |
| "num_tokens": 184646945.0, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.7415941286909029, | |
| "grad_norm": 0.8859400994034398, | |
| "learning_rate": 1.273686152776274e-05, | |
| "loss": 0.6085, | |
| "num_tokens": 185108387.0, | |
| "step": 4345 | |
| }, | |
| { | |
| "epoch": 0.742447516641065, | |
| "grad_norm": 0.7854784674766948, | |
| "learning_rate": 1.2689002368003539e-05, | |
| "loss": 0.535, | |
| "num_tokens": 185565586.0, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.7433009045912272, | |
| "grad_norm": 0.7570419777611505, | |
| "learning_rate": 1.2641261168953366e-05, | |
| "loss": 0.6244, | |
| "num_tokens": 186141614.0, | |
| "step": 4355 | |
| }, | |
| { | |
| "epoch": 0.7441542925413893, | |
| "grad_norm": 0.7757142596947014, | |
| "learning_rate": 1.2593638310842235e-05, | |
| "loss": 0.581, | |
| "num_tokens": 186661552.0, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.7450076804915514, | |
| "grad_norm": 0.8839499200264586, | |
| "learning_rate": 1.2546134172957619e-05, | |
| "loss": 0.6213, | |
| "num_tokens": 187144035.0, | |
| "step": 4365 | |
| }, | |
| { | |
| "epoch": 0.7458610684417136, | |
| "grad_norm": 0.816312832091104, | |
| "learning_rate": 1.2498749133641489e-05, | |
| "loss": 0.5561, | |
| "num_tokens": 187581120.0, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 0.7467144563918757, | |
| "grad_norm": 0.8213980561955142, | |
| "learning_rate": 1.245148357028725e-05, | |
| "loss": 0.5857, | |
| "num_tokens": 188075796.0, | |
| "step": 4375 | |
| }, | |
| { | |
| "epoch": 0.7475678443420379, | |
| "grad_norm": 0.9380415013983583, | |
| "learning_rate": 1.2404337859336743e-05, | |
| "loss": 0.5675, | |
| "num_tokens": 188555486.0, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.7484212322922, | |
| "grad_norm": 0.9725183192014755, | |
| "learning_rate": 1.2357312376277279e-05, | |
| "loss": 0.5502, | |
| "num_tokens": 188977941.0, | |
| "step": 4385 | |
| }, | |
| { | |
| "epoch": 0.7492746202423621, | |
| "grad_norm": 0.7992605376572952, | |
| "learning_rate": 1.2310407495638599e-05, | |
| "loss": 0.5645, | |
| "num_tokens": 189416481.0, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 0.7501280081925243, | |
| "grad_norm": 0.9374154424761773, | |
| "learning_rate": 1.226362359098995e-05, | |
| "loss": 0.5715, | |
| "num_tokens": 189868560.0, | |
| "step": 4395 | |
| }, | |
| { | |
| "epoch": 0.7509813961426864, | |
| "grad_norm": 0.891667619403936, | |
| "learning_rate": 1.2216961034937048e-05, | |
| "loss": 0.593, | |
| "num_tokens": 190341488.0, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.7518347840928487, | |
| "grad_norm": 0.8840376687918124, | |
| "learning_rate": 1.2170420199119151e-05, | |
| "loss": 0.5809, | |
| "num_tokens": 190774240.0, | |
| "step": 4405 | |
| }, | |
| { | |
| "epoch": 0.7526881720430108, | |
| "grad_norm": 0.8190274046973514, | |
| "learning_rate": 1.2124001454206102e-05, | |
| "loss": 0.5845, | |
| "num_tokens": 191282701.0, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.7535415599931728, | |
| "grad_norm": 0.9123518049084535, | |
| "learning_rate": 1.2077705169895338e-05, | |
| "loss": 0.5477, | |
| "num_tokens": 191713963.0, | |
| "step": 4415 | |
| }, | |
| { | |
| "epoch": 0.7543949479433351, | |
| "grad_norm": 0.8188073549864183, | |
| "learning_rate": 1.2031531714908997e-05, | |
| "loss": 0.5755, | |
| "num_tokens": 192195436.0, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.7552483358934972, | |
| "grad_norm": 0.7994958336328707, | |
| "learning_rate": 1.1985481456990928e-05, | |
| "loss": 0.6002, | |
| "num_tokens": 192609004.0, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 0.7561017238436594, | |
| "grad_norm": 0.9659539232854811, | |
| "learning_rate": 1.1939554762903813e-05, | |
| "loss": 0.5689, | |
| "num_tokens": 193040643.0, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 0.7569551117938215, | |
| "grad_norm": 0.8994805121484224, | |
| "learning_rate": 1.189375199842622e-05, | |
| "loss": 0.5746, | |
| "num_tokens": 193509769.0, | |
| "step": 4435 | |
| }, | |
| { | |
| "epoch": 0.7578084997439836, | |
| "grad_norm": 0.879957189035845, | |
| "learning_rate": 1.1848073528349676e-05, | |
| "loss": 0.5944, | |
| "num_tokens": 193975222.0, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.7586618876941458, | |
| "grad_norm": 0.7874543956440837, | |
| "learning_rate": 1.1802519716475786e-05, | |
| "loss": 0.5826, | |
| "num_tokens": 194477436.0, | |
| "step": 4445 | |
| }, | |
| { | |
| "epoch": 0.7595152756443079, | |
| "grad_norm": 0.7963664649172572, | |
| "learning_rate": 1.1757090925613323e-05, | |
| "loss": 0.5891, | |
| "num_tokens": 194908296.0, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.7603686635944701, | |
| "grad_norm": 0.7927650818238564, | |
| "learning_rate": 1.171178751757535e-05, | |
| "loss": 0.5368, | |
| "num_tokens": 195384337.0, | |
| "step": 4455 | |
| }, | |
| { | |
| "epoch": 0.7612220515446322, | |
| "grad_norm": 0.8590390896353224, | |
| "learning_rate": 1.1666609853176342e-05, | |
| "loss": 0.5844, | |
| "num_tokens": 195804956.0, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.7620754394947943, | |
| "grad_norm": 0.8261475530402624, | |
| "learning_rate": 1.1621558292229268e-05, | |
| "loss": 0.5832, | |
| "num_tokens": 196260094.0, | |
| "step": 4465 | |
| }, | |
| { | |
| "epoch": 0.7629288274449565, | |
| "grad_norm": 0.7838091102369364, | |
| "learning_rate": 1.1576633193542797e-05, | |
| "loss": 0.5663, | |
| "num_tokens": 196766634.0, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.7637822153951186, | |
| "grad_norm": 0.8734185248220019, | |
| "learning_rate": 1.1531834914918365e-05, | |
| "loss": 0.5917, | |
| "num_tokens": 197264377.0, | |
| "step": 4475 | |
| }, | |
| { | |
| "epoch": 0.7646356033452808, | |
| "grad_norm": 0.7423950377840982, | |
| "learning_rate": 1.14871638131474e-05, | |
| "loss": 0.5446, | |
| "num_tokens": 197695417.0, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.7654889912954429, | |
| "grad_norm": 0.8525858260701448, | |
| "learning_rate": 1.1442620244008409e-05, | |
| "loss": 0.5807, | |
| "num_tokens": 198119324.0, | |
| "step": 4485 | |
| }, | |
| { | |
| "epoch": 0.766342379245605, | |
| "grad_norm": 0.75437978397281, | |
| "learning_rate": 1.1398204562264192e-05, | |
| "loss": 0.5745, | |
| "num_tokens": 198624663.0, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 0.7671957671957672, | |
| "grad_norm": 0.7874050783277934, | |
| "learning_rate": 1.1353917121659017e-05, | |
| "loss": 0.5672, | |
| "num_tokens": 199080528.0, | |
| "step": 4495 | |
| }, | |
| { | |
| "epoch": 0.7680491551459293, | |
| "grad_norm": 0.8607808840435229, | |
| "learning_rate": 1.1309758274915756e-05, | |
| "loss": 0.5774, | |
| "num_tokens": 199619482.0, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.7689025430960915, | |
| "grad_norm": 0.7816525166605554, | |
| "learning_rate": 1.1265728373733138e-05, | |
| "loss": 0.5837, | |
| "num_tokens": 200153610.0, | |
| "step": 4505 | |
| }, | |
| { | |
| "epoch": 0.7697559310462536, | |
| "grad_norm": 0.8119337918394042, | |
| "learning_rate": 1.1221827768782909e-05, | |
| "loss": 0.5649, | |
| "num_tokens": 200611222.0, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 0.7706093189964157, | |
| "grad_norm": 0.8798648885674402, | |
| "learning_rate": 1.1178056809707035e-05, | |
| "loss": 0.5703, | |
| "num_tokens": 201102129.0, | |
| "step": 4515 | |
| }, | |
| { | |
| "epoch": 0.7714627069465779, | |
| "grad_norm": 0.7970258353455149, | |
| "learning_rate": 1.1134415845114954e-05, | |
| "loss": 0.5595, | |
| "num_tokens": 201577891.0, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.77231609489674, | |
| "grad_norm": 0.8186504328550432, | |
| "learning_rate": 1.1090905222580756e-05, | |
| "loss": 0.5512, | |
| "num_tokens": 202031474.0, | |
| "step": 4525 | |
| }, | |
| { | |
| "epoch": 0.7731694828469022, | |
| "grad_norm": 0.8906791572152828, | |
| "learning_rate": 1.104752528864044e-05, | |
| "loss": 0.6207, | |
| "num_tokens": 202513129.0, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.7740228707970643, | |
| "grad_norm": 0.783806411579549, | |
| "learning_rate": 1.1004276388789146e-05, | |
| "loss": 0.5761, | |
| "num_tokens": 203027497.0, | |
| "step": 4535 | |
| }, | |
| { | |
| "epoch": 0.7748762587472265, | |
| "grad_norm": 0.773588946716583, | |
| "learning_rate": 1.096115886747842e-05, | |
| "loss": 0.5752, | |
| "num_tokens": 203502830.0, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.7757296466973886, | |
| "grad_norm": 0.8894811028102954, | |
| "learning_rate": 1.0918173068113446e-05, | |
| "loss": 0.5971, | |
| "num_tokens": 203985185.0, | |
| "step": 4545 | |
| }, | |
| { | |
| "epoch": 0.7765830346475507, | |
| "grad_norm": 0.8716160687877994, | |
| "learning_rate": 1.0875319333050315e-05, | |
| "loss": 0.572, | |
| "num_tokens": 204502494.0, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.777436422597713, | |
| "grad_norm": 0.8108686527858917, | |
| "learning_rate": 1.0832598003593325e-05, | |
| "loss": 0.555, | |
| "num_tokens": 204918625.0, | |
| "step": 4555 | |
| }, | |
| { | |
| "epoch": 0.778289810547875, | |
| "grad_norm": 0.8765591223692875, | |
| "learning_rate": 1.079000941999222e-05, | |
| "loss": 0.5606, | |
| "num_tokens": 205367508.0, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.7791431984980373, | |
| "grad_norm": 0.7740345225147104, | |
| "learning_rate": 1.0747553921439515e-05, | |
| "loss": 0.5662, | |
| "num_tokens": 205882344.0, | |
| "step": 4565 | |
| }, | |
| { | |
| "epoch": 0.7799965864481994, | |
| "grad_norm": 0.8585554136024073, | |
| "learning_rate": 1.0705231846067792e-05, | |
| "loss": 0.5968, | |
| "num_tokens": 206375735.0, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 0.7808499743983615, | |
| "grad_norm": 0.7706847829265646, | |
| "learning_rate": 1.0663043530946979e-05, | |
| "loss": 0.5787, | |
| "num_tokens": 206887531.0, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 0.7817033623485237, | |
| "grad_norm": 0.8490061354179466, | |
| "learning_rate": 1.0620989312081695e-05, | |
| "loss": 0.5658, | |
| "num_tokens": 207352866.0, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.7825567502986858, | |
| "grad_norm": 0.8906414696480907, | |
| "learning_rate": 1.0579069524408547e-05, | |
| "loss": 0.5704, | |
| "num_tokens": 207800778.0, | |
| "step": 4585 | |
| }, | |
| { | |
| "epoch": 0.783410138248848, | |
| "grad_norm": 0.8075876336778882, | |
| "learning_rate": 1.0537284501793502e-05, | |
| "loss": 0.542, | |
| "num_tokens": 208278578.0, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.7842635261990101, | |
| "grad_norm": 0.8861436743675016, | |
| "learning_rate": 1.0495634577029192e-05, | |
| "loss": 0.5987, | |
| "num_tokens": 208767476.0, | |
| "step": 4595 | |
| }, | |
| { | |
| "epoch": 0.7851169141491722, | |
| "grad_norm": 0.9121941445826955, | |
| "learning_rate": 1.045412008183227e-05, | |
| "loss": 0.5826, | |
| "num_tokens": 209296377.0, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.7859703020993344, | |
| "grad_norm": 0.8336097249773415, | |
| "learning_rate": 1.0412741346840793e-05, | |
| "loss": 0.5885, | |
| "num_tokens": 209881821.0, | |
| "step": 4605 | |
| }, | |
| { | |
| "epoch": 0.7868236900494965, | |
| "grad_norm": 0.853499272795292, | |
| "learning_rate": 1.037149870161154e-05, | |
| "loss": 0.6218, | |
| "num_tokens": 210371052.0, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 0.7876770779996587, | |
| "grad_norm": 0.78391018928751, | |
| "learning_rate": 1.0330392474617448e-05, | |
| "loss": 0.5802, | |
| "num_tokens": 210913400.0, | |
| "step": 4615 | |
| }, | |
| { | |
| "epoch": 0.7885304659498208, | |
| "grad_norm": 0.8886541706025143, | |
| "learning_rate": 1.0289422993244942e-05, | |
| "loss": 0.5863, | |
| "num_tokens": 211412557.0, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.7893838538999829, | |
| "grad_norm": 0.8510065551834175, | |
| "learning_rate": 1.0248590583791355e-05, | |
| "loss": 0.5917, | |
| "num_tokens": 211919254.0, | |
| "step": 4625 | |
| }, | |
| { | |
| "epoch": 0.7902372418501451, | |
| "grad_norm": 0.782054120335503, | |
| "learning_rate": 1.0207895571462337e-05, | |
| "loss": 0.5897, | |
| "num_tokens": 212411021.0, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 0.7910906298003072, | |
| "grad_norm": 0.9616075507807224, | |
| "learning_rate": 1.0167338280369233e-05, | |
| "loss": 0.5961, | |
| "num_tokens": 212845993.0, | |
| "step": 4635 | |
| }, | |
| { | |
| "epoch": 0.7919440177504694, | |
| "grad_norm": 0.8511972746806536, | |
| "learning_rate": 1.0126919033526536e-05, | |
| "loss": 0.5938, | |
| "num_tokens": 213335201.0, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.7927974057006315, | |
| "grad_norm": 0.8778820941483984, | |
| "learning_rate": 1.0086638152849298e-05, | |
| "loss": 0.585, | |
| "num_tokens": 213824239.0, | |
| "step": 4645 | |
| }, | |
| { | |
| "epoch": 0.7936507936507936, | |
| "grad_norm": 0.9536384542083954, | |
| "learning_rate": 1.0046495959150554e-05, | |
| "loss": 0.5723, | |
| "num_tokens": 214265326.0, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.7945041816009558, | |
| "grad_norm": 0.8088021837609581, | |
| "learning_rate": 1.0006492772138798e-05, | |
| "loss": 0.5743, | |
| "num_tokens": 214758689.0, | |
| "step": 4655 | |
| }, | |
| { | |
| "epoch": 0.7953575695511179, | |
| "grad_norm": 1.0221419584239615, | |
| "learning_rate": 9.966628910415413e-06, | |
| "loss": 0.5904, | |
| "num_tokens": 215164364.0, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.7962109575012801, | |
| "grad_norm": 0.7525718089356042, | |
| "learning_rate": 9.926904691472134e-06, | |
| "loss": 0.5867, | |
| "num_tokens": 215687696.0, | |
| "step": 4665 | |
| }, | |
| { | |
| "epoch": 0.7970643454514422, | |
| "grad_norm": 0.8226639021102814, | |
| "learning_rate": 9.887320431688521e-06, | |
| "loss": 0.5931, | |
| "num_tokens": 216209019.0, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 0.7979177334016043, | |
| "grad_norm": 0.8644761017616382, | |
| "learning_rate": 9.847876446329457e-06, | |
| "loss": 0.5637, | |
| "num_tokens": 216694502.0, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 0.7987711213517665, | |
| "grad_norm": 0.8286164858966968, | |
| "learning_rate": 9.808573049542627e-06, | |
| "loss": 0.5649, | |
| "num_tokens": 217141531.0, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.7996245093019286, | |
| "grad_norm": 0.778990278771165, | |
| "learning_rate": 9.76941055435599e-06, | |
| "loss": 0.5836, | |
| "num_tokens": 217585422.0, | |
| "step": 4685 | |
| }, | |
| { | |
| "epoch": 0.8004778972520908, | |
| "grad_norm": 0.8636912543908317, | |
| "learning_rate": 9.730389272675331e-06, | |
| "loss": 0.5756, | |
| "num_tokens": 218014891.0, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 0.8013312852022529, | |
| "grad_norm": 0.8904461215069976, | |
| "learning_rate": 9.691509515281738e-06, | |
| "loss": 0.6071, | |
| "num_tokens": 218492528.0, | |
| "step": 4695 | |
| }, | |
| { | |
| "epoch": 0.802184673152415, | |
| "grad_norm": 0.8795915282441338, | |
| "learning_rate": 9.652771591829156e-06, | |
| "loss": 0.5754, | |
| "num_tokens": 218954717.0, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.8030380611025772, | |
| "grad_norm": 0.9801298504817358, | |
| "learning_rate": 9.614175810841896e-06, | |
| "loss": 0.5862, | |
| "num_tokens": 219452296.0, | |
| "step": 4705 | |
| }, | |
| { | |
| "epoch": 0.8038914490527393, | |
| "grad_norm": 0.9097635974714854, | |
| "learning_rate": 9.57572247971219e-06, | |
| "loss": 0.616, | |
| "num_tokens": 219967343.0, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.8047448370029016, | |
| "grad_norm": 0.9048513040998128, | |
| "learning_rate": 9.53741190469776e-06, | |
| "loss": 0.619, | |
| "num_tokens": 220415584.0, | |
| "step": 4715 | |
| }, | |
| { | |
| "epoch": 0.8055982249530637, | |
| "grad_norm": 0.8720022920338458, | |
| "learning_rate": 9.499244390919335e-06, | |
| "loss": 0.5949, | |
| "num_tokens": 220976800.0, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.8064516129032258, | |
| "grad_norm": 0.874631764508416, | |
| "learning_rate": 9.461220242358268e-06, | |
| "loss": 0.5847, | |
| "num_tokens": 221488239.0, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 0.807305000853388, | |
| "grad_norm": 0.9258057592837162, | |
| "learning_rate": 9.42333976185409e-06, | |
| "loss": 0.5974, | |
| "num_tokens": 221955584.0, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 0.8081583888035501, | |
| "grad_norm": 0.8881679738093546, | |
| "learning_rate": 9.385603251102084e-06, | |
| "loss": 0.5706, | |
| "num_tokens": 222421916.0, | |
| "step": 4735 | |
| }, | |
| { | |
| "epoch": 0.8090117767537123, | |
| "grad_norm": 0.8639817388609756, | |
| "learning_rate": 9.348011010650937e-06, | |
| "loss": 0.5617, | |
| "num_tokens": 222840212.0, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.8098651647038744, | |
| "grad_norm": 0.7798641236952212, | |
| "learning_rate": 9.310563339900272e-06, | |
| "loss": 0.5672, | |
| "num_tokens": 223329381.0, | |
| "step": 4745 | |
| }, | |
| { | |
| "epoch": 0.8107185526540365, | |
| "grad_norm": 0.9982648728468079, | |
| "learning_rate": 9.273260537098315e-06, | |
| "loss": 0.5857, | |
| "num_tokens": 223856776.0, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.8115719406041987, | |
| "grad_norm": 0.8385007623825974, | |
| "learning_rate": 9.236102899339519e-06, | |
| "loss": 0.5767, | |
| "num_tokens": 224327062.0, | |
| "step": 4755 | |
| }, | |
| { | |
| "epoch": 0.8124253285543608, | |
| "grad_norm": 0.8360601966958555, | |
| "learning_rate": 9.199090722562156e-06, | |
| "loss": 0.5808, | |
| "num_tokens": 224773409.0, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.813278716504523, | |
| "grad_norm": 0.8512635733155746, | |
| "learning_rate": 9.162224301546025e-06, | |
| "loss": 0.5538, | |
| "num_tokens": 225233757.0, | |
| "step": 4765 | |
| }, | |
| { | |
| "epoch": 0.8141321044546851, | |
| "grad_norm": 0.8816024503899602, | |
| "learning_rate": 9.125503929910035e-06, | |
| "loss": 0.5847, | |
| "num_tokens": 225737840.0, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 0.8149854924048472, | |
| "grad_norm": 0.8511656206892778, | |
| "learning_rate": 9.08892990010992e-06, | |
| "loss": 0.5321, | |
| "num_tokens": 226192735.0, | |
| "step": 4775 | |
| }, | |
| { | |
| "epoch": 0.8158388803550094, | |
| "grad_norm": 0.8822874517324099, | |
| "learning_rate": 9.052502503435873e-06, | |
| "loss": 0.5657, | |
| "num_tokens": 226708301.0, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.8166922683051715, | |
| "grad_norm": 1.4766944705049547, | |
| "learning_rate": 9.016222030010259e-06, | |
| "loss": 0.5858, | |
| "num_tokens": 227130751.0, | |
| "step": 4785 | |
| }, | |
| { | |
| "epoch": 0.8175456562553337, | |
| "grad_norm": 0.9592968439333777, | |
| "learning_rate": 8.980088768785271e-06, | |
| "loss": 0.5858, | |
| "num_tokens": 227615269.0, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 0.8183990442054958, | |
| "grad_norm": 0.8141808313694623, | |
| "learning_rate": 8.94410300754067e-06, | |
| "loss": 0.5629, | |
| "num_tokens": 228124400.0, | |
| "step": 4795 | |
| }, | |
| { | |
| "epoch": 0.819252432155658, | |
| "grad_norm": 0.8149199518274812, | |
| "learning_rate": 8.908265032881438e-06, | |
| "loss": 0.6274, | |
| "num_tokens": 228615374.0, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.8201058201058201, | |
| "grad_norm": 0.7908437674204951, | |
| "learning_rate": 8.872575130235533e-06, | |
| "loss": 0.5894, | |
| "num_tokens": 229119350.0, | |
| "step": 4805 | |
| }, | |
| { | |
| "epoch": 0.8209592080559822, | |
| "grad_norm": 0.8506638324105442, | |
| "learning_rate": 8.837033583851625e-06, | |
| "loss": 0.6181, | |
| "num_tokens": 229579333.0, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 0.8218125960061444, | |
| "grad_norm": 0.9547144252263968, | |
| "learning_rate": 8.801640676796811e-06, | |
| "loss": 0.57, | |
| "num_tokens": 230025206.0, | |
| "step": 4815 | |
| }, | |
| { | |
| "epoch": 0.8226659839563065, | |
| "grad_norm": 0.8678414895568546, | |
| "learning_rate": 8.76639669095434e-06, | |
| "loss": 0.6118, | |
| "num_tokens": 230517205.0, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.8235193719064687, | |
| "grad_norm": 0.7943130767209409, | |
| "learning_rate": 8.73130190702143e-06, | |
| "loss": 0.5543, | |
| "num_tokens": 230934827.0, | |
| "step": 4825 | |
| }, | |
| { | |
| "epoch": 0.8243727598566308, | |
| "grad_norm": 0.9543452649607513, | |
| "learning_rate": 8.696356604506964e-06, | |
| "loss": 0.6102, | |
| "num_tokens": 231385653.0, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 0.8252261478067929, | |
| "grad_norm": 0.8397955922079047, | |
| "learning_rate": 8.66156106172932e-06, | |
| "loss": 0.5431, | |
| "num_tokens": 231886693.0, | |
| "step": 4835 | |
| }, | |
| { | |
| "epoch": 0.8260795357569551, | |
| "grad_norm": 0.8996229035875397, | |
| "learning_rate": 8.62691555581411e-06, | |
| "loss": 0.55, | |
| "num_tokens": 232345381.0, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.8269329237071172, | |
| "grad_norm": 0.8009699462259144, | |
| "learning_rate": 8.592420362691994e-06, | |
| "loss": 0.5681, | |
| "num_tokens": 232881440.0, | |
| "step": 4845 | |
| }, | |
| { | |
| "epoch": 0.8277863116572794, | |
| "grad_norm": 0.8342452775283754, | |
| "learning_rate": 8.558075757096502e-06, | |
| "loss": 0.5901, | |
| "num_tokens": 233383237.0, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.8286396996074415, | |
| "grad_norm": 0.7805241990962766, | |
| "learning_rate": 8.523882012561792e-06, | |
| "loss": 0.5532, | |
| "num_tokens": 233853374.0, | |
| "step": 4855 | |
| }, | |
| { | |
| "epoch": 0.8294930875576036, | |
| "grad_norm": 0.8964409321897953, | |
| "learning_rate": 8.489839401420538e-06, | |
| "loss": 0.5426, | |
| "num_tokens": 234278092.0, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.8303464755077659, | |
| "grad_norm": 0.8811351414159762, | |
| "learning_rate": 8.455948194801706e-06, | |
| "loss": 0.5938, | |
| "num_tokens": 234751112.0, | |
| "step": 4865 | |
| }, | |
| { | |
| "epoch": 0.831199863457928, | |
| "grad_norm": 0.7503250035269912, | |
| "learning_rate": 8.422208662628415e-06, | |
| "loss": 0.5571, | |
| "num_tokens": 235269180.0, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 0.8320532514080902, | |
| "grad_norm": 0.8600870648780747, | |
| "learning_rate": 8.388621073615803e-06, | |
| "loss": 0.6083, | |
| "num_tokens": 235766057.0, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 0.8329066393582523, | |
| "grad_norm": 0.7556334243138391, | |
| "learning_rate": 8.355185695268858e-06, | |
| "loss": 0.5586, | |
| "num_tokens": 236263389.0, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.8337600273084144, | |
| "grad_norm": 0.7869421912173387, | |
| "learning_rate": 8.321902793880301e-06, | |
| "loss": 0.5719, | |
| "num_tokens": 236705262.0, | |
| "step": 4885 | |
| }, | |
| { | |
| "epoch": 0.8346134152585766, | |
| "grad_norm": 0.7798357656578301, | |
| "learning_rate": 8.28877263452848e-06, | |
| "loss": 0.6194, | |
| "num_tokens": 237259536.0, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 0.8354668032087387, | |
| "grad_norm": 0.767779208940415, | |
| "learning_rate": 8.255795481075228e-06, | |
| "loss": 0.5648, | |
| "num_tokens": 237769080.0, | |
| "step": 4895 | |
| }, | |
| { | |
| "epoch": 0.8363201911589009, | |
| "grad_norm": 0.9854382778495104, | |
| "learning_rate": 8.222971596163792e-06, | |
| "loss": 0.5442, | |
| "num_tokens": 238251727.0, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.837173579109063, | |
| "grad_norm": 0.833319716288937, | |
| "learning_rate": 8.19030124121671e-06, | |
| "loss": 0.5461, | |
| "num_tokens": 238648424.0, | |
| "step": 4905 | |
| }, | |
| { | |
| "epoch": 0.8380269670592251, | |
| "grad_norm": 0.854597338972544, | |
| "learning_rate": 8.157784676433764e-06, | |
| "loss": 0.571, | |
| "num_tokens": 239134618.0, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 0.8388803550093873, | |
| "grad_norm": 0.8500312839651374, | |
| "learning_rate": 8.125422160789878e-06, | |
| "loss": 0.5711, | |
| "num_tokens": 239603031.0, | |
| "step": 4915 | |
| }, | |
| { | |
| "epoch": 0.8397337429595494, | |
| "grad_norm": 0.7630165552793707, | |
| "learning_rate": 8.093213952033072e-06, | |
| "loss": 0.5828, | |
| "num_tokens": 240081011.0, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.8405871309097116, | |
| "grad_norm": 0.7564955060386162, | |
| "learning_rate": 8.061160306682406e-06, | |
| "loss": 0.57, | |
| "num_tokens": 240536075.0, | |
| "step": 4925 | |
| }, | |
| { | |
| "epoch": 0.8414405188598737, | |
| "grad_norm": 0.8295745453389126, | |
| "learning_rate": 8.029261480025922e-06, | |
| "loss": 0.5609, | |
| "num_tokens": 240964799.0, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 0.8422939068100358, | |
| "grad_norm": 0.8572743861594461, | |
| "learning_rate": 7.997517726118644e-06, | |
| "loss": 0.5735, | |
| "num_tokens": 241438662.0, | |
| "step": 4935 | |
| }, | |
| { | |
| "epoch": 0.843147294760198, | |
| "grad_norm": 0.8689601209375987, | |
| "learning_rate": 7.965929297780515e-06, | |
| "loss": 0.5735, | |
| "num_tokens": 241906415.0, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.8440006827103601, | |
| "grad_norm": 0.8096196074976684, | |
| "learning_rate": 7.934496446594417e-06, | |
| "loss": 0.5712, | |
| "num_tokens": 242373880.0, | |
| "step": 4945 | |
| }, | |
| { | |
| "epoch": 0.8448540706605223, | |
| "grad_norm": 0.7535492611876398, | |
| "learning_rate": 7.903219422904158e-06, | |
| "loss": 0.5612, | |
| "num_tokens": 242864038.0, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.8457074586106844, | |
| "grad_norm": 0.7504541427432688, | |
| "learning_rate": 7.87209847581245e-06, | |
| "loss": 0.5956, | |
| "num_tokens": 243431736.0, | |
| "step": 4955 | |
| }, | |
| { | |
| "epoch": 0.8465608465608465, | |
| "grad_norm": 0.7428325861767835, | |
| "learning_rate": 7.841133853178975e-06, | |
| "loss": 0.5354, | |
| "num_tokens": 243860094.0, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.8474142345110087, | |
| "grad_norm": 0.7861488676367593, | |
| "learning_rate": 7.810325801618365e-06, | |
| "loss": 0.6067, | |
| "num_tokens": 244395207.0, | |
| "step": 4965 | |
| }, | |
| { | |
| "epoch": 0.8482676224611708, | |
| "grad_norm": 0.7669494987596239, | |
| "learning_rate": 7.779674566498263e-06, | |
| "loss": 0.5757, | |
| "num_tokens": 244897324.0, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 0.849121010411333, | |
| "grad_norm": 0.7909377762889123, | |
| "learning_rate": 7.749180391937372e-06, | |
| "loss": 0.5762, | |
| "num_tokens": 245340722.0, | |
| "step": 4975 | |
| }, | |
| { | |
| "epoch": 0.8499743983614951, | |
| "grad_norm": 0.7883324294633338, | |
| "learning_rate": 7.718843520803487e-06, | |
| "loss": 0.572, | |
| "num_tokens": 245808193.0, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.8508277863116572, | |
| "grad_norm": 0.7826348291990151, | |
| "learning_rate": 7.688664194711592e-06, | |
| "loss": 0.5989, | |
| "num_tokens": 246277986.0, | |
| "step": 4985 | |
| }, | |
| { | |
| "epoch": 0.8516811742618194, | |
| "grad_norm": 0.8490644704046821, | |
| "learning_rate": 7.658642654021904e-06, | |
| "loss": 0.5914, | |
| "num_tokens": 246691850.0, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 0.8525345622119815, | |
| "grad_norm": 1.0022405608345875, | |
| "learning_rate": 7.628779137837981e-06, | |
| "loss": 0.58, | |
| "num_tokens": 247203770.0, | |
| "step": 4995 | |
| }, | |
| { | |
| "epoch": 0.8533879501621437, | |
| "grad_norm": 0.8084244306668955, | |
| "learning_rate": 7.5990738840048174e-06, | |
| "loss": 0.5894, | |
| "num_tokens": 247690866.0, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.8542413381123058, | |
| "grad_norm": 0.8985497283703163, | |
| "learning_rate": 7.569527129106931e-06, | |
| "loss": 0.5861, | |
| "num_tokens": 248161325.0, | |
| "step": 5005 | |
| }, | |
| { | |
| "epoch": 0.855094726062468, | |
| "grad_norm": 0.7873859767195008, | |
| "learning_rate": 7.540139108466496e-06, | |
| "loss": 0.588, | |
| "num_tokens": 248652400.0, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 0.8559481140126302, | |
| "grad_norm": 0.819987556398134, | |
| "learning_rate": 7.510910056141456e-06, | |
| "loss": 0.5709, | |
| "num_tokens": 249134729.0, | |
| "step": 5015 | |
| }, | |
| { | |
| "epoch": 0.8568015019627923, | |
| "grad_norm": 0.7492864937367162, | |
| "learning_rate": 7.481840204923681e-06, | |
| "loss": 0.585, | |
| "num_tokens": 249661501.0, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 0.8576548899129545, | |
| "grad_norm": 0.8040792030511285, | |
| "learning_rate": 7.452929786337096e-06, | |
| "loss": 0.5965, | |
| "num_tokens": 250217631.0, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 0.8585082778631166, | |
| "grad_norm": 0.900081931997727, | |
| "learning_rate": 7.424179030635831e-06, | |
| "loss": 0.5641, | |
| "num_tokens": 250690443.0, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 0.8593616658132787, | |
| "grad_norm": 0.8160666402757599, | |
| "learning_rate": 7.395588166802412e-06, | |
| "loss": 0.5955, | |
| "num_tokens": 251173634.0, | |
| "step": 5035 | |
| }, | |
| { | |
| "epoch": 0.8602150537634409, | |
| "grad_norm": 0.8391441098982425, | |
| "learning_rate": 7.367157422545904e-06, | |
| "loss": 0.5855, | |
| "num_tokens": 251638405.0, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 0.861068441713603, | |
| "grad_norm": 0.7940539632042635, | |
| "learning_rate": 7.338887024300134e-06, | |
| "loss": 0.5572, | |
| "num_tokens": 252139392.0, | |
| "step": 5045 | |
| }, | |
| { | |
| "epoch": 0.8619218296637652, | |
| "grad_norm": 0.7894583290273409, | |
| "learning_rate": 7.310777197221854e-06, | |
| "loss": 0.5523, | |
| "num_tokens": 252543885.0, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.8627752176139273, | |
| "grad_norm": 0.7911532831982829, | |
| "learning_rate": 7.282828165188976e-06, | |
| "loss": 0.5868, | |
| "num_tokens": 253005760.0, | |
| "step": 5055 | |
| }, | |
| { | |
| "epoch": 0.8636286055640894, | |
| "grad_norm": 0.8585063316932868, | |
| "learning_rate": 7.255040150798771e-06, | |
| "loss": 0.5848, | |
| "num_tokens": 253509550.0, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 0.8644819935142516, | |
| "grad_norm": 0.7240412622759334, | |
| "learning_rate": 7.227413375366089e-06, | |
| "loss": 0.5845, | |
| "num_tokens": 254008860.0, | |
| "step": 5065 | |
| }, | |
| { | |
| "epoch": 0.8653353814644137, | |
| "grad_norm": 0.8250225040930351, | |
| "learning_rate": 7.199948058921629e-06, | |
| "loss": 0.6073, | |
| "num_tokens": 254522809.0, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 0.8661887694145759, | |
| "grad_norm": 0.9326895880461082, | |
| "learning_rate": 7.1726444202101535e-06, | |
| "loss": 0.6093, | |
| "num_tokens": 255005160.0, | |
| "step": 5075 | |
| }, | |
| { | |
| "epoch": 0.867042157364738, | |
| "grad_norm": 0.7994442628252905, | |
| "learning_rate": 7.145502676688759e-06, | |
| "loss": 0.5745, | |
| "num_tokens": 255461423.0, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 0.8678955453149002, | |
| "grad_norm": 0.9707923615366039, | |
| "learning_rate": 7.1185230445251535e-06, | |
| "loss": 0.5934, | |
| "num_tokens": 255944452.0, | |
| "step": 5085 | |
| }, | |
| { | |
| "epoch": 0.8687489332650623, | |
| "grad_norm": 0.796895516510252, | |
| "learning_rate": 7.091705738595911e-06, | |
| "loss": 0.5487, | |
| "num_tokens": 256455794.0, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 0.8696023212152244, | |
| "grad_norm": 0.7400404717908586, | |
| "learning_rate": 7.065050972484788e-06, | |
| "loss": 0.5577, | |
| "num_tokens": 256935670.0, | |
| "step": 5095 | |
| }, | |
| { | |
| "epoch": 0.8704557091653866, | |
| "grad_norm": 0.7988037904803611, | |
| "learning_rate": 7.038558958481001e-06, | |
| "loss": 0.5727, | |
| "num_tokens": 257407700.0, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.8713090971155487, | |
| "grad_norm": 0.9147730742138965, | |
| "learning_rate": 7.01222990757754e-06, | |
| "loss": 0.5934, | |
| "num_tokens": 257838492.0, | |
| "step": 5105 | |
| }, | |
| { | |
| "epoch": 0.8721624850657109, | |
| "grad_norm": 0.7135980779624871, | |
| "learning_rate": 6.986064029469508e-06, | |
| "loss": 0.5336, | |
| "num_tokens": 258365498.0, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 0.873015873015873, | |
| "grad_norm": 0.7544369621525995, | |
| "learning_rate": 6.9600615325524115e-06, | |
| "loss": 0.559, | |
| "num_tokens": 258844796.0, | |
| "step": 5115 | |
| }, | |
| { | |
| "epoch": 0.8738692609660351, | |
| "grad_norm": 0.7444148640752304, | |
| "learning_rate": 6.934222623920547e-06, | |
| "loss": 0.5602, | |
| "num_tokens": 259349061.0, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 0.8747226489161973, | |
| "grad_norm": 0.9049523089278464, | |
| "learning_rate": 6.908547509365305e-06, | |
| "loss": 0.5395, | |
| "num_tokens": 259756016.0, | |
| "step": 5125 | |
| }, | |
| { | |
| "epoch": 0.8755760368663594, | |
| "grad_norm": 0.7527681089262206, | |
| "learning_rate": 6.883036393373579e-06, | |
| "loss": 0.5912, | |
| "num_tokens": 260313344.0, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 0.8764294248165216, | |
| "grad_norm": 0.9028818740280382, | |
| "learning_rate": 6.857689479126099e-06, | |
| "loss": 0.5528, | |
| "num_tokens": 260737923.0, | |
| "step": 5135 | |
| }, | |
| { | |
| "epoch": 0.8772828127666837, | |
| "grad_norm": 0.793355764631848, | |
| "learning_rate": 6.8325069684958235e-06, | |
| "loss": 0.5845, | |
| "num_tokens": 261263489.0, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 0.8781362007168458, | |
| "grad_norm": 0.785894856367147, | |
| "learning_rate": 6.8074890620463394e-06, | |
| "loss": 0.5578, | |
| "num_tokens": 261788584.0, | |
| "step": 5145 | |
| }, | |
| { | |
| "epoch": 0.878989588667008, | |
| "grad_norm": 0.8143390310414678, | |
| "learning_rate": 6.782635959030259e-06, | |
| "loss": 0.5688, | |
| "num_tokens": 262287010.0, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.8798429766171701, | |
| "grad_norm": 0.865141361269142, | |
| "learning_rate": 6.7579478573876366e-06, | |
| "loss": 0.5842, | |
| "num_tokens": 262746009.0, | |
| "step": 5155 | |
| }, | |
| { | |
| "epoch": 0.8806963645673324, | |
| "grad_norm": 0.9092110398635419, | |
| "learning_rate": 6.733424953744391e-06, | |
| "loss": 0.5753, | |
| "num_tokens": 263182542.0, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 0.8815497525174945, | |
| "grad_norm": 0.821394284330915, | |
| "learning_rate": 6.709067443410733e-06, | |
| "loss": 0.5951, | |
| "num_tokens": 263729263.0, | |
| "step": 5165 | |
| }, | |
| { | |
| "epoch": 0.8824031404676566, | |
| "grad_norm": 0.7488745968684951, | |
| "learning_rate": 6.684875520379618e-06, | |
| "loss": 0.5433, | |
| "num_tokens": 264227619.0, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 0.8832565284178188, | |
| "grad_norm": 0.976772151141878, | |
| "learning_rate": 6.66084937732519e-06, | |
| "loss": 0.6446, | |
| "num_tokens": 264718377.0, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 0.8841099163679809, | |
| "grad_norm": 0.7642964720283203, | |
| "learning_rate": 6.636989205601276e-06, | |
| "loss": 0.5897, | |
| "num_tokens": 265270188.0, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 0.8849633043181431, | |
| "grad_norm": 0.8028339592698639, | |
| "learning_rate": 6.613295195239816e-06, | |
| "loss": 0.561, | |
| "num_tokens": 265753973.0, | |
| "step": 5185 | |
| }, | |
| { | |
| "epoch": 0.8858166922683052, | |
| "grad_norm": 0.7776641112247487, | |
| "learning_rate": 6.589767534949384e-06, | |
| "loss": 0.5511, | |
| "num_tokens": 266247179.0, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 0.8866700802184673, | |
| "grad_norm": 0.7276998937386446, | |
| "learning_rate": 6.5664064121136865e-06, | |
| "loss": 0.5465, | |
| "num_tokens": 266736076.0, | |
| "step": 5195 | |
| }, | |
| { | |
| "epoch": 0.8875234681686295, | |
| "grad_norm": 0.8212452792937524, | |
| "learning_rate": 6.543212012790038e-06, | |
| "loss": 0.5752, | |
| "num_tokens": 267195187.0, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.8883768561187916, | |
| "grad_norm": 0.7939352940851623, | |
| "learning_rate": 6.520184521707923e-06, | |
| "loss": 0.5677, | |
| "num_tokens": 267663496.0, | |
| "step": 5205 | |
| }, | |
| { | |
| "epoch": 0.8892302440689538, | |
| "grad_norm": 0.9467450707219625, | |
| "learning_rate": 6.497324122267478e-06, | |
| "loss": 0.6117, | |
| "num_tokens": 268161021.0, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 0.8900836320191159, | |
| "grad_norm": 0.8245860155572584, | |
| "learning_rate": 6.474630996538078e-06, | |
| "loss": 0.5591, | |
| "num_tokens": 268673121.0, | |
| "step": 5215 | |
| }, | |
| { | |
| "epoch": 0.890937019969278, | |
| "grad_norm": 0.7590094334951016, | |
| "learning_rate": 6.452105325256852e-06, | |
| "loss": 0.568, | |
| "num_tokens": 269179279.0, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 0.8917904079194402, | |
| "grad_norm": 0.8217521579318463, | |
| "learning_rate": 6.429747287827254e-06, | |
| "loss": 0.5807, | |
| "num_tokens": 269693196.0, | |
| "step": 5225 | |
| }, | |
| { | |
| "epoch": 0.8926437958696023, | |
| "grad_norm": 0.8623514216744014, | |
| "learning_rate": 6.407557062317632e-06, | |
| "loss": 0.5582, | |
| "num_tokens": 270166668.0, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 0.8934971838197645, | |
| "grad_norm": 0.8533922181449645, | |
| "learning_rate": 6.38553482545982e-06, | |
| "loss": 0.5716, | |
| "num_tokens": 270670889.0, | |
| "step": 5235 | |
| }, | |
| { | |
| "epoch": 0.8943505717699266, | |
| "grad_norm": 0.7713411612044356, | |
| "learning_rate": 6.36368075264772e-06, | |
| "loss": 0.5555, | |
| "num_tokens": 271154055.0, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 0.8952039597200887, | |
| "grad_norm": 0.7525851117026119, | |
| "learning_rate": 6.341995017935916e-06, | |
| "loss": 0.5926, | |
| "num_tokens": 271718782.0, | |
| "step": 5245 | |
| }, | |
| { | |
| "epoch": 0.8960573476702509, | |
| "grad_norm": 0.8380438697228516, | |
| "learning_rate": 6.320477794038258e-06, | |
| "loss": 0.5639, | |
| "num_tokens": 272179817.0, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.896910735620413, | |
| "grad_norm": 0.7363618546018679, | |
| "learning_rate": 6.299129252326541e-06, | |
| "loss": 0.6016, | |
| "num_tokens": 272681025.0, | |
| "step": 5255 | |
| }, | |
| { | |
| "epoch": 0.8977641235705752, | |
| "grad_norm": 0.964225303945823, | |
| "learning_rate": 6.277949562829075e-06, | |
| "loss": 0.5891, | |
| "num_tokens": 273165716.0, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 0.8986175115207373, | |
| "grad_norm": 0.806927745635919, | |
| "learning_rate": 6.256938894229389e-06, | |
| "loss": 0.5705, | |
| "num_tokens": 273715683.0, | |
| "step": 5265 | |
| }, | |
| { | |
| "epoch": 0.8994708994708994, | |
| "grad_norm": 0.7821754782396905, | |
| "learning_rate": 6.236097413864841e-06, | |
| "loss": 0.5697, | |
| "num_tokens": 274153339.0, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 0.9003242874210616, | |
| "grad_norm": 0.9491365135771385, | |
| "learning_rate": 6.215425287725328e-06, | |
| "loss": 0.5629, | |
| "num_tokens": 274622449.0, | |
| "step": 5275 | |
| }, | |
| { | |
| "epoch": 0.9011776753712237, | |
| "grad_norm": 0.8490653535475988, | |
| "learning_rate": 6.194922680451922e-06, | |
| "loss": 0.5641, | |
| "num_tokens": 275099879.0, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 0.9020310633213859, | |
| "grad_norm": 0.7060635522815266, | |
| "learning_rate": 6.17458975533559e-06, | |
| "loss": 0.5723, | |
| "num_tokens": 275615026.0, | |
| "step": 5285 | |
| }, | |
| { | |
| "epoch": 0.902884451271548, | |
| "grad_norm": 0.8929218635182984, | |
| "learning_rate": 6.1544266743158805e-06, | |
| "loss": 0.5758, | |
| "num_tokens": 276105435.0, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 0.9037378392217101, | |
| "grad_norm": 0.8900936194728736, | |
| "learning_rate": 6.134433597979634e-06, | |
| "loss": 0.5814, | |
| "num_tokens": 276567753.0, | |
| "step": 5295 | |
| }, | |
| { | |
| "epoch": 0.9045912271718723, | |
| "grad_norm": 0.7450155592638278, | |
| "learning_rate": 6.114610685559708e-06, | |
| "loss": 0.5492, | |
| "num_tokens": 277067496.0, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.9054446151220344, | |
| "grad_norm": 0.822036818022831, | |
| "learning_rate": 6.09495809493371e-06, | |
| "loss": 0.5945, | |
| "num_tokens": 277531967.0, | |
| "step": 5305 | |
| }, | |
| { | |
| "epoch": 0.9062980030721967, | |
| "grad_norm": 0.7866502345790104, | |
| "learning_rate": 6.0754759826227225e-06, | |
| "loss": 0.5558, | |
| "num_tokens": 278023533.0, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 0.9071513910223588, | |
| "grad_norm": 0.8423352400084989, | |
| "learning_rate": 6.056164503790092e-06, | |
| "loss": 0.5801, | |
| "num_tokens": 278474312.0, | |
| "step": 5315 | |
| }, | |
| { | |
| "epoch": 0.9080047789725209, | |
| "grad_norm": 0.9494363071010109, | |
| "learning_rate": 6.0370238122401495e-06, | |
| "loss": 0.5597, | |
| "num_tokens": 278896745.0, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 0.9088581669226831, | |
| "grad_norm": 0.8282491259672817, | |
| "learning_rate": 6.01805406041702e-06, | |
| "loss": 0.5626, | |
| "num_tokens": 279349709.0, | |
| "step": 5325 | |
| }, | |
| { | |
| "epoch": 0.9097115548728452, | |
| "grad_norm": 0.8108165719871948, | |
| "learning_rate": 5.999255399403401e-06, | |
| "loss": 0.5843, | |
| "num_tokens": 279854864.0, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 0.9105649428230074, | |
| "grad_norm": 1.368074602666708, | |
| "learning_rate": 5.980627978919339e-06, | |
| "loss": 0.573, | |
| "num_tokens": 280321831.0, | |
| "step": 5335 | |
| }, | |
| { | |
| "epoch": 0.9114183307731695, | |
| "grad_norm": 0.865994558880661, | |
| "learning_rate": 5.962171947321067e-06, | |
| "loss": 0.5722, | |
| "num_tokens": 280833419.0, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 0.9122717187233317, | |
| "grad_norm": 0.8067476350501013, | |
| "learning_rate": 5.943887451599798e-06, | |
| "loss": 0.5628, | |
| "num_tokens": 281281758.0, | |
| "step": 5345 | |
| }, | |
| { | |
| "epoch": 0.9131251066734938, | |
| "grad_norm": 0.7854491662288046, | |
| "learning_rate": 5.925774637380573e-06, | |
| "loss": 0.5867, | |
| "num_tokens": 281796725.0, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.9139784946236559, | |
| "grad_norm": 0.7889550827339223, | |
| "learning_rate": 5.9078336489210895e-06, | |
| "loss": 0.5945, | |
| "num_tokens": 282294087.0, | |
| "step": 5355 | |
| }, | |
| { | |
| "epoch": 0.9148318825738181, | |
| "grad_norm": 0.8663650736244458, | |
| "learning_rate": 5.890064629110552e-06, | |
| "loss": 0.6051, | |
| "num_tokens": 282763597.0, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 0.9156852705239802, | |
| "grad_norm": 0.8701454268029402, | |
| "learning_rate": 5.8724677194685435e-06, | |
| "loss": 0.5515, | |
| "num_tokens": 283196267.0, | |
| "step": 5365 | |
| }, | |
| { | |
| "epoch": 0.9165386584741424, | |
| "grad_norm": 0.7957477705163372, | |
| "learning_rate": 5.855043060143887e-06, | |
| "loss": 0.5779, | |
| "num_tokens": 283679766.0, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 0.9173920464243045, | |
| "grad_norm": 0.7978470026361059, | |
| "learning_rate": 5.83779078991354e-06, | |
| "loss": 0.5329, | |
| "num_tokens": 284144403.0, | |
| "step": 5375 | |
| }, | |
| { | |
| "epoch": 0.9182454343744666, | |
| "grad_norm": 0.8716938721156829, | |
| "learning_rate": 5.820711046181488e-06, | |
| "loss": 0.5652, | |
| "num_tokens": 284612089.0, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 0.9190988223246288, | |
| "grad_norm": 0.7905417532595569, | |
| "learning_rate": 5.803803964977634e-06, | |
| "loss": 0.5381, | |
| "num_tokens": 285094795.0, | |
| "step": 5385 | |
| }, | |
| { | |
| "epoch": 0.9199522102747909, | |
| "grad_norm": 0.8101830629318585, | |
| "learning_rate": 5.7870696809567425e-06, | |
| "loss": 0.5626, | |
| "num_tokens": 285541742.0, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 0.9208055982249531, | |
| "grad_norm": 0.8322898676886099, | |
| "learning_rate": 5.770508327397339e-06, | |
| "loss": 0.587, | |
| "num_tokens": 286021840.0, | |
| "step": 5395 | |
| }, | |
| { | |
| "epoch": 0.9216589861751152, | |
| "grad_norm": 0.8824032296667493, | |
| "learning_rate": 5.754120036200669e-06, | |
| "loss": 0.6291, | |
| "num_tokens": 286505617.0, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.9225123741252773, | |
| "grad_norm": 0.7998666514749609, | |
| "learning_rate": 5.7379049378896406e-06, | |
| "loss": 0.548, | |
| "num_tokens": 286948538.0, | |
| "step": 5405 | |
| }, | |
| { | |
| "epoch": 0.9233657620754395, | |
| "grad_norm": 0.9057331244599551, | |
| "learning_rate": 5.721863161607775e-06, | |
| "loss": 0.5991, | |
| "num_tokens": 287462501.0, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 0.9242191500256016, | |
| "grad_norm": 0.8123865828941764, | |
| "learning_rate": 5.705994835118203e-06, | |
| "loss": 0.607, | |
| "num_tokens": 287916310.0, | |
| "step": 5415 | |
| }, | |
| { | |
| "epoch": 0.9250725379757638, | |
| "grad_norm": 0.8199261801620987, | |
| "learning_rate": 5.6903000848026165e-06, | |
| "loss": 0.5809, | |
| "num_tokens": 288427512.0, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 0.9259259259259259, | |
| "grad_norm": 0.7633767277902096, | |
| "learning_rate": 5.674779035660291e-06, | |
| "loss": 0.5555, | |
| "num_tokens": 288889633.0, | |
| "step": 5425 | |
| }, | |
| { | |
| "epoch": 0.926779313876088, | |
| "grad_norm": 0.7638580491242525, | |
| "learning_rate": 5.659431811307065e-06, | |
| "loss": 0.5673, | |
| "num_tokens": 289335039.0, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 0.9276327018262502, | |
| "grad_norm": 0.7529647509629525, | |
| "learning_rate": 5.644258533974374e-06, | |
| "loss": 0.5793, | |
| "num_tokens": 289786616.0, | |
| "step": 5435 | |
| }, | |
| { | |
| "epoch": 0.9284860897764123, | |
| "grad_norm": 0.836400859683071, | |
| "learning_rate": 5.629259324508267e-06, | |
| "loss": 0.5867, | |
| "num_tokens": 290255858.0, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 0.9293394777265745, | |
| "grad_norm": 0.8755510963029428, | |
| "learning_rate": 5.614434302368449e-06, | |
| "loss": 0.5546, | |
| "num_tokens": 290698740.0, | |
| "step": 5445 | |
| }, | |
| { | |
| "epoch": 0.9301928656767366, | |
| "grad_norm": 0.8332440496913869, | |
| "learning_rate": 5.599783585627322e-06, | |
| "loss": 0.5484, | |
| "num_tokens": 291138185.0, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.9310462536268987, | |
| "grad_norm": 0.8426341821246235, | |
| "learning_rate": 5.585307290969054e-06, | |
| "loss": 0.5719, | |
| "num_tokens": 291643871.0, | |
| "step": 5455 | |
| }, | |
| { | |
| "epoch": 0.931899641577061, | |
| "grad_norm": 0.8678942218248095, | |
| "learning_rate": 5.571005533688649e-06, | |
| "loss": 0.5553, | |
| "num_tokens": 292113768.0, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 0.932753029527223, | |
| "grad_norm": 0.7594279684222428, | |
| "learning_rate": 5.556878427691023e-06, | |
| "loss": 0.5625, | |
| "num_tokens": 292583353.0, | |
| "step": 5465 | |
| }, | |
| { | |
| "epoch": 0.9336064174773853, | |
| "grad_norm": 0.7842725351760226, | |
| "learning_rate": 5.542926085490093e-06, | |
| "loss": 0.5387, | |
| "num_tokens": 293080307.0, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 0.9344598054275474, | |
| "grad_norm": 0.826007070503848, | |
| "learning_rate": 5.529148618207897e-06, | |
| "loss": 0.5479, | |
| "num_tokens": 293565490.0, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 0.9353131933777095, | |
| "grad_norm": 0.8597332041464362, | |
| "learning_rate": 5.515546135573695e-06, | |
| "loss": 0.6213, | |
| "num_tokens": 294083350.0, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 0.9361665813278717, | |
| "grad_norm": 0.8219571384073128, | |
| "learning_rate": 5.5021187459230964e-06, | |
| "loss": 0.5651, | |
| "num_tokens": 294552165.0, | |
| "step": 5485 | |
| }, | |
| { | |
| "epoch": 0.9370199692780338, | |
| "grad_norm": 0.8559937062906311, | |
| "learning_rate": 5.4888665561972065e-06, | |
| "loss": 0.5765, | |
| "num_tokens": 294991459.0, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 0.937873357228196, | |
| "grad_norm": 0.7579381758259627, | |
| "learning_rate": 5.475789671941761e-06, | |
| "loss": 0.6216, | |
| "num_tokens": 295566010.0, | |
| "step": 5495 | |
| }, | |
| { | |
| "epoch": 0.9387267451783581, | |
| "grad_norm": 0.7621531478864432, | |
| "learning_rate": 5.462888197306301e-06, | |
| "loss": 0.5396, | |
| "num_tokens": 296035873.0, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.9395801331285202, | |
| "grad_norm": 0.8260330228111058, | |
| "learning_rate": 5.450162235043325e-06, | |
| "loss": 0.5954, | |
| "num_tokens": 296548316.0, | |
| "step": 5505 | |
| }, | |
| { | |
| "epoch": 0.9404335210786824, | |
| "grad_norm": 0.7709410941202116, | |
| "learning_rate": 5.43761188650749e-06, | |
| "loss": 0.5849, | |
| "num_tokens": 297048138.0, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 0.9412869090288445, | |
| "grad_norm": 0.7507662055546271, | |
| "learning_rate": 5.425237251654792e-06, | |
| "loss": 0.5376, | |
| "num_tokens": 297516443.0, | |
| "step": 5515 | |
| }, | |
| { | |
| "epoch": 0.9421402969790067, | |
| "grad_norm": 0.8591780584259089, | |
| "learning_rate": 5.41303842904177e-06, | |
| "loss": 0.6037, | |
| "num_tokens": 297956377.0, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 0.9429936849291688, | |
| "grad_norm": 0.7529771128056942, | |
| "learning_rate": 5.401015515824727e-06, | |
| "loss": 0.5635, | |
| "num_tokens": 298486321.0, | |
| "step": 5525 | |
| }, | |
| { | |
| "epoch": 0.9438470728793309, | |
| "grad_norm": 0.8051753782069618, | |
| "learning_rate": 5.389168607758956e-06, | |
| "loss": 0.586, | |
| "num_tokens": 299001151.0, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 0.9447004608294931, | |
| "grad_norm": 0.8410405788431461, | |
| "learning_rate": 5.377497799197965e-06, | |
| "loss": 0.5875, | |
| "num_tokens": 299468717.0, | |
| "step": 5535 | |
| }, | |
| { | |
| "epoch": 0.9455538487796552, | |
| "grad_norm": 0.8088038460686016, | |
| "learning_rate": 5.366003183092747e-06, | |
| "loss": 0.5779, | |
| "num_tokens": 299913523.0, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 0.9464072367298174, | |
| "grad_norm": 0.9065078781042983, | |
| "learning_rate": 5.354684850991019e-06, | |
| "loss": 0.5678, | |
| "num_tokens": 300388047.0, | |
| "step": 5545 | |
| }, | |
| { | |
| "epoch": 0.9472606246799795, | |
| "grad_norm": 0.915648202403995, | |
| "learning_rate": 5.343542893036508e-06, | |
| "loss": 0.5536, | |
| "num_tokens": 300834485.0, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.9481140126301416, | |
| "grad_norm": 0.8343832418261607, | |
| "learning_rate": 5.332577397968218e-06, | |
| "loss": 0.5899, | |
| "num_tokens": 301298744.0, | |
| "step": 5555 | |
| }, | |
| { | |
| "epoch": 0.9489674005803038, | |
| "grad_norm": 0.7289519169355831, | |
| "learning_rate": 5.321788453119741e-06, | |
| "loss": 0.5762, | |
| "num_tokens": 301755872.0, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 0.9498207885304659, | |
| "grad_norm": 0.8430411206851547, | |
| "learning_rate": 5.3111761444185486e-06, | |
| "loss": 0.5741, | |
| "num_tokens": 302223432.0, | |
| "step": 5565 | |
| }, | |
| { | |
| "epoch": 0.9506741764806281, | |
| "grad_norm": 0.7548208551690323, | |
| "learning_rate": 5.300740556385312e-06, | |
| "loss": 0.6051, | |
| "num_tokens": 302734843.0, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 0.9515275644307902, | |
| "grad_norm": 0.9660243301325951, | |
| "learning_rate": 5.29048177213323e-06, | |
| "loss": 0.6151, | |
| "num_tokens": 303156803.0, | |
| "step": 5575 | |
| }, | |
| { | |
| "epoch": 0.9523809523809523, | |
| "grad_norm": 0.7993982834474233, | |
| "learning_rate": 5.280399873367359e-06, | |
| "loss": 0.5612, | |
| "num_tokens": 303649911.0, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 0.9532343403311145, | |
| "grad_norm": 0.9320720781344985, | |
| "learning_rate": 5.270494940383981e-06, | |
| "loss": 0.5702, | |
| "num_tokens": 304114480.0, | |
| "step": 5585 | |
| }, | |
| { | |
| "epoch": 0.9540877282812766, | |
| "grad_norm": 0.8829913819415941, | |
| "learning_rate": 5.260767052069932e-06, | |
| "loss": 0.5835, | |
| "num_tokens": 304603280.0, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 0.9549411162314388, | |
| "grad_norm": 0.8333512888706996, | |
| "learning_rate": 5.251216285902014e-06, | |
| "loss": 0.5597, | |
| "num_tokens": 305082191.0, | |
| "step": 5595 | |
| }, | |
| { | |
| "epoch": 0.955794504181601, | |
| "grad_norm": 0.7482070040515338, | |
| "learning_rate": 5.241842717946349e-06, | |
| "loss": 0.5598, | |
| "num_tokens": 305542730.0, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.956647892131763, | |
| "grad_norm": 0.8316101679517978, | |
| "learning_rate": 5.232646422857779e-06, | |
| "loss": 0.5351, | |
| "num_tokens": 306042529.0, | |
| "step": 5605 | |
| }, | |
| { | |
| "epoch": 0.9575012800819253, | |
| "grad_norm": 0.8066685285673841, | |
| "learning_rate": 5.2236274738792755e-06, | |
| "loss": 0.56, | |
| "num_tokens": 306498168.0, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 0.9583546680320874, | |
| "grad_norm": 0.8308438231158456, | |
| "learning_rate": 5.214785942841354e-06, | |
| "loss": 0.5731, | |
| "num_tokens": 307010799.0, | |
| "step": 5615 | |
| }, | |
| { | |
| "epoch": 0.9592080559822496, | |
| "grad_norm": 0.7360643184841489, | |
| "learning_rate": 5.206121900161511e-06, | |
| "loss": 0.5608, | |
| "num_tokens": 307537300.0, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 0.9600614439324117, | |
| "grad_norm": 0.764899408471559, | |
| "learning_rate": 5.197635414843641e-06, | |
| "loss": 0.5587, | |
| "num_tokens": 308038692.0, | |
| "step": 5625 | |
| }, | |
| { | |
| "epoch": 0.9609148318825739, | |
| "grad_norm": 0.771106174868778, | |
| "learning_rate": 5.189326554477508e-06, | |
| "loss": 0.6011, | |
| "num_tokens": 308532583.0, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 0.961768219832736, | |
| "grad_norm": 0.8673796675256896, | |
| "learning_rate": 5.181195385238204e-06, | |
| "loss": 0.5583, | |
| "num_tokens": 308972109.0, | |
| "step": 5635 | |
| }, | |
| { | |
| "epoch": 0.9626216077828981, | |
| "grad_norm": 0.8023969340172891, | |
| "learning_rate": 5.173241971885606e-06, | |
| "loss": 0.5415, | |
| "num_tokens": 309426949.0, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 0.9634749957330603, | |
| "grad_norm": 0.7207659472227531, | |
| "learning_rate": 5.1654663777638825e-06, | |
| "loss": 0.5689, | |
| "num_tokens": 309918947.0, | |
| "step": 5645 | |
| }, | |
| { | |
| "epoch": 0.9643283836832224, | |
| "grad_norm": 0.9340552949332678, | |
| "learning_rate": 5.15786866480098e-06, | |
| "loss": 0.5728, | |
| "num_tokens": 310388816.0, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.9651817716333846, | |
| "grad_norm": 0.8775884081140152, | |
| "learning_rate": 5.150448893508114e-06, | |
| "loss": 0.5577, | |
| "num_tokens": 310788976.0, | |
| "step": 5655 | |
| }, | |
| { | |
| "epoch": 0.9660351595835467, | |
| "grad_norm": 0.862770324032615, | |
| "learning_rate": 5.143207122979319e-06, | |
| "loss": 0.5742, | |
| "num_tokens": 311244321.0, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 0.9668885475337088, | |
| "grad_norm": 0.8645037498593092, | |
| "learning_rate": 5.136143410890947e-06, | |
| "loss": 0.6224, | |
| "num_tokens": 311720200.0, | |
| "step": 5665 | |
| }, | |
| { | |
| "epoch": 0.967741935483871, | |
| "grad_norm": 0.7700789746051657, | |
| "learning_rate": 5.129257813501227e-06, | |
| "loss": 0.5511, | |
| "num_tokens": 312179788.0, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 0.9685953234340331, | |
| "grad_norm": 1.0031075094614286, | |
| "learning_rate": 5.122550385649811e-06, | |
| "loss": 0.5569, | |
| "num_tokens": 312622136.0, | |
| "step": 5675 | |
| }, | |
| { | |
| "epoch": 0.9694487113841953, | |
| "grad_norm": 0.8550909317145511, | |
| "learning_rate": 5.116021180757339e-06, | |
| "loss": 0.5757, | |
| "num_tokens": 313113611.0, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 0.9703020993343574, | |
| "grad_norm": 0.8269524914604716, | |
| "learning_rate": 5.1096702508250065e-06, | |
| "loss": 0.549, | |
| "num_tokens": 313614384.0, | |
| "step": 5685 | |
| }, | |
| { | |
| "epoch": 0.9711554872845195, | |
| "grad_norm": 0.8723888693586144, | |
| "learning_rate": 5.103497646434162e-06, | |
| "loss": 0.6138, | |
| "num_tokens": 314088892.0, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 0.9720088752346817, | |
| "grad_norm": 1.005485841547129, | |
| "learning_rate": 5.0975034167458985e-06, | |
| "loss": 0.6324, | |
| "num_tokens": 314588739.0, | |
| "step": 5695 | |
| }, | |
| { | |
| "epoch": 0.9728622631848438, | |
| "grad_norm": 0.9615462609522011, | |
| "learning_rate": 5.0916876095006525e-06, | |
| "loss": 0.5696, | |
| "num_tokens": 315051225.0, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.973715651135006, | |
| "grad_norm": 0.9044747223815061, | |
| "learning_rate": 5.086050271017843e-06, | |
| "loss": 0.5698, | |
| "num_tokens": 315515610.0, | |
| "step": 5705 | |
| }, | |
| { | |
| "epoch": 0.9745690390851681, | |
| "grad_norm": 0.9352208596040631, | |
| "learning_rate": 5.080591446195489e-06, | |
| "loss": 0.5828, | |
| "num_tokens": 315957594.0, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 0.9754224270353302, | |
| "grad_norm": 0.7315299597342139, | |
| "learning_rate": 5.075311178509852e-06, | |
| "loss": 0.5462, | |
| "num_tokens": 316377461.0, | |
| "step": 5715 | |
| }, | |
| { | |
| "epoch": 0.9762758149854924, | |
| "grad_norm": 0.9452274514037523, | |
| "learning_rate": 5.070209510015099e-06, | |
| "loss": 0.6164, | |
| "num_tokens": 316867560.0, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 0.9771292029356545, | |
| "grad_norm": 0.8371279249883385, | |
| "learning_rate": 5.065286481342953e-06, | |
| "loss": 0.5534, | |
| "num_tokens": 317298878.0, | |
| "step": 5725 | |
| }, | |
| { | |
| "epoch": 0.9779825908858167, | |
| "grad_norm": 0.8014870954247736, | |
| "learning_rate": 5.060542131702389e-06, | |
| "loss": 0.5693, | |
| "num_tokens": 317820296.0, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 0.9788359788359788, | |
| "grad_norm": 0.909292736397671, | |
| "learning_rate": 5.055976498879303e-06, | |
| "loss": 0.5763, | |
| "num_tokens": 318337653.0, | |
| "step": 5735 | |
| }, | |
| { | |
| "epoch": 0.9796893667861409, | |
| "grad_norm": 0.8132523377203885, | |
| "learning_rate": 5.05158961923622e-06, | |
| "loss": 0.5716, | |
| "num_tokens": 318780172.0, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 0.9805427547363031, | |
| "grad_norm": 0.8716679753297544, | |
| "learning_rate": 5.047381527712007e-06, | |
| "loss": 0.5827, | |
| "num_tokens": 319265582.0, | |
| "step": 5745 | |
| }, | |
| { | |
| "epoch": 0.9813961426864652, | |
| "grad_norm": 0.9185765003687043, | |
| "learning_rate": 5.0433522578215845e-06, | |
| "loss": 0.585, | |
| "num_tokens": 319710867.0, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.9822495306366275, | |
| "grad_norm": 0.682401230441357, | |
| "learning_rate": 5.039501841655672e-06, | |
| "loss": 0.5582, | |
| "num_tokens": 320197204.0, | |
| "step": 5755 | |
| }, | |
| { | |
| "epoch": 0.9831029185867896, | |
| "grad_norm": 0.9090161531763877, | |
| "learning_rate": 5.035830309880523e-06, | |
| "loss": 0.5464, | |
| "num_tokens": 320667324.0, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 0.9839563065369517, | |
| "grad_norm": 0.902665666969385, | |
| "learning_rate": 5.032337691737683e-06, | |
| "loss": 0.5731, | |
| "num_tokens": 321110869.0, | |
| "step": 5765 | |
| }, | |
| { | |
| "epoch": 0.9848096944871139, | |
| "grad_norm": 0.8605159320287843, | |
| "learning_rate": 5.0290240150437645e-06, | |
| "loss": 0.5466, | |
| "num_tokens": 321575498.0, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 0.985663082437276, | |
| "grad_norm": 0.7295626658824659, | |
| "learning_rate": 5.025889306190208e-06, | |
| "loss": 0.5792, | |
| "num_tokens": 322113584.0, | |
| "step": 5775 | |
| }, | |
| { | |
| "epoch": 0.9865164703874382, | |
| "grad_norm": 0.7249267684266418, | |
| "learning_rate": 5.0229335901430926e-06, | |
| "loss": 0.5929, | |
| "num_tokens": 322595635.0, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 0.9873698583376003, | |
| "grad_norm": 0.7282049411658805, | |
| "learning_rate": 5.020156890442924e-06, | |
| "loss": 0.5452, | |
| "num_tokens": 323066444.0, | |
| "step": 5785 | |
| }, | |
| { | |
| "epoch": 0.9882232462877624, | |
| "grad_norm": 0.7731368697074033, | |
| "learning_rate": 5.017559229204447e-06, | |
| "loss": 0.5625, | |
| "num_tokens": 323576730.0, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 0.9890766342379246, | |
| "grad_norm": 0.8143826121511745, | |
| "learning_rate": 5.015140627116475e-06, | |
| "loss": 0.5574, | |
| "num_tokens": 324026506.0, | |
| "step": 5795 | |
| }, | |
| { | |
| "epoch": 0.9899300221880867, | |
| "grad_norm": 0.8762523071927865, | |
| "learning_rate": 5.012901103441723e-06, | |
| "loss": 0.5586, | |
| "num_tokens": 324513537.0, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.9907834101382489, | |
| "grad_norm": 0.9521303946801305, | |
| "learning_rate": 5.010840676016651e-06, | |
| "loss": 0.5537, | |
| "num_tokens": 324953710.0, | |
| "step": 5805 | |
| }, | |
| { | |
| "epoch": 0.991636798088411, | |
| "grad_norm": 0.8448341812920022, | |
| "learning_rate": 5.008959361251331e-06, | |
| "loss": 0.5736, | |
| "num_tokens": 325415446.0, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 0.9924901860385731, | |
| "grad_norm": 0.8913613215917732, | |
| "learning_rate": 5.007257174129304e-06, | |
| "loss": 0.5843, | |
| "num_tokens": 325886164.0, | |
| "step": 5815 | |
| }, | |
| { | |
| "epoch": 0.9933435739887353, | |
| "grad_norm": 0.8125616463451861, | |
| "learning_rate": 5.00573412820747e-06, | |
| "loss": 0.6087, | |
| "num_tokens": 326345284.0, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 0.9941969619388974, | |
| "grad_norm": 0.8354108870278607, | |
| "learning_rate": 5.004390235615973e-06, | |
| "loss": 0.5695, | |
| "num_tokens": 326821080.0, | |
| "step": 5825 | |
| }, | |
| { | |
| "epoch": 0.9950503498890596, | |
| "grad_norm": 0.874952932031631, | |
| "learning_rate": 5.003225507058114e-06, | |
| "loss": 0.5765, | |
| "num_tokens": 327273974.0, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 0.9959037378392217, | |
| "grad_norm": 0.8005811663402586, | |
| "learning_rate": 5.002239951810257e-06, | |
| "loss": 0.5482, | |
| "num_tokens": 327795818.0, | |
| "step": 5835 | |
| }, | |
| { | |
| "epoch": 0.9967571257893838, | |
| "grad_norm": 0.733021531503028, | |
| "learning_rate": 5.001433577721758e-06, | |
| "loss": 0.5602, | |
| "num_tokens": 328290879.0, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 0.997610513739546, | |
| "grad_norm": 0.906113294495654, | |
| "learning_rate": 5.000806391214903e-06, | |
| "loss": 0.5923, | |
| "num_tokens": 328734615.0, | |
| "step": 5845 | |
| }, | |
| { | |
| "epoch": 0.9984639016897081, | |
| "grad_norm": 0.7589826908596056, | |
| "learning_rate": 5.000358397284853e-06, | |
| "loss": 0.5371, | |
| "num_tokens": 329195352.0, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.9993172896398703, | |
| "grad_norm": 0.9083053676708959, | |
| "learning_rate": 5.0000895994996155e-06, | |
| "loss": 0.5777, | |
| "num_tokens": 329702919.0, | |
| "step": 5855 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "num_tokens": 330043597.0, | |
| "step": 5859, | |
| "total_flos": 566103768268800.0, | |
| "train_loss": 0.349827840967499, | |
| "train_runtime": 19944.0659, | |
| "train_samples_per_second": 4.7, | |
| "train_steps_per_second": 0.294 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 5859, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 566103768268800.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |