{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 99.48249027237354, "eval_steps": 500, "global_step": 19200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.9961089494163424, "grad_norm": 0.9873842597007751, "learning_rate": 9.901562500000001e-06, "loss": 2.5953, "step": 192 }, { "epoch": 1.9909208819714657, "grad_norm": 1.439842700958252, "learning_rate": 9.8015625e-06, "loss": 2.462, "step": 384 }, { "epoch": 2.985732814526589, "grad_norm": 2.307729482650757, "learning_rate": 9.701562500000001e-06, "loss": 2.3659, "step": 576 }, { "epoch": 3.980544747081712, "grad_norm": 4.006823539733887, "learning_rate": 9.6015625e-06, "loss": 2.3475, "step": 768 }, { "epoch": 4.975356679636835, "grad_norm": 2.971815824508667, "learning_rate": 9.501562500000001e-06, "loss": 2.3162, "step": 960 }, { "epoch": 5.970168612191959, "grad_norm": 3.6012327671051025, "learning_rate": 9.4015625e-06, "loss": 2.3049, "step": 1152 }, { "epoch": 6.964980544747082, "grad_norm": 3.803527355194092, "learning_rate": 9.301562500000002e-06, "loss": 2.2733, "step": 1344 }, { "epoch": 7.959792477302205, "grad_norm": 3.0761234760284424, "learning_rate": 9.201562500000001e-06, "loss": 2.2576, "step": 1536 }, { "epoch": 8.954604409857328, "grad_norm": 2.6708500385284424, "learning_rate": 9.1015625e-06, "loss": 2.2367, "step": 1728 }, { "epoch": 9.949416342412452, "grad_norm": 3.482462167739868, "learning_rate": 9.001562500000001e-06, "loss": 2.176, "step": 1920 }, { "epoch": 10.944228274967575, "grad_norm": 2.826169967651367, "learning_rate": 8.902083333333334e-06, "loss": 2.2051, "step": 2112 }, { "epoch": 11.939040207522698, "grad_norm": 3.164243698120117, "learning_rate": 8.802083333333335e-06, "loss": 2.1874, "step": 2304 }, { "epoch": 12.933852140077821, "grad_norm": 4.725254058837891, "learning_rate": 8.702083333333334e-06, "loss": 2.1809, "step": 2496 }, { "epoch": 13.928664072632944, "grad_norm": 3.7806782722473145, "learning_rate": 8.602083333333333e-06, "loss": 2.16, "step": 2688 }, { "epoch": 14.923476005188068, "grad_norm": 3.735746145248413, "learning_rate": 8.502083333333334e-06, "loss": 2.1663, "step": 2880 }, { "epoch": 15.918287937743191, "grad_norm": 3.401533365249634, "learning_rate": 8.402083333333334e-06, "loss": 2.1353, "step": 3072 }, { "epoch": 16.913099870298314, "grad_norm": 3.1076550483703613, "learning_rate": 8.302083333333335e-06, "loss": 2.1277, "step": 3264 }, { "epoch": 17.907911802853437, "grad_norm": 3.2938179969787598, "learning_rate": 8.202083333333334e-06, "loss": 2.1165, "step": 3456 }, { "epoch": 18.90272373540856, "grad_norm": 3.4992620944976807, "learning_rate": 8.102083333333333e-06, "loss": 2.1407, "step": 3648 }, { "epoch": 19.897535667963684, "grad_norm": 5.144476413726807, "learning_rate": 8.002083333333334e-06, "loss": 2.1124, "step": 3840 }, { "epoch": 20.892347600518807, "grad_norm": 3.369130849838257, "learning_rate": 7.902083333333333e-06, "loss": 2.0994, "step": 4032 }, { "epoch": 21.88715953307393, "grad_norm": 2.5219101905822754, "learning_rate": 7.802604166666668e-06, "loss": 2.086, "step": 4224 }, { "epoch": 22.881971465629054, "grad_norm": 2.416583776473999, "learning_rate": 7.702604166666667e-06, "loss": 2.1011, "step": 4416 }, { "epoch": 23.876783398184177, "grad_norm": 2.880645513534546, "learning_rate": 7.602604166666667e-06, "loss": 2.069, "step": 4608 }, { "epoch": 24.8715953307393, "grad_norm": 3.6908700466156006, "learning_rate": 7.5026041666666675e-06, "loss": 2.0878, "step": 4800 }, { "epoch": 25.866407263294423, "grad_norm": 3.9834940433502197, "learning_rate": 7.402604166666667e-06, "loss": 2.0832, "step": 4992 }, { "epoch": 26.861219195849547, "grad_norm": 4.434403896331787, "learning_rate": 7.302604166666667e-06, "loss": 2.0618, "step": 5184 }, { "epoch": 27.85603112840467, "grad_norm": 2.6179683208465576, "learning_rate": 7.202604166666667e-06, "loss": 2.0698, "step": 5376 }, { "epoch": 28.850843060959793, "grad_norm": 3.7101306915283203, "learning_rate": 7.102604166666667e-06, "loss": 2.0561, "step": 5568 }, { "epoch": 29.845654993514916, "grad_norm": 2.940573215484619, "learning_rate": 7.002604166666668e-06, "loss": 2.0499, "step": 5760 }, { "epoch": 30.84046692607004, "grad_norm": 3.1798958778381348, "learning_rate": 6.902604166666667e-06, "loss": 2.049, "step": 5952 }, { "epoch": 31.835278858625163, "grad_norm": 3.288163423538208, "learning_rate": 6.802604166666667e-06, "loss": 2.0326, "step": 6144 }, { "epoch": 32.83009079118028, "grad_norm": 2.8764078617095947, "learning_rate": 6.703125e-06, "loss": 2.0371, "step": 6336 }, { "epoch": 33.824902723735406, "grad_norm": 2.9869441986083984, "learning_rate": 6.603125e-06, "loss": 2.0611, "step": 6528 }, { "epoch": 34.81971465629053, "grad_norm": 3.1720468997955322, "learning_rate": 6.5031250000000005e-06, "loss": 2.0541, "step": 6720 }, { "epoch": 35.81452658884565, "grad_norm": 3.7449281215667725, "learning_rate": 6.403125e-06, "loss": 2.0177, "step": 6912 }, { "epoch": 36.809338521400775, "grad_norm": 2.036007881164551, "learning_rate": 6.303125000000001e-06, "loss": 2.006, "step": 7104 }, { "epoch": 37.8041504539559, "grad_norm": 2.677340507507324, "learning_rate": 6.203125000000001e-06, "loss": 2.0329, "step": 7296 }, { "epoch": 38.79896238651102, "grad_norm": 2.8975298404693604, "learning_rate": 6.103125000000001e-06, "loss": 2.0347, "step": 7488 }, { "epoch": 39.793774319066145, "grad_norm": 1.8516769409179688, "learning_rate": 6.003125000000001e-06, "loss": 2.0179, "step": 7680 }, { "epoch": 40.78858625162127, "grad_norm": 2.984468698501587, "learning_rate": 5.903125e-06, "loss": 2.0002, "step": 7872 }, { "epoch": 41.78339818417639, "grad_norm": 4.378921031951904, "learning_rate": 5.8031250000000004e-06, "loss": 2.0032, "step": 8064 }, { "epoch": 42.778210116731515, "grad_norm": 3.600853204727173, "learning_rate": 5.7031250000000006e-06, "loss": 2.0131, "step": 8256 }, { "epoch": 43.77302204928664, "grad_norm": 3.475604295730591, "learning_rate": 5.603125000000001e-06, "loss": 2.0167, "step": 8448 }, { "epoch": 44.76783398184176, "grad_norm": 3.7388381958007812, "learning_rate": 5.503125e-06, "loss": 2.0286, "step": 8640 }, { "epoch": 45.762645914396884, "grad_norm": 3.0160775184631348, "learning_rate": 5.403125e-06, "loss": 2.0327, "step": 8832 }, { "epoch": 46.75745784695201, "grad_norm": 2.4288928508758545, "learning_rate": 5.303645833333334e-06, "loss": 1.9837, "step": 9024 }, { "epoch": 47.75226977950713, "grad_norm": 2.6072142124176025, "learning_rate": 5.203645833333334e-06, "loss": 2.022, "step": 9216 }, { "epoch": 48.747081712062254, "grad_norm": 3.0784308910369873, "learning_rate": 5.103645833333334e-06, "loss": 1.9946, "step": 9408 }, { "epoch": 49.74189364461738, "grad_norm": 2.552196979522705, "learning_rate": 5.003645833333334e-06, "loss": 1.9857, "step": 9600 }, { "epoch": 50.7367055771725, "grad_norm": 5.163851261138916, "learning_rate": 4.903645833333333e-06, "loss": 2.0038, "step": 9792 }, { "epoch": 51.731517509727624, "grad_norm": 3.9303905963897705, "learning_rate": 4.8036458333333335e-06, "loss": 2.0327, "step": 9984 }, { "epoch": 52.72632944228275, "grad_norm": 4.269293308258057, "learning_rate": 4.703645833333334e-06, "loss": 1.978, "step": 10176 }, { "epoch": 53.72114137483787, "grad_norm": 3.4755337238311768, "learning_rate": 4.603645833333334e-06, "loss": 2.0154, "step": 10368 }, { "epoch": 54.715953307392994, "grad_norm": 2.2201080322265625, "learning_rate": 4.503645833333334e-06, "loss": 1.9945, "step": 10560 }, { "epoch": 55.71076523994812, "grad_norm": 3.3768227100372314, "learning_rate": 4.403645833333334e-06, "loss": 1.9764, "step": 10752 }, { "epoch": 56.70557717250324, "grad_norm": 2.9863104820251465, "learning_rate": 4.303645833333334e-06, "loss": 2.0031, "step": 10944 }, { "epoch": 57.70038910505836, "grad_norm": 3.351330280303955, "learning_rate": 4.203645833333333e-06, "loss": 1.9915, "step": 11136 }, { "epoch": 58.69520103761349, "grad_norm": 2.745793104171753, "learning_rate": 4.103645833333333e-06, "loss": 2.0054, "step": 11328 }, { "epoch": 59.69001297016861, "grad_norm": 4.333703517913818, "learning_rate": 4.0036458333333335e-06, "loss": 1.9629, "step": 11520 }, { "epoch": 60.68482490272373, "grad_norm": 3.1686012744903564, "learning_rate": 3.903645833333334e-06, "loss": 2.0011, "step": 11712 }, { "epoch": 61.679636835278856, "grad_norm": 2.1179568767547607, "learning_rate": 3.804166666666667e-06, "loss": 1.9949, "step": 11904 }, { "epoch": 62.67444876783398, "grad_norm": 2.4998276233673096, "learning_rate": 3.704166666666667e-06, "loss": 1.9791, "step": 12096 }, { "epoch": 63.6692607003891, "grad_norm": 3.710357189178467, "learning_rate": 3.6041666666666667e-06, "loss": 1.9757, "step": 12288 }, { "epoch": 64.66407263294423, "grad_norm": 3.4601991176605225, "learning_rate": 3.504166666666667e-06, "loss": 2.025, "step": 12480 }, { "epoch": 65.65888456549935, "grad_norm": 2.626007318496704, "learning_rate": 3.4041666666666665e-06, "loss": 1.9755, "step": 12672 }, { "epoch": 66.65369649805447, "grad_norm": 3.1454813480377197, "learning_rate": 3.304166666666667e-06, "loss": 1.9984, "step": 12864 }, { "epoch": 67.6485084306096, "grad_norm": 7.283568859100342, "learning_rate": 3.204166666666667e-06, "loss": 1.9776, "step": 13056 }, { "epoch": 68.64332036316472, "grad_norm": 2.031538724899292, "learning_rate": 3.104166666666667e-06, "loss": 1.9939, "step": 13248 }, { "epoch": 69.63813229571984, "grad_norm": 2.370424270629883, "learning_rate": 3.004166666666667e-06, "loss": 1.9805, "step": 13440 }, { "epoch": 70.63294422827497, "grad_norm": 3.6357874870300293, "learning_rate": 2.9041666666666667e-06, "loss": 1.9998, "step": 13632 }, { "epoch": 71.62775616083009, "grad_norm": 3.61807918548584, "learning_rate": 2.8041666666666668e-06, "loss": 1.9615, "step": 13824 }, { "epoch": 72.62256809338521, "grad_norm": 2.440490484237671, "learning_rate": 2.7041666666666673e-06, "loss": 1.9715, "step": 14016 }, { "epoch": 73.61738002594034, "grad_norm": 2.490007162094116, "learning_rate": 2.6046875000000006e-06, "loss": 1.9765, "step": 14208 }, { "epoch": 74.61219195849546, "grad_norm": 2.368825674057007, "learning_rate": 2.5046875000000003e-06, "loss": 1.9819, "step": 14400 }, { "epoch": 75.60700389105058, "grad_norm": 2.9740967750549316, "learning_rate": 2.4046875000000004e-06, "loss": 1.9702, "step": 14592 }, { "epoch": 76.6018158236057, "grad_norm": 3.215397834777832, "learning_rate": 2.3046875e-06, "loss": 1.9913, "step": 14784 }, { "epoch": 77.59662775616083, "grad_norm": 2.465824604034424, "learning_rate": 2.2046875000000002e-06, "loss": 1.9751, "step": 14976 }, { "epoch": 78.59143968871595, "grad_norm": 2.2337405681610107, "learning_rate": 2.1046875000000003e-06, "loss": 1.9496, "step": 15168 }, { "epoch": 79.58625162127107, "grad_norm": 2.875598907470703, "learning_rate": 2.0046875e-06, "loss": 2.0066, "step": 15360 }, { "epoch": 80.5810635538262, "grad_norm": 2.8990750312805176, "learning_rate": 1.9046875000000001e-06, "loss": 1.9466, "step": 15552 }, { "epoch": 81.57587548638132, "grad_norm": 4.734499454498291, "learning_rate": 1.8046875000000002e-06, "loss": 1.9636, "step": 15744 }, { "epoch": 82.57068741893644, "grad_norm": 2.9932515621185303, "learning_rate": 1.7046875000000001e-06, "loss": 1.973, "step": 15936 }, { "epoch": 83.56549935149157, "grad_norm": 4.040909767150879, "learning_rate": 1.6046875e-06, "loss": 1.9678, "step": 16128 }, { "epoch": 84.56031128404669, "grad_norm": 3.0035159587860107, "learning_rate": 1.5046875000000002e-06, "loss": 1.9457, "step": 16320 }, { "epoch": 85.55512321660181, "grad_norm": 2.5469093322753906, "learning_rate": 1.4046875e-06, "loss": 1.9642, "step": 16512 }, { "epoch": 86.54993514915694, "grad_norm": 2.850858688354492, "learning_rate": 1.3046875e-06, "loss": 1.9638, "step": 16704 }, { "epoch": 87.54474708171206, "grad_norm": 2.5728836059570312, "learning_rate": 1.2046875e-06, "loss": 1.9681, "step": 16896 }, { "epoch": 88.53955901426718, "grad_norm": 1.4205690622329712, "learning_rate": 1.1046875000000002e-06, "loss": 1.9585, "step": 17088 }, { "epoch": 89.5343709468223, "grad_norm": 1.9591172933578491, "learning_rate": 1.0046875e-06, "loss": 1.9894, "step": 17280 }, { "epoch": 90.52918287937743, "grad_norm": 3.981717348098755, "learning_rate": 9.046875000000001e-07, "loss": 1.9388, "step": 17472 }, { "epoch": 91.52399481193255, "grad_norm": 2.348127841949463, "learning_rate": 8.046875000000001e-07, "loss": 1.9677, "step": 17664 }, { "epoch": 92.51880674448768, "grad_norm": 3.0600669384002686, "learning_rate": 7.046875e-07, "loss": 1.9747, "step": 17856 }, { "epoch": 93.5136186770428, "grad_norm": 3.865741491317749, "learning_rate": 6.046875000000001e-07, "loss": 1.958, "step": 18048 }, { "epoch": 94.50843060959792, "grad_norm": 2.755720853805542, "learning_rate": 5.052083333333334e-07, "loss": 1.975, "step": 18240 }, { "epoch": 95.50324254215305, "grad_norm": 3.4946985244750977, "learning_rate": 4.0520833333333335e-07, "loss": 1.9426, "step": 18432 }, { "epoch": 96.49805447470817, "grad_norm": 2.993678331375122, "learning_rate": 3.0520833333333336e-07, "loss": 1.9497, "step": 18624 }, { "epoch": 97.49286640726329, "grad_norm": 3.959327459335327, "learning_rate": 2.0520833333333334e-07, "loss": 1.9409, "step": 18816 }, { "epoch": 98.48767833981842, "grad_norm": 2.147462844848633, "learning_rate": 1.0520833333333334e-07, "loss": 1.9485, "step": 19008 }, { "epoch": 99.48249027237354, "grad_norm": 2.728996992111206, "learning_rate": 5.208333333333334e-09, "loss": 1.981, "step": 19200 }, { "epoch": 99.48249027237354, "step": 19200, "total_flos": 1.0064407234989773e+18, "train_loss": 2.0492228651046753, "train_runtime": 20851.1432, "train_samples_per_second": 184.834, "train_steps_per_second": 0.921 } ], "logging_steps": 192, "max_steps": 19200, "num_input_tokens_seen": 0, "num_train_epochs": 100, "save_steps": 6800, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.0064407234989773e+18, "train_batch_size": 50, "trial_name": null, "trial_params": null }