| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.0, | |
| "eval_steps": 500, | |
| "global_step": 225, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.044444444444444446, | |
| "grad_norm": 0.17873810231685638, | |
| "learning_rate": 8.695652173913044e-06, | |
| "loss": 0.9284, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.08888888888888889, | |
| "grad_norm": 0.1875726729631424, | |
| "learning_rate": 1.739130434782609e-05, | |
| "loss": 0.9365, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.13333333333333333, | |
| "grad_norm": 0.2115127593278885, | |
| "learning_rate": 2.608695652173913e-05, | |
| "loss": 0.9367, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.17777777777777778, | |
| "grad_norm": 0.2282930463552475, | |
| "learning_rate": 3.478260869565218e-05, | |
| "loss": 0.8894, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.2222222222222222, | |
| "grad_norm": 0.2383526861667633, | |
| "learning_rate": 4.347826086956522e-05, | |
| "loss": 0.9455, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.26666666666666666, | |
| "grad_norm": 0.23653796315193176, | |
| "learning_rate": 5.217391304347826e-05, | |
| "loss": 1.0121, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.3111111111111111, | |
| "grad_norm": 0.253218412399292, | |
| "learning_rate": 6.086956521739131e-05, | |
| "loss": 0.8756, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.35555555555555557, | |
| "grad_norm": 0.24895112216472626, | |
| "learning_rate": 6.956521739130436e-05, | |
| "loss": 0.9091, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.2350614368915558, | |
| "learning_rate": 7.82608695652174e-05, | |
| "loss": 0.9091, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.4444444444444444, | |
| "grad_norm": 0.27106258273124695, | |
| "learning_rate": 8.695652173913044e-05, | |
| "loss": 0.8018, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.4888888888888889, | |
| "grad_norm": 0.3147624731063843, | |
| "learning_rate": 9.565217391304348e-05, | |
| "loss": 0.796, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.5333333333333333, | |
| "grad_norm": 0.3592516779899597, | |
| "learning_rate": 9.999395316300748e-05, | |
| "loss": 0.8837, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.5777777777777777, | |
| "grad_norm": 0.27557510137557983, | |
| "learning_rate": 9.994558724213054e-05, | |
| "loss": 0.8367, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.6222222222222222, | |
| "grad_norm": 0.2757929861545563, | |
| "learning_rate": 9.984890219128146e-05, | |
| "loss": 0.8764, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.6666666666666666, | |
| "grad_norm": 0.30082619190216064, | |
| "learning_rate": 9.970399154700263e-05, | |
| "loss": 0.8687, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.7111111111111111, | |
| "grad_norm": 0.26034796237945557, | |
| "learning_rate": 9.951099550098349e-05, | |
| "loss": 0.8371, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.7555555555555555, | |
| "grad_norm": 0.22807522118091583, | |
| "learning_rate": 9.927010076443407e-05, | |
| "loss": 0.8085, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.22343170642852783, | |
| "learning_rate": 9.898154038745408e-05, | |
| "loss": 0.8812, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.8444444444444444, | |
| "grad_norm": 0.18960116803646088, | |
| "learning_rate": 9.864559353357187e-05, | |
| "loss": 0.7953, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.8888888888888888, | |
| "grad_norm": 0.21037450432777405, | |
| "learning_rate": 9.826258520967178e-05, | |
| "loss": 0.7484, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.9333333333333333, | |
| "grad_norm": 0.2212020456790924, | |
| "learning_rate": 9.783288595157098e-05, | |
| "loss": 0.7931, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.9777777777777777, | |
| "grad_norm": 0.23480449616909027, | |
| "learning_rate": 9.735691146555002e-05, | |
| "loss": 0.8413, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.0222222222222221, | |
| "grad_norm": 0.22327324748039246, | |
| "learning_rate": 9.683512222618377e-05, | |
| "loss": 0.8391, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.0666666666666667, | |
| "grad_norm": 0.24856683611869812, | |
| "learning_rate": 9.626802303086208e-05, | |
| "loss": 0.7997, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.1111111111111112, | |
| "grad_norm": 0.2144002616405487, | |
| "learning_rate": 9.565616251143094e-05, | |
| "loss": 0.9146, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.1555555555555554, | |
| "grad_norm": 0.24486525356769562, | |
| "learning_rate": 9.500013260342651e-05, | |
| "loss": 0.7813, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 0.23040799796581268, | |
| "learning_rate": 9.430056797341574e-05, | |
| "loss": 0.7282, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.2444444444444445, | |
| "grad_norm": 0.24283307790756226, | |
| "learning_rate": 9.355814540499752e-05, | |
| "loss": 0.7598, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.2888888888888888, | |
| "grad_norm": 0.25237399339675903, | |
| "learning_rate": 9.27735831440582e-05, | |
| "loss": 0.7959, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.3333333333333333, | |
| "grad_norm": 0.2991330325603485, | |
| "learning_rate": 9.194764020391506e-05, | |
| "loss": 0.7853, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.3777777777777778, | |
| "grad_norm": 0.27033334970474243, | |
| "learning_rate": 9.108111563102004e-05, | |
| "loss": 0.7553, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.4222222222222223, | |
| "grad_norm": 0.32717347145080566, | |
| "learning_rate": 9.017484773193378e-05, | |
| "loss": 0.7809, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.4666666666666668, | |
| "grad_norm": 0.29897943139076233, | |
| "learning_rate": 8.92297132623183e-05, | |
| "loss": 0.7628, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.511111111111111, | |
| "grad_norm": 0.30663371086120605, | |
| "learning_rate": 8.824662657873239e-05, | |
| "loss": 0.7763, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.5555555555555556, | |
| "grad_norm": 0.28747016191482544, | |
| "learning_rate": 8.722653875405075e-05, | |
| "loss": 0.737, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 0.30916616320610046, | |
| "learning_rate": 8.617043665736249e-05, | |
| "loss": 0.7879, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.6444444444444444, | |
| "grad_norm": 0.35312747955322266, | |
| "learning_rate": 8.507934199923884e-05, | |
| "loss": 0.7707, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.6888888888888889, | |
| "grad_norm": 0.36939722299575806, | |
| "learning_rate": 8.39543103432943e-05, | |
| "loss": 0.727, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.7333333333333334, | |
| "grad_norm": 0.3665863871574402, | |
| "learning_rate": 8.2796430084997e-05, | |
| "loss": 0.7193, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.7777777777777777, | |
| "grad_norm": 0.34428849816322327, | |
| "learning_rate": 8.160682139871633e-05, | |
| "loss": 0.7573, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.8222222222222222, | |
| "grad_norm": 0.3433472216129303, | |
| "learning_rate": 8.03866351540266e-05, | |
| "loss": 0.7711, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.8666666666666667, | |
| "grad_norm": 0.3693158030509949, | |
| "learning_rate": 7.913705180231505e-05, | |
| "loss": 0.7741, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.911111111111111, | |
| "grad_norm": 0.4667452573776245, | |
| "learning_rate": 7.785928023477142e-05, | |
| "loss": 0.812, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.9555555555555557, | |
| "grad_norm": 0.36592212319374084, | |
| "learning_rate": 7.655455661286376e-05, | |
| "loss": 0.7561, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.4149788022041321, | |
| "learning_rate": 7.5224143172432e-05, | |
| "loss": 0.7797, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.0444444444444443, | |
| "grad_norm": 0.39174893498420715, | |
| "learning_rate": 7.386932700255636e-05, | |
| "loss": 0.7651, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 2.088888888888889, | |
| "grad_norm": 0.4016912877559662, | |
| "learning_rate": 7.24914188003818e-05, | |
| "loss": 0.741, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.1333333333333333, | |
| "grad_norm": 0.46315836906433105, | |
| "learning_rate": 7.109175160310312e-05, | |
| "loss": 0.695, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.1777777777777776, | |
| "grad_norm": 0.48870378732681274, | |
| "learning_rate": 6.967167949833763e-05, | |
| "loss": 0.7259, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.2222222222222223, | |
| "grad_norm": 0.43603938817977905, | |
| "learning_rate": 6.823257631413276e-05, | |
| "loss": 0.7574, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.2666666666666666, | |
| "grad_norm": 0.46362167596817017, | |
| "learning_rate": 6.677583428987625e-05, | |
| "loss": 0.6457, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.311111111111111, | |
| "grad_norm": 0.5170242786407471, | |
| "learning_rate": 6.530286272939437e-05, | |
| "loss": 0.6498, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 2.3555555555555556, | |
| "grad_norm": 0.4572562575340271, | |
| "learning_rate": 6.381508663754153e-05, | |
| "loss": 0.6389, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 0.5652968287467957, | |
| "learning_rate": 6.231394534160008e-05, | |
| "loss": 0.7215, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.4444444444444446, | |
| "grad_norm": 0.5415116548538208, | |
| "learning_rate": 6.0800891098824186e-05, | |
| "loss": 0.7096, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.488888888888889, | |
| "grad_norm": 0.493282675743103, | |
| "learning_rate": 5.9277387691474676e-05, | |
| "loss": 0.6736, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 2.533333333333333, | |
| "grad_norm": 0.6388583183288574, | |
| "learning_rate": 5.774490901070424e-05, | |
| "loss": 0.6875, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 2.5777777777777775, | |
| "grad_norm": 0.4991196393966675, | |
| "learning_rate": 5.620493763066297e-05, | |
| "loss": 0.6023, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 2.6222222222222222, | |
| "grad_norm": 0.5749160051345825, | |
| "learning_rate": 5.465896337420359e-05, | |
| "loss": 0.6356, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 2.6666666666666665, | |
| "grad_norm": 0.6559845209121704, | |
| "learning_rate": 5.3108481871574036e-05, | |
| "loss": 0.6939, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.7111111111111112, | |
| "grad_norm": 0.6793063879013062, | |
| "learning_rate": 5.155499311349185e-05, | |
| "loss": 0.6981, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 2.7555555555555555, | |
| "grad_norm": 0.6073561310768127, | |
| "learning_rate": 5e-05, | |
| "loss": 0.6733, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "grad_norm": 0.6620696187019348, | |
| "learning_rate": 4.844500688650816e-05, | |
| "loss": 0.6851, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 2.8444444444444446, | |
| "grad_norm": 0.626446545124054, | |
| "learning_rate": 4.6891518128425976e-05, | |
| "loss": 0.654, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 2.888888888888889, | |
| "grad_norm": 0.6012407541275024, | |
| "learning_rate": 4.534103662579642e-05, | |
| "loss": 0.6303, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.9333333333333336, | |
| "grad_norm": 0.6335547566413879, | |
| "learning_rate": 4.3795062369337034e-05, | |
| "loss": 0.5819, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 2.977777777777778, | |
| "grad_norm": 0.6275166273117065, | |
| "learning_rate": 4.2255090989295764e-05, | |
| "loss": 0.6726, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 3.022222222222222, | |
| "grad_norm": 0.545259416103363, | |
| "learning_rate": 4.0722612308525335e-05, | |
| "loss": 0.6621, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 3.066666666666667, | |
| "grad_norm": 0.6598561406135559, | |
| "learning_rate": 3.919910890117584e-05, | |
| "loss": 0.5295, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 3.111111111111111, | |
| "grad_norm": 0.6240947842597961, | |
| "learning_rate": 3.7686054658399935e-05, | |
| "loss": 0.5466, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 3.1555555555555554, | |
| "grad_norm": 0.682789146900177, | |
| "learning_rate": 3.618491336245849e-05, | |
| "loss": 0.5419, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "grad_norm": 0.7448738217353821, | |
| "learning_rate": 3.469713727060564e-05, | |
| "loss": 0.6171, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 3.2444444444444445, | |
| "grad_norm": 0.7555476427078247, | |
| "learning_rate": 3.3224165710123756e-05, | |
| "loss": 0.5628, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 3.2888888888888888, | |
| "grad_norm": 0.7862750887870789, | |
| "learning_rate": 3.176742368586725e-05, | |
| "loss": 0.5923, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 3.3333333333333335, | |
| "grad_norm": 0.7729607224464417, | |
| "learning_rate": 3.032832050166239e-05, | |
| "loss": 0.5308, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 3.3777777777777778, | |
| "grad_norm": 0.7676399350166321, | |
| "learning_rate": 2.890824839689689e-05, | |
| "loss": 0.6444, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 3.422222222222222, | |
| "grad_norm": 0.8251731991767883, | |
| "learning_rate": 2.750858119961821e-05, | |
| "loss": 0.6573, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 3.466666666666667, | |
| "grad_norm": 0.8169701099395752, | |
| "learning_rate": 2.613067299744364e-05, | |
| "loss": 0.582, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 3.511111111111111, | |
| "grad_norm": 0.7788083553314209, | |
| "learning_rate": 2.4775856827568016e-05, | |
| "loss": 0.6335, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 3.5555555555555554, | |
| "grad_norm": 0.8565711379051208, | |
| "learning_rate": 2.3445443387136244e-05, | |
| "loss": 0.508, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "grad_norm": 0.8271569609642029, | |
| "learning_rate": 2.2140719765228584e-05, | |
| "loss": 0.5991, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 3.6444444444444444, | |
| "grad_norm": 0.812567412853241, | |
| "learning_rate": 2.0862948197684955e-05, | |
| "loss": 0.5808, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 3.688888888888889, | |
| "grad_norm": 0.7940819263458252, | |
| "learning_rate": 1.961336484597343e-05, | |
| "loss": 0.5703, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 3.7333333333333334, | |
| "grad_norm": 0.9031268358230591, | |
| "learning_rate": 1.8393178601283683e-05, | |
| "loss": 0.5002, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 3.7777777777777777, | |
| "grad_norm": 0.8260697722434998, | |
| "learning_rate": 1.7203569915003005e-05, | |
| "loss": 0.5269, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 3.822222222222222, | |
| "grad_norm": 0.9412121772766113, | |
| "learning_rate": 1.6045689656705716e-05, | |
| "loss": 0.4902, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 3.8666666666666667, | |
| "grad_norm": 0.9131683707237244, | |
| "learning_rate": 1.4920658000761174e-05, | |
| "loss": 0.6185, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 3.911111111111111, | |
| "grad_norm": 0.7923269271850586, | |
| "learning_rate": 1.3829563342637513e-05, | |
| "loss": 0.537, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 3.9555555555555557, | |
| "grad_norm": 0.8013989329338074, | |
| "learning_rate": 1.2773461245949247e-05, | |
| "loss": 0.5637, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 0.7401474118232727, | |
| "learning_rate": 1.1753373421267621e-05, | |
| "loss": 0.5124, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 4.044444444444444, | |
| "grad_norm": 0.8890312910079956, | |
| "learning_rate": 1.0770286737681701e-05, | |
| "loss": 0.5488, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 4.088888888888889, | |
| "grad_norm": 0.858996570110321, | |
| "learning_rate": 9.825152268066213e-06, | |
| "loss": 0.4998, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 4.133333333333334, | |
| "grad_norm": 0.755382239818573, | |
| "learning_rate": 8.91888436897997e-06, | |
| "loss": 0.4892, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 4.177777777777778, | |
| "grad_norm": 0.8359836935997009, | |
| "learning_rate": 8.052359796084951e-06, | |
| "loss": 0.5365, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 4.222222222222222, | |
| "grad_norm": 0.8235160112380981, | |
| "learning_rate": 7.226416855941814e-06, | |
| "loss": 0.4963, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 4.266666666666667, | |
| "grad_norm": 0.9951562881469727, | |
| "learning_rate": 6.441854595002477e-06, | |
| "loss": 0.5367, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 4.311111111111111, | |
| "grad_norm": 0.907408595085144, | |
| "learning_rate": 5.699432026584267e-06, | |
| "loss": 0.5051, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 4.355555555555555, | |
| "grad_norm": 0.9453915357589722, | |
| "learning_rate": 4.999867396573499e-06, | |
| "loss": 0.5326, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "grad_norm": 0.8306211829185486, | |
| "learning_rate": 4.343837488569058e-06, | |
| "loss": 0.4549, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 4.444444444444445, | |
| "grad_norm": 0.9665538668632507, | |
| "learning_rate": 3.731976969137929e-06, | |
| "loss": 0.4968, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 4.488888888888889, | |
| "grad_norm": 0.9233301281929016, | |
| "learning_rate": 3.1648777738162494e-06, | |
| "loss": 0.4623, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 4.533333333333333, | |
| "grad_norm": 0.8156710267066956, | |
| "learning_rate": 2.6430885344499946e-06, | |
| "loss": 0.4445, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 4.5777777777777775, | |
| "grad_norm": 0.8379063010215759, | |
| "learning_rate": 2.1671140484290142e-06, | |
| "loss": 0.4678, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 4.622222222222222, | |
| "grad_norm": 0.8422232866287231, | |
| "learning_rate": 1.7374147903282178e-06, | |
| "loss": 0.5377, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 4.666666666666667, | |
| "grad_norm": 0.9653432965278625, | |
| "learning_rate": 1.3544064664281265e-06, | |
| "loss": 0.5819, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 4.711111111111111, | |
| "grad_norm": 0.8152110576629639, | |
| "learning_rate": 1.0184596125459135e-06, | |
| "loss": 0.494, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 4.7555555555555555, | |
| "grad_norm": 0.9850694537162781, | |
| "learning_rate": 7.29899235565934e-07, | |
| "loss": 0.4983, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "grad_norm": 0.9584403038024902, | |
| "learning_rate": 4.890044990165321e-07, | |
| "loss": 0.5186, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 4.844444444444444, | |
| "grad_norm": 0.948579728603363, | |
| "learning_rate": 2.9600845299737056e-07, | |
| "loss": 0.5044, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 4.888888888888889, | |
| "grad_norm": 0.8661581873893738, | |
| "learning_rate": 1.5109780871853663e-07, | |
| "loss": 0.5062, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 4.933333333333334, | |
| "grad_norm": 0.9714407324790955, | |
| "learning_rate": 5.4412757869459763e-08, | |
| "loss": 0.5391, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 4.977777777777778, | |
| "grad_norm": 0.7409553527832031, | |
| "learning_rate": 6.04683699252373e-09, | |
| "loss": 0.5243, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "step": 225, | |
| "total_flos": 8.831352737326694e+16, | |
| "train_loss": 0.6794352424144745, | |
| "train_runtime": 806.654, | |
| "train_samples_per_second": 8.926, | |
| "train_steps_per_second": 0.279 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 225, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": false, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.831352737326694e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |