| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.8994865546949784, | |
| "eval_steps": 100, | |
| "global_step": 11200, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.017404925593943087, | |
| "grad_norm": 6.07357337417729, | |
| "learning_rate": 4.9e-05, | |
| "loss": 3.6087, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.034809851187886175, | |
| "grad_norm": 3.5663292481821296, | |
| "learning_rate": 4.999971170608479e-05, | |
| "loss": 2.545, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.034809851187886175, | |
| "eval_loss": 2.3234446048736572, | |
| "eval_runtime": 13.9956, | |
| "eval_samples_per_second": 71.451, | |
| "eval_steps_per_second": 2.286, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.05221477678182926, | |
| "grad_norm": 3.76347751871297, | |
| "learning_rate": 4.999882317878336e-05, | |
| "loss": 2.2721, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.06961970237577235, | |
| "grad_norm": 6.121069966490686, | |
| "learning_rate": 4.99973343246949e-05, | |
| "loss": 2.0721, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.06961970237577235, | |
| "eval_loss": 1.9963123798370361, | |
| "eval_runtime": 14.05, | |
| "eval_samples_per_second": 71.174, | |
| "eval_steps_per_second": 2.278, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.08702462796971543, | |
| "grad_norm": 2.0071155926776294, | |
| "learning_rate": 4.9995245188511975e-05, | |
| "loss": 1.9929, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.10442955356365852, | |
| "grad_norm": 1.7760725594180993, | |
| "learning_rate": 4.9992555832946495e-05, | |
| "loss": 1.9184, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.10442955356365852, | |
| "eval_loss": 1.8656612634658813, | |
| "eval_runtime": 13.9964, | |
| "eval_samples_per_second": 71.447, | |
| "eval_steps_per_second": 2.286, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.12183447915760161, | |
| "grad_norm": 1.8447005346692253, | |
| "learning_rate": 4.9989266338727845e-05, | |
| "loss": 1.8957, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.1392394047515447, | |
| "grad_norm": 2.304515920283501, | |
| "learning_rate": 4.998537680460039e-05, | |
| "loss": 1.8629, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.1392394047515447, | |
| "eval_loss": 1.8171635866165161, | |
| "eval_runtime": 14.0311, | |
| "eval_samples_per_second": 71.27, | |
| "eval_steps_per_second": 2.281, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.15664433034548778, | |
| "grad_norm": 2.028746062309368, | |
| "learning_rate": 4.998088734732058e-05, | |
| "loss": 1.8457, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.17404925593943085, | |
| "grad_norm": 1.5394435777337994, | |
| "learning_rate": 4.997579810165342e-05, | |
| "loss": 1.8232, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.17404925593943085, | |
| "eval_loss": 1.7938146591186523, | |
| "eval_runtime": 13.9792, | |
| "eval_samples_per_second": 71.535, | |
| "eval_steps_per_second": 2.289, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.19145418153337396, | |
| "grad_norm": 1.890814268154515, | |
| "learning_rate": 4.997010922036841e-05, | |
| "loss": 1.8101, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.20885910712731703, | |
| "grad_norm": 2.872076377134816, | |
| "learning_rate": 4.9963820874235e-05, | |
| "loss": 1.7842, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.20885910712731703, | |
| "eval_loss": 1.779714584350586, | |
| "eval_runtime": 13.9665, | |
| "eval_samples_per_second": 71.6, | |
| "eval_steps_per_second": 2.291, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.2262640327212601, | |
| "grad_norm": 2.481027223933055, | |
| "learning_rate": 4.995693325201745e-05, | |
| "loss": 1.794, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.24366895831520322, | |
| "grad_norm": 0.8094796094808727, | |
| "learning_rate": 4.99494465604691e-05, | |
| "loss": 1.7793, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.24366895831520322, | |
| "eval_loss": 1.7532519102096558, | |
| "eval_runtime": 14.0406, | |
| "eval_samples_per_second": 71.222, | |
| "eval_steps_per_second": 2.279, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.26107388390914626, | |
| "grad_norm": 1.2444472352367841, | |
| "learning_rate": 4.994136102432626e-05, | |
| "loss": 1.7694, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.2784788095030894, | |
| "grad_norm": 3.529061157225736, | |
| "learning_rate": 4.99326768863014e-05, | |
| "loss": 1.7661, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.2784788095030894, | |
| "eval_loss": 1.75984525680542, | |
| "eval_runtime": 13.9384, | |
| "eval_samples_per_second": 71.744, | |
| "eval_steps_per_second": 2.296, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.2958837350970325, | |
| "grad_norm": 1.377228521670443, | |
| "learning_rate": 4.992339440707591e-05, | |
| "loss": 1.7622, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.31328866069097555, | |
| "grad_norm": 1.6693867914447456, | |
| "learning_rate": 4.991351386529222e-05, | |
| "loss": 1.753, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.31328866069097555, | |
| "eval_loss": 1.72667396068573, | |
| "eval_runtime": 14.0823, | |
| "eval_samples_per_second": 71.011, | |
| "eval_steps_per_second": 2.272, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.33069358628491863, | |
| "grad_norm": 1.7888888121121684, | |
| "learning_rate": 4.9903035557545474e-05, | |
| "loss": 1.7394, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.3480985118788617, | |
| "grad_norm": 1.1475145633011983, | |
| "learning_rate": 4.9891959798374635e-05, | |
| "loss": 1.7418, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.3480985118788617, | |
| "eval_loss": 1.7262976169586182, | |
| "eval_runtime": 13.9499, | |
| "eval_samples_per_second": 71.685, | |
| "eval_steps_per_second": 2.294, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.3655034374728048, | |
| "grad_norm": 1.2690065543406053, | |
| "learning_rate": 4.9880286920252995e-05, | |
| "loss": 1.7323, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.3829083630667479, | |
| "grad_norm": 1.3554029672368741, | |
| "learning_rate": 4.986801727357825e-05, | |
| "loss": 1.7214, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.3829083630667479, | |
| "eval_loss": 1.7105458974838257, | |
| "eval_runtime": 14.0176, | |
| "eval_samples_per_second": 71.339, | |
| "eval_steps_per_second": 2.283, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.400313288660691, | |
| "grad_norm": 1.1363622763657797, | |
| "learning_rate": 4.9855151226661954e-05, | |
| "loss": 1.7249, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.41771821425463407, | |
| "grad_norm": 0.821944816072442, | |
| "learning_rate": 4.984168916571846e-05, | |
| "loss": 1.7315, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.41771821425463407, | |
| "eval_loss": 1.7037900686264038, | |
| "eval_runtime": 13.9751, | |
| "eval_samples_per_second": 71.556, | |
| "eval_steps_per_second": 2.29, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.43512313984857715, | |
| "grad_norm": 1.2782341420995915, | |
| "learning_rate": 4.9827631494853346e-05, | |
| "loss": 1.734, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.4525280654425202, | |
| "grad_norm": 0.9283632211124734, | |
| "learning_rate": 4.9812978636051235e-05, | |
| "loss": 1.7198, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.4525280654425202, | |
| "eval_loss": 1.6981171369552612, | |
| "eval_runtime": 14.0083, | |
| "eval_samples_per_second": 71.386, | |
| "eval_steps_per_second": 2.284, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.4699329910364633, | |
| "grad_norm": 1.0677942729129404, | |
| "learning_rate": 4.9797731029163206e-05, | |
| "loss": 1.7041, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.48733791663040643, | |
| "grad_norm": 1.1328848186202982, | |
| "learning_rate": 4.978188913189354e-05, | |
| "loss": 1.7106, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.48733791663040643, | |
| "eval_loss": 1.6910983324050903, | |
| "eval_runtime": 13.9325, | |
| "eval_samples_per_second": 71.775, | |
| "eval_steps_per_second": 2.297, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.5047428422243495, | |
| "grad_norm": 0.8854373258207832, | |
| "learning_rate": 4.976545341978598e-05, | |
| "loss": 1.7079, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.5221477678182925, | |
| "grad_norm": 1.5301565843095193, | |
| "learning_rate": 4.9748424386209455e-05, | |
| "loss": 1.7065, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.5221477678182925, | |
| "eval_loss": 1.6862553358078003, | |
| "eval_runtime": 14.0162, | |
| "eval_samples_per_second": 71.346, | |
| "eval_steps_per_second": 2.283, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.5395526934122357, | |
| "grad_norm": 1.4228459883867106, | |
| "learning_rate": 4.973080254234331e-05, | |
| "loss": 1.7037, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.5569576190061788, | |
| "grad_norm": 0.8012113010542086, | |
| "learning_rate": 4.971258841716191e-05, | |
| "loss": 1.695, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.5569576190061788, | |
| "eval_loss": 1.6799589395523071, | |
| "eval_runtime": 13.9995, | |
| "eval_samples_per_second": 71.431, | |
| "eval_steps_per_second": 2.286, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.5743625446001218, | |
| "grad_norm": 1.190690319783776, | |
| "learning_rate": 4.969378255741879e-05, | |
| "loss": 1.69, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.591767470194065, | |
| "grad_norm": 0.8887720624675962, | |
| "learning_rate": 4.967438552763023e-05, | |
| "loss": 1.7013, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.591767470194065, | |
| "eval_loss": 1.6794394254684448, | |
| "eval_runtime": 14.0089, | |
| "eval_samples_per_second": 71.383, | |
| "eval_steps_per_second": 2.284, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.609172395788008, | |
| "grad_norm": 0.7315204698775325, | |
| "learning_rate": 4.9654397910058326e-05, | |
| "loss": 1.6958, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.6265773213819511, | |
| "grad_norm": 0.9699559384796247, | |
| "learning_rate": 4.963382030469349e-05, | |
| "loss": 1.6788, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.6265773213819511, | |
| "eval_loss": 1.6714988946914673, | |
| "eval_runtime": 14.0096, | |
| "eval_samples_per_second": 71.379, | |
| "eval_steps_per_second": 2.284, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.6439822469758941, | |
| "grad_norm": 0.8473820751122184, | |
| "learning_rate": 4.9612653329236446e-05, | |
| "loss": 1.6879, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.6613871725698373, | |
| "grad_norm": 1.0060512118979683, | |
| "learning_rate": 4.959089761907972e-05, | |
| "loss": 1.6724, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.6613871725698373, | |
| "eval_loss": 1.6676414012908936, | |
| "eval_runtime": 13.9883, | |
| "eval_samples_per_second": 71.488, | |
| "eval_steps_per_second": 2.288, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.6787920981637804, | |
| "grad_norm": 1.6044912820626946, | |
| "learning_rate": 4.95685538272885e-05, | |
| "loss": 1.6811, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.6961970237577234, | |
| "grad_norm": 0.9205203899330789, | |
| "learning_rate": 4.95456226245811e-05, | |
| "loss": 1.6807, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.6961970237577234, | |
| "eval_loss": 1.664583444595337, | |
| "eval_runtime": 13.9774, | |
| "eval_samples_per_second": 71.544, | |
| "eval_steps_per_second": 2.289, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.7136019493516665, | |
| "grad_norm": 1.3251835254853215, | |
| "learning_rate": 4.952210469930877e-05, | |
| "loss": 1.6798, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.7310068749456096, | |
| "grad_norm": 0.9832051746397106, | |
| "learning_rate": 4.949800075743509e-05, | |
| "loss": 1.677, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.7310068749456096, | |
| "eval_loss": 1.6596543788909912, | |
| "eval_runtime": 14.0141, | |
| "eval_samples_per_second": 71.357, | |
| "eval_steps_per_second": 2.283, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.7484118005395527, | |
| "grad_norm": 1.2949214337555768, | |
| "learning_rate": 4.947331152251474e-05, | |
| "loss": 1.6645, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.7658167261334958, | |
| "grad_norm": 1.0901276287938682, | |
| "learning_rate": 4.9448037735671757e-05, | |
| "loss": 1.6681, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.7658167261334958, | |
| "eval_loss": 1.65700364112854, | |
| "eval_runtime": 14.0206, | |
| "eval_samples_per_second": 71.323, | |
| "eval_steps_per_second": 2.282, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.7832216517274389, | |
| "grad_norm": 1.048369032858913, | |
| "learning_rate": 4.942218015557737e-05, | |
| "loss": 1.6695, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.800626577321382, | |
| "grad_norm": 1.211637616168943, | |
| "learning_rate": 4.939573955842714e-05, | |
| "loss": 1.6642, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.800626577321382, | |
| "eval_loss": 1.6546015739440918, | |
| "eval_runtime": 13.9713, | |
| "eval_samples_per_second": 71.575, | |
| "eval_steps_per_second": 2.29, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.818031502915325, | |
| "grad_norm": 1.1547651007185253, | |
| "learning_rate": 4.93687167379177e-05, | |
| "loss": 1.6666, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.8354364285092681, | |
| "grad_norm": 0.8815838845151333, | |
| "learning_rate": 4.934111250522293e-05, | |
| "loss": 1.6704, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.8354364285092681, | |
| "eval_loss": 1.6485086679458618, | |
| "eval_runtime": 13.9864, | |
| "eval_samples_per_second": 71.498, | |
| "eval_steps_per_second": 2.288, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.8528413541032112, | |
| "grad_norm": 1.2788967756098748, | |
| "learning_rate": 4.9312927688969614e-05, | |
| "loss": 1.6665, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.8702462796971543, | |
| "grad_norm": 1.0757149545926434, | |
| "learning_rate": 4.9284163135212544e-05, | |
| "loss": 1.6586, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.8702462796971543, | |
| "eval_loss": 1.6465007066726685, | |
| "eval_runtime": 14.0157, | |
| "eval_samples_per_second": 71.348, | |
| "eval_steps_per_second": 2.283, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.8876512052910974, | |
| "grad_norm": 0.9368397945653969, | |
| "learning_rate": 4.925481970740913e-05, | |
| "loss": 1.6622, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.9050561308850404, | |
| "grad_norm": 1.2877310576716585, | |
| "learning_rate": 4.9224898286393486e-05, | |
| "loss": 1.6532, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.9050561308850404, | |
| "eval_loss": 1.6464436054229736, | |
| "eval_runtime": 14.0065, | |
| "eval_samples_per_second": 71.395, | |
| "eval_steps_per_second": 2.285, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.9224610564789836, | |
| "grad_norm": 1.1714391439428165, | |
| "learning_rate": 4.919439977034997e-05, | |
| "loss": 1.6512, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.9398659820729266, | |
| "grad_norm": 1.11375610693323, | |
| "learning_rate": 4.916332507478631e-05, | |
| "loss": 1.6408, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.9398659820729266, | |
| "eval_loss": 1.6425988674163818, | |
| "eval_runtime": 13.9384, | |
| "eval_samples_per_second": 71.744, | |
| "eval_steps_per_second": 2.296, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.9572709076668697, | |
| "grad_norm": 0.740049431872551, | |
| "learning_rate": 4.913167513250595e-05, | |
| "loss": 1.648, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.9746758332608129, | |
| "grad_norm": 1.353726058408381, | |
| "learning_rate": 4.909945089358022e-05, | |
| "loss": 1.6464, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.9746758332608129, | |
| "eval_loss": 1.6430168151855469, | |
| "eval_runtime": 13.9818, | |
| "eval_samples_per_second": 71.521, | |
| "eval_steps_per_second": 2.289, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.9920807588547559, | |
| "grad_norm": 0.7896432929313317, | |
| "learning_rate": 4.906665332531975e-05, | |
| "loss": 1.6523, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 1.009746758332608, | |
| "grad_norm": 0.7591020143028239, | |
| "learning_rate": 4.903328341224537e-05, | |
| "loss": 1.6531, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.009746758332608, | |
| "eval_loss": 1.636893630027771, | |
| "eval_runtime": 14.0408, | |
| "eval_samples_per_second": 71.221, | |
| "eval_steps_per_second": 2.279, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.0271516839265513, | |
| "grad_norm": 0.785567616684764, | |
| "learning_rate": 4.899934215605867e-05, | |
| "loss": 1.6095, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 1.0445566095204943, | |
| "grad_norm": 0.6496709588539313, | |
| "learning_rate": 4.8964830575611876e-05, | |
| "loss": 1.6169, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.0445566095204943, | |
| "eval_loss": 1.6327561140060425, | |
| "eval_runtime": 14.0121, | |
| "eval_samples_per_second": 71.367, | |
| "eval_steps_per_second": 2.284, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.0619615351144374, | |
| "grad_norm": 1.0307342387456608, | |
| "learning_rate": 4.8929749706877246e-05, | |
| "loss": 1.6061, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 1.0793664607083804, | |
| "grad_norm": 0.8079301722246791, | |
| "learning_rate": 4.8894100602916016e-05, | |
| "loss": 1.6199, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.0793664607083804, | |
| "eval_loss": 1.6337392330169678, | |
| "eval_runtime": 14.0185, | |
| "eval_samples_per_second": 71.334, | |
| "eval_steps_per_second": 2.283, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.0967713863023236, | |
| "grad_norm": 0.7473705601316266, | |
| "learning_rate": 4.8857884333846735e-05, | |
| "loss": 1.6114, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 1.1141763118962666, | |
| "grad_norm": 1.2192071702470884, | |
| "learning_rate": 4.882110198681321e-05, | |
| "loss": 1.6094, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.1141763118962666, | |
| "eval_loss": 1.6318168640136719, | |
| "eval_runtime": 13.9864, | |
| "eval_samples_per_second": 71.498, | |
| "eval_steps_per_second": 2.288, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.1315812374902097, | |
| "grad_norm": 0.9888531694645865, | |
| "learning_rate": 4.878375466595181e-05, | |
| "loss": 1.6139, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.148986163084153, | |
| "grad_norm": 0.8550909287275011, | |
| "learning_rate": 4.874584349235836e-05, | |
| "loss": 1.601, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.148986163084153, | |
| "eval_loss": 1.6282851696014404, | |
| "eval_runtime": 13.9933, | |
| "eval_samples_per_second": 71.463, | |
| "eval_steps_per_second": 2.287, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.166391088678096, | |
| "grad_norm": 1.2063759248291055, | |
| "learning_rate": 4.8707369604054474e-05, | |
| "loss": 1.5998, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 1.183796014272039, | |
| "grad_norm": 0.6341552384920974, | |
| "learning_rate": 4.866833415595341e-05, | |
| "loss": 1.6082, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.183796014272039, | |
| "eval_loss": 1.6227760314941406, | |
| "eval_runtime": 14.032, | |
| "eval_samples_per_second": 71.266, | |
| "eval_steps_per_second": 2.281, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.201200939865982, | |
| "grad_norm": 0.675165295355897, | |
| "learning_rate": 4.8628738319825365e-05, | |
| "loss": 1.5991, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 1.2186058654599252, | |
| "grad_norm": 0.9698012232943588, | |
| "learning_rate": 4.8588583284262324e-05, | |
| "loss": 1.6175, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.2186058654599252, | |
| "eval_loss": 1.6227885484695435, | |
| "eval_runtime": 13.9808, | |
| "eval_samples_per_second": 71.527, | |
| "eval_steps_per_second": 2.289, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.2360107910538682, | |
| "grad_norm": 0.7064806896703955, | |
| "learning_rate": 4.854787025464238e-05, | |
| "loss": 1.6003, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 1.2534157166478113, | |
| "grad_norm": 0.5984772772171928, | |
| "learning_rate": 4.850660045309358e-05, | |
| "loss": 1.6061, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.2534157166478113, | |
| "eval_loss": 1.619645118713379, | |
| "eval_runtime": 14.0435, | |
| "eval_samples_per_second": 71.207, | |
| "eval_steps_per_second": 2.279, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.2708206422417545, | |
| "grad_norm": 0.8428216369485679, | |
| "learning_rate": 4.8464775118457174e-05, | |
| "loss": 1.6033, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 1.2882255678356975, | |
| "grad_norm": 1.0213649998959717, | |
| "learning_rate": 4.842239550625048e-05, | |
| "loss": 1.6101, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.2882255678356975, | |
| "eval_loss": 1.619710922241211, | |
| "eval_runtime": 13.9393, | |
| "eval_samples_per_second": 71.739, | |
| "eval_steps_per_second": 2.296, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.3056304934296405, | |
| "grad_norm": 1.0089461514034055, | |
| "learning_rate": 4.8379462888629166e-05, | |
| "loss": 1.5997, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 1.3230354190235838, | |
| "grad_norm": 0.7527069966901813, | |
| "learning_rate": 4.8335978554349094e-05, | |
| "loss": 1.5978, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.3230354190235838, | |
| "eval_loss": 1.6153150796890259, | |
| "eval_runtime": 13.9935, | |
| "eval_samples_per_second": 71.462, | |
| "eval_steps_per_second": 2.287, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.3404403446175268, | |
| "grad_norm": 0.6599591226592018, | |
| "learning_rate": 4.8291943808727605e-05, | |
| "loss": 1.6024, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 1.3578452702114698, | |
| "grad_norm": 0.7292499200644743, | |
| "learning_rate": 4.8247359973604324e-05, | |
| "loss": 1.6123, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.3578452702114698, | |
| "eval_loss": 1.612821102142334, | |
| "eval_runtime": 13.9912, | |
| "eval_samples_per_second": 71.474, | |
| "eval_steps_per_second": 2.287, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.3752501958054129, | |
| "grad_norm": 1.1210984888321722, | |
| "learning_rate": 4.8202228387301554e-05, | |
| "loss": 1.595, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 1.392655121399356, | |
| "grad_norm": 0.5893087273399503, | |
| "learning_rate": 4.8156550404584e-05, | |
| "loss": 1.6056, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.392655121399356, | |
| "eval_loss": 1.6076596975326538, | |
| "eval_runtime": 14.0138, | |
| "eval_samples_per_second": 71.358, | |
| "eval_steps_per_second": 2.283, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.4100600469932991, | |
| "grad_norm": 1.1601240317764614, | |
| "learning_rate": 4.81103273966182e-05, | |
| "loss": 1.591, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 1.4274649725872421, | |
| "grad_norm": 0.6658774333699882, | |
| "learning_rate": 4.806356075093125e-05, | |
| "loss": 1.5944, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.4274649725872421, | |
| "eval_loss": 1.6094013452529907, | |
| "eval_runtime": 13.9967, | |
| "eval_samples_per_second": 71.445, | |
| "eval_steps_per_second": 2.286, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.4448698981811852, | |
| "grad_norm": 0.9722493291962953, | |
| "learning_rate": 4.801625187136928e-05, | |
| "loss": 1.5919, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 1.4622748237751284, | |
| "grad_norm": 0.6034495360815867, | |
| "learning_rate": 4.796840217805524e-05, | |
| "loss": 1.5971, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.4622748237751284, | |
| "eval_loss": 1.6052404642105103, | |
| "eval_runtime": 14.2403, | |
| "eval_samples_per_second": 70.223, | |
| "eval_steps_per_second": 2.247, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.4796797493690714, | |
| "grad_norm": 0.6329224233323587, | |
| "learning_rate": 4.792001310734627e-05, | |
| "loss": 1.5887, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 1.4970846749630145, | |
| "grad_norm": 0.4863367308791887, | |
| "learning_rate": 4.7871086111790634e-05, | |
| "loss": 1.5889, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.4970846749630145, | |
| "eval_loss": 1.6018826961517334, | |
| "eval_runtime": 13.9744, | |
| "eval_samples_per_second": 71.559, | |
| "eval_steps_per_second": 2.29, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.5144896005569577, | |
| "grad_norm": 0.9659581402091176, | |
| "learning_rate": 4.782162266008402e-05, | |
| "loss": 1.5973, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 1.5318945261509007, | |
| "grad_norm": 0.6230264173988969, | |
| "learning_rate": 4.7771624237025556e-05, | |
| "loss": 1.6011, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.5318945261509007, | |
| "eval_loss": 1.6019304990768433, | |
| "eval_runtime": 14.0183, | |
| "eval_samples_per_second": 71.335, | |
| "eval_steps_per_second": 2.283, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.5492994517448437, | |
| "grad_norm": 0.510207427178355, | |
| "learning_rate": 4.772109234347319e-05, | |
| "loss": 1.5908, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 1.566704377338787, | |
| "grad_norm": 0.93423236140836, | |
| "learning_rate": 4.767002849629866e-05, | |
| "loss": 1.591, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.566704377338787, | |
| "eval_loss": 1.6014844179153442, | |
| "eval_runtime": 13.9746, | |
| "eval_samples_per_second": 71.558, | |
| "eval_steps_per_second": 2.29, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.58410930293273, | |
| "grad_norm": 0.6758808109530141, | |
| "learning_rate": 4.761843422834188e-05, | |
| "loss": 1.5849, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 1.601514228526673, | |
| "grad_norm": 0.5379258818854082, | |
| "learning_rate": 4.756631108836508e-05, | |
| "loss": 1.5854, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.601514228526673, | |
| "eval_loss": 1.5978697538375854, | |
| "eval_runtime": 14.0029, | |
| "eval_samples_per_second": 71.414, | |
| "eval_steps_per_second": 2.285, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.6189191541206163, | |
| "grad_norm": 0.5909606065327258, | |
| "learning_rate": 4.7513660641006165e-05, | |
| "loss": 1.5833, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 1.636324079714559, | |
| "grad_norm": 1.023013373233777, | |
| "learning_rate": 4.746048446673184e-05, | |
| "loss": 1.5873, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.636324079714559, | |
| "eval_loss": 1.5967607498168945, | |
| "eval_runtime": 13.9738, | |
| "eval_samples_per_second": 71.562, | |
| "eval_steps_per_second": 2.29, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.6537290053085023, | |
| "grad_norm": 0.6372756642593507, | |
| "learning_rate": 4.740678416179009e-05, | |
| "loss": 1.581, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 1.6711339309024456, | |
| "grad_norm": 0.5784388626980227, | |
| "learning_rate": 4.735256133816237e-05, | |
| "loss": 1.5755, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.6711339309024456, | |
| "eval_loss": 1.5935559272766113, | |
| "eval_runtime": 14.0039, | |
| "eval_samples_per_second": 71.408, | |
| "eval_steps_per_second": 2.285, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.6885388564963884, | |
| "grad_norm": 0.8001549248943342, | |
| "learning_rate": 4.7297817623515114e-05, | |
| "loss": 1.5896, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 1.7059437820903316, | |
| "grad_norm": 0.9138283791586076, | |
| "learning_rate": 4.724255466115095e-05, | |
| "loss": 1.5898, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.7059437820903316, | |
| "eval_loss": 1.5930613279342651, | |
| "eval_runtime": 14.0797, | |
| "eval_samples_per_second": 71.024, | |
| "eval_steps_per_second": 2.273, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.7233487076842746, | |
| "grad_norm": 0.5583757810052768, | |
| "learning_rate": 4.7186774109959306e-05, | |
| "loss": 1.5719, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 1.7407536332782176, | |
| "grad_norm": 0.6698213257447616, | |
| "learning_rate": 4.7130477644366656e-05, | |
| "loss": 1.5842, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.7407536332782176, | |
| "eval_loss": 1.5888005495071411, | |
| "eval_runtime": 13.9853, | |
| "eval_samples_per_second": 71.503, | |
| "eval_steps_per_second": 2.288, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.7581585588721609, | |
| "grad_norm": 0.729413212458277, | |
| "learning_rate": 4.7073666954286275e-05, | |
| "loss": 1.5882, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 1.775563484466104, | |
| "grad_norm": 0.5635621433923584, | |
| "learning_rate": 4.7016343745067436e-05, | |
| "loss": 1.5845, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.775563484466104, | |
| "eval_loss": 1.5877995491027832, | |
| "eval_runtime": 13.9683, | |
| "eval_samples_per_second": 71.591, | |
| "eval_steps_per_second": 2.291, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.792968410060047, | |
| "grad_norm": 0.7101325374582853, | |
| "learning_rate": 4.69585097374443e-05, | |
| "loss": 1.5815, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 1.8103733356539902, | |
| "grad_norm": 1.0099442818196882, | |
| "learning_rate": 4.690016666748421e-05, | |
| "loss": 1.5754, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.8103733356539902, | |
| "eval_loss": 1.5878440141677856, | |
| "eval_runtime": 13.9816, | |
| "eval_samples_per_second": 71.522, | |
| "eval_steps_per_second": 2.289, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.8277782612479332, | |
| "grad_norm": 0.9861038562211553, | |
| "learning_rate": 4.684131628653562e-05, | |
| "loss": 1.5747, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 1.8451831868418762, | |
| "grad_norm": 0.6476902712711115, | |
| "learning_rate": 4.678196036117547e-05, | |
| "loss": 1.5822, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.8451831868418762, | |
| "eval_loss": 1.5809299945831299, | |
| "eval_runtime": 13.9435, | |
| "eval_samples_per_second": 71.718, | |
| "eval_steps_per_second": 2.295, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.8625881124358195, | |
| "grad_norm": 0.8920609219765819, | |
| "learning_rate": 4.672210067315619e-05, | |
| "loss": 1.5751, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 1.8799930380297625, | |
| "grad_norm": 1.0023436798205103, | |
| "learning_rate": 4.6661739019352244e-05, | |
| "loss": 1.5736, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.8799930380297625, | |
| "eval_loss": 1.5810316801071167, | |
| "eval_runtime": 14.0243, | |
| "eval_samples_per_second": 71.305, | |
| "eval_steps_per_second": 2.282, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.8973979636237055, | |
| "grad_norm": 0.6409767838575461, | |
| "learning_rate": 4.660087721170612e-05, | |
| "loss": 1.5592, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 1.9148028892176487, | |
| "grad_norm": 0.9696993404169995, | |
| "learning_rate": 4.6539517077173975e-05, | |
| "loss": 1.5639, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.9148028892176487, | |
| "eval_loss": 1.5775830745697021, | |
| "eval_runtime": 14.0018, | |
| "eval_samples_per_second": 71.419, | |
| "eval_steps_per_second": 2.285, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.9322078148115915, | |
| "grad_norm": 0.6681985853451967, | |
| "learning_rate": 4.6477660457670835e-05, | |
| "loss": 1.5788, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 1.9496127404055348, | |
| "grad_norm": 0.5491471084373808, | |
| "learning_rate": 4.641530921001523e-05, | |
| "loss": 1.5678, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.9496127404055348, | |
| "eval_loss": 1.575533151626587, | |
| "eval_runtime": 14.0249, | |
| "eval_samples_per_second": 71.302, | |
| "eval_steps_per_second": 2.282, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.9670176659994778, | |
| "grad_norm": 0.6947740445053832, | |
| "learning_rate": 4.63524652058735e-05, | |
| "loss": 1.5732, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 1.9844225915934208, | |
| "grad_norm": 0.597698832959584, | |
| "learning_rate": 4.628913033170359e-05, | |
| "loss": 1.5641, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.9844225915934208, | |
| "eval_loss": 1.5763437747955322, | |
| "eval_runtime": 13.9979, | |
| "eval_samples_per_second": 71.439, | |
| "eval_steps_per_second": 2.286, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 2.0020885910712733, | |
| "grad_norm": 1.3542755699053255, | |
| "learning_rate": 4.622530648869846e-05, | |
| "loss": 1.5802, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 2.019493516665216, | |
| "grad_norm": 0.5422084960348866, | |
| "learning_rate": 4.6160995592728964e-05, | |
| "loss": 1.5088, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 2.019493516665216, | |
| "eval_loss": 1.5745503902435303, | |
| "eval_runtime": 13.9995, | |
| "eval_samples_per_second": 71.431, | |
| "eval_steps_per_second": 2.286, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 2.0368984422591594, | |
| "grad_norm": 0.6504392686205402, | |
| "learning_rate": 4.609619957428636e-05, | |
| "loss": 1.5118, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 2.0543033678531026, | |
| "grad_norm": 1.2584747303351034, | |
| "learning_rate": 4.6030920378424405e-05, | |
| "loss": 1.5085, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 2.0543033678531026, | |
| "eval_loss": 1.5728845596313477, | |
| "eval_runtime": 14.0, | |
| "eval_samples_per_second": 71.429, | |
| "eval_steps_per_second": 2.286, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 2.0717082934470454, | |
| "grad_norm": 0.9297351113236545, | |
| "learning_rate": 4.596515996470088e-05, | |
| "loss": 1.5015, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 2.0891132190409887, | |
| "grad_norm": 0.6770476529726162, | |
| "learning_rate": 4.589892030711882e-05, | |
| "loss": 1.5009, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.0891132190409887, | |
| "eval_loss": 1.5689187049865723, | |
| "eval_runtime": 13.9922, | |
| "eval_samples_per_second": 71.468, | |
| "eval_steps_per_second": 2.287, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.1065181446349315, | |
| "grad_norm": 0.5333302425721433, | |
| "learning_rate": 4.583220339406727e-05, | |
| "loss": 1.5005, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 2.1239230702288747, | |
| "grad_norm": 0.7158614427338245, | |
| "learning_rate": 4.57650112282616e-05, | |
| "loss": 1.4947, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 2.1239230702288747, | |
| "eval_loss": 1.5668717622756958, | |
| "eval_runtime": 13.978, | |
| "eval_samples_per_second": 71.541, | |
| "eval_steps_per_second": 2.289, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 2.141327995822818, | |
| "grad_norm": 0.5418041632358056, | |
| "learning_rate": 4.5697345826683316e-05, | |
| "loss": 1.505, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 2.1587329214167608, | |
| "grad_norm": 0.6349654558761868, | |
| "learning_rate": 4.56292092205196e-05, | |
| "loss": 1.4941, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 2.1587329214167608, | |
| "eval_loss": 1.5645283460617065, | |
| "eval_runtime": 13.9969, | |
| "eval_samples_per_second": 71.445, | |
| "eval_steps_per_second": 2.286, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 2.176137847010704, | |
| "grad_norm": 0.6318855187216622, | |
| "learning_rate": 4.556060345510229e-05, | |
| "loss": 1.4964, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 2.1935427726046473, | |
| "grad_norm": 0.820231743084033, | |
| "learning_rate": 4.5491530589846496e-05, | |
| "loss": 1.5054, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 2.1935427726046473, | |
| "eval_loss": 1.565173864364624, | |
| "eval_runtime": 16.4402, | |
| "eval_samples_per_second": 60.826, | |
| "eval_steps_per_second": 1.946, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 2.21094769819859, | |
| "grad_norm": 0.5808489693725816, | |
| "learning_rate": 4.5421992698188806e-05, | |
| "loss": 1.5161, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 2.2283526237925333, | |
| "grad_norm": 0.7488137844322202, | |
| "learning_rate": 4.5351991867524976e-05, | |
| "loss": 1.4909, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 2.2283526237925333, | |
| "eval_loss": 1.5601595640182495, | |
| "eval_runtime": 13.9913, | |
| "eval_samples_per_second": 71.473, | |
| "eval_steps_per_second": 2.287, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 2.2457575493864765, | |
| "grad_norm": 0.6548395516938676, | |
| "learning_rate": 4.528153019914735e-05, | |
| "loss": 1.4985, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 2.2631624749804193, | |
| "grad_norm": 0.8203597618530264, | |
| "learning_rate": 4.5210609808181716e-05, | |
| "loss": 1.4944, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.2631624749804193, | |
| "eval_loss": 1.5601884126663208, | |
| "eval_runtime": 13.9999, | |
| "eval_samples_per_second": 71.429, | |
| "eval_steps_per_second": 2.286, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.2805674005743626, | |
| "grad_norm": 0.9291833431724984, | |
| "learning_rate": 4.513923282352388e-05, | |
| "loss": 1.4979, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 2.297972326168306, | |
| "grad_norm": 0.6192831769608078, | |
| "learning_rate": 4.506740138777571e-05, | |
| "loss": 1.5013, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 2.297972326168306, | |
| "eval_loss": 1.5595694780349731, | |
| "eval_runtime": 14.0125, | |
| "eval_samples_per_second": 71.365, | |
| "eval_steps_per_second": 2.284, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 2.3153772517622486, | |
| "grad_norm": 0.5252276325498895, | |
| "learning_rate": 4.499511765718082e-05, | |
| "loss": 1.4985, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 2.332782177356192, | |
| "grad_norm": 0.8809278512163949, | |
| "learning_rate": 4.492238380155988e-05, | |
| "loss": 1.5023, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 2.332782177356192, | |
| "eval_loss": 1.556013822555542, | |
| "eval_runtime": 14.0065, | |
| "eval_samples_per_second": 71.395, | |
| "eval_steps_per_second": 2.285, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 2.3501871029501347, | |
| "grad_norm": 2.493527077826374, | |
| "learning_rate": 4.484920200424544e-05, | |
| "loss": 1.5031, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 2.367592028544078, | |
| "grad_norm": 0.5602546682634547, | |
| "learning_rate": 4.4775574462016446e-05, | |
| "loss": 1.4949, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 2.367592028544078, | |
| "eval_loss": 1.5550192594528198, | |
| "eval_runtime": 13.9979, | |
| "eval_samples_per_second": 71.439, | |
| "eval_steps_per_second": 2.286, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 2.384996954138021, | |
| "grad_norm": 0.6370530516407709, | |
| "learning_rate": 4.470150338503223e-05, | |
| "loss": 1.487, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 2.402401879731964, | |
| "grad_norm": 0.6921040878467456, | |
| "learning_rate": 4.462699099676619e-05, | |
| "loss": 1.4834, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 2.402401879731964, | |
| "eval_loss": 1.5544353723526, | |
| "eval_runtime": 13.9865, | |
| "eval_samples_per_second": 71.497, | |
| "eval_steps_per_second": 2.288, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 2.419806805325907, | |
| "grad_norm": 0.8610525793818672, | |
| "learning_rate": 4.455203953393908e-05, | |
| "loss": 1.4912, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 2.4372117309198504, | |
| "grad_norm": 0.7729395434067594, | |
| "learning_rate": 4.447665124645185e-05, | |
| "loss": 1.4978, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.4372117309198504, | |
| "eval_loss": 1.5510929822921753, | |
| "eval_runtime": 13.9851, | |
| "eval_samples_per_second": 71.505, | |
| "eval_steps_per_second": 2.288, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.4546166565137932, | |
| "grad_norm": 0.60728121089851, | |
| "learning_rate": 4.440082839731805e-05, | |
| "loss": 1.4822, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 2.4720215821077365, | |
| "grad_norm": 0.5719460953022453, | |
| "learning_rate": 4.4324573262595994e-05, | |
| "loss": 1.488, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 2.4720215821077365, | |
| "eval_loss": 1.5491801500320435, | |
| "eval_runtime": 13.9637, | |
| "eval_samples_per_second": 71.614, | |
| "eval_steps_per_second": 2.292, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 2.4894265077016797, | |
| "grad_norm": 0.5494920086386093, | |
| "learning_rate": 4.424788813132036e-05, | |
| "loss": 1.4823, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 2.5068314332956225, | |
| "grad_norm": 0.6483803635153482, | |
| "learning_rate": 4.417077530543352e-05, | |
| "loss": 1.498, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 2.5068314332956225, | |
| "eval_loss": 1.5479047298431396, | |
| "eval_runtime": 14.032, | |
| "eval_samples_per_second": 71.266, | |
| "eval_steps_per_second": 2.28, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 2.5242363588895658, | |
| "grad_norm": 0.6182294157167969, | |
| "learning_rate": 4.409323709971643e-05, | |
| "loss": 1.5015, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 2.541641284483509, | |
| "grad_norm": 0.6672439728117495, | |
| "learning_rate": 4.401527584171915e-05, | |
| "loss": 1.4848, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 2.541641284483509, | |
| "eval_loss": 1.5437195301055908, | |
| "eval_runtime": 14.0431, | |
| "eval_samples_per_second": 71.209, | |
| "eval_steps_per_second": 2.279, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 2.559046210077452, | |
| "grad_norm": 0.6569153165541121, | |
| "learning_rate": 4.393689387169097e-05, | |
| "loss": 1.4844, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 2.576451135671395, | |
| "grad_norm": 0.70328233064019, | |
| "learning_rate": 4.38580935425101e-05, | |
| "loss": 1.4846, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 2.576451135671395, | |
| "eval_loss": 1.5450177192687988, | |
| "eval_runtime": 13.9709, | |
| "eval_samples_per_second": 71.578, | |
| "eval_steps_per_second": 2.29, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 2.5938560612653383, | |
| "grad_norm": 0.5537205492155896, | |
| "learning_rate": 4.37788772196132e-05, | |
| "loss": 1.4897, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 2.611260986859281, | |
| "grad_norm": 0.7853101261371833, | |
| "learning_rate": 4.3699247280924195e-05, | |
| "loss": 1.4849, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 2.611260986859281, | |
| "eval_loss": 1.543212890625, | |
| "eval_runtime": 13.9687, | |
| "eval_samples_per_second": 71.589, | |
| "eval_steps_per_second": 2.291, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 2.6286659124532243, | |
| "grad_norm": 0.7915889506298213, | |
| "learning_rate": 4.3619206116782994e-05, | |
| "loss": 1.4888, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 2.6460708380471676, | |
| "grad_norm": 0.5435421494025616, | |
| "learning_rate": 4.353875612987372e-05, | |
| "loss": 1.4903, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 2.6460708380471676, | |
| "eval_loss": 1.5421267747879028, | |
| "eval_runtime": 13.9954, | |
| "eval_samples_per_second": 71.452, | |
| "eval_steps_per_second": 2.286, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 2.6634757636411104, | |
| "grad_norm": 0.7150187738565811, | |
| "learning_rate": 4.3457899735152594e-05, | |
| "loss": 1.4856, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 2.6808806892350536, | |
| "grad_norm": 0.7286431872094431, | |
| "learning_rate": 4.337663935977541e-05, | |
| "loss": 1.4851, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 2.6808806892350536, | |
| "eval_loss": 1.537925362586975, | |
| "eval_runtime": 13.9886, | |
| "eval_samples_per_second": 71.487, | |
| "eval_steps_per_second": 2.288, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 2.698285614828997, | |
| "grad_norm": 0.5315377291160136, | |
| "learning_rate": 4.3294977443024674e-05, | |
| "loss": 1.4799, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 2.7156905404229397, | |
| "grad_norm": 0.7464300338497242, | |
| "learning_rate": 4.3212916436236474e-05, | |
| "loss": 1.4932, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 2.7156905404229397, | |
| "eval_loss": 1.5362157821655273, | |
| "eval_runtime": 13.9746, | |
| "eval_samples_per_second": 71.559, | |
| "eval_steps_per_second": 2.29, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 2.733095466016883, | |
| "grad_norm": 0.7735626869311409, | |
| "learning_rate": 4.313045880272675e-05, | |
| "loss": 1.4868, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 2.7505003916108257, | |
| "grad_norm": 0.7079246653886904, | |
| "learning_rate": 4.304760701771747e-05, | |
| "loss": 1.4736, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 2.7505003916108257, | |
| "eval_loss": 1.5344328880310059, | |
| "eval_runtime": 13.9862, | |
| "eval_samples_per_second": 71.499, | |
| "eval_steps_per_second": 2.288, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 2.767905317204769, | |
| "grad_norm": 0.5607638781051937, | |
| "learning_rate": 4.296436356826226e-05, | |
| "loss": 1.4828, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 2.785310242798712, | |
| "grad_norm": 0.6698483645730473, | |
| "learning_rate": 4.2880730953171786e-05, | |
| "loss": 1.4818, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 2.785310242798712, | |
| "eval_loss": 1.5319961309432983, | |
| "eval_runtime": 14.013, | |
| "eval_samples_per_second": 71.362, | |
| "eval_steps_per_second": 2.284, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 2.802715168392655, | |
| "grad_norm": 0.503273498716246, | |
| "learning_rate": 4.279671168293873e-05, | |
| "loss": 1.4715, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 2.8201200939865982, | |
| "grad_norm": 0.6138361015012737, | |
| "learning_rate": 4.2712308279662405e-05, | |
| "loss": 1.4833, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 2.8201200939865982, | |
| "eval_loss": 1.5272479057312012, | |
| "eval_runtime": 14.0366, | |
| "eval_samples_per_second": 71.242, | |
| "eval_steps_per_second": 2.28, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 2.837525019580541, | |
| "grad_norm": 0.6800768745825728, | |
| "learning_rate": 4.2627523276973114e-05, | |
| "loss": 1.4758, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 2.8549299451744843, | |
| "grad_norm": 0.7176069315448348, | |
| "learning_rate": 4.254235921995601e-05, | |
| "loss": 1.4746, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 2.8549299451744843, | |
| "eval_loss": 1.5255234241485596, | |
| "eval_runtime": 13.9681, | |
| "eval_samples_per_second": 71.592, | |
| "eval_steps_per_second": 2.291, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 2.8723348707684275, | |
| "grad_norm": 0.7304682952222467, | |
| "learning_rate": 4.245681866507474e-05, | |
| "loss": 1.4712, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 2.8897397963623703, | |
| "grad_norm": 0.6236989319439076, | |
| "learning_rate": 4.237090418009474e-05, | |
| "loss": 1.4726, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 2.8897397963623703, | |
| "eval_loss": 1.5243220329284668, | |
| "eval_runtime": 13.986, | |
| "eval_samples_per_second": 71.5, | |
| "eval_steps_per_second": 2.288, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 2.9071447219563136, | |
| "grad_norm": 0.615639176805716, | |
| "learning_rate": 4.228461834400609e-05, | |
| "loss": 1.4768, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 2.924549647550257, | |
| "grad_norm": 0.7789593131085253, | |
| "learning_rate": 4.219796374694612e-05, | |
| "loss": 1.4696, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 2.924549647550257, | |
| "eval_loss": 1.5228594541549683, | |
| "eval_runtime": 14.0002, | |
| "eval_samples_per_second": 71.428, | |
| "eval_steps_per_second": 2.286, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 2.9419545731441996, | |
| "grad_norm": 0.6541490416672233, | |
| "learning_rate": 4.2110942990121707e-05, | |
| "loss": 1.4683, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 2.959359498738143, | |
| "grad_norm": 0.8019396630848001, | |
| "learning_rate": 4.20235586857311e-05, | |
| "loss": 1.4769, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 2.959359498738143, | |
| "eval_loss": 1.5207167863845825, | |
| "eval_runtime": 14.002, | |
| "eval_samples_per_second": 71.419, | |
| "eval_steps_per_second": 2.285, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 2.976764424332086, | |
| "grad_norm": 0.4872815656219925, | |
| "learning_rate": 4.193581345688559e-05, | |
| "loss": 1.4769, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 2.994169349926029, | |
| "grad_norm": 0.48438009286578315, | |
| "learning_rate": 4.184770993753072e-05, | |
| "loss": 1.4727, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 2.994169349926029, | |
| "eval_loss": 1.5173397064208984, | |
| "eval_runtime": 14.0039, | |
| "eval_samples_per_second": 71.408, | |
| "eval_steps_per_second": 2.285, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 3.0118353494038814, | |
| "grad_norm": 0.5172037034713615, | |
| "learning_rate": 4.175925077236725e-05, | |
| "loss": 1.4341, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 3.0292402749978242, | |
| "grad_norm": 0.7660981512215144, | |
| "learning_rate": 4.167043861677175e-05, | |
| "loss": 1.388, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 3.0292402749978242, | |
| "eval_loss": 1.5206201076507568, | |
| "eval_runtime": 14.0092, | |
| "eval_samples_per_second": 71.382, | |
| "eval_steps_per_second": 2.284, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 3.0466452005917675, | |
| "grad_norm": 0.7259175655214019, | |
| "learning_rate": 4.158127613671687e-05, | |
| "loss": 1.3853, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 3.0640501261857107, | |
| "grad_norm": 0.759326231296325, | |
| "learning_rate": 4.149176600869138e-05, | |
| "loss": 1.3921, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 3.0640501261857107, | |
| "eval_loss": 1.5212202072143555, | |
| "eval_runtime": 13.9904, | |
| "eval_samples_per_second": 71.477, | |
| "eval_steps_per_second": 2.287, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 3.0814550517796535, | |
| "grad_norm": 0.6816156678791402, | |
| "learning_rate": 4.1401910919619735e-05, | |
| "loss": 1.3935, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 3.0988599773735968, | |
| "grad_norm": 0.6716116357078614, | |
| "learning_rate": 4.131171356678151e-05, | |
| "loss": 1.3949, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 3.0988599773735968, | |
| "eval_loss": 1.5215586423873901, | |
| "eval_runtime": 13.9752, | |
| "eval_samples_per_second": 71.555, | |
| "eval_steps_per_second": 2.29, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 3.11626490296754, | |
| "grad_norm": 0.585427370042915, | |
| "learning_rate": 4.1221176657730366e-05, | |
| "loss": 1.3937, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 3.133669828561483, | |
| "grad_norm": 0.5070038039825276, | |
| "learning_rate": 4.11303029102128e-05, | |
| "loss": 1.3905, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 3.133669828561483, | |
| "eval_loss": 1.5187056064605713, | |
| "eval_runtime": 13.9925, | |
| "eval_samples_per_second": 71.467, | |
| "eval_steps_per_second": 2.287, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 3.151074754155426, | |
| "grad_norm": 0.5782658425686112, | |
| "learning_rate": 4.103909505208659e-05, | |
| "loss": 1.3818, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 3.1684796797493693, | |
| "grad_norm": 0.6971195268104966, | |
| "learning_rate": 4.094755582123881e-05, | |
| "loss": 1.4096, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 3.1684796797493693, | |
| "eval_loss": 1.513555884361267, | |
| "eval_runtime": 14.0414, | |
| "eval_samples_per_second": 71.218, | |
| "eval_steps_per_second": 2.279, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 3.185884605343312, | |
| "grad_norm": 0.601333503904311, | |
| "learning_rate": 4.085568796550381e-05, | |
| "loss": 1.3914, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 3.2032895309372553, | |
| "grad_norm": 0.596102911328116, | |
| "learning_rate": 4.076349424258056e-05, | |
| "loss": 1.3898, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 3.2032895309372553, | |
| "eval_loss": 1.513745903968811, | |
| "eval_runtime": 14.012, | |
| "eval_samples_per_second": 71.367, | |
| "eval_steps_per_second": 2.284, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 3.220694456531198, | |
| "grad_norm": 0.6397943716093005, | |
| "learning_rate": 4.0670977419949985e-05, | |
| "loss": 1.3926, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 3.2380993821251414, | |
| "grad_norm": 0.5166013726068385, | |
| "learning_rate": 4.057814027479184e-05, | |
| "loss": 1.3874, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 3.2380993821251414, | |
| "eval_loss": 1.513095736503601, | |
| "eval_runtime": 13.9776, | |
| "eval_samples_per_second": 71.543, | |
| "eval_steps_per_second": 2.289, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 3.2555043077190846, | |
| "grad_norm": 0.5888457174066127, | |
| "learning_rate": 4.048498559390136e-05, | |
| "loss": 1.3818, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 3.2729092333130274, | |
| "grad_norm": 0.6156597160103343, | |
| "learning_rate": 4.0391516173605594e-05, | |
| "loss": 1.3935, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 3.2729092333130274, | |
| "eval_loss": 1.5117729902267456, | |
| "eval_runtime": 13.9905, | |
| "eval_samples_per_second": 71.477, | |
| "eval_steps_per_second": 2.287, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 3.2903141589069707, | |
| "grad_norm": 0.7131915538353129, | |
| "learning_rate": 4.029773481967948e-05, | |
| "loss": 1.3979, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 3.307719084500914, | |
| "grad_norm": 0.5758254019327259, | |
| "learning_rate": 4.020364434726159e-05, | |
| "loss": 1.4087, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 3.307719084500914, | |
| "eval_loss": 1.5071306228637695, | |
| "eval_runtime": 13.9596, | |
| "eval_samples_per_second": 71.636, | |
| "eval_steps_per_second": 2.292, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 3.3251240100948567, | |
| "grad_norm": 0.5799334943773048, | |
| "learning_rate": 4.010924758076968e-05, | |
| "loss": 1.3959, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 3.3425289356888, | |
| "grad_norm": 0.6537221695847955, | |
| "learning_rate": 4.001454735381587e-05, | |
| "loss": 1.3947, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 3.3425289356888, | |
| "eval_loss": 1.510569453239441, | |
| "eval_runtime": 14.0089, | |
| "eval_samples_per_second": 71.383, | |
| "eval_steps_per_second": 2.284, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 3.359933861282743, | |
| "grad_norm": 0.5409192150645737, | |
| "learning_rate": 3.9919546509121556e-05, | |
| "loss": 1.3855, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 3.377338786876686, | |
| "grad_norm": 0.5241132458551793, | |
| "learning_rate": 3.982424789843215e-05, | |
| "loss": 1.3831, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 3.377338786876686, | |
| "eval_loss": 1.5041130781173706, | |
| "eval_runtime": 14.0156, | |
| "eval_samples_per_second": 71.349, | |
| "eval_steps_per_second": 2.283, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 3.3947437124706292, | |
| "grad_norm": 1.3217192216589941, | |
| "learning_rate": 3.972865438243141e-05, | |
| "loss": 1.3951, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 3.412148638064572, | |
| "grad_norm": 0.6672183386467068, | |
| "learning_rate": 3.963276883065559e-05, | |
| "loss": 1.3984, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 3.412148638064572, | |
| "eval_loss": 1.5004818439483643, | |
| "eval_runtime": 13.9869, | |
| "eval_samples_per_second": 71.495, | |
| "eval_steps_per_second": 2.288, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 3.4295535636585153, | |
| "grad_norm": 0.6864261952368358, | |
| "learning_rate": 3.953659412140731e-05, | |
| "loss": 1.3919, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 3.4469584892524585, | |
| "grad_norm": 0.8020893138986491, | |
| "learning_rate": 3.944013314166918e-05, | |
| "loss": 1.3901, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 3.4469584892524585, | |
| "eval_loss": 1.4998400211334229, | |
| "eval_runtime": 13.9738, | |
| "eval_samples_per_second": 71.562, | |
| "eval_steps_per_second": 2.29, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 3.4643634148464013, | |
| "grad_norm": 0.5280983708253972, | |
| "learning_rate": 3.9343388787017045e-05, | |
| "loss": 1.3925, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 3.4817683404403446, | |
| "grad_norm": 0.6261045706297181, | |
| "learning_rate": 3.92463639615332e-05, | |
| "loss": 1.3803, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 3.4817683404403446, | |
| "eval_loss": 1.494409441947937, | |
| "eval_runtime": 13.9651, | |
| "eval_samples_per_second": 71.607, | |
| "eval_steps_per_second": 2.291, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 3.499173266034288, | |
| "grad_norm": 0.6077008396812058, | |
| "learning_rate": 3.914906157771907e-05, | |
| "loss": 1.3903, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 3.5165781916282306, | |
| "grad_norm": 0.6825591451005114, | |
| "learning_rate": 3.905148455640793e-05, | |
| "loss": 1.3869, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 3.5165781916282306, | |
| "eval_loss": 1.4927072525024414, | |
| "eval_runtime": 14.0115, | |
| "eval_samples_per_second": 71.37, | |
| "eval_steps_per_second": 2.284, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 3.533983117222174, | |
| "grad_norm": 0.7127239486947049, | |
| "learning_rate": 3.895363582667708e-05, | |
| "loss": 1.3877, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 3.551388042816117, | |
| "grad_norm": 0.6124269855188678, | |
| "learning_rate": 3.885551832576005e-05, | |
| "loss": 1.3976, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 3.551388042816117, | |
| "eval_loss": 1.4895879030227661, | |
| "eval_runtime": 14.0183, | |
| "eval_samples_per_second": 71.336, | |
| "eval_steps_per_second": 2.283, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 3.56879296841006, | |
| "grad_norm": 0.5087923983285213, | |
| "learning_rate": 3.875713499895833e-05, | |
| "loss": 1.3923, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 3.586197894004003, | |
| "grad_norm": 0.6500951563330185, | |
| "learning_rate": 3.865848879955305e-05, | |
| "loss": 1.3777, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 3.586197894004003, | |
| "eval_loss": 1.488155484199524, | |
| "eval_runtime": 13.9697, | |
| "eval_samples_per_second": 71.583, | |
| "eval_steps_per_second": 2.291, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 3.6036028195979464, | |
| "grad_norm": 0.620126750984765, | |
| "learning_rate": 3.8559582688716215e-05, | |
| "loss": 1.3705, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 3.621007745191889, | |
| "grad_norm": 0.593411564329284, | |
| "learning_rate": 3.8460419635421905e-05, | |
| "loss": 1.3854, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 3.621007745191889, | |
| "eval_loss": 1.490203857421875, | |
| "eval_runtime": 14.0195, | |
| "eval_samples_per_second": 71.329, | |
| "eval_steps_per_second": 2.283, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 3.6384126707858324, | |
| "grad_norm": 0.6082624242172648, | |
| "learning_rate": 3.8361002616357164e-05, | |
| "loss": 1.3963, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 3.6558175963797757, | |
| "grad_norm": 0.5432316719761062, | |
| "learning_rate": 3.8261334615832545e-05, | |
| "loss": 1.3906, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 3.6558175963797757, | |
| "eval_loss": 1.486801266670227, | |
| "eval_runtime": 14.0284, | |
| "eval_samples_per_second": 71.284, | |
| "eval_steps_per_second": 2.281, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 3.6732225219737185, | |
| "grad_norm": 0.45833443577890215, | |
| "learning_rate": 3.8161418625692626e-05, | |
| "loss": 1.3954, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 3.6906274475676617, | |
| "grad_norm": 0.6292682710059461, | |
| "learning_rate": 3.806125764522616e-05, | |
| "loss": 1.3899, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 3.6906274475676617, | |
| "eval_loss": 1.485595703125, | |
| "eval_runtime": 14.0352, | |
| "eval_samples_per_second": 71.249, | |
| "eval_steps_per_second": 2.28, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 3.708032373161605, | |
| "grad_norm": 0.6398479114321272, | |
| "learning_rate": 3.796085468107603e-05, | |
| "loss": 1.3785, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 3.7254372987555477, | |
| "grad_norm": 0.9194838860230665, | |
| "learning_rate": 3.7860212747149026e-05, | |
| "loss": 1.3943, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 3.7254372987555477, | |
| "eval_loss": 1.4841216802597046, | |
| "eval_runtime": 13.9765, | |
| "eval_samples_per_second": 71.548, | |
| "eval_steps_per_second": 2.29, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 3.742842224349491, | |
| "grad_norm": 0.6034510757807239, | |
| "learning_rate": 3.775933486452535e-05, | |
| "loss": 1.3871, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 3.7602471499434342, | |
| "grad_norm": 0.549592791064493, | |
| "learning_rate": 3.765822406136796e-05, | |
| "loss": 1.385, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 3.7602471499434342, | |
| "eval_loss": 1.4785528182983398, | |
| "eval_runtime": 14.0146, | |
| "eval_samples_per_second": 71.354, | |
| "eval_steps_per_second": 2.283, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 3.777652075537377, | |
| "grad_norm": 3.923057882484424, | |
| "learning_rate": 3.755688337283161e-05, | |
| "loss": 1.3735, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 3.7950570011313203, | |
| "grad_norm": 0.6208751963635791, | |
| "learning_rate": 3.745531584097179e-05, | |
| "loss": 1.3885, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 3.7950570011313203, | |
| "eval_loss": 1.4790866374969482, | |
| "eval_runtime": 13.9448, | |
| "eval_samples_per_second": 71.711, | |
| "eval_steps_per_second": 2.295, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 3.8124619267252635, | |
| "grad_norm": 0.514334144822583, | |
| "learning_rate": 3.735352451465342e-05, | |
| "loss": 1.3779, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 3.8298668523192063, | |
| "grad_norm": 0.5605077057085409, | |
| "learning_rate": 3.725151244945926e-05, | |
| "loss": 1.3871, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 3.8298668523192063, | |
| "eval_loss": 1.473836898803711, | |
| "eval_runtime": 13.9804, | |
| "eval_samples_per_second": 71.529, | |
| "eval_steps_per_second": 2.289, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 3.8472717779131496, | |
| "grad_norm": 0.5589049809633155, | |
| "learning_rate": 3.7149282707598284e-05, | |
| "loss": 1.3788, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 3.8646767035070924, | |
| "grad_norm": 0.8586729877222112, | |
| "learning_rate": 3.704683835781368e-05, | |
| "loss": 1.3913, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 3.8646767035070924, | |
| "eval_loss": 1.4766095876693726, | |
| "eval_runtime": 13.9861, | |
| "eval_samples_per_second": 71.5, | |
| "eval_steps_per_second": 2.288, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 3.8820816291010356, | |
| "grad_norm": 0.43899438984382605, | |
| "learning_rate": 3.6944182475290755e-05, | |
| "loss": 1.3803, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 3.8994865546949784, | |
| "grad_norm": 0.48403601878369473, | |
| "learning_rate": 3.684131814156465e-05, | |
| "loss": 1.3765, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 3.8994865546949784, | |
| "eval_loss": 1.471179723739624, | |
| "eval_runtime": 13.9803, | |
| "eval_samples_per_second": 71.529, | |
| "eval_steps_per_second": 2.289, | |
| "step": 11200 | |
| } | |
| ], | |
| "logging_steps": 50, | |
| "max_steps": 28720, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 800, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.086550139994112e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |