| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 1056, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.014204545454545454, | |
| "grad_norm": 16.233426918647123, | |
| "learning_rate": 3.7735849056603773e-06, | |
| "loss": 1.9367, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.028409090909090908, | |
| "grad_norm": 3.700378147836118, | |
| "learning_rate": 8.49056603773585e-06, | |
| "loss": 1.5623, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.04261363636363636, | |
| "grad_norm": 1.411828049687858, | |
| "learning_rate": 1.320754716981132e-05, | |
| "loss": 1.0659, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.056818181818181816, | |
| "grad_norm": 0.5502005679393165, | |
| "learning_rate": 1.7924528301886792e-05, | |
| "loss": 0.876, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07102272727272728, | |
| "grad_norm": 0.39964518009977995, | |
| "learning_rate": 2.2641509433962265e-05, | |
| "loss": 0.7709, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.08522727272727272, | |
| "grad_norm": 0.2819938310171901, | |
| "learning_rate": 2.7358490566037738e-05, | |
| "loss": 0.7212, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.09943181818181818, | |
| "grad_norm": 0.22596397421278658, | |
| "learning_rate": 3.207547169811321e-05, | |
| "loss": 0.6941, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.11363636363636363, | |
| "grad_norm": 0.19821070636934998, | |
| "learning_rate": 3.679245283018868e-05, | |
| "loss": 0.6626, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.1278409090909091, | |
| "grad_norm": 0.1776825768871472, | |
| "learning_rate": 4.150943396226415e-05, | |
| "loss": 0.6458, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.14204545454545456, | |
| "grad_norm": 0.15959969460809642, | |
| "learning_rate": 4.6226415094339625e-05, | |
| "loss": 0.6317, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.15625, | |
| "grad_norm": 0.16520239332693426, | |
| "learning_rate": 4.999988963025308e-05, | |
| "loss": 0.6111, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.17045454545454544, | |
| "grad_norm": 0.14604811373242252, | |
| "learning_rate": 4.999602680280325e-05, | |
| "loss": 0.6049, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1846590909090909, | |
| "grad_norm": 0.153405093129135, | |
| "learning_rate": 4.998664657075762e-05, | |
| "loss": 0.5878, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.19886363636363635, | |
| "grad_norm": 0.13290633271306487, | |
| "learning_rate": 4.9971751234724005e-05, | |
| "loss": 0.5872, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.21306818181818182, | |
| "grad_norm": 0.14464611604457972, | |
| "learning_rate": 4.995134444795177e-05, | |
| "loss": 0.5809, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.22727272727272727, | |
| "grad_norm": 0.1539967921458611, | |
| "learning_rate": 4.992543121543581e-05, | |
| "loss": 0.5687, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.24147727272727273, | |
| "grad_norm": 0.1872261556450498, | |
| "learning_rate": 4.9894017892689e-05, | |
| "loss": 0.5736, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.2556818181818182, | |
| "grad_norm": 0.1461192468496281, | |
| "learning_rate": 4.985711218418346e-05, | |
| "loss": 0.5729, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.26988636363636365, | |
| "grad_norm": 0.15510991989675932, | |
| "learning_rate": 4.981472314146093e-05, | |
| "loss": 0.5651, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.2840909090909091, | |
| "grad_norm": 0.15383826812713372, | |
| "learning_rate": 4.976686116091281e-05, | |
| "loss": 0.5669, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.29829545454545453, | |
| "grad_norm": 0.12937769599328974, | |
| "learning_rate": 4.971353798123026e-05, | |
| "loss": 0.5608, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.3125, | |
| "grad_norm": 0.14866942632641453, | |
| "learning_rate": 4.965476668052522e-05, | |
| "loss": 0.5547, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.32670454545454547, | |
| "grad_norm": 0.15767211601593334, | |
| "learning_rate": 4.959056167312283e-05, | |
| "loss": 0.5572, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.3409090909090909, | |
| "grad_norm": 0.16235663995148145, | |
| "learning_rate": 4.952093870602615e-05, | |
| "loss": 0.5613, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.35511363636363635, | |
| "grad_norm": 0.15329109428934323, | |
| "learning_rate": 4.944591485505405e-05, | |
| "loss": 0.5515, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.3693181818181818, | |
| "grad_norm": 0.16475626406284527, | |
| "learning_rate": 4.93655085206531e-05, | |
| "loss": 0.5538, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.3835227272727273, | |
| "grad_norm": 0.17396494811540575, | |
| "learning_rate": 4.92797394233848e-05, | |
| "loss": 0.5455, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.3977272727272727, | |
| "grad_norm": 0.1684931659310488, | |
| "learning_rate": 4.918862859908871e-05, | |
| "loss": 0.5439, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.4119318181818182, | |
| "grad_norm": 0.15970431189153267, | |
| "learning_rate": 4.909219839372333e-05, | |
| "loss": 0.5431, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.42613636363636365, | |
| "grad_norm": 0.1333165751370488, | |
| "learning_rate": 4.8990472457885386e-05, | |
| "loss": 0.5367, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.4403409090909091, | |
| "grad_norm": 0.15417885622445218, | |
| "learning_rate": 4.8883475741009324e-05, | |
| "loss": 0.5381, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.45454545454545453, | |
| "grad_norm": 0.1416173613974234, | |
| "learning_rate": 4.877123448524815e-05, | |
| "loss": 0.5356, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.46875, | |
| "grad_norm": 0.13087882893136601, | |
| "learning_rate": 4.865377621903726e-05, | |
| "loss": 0.5288, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.48295454545454547, | |
| "grad_norm": 0.14764636629826566, | |
| "learning_rate": 4.853112975034276e-05, | |
| "loss": 0.5347, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.4971590909090909, | |
| "grad_norm": 0.16758747023050016, | |
| "learning_rate": 4.840332515959601e-05, | |
| "loss": 0.5372, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.5113636363636364, | |
| "grad_norm": 0.18224306322957698, | |
| "learning_rate": 4.827039379231604e-05, | |
| "loss": 0.5275, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.5255681818181818, | |
| "grad_norm": 0.1662865774101808, | |
| "learning_rate": 4.813236825142173e-05, | |
| "loss": 0.5408, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.5397727272727273, | |
| "grad_norm": 0.17763881404547607, | |
| "learning_rate": 4.798928238923555e-05, | |
| "loss": 0.5289, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.5539772727272727, | |
| "grad_norm": 0.16091215915603133, | |
| "learning_rate": 4.7841171299180906e-05, | |
| "loss": 0.5397, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.5681818181818182, | |
| "grad_norm": 0.1404358353075446, | |
| "learning_rate": 4.768807130717508e-05, | |
| "loss": 0.5256, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.5823863636363636, | |
| "grad_norm": 0.166459644318507, | |
| "learning_rate": 4.753001996271991e-05, | |
| "loss": 0.5235, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.5965909090909091, | |
| "grad_norm": 0.13606607208723026, | |
| "learning_rate": 4.73670560296923e-05, | |
| "loss": 0.5291, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.6107954545454546, | |
| "grad_norm": 0.1370837785726949, | |
| "learning_rate": 4.7199219476837034e-05, | |
| "loss": 0.5254, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.625, | |
| "grad_norm": 0.130041325528082, | |
| "learning_rate": 4.702655146796385e-05, | |
| "loss": 0.5281, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.6392045454545454, | |
| "grad_norm": 0.13838008337473828, | |
| "learning_rate": 4.684909435185169e-05, | |
| "loss": 0.5163, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.6534090909090909, | |
| "grad_norm": 0.1249145523375738, | |
| "learning_rate": 4.6666891651862096e-05, | |
| "loss": 0.5164, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.6676136363636364, | |
| "grad_norm": 0.11918127255464575, | |
| "learning_rate": 4.6479988055264654e-05, | |
| "loss": 0.5156, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.6818181818181818, | |
| "grad_norm": 0.13410033100059227, | |
| "learning_rate": 4.628842940227693e-05, | |
| "loss": 0.5267, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.6960227272727273, | |
| "grad_norm": 0.14513137581304156, | |
| "learning_rate": 4.609226267482162e-05, | |
| "loss": 0.516, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.7102272727272727, | |
| "grad_norm": 0.17472387433508568, | |
| "learning_rate": 4.589153598500372e-05, | |
| "loss": 0.5103, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.7244318181818182, | |
| "grad_norm": 0.13807513906855917, | |
| "learning_rate": 4.568629856331049e-05, | |
| "loss": 0.5111, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.7386363636363636, | |
| "grad_norm": 0.1606554123531758, | |
| "learning_rate": 4.547660074653713e-05, | |
| "loss": 0.5106, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.7528409090909091, | |
| "grad_norm": 0.13512067917257073, | |
| "learning_rate": 4.526249396544107e-05, | |
| "loss": 0.52, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.7670454545454546, | |
| "grad_norm": 0.12966854507444983, | |
| "learning_rate": 4.504403073212805e-05, | |
| "loss": 0.5137, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.78125, | |
| "grad_norm": 0.14417839979685987, | |
| "learning_rate": 4.482126462717289e-05, | |
| "loss": 0.5131, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.7954545454545454, | |
| "grad_norm": 0.14967027976448785, | |
| "learning_rate": 4.45942502864783e-05, | |
| "loss": 0.5164, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.8096590909090909, | |
| "grad_norm": 0.13862324342167873, | |
| "learning_rate": 4.436304338787472e-05, | |
| "loss": 0.5083, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.8238636363636364, | |
| "grad_norm": 0.1285629705362091, | |
| "learning_rate": 4.4127700637464834e-05, | |
| "loss": 0.5196, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.8380681818181818, | |
| "grad_norm": 0.12976169635746898, | |
| "learning_rate": 4.388827975571561e-05, | |
| "loss": 0.5125, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.8522727272727273, | |
| "grad_norm": 0.14210550405367858, | |
| "learning_rate": 4.364483946330182e-05, | |
| "loss": 0.5051, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.8664772727272727, | |
| "grad_norm": 0.16907430743029442, | |
| "learning_rate": 4.3397439466704074e-05, | |
| "loss": 0.5135, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.8806818181818182, | |
| "grad_norm": 0.16584741007419773, | |
| "learning_rate": 4.314614044356517e-05, | |
| "loss": 0.5104, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.8948863636363636, | |
| "grad_norm": 0.15722645631531904, | |
| "learning_rate": 4.289100402780813e-05, | |
| "loss": 0.5044, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.9090909090909091, | |
| "grad_norm": 0.13402395269180703, | |
| "learning_rate": 4.2632092794519926e-05, | |
| "loss": 0.5126, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.9232954545454546, | |
| "grad_norm": 0.12897493665476578, | |
| "learning_rate": 4.236947024460409e-05, | |
| "loss": 0.5073, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.9375, | |
| "grad_norm": 0.12307110209575203, | |
| "learning_rate": 4.2103200789206496e-05, | |
| "loss": 0.5064, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.9517045454545454, | |
| "grad_norm": 0.13103687604123465, | |
| "learning_rate": 4.1833349733917774e-05, | |
| "loss": 0.5018, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.9659090909090909, | |
| "grad_norm": 0.1441008725503634, | |
| "learning_rate": 4.155998326275636e-05, | |
| "loss": 0.5016, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.9801136363636364, | |
| "grad_norm": 0.14676840920070153, | |
| "learning_rate": 4.128316842193612e-05, | |
| "loss": 0.5017, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.9943181818181818, | |
| "grad_norm": 0.1299772490214553, | |
| "learning_rate": 4.100297310342252e-05, | |
| "loss": 0.5084, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.0085227272727273, | |
| "grad_norm": 0.11841362182036683, | |
| "learning_rate": 4.071946602828136e-05, | |
| "loss": 0.484, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.0227272727272727, | |
| "grad_norm": 0.1320677520863094, | |
| "learning_rate": 4.043271672982416e-05, | |
| "loss": 0.4757, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.0369318181818181, | |
| "grad_norm": 0.1525515212145726, | |
| "learning_rate": 4.014279553655429e-05, | |
| "loss": 0.4712, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.0511363636363635, | |
| "grad_norm": 0.12388717475970465, | |
| "learning_rate": 3.984977355491816e-05, | |
| "loss": 0.4741, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.0653409090909092, | |
| "grad_norm": 0.12285428485981521, | |
| "learning_rate": 3.9553722651865525e-05, | |
| "loss": 0.4805, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.0795454545454546, | |
| "grad_norm": 0.14068486142094294, | |
| "learning_rate": 3.925471543722334e-05, | |
| "loss": 0.476, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.09375, | |
| "grad_norm": 0.11956188719769793, | |
| "learning_rate": 3.895282524588736e-05, | |
| "loss": 0.4744, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.1079545454545454, | |
| "grad_norm": 0.12566341647227572, | |
| "learning_rate": 3.8648126119835895e-05, | |
| "loss": 0.4728, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.1221590909090908, | |
| "grad_norm": 0.11910968307671285, | |
| "learning_rate": 3.834069278997025e-05, | |
| "loss": 0.4672, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.1363636363636362, | |
| "grad_norm": 0.1357524801191257, | |
| "learning_rate": 3.803060065778607e-05, | |
| "loss": 0.4765, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.1505681818181819, | |
| "grad_norm": 0.12366787866403461, | |
| "learning_rate": 3.77179257768803e-05, | |
| "loss": 0.4769, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.1647727272727273, | |
| "grad_norm": 0.1322634035357931, | |
| "learning_rate": 3.740274483429815e-05, | |
| "loss": 0.4808, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.1789772727272727, | |
| "grad_norm": 0.13258784372135785, | |
| "learning_rate": 3.708513513172481e-05, | |
| "loss": 0.4763, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.1931818181818181, | |
| "grad_norm": 0.1356892941793857, | |
| "learning_rate": 3.67651745665262e-05, | |
| "loss": 0.4772, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.2073863636363638, | |
| "grad_norm": 0.12523099567966772, | |
| "learning_rate": 3.644294161264394e-05, | |
| "loss": 0.4691, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.2215909090909092, | |
| "grad_norm": 0.12166675336360953, | |
| "learning_rate": 3.611851530134856e-05, | |
| "loss": 0.4763, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.2357954545454546, | |
| "grad_norm": 0.11908896334134955, | |
| "learning_rate": 3.579197520185628e-05, | |
| "loss": 0.4772, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 0.1273632443581102, | |
| "learning_rate": 3.546340140181376e-05, | |
| "loss": 0.4887, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.2642045454545454, | |
| "grad_norm": 0.11548733790030413, | |
| "learning_rate": 3.5132874487655685e-05, | |
| "loss": 0.4688, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.2784090909090908, | |
| "grad_norm": 0.1251859404517232, | |
| "learning_rate": 3.480047552484001e-05, | |
| "loss": 0.4695, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.2926136363636362, | |
| "grad_norm": 0.11622578447126647, | |
| "learning_rate": 3.4466286037965786e-05, | |
| "loss": 0.476, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.3068181818181819, | |
| "grad_norm": 0.12517918383447438, | |
| "learning_rate": 3.4130387990778286e-05, | |
| "loss": 0.4805, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.3210227272727273, | |
| "grad_norm": 0.12611864483533994, | |
| "learning_rate": 3.379286376606651e-05, | |
| "loss": 0.4754, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.3352272727272727, | |
| "grad_norm": 0.13719373565556317, | |
| "learning_rate": 3.3453796145457835e-05, | |
| "loss": 0.4659, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.3494318181818181, | |
| "grad_norm": 0.11875949906730647, | |
| "learning_rate": 3.311326828911494e-05, | |
| "loss": 0.4769, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.3636363636363638, | |
| "grad_norm": 0.13353175627047656, | |
| "learning_rate": 3.277136371533981e-05, | |
| "loss": 0.4759, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.3778409090909092, | |
| "grad_norm": 0.1417715943995476, | |
| "learning_rate": 3.2428166280089885e-05, | |
| "loss": 0.4744, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.3920454545454546, | |
| "grad_norm": 0.14007028648406247, | |
| "learning_rate": 3.20837601564115e-05, | |
| "loss": 0.4734, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.40625, | |
| "grad_norm": 0.12629456979175258, | |
| "learning_rate": 3.173822981379548e-05, | |
| "loss": 0.4724, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.4204545454545454, | |
| "grad_norm": 0.12349607254647936, | |
| "learning_rate": 3.1391659997460064e-05, | |
| "loss": 0.4713, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.4346590909090908, | |
| "grad_norm": 0.11464965836513225, | |
| "learning_rate": 3.104413570756612e-05, | |
| "loss": 0.4708, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.4488636363636362, | |
| "grad_norm": 0.1208312662969904, | |
| "learning_rate": 3.069574217836993e-05, | |
| "loss": 0.4803, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.4630681818181819, | |
| "grad_norm": 0.10698876943855158, | |
| "learning_rate": 3.0346564857318567e-05, | |
| "loss": 0.4633, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.4772727272727273, | |
| "grad_norm": 0.11384636090749484, | |
| "learning_rate": 2.9996689384092874e-05, | |
| "loss": 0.4699, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.4914772727272727, | |
| "grad_norm": 0.11145569232907346, | |
| "learning_rate": 2.964620156960342e-05, | |
| "loss": 0.4744, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.5056818181818183, | |
| "grad_norm": 0.13346919699354048, | |
| "learning_rate": 2.929518737494437e-05, | |
| "loss": 0.4747, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.5198863636363638, | |
| "grad_norm": 0.11742663104280518, | |
| "learning_rate": 2.8943732890310577e-05, | |
| "loss": 0.4713, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.5340909090909092, | |
| "grad_norm": 0.12233693509061423, | |
| "learning_rate": 2.859192431388298e-05, | |
| "loss": 0.4747, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.5482954545454546, | |
| "grad_norm": 0.11567941988020804, | |
| "learning_rate": 2.8239847930687522e-05, | |
| "loss": 0.4776, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.5625, | |
| "grad_norm": 0.12718334961619032, | |
| "learning_rate": 2.788759009143279e-05, | |
| "loss": 0.4686, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.5767045454545454, | |
| "grad_norm": 0.10446651216785705, | |
| "learning_rate": 2.753523719133152e-05, | |
| "loss": 0.4712, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.5909090909090908, | |
| "grad_norm": 0.11060336379300346, | |
| "learning_rate": 2.7182875648911177e-05, | |
| "loss": 0.4739, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.6051136363636362, | |
| "grad_norm": 0.11390898204723482, | |
| "learning_rate": 2.683059188481887e-05, | |
| "loss": 0.4638, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.6193181818181817, | |
| "grad_norm": 0.11190601188366513, | |
| "learning_rate": 2.647847230062567e-05, | |
| "loss": 0.4711, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.6335227272727273, | |
| "grad_norm": 0.11305263482331018, | |
| "learning_rate": 2.6126603257635733e-05, | |
| "loss": 0.4691, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.6477272727272727, | |
| "grad_norm": 0.1265534565322895, | |
| "learning_rate": 2.5775071055705075e-05, | |
| "loss": 0.4791, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.6619318181818183, | |
| "grad_norm": 0.11028857191690714, | |
| "learning_rate": 2.5423961912075712e-05, | |
| "loss": 0.4741, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.6761363636363638, | |
| "grad_norm": 0.11709279310297076, | |
| "learning_rate": 2.5073361940229773e-05, | |
| "loss": 0.4627, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.6903409090909092, | |
| "grad_norm": 0.12689675181012736, | |
| "learning_rate": 2.4723357128769286e-05, | |
| "loss": 0.4716, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.7045454545454546, | |
| "grad_norm": 0.12232567128887789, | |
| "learning_rate": 2.4374033320326506e-05, | |
| "loss": 0.4718, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.71875, | |
| "grad_norm": 0.11216636447930772, | |
| "learning_rate": 2.4025476190509938e-05, | |
| "loss": 0.4703, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.7329545454545454, | |
| "grad_norm": 0.11075120188097681, | |
| "learning_rate": 2.3677771226891564e-05, | |
| "loss": 0.4749, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.7471590909090908, | |
| "grad_norm": 0.1296834181423037, | |
| "learning_rate": 2.3331003708039932e-05, | |
| "loss": 0.4756, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.7613636363636362, | |
| "grad_norm": 0.10873440353655982, | |
| "learning_rate": 2.2985258682604706e-05, | |
| "loss": 0.4753, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.7755681818181817, | |
| "grad_norm": 0.11054731896051188, | |
| "learning_rate": 2.264062094845749e-05, | |
| "loss": 0.4681, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.7897727272727273, | |
| "grad_norm": 0.1077785017076799, | |
| "learning_rate": 2.229717503189418e-05, | |
| "loss": 0.4667, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.8039772727272727, | |
| "grad_norm": 0.12146741261267094, | |
| "learning_rate": 2.195500516690398e-05, | |
| "loss": 0.4667, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.8181818181818183, | |
| "grad_norm": 0.11013296161652279, | |
| "learning_rate": 2.1614195274510002e-05, | |
| "loss": 0.4692, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.8323863636363638, | |
| "grad_norm": 0.1122431509566975, | |
| "learning_rate": 2.1274828942186743e-05, | |
| "loss": 0.4761, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.8465909090909092, | |
| "grad_norm": 0.11152245824612014, | |
| "learning_rate": 2.09369894033593e-05, | |
| "loss": 0.4611, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.8607954545454546, | |
| "grad_norm": 0.10540289603858785, | |
| "learning_rate": 2.0600759516989397e-05, | |
| "loss": 0.4659, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.875, | |
| "grad_norm": 0.12268974044452592, | |
| "learning_rate": 2.026622174725336e-05, | |
| "loss": 0.4659, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.8892045454545454, | |
| "grad_norm": 0.11217674945544069, | |
| "learning_rate": 1.9933458143316748e-05, | |
| "loss": 0.4683, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.9034090909090908, | |
| "grad_norm": 0.11593305886838406, | |
| "learning_rate": 1.9602550319210967e-05, | |
| "loss": 0.4655, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.9176136363636362, | |
| "grad_norm": 0.11394452786176468, | |
| "learning_rate": 1.9273579433816452e-05, | |
| "loss": 0.4634, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.9318181818181817, | |
| "grad_norm": 0.12344588418657035, | |
| "learning_rate": 1.894662617095753e-05, | |
| "loss": 0.4605, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.9460227272727273, | |
| "grad_norm": 0.10588427958742584, | |
| "learning_rate": 1.8621770719613862e-05, | |
| "loss": 0.4636, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.9602272727272727, | |
| "grad_norm": 0.10548340772355062, | |
| "learning_rate": 1.829909275425309e-05, | |
| "loss": 0.4694, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.9744318181818183, | |
| "grad_norm": 0.10424071348650422, | |
| "learning_rate": 1.7978671415289878e-05, | |
| "loss": 0.4705, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.9886363636363638, | |
| "grad_norm": 0.1105427740992237, | |
| "learning_rate": 1.7660585289675758e-05, | |
| "loss": 0.4631, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 2.002840909090909, | |
| "grad_norm": 0.12132401905710084, | |
| "learning_rate": 1.7344912391624826e-05, | |
| "loss": 0.4573, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 2.0170454545454546, | |
| "grad_norm": 0.11150850216312608, | |
| "learning_rate": 1.7031730143479873e-05, | |
| "loss": 0.4495, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 2.03125, | |
| "grad_norm": 0.1062932021752816, | |
| "learning_rate": 1.672111535672367e-05, | |
| "loss": 0.4361, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 2.0454545454545454, | |
| "grad_norm": 0.10247863399966832, | |
| "learning_rate": 1.64131442131401e-05, | |
| "loss": 0.4437, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 2.059659090909091, | |
| "grad_norm": 0.10677014234090489, | |
| "learning_rate": 1.6107892246129692e-05, | |
| "loss": 0.4434, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 2.0738636363636362, | |
| "grad_norm": 0.10650883791134519, | |
| "learning_rate": 1.5805434322184224e-05, | |
| "loss": 0.4445, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 2.0880681818181817, | |
| "grad_norm": 0.10387373586373833, | |
| "learning_rate": 1.550584462252487e-05, | |
| "loss": 0.4509, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 2.102272727272727, | |
| "grad_norm": 0.1030086032533747, | |
| "learning_rate": 1.520919662490839e-05, | |
| "loss": 0.4438, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.116477272727273, | |
| "grad_norm": 0.10277657855159573, | |
| "learning_rate": 1.4915563085605922e-05, | |
| "loss": 0.4391, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 2.1306818181818183, | |
| "grad_norm": 0.09814581198724093, | |
| "learning_rate": 1.4625016021558654e-05, | |
| "loss": 0.4458, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 2.1448863636363638, | |
| "grad_norm": 0.11056576999353432, | |
| "learning_rate": 1.4337626692714862e-05, | |
| "loss": 0.4496, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 2.159090909090909, | |
| "grad_norm": 0.09890533126932854, | |
| "learning_rate": 1.4053465584552613e-05, | |
| "loss": 0.4395, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 2.1732954545454546, | |
| "grad_norm": 0.10516851231359878, | |
| "learning_rate": 1.3772602390792397e-05, | |
| "loss": 0.442, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 2.1875, | |
| "grad_norm": 0.10009393657351306, | |
| "learning_rate": 1.3495105996303959e-05, | |
| "loss": 0.4384, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 2.2017045454545454, | |
| "grad_norm": 0.0998627875200131, | |
| "learning_rate": 1.3221044460211496e-05, | |
| "loss": 0.4436, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 2.215909090909091, | |
| "grad_norm": 0.09883322341572116, | |
| "learning_rate": 1.2950484999201407e-05, | |
| "loss": 0.4508, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.2301136363636362, | |
| "grad_norm": 0.09827162855077891, | |
| "learning_rate": 1.2683493971036633e-05, | |
| "loss": 0.4394, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 2.2443181818181817, | |
| "grad_norm": 0.1028260573381402, | |
| "learning_rate": 1.2420136858281663e-05, | |
| "loss": 0.4442, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 2.2585227272727275, | |
| "grad_norm": 0.10185969414875563, | |
| "learning_rate": 1.2160478252242191e-05, | |
| "loss": 0.4522, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 2.2727272727272725, | |
| "grad_norm": 0.09897503980351197, | |
| "learning_rate": 1.1904581837123386e-05, | |
| "loss": 0.4395, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.2869318181818183, | |
| "grad_norm": 0.10179604173619207, | |
| "learning_rate": 1.1652510374410607e-05, | |
| "loss": 0.4476, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 2.3011363636363638, | |
| "grad_norm": 0.10087044008989485, | |
| "learning_rate": 1.1404325687476435e-05, | |
| "loss": 0.455, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.315340909090909, | |
| "grad_norm": 0.09796524933029822, | |
| "learning_rate": 1.1160088646417834e-05, | |
| "loss": 0.441, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 2.3295454545454546, | |
| "grad_norm": 0.10278194962659604, | |
| "learning_rate": 1.0919859153127004e-05, | |
| "loss": 0.4415, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.34375, | |
| "grad_norm": 0.10008666177717428, | |
| "learning_rate": 1.0683696126599816e-05, | |
| "loss": 0.4504, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 2.3579545454545454, | |
| "grad_norm": 0.10127145328521546, | |
| "learning_rate": 1.045165748848525e-05, | |
| "loss": 0.4432, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.372159090909091, | |
| "grad_norm": 0.09465727719419995, | |
| "learning_rate": 1.0223800148879456e-05, | |
| "loss": 0.4406, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 2.3863636363636362, | |
| "grad_norm": 0.10223238212578617, | |
| "learning_rate": 1.0000179992367925e-05, | |
| "loss": 0.4417, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.4005681818181817, | |
| "grad_norm": 0.10542315681224666, | |
| "learning_rate": 9.780851864319119e-06, | |
| "loss": 0.4447, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 2.4147727272727275, | |
| "grad_norm": 0.0963765340314958, | |
| "learning_rate": 9.565869557433065e-06, | |
| "loss": 0.4412, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.4289772727272725, | |
| "grad_norm": 0.10046217850351946, | |
| "learning_rate": 9.35528579854809e-06, | |
| "loss": 0.4484, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 2.4431818181818183, | |
| "grad_norm": 0.10147180197503058, | |
| "learning_rate": 9.149152235708938e-06, | |
| "loss": 0.4397, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.4573863636363638, | |
| "grad_norm": 0.11146703090708097, | |
| "learning_rate": 8.947519425499546e-06, | |
| "loss": 0.4457, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 2.471590909090909, | |
| "grad_norm": 0.0945703122512324, | |
| "learning_rate": 8.750436820643412e-06, | |
| "loss": 0.4392, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.4857954545454546, | |
| "grad_norm": 0.09908284212508756, | |
| "learning_rate": 8.557952757874824e-06, | |
| "loss": 0.442, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 0.09395206380523909, | |
| "learning_rate": 8.370114446083686e-06, | |
| "loss": 0.4354, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.5142045454545454, | |
| "grad_norm": 0.1005576468311754, | |
| "learning_rate": 8.186967954736992e-06, | |
| "loss": 0.4533, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 2.528409090909091, | |
| "grad_norm": 0.1020807372597039, | |
| "learning_rate": 8.008558202579839e-06, | |
| "loss": 0.443, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.5426136363636362, | |
| "grad_norm": 0.09991703967673085, | |
| "learning_rate": 7.834928946618499e-06, | |
| "loss": 0.4475, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 2.5568181818181817, | |
| "grad_norm": 0.10275386995373673, | |
| "learning_rate": 7.666122771388617e-06, | |
| "loss": 0.4484, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.5710227272727275, | |
| "grad_norm": 0.10616220162227955, | |
| "learning_rate": 7.502181078510801e-06, | |
| "loss": 0.441, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 2.5852272727272725, | |
| "grad_norm": 0.09678868754399234, | |
| "learning_rate": 7.343144076536432e-06, | |
| "loss": 0.4345, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.5994318181818183, | |
| "grad_norm": 0.09985443664593775, | |
| "learning_rate": 7.189050771086084e-06, | |
| "loss": 0.4432, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 2.6136363636363638, | |
| "grad_norm": 0.09670777253001571, | |
| "learning_rate": 7.039938955282907e-06, | |
| "loss": 0.4417, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.627840909090909, | |
| "grad_norm": 0.09783216540452833, | |
| "learning_rate": 6.895845200483489e-06, | |
| "loss": 0.4424, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 2.6420454545454546, | |
| "grad_norm": 0.09147945018928255, | |
| "learning_rate": 6.756804847308286e-06, | |
| "loss": 0.4434, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.65625, | |
| "grad_norm": 0.09802106237190225, | |
| "learning_rate": 6.622851996973932e-06, | |
| "loss": 0.4407, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 2.6704545454545454, | |
| "grad_norm": 0.09462692879617098, | |
| "learning_rate": 6.49401950292956e-06, | |
| "loss": 0.4433, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.684659090909091, | |
| "grad_norm": 0.09624748034407322, | |
| "learning_rate": 6.3703389627990844e-06, | |
| "loss": 0.4387, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 2.6988636363636362, | |
| "grad_norm": 0.09543251734275723, | |
| "learning_rate": 6.2518407106315695e-06, | |
| "loss": 0.4363, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.7130681818181817, | |
| "grad_norm": 0.09883095226273896, | |
| "learning_rate": 6.138553809461422e-06, | |
| "loss": 0.4386, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 2.7272727272727275, | |
| "grad_norm": 0.09756625437555602, | |
| "learning_rate": 6.030506044180375e-06, | |
| "loss": 0.4459, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.7414772727272725, | |
| "grad_norm": 0.09607808880251172, | |
| "learning_rate": 5.92772391472293e-06, | |
| "loss": 0.4359, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 2.7556818181818183, | |
| "grad_norm": 0.09216704090541984, | |
| "learning_rate": 5.830232629566935e-06, | |
| "loss": 0.45, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.7698863636363638, | |
| "grad_norm": 0.09291163689146356, | |
| "learning_rate": 5.738056099550936e-06, | |
| "loss": 0.4363, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 2.784090909090909, | |
| "grad_norm": 0.09345711945306576, | |
| "learning_rate": 5.6512169320097665e-06, | |
| "loss": 0.4497, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.7982954545454546, | |
| "grad_norm": 0.09335066659653964, | |
| "learning_rate": 5.569736425229844e-06, | |
| "loss": 0.4411, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 2.8125, | |
| "grad_norm": 0.09732221176443216, | |
| "learning_rate": 5.493634563225524e-06, | |
| "loss": 0.4453, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.8267045454545454, | |
| "grad_norm": 0.0954814046395625, | |
| "learning_rate": 5.422930010837793e-06, | |
| "loss": 0.4433, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 2.840909090909091, | |
| "grad_norm": 0.09445562287163194, | |
| "learning_rate": 5.357640109156516e-06, | |
| "loss": 0.4396, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.8551136363636362, | |
| "grad_norm": 0.09370114782801439, | |
| "learning_rate": 5.297780871267323e-06, | |
| "loss": 0.441, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 2.8693181818181817, | |
| "grad_norm": 0.09116071509216306, | |
| "learning_rate": 5.243366978324231e-06, | |
| "loss": 0.4422, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.8835227272727275, | |
| "grad_norm": 0.09885237977800734, | |
| "learning_rate": 5.1944117759489156e-06, | |
| "loss": 0.4404, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 2.8977272727272725, | |
| "grad_norm": 0.09995837383801799, | |
| "learning_rate": 5.1509272709575616e-06, | |
| "loss": 0.4456, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.9119318181818183, | |
| "grad_norm": 0.09596178401038803, | |
| "learning_rate": 5.112924128416038e-06, | |
| "loss": 0.4404, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 2.9261363636363638, | |
| "grad_norm": 0.09406970508523729, | |
| "learning_rate": 5.080411669024179e-06, | |
| "loss": 0.4391, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.940340909090909, | |
| "grad_norm": 0.0951093143907049, | |
| "learning_rate": 5.0533978668297775e-06, | |
| "loss": 0.4483, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 2.9545454545454546, | |
| "grad_norm": 0.09615998973664797, | |
| "learning_rate": 5.031889347272857e-06, | |
| "loss": 0.4408, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.96875, | |
| "grad_norm": 0.09428736632782057, | |
| "learning_rate": 5.015891385560713e-06, | |
| "loss": 0.4432, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 2.9829545454545454, | |
| "grad_norm": 0.09338757409389373, | |
| "learning_rate": 5.005407905374099e-06, | |
| "loss": 0.4372, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.997159090909091, | |
| "grad_norm": 0.09934638872755362, | |
| "learning_rate": 5.0004414779049116e-06, | |
| "loss": 0.4421, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 1056, | |
| "total_flos": 962656789856256.0, | |
| "train_loss": 0.5030564876329718, | |
| "train_runtime": 14725.7023, | |
| "train_samples_per_second": 2.293, | |
| "train_steps_per_second": 0.072 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 1056, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 962656789856256.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |