| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.9994378864530633, | |
| "eval_steps": 500, | |
| "global_step": 2001, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.014989694584972832, | |
| "grad_norm": 3.2230756947704684, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8065, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.029979389169945664, | |
| "grad_norm": 4.087000478453515, | |
| "learning_rate": 5e-06, | |
| "loss": 0.651, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.044969083754918496, | |
| "grad_norm": 2.503327595683403, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6152, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.05995877833989133, | |
| "grad_norm": 2.6733564049479916, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5859, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.07494847292486416, | |
| "grad_norm": 2.7328394257454245, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5786, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.08993816750983699, | |
| "grad_norm": 3.1797521945570693, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5685, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.10492786209480982, | |
| "grad_norm": 2.684949545099637, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5586, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.11991755667978266, | |
| "grad_norm": 2.9219404511166824, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5529, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.13490725126475547, | |
| "grad_norm": 2.3100236702034826, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5461, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.14989694584972832, | |
| "grad_norm": 2.2049571056422397, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5381, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.16488664043470114, | |
| "grad_norm": 1.7661938201026661, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5359, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.17987633501967398, | |
| "grad_norm": 2.1859179504620174, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5318, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.1948660296046468, | |
| "grad_norm": 1.6326069613621976, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5279, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.20985572418961965, | |
| "grad_norm": 1.5360852906450997, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5309, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.22484541877459246, | |
| "grad_norm": 1.7724258855816508, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5207, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.2398351133595653, | |
| "grad_norm": 1.7886911455233951, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5203, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.25482480794453816, | |
| "grad_norm": 1.436404813254963, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5179, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.26981450252951095, | |
| "grad_norm": 1.892553234562171, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5134, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.2848041971144838, | |
| "grad_norm": 1.6450239492265477, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5124, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.29979389169945664, | |
| "grad_norm": 1.894402097483371, | |
| "learning_rate": 5e-06, | |
| "loss": 0.508, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.31478358628442943, | |
| "grad_norm": 1.6348153643564973, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5032, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.3297732808694023, | |
| "grad_norm": 1.3715172854305366, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5051, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.3447629754543751, | |
| "grad_norm": 1.742102338514997, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5036, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.35975267003934797, | |
| "grad_norm": 1.7426571084822509, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5041, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.37474236462432076, | |
| "grad_norm": 2.1505290746647465, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5001, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.3897320592092936, | |
| "grad_norm": 1.793888447963483, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5015, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.40472175379426645, | |
| "grad_norm": 1.8644855732640748, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4958, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.4197114483792393, | |
| "grad_norm": 2.6049833173978825, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4996, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.4347011429642121, | |
| "grad_norm": 1.4405086279628676, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4929, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.44969083754918493, | |
| "grad_norm": 1.604860450889284, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4897, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.4646805321341578, | |
| "grad_norm": 1.662135718621203, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4927, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.4796702267191306, | |
| "grad_norm": 2.062302477917024, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4892, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.4946599213041034, | |
| "grad_norm": 1.4281806778360036, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4889, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.5096496158890763, | |
| "grad_norm": 1.4534145149693543, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4872, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.5246393104740491, | |
| "grad_norm": 1.8740427050796966, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4839, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.5396290050590219, | |
| "grad_norm": 1.5668501527936287, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4886, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.5546186996439948, | |
| "grad_norm": 1.3759231439903845, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4837, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.5696083942289676, | |
| "grad_norm": 1.498739905029643, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4822, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.5845980888139404, | |
| "grad_norm": 1.5404538522347584, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4798, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.5995877833989133, | |
| "grad_norm": 2.0684145258599544, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4831, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.6145774779838861, | |
| "grad_norm": 1.7365319201533058, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4803, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.6295671725688589, | |
| "grad_norm": 1.7707026206608905, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4759, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.6445568671538318, | |
| "grad_norm": 1.4832433269572964, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4759, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.6595465617388045, | |
| "grad_norm": 1.4849966253145612, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4731, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.6745362563237775, | |
| "grad_norm": 2.1406203761167037, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4748, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.6895259509087502, | |
| "grad_norm": 1.6432408076200942, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4692, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.704515645493723, | |
| "grad_norm": 2.110147852343459, | |
| "learning_rate": 5e-06, | |
| "loss": 0.469, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.7195053400786959, | |
| "grad_norm": 2.67027429796622, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4742, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.7344950346636687, | |
| "grad_norm": 2.1612548231072197, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4734, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.7494847292486415, | |
| "grad_norm": 1.9020717675158103, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4693, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.7644744238336144, | |
| "grad_norm": 1.8918141412998646, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4734, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.7794641184185872, | |
| "grad_norm": 1.368955726552673, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4663, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.7944538130035601, | |
| "grad_norm": 1.5459750086281097, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4645, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.8094435075885329, | |
| "grad_norm": 1.5230854042747586, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4647, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.8244332021735057, | |
| "grad_norm": 1.5090270259811516, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4707, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.8394228967584786, | |
| "grad_norm": 1.5558244270343213, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4674, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.8544125913434514, | |
| "grad_norm": 1.4739129842995866, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4678, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.8694022859284242, | |
| "grad_norm": 1.7227702869131336, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4614, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.8843919805133971, | |
| "grad_norm": 1.412974957600055, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4634, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.8993816750983699, | |
| "grad_norm": 1.3755393868079429, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4649, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.9143713696833426, | |
| "grad_norm": 1.3451129180593917, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4638, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.9293610642683156, | |
| "grad_norm": 1.4949083165414878, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4639, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.9443507588532883, | |
| "grad_norm": 1.3507178958809956, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4656, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.9593404534382612, | |
| "grad_norm": 1.572459968859256, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4592, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.974330148023234, | |
| "grad_norm": 1.2767150139708328, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4632, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.9893198426082068, | |
| "grad_norm": 1.5510115795503987, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4596, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.9998126288176878, | |
| "eval_loss": 0.05758751183748245, | |
| "eval_runtime": 454.8318, | |
| "eval_samples_per_second": 39.525, | |
| "eval_steps_per_second": 0.618, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 1.0043095371931796, | |
| "grad_norm": 2.7775612629774344, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4348, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.0192992317781526, | |
| "grad_norm": 1.9710376561762528, | |
| "learning_rate": 5e-06, | |
| "loss": 0.364, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.0342889263631254, | |
| "grad_norm": 1.5909045555502788, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3558, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.0492786209480982, | |
| "grad_norm": 1.3602748984083428, | |
| "learning_rate": 5e-06, | |
| "loss": 0.355, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.064268315533071, | |
| "grad_norm": 1.6434233988948692, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3519, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.0792580101180438, | |
| "grad_norm": 1.4242278638735506, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3545, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.0942477047030166, | |
| "grad_norm": 1.547462512141288, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3583, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.1092373992879896, | |
| "grad_norm": 1.4513237085400739, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3558, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.1242270938729624, | |
| "grad_norm": 2.1885817879178404, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3563, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.1392167884579352, | |
| "grad_norm": 1.695688809842665, | |
| "learning_rate": 5e-06, | |
| "loss": 0.358, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.154206483042908, | |
| "grad_norm": 2.289531376884916, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3574, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.1691961776278808, | |
| "grad_norm": 1.9357596883149255, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3584, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.1841858722128538, | |
| "grad_norm": 2.1618585974064395, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3564, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.1991755667978266, | |
| "grad_norm": 1.6373234010519413, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3527, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.2141652613827993, | |
| "grad_norm": 1.8399099410223894, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3566, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.2291549559677721, | |
| "grad_norm": 1.6753343512169294, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3601, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.244144650552745, | |
| "grad_norm": 1.505241949293805, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3583, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.259134345137718, | |
| "grad_norm": 1.5312665902757236, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3542, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.2741240397226907, | |
| "grad_norm": 1.8665087434383938, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3575, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.2891137343076635, | |
| "grad_norm": 1.7982799864263068, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3586, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.3041034288926363, | |
| "grad_norm": 1.5371290759554668, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3622, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.319093123477609, | |
| "grad_norm": 1.5167552093630703, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3588, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.334082818062582, | |
| "grad_norm": 1.6591098979757748, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3634, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.3490725126475547, | |
| "grad_norm": 1.5726838898036415, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3539, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.3640622072325277, | |
| "grad_norm": 1.7507238369968354, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3632, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.3790519018175005, | |
| "grad_norm": 1.475358404698048, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3625, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.3940415964024733, | |
| "grad_norm": 1.5954088726182911, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3595, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.409031290987446, | |
| "grad_norm": 1.520181904849792, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3597, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.4240209855724189, | |
| "grad_norm": 1.6317070098760134, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3611, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.4390106801573919, | |
| "grad_norm": 1.768300641616324, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3603, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.4540003747423647, | |
| "grad_norm": 1.5408964888063073, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3631, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.4689900693273374, | |
| "grad_norm": 1.452167086951644, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3609, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.4839797639123102, | |
| "grad_norm": 1.6249338530074764, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3625, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.498969458497283, | |
| "grad_norm": 1.7743633699281656, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3618, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.513959153082256, | |
| "grad_norm": 1.7741260951568605, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3612, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.5289488476672288, | |
| "grad_norm": 1.3987893388986192, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3628, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.5439385422522016, | |
| "grad_norm": 1.4893576161794635, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3626, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.5589282368371744, | |
| "grad_norm": 1.4112328226196433, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3618, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.5739179314221472, | |
| "grad_norm": 1.3858150270139942, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3615, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.5889076260071202, | |
| "grad_norm": 1.5535049280480435, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3618, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.6038973205920928, | |
| "grad_norm": 1.3757444255540903, | |
| "learning_rate": 5e-06, | |
| "loss": 0.365, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.6188870151770658, | |
| "grad_norm": 1.5995972790776192, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3632, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.6338767097620386, | |
| "grad_norm": 1.7108799556122343, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3643, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.6488664043470114, | |
| "grad_norm": 1.686754492103127, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3649, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.6638560989319844, | |
| "grad_norm": 1.7693708666917307, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3612, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.678845793516957, | |
| "grad_norm": 1.520817212897819, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3614, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.69383548810193, | |
| "grad_norm": 1.9406706141560084, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3601, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.7088251826869028, | |
| "grad_norm": 1.488137404772404, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3627, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.7238148772718755, | |
| "grad_norm": 1.8331126631936938, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3609, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.7388045718568486, | |
| "grad_norm": 1.5561141952783655, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3601, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.7537942664418211, | |
| "grad_norm": 1.4134693187494294, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3596, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.7687839610267941, | |
| "grad_norm": 1.696827607327685, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3614, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.783773655611767, | |
| "grad_norm": 1.5042455149771978, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3629, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.7987633501967397, | |
| "grad_norm": 1.619841917897079, | |
| "learning_rate": 5e-06, | |
| "loss": 0.364, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.8137530447817127, | |
| "grad_norm": 1.9678117854674795, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3623, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.8287427393666853, | |
| "grad_norm": 2.000250566254895, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3629, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.8437324339516583, | |
| "grad_norm": 1.7246823549895025, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3639, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.858722128536631, | |
| "grad_norm": 1.9341838486216245, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3614, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.873711823121604, | |
| "grad_norm": 1.6254030300504665, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3628, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.8887015177065767, | |
| "grad_norm": 1.5870549405395955, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3624, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.9036912122915495, | |
| "grad_norm": 1.5435102287374787, | |
| "learning_rate": 5e-06, | |
| "loss": 0.362, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.9186809068765225, | |
| "grad_norm": 1.4063275660945347, | |
| "learning_rate": 5e-06, | |
| "loss": 0.363, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.9336706014614953, | |
| "grad_norm": 1.5087142670105975, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3587, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.948660296046468, | |
| "grad_norm": 1.3583330142096965, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3608, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.9636499906314409, | |
| "grad_norm": 1.5780539376028968, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3621, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.9786396852164136, | |
| "grad_norm": 1.451758117798617, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3632, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.9936293798013867, | |
| "grad_norm": 1.428577561948826, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3617, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.9996252576353757, | |
| "eval_loss": 0.056735917925834656, | |
| "eval_runtime": 450.721, | |
| "eval_samples_per_second": 39.885, | |
| "eval_steps_per_second": 0.623, | |
| "step": 1334 | |
| }, | |
| { | |
| "epoch": 2.0086190743863592, | |
| "grad_norm": 2.693760600411948, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3009, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.0236087689713322, | |
| "grad_norm": 1.8123948443076334, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2445, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.0385984635563053, | |
| "grad_norm": 1.8233844327944795, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2404, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.053588158141278, | |
| "grad_norm": 1.659782737096887, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2396, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.068577852726251, | |
| "grad_norm": 1.6369760989914484, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2399, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.0835675473112234, | |
| "grad_norm": 1.6159958358268898, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2398, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.0985572418961964, | |
| "grad_norm": 1.8629592527737424, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2404, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.1135469364811694, | |
| "grad_norm": 1.9402493917715093, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2406, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.128536631066142, | |
| "grad_norm": 1.9914438604583353, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2445, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.143526325651115, | |
| "grad_norm": 1.6486122695508032, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2428, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.1585160202360876, | |
| "grad_norm": 1.5254691785643584, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2443, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.1735057148210606, | |
| "grad_norm": 1.9133206318932665, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2431, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 2.188495409406033, | |
| "grad_norm": 1.6550777901047227, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2448, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.203485103991006, | |
| "grad_norm": 1.8001235821461439, | |
| "learning_rate": 5e-06, | |
| "loss": 0.247, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 2.218474798575979, | |
| "grad_norm": 1.6442525813509412, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2491, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.2334644931609517, | |
| "grad_norm": 1.6551298330853466, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2485, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 2.2484541877459248, | |
| "grad_norm": 1.691060089285524, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2463, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.2634438823308973, | |
| "grad_norm": 1.680873308991847, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2486, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 2.2784335769158703, | |
| "grad_norm": 1.5264889834969293, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2497, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.2934232715008434, | |
| "grad_norm": 1.9490999202092631, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2492, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 2.308412966085816, | |
| "grad_norm": 1.5890774001111174, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2503, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.323402660670789, | |
| "grad_norm": 1.8381634501461115, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2505, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 2.3383923552557615, | |
| "grad_norm": 1.8112499776123423, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2518, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.3533820498407345, | |
| "grad_norm": 1.7701674873383388, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2504, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 2.3683717444257075, | |
| "grad_norm": 1.7787057235293418, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2511, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.38336143901068, | |
| "grad_norm": 1.6220489456862308, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2517, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 2.398351133595653, | |
| "grad_norm": 1.6092461385206882, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2519, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.4133408281806257, | |
| "grad_norm": 1.6892803378685273, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2537, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.4283305227655987, | |
| "grad_norm": 1.5514590659702057, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2525, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.4433202173505713, | |
| "grad_norm": 1.6463095287912044, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2519, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 2.4583099119355443, | |
| "grad_norm": 1.8642417904545223, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2542, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.4732996065205173, | |
| "grad_norm": 2.1833462931941483, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2523, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 2.48828930110549, | |
| "grad_norm": 1.7334334035624595, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2525, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.503278995690463, | |
| "grad_norm": 1.996879481861675, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2512, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 2.518268690275436, | |
| "grad_norm": 1.7646159599515048, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2543, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.5332583848604084, | |
| "grad_norm": 1.5280180383047155, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2542, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 2.5482480794453815, | |
| "grad_norm": 1.6854655209849605, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2543, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.563237774030354, | |
| "grad_norm": 1.614184312338993, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2547, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 2.578227468615327, | |
| "grad_norm": 1.7759088208026037, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2577, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.5932171632002996, | |
| "grad_norm": 1.7074767316934885, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2585, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 2.6082068577852726, | |
| "grad_norm": 1.666840211483153, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2566, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 2.6231965523702456, | |
| "grad_norm": 1.5612601726380533, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2577, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 2.638186246955218, | |
| "grad_norm": 1.6048678229351492, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2584, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 2.653175941540191, | |
| "grad_norm": 1.535012707234654, | |
| "learning_rate": 5e-06, | |
| "loss": 0.258, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 2.668165636125164, | |
| "grad_norm": 1.5594137106457704, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2596, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 2.683155330710137, | |
| "grad_norm": 1.7412507218140363, | |
| "learning_rate": 5e-06, | |
| "loss": 0.259, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 2.6981450252951094, | |
| "grad_norm": 1.8620927203026323, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2613, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.7131347198800824, | |
| "grad_norm": 1.5875808505192006, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2585, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 2.7281244144650554, | |
| "grad_norm": 1.6955850941699981, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2597, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 2.743114109050028, | |
| "grad_norm": 1.7569205126042602, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2571, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 2.758103803635001, | |
| "grad_norm": 1.519918624729941, | |
| "learning_rate": 5e-06, | |
| "loss": 0.263, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 2.773093498219974, | |
| "grad_norm": 1.5838501646665883, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2608, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 2.7880831928049465, | |
| "grad_norm": 1.6657424039076156, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2613, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 2.8030728873899196, | |
| "grad_norm": 1.595659803150772, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2611, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 2.818062581974892, | |
| "grad_norm": 1.7789456591238404, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2622, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 2.833052276559865, | |
| "grad_norm": 1.6114699686236196, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2615, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 2.8480419711448377, | |
| "grad_norm": 1.6465619591478904, | |
| "learning_rate": 5e-06, | |
| "loss": 0.263, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 2.8630316657298107, | |
| "grad_norm": 1.6736896325417563, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2649, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 2.8780213603147837, | |
| "grad_norm": 1.52817706604071, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2611, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 2.8930110548997563, | |
| "grad_norm": 1.7454962824468059, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2626, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 2.9080007494847293, | |
| "grad_norm": 1.6946473714595274, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2627, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 2.9229904440697023, | |
| "grad_norm": 1.6411356019714227, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2628, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 2.937980138654675, | |
| "grad_norm": 1.9705938078449712, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2627, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 2.952969833239648, | |
| "grad_norm": 1.9589063291434328, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2646, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 2.9679595278246205, | |
| "grad_norm": 1.6622226531628324, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2637, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 2.9829492224095935, | |
| "grad_norm": 1.7651652877550925, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2637, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 2.997938916994566, | |
| "grad_norm": 1.6491028582670904, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2629, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.9994378864530633, | |
| "eval_loss": 0.061618607491254807, | |
| "eval_runtime": 450.5684, | |
| "eval_samples_per_second": 39.898, | |
| "eval_steps_per_second": 0.624, | |
| "step": 2001 | |
| }, | |
| { | |
| "epoch": 2.9994378864530633, | |
| "step": 2001, | |
| "total_flos": 3351540148469760.0, | |
| "train_loss": 0.37255630833932246, | |
| "train_runtime": 64694.9321, | |
| "train_samples_per_second": 15.838, | |
| "train_steps_per_second": 0.031 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2001, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3351540148469760.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |