| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.998551424432641, | |
| "eval_steps": 500, | |
| "global_step": 3105, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.009657170449058426, | |
| "grad_norm": 3.7815811768441914, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6253, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01931434089811685, | |
| "grad_norm": 2.097627838391192, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5439, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.028971511347175277, | |
| "grad_norm": 1.991134226246532, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5217, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0386286817962337, | |
| "grad_norm": 1.786626694141357, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5096, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04828585224529213, | |
| "grad_norm": 1.717098828666389, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5039, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05794302269435055, | |
| "grad_norm": 1.8094322954355757, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5015, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.06760019314340898, | |
| "grad_norm": 2.0392174555732336, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4883, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.0772573635924674, | |
| "grad_norm": 1.9677308677580745, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4874, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.08691453404152583, | |
| "grad_norm": 1.7766414325618487, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4874, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.09657170449058426, | |
| "grad_norm": 1.6816893104701698, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4803, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.10622887493964268, | |
| "grad_norm": 1.4814895014838179, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4781, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.1158860453887011, | |
| "grad_norm": 1.809417684888777, | |
| "learning_rate": 5e-06, | |
| "loss": 0.477, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.12554321583775954, | |
| "grad_norm": 1.604499406008116, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4745, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.13520038628681796, | |
| "grad_norm": 1.5342478528722183, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4733, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.14485755673587639, | |
| "grad_norm": 1.7155349659817856, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4733, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.1545147271849348, | |
| "grad_norm": 1.4873226511627278, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4717, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.16417189763399323, | |
| "grad_norm": 1.5335342617066339, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4703, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.17382906808305165, | |
| "grad_norm": 1.692809745357256, | |
| "learning_rate": 5e-06, | |
| "loss": 0.473, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.1834862385321101, | |
| "grad_norm": 1.6215225290403987, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4713, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.19314340898116852, | |
| "grad_norm": 1.61595338484294, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4666, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.20280057943022695, | |
| "grad_norm": 1.535971453026849, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4644, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.21245774987928537, | |
| "grad_norm": 1.448430643756855, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4583, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.2221149203283438, | |
| "grad_norm": 1.6067011202263324, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4575, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.2317720907774022, | |
| "grad_norm": 1.4643615572820308, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4554, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.24142926122646063, | |
| "grad_norm": 1.5756879099170888, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4661, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.2510864316755191, | |
| "grad_norm": 1.6966284718047921, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4586, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.2607436021245775, | |
| "grad_norm": 1.6310274076314688, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4616, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.27040077257363593, | |
| "grad_norm": 1.5908628903198843, | |
| "learning_rate": 5e-06, | |
| "loss": 0.456, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2800579430226944, | |
| "grad_norm": 1.5321607540064421, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4632, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.28971511347175277, | |
| "grad_norm": 1.5432080707621136, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4498, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.2993722839208112, | |
| "grad_norm": 1.5174884163746274, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4547, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.3090294543698696, | |
| "grad_norm": 1.4431913116342714, | |
| "learning_rate": 5e-06, | |
| "loss": 0.448, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.31868662481892807, | |
| "grad_norm": 1.7071593067951876, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4493, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.32834379526798646, | |
| "grad_norm": 1.3857947339153756, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4545, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.3380009657170449, | |
| "grad_norm": 1.4688790780164955, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4465, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3476581361661033, | |
| "grad_norm": 1.4620765033414924, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4495, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.35731530661516175, | |
| "grad_norm": 1.4806186634505536, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4455, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.3669724770642202, | |
| "grad_norm": 1.3726165431422352, | |
| "learning_rate": 5e-06, | |
| "loss": 0.446, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.3766296475132786, | |
| "grad_norm": 1.902972844618989, | |
| "learning_rate": 5e-06, | |
| "loss": 0.447, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.38628681796233705, | |
| "grad_norm": 1.4905365903207408, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4446, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.39594398841139544, | |
| "grad_norm": 1.4193622471558158, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4481, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.4056011588604539, | |
| "grad_norm": 1.5124518593909522, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4417, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.4152583293095123, | |
| "grad_norm": 1.3927026205868633, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4469, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.42491549975857074, | |
| "grad_norm": 1.4506580301027323, | |
| "learning_rate": 5e-06, | |
| "loss": 0.442, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.4345726702076292, | |
| "grad_norm": 1.3495438486710758, | |
| "learning_rate": 5e-06, | |
| "loss": 0.446, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.4442298406566876, | |
| "grad_norm": 1.5206970544390066, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4419, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.45388701110574603, | |
| "grad_norm": 1.4858134923328608, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4543, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.4635441815548044, | |
| "grad_norm": 1.5043934135687507, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4459, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.4732013520038629, | |
| "grad_norm": 2.2482606991957463, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4426, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.48285852245292127, | |
| "grad_norm": 1.4045208076303264, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4326, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.4925156929019797, | |
| "grad_norm": 1.543363161753652, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4439, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.5021728633510382, | |
| "grad_norm": 1.3653299013378564, | |
| "learning_rate": 5e-06, | |
| "loss": 0.436, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.5118300338000966, | |
| "grad_norm": 1.4150482204508787, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4378, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.521487204249155, | |
| "grad_norm": 1.3973876180346887, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4395, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5311443746982134, | |
| "grad_norm": 1.4601941048467661, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4409, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.5408015451472719, | |
| "grad_norm": 1.367282105296493, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4365, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5504587155963303, | |
| "grad_norm": 1.4142842428381743, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4368, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5601158860453888, | |
| "grad_norm": 1.5238374949968105, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4351, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5697730564944471, | |
| "grad_norm": 1.4426020649884985, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4378, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5794302269435055, | |
| "grad_norm": 1.6034191206258377, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4359, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.589087397392564, | |
| "grad_norm": 1.3861019948416724, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4366, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5987445678416224, | |
| "grad_norm": 1.5575620851838663, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4293, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.6084017382906808, | |
| "grad_norm": 1.4940232627496768, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4308, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.6180589087397392, | |
| "grad_norm": 1.3940687230128848, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4344, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6277160791887977, | |
| "grad_norm": 1.4063585473842182, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4312, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.6373732496378561, | |
| "grad_norm": 1.4263057370102823, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4288, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6470304200869146, | |
| "grad_norm": 1.359338730182779, | |
| "learning_rate": 5e-06, | |
| "loss": 0.432, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6566875905359729, | |
| "grad_norm": 1.4201113340211817, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4271, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.6663447609850314, | |
| "grad_norm": 1.3945845556957837, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4297, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.6760019314340898, | |
| "grad_norm": 1.4879889305917073, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4275, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.6856591018831483, | |
| "grad_norm": 1.5684295486014674, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4318, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.6953162723322066, | |
| "grad_norm": 1.6080556978629539, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4327, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.7049734427812651, | |
| "grad_norm": 1.4517662054077567, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4316, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.7146306132303235, | |
| "grad_norm": 1.2929610515185452, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4294, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.724287783679382, | |
| "grad_norm": 1.4669951814767501, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4295, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.7339449541284404, | |
| "grad_norm": 1.3523170842518906, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4323, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.7436021245774987, | |
| "grad_norm": 1.4848119403639475, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4307, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.7532592950265572, | |
| "grad_norm": 1.2865027479600895, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4302, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.7629164654756156, | |
| "grad_norm": 1.3394197026037133, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4274, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.7725736359246741, | |
| "grad_norm": 1.4058992660952825, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4315, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.7822308063737325, | |
| "grad_norm": 1.4021237990828046, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4258, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.7918879768227909, | |
| "grad_norm": 1.4089771533232405, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4334, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.8015451472718493, | |
| "grad_norm": 1.3600046812066533, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4225, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.8112023177209078, | |
| "grad_norm": 1.4685840247095097, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4251, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.8208594881699662, | |
| "grad_norm": 1.3854952044022746, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4305, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.8305166586190246, | |
| "grad_norm": 1.3373112968229082, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4266, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.840173829068083, | |
| "grad_norm": 1.4141453052618997, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4276, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.8498309995171415, | |
| "grad_norm": 1.5086378917784917, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4247, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.8594881699661999, | |
| "grad_norm": 7.458182848059056, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4299, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.8691453404152584, | |
| "grad_norm": 1.5061946368232628, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4207, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.8788025108643167, | |
| "grad_norm": 1.2888259560849804, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4224, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.8884596813133752, | |
| "grad_norm": 1.4278585748957382, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4222, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.8981168517624336, | |
| "grad_norm": 1.4321653215479198, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4242, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.9077740222114921, | |
| "grad_norm": 1.3716154862902383, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4225, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.9174311926605505, | |
| "grad_norm": 1.3700235242916874, | |
| "learning_rate": 5e-06, | |
| "loss": 0.421, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.9270883631096088, | |
| "grad_norm": 1.3005175365880353, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4191, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.9367455335586673, | |
| "grad_norm": 1.3469399837188125, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4215, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.9464027040077257, | |
| "grad_norm": 1.3797432696350105, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4265, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.9560598744567842, | |
| "grad_norm": 1.2977225244391166, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4215, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.9657170449058425, | |
| "grad_norm": 1.2662429869365561, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4227, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.975374215354901, | |
| "grad_norm": 1.3086127516497834, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4178, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.9850313858039594, | |
| "grad_norm": 1.3519105272191883, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4186, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.9946885562530179, | |
| "grad_norm": 1.301344809598359, | |
| "learning_rate": 5e-06, | |
| "loss": 0.417, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.9995171414775471, | |
| "eval_loss": 0.4187374413013458, | |
| "eval_runtime": 182.0762, | |
| "eval_samples_per_second": 153.249, | |
| "eval_steps_per_second": 0.599, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 1.0043457267020763, | |
| "grad_norm": 2.0938717237725024, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3807, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.0140028971511348, | |
| "grad_norm": 1.6443261926343353, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3165, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.0236600676001932, | |
| "grad_norm": 1.3639427703718043, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3142, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.0333172380492515, | |
| "grad_norm": 1.4375989800345246, | |
| "learning_rate": 5e-06, | |
| "loss": 0.315, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.04297440849831, | |
| "grad_norm": 1.608018512856855, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3105, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.0526315789473684, | |
| "grad_norm": 1.6910926571603153, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3156, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.0622887493964268, | |
| "grad_norm": 1.501323197263165, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3214, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.0719459198454853, | |
| "grad_norm": 1.5198024912737862, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3187, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.0816030902945437, | |
| "grad_norm": 1.3967412194416047, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3177, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.0912602607436022, | |
| "grad_norm": 1.5169020279572174, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3194, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.1009174311926606, | |
| "grad_norm": 1.436092584626518, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3184, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.110574601641719, | |
| "grad_norm": 1.5133593820810858, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3209, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.1202317720907775, | |
| "grad_norm": 1.4358744153476208, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3162, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.1298889425398357, | |
| "grad_norm": 1.550919605032498, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3176, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.1395461129888942, | |
| "grad_norm": 1.4353404806567058, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3206, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.1492032834379526, | |
| "grad_norm": 1.6213115971080414, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3189, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.158860453887011, | |
| "grad_norm": 1.4735029754670448, | |
| "learning_rate": 5e-06, | |
| "loss": 0.318, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.1685176243360695, | |
| "grad_norm": 1.4914905722949483, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3259, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.178174794785128, | |
| "grad_norm": 1.4734223688553245, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3197, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.1878319652341864, | |
| "grad_norm": 1.4303986315832913, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3152, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.1974891356832449, | |
| "grad_norm": 1.641460216526965, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3203, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.2071463061323033, | |
| "grad_norm": 1.4854108071397198, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3187, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.2168034765813616, | |
| "grad_norm": 1.5657183004768043, | |
| "learning_rate": 5e-06, | |
| "loss": 0.321, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.22646064703042, | |
| "grad_norm": 1.5392245759709497, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3199, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.2361178174794785, | |
| "grad_norm": 1.4716569015272043, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3167, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.245774987928537, | |
| "grad_norm": 1.4273719929805762, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3203, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.2554321583775954, | |
| "grad_norm": 1.437096297299039, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3211, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.2650893288266538, | |
| "grad_norm": 1.4122099703029305, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3182, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.2747464992757123, | |
| "grad_norm": 1.62630879502805, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3217, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.2844036697247707, | |
| "grad_norm": 1.516391222863939, | |
| "learning_rate": 5e-06, | |
| "loss": 0.319, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.2940608401738292, | |
| "grad_norm": 1.9043021829524946, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3201, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.3037180106228874, | |
| "grad_norm": 1.468668927245736, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3185, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.3133751810719458, | |
| "grad_norm": 1.5192911853076112, | |
| "learning_rate": 5e-06, | |
| "loss": 0.321, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.3230323515210043, | |
| "grad_norm": 1.4775016904206435, | |
| "learning_rate": 5e-06, | |
| "loss": 0.323, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.3326895219700627, | |
| "grad_norm": 1.5738282394014576, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3244, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.3423466924191212, | |
| "grad_norm": 1.451975577099937, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3231, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.3520038628681796, | |
| "grad_norm": 1.509638768808278, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3238, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.361661033317238, | |
| "grad_norm": 1.4711514260625576, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3246, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.3713182037662965, | |
| "grad_norm": 1.5617390029956357, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3224, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.380975374215355, | |
| "grad_norm": 1.5052132010129833, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3225, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.3906325446644132, | |
| "grad_norm": 1.5656411569392583, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3151, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.4002897151134719, | |
| "grad_norm": 1.4546198678024953, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3186, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.4099468855625301, | |
| "grad_norm": 1.4514681189539558, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3252, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.4196040560115886, | |
| "grad_norm": 1.5187373109976763, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3218, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.429261226460647, | |
| "grad_norm": 1.4206099886595174, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3224, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.4389183969097055, | |
| "grad_norm": 1.3933469991291707, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3261, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.448575567358764, | |
| "grad_norm": 1.5275153572741944, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3196, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.4582327378078224, | |
| "grad_norm": 1.502387340983924, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3202, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.4678899082568808, | |
| "grad_norm": 1.4367426017329248, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3246, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.477547078705939, | |
| "grad_norm": 1.50109980877967, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3232, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.4872042491549977, | |
| "grad_norm": 1.6172108349800483, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3261, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.496861419604056, | |
| "grad_norm": 1.44998024463031, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3223, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.5065185900531144, | |
| "grad_norm": 1.4396628484827616, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3207, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.5161757605021728, | |
| "grad_norm": 1.4505846726645737, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3238, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.5258329309512313, | |
| "grad_norm": 1.3991156805400873, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3205, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.5354901014002897, | |
| "grad_norm": 1.8677762783077632, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3258, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.5451472718493482, | |
| "grad_norm": 1.5966624238860196, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3253, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.5548044422984066, | |
| "grad_norm": 1.4666709592985163, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3197, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.5644616127474649, | |
| "grad_norm": 1.5429686092632175, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3243, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.5741187831965235, | |
| "grad_norm": 1.6197267190074782, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3194, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.5837759536455818, | |
| "grad_norm": 1.5284451091184315, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3195, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.5934331240946402, | |
| "grad_norm": 1.4818015371419853, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3252, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.6030902945436987, | |
| "grad_norm": 1.5230446062018719, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3256, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.6127474649927571, | |
| "grad_norm": 1.3984458497949286, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3251, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.6224046354418156, | |
| "grad_norm": 1.3753762689022815, | |
| "learning_rate": 5e-06, | |
| "loss": 0.318, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.632061805890874, | |
| "grad_norm": 1.4229134627752649, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3206, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.6417189763399325, | |
| "grad_norm": 1.6325248655032423, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3254, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.6513761467889907, | |
| "grad_norm": 1.460452232456223, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3256, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.6610333172380494, | |
| "grad_norm": 1.4967322575333943, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3221, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.6706904876871076, | |
| "grad_norm": 1.4589595336063315, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3247, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.6803476581361663, | |
| "grad_norm": 1.535858835340216, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3218, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.6900048285852245, | |
| "grad_norm": 1.5876620484010096, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3232, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.699661999034283, | |
| "grad_norm": 1.4347951889442994, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3223, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.7093191694833414, | |
| "grad_norm": 1.462429783459939, | |
| "learning_rate": 5e-06, | |
| "loss": 0.321, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.7189763399323998, | |
| "grad_norm": 1.4922559510321622, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3231, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.7286335103814583, | |
| "grad_norm": 1.467959135285837, | |
| "learning_rate": 5e-06, | |
| "loss": 0.328, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.7382906808305165, | |
| "grad_norm": 1.49130361205012, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3266, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.7479478512795752, | |
| "grad_norm": 2.106032317323986, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3251, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.7576050217286334, | |
| "grad_norm": 1.543093994781833, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3243, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.767262192177692, | |
| "grad_norm": 1.4767806762226672, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3222, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.7769193626267503, | |
| "grad_norm": 1.6074675928640145, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3243, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.7865765330758088, | |
| "grad_norm": 1.586893521300227, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3271, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.7962337035248672, | |
| "grad_norm": 1.5199781278405553, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3226, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.8058908739739257, | |
| "grad_norm": 1.7072009333034235, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3251, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.8155480444229841, | |
| "grad_norm": 1.4657197536244577, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3232, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.8252052148720423, | |
| "grad_norm": 1.5502462503355572, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3241, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.834862385321101, | |
| "grad_norm": 1.4268551559294822, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3233, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.8445195557701592, | |
| "grad_norm": 1.5487349032646411, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3248, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.854176726219218, | |
| "grad_norm": 1.5587365739097951, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3262, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.8638338966682761, | |
| "grad_norm": 1.6410211222118756, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3243, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.8734910671173346, | |
| "grad_norm": 1.4412258241860316, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3242, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.883148237566393, | |
| "grad_norm": 1.408887524672963, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3207, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.8928054080154515, | |
| "grad_norm": 1.709744732252996, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3192, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.90246257846451, | |
| "grad_norm": 1.53624861095146, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3234, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.9121197489135682, | |
| "grad_norm": 1.4790016911744388, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3239, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.9217769193626268, | |
| "grad_norm": 1.4505596932174116, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3226, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.931434089811685, | |
| "grad_norm": 1.3770134690086049, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3253, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.9410912602607437, | |
| "grad_norm": 1.6550530625579105, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3255, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.950748430709802, | |
| "grad_norm": 1.5935808137814957, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3236, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.9604056011588604, | |
| "grad_norm": 1.5070382250631085, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3314, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.9700627716079189, | |
| "grad_norm": 1.3793762943653258, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3231, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.9797199420569773, | |
| "grad_norm": 1.4658674183812177, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3247, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.9893771125060358, | |
| "grad_norm": 1.5804425363669086, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3286, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.999034282955094, | |
| "grad_norm": 1.550360674186604, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3265, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.4136686325073242, | |
| "eval_runtime": 178.6997, | |
| "eval_samples_per_second": 156.145, | |
| "eval_steps_per_second": 0.61, | |
| "step": 2071 | |
| }, | |
| { | |
| "epoch": 2.0086914534041527, | |
| "grad_norm": 1.9820970989127087, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2226, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 2.018348623853211, | |
| "grad_norm": 1.746604108675904, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2022, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 2.0280057943022696, | |
| "grad_norm": 1.738922693622109, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2034, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 2.037662964751328, | |
| "grad_norm": 1.842908561465572, | |
| "learning_rate": 5e-06, | |
| "loss": 0.1987, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 2.0473201352003865, | |
| "grad_norm": 1.555963693945732, | |
| "learning_rate": 5e-06, | |
| "loss": 0.1998, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 2.0569773056494447, | |
| "grad_norm": 1.6105373644152785, | |
| "learning_rate": 5e-06, | |
| "loss": 0.1977, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 2.066634476098503, | |
| "grad_norm": 1.7277918775624639, | |
| "learning_rate": 5e-06, | |
| "loss": 0.1997, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 2.0762916465475616, | |
| "grad_norm": 1.7710659996843952, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2002, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 2.08594881699662, | |
| "grad_norm": 1.6510171974053387, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2025, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 2.0956059874456785, | |
| "grad_norm": 1.7020580751172847, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2005, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 2.1052631578947367, | |
| "grad_norm": 2.0101369980825474, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2076, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 2.1149203283437954, | |
| "grad_norm": 1.6487612820560622, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2031, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 2.1245774987928536, | |
| "grad_norm": 1.6667015280913091, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2042, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 2.1342346692419123, | |
| "grad_norm": 1.6702800446638546, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2016, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 2.1438918396909705, | |
| "grad_norm": 1.5921726575950466, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2017, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 2.153549010140029, | |
| "grad_norm": 1.6461690389920633, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2015, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 2.1632061805890874, | |
| "grad_norm": 1.7782309168905173, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2012, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 2.1728633510381457, | |
| "grad_norm": 1.768405330134614, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2058, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 2.1825205214872043, | |
| "grad_norm": 1.8294577091973667, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2039, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 2.1921776919362626, | |
| "grad_norm": 1.6757794760632077, | |
| "learning_rate": 5e-06, | |
| "loss": 0.206, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 2.2018348623853212, | |
| "grad_norm": 1.674322596519805, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2073, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 2.2114920328343795, | |
| "grad_norm": 1.729559352602187, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2082, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 2.221149203283438, | |
| "grad_norm": 1.7521877519603215, | |
| "learning_rate": 5e-06, | |
| "loss": 0.206, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 2.2308063737324964, | |
| "grad_norm": 1.6806851822041886, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2022, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 2.240463544181555, | |
| "grad_norm": 1.7855843714445576, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2092, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 2.2501207146306133, | |
| "grad_norm": 1.6890806520431167, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2089, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 2.2597778850796715, | |
| "grad_norm": 1.6876559538674738, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2076, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 2.26943505552873, | |
| "grad_norm": 1.866952527448764, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2101, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.2790922259777884, | |
| "grad_norm": 1.7778130347386727, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2092, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 2.288749396426847, | |
| "grad_norm": 1.7573632813881142, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2092, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 2.2984065668759053, | |
| "grad_norm": 1.670588490011592, | |
| "learning_rate": 5e-06, | |
| "loss": 0.21, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 2.308063737324964, | |
| "grad_norm": 2.0244107541981857, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2121, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 2.317720907774022, | |
| "grad_norm": 1.7228769196960538, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2065, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.327378078223081, | |
| "grad_norm": 1.7316110503319007, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2079, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 2.337035248672139, | |
| "grad_norm": 1.7859669994070049, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2106, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 2.3466924191211973, | |
| "grad_norm": 1.6790885546872112, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2134, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 2.356349589570256, | |
| "grad_norm": 1.658832098468064, | |
| "learning_rate": 5e-06, | |
| "loss": 0.209, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 2.366006760019314, | |
| "grad_norm": 1.737362445041373, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2086, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.375663930468373, | |
| "grad_norm": 1.627103813908181, | |
| "learning_rate": 5e-06, | |
| "loss": 0.214, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 2.385321100917431, | |
| "grad_norm": 1.7557722068636354, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2119, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 2.3949782713664898, | |
| "grad_norm": 1.676762758309891, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2143, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 2.404635441815548, | |
| "grad_norm": 1.8055582761883138, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2138, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 2.4142926122646067, | |
| "grad_norm": 1.8126267211804017, | |
| "learning_rate": 5e-06, | |
| "loss": 0.213, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.423949782713665, | |
| "grad_norm": 1.737365446135157, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2101, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 2.433606953162723, | |
| "grad_norm": 1.8601108967097126, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2113, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 2.443264123611782, | |
| "grad_norm": 1.7975172388473801, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2121, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 2.45292129406084, | |
| "grad_norm": 1.6548943065520165, | |
| "learning_rate": 5e-06, | |
| "loss": 0.213, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 2.4625784645098987, | |
| "grad_norm": 1.6142344009099507, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2101, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.472235634958957, | |
| "grad_norm": 1.7136093491402586, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2132, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.4818928054080156, | |
| "grad_norm": 1.6928112777667244, | |
| "learning_rate": 5e-06, | |
| "loss": 0.214, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 2.491549975857074, | |
| "grad_norm": 1.7487627188812938, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2121, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 2.5012071463061325, | |
| "grad_norm": 1.7611973606308302, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2146, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 2.5108643167551907, | |
| "grad_norm": 1.7432835398393358, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2131, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.520521487204249, | |
| "grad_norm": 1.914665414646886, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2135, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 2.5301786576533076, | |
| "grad_norm": 1.7235392993581606, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2139, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 2.539835828102366, | |
| "grad_norm": 1.8401761899916762, | |
| "learning_rate": 5e-06, | |
| "loss": 0.216, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 2.5494929985514245, | |
| "grad_norm": 1.7238390715314391, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2136, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 2.5591501690004828, | |
| "grad_norm": 1.6967163269012748, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2135, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.5688073394495414, | |
| "grad_norm": 1.6547200950158283, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2154, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 2.5784645098985997, | |
| "grad_norm": 1.7345144961212824, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2194, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 2.5881216803476583, | |
| "grad_norm": 1.695025030426855, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2148, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 2.5977788507967166, | |
| "grad_norm": 1.6586909849777336, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2177, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 2.607436021245775, | |
| "grad_norm": 1.6450495177327662, | |
| "learning_rate": 5e-06, | |
| "loss": 0.217, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.6170931916948335, | |
| "grad_norm": 1.75951484894314, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2161, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 2.6267503621438917, | |
| "grad_norm": 1.6771766744486507, | |
| "learning_rate": 5e-06, | |
| "loss": 0.216, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 2.6364075325929504, | |
| "grad_norm": 1.6522749567618924, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2164, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 2.6460647030420086, | |
| "grad_norm": 1.6904676125685896, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2183, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 2.6557218734910673, | |
| "grad_norm": 1.9074890209165016, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2185, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.6653790439401255, | |
| "grad_norm": 1.7817515462190572, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2166, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 2.675036214389184, | |
| "grad_norm": 1.7562563485880531, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2184, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 2.6846933848382424, | |
| "grad_norm": 1.998249733627222, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2207, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 2.6943505552873006, | |
| "grad_norm": 1.7695683095416619, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2154, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 2.7040077257363593, | |
| "grad_norm": 1.8078600319718185, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2194, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.7136648961854175, | |
| "grad_norm": 1.8027624003524778, | |
| "learning_rate": 5e-06, | |
| "loss": 0.216, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 2.723322066634476, | |
| "grad_norm": 1.6742700740217098, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2208, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 2.7329792370835344, | |
| "grad_norm": 1.773334112408393, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2188, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 2.742636407532593, | |
| "grad_norm": 1.6731213070854867, | |
| "learning_rate": 5e-06, | |
| "loss": 0.222, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 2.7522935779816513, | |
| "grad_norm": 1.7120912589030635, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2174, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.76195074843071, | |
| "grad_norm": 1.7123222485174125, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2177, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 2.771607918879768, | |
| "grad_norm": 1.6545568282400362, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2176, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 2.7812650893288264, | |
| "grad_norm": 2.0273088669056385, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2177, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 2.790922259777885, | |
| "grad_norm": 1.6066656179070626, | |
| "learning_rate": 5e-06, | |
| "loss": 0.219, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 2.8005794302269438, | |
| "grad_norm": 1.8053174623133685, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2176, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.810236600676002, | |
| "grad_norm": 1.70700256545585, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2196, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 2.8198937711250602, | |
| "grad_norm": 1.8175496500932709, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2207, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 2.829550941574119, | |
| "grad_norm": 1.7332734928998839, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2192, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 2.839208112023177, | |
| "grad_norm": 1.6737893973481872, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2198, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 2.848865282472236, | |
| "grad_norm": 1.861536272319189, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2211, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.858522452921294, | |
| "grad_norm": 1.7745633627018984, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2242, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 2.8681796233703523, | |
| "grad_norm": 1.7269651603359077, | |
| "learning_rate": 5e-06, | |
| "loss": 0.221, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 2.877836793819411, | |
| "grad_norm": 1.8445740625163256, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2205, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 2.8874939642684696, | |
| "grad_norm": 1.7256657735965188, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2187, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 2.897151134717528, | |
| "grad_norm": 1.7947034947358826, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2217, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.906808305166586, | |
| "grad_norm": 1.7396418325368443, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2183, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 2.9164654756156447, | |
| "grad_norm": 1.6566041516835608, | |
| "learning_rate": 5e-06, | |
| "loss": 0.224, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 2.926122646064703, | |
| "grad_norm": 1.7285392977136116, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2209, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 2.9357798165137616, | |
| "grad_norm": 1.6972125749702298, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2222, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 2.94543698696282, | |
| "grad_norm": 1.7429187695624144, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2197, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 2.955094157411878, | |
| "grad_norm": 1.7068313963220236, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2219, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 2.9647513278609368, | |
| "grad_norm": 1.6955378543746138, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2212, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 2.9744084983099954, | |
| "grad_norm": 1.685143060439258, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2233, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 2.9840656687590537, | |
| "grad_norm": 1.7308667414857535, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2226, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 2.993722839208112, | |
| "grad_norm": 1.7145293308716394, | |
| "learning_rate": 5e-06, | |
| "loss": 0.2217, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 2.998551424432641, | |
| "eval_loss": 0.44667962193489075, | |
| "eval_runtime": 176.8031, | |
| "eval_samples_per_second": 157.82, | |
| "eval_steps_per_second": 0.617, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 2.998551424432641, | |
| "step": 3105, | |
| "total_flos": 5200153128468480.0, | |
| "train_loss": 0.3276555372321087, | |
| "train_runtime": 26364.0416, | |
| "train_samples_per_second": 60.326, | |
| "train_steps_per_second": 0.118 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3105, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5200153128468480.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |