| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 1278, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.023508668821627974, | |
| "grad_norm": 84.37700653076172, | |
| "learning_rate": 7.031250000000001e-06, | |
| "loss": 115.85, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.04701733764325595, | |
| "grad_norm": 106.4520034790039, | |
| "learning_rate": 1.484375e-05, | |
| "loss": 116.5781, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07052600646488393, | |
| "grad_norm": 130.89254760742188, | |
| "learning_rate": 2.2656250000000002e-05, | |
| "loss": 103.3281, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0940346752865119, | |
| "grad_norm": 37.089256286621094, | |
| "learning_rate": 3.0468750000000002e-05, | |
| "loss": 60.0406, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.11754334410813988, | |
| "grad_norm": 44.051109313964844, | |
| "learning_rate": 3.828125e-05, | |
| "loss": 39.75, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.14105201292976785, | |
| "grad_norm": 43.561981201171875, | |
| "learning_rate": 4.609375e-05, | |
| "loss": 24.7984, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.16456068175139582, | |
| "grad_norm": 34.622520446777344, | |
| "learning_rate": 5.3906250000000006e-05, | |
| "loss": 11.9723, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.1880693505730238, | |
| "grad_norm": 22.26849937438965, | |
| "learning_rate": 6.171875e-05, | |
| "loss": 5.4578, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.2115780193946518, | |
| "grad_norm": 9.059814453125, | |
| "learning_rate": 6.953125e-05, | |
| "loss": 3.992, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.23508668821627976, | |
| "grad_norm": 4.332883834838867, | |
| "learning_rate": 7.734375e-05, | |
| "loss": 3.1058, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2585953570379077, | |
| "grad_norm": 6.844908237457275, | |
| "learning_rate": 8.515625e-05, | |
| "loss": 2.6654, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2821040258595357, | |
| "grad_norm": 4.539117813110352, | |
| "learning_rate": 9.296875e-05, | |
| "loss": 2.422, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.3056126946811637, | |
| "grad_norm": 3.098025321960449, | |
| "learning_rate": 9.999981342914437e-05, | |
| "loss": 2.1623, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.32912136350279164, | |
| "grad_norm": 1.3452321290969849, | |
| "learning_rate": 9.997742661115932e-05, | |
| "loss": 2.1429, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.35263003232441964, | |
| "grad_norm": 0.37226417660713196, | |
| "learning_rate": 9.991774476447404e-05, | |
| "loss": 2.1347, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.3761387011460476, | |
| "grad_norm": 3.523218870162964, | |
| "learning_rate": 9.982081242591919e-05, | |
| "loss": 2.0512, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3996473699676756, | |
| "grad_norm": 0.33414289355278015, | |
| "learning_rate": 9.968670193003843e-05, | |
| "loss": 2.047, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.4231560387893036, | |
| "grad_norm": 0.24106919765472412, | |
| "learning_rate": 9.951551335510978e-05, | |
| "loss": 2.1046, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.4466647076109315, | |
| "grad_norm": 0.328752726316452, | |
| "learning_rate": 9.930737444846331e-05, | |
| "loss": 2.1397, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4701733764325595, | |
| "grad_norm": 1.1409285068511963, | |
| "learning_rate": 9.906244053115143e-05, | |
| "loss": 2.1377, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.49368204525418746, | |
| "grad_norm": 0.0690290704369545, | |
| "learning_rate": 9.87808943820424e-05, | |
| "loss": 2.0992, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.5171907140758154, | |
| "grad_norm": 0.20527280867099762, | |
| "learning_rate": 9.846294610142398e-05, | |
| "loss": 2.1339, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5406993828974435, | |
| "grad_norm": 0.3991909623146057, | |
| "learning_rate": 9.810883295421864e-05, | |
| "loss": 2.1176, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5642080517190714, | |
| "grad_norm": 0.08950258791446686, | |
| "learning_rate": 9.771881919292765e-05, | |
| "loss": 2.0895, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5877167205406993, | |
| "grad_norm": 0.17848969995975494, | |
| "learning_rate": 9.729319586043591e-05, | |
| "loss": 2.0877, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.6112253893623274, | |
| "grad_norm": 0.30160897970199585, | |
| "learning_rate": 9.683228057282483e-05, | |
| "loss": 2.0648, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6347340581839553, | |
| "grad_norm": 0.11883804202079773, | |
| "learning_rate": 9.63364172823554e-05, | |
| "loss": 2.0591, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6582427270055833, | |
| "grad_norm": 0.1619461327791214, | |
| "learning_rate": 9.580597602079802e-05, | |
| "loss": 2.1386, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6817513958272113, | |
| "grad_norm": 0.11473017930984497, | |
| "learning_rate": 9.524135262330098e-05, | |
| "loss": 2.1028, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.7052600646488393, | |
| "grad_norm": 0.059143248945474625, | |
| "learning_rate": 9.464296843300342e-05, | |
| "loss": 2.0881, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7287687334704672, | |
| "grad_norm": 0.1624162495136261, | |
| "learning_rate": 9.401126998661328e-05, | |
| "loss": 2.0971, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.7522774022920952, | |
| "grad_norm": 0.15460653603076935, | |
| "learning_rate": 9.334672868118491e-05, | |
| "loss": 2.1154, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.7757860711137232, | |
| "grad_norm": 0.4213317334651947, | |
| "learning_rate": 9.26498404223449e-05, | |
| "loss": 2.1113, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7992947399353512, | |
| "grad_norm": 0.14529550075531006, | |
| "learning_rate": 9.192112525422868e-05, | |
| "loss": 2.0867, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.8228034087569791, | |
| "grad_norm": 0.06652400642633438, | |
| "learning_rate": 9.116112697140418e-05, | |
| "loss": 2.0807, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.8463120775786072, | |
| "grad_norm": 0.12574820220470428, | |
| "learning_rate": 9.037041271307188e-05, | |
| "loss": 2.1711, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.8698207464002351, | |
| "grad_norm": 0.10869150608778, | |
| "learning_rate": 8.954957253984426e-05, | |
| "loss": 2.1252, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.893329415221863, | |
| "grad_norm": 0.143524169921875, | |
| "learning_rate": 8.869921899342056e-05, | |
| "loss": 2.0468, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.916838084043491, | |
| "grad_norm": 0.27682727575302124, | |
| "learning_rate": 8.781998663948513e-05, | |
| "loss": 2.1023, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.940346752865119, | |
| "grad_norm": 0.08671136200428009, | |
| "learning_rate": 8.691253159417074e-05, | |
| "loss": 2.0991, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.963855421686747, | |
| "grad_norm": 0.14743109047412872, | |
| "learning_rate": 8.597753103444016e-05, | |
| "loss": 2.0858, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.9873640905083749, | |
| "grad_norm": 0.14369799196720123, | |
| "learning_rate": 8.501568269275126e-05, | |
| "loss": 2.1057, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.0094034675286512, | |
| "grad_norm": 0.09220755100250244, | |
| "learning_rate": 8.40277043363831e-05, | |
| "loss": 1.9405, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.0329121363502791, | |
| "grad_norm": 0.25697803497314453, | |
| "learning_rate": 8.301433323181076e-05, | |
| "loss": 2.1221, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.056420805171907, | |
| "grad_norm": 0.1321459412574768, | |
| "learning_rate": 8.19763255945298e-05, | |
| "loss": 2.1576, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.079929473993535, | |
| "grad_norm": 0.11310122162103653, | |
| "learning_rate": 8.091445602473972e-05, | |
| "loss": 2.0483, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.1034381428151632, | |
| "grad_norm": 0.08170254528522491, | |
| "learning_rate": 7.982951692930829e-05, | |
| "loss": 2.1367, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.1269468116367911, | |
| "grad_norm": 0.022492246702313423, | |
| "learning_rate": 7.87223179304479e-05, | |
| "loss": 2.1435, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.150455480458419, | |
| "grad_norm": 0.015893638134002686, | |
| "learning_rate": 7.759368526154509e-05, | |
| "loss": 2.0666, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.173964149280047, | |
| "grad_norm": 0.053727954626083374, | |
| "learning_rate": 7.644446115059425e-05, | |
| "loss": 2.0886, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.197472818101675, | |
| "grad_norm": 0.12717482447624207, | |
| "learning_rate": 7.527550319169546e-05, | |
| "loss": 2.1418, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.2209814869233029, | |
| "grad_norm": 0.030752327293157578, | |
| "learning_rate": 7.408768370508576e-05, | |
| "loss": 2.0688, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.244490155744931, | |
| "grad_norm": 0.07915918529033661, | |
| "learning_rate": 7.288188908618102e-05, | |
| "loss": 2.0857, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.267998824566559, | |
| "grad_norm": 0.09853250533342361, | |
| "learning_rate": 7.165901914411435e-05, | |
| "loss": 2.0392, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.291507493388187, | |
| "grad_norm": 0.21571961045265198, | |
| "learning_rate": 7.041998643026511e-05, | |
| "loss": 2.0921, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.3150161622098149, | |
| "grad_norm": 0.10290564596652985, | |
| "learning_rate": 6.916571555727852e-05, | |
| "loss": 2.0686, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.3385248310314428, | |
| "grad_norm": 0.08689925819635391, | |
| "learning_rate": 6.789714250908533e-05, | |
| "loss": 2.0845, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.3620334998530708, | |
| "grad_norm": 0.1267048567533493, | |
| "learning_rate": 6.661521394243533e-05, | |
| "loss": 2.1465, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.3855421686746987, | |
| "grad_norm": 0.18292003870010376, | |
| "learning_rate": 6.532088648046677e-05, | |
| "loss": 2.067, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.4090508374963266, | |
| "grad_norm": 0.051083244383335114, | |
| "learning_rate": 6.401512599883818e-05, | |
| "loss": 2.0697, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.4325595063179548, | |
| "grad_norm": 0.0492679663002491, | |
| "learning_rate": 6.269890690495593e-05, | |
| "loss": 2.1064, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.4560681751395828, | |
| "grad_norm": 0.051078781485557556, | |
| "learning_rate": 6.137321141083468e-05, | |
| "loss": 2.0533, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.4795768439612107, | |
| "grad_norm": 0.11715124547481537, | |
| "learning_rate": 6.003902880013416e-05, | |
| "loss": 2.1156, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.5030855127828386, | |
| "grad_norm": 0.041975561529397964, | |
| "learning_rate": 5.869735468991854e-05, | |
| "loss": 2.0543, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.5265941816044668, | |
| "grad_norm": 0.10740803927183151, | |
| "learning_rate": 5.73491902876897e-05, | |
| "loss": 2.0539, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.5501028504260947, | |
| "grad_norm": 0.07230595499277115, | |
| "learning_rate": 5.599554164424859e-05, | |
| "loss": 2.1078, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.5736115192477227, | |
| "grad_norm": 0.010545202530920506, | |
| "learning_rate": 5.463741890294257e-05, | |
| "loss": 2.1434, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.5971201880693506, | |
| "grad_norm": 0.0882532000541687, | |
| "learning_rate": 5.327583554585842e-05, | |
| "loss": 2.0696, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.6206288568909786, | |
| "grad_norm": 0.15730935335159302, | |
| "learning_rate": 5.191180763752418e-05, | |
| "loss": 2.0743, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.6441375257126065, | |
| "grad_norm": 0.16760171949863434, | |
| "learning_rate": 5.054635306668355e-05, | |
| "loss": 2.1303, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.6676461945342345, | |
| "grad_norm": 0.12746253609657288, | |
| "learning_rate": 4.9180490786709335e-05, | |
| "loss": 2.1359, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.6911548633558624, | |
| "grad_norm": 0.08576823770999908, | |
| "learning_rate": 4.781524005522216e-05, | |
| "loss": 2.0884, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.7146635321774903, | |
| "grad_norm": 0.01966356858611107, | |
| "learning_rate": 4.645161967348223e-05, | |
| "loss": 2.1102, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.7381722009991183, | |
| "grad_norm": 0.1140172928571701, | |
| "learning_rate": 4.509064722612169e-05, | |
| "loss": 2.134, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.7616808698207465, | |
| "grad_norm": 0.11263474822044373, | |
| "learning_rate": 4.373333832178478e-05, | |
| "loss": 2.212, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.7851895386423744, | |
| "grad_norm": 0.01955031044781208, | |
| "learning_rate": 4.2380705835242716e-05, | |
| "loss": 2.0876, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.8086982074640023, | |
| "grad_norm": 0.02556225284934044, | |
| "learning_rate": 4.10337591515484e-05, | |
| "loss": 2.064, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.8322068762856303, | |
| "grad_norm": 0.04094263166189194, | |
| "learning_rate": 3.969350341279566e-05, | |
| "loss": 2.1206, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.8557155451072584, | |
| "grad_norm": 0.0680101215839386, | |
| "learning_rate": 3.8360938768044405e-05, | |
| "loss": 2.0975, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.8792242139288864, | |
| "grad_norm": 0.08410263061523438, | |
| "learning_rate": 3.7037059626971944e-05, | |
| "loss": 2.0519, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.9027328827505143, | |
| "grad_norm": 0.02342092990875244, | |
| "learning_rate": 3.572285391780734e-05, | |
| "loss": 2.019, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.9262415515721423, | |
| "grad_norm": 0.02327684499323368, | |
| "learning_rate": 3.441930235010211e-05, | |
| "loss": 2.0966, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.9497502203937702, | |
| "grad_norm": 0.0357813723385334, | |
| "learning_rate": 3.312737768288827e-05, | |
| "loss": 2.0924, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.9732588892153982, | |
| "grad_norm": 0.07186655700206757, | |
| "learning_rate": 3.184804399876886e-05, | |
| "loss": 2.0315, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.996767558037026, | |
| "grad_norm": 0.09240284562110901, | |
| "learning_rate": 3.0582255984483535e-05, | |
| "loss": 2.0409, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.0188069350573024, | |
| "grad_norm": 0.04462951049208641, | |
| "learning_rate": 2.9330958218485392e-05, | |
| "loss": 1.9696, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.0423156038789303, | |
| "grad_norm": 0.029340475797653198, | |
| "learning_rate": 2.8095084466061367e-05, | |
| "loss": 2.1055, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.0658242727005582, | |
| "grad_norm": 0.048022590577602386, | |
| "learning_rate": 2.6875556982521248e-05, | |
| "loss": 2.0865, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.089332941522186, | |
| "grad_norm": 0.04670249670743942, | |
| "learning_rate": 2.5673285824976575e-05, | |
| "loss": 2.1081, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.112841610343814, | |
| "grad_norm": 0.015516964718699455, | |
| "learning_rate": 2.4489168173221567e-05, | |
| "loss": 2.1023, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.136350279165442, | |
| "grad_norm": 0.06762902438640594, | |
| "learning_rate": 2.33240876602242e-05, | |
| "loss": 2.0929, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.15985894798707, | |
| "grad_norm": 0.014990041963756084, | |
| "learning_rate": 2.2178913712726045e-05, | |
| "loss": 2.0817, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.1833676168086984, | |
| "grad_norm": 0.07376622408628464, | |
| "learning_rate": 2.1054500902443423e-05, | |
| "loss": 2.0868, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.2068762856303263, | |
| "grad_norm": 0.03299334645271301, | |
| "learning_rate": 1.9951688308354133e-05, | |
| "loss": 2.1123, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.2303849544519543, | |
| "grad_norm": 0.05409041792154312, | |
| "learning_rate": 1.887129889054503e-05, | |
| "loss": 2.051, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.2538936232735822, | |
| "grad_norm": 0.041158534586429596, | |
| "learning_rate": 1.7814138876088566e-05, | |
| "loss": 2.0458, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.27740229209521, | |
| "grad_norm": 0.03300878778100014, | |
| "learning_rate": 1.6780997157405846e-05, | |
| "loss": 2.0565, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.300910960916838, | |
| "grad_norm": 0.009902500547468662, | |
| "learning_rate": 1.5772644703565565e-05, | |
| "loss": 2.0952, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.324419629738466, | |
| "grad_norm": 0.44653964042663574, | |
| "learning_rate": 1.4789833984957962e-05, | |
| "loss": 2.0386, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.347928298560094, | |
| "grad_norm": 0.4100548326969147, | |
| "learning_rate": 1.3833298411773276e-05, | |
| "loss": 2.0977, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 2.371436967381722, | |
| "grad_norm": 0.00745609262958169, | |
| "learning_rate": 1.2903751786703482e-05, | |
| "loss": 2.0503, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 2.39494563620335, | |
| "grad_norm": 0.05212165042757988, | |
| "learning_rate": 1.2001887772276076e-05, | |
| "loss": 2.1437, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 2.418454305024978, | |
| "grad_norm": 0.0367308035492897, | |
| "learning_rate": 1.1128379373217058e-05, | |
| "loss": 2.0911, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 2.4419629738466058, | |
| "grad_norm": 0.02135683037340641, | |
| "learning_rate": 1.028387843422952e-05, | |
| "loss": 2.1183, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 2.465471642668234, | |
| "grad_norm": 0.02447247877717018, | |
| "learning_rate": 9.469015153562855e-06, | |
| "loss": 2.1205, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 2.488980311489862, | |
| "grad_norm": 0.011414138600230217, | |
| "learning_rate": 8.684397612735024e-06, | |
| "loss": 2.0027, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 2.51248898031149, | |
| "grad_norm": 0.0028556822799146175, | |
| "learning_rate": 7.930611322759462e-06, | |
| "loss": 2.0682, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 2.535997649133118, | |
| "grad_norm": 0.008712991140782833, | |
| "learning_rate": 7.2082187872147935e-06, | |
| "loss": 2.0895, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 2.559506317954746, | |
| "grad_norm": 0.03148069605231285, | |
| "learning_rate": 6.517759082483543e-06, | |
| "loss": 2.1298, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 2.583014986776374, | |
| "grad_norm": 0.03679489344358444, | |
| "learning_rate": 5.8597474554731745e-06, | |
| "loss": 2.0837, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 2.606523655598002, | |
| "grad_norm": 0.05236874148249626, | |
| "learning_rate": 5.234674939119538e-06, | |
| "loss": 2.1486, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 2.6300323244196298, | |
| "grad_norm": 0.03369889780879021, | |
| "learning_rate": 4.643007985959641e-06, | |
| "loss": 2.0959, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 2.6535409932412577, | |
| "grad_norm": 0.006195698864758015, | |
| "learning_rate": 4.085188120047362e-06, | |
| "loss": 2.0439, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 2.6770496620628856, | |
| "grad_norm": 0.03900829330086708, | |
| "learning_rate": 3.5616316074715995e-06, | |
| "loss": 2.111, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.7005583308845136, | |
| "grad_norm": 0.00790059007704258, | |
| "learning_rate": 3.0727291457229303e-06, | |
| "loss": 2.1078, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 2.7240669997061415, | |
| "grad_norm": 0.017095139250159264, | |
| "learning_rate": 2.618845572140527e-06, | |
| "loss": 2.1248, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.7475756685277695, | |
| "grad_norm": 0.006935155484825373, | |
| "learning_rate": 2.2003195916567855e-06, | |
| "loss": 2.0573, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 2.7710843373493974, | |
| "grad_norm": 0.01475539617240429, | |
| "learning_rate": 1.8174635240431282e-06, | |
| "loss": 2.0218, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.7945930061710254, | |
| "grad_norm": 0.08390510082244873, | |
| "learning_rate": 1.4705630708451712e-06, | |
| "loss": 2.0729, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 2.8181016749926533, | |
| "grad_norm": 0.04823232442140579, | |
| "learning_rate": 1.1598771021816058e-06, | |
| "loss": 2.0432, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.8416103438142817, | |
| "grad_norm": 0.017384668812155724, | |
| "learning_rate": 8.856374635655695e-07, | |
| "loss": 2.1005, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 2.8651190126359096, | |
| "grad_norm": 0.016273437067866325, | |
| "learning_rate": 6.480488028928422e-07, | |
| "loss": 2.1043, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.8886276814575376, | |
| "grad_norm": 0.004406425170600414, | |
| "learning_rate": 4.47288417725944e-07, | |
| "loss": 2.1214, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 2.9121363502791655, | |
| "grad_norm": 0.04330357536673546, | |
| "learning_rate": 2.8350612298801427e-07, | |
| "loss": 2.0679, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.9356450191007935, | |
| "grad_norm": 0.005912340711802244, | |
| "learning_rate": 1.5682413916531802e-07, | |
| "loss": 2.1937, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.9591536879224214, | |
| "grad_norm": 0.021025869995355606, | |
| "learning_rate": 6.733700110174157e-08, | |
| "loss": 2.1239, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.9826623567440493, | |
| "grad_norm": 0.006763559300452471, | |
| "learning_rate": 1.511148745335089e-08, | |
| "loss": 2.1434, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 1278, | |
| "total_flos": 2.0454733263908045e+17, | |
| "train_loss": 5.729273645344288, | |
| "train_runtime": 5128.0315, | |
| "train_samples_per_second": 1.991, | |
| "train_steps_per_second": 0.249 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1278, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.0454733263908045e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |