| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 20583, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004858378273332362, | |
| "grad_norm": 66137.25, | |
| "learning_rate": 4.9757081086333386e-05, | |
| "loss": 0.5598, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.009716756546664724, | |
| "grad_norm": 60222.27734375, | |
| "learning_rate": 4.951416217266677e-05, | |
| "loss": 0.5419, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.014575134819997084, | |
| "grad_norm": 67718.328125, | |
| "learning_rate": 4.927124325900015e-05, | |
| "loss": 0.5491, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.019433513093329448, | |
| "grad_norm": 73856.1015625, | |
| "learning_rate": 4.902832434533353e-05, | |
| "loss": 0.5496, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.024291891366661808, | |
| "grad_norm": 60560.8359375, | |
| "learning_rate": 4.878540543166691e-05, | |
| "loss": 0.5381, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.029150269639994168, | |
| "grad_norm": 76622.125, | |
| "learning_rate": 4.8542486518000295e-05, | |
| "loss": 0.5397, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.03400864791332653, | |
| "grad_norm": 55521.8203125, | |
| "learning_rate": 4.829956760433367e-05, | |
| "loss": 0.5339, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.038867026186658896, | |
| "grad_norm": 59273.4921875, | |
| "learning_rate": 4.8056648690667055e-05, | |
| "loss": 0.5342, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.043725404459991256, | |
| "grad_norm": 63513.421875, | |
| "learning_rate": 4.781372977700044e-05, | |
| "loss": 0.5247, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.048583782733323616, | |
| "grad_norm": 61088.77734375, | |
| "learning_rate": 4.757081086333382e-05, | |
| "loss": 0.5302, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.053442161006655976, | |
| "grad_norm": 71392.75, | |
| "learning_rate": 4.7327891949667204e-05, | |
| "loss": 0.5261, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.058300539279988337, | |
| "grad_norm": 72375.1328125, | |
| "learning_rate": 4.708497303600059e-05, | |
| "loss": 0.5153, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.0631589175533207, | |
| "grad_norm": 65540.171875, | |
| "learning_rate": 4.6842054122333964e-05, | |
| "loss": 0.5068, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.06801729582665306, | |
| "grad_norm": 63439.60546875, | |
| "learning_rate": 4.659913520866735e-05, | |
| "loss": 0.5087, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.07287567409998542, | |
| "grad_norm": 56190.2265625, | |
| "learning_rate": 4.635621629500073e-05, | |
| "loss": 0.5074, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.07773405237331779, | |
| "grad_norm": 72489.0234375, | |
| "learning_rate": 4.6113297381334114e-05, | |
| "loss": 0.5036, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.08259243064665014, | |
| "grad_norm": 55422.61328125, | |
| "learning_rate": 4.58703784676675e-05, | |
| "loss": 0.4979, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.08745080891998251, | |
| "grad_norm": 58258.125, | |
| "learning_rate": 4.562745955400088e-05, | |
| "loss": 0.5016, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.09230918719331486, | |
| "grad_norm": 57958.55859375, | |
| "learning_rate": 4.5384540640334264e-05, | |
| "loss": 0.508, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.09716756546664723, | |
| "grad_norm": 52961.26953125, | |
| "learning_rate": 4.514162172666764e-05, | |
| "loss": 0.4843, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.1020259437399796, | |
| "grad_norm": 57415.234375, | |
| "learning_rate": 4.489870281300102e-05, | |
| "loss": 0.4943, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.10688432201331195, | |
| "grad_norm": 49461.50390625, | |
| "learning_rate": 4.46557838993344e-05, | |
| "loss": 0.4817, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.11174270028664432, | |
| "grad_norm": 51018.04296875, | |
| "learning_rate": 4.441286498566778e-05, | |
| "loss": 0.4825, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.11660107855997667, | |
| "grad_norm": 52488.46484375, | |
| "learning_rate": 4.4169946072001166e-05, | |
| "loss": 0.4786, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.12145945683330904, | |
| "grad_norm": 61512.5078125, | |
| "learning_rate": 4.392702715833455e-05, | |
| "loss": 0.4775, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.1263178351066414, | |
| "grad_norm": 56490.4296875, | |
| "learning_rate": 4.368410824466793e-05, | |
| "loss": 0.4813, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.13117621337997376, | |
| "grad_norm": 63387.01171875, | |
| "learning_rate": 4.3441189331001316e-05, | |
| "loss": 0.4744, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.13603459165330611, | |
| "grad_norm": 62635.20703125, | |
| "learning_rate": 4.31982704173347e-05, | |
| "loss": 0.4758, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.1408929699266385, | |
| "grad_norm": 52577.96875, | |
| "learning_rate": 4.2955351503668076e-05, | |
| "loss": 0.4867, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.14575134819997085, | |
| "grad_norm": 54731.5, | |
| "learning_rate": 4.271243259000146e-05, | |
| "loss": 0.4658, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.1506097264733032, | |
| "grad_norm": 52055.84765625, | |
| "learning_rate": 4.246951367633484e-05, | |
| "loss": 0.4737, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.15546810474663558, | |
| "grad_norm": 48978.90234375, | |
| "learning_rate": 4.2226594762668225e-05, | |
| "loss": 0.4765, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.16032648301996794, | |
| "grad_norm": 54819.38671875, | |
| "learning_rate": 4.198367584900161e-05, | |
| "loss": 0.466, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.1651848612933003, | |
| "grad_norm": 58637.75390625, | |
| "learning_rate": 4.1740756935334985e-05, | |
| "loss": 0.4725, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.17004323956663267, | |
| "grad_norm": 55842.4609375, | |
| "learning_rate": 4.149783802166837e-05, | |
| "loss": 0.4724, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.17490161783996502, | |
| "grad_norm": 52083.8125, | |
| "learning_rate": 4.125491910800175e-05, | |
| "loss": 0.4796, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.17975999611329738, | |
| "grad_norm": 56764.9296875, | |
| "learning_rate": 4.1012000194335135e-05, | |
| "loss": 0.4736, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.18461837438662973, | |
| "grad_norm": 50524.82421875, | |
| "learning_rate": 4.076908128066851e-05, | |
| "loss": 0.4627, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.1894767526599621, | |
| "grad_norm": 50128.8359375, | |
| "learning_rate": 4.0526162367001894e-05, | |
| "loss": 0.4717, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.19433513093329446, | |
| "grad_norm": 72415.296875, | |
| "learning_rate": 4.028324345333528e-05, | |
| "loss": 0.456, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.19919350920662682, | |
| "grad_norm": 57324.921875, | |
| "learning_rate": 4.004032453966866e-05, | |
| "loss": 0.4505, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.2040518874799592, | |
| "grad_norm": 60287.41796875, | |
| "learning_rate": 3.9797405626002044e-05, | |
| "loss": 0.4441, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.20891026575329155, | |
| "grad_norm": 65722.1171875, | |
| "learning_rate": 3.955448671233543e-05, | |
| "loss": 0.4426, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.2137686440266239, | |
| "grad_norm": 57532.5625, | |
| "learning_rate": 3.931156779866881e-05, | |
| "loss": 0.4532, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.2186270222999563, | |
| "grad_norm": 60035.2265625, | |
| "learning_rate": 3.906864888500219e-05, | |
| "loss": 0.443, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.22348540057328864, | |
| "grad_norm": 50842.09375, | |
| "learning_rate": 3.882572997133557e-05, | |
| "loss": 0.4527, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.228343778846621, | |
| "grad_norm": 53098.82421875, | |
| "learning_rate": 3.858281105766895e-05, | |
| "loss": 0.441, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.23320215711995335, | |
| "grad_norm": 52847.52734375, | |
| "learning_rate": 3.833989214400233e-05, | |
| "loss": 0.4437, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.23806053539328573, | |
| "grad_norm": 55201.1953125, | |
| "learning_rate": 3.809697323033571e-05, | |
| "loss": 0.4534, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.24291891366661808, | |
| "grad_norm": 55352.77734375, | |
| "learning_rate": 3.7854054316669096e-05, | |
| "loss": 0.4559, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.24777729193995043, | |
| "grad_norm": 77840.125, | |
| "learning_rate": 3.761113540300248e-05, | |
| "loss": 0.4368, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.2526356702132828, | |
| "grad_norm": 69213.84375, | |
| "learning_rate": 3.736821648933586e-05, | |
| "loss": 0.4428, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.25749404848661517, | |
| "grad_norm": 61722.49609375, | |
| "learning_rate": 3.7125297575669246e-05, | |
| "loss": 0.444, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.2623524267599475, | |
| "grad_norm": 56948.3125, | |
| "learning_rate": 3.688237866200262e-05, | |
| "loss": 0.4423, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.2672108050332799, | |
| "grad_norm": 50691.4453125, | |
| "learning_rate": 3.6639459748336006e-05, | |
| "loss": 0.4371, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.27206918330661223, | |
| "grad_norm": 48622.8671875, | |
| "learning_rate": 3.639654083466939e-05, | |
| "loss": 0.4459, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.27692756157994464, | |
| "grad_norm": 58408.26171875, | |
| "learning_rate": 3.615362192100277e-05, | |
| "loss": 0.4215, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.281785939853277, | |
| "grad_norm": 48251.1953125, | |
| "learning_rate": 3.5910703007336156e-05, | |
| "loss": 0.4204, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.28664431812660934, | |
| "grad_norm": 73270.4453125, | |
| "learning_rate": 3.566778409366954e-05, | |
| "loss": 0.4421, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.2915026963999417, | |
| "grad_norm": 52048.3828125, | |
| "learning_rate": 3.5424865180002915e-05, | |
| "loss": 0.4438, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.29636107467327405, | |
| "grad_norm": 57475.203125, | |
| "learning_rate": 3.51819462663363e-05, | |
| "loss": 0.418, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.3012194529466064, | |
| "grad_norm": 55124.11328125, | |
| "learning_rate": 3.493902735266968e-05, | |
| "loss": 0.4347, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.3060778312199388, | |
| "grad_norm": 62276.3671875, | |
| "learning_rate": 3.469610843900306e-05, | |
| "loss": 0.4192, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.31093620949327117, | |
| "grad_norm": 52723.7421875, | |
| "learning_rate": 3.445318952533644e-05, | |
| "loss": 0.4413, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.3157945877666035, | |
| "grad_norm": 53744.6484375, | |
| "learning_rate": 3.4210270611669825e-05, | |
| "loss": 0.4324, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.32065296603993587, | |
| "grad_norm": 44454.984375, | |
| "learning_rate": 3.396735169800321e-05, | |
| "loss": 0.4269, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.3255113443132682, | |
| "grad_norm": 52088.671875, | |
| "learning_rate": 3.372443278433659e-05, | |
| "loss": 0.4317, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.3303697225866006, | |
| "grad_norm": 50359.18359375, | |
| "learning_rate": 3.3481513870669974e-05, | |
| "loss": 0.4102, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.33522810085993293, | |
| "grad_norm": 54712.453125, | |
| "learning_rate": 3.323859495700336e-05, | |
| "loss": 0.4194, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.34008647913326534, | |
| "grad_norm": 49144.39453125, | |
| "learning_rate": 3.2995676043336734e-05, | |
| "loss": 0.414, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.3449448574065977, | |
| "grad_norm": 48121.30078125, | |
| "learning_rate": 3.275275712967012e-05, | |
| "loss": 0.4203, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.34980323567993005, | |
| "grad_norm": 60245.484375, | |
| "learning_rate": 3.25098382160035e-05, | |
| "loss": 0.4227, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.3546616139532624, | |
| "grad_norm": 53339.56640625, | |
| "learning_rate": 3.226691930233688e-05, | |
| "loss": 0.4235, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.35951999222659475, | |
| "grad_norm": 61334.859375, | |
| "learning_rate": 3.202400038867026e-05, | |
| "loss": 0.4318, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.3643783704999271, | |
| "grad_norm": 61368.46484375, | |
| "learning_rate": 3.1781081475003643e-05, | |
| "loss": 0.4223, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.36923674877325946, | |
| "grad_norm": 67382.265625, | |
| "learning_rate": 3.153816256133703e-05, | |
| "loss": 0.4053, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.37409512704659187, | |
| "grad_norm": 56739.12890625, | |
| "learning_rate": 3.129524364767041e-05, | |
| "loss": 0.4121, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.3789535053199242, | |
| "grad_norm": 51203.3203125, | |
| "learning_rate": 3.105232473400379e-05, | |
| "loss": 0.432, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.3838118835932566, | |
| "grad_norm": 58815.6484375, | |
| "learning_rate": 3.080940582033717e-05, | |
| "loss": 0.405, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.38867026186658893, | |
| "grad_norm": 59941.88671875, | |
| "learning_rate": 3.056648690667055e-05, | |
| "loss": 0.425, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.3935286401399213, | |
| "grad_norm": 53543.46484375, | |
| "learning_rate": 3.0323567993003936e-05, | |
| "loss": 0.4103, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.39838701841325364, | |
| "grad_norm": 55630.40625, | |
| "learning_rate": 3.008064907933732e-05, | |
| "loss": 0.3964, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.40324539668658604, | |
| "grad_norm": 52717.6953125, | |
| "learning_rate": 2.9837730165670703e-05, | |
| "loss": 0.3989, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.4081037749599184, | |
| "grad_norm": 50914.84375, | |
| "learning_rate": 2.9594811252004086e-05, | |
| "loss": 0.4097, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.41296215323325075, | |
| "grad_norm": 61387.0390625, | |
| "learning_rate": 2.9351892338337466e-05, | |
| "loss": 0.4051, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.4178205315065831, | |
| "grad_norm": 45076.01953125, | |
| "learning_rate": 2.910897342467085e-05, | |
| "loss": 0.4012, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.42267890977991546, | |
| "grad_norm": 54953.07421875, | |
| "learning_rate": 2.8866054511004225e-05, | |
| "loss": 0.4202, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.4275372880532478, | |
| "grad_norm": 59405.3828125, | |
| "learning_rate": 2.862313559733761e-05, | |
| "loss": 0.3931, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.43239566632658016, | |
| "grad_norm": 61979.93359375, | |
| "learning_rate": 2.8380216683670992e-05, | |
| "loss": 0.3987, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.4372540445999126, | |
| "grad_norm": 46517.14453125, | |
| "learning_rate": 2.813729777000437e-05, | |
| "loss": 0.4135, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.4421124228732449, | |
| "grad_norm": 57222.25, | |
| "learning_rate": 2.7894378856337755e-05, | |
| "loss": 0.4048, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.4469708011465773, | |
| "grad_norm": 60419.09375, | |
| "learning_rate": 2.7651459942671138e-05, | |
| "loss": 0.4008, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.45182917941990963, | |
| "grad_norm": 42610.984375, | |
| "learning_rate": 2.740854102900452e-05, | |
| "loss": 0.3975, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.456687557693242, | |
| "grad_norm": 58822.0234375, | |
| "learning_rate": 2.71656221153379e-05, | |
| "loss": 0.3923, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.46154593596657434, | |
| "grad_norm": 47564.5234375, | |
| "learning_rate": 2.6922703201671284e-05, | |
| "loss": 0.391, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.4664043142399067, | |
| "grad_norm": 48113.20703125, | |
| "learning_rate": 2.6679784288004668e-05, | |
| "loss": 0.4064, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.4712626925132391, | |
| "grad_norm": 49079.30078125, | |
| "learning_rate": 2.6436865374338048e-05, | |
| "loss": 0.4077, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.47612107078657145, | |
| "grad_norm": 53815.7890625, | |
| "learning_rate": 2.619394646067143e-05, | |
| "loss": 0.3844, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.4809794490599038, | |
| "grad_norm": 45962.41796875, | |
| "learning_rate": 2.5951027547004814e-05, | |
| "loss": 0.3925, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.48583782733323616, | |
| "grad_norm": 50792.91015625, | |
| "learning_rate": 2.570810863333819e-05, | |
| "loss": 0.3838, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.4906962056065685, | |
| "grad_norm": 44864.3046875, | |
| "learning_rate": 2.5465189719671574e-05, | |
| "loss": 0.3924, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.49555458387990087, | |
| "grad_norm": 50477.6015625, | |
| "learning_rate": 2.5222270806004954e-05, | |
| "loss": 0.3875, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.5004129621532333, | |
| "grad_norm": 46136.0859375, | |
| "learning_rate": 2.4979351892338337e-05, | |
| "loss": 0.3956, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.5052713404265656, | |
| "grad_norm": 57640.59765625, | |
| "learning_rate": 2.473643297867172e-05, | |
| "loss": 0.38, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.510129718699898, | |
| "grad_norm": 60448.453125, | |
| "learning_rate": 2.4493514065005103e-05, | |
| "loss": 0.3925, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.5149880969732303, | |
| "grad_norm": 43668.453125, | |
| "learning_rate": 2.4250595151338483e-05, | |
| "loss": 0.3966, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.5198464752465627, | |
| "grad_norm": 51975.83203125, | |
| "learning_rate": 2.4007676237671866e-05, | |
| "loss": 0.399, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.524704853519895, | |
| "grad_norm": 60277.078125, | |
| "learning_rate": 2.376475732400525e-05, | |
| "loss": 0.3968, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.5295632317932274, | |
| "grad_norm": 47762.875, | |
| "learning_rate": 2.3521838410338633e-05, | |
| "loss": 0.3937, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.5344216100665597, | |
| "grad_norm": 62721.125, | |
| "learning_rate": 2.327891949667201e-05, | |
| "loss": 0.3871, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.5392799883398921, | |
| "grad_norm": 41137.8515625, | |
| "learning_rate": 2.3036000583005392e-05, | |
| "loss": 0.3814, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.5441383666132245, | |
| "grad_norm": 50033.2734375, | |
| "learning_rate": 2.2793081669338776e-05, | |
| "loss": 0.3987, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.5489967448865569, | |
| "grad_norm": 50574.140625, | |
| "learning_rate": 2.255016275567216e-05, | |
| "loss": 0.3741, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.5538551231598893, | |
| "grad_norm": 46732.82421875, | |
| "learning_rate": 2.230724384200554e-05, | |
| "loss": 0.3775, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.5587135014332216, | |
| "grad_norm": 57961.8671875, | |
| "learning_rate": 2.2064324928338922e-05, | |
| "loss": 0.3826, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.563571879706554, | |
| "grad_norm": 43375.69140625, | |
| "learning_rate": 2.1821406014672302e-05, | |
| "loss": 0.3823, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.5684302579798863, | |
| "grad_norm": 49148.40234375, | |
| "learning_rate": 2.1578487101005685e-05, | |
| "loss": 0.3782, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.5732886362532187, | |
| "grad_norm": 60182.703125, | |
| "learning_rate": 2.133556818733907e-05, | |
| "loss": 0.3804, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.578147014526551, | |
| "grad_norm": 47590.46484375, | |
| "learning_rate": 2.1092649273672448e-05, | |
| "loss": 0.3974, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.5830053927998834, | |
| "grad_norm": 65701.859375, | |
| "learning_rate": 2.084973036000583e-05, | |
| "loss": 0.3798, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.5878637710732157, | |
| "grad_norm": 43476.69140625, | |
| "learning_rate": 2.0606811446339215e-05, | |
| "loss": 0.378, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.5927221493465481, | |
| "grad_norm": 43104.83203125, | |
| "learning_rate": 2.0363892532672595e-05, | |
| "loss": 0.3718, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.5975805276198805, | |
| "grad_norm": 58024.0546875, | |
| "learning_rate": 2.0120973619005974e-05, | |
| "loss": 0.3837, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.6024389058932128, | |
| "grad_norm": 56071.4765625, | |
| "learning_rate": 1.9878054705339358e-05, | |
| "loss": 0.3789, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.6072972841665452, | |
| "grad_norm": 56979.65625, | |
| "learning_rate": 1.963513579167274e-05, | |
| "loss": 0.3726, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.6121556624398776, | |
| "grad_norm": 48315.94140625, | |
| "learning_rate": 1.9392216878006124e-05, | |
| "loss": 0.3697, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.61701404071321, | |
| "grad_norm": 48739.546875, | |
| "learning_rate": 1.9149297964339504e-05, | |
| "loss": 0.3849, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.6218724189865423, | |
| "grad_norm": 49214.71875, | |
| "learning_rate": 1.8906379050672887e-05, | |
| "loss": 0.3646, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.6267307972598747, | |
| "grad_norm": 46501.2265625, | |
| "learning_rate": 1.866346013700627e-05, | |
| "loss": 0.3693, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.631589175533207, | |
| "grad_norm": 69907.9609375, | |
| "learning_rate": 1.842054122333965e-05, | |
| "loss": 0.3737, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.6364475538065394, | |
| "grad_norm": 47973.73046875, | |
| "learning_rate": 1.817762230967303e-05, | |
| "loss": 0.3875, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.6413059320798717, | |
| "grad_norm": 58501.1796875, | |
| "learning_rate": 1.7934703396006413e-05, | |
| "loss": 0.3778, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.6461643103532041, | |
| "grad_norm": 61768.96875, | |
| "learning_rate": 1.7691784482339797e-05, | |
| "loss": 0.384, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.6510226886265364, | |
| "grad_norm": 46427.80859375, | |
| "learning_rate": 1.744886556867318e-05, | |
| "loss": 0.365, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.6558810668998688, | |
| "grad_norm": 58254.9453125, | |
| "learning_rate": 1.720594665500656e-05, | |
| "loss": 0.3641, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.6607394451732012, | |
| "grad_norm": 49971.00390625, | |
| "learning_rate": 1.696302774133994e-05, | |
| "loss": 0.363, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.6655978234465335, | |
| "grad_norm": 56874.12890625, | |
| "learning_rate": 1.6720108827673323e-05, | |
| "loss": 0.3795, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.6704562017198659, | |
| "grad_norm": 51813.33984375, | |
| "learning_rate": 1.6477189914006706e-05, | |
| "loss": 0.3809, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.6753145799931982, | |
| "grad_norm": 51388.640625, | |
| "learning_rate": 1.6234271000340086e-05, | |
| "loss": 0.3649, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.6801729582665307, | |
| "grad_norm": 50291.19921875, | |
| "learning_rate": 1.599135208667347e-05, | |
| "loss": 0.3609, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.685031336539863, | |
| "grad_norm": 57660.2578125, | |
| "learning_rate": 1.5748433173006852e-05, | |
| "loss": 0.3719, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.6898897148131954, | |
| "grad_norm": 67932.59375, | |
| "learning_rate": 1.5505514259340236e-05, | |
| "loss": 0.3685, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.6947480930865277, | |
| "grad_norm": 54271.5234375, | |
| "learning_rate": 1.5262595345673615e-05, | |
| "loss": 0.376, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.6996064713598601, | |
| "grad_norm": 45938.90234375, | |
| "learning_rate": 1.5019676432006995e-05, | |
| "loss": 0.3555, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.7044648496331924, | |
| "grad_norm": 65276.1328125, | |
| "learning_rate": 1.4776757518340378e-05, | |
| "loss": 0.382, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.7093232279065248, | |
| "grad_norm": 52780.53125, | |
| "learning_rate": 1.453383860467376e-05, | |
| "loss": 0.3627, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.7141816061798572, | |
| "grad_norm": 66279.4375, | |
| "learning_rate": 1.4290919691007143e-05, | |
| "loss": 0.3726, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.7190399844531895, | |
| "grad_norm": 48593.3359375, | |
| "learning_rate": 1.4048000777340525e-05, | |
| "loss": 0.3623, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.7238983627265219, | |
| "grad_norm": 47858.83203125, | |
| "learning_rate": 1.3805081863673908e-05, | |
| "loss": 0.3605, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.7287567409998542, | |
| "grad_norm": 55828.08203125, | |
| "learning_rate": 1.3562162950007288e-05, | |
| "loss": 0.3559, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.7336151192731866, | |
| "grad_norm": 50614.34765625, | |
| "learning_rate": 1.331924403634067e-05, | |
| "loss": 0.3534, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.7384734975465189, | |
| "grad_norm": 48779.75, | |
| "learning_rate": 1.3076325122674053e-05, | |
| "loss": 0.3593, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.7433318758198514, | |
| "grad_norm": 50229.625, | |
| "learning_rate": 1.2833406209007434e-05, | |
| "loss": 0.3594, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.7481902540931837, | |
| "grad_norm": 47616.875, | |
| "learning_rate": 1.2590487295340816e-05, | |
| "loss": 0.3417, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.7530486323665161, | |
| "grad_norm": 51565.13671875, | |
| "learning_rate": 1.2347568381674197e-05, | |
| "loss": 0.3556, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.7579070106398484, | |
| "grad_norm": 159462.1875, | |
| "learning_rate": 1.210464946800758e-05, | |
| "loss": 0.3641, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.7627653889131808, | |
| "grad_norm": 54541.4375, | |
| "learning_rate": 1.1861730554340962e-05, | |
| "loss": 0.362, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.7676237671865132, | |
| "grad_norm": 60987.12109375, | |
| "learning_rate": 1.1618811640674344e-05, | |
| "loss": 0.3573, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.7724821454598455, | |
| "grad_norm": 51971.9765625, | |
| "learning_rate": 1.1375892727007725e-05, | |
| "loss": 0.3595, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.7773405237331779, | |
| "grad_norm": 53555.78515625, | |
| "learning_rate": 1.1132973813341108e-05, | |
| "loss": 0.3617, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.7821989020065102, | |
| "grad_norm": 50965.8984375, | |
| "learning_rate": 1.0890054899674488e-05, | |
| "loss": 0.356, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.7870572802798426, | |
| "grad_norm": 52491.86328125, | |
| "learning_rate": 1.0647135986007871e-05, | |
| "loss": 0.3442, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.7919156585531749, | |
| "grad_norm": 46180.28515625, | |
| "learning_rate": 1.0404217072341253e-05, | |
| "loss": 0.3618, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.7967740368265073, | |
| "grad_norm": 43963.23828125, | |
| "learning_rate": 1.0161298158674636e-05, | |
| "loss": 0.3451, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.8016324150998396, | |
| "grad_norm": 50149.4140625, | |
| "learning_rate": 9.918379245008016e-06, | |
| "loss": 0.3515, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.8064907933731721, | |
| "grad_norm": 64846.9609375, | |
| "learning_rate": 9.6754603313414e-06, | |
| "loss": 0.3593, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.8113491716465044, | |
| "grad_norm": 59240.46484375, | |
| "learning_rate": 9.432541417674781e-06, | |
| "loss": 0.345, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.8162075499198368, | |
| "grad_norm": 51653.4765625, | |
| "learning_rate": 9.189622504008162e-06, | |
| "loss": 0.3486, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.8210659281931691, | |
| "grad_norm": 56696.0703125, | |
| "learning_rate": 8.946703590341544e-06, | |
| "loss": 0.3514, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.8259243064665015, | |
| "grad_norm": 44659.8515625, | |
| "learning_rate": 8.703784676674927e-06, | |
| "loss": 0.3538, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.8307826847398339, | |
| "grad_norm": 62480.4453125, | |
| "learning_rate": 8.460865763008307e-06, | |
| "loss": 0.3553, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.8356410630131662, | |
| "grad_norm": 49702.1796875, | |
| "learning_rate": 8.21794684934169e-06, | |
| "loss": 0.355, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.8404994412864986, | |
| "grad_norm": 51918.6171875, | |
| "learning_rate": 7.975027935675072e-06, | |
| "loss": 0.3516, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.8453578195598309, | |
| "grad_norm": 70460.1875, | |
| "learning_rate": 7.732109022008453e-06, | |
| "loss": 0.3485, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.8502161978331633, | |
| "grad_norm": 39541.1328125, | |
| "learning_rate": 7.489190108341836e-06, | |
| "loss": 0.3682, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.8550745761064956, | |
| "grad_norm": 55487.01171875, | |
| "learning_rate": 7.246271194675218e-06, | |
| "loss": 0.3634, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.859932954379828, | |
| "grad_norm": 61263.55859375, | |
| "learning_rate": 7.0033522810086005e-06, | |
| "loss": 0.3661, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.8647913326531603, | |
| "grad_norm": 51811.1875, | |
| "learning_rate": 6.760433367341981e-06, | |
| "loss": 0.3545, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.8696497109264928, | |
| "grad_norm": 58568.05078125, | |
| "learning_rate": 6.517514453675364e-06, | |
| "loss": 0.3511, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.8745080891998251, | |
| "grad_norm": 42418.73828125, | |
| "learning_rate": 6.274595540008746e-06, | |
| "loss": 0.3462, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.8793664674731575, | |
| "grad_norm": 66980.625, | |
| "learning_rate": 6.0316766263421275e-06, | |
| "loss": 0.3429, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.8842248457464899, | |
| "grad_norm": 51797.6875, | |
| "learning_rate": 5.788757712675509e-06, | |
| "loss": 0.3457, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.8890832240198222, | |
| "grad_norm": 51049.6640625, | |
| "learning_rate": 5.5458387990088915e-06, | |
| "loss": 0.3462, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.8939416022931546, | |
| "grad_norm": 53216.921875, | |
| "learning_rate": 5.302919885342273e-06, | |
| "loss": 0.3482, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.8987999805664869, | |
| "grad_norm": 52944.14453125, | |
| "learning_rate": 5.0600009716756545e-06, | |
| "loss": 0.3429, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.9036583588398193, | |
| "grad_norm": 55617.85546875, | |
| "learning_rate": 4.817082058009037e-06, | |
| "loss": 0.3496, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.9085167371131516, | |
| "grad_norm": 51810.25390625, | |
| "learning_rate": 4.5741631443424185e-06, | |
| "loss": 0.3406, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.913375115386484, | |
| "grad_norm": 62037.01171875, | |
| "learning_rate": 4.331244230675801e-06, | |
| "loss": 0.3393, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.9182334936598163, | |
| "grad_norm": 55568.06640625, | |
| "learning_rate": 4.088325317009182e-06, | |
| "loss": 0.3412, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.9230918719331487, | |
| "grad_norm": 55022.734375, | |
| "learning_rate": 3.845406403342564e-06, | |
| "loss": 0.3297, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.927950250206481, | |
| "grad_norm": 59076.30078125, | |
| "learning_rate": 3.6024874896759463e-06, | |
| "loss": 0.3311, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.9328086284798134, | |
| "grad_norm": 55583.625, | |
| "learning_rate": 3.3595685760093283e-06, | |
| "loss": 0.3334, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.9376670067531458, | |
| "grad_norm": 44107.91015625, | |
| "learning_rate": 3.1166496623427103e-06, | |
| "loss": 0.3287, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.9425253850264782, | |
| "grad_norm": 43654.265625, | |
| "learning_rate": 2.8737307486760922e-06, | |
| "loss": 0.3501, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.9473837632998106, | |
| "grad_norm": 53803.21875, | |
| "learning_rate": 2.630811835009474e-06, | |
| "loss": 0.3418, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.9522421415731429, | |
| "grad_norm": 67202.1953125, | |
| "learning_rate": 2.3878929213428557e-06, | |
| "loss": 0.3343, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.9571005198464753, | |
| "grad_norm": 57509.7734375, | |
| "learning_rate": 2.1449740076762377e-06, | |
| "loss": 0.3397, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.9619588981198076, | |
| "grad_norm": 45354.30859375, | |
| "learning_rate": 1.9020550940096197e-06, | |
| "loss": 0.3467, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.96681727639314, | |
| "grad_norm": 51548.08984375, | |
| "learning_rate": 1.6591361803430016e-06, | |
| "loss": 0.3619, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.9716756546664723, | |
| "grad_norm": 65316.8515625, | |
| "learning_rate": 1.4162172666763834e-06, | |
| "loss": 0.3216, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.9765340329398047, | |
| "grad_norm": 48864.44921875, | |
| "learning_rate": 1.1732983530097653e-06, | |
| "loss": 0.3303, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.981392411213137, | |
| "grad_norm": 48878.0, | |
| "learning_rate": 9.303794393431472e-07, | |
| "loss": 0.3492, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.9862507894864694, | |
| "grad_norm": 43658.50390625, | |
| "learning_rate": 6.874605256765293e-07, | |
| "loss": 0.3306, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.9911091677598017, | |
| "grad_norm": 58298.01171875, | |
| "learning_rate": 4.4454161200991114e-07, | |
| "loss": 0.3366, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.9959675460331341, | |
| "grad_norm": 49216.18359375, | |
| "learning_rate": 2.01622698343293e-07, | |
| "loss": 0.3288, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 20583, | |
| "total_flos": 3.1397459461543035e+18, | |
| "train_loss": 0.4062042611626282, | |
| "train_runtime": 286226.4367, | |
| "train_samples_per_second": 0.863, | |
| "train_steps_per_second": 0.072 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 20583, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 20583, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.1397459461543035e+18, | |
| "train_batch_size": 12, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |