| { | |
| "best_metric": 0.07686587423086166, | |
| "best_model_checkpoint": "/teamspace/studios/this_studio/save/LLama_End/checkpoint-3500", | |
| "epoch": 1.9991833401388321, | |
| "eval_steps": 500, | |
| "global_step": 3672, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0027221995372260785, | |
| "grad_norm": 0.3611927926540375, | |
| "learning_rate": 5.000000000000001e-07, | |
| "loss": 0.7652, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.005444399074452157, | |
| "grad_norm": 0.3633479177951813, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 0.7798, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.008166598611678236, | |
| "grad_norm": 0.31540995836257935, | |
| "learning_rate": 1.5e-06, | |
| "loss": 0.7635, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.010888798148904314, | |
| "grad_norm": 0.3314383924007416, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 0.7475, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.013610997686130393, | |
| "grad_norm": 0.391197144985199, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.7827, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.016333197223356473, | |
| "grad_norm": 0.39311718940734863, | |
| "learning_rate": 3e-06, | |
| "loss": 0.7985, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01905539676058255, | |
| "grad_norm": 0.34564024209976196, | |
| "learning_rate": 3.5000000000000004e-06, | |
| "loss": 0.7608, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.021777596297808628, | |
| "grad_norm": 0.43378740549087524, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.7566, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02449979583503471, | |
| "grad_norm": 0.38521063327789307, | |
| "learning_rate": 4.5e-06, | |
| "loss": 0.75, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.027221995372260787, | |
| "grad_norm": 0.4021587371826172, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7492, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.029944194909486865, | |
| "grad_norm": 0.44062522053718567, | |
| "learning_rate": 5.500000000000001e-06, | |
| "loss": 0.7208, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.032666394446712946, | |
| "grad_norm": 0.37620142102241516, | |
| "learning_rate": 6e-06, | |
| "loss": 0.6911, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03538859398393902, | |
| "grad_norm": 0.27264639735221863, | |
| "learning_rate": 6.5000000000000004e-06, | |
| "loss": 0.6624, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.0381107935211651, | |
| "grad_norm": 0.2566782832145691, | |
| "learning_rate": 7.000000000000001e-06, | |
| "loss": 0.6326, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.04083299305839118, | |
| "grad_norm": 0.27515003085136414, | |
| "learning_rate": 7.5e-06, | |
| "loss": 0.6135, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.043555192595617256, | |
| "grad_norm": 0.2592063248157501, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 0.5585, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.046277392132843334, | |
| "grad_norm": 0.22615216672420502, | |
| "learning_rate": 8.500000000000002e-06, | |
| "loss": 0.5482, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.04899959167006942, | |
| "grad_norm": 0.21914538741111755, | |
| "learning_rate": 9e-06, | |
| "loss": 0.5488, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.051721791207295496, | |
| "grad_norm": 0.23039759695529938, | |
| "learning_rate": 9.5e-06, | |
| "loss": 0.4816, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.054443990744521574, | |
| "grad_norm": 0.21080121397972107, | |
| "learning_rate": 1e-05, | |
| "loss": 0.4686, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.05716619028174765, | |
| "grad_norm": 0.19633150100708008, | |
| "learning_rate": 1.05e-05, | |
| "loss": 0.4426, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.05988838981897373, | |
| "grad_norm": 0.20176054537296295, | |
| "learning_rate": 1.1000000000000001e-05, | |
| "loss": 0.4131, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.06261058935619981, | |
| "grad_norm": 0.1965443640947342, | |
| "learning_rate": 1.1500000000000002e-05, | |
| "loss": 0.394, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.06533278889342589, | |
| "grad_norm": 0.18852515518665314, | |
| "learning_rate": 1.2e-05, | |
| "loss": 0.3621, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.06805498843065197, | |
| "grad_norm": 0.1972607970237732, | |
| "learning_rate": 1.25e-05, | |
| "loss": 0.3436, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.07077718796787805, | |
| "grad_norm": 0.1883157640695572, | |
| "learning_rate": 1.3000000000000001e-05, | |
| "loss": 0.3235, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.07349938750510412, | |
| "grad_norm": 0.20938365161418915, | |
| "learning_rate": 1.3500000000000001e-05, | |
| "loss": 0.3093, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.0762215870423302, | |
| "grad_norm": 0.2136233150959015, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 0.281, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.07894378657955628, | |
| "grad_norm": 0.23373591899871826, | |
| "learning_rate": 1.45e-05, | |
| "loss": 0.2645, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.08166598611678236, | |
| "grad_norm": 0.21917635202407837, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.2399, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.08438818565400844, | |
| "grad_norm": 0.2619529068470001, | |
| "learning_rate": 1.55e-05, | |
| "loss": 0.2381, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.08711038519123451, | |
| "grad_norm": 0.2083514928817749, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 0.2139, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.08983258472846059, | |
| "grad_norm": 0.22957220673561096, | |
| "learning_rate": 1.65e-05, | |
| "loss": 0.1956, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.09255478426568667, | |
| "grad_norm": 0.270378977060318, | |
| "learning_rate": 1.7000000000000003e-05, | |
| "loss": 0.1906, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.09527698380291276, | |
| "grad_norm": 0.24123062193393707, | |
| "learning_rate": 1.75e-05, | |
| "loss": 0.1811, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.09799918334013884, | |
| "grad_norm": 0.2726229429244995, | |
| "learning_rate": 1.8e-05, | |
| "loss": 0.1756, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.10072138287736492, | |
| "grad_norm": 0.24878880381584167, | |
| "learning_rate": 1.85e-05, | |
| "loss": 0.1745, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.10344358241459099, | |
| "grad_norm": 0.22229672968387604, | |
| "learning_rate": 1.9e-05, | |
| "loss": 0.1641, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.10616578195181707, | |
| "grad_norm": 0.25277239084243774, | |
| "learning_rate": 1.9500000000000003e-05, | |
| "loss": 0.1478, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.10888798148904315, | |
| "grad_norm": 0.23793649673461914, | |
| "learning_rate": 2e-05, | |
| "loss": 0.1621, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.11161018102626923, | |
| "grad_norm": 0.22567716240882874, | |
| "learning_rate": 2.05e-05, | |
| "loss": 0.1416, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.1143323805634953, | |
| "grad_norm": 0.24615788459777832, | |
| "learning_rate": 2.1e-05, | |
| "loss": 0.1418, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.11705458010072138, | |
| "grad_norm": 0.23229427635669708, | |
| "learning_rate": 2.15e-05, | |
| "loss": 0.137, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.11977677963794746, | |
| "grad_norm": 0.2679113447666168, | |
| "learning_rate": 2.2000000000000003e-05, | |
| "loss": 0.1186, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.12249897917517354, | |
| "grad_norm": 0.24723300337791443, | |
| "learning_rate": 2.25e-05, | |
| "loss": 0.1243, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.12522117871239963, | |
| "grad_norm": 0.24604055285453796, | |
| "learning_rate": 2.3000000000000003e-05, | |
| "loss": 0.1472, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.1279433782496257, | |
| "grad_norm": 0.2207736223936081, | |
| "learning_rate": 2.35e-05, | |
| "loss": 0.1258, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.13066557778685178, | |
| "grad_norm": 0.2685292065143585, | |
| "learning_rate": 2.4e-05, | |
| "loss": 0.1407, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.13338777732407786, | |
| "grad_norm": 0.21564778685569763, | |
| "learning_rate": 2.45e-05, | |
| "loss": 0.1355, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.13610997686130394, | |
| "grad_norm": 0.22922654449939728, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.1298, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.13883217639853002, | |
| "grad_norm": 0.21827565133571625, | |
| "learning_rate": 2.5500000000000003e-05, | |
| "loss": 0.1256, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.1415543759357561, | |
| "grad_norm": 0.2129678726196289, | |
| "learning_rate": 2.6000000000000002e-05, | |
| "loss": 0.1435, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.14427657547298217, | |
| "grad_norm": 0.196010559797287, | |
| "learning_rate": 2.6500000000000004e-05, | |
| "loss": 0.1353, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.14699877501020825, | |
| "grad_norm": 0.22913923859596252, | |
| "learning_rate": 2.7000000000000002e-05, | |
| "loss": 0.1187, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.14972097454743433, | |
| "grad_norm": 0.26281264424324036, | |
| "learning_rate": 2.7500000000000004e-05, | |
| "loss": 0.1697, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.1524431740846604, | |
| "grad_norm": 0.28295812010765076, | |
| "learning_rate": 2.8000000000000003e-05, | |
| "loss": 0.1229, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.15516537362188648, | |
| "grad_norm": 0.23690921068191528, | |
| "learning_rate": 2.8499999999999998e-05, | |
| "loss": 0.1086, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.15788757315911256, | |
| "grad_norm": 0.22863948345184326, | |
| "learning_rate": 2.9e-05, | |
| "loss": 0.1064, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.16060977269633864, | |
| "grad_norm": 0.1989349126815796, | |
| "learning_rate": 2.95e-05, | |
| "loss": 0.1145, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.16333197223356472, | |
| "grad_norm": 0.2325228750705719, | |
| "learning_rate": 3e-05, | |
| "loss": 0.1181, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.1660541717707908, | |
| "grad_norm": 0.2611011266708374, | |
| "learning_rate": 3.05e-05, | |
| "loss": 0.1093, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.16877637130801687, | |
| "grad_norm": 0.23333916068077087, | |
| "learning_rate": 3.1e-05, | |
| "loss": 0.1119, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.17149857084524295, | |
| "grad_norm": 0.22769390046596527, | |
| "learning_rate": 3.15e-05, | |
| "loss": 0.1079, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.17422077038246903, | |
| "grad_norm": 0.23360049724578857, | |
| "learning_rate": 3.2000000000000005e-05, | |
| "loss": 0.1035, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.1769429699196951, | |
| "grad_norm": 0.2405271828174591, | |
| "learning_rate": 3.2500000000000004e-05, | |
| "loss": 0.1208, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.17966516945692118, | |
| "grad_norm": 0.2959926128387451, | |
| "learning_rate": 3.3e-05, | |
| "loss": 0.113, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.18238736899414726, | |
| "grad_norm": 0.2246614694595337, | |
| "learning_rate": 3.35e-05, | |
| "loss": 0.1082, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.18510956853137334, | |
| "grad_norm": 0.23407892882823944, | |
| "learning_rate": 3.4000000000000007e-05, | |
| "loss": 0.103, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.18783176806859944, | |
| "grad_norm": 0.21018989384174347, | |
| "learning_rate": 3.45e-05, | |
| "loss": 0.1084, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.19055396760582552, | |
| "grad_norm": 0.1964080035686493, | |
| "learning_rate": 3.5e-05, | |
| "loss": 0.1021, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.1932761671430516, | |
| "grad_norm": 0.2037973254919052, | |
| "learning_rate": 3.55e-05, | |
| "loss": 0.1088, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.19599836668027767, | |
| "grad_norm": 0.2156139761209488, | |
| "learning_rate": 3.6e-05, | |
| "loss": 0.1104, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.19872056621750375, | |
| "grad_norm": 0.21694843471050262, | |
| "learning_rate": 3.65e-05, | |
| "loss": 0.1073, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.20144276575472983, | |
| "grad_norm": 0.21144092082977295, | |
| "learning_rate": 3.7e-05, | |
| "loss": 0.1365, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.2041649652919559, | |
| "grad_norm": 0.24178194999694824, | |
| "learning_rate": 3.7500000000000003e-05, | |
| "loss": 0.0984, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.20688716482918199, | |
| "grad_norm": 0.20173633098602295, | |
| "learning_rate": 3.8e-05, | |
| "loss": 0.1289, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.20960936436640806, | |
| "grad_norm": 0.2253100723028183, | |
| "learning_rate": 3.85e-05, | |
| "loss": 0.1055, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.21233156390363414, | |
| "grad_norm": 0.2524998188018799, | |
| "learning_rate": 3.9000000000000006e-05, | |
| "loss": 0.1203, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.21505376344086022, | |
| "grad_norm": 0.21095526218414307, | |
| "learning_rate": 3.9500000000000005e-05, | |
| "loss": 0.0965, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.2177759629780863, | |
| "grad_norm": 0.1919548511505127, | |
| "learning_rate": 4e-05, | |
| "loss": 0.0961, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.22049816251531237, | |
| "grad_norm": 0.2088468074798584, | |
| "learning_rate": 4.05e-05, | |
| "loss": 0.0945, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.22322036205253845, | |
| "grad_norm": 0.22967375814914703, | |
| "learning_rate": 4.1e-05, | |
| "loss": 0.0991, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.22594256158976453, | |
| "grad_norm": 0.19900205731391907, | |
| "learning_rate": 4.15e-05, | |
| "loss": 0.1069, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.2286647611269906, | |
| "grad_norm": 0.22003936767578125, | |
| "learning_rate": 4.2e-05, | |
| "loss": 0.1037, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.23138696066421668, | |
| "grad_norm": 0.21360959112644196, | |
| "learning_rate": 4.25e-05, | |
| "loss": 0.0952, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.23410916020144276, | |
| "grad_norm": 0.1812208890914917, | |
| "learning_rate": 4.3e-05, | |
| "loss": 0.0962, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.23683135973866884, | |
| "grad_norm": 0.19043520092964172, | |
| "learning_rate": 4.35e-05, | |
| "loss": 0.1075, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.23955355927589492, | |
| "grad_norm": 0.2013886421918869, | |
| "learning_rate": 4.4000000000000006e-05, | |
| "loss": 0.104, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.242275758813121, | |
| "grad_norm": 0.1831788420677185, | |
| "learning_rate": 4.4500000000000004e-05, | |
| "loss": 0.1012, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.24499795835034707, | |
| "grad_norm": 0.1884724646806717, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.1115, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.24772015788757315, | |
| "grad_norm": 0.1994744837284088, | |
| "learning_rate": 4.55e-05, | |
| "loss": 0.1006, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.25044235742479926, | |
| "grad_norm": 0.1783529371023178, | |
| "learning_rate": 4.600000000000001e-05, | |
| "loss": 0.0902, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.25316455696202533, | |
| "grad_norm": 0.180820032954216, | |
| "learning_rate": 4.6500000000000005e-05, | |
| "loss": 0.0937, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.2558867564992514, | |
| "grad_norm": 0.19279232621192932, | |
| "learning_rate": 4.7e-05, | |
| "loss": 0.0944, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.2586089560364775, | |
| "grad_norm": 0.1725597232580185, | |
| "learning_rate": 4.75e-05, | |
| "loss": 0.104, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.26133115557370357, | |
| "grad_norm": 0.17777124047279358, | |
| "learning_rate": 4.8e-05, | |
| "loss": 0.0889, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.26405335511092964, | |
| "grad_norm": 0.18942312896251678, | |
| "learning_rate": 4.85e-05, | |
| "loss": 0.0964, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.2667755546481557, | |
| "grad_norm": 0.17807991802692413, | |
| "learning_rate": 4.9e-05, | |
| "loss": 0.1038, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.2694977541853818, | |
| "grad_norm": 0.21202729642391205, | |
| "learning_rate": 4.9500000000000004e-05, | |
| "loss": 0.0941, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.2722199537226079, | |
| "grad_norm": 0.18360304832458496, | |
| "learning_rate": 5e-05, | |
| "loss": 0.1013, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.2722199537226079, | |
| "eval_loss": 0.09772992134094238, | |
| "eval_runtime": 271.0226, | |
| "eval_samples_per_second": 2.214, | |
| "eval_steps_per_second": 0.553, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.27494215325983395, | |
| "grad_norm": 0.1730509251356125, | |
| "learning_rate": 4.999969346326857e-05, | |
| "loss": 0.098, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.27766435279706003, | |
| "grad_norm": 0.17720815539360046, | |
| "learning_rate": 4.9998773860591444e-05, | |
| "loss": 0.0912, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.2803865523342861, | |
| "grad_norm": 0.16912730038166046, | |
| "learning_rate": 4.9997241214519986e-05, | |
| "loss": 0.0885, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.2831087518715122, | |
| "grad_norm": 0.166640505194664, | |
| "learning_rate": 4.999509556263919e-05, | |
| "loss": 0.1013, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.28583095140873827, | |
| "grad_norm": 0.1893271952867508, | |
| "learning_rate": 4.999233695756673e-05, | |
| "loss": 0.1005, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.28855315094596434, | |
| "grad_norm": 0.1966632455587387, | |
| "learning_rate": 4.998896546695172e-05, | |
| "loss": 0.0958, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.2912753504831904, | |
| "grad_norm": 0.16383662819862366, | |
| "learning_rate": 4.998498117347302e-05, | |
| "loss": 0.0971, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.2939975500204165, | |
| "grad_norm": 0.21534255146980286, | |
| "learning_rate": 4.998038417483721e-05, | |
| "loss": 0.1057, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.2967197495576426, | |
| "grad_norm": 0.16162091493606567, | |
| "learning_rate": 4.9975174583776196e-05, | |
| "loss": 0.0943, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.29944194909486865, | |
| "grad_norm": 0.1837451457977295, | |
| "learning_rate": 4.996935252804448e-05, | |
| "loss": 0.0968, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.30216414863209473, | |
| "grad_norm": 0.15664127469062805, | |
| "learning_rate": 4.9962918150415946e-05, | |
| "loss": 0.0975, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.3048863481693208, | |
| "grad_norm": 0.16413035988807678, | |
| "learning_rate": 4.995587160868047e-05, | |
| "loss": 0.0925, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3076085477065469, | |
| "grad_norm": 0.16700997948646545, | |
| "learning_rate": 4.994821307563995e-05, | |
| "loss": 0.097, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.31033074724377296, | |
| "grad_norm": 0.16121503710746765, | |
| "learning_rate": 4.9939942739104105e-05, | |
| "loss": 0.0996, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.31305294678099904, | |
| "grad_norm": 0.16212740540504456, | |
| "learning_rate": 4.993106080188592e-05, | |
| "loss": 0.0888, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.3157751463182251, | |
| "grad_norm": 0.1662980020046234, | |
| "learning_rate": 4.9921567481796585e-05, | |
| "loss": 0.0994, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.3184973458554512, | |
| "grad_norm": 0.1845940202474594, | |
| "learning_rate": 4.9911463011640195e-05, | |
| "loss": 0.0976, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.3212195453926773, | |
| "grad_norm": 0.17079226672649384, | |
| "learning_rate": 4.9900747639208044e-05, | |
| "loss": 0.0947, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.32394174492990335, | |
| "grad_norm": 0.1642676740884781, | |
| "learning_rate": 4.9889421627272575e-05, | |
| "loss": 0.1039, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.32666394446712943, | |
| "grad_norm": 0.20938318967819214, | |
| "learning_rate": 4.987748525358087e-05, | |
| "loss": 0.1005, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.3293861440043555, | |
| "grad_norm": 0.17017363011837006, | |
| "learning_rate": 4.9864938810847884e-05, | |
| "loss": 0.0941, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.3321083435415816, | |
| "grad_norm": 0.17874999344348907, | |
| "learning_rate": 4.985178260674927e-05, | |
| "loss": 0.1055, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.33483054307880766, | |
| "grad_norm": 0.16853144764900208, | |
| "learning_rate": 4.98380169639138e-05, | |
| "loss": 0.0941, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.33755274261603374, | |
| "grad_norm": 0.18321168422698975, | |
| "learning_rate": 4.98236422199155e-05, | |
| "loss": 0.0991, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.3402749421532598, | |
| "grad_norm": 0.16565477848052979, | |
| "learning_rate": 4.980865872726532e-05, | |
| "loss": 0.0979, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.3429971416904859, | |
| "grad_norm": 0.19292525947093964, | |
| "learning_rate": 4.9793066853402536e-05, | |
| "loss": 0.0957, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.345719341227712, | |
| "grad_norm": 0.1700926125049591, | |
| "learning_rate": 4.977686698068572e-05, | |
| "loss": 0.0889, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.34844154076493805, | |
| "grad_norm": 0.15266568958759308, | |
| "learning_rate": 4.976005950638334e-05, | |
| "loss": 0.0928, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.35116374030216413, | |
| "grad_norm": 0.15712082386016846, | |
| "learning_rate": 4.974264484266406e-05, | |
| "loss": 0.0927, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.3538859398393902, | |
| "grad_norm": 0.15186965465545654, | |
| "learning_rate": 4.972462341658661e-05, | |
| "loss": 0.0944, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.3566081393766163, | |
| "grad_norm": 0.18069899082183838, | |
| "learning_rate": 4.970599567008931e-05, | |
| "loss": 0.0998, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.35933033891384236, | |
| "grad_norm": 0.14533384144306183, | |
| "learning_rate": 4.968676205997925e-05, | |
| "loss": 0.0981, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.36205253845106844, | |
| "grad_norm": 0.16364873945713043, | |
| "learning_rate": 4.966692305792106e-05, | |
| "loss": 0.0932, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.3647747379882945, | |
| "grad_norm": 0.18472006916999817, | |
| "learning_rate": 4.9646479150425376e-05, | |
| "loss": 0.0866, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.3674969375255206, | |
| "grad_norm": 0.15847539901733398, | |
| "learning_rate": 4.962543083883687e-05, | |
| "loss": 0.0837, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.3702191370627467, | |
| "grad_norm": 0.155991330742836, | |
| "learning_rate": 4.9603778639322004e-05, | |
| "loss": 0.0906, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.37294133659997275, | |
| "grad_norm": 0.16835445165634155, | |
| "learning_rate": 4.958152308285633e-05, | |
| "loss": 0.0949, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.3756635361371989, | |
| "grad_norm": 0.18460558354854584, | |
| "learning_rate": 4.95586647152115e-05, | |
| "loss": 0.0931, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.37838573567442496, | |
| "grad_norm": 0.17255567014217377, | |
| "learning_rate": 4.9535204096941854e-05, | |
| "loss": 0.087, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.38110793521165104, | |
| "grad_norm": 0.15889868140220642, | |
| "learning_rate": 4.951114180337069e-05, | |
| "loss": 0.0965, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.3838301347488771, | |
| "grad_norm": 0.1403263658285141, | |
| "learning_rate": 4.948647842457615e-05, | |
| "loss": 0.0862, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.3865523342861032, | |
| "grad_norm": 0.16774417459964752, | |
| "learning_rate": 4.946121456537676e-05, | |
| "loss": 0.0911, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.38927453382332927, | |
| "grad_norm": 0.16498145461082458, | |
| "learning_rate": 4.9435350845316575e-05, | |
| "loss": 0.0851, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.39199673336055535, | |
| "grad_norm": 0.15889093279838562, | |
| "learning_rate": 4.9408887898650036e-05, | |
| "loss": 0.0836, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.3947189328977814, | |
| "grad_norm": 0.14469870924949646, | |
| "learning_rate": 4.9381826374326336e-05, | |
| "loss": 0.1033, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.3974411324350075, | |
| "grad_norm": 0.1625497043132782, | |
| "learning_rate": 4.935416693597358e-05, | |
| "loss": 0.0862, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.4001633319722336, | |
| "grad_norm": 0.15590594708919525, | |
| "learning_rate": 4.932591026188247e-05, | |
| "loss": 0.0946, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.40288553150945966, | |
| "grad_norm": 0.1706922948360443, | |
| "learning_rate": 4.929705704498969e-05, | |
| "loss": 0.0913, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.40560773104668574, | |
| "grad_norm": 0.15704989433288574, | |
| "learning_rate": 4.9267607992860906e-05, | |
| "loss": 0.0888, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.4083299305839118, | |
| "grad_norm": 0.14272324740886688, | |
| "learning_rate": 4.9237563827673416e-05, | |
| "loss": 0.0845, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.4110521301211379, | |
| "grad_norm": 0.1592000275850296, | |
| "learning_rate": 4.9206925286198426e-05, | |
| "loss": 0.09, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.41377432965836397, | |
| "grad_norm": 0.16091205179691315, | |
| "learning_rate": 4.9175693119783013e-05, | |
| "loss": 0.0855, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.41649652919559005, | |
| "grad_norm": 0.13851270079612732, | |
| "learning_rate": 4.914386809433167e-05, | |
| "loss": 0.0946, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.4192187287328161, | |
| "grad_norm": 0.15238268673419952, | |
| "learning_rate": 4.911145099028753e-05, | |
| "loss": 0.0861, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.4219409282700422, | |
| "grad_norm": 0.15689148008823395, | |
| "learning_rate": 4.9078442602613265e-05, | |
| "loss": 0.088, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.4246631278072683, | |
| "grad_norm": 0.17214879393577576, | |
| "learning_rate": 4.9044843740771505e-05, | |
| "loss": 0.0882, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.42738532734449436, | |
| "grad_norm": 0.15415912866592407, | |
| "learning_rate": 4.901065522870511e-05, | |
| "loss": 0.0978, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.43010752688172044, | |
| "grad_norm": 0.165154829621315, | |
| "learning_rate": 4.897587790481683e-05, | |
| "loss": 0.0833, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.4328297264189465, | |
| "grad_norm": 0.15241703391075134, | |
| "learning_rate": 4.894051262194885e-05, | |
| "loss": 0.0944, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.4355519259561726, | |
| "grad_norm": 0.13514836132526398, | |
| "learning_rate": 4.8904560247361833e-05, | |
| "loss": 0.1036, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.43827412549339867, | |
| "grad_norm": 0.14687085151672363, | |
| "learning_rate": 4.886802166271364e-05, | |
| "loss": 0.1042, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.44099632503062475, | |
| "grad_norm": 0.143589586019516, | |
| "learning_rate": 4.8830897764037744e-05, | |
| "loss": 0.0771, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.4437185245678508, | |
| "grad_norm": 0.16582971811294556, | |
| "learning_rate": 4.879318946172123e-05, | |
| "loss": 0.0992, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.4464407241050769, | |
| "grad_norm": 0.1646450310945511, | |
| "learning_rate": 4.875489768048247e-05, | |
| "loss": 0.0959, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.449162923642303, | |
| "grad_norm": 0.12446028739213943, | |
| "learning_rate": 4.871602335934847e-05, | |
| "loss": 0.0817, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.45188512317952906, | |
| "grad_norm": 0.15374915301799774, | |
| "learning_rate": 4.867656745163182e-05, | |
| "loss": 0.0763, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.45460732271675514, | |
| "grad_norm": 0.1652667224407196, | |
| "learning_rate": 4.8636530924907296e-05, | |
| "loss": 0.1026, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.4573295222539812, | |
| "grad_norm": 0.13503149151802063, | |
| "learning_rate": 4.85959147609882e-05, | |
| "loss": 0.0986, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.4600517217912073, | |
| "grad_norm": 0.16977562010288239, | |
| "learning_rate": 4.855471995590222e-05, | |
| "loss": 0.1099, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.46277392132843337, | |
| "grad_norm": 0.14991120994091034, | |
| "learning_rate": 4.851294751986702e-05, | |
| "loss": 0.0839, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.46549612086565945, | |
| "grad_norm": 0.15737441182136536, | |
| "learning_rate": 4.84705984772655e-05, | |
| "loss": 0.097, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.4682183204028855, | |
| "grad_norm": 0.14336347579956055, | |
| "learning_rate": 4.8427673866620615e-05, | |
| "loss": 0.0837, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.4709405199401116, | |
| "grad_norm": 0.12957650423049927, | |
| "learning_rate": 4.8384174740569944e-05, | |
| "loss": 0.0871, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.4736627194773377, | |
| "grad_norm": 0.15204955637454987, | |
| "learning_rate": 4.83401021658399e-05, | |
| "loss": 0.0905, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.47638491901456376, | |
| "grad_norm": 0.1311759650707245, | |
| "learning_rate": 4.82954572232195e-05, | |
| "loss": 0.1187, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.47910711855178983, | |
| "grad_norm": 0.14466483891010284, | |
| "learning_rate": 4.825024100753395e-05, | |
| "loss": 0.0892, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.4818293180890159, | |
| "grad_norm": 0.14845815300941467, | |
| "learning_rate": 4.820445462761771e-05, | |
| "loss": 0.0865, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.484551517626242, | |
| "grad_norm": 0.15656766295433044, | |
| "learning_rate": 4.815809920628738e-05, | |
| "loss": 0.0899, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.48727371716346807, | |
| "grad_norm": 0.11957818269729614, | |
| "learning_rate": 4.8111175880314084e-05, | |
| "loss": 0.0767, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.48999591670069415, | |
| "grad_norm": 0.14076441526412964, | |
| "learning_rate": 4.806368580039566e-05, | |
| "loss": 0.0832, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.4927181162379202, | |
| "grad_norm": 0.15563461184501648, | |
| "learning_rate": 4.801563013112844e-05, | |
| "loss": 0.0873, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.4954403157751463, | |
| "grad_norm": 0.13865019381046295, | |
| "learning_rate": 4.7967010050978635e-05, | |
| "loss": 0.1083, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.4981625153123724, | |
| "grad_norm": 0.1487104445695877, | |
| "learning_rate": 4.791782675225348e-05, | |
| "loss": 0.0853, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.5008847148495985, | |
| "grad_norm": 0.14385932683944702, | |
| "learning_rate": 4.7868081441071975e-05, | |
| "loss": 0.0882, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.5036069143868246, | |
| "grad_norm": 0.144570991396904, | |
| "learning_rate": 4.781777533733534e-05, | |
| "loss": 0.0859, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.5063291139240507, | |
| "grad_norm": 0.1305367350578308, | |
| "learning_rate": 4.776690967469708e-05, | |
| "loss": 0.0788, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.5090513134612767, | |
| "grad_norm": 0.14406158030033112, | |
| "learning_rate": 4.771548570053268e-05, | |
| "loss": 0.075, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.5117735129985028, | |
| "grad_norm": 0.14757996797561646, | |
| "learning_rate": 4.766350467590911e-05, | |
| "loss": 0.0991, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.5144957125357289, | |
| "grad_norm": 0.13629554212093353, | |
| "learning_rate": 4.7610967875553846e-05, | |
| "loss": 0.082, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.517217912072955, | |
| "grad_norm": 0.1259639412164688, | |
| "learning_rate": 4.755787658782361e-05, | |
| "loss": 0.0848, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.519940111610181, | |
| "grad_norm": 0.14720089733600616, | |
| "learning_rate": 4.750423211467278e-05, | |
| "loss": 0.083, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.5226623111474071, | |
| "grad_norm": 0.13509656488895416, | |
| "learning_rate": 4.745003577162148e-05, | |
| "loss": 0.0856, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5253845106846332, | |
| "grad_norm": 0.15491896867752075, | |
| "learning_rate": 4.7395288887723296e-05, | |
| "loss": 0.0953, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.5281067102218593, | |
| "grad_norm": 0.14023731648921967, | |
| "learning_rate": 4.73399928055327e-05, | |
| "loss": 0.0887, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.5308289097590854, | |
| "grad_norm": 0.13751854002475739, | |
| "learning_rate": 4.728414888107211e-05, | |
| "loss": 0.0908, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.5335511092963114, | |
| "grad_norm": 0.16226938366889954, | |
| "learning_rate": 4.722775848379866e-05, | |
| "loss": 0.0898, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.5362733088335375, | |
| "grad_norm": 0.12918660044670105, | |
| "learning_rate": 4.717082299657058e-05, | |
| "loss": 0.0985, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.5389955083707636, | |
| "grad_norm": 0.14950741827487946, | |
| "learning_rate": 4.711334381561333e-05, | |
| "loss": 0.0894, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.5417177079079897, | |
| "grad_norm": 0.158179372549057, | |
| "learning_rate": 4.7055322350485344e-05, | |
| "loss": 0.0823, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.5444399074452158, | |
| "grad_norm": 0.14156126976013184, | |
| "learning_rate": 4.699676002404342e-05, | |
| "loss": 0.0851, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.5444399074452158, | |
| "eval_loss": 0.08648888021707535, | |
| "eval_runtime": 254.9955, | |
| "eval_samples_per_second": 2.353, | |
| "eval_steps_per_second": 0.588, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.5471621069824418, | |
| "grad_norm": 0.14035391807556152, | |
| "learning_rate": 4.6937658272407905e-05, | |
| "loss": 0.0827, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.5498843065196679, | |
| "grad_norm": 0.1346469521522522, | |
| "learning_rate": 4.6878018544927415e-05, | |
| "loss": 0.0986, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.552606506056894, | |
| "grad_norm": 0.16363918781280518, | |
| "learning_rate": 4.681784230414332e-05, | |
| "loss": 0.0883, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.5553287055941201, | |
| "grad_norm": 0.14541316032409668, | |
| "learning_rate": 4.6757131025753886e-05, | |
| "loss": 0.0881, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.5580509051313461, | |
| "grad_norm": 0.1540316641330719, | |
| "learning_rate": 4.6695886198578034e-05, | |
| "loss": 0.0825, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.5607731046685722, | |
| "grad_norm": 0.15897440910339355, | |
| "learning_rate": 4.6634109324518914e-05, | |
| "loss": 0.0978, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.5634953042057983, | |
| "grad_norm": 0.13335077464580536, | |
| "learning_rate": 4.657180191852701e-05, | |
| "loss": 0.1565, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.5662175037430244, | |
| "grad_norm": 0.1566026508808136, | |
| "learning_rate": 4.6508965508562995e-05, | |
| "loss": 0.0927, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.5689397032802505, | |
| "grad_norm": 0.1556750237941742, | |
| "learning_rate": 4.644560163556031e-05, | |
| "loss": 0.0839, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.5716619028174765, | |
| "grad_norm": 0.15351781249046326, | |
| "learning_rate": 4.638171185338729e-05, | |
| "loss": 0.0812, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.5743841023547026, | |
| "grad_norm": 0.16494281589984894, | |
| "learning_rate": 4.6317297728809147e-05, | |
| "loss": 0.0889, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.5771063018919287, | |
| "grad_norm": 0.15591537952423096, | |
| "learning_rate": 4.62523608414495e-05, | |
| "loss": 0.0858, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.5798285014291548, | |
| "grad_norm": 0.1396636664867401, | |
| "learning_rate": 4.618690278375164e-05, | |
| "loss": 0.0856, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.5825507009663808, | |
| "grad_norm": 0.1384498029947281, | |
| "learning_rate": 4.61209251609395e-05, | |
| "loss": 0.082, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.5852729005036069, | |
| "grad_norm": 0.13901008665561676, | |
| "learning_rate": 4.605442959097826e-05, | |
| "loss": 0.0821, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.587995100040833, | |
| "grad_norm": 0.1609126627445221, | |
| "learning_rate": 4.5987417704534697e-05, | |
| "loss": 0.1273, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.5907172995780591, | |
| "grad_norm": 0.1475275307893753, | |
| "learning_rate": 4.591989114493718e-05, | |
| "loss": 0.0807, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.5934394991152852, | |
| "grad_norm": 0.14419402182102203, | |
| "learning_rate": 4.5851851568135376e-05, | |
| "loss": 0.0737, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.5961616986525112, | |
| "grad_norm": 0.13583903014659882, | |
| "learning_rate": 4.5783300642659644e-05, | |
| "loss": 0.0904, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.5988838981897373, | |
| "grad_norm": 0.1405334770679474, | |
| "learning_rate": 4.571424004958012e-05, | |
| "loss": 0.0863, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.6016060977269634, | |
| "grad_norm": 0.1326403170824051, | |
| "learning_rate": 4.564467148246548e-05, | |
| "loss": 0.0839, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.6043282972641895, | |
| "grad_norm": 0.14316369593143463, | |
| "learning_rate": 4.557459664734141e-05, | |
| "loss": 0.0862, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.6070504968014155, | |
| "grad_norm": 0.14497336745262146, | |
| "learning_rate": 4.550401726264879e-05, | |
| "loss": 0.0733, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.6097726963386416, | |
| "grad_norm": 0.15112735331058502, | |
| "learning_rate": 4.5432935059201544e-05, | |
| "loss": 0.0712, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.6124948958758677, | |
| "grad_norm": 0.1457735002040863, | |
| "learning_rate": 4.536135178014415e-05, | |
| "loss": 0.0808, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.6152170954130938, | |
| "grad_norm": 0.14395588636398315, | |
| "learning_rate": 4.528926918090898e-05, | |
| "loss": 0.0863, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.6179392949503199, | |
| "grad_norm": 0.13603851199150085, | |
| "learning_rate": 4.521668902917317e-05, | |
| "loss": 0.0848, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.6206614944875459, | |
| "grad_norm": 0.16027645766735077, | |
| "learning_rate": 4.514361310481533e-05, | |
| "loss": 0.0929, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.623383694024772, | |
| "grad_norm": 0.13826999068260193, | |
| "learning_rate": 4.507004319987185e-05, | |
| "loss": 0.0914, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.6261058935619981, | |
| "grad_norm": 0.14140823483467102, | |
| "learning_rate": 4.499598111849299e-05, | |
| "loss": 0.0885, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.6288280930992242, | |
| "grad_norm": 0.13926826417446136, | |
| "learning_rate": 4.492142867689861e-05, | |
| "loss": 0.083, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.6315502926364502, | |
| "grad_norm": 0.148858904838562, | |
| "learning_rate": 4.484638770333367e-05, | |
| "loss": 0.0866, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.6342724921736763, | |
| "grad_norm": 0.12866345047950745, | |
| "learning_rate": 4.4770860038023335e-05, | |
| "loss": 0.0991, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.6369946917109024, | |
| "grad_norm": 0.13434378802776337, | |
| "learning_rate": 4.4694847533127903e-05, | |
| "loss": 0.079, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.6397168912481285, | |
| "grad_norm": 0.13786840438842773, | |
| "learning_rate": 4.461835205269736e-05, | |
| "loss": 0.0842, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.6424390907853545, | |
| "grad_norm": 0.12938237190246582, | |
| "learning_rate": 4.454137547262566e-05, | |
| "loss": 0.0771, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.13730603456497192, | |
| "learning_rate": 4.446391968060475e-05, | |
| "loss": 0.0831, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.6478834898598067, | |
| "grad_norm": 0.142304465174675, | |
| "learning_rate": 4.4385986576078254e-05, | |
| "loss": 0.0918, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.6506056893970328, | |
| "grad_norm": 0.1353188157081604, | |
| "learning_rate": 4.43075780701949e-05, | |
| "loss": 0.0897, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.6533278889342589, | |
| "grad_norm": 0.1252220869064331, | |
| "learning_rate": 4.422869608576167e-05, | |
| "loss": 0.1353, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.6560500884714849, | |
| "grad_norm": 0.14703655242919922, | |
| "learning_rate": 4.4149342557196605e-05, | |
| "loss": 0.2083, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.658772288008711, | |
| "grad_norm": 0.13858124613761902, | |
| "learning_rate": 4.406951943048141e-05, | |
| "loss": 0.0824, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.6614944875459371, | |
| "grad_norm": 0.1428638994693756, | |
| "learning_rate": 4.3989228663113714e-05, | |
| "loss": 0.0774, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.6642166870831632, | |
| "grad_norm": 0.13073915243148804, | |
| "learning_rate": 4.3908472224059064e-05, | |
| "loss": 0.0788, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.6669388866203892, | |
| "grad_norm": 0.14101268351078033, | |
| "learning_rate": 4.3827252093702656e-05, | |
| "loss": 0.0882, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.6696610861576153, | |
| "grad_norm": 0.13420827686786652, | |
| "learning_rate": 4.374557026380075e-05, | |
| "loss": 0.0783, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.6723832856948414, | |
| "grad_norm": 0.12414020299911499, | |
| "learning_rate": 4.366342873743185e-05, | |
| "loss": 0.0874, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.6751054852320675, | |
| "grad_norm": 0.12837223708629608, | |
| "learning_rate": 4.358082952894753e-05, | |
| "loss": 0.0863, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.6778276847692936, | |
| "grad_norm": 0.14144425094127655, | |
| "learning_rate": 4.349777466392313e-05, | |
| "loss": 0.0784, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.6805498843065196, | |
| "grad_norm": 0.13347384333610535, | |
| "learning_rate": 4.341426617910798e-05, | |
| "loss": 0.0823, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.6832720838437457, | |
| "grad_norm": 0.1294572502374649, | |
| "learning_rate": 4.3330306122375516e-05, | |
| "loss": 0.0787, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.6859942833809718, | |
| "grad_norm": 0.13768264651298523, | |
| "learning_rate": 4.324589655267306e-05, | |
| "loss": 0.0812, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.6887164829181979, | |
| "grad_norm": 0.14431719481945038, | |
| "learning_rate": 4.3161039539971295e-05, | |
| "loss": 0.0917, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.691438682455424, | |
| "grad_norm": 0.13428707420825958, | |
| "learning_rate": 4.307573716521353e-05, | |
| "loss": 0.0866, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.69416088199265, | |
| "grad_norm": 0.14333197474479675, | |
| "learning_rate": 4.298999152026465e-05, | |
| "loss": 0.0822, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.6968830815298761, | |
| "grad_norm": 0.13545657694339752, | |
| "learning_rate": 4.2903804707859835e-05, | |
| "loss": 0.1482, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.6996052810671022, | |
| "grad_norm": 0.13254009187221527, | |
| "learning_rate": 4.281717884155298e-05, | |
| "loss": 0.0792, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.7023274806043283, | |
| "grad_norm": 0.12849317491054535, | |
| "learning_rate": 4.2730116045664905e-05, | |
| "loss": 0.0909, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.7050496801415543, | |
| "grad_norm": 0.15678727626800537, | |
| "learning_rate": 4.264261845523116e-05, | |
| "loss": 0.0902, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.7077718796787804, | |
| "grad_norm": 0.1492096483707428, | |
| "learning_rate": 4.255468821594981e-05, | |
| "loss": 0.0796, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.7104940792160065, | |
| "grad_norm": 0.12939919531345367, | |
| "learning_rate": 4.2466327484128685e-05, | |
| "loss": 0.0929, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.7132162787532326, | |
| "grad_norm": 0.13585573434829712, | |
| "learning_rate": 4.2377538426632595e-05, | |
| "loss": 0.0997, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.7159384782904586, | |
| "grad_norm": 0.15193824470043182, | |
| "learning_rate": 4.228832322083013e-05, | |
| "loss": 0.0806, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.7186606778276847, | |
| "grad_norm": 0.1575016975402832, | |
| "learning_rate": 4.2198684054540285e-05, | |
| "loss": 0.0877, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.7213828773649108, | |
| "grad_norm": 0.12813854217529297, | |
| "learning_rate": 4.210862312597884e-05, | |
| "loss": 0.0802, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.7241050769021369, | |
| "grad_norm": 0.13729339838027954, | |
| "learning_rate": 4.201814264370441e-05, | |
| "loss": 0.0816, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.726827276439363, | |
| "grad_norm": 0.12651780247688293, | |
| "learning_rate": 4.192724482656428e-05, | |
| "loss": 0.0812, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.729549475976589, | |
| "grad_norm": 0.13558164238929749, | |
| "learning_rate": 4.1835931903640046e-05, | |
| "loss": 0.0926, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.7322716755138151, | |
| "grad_norm": 0.1335664838552475, | |
| "learning_rate": 4.17442061141929e-05, | |
| "loss": 0.078, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.7349938750510412, | |
| "grad_norm": 0.12852634489536285, | |
| "learning_rate": 4.165206970760874e-05, | |
| "loss": 0.0805, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.7377160745882673, | |
| "grad_norm": 0.14529520273208618, | |
| "learning_rate": 4.1559524943342985e-05, | |
| "loss": 0.0837, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.7404382741254933, | |
| "grad_norm": 0.14452920854091644, | |
| "learning_rate": 4.1466574090865225e-05, | |
| "loss": 0.0857, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.7431604736627194, | |
| "grad_norm": 0.13917720317840576, | |
| "learning_rate": 4.1373219429603473e-05, | |
| "loss": 0.0814, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.7458826731999455, | |
| "grad_norm": 0.11816851049661636, | |
| "learning_rate": 4.127946324888836e-05, | |
| "loss": 0.0743, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.7486048727371717, | |
| "grad_norm": 0.14422592520713806, | |
| "learning_rate": 4.118530784789694e-05, | |
| "loss": 0.0768, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.7513270722743978, | |
| "grad_norm": 0.14289607107639313, | |
| "learning_rate": 4.109075553559633e-05, | |
| "loss": 0.0806, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.7540492718116238, | |
| "grad_norm": 0.14432035386562347, | |
| "learning_rate": 4.099580863068706e-05, | |
| "loss": 0.0915, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.7567714713488499, | |
| "grad_norm": 0.16059063374996185, | |
| "learning_rate": 4.0900469461546235e-05, | |
| "loss": 0.0799, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.759493670886076, | |
| "grad_norm": 0.14347511529922485, | |
| "learning_rate": 4.0804740366170454e-05, | |
| "loss": 0.0899, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.7622158704233021, | |
| "grad_norm": 0.12717047333717346, | |
| "learning_rate": 4.070862369211843e-05, | |
| "loss": 0.0827, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.7649380699605282, | |
| "grad_norm": 0.12884029746055603, | |
| "learning_rate": 4.061212179645345e-05, | |
| "loss": 0.0856, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.7676602694977542, | |
| "grad_norm": 0.14441439509391785, | |
| "learning_rate": 4.051523704568557e-05, | |
| "loss": 0.1416, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.7703824690349803, | |
| "grad_norm": 0.12841658294200897, | |
| "learning_rate": 4.0417971815713584e-05, | |
| "loss": 0.0827, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.7731046685722064, | |
| "grad_norm": 0.14411351084709167, | |
| "learning_rate": 4.032032849176673e-05, | |
| "loss": 0.0808, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.7758268681094325, | |
| "grad_norm": 0.13570953905582428, | |
| "learning_rate": 4.022230946834624e-05, | |
| "loss": 0.08, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.7785490676466585, | |
| "grad_norm": 0.13861487805843353, | |
| "learning_rate": 4.012391714916661e-05, | |
| "loss": 0.0867, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.7812712671838846, | |
| "grad_norm": 0.14143118262290955, | |
| "learning_rate": 4.0025153947096624e-05, | |
| "loss": 0.0865, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.7839934667211107, | |
| "grad_norm": 0.1405879408121109, | |
| "learning_rate": 3.992602228410023e-05, | |
| "loss": 0.0983, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.7867156662583368, | |
| "grad_norm": 0.14435367286205292, | |
| "learning_rate": 3.982652459117707e-05, | |
| "loss": 0.0829, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.7894378657955629, | |
| "grad_norm": 0.14033889770507812, | |
| "learning_rate": 3.972666330830299e-05, | |
| "loss": 0.0903, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.7921600653327889, | |
| "grad_norm": 0.13285259902477264, | |
| "learning_rate": 3.9626440884370065e-05, | |
| "loss": 0.0748, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.794882264870015, | |
| "grad_norm": 0.1237478256225586, | |
| "learning_rate": 3.952585977712664e-05, | |
| "loss": 0.083, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.7976044644072411, | |
| "grad_norm": 0.14227712154388428, | |
| "learning_rate": 3.942492245311703e-05, | |
| "loss": 0.0926, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.8003266639444672, | |
| "grad_norm": 0.13259918987751007, | |
| "learning_rate": 3.9323631387621015e-05, | |
| "loss": 0.0841, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.8030488634816932, | |
| "grad_norm": 0.12796998023986816, | |
| "learning_rate": 3.9221989064593175e-05, | |
| "loss": 0.076, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.8057710630189193, | |
| "grad_norm": 0.13424307107925415, | |
| "learning_rate": 3.9119997976601954e-05, | |
| "loss": 0.097, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.8084932625561454, | |
| "grad_norm": 0.12660135328769684, | |
| "learning_rate": 3.9017660624768515e-05, | |
| "loss": 0.0834, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.8112154620933715, | |
| "grad_norm": 0.12413927167654037, | |
| "learning_rate": 3.891497951870545e-05, | |
| "loss": 0.0765, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.8139376616305976, | |
| "grad_norm": 0.1326495110988617, | |
| "learning_rate": 3.8811957176455214e-05, | |
| "loss": 0.0832, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.8166598611678236, | |
| "grad_norm": 0.133337140083313, | |
| "learning_rate": 3.870859612442837e-05, | |
| "loss": 0.0819, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.8166598611678236, | |
| "eval_loss": 0.08255165070295334, | |
| "eval_runtime": 238.131, | |
| "eval_samples_per_second": 2.52, | |
| "eval_steps_per_second": 0.63, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.8193820607050497, | |
| "grad_norm": 0.14871715009212494, | |
| "learning_rate": 3.8604898897341644e-05, | |
| "loss": 0.1878, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.8221042602422758, | |
| "grad_norm": 0.1360073983669281, | |
| "learning_rate": 3.850086803815576e-05, | |
| "loss": 0.0879, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.8248264597795019, | |
| "grad_norm": 0.12314160168170929, | |
| "learning_rate": 3.8396506098013076e-05, | |
| "loss": 0.0767, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.8275486593167279, | |
| "grad_norm": 0.14723584055900574, | |
| "learning_rate": 3.829181563617503e-05, | |
| "loss": 0.0972, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.830270858853954, | |
| "grad_norm": 0.14206384122371674, | |
| "learning_rate": 3.81867992199594e-05, | |
| "loss": 0.0839, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.8329930583911801, | |
| "grad_norm": 0.11804749816656113, | |
| "learning_rate": 3.808145942467729e-05, | |
| "loss": 0.0923, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.8357152579284062, | |
| "grad_norm": 0.12313738465309143, | |
| "learning_rate": 3.797579883357002e-05, | |
| "loss": 0.0872, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.8384374574656323, | |
| "grad_norm": 0.13978321850299835, | |
| "learning_rate": 3.7869820037745776e-05, | |
| "loss": 0.1243, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.8411596570028583, | |
| "grad_norm": 0.12310109287500381, | |
| "learning_rate": 3.776352563611604e-05, | |
| "loss": 0.0839, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.8438818565400844, | |
| "grad_norm": 0.13480614125728607, | |
| "learning_rate": 3.765691823533191e-05, | |
| "loss": 0.0873, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.8466040560773105, | |
| "grad_norm": 0.13653361797332764, | |
| "learning_rate": 3.75500004497201e-05, | |
| "loss": 0.0771, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.8493262556145366, | |
| "grad_norm": 0.1283886879682541, | |
| "learning_rate": 3.74427749012189e-05, | |
| "loss": 0.0816, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.8520484551517626, | |
| "grad_norm": 0.12764351069927216, | |
| "learning_rate": 3.733524421931385e-05, | |
| "loss": 0.0859, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.8547706546889887, | |
| "grad_norm": 0.1281007081270218, | |
| "learning_rate": 3.722741104097323e-05, | |
| "loss": 0.0806, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.8574928542262148, | |
| "grad_norm": 0.1311092972755432, | |
| "learning_rate": 3.711927801058347e-05, | |
| "loss": 0.0792, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.8602150537634409, | |
| "grad_norm": 0.11376714706420898, | |
| "learning_rate": 3.7010847779884204e-05, | |
| "loss": 0.0763, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.862937253300667, | |
| "grad_norm": 0.14167912304401398, | |
| "learning_rate": 3.690212300790333e-05, | |
| "loss": 0.0871, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.865659452837893, | |
| "grad_norm": 0.13373591005802155, | |
| "learning_rate": 3.679310636089174e-05, | |
| "loss": 0.1189, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.8683816523751191, | |
| "grad_norm": 0.14792831242084503, | |
| "learning_rate": 3.668380051225794e-05, | |
| "loss": 0.0811, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.8711038519123452, | |
| "grad_norm": 0.14200051128864288, | |
| "learning_rate": 3.657420814250258e-05, | |
| "loss": 0.086, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.8738260514495713, | |
| "grad_norm": 0.1501314491033554, | |
| "learning_rate": 3.646433193915257e-05, | |
| "loss": 0.0885, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.8765482509867973, | |
| "grad_norm": 0.10102449357509613, | |
| "learning_rate": 3.635417459669532e-05, | |
| "loss": 0.0738, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.8792704505240234, | |
| "grad_norm": 0.12374872714281082, | |
| "learning_rate": 3.624373881651254e-05, | |
| "loss": 0.0883, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.8819926500612495, | |
| "grad_norm": 0.1358879953622818, | |
| "learning_rate": 3.6133027306814085e-05, | |
| "loss": 0.0869, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.8847148495984756, | |
| "grad_norm": 0.12947793304920197, | |
| "learning_rate": 3.6022042782571494e-05, | |
| "loss": 0.0868, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.8874370491357016, | |
| "grad_norm": 0.14391177892684937, | |
| "learning_rate": 3.591078796545144e-05, | |
| "loss": 0.0829, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.8901592486729277, | |
| "grad_norm": 0.13038809597492218, | |
| "learning_rate": 3.579926558374897e-05, | |
| "loss": 0.0888, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.8928814482101538, | |
| "grad_norm": 0.15558017790317535, | |
| "learning_rate": 3.5687478372320576e-05, | |
| "loss": 0.0853, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.8956036477473799, | |
| "grad_norm": 0.1262005716562271, | |
| "learning_rate": 3.557542907251718e-05, | |
| "loss": 0.0763, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.898325847284606, | |
| "grad_norm": 0.1222100704908371, | |
| "learning_rate": 3.546312043211687e-05, | |
| "loss": 0.0835, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.901048046821832, | |
| "grad_norm": 0.14368890225887299, | |
| "learning_rate": 3.535055520525753e-05, | |
| "loss": 0.0795, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.9037702463590581, | |
| "grad_norm": 0.13927970826625824, | |
| "learning_rate": 3.52377361523693e-05, | |
| "loss": 0.0838, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.9064924458962842, | |
| "grad_norm": 0.14974816143512726, | |
| "learning_rate": 3.512466604010688e-05, | |
| "loss": 0.0834, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.9092146454335103, | |
| "grad_norm": 0.14425551891326904, | |
| "learning_rate": 3.501134764128167e-05, | |
| "loss": 0.0933, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.9119368449707363, | |
| "grad_norm": 0.13006910681724548, | |
| "learning_rate": 3.4897783734793794e-05, | |
| "loss": 0.0829, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.9146590445079624, | |
| "grad_norm": 0.10913355648517609, | |
| "learning_rate": 3.478397710556397e-05, | |
| "loss": 0.0775, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.9173812440451885, | |
| "grad_norm": 0.13194523751735687, | |
| "learning_rate": 3.466993054446515e-05, | |
| "loss": 0.0762, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.9201034435824146, | |
| "grad_norm": 0.1317387819290161, | |
| "learning_rate": 3.4555646848254136e-05, | |
| "loss": 0.094, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.9228256431196407, | |
| "grad_norm": 0.1169929951429367, | |
| "learning_rate": 3.4441128819503e-05, | |
| "loss": 0.0839, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.9255478426568667, | |
| "grad_norm": 0.13555140793323517, | |
| "learning_rate": 3.4326379266530314e-05, | |
| "loss": 0.0831, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.9282700421940928, | |
| "grad_norm": 0.1173190027475357, | |
| "learning_rate": 3.421140100333231e-05, | |
| "loss": 0.0842, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.9309922417313189, | |
| "grad_norm": 0.1393524408340454, | |
| "learning_rate": 3.409619684951386e-05, | |
| "loss": 0.1127, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.933714441268545, | |
| "grad_norm": 0.12765948474407196, | |
| "learning_rate": 3.3980769630219354e-05, | |
| "loss": 0.0781, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.936436640805771, | |
| "grad_norm": 0.11775423586368561, | |
| "learning_rate": 3.386512217606339e-05, | |
| "loss": 0.071, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.9391588403429971, | |
| "grad_norm": 0.1447712928056717, | |
| "learning_rate": 3.3749257323061376e-05, | |
| "loss": 0.0852, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.9418810398802232, | |
| "grad_norm": 0.14077867567539215, | |
| "learning_rate": 3.3633177912559984e-05, | |
| "loss": 0.0793, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.9446032394174493, | |
| "grad_norm": 0.13743150234222412, | |
| "learning_rate": 3.3516886791167444e-05, | |
| "loss": 0.0804, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.9473254389546754, | |
| "grad_norm": 0.13769537210464478, | |
| "learning_rate": 3.34003868106838e-05, | |
| "loss": 0.0865, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.9500476384919014, | |
| "grad_norm": 0.1354909986257553, | |
| "learning_rate": 3.328368082803088e-05, | |
| "loss": 0.0793, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.9527698380291275, | |
| "grad_norm": 0.12832538783550262, | |
| "learning_rate": 3.316677170518235e-05, | |
| "loss": 0.0769, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.9554920375663536, | |
| "grad_norm": 0.1272445023059845, | |
| "learning_rate": 3.304966230909342e-05, | |
| "loss": 0.0931, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.9582142371035797, | |
| "grad_norm": 0.14885401725769043, | |
| "learning_rate": 3.293235551163063e-05, | |
| "loss": 0.0835, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.9609364366408057, | |
| "grad_norm": 0.13052543997764587, | |
| "learning_rate": 3.2814854189501346e-05, | |
| "loss": 0.0866, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.9636586361780318, | |
| "grad_norm": 0.12114793807268143, | |
| "learning_rate": 3.269716122418326e-05, | |
| "loss": 0.0851, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.9663808357152579, | |
| "grad_norm": 0.13619346916675568, | |
| "learning_rate": 3.2579279501853744e-05, | |
| "loss": 0.0834, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.969103035252484, | |
| "grad_norm": 0.13156457245349884, | |
| "learning_rate": 3.246121191331902e-05, | |
| "loss": 0.0868, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.9718252347897101, | |
| "grad_norm": 0.1379031538963318, | |
| "learning_rate": 3.234296135394329e-05, | |
| "loss": 0.081, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.9745474343269361, | |
| "grad_norm": 0.1381215900182724, | |
| "learning_rate": 3.222453072357777e-05, | |
| "loss": 0.0758, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.9772696338641622, | |
| "grad_norm": 0.11675938963890076, | |
| "learning_rate": 3.2105922926489504e-05, | |
| "loss": 0.1002, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.9799918334013883, | |
| "grad_norm": 0.12285912781953812, | |
| "learning_rate": 3.1987140871290236e-05, | |
| "loss": 0.0736, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.9827140329386144, | |
| "grad_norm": 0.13735820353031158, | |
| "learning_rate": 3.1868187470864984e-05, | |
| "loss": 0.0779, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.9854362324758404, | |
| "grad_norm": 0.13905146718025208, | |
| "learning_rate": 3.1749065642300674e-05, | |
| "loss": 0.0833, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.9881584320130665, | |
| "grad_norm": 0.13085448741912842, | |
| "learning_rate": 3.1629778306814585e-05, | |
| "loss": 0.0859, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.9908806315502926, | |
| "grad_norm": 0.13941799104213715, | |
| "learning_rate": 3.151032838968271e-05, | |
| "loss": 0.0845, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.9936028310875187, | |
| "grad_norm": 0.12749053537845612, | |
| "learning_rate": 3.139071882016802e-05, | |
| "loss": 0.0935, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.9963250306247448, | |
| "grad_norm": 0.12027429789304733, | |
| "learning_rate": 3.127095253144864e-05, | |
| "loss": 0.077, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.9990472301619708, | |
| "grad_norm": 0.13250844180583954, | |
| "learning_rate": 3.11510324605459e-05, | |
| "loss": 0.0795, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 1.001769429699197, | |
| "grad_norm": 0.12925724685192108, | |
| "learning_rate": 3.103096154825233e-05, | |
| "loss": 0.1021, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.004491629236423, | |
| "grad_norm": 0.1351374089717865, | |
| "learning_rate": 3.0910742739059527e-05, | |
| "loss": 0.0833, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 1.0072138287736492, | |
| "grad_norm": 0.13750001788139343, | |
| "learning_rate": 3.0790378981085956e-05, | |
| "loss": 0.076, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.0099360283108751, | |
| "grad_norm": 0.1203581839799881, | |
| "learning_rate": 3.0669873226004655e-05, | |
| "loss": 0.0769, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 1.0126582278481013, | |
| "grad_norm": 0.12651003897190094, | |
| "learning_rate": 3.054922842897084e-05, | |
| "loss": 0.0729, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.0153804273853273, | |
| "grad_norm": 0.14429447054862976, | |
| "learning_rate": 3.0428447548549467e-05, | |
| "loss": 0.0772, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 1.0181026269225535, | |
| "grad_norm": 0.13154245913028717, | |
| "learning_rate": 3.030753354664262e-05, | |
| "loss": 0.1681, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.0208248264597795, | |
| "grad_norm": 0.11439166218042374, | |
| "learning_rate": 3.018648938841695e-05, | |
| "loss": 0.0753, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 1.0235470259970056, | |
| "grad_norm": 0.12332076579332352, | |
| "learning_rate": 3.00653180422309e-05, | |
| "loss": 0.0746, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.0262692255342316, | |
| "grad_norm": 0.12956227362155914, | |
| "learning_rate": 2.994402247956194e-05, | |
| "loss": 0.0733, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 1.0289914250714578, | |
| "grad_norm": 0.13542629778385162, | |
| "learning_rate": 2.9822605674933696e-05, | |
| "loss": 0.076, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.0317136246086838, | |
| "grad_norm": 0.14228489995002747, | |
| "learning_rate": 2.9701070605843e-05, | |
| "loss": 0.0756, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 1.03443582414591, | |
| "grad_norm": 0.13206136226654053, | |
| "learning_rate": 2.957942025268689e-05, | |
| "loss": 0.0733, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.037158023683136, | |
| "grad_norm": 0.14194877445697784, | |
| "learning_rate": 2.945765759868949e-05, | |
| "loss": 0.0837, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 1.039880223220362, | |
| "grad_norm": 0.1430080085992813, | |
| "learning_rate": 2.933578562982888e-05, | |
| "loss": 0.0814, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.042602422757588, | |
| "grad_norm": 0.12564821541309357, | |
| "learning_rate": 2.9213807334763854e-05, | |
| "loss": 0.0748, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 1.0453246222948143, | |
| "grad_norm": 0.12261082231998444, | |
| "learning_rate": 2.9091725704760638e-05, | |
| "loss": 0.0668, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.0480468218320402, | |
| "grad_norm": 0.14175981283187866, | |
| "learning_rate": 2.8969543733619554e-05, | |
| "loss": 0.0742, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 1.0507690213692664, | |
| "grad_norm": 0.1397016942501068, | |
| "learning_rate": 2.884726441760155e-05, | |
| "loss": 0.0716, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.0534912209064924, | |
| "grad_norm": 0.1348811835050583, | |
| "learning_rate": 2.8724890755354783e-05, | |
| "loss": 0.0882, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 1.0562134204437186, | |
| "grad_norm": 0.11297186464071274, | |
| "learning_rate": 2.8602425747841057e-05, | |
| "loss": 0.0715, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.0589356199809445, | |
| "grad_norm": 0.1416822224855423, | |
| "learning_rate": 2.8479872398262198e-05, | |
| "loss": 0.0805, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 1.0616578195181707, | |
| "grad_norm": 0.147059828042984, | |
| "learning_rate": 2.8357233711986487e-05, | |
| "loss": 0.0764, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.0643800190553967, | |
| "grad_norm": 0.137193962931633, | |
| "learning_rate": 2.8234512696474875e-05, | |
| "loss": 0.0733, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 1.0671022185926229, | |
| "grad_norm": 0.12843886017799377, | |
| "learning_rate": 2.8111712361207283e-05, | |
| "loss": 0.0751, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.0698244181298489, | |
| "grad_norm": 0.1351711004972458, | |
| "learning_rate": 2.7988835717608784e-05, | |
| "loss": 0.0776, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 1.072546617667075, | |
| "grad_norm": 0.1511557698249817, | |
| "learning_rate": 2.7865885778975743e-05, | |
| "loss": 0.0764, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.075268817204301, | |
| "grad_norm": 0.12794414162635803, | |
| "learning_rate": 2.774286556040196e-05, | |
| "loss": 0.0792, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 1.0779910167415272, | |
| "grad_norm": 0.13065066933631897, | |
| "learning_rate": 2.7619778078704685e-05, | |
| "loss": 0.0706, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.0807132162787532, | |
| "grad_norm": 0.13252925872802734, | |
| "learning_rate": 2.7496626352350664e-05, | |
| "loss": 0.0826, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 1.0834354158159794, | |
| "grad_norm": 0.1248321682214737, | |
| "learning_rate": 2.7373413401382104e-05, | |
| "loss": 0.076, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.0861576153532053, | |
| "grad_norm": 0.14091134071350098, | |
| "learning_rate": 2.7250142247342637e-05, | |
| "loss": 0.1298, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 1.0888798148904315, | |
| "grad_norm": 0.13955925405025482, | |
| "learning_rate": 2.7126815913203178e-05, | |
| "loss": 0.0753, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.0888798148904315, | |
| "eval_loss": 0.08035612851381302, | |
| "eval_runtime": 228.5599, | |
| "eval_samples_per_second": 2.625, | |
| "eval_steps_per_second": 0.656, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.0916020144276575, | |
| "grad_norm": 0.1300753802061081, | |
| "learning_rate": 2.7003437423287857e-05, | |
| "loss": 0.1036, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 1.0943242139648837, | |
| "grad_norm": 0.12375082820653915, | |
| "learning_rate": 2.688000980319979e-05, | |
| "loss": 0.0744, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.0970464135021096, | |
| "grad_norm": 0.14749085903167725, | |
| "learning_rate": 2.6756536079746907e-05, | |
| "loss": 0.0738, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 1.0997686130393358, | |
| "grad_norm": 0.13898104429244995, | |
| "learning_rate": 2.663301928086774e-05, | |
| "loss": 0.0701, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.1024908125765618, | |
| "grad_norm": 0.12244424223899841, | |
| "learning_rate": 2.6509462435557152e-05, | |
| "loss": 0.0707, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 1.105213012113788, | |
| "grad_norm": 0.13717930018901825, | |
| "learning_rate": 2.6385868573792072e-05, | |
| "loss": 0.07, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.107935211651014, | |
| "grad_norm": 0.16028448939323425, | |
| "learning_rate": 2.6262240726457167e-05, | |
| "loss": 0.0815, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 1.1106574111882401, | |
| "grad_norm": 0.13453496992588043, | |
| "learning_rate": 2.6138581925270533e-05, | |
| "loss": 0.0802, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.113379610725466, | |
| "grad_norm": 0.14346693456172943, | |
| "learning_rate": 2.6014895202709354e-05, | |
| "loss": 0.0777, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 1.1161018102626923, | |
| "grad_norm": 0.12791860103607178, | |
| "learning_rate": 2.5891183591935515e-05, | |
| "loss": 0.0798, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.1188240097999183, | |
| "grad_norm": 0.1344967633485794, | |
| "learning_rate": 2.5767450126721255e-05, | |
| "loss": 0.0679, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 1.1215462093371444, | |
| "grad_norm": 0.1335950344800949, | |
| "learning_rate": 2.564369784137472e-05, | |
| "loss": 0.0742, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.1242684088743704, | |
| "grad_norm": 0.12762115895748138, | |
| "learning_rate": 2.5519929770665597e-05, | |
| "loss": 0.0885, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 1.1269906084115966, | |
| "grad_norm": 0.1343313604593277, | |
| "learning_rate": 2.539614894975067e-05, | |
| "loss": 0.0672, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.1297128079488226, | |
| "grad_norm": 0.13929401338100433, | |
| "learning_rate": 2.5272358414099413e-05, | |
| "loss": 0.0765, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 1.1324350074860488, | |
| "grad_norm": 0.1336519867181778, | |
| "learning_rate": 2.5148561199419516e-05, | |
| "loss": 0.0728, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.1351572070232747, | |
| "grad_norm": 0.13433049619197845, | |
| "learning_rate": 2.5024760341582453e-05, | |
| "loss": 0.0785, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 1.137879406560501, | |
| "grad_norm": 0.12692618370056152, | |
| "learning_rate": 2.490095887654906e-05, | |
| "loss": 0.0971, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.1406016060977269, | |
| "grad_norm": 0.1449315845966339, | |
| "learning_rate": 2.4777159840295046e-05, | |
| "loss": 0.0669, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 1.143323805634953, | |
| "grad_norm": 0.14388057589530945, | |
| "learning_rate": 2.4653366268736565e-05, | |
| "loss": 0.0776, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.146046005172179, | |
| "grad_norm": 0.12486547976732254, | |
| "learning_rate": 2.452958119765577e-05, | |
| "loss": 0.0782, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 1.1487682047094052, | |
| "grad_norm": 0.135404571890831, | |
| "learning_rate": 2.4405807662626338e-05, | |
| "loss": 0.0792, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.1514904042466312, | |
| "grad_norm": 0.14270424842834473, | |
| "learning_rate": 2.4282048698939062e-05, | |
| "loss": 0.0729, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 1.1542126037838574, | |
| "grad_norm": 0.13348889350891113, | |
| "learning_rate": 2.41583073415274e-05, | |
| "loss": 0.0842, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.1569348033210836, | |
| "grad_norm": 0.15408015251159668, | |
| "learning_rate": 2.403458662489304e-05, | |
| "loss": 0.0903, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 1.1596570028583095, | |
| "grad_norm": 0.12234120815992355, | |
| "learning_rate": 2.3910889583031533e-05, | |
| "loss": 0.0793, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.1623792023955355, | |
| "grad_norm": 0.15008710324764252, | |
| "learning_rate": 2.3787219249357803e-05, | |
| "loss": 0.0759, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 1.1651014019327617, | |
| "grad_norm": 0.133488729596138, | |
| "learning_rate": 2.3663578656631858e-05, | |
| "loss": 0.0783, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.1678236014699879, | |
| "grad_norm": 0.12147729843854904, | |
| "learning_rate": 2.3539970836884347e-05, | |
| "loss": 0.0816, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 1.1705458010072138, | |
| "grad_norm": 0.14398206770420074, | |
| "learning_rate": 2.3416398821342238e-05, | |
| "loss": 0.0739, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.1732680005444398, | |
| "grad_norm": 0.14384213089942932, | |
| "learning_rate": 2.329286564035446e-05, | |
| "loss": 0.0764, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 1.175990200081666, | |
| "grad_norm": 0.1491377055644989, | |
| "learning_rate": 2.3169374323317624e-05, | |
| "loss": 0.0716, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.1787123996188922, | |
| "grad_norm": 0.13119584321975708, | |
| "learning_rate": 2.3045927898601702e-05, | |
| "loss": 0.0737, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 1.1814345991561181, | |
| "grad_norm": 0.13395731151103973, | |
| "learning_rate": 2.292252939347577e-05, | |
| "loss": 0.0674, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.1841567986933441, | |
| "grad_norm": 0.1500498354434967, | |
| "learning_rate": 2.27991818340338e-05, | |
| "loss": 0.0872, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 1.1868789982305703, | |
| "grad_norm": 0.15766474604606628, | |
| "learning_rate": 2.2675888245120382e-05, | |
| "loss": 0.1069, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.1896011977677965, | |
| "grad_norm": 0.1266118288040161, | |
| "learning_rate": 2.255265165025663e-05, | |
| "loss": 0.0753, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 1.1923233973050225, | |
| "grad_norm": 0.12611445784568787, | |
| "learning_rate": 2.2429475071565987e-05, | |
| "loss": 0.0704, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.1950455968422484, | |
| "grad_norm": 0.1382066011428833, | |
| "learning_rate": 2.2306361529700125e-05, | |
| "loss": 0.081, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 1.1977677963794746, | |
| "grad_norm": 0.13451717793941498, | |
| "learning_rate": 2.218331404376484e-05, | |
| "loss": 0.0813, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.2004899959167008, | |
| "grad_norm": 0.1681748777627945, | |
| "learning_rate": 2.2060335631246075e-05, | |
| "loss": 0.0793, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 1.2032121954539268, | |
| "grad_norm": 0.14408931136131287, | |
| "learning_rate": 2.1937429307935887e-05, | |
| "loss": 0.0763, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.2059343949911527, | |
| "grad_norm": 0.14158952236175537, | |
| "learning_rate": 2.1814598087858476e-05, | |
| "loss": 0.0882, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 1.208656594528379, | |
| "grad_norm": 0.12508870661258698, | |
| "learning_rate": 2.16918449831963e-05, | |
| "loss": 0.0762, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.2113787940656051, | |
| "grad_norm": 0.13746081292629242, | |
| "learning_rate": 2.15691730042162e-05, | |
| "loss": 0.0802, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 1.214100993602831, | |
| "grad_norm": 0.12565705180168152, | |
| "learning_rate": 2.144658515919557e-05, | |
| "loss": 0.071, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.216823193140057, | |
| "grad_norm": 0.13257178664207458, | |
| "learning_rate": 2.1324084454348592e-05, | |
| "loss": 0.069, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 1.2195453926772832, | |
| "grad_norm": 0.15064120292663574, | |
| "learning_rate": 2.1201673893752534e-05, | |
| "loss": 0.0793, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.2222675922145094, | |
| "grad_norm": 0.1452379673719406, | |
| "learning_rate": 2.107935647927404e-05, | |
| "loss": 0.0653, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 1.2249897917517354, | |
| "grad_norm": 0.16073830425739288, | |
| "learning_rate": 2.095713521049554e-05, | |
| "loss": 0.0726, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.2277119912889614, | |
| "grad_norm": 0.15698719024658203, | |
| "learning_rate": 2.0835013084641704e-05, | |
| "loss": 0.0781, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 1.2304341908261875, | |
| "grad_norm": 0.13535834848880768, | |
| "learning_rate": 2.07129930965059e-05, | |
| "loss": 0.0754, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.2331563903634137, | |
| "grad_norm": 0.13440349698066711, | |
| "learning_rate": 2.0591078238376803e-05, | |
| "loss": 0.1275, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 1.2358785899006397, | |
| "grad_norm": 0.15323837101459503, | |
| "learning_rate": 2.0469271499964993e-05, | |
| "loss": 0.0738, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 1.238600789437866, | |
| "grad_norm": 0.14659421145915985, | |
| "learning_rate": 2.034757586832961e-05, | |
| "loss": 0.0719, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 1.2413229889750919, | |
| "grad_norm": 0.14592771232128143, | |
| "learning_rate": 2.022599432780515e-05, | |
| "loss": 0.0793, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.244045188512318, | |
| "grad_norm": 0.12547121942043304, | |
| "learning_rate": 2.0104529859928254e-05, | |
| "loss": 0.0754, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 1.246767388049544, | |
| "grad_norm": 0.13453277945518494, | |
| "learning_rate": 1.9983185443364615e-05, | |
| "loss": 0.0679, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 1.2494895875867702, | |
| "grad_norm": 0.13268929719924927, | |
| "learning_rate": 1.9861964053835885e-05, | |
| "loss": 0.0778, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 1.2522117871239962, | |
| "grad_norm": 0.14037257432937622, | |
| "learning_rate": 1.974086866404675e-05, | |
| "loss": 0.0874, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.2549339866612224, | |
| "grad_norm": 0.1446313112974167, | |
| "learning_rate": 1.961990224361201e-05, | |
| "loss": 0.0708, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 1.2576561861984483, | |
| "grad_norm": 0.15174619853496552, | |
| "learning_rate": 1.949906775898375e-05, | |
| "loss": 0.0841, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 1.2603783857356743, | |
| "grad_norm": 0.14297954738140106, | |
| "learning_rate": 1.9378368173378618e-05, | |
| "loss": 0.0802, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 1.2631005852729005, | |
| "grad_norm": 0.1344052255153656, | |
| "learning_rate": 1.9257806446705116e-05, | |
| "loss": 0.0675, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 1.2658227848101267, | |
| "grad_norm": 0.14959311485290527, | |
| "learning_rate": 1.913738553549106e-05, | |
| "loss": 0.0848, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 1.2685449843473526, | |
| "grad_norm": 0.12404653429985046, | |
| "learning_rate": 1.9017108392811065e-05, | |
| "loss": 0.0687, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 1.2712671838845788, | |
| "grad_norm": 0.13011127710342407, | |
| "learning_rate": 1.8896977968214078e-05, | |
| "loss": 0.113, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 1.2739893834218048, | |
| "grad_norm": 0.14289557933807373, | |
| "learning_rate": 1.877699720765114e-05, | |
| "loss": 0.0771, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 1.276711582959031, | |
| "grad_norm": 0.14011026918888092, | |
| "learning_rate": 1.8657169053403052e-05, | |
| "loss": 0.0771, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 1.279433782496257, | |
| "grad_norm": 0.1479315608739853, | |
| "learning_rate": 1.8537496444008283e-05, | |
| "loss": 0.07, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 1.2821559820334831, | |
| "grad_norm": 0.1444329172372818, | |
| "learning_rate": 1.841798231419087e-05, | |
| "loss": 0.0815, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 1.284878181570709, | |
| "grad_norm": 0.11986621469259262, | |
| "learning_rate": 1.8298629594788467e-05, | |
| "loss": 0.0689, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 1.2876003811079353, | |
| "grad_norm": 0.13101747632026672, | |
| "learning_rate": 1.817944121268048e-05, | |
| "loss": 0.0728, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 1.2903225806451613, | |
| "grad_norm": 0.1301647424697876, | |
| "learning_rate": 1.8060420090716265e-05, | |
| "loss": 0.0789, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 1.2930447801823874, | |
| "grad_norm": 0.14230471849441528, | |
| "learning_rate": 1.794156914764349e-05, | |
| "loss": 0.0702, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 1.2957669797196134, | |
| "grad_norm": 0.15737561881542206, | |
| "learning_rate": 1.7822891298036515e-05, | |
| "loss": 0.0781, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 1.2984891792568396, | |
| "grad_norm": 0.16721278429031372, | |
| "learning_rate": 1.7704389452224944e-05, | |
| "loss": 0.0805, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 1.3012113787940656, | |
| "grad_norm": 0.1321476846933365, | |
| "learning_rate": 1.7586066516222276e-05, | |
| "loss": 0.0768, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 1.3039335783312918, | |
| "grad_norm": 0.15134398639202118, | |
| "learning_rate": 1.7467925391654585e-05, | |
| "loss": 0.0825, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 1.3066557778685177, | |
| "grad_norm": 0.14100222289562225, | |
| "learning_rate": 1.73499689756894e-05, | |
| "loss": 0.0823, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.309377977405744, | |
| "grad_norm": 0.1558840423822403, | |
| "learning_rate": 1.7232200160964657e-05, | |
| "loss": 0.0832, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 1.3121001769429699, | |
| "grad_norm": 0.14366985857486725, | |
| "learning_rate": 1.7114621835517773e-05, | |
| "loss": 0.0791, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 1.314822376480196, | |
| "grad_norm": 0.12937745451927185, | |
| "learning_rate": 1.699723688271477e-05, | |
| "loss": 0.0703, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 1.317544576017422, | |
| "grad_norm": 0.12883788347244263, | |
| "learning_rate": 1.6880048181179652e-05, | |
| "loss": 0.0797, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 1.3202667755546482, | |
| "grad_norm": 0.13123267889022827, | |
| "learning_rate": 1.6763058604723723e-05, | |
| "loss": 0.0727, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 1.3229889750918742, | |
| "grad_norm": 0.10518030822277069, | |
| "learning_rate": 1.6646271022275185e-05, | |
| "loss": 0.066, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 1.3257111746291004, | |
| "grad_norm": 0.13036532700061798, | |
| "learning_rate": 1.6529688297808726e-05, | |
| "loss": 0.078, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 1.3284333741663263, | |
| "grad_norm": 0.14793965220451355, | |
| "learning_rate": 1.6413313290275355e-05, | |
| "loss": 0.0779, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.3311555737035525, | |
| "grad_norm": 0.15071183443069458, | |
| "learning_rate": 1.629714885353221e-05, | |
| "loss": 0.0835, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 1.3338777732407785, | |
| "grad_norm": 0.13086527585983276, | |
| "learning_rate": 1.618119783627263e-05, | |
| "loss": 0.0759, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 1.3365999727780047, | |
| "grad_norm": 0.13066627085208893, | |
| "learning_rate": 1.6065463081956292e-05, | |
| "loss": 0.0745, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 1.3393221723152307, | |
| "grad_norm": 0.15003521740436554, | |
| "learning_rate": 1.5949947428739448e-05, | |
| "loss": 0.084, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.3420443718524568, | |
| "grad_norm": 0.12341570854187012, | |
| "learning_rate": 1.5834653709405368e-05, | |
| "loss": 0.0707, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 1.3447665713896828, | |
| "grad_norm": 0.1332896649837494, | |
| "learning_rate": 1.571958475129484e-05, | |
| "loss": 0.0701, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 1.347488770926909, | |
| "grad_norm": 0.12761184573173523, | |
| "learning_rate": 1.5604743376236847e-05, | |
| "loss": 0.0708, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 1.350210970464135, | |
| "grad_norm": 0.12873777747154236, | |
| "learning_rate": 1.549013240047937e-05, | |
| "loss": 0.0731, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.3529331700013612, | |
| "grad_norm": 0.1278139352798462, | |
| "learning_rate": 1.537575463462031e-05, | |
| "loss": 0.0712, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 1.3556553695385871, | |
| "grad_norm": 0.16153199970722198, | |
| "learning_rate": 1.526161288353861e-05, | |
| "loss": 0.0825, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 1.3583775690758133, | |
| "grad_norm": 0.13839781284332275, | |
| "learning_rate": 1.5147709946325395e-05, | |
| "loss": 0.0769, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 1.3610997686130393, | |
| "grad_norm": 0.1404309719800949, | |
| "learning_rate": 1.5034048616215402e-05, | |
| "loss": 0.0737, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.3610997686130393, | |
| "eval_loss": 0.07860012352466583, | |
| "eval_runtime": 248.5723, | |
| "eval_samples_per_second": 2.414, | |
| "eval_steps_per_second": 0.603, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.3638219681502655, | |
| "grad_norm": 0.14883238077163696, | |
| "learning_rate": 1.4920631680518432e-05, | |
| "loss": 0.0695, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 1.3665441676874914, | |
| "grad_norm": 0.14661747217178345, | |
| "learning_rate": 1.4807461920551028e-05, | |
| "loss": 0.0752, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 1.3692663672247176, | |
| "grad_norm": 0.13054049015045166, | |
| "learning_rate": 1.469454211156826e-05, | |
| "loss": 0.0698, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 1.3719885667619436, | |
| "grad_norm": 0.13900673389434814, | |
| "learning_rate": 1.4581875022695653e-05, | |
| "loss": 0.0897, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.3747107662991698, | |
| "grad_norm": 0.1521766632795334, | |
| "learning_rate": 1.4469463416861307e-05, | |
| "loss": 0.0751, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 1.3774329658363957, | |
| "grad_norm": 0.14469105005264282, | |
| "learning_rate": 1.4357310050728115e-05, | |
| "loss": 0.0807, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 1.380155165373622, | |
| "grad_norm": 0.132577583193779, | |
| "learning_rate": 1.4245417674626183e-05, | |
| "loss": 0.0693, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 1.382877364910848, | |
| "grad_norm": 0.12606105208396912, | |
| "learning_rate": 1.4133789032485367e-05, | |
| "loss": 0.0739, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.385599564448074, | |
| "grad_norm": 0.1424499899148941, | |
| "learning_rate": 1.4022426861767998e-05, | |
| "loss": 0.0806, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 1.3883217639853, | |
| "grad_norm": 0.15580779314041138, | |
| "learning_rate": 1.391133389340174e-05, | |
| "loss": 0.0737, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 1.3910439635225262, | |
| "grad_norm": 0.14672575891017914, | |
| "learning_rate": 1.3800512851712635e-05, | |
| "loss": 0.0746, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 1.3937661630597522, | |
| "grad_norm": 0.15802086889743805, | |
| "learning_rate": 1.3689966454358256e-05, | |
| "loss": 0.0768, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.3964883625969784, | |
| "grad_norm": 0.1357397735118866, | |
| "learning_rate": 1.3579697412261117e-05, | |
| "loss": 0.0724, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 1.3992105621342044, | |
| "grad_norm": 0.14769281446933746, | |
| "learning_rate": 1.3469708429542157e-05, | |
| "loss": 0.0738, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 1.4019327616714305, | |
| "grad_norm": 0.1663895845413208, | |
| "learning_rate": 1.3360002203454442e-05, | |
| "loss": 0.0739, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 1.4046549612086565, | |
| "grad_norm": 0.13682472705841064, | |
| "learning_rate": 1.325058142431701e-05, | |
| "loss": 0.0732, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.4073771607458827, | |
| "grad_norm": 0.1464831680059433, | |
| "learning_rate": 1.3141448775448875e-05, | |
| "loss": 0.0781, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 1.4100993602831087, | |
| "grad_norm": 0.14444297552108765, | |
| "learning_rate": 1.3032606933103305e-05, | |
| "loss": 0.0742, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 1.4128215598203349, | |
| "grad_norm": 0.14417661726474762, | |
| "learning_rate": 1.2924058566402098e-05, | |
| "loss": 0.0761, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 1.4155437593575608, | |
| "grad_norm": 0.12622784078121185, | |
| "learning_rate": 1.2815806337270186e-05, | |
| "loss": 0.0693, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.418265958894787, | |
| "grad_norm": 0.1415053904056549, | |
| "learning_rate": 1.270785290037031e-05, | |
| "loss": 0.0771, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 1.4209881584320132, | |
| "grad_norm": 0.14631399512290955, | |
| "learning_rate": 1.260020090303797e-05, | |
| "loss": 0.1018, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 1.4237103579692392, | |
| "grad_norm": 0.1438084840774536, | |
| "learning_rate": 1.2492852985216483e-05, | |
| "loss": 0.0708, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 1.4264325575064651, | |
| "grad_norm": 0.12973198294639587, | |
| "learning_rate": 1.2385811779392236e-05, | |
| "loss": 0.0795, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.4291547570436913, | |
| "grad_norm": 0.15162228047847748, | |
| "learning_rate": 1.2279079910530147e-05, | |
| "loss": 0.0811, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 1.4318769565809175, | |
| "grad_norm": 0.14159157872200012, | |
| "learning_rate": 1.2172659996009254e-05, | |
| "loss": 0.0762, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 1.4345991561181435, | |
| "grad_norm": 0.13606242835521698, | |
| "learning_rate": 1.2066554645558578e-05, | |
| "loss": 0.0739, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 1.4373213556553694, | |
| "grad_norm": 0.15046873688697815, | |
| "learning_rate": 1.1960766461193124e-05, | |
| "loss": 0.0663, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.4400435551925956, | |
| "grad_norm": 0.12745556235313416, | |
| "learning_rate": 1.1855298037150022e-05, | |
| "loss": 0.073, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 1.4427657547298218, | |
| "grad_norm": 0.156600683927536, | |
| "learning_rate": 1.1750151959824961e-05, | |
| "loss": 0.0746, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 1.4454879542670478, | |
| "grad_norm": 0.13598720729351044, | |
| "learning_rate": 1.1645330807708713e-05, | |
| "loss": 0.0687, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 1.4482101538042738, | |
| "grad_norm": 0.13264238834381104, | |
| "learning_rate": 1.1540837151323951e-05, | |
| "loss": 0.0814, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.4509323533415, | |
| "grad_norm": 0.13126635551452637, | |
| "learning_rate": 1.143667355316219e-05, | |
| "loss": 0.0768, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 1.4536545528787261, | |
| "grad_norm": 0.14652326703071594, | |
| "learning_rate": 1.1332842567620941e-05, | |
| "loss": 0.0731, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 1.456376752415952, | |
| "grad_norm": 0.14913715422153473, | |
| "learning_rate": 1.1229346740941088e-05, | |
| "loss": 0.0893, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 1.459098951953178, | |
| "grad_norm": 0.15281103551387787, | |
| "learning_rate": 1.1126188611144406e-05, | |
| "loss": 0.0981, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.4618211514904043, | |
| "grad_norm": 0.14536678791046143, | |
| "learning_rate": 1.102337070797137e-05, | |
| "loss": 0.0775, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 1.4645433510276304, | |
| "grad_norm": 0.14888562262058258, | |
| "learning_rate": 1.0920895552819118e-05, | |
| "loss": 0.0828, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 1.4672655505648564, | |
| "grad_norm": 0.12814417481422424, | |
| "learning_rate": 1.0818765658679576e-05, | |
| "loss": 0.0766, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 1.4699877501020824, | |
| "grad_norm": 0.12727950513362885, | |
| "learning_rate": 1.0716983530077843e-05, | |
| "loss": 0.0776, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.4727099496393086, | |
| "grad_norm": 0.12727922201156616, | |
| "learning_rate": 1.0615551663010806e-05, | |
| "loss": 0.0689, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 1.4754321491765348, | |
| "grad_norm": 0.14102789759635925, | |
| "learning_rate": 1.051447254488591e-05, | |
| "loss": 0.0634, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 1.4781543487137607, | |
| "grad_norm": 0.14436770975589752, | |
| "learning_rate": 1.0413748654460149e-05, | |
| "loss": 0.0796, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 1.4808765482509867, | |
| "grad_norm": 0.1291683316230774, | |
| "learning_rate": 1.0313382461779306e-05, | |
| "loss": 0.0755, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.4835987477882129, | |
| "grad_norm": 0.12600617110729218, | |
| "learning_rate": 1.0213376428117333e-05, | |
| "loss": 0.0658, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 1.486320947325439, | |
| "grad_norm": 0.14813588559627533, | |
| "learning_rate": 1.0113733005916057e-05, | |
| "loss": 0.075, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 1.489043146862665, | |
| "grad_norm": 0.15321391820907593, | |
| "learning_rate": 1.0014454638724982e-05, | |
| "loss": 0.0774, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 1.491765346399891, | |
| "grad_norm": 0.13844044506549835, | |
| "learning_rate": 9.915543761141432e-06, | |
| "loss": 0.074, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.4944875459371172, | |
| "grad_norm": 0.14296895265579224, | |
| "learning_rate": 9.81700279875075e-06, | |
| "loss": 0.0796, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 1.4972097454743434, | |
| "grad_norm": 0.14961597323417664, | |
| "learning_rate": 9.718834168066904e-06, | |
| "loss": 0.0713, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.4999319450115693, | |
| "grad_norm": 0.1663280427455902, | |
| "learning_rate": 9.62104027647319e-06, | |
| "loss": 0.077, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 1.5026541445487953, | |
| "grad_norm": 0.14651760458946228, | |
| "learning_rate": 9.523623522163197e-06, | |
| "loss": 0.0719, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.5053763440860215, | |
| "grad_norm": 0.14970749616622925, | |
| "learning_rate": 9.426586294082013e-06, | |
| "loss": 0.0776, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 1.5080985436232477, | |
| "grad_norm": 0.15019264817237854, | |
| "learning_rate": 9.329930971867595e-06, | |
| "loss": 0.0771, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 1.5108207431604737, | |
| "grad_norm": 0.14440083503723145, | |
| "learning_rate": 9.233659925792477e-06, | |
| "loss": 0.072, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 1.5135429426976996, | |
| "grad_norm": 0.138129323720932, | |
| "learning_rate": 9.137775516705604e-06, | |
| "loss": 0.0758, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.5162651422349258, | |
| "grad_norm": 0.14640666544437408, | |
| "learning_rate": 9.042280095974434e-06, | |
| "loss": 0.0844, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 1.518987341772152, | |
| "grad_norm": 0.14141102135181427, | |
| "learning_rate": 8.947176005427324e-06, | |
| "loss": 0.0798, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 1.521709541309378, | |
| "grad_norm": 0.14003820717334747, | |
| "learning_rate": 8.852465577296015e-06, | |
| "loss": 0.0761, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 1.524431740846604, | |
| "grad_norm": 0.13643573224544525, | |
| "learning_rate": 8.75815113415852e-06, | |
| "loss": 0.0753, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.5271539403838301, | |
| "grad_norm": 0.13099689781665802, | |
| "learning_rate": 8.66423498888213e-06, | |
| "loss": 0.0764, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 1.5298761399210563, | |
| "grad_norm": 0.16020213067531586, | |
| "learning_rate": 8.570719444566702e-06, | |
| "loss": 0.0733, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 1.5325983394582823, | |
| "grad_norm": 0.1546473205089569, | |
| "learning_rate": 8.477606794488183e-06, | |
| "loss": 0.0839, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 1.5353205389955082, | |
| "grad_norm": 0.14650870859622955, | |
| "learning_rate": 8.384899322042356e-06, | |
| "loss": 0.0837, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.5380427385327344, | |
| "grad_norm": 0.11506952345371246, | |
| "learning_rate": 8.29259930068887e-06, | |
| "loss": 0.0775, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 1.5407649380699606, | |
| "grad_norm": 0.14615637063980103, | |
| "learning_rate": 8.200708993895476e-06, | |
| "loss": 0.0744, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 1.5434871376071866, | |
| "grad_norm": 0.13769914209842682, | |
| "learning_rate": 8.109230655082517e-06, | |
| "loss": 0.0795, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 1.5462093371444126, | |
| "grad_norm": 0.13748113811016083, | |
| "learning_rate": 8.018166527567672e-06, | |
| "loss": 0.0675, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.5489315366816387, | |
| "grad_norm": 0.15049228072166443, | |
| "learning_rate": 7.927518844510941e-06, | |
| "loss": 0.0703, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 1.551653736218865, | |
| "grad_norm": 0.13132914900779724, | |
| "learning_rate": 7.837289828859884e-06, | |
| "loss": 0.0742, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 1.554375935756091, | |
| "grad_norm": 0.14202824234962463, | |
| "learning_rate": 7.7474816932951e-06, | |
| "loss": 0.0699, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 1.5570981352933169, | |
| "grad_norm": 0.16140979528427124, | |
| "learning_rate": 7.658096640175985e-06, | |
| "loss": 0.0777, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.559820334830543, | |
| "grad_norm": 0.1300312578678131, | |
| "learning_rate": 7.569136861486686e-06, | |
| "loss": 0.0648, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 1.5625425343677692, | |
| "grad_norm": 0.14492963254451752, | |
| "learning_rate": 7.4806045387823935e-06, | |
| "loss": 0.0713, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 1.5652647339049952, | |
| "grad_norm": 0.13518249988555908, | |
| "learning_rate": 7.392501843135816e-06, | |
| "loss": 0.0649, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 1.5679869334422212, | |
| "grad_norm": 0.14563652873039246, | |
| "learning_rate": 7.304830935083934e-06, | |
| "loss": 0.0724, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.5707091329794474, | |
| "grad_norm": 0.1495407372713089, | |
| "learning_rate": 7.2175939645750454e-06, | |
| "loss": 0.0695, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 1.5734313325166736, | |
| "grad_norm": 0.15107296407222748, | |
| "learning_rate": 7.130793070916006e-06, | |
| "loss": 0.0762, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 1.5761535320538995, | |
| "grad_norm": 0.13882042467594147, | |
| "learning_rate": 7.0444303827197994e-06, | |
| "loss": 0.0698, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 1.5788757315911255, | |
| "grad_norm": 0.15497738122940063, | |
| "learning_rate": 6.9585080178533155e-06, | |
| "loss": 0.0729, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.5815979311283517, | |
| "grad_norm": 0.14168910682201385, | |
| "learning_rate": 6.873028083385435e-06, | |
| "loss": 0.0681, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 1.5843201306655779, | |
| "grad_norm": 0.1269499659538269, | |
| "learning_rate": 6.787992675535318e-06, | |
| "loss": 0.0803, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 1.5870423302028038, | |
| "grad_norm": 0.14480414986610413, | |
| "learning_rate": 6.703403879621048e-06, | |
| "loss": 0.0719, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 1.5897645297400298, | |
| "grad_norm": 0.14816336333751678, | |
| "learning_rate": 6.619263770008471e-06, | |
| "loss": 0.0764, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.592486729277256, | |
| "grad_norm": 0.13959960639476776, | |
| "learning_rate": 6.535574410060321e-06, | |
| "loss": 0.0777, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 1.5952089288144822, | |
| "grad_norm": 0.13289447128772736, | |
| "learning_rate": 6.452337852085621e-06, | |
| "loss": 0.077, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 1.5979311283517081, | |
| "grad_norm": 0.14249089360237122, | |
| "learning_rate": 6.369556137289373e-06, | |
| "loss": 0.0678, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 1.600653327888934, | |
| "grad_norm": 0.14189301431179047, | |
| "learning_rate": 6.28723129572247e-06, | |
| "loss": 0.077, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.6033755274261603, | |
| "grad_norm": 0.13121193647384644, | |
| "learning_rate": 6.2053653462319475e-06, | |
| "loss": 0.0651, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 1.6060977269633865, | |
| "grad_norm": 0.13624070584774017, | |
| "learning_rate": 6.123960296411449e-06, | |
| "loss": 0.0726, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 1.6088199265006125, | |
| "grad_norm": 0.1546383649110794, | |
| "learning_rate": 6.043018142552009e-06, | |
| "loss": 0.0747, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 1.6115421260378384, | |
| "grad_norm": 0.14421793818473816, | |
| "learning_rate": 5.962540869593081e-06, | |
| "loss": 0.0818, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.6142643255750646, | |
| "grad_norm": 0.13135267794132233, | |
| "learning_rate": 5.882530451073886e-06, | |
| "loss": 0.0724, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 1.6169865251122908, | |
| "grad_norm": 0.14589843153953552, | |
| "learning_rate": 5.8029888490850005e-06, | |
| "loss": 0.0769, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 1.6197087246495168, | |
| "grad_norm": 0.1423359513282776, | |
| "learning_rate": 5.723918014220236e-06, | |
| "loss": 0.0662, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 1.6224309241867427, | |
| "grad_norm": 0.12941160798072815, | |
| "learning_rate": 5.645319885528824e-06, | |
| "loss": 0.0938, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.625153123723969, | |
| "grad_norm": 0.16445757448673248, | |
| "learning_rate": 5.5671963904678185e-06, | |
| "loss": 0.0752, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 1.627875323261195, | |
| "grad_norm": 0.15879502892494202, | |
| "learning_rate": 5.489549444854908e-06, | |
| "loss": 0.0875, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 1.630597522798421, | |
| "grad_norm": 0.1418047547340393, | |
| "learning_rate": 5.412380952821358e-06, | |
| "loss": 0.0778, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 1.633319722335647, | |
| "grad_norm": 0.1362149566411972, | |
| "learning_rate": 5.33569280676536e-06, | |
| "loss": 0.0692, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.633319722335647, | |
| "eval_loss": 0.07731131464242935, | |
| "eval_runtime": 262.646, | |
| "eval_samples_per_second": 2.284, | |
| "eval_steps_per_second": 0.571, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.6360419218728732, | |
| "grad_norm": 0.14646309614181519, | |
| "learning_rate": 5.259486887305592e-06, | |
| "loss": 0.068, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 1.6387641214100994, | |
| "grad_norm": 0.13968642055988312, | |
| "learning_rate": 5.183765063235138e-06, | |
| "loss": 0.0727, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 1.6414863209473254, | |
| "grad_norm": 0.1375684291124344, | |
| "learning_rate": 5.108529191475636e-06, | |
| "loss": 0.0794, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 1.6442085204845516, | |
| "grad_norm": 0.1326666921377182, | |
| "learning_rate": 5.033781117031738e-06, | |
| "loss": 0.072, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.6469307200217775, | |
| "grad_norm": 0.14453214406967163, | |
| "learning_rate": 4.959522672945891e-06, | |
| "loss": 0.0763, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 1.6496529195590037, | |
| "grad_norm": 0.15132352709770203, | |
| "learning_rate": 4.885755680253334e-06, | |
| "loss": 0.1175, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 1.65237511909623, | |
| "grad_norm": 0.15359079837799072, | |
| "learning_rate": 4.812481947937497e-06, | |
| "loss": 0.0837, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 1.6550973186334559, | |
| "grad_norm": 0.13241231441497803, | |
| "learning_rate": 4.739703272885626e-06, | |
| "loss": 0.0738, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.6578195181706818, | |
| "grad_norm": 0.14873500168323517, | |
| "learning_rate": 4.667421439844691e-06, | |
| "loss": 0.0708, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 1.660541717707908, | |
| "grad_norm": 0.13607364892959595, | |
| "learning_rate": 4.5956382213776295e-06, | |
| "loss": 0.1151, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 1.6632639172451342, | |
| "grad_norm": 0.15463878214359283, | |
| "learning_rate": 4.524355377819897e-06, | |
| "loss": 0.0831, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 1.6659861167823602, | |
| "grad_norm": 0.1272389441728592, | |
| "learning_rate": 4.45357465723629e-06, | |
| "loss": 0.0714, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.6687083163195862, | |
| "grad_norm": 0.14218097925186157, | |
| "learning_rate": 4.383297795378061e-06, | |
| "loss": 0.0675, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 1.6714305158568123, | |
| "grad_norm": 0.1373002678155899, | |
| "learning_rate": 4.3135265156403756e-06, | |
| "loss": 0.0791, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 1.6741527153940385, | |
| "grad_norm": 0.16303496062755585, | |
| "learning_rate": 4.244262529020041e-06, | |
| "loss": 0.0757, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 1.6768749149312645, | |
| "grad_norm": 0.1444956660270691, | |
| "learning_rate": 4.175507534073533e-06, | |
| "loss": 0.0732, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.6795971144684905, | |
| "grad_norm": 0.1505775898694992, | |
| "learning_rate": 4.107263216875362e-06, | |
| "loss": 0.0751, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 1.6823193140057167, | |
| "grad_norm": 0.14044207334518433, | |
| "learning_rate": 4.039531250976735e-06, | |
| "loss": 0.069, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 1.6850415135429428, | |
| "grad_norm": 0.14999620616436005, | |
| "learning_rate": 3.972313297364494e-06, | |
| "loss": 0.0708, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 1.6877637130801688, | |
| "grad_norm": 0.1265767216682434, | |
| "learning_rate": 3.90561100442036e-06, | |
| "loss": 0.0655, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.6904859126173948, | |
| "grad_norm": 0.16221390664577484, | |
| "learning_rate": 3.8394260078805755e-06, | |
| "loss": 0.0736, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 1.693208112154621, | |
| "grad_norm": 0.15090176463127136, | |
| "learning_rate": 3.7737599307957365e-06, | |
| "loss": 0.0669, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 1.6959303116918472, | |
| "grad_norm": 0.14406718313694, | |
| "learning_rate": 3.708614383491016e-06, | |
| "loss": 0.0737, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 1.6986525112290731, | |
| "grad_norm": 0.12115464359521866, | |
| "learning_rate": 3.643990963526672e-06, | |
| "loss": 0.0627, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.701374710766299, | |
| "grad_norm": 0.16066431999206543, | |
| "learning_rate": 3.579891255658846e-06, | |
| "loss": 0.0757, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 1.7040969103035253, | |
| "grad_norm": 0.13997657597064972, | |
| "learning_rate": 3.5163168318007365e-06, | |
| "loss": 0.1398, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 1.7068191098407515, | |
| "grad_norm": 0.13934873044490814, | |
| "learning_rate": 3.4532692509840313e-06, | |
| "loss": 0.0723, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 1.7095413093779774, | |
| "grad_norm": 0.13022823631763458, | |
| "learning_rate": 3.390750059320688e-06, | |
| "loss": 0.0789, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.7122635089152034, | |
| "grad_norm": 0.13600829243659973, | |
| "learning_rate": 3.328760789964988e-06, | |
| "loss": 0.0728, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 1.7149857084524296, | |
| "grad_norm": 0.1362736076116562, | |
| "learning_rate": 3.2673029630759745e-06, | |
| "loss": 0.0694, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 1.7177079079896558, | |
| "grad_norm": 0.14168758690357208, | |
| "learning_rate": 3.2063780857801596e-06, | |
| "loss": 0.0726, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 1.7204301075268817, | |
| "grad_norm": 0.14045333862304688, | |
| "learning_rate": 3.145987652134563e-06, | |
| "loss": 0.0666, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.7231523070641077, | |
| "grad_norm": 0.15181277692317963, | |
| "learning_rate": 3.0861331430900807e-06, | |
| "loss": 0.0788, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 1.725874506601334, | |
| "grad_norm": 0.14058373868465424, | |
| "learning_rate": 3.0268160264551484e-06, | |
| "loss": 0.0701, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 1.72859670613856, | |
| "grad_norm": 0.1474706381559372, | |
| "learning_rate": 2.968037756859776e-06, | |
| "loss": 0.0745, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 1.731318905675786, | |
| "grad_norm": 0.13882222771644592, | |
| "learning_rate": 2.9097997757198516e-06, | |
| "loss": 0.0717, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.734041105213012, | |
| "grad_norm": 0.14754945039749146, | |
| "learning_rate": 2.8521035112018063e-06, | |
| "loss": 0.0644, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 1.7367633047502382, | |
| "grad_norm": 0.1717384308576584, | |
| "learning_rate": 2.7949503781875912e-06, | |
| "loss": 0.0768, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 1.7394855042874644, | |
| "grad_norm": 0.1543627679347992, | |
| "learning_rate": 2.7383417782399693e-06, | |
| "loss": 0.0709, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 1.7422077038246904, | |
| "grad_norm": 0.1619829684495926, | |
| "learning_rate": 2.6822790995681577e-06, | |
| "loss": 0.0814, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.7449299033619163, | |
| "grad_norm": 0.13856656849384308, | |
| "learning_rate": 2.62676371699378e-06, | |
| "loss": 0.0701, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 1.7476521028991425, | |
| "grad_norm": 0.14767934381961823, | |
| "learning_rate": 2.5717969919171553e-06, | |
| "loss": 0.076, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 1.7503743024363687, | |
| "grad_norm": 0.1731363832950592, | |
| "learning_rate": 2.5173802722838867e-06, | |
| "loss": 0.0711, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 1.7530965019735947, | |
| "grad_norm": 0.14000795781612396, | |
| "learning_rate": 2.4635148925518577e-06, | |
| "loss": 0.0675, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.7558187015108206, | |
| "grad_norm": 0.14177638292312622, | |
| "learning_rate": 2.4102021736584575e-06, | |
| "loss": 0.0659, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 1.7585409010480468, | |
| "grad_norm": 0.1450289785861969, | |
| "learning_rate": 2.3574434229882145e-06, | |
| "loss": 0.0836, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 1.761263100585273, | |
| "grad_norm": 0.15300573408603668, | |
| "learning_rate": 2.3052399343407277e-06, | |
| "loss": 0.079, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 1.763985300122499, | |
| "grad_norm": 0.12552043795585632, | |
| "learning_rate": 2.253592987898942e-06, | |
| "loss": 0.075, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.766707499659725, | |
| "grad_norm": 0.13973963260650635, | |
| "learning_rate": 2.2025038501977486e-06, | |
| "loss": 0.075, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 1.7694296991969511, | |
| "grad_norm": 0.14422014355659485, | |
| "learning_rate": 2.1519737740929274e-06, | |
| "loss": 0.0847, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.7721518987341773, | |
| "grad_norm": 0.15774239599704742, | |
| "learning_rate": 2.1020039987304285e-06, | |
| "loss": 0.0852, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 1.7748740982714033, | |
| "grad_norm": 0.14551898837089539, | |
| "learning_rate": 2.05259574951599e-06, | |
| "loss": 0.0726, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.7775962978086293, | |
| "grad_norm": 0.1303868442773819, | |
| "learning_rate": 2.003750238085053e-06, | |
| "loss": 0.076, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 1.7803184973458555, | |
| "grad_norm": 0.16554243862628937, | |
| "learning_rate": 1.9554686622730993e-06, | |
| "loss": 0.0992, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 1.7830406968830816, | |
| "grad_norm": 0.15280041098594666, | |
| "learning_rate": 1.907752206086247e-06, | |
| "loss": 0.0819, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 1.7857628964203076, | |
| "grad_norm": 0.14237718284130096, | |
| "learning_rate": 1.8606020396722129e-06, | |
| "loss": 0.0726, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.7884850959575336, | |
| "grad_norm": 0.15816614031791687, | |
| "learning_rate": 1.814019319291635e-06, | |
| "loss": 0.0754, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 1.7912072954947598, | |
| "grad_norm": 0.13710108399391174, | |
| "learning_rate": 1.7680051872896869e-06, | |
| "loss": 0.0658, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 1.793929495031986, | |
| "grad_norm": 0.15852783620357513, | |
| "learning_rate": 1.7225607720681131e-06, | |
| "loss": 0.0732, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 1.796651694569212, | |
| "grad_norm": 0.15130798518657684, | |
| "learning_rate": 1.6776871880575084e-06, | |
| "loss": 0.0833, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.7993738941064379, | |
| "grad_norm": 0.15700189769268036, | |
| "learning_rate": 1.6333855356900185e-06, | |
| "loss": 0.1652, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 1.802096093643664, | |
| "grad_norm": 0.14298690855503082, | |
| "learning_rate": 1.5896569013723329e-06, | |
| "loss": 0.0797, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 1.8048182931808903, | |
| "grad_norm": 0.14160804450511932, | |
| "learning_rate": 1.5465023574590676e-06, | |
| "loss": 0.0757, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 1.8075404927181162, | |
| "grad_norm": 0.15241742134094238, | |
| "learning_rate": 1.5039229622264479e-06, | |
| "loss": 0.0781, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.8102626922553422, | |
| "grad_norm": 0.15816597640514374, | |
| "learning_rate": 1.4619197598463641e-06, | |
| "loss": 0.0764, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 1.8129848917925684, | |
| "grad_norm": 0.15082836151123047, | |
| "learning_rate": 1.420493780360771e-06, | |
| "loss": 0.0708, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 1.8157070913297946, | |
| "grad_norm": 0.13604667782783508, | |
| "learning_rate": 1.3796460396564098e-06, | |
| "loss": 0.0711, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 1.8184292908670205, | |
| "grad_norm": 0.14523722231388092, | |
| "learning_rate": 1.3393775394399123e-06, | |
| "loss": 0.0797, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.8211514904042465, | |
| "grad_norm": 0.14779233932495117, | |
| "learning_rate": 1.2996892672132416e-06, | |
| "loss": 0.0845, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 1.8238736899414727, | |
| "grad_norm": 0.13566653430461884, | |
| "learning_rate": 1.2605821962494574e-06, | |
| "loss": 0.1118, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 1.8265958894786989, | |
| "grad_norm": 0.15038765966892242, | |
| "learning_rate": 1.2220572855688507e-06, | |
| "loss": 0.0782, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 1.8293180890159249, | |
| "grad_norm": 0.14022061228752136, | |
| "learning_rate": 1.1841154799154374e-06, | |
| "loss": 0.0737, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.8320402885531508, | |
| "grad_norm": 0.14298531413078308, | |
| "learning_rate": 1.1467577097337867e-06, | |
| "loss": 0.0939, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 1.834762488090377, | |
| "grad_norm": 0.15060581266880035, | |
| "learning_rate": 1.1099848911462014e-06, | |
| "loss": 0.0752, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 1.8374846876276032, | |
| "grad_norm": 0.1449151188135147, | |
| "learning_rate": 1.0737979259302478e-06, | |
| "loss": 0.0688, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 1.8402068871648292, | |
| "grad_norm": 0.13491316139698029, | |
| "learning_rate": 1.0381977014966543e-06, | |
| "loss": 0.0996, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.8429290867020551, | |
| "grad_norm": 0.128363698720932, | |
| "learning_rate": 1.0031850908675283e-06, | |
| "loss": 0.1217, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 1.8456512862392813, | |
| "grad_norm": 0.14816632866859436, | |
| "learning_rate": 9.68760952654968e-07, | |
| "loss": 0.0653, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 1.8483734857765075, | |
| "grad_norm": 0.16936026513576508, | |
| "learning_rate": 9.349261310400037e-07, | |
| "loss": 0.0735, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 1.8510956853137335, | |
| "grad_norm": 0.14492131769657135, | |
| "learning_rate": 9.016814557518849e-07, | |
| "loss": 0.0683, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.8538178848509594, | |
| "grad_norm": 0.14610859751701355, | |
| "learning_rate": 8.690277420477372e-07, | |
| "loss": 0.079, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 1.8565400843881856, | |
| "grad_norm": 0.1451197862625122, | |
| "learning_rate": 8.369657906925732e-07, | |
| "loss": 0.0753, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 1.8592622839254118, | |
| "grad_norm": 0.14033135771751404, | |
| "learning_rate": 8.054963879396554e-07, | |
| "loss": 0.0646, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 1.8619844834626378, | |
| "grad_norm": 0.13456685841083527, | |
| "learning_rate": 7.746203055112145e-07, | |
| "loss": 0.1086, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.8647066829998638, | |
| "grad_norm": 0.15288980305194855, | |
| "learning_rate": 7.443383005795224e-07, | |
| "loss": 0.0946, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 1.86742888253709, | |
| "grad_norm": 0.14935944974422455, | |
| "learning_rate": 7.146511157483216e-07, | |
| "loss": 0.0829, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 1.8701510820743161, | |
| "grad_norm": 0.14166168868541718, | |
| "learning_rate": 6.85559479034617e-07, | |
| "loss": 0.0679, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 1.872873281611542, | |
| "grad_norm": 0.14961621165275574, | |
| "learning_rate": 6.570641038508296e-07, | |
| "loss": 0.0762, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.875595481148768, | |
| "grad_norm": 0.1400919258594513, | |
| "learning_rate": 6.291656889873016e-07, | |
| "loss": 0.0803, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 1.8783176806859943, | |
| "grad_norm": 0.13169647753238678, | |
| "learning_rate": 6.018649185951325e-07, | |
| "loss": 0.0707, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 1.8810398802232204, | |
| "grad_norm": 0.13241390883922577, | |
| "learning_rate": 5.751624621694429e-07, | |
| "loss": 0.0636, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 1.8837620797604464, | |
| "grad_norm": 0.16256216168403625, | |
| "learning_rate": 5.490589745329261e-07, | |
| "loss": 0.0774, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.8864842792976724, | |
| "grad_norm": 0.15392127633094788, | |
| "learning_rate": 5.235550958198083e-07, | |
| "loss": 0.073, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 1.8892064788348986, | |
| "grad_norm": 0.14630888402462006, | |
| "learning_rate": 4.986514514601415e-07, | |
| "loss": 0.0742, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 1.8919286783721248, | |
| "grad_norm": 0.16385363042354584, | |
| "learning_rate": 4.74348652164458e-07, | |
| "loss": 0.083, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 1.8946508779093507, | |
| "grad_norm": 0.16106607019901276, | |
| "learning_rate": 4.5064729390881246e-07, | |
| "loss": 0.0755, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 1.8973730774465767, | |
| "grad_norm": 0.13656839728355408, | |
| "learning_rate": 4.275479579201602e-07, | |
| "loss": 0.0703, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 1.9000952769838029, | |
| "grad_norm": 0.12616074085235596, | |
| "learning_rate": 4.050512106620913e-07, | |
| "loss": 0.0692, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 1.902817476521029, | |
| "grad_norm": 0.15891359746456146, | |
| "learning_rate": 3.8315760382096057e-07, | |
| "loss": 0.0794, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 1.905539676058255, | |
| "grad_norm": 0.1297856718301773, | |
| "learning_rate": 3.6186767429234323e-07, | |
| "loss": 0.0671, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.905539676058255, | |
| "eval_loss": 0.07686587423086166, | |
| "eval_runtime": 218.8338, | |
| "eval_samples_per_second": 2.742, | |
| "eval_steps_per_second": 0.685, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.908261875595481, | |
| "grad_norm": 0.16892676055431366, | |
| "learning_rate": 3.4118194416786764e-07, | |
| "loss": 0.076, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 1.9109840751327072, | |
| "grad_norm": 0.14325669407844543, | |
| "learning_rate": 3.211009207224225e-07, | |
| "loss": 0.0723, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 1.9137062746699334, | |
| "grad_norm": 0.12773250043392181, | |
| "learning_rate": 3.016250964017142e-07, | |
| "loss": 0.0727, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 1.9164284742071593, | |
| "grad_norm": 0.15402384102344513, | |
| "learning_rate": 2.827549488101849e-07, | |
| "loss": 0.0699, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 1.9191506737443853, | |
| "grad_norm": 0.13994073867797852, | |
| "learning_rate": 2.644909406992996e-07, | |
| "loss": 0.0674, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 1.9218728732816115, | |
| "grad_norm": 0.15044575929641724, | |
| "learning_rate": 2.468335199562105e-07, | |
| "loss": 0.0779, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 1.9245950728188377, | |
| "grad_norm": 0.13634447753429413, | |
| "learning_rate": 2.29783119592758e-07, | |
| "loss": 0.0697, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 1.9273172723560636, | |
| "grad_norm": 0.1296970546245575, | |
| "learning_rate": 2.1334015773486203e-07, | |
| "loss": 0.0682, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 1.9300394718932896, | |
| "grad_norm": 0.14532914757728577, | |
| "learning_rate": 1.975050376122667e-07, | |
| "loss": 0.0787, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 1.9327616714305158, | |
| "grad_norm": 0.15663999319076538, | |
| "learning_rate": 1.8227814754865068e-07, | |
| "loss": 0.0735, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 1.935483870967742, | |
| "grad_norm": 0.13645371794700623, | |
| "learning_rate": 1.6765986095209908e-07, | |
| "loss": 0.0684, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 1.938206070504968, | |
| "grad_norm": 0.14688655734062195, | |
| "learning_rate": 1.536505363059576e-07, | |
| "loss": 0.0738, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 1.9409282700421941, | |
| "grad_norm": 0.14657242596149445, | |
| "learning_rate": 1.4025051716003157e-07, | |
| "loss": 0.083, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 1.9436504695794201, | |
| "grad_norm": 0.1277332305908203, | |
| "learning_rate": 1.2746013212217022e-07, | |
| "loss": 0.0758, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 1.9463726691166463, | |
| "grad_norm": 0.1419767141342163, | |
| "learning_rate": 1.1527969485019275e-07, | |
| "loss": 0.0755, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 1.9490948686538725, | |
| "grad_norm": 0.15132427215576172, | |
| "learning_rate": 1.0370950404421931e-07, | |
| "loss": 0.0651, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 1.9518170681910985, | |
| "grad_norm": 0.1558247208595276, | |
| "learning_rate": 9.274984343932702e-08, | |
| "loss": 0.0785, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 1.9545392677283244, | |
| "grad_norm": 0.1476377248764038, | |
| "learning_rate": 8.240098179859712e-08, | |
| "loss": 0.0826, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 1.9572614672655506, | |
| "grad_norm": 0.14471116662025452, | |
| "learning_rate": 7.26631729065258e-08, | |
| "loss": 0.0816, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 1.9599836668027768, | |
| "grad_norm": 0.14453163743019104, | |
| "learning_rate": 6.353665556280697e-08, | |
| "loss": 0.0699, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.9627058663400028, | |
| "grad_norm": 0.13174843788146973, | |
| "learning_rate": 5.5021653576459164e-08, | |
| "loss": 0.062, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 1.9654280658772287, | |
| "grad_norm": 0.1312907189130783, | |
| "learning_rate": 4.7118375760357716e-08, | |
| "loss": 0.0804, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 1.968150265414455, | |
| "grad_norm": 0.15283526480197906, | |
| "learning_rate": 3.982701592609439e-08, | |
| "loss": 0.1359, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 1.9708724649516811, | |
| "grad_norm": 0.14205031096935272, | |
| "learning_rate": 3.314775287923677e-08, | |
| "loss": 0.071, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 1.973594664488907, | |
| "grad_norm": 0.16160933673381805, | |
| "learning_rate": 2.708075041494562e-08, | |
| "loss": 0.0784, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 1.976316864026133, | |
| "grad_norm": 0.12819606065750122, | |
| "learning_rate": 2.1626157313950345e-08, | |
| "loss": 0.097, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 1.9790390635633592, | |
| "grad_norm": 0.1536770612001419, | |
| "learning_rate": 1.6784107338899124e-08, | |
| "loss": 0.0775, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 1.9817612631005854, | |
| "grad_norm": 0.12998686730861664, | |
| "learning_rate": 1.2554719231083755e-08, | |
| "loss": 0.0696, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 1.9844834626378114, | |
| "grad_norm": 0.11487606167793274, | |
| "learning_rate": 8.93809670753365e-09, | |
| "loss": 0.0674, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 1.9872056621750374, | |
| "grad_norm": 0.13673199713230133, | |
| "learning_rate": 5.934328458459537e-09, | |
| "loss": 0.0712, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 1.9899278617122635, | |
| "grad_norm": 0.1361059844493866, | |
| "learning_rate": 3.543488145082985e-09, | |
| "loss": 0.0614, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 1.9926500612494897, | |
| "grad_norm": 0.15082420408725739, | |
| "learning_rate": 1.7656343978378342e-09, | |
| "loss": 0.0762, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 1.9953722607867157, | |
| "grad_norm": 0.1459612250328064, | |
| "learning_rate": 6.008108149185843e-10, | |
| "loss": 0.0841, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 1.9980944603239417, | |
| "grad_norm": 0.15308088064193726, | |
| "learning_rate": 4.904596122290439e-11, | |
| "loss": 0.0805, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 1.9991833401388321, | |
| "step": 3672, | |
| "total_flos": 2.724210105233965e+18, | |
| "train_loss": 0.10618524583693684, | |
| "train_runtime": 81161.5668, | |
| "train_samples_per_second": 0.724, | |
| "train_steps_per_second": 0.045 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 3672, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.724210105233965e+18, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |