| { |
| "best_metric": 3.3056321144104004, |
| "best_model_checkpoint": "/scratch/cl5625/exceptions/models/100M_8397/checkpoint-90000", |
| "epoch": 10.0, |
| "eval_steps": 1000, |
| "global_step": 92910, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.005381552039608223, |
| "grad_norm": 2.765322685241699, |
| "learning_rate": 0.0003, |
| "loss": 8.4843, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.010763104079216447, |
| "grad_norm": 4.5047125816345215, |
| "learning_rate": 0.0006, |
| "loss": 6.9314, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.01614465611882467, |
| "grad_norm": 1.1575511693954468, |
| "learning_rate": 0.0005996767589699385, |
| "loss": 6.5187, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.021526208158432893, |
| "grad_norm": 2.314443826675415, |
| "learning_rate": 0.0005993535179398771, |
| "loss": 6.2703, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.026907760198041114, |
| "grad_norm": 0.848477303981781, |
| "learning_rate": 0.0005990302769098158, |
| "loss": 6.1093, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.03228931223764934, |
| "grad_norm": 1.2320305109024048, |
| "learning_rate": 0.0005987070358797543, |
| "loss": 5.9673, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.03767086427725756, |
| "grad_norm": 1.2428854703903198, |
| "learning_rate": 0.0005983837948496929, |
| "loss": 5.8876, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.04305241631686579, |
| "grad_norm": 1.5372645854949951, |
| "learning_rate": 0.0005980605538196314, |
| "loss": 5.7962, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.048433968356474004, |
| "grad_norm": 1.0138875246047974, |
| "learning_rate": 0.0005977373127895701, |
| "loss": 5.7055, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.05381552039608223, |
| "grad_norm": 1.1828784942626953, |
| "learning_rate": 0.0005974140717595086, |
| "loss": 5.6615, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.05919707243569045, |
| "grad_norm": 1.8574672937393188, |
| "learning_rate": 0.0005970908307294472, |
| "loss": 5.564, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.06457862447529868, |
| "grad_norm": 1.2299022674560547, |
| "learning_rate": 0.0005967675896993858, |
| "loss": 5.4985, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.0699601765149069, |
| "grad_norm": 1.4707765579223633, |
| "learning_rate": 0.0005964443486693243, |
| "loss": 5.4515, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.07534172855451512, |
| "grad_norm": 1.6264588832855225, |
| "learning_rate": 0.000596121107639263, |
| "loss": 5.3654, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.08072328059412334, |
| "grad_norm": 1.0290075540542603, |
| "learning_rate": 0.0005957978666092015, |
| "loss": 5.3191, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.08610483263373157, |
| "grad_norm": 1.0803887844085693, |
| "learning_rate": 0.0005954746255791401, |
| "loss": 5.2614, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.09148638467333979, |
| "grad_norm": 1.364003300666809, |
| "learning_rate": 0.0005951513845490787, |
| "loss": 5.2103, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.09686793671294801, |
| "grad_norm": 1.0939010381698608, |
| "learning_rate": 0.0005948281435190174, |
| "loss": 5.1724, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.10224948875255624, |
| "grad_norm": 1.1524903774261475, |
| "learning_rate": 0.0005945049024889559, |
| "loss": 5.1273, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.10763104079216446, |
| "grad_norm": 1.0799986124038696, |
| "learning_rate": 0.0005941816614588944, |
| "loss": 5.0654, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.10763104079216446, |
| "eval_accuracy": 0.22742907383952468, |
| "eval_loss": 5.017577648162842, |
| "eval_runtime": 181.071, |
| "eval_samples_per_second": 99.469, |
| "eval_steps_per_second": 6.219, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.11301259283177269, |
| "grad_norm": 1.048494577407837, |
| "learning_rate": 0.000593858420428833, |
| "loss": 5.0541, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.1183941448713809, |
| "grad_norm": 1.2681626081466675, |
| "learning_rate": 0.0005935351793987716, |
| "loss": 5.0134, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.12377569691098914, |
| "grad_norm": 1.3078149557113647, |
| "learning_rate": 0.0005932119383687103, |
| "loss": 5.0027, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.12915724895059735, |
| "grad_norm": 0.8166284561157227, |
| "learning_rate": 0.0005928886973386488, |
| "loss": 4.9447, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.13453880099020557, |
| "grad_norm": 1.0351450443267822, |
| "learning_rate": 0.0005925654563085874, |
| "loss": 4.93, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.1399203530298138, |
| "grad_norm": 1.2266950607299805, |
| "learning_rate": 0.000592242215278526, |
| "loss": 4.8866, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.14530190506942203, |
| "grad_norm": 1.2586219310760498, |
| "learning_rate": 0.0005919189742484645, |
| "loss": 4.8748, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.15068345710903025, |
| "grad_norm": 1.322376012802124, |
| "learning_rate": 0.0005915957332184032, |
| "loss": 4.8629, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.15606500914863847, |
| "grad_norm": 0.9112418293952942, |
| "learning_rate": 0.0005912724921883417, |
| "loss": 4.8164, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.16144656118824668, |
| "grad_norm": 1.3490499258041382, |
| "learning_rate": 0.0005909492511582803, |
| "loss": 4.8008, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.1668281132278549, |
| "grad_norm": 1.0545638799667358, |
| "learning_rate": 0.0005906260101282189, |
| "loss": 4.7662, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.17220966526746315, |
| "grad_norm": 0.9380106925964355, |
| "learning_rate": 0.0005903027690981575, |
| "loss": 4.7539, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.17759121730707136, |
| "grad_norm": 0.9979204535484314, |
| "learning_rate": 0.000589979528068096, |
| "loss": 4.7435, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.18297276934667958, |
| "grad_norm": 0.7762619256973267, |
| "learning_rate": 0.0005896562870380347, |
| "loss": 4.7269, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.1883543213862878, |
| "grad_norm": 1.121840000152588, |
| "learning_rate": 0.0005893330460079732, |
| "loss": 4.6665, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.19373587342589602, |
| "grad_norm": 0.9376957416534424, |
| "learning_rate": 0.0005890098049779118, |
| "loss": 4.6738, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.19911742546550426, |
| "grad_norm": 1.1272518634796143, |
| "learning_rate": 0.0005886865639478504, |
| "loss": 4.6606, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.20449897750511248, |
| "grad_norm": 0.9599424600601196, |
| "learning_rate": 0.0005883633229177889, |
| "loss": 4.6454, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.2098805295447207, |
| "grad_norm": 1.1271470785140991, |
| "learning_rate": 0.0005880400818877276, |
| "loss": 4.5919, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.2152620815843289, |
| "grad_norm": 1.019126534461975, |
| "learning_rate": 0.0005877168408576662, |
| "loss": 4.5787, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.2152620815843289, |
| "eval_accuracy": 0.2694480791853554, |
| "eval_loss": 4.517198085784912, |
| "eval_runtime": 179.8348, |
| "eval_samples_per_second": 100.153, |
| "eval_steps_per_second": 6.261, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.22064363362393713, |
| "grad_norm": 0.9893919825553894, |
| "learning_rate": 0.0005873935998276048, |
| "loss": 4.5569, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.22602518566354537, |
| "grad_norm": 0.9964181184768677, |
| "learning_rate": 0.0005870703587975433, |
| "loss": 4.5256, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.2314067377031536, |
| "grad_norm": 0.8266022801399231, |
| "learning_rate": 0.0005867471177674818, |
| "loss": 4.5383, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.2367882897427618, |
| "grad_norm": 1.0192140340805054, |
| "learning_rate": 0.0005864238767374205, |
| "loss": 4.4971, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.24216984178237003, |
| "grad_norm": 1.1359943151474, |
| "learning_rate": 0.0005861006357073591, |
| "loss": 4.4988, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.24755139382197827, |
| "grad_norm": 0.9294039011001587, |
| "learning_rate": 0.0005857773946772977, |
| "loss": 4.4732, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.2529329458615865, |
| "grad_norm": 0.8425009846687317, |
| "learning_rate": 0.0005854541536472362, |
| "loss": 4.4396, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.2583144979011947, |
| "grad_norm": 0.6667693853378296, |
| "learning_rate": 0.0005851309126171749, |
| "loss": 4.4376, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.2636960499408029, |
| "grad_norm": 0.9736728668212891, |
| "learning_rate": 0.0005848076715871134, |
| "loss": 4.4274, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.26907760198041114, |
| "grad_norm": 0.9293293356895447, |
| "learning_rate": 0.000584484430557052, |
| "loss": 4.4234, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.27445915402001936, |
| "grad_norm": 0.9411072134971619, |
| "learning_rate": 0.0005841611895269906, |
| "loss": 4.4028, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.2798407060596276, |
| "grad_norm": 0.9999932646751404, |
| "learning_rate": 0.0005838379484969291, |
| "loss": 4.4162, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.2852222580992358, |
| "grad_norm": 0.8416957855224609, |
| "learning_rate": 0.0005835147074668678, |
| "loss": 4.3881, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.29060381013884407, |
| "grad_norm": 0.708859384059906, |
| "learning_rate": 0.0005831914664368063, |
| "loss": 4.3881, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.2959853621784523, |
| "grad_norm": 0.7405576109886169, |
| "learning_rate": 0.0005828682254067449, |
| "loss": 4.3532, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.3013669142180605, |
| "grad_norm": 0.8557761907577515, |
| "learning_rate": 0.0005825449843766835, |
| "loss": 4.3477, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.3067484662576687, |
| "grad_norm": 0.8067885637283325, |
| "learning_rate": 0.0005822217433466221, |
| "loss": 4.3309, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.31213001829727693, |
| "grad_norm": 1.0514090061187744, |
| "learning_rate": 0.0005818985023165607, |
| "loss": 4.3391, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.31751157033688515, |
| "grad_norm": 0.6951010227203369, |
| "learning_rate": 0.0005815752612864992, |
| "loss": 4.2988, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.32289312237649337, |
| "grad_norm": 0.7815271615982056, |
| "learning_rate": 0.0005812520202564378, |
| "loss": 4.3112, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.32289312237649337, |
| "eval_accuracy": 0.29842786877034594, |
| "eval_loss": 4.237146854400635, |
| "eval_runtime": 179.5413, |
| "eval_samples_per_second": 100.317, |
| "eval_steps_per_second": 6.272, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.3282746744161016, |
| "grad_norm": 0.7033036351203918, |
| "learning_rate": 0.0005809287792263764, |
| "loss": 4.3185, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.3336562264557098, |
| "grad_norm": 0.8528683185577393, |
| "learning_rate": 0.0005806055381963151, |
| "loss": 4.2874, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.3390377784953181, |
| "grad_norm": 0.7960010766983032, |
| "learning_rate": 0.0005802822971662536, |
| "loss": 4.2738, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.3444193305349263, |
| "grad_norm": 0.638249397277832, |
| "learning_rate": 0.0005799590561361922, |
| "loss": 4.2911, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.3498008825745345, |
| "grad_norm": 0.8658631443977356, |
| "learning_rate": 0.0005796358151061307, |
| "loss": 4.2494, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.35518243461414273, |
| "grad_norm": 0.7108155488967896, |
| "learning_rate": 0.0005793125740760694, |
| "loss": 4.27, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.36056398665375095, |
| "grad_norm": 0.7243547439575195, |
| "learning_rate": 0.0005789893330460079, |
| "loss": 4.2532, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.36594553869335916, |
| "grad_norm": 0.7133619785308838, |
| "learning_rate": 0.0005786660920159465, |
| "loss": 4.2336, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.3713270907329674, |
| "grad_norm": 0.7307054400444031, |
| "learning_rate": 0.0005783428509858851, |
| "loss": 4.2161, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.3767086427725756, |
| "grad_norm": 0.783027708530426, |
| "learning_rate": 0.0005780196099558237, |
| "loss": 4.2252, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.3820901948121838, |
| "grad_norm": 0.774836540222168, |
| "learning_rate": 0.0005776963689257623, |
| "loss": 4.2057, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.38747174685179203, |
| "grad_norm": 0.7645639777183533, |
| "learning_rate": 0.0005773731278957008, |
| "loss": 4.2043, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.3928532988914003, |
| "grad_norm": 0.6912294626235962, |
| "learning_rate": 0.0005770498868656394, |
| "loss": 4.176, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.3982348509310085, |
| "grad_norm": 1.0210829973220825, |
| "learning_rate": 0.000576726645835578, |
| "loss": 4.1892, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.40361640297061674, |
| "grad_norm": 0.7821992635726929, |
| "learning_rate": 0.0005764034048055167, |
| "loss": 4.1735, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.40899795501022496, |
| "grad_norm": 0.8775643110275269, |
| "learning_rate": 0.0005760801637754552, |
| "loss": 4.1808, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.4143795070498332, |
| "grad_norm": 0.787700891494751, |
| "learning_rate": 0.0005757569227453937, |
| "loss": 4.1648, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.4197610590894414, |
| "grad_norm": 0.8575156927108765, |
| "learning_rate": 0.0005754336817153324, |
| "loss": 4.1757, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.4251426111290496, |
| "grad_norm": 0.7350611090660095, |
| "learning_rate": 0.0005751104406852709, |
| "loss": 4.1602, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.4305241631686578, |
| "grad_norm": 0.6890688538551331, |
| "learning_rate": 0.0005747871996552096, |
| "loss": 4.164, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.4305241631686578, |
| "eval_accuracy": 0.31287316155239336, |
| "eval_loss": 4.085758209228516, |
| "eval_runtime": 179.588, |
| "eval_samples_per_second": 100.291, |
| "eval_steps_per_second": 6.27, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.43590571520826604, |
| "grad_norm": 0.7492932677268982, |
| "learning_rate": 0.0005744639586251481, |
| "loss": 4.1546, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.44128726724787426, |
| "grad_norm": 0.6462448239326477, |
| "learning_rate": 0.0005741407175950867, |
| "loss": 4.159, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.44666881928748253, |
| "grad_norm": 0.7062949538230896, |
| "learning_rate": 0.0005738174765650253, |
| "loss": 4.1492, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.45205037132709075, |
| "grad_norm": 0.8296352624893188, |
| "learning_rate": 0.0005734942355349638, |
| "loss": 4.1268, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.45743192336669897, |
| "grad_norm": 0.7677702307701111, |
| "learning_rate": 0.0005731709945049025, |
| "loss": 4.1527, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.4628134754063072, |
| "grad_norm": 0.5997741222381592, |
| "learning_rate": 0.000572847753474841, |
| "loss": 4.1248, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.4681950274459154, |
| "grad_norm": 0.7054913640022278, |
| "learning_rate": 0.0005725245124447796, |
| "loss": 4.1242, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.4735765794855236, |
| "grad_norm": 0.6141555309295654, |
| "learning_rate": 0.0005722012714147182, |
| "loss": 4.102, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.47895813152513184, |
| "grad_norm": 0.9202010035514832, |
| "learning_rate": 0.0005718780303846568, |
| "loss": 4.1078, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.48433968356474005, |
| "grad_norm": 0.7407594919204712, |
| "learning_rate": 0.0005715547893545953, |
| "loss": 4.1116, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.48972123560434827, |
| "grad_norm": 0.7345452308654785, |
| "learning_rate": 0.000571231548324534, |
| "loss": 4.108, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.49510278764395654, |
| "grad_norm": 0.6181378364562988, |
| "learning_rate": 0.0005709083072944725, |
| "loss": 4.0897, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.5004843396835648, |
| "grad_norm": 0.5611840486526489, |
| "learning_rate": 0.0005705850662644111, |
| "loss": 4.0879, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.505865891723173, |
| "grad_norm": 0.7019966840744019, |
| "learning_rate": 0.0005702618252343497, |
| "loss": 4.1121, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.5112474437627812, |
| "grad_norm": 0.5246710777282715, |
| "learning_rate": 0.0005699385842042882, |
| "loss": 4.0842, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.5166289958023894, |
| "grad_norm": 0.7353329062461853, |
| "learning_rate": 0.0005696153431742269, |
| "loss": 4.0774, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.5220105478419976, |
| "grad_norm": 0.674192488193512, |
| "learning_rate": 0.0005692921021441655, |
| "loss": 4.0947, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.5273920998816058, |
| "grad_norm": 0.7477933168411255, |
| "learning_rate": 0.0005689688611141041, |
| "loss": 4.0597, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.5327736519212141, |
| "grad_norm": 0.8160755634307861, |
| "learning_rate": 0.0005686456200840426, |
| "loss": 4.0517, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.5381552039608223, |
| "grad_norm": 0.5745596885681152, |
| "learning_rate": 0.0005683223790539811, |
| "loss": 4.0509, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5381552039608223, |
| "eval_accuracy": 0.32170837977791567, |
| "eval_loss": 3.98921537399292, |
| "eval_runtime": 179.5969, |
| "eval_samples_per_second": 100.286, |
| "eval_steps_per_second": 6.27, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5435367560004305, |
| "grad_norm": 0.701214611530304, |
| "learning_rate": 0.0005679991380239198, |
| "loss": 4.0438, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.5489183080400387, |
| "grad_norm": 0.7315069437026978, |
| "learning_rate": 0.0005676758969938584, |
| "loss": 4.0379, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.5542998600796469, |
| "grad_norm": 0.7453345656394958, |
| "learning_rate": 0.000567352655963797, |
| "loss": 4.0327, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.5596814121192552, |
| "grad_norm": 0.7549764513969421, |
| "learning_rate": 0.0005670294149337355, |
| "loss": 4.0571, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.5650629641588634, |
| "grad_norm": 0.6196472644805908, |
| "learning_rate": 0.0005667061739036742, |
| "loss": 4.0388, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.5704445161984716, |
| "grad_norm": 0.6065132021903992, |
| "learning_rate": 0.0005663829328736127, |
| "loss": 4.023, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.5758260682380799, |
| "grad_norm": 0.7559676170349121, |
| "learning_rate": 0.0005660596918435512, |
| "loss": 4.0224, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.5812076202776881, |
| "grad_norm": 0.5514401793479919, |
| "learning_rate": 0.0005657364508134899, |
| "loss": 4.0343, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.5865891723172963, |
| "grad_norm": 0.6436402201652527, |
| "learning_rate": 0.0005654132097834284, |
| "loss": 4.0141, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.5919707243569046, |
| "grad_norm": 0.6794455647468567, |
| "learning_rate": 0.0005650899687533671, |
| "loss": 4.0005, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.5973522763965128, |
| "grad_norm": 0.6911609768867493, |
| "learning_rate": 0.0005647667277233056, |
| "loss": 4.0005, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.602733828436121, |
| "grad_norm": 0.6370590329170227, |
| "learning_rate": 0.0005644434866932442, |
| "loss": 4.0197, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.6081153804757292, |
| "grad_norm": 0.5897583365440369, |
| "learning_rate": 0.0005641202456631828, |
| "loss": 4.0107, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.6134969325153374, |
| "grad_norm": 0.6555122137069702, |
| "learning_rate": 0.0005637970046331214, |
| "loss": 4.0186, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.6188784845549457, |
| "grad_norm": 0.7258750200271606, |
| "learning_rate": 0.00056347376360306, |
| "loss": 4.0089, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.6242600365945539, |
| "grad_norm": 0.7162750363349915, |
| "learning_rate": 0.0005631505225729985, |
| "loss": 4.0215, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.6296415886341621, |
| "grad_norm": 0.5979080200195312, |
| "learning_rate": 0.0005628272815429371, |
| "loss": 3.991, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.6350231406737703, |
| "grad_norm": 0.5397379398345947, |
| "learning_rate": 0.0005625040405128757, |
| "loss": 4.0112, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.6404046927133785, |
| "grad_norm": 0.8027802109718323, |
| "learning_rate": 0.0005621807994828143, |
| "loss": 3.9821, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.6457862447529867, |
| "grad_norm": 0.7149248719215393, |
| "learning_rate": 0.0005618575584527529, |
| "loss": 3.9733, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.6457862447529867, |
| "eval_accuracy": 0.3285034227289455, |
| "eval_loss": 3.9209306240081787, |
| "eval_runtime": 179.6293, |
| "eval_samples_per_second": 100.268, |
| "eval_steps_per_second": 6.268, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.651167796792595, |
| "grad_norm": 0.5952973365783691, |
| "learning_rate": 0.0005615343174226915, |
| "loss": 3.9764, |
| "step": 6050 |
| }, |
| { |
| "epoch": 0.6565493488322032, |
| "grad_norm": 0.5565899014472961, |
| "learning_rate": 0.00056121107639263, |
| "loss": 3.9602, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.6619309008718114, |
| "grad_norm": 0.48603519797325134, |
| "learning_rate": 0.0005608878353625687, |
| "loss": 3.9707, |
| "step": 6150 |
| }, |
| { |
| "epoch": 0.6673124529114196, |
| "grad_norm": 0.6118786334991455, |
| "learning_rate": 0.0005605645943325072, |
| "loss": 3.9832, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.6726940049510278, |
| "grad_norm": 0.6439682245254517, |
| "learning_rate": 0.0005602413533024458, |
| "loss": 3.9652, |
| "step": 6250 |
| }, |
| { |
| "epoch": 0.6780755569906362, |
| "grad_norm": 0.5692785978317261, |
| "learning_rate": 0.0005599181122723844, |
| "loss": 3.9682, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.6834571090302444, |
| "grad_norm": 0.5279796719551086, |
| "learning_rate": 0.000559594871242323, |
| "loss": 3.9575, |
| "step": 6350 |
| }, |
| { |
| "epoch": 0.6888386610698526, |
| "grad_norm": 0.5329293012619019, |
| "learning_rate": 0.0005592716302122616, |
| "loss": 3.9661, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.6942202131094608, |
| "grad_norm": 0.59842848777771, |
| "learning_rate": 0.0005589483891822001, |
| "loss": 3.945, |
| "step": 6450 |
| }, |
| { |
| "epoch": 0.699601765149069, |
| "grad_norm": 0.6459380984306335, |
| "learning_rate": 0.0005586251481521387, |
| "loss": 3.9708, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.7049833171886772, |
| "grad_norm": 0.6028556227684021, |
| "learning_rate": 0.0005583083719426786, |
| "loss": 3.9382, |
| "step": 6550 |
| }, |
| { |
| "epoch": 0.7103648692282855, |
| "grad_norm": 0.6786797046661377, |
| "learning_rate": 0.0005579851309126171, |
| "loss": 3.9489, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.7157464212678937, |
| "grad_norm": 0.5982097387313843, |
| "learning_rate": 0.0005576618898825558, |
| "loss": 3.9588, |
| "step": 6650 |
| }, |
| { |
| "epoch": 0.7211279733075019, |
| "grad_norm": 0.668205738067627, |
| "learning_rate": 0.0005573386488524943, |
| "loss": 3.951, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.7265095253471101, |
| "grad_norm": 0.5475454330444336, |
| "learning_rate": 0.0005570154078224328, |
| "loss": 3.9315, |
| "step": 6750 |
| }, |
| { |
| "epoch": 0.7318910773867183, |
| "grad_norm": 0.5718148946762085, |
| "learning_rate": 0.0005566921667923715, |
| "loss": 3.9428, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.7372726294263265, |
| "grad_norm": 0.604880690574646, |
| "learning_rate": 0.00055636892576231, |
| "loss": 3.9552, |
| "step": 6850 |
| }, |
| { |
| "epoch": 0.7426541814659348, |
| "grad_norm": 0.5904679298400879, |
| "learning_rate": 0.0005560456847322487, |
| "loss": 3.9205, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.748035733505543, |
| "grad_norm": 0.6117699146270752, |
| "learning_rate": 0.0005557224437021872, |
| "loss": 3.943, |
| "step": 6950 |
| }, |
| { |
| "epoch": 0.7534172855451512, |
| "grad_norm": 0.5624613761901855, |
| "learning_rate": 0.0005553992026721258, |
| "loss": 3.9243, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7534172855451512, |
| "eval_accuracy": 0.33315843840158127, |
| "eval_loss": 3.8613877296447754, |
| "eval_runtime": 179.8814, |
| "eval_samples_per_second": 100.127, |
| "eval_steps_per_second": 6.26, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7587988375847594, |
| "grad_norm": 0.6362348794937134, |
| "learning_rate": 0.0005550759616420644, |
| "loss": 3.9187, |
| "step": 7050 |
| }, |
| { |
| "epoch": 0.7641803896243676, |
| "grad_norm": 0.6065320372581482, |
| "learning_rate": 0.000554752720612003, |
| "loss": 3.9133, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.7695619416639758, |
| "grad_norm": 0.6070377826690674, |
| "learning_rate": 0.0005544294795819415, |
| "loss": 3.9235, |
| "step": 7150 |
| }, |
| { |
| "epoch": 0.7749434937035841, |
| "grad_norm": 0.6300147771835327, |
| "learning_rate": 0.0005541062385518801, |
| "loss": 3.9364, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.7803250457431924, |
| "grad_norm": 0.6292157769203186, |
| "learning_rate": 0.0005537829975218188, |
| "loss": 3.9266, |
| "step": 7250 |
| }, |
| { |
| "epoch": 0.7857065977828006, |
| "grad_norm": 0.6281834840774536, |
| "learning_rate": 0.0005534597564917573, |
| "loss": 3.9215, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.7910881498224088, |
| "grad_norm": 0.6987627148628235, |
| "learning_rate": 0.0005531365154616959, |
| "loss": 3.932, |
| "step": 7350 |
| }, |
| { |
| "epoch": 0.796469701862017, |
| "grad_norm": 0.5451225638389587, |
| "learning_rate": 0.0005528132744316344, |
| "loss": 3.91, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.8018512539016253, |
| "grad_norm": 0.610156238079071, |
| "learning_rate": 0.0005524900334015731, |
| "loss": 3.9156, |
| "step": 7450 |
| }, |
| { |
| "epoch": 0.8072328059412335, |
| "grad_norm": 0.5881654620170593, |
| "learning_rate": 0.0005521667923715117, |
| "loss": 3.8981, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.8126143579808417, |
| "grad_norm": 0.6110484004020691, |
| "learning_rate": 0.0005518435513414502, |
| "loss": 3.9237, |
| "step": 7550 |
| }, |
| { |
| "epoch": 0.8179959100204499, |
| "grad_norm": 0.6429305076599121, |
| "learning_rate": 0.0005515203103113888, |
| "loss": 3.8959, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.8233774620600581, |
| "grad_norm": 0.6395870447158813, |
| "learning_rate": 0.0005511970692813274, |
| "loss": 3.9084, |
| "step": 7650 |
| }, |
| { |
| "epoch": 0.8287590140996663, |
| "grad_norm": 0.6113373041152954, |
| "learning_rate": 0.000550873828251266, |
| "loss": 3.9029, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.8341405661392746, |
| "grad_norm": 0.641963005065918, |
| "learning_rate": 0.0005505505872212045, |
| "loss": 3.883, |
| "step": 7750 |
| }, |
| { |
| "epoch": 0.8395221181788828, |
| "grad_norm": 0.6410965323448181, |
| "learning_rate": 0.0005502273461911432, |
| "loss": 3.8844, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.844903670218491, |
| "grad_norm": 0.604951024055481, |
| "learning_rate": 0.0005499041051610817, |
| "loss": 3.8943, |
| "step": 7850 |
| }, |
| { |
| "epoch": 0.8502852222580992, |
| "grad_norm": 0.6073367595672607, |
| "learning_rate": 0.0005495808641310204, |
| "loss": 3.9, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.8556667742977074, |
| "grad_norm": 0.5822165012359619, |
| "learning_rate": 0.0005492576231009589, |
| "loss": 3.898, |
| "step": 7950 |
| }, |
| { |
| "epoch": 0.8610483263373157, |
| "grad_norm": 0.618316650390625, |
| "learning_rate": 0.0005489343820708974, |
| "loss": 3.8774, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.8610483263373157, |
| "eval_accuracy": 0.3376425430816849, |
| "eval_loss": 3.820213556289673, |
| "eval_runtime": 179.9142, |
| "eval_samples_per_second": 100.109, |
| "eval_steps_per_second": 6.259, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.8664298783769239, |
| "grad_norm": 0.5604105591773987, |
| "learning_rate": 0.0005486111410408361, |
| "loss": 3.8929, |
| "step": 8050 |
| }, |
| { |
| "epoch": 0.8718114304165321, |
| "grad_norm": 0.6018927097320557, |
| "learning_rate": 0.0005482879000107746, |
| "loss": 3.8857, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.8771929824561403, |
| "grad_norm": 0.5374700427055359, |
| "learning_rate": 0.0005479646589807133, |
| "loss": 3.8744, |
| "step": 8150 |
| }, |
| { |
| "epoch": 0.8825745344957485, |
| "grad_norm": 0.5941512584686279, |
| "learning_rate": 0.0005476414179506518, |
| "loss": 3.8599, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.8879560865353568, |
| "grad_norm": 0.5862547159194946, |
| "learning_rate": 0.0005473181769205904, |
| "loss": 3.8706, |
| "step": 8250 |
| }, |
| { |
| "epoch": 0.8933376385749651, |
| "grad_norm": 0.5117703080177307, |
| "learning_rate": 0.000546994935890529, |
| "loss": 3.8804, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.8987191906145733, |
| "grad_norm": 0.5411306023597717, |
| "learning_rate": 0.0005466716948604677, |
| "loss": 3.8732, |
| "step": 8350 |
| }, |
| { |
| "epoch": 0.9041007426541815, |
| "grad_norm": 0.624138355255127, |
| "learning_rate": 0.0005463484538304062, |
| "loss": 3.8787, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.9094822946937897, |
| "grad_norm": 0.5695147514343262, |
| "learning_rate": 0.0005460252128003447, |
| "loss": 3.8548, |
| "step": 8450 |
| }, |
| { |
| "epoch": 0.9148638467333979, |
| "grad_norm": 0.648750901222229, |
| "learning_rate": 0.0005457019717702833, |
| "loss": 3.8585, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.9202453987730062, |
| "grad_norm": 0.6337014436721802, |
| "learning_rate": 0.0005453787307402219, |
| "loss": 3.8518, |
| "step": 8550 |
| }, |
| { |
| "epoch": 0.9256269508126144, |
| "grad_norm": 0.5613633990287781, |
| "learning_rate": 0.0005450619545307617, |
| "loss": 3.8631, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.9310085028522226, |
| "grad_norm": 0.6017595529556274, |
| "learning_rate": 0.0005447387135007003, |
| "loss": 3.8429, |
| "step": 8650 |
| }, |
| { |
| "epoch": 0.9363900548918308, |
| "grad_norm": 0.6081689596176147, |
| "learning_rate": 0.0005444154724706389, |
| "loss": 3.854, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.941771606931439, |
| "grad_norm": 0.5572118163108826, |
| "learning_rate": 0.0005440922314405775, |
| "loss": 3.8723, |
| "step": 8750 |
| }, |
| { |
| "epoch": 0.9471531589710472, |
| "grad_norm": 0.5897894501686096, |
| "learning_rate": 0.0005437689904105161, |
| "loss": 3.8473, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.9525347110106555, |
| "grad_norm": 0.5687997341156006, |
| "learning_rate": 0.0005434457493804546, |
| "loss": 3.8436, |
| "step": 8850 |
| }, |
| { |
| "epoch": 0.9579162630502637, |
| "grad_norm": 0.624033510684967, |
| "learning_rate": 0.0005431225083503932, |
| "loss": 3.84, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.9632978150898719, |
| "grad_norm": 0.5875102877616882, |
| "learning_rate": 0.0005427992673203318, |
| "loss": 3.866, |
| "step": 8950 |
| }, |
| { |
| "epoch": 0.9686793671294801, |
| "grad_norm": 0.6378977298736572, |
| "learning_rate": 0.0005424760262902704, |
| "loss": 3.8444, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.9686793671294801, |
| "eval_accuracy": 0.3414037800991588, |
| "eval_loss": 3.7785165309906006, |
| "eval_runtime": 179.7193, |
| "eval_samples_per_second": 100.217, |
| "eval_steps_per_second": 6.265, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.9740609191690883, |
| "grad_norm": 0.6483079195022583, |
| "learning_rate": 0.000542152785260209, |
| "loss": 3.8263, |
| "step": 9050 |
| }, |
| { |
| "epoch": 0.9794424712086965, |
| "grad_norm": 0.575869083404541, |
| "learning_rate": 0.0005418295442301476, |
| "loss": 3.8347, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.9848240232483048, |
| "grad_norm": 0.6090970635414124, |
| "learning_rate": 0.0005415063032000861, |
| "loss": 3.8418, |
| "step": 9150 |
| }, |
| { |
| "epoch": 0.9902055752879131, |
| "grad_norm": 0.5478109121322632, |
| "learning_rate": 0.0005411830621700248, |
| "loss": 3.8362, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.9955871273275213, |
| "grad_norm": 0.5703595280647278, |
| "learning_rate": 0.0005408598211399633, |
| "loss": 3.8135, |
| "step": 9250 |
| }, |
| { |
| "epoch": 1.0009686793671295, |
| "grad_norm": 0.6738930940628052, |
| "learning_rate": 0.0005405365801099019, |
| "loss": 3.8282, |
| "step": 9300 |
| }, |
| { |
| "epoch": 1.0063502314067376, |
| "grad_norm": 0.5832762718200684, |
| "learning_rate": 0.0005402133390798405, |
| "loss": 3.7677, |
| "step": 9350 |
| }, |
| { |
| "epoch": 1.011731783446346, |
| "grad_norm": 0.6429535150527954, |
| "learning_rate": 0.000539890098049779, |
| "loss": 3.7784, |
| "step": 9400 |
| }, |
| { |
| "epoch": 1.017113335485954, |
| "grad_norm": 0.5857776999473572, |
| "learning_rate": 0.0005395668570197177, |
| "loss": 3.753, |
| "step": 9450 |
| }, |
| { |
| "epoch": 1.0224948875255624, |
| "grad_norm": 0.553745448589325, |
| "learning_rate": 0.0005392436159896562, |
| "loss": 3.7717, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.0278764395651705, |
| "grad_norm": 0.592886209487915, |
| "learning_rate": 0.0005389203749595948, |
| "loss": 3.7701, |
| "step": 9550 |
| }, |
| { |
| "epoch": 1.0332579916047788, |
| "grad_norm": 0.5704110264778137, |
| "learning_rate": 0.0005385971339295334, |
| "loss": 3.7672, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.0386395436443872, |
| "grad_norm": 0.5219415426254272, |
| "learning_rate": 0.000538273892899472, |
| "loss": 3.7743, |
| "step": 9650 |
| }, |
| { |
| "epoch": 1.0440210956839953, |
| "grad_norm": 0.5501973628997803, |
| "learning_rate": 0.0005379506518694106, |
| "loss": 3.7665, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.0494026477236036, |
| "grad_norm": 0.5235220789909363, |
| "learning_rate": 0.0005376274108393491, |
| "loss": 3.772, |
| "step": 9750 |
| }, |
| { |
| "epoch": 1.0547841997632117, |
| "grad_norm": 0.6010895371437073, |
| "learning_rate": 0.0005373041698092877, |
| "loss": 3.7833, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.06016575180282, |
| "grad_norm": 0.5482034087181091, |
| "learning_rate": 0.0005369809287792263, |
| "loss": 3.7599, |
| "step": 9850 |
| }, |
| { |
| "epoch": 1.0655473038424281, |
| "grad_norm": 0.6096392869949341, |
| "learning_rate": 0.000536657687749165, |
| "loss": 3.7603, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.0709288558820365, |
| "grad_norm": 0.5938239693641663, |
| "learning_rate": 0.0005363344467191035, |
| "loss": 3.7484, |
| "step": 9950 |
| }, |
| { |
| "epoch": 1.0763104079216446, |
| "grad_norm": 0.6193915605545044, |
| "learning_rate": 0.000536011205689042, |
| "loss": 3.7673, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.0763104079216446, |
| "eval_accuracy": 0.3440351357362499, |
| "eval_loss": 3.752418279647827, |
| "eval_runtime": 179.8923, |
| "eval_samples_per_second": 100.121, |
| "eval_steps_per_second": 6.259, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.081691959961253, |
| "grad_norm": 0.6629969477653503, |
| "learning_rate": 0.0005356879646589807, |
| "loss": 3.773, |
| "step": 10050 |
| }, |
| { |
| "epoch": 1.087073512000861, |
| "grad_norm": 0.6844997406005859, |
| "learning_rate": 0.0005353647236289192, |
| "loss": 3.7536, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.0924550640404693, |
| "grad_norm": 0.5414447784423828, |
| "learning_rate": 0.0005350414825988579, |
| "loss": 3.7628, |
| "step": 10150 |
| }, |
| { |
| "epoch": 1.0978366160800774, |
| "grad_norm": 0.6185393929481506, |
| "learning_rate": 0.0005347182415687964, |
| "loss": 3.7544, |
| "step": 10200 |
| }, |
| { |
| "epoch": 1.1032181681196858, |
| "grad_norm": 0.6166486740112305, |
| "learning_rate": 0.000534395000538735, |
| "loss": 3.7621, |
| "step": 10250 |
| }, |
| { |
| "epoch": 1.1085997201592939, |
| "grad_norm": 0.5102859735488892, |
| "learning_rate": 0.0005340717595086736, |
| "loss": 3.7568, |
| "step": 10300 |
| }, |
| { |
| "epoch": 1.1139812721989022, |
| "grad_norm": 0.5167948007583618, |
| "learning_rate": 0.0005337485184786122, |
| "loss": 3.762, |
| "step": 10350 |
| }, |
| { |
| "epoch": 1.1193628242385103, |
| "grad_norm": 0.5715023279190063, |
| "learning_rate": 0.0005334252774485507, |
| "loss": 3.7745, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.1247443762781186, |
| "grad_norm": 0.6422619223594666, |
| "learning_rate": 0.0005331020364184894, |
| "loss": 3.7554, |
| "step": 10450 |
| }, |
| { |
| "epoch": 1.1301259283177267, |
| "grad_norm": 0.5923563838005066, |
| "learning_rate": 0.0005327787953884279, |
| "loss": 3.7443, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.135507480357335, |
| "grad_norm": 0.6288694143295288, |
| "learning_rate": 0.0005324555543583665, |
| "loss": 3.7642, |
| "step": 10550 |
| }, |
| { |
| "epoch": 1.1408890323969434, |
| "grad_norm": 0.626955509185791, |
| "learning_rate": 0.0005321323133283051, |
| "loss": 3.7581, |
| "step": 10600 |
| }, |
| { |
| "epoch": 1.1462705844365515, |
| "grad_norm": 0.6043279767036438, |
| "learning_rate": 0.0005318090722982436, |
| "loss": 3.7467, |
| "step": 10650 |
| }, |
| { |
| "epoch": 1.1516521364761596, |
| "grad_norm": 0.5194068551063538, |
| "learning_rate": 0.0005314922960887834, |
| "loss": 3.747, |
| "step": 10700 |
| }, |
| { |
| "epoch": 1.157033688515768, |
| "grad_norm": 0.5931900143623352, |
| "learning_rate": 0.0005311690550587221, |
| "loss": 3.7423, |
| "step": 10750 |
| }, |
| { |
| "epoch": 1.1624152405553763, |
| "grad_norm": 0.6204814910888672, |
| "learning_rate": 0.0005308458140286607, |
| "loss": 3.7472, |
| "step": 10800 |
| }, |
| { |
| "epoch": 1.1677967925949844, |
| "grad_norm": 0.5869725346565247, |
| "learning_rate": 0.0005305225729985993, |
| "loss": 3.7455, |
| "step": 10850 |
| }, |
| { |
| "epoch": 1.1731783446345927, |
| "grad_norm": 0.616137683391571, |
| "learning_rate": 0.0005301993319685378, |
| "loss": 3.7662, |
| "step": 10900 |
| }, |
| { |
| "epoch": 1.1785598966742008, |
| "grad_norm": 0.5727664232254028, |
| "learning_rate": 0.0005298760909384765, |
| "loss": 3.7566, |
| "step": 10950 |
| }, |
| { |
| "epoch": 1.1839414487138091, |
| "grad_norm": 0.5144981145858765, |
| "learning_rate": 0.000529552849908415, |
| "loss": 3.753, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.1839414487138091, |
| "eval_accuracy": 0.34677949037666805, |
| "eval_loss": 3.7220773696899414, |
| "eval_runtime": 179.7359, |
| "eval_samples_per_second": 100.208, |
| "eval_steps_per_second": 6.265, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.1893230007534172, |
| "grad_norm": 0.5506604909896851, |
| "learning_rate": 0.0005292296088783535, |
| "loss": 3.7472, |
| "step": 11050 |
| }, |
| { |
| "epoch": 1.1947045527930256, |
| "grad_norm": 0.5824575424194336, |
| "learning_rate": 0.0005289063678482922, |
| "loss": 3.753, |
| "step": 11100 |
| }, |
| { |
| "epoch": 1.2000861048326337, |
| "grad_norm": 0.5471564531326294, |
| "learning_rate": 0.0005285831268182307, |
| "loss": 3.7553, |
| "step": 11150 |
| }, |
| { |
| "epoch": 1.205467656872242, |
| "grad_norm": 0.5185365080833435, |
| "learning_rate": 0.0005282598857881694, |
| "loss": 3.7437, |
| "step": 11200 |
| }, |
| { |
| "epoch": 1.21084920891185, |
| "grad_norm": 0.595317006111145, |
| "learning_rate": 0.0005279366447581079, |
| "loss": 3.7284, |
| "step": 11250 |
| }, |
| { |
| "epoch": 1.2162307609514584, |
| "grad_norm": 0.53959721326828, |
| "learning_rate": 0.0005276134037280465, |
| "loss": 3.7602, |
| "step": 11300 |
| }, |
| { |
| "epoch": 1.2216123129910665, |
| "grad_norm": 0.5472400784492493, |
| "learning_rate": 0.0005272901626979851, |
| "loss": 3.7386, |
| "step": 11350 |
| }, |
| { |
| "epoch": 1.2269938650306749, |
| "grad_norm": 0.5732322335243225, |
| "learning_rate": 0.0005269669216679236, |
| "loss": 3.7517, |
| "step": 11400 |
| }, |
| { |
| "epoch": 1.232375417070283, |
| "grad_norm": 0.5511186122894287, |
| "learning_rate": 0.0005266436806378623, |
| "loss": 3.739, |
| "step": 11450 |
| }, |
| { |
| "epoch": 1.2377569691098913, |
| "grad_norm": 0.5673669576644897, |
| "learning_rate": 0.0005263204396078008, |
| "loss": 3.732, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.2431385211494996, |
| "grad_norm": 0.5002085566520691, |
| "learning_rate": 0.0005259971985777394, |
| "loss": 3.7578, |
| "step": 11550 |
| }, |
| { |
| "epoch": 1.2485200731891077, |
| "grad_norm": 0.5736905336380005, |
| "learning_rate": 0.000525673957547678, |
| "loss": 3.7448, |
| "step": 11600 |
| }, |
| { |
| "epoch": 1.2539016252287158, |
| "grad_norm": 0.5424861907958984, |
| "learning_rate": 0.0005253507165176167, |
| "loss": 3.7202, |
| "step": 11650 |
| }, |
| { |
| "epoch": 1.2592831772683242, |
| "grad_norm": 0.542267382144928, |
| "learning_rate": 0.0005250274754875552, |
| "loss": 3.7241, |
| "step": 11700 |
| }, |
| { |
| "epoch": 1.2646647293079325, |
| "grad_norm": 0.5920253396034241, |
| "learning_rate": 0.0005247042344574938, |
| "loss": 3.7427, |
| "step": 11750 |
| }, |
| { |
| "epoch": 1.2700462813475406, |
| "grad_norm": 0.5953553915023804, |
| "learning_rate": 0.0005243809934274323, |
| "loss": 3.7356, |
| "step": 11800 |
| }, |
| { |
| "epoch": 1.275427833387149, |
| "grad_norm": 0.574566662311554, |
| "learning_rate": 0.0005240577523973709, |
| "loss": 3.7172, |
| "step": 11850 |
| }, |
| { |
| "epoch": 1.280809385426757, |
| "grad_norm": 0.6085739731788635, |
| "learning_rate": 0.0005237345113673095, |
| "loss": 3.7379, |
| "step": 11900 |
| }, |
| { |
| "epoch": 1.2861909374663654, |
| "grad_norm": 0.5501704812049866, |
| "learning_rate": 0.0005234112703372481, |
| "loss": 3.7723, |
| "step": 11950 |
| }, |
| { |
| "epoch": 1.2915724895059735, |
| "grad_norm": 0.5961480140686035, |
| "learning_rate": 0.0005230880293071867, |
| "loss": 3.7378, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.2915724895059735, |
| "eval_accuracy": 0.34941084601375916, |
| "eval_loss": 3.695894241333008, |
| "eval_runtime": 179.7781, |
| "eval_samples_per_second": 100.185, |
| "eval_steps_per_second": 6.263, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.2969540415455818, |
| "grad_norm": 0.587378740310669, |
| "learning_rate": 0.0005227647882771253, |
| "loss": 3.7458, |
| "step": 12050 |
| }, |
| { |
| "epoch": 1.30233559358519, |
| "grad_norm": 0.5950103402137756, |
| "learning_rate": 0.0005224415472470639, |
| "loss": 3.7297, |
| "step": 12100 |
| }, |
| { |
| "epoch": 1.3077171456247982, |
| "grad_norm": 0.5634168386459351, |
| "learning_rate": 0.0005221183062170024, |
| "loss": 3.7124, |
| "step": 12150 |
| }, |
| { |
| "epoch": 1.3130986976644063, |
| "grad_norm": 0.5455742478370667, |
| "learning_rate": 0.0005217950651869409, |
| "loss": 3.7115, |
| "step": 12200 |
| }, |
| { |
| "epoch": 1.3184802497040147, |
| "grad_norm": 0.5473319888114929, |
| "learning_rate": 0.0005214718241568796, |
| "loss": 3.7046, |
| "step": 12250 |
| }, |
| { |
| "epoch": 1.3238618017436228, |
| "grad_norm": 0.5351397395133972, |
| "learning_rate": 0.0005211485831268182, |
| "loss": 3.7516, |
| "step": 12300 |
| }, |
| { |
| "epoch": 1.329243353783231, |
| "grad_norm": 0.5263974070549011, |
| "learning_rate": 0.0005208253420967568, |
| "loss": 3.7259, |
| "step": 12350 |
| }, |
| { |
| "epoch": 1.3346249058228392, |
| "grad_norm": 0.544121503829956, |
| "learning_rate": 0.0005205021010666953, |
| "loss": 3.7135, |
| "step": 12400 |
| }, |
| { |
| "epoch": 1.3400064578624475, |
| "grad_norm": 0.5584757924079895, |
| "learning_rate": 0.0005201788600366339, |
| "loss": 3.7257, |
| "step": 12450 |
| }, |
| { |
| "epoch": 1.3453880099020559, |
| "grad_norm": 0.5996146202087402, |
| "learning_rate": 0.0005198556190065725, |
| "loss": 3.7364, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.350769561941664, |
| "grad_norm": 0.5599164366722107, |
| "learning_rate": 0.0005195323779765112, |
| "loss": 3.7402, |
| "step": 12550 |
| }, |
| { |
| "epoch": 1.356151113981272, |
| "grad_norm": 0.6167038083076477, |
| "learning_rate": 0.0005192091369464497, |
| "loss": 3.7381, |
| "step": 12600 |
| }, |
| { |
| "epoch": 1.3615326660208804, |
| "grad_norm": 0.5943508744239807, |
| "learning_rate": 0.0005188858959163882, |
| "loss": 3.7255, |
| "step": 12650 |
| }, |
| { |
| "epoch": 1.3669142180604887, |
| "grad_norm": 0.5350813269615173, |
| "learning_rate": 0.0005185626548863269, |
| "loss": 3.7235, |
| "step": 12700 |
| }, |
| { |
| "epoch": 1.3722957701000968, |
| "grad_norm": 0.5343997478485107, |
| "learning_rate": 0.0005182458786768667, |
| "loss": 3.7154, |
| "step": 12750 |
| }, |
| { |
| "epoch": 1.3776773221397052, |
| "grad_norm": 0.6642977595329285, |
| "learning_rate": 0.0005179226376468052, |
| "loss": 3.7167, |
| "step": 12800 |
| }, |
| { |
| "epoch": 1.3830588741793133, |
| "grad_norm": 0.56437748670578, |
| "learning_rate": 0.0005175993966167438, |
| "loss": 3.7007, |
| "step": 12850 |
| }, |
| { |
| "epoch": 1.3884404262189216, |
| "grad_norm": 0.5799341797828674, |
| "learning_rate": 0.0005172761555866824, |
| "loss": 3.7025, |
| "step": 12900 |
| }, |
| { |
| "epoch": 1.3938219782585297, |
| "grad_norm": 0.6053862571716309, |
| "learning_rate": 0.0005169529145566211, |
| "loss": 3.6903, |
| "step": 12950 |
| }, |
| { |
| "epoch": 1.399203530298138, |
| "grad_norm": 0.5302448868751526, |
| "learning_rate": 0.0005166296735265596, |
| "loss": 3.711, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.399203530298138, |
| "eval_accuracy": 0.3507369545095349, |
| "eval_loss": 3.6787874698638916, |
| "eval_runtime": 179.7451, |
| "eval_samples_per_second": 100.203, |
| "eval_steps_per_second": 6.264, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.4045850823377461, |
| "grad_norm": 0.5383056402206421, |
| "learning_rate": 0.0005163064324964982, |
| "loss": 3.7115, |
| "step": 13050 |
| }, |
| { |
| "epoch": 1.4099666343773545, |
| "grad_norm": 0.5403004288673401, |
| "learning_rate": 0.0005159831914664367, |
| "loss": 3.7064, |
| "step": 13100 |
| }, |
| { |
| "epoch": 1.4153481864169626, |
| "grad_norm": 0.5427159070968628, |
| "learning_rate": 0.0005156599504363753, |
| "loss": 3.7023, |
| "step": 13150 |
| }, |
| { |
| "epoch": 1.420729738456571, |
| "grad_norm": 0.5563752055168152, |
| "learning_rate": 0.000515336709406314, |
| "loss": 3.7225, |
| "step": 13200 |
| }, |
| { |
| "epoch": 1.426111290496179, |
| "grad_norm": 0.6402376890182495, |
| "learning_rate": 0.0005150134683762525, |
| "loss": 3.7011, |
| "step": 13250 |
| }, |
| { |
| "epoch": 1.4314928425357873, |
| "grad_norm": 0.5857889652252197, |
| "learning_rate": 0.0005146902273461911, |
| "loss": 3.7031, |
| "step": 13300 |
| }, |
| { |
| "epoch": 1.4368743945753955, |
| "grad_norm": 0.6158170104026794, |
| "learning_rate": 0.0005143669863161297, |
| "loss": 3.6897, |
| "step": 13350 |
| }, |
| { |
| "epoch": 1.4422559466150038, |
| "grad_norm": 0.6112419962882996, |
| "learning_rate": 0.0005140437452860683, |
| "loss": 3.7072, |
| "step": 13400 |
| }, |
| { |
| "epoch": 1.447637498654612, |
| "grad_norm": 0.5744062662124634, |
| "learning_rate": 0.0005137205042560069, |
| "loss": 3.7056, |
| "step": 13450 |
| }, |
| { |
| "epoch": 1.4530190506942202, |
| "grad_norm": 0.6232595443725586, |
| "learning_rate": 0.0005133972632259455, |
| "loss": 3.7002, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.4584006027338283, |
| "grad_norm": 0.567087709903717, |
| "learning_rate": 0.000513074022195884, |
| "loss": 3.704, |
| "step": 13550 |
| }, |
| { |
| "epoch": 1.4637821547734367, |
| "grad_norm": 0.5451653003692627, |
| "learning_rate": 0.0005127507811658226, |
| "loss": 3.7109, |
| "step": 13600 |
| }, |
| { |
| "epoch": 1.469163706813045, |
| "grad_norm": 0.671532154083252, |
| "learning_rate": 0.0005124275401357612, |
| "loss": 3.6943, |
| "step": 13650 |
| }, |
| { |
| "epoch": 1.474545258852653, |
| "grad_norm": 0.542718768119812, |
| "learning_rate": 0.0005121042991056997, |
| "loss": 3.6789, |
| "step": 13700 |
| }, |
| { |
| "epoch": 1.4799268108922612, |
| "grad_norm": 0.5489877462387085, |
| "learning_rate": 0.0005117810580756384, |
| "loss": 3.714, |
| "step": 13750 |
| }, |
| { |
| "epoch": 1.4853083629318695, |
| "grad_norm": 0.5520055890083313, |
| "learning_rate": 0.0005114578170455769, |
| "loss": 3.6945, |
| "step": 13800 |
| }, |
| { |
| "epoch": 1.4906899149714778, |
| "grad_norm": 0.5908713340759277, |
| "learning_rate": 0.0005111345760155156, |
| "loss": 3.6954, |
| "step": 13850 |
| }, |
| { |
| "epoch": 1.496071467011086, |
| "grad_norm": 0.5579512119293213, |
| "learning_rate": 0.0005108113349854541, |
| "loss": 3.7161, |
| "step": 13900 |
| }, |
| { |
| "epoch": 1.501453019050694, |
| "grad_norm": 0.5487799048423767, |
| "learning_rate": 0.0005104880939553926, |
| "loss": 3.7058, |
| "step": 13950 |
| }, |
| { |
| "epoch": 1.5068345710903024, |
| "grad_norm": 0.5599663257598877, |
| "learning_rate": 0.0005101648529253313, |
| "loss": 3.7049, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.5068345710903024, |
| "eval_accuracy": 0.35338612922022755, |
| "eval_loss": 3.6561355590820312, |
| "eval_runtime": 179.7601, |
| "eval_samples_per_second": 100.195, |
| "eval_steps_per_second": 6.264, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.5122161231299107, |
| "grad_norm": 0.5696370601654053, |
| "learning_rate": 0.0005098416118952699, |
| "loss": 3.6653, |
| "step": 14050 |
| }, |
| { |
| "epoch": 1.5175976751695188, |
| "grad_norm": 0.7141689658164978, |
| "learning_rate": 0.0005095183708652085, |
| "loss": 3.6826, |
| "step": 14100 |
| }, |
| { |
| "epoch": 1.5229792272091272, |
| "grad_norm": 0.6532941460609436, |
| "learning_rate": 0.000509195129835147, |
| "loss": 3.7016, |
| "step": 14150 |
| }, |
| { |
| "epoch": 1.5283607792487355, |
| "grad_norm": 0.5246692299842834, |
| "learning_rate": 0.0005088718888050856, |
| "loss": 3.6981, |
| "step": 14200 |
| }, |
| { |
| "epoch": 1.5337423312883436, |
| "grad_norm": 0.5713950991630554, |
| "learning_rate": 0.0005085486477750242, |
| "loss": 3.6811, |
| "step": 14250 |
| }, |
| { |
| "epoch": 1.5391238833279517, |
| "grad_norm": 0.5982765555381775, |
| "learning_rate": 0.0005082254067449629, |
| "loss": 3.7083, |
| "step": 14300 |
| }, |
| { |
| "epoch": 1.54450543536756, |
| "grad_norm": 0.5973095893859863, |
| "learning_rate": 0.0005079021657149014, |
| "loss": 3.6941, |
| "step": 14350 |
| }, |
| { |
| "epoch": 1.5498869874071683, |
| "grad_norm": 0.555167555809021, |
| "learning_rate": 0.0005075789246848399, |
| "loss": 3.6879, |
| "step": 14400 |
| }, |
| { |
| "epoch": 1.5552685394467765, |
| "grad_norm": 0.5746867060661316, |
| "learning_rate": 0.0005072556836547785, |
| "loss": 3.6938, |
| "step": 14450 |
| }, |
| { |
| "epoch": 1.5606500914863846, |
| "grad_norm": 0.5191563367843628, |
| "learning_rate": 0.0005069324426247171, |
| "loss": 3.6978, |
| "step": 14500 |
| }, |
| { |
| "epoch": 1.566031643525993, |
| "grad_norm": 0.5576121807098389, |
| "learning_rate": 0.0005066092015946557, |
| "loss": 3.7036, |
| "step": 14550 |
| }, |
| { |
| "epoch": 1.5714131955656012, |
| "grad_norm": 0.5687995553016663, |
| "learning_rate": 0.0005062859605645943, |
| "loss": 3.6899, |
| "step": 14600 |
| }, |
| { |
| "epoch": 1.5767947476052093, |
| "grad_norm": 0.6132740378379822, |
| "learning_rate": 0.0005059627195345329, |
| "loss": 3.6664, |
| "step": 14650 |
| }, |
| { |
| "epoch": 1.5821762996448174, |
| "grad_norm": 0.5791540145874023, |
| "learning_rate": 0.0005056394785044715, |
| "loss": 3.6893, |
| "step": 14700 |
| }, |
| { |
| "epoch": 1.5875578516844258, |
| "grad_norm": 0.5374217629432678, |
| "learning_rate": 0.00050531623747441, |
| "loss": 3.7027, |
| "step": 14750 |
| }, |
| { |
| "epoch": 1.592939403724034, |
| "grad_norm": 0.5353657603263855, |
| "learning_rate": 0.0005049994612649499, |
| "loss": 3.6779, |
| "step": 14800 |
| }, |
| { |
| "epoch": 1.5983209557636422, |
| "grad_norm": 0.6020593047142029, |
| "learning_rate": 0.0005046762202348884, |
| "loss": 3.6988, |
| "step": 14850 |
| }, |
| { |
| "epoch": 1.6037025078032503, |
| "grad_norm": 0.6164742708206177, |
| "learning_rate": 0.000504352979204827, |
| "loss": 3.711, |
| "step": 14900 |
| }, |
| { |
| "epoch": 1.6090840598428586, |
| "grad_norm": 0.5817080140113831, |
| "learning_rate": 0.0005040297381747656, |
| "loss": 3.6941, |
| "step": 14950 |
| }, |
| { |
| "epoch": 1.614465611882467, |
| "grad_norm": 0.5781374573707581, |
| "learning_rate": 0.0005037064971447042, |
| "loss": 3.6648, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.614465611882467, |
| "eval_accuracy": 0.35519965457073904, |
| "eval_loss": 3.6373958587646484, |
| "eval_runtime": 179.8191, |
| "eval_samples_per_second": 100.162, |
| "eval_steps_per_second": 6.262, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.619847163922075, |
| "grad_norm": 0.5674581527709961, |
| "learning_rate": 0.0005033832561146428, |
| "loss": 3.7027, |
| "step": 15050 |
| }, |
| { |
| "epoch": 1.6252287159616834, |
| "grad_norm": 0.5419674515724182, |
| "learning_rate": 0.0005030600150845813, |
| "loss": 3.6874, |
| "step": 15100 |
| }, |
| { |
| "epoch": 1.6306102680012917, |
| "grad_norm": 0.5471088886260986, |
| "learning_rate": 0.00050273677405452, |
| "loss": 3.6781, |
| "step": 15150 |
| }, |
| { |
| "epoch": 1.6359918200408998, |
| "grad_norm": 0.5623396039009094, |
| "learning_rate": 0.0005024135330244585, |
| "loss": 3.6821, |
| "step": 15200 |
| }, |
| { |
| "epoch": 1.641373372080508, |
| "grad_norm": 0.5248699188232422, |
| "learning_rate": 0.0005020902919943972, |
| "loss": 3.6966, |
| "step": 15250 |
| }, |
| { |
| "epoch": 1.6467549241201163, |
| "grad_norm": 0.5599279403686523, |
| "learning_rate": 0.0005017670509643357, |
| "loss": 3.6825, |
| "step": 15300 |
| }, |
| { |
| "epoch": 1.6521364761597246, |
| "grad_norm": 0.5801636576652527, |
| "learning_rate": 0.0005014438099342743, |
| "loss": 3.6487, |
| "step": 15350 |
| }, |
| { |
| "epoch": 1.6575180281993327, |
| "grad_norm": 0.5466539263725281, |
| "learning_rate": 0.0005011205689042129, |
| "loss": 3.6799, |
| "step": 15400 |
| }, |
| { |
| "epoch": 1.6628995802389408, |
| "grad_norm": 0.5191233158111572, |
| "learning_rate": 0.0005007973278741514, |
| "loss": 3.6652, |
| "step": 15450 |
| }, |
| { |
| "epoch": 1.6682811322785491, |
| "grad_norm": 0.538023829460144, |
| "learning_rate": 0.00050047408684409, |
| "loss": 3.6783, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.6736626843181575, |
| "grad_norm": 0.5874162316322327, |
| "learning_rate": 0.0005001508458140286, |
| "loss": 3.6757, |
| "step": 15550 |
| }, |
| { |
| "epoch": 1.6790442363577656, |
| "grad_norm": 0.5796945691108704, |
| "learning_rate": 0.0004998276047839673, |
| "loss": 3.6699, |
| "step": 15600 |
| }, |
| { |
| "epoch": 1.6844257883973737, |
| "grad_norm": 0.6127719879150391, |
| "learning_rate": 0.0004995043637539058, |
| "loss": 3.6761, |
| "step": 15650 |
| }, |
| { |
| "epoch": 1.689807340436982, |
| "grad_norm": 0.5545514822006226, |
| "learning_rate": 0.0004991811227238443, |
| "loss": 3.6704, |
| "step": 15700 |
| }, |
| { |
| "epoch": 1.6951888924765903, |
| "grad_norm": 0.59245765209198, |
| "learning_rate": 0.0004988578816937829, |
| "loss": 3.6674, |
| "step": 15750 |
| }, |
| { |
| "epoch": 1.7005704445161984, |
| "grad_norm": 0.5616726279258728, |
| "learning_rate": 0.0004985346406637215, |
| "loss": 3.6786, |
| "step": 15800 |
| }, |
| { |
| "epoch": 1.7059519965558065, |
| "grad_norm": 0.5251116156578064, |
| "learning_rate": 0.0004982113996336602, |
| "loss": 3.6491, |
| "step": 15850 |
| }, |
| { |
| "epoch": 1.7113335485954149, |
| "grad_norm": 0.5580370426177979, |
| "learning_rate": 0.0004978881586035987, |
| "loss": 3.6599, |
| "step": 15900 |
| }, |
| { |
| "epoch": 1.7167151006350232, |
| "grad_norm": 0.5888770818710327, |
| "learning_rate": 0.0004975649175735373, |
| "loss": 3.6759, |
| "step": 15950 |
| }, |
| { |
| "epoch": 1.7220966526746313, |
| "grad_norm": 0.561681866645813, |
| "learning_rate": 0.0004972416765434759, |
| "loss": 3.6552, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.7220966526746313, |
| "eval_accuracy": 0.3567330727764648, |
| "eval_loss": 3.6214287281036377, |
| "eval_runtime": 179.9334, |
| "eval_samples_per_second": 100.098, |
| "eval_steps_per_second": 6.258, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.7274782047142396, |
| "grad_norm": 0.5320652723312378, |
| "learning_rate": 0.0004969184355134145, |
| "loss": 3.6572, |
| "step": 16050 |
| }, |
| { |
| "epoch": 1.732859756753848, |
| "grad_norm": 0.5540884733200073, |
| "learning_rate": 0.0004965951944833531, |
| "loss": 3.6749, |
| "step": 16100 |
| }, |
| { |
| "epoch": 1.738241308793456, |
| "grad_norm": 0.6164997220039368, |
| "learning_rate": 0.0004962719534532916, |
| "loss": 3.6743, |
| "step": 16150 |
| }, |
| { |
| "epoch": 1.7436228608330642, |
| "grad_norm": 0.5549599528312683, |
| "learning_rate": 0.0004959487124232302, |
| "loss": 3.6486, |
| "step": 16200 |
| }, |
| { |
| "epoch": 1.7490044128726725, |
| "grad_norm": 0.5784186720848083, |
| "learning_rate": 0.0004956254713931688, |
| "loss": 3.655, |
| "step": 16250 |
| }, |
| { |
| "epoch": 1.7543859649122808, |
| "grad_norm": 0.6285885572433472, |
| "learning_rate": 0.0004953022303631074, |
| "loss": 3.6737, |
| "step": 16300 |
| }, |
| { |
| "epoch": 1.759767516951889, |
| "grad_norm": 0.5278353095054626, |
| "learning_rate": 0.0004949789893330459, |
| "loss": 3.6691, |
| "step": 16350 |
| }, |
| { |
| "epoch": 1.765149068991497, |
| "grad_norm": 0.5532869100570679, |
| "learning_rate": 0.0004946557483029846, |
| "loss": 3.6645, |
| "step": 16400 |
| }, |
| { |
| "epoch": 1.7705306210311054, |
| "grad_norm": 0.5599241852760315, |
| "learning_rate": 0.0004943325072729231, |
| "loss": 3.6569, |
| "step": 16450 |
| }, |
| { |
| "epoch": 1.7759121730707137, |
| "grad_norm": 0.5885684490203857, |
| "learning_rate": 0.0004940092662428617, |
| "loss": 3.6504, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.7812937251103218, |
| "grad_norm": 0.5403637290000916, |
| "learning_rate": 0.0004936860252128003, |
| "loss": 3.667, |
| "step": 16550 |
| }, |
| { |
| "epoch": 1.78667527714993, |
| "grad_norm": 0.5659014582633972, |
| "learning_rate": 0.0004933627841827388, |
| "loss": 3.6669, |
| "step": 16600 |
| }, |
| { |
| "epoch": 1.7920568291895382, |
| "grad_norm": 0.5648633241653442, |
| "learning_rate": 0.0004930395431526775, |
| "loss": 3.6579, |
| "step": 16650 |
| }, |
| { |
| "epoch": 1.7974383812291466, |
| "grad_norm": 0.5897268056869507, |
| "learning_rate": 0.0004927163021226161, |
| "loss": 3.6617, |
| "step": 16700 |
| }, |
| { |
| "epoch": 1.8028199332687547, |
| "grad_norm": 0.5169820189476013, |
| "learning_rate": 0.0004923930610925547, |
| "loss": 3.6545, |
| "step": 16750 |
| }, |
| { |
| "epoch": 1.8082014853083628, |
| "grad_norm": 0.5676159858703613, |
| "learning_rate": 0.0004920698200624932, |
| "loss": 3.6502, |
| "step": 16800 |
| }, |
| { |
| "epoch": 1.813583037347971, |
| "grad_norm": 0.629549503326416, |
| "learning_rate": 0.000491753043853033, |
| "loss": 3.6371, |
| "step": 16850 |
| }, |
| { |
| "epoch": 1.8189645893875794, |
| "grad_norm": 0.5634384155273438, |
| "learning_rate": 0.0004914298028229717, |
| "loss": 3.6592, |
| "step": 16900 |
| }, |
| { |
| "epoch": 1.8243461414271875, |
| "grad_norm": 0.5446241497993469, |
| "learning_rate": 0.0004911065617929102, |
| "loss": 3.6693, |
| "step": 16950 |
| }, |
| { |
| "epoch": 1.8297276934667959, |
| "grad_norm": 0.5624287128448486, |
| "learning_rate": 0.0004907833207628487, |
| "loss": 3.6525, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.8297276934667959, |
| "eval_accuracy": 0.3582982176254324, |
| "eval_loss": 3.606470823287964, |
| "eval_runtime": 179.8801, |
| "eval_samples_per_second": 100.128, |
| "eval_steps_per_second": 6.26, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.8351092455064042, |
| "grad_norm": 0.5610666275024414, |
| "learning_rate": 0.0004904600797327874, |
| "loss": 3.6374, |
| "step": 17050 |
| }, |
| { |
| "epoch": 1.8404907975460123, |
| "grad_norm": 0.5537661910057068, |
| "learning_rate": 0.0004901368387027259, |
| "loss": 3.6434, |
| "step": 17100 |
| }, |
| { |
| "epoch": 1.8458723495856204, |
| "grad_norm": 0.5830965638160706, |
| "learning_rate": 0.0004898135976726646, |
| "loss": 3.6273, |
| "step": 17150 |
| }, |
| { |
| "epoch": 1.8512539016252287, |
| "grad_norm": 0.635771632194519, |
| "learning_rate": 0.0004894903566426031, |
| "loss": 3.6567, |
| "step": 17200 |
| }, |
| { |
| "epoch": 1.856635453664837, |
| "grad_norm": 0.6146743297576904, |
| "learning_rate": 0.0004891671156125417, |
| "loss": 3.6323, |
| "step": 17250 |
| }, |
| { |
| "epoch": 1.8620170057044452, |
| "grad_norm": 0.5971278548240662, |
| "learning_rate": 0.0004888438745824803, |
| "loss": 3.6194, |
| "step": 17300 |
| }, |
| { |
| "epoch": 1.8673985577440533, |
| "grad_norm": 0.4955149292945862, |
| "learning_rate": 0.0004885206335524189, |
| "loss": 3.6616, |
| "step": 17350 |
| }, |
| { |
| "epoch": 1.8727801097836616, |
| "grad_norm": 0.5917460322380066, |
| "learning_rate": 0.0004881973925223575, |
| "loss": 3.6489, |
| "step": 17400 |
| }, |
| { |
| "epoch": 1.87816166182327, |
| "grad_norm": 0.5226415991783142, |
| "learning_rate": 0.0004878806163128973, |
| "loss": 3.6379, |
| "step": 17450 |
| }, |
| { |
| "epoch": 1.883543213862878, |
| "grad_norm": 0.5437645316123962, |
| "learning_rate": 0.00048755737528283585, |
| "loss": 3.6462, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.8889247659024861, |
| "grad_norm": 0.5848841667175293, |
| "learning_rate": 0.00048723413425277444, |
| "loss": 3.6338, |
| "step": 17550 |
| }, |
| { |
| "epoch": 1.8943063179420945, |
| "grad_norm": 0.5598640441894531, |
| "learning_rate": 0.00048691089322271304, |
| "loss": 3.6487, |
| "step": 17600 |
| }, |
| { |
| "epoch": 1.8996878699817028, |
| "grad_norm": 0.626156747341156, |
| "learning_rate": 0.0004865876521926516, |
| "loss": 3.6352, |
| "step": 17650 |
| }, |
| { |
| "epoch": 1.905069422021311, |
| "grad_norm": 0.589815080165863, |
| "learning_rate": 0.0004862644111625902, |
| "loss": 3.6364, |
| "step": 17700 |
| }, |
| { |
| "epoch": 1.910450974060919, |
| "grad_norm": 0.6275328993797302, |
| "learning_rate": 0.0004859411701325288, |
| "loss": 3.6526, |
| "step": 17750 |
| }, |
| { |
| "epoch": 1.9158325261005273, |
| "grad_norm": 0.5926043391227722, |
| "learning_rate": 0.00048561792910246736, |
| "loss": 3.6384, |
| "step": 17800 |
| }, |
| { |
| "epoch": 1.9212140781401357, |
| "grad_norm": 0.5851974487304688, |
| "learning_rate": 0.00048529468807240596, |
| "loss": 3.6396, |
| "step": 17850 |
| }, |
| { |
| "epoch": 1.9265956301797438, |
| "grad_norm": 0.5288086533546448, |
| "learning_rate": 0.0004849714470423445, |
| "loss": 3.6497, |
| "step": 17900 |
| }, |
| { |
| "epoch": 1.931977182219352, |
| "grad_norm": 0.5444283485412598, |
| "learning_rate": 0.0004846482060122831, |
| "loss": 3.6326, |
| "step": 17950 |
| }, |
| { |
| "epoch": 1.9373587342589604, |
| "grad_norm": 0.507646918296814, |
| "learning_rate": 0.00048432496498222174, |
| "loss": 3.6271, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.9373587342589604, |
| "eval_accuracy": 0.35979230348576935, |
| "eval_loss": 3.591550827026367, |
| "eval_runtime": 179.8657, |
| "eval_samples_per_second": 100.136, |
| "eval_steps_per_second": 6.26, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.9427402862985685, |
| "grad_norm": 0.5543516278266907, |
| "learning_rate": 0.0004840017239521603, |
| "loss": 3.6373, |
| "step": 18050 |
| }, |
| { |
| "epoch": 1.9481218383381766, |
| "grad_norm": 0.5450665950775146, |
| "learning_rate": 0.0004836784829220989, |
| "loss": 3.6373, |
| "step": 18100 |
| }, |
| { |
| "epoch": 1.953503390377785, |
| "grad_norm": 0.5807346701622009, |
| "learning_rate": 0.00048335524189203747, |
| "loss": 3.6481, |
| "step": 18150 |
| }, |
| { |
| "epoch": 1.9588849424173933, |
| "grad_norm": 0.5595925450325012, |
| "learning_rate": 0.000483032000861976, |
| "loss": 3.6223, |
| "step": 18200 |
| }, |
| { |
| "epoch": 1.9642664944570014, |
| "grad_norm": 0.6111944913864136, |
| "learning_rate": 0.0004827087598319146, |
| "loss": 3.6431, |
| "step": 18250 |
| }, |
| { |
| "epoch": 1.9696480464966095, |
| "grad_norm": 0.5740875601768494, |
| "learning_rate": 0.00048238551880185325, |
| "loss": 3.6397, |
| "step": 18300 |
| }, |
| { |
| "epoch": 1.9750295985362178, |
| "grad_norm": 0.5665279626846313, |
| "learning_rate": 0.0004820622777717918, |
| "loss": 3.6398, |
| "step": 18350 |
| }, |
| { |
| "epoch": 1.9804111505758262, |
| "grad_norm": 0.5787563323974609, |
| "learning_rate": 0.0004817390367417304, |
| "loss": 3.6241, |
| "step": 18400 |
| }, |
| { |
| "epoch": 1.9857927026154343, |
| "grad_norm": 0.5683515071868896, |
| "learning_rate": 0.00048141579571166893, |
| "loss": 3.6398, |
| "step": 18450 |
| }, |
| { |
| "epoch": 1.9911742546550424, |
| "grad_norm": 0.5649462938308716, |
| "learning_rate": 0.0004810925546816075, |
| "loss": 3.6431, |
| "step": 18500 |
| }, |
| { |
| "epoch": 1.9965558066946507, |
| "grad_norm": 0.5490776300430298, |
| "learning_rate": 0.0004807693136515461, |
| "loss": 3.6415, |
| "step": 18550 |
| }, |
| { |
| "epoch": 2.001937358734259, |
| "grad_norm": 0.5904884934425354, |
| "learning_rate": 0.0004804460726214847, |
| "loss": 3.5982, |
| "step": 18600 |
| }, |
| { |
| "epoch": 2.007318910773867, |
| "grad_norm": 0.5800067186355591, |
| "learning_rate": 0.0004801228315914233, |
| "loss": 3.5538, |
| "step": 18650 |
| }, |
| { |
| "epoch": 2.0127004628134753, |
| "grad_norm": 0.5690127611160278, |
| "learning_rate": 0.0004797995905613619, |
| "loss": 3.5412, |
| "step": 18700 |
| }, |
| { |
| "epoch": 2.018082014853084, |
| "grad_norm": 0.5800209641456604, |
| "learning_rate": 0.00047947634953130044, |
| "loss": 3.554, |
| "step": 18750 |
| }, |
| { |
| "epoch": 2.023463566892692, |
| "grad_norm": 0.595440685749054, |
| "learning_rate": 0.00047915310850123904, |
| "loss": 3.538, |
| "step": 18800 |
| }, |
| { |
| "epoch": 2.0288451189323, |
| "grad_norm": 0.5423743724822998, |
| "learning_rate": 0.0004788298674711777, |
| "loss": 3.544, |
| "step": 18850 |
| }, |
| { |
| "epoch": 2.034226670971908, |
| "grad_norm": 0.5956675410270691, |
| "learning_rate": 0.00047850662644111623, |
| "loss": 3.5531, |
| "step": 18900 |
| }, |
| { |
| "epoch": 2.0396082230115167, |
| "grad_norm": 0.531262993812561, |
| "learning_rate": 0.0004781833854110548, |
| "loss": 3.5422, |
| "step": 18950 |
| }, |
| { |
| "epoch": 2.044989775051125, |
| "grad_norm": 0.6084029674530029, |
| "learning_rate": 0.00047786014438099336, |
| "loss": 3.5643, |
| "step": 19000 |
| }, |
| { |
| "epoch": 2.044989775051125, |
| "eval_accuracy": 0.3612721558178026, |
| "eval_loss": 3.5829977989196777, |
| "eval_runtime": 179.5266, |
| "eval_samples_per_second": 100.325, |
| "eval_steps_per_second": 6.272, |
| "step": 19000 |
| }, |
| { |
| "epoch": 2.050371327090733, |
| "grad_norm": 0.5787180662155151, |
| "learning_rate": 0.00047753690335093196, |
| "loss": 3.5347, |
| "step": 19050 |
| }, |
| { |
| "epoch": 2.055752879130341, |
| "grad_norm": 0.5521721839904785, |
| "learning_rate": 0.00047721366232087055, |
| "loss": 3.555, |
| "step": 19100 |
| }, |
| { |
| "epoch": 2.0611344311699495, |
| "grad_norm": 0.584023118019104, |
| "learning_rate": 0.00047689042129080915, |
| "loss": 3.556, |
| "step": 19150 |
| }, |
| { |
| "epoch": 2.0665159832095576, |
| "grad_norm": 0.6237092018127441, |
| "learning_rate": 0.00047656718026074774, |
| "loss": 3.5654, |
| "step": 19200 |
| }, |
| { |
| "epoch": 2.0718975352491658, |
| "grad_norm": 0.5408397316932678, |
| "learning_rate": 0.00047624393923068634, |
| "loss": 3.5346, |
| "step": 19250 |
| }, |
| { |
| "epoch": 2.0772790872887743, |
| "grad_norm": 0.6254966855049133, |
| "learning_rate": 0.0004759206982006249, |
| "loss": 3.5376, |
| "step": 19300 |
| }, |
| { |
| "epoch": 2.0826606393283824, |
| "grad_norm": 0.5704235434532166, |
| "learning_rate": 0.00047559745717056347, |
| "loss": 3.5709, |
| "step": 19350 |
| }, |
| { |
| "epoch": 2.0880421913679905, |
| "grad_norm": 0.5291028022766113, |
| "learning_rate": 0.000475274216140502, |
| "loss": 3.5693, |
| "step": 19400 |
| }, |
| { |
| "epoch": 2.0934237434075986, |
| "grad_norm": 0.5555849671363831, |
| "learning_rate": 0.00047495097511044066, |
| "loss": 3.552, |
| "step": 19450 |
| }, |
| { |
| "epoch": 2.098805295447207, |
| "grad_norm": 0.5503198504447937, |
| "learning_rate": 0.00047462773408037925, |
| "loss": 3.5421, |
| "step": 19500 |
| }, |
| { |
| "epoch": 2.1041868474868153, |
| "grad_norm": 0.5942191481590271, |
| "learning_rate": 0.0004743044930503178, |
| "loss": 3.5594, |
| "step": 19550 |
| }, |
| { |
| "epoch": 2.1095683995264234, |
| "grad_norm": 0.5651226043701172, |
| "learning_rate": 0.0004739812520202564, |
| "loss": 3.554, |
| "step": 19600 |
| }, |
| { |
| "epoch": 2.1149499515660315, |
| "grad_norm": 0.5572126507759094, |
| "learning_rate": 0.000473658010990195, |
| "loss": 3.5501, |
| "step": 19650 |
| }, |
| { |
| "epoch": 2.12033150360564, |
| "grad_norm": 0.5130614042282104, |
| "learning_rate": 0.0004733347699601336, |
| "loss": 3.5539, |
| "step": 19700 |
| }, |
| { |
| "epoch": 2.125713055645248, |
| "grad_norm": 0.5788480639457703, |
| "learning_rate": 0.0004730115289300722, |
| "loss": 3.5763, |
| "step": 19750 |
| }, |
| { |
| "epoch": 2.1310946076848563, |
| "grad_norm": 0.5898392796516418, |
| "learning_rate": 0.00047268828790001077, |
| "loss": 3.563, |
| "step": 19800 |
| }, |
| { |
| "epoch": 2.1364761597244644, |
| "grad_norm": 0.5292854309082031, |
| "learning_rate": 0.0004723650468699493, |
| "loss": 3.5473, |
| "step": 19850 |
| }, |
| { |
| "epoch": 2.141857711764073, |
| "grad_norm": 0.5742781758308411, |
| "learning_rate": 0.0004720418058398879, |
| "loss": 3.5564, |
| "step": 19900 |
| }, |
| { |
| "epoch": 2.147239263803681, |
| "grad_norm": 0.5774244666099548, |
| "learning_rate": 0.00047171856480982644, |
| "loss": 3.565, |
| "step": 19950 |
| }, |
| { |
| "epoch": 2.152620815843289, |
| "grad_norm": 0.652957558631897, |
| "learning_rate": 0.0004713953237797651, |
| "loss": 3.5546, |
| "step": 20000 |
| }, |
| { |
| "epoch": 2.152620815843289, |
| "eval_accuracy": 0.3621433346722991, |
| "eval_loss": 3.573176860809326, |
| "eval_runtime": 180.022, |
| "eval_samples_per_second": 100.049, |
| "eval_steps_per_second": 6.255, |
| "step": 20000 |
| }, |
| { |
| "epoch": 2.1580023678828972, |
| "grad_norm": 0.5954522490501404, |
| "learning_rate": 0.0004710720827497037, |
| "loss": 3.5618, |
| "step": 20050 |
| }, |
| { |
| "epoch": 2.163383919922506, |
| "grad_norm": 0.5455233454704285, |
| "learning_rate": 0.00047074884171964223, |
| "loss": 3.5538, |
| "step": 20100 |
| }, |
| { |
| "epoch": 2.168765471962114, |
| "grad_norm": 0.5485193133354187, |
| "learning_rate": 0.0004704256006895808, |
| "loss": 3.5532, |
| "step": 20150 |
| }, |
| { |
| "epoch": 2.174147024001722, |
| "grad_norm": 0.5491883158683777, |
| "learning_rate": 0.00047010235965951936, |
| "loss": 3.546, |
| "step": 20200 |
| }, |
| { |
| "epoch": 2.1795285760413305, |
| "grad_norm": 0.627436101436615, |
| "learning_rate": 0.00046977911862945796, |
| "loss": 3.5433, |
| "step": 20250 |
| }, |
| { |
| "epoch": 2.1849101280809387, |
| "grad_norm": 0.5892840027809143, |
| "learning_rate": 0.0004694558775993966, |
| "loss": 3.565, |
| "step": 20300 |
| }, |
| { |
| "epoch": 2.1902916801205468, |
| "grad_norm": 0.594659686088562, |
| "learning_rate": 0.00046913263656933515, |
| "loss": 3.5551, |
| "step": 20350 |
| }, |
| { |
| "epoch": 2.195673232160155, |
| "grad_norm": 0.578382670879364, |
| "learning_rate": 0.00046880939553927374, |
| "loss": 3.5707, |
| "step": 20400 |
| }, |
| { |
| "epoch": 2.2010547841997634, |
| "grad_norm": 0.5580980181694031, |
| "learning_rate": 0.00046848615450921234, |
| "loss": 3.5548, |
| "step": 20450 |
| }, |
| { |
| "epoch": 2.2064363362393715, |
| "grad_norm": 0.5704784989356995, |
| "learning_rate": 0.0004681629134791509, |
| "loss": 3.5635, |
| "step": 20500 |
| }, |
| { |
| "epoch": 2.2118178882789796, |
| "grad_norm": 0.5783650279045105, |
| "learning_rate": 0.00046783967244908947, |
| "loss": 3.5705, |
| "step": 20550 |
| }, |
| { |
| "epoch": 2.2171994403185877, |
| "grad_norm": 0.5566691160202026, |
| "learning_rate": 0.0004675164314190281, |
| "loss": 3.5542, |
| "step": 20600 |
| }, |
| { |
| "epoch": 2.2225809923581963, |
| "grad_norm": 0.5815993547439575, |
| "learning_rate": 0.00046719319038896666, |
| "loss": 3.5566, |
| "step": 20650 |
| }, |
| { |
| "epoch": 2.2279625443978044, |
| "grad_norm": 0.5823930501937866, |
| "learning_rate": 0.00046686994935890526, |
| "loss": 3.5397, |
| "step": 20700 |
| }, |
| { |
| "epoch": 2.2333440964374125, |
| "grad_norm": 0.624121904373169, |
| "learning_rate": 0.0004665467083288438, |
| "loss": 3.5692, |
| "step": 20750 |
| }, |
| { |
| "epoch": 2.2387256484770206, |
| "grad_norm": 0.6141852140426636, |
| "learning_rate": 0.0004662234672987824, |
| "loss": 3.5432, |
| "step": 20800 |
| }, |
| { |
| "epoch": 2.244107200516629, |
| "grad_norm": 0.5725633502006531, |
| "learning_rate": 0.00046590022626872104, |
| "loss": 3.5536, |
| "step": 20850 |
| }, |
| { |
| "epoch": 2.2494887525562373, |
| "grad_norm": 0.6109887957572937, |
| "learning_rate": 0.0004655769852386596, |
| "loss": 3.5639, |
| "step": 20900 |
| }, |
| { |
| "epoch": 2.2548703045958454, |
| "grad_norm": 0.6038004159927368, |
| "learning_rate": 0.0004652537442085982, |
| "loss": 3.5691, |
| "step": 20950 |
| }, |
| { |
| "epoch": 2.2602518566354535, |
| "grad_norm": 0.5455430150032043, |
| "learning_rate": 0.00046493050317853677, |
| "loss": 3.5598, |
| "step": 21000 |
| }, |
| { |
| "epoch": 2.2602518566354535, |
| "eval_accuracy": 0.3633288463312429, |
| "eval_loss": 3.563311815261841, |
| "eval_runtime": 179.7973, |
| "eval_samples_per_second": 100.174, |
| "eval_steps_per_second": 6.263, |
| "step": 21000 |
| }, |
| { |
| "epoch": 2.265633408675062, |
| "grad_norm": 0.5887195467948914, |
| "learning_rate": 0.0004646072621484753, |
| "loss": 3.5452, |
| "step": 21050 |
| }, |
| { |
| "epoch": 2.27101496071467, |
| "grad_norm": 0.5390068888664246, |
| "learning_rate": 0.0004642840211184139, |
| "loss": 3.5495, |
| "step": 21100 |
| }, |
| { |
| "epoch": 2.2763965127542782, |
| "grad_norm": 0.5845677852630615, |
| "learning_rate": 0.00046396078008835255, |
| "loss": 3.5614, |
| "step": 21150 |
| }, |
| { |
| "epoch": 2.281778064793887, |
| "grad_norm": 0.5742790102958679, |
| "learning_rate": 0.0004636375390582911, |
| "loss": 3.569, |
| "step": 21200 |
| }, |
| { |
| "epoch": 2.287159616833495, |
| "grad_norm": 0.5437269806861877, |
| "learning_rate": 0.0004633142980282297, |
| "loss": 3.565, |
| "step": 21250 |
| }, |
| { |
| "epoch": 2.292541168873103, |
| "grad_norm": 0.6048343181610107, |
| "learning_rate": 0.00046299105699816823, |
| "loss": 3.5613, |
| "step": 21300 |
| }, |
| { |
| "epoch": 2.297922720912711, |
| "grad_norm": 0.6674041748046875, |
| "learning_rate": 0.0004626678159681068, |
| "loss": 3.5655, |
| "step": 21350 |
| }, |
| { |
| "epoch": 2.303304272952319, |
| "grad_norm": 0.6118209958076477, |
| "learning_rate": 0.0004623445749380454, |
| "loss": 3.5589, |
| "step": 21400 |
| }, |
| { |
| "epoch": 2.3086858249919278, |
| "grad_norm": 0.5320875644683838, |
| "learning_rate": 0.000462021333907984, |
| "loss": 3.5578, |
| "step": 21450 |
| }, |
| { |
| "epoch": 2.314067377031536, |
| "grad_norm": 0.5597154498100281, |
| "learning_rate": 0.0004616980928779226, |
| "loss": 3.5686, |
| "step": 21500 |
| }, |
| { |
| "epoch": 2.319448929071144, |
| "grad_norm": 0.6195567846298218, |
| "learning_rate": 0.0004613748518478612, |
| "loss": 3.5259, |
| "step": 21550 |
| }, |
| { |
| "epoch": 2.3248304811107525, |
| "grad_norm": 0.5434510111808777, |
| "learning_rate": 0.000461058075638401, |
| "loss": 3.5511, |
| "step": 21600 |
| }, |
| { |
| "epoch": 2.3302120331503606, |
| "grad_norm": 0.5677668452262878, |
| "learning_rate": 0.00046073483460833955, |
| "loss": 3.5598, |
| "step": 21650 |
| }, |
| { |
| "epoch": 2.3355935851899687, |
| "grad_norm": 0.5577279925346375, |
| "learning_rate": 0.00046041159357827814, |
| "loss": 3.551, |
| "step": 21700 |
| }, |
| { |
| "epoch": 2.340975137229577, |
| "grad_norm": 0.5741985440254211, |
| "learning_rate": 0.00046008835254821674, |
| "loss": 3.5495, |
| "step": 21750 |
| }, |
| { |
| "epoch": 2.3463566892691854, |
| "grad_norm": 0.5373184680938721, |
| "learning_rate": 0.0004597651115181554, |
| "loss": 3.559, |
| "step": 21800 |
| }, |
| { |
| "epoch": 2.3517382413087935, |
| "grad_norm": 0.6047030091285706, |
| "learning_rate": 0.00045944187048809393, |
| "loss": 3.5594, |
| "step": 21850 |
| }, |
| { |
| "epoch": 2.3571197933484016, |
| "grad_norm": 0.5527496933937073, |
| "learning_rate": 0.0004591186294580325, |
| "loss": 3.5631, |
| "step": 21900 |
| }, |
| { |
| "epoch": 2.3625013453880097, |
| "grad_norm": 0.6777981519699097, |
| "learning_rate": 0.00045879538842797106, |
| "loss": 3.5531, |
| "step": 21950 |
| }, |
| { |
| "epoch": 2.3678828974276183, |
| "grad_norm": 0.5738788843154907, |
| "learning_rate": 0.00045847214739790966, |
| "loss": 3.5472, |
| "step": 22000 |
| }, |
| { |
| "epoch": 2.3678828974276183, |
| "eval_accuracy": 0.3643889725576488, |
| "eval_loss": 3.552778720855713, |
| "eval_runtime": 179.721, |
| "eval_samples_per_second": 100.216, |
| "eval_steps_per_second": 6.265, |
| "step": 22000 |
| }, |
| { |
| "epoch": 2.3732644494672264, |
| "grad_norm": 0.5943243503570557, |
| "learning_rate": 0.0004581489063678482, |
| "loss": 3.5544, |
| "step": 22050 |
| }, |
| { |
| "epoch": 2.3786460015068345, |
| "grad_norm": 0.5966238379478455, |
| "learning_rate": 0.00045782566533778685, |
| "loss": 3.5685, |
| "step": 22100 |
| }, |
| { |
| "epoch": 2.384027553546443, |
| "grad_norm": 0.6006495952606201, |
| "learning_rate": 0.00045750242430772544, |
| "loss": 3.5509, |
| "step": 22150 |
| }, |
| { |
| "epoch": 2.389409105586051, |
| "grad_norm": 0.5800380706787109, |
| "learning_rate": 0.000457179183277664, |
| "loss": 3.5442, |
| "step": 22200 |
| }, |
| { |
| "epoch": 2.3947906576256592, |
| "grad_norm": 0.6179288029670715, |
| "learning_rate": 0.0004568559422476026, |
| "loss": 3.5578, |
| "step": 22250 |
| }, |
| { |
| "epoch": 2.4001722096652673, |
| "grad_norm": 0.5599116683006287, |
| "learning_rate": 0.00045653270121754117, |
| "loss": 3.5604, |
| "step": 22300 |
| }, |
| { |
| "epoch": 2.4055537617048754, |
| "grad_norm": 0.5807077884674072, |
| "learning_rate": 0.0004562094601874797, |
| "loss": 3.5322, |
| "step": 22350 |
| }, |
| { |
| "epoch": 2.410935313744484, |
| "grad_norm": 0.6106342077255249, |
| "learning_rate": 0.00045588621915741836, |
| "loss": 3.5576, |
| "step": 22400 |
| }, |
| { |
| "epoch": 2.416316865784092, |
| "grad_norm": 0.5440487861633301, |
| "learning_rate": 0.00045556297812735696, |
| "loss": 3.5573, |
| "step": 22450 |
| }, |
| { |
| "epoch": 2.4216984178237, |
| "grad_norm": 0.5883488655090332, |
| "learning_rate": 0.0004552397370972955, |
| "loss": 3.5533, |
| "step": 22500 |
| }, |
| { |
| "epoch": 2.4270799698633088, |
| "grad_norm": 0.5989765524864197, |
| "learning_rate": 0.0004549164960672341, |
| "loss": 3.5697, |
| "step": 22550 |
| }, |
| { |
| "epoch": 2.432461521902917, |
| "grad_norm": 0.5756821632385254, |
| "learning_rate": 0.00045459325503717263, |
| "loss": 3.5493, |
| "step": 22600 |
| }, |
| { |
| "epoch": 2.437843073942525, |
| "grad_norm": 0.5626195669174194, |
| "learning_rate": 0.0004542700140071113, |
| "loss": 3.5562, |
| "step": 22650 |
| }, |
| { |
| "epoch": 2.443224625982133, |
| "grad_norm": 0.5613085627555847, |
| "learning_rate": 0.0004539467729770499, |
| "loss": 3.5262, |
| "step": 22700 |
| }, |
| { |
| "epoch": 2.4486061780217416, |
| "grad_norm": 0.6236090064048767, |
| "learning_rate": 0.0004536235319469884, |
| "loss": 3.5649, |
| "step": 22750 |
| }, |
| { |
| "epoch": 2.4539877300613497, |
| "grad_norm": 0.5902277827262878, |
| "learning_rate": 0.000453300290916927, |
| "loss": 3.553, |
| "step": 22800 |
| }, |
| { |
| "epoch": 2.459369282100958, |
| "grad_norm": 0.6016800403594971, |
| "learning_rate": 0.0004529770498868656, |
| "loss": 3.5478, |
| "step": 22850 |
| }, |
| { |
| "epoch": 2.464750834140566, |
| "grad_norm": 0.6465705633163452, |
| "learning_rate": 0.00045265380885680414, |
| "loss": 3.5654, |
| "step": 22900 |
| }, |
| { |
| "epoch": 2.4701323861801745, |
| "grad_norm": 0.5794358849525452, |
| "learning_rate": 0.0004523305678267428, |
| "loss": 3.5521, |
| "step": 22950 |
| }, |
| { |
| "epoch": 2.4755139382197826, |
| "grad_norm": 0.5933077335357666, |
| "learning_rate": 0.0004520073267966814, |
| "loss": 3.5516, |
| "step": 23000 |
| }, |
| { |
| "epoch": 2.4755139382197826, |
| "eval_accuracy": 0.3657163848880783, |
| "eval_loss": 3.5414011478424072, |
| "eval_runtime": 180.0044, |
| "eval_samples_per_second": 100.059, |
| "eval_steps_per_second": 6.255, |
| "step": 23000 |
| }, |
| { |
| "epoch": 2.4808954902593907, |
| "grad_norm": 0.5735173225402832, |
| "learning_rate": 0.00045168408576661993, |
| "loss": 3.5382, |
| "step": 23050 |
| }, |
| { |
| "epoch": 2.4862770422989993, |
| "grad_norm": 0.5812858939170837, |
| "learning_rate": 0.0004513608447365585, |
| "loss": 3.5504, |
| "step": 23100 |
| }, |
| { |
| "epoch": 2.4916585943386074, |
| "grad_norm": 0.5932419300079346, |
| "learning_rate": 0.00045103760370649706, |
| "loss": 3.5628, |
| "step": 23150 |
| }, |
| { |
| "epoch": 2.4970401463782155, |
| "grad_norm": 0.5909202098846436, |
| "learning_rate": 0.00045071436267643566, |
| "loss": 3.5538, |
| "step": 23200 |
| }, |
| { |
| "epoch": 2.5024216984178236, |
| "grad_norm": 0.5628249645233154, |
| "learning_rate": 0.0004503911216463743, |
| "loss": 3.5681, |
| "step": 23250 |
| }, |
| { |
| "epoch": 2.5078032504574317, |
| "grad_norm": 0.5927748680114746, |
| "learning_rate": 0.00045006788061631285, |
| "loss": 3.5467, |
| "step": 23300 |
| }, |
| { |
| "epoch": 2.5131848024970402, |
| "grad_norm": 0.6498246192932129, |
| "learning_rate": 0.00044974463958625144, |
| "loss": 3.5468, |
| "step": 23350 |
| }, |
| { |
| "epoch": 2.5185663545366483, |
| "grad_norm": 0.5863044857978821, |
| "learning_rate": 0.00044942139855619004, |
| "loss": 3.5537, |
| "step": 23400 |
| }, |
| { |
| "epoch": 2.5239479065762565, |
| "grad_norm": 0.5949398279190063, |
| "learning_rate": 0.0004490981575261286, |
| "loss": 3.5465, |
| "step": 23450 |
| }, |
| { |
| "epoch": 2.529329458615865, |
| "grad_norm": 0.6202422380447388, |
| "learning_rate": 0.0004487749164960672, |
| "loss": 3.5471, |
| "step": 23500 |
| }, |
| { |
| "epoch": 2.534711010655473, |
| "grad_norm": 0.5920185446739197, |
| "learning_rate": 0.0004484516754660058, |
| "loss": 3.5452, |
| "step": 23550 |
| }, |
| { |
| "epoch": 2.540092562695081, |
| "grad_norm": 0.5713174939155579, |
| "learning_rate": 0.00044812843443594436, |
| "loss": 3.5476, |
| "step": 23600 |
| }, |
| { |
| "epoch": 2.5454741147346893, |
| "grad_norm": 0.5799892544746399, |
| "learning_rate": 0.00044781165822648417, |
| "loss": 3.5464, |
| "step": 23650 |
| }, |
| { |
| "epoch": 2.550855666774298, |
| "grad_norm": 0.5944963693618774, |
| "learning_rate": 0.00044748841719642276, |
| "loss": 3.5446, |
| "step": 23700 |
| }, |
| { |
| "epoch": 2.556237218813906, |
| "grad_norm": 0.5628204941749573, |
| "learning_rate": 0.00044716517616636136, |
| "loss": 3.5501, |
| "step": 23750 |
| }, |
| { |
| "epoch": 2.561618770853514, |
| "grad_norm": 0.5771909952163696, |
| "learning_rate": 0.0004468419351362999, |
| "loss": 3.5373, |
| "step": 23800 |
| }, |
| { |
| "epoch": 2.567000322893122, |
| "grad_norm": 0.5741766691207886, |
| "learning_rate": 0.0004465186941062385, |
| "loss": 3.5667, |
| "step": 23850 |
| }, |
| { |
| "epoch": 2.5723818749327307, |
| "grad_norm": 0.5425882339477539, |
| "learning_rate": 0.00044619545307617714, |
| "loss": 3.5707, |
| "step": 23900 |
| }, |
| { |
| "epoch": 2.577763426972339, |
| "grad_norm": 0.6335693001747131, |
| "learning_rate": 0.0004458722120461157, |
| "loss": 3.5424, |
| "step": 23950 |
| }, |
| { |
| "epoch": 2.583144979011947, |
| "grad_norm": 0.6267858743667603, |
| "learning_rate": 0.0004455489710160543, |
| "loss": 3.5437, |
| "step": 24000 |
| }, |
| { |
| "epoch": 2.583144979011947, |
| "eval_accuracy": 0.36688157845700076, |
| "eval_loss": 3.532123565673828, |
| "eval_runtime": 179.6516, |
| "eval_samples_per_second": 100.255, |
| "eval_steps_per_second": 6.268, |
| "step": 24000 |
| }, |
| { |
| "epoch": 2.5885265310515555, |
| "grad_norm": 0.6489107608795166, |
| "learning_rate": 0.0004452257299859928, |
| "loss": 3.5309, |
| "step": 24050 |
| }, |
| { |
| "epoch": 2.5939080830911636, |
| "grad_norm": 0.6115685105323792, |
| "learning_rate": 0.0004449024889559314, |
| "loss": 3.5566, |
| "step": 24100 |
| }, |
| { |
| "epoch": 2.5992896351307717, |
| "grad_norm": 0.5589924454689026, |
| "learning_rate": 0.00044457924792587, |
| "loss": 3.5505, |
| "step": 24150 |
| }, |
| { |
| "epoch": 2.60467118717038, |
| "grad_norm": 0.6251258254051208, |
| "learning_rate": 0.0004442560068958086, |
| "loss": 3.5464, |
| "step": 24200 |
| }, |
| { |
| "epoch": 2.610052739209988, |
| "grad_norm": 0.5544481873512268, |
| "learning_rate": 0.0004439327658657472, |
| "loss": 3.556, |
| "step": 24250 |
| }, |
| { |
| "epoch": 2.6154342912495965, |
| "grad_norm": 0.5777297616004944, |
| "learning_rate": 0.0004436095248356858, |
| "loss": 3.5515, |
| "step": 24300 |
| }, |
| { |
| "epoch": 2.6208158432892046, |
| "grad_norm": 0.6586969494819641, |
| "learning_rate": 0.00044328628380562433, |
| "loss": 3.5414, |
| "step": 24350 |
| }, |
| { |
| "epoch": 2.6261973953288127, |
| "grad_norm": 0.6207393407821655, |
| "learning_rate": 0.0004429630427755629, |
| "loss": 3.5585, |
| "step": 24400 |
| }, |
| { |
| "epoch": 2.6315789473684212, |
| "grad_norm": 0.5969880819320679, |
| "learning_rate": 0.0004426398017455016, |
| "loss": 3.5571, |
| "step": 24450 |
| }, |
| { |
| "epoch": 2.6369604994080293, |
| "grad_norm": 0.6068384051322937, |
| "learning_rate": 0.0004423165607154401, |
| "loss": 3.5417, |
| "step": 24500 |
| }, |
| { |
| "epoch": 2.6423420514476375, |
| "grad_norm": 0.582072913646698, |
| "learning_rate": 0.0004419933196853787, |
| "loss": 3.5233, |
| "step": 24550 |
| }, |
| { |
| "epoch": 2.6477236034872456, |
| "grad_norm": 0.6101149320602417, |
| "learning_rate": 0.00044167007865531725, |
| "loss": 3.5402, |
| "step": 24600 |
| }, |
| { |
| "epoch": 2.653105155526854, |
| "grad_norm": 0.6337757110595703, |
| "learning_rate": 0.00044134683762525584, |
| "loss": 3.5594, |
| "step": 24650 |
| }, |
| { |
| "epoch": 2.658486707566462, |
| "grad_norm": 0.6083040833473206, |
| "learning_rate": 0.00044102359659519444, |
| "loss": 3.5444, |
| "step": 24700 |
| }, |
| { |
| "epoch": 2.6638682596060703, |
| "grad_norm": 0.6233500242233276, |
| "learning_rate": 0.00044070035556513303, |
| "loss": 3.5584, |
| "step": 24750 |
| }, |
| { |
| "epoch": 2.6692498116456784, |
| "grad_norm": 0.5510074496269226, |
| "learning_rate": 0.00044037711453507163, |
| "loss": 3.5389, |
| "step": 24800 |
| }, |
| { |
| "epoch": 2.674631363685287, |
| "grad_norm": 0.6422367095947266, |
| "learning_rate": 0.0004400538735050102, |
| "loss": 3.5267, |
| "step": 24850 |
| }, |
| { |
| "epoch": 2.680012915724895, |
| "grad_norm": 0.5510942339897156, |
| "learning_rate": 0.00043973063247494876, |
| "loss": 3.5431, |
| "step": 24900 |
| }, |
| { |
| "epoch": 2.685394467764503, |
| "grad_norm": 0.6150233149528503, |
| "learning_rate": 0.00043940739144488736, |
| "loss": 3.5392, |
| "step": 24950 |
| }, |
| { |
| "epoch": 2.6907760198041117, |
| "grad_norm": 0.5818309783935547, |
| "learning_rate": 0.0004390841504148259, |
| "loss": 3.5452, |
| "step": 25000 |
| }, |
| { |
| "epoch": 2.6907760198041117, |
| "eval_accuracy": 0.36736758282419496, |
| "eval_loss": 3.5232725143432617, |
| "eval_runtime": 180.1337, |
| "eval_samples_per_second": 99.987, |
| "eval_steps_per_second": 6.251, |
| "step": 25000 |
| }, |
| { |
| "epoch": 2.69615757184372, |
| "grad_norm": 0.5925581455230713, |
| "learning_rate": 0.00043876090938476455, |
| "loss": 3.5541, |
| "step": 25050 |
| }, |
| { |
| "epoch": 2.701539123883328, |
| "grad_norm": 0.5792503952980042, |
| "learning_rate": 0.00043843766835470314, |
| "loss": 3.5547, |
| "step": 25100 |
| }, |
| { |
| "epoch": 2.706920675922936, |
| "grad_norm": 0.6317951083183289, |
| "learning_rate": 0.0004381144273246417, |
| "loss": 3.5403, |
| "step": 25150 |
| }, |
| { |
| "epoch": 2.712302227962544, |
| "grad_norm": 0.5946325659751892, |
| "learning_rate": 0.0004377911862945803, |
| "loss": 3.5453, |
| "step": 25200 |
| }, |
| { |
| "epoch": 2.7176837800021527, |
| "grad_norm": 0.59228515625, |
| "learning_rate": 0.00043746794526451887, |
| "loss": 3.5341, |
| "step": 25250 |
| }, |
| { |
| "epoch": 2.723065332041761, |
| "grad_norm": 0.5823553800582886, |
| "learning_rate": 0.00043714470423445747, |
| "loss": 3.5449, |
| "step": 25300 |
| }, |
| { |
| "epoch": 2.728446884081369, |
| "grad_norm": 0.6187898516654968, |
| "learning_rate": 0.00043682146320439606, |
| "loss": 3.5427, |
| "step": 25350 |
| }, |
| { |
| "epoch": 2.7338284361209775, |
| "grad_norm": 0.6080958247184753, |
| "learning_rate": 0.00043649822217433466, |
| "loss": 3.5392, |
| "step": 25400 |
| }, |
| { |
| "epoch": 2.7392099881605856, |
| "grad_norm": 0.5838017463684082, |
| "learning_rate": 0.0004361749811442732, |
| "loss": 3.5302, |
| "step": 25450 |
| }, |
| { |
| "epoch": 2.7445915402001937, |
| "grad_norm": 0.5467104911804199, |
| "learning_rate": 0.0004358517401142118, |
| "loss": 3.5391, |
| "step": 25500 |
| }, |
| { |
| "epoch": 2.749973092239802, |
| "grad_norm": 0.5581143498420715, |
| "learning_rate": 0.00043552849908415033, |
| "loss": 3.5437, |
| "step": 25550 |
| }, |
| { |
| "epoch": 2.7553546442794103, |
| "grad_norm": 0.5552875399589539, |
| "learning_rate": 0.000435205258054089, |
| "loss": 3.5394, |
| "step": 25600 |
| }, |
| { |
| "epoch": 2.7607361963190185, |
| "grad_norm": 0.6552839875221252, |
| "learning_rate": 0.00043488848184462873, |
| "loss": 3.5327, |
| "step": 25650 |
| }, |
| { |
| "epoch": 2.7661177483586266, |
| "grad_norm": 0.648295521736145, |
| "learning_rate": 0.0004345652408145674, |
| "loss": 3.5352, |
| "step": 25700 |
| }, |
| { |
| "epoch": 2.7714993003982347, |
| "grad_norm": 0.6237471699714661, |
| "learning_rate": 0.000434241999784506, |
| "loss": 3.5333, |
| "step": 25750 |
| }, |
| { |
| "epoch": 2.776880852437843, |
| "grad_norm": 0.5941251516342163, |
| "learning_rate": 0.0004339187587544445, |
| "loss": 3.5268, |
| "step": 25800 |
| }, |
| { |
| "epoch": 2.7822624044774513, |
| "grad_norm": 0.6055666208267212, |
| "learning_rate": 0.0004335955177243831, |
| "loss": 3.5436, |
| "step": 25850 |
| }, |
| { |
| "epoch": 2.7876439565170594, |
| "grad_norm": 0.6202614903450012, |
| "learning_rate": 0.00043327227669432165, |
| "loss": 3.5259, |
| "step": 25900 |
| }, |
| { |
| "epoch": 2.793025508556668, |
| "grad_norm": 0.579383909702301, |
| "learning_rate": 0.00043294903566426025, |
| "loss": 3.5166, |
| "step": 25950 |
| }, |
| { |
| "epoch": 2.798407060596276, |
| "grad_norm": 0.5871159434318542, |
| "learning_rate": 0.0004326257946341989, |
| "loss": 3.5385, |
| "step": 26000 |
| }, |
| { |
| "epoch": 2.798407060596276, |
| "eval_accuracy": 0.3681571633599428, |
| "eval_loss": 3.51601505279541, |
| "eval_runtime": 179.8655, |
| "eval_samples_per_second": 100.136, |
| "eval_steps_per_second": 6.26, |
| "step": 26000 |
| }, |
| { |
| "epoch": 2.803788612635884, |
| "grad_norm": 0.6053968071937561, |
| "learning_rate": 0.00043230255360413744, |
| "loss": 3.5358, |
| "step": 26050 |
| }, |
| { |
| "epoch": 2.8091701646754923, |
| "grad_norm": 0.5837214589118958, |
| "learning_rate": 0.00043197931257407603, |
| "loss": 3.5295, |
| "step": 26100 |
| }, |
| { |
| "epoch": 2.8145517167151004, |
| "grad_norm": 0.6265986561775208, |
| "learning_rate": 0.0004316560715440146, |
| "loss": 3.5275, |
| "step": 26150 |
| }, |
| { |
| "epoch": 2.819933268754709, |
| "grad_norm": 0.6679002642631531, |
| "learning_rate": 0.00043133929533455443, |
| "loss": 3.534, |
| "step": 26200 |
| }, |
| { |
| "epoch": 2.825314820794317, |
| "grad_norm": 0.6008105874061584, |
| "learning_rate": 0.00043101605430449297, |
| "loss": 3.5266, |
| "step": 26250 |
| }, |
| { |
| "epoch": 2.830696372833925, |
| "grad_norm": 0.6022807955741882, |
| "learning_rate": 0.00043069281327443157, |
| "loss": 3.5404, |
| "step": 26300 |
| }, |
| { |
| "epoch": 2.8360779248735337, |
| "grad_norm": 0.5778976678848267, |
| "learning_rate": 0.0004303695722443702, |
| "loss": 3.5387, |
| "step": 26350 |
| }, |
| { |
| "epoch": 2.841459476913142, |
| "grad_norm": 0.6605262160301208, |
| "learning_rate": 0.00043004633121430876, |
| "loss": 3.5403, |
| "step": 26400 |
| }, |
| { |
| "epoch": 2.84684102895275, |
| "grad_norm": 0.5825318098068237, |
| "learning_rate": 0.00042972309018424735, |
| "loss": 3.5128, |
| "step": 26450 |
| }, |
| { |
| "epoch": 2.852222580992358, |
| "grad_norm": 0.612047553062439, |
| "learning_rate": 0.00042939984915418594, |
| "loss": 3.54, |
| "step": 26500 |
| }, |
| { |
| "epoch": 2.857604133031966, |
| "grad_norm": 0.6171521544456482, |
| "learning_rate": 0.0004290766081241245, |
| "loss": 3.5308, |
| "step": 26550 |
| }, |
| { |
| "epoch": 2.8629856850715747, |
| "grad_norm": 0.5376119017601013, |
| "learning_rate": 0.0004287533670940631, |
| "loss": 3.5204, |
| "step": 26600 |
| }, |
| { |
| "epoch": 2.868367237111183, |
| "grad_norm": 0.6116226315498352, |
| "learning_rate": 0.00042843012606400173, |
| "loss": 3.5264, |
| "step": 26650 |
| }, |
| { |
| "epoch": 2.873748789150791, |
| "grad_norm": 0.5715238451957703, |
| "learning_rate": 0.00042810688503394027, |
| "loss": 3.5308, |
| "step": 26700 |
| }, |
| { |
| "epoch": 2.8791303411903995, |
| "grad_norm": 0.5794273018836975, |
| "learning_rate": 0.00042778364400387886, |
| "loss": 3.5252, |
| "step": 26750 |
| }, |
| { |
| "epoch": 2.8845118932300076, |
| "grad_norm": 0.6008096933364868, |
| "learning_rate": 0.0004274604029738174, |
| "loss": 3.5288, |
| "step": 26800 |
| }, |
| { |
| "epoch": 2.8898934452696157, |
| "grad_norm": 0.6101154088973999, |
| "learning_rate": 0.000427137161943756, |
| "loss": 3.5337, |
| "step": 26850 |
| }, |
| { |
| "epoch": 2.895274997309224, |
| "grad_norm": 0.5782980918884277, |
| "learning_rate": 0.00042681392091369465, |
| "loss": 3.531, |
| "step": 26900 |
| }, |
| { |
| "epoch": 2.9006565493488323, |
| "grad_norm": 0.5600022077560425, |
| "learning_rate": 0.0004264906798836332, |
| "loss": 3.509, |
| "step": 26950 |
| }, |
| { |
| "epoch": 2.9060381013884404, |
| "grad_norm": 0.5735082030296326, |
| "learning_rate": 0.0004261674388535718, |
| "loss": 3.5233, |
| "step": 27000 |
| }, |
| { |
| "epoch": 2.9060381013884404, |
| "eval_accuracy": 0.3694485029316179, |
| "eval_loss": 3.506122589111328, |
| "eval_runtime": 179.9881, |
| "eval_samples_per_second": 100.068, |
| "eval_steps_per_second": 6.256, |
| "step": 27000 |
| }, |
| { |
| "epoch": 2.9114196534280485, |
| "grad_norm": 0.6081140041351318, |
| "learning_rate": 0.0004258441978235104, |
| "loss": 3.5229, |
| "step": 27050 |
| }, |
| { |
| "epoch": 2.9168012054676566, |
| "grad_norm": 0.591622531414032, |
| "learning_rate": 0.0004255209567934489, |
| "loss": 3.5366, |
| "step": 27100 |
| }, |
| { |
| "epoch": 2.922182757507265, |
| "grad_norm": 0.5882881283760071, |
| "learning_rate": 0.0004251977157633875, |
| "loss": 3.528, |
| "step": 27150 |
| }, |
| { |
| "epoch": 2.9275643095468733, |
| "grad_norm": 0.6217557191848755, |
| "learning_rate": 0.00042487447473332616, |
| "loss": 3.5435, |
| "step": 27200 |
| }, |
| { |
| "epoch": 2.9329458615864814, |
| "grad_norm": 0.5857243537902832, |
| "learning_rate": 0.0004245512337032647, |
| "loss": 3.514, |
| "step": 27250 |
| }, |
| { |
| "epoch": 2.93832741362609, |
| "grad_norm": 0.5875625610351562, |
| "learning_rate": 0.0004242279926732033, |
| "loss": 3.5184, |
| "step": 27300 |
| }, |
| { |
| "epoch": 2.943708965665698, |
| "grad_norm": 0.5514461398124695, |
| "learning_rate": 0.00042390475164314184, |
| "loss": 3.5367, |
| "step": 27350 |
| }, |
| { |
| "epoch": 2.949090517705306, |
| "grad_norm": 0.5746405124664307, |
| "learning_rate": 0.00042358151061308043, |
| "loss": 3.5174, |
| "step": 27400 |
| }, |
| { |
| "epoch": 2.9544720697449143, |
| "grad_norm": 0.5785201191902161, |
| "learning_rate": 0.000423258269583019, |
| "loss": 3.5364, |
| "step": 27450 |
| }, |
| { |
| "epoch": 2.9598536217845224, |
| "grad_norm": 0.6101451516151428, |
| "learning_rate": 0.0004229350285529576, |
| "loss": 3.5128, |
| "step": 27500 |
| }, |
| { |
| "epoch": 2.965235173824131, |
| "grad_norm": 0.607959508895874, |
| "learning_rate": 0.0004226117875228962, |
| "loss": 3.5115, |
| "step": 27550 |
| }, |
| { |
| "epoch": 2.970616725863739, |
| "grad_norm": 0.5360074043273926, |
| "learning_rate": 0.0004222885464928348, |
| "loss": 3.5161, |
| "step": 27600 |
| }, |
| { |
| "epoch": 2.975998277903347, |
| "grad_norm": 0.6007967591285706, |
| "learning_rate": 0.00042196530546277335, |
| "loss": 3.5294, |
| "step": 27650 |
| }, |
| { |
| "epoch": 2.9813798299429557, |
| "grad_norm": 0.679694652557373, |
| "learning_rate": 0.00042164206443271195, |
| "loss": 3.5162, |
| "step": 27700 |
| }, |
| { |
| "epoch": 2.986761381982564, |
| "grad_norm": 0.589306116104126, |
| "learning_rate": 0.0004213188234026505, |
| "loss": 3.5213, |
| "step": 27750 |
| }, |
| { |
| "epoch": 2.992142934022172, |
| "grad_norm": 0.5684702396392822, |
| "learning_rate": 0.00042099558237258914, |
| "loss": 3.5406, |
| "step": 27800 |
| }, |
| { |
| "epoch": 2.9975244860617805, |
| "grad_norm": 0.6052414774894714, |
| "learning_rate": 0.00042067234134252773, |
| "loss": 3.5357, |
| "step": 27850 |
| }, |
| { |
| "epoch": 3.0029060381013886, |
| "grad_norm": 0.5678727030754089, |
| "learning_rate": 0.00042034910031246627, |
| "loss": 3.486, |
| "step": 27900 |
| }, |
| { |
| "epoch": 3.0082875901409967, |
| "grad_norm": 0.6087445020675659, |
| "learning_rate": 0.00042002585928240486, |
| "loss": 3.4427, |
| "step": 27950 |
| }, |
| { |
| "epoch": 3.0136691421806048, |
| "grad_norm": 0.6876854300498962, |
| "learning_rate": 0.00041970261825234346, |
| "loss": 3.4446, |
| "step": 28000 |
| }, |
| { |
| "epoch": 3.0136691421806048, |
| "eval_accuracy": 0.37010759134910054, |
| "eval_loss": 3.5012426376342773, |
| "eval_runtime": 179.7175, |
| "eval_samples_per_second": 100.218, |
| "eval_steps_per_second": 6.265, |
| "step": 28000 |
| }, |
| { |
| "epoch": 3.0190506942202133, |
| "grad_norm": 0.5857146978378296, |
| "learning_rate": 0.00041937937722228205, |
| "loss": 3.4242, |
| "step": 28050 |
| }, |
| { |
| "epoch": 3.0244322462598214, |
| "grad_norm": 0.6560085415840149, |
| "learning_rate": 0.00041905613619222065, |
| "loss": 3.4337, |
| "step": 28100 |
| }, |
| { |
| "epoch": 3.0298137982994295, |
| "grad_norm": 0.6381294131278992, |
| "learning_rate": 0.00041873289516215924, |
| "loss": 3.4334, |
| "step": 28150 |
| }, |
| { |
| "epoch": 3.0351953503390376, |
| "grad_norm": 0.6040878891944885, |
| "learning_rate": 0.0004184096541320978, |
| "loss": 3.4403, |
| "step": 28200 |
| }, |
| { |
| "epoch": 3.040576902378646, |
| "grad_norm": 0.5884063839912415, |
| "learning_rate": 0.0004180864131020364, |
| "loss": 3.4367, |
| "step": 28250 |
| }, |
| { |
| "epoch": 3.0459584544182543, |
| "grad_norm": 0.574171781539917, |
| "learning_rate": 0.0004177631720719749, |
| "loss": 3.4463, |
| "step": 28300 |
| }, |
| { |
| "epoch": 3.0513400064578624, |
| "grad_norm": 0.6166306734085083, |
| "learning_rate": 0.00041743993104191357, |
| "loss": 3.4219, |
| "step": 28350 |
| }, |
| { |
| "epoch": 3.0567215584974705, |
| "grad_norm": 0.5581924915313721, |
| "learning_rate": 0.00041711669001185216, |
| "loss": 3.4681, |
| "step": 28400 |
| }, |
| { |
| "epoch": 3.062103110537079, |
| "grad_norm": 0.5798004269599915, |
| "learning_rate": 0.0004167934489817907, |
| "loss": 3.402, |
| "step": 28450 |
| }, |
| { |
| "epoch": 3.067484662576687, |
| "grad_norm": 0.5792421102523804, |
| "learning_rate": 0.0004164702079517293, |
| "loss": 3.4292, |
| "step": 28500 |
| }, |
| { |
| "epoch": 3.0728662146162953, |
| "grad_norm": 0.5974885821342468, |
| "learning_rate": 0.0004161469669216679, |
| "loss": 3.4502, |
| "step": 28550 |
| }, |
| { |
| "epoch": 3.0782477666559034, |
| "grad_norm": 0.6063544750213623, |
| "learning_rate": 0.00041582372589160643, |
| "loss": 3.4676, |
| "step": 28600 |
| }, |
| { |
| "epoch": 3.083629318695512, |
| "grad_norm": 0.6233583688735962, |
| "learning_rate": 0.0004155004848615451, |
| "loss": 3.4665, |
| "step": 28650 |
| }, |
| { |
| "epoch": 3.08901087073512, |
| "grad_norm": 0.6059950590133667, |
| "learning_rate": 0.0004151772438314837, |
| "loss": 3.4411, |
| "step": 28700 |
| }, |
| { |
| "epoch": 3.094392422774728, |
| "grad_norm": 0.634493887424469, |
| "learning_rate": 0.0004148540028014222, |
| "loss": 3.4556, |
| "step": 28750 |
| }, |
| { |
| "epoch": 3.0997739748143363, |
| "grad_norm": 0.6094062328338623, |
| "learning_rate": 0.0004145307617713608, |
| "loss": 3.4545, |
| "step": 28800 |
| }, |
| { |
| "epoch": 3.105155526853945, |
| "grad_norm": 0.6499791145324707, |
| "learning_rate": 0.00041420752074129935, |
| "loss": 3.4478, |
| "step": 28850 |
| }, |
| { |
| "epoch": 3.110537078893553, |
| "grad_norm": 0.5969070196151733, |
| "learning_rate": 0.000413884279711238, |
| "loss": 3.4373, |
| "step": 28900 |
| }, |
| { |
| "epoch": 3.115918630933161, |
| "grad_norm": 0.5845366716384888, |
| "learning_rate": 0.0004135610386811766, |
| "loss": 3.4629, |
| "step": 28950 |
| }, |
| { |
| "epoch": 3.121300182972769, |
| "grad_norm": 0.609656035900116, |
| "learning_rate": 0.00041323779765111514, |
| "loss": 3.4549, |
| "step": 29000 |
| }, |
| { |
| "epoch": 3.121300182972769, |
| "eval_accuracy": 0.37079699392228344, |
| "eval_loss": 3.4950499534606934, |
| "eval_runtime": 179.9053, |
| "eval_samples_per_second": 100.114, |
| "eval_steps_per_second": 6.259, |
| "step": 29000 |
| }, |
| { |
| "epoch": 3.1266817350123777, |
| "grad_norm": 0.5781153440475464, |
| "learning_rate": 0.00041291455662105373, |
| "loss": 3.4505, |
| "step": 29050 |
| }, |
| { |
| "epoch": 3.132063287051986, |
| "grad_norm": 0.6373555064201355, |
| "learning_rate": 0.0004125913155909923, |
| "loss": 3.4634, |
| "step": 29100 |
| }, |
| { |
| "epoch": 3.137444839091594, |
| "grad_norm": 0.6015342473983765, |
| "learning_rate": 0.00041226807456093087, |
| "loss": 3.4469, |
| "step": 29150 |
| }, |
| { |
| "epoch": 3.1428263911312024, |
| "grad_norm": 0.644429624080658, |
| "learning_rate": 0.0004119448335308695, |
| "loss": 3.4523, |
| "step": 29200 |
| }, |
| { |
| "epoch": 3.1482079431708105, |
| "grad_norm": 0.5962895750999451, |
| "learning_rate": 0.0004116215925008081, |
| "loss": 3.4383, |
| "step": 29250 |
| }, |
| { |
| "epoch": 3.1535894952104186, |
| "grad_norm": 0.6676060557365417, |
| "learning_rate": 0.00041129835147074665, |
| "loss": 3.4608, |
| "step": 29300 |
| }, |
| { |
| "epoch": 3.1589710472500268, |
| "grad_norm": 0.6141141057014465, |
| "learning_rate": 0.00041097511044068524, |
| "loss": 3.4517, |
| "step": 29350 |
| }, |
| { |
| "epoch": 3.1643525992896353, |
| "grad_norm": 0.5669366121292114, |
| "learning_rate": 0.0004106518694106238, |
| "loss": 3.4687, |
| "step": 29400 |
| }, |
| { |
| "epoch": 3.1697341513292434, |
| "grad_norm": 0.5981097221374512, |
| "learning_rate": 0.0004103286283805624, |
| "loss": 3.4496, |
| "step": 29450 |
| }, |
| { |
| "epoch": 3.1751157033688515, |
| "grad_norm": 0.5925617814064026, |
| "learning_rate": 0.0004100118521711022, |
| "loss": 3.4593, |
| "step": 29500 |
| }, |
| { |
| "epoch": 3.1804972554084596, |
| "grad_norm": 0.5993756055831909, |
| "learning_rate": 0.0004096886111410408, |
| "loss": 3.4616, |
| "step": 29550 |
| }, |
| { |
| "epoch": 3.185878807448068, |
| "grad_norm": 0.6227186322212219, |
| "learning_rate": 0.00040936537011097943, |
| "loss": 3.46, |
| "step": 29600 |
| }, |
| { |
| "epoch": 3.1912603594876763, |
| "grad_norm": 0.6608783006668091, |
| "learning_rate": 0.00040904212908091797, |
| "loss": 3.4731, |
| "step": 29650 |
| }, |
| { |
| "epoch": 3.1966419115272844, |
| "grad_norm": 0.6810404062271118, |
| "learning_rate": 0.00040871888805085656, |
| "loss": 3.4522, |
| "step": 29700 |
| }, |
| { |
| "epoch": 3.2020234635668925, |
| "grad_norm": 0.6131249666213989, |
| "learning_rate": 0.0004083956470207951, |
| "loss": 3.4594, |
| "step": 29750 |
| }, |
| { |
| "epoch": 3.207405015606501, |
| "grad_norm": 0.6485390663146973, |
| "learning_rate": 0.0004080724059907337, |
| "loss": 3.4409, |
| "step": 29800 |
| }, |
| { |
| "epoch": 3.212786567646109, |
| "grad_norm": 0.6126716136932373, |
| "learning_rate": 0.00040774916496067235, |
| "loss": 3.4303, |
| "step": 29850 |
| }, |
| { |
| "epoch": 3.2181681196857173, |
| "grad_norm": 0.626392662525177, |
| "learning_rate": 0.0004074259239306109, |
| "loss": 3.4539, |
| "step": 29900 |
| }, |
| { |
| "epoch": 3.2235496717253254, |
| "grad_norm": 0.5780983567237854, |
| "learning_rate": 0.0004071026829005495, |
| "loss": 3.4479, |
| "step": 29950 |
| }, |
| { |
| "epoch": 3.228931223764934, |
| "grad_norm": 0.5982878804206848, |
| "learning_rate": 0.0004067794418704881, |
| "loss": 3.445, |
| "step": 30000 |
| }, |
| { |
| "epoch": 3.228931223764934, |
| "eval_accuracy": 0.37137198500459767, |
| "eval_loss": 3.49147367477417, |
| "eval_runtime": 179.7703, |
| "eval_samples_per_second": 100.189, |
| "eval_steps_per_second": 6.264, |
| "step": 30000 |
| }, |
| { |
| "epoch": 3.234312775804542, |
| "grad_norm": 0.5977009534835815, |
| "learning_rate": 0.0004064562008404266, |
| "loss": 3.4542, |
| "step": 30050 |
| }, |
| { |
| "epoch": 3.23969432784415, |
| "grad_norm": 0.5981681942939758, |
| "learning_rate": 0.0004061329598103652, |
| "loss": 3.4456, |
| "step": 30100 |
| }, |
| { |
| "epoch": 3.2450758798837587, |
| "grad_norm": 0.5974931120872498, |
| "learning_rate": 0.00040580971878030386, |
| "loss": 3.4531, |
| "step": 30150 |
| }, |
| { |
| "epoch": 3.250457431923367, |
| "grad_norm": 0.6247798800468445, |
| "learning_rate": 0.0004054864777502424, |
| "loss": 3.4451, |
| "step": 30200 |
| }, |
| { |
| "epoch": 3.255838983962975, |
| "grad_norm": 0.6065713167190552, |
| "learning_rate": 0.000405163236720181, |
| "loss": 3.4398, |
| "step": 30250 |
| }, |
| { |
| "epoch": 3.261220536002583, |
| "grad_norm": 0.5848674774169922, |
| "learning_rate": 0.00040483999569011954, |
| "loss": 3.4452, |
| "step": 30300 |
| }, |
| { |
| "epoch": 3.2666020880421915, |
| "grad_norm": 0.6002523899078369, |
| "learning_rate": 0.00040451675466005813, |
| "loss": 3.4577, |
| "step": 30350 |
| }, |
| { |
| "epoch": 3.2719836400817996, |
| "grad_norm": 0.6616047024726868, |
| "learning_rate": 0.0004041935136299967, |
| "loss": 3.4653, |
| "step": 30400 |
| }, |
| { |
| "epoch": 3.2773651921214078, |
| "grad_norm": 0.610581636428833, |
| "learning_rate": 0.0004038702725999353, |
| "loss": 3.4604, |
| "step": 30450 |
| }, |
| { |
| "epoch": 3.282746744161016, |
| "grad_norm": 0.6293004155158997, |
| "learning_rate": 0.0004035470315698739, |
| "loss": 3.4615, |
| "step": 30500 |
| }, |
| { |
| "epoch": 3.2881282962006244, |
| "grad_norm": 0.5750092267990112, |
| "learning_rate": 0.0004032237905398125, |
| "loss": 3.4553, |
| "step": 30550 |
| }, |
| { |
| "epoch": 3.2935098482402325, |
| "grad_norm": 0.5964309573173523, |
| "learning_rate": 0.00040290054950975105, |
| "loss": 3.4681, |
| "step": 30600 |
| }, |
| { |
| "epoch": 3.2988914002798406, |
| "grad_norm": 0.5653200149536133, |
| "learning_rate": 0.00040257730847968965, |
| "loss": 3.455, |
| "step": 30650 |
| }, |
| { |
| "epoch": 3.304272952319449, |
| "grad_norm": 0.5926148891448975, |
| "learning_rate": 0.0004022540674496283, |
| "loss": 3.4585, |
| "step": 30700 |
| }, |
| { |
| "epoch": 3.3096545043590573, |
| "grad_norm": 0.6698280572891235, |
| "learning_rate": 0.00040193082641956684, |
| "loss": 3.4393, |
| "step": 30750 |
| }, |
| { |
| "epoch": 3.3150360563986654, |
| "grad_norm": 0.6018702983856201, |
| "learning_rate": 0.00040160758538950543, |
| "loss": 3.4479, |
| "step": 30800 |
| }, |
| { |
| "epoch": 3.3204176084382735, |
| "grad_norm": 0.6161127090454102, |
| "learning_rate": 0.00040128434435944397, |
| "loss": 3.4725, |
| "step": 30850 |
| }, |
| { |
| "epoch": 3.3257991604778816, |
| "grad_norm": 0.6327736973762512, |
| "learning_rate": 0.00040096110332938257, |
| "loss": 3.4554, |
| "step": 30900 |
| }, |
| { |
| "epoch": 3.33118071251749, |
| "grad_norm": 0.55366051197052, |
| "learning_rate": 0.00040063786229932116, |
| "loss": 3.4484, |
| "step": 30950 |
| }, |
| { |
| "epoch": 3.3365622645570983, |
| "grad_norm": 0.6205126643180847, |
| "learning_rate": 0.00040031462126925975, |
| "loss": 3.4523, |
| "step": 31000 |
| }, |
| { |
| "epoch": 3.3365622645570983, |
| "eval_accuracy": 0.3720559549333898, |
| "eval_loss": 3.4844534397125244, |
| "eval_runtime": 180.3473, |
| "eval_samples_per_second": 99.868, |
| "eval_steps_per_second": 6.244, |
| "step": 31000 |
| }, |
| { |
| "epoch": 3.3419438165967064, |
| "grad_norm": 0.6246373057365417, |
| "learning_rate": 0.00039999138023919835, |
| "loss": 3.455, |
| "step": 31050 |
| }, |
| { |
| "epoch": 3.347325368636315, |
| "grad_norm": 0.5978841781616211, |
| "learning_rate": 0.00039966813920913694, |
| "loss": 3.4602, |
| "step": 31100 |
| }, |
| { |
| "epoch": 3.352706920675923, |
| "grad_norm": 0.6038092374801636, |
| "learning_rate": 0.0003993448981790755, |
| "loss": 3.4566, |
| "step": 31150 |
| }, |
| { |
| "epoch": 3.358088472715531, |
| "grad_norm": 0.6315546631813049, |
| "learning_rate": 0.0003990216571490141, |
| "loss": 3.4436, |
| "step": 31200 |
| }, |
| { |
| "epoch": 3.3634700247551392, |
| "grad_norm": 0.6053710579872131, |
| "learning_rate": 0.0003986984161189526, |
| "loss": 3.4482, |
| "step": 31250 |
| }, |
| { |
| "epoch": 3.368851576794748, |
| "grad_norm": 0.6105901598930359, |
| "learning_rate": 0.00039837517508889127, |
| "loss": 3.4674, |
| "step": 31300 |
| }, |
| { |
| "epoch": 3.374233128834356, |
| "grad_norm": 0.5860775113105774, |
| "learning_rate": 0.00039805193405882986, |
| "loss": 3.4504, |
| "step": 31350 |
| }, |
| { |
| "epoch": 3.379614680873964, |
| "grad_norm": 0.6133190989494324, |
| "learning_rate": 0.0003977286930287684, |
| "loss": 3.455, |
| "step": 31400 |
| }, |
| { |
| "epoch": 3.384996232913572, |
| "grad_norm": 0.6758852005004883, |
| "learning_rate": 0.000397405451998707, |
| "loss": 3.4575, |
| "step": 31450 |
| }, |
| { |
| "epoch": 3.3903777849531807, |
| "grad_norm": 0.5996787548065186, |
| "learning_rate": 0.00039708221096864554, |
| "loss": 3.4638, |
| "step": 31500 |
| }, |
| { |
| "epoch": 3.3957593369927888, |
| "grad_norm": 0.6079868674278259, |
| "learning_rate": 0.00039675896993858413, |
| "loss": 3.4638, |
| "step": 31550 |
| }, |
| { |
| "epoch": 3.401140889032397, |
| "grad_norm": 0.5919036865234375, |
| "learning_rate": 0.0003964357289085228, |
| "loss": 3.4436, |
| "step": 31600 |
| }, |
| { |
| "epoch": 3.4065224410720054, |
| "grad_norm": 0.6195621490478516, |
| "learning_rate": 0.0003961124878784613, |
| "loss": 3.4653, |
| "step": 31650 |
| }, |
| { |
| "epoch": 3.4119039931116135, |
| "grad_norm": 0.6306630969047546, |
| "learning_rate": 0.0003957892468483999, |
| "loss": 3.4508, |
| "step": 31700 |
| }, |
| { |
| "epoch": 3.4172855451512216, |
| "grad_norm": 0.6084063053131104, |
| "learning_rate": 0.0003954660058183385, |
| "loss": 3.4406, |
| "step": 31750 |
| }, |
| { |
| "epoch": 3.4226670971908297, |
| "grad_norm": 0.6287810206413269, |
| "learning_rate": 0.00039514276478827705, |
| "loss": 3.4555, |
| "step": 31800 |
| }, |
| { |
| "epoch": 3.428048649230438, |
| "grad_norm": 0.598998486995697, |
| "learning_rate": 0.0003948195237582157, |
| "loss": 3.4591, |
| "step": 31850 |
| }, |
| { |
| "epoch": 3.4334302012700464, |
| "grad_norm": 0.6260305047035217, |
| "learning_rate": 0.0003944962827281543, |
| "loss": 3.4522, |
| "step": 31900 |
| }, |
| { |
| "epoch": 3.4388117533096545, |
| "grad_norm": 0.6597923040390015, |
| "learning_rate": 0.00039417304169809284, |
| "loss": 3.4599, |
| "step": 31950 |
| }, |
| { |
| "epoch": 3.4441933053492626, |
| "grad_norm": 0.6098554134368896, |
| "learning_rate": 0.00039384980066803143, |
| "loss": 3.4609, |
| "step": 32000 |
| }, |
| { |
| "epoch": 3.4441933053492626, |
| "eval_accuracy": 0.37268233883164026, |
| "eval_loss": 3.4774160385131836, |
| "eval_runtime": 179.8913, |
| "eval_samples_per_second": 100.122, |
| "eval_steps_per_second": 6.259, |
| "step": 32000 |
| }, |
| { |
| "epoch": 3.449574857388871, |
| "grad_norm": 0.574020504951477, |
| "learning_rate": 0.00039352655963796997, |
| "loss": 3.4806, |
| "step": 32050 |
| }, |
| { |
| "epoch": 3.4549564094284793, |
| "grad_norm": 0.6245083808898926, |
| "learning_rate": 0.00039320331860790857, |
| "loss": 3.4747, |
| "step": 32100 |
| }, |
| { |
| "epoch": 3.4603379614680874, |
| "grad_norm": 0.5653278827667236, |
| "learning_rate": 0.0003928800775778472, |
| "loss": 3.4557, |
| "step": 32150 |
| }, |
| { |
| "epoch": 3.4657195135076955, |
| "grad_norm": 0.6541618704795837, |
| "learning_rate": 0.00039255683654778576, |
| "loss": 3.4788, |
| "step": 32200 |
| }, |
| { |
| "epoch": 3.471101065547304, |
| "grad_norm": 0.5819790363311768, |
| "learning_rate": 0.00039223359551772435, |
| "loss": 3.4577, |
| "step": 32250 |
| }, |
| { |
| "epoch": 3.476482617586912, |
| "grad_norm": 0.612333357334137, |
| "learning_rate": 0.00039191035448766294, |
| "loss": 3.4452, |
| "step": 32300 |
| }, |
| { |
| "epoch": 3.4818641696265202, |
| "grad_norm": 0.616150975227356, |
| "learning_rate": 0.0003915871134576015, |
| "loss": 3.445, |
| "step": 32350 |
| }, |
| { |
| "epoch": 3.4872457216661283, |
| "grad_norm": 0.6523071527481079, |
| "learning_rate": 0.0003912638724275401, |
| "loss": 3.4603, |
| "step": 32400 |
| }, |
| { |
| "epoch": 3.492627273705737, |
| "grad_norm": 0.6022084951400757, |
| "learning_rate": 0.00039094063139747873, |
| "loss": 3.4648, |
| "step": 32450 |
| }, |
| { |
| "epoch": 3.498008825745345, |
| "grad_norm": 0.6120687127113342, |
| "learning_rate": 0.00039061739036741727, |
| "loss": 3.4401, |
| "step": 32500 |
| }, |
| { |
| "epoch": 3.503390377784953, |
| "grad_norm": 0.6045283675193787, |
| "learning_rate": 0.00039029414933735586, |
| "loss": 3.4597, |
| "step": 32550 |
| }, |
| { |
| "epoch": 3.5087719298245617, |
| "grad_norm": 0.5807498097419739, |
| "learning_rate": 0.0003899709083072944, |
| "loss": 3.4839, |
| "step": 32600 |
| }, |
| { |
| "epoch": 3.5141534818641698, |
| "grad_norm": 0.6208375692367554, |
| "learning_rate": 0.000389647667277233, |
| "loss": 3.4497, |
| "step": 32650 |
| }, |
| { |
| "epoch": 3.519535033903778, |
| "grad_norm": 0.6369626522064209, |
| "learning_rate": 0.00038932442624717165, |
| "loss": 3.4686, |
| "step": 32700 |
| }, |
| { |
| "epoch": 3.524916585943386, |
| "grad_norm": 0.5826206207275391, |
| "learning_rate": 0.0003890011852171102, |
| "loss": 3.4625, |
| "step": 32750 |
| }, |
| { |
| "epoch": 3.530298137982994, |
| "grad_norm": 0.6600220203399658, |
| "learning_rate": 0.0003886779441870488, |
| "loss": 3.4547, |
| "step": 32800 |
| }, |
| { |
| "epoch": 3.5356796900226026, |
| "grad_norm": 0.6286580562591553, |
| "learning_rate": 0.0003883547031569874, |
| "loss": 3.4523, |
| "step": 32850 |
| }, |
| { |
| "epoch": 3.5410612420622107, |
| "grad_norm": 0.5954705476760864, |
| "learning_rate": 0.0003880314621269259, |
| "loss": 3.4432, |
| "step": 32900 |
| }, |
| { |
| "epoch": 3.546442794101819, |
| "grad_norm": 0.5685660243034363, |
| "learning_rate": 0.0003877082210968645, |
| "loss": 3.4673, |
| "step": 32950 |
| }, |
| { |
| "epoch": 3.5518243461414274, |
| "grad_norm": 0.5767381191253662, |
| "learning_rate": 0.00038738498006680316, |
| "loss": 3.4729, |
| "step": 33000 |
| }, |
| { |
| "epoch": 3.5518243461414274, |
| "eval_accuracy": 0.3732153898992581, |
| "eval_loss": 3.4731204509735107, |
| "eval_runtime": 180.0807, |
| "eval_samples_per_second": 100.016, |
| "eval_steps_per_second": 6.253, |
| "step": 33000 |
| }, |
| { |
| "epoch": 3.5572058981810355, |
| "grad_norm": 0.5783354043960571, |
| "learning_rate": 0.0003870617390367417, |
| "loss": 3.4849, |
| "step": 33050 |
| }, |
| { |
| "epoch": 3.5625874502206436, |
| "grad_norm": 0.6520091891288757, |
| "learning_rate": 0.0003867384980066803, |
| "loss": 3.4411, |
| "step": 33100 |
| }, |
| { |
| "epoch": 3.5679690022602517, |
| "grad_norm": 0.6067547798156738, |
| "learning_rate": 0.00038641525697661884, |
| "loss": 3.4514, |
| "step": 33150 |
| }, |
| { |
| "epoch": 3.57335055429986, |
| "grad_norm": 0.6351250410079956, |
| "learning_rate": 0.00038609201594655743, |
| "loss": 3.4697, |
| "step": 33200 |
| }, |
| { |
| "epoch": 3.5787321063394684, |
| "grad_norm": 0.5866857171058655, |
| "learning_rate": 0.000385768774916496, |
| "loss": 3.4711, |
| "step": 33250 |
| }, |
| { |
| "epoch": 3.5841136583790765, |
| "grad_norm": 0.6151683330535889, |
| "learning_rate": 0.0003854455338864346, |
| "loss": 3.4429, |
| "step": 33300 |
| }, |
| { |
| "epoch": 3.5894952104186846, |
| "grad_norm": 0.580003023147583, |
| "learning_rate": 0.0003851222928563732, |
| "loss": 3.4521, |
| "step": 33350 |
| }, |
| { |
| "epoch": 3.594876762458293, |
| "grad_norm": 0.6485602259635925, |
| "learning_rate": 0.0003847990518263118, |
| "loss": 3.4508, |
| "step": 33400 |
| }, |
| { |
| "epoch": 3.6002583144979012, |
| "grad_norm": 0.6691255569458008, |
| "learning_rate": 0.00038447581079625035, |
| "loss": 3.4681, |
| "step": 33450 |
| }, |
| { |
| "epoch": 3.6056398665375093, |
| "grad_norm": 0.5927985310554504, |
| "learning_rate": 0.00038415903458679016, |
| "loss": 3.4552, |
| "step": 33500 |
| }, |
| { |
| "epoch": 3.611021418577118, |
| "grad_norm": 0.6414245367050171, |
| "learning_rate": 0.00038383579355672875, |
| "loss": 3.4548, |
| "step": 33550 |
| }, |
| { |
| "epoch": 3.616402970616726, |
| "grad_norm": 0.6126065850257874, |
| "learning_rate": 0.00038351255252666735, |
| "loss": 3.4498, |
| "step": 33600 |
| }, |
| { |
| "epoch": 3.621784522656334, |
| "grad_norm": 0.5922813415527344, |
| "learning_rate": 0.00038318931149660594, |
| "loss": 3.4548, |
| "step": 33650 |
| }, |
| { |
| "epoch": 3.627166074695942, |
| "grad_norm": 0.6042068004608154, |
| "learning_rate": 0.00038286607046654454, |
| "loss": 3.4562, |
| "step": 33700 |
| }, |
| { |
| "epoch": 3.6325476267355503, |
| "grad_norm": 0.6460469961166382, |
| "learning_rate": 0.00038254929425708434, |
| "loss": 3.4569, |
| "step": 33750 |
| }, |
| { |
| "epoch": 3.637929178775159, |
| "grad_norm": 0.6714599132537842, |
| "learning_rate": 0.00038222605322702294, |
| "loss": 3.4775, |
| "step": 33800 |
| }, |
| { |
| "epoch": 3.643310730814767, |
| "grad_norm": 0.6167879104614258, |
| "learning_rate": 0.00038190281219696153, |
| "loss": 3.4382, |
| "step": 33850 |
| }, |
| { |
| "epoch": 3.648692282854375, |
| "grad_norm": 0.5907341241836548, |
| "learning_rate": 0.00038157957116690007, |
| "loss": 3.4661, |
| "step": 33900 |
| }, |
| { |
| "epoch": 3.6540738348939836, |
| "grad_norm": 0.6232337355613708, |
| "learning_rate": 0.00038125633013683867, |
| "loss": 3.4681, |
| "step": 33950 |
| }, |
| { |
| "epoch": 3.6594553869335917, |
| "grad_norm": 0.6158431768417358, |
| "learning_rate": 0.0003809330891067772, |
| "loss": 3.4414, |
| "step": 34000 |
| }, |
| { |
| "epoch": 3.6594553869335917, |
| "eval_accuracy": 0.3739183740834178, |
| "eval_loss": 3.4675493240356445, |
| "eval_runtime": 179.9271, |
| "eval_samples_per_second": 100.102, |
| "eval_steps_per_second": 6.258, |
| "step": 34000 |
| }, |
| { |
| "epoch": 3.6648369389732, |
| "grad_norm": 0.6306874752044678, |
| "learning_rate": 0.00038060984807671586, |
| "loss": 3.4467, |
| "step": 34050 |
| }, |
| { |
| "epoch": 3.670218491012808, |
| "grad_norm": 0.6371271014213562, |
| "learning_rate": 0.00038028660704665445, |
| "loss": 3.4585, |
| "step": 34100 |
| }, |
| { |
| "epoch": 3.675600043052416, |
| "grad_norm": 0.5911433696746826, |
| "learning_rate": 0.000379963366016593, |
| "loss": 3.4271, |
| "step": 34150 |
| }, |
| { |
| "epoch": 3.6809815950920246, |
| "grad_norm": 0.610369086265564, |
| "learning_rate": 0.0003796401249865316, |
| "loss": 3.4656, |
| "step": 34200 |
| }, |
| { |
| "epoch": 3.6863631471316327, |
| "grad_norm": 0.6166794300079346, |
| "learning_rate": 0.0003793168839564701, |
| "loss": 3.4541, |
| "step": 34250 |
| }, |
| { |
| "epoch": 3.691744699171241, |
| "grad_norm": 0.6205360889434814, |
| "learning_rate": 0.0003789936429264088, |
| "loss": 3.4503, |
| "step": 34300 |
| }, |
| { |
| "epoch": 3.6971262512108494, |
| "grad_norm": 0.6370207071304321, |
| "learning_rate": 0.00037867040189634737, |
| "loss": 3.4441, |
| "step": 34350 |
| }, |
| { |
| "epoch": 3.7025078032504575, |
| "grad_norm": 0.6453512907028198, |
| "learning_rate": 0.0003783471608662859, |
| "loss": 3.4555, |
| "step": 34400 |
| }, |
| { |
| "epoch": 3.7078893552900656, |
| "grad_norm": 0.5642420649528503, |
| "learning_rate": 0.0003780239198362245, |
| "loss": 3.45, |
| "step": 34450 |
| }, |
| { |
| "epoch": 3.713270907329674, |
| "grad_norm": 0.5417167544364929, |
| "learning_rate": 0.0003777006788061631, |
| "loss": 3.4647, |
| "step": 34500 |
| }, |
| { |
| "epoch": 3.7186524593692822, |
| "grad_norm": 0.5547471642494202, |
| "learning_rate": 0.00037737743777610164, |
| "loss": 3.4465, |
| "step": 34550 |
| }, |
| { |
| "epoch": 3.7240340114088903, |
| "grad_norm": 0.5820665955543518, |
| "learning_rate": 0.0003770541967460403, |
| "loss": 3.4561, |
| "step": 34600 |
| }, |
| { |
| "epoch": 3.7294155634484984, |
| "grad_norm": 0.5901154279708862, |
| "learning_rate": 0.0003767309557159789, |
| "loss": 3.4465, |
| "step": 34650 |
| }, |
| { |
| "epoch": 3.7347971154881066, |
| "grad_norm": 0.5940930247306824, |
| "learning_rate": 0.0003764077146859174, |
| "loss": 3.4482, |
| "step": 34700 |
| }, |
| { |
| "epoch": 3.740178667527715, |
| "grad_norm": 0.6073951721191406, |
| "learning_rate": 0.000376084473655856, |
| "loss": 3.4522, |
| "step": 34750 |
| }, |
| { |
| "epoch": 3.745560219567323, |
| "grad_norm": 0.573229968547821, |
| "learning_rate": 0.00037576123262579456, |
| "loss": 3.4248, |
| "step": 34800 |
| }, |
| { |
| "epoch": 3.7509417716069313, |
| "grad_norm": 0.6249234676361084, |
| "learning_rate": 0.00037543799159573315, |
| "loss": 3.4414, |
| "step": 34850 |
| }, |
| { |
| "epoch": 3.75632332364654, |
| "grad_norm": 0.5785616040229797, |
| "learning_rate": 0.0003751147505656718, |
| "loss": 3.4517, |
| "step": 34900 |
| }, |
| { |
| "epoch": 3.761704875686148, |
| "grad_norm": 0.5984014272689819, |
| "learning_rate": 0.00037479150953561034, |
| "loss": 3.4535, |
| "step": 34950 |
| }, |
| { |
| "epoch": 3.767086427725756, |
| "grad_norm": 0.575398862361908, |
| "learning_rate": 0.00037446826850554894, |
| "loss": 3.4656, |
| "step": 35000 |
| }, |
| { |
| "epoch": 3.767086427725756, |
| "eval_accuracy": 0.3745656193361287, |
| "eval_loss": 3.462155818939209, |
| "eval_runtime": 180.1892, |
| "eval_samples_per_second": 99.956, |
| "eval_steps_per_second": 6.249, |
| "step": 35000 |
| }, |
| { |
| "epoch": 3.772467979765364, |
| "grad_norm": 0.5907803177833557, |
| "learning_rate": 0.00037414502747548753, |
| "loss": 3.4421, |
| "step": 35050 |
| }, |
| { |
| "epoch": 3.7778495318049723, |
| "grad_norm": 0.5853913426399231, |
| "learning_rate": 0.00037382178644542607, |
| "loss": 3.4405, |
| "step": 35100 |
| }, |
| { |
| "epoch": 3.783231083844581, |
| "grad_norm": 0.6257376074790955, |
| "learning_rate": 0.00037349854541536467, |
| "loss": 3.4686, |
| "step": 35150 |
| }, |
| { |
| "epoch": 3.788612635884189, |
| "grad_norm": 0.6458983421325684, |
| "learning_rate": 0.0003731753043853033, |
| "loss": 3.4546, |
| "step": 35200 |
| }, |
| { |
| "epoch": 3.793994187923797, |
| "grad_norm": 0.612881064414978, |
| "learning_rate": 0.00037285206335524186, |
| "loss": 3.4791, |
| "step": 35250 |
| }, |
| { |
| "epoch": 3.7993757399634056, |
| "grad_norm": 0.5735159516334534, |
| "learning_rate": 0.00037252882232518045, |
| "loss": 3.4668, |
| "step": 35300 |
| }, |
| { |
| "epoch": 3.8047572920030137, |
| "grad_norm": 0.6126082539558411, |
| "learning_rate": 0.000372205581295119, |
| "loss": 3.456, |
| "step": 35350 |
| }, |
| { |
| "epoch": 3.810138844042622, |
| "grad_norm": 0.5752865672111511, |
| "learning_rate": 0.0003718823402650576, |
| "loss": 3.4603, |
| "step": 35400 |
| }, |
| { |
| "epoch": 3.8155203960822304, |
| "grad_norm": 0.6268784999847412, |
| "learning_rate": 0.00037155909923499624, |
| "loss": 3.4538, |
| "step": 35450 |
| }, |
| { |
| "epoch": 3.8209019481218385, |
| "grad_norm": 0.613398015499115, |
| "learning_rate": 0.0003712358582049348, |
| "loss": 3.4408, |
| "step": 35500 |
| }, |
| { |
| "epoch": 3.8262835001614466, |
| "grad_norm": 0.6554962992668152, |
| "learning_rate": 0.00037091261717487337, |
| "loss": 3.4481, |
| "step": 35550 |
| }, |
| { |
| "epoch": 3.8316650522010547, |
| "grad_norm": 0.6721749901771545, |
| "learning_rate": 0.00037058937614481197, |
| "loss": 3.4502, |
| "step": 35600 |
| }, |
| { |
| "epoch": 3.837046604240663, |
| "grad_norm": 0.630124568939209, |
| "learning_rate": 0.0003702661351147505, |
| "loss": 3.4629, |
| "step": 35650 |
| }, |
| { |
| "epoch": 3.8424281562802713, |
| "grad_norm": 0.5516995787620544, |
| "learning_rate": 0.0003699428940846891, |
| "loss": 3.4512, |
| "step": 35700 |
| }, |
| { |
| "epoch": 3.8478097083198795, |
| "grad_norm": 0.7088376879692078, |
| "learning_rate": 0.00036961965305462775, |
| "loss": 3.4578, |
| "step": 35750 |
| }, |
| { |
| "epoch": 3.8531912603594876, |
| "grad_norm": 0.6366024613380432, |
| "learning_rate": 0.0003692964120245663, |
| "loss": 3.4455, |
| "step": 35800 |
| }, |
| { |
| "epoch": 3.858572812399096, |
| "grad_norm": 0.6255238056182861, |
| "learning_rate": 0.0003689731709945049, |
| "loss": 3.4399, |
| "step": 35850 |
| }, |
| { |
| "epoch": 3.863954364438704, |
| "grad_norm": 0.5794286131858826, |
| "learning_rate": 0.0003686499299644434, |
| "loss": 3.4501, |
| "step": 35900 |
| }, |
| { |
| "epoch": 3.8693359164783123, |
| "grad_norm": 0.6042854189872742, |
| "learning_rate": 0.000368326688934382, |
| "loss": 3.4426, |
| "step": 35950 |
| }, |
| { |
| "epoch": 3.8747174685179204, |
| "grad_norm": 0.6024532914161682, |
| "learning_rate": 0.0003680034479043206, |
| "loss": 3.4696, |
| "step": 36000 |
| }, |
| { |
| "epoch": 3.8747174685179204, |
| "eval_accuracy": 0.3750145730685781, |
| "eval_loss": 3.4540035724639893, |
| "eval_runtime": 179.8843, |
| "eval_samples_per_second": 100.125, |
| "eval_steps_per_second": 6.26, |
| "step": 36000 |
| }, |
| { |
| "epoch": 3.8800990205575285, |
| "grad_norm": 0.5712722539901733, |
| "learning_rate": 0.0003676802068742592, |
| "loss": 3.4696, |
| "step": 36050 |
| }, |
| { |
| "epoch": 3.885480572597137, |
| "grad_norm": 0.6209331750869751, |
| "learning_rate": 0.0003673569658441978, |
| "loss": 3.4485, |
| "step": 36100 |
| }, |
| { |
| "epoch": 3.890862124636745, |
| "grad_norm": 0.573322594165802, |
| "learning_rate": 0.0003670337248141364, |
| "loss": 3.4665, |
| "step": 36150 |
| }, |
| { |
| "epoch": 3.8962436766763533, |
| "grad_norm": 0.6659927368164062, |
| "learning_rate": 0.00036671048378407494, |
| "loss": 3.437, |
| "step": 36200 |
| }, |
| { |
| "epoch": 3.901625228715962, |
| "grad_norm": 0.6092972159385681, |
| "learning_rate": 0.00036638724275401353, |
| "loss": 3.4379, |
| "step": 36250 |
| }, |
| { |
| "epoch": 3.90700678075557, |
| "grad_norm": 0.5966007113456726, |
| "learning_rate": 0.0003660640017239522, |
| "loss": 3.4472, |
| "step": 36300 |
| }, |
| { |
| "epoch": 3.912388332795178, |
| "grad_norm": 0.6116428971290588, |
| "learning_rate": 0.0003657407606938907, |
| "loss": 3.4242, |
| "step": 36350 |
| }, |
| { |
| "epoch": 3.9177698848347866, |
| "grad_norm": 0.6971980333328247, |
| "learning_rate": 0.0003654175196638293, |
| "loss": 3.4542, |
| "step": 36400 |
| }, |
| { |
| "epoch": 3.9231514368743947, |
| "grad_norm": 0.5674816966056824, |
| "learning_rate": 0.00036509427863376786, |
| "loss": 3.4674, |
| "step": 36450 |
| }, |
| { |
| "epoch": 3.928532988914003, |
| "grad_norm": 0.596181333065033, |
| "learning_rate": 0.00036477103760370645, |
| "loss": 3.4494, |
| "step": 36500 |
| }, |
| { |
| "epoch": 3.933914540953611, |
| "grad_norm": 0.6476390957832336, |
| "learning_rate": 0.00036444779657364505, |
| "loss": 3.4458, |
| "step": 36550 |
| }, |
| { |
| "epoch": 3.939296092993219, |
| "grad_norm": 0.6111776828765869, |
| "learning_rate": 0.00036412455554358364, |
| "loss": 3.4511, |
| "step": 36600 |
| }, |
| { |
| "epoch": 3.9446776450328276, |
| "grad_norm": 0.6062666773796082, |
| "learning_rate": 0.00036380131451352224, |
| "loss": 3.4604, |
| "step": 36650 |
| }, |
| { |
| "epoch": 3.9500591970724357, |
| "grad_norm": 0.6287407279014587, |
| "learning_rate": 0.00036347807348346083, |
| "loss": 3.4553, |
| "step": 36700 |
| }, |
| { |
| "epoch": 3.955440749112044, |
| "grad_norm": 0.6396850347518921, |
| "learning_rate": 0.00036315483245339937, |
| "loss": 3.4422, |
| "step": 36750 |
| }, |
| { |
| "epoch": 3.9608223011516523, |
| "grad_norm": 0.6441523432731628, |
| "learning_rate": 0.00036283159142333797, |
| "loss": 3.4455, |
| "step": 36800 |
| }, |
| { |
| "epoch": 3.9662038531912605, |
| "grad_norm": 0.5901207327842712, |
| "learning_rate": 0.0003625083503932765, |
| "loss": 3.4434, |
| "step": 36850 |
| }, |
| { |
| "epoch": 3.9715854052308686, |
| "grad_norm": 0.6204649209976196, |
| "learning_rate": 0.00036218510936321516, |
| "loss": 3.4559, |
| "step": 36900 |
| }, |
| { |
| "epoch": 3.9769669572704767, |
| "grad_norm": 0.6309411525726318, |
| "learning_rate": 0.00036186186833315375, |
| "loss": 3.4507, |
| "step": 36950 |
| }, |
| { |
| "epoch": 3.9823485093100848, |
| "grad_norm": 0.6167184114456177, |
| "learning_rate": 0.0003615386273030923, |
| "loss": 3.4349, |
| "step": 37000 |
| }, |
| { |
| "epoch": 3.9823485093100848, |
| "eval_accuracy": 0.3759814751172392, |
| "eval_loss": 3.4468631744384766, |
| "eval_runtime": 179.8588, |
| "eval_samples_per_second": 100.14, |
| "eval_steps_per_second": 6.26, |
| "step": 37000 |
| }, |
| { |
| "epoch": 3.9877300613496933, |
| "grad_norm": 0.6533148884773254, |
| "learning_rate": 0.0003612153862730309, |
| "loss": 3.4514, |
| "step": 37050 |
| }, |
| { |
| "epoch": 3.9931116133893014, |
| "grad_norm": 0.5957464575767517, |
| "learning_rate": 0.0003608921452429695, |
| "loss": 3.4542, |
| "step": 37100 |
| }, |
| { |
| "epoch": 3.9984931654289095, |
| "grad_norm": 0.6055651903152466, |
| "learning_rate": 0.000360568904212908, |
| "loss": 3.4395, |
| "step": 37150 |
| }, |
| { |
| "epoch": 4.003874717468518, |
| "grad_norm": 0.579195499420166, |
| "learning_rate": 0.00036024566318284667, |
| "loss": 3.3731, |
| "step": 37200 |
| }, |
| { |
| "epoch": 4.009256269508126, |
| "grad_norm": 0.6339341998100281, |
| "learning_rate": 0.00035992242215278526, |
| "loss": 3.3399, |
| "step": 37250 |
| }, |
| { |
| "epoch": 4.014637821547734, |
| "grad_norm": 0.6784026622772217, |
| "learning_rate": 0.0003595991811227238, |
| "loss": 3.3643, |
| "step": 37300 |
| }, |
| { |
| "epoch": 4.020019373587343, |
| "grad_norm": 0.5875265598297119, |
| "learning_rate": 0.0003592759400926624, |
| "loss": 3.356, |
| "step": 37350 |
| }, |
| { |
| "epoch": 4.0254009256269505, |
| "grad_norm": 0.6232568621635437, |
| "learning_rate": 0.00035895269906260094, |
| "loss": 3.3561, |
| "step": 37400 |
| }, |
| { |
| "epoch": 4.030782477666559, |
| "grad_norm": 0.5812575221061707, |
| "learning_rate": 0.0003586294580325396, |
| "loss": 3.3639, |
| "step": 37450 |
| }, |
| { |
| "epoch": 4.036164029706168, |
| "grad_norm": 0.622757613658905, |
| "learning_rate": 0.0003583062170024782, |
| "loss": 3.34, |
| "step": 37500 |
| }, |
| { |
| "epoch": 4.041545581745775, |
| "grad_norm": 0.6129567623138428, |
| "learning_rate": 0.0003579829759724167, |
| "loss": 3.3655, |
| "step": 37550 |
| }, |
| { |
| "epoch": 4.046927133785384, |
| "grad_norm": 0.6567339301109314, |
| "learning_rate": 0.0003576597349423553, |
| "loss": 3.3565, |
| "step": 37600 |
| }, |
| { |
| "epoch": 4.0523086858249915, |
| "grad_norm": 0.6535717248916626, |
| "learning_rate": 0.0003573364939122939, |
| "loss": 3.3767, |
| "step": 37650 |
| }, |
| { |
| "epoch": 4.0576902378646, |
| "grad_norm": 0.6005692481994629, |
| "learning_rate": 0.00035701325288223245, |
| "loss": 3.3669, |
| "step": 37700 |
| }, |
| { |
| "epoch": 4.063071789904209, |
| "grad_norm": 0.5821049213409424, |
| "learning_rate": 0.0003566900118521711, |
| "loss": 3.3626, |
| "step": 37750 |
| }, |
| { |
| "epoch": 4.068453341943816, |
| "grad_norm": 0.6144912838935852, |
| "learning_rate": 0.0003563667708221097, |
| "loss": 3.3676, |
| "step": 37800 |
| }, |
| { |
| "epoch": 4.073834893983425, |
| "grad_norm": 0.5643580555915833, |
| "learning_rate": 0.00035604352979204824, |
| "loss": 3.3654, |
| "step": 37850 |
| }, |
| { |
| "epoch": 4.079216446023033, |
| "grad_norm": 0.5872395634651184, |
| "learning_rate": 0.00035572675358258804, |
| "loss": 3.3534, |
| "step": 37900 |
| }, |
| { |
| "epoch": 4.084597998062641, |
| "grad_norm": 0.6200112700462341, |
| "learning_rate": 0.00035540351255252664, |
| "loss": 3.368, |
| "step": 37950 |
| }, |
| { |
| "epoch": 4.08997955010225, |
| "grad_norm": 0.6187693476676941, |
| "learning_rate": 0.00035508027152246523, |
| "loss": 3.3756, |
| "step": 38000 |
| }, |
| { |
| "epoch": 4.08997955010225, |
| "eval_accuracy": 0.3760597051964656, |
| "eval_loss": 3.4495081901550293, |
| "eval_runtime": 179.77, |
| "eval_samples_per_second": 100.189, |
| "eval_steps_per_second": 6.264, |
| "step": 38000 |
| }, |
| { |
| "epoch": 4.095361102141858, |
| "grad_norm": 0.6478372812271118, |
| "learning_rate": 0.0003547570304924038, |
| "loss": 3.3678, |
| "step": 38050 |
| }, |
| { |
| "epoch": 4.100742654181466, |
| "grad_norm": 0.6313759088516235, |
| "learning_rate": 0.0003544337894623424, |
| "loss": 3.3722, |
| "step": 38100 |
| }, |
| { |
| "epoch": 4.106124206221074, |
| "grad_norm": 0.6767866611480713, |
| "learning_rate": 0.000354110548432281, |
| "loss": 3.3723, |
| "step": 38150 |
| }, |
| { |
| "epoch": 4.111505758260682, |
| "grad_norm": 0.5833486318588257, |
| "learning_rate": 0.00035378730740221956, |
| "loss": 3.3513, |
| "step": 38200 |
| }, |
| { |
| "epoch": 4.1168873103002905, |
| "grad_norm": 0.5952380299568176, |
| "learning_rate": 0.00035346406637215815, |
| "loss": 3.3763, |
| "step": 38250 |
| }, |
| { |
| "epoch": 4.122268862339899, |
| "grad_norm": 0.634947657585144, |
| "learning_rate": 0.0003531408253420967, |
| "loss": 3.3668, |
| "step": 38300 |
| }, |
| { |
| "epoch": 4.127650414379507, |
| "grad_norm": 0.6541725397109985, |
| "learning_rate": 0.0003528175843120353, |
| "loss": 3.3724, |
| "step": 38350 |
| }, |
| { |
| "epoch": 4.133031966419115, |
| "grad_norm": 0.6105016469955444, |
| "learning_rate": 0.00035249434328197394, |
| "loss": 3.367, |
| "step": 38400 |
| }, |
| { |
| "epoch": 4.138413518458724, |
| "grad_norm": 0.6355590224266052, |
| "learning_rate": 0.0003521711022519125, |
| "loss": 3.3708, |
| "step": 38450 |
| }, |
| { |
| "epoch": 4.1437950704983315, |
| "grad_norm": 0.5767462253570557, |
| "learning_rate": 0.00035185432604245234, |
| "loss": 3.39, |
| "step": 38500 |
| }, |
| { |
| "epoch": 4.14917662253794, |
| "grad_norm": 0.6215607523918152, |
| "learning_rate": 0.0003515310850123909, |
| "loss": 3.3777, |
| "step": 38550 |
| }, |
| { |
| "epoch": 4.154558174577549, |
| "grad_norm": 0.6338831186294556, |
| "learning_rate": 0.00035120784398232947, |
| "loss": 3.3888, |
| "step": 38600 |
| }, |
| { |
| "epoch": 4.159939726617156, |
| "grad_norm": 0.6333625912666321, |
| "learning_rate": 0.000350884602952268, |
| "loss": 3.3751, |
| "step": 38650 |
| }, |
| { |
| "epoch": 4.165321278656765, |
| "grad_norm": 0.6430308222770691, |
| "learning_rate": 0.0003505613619222066, |
| "loss": 3.385, |
| "step": 38700 |
| }, |
| { |
| "epoch": 4.1707028306963725, |
| "grad_norm": 0.5828115344047546, |
| "learning_rate": 0.0003502381208921452, |
| "loss": 3.4078, |
| "step": 38750 |
| }, |
| { |
| "epoch": 4.176084382735981, |
| "grad_norm": 0.652651846408844, |
| "learning_rate": 0.0003499148798620838, |
| "loss": 3.3657, |
| "step": 38800 |
| }, |
| { |
| "epoch": 4.18146593477559, |
| "grad_norm": 0.6599338054656982, |
| "learning_rate": 0.0003495916388320224, |
| "loss": 3.3897, |
| "step": 38850 |
| }, |
| { |
| "epoch": 4.186847486815197, |
| "grad_norm": 0.5994952917098999, |
| "learning_rate": 0.000349268397801961, |
| "loss": 3.3771, |
| "step": 38900 |
| }, |
| { |
| "epoch": 4.192229038854806, |
| "grad_norm": 0.6271725296974182, |
| "learning_rate": 0.0003489451567718995, |
| "loss": 3.3818, |
| "step": 38950 |
| }, |
| { |
| "epoch": 4.197610590894414, |
| "grad_norm": 0.617135226726532, |
| "learning_rate": 0.0003486219157418381, |
| "loss": 3.3611, |
| "step": 39000 |
| }, |
| { |
| "epoch": 4.197610590894414, |
| "eval_accuracy": 0.37661372627143164, |
| "eval_loss": 3.4469828605651855, |
| "eval_runtime": 179.8408, |
| "eval_samples_per_second": 100.15, |
| "eval_steps_per_second": 6.261, |
| "step": 39000 |
| }, |
| { |
| "epoch": 4.202992142934022, |
| "grad_norm": 0.61525958776474, |
| "learning_rate": 0.00034829867471177677, |
| "loss": 3.3823, |
| "step": 39050 |
| }, |
| { |
| "epoch": 4.208373694973631, |
| "grad_norm": 0.638131320476532, |
| "learning_rate": 0.0003479754336817153, |
| "loss": 3.384, |
| "step": 39100 |
| }, |
| { |
| "epoch": 4.213755247013238, |
| "grad_norm": 0.6497801542282104, |
| "learning_rate": 0.0003476521926516539, |
| "loss": 3.3798, |
| "step": 39150 |
| }, |
| { |
| "epoch": 4.219136799052847, |
| "grad_norm": 0.6336971521377563, |
| "learning_rate": 0.00034732895162159245, |
| "loss": 3.3859, |
| "step": 39200 |
| }, |
| { |
| "epoch": 4.224518351092455, |
| "grad_norm": 0.7161690592765808, |
| "learning_rate": 0.00034700571059153104, |
| "loss": 3.3821, |
| "step": 39250 |
| }, |
| { |
| "epoch": 4.229899903132063, |
| "grad_norm": 0.5993850231170654, |
| "learning_rate": 0.00034668246956146963, |
| "loss": 3.3726, |
| "step": 39300 |
| }, |
| { |
| "epoch": 4.2352814551716715, |
| "grad_norm": 0.6245800256729126, |
| "learning_rate": 0.00034635922853140823, |
| "loss": 3.398, |
| "step": 39350 |
| }, |
| { |
| "epoch": 4.24066300721128, |
| "grad_norm": 0.599709689617157, |
| "learning_rate": 0.0003460359875013468, |
| "loss": 3.3838, |
| "step": 39400 |
| }, |
| { |
| "epoch": 4.246044559250888, |
| "grad_norm": 0.6475380063056946, |
| "learning_rate": 0.0003457127464712854, |
| "loss": 3.3766, |
| "step": 39450 |
| }, |
| { |
| "epoch": 4.251426111290496, |
| "grad_norm": 0.6066352128982544, |
| "learning_rate": 0.00034538950544122396, |
| "loss": 3.3921, |
| "step": 39500 |
| }, |
| { |
| "epoch": 4.256807663330104, |
| "grad_norm": 0.5933494567871094, |
| "learning_rate": 0.00034506626441116255, |
| "loss": 3.3822, |
| "step": 39550 |
| }, |
| { |
| "epoch": 4.2621892153697125, |
| "grad_norm": 0.6642129421234131, |
| "learning_rate": 0.0003447430233811011, |
| "loss": 3.3944, |
| "step": 39600 |
| }, |
| { |
| "epoch": 4.267570767409321, |
| "grad_norm": 0.6359052062034607, |
| "learning_rate": 0.00034441978235103974, |
| "loss": 3.3894, |
| "step": 39650 |
| }, |
| { |
| "epoch": 4.272952319448929, |
| "grad_norm": 0.6086246967315674, |
| "learning_rate": 0.00034409654132097834, |
| "loss": 3.3938, |
| "step": 39700 |
| }, |
| { |
| "epoch": 4.278333871488537, |
| "grad_norm": 0.6143749356269836, |
| "learning_rate": 0.0003437733002909169, |
| "loss": 3.3853, |
| "step": 39750 |
| }, |
| { |
| "epoch": 4.283715423528146, |
| "grad_norm": 0.6609334945678711, |
| "learning_rate": 0.00034345005926085547, |
| "loss": 3.3985, |
| "step": 39800 |
| }, |
| { |
| "epoch": 4.2890969755677535, |
| "grad_norm": 0.6083665490150452, |
| "learning_rate": 0.00034312681823079407, |
| "loss": 3.361, |
| "step": 39850 |
| }, |
| { |
| "epoch": 4.294478527607362, |
| "grad_norm": 0.6407836675643921, |
| "learning_rate": 0.00034280357720073266, |
| "loss": 3.3881, |
| "step": 39900 |
| }, |
| { |
| "epoch": 4.299860079646971, |
| "grad_norm": 0.6500919461250305, |
| "learning_rate": 0.00034248033617067126, |
| "loss": 3.4023, |
| "step": 39950 |
| }, |
| { |
| "epoch": 4.305241631686578, |
| "grad_norm": 0.6269468665122986, |
| "learning_rate": 0.00034215709514060985, |
| "loss": 3.3894, |
| "step": 40000 |
| }, |
| { |
| "epoch": 4.305241631686578, |
| "eval_accuracy": 0.37716155413179225, |
| "eval_loss": 3.440868377685547, |
| "eval_runtime": 179.5856, |
| "eval_samples_per_second": 100.292, |
| "eval_steps_per_second": 6.27, |
| "step": 40000 |
| }, |
| { |
| "epoch": 4.310623183726187, |
| "grad_norm": 0.6156846284866333, |
| "learning_rate": 0.0003418338541105484, |
| "loss": 3.3992, |
| "step": 40050 |
| }, |
| { |
| "epoch": 4.3160047357657945, |
| "grad_norm": 0.6544471979141235, |
| "learning_rate": 0.000341510613080487, |
| "loss": 3.3941, |
| "step": 40100 |
| }, |
| { |
| "epoch": 4.321386287805403, |
| "grad_norm": 0.6093199849128723, |
| "learning_rate": 0.0003411873720504255, |
| "loss": 3.3693, |
| "step": 40150 |
| }, |
| { |
| "epoch": 4.326767839845012, |
| "grad_norm": 0.5988791584968567, |
| "learning_rate": 0.0003408641310203642, |
| "loss": 3.3889, |
| "step": 40200 |
| }, |
| { |
| "epoch": 4.332149391884619, |
| "grad_norm": 0.640820324420929, |
| "learning_rate": 0.00034054088999030277, |
| "loss": 3.3808, |
| "step": 40250 |
| }, |
| { |
| "epoch": 4.337530943924228, |
| "grad_norm": 0.607883632183075, |
| "learning_rate": 0.0003402176489602413, |
| "loss": 3.3864, |
| "step": 40300 |
| }, |
| { |
| "epoch": 4.342912495963836, |
| "grad_norm": 0.6579638123512268, |
| "learning_rate": 0.0003398944079301799, |
| "loss": 3.394, |
| "step": 40350 |
| }, |
| { |
| "epoch": 4.348294048003444, |
| "grad_norm": 0.587094247341156, |
| "learning_rate": 0.0003395711669001185, |
| "loss": 3.3886, |
| "step": 40400 |
| }, |
| { |
| "epoch": 4.3536756000430525, |
| "grad_norm": 0.6616682410240173, |
| "learning_rate": 0.00033924792587005704, |
| "loss": 3.3742, |
| "step": 40450 |
| }, |
| { |
| "epoch": 4.359057152082661, |
| "grad_norm": 0.6799524426460266, |
| "learning_rate": 0.0003389246848399957, |
| "loss": 3.3839, |
| "step": 40500 |
| }, |
| { |
| "epoch": 4.364438704122269, |
| "grad_norm": 0.6899614930152893, |
| "learning_rate": 0.0003386014438099343, |
| "loss": 3.3992, |
| "step": 40550 |
| }, |
| { |
| "epoch": 4.369820256161877, |
| "grad_norm": 0.6630707383155823, |
| "learning_rate": 0.0003382782027798728, |
| "loss": 3.3764, |
| "step": 40600 |
| }, |
| { |
| "epoch": 4.375201808201485, |
| "grad_norm": 0.598388671875, |
| "learning_rate": 0.0003379549617498114, |
| "loss": 3.3942, |
| "step": 40650 |
| }, |
| { |
| "epoch": 4.3805833602410935, |
| "grad_norm": 0.6070818305015564, |
| "learning_rate": 0.00033763172071974996, |
| "loss": 3.3867, |
| "step": 40700 |
| }, |
| { |
| "epoch": 4.385964912280702, |
| "grad_norm": 0.5876506567001343, |
| "learning_rate": 0.0003373149445102898, |
| "loss": 3.3963, |
| "step": 40750 |
| }, |
| { |
| "epoch": 4.39134646432031, |
| "grad_norm": 0.6956170797348022, |
| "learning_rate": 0.00033699170348022836, |
| "loss": 3.3942, |
| "step": 40800 |
| }, |
| { |
| "epoch": 4.396728016359918, |
| "grad_norm": 0.6381921172142029, |
| "learning_rate": 0.000336668462450167, |
| "loss": 3.3903, |
| "step": 40850 |
| }, |
| { |
| "epoch": 4.402109568399527, |
| "grad_norm": 0.5954846143722534, |
| "learning_rate": 0.0003363452214201056, |
| "loss": 3.3904, |
| "step": 40900 |
| }, |
| { |
| "epoch": 4.4074911204391345, |
| "grad_norm": 0.6178380250930786, |
| "learning_rate": 0.00033602198039004415, |
| "loss": 3.3963, |
| "step": 40950 |
| }, |
| { |
| "epoch": 4.412872672478743, |
| "grad_norm": 0.6097917556762695, |
| "learning_rate": 0.00033569873935998274, |
| "loss": 3.3884, |
| "step": 41000 |
| }, |
| { |
| "epoch": 4.412872672478743, |
| "eval_accuracy": 0.3774454641276515, |
| "eval_loss": 3.4369654655456543, |
| "eval_runtime": 180.2414, |
| "eval_samples_per_second": 99.927, |
| "eval_steps_per_second": 6.247, |
| "step": 41000 |
| }, |
| { |
| "epoch": 4.418254224518351, |
| "grad_norm": 0.6266130208969116, |
| "learning_rate": 0.0003353754983299213, |
| "loss": 3.3913, |
| "step": 41050 |
| }, |
| { |
| "epoch": 4.423635776557959, |
| "grad_norm": 0.703901469707489, |
| "learning_rate": 0.0003350522572998599, |
| "loss": 3.3757, |
| "step": 41100 |
| }, |
| { |
| "epoch": 4.429017328597568, |
| "grad_norm": 0.6588935256004333, |
| "learning_rate": 0.0003347290162697985, |
| "loss": 3.4056, |
| "step": 41150 |
| }, |
| { |
| "epoch": 4.4343988806371755, |
| "grad_norm": 0.6265673041343689, |
| "learning_rate": 0.00033440577523973706, |
| "loss": 3.3899, |
| "step": 41200 |
| }, |
| { |
| "epoch": 4.439780432676784, |
| "grad_norm": 0.6723883152008057, |
| "learning_rate": 0.00033408253420967566, |
| "loss": 3.3989, |
| "step": 41250 |
| }, |
| { |
| "epoch": 4.445161984716393, |
| "grad_norm": 0.5973272323608398, |
| "learning_rate": 0.00033375929317961425, |
| "loss": 3.3731, |
| "step": 41300 |
| }, |
| { |
| "epoch": 4.450543536756, |
| "grad_norm": 0.6177666783332825, |
| "learning_rate": 0.0003334360521495528, |
| "loss": 3.4056, |
| "step": 41350 |
| }, |
| { |
| "epoch": 4.455925088795609, |
| "grad_norm": 0.6231200695037842, |
| "learning_rate": 0.0003331128111194914, |
| "loss": 3.3935, |
| "step": 41400 |
| }, |
| { |
| "epoch": 4.461306640835216, |
| "grad_norm": 0.6817479729652405, |
| "learning_rate": 0.00033278957008943004, |
| "loss": 3.3891, |
| "step": 41450 |
| }, |
| { |
| "epoch": 4.466688192874825, |
| "grad_norm": 0.6218167543411255, |
| "learning_rate": 0.0003324663290593686, |
| "loss": 3.3724, |
| "step": 41500 |
| }, |
| { |
| "epoch": 4.4720697449144335, |
| "grad_norm": 0.6025295257568359, |
| "learning_rate": 0.00033214308802930717, |
| "loss": 3.3758, |
| "step": 41550 |
| }, |
| { |
| "epoch": 4.477451296954041, |
| "grad_norm": 0.6618101000785828, |
| "learning_rate": 0.0003318198469992457, |
| "loss": 3.3954, |
| "step": 41600 |
| }, |
| { |
| "epoch": 4.48283284899365, |
| "grad_norm": 0.6926753520965576, |
| "learning_rate": 0.0003314966059691843, |
| "loss": 3.3904, |
| "step": 41650 |
| }, |
| { |
| "epoch": 4.488214401033258, |
| "grad_norm": 0.6272509098052979, |
| "learning_rate": 0.00033117336493912296, |
| "loss": 3.4091, |
| "step": 41700 |
| }, |
| { |
| "epoch": 4.493595953072866, |
| "grad_norm": 0.6476884484291077, |
| "learning_rate": 0.0003308501239090615, |
| "loss": 3.3659, |
| "step": 41750 |
| }, |
| { |
| "epoch": 4.4989775051124745, |
| "grad_norm": 0.6307544708251953, |
| "learning_rate": 0.0003305268828790001, |
| "loss": 3.4048, |
| "step": 41800 |
| }, |
| { |
| "epoch": 4.504359057152083, |
| "grad_norm": 0.6108989715576172, |
| "learning_rate": 0.0003302036418489387, |
| "loss": 3.383, |
| "step": 41850 |
| }, |
| { |
| "epoch": 4.509740609191691, |
| "grad_norm": 0.6389515995979309, |
| "learning_rate": 0.0003298804008188772, |
| "loss": 3.3898, |
| "step": 41900 |
| }, |
| { |
| "epoch": 4.515122161231299, |
| "grad_norm": 0.6314558386802673, |
| "learning_rate": 0.0003295571597888158, |
| "loss": 3.3848, |
| "step": 41950 |
| }, |
| { |
| "epoch": 4.520503713270907, |
| "grad_norm": 0.6694413423538208, |
| "learning_rate": 0.00032923391875875447, |
| "loss": 3.3698, |
| "step": 42000 |
| }, |
| { |
| "epoch": 4.520503713270907, |
| "eval_accuracy": 0.3781866941283219, |
| "eval_loss": 3.4318511486053467, |
| "eval_runtime": 179.6259, |
| "eval_samples_per_second": 100.27, |
| "eval_steps_per_second": 6.269, |
| "step": 42000 |
| }, |
| { |
| "epoch": 4.5258852653105155, |
| "grad_norm": 0.6342461705207825, |
| "learning_rate": 0.000328910677728693, |
| "loss": 3.3916, |
| "step": 42050 |
| }, |
| { |
| "epoch": 4.531266817350124, |
| "grad_norm": 0.6582828760147095, |
| "learning_rate": 0.0003285874366986316, |
| "loss": 3.4096, |
| "step": 42100 |
| }, |
| { |
| "epoch": 4.536648369389732, |
| "grad_norm": 0.6246548295021057, |
| "learning_rate": 0.00032826419566857015, |
| "loss": 3.3845, |
| "step": 42150 |
| }, |
| { |
| "epoch": 4.54202992142934, |
| "grad_norm": 0.6190484762191772, |
| "learning_rate": 0.00032794095463850874, |
| "loss": 3.396, |
| "step": 42200 |
| }, |
| { |
| "epoch": 4.547411473468949, |
| "grad_norm": 0.6867701411247253, |
| "learning_rate": 0.0003276177136084473, |
| "loss": 3.4047, |
| "step": 42250 |
| }, |
| { |
| "epoch": 4.5527930255085565, |
| "grad_norm": 0.6821978688240051, |
| "learning_rate": 0.00032729447257838593, |
| "loss": 3.3873, |
| "step": 42300 |
| }, |
| { |
| "epoch": 4.558174577548165, |
| "grad_norm": 0.6484827995300293, |
| "learning_rate": 0.0003269712315483245, |
| "loss": 3.382, |
| "step": 42350 |
| }, |
| { |
| "epoch": 4.563556129587774, |
| "grad_norm": 0.611552357673645, |
| "learning_rate": 0.00032664799051826306, |
| "loss": 3.3833, |
| "step": 42400 |
| }, |
| { |
| "epoch": 4.568937681627381, |
| "grad_norm": 0.6681063771247864, |
| "learning_rate": 0.00032632474948820166, |
| "loss": 3.3999, |
| "step": 42450 |
| }, |
| { |
| "epoch": 4.57431923366699, |
| "grad_norm": 0.6429010033607483, |
| "learning_rate": 0.00032600150845814025, |
| "loss": 3.3881, |
| "step": 42500 |
| }, |
| { |
| "epoch": 4.579700785706597, |
| "grad_norm": 0.6277284026145935, |
| "learning_rate": 0.0003256782674280788, |
| "loss": 3.394, |
| "step": 42550 |
| }, |
| { |
| "epoch": 4.585082337746206, |
| "grad_norm": 0.6517542600631714, |
| "learning_rate": 0.00032535502639801744, |
| "loss": 3.3899, |
| "step": 42600 |
| }, |
| { |
| "epoch": 4.5904638897858145, |
| "grad_norm": 0.5616545677185059, |
| "learning_rate": 0.00032503178536795604, |
| "loss": 3.3953, |
| "step": 42650 |
| }, |
| { |
| "epoch": 4.595845441825422, |
| "grad_norm": 0.6394187211990356, |
| "learning_rate": 0.0003247085443378946, |
| "loss": 3.3838, |
| "step": 42700 |
| }, |
| { |
| "epoch": 4.601226993865031, |
| "grad_norm": 0.604969322681427, |
| "learning_rate": 0.0003243853033078332, |
| "loss": 3.4029, |
| "step": 42750 |
| }, |
| { |
| "epoch": 4.606608545904638, |
| "grad_norm": 0.6396239995956421, |
| "learning_rate": 0.0003240620622777717, |
| "loss": 3.3879, |
| "step": 42800 |
| }, |
| { |
| "epoch": 4.611990097944247, |
| "grad_norm": 0.625914454460144, |
| "learning_rate": 0.00032373882124771036, |
| "loss": 3.39, |
| "step": 42850 |
| }, |
| { |
| "epoch": 4.6173716499838555, |
| "grad_norm": 0.6272422671318054, |
| "learning_rate": 0.00032341558021764896, |
| "loss": 3.3859, |
| "step": 42900 |
| }, |
| { |
| "epoch": 4.622753202023463, |
| "grad_norm": 0.628781795501709, |
| "learning_rate": 0.0003230923391875875, |
| "loss": 3.3976, |
| "step": 42950 |
| }, |
| { |
| "epoch": 4.628134754063072, |
| "grad_norm": 0.690066933631897, |
| "learning_rate": 0.0003227690981575261, |
| "loss": 3.3775, |
| "step": 43000 |
| }, |
| { |
| "epoch": 4.628134754063072, |
| "eval_accuracy": 0.3786348872905566, |
| "eval_loss": 3.4256374835968018, |
| "eval_runtime": 179.7124, |
| "eval_samples_per_second": 100.221, |
| "eval_steps_per_second": 6.266, |
| "step": 43000 |
| }, |
| { |
| "epoch": 4.63351630610268, |
| "grad_norm": 0.6105633974075317, |
| "learning_rate": 0.0003224458571274647, |
| "loss": 3.4001, |
| "step": 43050 |
| }, |
| { |
| "epoch": 4.638897858142288, |
| "grad_norm": 0.6365822553634644, |
| "learning_rate": 0.00032212261609740323, |
| "loss": 3.3949, |
| "step": 43100 |
| }, |
| { |
| "epoch": 4.6442794101818965, |
| "grad_norm": 0.6391448378562927, |
| "learning_rate": 0.0003217993750673419, |
| "loss": 3.3986, |
| "step": 43150 |
| }, |
| { |
| "epoch": 4.649660962221505, |
| "grad_norm": 0.645176351070404, |
| "learning_rate": 0.00032147613403728047, |
| "loss": 3.3902, |
| "step": 43200 |
| }, |
| { |
| "epoch": 4.655042514261113, |
| "grad_norm": 0.6706348657608032, |
| "learning_rate": 0.000321152893007219, |
| "loss": 3.3902, |
| "step": 43250 |
| }, |
| { |
| "epoch": 4.660424066300721, |
| "grad_norm": 0.6370499730110168, |
| "learning_rate": 0.0003208296519771576, |
| "loss": 3.4003, |
| "step": 43300 |
| }, |
| { |
| "epoch": 4.665805618340329, |
| "grad_norm": 0.6663179397583008, |
| "learning_rate": 0.00032050641094709615, |
| "loss": 3.4038, |
| "step": 43350 |
| }, |
| { |
| "epoch": 4.6711871703799375, |
| "grad_norm": 0.6630804538726807, |
| "learning_rate": 0.00032018316991703474, |
| "loss": 3.3995, |
| "step": 43400 |
| }, |
| { |
| "epoch": 4.676568722419546, |
| "grad_norm": 0.6663694977760315, |
| "learning_rate": 0.0003198599288869734, |
| "loss": 3.3829, |
| "step": 43450 |
| }, |
| { |
| "epoch": 4.681950274459154, |
| "grad_norm": 0.6295120120048523, |
| "learning_rate": 0.00031953668785691193, |
| "loss": 3.4082, |
| "step": 43500 |
| }, |
| { |
| "epoch": 4.687331826498762, |
| "grad_norm": 0.6530965566635132, |
| "learning_rate": 0.0003192134468268505, |
| "loss": 3.4099, |
| "step": 43550 |
| }, |
| { |
| "epoch": 4.692713378538371, |
| "grad_norm": 0.6678366661071777, |
| "learning_rate": 0.0003188902057967891, |
| "loss": 3.3738, |
| "step": 43600 |
| }, |
| { |
| "epoch": 4.6980949305779784, |
| "grad_norm": 0.650449275970459, |
| "learning_rate": 0.00031856696476672766, |
| "loss": 3.4114, |
| "step": 43650 |
| }, |
| { |
| "epoch": 4.703476482617587, |
| "grad_norm": 0.6517276763916016, |
| "learning_rate": 0.0003182437237366663, |
| "loss": 3.3892, |
| "step": 43700 |
| }, |
| { |
| "epoch": 4.7088580346571955, |
| "grad_norm": 0.651042103767395, |
| "learning_rate": 0.0003179204827066049, |
| "loss": 3.4098, |
| "step": 43750 |
| }, |
| { |
| "epoch": 4.714239586696803, |
| "grad_norm": 0.6547741293907166, |
| "learning_rate": 0.00031759724167654344, |
| "loss": 3.3836, |
| "step": 43800 |
| }, |
| { |
| "epoch": 4.719621138736412, |
| "grad_norm": 0.6325539946556091, |
| "learning_rate": 0.00031727400064648204, |
| "loss": 3.387, |
| "step": 43850 |
| }, |
| { |
| "epoch": 4.725002690776019, |
| "grad_norm": 0.6404955387115479, |
| "learning_rate": 0.0003169507596164206, |
| "loss": 3.3987, |
| "step": 43900 |
| }, |
| { |
| "epoch": 4.730384242815628, |
| "grad_norm": 0.6595001220703125, |
| "learning_rate": 0.0003166275185863592, |
| "loss": 3.3737, |
| "step": 43950 |
| }, |
| { |
| "epoch": 4.7357657948552365, |
| "grad_norm": 0.6610948443412781, |
| "learning_rate": 0.0003163042775562978, |
| "loss": 3.3787, |
| "step": 44000 |
| }, |
| { |
| "epoch": 4.7357657948552365, |
| "eval_accuracy": 0.37884719503334613, |
| "eval_loss": 3.4220328330993652, |
| "eval_runtime": 180.1789, |
| "eval_samples_per_second": 99.962, |
| "eval_steps_per_second": 6.249, |
| "step": 44000 |
| }, |
| { |
| "epoch": 4.741147346894844, |
| "grad_norm": 0.7104712724685669, |
| "learning_rate": 0.00031598103652623636, |
| "loss": 3.3982, |
| "step": 44050 |
| }, |
| { |
| "epoch": 4.746528898934453, |
| "grad_norm": 0.6293849945068359, |
| "learning_rate": 0.00031565779549617496, |
| "loss": 3.3637, |
| "step": 44100 |
| }, |
| { |
| "epoch": 4.751910450974061, |
| "grad_norm": 0.6018164753913879, |
| "learning_rate": 0.00031533455446611355, |
| "loss": 3.3908, |
| "step": 44150 |
| }, |
| { |
| "epoch": 4.757292003013669, |
| "grad_norm": 0.646990180015564, |
| "learning_rate": 0.0003150113134360521, |
| "loss": 3.3998, |
| "step": 44200 |
| }, |
| { |
| "epoch": 4.7626735550532775, |
| "grad_norm": 0.6214666366577148, |
| "learning_rate": 0.0003146880724059907, |
| "loss": 3.3958, |
| "step": 44250 |
| }, |
| { |
| "epoch": 4.768055107092886, |
| "grad_norm": 0.652014434337616, |
| "learning_rate": 0.00031436483137592934, |
| "loss": 3.3836, |
| "step": 44300 |
| }, |
| { |
| "epoch": 4.773436659132494, |
| "grad_norm": 0.6798874139785767, |
| "learning_rate": 0.0003140415903458679, |
| "loss": 3.3979, |
| "step": 44350 |
| }, |
| { |
| "epoch": 4.778818211172102, |
| "grad_norm": 0.6357735991477966, |
| "learning_rate": 0.00031371834931580647, |
| "loss": 3.3785, |
| "step": 44400 |
| }, |
| { |
| "epoch": 4.78419976321171, |
| "grad_norm": 0.661415696144104, |
| "learning_rate": 0.000313395108285745, |
| "loss": 3.3966, |
| "step": 44450 |
| }, |
| { |
| "epoch": 4.7895813152513185, |
| "grad_norm": 0.6242609024047852, |
| "learning_rate": 0.0003130718672556836, |
| "loss": 3.3797, |
| "step": 44500 |
| }, |
| { |
| "epoch": 4.794962867290927, |
| "grad_norm": 0.6368847489356995, |
| "learning_rate": 0.0003127486262256222, |
| "loss": 3.3922, |
| "step": 44550 |
| }, |
| { |
| "epoch": 4.800344419330535, |
| "grad_norm": 0.626862108707428, |
| "learning_rate": 0.0003124253851955608, |
| "loss": 3.3845, |
| "step": 44600 |
| }, |
| { |
| "epoch": 4.805725971370143, |
| "grad_norm": 0.6585181355476379, |
| "learning_rate": 0.0003121021441654994, |
| "loss": 3.411, |
| "step": 44650 |
| }, |
| { |
| "epoch": 4.811107523409751, |
| "grad_norm": 0.6374171376228333, |
| "learning_rate": 0.000311778903135438, |
| "loss": 3.3719, |
| "step": 44700 |
| }, |
| { |
| "epoch": 4.8164890754493594, |
| "grad_norm": 0.6377184987068176, |
| "learning_rate": 0.0003114556621053765, |
| "loss": 3.4087, |
| "step": 44750 |
| }, |
| { |
| "epoch": 4.821870627488968, |
| "grad_norm": 0.6534616351127625, |
| "learning_rate": 0.00031113888589591633, |
| "loss": 3.3769, |
| "step": 44800 |
| }, |
| { |
| "epoch": 4.827252179528576, |
| "grad_norm": 0.6503520607948303, |
| "learning_rate": 0.00031081564486585493, |
| "loss": 3.3823, |
| "step": 44850 |
| }, |
| { |
| "epoch": 4.832633731568184, |
| "grad_norm": 0.6115184426307678, |
| "learning_rate": 0.0003104924038357935, |
| "loss": 3.3727, |
| "step": 44900 |
| }, |
| { |
| "epoch": 4.838015283607793, |
| "grad_norm": 0.6773943901062012, |
| "learning_rate": 0.0003101691628057321, |
| "loss": 3.382, |
| "step": 44950 |
| }, |
| { |
| "epoch": 4.8433968356474, |
| "grad_norm": 0.64551842212677, |
| "learning_rate": 0.0003098459217756707, |
| "loss": 3.3814, |
| "step": 45000 |
| }, |
| { |
| "epoch": 4.8433968356474, |
| "eval_accuracy": 0.37974640633289874, |
| "eval_loss": 3.416212558746338, |
| "eval_runtime": 179.8221, |
| "eval_samples_per_second": 100.16, |
| "eval_steps_per_second": 6.262, |
| "step": 45000 |
| }, |
| { |
| "epoch": 4.848778387687009, |
| "grad_norm": 0.620168924331665, |
| "learning_rate": 0.0003095226807456093, |
| "loss": 3.3979, |
| "step": 45050 |
| }, |
| { |
| "epoch": 4.8541599397266175, |
| "grad_norm": 0.6280816197395325, |
| "learning_rate": 0.00030919943971554785, |
| "loss": 3.3942, |
| "step": 45100 |
| }, |
| { |
| "epoch": 4.859541491766225, |
| "grad_norm": 0.6292531490325928, |
| "learning_rate": 0.00030888266350608765, |
| "loss": 3.3716, |
| "step": 45150 |
| }, |
| { |
| "epoch": 4.864923043805834, |
| "grad_norm": 0.64167320728302, |
| "learning_rate": 0.00030855942247602625, |
| "loss": 3.3863, |
| "step": 45200 |
| }, |
| { |
| "epoch": 4.870304595845441, |
| "grad_norm": 0.6454716920852661, |
| "learning_rate": 0.00030823618144596484, |
| "loss": 3.3806, |
| "step": 45250 |
| }, |
| { |
| "epoch": 4.87568614788505, |
| "grad_norm": 0.6556320190429688, |
| "learning_rate": 0.0003079129404159035, |
| "loss": 3.4006, |
| "step": 45300 |
| }, |
| { |
| "epoch": 4.8810676999246585, |
| "grad_norm": 0.7040285468101501, |
| "learning_rate": 0.00030758969938584203, |
| "loss": 3.3745, |
| "step": 45350 |
| }, |
| { |
| "epoch": 4.886449251964266, |
| "grad_norm": 0.6378863453865051, |
| "learning_rate": 0.0003072664583557806, |
| "loss": 3.3978, |
| "step": 45400 |
| }, |
| { |
| "epoch": 4.891830804003875, |
| "grad_norm": 0.654747486114502, |
| "learning_rate": 0.00030694321732571917, |
| "loss": 3.4139, |
| "step": 45450 |
| }, |
| { |
| "epoch": 4.897212356043483, |
| "grad_norm": 0.7145013213157654, |
| "learning_rate": 0.00030661997629565776, |
| "loss": 3.3888, |
| "step": 45500 |
| }, |
| { |
| "epoch": 4.902593908083091, |
| "grad_norm": 0.7178427577018738, |
| "learning_rate": 0.0003062967352655963, |
| "loss": 3.3789, |
| "step": 45550 |
| }, |
| { |
| "epoch": 4.9079754601226995, |
| "grad_norm": 0.6693344712257385, |
| "learning_rate": 0.00030597349423553495, |
| "loss": 3.3774, |
| "step": 45600 |
| }, |
| { |
| "epoch": 4.913357012162308, |
| "grad_norm": 0.6698407530784607, |
| "learning_rate": 0.00030565025320547355, |
| "loss": 3.3774, |
| "step": 45650 |
| }, |
| { |
| "epoch": 4.918738564201916, |
| "grad_norm": 0.625595211982727, |
| "learning_rate": 0.0003053270121754121, |
| "loss": 3.3888, |
| "step": 45700 |
| }, |
| { |
| "epoch": 4.924120116241524, |
| "grad_norm": 0.6924371719360352, |
| "learning_rate": 0.0003050037711453507, |
| "loss": 3.3928, |
| "step": 45750 |
| }, |
| { |
| "epoch": 4.929501668281132, |
| "grad_norm": 0.6393285989761353, |
| "learning_rate": 0.0003046805301152893, |
| "loss": 3.3896, |
| "step": 45800 |
| }, |
| { |
| "epoch": 4.9348832203207404, |
| "grad_norm": 0.6794965267181396, |
| "learning_rate": 0.0003043572890852278, |
| "loss": 3.3805, |
| "step": 45850 |
| }, |
| { |
| "epoch": 4.940264772360349, |
| "grad_norm": 0.6193264722824097, |
| "learning_rate": 0.00030403404805516646, |
| "loss": 3.3845, |
| "step": 45900 |
| }, |
| { |
| "epoch": 4.945646324399957, |
| "grad_norm": 0.8187567591667175, |
| "learning_rate": 0.00030371080702510506, |
| "loss": 3.3807, |
| "step": 45950 |
| }, |
| { |
| "epoch": 4.951027876439565, |
| "grad_norm": 0.7009637951850891, |
| "learning_rate": 0.0003033875659950436, |
| "loss": 3.3892, |
| "step": 46000 |
| }, |
| { |
| "epoch": 4.951027876439565, |
| "eval_accuracy": 0.38006943136837124, |
| "eval_loss": 3.4113309383392334, |
| "eval_runtime": 179.7172, |
| "eval_samples_per_second": 100.219, |
| "eval_steps_per_second": 6.265, |
| "step": 46000 |
| }, |
| { |
| "epoch": 4.956409428479174, |
| "grad_norm": 0.7120588421821594, |
| "learning_rate": 0.0003030643249649822, |
| "loss": 3.3983, |
| "step": 46050 |
| }, |
| { |
| "epoch": 4.961790980518781, |
| "grad_norm": 0.66907799243927, |
| "learning_rate": 0.00030274108393492073, |
| "loss": 3.3768, |
| "step": 46100 |
| }, |
| { |
| "epoch": 4.96717253255839, |
| "grad_norm": 0.6506483554840088, |
| "learning_rate": 0.00030241784290485933, |
| "loss": 3.3993, |
| "step": 46150 |
| }, |
| { |
| "epoch": 4.9725540845979985, |
| "grad_norm": 0.6320881843566895, |
| "learning_rate": 0.000302094601874798, |
| "loss": 3.4016, |
| "step": 46200 |
| }, |
| { |
| "epoch": 4.977935636637606, |
| "grad_norm": 0.6613715291023254, |
| "learning_rate": 0.0003017713608447365, |
| "loss": 3.3842, |
| "step": 46250 |
| }, |
| { |
| "epoch": 4.983317188677215, |
| "grad_norm": 0.6871373653411865, |
| "learning_rate": 0.0003014481198146751, |
| "loss": 3.39, |
| "step": 46300 |
| }, |
| { |
| "epoch": 4.988698740716822, |
| "grad_norm": 0.6517271399497986, |
| "learning_rate": 0.0003011248787846137, |
| "loss": 3.3789, |
| "step": 46350 |
| }, |
| { |
| "epoch": 4.994080292756431, |
| "grad_norm": 0.6434972882270813, |
| "learning_rate": 0.00030080163775455225, |
| "loss": 3.4122, |
| "step": 46400 |
| }, |
| { |
| "epoch": 4.9994618447960395, |
| "grad_norm": 0.6256486773490906, |
| "learning_rate": 0.0003004783967244909, |
| "loss": 3.4034, |
| "step": 46450 |
| }, |
| { |
| "epoch": 5.004843396835647, |
| "grad_norm": 0.6456863284111023, |
| "learning_rate": 0.0003001551556944295, |
| "loss": 3.3024, |
| "step": 46500 |
| }, |
| { |
| "epoch": 5.010224948875256, |
| "grad_norm": 0.7082270383834839, |
| "learning_rate": 0.00029983191466436803, |
| "loss": 3.3113, |
| "step": 46550 |
| }, |
| { |
| "epoch": 5.015606500914864, |
| "grad_norm": 0.6580175161361694, |
| "learning_rate": 0.0002995086736343066, |
| "loss": 3.2953, |
| "step": 46600 |
| }, |
| { |
| "epoch": 5.020988052954472, |
| "grad_norm": 0.6615280508995056, |
| "learning_rate": 0.00029918543260424517, |
| "loss": 3.2981, |
| "step": 46650 |
| }, |
| { |
| "epoch": 5.0263696049940805, |
| "grad_norm": 0.6606518030166626, |
| "learning_rate": 0.0002988621915741838, |
| "loss": 3.2933, |
| "step": 46700 |
| }, |
| { |
| "epoch": 5.031751157033688, |
| "grad_norm": 0.6257070302963257, |
| "learning_rate": 0.00029853895054412236, |
| "loss": 3.2914, |
| "step": 46750 |
| }, |
| { |
| "epoch": 5.037132709073297, |
| "grad_norm": 0.6961430311203003, |
| "learning_rate": 0.00029821570951406095, |
| "loss": 3.3039, |
| "step": 46800 |
| }, |
| { |
| "epoch": 5.042514261112905, |
| "grad_norm": 0.6646069288253784, |
| "learning_rate": 0.00029789246848399955, |
| "loss": 3.3104, |
| "step": 46850 |
| }, |
| { |
| "epoch": 5.047895813152513, |
| "grad_norm": 0.6522386074066162, |
| "learning_rate": 0.00029756922745393814, |
| "loss": 3.2756, |
| "step": 46900 |
| }, |
| { |
| "epoch": 5.0532773651921215, |
| "grad_norm": 0.6988029479980469, |
| "learning_rate": 0.00029724598642387674, |
| "loss": 3.32, |
| "step": 46950 |
| }, |
| { |
| "epoch": 5.05865891723173, |
| "grad_norm": 0.6169154644012451, |
| "learning_rate": 0.0002969227453938153, |
| "loss": 3.3194, |
| "step": 47000 |
| }, |
| { |
| "epoch": 5.05865891723173, |
| "eval_accuracy": 0.3803127051841878, |
| "eval_loss": 3.414651393890381, |
| "eval_runtime": 179.7017, |
| "eval_samples_per_second": 100.227, |
| "eval_steps_per_second": 6.266, |
| "step": 47000 |
| }, |
| { |
| "epoch": 5.064040469271338, |
| "grad_norm": 0.6657958030700684, |
| "learning_rate": 0.00029659950436375387, |
| "loss": 3.3005, |
| "step": 47050 |
| }, |
| { |
| "epoch": 5.069422021310946, |
| "grad_norm": 0.6399413347244263, |
| "learning_rate": 0.00029627626333369246, |
| "loss": 3.301, |
| "step": 47100 |
| }, |
| { |
| "epoch": 5.074803573350554, |
| "grad_norm": 0.6759291887283325, |
| "learning_rate": 0.00029595302230363106, |
| "loss": 3.3139, |
| "step": 47150 |
| }, |
| { |
| "epoch": 5.080185125390162, |
| "grad_norm": 0.6591566205024719, |
| "learning_rate": 0.0002956297812735696, |
| "loss": 3.3166, |
| "step": 47200 |
| }, |
| { |
| "epoch": 5.085566677429771, |
| "grad_norm": 0.6507174372673035, |
| "learning_rate": 0.00029530654024350825, |
| "loss": 3.3268, |
| "step": 47250 |
| }, |
| { |
| "epoch": 5.090948229469379, |
| "grad_norm": 0.6434475779533386, |
| "learning_rate": 0.0002949832992134468, |
| "loss": 3.2906, |
| "step": 47300 |
| }, |
| { |
| "epoch": 5.096329781508987, |
| "grad_norm": 0.6955178380012512, |
| "learning_rate": 0.0002946600581833854, |
| "loss": 3.3277, |
| "step": 47350 |
| }, |
| { |
| "epoch": 5.101711333548596, |
| "grad_norm": 0.6766684055328369, |
| "learning_rate": 0.000294336817153324, |
| "loss": 3.3073, |
| "step": 47400 |
| }, |
| { |
| "epoch": 5.107092885588203, |
| "grad_norm": 0.655519962310791, |
| "learning_rate": 0.0002940135761232626, |
| "loss": 3.3137, |
| "step": 47450 |
| }, |
| { |
| "epoch": 5.112474437627812, |
| "grad_norm": 0.6430268287658691, |
| "learning_rate": 0.0002936903350932011, |
| "loss": 3.3108, |
| "step": 47500 |
| }, |
| { |
| "epoch": 5.1178559896674205, |
| "grad_norm": 0.6219215989112854, |
| "learning_rate": 0.0002933670940631397, |
| "loss": 3.2839, |
| "step": 47550 |
| }, |
| { |
| "epoch": 5.123237541707028, |
| "grad_norm": 0.7049906253814697, |
| "learning_rate": 0.0002930438530330783, |
| "loss": 3.3125, |
| "step": 47600 |
| }, |
| { |
| "epoch": 5.128619093746637, |
| "grad_norm": 0.630456805229187, |
| "learning_rate": 0.0002927206120030169, |
| "loss": 3.3006, |
| "step": 47650 |
| }, |
| { |
| "epoch": 5.134000645786244, |
| "grad_norm": 0.676864504814148, |
| "learning_rate": 0.0002923973709729555, |
| "loss": 3.3178, |
| "step": 47700 |
| }, |
| { |
| "epoch": 5.139382197825853, |
| "grad_norm": 0.6149758696556091, |
| "learning_rate": 0.00029207412994289403, |
| "loss": 3.3064, |
| "step": 47750 |
| }, |
| { |
| "epoch": 5.1447637498654615, |
| "grad_norm": 0.6830527782440186, |
| "learning_rate": 0.0002917508889128327, |
| "loss": 3.3164, |
| "step": 47800 |
| }, |
| { |
| "epoch": 5.150145301905069, |
| "grad_norm": 0.657446563243866, |
| "learning_rate": 0.0002914276478827712, |
| "loss": 3.2956, |
| "step": 47850 |
| }, |
| { |
| "epoch": 5.155526853944678, |
| "grad_norm": 0.67881840467453, |
| "learning_rate": 0.0002911044068527098, |
| "loss": 3.3083, |
| "step": 47900 |
| }, |
| { |
| "epoch": 5.160908405984286, |
| "grad_norm": 0.6833150386810303, |
| "learning_rate": 0.0002907811658226484, |
| "loss": 3.3295, |
| "step": 47950 |
| }, |
| { |
| "epoch": 5.166289958023894, |
| "grad_norm": 0.7698062062263489, |
| "learning_rate": 0.000290457924792587, |
| "loss": 3.306, |
| "step": 48000 |
| }, |
| { |
| "epoch": 5.166289958023894, |
| "eval_accuracy": 0.38088324149810165, |
| "eval_loss": 3.413996934890747, |
| "eval_runtime": 180.0152, |
| "eval_samples_per_second": 100.053, |
| "eval_steps_per_second": 6.255, |
| "step": 48000 |
| }, |
| { |
| "epoch": 5.1716715100635025, |
| "grad_norm": 0.6625930666923523, |
| "learning_rate": 0.00029013468376252555, |
| "loss": 3.3147, |
| "step": 48050 |
| }, |
| { |
| "epoch": 5.17705306210311, |
| "grad_norm": 0.6521696448326111, |
| "learning_rate": 0.00028981144273246414, |
| "loss": 3.3087, |
| "step": 48100 |
| }, |
| { |
| "epoch": 5.182434614142719, |
| "grad_norm": 0.6896630525588989, |
| "learning_rate": 0.00028948820170240274, |
| "loss": 3.3195, |
| "step": 48150 |
| }, |
| { |
| "epoch": 5.187816166182327, |
| "grad_norm": 0.6332021951675415, |
| "learning_rate": 0.00028916496067234133, |
| "loss": 3.3183, |
| "step": 48200 |
| }, |
| { |
| "epoch": 5.193197718221935, |
| "grad_norm": 0.6617722511291504, |
| "learning_rate": 0.0002888417196422799, |
| "loss": 3.3017, |
| "step": 48250 |
| }, |
| { |
| "epoch": 5.198579270261543, |
| "grad_norm": 0.6539857983589172, |
| "learning_rate": 0.00028851847861221847, |
| "loss": 3.3423, |
| "step": 48300 |
| }, |
| { |
| "epoch": 5.203960822301152, |
| "grad_norm": 0.6800119280815125, |
| "learning_rate": 0.00028819523758215706, |
| "loss": 3.3171, |
| "step": 48350 |
| }, |
| { |
| "epoch": 5.20934237434076, |
| "grad_norm": 0.6877304315567017, |
| "learning_rate": 0.00028787199655209566, |
| "loss": 3.3214, |
| "step": 48400 |
| }, |
| { |
| "epoch": 5.214723926380368, |
| "grad_norm": 0.6638197302818298, |
| "learning_rate": 0.00028754875552203425, |
| "loss": 3.309, |
| "step": 48450 |
| }, |
| { |
| "epoch": 5.220105478419977, |
| "grad_norm": 0.6807978749275208, |
| "learning_rate": 0.0002872255144919728, |
| "loss": 3.3352, |
| "step": 48500 |
| }, |
| { |
| "epoch": 5.225487030459584, |
| "grad_norm": 0.6976240277290344, |
| "learning_rate": 0.00028690227346191144, |
| "loss": 3.3248, |
| "step": 48550 |
| }, |
| { |
| "epoch": 5.230868582499193, |
| "grad_norm": 0.713167130947113, |
| "learning_rate": 0.00028657903243185, |
| "loss": 3.3203, |
| "step": 48600 |
| }, |
| { |
| "epoch": 5.236250134538801, |
| "grad_norm": 0.7144914269447327, |
| "learning_rate": 0.0002862557914017886, |
| "loss": 3.3233, |
| "step": 48650 |
| }, |
| { |
| "epoch": 5.241631686578409, |
| "grad_norm": 0.679750919342041, |
| "learning_rate": 0.00028593255037172717, |
| "loss": 3.3142, |
| "step": 48700 |
| }, |
| { |
| "epoch": 5.247013238618018, |
| "grad_norm": 0.6690563559532166, |
| "learning_rate": 0.00028560930934166576, |
| "loss": 3.3248, |
| "step": 48750 |
| }, |
| { |
| "epoch": 5.252394790657625, |
| "grad_norm": 0.6724905967712402, |
| "learning_rate": 0.00028528606831160436, |
| "loss": 3.3239, |
| "step": 48800 |
| }, |
| { |
| "epoch": 5.257776342697234, |
| "grad_norm": 0.6915743947029114, |
| "learning_rate": 0.0002849628272815429, |
| "loss": 3.3224, |
| "step": 48850 |
| }, |
| { |
| "epoch": 5.2631578947368425, |
| "grad_norm": 0.7154117822647095, |
| "learning_rate": 0.0002846395862514815, |
| "loss": 3.3256, |
| "step": 48900 |
| }, |
| { |
| "epoch": 5.26853944677645, |
| "grad_norm": 0.6927023530006409, |
| "learning_rate": 0.0002843163452214201, |
| "loss": 3.332, |
| "step": 48950 |
| }, |
| { |
| "epoch": 5.273920998816059, |
| "grad_norm": 0.7423051595687866, |
| "learning_rate": 0.0002839931041913587, |
| "loss": 3.3269, |
| "step": 49000 |
| }, |
| { |
| "epoch": 5.273920998816059, |
| "eval_accuracy": 0.38093919773532614, |
| "eval_loss": 3.410620927810669, |
| "eval_runtime": 179.5074, |
| "eval_samples_per_second": 100.336, |
| "eval_steps_per_second": 6.273, |
| "step": 49000 |
| }, |
| { |
| "epoch": 5.279302550855666, |
| "grad_norm": 0.6497479677200317, |
| "learning_rate": 0.0002836698631612972, |
| "loss": 3.3177, |
| "step": 49050 |
| }, |
| { |
| "epoch": 5.284684102895275, |
| "grad_norm": 0.7746565937995911, |
| "learning_rate": 0.00028334662213123587, |
| "loss": 3.3411, |
| "step": 49100 |
| }, |
| { |
| "epoch": 5.2900656549348835, |
| "grad_norm": 0.702002763748169, |
| "learning_rate": 0.0002830233811011744, |
| "loss": 3.3387, |
| "step": 49150 |
| }, |
| { |
| "epoch": 5.295447206974491, |
| "grad_norm": 0.636340320110321, |
| "learning_rate": 0.0002827066048917142, |
| "loss": 3.3153, |
| "step": 49200 |
| }, |
| { |
| "epoch": 5.3008287590141, |
| "grad_norm": 0.7339147329330444, |
| "learning_rate": 0.0002823833638616528, |
| "loss": 3.3402, |
| "step": 49250 |
| }, |
| { |
| "epoch": 5.306210311053708, |
| "grad_norm": 0.6532171964645386, |
| "learning_rate": 0.0002820601228315914, |
| "loss": 3.3329, |
| "step": 49300 |
| }, |
| { |
| "epoch": 5.311591863093316, |
| "grad_norm": 0.7197509407997131, |
| "learning_rate": 0.00028173688180153, |
| "loss": 3.3392, |
| "step": 49350 |
| }, |
| { |
| "epoch": 5.316973415132924, |
| "grad_norm": 0.7137599587440491, |
| "learning_rate": 0.00028141364077146854, |
| "loss": 3.3024, |
| "step": 49400 |
| }, |
| { |
| "epoch": 5.322354967172533, |
| "grad_norm": 0.6693756580352783, |
| "learning_rate": 0.00028109039974140714, |
| "loss": 3.3333, |
| "step": 49450 |
| }, |
| { |
| "epoch": 5.327736519212141, |
| "grad_norm": 0.6450282335281372, |
| "learning_rate": 0.00028076715871134573, |
| "loss": 3.3379, |
| "step": 49500 |
| }, |
| { |
| "epoch": 5.333118071251749, |
| "grad_norm": 0.7221931219100952, |
| "learning_rate": 0.00028044391768128433, |
| "loss": 3.3119, |
| "step": 49550 |
| }, |
| { |
| "epoch": 5.338499623291357, |
| "grad_norm": 0.7065942287445068, |
| "learning_rate": 0.0002801206766512229, |
| "loss": 3.3233, |
| "step": 49600 |
| }, |
| { |
| "epoch": 5.343881175330965, |
| "grad_norm": 0.6537820100784302, |
| "learning_rate": 0.0002797974356211615, |
| "loss": 3.3253, |
| "step": 49650 |
| }, |
| { |
| "epoch": 5.349262727370574, |
| "grad_norm": 0.6890376210212708, |
| "learning_rate": 0.00027947419459110006, |
| "loss": 3.3457, |
| "step": 49700 |
| }, |
| { |
| "epoch": 5.354644279410182, |
| "grad_norm": 0.6945475339889526, |
| "learning_rate": 0.00027915095356103865, |
| "loss": 3.3181, |
| "step": 49750 |
| }, |
| { |
| "epoch": 5.36002583144979, |
| "grad_norm": 0.6715689301490784, |
| "learning_rate": 0.00027882771253097725, |
| "loss": 3.3325, |
| "step": 49800 |
| }, |
| { |
| "epoch": 5.365407383489399, |
| "grad_norm": 0.6894263625144958, |
| "learning_rate": 0.00027850447150091584, |
| "loss": 3.345, |
| "step": 49850 |
| }, |
| { |
| "epoch": 5.370788935529006, |
| "grad_norm": 0.6533727645874023, |
| "learning_rate": 0.00027818123047085444, |
| "loss": 3.347, |
| "step": 49900 |
| }, |
| { |
| "epoch": 5.376170487568615, |
| "grad_norm": 0.7649930715560913, |
| "learning_rate": 0.000277857989440793, |
| "loss": 3.339, |
| "step": 49950 |
| }, |
| { |
| "epoch": 5.3815520396082235, |
| "grad_norm": 0.6463519930839539, |
| "learning_rate": 0.00027753474841073157, |
| "loss": 3.3199, |
| "step": 50000 |
| }, |
| { |
| "epoch": 5.3815520396082235, |
| "eval_accuracy": 0.3816024149625457, |
| "eval_loss": 3.4057817459106445, |
| "eval_runtime": 179.819, |
| "eval_samples_per_second": 100.162, |
| "eval_steps_per_second": 6.262, |
| "step": 50000 |
| }, |
| { |
| "epoch": 5.386933591647831, |
| "grad_norm": 0.6719322800636292, |
| "learning_rate": 0.00027721150738067017, |
| "loss": 3.335, |
| "step": 50050 |
| }, |
| { |
| "epoch": 5.39231514368744, |
| "grad_norm": 0.7052873373031616, |
| "learning_rate": 0.00027688826635060876, |
| "loss": 3.3272, |
| "step": 50100 |
| }, |
| { |
| "epoch": 5.397696695727047, |
| "grad_norm": 0.6729222536087036, |
| "learning_rate": 0.00027657149014114857, |
| "loss": 3.3287, |
| "step": 50150 |
| }, |
| { |
| "epoch": 5.403078247766656, |
| "grad_norm": 0.6733786463737488, |
| "learning_rate": 0.00027624824911108716, |
| "loss": 3.3366, |
| "step": 50200 |
| }, |
| { |
| "epoch": 5.4084597998062645, |
| "grad_norm": 0.6847931146621704, |
| "learning_rate": 0.0002759250080810257, |
| "loss": 3.3309, |
| "step": 50250 |
| }, |
| { |
| "epoch": 5.413841351845872, |
| "grad_norm": 0.6446109414100647, |
| "learning_rate": 0.0002756017670509643, |
| "loss": 3.3295, |
| "step": 50300 |
| }, |
| { |
| "epoch": 5.419222903885481, |
| "grad_norm": 0.7692701816558838, |
| "learning_rate": 0.0002752785260209029, |
| "loss": 3.3344, |
| "step": 50350 |
| }, |
| { |
| "epoch": 5.424604455925088, |
| "grad_norm": 0.7050197720527649, |
| "learning_rate": 0.0002749552849908415, |
| "loss": 3.3242, |
| "step": 50400 |
| }, |
| { |
| "epoch": 5.429986007964697, |
| "grad_norm": 0.7438011765480042, |
| "learning_rate": 0.0002746320439607801, |
| "loss": 3.341, |
| "step": 50450 |
| }, |
| { |
| "epoch": 5.435367560004305, |
| "grad_norm": 0.6816525459289551, |
| "learning_rate": 0.0002743088029307186, |
| "loss": 3.3404, |
| "step": 50500 |
| }, |
| { |
| "epoch": 5.440749112043913, |
| "grad_norm": 0.6949660778045654, |
| "learning_rate": 0.00027398556190065727, |
| "loss": 3.3538, |
| "step": 50550 |
| }, |
| { |
| "epoch": 5.446130664083522, |
| "grad_norm": 0.7127841114997864, |
| "learning_rate": 0.0002736623208705958, |
| "loss": 3.323, |
| "step": 50600 |
| }, |
| { |
| "epoch": 5.45151221612313, |
| "grad_norm": 0.7076895833015442, |
| "learning_rate": 0.0002733390798405344, |
| "loss": 3.3439, |
| "step": 50650 |
| }, |
| { |
| "epoch": 5.456893768162738, |
| "grad_norm": 0.6947559714317322, |
| "learning_rate": 0.000273015838810473, |
| "loss": 3.3573, |
| "step": 50700 |
| }, |
| { |
| "epoch": 5.462275320202346, |
| "grad_norm": 0.7545173764228821, |
| "learning_rate": 0.0002726925977804116, |
| "loss": 3.3316, |
| "step": 50750 |
| }, |
| { |
| "epoch": 5.467656872241955, |
| "grad_norm": 0.6546699404716492, |
| "learning_rate": 0.00027236935675035013, |
| "loss": 3.3368, |
| "step": 50800 |
| }, |
| { |
| "epoch": 5.473038424281563, |
| "grad_norm": 0.6620163917541504, |
| "learning_rate": 0.00027204611572028873, |
| "loss": 3.3503, |
| "step": 50850 |
| }, |
| { |
| "epoch": 5.478419976321171, |
| "grad_norm": 0.6684191226959229, |
| "learning_rate": 0.0002717228746902273, |
| "loss": 3.3254, |
| "step": 50900 |
| }, |
| { |
| "epoch": 5.483801528360779, |
| "grad_norm": 0.6487999558448792, |
| "learning_rate": 0.0002713996336601659, |
| "loss": 3.3357, |
| "step": 50950 |
| }, |
| { |
| "epoch": 5.489183080400387, |
| "grad_norm": 0.6728455424308777, |
| "learning_rate": 0.0002710763926301045, |
| "loss": 3.3263, |
| "step": 51000 |
| }, |
| { |
| "epoch": 5.489183080400387, |
| "eval_accuracy": 0.38202127184507056, |
| "eval_loss": 3.400080680847168, |
| "eval_runtime": 179.9731, |
| "eval_samples_per_second": 100.076, |
| "eval_steps_per_second": 6.256, |
| "step": 51000 |
| }, |
| { |
| "epoch": 5.494564632439996, |
| "grad_norm": 0.6715853810310364, |
| "learning_rate": 0.00027075315160004305, |
| "loss": 3.3271, |
| "step": 51050 |
| }, |
| { |
| "epoch": 5.499946184479604, |
| "grad_norm": 0.6638278961181641, |
| "learning_rate": 0.00027042991056998165, |
| "loss": 3.3207, |
| "step": 51100 |
| }, |
| { |
| "epoch": 5.505327736519212, |
| "grad_norm": 0.683484673500061, |
| "learning_rate": 0.00027010666953992024, |
| "loss": 3.3383, |
| "step": 51150 |
| }, |
| { |
| "epoch": 5.510709288558821, |
| "grad_norm": 0.6461046934127808, |
| "learning_rate": 0.00026978342850985884, |
| "loss": 3.3287, |
| "step": 51200 |
| }, |
| { |
| "epoch": 5.516090840598428, |
| "grad_norm": 0.6551113724708557, |
| "learning_rate": 0.0002694601874797974, |
| "loss": 3.3486, |
| "step": 51250 |
| }, |
| { |
| "epoch": 5.521472392638037, |
| "grad_norm": 0.681959331035614, |
| "learning_rate": 0.000269136946449736, |
| "loss": 3.3301, |
| "step": 51300 |
| }, |
| { |
| "epoch": 5.5268539446776455, |
| "grad_norm": 0.7404993772506714, |
| "learning_rate": 0.00026881370541967457, |
| "loss": 3.3349, |
| "step": 51350 |
| }, |
| { |
| "epoch": 5.532235496717253, |
| "grad_norm": 0.6710616946220398, |
| "learning_rate": 0.00026849046438961316, |
| "loss": 3.3252, |
| "step": 51400 |
| }, |
| { |
| "epoch": 5.537617048756862, |
| "grad_norm": 0.670371949672699, |
| "learning_rate": 0.00026816722335955176, |
| "loss": 3.3355, |
| "step": 51450 |
| }, |
| { |
| "epoch": 5.542998600796469, |
| "grad_norm": 0.6416595578193665, |
| "learning_rate": 0.00026784398232949035, |
| "loss": 3.3277, |
| "step": 51500 |
| }, |
| { |
| "epoch": 5.548380152836078, |
| "grad_norm": 0.7234741449356079, |
| "learning_rate": 0.00026752074129942895, |
| "loss": 3.3327, |
| "step": 51550 |
| }, |
| { |
| "epoch": 5.553761704875686, |
| "grad_norm": 0.6543775796890259, |
| "learning_rate": 0.0002671975002693675, |
| "loss": 3.3349, |
| "step": 51600 |
| }, |
| { |
| "epoch": 5.559143256915294, |
| "grad_norm": 0.6863155961036682, |
| "learning_rate": 0.0002668742592393061, |
| "loss": 3.3377, |
| "step": 51650 |
| }, |
| { |
| "epoch": 5.564524808954903, |
| "grad_norm": 0.6523829102516174, |
| "learning_rate": 0.0002665510182092447, |
| "loss": 3.3378, |
| "step": 51700 |
| }, |
| { |
| "epoch": 5.569906360994511, |
| "grad_norm": 0.6852605938911438, |
| "learning_rate": 0.00026622777717918327, |
| "loss": 3.3293, |
| "step": 51750 |
| }, |
| { |
| "epoch": 5.575287913034119, |
| "grad_norm": 0.7273456454277039, |
| "learning_rate": 0.0002659045361491218, |
| "loss": 3.3322, |
| "step": 51800 |
| }, |
| { |
| "epoch": 5.580669465073727, |
| "grad_norm": 0.671544075012207, |
| "learning_rate": 0.00026558129511906046, |
| "loss": 3.326, |
| "step": 51850 |
| }, |
| { |
| "epoch": 5.586051017113336, |
| "grad_norm": 0.6995226740837097, |
| "learning_rate": 0.000265258054088999, |
| "loss": 3.3282, |
| "step": 51900 |
| }, |
| { |
| "epoch": 5.591432569152944, |
| "grad_norm": 0.6662087440490723, |
| "learning_rate": 0.0002649348130589376, |
| "loss": 3.3384, |
| "step": 51950 |
| }, |
| { |
| "epoch": 5.596814121192552, |
| "grad_norm": 0.90606689453125, |
| "learning_rate": 0.0002646115720288762, |
| "loss": 3.3439, |
| "step": 52000 |
| }, |
| { |
| "epoch": 5.596814121192552, |
| "eval_accuracy": 0.3824698996188565, |
| "eval_loss": 3.3940930366516113, |
| "eval_runtime": 180.0459, |
| "eval_samples_per_second": 100.036, |
| "eval_steps_per_second": 6.254, |
| "step": 52000 |
| }, |
| { |
| "epoch": 5.60219567323216, |
| "grad_norm": 0.670542299747467, |
| "learning_rate": 0.00026428833099881473, |
| "loss": 3.3299, |
| "step": 52050 |
| }, |
| { |
| "epoch": 5.607577225271768, |
| "grad_norm": 0.6939104795455933, |
| "learning_rate": 0.0002639650899687533, |
| "loss": 3.3458, |
| "step": 52100 |
| }, |
| { |
| "epoch": 5.612958777311377, |
| "grad_norm": 0.6510215997695923, |
| "learning_rate": 0.0002636418489386919, |
| "loss": 3.3294, |
| "step": 52150 |
| }, |
| { |
| "epoch": 5.618340329350985, |
| "grad_norm": 0.6817855834960938, |
| "learning_rate": 0.0002633186079086305, |
| "loss": 3.3445, |
| "step": 52200 |
| }, |
| { |
| "epoch": 5.623721881390593, |
| "grad_norm": 0.6868756413459778, |
| "learning_rate": 0.00026299536687856905, |
| "loss": 3.3462, |
| "step": 52250 |
| }, |
| { |
| "epoch": 5.629103433430201, |
| "grad_norm": 0.6777948141098022, |
| "learning_rate": 0.0002626721258485077, |
| "loss": 3.3293, |
| "step": 52300 |
| }, |
| { |
| "epoch": 5.634484985469809, |
| "grad_norm": 0.6776952743530273, |
| "learning_rate": 0.00026234888481844624, |
| "loss": 3.3273, |
| "step": 52350 |
| }, |
| { |
| "epoch": 5.639866537509418, |
| "grad_norm": 0.7230375409126282, |
| "learning_rate": 0.00026202564378838484, |
| "loss": 3.3376, |
| "step": 52400 |
| }, |
| { |
| "epoch": 5.645248089549026, |
| "grad_norm": 0.6413059830665588, |
| "learning_rate": 0.00026170240275832343, |
| "loss": 3.3498, |
| "step": 52450 |
| }, |
| { |
| "epoch": 5.650629641588634, |
| "grad_norm": 0.6590431332588196, |
| "learning_rate": 0.00026137916172826203, |
| "loss": 3.3422, |
| "step": 52500 |
| }, |
| { |
| "epoch": 5.656011193628243, |
| "grad_norm": 0.7001703977584839, |
| "learning_rate": 0.0002610559206982006, |
| "loss": 3.3453, |
| "step": 52550 |
| }, |
| { |
| "epoch": 5.66139274566785, |
| "grad_norm": 0.7514291405677795, |
| "learning_rate": 0.00026073267966813916, |
| "loss": 3.3427, |
| "step": 52600 |
| }, |
| { |
| "epoch": 5.666774297707459, |
| "grad_norm": 0.6938708424568176, |
| "learning_rate": 0.00026040943863807776, |
| "loss": 3.3422, |
| "step": 52650 |
| }, |
| { |
| "epoch": 5.672155849747067, |
| "grad_norm": 0.6692508459091187, |
| "learning_rate": 0.00026008619760801635, |
| "loss": 3.3259, |
| "step": 52700 |
| }, |
| { |
| "epoch": 5.677537401786675, |
| "grad_norm": 0.8175403475761414, |
| "learning_rate": 0.00025976295657795495, |
| "loss": 3.3254, |
| "step": 52750 |
| }, |
| { |
| "epoch": 5.682918953826284, |
| "grad_norm": 0.7053045034408569, |
| "learning_rate": 0.0002594397155478935, |
| "loss": 3.357, |
| "step": 52800 |
| }, |
| { |
| "epoch": 5.688300505865891, |
| "grad_norm": 0.7577919363975525, |
| "learning_rate": 0.00025911647451783214, |
| "loss": 3.3368, |
| "step": 52850 |
| }, |
| { |
| "epoch": 5.6936820579055, |
| "grad_norm": 0.6919559836387634, |
| "learning_rate": 0.0002587932334877707, |
| "loss": 3.3258, |
| "step": 52900 |
| }, |
| { |
| "epoch": 5.699063609945108, |
| "grad_norm": 0.7500218152999878, |
| "learning_rate": 0.00025846999245770927, |
| "loss": 3.3301, |
| "step": 52950 |
| }, |
| { |
| "epoch": 5.704445161984716, |
| "grad_norm": 0.7383402585983276, |
| "learning_rate": 0.00025814675142764787, |
| "loss": 3.3445, |
| "step": 53000 |
| }, |
| { |
| "epoch": 5.704445161984716, |
| "eval_accuracy": 0.38280150823246634, |
| "eval_loss": 3.3909990787506104, |
| "eval_runtime": 179.6204, |
| "eval_samples_per_second": 100.273, |
| "eval_steps_per_second": 6.269, |
| "step": 53000 |
| }, |
| { |
| "epoch": 5.709826714024325, |
| "grad_norm": 0.6965733170509338, |
| "learning_rate": 0.00025782351039758646, |
| "loss": 3.3443, |
| "step": 53050 |
| }, |
| { |
| "epoch": 5.715208266063933, |
| "grad_norm": 0.7147427797317505, |
| "learning_rate": 0.000257500269367525, |
| "loss": 3.3313, |
| "step": 53100 |
| }, |
| { |
| "epoch": 5.720589818103541, |
| "grad_norm": 0.692800760269165, |
| "learning_rate": 0.0002571770283374636, |
| "loss": 3.326, |
| "step": 53150 |
| }, |
| { |
| "epoch": 5.725971370143149, |
| "grad_norm": 0.6957950592041016, |
| "learning_rate": 0.0002568537873074022, |
| "loss": 3.338, |
| "step": 53200 |
| }, |
| { |
| "epoch": 5.731352922182758, |
| "grad_norm": 0.6565383076667786, |
| "learning_rate": 0.0002565305462773408, |
| "loss": 3.3311, |
| "step": 53250 |
| }, |
| { |
| "epoch": 5.736734474222366, |
| "grad_norm": 0.6763213872909546, |
| "learning_rate": 0.0002562073052472794, |
| "loss": 3.3413, |
| "step": 53300 |
| }, |
| { |
| "epoch": 5.742116026261974, |
| "grad_norm": 0.6820769309997559, |
| "learning_rate": 0.0002558840642172179, |
| "loss": 3.3296, |
| "step": 53350 |
| }, |
| { |
| "epoch": 5.747497578301582, |
| "grad_norm": 0.7125808000564575, |
| "learning_rate": 0.00025556082318715657, |
| "loss": 3.323, |
| "step": 53400 |
| }, |
| { |
| "epoch": 5.75287913034119, |
| "grad_norm": 0.6912208795547485, |
| "learning_rate": 0.0002552375821570951, |
| "loss": 3.3392, |
| "step": 53450 |
| }, |
| { |
| "epoch": 5.758260682380799, |
| "grad_norm": 0.6613236665725708, |
| "learning_rate": 0.0002549143411270337, |
| "loss": 3.3342, |
| "step": 53500 |
| }, |
| { |
| "epoch": 5.763642234420407, |
| "grad_norm": 0.681589663028717, |
| "learning_rate": 0.0002545911000969723, |
| "loss": 3.3559, |
| "step": 53550 |
| }, |
| { |
| "epoch": 5.769023786460015, |
| "grad_norm": 0.7078532576560974, |
| "learning_rate": 0.0002542678590669109, |
| "loss": 3.3391, |
| "step": 53600 |
| }, |
| { |
| "epoch": 5.774405338499624, |
| "grad_norm": 0.6990439891815186, |
| "learning_rate": 0.00025394461803684943, |
| "loss": 3.3236, |
| "step": 53650 |
| }, |
| { |
| "epoch": 5.779786890539231, |
| "grad_norm": 0.677062451839447, |
| "learning_rate": 0.00025362137700678803, |
| "loss": 3.3194, |
| "step": 53700 |
| }, |
| { |
| "epoch": 5.78516844257884, |
| "grad_norm": 0.8788602948188782, |
| "learning_rate": 0.0002532981359767266, |
| "loss": 3.3314, |
| "step": 53750 |
| }, |
| { |
| "epoch": 5.790549994618448, |
| "grad_norm": 0.7403194904327393, |
| "learning_rate": 0.0002529748949466652, |
| "loss": 3.3294, |
| "step": 53800 |
| }, |
| { |
| "epoch": 5.795931546658056, |
| "grad_norm": 0.7146193981170654, |
| "learning_rate": 0.0002526516539166038, |
| "loss": 3.3165, |
| "step": 53850 |
| }, |
| { |
| "epoch": 5.801313098697665, |
| "grad_norm": 0.7052546739578247, |
| "learning_rate": 0.00025232841288654235, |
| "loss": 3.3448, |
| "step": 53900 |
| }, |
| { |
| "epoch": 5.806694650737272, |
| "grad_norm": 0.7031681537628174, |
| "learning_rate": 0.00025200517185648095, |
| "loss": 3.3405, |
| "step": 53950 |
| }, |
| { |
| "epoch": 5.812076202776881, |
| "grad_norm": 0.7235373258590698, |
| "learning_rate": 0.00025168193082641954, |
| "loss": 3.3478, |
| "step": 54000 |
| }, |
| { |
| "epoch": 5.812076202776881, |
| "eval_accuracy": 0.38333303815965475, |
| "eval_loss": 3.3870346546173096, |
| "eval_runtime": 180.071, |
| "eval_samples_per_second": 100.022, |
| "eval_steps_per_second": 6.253, |
| "step": 54000 |
| }, |
| { |
| "epoch": 5.817457754816489, |
| "grad_norm": 0.6720332503318787, |
| "learning_rate": 0.00025136515461695935, |
| "loss": 3.3463, |
| "step": 54050 |
| }, |
| { |
| "epoch": 5.822839306856097, |
| "grad_norm": 0.6974610090255737, |
| "learning_rate": 0.00025104191358689794, |
| "loss": 3.3252, |
| "step": 54100 |
| }, |
| { |
| "epoch": 5.828220858895706, |
| "grad_norm": 0.679633378982544, |
| "learning_rate": 0.00025071867255683654, |
| "loss": 3.3429, |
| "step": 54150 |
| }, |
| { |
| "epoch": 5.833602410935313, |
| "grad_norm": 0.7177897095680237, |
| "learning_rate": 0.00025039543152677513, |
| "loss": 3.3452, |
| "step": 54200 |
| }, |
| { |
| "epoch": 5.838983962974922, |
| "grad_norm": 0.7383708357810974, |
| "learning_rate": 0.0002500721904967137, |
| "loss": 3.3431, |
| "step": 54250 |
| }, |
| { |
| "epoch": 5.84436551501453, |
| "grad_norm": 0.7053207159042358, |
| "learning_rate": 0.00024974894946665227, |
| "loss": 3.3192, |
| "step": 54300 |
| }, |
| { |
| "epoch": 5.849747067054138, |
| "grad_norm": 0.6786372065544128, |
| "learning_rate": 0.00024942570843659086, |
| "loss": 3.3387, |
| "step": 54350 |
| }, |
| { |
| "epoch": 5.855128619093747, |
| "grad_norm": 0.7125976085662842, |
| "learning_rate": 0.00024910246740652946, |
| "loss": 3.3389, |
| "step": 54400 |
| }, |
| { |
| "epoch": 5.860510171133355, |
| "grad_norm": 0.6816068887710571, |
| "learning_rate": 0.000248779226376468, |
| "loss": 3.3362, |
| "step": 54450 |
| }, |
| { |
| "epoch": 5.865891723172963, |
| "grad_norm": 0.691918671131134, |
| "learning_rate": 0.00024845598534640665, |
| "loss": 3.3445, |
| "step": 54500 |
| }, |
| { |
| "epoch": 5.871273275212571, |
| "grad_norm": 0.7392534017562866, |
| "learning_rate": 0.0002481327443163452, |
| "loss": 3.3357, |
| "step": 54550 |
| }, |
| { |
| "epoch": 5.87665482725218, |
| "grad_norm": 0.6493129730224609, |
| "learning_rate": 0.0002478095032862838, |
| "loss": 3.3564, |
| "step": 54600 |
| }, |
| { |
| "epoch": 5.882036379291788, |
| "grad_norm": 0.6863617300987244, |
| "learning_rate": 0.0002474862622562224, |
| "loss": 3.3392, |
| "step": 54650 |
| }, |
| { |
| "epoch": 5.887417931331396, |
| "grad_norm": 0.683229923248291, |
| "learning_rate": 0.00024716302122616097, |
| "loss": 3.3235, |
| "step": 54700 |
| }, |
| { |
| "epoch": 5.892799483371004, |
| "grad_norm": 0.6831369400024414, |
| "learning_rate": 0.0002468397801960995, |
| "loss": 3.3386, |
| "step": 54750 |
| }, |
| { |
| "epoch": 5.898181035410612, |
| "grad_norm": 0.7189830541610718, |
| "learning_rate": 0.0002465165391660381, |
| "loss": 3.3238, |
| "step": 54800 |
| }, |
| { |
| "epoch": 5.903562587450221, |
| "grad_norm": 0.7210323214530945, |
| "learning_rate": 0.0002461932981359767, |
| "loss": 3.3382, |
| "step": 54850 |
| }, |
| { |
| "epoch": 5.9089441394898286, |
| "grad_norm": 0.7270302772521973, |
| "learning_rate": 0.0002458700571059153, |
| "loss": 3.3289, |
| "step": 54900 |
| }, |
| { |
| "epoch": 5.914325691529437, |
| "grad_norm": 0.7537262439727783, |
| "learning_rate": 0.0002455468160758539, |
| "loss": 3.3261, |
| "step": 54950 |
| }, |
| { |
| "epoch": 5.919707243569046, |
| "grad_norm": 0.6871313452720642, |
| "learning_rate": 0.00024522357504579243, |
| "loss": 3.3193, |
| "step": 55000 |
| }, |
| { |
| "epoch": 5.919707243569046, |
| "eval_accuracy": 0.38383121165028417, |
| "eval_loss": 3.383720636367798, |
| "eval_runtime": 179.8106, |
| "eval_samples_per_second": 100.166, |
| "eval_steps_per_second": 6.262, |
| "step": 55000 |
| }, |
| { |
| "epoch": 5.925088795608653, |
| "grad_norm": 0.7151455879211426, |
| "learning_rate": 0.000244900334015731, |
| "loss": 3.3414, |
| "step": 55050 |
| }, |
| { |
| "epoch": 5.930470347648262, |
| "grad_norm": 0.6931750178337097, |
| "learning_rate": 0.0002445770929856696, |
| "loss": 3.3263, |
| "step": 55100 |
| }, |
| { |
| "epoch": 5.93585189968787, |
| "grad_norm": 0.6594939231872559, |
| "learning_rate": 0.0002442538519556082, |
| "loss": 3.326, |
| "step": 55150 |
| }, |
| { |
| "epoch": 5.941233451727478, |
| "grad_norm": 0.6955778002738953, |
| "learning_rate": 0.0002439306109255468, |
| "loss": 3.3455, |
| "step": 55200 |
| }, |
| { |
| "epoch": 5.946615003767087, |
| "grad_norm": 0.70442795753479, |
| "learning_rate": 0.00024360736989548538, |
| "loss": 3.3577, |
| "step": 55250 |
| }, |
| { |
| "epoch": 5.951996555806694, |
| "grad_norm": 0.7052885293960571, |
| "learning_rate": 0.00024328412886542394, |
| "loss": 3.3171, |
| "step": 55300 |
| }, |
| { |
| "epoch": 5.957378107846303, |
| "grad_norm": 0.7235458493232727, |
| "learning_rate": 0.00024296088783536257, |
| "loss": 3.3458, |
| "step": 55350 |
| }, |
| { |
| "epoch": 5.962759659885911, |
| "grad_norm": 0.7326828837394714, |
| "learning_rate": 0.00024263764680530113, |
| "loss": 3.3204, |
| "step": 55400 |
| }, |
| { |
| "epoch": 5.968141211925519, |
| "grad_norm": 0.724249005317688, |
| "learning_rate": 0.0002423144057752397, |
| "loss": 3.3413, |
| "step": 55450 |
| }, |
| { |
| "epoch": 5.973522763965128, |
| "grad_norm": 0.6884929537773132, |
| "learning_rate": 0.00024199116474517832, |
| "loss": 3.3452, |
| "step": 55500 |
| }, |
| { |
| "epoch": 5.978904316004736, |
| "grad_norm": 0.714123010635376, |
| "learning_rate": 0.0002416679237151169, |
| "loss": 3.3299, |
| "step": 55550 |
| }, |
| { |
| "epoch": 5.984285868044344, |
| "grad_norm": 0.7497240900993347, |
| "learning_rate": 0.00024134468268505546, |
| "loss": 3.3456, |
| "step": 55600 |
| }, |
| { |
| "epoch": 5.989667420083952, |
| "grad_norm": 0.7143298387527466, |
| "learning_rate": 0.00024102144165499405, |
| "loss": 3.3301, |
| "step": 55650 |
| }, |
| { |
| "epoch": 5.995048972123561, |
| "grad_norm": 0.6992262601852417, |
| "learning_rate": 0.00024069820062493265, |
| "loss": 3.3372, |
| "step": 55700 |
| }, |
| { |
| "epoch": 6.000430524163169, |
| "grad_norm": 0.6653108596801758, |
| "learning_rate": 0.00024037495959487121, |
| "loss": 3.318, |
| "step": 55750 |
| }, |
| { |
| "epoch": 6.005812076202777, |
| "grad_norm": 0.7049205303192139, |
| "learning_rate": 0.0002400517185648098, |
| "loss": 3.2475, |
| "step": 55800 |
| }, |
| { |
| "epoch": 6.011193628242385, |
| "grad_norm": 0.71421217918396, |
| "learning_rate": 0.00023972847753474838, |
| "loss": 3.2392, |
| "step": 55850 |
| }, |
| { |
| "epoch": 6.016575180281993, |
| "grad_norm": 0.6809349656105042, |
| "learning_rate": 0.00023940523650468697, |
| "loss": 3.2473, |
| "step": 55900 |
| }, |
| { |
| "epoch": 6.021956732321602, |
| "grad_norm": 0.7392157316207886, |
| "learning_rate": 0.00023908199547462557, |
| "loss": 3.25, |
| "step": 55950 |
| }, |
| { |
| "epoch": 6.0273382843612096, |
| "grad_norm": 0.7034648060798645, |
| "learning_rate": 0.00023875875444456413, |
| "loss": 3.2467, |
| "step": 56000 |
| }, |
| { |
| "epoch": 6.0273382843612096, |
| "eval_accuracy": 0.38408567671354565, |
| "eval_loss": 3.383943796157837, |
| "eval_runtime": 179.8915, |
| "eval_samples_per_second": 100.121, |
| "eval_steps_per_second": 6.259, |
| "step": 56000 |
| }, |
| { |
| "epoch": 6.032719836400818, |
| "grad_norm": 0.7213525176048279, |
| "learning_rate": 0.0002384355134145027, |
| "loss": 3.2419, |
| "step": 56050 |
| }, |
| { |
| "epoch": 6.038101388440427, |
| "grad_norm": 0.7486132979393005, |
| "learning_rate": 0.00023811227238444132, |
| "loss": 3.245, |
| "step": 56100 |
| }, |
| { |
| "epoch": 6.043482940480034, |
| "grad_norm": 0.6930773258209229, |
| "learning_rate": 0.0002377890313543799, |
| "loss": 3.2324, |
| "step": 56150 |
| }, |
| { |
| "epoch": 6.048864492519643, |
| "grad_norm": 0.7631097435951233, |
| "learning_rate": 0.00023746579032431849, |
| "loss": 3.253, |
| "step": 56200 |
| }, |
| { |
| "epoch": 6.0542460445592505, |
| "grad_norm": 0.7404524087905884, |
| "learning_rate": 0.00023714254929425708, |
| "loss": 3.2553, |
| "step": 56250 |
| }, |
| { |
| "epoch": 6.059627596598859, |
| "grad_norm": 0.7180728912353516, |
| "learning_rate": 0.00023681930826419565, |
| "loss": 3.2438, |
| "step": 56300 |
| }, |
| { |
| "epoch": 6.065009148638468, |
| "grad_norm": 0.7711812853813171, |
| "learning_rate": 0.00023649606723413424, |
| "loss": 3.254, |
| "step": 56350 |
| }, |
| { |
| "epoch": 6.070390700678075, |
| "grad_norm": 0.7722243666648865, |
| "learning_rate": 0.0002361728262040728, |
| "loss": 3.2458, |
| "step": 56400 |
| }, |
| { |
| "epoch": 6.075772252717684, |
| "grad_norm": 0.6931943893432617, |
| "learning_rate": 0.00023584958517401138, |
| "loss": 3.2818, |
| "step": 56450 |
| }, |
| { |
| "epoch": 6.081153804757292, |
| "grad_norm": 0.6928170323371887, |
| "learning_rate": 0.00023552634414395, |
| "loss": 3.258, |
| "step": 56500 |
| }, |
| { |
| "epoch": 6.0865353567969, |
| "grad_norm": 0.7071146368980408, |
| "learning_rate": 0.00023520310311388857, |
| "loss": 3.2763, |
| "step": 56550 |
| }, |
| { |
| "epoch": 6.091916908836509, |
| "grad_norm": 0.7333793044090271, |
| "learning_rate": 0.00023487986208382713, |
| "loss": 3.2601, |
| "step": 56600 |
| }, |
| { |
| "epoch": 6.097298460876116, |
| "grad_norm": 0.7094366550445557, |
| "learning_rate": 0.00023455662105376576, |
| "loss": 3.2547, |
| "step": 56650 |
| }, |
| { |
| "epoch": 6.102680012915725, |
| "grad_norm": 0.7214093208312988, |
| "learning_rate": 0.00023423338002370432, |
| "loss": 3.2599, |
| "step": 56700 |
| }, |
| { |
| "epoch": 6.108061564955333, |
| "grad_norm": 0.715020477771759, |
| "learning_rate": 0.0002339101389936429, |
| "loss": 3.2696, |
| "step": 56750 |
| }, |
| { |
| "epoch": 6.113443116994941, |
| "grad_norm": 0.7385733127593994, |
| "learning_rate": 0.00023358689796358149, |
| "loss": 3.2711, |
| "step": 56800 |
| }, |
| { |
| "epoch": 6.11882466903455, |
| "grad_norm": 0.7252424359321594, |
| "learning_rate": 0.00023326365693352008, |
| "loss": 3.2499, |
| "step": 56850 |
| }, |
| { |
| "epoch": 6.124206221074158, |
| "grad_norm": 0.7170380353927612, |
| "learning_rate": 0.00023294041590345865, |
| "loss": 3.2568, |
| "step": 56900 |
| }, |
| { |
| "epoch": 6.129587773113766, |
| "grad_norm": 0.7241175174713135, |
| "learning_rate": 0.00023261717487339724, |
| "loss": 3.2735, |
| "step": 56950 |
| }, |
| { |
| "epoch": 6.134969325153374, |
| "grad_norm": 0.6951444149017334, |
| "learning_rate": 0.0002322939338433358, |
| "loss": 3.2572, |
| "step": 57000 |
| }, |
| { |
| "epoch": 6.134969325153374, |
| "eval_accuracy": 0.38414315409119953, |
| "eval_loss": 3.3844387531280518, |
| "eval_runtime": 179.7181, |
| "eval_samples_per_second": 100.218, |
| "eval_steps_per_second": 6.265, |
| "step": 57000 |
| }, |
| { |
| "epoch": 6.140350877192983, |
| "grad_norm": 0.7219548225402832, |
| "learning_rate": 0.0002319706928132744, |
| "loss": 3.2531, |
| "step": 57050 |
| }, |
| { |
| "epoch": 6.1457324292325906, |
| "grad_norm": 0.6793873310089111, |
| "learning_rate": 0.000231647451783213, |
| "loss": 3.2661, |
| "step": 57100 |
| }, |
| { |
| "epoch": 6.151113981272199, |
| "grad_norm": 0.6982064247131348, |
| "learning_rate": 0.00023132421075315157, |
| "loss": 3.2649, |
| "step": 57150 |
| }, |
| { |
| "epoch": 6.156495533311807, |
| "grad_norm": 0.7467171549797058, |
| "learning_rate": 0.0002310009697230902, |
| "loss": 3.2689, |
| "step": 57200 |
| }, |
| { |
| "epoch": 6.161877085351415, |
| "grad_norm": 0.6914950609207153, |
| "learning_rate": 0.00023067772869302876, |
| "loss": 3.2709, |
| "step": 57250 |
| }, |
| { |
| "epoch": 6.167258637391024, |
| "grad_norm": 0.7428169250488281, |
| "learning_rate": 0.00023035448766296732, |
| "loss": 3.2955, |
| "step": 57300 |
| }, |
| { |
| "epoch": 6.1726401894306315, |
| "grad_norm": 0.7214995622634888, |
| "learning_rate": 0.00023003124663290592, |
| "loss": 3.269, |
| "step": 57350 |
| }, |
| { |
| "epoch": 6.17802174147024, |
| "grad_norm": 0.7301390767097473, |
| "learning_rate": 0.0002297080056028445, |
| "loss": 3.271, |
| "step": 57400 |
| }, |
| { |
| "epoch": 6.183403293509849, |
| "grad_norm": 0.7483002543449402, |
| "learning_rate": 0.00022938476457278308, |
| "loss": 3.2873, |
| "step": 57450 |
| }, |
| { |
| "epoch": 6.188784845549456, |
| "grad_norm": 0.7099791765213013, |
| "learning_rate": 0.00022906152354272168, |
| "loss": 3.2824, |
| "step": 57500 |
| }, |
| { |
| "epoch": 6.194166397589065, |
| "grad_norm": 0.7995650768280029, |
| "learning_rate": 0.00022873828251266024, |
| "loss": 3.2704, |
| "step": 57550 |
| }, |
| { |
| "epoch": 6.1995479496286725, |
| "grad_norm": 0.7128803730010986, |
| "learning_rate": 0.00022841504148259884, |
| "loss": 3.2638, |
| "step": 57600 |
| }, |
| { |
| "epoch": 6.204929501668281, |
| "grad_norm": 0.7057095170021057, |
| "learning_rate": 0.00022809180045253743, |
| "loss": 3.2757, |
| "step": 57650 |
| }, |
| { |
| "epoch": 6.21031105370789, |
| "grad_norm": 0.7325379252433777, |
| "learning_rate": 0.000227768559422476, |
| "loss": 3.2648, |
| "step": 57700 |
| }, |
| { |
| "epoch": 6.215692605747497, |
| "grad_norm": 0.710580050945282, |
| "learning_rate": 0.00022744531839241457, |
| "loss": 3.2753, |
| "step": 57750 |
| }, |
| { |
| "epoch": 6.221074157787106, |
| "grad_norm": 0.7522643208503723, |
| "learning_rate": 0.0002271220773623532, |
| "loss": 3.2799, |
| "step": 57800 |
| }, |
| { |
| "epoch": 6.226455709826714, |
| "grad_norm": 0.7007228136062622, |
| "learning_rate": 0.00022679883633229176, |
| "loss": 3.2747, |
| "step": 57850 |
| }, |
| { |
| "epoch": 6.231837261866322, |
| "grad_norm": 0.7241410613059998, |
| "learning_rate": 0.00022647559530223032, |
| "loss": 3.2759, |
| "step": 57900 |
| }, |
| { |
| "epoch": 6.237218813905931, |
| "grad_norm": 0.7449703216552734, |
| "learning_rate": 0.00022615235427216895, |
| "loss": 3.2641, |
| "step": 57950 |
| }, |
| { |
| "epoch": 6.242600365945538, |
| "grad_norm": 0.7371934652328491, |
| "learning_rate": 0.0002258291132421075, |
| "loss": 3.2651, |
| "step": 58000 |
| }, |
| { |
| "epoch": 6.242600365945538, |
| "eval_accuracy": 0.3842598472927123, |
| "eval_loss": 3.381721258163452, |
| "eval_runtime": 179.7671, |
| "eval_samples_per_second": 100.191, |
| "eval_steps_per_second": 6.264, |
| "step": 58000 |
| }, |
| { |
| "epoch": 6.247981917985147, |
| "grad_norm": 0.7152185440063477, |
| "learning_rate": 0.00022551233703264732, |
| "loss": 3.2728, |
| "step": 58050 |
| }, |
| { |
| "epoch": 6.253363470024755, |
| "grad_norm": 0.7568090558052063, |
| "learning_rate": 0.0002251890960025859, |
| "loss": 3.2583, |
| "step": 58100 |
| }, |
| { |
| "epoch": 6.258745022064363, |
| "grad_norm": 0.7317862510681152, |
| "learning_rate": 0.0002248658549725245, |
| "loss": 3.2633, |
| "step": 58150 |
| }, |
| { |
| "epoch": 6.264126574103972, |
| "grad_norm": 0.7428773045539856, |
| "learning_rate": 0.00022454261394246308, |
| "loss": 3.2919, |
| "step": 58200 |
| }, |
| { |
| "epoch": 6.26950812614358, |
| "grad_norm": 0.7505496144294739, |
| "learning_rate": 0.00022421937291240164, |
| "loss": 3.28, |
| "step": 58250 |
| }, |
| { |
| "epoch": 6.274889678183188, |
| "grad_norm": 0.7657462954521179, |
| "learning_rate": 0.00022389613188234027, |
| "loss": 3.2791, |
| "step": 58300 |
| }, |
| { |
| "epoch": 6.280271230222796, |
| "grad_norm": 0.7743219137191772, |
| "learning_rate": 0.00022357289085227883, |
| "loss": 3.2836, |
| "step": 58350 |
| }, |
| { |
| "epoch": 6.285652782262405, |
| "grad_norm": 0.7271228432655334, |
| "learning_rate": 0.0002232496498222174, |
| "loss": 3.2818, |
| "step": 58400 |
| }, |
| { |
| "epoch": 6.2910343343020125, |
| "grad_norm": 0.7528961896896362, |
| "learning_rate": 0.000222926408792156, |
| "loss": 3.2631, |
| "step": 58450 |
| }, |
| { |
| "epoch": 6.296415886341621, |
| "grad_norm": 0.8071222305297852, |
| "learning_rate": 0.0002226031677620946, |
| "loss": 3.2787, |
| "step": 58500 |
| }, |
| { |
| "epoch": 6.301797438381229, |
| "grad_norm": 0.7826973795890808, |
| "learning_rate": 0.00022227992673203316, |
| "loss": 3.2795, |
| "step": 58550 |
| }, |
| { |
| "epoch": 6.307178990420837, |
| "grad_norm": 0.7677108645439148, |
| "learning_rate": 0.00022195668570197175, |
| "loss": 3.2784, |
| "step": 58600 |
| }, |
| { |
| "epoch": 6.312560542460446, |
| "grad_norm": 0.7295692563056946, |
| "learning_rate": 0.00022163344467191032, |
| "loss": 3.2896, |
| "step": 58650 |
| }, |
| { |
| "epoch": 6.3179420945000535, |
| "grad_norm": 0.7121778130531311, |
| "learning_rate": 0.00022131020364184891, |
| "loss": 3.2813, |
| "step": 58700 |
| }, |
| { |
| "epoch": 6.323323646539662, |
| "grad_norm": 0.743169903755188, |
| "learning_rate": 0.0002209869626117875, |
| "loss": 3.2768, |
| "step": 58750 |
| }, |
| { |
| "epoch": 6.328705198579271, |
| "grad_norm": 0.7486587762832642, |
| "learning_rate": 0.00022066372158172608, |
| "loss": 3.2684, |
| "step": 58800 |
| }, |
| { |
| "epoch": 6.334086750618878, |
| "grad_norm": 0.738730788230896, |
| "learning_rate": 0.00022034048055166464, |
| "loss": 3.2697, |
| "step": 58850 |
| }, |
| { |
| "epoch": 6.339468302658487, |
| "grad_norm": 0.7581406831741333, |
| "learning_rate": 0.00022001723952160327, |
| "loss": 3.2812, |
| "step": 58900 |
| }, |
| { |
| "epoch": 6.344849854698095, |
| "grad_norm": 0.7450671792030334, |
| "learning_rate": 0.00021969399849154183, |
| "loss": 3.2936, |
| "step": 58950 |
| }, |
| { |
| "epoch": 6.350231406737703, |
| "grad_norm": 0.8056238293647766, |
| "learning_rate": 0.00021937075746148043, |
| "loss": 3.2875, |
| "step": 59000 |
| }, |
| { |
| "epoch": 6.350231406737703, |
| "eval_accuracy": 0.38507724296774065, |
| "eval_loss": 3.3782382011413574, |
| "eval_runtime": 179.996, |
| "eval_samples_per_second": 100.063, |
| "eval_steps_per_second": 6.256, |
| "step": 59000 |
| }, |
| { |
| "epoch": 6.355612958777312, |
| "grad_norm": 0.7694650888442993, |
| "learning_rate": 0.00021904751643141902, |
| "loss": 3.2736, |
| "step": 59050 |
| }, |
| { |
| "epoch": 6.360994510816919, |
| "grad_norm": 0.8028887510299683, |
| "learning_rate": 0.0002187242754013576, |
| "loss": 3.2917, |
| "step": 59100 |
| }, |
| { |
| "epoch": 6.366376062856528, |
| "grad_norm": 0.7098097205162048, |
| "learning_rate": 0.00021840103437129619, |
| "loss": 3.2609, |
| "step": 59150 |
| }, |
| { |
| "epoch": 6.371757614896136, |
| "grad_norm": 0.7649843096733093, |
| "learning_rate": 0.00021807779334123475, |
| "loss": 3.2861, |
| "step": 59200 |
| }, |
| { |
| "epoch": 6.377139166935744, |
| "grad_norm": 0.7714524865150452, |
| "learning_rate": 0.00021775455231117335, |
| "loss": 3.2928, |
| "step": 59250 |
| }, |
| { |
| "epoch": 6.382520718975353, |
| "grad_norm": 0.7555143237113953, |
| "learning_rate": 0.00021743131128111194, |
| "loss": 3.2814, |
| "step": 59300 |
| }, |
| { |
| "epoch": 6.387902271014961, |
| "grad_norm": 0.7391084432601929, |
| "learning_rate": 0.0002171080702510505, |
| "loss": 3.277, |
| "step": 59350 |
| }, |
| { |
| "epoch": 6.393283823054569, |
| "grad_norm": 0.7586864233016968, |
| "learning_rate": 0.00021678482922098908, |
| "loss": 3.2753, |
| "step": 59400 |
| }, |
| { |
| "epoch": 6.398665375094177, |
| "grad_norm": 0.7793105840682983, |
| "learning_rate": 0.0002164615881909277, |
| "loss": 3.283, |
| "step": 59450 |
| }, |
| { |
| "epoch": 6.404046927133785, |
| "grad_norm": 0.7611179351806641, |
| "learning_rate": 0.00021613834716086627, |
| "loss": 3.2817, |
| "step": 59500 |
| }, |
| { |
| "epoch": 6.4094284791733935, |
| "grad_norm": 0.7332265973091125, |
| "learning_rate": 0.00021581510613080483, |
| "loss": 3.2797, |
| "step": 59550 |
| }, |
| { |
| "epoch": 6.414810031213002, |
| "grad_norm": 0.7599790096282959, |
| "learning_rate": 0.00021549186510074346, |
| "loss": 3.2779, |
| "step": 59600 |
| }, |
| { |
| "epoch": 6.42019158325261, |
| "grad_norm": 0.7514798045158386, |
| "learning_rate": 0.00021516862407068202, |
| "loss": 3.2788, |
| "step": 59650 |
| }, |
| { |
| "epoch": 6.425573135292218, |
| "grad_norm": 0.7226225137710571, |
| "learning_rate": 0.0002148453830406206, |
| "loss": 3.3054, |
| "step": 59700 |
| }, |
| { |
| "epoch": 6.430954687331827, |
| "grad_norm": 0.7565598487854004, |
| "learning_rate": 0.00021452214201055919, |
| "loss": 3.2657, |
| "step": 59750 |
| }, |
| { |
| "epoch": 6.4363362393714345, |
| "grad_norm": 0.7199046015739441, |
| "learning_rate": 0.00021419890098049778, |
| "loss": 3.2656, |
| "step": 59800 |
| }, |
| { |
| "epoch": 6.441717791411043, |
| "grad_norm": 0.7737821340560913, |
| "learning_rate": 0.00021387565995043638, |
| "loss": 3.285, |
| "step": 59850 |
| }, |
| { |
| "epoch": 6.447099343450651, |
| "grad_norm": 0.7489638924598694, |
| "learning_rate": 0.00021355241892037494, |
| "loss": 3.2728, |
| "step": 59900 |
| }, |
| { |
| "epoch": 6.452480895490259, |
| "grad_norm": 0.8023089170455933, |
| "learning_rate": 0.0002132291778903135, |
| "loss": 3.2688, |
| "step": 59950 |
| }, |
| { |
| "epoch": 6.457862447529868, |
| "grad_norm": 0.7359650731086731, |
| "learning_rate": 0.00021290593686025213, |
| "loss": 3.2697, |
| "step": 60000 |
| }, |
| { |
| "epoch": 6.457862447529868, |
| "eval_accuracy": 0.38522142534987047, |
| "eval_loss": 3.3750526905059814, |
| "eval_runtime": 179.9044, |
| "eval_samples_per_second": 100.114, |
| "eval_steps_per_second": 6.259, |
| "step": 60000 |
| }, |
| { |
| "epoch": 6.4632439995694755, |
| "grad_norm": 0.739447832107544, |
| "learning_rate": 0.0002125891606507919, |
| "loss": 3.2756, |
| "step": 60050 |
| }, |
| { |
| "epoch": 6.468625551609084, |
| "grad_norm": 0.7175700068473816, |
| "learning_rate": 0.0002122659196207305, |
| "loss": 3.2523, |
| "step": 60100 |
| }, |
| { |
| "epoch": 6.474007103648693, |
| "grad_norm": 0.7286744713783264, |
| "learning_rate": 0.0002119426785906691, |
| "loss": 3.2838, |
| "step": 60150 |
| }, |
| { |
| "epoch": 6.4793886556883, |
| "grad_norm": 0.765244722366333, |
| "learning_rate": 0.0002116259023812089, |
| "loss": 3.2948, |
| "step": 60200 |
| }, |
| { |
| "epoch": 6.484770207727909, |
| "grad_norm": 0.7369528412818909, |
| "learning_rate": 0.00021130266135114747, |
| "loss": 3.2776, |
| "step": 60250 |
| }, |
| { |
| "epoch": 6.490151759767517, |
| "grad_norm": 0.7986508011817932, |
| "learning_rate": 0.00021097942032108607, |
| "loss": 3.3132, |
| "step": 60300 |
| }, |
| { |
| "epoch": 6.495533311807125, |
| "grad_norm": 0.7739220261573792, |
| "learning_rate": 0.00021065617929102466, |
| "loss": 3.2838, |
| "step": 60350 |
| }, |
| { |
| "epoch": 6.500914863846734, |
| "grad_norm": 0.7373184561729431, |
| "learning_rate": 0.00021033293826096323, |
| "loss": 3.297, |
| "step": 60400 |
| }, |
| { |
| "epoch": 6.506296415886341, |
| "grad_norm": 0.7406927347183228, |
| "learning_rate": 0.0002100096972309018, |
| "loss": 3.2618, |
| "step": 60450 |
| }, |
| { |
| "epoch": 6.51167796792595, |
| "grad_norm": 0.7355625033378601, |
| "learning_rate": 0.00020968645620084042, |
| "loss": 3.2744, |
| "step": 60500 |
| }, |
| { |
| "epoch": 6.517059519965558, |
| "grad_norm": 0.7376171946525574, |
| "learning_rate": 0.000209363215170779, |
| "loss": 3.2761, |
| "step": 60550 |
| }, |
| { |
| "epoch": 6.522441072005166, |
| "grad_norm": 0.76179039478302, |
| "learning_rate": 0.00020903997414071758, |
| "loss": 3.2946, |
| "step": 60600 |
| }, |
| { |
| "epoch": 6.5278226240447745, |
| "grad_norm": 0.7716966271400452, |
| "learning_rate": 0.00020871673311065618, |
| "loss": 3.2846, |
| "step": 60650 |
| }, |
| { |
| "epoch": 6.533204176084383, |
| "grad_norm": 0.7968785166740417, |
| "learning_rate": 0.00020839349208059475, |
| "loss": 3.2766, |
| "step": 60700 |
| }, |
| { |
| "epoch": 6.538585728123991, |
| "grad_norm": 0.7669464349746704, |
| "learning_rate": 0.00020807025105053334, |
| "loss": 3.2771, |
| "step": 60750 |
| }, |
| { |
| "epoch": 6.543967280163599, |
| "grad_norm": 0.7410077452659607, |
| "learning_rate": 0.0002077470100204719, |
| "loss": 3.2788, |
| "step": 60800 |
| }, |
| { |
| "epoch": 6.549348832203208, |
| "grad_norm": 0.7026287913322449, |
| "learning_rate": 0.00020742376899041048, |
| "loss": 3.2826, |
| "step": 60850 |
| }, |
| { |
| "epoch": 6.5547303842428155, |
| "grad_norm": 0.7268006205558777, |
| "learning_rate": 0.0002071005279603491, |
| "loss": 3.2779, |
| "step": 60900 |
| }, |
| { |
| "epoch": 6.560111936282424, |
| "grad_norm": 0.7330572605133057, |
| "learning_rate": 0.00020677728693028766, |
| "loss": 3.2736, |
| "step": 60950 |
| }, |
| { |
| "epoch": 6.565493488322032, |
| "grad_norm": 0.7642152905464172, |
| "learning_rate": 0.00020645404590022623, |
| "loss": 3.2909, |
| "step": 61000 |
| }, |
| { |
| "epoch": 6.565493488322032, |
| "eval_accuracy": 0.3857069951055134, |
| "eval_loss": 3.372335433959961, |
| "eval_runtime": 179.717, |
| "eval_samples_per_second": 100.219, |
| "eval_steps_per_second": 6.265, |
| "step": 61000 |
| }, |
| { |
| "epoch": 6.57087504036164, |
| "grad_norm": 0.7624441981315613, |
| "learning_rate": 0.00020613080487016485, |
| "loss": 3.2991, |
| "step": 61050 |
| }, |
| { |
| "epoch": 6.576256592401249, |
| "grad_norm": 0.7940060496330261, |
| "learning_rate": 0.00020580756384010342, |
| "loss": 3.302, |
| "step": 61100 |
| }, |
| { |
| "epoch": 6.5816381444408565, |
| "grad_norm": 0.6994228959083557, |
| "learning_rate": 0.000205484322810042, |
| "loss": 3.2817, |
| "step": 61150 |
| }, |
| { |
| "epoch": 6.587019696480465, |
| "grad_norm": 0.7420439124107361, |
| "learning_rate": 0.00020516108177998058, |
| "loss": 3.2781, |
| "step": 61200 |
| }, |
| { |
| "epoch": 6.592401248520073, |
| "grad_norm": 0.7529013156890869, |
| "learning_rate": 0.00020483784074991918, |
| "loss": 3.2846, |
| "step": 61250 |
| }, |
| { |
| "epoch": 6.597782800559681, |
| "grad_norm": 0.7393249273300171, |
| "learning_rate": 0.00020451459971985775, |
| "loss": 3.2994, |
| "step": 61300 |
| }, |
| { |
| "epoch": 6.60316435259929, |
| "grad_norm": 0.8019313216209412, |
| "learning_rate": 0.00020419135868979634, |
| "loss": 3.2977, |
| "step": 61350 |
| }, |
| { |
| "epoch": 6.608545904638898, |
| "grad_norm": 0.7454643845558167, |
| "learning_rate": 0.0002038681176597349, |
| "loss": 3.2705, |
| "step": 61400 |
| }, |
| { |
| "epoch": 6.613927456678506, |
| "grad_norm": 0.7411048412322998, |
| "learning_rate": 0.0002035448766296735, |
| "loss": 3.2781, |
| "step": 61450 |
| }, |
| { |
| "epoch": 6.619309008718115, |
| "grad_norm": 0.747560441493988, |
| "learning_rate": 0.0002032216355996121, |
| "loss": 3.2803, |
| "step": 61500 |
| }, |
| { |
| "epoch": 6.624690560757722, |
| "grad_norm": 0.7290646433830261, |
| "learning_rate": 0.00020289839456955067, |
| "loss": 3.2812, |
| "step": 61550 |
| }, |
| { |
| "epoch": 6.630072112797331, |
| "grad_norm": 0.757982611656189, |
| "learning_rate": 0.0002025751535394893, |
| "loss": 3.2954, |
| "step": 61600 |
| }, |
| { |
| "epoch": 6.635453664836939, |
| "grad_norm": 0.7968859672546387, |
| "learning_rate": 0.00020225191250942785, |
| "loss": 3.2833, |
| "step": 61650 |
| }, |
| { |
| "epoch": 6.640835216876547, |
| "grad_norm": 0.7631063461303711, |
| "learning_rate": 0.00020192867147936642, |
| "loss": 3.2753, |
| "step": 61700 |
| }, |
| { |
| "epoch": 6.6462167689161555, |
| "grad_norm": 0.781497061252594, |
| "learning_rate": 0.00020160543044930502, |
| "loss": 3.2842, |
| "step": 61750 |
| }, |
| { |
| "epoch": 6.651598320955763, |
| "grad_norm": 0.7126927375793457, |
| "learning_rate": 0.0002012821894192436, |
| "loss": 3.2704, |
| "step": 61800 |
| }, |
| { |
| "epoch": 6.656979872995372, |
| "grad_norm": 0.7944461703300476, |
| "learning_rate": 0.00020095894838918218, |
| "loss": 3.2639, |
| "step": 61850 |
| }, |
| { |
| "epoch": 6.66236142503498, |
| "grad_norm": 0.7860047817230225, |
| "learning_rate": 0.00020063570735912077, |
| "loss": 3.296, |
| "step": 61900 |
| }, |
| { |
| "epoch": 6.667742977074588, |
| "grad_norm": 0.8037759065628052, |
| "learning_rate": 0.00020031246632905934, |
| "loss": 3.2693, |
| "step": 61950 |
| }, |
| { |
| "epoch": 6.6731245291141965, |
| "grad_norm": 0.7314351797103882, |
| "learning_rate": 0.00019998922529899794, |
| "loss": 3.2797, |
| "step": 62000 |
| }, |
| { |
| "epoch": 6.6731245291141965, |
| "eval_accuracy": 0.38599720696886586, |
| "eval_loss": 3.369133234024048, |
| "eval_runtime": 180.0106, |
| "eval_samples_per_second": 100.055, |
| "eval_steps_per_second": 6.255, |
| "step": 62000 |
| }, |
| { |
| "epoch": 6.678506081153805, |
| "grad_norm": 0.761888861656189, |
| "learning_rate": 0.00019966598426893653, |
| "loss": 3.2843, |
| "step": 62050 |
| }, |
| { |
| "epoch": 6.683887633193413, |
| "grad_norm": 0.715152382850647, |
| "learning_rate": 0.0001993427432388751, |
| "loss": 3.2712, |
| "step": 62100 |
| }, |
| { |
| "epoch": 6.689269185233021, |
| "grad_norm": 0.7696418166160583, |
| "learning_rate": 0.00019901950220881367, |
| "loss": 3.3034, |
| "step": 62150 |
| }, |
| { |
| "epoch": 6.69465073727263, |
| "grad_norm": 0.7413962483406067, |
| "learning_rate": 0.0001986962611787523, |
| "loss": 3.2882, |
| "step": 62200 |
| }, |
| { |
| "epoch": 6.7000322893122375, |
| "grad_norm": 0.7677690386772156, |
| "learning_rate": 0.00019837302014869085, |
| "loss": 3.2869, |
| "step": 62250 |
| }, |
| { |
| "epoch": 6.705413841351846, |
| "grad_norm": 0.7569989562034607, |
| "learning_rate": 0.00019804977911862942, |
| "loss": 3.2869, |
| "step": 62300 |
| }, |
| { |
| "epoch": 6.710795393391454, |
| "grad_norm": 0.7461286783218384, |
| "learning_rate": 0.00019772653808856804, |
| "loss": 3.2899, |
| "step": 62350 |
| }, |
| { |
| "epoch": 6.716176945431062, |
| "grad_norm": 0.7389951944351196, |
| "learning_rate": 0.0001974032970585066, |
| "loss": 3.2635, |
| "step": 62400 |
| }, |
| { |
| "epoch": 6.721558497470671, |
| "grad_norm": 0.7195466756820679, |
| "learning_rate": 0.00019708005602844518, |
| "loss": 3.2589, |
| "step": 62450 |
| }, |
| { |
| "epoch": 6.7269400495102785, |
| "grad_norm": 0.7346585988998413, |
| "learning_rate": 0.00019675681499838377, |
| "loss": 3.2894, |
| "step": 62500 |
| }, |
| { |
| "epoch": 6.732321601549887, |
| "grad_norm": 0.7372649312019348, |
| "learning_rate": 0.00019643357396832237, |
| "loss": 3.2776, |
| "step": 62550 |
| }, |
| { |
| "epoch": 6.737703153589496, |
| "grad_norm": 0.7496547102928162, |
| "learning_rate": 0.00019611033293826096, |
| "loss": 3.2765, |
| "step": 62600 |
| }, |
| { |
| "epoch": 6.743084705629103, |
| "grad_norm": 0.7103299498558044, |
| "learning_rate": 0.00019578709190819953, |
| "loss": 3.3054, |
| "step": 62650 |
| }, |
| { |
| "epoch": 6.748466257668712, |
| "grad_norm": 0.7128199934959412, |
| "learning_rate": 0.0001954638508781381, |
| "loss": 3.2887, |
| "step": 62700 |
| }, |
| { |
| "epoch": 6.75384780970832, |
| "grad_norm": 0.7775223255157471, |
| "learning_rate": 0.00019514060984807672, |
| "loss": 3.2817, |
| "step": 62750 |
| }, |
| { |
| "epoch": 6.759229361747928, |
| "grad_norm": 0.7413020133972168, |
| "learning_rate": 0.0001948173688180153, |
| "loss": 3.2883, |
| "step": 62800 |
| }, |
| { |
| "epoch": 6.7646109137875365, |
| "grad_norm": 0.7511486411094666, |
| "learning_rate": 0.00019449412778795386, |
| "loss": 3.2796, |
| "step": 62850 |
| }, |
| { |
| "epoch": 6.769992465827144, |
| "grad_norm": 0.7225663661956787, |
| "learning_rate": 0.00019417088675789245, |
| "loss": 3.2843, |
| "step": 62900 |
| }, |
| { |
| "epoch": 6.775374017866753, |
| "grad_norm": 0.7489200830459595, |
| "learning_rate": 0.00019384764572783104, |
| "loss": 3.2977, |
| "step": 62950 |
| }, |
| { |
| "epoch": 6.780755569906361, |
| "grad_norm": 0.7403072714805603, |
| "learning_rate": 0.0001935244046977696, |
| "loss": 3.2781, |
| "step": 63000 |
| }, |
| { |
| "epoch": 6.780755569906361, |
| "eval_accuracy": 0.3861868062581021, |
| "eval_loss": 3.361095666885376, |
| "eval_runtime": 179.7674, |
| "eval_samples_per_second": 100.191, |
| "eval_steps_per_second": 6.264, |
| "step": 63000 |
| }, |
| { |
| "epoch": 6.786137121945969, |
| "grad_norm": 0.77875155210495, |
| "learning_rate": 0.0001932011636677082, |
| "loss": 3.3193, |
| "step": 63050 |
| }, |
| { |
| "epoch": 6.7915186739855775, |
| "grad_norm": 0.8084599375724792, |
| "learning_rate": 0.00019287792263764677, |
| "loss": 3.294, |
| "step": 63100 |
| }, |
| { |
| "epoch": 6.796900226025185, |
| "grad_norm": 0.7540778517723083, |
| "learning_rate": 0.00019255468160758537, |
| "loss": 3.284, |
| "step": 63150 |
| }, |
| { |
| "epoch": 6.802281778064794, |
| "grad_norm": 0.7482621669769287, |
| "learning_rate": 0.00019223144057752396, |
| "loss": 3.2912, |
| "step": 63200 |
| }, |
| { |
| "epoch": 6.807663330104402, |
| "grad_norm": 0.7529076933860779, |
| "learning_rate": 0.00019190819954746253, |
| "loss": 3.2768, |
| "step": 63250 |
| }, |
| { |
| "epoch": 6.813044882144011, |
| "grad_norm": 0.7714017033576965, |
| "learning_rate": 0.0001915849585174011, |
| "loss": 3.2763, |
| "step": 63300 |
| }, |
| { |
| "epoch": 6.8184264341836185, |
| "grad_norm": 0.7732032537460327, |
| "learning_rate": 0.00019126171748733972, |
| "loss": 3.2988, |
| "step": 63350 |
| }, |
| { |
| "epoch": 6.823807986223227, |
| "grad_norm": 0.7607737183570862, |
| "learning_rate": 0.0001909384764572783, |
| "loss": 3.2844, |
| "step": 63400 |
| }, |
| { |
| "epoch": 6.829189538262835, |
| "grad_norm": 0.7674480080604553, |
| "learning_rate": 0.00019061523542721688, |
| "loss": 3.2795, |
| "step": 63450 |
| }, |
| { |
| "epoch": 6.834571090302443, |
| "grad_norm": 0.7431223392486572, |
| "learning_rate": 0.00019029199439715548, |
| "loss": 3.2974, |
| "step": 63500 |
| }, |
| { |
| "epoch": 6.839952642342052, |
| "grad_norm": 0.7607661485671997, |
| "learning_rate": 0.00018996875336709404, |
| "loss": 3.2843, |
| "step": 63550 |
| }, |
| { |
| "epoch": 6.8453341943816595, |
| "grad_norm": 0.7642389535903931, |
| "learning_rate": 0.00018964551233703264, |
| "loss": 3.3047, |
| "step": 63600 |
| }, |
| { |
| "epoch": 6.850715746421268, |
| "grad_norm": 0.760985791683197, |
| "learning_rate": 0.0001893222713069712, |
| "loss": 3.285, |
| "step": 63650 |
| }, |
| { |
| "epoch": 6.856097298460876, |
| "grad_norm": 0.7435808777809143, |
| "learning_rate": 0.0001889990302769098, |
| "loss": 3.2766, |
| "step": 63700 |
| }, |
| { |
| "epoch": 6.861478850500484, |
| "grad_norm": 0.7285375595092773, |
| "learning_rate": 0.0001886757892468484, |
| "loss": 3.3006, |
| "step": 63750 |
| }, |
| { |
| "epoch": 6.866860402540093, |
| "grad_norm": 0.7190316915512085, |
| "learning_rate": 0.00018835254821678696, |
| "loss": 3.2846, |
| "step": 63800 |
| }, |
| { |
| "epoch": 6.8722419545797, |
| "grad_norm": 0.7786919474601746, |
| "learning_rate": 0.00018802930718672553, |
| "loss": 3.2925, |
| "step": 63850 |
| }, |
| { |
| "epoch": 6.877623506619309, |
| "grad_norm": 0.7862507104873657, |
| "learning_rate": 0.00018770606615666415, |
| "loss": 3.2824, |
| "step": 63900 |
| }, |
| { |
| "epoch": 6.8830050586589175, |
| "grad_norm": 0.8249383568763733, |
| "learning_rate": 0.00018738282512660272, |
| "loss": 3.284, |
| "step": 63950 |
| }, |
| { |
| "epoch": 6.888386610698525, |
| "grad_norm": 0.7981451749801636, |
| "learning_rate": 0.0001870595840965413, |
| "loss": 3.2945, |
| "step": 64000 |
| }, |
| { |
| "epoch": 6.888386610698525, |
| "eval_accuracy": 0.38657045960497505, |
| "eval_loss": 3.35958194732666, |
| "eval_runtime": 179.9583, |
| "eval_samples_per_second": 100.084, |
| "eval_steps_per_second": 6.257, |
| "step": 64000 |
| }, |
| { |
| "epoch": 6.893768162738134, |
| "grad_norm": 0.7628564834594727, |
| "learning_rate": 0.0001867363430664799, |
| "loss": 3.292, |
| "step": 64050 |
| }, |
| { |
| "epoch": 6.899149714777742, |
| "grad_norm": 0.8125451803207397, |
| "learning_rate": 0.00018641310203641848, |
| "loss": 3.2681, |
| "step": 64100 |
| }, |
| { |
| "epoch": 6.90453126681735, |
| "grad_norm": 0.7533065676689148, |
| "learning_rate": 0.00018608986100635705, |
| "loss": 3.2843, |
| "step": 64150 |
| }, |
| { |
| "epoch": 6.9099128188569585, |
| "grad_norm": 0.8026787638664246, |
| "learning_rate": 0.00018577308479689685, |
| "loss": 3.3001, |
| "step": 64200 |
| }, |
| { |
| "epoch": 6.915294370896566, |
| "grad_norm": 0.7561555504798889, |
| "learning_rate": 0.00018544984376683545, |
| "loss": 3.2767, |
| "step": 64250 |
| }, |
| { |
| "epoch": 6.920675922936175, |
| "grad_norm": 0.764043927192688, |
| "learning_rate": 0.00018512660273677404, |
| "loss": 3.2937, |
| "step": 64300 |
| }, |
| { |
| "epoch": 6.926057474975783, |
| "grad_norm": 0.7226858139038086, |
| "learning_rate": 0.0001848033617067126, |
| "loss": 3.2866, |
| "step": 64350 |
| }, |
| { |
| "epoch": 6.931439027015391, |
| "grad_norm": 0.7623168230056763, |
| "learning_rate": 0.00018448012067665123, |
| "loss": 3.2637, |
| "step": 64400 |
| }, |
| { |
| "epoch": 6.9368205790549995, |
| "grad_norm": 0.7732171416282654, |
| "learning_rate": 0.0001841568796465898, |
| "loss": 3.2752, |
| "step": 64450 |
| }, |
| { |
| "epoch": 6.942202131094608, |
| "grad_norm": 0.7387587428092957, |
| "learning_rate": 0.00018383363861652837, |
| "loss": 3.2894, |
| "step": 64500 |
| }, |
| { |
| "epoch": 6.947583683134216, |
| "grad_norm": 0.7186091542243958, |
| "learning_rate": 0.00018351039758646696, |
| "loss": 3.2944, |
| "step": 64550 |
| }, |
| { |
| "epoch": 6.952965235173824, |
| "grad_norm": 0.7569155097007751, |
| "learning_rate": 0.00018318715655640555, |
| "loss": 3.2855, |
| "step": 64600 |
| }, |
| { |
| "epoch": 6.958346787213433, |
| "grad_norm": 0.7623655200004578, |
| "learning_rate": 0.00018286391552634412, |
| "loss": 3.281, |
| "step": 64650 |
| }, |
| { |
| "epoch": 6.9637283392530405, |
| "grad_norm": 0.8205636143684387, |
| "learning_rate": 0.00018254067449628272, |
| "loss": 3.281, |
| "step": 64700 |
| }, |
| { |
| "epoch": 6.969109891292649, |
| "grad_norm": 0.7782757878303528, |
| "learning_rate": 0.00018221743346622128, |
| "loss": 3.2977, |
| "step": 64750 |
| }, |
| { |
| "epoch": 6.974491443332257, |
| "grad_norm": 0.7646986246109009, |
| "learning_rate": 0.00018189419243615988, |
| "loss": 3.2798, |
| "step": 64800 |
| }, |
| { |
| "epoch": 6.979872995371865, |
| "grad_norm": 0.7757244110107422, |
| "learning_rate": 0.00018157095140609847, |
| "loss": 3.287, |
| "step": 64850 |
| }, |
| { |
| "epoch": 6.985254547411474, |
| "grad_norm": 0.7472320199012756, |
| "learning_rate": 0.00018125417519663825, |
| "loss": 3.2609, |
| "step": 64900 |
| }, |
| { |
| "epoch": 6.990636099451081, |
| "grad_norm": 0.7466889023780823, |
| "learning_rate": 0.00018093093416657687, |
| "loss": 3.2883, |
| "step": 64950 |
| }, |
| { |
| "epoch": 6.99601765149069, |
| "grad_norm": 0.7409732937812805, |
| "learning_rate": 0.00018060769313651544, |
| "loss": 3.284, |
| "step": 65000 |
| }, |
| { |
| "epoch": 6.99601765149069, |
| "eval_accuracy": 0.3868794684179194, |
| "eval_loss": 3.35493803024292, |
| "eval_runtime": 179.8236, |
| "eval_samples_per_second": 100.159, |
| "eval_steps_per_second": 6.262, |
| "step": 65000 |
| }, |
| { |
| "epoch": 7.0013992035302985, |
| "grad_norm": 0.7562103867530823, |
| "learning_rate": 0.000180284452106454, |
| "loss": 3.2676, |
| "step": 65050 |
| }, |
| { |
| "epoch": 7.006780755569906, |
| "grad_norm": 0.7899980545043945, |
| "learning_rate": 0.00017996121107639263, |
| "loss": 3.223, |
| "step": 65100 |
| }, |
| { |
| "epoch": 7.012162307609515, |
| "grad_norm": 0.7686735391616821, |
| "learning_rate": 0.0001796379700463312, |
| "loss": 3.2053, |
| "step": 65150 |
| }, |
| { |
| "epoch": 7.017543859649122, |
| "grad_norm": 0.7950847744941711, |
| "learning_rate": 0.0001793147290162698, |
| "loss": 3.1919, |
| "step": 65200 |
| }, |
| { |
| "epoch": 7.022925411688731, |
| "grad_norm": 0.8973643183708191, |
| "learning_rate": 0.00017899148798620836, |
| "loss": 3.1952, |
| "step": 65250 |
| }, |
| { |
| "epoch": 7.0283069637283395, |
| "grad_norm": 0.7614500522613525, |
| "learning_rate": 0.00017866824695614696, |
| "loss": 3.1957, |
| "step": 65300 |
| }, |
| { |
| "epoch": 7.033688515767947, |
| "grad_norm": 0.8014223575592041, |
| "learning_rate": 0.00017834500592608555, |
| "loss": 3.1986, |
| "step": 65350 |
| }, |
| { |
| "epoch": 7.039070067807556, |
| "grad_norm": 0.7725815773010254, |
| "learning_rate": 0.00017802176489602412, |
| "loss": 3.1856, |
| "step": 65400 |
| }, |
| { |
| "epoch": 7.044451619847164, |
| "grad_norm": 0.7577409744262695, |
| "learning_rate": 0.00017769852386596269, |
| "loss": 3.2132, |
| "step": 65450 |
| }, |
| { |
| "epoch": 7.049833171886772, |
| "grad_norm": 0.7576920986175537, |
| "learning_rate": 0.0001773752828359013, |
| "loss": 3.1937, |
| "step": 65500 |
| }, |
| { |
| "epoch": 7.0552147239263805, |
| "grad_norm": 0.7937734723091125, |
| "learning_rate": 0.00017705204180583988, |
| "loss": 3.2073, |
| "step": 65550 |
| }, |
| { |
| "epoch": 7.060596275965988, |
| "grad_norm": 0.7605912685394287, |
| "learning_rate": 0.00017672880077577844, |
| "loss": 3.2075, |
| "step": 65600 |
| }, |
| { |
| "epoch": 7.065977828005597, |
| "grad_norm": 0.7895816564559937, |
| "learning_rate": 0.00017640555974571704, |
| "loss": 3.2088, |
| "step": 65650 |
| }, |
| { |
| "epoch": 7.071359380045205, |
| "grad_norm": 0.8300418853759766, |
| "learning_rate": 0.00017608231871565563, |
| "loss": 3.206, |
| "step": 65700 |
| }, |
| { |
| "epoch": 7.076740932084813, |
| "grad_norm": 0.7783611416816711, |
| "learning_rate": 0.0001757590776855942, |
| "loss": 3.2163, |
| "step": 65750 |
| }, |
| { |
| "epoch": 7.0821224841244215, |
| "grad_norm": 0.7723811864852905, |
| "learning_rate": 0.0001754358366555328, |
| "loss": 3.215, |
| "step": 65800 |
| }, |
| { |
| "epoch": 7.08750403616403, |
| "grad_norm": 0.7977986931800842, |
| "learning_rate": 0.00017511259562547136, |
| "loss": 3.2125, |
| "step": 65850 |
| }, |
| { |
| "epoch": 7.092885588203638, |
| "grad_norm": 0.7651707530021667, |
| "learning_rate": 0.00017478935459540996, |
| "loss": 3.2135, |
| "step": 65900 |
| }, |
| { |
| "epoch": 7.098267140243246, |
| "grad_norm": 0.812610924243927, |
| "learning_rate": 0.00017446611356534855, |
| "loss": 3.2242, |
| "step": 65950 |
| }, |
| { |
| "epoch": 7.103648692282855, |
| "grad_norm": 0.7896516919136047, |
| "learning_rate": 0.00017414287253528712, |
| "loss": 3.1994, |
| "step": 66000 |
| }, |
| { |
| "epoch": 7.103648692282855, |
| "eval_accuracy": 0.3869793204218209, |
| "eval_loss": 3.361632823944092, |
| "eval_runtime": 179.86, |
| "eval_samples_per_second": 100.139, |
| "eval_steps_per_second": 6.26, |
| "step": 66000 |
| }, |
| { |
| "epoch": 7.109030244322462, |
| "grad_norm": 0.7712588310241699, |
| "learning_rate": 0.00017381963150522569, |
| "loss": 3.2167, |
| "step": 66050 |
| }, |
| { |
| "epoch": 7.114411796362071, |
| "grad_norm": 0.7564050555229187, |
| "learning_rate": 0.0001734963904751643, |
| "loss": 3.2134, |
| "step": 66100 |
| }, |
| { |
| "epoch": 7.119793348401679, |
| "grad_norm": 0.7921330332756042, |
| "learning_rate": 0.00017317314944510288, |
| "loss": 3.2075, |
| "step": 66150 |
| }, |
| { |
| "epoch": 7.125174900441287, |
| "grad_norm": 0.8234257698059082, |
| "learning_rate": 0.00017284990841504147, |
| "loss": 3.2197, |
| "step": 66200 |
| }, |
| { |
| "epoch": 7.130556452480896, |
| "grad_norm": 0.7843002080917358, |
| "learning_rate": 0.00017252666738498007, |
| "loss": 3.2197, |
| "step": 66250 |
| }, |
| { |
| "epoch": 7.135938004520503, |
| "grad_norm": 0.7515401244163513, |
| "learning_rate": 0.00017220342635491863, |
| "loss": 3.209, |
| "step": 66300 |
| }, |
| { |
| "epoch": 7.141319556560112, |
| "grad_norm": 0.841936469078064, |
| "learning_rate": 0.00017188018532485723, |
| "loss": 3.2314, |
| "step": 66350 |
| }, |
| { |
| "epoch": 7.1467011085997205, |
| "grad_norm": 0.7923972010612488, |
| "learning_rate": 0.0001715569442947958, |
| "loss": 3.2195, |
| "step": 66400 |
| }, |
| { |
| "epoch": 7.152082660639328, |
| "grad_norm": 0.7993304133415222, |
| "learning_rate": 0.0001712337032647344, |
| "loss": 3.2214, |
| "step": 66450 |
| }, |
| { |
| "epoch": 7.157464212678937, |
| "grad_norm": 0.7993540167808533, |
| "learning_rate": 0.00017091046223467298, |
| "loss": 3.2377, |
| "step": 66500 |
| }, |
| { |
| "epoch": 7.162845764718545, |
| "grad_norm": 0.8128913640975952, |
| "learning_rate": 0.00017058722120461155, |
| "loss": 3.2127, |
| "step": 66550 |
| }, |
| { |
| "epoch": 7.168227316758153, |
| "grad_norm": 0.7577317357063293, |
| "learning_rate": 0.00017026398017455012, |
| "loss": 3.2401, |
| "step": 66600 |
| }, |
| { |
| "epoch": 7.1736088687977615, |
| "grad_norm": 0.8127763271331787, |
| "learning_rate": 0.00016994073914448874, |
| "loss": 3.2001, |
| "step": 66650 |
| }, |
| { |
| "epoch": 7.178990420837369, |
| "grad_norm": 0.7880443930625916, |
| "learning_rate": 0.0001696174981144273, |
| "loss": 3.2313, |
| "step": 66700 |
| }, |
| { |
| "epoch": 7.184371972876978, |
| "grad_norm": 0.7774691581726074, |
| "learning_rate": 0.00016929425708436588, |
| "loss": 3.2234, |
| "step": 66750 |
| }, |
| { |
| "epoch": 7.189753524916586, |
| "grad_norm": 0.8356835246086121, |
| "learning_rate": 0.0001689710160543045, |
| "loss": 3.203, |
| "step": 66800 |
| }, |
| { |
| "epoch": 7.195135076956194, |
| "grad_norm": 0.8748675584793091, |
| "learning_rate": 0.00016864777502424307, |
| "loss": 3.2245, |
| "step": 66850 |
| }, |
| { |
| "epoch": 7.2005166289958025, |
| "grad_norm": 0.7685278654098511, |
| "learning_rate": 0.00016832453399418163, |
| "loss": 3.225, |
| "step": 66900 |
| }, |
| { |
| "epoch": 7.205898181035411, |
| "grad_norm": 0.7593712210655212, |
| "learning_rate": 0.00016800129296412023, |
| "loss": 3.2192, |
| "step": 66950 |
| }, |
| { |
| "epoch": 7.211279733075019, |
| "grad_norm": 0.7982090711593628, |
| "learning_rate": 0.00016767805193405882, |
| "loss": 3.2195, |
| "step": 67000 |
| }, |
| { |
| "epoch": 7.211279733075019, |
| "eval_accuracy": 0.3874346846746514, |
| "eval_loss": 3.359243631362915, |
| "eval_runtime": 179.5957, |
| "eval_samples_per_second": 100.286, |
| "eval_steps_per_second": 6.27, |
| "step": 67000 |
| }, |
| { |
| "epoch": 7.216661285114627, |
| "grad_norm": 0.7449146509170532, |
| "learning_rate": 0.0001673548109039974, |
| "loss": 3.2246, |
| "step": 67050 |
| }, |
| { |
| "epoch": 7.222042837154235, |
| "grad_norm": 0.7682053446769714, |
| "learning_rate": 0.00016703156987393598, |
| "loss": 3.2388, |
| "step": 67100 |
| }, |
| { |
| "epoch": 7.2274243891938434, |
| "grad_norm": 0.7306789755821228, |
| "learning_rate": 0.00016670832884387455, |
| "loss": 3.2375, |
| "step": 67150 |
| }, |
| { |
| "epoch": 7.232805941233452, |
| "grad_norm": 0.7582760453224182, |
| "learning_rate": 0.00016638508781381317, |
| "loss": 3.2174, |
| "step": 67200 |
| }, |
| { |
| "epoch": 7.23818749327306, |
| "grad_norm": 0.7850392460823059, |
| "learning_rate": 0.00016606184678375174, |
| "loss": 3.2253, |
| "step": 67250 |
| }, |
| { |
| "epoch": 7.243569045312668, |
| "grad_norm": 0.7982510924339294, |
| "learning_rate": 0.0001657386057536903, |
| "loss": 3.2302, |
| "step": 67300 |
| }, |
| { |
| "epoch": 7.248950597352277, |
| "grad_norm": 0.8140314817428589, |
| "learning_rate": 0.00016541536472362893, |
| "loss": 3.2346, |
| "step": 67350 |
| }, |
| { |
| "epoch": 7.254332149391884, |
| "grad_norm": 0.7745968699455261, |
| "learning_rate": 0.0001650921236935675, |
| "loss": 3.2229, |
| "step": 67400 |
| }, |
| { |
| "epoch": 7.259713701431493, |
| "grad_norm": 0.7968515753746033, |
| "learning_rate": 0.00016476888266350607, |
| "loss": 3.2248, |
| "step": 67450 |
| }, |
| { |
| "epoch": 7.265095253471101, |
| "grad_norm": 0.7656205892562866, |
| "learning_rate": 0.00016444564163344466, |
| "loss": 3.2252, |
| "step": 67500 |
| }, |
| { |
| "epoch": 7.270476805510709, |
| "grad_norm": 0.8577672243118286, |
| "learning_rate": 0.00016412240060338326, |
| "loss": 3.2215, |
| "step": 67550 |
| }, |
| { |
| "epoch": 7.275858357550318, |
| "grad_norm": 0.795576274394989, |
| "learning_rate": 0.00016379915957332182, |
| "loss": 3.2365, |
| "step": 67600 |
| }, |
| { |
| "epoch": 7.281239909589925, |
| "grad_norm": 0.7914314866065979, |
| "learning_rate": 0.00016347591854326042, |
| "loss": 3.2276, |
| "step": 67650 |
| }, |
| { |
| "epoch": 7.286621461629534, |
| "grad_norm": 0.7388089299201965, |
| "learning_rate": 0.00016315267751319898, |
| "loss": 3.2328, |
| "step": 67700 |
| }, |
| { |
| "epoch": 7.2920030136691425, |
| "grad_norm": 0.8092725276947021, |
| "learning_rate": 0.00016282943648313755, |
| "loss": 3.221, |
| "step": 67750 |
| }, |
| { |
| "epoch": 7.29738456570875, |
| "grad_norm": 0.7859850525856018, |
| "learning_rate": 0.00016250619545307617, |
| "loss": 3.2308, |
| "step": 67800 |
| }, |
| { |
| "epoch": 7.302766117748359, |
| "grad_norm": 0.7894166111946106, |
| "learning_rate": 0.00016218295442301474, |
| "loss": 3.2362, |
| "step": 67850 |
| }, |
| { |
| "epoch": 7.308147669787967, |
| "grad_norm": 0.8145300149917603, |
| "learning_rate": 0.0001618597133929533, |
| "loss": 3.2318, |
| "step": 67900 |
| }, |
| { |
| "epoch": 7.313529221827575, |
| "grad_norm": 0.8098170757293701, |
| "learning_rate": 0.00016153647236289193, |
| "loss": 3.2284, |
| "step": 67950 |
| }, |
| { |
| "epoch": 7.3189107738671835, |
| "grad_norm": 0.8146171569824219, |
| "learning_rate": 0.0001612132313328305, |
| "loss": 3.2313, |
| "step": 68000 |
| }, |
| { |
| "epoch": 7.3189107738671835, |
| "eval_accuracy": 0.3875706094373073, |
| "eval_loss": 3.3579890727996826, |
| "eval_runtime": 179.851, |
| "eval_samples_per_second": 100.144, |
| "eval_steps_per_second": 6.261, |
| "step": 68000 |
| }, |
| { |
| "epoch": 7.324292325906791, |
| "grad_norm": 0.8207062482833862, |
| "learning_rate": 0.0001608899903027691, |
| "loss": 3.238, |
| "step": 68050 |
| }, |
| { |
| "epoch": 7.3296738779464, |
| "grad_norm": 1.0600287914276123, |
| "learning_rate": 0.00016056674927270766, |
| "loss": 3.2332, |
| "step": 68100 |
| }, |
| { |
| "epoch": 7.335055429986008, |
| "grad_norm": 0.7621212601661682, |
| "learning_rate": 0.00016024350824264626, |
| "loss": 3.2204, |
| "step": 68150 |
| }, |
| { |
| "epoch": 7.340436982025616, |
| "grad_norm": 0.804821789264679, |
| "learning_rate": 0.00015992026721258485, |
| "loss": 3.2355, |
| "step": 68200 |
| }, |
| { |
| "epoch": 7.3458185340652244, |
| "grad_norm": 0.7746620774269104, |
| "learning_rate": 0.00015959702618252342, |
| "loss": 3.2425, |
| "step": 68250 |
| }, |
| { |
| "epoch": 7.351200086104833, |
| "grad_norm": 0.7920762896537781, |
| "learning_rate": 0.00015927378515246199, |
| "loss": 3.2154, |
| "step": 68300 |
| }, |
| { |
| "epoch": 7.356581638144441, |
| "grad_norm": 0.8123354315757751, |
| "learning_rate": 0.0001589505441224006, |
| "loss": 3.2222, |
| "step": 68350 |
| }, |
| { |
| "epoch": 7.361963190184049, |
| "grad_norm": 0.7643717527389526, |
| "learning_rate": 0.00015862730309233917, |
| "loss": 3.2192, |
| "step": 68400 |
| }, |
| { |
| "epoch": 7.367344742223658, |
| "grad_norm": 0.7962055802345276, |
| "learning_rate": 0.00015830406206227774, |
| "loss": 3.2198, |
| "step": 68450 |
| }, |
| { |
| "epoch": 7.372726294263265, |
| "grad_norm": 0.7841677069664001, |
| "learning_rate": 0.00015798082103221636, |
| "loss": 3.2349, |
| "step": 68500 |
| }, |
| { |
| "epoch": 7.378107846302874, |
| "grad_norm": 0.8586899042129517, |
| "learning_rate": 0.00015765758000215493, |
| "loss": 3.2325, |
| "step": 68550 |
| }, |
| { |
| "epoch": 7.383489398342482, |
| "grad_norm": 0.7646048665046692, |
| "learning_rate": 0.0001573343389720935, |
| "loss": 3.2206, |
| "step": 68600 |
| }, |
| { |
| "epoch": 7.38887095038209, |
| "grad_norm": 0.7651141285896301, |
| "learning_rate": 0.0001570110979420321, |
| "loss": 3.2242, |
| "step": 68650 |
| }, |
| { |
| "epoch": 7.394252502421699, |
| "grad_norm": 0.7675405144691467, |
| "learning_rate": 0.0001566943217325719, |
| "loss": 3.221, |
| "step": 68700 |
| }, |
| { |
| "epoch": 7.399634054461306, |
| "grad_norm": 0.7907142639160156, |
| "learning_rate": 0.0001563710807025105, |
| "loss": 3.2205, |
| "step": 68750 |
| }, |
| { |
| "epoch": 7.405015606500915, |
| "grad_norm": 0.7915471792221069, |
| "learning_rate": 0.00015604783967244906, |
| "loss": 3.2347, |
| "step": 68800 |
| }, |
| { |
| "epoch": 7.4103971585405235, |
| "grad_norm": 0.8003562092781067, |
| "learning_rate": 0.00015572459864238763, |
| "loss": 3.2305, |
| "step": 68850 |
| }, |
| { |
| "epoch": 7.415778710580131, |
| "grad_norm": 0.7971240282058716, |
| "learning_rate": 0.00015540135761232625, |
| "loss": 3.2337, |
| "step": 68900 |
| }, |
| { |
| "epoch": 7.42116026261974, |
| "grad_norm": 0.7408497333526611, |
| "learning_rate": 0.00015507811658226482, |
| "loss": 3.2173, |
| "step": 68950 |
| }, |
| { |
| "epoch": 7.426541814659347, |
| "grad_norm": 0.803286075592041, |
| "learning_rate": 0.00015475487555220344, |
| "loss": 3.2322, |
| "step": 69000 |
| }, |
| { |
| "epoch": 7.426541814659347, |
| "eval_accuracy": 0.38788418167153993, |
| "eval_loss": 3.3515703678131104, |
| "eval_runtime": 179.8641, |
| "eval_samples_per_second": 100.137, |
| "eval_steps_per_second": 6.26, |
| "step": 69000 |
| }, |
| { |
| "epoch": 7.431923366698956, |
| "grad_norm": 0.7986598014831543, |
| "learning_rate": 0.000154431634522142, |
| "loss": 3.2336, |
| "step": 69050 |
| }, |
| { |
| "epoch": 7.4373049187385645, |
| "grad_norm": 0.7994623184204102, |
| "learning_rate": 0.00015410839349208058, |
| "loss": 3.2172, |
| "step": 69100 |
| }, |
| { |
| "epoch": 7.442686470778172, |
| "grad_norm": 0.802762508392334, |
| "learning_rate": 0.00015378515246201917, |
| "loss": 3.2326, |
| "step": 69150 |
| }, |
| { |
| "epoch": 7.448068022817781, |
| "grad_norm": 0.768039882183075, |
| "learning_rate": 0.00015346191143195774, |
| "loss": 3.2339, |
| "step": 69200 |
| }, |
| { |
| "epoch": 7.453449574857389, |
| "grad_norm": 0.818593442440033, |
| "learning_rate": 0.00015313867040189633, |
| "loss": 3.2376, |
| "step": 69250 |
| }, |
| { |
| "epoch": 7.458831126896997, |
| "grad_norm": 0.8104583621025085, |
| "learning_rate": 0.00015281542937183493, |
| "loss": 3.2267, |
| "step": 69300 |
| }, |
| { |
| "epoch": 7.4642126789366054, |
| "grad_norm": 0.806210458278656, |
| "learning_rate": 0.0001524921883417735, |
| "loss": 3.2389, |
| "step": 69350 |
| }, |
| { |
| "epoch": 7.469594230976213, |
| "grad_norm": 0.7950648665428162, |
| "learning_rate": 0.00015216894731171206, |
| "loss": 3.2193, |
| "step": 69400 |
| }, |
| { |
| "epoch": 7.474975783015822, |
| "grad_norm": 0.8381013870239258, |
| "learning_rate": 0.00015184570628165068, |
| "loss": 3.2313, |
| "step": 69450 |
| }, |
| { |
| "epoch": 7.48035733505543, |
| "grad_norm": 0.7694802284240723, |
| "learning_rate": 0.00015152246525158925, |
| "loss": 3.2385, |
| "step": 69500 |
| }, |
| { |
| "epoch": 7.485738887095038, |
| "grad_norm": 0.8459978699684143, |
| "learning_rate": 0.00015119922422152782, |
| "loss": 3.2132, |
| "step": 69550 |
| }, |
| { |
| "epoch": 7.491120439134646, |
| "grad_norm": 0.8075374960899353, |
| "learning_rate": 0.00015087598319146644, |
| "loss": 3.2399, |
| "step": 69600 |
| }, |
| { |
| "epoch": 7.496501991174255, |
| "grad_norm": 0.7973178029060364, |
| "learning_rate": 0.000150552742161405, |
| "loss": 3.2294, |
| "step": 69650 |
| }, |
| { |
| "epoch": 7.501883543213863, |
| "grad_norm": 0.7813896536827087, |
| "learning_rate": 0.00015022950113134358, |
| "loss": 3.228, |
| "step": 69700 |
| }, |
| { |
| "epoch": 7.507265095253471, |
| "grad_norm": 0.8545538187026978, |
| "learning_rate": 0.00014990626010128217, |
| "loss": 3.2443, |
| "step": 69750 |
| }, |
| { |
| "epoch": 7.51264664729308, |
| "grad_norm": 0.7499383091926575, |
| "learning_rate": 0.00014958301907122077, |
| "loss": 3.2401, |
| "step": 69800 |
| }, |
| { |
| "epoch": 7.518028199332687, |
| "grad_norm": 0.8156957030296326, |
| "learning_rate": 0.00014925977804115933, |
| "loss": 3.2466, |
| "step": 69850 |
| }, |
| { |
| "epoch": 7.523409751372296, |
| "grad_norm": 0.8415682315826416, |
| "learning_rate": 0.00014893653701109793, |
| "loss": 3.2477, |
| "step": 69900 |
| }, |
| { |
| "epoch": 7.528791303411904, |
| "grad_norm": 0.7899500131607056, |
| "learning_rate": 0.0001486132959810365, |
| "loss": 3.2335, |
| "step": 69950 |
| }, |
| { |
| "epoch": 7.534172855451512, |
| "grad_norm": 0.819366991519928, |
| "learning_rate": 0.0001482900549509751, |
| "loss": 3.2278, |
| "step": 70000 |
| }, |
| { |
| "epoch": 7.534172855451512, |
| "eval_accuracy": 0.38836833893964123, |
| "eval_loss": 3.3499035835266113, |
| "eval_runtime": 180.0283, |
| "eval_samples_per_second": 100.045, |
| "eval_steps_per_second": 6.255, |
| "step": 70000 |
| }, |
| { |
| "epoch": 7.539554407491121, |
| "grad_norm": 0.8849138021469116, |
| "learning_rate": 0.00014796681392091368, |
| "loss": 3.2443, |
| "step": 70050 |
| }, |
| { |
| "epoch": 7.544935959530728, |
| "grad_norm": 0.8181607723236084, |
| "learning_rate": 0.00014764357289085228, |
| "loss": 3.2384, |
| "step": 70100 |
| }, |
| { |
| "epoch": 7.550317511570337, |
| "grad_norm": 0.9902865290641785, |
| "learning_rate": 0.00014732033186079085, |
| "loss": 3.2333, |
| "step": 70150 |
| }, |
| { |
| "epoch": 7.5556990636099455, |
| "grad_norm": 0.7719454169273376, |
| "learning_rate": 0.00014699709083072944, |
| "loss": 3.2358, |
| "step": 70200 |
| }, |
| { |
| "epoch": 7.561080615649553, |
| "grad_norm": 0.830987274646759, |
| "learning_rate": 0.00014667384980066804, |
| "loss": 3.2325, |
| "step": 70250 |
| }, |
| { |
| "epoch": 7.566462167689162, |
| "grad_norm": 0.8358805775642395, |
| "learning_rate": 0.0001463506087706066, |
| "loss": 3.2332, |
| "step": 70300 |
| }, |
| { |
| "epoch": 7.57184371972877, |
| "grad_norm": 0.7925933003425598, |
| "learning_rate": 0.0001460273677405452, |
| "loss": 3.2502, |
| "step": 70350 |
| }, |
| { |
| "epoch": 7.577225271768378, |
| "grad_norm": 0.8156459927558899, |
| "learning_rate": 0.00014570412671048377, |
| "loss": 3.2425, |
| "step": 70400 |
| }, |
| { |
| "epoch": 7.5826068238079865, |
| "grad_norm": 0.8065808415412903, |
| "learning_rate": 0.00014538088568042236, |
| "loss": 3.2272, |
| "step": 70450 |
| }, |
| { |
| "epoch": 7.587988375847594, |
| "grad_norm": 0.8279649019241333, |
| "learning_rate": 0.00014505764465036093, |
| "loss": 3.232, |
| "step": 70500 |
| }, |
| { |
| "epoch": 7.593369927887203, |
| "grad_norm": 0.8306601047515869, |
| "learning_rate": 0.00014473440362029952, |
| "loss": 3.2414, |
| "step": 70550 |
| }, |
| { |
| "epoch": 7.598751479926811, |
| "grad_norm": 0.8502399325370789, |
| "learning_rate": 0.00014441116259023812, |
| "loss": 3.2204, |
| "step": 70600 |
| }, |
| { |
| "epoch": 7.604133031966419, |
| "grad_norm": 0.8230718970298767, |
| "learning_rate": 0.00014408792156017669, |
| "loss": 3.2361, |
| "step": 70650 |
| }, |
| { |
| "epoch": 7.609514584006027, |
| "grad_norm": 0.8219791650772095, |
| "learning_rate": 0.00014376468053011528, |
| "loss": 3.2281, |
| "step": 70700 |
| }, |
| { |
| "epoch": 7.614896136045635, |
| "grad_norm": 0.8464500904083252, |
| "learning_rate": 0.00014344143950005387, |
| "loss": 3.2384, |
| "step": 70750 |
| }, |
| { |
| "epoch": 7.620277688085244, |
| "grad_norm": 0.8252779245376587, |
| "learning_rate": 0.00014311819846999244, |
| "loss": 3.2316, |
| "step": 70800 |
| }, |
| { |
| "epoch": 7.625659240124852, |
| "grad_norm": 0.808069109916687, |
| "learning_rate": 0.00014279495743993104, |
| "loss": 3.2265, |
| "step": 70850 |
| }, |
| { |
| "epoch": 7.63104079216446, |
| "grad_norm": 0.8365920782089233, |
| "learning_rate": 0.00014247171640986963, |
| "loss": 3.2415, |
| "step": 70900 |
| }, |
| { |
| "epoch": 7.636422344204068, |
| "grad_norm": 0.7884583473205566, |
| "learning_rate": 0.0001421484753798082, |
| "loss": 3.2226, |
| "step": 70950 |
| }, |
| { |
| "epoch": 7.641803896243677, |
| "grad_norm": 0.8387513160705566, |
| "learning_rate": 0.0001418252343497468, |
| "loss": 3.253, |
| "step": 71000 |
| }, |
| { |
| "epoch": 7.641803896243677, |
| "eval_accuracy": 0.3887576422366805, |
| "eval_loss": 3.345111608505249, |
| "eval_runtime": 179.568, |
| "eval_samples_per_second": 100.302, |
| "eval_steps_per_second": 6.271, |
| "step": 71000 |
| }, |
| { |
| "epoch": 7.647185448283285, |
| "grad_norm": 0.7808358669281006, |
| "learning_rate": 0.00014150199331968536, |
| "loss": 3.2328, |
| "step": 71050 |
| }, |
| { |
| "epoch": 7.652567000322893, |
| "grad_norm": 0.7849861979484558, |
| "learning_rate": 0.00014117875228962396, |
| "loss": 3.2317, |
| "step": 71100 |
| }, |
| { |
| "epoch": 7.657948552362502, |
| "grad_norm": 0.8119670152664185, |
| "learning_rate": 0.00014085551125956252, |
| "loss": 3.233, |
| "step": 71150 |
| }, |
| { |
| "epoch": 7.663330104402109, |
| "grad_norm": 0.8060752749443054, |
| "learning_rate": 0.00014053227022950112, |
| "loss": 3.2309, |
| "step": 71200 |
| }, |
| { |
| "epoch": 7.668711656441718, |
| "grad_norm": 0.8241363167762756, |
| "learning_rate": 0.0001402090291994397, |
| "loss": 3.238, |
| "step": 71250 |
| }, |
| { |
| "epoch": 7.674093208481326, |
| "grad_norm": 0.8252966403961182, |
| "learning_rate": 0.00013988578816937828, |
| "loss": 3.2444, |
| "step": 71300 |
| }, |
| { |
| "epoch": 7.679474760520934, |
| "grad_norm": 0.8020745515823364, |
| "learning_rate": 0.00013956254713931687, |
| "loss": 3.2285, |
| "step": 71350 |
| }, |
| { |
| "epoch": 7.684856312560543, |
| "grad_norm": 0.8435476422309875, |
| "learning_rate": 0.00013923930610925547, |
| "loss": 3.2405, |
| "step": 71400 |
| }, |
| { |
| "epoch": 7.69023786460015, |
| "grad_norm": 0.7923038005828857, |
| "learning_rate": 0.00013891606507919404, |
| "loss": 3.2201, |
| "step": 71450 |
| }, |
| { |
| "epoch": 7.695619416639759, |
| "grad_norm": 0.8090280890464783, |
| "learning_rate": 0.00013859282404913263, |
| "loss": 3.2247, |
| "step": 71500 |
| }, |
| { |
| "epoch": 7.7010009686793675, |
| "grad_norm": 0.8997794985771179, |
| "learning_rate": 0.0001382695830190712, |
| "loss": 3.2194, |
| "step": 71550 |
| }, |
| { |
| "epoch": 7.706382520718975, |
| "grad_norm": 0.8104215264320374, |
| "learning_rate": 0.0001379463419890098, |
| "loss": 3.2425, |
| "step": 71600 |
| }, |
| { |
| "epoch": 7.711764072758584, |
| "grad_norm": 0.8430519104003906, |
| "learning_rate": 0.00013762310095894836, |
| "loss": 3.242, |
| "step": 71650 |
| }, |
| { |
| "epoch": 7.717145624798192, |
| "grad_norm": 0.8335681557655334, |
| "learning_rate": 0.00013729985992888696, |
| "loss": 3.237, |
| "step": 71700 |
| }, |
| { |
| "epoch": 7.7225271768378, |
| "grad_norm": 0.8195766806602478, |
| "learning_rate": 0.00013697661889882555, |
| "loss": 3.2369, |
| "step": 71750 |
| }, |
| { |
| "epoch": 7.727908728877408, |
| "grad_norm": 0.8362144231796265, |
| "learning_rate": 0.00013665337786876412, |
| "loss": 3.226, |
| "step": 71800 |
| }, |
| { |
| "epoch": 7.733290280917016, |
| "grad_norm": 0.8281188011169434, |
| "learning_rate": 0.0001363301368387027, |
| "loss": 3.2287, |
| "step": 71850 |
| }, |
| { |
| "epoch": 7.738671832956625, |
| "grad_norm": 0.7755012512207031, |
| "learning_rate": 0.0001360068958086413, |
| "loss": 3.2165, |
| "step": 71900 |
| }, |
| { |
| "epoch": 7.744053384996233, |
| "grad_norm": 0.8271076083183289, |
| "learning_rate": 0.00013568365477857988, |
| "loss": 3.234, |
| "step": 71950 |
| }, |
| { |
| "epoch": 7.749434937035841, |
| "grad_norm": 0.7885873913764954, |
| "learning_rate": 0.00013536041374851847, |
| "loss": 3.2342, |
| "step": 72000 |
| }, |
| { |
| "epoch": 7.749434937035841, |
| "eval_accuracy": 0.38911706598957085, |
| "eval_loss": 3.3421249389648438, |
| "eval_runtime": 179.9591, |
| "eval_samples_per_second": 100.084, |
| "eval_steps_per_second": 6.257, |
| "step": 72000 |
| }, |
| { |
| "epoch": 7.754816489075449, |
| "grad_norm": 0.7660197615623474, |
| "learning_rate": 0.00013503717271845706, |
| "loss": 3.2277, |
| "step": 72050 |
| }, |
| { |
| "epoch": 7.760198041115058, |
| "grad_norm": 0.9116640090942383, |
| "learning_rate": 0.00013471393168839563, |
| "loss": 3.2477, |
| "step": 72100 |
| }, |
| { |
| "epoch": 7.765579593154666, |
| "grad_norm": 0.8500056862831116, |
| "learning_rate": 0.00013439069065833423, |
| "loss": 3.2316, |
| "step": 72150 |
| }, |
| { |
| "epoch": 7.770961145194274, |
| "grad_norm": 0.8157951831817627, |
| "learning_rate": 0.0001340674496282728, |
| "loss": 3.2313, |
| "step": 72200 |
| }, |
| { |
| "epoch": 7.776342697233883, |
| "grad_norm": 0.7920417785644531, |
| "learning_rate": 0.0001337442085982114, |
| "loss": 3.2379, |
| "step": 72250 |
| }, |
| { |
| "epoch": 7.78172424927349, |
| "grad_norm": 0.8302222490310669, |
| "learning_rate": 0.00013342096756814996, |
| "loss": 3.2365, |
| "step": 72300 |
| }, |
| { |
| "epoch": 7.787105801313099, |
| "grad_norm": 0.8150511980056763, |
| "learning_rate": 0.00013309772653808855, |
| "loss": 3.2492, |
| "step": 72350 |
| }, |
| { |
| "epoch": 7.792487353352707, |
| "grad_norm": 0.8173815608024597, |
| "learning_rate": 0.00013277448550802715, |
| "loss": 3.2356, |
| "step": 72400 |
| }, |
| { |
| "epoch": 7.797868905392315, |
| "grad_norm": 0.7868077754974365, |
| "learning_rate": 0.0001324512444779657, |
| "loss": 3.2376, |
| "step": 72450 |
| }, |
| { |
| "epoch": 7.803250457431924, |
| "grad_norm": 0.7838640213012695, |
| "learning_rate": 0.0001321280034479043, |
| "loss": 3.2176, |
| "step": 72500 |
| }, |
| { |
| "epoch": 7.808632009471531, |
| "grad_norm": 0.8046959042549133, |
| "learning_rate": 0.0001318047624178429, |
| "loss": 3.2453, |
| "step": 72550 |
| }, |
| { |
| "epoch": 7.81401356151114, |
| "grad_norm": 0.8591763377189636, |
| "learning_rate": 0.0001314815213877815, |
| "loss": 3.233, |
| "step": 72600 |
| }, |
| { |
| "epoch": 7.819395113550748, |
| "grad_norm": 0.7944203019142151, |
| "learning_rate": 0.00013115828035772007, |
| "loss": 3.2246, |
| "step": 72650 |
| }, |
| { |
| "epoch": 7.824776665590356, |
| "grad_norm": 0.8305500149726868, |
| "learning_rate": 0.00013084150414825987, |
| "loss": 3.2299, |
| "step": 72700 |
| }, |
| { |
| "epoch": 7.830158217629965, |
| "grad_norm": 0.8165892362594604, |
| "learning_rate": 0.00013051826311819844, |
| "loss": 3.247, |
| "step": 72750 |
| }, |
| { |
| "epoch": 7.835539769669572, |
| "grad_norm": 0.8104371428489685, |
| "learning_rate": 0.00013019502208813703, |
| "loss": 3.2239, |
| "step": 72800 |
| }, |
| { |
| "epoch": 7.840921321709181, |
| "grad_norm": 0.808006227016449, |
| "learning_rate": 0.00012987178105807563, |
| "loss": 3.231, |
| "step": 72850 |
| }, |
| { |
| "epoch": 7.846302873748789, |
| "grad_norm": 0.7946688532829285, |
| "learning_rate": 0.00012954854002801422, |
| "loss": 3.2496, |
| "step": 72900 |
| }, |
| { |
| "epoch": 7.851684425788397, |
| "grad_norm": 0.7898575663566589, |
| "learning_rate": 0.00012923176381855403, |
| "loss": 3.2305, |
| "step": 72950 |
| }, |
| { |
| "epoch": 7.857065977828006, |
| "grad_norm": 0.8683642148971558, |
| "learning_rate": 0.00012890852278849262, |
| "loss": 3.2291, |
| "step": 73000 |
| }, |
| { |
| "epoch": 7.857065977828006, |
| "eval_accuracy": 0.38946323409014777, |
| "eval_loss": 3.33805513381958, |
| "eval_runtime": 179.9574, |
| "eval_samples_per_second": 100.085, |
| "eval_steps_per_second": 6.257, |
| "step": 73000 |
| }, |
| { |
| "epoch": 7.862447529867614, |
| "grad_norm": 0.8216552138328552, |
| "learning_rate": 0.0001285852817584312, |
| "loss": 3.2354, |
| "step": 73050 |
| }, |
| { |
| "epoch": 7.867829081907222, |
| "grad_norm": 0.8108968138694763, |
| "learning_rate": 0.0001282620407283698, |
| "loss": 3.2364, |
| "step": 73100 |
| }, |
| { |
| "epoch": 7.87321063394683, |
| "grad_norm": 0.8014196157455444, |
| "learning_rate": 0.00012793879969830835, |
| "loss": 3.2347, |
| "step": 73150 |
| }, |
| { |
| "epoch": 7.878592185986438, |
| "grad_norm": 0.7955536246299744, |
| "learning_rate": 0.00012761555866824695, |
| "loss": 3.2308, |
| "step": 73200 |
| }, |
| { |
| "epoch": 7.883973738026047, |
| "grad_norm": 0.9082058072090149, |
| "learning_rate": 0.00012729231763818552, |
| "loss": 3.2434, |
| "step": 73250 |
| }, |
| { |
| "epoch": 7.889355290065655, |
| "grad_norm": 0.842169463634491, |
| "learning_rate": 0.0001269690766081241, |
| "loss": 3.2563, |
| "step": 73300 |
| }, |
| { |
| "epoch": 7.894736842105263, |
| "grad_norm": 0.8316313028335571, |
| "learning_rate": 0.0001266458355780627, |
| "loss": 3.2582, |
| "step": 73350 |
| }, |
| { |
| "epoch": 7.900118394144871, |
| "grad_norm": 0.8240712285041809, |
| "learning_rate": 0.00012632259454800127, |
| "loss": 3.2216, |
| "step": 73400 |
| }, |
| { |
| "epoch": 7.90549994618448, |
| "grad_norm": 0.803352952003479, |
| "learning_rate": 0.00012599935351793987, |
| "loss": 3.2502, |
| "step": 73450 |
| }, |
| { |
| "epoch": 7.910881498224088, |
| "grad_norm": 0.8391147255897522, |
| "learning_rate": 0.00012567611248787846, |
| "loss": 3.2302, |
| "step": 73500 |
| }, |
| { |
| "epoch": 7.916263050263696, |
| "grad_norm": 0.8049103617668152, |
| "learning_rate": 0.00012535287145781703, |
| "loss": 3.2284, |
| "step": 73550 |
| }, |
| { |
| "epoch": 7.921644602303305, |
| "grad_norm": 0.8465777635574341, |
| "learning_rate": 0.00012502963042775562, |
| "loss": 3.2373, |
| "step": 73600 |
| }, |
| { |
| "epoch": 7.927026154342912, |
| "grad_norm": 0.8268304467201233, |
| "learning_rate": 0.00012470638939769422, |
| "loss": 3.2596, |
| "step": 73650 |
| }, |
| { |
| "epoch": 7.932407706382521, |
| "grad_norm": 0.8311086297035217, |
| "learning_rate": 0.0001243831483676328, |
| "loss": 3.2529, |
| "step": 73700 |
| }, |
| { |
| "epoch": 7.937789258422129, |
| "grad_norm": 0.8166625499725342, |
| "learning_rate": 0.00012405990733757138, |
| "loss": 3.2289, |
| "step": 73750 |
| }, |
| { |
| "epoch": 7.943170810461737, |
| "grad_norm": 0.805748462677002, |
| "learning_rate": 0.00012373666630750995, |
| "loss": 3.251, |
| "step": 73800 |
| }, |
| { |
| "epoch": 7.948552362501346, |
| "grad_norm": 0.8548930287361145, |
| "learning_rate": 0.00012341342527744854, |
| "loss": 3.2249, |
| "step": 73850 |
| }, |
| { |
| "epoch": 7.953933914540953, |
| "grad_norm": 0.8273134827613831, |
| "learning_rate": 0.0001230901842473871, |
| "loss": 3.2495, |
| "step": 73900 |
| }, |
| { |
| "epoch": 7.959315466580562, |
| "grad_norm": 0.782123327255249, |
| "learning_rate": 0.0001227669432173257, |
| "loss": 3.2184, |
| "step": 73950 |
| }, |
| { |
| "epoch": 7.96469701862017, |
| "grad_norm": 0.8412731885910034, |
| "learning_rate": 0.0001224437021872643, |
| "loss": 3.2333, |
| "step": 74000 |
| }, |
| { |
| "epoch": 7.96469701862017, |
| "eval_accuracy": 0.3899255521277984, |
| "eval_loss": 3.3341968059539795, |
| "eval_runtime": 180.0793, |
| "eval_samples_per_second": 100.017, |
| "eval_steps_per_second": 6.253, |
| "step": 74000 |
| }, |
| { |
| "epoch": 7.970078570659778, |
| "grad_norm": 0.8382527232170105, |
| "learning_rate": 0.00012212046115720287, |
| "loss": 3.2202, |
| "step": 74050 |
| }, |
| { |
| "epoch": 7.975460122699387, |
| "grad_norm": 0.8564316630363464, |
| "learning_rate": 0.00012179722012714146, |
| "loss": 3.2181, |
| "step": 74100 |
| }, |
| { |
| "epoch": 7.980841674738995, |
| "grad_norm": 0.9008880257606506, |
| "learning_rate": 0.00012147397909708004, |
| "loss": 3.2314, |
| "step": 74150 |
| }, |
| { |
| "epoch": 7.986223226778603, |
| "grad_norm": 0.7987130284309387, |
| "learning_rate": 0.00012115073806701864, |
| "loss": 3.2211, |
| "step": 74200 |
| }, |
| { |
| "epoch": 7.991604778818211, |
| "grad_norm": 0.8617604970932007, |
| "learning_rate": 0.0001208274970369572, |
| "loss": 3.2507, |
| "step": 74250 |
| }, |
| { |
| "epoch": 7.996986330857819, |
| "grad_norm": 0.8378741145133972, |
| "learning_rate": 0.0001205042560068958, |
| "loss": 3.2028, |
| "step": 74300 |
| }, |
| { |
| "epoch": 8.002367882897428, |
| "grad_norm": 0.8566402196884155, |
| "learning_rate": 0.0001201810149768344, |
| "loss": 3.2122, |
| "step": 74350 |
| }, |
| { |
| "epoch": 8.007749434937036, |
| "grad_norm": 0.8603155016899109, |
| "learning_rate": 0.00011985777394677296, |
| "loss": 3.1329, |
| "step": 74400 |
| }, |
| { |
| "epoch": 8.013130986976645, |
| "grad_norm": 0.8972188830375671, |
| "learning_rate": 0.00011953453291671154, |
| "loss": 3.162, |
| "step": 74450 |
| }, |
| { |
| "epoch": 8.018512539016251, |
| "grad_norm": 0.8691694140434265, |
| "learning_rate": 0.00011921129188665014, |
| "loss": 3.1586, |
| "step": 74500 |
| }, |
| { |
| "epoch": 8.02389409105586, |
| "grad_norm": 0.8346155285835266, |
| "learning_rate": 0.0001188880508565887, |
| "loss": 3.1576, |
| "step": 74550 |
| }, |
| { |
| "epoch": 8.029275643095469, |
| "grad_norm": 0.8023788332939148, |
| "learning_rate": 0.0001185648098265273, |
| "loss": 3.1625, |
| "step": 74600 |
| }, |
| { |
| "epoch": 8.034657195135077, |
| "grad_norm": 0.8025177121162415, |
| "learning_rate": 0.0001182415687964659, |
| "loss": 3.1463, |
| "step": 74650 |
| }, |
| { |
| "epoch": 8.040038747174686, |
| "grad_norm": 0.827228844165802, |
| "learning_rate": 0.00011791832776640448, |
| "loss": 3.1827, |
| "step": 74700 |
| }, |
| { |
| "epoch": 8.045420299214294, |
| "grad_norm": 0.8884971737861633, |
| "learning_rate": 0.00011759508673634306, |
| "loss": 3.1788, |
| "step": 74750 |
| }, |
| { |
| "epoch": 8.050801851253901, |
| "grad_norm": 0.8682385087013245, |
| "learning_rate": 0.00011727184570628164, |
| "loss": 3.1658, |
| "step": 74800 |
| }, |
| { |
| "epoch": 8.05618340329351, |
| "grad_norm": 0.847655713558197, |
| "learning_rate": 0.00011694860467622023, |
| "loss": 3.1655, |
| "step": 74850 |
| }, |
| { |
| "epoch": 8.061564955333118, |
| "grad_norm": 0.8776028156280518, |
| "learning_rate": 0.0001166253636461588, |
| "loss": 3.1699, |
| "step": 74900 |
| }, |
| { |
| "epoch": 8.066946507372727, |
| "grad_norm": 0.8304228782653809, |
| "learning_rate": 0.0001163021226160974, |
| "loss": 3.1682, |
| "step": 74950 |
| }, |
| { |
| "epoch": 8.072328059412335, |
| "grad_norm": 0.799704909324646, |
| "learning_rate": 0.00011597888158603598, |
| "loss": 3.1684, |
| "step": 75000 |
| }, |
| { |
| "epoch": 8.072328059412335, |
| "eval_accuracy": 0.3898515595111967, |
| "eval_loss": 3.3377063274383545, |
| "eval_runtime": 180.2546, |
| "eval_samples_per_second": 99.92, |
| "eval_steps_per_second": 6.247, |
| "step": 75000 |
| }, |
| { |
| "epoch": 8.077709611451942, |
| "grad_norm": 0.8407004475593567, |
| "learning_rate": 0.00011565564055597456, |
| "loss": 3.1608, |
| "step": 75050 |
| }, |
| { |
| "epoch": 8.08309116349155, |
| "grad_norm": 0.878447949886322, |
| "learning_rate": 0.00011533239952591314, |
| "loss": 3.1724, |
| "step": 75100 |
| }, |
| { |
| "epoch": 8.088472715531159, |
| "grad_norm": 0.8164194822311401, |
| "learning_rate": 0.00011500915849585173, |
| "loss": 3.1569, |
| "step": 75150 |
| }, |
| { |
| "epoch": 8.093854267570768, |
| "grad_norm": 0.8630013465881348, |
| "learning_rate": 0.00011469238228639154, |
| "loss": 3.1556, |
| "step": 75200 |
| }, |
| { |
| "epoch": 8.099235819610376, |
| "grad_norm": 0.8045172095298767, |
| "learning_rate": 0.00011436914125633012, |
| "loss": 3.1652, |
| "step": 75250 |
| }, |
| { |
| "epoch": 8.104617371649983, |
| "grad_norm": 0.8801988959312439, |
| "learning_rate": 0.00011404590022626872, |
| "loss": 3.1672, |
| "step": 75300 |
| }, |
| { |
| "epoch": 8.109998923689592, |
| "grad_norm": 0.8112406730651855, |
| "learning_rate": 0.00011372265919620728, |
| "loss": 3.1884, |
| "step": 75350 |
| }, |
| { |
| "epoch": 8.1153804757292, |
| "grad_norm": 0.8461372256278992, |
| "learning_rate": 0.00011339941816614588, |
| "loss": 3.1602, |
| "step": 75400 |
| }, |
| { |
| "epoch": 8.120762027768809, |
| "grad_norm": 0.9137732982635498, |
| "learning_rate": 0.00011307617713608447, |
| "loss": 3.1663, |
| "step": 75450 |
| }, |
| { |
| "epoch": 8.126143579808417, |
| "grad_norm": 0.8456742763519287, |
| "learning_rate": 0.00011275293610602305, |
| "loss": 3.1589, |
| "step": 75500 |
| }, |
| { |
| "epoch": 8.131525131848026, |
| "grad_norm": 0.8249267935752869, |
| "learning_rate": 0.00011242969507596164, |
| "loss": 3.1508, |
| "step": 75550 |
| }, |
| { |
| "epoch": 8.136906683887632, |
| "grad_norm": 0.8262121081352234, |
| "learning_rate": 0.00011210645404590022, |
| "loss": 3.1733, |
| "step": 75600 |
| }, |
| { |
| "epoch": 8.142288235927241, |
| "grad_norm": 0.8334835171699524, |
| "learning_rate": 0.00011178321301583881, |
| "loss": 3.1677, |
| "step": 75650 |
| }, |
| { |
| "epoch": 8.14766978796685, |
| "grad_norm": 0.8308705687522888, |
| "learning_rate": 0.00011145997198577738, |
| "loss": 3.1778, |
| "step": 75700 |
| }, |
| { |
| "epoch": 8.153051340006458, |
| "grad_norm": 0.8552451729774475, |
| "learning_rate": 0.00011113673095571597, |
| "loss": 3.1685, |
| "step": 75750 |
| }, |
| { |
| "epoch": 8.158432892046067, |
| "grad_norm": 0.8798450231552124, |
| "learning_rate": 0.00011081348992565455, |
| "loss": 3.1637, |
| "step": 75800 |
| }, |
| { |
| "epoch": 8.163814444085673, |
| "grad_norm": 0.839418888092041, |
| "learning_rate": 0.00011049024889559314, |
| "loss": 3.165, |
| "step": 75850 |
| }, |
| { |
| "epoch": 8.169195996125282, |
| "grad_norm": 0.8186075091362, |
| "learning_rate": 0.00011016700786553172, |
| "loss": 3.1782, |
| "step": 75900 |
| }, |
| { |
| "epoch": 8.17457754816489, |
| "grad_norm": 0.836857795715332, |
| "learning_rate": 0.00010984376683547031, |
| "loss": 3.1717, |
| "step": 75950 |
| }, |
| { |
| "epoch": 8.1799591002045, |
| "grad_norm": 0.78592449426651, |
| "learning_rate": 0.00010952052580540889, |
| "loss": 3.1849, |
| "step": 76000 |
| }, |
| { |
| "epoch": 8.1799591002045, |
| "eval_accuracy": 0.39025553094809096, |
| "eval_loss": 3.335026979446411, |
| "eval_runtime": 179.5655, |
| "eval_samples_per_second": 100.303, |
| "eval_steps_per_second": 6.271, |
| "step": 76000 |
| }, |
| { |
| "epoch": 8.185340652244108, |
| "grad_norm": 0.8572103381156921, |
| "learning_rate": 0.00010919728477534747, |
| "loss": 3.18, |
| "step": 76050 |
| }, |
| { |
| "epoch": 8.190722204283716, |
| "grad_norm": 0.8487206101417542, |
| "learning_rate": 0.0001088805085658873, |
| "loss": 3.165, |
| "step": 76100 |
| }, |
| { |
| "epoch": 8.196103756323323, |
| "grad_norm": 0.8511477112770081, |
| "learning_rate": 0.00010855726753582586, |
| "loss": 3.1883, |
| "step": 76150 |
| }, |
| { |
| "epoch": 8.201485308362932, |
| "grad_norm": 0.8475496172904968, |
| "learning_rate": 0.00010823402650576446, |
| "loss": 3.1702, |
| "step": 76200 |
| }, |
| { |
| "epoch": 8.20686686040254, |
| "grad_norm": 0.8378956317901611, |
| "learning_rate": 0.00010791078547570304, |
| "loss": 3.1585, |
| "step": 76250 |
| }, |
| { |
| "epoch": 8.212248412442149, |
| "grad_norm": 0.8379483222961426, |
| "learning_rate": 0.00010758754444564162, |
| "loss": 3.1581, |
| "step": 76300 |
| }, |
| { |
| "epoch": 8.217629964481757, |
| "grad_norm": 0.8493615388870239, |
| "learning_rate": 0.0001072643034155802, |
| "loss": 3.1709, |
| "step": 76350 |
| }, |
| { |
| "epoch": 8.223011516521364, |
| "grad_norm": 0.8749557137489319, |
| "learning_rate": 0.0001069410623855188, |
| "loss": 3.1636, |
| "step": 76400 |
| }, |
| { |
| "epoch": 8.228393068560973, |
| "grad_norm": 0.8558868765830994, |
| "learning_rate": 0.00010661782135545739, |
| "loss": 3.1669, |
| "step": 76450 |
| }, |
| { |
| "epoch": 8.233774620600581, |
| "grad_norm": 0.813459575176239, |
| "learning_rate": 0.00010629458032539596, |
| "loss": 3.1802, |
| "step": 76500 |
| }, |
| { |
| "epoch": 8.23915617264019, |
| "grad_norm": 0.8157198429107666, |
| "learning_rate": 0.00010597133929533455, |
| "loss": 3.1721, |
| "step": 76550 |
| }, |
| { |
| "epoch": 8.244537724679798, |
| "grad_norm": 0.8524532318115234, |
| "learning_rate": 0.00010564809826527313, |
| "loss": 3.1772, |
| "step": 76600 |
| }, |
| { |
| "epoch": 8.249919276719407, |
| "grad_norm": 0.813799262046814, |
| "learning_rate": 0.00010532485723521171, |
| "loss": 3.1585, |
| "step": 76650 |
| }, |
| { |
| "epoch": 8.255300828759013, |
| "grad_norm": 0.8357850909233093, |
| "learning_rate": 0.0001050016162051503, |
| "loss": 3.1791, |
| "step": 76700 |
| }, |
| { |
| "epoch": 8.260682380798622, |
| "grad_norm": 0.8634746670722961, |
| "learning_rate": 0.00010467837517508889, |
| "loss": 3.1795, |
| "step": 76750 |
| }, |
| { |
| "epoch": 8.26606393283823, |
| "grad_norm": 0.8267971277236938, |
| "learning_rate": 0.00010435513414502746, |
| "loss": 3.1719, |
| "step": 76800 |
| }, |
| { |
| "epoch": 8.27144548487784, |
| "grad_norm": 0.8415201306343079, |
| "learning_rate": 0.00010403189311496605, |
| "loss": 3.19, |
| "step": 76850 |
| }, |
| { |
| "epoch": 8.276827036917448, |
| "grad_norm": 0.8072708249092102, |
| "learning_rate": 0.00010370865208490463, |
| "loss": 3.1761, |
| "step": 76900 |
| }, |
| { |
| "epoch": 8.282208588957054, |
| "grad_norm": 0.8550540208816528, |
| "learning_rate": 0.00010338541105484323, |
| "loss": 3.1753, |
| "step": 76950 |
| }, |
| { |
| "epoch": 8.287590140996663, |
| "grad_norm": 0.8370460271835327, |
| "learning_rate": 0.0001030621700247818, |
| "loss": 3.1775, |
| "step": 77000 |
| }, |
| { |
| "epoch": 8.287590140996663, |
| "eval_accuracy": 0.3904935894252925, |
| "eval_loss": 3.3347671031951904, |
| "eval_runtime": 179.9264, |
| "eval_samples_per_second": 100.102, |
| "eval_steps_per_second": 6.258, |
| "step": 77000 |
| }, |
| { |
| "epoch": 8.292971693036272, |
| "grad_norm": 0.8297492265701294, |
| "learning_rate": 0.00010273892899472039, |
| "loss": 3.1716, |
| "step": 77050 |
| }, |
| { |
| "epoch": 8.29835324507588, |
| "grad_norm": 0.8193721175193787, |
| "learning_rate": 0.00010241568796465898, |
| "loss": 3.1875, |
| "step": 77100 |
| }, |
| { |
| "epoch": 8.303734797115489, |
| "grad_norm": 0.8413053154945374, |
| "learning_rate": 0.00010209244693459755, |
| "loss": 3.1671, |
| "step": 77150 |
| }, |
| { |
| "epoch": 8.309116349155097, |
| "grad_norm": 0.8616971969604492, |
| "learning_rate": 0.00010176920590453613, |
| "loss": 3.1864, |
| "step": 77200 |
| }, |
| { |
| "epoch": 8.314497901194704, |
| "grad_norm": 0.8818926811218262, |
| "learning_rate": 0.00010144596487447473, |
| "loss": 3.175, |
| "step": 77250 |
| }, |
| { |
| "epoch": 8.319879453234313, |
| "grad_norm": 0.8248533606529236, |
| "learning_rate": 0.00010112272384441332, |
| "loss": 3.1903, |
| "step": 77300 |
| }, |
| { |
| "epoch": 8.325261005273921, |
| "grad_norm": 0.8652777075767517, |
| "learning_rate": 0.00010079948281435189, |
| "loss": 3.1636, |
| "step": 77350 |
| }, |
| { |
| "epoch": 8.33064255731353, |
| "grad_norm": 0.819579541683197, |
| "learning_rate": 0.00010047624178429048, |
| "loss": 3.1682, |
| "step": 77400 |
| }, |
| { |
| "epoch": 8.336024109353138, |
| "grad_norm": 0.795547366142273, |
| "learning_rate": 0.00010015300075422906, |
| "loss": 3.1727, |
| "step": 77450 |
| }, |
| { |
| "epoch": 8.341405661392745, |
| "grad_norm": 0.8604868650436401, |
| "learning_rate": 9.982975972416765e-05, |
| "loss": 3.1878, |
| "step": 77500 |
| }, |
| { |
| "epoch": 8.346787213432354, |
| "grad_norm": 0.8407930135726929, |
| "learning_rate": 9.950651869410623e-05, |
| "loss": 3.1736, |
| "step": 77550 |
| }, |
| { |
| "epoch": 8.352168765471962, |
| "grad_norm": 0.8258522748947144, |
| "learning_rate": 9.918327766404482e-05, |
| "loss": 3.1694, |
| "step": 77600 |
| }, |
| { |
| "epoch": 8.35755031751157, |
| "grad_norm": 0.8680301308631897, |
| "learning_rate": 9.886003663398339e-05, |
| "loss": 3.1864, |
| "step": 77650 |
| }, |
| { |
| "epoch": 8.36293186955118, |
| "grad_norm": 0.8190805912017822, |
| "learning_rate": 9.853679560392198e-05, |
| "loss": 3.1861, |
| "step": 77700 |
| }, |
| { |
| "epoch": 8.368313421590786, |
| "grad_norm": 0.8155598044395447, |
| "learning_rate": 9.821355457386056e-05, |
| "loss": 3.1661, |
| "step": 77750 |
| }, |
| { |
| "epoch": 8.373694973630395, |
| "grad_norm": 0.850091814994812, |
| "learning_rate": 9.789031354379916e-05, |
| "loss": 3.1893, |
| "step": 77800 |
| }, |
| { |
| "epoch": 8.379076525670003, |
| "grad_norm": 0.8463276028633118, |
| "learning_rate": 9.756707251373773e-05, |
| "loss": 3.1879, |
| "step": 77850 |
| }, |
| { |
| "epoch": 8.384458077709612, |
| "grad_norm": 0.8920957446098328, |
| "learning_rate": 9.724383148367632e-05, |
| "loss": 3.1779, |
| "step": 77900 |
| }, |
| { |
| "epoch": 8.38983962974922, |
| "grad_norm": 0.8446083068847656, |
| "learning_rate": 9.692059045361492e-05, |
| "loss": 3.1744, |
| "step": 77950 |
| }, |
| { |
| "epoch": 8.395221181788829, |
| "grad_norm": 0.864314079284668, |
| "learning_rate": 9.659734942355348e-05, |
| "loss": 3.1833, |
| "step": 78000 |
| }, |
| { |
| "epoch": 8.395221181788829, |
| "eval_accuracy": 0.3908043366844419, |
| "eval_loss": 3.33040189743042, |
| "eval_runtime": 179.9825, |
| "eval_samples_per_second": 100.071, |
| "eval_steps_per_second": 6.256, |
| "step": 78000 |
| }, |
| { |
| "epoch": 8.400602733828435, |
| "grad_norm": 0.8862897157669067, |
| "learning_rate": 9.627410839349208e-05, |
| "loss": 3.1755, |
| "step": 78050 |
| }, |
| { |
| "epoch": 8.405984285868044, |
| "grad_norm": 0.819072961807251, |
| "learning_rate": 9.595086736343066e-05, |
| "loss": 3.176, |
| "step": 78100 |
| }, |
| { |
| "epoch": 8.411365837907653, |
| "grad_norm": 0.8221203088760376, |
| "learning_rate": 9.562762633336923e-05, |
| "loss": 3.1816, |
| "step": 78150 |
| }, |
| { |
| "epoch": 8.416747389947261, |
| "grad_norm": 0.8711651563644409, |
| "learning_rate": 9.530438530330782e-05, |
| "loss": 3.1838, |
| "step": 78200 |
| }, |
| { |
| "epoch": 8.42212894198687, |
| "grad_norm": 0.8410494923591614, |
| "learning_rate": 9.498114427324642e-05, |
| "loss": 3.1768, |
| "step": 78250 |
| }, |
| { |
| "epoch": 8.427510494026476, |
| "grad_norm": 0.8514099717140198, |
| "learning_rate": 9.4657903243185e-05, |
| "loss": 3.166, |
| "step": 78300 |
| }, |
| { |
| "epoch": 8.432892046066085, |
| "grad_norm": 0.8536844253540039, |
| "learning_rate": 9.433466221312358e-05, |
| "loss": 3.172, |
| "step": 78350 |
| }, |
| { |
| "epoch": 8.438273598105694, |
| "grad_norm": 0.8886139988899231, |
| "learning_rate": 9.401142118306216e-05, |
| "loss": 3.1935, |
| "step": 78400 |
| }, |
| { |
| "epoch": 8.443655150145302, |
| "grad_norm": 0.8296248912811279, |
| "learning_rate": 9.368818015300075e-05, |
| "loss": 3.1878, |
| "step": 78450 |
| }, |
| { |
| "epoch": 8.44903670218491, |
| "grad_norm": 0.88595050573349, |
| "learning_rate": 9.336493912293932e-05, |
| "loss": 3.1633, |
| "step": 78500 |
| }, |
| { |
| "epoch": 8.45441825422452, |
| "grad_norm": 0.8764588832855225, |
| "learning_rate": 9.304169809287792e-05, |
| "loss": 3.1798, |
| "step": 78550 |
| }, |
| { |
| "epoch": 8.459799806264126, |
| "grad_norm": 0.830272912979126, |
| "learning_rate": 9.27184570628165e-05, |
| "loss": 3.1857, |
| "step": 78600 |
| }, |
| { |
| "epoch": 8.465181358303735, |
| "grad_norm": 0.8409690856933594, |
| "learning_rate": 9.239521603275508e-05, |
| "loss": 3.1858, |
| "step": 78650 |
| }, |
| { |
| "epoch": 8.470562910343343, |
| "grad_norm": 0.8955550789833069, |
| "learning_rate": 9.207197500269366e-05, |
| "loss": 3.1948, |
| "step": 78700 |
| }, |
| { |
| "epoch": 8.475944462382952, |
| "grad_norm": 0.8606041073799133, |
| "learning_rate": 9.174873397263225e-05, |
| "loss": 3.2018, |
| "step": 78750 |
| }, |
| { |
| "epoch": 8.48132601442256, |
| "grad_norm": 0.8329320549964905, |
| "learning_rate": 9.142549294257085e-05, |
| "loss": 3.1797, |
| "step": 78800 |
| }, |
| { |
| "epoch": 8.486707566462167, |
| "grad_norm": 0.8341664671897888, |
| "learning_rate": 9.110225191250942e-05, |
| "loss": 3.1785, |
| "step": 78850 |
| }, |
| { |
| "epoch": 8.492089118501776, |
| "grad_norm": 0.8955467939376831, |
| "learning_rate": 9.077901088244801e-05, |
| "loss": 3.1911, |
| "step": 78900 |
| }, |
| { |
| "epoch": 8.497470670541384, |
| "grad_norm": 0.8430368304252625, |
| "learning_rate": 9.045576985238659e-05, |
| "loss": 3.1755, |
| "step": 78950 |
| }, |
| { |
| "epoch": 8.502852222580993, |
| "grad_norm": 0.8482922315597534, |
| "learning_rate": 9.013252882232517e-05, |
| "loss": 3.183, |
| "step": 79000 |
| }, |
| { |
| "epoch": 8.502852222580993, |
| "eval_accuracy": 0.39098600431286773, |
| "eval_loss": 3.328646183013916, |
| "eval_runtime": 180.2849, |
| "eval_samples_per_second": 99.903, |
| "eval_steps_per_second": 6.246, |
| "step": 79000 |
| }, |
| { |
| "epoch": 8.508233774620601, |
| "grad_norm": 0.860719621181488, |
| "learning_rate": 8.980928779226375e-05, |
| "loss": 3.2104, |
| "step": 79050 |
| }, |
| { |
| "epoch": 8.513615326660208, |
| "grad_norm": 0.836053729057312, |
| "learning_rate": 8.948604676220235e-05, |
| "loss": 3.1859, |
| "step": 79100 |
| }, |
| { |
| "epoch": 8.518996878699816, |
| "grad_norm": 0.8130183815956116, |
| "learning_rate": 8.916280573214092e-05, |
| "loss": 3.1714, |
| "step": 79150 |
| }, |
| { |
| "epoch": 8.524378430739425, |
| "grad_norm": 0.927423357963562, |
| "learning_rate": 8.883956470207951e-05, |
| "loss": 3.2046, |
| "step": 79200 |
| }, |
| { |
| "epoch": 8.529759982779034, |
| "grad_norm": 0.8596153855323792, |
| "learning_rate": 8.851632367201809e-05, |
| "loss": 3.183, |
| "step": 79250 |
| }, |
| { |
| "epoch": 8.535141534818642, |
| "grad_norm": 0.8415372371673584, |
| "learning_rate": 8.819308264195669e-05, |
| "loss": 3.1731, |
| "step": 79300 |
| }, |
| { |
| "epoch": 8.54052308685825, |
| "grad_norm": 0.8734657168388367, |
| "learning_rate": 8.786984161189526e-05, |
| "loss": 3.1889, |
| "step": 79350 |
| }, |
| { |
| "epoch": 8.545904638897857, |
| "grad_norm": 0.838965654373169, |
| "learning_rate": 8.754660058183385e-05, |
| "loss": 3.1763, |
| "step": 79400 |
| }, |
| { |
| "epoch": 8.551286190937466, |
| "grad_norm": 0.8917363286018372, |
| "learning_rate": 8.722335955177243e-05, |
| "loss": 3.1875, |
| "step": 79450 |
| }, |
| { |
| "epoch": 8.556667742977075, |
| "grad_norm": 0.8633471727371216, |
| "learning_rate": 8.690011852171101e-05, |
| "loss": 3.1781, |
| "step": 79500 |
| }, |
| { |
| "epoch": 8.562049295016683, |
| "grad_norm": 0.8390756845474243, |
| "learning_rate": 8.657687749164959e-05, |
| "loss": 3.1836, |
| "step": 79550 |
| }, |
| { |
| "epoch": 8.567430847056292, |
| "grad_norm": 0.8166258931159973, |
| "learning_rate": 8.625363646158819e-05, |
| "loss": 3.1666, |
| "step": 79600 |
| }, |
| { |
| "epoch": 8.572812399095898, |
| "grad_norm": 0.857069194316864, |
| "learning_rate": 8.593039543152676e-05, |
| "loss": 3.1912, |
| "step": 79650 |
| }, |
| { |
| "epoch": 8.578193951135507, |
| "grad_norm": 0.8506762385368347, |
| "learning_rate": 8.560715440146535e-05, |
| "loss": 3.1866, |
| "step": 79700 |
| }, |
| { |
| "epoch": 8.583575503175116, |
| "grad_norm": 0.8568733334541321, |
| "learning_rate": 8.528391337140394e-05, |
| "loss": 3.1747, |
| "step": 79750 |
| }, |
| { |
| "epoch": 8.588957055214724, |
| "grad_norm": 0.8337421417236328, |
| "learning_rate": 8.496067234134253e-05, |
| "loss": 3.1741, |
| "step": 79800 |
| }, |
| { |
| "epoch": 8.594338607254333, |
| "grad_norm": 0.8349370360374451, |
| "learning_rate": 8.46374313112811e-05, |
| "loss": 3.1783, |
| "step": 79850 |
| }, |
| { |
| "epoch": 8.599720159293941, |
| "grad_norm": 0.9050174355506897, |
| "learning_rate": 8.431419028121969e-05, |
| "loss": 3.1898, |
| "step": 79900 |
| }, |
| { |
| "epoch": 8.605101711333548, |
| "grad_norm": 0.8692119717597961, |
| "learning_rate": 8.399094925115828e-05, |
| "loss": 3.1776, |
| "step": 79950 |
| }, |
| { |
| "epoch": 8.610483263373157, |
| "grad_norm": 0.8102073669433594, |
| "learning_rate": 8.366770822109685e-05, |
| "loss": 3.1628, |
| "step": 80000 |
| }, |
| { |
| "epoch": 8.610483263373157, |
| "eval_accuracy": 0.39150384397619153, |
| "eval_loss": 3.32488751411438, |
| "eval_runtime": 179.9701, |
| "eval_samples_per_second": 100.078, |
| "eval_steps_per_second": 6.257, |
| "step": 80000 |
| }, |
| { |
| "epoch": 8.615864815412765, |
| "grad_norm": 0.8268591165542603, |
| "learning_rate": 8.334446719103544e-05, |
| "loss": 3.1842, |
| "step": 80050 |
| }, |
| { |
| "epoch": 8.621246367452374, |
| "grad_norm": 0.877962052822113, |
| "learning_rate": 8.302122616097403e-05, |
| "loss": 3.2023, |
| "step": 80100 |
| }, |
| { |
| "epoch": 8.626627919491982, |
| "grad_norm": 0.8843813538551331, |
| "learning_rate": 8.269798513091261e-05, |
| "loss": 3.1682, |
| "step": 80150 |
| }, |
| { |
| "epoch": 8.632009471531589, |
| "grad_norm": 0.8113887906074524, |
| "learning_rate": 8.237474410085119e-05, |
| "loss": 3.1886, |
| "step": 80200 |
| }, |
| { |
| "epoch": 8.637391023571197, |
| "grad_norm": 0.8340319991111755, |
| "learning_rate": 8.205150307078978e-05, |
| "loss": 3.2039, |
| "step": 80250 |
| }, |
| { |
| "epoch": 8.642772575610806, |
| "grad_norm": 0.8976853489875793, |
| "learning_rate": 8.172826204072838e-05, |
| "loss": 3.2052, |
| "step": 80300 |
| }, |
| { |
| "epoch": 8.648154127650415, |
| "grad_norm": 0.8994436264038086, |
| "learning_rate": 8.140502101066694e-05, |
| "loss": 3.1815, |
| "step": 80350 |
| }, |
| { |
| "epoch": 8.653535679690023, |
| "grad_norm": 0.8481301665306091, |
| "learning_rate": 8.108177998060553e-05, |
| "loss": 3.1876, |
| "step": 80400 |
| }, |
| { |
| "epoch": 8.658917231729632, |
| "grad_norm": 0.8830190896987915, |
| "learning_rate": 8.075853895054412e-05, |
| "loss": 3.1799, |
| "step": 80450 |
| }, |
| { |
| "epoch": 8.664298783769238, |
| "grad_norm": 0.8312429785728455, |
| "learning_rate": 8.043529792048269e-05, |
| "loss": 3.1841, |
| "step": 80500 |
| }, |
| { |
| "epoch": 8.669680335808847, |
| "grad_norm": 0.8653261661529541, |
| "learning_rate": 8.011205689042128e-05, |
| "loss": 3.1789, |
| "step": 80550 |
| }, |
| { |
| "epoch": 8.675061887848456, |
| "grad_norm": 0.8662141561508179, |
| "learning_rate": 7.978881586035988e-05, |
| "loss": 3.171, |
| "step": 80600 |
| }, |
| { |
| "epoch": 8.680443439888064, |
| "grad_norm": 0.9127115607261658, |
| "learning_rate": 7.946557483029845e-05, |
| "loss": 3.1847, |
| "step": 80650 |
| }, |
| { |
| "epoch": 8.685824991927673, |
| "grad_norm": 0.8367604613304138, |
| "learning_rate": 7.914233380023704e-05, |
| "loss": 3.1722, |
| "step": 80700 |
| }, |
| { |
| "epoch": 8.69120654396728, |
| "grad_norm": 0.8278328776359558, |
| "learning_rate": 7.881909277017562e-05, |
| "loss": 3.1743, |
| "step": 80750 |
| }, |
| { |
| "epoch": 8.696588096006888, |
| "grad_norm": 0.8759754300117493, |
| "learning_rate": 7.849585174011422e-05, |
| "loss": 3.1851, |
| "step": 80800 |
| }, |
| { |
| "epoch": 8.701969648046497, |
| "grad_norm": 0.8897731304168701, |
| "learning_rate": 7.817261071005278e-05, |
| "loss": 3.1884, |
| "step": 80850 |
| }, |
| { |
| "epoch": 8.707351200086105, |
| "grad_norm": 0.846344530582428, |
| "learning_rate": 7.784936967999138e-05, |
| "loss": 3.2009, |
| "step": 80900 |
| }, |
| { |
| "epoch": 8.712732752125714, |
| "grad_norm": 0.832398533821106, |
| "learning_rate": 7.752612864992996e-05, |
| "loss": 3.188, |
| "step": 80950 |
| }, |
| { |
| "epoch": 8.718114304165322, |
| "grad_norm": 0.8839730024337769, |
| "learning_rate": 7.720288761986854e-05, |
| "loss": 3.1654, |
| "step": 81000 |
| }, |
| { |
| "epoch": 8.718114304165322, |
| "eval_accuracy": 0.3917791703939134, |
| "eval_loss": 3.3209238052368164, |
| "eval_runtime": 179.8859, |
| "eval_samples_per_second": 100.125, |
| "eval_steps_per_second": 6.26, |
| "step": 81000 |
| }, |
| { |
| "epoch": 8.723495856204929, |
| "grad_norm": 0.8395397067070007, |
| "learning_rate": 7.687964658980712e-05, |
| "loss": 3.1755, |
| "step": 81050 |
| }, |
| { |
| "epoch": 8.728877408244538, |
| "grad_norm": 0.8440877199172974, |
| "learning_rate": 7.655640555974572e-05, |
| "loss": 3.2005, |
| "step": 81100 |
| }, |
| { |
| "epoch": 8.734258960284146, |
| "grad_norm": 0.8579280376434326, |
| "learning_rate": 7.623316452968428e-05, |
| "loss": 3.1813, |
| "step": 81150 |
| }, |
| { |
| "epoch": 8.739640512323755, |
| "grad_norm": 0.8807401657104492, |
| "learning_rate": 7.590992349962288e-05, |
| "loss": 3.1917, |
| "step": 81200 |
| }, |
| { |
| "epoch": 8.745022064363363, |
| "grad_norm": 0.8908937573432922, |
| "learning_rate": 7.558668246956147e-05, |
| "loss": 3.1825, |
| "step": 81250 |
| }, |
| { |
| "epoch": 8.75040361640297, |
| "grad_norm": 0.9015775322914124, |
| "learning_rate": 7.526990626010127e-05, |
| "loss": 3.1864, |
| "step": 81300 |
| }, |
| { |
| "epoch": 8.755785168442578, |
| "grad_norm": 0.823549211025238, |
| "learning_rate": 7.494666523003986e-05, |
| "loss": 3.1703, |
| "step": 81350 |
| }, |
| { |
| "epoch": 8.761166720482187, |
| "grad_norm": 0.9204948544502258, |
| "learning_rate": 7.462342419997844e-05, |
| "loss": 3.1872, |
| "step": 81400 |
| }, |
| { |
| "epoch": 8.766548272521796, |
| "grad_norm": 0.8299639225006104, |
| "learning_rate": 7.430018316991704e-05, |
| "loss": 3.2054, |
| "step": 81450 |
| }, |
| { |
| "epoch": 8.771929824561404, |
| "grad_norm": 0.8888871669769287, |
| "learning_rate": 7.397694213985562e-05, |
| "loss": 3.1921, |
| "step": 81500 |
| }, |
| { |
| "epoch": 8.777311376601011, |
| "grad_norm": 0.8606129288673401, |
| "learning_rate": 7.36537011097942e-05, |
| "loss": 3.1933, |
| "step": 81550 |
| }, |
| { |
| "epoch": 8.78269292864062, |
| "grad_norm": 0.8673796653747559, |
| "learning_rate": 7.333046007973278e-05, |
| "loss": 3.1882, |
| "step": 81600 |
| }, |
| { |
| "epoch": 8.788074480680228, |
| "grad_norm": 0.8238900899887085, |
| "learning_rate": 7.300721904967136e-05, |
| "loss": 3.1728, |
| "step": 81650 |
| }, |
| { |
| "epoch": 8.793456032719837, |
| "grad_norm": 0.8531633615493774, |
| "learning_rate": 7.268397801960996e-05, |
| "loss": 3.1705, |
| "step": 81700 |
| }, |
| { |
| "epoch": 8.798837584759445, |
| "grad_norm": 0.84257972240448, |
| "learning_rate": 7.236073698954854e-05, |
| "loss": 3.1883, |
| "step": 81750 |
| }, |
| { |
| "epoch": 8.804219136799054, |
| "grad_norm": 0.9469799399375916, |
| "learning_rate": 7.203749595948712e-05, |
| "loss": 3.1883, |
| "step": 81800 |
| }, |
| { |
| "epoch": 8.80960068883866, |
| "grad_norm": 0.9426635503768921, |
| "learning_rate": 7.17142549294257e-05, |
| "loss": 3.1985, |
| "step": 81850 |
| }, |
| { |
| "epoch": 8.814982240878269, |
| "grad_norm": 0.8460460305213928, |
| "learning_rate": 7.139101389936428e-05, |
| "loss": 3.1817, |
| "step": 81900 |
| }, |
| { |
| "epoch": 8.820363792917878, |
| "grad_norm": 0.8010488748550415, |
| "learning_rate": 7.106777286930287e-05, |
| "loss": 3.1964, |
| "step": 81950 |
| }, |
| { |
| "epoch": 8.825745344957486, |
| "grad_norm": 0.9217504262924194, |
| "learning_rate": 7.074453183924146e-05, |
| "loss": 3.1983, |
| "step": 82000 |
| }, |
| { |
| "epoch": 8.825745344957486, |
| "eval_accuracy": 0.3922471383817304, |
| "eval_loss": 3.3171520233154297, |
| "eval_runtime": 179.8992, |
| "eval_samples_per_second": 100.117, |
| "eval_steps_per_second": 6.259, |
| "step": 82000 |
| }, |
| { |
| "epoch": 8.831126896997095, |
| "grad_norm": 0.8669100403785706, |
| "learning_rate": 7.042129080918004e-05, |
| "loss": 3.1705, |
| "step": 82050 |
| }, |
| { |
| "epoch": 8.836508449036701, |
| "grad_norm": 0.9114564657211304, |
| "learning_rate": 7.009804977911862e-05, |
| "loss": 3.1815, |
| "step": 82100 |
| }, |
| { |
| "epoch": 8.84189000107631, |
| "grad_norm": 0.8626256585121155, |
| "learning_rate": 6.97748087490572e-05, |
| "loss": 3.175, |
| "step": 82150 |
| }, |
| { |
| "epoch": 8.847271553115919, |
| "grad_norm": 0.842830240726471, |
| "learning_rate": 6.945156771899579e-05, |
| "loss": 3.1945, |
| "step": 82200 |
| }, |
| { |
| "epoch": 8.852653105155527, |
| "grad_norm": 0.8612439036369324, |
| "learning_rate": 6.912832668893437e-05, |
| "loss": 3.1973, |
| "step": 82250 |
| }, |
| { |
| "epoch": 8.858034657195136, |
| "grad_norm": 0.8703390955924988, |
| "learning_rate": 6.880508565887297e-05, |
| "loss": 3.1825, |
| "step": 82300 |
| }, |
| { |
| "epoch": 8.863416209234742, |
| "grad_norm": 0.9062275290489197, |
| "learning_rate": 6.848184462881155e-05, |
| "loss": 3.1814, |
| "step": 82350 |
| }, |
| { |
| "epoch": 8.868797761274351, |
| "grad_norm": 0.895692765712738, |
| "learning_rate": 6.815860359875013e-05, |
| "loss": 3.1771, |
| "step": 82400 |
| }, |
| { |
| "epoch": 8.87417931331396, |
| "grad_norm": 0.8932182788848877, |
| "learning_rate": 6.783536256868871e-05, |
| "loss": 3.1825, |
| "step": 82450 |
| }, |
| { |
| "epoch": 8.879560865353568, |
| "grad_norm": 0.8720111846923828, |
| "learning_rate": 6.75121215386273e-05, |
| "loss": 3.1972, |
| "step": 82500 |
| }, |
| { |
| "epoch": 8.884942417393177, |
| "grad_norm": 0.8333724737167358, |
| "learning_rate": 6.718888050856589e-05, |
| "loss": 3.199, |
| "step": 82550 |
| }, |
| { |
| "epoch": 8.890323969432785, |
| "grad_norm": 0.8801641464233398, |
| "learning_rate": 6.686563947850447e-05, |
| "loss": 3.1722, |
| "step": 82600 |
| }, |
| { |
| "epoch": 8.895705521472392, |
| "grad_norm": 0.8496854305267334, |
| "learning_rate": 6.654239844844305e-05, |
| "loss": 3.1837, |
| "step": 82650 |
| }, |
| { |
| "epoch": 8.901087073512, |
| "grad_norm": 0.888050377368927, |
| "learning_rate": 6.621915741838163e-05, |
| "loss": 3.1976, |
| "step": 82700 |
| }, |
| { |
| "epoch": 8.906468625551609, |
| "grad_norm": 0.8081518411636353, |
| "learning_rate": 6.589591638832021e-05, |
| "loss": 3.188, |
| "step": 82750 |
| }, |
| { |
| "epoch": 8.911850177591218, |
| "grad_norm": 0.8638512492179871, |
| "learning_rate": 6.557267535825881e-05, |
| "loss": 3.1785, |
| "step": 82800 |
| }, |
| { |
| "epoch": 8.917231729630826, |
| "grad_norm": 0.8664565682411194, |
| "learning_rate": 6.524943432819739e-05, |
| "loss": 3.1917, |
| "step": 82850 |
| }, |
| { |
| "epoch": 8.922613281670433, |
| "grad_norm": 0.8508527874946594, |
| "learning_rate": 6.492619329813597e-05, |
| "loss": 3.1835, |
| "step": 82900 |
| }, |
| { |
| "epoch": 8.927994833710041, |
| "grad_norm": 0.8651978373527527, |
| "learning_rate": 6.460295226807455e-05, |
| "loss": 3.1777, |
| "step": 82950 |
| }, |
| { |
| "epoch": 8.93337638574965, |
| "grad_norm": 0.8856733441352844, |
| "learning_rate": 6.427971123801313e-05, |
| "loss": 3.1908, |
| "step": 83000 |
| }, |
| { |
| "epoch": 8.93337638574965, |
| "eval_accuracy": 0.3923557912695449, |
| "eval_loss": 3.314030170440674, |
| "eval_runtime": 179.6808, |
| "eval_samples_per_second": 100.239, |
| "eval_steps_per_second": 6.267, |
| "step": 83000 |
| }, |
| { |
| "epoch": 8.938757937789259, |
| "grad_norm": 0.8681096434593201, |
| "learning_rate": 6.395647020795173e-05, |
| "loss": 3.1862, |
| "step": 83050 |
| }, |
| { |
| "epoch": 8.944139489828867, |
| "grad_norm": 0.8370834589004517, |
| "learning_rate": 6.363322917789031e-05, |
| "loss": 3.178, |
| "step": 83100 |
| }, |
| { |
| "epoch": 8.949521041868476, |
| "grad_norm": 0.8382720947265625, |
| "learning_rate": 6.330998814782889e-05, |
| "loss": 3.1773, |
| "step": 83150 |
| }, |
| { |
| "epoch": 8.954902593908082, |
| "grad_norm": 0.918395459651947, |
| "learning_rate": 6.298674711776748e-05, |
| "loss": 3.1927, |
| "step": 83200 |
| }, |
| { |
| "epoch": 8.960284145947691, |
| "grad_norm": 0.8420078754425049, |
| "learning_rate": 6.266350608770606e-05, |
| "loss": 3.1879, |
| "step": 83250 |
| }, |
| { |
| "epoch": 8.9656656979873, |
| "grad_norm": 0.8565245866775513, |
| "learning_rate": 6.234026505764465e-05, |
| "loss": 3.1709, |
| "step": 83300 |
| }, |
| { |
| "epoch": 8.971047250026908, |
| "grad_norm": 0.8625103235244751, |
| "learning_rate": 6.201702402758323e-05, |
| "loss": 3.1964, |
| "step": 83350 |
| }, |
| { |
| "epoch": 8.976428802066517, |
| "grad_norm": 0.8943954706192017, |
| "learning_rate": 6.169378299752181e-05, |
| "loss": 3.2013, |
| "step": 83400 |
| }, |
| { |
| "epoch": 8.981810354106123, |
| "grad_norm": 0.9224780201911926, |
| "learning_rate": 6.13705419674604e-05, |
| "loss": 3.1733, |
| "step": 83450 |
| }, |
| { |
| "epoch": 8.987191906145732, |
| "grad_norm": 0.8462356925010681, |
| "learning_rate": 6.104730093739898e-05, |
| "loss": 3.1706, |
| "step": 83500 |
| }, |
| { |
| "epoch": 8.99257345818534, |
| "grad_norm": 0.8774244785308838, |
| "learning_rate": 6.072405990733757e-05, |
| "loss": 3.1717, |
| "step": 83550 |
| }, |
| { |
| "epoch": 8.997955010224949, |
| "grad_norm": 0.8515488505363464, |
| "learning_rate": 6.040081887727615e-05, |
| "loss": 3.1694, |
| "step": 83600 |
| }, |
| { |
| "epoch": 9.003336562264558, |
| "grad_norm": 0.8341154456138611, |
| "learning_rate": 6.007757784721473e-05, |
| "loss": 3.1465, |
| "step": 83650 |
| }, |
| { |
| "epoch": 9.008718114304166, |
| "grad_norm": 0.8942939043045044, |
| "learning_rate": 5.975433681715332e-05, |
| "loss": 3.1279, |
| "step": 83700 |
| }, |
| { |
| "epoch": 9.014099666343773, |
| "grad_norm": 0.8541099429130554, |
| "learning_rate": 5.94310957870919e-05, |
| "loss": 3.0992, |
| "step": 83750 |
| }, |
| { |
| "epoch": 9.019481218383381, |
| "grad_norm": 0.8721352219581604, |
| "learning_rate": 5.910785475703049e-05, |
| "loss": 3.1244, |
| "step": 83800 |
| }, |
| { |
| "epoch": 9.02486277042299, |
| "grad_norm": 0.877285897731781, |
| "learning_rate": 5.878461372696907e-05, |
| "loss": 3.1187, |
| "step": 83850 |
| }, |
| { |
| "epoch": 9.030244322462599, |
| "grad_norm": 0.9021511673927307, |
| "learning_rate": 5.846137269690765e-05, |
| "loss": 3.1431, |
| "step": 83900 |
| }, |
| { |
| "epoch": 9.035625874502207, |
| "grad_norm": 0.9125701785087585, |
| "learning_rate": 5.813813166684624e-05, |
| "loss": 3.1139, |
| "step": 83950 |
| }, |
| { |
| "epoch": 9.041007426541814, |
| "grad_norm": 0.8349764347076416, |
| "learning_rate": 5.781489063678482e-05, |
| "loss": 3.126, |
| "step": 84000 |
| }, |
| { |
| "epoch": 9.041007426541814, |
| "eval_accuracy": 0.3922756054383378, |
| "eval_loss": 3.317336320877075, |
| "eval_runtime": 180.0088, |
| "eval_samples_per_second": 100.056, |
| "eval_steps_per_second": 6.255, |
| "step": 84000 |
| }, |
| { |
| "epoch": 9.046388978581422, |
| "grad_norm": 0.9072763323783875, |
| "learning_rate": 5.749164960672341e-05, |
| "loss": 3.1405, |
| "step": 84050 |
| }, |
| { |
| "epoch": 9.051770530621031, |
| "grad_norm": 0.9203513264656067, |
| "learning_rate": 5.716840857666199e-05, |
| "loss": 3.1331, |
| "step": 84100 |
| }, |
| { |
| "epoch": 9.05715208266064, |
| "grad_norm": 0.8729878664016724, |
| "learning_rate": 5.684516754660057e-05, |
| "loss": 3.1335, |
| "step": 84150 |
| }, |
| { |
| "epoch": 9.062533634700248, |
| "grad_norm": 0.9056902527809143, |
| "learning_rate": 5.652192651653916e-05, |
| "loss": 3.1256, |
| "step": 84200 |
| }, |
| { |
| "epoch": 9.067915186739857, |
| "grad_norm": 0.8653237223625183, |
| "learning_rate": 5.619868548647774e-05, |
| "loss": 3.1253, |
| "step": 84250 |
| }, |
| { |
| "epoch": 9.073296738779463, |
| "grad_norm": 0.8897874355316162, |
| "learning_rate": 5.5875444456416335e-05, |
| "loss": 3.1121, |
| "step": 84300 |
| }, |
| { |
| "epoch": 9.078678290819072, |
| "grad_norm": 0.873859167098999, |
| "learning_rate": 5.5552203426354916e-05, |
| "loss": 3.1191, |
| "step": 84350 |
| }, |
| { |
| "epoch": 9.08405984285868, |
| "grad_norm": 0.8796614408493042, |
| "learning_rate": 5.52289623962935e-05, |
| "loss": 3.1252, |
| "step": 84400 |
| }, |
| { |
| "epoch": 9.089441394898289, |
| "grad_norm": 0.9063519835472107, |
| "learning_rate": 5.4905721366232085e-05, |
| "loss": 3.1188, |
| "step": 84450 |
| }, |
| { |
| "epoch": 9.094822946937898, |
| "grad_norm": 0.8301732540130615, |
| "learning_rate": 5.4582480336170666e-05, |
| "loss": 3.117, |
| "step": 84500 |
| }, |
| { |
| "epoch": 9.100204498977504, |
| "grad_norm": 0.8899036049842834, |
| "learning_rate": 5.4259239306109254e-05, |
| "loss": 3.1182, |
| "step": 84550 |
| }, |
| { |
| "epoch": 9.105586051017113, |
| "grad_norm": 0.876559317111969, |
| "learning_rate": 5.3935998276047835e-05, |
| "loss": 3.1158, |
| "step": 84600 |
| }, |
| { |
| "epoch": 9.110967603056721, |
| "grad_norm": 0.8589257597923279, |
| "learning_rate": 5.3612757245986416e-05, |
| "loss": 3.1303, |
| "step": 84650 |
| }, |
| { |
| "epoch": 9.11634915509633, |
| "grad_norm": 0.9055981636047363, |
| "learning_rate": 5.3289516215925004e-05, |
| "loss": 3.113, |
| "step": 84700 |
| }, |
| { |
| "epoch": 9.121730707135939, |
| "grad_norm": 0.898881196975708, |
| "learning_rate": 5.2966275185863585e-05, |
| "loss": 3.1289, |
| "step": 84750 |
| }, |
| { |
| "epoch": 9.127112259175545, |
| "grad_norm": 0.8705615401268005, |
| "learning_rate": 5.264303415580217e-05, |
| "loss": 3.1179, |
| "step": 84800 |
| }, |
| { |
| "epoch": 9.132493811215154, |
| "grad_norm": 0.8401497006416321, |
| "learning_rate": 5.2319793125740754e-05, |
| "loss": 3.1123, |
| "step": 84850 |
| }, |
| { |
| "epoch": 9.137875363254762, |
| "grad_norm": 0.8974107503890991, |
| "learning_rate": 5.1996552095679336e-05, |
| "loss": 3.1348, |
| "step": 84900 |
| }, |
| { |
| "epoch": 9.143256915294371, |
| "grad_norm": 0.8535527586936951, |
| "learning_rate": 5.1673311065617923e-05, |
| "loss": 3.1261, |
| "step": 84950 |
| }, |
| { |
| "epoch": 9.14863846733398, |
| "grad_norm": 0.8549533486366272, |
| "learning_rate": 5.135653485615774e-05, |
| "loss": 3.1465, |
| "step": 85000 |
| }, |
| { |
| "epoch": 9.14863846733398, |
| "eval_accuracy": 0.39245216638103636, |
| "eval_loss": 3.317115545272827, |
| "eval_runtime": 179.8863, |
| "eval_samples_per_second": 100.124, |
| "eval_steps_per_second": 6.26, |
| "step": 85000 |
| }, |
| { |
| "epoch": 9.154020019373588, |
| "grad_norm": 0.8901803493499756, |
| "learning_rate": 5.103329382609632e-05, |
| "loss": 3.1413, |
| "step": 85050 |
| }, |
| { |
| "epoch": 9.159401571413195, |
| "grad_norm": 0.8346094489097595, |
| "learning_rate": 5.0710052796034906e-05, |
| "loss": 3.1247, |
| "step": 85100 |
| }, |
| { |
| "epoch": 9.164783123452803, |
| "grad_norm": 0.8812398910522461, |
| "learning_rate": 5.038681176597349e-05, |
| "loss": 3.137, |
| "step": 85150 |
| }, |
| { |
| "epoch": 9.170164675492412, |
| "grad_norm": 0.8596518635749817, |
| "learning_rate": 5.006357073591207e-05, |
| "loss": 3.1351, |
| "step": 85200 |
| }, |
| { |
| "epoch": 9.17554622753202, |
| "grad_norm": 0.8846608400344849, |
| "learning_rate": 4.974032970585066e-05, |
| "loss": 3.1143, |
| "step": 85250 |
| }, |
| { |
| "epoch": 9.180927779571629, |
| "grad_norm": 0.8773012161254883, |
| "learning_rate": 4.9417088675789244e-05, |
| "loss": 3.1483, |
| "step": 85300 |
| }, |
| { |
| "epoch": 9.186309331611236, |
| "grad_norm": 0.855322003364563, |
| "learning_rate": 4.909384764572783e-05, |
| "loss": 3.1371, |
| "step": 85350 |
| }, |
| { |
| "epoch": 9.191690883650844, |
| "grad_norm": 0.8355333209037781, |
| "learning_rate": 4.877060661566641e-05, |
| "loss": 3.1179, |
| "step": 85400 |
| }, |
| { |
| "epoch": 9.197072435690453, |
| "grad_norm": 0.8519471883773804, |
| "learning_rate": 4.8447365585604994e-05, |
| "loss": 3.1286, |
| "step": 85450 |
| }, |
| { |
| "epoch": 9.202453987730062, |
| "grad_norm": 0.8353000283241272, |
| "learning_rate": 4.812412455554358e-05, |
| "loss": 3.118, |
| "step": 85500 |
| }, |
| { |
| "epoch": 9.20783553976967, |
| "grad_norm": 0.8422732353210449, |
| "learning_rate": 4.780088352548216e-05, |
| "loss": 3.1536, |
| "step": 85550 |
| }, |
| { |
| "epoch": 9.213217091809279, |
| "grad_norm": 0.8970634937286377, |
| "learning_rate": 4.747764249542075e-05, |
| "loss": 3.1394, |
| "step": 85600 |
| }, |
| { |
| "epoch": 9.218598643848885, |
| "grad_norm": 0.8779487013816833, |
| "learning_rate": 4.715440146535933e-05, |
| "loss": 3.1236, |
| "step": 85650 |
| }, |
| { |
| "epoch": 9.223980195888494, |
| "grad_norm": 0.887233316898346, |
| "learning_rate": 4.683116043529791e-05, |
| "loss": 3.1297, |
| "step": 85700 |
| }, |
| { |
| "epoch": 9.229361747928102, |
| "grad_norm": 0.8759692907333374, |
| "learning_rate": 4.65079194052365e-05, |
| "loss": 3.1114, |
| "step": 85750 |
| }, |
| { |
| "epoch": 9.234743299967711, |
| "grad_norm": 0.8891520500183105, |
| "learning_rate": 4.618467837517508e-05, |
| "loss": 3.1221, |
| "step": 85800 |
| }, |
| { |
| "epoch": 9.24012485200732, |
| "grad_norm": 0.867538571357727, |
| "learning_rate": 4.586143734511367e-05, |
| "loss": 3.1237, |
| "step": 85850 |
| }, |
| { |
| "epoch": 9.245506404046926, |
| "grad_norm": 0.8855520486831665, |
| "learning_rate": 4.553819631505225e-05, |
| "loss": 3.1306, |
| "step": 85900 |
| }, |
| { |
| "epoch": 9.250887956086535, |
| "grad_norm": 0.8678146600723267, |
| "learning_rate": 4.521495528499083e-05, |
| "loss": 3.1267, |
| "step": 85950 |
| }, |
| { |
| "epoch": 9.256269508126143, |
| "grad_norm": 0.8817727565765381, |
| "learning_rate": 4.489171425492942e-05, |
| "loss": 3.1352, |
| "step": 86000 |
| }, |
| { |
| "epoch": 9.256269508126143, |
| "eval_accuracy": 0.39275030990519927, |
| "eval_loss": 3.3147568702697754, |
| "eval_runtime": 179.4526, |
| "eval_samples_per_second": 100.366, |
| "eval_steps_per_second": 6.275, |
| "step": 86000 |
| }, |
| { |
| "epoch": 9.261651060165752, |
| "grad_norm": 0.8901002407073975, |
| "learning_rate": 4.4568473224868e-05, |
| "loss": 3.1272, |
| "step": 86050 |
| }, |
| { |
| "epoch": 9.26703261220536, |
| "grad_norm": 0.882002592086792, |
| "learning_rate": 4.4245232194806596e-05, |
| "loss": 3.1305, |
| "step": 86100 |
| }, |
| { |
| "epoch": 9.272414164244967, |
| "grad_norm": 0.8344738483428955, |
| "learning_rate": 4.392199116474518e-05, |
| "loss": 3.1374, |
| "step": 86150 |
| }, |
| { |
| "epoch": 9.277795716284576, |
| "grad_norm": 0.8456497192382812, |
| "learning_rate": 4.359875013468376e-05, |
| "loss": 3.1337, |
| "step": 86200 |
| }, |
| { |
| "epoch": 9.283177268324184, |
| "grad_norm": 0.8751317262649536, |
| "learning_rate": 4.3275509104622346e-05, |
| "loss": 3.1329, |
| "step": 86250 |
| }, |
| { |
| "epoch": 9.288558820363793, |
| "grad_norm": 0.8995939493179321, |
| "learning_rate": 4.295226807456093e-05, |
| "loss": 3.1315, |
| "step": 86300 |
| }, |
| { |
| "epoch": 9.293940372403402, |
| "grad_norm": 0.8733546137809753, |
| "learning_rate": 4.2629027044499515e-05, |
| "loss": 3.1177, |
| "step": 86350 |
| }, |
| { |
| "epoch": 9.29932192444301, |
| "grad_norm": 0.837860643863678, |
| "learning_rate": 4.2305786014438096e-05, |
| "loss": 3.1346, |
| "step": 86400 |
| }, |
| { |
| "epoch": 9.304703476482617, |
| "grad_norm": 0.8571630120277405, |
| "learning_rate": 4.198254498437668e-05, |
| "loss": 3.1312, |
| "step": 86450 |
| }, |
| { |
| "epoch": 9.310085028522225, |
| "grad_norm": 0.8765683174133301, |
| "learning_rate": 4.1659303954315265e-05, |
| "loss": 3.1339, |
| "step": 86500 |
| }, |
| { |
| "epoch": 9.315466580561834, |
| "grad_norm": 0.8633087873458862, |
| "learning_rate": 4.1336062924253846e-05, |
| "loss": 3.146, |
| "step": 86550 |
| }, |
| { |
| "epoch": 9.320848132601443, |
| "grad_norm": 0.8428789377212524, |
| "learning_rate": 4.1012821894192434e-05, |
| "loss": 3.1403, |
| "step": 86600 |
| }, |
| { |
| "epoch": 9.326229684641051, |
| "grad_norm": 0.815321683883667, |
| "learning_rate": 4.0689580864131015e-05, |
| "loss": 3.1163, |
| "step": 86650 |
| }, |
| { |
| "epoch": 9.331611236680658, |
| "grad_norm": 0.8932182788848877, |
| "learning_rate": 4.0366339834069596e-05, |
| "loss": 3.142, |
| "step": 86700 |
| }, |
| { |
| "epoch": 9.336992788720266, |
| "grad_norm": 0.8550369143486023, |
| "learning_rate": 4.0043098804008184e-05, |
| "loss": 3.1437, |
| "step": 86750 |
| }, |
| { |
| "epoch": 9.342374340759875, |
| "grad_norm": 0.8777996301651001, |
| "learning_rate": 3.9719857773946765e-05, |
| "loss": 3.1413, |
| "step": 86800 |
| }, |
| { |
| "epoch": 9.347755892799483, |
| "grad_norm": 0.8686240315437317, |
| "learning_rate": 3.939661674388536e-05, |
| "loss": 3.1265, |
| "step": 86850 |
| }, |
| { |
| "epoch": 9.353137444839092, |
| "grad_norm": 0.8639189600944519, |
| "learning_rate": 3.9073375713823934e-05, |
| "loss": 3.1314, |
| "step": 86900 |
| }, |
| { |
| "epoch": 9.3585189968787, |
| "grad_norm": 0.8951987028121948, |
| "learning_rate": 3.8750134683762515e-05, |
| "loss": 3.1329, |
| "step": 86950 |
| }, |
| { |
| "epoch": 9.363900548918307, |
| "grad_norm": 0.8524070382118225, |
| "learning_rate": 3.842689365370111e-05, |
| "loss": 3.129, |
| "step": 87000 |
| }, |
| { |
| "epoch": 9.363900548918307, |
| "eval_accuracy": 0.39313265941741843, |
| "eval_loss": 3.3121142387390137, |
| "eval_runtime": 179.955, |
| "eval_samples_per_second": 100.086, |
| "eval_steps_per_second": 6.257, |
| "step": 87000 |
| }, |
| { |
| "epoch": 9.369282100957916, |
| "grad_norm": 0.8834142088890076, |
| "learning_rate": 3.811011744424092e-05, |
| "loss": 3.1444, |
| "step": 87050 |
| }, |
| { |
| "epoch": 9.374663652997524, |
| "grad_norm": 0.8924848437309265, |
| "learning_rate": 3.7786876414179504e-05, |
| "loss": 3.1386, |
| "step": 87100 |
| }, |
| { |
| "epoch": 9.380045205037133, |
| "grad_norm": 0.8910598158836365, |
| "learning_rate": 3.7463635384118085e-05, |
| "loss": 3.1267, |
| "step": 87150 |
| }, |
| { |
| "epoch": 9.385426757076742, |
| "grad_norm": 0.9576981067657471, |
| "learning_rate": 3.714039435405667e-05, |
| "loss": 3.1285, |
| "step": 87200 |
| }, |
| { |
| "epoch": 9.390808309116348, |
| "grad_norm": 0.8550928235054016, |
| "learning_rate": 3.6817153323995254e-05, |
| "loss": 3.1348, |
| "step": 87250 |
| }, |
| { |
| "epoch": 9.396189861155957, |
| "grad_norm": 0.8894127607345581, |
| "learning_rate": 3.649391229393384e-05, |
| "loss": 3.1176, |
| "step": 87300 |
| }, |
| { |
| "epoch": 9.401571413195565, |
| "grad_norm": 0.8873329758644104, |
| "learning_rate": 3.617067126387242e-05, |
| "loss": 3.1366, |
| "step": 87350 |
| }, |
| { |
| "epoch": 9.406952965235174, |
| "grad_norm": 0.916664719581604, |
| "learning_rate": 3.5847430233811004e-05, |
| "loss": 3.1222, |
| "step": 87400 |
| }, |
| { |
| "epoch": 9.412334517274783, |
| "grad_norm": 0.8714802265167236, |
| "learning_rate": 3.552418920374959e-05, |
| "loss": 3.1415, |
| "step": 87450 |
| }, |
| { |
| "epoch": 9.417716069314391, |
| "grad_norm": 0.8508152961730957, |
| "learning_rate": 3.520094817368818e-05, |
| "loss": 3.134, |
| "step": 87500 |
| }, |
| { |
| "epoch": 9.423097621353998, |
| "grad_norm": 0.8734935522079468, |
| "learning_rate": 3.487770714362676e-05, |
| "loss": 3.1225, |
| "step": 87550 |
| }, |
| { |
| "epoch": 9.428479173393606, |
| "grad_norm": 0.8918778300285339, |
| "learning_rate": 3.455446611356535e-05, |
| "loss": 3.1418, |
| "step": 87600 |
| }, |
| { |
| "epoch": 9.433860725433215, |
| "grad_norm": 0.848052442073822, |
| "learning_rate": 3.423122508350393e-05, |
| "loss": 3.1324, |
| "step": 87650 |
| }, |
| { |
| "epoch": 9.439242277472824, |
| "grad_norm": 0.9284157156944275, |
| "learning_rate": 3.390798405344251e-05, |
| "loss": 3.1299, |
| "step": 87700 |
| }, |
| { |
| "epoch": 9.444623829512432, |
| "grad_norm": 0.8494699597358704, |
| "learning_rate": 3.35847430233811e-05, |
| "loss": 3.1307, |
| "step": 87750 |
| }, |
| { |
| "epoch": 9.450005381552039, |
| "grad_norm": 0.8794162273406982, |
| "learning_rate": 3.326150199331968e-05, |
| "loss": 3.1361, |
| "step": 87800 |
| }, |
| { |
| "epoch": 9.455386933591647, |
| "grad_norm": 0.8918963670730591, |
| "learning_rate": 3.293826096325827e-05, |
| "loss": 3.1246, |
| "step": 87850 |
| }, |
| { |
| "epoch": 9.460768485631256, |
| "grad_norm": 0.8786701560020447, |
| "learning_rate": 3.261501993319685e-05, |
| "loss": 3.1148, |
| "step": 87900 |
| }, |
| { |
| "epoch": 9.466150037670864, |
| "grad_norm": 0.8807184100151062, |
| "learning_rate": 3.229177890313544e-05, |
| "loss": 3.135, |
| "step": 87950 |
| }, |
| { |
| "epoch": 9.471531589710473, |
| "grad_norm": 0.8800942301750183, |
| "learning_rate": 3.196853787307402e-05, |
| "loss": 3.1379, |
| "step": 88000 |
| }, |
| { |
| "epoch": 9.471531589710473, |
| "eval_accuracy": 0.39338103991896234, |
| "eval_loss": 3.3090343475341797, |
| "eval_runtime": 179.8302, |
| "eval_samples_per_second": 100.156, |
| "eval_steps_per_second": 6.261, |
| "step": 88000 |
| }, |
| { |
| "epoch": 9.476913141750082, |
| "grad_norm": 0.9230080246925354, |
| "learning_rate": 3.1645296843012606e-05, |
| "loss": 3.1339, |
| "step": 88050 |
| }, |
| { |
| "epoch": 9.482294693789688, |
| "grad_norm": 0.8794991970062256, |
| "learning_rate": 3.132205581295119e-05, |
| "loss": 3.128, |
| "step": 88100 |
| }, |
| { |
| "epoch": 9.487676245829297, |
| "grad_norm": 0.8440155386924744, |
| "learning_rate": 3.099881478288977e-05, |
| "loss": 3.1349, |
| "step": 88150 |
| }, |
| { |
| "epoch": 9.493057797868905, |
| "grad_norm": 0.8837097883224487, |
| "learning_rate": 3.0675573752828356e-05, |
| "loss": 3.1294, |
| "step": 88200 |
| }, |
| { |
| "epoch": 9.498439349908514, |
| "grad_norm": 0.8591768741607666, |
| "learning_rate": 3.035233272276694e-05, |
| "loss": 3.1204, |
| "step": 88250 |
| }, |
| { |
| "epoch": 9.503820901948123, |
| "grad_norm": 0.9089881181716919, |
| "learning_rate": 3.0029091692705525e-05, |
| "loss": 3.1381, |
| "step": 88300 |
| }, |
| { |
| "epoch": 9.50920245398773, |
| "grad_norm": 0.8778650760650635, |
| "learning_rate": 2.970585066264411e-05, |
| "loss": 3.1344, |
| "step": 88350 |
| }, |
| { |
| "epoch": 9.514584006027338, |
| "grad_norm": 0.8694146871566772, |
| "learning_rate": 2.938260963258269e-05, |
| "loss": 3.1233, |
| "step": 88400 |
| }, |
| { |
| "epoch": 9.519965558066946, |
| "grad_norm": 0.8793097734451294, |
| "learning_rate": 2.9059368602521275e-05, |
| "loss": 3.1229, |
| "step": 88450 |
| }, |
| { |
| "epoch": 9.525347110106555, |
| "grad_norm": 0.896824300289154, |
| "learning_rate": 2.8736127572459863e-05, |
| "loss": 3.1412, |
| "step": 88500 |
| }, |
| { |
| "epoch": 9.530728662146164, |
| "grad_norm": 0.8718636631965637, |
| "learning_rate": 2.8412886542398448e-05, |
| "loss": 3.1335, |
| "step": 88550 |
| }, |
| { |
| "epoch": 9.536110214185772, |
| "grad_norm": 0.8667758703231812, |
| "learning_rate": 2.8089645512337032e-05, |
| "loss": 3.13, |
| "step": 88600 |
| }, |
| { |
| "epoch": 9.541491766225379, |
| "grad_norm": 0.8904845118522644, |
| "learning_rate": 2.7766404482275613e-05, |
| "loss": 3.1138, |
| "step": 88650 |
| }, |
| { |
| "epoch": 9.546873318264987, |
| "grad_norm": 0.9374863505363464, |
| "learning_rate": 2.7443163452214198e-05, |
| "loss": 3.1137, |
| "step": 88700 |
| }, |
| { |
| "epoch": 9.552254870304596, |
| "grad_norm": 0.9184094071388245, |
| "learning_rate": 2.7119922422152782e-05, |
| "loss": 3.148, |
| "step": 88750 |
| }, |
| { |
| "epoch": 9.557636422344205, |
| "grad_norm": 0.8584219217300415, |
| "learning_rate": 2.6796681392091367e-05, |
| "loss": 3.1272, |
| "step": 88800 |
| }, |
| { |
| "epoch": 9.563017974383813, |
| "grad_norm": 0.8915684819221497, |
| "learning_rate": 2.6473440362029955e-05, |
| "loss": 3.1465, |
| "step": 88850 |
| }, |
| { |
| "epoch": 9.56839952642342, |
| "grad_norm": 0.9296562671661377, |
| "learning_rate": 2.6150199331968536e-05, |
| "loss": 3.1323, |
| "step": 88900 |
| }, |
| { |
| "epoch": 9.573781078463028, |
| "grad_norm": 0.888725757598877, |
| "learning_rate": 2.582695830190712e-05, |
| "loss": 3.134, |
| "step": 88950 |
| }, |
| { |
| "epoch": 9.579162630502637, |
| "grad_norm": 0.9244557619094849, |
| "learning_rate": 2.5503717271845705e-05, |
| "loss": 3.1373, |
| "step": 89000 |
| }, |
| { |
| "epoch": 9.579162630502637, |
| "eval_accuracy": 0.393602039892777, |
| "eval_loss": 3.307893753051758, |
| "eval_runtime": 180.0866, |
| "eval_samples_per_second": 100.013, |
| "eval_steps_per_second": 6.253, |
| "step": 89000 |
| }, |
| { |
| "epoch": 9.584544182542245, |
| "grad_norm": 0.8654752373695374, |
| "learning_rate": 2.518047624178429e-05, |
| "loss": 3.1271, |
| "step": 89050 |
| }, |
| { |
| "epoch": 9.589925734581854, |
| "grad_norm": 0.8975775837898254, |
| "learning_rate": 2.4857235211722874e-05, |
| "loss": 3.1469, |
| "step": 89100 |
| }, |
| { |
| "epoch": 9.59530728662146, |
| "grad_norm": 0.8581560254096985, |
| "learning_rate": 2.4533994181661455e-05, |
| "loss": 3.1602, |
| "step": 89150 |
| }, |
| { |
| "epoch": 9.60068883866107, |
| "grad_norm": 0.8734593391418457, |
| "learning_rate": 2.421075315160004e-05, |
| "loss": 3.121, |
| "step": 89200 |
| }, |
| { |
| "epoch": 9.606070390700678, |
| "grad_norm": 0.8599057197570801, |
| "learning_rate": 2.3887512121538624e-05, |
| "loss": 3.1476, |
| "step": 89250 |
| }, |
| { |
| "epoch": 9.611451942740286, |
| "grad_norm": 0.8804906606674194, |
| "learning_rate": 2.3564271091477212e-05, |
| "loss": 3.1037, |
| "step": 89300 |
| }, |
| { |
| "epoch": 9.616833494779895, |
| "grad_norm": 0.8924012184143066, |
| "learning_rate": 2.3241030061415796e-05, |
| "loss": 3.1147, |
| "step": 89350 |
| }, |
| { |
| "epoch": 9.622215046819504, |
| "grad_norm": 0.8790477514266968, |
| "learning_rate": 2.2917789031354377e-05, |
| "loss": 3.1474, |
| "step": 89400 |
| }, |
| { |
| "epoch": 9.62759659885911, |
| "grad_norm": 0.8938375115394592, |
| "learning_rate": 2.2594548001292962e-05, |
| "loss": 3.1409, |
| "step": 89450 |
| }, |
| { |
| "epoch": 9.632978150898719, |
| "grad_norm": 0.849459171295166, |
| "learning_rate": 2.2271306971231546e-05, |
| "loss": 3.1288, |
| "step": 89500 |
| }, |
| { |
| "epoch": 9.638359702938327, |
| "grad_norm": 0.8806934952735901, |
| "learning_rate": 2.194806594117013e-05, |
| "loss": 3.128, |
| "step": 89550 |
| }, |
| { |
| "epoch": 9.643741254977936, |
| "grad_norm": 0.8296778202056885, |
| "learning_rate": 2.1624824911108715e-05, |
| "loss": 3.1499, |
| "step": 89600 |
| }, |
| { |
| "epoch": 9.649122807017545, |
| "grad_norm": 0.8698211908340454, |
| "learning_rate": 2.1301583881047296e-05, |
| "loss": 3.131, |
| "step": 89650 |
| }, |
| { |
| "epoch": 9.654504359057151, |
| "grad_norm": 0.9121602177619934, |
| "learning_rate": 2.097834285098588e-05, |
| "loss": 3.1383, |
| "step": 89700 |
| }, |
| { |
| "epoch": 9.65988591109676, |
| "grad_norm": 0.8963391780853271, |
| "learning_rate": 2.065510182092447e-05, |
| "loss": 3.1249, |
| "step": 89750 |
| }, |
| { |
| "epoch": 9.665267463136368, |
| "grad_norm": 0.9141267538070679, |
| "learning_rate": 2.0331860790863053e-05, |
| "loss": 3.1462, |
| "step": 89800 |
| }, |
| { |
| "epoch": 9.670649015175977, |
| "grad_norm": 0.869062602519989, |
| "learning_rate": 2.0008619760801638e-05, |
| "loss": 3.1401, |
| "step": 89850 |
| }, |
| { |
| "epoch": 9.676030567215586, |
| "grad_norm": 0.897432804107666, |
| "learning_rate": 1.968537873074022e-05, |
| "loss": 3.1304, |
| "step": 89900 |
| }, |
| { |
| "epoch": 9.681412119255192, |
| "grad_norm": 0.8654459118843079, |
| "learning_rate": 1.9362137700678803e-05, |
| "loss": 3.1216, |
| "step": 89950 |
| }, |
| { |
| "epoch": 9.6867936712948, |
| "grad_norm": 0.9050065875053406, |
| "learning_rate": 1.9038896670617388e-05, |
| "loss": 3.1356, |
| "step": 90000 |
| }, |
| { |
| "epoch": 9.6867936712948, |
| "eval_accuracy": 0.39384498774993015, |
| "eval_loss": 3.3056321144104004, |
| "eval_runtime": 179.7547, |
| "eval_samples_per_second": 100.198, |
| "eval_steps_per_second": 6.264, |
| "step": 90000 |
| }, |
| { |
| "epoch": 9.69217522333441, |
| "grad_norm": 0.9278258681297302, |
| "learning_rate": 1.8715655640555972e-05, |
| "loss": 3.15, |
| "step": 90050 |
| }, |
| { |
| "epoch": 9.697556775374018, |
| "grad_norm": 0.8931335806846619, |
| "learning_rate": 1.8392414610494557e-05, |
| "loss": 3.1421, |
| "step": 90100 |
| }, |
| { |
| "epoch": 9.702938327413626, |
| "grad_norm": 0.8307539820671082, |
| "learning_rate": 1.806917358043314e-05, |
| "loss": 3.1121, |
| "step": 90150 |
| }, |
| { |
| "epoch": 9.708319879453235, |
| "grad_norm": 0.8717964887619019, |
| "learning_rate": 1.7745932550371726e-05, |
| "loss": 3.1036, |
| "step": 90200 |
| }, |
| { |
| "epoch": 9.713701431492842, |
| "grad_norm": 0.8405609726905823, |
| "learning_rate": 1.742269152031031e-05, |
| "loss": 3.1154, |
| "step": 90250 |
| }, |
| { |
| "epoch": 9.71908298353245, |
| "grad_norm": 0.8634082078933716, |
| "learning_rate": 1.7099450490248895e-05, |
| "loss": 3.1389, |
| "step": 90300 |
| }, |
| { |
| "epoch": 9.724464535572059, |
| "grad_norm": 0.905585527420044, |
| "learning_rate": 1.677620946018748e-05, |
| "loss": 3.1538, |
| "step": 90350 |
| }, |
| { |
| "epoch": 9.729846087611667, |
| "grad_norm": 0.8903700709342957, |
| "learning_rate": 1.6452968430126064e-05, |
| "loss": 3.1403, |
| "step": 90400 |
| }, |
| { |
| "epoch": 9.735227639651276, |
| "grad_norm": 0.8525581359863281, |
| "learning_rate": 1.6129727400064645e-05, |
| "loss": 3.1337, |
| "step": 90450 |
| }, |
| { |
| "epoch": 9.740609191690883, |
| "grad_norm": 0.8256161212921143, |
| "learning_rate": 1.580648637000323e-05, |
| "loss": 3.1202, |
| "step": 90500 |
| }, |
| { |
| "epoch": 9.745990743730491, |
| "grad_norm": 0.8859636783599854, |
| "learning_rate": 1.5483245339941817e-05, |
| "loss": 3.1281, |
| "step": 90550 |
| }, |
| { |
| "epoch": 9.7513722957701, |
| "grad_norm": 0.879533052444458, |
| "learning_rate": 1.5160004309880398e-05, |
| "loss": 3.1193, |
| "step": 90600 |
| }, |
| { |
| "epoch": 9.756753847809708, |
| "grad_norm": 0.8815825581550598, |
| "learning_rate": 1.4836763279818985e-05, |
| "loss": 3.148, |
| "step": 90650 |
| }, |
| { |
| "epoch": 9.762135399849317, |
| "grad_norm": 0.8480821847915649, |
| "learning_rate": 1.4513522249757567e-05, |
| "loss": 3.1323, |
| "step": 90700 |
| }, |
| { |
| "epoch": 9.767516951888926, |
| "grad_norm": 0.9032825827598572, |
| "learning_rate": 1.4190281219696152e-05, |
| "loss": 3.1242, |
| "step": 90750 |
| }, |
| { |
| "epoch": 9.772898503928532, |
| "grad_norm": 0.8777044415473938, |
| "learning_rate": 1.3867040189634736e-05, |
| "loss": 3.1248, |
| "step": 90800 |
| }, |
| { |
| "epoch": 9.77828005596814, |
| "grad_norm": 0.8846954107284546, |
| "learning_rate": 1.3543799159573321e-05, |
| "loss": 3.1226, |
| "step": 90850 |
| }, |
| { |
| "epoch": 9.78366160800775, |
| "grad_norm": 0.9044235944747925, |
| "learning_rate": 1.3220558129511905e-05, |
| "loss": 3.1235, |
| "step": 90900 |
| }, |
| { |
| "epoch": 9.789043160047358, |
| "grad_norm": 0.8546289205551147, |
| "learning_rate": 1.2897317099450488e-05, |
| "loss": 3.1282, |
| "step": 90950 |
| }, |
| { |
| "epoch": 9.794424712086967, |
| "grad_norm": 0.8554105758666992, |
| "learning_rate": 1.2574076069389073e-05, |
| "loss": 3.1494, |
| "step": 91000 |
| }, |
| { |
| "epoch": 9.794424712086967, |
| "eval_accuracy": 0.39412846313423816, |
| "eval_loss": 3.3036842346191406, |
| "eval_runtime": 179.8656, |
| "eval_samples_per_second": 100.136, |
| "eval_steps_per_second": 6.26, |
| "step": 91000 |
| }, |
| { |
| "epoch": 9.799806264126573, |
| "grad_norm": 0.8401182293891907, |
| "learning_rate": 1.2257299859928886e-05, |
| "loss": 3.1188, |
| "step": 91050 |
| }, |
| { |
| "epoch": 9.805187816166182, |
| "grad_norm": 0.8869813084602356, |
| "learning_rate": 1.193405882986747e-05, |
| "loss": 3.1306, |
| "step": 91100 |
| }, |
| { |
| "epoch": 9.81056936820579, |
| "grad_norm": 0.9146585464477539, |
| "learning_rate": 1.1610817799806053e-05, |
| "loss": 3.1238, |
| "step": 91150 |
| }, |
| { |
| "epoch": 9.815950920245399, |
| "grad_norm": 0.8608843088150024, |
| "learning_rate": 1.128757676974464e-05, |
| "loss": 3.1311, |
| "step": 91200 |
| }, |
| { |
| "epoch": 9.821332472285007, |
| "grad_norm": 0.8796128630638123, |
| "learning_rate": 1.0964335739683222e-05, |
| "loss": 3.1301, |
| "step": 91250 |
| }, |
| { |
| "epoch": 9.826714024324616, |
| "grad_norm": 0.8590572476387024, |
| "learning_rate": 1.0641094709621807e-05, |
| "loss": 3.1268, |
| "step": 91300 |
| }, |
| { |
| "epoch": 9.832095576364223, |
| "grad_norm": 0.841092586517334, |
| "learning_rate": 1.0317853679560393e-05, |
| "loss": 3.1468, |
| "step": 91350 |
| }, |
| { |
| "epoch": 9.837477128403831, |
| "grad_norm": 0.879798412322998, |
| "learning_rate": 9.994612649498976e-06, |
| "loss": 3.124, |
| "step": 91400 |
| }, |
| { |
| "epoch": 9.84285868044344, |
| "grad_norm": 0.8939313292503357, |
| "learning_rate": 9.67137161943756e-06, |
| "loss": 3.1264, |
| "step": 91450 |
| }, |
| { |
| "epoch": 9.848240232483048, |
| "grad_norm": 0.8775787949562073, |
| "learning_rate": 9.348130589376145e-06, |
| "loss": 3.1319, |
| "step": 91500 |
| }, |
| { |
| "epoch": 9.853621784522657, |
| "grad_norm": 0.8580185174942017, |
| "learning_rate": 9.024889559314727e-06, |
| "loss": 3.1292, |
| "step": 91550 |
| }, |
| { |
| "epoch": 9.859003336562264, |
| "grad_norm": 0.9018335342407227, |
| "learning_rate": 8.701648529253312e-06, |
| "loss": 3.1408, |
| "step": 91600 |
| }, |
| { |
| "epoch": 9.864384888601872, |
| "grad_norm": 0.8868499398231506, |
| "learning_rate": 8.378407499191896e-06, |
| "loss": 3.1549, |
| "step": 91650 |
| }, |
| { |
| "epoch": 9.869766440641481, |
| "grad_norm": 0.8963901400566101, |
| "learning_rate": 8.055166469130481e-06, |
| "loss": 3.1332, |
| "step": 91700 |
| }, |
| { |
| "epoch": 9.87514799268109, |
| "grad_norm": 0.8725941777229309, |
| "learning_rate": 7.731925439069065e-06, |
| "loss": 3.1298, |
| "step": 91750 |
| }, |
| { |
| "epoch": 9.880529544720698, |
| "grad_norm": 0.8279412984848022, |
| "learning_rate": 7.408684409007649e-06, |
| "loss": 3.1264, |
| "step": 91800 |
| }, |
| { |
| "epoch": 9.885911096760307, |
| "grad_norm": 0.8716703057289124, |
| "learning_rate": 7.085443378946234e-06, |
| "loss": 3.121, |
| "step": 91850 |
| }, |
| { |
| "epoch": 9.891292648799913, |
| "grad_norm": 0.880731463432312, |
| "learning_rate": 6.762202348884817e-06, |
| "loss": 3.1255, |
| "step": 91900 |
| }, |
| { |
| "epoch": 9.896674200839522, |
| "grad_norm": 0.8716903328895569, |
| "learning_rate": 6.438961318823403e-06, |
| "loss": 3.1238, |
| "step": 91950 |
| }, |
| { |
| "epoch": 9.90205575287913, |
| "grad_norm": 0.8360830545425415, |
| "learning_rate": 6.115720288761986e-06, |
| "loss": 3.1415, |
| "step": 92000 |
| }, |
| { |
| "epoch": 9.90205575287913, |
| "eval_accuracy": 0.3942502630214782, |
| "eval_loss": 3.3022382259368896, |
| "eval_runtime": 179.9303, |
| "eval_samples_per_second": 100.1, |
| "eval_steps_per_second": 6.258, |
| "step": 92000 |
| }, |
| { |
| "epoch": 9.907437304918739, |
| "grad_norm": 0.8879337906837463, |
| "learning_rate": 5.792479258700571e-06, |
| "loss": 3.1039, |
| "step": 92050 |
| }, |
| { |
| "epoch": 9.912818856958348, |
| "grad_norm": 0.8420632481575012, |
| "learning_rate": 5.469238228639154e-06, |
| "loss": 3.1355, |
| "step": 92100 |
| }, |
| { |
| "epoch": 9.918200408997954, |
| "grad_norm": 0.8572197556495667, |
| "learning_rate": 5.145997198577739e-06, |
| "loss": 3.1128, |
| "step": 92150 |
| }, |
| { |
| "epoch": 9.923581961037563, |
| "grad_norm": 0.9185923337936401, |
| "learning_rate": 4.829220989117551e-06, |
| "loss": 3.1528, |
| "step": 92200 |
| }, |
| { |
| "epoch": 9.928963513077171, |
| "grad_norm": 0.8786590695381165, |
| "learning_rate": 4.505979959056136e-06, |
| "loss": 3.123, |
| "step": 92250 |
| }, |
| { |
| "epoch": 9.93434506511678, |
| "grad_norm": 0.8364728689193726, |
| "learning_rate": 4.18273892899472e-06, |
| "loss": 3.1189, |
| "step": 92300 |
| }, |
| { |
| "epoch": 9.939726617156388, |
| "grad_norm": 0.9233676791191101, |
| "learning_rate": 3.859497898933305e-06, |
| "loss": 3.1193, |
| "step": 92350 |
| }, |
| { |
| "epoch": 9.945108169195997, |
| "grad_norm": 0.874893844127655, |
| "learning_rate": 3.5362568688718884e-06, |
| "loss": 3.1365, |
| "step": 92400 |
| }, |
| { |
| "epoch": 9.950489721235604, |
| "grad_norm": 0.8843012452125549, |
| "learning_rate": 3.2130158388104725e-06, |
| "loss": 3.1061, |
| "step": 92450 |
| }, |
| { |
| "epoch": 9.955871273275212, |
| "grad_norm": 0.9005907773971558, |
| "learning_rate": 2.889774808749057e-06, |
| "loss": 3.1091, |
| "step": 92500 |
| }, |
| { |
| "epoch": 9.961252825314821, |
| "grad_norm": 0.881787896156311, |
| "learning_rate": 2.566533778687641e-06, |
| "loss": 3.1411, |
| "step": 92550 |
| }, |
| { |
| "epoch": 9.96663437735443, |
| "grad_norm": 0.8765003681182861, |
| "learning_rate": 2.243292748626225e-06, |
| "loss": 3.1421, |
| "step": 92600 |
| }, |
| { |
| "epoch": 9.972015929394038, |
| "grad_norm": 0.8894038200378418, |
| "learning_rate": 1.9200517185648096e-06, |
| "loss": 3.1143, |
| "step": 92650 |
| }, |
| { |
| "epoch": 9.977397481433645, |
| "grad_norm": 0.9114412069320679, |
| "learning_rate": 1.596810688503394e-06, |
| "loss": 3.1172, |
| "step": 92700 |
| }, |
| { |
| "epoch": 9.982779033473253, |
| "grad_norm": 0.8864458203315735, |
| "learning_rate": 1.2735696584419782e-06, |
| "loss": 3.1232, |
| "step": 92750 |
| }, |
| { |
| "epoch": 9.988160585512862, |
| "grad_norm": 0.8886809349060059, |
| "learning_rate": 9.503286283805623e-07, |
| "loss": 3.1238, |
| "step": 92800 |
| }, |
| { |
| "epoch": 9.99354213755247, |
| "grad_norm": 0.8665906190872192, |
| "learning_rate": 6.270875983191466e-07, |
| "loss": 3.1535, |
| "step": 92850 |
| }, |
| { |
| "epoch": 9.998923689592079, |
| "grad_norm": 0.8708318471908569, |
| "learning_rate": 3.038465682577308e-07, |
| "loss": 3.1364, |
| "step": 92900 |
| }, |
| { |
| "epoch": 10.0, |
| "step": 92910, |
| "total_flos": 7.7681598529536e+17, |
| "train_loss": 3.4556805386993372, |
| "train_runtime": 79114.6094, |
| "train_samples_per_second": 37.578, |
| "train_steps_per_second": 1.174 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 92910, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 10000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 7.7681598529536e+17, |
| "train_batch_size": 32, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|