{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 4824, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0020730759264058047, "grad_norm": 0.9091788530349731, "learning_rate": 0.0001996268656716418, "loss": 2.3429805755615236, "step": 10 }, { "epoch": 0.004146151852811609, "grad_norm": 1.014137625694275, "learning_rate": 0.000199212271973466, "loss": 2.083993911743164, "step": 20 }, { "epoch": 0.006219227779217414, "grad_norm": 0.9813042283058167, "learning_rate": 0.00019879767827529022, "loss": 2.0440767288208006, "step": 30 }, { "epoch": 0.008292303705623219, "grad_norm": 0.8600534200668335, "learning_rate": 0.00019838308457711444, "loss": 2.0040666580200197, "step": 40 }, { "epoch": 0.010365379632029024, "grad_norm": 0.9713129997253418, "learning_rate": 0.00019796849087893865, "loss": 1.8833141326904297, "step": 50 }, { "epoch": 0.012438455558434827, "grad_norm": 0.8239923119544983, "learning_rate": 0.00019755389718076287, "loss": 1.9487165451049804, "step": 60 }, { "epoch": 0.014511531484840632, "grad_norm": 0.970037579536438, "learning_rate": 0.00019713930348258708, "loss": 1.8362905502319335, "step": 70 }, { "epoch": 0.016584607411246437, "grad_norm": 0.8923717141151428, "learning_rate": 0.0001967247097844113, "loss": 1.8545101165771485, "step": 80 }, { "epoch": 0.01865768333765224, "grad_norm": 1.0183212757110596, "learning_rate": 0.0001963101160862355, "loss": 1.8110658645629882, "step": 90 }, { "epoch": 0.020730759264058048, "grad_norm": 0.8115273118019104, "learning_rate": 0.0001958955223880597, "loss": 1.8285026550292969, "step": 100 }, { "epoch": 0.02280383519046385, "grad_norm": 0.9811990857124329, "learning_rate": 0.0001954809286898839, "loss": 1.8877086639404297, "step": 110 }, { "epoch": 0.024876911116869654, "grad_norm": 0.9637799859046936, "learning_rate": 0.00019506633499170815, "loss": 1.767138671875, "step": 120 }, { "epoch": 0.02694998704327546, "grad_norm": 0.8544086217880249, "learning_rate": 0.00019465174129353234, "loss": 1.8863643646240233, "step": 130 }, { "epoch": 0.029023062969681265, "grad_norm": 0.8133947253227234, "learning_rate": 0.00019423714759535655, "loss": 1.763634490966797, "step": 140 }, { "epoch": 0.031096138896087068, "grad_norm": 0.8301745057106018, "learning_rate": 0.00019382255389718077, "loss": 1.7863468170166015, "step": 150 }, { "epoch": 0.033169214822492875, "grad_norm": 1.0789601802825928, "learning_rate": 0.00019340796019900498, "loss": 1.7523063659667968, "step": 160 }, { "epoch": 0.03524229074889868, "grad_norm": 0.9485411047935486, "learning_rate": 0.0001929933665008292, "loss": 1.782190704345703, "step": 170 }, { "epoch": 0.03731536667530448, "grad_norm": 0.9296794533729553, "learning_rate": 0.0001925787728026534, "loss": 1.7600175857543945, "step": 180 }, { "epoch": 0.03938844260171029, "grad_norm": 0.8409883379936218, "learning_rate": 0.00019216417910447763, "loss": 1.7022655487060547, "step": 190 }, { "epoch": 0.041461518528116095, "grad_norm": 1.083732008934021, "learning_rate": 0.00019174958540630184, "loss": 1.798708152770996, "step": 200 }, { "epoch": 0.043534594454521895, "grad_norm": 0.9614971280097961, "learning_rate": 0.00019133499170812605, "loss": 1.755731201171875, "step": 210 }, { "epoch": 0.0456076703809277, "grad_norm": 0.8356192708015442, "learning_rate": 0.00019092039800995024, "loss": 1.7372133255004882, "step": 220 }, { "epoch": 0.04768074630733351, "grad_norm": 0.8859673142433167, "learning_rate": 0.00019050580431177448, "loss": 1.709942626953125, "step": 230 }, { "epoch": 0.04975382223373931, "grad_norm": 0.8417310118675232, "learning_rate": 0.0001900912106135987, "loss": 1.7145137786865234, "step": 240 }, { "epoch": 0.051826898160145116, "grad_norm": 1.0017660856246948, "learning_rate": 0.00018967661691542289, "loss": 1.7842830657958983, "step": 250 }, { "epoch": 0.05389997408655092, "grad_norm": 0.8389541506767273, "learning_rate": 0.0001892620232172471, "loss": 1.696818733215332, "step": 260 }, { "epoch": 0.05597305001295672, "grad_norm": 0.9801367521286011, "learning_rate": 0.00018884742951907134, "loss": 1.6806442260742187, "step": 270 }, { "epoch": 0.05804612593936253, "grad_norm": 1.0814357995986938, "learning_rate": 0.00018843283582089553, "loss": 1.7625520706176758, "step": 280 }, { "epoch": 0.060119201865768336, "grad_norm": 1.0982401371002197, "learning_rate": 0.00018801824212271974, "loss": 1.7068971633911132, "step": 290 }, { "epoch": 0.062192277792174136, "grad_norm": 0.8793812394142151, "learning_rate": 0.00018760364842454396, "loss": 1.6139469146728516, "step": 300 }, { "epoch": 0.06426535371857994, "grad_norm": 1.1414549350738525, "learning_rate": 0.00018718905472636817, "loss": 1.7397516250610352, "step": 310 }, { "epoch": 0.06633842964498575, "grad_norm": 0.8776243329048157, "learning_rate": 0.00018677446102819239, "loss": 1.7217279434204102, "step": 320 }, { "epoch": 0.06841150557139156, "grad_norm": 1.0756258964538574, "learning_rate": 0.0001863598673300166, "loss": 1.757959747314453, "step": 330 }, { "epoch": 0.07048458149779736, "grad_norm": 0.9709790349006653, "learning_rate": 0.0001859452736318408, "loss": 1.7738918304443358, "step": 340 }, { "epoch": 0.07255765742420316, "grad_norm": 0.852245569229126, "learning_rate": 0.00018553067993366503, "loss": 1.718656349182129, "step": 350 }, { "epoch": 0.07463073335060896, "grad_norm": 0.9301595091819763, "learning_rate": 0.00018511608623548924, "loss": 1.6480951309204102, "step": 360 }, { "epoch": 0.07670380927701477, "grad_norm": 0.9946972131729126, "learning_rate": 0.00018470149253731343, "loss": 1.6211790084838866, "step": 370 }, { "epoch": 0.07877688520342058, "grad_norm": 0.9088757038116455, "learning_rate": 0.00018428689883913764, "loss": 1.6915576934814454, "step": 380 }, { "epoch": 0.08084996112982638, "grad_norm": 1.0271170139312744, "learning_rate": 0.0001838723051409619, "loss": 1.7419628143310546, "step": 390 }, { "epoch": 0.08292303705623219, "grad_norm": 0.850945770740509, "learning_rate": 0.00018345771144278607, "loss": 1.7296173095703125, "step": 400 }, { "epoch": 0.08499611298263798, "grad_norm": 0.8624365925788879, "learning_rate": 0.0001830431177446103, "loss": 1.6714574813842773, "step": 410 }, { "epoch": 0.08706918890904379, "grad_norm": 0.9081392288208008, "learning_rate": 0.0001826285240464345, "loss": 1.7895374298095703, "step": 420 }, { "epoch": 0.0891422648354496, "grad_norm": 0.8685234189033508, "learning_rate": 0.00018221393034825872, "loss": 1.6459678649902343, "step": 430 }, { "epoch": 0.0912153407618554, "grad_norm": 0.9042524695396423, "learning_rate": 0.00018179933665008293, "loss": 1.652775764465332, "step": 440 }, { "epoch": 0.09328841668826121, "grad_norm": 0.9973405003547668, "learning_rate": 0.00018138474295190715, "loss": 1.7169090270996095, "step": 450 }, { "epoch": 0.09536149261466702, "grad_norm": 0.8893020749092102, "learning_rate": 0.00018097014925373136, "loss": 1.6703657150268554, "step": 460 }, { "epoch": 0.09743456854107281, "grad_norm": 0.8917216658592224, "learning_rate": 0.00018055555555555557, "loss": 1.7180200576782227, "step": 470 }, { "epoch": 0.09950764446747862, "grad_norm": 0.9904600381851196, "learning_rate": 0.0001801409618573798, "loss": 1.7480133056640625, "step": 480 }, { "epoch": 0.10158072039388442, "grad_norm": 0.8615960478782654, "learning_rate": 0.00017972636815920398, "loss": 1.6338689804077149, "step": 490 }, { "epoch": 0.10365379632029023, "grad_norm": 0.8724789619445801, "learning_rate": 0.0001793117744610282, "loss": 1.7188920974731445, "step": 500 }, { "epoch": 0.10572687224669604, "grad_norm": 0.9289916753768921, "learning_rate": 0.00017889718076285243, "loss": 1.711186408996582, "step": 510 }, { "epoch": 0.10779994817310184, "grad_norm": 0.9007911682128906, "learning_rate": 0.00017848258706467662, "loss": 1.691291046142578, "step": 520 }, { "epoch": 0.10987302409950764, "grad_norm": 0.9181972742080688, "learning_rate": 0.00017806799336650083, "loss": 1.7039215087890625, "step": 530 }, { "epoch": 0.11194610002591344, "grad_norm": 0.9199845194816589, "learning_rate": 0.00017765339966832505, "loss": 1.6865100860595703, "step": 540 }, { "epoch": 0.11401917595231925, "grad_norm": 0.986028254032135, "learning_rate": 0.00017723880597014926, "loss": 1.6208017349243165, "step": 550 }, { "epoch": 0.11609225187872506, "grad_norm": 1.0031794309616089, "learning_rate": 0.00017682421227197348, "loss": 1.7129045486450196, "step": 560 }, { "epoch": 0.11816532780513087, "grad_norm": 0.9297766089439392, "learning_rate": 0.0001764096185737977, "loss": 1.6912664413452148, "step": 570 }, { "epoch": 0.12023840373153667, "grad_norm": 0.9929037690162659, "learning_rate": 0.0001759950248756219, "loss": 1.5491928100585937, "step": 580 }, { "epoch": 0.12231147965794248, "grad_norm": 0.9651626944541931, "learning_rate": 0.00017558043117744612, "loss": 1.611497688293457, "step": 590 }, { "epoch": 0.12438455558434827, "grad_norm": 0.9021316766738892, "learning_rate": 0.00017516583747927033, "loss": 1.67208251953125, "step": 600 }, { "epoch": 0.1264576315107541, "grad_norm": 1.012869954109192, "learning_rate": 0.00017475124378109452, "loss": 1.6646055221557616, "step": 610 }, { "epoch": 0.12853070743715989, "grad_norm": 0.9233406186103821, "learning_rate": 0.00017433665008291876, "loss": 1.7033281326293945, "step": 620 }, { "epoch": 0.13060378336356568, "grad_norm": 0.9368692636489868, "learning_rate": 0.00017392205638474298, "loss": 1.705314826965332, "step": 630 }, { "epoch": 0.1326768592899715, "grad_norm": 1.06294846534729, "learning_rate": 0.00017350746268656716, "loss": 1.637823486328125, "step": 640 }, { "epoch": 0.1347499352163773, "grad_norm": 1.02590811252594, "learning_rate": 0.00017309286898839138, "loss": 1.5813873291015625, "step": 650 }, { "epoch": 0.1368230111427831, "grad_norm": 1.0086368322372437, "learning_rate": 0.00017267827529021562, "loss": 1.670191764831543, "step": 660 }, { "epoch": 0.1388960870691889, "grad_norm": 0.9091687798500061, "learning_rate": 0.0001722636815920398, "loss": 1.6596086502075196, "step": 670 }, { "epoch": 0.14096916299559473, "grad_norm": 0.9156310558319092, "learning_rate": 0.00017184908789386402, "loss": 1.636850929260254, "step": 680 }, { "epoch": 0.14304223892200052, "grad_norm": 0.920555055141449, "learning_rate": 0.00017143449419568824, "loss": 1.6622644424438477, "step": 690 }, { "epoch": 0.1451153148484063, "grad_norm": 0.8024299740791321, "learning_rate": 0.00017101990049751245, "loss": 1.6317838668823241, "step": 700 }, { "epoch": 0.14718839077481213, "grad_norm": 0.9355529546737671, "learning_rate": 0.00017060530679933667, "loss": 1.6054956436157226, "step": 710 }, { "epoch": 0.14926146670121793, "grad_norm": 0.9528997540473938, "learning_rate": 0.00017019071310116088, "loss": 1.5931829452514648, "step": 720 }, { "epoch": 0.15133454262762375, "grad_norm": 0.8880454301834106, "learning_rate": 0.00016977611940298507, "loss": 1.7470476150512695, "step": 730 }, { "epoch": 0.15340761855402954, "grad_norm": 0.9789003133773804, "learning_rate": 0.0001693615257048093, "loss": 1.5642525672912597, "step": 740 }, { "epoch": 0.15548069448043533, "grad_norm": 0.9722698330879211, "learning_rate": 0.00016894693200663352, "loss": 1.5884984016418457, "step": 750 }, { "epoch": 0.15755377040684115, "grad_norm": 0.9086152911186218, "learning_rate": 0.0001685323383084577, "loss": 1.699564552307129, "step": 760 }, { "epoch": 0.15962684633324695, "grad_norm": 0.9152286648750305, "learning_rate": 0.00016811774461028192, "loss": 1.6350950241088866, "step": 770 }, { "epoch": 0.16169992225965277, "grad_norm": 0.999038815498352, "learning_rate": 0.00016770315091210617, "loss": 1.6626720428466797, "step": 780 }, { "epoch": 0.16377299818605856, "grad_norm": 0.9354605674743652, "learning_rate": 0.00016728855721393035, "loss": 1.6083240509033203, "step": 790 }, { "epoch": 0.16584607411246438, "grad_norm": 0.9644563794136047, "learning_rate": 0.00016687396351575457, "loss": 1.584075927734375, "step": 800 }, { "epoch": 0.16791915003887017, "grad_norm": 1.0484062433242798, "learning_rate": 0.00016645936981757878, "loss": 1.5805594444274902, "step": 810 }, { "epoch": 0.16999222596527597, "grad_norm": 0.9263212084770203, "learning_rate": 0.000166044776119403, "loss": 1.5858475685119628, "step": 820 }, { "epoch": 0.1720653018916818, "grad_norm": 0.8885413408279419, "learning_rate": 0.0001656301824212272, "loss": 1.6233335494995118, "step": 830 }, { "epoch": 0.17413837781808758, "grad_norm": 1.003126621246338, "learning_rate": 0.0001652155887230514, "loss": 1.5958114624023438, "step": 840 }, { "epoch": 0.1762114537444934, "grad_norm": 0.9760786294937134, "learning_rate": 0.00016480099502487564, "loss": 1.5831708908081055, "step": 850 }, { "epoch": 0.1782845296708992, "grad_norm": 0.9633323550224304, "learning_rate": 0.00016438640132669985, "loss": 1.6611177444458007, "step": 860 }, { "epoch": 0.180357605597305, "grad_norm": 0.9381558895111084, "learning_rate": 0.00016397180762852404, "loss": 1.6489791870117188, "step": 870 }, { "epoch": 0.1824306815237108, "grad_norm": 0.9603164196014404, "learning_rate": 0.00016355721393034826, "loss": 1.5972728729248047, "step": 880 }, { "epoch": 0.1845037574501166, "grad_norm": 1.02583909034729, "learning_rate": 0.00016314262023217247, "loss": 1.5938024520874023, "step": 890 }, { "epoch": 0.18657683337652242, "grad_norm": 1.0646976232528687, "learning_rate": 0.00016272802653399668, "loss": 1.6365236282348632, "step": 900 }, { "epoch": 0.18864990930292821, "grad_norm": 1.0200999975204468, "learning_rate": 0.0001623134328358209, "loss": 1.6409852981567383, "step": 910 }, { "epoch": 0.19072298522933404, "grad_norm": 0.911371648311615, "learning_rate": 0.0001618988391376451, "loss": 1.647599220275879, "step": 920 }, { "epoch": 0.19279606115573983, "grad_norm": 1.0000650882720947, "learning_rate": 0.00016148424543946933, "loss": 1.6000936508178711, "step": 930 }, { "epoch": 0.19486913708214562, "grad_norm": 0.9260374307632446, "learning_rate": 0.00016106965174129354, "loss": 1.6430105209350585, "step": 940 }, { "epoch": 0.19694221300855144, "grad_norm": 0.9303072690963745, "learning_rate": 0.00016065505804311776, "loss": 1.650974655151367, "step": 950 }, { "epoch": 0.19901528893495724, "grad_norm": 0.9936369061470032, "learning_rate": 0.00016024046434494194, "loss": 1.6197996139526367, "step": 960 }, { "epoch": 0.20108836486136306, "grad_norm": 1.029943823814392, "learning_rate": 0.00015982587064676618, "loss": 1.6034399032592774, "step": 970 }, { "epoch": 0.20316144078776885, "grad_norm": 0.9719572067260742, "learning_rate": 0.0001594112769485904, "loss": 1.6465234756469727, "step": 980 }, { "epoch": 0.20523451671417467, "grad_norm": 0.950306236743927, "learning_rate": 0.00015899668325041459, "loss": 1.5650517463684082, "step": 990 }, { "epoch": 0.20730759264058046, "grad_norm": 0.9367968440055847, "learning_rate": 0.0001585820895522388, "loss": 1.5462970733642578, "step": 1000 }, { "epoch": 0.20938066856698626, "grad_norm": 0.9189471006393433, "learning_rate": 0.00015816749585406304, "loss": 1.5616193771362306, "step": 1010 }, { "epoch": 0.21145374449339208, "grad_norm": 0.9336299896240234, "learning_rate": 0.00015775290215588723, "loss": 1.5709722518920899, "step": 1020 }, { "epoch": 0.21352682041979787, "grad_norm": 0.842507004737854, "learning_rate": 0.00015733830845771144, "loss": 1.5571642875671388, "step": 1030 }, { "epoch": 0.2155998963462037, "grad_norm": 1.0135430097579956, "learning_rate": 0.00015692371475953566, "loss": 1.5800284385681151, "step": 1040 }, { "epoch": 0.21767297227260948, "grad_norm": 1.0689647197723389, "learning_rate": 0.00015650912106135987, "loss": 1.621286392211914, "step": 1050 }, { "epoch": 0.21974604819901528, "grad_norm": 0.9640220999717712, "learning_rate": 0.0001560945273631841, "loss": 1.6154445648193358, "step": 1060 }, { "epoch": 0.2218191241254211, "grad_norm": 1.0506129264831543, "learning_rate": 0.0001556799336650083, "loss": 1.6239688873291016, "step": 1070 }, { "epoch": 0.2238922000518269, "grad_norm": 0.9619534015655518, "learning_rate": 0.0001552653399668325, "loss": 1.6208015441894532, "step": 1080 }, { "epoch": 0.2259652759782327, "grad_norm": 0.9947149753570557, "learning_rate": 0.00015485074626865673, "loss": 1.559681224822998, "step": 1090 }, { "epoch": 0.2280383519046385, "grad_norm": 0.9842751622200012, "learning_rate": 0.00015443615257048094, "loss": 1.593973731994629, "step": 1100 }, { "epoch": 0.23011142783104432, "grad_norm": 1.0848839282989502, "learning_rate": 0.00015402155887230513, "loss": 1.6429691314697266, "step": 1110 }, { "epoch": 0.23218450375745012, "grad_norm": 0.9902291297912598, "learning_rate": 0.00015360696517412935, "loss": 1.560030174255371, "step": 1120 }, { "epoch": 0.2342575796838559, "grad_norm": 1.0478103160858154, "learning_rate": 0.0001531923714759536, "loss": 1.5744388580322266, "step": 1130 }, { "epoch": 0.23633065561026173, "grad_norm": 1.0551514625549316, "learning_rate": 0.00015277777777777777, "loss": 1.5343515396118164, "step": 1140 }, { "epoch": 0.23840373153666752, "grad_norm": 0.9508061408996582, "learning_rate": 0.000152363184079602, "loss": 1.64904842376709, "step": 1150 }, { "epoch": 0.24047680746307334, "grad_norm": 1.0218685865402222, "learning_rate": 0.0001519485903814262, "loss": 1.598756980895996, "step": 1160 }, { "epoch": 0.24254988338947914, "grad_norm": 1.0358250141143799, "learning_rate": 0.00015153399668325042, "loss": 1.5701421737670898, "step": 1170 }, { "epoch": 0.24462295931588496, "grad_norm": 0.9931305646896362, "learning_rate": 0.00015111940298507463, "loss": 1.6070541381835937, "step": 1180 }, { "epoch": 0.24669603524229075, "grad_norm": 0.9626891016960144, "learning_rate": 0.00015070480928689885, "loss": 1.4579319953918457, "step": 1190 }, { "epoch": 0.24876911116869654, "grad_norm": 0.9033750295639038, "learning_rate": 0.00015029021558872306, "loss": 1.5091530799865722, "step": 1200 }, { "epoch": 0.25084218709510236, "grad_norm": 0.9851518869400024, "learning_rate": 0.00014987562189054728, "loss": 1.5484381675720216, "step": 1210 }, { "epoch": 0.2529152630215082, "grad_norm": 1.124658226966858, "learning_rate": 0.0001494610281923715, "loss": 1.6561983108520508, "step": 1220 }, { "epoch": 0.25498833894791395, "grad_norm": 1.1919487714767456, "learning_rate": 0.00014904643449419568, "loss": 1.5920031547546387, "step": 1230 }, { "epoch": 0.25706141487431977, "grad_norm": 1.0409331321716309, "learning_rate": 0.00014863184079601992, "loss": 1.5242692947387695, "step": 1240 }, { "epoch": 0.2591344908007256, "grad_norm": 0.9168948531150818, "learning_rate": 0.00014821724709784413, "loss": 1.523965549468994, "step": 1250 }, { "epoch": 0.26120756672713136, "grad_norm": 0.9807767868041992, "learning_rate": 0.00014780265339966832, "loss": 1.555476188659668, "step": 1260 }, { "epoch": 0.2632806426535372, "grad_norm": 0.905844509601593, "learning_rate": 0.00014738805970149253, "loss": 1.526420497894287, "step": 1270 }, { "epoch": 0.265353718579943, "grad_norm": 0.9614242315292358, "learning_rate": 0.00014697346600331675, "loss": 1.5558547973632812, "step": 1280 }, { "epoch": 0.2674267945063488, "grad_norm": 0.8911798596382141, "learning_rate": 0.00014655887230514096, "loss": 1.5765740394592285, "step": 1290 }, { "epoch": 0.2694998704327546, "grad_norm": 0.9637242555618286, "learning_rate": 0.00014614427860696518, "loss": 1.5548508644104004, "step": 1300 }, { "epoch": 0.2715729463591604, "grad_norm": 1.0161306858062744, "learning_rate": 0.0001457296849087894, "loss": 1.5633068084716797, "step": 1310 }, { "epoch": 0.2736460222855662, "grad_norm": 0.9732279181480408, "learning_rate": 0.0001453150912106136, "loss": 1.5484658241271974, "step": 1320 }, { "epoch": 0.275719098211972, "grad_norm": 1.1542972326278687, "learning_rate": 0.00014490049751243782, "loss": 1.661750030517578, "step": 1330 }, { "epoch": 0.2777921741383778, "grad_norm": 0.9862635731697083, "learning_rate": 0.00014448590381426204, "loss": 1.6615022659301757, "step": 1340 }, { "epoch": 0.27986525006478363, "grad_norm": 0.9343971014022827, "learning_rate": 0.00014407131011608622, "loss": 1.5562618255615235, "step": 1350 }, { "epoch": 0.28193832599118945, "grad_norm": 0.9783313870429993, "learning_rate": 0.00014365671641791046, "loss": 1.6128976821899415, "step": 1360 }, { "epoch": 0.2840114019175952, "grad_norm": 0.9503439664840698, "learning_rate": 0.00014324212271973468, "loss": 1.6037235260009766, "step": 1370 }, { "epoch": 0.28608447784400104, "grad_norm": 1.033592700958252, "learning_rate": 0.00014282752902155887, "loss": 1.4874424934387207, "step": 1380 }, { "epoch": 0.28815755377040686, "grad_norm": 1.0737738609313965, "learning_rate": 0.00014241293532338308, "loss": 1.5970182418823242, "step": 1390 }, { "epoch": 0.2902306296968126, "grad_norm": 1.0103057622909546, "learning_rate": 0.00014199834162520732, "loss": 1.5805618286132812, "step": 1400 }, { "epoch": 0.29230370562321845, "grad_norm": 0.9714734554290771, "learning_rate": 0.0001415837479270315, "loss": 1.4544689178466796, "step": 1410 }, { "epoch": 0.29437678154962427, "grad_norm": 1.0983737707138062, "learning_rate": 0.00014116915422885572, "loss": 1.4936376571655274, "step": 1420 }, { "epoch": 0.29644985747603003, "grad_norm": 1.0537210702896118, "learning_rate": 0.00014075456053067994, "loss": 1.531502628326416, "step": 1430 }, { "epoch": 0.29852293340243585, "grad_norm": 1.0152987241744995, "learning_rate": 0.00014033996683250415, "loss": 1.5243349075317383, "step": 1440 }, { "epoch": 0.3005960093288417, "grad_norm": 0.9748072028160095, "learning_rate": 0.00013992537313432837, "loss": 1.552549457550049, "step": 1450 }, { "epoch": 0.3026690852552475, "grad_norm": 0.9957873225212097, "learning_rate": 0.00013951077943615258, "loss": 1.5652660369873046, "step": 1460 }, { "epoch": 0.30474216118165326, "grad_norm": 1.0181517601013184, "learning_rate": 0.00013909618573797677, "loss": 1.5933172225952148, "step": 1470 }, { "epoch": 0.3068152371080591, "grad_norm": 1.013455867767334, "learning_rate": 0.000138681592039801, "loss": 1.6291587829589844, "step": 1480 }, { "epoch": 0.3088883130344649, "grad_norm": 1.0690240859985352, "learning_rate": 0.00013826699834162522, "loss": 1.5185072898864747, "step": 1490 }, { "epoch": 0.31096138896087067, "grad_norm": 1.0201985836029053, "learning_rate": 0.0001378524046434494, "loss": 1.6345909118652344, "step": 1500 }, { "epoch": 0.3130344648872765, "grad_norm": 1.205460548400879, "learning_rate": 0.00013743781094527363, "loss": 1.4923598289489746, "step": 1510 }, { "epoch": 0.3151075408136823, "grad_norm": 1.0943695306777954, "learning_rate": 0.00013702321724709787, "loss": 1.551824951171875, "step": 1520 }, { "epoch": 0.31718061674008813, "grad_norm": 0.9574096202850342, "learning_rate": 0.00013660862354892205, "loss": 1.563786506652832, "step": 1530 }, { "epoch": 0.3192536926664939, "grad_norm": 1.01716947555542, "learning_rate": 0.00013619402985074627, "loss": 1.5811141014099122, "step": 1540 }, { "epoch": 0.3213267685928997, "grad_norm": 1.0344488620758057, "learning_rate": 0.00013577943615257048, "loss": 1.4693257331848144, "step": 1550 }, { "epoch": 0.32339984451930553, "grad_norm": 1.1056243181228638, "learning_rate": 0.0001353648424543947, "loss": 1.5346773147583008, "step": 1560 }, { "epoch": 0.3254729204457113, "grad_norm": 1.1026792526245117, "learning_rate": 0.0001349502487562189, "loss": 1.4362568855285645, "step": 1570 }, { "epoch": 0.3275459963721171, "grad_norm": 1.0443516969680786, "learning_rate": 0.00013453565505804313, "loss": 1.5090986251831056, "step": 1580 }, { "epoch": 0.32961907229852294, "grad_norm": 1.006152629852295, "learning_rate": 0.00013412106135986734, "loss": 1.5561832427978515, "step": 1590 }, { "epoch": 0.33169214822492876, "grad_norm": 1.0812287330627441, "learning_rate": 0.00013370646766169155, "loss": 1.5622328758239745, "step": 1600 }, { "epoch": 0.3337652241513345, "grad_norm": 1.0125651359558105, "learning_rate": 0.00013329187396351577, "loss": 1.5935446739196777, "step": 1610 }, { "epoch": 0.33583830007774035, "grad_norm": 1.0166149139404297, "learning_rate": 0.00013287728026533996, "loss": 1.5660451889038085, "step": 1620 }, { "epoch": 0.33791137600414617, "grad_norm": 1.0273667573928833, "learning_rate": 0.0001324626865671642, "loss": 1.556844139099121, "step": 1630 }, { "epoch": 0.33998445193055193, "grad_norm": 1.0821034908294678, "learning_rate": 0.0001320480928689884, "loss": 1.5019357681274415, "step": 1640 }, { "epoch": 0.34205752785695775, "grad_norm": 1.082353949546814, "learning_rate": 0.0001316334991708126, "loss": 1.5400747299194335, "step": 1650 }, { "epoch": 0.3441306037833636, "grad_norm": 1.059366226196289, "learning_rate": 0.00013121890547263681, "loss": 1.5414251327514648, "step": 1660 }, { "epoch": 0.3462036797097694, "grad_norm": 0.996767520904541, "learning_rate": 0.00013080431177446103, "loss": 1.5605695724487305, "step": 1670 }, { "epoch": 0.34827675563617516, "grad_norm": 0.9471081495285034, "learning_rate": 0.00013038971807628524, "loss": 1.550748634338379, "step": 1680 }, { "epoch": 0.350349831562581, "grad_norm": 1.0039408206939697, "learning_rate": 0.00012997512437810946, "loss": 1.5467573165893556, "step": 1690 }, { "epoch": 0.3524229074889868, "grad_norm": 1.1023355722427368, "learning_rate": 0.00012956053067993367, "loss": 1.6264640808105468, "step": 1700 }, { "epoch": 0.35449598341539257, "grad_norm": 0.98204505443573, "learning_rate": 0.00012914593698175789, "loss": 1.5399457931518554, "step": 1710 }, { "epoch": 0.3565690593417984, "grad_norm": 1.1038545370101929, "learning_rate": 0.0001287313432835821, "loss": 1.5933405876159668, "step": 1720 }, { "epoch": 0.3586421352682042, "grad_norm": 1.0126818418502808, "learning_rate": 0.00012831674958540631, "loss": 1.5504828453063966, "step": 1730 }, { "epoch": 0.36071521119461, "grad_norm": 1.2247607707977295, "learning_rate": 0.0001279021558872305, "loss": 1.5725330352783202, "step": 1740 }, { "epoch": 0.3627882871210158, "grad_norm": 1.027273416519165, "learning_rate": 0.00012748756218905474, "loss": 1.5412575721740722, "step": 1750 }, { "epoch": 0.3648613630474216, "grad_norm": 1.1408429145812988, "learning_rate": 0.00012707296849087896, "loss": 1.5795598030090332, "step": 1760 }, { "epoch": 0.36693443897382744, "grad_norm": 0.9655275344848633, "learning_rate": 0.00012665837479270315, "loss": 1.5225658416748047, "step": 1770 }, { "epoch": 0.3690075149002332, "grad_norm": 1.0792255401611328, "learning_rate": 0.00012624378109452736, "loss": 1.4800280570983886, "step": 1780 }, { "epoch": 0.371080590826639, "grad_norm": 1.1403069496154785, "learning_rate": 0.0001258291873963516, "loss": 1.5279869079589843, "step": 1790 }, { "epoch": 0.37315366675304484, "grad_norm": 0.9717866778373718, "learning_rate": 0.0001254145936981758, "loss": 1.5097025871276855, "step": 1800 }, { "epoch": 0.3752267426794506, "grad_norm": 1.2142555713653564, "learning_rate": 0.000125, "loss": 1.5517022132873535, "step": 1810 }, { "epoch": 0.37729981860585643, "grad_norm": 1.0863559246063232, "learning_rate": 0.00012458540630182422, "loss": 1.6459320068359375, "step": 1820 }, { "epoch": 0.37937289453226225, "grad_norm": 0.935858428478241, "learning_rate": 0.00012417081260364843, "loss": 1.4541780471801757, "step": 1830 }, { "epoch": 0.38144597045866807, "grad_norm": 1.0745552778244019, "learning_rate": 0.00012375621890547265, "loss": 1.5062150001525878, "step": 1840 }, { "epoch": 0.38351904638507384, "grad_norm": 0.98747718334198, "learning_rate": 0.00012334162520729686, "loss": 1.450760841369629, "step": 1850 }, { "epoch": 0.38559212231147966, "grad_norm": 1.017136573791504, "learning_rate": 0.00012292703150912105, "loss": 1.5897736549377441, "step": 1860 }, { "epoch": 0.3876651982378855, "grad_norm": 1.0018948316574097, "learning_rate": 0.0001225124378109453, "loss": 1.5147670745849608, "step": 1870 }, { "epoch": 0.38973827416429124, "grad_norm": 1.0614641904830933, "learning_rate": 0.0001220978441127695, "loss": 1.558132266998291, "step": 1880 }, { "epoch": 0.39181135009069706, "grad_norm": 1.192165732383728, "learning_rate": 0.00012168325041459369, "loss": 1.5617985725402832, "step": 1890 }, { "epoch": 0.3938844260171029, "grad_norm": 1.1147258281707764, "learning_rate": 0.00012126865671641792, "loss": 1.5168773651123046, "step": 1900 }, { "epoch": 0.3959575019435087, "grad_norm": 1.129460096359253, "learning_rate": 0.00012085406301824213, "loss": 1.5103120803833008, "step": 1910 }, { "epoch": 0.39803057786991447, "grad_norm": 1.0588513612747192, "learning_rate": 0.00012043946932006633, "loss": 1.5003737449645995, "step": 1920 }, { "epoch": 0.4001036537963203, "grad_norm": 0.9957575798034668, "learning_rate": 0.00012002487562189055, "loss": 1.5337603569030762, "step": 1930 }, { "epoch": 0.4021767297227261, "grad_norm": 0.996632993221283, "learning_rate": 0.00011961028192371478, "loss": 1.6394582748413087, "step": 1940 }, { "epoch": 0.4042498056491319, "grad_norm": 1.086541771888733, "learning_rate": 0.00011919568822553898, "loss": 1.5516023635864258, "step": 1950 }, { "epoch": 0.4063228815755377, "grad_norm": 0.9926400184631348, "learning_rate": 0.00011878109452736319, "loss": 1.5786477088928224, "step": 1960 }, { "epoch": 0.4083959575019435, "grad_norm": 1.1575102806091309, "learning_rate": 0.0001183665008291874, "loss": 1.5651155471801759, "step": 1970 }, { "epoch": 0.41046903342834934, "grad_norm": 0.964144766330719, "learning_rate": 0.0001179519071310116, "loss": 1.5103749275207519, "step": 1980 }, { "epoch": 0.4125421093547551, "grad_norm": 1.1728180646896362, "learning_rate": 0.00011753731343283582, "loss": 1.5683252334594726, "step": 1990 }, { "epoch": 0.4146151852811609, "grad_norm": 1.0341147184371948, "learning_rate": 0.00011712271973466005, "loss": 1.5328519821166993, "step": 2000 }, { "epoch": 0.41668826120756675, "grad_norm": 1.220893383026123, "learning_rate": 0.00011670812603648425, "loss": 1.569212055206299, "step": 2010 }, { "epoch": 0.4187613371339725, "grad_norm": 1.287847638130188, "learning_rate": 0.00011629353233830846, "loss": 1.4834338188171388, "step": 2020 }, { "epoch": 0.42083441306037833, "grad_norm": 0.9762222766876221, "learning_rate": 0.00011587893864013268, "loss": 1.5161984443664551, "step": 2030 }, { "epoch": 0.42290748898678415, "grad_norm": 1.0173159837722778, "learning_rate": 0.00011546434494195688, "loss": 1.531379795074463, "step": 2040 }, { "epoch": 0.42498056491319, "grad_norm": 1.1367406845092773, "learning_rate": 0.00011504975124378111, "loss": 1.4666010856628418, "step": 2050 }, { "epoch": 0.42705364083959574, "grad_norm": 1.034964919090271, "learning_rate": 0.00011463515754560532, "loss": 1.4785846710205077, "step": 2060 }, { "epoch": 0.42912671676600156, "grad_norm": 1.1348105669021606, "learning_rate": 0.00011422056384742952, "loss": 1.5278627395629882, "step": 2070 }, { "epoch": 0.4311997926924074, "grad_norm": 1.092238426208496, "learning_rate": 0.00011380597014925374, "loss": 1.54183349609375, "step": 2080 }, { "epoch": 0.43327286861881315, "grad_norm": 1.1732505559921265, "learning_rate": 0.00011339137645107795, "loss": 1.5574755668640137, "step": 2090 }, { "epoch": 0.43534594454521897, "grad_norm": 1.108764410018921, "learning_rate": 0.00011297678275290215, "loss": 1.5414604187011718, "step": 2100 }, { "epoch": 0.4374190204716248, "grad_norm": 0.9649912118911743, "learning_rate": 0.00011256218905472638, "loss": 1.4722467422485352, "step": 2110 }, { "epoch": 0.43949209639803055, "grad_norm": 1.0946617126464844, "learning_rate": 0.0001121475953565506, "loss": 1.5284259796142579, "step": 2120 }, { "epoch": 0.4415651723244364, "grad_norm": 0.9634913206100464, "learning_rate": 0.0001117330016583748, "loss": 1.517714500427246, "step": 2130 }, { "epoch": 0.4436382482508422, "grad_norm": 1.1210685968399048, "learning_rate": 0.00011131840796019901, "loss": 1.463743305206299, "step": 2140 }, { "epoch": 0.445711324177248, "grad_norm": 1.065474510192871, "learning_rate": 0.00011090381426202324, "loss": 1.5272763252258301, "step": 2150 }, { "epoch": 0.4477844001036538, "grad_norm": 1.0351336002349854, "learning_rate": 0.00011048922056384742, "loss": 1.4388788223266602, "step": 2160 }, { "epoch": 0.4498574760300596, "grad_norm": 1.0451213121414185, "learning_rate": 0.00011007462686567165, "loss": 1.495101261138916, "step": 2170 }, { "epoch": 0.4519305519564654, "grad_norm": 1.2555310726165771, "learning_rate": 0.00010966003316749587, "loss": 1.4896994590759278, "step": 2180 }, { "epoch": 0.4540036278828712, "grad_norm": 1.059262752532959, "learning_rate": 0.00010924543946932007, "loss": 1.4970458030700684, "step": 2190 }, { "epoch": 0.456076703809277, "grad_norm": 0.9911238551139832, "learning_rate": 0.00010883084577114428, "loss": 1.5557072639465332, "step": 2200 }, { "epoch": 0.4581497797356828, "grad_norm": 1.1385858058929443, "learning_rate": 0.00010841625207296851, "loss": 1.4651107788085938, "step": 2210 }, { "epoch": 0.46022285566208865, "grad_norm": 1.1184098720550537, "learning_rate": 0.0001080016583747927, "loss": 1.5804038047790527, "step": 2220 }, { "epoch": 0.4622959315884944, "grad_norm": 1.0929696559906006, "learning_rate": 0.00010758706467661693, "loss": 1.5022403717041015, "step": 2230 }, { "epoch": 0.46436900751490023, "grad_norm": 0.9907273650169373, "learning_rate": 0.00010717247097844114, "loss": 1.4441984176635743, "step": 2240 }, { "epoch": 0.46644208344130605, "grad_norm": 0.9875295758247375, "learning_rate": 0.00010675787728026534, "loss": 1.4747949600219727, "step": 2250 }, { "epoch": 0.4685151593677118, "grad_norm": 0.9860086441040039, "learning_rate": 0.00010634328358208955, "loss": 1.506369400024414, "step": 2260 }, { "epoch": 0.47058823529411764, "grad_norm": 1.272979497909546, "learning_rate": 0.00010592868988391378, "loss": 1.4900493621826172, "step": 2270 }, { "epoch": 0.47266131122052346, "grad_norm": 1.1522997617721558, "learning_rate": 0.00010551409618573797, "loss": 1.4898492813110351, "step": 2280 }, { "epoch": 0.4747343871469293, "grad_norm": 1.0061860084533691, "learning_rate": 0.0001050995024875622, "loss": 1.5466057777404785, "step": 2290 }, { "epoch": 0.47680746307333505, "grad_norm": 1.1039366722106934, "learning_rate": 0.00010468490878938641, "loss": 1.4958653450012207, "step": 2300 }, { "epoch": 0.47888053899974087, "grad_norm": 1.064855933189392, "learning_rate": 0.00010427031509121061, "loss": 1.4066000938415528, "step": 2310 }, { "epoch": 0.4809536149261467, "grad_norm": 1.2152620553970337, "learning_rate": 0.00010385572139303483, "loss": 1.532679271697998, "step": 2320 }, { "epoch": 0.48302669085255245, "grad_norm": 1.1034210920333862, "learning_rate": 0.00010344112769485906, "loss": 1.4218317985534668, "step": 2330 }, { "epoch": 0.4850997667789583, "grad_norm": 1.0778026580810547, "learning_rate": 0.00010302653399668326, "loss": 1.4759449005126952, "step": 2340 }, { "epoch": 0.4871728427053641, "grad_norm": 1.0956099033355713, "learning_rate": 0.00010261194029850747, "loss": 1.510129165649414, "step": 2350 }, { "epoch": 0.4892459186317699, "grad_norm": 1.026975393295288, "learning_rate": 0.00010219734660033168, "loss": 1.450872802734375, "step": 2360 }, { "epoch": 0.4913189945581757, "grad_norm": 1.0078274011611938, "learning_rate": 0.00010178275290215589, "loss": 1.546852207183838, "step": 2370 }, { "epoch": 0.4933920704845815, "grad_norm": 0.9396551847457886, "learning_rate": 0.0001013681592039801, "loss": 1.4956705093383789, "step": 2380 }, { "epoch": 0.4954651464109873, "grad_norm": 1.0600014925003052, "learning_rate": 0.00010095356550580433, "loss": 1.4916232109069825, "step": 2390 }, { "epoch": 0.4975382223373931, "grad_norm": 0.9873512387275696, "learning_rate": 0.00010053897180762853, "loss": 1.383552074432373, "step": 2400 }, { "epoch": 0.4996112982637989, "grad_norm": 1.2188613414764404, "learning_rate": 0.00010012437810945274, "loss": 1.529955005645752, "step": 2410 }, { "epoch": 0.5016843741902047, "grad_norm": 1.0167559385299683, "learning_rate": 9.970978441127696e-05, "loss": 1.5507872581481934, "step": 2420 }, { "epoch": 0.5037574501166106, "grad_norm": 1.1081862449645996, "learning_rate": 9.929519071310116e-05, "loss": 1.4936115264892578, "step": 2430 }, { "epoch": 0.5058305260430164, "grad_norm": 1.160117268562317, "learning_rate": 9.888059701492539e-05, "loss": 1.479538917541504, "step": 2440 }, { "epoch": 0.5079036019694221, "grad_norm": 1.0124058723449707, "learning_rate": 9.846600331674959e-05, "loss": 1.407692050933838, "step": 2450 }, { "epoch": 0.5099766778958279, "grad_norm": 0.9970583319664001, "learning_rate": 9.80514096185738e-05, "loss": 1.5096822738647462, "step": 2460 }, { "epoch": 0.5120497538222337, "grad_norm": 1.0601921081542969, "learning_rate": 9.763681592039802e-05, "loss": 1.5124183654785157, "step": 2470 }, { "epoch": 0.5141228297486395, "grad_norm": 1.2017216682434082, "learning_rate": 9.722222222222223e-05, "loss": 1.4738014221191407, "step": 2480 }, { "epoch": 0.5161959056750454, "grad_norm": 1.1182701587677002, "learning_rate": 9.680762852404643e-05, "loss": 1.4870125770568847, "step": 2490 }, { "epoch": 0.5182689816014512, "grad_norm": 0.9587567448616028, "learning_rate": 9.639303482587066e-05, "loss": 1.5147294998168945, "step": 2500 }, { "epoch": 0.520342057527857, "grad_norm": 1.177620530128479, "learning_rate": 9.597844112769486e-05, "loss": 1.519681739807129, "step": 2510 }, { "epoch": 0.5224151334542627, "grad_norm": 1.1010137796401978, "learning_rate": 9.556384742951907e-05, "loss": 1.4474271774291991, "step": 2520 }, { "epoch": 0.5244882093806685, "grad_norm": 0.9620181918144226, "learning_rate": 9.514925373134329e-05, "loss": 1.5070289611816405, "step": 2530 }, { "epoch": 0.5265612853070744, "grad_norm": 1.0856534242630005, "learning_rate": 9.47346600331675e-05, "loss": 1.5071434020996093, "step": 2540 }, { "epoch": 0.5286343612334802, "grad_norm": 1.1289268732070923, "learning_rate": 9.43200663349917e-05, "loss": 1.6005094528198243, "step": 2550 }, { "epoch": 0.530707437159886, "grad_norm": 1.063301682472229, "learning_rate": 9.390547263681593e-05, "loss": 1.490926742553711, "step": 2560 }, { "epoch": 0.5327805130862918, "grad_norm": 1.0895897150039673, "learning_rate": 9.349087893864013e-05, "loss": 1.4283631324768067, "step": 2570 }, { "epoch": 0.5348535890126976, "grad_norm": 1.0446993112564087, "learning_rate": 9.307628524046435e-05, "loss": 1.5121649742126464, "step": 2580 }, { "epoch": 0.5369266649391033, "grad_norm": 1.118340253829956, "learning_rate": 9.266169154228856e-05, "loss": 1.5203582763671875, "step": 2590 }, { "epoch": 0.5389997408655092, "grad_norm": 0.9693984389305115, "learning_rate": 9.224709784411278e-05, "loss": 1.493481731414795, "step": 2600 }, { "epoch": 0.541072816791915, "grad_norm": 1.027601957321167, "learning_rate": 9.183250414593698e-05, "loss": 1.3991009712219238, "step": 2610 }, { "epoch": 0.5431458927183208, "grad_norm": 1.0012259483337402, "learning_rate": 9.14179104477612e-05, "loss": 1.4645806312561036, "step": 2620 }, { "epoch": 0.5452189686447266, "grad_norm": 1.021893858909607, "learning_rate": 9.10033167495854e-05, "loss": 1.4260982513427733, "step": 2630 }, { "epoch": 0.5472920445711325, "grad_norm": 1.1559139490127563, "learning_rate": 9.058872305140962e-05, "loss": 1.436918354034424, "step": 2640 }, { "epoch": 0.5493651204975383, "grad_norm": 1.0691229104995728, "learning_rate": 9.017412935323383e-05, "loss": 1.5417165756225586, "step": 2650 }, { "epoch": 0.551438196423944, "grad_norm": 1.0778980255126953, "learning_rate": 8.975953565505805e-05, "loss": 1.4646015167236328, "step": 2660 }, { "epoch": 0.5535112723503498, "grad_norm": 1.128537654876709, "learning_rate": 8.934494195688225e-05, "loss": 1.4791038513183594, "step": 2670 }, { "epoch": 0.5555843482767556, "grad_norm": 1.1227152347564697, "learning_rate": 8.893034825870648e-05, "loss": 1.477587890625, "step": 2680 }, { "epoch": 0.5576574242031614, "grad_norm": 1.1616007089614868, "learning_rate": 8.851575456053068e-05, "loss": 1.5081703186035156, "step": 2690 }, { "epoch": 0.5597305001295673, "grad_norm": 1.071717381477356, "learning_rate": 8.810116086235489e-05, "loss": 1.478490161895752, "step": 2700 }, { "epoch": 0.5618035760559731, "grad_norm": 0.953491747379303, "learning_rate": 8.76865671641791e-05, "loss": 1.486307144165039, "step": 2710 }, { "epoch": 0.5638766519823789, "grad_norm": 1.107778787612915, "learning_rate": 8.727197346600332e-05, "loss": 1.4863730430603028, "step": 2720 }, { "epoch": 0.5659497279087846, "grad_norm": 1.1062201261520386, "learning_rate": 8.685737976782754e-05, "loss": 1.4807310104370117, "step": 2730 }, { "epoch": 0.5680228038351904, "grad_norm": 1.1240142583847046, "learning_rate": 8.644278606965175e-05, "loss": 1.4995529174804687, "step": 2740 }, { "epoch": 0.5700958797615963, "grad_norm": 1.0423475503921509, "learning_rate": 8.602819237147596e-05, "loss": 1.4720562934875487, "step": 2750 }, { "epoch": 0.5721689556880021, "grad_norm": 1.1449542045593262, "learning_rate": 8.561359867330017e-05, "loss": 1.5412587165832519, "step": 2760 }, { "epoch": 0.5742420316144079, "grad_norm": 1.1100146770477295, "learning_rate": 8.519900497512438e-05, "loss": 1.4932291030883789, "step": 2770 }, { "epoch": 0.5763151075408137, "grad_norm": 1.066078543663025, "learning_rate": 8.47844112769486e-05, "loss": 1.4360157012939454, "step": 2780 }, { "epoch": 0.5783881834672195, "grad_norm": 1.0685060024261475, "learning_rate": 8.436981757877281e-05, "loss": 1.5007810592651367, "step": 2790 }, { "epoch": 0.5804612593936253, "grad_norm": 1.337509274482727, "learning_rate": 8.395522388059702e-05, "loss": 1.4985486030578614, "step": 2800 }, { "epoch": 0.5825343353200311, "grad_norm": 1.0489015579223633, "learning_rate": 8.354063018242124e-05, "loss": 1.4878012657165527, "step": 2810 }, { "epoch": 0.5846074112464369, "grad_norm": 1.1977568864822388, "learning_rate": 8.312603648424544e-05, "loss": 1.4938278198242188, "step": 2820 }, { "epoch": 0.5866804871728427, "grad_norm": 1.0384935140609741, "learning_rate": 8.271144278606967e-05, "loss": 1.484629249572754, "step": 2830 }, { "epoch": 0.5887535630992485, "grad_norm": 1.0145131349563599, "learning_rate": 8.229684908789387e-05, "loss": 1.4890990257263184, "step": 2840 }, { "epoch": 0.5908266390256544, "grad_norm": 1.0456562042236328, "learning_rate": 8.188225538971808e-05, "loss": 1.4158708572387695, "step": 2850 }, { "epoch": 0.5928997149520601, "grad_norm": 1.0825212001800537, "learning_rate": 8.14676616915423e-05, "loss": 1.456554126739502, "step": 2860 }, { "epoch": 0.5949727908784659, "grad_norm": 1.0650807619094849, "learning_rate": 8.105306799336651e-05, "loss": 1.473994255065918, "step": 2870 }, { "epoch": 0.5970458668048717, "grad_norm": 1.060115098953247, "learning_rate": 8.063847429519071e-05, "loss": 1.4481081008911132, "step": 2880 }, { "epoch": 0.5991189427312775, "grad_norm": 0.9253358840942383, "learning_rate": 8.022388059701494e-05, "loss": 1.4642643928527832, "step": 2890 }, { "epoch": 0.6011920186576833, "grad_norm": 1.0225350856781006, "learning_rate": 7.980928689883914e-05, "loss": 1.3740483283996583, "step": 2900 }, { "epoch": 0.6032650945840892, "grad_norm": 1.0211741924285889, "learning_rate": 7.939469320066335e-05, "loss": 1.4753257751464843, "step": 2910 }, { "epoch": 0.605338170510495, "grad_norm": 1.0062146186828613, "learning_rate": 7.898009950248757e-05, "loss": 1.4432376861572265, "step": 2920 }, { "epoch": 0.6074112464369007, "grad_norm": 1.1036884784698486, "learning_rate": 7.856550580431178e-05, "loss": 1.3941972732543946, "step": 2930 }, { "epoch": 0.6094843223633065, "grad_norm": 1.0829540491104126, "learning_rate": 7.815091210613598e-05, "loss": 1.5212538719177247, "step": 2940 }, { "epoch": 0.6115573982897123, "grad_norm": 1.1836856603622437, "learning_rate": 7.773631840796021e-05, "loss": 1.5671442031860352, "step": 2950 }, { "epoch": 0.6136304742161182, "grad_norm": 0.9901530742645264, "learning_rate": 7.732172470978441e-05, "loss": 1.4176819801330567, "step": 2960 }, { "epoch": 0.615703550142524, "grad_norm": 1.0980446338653564, "learning_rate": 7.690713101160863e-05, "loss": 1.47709379196167, "step": 2970 }, { "epoch": 0.6177766260689298, "grad_norm": 1.0742297172546387, "learning_rate": 7.649253731343284e-05, "loss": 1.3669164657592774, "step": 2980 }, { "epoch": 0.6198497019953356, "grad_norm": 1.0716501474380493, "learning_rate": 7.607794361525706e-05, "loss": 1.5048974990844726, "step": 2990 }, { "epoch": 0.6219227779217413, "grad_norm": 1.2103937864303589, "learning_rate": 7.566334991708126e-05, "loss": 1.4739879608154296, "step": 3000 }, { "epoch": 0.6239958538481472, "grad_norm": 1.0737628936767578, "learning_rate": 7.524875621890548e-05, "loss": 1.493875503540039, "step": 3010 }, { "epoch": 0.626068929774553, "grad_norm": 1.0744723081588745, "learning_rate": 7.483416252072968e-05, "loss": 1.50384578704834, "step": 3020 }, { "epoch": 0.6281420057009588, "grad_norm": 1.070535659790039, "learning_rate": 7.44195688225539e-05, "loss": 1.4305116653442382, "step": 3030 }, { "epoch": 0.6302150816273646, "grad_norm": 1.0274264812469482, "learning_rate": 7.400497512437811e-05, "loss": 1.452561092376709, "step": 3040 }, { "epoch": 0.6322881575537704, "grad_norm": 1.1508046388626099, "learning_rate": 7.359038142620233e-05, "loss": 1.4898262023925781, "step": 3050 }, { "epoch": 0.6343612334801763, "grad_norm": 1.0203231573104858, "learning_rate": 7.317578772802653e-05, "loss": 1.490403175354004, "step": 3060 }, { "epoch": 0.636434309406582, "grad_norm": 1.3003417253494263, "learning_rate": 7.276119402985076e-05, "loss": 1.4818619728088378, "step": 3070 }, { "epoch": 0.6385073853329878, "grad_norm": 1.1255285739898682, "learning_rate": 7.234660033167496e-05, "loss": 1.5162216186523438, "step": 3080 }, { "epoch": 0.6405804612593936, "grad_norm": 1.0838392972946167, "learning_rate": 7.193200663349917e-05, "loss": 1.4879429817199707, "step": 3090 }, { "epoch": 0.6426535371857994, "grad_norm": 1.036298394203186, "learning_rate": 7.151741293532339e-05, "loss": 1.5392193794250488, "step": 3100 }, { "epoch": 0.6447266131122052, "grad_norm": 1.0307782888412476, "learning_rate": 7.11028192371476e-05, "loss": 1.4172185897827148, "step": 3110 }, { "epoch": 0.6467996890386111, "grad_norm": 1.0826961994171143, "learning_rate": 7.068822553897181e-05, "loss": 1.5217445373535157, "step": 3120 }, { "epoch": 0.6488727649650169, "grad_norm": 1.1925781965255737, "learning_rate": 7.027363184079603e-05, "loss": 1.4902685165405274, "step": 3130 }, { "epoch": 0.6509458408914226, "grad_norm": 1.1316440105438232, "learning_rate": 6.985903814262023e-05, "loss": 1.4429686546325684, "step": 3140 }, { "epoch": 0.6530189168178284, "grad_norm": 1.0991287231445312, "learning_rate": 6.944444444444444e-05, "loss": 1.5057510375976562, "step": 3150 }, { "epoch": 0.6550919927442342, "grad_norm": 0.997954249382019, "learning_rate": 6.902985074626866e-05, "loss": 1.4608537673950195, "step": 3160 }, { "epoch": 0.6571650686706401, "grad_norm": 1.3616468906402588, "learning_rate": 6.861525704809287e-05, "loss": 1.4557350158691407, "step": 3170 }, { "epoch": 0.6592381445970459, "grad_norm": 1.106767177581787, "learning_rate": 6.820066334991709e-05, "loss": 1.486179733276367, "step": 3180 }, { "epoch": 0.6613112205234517, "grad_norm": 1.0956076383590698, "learning_rate": 6.77860696517413e-05, "loss": 1.4668097496032715, "step": 3190 }, { "epoch": 0.6633842964498575, "grad_norm": 1.1206140518188477, "learning_rate": 6.737147595356552e-05, "loss": 1.4768742561340331, "step": 3200 }, { "epoch": 0.6654573723762632, "grad_norm": 0.9833910465240479, "learning_rate": 6.695688225538972e-05, "loss": 1.5204994201660156, "step": 3210 }, { "epoch": 0.667530448302669, "grad_norm": 1.1817854642868042, "learning_rate": 6.654228855721395e-05, "loss": 1.5230501174926758, "step": 3220 }, { "epoch": 0.6696035242290749, "grad_norm": 1.0882384777069092, "learning_rate": 6.612769485903815e-05, "loss": 1.3979837417602539, "step": 3230 }, { "epoch": 0.6716766001554807, "grad_norm": 1.1264276504516602, "learning_rate": 6.571310116086236e-05, "loss": 1.4921886444091796, "step": 3240 }, { "epoch": 0.6737496760818865, "grad_norm": 1.1087210178375244, "learning_rate": 6.529850746268657e-05, "loss": 1.4735729217529296, "step": 3250 }, { "epoch": 0.6758227520082923, "grad_norm": 1.0597937107086182, "learning_rate": 6.488391376451079e-05, "loss": 1.5197007179260253, "step": 3260 }, { "epoch": 0.6778958279346982, "grad_norm": 1.0482463836669922, "learning_rate": 6.446932006633499e-05, "loss": 1.465884017944336, "step": 3270 }, { "epoch": 0.6799689038611039, "grad_norm": 1.035261869430542, "learning_rate": 6.405472636815922e-05, "loss": 1.4476487159729003, "step": 3280 }, { "epoch": 0.6820419797875097, "grad_norm": 1.1454598903656006, "learning_rate": 6.364013266998342e-05, "loss": 1.453122329711914, "step": 3290 }, { "epoch": 0.6841150557139155, "grad_norm": 1.1053162813186646, "learning_rate": 6.322553897180763e-05, "loss": 1.419975471496582, "step": 3300 }, { "epoch": 0.6861881316403213, "grad_norm": 1.046738624572754, "learning_rate": 6.281094527363185e-05, "loss": 1.4719730377197267, "step": 3310 }, { "epoch": 0.6882612075667272, "grad_norm": 1.2806482315063477, "learning_rate": 6.239635157545606e-05, "loss": 1.4755671501159668, "step": 3320 }, { "epoch": 0.690334283493133, "grad_norm": 0.9969695210456848, "learning_rate": 6.198175787728026e-05, "loss": 1.4757104873657227, "step": 3330 }, { "epoch": 0.6924073594195388, "grad_norm": 1.1298997402191162, "learning_rate": 6.156716417910448e-05, "loss": 1.39326171875, "step": 3340 }, { "epoch": 0.6944804353459445, "grad_norm": 1.1446443796157837, "learning_rate": 6.115257048092869e-05, "loss": 1.518898105621338, "step": 3350 }, { "epoch": 0.6965535112723503, "grad_norm": 1.145512580871582, "learning_rate": 6.0737976782752906e-05, "loss": 1.533563995361328, "step": 3360 }, { "epoch": 0.6986265871987561, "grad_norm": 1.1053916215896606, "learning_rate": 6.032338308457711e-05, "loss": 1.5256379127502442, "step": 3370 }, { "epoch": 0.700699663125162, "grad_norm": 1.1185383796691895, "learning_rate": 5.9908789386401334e-05, "loss": 1.4893531799316406, "step": 3380 }, { "epoch": 0.7027727390515678, "grad_norm": 1.0875242948532104, "learning_rate": 5.949419568822554e-05, "loss": 1.5119455337524415, "step": 3390 }, { "epoch": 0.7048458149779736, "grad_norm": 0.9813434481620789, "learning_rate": 5.907960199004975e-05, "loss": 1.4083710670471192, "step": 3400 }, { "epoch": 0.7069188909043794, "grad_norm": 1.216704249382019, "learning_rate": 5.866500829187397e-05, "loss": 1.3952042579650878, "step": 3410 }, { "epoch": 0.7089919668307851, "grad_norm": 0.9959679841995239, "learning_rate": 5.825041459369818e-05, "loss": 1.4536603927612304, "step": 3420 }, { "epoch": 0.711065042757191, "grad_norm": 1.088835597038269, "learning_rate": 5.7835820895522386e-05, "loss": 1.5083325386047364, "step": 3430 }, { "epoch": 0.7131381186835968, "grad_norm": 1.0195202827453613, "learning_rate": 5.742122719734661e-05, "loss": 1.3630632400512694, "step": 3440 }, { "epoch": 0.7152111946100026, "grad_norm": 1.1602747440338135, "learning_rate": 5.7006633499170815e-05, "loss": 1.4151906967163086, "step": 3450 }, { "epoch": 0.7172842705364084, "grad_norm": 1.0450451374053955, "learning_rate": 5.659203980099502e-05, "loss": 1.4249334335327148, "step": 3460 }, { "epoch": 0.7193573464628142, "grad_norm": 1.100573182106018, "learning_rate": 5.6177446102819243e-05, "loss": 1.4444709777832032, "step": 3470 }, { "epoch": 0.72143042238922, "grad_norm": 1.190826654434204, "learning_rate": 5.576285240464345e-05, "loss": 1.4221847534179688, "step": 3480 }, { "epoch": 0.7235034983156258, "grad_norm": 1.1273324489593506, "learning_rate": 5.534825870646766e-05, "loss": 1.4116867065429688, "step": 3490 }, { "epoch": 0.7255765742420316, "grad_norm": 1.132863163948059, "learning_rate": 5.493366500829188e-05, "loss": 1.5136717796325683, "step": 3500 }, { "epoch": 0.7276496501684374, "grad_norm": 1.256491780281067, "learning_rate": 5.451907131011609e-05, "loss": 1.4937585830688476, "step": 3510 }, { "epoch": 0.7297227260948432, "grad_norm": 1.0635231733322144, "learning_rate": 5.4104477611940295e-05, "loss": 1.5216801643371582, "step": 3520 }, { "epoch": 0.731795802021249, "grad_norm": 1.0570056438446045, "learning_rate": 5.3689883913764516e-05, "loss": 1.4815841674804688, "step": 3530 }, { "epoch": 0.7338688779476549, "grad_norm": 1.1593701839447021, "learning_rate": 5.3275290215588724e-05, "loss": 1.4530771255493165, "step": 3540 }, { "epoch": 0.7359419538740606, "grad_norm": 1.1562939882278442, "learning_rate": 5.286069651741293e-05, "loss": 1.5589415550231933, "step": 3550 }, { "epoch": 0.7380150298004664, "grad_norm": 1.1924159526824951, "learning_rate": 5.244610281923715e-05, "loss": 1.3856592178344727, "step": 3560 }, { "epoch": 0.7400881057268722, "grad_norm": 1.2092247009277344, "learning_rate": 5.203150912106136e-05, "loss": 1.402101993560791, "step": 3570 }, { "epoch": 0.742161181653278, "grad_norm": 1.155090093612671, "learning_rate": 5.161691542288557e-05, "loss": 1.4879738807678222, "step": 3580 }, { "epoch": 0.7442342575796839, "grad_norm": 0.9951611161231995, "learning_rate": 5.120232172470979e-05, "loss": 1.463954544067383, "step": 3590 }, { "epoch": 0.7463073335060897, "grad_norm": 1.038360595703125, "learning_rate": 5.0787728026533996e-05, "loss": 1.514687728881836, "step": 3600 }, { "epoch": 0.7483804094324955, "grad_norm": 1.1276081800460815, "learning_rate": 5.0373134328358204e-05, "loss": 1.383653736114502, "step": 3610 }, { "epoch": 0.7504534853589012, "grad_norm": 1.2908998727798462, "learning_rate": 4.995854063018242e-05, "loss": 1.431437873840332, "step": 3620 }, { "epoch": 0.752526561285307, "grad_norm": 1.1836334466934204, "learning_rate": 4.954394693200663e-05, "loss": 1.495081615447998, "step": 3630 }, { "epoch": 0.7545996372117129, "grad_norm": 1.0293713808059692, "learning_rate": 4.912935323383085e-05, "loss": 1.4396793365478515, "step": 3640 }, { "epoch": 0.7566727131381187, "grad_norm": 1.1716481447219849, "learning_rate": 4.8714759535655055e-05, "loss": 1.5675559997558595, "step": 3650 }, { "epoch": 0.7587457890645245, "grad_norm": 1.090667963027954, "learning_rate": 4.830016583747927e-05, "loss": 1.4259960174560546, "step": 3660 }, { "epoch": 0.7608188649909303, "grad_norm": 1.1595834493637085, "learning_rate": 4.7885572139303484e-05, "loss": 1.506321907043457, "step": 3670 }, { "epoch": 0.7628919409173361, "grad_norm": 1.2258275747299194, "learning_rate": 4.74709784411277e-05, "loss": 1.4344300270080566, "step": 3680 }, { "epoch": 0.7649650168437419, "grad_norm": 1.059180498123169, "learning_rate": 4.705638474295191e-05, "loss": 1.4194180488586425, "step": 3690 }, { "epoch": 0.7670380927701477, "grad_norm": 1.0283405780792236, "learning_rate": 4.664179104477612e-05, "loss": 1.466190242767334, "step": 3700 }, { "epoch": 0.7691111686965535, "grad_norm": 1.199347734451294, "learning_rate": 4.6227197346600334e-05, "loss": 1.502589797973633, "step": 3710 }, { "epoch": 0.7711842446229593, "grad_norm": 1.0684884786605835, "learning_rate": 4.581260364842455e-05, "loss": 1.3968034744262696, "step": 3720 }, { "epoch": 0.7732573205493651, "grad_norm": 1.240468978881836, "learning_rate": 4.539800995024876e-05, "loss": 1.4589550971984864, "step": 3730 }, { "epoch": 0.775330396475771, "grad_norm": 1.2167885303497314, "learning_rate": 4.498341625207297e-05, "loss": 1.4414511680603028, "step": 3740 }, { "epoch": 0.7774034724021768, "grad_norm": 1.1880028247833252, "learning_rate": 4.4568822553897185e-05, "loss": 1.4229560852050782, "step": 3750 }, { "epoch": 0.7794765483285825, "grad_norm": 1.023346185684204, "learning_rate": 4.41542288557214e-05, "loss": 1.414274311065674, "step": 3760 }, { "epoch": 0.7815496242549883, "grad_norm": 1.0751137733459473, "learning_rate": 4.373963515754561e-05, "loss": 1.4005146980285645, "step": 3770 }, { "epoch": 0.7836227001813941, "grad_norm": 1.1147372722625732, "learning_rate": 4.332504145936982e-05, "loss": 1.4554076194763184, "step": 3780 }, { "epoch": 0.7856957761078, "grad_norm": 1.1774779558181763, "learning_rate": 4.2910447761194036e-05, "loss": 1.4009012222290038, "step": 3790 }, { "epoch": 0.7877688520342058, "grad_norm": 1.0493589639663696, "learning_rate": 4.249585406301824e-05, "loss": 1.4634977340698243, "step": 3800 }, { "epoch": 0.7898419279606116, "grad_norm": 1.1655833721160889, "learning_rate": 4.208126036484246e-05, "loss": 1.4797652244567872, "step": 3810 }, { "epoch": 0.7919150038870174, "grad_norm": 1.1000230312347412, "learning_rate": 4.166666666666667e-05, "loss": 1.4450549125671386, "step": 3820 }, { "epoch": 0.7939880798134231, "grad_norm": 1.0554734468460083, "learning_rate": 4.125207296849088e-05, "loss": 1.4720548629760741, "step": 3830 }, { "epoch": 0.7960611557398289, "grad_norm": 1.1296530961990356, "learning_rate": 4.0837479270315094e-05, "loss": 1.5152462005615235, "step": 3840 }, { "epoch": 0.7981342316662348, "grad_norm": 1.1453872919082642, "learning_rate": 4.042288557213931e-05, "loss": 1.3691265106201171, "step": 3850 }, { "epoch": 0.8002073075926406, "grad_norm": 1.1940151453018188, "learning_rate": 4.0008291873963516e-05, "loss": 1.3959809303283692, "step": 3860 }, { "epoch": 0.8022803835190464, "grad_norm": 1.1130040884017944, "learning_rate": 3.959369817578773e-05, "loss": 1.4111416816711426, "step": 3870 }, { "epoch": 0.8043534594454522, "grad_norm": 1.1695324182510376, "learning_rate": 3.9179104477611945e-05, "loss": 1.507277011871338, "step": 3880 }, { "epoch": 0.806426535371858, "grad_norm": 1.2983999252319336, "learning_rate": 3.876451077943615e-05, "loss": 1.3935456275939941, "step": 3890 }, { "epoch": 0.8084996112982638, "grad_norm": 1.1417698860168457, "learning_rate": 3.834991708126037e-05, "loss": 1.464134407043457, "step": 3900 }, { "epoch": 0.8105726872246696, "grad_norm": 1.0782017707824707, "learning_rate": 3.793532338308458e-05, "loss": 1.4875127792358398, "step": 3910 }, { "epoch": 0.8126457631510754, "grad_norm": 1.2534525394439697, "learning_rate": 3.752072968490879e-05, "loss": 1.5353063583374023, "step": 3920 }, { "epoch": 0.8147188390774812, "grad_norm": 1.2145850658416748, "learning_rate": 3.7106135986733e-05, "loss": 1.4504884719848632, "step": 3930 }, { "epoch": 0.816791915003887, "grad_norm": 1.1368614435195923, "learning_rate": 3.669154228855722e-05, "loss": 1.4631189346313476, "step": 3940 }, { "epoch": 0.8188649909302929, "grad_norm": 1.1054575443267822, "learning_rate": 3.6276948590381425e-05, "loss": 1.44279203414917, "step": 3950 }, { "epoch": 0.8209380668566987, "grad_norm": 1.202141523361206, "learning_rate": 3.586235489220564e-05, "loss": 1.5331624031066895, "step": 3960 }, { "epoch": 0.8230111427831044, "grad_norm": 1.2692559957504272, "learning_rate": 3.5447761194029854e-05, "loss": 1.4537653923034668, "step": 3970 }, { "epoch": 0.8250842187095102, "grad_norm": 1.2438832521438599, "learning_rate": 3.503316749585406e-05, "loss": 1.4415541648864747, "step": 3980 }, { "epoch": 0.827157294635916, "grad_norm": 1.1327502727508545, "learning_rate": 3.4618573797678276e-05, "loss": 1.4588258743286133, "step": 3990 }, { "epoch": 0.8292303705623219, "grad_norm": 1.1558505296707153, "learning_rate": 3.420398009950249e-05, "loss": 1.449343776702881, "step": 4000 }, { "epoch": 0.8313034464887277, "grad_norm": 1.146558165550232, "learning_rate": 0.00011689469320066336, "loss": 1.5187210083007812, "step": 4010 }, { "epoch": 0.8333765224151335, "grad_norm": 1.154318928718567, "learning_rate": 0.00011668739635157546, "loss": 1.4608290672302247, "step": 4020 }, { "epoch": 0.8354495983415393, "grad_norm": 1.3546490669250488, "learning_rate": 0.00011648009950248757, "loss": 1.5246877670288086, "step": 4030 }, { "epoch": 0.837522674267945, "grad_norm": 1.1111828088760376, "learning_rate": 0.00011627280265339968, "loss": 1.3747848510742187, "step": 4040 }, { "epoch": 0.8395957501943508, "grad_norm": 1.0534316301345825, "learning_rate": 0.00011606550580431177, "loss": 1.4230124473571777, "step": 4050 }, { "epoch": 0.8416688261207567, "grad_norm": 1.1041001081466675, "learning_rate": 0.00011585820895522388, "loss": 1.5086248397827149, "step": 4060 }, { "epoch": 0.8437419020471625, "grad_norm": 1.1933906078338623, "learning_rate": 0.00011565091210613599, "loss": 1.401639175415039, "step": 4070 }, { "epoch": 0.8458149779735683, "grad_norm": 1.0717347860336304, "learning_rate": 0.0001154436152570481, "loss": 1.4698259353637695, "step": 4080 }, { "epoch": 0.8478880538999741, "grad_norm": 0.955698549747467, "learning_rate": 0.00011523631840796021, "loss": 1.375072193145752, "step": 4090 }, { "epoch": 0.84996112982638, "grad_norm": 1.2389768362045288, "learning_rate": 0.00011502902155887232, "loss": 1.379574966430664, "step": 4100 }, { "epoch": 0.8520342057527857, "grad_norm": 1.278762698173523, "learning_rate": 0.0001148217247097844, "loss": 1.4193490982055663, "step": 4110 }, { "epoch": 0.8541072816791915, "grad_norm": 1.1002545356750488, "learning_rate": 0.00011461442786069652, "loss": 1.3836387634277343, "step": 4120 }, { "epoch": 0.8561803576055973, "grad_norm": 1.1513651609420776, "learning_rate": 0.00011440713101160863, "loss": 1.5280473709106446, "step": 4130 }, { "epoch": 0.8582534335320031, "grad_norm": 0.9518608450889587, "learning_rate": 0.00011419983416252074, "loss": 1.4227890968322754, "step": 4140 }, { "epoch": 0.8603265094584089, "grad_norm": 1.2304730415344238, "learning_rate": 0.00011399253731343284, "loss": 1.5353459358215331, "step": 4150 }, { "epoch": 0.8623995853848148, "grad_norm": 1.1575299501419067, "learning_rate": 0.00011378524046434495, "loss": 1.362869644165039, "step": 4160 }, { "epoch": 0.8644726613112205, "grad_norm": 1.1317075490951538, "learning_rate": 0.00011357794361525704, "loss": 1.4479413986206056, "step": 4170 }, { "epoch": 0.8665457372376263, "grad_norm": 1.2143278121948242, "learning_rate": 0.00011337064676616915, "loss": 1.4484575271606446, "step": 4180 }, { "epoch": 0.8686188131640321, "grad_norm": 1.1178550720214844, "learning_rate": 0.00011316334991708126, "loss": 1.5304969787597655, "step": 4190 }, { "epoch": 0.8706918890904379, "grad_norm": 1.2381987571716309, "learning_rate": 0.00011295605306799337, "loss": 1.398491668701172, "step": 4200 }, { "epoch": 0.8727649650168438, "grad_norm": 1.288427710533142, "learning_rate": 0.00011274875621890549, "loss": 1.481517219543457, "step": 4210 }, { "epoch": 0.8748380409432496, "grad_norm": 1.2424670457839966, "learning_rate": 0.0001125414593698176, "loss": 1.4245829582214355, "step": 4220 }, { "epoch": 0.8769111168696554, "grad_norm": 1.0825245380401611, "learning_rate": 0.00011233416252072967, "loss": 1.5052209854125977, "step": 4230 }, { "epoch": 0.8789841927960611, "grad_norm": 1.1469889879226685, "learning_rate": 0.0001121268656716418, "loss": 1.432518768310547, "step": 4240 }, { "epoch": 0.8810572687224669, "grad_norm": 1.1925941705703735, "learning_rate": 0.0001119195688225539, "loss": 1.460960865020752, "step": 4250 }, { "epoch": 0.8831303446488727, "grad_norm": 1.0974957942962646, "learning_rate": 0.00011171227197346601, "loss": 1.4701688766479493, "step": 4260 }, { "epoch": 0.8852034205752786, "grad_norm": 1.19304358959198, "learning_rate": 0.00011150497512437812, "loss": 1.4184856414794922, "step": 4270 }, { "epoch": 0.8872764965016844, "grad_norm": 1.2636064291000366, "learning_rate": 0.00011129767827529022, "loss": 1.4618330001831055, "step": 4280 }, { "epoch": 0.8893495724280902, "grad_norm": 1.1047471761703491, "learning_rate": 0.00011109038142620232, "loss": 1.4464719772338868, "step": 4290 }, { "epoch": 0.891422648354496, "grad_norm": 1.1815509796142578, "learning_rate": 0.00011088308457711442, "loss": 1.4445013999938965, "step": 4300 }, { "epoch": 0.8934957242809017, "grad_norm": 1.15771484375, "learning_rate": 0.00011067578772802653, "loss": 1.4567338943481445, "step": 4310 }, { "epoch": 0.8955688002073076, "grad_norm": 1.094064474105835, "learning_rate": 0.00011046849087893865, "loss": 1.4705141067504883, "step": 4320 }, { "epoch": 0.8976418761337134, "grad_norm": 1.1300767660140991, "learning_rate": 0.00011026119402985076, "loss": 1.3876087188720703, "step": 4330 }, { "epoch": 0.8997149520601192, "grad_norm": 1.255149483680725, "learning_rate": 0.00011005389718076287, "loss": 1.3945764541625976, "step": 4340 }, { "epoch": 0.901788027986525, "grad_norm": 1.1853265762329102, "learning_rate": 0.00010984660033167495, "loss": 1.4600879669189453, "step": 4350 }, { "epoch": 0.9038611039129308, "grad_norm": 1.3775529861450195, "learning_rate": 0.00010963930348258707, "loss": 1.3908918380737305, "step": 4360 }, { "epoch": 0.9059341798393367, "grad_norm": 1.178391695022583, "learning_rate": 0.00010943200663349918, "loss": 1.418002223968506, "step": 4370 }, { "epoch": 0.9080072557657424, "grad_norm": 1.3530340194702148, "learning_rate": 0.00010922470978441128, "loss": 1.360008716583252, "step": 4380 }, { "epoch": 0.9100803316921482, "grad_norm": 1.3142609596252441, "learning_rate": 0.00010901741293532339, "loss": 1.3430272102355958, "step": 4390 }, { "epoch": 0.912153407618554, "grad_norm": 1.1689645051956177, "learning_rate": 0.0001088101160862355, "loss": 1.5633913040161134, "step": 4400 }, { "epoch": 0.9142264835449598, "grad_norm": 1.2114495038986206, "learning_rate": 0.00010860281923714759, "loss": 1.4697650909423827, "step": 4410 }, { "epoch": 0.9162995594713657, "grad_norm": 1.2312393188476562, "learning_rate": 0.0001083955223880597, "loss": 1.4190072059631347, "step": 4420 }, { "epoch": 0.9183726353977715, "grad_norm": 1.3181791305541992, "learning_rate": 0.0001081882255389718, "loss": 1.3760835647583007, "step": 4430 }, { "epoch": 0.9204457113241773, "grad_norm": 1.1612746715545654, "learning_rate": 0.00010798092868988393, "loss": 1.4294224739074708, "step": 4440 }, { "epoch": 0.922518787250583, "grad_norm": 1.0622650384902954, "learning_rate": 0.00010777363184079603, "loss": 1.3979280471801758, "step": 4450 }, { "epoch": 0.9245918631769888, "grad_norm": 1.1344703435897827, "learning_rate": 0.00010756633499170814, "loss": 1.432724380493164, "step": 4460 }, { "epoch": 0.9266649391033946, "grad_norm": 1.3609569072723389, "learning_rate": 0.00010735903814262023, "loss": 1.4266067504882813, "step": 4470 }, { "epoch": 0.9287380150298005, "grad_norm": 1.2317551374435425, "learning_rate": 0.00010715174129353234, "loss": 1.3972136497497558, "step": 4480 }, { "epoch": 0.9308110909562063, "grad_norm": 1.1485719680786133, "learning_rate": 0.00010694444444444445, "loss": 1.4437122344970703, "step": 4490 }, { "epoch": 0.9328841668826121, "grad_norm": 1.1265562772750854, "learning_rate": 0.00010673714759535656, "loss": 1.3951814651489258, "step": 4500 }, { "epoch": 0.9349572428090179, "grad_norm": 1.2382943630218506, "learning_rate": 0.00010652985074626866, "loss": 1.47876033782959, "step": 4510 }, { "epoch": 0.9370303187354236, "grad_norm": 1.1904460191726685, "learning_rate": 0.00010632255389718078, "loss": 1.5116327285766602, "step": 4520 }, { "epoch": 0.9391033946618295, "grad_norm": 1.2784380912780762, "learning_rate": 0.00010611525704809286, "loss": 1.4328707695007323, "step": 4530 }, { "epoch": 0.9411764705882353, "grad_norm": 1.1920058727264404, "learning_rate": 0.00010590796019900497, "loss": 1.4105209350585937, "step": 4540 }, { "epoch": 0.9432495465146411, "grad_norm": 1.3006956577301025, "learning_rate": 0.00010570066334991708, "loss": 1.4261651039123535, "step": 4550 }, { "epoch": 0.9453226224410469, "grad_norm": 1.1092578172683716, "learning_rate": 0.0001054933665008292, "loss": 1.3674508094787599, "step": 4560 }, { "epoch": 0.9473956983674527, "grad_norm": 1.2003893852233887, "learning_rate": 0.0001052860696517413, "loss": 1.411275577545166, "step": 4570 }, { "epoch": 0.9494687742938586, "grad_norm": 1.1365995407104492, "learning_rate": 0.00010507877280265341, "loss": 1.4263607025146485, "step": 4580 }, { "epoch": 0.9515418502202643, "grad_norm": 1.1568050384521484, "learning_rate": 0.0001048714759535655, "loss": 1.4921070098876954, "step": 4590 }, { "epoch": 0.9536149261466701, "grad_norm": 1.249493956565857, "learning_rate": 0.00010466417910447761, "loss": 1.4664103507995605, "step": 4600 }, { "epoch": 0.9556880020730759, "grad_norm": 1.1990628242492676, "learning_rate": 0.00010445688225538972, "loss": 1.4022604942321777, "step": 4610 }, { "epoch": 0.9577610779994817, "grad_norm": 1.198508858680725, "learning_rate": 0.00010424958540630183, "loss": 1.3575029373168945, "step": 4620 }, { "epoch": 0.9598341539258876, "grad_norm": 1.4220664501190186, "learning_rate": 0.00010404228855721393, "loss": 1.506014347076416, "step": 4630 }, { "epoch": 0.9619072298522934, "grad_norm": 1.1849695444107056, "learning_rate": 0.00010383499170812606, "loss": 1.3853163719177246, "step": 4640 }, { "epoch": 0.9639803057786992, "grad_norm": 1.183779001235962, "learning_rate": 0.00010362769485903814, "loss": 1.378060531616211, "step": 4650 }, { "epoch": 0.9660533817051049, "grad_norm": 1.159185767173767, "learning_rate": 0.00010342039800995024, "loss": 1.5042824745178223, "step": 4660 }, { "epoch": 0.9681264576315107, "grad_norm": 1.225111484527588, "learning_rate": 0.00010321310116086236, "loss": 1.4617128372192383, "step": 4670 }, { "epoch": 0.9701995335579165, "grad_norm": 1.1813310384750366, "learning_rate": 0.00010300580431177447, "loss": 1.4024372100830078, "step": 4680 }, { "epoch": 0.9722726094843224, "grad_norm": 1.261643648147583, "learning_rate": 0.00010279850746268658, "loss": 1.360646915435791, "step": 4690 }, { "epoch": 0.9743456854107282, "grad_norm": 1.0765501260757446, "learning_rate": 0.00010259121061359869, "loss": 1.3550106048583985, "step": 4700 }, { "epoch": 0.976418761337134, "grad_norm": 1.1051616668701172, "learning_rate": 0.00010238391376451078, "loss": 1.395486831665039, "step": 4710 }, { "epoch": 0.9784918372635398, "grad_norm": 1.2162697315216064, "learning_rate": 0.00010217661691542289, "loss": 1.4161664962768554, "step": 4720 }, { "epoch": 0.9805649131899455, "grad_norm": 1.2934092283248901, "learning_rate": 0.000101969320066335, "loss": 1.4680258750915527, "step": 4730 }, { "epoch": 0.9826379891163514, "grad_norm": 1.1861159801483154, "learning_rate": 0.0001017620232172471, "loss": 1.4943526268005372, "step": 4740 }, { "epoch": 0.9847110650427572, "grad_norm": 1.2565635442733765, "learning_rate": 0.00010155472636815921, "loss": 1.4541117668151855, "step": 4750 }, { "epoch": 0.986784140969163, "grad_norm": 1.38951575756073, "learning_rate": 0.00010134742951907133, "loss": 1.3703868865966797, "step": 4760 }, { "epoch": 0.9888572168955688, "grad_norm": 1.1929153203964233, "learning_rate": 0.00010114013266998341, "loss": 1.4121877670288085, "step": 4770 }, { "epoch": 0.9909302928219746, "grad_norm": 1.1389548778533936, "learning_rate": 0.00010093283582089552, "loss": 1.371178436279297, "step": 4780 }, { "epoch": 0.9930033687483804, "grad_norm": 0.9990752339363098, "learning_rate": 0.00010072553897180764, "loss": 1.394187831878662, "step": 4790 }, { "epoch": 0.9950764446747862, "grad_norm": 1.1418439149856567, "learning_rate": 0.00010051824212271974, "loss": 1.3987725257873536, "step": 4800 }, { "epoch": 0.9952837522674268, "grad_norm": 1.3999691009521484, "learning_rate": 0.00010049751243781096, "loss": 1.56894850730896, "step": 4801 }, { "epoch": 0.9954910598600674, "grad_norm": 1.1055763959884644, "learning_rate": 9.535655058043119e-07, "loss": 1.3579680919647217, "step": 4802 }, { "epoch": 0.995698367452708, "grad_norm": 1.2521191835403442, "learning_rate": 9.121061359867331e-07, "loss": 1.6727182865142822, "step": 4803 }, { "epoch": 0.9959056750453485, "grad_norm": 1.1679822206497192, "learning_rate": 8.706467661691543e-07, "loss": 1.5550488233566284, "step": 4804 }, { "epoch": 0.9961129826379891, "grad_norm": 1.414447546005249, "learning_rate": 8.291873963515756e-07, "loss": 1.2959469556808472, "step": 4805 }, { "epoch": 0.9963202902306297, "grad_norm": 1.2053686380386353, "learning_rate": 7.877280265339968e-07, "loss": 1.6059751510620117, "step": 4806 }, { "epoch": 0.9965275978232703, "grad_norm": 1.3234610557556152, "learning_rate": 7.462686567164179e-07, "loss": 1.3246705532073975, "step": 4807 }, { "epoch": 0.9967349054159108, "grad_norm": 1.275701880455017, "learning_rate": 7.048092868988392e-07, "loss": 1.456787109375, "step": 4808 }, { "epoch": 0.9969422130085515, "grad_norm": 1.0960595607757568, "learning_rate": 6.633499170812604e-07, "loss": 1.6712136268615723, "step": 4809 }, { "epoch": 0.997149520601192, "grad_norm": 1.1167244911193848, "learning_rate": 6.218905472636816e-07, "loss": 1.2081254720687866, "step": 4810 }, { "epoch": 0.9973568281938326, "grad_norm": 1.1857860088348389, "learning_rate": 5.804311774461028e-07, "loss": 1.4792537689208984, "step": 4811 }, { "epoch": 0.9975641357864732, "grad_norm": 1.1021283864974976, "learning_rate": 5.38971807628524e-07, "loss": 1.2333704233169556, "step": 4812 }, { "epoch": 0.9977714433791137, "grad_norm": 1.0850183963775635, "learning_rate": 4.975124378109453e-07, "loss": 1.3954815864562988, "step": 4813 }, { "epoch": 0.9979787509717544, "grad_norm": 1.2729003429412842, "learning_rate": 4.5605306799336654e-07, "loss": 1.5298696756362915, "step": 4814 }, { "epoch": 0.9981860585643949, "grad_norm": 1.2735594511032104, "learning_rate": 4.145936981757878e-07, "loss": 1.3652870655059814, "step": 4815 }, { "epoch": 0.9983933661570356, "grad_norm": 1.3938478231430054, "learning_rate": 3.7313432835820895e-07, "loss": 1.5470430850982666, "step": 4816 }, { "epoch": 0.9986006737496761, "grad_norm": 1.2034921646118164, "learning_rate": 3.316749585406302e-07, "loss": 1.539093017578125, "step": 4817 }, { "epoch": 0.9988079813423166, "grad_norm": 1.165834903717041, "learning_rate": 2.902155887230514e-07, "loss": 1.3995617628097534, "step": 4818 }, { "epoch": 0.9990152889349573, "grad_norm": 1.1757153272628784, "learning_rate": 2.4875621890547267e-07, "loss": 1.4894084930419922, "step": 4819 }, { "epoch": 0.9992225965275978, "grad_norm": 1.3626255989074707, "learning_rate": 2.072968490878939e-07, "loss": 1.5547418594360352, "step": 4820 }, { "epoch": 0.9994299041202384, "grad_norm": 1.1462852954864502, "learning_rate": 1.658374792703151e-07, "loss": 1.5449320077896118, "step": 4821 }, { "epoch": 0.999637211712879, "grad_norm": 1.122637152671814, "learning_rate": 1.2437810945273633e-07, "loss": 1.3381319046020508, "step": 4822 }, { "epoch": 0.9998445193055195, "grad_norm": 1.303135871887207, "learning_rate": 8.291873963515755e-08, "loss": 1.4987614154815674, "step": 4823 }, { "epoch": 1.0, "grad_norm": 1.30817711353302, "learning_rate": 4.1459369817578775e-08, "loss": 1.4066091775894165, "step": 4824 } ], "logging_steps": 1, "max_steps": 4824, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.698521011117097e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }