| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 8.889877641824249, | |
| "eval_steps": 500, | |
| "global_step": 2000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "entropy": 0.8774386405944824, | |
| "epoch": 0.04449388209121246, | |
| "grad_norm": 0.12704895436763763, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2824, | |
| "mean_token_accuracy": 0.6995065063238144, | |
| "num_tokens": 327680.0, | |
| "step": 10 | |
| }, | |
| { | |
| "entropy": 0.990321047604084, | |
| "epoch": 0.08898776418242492, | |
| "grad_norm": 0.06498447060585022, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9436, | |
| "mean_token_accuracy": 0.7361853331327438, | |
| "num_tokens": 655360.0, | |
| "step": 20 | |
| }, | |
| { | |
| "entropy": 0.8516890451312065, | |
| "epoch": 0.13348164627363737, | |
| "grad_norm": 0.08354150503873825, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8984, | |
| "mean_token_accuracy": 0.7455797985196113, | |
| "num_tokens": 983040.0, | |
| "step": 30 | |
| }, | |
| { | |
| "entropy": 0.8647283613681793, | |
| "epoch": 0.17797552836484984, | |
| "grad_norm": 0.06878677755594254, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8596, | |
| "mean_token_accuracy": 0.7515453889966011, | |
| "num_tokens": 1310620.0, | |
| "step": 40 | |
| }, | |
| { | |
| "entropy": 0.8749317795038223, | |
| "epoch": 0.22246941045606228, | |
| "grad_norm": 0.12942193448543549, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8933, | |
| "mean_token_accuracy": 0.7529007002711297, | |
| "num_tokens": 1623688.0, | |
| "step": 50 | |
| }, | |
| { | |
| "entropy": 0.8274541616439819, | |
| "epoch": 0.26696329254727474, | |
| "grad_norm": 0.0724002867937088, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8346, | |
| "mean_token_accuracy": 0.7569744408130645, | |
| "num_tokens": 1951368.0, | |
| "step": 60 | |
| }, | |
| { | |
| "entropy": 0.8037244468927384, | |
| "epoch": 0.3114571746384872, | |
| "grad_norm": 0.06898455321788788, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7998, | |
| "mean_token_accuracy": 0.765008519589901, | |
| "num_tokens": 2279048.0, | |
| "step": 70 | |
| }, | |
| { | |
| "entropy": 0.7850529387593269, | |
| "epoch": 0.3559510567296997, | |
| "grad_norm": 0.06743492931127548, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7893, | |
| "mean_token_accuracy": 0.7665843173861504, | |
| "num_tokens": 2606728.0, | |
| "step": 80 | |
| }, | |
| { | |
| "entropy": 0.7824164897203445, | |
| "epoch": 0.40044493882091214, | |
| "grad_norm": 0.0739995464682579, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7853, | |
| "mean_token_accuracy": 0.7667212918400764, | |
| "num_tokens": 2934287.0, | |
| "step": 90 | |
| }, | |
| { | |
| "entropy": 0.8248968973755837, | |
| "epoch": 0.44493882091212456, | |
| "grad_norm": 0.07532794028520584, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8328, | |
| "mean_token_accuracy": 0.7640984863042831, | |
| "num_tokens": 3253000.0, | |
| "step": 100 | |
| }, | |
| { | |
| "entropy": 0.7598924160003662, | |
| "epoch": 0.489432703003337, | |
| "grad_norm": 0.07987581938505173, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7672, | |
| "mean_token_accuracy": 0.7705939218401909, | |
| "num_tokens": 3580680.0, | |
| "step": 110 | |
| }, | |
| { | |
| "entropy": 0.7683228522539138, | |
| "epoch": 0.5339265850945495, | |
| "grad_norm": 0.0910281166434288, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7702, | |
| "mean_token_accuracy": 0.7697535201907157, | |
| "num_tokens": 3908360.0, | |
| "step": 120 | |
| }, | |
| { | |
| "entropy": 0.7364083752036095, | |
| "epoch": 0.578420467185762, | |
| "grad_norm": 0.0787215381860733, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7448, | |
| "mean_token_accuracy": 0.775777031481266, | |
| "num_tokens": 4236040.0, | |
| "step": 130 | |
| }, | |
| { | |
| "entropy": 0.7472490265965461, | |
| "epoch": 0.6229143492769744, | |
| "grad_norm": 0.08141525834798813, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7503, | |
| "mean_token_accuracy": 0.7719789668917656, | |
| "num_tokens": 4563587.0, | |
| "step": 140 | |
| }, | |
| { | |
| "entropy": 0.7923195600509644, | |
| "epoch": 0.6674082313681868, | |
| "grad_norm": 0.1027020588517189, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7961, | |
| "mean_token_accuracy": 0.7704089671373368, | |
| "num_tokens": 4884278.0, | |
| "step": 150 | |
| }, | |
| { | |
| "entropy": 0.7354328155517578, | |
| "epoch": 0.7119021134593994, | |
| "grad_norm": 0.08376702666282654, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7421, | |
| "mean_token_accuracy": 0.7752418130636215, | |
| "num_tokens": 5211958.0, | |
| "step": 160 | |
| }, | |
| { | |
| "entropy": 0.7320004045963288, | |
| "epoch": 0.7563959955506118, | |
| "grad_norm": 0.07954639941453934, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7347, | |
| "mean_token_accuracy": 0.7771021112799644, | |
| "num_tokens": 5539638.0, | |
| "step": 170 | |
| }, | |
| { | |
| "entropy": 0.726919624209404, | |
| "epoch": 0.8008898776418243, | |
| "grad_norm": 0.08581581711769104, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7283, | |
| "mean_token_accuracy": 0.7800618261098862, | |
| "num_tokens": 5867318.0, | |
| "step": 180 | |
| }, | |
| { | |
| "entropy": 0.7299349635839463, | |
| "epoch": 0.8453837597330367, | |
| "grad_norm": 0.08491721004247665, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7291, | |
| "mean_token_accuracy": 0.778104068338871, | |
| "num_tokens": 6194899.0, | |
| "step": 190 | |
| }, | |
| { | |
| "entropy": 0.7759820625185967, | |
| "epoch": 0.8898776418242491, | |
| "grad_norm": 0.08337808400392532, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7852, | |
| "mean_token_accuracy": 0.772493363916874, | |
| "num_tokens": 6517162.0, | |
| "step": 200 | |
| }, | |
| { | |
| "entropy": 0.7255487650632858, | |
| "epoch": 0.9343715239154616, | |
| "grad_norm": 0.08438117802143097, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7265, | |
| "mean_token_accuracy": 0.7796106413006783, | |
| "num_tokens": 6844842.0, | |
| "step": 210 | |
| }, | |
| { | |
| "entropy": 0.719422459602356, | |
| "epoch": 0.978865406006674, | |
| "grad_norm": 0.08568471670150757, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7188, | |
| "mean_token_accuracy": 0.7807851985096932, | |
| "num_tokens": 7172484.0, | |
| "step": 220 | |
| }, | |
| { | |
| "entropy": 0.7186581477140769, | |
| "epoch": 1.0222469410456063, | |
| "grad_norm": 0.08594680577516556, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7087, | |
| "mean_token_accuracy": 0.7901185231331067, | |
| "num_tokens": 7484161.0, | |
| "step": 230 | |
| }, | |
| { | |
| "entropy": 0.6627673536539078, | |
| "epoch": 1.0667408231368187, | |
| "grad_norm": 0.0888429582118988, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6608, | |
| "mean_token_accuracy": 0.796707707643509, | |
| "num_tokens": 7811841.0, | |
| "step": 240 | |
| }, | |
| { | |
| "entropy": 0.6725413352251053, | |
| "epoch": 1.1112347052280311, | |
| "grad_norm": 0.09360355883836746, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6687, | |
| "mean_token_accuracy": 0.7943419247865677, | |
| "num_tokens": 8139521.0, | |
| "step": 250 | |
| }, | |
| { | |
| "entropy": 0.6608762927353382, | |
| "epoch": 1.1557285873192435, | |
| "grad_norm": 0.09271729737520218, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6528, | |
| "mean_token_accuracy": 0.7991992741823196, | |
| "num_tokens": 8467165.0, | |
| "step": 260 | |
| }, | |
| { | |
| "entropy": 0.6472577676177025, | |
| "epoch": 1.200222469410456, | |
| "grad_norm": 0.0905335545539856, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6395, | |
| "mean_token_accuracy": 0.8023400634527207, | |
| "num_tokens": 8794479.0, | |
| "step": 270 | |
| }, | |
| { | |
| "entropy": 0.7025618955492974, | |
| "epoch": 1.2447163515016686, | |
| "grad_norm": 0.09981676936149597, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6969, | |
| "mean_token_accuracy": 0.7927162751555443, | |
| "num_tokens": 9115502.0, | |
| "step": 280 | |
| }, | |
| { | |
| "entropy": 0.6550201892852783, | |
| "epoch": 1.289210233592881, | |
| "grad_norm": 0.09637434035539627, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6472, | |
| "mean_token_accuracy": 0.7994350507855416, | |
| "num_tokens": 9443182.0, | |
| "step": 290 | |
| }, | |
| { | |
| "entropy": 0.6535603314638138, | |
| "epoch": 1.3337041156840934, | |
| "grad_norm": 0.09058881551027298, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6505, | |
| "mean_token_accuracy": 0.8004654511809349, | |
| "num_tokens": 9770862.0, | |
| "step": 300 | |
| }, | |
| { | |
| "entropy": 0.6676291272044181, | |
| "epoch": 1.3781979977753058, | |
| "grad_norm": 0.10104376822710037, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6605, | |
| "mean_token_accuracy": 0.7970087364315986, | |
| "num_tokens": 10098521.0, | |
| "step": 310 | |
| }, | |
| { | |
| "entropy": 0.6632239744067192, | |
| "epoch": 1.4226918798665182, | |
| "grad_norm": 0.09252401441335678, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6567, | |
| "mean_token_accuracy": 0.7999079167842865, | |
| "num_tokens": 10425894.0, | |
| "step": 320 | |
| }, | |
| { | |
| "entropy": 0.6941266790032387, | |
| "epoch": 1.4671857619577309, | |
| "grad_norm": 0.10002182424068451, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6904, | |
| "mean_token_accuracy": 0.7945976346731186, | |
| "num_tokens": 10746167.0, | |
| "step": 330 | |
| }, | |
| { | |
| "entropy": 0.6502977207303047, | |
| "epoch": 1.5116796440489433, | |
| "grad_norm": 0.10251280665397644, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6447, | |
| "mean_token_accuracy": 0.8009829133749008, | |
| "num_tokens": 11073847.0, | |
| "step": 340 | |
| }, | |
| { | |
| "entropy": 0.6449682429432869, | |
| "epoch": 1.5561735261401557, | |
| "grad_norm": 0.10051094740629196, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6297, | |
| "mean_token_accuracy": 0.8034362852573395, | |
| "num_tokens": 11401527.0, | |
| "step": 350 | |
| }, | |
| { | |
| "entropy": 0.647579725086689, | |
| "epoch": 1.6006674082313683, | |
| "grad_norm": 0.09883299469947815, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6439, | |
| "mean_token_accuracy": 0.8008567914366722, | |
| "num_tokens": 11729207.0, | |
| "step": 360 | |
| }, | |
| { | |
| "entropy": 0.6341586455702781, | |
| "epoch": 1.6451612903225805, | |
| "grad_norm": 0.0934380292892456, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6273, | |
| "mean_token_accuracy": 0.805011647939682, | |
| "num_tokens": 12056607.0, | |
| "step": 370 | |
| }, | |
| { | |
| "entropy": 0.7022602126002312, | |
| "epoch": 1.6896551724137931, | |
| "grad_norm": 0.10245587676763535, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6997, | |
| "mean_token_accuracy": 0.7951286539435387, | |
| "num_tokens": 12375719.0, | |
| "step": 380 | |
| }, | |
| { | |
| "entropy": 0.6635140493512154, | |
| "epoch": 1.7341490545050056, | |
| "grad_norm": 0.10037633031606674, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6532, | |
| "mean_token_accuracy": 0.7992691040039063, | |
| "num_tokens": 12703399.0, | |
| "step": 390 | |
| }, | |
| { | |
| "entropy": 0.6396854467689991, | |
| "epoch": 1.778642936596218, | |
| "grad_norm": 0.09362801909446716, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6319, | |
| "mean_token_accuracy": 0.8054976046085358, | |
| "num_tokens": 13031079.0, | |
| "step": 400 | |
| }, | |
| { | |
| "entropy": 0.645181667804718, | |
| "epoch": 1.8231368186874306, | |
| "grad_norm": 0.09476775676012039, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6366, | |
| "mean_token_accuracy": 0.8037867248058319, | |
| "num_tokens": 13358733.0, | |
| "step": 410 | |
| }, | |
| { | |
| "entropy": 0.652097150683403, | |
| "epoch": 1.8676307007786428, | |
| "grad_norm": 0.09551919251680374, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6399, | |
| "mean_token_accuracy": 0.8037777736783027, | |
| "num_tokens": 13685998.0, | |
| "step": 420 | |
| }, | |
| { | |
| "entropy": 0.6837869822978974, | |
| "epoch": 1.9121245828698554, | |
| "grad_norm": 0.10013949871063232, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6841, | |
| "mean_token_accuracy": 0.7974902361631393, | |
| "num_tokens": 14005579.0, | |
| "step": 430 | |
| }, | |
| { | |
| "entropy": 0.628454002737999, | |
| "epoch": 1.9566184649610678, | |
| "grad_norm": 0.13781693577766418, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6213, | |
| "mean_token_accuracy": 0.8075341194868088, | |
| "num_tokens": 14333259.0, | |
| "step": 440 | |
| }, | |
| { | |
| "entropy": 0.6827295254438351, | |
| "epoch": 2.0, | |
| "grad_norm": 0.15781015157699585, | |
| "learning_rate": 0.0002, | |
| "loss": 0.68, | |
| "mean_token_accuracy": 0.7996979187696408, | |
| "num_tokens": 14640642.0, | |
| "step": 450 | |
| }, | |
| { | |
| "entropy": 0.5626548573374748, | |
| "epoch": 2.0444938820912126, | |
| "grad_norm": 0.10359702259302139, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5339, | |
| "mean_token_accuracy": 0.8331230849027633, | |
| "num_tokens": 14968322.0, | |
| "step": 460 | |
| }, | |
| { | |
| "entropy": 0.5715467736124993, | |
| "epoch": 2.088987764182425, | |
| "grad_norm": 0.10643124580383301, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5351, | |
| "mean_token_accuracy": 0.8342131584882736, | |
| "num_tokens": 15296002.0, | |
| "step": 470 | |
| }, | |
| { | |
| "entropy": 0.5666955970227718, | |
| "epoch": 2.1334816462736375, | |
| "grad_norm": 0.10716599971055984, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5381, | |
| "mean_token_accuracy": 0.8319605439901352, | |
| "num_tokens": 15623682.0, | |
| "step": 480 | |
| }, | |
| { | |
| "entropy": 0.5521194256842137, | |
| "epoch": 2.1779755283648496, | |
| "grad_norm": 0.11130277067422867, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5209, | |
| "mean_token_accuracy": 0.8368723064661026, | |
| "num_tokens": 15951273.0, | |
| "step": 490 | |
| }, | |
| { | |
| "entropy": 0.6372111544013024, | |
| "epoch": 2.2224694104560623, | |
| "grad_norm": 0.10940540581941605, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6048, | |
| "mean_token_accuracy": 0.8223052099347115, | |
| "num_tokens": 16269048.0, | |
| "step": 500 | |
| }, | |
| { | |
| "entropy": 0.5748868621885777, | |
| "epoch": 2.266963292547275, | |
| "grad_norm": 0.1133793443441391, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5473, | |
| "mean_token_accuracy": 0.8281539842486382, | |
| "num_tokens": 16596728.0, | |
| "step": 510 | |
| }, | |
| { | |
| "entropy": 0.5760611467063427, | |
| "epoch": 2.311457174638487, | |
| "grad_norm": 0.11276053637266159, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5512, | |
| "mean_token_accuracy": 0.8286992073059082, | |
| "num_tokens": 16924408.0, | |
| "step": 520 | |
| }, | |
| { | |
| "entropy": 0.5595883250236511, | |
| "epoch": 2.3559510567296997, | |
| "grad_norm": 0.10785238444805145, | |
| "learning_rate": 0.0002, | |
| "loss": 0.534, | |
| "mean_token_accuracy": 0.8329534187912941, | |
| "num_tokens": 17252088.0, | |
| "step": 530 | |
| }, | |
| { | |
| "entropy": 0.5623729139566421, | |
| "epoch": 2.400444938820912, | |
| "grad_norm": 0.11915665119886398, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5318, | |
| "mean_token_accuracy": 0.833781997859478, | |
| "num_tokens": 17579671.0, | |
| "step": 540 | |
| }, | |
| { | |
| "entropy": 0.6471138700842858, | |
| "epoch": 2.4449388209121246, | |
| "grad_norm": 0.09796369075775146, | |
| "learning_rate": 0.0002, | |
| "loss": 0.62, | |
| "mean_token_accuracy": 0.8189012482762337, | |
| "num_tokens": 17899138.0, | |
| "step": 550 | |
| }, | |
| { | |
| "entropy": 0.5706302203238011, | |
| "epoch": 2.489432703003337, | |
| "grad_norm": 0.11501935124397278, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5443, | |
| "mean_token_accuracy": 0.8297176495194435, | |
| "num_tokens": 18226818.0, | |
| "step": 560 | |
| }, | |
| { | |
| "entropy": 0.5607033468782902, | |
| "epoch": 2.5339265850945494, | |
| "grad_norm": 0.10209450125694275, | |
| "learning_rate": 0.0002, | |
| "loss": 0.537, | |
| "mean_token_accuracy": 0.8327666461467743, | |
| "num_tokens": 18554498.0, | |
| "step": 570 | |
| }, | |
| { | |
| "entropy": 0.5697788536548615, | |
| "epoch": 2.578420467185762, | |
| "grad_norm": 0.11915192008018494, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5417, | |
| "mean_token_accuracy": 0.8292823225259781, | |
| "num_tokens": 18882178.0, | |
| "step": 580 | |
| }, | |
| { | |
| "entropy": 0.5598745912313461, | |
| "epoch": 2.622914349276974, | |
| "grad_norm": 0.12072969228029251, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5318, | |
| "mean_token_accuracy": 0.8316263392567634, | |
| "num_tokens": 19209735.0, | |
| "step": 590 | |
| }, | |
| { | |
| "entropy": 0.6062877185642719, | |
| "epoch": 2.667408231368187, | |
| "grad_norm": 0.11766766756772995, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5785, | |
| "mean_token_accuracy": 0.8279747024178505, | |
| "num_tokens": 19531022.0, | |
| "step": 600 | |
| }, | |
| { | |
| "entropy": 0.5801788948476314, | |
| "epoch": 2.7119021134593995, | |
| "grad_norm": 0.11248841881752014, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5555, | |
| "mean_token_accuracy": 0.8277870997786522, | |
| "num_tokens": 19858702.0, | |
| "step": 610 | |
| }, | |
| { | |
| "entropy": 0.5594978570938111, | |
| "epoch": 2.7563959955506117, | |
| "grad_norm": 0.11471142619848251, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5343, | |
| "mean_token_accuracy": 0.8331717640161515, | |
| "num_tokens": 20186382.0, | |
| "step": 620 | |
| }, | |
| { | |
| "entropy": 0.5783086307346821, | |
| "epoch": 2.8008898776418243, | |
| "grad_norm": 0.12316859513521194, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5514, | |
| "mean_token_accuracy": 0.8270430579781533, | |
| "num_tokens": 20514062.0, | |
| "step": 630 | |
| }, | |
| { | |
| "entropy": 0.5552236251533031, | |
| "epoch": 2.8453837597330365, | |
| "grad_norm": 0.1180945634841919, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5254, | |
| "mean_token_accuracy": 0.833421728014946, | |
| "num_tokens": 20841607.0, | |
| "step": 640 | |
| }, | |
| { | |
| "entropy": 0.6508079275488854, | |
| "epoch": 2.889877641824249, | |
| "grad_norm": 0.11594584584236145, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6277, | |
| "mean_token_accuracy": 0.8178864538669586, | |
| "num_tokens": 21156705.0, | |
| "step": 650 | |
| }, | |
| { | |
| "entropy": 0.5584865964949131, | |
| "epoch": 2.9343715239154617, | |
| "grad_norm": 0.10876645147800446, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5341, | |
| "mean_token_accuracy": 0.8327999398112297, | |
| "num_tokens": 21484385.0, | |
| "step": 660 | |
| }, | |
| { | |
| "entropy": 0.5605839185416699, | |
| "epoch": 2.978865406006674, | |
| "grad_norm": 0.12009165436029434, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5317, | |
| "mean_token_accuracy": 0.8333747163414955, | |
| "num_tokens": 21812022.0, | |
| "step": 670 | |
| }, | |
| { | |
| "entropy": 0.5539867511162391, | |
| "epoch": 3.0222469410456063, | |
| "grad_norm": 0.12213966995477676, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4962, | |
| "mean_token_accuracy": 0.8499505198918856, | |
| "num_tokens": 22124803.0, | |
| "step": 680 | |
| }, | |
| { | |
| "entropy": 0.48398632630705835, | |
| "epoch": 3.0667408231368185, | |
| "grad_norm": 0.1170133501291275, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4345, | |
| "mean_token_accuracy": 0.8670241564512253, | |
| "num_tokens": 22452483.0, | |
| "step": 690 | |
| }, | |
| { | |
| "entropy": 0.4720599815249443, | |
| "epoch": 3.111234705228031, | |
| "grad_norm": 0.12524375319480896, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4173, | |
| "mean_token_accuracy": 0.8702847585082054, | |
| "num_tokens": 22780163.0, | |
| "step": 700 | |
| }, | |
| { | |
| "entropy": 0.4982055649161339, | |
| "epoch": 3.1557285873192438, | |
| "grad_norm": 0.11435994505882263, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4423, | |
| "mean_token_accuracy": 0.864942966401577, | |
| "num_tokens": 23107835.0, | |
| "step": 710 | |
| }, | |
| { | |
| "entropy": 0.48039583414793013, | |
| "epoch": 3.200222469410456, | |
| "grad_norm": 0.12139066308736801, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4226, | |
| "mean_token_accuracy": 0.8702625542879104, | |
| "num_tokens": 23435179.0, | |
| "step": 720 | |
| }, | |
| { | |
| "entropy": 0.5418343998491764, | |
| "epoch": 3.2447163515016686, | |
| "grad_norm": 0.13233190774917603, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4924, | |
| "mean_token_accuracy": 0.8546271666884422, | |
| "num_tokens": 23753029.0, | |
| "step": 730 | |
| }, | |
| { | |
| "entropy": 0.4849280208349228, | |
| "epoch": 3.289210233592881, | |
| "grad_norm": 0.12592358887195587, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4278, | |
| "mean_token_accuracy": 0.8678036436438561, | |
| "num_tokens": 24080709.0, | |
| "step": 740 | |
| }, | |
| { | |
| "entropy": 0.5037459127604962, | |
| "epoch": 3.3337041156840934, | |
| "grad_norm": 0.12919606268405914, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4524, | |
| "mean_token_accuracy": 0.8612239480018615, | |
| "num_tokens": 24408389.0, | |
| "step": 750 | |
| }, | |
| { | |
| "entropy": 0.4822153463959694, | |
| "epoch": 3.378197997775306, | |
| "grad_norm": 0.12449780106544495, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4291, | |
| "mean_token_accuracy": 0.8676740288734436, | |
| "num_tokens": 24736036.0, | |
| "step": 760 | |
| }, | |
| { | |
| "entropy": 0.4992881424725056, | |
| "epoch": 3.4226918798665182, | |
| "grad_norm": 0.1114158034324646, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4433, | |
| "mean_token_accuracy": 0.8635369181632996, | |
| "num_tokens": 25063393.0, | |
| "step": 770 | |
| }, | |
| { | |
| "entropy": 0.5634101495146752, | |
| "epoch": 3.467185761957731, | |
| "grad_norm": 0.13151301443576813, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5292, | |
| "mean_token_accuracy": 0.848271718621254, | |
| "num_tokens": 25383789.0, | |
| "step": 780 | |
| }, | |
| { | |
| "entropy": 0.5000352062284946, | |
| "epoch": 3.511679644048943, | |
| "grad_norm": 0.1317589432001114, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4488, | |
| "mean_token_accuracy": 0.8623447984457016, | |
| "num_tokens": 25711469.0, | |
| "step": 790 | |
| }, | |
| { | |
| "entropy": 0.5025811523199082, | |
| "epoch": 3.5561735261401557, | |
| "grad_norm": 0.12309940159320831, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4515, | |
| "mean_token_accuracy": 0.8600758820772171, | |
| "num_tokens": 26039149.0, | |
| "step": 800 | |
| }, | |
| { | |
| "entropy": 0.4819103963673115, | |
| "epoch": 3.6006674082313683, | |
| "grad_norm": 0.12250859290361404, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4342, | |
| "mean_token_accuracy": 0.8647636532783508, | |
| "num_tokens": 26366815.0, | |
| "step": 810 | |
| }, | |
| { | |
| "entropy": 0.4777693450450897, | |
| "epoch": 3.6451612903225805, | |
| "grad_norm": 0.11709456145763397, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4249, | |
| "mean_token_accuracy": 0.8680423229932785, | |
| "num_tokens": 26694191.0, | |
| "step": 820 | |
| }, | |
| { | |
| "entropy": 0.5633887559175491, | |
| "epoch": 3.689655172413793, | |
| "grad_norm": 0.12784965336322784, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5239, | |
| "mean_token_accuracy": 0.8493879333138465, | |
| "num_tokens": 27013314.0, | |
| "step": 830 | |
| }, | |
| { | |
| "entropy": 0.48369882851839063, | |
| "epoch": 3.7341490545050053, | |
| "grad_norm": 0.12846286594867706, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4369, | |
| "mean_token_accuracy": 0.8636931553483009, | |
| "num_tokens": 27340994.0, | |
| "step": 840 | |
| }, | |
| { | |
| "entropy": 0.5082808144390583, | |
| "epoch": 3.778642936596218, | |
| "grad_norm": 0.12857837975025177, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4617, | |
| "mean_token_accuracy": 0.8581709936261177, | |
| "num_tokens": 27668674.0, | |
| "step": 850 | |
| }, | |
| { | |
| "entropy": 0.47558383345603944, | |
| "epoch": 3.8231368186874306, | |
| "grad_norm": 0.12611748278141022, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4235, | |
| "mean_token_accuracy": 0.866823847591877, | |
| "num_tokens": 27996329.0, | |
| "step": 860 | |
| }, | |
| { | |
| "entropy": 0.4854692369699478, | |
| "epoch": 3.867630700778643, | |
| "grad_norm": 0.12649761140346527, | |
| "learning_rate": 0.0002, | |
| "loss": 0.44, | |
| "mean_token_accuracy": 0.863289874792099, | |
| "num_tokens": 28323622.0, | |
| "step": 870 | |
| }, | |
| { | |
| "entropy": 0.5147525526583194, | |
| "epoch": 3.9121245828698554, | |
| "grad_norm": 0.1261458545923233, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4702, | |
| "mean_token_accuracy": 0.8576680824160576, | |
| "num_tokens": 28644498.0, | |
| "step": 880 | |
| }, | |
| { | |
| "entropy": 0.5058113448321819, | |
| "epoch": 3.9566184649610676, | |
| "grad_norm": 0.12469454854726791, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4509, | |
| "mean_token_accuracy": 0.8585198208689689, | |
| "num_tokens": 28972178.0, | |
| "step": 890 | |
| }, | |
| { | |
| "entropy": 0.5309454324917916, | |
| "epoch": 4.0, | |
| "grad_norm": 0.14466610550880432, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4892, | |
| "mean_token_accuracy": 0.8545252420963385, | |
| "num_tokens": 29281284.0, | |
| "step": 900 | |
| }, | |
| { | |
| "entropy": 0.4318658284842968, | |
| "epoch": 4.044493882091213, | |
| "grad_norm": 0.1319647878408432, | |
| "learning_rate": 0.0002, | |
| "loss": 0.361, | |
| "mean_token_accuracy": 0.8950821414589882, | |
| "num_tokens": 29608964.0, | |
| "step": 910 | |
| }, | |
| { | |
| "entropy": 0.4118408665060997, | |
| "epoch": 4.088987764182425, | |
| "grad_norm": 0.12470532953739166, | |
| "learning_rate": 0.0002, | |
| "loss": 0.333, | |
| "mean_token_accuracy": 0.901517990231514, | |
| "num_tokens": 29936644.0, | |
| "step": 920 | |
| }, | |
| { | |
| "entropy": 0.4389042042195797, | |
| "epoch": 4.133481646273637, | |
| "grad_norm": 0.1262277513742447, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3689, | |
| "mean_token_accuracy": 0.8918602123856545, | |
| "num_tokens": 30264324.0, | |
| "step": 930 | |
| }, | |
| { | |
| "entropy": 0.40702272206544876, | |
| "epoch": 4.17797552836485, | |
| "grad_norm": 0.1268932968378067, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3338, | |
| "mean_token_accuracy": 0.9011573985219001, | |
| "num_tokens": 30591882.0, | |
| "step": 940 | |
| }, | |
| { | |
| "entropy": 0.4825476065278053, | |
| "epoch": 4.222469410456062, | |
| "grad_norm": 0.11590472608804703, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4132, | |
| "mean_token_accuracy": 0.8850692108273506, | |
| "num_tokens": 30911133.0, | |
| "step": 950 | |
| }, | |
| { | |
| "entropy": 0.4279182717204094, | |
| "epoch": 4.266963292547275, | |
| "grad_norm": 0.1330350935459137, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3625, | |
| "mean_token_accuracy": 0.893429034948349, | |
| "num_tokens": 31238813.0, | |
| "step": 960 | |
| }, | |
| { | |
| "entropy": 0.41735711842775347, | |
| "epoch": 4.3114571746384875, | |
| "grad_norm": 0.13147839903831482, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3465, | |
| "mean_token_accuracy": 0.896250243484974, | |
| "num_tokens": 31566493.0, | |
| "step": 970 | |
| }, | |
| { | |
| "entropy": 0.41771056577563287, | |
| "epoch": 4.355951056729699, | |
| "grad_norm": 0.1316802203655243, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3503, | |
| "mean_token_accuracy": 0.8955738350749016, | |
| "num_tokens": 31894173.0, | |
| "step": 980 | |
| }, | |
| { | |
| "entropy": 0.41876513287425043, | |
| "epoch": 4.400444938820912, | |
| "grad_norm": 0.13833066821098328, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3486, | |
| "mean_token_accuracy": 0.8961227789521218, | |
| "num_tokens": 32221752.0, | |
| "step": 990 | |
| }, | |
| { | |
| "entropy": 0.514089448004961, | |
| "epoch": 4.4449388209121246, | |
| "grad_norm": 0.12354996800422668, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4487, | |
| "mean_token_accuracy": 0.8752555340528488, | |
| "num_tokens": 32538458.0, | |
| "step": 1000 | |
| }, | |
| { | |
| "entropy": 0.4165398746728897, | |
| "epoch": 4.489432703003337, | |
| "grad_norm": 0.13673128187656403, | |
| "learning_rate": 0.0002, | |
| "loss": 0.355, | |
| "mean_token_accuracy": 0.8935215994715691, | |
| "num_tokens": 32866138.0, | |
| "step": 1010 | |
| }, | |
| { | |
| "entropy": 0.4195031985640526, | |
| "epoch": 4.53392658509455, | |
| "grad_norm": 0.12385623902082443, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3561, | |
| "mean_token_accuracy": 0.8933135420084, | |
| "num_tokens": 33193818.0, | |
| "step": 1020 | |
| }, | |
| { | |
| "entropy": 0.4167650781571865, | |
| "epoch": 4.578420467185762, | |
| "grad_norm": 0.1269628405570984, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3523, | |
| "mean_token_accuracy": 0.8935295462608337, | |
| "num_tokens": 33521498.0, | |
| "step": 1030 | |
| }, | |
| { | |
| "entropy": 0.4086064472794533, | |
| "epoch": 4.622914349276974, | |
| "grad_norm": 0.13126298785209656, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3432, | |
| "mean_token_accuracy": 0.8961806997656823, | |
| "num_tokens": 33849067.0, | |
| "step": 1040 | |
| }, | |
| { | |
| "entropy": 0.499802465736866, | |
| "epoch": 4.667408231368187, | |
| "grad_norm": 0.1461581140756607, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4409, | |
| "mean_token_accuracy": 0.8770750179886818, | |
| "num_tokens": 34165797.0, | |
| "step": 1050 | |
| }, | |
| { | |
| "entropy": 0.41772385239601134, | |
| "epoch": 4.7119021134593995, | |
| "grad_norm": 0.13139505684375763, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3549, | |
| "mean_token_accuracy": 0.8931645408272744, | |
| "num_tokens": 34493477.0, | |
| "step": 1060 | |
| }, | |
| { | |
| "entropy": 0.4297786563634872, | |
| "epoch": 4.756395995550612, | |
| "grad_norm": 0.13271018862724304, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3666, | |
| "mean_token_accuracy": 0.8901869565248489, | |
| "num_tokens": 34821157.0, | |
| "step": 1070 | |
| }, | |
| { | |
| "entropy": 0.41754806637763975, | |
| "epoch": 4.800889877641824, | |
| "grad_norm": 0.1331576555967331, | |
| "learning_rate": 0.0002, | |
| "loss": 0.359, | |
| "mean_token_accuracy": 0.8924819275736808, | |
| "num_tokens": 35148837.0, | |
| "step": 1080 | |
| }, | |
| { | |
| "entropy": 0.406176183372736, | |
| "epoch": 4.8453837597330365, | |
| "grad_norm": 0.13308922946453094, | |
| "learning_rate": 0.0002, | |
| "loss": 0.34, | |
| "mean_token_accuracy": 0.8957495510578155, | |
| "num_tokens": 35476418.0, | |
| "step": 1090 | |
| }, | |
| { | |
| "entropy": 0.4817497693002224, | |
| "epoch": 4.889877641824249, | |
| "grad_norm": 0.11259204149246216, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4197, | |
| "mean_token_accuracy": 0.8795695826411247, | |
| "num_tokens": 35799296.0, | |
| "step": 1100 | |
| }, | |
| { | |
| "entropy": 0.42334735170006754, | |
| "epoch": 4.934371523915462, | |
| "grad_norm": 0.13311080634593964, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3669, | |
| "mean_token_accuracy": 0.8901919454336167, | |
| "num_tokens": 36126976.0, | |
| "step": 1110 | |
| }, | |
| { | |
| "entropy": 0.41451009213924406, | |
| "epoch": 4.978865406006674, | |
| "grad_norm": 0.1383262574672699, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3487, | |
| "mean_token_accuracy": 0.8940043538808823, | |
| "num_tokens": 36454597.0, | |
| "step": 1120 | |
| }, | |
| { | |
| "entropy": 0.4311733207641504, | |
| "epoch": 5.022246941045606, | |
| "grad_norm": 0.1265476644039154, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3571, | |
| "mean_token_accuracy": 0.9023441947423495, | |
| "num_tokens": 36765445.0, | |
| "step": 1130 | |
| }, | |
| { | |
| "entropy": 0.3527758061885834, | |
| "epoch": 5.0667408231368185, | |
| "grad_norm": 0.11817038059234619, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2815, | |
| "mean_token_accuracy": 0.9223723724484444, | |
| "num_tokens": 37093125.0, | |
| "step": 1140 | |
| }, | |
| { | |
| "entropy": 0.3410337470471859, | |
| "epoch": 5.111234705228031, | |
| "grad_norm": 0.1334749162197113, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2625, | |
| "mean_token_accuracy": 0.925548429787159, | |
| "num_tokens": 37420805.0, | |
| "step": 1150 | |
| }, | |
| { | |
| "entropy": 0.3618052005767822, | |
| "epoch": 5.155728587319244, | |
| "grad_norm": 0.12979310750961304, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2859, | |
| "mean_token_accuracy": 0.9212473854422569, | |
| "num_tokens": 37748485.0, | |
| "step": 1160 | |
| }, | |
| { | |
| "entropy": 0.3468689829111099, | |
| "epoch": 5.200222469410456, | |
| "grad_norm": 0.12981602549552917, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2635, | |
| "mean_token_accuracy": 0.9270615592598915, | |
| "num_tokens": 38075894.0, | |
| "step": 1170 | |
| }, | |
| { | |
| "entropy": 0.43322302848100663, | |
| "epoch": 5.244716351501668, | |
| "grad_norm": 0.1404024064540863, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3655, | |
| "mean_token_accuracy": 0.9022073581814766, | |
| "num_tokens": 38396794.0, | |
| "step": 1180 | |
| }, | |
| { | |
| "entropy": 0.35545270666480067, | |
| "epoch": 5.289210233592881, | |
| "grad_norm": 0.129226952791214, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2817, | |
| "mean_token_accuracy": 0.9216006860136986, | |
| "num_tokens": 38724474.0, | |
| "step": 1190 | |
| }, | |
| { | |
| "entropy": 0.3626397028565407, | |
| "epoch": 5.333704115684093, | |
| "grad_norm": 0.12991155683994293, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2881, | |
| "mean_token_accuracy": 0.9190065816044808, | |
| "num_tokens": 39052154.0, | |
| "step": 1200 | |
| }, | |
| { | |
| "entropy": 0.3538004279136658, | |
| "epoch": 5.378197997775306, | |
| "grad_norm": 0.13420861959457397, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2784, | |
| "mean_token_accuracy": 0.9214035272598267, | |
| "num_tokens": 39379796.0, | |
| "step": 1210 | |
| }, | |
| { | |
| "entropy": 0.36458633579313754, | |
| "epoch": 5.422691879866519, | |
| "grad_norm": 0.12140852212905884, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2919, | |
| "mean_token_accuracy": 0.9194388329982758, | |
| "num_tokens": 39707050.0, | |
| "step": 1220 | |
| }, | |
| { | |
| "entropy": 0.40958601981401443, | |
| "epoch": 5.46718576195773, | |
| "grad_norm": 0.1385655701160431, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3391, | |
| "mean_token_accuracy": 0.9059785082936287, | |
| "num_tokens": 40022495.0, | |
| "step": 1230 | |
| }, | |
| { | |
| "entropy": 0.3671642690896988, | |
| "epoch": 5.511679644048943, | |
| "grad_norm": 0.15173068642616272, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2929, | |
| "mean_token_accuracy": 0.9173887968063354, | |
| "num_tokens": 40350175.0, | |
| "step": 1240 | |
| }, | |
| { | |
| "entropy": 0.3846460208296776, | |
| "epoch": 5.556173526140156, | |
| "grad_norm": 0.13141454756259918, | |
| "learning_rate": 0.0002, | |
| "loss": 0.313, | |
| "mean_token_accuracy": 0.9108256429433823, | |
| "num_tokens": 40677855.0, | |
| "step": 1250 | |
| }, | |
| { | |
| "entropy": 0.3476515613496304, | |
| "epoch": 5.600667408231368, | |
| "grad_norm": 0.14480914175510406, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2755, | |
| "mean_token_accuracy": 0.9198622301220893, | |
| "num_tokens": 41005532.0, | |
| "step": 1260 | |
| }, | |
| { | |
| "entropy": 0.35671931579709054, | |
| "epoch": 5.645161290322581, | |
| "grad_norm": 0.1569768637418747, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2837, | |
| "mean_token_accuracy": 0.9195132747292518, | |
| "num_tokens": 41332938.0, | |
| "step": 1270 | |
| }, | |
| { | |
| "entropy": 0.4298785075545311, | |
| "epoch": 5.689655172413794, | |
| "grad_norm": 0.13758735358715057, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3626, | |
| "mean_token_accuracy": 0.9006465882062912, | |
| "num_tokens": 41652505.0, | |
| "step": 1280 | |
| }, | |
| { | |
| "entropy": 0.36134291216731074, | |
| "epoch": 5.734149054505005, | |
| "grad_norm": 0.14216244220733643, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2935, | |
| "mean_token_accuracy": 0.9152890175580979, | |
| "num_tokens": 41980185.0, | |
| "step": 1290 | |
| }, | |
| { | |
| "entropy": 0.3726960562169552, | |
| "epoch": 5.778642936596218, | |
| "grad_norm": 0.14236606657505035, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3048, | |
| "mean_token_accuracy": 0.9139653652906418, | |
| "num_tokens": 42307865.0, | |
| "step": 1300 | |
| }, | |
| { | |
| "entropy": 0.36691512912511826, | |
| "epoch": 5.823136818687431, | |
| "grad_norm": 0.1484624594449997, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2968, | |
| "mean_token_accuracy": 0.915194633603096, | |
| "num_tokens": 42635518.0, | |
| "step": 1310 | |
| }, | |
| { | |
| "entropy": 0.3883252374827862, | |
| "epoch": 5.867630700778643, | |
| "grad_norm": 0.13436342775821686, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3196, | |
| "mean_token_accuracy": 0.9100908860564232, | |
| "num_tokens": 42962763.0, | |
| "step": 1320 | |
| }, | |
| { | |
| "entropy": 0.43875667452812195, | |
| "epoch": 5.912124582869856, | |
| "grad_norm": 0.14851780235767365, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3816, | |
| "mean_token_accuracy": 0.8956333592534065, | |
| "num_tokens": 43280632.0, | |
| "step": 1330 | |
| }, | |
| { | |
| "entropy": 0.38048830032348635, | |
| "epoch": 5.956618464961068, | |
| "grad_norm": 0.13980592787265778, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3082, | |
| "mean_token_accuracy": 0.9120747581124306, | |
| "num_tokens": 43608312.0, | |
| "step": 1340 | |
| }, | |
| { | |
| "entropy": 0.4035546894256885, | |
| "epoch": 6.0, | |
| "grad_norm": 0.1455690562725067, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3405, | |
| "mean_token_accuracy": 0.9065795892324203, | |
| "num_tokens": 43921926.0, | |
| "step": 1350 | |
| }, | |
| { | |
| "entropy": 0.31527973636984824, | |
| "epoch": 6.044493882091213, | |
| "grad_norm": 0.14178436994552612, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2397, | |
| "mean_token_accuracy": 0.9377185016870498, | |
| "num_tokens": 44249606.0, | |
| "step": 1360 | |
| }, | |
| { | |
| "entropy": 0.3210640762001276, | |
| "epoch": 6.088987764182425, | |
| "grad_norm": 0.12400421500205994, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2384, | |
| "mean_token_accuracy": 0.937873150408268, | |
| "num_tokens": 44577286.0, | |
| "step": 1370 | |
| }, | |
| { | |
| "entropy": 0.31932172551751137, | |
| "epoch": 6.133481646273637, | |
| "grad_norm": 0.14534889161586761, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2461, | |
| "mean_token_accuracy": 0.9356767356395721, | |
| "num_tokens": 44904966.0, | |
| "step": 1380 | |
| }, | |
| { | |
| "entropy": 0.3014022711664438, | |
| "epoch": 6.17797552836485, | |
| "grad_norm": 0.1284133493900299, | |
| "learning_rate": 0.0002, | |
| "loss": 0.225, | |
| "mean_token_accuracy": 0.9409583687782288, | |
| "num_tokens": 45232502.0, | |
| "step": 1390 | |
| }, | |
| { | |
| "entropy": 0.3887506004422903, | |
| "epoch": 6.222469410456062, | |
| "grad_norm": 0.11520787328481674, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3126, | |
| "mean_token_accuracy": 0.920306672155857, | |
| "num_tokens": 45550770.0, | |
| "step": 1400 | |
| }, | |
| { | |
| "entropy": 0.3105563811957836, | |
| "epoch": 6.266963292547275, | |
| "grad_norm": 0.12666074931621552, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2376, | |
| "mean_token_accuracy": 0.9375046223402024, | |
| "num_tokens": 45878450.0, | |
| "step": 1410 | |
| }, | |
| { | |
| "entropy": 0.31831442639231683, | |
| "epoch": 6.3114571746384875, | |
| "grad_norm": 0.13169772922992706, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2445, | |
| "mean_token_accuracy": 0.9347741514444351, | |
| "num_tokens": 46206130.0, | |
| "step": 1420 | |
| }, | |
| { | |
| "entropy": 0.3245772857218981, | |
| "epoch": 6.355951056729699, | |
| "grad_norm": 0.1360885202884674, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2579, | |
| "mean_token_accuracy": 0.9318127825856208, | |
| "num_tokens": 46533810.0, | |
| "step": 1430 | |
| }, | |
| { | |
| "entropy": 0.2980411469936371, | |
| "epoch": 6.400444938820912, | |
| "grad_norm": 0.13366331160068512, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2218, | |
| "mean_token_accuracy": 0.9406819701194763, | |
| "num_tokens": 46861397.0, | |
| "step": 1440 | |
| }, | |
| { | |
| "entropy": 0.3910529091954231, | |
| "epoch": 6.4449388209121246, | |
| "grad_norm": 0.1334383487701416, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3195, | |
| "mean_token_accuracy": 0.9170822516083718, | |
| "num_tokens": 47177802.0, | |
| "step": 1450 | |
| }, | |
| { | |
| "entropy": 0.29828712679445746, | |
| "epoch": 6.489432703003337, | |
| "grad_norm": 0.13855065405368805, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2263, | |
| "mean_token_accuracy": 0.9390584915876389, | |
| "num_tokens": 47505482.0, | |
| "step": 1460 | |
| }, | |
| { | |
| "entropy": 0.31491950489580633, | |
| "epoch": 6.53392658509455, | |
| "grad_norm": 0.1417187601327896, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2428, | |
| "mean_token_accuracy": 0.9350977450609207, | |
| "num_tokens": 47833162.0, | |
| "step": 1470 | |
| }, | |
| { | |
| "entropy": 0.31174046322703364, | |
| "epoch": 6.578420467185762, | |
| "grad_norm": 0.14676059782505035, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2374, | |
| "mean_token_accuracy": 0.936232116818428, | |
| "num_tokens": 48160842.0, | |
| "step": 1480 | |
| }, | |
| { | |
| "entropy": 0.2903703946620226, | |
| "epoch": 6.622914349276974, | |
| "grad_norm": 0.1506464034318924, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2187, | |
| "mean_token_accuracy": 0.9405747085809708, | |
| "num_tokens": 48488428.0, | |
| "step": 1490 | |
| }, | |
| { | |
| "entropy": 0.38095769509673116, | |
| "epoch": 6.667408231368187, | |
| "grad_norm": 0.1296055167913437, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3102, | |
| "mean_token_accuracy": 0.9190786957740784, | |
| "num_tokens": 48808893.0, | |
| "step": 1500 | |
| }, | |
| { | |
| "entropy": 0.31492537409067156, | |
| "epoch": 6.7119021134593995, | |
| "grad_norm": 0.13277411460876465, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2465, | |
| "mean_token_accuracy": 0.9332764238119126, | |
| "num_tokens": 49136573.0, | |
| "step": 1510 | |
| }, | |
| { | |
| "entropy": 0.33220178112387655, | |
| "epoch": 6.756395995550612, | |
| "grad_norm": 0.14153026044368744, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2567, | |
| "mean_token_accuracy": 0.9304034173488617, | |
| "num_tokens": 49464253.0, | |
| "step": 1520 | |
| }, | |
| { | |
| "entropy": 0.32433032616972923, | |
| "epoch": 6.800889877641824, | |
| "grad_norm": 0.1524932086467743, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2532, | |
| "mean_token_accuracy": 0.931054875254631, | |
| "num_tokens": 49791933.0, | |
| "step": 1530 | |
| }, | |
| { | |
| "entropy": 0.3049290366470814, | |
| "epoch": 6.8453837597330365, | |
| "grad_norm": 0.1372746080160141, | |
| "learning_rate": 0.0002, | |
| "loss": 0.231, | |
| "mean_token_accuracy": 0.9373004108667373, | |
| "num_tokens": 50119504.0, | |
| "step": 1540 | |
| }, | |
| { | |
| "entropy": 0.41487769819796083, | |
| "epoch": 6.889877641824249, | |
| "grad_norm": 0.12199725210666656, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3525, | |
| "mean_token_accuracy": 0.9089711382985115, | |
| "num_tokens": 50440945.0, | |
| "step": 1550 | |
| }, | |
| { | |
| "entropy": 0.3157193776220083, | |
| "epoch": 6.934371523915462, | |
| "grad_norm": 0.1375955194234848, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2483, | |
| "mean_token_accuracy": 0.931727260351181, | |
| "num_tokens": 50768625.0, | |
| "step": 1560 | |
| }, | |
| { | |
| "entropy": 0.32167453654110434, | |
| "epoch": 6.978865406006674, | |
| "grad_norm": 0.13950124382972717, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2491, | |
| "mean_token_accuracy": 0.9321405410766601, | |
| "num_tokens": 51096247.0, | |
| "step": 1570 | |
| }, | |
| { | |
| "entropy": 0.3727147991840656, | |
| "epoch": 7.022246941045606, | |
| "grad_norm": 0.1275571882724762, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3099, | |
| "mean_token_accuracy": 0.9235704464790149, | |
| "num_tokens": 51406087.0, | |
| "step": 1580 | |
| }, | |
| { | |
| "entropy": 0.2698149785399437, | |
| "epoch": 7.0667408231368185, | |
| "grad_norm": 0.11617311090230942, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2024, | |
| "mean_token_accuracy": 0.9503593191504478, | |
| "num_tokens": 51733767.0, | |
| "step": 1590 | |
| }, | |
| { | |
| "entropy": 0.27310304902493954, | |
| "epoch": 7.111234705228031, | |
| "grad_norm": 0.13043957948684692, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1976, | |
| "mean_token_accuracy": 0.95060543268919, | |
| "num_tokens": 52061447.0, | |
| "step": 1600 | |
| }, | |
| { | |
| "entropy": 0.2578814696520567, | |
| "epoch": 7.155728587319244, | |
| "grad_norm": 0.12471891194581985, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1861, | |
| "mean_token_accuracy": 0.9535190463066101, | |
| "num_tokens": 52389107.0, | |
| "step": 1610 | |
| }, | |
| { | |
| "entropy": 0.27234206683933737, | |
| "epoch": 7.200222469410456, | |
| "grad_norm": 0.11977366358041763, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2019, | |
| "mean_token_accuracy": 0.9503579303622246, | |
| "num_tokens": 52716396.0, | |
| "step": 1620 | |
| }, | |
| { | |
| "entropy": 0.3498981598764658, | |
| "epoch": 7.244716351501668, | |
| "grad_norm": 0.13062220811843872, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2855, | |
| "mean_token_accuracy": 0.9283822610974312, | |
| "num_tokens": 53034172.0, | |
| "step": 1630 | |
| }, | |
| { | |
| "entropy": 0.28183968141675, | |
| "epoch": 7.289210233592881, | |
| "grad_norm": 0.13094866275787354, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2094, | |
| "mean_token_accuracy": 0.9477073162794113, | |
| "num_tokens": 53361852.0, | |
| "step": 1640 | |
| }, | |
| { | |
| "entropy": 0.2699905507266521, | |
| "epoch": 7.333704115684093, | |
| "grad_norm": 0.13447000086307526, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1977, | |
| "mean_token_accuracy": 0.9505031540989876, | |
| "num_tokens": 53689532.0, | |
| "step": 1650 | |
| }, | |
| { | |
| "entropy": 0.2764996990561485, | |
| "epoch": 7.378197997775306, | |
| "grad_norm": 0.12072061747312546, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2087, | |
| "mean_token_accuracy": 0.9473500683903694, | |
| "num_tokens": 54017194.0, | |
| "step": 1660 | |
| }, | |
| { | |
| "entropy": 0.26075094416737554, | |
| "epoch": 7.422691879866519, | |
| "grad_norm": 0.13177797198295593, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1886, | |
| "mean_token_accuracy": 0.9532095476984978, | |
| "num_tokens": 54344528.0, | |
| "step": 1670 | |
| }, | |
| { | |
| "entropy": 0.3465573750436306, | |
| "epoch": 7.46718576195773, | |
| "grad_norm": 0.14259563386440277, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2826, | |
| "mean_token_accuracy": 0.9294228002429008, | |
| "num_tokens": 54665549.0, | |
| "step": 1680 | |
| }, | |
| { | |
| "entropy": 0.2741903692483902, | |
| "epoch": 7.511679644048943, | |
| "grad_norm": 0.1369532197713852, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2043, | |
| "mean_token_accuracy": 0.9481862053275109, | |
| "num_tokens": 54993229.0, | |
| "step": 1690 | |
| }, | |
| { | |
| "entropy": 0.28000511340796946, | |
| "epoch": 7.556173526140156, | |
| "grad_norm": 0.1311381757259369, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2101, | |
| "mean_token_accuracy": 0.9463398531079292, | |
| "num_tokens": 55320909.0, | |
| "step": 1700 | |
| }, | |
| { | |
| "entropy": 0.27105078287422657, | |
| "epoch": 7.600667408231368, | |
| "grad_norm": 0.1463298499584198, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2019, | |
| "mean_token_accuracy": 0.9486314669251442, | |
| "num_tokens": 55648555.0, | |
| "step": 1710 | |
| }, | |
| { | |
| "entropy": 0.27910977341234683, | |
| "epoch": 7.645161290322581, | |
| "grad_norm": 0.12350628525018692, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2108, | |
| "mean_token_accuracy": 0.9471294045448303, | |
| "num_tokens": 55975883.0, | |
| "step": 1720 | |
| }, | |
| { | |
| "entropy": 0.36826528087258337, | |
| "epoch": 7.689655172413794, | |
| "grad_norm": 0.14085908234119415, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3072, | |
| "mean_token_accuracy": 0.9231738716363906, | |
| "num_tokens": 56297600.0, | |
| "step": 1730 | |
| }, | |
| { | |
| "entropy": 0.27950247190892696, | |
| "epoch": 7.734149054505005, | |
| "grad_norm": 0.13358551263809204, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2123, | |
| "mean_token_accuracy": 0.9456962540745735, | |
| "num_tokens": 56625280.0, | |
| "step": 1740 | |
| }, | |
| { | |
| "entropy": 0.2891340095549822, | |
| "epoch": 7.778642936596218, | |
| "grad_norm": 0.12801247835159302, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2208, | |
| "mean_token_accuracy": 0.9434458807110786, | |
| "num_tokens": 56952960.0, | |
| "step": 1750 | |
| }, | |
| { | |
| "entropy": 0.26483893804252145, | |
| "epoch": 7.823136818687431, | |
| "grad_norm": 0.1359819769859314, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1942, | |
| "mean_token_accuracy": 0.9500894829630852, | |
| "num_tokens": 57280629.0, | |
| "step": 1760 | |
| }, | |
| { | |
| "entropy": 0.30607642270624635, | |
| "epoch": 7.867630700778643, | |
| "grad_norm": 0.1317957043647766, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2363, | |
| "mean_token_accuracy": 0.9401613309979439, | |
| "num_tokens": 57607998.0, | |
| "step": 1770 | |
| }, | |
| { | |
| "entropy": 0.3491332869976759, | |
| "epoch": 7.912124582869856, | |
| "grad_norm": 0.15188415348529816, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2874, | |
| "mean_token_accuracy": 0.9255853906273842, | |
| "num_tokens": 57923119.0, | |
| "step": 1780 | |
| }, | |
| { | |
| "entropy": 0.290904039517045, | |
| "epoch": 7.956618464961068, | |
| "grad_norm": 0.13923445343971252, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2204, | |
| "mean_token_accuracy": 0.9437047794461251, | |
| "num_tokens": 58250799.0, | |
| "step": 1790 | |
| }, | |
| { | |
| "entropy": 0.31344421742818296, | |
| "epoch": 8.0, | |
| "grad_norm": 0.1466398537158966, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2446, | |
| "mean_token_accuracy": 0.9374472070962955, | |
| "num_tokens": 58562568.0, | |
| "step": 1800 | |
| }, | |
| { | |
| "entropy": 0.24779046475887298, | |
| "epoch": 8.044493882091212, | |
| "grad_norm": 0.13469737768173218, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1839, | |
| "mean_token_accuracy": 0.9569739550352097, | |
| "num_tokens": 58890248.0, | |
| "step": 1810 | |
| }, | |
| { | |
| "entropy": 0.247482842579484, | |
| "epoch": 8.088987764182425, | |
| "grad_norm": 0.11496013402938843, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1769, | |
| "mean_token_accuracy": 0.9584277510643006, | |
| "num_tokens": 59217928.0, | |
| "step": 1820 | |
| }, | |
| { | |
| "entropy": 0.23934231325984, | |
| "epoch": 8.133481646273637, | |
| "grad_norm": 0.11410822719335556, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1714, | |
| "mean_token_accuracy": 0.9596161171793938, | |
| "num_tokens": 59545608.0, | |
| "step": 1830 | |
| }, | |
| { | |
| "entropy": 0.23615190275013448, | |
| "epoch": 8.17797552836485, | |
| "grad_norm": 0.1125149056315422, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1712, | |
| "mean_token_accuracy": 0.9590911790728569, | |
| "num_tokens": 59873186.0, | |
| "step": 1840 | |
| }, | |
| { | |
| "entropy": 0.31146894954144955, | |
| "epoch": 8.222469410456062, | |
| "grad_norm": 0.11172334104776382, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2454, | |
| "mean_token_accuracy": 0.9410143241286277, | |
| "num_tokens": 60193844.0, | |
| "step": 1850 | |
| }, | |
| { | |
| "entropy": 0.2522707603871822, | |
| "epoch": 8.266963292547274, | |
| "grad_norm": 0.1153436079621315, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1891, | |
| "mean_token_accuracy": 0.9543200537562371, | |
| "num_tokens": 60521524.0, | |
| "step": 1860 | |
| }, | |
| { | |
| "entropy": 0.23912321217358112, | |
| "epoch": 8.311457174638488, | |
| "grad_norm": 0.11682998389005661, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1752, | |
| "mean_token_accuracy": 0.9578092664480209, | |
| "num_tokens": 60849204.0, | |
| "step": 1870 | |
| }, | |
| { | |
| "entropy": 0.23446308337152005, | |
| "epoch": 8.3559510567297, | |
| "grad_norm": 0.11366456747055054, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1715, | |
| "mean_token_accuracy": 0.9589567482471466, | |
| "num_tokens": 61176884.0, | |
| "step": 1880 | |
| }, | |
| { | |
| "entropy": 0.23881982676684857, | |
| "epoch": 8.400444938820913, | |
| "grad_norm": 0.12823918461799622, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1742, | |
| "mean_token_accuracy": 0.9584661111235618, | |
| "num_tokens": 61504420.0, | |
| "step": 1890 | |
| }, | |
| { | |
| "entropy": 0.3114849112927914, | |
| "epoch": 8.444938820912125, | |
| "grad_norm": 0.12093894183635712, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2493, | |
| "mean_token_accuracy": 0.9392594262957573, | |
| "num_tokens": 61822339.0, | |
| "step": 1900 | |
| }, | |
| { | |
| "entropy": 0.238529834151268, | |
| "epoch": 8.489432703003336, | |
| "grad_norm": 0.12336688488721848, | |
| "learning_rate": 0.0002, | |
| "loss": 0.178, | |
| "mean_token_accuracy": 0.9569561630487442, | |
| "num_tokens": 62150019.0, | |
| "step": 1910 | |
| }, | |
| { | |
| "entropy": 0.23640490621328353, | |
| "epoch": 8.53392658509455, | |
| "grad_norm": 0.1193024292588234, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1712, | |
| "mean_token_accuracy": 0.9583990573883057, | |
| "num_tokens": 62477699.0, | |
| "step": 1920 | |
| }, | |
| { | |
| "entropy": 0.2368166845291853, | |
| "epoch": 8.578420467185762, | |
| "grad_norm": 0.1242118701338768, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1697, | |
| "mean_token_accuracy": 0.9590780705213546, | |
| "num_tokens": 62805379.0, | |
| "step": 1930 | |
| }, | |
| { | |
| "entropy": 0.23472826182842255, | |
| "epoch": 8.622914349276975, | |
| "grad_norm": 0.1342175304889679, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1686, | |
| "mean_token_accuracy": 0.9593432620167732, | |
| "num_tokens": 63132961.0, | |
| "step": 1940 | |
| }, | |
| { | |
| "entropy": 0.3145587895065546, | |
| "epoch": 8.667408231368187, | |
| "grad_norm": 0.11396913975477219, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2531, | |
| "mean_token_accuracy": 0.937441036105156, | |
| "num_tokens": 63451244.0, | |
| "step": 1950 | |
| }, | |
| { | |
| "entropy": 0.24804712384939193, | |
| "epoch": 8.711902113459399, | |
| "grad_norm": 0.12833833694458008, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1878, | |
| "mean_token_accuracy": 0.9542065858840942, | |
| "num_tokens": 63778924.0, | |
| "step": 1960 | |
| }, | |
| { | |
| "entropy": 0.238729153200984, | |
| "epoch": 8.756395995550612, | |
| "grad_norm": 0.12618130445480347, | |
| "learning_rate": 0.0002, | |
| "loss": 0.173, | |
| "mean_token_accuracy": 0.9577292338013649, | |
| "num_tokens": 64106604.0, | |
| "step": 1970 | |
| }, | |
| { | |
| "entropy": 0.25689107105135917, | |
| "epoch": 8.800889877641824, | |
| "grad_norm": 0.12338786572217941, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1924, | |
| "mean_token_accuracy": 0.9526981726288796, | |
| "num_tokens": 64434284.0, | |
| "step": 1980 | |
| }, | |
| { | |
| "entropy": 0.24375672303140164, | |
| "epoch": 8.845383759733037, | |
| "grad_norm": 0.1313425898551941, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1841, | |
| "mean_token_accuracy": 0.9547993883490562, | |
| "num_tokens": 64761880.0, | |
| "step": 1990 | |
| }, | |
| { | |
| "entropy": 0.3665214736014605, | |
| "epoch": 8.889877641824249, | |
| "grad_norm": 0.13209204375743866, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3079, | |
| "mean_token_accuracy": 0.924127696454525, | |
| "num_tokens": 65080688.0, | |
| "step": 2000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2250, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.4704197278167695e+18, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |