{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.99749049974905, "eval_steps": 50, "global_step": 8715, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005736000573600057, "grad_norm": 2.3000271320343018, "learning_rate": 1.1467889908256882e-07, "loss": 0.3103, "step": 10 }, { "epoch": 0.011472001147200114, "grad_norm": 0.23891927301883698, "learning_rate": 2.2935779816513764e-07, "loss": 0.2101, "step": 20 }, { "epoch": 0.017208001720800174, "grad_norm": 3.7850193977355957, "learning_rate": 3.4403669724770646e-07, "loss": 0.3158, "step": 30 }, { "epoch": 0.022944002294400228, "grad_norm": 10.902544021606445, "learning_rate": 4.587155963302753e-07, "loss": 0.3107, "step": 40 }, { "epoch": 0.028680002868000286, "grad_norm": 12.402349472045898, "learning_rate": 5.733944954128441e-07, "loss": 0.3895, "step": 50 }, { "epoch": 0.028680002868000286, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 1.0573054552078247, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 17.8757, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 36.53, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 18.293, "step": 50 }, { "epoch": 0.03441600344160035, "grad_norm": 1.5937564373016357, "learning_rate": 6.880733944954129e-07, "loss": 0.2534, "step": 60 }, { "epoch": 0.0401520040152004, "grad_norm": 3.687685012817383, "learning_rate": 8.027522935779817e-07, "loss": 0.3795, "step": 70 }, { "epoch": 0.045888004588800456, "grad_norm": 0.03578287363052368, "learning_rate": 9.174311926605506e-07, "loss": 0.4276, "step": 80 }, { "epoch": 0.05162400516240052, "grad_norm": 0.022680282592773438, "learning_rate": 1.0321100917431195e-06, "loss": 0.277, "step": 90 }, { "epoch": 0.05736000573600057, "grad_norm": 1.7502835988998413, "learning_rate": 1.1467889908256882e-06, "loss": 0.2279, "step": 100 }, { "epoch": 0.05736000573600057, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.9872124791145325, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.223, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.768, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.408, "step": 100 }, { "epoch": 0.06309600630960063, "grad_norm": 4.763681411743164, "learning_rate": 1.261467889908257e-06, "loss": 0.3441, "step": 110 }, { "epoch": 0.0688320068832007, "grad_norm": 7.319093704223633, "learning_rate": 1.3761467889908258e-06, "loss": 0.4273, "step": 120 }, { "epoch": 0.07456800745680074, "grad_norm": 9.32235336303711, "learning_rate": 1.4908256880733945e-06, "loss": 0.3338, "step": 130 }, { "epoch": 0.0803040080304008, "grad_norm": 2.0044586658477783, "learning_rate": 1.6055045871559635e-06, "loss": 0.1364, "step": 140 }, { "epoch": 0.08604000860400086, "grad_norm": 10.962570190429688, "learning_rate": 1.7201834862385322e-06, "loss": 0.2404, "step": 150 }, { "epoch": 0.08604000860400086, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.7815765738487244, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2921, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.669, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.358, "step": 150 }, { "epoch": 0.09177600917760091, "grad_norm": 4.845127105712891, "learning_rate": 1.8348623853211011e-06, "loss": 0.3601, "step": 160 }, { "epoch": 0.09751200975120097, "grad_norm": 3.165642738342285, "learning_rate": 1.94954128440367e-06, "loss": 0.1938, "step": 170 }, { "epoch": 0.10324801032480103, "grad_norm": 3.4738926887512207, "learning_rate": 2.064220183486239e-06, "loss": 0.1979, "step": 180 }, { "epoch": 0.1089840108984011, "grad_norm": 0.34281352162361145, "learning_rate": 2.1788990825688075e-06, "loss": 0.2278, "step": 190 }, { "epoch": 0.11472001147200114, "grad_norm": 9.631810188293457, "learning_rate": 2.2935779816513764e-06, "loss": 0.1642, "step": 200 }, { "epoch": 0.11472001147200114, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.32764628529548645, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1681, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.848, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.448, "step": 200 }, { "epoch": 0.1204560120456012, "grad_norm": 4.584328651428223, "learning_rate": 2.4082568807339453e-06, "loss": 0.1721, "step": 210 }, { "epoch": 0.12619201261920127, "grad_norm": 7.435701370239258, "learning_rate": 2.522935779816514e-06, "loss": 0.1632, "step": 220 }, { "epoch": 0.1319280131928013, "grad_norm": 2.957846164703369, "learning_rate": 2.6376146788990823e-06, "loss": 0.1175, "step": 230 }, { "epoch": 0.1376640137664014, "grad_norm": 4.342548847198486, "learning_rate": 2.7522935779816517e-06, "loss": 0.0877, "step": 240 }, { "epoch": 0.14340001434000144, "grad_norm": 2.2145071029663086, "learning_rate": 2.8669724770642206e-06, "loss": 0.0692, "step": 250 }, { "epoch": 0.14340001434000144, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.14406049251556396, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2632, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.71, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.379, "step": 250 }, { "epoch": 0.14913601491360148, "grad_norm": 0.3610692620277405, "learning_rate": 2.981651376146789e-06, "loss": 0.1079, "step": 260 }, { "epoch": 0.15487201548720156, "grad_norm": 6.997246265411377, "learning_rate": 3.0963302752293576e-06, "loss": 0.0748, "step": 270 }, { "epoch": 0.1606080160608016, "grad_norm": 2.0748515129089355, "learning_rate": 3.211009174311927e-06, "loss": 0.0749, "step": 280 }, { "epoch": 0.16634401663440165, "grad_norm": 3.6645171642303467, "learning_rate": 3.325688073394496e-06, "loss": 0.097, "step": 290 }, { "epoch": 0.17208001720800173, "grad_norm": 8.085243225097656, "learning_rate": 3.4403669724770644e-06, "loss": 0.0722, "step": 300 }, { "epoch": 0.17208001720800173, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11341650038957596, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2959, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.663, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.355, "step": 300 }, { "epoch": 0.17781601778160178, "grad_norm": 2.7256767749786377, "learning_rate": 3.5550458715596333e-06, "loss": 0.0707, "step": 310 }, { "epoch": 0.18355201835520182, "grad_norm": 2.6183135509490967, "learning_rate": 3.6697247706422022e-06, "loss": 0.0418, "step": 320 }, { "epoch": 0.1892880189288019, "grad_norm": 1.0414828062057495, "learning_rate": 3.784403669724771e-06, "loss": 0.041, "step": 330 }, { "epoch": 0.19502401950240195, "grad_norm": 1.9375965595245361, "learning_rate": 3.89908256880734e-06, "loss": 0.0742, "step": 340 }, { "epoch": 0.200760020076002, "grad_norm": 1.6328781843185425, "learning_rate": 4.013761467889909e-06, "loss": 0.0426, "step": 350 }, { "epoch": 0.200760020076002, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11491454392671585, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2047, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.795, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.421, "step": 350 }, { "epoch": 0.20649602064960207, "grad_norm": 0.45767122507095337, "learning_rate": 4.128440366972478e-06, "loss": 0.0771, "step": 360 }, { "epoch": 0.21223202122320212, "grad_norm": 2.768129825592041, "learning_rate": 4.2431192660550464e-06, "loss": 0.0869, "step": 370 }, { "epoch": 0.2179680217968022, "grad_norm": 1.4702483415603638, "learning_rate": 4.357798165137615e-06, "loss": 0.0536, "step": 380 }, { "epoch": 0.22370402237040224, "grad_norm": 1.277462124824524, "learning_rate": 4.4724770642201834e-06, "loss": 0.0329, "step": 390 }, { "epoch": 0.2294400229440023, "grad_norm": 1.8748779296875, "learning_rate": 4.587155963302753e-06, "loss": 0.045, "step": 400 }, { "epoch": 0.2294400229440023, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.12494589388370514, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2877, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.675, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.361, "step": 400 }, { "epoch": 0.23517602351760236, "grad_norm": 2.339601516723633, "learning_rate": 4.701834862385321e-06, "loss": 0.0546, "step": 410 }, { "epoch": 0.2409120240912024, "grad_norm": 0.9202137589454651, "learning_rate": 4.816513761467891e-06, "loss": 0.1038, "step": 420 }, { "epoch": 0.24664802466480246, "grad_norm": 2.0663821697235107, "learning_rate": 4.931192660550459e-06, "loss": 0.0418, "step": 430 }, { "epoch": 0.25238402523840253, "grad_norm": 2.6615207195281982, "learning_rate": 5.045871559633028e-06, "loss": 0.0414, "step": 440 }, { "epoch": 0.2581200258120026, "grad_norm": 2.298173427581787, "learning_rate": 5.160550458715596e-06, "loss": 0.063, "step": 450 }, { "epoch": 0.2581200258120026, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.1194896399974823, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3177, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.632, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.339, "step": 450 }, { "epoch": 0.2638560263856026, "grad_norm": 0.9054094552993774, "learning_rate": 5.275229357798165e-06, "loss": 0.0572, "step": 460 }, { "epoch": 0.2695920269592027, "grad_norm": 0.5312202572822571, "learning_rate": 5.389908256880735e-06, "loss": 0.0697, "step": 470 }, { "epoch": 0.2753280275328028, "grad_norm": 1.5197083950042725, "learning_rate": 5.504587155963303e-06, "loss": 0.047, "step": 480 }, { "epoch": 0.2810640281064028, "grad_norm": 1.341577410697937, "learning_rate": 5.619266055045872e-06, "loss": 0.0285, "step": 490 }, { "epoch": 0.28680002868000287, "grad_norm": 1.6182245016098022, "learning_rate": 5.733944954128441e-06, "loss": 0.0402, "step": 500 }, { "epoch": 0.28680002868000287, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.12436391413211823, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3365, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.605, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.326, "step": 500 }, { "epoch": 0.29253602925360295, "grad_norm": 2.623664140701294, "learning_rate": 5.84862385321101e-06, "loss": 0.0552, "step": 510 }, { "epoch": 0.29827202982720297, "grad_norm": 1.9711512327194214, "learning_rate": 5.963302752293578e-06, "loss": 0.0534, "step": 520 }, { "epoch": 0.30400803040080304, "grad_norm": 1.4279474020004272, "learning_rate": 6.077981651376147e-06, "loss": 0.0356, "step": 530 }, { "epoch": 0.3097440309744031, "grad_norm": 1.5085383653640747, "learning_rate": 6.192660550458715e-06, "loss": 0.0389, "step": 540 }, { "epoch": 0.31548003154800314, "grad_norm": 1.8550148010253906, "learning_rate": 6.307339449541285e-06, "loss": 0.0342, "step": 550 }, { "epoch": 0.31548003154800314, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.16011035442352295, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2066, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.792, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.42, "step": 550 }, { "epoch": 0.3212160321216032, "grad_norm": 0.8617078065872192, "learning_rate": 6.422018348623854e-06, "loss": 0.0351, "step": 560 }, { "epoch": 0.3269520326952033, "grad_norm": 1.7970987558364868, "learning_rate": 6.536697247706422e-06, "loss": 0.0328, "step": 570 }, { "epoch": 0.3326880332688033, "grad_norm": 1.319393515586853, "learning_rate": 6.651376146788992e-06, "loss": 0.0348, "step": 580 }, { "epoch": 0.3384240338424034, "grad_norm": 3.622856855392456, "learning_rate": 6.76605504587156e-06, "loss": 0.035, "step": 590 }, { "epoch": 0.34416003441600346, "grad_norm": 2.2089531421661377, "learning_rate": 6.880733944954129e-06, "loss": 0.0485, "step": 600 }, { "epoch": 0.34416003441600346, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.12377385795116425, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2989, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.659, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.353, "step": 600 }, { "epoch": 0.3498960349896035, "grad_norm": 1.3465012311935425, "learning_rate": 6.995412844036697e-06, "loss": 0.0448, "step": 610 }, { "epoch": 0.35563203556320355, "grad_norm": 1.0101594924926758, "learning_rate": 7.110091743119267e-06, "loss": 0.0298, "step": 620 }, { "epoch": 0.3613680361368036, "grad_norm": 3.003561019897461, "learning_rate": 7.224770642201836e-06, "loss": 0.052, "step": 630 }, { "epoch": 0.36710403671040365, "grad_norm": 2.2370080947875977, "learning_rate": 7.3394495412844045e-06, "loss": 0.0241, "step": 640 }, { "epoch": 0.3728400372840037, "grad_norm": 3.373074531555176, "learning_rate": 7.454128440366973e-06, "loss": 0.0446, "step": 650 }, { "epoch": 0.3728400372840037, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11963505297899246, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1731, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.841, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.444, "step": 650 }, { "epoch": 0.3785760378576038, "grad_norm": 0.6863628029823303, "learning_rate": 7.568807339449542e-06, "loss": 0.0116, "step": 660 }, { "epoch": 0.3843120384312038, "grad_norm": 2.9094369411468506, "learning_rate": 7.68348623853211e-06, "loss": 0.0657, "step": 670 }, { "epoch": 0.3900480390048039, "grad_norm": 0.2047434151172638, "learning_rate": 7.79816513761468e-06, "loss": 0.0319, "step": 680 }, { "epoch": 0.39578403957840397, "grad_norm": 1.523390531539917, "learning_rate": 7.912844036697249e-06, "loss": 0.0492, "step": 690 }, { "epoch": 0.401520040152004, "grad_norm": 0.10014928132295609, "learning_rate": 8.027522935779817e-06, "loss": 0.0333, "step": 700 }, { "epoch": 0.401520040152004, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10763221979141235, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.342, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.597, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.322, "step": 700 }, { "epoch": 0.40725604072560406, "grad_norm": 0.24312058091163635, "learning_rate": 8.142201834862386e-06, "loss": 0.0248, "step": 710 }, { "epoch": 0.41299204129920414, "grad_norm": 0.08946532011032104, "learning_rate": 8.256880733944956e-06, "loss": 0.0133, "step": 720 }, { "epoch": 0.4187280418728042, "grad_norm": 0.13466329872608185, "learning_rate": 8.371559633027524e-06, "loss": 0.0226, "step": 730 }, { "epoch": 0.42446404244640423, "grad_norm": 0.4071923792362213, "learning_rate": 8.486238532110093e-06, "loss": 0.0353, "step": 740 }, { "epoch": 0.4302000430200043, "grad_norm": 1.2394853830337524, "learning_rate": 8.600917431192661e-06, "loss": 0.0199, "step": 750 }, { "epoch": 0.4302000430200043, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09811359643936157, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3072, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.647, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.347, "step": 750 }, { "epoch": 0.4359360435936044, "grad_norm": 2.9703874588012695, "learning_rate": 8.71559633027523e-06, "loss": 0.0259, "step": 760 }, { "epoch": 0.4416720441672044, "grad_norm": 1.3356784582138062, "learning_rate": 8.830275229357798e-06, "loss": 0.0496, "step": 770 }, { "epoch": 0.4474080447408045, "grad_norm": 0.7478968501091003, "learning_rate": 8.944954128440367e-06, "loss": 0.0383, "step": 780 }, { "epoch": 0.45314404531440455, "grad_norm": 0.418811172246933, "learning_rate": 9.059633027522935e-06, "loss": 0.0173, "step": 790 }, { "epoch": 0.4588800458880046, "grad_norm": 0.6638255715370178, "learning_rate": 9.174311926605506e-06, "loss": 0.0271, "step": 800 }, { "epoch": 0.4588800458880046, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11180277168750763, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2311, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.757, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.402, "step": 800 }, { "epoch": 0.46461604646160465, "grad_norm": 0.21615058183670044, "learning_rate": 9.288990825688074e-06, "loss": 0.0175, "step": 810 }, { "epoch": 0.4703520470352047, "grad_norm": 1.2093310356140137, "learning_rate": 9.403669724770643e-06, "loss": 0.0215, "step": 820 }, { "epoch": 0.47608804760880474, "grad_norm": 0.2247300148010254, "learning_rate": 9.518348623853211e-06, "loss": 0.0486, "step": 830 }, { "epoch": 0.4818240481824048, "grad_norm": 0.6988762617111206, "learning_rate": 9.633027522935781e-06, "loss": 0.0161, "step": 840 }, { "epoch": 0.4875600487560049, "grad_norm": 1.2162249088287354, "learning_rate": 9.74770642201835e-06, "loss": 0.0277, "step": 850 }, { "epoch": 0.4875600487560049, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08368796855211258, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2199, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.773, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.41, "step": 850 }, { "epoch": 0.4932960493296049, "grad_norm": 1.0620274543762207, "learning_rate": 9.862385321100918e-06, "loss": 0.0566, "step": 860 }, { "epoch": 0.499032049903205, "grad_norm": 0.23082388937473297, "learning_rate": 9.977064220183487e-06, "loss": 0.0247, "step": 870 }, { "epoch": 0.5047680504768051, "grad_norm": 0.9556642770767212, "learning_rate": 9.999974328282418e-06, "loss": 0.0242, "step": 880 }, { "epoch": 0.5105040510504051, "grad_norm": 0.6304249167442322, "learning_rate": 9.999870037381541e-06, "loss": 0.0268, "step": 890 }, { "epoch": 0.5162400516240052, "grad_norm": 0.07784967124462128, "learning_rate": 9.99968552448706e-06, "loss": 0.033, "step": 900 }, { "epoch": 0.5162400516240052, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06569766253232956, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2473, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.733, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.39, "step": 900 }, { "epoch": 0.5219760521976052, "grad_norm": 1.320080280303955, "learning_rate": 9.999420792559453e-06, "loss": 0.0623, "step": 910 }, { "epoch": 0.5277120527712053, "grad_norm": 0.24600175023078918, "learning_rate": 9.999075845846292e-06, "loss": 0.022, "step": 920 }, { "epoch": 0.5334480533448054, "grad_norm": 1.860329031944275, "learning_rate": 9.998650689882184e-06, "loss": 0.0186, "step": 930 }, { "epoch": 0.5391840539184054, "grad_norm": 1.9470009803771973, "learning_rate": 9.99814533148868e-06, "loss": 0.023, "step": 940 }, { "epoch": 0.5449200544920054, "grad_norm": 0.3151063919067383, "learning_rate": 9.997559778774162e-06, "loss": 0.0111, "step": 950 }, { "epoch": 0.5449200544920054, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07853744179010391, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2659, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.706, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.377, "step": 950 }, { "epoch": 0.5506560550656056, "grad_norm": 1.4308031797409058, "learning_rate": 9.996894041133715e-06, "loss": 0.0239, "step": 960 }, { "epoch": 0.5563920556392056, "grad_norm": 0.32267820835113525, "learning_rate": 9.996148129248975e-06, "loss": 0.0365, "step": 970 }, { "epoch": 0.5621280562128056, "grad_norm": 0.026730485260486603, "learning_rate": 9.995322055087963e-06, "loss": 0.0255, "step": 980 }, { "epoch": 0.5678640567864057, "grad_norm": 4.146325588226318, "learning_rate": 9.994415831904886e-06, "loss": 0.06, "step": 990 }, { "epoch": 0.5736000573600057, "grad_norm": 0.9087633490562439, "learning_rate": 9.993429474239928e-06, "loss": 0.0371, "step": 1000 }, { "epoch": 0.5736000573600057, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08875266462564468, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2636, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.71, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.378, "step": 1000 }, { "epoch": 0.5793360579336058, "grad_norm": 1.1552188396453857, "learning_rate": 9.992362997919016e-06, "loss": 0.0419, "step": 1010 }, { "epoch": 0.5850720585072059, "grad_norm": 0.8535066246986389, "learning_rate": 9.991216420053565e-06, "loss": 0.0162, "step": 1020 }, { "epoch": 0.5908080590808059, "grad_norm": 1.3267903327941895, "learning_rate": 9.989989759040208e-06, "loss": 0.0375, "step": 1030 }, { "epoch": 0.5965440596544059, "grad_norm": 2.629297971725464, "learning_rate": 9.988683034560494e-06, "loss": 0.0097, "step": 1040 }, { "epoch": 0.6022800602280061, "grad_norm": 2.267737627029419, "learning_rate": 9.987296267580575e-06, "loss": 0.0267, "step": 1050 }, { "epoch": 0.6022800602280061, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10149678587913513, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2648, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.708, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.377, "step": 1050 }, { "epoch": 0.6080160608016061, "grad_norm": 0.04809356853365898, "learning_rate": 9.985829480350874e-06, "loss": 0.0104, "step": 1060 }, { "epoch": 0.6137520613752061, "grad_norm": 2.897101640701294, "learning_rate": 9.98428269640572e-06, "loss": 0.0313, "step": 1070 }, { "epoch": 0.6194880619488062, "grad_norm": 0.2844018340110779, "learning_rate": 9.982655940562978e-06, "loss": 0.0202, "step": 1080 }, { "epoch": 0.6252240625224063, "grad_norm": 1.7705107927322388, "learning_rate": 9.980949238923646e-06, "loss": 0.0151, "step": 1090 }, { "epoch": 0.6309600630960063, "grad_norm": 2.162916421890259, "learning_rate": 9.979162618871435e-06, "loss": 0.021, "step": 1100 }, { "epoch": 0.6309600630960063, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.087111696600914, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3134, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.638, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.342, "step": 1100 }, { "epoch": 0.6366960636696064, "grad_norm": 3.3192200660705566, "learning_rate": 9.977296109072337e-06, "loss": 0.0182, "step": 1110 }, { "epoch": 0.6424320642432064, "grad_norm": 0.3482876718044281, "learning_rate": 9.975349739474156e-06, "loss": 0.0309, "step": 1120 }, { "epoch": 0.6481680648168064, "grad_norm": 2.682620048522949, "learning_rate": 9.973323541306032e-06, "loss": 0.0299, "step": 1130 }, { "epoch": 0.6539040653904066, "grad_norm": 2.5474138259887695, "learning_rate": 9.971217547077947e-06, "loss": 0.0337, "step": 1140 }, { "epoch": 0.6596400659640066, "grad_norm": 0.5940343141555786, "learning_rate": 9.969031790580185e-06, "loss": 0.0294, "step": 1150 }, { "epoch": 0.6596400659640066, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07855822145938873, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1855, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.823, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.435, "step": 1150 }, { "epoch": 0.6653760665376066, "grad_norm": 1.4965182542800903, "learning_rate": 9.966766306882811e-06, "loss": 0.0257, "step": 1160 }, { "epoch": 0.6711120671112067, "grad_norm": 0.10210319608449936, "learning_rate": 9.964421132335091e-06, "loss": 0.0214, "step": 1170 }, { "epoch": 0.6768480676848068, "grad_norm": 1.594895362854004, "learning_rate": 9.961996304564916e-06, "loss": 0.0182, "step": 1180 }, { "epoch": 0.6825840682584068, "grad_norm": 1.8244154453277588, "learning_rate": 9.959491862478206e-06, "loss": 0.0195, "step": 1190 }, { "epoch": 0.6883200688320069, "grad_norm": 3.9240288734436035, "learning_rate": 9.956907846258268e-06, "loss": 0.0426, "step": 1200 }, { "epoch": 0.6883200688320069, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08573708683252335, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1456, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.881, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.464, "step": 1200 }, { "epoch": 0.6940560694056069, "grad_norm": 0.06307424604892731, "learning_rate": 9.954244297365169e-06, "loss": 0.0592, "step": 1210 }, { "epoch": 0.699792069979207, "grad_norm": 0.44088277220726013, "learning_rate": 9.951501258535061e-06, "loss": 0.0267, "step": 1220 }, { "epoch": 0.7055280705528071, "grad_norm": 2.217980146408081, "learning_rate": 9.948678773779495e-06, "loss": 0.0171, "step": 1230 }, { "epoch": 0.7112640711264071, "grad_norm": 1.3680331707000732, "learning_rate": 9.94577688838472e-06, "loss": 0.0369, "step": 1240 }, { "epoch": 0.7170000717000071, "grad_norm": 0.043984800577163696, "learning_rate": 9.94279564891095e-06, "loss": 0.0173, "step": 1250 }, { "epoch": 0.7170000717000071, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06791304796934128, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1924, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.813, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.43, "step": 1250 }, { "epoch": 0.7227360722736073, "grad_norm": 0.05398892983794212, "learning_rate": 9.939735103191625e-06, "loss": 0.0151, "step": 1260 }, { "epoch": 0.7284720728472073, "grad_norm": 0.26481369137763977, "learning_rate": 9.936595300332638e-06, "loss": 0.0242, "step": 1270 }, { "epoch": 0.7342080734208073, "grad_norm": 0.0726240873336792, "learning_rate": 9.933376290711546e-06, "loss": 0.0128, "step": 1280 }, { "epoch": 0.7399440739944074, "grad_norm": 0.21732382476329803, "learning_rate": 9.930078125976767e-06, "loss": 0.0268, "step": 1290 }, { "epoch": 0.7456800745680074, "grad_norm": 1.968591332435608, "learning_rate": 9.926700859046745e-06, "loss": 0.0264, "step": 1300 }, { "epoch": 0.7456800745680074, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08293525129556656, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2805, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.685, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.366, "step": 1300 }, { "epoch": 0.7514160751416075, "grad_norm": 0.12683559954166412, "learning_rate": 9.923244544109107e-06, "loss": 0.0107, "step": 1310 }, { "epoch": 0.7571520757152076, "grad_norm": 0.07375319302082062, "learning_rate": 9.919709236619786e-06, "loss": 0.0237, "step": 1320 }, { "epoch": 0.7628880762888076, "grad_norm": 0.09353283792734146, "learning_rate": 9.916094993302144e-06, "loss": 0.0247, "step": 1330 }, { "epoch": 0.7686240768624076, "grad_norm": 2.4819159507751465, "learning_rate": 9.912401872146047e-06, "loss": 0.0361, "step": 1340 }, { "epoch": 0.7743600774360078, "grad_norm": 1.608819842338562, "learning_rate": 9.908629932406943e-06, "loss": 0.0123, "step": 1350 }, { "epoch": 0.7743600774360078, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11693766713142395, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2248, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.766, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.407, "step": 1350 }, { "epoch": 0.7800960780096078, "grad_norm": 0.044521648436784744, "learning_rate": 9.90477923460491e-06, "loss": 0.0559, "step": 1360 }, { "epoch": 0.7858320785832078, "grad_norm": 1.6780263185501099, "learning_rate": 9.900849840523686e-06, "loss": 0.016, "step": 1370 }, { "epoch": 0.7915680791568079, "grad_norm": 0.20249976217746735, "learning_rate": 9.896841813209675e-06, "loss": 0.0103, "step": 1380 }, { "epoch": 0.797304079730408, "grad_norm": 0.5444568991661072, "learning_rate": 9.892755216970937e-06, "loss": 0.0265, "step": 1390 }, { "epoch": 0.803040080304008, "grad_norm": 0.39745691418647766, "learning_rate": 9.888590117376154e-06, "loss": 0.0305, "step": 1400 }, { "epoch": 0.803040080304008, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07220673561096191, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2419, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.741, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.394, "step": 1400 }, { "epoch": 0.8087760808776081, "grad_norm": 3.1255667209625244, "learning_rate": 9.884346581253584e-06, "loss": 0.0155, "step": 1410 }, { "epoch": 0.8145120814512081, "grad_norm": 0.28445881605148315, "learning_rate": 9.88002467668998e-06, "loss": 0.0543, "step": 1420 }, { "epoch": 0.8202480820248083, "grad_norm": 2.316286087036133, "learning_rate": 9.875624473029508e-06, "loss": 0.0208, "step": 1430 }, { "epoch": 0.8259840825984083, "grad_norm": 0.3444057106971741, "learning_rate": 9.871146040872621e-06, "loss": 0.0375, "step": 1440 }, { "epoch": 0.8317200831720083, "grad_norm": 0.36751267313957214, "learning_rate": 9.86658945207494e-06, "loss": 0.0273, "step": 1450 }, { "epoch": 0.8317200831720083, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06544554233551025, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1664, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.851, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.449, "step": 1450 }, { "epoch": 0.8374560837456084, "grad_norm": 0.6214566230773926, "learning_rate": 9.861954779746092e-06, "loss": 0.0095, "step": 1460 }, { "epoch": 0.8431920843192084, "grad_norm": 0.6569352149963379, "learning_rate": 9.857242098248543e-06, "loss": 0.024, "step": 1470 }, { "epoch": 0.8489280848928085, "grad_norm": 1.6928707361221313, "learning_rate": 9.852451483196394e-06, "loss": 0.0158, "step": 1480 }, { "epoch": 0.8546640854664086, "grad_norm": 1.0950285196304321, "learning_rate": 9.847583011454187e-06, "loss": 0.0099, "step": 1490 }, { "epoch": 0.8604000860400086, "grad_norm": 1.4251108169555664, "learning_rate": 9.842636761135651e-06, "loss": 0.0677, "step": 1500 }, { "epoch": 0.8604000860400086, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11534915119409561, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2226, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.769, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.408, "step": 1500 }, { "epoch": 0.8661360866136086, "grad_norm": 0.9497309923171997, "learning_rate": 9.837612811602462e-06, "loss": 0.0282, "step": 1510 }, { "epoch": 0.8718720871872088, "grad_norm": 0.9512405395507812, "learning_rate": 9.832511243462962e-06, "loss": 0.0079, "step": 1520 }, { "epoch": 0.8776080877608088, "grad_norm": 0.04131248593330383, "learning_rate": 9.827332138570878e-06, "loss": 0.0051, "step": 1530 }, { "epoch": 0.8833440883344088, "grad_norm": 1.3209096193313599, "learning_rate": 9.822075580023987e-06, "loss": 0.0132, "step": 1540 }, { "epoch": 0.8890800889080089, "grad_norm": 0.05282355472445488, "learning_rate": 9.816741652162807e-06, "loss": 0.0592, "step": 1550 }, { "epoch": 0.8890800889080089, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08604957163333893, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3059, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.649, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.348, "step": 1550 }, { "epoch": 0.894816089481609, "grad_norm": 0.5249603390693665, "learning_rate": 9.811330440569226e-06, "loss": 0.0169, "step": 1560 }, { "epoch": 0.900552090055209, "grad_norm": 1.4722495079040527, "learning_rate": 9.805842032065135e-06, "loss": 0.0257, "step": 1570 }, { "epoch": 0.9062880906288091, "grad_norm": 1.664174199104309, "learning_rate": 9.800276514711044e-06, "loss": 0.0318, "step": 1580 }, { "epoch": 0.9120240912024091, "grad_norm": 0.017392676323652267, "learning_rate": 9.794633977804646e-06, "loss": 0.0165, "step": 1590 }, { "epoch": 0.9177600917760091, "grad_norm": 0.18038752675056458, "learning_rate": 9.788914511879412e-06, "loss": 0.0488, "step": 1600 }, { "epoch": 0.9177600917760091, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08364205807447433, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2219, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.77, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.409, "step": 1600 }, { "epoch": 0.9234960923496093, "grad_norm": 1.785083293914795, "learning_rate": 9.783118208703115e-06, "loss": 0.0212, "step": 1610 }, { "epoch": 0.9292320929232093, "grad_norm": 0.36738669872283936, "learning_rate": 9.777245161276372e-06, "loss": 0.0221, "step": 1620 }, { "epoch": 0.9349680934968093, "grad_norm": 0.9083078503608704, "learning_rate": 9.77129546383115e-06, "loss": 0.0261, "step": 1630 }, { "epoch": 0.9407040940704094, "grad_norm": 0.5853396058082581, "learning_rate": 9.765269211829245e-06, "loss": 0.011, "step": 1640 }, { "epoch": 0.9464400946440095, "grad_norm": 0.027249574661254883, "learning_rate": 9.759166501960762e-06, "loss": 0.0156, "step": 1650 }, { "epoch": 0.9464400946440095, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07951950281858444, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2513, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.727, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.387, "step": 1650 }, { "epoch": 0.9521760952176095, "grad_norm": 0.5655274987220764, "learning_rate": 9.752987432142556e-06, "loss": 0.0141, "step": 1660 }, { "epoch": 0.9579120957912096, "grad_norm": 0.20478393137454987, "learning_rate": 9.746732101516665e-06, "loss": 0.0225, "step": 1670 }, { "epoch": 0.9636480963648096, "grad_norm": 0.38293856382369995, "learning_rate": 9.740400610448714e-06, "loss": 0.0228, "step": 1680 }, { "epoch": 0.9693840969384097, "grad_norm": 0.4184674918651581, "learning_rate": 9.733993060526313e-06, "loss": 0.0278, "step": 1690 }, { "epoch": 0.9751200975120098, "grad_norm": 0.6761838793754578, "learning_rate": 9.727509554557416e-06, "loss": 0.0148, "step": 1700 }, { "epoch": 0.9751200975120098, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0835602656006813, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2977, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.661, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.354, "step": 1700 }, { "epoch": 0.9808560980856098, "grad_norm": 0.10887222737073898, "learning_rate": 9.720950196568689e-06, "loss": 0.0065, "step": 1710 }, { "epoch": 0.9865920986592098, "grad_norm": 0.13018013536930084, "learning_rate": 9.714315091803815e-06, "loss": 0.0083, "step": 1720 }, { "epoch": 0.99232809923281, "grad_norm": 0.1999017894268036, "learning_rate": 9.707604346721833e-06, "loss": 0.0097, "step": 1730 }, { "epoch": 0.99806409980641, "grad_norm": 0.06055552884936333, "learning_rate": 9.700818068995407e-06, "loss": 0.0255, "step": 1740 }, { "epoch": 1.00344160034416, "grad_norm": 2.305107831954956, "learning_rate": 9.693956367509117e-06, "loss": 0.021, "step": 1750 }, { "epoch": 1.00344160034416, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11077206581830978, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3027, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.653, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.35, "step": 1750 }, { "epoch": 1.00917760091776, "grad_norm": 2.2754151821136475, "learning_rate": 9.687019352357699e-06, "loss": 0.0398, "step": 1760 }, { "epoch": 1.0149136014913602, "grad_norm": 0.03042173944413662, "learning_rate": 9.680007134844279e-06, "loss": 0.0189, "step": 1770 }, { "epoch": 1.0206496020649602, "grad_norm": 1.6503928899765015, "learning_rate": 9.672919827478598e-06, "loss": 0.0111, "step": 1780 }, { "epoch": 1.0263856026385603, "grad_norm": 1.7391407489776611, "learning_rate": 9.665757543975196e-06, "loss": 0.0204, "step": 1790 }, { "epoch": 1.0321216032121603, "grad_norm": 1.6538548469543457, "learning_rate": 9.658520399251592e-06, "loss": 0.0165, "step": 1800 }, { "epoch": 1.0321216032121603, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07668553292751312, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1222, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.915, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.481, "step": 1800 }, { "epoch": 1.0378576037857603, "grad_norm": 0.022262316197156906, "learning_rate": 9.651208509426442e-06, "loss": 0.0053, "step": 1810 }, { "epoch": 1.0435936043593603, "grad_norm": 1.768957495689392, "learning_rate": 9.64382199181767e-06, "loss": 0.0164, "step": 1820 }, { "epoch": 1.0493296049329606, "grad_norm": 0.11224523931741714, "learning_rate": 9.63636096494059e-06, "loss": 0.0216, "step": 1830 }, { "epoch": 1.0550656055065606, "grad_norm": 0.022123126313090324, "learning_rate": 9.628825548506002e-06, "loss": 0.0069, "step": 1840 }, { "epoch": 1.0608016060801606, "grad_norm": 4.303464412689209, "learning_rate": 9.621215863418276e-06, "loss": 0.0253, "step": 1850 }, { "epoch": 1.0608016060801606, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0671575739979744, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2804, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.686, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.366, "step": 1850 }, { "epoch": 1.0665376066537606, "grad_norm": 0.08216200768947601, "learning_rate": 9.61353203177341e-06, "loss": 0.0052, "step": 1860 }, { "epoch": 1.0722736072273606, "grad_norm": 1.3728184700012207, "learning_rate": 9.605774176857062e-06, "loss": 0.0176, "step": 1870 }, { "epoch": 1.078009607800961, "grad_norm": 0.0963163748383522, "learning_rate": 9.597942423142586e-06, "loss": 0.0165, "step": 1880 }, { "epoch": 1.083745608374561, "grad_norm": 0.09453441202640533, "learning_rate": 9.59003689628903e-06, "loss": 0.0212, "step": 1890 }, { "epoch": 1.089481608948161, "grad_norm": 0.0946192815899849, "learning_rate": 9.582057723139115e-06, "loss": 0.0258, "step": 1900 }, { "epoch": 1.089481608948161, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07539978623390198, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2173, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.777, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.412, "step": 1900 }, { "epoch": 1.095217609521761, "grad_norm": 0.011775615625083447, "learning_rate": 9.574005031717203e-06, "loss": 0.0093, "step": 1910 }, { "epoch": 1.100953610095361, "grad_norm": 0.06459964066743851, "learning_rate": 9.565878951227247e-06, "loss": 0.0082, "step": 1920 }, { "epoch": 1.106689610668961, "grad_norm": 0.020801063627004623, "learning_rate": 9.557679612050708e-06, "loss": 0.0067, "step": 1930 }, { "epoch": 1.112425611242561, "grad_norm": 0.048270970582962036, "learning_rate": 9.549407145744473e-06, "loss": 0.0442, "step": 1940 }, { "epoch": 1.1181616118161612, "grad_norm": 1.6280121803283691, "learning_rate": 9.541061685038742e-06, "loss": 0.0258, "step": 1950 }, { "epoch": 1.1181616118161612, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0749003142118454, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1385, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.891, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.469, "step": 1950 }, { "epoch": 1.1238976123897613, "grad_norm": 0.4389003813266754, "learning_rate": 9.532643363834891e-06, "loss": 0.0165, "step": 1960 }, { "epoch": 1.1296336129633613, "grad_norm": 0.07638214528560638, "learning_rate": 9.524152317203337e-06, "loss": 0.0339, "step": 1970 }, { "epoch": 1.1353696135369613, "grad_norm": 0.26499396562576294, "learning_rate": 9.515588681381356e-06, "loss": 0.0229, "step": 1980 }, { "epoch": 1.1411056141105613, "grad_norm": 0.7412366271018982, "learning_rate": 9.506952593770908e-06, "loss": 0.0157, "step": 1990 }, { "epoch": 1.1468416146841616, "grad_norm": 0.16746670007705688, "learning_rate": 9.498244192936428e-06, "loss": 0.0102, "step": 2000 }, { "epoch": 1.1468416146841616, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09212490171194077, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1271, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.908, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.478, "step": 2000 }, { "epoch": 1.1525776152577616, "grad_norm": 0.182969868183136, "learning_rate": 9.489463618602602e-06, "loss": 0.0111, "step": 2010 }, { "epoch": 1.1583136158313616, "grad_norm": 1.3869882822036743, "learning_rate": 9.480611011652128e-06, "loss": 0.0146, "step": 2020 }, { "epoch": 1.1640496164049616, "grad_norm": 0.0466344878077507, "learning_rate": 9.471686514123455e-06, "loss": 0.009, "step": 2030 }, { "epoch": 1.1697856169785616, "grad_norm": 0.07296488434076309, "learning_rate": 9.462690269208498e-06, "loss": 0.0354, "step": 2040 }, { "epoch": 1.1755216175521617, "grad_norm": 1.9699822664260864, "learning_rate": 9.453622421250353e-06, "loss": 0.0114, "step": 2050 }, { "epoch": 1.1755216175521617, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09612752497196198, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1568, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.865, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.456, "step": 2050 }, { "epoch": 1.1812576181257617, "grad_norm": 2.4988596439361572, "learning_rate": 9.444483115740968e-06, "loss": 0.0235, "step": 2060 }, { "epoch": 1.186993618699362, "grad_norm": 0.0865960419178009, "learning_rate": 9.435272499318815e-06, "loss": 0.0197, "step": 2070 }, { "epoch": 1.192729619272962, "grad_norm": 0.7196880578994751, "learning_rate": 9.425990719766542e-06, "loss": 0.0085, "step": 2080 }, { "epoch": 1.198465619846562, "grad_norm": 0.4979308843612671, "learning_rate": 9.416637926008587e-06, "loss": 0.0177, "step": 2090 }, { "epoch": 1.204201620420162, "grad_norm": 1.3404853343963623, "learning_rate": 9.407214268108805e-06, "loss": 0.0079, "step": 2100 }, { "epoch": 1.204201620420162, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07606372237205505, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2379, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.747, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.397, "step": 2100 }, { "epoch": 1.209937620993762, "grad_norm": 0.007559036836028099, "learning_rate": 9.397719897268049e-06, "loss": 0.0122, "step": 2110 }, { "epoch": 1.2156736215673623, "grad_norm": 0.036009229719638824, "learning_rate": 9.388154965821754e-06, "loss": 0.0195, "step": 2120 }, { "epoch": 1.2214096221409623, "grad_norm": 0.23815706372261047, "learning_rate": 9.37851962723748e-06, "loss": 0.0311, "step": 2130 }, { "epoch": 1.2271456227145623, "grad_norm": 0.32620975375175476, "learning_rate": 9.36881403611246e-06, "loss": 0.0235, "step": 2140 }, { "epoch": 1.2328816232881623, "grad_norm": 0.6975133419036865, "learning_rate": 9.359038348171113e-06, "loss": 0.0398, "step": 2150 }, { "epoch": 1.2328816232881623, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07932226359844208, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1766, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.836, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.442, "step": 2150 }, { "epoch": 1.2386176238617623, "grad_norm": 0.055091869086027145, "learning_rate": 9.349192720262556e-06, "loss": 0.031, "step": 2160 }, { "epoch": 1.2443536244353623, "grad_norm": 0.07653547823429108, "learning_rate": 9.33927731035807e-06, "loss": 0.0188, "step": 2170 }, { "epoch": 1.2500896250089624, "grad_norm": 1.0113948583602905, "learning_rate": 9.329292277548584e-06, "loss": 0.023, "step": 2180 }, { "epoch": 1.2558256255825626, "grad_norm": 0.38709500432014465, "learning_rate": 9.319237782042108e-06, "loss": 0.0134, "step": 2190 }, { "epoch": 1.2615616261561626, "grad_norm": 1.0346875190734863, "learning_rate": 9.309113985161169e-06, "loss": 0.0272, "step": 2200 }, { "epoch": 1.2615616261561626, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06406532227993011, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1643, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.854, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.451, "step": 2200 }, { "epoch": 1.2672976267297626, "grad_norm": 3.084003448486328, "learning_rate": 9.298921049340226e-06, "loss": 0.0134, "step": 2210 }, { "epoch": 1.2730336273033627, "grad_norm": 0.3186517059803009, "learning_rate": 9.288659138123053e-06, "loss": 0.0142, "step": 2220 }, { "epoch": 1.278769627876963, "grad_norm": 0.02311491221189499, "learning_rate": 9.278328416160125e-06, "loss": 0.0116, "step": 2230 }, { "epoch": 1.284505628450563, "grad_norm": 0.021789675578475, "learning_rate": 9.267929049205976e-06, "loss": 0.003, "step": 2240 }, { "epoch": 1.290241629024163, "grad_norm": 0.004828541073948145, "learning_rate": 9.25746120411653e-06, "loss": 0.0074, "step": 2250 }, { "epoch": 1.290241629024163, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09356704354286194, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1187, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.92, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.484, "step": 2250 }, { "epoch": 1.295977629597763, "grad_norm": 4.326236248016357, "learning_rate": 9.246925048846433e-06, "loss": 0.0318, "step": 2260 }, { "epoch": 1.301713630171363, "grad_norm": 0.10032517462968826, "learning_rate": 9.236320752446357e-06, "loss": 0.0248, "step": 2270 }, { "epoch": 1.307449630744963, "grad_norm": 1.968422293663025, "learning_rate": 9.225648485060283e-06, "loss": 0.0245, "step": 2280 }, { "epoch": 1.313185631318563, "grad_norm": 0.9600183367729187, "learning_rate": 9.214908417922774e-06, "loss": 0.0169, "step": 2290 }, { "epoch": 1.3189216318921633, "grad_norm": 0.41536957025527954, "learning_rate": 9.20410072335623e-06, "loss": 0.0103, "step": 2300 }, { "epoch": 1.3189216318921633, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06794232130050659, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1585, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.862, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.455, "step": 2300 }, { "epoch": 1.3246576324657633, "grad_norm": 0.7403652667999268, "learning_rate": 9.19322557476812e-06, "loss": 0.0141, "step": 2310 }, { "epoch": 1.3303936330393633, "grad_norm": 0.03548659384250641, "learning_rate": 9.182283146648197e-06, "loss": 0.0214, "step": 2320 }, { "epoch": 1.3361296336129633, "grad_norm": 2.034658193588257, "learning_rate": 9.171273614565704e-06, "loss": 0.008, "step": 2330 }, { "epoch": 1.3418656341865633, "grad_norm": 1.0196254253387451, "learning_rate": 9.160197155166559e-06, "loss": 0.0459, "step": 2340 }, { "epoch": 1.3476016347601636, "grad_norm": 0.014422023668885231, "learning_rate": 9.149053946170507e-06, "loss": 0.0198, "step": 2350 }, { "epoch": 1.3476016347601636, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.058592408895492554, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2196, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.773, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.41, "step": 2350 }, { "epoch": 1.3533376353337636, "grad_norm": 0.019738655537366867, "learning_rate": 9.137844166368289e-06, "loss": 0.0233, "step": 2360 }, { "epoch": 1.3590736359073636, "grad_norm": 0.03966844826936722, "learning_rate": 9.126567995618752e-06, "loss": 0.0082, "step": 2370 }, { "epoch": 1.3648096364809637, "grad_norm": 0.10851484537124634, "learning_rate": 9.115225614845979e-06, "loss": 0.0132, "step": 2380 }, { "epoch": 1.3705456370545637, "grad_norm": 2.611233711242676, "learning_rate": 9.103817206036383e-06, "loss": 0.0062, "step": 2390 }, { "epoch": 1.3762816376281637, "grad_norm": 2.288485050201416, "learning_rate": 9.092342952235777e-06, "loss": 0.0476, "step": 2400 }, { "epoch": 1.3762816376281637, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07158029079437256, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1533, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.87, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.459, "step": 2400 }, { "epoch": 1.3820176382017637, "grad_norm": 0.04291702061891556, "learning_rate": 9.080803037546454e-06, "loss": 0.0031, "step": 2410 }, { "epoch": 1.387753638775364, "grad_norm": 0.008397878147661686, "learning_rate": 9.069197647124216e-06, "loss": 0.0121, "step": 2420 }, { "epoch": 1.393489639348964, "grad_norm": 0.015200245194137096, "learning_rate": 9.057526967175415e-06, "loss": 0.0047, "step": 2430 }, { "epoch": 1.399225639922564, "grad_norm": 3.2773385047912598, "learning_rate": 9.04579118495396e-06, "loss": 0.005, "step": 2440 }, { "epoch": 1.404961640496164, "grad_norm": 0.5176500678062439, "learning_rate": 9.033990488758317e-06, "loss": 0.0329, "step": 2450 }, { "epoch": 1.404961640496164, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.1131686270236969, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1939, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.811, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.429, "step": 2450 }, { "epoch": 1.410697641069764, "grad_norm": 0.002755386522039771, "learning_rate": 9.02212506792848e-06, "loss": 0.0164, "step": 2460 }, { "epoch": 1.4164336416433643, "grad_norm": 0.0466199554502964, "learning_rate": 9.01019511284294e-06, "loss": 0.0233, "step": 2470 }, { "epoch": 1.4221696422169643, "grad_norm": 0.8431953191757202, "learning_rate": 8.99820081491563e-06, "loss": 0.0051, "step": 2480 }, { "epoch": 1.4279056427905643, "grad_norm": 0.035143181681632996, "learning_rate": 8.98614236659285e-06, "loss": 0.0283, "step": 2490 }, { "epoch": 1.4336416433641643, "grad_norm": 0.05582081899046898, "learning_rate": 8.97401996135018e-06, "loss": 0.0073, "step": 2500 }, { "epoch": 1.4336416433641643, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09689504653215408, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.294, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.666, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.356, "step": 2500 }, { "epoch": 1.4393776439377644, "grad_norm": 2.2861459255218506, "learning_rate": 8.961833793689384e-06, "loss": 0.0107, "step": 2510 }, { "epoch": 1.4451136445113644, "grad_norm": 2.3350398540496826, "learning_rate": 8.94958405913527e-06, "loss": 0.0128, "step": 2520 }, { "epoch": 1.4508496450849644, "grad_norm": 0.019439250230789185, "learning_rate": 8.937270954232576e-06, "loss": 0.0061, "step": 2530 }, { "epoch": 1.4565856456585646, "grad_norm": 0.002678696997463703, "learning_rate": 8.924894676542801e-06, "loss": 0.0085, "step": 2540 }, { "epoch": 1.4623216462321647, "grad_norm": 0.03475391864776611, "learning_rate": 8.91245542464104e-06, "loss": 0.0234, "step": 2550 }, { "epoch": 1.4623216462321647, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10303693264722824, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1871, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.821, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, "step": 2550 }, { "epoch": 1.4680576468057647, "grad_norm": 0.09974029660224915, "learning_rate": 8.8999533981128e-06, "loss": 0.013, "step": 2560 }, { "epoch": 1.4737936473793647, "grad_norm": 0.19927239418029785, "learning_rate": 8.887388797550791e-06, "loss": 0.0033, "step": 2570 }, { "epoch": 1.479529647952965, "grad_norm": 0.03026706352829933, "learning_rate": 8.874761824551717e-06, "loss": 0.0219, "step": 2580 }, { "epoch": 1.485265648526565, "grad_norm": 0.8363674879074097, "learning_rate": 8.862072681713027e-06, "loss": 0.0167, "step": 2590 }, { "epoch": 1.491001649100165, "grad_norm": 2.117938280105591, "learning_rate": 8.84932157262969e-06, "loss": 0.0365, "step": 2600 }, { "epoch": 1.491001649100165, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11520245671272278, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2361, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.75, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.398, "step": 2600 }, { "epoch": 1.496737649673765, "grad_norm": 0.14108580350875854, "learning_rate": 8.836508701890892e-06, "loss": 0.0196, "step": 2610 }, { "epoch": 1.502473650247365, "grad_norm": 2.1537435054779053, "learning_rate": 8.823634275076792e-06, "loss": 0.0207, "step": 2620 }, { "epoch": 1.508209650820965, "grad_norm": 0.93586665391922, "learning_rate": 8.81069849875519e-06, "loss": 0.0125, "step": 2630 }, { "epoch": 1.513945651394565, "grad_norm": 3.1711156368255615, "learning_rate": 8.797701580478235e-06, "loss": 0.0094, "step": 2640 }, { "epoch": 1.519681651968165, "grad_norm": 0.0233930516988039, "learning_rate": 8.784643728779089e-06, "loss": 0.0152, "step": 2650 }, { "epoch": 1.519681651968165, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07520527392625809, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2328, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.754, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.401, "step": 2650 }, { "epoch": 1.525417652541765, "grad_norm": 0.14535315334796906, "learning_rate": 8.771525153168576e-06, "loss": 0.0035, "step": 2660 }, { "epoch": 1.5311536531153653, "grad_norm": 0.3828493654727936, "learning_rate": 8.758346064131824e-06, "loss": 0.0026, "step": 2670 }, { "epoch": 1.5368896536889654, "grad_norm": 2.7801029682159424, "learning_rate": 8.745106673124888e-06, "loss": 0.0215, "step": 2680 }, { "epoch": 1.5426256542625654, "grad_norm": 1.6995360851287842, "learning_rate": 8.731807192571359e-06, "loss": 0.022, "step": 2690 }, { "epoch": 1.5483616548361656, "grad_norm": 0.24279280006885529, "learning_rate": 8.718447835858951e-06, "loss": 0.0109, "step": 2700 }, { "epoch": 1.5483616548361656, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08999116718769073, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1893, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.817, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.432, "step": 2700 }, { "epoch": 1.5540976554097656, "grad_norm": 0.015813475474715233, "learning_rate": 8.705028817336083e-06, "loss": 0.0019, "step": 2710 }, { "epoch": 1.5598336559833657, "grad_norm": 0.017328623682260513, "learning_rate": 8.691550352308431e-06, "loss": 0.0044, "step": 2720 }, { "epoch": 1.5655696565569657, "grad_norm": 1.36957585811615, "learning_rate": 8.678012657035487e-06, "loss": 0.0033, "step": 2730 }, { "epoch": 1.5713056571305657, "grad_norm": 0.8026090264320374, "learning_rate": 8.664415948727076e-06, "loss": 0.0063, "step": 2740 }, { "epoch": 1.5770416577041657, "grad_norm": 0.038070451468229294, "learning_rate": 8.65076044553988e-06, "loss": 0.02, "step": 2750 }, { "epoch": 1.5770416577041657, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09743982553482056, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1673, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.849, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.448, "step": 2750 }, { "epoch": 1.5827776582777657, "grad_norm": 2.638280153274536, "learning_rate": 8.63704636657393e-06, "loss": 0.0205, "step": 2760 }, { "epoch": 1.5885136588513658, "grad_norm": 0.08616320043802261, "learning_rate": 8.623273931869094e-06, "loss": 0.0151, "step": 2770 }, { "epoch": 1.594249659424966, "grad_norm": 0.0016790858935564756, "learning_rate": 8.609443362401553e-06, "loss": 0.0077, "step": 2780 }, { "epoch": 1.599985659998566, "grad_norm": 0.0031273786444216967, "learning_rate": 8.595554880080246e-06, "loss": 0.0014, "step": 2790 }, { "epoch": 1.605721660572166, "grad_norm": 2.3747873306274414, "learning_rate": 8.581608707743312e-06, "loss": 0.0387, "step": 2800 }, { "epoch": 1.605721660572166, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10180775076150894, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1502, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.874, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.461, "step": 2800 }, { "epoch": 1.6114576611457663, "grad_norm": 0.03558593988418579, "learning_rate": 8.567605069154517e-06, "loss": 0.0009, "step": 2810 }, { "epoch": 1.6171936617193663, "grad_norm": 0.005191161762923002, "learning_rate": 8.55354418899966e-06, "loss": 0.0349, "step": 2820 }, { "epoch": 1.6229296622929663, "grad_norm": 1.5500961542129517, "learning_rate": 8.539426292882976e-06, "loss": 0.0296, "step": 2830 }, { "epoch": 1.6286656628665663, "grad_norm": 0.12692737579345703, "learning_rate": 8.525251607323506e-06, "loss": 0.0076, "step": 2840 }, { "epoch": 1.6344016634401664, "grad_norm": 0.034373391419649124, "learning_rate": 8.511020359751467e-06, "loss": 0.0125, "step": 2850 }, { "epoch": 1.6344016634401664, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07807676494121552, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1709, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.844, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.446, "step": 2850 }, { "epoch": 1.6401376640137664, "grad_norm": 0.3334125876426697, "learning_rate": 8.496732778504608e-06, "loss": 0.007, "step": 2860 }, { "epoch": 1.6458736645873664, "grad_norm": 1.8848798274993896, "learning_rate": 8.482389092824535e-06, "loss": 0.0279, "step": 2870 }, { "epoch": 1.6516096651609664, "grad_norm": 0.0071907006204128265, "learning_rate": 8.46798953285304e-06, "loss": 0.0058, "step": 2880 }, { "epoch": 1.6573456657345664, "grad_norm": 0.01580858789384365, "learning_rate": 8.453534329628414e-06, "loss": 0.0111, "step": 2890 }, { "epoch": 1.6630816663081667, "grad_norm": 0.5010847449302673, "learning_rate": 8.439023715081729e-06, "loss": 0.0205, "step": 2900 }, { "epoch": 1.6630816663081667, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0824664756655693, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2284, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.761, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.404, "step": 2900 }, { "epoch": 1.6688176668817667, "grad_norm": 0.16337744891643524, "learning_rate": 8.42445792203312e-06, "loss": 0.0175, "step": 2910 }, { "epoch": 1.6745536674553667, "grad_norm": 0.054014191031455994, "learning_rate": 8.409837184188056e-06, "loss": 0.0073, "step": 2920 }, { "epoch": 1.680289668028967, "grad_norm": 3.6512198448181152, "learning_rate": 8.395161736133579e-06, "loss": 0.0226, "step": 2930 }, { "epoch": 1.686025668602567, "grad_norm": 0.6102538704872131, "learning_rate": 8.380431813334548e-06, "loss": 0.0033, "step": 2940 }, { "epoch": 1.691761669176167, "grad_norm": 2.3051533699035645, "learning_rate": 8.365647652129865e-06, "loss": 0.0327, "step": 2950 }, { "epoch": 1.691761669176167, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0759689062833786, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.5562, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.293, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.17, "step": 2950 }, { "epoch": 1.697497669749767, "grad_norm": 0.008392853662371635, "learning_rate": 8.350809489728673e-06, "loss": 0.0279, "step": 2960 }, { "epoch": 1.703233670323367, "grad_norm": 1.614988088607788, "learning_rate": 8.33591756420655e-06, "loss": 0.0102, "step": 2970 }, { "epoch": 1.708969670896967, "grad_norm": 0.04466373473405838, "learning_rate": 8.320972114501698e-06, "loss": 0.0158, "step": 2980 }, { "epoch": 1.714705671470567, "grad_norm": 1.4472012519836426, "learning_rate": 8.305973380411107e-06, "loss": 0.0437, "step": 2990 }, { "epoch": 1.720441672044167, "grad_norm": 0.0326494500041008, "learning_rate": 8.290921602586699e-06, "loss": 0.0188, "step": 3000 }, { "epoch": 1.720441672044167, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06277668476104736, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2361, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.75, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.398, "step": 3000 }, { "epoch": 1.7261776726177671, "grad_norm": 2.090261697769165, "learning_rate": 8.275817022531479e-06, "loss": 0.0229, "step": 3010 }, { "epoch": 1.7319136731913674, "grad_norm": 1.1351670026779175, "learning_rate": 8.260659882595647e-06, "loss": 0.005, "step": 3020 }, { "epoch": 1.7376496737649674, "grad_norm": 0.0521712526679039, "learning_rate": 8.245450425972728e-06, "loss": 0.0126, "step": 3030 }, { "epoch": 1.7433856743385674, "grad_norm": 0.04499056562781334, "learning_rate": 8.230188896695643e-06, "loss": 0.0133, "step": 3040 }, { "epoch": 1.7491216749121676, "grad_norm": 0.007846461609005928, "learning_rate": 8.214875539632825e-06, "loss": 0.0191, "step": 3050 }, { "epoch": 1.7491216749121676, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07422397285699844, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1752, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.838, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.443, "step": 3050 }, { "epoch": 1.7548576754857677, "grad_norm": 0.055605433881282806, "learning_rate": 8.199510600484261e-06, "loss": 0.0148, "step": 3060 }, { "epoch": 1.7605936760593677, "grad_norm": 0.27163052558898926, "learning_rate": 8.184094325777573e-06, "loss": 0.0075, "step": 3070 }, { "epoch": 1.7663296766329677, "grad_norm": 0.11399796605110168, "learning_rate": 8.168626962864045e-06, "loss": 0.0016, "step": 3080 }, { "epoch": 1.7720656772065677, "grad_norm": 0.07731972634792328, "learning_rate": 8.153108759914669e-06, "loss": 0.0141, "step": 3090 }, { "epoch": 1.7778016777801677, "grad_norm": 0.2923997938632965, "learning_rate": 8.13753996591615e-06, "loss": 0.0047, "step": 3100 }, { "epoch": 1.7778016777801677, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06707713752985, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.4009, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.513, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.28, "step": 3100 }, { "epoch": 1.7835376783537678, "grad_norm": 0.26035454869270325, "learning_rate": 8.121920830666918e-06, "loss": 0.0092, "step": 3110 }, { "epoch": 1.7892736789273678, "grad_norm": 3.143533706665039, "learning_rate": 8.106251604773125e-06, "loss": 0.0141, "step": 3120 }, { "epoch": 1.7950096795009678, "grad_norm": 1.8246173858642578, "learning_rate": 8.090532539644608e-06, "loss": 0.0249, "step": 3130 }, { "epoch": 1.800745680074568, "grad_norm": 0.01822010800242424, "learning_rate": 8.074763887490878e-06, "loss": 0.0182, "step": 3140 }, { "epoch": 1.806481680648168, "grad_norm": 1.8796803951263428, "learning_rate": 8.058945901317047e-06, "loss": 0.0276, "step": 3150 }, { "epoch": 1.806481680648168, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06196051836013794, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1882, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.819, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.433, "step": 3150 }, { "epoch": 1.812217681221768, "grad_norm": 0.3315775692462921, "learning_rate": 8.043078834919792e-06, "loss": 0.0053, "step": 3160 }, { "epoch": 1.8179536817953683, "grad_norm": 0.6810622215270996, "learning_rate": 8.027162942883271e-06, "loss": 0.0027, "step": 3170 }, { "epoch": 1.8236896823689683, "grad_norm": 0.04800894856452942, "learning_rate": 8.011198480575037e-06, "loss": 0.0104, "step": 3180 }, { "epoch": 1.8294256829425684, "grad_norm": 0.21183735132217407, "learning_rate": 7.995185704141948e-06, "loss": 0.0139, "step": 3190 }, { "epoch": 1.8351616835161684, "grad_norm": 0.05010043457150459, "learning_rate": 7.979124870506052e-06, "loss": 0.013, "step": 3200 }, { "epoch": 1.8351616835161684, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08539190888404846, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2295, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.759, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.403, "step": 3200 }, { "epoch": 1.8408976840897684, "grad_norm": 0.05234100669622421, "learning_rate": 7.963016237360465e-06, "loss": 0.0113, "step": 3210 }, { "epoch": 1.8466336846633684, "grad_norm": 0.7740063071250916, "learning_rate": 7.946860063165238e-06, "loss": 0.0115, "step": 3220 }, { "epoch": 1.8523696852369684, "grad_norm": 0.007312777452170849, "learning_rate": 7.93065660714321e-06, "loss": 0.0378, "step": 3230 }, { "epoch": 1.8581056858105685, "grad_norm": 0.014250587671995163, "learning_rate": 7.914406129275847e-06, "loss": 0.0188, "step": 3240 }, { "epoch": 1.8638416863841685, "grad_norm": 0.0380655862390995, "learning_rate": 7.898108890299072e-06, "loss": 0.013, "step": 3250 }, { "epoch": 1.8638416863841685, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07671073824167252, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.182, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.828, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.438, "step": 3250 }, { "epoch": 1.8695776869577687, "grad_norm": 0.034287262707948685, "learning_rate": 7.881765151699085e-06, "loss": 0.0237, "step": 3260 }, { "epoch": 1.8753136875313687, "grad_norm": 0.014141724444925785, "learning_rate": 7.865375175708158e-06, "loss": 0.0294, "step": 3270 }, { "epoch": 1.881049688104969, "grad_norm": 2.08565616607666, "learning_rate": 7.848939225300436e-06, "loss": 0.0229, "step": 3280 }, { "epoch": 1.886785688678569, "grad_norm": 1.476799726486206, "learning_rate": 7.832457564187715e-06, "loss": 0.0257, "step": 3290 }, { "epoch": 1.892521689252169, "grad_norm": 1.641891598701477, "learning_rate": 7.815930456815212e-06, "loss": 0.0171, "step": 3300 }, { "epoch": 1.892521689252169, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06915121525526047, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.0913, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.961, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.504, "step": 3300 }, { "epoch": 1.898257689825769, "grad_norm": 0.03908121585845947, "learning_rate": 7.799358168357323e-06, "loss": 0.0208, "step": 3310 }, { "epoch": 1.903993690399369, "grad_norm": 0.27976372838020325, "learning_rate": 7.782740964713358e-06, "loss": 0.0079, "step": 3320 }, { "epoch": 1.909729690972969, "grad_norm": 0.1647738814353943, "learning_rate": 7.76607911250329e-06, "loss": 0.0284, "step": 3330 }, { "epoch": 1.915465691546569, "grad_norm": 0.009193528443574905, "learning_rate": 7.749372879063468e-06, "loss": 0.0219, "step": 3340 }, { "epoch": 1.9212016921201691, "grad_norm": 0.08929289132356644, "learning_rate": 7.73262253244233e-06, "loss": 0.0136, "step": 3350 }, { "epoch": 1.9212016921201691, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06728994846343994, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1919, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.814, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.43, "step": 3350 }, { "epoch": 1.9269376926937691, "grad_norm": 1.6335256099700928, "learning_rate": 7.715828341396102e-06, "loss": 0.0322, "step": 3360 }, { "epoch": 1.9326736932673694, "grad_norm": 0.2756313979625702, "learning_rate": 7.698990575384483e-06, "loss": 0.0282, "step": 3370 }, { "epoch": 1.9384096938409694, "grad_norm": 0.5282019376754761, "learning_rate": 7.68210950456633e-06, "loss": 0.0059, "step": 3380 }, { "epoch": 1.9441456944145694, "grad_norm": 0.013134743086993694, "learning_rate": 7.66518539979531e-06, "loss": 0.0088, "step": 3390 }, { "epoch": 1.9498816949881697, "grad_norm": 0.020495153963565826, "learning_rate": 7.648218532615572e-06, "loss": 0.0275, "step": 3400 }, { "epoch": 1.9498816949881697, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06392066925764084, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1842, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.825, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.436, "step": 3400 }, { "epoch": 1.9556176955617697, "grad_norm": 0.030520539730787277, "learning_rate": 7.631209175257368e-06, "loss": 0.0176, "step": 3410 }, { "epoch": 1.9613536961353697, "grad_norm": 0.356724351644516, "learning_rate": 7.614157600632706e-06, "loss": 0.0306, "step": 3420 }, { "epoch": 1.9670896967089697, "grad_norm": 0.5876423716545105, "learning_rate": 7.597064082330961e-06, "loss": 0.0178, "step": 3430 }, { "epoch": 1.9728256972825697, "grad_norm": 0.04390386864542961, "learning_rate": 7.579928894614479e-06, "loss": 0.0099, "step": 3440 }, { "epoch": 1.9785616978561698, "grad_norm": 0.026595573872327805, "learning_rate": 7.562752312414196e-06, "loss": 0.0097, "step": 3450 }, { "epoch": 1.9785616978561698, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08633749186992645, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1407, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.888, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.468, "step": 3450 }, { "epoch": 1.9842976984297698, "grad_norm": 0.4078025221824646, "learning_rate": 7.545534611325207e-06, "loss": 0.0189, "step": 3460 }, { "epoch": 1.9900336990033698, "grad_norm": 0.4737612307071686, "learning_rate": 7.528276067602352e-06, "loss": 0.0019, "step": 3470 }, { "epoch": 1.9957696995769698, "grad_norm": 0.0031162798404693604, "learning_rate": 7.510976958155789e-06, "loss": 0.0037, "step": 3480 }, { "epoch": 2.00114720011472, "grad_norm": 1.097615361213684, "learning_rate": 7.493637560546539e-06, "loss": 0.0127, "step": 3490 }, { "epoch": 2.00688320068832, "grad_norm": 0.005019227508455515, "learning_rate": 7.476258152982043e-06, "loss": 0.0019, "step": 3500 }, { "epoch": 2.00688320068832, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07848720997571945, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.177, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.835, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.441, "step": 3500 }, { "epoch": 2.01261920126192, "grad_norm": 0.016102029010653496, "learning_rate": 7.458839014311696e-06, "loss": 0.0057, "step": 3510 }, { "epoch": 2.01835520183552, "grad_norm": 0.009885331615805626, "learning_rate": 7.441380424022364e-06, "loss": 0.0089, "step": 3520 }, { "epoch": 2.0240912024091204, "grad_norm": 2.9838740825653076, "learning_rate": 7.423882662233915e-06, "loss": 0.0142, "step": 3530 }, { "epoch": 2.0298272029827205, "grad_norm": 0.29027923941612244, "learning_rate": 7.406346009694713e-06, "loss": 0.0033, "step": 3540 }, { "epoch": 2.0355632035563205, "grad_norm": 0.004921222571283579, "learning_rate": 7.388770747777116e-06, "loss": 0.0108, "step": 3550 }, { "epoch": 2.0355632035563205, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0910039097070694, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.144, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.883, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.465, "step": 3550 }, { "epoch": 2.0412992041299205, "grad_norm": 0.012345495633780956, "learning_rate": 7.371157158472965e-06, "loss": 0.0186, "step": 3560 }, { "epoch": 2.0470352047035205, "grad_norm": 0.013729539699852467, "learning_rate": 7.353505524389052e-06, "loss": 0.0038, "step": 3570 }, { "epoch": 2.0527712052771205, "grad_norm": 3.2665841579437256, "learning_rate": 7.335816128742599e-06, "loss": 0.0203, "step": 3580 }, { "epoch": 2.0585072058507206, "grad_norm": 0.0018614591099321842, "learning_rate": 7.318089255356695e-06, "loss": 0.0093, "step": 3590 }, { "epoch": 2.0642432064243206, "grad_norm": 0.10742621123790741, "learning_rate": 7.300325188655762e-06, "loss": 0.0093, "step": 3600 }, { "epoch": 2.0642432064243206, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07848691940307617, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1618, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.858, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.452, "step": 3600 }, { "epoch": 2.0699792069979206, "grad_norm": 0.6563161611557007, "learning_rate": 7.282524213660974e-06, "loss": 0.001, "step": 3610 }, { "epoch": 2.0757152075715206, "grad_norm": 0.12672875821590424, "learning_rate": 7.264686615985697e-06, "loss": 0.0008, "step": 3620 }, { "epoch": 2.0814512081451206, "grad_norm": 0.004448240157216787, "learning_rate": 7.246812681830899e-06, "loss": 0.0139, "step": 3630 }, { "epoch": 2.0871872087187207, "grad_norm": 0.007188173942267895, "learning_rate": 7.228902697980562e-06, "loss": 0.0106, "step": 3640 }, { "epoch": 2.092923209292321, "grad_norm": 0.006886759772896767, "learning_rate": 7.210956951797074e-06, "loss": 0.0076, "step": 3650 }, { "epoch": 2.092923209292321, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08482418954372406, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1827, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.827, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.437, "step": 3650 }, { "epoch": 2.098659209865921, "grad_norm": 3.8101847171783447, "learning_rate": 7.19297573121663e-06, "loss": 0.0089, "step": 3660 }, { "epoch": 2.104395210439521, "grad_norm": 0.002013096585869789, "learning_rate": 7.174959324744599e-06, "loss": 0.0123, "step": 3670 }, { "epoch": 2.110131211013121, "grad_norm": 3.1393046379089355, "learning_rate": 7.156908021450904e-06, "loss": 0.01, "step": 3680 }, { "epoch": 2.115867211586721, "grad_norm": 0.07943509519100189, "learning_rate": 7.138822110965381e-06, "loss": 0.0016, "step": 3690 }, { "epoch": 2.121603212160321, "grad_norm": 2.5405237674713135, "learning_rate": 7.120701883473131e-06, "loss": 0.0148, "step": 3700 }, { "epoch": 2.121603212160321, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10251594334840775, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2361, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.75, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.398, "step": 3700 }, { "epoch": 2.1273392127339212, "grad_norm": 0.006140326615422964, "learning_rate": 7.102547629709867e-06, "loss": 0.0163, "step": 3710 }, { "epoch": 2.1330752133075213, "grad_norm": 0.04222610965371132, "learning_rate": 7.084359640957246e-06, "loss": 0.017, "step": 3720 }, { "epoch": 2.1388112138811213, "grad_norm": 0.03461828827857971, "learning_rate": 7.066138209038194e-06, "loss": 0.0162, "step": 3730 }, { "epoch": 2.1445472144547213, "grad_norm": 0.9535601139068604, "learning_rate": 7.047883626312233e-06, "loss": 0.0266, "step": 3740 }, { "epoch": 2.1502832150283213, "grad_norm": 0.8345515131950378, "learning_rate": 7.029596185670778e-06, "loss": 0.0446, "step": 3750 }, { "epoch": 2.1502832150283213, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07697154581546783, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1863, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.822, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.435, "step": 3750 }, { "epoch": 2.156019215601922, "grad_norm": 0.14290325343608856, "learning_rate": 7.011276180532445e-06, "loss": 0.0218, "step": 3760 }, { "epoch": 2.161755216175522, "grad_norm": 0.05109575018286705, "learning_rate": 6.992923904838341e-06, "loss": 0.0032, "step": 3770 }, { "epoch": 2.167491216749122, "grad_norm": 3.4320428371429443, "learning_rate": 6.974539653047346e-06, "loss": 0.0107, "step": 3780 }, { "epoch": 2.173227217322722, "grad_norm": 0.9100053310394287, "learning_rate": 6.956123720131398e-06, "loss": 0.0162, "step": 3790 }, { "epoch": 2.178963217896322, "grad_norm": 0.09293865412473679, "learning_rate": 6.937676401570744e-06, "loss": 0.0175, "step": 3800 }, { "epoch": 2.178963217896322, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07296475023031235, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.156, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.866, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.457, "step": 3800 }, { "epoch": 2.184699218469922, "grad_norm": 0.25332963466644287, "learning_rate": 6.9191979933492135e-06, "loss": 0.0107, "step": 3810 }, { "epoch": 2.190435219043522, "grad_norm": 0.47701334953308105, "learning_rate": 6.900688791949463e-06, "loss": 0.0087, "step": 3820 }, { "epoch": 2.196171219617122, "grad_norm": 0.019721075892448425, "learning_rate": 6.882149094348215e-06, "loss": 0.0178, "step": 3830 }, { "epoch": 2.201907220190722, "grad_norm": 0.057930897921323776, "learning_rate": 6.863579198011506e-06, "loss": 0.0291, "step": 3840 }, { "epoch": 2.207643220764322, "grad_norm": 0.03594127669930458, "learning_rate": 6.8449794008899e-06, "loss": 0.005, "step": 3850 }, { "epoch": 2.207643220764322, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07210449874401093, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1052, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.94, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.494, "step": 3850 }, { "epoch": 2.213379221337922, "grad_norm": 2.532087802886963, "learning_rate": 6.826350001413713e-06, "loss": 0.0085, "step": 3860 }, { "epoch": 2.219115221911522, "grad_norm": 0.3755287230014801, "learning_rate": 6.807691298488231e-06, "loss": 0.0014, "step": 3870 }, { "epoch": 2.224851222485122, "grad_norm": 0.09285666048526764, "learning_rate": 6.789003591488902e-06, "loss": 0.0013, "step": 3880 }, { "epoch": 2.2305872230587225, "grad_norm": 3.115718126296997, "learning_rate": 6.770287180256545e-06, "loss": 0.0045, "step": 3890 }, { "epoch": 2.2363232236323225, "grad_norm": 1.7172398567199707, "learning_rate": 6.751542365092527e-06, "loss": 0.0085, "step": 3900 }, { "epoch": 2.2363232236323225, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08293686807155609, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1262, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.909, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.478, "step": 3900 }, { "epoch": 2.2420592242059225, "grad_norm": 0.004544651135802269, "learning_rate": 6.732769446753954e-06, "loss": 0.021, "step": 3910 }, { "epoch": 2.2477952247795225, "grad_norm": 3.599937677383423, "learning_rate": 6.713968726448844e-06, "loss": 0.0067, "step": 3920 }, { "epoch": 2.2535312253531226, "grad_norm": 0.9228914976119995, "learning_rate": 6.69514050583129e-06, "loss": 0.0097, "step": 3930 }, { "epoch": 2.2592672259267226, "grad_norm": 0.005351903382688761, "learning_rate": 6.676285086996623e-06, "loss": 0.0353, "step": 3940 }, { "epoch": 2.2650032265003226, "grad_norm": 0.07884957641363144, "learning_rate": 6.657402772476563e-06, "loss": 0.0092, "step": 3950 }, { "epoch": 2.2650032265003226, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09563428908586502, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.7675, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 29.999, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.022, "step": 3950 }, { "epoch": 2.2707392270739226, "grad_norm": 1.1882566213607788, "learning_rate": 6.638493865234369e-06, "loss": 0.0132, "step": 3960 }, { "epoch": 2.2764752276475226, "grad_norm": 0.13506627082824707, "learning_rate": 6.619558668659975e-06, "loss": 0.0219, "step": 3970 }, { "epoch": 2.2822112282211227, "grad_norm": 0.007426030468195677, "learning_rate": 6.600597486565119e-06, "loss": 0.0216, "step": 3980 }, { "epoch": 2.2879472287947227, "grad_norm": 0.27120983600616455, "learning_rate": 6.581610623178476e-06, "loss": 0.0153, "step": 3990 }, { "epoch": 2.293683229368323, "grad_norm": 1.4845871925354004, "learning_rate": 6.562598383140773e-06, "loss": 0.0059, "step": 4000 }, { "epoch": 2.293683229368323, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07171830534934998, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2153, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.78, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.413, "step": 4000 }, { "epoch": 2.299419229941923, "grad_norm": 1.7247122526168823, "learning_rate": 6.543561071499893e-06, "loss": 0.0216, "step": 4010 }, { "epoch": 2.305155230515523, "grad_norm": 0.02364450879395008, "learning_rate": 6.524498993706e-06, "loss": 0.0244, "step": 4020 }, { "epoch": 2.310891231089123, "grad_norm": 0.048632461577653885, "learning_rate": 6.505412455606617e-06, "loss": 0.0029, "step": 4030 }, { "epoch": 2.316627231662723, "grad_norm": 0.12222401797771454, "learning_rate": 6.486301763441732e-06, "loss": 0.0027, "step": 4040 }, { "epoch": 2.3223632322363232, "grad_norm": 0.07429647445678711, "learning_rate": 6.467167223838879e-06, "loss": 0.0119, "step": 4050 }, { "epoch": 2.3223632322363232, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09189260751008987, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1514, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.873, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.46, "step": 4050 }, { "epoch": 2.3280992328099233, "grad_norm": 0.10564873367547989, "learning_rate": 6.4480091438082225e-06, "loss": 0.0328, "step": 4060 }, { "epoch": 2.3338352333835233, "grad_norm": 0.01708606630563736, "learning_rate": 6.4288278307376265e-06, "loss": 0.015, "step": 4070 }, { "epoch": 2.3395712339571233, "grad_norm": 0.25011512637138367, "learning_rate": 6.4096235923877256e-06, "loss": 0.0061, "step": 4080 }, { "epoch": 2.3453072345307233, "grad_norm": 0.055370282381772995, "learning_rate": 6.390396736886986e-06, "loss": 0.007, "step": 4090 }, { "epoch": 2.3510432351043233, "grad_norm": 0.030733373016119003, "learning_rate": 6.371147572726761e-06, "loss": 0.0017, "step": 4100 }, { "epoch": 2.3510432351043233, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08482550084590912, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1309, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.903, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.475, "step": 4100 }, { "epoch": 2.356779235677924, "grad_norm": 0.17796050012111664, "learning_rate": 6.351876408756344e-06, "loss": 0.0053, "step": 4110 }, { "epoch": 2.3625152362515234, "grad_norm": 0.0136833805590868, "learning_rate": 6.332583554178009e-06, "loss": 0.0044, "step": 4120 }, { "epoch": 2.368251236825124, "grad_norm": 4.209476470947266, "learning_rate": 6.313269318542057e-06, "loss": 0.0259, "step": 4130 }, { "epoch": 2.373987237398724, "grad_norm": 3.9427716732025146, "learning_rate": 6.2939340117418355e-06, "loss": 0.0237, "step": 4140 }, { "epoch": 2.379723237972324, "grad_norm": 0.008540840819478035, "learning_rate": 6.274577944008785e-06, "loss": 0.0044, "step": 4150 }, { "epoch": 2.379723237972324, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07825496047735214, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1875, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.82, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, "step": 4150 }, { "epoch": 2.385459238545924, "grad_norm": 0.01801919937133789, "learning_rate": 6.255201425907442e-06, "loss": 0.0002, "step": 4160 }, { "epoch": 2.391195239119524, "grad_norm": 0.7031999230384827, "learning_rate": 6.235804768330472e-06, "loss": 0.0198, "step": 4170 }, { "epoch": 2.396931239693124, "grad_norm": 0.03624948486685753, "learning_rate": 6.21638828249367e-06, "loss": 0.0096, "step": 4180 }, { "epoch": 2.402667240266724, "grad_norm": 0.010821559466421604, "learning_rate": 6.196952279930977e-06, "loss": 0.0089, "step": 4190 }, { "epoch": 2.408403240840324, "grad_norm": 0.01340206153690815, "learning_rate": 6.177497072489473e-06, "loss": 0.0227, "step": 4200 }, { "epoch": 2.408403240840324, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07348810136318207, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1832, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.826, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.437, "step": 4200 }, { "epoch": 2.414139241413924, "grad_norm": 0.006387198343873024, "learning_rate": 6.158022972324375e-06, "loss": 0.0112, "step": 4210 }, { "epoch": 2.419875241987524, "grad_norm": 0.1088046282529831, "learning_rate": 6.138530291894033e-06, "loss": 0.015, "step": 4220 }, { "epoch": 2.425611242561124, "grad_norm": 0.010532280430197716, "learning_rate": 6.119019343954914e-06, "loss": 0.0012, "step": 4230 }, { "epoch": 2.4313472431347245, "grad_norm": 2.5161547660827637, "learning_rate": 6.099490441556582e-06, "loss": 0.0054, "step": 4240 }, { "epoch": 2.4370832437083245, "grad_norm": 0.41214823722839355, "learning_rate": 6.07994389803668e-06, "loss": 0.0067, "step": 4250 }, { "epoch": 2.4370832437083245, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08171182125806808, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1867, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.821, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, "step": 4250 }, { "epoch": 2.4428192442819245, "grad_norm": 3.9505116939544678, "learning_rate": 6.060380027015897e-06, "loss": 0.0205, "step": 4260 }, { "epoch": 2.4485552448555246, "grad_norm": 2.172741651535034, "learning_rate": 6.040799142392937e-06, "loss": 0.015, "step": 4270 }, { "epoch": 2.4542912454291246, "grad_norm": 0.003915698733180761, "learning_rate": 6.02120155833949e-06, "loss": 0.0047, "step": 4280 }, { "epoch": 2.4600272460027246, "grad_norm": 0.027446260675787926, "learning_rate": 6.001587589295179e-06, "loss": 0.0172, "step": 4290 }, { "epoch": 2.4657632465763246, "grad_norm": 0.9420067667961121, "learning_rate": 5.981957549962524e-06, "loss": 0.0132, "step": 4300 }, { "epoch": 2.4657632465763246, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06599417328834534, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2262, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.764, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.405, "step": 4300 }, { "epoch": 2.4714992471499246, "grad_norm": 2.6476902961730957, "learning_rate": 5.96231175530189e-06, "loss": 0.01, "step": 4310 }, { "epoch": 2.4772352477235247, "grad_norm": 0.03741572052240372, "learning_rate": 5.942650520526432e-06, "loss": 0.0102, "step": 4320 }, { "epoch": 2.4829712482971247, "grad_norm": 0.012345872819423676, "learning_rate": 5.9229741610970425e-06, "loss": 0.0154, "step": 4330 }, { "epoch": 2.4887072488707247, "grad_norm": 0.006493984255939722, "learning_rate": 5.903282992717281e-06, "loss": 0.0063, "step": 4340 }, { "epoch": 2.494443249444325, "grad_norm": 0.011486309580504894, "learning_rate": 5.883577331328316e-06, "loss": 0.0005, "step": 4350 }, { "epoch": 2.494443249444325, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0845269188284874, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1186, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.921, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.484, "step": 4350 }, { "epoch": 2.5001792500179247, "grad_norm": 0.05512839928269386, "learning_rate": 5.863857493103855e-06, "loss": 0.0249, "step": 4360 }, { "epoch": 2.505915250591525, "grad_norm": 0.010474382899701595, "learning_rate": 5.84412379444507e-06, "loss": 0.0007, "step": 4370 }, { "epoch": 2.511651251165125, "grad_norm": 0.8734518885612488, "learning_rate": 5.824376551975519e-06, "loss": 0.0081, "step": 4380 }, { "epoch": 2.5173872517387252, "grad_norm": 0.11771434545516968, "learning_rate": 5.804616082536071e-06, "loss": 0.0074, "step": 4390 }, { "epoch": 2.5231232523123253, "grad_norm": 1.922683835029602, "learning_rate": 5.784842703179814e-06, "loss": 0.0328, "step": 4400 }, { "epoch": 2.5231232523123253, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07923904061317444, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.0857, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.969, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.508, "step": 4400 }, { "epoch": 2.5288592528859253, "grad_norm": 4.321990013122559, "learning_rate": 5.765056731166982e-06, "loss": 0.0237, "step": 4410 }, { "epoch": 2.5345952534595253, "grad_norm": 0.01736987754702568, "learning_rate": 5.745258483959847e-06, "loss": 0.0064, "step": 4420 }, { "epoch": 2.5403312540331253, "grad_norm": 0.07172615826129913, "learning_rate": 5.725448279217638e-06, "loss": 0.0118, "step": 4430 }, { "epoch": 2.5460672546067253, "grad_norm": 0.1578182578086853, "learning_rate": 5.705626434791444e-06, "loss": 0.0016, "step": 4440 }, { "epoch": 2.5518032551803254, "grad_norm": 0.005266325548291206, "learning_rate": 5.6857932687191e-06, "loss": 0.0023, "step": 4450 }, { "epoch": 2.5518032551803254, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07244950532913208, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2865, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.677, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, "step": 4450 }, { "epoch": 2.557539255753926, "grad_norm": 0.21443744003772736, "learning_rate": 5.66594909922011e-06, "loss": 0.0114, "step": 4460 }, { "epoch": 2.5632752563275254, "grad_norm": 0.08916337788105011, "learning_rate": 5.646094244690511e-06, "loss": 0.0064, "step": 4470 }, { "epoch": 2.569011256901126, "grad_norm": 2.983360528945923, "learning_rate": 5.626229023697789e-06, "loss": 0.0159, "step": 4480 }, { "epoch": 2.574747257474726, "grad_norm": 0.06603588908910751, "learning_rate": 5.606353754975755e-06, "loss": 0.0104, "step": 4490 }, { "epoch": 2.580483258048326, "grad_norm": 0.9153593182563782, "learning_rate": 5.586468757419433e-06, "loss": 0.0012, "step": 4500 }, { "epoch": 2.580483258048326, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07681399583816528, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3831, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.538, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.292, "step": 4500 }, { "epoch": 2.586219258621926, "grad_norm": 0.030913453549146652, "learning_rate": 5.566574350079946e-06, "loss": 0.0139, "step": 4510 }, { "epoch": 2.591955259195526, "grad_norm": 3.312136650085449, "learning_rate": 5.546670852159396e-06, "loss": 0.0074, "step": 4520 }, { "epoch": 2.597691259769126, "grad_norm": 0.008490420877933502, "learning_rate": 5.526758583005736e-06, "loss": 0.008, "step": 4530 }, { "epoch": 2.603427260342726, "grad_norm": 0.010648602619767189, "learning_rate": 5.50683786210766e-06, "loss": 0.0051, "step": 4540 }, { "epoch": 2.609163260916326, "grad_norm": 2.533731460571289, "learning_rate": 5.486909009089463e-06, "loss": 0.0185, "step": 4550 }, { "epoch": 2.609163260916326, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07284829020500183, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1938, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.811, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.429, "step": 4550 }, { "epoch": 2.614899261489926, "grad_norm": 0.0175775233656168, "learning_rate": 5.4669723437059194e-06, "loss": 0.0015, "step": 4560 }, { "epoch": 2.6206352620635265, "grad_norm": 0.00764029985293746, "learning_rate": 5.4470281858371555e-06, "loss": 0.0014, "step": 4570 }, { "epoch": 2.626371262637126, "grad_norm": 0.4708549380302429, "learning_rate": 5.4270768554835056e-06, "loss": 0.0023, "step": 4580 }, { "epoch": 2.6321072632107265, "grad_norm": 0.06411808729171753, "learning_rate": 5.407118672760393e-06, "loss": 0.0037, "step": 4590 }, { "epoch": 2.6378432637843265, "grad_norm": 0.009167805314064026, "learning_rate": 5.387153957893181e-06, "loss": 0.0016, "step": 4600 }, { "epoch": 2.6378432637843265, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09603765606880188, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2377, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.747, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.397, "step": 4600 }, { "epoch": 2.6435792643579266, "grad_norm": 0.0018381074769422412, "learning_rate": 5.367183031212041e-06, "loss": 0.007, "step": 4610 }, { "epoch": 2.6493152649315266, "grad_norm": 0.0011304231593385339, "learning_rate": 5.347206213146813e-06, "loss": 0.0277, "step": 4620 }, { "epoch": 2.6550512655051266, "grad_norm": 4.376464366912842, "learning_rate": 5.327223824221862e-06, "loss": 0.0093, "step": 4630 }, { "epoch": 2.6607872660787266, "grad_norm": 0.031111005693674088, "learning_rate": 5.307236185050937e-06, "loss": 0.0029, "step": 4640 }, { "epoch": 2.6665232666523266, "grad_norm": 2.6803221702575684, "learning_rate": 5.287243616332027e-06, "loss": 0.0145, "step": 4650 }, { "epoch": 2.6665232666523266, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10255688428878784, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2291, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.76, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.403, "step": 4650 }, { "epoch": 2.6722592672259267, "grad_norm": 0.03457150608301163, "learning_rate": 5.267246438842213e-06, "loss": 0.005, "step": 4660 }, { "epoch": 2.6779952677995267, "grad_norm": 4.120417594909668, "learning_rate": 5.247244973432524e-06, "loss": 0.0141, "step": 4670 }, { "epoch": 2.6837312683731267, "grad_norm": 0.005652038846164942, "learning_rate": 5.227239541022786e-06, "loss": 0.0034, "step": 4680 }, { "epoch": 2.6894672689467267, "grad_norm": 2.9149153232574463, "learning_rate": 5.2072304625964785e-06, "loss": 0.0061, "step": 4690 }, { "epoch": 2.695203269520327, "grad_norm": 0.0008329672855325043, "learning_rate": 5.187218059195578e-06, "loss": 0.0047, "step": 4700 }, { "epoch": 2.695203269520327, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10345083475112915, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2387, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.746, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.396, "step": 4700 }, { "epoch": 2.7009392700939268, "grad_norm": 1.024248480796814, "learning_rate": 5.167202651915409e-06, "loss": 0.0071, "step": 4710 }, { "epoch": 2.7066752706675272, "grad_norm": 0.014447176828980446, "learning_rate": 5.147184561899495e-06, "loss": 0.0076, "step": 4720 }, { "epoch": 2.7124112712411272, "grad_norm": 5.958902835845947, "learning_rate": 5.1271641103344e-06, "loss": 0.0116, "step": 4730 }, { "epoch": 2.7181472718147273, "grad_norm": 0.016329055652022362, "learning_rate": 5.1071416184445845e-06, "loss": 0.0105, "step": 4740 }, { "epoch": 2.7238832723883273, "grad_norm": 0.4175659716129303, "learning_rate": 5.087117407487239e-06, "loss": 0.0166, "step": 4750 }, { "epoch": 2.7238832723883273, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08455558121204376, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2591, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.716, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.382, "step": 4750 }, { "epoch": 2.7296192729619273, "grad_norm": 0.022596385329961777, "learning_rate": 5.06709179874714e-06, "loss": 0.0009, "step": 4760 }, { "epoch": 2.7353552735355273, "grad_norm": 0.0726088285446167, "learning_rate": 5.047065113531493e-06, "loss": 0.0254, "step": 4770 }, { "epoch": 2.7410912741091273, "grad_norm": 0.035875000059604645, "learning_rate": 5.02703767316477e-06, "loss": 0.0244, "step": 4780 }, { "epoch": 2.7468272746827274, "grad_norm": 3.32077956199646, "learning_rate": 5.00700979898356e-06, "loss": 0.006, "step": 4790 }, { "epoch": 2.7525632752563274, "grad_norm": 1.0301790237426758, "learning_rate": 4.98698181233142e-06, "loss": 0.0059, "step": 4800 }, { "epoch": 2.7525632752563274, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07655075937509537, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2238, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.767, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.407, "step": 4800 }, { "epoch": 2.758299275829928, "grad_norm": 0.446961373090744, "learning_rate": 4.966954034553699e-06, "loss": 0.0087, "step": 4810 }, { "epoch": 2.7640352764035274, "grad_norm": 0.33151447772979736, "learning_rate": 4.946926786992407e-06, "loss": 0.0009, "step": 4820 }, { "epoch": 2.769771276977128, "grad_norm": 0.12154054641723633, "learning_rate": 4.9269003909810405e-06, "loss": 0.0055, "step": 4830 }, { "epoch": 2.775507277550728, "grad_norm": 0.016543418169021606, "learning_rate": 4.906875167839433e-06, "loss": 0.0373, "step": 4840 }, { "epoch": 2.781243278124328, "grad_norm": 0.018002478405833244, "learning_rate": 4.886851438868599e-06, "loss": 0.0133, "step": 4850 }, { "epoch": 2.781243278124328, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07717321813106537, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.202, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.799, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.423, "step": 4850 }, { "epoch": 2.786979278697928, "grad_norm": 0.01707332953810692, "learning_rate": 4.866829525345585e-06, "loss": 0.0166, "step": 4860 }, { "epoch": 2.792715279271528, "grad_norm": 0.00804219115525484, "learning_rate": 4.846809748518302e-06, "loss": 0.0015, "step": 4870 }, { "epoch": 2.798451279845128, "grad_norm": 0.02389547787606716, "learning_rate": 4.826792429600381e-06, "loss": 0.0021, "step": 4880 }, { "epoch": 2.804187280418728, "grad_norm": 0.005223631393164396, "learning_rate": 4.806777889766016e-06, "loss": 0.0136, "step": 4890 }, { "epoch": 2.809923280992328, "grad_norm": 0.20156431198120117, "learning_rate": 4.78676645014481e-06, "loss": 0.0093, "step": 4900 }, { "epoch": 2.809923280992328, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0736735612154007, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.151, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.873, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.46, "step": 4900 }, { "epoch": 2.815659281565928, "grad_norm": 0.0024203194770962, "learning_rate": 4.766758431816629e-06, "loss": 0.0052, "step": 4910 }, { "epoch": 2.821395282139528, "grad_norm": 0.031604766845703125, "learning_rate": 4.746754155806437e-06, "loss": 0.0095, "step": 4920 }, { "epoch": 2.827131282713128, "grad_norm": 3.787029266357422, "learning_rate": 4.72675394307916e-06, "loss": 0.0139, "step": 4930 }, { "epoch": 2.8328672832867285, "grad_norm": 0.07122969627380371, "learning_rate": 4.7067581145345226e-06, "loss": 0.0139, "step": 4940 }, { "epoch": 2.838603283860328, "grad_norm": 3.0758485794067383, "learning_rate": 4.686766991001913e-06, "loss": 0.0161, "step": 4950 }, { "epoch": 2.838603283860328, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08140452206134796, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.151, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.873, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.46, "step": 4950 }, { "epoch": 2.8443392844339286, "grad_norm": 0.07921534031629562, "learning_rate": 4.666780893235227e-06, "loss": 0.0006, "step": 4960 }, { "epoch": 2.8500752850075286, "grad_norm": 0.12975841760635376, "learning_rate": 4.646800141907717e-06, "loss": 0.0154, "step": 4970 }, { "epoch": 2.8558112855811286, "grad_norm": 0.0262366384267807, "learning_rate": 4.626825057606859e-06, "loss": 0.0093, "step": 4980 }, { "epoch": 2.8615472861547286, "grad_norm": 0.00262279505841434, "learning_rate": 4.606855960829199e-06, "loss": 0.0202, "step": 4990 }, { "epoch": 2.8672832867283287, "grad_norm": 0.01955203339457512, "learning_rate": 4.586893171975218e-06, "loss": 0.0266, "step": 5000 }, { "epoch": 2.8672832867283287, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07921645790338516, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3258, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.62, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.334, "step": 5000 }, { "epoch": 2.8730192873019287, "grad_norm": 0.10377766937017441, "learning_rate": 4.566937011344184e-06, "loss": 0.0017, "step": 5010 }, { "epoch": 2.8787552878755287, "grad_norm": 0.784302294254303, "learning_rate": 4.546987799129018e-06, "loss": 0.0046, "step": 5020 }, { "epoch": 2.8844912884491287, "grad_norm": 2.035209894180298, "learning_rate": 4.527045855411153e-06, "loss": 0.0119, "step": 5030 }, { "epoch": 2.8902272890227287, "grad_norm": 3.6558942794799805, "learning_rate": 4.507111500155407e-06, "loss": 0.0123, "step": 5040 }, { "epoch": 2.895963289596329, "grad_norm": 3.73490309715271, "learning_rate": 4.487185053204832e-06, "loss": 0.0073, "step": 5050 }, { "epoch": 2.895963289596329, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07410901039838791, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3293, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.615, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.331, "step": 5050 }, { "epoch": 2.901699290169929, "grad_norm": 0.008407690562307835, "learning_rate": 4.467266834275601e-06, "loss": 0.0191, "step": 5060 }, { "epoch": 2.9074352907435292, "grad_norm": 0.03255997970700264, "learning_rate": 4.447357162951865e-06, "loss": 0.0458, "step": 5070 }, { "epoch": 2.9131712913171293, "grad_norm": 1.6957069635391235, "learning_rate": 4.427456358680635e-06, "loss": 0.0236, "step": 5080 }, { "epoch": 2.9189072918907293, "grad_norm": 0.1675146073102951, "learning_rate": 4.407564740766648e-06, "loss": 0.0177, "step": 5090 }, { "epoch": 2.9246432924643293, "grad_norm": 0.04229651764035225, "learning_rate": 4.3876826283672485e-06, "loss": 0.0116, "step": 5100 }, { "epoch": 2.9246432924643293, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06259483098983765, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.178, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.834, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.441, "step": 5100 }, { "epoch": 2.9303792930379293, "grad_norm": 0.04005512595176697, "learning_rate": 4.367810340487267e-06, "loss": 0.0099, "step": 5110 }, { "epoch": 2.9361152936115293, "grad_norm": 0.1425451785326004, "learning_rate": 4.347948195973901e-06, "loss": 0.0033, "step": 5120 }, { "epoch": 2.9418512941851294, "grad_norm": 0.10147175192832947, "learning_rate": 4.328096513511601e-06, "loss": 0.0076, "step": 5130 }, { "epoch": 2.9475872947587294, "grad_norm": 0.13407284021377563, "learning_rate": 4.308255611616954e-06, "loss": 0.0034, "step": 5140 }, { "epoch": 2.9533232953323294, "grad_norm": 0.12246271222829819, "learning_rate": 4.2884258086335755e-06, "loss": 0.0077, "step": 5150 }, { "epoch": 2.9533232953323294, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07015232741832733, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2126, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.784, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.415, "step": 5150 }, { "epoch": 2.95905929590593, "grad_norm": 0.06634827703237534, "learning_rate": 4.268607422727e-06, "loss": 0.0131, "step": 5160 }, { "epoch": 2.9647952964795294, "grad_norm": 2.3556761741638184, "learning_rate": 4.248800771879581e-06, "loss": 0.029, "step": 5170 }, { "epoch": 2.97053129705313, "grad_norm": 0.03471173346042633, "learning_rate": 4.229006173885381e-06, "loss": 0.0021, "step": 5180 }, { "epoch": 2.97626729762673, "grad_norm": 0.006328817922621965, "learning_rate": 4.2092239463450775e-06, "loss": 0.0212, "step": 5190 }, { "epoch": 2.98200329820033, "grad_norm": 0.02347049117088318, "learning_rate": 4.189454406660865e-06, "loss": 0.0056, "step": 5200 }, { "epoch": 2.98200329820033, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06551730632781982, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.5411, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.314, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.18, "step": 5200 }, { "epoch": 2.98773929877393, "grad_norm": 0.012669041752815247, "learning_rate": 4.169697872031368e-06, "loss": 0.012, "step": 5210 }, { "epoch": 2.99347529934753, "grad_norm": 0.07477905601263046, "learning_rate": 4.1499546594465465e-06, "loss": 0.0215, "step": 5220 }, { "epoch": 2.99921129992113, "grad_norm": 1.7680764198303223, "learning_rate": 4.1302250856826045e-06, "loss": 0.0107, "step": 5230 }, { "epoch": 3.00458880045888, "grad_norm": 0.19178365170955658, "learning_rate": 4.110509467296923e-06, "loss": 0.006, "step": 5240 }, { "epoch": 3.01032480103248, "grad_norm": 0.025907142087817192, "learning_rate": 4.090808120622961e-06, "loss": 0.0139, "step": 5250 }, { "epoch": 3.01032480103248, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07205013185739517, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2727, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.697, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.372, "step": 5250 }, { "epoch": 3.01606080160608, "grad_norm": 0.018714534118771553, "learning_rate": 4.071121361765201e-06, "loss": 0.0033, "step": 5260 }, { "epoch": 3.02179680217968, "grad_norm": 2.1166880130767822, "learning_rate": 4.05144950659406e-06, "loss": 0.0081, "step": 5270 }, { "epoch": 3.02753280275328, "grad_norm": 0.07031437754631042, "learning_rate": 4.031792870740831e-06, "loss": 0.0086, "step": 5280 }, { "epoch": 3.03326880332688, "grad_norm": 0.24249830842018127, "learning_rate": 4.012151769592612e-06, "loss": 0.0132, "step": 5290 }, { "epoch": 3.0390048039004802, "grad_norm": 0.009602251462638378, "learning_rate": 3.992526518287258e-06, "loss": 0.0075, "step": 5300 }, { "epoch": 3.0390048039004802, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07945062220096588, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1773, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.835, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.441, "step": 5300 }, { "epoch": 3.0447408044740802, "grad_norm": 0.1440117210149765, "learning_rate": 3.972917431708305e-06, "loss": 0.0013, "step": 5310 }, { "epoch": 3.0504768050476807, "grad_norm": 0.013155910186469555, "learning_rate": 3.9533248244799375e-06, "loss": 0.0049, "step": 5320 }, { "epoch": 3.0562128056212807, "grad_norm": 0.1568404585123062, "learning_rate": 3.933749010961927e-06, "loss": 0.0024, "step": 5330 }, { "epoch": 3.0619488061948807, "grad_norm": 0.00445369491353631, "learning_rate": 3.914190305244595e-06, "loss": 0.0024, "step": 5340 }, { "epoch": 3.0676848067684808, "grad_norm": 0.15039828419685364, "learning_rate": 3.8946490211437735e-06, "loss": 0.02, "step": 5350 }, { "epoch": 3.0676848067684808, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08091636747121811, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2042, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.796, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.421, "step": 5350 }, { "epoch": 3.073420807342081, "grad_norm": 0.13300937414169312, "learning_rate": 3.875125472195764e-06, "loss": 0.0012, "step": 5360 }, { "epoch": 3.079156807915681, "grad_norm": 2.184394121170044, "learning_rate": 3.855619971652314e-06, "loss": 0.0117, "step": 5370 }, { "epoch": 3.084892808489281, "grad_norm": 0.02020282857120037, "learning_rate": 3.836132832475583e-06, "loss": 0.0014, "step": 5380 }, { "epoch": 3.090628809062881, "grad_norm": 0.008466456085443497, "learning_rate": 3.816664367333131e-06, "loss": 0.009, "step": 5390 }, { "epoch": 3.096364809636481, "grad_norm": 5.091494560241699, "learning_rate": 3.797214888592896e-06, "loss": 0.0128, "step": 5400 }, { "epoch": 3.096364809636481, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07716992497444153, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2065, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.792, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.42, "step": 5400 }, { "epoch": 3.102100810210081, "grad_norm": 0.20820866525173187, "learning_rate": 3.777784708318178e-06, "loss": 0.0184, "step": 5410 }, { "epoch": 3.107836810783681, "grad_norm": 1.1382921934127808, "learning_rate": 3.7583741382626402e-06, "loss": 0.0093, "step": 5420 }, { "epoch": 3.113572811357281, "grad_norm": 0.38132885098457336, "learning_rate": 3.7389834898653067e-06, "loss": 0.0117, "step": 5430 }, { "epoch": 3.1193088119308814, "grad_norm": 0.009956207126379013, "learning_rate": 3.719613074245555e-06, "loss": 0.0017, "step": 5440 }, { "epoch": 3.1250448125044814, "grad_norm": 5.792750358581543, "learning_rate": 3.7002632021981368e-06, "loss": 0.0146, "step": 5450 }, { "epoch": 3.1250448125044814, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0894627794623375, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2897, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.672, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.36, "step": 5450 }, { "epoch": 3.1307808130780814, "grad_norm": 1.089165210723877, "learning_rate": 3.680934184188182e-06, "loss": 0.0032, "step": 5460 }, { "epoch": 3.1365168136516814, "grad_norm": 0.020885087549686432, "learning_rate": 3.661626330346224e-06, "loss": 0.0037, "step": 5470 }, { "epoch": 3.1422528142252815, "grad_norm": 0.09390339255332947, "learning_rate": 3.642339950463224e-06, "loss": 0.0083, "step": 5480 }, { "epoch": 3.1479888147988815, "grad_norm": 0.0025120277423411608, "learning_rate": 3.62307535398559e-06, "loss": 0.0108, "step": 5490 }, { "epoch": 3.1537248153724815, "grad_norm": 0.06790509074926376, "learning_rate": 3.603832850010226e-06, "loss": 0.0012, "step": 5500 }, { "epoch": 3.1537248153724815, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08822309225797653, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3502, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.585, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.316, "step": 5500 }, { "epoch": 3.1594608159460815, "grad_norm": 0.07959645241498947, "learning_rate": 3.58461274727956e-06, "loss": 0.0075, "step": 5510 }, { "epoch": 3.1651968165196815, "grad_norm": 3.0567445755004883, "learning_rate": 3.5654153541766023e-06, "loss": 0.008, "step": 5520 }, { "epoch": 3.1709328170932816, "grad_norm": 0.20890028774738312, "learning_rate": 3.546240978719988e-06, "loss": 0.0006, "step": 5530 }, { "epoch": 3.1766688176668816, "grad_norm": 2.681074619293213, "learning_rate": 3.5270899285590375e-06, "loss": 0.0027, "step": 5540 }, { "epoch": 3.1824048182404816, "grad_norm": 0.04251394048333168, "learning_rate": 3.5079625109688243e-06, "loss": 0.0005, "step": 5550 }, { "epoch": 3.1824048182404816, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09866516292095184, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2324, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.755, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.401, "step": 5550 }, { "epoch": 3.188140818814082, "grad_norm": 4.062971591949463, "learning_rate": 3.4888590328452353e-06, "loss": 0.0178, "step": 5560 }, { "epoch": 3.193876819387682, "grad_norm": 0.0014892960898578167, "learning_rate": 3.4697798007000624e-06, "loss": 0.0126, "step": 5570 }, { "epoch": 3.199612819961282, "grad_norm": 0.005618450231850147, "learning_rate": 3.450725120656069e-06, "loss": 0.0261, "step": 5580 }, { "epoch": 3.205348820534882, "grad_norm": 1.2523528337478638, "learning_rate": 3.431695298442084e-06, "loss": 0.0061, "step": 5590 }, { "epoch": 3.211084821108482, "grad_norm": 3.6290197372436523, "learning_rate": 3.4126906393881e-06, "loss": 0.0177, "step": 5600 }, { "epoch": 3.211084821108482, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07931295037269592, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2163, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.778, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.413, "step": 5600 }, { "epoch": 3.216820821682082, "grad_norm": 0.015163871459662914, "learning_rate": 3.393711448420372e-06, "loss": 0.0057, "step": 5610 }, { "epoch": 3.222556822255682, "grad_norm": 0.059904176741838455, "learning_rate": 3.37475803005652e-06, "loss": 0.0017, "step": 5620 }, { "epoch": 3.228292822829282, "grad_norm": 0.3671509623527527, "learning_rate": 3.355830688400652e-06, "loss": 0.0211, "step": 5630 }, { "epoch": 3.2340288234028822, "grad_norm": 2.3451545238494873, "learning_rate": 3.336929727138474e-06, "loss": 0.0221, "step": 5640 }, { "epoch": 3.2397648239764822, "grad_norm": 0.00577126070857048, "learning_rate": 3.31805544953243e-06, "loss": 0.0082, "step": 5650 }, { "epoch": 3.2397648239764822, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07332266122102737, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1867, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.821, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, "step": 5650 }, { "epoch": 3.2455008245500823, "grad_norm": 1.244507074356079, "learning_rate": 3.299208158416829e-06, "loss": 0.0177, "step": 5660 }, { "epoch": 3.2512368251236827, "grad_norm": 1.7493702173233032, "learning_rate": 3.2803881561929806e-06, "loss": 0.0182, "step": 5670 }, { "epoch": 3.2569728256972827, "grad_norm": 0.03632393479347229, "learning_rate": 3.2615957448243562e-06, "loss": 0.0075, "step": 5680 }, { "epoch": 3.2627088262708828, "grad_norm": 0.021715901792049408, "learning_rate": 3.2428312258317306e-06, "loss": 0.0044, "step": 5690 }, { "epoch": 3.268444826844483, "grad_norm": 0.7747498154640198, "learning_rate": 3.224094900288357e-06, "loss": 0.0071, "step": 5700 }, { "epoch": 3.268444826844483, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08630760759115219, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1329, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.9, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.474, "step": 5700 }, { "epoch": 3.274180827418083, "grad_norm": 0.003438598709180951, "learning_rate": 3.205387068815127e-06, "loss": 0.0025, "step": 5710 }, { "epoch": 3.279916827991683, "grad_norm": 0.02451498433947563, "learning_rate": 3.1867080315757477e-06, "loss": 0.0004, "step": 5720 }, { "epoch": 3.285652828565283, "grad_norm": 0.4422462582588196, "learning_rate": 3.1680580882719304e-06, "loss": 0.0206, "step": 5730 }, { "epoch": 3.291388829138883, "grad_norm": 3.2265453338623047, "learning_rate": 3.149437538138583e-06, "loss": 0.0142, "step": 5740 }, { "epoch": 3.297124829712483, "grad_norm": 0.015962867066264153, "learning_rate": 3.130846679938998e-06, "loss": 0.0099, "step": 5750 }, { "epoch": 3.297124829712483, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07895597815513611, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1289, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.906, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.476, "step": 5750 }, { "epoch": 3.302860830286083, "grad_norm": 0.470051109790802, "learning_rate": 3.1122858119600726e-06, "loss": 0.0031, "step": 5760 }, { "epoch": 3.308596830859683, "grad_norm": 0.002947809174656868, "learning_rate": 3.0937552320075116e-06, "loss": 0.0005, "step": 5770 }, { "epoch": 3.3143328314332834, "grad_norm": 4.272603511810303, "learning_rate": 3.0752552374010567e-06, "loss": 0.0185, "step": 5780 }, { "epoch": 3.320068832006883, "grad_norm": 0.0015301775420084596, "learning_rate": 3.056786124969716e-06, "loss": 0.0012, "step": 5790 }, { "epoch": 3.3258048325804834, "grad_norm": 0.002659064019098878, "learning_rate": 3.0383481910469936e-06, "loss": 0.004, "step": 5800 }, { "epoch": 3.3258048325804834, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08382538706064224, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1553, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.867, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.457, "step": 5800 }, { "epoch": 3.3315408331540834, "grad_norm": 0.25469204783439636, "learning_rate": 3.019941731466143e-06, "loss": 0.0005, "step": 5810 }, { "epoch": 3.3372768337276835, "grad_norm": 0.5178417563438416, "learning_rate": 3.0015670415554143e-06, "loss": 0.016, "step": 5820 }, { "epoch": 3.3430128343012835, "grad_norm": 0.07551216334104538, "learning_rate": 2.9832244161333257e-06, "loss": 0.0202, "step": 5830 }, { "epoch": 3.3487488348748835, "grad_norm": 0.008605745621025562, "learning_rate": 2.9649141495039225e-06, "loss": 0.0009, "step": 5840 }, { "epoch": 3.3544848354484835, "grad_norm": 0.0697212815284729, "learning_rate": 2.9466365354520564e-06, "loss": 0.0204, "step": 5850 }, { "epoch": 3.3544848354484835, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07445970177650452, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2129, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.783, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.415, "step": 5850 }, { "epoch": 3.3602208360220835, "grad_norm": 0.017461596056818962, "learning_rate": 2.928391867238679e-06, "loss": 0.0012, "step": 5860 }, { "epoch": 3.3659568365956836, "grad_norm": 0.0013357176212593913, "learning_rate": 2.910180437596132e-06, "loss": 0.0036, "step": 5870 }, { "epoch": 3.3716928371692836, "grad_norm": 0.00179756130091846, "learning_rate": 2.8920025387234484e-06, "loss": 0.0131, "step": 5880 }, { "epoch": 3.3774288377428836, "grad_norm": 0.005055127199739218, "learning_rate": 2.8738584622816656e-06, "loss": 0.0133, "step": 5890 }, { "epoch": 3.3831648383164836, "grad_norm": 0.778300940990448, "learning_rate": 2.8557484993891484e-06, "loss": 0.007, "step": 5900 }, { "epoch": 3.3831648383164836, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08228859305381775, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2282, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.761, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.404, "step": 5900 }, { "epoch": 3.388900838890084, "grad_norm": 0.05169231817126274, "learning_rate": 2.837672940616911e-06, "loss": 0.0088, "step": 5910 }, { "epoch": 3.394636839463684, "grad_norm": 0.0016457217279821634, "learning_rate": 2.8196320759839677e-06, "loss": 0.0173, "step": 5920 }, { "epoch": 3.400372840037284, "grad_norm": 0.11143210530281067, "learning_rate": 2.801626194952669e-06, "loss": 0.0015, "step": 5930 }, { "epoch": 3.406108840610884, "grad_norm": 0.03351111710071564, "learning_rate": 2.7836555864240566e-06, "loss": 0.0079, "step": 5940 }, { "epoch": 3.411844841184484, "grad_norm": 0.22604842483997345, "learning_rate": 2.7657205387332313e-06, "loss": 0.0124, "step": 5950 }, { "epoch": 3.411844841184484, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08073882758617401, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2042, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.796, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.421, "step": 5950 }, { "epoch": 3.417580841758084, "grad_norm": 0.1360362470149994, "learning_rate": 2.747821339644734e-06, "loss": 0.028, "step": 5960 }, { "epoch": 3.423316842331684, "grad_norm": 0.015132960863411427, "learning_rate": 2.7299582763479115e-06, "loss": 0.0041, "step": 5970 }, { "epoch": 3.4290528429052842, "grad_norm": 0.06472807377576828, "learning_rate": 2.7121316354523286e-06, "loss": 0.0017, "step": 5980 }, { "epoch": 3.4347888434788842, "grad_norm": 3.1584267616271973, "learning_rate": 2.6943417029831504e-06, "loss": 0.0232, "step": 5990 }, { "epoch": 3.4405248440524843, "grad_norm": 0.004198137205094099, "learning_rate": 2.6765887643765653e-06, "loss": 0.0052, "step": 6000 }, { "epoch": 3.4405248440524843, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08295302093029022, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2169, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.777, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.412, "step": 6000 }, { "epoch": 3.4462608446260843, "grad_norm": 0.018915435299277306, "learning_rate": 2.6588731044752057e-06, "loss": 0.0153, "step": 6010 }, { "epoch": 3.4519968451996847, "grad_norm": 3.4281928539276123, "learning_rate": 2.641195007523568e-06, "loss": 0.0099, "step": 6020 }, { "epoch": 3.4577328457732848, "grad_norm": 0.0030478844419121742, "learning_rate": 2.623554757163464e-06, "loss": 0.0004, "step": 6030 }, { "epoch": 3.463468846346885, "grad_norm": 0.022162869572639465, "learning_rate": 2.605952636429456e-06, "loss": 0.0114, "step": 6040 }, { "epoch": 3.469204846920485, "grad_norm": 0.009365621022880077, "learning_rate": 2.5883889277443347e-06, "loss": 0.0009, "step": 6050 }, { "epoch": 3.469204846920485, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07579995691776276, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2596, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.716, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.381, "step": 6050 }, { "epoch": 3.474940847494085, "grad_norm": 0.29800668358802795, "learning_rate": 2.570863912914566e-06, "loss": 0.018, "step": 6060 }, { "epoch": 3.480676848067685, "grad_norm": 3.4592084884643555, "learning_rate": 2.5533778731257824e-06, "loss": 0.0041, "step": 6070 }, { "epoch": 3.486412848641285, "grad_norm": 0.03675074130296707, "learning_rate": 2.535931088938274e-06, "loss": 0.0129, "step": 6080 }, { "epoch": 3.492148849214885, "grad_norm": 0.5131600499153137, "learning_rate": 2.518523840282479e-06, "loss": 0.0069, "step": 6090 }, { "epoch": 3.497884849788485, "grad_norm": 0.19048012793064117, "learning_rate": 2.5011564064544945e-06, "loss": 0.0061, "step": 6100 }, { "epoch": 3.497884849788485, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07761486619710922, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2658, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.707, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.377, "step": 6100 }, { "epoch": 3.503620850362085, "grad_norm": 0.37152099609375, "learning_rate": 2.483829066111596e-06, "loss": 0.0108, "step": 6110 }, { "epoch": 3.509356850935685, "grad_norm": 4.403800964355469, "learning_rate": 2.466542097267765e-06, "loss": 0.0119, "step": 6120 }, { "epoch": 3.5150928515092854, "grad_norm": 0.008035624399781227, "learning_rate": 2.4492957772892345e-06, "loss": 0.0192, "step": 6130 }, { "epoch": 3.520828852082885, "grad_norm": 1.8271785974502563, "learning_rate": 2.432090382890032e-06, "loss": 0.0159, "step": 6140 }, { "epoch": 3.5265648526564854, "grad_norm": 0.00973561592400074, "learning_rate": 2.4149261901275393e-06, "loss": 0.0184, "step": 6150 }, { "epoch": 3.5265648526564854, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07000607997179031, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3198, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.629, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.338, "step": 6150 }, { "epoch": 3.5323008532300855, "grad_norm": 0.3272586464881897, "learning_rate": 2.3978034743980676e-06, "loss": 0.0014, "step": 6160 }, { "epoch": 3.5380368538036855, "grad_norm": 0.015382179990410805, "learning_rate": 2.3807225104324337e-06, "loss": 0.0099, "step": 6170 }, { "epoch": 3.5437728543772855, "grad_norm": 0.016446420922875404, "learning_rate": 2.36368357229156e-06, "loss": 0.0039, "step": 6180 }, { "epoch": 3.5495088549508855, "grad_norm": 0.006563331466168165, "learning_rate": 2.34668693336207e-06, "loss": 0.0035, "step": 6190 }, { "epoch": 3.5552448555244855, "grad_norm": 0.024284416809678078, "learning_rate": 2.3297328663519044e-06, "loss": 0.0117, "step": 6200 }, { "epoch": 3.5552448555244855, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07584439218044281, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3203, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.628, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.337, "step": 6200 }, { "epoch": 3.5609808560980856, "grad_norm": 0.011505261063575745, "learning_rate": 2.3128216432859435e-06, "loss": 0.0192, "step": 6210 }, { "epoch": 3.5667168566716856, "grad_norm": 0.01848476566374302, "learning_rate": 2.2959535355016436e-06, "loss": 0.0003, "step": 6220 }, { "epoch": 3.5724528572452856, "grad_norm": 0.007101431954652071, "learning_rate": 2.27912881364469e-06, "loss": 0.0098, "step": 6230 }, { "epoch": 3.5781888578188856, "grad_norm": 0.9875366687774658, "learning_rate": 2.2623477476646447e-06, "loss": 0.0014, "step": 6240 }, { "epoch": 3.5839248583924856, "grad_norm": 0.07821710407733917, "learning_rate": 2.2456106068106206e-06, "loss": 0.0007, "step": 6250 }, { "epoch": 3.5839248583924856, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08138631284236908, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2288, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.76, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.404, "step": 6250 }, { "epoch": 3.589660858966086, "grad_norm": 0.8070231676101685, "learning_rate": 2.2289176596269545e-06, "loss": 0.0015, "step": 6260 }, { "epoch": 3.5953968595396857, "grad_norm": 0.0035041391383856535, "learning_rate": 2.2122691739489143e-06, "loss": 0.0021, "step": 6270 }, { "epoch": 3.601132860113286, "grad_norm": 0.013247767463326454, "learning_rate": 2.195665416898381e-06, "loss": 0.0026, "step": 6280 }, { "epoch": 3.606868860686886, "grad_norm": 3.3201675415039062, "learning_rate": 2.179106654879581e-06, "loss": 0.0236, "step": 6290 }, { "epoch": 3.612604861260486, "grad_norm": 0.00799557100981474, "learning_rate": 2.1625931535747964e-06, "loss": 0.0078, "step": 6300 }, { "epoch": 3.612604861260486, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07524015754461288, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3312, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.613, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.33, "step": 6300 }, { "epoch": 3.618340861834086, "grad_norm": 0.796405553817749, "learning_rate": 2.1461251779401175e-06, "loss": 0.0028, "step": 6310 }, { "epoch": 3.6240768624076862, "grad_norm": 0.0037477288860827684, "learning_rate": 2.1297029922011775e-06, "loss": 0.0093, "step": 6320 }, { "epoch": 3.6298128629812862, "grad_norm": 0.07188570499420166, "learning_rate": 2.113326859848919e-06, "loss": 0.0031, "step": 6330 }, { "epoch": 3.6355488635548863, "grad_norm": 2.5588748455047607, "learning_rate": 2.0969970436353725e-06, "loss": 0.0074, "step": 6340 }, { "epoch": 3.6412848641284863, "grad_norm": 0.008328588679432869, "learning_rate": 2.080713805569427e-06, "loss": 0.0025, "step": 6350 }, { "epoch": 3.6412848641284863, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07739181071519852, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2601, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.715, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.381, "step": 6350 }, { "epoch": 3.6470208647020863, "grad_norm": 0.0075204516761004925, "learning_rate": 2.0644774069126406e-06, "loss": 0.0109, "step": 6360 }, { "epoch": 3.6527568652756868, "grad_norm": 0.003392142942175269, "learning_rate": 2.048288108175036e-06, "loss": 0.0091, "step": 6370 }, { "epoch": 3.6584928658492863, "grad_norm": 0.00919413473457098, "learning_rate": 2.0321461691109288e-06, "loss": 0.007, "step": 6380 }, { "epoch": 3.664228866422887, "grad_norm": 0.020889578387141228, "learning_rate": 2.016051848714758e-06, "loss": 0.0012, "step": 6390 }, { "epoch": 3.669964866996487, "grad_norm": 0.09611943364143372, "learning_rate": 2.000005405216931e-06, "loss": 0.0011, "step": 6400 }, { "epoch": 3.669964866996487, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08481152355670929, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2176, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.776, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.412, "step": 6400 }, { "epoch": 3.675700867570087, "grad_norm": 0.004153054673224688, "learning_rate": 1.984007096079676e-06, "loss": 0.0175, "step": 6410 }, { "epoch": 3.681436868143687, "grad_norm": 0.0063193319365382195, "learning_rate": 1.968057177992915e-06, "loss": 0.0018, "step": 6420 }, { "epoch": 3.687172868717287, "grad_norm": 0.05629734694957733, "learning_rate": 1.9521559068701447e-06, "loss": 0.0005, "step": 6430 }, { "epoch": 3.692908869290887, "grad_norm": 0.042965829372406006, "learning_rate": 1.9363035378443313e-06, "loss": 0.0017, "step": 6440 }, { "epoch": 3.698644869864487, "grad_norm": 0.029728276655077934, "learning_rate": 1.9205003252638176e-06, "loss": 0.0041, "step": 6450 }, { "epoch": 3.698644869864487, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08883155882358551, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1133, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.928, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.488, "step": 6450 }, { "epoch": 3.704380870438087, "grad_norm": 0.014888194389641285, "learning_rate": 1.904746522688236e-06, "loss": 0.0018, "step": 6460 }, { "epoch": 3.710116871011687, "grad_norm": 4.221490859985352, "learning_rate": 1.8890423828844462e-06, "loss": 0.0047, "step": 6470 }, { "epoch": 3.7158528715852874, "grad_norm": 2.9691474437713623, "learning_rate": 1.873388157822477e-06, "loss": 0.0087, "step": 6480 }, { "epoch": 3.721588872158887, "grad_norm": 0.1171032041311264, "learning_rate": 1.8577840986714868e-06, "loss": 0.0019, "step": 6490 }, { "epoch": 3.7273248727324875, "grad_norm": 0.006826687604188919, "learning_rate": 1.8422304557957315e-06, "loss": 0.0123, "step": 6500 }, { "epoch": 3.7273248727324875, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08897148817777634, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2955, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.664, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.355, "step": 6500 }, { "epoch": 3.7330608733060875, "grad_norm": 0.022587748244404793, "learning_rate": 1.8267274787505446e-06, "loss": 0.0087, "step": 6510 }, { "epoch": 3.7387968738796875, "grad_norm": 0.5572834014892578, "learning_rate": 1.8112754162783336e-06, "loss": 0.0084, "step": 6520 }, { "epoch": 3.7445328744532875, "grad_norm": 0.015338007360696793, "learning_rate": 1.7958745163045987e-06, "loss": 0.0085, "step": 6530 }, { "epoch": 3.7502688750268875, "grad_norm": 0.004370348062366247, "learning_rate": 1.7805250259339379e-06, "loss": 0.0075, "step": 6540 }, { "epoch": 3.7560048756004876, "grad_norm": 0.006390043999999762, "learning_rate": 1.765227191446101e-06, "loss": 0.0074, "step": 6550 }, { "epoch": 3.7560048756004876, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08355527371168137, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1302, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.904, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.475, "step": 6550 }, { "epoch": 3.7617408761740876, "grad_norm": 0.09934690594673157, "learning_rate": 1.7499812582920222e-06, "loss": 0.0038, "step": 6560 }, { "epoch": 3.7674768767476876, "grad_norm": 0.0005312475841492414, "learning_rate": 1.734787471089887e-06, "loss": 0.005, "step": 6570 }, { "epoch": 3.7732128773212876, "grad_norm": 0.5943177938461304, "learning_rate": 1.7196460736212167e-06, "loss": 0.0111, "step": 6580 }, { "epoch": 3.7789488778948876, "grad_norm": 0.03382878378033638, "learning_rate": 1.7045573088269408e-06, "loss": 0.0022, "step": 6590 }, { "epoch": 3.7846848784684877, "grad_norm": 0.018695516511797905, "learning_rate": 1.6895214188035125e-06, "loss": 0.0038, "step": 6600 }, { "epoch": 3.7846848784684877, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0897902175784111, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2015, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.8, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.423, "step": 6600 }, { "epoch": 3.790420879042088, "grad_norm": 0.03567569702863693, "learning_rate": 1.6745386447990153e-06, "loss": 0.0149, "step": 6610 }, { "epoch": 3.7961568796156877, "grad_norm": 0.005937593523412943, "learning_rate": 1.6596092272093e-06, "loss": 0.0057, "step": 6620 }, { "epoch": 3.801892880189288, "grad_norm": 0.001366928219795227, "learning_rate": 1.6447334055741198e-06, "loss": 0.0019, "step": 6630 }, { "epoch": 3.807628880762888, "grad_norm": 0.017239965498447418, "learning_rate": 1.6299114185732918e-06, "loss": 0.0068, "step": 6640 }, { "epoch": 3.813364881336488, "grad_norm": 0.09475360810756683, "learning_rate": 1.6151435040228663e-06, "loss": 0.0084, "step": 6650 }, { "epoch": 3.813364881336488, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09376490861177444, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2936, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.667, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.357, "step": 6650 }, { "epoch": 3.8191008819100882, "grad_norm": 0.004428621847182512, "learning_rate": 1.6004298988713153e-06, "loss": 0.0003, "step": 6660 }, { "epoch": 3.8248368824836882, "grad_norm": 0.0019930372945964336, "learning_rate": 1.5857708391957194e-06, "loss": 0.0012, "step": 6670 }, { "epoch": 3.8305728830572883, "grad_norm": 0.24852816760540009, "learning_rate": 1.571166560197991e-06, "loss": 0.0018, "step": 6680 }, { "epoch": 3.8363088836308883, "grad_norm": 0.05106634274125099, "learning_rate": 1.5566172962010944e-06, "loss": 0.0088, "step": 6690 }, { "epoch": 3.8420448842044883, "grad_norm": 0.957535982131958, "learning_rate": 1.542123280645292e-06, "loss": 0.0013, "step": 6700 }, { "epoch": 3.8420448842044883, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09245079010725021, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1154, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.925, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.486, "step": 6700 }, { "epoch": 3.8477808847780883, "grad_norm": 0.022298647090792656, "learning_rate": 1.527684746084394e-06, "loss": 0.0103, "step": 6710 }, { "epoch": 3.853516885351689, "grad_norm": 0.21422079205513, "learning_rate": 1.5133019241820257e-06, "loss": 0.0341, "step": 6720 }, { "epoch": 3.8592528859252884, "grad_norm": 0.08396855741739273, "learning_rate": 1.4989750457079156e-06, "loss": 0.0033, "step": 6730 }, { "epoch": 3.864988886498889, "grad_norm": 0.12731043994426727, "learning_rate": 1.484704340534187e-06, "loss": 0.0028, "step": 6740 }, { "epoch": 3.870724887072489, "grad_norm": 0.4012109637260437, "learning_rate": 1.4704900376316773e-06, "loss": 0.0008, "step": 6750 }, { "epoch": 3.870724887072489, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09105575829744339, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2863, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.677, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, "step": 6750 }, { "epoch": 3.876460887646089, "grad_norm": 0.4341152012348175, "learning_rate": 1.4563323650662586e-06, "loss": 0.0013, "step": 6760 }, { "epoch": 3.882196888219689, "grad_norm": 0.14372624456882477, "learning_rate": 1.4422315499951783e-06, "loss": 0.0059, "step": 6770 }, { "epoch": 3.887932888793289, "grad_norm": 0.010315956547856331, "learning_rate": 1.4281878186634157e-06, "loss": 0.01, "step": 6780 }, { "epoch": 3.893668889366889, "grad_norm": 0.1287292093038559, "learning_rate": 1.4142013964000513e-06, "loss": 0.01, "step": 6790 }, { "epoch": 3.899404889940489, "grad_norm": 0.4034357964992523, "learning_rate": 1.400272507614655e-06, "loss": 0.0012, "step": 6800 }, { "epoch": 3.899404889940489, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09034840762615204, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1486, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.877, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.462, "step": 6800 }, { "epoch": 3.905140890514089, "grad_norm": 0.15965530276298523, "learning_rate": 1.3864013757936828e-06, "loss": 0.0009, "step": 6810 }, { "epoch": 3.910876891087689, "grad_norm": 0.07124567031860352, "learning_rate": 1.3725882234968879e-06, "loss": 0.0118, "step": 6820 }, { "epoch": 3.916612891661289, "grad_norm": 0.022812863811850548, "learning_rate": 1.3588332723537523e-06, "loss": 0.0133, "step": 6830 }, { "epoch": 3.922348892234889, "grad_norm": 0.0023223236203193665, "learning_rate": 1.345136743059936e-06, "loss": 0.013, "step": 6840 }, { "epoch": 3.9280848928084895, "grad_norm": 0.08886972069740295, "learning_rate": 1.3314988553737256e-06, "loss": 0.0028, "step": 6850 }, { "epoch": 3.9280848928084895, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0912175104022026, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2896, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.672, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.36, "step": 6850 }, { "epoch": 3.933820893382089, "grad_norm": 0.022729717195034027, "learning_rate": 1.3179198281125188e-06, "loss": 0.0069, "step": 6860 }, { "epoch": 3.9395568939556895, "grad_norm": 2.09700083732605, "learning_rate": 1.3043998791493034e-06, "loss": 0.006, "step": 6870 }, { "epoch": 3.9452928945292896, "grad_norm": 0.026967119425535202, "learning_rate": 1.290939225409173e-06, "loss": 0.0018, "step": 6880 }, { "epoch": 3.9510288951028896, "grad_norm": 0.010416003875434399, "learning_rate": 1.277538082865835e-06, "loss": 0.012, "step": 6890 }, { "epoch": 3.9567648956764896, "grad_norm": 0.04778837040066719, "learning_rate": 1.2641966665381517e-06, "loss": 0.0117, "step": 6900 }, { "epoch": 3.9567648956764896, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09045156091451645, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1714, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.844, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.445, "step": 6900 }, { "epoch": 3.9625008962500896, "grad_norm": 0.0013466872042044997, "learning_rate": 1.2509151904866922e-06, "loss": 0.0012, "step": 6910 }, { "epoch": 3.9682368968236896, "grad_norm": 0.057207245379686356, "learning_rate": 1.2376938678102913e-06, "loss": 0.0172, "step": 6920 }, { "epoch": 3.9739728973972896, "grad_norm": 2.2670977115631104, "learning_rate": 1.2245329106426367e-06, "loss": 0.013, "step": 6930 }, { "epoch": 3.9797088979708897, "grad_norm": 0.0011907550506293774, "learning_rate": 1.2114325301488616e-06, "loss": 0.0117, "step": 6940 }, { "epoch": 3.9854448985444897, "grad_norm": 0.023963019251823425, "learning_rate": 1.1983929365221548e-06, "loss": 0.0096, "step": 6950 }, { "epoch": 3.9854448985444897, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08352689445018768, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2306, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.757, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.402, "step": 6950 }, { "epoch": 3.99118089911809, "grad_norm": 0.02613544650375843, "learning_rate": 1.1854143389803962e-06, "loss": 0.0104, "step": 6960 }, { "epoch": 3.9969168996916897, "grad_norm": 0.004353563766926527, "learning_rate": 1.1724969457627928e-06, "loss": 0.0004, "step": 6970 }, { "epoch": 4.00229440022944, "grad_norm": 0.013296003453433514, "learning_rate": 1.1596409641265376e-06, "loss": 0.0053, "step": 6980 }, { "epoch": 4.00803040080304, "grad_norm": 0.01596342958509922, "learning_rate": 1.146846600343488e-06, "loss": 0.0013, "step": 6990 }, { "epoch": 4.01376640137664, "grad_norm": 0.003088761121034622, "learning_rate": 1.1341140596968525e-06, "loss": 0.0005, "step": 7000 }, { "epoch": 4.01376640137664, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08817622810602188, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2983, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.66, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.353, "step": 7000 }, { "epoch": 4.01950240195024, "grad_norm": 0.0804172232747078, "learning_rate": 1.1214435464779006e-06, "loss": 0.0003, "step": 7010 }, { "epoch": 4.02523840252384, "grad_norm": 0.13886569440364838, "learning_rate": 1.1088352639826844e-06, "loss": 0.0006, "step": 7020 }, { "epoch": 4.03097440309744, "grad_norm": 0.029498988762497902, "learning_rate": 1.0962894145087715e-06, "loss": 0.0058, "step": 7030 }, { "epoch": 4.03671040367104, "grad_norm": 3.7734415531158447, "learning_rate": 1.0838061993520072e-06, "loss": 0.0125, "step": 7040 }, { "epoch": 4.04244640424464, "grad_norm": 0.12142331153154373, "learning_rate": 1.0713858188032755e-06, "loss": 0.0161, "step": 7050 }, { "epoch": 4.04244640424464, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09039638936519623, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2759, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.692, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.369, "step": 7050 }, { "epoch": 4.048182404818241, "grad_norm": 0.003838218515738845, "learning_rate": 1.0590284721452965e-06, "loss": 0.0076, "step": 7060 }, { "epoch": 4.0539184053918405, "grad_norm": 0.0031382606830447912, "learning_rate": 1.0467343576494215e-06, "loss": 0.0007, "step": 7070 }, { "epoch": 4.059654405965441, "grad_norm": 0.0018334095366299152, "learning_rate": 1.0345036725724517e-06, "loss": 0.0206, "step": 7080 }, { "epoch": 4.0653904065390405, "grad_norm": 0.06359425187110901, "learning_rate": 1.0223366131534746e-06, "loss": 0.0004, "step": 7090 }, { "epoch": 4.071126407112641, "grad_norm": 0.08629762381315231, "learning_rate": 1.010233374610719e-06, "loss": 0.0003, "step": 7100 }, { "epoch": 4.071126407112641, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09012720733880997, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2717, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.698, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.373, "step": 7100 }, { "epoch": 4.0768624076862405, "grad_norm": 0.16138513386249542, "learning_rate": 9.981941511384152e-07, "loss": 0.0042, "step": 7110 }, { "epoch": 4.082598408259841, "grad_norm": 2.28838849067688, "learning_rate": 9.862191359036883e-07, "loss": 0.0019, "step": 7120 }, { "epoch": 4.088334408833441, "grad_norm": 0.003568111453205347, "learning_rate": 9.743085210434505e-07, "loss": 0.0136, "step": 7130 }, { "epoch": 4.094070409407041, "grad_norm": 2.805901050567627, "learning_rate": 9.624624976613217e-07, "loss": 0.0083, "step": 7140 }, { "epoch": 4.099806409980641, "grad_norm": 0.009448690339922905, "learning_rate": 9.506812558245665e-07, "loss": 0.0004, "step": 7150 }, { "epoch": 4.099806409980641, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09947969764471054, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2914, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.67, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.358, "step": 7150 }, { "epoch": 4.105542410554241, "grad_norm": 0.007924864068627357, "learning_rate": 9.389649845610376e-07, "loss": 0.0083, "step": 7160 }, { "epoch": 4.1112784111278415, "grad_norm": 0.03764890506863594, "learning_rate": 9.273138718561519e-07, "loss": 0.0003, "step": 7170 }, { "epoch": 4.117014411701441, "grad_norm": 0.015703987330198288, "learning_rate": 9.157281046498628e-07, "loss": 0.0141, "step": 7180 }, { "epoch": 4.122750412275042, "grad_norm": 0.04593523591756821, "learning_rate": 9.042078688336737e-07, "loss": 0.0005, "step": 7190 }, { "epoch": 4.128486412848641, "grad_norm": 0.08672873675823212, "learning_rate": 8.927533492476437e-07, "loss": 0.0088, "step": 7200 }, { "epoch": 4.128486412848641, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09371308237314224, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2865, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.677, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, "step": 7200 }, { "epoch": 4.134222413422242, "grad_norm": 0.0006487391074188054, "learning_rate": 8.813647296774291e-07, "loss": 0.0033, "step": 7210 }, { "epoch": 4.139958413995841, "grad_norm": 2.7435028553009033, "learning_rate": 8.700421928513353e-07, "loss": 0.0104, "step": 7220 }, { "epoch": 4.145694414569442, "grad_norm": 0.12825919687747955, "learning_rate": 8.58785920437376e-07, "loss": 0.0131, "step": 7230 }, { "epoch": 4.151430415143041, "grad_norm": 0.019473344087600708, "learning_rate": 8.475960930403721e-07, "loss": 0.0123, "step": 7240 }, { "epoch": 4.157166415716642, "grad_norm": 0.019103556871414185, "learning_rate": 8.364728901990404e-07, "loss": 0.0161, "step": 7250 }, { "epoch": 4.157166415716642, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08745884150266647, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3788, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.544, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.296, "step": 7250 }, { "epoch": 4.162902416290241, "grad_norm": 0.001977517269551754, "learning_rate": 8.254164903831203e-07, "loss": 0.009, "step": 7260 }, { "epoch": 4.168638416863842, "grad_norm": 0.00919981300830841, "learning_rate": 8.144270709905117e-07, "loss": 0.0185, "step": 7270 }, { "epoch": 4.174374417437441, "grad_norm": 0.007627399172633886, "learning_rate": 8.035048083444241e-07, "loss": 0.0013, "step": 7280 }, { "epoch": 4.180110418011042, "grad_norm": 0.006586376577615738, "learning_rate": 7.926498776905495e-07, "loss": 0.0013, "step": 7290 }, { "epoch": 4.185846418584642, "grad_norm": 0.0025622497778385878, "learning_rate": 7.818624531942493e-07, "loss": 0.0003, "step": 7300 }, { "epoch": 4.185846418584642, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08901920914649963, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2083, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.79, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.419, "step": 7300 }, { "epoch": 4.191582419158242, "grad_norm": 0.07510726898908615, "learning_rate": 7.711427079377614e-07, "loss": 0.0055, "step": 7310 }, { "epoch": 4.197318419731842, "grad_norm": 0.002239730441942811, "learning_rate": 7.604908139174255e-07, "loss": 0.0007, "step": 7320 }, { "epoch": 4.203054420305442, "grad_norm": 0.0018331691389903426, "learning_rate": 7.499069420409183e-07, "loss": 0.0104, "step": 7330 }, { "epoch": 4.208790420879042, "grad_norm": 0.06228185445070267, "learning_rate": 7.393912621245142e-07, "loss": 0.015, "step": 7340 }, { "epoch": 4.214526421452642, "grad_norm": 0.0020015144255012274, "learning_rate": 7.289439428903599e-07, "loss": 0.0015, "step": 7350 }, { "epoch": 4.214526421452642, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08976524323225021, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3011, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.656, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.351, "step": 7350 }, { "epoch": 4.220262422026242, "grad_norm": 3.2421772480010986, "learning_rate": 7.185651519637659e-07, "loss": 0.0083, "step": 7360 }, { "epoch": 4.225998422599842, "grad_norm": 0.0900789201259613, "learning_rate": 7.082550558705225e-07, "loss": 0.0084, "step": 7370 }, { "epoch": 4.231734423173442, "grad_norm": 0.0065153795294463634, "learning_rate": 6.980138200342229e-07, "loss": 0.0003, "step": 7380 }, { "epoch": 4.237470423747042, "grad_norm": 0.041670165956020355, "learning_rate": 6.878416087736079e-07, "loss": 0.0006, "step": 7390 }, { "epoch": 4.243206424320642, "grad_norm": 0.005517491605132818, "learning_rate": 6.777385852999319e-07, "loss": 0.0005, "step": 7400 }, { "epoch": 4.243206424320642, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08851981908082962, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2967, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.662, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.355, "step": 7400 }, { "epoch": 4.248942424894243, "grad_norm": 0.007675142493098974, "learning_rate": 6.67704911714348e-07, "loss": 0.0127, "step": 7410 }, { "epoch": 4.2546784254678425, "grad_norm": 2.6328227519989014, "learning_rate": 6.577407490052978e-07, "loss": 0.0152, "step": 7420 }, { "epoch": 4.260414426041443, "grad_norm": 0.14412537217140198, "learning_rate": 6.478462570459388e-07, "loss": 0.0012, "step": 7430 }, { "epoch": 4.2661504266150425, "grad_norm": 3.322977066040039, "learning_rate": 6.380215945915702e-07, "loss": 0.0065, "step": 7440 }, { "epoch": 4.271886427188643, "grad_norm": 0.0006975011201575398, "learning_rate": 6.282669192770896e-07, "loss": 0.0008, "step": 7450 }, { "epoch": 4.271886427188643, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09144297242164612, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2868, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.676, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, "step": 7450 }, { "epoch": 4.2776224277622426, "grad_norm": 0.04776507988572121, "learning_rate": 6.185823876144675e-07, "loss": 0.0052, "step": 7460 }, { "epoch": 4.283358428335843, "grad_norm": 4.260890483856201, "learning_rate": 6.089681549902287e-07, "loss": 0.0172, "step": 7470 }, { "epoch": 4.289094428909443, "grad_norm": 0.0017789709381759167, "learning_rate": 5.994243756629664e-07, "loss": 0.0045, "step": 7480 }, { "epoch": 4.294830429483043, "grad_norm": 0.020361248403787613, "learning_rate": 5.89951202760859e-07, "loss": 0.0013, "step": 7490 }, { "epoch": 4.300566430056643, "grad_norm": 0.01590687222778797, "learning_rate": 5.805487882792227e-07, "loss": 0.0079, "step": 7500 }, { "epoch": 4.300566430056643, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08803683519363403, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2002, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.802, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.424, "step": 7500 }, { "epoch": 4.306302430630243, "grad_norm": 0.016193749383091927, "learning_rate": 5.712172830780649e-07, "loss": 0.0032, "step": 7510 }, { "epoch": 4.312038431203844, "grad_norm": 0.2794877588748932, "learning_rate": 5.619568368796657e-07, "loss": 0.0046, "step": 7520 }, { "epoch": 4.317774431777443, "grad_norm": 0.3345611095428467, "learning_rate": 5.527675982661801e-07, "loss": 0.0028, "step": 7530 }, { "epoch": 4.323510432351044, "grad_norm": 0.10426267981529236, "learning_rate": 5.43649714677249e-07, "loss": 0.0013, "step": 7540 }, { "epoch": 4.329246432924643, "grad_norm": 0.0005553574301302433, "learning_rate": 5.346033324076338e-07, "loss": 0.006, "step": 7550 }, { "epoch": 4.329246432924643, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.087630994617939, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2554, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.722, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.384, "step": 7550 }, { "epoch": 4.334982433498244, "grad_norm": 0.08077429980039597, "learning_rate": 5.256285966048719e-07, "loss": 0.0067, "step": 7560 }, { "epoch": 4.340718434071843, "grad_norm": 0.004632384981960058, "learning_rate": 5.167256512669444e-07, "loss": 0.0084, "step": 7570 }, { "epoch": 4.346454434645444, "grad_norm": 0.0067880041897296906, "learning_rate": 5.078946392399703e-07, "loss": 0.0021, "step": 7580 }, { "epoch": 4.352190435219043, "grad_norm": 0.001848329440690577, "learning_rate": 4.991357022159111e-07, "loss": 0.0007, "step": 7590 }, { "epoch": 4.357926435792644, "grad_norm": 0.09323499351739883, "learning_rate": 4.90448980730297e-07, "loss": 0.0024, "step": 7600 }, { "epoch": 4.357926435792644, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08750669658184052, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2509, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.728, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.388, "step": 7600 }, { "epoch": 4.363662436366243, "grad_norm": 0.0019721076823771, "learning_rate": 4.818346141599739e-07, "loss": 0.0037, "step": 7610 }, { "epoch": 4.369398436939844, "grad_norm": 0.26007330417633057, "learning_rate": 4.732927407208654e-07, "loss": 0.0011, "step": 7620 }, { "epoch": 4.375134437513443, "grad_norm": 3.1276211738586426, "learning_rate": 4.6482349746575783e-07, "loss": 0.0274, "step": 7630 }, { "epoch": 4.380870438087044, "grad_norm": 0.004567150957882404, "learning_rate": 4.564270202820997e-07, "loss": 0.0004, "step": 7640 }, { "epoch": 4.386606438660644, "grad_norm": 0.033358391374349594, "learning_rate": 4.4810344388982017e-07, "loss": 0.0078, "step": 7650 }, { "epoch": 4.386606438660644, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08857561647891998, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2607, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.714, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.381, "step": 7650 }, { "epoch": 4.392342439234244, "grad_norm": 0.3984158933162689, "learning_rate": 4.3985290183916893e-07, "loss": 0.003, "step": 7660 }, { "epoch": 4.398078439807844, "grad_norm": 0.09813853353261948, "learning_rate": 4.316755265085715e-07, "loss": 0.002, "step": 7670 }, { "epoch": 4.403814440381444, "grad_norm": 0.013608184643089771, "learning_rate": 4.2357144910251003e-07, "loss": 0.0007, "step": 7680 }, { "epoch": 4.409550440955044, "grad_norm": 0.012810665182769299, "learning_rate": 4.155407996494143e-07, "loss": 0.0003, "step": 7690 }, { "epoch": 4.415286441528644, "grad_norm": 0.5154911279678345, "learning_rate": 4.0758370699957416e-07, "loss": 0.0135, "step": 7700 }, { "epoch": 4.415286441528644, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08959878236055374, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2846, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.679, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.363, "step": 7700 }, { "epoch": 4.421022442102244, "grad_norm": 0.0007974316249601543, "learning_rate": 3.997002988230747e-07, "loss": 0.0026, "step": 7710 }, { "epoch": 4.426758442675844, "grad_norm": 0.032501526176929474, "learning_rate": 3.918907016077489e-07, "loss": 0.0051, "step": 7720 }, { "epoch": 4.432494443249444, "grad_norm": 0.2708737552165985, "learning_rate": 3.841550406571443e-07, "loss": 0.0009, "step": 7730 }, { "epoch": 4.438230443823044, "grad_norm": 0.6233472228050232, "learning_rate": 3.764934400885162e-07, "loss": 0.0012, "step": 7740 }, { "epoch": 4.4439664443966445, "grad_norm": 0.00045903949649073184, "learning_rate": 3.689060228308339e-07, "loss": 0.0016, "step": 7750 }, { "epoch": 4.4439664443966445, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08894706517457962, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3241, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.623, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.335, "step": 7750 }, { "epoch": 4.449702444970244, "grad_norm": 0.02216159924864769, "learning_rate": 3.6139291062281036e-07, "loss": 0.0112, "step": 7760 }, { "epoch": 4.4554384455438445, "grad_norm": 0.028067264705896378, "learning_rate": 3.539542240109456e-07, "loss": 0.0044, "step": 7770 }, { "epoch": 4.461174446117445, "grad_norm": 3.8059170246124268, "learning_rate": 3.4659008234759597e-07, "loss": 0.0063, "step": 7780 }, { "epoch": 4.4669104466910445, "grad_norm": 0.0011888346634805202, "learning_rate": 3.3930060378905893e-07, "loss": 0.0003, "step": 7790 }, { "epoch": 4.472646447264645, "grad_norm": 0.0015550279058516026, "learning_rate": 3.320859052936731e-07, "loss": 0.0009, "step": 7800 }, { "epoch": 4.472646447264645, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09025800973176956, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1333, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.899, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.473, "step": 7800 }, { "epoch": 4.478382447838245, "grad_norm": 0.039686888456344604, "learning_rate": 3.249461026199485e-07, "loss": 0.0003, "step": 7810 }, { "epoch": 4.484118448411845, "grad_norm": 3.7747395038604736, "learning_rate": 3.1788131032470306e-07, "loss": 0.0124, "step": 7820 }, { "epoch": 4.489854448985445, "grad_norm": 0.000805154733825475, "learning_rate": 3.108916417612262e-07, "loss": 0.0031, "step": 7830 }, { "epoch": 4.495590449559045, "grad_norm": 0.0011434925254434347, "learning_rate": 3.039772090774634e-07, "loss": 0.0009, "step": 7840 }, { "epoch": 4.501326450132645, "grad_norm": 0.019072147086262703, "learning_rate": 2.97138123214214e-07, "loss": 0.0003, "step": 7850 }, { "epoch": 4.501326450132645, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09056304395198822, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.169, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.847, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.447, "step": 7850 }, { "epoch": 4.507062450706245, "grad_norm": 0.005647636018693447, "learning_rate": 2.903744939033504e-07, "loss": 0.0001, "step": 7860 }, { "epoch": 4.512798451279846, "grad_norm": 2.6264231204986572, "learning_rate": 2.836864296660574e-07, "loss": 0.0132, "step": 7870 }, { "epoch": 4.518534451853445, "grad_norm": 0.010306664742529392, "learning_rate": 2.770740378110942e-07, "loss": 0.0145, "step": 7880 }, { "epoch": 4.524270452427046, "grad_norm": 0.18294845521450043, "learning_rate": 2.7053742443307054e-07, "loss": 0.0041, "step": 7890 }, { "epoch": 4.530006453000645, "grad_norm": 0.017165735363960266, "learning_rate": 2.640766944107431e-07, "loss": 0.0007, "step": 7900 }, { "epoch": 4.530006453000645, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08944171667098999, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2567, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.72, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.383, "step": 7900 }, { "epoch": 4.535742453574246, "grad_norm": 0.04896865785121918, "learning_rate": 2.5769195140533556e-07, "loss": 0.0125, "step": 7910 }, { "epoch": 4.541478454147845, "grad_norm": 0.022854099050164223, "learning_rate": 2.5138329785887263e-07, "loss": 0.0005, "step": 7920 }, { "epoch": 4.547214454721446, "grad_norm": 0.0018585673533380032, "learning_rate": 2.4515083499253743e-07, "loss": 0.0016, "step": 7930 }, { "epoch": 4.552950455295045, "grad_norm": 0.003768256865441799, "learning_rate": 2.3899466280504936e-07, "loss": 0.0217, "step": 7940 }, { "epoch": 4.558686455868646, "grad_norm": 0.05598225072026253, "learning_rate": 2.3291488007105734e-07, "loss": 0.009, "step": 7950 }, { "epoch": 4.558686455868646, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09063292294740677, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1171, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.923, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.485, "step": 7950 }, { "epoch": 4.564422456442245, "grad_norm": 3.3265416622161865, "learning_rate": 2.269115843395553e-07, "loss": 0.0151, "step": 7960 }, { "epoch": 4.570158457015846, "grad_norm": 0.0049064746126532555, "learning_rate": 2.2098487193231543e-07, "loss": 0.0208, "step": 7970 }, { "epoch": 4.575894457589445, "grad_norm": 0.0036308506969362497, "learning_rate": 2.1513483794234847e-07, "loss": 0.0068, "step": 7980 }, { "epoch": 4.581630458163046, "grad_norm": 0.0013542186934500933, "learning_rate": 2.093615762323703e-07, "loss": 0.0022, "step": 7990 }, { "epoch": 4.587366458736646, "grad_norm": 2.7696778774261475, "learning_rate": 2.0366517943330278e-07, "loss": 0.0014, "step": 8000 }, { "epoch": 4.587366458736646, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08961217105388641, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3584, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.574, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.31, "step": 8000 }, { "epoch": 4.593102459310246, "grad_norm": 0.0006430260837078094, "learning_rate": 1.9804573894278311e-07, "loss": 0.0137, "step": 8010 }, { "epoch": 4.598838459883846, "grad_norm": 0.49191415309906006, "learning_rate": 1.925033449236974e-07, "loss": 0.0127, "step": 8020 }, { "epoch": 4.604574460457446, "grad_norm": 0.4231385886669159, "learning_rate": 1.8703808630273768e-07, "loss": 0.0108, "step": 8030 }, { "epoch": 4.610310461031046, "grad_norm": 5.982341289520264, "learning_rate": 1.8165005076897114e-07, "loss": 0.0034, "step": 8040 }, { "epoch": 4.616046461604646, "grad_norm": 0.01885843276977539, "learning_rate": 1.7633932477243599e-07, "loss": 0.0227, "step": 8050 }, { "epoch": 4.616046461604646, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09145788848400116, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2115, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.785, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.416, "step": 8050 }, { "epoch": 4.621782462178246, "grad_norm": 0.004457383882254362, "learning_rate": 1.7110599352275104e-07, "loss": 0.0066, "step": 8060 }, { "epoch": 4.627518462751846, "grad_norm": 0.005887819454073906, "learning_rate": 1.6595014098775342e-07, "loss": 0.004, "step": 8070 }, { "epoch": 4.633254463325446, "grad_norm": 0.023113150149583817, "learning_rate": 1.6087184989214687e-07, "loss": 0.0258, "step": 8080 }, { "epoch": 4.638990463899046, "grad_norm": 0.020028818398714066, "learning_rate": 1.5587120171617509e-07, "loss": 0.0005, "step": 8090 }, { "epoch": 4.6447264644726465, "grad_norm": 0.01912406086921692, "learning_rate": 1.509482766943182e-07, "loss": 0.0006, "step": 8100 }, { "epoch": 4.6447264644726465, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09051110595464706, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.4316, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.469, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.258, "step": 8100 }, { "epoch": 4.650462465046246, "grad_norm": 3.4732401371002197, "learning_rate": 1.4610315381400175e-07, "loss": 0.0149, "step": 8110 }, { "epoch": 4.6561984656198465, "grad_norm": 0.0006817178218625486, "learning_rate": 1.4133591081432806e-07, "loss": 0.0008, "step": 8120 }, { "epoch": 4.661934466193447, "grad_norm": 0.0062561482191085815, "learning_rate": 1.366466241848341e-07, "loss": 0.0004, "step": 8130 }, { "epoch": 4.6676704667670466, "grad_norm": 0.022234557196497917, "learning_rate": 1.3203536916425842e-07, "loss": 0.001, "step": 8140 }, { "epoch": 4.673406467340647, "grad_norm": 4.0328369140625, "learning_rate": 1.275022197393394e-07, "loss": 0.0047, "step": 8150 }, { "epoch": 4.673406467340647, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09013617783784866, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1698, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.846, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.447, "step": 8150 }, { "epoch": 4.679142467914247, "grad_norm": 0.15110693871974945, "learning_rate": 1.230472486436246e-07, "loss": 0.0004, "step": 8160 }, { "epoch": 4.684878468487847, "grad_norm": 0.009227008558809757, "learning_rate": 1.1867052735630435e-07, "loss": 0.0014, "step": 8170 }, { "epoch": 4.690614469061447, "grad_norm": 0.2895458936691284, "learning_rate": 1.1437212610106496e-07, "loss": 0.0013, "step": 8180 }, { "epoch": 4.696350469635047, "grad_norm": 0.541077733039856, "learning_rate": 1.1015211384496238e-07, "loss": 0.0102, "step": 8190 }, { "epoch": 4.702086470208647, "grad_norm": 2.9748380184173584, "learning_rate": 1.0601055829731588e-07, "loss": 0.023, "step": 8200 }, { "epoch": 4.702086470208647, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09088833630084991, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2127, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.783, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.415, "step": 8200 }, { "epoch": 4.707822470782247, "grad_norm": 5.7777838706970215, "learning_rate": 1.0194752590862223e-07, "loss": 0.0119, "step": 8210 }, { "epoch": 4.713558471355848, "grad_norm": 0.06003111973404884, "learning_rate": 9.796308186948711e-08, "loss": 0.0007, "step": 8220 }, { "epoch": 4.719294471929447, "grad_norm": 0.005556735210120678, "learning_rate": 9.405729010958043e-08, "loss": 0.001, "step": 8230 }, { "epoch": 4.725030472503047, "grad_norm": 2.6498606204986572, "learning_rate": 9.023021329661152e-08, "loss": 0.0095, "step": 8240 }, { "epoch": 4.730766473076647, "grad_norm": 0.03512834012508392, "learning_rate": 8.648191283532337e-08, "loss": 0.0081, "step": 8250 }, { "epoch": 4.730766473076647, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0907173827290535, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.316, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.634, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.341, "step": 8250 }, { "epoch": 4.736502473650248, "grad_norm": 0.009284062311053276, "learning_rate": 8.281244886650607e-08, "loss": 0.0051, "step": 8260 }, { "epoch": 4.742238474223847, "grad_norm": 0.02183857187628746, "learning_rate": 7.922188026603273e-08, "loss": 0.0022, "step": 8270 }, { "epoch": 4.747974474797448, "grad_norm": 0.0035564859863370657, "learning_rate": 7.571026464391451e-08, "loss": 0.0025, "step": 8280 }, { "epoch": 4.753710475371047, "grad_norm": 0.04188934341073036, "learning_rate": 7.227765834337874e-08, "loss": 0.0004, "step": 8290 }, { "epoch": 4.759446475944648, "grad_norm": 0.6253750324249268, "learning_rate": 6.892411643995955e-08, "loss": 0.0006, "step": 8300 }, { "epoch": 4.759446475944648, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.090791717171669, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3622, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.568, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.307, "step": 8300 }, { "epoch": 4.765182476518247, "grad_norm": 0.3705517053604126, "learning_rate": 6.564969274061972e-08, "loss": 0.0042, "step": 8310 }, { "epoch": 4.770918477091848, "grad_norm": 0.0012910285731777549, "learning_rate": 6.245443978288413e-08, "loss": 0.0023, "step": 8320 }, { "epoch": 4.776654477665447, "grad_norm": 0.5175131559371948, "learning_rate": 5.933840883399766e-08, "loss": 0.0011, "step": 8330 }, { "epoch": 4.782390478239048, "grad_norm": 3.527125358581543, "learning_rate": 5.630164989010312e-08, "loss": 0.0023, "step": 8340 }, { "epoch": 4.788126478812648, "grad_norm": 0.3973446190357208, "learning_rate": 5.334421167543735e-08, "loss": 0.0026, "step": 8350 }, { "epoch": 4.788126478812648, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09068261086940765, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1697, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.846, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.447, "step": 8350 }, { "epoch": 4.793862479386248, "grad_norm": 0.003769023111090064, "learning_rate": 5.0466141641553056e-08, "loss": 0.0001, "step": 8360 }, { "epoch": 4.799598479959848, "grad_norm": 0.08687523007392883, "learning_rate": 4.766748596655268e-08, "loss": 0.0061, "step": 8370 }, { "epoch": 4.805334480533448, "grad_norm": 5.824708461761475, "learning_rate": 4.494828955435126e-08, "loss": 0.0047, "step": 8380 }, { "epoch": 4.811070481107048, "grad_norm": 2.9158382415771484, "learning_rate": 4.230859603395421e-08, "loss": 0.011, "step": 8390 }, { "epoch": 4.816806481680648, "grad_norm": 0.04298550263047218, "learning_rate": 3.974844775875786e-08, "loss": 0.0016, "step": 8400 }, { "epoch": 4.816806481680648, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09151580929756165, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2757, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.692, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.37, "step": 8400 }, { "epoch": 4.822542482254248, "grad_norm": 0.005364727228879929, "learning_rate": 3.726788580586893e-08, "loss": 0.0099, "step": 8410 }, { "epoch": 4.828278482827848, "grad_norm": 0.0018757123034447432, "learning_rate": 3.4866949975448374e-08, "loss": 0.0009, "step": 8420 }, { "epoch": 4.8340144834014485, "grad_norm": 0.008776834234595299, "learning_rate": 3.254567879006798e-08, "loss": 0.0068, "step": 8430 }, { "epoch": 4.839750483975048, "grad_norm": 0.009764501824975014, "learning_rate": 3.030410949409701e-08, "loss": 0.0088, "step": 8440 }, { "epoch": 4.8454864845486485, "grad_norm": 4.281154632568359, "learning_rate": 2.8142278053101545e-08, "loss": 0.0135, "step": 8450 }, { "epoch": 4.8454864845486485, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0919208973646164, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1321, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.901, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.474, "step": 8450 }, { "epoch": 4.851222485122248, "grad_norm": 0.09912008792161942, "learning_rate": 2.6060219153268286e-08, "loss": 0.0005, "step": 8460 }, { "epoch": 4.8569584856958485, "grad_norm": 0.019211160019040108, "learning_rate": 2.4057966200849437e-08, "loss": 0.0005, "step": 8470 }, { "epoch": 4.862694486269449, "grad_norm": 0.05037353187799454, "learning_rate": 2.213555132162426e-08, "loss": 0.0012, "step": 8480 }, { "epoch": 4.868430486843049, "grad_norm": 0.00411713682115078, "learning_rate": 2.0293005360386142e-08, "loss": 0.0003, "step": 8490 }, { "epoch": 4.874166487416649, "grad_norm": 0.0006664736429229379, "learning_rate": 1.8530357880444105e-08, "loss": 0.0202, "step": 8500 }, { "epoch": 4.874166487416649, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09091775864362717, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2769, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.69, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.369, "step": 8500 }, { "epoch": 4.879902487990249, "grad_norm": 1.6142276525497437, "learning_rate": 1.684763716315374e-08, "loss": 0.0011, "step": 8510 }, { "epoch": 4.885638488563849, "grad_norm": 0.000510135549120605, "learning_rate": 1.524487020745813e-08, "loss": 0.0011, "step": 8520 }, { "epoch": 4.891374489137449, "grad_norm": 0.001830592635087669, "learning_rate": 1.372208272945763e-08, "loss": 0.0187, "step": 8530 }, { "epoch": 4.897110489711049, "grad_norm": 0.2799864709377289, "learning_rate": 1.2279299161997438e-08, "loss": 0.0157, "step": 8540 }, { "epoch": 4.902846490284649, "grad_norm": 0.009001590311527252, "learning_rate": 1.0916542654273443e-08, "loss": 0.0272, "step": 8550 }, { "epoch": 4.902846490284649, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09087590873241425, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3382, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.602, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.325, "step": 8550 }, { "epoch": 4.908582490858249, "grad_norm": 0.17294363677501678, "learning_rate": 9.633835071463094e-09, "loss": 0.0052, "step": 8560 }, { "epoch": 4.91431849143185, "grad_norm": 0.008691389113664627, "learning_rate": 8.431196994373447e-09, "loss": 0.0065, "step": 8570 }, { "epoch": 4.920054492005449, "grad_norm": 0.010057262144982815, "learning_rate": 7.3086477191103285e-09, "loss": 0.02, "step": 8580 }, { "epoch": 4.925790492579049, "grad_norm": 0.19418714940547943, "learning_rate": 6.2662052567702414e-09, "loss": 0.0003, "step": 8590 }, { "epoch": 4.931526493152649, "grad_norm": 0.033849816769361496, "learning_rate": 5.303886333151154e-09, "loss": 0.0126, "step": 8600 }, { "epoch": 4.931526493152649, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09129200130701065, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2495, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.73, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.389, "step": 8600 }, { "epoch": 4.93726249372625, "grad_norm": 0.024734610691666603, "learning_rate": 4.421706388481606e-09, "loss": 0.0171, "step": 8610 }, { "epoch": 4.942998494299849, "grad_norm": 0.0016847607912495732, "learning_rate": 3.6196795771770156e-09, "loss": 0.0126, "step": 8620 }, { "epoch": 4.94873449487345, "grad_norm": 0.004256678279489279, "learning_rate": 2.897818767609861e-09, "loss": 0.0161, "step": 8630 }, { "epoch": 4.954470495447049, "grad_norm": 4.404531002044678, "learning_rate": 2.2561355419037368e-09, "loss": 0.0087, "step": 8640 }, { "epoch": 4.96020649602065, "grad_norm": 3.0091116428375244, "learning_rate": 1.6946401957479431e-09, "loss": 0.0145, "step": 8650 }, { "epoch": 4.96020649602065, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09181646257638931, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2421, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.741, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.394, "step": 8650 }, { "epoch": 4.965942496594249, "grad_norm": 0.04898509010672569, "learning_rate": 1.2133417382320656e-09, "loss": 0.0032, "step": 8660 }, { "epoch": 4.97167849716785, "grad_norm": 0.002836048137396574, "learning_rate": 8.122478916999799e-10, "loss": 0.0074, "step": 8670 }, { "epoch": 4.977414497741449, "grad_norm": 0.02022160217165947, "learning_rate": 4.913650916299473e-10, "loss": 0.0098, "step": 8680 }, { "epoch": 4.98315049831505, "grad_norm": 0.007077803369611502, "learning_rate": 2.506984865263684e-10, "loss": 0.0016, "step": 8690 }, { "epoch": 4.98888649888865, "grad_norm": 0.1649414747953415, "learning_rate": 9.025193784151232e-11, "loss": 0.0041, "step": 8700 }, { "epoch": 4.98888649888865, "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09170719236135483, "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1606, "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.859, "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.453, "step": 8700 }, { "epoch": 4.99462249946225, "grad_norm": 0.00366253312677145, "learning_rate": 1.0028019910013342e-11, "loss": 0.0032, "step": 8710 }, { "epoch": 4.99749049974905, "step": 8715, "total_flos": 6.758260289254195e+17, "train_loss": 0.02162364056710329, "train_runtime": 21813.1277, "train_samples_per_second": 6.394, "train_steps_per_second": 0.4 } ], "logging_steps": 10, "max_steps": 8715, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 250, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.758260289254195e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }