diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7531 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.99749049974905, + "eval_steps": 50, + "global_step": 8715, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.005736000573600057, + "grad_norm": 2.3000271320343018, + "learning_rate": 1.1467889908256882e-07, + "loss": 0.3103, + "step": 10 + }, + { + "epoch": 0.011472001147200114, + "grad_norm": 0.23891927301883698, + "learning_rate": 2.2935779816513764e-07, + "loss": 0.2101, + "step": 20 + }, + { + "epoch": 0.017208001720800174, + "grad_norm": 3.7850193977355957, + "learning_rate": 3.4403669724770646e-07, + "loss": 0.3158, + "step": 30 + }, + { + "epoch": 0.022944002294400228, + "grad_norm": 10.902544021606445, + "learning_rate": 4.587155963302753e-07, + "loss": 0.3107, + "step": 40 + }, + { + "epoch": 0.028680002868000286, + "grad_norm": 12.402349472045898, + "learning_rate": 5.733944954128441e-07, + "loss": 0.3895, + "step": 50 + }, + { + "epoch": 0.028680002868000286, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 1.0573054552078247, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 17.8757, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 36.53, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 18.293, + "step": 50 + }, + { + "epoch": 0.03441600344160035, + "grad_norm": 1.5937564373016357, + "learning_rate": 6.880733944954129e-07, + "loss": 0.2534, + "step": 60 + }, + { + "epoch": 0.0401520040152004, + "grad_norm": 3.687685012817383, + "learning_rate": 8.027522935779817e-07, + "loss": 0.3795, + "step": 70 + }, + { + "epoch": 0.045888004588800456, + "grad_norm": 0.03578287363052368, + "learning_rate": 9.174311926605506e-07, + "loss": 0.4276, + "step": 80 + }, + { + "epoch": 0.05162400516240052, + "grad_norm": 0.022680282592773438, + "learning_rate": 1.0321100917431195e-06, + "loss": 0.277, + "step": 90 + }, + { + "epoch": 0.05736000573600057, + "grad_norm": 1.7502835988998413, + "learning_rate": 1.1467889908256882e-06, + "loss": 0.2279, + "step": 100 + }, + { + "epoch": 0.05736000573600057, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.9872124791145325, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.223, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.768, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.408, + "step": 100 + }, + { + "epoch": 0.06309600630960063, + "grad_norm": 4.763681411743164, + "learning_rate": 1.261467889908257e-06, + "loss": 0.3441, + "step": 110 + }, + { + "epoch": 0.0688320068832007, + "grad_norm": 7.319093704223633, + "learning_rate": 1.3761467889908258e-06, + "loss": 0.4273, + "step": 120 + }, + { + "epoch": 0.07456800745680074, + "grad_norm": 9.32235336303711, + "learning_rate": 1.4908256880733945e-06, + "loss": 0.3338, + "step": 130 + }, + { + "epoch": 0.0803040080304008, + "grad_norm": 2.0044586658477783, + "learning_rate": 1.6055045871559635e-06, + "loss": 0.1364, + "step": 140 + }, + { + "epoch": 0.08604000860400086, + "grad_norm": 10.962570190429688, + "learning_rate": 1.7201834862385322e-06, + "loss": 0.2404, + "step": 150 + }, + { + "epoch": 0.08604000860400086, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.7815765738487244, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2921, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.669, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.358, + "step": 150 + }, + { + "epoch": 0.09177600917760091, + "grad_norm": 4.845127105712891, + "learning_rate": 1.8348623853211011e-06, + "loss": 0.3601, + "step": 160 + }, + { + "epoch": 0.09751200975120097, + "grad_norm": 3.165642738342285, + "learning_rate": 1.94954128440367e-06, + "loss": 0.1938, + "step": 170 + }, + { + "epoch": 0.10324801032480103, + "grad_norm": 3.4738926887512207, + "learning_rate": 2.064220183486239e-06, + "loss": 0.1979, + "step": 180 + }, + { + "epoch": 0.1089840108984011, + "grad_norm": 0.34281352162361145, + "learning_rate": 2.1788990825688075e-06, + "loss": 0.2278, + "step": 190 + }, + { + "epoch": 0.11472001147200114, + "grad_norm": 9.631810188293457, + "learning_rate": 2.2935779816513764e-06, + "loss": 0.1642, + "step": 200 + }, + { + "epoch": 0.11472001147200114, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.32764628529548645, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1681, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.848, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.448, + "step": 200 + }, + { + "epoch": 0.1204560120456012, + "grad_norm": 4.584328651428223, + "learning_rate": 2.4082568807339453e-06, + "loss": 0.1721, + "step": 210 + }, + { + "epoch": 0.12619201261920127, + "grad_norm": 7.435701370239258, + "learning_rate": 2.522935779816514e-06, + "loss": 0.1632, + "step": 220 + }, + { + "epoch": 0.1319280131928013, + "grad_norm": 2.957846164703369, + "learning_rate": 2.6376146788990823e-06, + "loss": 0.1175, + "step": 230 + }, + { + "epoch": 0.1376640137664014, + "grad_norm": 4.342548847198486, + "learning_rate": 2.7522935779816517e-06, + "loss": 0.0877, + "step": 240 + }, + { + "epoch": 0.14340001434000144, + "grad_norm": 2.2145071029663086, + "learning_rate": 2.8669724770642206e-06, + "loss": 0.0692, + "step": 250 + }, + { + "epoch": 0.14340001434000144, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.14406049251556396, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2632, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.71, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.379, + "step": 250 + }, + { + "epoch": 0.14913601491360148, + "grad_norm": 0.3610692620277405, + "learning_rate": 2.981651376146789e-06, + "loss": 0.1079, + "step": 260 + }, + { + "epoch": 0.15487201548720156, + "grad_norm": 6.997246265411377, + "learning_rate": 3.0963302752293576e-06, + "loss": 0.0748, + "step": 270 + }, + { + "epoch": 0.1606080160608016, + "grad_norm": 2.0748515129089355, + "learning_rate": 3.211009174311927e-06, + "loss": 0.0749, + "step": 280 + }, + { + "epoch": 0.16634401663440165, + "grad_norm": 3.6645171642303467, + "learning_rate": 3.325688073394496e-06, + "loss": 0.097, + "step": 290 + }, + { + "epoch": 0.17208001720800173, + "grad_norm": 8.085243225097656, + "learning_rate": 3.4403669724770644e-06, + "loss": 0.0722, + "step": 300 + }, + { + "epoch": 0.17208001720800173, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11341650038957596, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2959, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.663, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.355, + "step": 300 + }, + { + "epoch": 0.17781601778160178, + "grad_norm": 2.7256767749786377, + "learning_rate": 3.5550458715596333e-06, + "loss": 0.0707, + "step": 310 + }, + { + "epoch": 0.18355201835520182, + "grad_norm": 2.6183135509490967, + "learning_rate": 3.6697247706422022e-06, + "loss": 0.0418, + "step": 320 + }, + { + "epoch": 0.1892880189288019, + "grad_norm": 1.0414828062057495, + "learning_rate": 3.784403669724771e-06, + "loss": 0.041, + "step": 330 + }, + { + "epoch": 0.19502401950240195, + "grad_norm": 1.9375965595245361, + "learning_rate": 3.89908256880734e-06, + "loss": 0.0742, + "step": 340 + }, + { + "epoch": 0.200760020076002, + "grad_norm": 1.6328781843185425, + "learning_rate": 4.013761467889909e-06, + "loss": 0.0426, + "step": 350 + }, + { + "epoch": 0.200760020076002, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11491454392671585, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2047, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.795, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.421, + "step": 350 + }, + { + "epoch": 0.20649602064960207, + "grad_norm": 0.45767122507095337, + "learning_rate": 4.128440366972478e-06, + "loss": 0.0771, + "step": 360 + }, + { + "epoch": 0.21223202122320212, + "grad_norm": 2.768129825592041, + "learning_rate": 4.2431192660550464e-06, + "loss": 0.0869, + "step": 370 + }, + { + "epoch": 0.2179680217968022, + "grad_norm": 1.4702483415603638, + "learning_rate": 4.357798165137615e-06, + "loss": 0.0536, + "step": 380 + }, + { + "epoch": 0.22370402237040224, + "grad_norm": 1.277462124824524, + "learning_rate": 4.4724770642201834e-06, + "loss": 0.0329, + "step": 390 + }, + { + "epoch": 0.2294400229440023, + "grad_norm": 1.8748779296875, + "learning_rate": 4.587155963302753e-06, + "loss": 0.045, + "step": 400 + }, + { + "epoch": 0.2294400229440023, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.12494589388370514, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2877, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.675, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.361, + "step": 400 + }, + { + "epoch": 0.23517602351760236, + "grad_norm": 2.339601516723633, + "learning_rate": 4.701834862385321e-06, + "loss": 0.0546, + "step": 410 + }, + { + "epoch": 0.2409120240912024, + "grad_norm": 0.9202137589454651, + "learning_rate": 4.816513761467891e-06, + "loss": 0.1038, + "step": 420 + }, + { + "epoch": 0.24664802466480246, + "grad_norm": 2.0663821697235107, + "learning_rate": 4.931192660550459e-06, + "loss": 0.0418, + "step": 430 + }, + { + "epoch": 0.25238402523840253, + "grad_norm": 2.6615207195281982, + "learning_rate": 5.045871559633028e-06, + "loss": 0.0414, + "step": 440 + }, + { + "epoch": 0.2581200258120026, + "grad_norm": 2.298173427581787, + "learning_rate": 5.160550458715596e-06, + "loss": 0.063, + "step": 450 + }, + { + "epoch": 0.2581200258120026, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.1194896399974823, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3177, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.632, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.339, + "step": 450 + }, + { + "epoch": 0.2638560263856026, + "grad_norm": 0.9054094552993774, + "learning_rate": 5.275229357798165e-06, + "loss": 0.0572, + "step": 460 + }, + { + "epoch": 0.2695920269592027, + "grad_norm": 0.5312202572822571, + "learning_rate": 5.389908256880735e-06, + "loss": 0.0697, + "step": 470 + }, + { + "epoch": 0.2753280275328028, + "grad_norm": 1.5197083950042725, + "learning_rate": 5.504587155963303e-06, + "loss": 0.047, + "step": 480 + }, + { + "epoch": 0.2810640281064028, + "grad_norm": 1.341577410697937, + "learning_rate": 5.619266055045872e-06, + "loss": 0.0285, + "step": 490 + }, + { + "epoch": 0.28680002868000287, + "grad_norm": 1.6182245016098022, + "learning_rate": 5.733944954128441e-06, + "loss": 0.0402, + "step": 500 + }, + { + "epoch": 0.28680002868000287, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.12436391413211823, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3365, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.605, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.326, + "step": 500 + }, + { + "epoch": 0.29253602925360295, + "grad_norm": 2.623664140701294, + "learning_rate": 5.84862385321101e-06, + "loss": 0.0552, + "step": 510 + }, + { + "epoch": 0.29827202982720297, + "grad_norm": 1.9711512327194214, + "learning_rate": 5.963302752293578e-06, + "loss": 0.0534, + "step": 520 + }, + { + "epoch": 0.30400803040080304, + "grad_norm": 1.4279474020004272, + "learning_rate": 6.077981651376147e-06, + "loss": 0.0356, + "step": 530 + }, + { + "epoch": 0.3097440309744031, + "grad_norm": 1.5085383653640747, + "learning_rate": 6.192660550458715e-06, + "loss": 0.0389, + "step": 540 + }, + { + "epoch": 0.31548003154800314, + "grad_norm": 1.8550148010253906, + "learning_rate": 6.307339449541285e-06, + "loss": 0.0342, + "step": 550 + }, + { + "epoch": 0.31548003154800314, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.16011035442352295, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2066, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.792, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.42, + "step": 550 + }, + { + "epoch": 0.3212160321216032, + "grad_norm": 0.8617078065872192, + "learning_rate": 6.422018348623854e-06, + "loss": 0.0351, + "step": 560 + }, + { + "epoch": 0.3269520326952033, + "grad_norm": 1.7970987558364868, + "learning_rate": 6.536697247706422e-06, + "loss": 0.0328, + "step": 570 + }, + { + "epoch": 0.3326880332688033, + "grad_norm": 1.319393515586853, + "learning_rate": 6.651376146788992e-06, + "loss": 0.0348, + "step": 580 + }, + { + "epoch": 0.3384240338424034, + "grad_norm": 3.622856855392456, + "learning_rate": 6.76605504587156e-06, + "loss": 0.035, + "step": 590 + }, + { + "epoch": 0.34416003441600346, + "grad_norm": 2.2089531421661377, + "learning_rate": 6.880733944954129e-06, + "loss": 0.0485, + "step": 600 + }, + { + "epoch": 0.34416003441600346, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.12377385795116425, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2989, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.659, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.353, + "step": 600 + }, + { + "epoch": 0.3498960349896035, + "grad_norm": 1.3465012311935425, + "learning_rate": 6.995412844036697e-06, + "loss": 0.0448, + "step": 610 + }, + { + "epoch": 0.35563203556320355, + "grad_norm": 1.0101594924926758, + "learning_rate": 7.110091743119267e-06, + "loss": 0.0298, + "step": 620 + }, + { + "epoch": 0.3613680361368036, + "grad_norm": 3.003561019897461, + "learning_rate": 7.224770642201836e-06, + "loss": 0.052, + "step": 630 + }, + { + "epoch": 0.36710403671040365, + "grad_norm": 2.2370080947875977, + "learning_rate": 7.3394495412844045e-06, + "loss": 0.0241, + "step": 640 + }, + { + "epoch": 0.3728400372840037, + "grad_norm": 3.373074531555176, + "learning_rate": 7.454128440366973e-06, + "loss": 0.0446, + "step": 650 + }, + { + "epoch": 0.3728400372840037, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11963505297899246, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1731, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.841, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.444, + "step": 650 + }, + { + "epoch": 0.3785760378576038, + "grad_norm": 0.6863628029823303, + "learning_rate": 7.568807339449542e-06, + "loss": 0.0116, + "step": 660 + }, + { + "epoch": 0.3843120384312038, + "grad_norm": 2.9094369411468506, + "learning_rate": 7.68348623853211e-06, + "loss": 0.0657, + "step": 670 + }, + { + "epoch": 0.3900480390048039, + "grad_norm": 0.2047434151172638, + "learning_rate": 7.79816513761468e-06, + "loss": 0.0319, + "step": 680 + }, + { + "epoch": 0.39578403957840397, + "grad_norm": 1.523390531539917, + "learning_rate": 7.912844036697249e-06, + "loss": 0.0492, + "step": 690 + }, + { + "epoch": 0.401520040152004, + "grad_norm": 0.10014928132295609, + "learning_rate": 8.027522935779817e-06, + "loss": 0.0333, + "step": 700 + }, + { + "epoch": 0.401520040152004, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10763221979141235, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.342, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.597, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.322, + "step": 700 + }, + { + "epoch": 0.40725604072560406, + "grad_norm": 0.24312058091163635, + "learning_rate": 8.142201834862386e-06, + "loss": 0.0248, + "step": 710 + }, + { + "epoch": 0.41299204129920414, + "grad_norm": 0.08946532011032104, + "learning_rate": 8.256880733944956e-06, + "loss": 0.0133, + "step": 720 + }, + { + "epoch": 0.4187280418728042, + "grad_norm": 0.13466329872608185, + "learning_rate": 8.371559633027524e-06, + "loss": 0.0226, + "step": 730 + }, + { + "epoch": 0.42446404244640423, + "grad_norm": 0.4071923792362213, + "learning_rate": 8.486238532110093e-06, + "loss": 0.0353, + "step": 740 + }, + { + "epoch": 0.4302000430200043, + "grad_norm": 1.2394853830337524, + "learning_rate": 8.600917431192661e-06, + "loss": 0.0199, + "step": 750 + }, + { + "epoch": 0.4302000430200043, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09811359643936157, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3072, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.647, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.347, + "step": 750 + }, + { + "epoch": 0.4359360435936044, + "grad_norm": 2.9703874588012695, + "learning_rate": 8.71559633027523e-06, + "loss": 0.0259, + "step": 760 + }, + { + "epoch": 0.4416720441672044, + "grad_norm": 1.3356784582138062, + "learning_rate": 8.830275229357798e-06, + "loss": 0.0496, + "step": 770 + }, + { + "epoch": 0.4474080447408045, + "grad_norm": 0.7478968501091003, + "learning_rate": 8.944954128440367e-06, + "loss": 0.0383, + "step": 780 + }, + { + "epoch": 0.45314404531440455, + "grad_norm": 0.418811172246933, + "learning_rate": 9.059633027522935e-06, + "loss": 0.0173, + "step": 790 + }, + { + "epoch": 0.4588800458880046, + "grad_norm": 0.6638255715370178, + "learning_rate": 9.174311926605506e-06, + "loss": 0.0271, + "step": 800 + }, + { + "epoch": 0.4588800458880046, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11180277168750763, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2311, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.757, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.402, + "step": 800 + }, + { + "epoch": 0.46461604646160465, + "grad_norm": 0.21615058183670044, + "learning_rate": 9.288990825688074e-06, + "loss": 0.0175, + "step": 810 + }, + { + "epoch": 0.4703520470352047, + "grad_norm": 1.2093310356140137, + "learning_rate": 9.403669724770643e-06, + "loss": 0.0215, + "step": 820 + }, + { + "epoch": 0.47608804760880474, + "grad_norm": 0.2247300148010254, + "learning_rate": 9.518348623853211e-06, + "loss": 0.0486, + "step": 830 + }, + { + "epoch": 0.4818240481824048, + "grad_norm": 0.6988762617111206, + "learning_rate": 9.633027522935781e-06, + "loss": 0.0161, + "step": 840 + }, + { + "epoch": 0.4875600487560049, + "grad_norm": 1.2162249088287354, + "learning_rate": 9.74770642201835e-06, + "loss": 0.0277, + "step": 850 + }, + { + "epoch": 0.4875600487560049, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08368796855211258, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2199, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.773, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.41, + "step": 850 + }, + { + "epoch": 0.4932960493296049, + "grad_norm": 1.0620274543762207, + "learning_rate": 9.862385321100918e-06, + "loss": 0.0566, + "step": 860 + }, + { + "epoch": 0.499032049903205, + "grad_norm": 0.23082388937473297, + "learning_rate": 9.977064220183487e-06, + "loss": 0.0247, + "step": 870 + }, + { + "epoch": 0.5047680504768051, + "grad_norm": 0.9556642770767212, + "learning_rate": 9.999974328282418e-06, + "loss": 0.0242, + "step": 880 + }, + { + "epoch": 0.5105040510504051, + "grad_norm": 0.6304249167442322, + "learning_rate": 9.999870037381541e-06, + "loss": 0.0268, + "step": 890 + }, + { + "epoch": 0.5162400516240052, + "grad_norm": 0.07784967124462128, + "learning_rate": 9.99968552448706e-06, + "loss": 0.033, + "step": 900 + }, + { + "epoch": 0.5162400516240052, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06569766253232956, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2473, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.733, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.39, + "step": 900 + }, + { + "epoch": 0.5219760521976052, + "grad_norm": 1.320080280303955, + "learning_rate": 9.999420792559453e-06, + "loss": 0.0623, + "step": 910 + }, + { + "epoch": 0.5277120527712053, + "grad_norm": 0.24600175023078918, + "learning_rate": 9.999075845846292e-06, + "loss": 0.022, + "step": 920 + }, + { + "epoch": 0.5334480533448054, + "grad_norm": 1.860329031944275, + "learning_rate": 9.998650689882184e-06, + "loss": 0.0186, + "step": 930 + }, + { + "epoch": 0.5391840539184054, + "grad_norm": 1.9470009803771973, + "learning_rate": 9.99814533148868e-06, + "loss": 0.023, + "step": 940 + }, + { + "epoch": 0.5449200544920054, + "grad_norm": 0.3151063919067383, + "learning_rate": 9.997559778774162e-06, + "loss": 0.0111, + "step": 950 + }, + { + "epoch": 0.5449200544920054, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07853744179010391, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2659, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.706, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.377, + "step": 950 + }, + { + "epoch": 0.5506560550656056, + "grad_norm": 1.4308031797409058, + "learning_rate": 9.996894041133715e-06, + "loss": 0.0239, + "step": 960 + }, + { + "epoch": 0.5563920556392056, + "grad_norm": 0.32267820835113525, + "learning_rate": 9.996148129248975e-06, + "loss": 0.0365, + "step": 970 + }, + { + "epoch": 0.5621280562128056, + "grad_norm": 0.026730485260486603, + "learning_rate": 9.995322055087963e-06, + "loss": 0.0255, + "step": 980 + }, + { + "epoch": 0.5678640567864057, + "grad_norm": 4.146325588226318, + "learning_rate": 9.994415831904886e-06, + "loss": 0.06, + "step": 990 + }, + { + "epoch": 0.5736000573600057, + "grad_norm": 0.9087633490562439, + "learning_rate": 9.993429474239928e-06, + "loss": 0.0371, + "step": 1000 + }, + { + "epoch": 0.5736000573600057, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08875266462564468, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2636, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.71, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.378, + "step": 1000 + }, + { + "epoch": 0.5793360579336058, + "grad_norm": 1.1552188396453857, + "learning_rate": 9.992362997919016e-06, + "loss": 0.0419, + "step": 1010 + }, + { + "epoch": 0.5850720585072059, + "grad_norm": 0.8535066246986389, + "learning_rate": 9.991216420053565e-06, + "loss": 0.0162, + "step": 1020 + }, + { + "epoch": 0.5908080590808059, + "grad_norm": 1.3267903327941895, + "learning_rate": 9.989989759040208e-06, + "loss": 0.0375, + "step": 1030 + }, + { + "epoch": 0.5965440596544059, + "grad_norm": 2.629297971725464, + "learning_rate": 9.988683034560494e-06, + "loss": 0.0097, + "step": 1040 + }, + { + "epoch": 0.6022800602280061, + "grad_norm": 2.267737627029419, + "learning_rate": 9.987296267580575e-06, + "loss": 0.0267, + "step": 1050 + }, + { + "epoch": 0.6022800602280061, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10149678587913513, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2648, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.708, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.377, + "step": 1050 + }, + { + "epoch": 0.6080160608016061, + "grad_norm": 0.04809356853365898, + "learning_rate": 9.985829480350874e-06, + "loss": 0.0104, + "step": 1060 + }, + { + "epoch": 0.6137520613752061, + "grad_norm": 2.897101640701294, + "learning_rate": 9.98428269640572e-06, + "loss": 0.0313, + "step": 1070 + }, + { + "epoch": 0.6194880619488062, + "grad_norm": 0.2844018340110779, + "learning_rate": 9.982655940562978e-06, + "loss": 0.0202, + "step": 1080 + }, + { + "epoch": 0.6252240625224063, + "grad_norm": 1.7705107927322388, + "learning_rate": 9.980949238923646e-06, + "loss": 0.0151, + "step": 1090 + }, + { + "epoch": 0.6309600630960063, + "grad_norm": 2.162916421890259, + "learning_rate": 9.979162618871435e-06, + "loss": 0.021, + "step": 1100 + }, + { + "epoch": 0.6309600630960063, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.087111696600914, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3134, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.638, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.342, + "step": 1100 + }, + { + "epoch": 0.6366960636696064, + "grad_norm": 3.3192200660705566, + "learning_rate": 9.977296109072337e-06, + "loss": 0.0182, + "step": 1110 + }, + { + "epoch": 0.6424320642432064, + "grad_norm": 0.3482876718044281, + "learning_rate": 9.975349739474156e-06, + "loss": 0.0309, + "step": 1120 + }, + { + "epoch": 0.6481680648168064, + "grad_norm": 2.682620048522949, + "learning_rate": 9.973323541306032e-06, + "loss": 0.0299, + "step": 1130 + }, + { + "epoch": 0.6539040653904066, + "grad_norm": 2.5474138259887695, + "learning_rate": 9.971217547077947e-06, + "loss": 0.0337, + "step": 1140 + }, + { + "epoch": 0.6596400659640066, + "grad_norm": 0.5940343141555786, + "learning_rate": 9.969031790580185e-06, + "loss": 0.0294, + "step": 1150 + }, + { + "epoch": 0.6596400659640066, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07855822145938873, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1855, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.823, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.435, + "step": 1150 + }, + { + "epoch": 0.6653760665376066, + "grad_norm": 1.4965182542800903, + "learning_rate": 9.966766306882811e-06, + "loss": 0.0257, + "step": 1160 + }, + { + "epoch": 0.6711120671112067, + "grad_norm": 0.10210319608449936, + "learning_rate": 9.964421132335091e-06, + "loss": 0.0214, + "step": 1170 + }, + { + "epoch": 0.6768480676848068, + "grad_norm": 1.594895362854004, + "learning_rate": 9.961996304564916e-06, + "loss": 0.0182, + "step": 1180 + }, + { + "epoch": 0.6825840682584068, + "grad_norm": 1.8244154453277588, + "learning_rate": 9.959491862478206e-06, + "loss": 0.0195, + "step": 1190 + }, + { + "epoch": 0.6883200688320069, + "grad_norm": 3.9240288734436035, + "learning_rate": 9.956907846258268e-06, + "loss": 0.0426, + "step": 1200 + }, + { + "epoch": 0.6883200688320069, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08573708683252335, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1456, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.881, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.464, + "step": 1200 + }, + { + "epoch": 0.6940560694056069, + "grad_norm": 0.06307424604892731, + "learning_rate": 9.954244297365169e-06, + "loss": 0.0592, + "step": 1210 + }, + { + "epoch": 0.699792069979207, + "grad_norm": 0.44088277220726013, + "learning_rate": 9.951501258535061e-06, + "loss": 0.0267, + "step": 1220 + }, + { + "epoch": 0.7055280705528071, + "grad_norm": 2.217980146408081, + "learning_rate": 9.948678773779495e-06, + "loss": 0.0171, + "step": 1230 + }, + { + "epoch": 0.7112640711264071, + "grad_norm": 1.3680331707000732, + "learning_rate": 9.94577688838472e-06, + "loss": 0.0369, + "step": 1240 + }, + { + "epoch": 0.7170000717000071, + "grad_norm": 0.043984800577163696, + "learning_rate": 9.94279564891095e-06, + "loss": 0.0173, + "step": 1250 + }, + { + "epoch": 0.7170000717000071, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06791304796934128, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1924, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.813, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.43, + "step": 1250 + }, + { + "epoch": 0.7227360722736073, + "grad_norm": 0.05398892983794212, + "learning_rate": 9.939735103191625e-06, + "loss": 0.0151, + "step": 1260 + }, + { + "epoch": 0.7284720728472073, + "grad_norm": 0.26481369137763977, + "learning_rate": 9.936595300332638e-06, + "loss": 0.0242, + "step": 1270 + }, + { + "epoch": 0.7342080734208073, + "grad_norm": 0.0726240873336792, + "learning_rate": 9.933376290711546e-06, + "loss": 0.0128, + "step": 1280 + }, + { + "epoch": 0.7399440739944074, + "grad_norm": 0.21732382476329803, + "learning_rate": 9.930078125976767e-06, + "loss": 0.0268, + "step": 1290 + }, + { + "epoch": 0.7456800745680074, + "grad_norm": 1.968591332435608, + "learning_rate": 9.926700859046745e-06, + "loss": 0.0264, + "step": 1300 + }, + { + "epoch": 0.7456800745680074, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08293525129556656, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2805, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.685, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.366, + "step": 1300 + }, + { + "epoch": 0.7514160751416075, + "grad_norm": 0.12683559954166412, + "learning_rate": 9.923244544109107e-06, + "loss": 0.0107, + "step": 1310 + }, + { + "epoch": 0.7571520757152076, + "grad_norm": 0.07375319302082062, + "learning_rate": 9.919709236619786e-06, + "loss": 0.0237, + "step": 1320 + }, + { + "epoch": 0.7628880762888076, + "grad_norm": 0.09353283792734146, + "learning_rate": 9.916094993302144e-06, + "loss": 0.0247, + "step": 1330 + }, + { + "epoch": 0.7686240768624076, + "grad_norm": 2.4819159507751465, + "learning_rate": 9.912401872146047e-06, + "loss": 0.0361, + "step": 1340 + }, + { + "epoch": 0.7743600774360078, + "grad_norm": 1.608819842338562, + "learning_rate": 9.908629932406943e-06, + "loss": 0.0123, + "step": 1350 + }, + { + "epoch": 0.7743600774360078, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11693766713142395, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2248, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.766, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.407, + "step": 1350 + }, + { + "epoch": 0.7800960780096078, + "grad_norm": 0.044521648436784744, + "learning_rate": 9.90477923460491e-06, + "loss": 0.0559, + "step": 1360 + }, + { + "epoch": 0.7858320785832078, + "grad_norm": 1.6780263185501099, + "learning_rate": 9.900849840523686e-06, + "loss": 0.016, + "step": 1370 + }, + { + "epoch": 0.7915680791568079, + "grad_norm": 0.20249976217746735, + "learning_rate": 9.896841813209675e-06, + "loss": 0.0103, + "step": 1380 + }, + { + "epoch": 0.797304079730408, + "grad_norm": 0.5444568991661072, + "learning_rate": 9.892755216970937e-06, + "loss": 0.0265, + "step": 1390 + }, + { + "epoch": 0.803040080304008, + "grad_norm": 0.39745691418647766, + "learning_rate": 9.888590117376154e-06, + "loss": 0.0305, + "step": 1400 + }, + { + "epoch": 0.803040080304008, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07220673561096191, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2419, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.741, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.394, + "step": 1400 + }, + { + "epoch": 0.8087760808776081, + "grad_norm": 3.1255667209625244, + "learning_rate": 9.884346581253584e-06, + "loss": 0.0155, + "step": 1410 + }, + { + "epoch": 0.8145120814512081, + "grad_norm": 0.28445881605148315, + "learning_rate": 9.88002467668998e-06, + "loss": 0.0543, + "step": 1420 + }, + { + "epoch": 0.8202480820248083, + "grad_norm": 2.316286087036133, + "learning_rate": 9.875624473029508e-06, + "loss": 0.0208, + "step": 1430 + }, + { + "epoch": 0.8259840825984083, + "grad_norm": 0.3444057106971741, + "learning_rate": 9.871146040872621e-06, + "loss": 0.0375, + "step": 1440 + }, + { + "epoch": 0.8317200831720083, + "grad_norm": 0.36751267313957214, + "learning_rate": 9.86658945207494e-06, + "loss": 0.0273, + "step": 1450 + }, + { + "epoch": 0.8317200831720083, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06544554233551025, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1664, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.851, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.449, + "step": 1450 + }, + { + "epoch": 0.8374560837456084, + "grad_norm": 0.6214566230773926, + "learning_rate": 9.861954779746092e-06, + "loss": 0.0095, + "step": 1460 + }, + { + "epoch": 0.8431920843192084, + "grad_norm": 0.6569352149963379, + "learning_rate": 9.857242098248543e-06, + "loss": 0.024, + "step": 1470 + }, + { + "epoch": 0.8489280848928085, + "grad_norm": 1.6928707361221313, + "learning_rate": 9.852451483196394e-06, + "loss": 0.0158, + "step": 1480 + }, + { + "epoch": 0.8546640854664086, + "grad_norm": 1.0950285196304321, + "learning_rate": 9.847583011454187e-06, + "loss": 0.0099, + "step": 1490 + }, + { + "epoch": 0.8604000860400086, + "grad_norm": 1.4251108169555664, + "learning_rate": 9.842636761135651e-06, + "loss": 0.0677, + "step": 1500 + }, + { + "epoch": 0.8604000860400086, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11534915119409561, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2226, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.769, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.408, + "step": 1500 + }, + { + "epoch": 0.8661360866136086, + "grad_norm": 0.9497309923171997, + "learning_rate": 9.837612811602462e-06, + "loss": 0.0282, + "step": 1510 + }, + { + "epoch": 0.8718720871872088, + "grad_norm": 0.9512405395507812, + "learning_rate": 9.832511243462962e-06, + "loss": 0.0079, + "step": 1520 + }, + { + "epoch": 0.8776080877608088, + "grad_norm": 0.04131248593330383, + "learning_rate": 9.827332138570878e-06, + "loss": 0.0051, + "step": 1530 + }, + { + "epoch": 0.8833440883344088, + "grad_norm": 1.3209096193313599, + "learning_rate": 9.822075580023987e-06, + "loss": 0.0132, + "step": 1540 + }, + { + "epoch": 0.8890800889080089, + "grad_norm": 0.05282355472445488, + "learning_rate": 9.816741652162807e-06, + "loss": 0.0592, + "step": 1550 + }, + { + "epoch": 0.8890800889080089, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08604957163333893, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3059, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.649, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.348, + "step": 1550 + }, + { + "epoch": 0.894816089481609, + "grad_norm": 0.5249603390693665, + "learning_rate": 9.811330440569226e-06, + "loss": 0.0169, + "step": 1560 + }, + { + "epoch": 0.900552090055209, + "grad_norm": 1.4722495079040527, + "learning_rate": 9.805842032065135e-06, + "loss": 0.0257, + "step": 1570 + }, + { + "epoch": 0.9062880906288091, + "grad_norm": 1.664174199104309, + "learning_rate": 9.800276514711044e-06, + "loss": 0.0318, + "step": 1580 + }, + { + "epoch": 0.9120240912024091, + "grad_norm": 0.017392676323652267, + "learning_rate": 9.794633977804646e-06, + "loss": 0.0165, + "step": 1590 + }, + { + "epoch": 0.9177600917760091, + "grad_norm": 0.18038752675056458, + "learning_rate": 9.788914511879412e-06, + "loss": 0.0488, + "step": 1600 + }, + { + "epoch": 0.9177600917760091, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08364205807447433, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2219, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.77, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.409, + "step": 1600 + }, + { + "epoch": 0.9234960923496093, + "grad_norm": 1.785083293914795, + "learning_rate": 9.783118208703115e-06, + "loss": 0.0212, + "step": 1610 + }, + { + "epoch": 0.9292320929232093, + "grad_norm": 0.36738669872283936, + "learning_rate": 9.777245161276372e-06, + "loss": 0.0221, + "step": 1620 + }, + { + "epoch": 0.9349680934968093, + "grad_norm": 0.9083078503608704, + "learning_rate": 9.77129546383115e-06, + "loss": 0.0261, + "step": 1630 + }, + { + "epoch": 0.9407040940704094, + "grad_norm": 0.5853396058082581, + "learning_rate": 9.765269211829245e-06, + "loss": 0.011, + "step": 1640 + }, + { + "epoch": 0.9464400946440095, + "grad_norm": 0.027249574661254883, + "learning_rate": 9.759166501960762e-06, + "loss": 0.0156, + "step": 1650 + }, + { + "epoch": 0.9464400946440095, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07951950281858444, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2513, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.727, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.387, + "step": 1650 + }, + { + "epoch": 0.9521760952176095, + "grad_norm": 0.5655274987220764, + "learning_rate": 9.752987432142556e-06, + "loss": 0.0141, + "step": 1660 + }, + { + "epoch": 0.9579120957912096, + "grad_norm": 0.20478393137454987, + "learning_rate": 9.746732101516665e-06, + "loss": 0.0225, + "step": 1670 + }, + { + "epoch": 0.9636480963648096, + "grad_norm": 0.38293856382369995, + "learning_rate": 9.740400610448714e-06, + "loss": 0.0228, + "step": 1680 + }, + { + "epoch": 0.9693840969384097, + "grad_norm": 0.4184674918651581, + "learning_rate": 9.733993060526313e-06, + "loss": 0.0278, + "step": 1690 + }, + { + "epoch": 0.9751200975120098, + "grad_norm": 0.6761838793754578, + "learning_rate": 9.727509554557416e-06, + "loss": 0.0148, + "step": 1700 + }, + { + "epoch": 0.9751200975120098, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0835602656006813, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2977, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.661, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.354, + "step": 1700 + }, + { + "epoch": 0.9808560980856098, + "grad_norm": 0.10887222737073898, + "learning_rate": 9.720950196568689e-06, + "loss": 0.0065, + "step": 1710 + }, + { + "epoch": 0.9865920986592098, + "grad_norm": 0.13018013536930084, + "learning_rate": 9.714315091803815e-06, + "loss": 0.0083, + "step": 1720 + }, + { + "epoch": 0.99232809923281, + "grad_norm": 0.1999017894268036, + "learning_rate": 9.707604346721833e-06, + "loss": 0.0097, + "step": 1730 + }, + { + "epoch": 0.99806409980641, + "grad_norm": 0.06055552884936333, + "learning_rate": 9.700818068995407e-06, + "loss": 0.0255, + "step": 1740 + }, + { + "epoch": 1.00344160034416, + "grad_norm": 2.305107831954956, + "learning_rate": 9.693956367509117e-06, + "loss": 0.021, + "step": 1750 + }, + { + "epoch": 1.00344160034416, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11077206581830978, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3027, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.653, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.35, + "step": 1750 + }, + { + "epoch": 1.00917760091776, + "grad_norm": 2.2754151821136475, + "learning_rate": 9.687019352357699e-06, + "loss": 0.0398, + "step": 1760 + }, + { + "epoch": 1.0149136014913602, + "grad_norm": 0.03042173944413662, + "learning_rate": 9.680007134844279e-06, + "loss": 0.0189, + "step": 1770 + }, + { + "epoch": 1.0206496020649602, + "grad_norm": 1.6503928899765015, + "learning_rate": 9.672919827478598e-06, + "loss": 0.0111, + "step": 1780 + }, + { + "epoch": 1.0263856026385603, + "grad_norm": 1.7391407489776611, + "learning_rate": 9.665757543975196e-06, + "loss": 0.0204, + "step": 1790 + }, + { + "epoch": 1.0321216032121603, + "grad_norm": 1.6538548469543457, + "learning_rate": 9.658520399251592e-06, + "loss": 0.0165, + "step": 1800 + }, + { + "epoch": 1.0321216032121603, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07668553292751312, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1222, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.915, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.481, + "step": 1800 + }, + { + "epoch": 1.0378576037857603, + "grad_norm": 0.022262316197156906, + "learning_rate": 9.651208509426442e-06, + "loss": 0.0053, + "step": 1810 + }, + { + "epoch": 1.0435936043593603, + "grad_norm": 1.768957495689392, + "learning_rate": 9.64382199181767e-06, + "loss": 0.0164, + "step": 1820 + }, + { + "epoch": 1.0493296049329606, + "grad_norm": 0.11224523931741714, + "learning_rate": 9.63636096494059e-06, + "loss": 0.0216, + "step": 1830 + }, + { + "epoch": 1.0550656055065606, + "grad_norm": 0.022123126313090324, + "learning_rate": 9.628825548506002e-06, + "loss": 0.0069, + "step": 1840 + }, + { + "epoch": 1.0608016060801606, + "grad_norm": 4.303464412689209, + "learning_rate": 9.621215863418276e-06, + "loss": 0.0253, + "step": 1850 + }, + { + "epoch": 1.0608016060801606, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0671575739979744, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2804, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.686, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.366, + "step": 1850 + }, + { + "epoch": 1.0665376066537606, + "grad_norm": 0.08216200768947601, + "learning_rate": 9.61353203177341e-06, + "loss": 0.0052, + "step": 1860 + }, + { + "epoch": 1.0722736072273606, + "grad_norm": 1.3728184700012207, + "learning_rate": 9.605774176857062e-06, + "loss": 0.0176, + "step": 1870 + }, + { + "epoch": 1.078009607800961, + "grad_norm": 0.0963163748383522, + "learning_rate": 9.597942423142586e-06, + "loss": 0.0165, + "step": 1880 + }, + { + "epoch": 1.083745608374561, + "grad_norm": 0.09453441202640533, + "learning_rate": 9.59003689628903e-06, + "loss": 0.0212, + "step": 1890 + }, + { + "epoch": 1.089481608948161, + "grad_norm": 0.0946192815899849, + "learning_rate": 9.582057723139115e-06, + "loss": 0.0258, + "step": 1900 + }, + { + "epoch": 1.089481608948161, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07539978623390198, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2173, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.777, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.412, + "step": 1900 + }, + { + "epoch": 1.095217609521761, + "grad_norm": 0.011775615625083447, + "learning_rate": 9.574005031717203e-06, + "loss": 0.0093, + "step": 1910 + }, + { + "epoch": 1.100953610095361, + "grad_norm": 0.06459964066743851, + "learning_rate": 9.565878951227247e-06, + "loss": 0.0082, + "step": 1920 + }, + { + "epoch": 1.106689610668961, + "grad_norm": 0.020801063627004623, + "learning_rate": 9.557679612050708e-06, + "loss": 0.0067, + "step": 1930 + }, + { + "epoch": 1.112425611242561, + "grad_norm": 0.048270970582962036, + "learning_rate": 9.549407145744473e-06, + "loss": 0.0442, + "step": 1940 + }, + { + "epoch": 1.1181616118161612, + "grad_norm": 1.6280121803283691, + "learning_rate": 9.541061685038742e-06, + "loss": 0.0258, + "step": 1950 + }, + { + "epoch": 1.1181616118161612, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0749003142118454, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1385, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.891, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.469, + "step": 1950 + }, + { + "epoch": 1.1238976123897613, + "grad_norm": 0.4389003813266754, + "learning_rate": 9.532643363834891e-06, + "loss": 0.0165, + "step": 1960 + }, + { + "epoch": 1.1296336129633613, + "grad_norm": 0.07638214528560638, + "learning_rate": 9.524152317203337e-06, + "loss": 0.0339, + "step": 1970 + }, + { + "epoch": 1.1353696135369613, + "grad_norm": 0.26499396562576294, + "learning_rate": 9.515588681381356e-06, + "loss": 0.0229, + "step": 1980 + }, + { + "epoch": 1.1411056141105613, + "grad_norm": 0.7412366271018982, + "learning_rate": 9.506952593770908e-06, + "loss": 0.0157, + "step": 1990 + }, + { + "epoch": 1.1468416146841616, + "grad_norm": 0.16746670007705688, + "learning_rate": 9.498244192936428e-06, + "loss": 0.0102, + "step": 2000 + }, + { + "epoch": 1.1468416146841616, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09212490171194077, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1271, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.908, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.478, + "step": 2000 + }, + { + "epoch": 1.1525776152577616, + "grad_norm": 0.182969868183136, + "learning_rate": 9.489463618602602e-06, + "loss": 0.0111, + "step": 2010 + }, + { + "epoch": 1.1583136158313616, + "grad_norm": 1.3869882822036743, + "learning_rate": 9.480611011652128e-06, + "loss": 0.0146, + "step": 2020 + }, + { + "epoch": 1.1640496164049616, + "grad_norm": 0.0466344878077507, + "learning_rate": 9.471686514123455e-06, + "loss": 0.009, + "step": 2030 + }, + { + "epoch": 1.1697856169785616, + "grad_norm": 0.07296488434076309, + "learning_rate": 9.462690269208498e-06, + "loss": 0.0354, + "step": 2040 + }, + { + "epoch": 1.1755216175521617, + "grad_norm": 1.9699822664260864, + "learning_rate": 9.453622421250353e-06, + "loss": 0.0114, + "step": 2050 + }, + { + "epoch": 1.1755216175521617, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09612752497196198, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1568, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.865, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.456, + "step": 2050 + }, + { + "epoch": 1.1812576181257617, + "grad_norm": 2.4988596439361572, + "learning_rate": 9.444483115740968e-06, + "loss": 0.0235, + "step": 2060 + }, + { + "epoch": 1.186993618699362, + "grad_norm": 0.0865960419178009, + "learning_rate": 9.435272499318815e-06, + "loss": 0.0197, + "step": 2070 + }, + { + "epoch": 1.192729619272962, + "grad_norm": 0.7196880578994751, + "learning_rate": 9.425990719766542e-06, + "loss": 0.0085, + "step": 2080 + }, + { + "epoch": 1.198465619846562, + "grad_norm": 0.4979308843612671, + "learning_rate": 9.416637926008587e-06, + "loss": 0.0177, + "step": 2090 + }, + { + "epoch": 1.204201620420162, + "grad_norm": 1.3404853343963623, + "learning_rate": 9.407214268108805e-06, + "loss": 0.0079, + "step": 2100 + }, + { + "epoch": 1.204201620420162, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07606372237205505, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2379, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.747, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.397, + "step": 2100 + }, + { + "epoch": 1.209937620993762, + "grad_norm": 0.007559036836028099, + "learning_rate": 9.397719897268049e-06, + "loss": 0.0122, + "step": 2110 + }, + { + "epoch": 1.2156736215673623, + "grad_norm": 0.036009229719638824, + "learning_rate": 9.388154965821754e-06, + "loss": 0.0195, + "step": 2120 + }, + { + "epoch": 1.2214096221409623, + "grad_norm": 0.23815706372261047, + "learning_rate": 9.37851962723748e-06, + "loss": 0.0311, + "step": 2130 + }, + { + "epoch": 1.2271456227145623, + "grad_norm": 0.32620975375175476, + "learning_rate": 9.36881403611246e-06, + "loss": 0.0235, + "step": 2140 + }, + { + "epoch": 1.2328816232881623, + "grad_norm": 0.6975133419036865, + "learning_rate": 9.359038348171113e-06, + "loss": 0.0398, + "step": 2150 + }, + { + "epoch": 1.2328816232881623, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07932226359844208, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1766, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.836, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.442, + "step": 2150 + }, + { + "epoch": 1.2386176238617623, + "grad_norm": 0.055091869086027145, + "learning_rate": 9.349192720262556e-06, + "loss": 0.031, + "step": 2160 + }, + { + "epoch": 1.2443536244353623, + "grad_norm": 0.07653547823429108, + "learning_rate": 9.33927731035807e-06, + "loss": 0.0188, + "step": 2170 + }, + { + "epoch": 1.2500896250089624, + "grad_norm": 1.0113948583602905, + "learning_rate": 9.329292277548584e-06, + "loss": 0.023, + "step": 2180 + }, + { + "epoch": 1.2558256255825626, + "grad_norm": 0.38709500432014465, + "learning_rate": 9.319237782042108e-06, + "loss": 0.0134, + "step": 2190 + }, + { + "epoch": 1.2615616261561626, + "grad_norm": 1.0346875190734863, + "learning_rate": 9.309113985161169e-06, + "loss": 0.0272, + "step": 2200 + }, + { + "epoch": 1.2615616261561626, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06406532227993011, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1643, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.854, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.451, + "step": 2200 + }, + { + "epoch": 1.2672976267297626, + "grad_norm": 3.084003448486328, + "learning_rate": 9.298921049340226e-06, + "loss": 0.0134, + "step": 2210 + }, + { + "epoch": 1.2730336273033627, + "grad_norm": 0.3186517059803009, + "learning_rate": 9.288659138123053e-06, + "loss": 0.0142, + "step": 2220 + }, + { + "epoch": 1.278769627876963, + "grad_norm": 0.02311491221189499, + "learning_rate": 9.278328416160125e-06, + "loss": 0.0116, + "step": 2230 + }, + { + "epoch": 1.284505628450563, + "grad_norm": 0.021789675578475, + "learning_rate": 9.267929049205976e-06, + "loss": 0.003, + "step": 2240 + }, + { + "epoch": 1.290241629024163, + "grad_norm": 0.004828541073948145, + "learning_rate": 9.25746120411653e-06, + "loss": 0.0074, + "step": 2250 + }, + { + "epoch": 1.290241629024163, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09356704354286194, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1187, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.92, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.484, + "step": 2250 + }, + { + "epoch": 1.295977629597763, + "grad_norm": 4.326236248016357, + "learning_rate": 9.246925048846433e-06, + "loss": 0.0318, + "step": 2260 + }, + { + "epoch": 1.301713630171363, + "grad_norm": 0.10032517462968826, + "learning_rate": 9.236320752446357e-06, + "loss": 0.0248, + "step": 2270 + }, + { + "epoch": 1.307449630744963, + "grad_norm": 1.968422293663025, + "learning_rate": 9.225648485060283e-06, + "loss": 0.0245, + "step": 2280 + }, + { + "epoch": 1.313185631318563, + "grad_norm": 0.9600183367729187, + "learning_rate": 9.214908417922774e-06, + "loss": 0.0169, + "step": 2290 + }, + { + "epoch": 1.3189216318921633, + "grad_norm": 0.41536957025527954, + "learning_rate": 9.20410072335623e-06, + "loss": 0.0103, + "step": 2300 + }, + { + "epoch": 1.3189216318921633, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06794232130050659, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1585, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.862, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.455, + "step": 2300 + }, + { + "epoch": 1.3246576324657633, + "grad_norm": 0.7403652667999268, + "learning_rate": 9.19322557476812e-06, + "loss": 0.0141, + "step": 2310 + }, + { + "epoch": 1.3303936330393633, + "grad_norm": 0.03548659384250641, + "learning_rate": 9.182283146648197e-06, + "loss": 0.0214, + "step": 2320 + }, + { + "epoch": 1.3361296336129633, + "grad_norm": 2.034658193588257, + "learning_rate": 9.171273614565704e-06, + "loss": 0.008, + "step": 2330 + }, + { + "epoch": 1.3418656341865633, + "grad_norm": 1.0196254253387451, + "learning_rate": 9.160197155166559e-06, + "loss": 0.0459, + "step": 2340 + }, + { + "epoch": 1.3476016347601636, + "grad_norm": 0.014422023668885231, + "learning_rate": 9.149053946170507e-06, + "loss": 0.0198, + "step": 2350 + }, + { + "epoch": 1.3476016347601636, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.058592408895492554, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2196, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.773, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.41, + "step": 2350 + }, + { + "epoch": 1.3533376353337636, + "grad_norm": 0.019738655537366867, + "learning_rate": 9.137844166368289e-06, + "loss": 0.0233, + "step": 2360 + }, + { + "epoch": 1.3590736359073636, + "grad_norm": 0.03966844826936722, + "learning_rate": 9.126567995618752e-06, + "loss": 0.0082, + "step": 2370 + }, + { + "epoch": 1.3648096364809637, + "grad_norm": 0.10851484537124634, + "learning_rate": 9.115225614845979e-06, + "loss": 0.0132, + "step": 2380 + }, + { + "epoch": 1.3705456370545637, + "grad_norm": 2.611233711242676, + "learning_rate": 9.103817206036383e-06, + "loss": 0.0062, + "step": 2390 + }, + { + "epoch": 1.3762816376281637, + "grad_norm": 2.288485050201416, + "learning_rate": 9.092342952235777e-06, + "loss": 0.0476, + "step": 2400 + }, + { + "epoch": 1.3762816376281637, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07158029079437256, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1533, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.87, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.459, + "step": 2400 + }, + { + "epoch": 1.3820176382017637, + "grad_norm": 0.04291702061891556, + "learning_rate": 9.080803037546454e-06, + "loss": 0.0031, + "step": 2410 + }, + { + "epoch": 1.387753638775364, + "grad_norm": 0.008397878147661686, + "learning_rate": 9.069197647124216e-06, + "loss": 0.0121, + "step": 2420 + }, + { + "epoch": 1.393489639348964, + "grad_norm": 0.015200245194137096, + "learning_rate": 9.057526967175415e-06, + "loss": 0.0047, + "step": 2430 + }, + { + "epoch": 1.399225639922564, + "grad_norm": 3.2773385047912598, + "learning_rate": 9.04579118495396e-06, + "loss": 0.005, + "step": 2440 + }, + { + "epoch": 1.404961640496164, + "grad_norm": 0.5176500678062439, + "learning_rate": 9.033990488758317e-06, + "loss": 0.0329, + "step": 2450 + }, + { + "epoch": 1.404961640496164, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.1131686270236969, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1939, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.811, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.429, + "step": 2450 + }, + { + "epoch": 1.410697641069764, + "grad_norm": 0.002755386522039771, + "learning_rate": 9.02212506792848e-06, + "loss": 0.0164, + "step": 2460 + }, + { + "epoch": 1.4164336416433643, + "grad_norm": 0.0466199554502964, + "learning_rate": 9.01019511284294e-06, + "loss": 0.0233, + "step": 2470 + }, + { + "epoch": 1.4221696422169643, + "grad_norm": 0.8431953191757202, + "learning_rate": 8.99820081491563e-06, + "loss": 0.0051, + "step": 2480 + }, + { + "epoch": 1.4279056427905643, + "grad_norm": 0.035143181681632996, + "learning_rate": 8.98614236659285e-06, + "loss": 0.0283, + "step": 2490 + }, + { + "epoch": 1.4336416433641643, + "grad_norm": 0.05582081899046898, + "learning_rate": 8.97401996135018e-06, + "loss": 0.0073, + "step": 2500 + }, + { + "epoch": 1.4336416433641643, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09689504653215408, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.294, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.666, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.356, + "step": 2500 + }, + { + "epoch": 1.4393776439377644, + "grad_norm": 2.2861459255218506, + "learning_rate": 8.961833793689384e-06, + "loss": 0.0107, + "step": 2510 + }, + { + "epoch": 1.4451136445113644, + "grad_norm": 2.3350398540496826, + "learning_rate": 8.94958405913527e-06, + "loss": 0.0128, + "step": 2520 + }, + { + "epoch": 1.4508496450849644, + "grad_norm": 0.019439250230789185, + "learning_rate": 8.937270954232576e-06, + "loss": 0.0061, + "step": 2530 + }, + { + "epoch": 1.4565856456585646, + "grad_norm": 0.002678696997463703, + "learning_rate": 8.924894676542801e-06, + "loss": 0.0085, + "step": 2540 + }, + { + "epoch": 1.4623216462321647, + "grad_norm": 0.03475391864776611, + "learning_rate": 8.91245542464104e-06, + "loss": 0.0234, + "step": 2550 + }, + { + "epoch": 1.4623216462321647, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10303693264722824, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1871, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.821, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, + "step": 2550 + }, + { + "epoch": 1.4680576468057647, + "grad_norm": 0.09974029660224915, + "learning_rate": 8.8999533981128e-06, + "loss": 0.013, + "step": 2560 + }, + { + "epoch": 1.4737936473793647, + "grad_norm": 0.19927239418029785, + "learning_rate": 8.887388797550791e-06, + "loss": 0.0033, + "step": 2570 + }, + { + "epoch": 1.479529647952965, + "grad_norm": 0.03026706352829933, + "learning_rate": 8.874761824551717e-06, + "loss": 0.0219, + "step": 2580 + }, + { + "epoch": 1.485265648526565, + "grad_norm": 0.8363674879074097, + "learning_rate": 8.862072681713027e-06, + "loss": 0.0167, + "step": 2590 + }, + { + "epoch": 1.491001649100165, + "grad_norm": 2.117938280105591, + "learning_rate": 8.84932157262969e-06, + "loss": 0.0365, + "step": 2600 + }, + { + "epoch": 1.491001649100165, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.11520245671272278, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2361, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.75, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.398, + "step": 2600 + }, + { + "epoch": 1.496737649673765, + "grad_norm": 0.14108580350875854, + "learning_rate": 8.836508701890892e-06, + "loss": 0.0196, + "step": 2610 + }, + { + "epoch": 1.502473650247365, + "grad_norm": 2.1537435054779053, + "learning_rate": 8.823634275076792e-06, + "loss": 0.0207, + "step": 2620 + }, + { + "epoch": 1.508209650820965, + "grad_norm": 0.93586665391922, + "learning_rate": 8.81069849875519e-06, + "loss": 0.0125, + "step": 2630 + }, + { + "epoch": 1.513945651394565, + "grad_norm": 3.1711156368255615, + "learning_rate": 8.797701580478235e-06, + "loss": 0.0094, + "step": 2640 + }, + { + "epoch": 1.519681651968165, + "grad_norm": 0.0233930516988039, + "learning_rate": 8.784643728779089e-06, + "loss": 0.0152, + "step": 2650 + }, + { + "epoch": 1.519681651968165, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07520527392625809, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2328, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.754, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.401, + "step": 2650 + }, + { + "epoch": 1.525417652541765, + "grad_norm": 0.14535315334796906, + "learning_rate": 8.771525153168576e-06, + "loss": 0.0035, + "step": 2660 + }, + { + "epoch": 1.5311536531153653, + "grad_norm": 0.3828493654727936, + "learning_rate": 8.758346064131824e-06, + "loss": 0.0026, + "step": 2670 + }, + { + "epoch": 1.5368896536889654, + "grad_norm": 2.7801029682159424, + "learning_rate": 8.745106673124888e-06, + "loss": 0.0215, + "step": 2680 + }, + { + "epoch": 1.5426256542625654, + "grad_norm": 1.6995360851287842, + "learning_rate": 8.731807192571359e-06, + "loss": 0.022, + "step": 2690 + }, + { + "epoch": 1.5483616548361656, + "grad_norm": 0.24279280006885529, + "learning_rate": 8.718447835858951e-06, + "loss": 0.0109, + "step": 2700 + }, + { + "epoch": 1.5483616548361656, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08999116718769073, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1893, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.817, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.432, + "step": 2700 + }, + { + "epoch": 1.5540976554097656, + "grad_norm": 0.015813475474715233, + "learning_rate": 8.705028817336083e-06, + "loss": 0.0019, + "step": 2710 + }, + { + "epoch": 1.5598336559833657, + "grad_norm": 0.017328623682260513, + "learning_rate": 8.691550352308431e-06, + "loss": 0.0044, + "step": 2720 + }, + { + "epoch": 1.5655696565569657, + "grad_norm": 1.36957585811615, + "learning_rate": 8.678012657035487e-06, + "loss": 0.0033, + "step": 2730 + }, + { + "epoch": 1.5713056571305657, + "grad_norm": 0.8026090264320374, + "learning_rate": 8.664415948727076e-06, + "loss": 0.0063, + "step": 2740 + }, + { + "epoch": 1.5770416577041657, + "grad_norm": 0.038070451468229294, + "learning_rate": 8.65076044553988e-06, + "loss": 0.02, + "step": 2750 + }, + { + "epoch": 1.5770416577041657, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09743982553482056, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1673, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.849, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.448, + "step": 2750 + }, + { + "epoch": 1.5827776582777657, + "grad_norm": 2.638280153274536, + "learning_rate": 8.63704636657393e-06, + "loss": 0.0205, + "step": 2760 + }, + { + "epoch": 1.5885136588513658, + "grad_norm": 0.08616320043802261, + "learning_rate": 8.623273931869094e-06, + "loss": 0.0151, + "step": 2770 + }, + { + "epoch": 1.594249659424966, + "grad_norm": 0.0016790858935564756, + "learning_rate": 8.609443362401553e-06, + "loss": 0.0077, + "step": 2780 + }, + { + "epoch": 1.599985659998566, + "grad_norm": 0.0031273786444216967, + "learning_rate": 8.595554880080246e-06, + "loss": 0.0014, + "step": 2790 + }, + { + "epoch": 1.605721660572166, + "grad_norm": 2.3747873306274414, + "learning_rate": 8.581608707743312e-06, + "loss": 0.0387, + "step": 2800 + }, + { + "epoch": 1.605721660572166, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10180775076150894, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1502, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.874, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.461, + "step": 2800 + }, + { + "epoch": 1.6114576611457663, + "grad_norm": 0.03558593988418579, + "learning_rate": 8.567605069154517e-06, + "loss": 0.0009, + "step": 2810 + }, + { + "epoch": 1.6171936617193663, + "grad_norm": 0.005191161762923002, + "learning_rate": 8.55354418899966e-06, + "loss": 0.0349, + "step": 2820 + }, + { + "epoch": 1.6229296622929663, + "grad_norm": 1.5500961542129517, + "learning_rate": 8.539426292882976e-06, + "loss": 0.0296, + "step": 2830 + }, + { + "epoch": 1.6286656628665663, + "grad_norm": 0.12692737579345703, + "learning_rate": 8.525251607323506e-06, + "loss": 0.0076, + "step": 2840 + }, + { + "epoch": 1.6344016634401664, + "grad_norm": 0.034373391419649124, + "learning_rate": 8.511020359751467e-06, + "loss": 0.0125, + "step": 2850 + }, + { + "epoch": 1.6344016634401664, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07807676494121552, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1709, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.844, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.446, + "step": 2850 + }, + { + "epoch": 1.6401376640137664, + "grad_norm": 0.3334125876426697, + "learning_rate": 8.496732778504608e-06, + "loss": 0.007, + "step": 2860 + }, + { + "epoch": 1.6458736645873664, + "grad_norm": 1.8848798274993896, + "learning_rate": 8.482389092824535e-06, + "loss": 0.0279, + "step": 2870 + }, + { + "epoch": 1.6516096651609664, + "grad_norm": 0.0071907006204128265, + "learning_rate": 8.46798953285304e-06, + "loss": 0.0058, + "step": 2880 + }, + { + "epoch": 1.6573456657345664, + "grad_norm": 0.01580858789384365, + "learning_rate": 8.453534329628414e-06, + "loss": 0.0111, + "step": 2890 + }, + { + "epoch": 1.6630816663081667, + "grad_norm": 0.5010847449302673, + "learning_rate": 8.439023715081729e-06, + "loss": 0.0205, + "step": 2900 + }, + { + "epoch": 1.6630816663081667, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0824664756655693, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2284, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.761, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.404, + "step": 2900 + }, + { + "epoch": 1.6688176668817667, + "grad_norm": 0.16337744891643524, + "learning_rate": 8.42445792203312e-06, + "loss": 0.0175, + "step": 2910 + }, + { + "epoch": 1.6745536674553667, + "grad_norm": 0.054014191031455994, + "learning_rate": 8.409837184188056e-06, + "loss": 0.0073, + "step": 2920 + }, + { + "epoch": 1.680289668028967, + "grad_norm": 3.6512198448181152, + "learning_rate": 8.395161736133579e-06, + "loss": 0.0226, + "step": 2930 + }, + { + "epoch": 1.686025668602567, + "grad_norm": 0.6102538704872131, + "learning_rate": 8.380431813334548e-06, + "loss": 0.0033, + "step": 2940 + }, + { + "epoch": 1.691761669176167, + "grad_norm": 2.3051533699035645, + "learning_rate": 8.365647652129865e-06, + "loss": 0.0327, + "step": 2950 + }, + { + "epoch": 1.691761669176167, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0759689062833786, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.5562, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.293, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.17, + "step": 2950 + }, + { + "epoch": 1.697497669749767, + "grad_norm": 0.008392853662371635, + "learning_rate": 8.350809489728673e-06, + "loss": 0.0279, + "step": 2960 + }, + { + "epoch": 1.703233670323367, + "grad_norm": 1.614988088607788, + "learning_rate": 8.33591756420655e-06, + "loss": 0.0102, + "step": 2970 + }, + { + "epoch": 1.708969670896967, + "grad_norm": 0.04466373473405838, + "learning_rate": 8.320972114501698e-06, + "loss": 0.0158, + "step": 2980 + }, + { + "epoch": 1.714705671470567, + "grad_norm": 1.4472012519836426, + "learning_rate": 8.305973380411107e-06, + "loss": 0.0437, + "step": 2990 + }, + { + "epoch": 1.720441672044167, + "grad_norm": 0.0326494500041008, + "learning_rate": 8.290921602586699e-06, + "loss": 0.0188, + "step": 3000 + }, + { + "epoch": 1.720441672044167, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06277668476104736, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2361, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.75, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.398, + "step": 3000 + }, + { + "epoch": 1.7261776726177671, + "grad_norm": 2.090261697769165, + "learning_rate": 8.275817022531479e-06, + "loss": 0.0229, + "step": 3010 + }, + { + "epoch": 1.7319136731913674, + "grad_norm": 1.1351670026779175, + "learning_rate": 8.260659882595647e-06, + "loss": 0.005, + "step": 3020 + }, + { + "epoch": 1.7376496737649674, + "grad_norm": 0.0521712526679039, + "learning_rate": 8.245450425972728e-06, + "loss": 0.0126, + "step": 3030 + }, + { + "epoch": 1.7433856743385674, + "grad_norm": 0.04499056562781334, + "learning_rate": 8.230188896695643e-06, + "loss": 0.0133, + "step": 3040 + }, + { + "epoch": 1.7491216749121676, + "grad_norm": 0.007846461609005928, + "learning_rate": 8.214875539632825e-06, + "loss": 0.0191, + "step": 3050 + }, + { + "epoch": 1.7491216749121676, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07422397285699844, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1752, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.838, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.443, + "step": 3050 + }, + { + "epoch": 1.7548576754857677, + "grad_norm": 0.055605433881282806, + "learning_rate": 8.199510600484261e-06, + "loss": 0.0148, + "step": 3060 + }, + { + "epoch": 1.7605936760593677, + "grad_norm": 0.27163052558898926, + "learning_rate": 8.184094325777573e-06, + "loss": 0.0075, + "step": 3070 + }, + { + "epoch": 1.7663296766329677, + "grad_norm": 0.11399796605110168, + "learning_rate": 8.168626962864045e-06, + "loss": 0.0016, + "step": 3080 + }, + { + "epoch": 1.7720656772065677, + "grad_norm": 0.07731972634792328, + "learning_rate": 8.153108759914669e-06, + "loss": 0.0141, + "step": 3090 + }, + { + "epoch": 1.7778016777801677, + "grad_norm": 0.2923997938632965, + "learning_rate": 8.13753996591615e-06, + "loss": 0.0047, + "step": 3100 + }, + { + "epoch": 1.7778016777801677, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06707713752985, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.4009, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.513, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.28, + "step": 3100 + }, + { + "epoch": 1.7835376783537678, + "grad_norm": 0.26035454869270325, + "learning_rate": 8.121920830666918e-06, + "loss": 0.0092, + "step": 3110 + }, + { + "epoch": 1.7892736789273678, + "grad_norm": 3.143533706665039, + "learning_rate": 8.106251604773125e-06, + "loss": 0.0141, + "step": 3120 + }, + { + "epoch": 1.7950096795009678, + "grad_norm": 1.8246173858642578, + "learning_rate": 8.090532539644608e-06, + "loss": 0.0249, + "step": 3130 + }, + { + "epoch": 1.800745680074568, + "grad_norm": 0.01822010800242424, + "learning_rate": 8.074763887490878e-06, + "loss": 0.0182, + "step": 3140 + }, + { + "epoch": 1.806481680648168, + "grad_norm": 1.8796803951263428, + "learning_rate": 8.058945901317047e-06, + "loss": 0.0276, + "step": 3150 + }, + { + "epoch": 1.806481680648168, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06196051836013794, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1882, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.819, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.433, + "step": 3150 + }, + { + "epoch": 1.812217681221768, + "grad_norm": 0.3315775692462921, + "learning_rate": 8.043078834919792e-06, + "loss": 0.0053, + "step": 3160 + }, + { + "epoch": 1.8179536817953683, + "grad_norm": 0.6810622215270996, + "learning_rate": 8.027162942883271e-06, + "loss": 0.0027, + "step": 3170 + }, + { + "epoch": 1.8236896823689683, + "grad_norm": 0.04800894856452942, + "learning_rate": 8.011198480575037e-06, + "loss": 0.0104, + "step": 3180 + }, + { + "epoch": 1.8294256829425684, + "grad_norm": 0.21183735132217407, + "learning_rate": 7.995185704141948e-06, + "loss": 0.0139, + "step": 3190 + }, + { + "epoch": 1.8351616835161684, + "grad_norm": 0.05010043457150459, + "learning_rate": 7.979124870506052e-06, + "loss": 0.013, + "step": 3200 + }, + { + "epoch": 1.8351616835161684, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08539190888404846, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2295, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.759, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.403, + "step": 3200 + }, + { + "epoch": 1.8408976840897684, + "grad_norm": 0.05234100669622421, + "learning_rate": 7.963016237360465e-06, + "loss": 0.0113, + "step": 3210 + }, + { + "epoch": 1.8466336846633684, + "grad_norm": 0.7740063071250916, + "learning_rate": 7.946860063165238e-06, + "loss": 0.0115, + "step": 3220 + }, + { + "epoch": 1.8523696852369684, + "grad_norm": 0.007312777452170849, + "learning_rate": 7.93065660714321e-06, + "loss": 0.0378, + "step": 3230 + }, + { + "epoch": 1.8581056858105685, + "grad_norm": 0.014250587671995163, + "learning_rate": 7.914406129275847e-06, + "loss": 0.0188, + "step": 3240 + }, + { + "epoch": 1.8638416863841685, + "grad_norm": 0.0380655862390995, + "learning_rate": 7.898108890299072e-06, + "loss": 0.013, + "step": 3250 + }, + { + "epoch": 1.8638416863841685, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07671073824167252, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.182, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.828, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.438, + "step": 3250 + }, + { + "epoch": 1.8695776869577687, + "grad_norm": 0.034287262707948685, + "learning_rate": 7.881765151699085e-06, + "loss": 0.0237, + "step": 3260 + }, + { + "epoch": 1.8753136875313687, + "grad_norm": 0.014141724444925785, + "learning_rate": 7.865375175708158e-06, + "loss": 0.0294, + "step": 3270 + }, + { + "epoch": 1.881049688104969, + "grad_norm": 2.08565616607666, + "learning_rate": 7.848939225300436e-06, + "loss": 0.0229, + "step": 3280 + }, + { + "epoch": 1.886785688678569, + "grad_norm": 1.476799726486206, + "learning_rate": 7.832457564187715e-06, + "loss": 0.0257, + "step": 3290 + }, + { + "epoch": 1.892521689252169, + "grad_norm": 1.641891598701477, + "learning_rate": 7.815930456815212e-06, + "loss": 0.0171, + "step": 3300 + }, + { + "epoch": 1.892521689252169, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06915121525526047, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.0913, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.961, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.504, + "step": 3300 + }, + { + "epoch": 1.898257689825769, + "grad_norm": 0.03908121585845947, + "learning_rate": 7.799358168357323e-06, + "loss": 0.0208, + "step": 3310 + }, + { + "epoch": 1.903993690399369, + "grad_norm": 0.27976372838020325, + "learning_rate": 7.782740964713358e-06, + "loss": 0.0079, + "step": 3320 + }, + { + "epoch": 1.909729690972969, + "grad_norm": 0.1647738814353943, + "learning_rate": 7.76607911250329e-06, + "loss": 0.0284, + "step": 3330 + }, + { + "epoch": 1.915465691546569, + "grad_norm": 0.009193528443574905, + "learning_rate": 7.749372879063468e-06, + "loss": 0.0219, + "step": 3340 + }, + { + "epoch": 1.9212016921201691, + "grad_norm": 0.08929289132356644, + "learning_rate": 7.73262253244233e-06, + "loss": 0.0136, + "step": 3350 + }, + { + "epoch": 1.9212016921201691, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06728994846343994, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1919, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.814, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.43, + "step": 3350 + }, + { + "epoch": 1.9269376926937691, + "grad_norm": 1.6335256099700928, + "learning_rate": 7.715828341396102e-06, + "loss": 0.0322, + "step": 3360 + }, + { + "epoch": 1.9326736932673694, + "grad_norm": 0.2756313979625702, + "learning_rate": 7.698990575384483e-06, + "loss": 0.0282, + "step": 3370 + }, + { + "epoch": 1.9384096938409694, + "grad_norm": 0.5282019376754761, + "learning_rate": 7.68210950456633e-06, + "loss": 0.0059, + "step": 3380 + }, + { + "epoch": 1.9441456944145694, + "grad_norm": 0.013134743086993694, + "learning_rate": 7.66518539979531e-06, + "loss": 0.0088, + "step": 3390 + }, + { + "epoch": 1.9498816949881697, + "grad_norm": 0.020495153963565826, + "learning_rate": 7.648218532615572e-06, + "loss": 0.0275, + "step": 3400 + }, + { + "epoch": 1.9498816949881697, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06392066925764084, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1842, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.825, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.436, + "step": 3400 + }, + { + "epoch": 1.9556176955617697, + "grad_norm": 0.030520539730787277, + "learning_rate": 7.631209175257368e-06, + "loss": 0.0176, + "step": 3410 + }, + { + "epoch": 1.9613536961353697, + "grad_norm": 0.356724351644516, + "learning_rate": 7.614157600632706e-06, + "loss": 0.0306, + "step": 3420 + }, + { + "epoch": 1.9670896967089697, + "grad_norm": 0.5876423716545105, + "learning_rate": 7.597064082330961e-06, + "loss": 0.0178, + "step": 3430 + }, + { + "epoch": 1.9728256972825697, + "grad_norm": 0.04390386864542961, + "learning_rate": 7.579928894614479e-06, + "loss": 0.0099, + "step": 3440 + }, + { + "epoch": 1.9785616978561698, + "grad_norm": 0.026595573872327805, + "learning_rate": 7.562752312414196e-06, + "loss": 0.0097, + "step": 3450 + }, + { + "epoch": 1.9785616978561698, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08633749186992645, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1407, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.888, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.468, + "step": 3450 + }, + { + "epoch": 1.9842976984297698, + "grad_norm": 0.4078025221824646, + "learning_rate": 7.545534611325207e-06, + "loss": 0.0189, + "step": 3460 + }, + { + "epoch": 1.9900336990033698, + "grad_norm": 0.4737612307071686, + "learning_rate": 7.528276067602352e-06, + "loss": 0.0019, + "step": 3470 + }, + { + "epoch": 1.9957696995769698, + "grad_norm": 0.0031162798404693604, + "learning_rate": 7.510976958155789e-06, + "loss": 0.0037, + "step": 3480 + }, + { + "epoch": 2.00114720011472, + "grad_norm": 1.097615361213684, + "learning_rate": 7.493637560546539e-06, + "loss": 0.0127, + "step": 3490 + }, + { + "epoch": 2.00688320068832, + "grad_norm": 0.005019227508455515, + "learning_rate": 7.476258152982043e-06, + "loss": 0.0019, + "step": 3500 + }, + { + "epoch": 2.00688320068832, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07848720997571945, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.177, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.835, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.441, + "step": 3500 + }, + { + "epoch": 2.01261920126192, + "grad_norm": 0.016102029010653496, + "learning_rate": 7.458839014311696e-06, + "loss": 0.0057, + "step": 3510 + }, + { + "epoch": 2.01835520183552, + "grad_norm": 0.009885331615805626, + "learning_rate": 7.441380424022364e-06, + "loss": 0.0089, + "step": 3520 + }, + { + "epoch": 2.0240912024091204, + "grad_norm": 2.9838740825653076, + "learning_rate": 7.423882662233915e-06, + "loss": 0.0142, + "step": 3530 + }, + { + "epoch": 2.0298272029827205, + "grad_norm": 0.29027923941612244, + "learning_rate": 7.406346009694713e-06, + "loss": 0.0033, + "step": 3540 + }, + { + "epoch": 2.0355632035563205, + "grad_norm": 0.004921222571283579, + "learning_rate": 7.388770747777116e-06, + "loss": 0.0108, + "step": 3550 + }, + { + "epoch": 2.0355632035563205, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0910039097070694, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.144, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.883, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.465, + "step": 3550 + }, + { + "epoch": 2.0412992041299205, + "grad_norm": 0.012345495633780956, + "learning_rate": 7.371157158472965e-06, + "loss": 0.0186, + "step": 3560 + }, + { + "epoch": 2.0470352047035205, + "grad_norm": 0.013729539699852467, + "learning_rate": 7.353505524389052e-06, + "loss": 0.0038, + "step": 3570 + }, + { + "epoch": 2.0527712052771205, + "grad_norm": 3.2665841579437256, + "learning_rate": 7.335816128742599e-06, + "loss": 0.0203, + "step": 3580 + }, + { + "epoch": 2.0585072058507206, + "grad_norm": 0.0018614591099321842, + "learning_rate": 7.318089255356695e-06, + "loss": 0.0093, + "step": 3590 + }, + { + "epoch": 2.0642432064243206, + "grad_norm": 0.10742621123790741, + "learning_rate": 7.300325188655762e-06, + "loss": 0.0093, + "step": 3600 + }, + { + "epoch": 2.0642432064243206, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07848691940307617, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1618, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.858, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.452, + "step": 3600 + }, + { + "epoch": 2.0699792069979206, + "grad_norm": 0.6563161611557007, + "learning_rate": 7.282524213660974e-06, + "loss": 0.001, + "step": 3610 + }, + { + "epoch": 2.0757152075715206, + "grad_norm": 0.12672875821590424, + "learning_rate": 7.264686615985697e-06, + "loss": 0.0008, + "step": 3620 + }, + { + "epoch": 2.0814512081451206, + "grad_norm": 0.004448240157216787, + "learning_rate": 7.246812681830899e-06, + "loss": 0.0139, + "step": 3630 + }, + { + "epoch": 2.0871872087187207, + "grad_norm": 0.007188173942267895, + "learning_rate": 7.228902697980562e-06, + "loss": 0.0106, + "step": 3640 + }, + { + "epoch": 2.092923209292321, + "grad_norm": 0.006886759772896767, + "learning_rate": 7.210956951797074e-06, + "loss": 0.0076, + "step": 3650 + }, + { + "epoch": 2.092923209292321, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08482418954372406, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1827, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.827, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.437, + "step": 3650 + }, + { + "epoch": 2.098659209865921, + "grad_norm": 3.8101847171783447, + "learning_rate": 7.19297573121663e-06, + "loss": 0.0089, + "step": 3660 + }, + { + "epoch": 2.104395210439521, + "grad_norm": 0.002013096585869789, + "learning_rate": 7.174959324744599e-06, + "loss": 0.0123, + "step": 3670 + }, + { + "epoch": 2.110131211013121, + "grad_norm": 3.1393046379089355, + "learning_rate": 7.156908021450904e-06, + "loss": 0.01, + "step": 3680 + }, + { + "epoch": 2.115867211586721, + "grad_norm": 0.07943509519100189, + "learning_rate": 7.138822110965381e-06, + "loss": 0.0016, + "step": 3690 + }, + { + "epoch": 2.121603212160321, + "grad_norm": 2.5405237674713135, + "learning_rate": 7.120701883473131e-06, + "loss": 0.0148, + "step": 3700 + }, + { + "epoch": 2.121603212160321, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10251594334840775, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2361, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.75, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.398, + "step": 3700 + }, + { + "epoch": 2.1273392127339212, + "grad_norm": 0.006140326615422964, + "learning_rate": 7.102547629709867e-06, + "loss": 0.0163, + "step": 3710 + }, + { + "epoch": 2.1330752133075213, + "grad_norm": 0.04222610965371132, + "learning_rate": 7.084359640957246e-06, + "loss": 0.017, + "step": 3720 + }, + { + "epoch": 2.1388112138811213, + "grad_norm": 0.03461828827857971, + "learning_rate": 7.066138209038194e-06, + "loss": 0.0162, + "step": 3730 + }, + { + "epoch": 2.1445472144547213, + "grad_norm": 0.9535601139068604, + "learning_rate": 7.047883626312233e-06, + "loss": 0.0266, + "step": 3740 + }, + { + "epoch": 2.1502832150283213, + "grad_norm": 0.8345515131950378, + "learning_rate": 7.029596185670778e-06, + "loss": 0.0446, + "step": 3750 + }, + { + "epoch": 2.1502832150283213, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07697154581546783, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1863, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.822, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.435, + "step": 3750 + }, + { + "epoch": 2.156019215601922, + "grad_norm": 0.14290325343608856, + "learning_rate": 7.011276180532445e-06, + "loss": 0.0218, + "step": 3760 + }, + { + "epoch": 2.161755216175522, + "grad_norm": 0.05109575018286705, + "learning_rate": 6.992923904838341e-06, + "loss": 0.0032, + "step": 3770 + }, + { + "epoch": 2.167491216749122, + "grad_norm": 3.4320428371429443, + "learning_rate": 6.974539653047346e-06, + "loss": 0.0107, + "step": 3780 + }, + { + "epoch": 2.173227217322722, + "grad_norm": 0.9100053310394287, + "learning_rate": 6.956123720131398e-06, + "loss": 0.0162, + "step": 3790 + }, + { + "epoch": 2.178963217896322, + "grad_norm": 0.09293865412473679, + "learning_rate": 6.937676401570744e-06, + "loss": 0.0175, + "step": 3800 + }, + { + "epoch": 2.178963217896322, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07296475023031235, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.156, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.866, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.457, + "step": 3800 + }, + { + "epoch": 2.184699218469922, + "grad_norm": 0.25332963466644287, + "learning_rate": 6.9191979933492135e-06, + "loss": 0.0107, + "step": 3810 + }, + { + "epoch": 2.190435219043522, + "grad_norm": 0.47701334953308105, + "learning_rate": 6.900688791949463e-06, + "loss": 0.0087, + "step": 3820 + }, + { + "epoch": 2.196171219617122, + "grad_norm": 0.019721075892448425, + "learning_rate": 6.882149094348215e-06, + "loss": 0.0178, + "step": 3830 + }, + { + "epoch": 2.201907220190722, + "grad_norm": 0.057930897921323776, + "learning_rate": 6.863579198011506e-06, + "loss": 0.0291, + "step": 3840 + }, + { + "epoch": 2.207643220764322, + "grad_norm": 0.03594127669930458, + "learning_rate": 6.8449794008899e-06, + "loss": 0.005, + "step": 3850 + }, + { + "epoch": 2.207643220764322, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07210449874401093, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1052, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.94, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.494, + "step": 3850 + }, + { + "epoch": 2.213379221337922, + "grad_norm": 2.532087802886963, + "learning_rate": 6.826350001413713e-06, + "loss": 0.0085, + "step": 3860 + }, + { + "epoch": 2.219115221911522, + "grad_norm": 0.3755287230014801, + "learning_rate": 6.807691298488231e-06, + "loss": 0.0014, + "step": 3870 + }, + { + "epoch": 2.224851222485122, + "grad_norm": 0.09285666048526764, + "learning_rate": 6.789003591488902e-06, + "loss": 0.0013, + "step": 3880 + }, + { + "epoch": 2.2305872230587225, + "grad_norm": 3.115718126296997, + "learning_rate": 6.770287180256545e-06, + "loss": 0.0045, + "step": 3890 + }, + { + "epoch": 2.2363232236323225, + "grad_norm": 1.7172398567199707, + "learning_rate": 6.751542365092527e-06, + "loss": 0.0085, + "step": 3900 + }, + { + "epoch": 2.2363232236323225, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08293686807155609, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1262, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.909, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.478, + "step": 3900 + }, + { + "epoch": 2.2420592242059225, + "grad_norm": 0.004544651135802269, + "learning_rate": 6.732769446753954e-06, + "loss": 0.021, + "step": 3910 + }, + { + "epoch": 2.2477952247795225, + "grad_norm": 3.599937677383423, + "learning_rate": 6.713968726448844e-06, + "loss": 0.0067, + "step": 3920 + }, + { + "epoch": 2.2535312253531226, + "grad_norm": 0.9228914976119995, + "learning_rate": 6.69514050583129e-06, + "loss": 0.0097, + "step": 3930 + }, + { + "epoch": 2.2592672259267226, + "grad_norm": 0.005351903382688761, + "learning_rate": 6.676285086996623e-06, + "loss": 0.0353, + "step": 3940 + }, + { + "epoch": 2.2650032265003226, + "grad_norm": 0.07884957641363144, + "learning_rate": 6.657402772476563e-06, + "loss": 0.0092, + "step": 3950 + }, + { + "epoch": 2.2650032265003226, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09563428908586502, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.7675, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 29.999, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.022, + "step": 3950 + }, + { + "epoch": 2.2707392270739226, + "grad_norm": 1.1882566213607788, + "learning_rate": 6.638493865234369e-06, + "loss": 0.0132, + "step": 3960 + }, + { + "epoch": 2.2764752276475226, + "grad_norm": 0.13506627082824707, + "learning_rate": 6.619558668659975e-06, + "loss": 0.0219, + "step": 3970 + }, + { + "epoch": 2.2822112282211227, + "grad_norm": 0.007426030468195677, + "learning_rate": 6.600597486565119e-06, + "loss": 0.0216, + "step": 3980 + }, + { + "epoch": 2.2879472287947227, + "grad_norm": 0.27120983600616455, + "learning_rate": 6.581610623178476e-06, + "loss": 0.0153, + "step": 3990 + }, + { + "epoch": 2.293683229368323, + "grad_norm": 1.4845871925354004, + "learning_rate": 6.562598383140773e-06, + "loss": 0.0059, + "step": 4000 + }, + { + "epoch": 2.293683229368323, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07171830534934998, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2153, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.78, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.413, + "step": 4000 + }, + { + "epoch": 2.299419229941923, + "grad_norm": 1.7247122526168823, + "learning_rate": 6.543561071499893e-06, + "loss": 0.0216, + "step": 4010 + }, + { + "epoch": 2.305155230515523, + "grad_norm": 0.02364450879395008, + "learning_rate": 6.524498993706e-06, + "loss": 0.0244, + "step": 4020 + }, + { + "epoch": 2.310891231089123, + "grad_norm": 0.048632461577653885, + "learning_rate": 6.505412455606617e-06, + "loss": 0.0029, + "step": 4030 + }, + { + "epoch": 2.316627231662723, + "grad_norm": 0.12222401797771454, + "learning_rate": 6.486301763441732e-06, + "loss": 0.0027, + "step": 4040 + }, + { + "epoch": 2.3223632322363232, + "grad_norm": 0.07429647445678711, + "learning_rate": 6.467167223838879e-06, + "loss": 0.0119, + "step": 4050 + }, + { + "epoch": 2.3223632322363232, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09189260751008987, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1514, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.873, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.46, + "step": 4050 + }, + { + "epoch": 2.3280992328099233, + "grad_norm": 0.10564873367547989, + "learning_rate": 6.4480091438082225e-06, + "loss": 0.0328, + "step": 4060 + }, + { + "epoch": 2.3338352333835233, + "grad_norm": 0.01708606630563736, + "learning_rate": 6.4288278307376265e-06, + "loss": 0.015, + "step": 4070 + }, + { + "epoch": 2.3395712339571233, + "grad_norm": 0.25011512637138367, + "learning_rate": 6.4096235923877256e-06, + "loss": 0.0061, + "step": 4080 + }, + { + "epoch": 2.3453072345307233, + "grad_norm": 0.055370282381772995, + "learning_rate": 6.390396736886986e-06, + "loss": 0.007, + "step": 4090 + }, + { + "epoch": 2.3510432351043233, + "grad_norm": 0.030733373016119003, + "learning_rate": 6.371147572726761e-06, + "loss": 0.0017, + "step": 4100 + }, + { + "epoch": 2.3510432351043233, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08482550084590912, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1309, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.903, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.475, + "step": 4100 + }, + { + "epoch": 2.356779235677924, + "grad_norm": 0.17796050012111664, + "learning_rate": 6.351876408756344e-06, + "loss": 0.0053, + "step": 4110 + }, + { + "epoch": 2.3625152362515234, + "grad_norm": 0.0136833805590868, + "learning_rate": 6.332583554178009e-06, + "loss": 0.0044, + "step": 4120 + }, + { + "epoch": 2.368251236825124, + "grad_norm": 4.209476470947266, + "learning_rate": 6.313269318542057e-06, + "loss": 0.0259, + "step": 4130 + }, + { + "epoch": 2.373987237398724, + "grad_norm": 3.9427716732025146, + "learning_rate": 6.2939340117418355e-06, + "loss": 0.0237, + "step": 4140 + }, + { + "epoch": 2.379723237972324, + "grad_norm": 0.008540840819478035, + "learning_rate": 6.274577944008785e-06, + "loss": 0.0044, + "step": 4150 + }, + { + "epoch": 2.379723237972324, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07825496047735214, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1875, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.82, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, + "step": 4150 + }, + { + "epoch": 2.385459238545924, + "grad_norm": 0.01801919937133789, + "learning_rate": 6.255201425907442e-06, + "loss": 0.0002, + "step": 4160 + }, + { + "epoch": 2.391195239119524, + "grad_norm": 0.7031999230384827, + "learning_rate": 6.235804768330472e-06, + "loss": 0.0198, + "step": 4170 + }, + { + "epoch": 2.396931239693124, + "grad_norm": 0.03624948486685753, + "learning_rate": 6.21638828249367e-06, + "loss": 0.0096, + "step": 4180 + }, + { + "epoch": 2.402667240266724, + "grad_norm": 0.010821559466421604, + "learning_rate": 6.196952279930977e-06, + "loss": 0.0089, + "step": 4190 + }, + { + "epoch": 2.408403240840324, + "grad_norm": 0.01340206153690815, + "learning_rate": 6.177497072489473e-06, + "loss": 0.0227, + "step": 4200 + }, + { + "epoch": 2.408403240840324, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07348810136318207, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1832, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.826, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.437, + "step": 4200 + }, + { + "epoch": 2.414139241413924, + "grad_norm": 0.006387198343873024, + "learning_rate": 6.158022972324375e-06, + "loss": 0.0112, + "step": 4210 + }, + { + "epoch": 2.419875241987524, + "grad_norm": 0.1088046282529831, + "learning_rate": 6.138530291894033e-06, + "loss": 0.015, + "step": 4220 + }, + { + "epoch": 2.425611242561124, + "grad_norm": 0.010532280430197716, + "learning_rate": 6.119019343954914e-06, + "loss": 0.0012, + "step": 4230 + }, + { + "epoch": 2.4313472431347245, + "grad_norm": 2.5161547660827637, + "learning_rate": 6.099490441556582e-06, + "loss": 0.0054, + "step": 4240 + }, + { + "epoch": 2.4370832437083245, + "grad_norm": 0.41214823722839355, + "learning_rate": 6.07994389803668e-06, + "loss": 0.0067, + "step": 4250 + }, + { + "epoch": 2.4370832437083245, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08171182125806808, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1867, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.821, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, + "step": 4250 + }, + { + "epoch": 2.4428192442819245, + "grad_norm": 3.9505116939544678, + "learning_rate": 6.060380027015897e-06, + "loss": 0.0205, + "step": 4260 + }, + { + "epoch": 2.4485552448555246, + "grad_norm": 2.172741651535034, + "learning_rate": 6.040799142392937e-06, + "loss": 0.015, + "step": 4270 + }, + { + "epoch": 2.4542912454291246, + "grad_norm": 0.003915698733180761, + "learning_rate": 6.02120155833949e-06, + "loss": 0.0047, + "step": 4280 + }, + { + "epoch": 2.4600272460027246, + "grad_norm": 0.027446260675787926, + "learning_rate": 6.001587589295179e-06, + "loss": 0.0172, + "step": 4290 + }, + { + "epoch": 2.4657632465763246, + "grad_norm": 0.9420067667961121, + "learning_rate": 5.981957549962524e-06, + "loss": 0.0132, + "step": 4300 + }, + { + "epoch": 2.4657632465763246, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06599417328834534, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2262, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.764, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.405, + "step": 4300 + }, + { + "epoch": 2.4714992471499246, + "grad_norm": 2.6476902961730957, + "learning_rate": 5.96231175530189e-06, + "loss": 0.01, + "step": 4310 + }, + { + "epoch": 2.4772352477235247, + "grad_norm": 0.03741572052240372, + "learning_rate": 5.942650520526432e-06, + "loss": 0.0102, + "step": 4320 + }, + { + "epoch": 2.4829712482971247, + "grad_norm": 0.012345872819423676, + "learning_rate": 5.9229741610970425e-06, + "loss": 0.0154, + "step": 4330 + }, + { + "epoch": 2.4887072488707247, + "grad_norm": 0.006493984255939722, + "learning_rate": 5.903282992717281e-06, + "loss": 0.0063, + "step": 4340 + }, + { + "epoch": 2.494443249444325, + "grad_norm": 0.011486309580504894, + "learning_rate": 5.883577331328316e-06, + "loss": 0.0005, + "step": 4350 + }, + { + "epoch": 2.494443249444325, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0845269188284874, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1186, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.921, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.484, + "step": 4350 + }, + { + "epoch": 2.5001792500179247, + "grad_norm": 0.05512839928269386, + "learning_rate": 5.863857493103855e-06, + "loss": 0.0249, + "step": 4360 + }, + { + "epoch": 2.505915250591525, + "grad_norm": 0.010474382899701595, + "learning_rate": 5.84412379444507e-06, + "loss": 0.0007, + "step": 4370 + }, + { + "epoch": 2.511651251165125, + "grad_norm": 0.8734518885612488, + "learning_rate": 5.824376551975519e-06, + "loss": 0.0081, + "step": 4380 + }, + { + "epoch": 2.5173872517387252, + "grad_norm": 0.11771434545516968, + "learning_rate": 5.804616082536071e-06, + "loss": 0.0074, + "step": 4390 + }, + { + "epoch": 2.5231232523123253, + "grad_norm": 1.922683835029602, + "learning_rate": 5.784842703179814e-06, + "loss": 0.0328, + "step": 4400 + }, + { + "epoch": 2.5231232523123253, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07923904061317444, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.0857, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.969, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.508, + "step": 4400 + }, + { + "epoch": 2.5288592528859253, + "grad_norm": 4.321990013122559, + "learning_rate": 5.765056731166982e-06, + "loss": 0.0237, + "step": 4410 + }, + { + "epoch": 2.5345952534595253, + "grad_norm": 0.01736987754702568, + "learning_rate": 5.745258483959847e-06, + "loss": 0.0064, + "step": 4420 + }, + { + "epoch": 2.5403312540331253, + "grad_norm": 0.07172615826129913, + "learning_rate": 5.725448279217638e-06, + "loss": 0.0118, + "step": 4430 + }, + { + "epoch": 2.5460672546067253, + "grad_norm": 0.1578182578086853, + "learning_rate": 5.705626434791444e-06, + "loss": 0.0016, + "step": 4440 + }, + { + "epoch": 2.5518032551803254, + "grad_norm": 0.005266325548291206, + "learning_rate": 5.6857932687191e-06, + "loss": 0.0023, + "step": 4450 + }, + { + "epoch": 2.5518032551803254, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07244950532913208, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2865, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.677, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, + "step": 4450 + }, + { + "epoch": 2.557539255753926, + "grad_norm": 0.21443744003772736, + "learning_rate": 5.66594909922011e-06, + "loss": 0.0114, + "step": 4460 + }, + { + "epoch": 2.5632752563275254, + "grad_norm": 0.08916337788105011, + "learning_rate": 5.646094244690511e-06, + "loss": 0.0064, + "step": 4470 + }, + { + "epoch": 2.569011256901126, + "grad_norm": 2.983360528945923, + "learning_rate": 5.626229023697789e-06, + "loss": 0.0159, + "step": 4480 + }, + { + "epoch": 2.574747257474726, + "grad_norm": 0.06603588908910751, + "learning_rate": 5.606353754975755e-06, + "loss": 0.0104, + "step": 4490 + }, + { + "epoch": 2.580483258048326, + "grad_norm": 0.9153593182563782, + "learning_rate": 5.586468757419433e-06, + "loss": 0.0012, + "step": 4500 + }, + { + "epoch": 2.580483258048326, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07681399583816528, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3831, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.538, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.292, + "step": 4500 + }, + { + "epoch": 2.586219258621926, + "grad_norm": 0.030913453549146652, + "learning_rate": 5.566574350079946e-06, + "loss": 0.0139, + "step": 4510 + }, + { + "epoch": 2.591955259195526, + "grad_norm": 3.312136650085449, + "learning_rate": 5.546670852159396e-06, + "loss": 0.0074, + "step": 4520 + }, + { + "epoch": 2.597691259769126, + "grad_norm": 0.008490420877933502, + "learning_rate": 5.526758583005736e-06, + "loss": 0.008, + "step": 4530 + }, + { + "epoch": 2.603427260342726, + "grad_norm": 0.010648602619767189, + "learning_rate": 5.50683786210766e-06, + "loss": 0.0051, + "step": 4540 + }, + { + "epoch": 2.609163260916326, + "grad_norm": 2.533731460571289, + "learning_rate": 5.486909009089463e-06, + "loss": 0.0185, + "step": 4550 + }, + { + "epoch": 2.609163260916326, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07284829020500183, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1938, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.811, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.429, + "step": 4550 + }, + { + "epoch": 2.614899261489926, + "grad_norm": 0.0175775233656168, + "learning_rate": 5.4669723437059194e-06, + "loss": 0.0015, + "step": 4560 + }, + { + "epoch": 2.6206352620635265, + "grad_norm": 0.00764029985293746, + "learning_rate": 5.4470281858371555e-06, + "loss": 0.0014, + "step": 4570 + }, + { + "epoch": 2.626371262637126, + "grad_norm": 0.4708549380302429, + "learning_rate": 5.4270768554835056e-06, + "loss": 0.0023, + "step": 4580 + }, + { + "epoch": 2.6321072632107265, + "grad_norm": 0.06411808729171753, + "learning_rate": 5.407118672760393e-06, + "loss": 0.0037, + "step": 4590 + }, + { + "epoch": 2.6378432637843265, + "grad_norm": 0.009167805314064026, + "learning_rate": 5.387153957893181e-06, + "loss": 0.0016, + "step": 4600 + }, + { + "epoch": 2.6378432637843265, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09603765606880188, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2377, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.747, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.397, + "step": 4600 + }, + { + "epoch": 2.6435792643579266, + "grad_norm": 0.0018381074769422412, + "learning_rate": 5.367183031212041e-06, + "loss": 0.007, + "step": 4610 + }, + { + "epoch": 2.6493152649315266, + "grad_norm": 0.0011304231593385339, + "learning_rate": 5.347206213146813e-06, + "loss": 0.0277, + "step": 4620 + }, + { + "epoch": 2.6550512655051266, + "grad_norm": 4.376464366912842, + "learning_rate": 5.327223824221862e-06, + "loss": 0.0093, + "step": 4630 + }, + { + "epoch": 2.6607872660787266, + "grad_norm": 0.031111005693674088, + "learning_rate": 5.307236185050937e-06, + "loss": 0.0029, + "step": 4640 + }, + { + "epoch": 2.6665232666523266, + "grad_norm": 2.6803221702575684, + "learning_rate": 5.287243616332027e-06, + "loss": 0.0145, + "step": 4650 + }, + { + "epoch": 2.6665232666523266, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10255688428878784, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2291, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.76, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.403, + "step": 4650 + }, + { + "epoch": 2.6722592672259267, + "grad_norm": 0.03457150608301163, + "learning_rate": 5.267246438842213e-06, + "loss": 0.005, + "step": 4660 + }, + { + "epoch": 2.6779952677995267, + "grad_norm": 4.120417594909668, + "learning_rate": 5.247244973432524e-06, + "loss": 0.0141, + "step": 4670 + }, + { + "epoch": 2.6837312683731267, + "grad_norm": 0.005652038846164942, + "learning_rate": 5.227239541022786e-06, + "loss": 0.0034, + "step": 4680 + }, + { + "epoch": 2.6894672689467267, + "grad_norm": 2.9149153232574463, + "learning_rate": 5.2072304625964785e-06, + "loss": 0.0061, + "step": 4690 + }, + { + "epoch": 2.695203269520327, + "grad_norm": 0.0008329672855325043, + "learning_rate": 5.187218059195578e-06, + "loss": 0.0047, + "step": 4700 + }, + { + "epoch": 2.695203269520327, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.10345083475112915, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2387, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.746, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.396, + "step": 4700 + }, + { + "epoch": 2.7009392700939268, + "grad_norm": 1.024248480796814, + "learning_rate": 5.167202651915409e-06, + "loss": 0.0071, + "step": 4710 + }, + { + "epoch": 2.7066752706675272, + "grad_norm": 0.014447176828980446, + "learning_rate": 5.147184561899495e-06, + "loss": 0.0076, + "step": 4720 + }, + { + "epoch": 2.7124112712411272, + "grad_norm": 5.958902835845947, + "learning_rate": 5.1271641103344e-06, + "loss": 0.0116, + "step": 4730 + }, + { + "epoch": 2.7181472718147273, + "grad_norm": 0.016329055652022362, + "learning_rate": 5.1071416184445845e-06, + "loss": 0.0105, + "step": 4740 + }, + { + "epoch": 2.7238832723883273, + "grad_norm": 0.4175659716129303, + "learning_rate": 5.087117407487239e-06, + "loss": 0.0166, + "step": 4750 + }, + { + "epoch": 2.7238832723883273, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08455558121204376, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2591, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.716, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.382, + "step": 4750 + }, + { + "epoch": 2.7296192729619273, + "grad_norm": 0.022596385329961777, + "learning_rate": 5.06709179874714e-06, + "loss": 0.0009, + "step": 4760 + }, + { + "epoch": 2.7353552735355273, + "grad_norm": 0.0726088285446167, + "learning_rate": 5.047065113531493e-06, + "loss": 0.0254, + "step": 4770 + }, + { + "epoch": 2.7410912741091273, + "grad_norm": 0.035875000059604645, + "learning_rate": 5.02703767316477e-06, + "loss": 0.0244, + "step": 4780 + }, + { + "epoch": 2.7468272746827274, + "grad_norm": 3.32077956199646, + "learning_rate": 5.00700979898356e-06, + "loss": 0.006, + "step": 4790 + }, + { + "epoch": 2.7525632752563274, + "grad_norm": 1.0301790237426758, + "learning_rate": 4.98698181233142e-06, + "loss": 0.0059, + "step": 4800 + }, + { + "epoch": 2.7525632752563274, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07655075937509537, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2238, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.767, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.407, + "step": 4800 + }, + { + "epoch": 2.758299275829928, + "grad_norm": 0.446961373090744, + "learning_rate": 4.966954034553699e-06, + "loss": 0.0087, + "step": 4810 + }, + { + "epoch": 2.7640352764035274, + "grad_norm": 0.33151447772979736, + "learning_rate": 4.946926786992407e-06, + "loss": 0.0009, + "step": 4820 + }, + { + "epoch": 2.769771276977128, + "grad_norm": 0.12154054641723633, + "learning_rate": 4.9269003909810405e-06, + "loss": 0.0055, + "step": 4830 + }, + { + "epoch": 2.775507277550728, + "grad_norm": 0.016543418169021606, + "learning_rate": 4.906875167839433e-06, + "loss": 0.0373, + "step": 4840 + }, + { + "epoch": 2.781243278124328, + "grad_norm": 0.018002478405833244, + "learning_rate": 4.886851438868599e-06, + "loss": 0.0133, + "step": 4850 + }, + { + "epoch": 2.781243278124328, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07717321813106537, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.202, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.799, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.423, + "step": 4850 + }, + { + "epoch": 2.786979278697928, + "grad_norm": 0.01707332953810692, + "learning_rate": 4.866829525345585e-06, + "loss": 0.0166, + "step": 4860 + }, + { + "epoch": 2.792715279271528, + "grad_norm": 0.00804219115525484, + "learning_rate": 4.846809748518302e-06, + "loss": 0.0015, + "step": 4870 + }, + { + "epoch": 2.798451279845128, + "grad_norm": 0.02389547787606716, + "learning_rate": 4.826792429600381e-06, + "loss": 0.0021, + "step": 4880 + }, + { + "epoch": 2.804187280418728, + "grad_norm": 0.005223631393164396, + "learning_rate": 4.806777889766016e-06, + "loss": 0.0136, + "step": 4890 + }, + { + "epoch": 2.809923280992328, + "grad_norm": 0.20156431198120117, + "learning_rate": 4.78676645014481e-06, + "loss": 0.0093, + "step": 4900 + }, + { + "epoch": 2.809923280992328, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0736735612154007, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.151, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.873, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.46, + "step": 4900 + }, + { + "epoch": 2.815659281565928, + "grad_norm": 0.0024203194770962, + "learning_rate": 4.766758431816629e-06, + "loss": 0.0052, + "step": 4910 + }, + { + "epoch": 2.821395282139528, + "grad_norm": 0.031604766845703125, + "learning_rate": 4.746754155806437e-06, + "loss": 0.0095, + "step": 4920 + }, + { + "epoch": 2.827131282713128, + "grad_norm": 3.787029266357422, + "learning_rate": 4.72675394307916e-06, + "loss": 0.0139, + "step": 4930 + }, + { + "epoch": 2.8328672832867285, + "grad_norm": 0.07122969627380371, + "learning_rate": 4.7067581145345226e-06, + "loss": 0.0139, + "step": 4940 + }, + { + "epoch": 2.838603283860328, + "grad_norm": 3.0758485794067383, + "learning_rate": 4.686766991001913e-06, + "loss": 0.0161, + "step": 4950 + }, + { + "epoch": 2.838603283860328, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08140452206134796, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.151, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.873, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.46, + "step": 4950 + }, + { + "epoch": 2.8443392844339286, + "grad_norm": 0.07921534031629562, + "learning_rate": 4.666780893235227e-06, + "loss": 0.0006, + "step": 4960 + }, + { + "epoch": 2.8500752850075286, + "grad_norm": 0.12975841760635376, + "learning_rate": 4.646800141907717e-06, + "loss": 0.0154, + "step": 4970 + }, + { + "epoch": 2.8558112855811286, + "grad_norm": 0.0262366384267807, + "learning_rate": 4.626825057606859e-06, + "loss": 0.0093, + "step": 4980 + }, + { + "epoch": 2.8615472861547286, + "grad_norm": 0.00262279505841434, + "learning_rate": 4.606855960829199e-06, + "loss": 0.0202, + "step": 4990 + }, + { + "epoch": 2.8672832867283287, + "grad_norm": 0.01955203339457512, + "learning_rate": 4.586893171975218e-06, + "loss": 0.0266, + "step": 5000 + }, + { + "epoch": 2.8672832867283287, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07921645790338516, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3258, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.62, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.334, + "step": 5000 + }, + { + "epoch": 2.8730192873019287, + "grad_norm": 0.10377766937017441, + "learning_rate": 4.566937011344184e-06, + "loss": 0.0017, + "step": 5010 + }, + { + "epoch": 2.8787552878755287, + "grad_norm": 0.784302294254303, + "learning_rate": 4.546987799129018e-06, + "loss": 0.0046, + "step": 5020 + }, + { + "epoch": 2.8844912884491287, + "grad_norm": 2.035209894180298, + "learning_rate": 4.527045855411153e-06, + "loss": 0.0119, + "step": 5030 + }, + { + "epoch": 2.8902272890227287, + "grad_norm": 3.6558942794799805, + "learning_rate": 4.507111500155407e-06, + "loss": 0.0123, + "step": 5040 + }, + { + "epoch": 2.895963289596329, + "grad_norm": 3.73490309715271, + "learning_rate": 4.487185053204832e-06, + "loss": 0.0073, + "step": 5050 + }, + { + "epoch": 2.895963289596329, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07410901039838791, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3293, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.615, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.331, + "step": 5050 + }, + { + "epoch": 2.901699290169929, + "grad_norm": 0.008407690562307835, + "learning_rate": 4.467266834275601e-06, + "loss": 0.0191, + "step": 5060 + }, + { + "epoch": 2.9074352907435292, + "grad_norm": 0.03255997970700264, + "learning_rate": 4.447357162951865e-06, + "loss": 0.0458, + "step": 5070 + }, + { + "epoch": 2.9131712913171293, + "grad_norm": 1.6957069635391235, + "learning_rate": 4.427456358680635e-06, + "loss": 0.0236, + "step": 5080 + }, + { + "epoch": 2.9189072918907293, + "grad_norm": 0.1675146073102951, + "learning_rate": 4.407564740766648e-06, + "loss": 0.0177, + "step": 5090 + }, + { + "epoch": 2.9246432924643293, + "grad_norm": 0.04229651764035225, + "learning_rate": 4.3876826283672485e-06, + "loss": 0.0116, + "step": 5100 + }, + { + "epoch": 2.9246432924643293, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06259483098983765, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.178, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.834, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.441, + "step": 5100 + }, + { + "epoch": 2.9303792930379293, + "grad_norm": 0.04005512595176697, + "learning_rate": 4.367810340487267e-06, + "loss": 0.0099, + "step": 5110 + }, + { + "epoch": 2.9361152936115293, + "grad_norm": 0.1425451785326004, + "learning_rate": 4.347948195973901e-06, + "loss": 0.0033, + "step": 5120 + }, + { + "epoch": 2.9418512941851294, + "grad_norm": 0.10147175192832947, + "learning_rate": 4.328096513511601e-06, + "loss": 0.0076, + "step": 5130 + }, + { + "epoch": 2.9475872947587294, + "grad_norm": 0.13407284021377563, + "learning_rate": 4.308255611616954e-06, + "loss": 0.0034, + "step": 5140 + }, + { + "epoch": 2.9533232953323294, + "grad_norm": 0.12246271222829819, + "learning_rate": 4.2884258086335755e-06, + "loss": 0.0077, + "step": 5150 + }, + { + "epoch": 2.9533232953323294, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07015232741832733, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2126, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.784, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.415, + "step": 5150 + }, + { + "epoch": 2.95905929590593, + "grad_norm": 0.06634827703237534, + "learning_rate": 4.268607422727e-06, + "loss": 0.0131, + "step": 5160 + }, + { + "epoch": 2.9647952964795294, + "grad_norm": 2.3556761741638184, + "learning_rate": 4.248800771879581e-06, + "loss": 0.029, + "step": 5170 + }, + { + "epoch": 2.97053129705313, + "grad_norm": 0.03471173346042633, + "learning_rate": 4.229006173885381e-06, + "loss": 0.0021, + "step": 5180 + }, + { + "epoch": 2.97626729762673, + "grad_norm": 0.006328817922621965, + "learning_rate": 4.2092239463450775e-06, + "loss": 0.0212, + "step": 5190 + }, + { + "epoch": 2.98200329820033, + "grad_norm": 0.02347049117088318, + "learning_rate": 4.189454406660865e-06, + "loss": 0.0056, + "step": 5200 + }, + { + "epoch": 2.98200329820033, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.06551730632781982, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.5411, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.314, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.18, + "step": 5200 + }, + { + "epoch": 2.98773929877393, + "grad_norm": 0.012669041752815247, + "learning_rate": 4.169697872031368e-06, + "loss": 0.012, + "step": 5210 + }, + { + "epoch": 2.99347529934753, + "grad_norm": 0.07477905601263046, + "learning_rate": 4.1499546594465465e-06, + "loss": 0.0215, + "step": 5220 + }, + { + "epoch": 2.99921129992113, + "grad_norm": 1.7680764198303223, + "learning_rate": 4.1302250856826045e-06, + "loss": 0.0107, + "step": 5230 + }, + { + "epoch": 3.00458880045888, + "grad_norm": 0.19178365170955658, + "learning_rate": 4.110509467296923e-06, + "loss": 0.006, + "step": 5240 + }, + { + "epoch": 3.01032480103248, + "grad_norm": 0.025907142087817192, + "learning_rate": 4.090808120622961e-06, + "loss": 0.0139, + "step": 5250 + }, + { + "epoch": 3.01032480103248, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07205013185739517, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2727, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.697, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.372, + "step": 5250 + }, + { + "epoch": 3.01606080160608, + "grad_norm": 0.018714534118771553, + "learning_rate": 4.071121361765201e-06, + "loss": 0.0033, + "step": 5260 + }, + { + "epoch": 3.02179680217968, + "grad_norm": 2.1166880130767822, + "learning_rate": 4.05144950659406e-06, + "loss": 0.0081, + "step": 5270 + }, + { + "epoch": 3.02753280275328, + "grad_norm": 0.07031437754631042, + "learning_rate": 4.031792870740831e-06, + "loss": 0.0086, + "step": 5280 + }, + { + "epoch": 3.03326880332688, + "grad_norm": 0.24249830842018127, + "learning_rate": 4.012151769592612e-06, + "loss": 0.0132, + "step": 5290 + }, + { + "epoch": 3.0390048039004802, + "grad_norm": 0.009602251462638378, + "learning_rate": 3.992526518287258e-06, + "loss": 0.0075, + "step": 5300 + }, + { + "epoch": 3.0390048039004802, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07945062220096588, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1773, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.835, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.441, + "step": 5300 + }, + { + "epoch": 3.0447408044740802, + "grad_norm": 0.1440117210149765, + "learning_rate": 3.972917431708305e-06, + "loss": 0.0013, + "step": 5310 + }, + { + "epoch": 3.0504768050476807, + "grad_norm": 0.013155910186469555, + "learning_rate": 3.9533248244799375e-06, + "loss": 0.0049, + "step": 5320 + }, + { + "epoch": 3.0562128056212807, + "grad_norm": 0.1568404585123062, + "learning_rate": 3.933749010961927e-06, + "loss": 0.0024, + "step": 5330 + }, + { + "epoch": 3.0619488061948807, + "grad_norm": 0.00445369491353631, + "learning_rate": 3.914190305244595e-06, + "loss": 0.0024, + "step": 5340 + }, + { + "epoch": 3.0676848067684808, + "grad_norm": 0.15039828419685364, + "learning_rate": 3.8946490211437735e-06, + "loss": 0.02, + "step": 5350 + }, + { + "epoch": 3.0676848067684808, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08091636747121811, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2042, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.796, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.421, + "step": 5350 + }, + { + "epoch": 3.073420807342081, + "grad_norm": 0.13300937414169312, + "learning_rate": 3.875125472195764e-06, + "loss": 0.0012, + "step": 5360 + }, + { + "epoch": 3.079156807915681, + "grad_norm": 2.184394121170044, + "learning_rate": 3.855619971652314e-06, + "loss": 0.0117, + "step": 5370 + }, + { + "epoch": 3.084892808489281, + "grad_norm": 0.02020282857120037, + "learning_rate": 3.836132832475583e-06, + "loss": 0.0014, + "step": 5380 + }, + { + "epoch": 3.090628809062881, + "grad_norm": 0.008466456085443497, + "learning_rate": 3.816664367333131e-06, + "loss": 0.009, + "step": 5390 + }, + { + "epoch": 3.096364809636481, + "grad_norm": 5.091494560241699, + "learning_rate": 3.797214888592896e-06, + "loss": 0.0128, + "step": 5400 + }, + { + "epoch": 3.096364809636481, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07716992497444153, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2065, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.792, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.42, + "step": 5400 + }, + { + "epoch": 3.102100810210081, + "grad_norm": 0.20820866525173187, + "learning_rate": 3.777784708318178e-06, + "loss": 0.0184, + "step": 5410 + }, + { + "epoch": 3.107836810783681, + "grad_norm": 1.1382921934127808, + "learning_rate": 3.7583741382626402e-06, + "loss": 0.0093, + "step": 5420 + }, + { + "epoch": 3.113572811357281, + "grad_norm": 0.38132885098457336, + "learning_rate": 3.7389834898653067e-06, + "loss": 0.0117, + "step": 5430 + }, + { + "epoch": 3.1193088119308814, + "grad_norm": 0.009956207126379013, + "learning_rate": 3.719613074245555e-06, + "loss": 0.0017, + "step": 5440 + }, + { + "epoch": 3.1250448125044814, + "grad_norm": 5.792750358581543, + "learning_rate": 3.7002632021981368e-06, + "loss": 0.0146, + "step": 5450 + }, + { + "epoch": 3.1250448125044814, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0894627794623375, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2897, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.672, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.36, + "step": 5450 + }, + { + "epoch": 3.1307808130780814, + "grad_norm": 1.089165210723877, + "learning_rate": 3.680934184188182e-06, + "loss": 0.0032, + "step": 5460 + }, + { + "epoch": 3.1365168136516814, + "grad_norm": 0.020885087549686432, + "learning_rate": 3.661626330346224e-06, + "loss": 0.0037, + "step": 5470 + }, + { + "epoch": 3.1422528142252815, + "grad_norm": 0.09390339255332947, + "learning_rate": 3.642339950463224e-06, + "loss": 0.0083, + "step": 5480 + }, + { + "epoch": 3.1479888147988815, + "grad_norm": 0.0025120277423411608, + "learning_rate": 3.62307535398559e-06, + "loss": 0.0108, + "step": 5490 + }, + { + "epoch": 3.1537248153724815, + "grad_norm": 0.06790509074926376, + "learning_rate": 3.603832850010226e-06, + "loss": 0.0012, + "step": 5500 + }, + { + "epoch": 3.1537248153724815, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08822309225797653, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3502, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.585, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.316, + "step": 5500 + }, + { + "epoch": 3.1594608159460815, + "grad_norm": 0.07959645241498947, + "learning_rate": 3.58461274727956e-06, + "loss": 0.0075, + "step": 5510 + }, + { + "epoch": 3.1651968165196815, + "grad_norm": 3.0567445755004883, + "learning_rate": 3.5654153541766023e-06, + "loss": 0.008, + "step": 5520 + }, + { + "epoch": 3.1709328170932816, + "grad_norm": 0.20890028774738312, + "learning_rate": 3.546240978719988e-06, + "loss": 0.0006, + "step": 5530 + }, + { + "epoch": 3.1766688176668816, + "grad_norm": 2.681074619293213, + "learning_rate": 3.5270899285590375e-06, + "loss": 0.0027, + "step": 5540 + }, + { + "epoch": 3.1824048182404816, + "grad_norm": 0.04251394048333168, + "learning_rate": 3.5079625109688243e-06, + "loss": 0.0005, + "step": 5550 + }, + { + "epoch": 3.1824048182404816, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09866516292095184, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2324, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.755, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.401, + "step": 5550 + }, + { + "epoch": 3.188140818814082, + "grad_norm": 4.062971591949463, + "learning_rate": 3.4888590328452353e-06, + "loss": 0.0178, + "step": 5560 + }, + { + "epoch": 3.193876819387682, + "grad_norm": 0.0014892960898578167, + "learning_rate": 3.4697798007000624e-06, + "loss": 0.0126, + "step": 5570 + }, + { + "epoch": 3.199612819961282, + "grad_norm": 0.005618450231850147, + "learning_rate": 3.450725120656069e-06, + "loss": 0.0261, + "step": 5580 + }, + { + "epoch": 3.205348820534882, + "grad_norm": 1.2523528337478638, + "learning_rate": 3.431695298442084e-06, + "loss": 0.0061, + "step": 5590 + }, + { + "epoch": 3.211084821108482, + "grad_norm": 3.6290197372436523, + "learning_rate": 3.4126906393881e-06, + "loss": 0.0177, + "step": 5600 + }, + { + "epoch": 3.211084821108482, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07931295037269592, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2163, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.778, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.413, + "step": 5600 + }, + { + "epoch": 3.216820821682082, + "grad_norm": 0.015163871459662914, + "learning_rate": 3.393711448420372e-06, + "loss": 0.0057, + "step": 5610 + }, + { + "epoch": 3.222556822255682, + "grad_norm": 0.059904176741838455, + "learning_rate": 3.37475803005652e-06, + "loss": 0.0017, + "step": 5620 + }, + { + "epoch": 3.228292822829282, + "grad_norm": 0.3671509623527527, + "learning_rate": 3.355830688400652e-06, + "loss": 0.0211, + "step": 5630 + }, + { + "epoch": 3.2340288234028822, + "grad_norm": 2.3451545238494873, + "learning_rate": 3.336929727138474e-06, + "loss": 0.0221, + "step": 5640 + }, + { + "epoch": 3.2397648239764822, + "grad_norm": 0.00577126070857048, + "learning_rate": 3.31805544953243e-06, + "loss": 0.0082, + "step": 5650 + }, + { + "epoch": 3.2397648239764822, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07332266122102737, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1867, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.821, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.434, + "step": 5650 + }, + { + "epoch": 3.2455008245500823, + "grad_norm": 1.244507074356079, + "learning_rate": 3.299208158416829e-06, + "loss": 0.0177, + "step": 5660 + }, + { + "epoch": 3.2512368251236827, + "grad_norm": 1.7493702173233032, + "learning_rate": 3.2803881561929806e-06, + "loss": 0.0182, + "step": 5670 + }, + { + "epoch": 3.2569728256972827, + "grad_norm": 0.03632393479347229, + "learning_rate": 3.2615957448243562e-06, + "loss": 0.0075, + "step": 5680 + }, + { + "epoch": 3.2627088262708828, + "grad_norm": 0.021715901792049408, + "learning_rate": 3.2428312258317306e-06, + "loss": 0.0044, + "step": 5690 + }, + { + "epoch": 3.268444826844483, + "grad_norm": 0.7747498154640198, + "learning_rate": 3.224094900288357e-06, + "loss": 0.0071, + "step": 5700 + }, + { + "epoch": 3.268444826844483, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08630760759115219, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1329, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.9, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.474, + "step": 5700 + }, + { + "epoch": 3.274180827418083, + "grad_norm": 0.003438598709180951, + "learning_rate": 3.205387068815127e-06, + "loss": 0.0025, + "step": 5710 + }, + { + "epoch": 3.279916827991683, + "grad_norm": 0.02451498433947563, + "learning_rate": 3.1867080315757477e-06, + "loss": 0.0004, + "step": 5720 + }, + { + "epoch": 3.285652828565283, + "grad_norm": 0.4422462582588196, + "learning_rate": 3.1680580882719304e-06, + "loss": 0.0206, + "step": 5730 + }, + { + "epoch": 3.291388829138883, + "grad_norm": 3.2265453338623047, + "learning_rate": 3.149437538138583e-06, + "loss": 0.0142, + "step": 5740 + }, + { + "epoch": 3.297124829712483, + "grad_norm": 0.015962867066264153, + "learning_rate": 3.130846679938998e-06, + "loss": 0.0099, + "step": 5750 + }, + { + "epoch": 3.297124829712483, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07895597815513611, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1289, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.906, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.476, + "step": 5750 + }, + { + "epoch": 3.302860830286083, + "grad_norm": 0.470051109790802, + "learning_rate": 3.1122858119600726e-06, + "loss": 0.0031, + "step": 5760 + }, + { + "epoch": 3.308596830859683, + "grad_norm": 0.002947809174656868, + "learning_rate": 3.0937552320075116e-06, + "loss": 0.0005, + "step": 5770 + }, + { + "epoch": 3.3143328314332834, + "grad_norm": 4.272603511810303, + "learning_rate": 3.0752552374010567e-06, + "loss": 0.0185, + "step": 5780 + }, + { + "epoch": 3.320068832006883, + "grad_norm": 0.0015301775420084596, + "learning_rate": 3.056786124969716e-06, + "loss": 0.0012, + "step": 5790 + }, + { + "epoch": 3.3258048325804834, + "grad_norm": 0.002659064019098878, + "learning_rate": 3.0383481910469936e-06, + "loss": 0.004, + "step": 5800 + }, + { + "epoch": 3.3258048325804834, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08382538706064224, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1553, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.867, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.457, + "step": 5800 + }, + { + "epoch": 3.3315408331540834, + "grad_norm": 0.25469204783439636, + "learning_rate": 3.019941731466143e-06, + "loss": 0.0005, + "step": 5810 + }, + { + "epoch": 3.3372768337276835, + "grad_norm": 0.5178417563438416, + "learning_rate": 3.0015670415554143e-06, + "loss": 0.016, + "step": 5820 + }, + { + "epoch": 3.3430128343012835, + "grad_norm": 0.07551216334104538, + "learning_rate": 2.9832244161333257e-06, + "loss": 0.0202, + "step": 5830 + }, + { + "epoch": 3.3487488348748835, + "grad_norm": 0.008605745621025562, + "learning_rate": 2.9649141495039225e-06, + "loss": 0.0009, + "step": 5840 + }, + { + "epoch": 3.3544848354484835, + "grad_norm": 0.0697212815284729, + "learning_rate": 2.9466365354520564e-06, + "loss": 0.0204, + "step": 5850 + }, + { + "epoch": 3.3544848354484835, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07445970177650452, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2129, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.783, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.415, + "step": 5850 + }, + { + "epoch": 3.3602208360220835, + "grad_norm": 0.017461596056818962, + "learning_rate": 2.928391867238679e-06, + "loss": 0.0012, + "step": 5860 + }, + { + "epoch": 3.3659568365956836, + "grad_norm": 0.0013357176212593913, + "learning_rate": 2.910180437596132e-06, + "loss": 0.0036, + "step": 5870 + }, + { + "epoch": 3.3716928371692836, + "grad_norm": 0.00179756130091846, + "learning_rate": 2.8920025387234484e-06, + "loss": 0.0131, + "step": 5880 + }, + { + "epoch": 3.3774288377428836, + "grad_norm": 0.005055127199739218, + "learning_rate": 2.8738584622816656e-06, + "loss": 0.0133, + "step": 5890 + }, + { + "epoch": 3.3831648383164836, + "grad_norm": 0.778300940990448, + "learning_rate": 2.8557484993891484e-06, + "loss": 0.007, + "step": 5900 + }, + { + "epoch": 3.3831648383164836, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08228859305381775, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2282, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.761, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.404, + "step": 5900 + }, + { + "epoch": 3.388900838890084, + "grad_norm": 0.05169231817126274, + "learning_rate": 2.837672940616911e-06, + "loss": 0.0088, + "step": 5910 + }, + { + "epoch": 3.394636839463684, + "grad_norm": 0.0016457217279821634, + "learning_rate": 2.8196320759839677e-06, + "loss": 0.0173, + "step": 5920 + }, + { + "epoch": 3.400372840037284, + "grad_norm": 0.11143210530281067, + "learning_rate": 2.801626194952669e-06, + "loss": 0.0015, + "step": 5930 + }, + { + "epoch": 3.406108840610884, + "grad_norm": 0.03351111710071564, + "learning_rate": 2.7836555864240566e-06, + "loss": 0.0079, + "step": 5940 + }, + { + "epoch": 3.411844841184484, + "grad_norm": 0.22604842483997345, + "learning_rate": 2.7657205387332313e-06, + "loss": 0.0124, + "step": 5950 + }, + { + "epoch": 3.411844841184484, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08073882758617401, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2042, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.796, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.421, + "step": 5950 + }, + { + "epoch": 3.417580841758084, + "grad_norm": 0.1360362470149994, + "learning_rate": 2.747821339644734e-06, + "loss": 0.028, + "step": 5960 + }, + { + "epoch": 3.423316842331684, + "grad_norm": 0.015132960863411427, + "learning_rate": 2.7299582763479115e-06, + "loss": 0.0041, + "step": 5970 + }, + { + "epoch": 3.4290528429052842, + "grad_norm": 0.06472807377576828, + "learning_rate": 2.7121316354523286e-06, + "loss": 0.0017, + "step": 5980 + }, + { + "epoch": 3.4347888434788842, + "grad_norm": 3.1584267616271973, + "learning_rate": 2.6943417029831504e-06, + "loss": 0.0232, + "step": 5990 + }, + { + "epoch": 3.4405248440524843, + "grad_norm": 0.004198137205094099, + "learning_rate": 2.6765887643765653e-06, + "loss": 0.0052, + "step": 6000 + }, + { + "epoch": 3.4405248440524843, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08295302093029022, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2169, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.777, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.412, + "step": 6000 + }, + { + "epoch": 3.4462608446260843, + "grad_norm": 0.018915435299277306, + "learning_rate": 2.6588731044752057e-06, + "loss": 0.0153, + "step": 6010 + }, + { + "epoch": 3.4519968451996847, + "grad_norm": 3.4281928539276123, + "learning_rate": 2.641195007523568e-06, + "loss": 0.0099, + "step": 6020 + }, + { + "epoch": 3.4577328457732848, + "grad_norm": 0.0030478844419121742, + "learning_rate": 2.623554757163464e-06, + "loss": 0.0004, + "step": 6030 + }, + { + "epoch": 3.463468846346885, + "grad_norm": 0.022162869572639465, + "learning_rate": 2.605952636429456e-06, + "loss": 0.0114, + "step": 6040 + }, + { + "epoch": 3.469204846920485, + "grad_norm": 0.009365621022880077, + "learning_rate": 2.5883889277443347e-06, + "loss": 0.0009, + "step": 6050 + }, + { + "epoch": 3.469204846920485, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07579995691776276, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2596, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.716, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.381, + "step": 6050 + }, + { + "epoch": 3.474940847494085, + "grad_norm": 0.29800668358802795, + "learning_rate": 2.570863912914566e-06, + "loss": 0.018, + "step": 6060 + }, + { + "epoch": 3.480676848067685, + "grad_norm": 3.4592084884643555, + "learning_rate": 2.5533778731257824e-06, + "loss": 0.0041, + "step": 6070 + }, + { + "epoch": 3.486412848641285, + "grad_norm": 0.03675074130296707, + "learning_rate": 2.535931088938274e-06, + "loss": 0.0129, + "step": 6080 + }, + { + "epoch": 3.492148849214885, + "grad_norm": 0.5131600499153137, + "learning_rate": 2.518523840282479e-06, + "loss": 0.0069, + "step": 6090 + }, + { + "epoch": 3.497884849788485, + "grad_norm": 0.19048012793064117, + "learning_rate": 2.5011564064544945e-06, + "loss": 0.0061, + "step": 6100 + }, + { + "epoch": 3.497884849788485, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07761486619710922, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2658, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.707, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.377, + "step": 6100 + }, + { + "epoch": 3.503620850362085, + "grad_norm": 0.37152099609375, + "learning_rate": 2.483829066111596e-06, + "loss": 0.0108, + "step": 6110 + }, + { + "epoch": 3.509356850935685, + "grad_norm": 4.403800964355469, + "learning_rate": 2.466542097267765e-06, + "loss": 0.0119, + "step": 6120 + }, + { + "epoch": 3.5150928515092854, + "grad_norm": 0.008035624399781227, + "learning_rate": 2.4492957772892345e-06, + "loss": 0.0192, + "step": 6130 + }, + { + "epoch": 3.520828852082885, + "grad_norm": 1.8271785974502563, + "learning_rate": 2.432090382890032e-06, + "loss": 0.0159, + "step": 6140 + }, + { + "epoch": 3.5265648526564854, + "grad_norm": 0.00973561592400074, + "learning_rate": 2.4149261901275393e-06, + "loss": 0.0184, + "step": 6150 + }, + { + "epoch": 3.5265648526564854, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07000607997179031, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3198, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.629, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.338, + "step": 6150 + }, + { + "epoch": 3.5323008532300855, + "grad_norm": 0.3272586464881897, + "learning_rate": 2.3978034743980676e-06, + "loss": 0.0014, + "step": 6160 + }, + { + "epoch": 3.5380368538036855, + "grad_norm": 0.015382179990410805, + "learning_rate": 2.3807225104324337e-06, + "loss": 0.0099, + "step": 6170 + }, + { + "epoch": 3.5437728543772855, + "grad_norm": 0.016446420922875404, + "learning_rate": 2.36368357229156e-06, + "loss": 0.0039, + "step": 6180 + }, + { + "epoch": 3.5495088549508855, + "grad_norm": 0.006563331466168165, + "learning_rate": 2.34668693336207e-06, + "loss": 0.0035, + "step": 6190 + }, + { + "epoch": 3.5552448555244855, + "grad_norm": 0.024284416809678078, + "learning_rate": 2.3297328663519044e-06, + "loss": 0.0117, + "step": 6200 + }, + { + "epoch": 3.5552448555244855, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07584439218044281, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3203, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.628, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.337, + "step": 6200 + }, + { + "epoch": 3.5609808560980856, + "grad_norm": 0.011505261063575745, + "learning_rate": 2.3128216432859435e-06, + "loss": 0.0192, + "step": 6210 + }, + { + "epoch": 3.5667168566716856, + "grad_norm": 0.01848476566374302, + "learning_rate": 2.2959535355016436e-06, + "loss": 0.0003, + "step": 6220 + }, + { + "epoch": 3.5724528572452856, + "grad_norm": 0.007101431954652071, + "learning_rate": 2.27912881364469e-06, + "loss": 0.0098, + "step": 6230 + }, + { + "epoch": 3.5781888578188856, + "grad_norm": 0.9875366687774658, + "learning_rate": 2.2623477476646447e-06, + "loss": 0.0014, + "step": 6240 + }, + { + "epoch": 3.5839248583924856, + "grad_norm": 0.07821710407733917, + "learning_rate": 2.2456106068106206e-06, + "loss": 0.0007, + "step": 6250 + }, + { + "epoch": 3.5839248583924856, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08138631284236908, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2288, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.76, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.404, + "step": 6250 + }, + { + "epoch": 3.589660858966086, + "grad_norm": 0.8070231676101685, + "learning_rate": 2.2289176596269545e-06, + "loss": 0.0015, + "step": 6260 + }, + { + "epoch": 3.5953968595396857, + "grad_norm": 0.0035041391383856535, + "learning_rate": 2.2122691739489143e-06, + "loss": 0.0021, + "step": 6270 + }, + { + "epoch": 3.601132860113286, + "grad_norm": 0.013247767463326454, + "learning_rate": 2.195665416898381e-06, + "loss": 0.0026, + "step": 6280 + }, + { + "epoch": 3.606868860686886, + "grad_norm": 3.3201675415039062, + "learning_rate": 2.179106654879581e-06, + "loss": 0.0236, + "step": 6290 + }, + { + "epoch": 3.612604861260486, + "grad_norm": 0.00799557100981474, + "learning_rate": 2.1625931535747964e-06, + "loss": 0.0078, + "step": 6300 + }, + { + "epoch": 3.612604861260486, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07524015754461288, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3312, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.613, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.33, + "step": 6300 + }, + { + "epoch": 3.618340861834086, + "grad_norm": 0.796405553817749, + "learning_rate": 2.1461251779401175e-06, + "loss": 0.0028, + "step": 6310 + }, + { + "epoch": 3.6240768624076862, + "grad_norm": 0.0037477288860827684, + "learning_rate": 2.1297029922011775e-06, + "loss": 0.0093, + "step": 6320 + }, + { + "epoch": 3.6298128629812862, + "grad_norm": 0.07188570499420166, + "learning_rate": 2.113326859848919e-06, + "loss": 0.0031, + "step": 6330 + }, + { + "epoch": 3.6355488635548863, + "grad_norm": 2.5588748455047607, + "learning_rate": 2.0969970436353725e-06, + "loss": 0.0074, + "step": 6340 + }, + { + "epoch": 3.6412848641284863, + "grad_norm": 0.008328588679432869, + "learning_rate": 2.080713805569427e-06, + "loss": 0.0025, + "step": 6350 + }, + { + "epoch": 3.6412848641284863, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.07739181071519852, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2601, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.715, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.381, + "step": 6350 + }, + { + "epoch": 3.6470208647020863, + "grad_norm": 0.0075204516761004925, + "learning_rate": 2.0644774069126406e-06, + "loss": 0.0109, + "step": 6360 + }, + { + "epoch": 3.6527568652756868, + "grad_norm": 0.003392142942175269, + "learning_rate": 2.048288108175036e-06, + "loss": 0.0091, + "step": 6370 + }, + { + "epoch": 3.6584928658492863, + "grad_norm": 0.00919413473457098, + "learning_rate": 2.0321461691109288e-06, + "loss": 0.007, + "step": 6380 + }, + { + "epoch": 3.664228866422887, + "grad_norm": 0.020889578387141228, + "learning_rate": 2.016051848714758e-06, + "loss": 0.0012, + "step": 6390 + }, + { + "epoch": 3.669964866996487, + "grad_norm": 0.09611943364143372, + "learning_rate": 2.000005405216931e-06, + "loss": 0.0011, + "step": 6400 + }, + { + "epoch": 3.669964866996487, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08481152355670929, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2176, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.776, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.412, + "step": 6400 + }, + { + "epoch": 3.675700867570087, + "grad_norm": 0.004153054673224688, + "learning_rate": 1.984007096079676e-06, + "loss": 0.0175, + "step": 6410 + }, + { + "epoch": 3.681436868143687, + "grad_norm": 0.0063193319365382195, + "learning_rate": 1.968057177992915e-06, + "loss": 0.0018, + "step": 6420 + }, + { + "epoch": 3.687172868717287, + "grad_norm": 0.05629734694957733, + "learning_rate": 1.9521559068701447e-06, + "loss": 0.0005, + "step": 6430 + }, + { + "epoch": 3.692908869290887, + "grad_norm": 0.042965829372406006, + "learning_rate": 1.9363035378443313e-06, + "loss": 0.0017, + "step": 6440 + }, + { + "epoch": 3.698644869864487, + "grad_norm": 0.029728276655077934, + "learning_rate": 1.9205003252638176e-06, + "loss": 0.0041, + "step": 6450 + }, + { + "epoch": 3.698644869864487, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08883155882358551, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1133, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.928, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.488, + "step": 6450 + }, + { + "epoch": 3.704380870438087, + "grad_norm": 0.014888194389641285, + "learning_rate": 1.904746522688236e-06, + "loss": 0.0018, + "step": 6460 + }, + { + "epoch": 3.710116871011687, + "grad_norm": 4.221490859985352, + "learning_rate": 1.8890423828844462e-06, + "loss": 0.0047, + "step": 6470 + }, + { + "epoch": 3.7158528715852874, + "grad_norm": 2.9691474437713623, + "learning_rate": 1.873388157822477e-06, + "loss": 0.0087, + "step": 6480 + }, + { + "epoch": 3.721588872158887, + "grad_norm": 0.1171032041311264, + "learning_rate": 1.8577840986714868e-06, + "loss": 0.0019, + "step": 6490 + }, + { + "epoch": 3.7273248727324875, + "grad_norm": 0.006826687604188919, + "learning_rate": 1.8422304557957315e-06, + "loss": 0.0123, + "step": 6500 + }, + { + "epoch": 3.7273248727324875, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08897148817777634, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2955, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.664, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.355, + "step": 6500 + }, + { + "epoch": 3.7330608733060875, + "grad_norm": 0.022587748244404793, + "learning_rate": 1.8267274787505446e-06, + "loss": 0.0087, + "step": 6510 + }, + { + "epoch": 3.7387968738796875, + "grad_norm": 0.5572834014892578, + "learning_rate": 1.8112754162783336e-06, + "loss": 0.0084, + "step": 6520 + }, + { + "epoch": 3.7445328744532875, + "grad_norm": 0.015338007360696793, + "learning_rate": 1.7958745163045987e-06, + "loss": 0.0085, + "step": 6530 + }, + { + "epoch": 3.7502688750268875, + "grad_norm": 0.004370348062366247, + "learning_rate": 1.7805250259339379e-06, + "loss": 0.0075, + "step": 6540 + }, + { + "epoch": 3.7560048756004876, + "grad_norm": 0.006390043999999762, + "learning_rate": 1.765227191446101e-06, + "loss": 0.0074, + "step": 6550 + }, + { + "epoch": 3.7560048756004876, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08355527371168137, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1302, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.904, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.475, + "step": 6550 + }, + { + "epoch": 3.7617408761740876, + "grad_norm": 0.09934690594673157, + "learning_rate": 1.7499812582920222e-06, + "loss": 0.0038, + "step": 6560 + }, + { + "epoch": 3.7674768767476876, + "grad_norm": 0.0005312475841492414, + "learning_rate": 1.734787471089887e-06, + "loss": 0.005, + "step": 6570 + }, + { + "epoch": 3.7732128773212876, + "grad_norm": 0.5943177938461304, + "learning_rate": 1.7196460736212167e-06, + "loss": 0.0111, + "step": 6580 + }, + { + "epoch": 3.7789488778948876, + "grad_norm": 0.03382878378033638, + "learning_rate": 1.7045573088269408e-06, + "loss": 0.0022, + "step": 6590 + }, + { + "epoch": 3.7846848784684877, + "grad_norm": 0.018695516511797905, + "learning_rate": 1.6895214188035125e-06, + "loss": 0.0038, + "step": 6600 + }, + { + "epoch": 3.7846848784684877, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0897902175784111, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2015, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.8, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.423, + "step": 6600 + }, + { + "epoch": 3.790420879042088, + "grad_norm": 0.03567569702863693, + "learning_rate": 1.6745386447990153e-06, + "loss": 0.0149, + "step": 6610 + }, + { + "epoch": 3.7961568796156877, + "grad_norm": 0.005937593523412943, + "learning_rate": 1.6596092272093e-06, + "loss": 0.0057, + "step": 6620 + }, + { + "epoch": 3.801892880189288, + "grad_norm": 0.001366928219795227, + "learning_rate": 1.6447334055741198e-06, + "loss": 0.0019, + "step": 6630 + }, + { + "epoch": 3.807628880762888, + "grad_norm": 0.017239965498447418, + "learning_rate": 1.6299114185732918e-06, + "loss": 0.0068, + "step": 6640 + }, + { + "epoch": 3.813364881336488, + "grad_norm": 0.09475360810756683, + "learning_rate": 1.6151435040228663e-06, + "loss": 0.0084, + "step": 6650 + }, + { + "epoch": 3.813364881336488, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09376490861177444, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2936, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.667, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.357, + "step": 6650 + }, + { + "epoch": 3.8191008819100882, + "grad_norm": 0.004428621847182512, + "learning_rate": 1.6004298988713153e-06, + "loss": 0.0003, + "step": 6660 + }, + { + "epoch": 3.8248368824836882, + "grad_norm": 0.0019930372945964336, + "learning_rate": 1.5857708391957194e-06, + "loss": 0.0012, + "step": 6670 + }, + { + "epoch": 3.8305728830572883, + "grad_norm": 0.24852816760540009, + "learning_rate": 1.571166560197991e-06, + "loss": 0.0018, + "step": 6680 + }, + { + "epoch": 3.8363088836308883, + "grad_norm": 0.05106634274125099, + "learning_rate": 1.5566172962010944e-06, + "loss": 0.0088, + "step": 6690 + }, + { + "epoch": 3.8420448842044883, + "grad_norm": 0.957535982131958, + "learning_rate": 1.542123280645292e-06, + "loss": 0.0013, + "step": 6700 + }, + { + "epoch": 3.8420448842044883, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09245079010725021, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1154, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.925, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.486, + "step": 6700 + }, + { + "epoch": 3.8477808847780883, + "grad_norm": 0.022298647090792656, + "learning_rate": 1.527684746084394e-06, + "loss": 0.0103, + "step": 6710 + }, + { + "epoch": 3.853516885351689, + "grad_norm": 0.21422079205513, + "learning_rate": 1.5133019241820257e-06, + "loss": 0.0341, + "step": 6720 + }, + { + "epoch": 3.8592528859252884, + "grad_norm": 0.08396855741739273, + "learning_rate": 1.4989750457079156e-06, + "loss": 0.0033, + "step": 6730 + }, + { + "epoch": 3.864988886498889, + "grad_norm": 0.12731043994426727, + "learning_rate": 1.484704340534187e-06, + "loss": 0.0028, + "step": 6740 + }, + { + "epoch": 3.870724887072489, + "grad_norm": 0.4012109637260437, + "learning_rate": 1.4704900376316773e-06, + "loss": 0.0008, + "step": 6750 + }, + { + "epoch": 3.870724887072489, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09105575829744339, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2863, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.677, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, + "step": 6750 + }, + { + "epoch": 3.876460887646089, + "grad_norm": 0.4341152012348175, + "learning_rate": 1.4563323650662586e-06, + "loss": 0.0013, + "step": 6760 + }, + { + "epoch": 3.882196888219689, + "grad_norm": 0.14372624456882477, + "learning_rate": 1.4422315499951783e-06, + "loss": 0.0059, + "step": 6770 + }, + { + "epoch": 3.887932888793289, + "grad_norm": 0.010315956547856331, + "learning_rate": 1.4281878186634157e-06, + "loss": 0.01, + "step": 6780 + }, + { + "epoch": 3.893668889366889, + "grad_norm": 0.1287292093038559, + "learning_rate": 1.4142013964000513e-06, + "loss": 0.01, + "step": 6790 + }, + { + "epoch": 3.899404889940489, + "grad_norm": 0.4034357964992523, + "learning_rate": 1.400272507614655e-06, + "loss": 0.0012, + "step": 6800 + }, + { + "epoch": 3.899404889940489, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09034840762615204, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1486, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.877, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.462, + "step": 6800 + }, + { + "epoch": 3.905140890514089, + "grad_norm": 0.15965530276298523, + "learning_rate": 1.3864013757936828e-06, + "loss": 0.0009, + "step": 6810 + }, + { + "epoch": 3.910876891087689, + "grad_norm": 0.07124567031860352, + "learning_rate": 1.3725882234968879e-06, + "loss": 0.0118, + "step": 6820 + }, + { + "epoch": 3.916612891661289, + "grad_norm": 0.022812863811850548, + "learning_rate": 1.3588332723537523e-06, + "loss": 0.0133, + "step": 6830 + }, + { + "epoch": 3.922348892234889, + "grad_norm": 0.0023223236203193665, + "learning_rate": 1.345136743059936e-06, + "loss": 0.013, + "step": 6840 + }, + { + "epoch": 3.9280848928084895, + "grad_norm": 0.08886972069740295, + "learning_rate": 1.3314988553737256e-06, + "loss": 0.0028, + "step": 6850 + }, + { + "epoch": 3.9280848928084895, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0912175104022026, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2896, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.672, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.36, + "step": 6850 + }, + { + "epoch": 3.933820893382089, + "grad_norm": 0.022729717195034027, + "learning_rate": 1.3179198281125188e-06, + "loss": 0.0069, + "step": 6860 + }, + { + "epoch": 3.9395568939556895, + "grad_norm": 2.09700083732605, + "learning_rate": 1.3043998791493034e-06, + "loss": 0.006, + "step": 6870 + }, + { + "epoch": 3.9452928945292896, + "grad_norm": 0.026967119425535202, + "learning_rate": 1.290939225409173e-06, + "loss": 0.0018, + "step": 6880 + }, + { + "epoch": 3.9510288951028896, + "grad_norm": 0.010416003875434399, + "learning_rate": 1.277538082865835e-06, + "loss": 0.012, + "step": 6890 + }, + { + "epoch": 3.9567648956764896, + "grad_norm": 0.04778837040066719, + "learning_rate": 1.2641966665381517e-06, + "loss": 0.0117, + "step": 6900 + }, + { + "epoch": 3.9567648956764896, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09045156091451645, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1714, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.844, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.445, + "step": 6900 + }, + { + "epoch": 3.9625008962500896, + "grad_norm": 0.0013466872042044997, + "learning_rate": 1.2509151904866922e-06, + "loss": 0.0012, + "step": 6910 + }, + { + "epoch": 3.9682368968236896, + "grad_norm": 0.057207245379686356, + "learning_rate": 1.2376938678102913e-06, + "loss": 0.0172, + "step": 6920 + }, + { + "epoch": 3.9739728973972896, + "grad_norm": 2.2670977115631104, + "learning_rate": 1.2245329106426367e-06, + "loss": 0.013, + "step": 6930 + }, + { + "epoch": 3.9797088979708897, + "grad_norm": 0.0011907550506293774, + "learning_rate": 1.2114325301488616e-06, + "loss": 0.0117, + "step": 6940 + }, + { + "epoch": 3.9854448985444897, + "grad_norm": 0.023963019251823425, + "learning_rate": 1.1983929365221548e-06, + "loss": 0.0096, + "step": 6950 + }, + { + "epoch": 3.9854448985444897, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08352689445018768, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2306, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.757, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.402, + "step": 6950 + }, + { + "epoch": 3.99118089911809, + "grad_norm": 0.02613544650375843, + "learning_rate": 1.1854143389803962e-06, + "loss": 0.0104, + "step": 6960 + }, + { + "epoch": 3.9969168996916897, + "grad_norm": 0.004353563766926527, + "learning_rate": 1.1724969457627928e-06, + "loss": 0.0004, + "step": 6970 + }, + { + "epoch": 4.00229440022944, + "grad_norm": 0.013296003453433514, + "learning_rate": 1.1596409641265376e-06, + "loss": 0.0053, + "step": 6980 + }, + { + "epoch": 4.00803040080304, + "grad_norm": 0.01596342958509922, + "learning_rate": 1.146846600343488e-06, + "loss": 0.0013, + "step": 6990 + }, + { + "epoch": 4.01376640137664, + "grad_norm": 0.003088761121034622, + "learning_rate": 1.1341140596968525e-06, + "loss": 0.0005, + "step": 7000 + }, + { + "epoch": 4.01376640137664, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08817622810602188, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2983, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.66, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.353, + "step": 7000 + }, + { + "epoch": 4.01950240195024, + "grad_norm": 0.0804172232747078, + "learning_rate": 1.1214435464779006e-06, + "loss": 0.0003, + "step": 7010 + }, + { + "epoch": 4.02523840252384, + "grad_norm": 0.13886569440364838, + "learning_rate": 1.1088352639826844e-06, + "loss": 0.0006, + "step": 7020 + }, + { + "epoch": 4.03097440309744, + "grad_norm": 0.029498988762497902, + "learning_rate": 1.0962894145087715e-06, + "loss": 0.0058, + "step": 7030 + }, + { + "epoch": 4.03671040367104, + "grad_norm": 3.7734415531158447, + "learning_rate": 1.0838061993520072e-06, + "loss": 0.0125, + "step": 7040 + }, + { + "epoch": 4.04244640424464, + "grad_norm": 0.12142331153154373, + "learning_rate": 1.0713858188032755e-06, + "loss": 0.0161, + "step": 7050 + }, + { + "epoch": 4.04244640424464, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09039638936519623, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2759, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.692, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.369, + "step": 7050 + }, + { + "epoch": 4.048182404818241, + "grad_norm": 0.003838218515738845, + "learning_rate": 1.0590284721452965e-06, + "loss": 0.0076, + "step": 7060 + }, + { + "epoch": 4.0539184053918405, + "grad_norm": 0.0031382606830447912, + "learning_rate": 1.0467343576494215e-06, + "loss": 0.0007, + "step": 7070 + }, + { + "epoch": 4.059654405965441, + "grad_norm": 0.0018334095366299152, + "learning_rate": 1.0345036725724517e-06, + "loss": 0.0206, + "step": 7080 + }, + { + "epoch": 4.0653904065390405, + "grad_norm": 0.06359425187110901, + "learning_rate": 1.0223366131534746e-06, + "loss": 0.0004, + "step": 7090 + }, + { + "epoch": 4.071126407112641, + "grad_norm": 0.08629762381315231, + "learning_rate": 1.010233374610719e-06, + "loss": 0.0003, + "step": 7100 + }, + { + "epoch": 4.071126407112641, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09012720733880997, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2717, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.698, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.373, + "step": 7100 + }, + { + "epoch": 4.0768624076862405, + "grad_norm": 0.16138513386249542, + "learning_rate": 9.981941511384152e-07, + "loss": 0.0042, + "step": 7110 + }, + { + "epoch": 4.082598408259841, + "grad_norm": 2.28838849067688, + "learning_rate": 9.862191359036883e-07, + "loss": 0.0019, + "step": 7120 + }, + { + "epoch": 4.088334408833441, + "grad_norm": 0.003568111453205347, + "learning_rate": 9.743085210434505e-07, + "loss": 0.0136, + "step": 7130 + }, + { + "epoch": 4.094070409407041, + "grad_norm": 2.805901050567627, + "learning_rate": 9.624624976613217e-07, + "loss": 0.0083, + "step": 7140 + }, + { + "epoch": 4.099806409980641, + "grad_norm": 0.009448690339922905, + "learning_rate": 9.506812558245665e-07, + "loss": 0.0004, + "step": 7150 + }, + { + "epoch": 4.099806409980641, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09947969764471054, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2914, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.67, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.358, + "step": 7150 + }, + { + "epoch": 4.105542410554241, + "grad_norm": 0.007924864068627357, + "learning_rate": 9.389649845610376e-07, + "loss": 0.0083, + "step": 7160 + }, + { + "epoch": 4.1112784111278415, + "grad_norm": 0.03764890506863594, + "learning_rate": 9.273138718561519e-07, + "loss": 0.0003, + "step": 7170 + }, + { + "epoch": 4.117014411701441, + "grad_norm": 0.015703987330198288, + "learning_rate": 9.157281046498628e-07, + "loss": 0.0141, + "step": 7180 + }, + { + "epoch": 4.122750412275042, + "grad_norm": 0.04593523591756821, + "learning_rate": 9.042078688336737e-07, + "loss": 0.0005, + "step": 7190 + }, + { + "epoch": 4.128486412848641, + "grad_norm": 0.08672873675823212, + "learning_rate": 8.927533492476437e-07, + "loss": 0.0088, + "step": 7200 + }, + { + "epoch": 4.128486412848641, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09371308237314224, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2865, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.677, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, + "step": 7200 + }, + { + "epoch": 4.134222413422242, + "grad_norm": 0.0006487391074188054, + "learning_rate": 8.813647296774291e-07, + "loss": 0.0033, + "step": 7210 + }, + { + "epoch": 4.139958413995841, + "grad_norm": 2.7435028553009033, + "learning_rate": 8.700421928513353e-07, + "loss": 0.0104, + "step": 7220 + }, + { + "epoch": 4.145694414569442, + "grad_norm": 0.12825919687747955, + "learning_rate": 8.58785920437376e-07, + "loss": 0.0131, + "step": 7230 + }, + { + "epoch": 4.151430415143041, + "grad_norm": 0.019473344087600708, + "learning_rate": 8.475960930403721e-07, + "loss": 0.0123, + "step": 7240 + }, + { + "epoch": 4.157166415716642, + "grad_norm": 0.019103556871414185, + "learning_rate": 8.364728901990404e-07, + "loss": 0.0161, + "step": 7250 + }, + { + "epoch": 4.157166415716642, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08745884150266647, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3788, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.544, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.296, + "step": 7250 + }, + { + "epoch": 4.162902416290241, + "grad_norm": 0.001977517269551754, + "learning_rate": 8.254164903831203e-07, + "loss": 0.009, + "step": 7260 + }, + { + "epoch": 4.168638416863842, + "grad_norm": 0.00919981300830841, + "learning_rate": 8.144270709905117e-07, + "loss": 0.0185, + "step": 7270 + }, + { + "epoch": 4.174374417437441, + "grad_norm": 0.007627399172633886, + "learning_rate": 8.035048083444241e-07, + "loss": 0.0013, + "step": 7280 + }, + { + "epoch": 4.180110418011042, + "grad_norm": 0.006586376577615738, + "learning_rate": 7.926498776905495e-07, + "loss": 0.0013, + "step": 7290 + }, + { + "epoch": 4.185846418584642, + "grad_norm": 0.0025622497778385878, + "learning_rate": 7.818624531942493e-07, + "loss": 0.0003, + "step": 7300 + }, + { + "epoch": 4.185846418584642, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08901920914649963, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2083, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.79, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.419, + "step": 7300 + }, + { + "epoch": 4.191582419158242, + "grad_norm": 0.07510726898908615, + "learning_rate": 7.711427079377614e-07, + "loss": 0.0055, + "step": 7310 + }, + { + "epoch": 4.197318419731842, + "grad_norm": 0.002239730441942811, + "learning_rate": 7.604908139174255e-07, + "loss": 0.0007, + "step": 7320 + }, + { + "epoch": 4.203054420305442, + "grad_norm": 0.0018331691389903426, + "learning_rate": 7.499069420409183e-07, + "loss": 0.0104, + "step": 7330 + }, + { + "epoch": 4.208790420879042, + "grad_norm": 0.06228185445070267, + "learning_rate": 7.393912621245142e-07, + "loss": 0.015, + "step": 7340 + }, + { + "epoch": 4.214526421452642, + "grad_norm": 0.0020015144255012274, + "learning_rate": 7.289439428903599e-07, + "loss": 0.0015, + "step": 7350 + }, + { + "epoch": 4.214526421452642, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08976524323225021, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3011, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.656, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.351, + "step": 7350 + }, + { + "epoch": 4.220262422026242, + "grad_norm": 3.2421772480010986, + "learning_rate": 7.185651519637659e-07, + "loss": 0.0083, + "step": 7360 + }, + { + "epoch": 4.225998422599842, + "grad_norm": 0.0900789201259613, + "learning_rate": 7.082550558705225e-07, + "loss": 0.0084, + "step": 7370 + }, + { + "epoch": 4.231734423173442, + "grad_norm": 0.0065153795294463634, + "learning_rate": 6.980138200342229e-07, + "loss": 0.0003, + "step": 7380 + }, + { + "epoch": 4.237470423747042, + "grad_norm": 0.041670165956020355, + "learning_rate": 6.878416087736079e-07, + "loss": 0.0006, + "step": 7390 + }, + { + "epoch": 4.243206424320642, + "grad_norm": 0.005517491605132818, + "learning_rate": 6.777385852999319e-07, + "loss": 0.0005, + "step": 7400 + }, + { + "epoch": 4.243206424320642, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08851981908082962, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2967, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.662, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.355, + "step": 7400 + }, + { + "epoch": 4.248942424894243, + "grad_norm": 0.007675142493098974, + "learning_rate": 6.67704911714348e-07, + "loss": 0.0127, + "step": 7410 + }, + { + "epoch": 4.2546784254678425, + "grad_norm": 2.6328227519989014, + "learning_rate": 6.577407490052978e-07, + "loss": 0.0152, + "step": 7420 + }, + { + "epoch": 4.260414426041443, + "grad_norm": 0.14412537217140198, + "learning_rate": 6.478462570459388e-07, + "loss": 0.0012, + "step": 7430 + }, + { + "epoch": 4.2661504266150425, + "grad_norm": 3.322977066040039, + "learning_rate": 6.380215945915702e-07, + "loss": 0.0065, + "step": 7440 + }, + { + "epoch": 4.271886427188643, + "grad_norm": 0.0006975011201575398, + "learning_rate": 6.282669192770896e-07, + "loss": 0.0008, + "step": 7450 + }, + { + "epoch": 4.271886427188643, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09144297242164612, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2868, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.676, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.362, + "step": 7450 + }, + { + "epoch": 4.2776224277622426, + "grad_norm": 0.04776507988572121, + "learning_rate": 6.185823876144675e-07, + "loss": 0.0052, + "step": 7460 + }, + { + "epoch": 4.283358428335843, + "grad_norm": 4.260890483856201, + "learning_rate": 6.089681549902287e-07, + "loss": 0.0172, + "step": 7470 + }, + { + "epoch": 4.289094428909443, + "grad_norm": 0.0017789709381759167, + "learning_rate": 5.994243756629664e-07, + "loss": 0.0045, + "step": 7480 + }, + { + "epoch": 4.294830429483043, + "grad_norm": 0.020361248403787613, + "learning_rate": 5.89951202760859e-07, + "loss": 0.0013, + "step": 7490 + }, + { + "epoch": 4.300566430056643, + "grad_norm": 0.01590687222778797, + "learning_rate": 5.805487882792227e-07, + "loss": 0.0079, + "step": 7500 + }, + { + "epoch": 4.300566430056643, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08803683519363403, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2002, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.802, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.424, + "step": 7500 + }, + { + "epoch": 4.306302430630243, + "grad_norm": 0.016193749383091927, + "learning_rate": 5.712172830780649e-07, + "loss": 0.0032, + "step": 7510 + }, + { + "epoch": 4.312038431203844, + "grad_norm": 0.2794877588748932, + "learning_rate": 5.619568368796657e-07, + "loss": 0.0046, + "step": 7520 + }, + { + "epoch": 4.317774431777443, + "grad_norm": 0.3345611095428467, + "learning_rate": 5.527675982661801e-07, + "loss": 0.0028, + "step": 7530 + }, + { + "epoch": 4.323510432351044, + "grad_norm": 0.10426267981529236, + "learning_rate": 5.43649714677249e-07, + "loss": 0.0013, + "step": 7540 + }, + { + "epoch": 4.329246432924643, + "grad_norm": 0.0005553574301302433, + "learning_rate": 5.346033324076338e-07, + "loss": 0.006, + "step": 7550 + }, + { + "epoch": 4.329246432924643, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.087630994617939, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2554, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.722, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.384, + "step": 7550 + }, + { + "epoch": 4.334982433498244, + "grad_norm": 0.08077429980039597, + "learning_rate": 5.256285966048719e-07, + "loss": 0.0067, + "step": 7560 + }, + { + "epoch": 4.340718434071843, + "grad_norm": 0.004632384981960058, + "learning_rate": 5.167256512669444e-07, + "loss": 0.0084, + "step": 7570 + }, + { + "epoch": 4.346454434645444, + "grad_norm": 0.0067880041897296906, + "learning_rate": 5.078946392399703e-07, + "loss": 0.0021, + "step": 7580 + }, + { + "epoch": 4.352190435219043, + "grad_norm": 0.001848329440690577, + "learning_rate": 4.991357022159111e-07, + "loss": 0.0007, + "step": 7590 + }, + { + "epoch": 4.357926435792644, + "grad_norm": 0.09323499351739883, + "learning_rate": 4.90448980730297e-07, + "loss": 0.0024, + "step": 7600 + }, + { + "epoch": 4.357926435792644, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08750669658184052, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2509, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.728, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.388, + "step": 7600 + }, + { + "epoch": 4.363662436366243, + "grad_norm": 0.0019721076823771, + "learning_rate": 4.818346141599739e-07, + "loss": 0.0037, + "step": 7610 + }, + { + "epoch": 4.369398436939844, + "grad_norm": 0.26007330417633057, + "learning_rate": 4.732927407208654e-07, + "loss": 0.0011, + "step": 7620 + }, + { + "epoch": 4.375134437513443, + "grad_norm": 3.1276211738586426, + "learning_rate": 4.6482349746575783e-07, + "loss": 0.0274, + "step": 7630 + }, + { + "epoch": 4.380870438087044, + "grad_norm": 0.004567150957882404, + "learning_rate": 4.564270202820997e-07, + "loss": 0.0004, + "step": 7640 + }, + { + "epoch": 4.386606438660644, + "grad_norm": 0.033358391374349594, + "learning_rate": 4.4810344388982017e-07, + "loss": 0.0078, + "step": 7650 + }, + { + "epoch": 4.386606438660644, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08857561647891998, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2607, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.714, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.381, + "step": 7650 + }, + { + "epoch": 4.392342439234244, + "grad_norm": 0.3984158933162689, + "learning_rate": 4.3985290183916893e-07, + "loss": 0.003, + "step": 7660 + }, + { + "epoch": 4.398078439807844, + "grad_norm": 0.09813853353261948, + "learning_rate": 4.316755265085715e-07, + "loss": 0.002, + "step": 7670 + }, + { + "epoch": 4.403814440381444, + "grad_norm": 0.013608184643089771, + "learning_rate": 4.2357144910251003e-07, + "loss": 0.0007, + "step": 7680 + }, + { + "epoch": 4.409550440955044, + "grad_norm": 0.012810665182769299, + "learning_rate": 4.155407996494143e-07, + "loss": 0.0003, + "step": 7690 + }, + { + "epoch": 4.415286441528644, + "grad_norm": 0.5154911279678345, + "learning_rate": 4.0758370699957416e-07, + "loss": 0.0135, + "step": 7700 + }, + { + "epoch": 4.415286441528644, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08959878236055374, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2846, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.679, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.363, + "step": 7700 + }, + { + "epoch": 4.421022442102244, + "grad_norm": 0.0007974316249601543, + "learning_rate": 3.997002988230747e-07, + "loss": 0.0026, + "step": 7710 + }, + { + "epoch": 4.426758442675844, + "grad_norm": 0.032501526176929474, + "learning_rate": 3.918907016077489e-07, + "loss": 0.0051, + "step": 7720 + }, + { + "epoch": 4.432494443249444, + "grad_norm": 0.2708737552165985, + "learning_rate": 3.841550406571443e-07, + "loss": 0.0009, + "step": 7730 + }, + { + "epoch": 4.438230443823044, + "grad_norm": 0.6233472228050232, + "learning_rate": 3.764934400885162e-07, + "loss": 0.0012, + "step": 7740 + }, + { + "epoch": 4.4439664443966445, + "grad_norm": 0.00045903949649073184, + "learning_rate": 3.689060228308339e-07, + "loss": 0.0016, + "step": 7750 + }, + { + "epoch": 4.4439664443966445, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08894706517457962, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3241, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.623, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.335, + "step": 7750 + }, + { + "epoch": 4.449702444970244, + "grad_norm": 0.02216159924864769, + "learning_rate": 3.6139291062281036e-07, + "loss": 0.0112, + "step": 7760 + }, + { + "epoch": 4.4554384455438445, + "grad_norm": 0.028067264705896378, + "learning_rate": 3.539542240109456e-07, + "loss": 0.0044, + "step": 7770 + }, + { + "epoch": 4.461174446117445, + "grad_norm": 3.8059170246124268, + "learning_rate": 3.4659008234759597e-07, + "loss": 0.0063, + "step": 7780 + }, + { + "epoch": 4.4669104466910445, + "grad_norm": 0.0011888346634805202, + "learning_rate": 3.3930060378905893e-07, + "loss": 0.0003, + "step": 7790 + }, + { + "epoch": 4.472646447264645, + "grad_norm": 0.0015550279058516026, + "learning_rate": 3.320859052936731e-07, + "loss": 0.0009, + "step": 7800 + }, + { + "epoch": 4.472646447264645, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09025800973176956, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1333, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.899, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.473, + "step": 7800 + }, + { + "epoch": 4.478382447838245, + "grad_norm": 0.039686888456344604, + "learning_rate": 3.249461026199485e-07, + "loss": 0.0003, + "step": 7810 + }, + { + "epoch": 4.484118448411845, + "grad_norm": 3.7747395038604736, + "learning_rate": 3.1788131032470306e-07, + "loss": 0.0124, + "step": 7820 + }, + { + "epoch": 4.489854448985445, + "grad_norm": 0.000805154733825475, + "learning_rate": 3.108916417612262e-07, + "loss": 0.0031, + "step": 7830 + }, + { + "epoch": 4.495590449559045, + "grad_norm": 0.0011434925254434347, + "learning_rate": 3.039772090774634e-07, + "loss": 0.0009, + "step": 7840 + }, + { + "epoch": 4.501326450132645, + "grad_norm": 0.019072147086262703, + "learning_rate": 2.97138123214214e-07, + "loss": 0.0003, + "step": 7850 + }, + { + "epoch": 4.501326450132645, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09056304395198822, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.169, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.847, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.447, + "step": 7850 + }, + { + "epoch": 4.507062450706245, + "grad_norm": 0.005647636018693447, + "learning_rate": 2.903744939033504e-07, + "loss": 0.0001, + "step": 7860 + }, + { + "epoch": 4.512798451279846, + "grad_norm": 2.6264231204986572, + "learning_rate": 2.836864296660574e-07, + "loss": 0.0132, + "step": 7870 + }, + { + "epoch": 4.518534451853445, + "grad_norm": 0.010306664742529392, + "learning_rate": 2.770740378110942e-07, + "loss": 0.0145, + "step": 7880 + }, + { + "epoch": 4.524270452427046, + "grad_norm": 0.18294845521450043, + "learning_rate": 2.7053742443307054e-07, + "loss": 0.0041, + "step": 7890 + }, + { + "epoch": 4.530006453000645, + "grad_norm": 0.017165735363960266, + "learning_rate": 2.640766944107431e-07, + "loss": 0.0007, + "step": 7900 + }, + { + "epoch": 4.530006453000645, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08944171667098999, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2567, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.72, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.383, + "step": 7900 + }, + { + "epoch": 4.535742453574246, + "grad_norm": 0.04896865785121918, + "learning_rate": 2.5769195140533556e-07, + "loss": 0.0125, + "step": 7910 + }, + { + "epoch": 4.541478454147845, + "grad_norm": 0.022854099050164223, + "learning_rate": 2.5138329785887263e-07, + "loss": 0.0005, + "step": 7920 + }, + { + "epoch": 4.547214454721446, + "grad_norm": 0.0018585673533380032, + "learning_rate": 2.4515083499253743e-07, + "loss": 0.0016, + "step": 7930 + }, + { + "epoch": 4.552950455295045, + "grad_norm": 0.003768256865441799, + "learning_rate": 2.3899466280504936e-07, + "loss": 0.0217, + "step": 7940 + }, + { + "epoch": 4.558686455868646, + "grad_norm": 0.05598225072026253, + "learning_rate": 2.3291488007105734e-07, + "loss": 0.009, + "step": 7950 + }, + { + "epoch": 4.558686455868646, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09063292294740677, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1171, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.923, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.485, + "step": 7950 + }, + { + "epoch": 4.564422456442245, + "grad_norm": 3.3265416622161865, + "learning_rate": 2.269115843395553e-07, + "loss": 0.0151, + "step": 7960 + }, + { + "epoch": 4.570158457015846, + "grad_norm": 0.0049064746126532555, + "learning_rate": 2.2098487193231543e-07, + "loss": 0.0208, + "step": 7970 + }, + { + "epoch": 4.575894457589445, + "grad_norm": 0.0036308506969362497, + "learning_rate": 2.1513483794234847e-07, + "loss": 0.0068, + "step": 7980 + }, + { + "epoch": 4.581630458163046, + "grad_norm": 0.0013542186934500933, + "learning_rate": 2.093615762323703e-07, + "loss": 0.0022, + "step": 7990 + }, + { + "epoch": 4.587366458736646, + "grad_norm": 2.7696778774261475, + "learning_rate": 2.0366517943330278e-07, + "loss": 0.0014, + "step": 8000 + }, + { + "epoch": 4.587366458736646, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.08961217105388641, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3584, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.574, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.31, + "step": 8000 + }, + { + "epoch": 4.593102459310246, + "grad_norm": 0.0006430260837078094, + "learning_rate": 1.9804573894278311e-07, + "loss": 0.0137, + "step": 8010 + }, + { + "epoch": 4.598838459883846, + "grad_norm": 0.49191415309906006, + "learning_rate": 1.925033449236974e-07, + "loss": 0.0127, + "step": 8020 + }, + { + "epoch": 4.604574460457446, + "grad_norm": 0.4231385886669159, + "learning_rate": 1.8703808630273768e-07, + "loss": 0.0108, + "step": 8030 + }, + { + "epoch": 4.610310461031046, + "grad_norm": 5.982341289520264, + "learning_rate": 1.8165005076897114e-07, + "loss": 0.0034, + "step": 8040 + }, + { + "epoch": 4.616046461604646, + "grad_norm": 0.01885843276977539, + "learning_rate": 1.7633932477243599e-07, + "loss": 0.0227, + "step": 8050 + }, + { + "epoch": 4.616046461604646, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09145788848400116, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2115, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.785, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.416, + "step": 8050 + }, + { + "epoch": 4.621782462178246, + "grad_norm": 0.004457383882254362, + "learning_rate": 1.7110599352275104e-07, + "loss": 0.0066, + "step": 8060 + }, + { + "epoch": 4.627518462751846, + "grad_norm": 0.005887819454073906, + "learning_rate": 1.6595014098775342e-07, + "loss": 0.004, + "step": 8070 + }, + { + "epoch": 4.633254463325446, + "grad_norm": 0.023113150149583817, + "learning_rate": 1.6087184989214687e-07, + "loss": 0.0258, + "step": 8080 + }, + { + "epoch": 4.638990463899046, + "grad_norm": 0.020028818398714066, + "learning_rate": 1.5587120171617509e-07, + "loss": 0.0005, + "step": 8090 + }, + { + "epoch": 4.6447264644726465, + "grad_norm": 0.01912406086921692, + "learning_rate": 1.509482766943182e-07, + "loss": 0.0006, + "step": 8100 + }, + { + "epoch": 4.6447264644726465, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09051110595464706, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.4316, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.469, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.258, + "step": 8100 + }, + { + "epoch": 4.650462465046246, + "grad_norm": 3.4732401371002197, + "learning_rate": 1.4610315381400175e-07, + "loss": 0.0149, + "step": 8110 + }, + { + "epoch": 4.6561984656198465, + "grad_norm": 0.0006817178218625486, + "learning_rate": 1.4133591081432806e-07, + "loss": 0.0008, + "step": 8120 + }, + { + "epoch": 4.661934466193447, + "grad_norm": 0.0062561482191085815, + "learning_rate": 1.366466241848341e-07, + "loss": 0.0004, + "step": 8130 + }, + { + "epoch": 4.6676704667670466, + "grad_norm": 0.022234557196497917, + "learning_rate": 1.3203536916425842e-07, + "loss": 0.001, + "step": 8140 + }, + { + "epoch": 4.673406467340647, + "grad_norm": 4.0328369140625, + "learning_rate": 1.275022197393394e-07, + "loss": 0.0047, + "step": 8150 + }, + { + "epoch": 4.673406467340647, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09013617783784866, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1698, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.846, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.447, + "step": 8150 + }, + { + "epoch": 4.679142467914247, + "grad_norm": 0.15110693871974945, + "learning_rate": 1.230472486436246e-07, + "loss": 0.0004, + "step": 8160 + }, + { + "epoch": 4.684878468487847, + "grad_norm": 0.009227008558809757, + "learning_rate": 1.1867052735630435e-07, + "loss": 0.0014, + "step": 8170 + }, + { + "epoch": 4.690614469061447, + "grad_norm": 0.2895458936691284, + "learning_rate": 1.1437212610106496e-07, + "loss": 0.0013, + "step": 8180 + }, + { + "epoch": 4.696350469635047, + "grad_norm": 0.541077733039856, + "learning_rate": 1.1015211384496238e-07, + "loss": 0.0102, + "step": 8190 + }, + { + "epoch": 4.702086470208647, + "grad_norm": 2.9748380184173584, + "learning_rate": 1.0601055829731588e-07, + "loss": 0.023, + "step": 8200 + }, + { + "epoch": 4.702086470208647, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09088833630084991, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2127, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.783, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.415, + "step": 8200 + }, + { + "epoch": 4.707822470782247, + "grad_norm": 5.7777838706970215, + "learning_rate": 1.0194752590862223e-07, + "loss": 0.0119, + "step": 8210 + }, + { + "epoch": 4.713558471355848, + "grad_norm": 0.06003111973404884, + "learning_rate": 9.796308186948711e-08, + "loss": 0.0007, + "step": 8220 + }, + { + "epoch": 4.719294471929447, + "grad_norm": 0.005556735210120678, + "learning_rate": 9.405729010958043e-08, + "loss": 0.001, + "step": 8230 + }, + { + "epoch": 4.725030472503047, + "grad_norm": 2.6498606204986572, + "learning_rate": 9.023021329661152e-08, + "loss": 0.0095, + "step": 8240 + }, + { + "epoch": 4.730766473076647, + "grad_norm": 0.03512834012508392, + "learning_rate": 8.648191283532337e-08, + "loss": 0.0081, + "step": 8250 + }, + { + "epoch": 4.730766473076647, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0907173827290535, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.316, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.634, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.341, + "step": 8250 + }, + { + "epoch": 4.736502473650248, + "grad_norm": 0.009284062311053276, + "learning_rate": 8.281244886650607e-08, + "loss": 0.0051, + "step": 8260 + }, + { + "epoch": 4.742238474223847, + "grad_norm": 0.02183857187628746, + "learning_rate": 7.922188026603273e-08, + "loss": 0.0022, + "step": 8270 + }, + { + "epoch": 4.747974474797448, + "grad_norm": 0.0035564859863370657, + "learning_rate": 7.571026464391451e-08, + "loss": 0.0025, + "step": 8280 + }, + { + "epoch": 4.753710475371047, + "grad_norm": 0.04188934341073036, + "learning_rate": 7.227765834337874e-08, + "loss": 0.0004, + "step": 8290 + }, + { + "epoch": 4.759446475944648, + "grad_norm": 0.6253750324249268, + "learning_rate": 6.892411643995955e-08, + "loss": 0.0006, + "step": 8300 + }, + { + "epoch": 4.759446475944648, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.090791717171669, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3622, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.568, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.307, + "step": 8300 + }, + { + "epoch": 4.765182476518247, + "grad_norm": 0.3705517053604126, + "learning_rate": 6.564969274061972e-08, + "loss": 0.0042, + "step": 8310 + }, + { + "epoch": 4.770918477091848, + "grad_norm": 0.0012910285731777549, + "learning_rate": 6.245443978288413e-08, + "loss": 0.0023, + "step": 8320 + }, + { + "epoch": 4.776654477665447, + "grad_norm": 0.5175131559371948, + "learning_rate": 5.933840883399766e-08, + "loss": 0.0011, + "step": 8330 + }, + { + "epoch": 4.782390478239048, + "grad_norm": 3.527125358581543, + "learning_rate": 5.630164989010312e-08, + "loss": 0.0023, + "step": 8340 + }, + { + "epoch": 4.788126478812648, + "grad_norm": 0.3973446190357208, + "learning_rate": 5.334421167543735e-08, + "loss": 0.0026, + "step": 8350 + }, + { + "epoch": 4.788126478812648, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09068261086940765, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1697, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.846, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.447, + "step": 8350 + }, + { + "epoch": 4.793862479386248, + "grad_norm": 0.003769023111090064, + "learning_rate": 5.0466141641553056e-08, + "loss": 0.0001, + "step": 8360 + }, + { + "epoch": 4.799598479959848, + "grad_norm": 0.08687523007392883, + "learning_rate": 4.766748596655268e-08, + "loss": 0.0061, + "step": 8370 + }, + { + "epoch": 4.805334480533448, + "grad_norm": 5.824708461761475, + "learning_rate": 4.494828955435126e-08, + "loss": 0.0047, + "step": 8380 + }, + { + "epoch": 4.811070481107048, + "grad_norm": 2.9158382415771484, + "learning_rate": 4.230859603395421e-08, + "loss": 0.011, + "step": 8390 + }, + { + "epoch": 4.816806481680648, + "grad_norm": 0.04298550263047218, + "learning_rate": 3.974844775875786e-08, + "loss": 0.0016, + "step": 8400 + }, + { + "epoch": 4.816806481680648, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09151580929756165, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2757, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.692, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.37, + "step": 8400 + }, + { + "epoch": 4.822542482254248, + "grad_norm": 0.005364727228879929, + "learning_rate": 3.726788580586893e-08, + "loss": 0.0099, + "step": 8410 + }, + { + "epoch": 4.828278482827848, + "grad_norm": 0.0018757123034447432, + "learning_rate": 3.4866949975448374e-08, + "loss": 0.0009, + "step": 8420 + }, + { + "epoch": 4.8340144834014485, + "grad_norm": 0.008776834234595299, + "learning_rate": 3.254567879006798e-08, + "loss": 0.0068, + "step": 8430 + }, + { + "epoch": 4.839750483975048, + "grad_norm": 0.009764501824975014, + "learning_rate": 3.030410949409701e-08, + "loss": 0.0088, + "step": 8440 + }, + { + "epoch": 4.8454864845486485, + "grad_norm": 4.281154632568359, + "learning_rate": 2.8142278053101545e-08, + "loss": 0.0135, + "step": 8450 + }, + { + "epoch": 4.8454864845486485, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.0919208973646164, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1321, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.901, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.474, + "step": 8450 + }, + { + "epoch": 4.851222485122248, + "grad_norm": 0.09912008792161942, + "learning_rate": 2.6060219153268286e-08, + "loss": 0.0005, + "step": 8460 + }, + { + "epoch": 4.8569584856958485, + "grad_norm": 0.019211160019040108, + "learning_rate": 2.4057966200849437e-08, + "loss": 0.0005, + "step": 8470 + }, + { + "epoch": 4.862694486269449, + "grad_norm": 0.05037353187799454, + "learning_rate": 2.213555132162426e-08, + "loss": 0.0012, + "step": 8480 + }, + { + "epoch": 4.868430486843049, + "grad_norm": 0.00411713682115078, + "learning_rate": 2.0293005360386142e-08, + "loss": 0.0003, + "step": 8490 + }, + { + "epoch": 4.874166487416649, + "grad_norm": 0.0006664736429229379, + "learning_rate": 1.8530357880444105e-08, + "loss": 0.0202, + "step": 8500 + }, + { + "epoch": 4.874166487416649, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09091775864362717, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2769, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.69, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.369, + "step": 8500 + }, + { + "epoch": 4.879902487990249, + "grad_norm": 1.6142276525497437, + "learning_rate": 1.684763716315374e-08, + "loss": 0.0011, + "step": 8510 + }, + { + "epoch": 4.885638488563849, + "grad_norm": 0.000510135549120605, + "learning_rate": 1.524487020745813e-08, + "loss": 0.0011, + "step": 8520 + }, + { + "epoch": 4.891374489137449, + "grad_norm": 0.001830592635087669, + "learning_rate": 1.372208272945763e-08, + "loss": 0.0187, + "step": 8530 + }, + { + "epoch": 4.897110489711049, + "grad_norm": 0.2799864709377289, + "learning_rate": 1.2279299161997438e-08, + "loss": 0.0157, + "step": 8540 + }, + { + "epoch": 4.902846490284649, + "grad_norm": 0.009001590311527252, + "learning_rate": 1.0916542654273443e-08, + "loss": 0.0272, + "step": 8550 + }, + { + "epoch": 4.902846490284649, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09087590873241425, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.3382, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.602, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.325, + "step": 8550 + }, + { + "epoch": 4.908582490858249, + "grad_norm": 0.17294363677501678, + "learning_rate": 9.633835071463094e-09, + "loss": 0.0052, + "step": 8560 + }, + { + "epoch": 4.91431849143185, + "grad_norm": 0.008691389113664627, + "learning_rate": 8.431196994373447e-09, + "loss": 0.0065, + "step": 8570 + }, + { + "epoch": 4.920054492005449, + "grad_norm": 0.010057262144982815, + "learning_rate": 7.3086477191103285e-09, + "loss": 0.02, + "step": 8580 + }, + { + "epoch": 4.925790492579049, + "grad_norm": 0.19418714940547943, + "learning_rate": 6.2662052567702414e-09, + "loss": 0.0003, + "step": 8590 + }, + { + "epoch": 4.931526493152649, + "grad_norm": 0.033849816769361496, + "learning_rate": 5.303886333151154e-09, + "loss": 0.0126, + "step": 8600 + }, + { + "epoch": 4.931526493152649, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09129200130701065, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2495, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.73, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.389, + "step": 8600 + }, + { + "epoch": 4.93726249372625, + "grad_norm": 0.024734610691666603, + "learning_rate": 4.421706388481606e-09, + "loss": 0.0171, + "step": 8610 + }, + { + "epoch": 4.942998494299849, + "grad_norm": 0.0016847607912495732, + "learning_rate": 3.6196795771770156e-09, + "loss": 0.0126, + "step": 8620 + }, + { + "epoch": 4.94873449487345, + "grad_norm": 0.004256678279489279, + "learning_rate": 2.897818767609861e-09, + "loss": 0.0161, + "step": 8630 + }, + { + "epoch": 4.954470495447049, + "grad_norm": 4.404531002044678, + "learning_rate": 2.2561355419037368e-09, + "loss": 0.0087, + "step": 8640 + }, + { + "epoch": 4.96020649602065, + "grad_norm": 3.0091116428375244, + "learning_rate": 1.6946401957479431e-09, + "loss": 0.0145, + "step": 8650 + }, + { + "epoch": 4.96020649602065, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09181646257638931, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.2421, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.741, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.394, + "step": 8650 + }, + { + "epoch": 4.965942496594249, + "grad_norm": 0.04898509010672569, + "learning_rate": 1.2133417382320656e-09, + "loss": 0.0032, + "step": 8660 + }, + { + "epoch": 4.97167849716785, + "grad_norm": 0.002836048137396574, + "learning_rate": 8.122478916999799e-10, + "loss": 0.0074, + "step": 8670 + }, + { + "epoch": 4.977414497741449, + "grad_norm": 0.02022160217165947, + "learning_rate": 4.913650916299473e-10, + "loss": 0.0098, + "step": 8680 + }, + { + "epoch": 4.98315049831505, + "grad_norm": 0.007077803369611502, + "learning_rate": 2.506984865263684e-10, + "loss": 0.0016, + "step": 8690 + }, + { + "epoch": 4.98888649888865, + "grad_norm": 0.1649414747953415, + "learning_rate": 9.025193784151232e-11, + "loss": 0.0041, + "step": 8700 + }, + { + "epoch": 4.98888649888865, + "eval_thought_eval_raw_data_english_72b_clean_1117_loss": 0.09170719236135483, + "eval_thought_eval_raw_data_english_72b_clean_1117_runtime": 21.1606, + "eval_thought_eval_raw_data_english_72b_clean_1117_samples_per_second": 30.859, + "eval_thought_eval_raw_data_english_72b_clean_1117_steps_per_second": 15.453, + "step": 8700 + }, + { + "epoch": 4.99462249946225, + "grad_norm": 0.00366253312677145, + "learning_rate": 1.0028019910013342e-11, + "loss": 0.0032, + "step": 8710 + }, + { + "epoch": 4.99749049974905, + "step": 8715, + "total_flos": 6.758260289254195e+17, + "train_loss": 0.02162364056710329, + "train_runtime": 21813.1277, + "train_samples_per_second": 6.394, + "train_steps_per_second": 0.4 + } + ], + "logging_steps": 10, + "max_steps": 8715, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 250, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 6.758260289254195e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}