{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 395, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.012658227848101266, "grad_norm": 6.695639272540547, "learning_rate": 5.000000000000001e-07, "loss": 1.1195, "step": 1 }, { "epoch": 0.06329113924050633, "grad_norm": 13.683866082279918, "learning_rate": 2.5e-06, "loss": 1.0968, "step": 5 }, { "epoch": 0.12658227848101267, "grad_norm": 2.7914110872997275, "learning_rate": 5e-06, "loss": 1.0868, "step": 10 }, { "epoch": 0.189873417721519, "grad_norm": 2.2869962494889995, "learning_rate": 7.500000000000001e-06, "loss": 1.0771, "step": 15 }, { "epoch": 0.25316455696202533, "grad_norm": 2.315595057148159, "learning_rate": 1e-05, "loss": 1.0546, "step": 20 }, { "epoch": 0.31645569620253167, "grad_norm": 1.9266353743480364, "learning_rate": 1.25e-05, "loss": 1.084, "step": 25 }, { "epoch": 0.379746835443038, "grad_norm": 2.150532379708014, "learning_rate": 1.5000000000000002e-05, "loss": 1.0654, "step": 30 }, { "epoch": 0.4430379746835443, "grad_norm": 2.1018953026935905, "learning_rate": 1.7500000000000002e-05, "loss": 1.0817, "step": 35 }, { "epoch": 0.5063291139240507, "grad_norm": 2.3849545878679272, "learning_rate": 2e-05, "loss": 1.0971, "step": 40 }, { "epoch": 0.569620253164557, "grad_norm": 2.345688862339495, "learning_rate": 1.9990212265199738e-05, "loss": 1.086, "step": 45 }, { "epoch": 0.6329113924050633, "grad_norm": 2.5866807131537906, "learning_rate": 1.996086822074945e-05, "loss": 1.0656, "step": 50 }, { "epoch": 0.6962025316455697, "grad_norm": 2.4392627537159655, "learning_rate": 1.9912025308994146e-05, "loss": 1.1059, "step": 55 }, { "epoch": 0.759493670886076, "grad_norm": 2.2706485443813276, "learning_rate": 1.9843779142227258e-05, "loss": 1.0973, "step": 60 }, { "epoch": 0.8227848101265823, "grad_norm": 2.3020097653483496, "learning_rate": 1.975626331552507e-05, "loss": 1.1083, "step": 65 }, { "epoch": 0.8860759493670886, "grad_norm": 2.0670888705685955, "learning_rate": 1.96496491452281e-05, "loss": 1.1039, "step": 70 }, { "epoch": 0.9493670886075949, "grad_norm": 2.371742516361207, "learning_rate": 1.9524145333581315e-05, "loss": 1.0895, "step": 75 }, { "epoch": 1.0, "eval_loss": 0.7223408222198486, "eval_runtime": 213.2984, "eval_samples_per_second": 47.375, "eval_steps_per_second": 0.741, "step": 79 }, { "epoch": 1.0126582278481013, "grad_norm": 8.106924593883994, "learning_rate": 1.9379997560189677e-05, "loss": 1.0198, "step": 80 }, { "epoch": 1.0759493670886076, "grad_norm": 3.486817193319529, "learning_rate": 1.9217488001088784e-05, "loss": 0.6812, "step": 85 }, { "epoch": 1.139240506329114, "grad_norm": 2.921219837737029, "learning_rate": 1.903693477637204e-05, "loss": 0.6683, "step": 90 }, { "epoch": 1.2025316455696202, "grad_norm": 3.1078224489011683, "learning_rate": 1.883869132745561e-05, "loss": 0.653, "step": 95 }, { "epoch": 1.2658227848101267, "grad_norm": 2.8280645620621874, "learning_rate": 1.862314572520028e-05, "loss": 0.6501, "step": 100 }, { "epoch": 1.3291139240506329, "grad_norm": 2.920193489057834, "learning_rate": 1.8390719910244487e-05, "loss": 0.647, "step": 105 }, { "epoch": 1.3924050632911391, "grad_norm": 2.647561794905359, "learning_rate": 1.8141868867035745e-05, "loss": 0.6531, "step": 110 }, { "epoch": 1.4556962025316456, "grad_norm": 2.7660605803019447, "learning_rate": 1.7877079733177185e-05, "loss": 0.6232, "step": 115 }, { "epoch": 1.518987341772152, "grad_norm": 2.497219995540892, "learning_rate": 1.759687084583285e-05, "loss": 0.6401, "step": 120 }, { "epoch": 1.5822784810126582, "grad_norm": 2.447623483254921, "learning_rate": 1.7301790727058344e-05, "loss": 0.6592, "step": 125 }, { "epoch": 1.6455696202531644, "grad_norm": 2.4783693483832563, "learning_rate": 1.6992417010043144e-05, "loss": 0.6441, "step": 130 }, { "epoch": 1.7088607594936709, "grad_norm": 2.348228472338975, "learning_rate": 1.666935530836651e-05, "loss": 0.6439, "step": 135 }, { "epoch": 1.7721518987341773, "grad_norm": 2.316917620320839, "learning_rate": 1.6333238030480473e-05, "loss": 0.6572, "step": 140 }, { "epoch": 1.8354430379746836, "grad_norm": 2.614503810192861, "learning_rate": 1.5984723141740578e-05, "loss": 0.6764, "step": 145 }, { "epoch": 1.8987341772151898, "grad_norm": 2.452736422029602, "learning_rate": 1.562449287640781e-05, "loss": 0.6555, "step": 150 }, { "epoch": 1.9620253164556962, "grad_norm": 2.737510300725156, "learning_rate": 1.5253252402142989e-05, "loss": 0.6454, "step": 155 }, { "epoch": 2.0, "eval_loss": 0.33173683285713196, "eval_runtime": 213.2106, "eval_samples_per_second": 47.394, "eval_steps_per_second": 0.741, "step": 158 }, { "epoch": 2.0253164556962027, "grad_norm": 3.9057198522311167, "learning_rate": 1.4871728439607967e-05, "loss": 0.5134, "step": 160 }, { "epoch": 2.088607594936709, "grad_norm": 3.4868328352161817, "learning_rate": 1.4480667839875786e-05, "loss": 0.3096, "step": 165 }, { "epoch": 2.151898734177215, "grad_norm": 2.9717493172755014, "learning_rate": 1.408083612243465e-05, "loss": 0.2909, "step": 170 }, { "epoch": 2.2151898734177213, "grad_norm": 2.642444961402327, "learning_rate": 1.367301597664757e-05, "loss": 0.2837, "step": 175 }, { "epoch": 2.278481012658228, "grad_norm": 2.3997899647768346, "learning_rate": 1.3258005729601178e-05, "loss": 0.2818, "step": 180 }, { "epoch": 2.3417721518987342, "grad_norm": 2.7423823494323774, "learning_rate": 1.2836617783342968e-05, "loss": 0.2907, "step": 185 }, { "epoch": 2.4050632911392404, "grad_norm": 2.3807259553082276, "learning_rate": 1.2409677024566145e-05, "loss": 0.2872, "step": 190 }, { "epoch": 2.4683544303797467, "grad_norm": 2.336572941961749, "learning_rate": 1.1978019209855174e-05, "loss": 0.2866, "step": 195 }, { "epoch": 2.5316455696202533, "grad_norm": 2.2601309613900757, "learning_rate": 1.1542489329653024e-05, "loss": 0.2945, "step": 200 }, { "epoch": 2.5949367088607596, "grad_norm": 2.3050926972713968, "learning_rate": 1.11039399541527e-05, "loss": 0.2937, "step": 205 }, { "epoch": 2.6582278481012658, "grad_norm": 2.221130883158219, "learning_rate": 1.066322956435104e-05, "loss": 0.2961, "step": 210 }, { "epoch": 2.721518987341772, "grad_norm": 2.270387701106379, "learning_rate": 1.022122087153187e-05, "loss": 0.2965, "step": 215 }, { "epoch": 2.7848101265822782, "grad_norm": 2.279056997659604, "learning_rate": 9.778779128468133e-06, "loss": 0.2964, "step": 220 }, { "epoch": 2.848101265822785, "grad_norm": 2.1417041666216696, "learning_rate": 9.336770435648963e-06, "loss": 0.2931, "step": 225 }, { "epoch": 2.911392405063291, "grad_norm": 2.439196726609028, "learning_rate": 8.896060045847305e-06, "loss": 0.2977, "step": 230 }, { "epoch": 2.9746835443037973, "grad_norm": 2.2397521488796417, "learning_rate": 8.457510670346976e-06, "loss": 0.2926, "step": 235 }, { "epoch": 3.0, "eval_loss": 0.12934371829032898, "eval_runtime": 213.237, "eval_samples_per_second": 47.389, "eval_steps_per_second": 0.741, "step": 237 }, { "epoch": 3.037974683544304, "grad_norm": 2.2704813832171142, "learning_rate": 8.021980790144828e-06, "loss": 0.199, "step": 240 }, { "epoch": 3.1012658227848102, "grad_norm": 2.132226843273487, "learning_rate": 7.590322975433857e-06, "loss": 0.1128, "step": 245 }, { "epoch": 3.1645569620253164, "grad_norm": 2.0782142465832236, "learning_rate": 7.163382216657033e-06, "loss": 0.1129, "step": 250 }, { "epoch": 3.2278481012658227, "grad_norm": 1.761669140987703, "learning_rate": 6.741994270398826e-06, "loss": 0.1099, "step": 255 }, { "epoch": 3.291139240506329, "grad_norm": 1.717634933034142, "learning_rate": 6.326984023352435e-06, "loss": 0.1086, "step": 260 }, { "epoch": 3.3544303797468356, "grad_norm": 1.6608157157785954, "learning_rate": 5.919163877565351e-06, "loss": 0.1068, "step": 265 }, { "epoch": 3.4177215189873418, "grad_norm": 1.6064024695135044, "learning_rate": 5.519332160124215e-06, "loss": 0.1101, "step": 270 }, { "epoch": 3.481012658227848, "grad_norm": 1.6349645931027217, "learning_rate": 5.128271560392037e-06, "loss": 0.1074, "step": 275 }, { "epoch": 3.5443037974683547, "grad_norm": 1.7679069294086043, "learning_rate": 4.746747597857014e-06, "loss": 0.1118, "step": 280 }, { "epoch": 3.607594936708861, "grad_norm": 1.5530405205049536, "learning_rate": 4.375507123592194e-06, "loss": 0.1077, "step": 285 }, { "epoch": 3.670886075949367, "grad_norm": 1.5871889705183277, "learning_rate": 4.015276858259427e-06, "loss": 0.1082, "step": 290 }, { "epoch": 3.7341772151898733, "grad_norm": 1.5985959451025067, "learning_rate": 3.6667619695195287e-06, "loss": 0.1063, "step": 295 }, { "epoch": 3.7974683544303796, "grad_norm": 1.567040797684349, "learning_rate": 3.330644691633492e-06, "loss": 0.1067, "step": 300 }, { "epoch": 3.8607594936708862, "grad_norm": 1.6098560275600409, "learning_rate": 3.00758298995686e-06, "loss": 0.1045, "step": 305 }, { "epoch": 3.9240506329113924, "grad_norm": 1.495424841878041, "learning_rate": 2.698209272941659e-06, "loss": 0.1076, "step": 310 }, { "epoch": 3.9873417721518987, "grad_norm": 1.4612235257248143, "learning_rate": 2.403129154167153e-06, "loss": 0.1048, "step": 315 }, { "epoch": 4.0, "eval_loss": 0.054249346256256104, "eval_runtime": 213.4038, "eval_samples_per_second": 47.352, "eval_steps_per_second": 0.74, "step": 316 }, { "epoch": 4.050632911392405, "grad_norm": 0.8970304833218358, "learning_rate": 2.1229202668228197e-06, "loss": 0.0643, "step": 320 }, { "epoch": 4.113924050632911, "grad_norm": 1.3031778574429476, "learning_rate": 1.8581311329642592e-06, "loss": 0.0496, "step": 325 }, { "epoch": 4.177215189873418, "grad_norm": 1.0494986965992847, "learning_rate": 1.609280089755515e-06, "loss": 0.0498, "step": 330 }, { "epoch": 4.2405063291139244, "grad_norm": 0.8871889072165642, "learning_rate": 1.3768542747997215e-06, "loss": 0.048, "step": 335 }, { "epoch": 4.30379746835443, "grad_norm": 0.8947823783464494, "learning_rate": 1.161308672544389e-06, "loss": 0.0491, "step": 340 }, { "epoch": 4.367088607594937, "grad_norm": 0.8681171283598373, "learning_rate": 9.630652236279626e-07, "loss": 0.0467, "step": 345 }, { "epoch": 4.430379746835443, "grad_norm": 0.8483515397273593, "learning_rate": 7.825119989112173e-07, "loss": 0.0474, "step": 350 }, { "epoch": 4.493670886075949, "grad_norm": 0.883073997111795, "learning_rate": 6.200024398103255e-07, "loss": 0.0461, "step": 355 }, { "epoch": 4.556962025316456, "grad_norm": 0.8423665781515933, "learning_rate": 4.7585466641868696e-07, "loss": 0.0474, "step": 360 }, { "epoch": 4.620253164556962, "grad_norm": 1.0038851569615406, "learning_rate": 3.5035085477190143e-07, "loss": 0.0476, "step": 365 }, { "epoch": 4.6835443037974684, "grad_norm": 0.9089402076181121, "learning_rate": 2.4373668447493225e-07, "loss": 0.0467, "step": 370 }, { "epoch": 4.746835443037975, "grad_norm": 0.8530919153878059, "learning_rate": 1.562208577727442e-07, "loss": 0.0469, "step": 375 }, { "epoch": 4.810126582278481, "grad_norm": 0.8298393668849166, "learning_rate": 8.797469100585432e-08, "loss": 0.0481, "step": 380 }, { "epoch": 4.8734177215189876, "grad_norm": 0.986632852960438, "learning_rate": 3.913177925055189e-08, "loss": 0.0471, "step": 385 }, { "epoch": 4.936708860759493, "grad_norm": 0.9075462948662969, "learning_rate": 9.78773480026396e-09, "loss": 0.0474, "step": 390 }, { "epoch": 5.0, "grad_norm": 0.7480159011876264, "learning_rate": 0.0, "loss": 0.0465, "step": 395 }, { "epoch": 5.0, "eval_loss": 0.039588429033756256, "eval_runtime": 213.3054, "eval_samples_per_second": 47.373, "eval_steps_per_second": 0.741, "step": 395 }, { "epoch": 5.0, "step": 395, "total_flos": 165305238159360.0, "train_loss": 0.43836132314386245, "train_runtime": 6288.8643, "train_samples_per_second": 8.034, "train_steps_per_second": 0.063 } ], "logging_steps": 5, "max_steps": 395, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 165305238159360.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }