| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.6597394029358403, | |
| "eval_steps": 10, | |
| "global_step": 1000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.006597394029358403, | |
| "grad_norm": 1.2868798971176147, | |
| "learning_rate": 3.2894736842105265e-06, | |
| "loss": 4.6934, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.006597394029358403, | |
| "eval_loss": 4.687362194061279, | |
| "eval_runtime": 282.9655, | |
| "eval_samples_per_second": 18.045, | |
| "eval_steps_per_second": 0.565, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.013194788058716806, | |
| "grad_norm": 1.302221417427063, | |
| "learning_rate": 6.578947368421053e-06, | |
| "loss": 4.6651, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.013194788058716806, | |
| "eval_loss": 4.648743152618408, | |
| "eval_runtime": 282.9326, | |
| "eval_samples_per_second": 18.047, | |
| "eval_steps_per_second": 0.566, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01979218208807521, | |
| "grad_norm": 1.4535945653915405, | |
| "learning_rate": 9.868421052631579e-06, | |
| "loss": 4.6063, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01979218208807521, | |
| "eval_loss": 4.54951286315918, | |
| "eval_runtime": 282.3584, | |
| "eval_samples_per_second": 18.083, | |
| "eval_steps_per_second": 0.567, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.026389576117433613, | |
| "grad_norm": 1.714498519897461, | |
| "learning_rate": 1.3157894736842106e-05, | |
| "loss": 4.4527, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.026389576117433613, | |
| "eval_loss": 4.304062366485596, | |
| "eval_runtime": 282.5935, | |
| "eval_samples_per_second": 18.068, | |
| "eval_steps_per_second": 0.566, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.03298697014679202, | |
| "grad_norm": 1.832472324371338, | |
| "learning_rate": 1.6447368421052635e-05, | |
| "loss": 4.1216, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03298697014679202, | |
| "eval_loss": 3.8426761627197266, | |
| "eval_runtime": 281.3902, | |
| "eval_samples_per_second": 18.146, | |
| "eval_steps_per_second": 0.569, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03958436417615042, | |
| "grad_norm": 1.1160222291946411, | |
| "learning_rate": 1.9736842105263158e-05, | |
| "loss": 3.6561, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03958436417615042, | |
| "eval_loss": 3.4654476642608643, | |
| "eval_runtime": 282.2746, | |
| "eval_samples_per_second": 18.089, | |
| "eval_steps_per_second": 0.567, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.04618175820550882, | |
| "grad_norm": 1.1160507202148438, | |
| "learning_rate": 2.3026315789473685e-05, | |
| "loss": 3.2993, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.04618175820550882, | |
| "eval_loss": 3.0560474395751953, | |
| "eval_runtime": 282.4864, | |
| "eval_samples_per_second": 18.075, | |
| "eval_steps_per_second": 0.566, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.052779152234867226, | |
| "grad_norm": 0.7468234896659851, | |
| "learning_rate": 2.6315789473684212e-05, | |
| "loss": 2.8777, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.052779152234867226, | |
| "eval_loss": 2.7082414627075195, | |
| "eval_runtime": 282.279, | |
| "eval_samples_per_second": 18.088, | |
| "eval_steps_per_second": 0.567, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.05937654626422563, | |
| "grad_norm": 0.7456527948379517, | |
| "learning_rate": 2.9605263157894735e-05, | |
| "loss": 2.5844, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.05937654626422563, | |
| "eval_loss": 2.400399923324585, | |
| "eval_runtime": 282.4234, | |
| "eval_samples_per_second": 18.079, | |
| "eval_steps_per_second": 0.567, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.06597394029358404, | |
| "grad_norm": 0.7843755483627319, | |
| "learning_rate": 3.289473684210527e-05, | |
| "loss": 2.2405, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.06597394029358404, | |
| "eval_loss": 2.043464422225952, | |
| "eval_runtime": 282.4971, | |
| "eval_samples_per_second": 18.075, | |
| "eval_steps_per_second": 0.566, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.07257133432294244, | |
| "grad_norm": 0.7848460078239441, | |
| "learning_rate": 3.618421052631579e-05, | |
| "loss": 1.8797, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.07257133432294244, | |
| "eval_loss": 1.6749211549758911, | |
| "eval_runtime": 282.4596, | |
| "eval_samples_per_second": 18.077, | |
| "eval_steps_per_second": 0.566, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.07916872835230084, | |
| "grad_norm": 0.8517773151397705, | |
| "learning_rate": 3.9473684210526316e-05, | |
| "loss": 1.5041, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.07916872835230084, | |
| "eval_loss": 1.322484016418457, | |
| "eval_runtime": 282.425, | |
| "eval_samples_per_second": 18.079, | |
| "eval_steps_per_second": 0.567, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.08576612238165925, | |
| "grad_norm": 0.924010694026947, | |
| "learning_rate": 4.2763157894736847e-05, | |
| "loss": 1.1712, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.08576612238165925, | |
| "eval_loss": 1.0092010498046875, | |
| "eval_runtime": 282.4135, | |
| "eval_samples_per_second": 18.08, | |
| "eval_steps_per_second": 0.567, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.09236351641101764, | |
| "grad_norm": 0.599291205406189, | |
| "learning_rate": 4.605263157894737e-05, | |
| "loss": 0.9068, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.09236351641101764, | |
| "eval_loss": 0.8262147307395935, | |
| "eval_runtime": 282.3089, | |
| "eval_samples_per_second": 18.087, | |
| "eval_steps_per_second": 0.567, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.09896091044037605, | |
| "grad_norm": 0.4690685570240021, | |
| "learning_rate": 4.9342105263157894e-05, | |
| "loss": 0.7821, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.09896091044037605, | |
| "eval_loss": 0.7165087461471558, | |
| "eval_runtime": 282.1805, | |
| "eval_samples_per_second": 18.095, | |
| "eval_steps_per_second": 0.567, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.10555830446973445, | |
| "grad_norm": 0.5129504203796387, | |
| "learning_rate": 4.9706529713866475e-05, | |
| "loss": 0.6867, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.10555830446973445, | |
| "eval_loss": 0.6483914256095886, | |
| "eval_runtime": 282.207, | |
| "eval_samples_per_second": 18.093, | |
| "eval_steps_per_second": 0.567, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.11215569849909286, | |
| "grad_norm": 0.4116060435771942, | |
| "learning_rate": 4.933969185619956e-05, | |
| "loss": 0.6315, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.11215569849909286, | |
| "eval_loss": 0.6029236912727356, | |
| "eval_runtime": 282.3594, | |
| "eval_samples_per_second": 18.083, | |
| "eval_steps_per_second": 0.567, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.11875309252845126, | |
| "grad_norm": 0.4301941990852356, | |
| "learning_rate": 4.8972853998532655e-05, | |
| "loss": 0.5833, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.11875309252845126, | |
| "eval_loss": 0.5720704197883606, | |
| "eval_runtime": 282.0439, | |
| "eval_samples_per_second": 18.104, | |
| "eval_steps_per_second": 0.567, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.12535048655780967, | |
| "grad_norm": 0.5514675378799438, | |
| "learning_rate": 4.860601614086574e-05, | |
| "loss": 0.5667, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.12535048655780967, | |
| "eval_loss": 0.5490744709968567, | |
| "eval_runtime": 282.3239, | |
| "eval_samples_per_second": 18.086, | |
| "eval_steps_per_second": 0.567, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.13194788058716808, | |
| "grad_norm": 0.4441012144088745, | |
| "learning_rate": 4.823917828319883e-05, | |
| "loss": 0.5303, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.13194788058716808, | |
| "eval_loss": 0.5291892886161804, | |
| "eval_runtime": 282.4233, | |
| "eval_samples_per_second": 18.079, | |
| "eval_steps_per_second": 0.567, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.13854527461652646, | |
| "grad_norm": 0.4900895059108734, | |
| "learning_rate": 4.787234042553192e-05, | |
| "loss": 0.5109, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.13854527461652646, | |
| "eval_loss": 0.5132142901420593, | |
| "eval_runtime": 282.4262, | |
| "eval_samples_per_second": 18.079, | |
| "eval_steps_per_second": 0.567, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.14514266864588488, | |
| "grad_norm": 0.4287125766277313, | |
| "learning_rate": 4.750550256786501e-05, | |
| "loss": 0.5109, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.14514266864588488, | |
| "eval_loss": 0.4994109869003296, | |
| "eval_runtime": 282.5477, | |
| "eval_samples_per_second": 18.071, | |
| "eval_steps_per_second": 0.566, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1517400626752433, | |
| "grad_norm": 0.4390946626663208, | |
| "learning_rate": 4.713866471019809e-05, | |
| "loss": 0.4952, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.1517400626752433, | |
| "eval_loss": 0.4881007969379425, | |
| "eval_runtime": 282.5158, | |
| "eval_samples_per_second": 18.073, | |
| "eval_steps_per_second": 0.566, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.15833745670460167, | |
| "grad_norm": 0.5645927786827087, | |
| "learning_rate": 4.677182685253118e-05, | |
| "loss": 0.488, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.15833745670460167, | |
| "eval_loss": 0.4788074791431427, | |
| "eval_runtime": 282.6337, | |
| "eval_samples_per_second": 18.066, | |
| "eval_steps_per_second": 0.566, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.16493485073396008, | |
| "grad_norm": 0.5930696129798889, | |
| "learning_rate": 4.6404988994864274e-05, | |
| "loss": 0.4833, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.16493485073396008, | |
| "eval_loss": 0.4700024425983429, | |
| "eval_runtime": 282.5674, | |
| "eval_samples_per_second": 18.07, | |
| "eval_steps_per_second": 0.566, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.1715322447633185, | |
| "grad_norm": 0.408792644739151, | |
| "learning_rate": 4.6038151137197364e-05, | |
| "loss": 0.4696, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.1715322447633185, | |
| "eval_loss": 0.46224626898765564, | |
| "eval_runtime": 282.2108, | |
| "eval_samples_per_second": 18.093, | |
| "eval_steps_per_second": 0.567, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.1781296387926769, | |
| "grad_norm": 0.5276111960411072, | |
| "learning_rate": 4.567131327953045e-05, | |
| "loss": 0.4579, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.1781296387926769, | |
| "eval_loss": 0.455151230096817, | |
| "eval_runtime": 282.6419, | |
| "eval_samples_per_second": 18.065, | |
| "eval_steps_per_second": 0.566, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.1847270328220353, | |
| "grad_norm": 0.5726278424263, | |
| "learning_rate": 4.530447542186354e-05, | |
| "loss": 0.4547, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.1847270328220353, | |
| "eval_loss": 0.44886669516563416, | |
| "eval_runtime": 282.6094, | |
| "eval_samples_per_second": 18.067, | |
| "eval_steps_per_second": 0.566, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.1913244268513937, | |
| "grad_norm": 0.4161057770252228, | |
| "learning_rate": 4.493763756419663e-05, | |
| "loss": 0.4513, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1913244268513937, | |
| "eval_loss": 0.44479382038116455, | |
| "eval_runtime": 282.5015, | |
| "eval_samples_per_second": 18.074, | |
| "eval_steps_per_second": 0.566, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1979218208807521, | |
| "grad_norm": 0.4173205494880676, | |
| "learning_rate": 4.457079970652971e-05, | |
| "loss": 0.4364, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.1979218208807521, | |
| "eval_loss": 0.4391051232814789, | |
| "eval_runtime": 282.6206, | |
| "eval_samples_per_second": 18.067, | |
| "eval_steps_per_second": 0.566, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.20451921491011052, | |
| "grad_norm": 0.4726370871067047, | |
| "learning_rate": 4.420396184886281e-05, | |
| "loss": 0.4413, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.20451921491011052, | |
| "eval_loss": 0.43503668904304504, | |
| "eval_runtime": 282.4681, | |
| "eval_samples_per_second": 18.076, | |
| "eval_steps_per_second": 0.566, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.2111166089394689, | |
| "grad_norm": 0.528791069984436, | |
| "learning_rate": 4.383712399119589e-05, | |
| "loss": 0.4347, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.2111166089394689, | |
| "eval_loss": 0.4299105107784271, | |
| "eval_runtime": 282.3837, | |
| "eval_samples_per_second": 18.082, | |
| "eval_steps_per_second": 0.567, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.2177140029688273, | |
| "grad_norm": 0.4637807607650757, | |
| "learning_rate": 4.347028613352898e-05, | |
| "loss": 0.4263, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.2177140029688273, | |
| "eval_loss": 0.42648786306381226, | |
| "eval_runtime": 282.4045, | |
| "eval_samples_per_second": 18.08, | |
| "eval_steps_per_second": 0.567, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.22431139699818572, | |
| "grad_norm": 0.5652016401290894, | |
| "learning_rate": 4.3103448275862066e-05, | |
| "loss": 0.424, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.22431139699818572, | |
| "eval_loss": 0.42295747995376587, | |
| "eval_runtime": 282.132, | |
| "eval_samples_per_second": 18.098, | |
| "eval_steps_per_second": 0.567, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.2309087910275441, | |
| "grad_norm": 0.5480278134346008, | |
| "learning_rate": 4.273661041819516e-05, | |
| "loss": 0.4246, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.2309087910275441, | |
| "eval_loss": 0.4191484749317169, | |
| "eval_runtime": 282.8434, | |
| "eval_samples_per_second": 18.052, | |
| "eval_steps_per_second": 0.566, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.23750618505690252, | |
| "grad_norm": 0.5954585671424866, | |
| "learning_rate": 4.2369772560528246e-05, | |
| "loss": 0.4194, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.23750618505690252, | |
| "eval_loss": 0.4160284399986267, | |
| "eval_runtime": 282.0609, | |
| "eval_samples_per_second": 18.102, | |
| "eval_steps_per_second": 0.567, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.24410357908626093, | |
| "grad_norm": 0.6444059610366821, | |
| "learning_rate": 4.2002934702861336e-05, | |
| "loss": 0.4151, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.24410357908626093, | |
| "eval_loss": 0.4145909547805786, | |
| "eval_runtime": 282.2951, | |
| "eval_samples_per_second": 18.087, | |
| "eval_steps_per_second": 0.567, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.25070097311561934, | |
| "grad_norm": 0.5506740212440491, | |
| "learning_rate": 4.163609684519443e-05, | |
| "loss": 0.4143, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.25070097311561934, | |
| "eval_loss": 0.41122376918792725, | |
| "eval_runtime": 282.3323, | |
| "eval_samples_per_second": 18.085, | |
| "eval_steps_per_second": 0.567, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.2572983671449777, | |
| "grad_norm": 0.5273014307022095, | |
| "learning_rate": 4.126925898752752e-05, | |
| "loss": 0.4059, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2572983671449777, | |
| "eval_loss": 0.41092923283576965, | |
| "eval_runtime": 282.2453, | |
| "eval_samples_per_second": 18.091, | |
| "eval_steps_per_second": 0.567, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.26389576117433616, | |
| "grad_norm": 0.5662177205085754, | |
| "learning_rate": 4.09024211298606e-05, | |
| "loss": 0.411, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.26389576117433616, | |
| "eval_loss": 0.40584176778793335, | |
| "eval_runtime": 282.2166, | |
| "eval_samples_per_second": 18.092, | |
| "eval_steps_per_second": 0.567, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.27049315520369455, | |
| "grad_norm": 0.6541048884391785, | |
| "learning_rate": 4.05355832721937e-05, | |
| "loss": 0.4017, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.27049315520369455, | |
| "eval_loss": 0.4048697054386139, | |
| "eval_runtime": 282.3287, | |
| "eval_samples_per_second": 18.085, | |
| "eval_steps_per_second": 0.567, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.27709054923305293, | |
| "grad_norm": 0.490032821893692, | |
| "learning_rate": 4.016874541452678e-05, | |
| "loss": 0.3986, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.27709054923305293, | |
| "eval_loss": 0.4038483202457428, | |
| "eval_runtime": 282.2547, | |
| "eval_samples_per_second": 18.09, | |
| "eval_steps_per_second": 0.567, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.28368794326241137, | |
| "grad_norm": 0.4884147644042969, | |
| "learning_rate": 3.980190755685987e-05, | |
| "loss": 0.4003, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.28368794326241137, | |
| "eval_loss": 0.40054452419281006, | |
| "eval_runtime": 282.2046, | |
| "eval_samples_per_second": 18.093, | |
| "eval_steps_per_second": 0.567, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.29028533729176975, | |
| "grad_norm": 0.5234985947608948, | |
| "learning_rate": 3.943506969919296e-05, | |
| "loss": 0.4005, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.29028533729176975, | |
| "eval_loss": 0.39862698316574097, | |
| "eval_runtime": 282.1876, | |
| "eval_samples_per_second": 18.094, | |
| "eval_steps_per_second": 0.567, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.29688273132112813, | |
| "grad_norm": 0.512843668460846, | |
| "learning_rate": 3.9068231841526045e-05, | |
| "loss": 0.3952, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.29688273132112813, | |
| "eval_loss": 0.3970402181148529, | |
| "eval_runtime": 282.2615, | |
| "eval_samples_per_second": 18.09, | |
| "eval_steps_per_second": 0.567, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.3034801253504866, | |
| "grad_norm": 0.46752044558525085, | |
| "learning_rate": 3.8701393983859135e-05, | |
| "loss": 0.3967, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.3034801253504866, | |
| "eval_loss": 0.39410126209259033, | |
| "eval_runtime": 282.2669, | |
| "eval_samples_per_second": 18.089, | |
| "eval_steps_per_second": 0.567, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.31007751937984496, | |
| "grad_norm": 0.49608954787254333, | |
| "learning_rate": 3.8334556126192226e-05, | |
| "loss": 0.3893, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.31007751937984496, | |
| "eval_loss": 0.3944699168205261, | |
| "eval_runtime": 282.3461, | |
| "eval_samples_per_second": 18.084, | |
| "eval_steps_per_second": 0.567, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.31667491340920334, | |
| "grad_norm": 0.4948144853115082, | |
| "learning_rate": 3.7967718268525316e-05, | |
| "loss": 0.3883, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.31667491340920334, | |
| "eval_loss": 0.3926471471786499, | |
| "eval_runtime": 282.4921, | |
| "eval_samples_per_second": 18.075, | |
| "eval_steps_per_second": 0.566, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.3232723074385618, | |
| "grad_norm": 0.5696155428886414, | |
| "learning_rate": 3.76008804108584e-05, | |
| "loss": 0.3921, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.3232723074385618, | |
| "eval_loss": 0.39166340231895447, | |
| "eval_runtime": 282.2989, | |
| "eval_samples_per_second": 18.087, | |
| "eval_steps_per_second": 0.567, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.32986970146792016, | |
| "grad_norm": 0.5400599241256714, | |
| "learning_rate": 3.723404255319149e-05, | |
| "loss": 0.3927, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.32986970146792016, | |
| "eval_loss": 0.3899206817150116, | |
| "eval_runtime": 282.6164, | |
| "eval_samples_per_second": 18.067, | |
| "eval_steps_per_second": 0.566, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.3364670954972786, | |
| "grad_norm": 0.5328640937805176, | |
| "learning_rate": 3.686720469552458e-05, | |
| "loss": 0.3893, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.3364670954972786, | |
| "eval_loss": 0.388072669506073, | |
| "eval_runtime": 282.6361, | |
| "eval_samples_per_second": 18.066, | |
| "eval_steps_per_second": 0.566, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.343064489526637, | |
| "grad_norm": 0.6197268962860107, | |
| "learning_rate": 3.650036683785767e-05, | |
| "loss": 0.3896, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.343064489526637, | |
| "eval_loss": 0.38577401638031006, | |
| "eval_runtime": 282.5926, | |
| "eval_samples_per_second": 18.068, | |
| "eval_steps_per_second": 0.566, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.34966188355599537, | |
| "grad_norm": 0.5348854064941406, | |
| "learning_rate": 3.6133528980190754e-05, | |
| "loss": 0.3824, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.34966188355599537, | |
| "eval_loss": 0.38608622550964355, | |
| "eval_runtime": 282.758, | |
| "eval_samples_per_second": 18.058, | |
| "eval_steps_per_second": 0.566, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.3562592775853538, | |
| "grad_norm": 0.46607691049575806, | |
| "learning_rate": 3.576669112252385e-05, | |
| "loss": 0.39, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.3562592775853538, | |
| "eval_loss": 0.3843710720539093, | |
| "eval_runtime": 282.6255, | |
| "eval_samples_per_second": 18.066, | |
| "eval_steps_per_second": 0.566, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.3628566716147122, | |
| "grad_norm": 0.5211663246154785, | |
| "learning_rate": 3.5399853264856934e-05, | |
| "loss": 0.3831, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.3628566716147122, | |
| "eval_loss": 0.38406872749328613, | |
| "eval_runtime": 282.6083, | |
| "eval_samples_per_second": 18.067, | |
| "eval_steps_per_second": 0.566, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.3694540656440706, | |
| "grad_norm": 0.4613727927207947, | |
| "learning_rate": 3.5033015407190025e-05, | |
| "loss": 0.3848, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3694540656440706, | |
| "eval_loss": 0.38266727328300476, | |
| "eval_runtime": 282.5681, | |
| "eval_samples_per_second": 18.07, | |
| "eval_steps_per_second": 0.566, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.376051459673429, | |
| "grad_norm": 0.4561725854873657, | |
| "learning_rate": 3.466617754952311e-05, | |
| "loss": 0.3771, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.376051459673429, | |
| "eval_loss": 0.38093405961990356, | |
| "eval_runtime": 282.579, | |
| "eval_samples_per_second": 18.069, | |
| "eval_steps_per_second": 0.566, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.3826488537027874, | |
| "grad_norm": 0.5360392928123474, | |
| "learning_rate": 3.4299339691856205e-05, | |
| "loss": 0.3819, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.3826488537027874, | |
| "eval_loss": 0.380355566740036, | |
| "eval_runtime": 282.5981, | |
| "eval_samples_per_second": 18.068, | |
| "eval_steps_per_second": 0.566, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.3892462477321458, | |
| "grad_norm": 0.6373459696769714, | |
| "learning_rate": 3.393250183418929e-05, | |
| "loss": 0.3773, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.3892462477321458, | |
| "eval_loss": 0.37904950976371765, | |
| "eval_runtime": 282.4657, | |
| "eval_samples_per_second": 18.077, | |
| "eval_steps_per_second": 0.566, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.3958436417615042, | |
| "grad_norm": 0.4231775403022766, | |
| "learning_rate": 3.356566397652238e-05, | |
| "loss": 0.3736, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.3958436417615042, | |
| "eval_loss": 0.37867945432662964, | |
| "eval_runtime": 282.479, | |
| "eval_samples_per_second": 18.076, | |
| "eval_steps_per_second": 0.566, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.4024410357908626, | |
| "grad_norm": 0.5430477261543274, | |
| "learning_rate": 3.319882611885547e-05, | |
| "loss": 0.3778, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.4024410357908626, | |
| "eval_loss": 0.3770335912704468, | |
| "eval_runtime": 282.4938, | |
| "eval_samples_per_second": 18.075, | |
| "eval_steps_per_second": 0.566, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.40903842982022104, | |
| "grad_norm": 0.5898250341415405, | |
| "learning_rate": 3.283198826118856e-05, | |
| "loss": 0.3783, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.40903842982022104, | |
| "eval_loss": 0.37556448578834534, | |
| "eval_runtime": 282.4757, | |
| "eval_samples_per_second": 18.076, | |
| "eval_steps_per_second": 0.566, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.4156358238495794, | |
| "grad_norm": 0.5367993116378784, | |
| "learning_rate": 3.246515040352164e-05, | |
| "loss": 0.3726, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.4156358238495794, | |
| "eval_loss": 0.37541213631629944, | |
| "eval_runtime": 282.4621, | |
| "eval_samples_per_second": 18.077, | |
| "eval_steps_per_second": 0.566, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.4222332178789378, | |
| "grad_norm": 0.5479997992515564, | |
| "learning_rate": 3.209831254585473e-05, | |
| "loss": 0.3747, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.4222332178789378, | |
| "eval_loss": 0.37338846921920776, | |
| "eval_runtime": 281.8183, | |
| "eval_samples_per_second": 18.118, | |
| "eval_steps_per_second": 0.568, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.42883061190829624, | |
| "grad_norm": 0.5168615579605103, | |
| "learning_rate": 3.1731474688187823e-05, | |
| "loss": 0.3748, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.42883061190829624, | |
| "eval_loss": 0.373104453086853, | |
| "eval_runtime": 282.3532, | |
| "eval_samples_per_second": 18.084, | |
| "eval_steps_per_second": 0.567, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.4354280059376546, | |
| "grad_norm": 0.4789239764213562, | |
| "learning_rate": 3.1364636830520914e-05, | |
| "loss": 0.3725, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.4354280059376546, | |
| "eval_loss": 0.37169981002807617, | |
| "eval_runtime": 282.5035, | |
| "eval_samples_per_second": 18.074, | |
| "eval_steps_per_second": 0.566, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.442025399967013, | |
| "grad_norm": 0.5288018584251404, | |
| "learning_rate": 3.0997798972854e-05, | |
| "loss": 0.3715, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.442025399967013, | |
| "eval_loss": 0.3723226487636566, | |
| "eval_runtime": 282.4315, | |
| "eval_samples_per_second": 18.079, | |
| "eval_steps_per_second": 0.567, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.44862279399637145, | |
| "grad_norm": 0.5399067997932434, | |
| "learning_rate": 3.063096111518709e-05, | |
| "loss": 0.3672, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.44862279399637145, | |
| "eval_loss": 0.3709748387336731, | |
| "eval_runtime": 282.3214, | |
| "eval_samples_per_second": 18.086, | |
| "eval_steps_per_second": 0.567, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.45522018802572983, | |
| "grad_norm": 0.5560658574104309, | |
| "learning_rate": 3.0264123257520178e-05, | |
| "loss": 0.3703, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.45522018802572983, | |
| "eval_loss": 0.3694921135902405, | |
| "eval_runtime": 282.225, | |
| "eval_samples_per_second": 18.092, | |
| "eval_steps_per_second": 0.567, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.4618175820550882, | |
| "grad_norm": 0.46738743782043457, | |
| "learning_rate": 2.9897285399853265e-05, | |
| "loss": 0.3691, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.4618175820550882, | |
| "eval_loss": 0.36861899495124817, | |
| "eval_runtime": 281.4833, | |
| "eval_samples_per_second": 18.14, | |
| "eval_steps_per_second": 0.568, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.46841497608444665, | |
| "grad_norm": 0.5636041760444641, | |
| "learning_rate": 2.953044754218636e-05, | |
| "loss": 0.3674, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.46841497608444665, | |
| "eval_loss": 0.3687904477119446, | |
| "eval_runtime": 281.7696, | |
| "eval_samples_per_second": 18.121, | |
| "eval_steps_per_second": 0.568, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.47501237011380504, | |
| "grad_norm": 0.6417700052261353, | |
| "learning_rate": 2.9163609684519445e-05, | |
| "loss": 0.3676, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.47501237011380504, | |
| "eval_loss": 0.36721163988113403, | |
| "eval_runtime": 281.7105, | |
| "eval_samples_per_second": 18.125, | |
| "eval_steps_per_second": 0.568, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.4816097641431635, | |
| "grad_norm": 0.5903835892677307, | |
| "learning_rate": 2.8796771826852532e-05, | |
| "loss": 0.3687, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.4816097641431635, | |
| "eval_loss": 0.36735713481903076, | |
| "eval_runtime": 280.4935, | |
| "eval_samples_per_second": 18.204, | |
| "eval_steps_per_second": 0.57, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.48820715817252186, | |
| "grad_norm": 0.6300948858261108, | |
| "learning_rate": 2.8429933969185622e-05, | |
| "loss": 0.3649, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.48820715817252186, | |
| "eval_loss": 0.3654569089412689, | |
| "eval_runtime": 282.5933, | |
| "eval_samples_per_second": 18.068, | |
| "eval_steps_per_second": 0.566, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.49480455220188024, | |
| "grad_norm": 0.5772879123687744, | |
| "learning_rate": 2.806309611151871e-05, | |
| "loss": 0.3636, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.49480455220188024, | |
| "eval_loss": 0.3652118444442749, | |
| "eval_runtime": 281.9717, | |
| "eval_samples_per_second": 18.108, | |
| "eval_steps_per_second": 0.567, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.5014019462312387, | |
| "grad_norm": 0.46134456992149353, | |
| "learning_rate": 2.7696258253851796e-05, | |
| "loss": 0.3626, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.5014019462312387, | |
| "eval_loss": 0.364380419254303, | |
| "eval_runtime": 281.1459, | |
| "eval_samples_per_second": 18.161, | |
| "eval_steps_per_second": 0.569, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.507999340260597, | |
| "grad_norm": 0.5804652571678162, | |
| "learning_rate": 2.7329420396184886e-05, | |
| "loss": 0.3611, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.507999340260597, | |
| "eval_loss": 0.3632882237434387, | |
| "eval_runtime": 281.5379, | |
| "eval_samples_per_second": 18.136, | |
| "eval_steps_per_second": 0.568, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.5145967342899554, | |
| "grad_norm": 0.45915406942367554, | |
| "learning_rate": 2.6962582538517977e-05, | |
| "loss": 0.3599, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.5145967342899554, | |
| "eval_loss": 0.3634037375450134, | |
| "eval_runtime": 281.4618, | |
| "eval_samples_per_second": 18.141, | |
| "eval_steps_per_second": 0.568, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.5211941283193139, | |
| "grad_norm": 0.4998124837875366, | |
| "learning_rate": 2.6595744680851064e-05, | |
| "loss": 0.3619, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.5211941283193139, | |
| "eval_loss": 0.3615591526031494, | |
| "eval_runtime": 282.3789, | |
| "eval_samples_per_second": 18.082, | |
| "eval_steps_per_second": 0.567, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.5277915223486723, | |
| "grad_norm": 0.5479339361190796, | |
| "learning_rate": 2.622890682318415e-05, | |
| "loss": 0.3577, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.5277915223486723, | |
| "eval_loss": 0.3610810339450836, | |
| "eval_runtime": 282.2307, | |
| "eval_samples_per_second": 18.092, | |
| "eval_steps_per_second": 0.567, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.5343889163780307, | |
| "grad_norm": 0.5121034383773804, | |
| "learning_rate": 2.5862068965517244e-05, | |
| "loss": 0.3621, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.5343889163780307, | |
| "eval_loss": 0.3605840802192688, | |
| "eval_runtime": 281.2796, | |
| "eval_samples_per_second": 18.153, | |
| "eval_steps_per_second": 0.569, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.5409863104073891, | |
| "grad_norm": 0.5326396822929382, | |
| "learning_rate": 2.549523110785033e-05, | |
| "loss": 0.3575, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.5409863104073891, | |
| "eval_loss": 0.359646201133728, | |
| "eval_runtime": 281.2148, | |
| "eval_samples_per_second": 18.157, | |
| "eval_steps_per_second": 0.569, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.5475837044367475, | |
| "grad_norm": 0.5050413608551025, | |
| "learning_rate": 2.5128393250183418e-05, | |
| "loss": 0.3545, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.5475837044367475, | |
| "eval_loss": 0.3594898581504822, | |
| "eval_runtime": 281.2594, | |
| "eval_samples_per_second": 18.154, | |
| "eval_steps_per_second": 0.569, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.5541810984661059, | |
| "grad_norm": 0.45093920826911926, | |
| "learning_rate": 2.4761555392516508e-05, | |
| "loss": 0.3585, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.5541810984661059, | |
| "eval_loss": 0.35947272181510925, | |
| "eval_runtime": 280.5764, | |
| "eval_samples_per_second": 18.198, | |
| "eval_steps_per_second": 0.57, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.5607784924954643, | |
| "grad_norm": 0.44670212268829346, | |
| "learning_rate": 2.43947175348496e-05, | |
| "loss": 0.3569, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.5607784924954643, | |
| "eval_loss": 0.3579043745994568, | |
| "eval_runtime": 281.4427, | |
| "eval_samples_per_second": 18.142, | |
| "eval_steps_per_second": 0.568, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.5673758865248227, | |
| "grad_norm": 0.5422260761260986, | |
| "learning_rate": 2.4027879677182685e-05, | |
| "loss": 0.3525, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.5673758865248227, | |
| "eval_loss": 0.35847437381744385, | |
| "eval_runtime": 280.6628, | |
| "eval_samples_per_second": 18.193, | |
| "eval_steps_per_second": 0.57, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.5739732805541811, | |
| "grad_norm": 0.49470987915992737, | |
| "learning_rate": 2.3661041819515776e-05, | |
| "loss": 0.3528, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5739732805541811, | |
| "eval_loss": 0.35762321949005127, | |
| "eval_runtime": 281.4774, | |
| "eval_samples_per_second": 18.14, | |
| "eval_steps_per_second": 0.568, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5805706745835395, | |
| "grad_norm": 0.5388288497924805, | |
| "learning_rate": 2.3294203961848866e-05, | |
| "loss": 0.3533, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.5805706745835395, | |
| "eval_loss": 0.3578256666660309, | |
| "eval_runtime": 280.4896, | |
| "eval_samples_per_second": 18.204, | |
| "eval_steps_per_second": 0.57, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.5871680686128979, | |
| "grad_norm": 0.4280325174331665, | |
| "learning_rate": 2.2927366104181953e-05, | |
| "loss": 0.3574, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5871680686128979, | |
| "eval_loss": 0.35690200328826904, | |
| "eval_runtime": 280.7823, | |
| "eval_samples_per_second": 18.185, | |
| "eval_steps_per_second": 0.57, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5937654626422563, | |
| "grad_norm": 0.6243644952774048, | |
| "learning_rate": 2.2560528246515043e-05, | |
| "loss": 0.356, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.5937654626422563, | |
| "eval_loss": 0.3565099537372589, | |
| "eval_runtime": 280.4402, | |
| "eval_samples_per_second": 18.207, | |
| "eval_steps_per_second": 0.571, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.6003628566716147, | |
| "grad_norm": 0.5401405096054077, | |
| "learning_rate": 2.219369038884813e-05, | |
| "loss": 0.3594, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.6003628566716147, | |
| "eval_loss": 0.3559376895427704, | |
| "eval_runtime": 281.1004, | |
| "eval_samples_per_second": 18.164, | |
| "eval_steps_per_second": 0.569, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.6069602507009731, | |
| "grad_norm": 0.578926682472229, | |
| "learning_rate": 2.182685253118122e-05, | |
| "loss": 0.3542, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.6069602507009731, | |
| "eval_loss": 0.35559171438217163, | |
| "eval_runtime": 281.4745, | |
| "eval_samples_per_second": 18.14, | |
| "eval_steps_per_second": 0.568, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.6135576447303315, | |
| "grad_norm": 0.460857629776001, | |
| "learning_rate": 2.1460014673514307e-05, | |
| "loss": 0.3587, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.6135576447303315, | |
| "eval_loss": 0.35534095764160156, | |
| "eval_runtime": 280.8406, | |
| "eval_samples_per_second": 18.181, | |
| "eval_steps_per_second": 0.57, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.6201550387596899, | |
| "grad_norm": 0.5364406704902649, | |
| "learning_rate": 2.1093176815847397e-05, | |
| "loss": 0.353, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.6201550387596899, | |
| "eval_loss": 0.3539869785308838, | |
| "eval_runtime": 280.7515, | |
| "eval_samples_per_second": 18.187, | |
| "eval_steps_per_second": 0.57, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.6267524327890484, | |
| "grad_norm": 0.50699782371521, | |
| "learning_rate": 2.0726338958180484e-05, | |
| "loss": 0.3525, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.6267524327890484, | |
| "eval_loss": 0.35391008853912354, | |
| "eval_runtime": 281.3705, | |
| "eval_samples_per_second": 18.147, | |
| "eval_steps_per_second": 0.569, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.6333498268184067, | |
| "grad_norm": 0.6315603852272034, | |
| "learning_rate": 2.035950110051357e-05, | |
| "loss": 0.3546, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.6333498268184067, | |
| "eval_loss": 0.35362812876701355, | |
| "eval_runtime": 281.4224, | |
| "eval_samples_per_second": 18.144, | |
| "eval_steps_per_second": 0.569, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.6399472208477651, | |
| "grad_norm": 0.5252578258514404, | |
| "learning_rate": 1.999266324284666e-05, | |
| "loss": 0.3521, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.6399472208477651, | |
| "eval_loss": 0.3530423045158386, | |
| "eval_runtime": 280.5129, | |
| "eval_samples_per_second": 18.202, | |
| "eval_steps_per_second": 0.57, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.6465446148771236, | |
| "grad_norm": 0.5415360331535339, | |
| "learning_rate": 1.962582538517975e-05, | |
| "loss": 0.3543, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.6465446148771236, | |
| "eval_loss": 0.3528820276260376, | |
| "eval_runtime": 280.8145, | |
| "eval_samples_per_second": 18.183, | |
| "eval_steps_per_second": 0.57, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.6531420089064819, | |
| "grad_norm": 0.5389861464500427, | |
| "learning_rate": 1.925898752751284e-05, | |
| "loss": 0.3528, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.6531420089064819, | |
| "eval_loss": 0.3525787889957428, | |
| "eval_runtime": 280.6494, | |
| "eval_samples_per_second": 18.194, | |
| "eval_steps_per_second": 0.57, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.6597394029358403, | |
| "grad_norm": 0.47096070647239685, | |
| "learning_rate": 1.889214966984593e-05, | |
| "loss": 0.3514, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.6597394029358403, | |
| "eval_loss": 0.352372944355011, | |
| "eval_runtime": 280.6707, | |
| "eval_samples_per_second": 18.192, | |
| "eval_steps_per_second": 0.57, | |
| "step": 1000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1515, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.3527241133195264e+17, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |