| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.20179496622456752, | |
| "eval_steps": 500, | |
| "global_step": 30000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00033632494370761253, | |
| "grad_norm": 5.551640033721924, | |
| "learning_rate": 4.9999999999999996e-05, | |
| "loss": 40.646, | |
| "num_input_tokens_seen": 13107200, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.0006726498874152251, | |
| "grad_norm": 3.0933010578155518, | |
| "learning_rate": 9.999999999999999e-05, | |
| "loss": 35.6719, | |
| "num_input_tokens_seen": 26214400, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0010089748311228376, | |
| "grad_norm": 1.8445225954055786, | |
| "learning_rate": 0.00015, | |
| "loss": 31.7929, | |
| "num_input_tokens_seen": 39321600, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.0013452997748304501, | |
| "grad_norm": 3.410053253173828, | |
| "learning_rate": 0.00019999999999999998, | |
| "loss": 29.7717, | |
| "num_input_tokens_seen": 52428800, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.0016816247185380627, | |
| "grad_norm": 3.2875728607177734, | |
| "learning_rate": 0.00025, | |
| "loss": 28.0852, | |
| "num_input_tokens_seen": 65536000, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.002017949662245675, | |
| "grad_norm": 3.1337997913360596, | |
| "learning_rate": 0.0003, | |
| "loss": 26.681, | |
| "num_input_tokens_seen": 78643200, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.002354274605953288, | |
| "grad_norm": 3.2058675289154053, | |
| "learning_rate": 0.00035, | |
| "loss": 25.5272, | |
| "num_input_tokens_seen": 91750400, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.0026905995496609002, | |
| "grad_norm": 2.94515323638916, | |
| "learning_rate": 0.00039999999999999996, | |
| "loss": 24.5054, | |
| "num_input_tokens_seen": 104857600, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.003026924493368513, | |
| "grad_norm": 1.952689528465271, | |
| "learning_rate": 0.00045, | |
| "loss": 23.6095, | |
| "num_input_tokens_seen": 117964800, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.0033632494370761253, | |
| "grad_norm": 2.3968894481658936, | |
| "learning_rate": 0.0005, | |
| "loss": 22.7795, | |
| "num_input_tokens_seen": 131072000, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0033632494370761253, | |
| "eval_loss": 5.559806823730469, | |
| "eval_runtime": 144.2401, | |
| "eval_samples_per_second": 34.664, | |
| "eval_steps_per_second": 8.666, | |
| "num_input_tokens_seen": 131072000, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.003699574380783738, | |
| "grad_norm": 2.3432235717773438, | |
| "learning_rate": 0.0005499999999999999, | |
| "loss": 22.0823, | |
| "num_input_tokens_seen": 144179200, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.00403589932449135, | |
| "grad_norm": 2.5011863708496094, | |
| "learning_rate": 0.0006, | |
| "loss": 21.3173, | |
| "num_input_tokens_seen": 157286400, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.004372224268198963, | |
| "grad_norm": 2.038031816482544, | |
| "learning_rate": 0.0005999957181118445, | |
| "loss": 20.6294, | |
| "num_input_tokens_seen": 170393600, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.004708549211906576, | |
| "grad_norm": 1.8921018838882446, | |
| "learning_rate": 0.0005999828725696082, | |
| "loss": 19.8806, | |
| "num_input_tokens_seen": 183500800, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.005044874155614189, | |
| "grad_norm": 2.1825079917907715, | |
| "learning_rate": 0.0005999614637399793, | |
| "loss": 19.2902, | |
| "num_input_tokens_seen": 196608000, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.0053811990993218005, | |
| "grad_norm": 1.6200745105743408, | |
| "learning_rate": 0.0005999314922340923, | |
| "loss": 18.7553, | |
| "num_input_tokens_seen": 209715200, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.005717524043029413, | |
| "grad_norm": 1.4547795057296753, | |
| "learning_rate": 0.0005998929589075115, | |
| "loss": 18.3636, | |
| "num_input_tokens_seen": 222822400, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.006053848986737026, | |
| "grad_norm": 1.7445319890975952, | |
| "learning_rate": 0.0005998458648602063, | |
| "loss": 18.0002, | |
| "num_input_tokens_seen": 235929600, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.006390173930444639, | |
| "grad_norm": 1.5439883470535278, | |
| "learning_rate": 0.0005997902114365196, | |
| "loss": 17.6987, | |
| "num_input_tokens_seen": 249036800, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.006726498874152251, | |
| "grad_norm": 1.5094984769821167, | |
| "learning_rate": 0.0005997260002251293, | |
| "loss": 17.4367, | |
| "num_input_tokens_seen": 262144000, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.006726498874152251, | |
| "eval_loss": 4.2485198974609375, | |
| "eval_runtime": 143.1933, | |
| "eval_samples_per_second": 34.918, | |
| "eval_steps_per_second": 8.729, | |
| "num_input_tokens_seen": 262144000, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.007062823817859863, | |
| "grad_norm": 1.380916714668274, | |
| "learning_rate": 0.0005996532330590042, | |
| "loss": 17.1672, | |
| "num_input_tokens_seen": 275251200, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.007399148761567476, | |
| "grad_norm": 1.2637253999710083, | |
| "learning_rate": 0.0005995719120153497, | |
| "loss": 16.9309, | |
| "num_input_tokens_seen": 288358400, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.007735473705275089, | |
| "grad_norm": 1.6453403234481812, | |
| "learning_rate": 0.0005994820394155497, | |
| "loss": 16.7436, | |
| "num_input_tokens_seen": 301465600, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.0080717986489827, | |
| "grad_norm": 1.4340417385101318, | |
| "learning_rate": 0.0005993836178251009, | |
| "loss": 16.507, | |
| "num_input_tokens_seen": 314572800, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.008408123592690313, | |
| "grad_norm": 1.1437392234802246, | |
| "learning_rate": 0.0005992766500535377, | |
| "loss": 16.4345, | |
| "num_input_tokens_seen": 327680000, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.008744448536397926, | |
| "grad_norm": 1.3861935138702393, | |
| "learning_rate": 0.0005991611391543539, | |
| "loss": 16.2523, | |
| "num_input_tokens_seen": 340787200, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.009080773480105539, | |
| "grad_norm": 1.2690448760986328, | |
| "learning_rate": 0.0005990370884249146, | |
| "loss": 16.1004, | |
| "num_input_tokens_seen": 353894400, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.009417098423813152, | |
| "grad_norm": 1.2439777851104736, | |
| "learning_rate": 0.000598904501406362, | |
| "loss": 16.0162, | |
| "num_input_tokens_seen": 367001600, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.009753423367520764, | |
| "grad_norm": 1.1519834995269775, | |
| "learning_rate": 0.0005987633818835147, | |
| "loss": 15.8826, | |
| "num_input_tokens_seen": 380108800, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.010089748311228377, | |
| "grad_norm": 1.101974606513977, | |
| "learning_rate": 0.0005986137338847594, | |
| "loss": 15.7688, | |
| "num_input_tokens_seen": 393216000, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.010089748311228377, | |
| "eval_loss": 3.8665707111358643, | |
| "eval_runtime": 143.1117, | |
| "eval_samples_per_second": 34.938, | |
| "eval_steps_per_second": 8.734, | |
| "num_input_tokens_seen": 393216000, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.010426073254935988, | |
| "grad_norm": 1.3389586210250854, | |
| "learning_rate": 0.0005984555616819361, | |
| "loss": 15.6984, | |
| "num_input_tokens_seen": 406323200, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.010762398198643601, | |
| "grad_norm": 1.175673246383667, | |
| "learning_rate": 0.0005982888697902161, | |
| "loss": 15.6319, | |
| "num_input_tokens_seen": 419430400, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.011098723142351214, | |
| "grad_norm": 1.122758388519287, | |
| "learning_rate": 0.0005981136629679728, | |
| "loss": 15.4898, | |
| "num_input_tokens_seen": 432537600, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.011435048086058826, | |
| "grad_norm": 1.0009791851043701, | |
| "learning_rate": 0.0005979299462166464, | |
| "loss": 15.4399, | |
| "num_input_tokens_seen": 445644800, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.01177137302976644, | |
| "grad_norm": 1.0287221670150757, | |
| "learning_rate": 0.0005977377247806006, | |
| "loss": 15.3713, | |
| "num_input_tokens_seen": 458752000, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.012107697973474052, | |
| "grad_norm": 1.061454176902771, | |
| "learning_rate": 0.0005975370041469738, | |
| "loss": 15.266, | |
| "num_input_tokens_seen": 471859200, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.012444022917181665, | |
| "grad_norm": 1.2214291095733643, | |
| "learning_rate": 0.0005973277900455209, | |
| "loss": 15.2011, | |
| "num_input_tokens_seen": 484966400, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.012780347860889277, | |
| "grad_norm": 1.0759477615356445, | |
| "learning_rate": 0.0005971100884484513, | |
| "loss": 15.153, | |
| "num_input_tokens_seen": 498073600, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.013116672804596888, | |
| "grad_norm": 0.8617029190063477, | |
| "learning_rate": 0.0005968839055702578, | |
| "loss": 15.1029, | |
| "num_input_tokens_seen": 511180800, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.013452997748304501, | |
| "grad_norm": 1.063086748123169, | |
| "learning_rate": 0.0005966492478675384, | |
| "loss": 14.9894, | |
| "num_input_tokens_seen": 524288000, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.013452997748304501, | |
| "eval_loss": 3.6751515865325928, | |
| "eval_runtime": 143.2457, | |
| "eval_samples_per_second": 34.905, | |
| "eval_steps_per_second": 8.726, | |
| "num_input_tokens_seen": 524288000, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.013789322692012114, | |
| "grad_norm": 0.8963438868522644, | |
| "learning_rate": 0.000596406122038814, | |
| "loss": 14.9472, | |
| "num_input_tokens_seen": 537395200, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.014125647635719727, | |
| "grad_norm": 0.8694930672645569, | |
| "learning_rate": 0.0005961545350243351, | |
| "loss": 14.8887, | |
| "num_input_tokens_seen": 550502400, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.01446197257942734, | |
| "grad_norm": 0.9276862144470215, | |
| "learning_rate": 0.0005958944940058844, | |
| "loss": 14.8208, | |
| "num_input_tokens_seen": 563609600, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.014798297523134952, | |
| "grad_norm": 0.8817610144615173, | |
| "learning_rate": 0.0005956260064065727, | |
| "loss": 14.7679, | |
| "num_input_tokens_seen": 576716800, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.015134622466842565, | |
| "grad_norm": 0.888661801815033, | |
| "learning_rate": 0.0005953490798906257, | |
| "loss": 14.7253, | |
| "num_input_tokens_seen": 589824000, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.015470947410550178, | |
| "grad_norm": 0.8768919706344604, | |
| "learning_rate": 0.0005950637223631658, | |
| "loss": 14.6678, | |
| "num_input_tokens_seen": 602931200, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.01580727235425779, | |
| "grad_norm": 0.8688133955001831, | |
| "learning_rate": 0.0005947699419699865, | |
| "loss": 14.6422, | |
| "num_input_tokens_seen": 616038400, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.0161435972979654, | |
| "grad_norm": 0.8557626008987427, | |
| "learning_rate": 0.0005944677470973196, | |
| "loss": 14.6511, | |
| "num_input_tokens_seen": 629145600, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.016479922241673016, | |
| "grad_norm": 0.956565260887146, | |
| "learning_rate": 0.0005941571463715962, | |
| "loss": 14.5594, | |
| "num_input_tokens_seen": 642252800, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.016816247185380627, | |
| "grad_norm": 0.8760116100311279, | |
| "learning_rate": 0.0005938381486591999, | |
| "loss": 14.5031, | |
| "num_input_tokens_seen": 655360000, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.016816247185380627, | |
| "eval_loss": 3.5482449531555176, | |
| "eval_runtime": 143.9882, | |
| "eval_samples_per_second": 34.725, | |
| "eval_steps_per_second": 8.681, | |
| "num_input_tokens_seen": 655360000, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.01715257212908824, | |
| "grad_norm": 0.8938534259796143, | |
| "learning_rate": 0.0005935107630662145, | |
| "loss": 14.4733, | |
| "num_input_tokens_seen": 668467200, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.017488897072795852, | |
| "grad_norm": 0.8379454016685486, | |
| "learning_rate": 0.0005931749989381631, | |
| "loss": 14.386, | |
| "num_input_tokens_seen": 681574400, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.017825222016503463, | |
| "grad_norm": 0.7709890007972717, | |
| "learning_rate": 0.000592830865859742, | |
| "loss": 14.3883, | |
| "num_input_tokens_seen": 694681600, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.018161546960211078, | |
| "grad_norm": 0.8483361601829529, | |
| "learning_rate": 0.000592478373654547, | |
| "loss": 14.4122, | |
| "num_input_tokens_seen": 707788800, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.01849787190391869, | |
| "grad_norm": 0.8239767551422119, | |
| "learning_rate": 0.0005921175323847927, | |
| "loss": 14.3169, | |
| "num_input_tokens_seen": 720896000, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.018834196847626303, | |
| "grad_norm": 0.7901423573493958, | |
| "learning_rate": 0.0005917483523510252, | |
| "loss": 14.263, | |
| "num_input_tokens_seen": 734003200, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.019170521791333914, | |
| "grad_norm": 0.7838689088821411, | |
| "learning_rate": 0.0005913708440918291, | |
| "loss": 14.2589, | |
| "num_input_tokens_seen": 747110400, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.01950684673504153, | |
| "grad_norm": 0.90792316198349, | |
| "learning_rate": 0.000590985018383525, | |
| "loss": 14.2538, | |
| "num_input_tokens_seen": 760217600, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.01984317167874914, | |
| "grad_norm": 0.7609734535217285, | |
| "learning_rate": 0.0005905908862398632, | |
| "loss": 14.208, | |
| "num_input_tokens_seen": 773324800, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.020179496622456754, | |
| "grad_norm": 0.7796016335487366, | |
| "learning_rate": 0.0005901884589117088, | |
| "loss": 14.2405, | |
| "num_input_tokens_seen": 786432000, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.020179496622456754, | |
| "eval_loss": 3.469104766845703, | |
| "eval_runtime": 142.6639, | |
| "eval_samples_per_second": 35.047, | |
| "eval_steps_per_second": 8.762, | |
| "num_input_tokens_seen": 786432000, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.020515821566164365, | |
| "grad_norm": 0.7596002221107483, | |
| "learning_rate": 0.0005897777478867204, | |
| "loss": 14.1367, | |
| "num_input_tokens_seen": 799539200, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.020852146509871976, | |
| "grad_norm": 0.7429248094558716, | |
| "learning_rate": 0.0005893587648890227, | |
| "loss": 14.1394, | |
| "num_input_tokens_seen": 812646400, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.02118847145357959, | |
| "grad_norm": 0.8267149329185486, | |
| "learning_rate": 0.0005889315218788711, | |
| "loss": 14.1218, | |
| "num_input_tokens_seen": 825753600, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.021524796397287202, | |
| "grad_norm": 0.7618885636329651, | |
| "learning_rate": 0.0005884960310523109, | |
| "loss": 14.0575, | |
| "num_input_tokens_seen": 838860800, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.021861121340994816, | |
| "grad_norm": 0.7333565950393677, | |
| "learning_rate": 0.0005880523048408287, | |
| "loss": 14.0723, | |
| "num_input_tokens_seen": 851968000, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.022197446284702427, | |
| "grad_norm": 0.7767319083213806, | |
| "learning_rate": 0.0005876003559109981, | |
| "loss": 14.0067, | |
| "num_input_tokens_seen": 865075200, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.022533771228410042, | |
| "grad_norm": 0.7478107213973999, | |
| "learning_rate": 0.0005871401971641175, | |
| "loss": 14.0154, | |
| "num_input_tokens_seen": 878182400, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.022870096172117653, | |
| "grad_norm": 0.7439610958099365, | |
| "learning_rate": 0.0005866718417358421, | |
| "loss": 13.9922, | |
| "num_input_tokens_seen": 891289600, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.023206421115825267, | |
| "grad_norm": 0.7624334096908569, | |
| "learning_rate": 0.0005861953029958091, | |
| "loss": 13.9456, | |
| "num_input_tokens_seen": 904396800, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.02354274605953288, | |
| "grad_norm": 0.7594953775405884, | |
| "learning_rate": 0.0005857105945472556, | |
| "loss": 13.9742, | |
| "num_input_tokens_seen": 917504000, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.02354274605953288, | |
| "eval_loss": 3.410249710083008, | |
| "eval_runtime": 143.2447, | |
| "eval_samples_per_second": 34.905, | |
| "eval_steps_per_second": 8.726, | |
| "num_input_tokens_seen": 917504000, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.02387907100324049, | |
| "grad_norm": 0.7231994867324829, | |
| "learning_rate": 0.0005852177302266308, | |
| "loss": 13.959, | |
| "num_input_tokens_seen": 930611200, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.024215395946948104, | |
| "grad_norm": 0.7433524131774902, | |
| "learning_rate": 0.0005847167241032006, | |
| "loss": 13.8909, | |
| "num_input_tokens_seen": 943718400, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.024551720890655715, | |
| "grad_norm": 0.6849333643913269, | |
| "learning_rate": 0.0005842075904786462, | |
| "loss": 13.8984, | |
| "num_input_tokens_seen": 956825600, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.02488804583436333, | |
| "grad_norm": 0.7122375965118408, | |
| "learning_rate": 0.000583690343886656, | |
| "loss": 13.8611, | |
| "num_input_tokens_seen": 969932800, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.02522437077807094, | |
| "grad_norm": 0.7722771763801575, | |
| "learning_rate": 0.0005831649990925102, | |
| "loss": 13.862, | |
| "num_input_tokens_seen": 983040000, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.025560695721778555, | |
| "grad_norm": 0.7184539437294006, | |
| "learning_rate": 0.0005826315710926599, | |
| "loss": 13.8641, | |
| "num_input_tokens_seen": 996147200, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.025897020665486166, | |
| "grad_norm": 0.7419592142105103, | |
| "learning_rate": 0.0005820900751142987, | |
| "loss": 13.808, | |
| "num_input_tokens_seen": 1009254400, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.026233345609193777, | |
| "grad_norm": 0.7380815148353577, | |
| "learning_rate": 0.0005815405266149281, | |
| "loss": 13.7751, | |
| "num_input_tokens_seen": 1022361600, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.02656967055290139, | |
| "grad_norm": 0.7219839692115784, | |
| "learning_rate": 0.000580982941281916, | |
| "loss": 13.8042, | |
| "num_input_tokens_seen": 1035468800, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.026905995496609002, | |
| "grad_norm": 0.7262866497039795, | |
| "learning_rate": 0.0005804173350320493, | |
| "loss": 13.7434, | |
| "num_input_tokens_seen": 1048576000, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.026905995496609002, | |
| "eval_loss": 3.364596128463745, | |
| "eval_runtime": 143.5464, | |
| "eval_samples_per_second": 34.832, | |
| "eval_steps_per_second": 8.708, | |
| "num_input_tokens_seen": 1048576000, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.027242320440316617, | |
| "grad_norm": 0.7669729590415955, | |
| "learning_rate": 0.0005798437240110794, | |
| "loss": 13.759, | |
| "num_input_tokens_seen": 1061683200, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.027578645384024228, | |
| "grad_norm": 0.75081866979599, | |
| "learning_rate": 0.0005792621245932613, | |
| "loss": 13.8008, | |
| "num_input_tokens_seen": 1074790400, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.027914970327731842, | |
| "grad_norm": 0.6844099164009094, | |
| "learning_rate": 0.0005786725533808858, | |
| "loss": 13.7462, | |
| "num_input_tokens_seen": 1087897600, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.028251295271439453, | |
| "grad_norm": 0.6971242427825928, | |
| "learning_rate": 0.0005780750272038064, | |
| "loss": 13.7535, | |
| "num_input_tokens_seen": 1101004800, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.028587620215147068, | |
| "grad_norm": 0.7153123021125793, | |
| "learning_rate": 0.0005774695631189582, | |
| "loss": 13.7085, | |
| "num_input_tokens_seen": 1114112000, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.02892394515885468, | |
| "grad_norm": 0.6920833587646484, | |
| "learning_rate": 0.0005768561784098711, | |
| "loss": 13.6495, | |
| "num_input_tokens_seen": 1127219200, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.02926027010256229, | |
| "grad_norm": 0.7521312236785889, | |
| "learning_rate": 0.0005762348905861764, | |
| "loss": 13.6559, | |
| "num_input_tokens_seen": 1140326400, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.029596595046269904, | |
| "grad_norm": 0.676365077495575, | |
| "learning_rate": 0.0005756057173831074, | |
| "loss": 13.6069, | |
| "num_input_tokens_seen": 1153433600, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.029932919989977515, | |
| "grad_norm": 0.6946042776107788, | |
| "learning_rate": 0.0005749686767609928, | |
| "loss": 13.6218, | |
| "num_input_tokens_seen": 1166540800, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.03026924493368513, | |
| "grad_norm": 0.7127935886383057, | |
| "learning_rate": 0.0005743237869047437, | |
| "loss": 13.6039, | |
| "num_input_tokens_seen": 1179648000, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.03026924493368513, | |
| "eval_loss": 3.3240578174591064, | |
| "eval_runtime": 143.6605, | |
| "eval_samples_per_second": 34.804, | |
| "eval_steps_per_second": 8.701, | |
| "num_input_tokens_seen": 1179648000, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.03060556987739274, | |
| "grad_norm": 0.7366272807121277, | |
| "learning_rate": 0.0005736710662233351, | |
| "loss": 13.604, | |
| "num_input_tokens_seen": 1192755200, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.030941894821100355, | |
| "grad_norm": 0.6478694081306458, | |
| "learning_rate": 0.0005730105333492799, | |
| "loss": 13.5717, | |
| "num_input_tokens_seen": 1205862400, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.03127821976480797, | |
| "grad_norm": 0.7228036522865295, | |
| "learning_rate": 0.0005723422071380976, | |
| "loss": 13.5385, | |
| "num_input_tokens_seen": 1218969600, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.03161454470851558, | |
| "grad_norm": 0.6998932957649231, | |
| "learning_rate": 0.0005716661066677753, | |
| "loss": 13.5237, | |
| "num_input_tokens_seen": 1232076800, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.03195086965222319, | |
| "grad_norm": 0.6933197379112244, | |
| "learning_rate": 0.0005709822512382236, | |
| "loss": 13.5417, | |
| "num_input_tokens_seen": 1245184000, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.0322871945959308, | |
| "grad_norm": 0.7209503054618835, | |
| "learning_rate": 0.0005702906603707256, | |
| "loss": 13.5653, | |
| "num_input_tokens_seen": 1258291200, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.032623519539638414, | |
| "grad_norm": 0.6796743273735046, | |
| "learning_rate": 0.0005695913538073798, | |
| "loss": 13.557, | |
| "num_input_tokens_seen": 1271398400, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.03295984448334603, | |
| "grad_norm": 0.6676873564720154, | |
| "learning_rate": 0.0005688843515105359, | |
| "loss": 13.4965, | |
| "num_input_tokens_seen": 1284505600, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.03329616942705364, | |
| "grad_norm": 0.6413120627403259, | |
| "learning_rate": 0.0005681696736622258, | |
| "loss": 13.5013, | |
| "num_input_tokens_seen": 1297612800, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.033632494370761254, | |
| "grad_norm": 0.6716573238372803, | |
| "learning_rate": 0.0005674473406635868, | |
| "loss": 13.4891, | |
| "num_input_tokens_seen": 1310720000, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.033632494370761254, | |
| "eval_loss": 3.2916977405548096, | |
| "eval_runtime": 143.1949, | |
| "eval_samples_per_second": 34.917, | |
| "eval_steps_per_second": 8.729, | |
| "num_input_tokens_seen": 1310720000, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.033968819314468865, | |
| "grad_norm": 0.6855191588401794, | |
| "learning_rate": 0.0005667173731342798, | |
| "loss": 13.4753, | |
| "num_input_tokens_seen": 1323827200, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.03430514425817648, | |
| "grad_norm": 0.6939865350723267, | |
| "learning_rate": 0.0005659797919119, | |
| "loss": 13.4583, | |
| "num_input_tokens_seen": 1336934400, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.034641469201884094, | |
| "grad_norm": 0.6979573369026184, | |
| "learning_rate": 0.0005652346180513829, | |
| "loss": 13.4339, | |
| "num_input_tokens_seen": 1350041600, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.034977794145591705, | |
| "grad_norm": 0.6683679819107056, | |
| "learning_rate": 0.0005644818728244026, | |
| "loss": 13.4496, | |
| "num_input_tokens_seen": 1363148800, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.035314119089299316, | |
| "grad_norm": 0.6824884414672852, | |
| "learning_rate": 0.0005637215777187651, | |
| "loss": 13.4705, | |
| "num_input_tokens_seen": 1376256000, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.03565044403300693, | |
| "grad_norm": 0.6840626001358032, | |
| "learning_rate": 0.0005629537544377942, | |
| "loss": 13.4349, | |
| "num_input_tokens_seen": 1389363200, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.035986768976714545, | |
| "grad_norm": 0.6613268852233887, | |
| "learning_rate": 0.0005621784248997128, | |
| "loss": 13.46, | |
| "num_input_tokens_seen": 1402470400, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.036323093920422156, | |
| "grad_norm": 0.6920462846755981, | |
| "learning_rate": 0.0005613956112370167, | |
| "loss": 13.4035, | |
| "num_input_tokens_seen": 1415577600, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.03665941886412977, | |
| "grad_norm": 0.6671062111854553, | |
| "learning_rate": 0.0005606053357958429, | |
| "loss": 13.3312, | |
| "num_input_tokens_seen": 1428684800, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.03699574380783738, | |
| "grad_norm": 0.6575270295143127, | |
| "learning_rate": 0.0005598076211353316, | |
| "loss": 13.3718, | |
| "num_input_tokens_seen": 1441792000, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.03699574380783738, | |
| "eval_loss": 3.2621724605560303, | |
| "eval_runtime": 143.0547, | |
| "eval_samples_per_second": 34.952, | |
| "eval_steps_per_second": 8.738, | |
| "num_input_tokens_seen": 1441792000, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.037332068751544996, | |
| "grad_norm": 0.6697873473167419, | |
| "learning_rate": 0.0005590024900269825, | |
| "loss": 13.3337, | |
| "num_input_tokens_seen": 1454899200, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.03766839369525261, | |
| "grad_norm": 0.6501484513282776, | |
| "learning_rate": 0.0005581899654540048, | |
| "loss": 13.3573, | |
| "num_input_tokens_seen": 1468006400, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.03800471863896022, | |
| "grad_norm": 0.6328523755073547, | |
| "learning_rate": 0.0005573700706106607, | |
| "loss": 13.3513, | |
| "num_input_tokens_seen": 1481113600, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.03834104358266783, | |
| "grad_norm": 0.6470258831977844, | |
| "learning_rate": 0.0005565428289016039, | |
| "loss": 13.2964, | |
| "num_input_tokens_seen": 1494220800, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.03867736852637544, | |
| "grad_norm": 0.6687533855438232, | |
| "learning_rate": 0.0005557082639412105, | |
| "loss": 13.3508, | |
| "num_input_tokens_seen": 1507328000, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.03901369347008306, | |
| "grad_norm": 0.6524744629859924, | |
| "learning_rate": 0.0005548663995529062, | |
| "loss": 13.3254, | |
| "num_input_tokens_seen": 1520435200, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.03935001841379067, | |
| "grad_norm": 0.6340435147285461, | |
| "learning_rate": 0.0005540172597684852, | |
| "loss": 13.3107, | |
| "num_input_tokens_seen": 1533542400, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.03968634335749828, | |
| "grad_norm": 0.6675236225128174, | |
| "learning_rate": 0.000553160868827425, | |
| "loss": 13.264, | |
| "num_input_tokens_seen": 1546649600, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.04002266830120589, | |
| "grad_norm": 0.6894492506980896, | |
| "learning_rate": 0.0005522972511761935, | |
| "loss": 13.2815, | |
| "num_input_tokens_seen": 1559756800, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.04035899324491351, | |
| "grad_norm": 0.6689581274986267, | |
| "learning_rate": 0.000551426431467552, | |
| "loss": 13.3443, | |
| "num_input_tokens_seen": 1572864000, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.04035899324491351, | |
| "eval_loss": 3.237107276916504, | |
| "eval_runtime": 142.4946, | |
| "eval_samples_per_second": 35.089, | |
| "eval_steps_per_second": 8.772, | |
| "num_input_tokens_seen": 1572864000, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.04069531818862112, | |
| "grad_norm": 0.6856837868690491, | |
| "learning_rate": 0.0005505484345598515, | |
| "loss": 13.2681, | |
| "num_input_tokens_seen": 1585971200, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.04103164313232873, | |
| "grad_norm": 0.6631260514259338, | |
| "learning_rate": 0.0005496632855163221, | |
| "loss": 13.2594, | |
| "num_input_tokens_seen": 1599078400, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.04136796807603634, | |
| "grad_norm": 0.6479213833808899, | |
| "learning_rate": 0.0005487710096043584, | |
| "loss": 13.2822, | |
| "num_input_tokens_seen": 1612185600, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.04170429301974395, | |
| "grad_norm": 0.6974468231201172, | |
| "learning_rate": 0.0005478716322947985, | |
| "loss": 13.2206, | |
| "num_input_tokens_seen": 1625292800, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.04204061796345157, | |
| "grad_norm": 0.633106529712677, | |
| "learning_rate": 0.0005469651792611956, | |
| "loss": 13.2054, | |
| "num_input_tokens_seen": 1638400000, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.04237694290715918, | |
| "grad_norm": 0.6755931377410889, | |
| "learning_rate": 0.0005460516763790867, | |
| "loss": 13.206, | |
| "num_input_tokens_seen": 1651507200, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.04271326785086679, | |
| "grad_norm": 0.6707047820091248, | |
| "learning_rate": 0.0005451311497252529, | |
| "loss": 13.2538, | |
| "num_input_tokens_seen": 1664614400, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.043049592794574404, | |
| "grad_norm": 0.6525476574897766, | |
| "learning_rate": 0.0005442036255769754, | |
| "loss": 13.1984, | |
| "num_input_tokens_seen": 1677721600, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.04338591773828202, | |
| "grad_norm": 0.6575285196304321, | |
| "learning_rate": 0.0005432691304112853, | |
| "loss": 13.1798, | |
| "num_input_tokens_seen": 1690828800, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.04372224268198963, | |
| "grad_norm": 0.6678348183631897, | |
| "learning_rate": 0.0005423276909042077, | |
| "loss": 13.1945, | |
| "num_input_tokens_seen": 1703936000, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.04372224268198963, | |
| "eval_loss": 3.2158761024475098, | |
| "eval_runtime": 143.477, | |
| "eval_samples_per_second": 34.849, | |
| "eval_steps_per_second": 8.712, | |
| "num_input_tokens_seen": 1703936000, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.044058567625697244, | |
| "grad_norm": 0.6580634713172913, | |
| "learning_rate": 0.0005413793339300004, | |
| "loss": 13.1733, | |
| "num_input_tokens_seen": 1717043200, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.044394892569404855, | |
| "grad_norm": 0.7573990821838379, | |
| "learning_rate": 0.000540424086560387, | |
| "loss": 13.1998, | |
| "num_input_tokens_seen": 1730150400, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.044731217513112466, | |
| "grad_norm": 0.6535853147506714, | |
| "learning_rate": 0.000539461976063783, | |
| "loss": 13.1668, | |
| "num_input_tokens_seen": 1743257600, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.045067542456820084, | |
| "grad_norm": 0.6875225305557251, | |
| "learning_rate": 0.0005384930299045193, | |
| "loss": 13.1695, | |
| "num_input_tokens_seen": 1756364800, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.045403867400527695, | |
| "grad_norm": 0.6734049320220947, | |
| "learning_rate": 0.0005375172757420559, | |
| "loss": 13.1982, | |
| "num_input_tokens_seen": 1769472000, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.045740192344235306, | |
| "grad_norm": 0.6594141721725464, | |
| "learning_rate": 0.0005365347414301942, | |
| "loss": 13.132, | |
| "num_input_tokens_seen": 1782579200, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.04607651728794292, | |
| "grad_norm": 0.606103777885437, | |
| "learning_rate": 0.0005355454550162814, | |
| "loss": 13.15, | |
| "num_input_tokens_seen": 1795686400, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.046412842231650535, | |
| "grad_norm": 0.6484935879707336, | |
| "learning_rate": 0.0005345494447404089, | |
| "loss": 13.1301, | |
| "num_input_tokens_seen": 1808793600, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.046749167175358146, | |
| "grad_norm": 0.7227681279182434, | |
| "learning_rate": 0.0005335467390346076, | |
| "loss": 13.1443, | |
| "num_input_tokens_seen": 1821900800, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.04708549211906576, | |
| "grad_norm": 0.6700535416603088, | |
| "learning_rate": 0.0005325373665220355, | |
| "loss": 13.0997, | |
| "num_input_tokens_seen": 1835008000, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.04708549211906576, | |
| "eval_loss": 3.194380044937134, | |
| "eval_runtime": 142.7031, | |
| "eval_samples_per_second": 35.038, | |
| "eval_steps_per_second": 8.759, | |
| "num_input_tokens_seen": 1835008000, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.04742181706277337, | |
| "grad_norm": 0.6384073495864868, | |
| "learning_rate": 0.0005315213560161604, | |
| "loss": 13.0959, | |
| "num_input_tokens_seen": 1848115200, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 0.04775814200648098, | |
| "grad_norm": 0.6730595231056213, | |
| "learning_rate": 0.0005304987365199383, | |
| "loss": 13.081, | |
| "num_input_tokens_seen": 1861222400, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.0480944669501886, | |
| "grad_norm": 0.6358413696289062, | |
| "learning_rate": 0.0005294695372249843, | |
| "loss": 13.0862, | |
| "num_input_tokens_seen": 1874329600, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 0.04843079189389621, | |
| "grad_norm": 0.6400682926177979, | |
| "learning_rate": 0.0005284337875107402, | |
| "loss": 13.0959, | |
| "num_input_tokens_seen": 1887436800, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.04876711683760382, | |
| "grad_norm": 0.6422862410545349, | |
| "learning_rate": 0.0005273915169436359, | |
| "loss": 13.0957, | |
| "num_input_tokens_seen": 1900544000, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 0.04910344178131143, | |
| "grad_norm": 0.6517816185951233, | |
| "learning_rate": 0.0005263427552762443, | |
| "loss": 13.0312, | |
| "num_input_tokens_seen": 1913651200, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.04943976672501905, | |
| "grad_norm": 0.6352054476737976, | |
| "learning_rate": 0.0005252875324464333, | |
| "loss": 13.0642, | |
| "num_input_tokens_seen": 1926758400, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 0.04977609166872666, | |
| "grad_norm": 0.6357077956199646, | |
| "learning_rate": 0.0005242258785765105, | |
| "loss": 13.0704, | |
| "num_input_tokens_seen": 1939865600, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.05011241661243427, | |
| "grad_norm": 0.6192994713783264, | |
| "learning_rate": 0.0005231578239723635, | |
| "loss": 13.0549, | |
| "num_input_tokens_seen": 1952972800, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 0.05044874155614188, | |
| "grad_norm": 0.6180127859115601, | |
| "learning_rate": 0.0005220833991225946, | |
| "loss": 13.1213, | |
| "num_input_tokens_seen": 1966080000, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.05044874155614188, | |
| "eval_loss": 3.1755564212799072, | |
| "eval_runtime": 142.155, | |
| "eval_samples_per_second": 35.173, | |
| "eval_steps_per_second": 8.793, | |
| "num_input_tokens_seen": 1966080000, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.05078506649984949, | |
| "grad_norm": 0.663218080997467, | |
| "learning_rate": 0.0005210026346976507, | |
| "loss": 13.0441, | |
| "num_input_tokens_seen": 1979187200, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 0.05112139144355711, | |
| "grad_norm": 0.6263464093208313, | |
| "learning_rate": 0.0005199155615489478, | |
| "loss": 13.0148, | |
| "num_input_tokens_seen": 1992294400, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.05145771638726472, | |
| "grad_norm": 0.6272994875907898, | |
| "learning_rate": 0.0005188222107079903, | |
| "loss": 13.0467, | |
| "num_input_tokens_seen": 2005401600, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 0.05179404133097233, | |
| "grad_norm": 0.6265645623207092, | |
| "learning_rate": 0.0005177226133854845, | |
| "loss": 13.0346, | |
| "num_input_tokens_seen": 2018508800, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.05213036627467994, | |
| "grad_norm": 0.6151268482208252, | |
| "learning_rate": 0.0005166168009704493, | |
| "loss": 13.0065, | |
| "num_input_tokens_seen": 2031616000, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 0.052466691218387554, | |
| "grad_norm": 0.6234976649284363, | |
| "learning_rate": 0.0005155048050293182, | |
| "loss": 13.0419, | |
| "num_input_tokens_seen": 2044723200, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.05280301616209517, | |
| "grad_norm": 0.6200417280197144, | |
| "learning_rate": 0.0005143866573050397, | |
| "loss": 12.9675, | |
| "num_input_tokens_seen": 2057830400, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 0.05313934110580278, | |
| "grad_norm": 0.6281518340110779, | |
| "learning_rate": 0.0005132623897161705, | |
| "loss": 12.9652, | |
| "num_input_tokens_seen": 2070937600, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.053475666049510394, | |
| "grad_norm": 0.6501129269599915, | |
| "learning_rate": 0.0005121320343559641, | |
| "loss": 13.0074, | |
| "num_input_tokens_seen": 2084044800, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 0.053811990993218005, | |
| "grad_norm": 0.6317852139472961, | |
| "learning_rate": 0.0005109956234914558, | |
| "loss": 12.977, | |
| "num_input_tokens_seen": 2097152000, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.053811990993218005, | |
| "eval_loss": 3.1572272777557373, | |
| "eval_runtime": 142.9574, | |
| "eval_samples_per_second": 34.975, | |
| "eval_steps_per_second": 8.744, | |
| "num_input_tokens_seen": 2097152000, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.05414831593692562, | |
| "grad_norm": 0.6210319995880127, | |
| "learning_rate": 0.0005098531895625401, | |
| "loss": 12.9927, | |
| "num_input_tokens_seen": 2110259200, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 0.054484640880633234, | |
| "grad_norm": 0.6362951397895813, | |
| "learning_rate": 0.0005087047651810459, | |
| "loss": 12.9658, | |
| "num_input_tokens_seen": 2123366400, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.054820965824340845, | |
| "grad_norm": 0.6348525285720825, | |
| "learning_rate": 0.0005075503831298047, | |
| "loss": 12.9523, | |
| "num_input_tokens_seen": 2136473600, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 0.055157290768048456, | |
| "grad_norm": 0.6554312705993652, | |
| "learning_rate": 0.0005063900763617156, | |
| "loss": 12.9581, | |
| "num_input_tokens_seen": 2149580800, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.05549361571175607, | |
| "grad_norm": 0.6416252851486206, | |
| "learning_rate": 0.0005052238779988038, | |
| "loss": 12.9369, | |
| "num_input_tokens_seen": 2162688000, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 0.055829940655463685, | |
| "grad_norm": 0.653998076915741, | |
| "learning_rate": 0.0005040518213312757, | |
| "loss": 12.9279, | |
| "num_input_tokens_seen": 2175795200, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.056166265599171296, | |
| "grad_norm": 0.6076102256774902, | |
| "learning_rate": 0.0005028739398165686, | |
| "loss": 12.9306, | |
| "num_input_tokens_seen": 2188902400, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 0.05650259054287891, | |
| "grad_norm": 0.6263251304626465, | |
| "learning_rate": 0.0005016902670783949, | |
| "loss": 12.9367, | |
| "num_input_tokens_seen": 2202009600, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.05683891548658652, | |
| "grad_norm": 0.6503254175186157, | |
| "learning_rate": 0.0005005008369057835, | |
| "loss": 12.8458, | |
| "num_input_tokens_seen": 2215116800, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 0.057175240430294136, | |
| "grad_norm": 0.6300747394561768, | |
| "learning_rate": 0.0004993056832521138, | |
| "loss": 12.8892, | |
| "num_input_tokens_seen": 2228224000, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.057175240430294136, | |
| "eval_loss": 3.1433920860290527, | |
| "eval_runtime": 145.793, | |
| "eval_samples_per_second": 34.295, | |
| "eval_steps_per_second": 8.574, | |
| "num_input_tokens_seen": 2228224000, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.05751156537400175, | |
| "grad_norm": 0.6060501337051392, | |
| "learning_rate": 0.0004981048402341477, | |
| "loss": 12.9441, | |
| "num_input_tokens_seen": 2241331200, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 0.05784789031770936, | |
| "grad_norm": 0.650665819644928, | |
| "learning_rate": 0.0004968983421310554, | |
| "loss": 12.8715, | |
| "num_input_tokens_seen": 2254438400, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.05818421526141697, | |
| "grad_norm": 0.6535419821739197, | |
| "learning_rate": 0.0004956862233834363, | |
| "loss": 12.8842, | |
| "num_input_tokens_seen": 2267545600, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 0.05852054020512458, | |
| "grad_norm": 0.6426728367805481, | |
| "learning_rate": 0.0004944685185923365, | |
| "loss": 12.9156, | |
| "num_input_tokens_seen": 2280652800, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.0588568651488322, | |
| "grad_norm": 0.6501869559288025, | |
| "learning_rate": 0.0004932452625182607, | |
| "loss": 12.8779, | |
| "num_input_tokens_seen": 2293760000, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 0.05919319009253981, | |
| "grad_norm": 0.6345402002334595, | |
| "learning_rate": 0.0004920164900801805, | |
| "loss": 12.8629, | |
| "num_input_tokens_seen": 2306867200, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.05952951503624742, | |
| "grad_norm": 0.6712388396263123, | |
| "learning_rate": 0.0004907822363545365, | |
| "loss": 12.876, | |
| "num_input_tokens_seen": 2319974400, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 0.05986583997995503, | |
| "grad_norm": 0.6922229528427124, | |
| "learning_rate": 0.0004895425365742384, | |
| "loss": 12.8556, | |
| "num_input_tokens_seen": 2333081600, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.06020216492366265, | |
| "grad_norm": 0.6531935334205627, | |
| "learning_rate": 0.0004882974261276581, | |
| "loss": 12.8296, | |
| "num_input_tokens_seen": 2346188800, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 0.06053848986737026, | |
| "grad_norm": 0.6729333400726318, | |
| "learning_rate": 0.00048704694055762005, | |
| "loss": 12.8258, | |
| "num_input_tokens_seen": 2359296000, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.06053848986737026, | |
| "eval_loss": 3.1286280155181885, | |
| "eval_runtime": 143.0292, | |
| "eval_samples_per_second": 34.958, | |
| "eval_steps_per_second": 8.739, | |
| "num_input_tokens_seen": 2359296000, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.06087481481107787, | |
| "grad_norm": 0.6235183477401733, | |
| "learning_rate": 0.0004857911155603867, | |
| "loss": 12.8588, | |
| "num_input_tokens_seen": 2372403200, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 0.06121113975478548, | |
| "grad_norm": 0.642000138759613, | |
| "learning_rate": 0.0004845299869846392, | |
| "loss": 12.8232, | |
| "num_input_tokens_seen": 2385510400, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.06154746469849309, | |
| "grad_norm": 0.6252527236938477, | |
| "learning_rate": 0.0004832635908304543, | |
| "loss": 12.8595, | |
| "num_input_tokens_seen": 2398617600, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 0.06188378964220071, | |
| "grad_norm": 0.6228143572807312, | |
| "learning_rate": 0.0004819919632482766, | |
| "loss": 12.8152, | |
| "num_input_tokens_seen": 2411724800, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.06222011458590832, | |
| "grad_norm": 0.661567211151123, | |
| "learning_rate": 0.00048071514053788666, | |
| "loss": 12.8356, | |
| "num_input_tokens_seen": 2424832000, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 0.06255643952961594, | |
| "grad_norm": 0.6318378448486328, | |
| "learning_rate": 0.00047943315914736475, | |
| "loss": 12.831, | |
| "num_input_tokens_seen": 2437939200, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.06289276447332355, | |
| "grad_norm": 0.6098783612251282, | |
| "learning_rate": 0.0004781460556720504, | |
| "loss": 12.8363, | |
| "num_input_tokens_seen": 2451046400, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 0.06322908941703116, | |
| "grad_norm": 0.643997073173523, | |
| "learning_rate": 0.00047685386685349796, | |
| "loss": 12.8267, | |
| "num_input_tokens_seen": 2464153600, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.06356541436073877, | |
| "grad_norm": 0.6287397146224976, | |
| "learning_rate": 0.000475556629578427, | |
| "loss": 12.8131, | |
| "num_input_tokens_seen": 2477260800, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 0.06390173930444638, | |
| "grad_norm": 0.6625655889511108, | |
| "learning_rate": 0.0004742543808776708, | |
| "loss": 12.8312, | |
| "num_input_tokens_seen": 2490368000, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.06390173930444638, | |
| "eval_loss": 3.1130659580230713, | |
| "eval_runtime": 143.1847, | |
| "eval_samples_per_second": 34.92, | |
| "eval_steps_per_second": 8.73, | |
| "num_input_tokens_seen": 2490368000, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.064238064248154, | |
| "grad_norm": 0.6380253434181213, | |
| "learning_rate": 0.0004729471579251177, | |
| "loss": 12.8645, | |
| "num_input_tokens_seen": 2503475200, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 0.0645743891918616, | |
| "grad_norm": 0.6287338137626648, | |
| "learning_rate": 0.00047163499803665085, | |
| "loss": 12.7931, | |
| "num_input_tokens_seen": 2516582400, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.06491071413556922, | |
| "grad_norm": 0.6729796528816223, | |
| "learning_rate": 0.00047031793866908294, | |
| "loss": 12.7903, | |
| "num_input_tokens_seen": 2529689600, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 0.06524703907927683, | |
| "grad_norm": 0.6398154497146606, | |
| "learning_rate": 0.0004689960174190865, | |
| "loss": 12.7746, | |
| "num_input_tokens_seen": 2542796800, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.06558336402298445, | |
| "grad_norm": 0.6751012206077576, | |
| "learning_rate": 0.00046766927202212145, | |
| "loss": 12.7655, | |
| "num_input_tokens_seen": 2555904000, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 0.06591968896669206, | |
| "grad_norm": 0.6046076416969299, | |
| "learning_rate": 0.0004663377403513568, | |
| "loss": 12.8018, | |
| "num_input_tokens_seen": 2569011200, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.06625601391039967, | |
| "grad_norm": 0.6526479721069336, | |
| "learning_rate": 0.0004650014604165907, | |
| "loss": 12.7394, | |
| "num_input_tokens_seen": 2582118400, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 0.06659233885410729, | |
| "grad_norm": 0.6187541484832764, | |
| "learning_rate": 0.00046366047036316456, | |
| "loss": 12.7346, | |
| "num_input_tokens_seen": 2595225600, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.0669286637978149, | |
| "grad_norm": 0.6106886267662048, | |
| "learning_rate": 0.0004623148084708745, | |
| "loss": 12.7597, | |
| "num_input_tokens_seen": 2608332800, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 0.06726498874152251, | |
| "grad_norm": 0.6358317136764526, | |
| "learning_rate": 0.0004609645131528788, | |
| "loss": 12.7303, | |
| "num_input_tokens_seen": 2621440000, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.06726498874152251, | |
| "eval_loss": 3.1026949882507324, | |
| "eval_runtime": 142.2834, | |
| "eval_samples_per_second": 35.141, | |
| "eval_steps_per_second": 8.785, | |
| "num_input_tokens_seen": 2621440000, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.06760131368523012, | |
| "grad_norm": 0.6313095688819885, | |
| "learning_rate": 0.0004596096229546009, | |
| "loss": 12.7336, | |
| "num_input_tokens_seen": 2634547200, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 0.06793763862893773, | |
| "grad_norm": 0.6490457057952881, | |
| "learning_rate": 0.00045825017655262934, | |
| "loss": 12.7727, | |
| "num_input_tokens_seen": 2647654400, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.06827396357264534, | |
| "grad_norm": 0.6609966158866882, | |
| "learning_rate": 0.000456886212753614, | |
| "loss": 12.759, | |
| "num_input_tokens_seen": 2660761600, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 0.06861028851635297, | |
| "grad_norm": 0.6392827033996582, | |
| "learning_rate": 0.00045551777049315757, | |
| "loss": 12.7189, | |
| "num_input_tokens_seen": 2673868800, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.06894661346006058, | |
| "grad_norm": 0.6272814273834229, | |
| "learning_rate": 0.0004541448888347047, | |
| "loss": 12.6948, | |
| "num_input_tokens_seen": 2686976000, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 0.06928293840376819, | |
| "grad_norm": 0.6286495327949524, | |
| "learning_rate": 0.00045276760696842693, | |
| "loss": 12.7224, | |
| "num_input_tokens_seen": 2700083200, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.0696192633474758, | |
| "grad_norm": 0.6213704943656921, | |
| "learning_rate": 0.00045138596421010374, | |
| "loss": 12.778, | |
| "num_input_tokens_seen": 2713190400, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 0.06995558829118341, | |
| "grad_norm": 0.6061195731163025, | |
| "learning_rate": 0.00045, | |
| "loss": 12.7403, | |
| "num_input_tokens_seen": 2726297600, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.07029191323489102, | |
| "grad_norm": 0.6419244408607483, | |
| "learning_rate": 0.0004486097539017407, | |
| "loss": 12.7137, | |
| "num_input_tokens_seen": 2739404800, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 0.07062823817859863, | |
| "grad_norm": 0.6618810892105103, | |
| "learning_rate": 0.00044721526560118134, | |
| "loss": 12.6896, | |
| "num_input_tokens_seen": 2752512000, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.07062823817859863, | |
| "eval_loss": 3.0883917808532715, | |
| "eval_runtime": 142.4547, | |
| "eval_samples_per_second": 35.099, | |
| "eval_steps_per_second": 8.775, | |
| "num_input_tokens_seen": 2752512000, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.07096456312230624, | |
| "grad_norm": 0.6856646537780762, | |
| "learning_rate": 0.00044581657490527473, | |
| "loss": 12.6825, | |
| "num_input_tokens_seen": 2765619200, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 0.07130088806601385, | |
| "grad_norm": 0.6331352591514587, | |
| "learning_rate": 0.00044441372174093487, | |
| "loss": 12.675, | |
| "num_input_tokens_seen": 2778726400, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.07163721300972148, | |
| "grad_norm": 0.6496602296829224, | |
| "learning_rate": 0.0004430067461538976, | |
| "loss": 12.6842, | |
| "num_input_tokens_seen": 2791833600, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 0.07197353795342909, | |
| "grad_norm": 0.6698866486549377, | |
| "learning_rate": 0.00044159568830757687, | |
| "loss": 12.6498, | |
| "num_input_tokens_seen": 2804940800, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.0723098628971367, | |
| "grad_norm": 0.6556456089019775, | |
| "learning_rate": 0.00044018058848191855, | |
| "loss": 12.7073, | |
| "num_input_tokens_seen": 2818048000, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 0.07264618784084431, | |
| "grad_norm": 0.6554015278816223, | |
| "learning_rate": 0.0004387614870722506, | |
| "loss": 12.6515, | |
| "num_input_tokens_seen": 2831155200, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.07298251278455192, | |
| "grad_norm": 0.6356109380722046, | |
| "learning_rate": 0.0004373384245881296, | |
| "loss": 12.6759, | |
| "num_input_tokens_seen": 2844262400, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 0.07331883772825953, | |
| "grad_norm": 0.6429396271705627, | |
| "learning_rate": 0.0004359114416521851, | |
| "loss": 12.6469, | |
| "num_input_tokens_seen": 2857369600, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.07365516267196714, | |
| "grad_norm": 0.6229676604270935, | |
| "learning_rate": 0.0004344805789989591, | |
| "loss": 12.6783, | |
| "num_input_tokens_seen": 2870476800, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 0.07399148761567476, | |
| "grad_norm": 0.6383066177368164, | |
| "learning_rate": 0.000433045877473744, | |
| "loss": 12.6273, | |
| "num_input_tokens_seen": 2883584000, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.07399148761567476, | |
| "eval_loss": 3.076796054840088, | |
| "eval_runtime": 143.2443, | |
| "eval_samples_per_second": 34.905, | |
| "eval_steps_per_second": 8.726, | |
| "num_input_tokens_seen": 2883584000, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.07432781255938237, | |
| "grad_norm": 0.612218976020813, | |
| "learning_rate": 0.0004316073780314163, | |
| "loss": 12.6729, | |
| "num_input_tokens_seen": 2896691200, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 0.07466413750308999, | |
| "grad_norm": 0.6343071460723877, | |
| "learning_rate": 0.00043016512173526736, | |
| "loss": 12.6507, | |
| "num_input_tokens_seen": 2909798400, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.0750004624467976, | |
| "grad_norm": 0.6494725942611694, | |
| "learning_rate": 0.0004287191497558317, | |
| "loss": 12.6271, | |
| "num_input_tokens_seen": 2922905600, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 0.07533678739050521, | |
| "grad_norm": 0.6436727046966553, | |
| "learning_rate": 0.0004272695033697111, | |
| "loss": 12.6529, | |
| "num_input_tokens_seen": 2936012800, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.07567311233421282, | |
| "grad_norm": 0.6481876373291016, | |
| "learning_rate": 0.00042581622395839705, | |
| "loss": 12.6528, | |
| "num_input_tokens_seen": 2949120000, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 0.07600943727792044, | |
| "grad_norm": 0.6492651104927063, | |
| "learning_rate": 0.0004243593530070886, | |
| "loss": 12.6312, | |
| "num_input_tokens_seen": 2962227200, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.07634576222162805, | |
| "grad_norm": 0.6570179462432861, | |
| "learning_rate": 0.00042289893210350907, | |
| "loss": 12.6428, | |
| "num_input_tokens_seen": 2975334400, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 0.07668208716533566, | |
| "grad_norm": 0.6505069732666016, | |
| "learning_rate": 0.0004214350029367181, | |
| "loss": 12.6549, | |
| "num_input_tokens_seen": 2988441600, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.07701841210904327, | |
| "grad_norm": 0.6301828026771545, | |
| "learning_rate": 0.0004199676072959222, | |
| "loss": 12.5838, | |
| "num_input_tokens_seen": 3001548800, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 0.07735473705275088, | |
| "grad_norm": 0.625487208366394, | |
| "learning_rate": 0.0004184967870692816, | |
| "loss": 12.6166, | |
| "num_input_tokens_seen": 3014656000, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.07735473705275088, | |
| "eval_loss": 3.0652644634246826, | |
| "eval_runtime": 142.6807, | |
| "eval_samples_per_second": 35.043, | |
| "eval_steps_per_second": 8.761, | |
| "num_input_tokens_seen": 3014656000, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.0776910619964585, | |
| "grad_norm": 0.6678441762924194, | |
| "learning_rate": 0.000417022584242714, | |
| "loss": 12.6271, | |
| "num_input_tokens_seen": 3027763200, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 0.07802738694016612, | |
| "grad_norm": 0.6448168754577637, | |
| "learning_rate": 0.00041554504089869716, | |
| "loss": 12.6012, | |
| "num_input_tokens_seen": 3040870400, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.07836371188387373, | |
| "grad_norm": 0.6791290640830994, | |
| "learning_rate": 0.0004140641992150667, | |
| "loss": 12.5798, | |
| "num_input_tokens_seen": 3053977600, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 0.07870003682758134, | |
| "grad_norm": 0.8807069659233093, | |
| "learning_rate": 0.00041258010146381224, | |
| "loss": 12.6015, | |
| "num_input_tokens_seen": 3067084800, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.07903636177128895, | |
| "grad_norm": 0.6284939646720886, | |
| "learning_rate": 0.00041109279000987105, | |
| "loss": 12.6183, | |
| "num_input_tokens_seen": 3080192000, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 0.07937268671499656, | |
| "grad_norm": 0.6453195810317993, | |
| "learning_rate": 0.0004096023073099185, | |
| "loss": 12.6, | |
| "num_input_tokens_seen": 3093299200, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.07970901165870417, | |
| "grad_norm": 0.6511227488517761, | |
| "learning_rate": 0.00040810869591115603, | |
| "loss": 12.5952, | |
| "num_input_tokens_seen": 3106406400, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 0.08004533660241178, | |
| "grad_norm": 0.6701833009719849, | |
| "learning_rate": 0.0004066119984500966, | |
| "loss": 12.5674, | |
| "num_input_tokens_seen": 3119513600, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.08038166154611939, | |
| "grad_norm": 0.6320140957832336, | |
| "learning_rate": 0.0004051122576513479, | |
| "loss": 12.5772, | |
| "num_input_tokens_seen": 3132620800, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 0.08071798648982702, | |
| "grad_norm": 0.6579756736755371, | |
| "learning_rate": 0.00040360951632639226, | |
| "loss": 12.57, | |
| "num_input_tokens_seen": 3145728000, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.08071798648982702, | |
| "eval_loss": 3.0548510551452637, | |
| "eval_runtime": 142.805, | |
| "eval_samples_per_second": 35.013, | |
| "eval_steps_per_second": 8.753, | |
| "num_input_tokens_seen": 3145728000, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.08105431143353463, | |
| "grad_norm": 0.6717228293418884, | |
| "learning_rate": 0.0004021038173723649, | |
| "loss": 12.5689, | |
| "num_input_tokens_seen": 3158835200, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 0.08139063637724224, | |
| "grad_norm": 0.6350929141044617, | |
| "learning_rate": 0.0004005952037708293, | |
| "loss": 12.5709, | |
| "num_input_tokens_seen": 3171942400, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.08172696132094985, | |
| "grad_norm": 0.6500872373580933, | |
| "learning_rate": 0.00039908371858655013, | |
| "loss": 12.576, | |
| "num_input_tokens_seen": 3185049600, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 0.08206328626465746, | |
| "grad_norm": 0.6404949426651001, | |
| "learning_rate": 0.00039756940496626415, | |
| "loss": 12.5173, | |
| "num_input_tokens_seen": 3198156800, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.08239961120836507, | |
| "grad_norm": 0.6140453219413757, | |
| "learning_rate": 0.0003960523061374484, | |
| "loss": 12.5427, | |
| "num_input_tokens_seen": 3211264000, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 0.08273593615207268, | |
| "grad_norm": 0.6440966725349426, | |
| "learning_rate": 0.00039453246540708625, | |
| "loss": 12.5706, | |
| "num_input_tokens_seen": 3224371200, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.0830722610957803, | |
| "grad_norm": 0.6301671862602234, | |
| "learning_rate": 0.00039300992616043105, | |
| "loss": 12.5483, | |
| "num_input_tokens_seen": 3237478400, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 0.0834085860394879, | |
| "grad_norm": 0.628695547580719, | |
| "learning_rate": 0.00039148473185976815, | |
| "loss": 12.5334, | |
| "num_input_tokens_seen": 3250585600, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.08374491098319553, | |
| "grad_norm": 0.6627179980278015, | |
| "learning_rate": 0.0003899569260431734, | |
| "loss": 12.565, | |
| "num_input_tokens_seen": 3263692800, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 0.08408123592690314, | |
| "grad_norm": 0.6234163045883179, | |
| "learning_rate": 0.00038842655232327125, | |
| "loss": 12.5742, | |
| "num_input_tokens_seen": 3276800000, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.08408123592690314, | |
| "eval_loss": 3.0441489219665527, | |
| "eval_runtime": 141.8038, | |
| "eval_samples_per_second": 35.26, | |
| "eval_steps_per_second": 8.815, | |
| "num_input_tokens_seen": 3276800000, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.08441756087061075, | |
| "grad_norm": 0.6204286217689514, | |
| "learning_rate": 0.0003868936543859888, | |
| "loss": 12.5493, | |
| "num_input_tokens_seen": 3289907200, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 0.08475388581431836, | |
| "grad_norm": 0.6237512230873108, | |
| "learning_rate": 0.00038535827598930967, | |
| "loss": 12.5179, | |
| "num_input_tokens_seen": 3303014400, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.08509021075802597, | |
| "grad_norm": 0.6418094635009766, | |
| "learning_rate": 0.00038382046096202435, | |
| "loss": 12.5096, | |
| "num_input_tokens_seen": 3316121600, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 0.08542653570173359, | |
| "grad_norm": 0.6306421160697937, | |
| "learning_rate": 0.0003822802532024791, | |
| "loss": 12.5202, | |
| "num_input_tokens_seen": 3329228800, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.0857628606454412, | |
| "grad_norm": 0.6436113715171814, | |
| "learning_rate": 0.000380737696677323, | |
| "loss": 12.4871, | |
| "num_input_tokens_seen": 3342336000, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 0.08609918558914881, | |
| "grad_norm": 1.0079458951950073, | |
| "learning_rate": 0.00037919283542025287, | |
| "loss": 12.4992, | |
| "num_input_tokens_seen": 3355443200, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.08643551053285642, | |
| "grad_norm": 0.6185023188591003, | |
| "learning_rate": 0.0003776457135307562, | |
| "loss": 12.4876, | |
| "num_input_tokens_seen": 3368550400, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 0.08677183547656404, | |
| "grad_norm": 0.6664910912513733, | |
| "learning_rate": 0.0003760963751728521, | |
| "loss": 12.4876, | |
| "num_input_tokens_seen": 3381657600, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.08710816042027165, | |
| "grad_norm": 0.6112196445465088, | |
| "learning_rate": 0.00037454486457383124, | |
| "loss": 12.4972, | |
| "num_input_tokens_seen": 3394764800, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 0.08744448536397927, | |
| "grad_norm": 0.6308513879776001, | |
| "learning_rate": 0.00037299122602299257, | |
| "loss": 12.4583, | |
| "num_input_tokens_seen": 3407872000, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.08744448536397927, | |
| "eval_loss": 3.034710645675659, | |
| "eval_runtime": 182.8724, | |
| "eval_samples_per_second": 27.341, | |
| "eval_steps_per_second": 6.835, | |
| "num_input_tokens_seen": 3407872000, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.08778081030768688, | |
| "grad_norm": 0.6481872200965881, | |
| "learning_rate": 0.00037143550387037943, | |
| "loss": 12.4646, | |
| "num_input_tokens_seen": 3420979200, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 0.08811713525139449, | |
| "grad_norm": 0.6672606468200684, | |
| "learning_rate": 0.0003698777425255136, | |
| "loss": 12.4237, | |
| "num_input_tokens_seen": 3434086400, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.0884534601951021, | |
| "grad_norm": 0.6188272833824158, | |
| "learning_rate": 0.00036831798645612735, | |
| "loss": 12.4983, | |
| "num_input_tokens_seen": 3447193600, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 0.08878978513880971, | |
| "grad_norm": 0.6584819555282593, | |
| "learning_rate": 0.0003667562801868943, | |
| "loss": 12.4316, | |
| "num_input_tokens_seen": 3460300800, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.08912611008251732, | |
| "grad_norm": 0.6392587423324585, | |
| "learning_rate": 0.0003651926682981584, | |
| "loss": 12.4541, | |
| "num_input_tokens_seen": 3473408000, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 0.08946243502622493, | |
| "grad_norm": 0.6473196148872375, | |
| "learning_rate": 0.00036362719542466104, | |
| "loss": 12.4921, | |
| "num_input_tokens_seen": 3486515200, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.08979875996993256, | |
| "grad_norm": 0.6527711153030396, | |
| "learning_rate": 0.00036205990625426724, | |
| "loss": 12.4578, | |
| "num_input_tokens_seen": 3499622400, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 0.09013508491364017, | |
| "grad_norm": 0.6588818430900574, | |
| "learning_rate": 0.00036049084552669, | |
| "loss": 12.4449, | |
| "num_input_tokens_seen": 3512729600, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.09047140985734778, | |
| "grad_norm": 0.6333611011505127, | |
| "learning_rate": 0.00035892005803221286, | |
| "loss": 12.4364, | |
| "num_input_tokens_seen": 3525836800, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 0.09080773480105539, | |
| "grad_norm": 0.6385447978973389, | |
| "learning_rate": 0.0003573475886104117, | |
| "loss": 12.4483, | |
| "num_input_tokens_seen": 3538944000, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.09080773480105539, | |
| "eval_loss": 3.0267038345336914, | |
| "eval_runtime": 143.0402, | |
| "eval_samples_per_second": 34.955, | |
| "eval_steps_per_second": 8.739, | |
| "num_input_tokens_seen": 3538944000, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.091144059744763, | |
| "grad_norm": 0.652103066444397, | |
| "learning_rate": 0.0003557734821488744, | |
| "loss": 12.3973, | |
| "num_input_tokens_seen": 3552051200, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 0.09148038468847061, | |
| "grad_norm": 0.629550576210022, | |
| "learning_rate": 0.00035419778358191967, | |
| "loss": 12.4529, | |
| "num_input_tokens_seen": 3565158400, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.09181670963217822, | |
| "grad_norm": 0.646165132522583, | |
| "learning_rate": 0.00035262053788931446, | |
| "loss": 12.4602, | |
| "num_input_tokens_seen": 3578265600, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 0.09215303457588583, | |
| "grad_norm": 0.6328135132789612, | |
| "learning_rate": 0.0003510417900949898, | |
| "loss": 12.4859, | |
| "num_input_tokens_seen": 3591372800, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.09248935951959344, | |
| "grad_norm": 0.6435760259628296, | |
| "learning_rate": 0.0003494615852657555, | |
| "loss": 12.4747, | |
| "num_input_tokens_seen": 3604480000, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 0.09282568446330107, | |
| "grad_norm": 0.6149182915687561, | |
| "learning_rate": 0.0003478799685100137, | |
| "loss": 12.4353, | |
| "num_input_tokens_seen": 3617587200, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.09316200940700868, | |
| "grad_norm": 0.6365089416503906, | |
| "learning_rate": 0.00034629698497647176, | |
| "loss": 12.4255, | |
| "num_input_tokens_seen": 3630694400, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 0.09349833435071629, | |
| "grad_norm": 0.6469732522964478, | |
| "learning_rate": 0.0003447126798528523, | |
| "loss": 12.4259, | |
| "num_input_tokens_seen": 3643801600, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.0938346592944239, | |
| "grad_norm": 0.6317386031150818, | |
| "learning_rate": 0.00034312709836460453, | |
| "loss": 12.4626, | |
| "num_input_tokens_seen": 3656908800, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 0.09417098423813151, | |
| "grad_norm": 0.6267306208610535, | |
| "learning_rate": 0.00034154028577361217, | |
| "loss": 12.3991, | |
| "num_input_tokens_seen": 3670016000, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.09417098423813151, | |
| "eval_loss": 3.016310691833496, | |
| "eval_runtime": 142.725, | |
| "eval_samples_per_second": 35.032, | |
| "eval_steps_per_second": 8.758, | |
| "num_input_tokens_seen": 3670016000, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.09450730918183912, | |
| "grad_norm": 0.6656507849693298, | |
| "learning_rate": 0.0003399522873769023, | |
| "loss": 12.4213, | |
| "num_input_tokens_seen": 3683123200, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 0.09484363412554674, | |
| "grad_norm": 0.6371810436248779, | |
| "learning_rate": 0.0003383631485053518, | |
| "loss": 12.4092, | |
| "num_input_tokens_seen": 3696230400, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.09517995906925435, | |
| "grad_norm": 0.6278609037399292, | |
| "learning_rate": 0.0003367729145223933, | |
| "loss": 12.3764, | |
| "num_input_tokens_seen": 3709337600, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 0.09551628401296196, | |
| "grad_norm": 0.6190541982650757, | |
| "learning_rate": 0.00033518163082272055, | |
| "loss": 12.4095, | |
| "num_input_tokens_seen": 3722444800, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.09585260895666958, | |
| "grad_norm": 0.6580514907836914, | |
| "learning_rate": 0.00033358934283099235, | |
| "loss": 12.3431, | |
| "num_input_tokens_seen": 3735552000, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 0.0961889339003772, | |
| "grad_norm": 0.6620698571205139, | |
| "learning_rate": 0.000331996096000536, | |
| "loss": 12.3971, | |
| "num_input_tokens_seen": 3748659200, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.0965252588440848, | |
| "grad_norm": 0.61739182472229, | |
| "learning_rate": 0.00033040193581204973, | |
| "loss": 12.3897, | |
| "num_input_tokens_seen": 3761766400, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 0.09686158378779242, | |
| "grad_norm": 0.6852706670761108, | |
| "learning_rate": 0.0003288069077723045, | |
| "loss": 12.4072, | |
| "num_input_tokens_seen": 3774873600, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.09719790873150003, | |
| "grad_norm": 0.6366174817085266, | |
| "learning_rate": 0.00032721105741284466, | |
| "loss": 12.3834, | |
| "num_input_tokens_seen": 3787980800, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 0.09753423367520764, | |
| "grad_norm": 0.685984194278717, | |
| "learning_rate": 0.0003256144302886885, | |
| "loss": 12.4215, | |
| "num_input_tokens_seen": 3801088000, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.09753423367520764, | |
| "eval_loss": 3.0072007179260254, | |
| "eval_runtime": 142.0382, | |
| "eval_samples_per_second": 35.202, | |
| "eval_steps_per_second": 8.8, | |
| "num_input_tokens_seen": 3801088000, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.09787055861891525, | |
| "grad_norm": 0.633934736251831, | |
| "learning_rate": 0.000324017071977028, | |
| "loss": 12.3848, | |
| "num_input_tokens_seen": 3814195200, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 0.09820688356262286, | |
| "grad_norm": 0.6223523020744324, | |
| "learning_rate": 0.0003224190280759273, | |
| "loss": 12.389, | |
| "num_input_tokens_seen": 3827302400, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.09854320850633047, | |
| "grad_norm": 0.6419284343719482, | |
| "learning_rate": 0.00032082034420302137, | |
| "loss": 12.3622, | |
| "num_input_tokens_seen": 3840409600, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 0.0988795334500381, | |
| "grad_norm": 0.6162405610084534, | |
| "learning_rate": 0.0003192210659942139, | |
| "loss": 12.4409, | |
| "num_input_tokens_seen": 3853516800, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.0992158583937457, | |
| "grad_norm": 0.6561248898506165, | |
| "learning_rate": 0.0003176212391023743, | |
| "loss": 12.4152, | |
| "num_input_tokens_seen": 3866624000, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 0.09955218333745332, | |
| "grad_norm": 0.6575373411178589, | |
| "learning_rate": 0.0003160209091960347, | |
| "loss": 12.3603, | |
| "num_input_tokens_seen": 3879731200, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.09988850828116093, | |
| "grad_norm": 0.6060482859611511, | |
| "learning_rate": 0.0003144201219580862, | |
| "loss": 12.3752, | |
| "num_input_tokens_seen": 3892838400, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 0.10022483322486854, | |
| "grad_norm": 0.6433590650558472, | |
| "learning_rate": 0.000312818923084475, | |
| "loss": 12.3568, | |
| "num_input_tokens_seen": 3905945600, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.10056115816857615, | |
| "grad_norm": 0.626518189907074, | |
| "learning_rate": 0.00031121735828289773, | |
| "loss": 12.3327, | |
| "num_input_tokens_seen": 3919052800, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 0.10089748311228376, | |
| "grad_norm": 0.6467755436897278, | |
| "learning_rate": 0.0003096154732714966, | |
| "loss": 12.367, | |
| "num_input_tokens_seen": 3932160000, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.10089748311228376, | |
| "eval_loss": 2.9978182315826416, | |
| "eval_runtime": 143.1987, | |
| "eval_samples_per_second": 34.917, | |
| "eval_steps_per_second": 8.729, | |
| "num_input_tokens_seen": 3932160000, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.10123380805599137, | |
| "grad_norm": 0.6437516808509827, | |
| "learning_rate": 0.00030801331377755466, | |
| "loss": 12.3776, | |
| "num_input_tokens_seen": 3945267200, | |
| "step": 15050 | |
| }, | |
| { | |
| "epoch": 0.10157013299969898, | |
| "grad_norm": 0.6743655204772949, | |
| "learning_rate": 0.0003064109255361904, | |
| "loss": 12.326, | |
| "num_input_tokens_seen": 3958374400, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.1019064579434066, | |
| "grad_norm": 0.6296969056129456, | |
| "learning_rate": 0.00030480835428905214, | |
| "loss": 12.3444, | |
| "num_input_tokens_seen": 3971481600, | |
| "step": 15150 | |
| }, | |
| { | |
| "epoch": 0.10224278288711422, | |
| "grad_norm": 0.648457407951355, | |
| "learning_rate": 0.000303205645783012, | |
| "loss": 12.3422, | |
| "num_input_tokens_seen": 3984588800, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.10257910783082183, | |
| "grad_norm": 0.6306461691856384, | |
| "learning_rate": 0.0003016028457688604, | |
| "loss": 12.3452, | |
| "num_input_tokens_seen": 3997696000, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 0.10291543277452944, | |
| "grad_norm": 0.6481978893280029, | |
| "learning_rate": 0.0003, | |
| "loss": 12.3079, | |
| "num_input_tokens_seen": 4010803200, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.10325175771823705, | |
| "grad_norm": 0.7946459650993347, | |
| "learning_rate": 0.0002983971542311397, | |
| "loss": 12.3674, | |
| "num_input_tokens_seen": 4023910400, | |
| "step": 15350 | |
| }, | |
| { | |
| "epoch": 0.10358808266194466, | |
| "grad_norm": 0.6375327706336975, | |
| "learning_rate": 0.000296794354216988, | |
| "loss": 12.3125, | |
| "num_input_tokens_seen": 4037017600, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.10392440760565227, | |
| "grad_norm": 0.6338579058647156, | |
| "learning_rate": 0.0002951916457109479, | |
| "loss": 12.3305, | |
| "num_input_tokens_seen": 4050124800, | |
| "step": 15450 | |
| }, | |
| { | |
| "epoch": 0.10426073254935989, | |
| "grad_norm": 0.642365038394928, | |
| "learning_rate": 0.00029358907446380955, | |
| "loss": 12.3038, | |
| "num_input_tokens_seen": 4063232000, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.10426073254935989, | |
| "eval_loss": 2.9912989139556885, | |
| "eval_runtime": 142.6952, | |
| "eval_samples_per_second": 35.04, | |
| "eval_steps_per_second": 8.76, | |
| "num_input_tokens_seen": 4063232000, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.1045970574930675, | |
| "grad_norm": 0.6200032830238342, | |
| "learning_rate": 0.00029198668622244534, | |
| "loss": 12.3153, | |
| "num_input_tokens_seen": 4076339200, | |
| "step": 15550 | |
| }, | |
| { | |
| "epoch": 0.10493338243677511, | |
| "grad_norm": 0.6352826356887817, | |
| "learning_rate": 0.0002903845267285034, | |
| "loss": 12.3094, | |
| "num_input_tokens_seen": 4089446400, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.10526970738048273, | |
| "grad_norm": 0.6530657410621643, | |
| "learning_rate": 0.0002887826417171023, | |
| "loss": 12.3094, | |
| "num_input_tokens_seen": 4102553600, | |
| "step": 15650 | |
| }, | |
| { | |
| "epoch": 0.10560603232419034, | |
| "grad_norm": 0.6631893515586853, | |
| "learning_rate": 0.00028718107691552496, | |
| "loss": 12.2943, | |
| "num_input_tokens_seen": 4115660800, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.10594235726789795, | |
| "grad_norm": 0.6634914875030518, | |
| "learning_rate": 0.0002855798780419138, | |
| "loss": 12.2738, | |
| "num_input_tokens_seen": 4128768000, | |
| "step": 15750 | |
| }, | |
| { | |
| "epoch": 0.10627868221160557, | |
| "grad_norm": 0.6240889430046082, | |
| "learning_rate": 0.00028397909080396527, | |
| "loss": 12.3316, | |
| "num_input_tokens_seen": 4141875200, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.10661500715531318, | |
| "grad_norm": 0.6263941526412964, | |
| "learning_rate": 0.00028237876089762574, | |
| "loss": 12.2874, | |
| "num_input_tokens_seen": 4154982400, | |
| "step": 15850 | |
| }, | |
| { | |
| "epoch": 0.10695133209902079, | |
| "grad_norm": 0.629359245300293, | |
| "learning_rate": 0.00028077893400578615, | |
| "loss": 12.3043, | |
| "num_input_tokens_seen": 4168089600, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.1072876570427284, | |
| "grad_norm": 0.6163947582244873, | |
| "learning_rate": 0.0002791796557969787, | |
| "loss": 12.3009, | |
| "num_input_tokens_seen": 4181196800, | |
| "step": 15950 | |
| }, | |
| { | |
| "epoch": 0.10762398198643601, | |
| "grad_norm": 0.6394808888435364, | |
| "learning_rate": 0.0002775809719240727, | |
| "loss": 12.2584, | |
| "num_input_tokens_seen": 4194304000, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.10762398198643601, | |
| "eval_loss": 2.984212875366211, | |
| "eval_runtime": 142.4659, | |
| "eval_samples_per_second": 35.096, | |
| "eval_steps_per_second": 8.774, | |
| "num_input_tokens_seen": 4194304000, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.10796030693014362, | |
| "grad_norm": 0.6504441499710083, | |
| "learning_rate": 0.00027598292802297203, | |
| "loss": 12.301, | |
| "num_input_tokens_seen": 4207411200, | |
| "step": 16050 | |
| }, | |
| { | |
| "epoch": 0.10829663187385125, | |
| "grad_norm": 0.6610515117645264, | |
| "learning_rate": 0.00027438556971131137, | |
| "loss": 12.2809, | |
| "num_input_tokens_seen": 4220518400, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.10863295681755886, | |
| "grad_norm": 0.6400002837181091, | |
| "learning_rate": 0.00027278894258715535, | |
| "loss": 12.2821, | |
| "num_input_tokens_seen": 4233625600, | |
| "step": 16150 | |
| }, | |
| { | |
| "epoch": 0.10896928176126647, | |
| "grad_norm": 0.6517115831375122, | |
| "learning_rate": 0.00027119309222769546, | |
| "loss": 12.2722, | |
| "num_input_tokens_seen": 4246732800, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.10930560670497408, | |
| "grad_norm": 0.6387389898300171, | |
| "learning_rate": 0.0002695980641879502, | |
| "loss": 12.2715, | |
| "num_input_tokens_seen": 4259840000, | |
| "step": 16250 | |
| }, | |
| { | |
| "epoch": 0.10964193164868169, | |
| "grad_norm": 0.6440519094467163, | |
| "learning_rate": 0.0002680039039994639, | |
| "loss": 12.25, | |
| "num_input_tokens_seen": 4272947200, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.1099782565923893, | |
| "grad_norm": 0.6389286518096924, | |
| "learning_rate": 0.0002664106571690076, | |
| "loss": 12.2565, | |
| "num_input_tokens_seen": 4286054400, | |
| "step": 16350 | |
| }, | |
| { | |
| "epoch": 0.11031458153609691, | |
| "grad_norm": 0.6398110389709473, | |
| "learning_rate": 0.00026481836917727946, | |
| "loss": 12.2356, | |
| "num_input_tokens_seen": 4299161600, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.11065090647980452, | |
| "grad_norm": 0.6471937298774719, | |
| "learning_rate": 0.00026322708547760676, | |
| "loss": 12.269, | |
| "num_input_tokens_seen": 4312268800, | |
| "step": 16450 | |
| }, | |
| { | |
| "epoch": 0.11098723142351213, | |
| "grad_norm": 0.6105075478553772, | |
| "learning_rate": 0.00026163685149464816, | |
| "loss": 12.2762, | |
| "num_input_tokens_seen": 4325376000, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.11098723142351213, | |
| "eval_loss": 2.9754507541656494, | |
| "eval_runtime": 142.3479, | |
| "eval_samples_per_second": 35.125, | |
| "eval_steps_per_second": 8.781, | |
| "num_input_tokens_seen": 4325376000, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.11132355636721976, | |
| "grad_norm": 0.6111390590667725, | |
| "learning_rate": 0.00026004771262309764, | |
| "loss": 12.2253, | |
| "num_input_tokens_seen": 4338483200, | |
| "step": 16550 | |
| }, | |
| { | |
| "epoch": 0.11165988131092737, | |
| "grad_norm": 0.6871252059936523, | |
| "learning_rate": 0.0002584597142263877, | |
| "loss": 12.2595, | |
| "num_input_tokens_seen": 4351590400, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.11199620625463498, | |
| "grad_norm": 0.6608724594116211, | |
| "learning_rate": 0.00025687290163539547, | |
| "loss": 12.2838, | |
| "num_input_tokens_seen": 4364697600, | |
| "step": 16650 | |
| }, | |
| { | |
| "epoch": 0.11233253119834259, | |
| "grad_norm": 0.634148895740509, | |
| "learning_rate": 0.0002552873201471476, | |
| "loss": 12.2522, | |
| "num_input_tokens_seen": 4377804800, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.1126688561420502, | |
| "grad_norm": 0.6481145620346069, | |
| "learning_rate": 0.00025370301502352825, | |
| "loss": 12.2185, | |
| "num_input_tokens_seen": 4390912000, | |
| "step": 16750 | |
| }, | |
| { | |
| "epoch": 0.11300518108575781, | |
| "grad_norm": 0.6411675810813904, | |
| "learning_rate": 0.0002521200314899863, | |
| "loss": 12.2566, | |
| "num_input_tokens_seen": 4404019200, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.11334150602946542, | |
| "grad_norm": 0.6985258460044861, | |
| "learning_rate": 0.00025053841473424447, | |
| "loss": 12.3036, | |
| "num_input_tokens_seen": 4417126400, | |
| "step": 16850 | |
| }, | |
| { | |
| "epoch": 0.11367783097317304, | |
| "grad_norm": 0.6223846673965454, | |
| "learning_rate": 0.0002489582099050102, | |
| "loss": 12.1942, | |
| "num_input_tokens_seen": 4430233600, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.11401415591688065, | |
| "grad_norm": 0.637690007686615, | |
| "learning_rate": 0.00024737946211068554, | |
| "loss": 12.2711, | |
| "num_input_tokens_seen": 4443340800, | |
| "step": 16950 | |
| }, | |
| { | |
| "epoch": 0.11435048086058827, | |
| "grad_norm": 0.6184976696968079, | |
| "learning_rate": 0.00024580221641808033, | |
| "loss": 12.2252, | |
| "num_input_tokens_seen": 4456448000, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.11435048086058827, | |
| "eval_loss": 2.9684932231903076, | |
| "eval_runtime": 142.7182, | |
| "eval_samples_per_second": 35.034, | |
| "eval_steps_per_second": 8.759, | |
| "num_input_tokens_seen": 4456448000, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.11468680580429588, | |
| "grad_norm": 0.618519127368927, | |
| "learning_rate": 0.0002442265178511256, | |
| "loss": 12.2066, | |
| "num_input_tokens_seen": 4469555200, | |
| "step": 17050 | |
| }, | |
| { | |
| "epoch": 0.1150231307480035, | |
| "grad_norm": 0.6514145731925964, | |
| "learning_rate": 0.00024265241138958835, | |
| "loss": 12.2228, | |
| "num_input_tokens_seen": 4482662400, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.1153594556917111, | |
| "grad_norm": 0.6823457479476929, | |
| "learning_rate": 0.00024107994196778714, | |
| "loss": 12.2507, | |
| "num_input_tokens_seen": 4495769600, | |
| "step": 17150 | |
| }, | |
| { | |
| "epoch": 0.11569578063541872, | |
| "grad_norm": 0.6243106722831726, | |
| "learning_rate": 0.0002395091544733101, | |
| "loss": 12.1857, | |
| "num_input_tokens_seen": 4508876800, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.11603210557912633, | |
| "grad_norm": 0.6370251774787903, | |
| "learning_rate": 0.00023794009374573274, | |
| "loss": 12.2309, | |
| "num_input_tokens_seen": 4521984000, | |
| "step": 17250 | |
| }, | |
| { | |
| "epoch": 0.11636843052283394, | |
| "grad_norm": 0.6504274010658264, | |
| "learning_rate": 0.00023637280457533902, | |
| "loss": 12.2132, | |
| "num_input_tokens_seen": 4535091200, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.11670475546654155, | |
| "grad_norm": 0.6156638860702515, | |
| "learning_rate": 0.00023480733170184158, | |
| "loss": 12.199, | |
| "num_input_tokens_seen": 4548198400, | |
| "step": 17350 | |
| }, | |
| { | |
| "epoch": 0.11704108041024916, | |
| "grad_norm": 0.6292795538902283, | |
| "learning_rate": 0.0002332437198131057, | |
| "loss": 12.2122, | |
| "num_input_tokens_seen": 4561305600, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.11737740535395678, | |
| "grad_norm": 0.6368102431297302, | |
| "learning_rate": 0.00023168201354387266, | |
| "loss": 12.2453, | |
| "num_input_tokens_seen": 4574412800, | |
| "step": 17450 | |
| }, | |
| { | |
| "epoch": 0.1177137302976644, | |
| "grad_norm": 0.6373352408409119, | |
| "learning_rate": 0.00023012225747448645, | |
| "loss": 12.2031, | |
| "num_input_tokens_seen": 4587520000, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.1177137302976644, | |
| "eval_loss": 2.961634635925293, | |
| "eval_runtime": 142.67, | |
| "eval_samples_per_second": 35.046, | |
| "eval_steps_per_second": 8.761, | |
| "num_input_tokens_seen": 4587520000, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.118050055241372, | |
| "grad_norm": 0.6781473755836487, | |
| "learning_rate": 0.0002285644961296205, | |
| "loss": 12.1939, | |
| "num_input_tokens_seen": 4600627200, | |
| "step": 17550 | |
| }, | |
| { | |
| "epoch": 0.11838638018507962, | |
| "grad_norm": 0.6434431076049805, | |
| "learning_rate": 0.0002270087739770074, | |
| "loss": 12.1876, | |
| "num_input_tokens_seen": 4613734400, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.11872270512878723, | |
| "grad_norm": 0.6625823974609375, | |
| "learning_rate": 0.00022545513542616865, | |
| "loss": 12.1683, | |
| "num_input_tokens_seen": 4626841600, | |
| "step": 17650 | |
| }, | |
| { | |
| "epoch": 0.11905903007249484, | |
| "grad_norm": 0.6367326974868774, | |
| "learning_rate": 0.0002239036248271478, | |
| "loss": 12.1769, | |
| "num_input_tokens_seen": 4639948800, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.11939535501620245, | |
| "grad_norm": 0.648065447807312, | |
| "learning_rate": 0.00022235428646924372, | |
| "loss": 12.2213, | |
| "num_input_tokens_seen": 4653056000, | |
| "step": 17750 | |
| }, | |
| { | |
| "epoch": 0.11973167995991006, | |
| "grad_norm": 0.648695170879364, | |
| "learning_rate": 0.00022080716457974705, | |
| "loss": 12.1699, | |
| "num_input_tokens_seen": 4666163200, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.12006800490361767, | |
| "grad_norm": 0.6311103105545044, | |
| "learning_rate": 0.00021926230332267694, | |
| "loss": 12.1912, | |
| "num_input_tokens_seen": 4679270400, | |
| "step": 17850 | |
| }, | |
| { | |
| "epoch": 0.1204043298473253, | |
| "grad_norm": 0.6318332552909851, | |
| "learning_rate": 0.00021771974679752094, | |
| "loss": 12.1242, | |
| "num_input_tokens_seen": 4692377600, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.12074065479103291, | |
| "grad_norm": 0.6513566374778748, | |
| "learning_rate": 0.0002161795390379756, | |
| "loss": 12.2068, | |
| "num_input_tokens_seen": 4705484800, | |
| "step": 17950 | |
| }, | |
| { | |
| "epoch": 0.12107697973474052, | |
| "grad_norm": 0.6865115761756897, | |
| "learning_rate": 0.00021464172401069027, | |
| "loss": 12.1477, | |
| "num_input_tokens_seen": 4718592000, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.12107697973474052, | |
| "eval_loss": 2.954843044281006, | |
| "eval_runtime": 142.8465, | |
| "eval_samples_per_second": 35.003, | |
| "eval_steps_per_second": 8.751, | |
| "num_input_tokens_seen": 4718592000, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.12141330467844813, | |
| "grad_norm": 0.622513473033905, | |
| "learning_rate": 0.00021310634561401109, | |
| "loss": 12.1664, | |
| "num_input_tokens_seen": 4731699200, | |
| "step": 18050 | |
| }, | |
| { | |
| "epoch": 0.12174962962215574, | |
| "grad_norm": 0.6387473344802856, | |
| "learning_rate": 0.0002115734476767287, | |
| "loss": 12.1838, | |
| "num_input_tokens_seen": 4744806400, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.12208595456586335, | |
| "grad_norm": 0.6974210739135742, | |
| "learning_rate": 0.00021004307395682648, | |
| "loss": 12.201, | |
| "num_input_tokens_seen": 4757913600, | |
| "step": 18150 | |
| }, | |
| { | |
| "epoch": 0.12242227950957096, | |
| "grad_norm": 0.6665675640106201, | |
| "learning_rate": 0.00020851526814023185, | |
| "loss": 12.1154, | |
| "num_input_tokens_seen": 4771020800, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.12275860445327857, | |
| "grad_norm": 0.6340165734291077, | |
| "learning_rate": 0.00020699007383956895, | |
| "loss": 12.19, | |
| "num_input_tokens_seen": 4784128000, | |
| "step": 18250 | |
| }, | |
| { | |
| "epoch": 0.12309492939698619, | |
| "grad_norm": 0.6360442638397217, | |
| "learning_rate": 0.00020546753459291378, | |
| "loss": 12.1872, | |
| "num_input_tokens_seen": 4797235200, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.12343125434069381, | |
| "grad_norm": 0.6116852760314941, | |
| "learning_rate": 0.00020394769386255162, | |
| "loss": 12.1645, | |
| "num_input_tokens_seen": 4810342400, | |
| "step": 18350 | |
| }, | |
| { | |
| "epoch": 0.12376757928440142, | |
| "grad_norm": 0.6432573795318604, | |
| "learning_rate": 0.00020243059503373588, | |
| "loss": 12.1537, | |
| "num_input_tokens_seen": 4823449600, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.12410390422810903, | |
| "grad_norm": 0.6480187773704529, | |
| "learning_rate": 0.00020091628141344996, | |
| "loss": 12.155, | |
| "num_input_tokens_seen": 4836556800, | |
| "step": 18450 | |
| }, | |
| { | |
| "epoch": 0.12444022917181664, | |
| "grad_norm": 0.643993616104126, | |
| "learning_rate": 0.00019940479622917068, | |
| "loss": 12.1604, | |
| "num_input_tokens_seen": 4849664000, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.12444022917181664, | |
| "eval_loss": 2.9489145278930664, | |
| "eval_runtime": 142.7319, | |
| "eval_samples_per_second": 35.031, | |
| "eval_steps_per_second": 8.758, | |
| "num_input_tokens_seen": 4849664000, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.12477655411552425, | |
| "grad_norm": 0.6500803232192993, | |
| "learning_rate": 0.00019789618262763508, | |
| "loss": 12.1604, | |
| "num_input_tokens_seen": 4862771200, | |
| "step": 18550 | |
| }, | |
| { | |
| "epoch": 0.12511287905923188, | |
| "grad_norm": 0.6314743161201477, | |
| "learning_rate": 0.00019639048367360774, | |
| "loss": 12.1107, | |
| "num_input_tokens_seen": 4875878400, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.1254492040029395, | |
| "grad_norm": 0.6902073621749878, | |
| "learning_rate": 0.00019488774234865217, | |
| "loss": 12.1634, | |
| "num_input_tokens_seen": 4888985600, | |
| "step": 18650 | |
| }, | |
| { | |
| "epoch": 0.1257855289466471, | |
| "grad_norm": 0.6349673867225647, | |
| "learning_rate": 0.00019338800154990337, | |
| "loss": 12.1828, | |
| "num_input_tokens_seen": 4902092800, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.1261218538903547, | |
| "grad_norm": 0.639392614364624, | |
| "learning_rate": 0.000191891304088844, | |
| "loss": 12.1314, | |
| "num_input_tokens_seen": 4915200000, | |
| "step": 18750 | |
| }, | |
| { | |
| "epoch": 0.12645817883406232, | |
| "grad_norm": 0.6466573476791382, | |
| "learning_rate": 0.00019039769269008148, | |
| "loss": 12.1521, | |
| "num_input_tokens_seen": 4928307200, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.12679450377776993, | |
| "grad_norm": 0.6457189917564392, | |
| "learning_rate": 0.00018890720999012895, | |
| "loss": 12.1631, | |
| "num_input_tokens_seen": 4941414400, | |
| "step": 18850 | |
| }, | |
| { | |
| "epoch": 0.12713082872147755, | |
| "grad_norm": 0.648733377456665, | |
| "learning_rate": 0.00018741989853618779, | |
| "loss": 12.1553, | |
| "num_input_tokens_seen": 4954521600, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.12746715366518516, | |
| "grad_norm": 0.6314489841461182, | |
| "learning_rate": 0.00018593580078493335, | |
| "loss": 12.1703, | |
| "num_input_tokens_seen": 4967628800, | |
| "step": 18950 | |
| }, | |
| { | |
| "epoch": 0.12780347860889277, | |
| "grad_norm": 0.6238834857940674, | |
| "learning_rate": 0.0001844549591013027, | |
| "loss": 12.0931, | |
| "num_input_tokens_seen": 4980736000, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.12780347860889277, | |
| "eval_loss": 2.943131923675537, | |
| "eval_runtime": 142.8812, | |
| "eval_samples_per_second": 34.994, | |
| "eval_steps_per_second": 8.749, | |
| "num_input_tokens_seen": 4980736000, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.12813980355260038, | |
| "grad_norm": 0.6120157241821289, | |
| "learning_rate": 0.00018297741575728593, | |
| "loss": 12.1415, | |
| "num_input_tokens_seen": 4993843200, | |
| "step": 19050 | |
| }, | |
| { | |
| "epoch": 0.128476128496308, | |
| "grad_norm": 0.6346642374992371, | |
| "learning_rate": 0.00018150321293071843, | |
| "loss": 12.1464, | |
| "num_input_tokens_seen": 5006950400, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.1288124534400156, | |
| "grad_norm": 0.6268289685249329, | |
| "learning_rate": 0.00018003239270407775, | |
| "loss": 12.1105, | |
| "num_input_tokens_seen": 5020057600, | |
| "step": 19150 | |
| }, | |
| { | |
| "epoch": 0.1291487783837232, | |
| "grad_norm": 0.6437589526176453, | |
| "learning_rate": 0.00017856499706328183, | |
| "loss": 12.1208, | |
| "num_input_tokens_seen": 5033164800, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.12948510332743082, | |
| "grad_norm": 0.6311147809028625, | |
| "learning_rate": 0.00017710106789649096, | |
| "loss": 12.1137, | |
| "num_input_tokens_seen": 5046272000, | |
| "step": 19250 | |
| }, | |
| { | |
| "epoch": 0.12982142827113843, | |
| "grad_norm": 0.646539568901062, | |
| "learning_rate": 0.00017564064699291133, | |
| "loss": 12.1824, | |
| "num_input_tokens_seen": 5059379200, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.13015775321484604, | |
| "grad_norm": 0.6385849714279175, | |
| "learning_rate": 0.00017418377604160295, | |
| "loss": 12.1106, | |
| "num_input_tokens_seen": 5072486400, | |
| "step": 19350 | |
| }, | |
| { | |
| "epoch": 0.13049407815855366, | |
| "grad_norm": 0.6449156403541565, | |
| "learning_rate": 0.0001727304966302887, | |
| "loss": 12.0996, | |
| "num_input_tokens_seen": 5085593600, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.1308304031022613, | |
| "grad_norm": 0.6219010949134827, | |
| "learning_rate": 0.0001712808502441682, | |
| "loss": 12.1306, | |
| "num_input_tokens_seen": 5098700800, | |
| "step": 19450 | |
| }, | |
| { | |
| "epoch": 0.1311667280459689, | |
| "grad_norm": 0.6273418664932251, | |
| "learning_rate": 0.00016983487826473256, | |
| "loss": 12.0719, | |
| "num_input_tokens_seen": 5111808000, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.1311667280459689, | |
| "eval_loss": 2.937514066696167, | |
| "eval_runtime": 142.6532, | |
| "eval_samples_per_second": 35.05, | |
| "eval_steps_per_second": 8.763, | |
| "num_input_tokens_seen": 5111808000, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.13150305298967652, | |
| "grad_norm": 0.6366037130355835, | |
| "learning_rate": 0.00016839262196858374, | |
| "loss": 12.1143, | |
| "num_input_tokens_seen": 5124915200, | |
| "step": 19550 | |
| }, | |
| { | |
| "epoch": 0.13183937793338413, | |
| "grad_norm": 0.6395111083984375, | |
| "learning_rate": 0.00016695412252625596, | |
| "loss": 12.0524, | |
| "num_input_tokens_seen": 5138022400, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.13217570287709174, | |
| "grad_norm": 0.639202892780304, | |
| "learning_rate": 0.0001655194210010409, | |
| "loss": 12.1006, | |
| "num_input_tokens_seen": 5151129600, | |
| "step": 19650 | |
| }, | |
| { | |
| "epoch": 0.13251202782079935, | |
| "grad_norm": 0.6547548174858093, | |
| "learning_rate": 0.00016408855834781487, | |
| "loss": 12.0684, | |
| "num_input_tokens_seen": 5164236800, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.13284835276450696, | |
| "grad_norm": 0.6669015884399414, | |
| "learning_rate": 0.00016266157541187034, | |
| "loss": 12.1204, | |
| "num_input_tokens_seen": 5177344000, | |
| "step": 19750 | |
| }, | |
| { | |
| "epoch": 0.13318467770821457, | |
| "grad_norm": 0.637744665145874, | |
| "learning_rate": 0.00016123851292774947, | |
| "loss": 12.1164, | |
| "num_input_tokens_seen": 5190451200, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.13352100265192218, | |
| "grad_norm": 0.6337763667106628, | |
| "learning_rate": 0.00015981941151808137, | |
| "loss": 12.1213, | |
| "num_input_tokens_seen": 5203558400, | |
| "step": 19850 | |
| }, | |
| { | |
| "epoch": 0.1338573275956298, | |
| "grad_norm": 0.651337742805481, | |
| "learning_rate": 0.0001584043116924231, | |
| "loss": 12.1115, | |
| "num_input_tokens_seen": 5216665600, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.1341936525393374, | |
| "grad_norm": 0.6313726902008057, | |
| "learning_rate": 0.00015699325384610244, | |
| "loss": 12.1078, | |
| "num_input_tokens_seen": 5229772800, | |
| "step": 19950 | |
| }, | |
| { | |
| "epoch": 0.13452997748304502, | |
| "grad_norm": 0.6925057768821716, | |
| "learning_rate": 0.00015558627825906524, | |
| "loss": 12.0672, | |
| "num_input_tokens_seen": 5242880000, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.13452997748304502, | |
| "eval_loss": 2.931644916534424, | |
| "eval_runtime": 143.0863, | |
| "eval_samples_per_second": 34.944, | |
| "eval_steps_per_second": 8.736, | |
| "num_input_tokens_seen": 5242880000, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.13486630242675263, | |
| "grad_norm": 0.6415194272994995, | |
| "learning_rate": 0.00015418342509472535, | |
| "loss": 12.1005, | |
| "num_input_tokens_seen": 5255987200, | |
| "step": 20050 | |
| }, | |
| { | |
| "epoch": 0.13520262737046024, | |
| "grad_norm": 0.6401641368865967, | |
| "learning_rate": 0.00015278473439881874, | |
| "loss": 12.0935, | |
| "num_input_tokens_seen": 5269094400, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.13553895231416785, | |
| "grad_norm": 0.6700222492218018, | |
| "learning_rate": 0.0001513902460982592, | |
| "loss": 12.0946, | |
| "num_input_tokens_seen": 5282201600, | |
| "step": 20150 | |
| }, | |
| { | |
| "epoch": 0.13587527725787546, | |
| "grad_norm": 0.6184066534042358, | |
| "learning_rate": 0.00015000000000000004, | |
| "loss": 12.058, | |
| "num_input_tokens_seen": 5295308800, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.13621160220158307, | |
| "grad_norm": 0.6642903685569763, | |
| "learning_rate": 0.00014861403578989629, | |
| "loss": 12.0421, | |
| "num_input_tokens_seen": 5308416000, | |
| "step": 20250 | |
| }, | |
| { | |
| "epoch": 0.13654792714529068, | |
| "grad_norm": 0.651897668838501, | |
| "learning_rate": 0.00014723239303157307, | |
| "loss": 12.0393, | |
| "num_input_tokens_seen": 5321523200, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.1368842520889983, | |
| "grad_norm": 0.616648256778717, | |
| "learning_rate": 0.00014585511116529528, | |
| "loss": 12.0737, | |
| "num_input_tokens_seen": 5334630400, | |
| "step": 20350 | |
| }, | |
| { | |
| "epoch": 0.13722057703270593, | |
| "grad_norm": 0.6298686861991882, | |
| "learning_rate": 0.00014448222950684246, | |
| "loss": 12.0721, | |
| "num_input_tokens_seen": 5347737600, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.13755690197641354, | |
| "grad_norm": 0.6637253165245056, | |
| "learning_rate": 0.00014311378724638605, | |
| "loss": 12.0921, | |
| "num_input_tokens_seen": 5360844800, | |
| "step": 20450 | |
| }, | |
| { | |
| "epoch": 0.13789322692012115, | |
| "grad_norm": 0.6153833866119385, | |
| "learning_rate": 0.0001417498234473706, | |
| "loss": 12.0664, | |
| "num_input_tokens_seen": 5373952000, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.13789322692012115, | |
| "eval_loss": 2.9268288612365723, | |
| "eval_runtime": 143.0059, | |
| "eval_samples_per_second": 34.964, | |
| "eval_steps_per_second": 8.741, | |
| "num_input_tokens_seen": 5373952000, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.13822955186382876, | |
| "grad_norm": 0.6536301374435425, | |
| "learning_rate": 0.00014039037704539906, | |
| "loss": 12.0644, | |
| "num_input_tokens_seen": 5387059200, | |
| "step": 20550 | |
| }, | |
| { | |
| "epoch": 0.13856587680753638, | |
| "grad_norm": 0.678833544254303, | |
| "learning_rate": 0.00013903548684712116, | |
| "loss": 12.0616, | |
| "num_input_tokens_seen": 5400166400, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.138902201751244, | |
| "grad_norm": 0.6597055792808533, | |
| "learning_rate": 0.00013768519152912537, | |
| "loss": 12.0914, | |
| "num_input_tokens_seen": 5413273600, | |
| "step": 20650 | |
| }, | |
| { | |
| "epoch": 0.1392385266949516, | |
| "grad_norm": 0.6703686714172363, | |
| "learning_rate": 0.00013633952963683542, | |
| "loss": 12.0582, | |
| "num_input_tokens_seen": 5426380800, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.1395748516386592, | |
| "grad_norm": 0.6616584062576294, | |
| "learning_rate": 0.00013499853958340923, | |
| "loss": 12.105, | |
| "num_input_tokens_seen": 5439488000, | |
| "step": 20750 | |
| }, | |
| { | |
| "epoch": 0.13991117658236682, | |
| "grad_norm": 0.6584370136260986, | |
| "learning_rate": 0.00013366225964864313, | |
| "loss": 12.0616, | |
| "num_input_tokens_seen": 5452595200, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.14024750152607443, | |
| "grad_norm": 0.6238560676574707, | |
| "learning_rate": 0.00013233072797787847, | |
| "loss": 12.074, | |
| "num_input_tokens_seen": 5465702400, | |
| "step": 20850 | |
| }, | |
| { | |
| "epoch": 0.14058382646978204, | |
| "grad_norm": 0.6119787096977234, | |
| "learning_rate": 0.00013100398258091337, | |
| "loss": 12.0441, | |
| "num_input_tokens_seen": 5478809600, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.14092015141348965, | |
| "grad_norm": 0.6162968873977661, | |
| "learning_rate": 0.00012968206133091707, | |
| "loss": 12.0726, | |
| "num_input_tokens_seen": 5491916800, | |
| "step": 20950 | |
| }, | |
| { | |
| "epoch": 0.14125647635719726, | |
| "grad_norm": 0.6324203014373779, | |
| "learning_rate": 0.00012836500196334916, | |
| "loss": 12.029, | |
| "num_input_tokens_seen": 5505024000, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.14125647635719726, | |
| "eval_loss": 2.9219326972961426, | |
| "eval_runtime": 142.444, | |
| "eval_samples_per_second": 35.102, | |
| "eval_steps_per_second": 8.775, | |
| "num_input_tokens_seen": 5505024000, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.14159280130090487, | |
| "grad_norm": 0.6728281378746033, | |
| "learning_rate": 0.0001270528420748823, | |
| "loss": 12.0576, | |
| "num_input_tokens_seen": 5518131200, | |
| "step": 21050 | |
| }, | |
| { | |
| "epoch": 0.14192912624461249, | |
| "grad_norm": 0.6371399164199829, | |
| "learning_rate": 0.0001257456191223292, | |
| "loss": 12.0809, | |
| "num_input_tokens_seen": 5531238400, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.1422654511883201, | |
| "grad_norm": 0.6416388750076294, | |
| "learning_rate": 0.00012444337042157285, | |
| "loss": 12.0472, | |
| "num_input_tokens_seen": 5544345600, | |
| "step": 21150 | |
| }, | |
| { | |
| "epoch": 0.1426017761320277, | |
| "grad_norm": 0.672295093536377, | |
| "learning_rate": 0.00012314613314650207, | |
| "loss": 12.0615, | |
| "num_input_tokens_seen": 5557452800, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.14293810107573532, | |
| "grad_norm": 0.6460967063903809, | |
| "learning_rate": 0.00012185394432794955, | |
| "loss": 12.0439, | |
| "num_input_tokens_seen": 5570560000, | |
| "step": 21250 | |
| }, | |
| { | |
| "epoch": 0.14327442601944296, | |
| "grad_norm": 0.6483781337738037, | |
| "learning_rate": 0.0001205668408526352, | |
| "loss": 12.0767, | |
| "num_input_tokens_seen": 5583667200, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.14361075096315057, | |
| "grad_norm": 0.6515306830406189, | |
| "learning_rate": 0.00011928485946211334, | |
| "loss": 12.0398, | |
| "num_input_tokens_seen": 5596774400, | |
| "step": 21350 | |
| }, | |
| { | |
| "epoch": 0.14394707590685818, | |
| "grad_norm": 0.6355323791503906, | |
| "learning_rate": 0.00011800803675172337, | |
| "loss": 12.0792, | |
| "num_input_tokens_seen": 5609881600, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 0.1442834008505658, | |
| "grad_norm": 0.6724342107772827, | |
| "learning_rate": 0.00011673640916954571, | |
| "loss": 12.0238, | |
| "num_input_tokens_seen": 5622988800, | |
| "step": 21450 | |
| }, | |
| { | |
| "epoch": 0.1446197257942734, | |
| "grad_norm": 0.6570594310760498, | |
| "learning_rate": 0.00011547001301536085, | |
| "loss": 12.0514, | |
| "num_input_tokens_seen": 5636096000, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.1446197257942734, | |
| "eval_loss": 2.91679310798645, | |
| "eval_runtime": 142.6116, | |
| "eval_samples_per_second": 35.06, | |
| "eval_steps_per_second": 8.765, | |
| "num_input_tokens_seen": 5636096000, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.144956050737981, | |
| "grad_norm": 0.6420052647590637, | |
| "learning_rate": 0.00011420888443961337, | |
| "loss": 12.02, | |
| "num_input_tokens_seen": 5649203200, | |
| "step": 21550 | |
| }, | |
| { | |
| "epoch": 0.14529237568168862, | |
| "grad_norm": 0.6295548677444458, | |
| "learning_rate": 0.00011295305944237995, | |
| "loss": 12.0275, | |
| "num_input_tokens_seen": 5662310400, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 0.14562870062539623, | |
| "grad_norm": 0.6434178352355957, | |
| "learning_rate": 0.00011170257387234198, | |
| "loss": 12.0421, | |
| "num_input_tokens_seen": 5675417600, | |
| "step": 21650 | |
| }, | |
| { | |
| "epoch": 0.14596502556910385, | |
| "grad_norm": 0.6139717102050781, | |
| "learning_rate": 0.0001104574634257616, | |
| "loss": 12.0342, | |
| "num_input_tokens_seen": 5688524800, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 0.14630135051281146, | |
| "grad_norm": 0.6519197225570679, | |
| "learning_rate": 0.00010921776364546347, | |
| "loss": 12.0328, | |
| "num_input_tokens_seen": 5701632000, | |
| "step": 21750 | |
| }, | |
| { | |
| "epoch": 0.14663767545651907, | |
| "grad_norm": 0.6653149724006653, | |
| "learning_rate": 0.00010798350991981948, | |
| "loss": 12.0151, | |
| "num_input_tokens_seen": 5714739200, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 0.14697400040022668, | |
| "grad_norm": 0.6633841395378113, | |
| "learning_rate": 0.0001067547374817392, | |
| "loss": 11.9882, | |
| "num_input_tokens_seen": 5727846400, | |
| "step": 21850 | |
| }, | |
| { | |
| "epoch": 0.1473103253439343, | |
| "grad_norm": 0.6165183186531067, | |
| "learning_rate": 0.00010553148140766353, | |
| "loss": 12.0242, | |
| "num_input_tokens_seen": 5740953600, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 0.1476466502876419, | |
| "grad_norm": 0.6443773508071899, | |
| "learning_rate": 0.00010431377661656374, | |
| "loss": 12.0166, | |
| "num_input_tokens_seen": 5754060800, | |
| "step": 21950 | |
| }, | |
| { | |
| "epoch": 0.1479829752313495, | |
| "grad_norm": 0.6805723905563354, | |
| "learning_rate": 0.00010310165786894456, | |
| "loss": 12.0284, | |
| "num_input_tokens_seen": 5767168000, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.1479829752313495, | |
| "eval_loss": 2.9135851860046387, | |
| "eval_runtime": 143.3087, | |
| "eval_samples_per_second": 34.89, | |
| "eval_steps_per_second": 8.722, | |
| "num_input_tokens_seen": 5767168000, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.14831930017505712, | |
| "grad_norm": 0.6574228405952454, | |
| "learning_rate": 0.00010189515976585224, | |
| "loss": 11.9953, | |
| "num_input_tokens_seen": 5780275200, | |
| "step": 22050 | |
| }, | |
| { | |
| "epoch": 0.14865562511876473, | |
| "grad_norm": 0.630247175693512, | |
| "learning_rate": 0.00010069431674788618, | |
| "loss": 12.0309, | |
| "num_input_tokens_seen": 5793382400, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 0.14899195006247234, | |
| "grad_norm": 0.6254024505615234, | |
| "learning_rate": 9.949916309421655e-05, | |
| "loss": 11.9972, | |
| "num_input_tokens_seen": 5806489600, | |
| "step": 22150 | |
| }, | |
| { | |
| "epoch": 0.14932827500617998, | |
| "grad_norm": 0.6164761781692505, | |
| "learning_rate": 9.830973292160493e-05, | |
| "loss": 12.0382, | |
| "num_input_tokens_seen": 5819596800, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 0.1496645999498876, | |
| "grad_norm": 0.6174560189247131, | |
| "learning_rate": 9.712606018343136e-05, | |
| "loss": 11.981, | |
| "num_input_tokens_seen": 5832704000, | |
| "step": 22250 | |
| }, | |
| { | |
| "epoch": 0.1500009248935952, | |
| "grad_norm": 0.6346741914749146, | |
| "learning_rate": 9.594817866872411e-05, | |
| "loss": 12.0161, | |
| "num_input_tokens_seen": 5845811200, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 0.15033724983730282, | |
| "grad_norm": 0.6521451473236084, | |
| "learning_rate": 9.477612200119616e-05, | |
| "loss": 12.0022, | |
| "num_input_tokens_seen": 5858918400, | |
| "step": 22350 | |
| }, | |
| { | |
| "epoch": 0.15067357478101043, | |
| "grad_norm": 0.6211933493614197, | |
| "learning_rate": 9.360992363828442e-05, | |
| "loss": 12.0695, | |
| "num_input_tokens_seen": 5872025600, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 0.15100989972471804, | |
| "grad_norm": 0.6488197445869446, | |
| "learning_rate": 9.244961687019529e-05, | |
| "loss": 12.0477, | |
| "num_input_tokens_seen": 5885132800, | |
| "step": 22450 | |
| }, | |
| { | |
| "epoch": 0.15134622466842565, | |
| "grad_norm": 0.6073492169380188, | |
| "learning_rate": 9.129523481895408e-05, | |
| "loss": 11.9863, | |
| "num_input_tokens_seen": 5898240000, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.15134622466842565, | |
| "eval_loss": 2.9087352752685547, | |
| "eval_runtime": 141.7599, | |
| "eval_samples_per_second": 35.271, | |
| "eval_steps_per_second": 8.818, | |
| "num_input_tokens_seen": 5898240000, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.15168254961213326, | |
| "grad_norm": 0.6158032417297363, | |
| "learning_rate": 9.014681043745983e-05, | |
| "loss": 12.0428, | |
| "num_input_tokens_seen": 5911347200, | |
| "step": 22550 | |
| }, | |
| { | |
| "epoch": 0.15201887455584087, | |
| "grad_norm": 0.6146510243415833, | |
| "learning_rate": 8.900437650854409e-05, | |
| "loss": 12.0035, | |
| "num_input_tokens_seen": 5924454400, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 0.15235519949954848, | |
| "grad_norm": 0.7664083242416382, | |
| "learning_rate": 8.786796564403575e-05, | |
| "loss": 12.0481, | |
| "num_input_tokens_seen": 5937561600, | |
| "step": 22650 | |
| }, | |
| { | |
| "epoch": 0.1526915244432561, | |
| "grad_norm": 0.6329470872879028, | |
| "learning_rate": 8.673761028382955e-05, | |
| "loss": 11.9683, | |
| "num_input_tokens_seen": 5950668800, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 0.1530278493869637, | |
| "grad_norm": 0.617677628993988, | |
| "learning_rate": 8.561334269496019e-05, | |
| "loss": 11.9993, | |
| "num_input_tokens_seen": 5963776000, | |
| "step": 22750 | |
| }, | |
| { | |
| "epoch": 0.15336417433067132, | |
| "grad_norm": 0.6368398070335388, | |
| "learning_rate": 8.449519497068174e-05, | |
| "loss": 11.9881, | |
| "num_input_tokens_seen": 5976883200, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 0.15370049927437893, | |
| "grad_norm": 0.6332319974899292, | |
| "learning_rate": 8.338319902955062e-05, | |
| "loss": 12.0005, | |
| "num_input_tokens_seen": 5989990400, | |
| "step": 22850 | |
| }, | |
| { | |
| "epoch": 0.15403682421808654, | |
| "grad_norm": 0.6333373188972473, | |
| "learning_rate": 8.227738661451541e-05, | |
| "loss": 12.0081, | |
| "num_input_tokens_seen": 6003097600, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 0.15437314916179415, | |
| "grad_norm": 0.6376117467880249, | |
| "learning_rate": 8.117778929200977e-05, | |
| "loss": 11.9789, | |
| "num_input_tokens_seen": 6016204800, | |
| "step": 22950 | |
| }, | |
| { | |
| "epoch": 0.15470947410550176, | |
| "grad_norm": 0.6416700482368469, | |
| "learning_rate": 8.008443845105216e-05, | |
| "loss": 11.9845, | |
| "num_input_tokens_seen": 6029312000, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.15470947410550176, | |
| "eval_loss": 2.9053738117218018, | |
| "eval_runtime": 142.1312, | |
| "eval_samples_per_second": 35.179, | |
| "eval_steps_per_second": 8.795, | |
| "num_input_tokens_seen": 6029312000, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.15504579904920937, | |
| "grad_norm": 0.6468757390975952, | |
| "learning_rate": 7.899736530234923e-05, | |
| "loss": 11.991, | |
| "num_input_tokens_seen": 6042419200, | |
| "step": 23050 | |
| }, | |
| { | |
| "epoch": 0.155382123992917, | |
| "grad_norm": 0.6658541560173035, | |
| "learning_rate": 7.791660087740537e-05, | |
| "loss": 11.9583, | |
| "num_input_tokens_seen": 6055526400, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 0.15571844893662462, | |
| "grad_norm": 0.6665578484535217, | |
| "learning_rate": 7.68421760276364e-05, | |
| "loss": 12.0004, | |
| "num_input_tokens_seen": 6068633600, | |
| "step": 23150 | |
| }, | |
| { | |
| "epoch": 0.15605477388033223, | |
| "grad_norm": 0.6088104844093323, | |
| "learning_rate": 7.577412142348944e-05, | |
| "loss": 11.9758, | |
| "num_input_tokens_seen": 6081740800, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 0.15639109882403984, | |
| "grad_norm": 0.6299030184745789, | |
| "learning_rate": 7.47124675535666e-05, | |
| "loss": 12.036, | |
| "num_input_tokens_seen": 6094848000, | |
| "step": 23250 | |
| }, | |
| { | |
| "epoch": 0.15672742376774745, | |
| "grad_norm": 0.642490565776825, | |
| "learning_rate": 7.365724472375568e-05, | |
| "loss": 11.9951, | |
| "num_input_tokens_seen": 6107955200, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 0.15706374871145506, | |
| "grad_norm": 0.6375728845596313, | |
| "learning_rate": 7.260848305636405e-05, | |
| "loss": 11.9859, | |
| "num_input_tokens_seen": 6121062400, | |
| "step": 23350 | |
| }, | |
| { | |
| "epoch": 0.15740007365516268, | |
| "grad_norm": 0.6122708320617676, | |
| "learning_rate": 7.156621248925967e-05, | |
| "loss": 11.9532, | |
| "num_input_tokens_seen": 6134169600, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 0.1577363985988703, | |
| "grad_norm": 0.6512198448181152, | |
| "learning_rate": 7.05304627750157e-05, | |
| "loss": 11.9962, | |
| "num_input_tokens_seen": 6147276800, | |
| "step": 23450 | |
| }, | |
| { | |
| "epoch": 0.1580727235425779, | |
| "grad_norm": 0.6488016247749329, | |
| "learning_rate": 6.950126348006171e-05, | |
| "loss": 11.9285, | |
| "num_input_tokens_seen": 6160384000, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.1580727235425779, | |
| "eval_loss": 2.9018726348876953, | |
| "eval_runtime": 143.4685, | |
| "eval_samples_per_second": 34.851, | |
| "eval_steps_per_second": 8.713, | |
| "num_input_tokens_seen": 6160384000, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.1584090484862855, | |
| "grad_norm": 0.6251162886619568, | |
| "learning_rate": 6.847864398383946e-05, | |
| "loss": 11.9805, | |
| "num_input_tokens_seen": 6173491200, | |
| "step": 23550 | |
| }, | |
| { | |
| "epoch": 0.15874537342999312, | |
| "grad_norm": 0.6338608264923096, | |
| "learning_rate": 6.746263347796449e-05, | |
| "loss": 11.9775, | |
| "num_input_tokens_seen": 6186598400, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 0.15908169837370073, | |
| "grad_norm": 0.6400789618492126, | |
| "learning_rate": 6.645326096539229e-05, | |
| "loss": 11.9472, | |
| "num_input_tokens_seen": 6199705600, | |
| "step": 23650 | |
| }, | |
| { | |
| "epoch": 0.15941802331740834, | |
| "grad_norm": 0.6252830624580383, | |
| "learning_rate": 6.545055525959105e-05, | |
| "loss": 11.9752, | |
| "num_input_tokens_seen": 6212812800, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 0.15975434826111595, | |
| "grad_norm": 0.6790284514427185, | |
| "learning_rate": 6.445454498371857e-05, | |
| "loss": 11.957, | |
| "num_input_tokens_seen": 6225920000, | |
| "step": 23750 | |
| }, | |
| { | |
| "epoch": 0.16009067320482356, | |
| "grad_norm": 0.621303379535675, | |
| "learning_rate": 6.346525856980567e-05, | |
| "loss": 11.9433, | |
| "num_input_tokens_seen": 6239027200, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 0.16042699814853117, | |
| "grad_norm": 0.6379457712173462, | |
| "learning_rate": 6.248272425794411e-05, | |
| "loss": 11.9516, | |
| "num_input_tokens_seen": 6252134400, | |
| "step": 23850 | |
| }, | |
| { | |
| "epoch": 0.16076332309223879, | |
| "grad_norm": 0.6223682761192322, | |
| "learning_rate": 6.150697009548073e-05, | |
| "loss": 11.9856, | |
| "num_input_tokens_seen": 6265241600, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 0.1610996480359464, | |
| "grad_norm": 0.6040588021278381, | |
| "learning_rate": 6.0538023936216814e-05, | |
| "loss": 11.9921, | |
| "num_input_tokens_seen": 6278348800, | |
| "step": 23950 | |
| }, | |
| { | |
| "epoch": 0.16143597297965404, | |
| "grad_norm": 0.6390047669410706, | |
| "learning_rate": 5.957591343961304e-05, | |
| "loss": 11.9322, | |
| "num_input_tokens_seen": 6291456000, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.16143597297965404, | |
| "eval_loss": 2.8987817764282227, | |
| "eval_runtime": 142.5945, | |
| "eval_samples_per_second": 35.064, | |
| "eval_steps_per_second": 8.766, | |
| "num_input_tokens_seen": 6291456000, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.16177229792336165, | |
| "grad_norm": 0.6663207411766052, | |
| "learning_rate": 5.862066606999949e-05, | |
| "loss": 11.9835, | |
| "num_input_tokens_seen": 6304563200, | |
| "step": 24050 | |
| }, | |
| { | |
| "epoch": 0.16210862286706926, | |
| "grad_norm": 0.6109934449195862, | |
| "learning_rate": 5.7672309095792316e-05, | |
| "loss": 11.933, | |
| "num_input_tokens_seen": 6317670400, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 0.16244494781077687, | |
| "grad_norm": 0.6243853569030762, | |
| "learning_rate": 5.6730869588714744e-05, | |
| "loss": 12.0097, | |
| "num_input_tokens_seen": 6330777600, | |
| "step": 24150 | |
| }, | |
| { | |
| "epoch": 0.16278127275448448, | |
| "grad_norm": 0.6165538430213928, | |
| "learning_rate": 5.579637442302454e-05, | |
| "loss": 11.9705, | |
| "num_input_tokens_seen": 6343884800, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 0.1631175976981921, | |
| "grad_norm": 0.5966577529907227, | |
| "learning_rate": 5.4868850274747045e-05, | |
| "loss": 11.9362, | |
| "num_input_tokens_seen": 6356992000, | |
| "step": 24250 | |
| }, | |
| { | |
| "epoch": 0.1634539226418997, | |
| "grad_norm": 0.6405600309371948, | |
| "learning_rate": 5.39483236209132e-05, | |
| "loss": 11.987, | |
| "num_input_tokens_seen": 6370099200, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 0.1637902475856073, | |
| "grad_norm": 0.6266763210296631, | |
| "learning_rate": 5.303482073880436e-05, | |
| "loss": 11.9779, | |
| "num_input_tokens_seen": 6383206400, | |
| "step": 24350 | |
| }, | |
| { | |
| "epoch": 0.16412657252931492, | |
| "grad_norm": 0.6331851482391357, | |
| "learning_rate": 5.2128367705201594e-05, | |
| "loss": 11.921, | |
| "num_input_tokens_seen": 6396313600, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 0.16446289747302253, | |
| "grad_norm": 0.6594439744949341, | |
| "learning_rate": 5.122899039564157e-05, | |
| "loss": 11.9332, | |
| "num_input_tokens_seen": 6409420800, | |
| "step": 24450 | |
| }, | |
| { | |
| "epoch": 0.16479922241673015, | |
| "grad_norm": 0.6269896626472473, | |
| "learning_rate": 5.033671448367788e-05, | |
| "loss": 11.9627, | |
| "num_input_tokens_seen": 6422528000, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.16479922241673015, | |
| "eval_loss": 2.896472454071045, | |
| "eval_runtime": 143.2784, | |
| "eval_samples_per_second": 34.897, | |
| "eval_steps_per_second": 8.724, | |
| "num_input_tokens_seen": 6422528000, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.16513554736043776, | |
| "grad_norm": 0.6272408962249756, | |
| "learning_rate": 4.945156544014846e-05, | |
| "loss": 11.9879, | |
| "num_input_tokens_seen": 6435635200, | |
| "step": 24550 | |
| }, | |
| { | |
| "epoch": 0.16547187230414537, | |
| "grad_norm": 0.6145939826965332, | |
| "learning_rate": 4.8573568532447815e-05, | |
| "loss": 11.964, | |
| "num_input_tokens_seen": 6448742400, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 0.16580819724785298, | |
| "grad_norm": 0.6379438638687134, | |
| "learning_rate": 4.770274882380648e-05, | |
| "loss": 11.9384, | |
| "num_input_tokens_seen": 6461849600, | |
| "step": 24650 | |
| }, | |
| { | |
| "epoch": 0.1661445221915606, | |
| "grad_norm": 0.6524396538734436, | |
| "learning_rate": 4.6839131172574996e-05, | |
| "loss": 11.9477, | |
| "num_input_tokens_seen": 6474956800, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 0.1664808471352682, | |
| "grad_norm": 0.6290236711502075, | |
| "learning_rate": 4.598274023151476e-05, | |
| "loss": 11.9441, | |
| "num_input_tokens_seen": 6488064000, | |
| "step": 24750 | |
| }, | |
| { | |
| "epoch": 0.1668171720789758, | |
| "grad_norm": 0.6329859495162964, | |
| "learning_rate": 4.513360044709382e-05, | |
| "loss": 11.9686, | |
| "num_input_tokens_seen": 6501171200, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 0.16715349702268342, | |
| "grad_norm": 0.6199634671211243, | |
| "learning_rate": 4.429173605878951e-05, | |
| "loss": 11.907, | |
| "num_input_tokens_seen": 6514278400, | |
| "step": 24850 | |
| }, | |
| { | |
| "epoch": 0.16748982196639106, | |
| "grad_norm": 0.6326203346252441, | |
| "learning_rate": 4.3457171098396174e-05, | |
| "loss": 11.9358, | |
| "num_input_tokens_seen": 6527385600, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 0.16782614691009867, | |
| "grad_norm": 0.647875189781189, | |
| "learning_rate": 4.2629929389339246e-05, | |
| "loss": 11.9304, | |
| "num_input_tokens_seen": 6540492800, | |
| "step": 24950 | |
| }, | |
| { | |
| "epoch": 0.16816247185380628, | |
| "grad_norm": 0.6240447759628296, | |
| "learning_rate": 4.181003454599512e-05, | |
| "loss": 11.9144, | |
| "num_input_tokens_seen": 6553600000, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.16816247185380628, | |
| "eval_loss": 2.8938522338867188, | |
| "eval_runtime": 143.6338, | |
| "eval_samples_per_second": 34.811, | |
| "eval_steps_per_second": 8.703, | |
| "num_input_tokens_seen": 6553600000, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.1684987967975139, | |
| "grad_norm": 0.6343597173690796, | |
| "learning_rate": 4.099750997301747e-05, | |
| "loss": 11.9949, | |
| "num_input_tokens_seen": 6566707200, | |
| "step": 25050 | |
| }, | |
| { | |
| "epoch": 0.1688351217412215, | |
| "grad_norm": 0.626124918460846, | |
| "learning_rate": 4.019237886466838e-05, | |
| "loss": 11.9272, | |
| "num_input_tokens_seen": 6579814400, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 0.16917144668492912, | |
| "grad_norm": 0.6266665458679199, | |
| "learning_rate": 3.939466420415709e-05, | |
| "loss": 11.935, | |
| "num_input_tokens_seen": 6592921600, | |
| "step": 25150 | |
| }, | |
| { | |
| "epoch": 0.16950777162863673, | |
| "grad_norm": 0.6637131571769714, | |
| "learning_rate": 3.8604388762983175e-05, | |
| "loss": 11.9444, | |
| "num_input_tokens_seen": 6606028800, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 0.16984409657234434, | |
| "grad_norm": 0.6241376399993896, | |
| "learning_rate": 3.782157510028706e-05, | |
| "loss": 11.9235, | |
| "num_input_tokens_seen": 6619136000, | |
| "step": 25250 | |
| }, | |
| { | |
| "epoch": 0.17018042151605195, | |
| "grad_norm": 0.617912232875824, | |
| "learning_rate": 3.704624556220566e-05, | |
| "loss": 11.9165, | |
| "num_input_tokens_seen": 6632243200, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 0.17051674645975956, | |
| "grad_norm": 0.6167590022087097, | |
| "learning_rate": 3.627842228123483e-05, | |
| "loss": 11.9636, | |
| "num_input_tokens_seen": 6645350400, | |
| "step": 25350 | |
| }, | |
| { | |
| "epoch": 0.17085307140346717, | |
| "grad_norm": 0.6313674449920654, | |
| "learning_rate": 3.551812717559729e-05, | |
| "loss": 11.9304, | |
| "num_input_tokens_seen": 6658457600, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 0.17118939634717478, | |
| "grad_norm": 0.6143530607223511, | |
| "learning_rate": 3.47653819486171e-05, | |
| "loss": 11.9495, | |
| "num_input_tokens_seen": 6671564800, | |
| "step": 25450 | |
| }, | |
| { | |
| "epoch": 0.1715257212908824, | |
| "grad_norm": 0.6127185821533203, | |
| "learning_rate": 3.402020808809996e-05, | |
| "loss": 11.926, | |
| "num_input_tokens_seen": 6684672000, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.1715257212908824, | |
| "eval_loss": 2.8920793533325195, | |
| "eval_runtime": 142.7714, | |
| "eval_samples_per_second": 35.021, | |
| "eval_steps_per_second": 8.755, | |
| "num_input_tokens_seen": 6684672000, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.17186204623459, | |
| "grad_norm": 0.621300995349884, | |
| "learning_rate": 3.328262686572024e-05, | |
| "loss": 11.9852, | |
| "num_input_tokens_seen": 6697779200, | |
| "step": 25550 | |
| }, | |
| { | |
| "epoch": 0.17219837117829762, | |
| "grad_norm": 0.6242550015449524, | |
| "learning_rate": 3.2552659336413154e-05, | |
| "loss": 11.9132, | |
| "num_input_tokens_seen": 6710886400, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 0.17253469612200523, | |
| "grad_norm": 0.6490415930747986, | |
| "learning_rate": 3.1830326337774124e-05, | |
| "loss": 11.9529, | |
| "num_input_tokens_seen": 6723993600, | |
| "step": 25650 | |
| }, | |
| { | |
| "epoch": 0.17287102106571284, | |
| "grad_norm": 0.5997505187988281, | |
| "learning_rate": 3.111564848946403e-05, | |
| "loss": 11.948, | |
| "num_input_tokens_seen": 6737100800, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 0.17320734600942045, | |
| "grad_norm": 0.6490405797958374, | |
| "learning_rate": 3.040864619262011e-05, | |
| "loss": 11.9353, | |
| "num_input_tokens_seen": 6750208000, | |
| "step": 25750 | |
| }, | |
| { | |
| "epoch": 0.1735436709531281, | |
| "grad_norm": 0.6102951169013977, | |
| "learning_rate": 2.9709339629274285e-05, | |
| "loss": 11.97, | |
| "num_input_tokens_seen": 6763315200, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 0.1738799958968357, | |
| "grad_norm": 0.6121110916137695, | |
| "learning_rate": 2.9017748761776394e-05, | |
| "loss": 11.9342, | |
| "num_input_tokens_seen": 6776422400, | |
| "step": 25850 | |
| }, | |
| { | |
| "epoch": 0.1742163208405433, | |
| "grad_norm": 0.6192799806594849, | |
| "learning_rate": 2.8333893332224754e-05, | |
| "loss": 11.928, | |
| "num_input_tokens_seen": 6789529600, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 0.17455264578425092, | |
| "grad_norm": 0.6458452939987183, | |
| "learning_rate": 2.7657792861902393e-05, | |
| "loss": 11.9213, | |
| "num_input_tokens_seen": 6802636800, | |
| "step": 25950 | |
| }, | |
| { | |
| "epoch": 0.17488897072795853, | |
| "grad_norm": 0.6549943089485168, | |
| "learning_rate": 2.6989466650720048e-05, | |
| "loss": 11.9298, | |
| "num_input_tokens_seen": 6815744000, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.17488897072795853, | |
| "eval_loss": 2.890101671218872, | |
| "eval_runtime": 142.6241, | |
| "eval_samples_per_second": 35.057, | |
| "eval_steps_per_second": 8.764, | |
| "num_input_tokens_seen": 6815744000, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.17522529567166614, | |
| "grad_norm": 0.6231434941291809, | |
| "learning_rate": 2.6328933776664907e-05, | |
| "loss": 11.8924, | |
| "num_input_tokens_seen": 6828851200, | |
| "step": 26050 | |
| }, | |
| { | |
| "epoch": 0.17556162061537375, | |
| "grad_norm": 0.6445599794387817, | |
| "learning_rate": 2.567621309525628e-05, | |
| "loss": 11.9639, | |
| "num_input_tokens_seen": 6841958400, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 0.17589794555908136, | |
| "grad_norm": 0.6182544827461243, | |
| "learning_rate": 2.503132323900714e-05, | |
| "loss": 11.8955, | |
| "num_input_tokens_seen": 6855065600, | |
| "step": 26150 | |
| }, | |
| { | |
| "epoch": 0.17623427050278898, | |
| "grad_norm": 0.6308871507644653, | |
| "learning_rate": 2.439428261689249e-05, | |
| "loss": 11.898, | |
| "num_input_tokens_seen": 6868172800, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 0.1765705954464966, | |
| "grad_norm": 0.6257124543190002, | |
| "learning_rate": 2.376510941382351e-05, | |
| "loss": 11.9309, | |
| "num_input_tokens_seen": 6881280000, | |
| "step": 26250 | |
| }, | |
| { | |
| "epoch": 0.1769069203902042, | |
| "grad_norm": 0.6235978603363037, | |
| "learning_rate": 2.3143821590128896e-05, | |
| "loss": 11.9587, | |
| "num_input_tokens_seen": 6894387200, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 0.1772432453339118, | |
| "grad_norm": 0.6002153158187866, | |
| "learning_rate": 2.2530436881041725e-05, | |
| "loss": 11.9336, | |
| "num_input_tokens_seen": 6907494400, | |
| "step": 26350 | |
| }, | |
| { | |
| "epoch": 0.17757957027761942, | |
| "grad_norm": 0.6364301443099976, | |
| "learning_rate": 2.1924972796193506e-05, | |
| "loss": 11.9054, | |
| "num_input_tokens_seen": 6920601600, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 0.17791589522132703, | |
| "grad_norm": 0.6437053680419922, | |
| "learning_rate": 2.132744661911412e-05, | |
| "loss": 11.9355, | |
| "num_input_tokens_seen": 6933708800, | |
| "step": 26450 | |
| }, | |
| { | |
| "epoch": 0.17825222016503464, | |
| "grad_norm": 0.6307169795036316, | |
| "learning_rate": 2.073787540673876e-05, | |
| "loss": 11.9117, | |
| "num_input_tokens_seen": 6946816000, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.17825222016503464, | |
| "eval_loss": 2.888777732849121, | |
| "eval_runtime": 142.3367, | |
| "eval_samples_per_second": 35.128, | |
| "eval_steps_per_second": 8.782, | |
| "num_input_tokens_seen": 6946816000, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.17858854510874225, | |
| "grad_norm": 0.6065594553947449, | |
| "learning_rate": 2.0156275988920568e-05, | |
| "loss": 11.9054, | |
| "num_input_tokens_seen": 6959923200, | |
| "step": 26550 | |
| }, | |
| { | |
| "epoch": 0.17892487005244986, | |
| "grad_norm": 0.6257479190826416, | |
| "learning_rate": 1.958266496795069e-05, | |
| "loss": 11.8735, | |
| "num_input_tokens_seen": 6973030400, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 0.17926119499615747, | |
| "grad_norm": 0.6354475617408752, | |
| "learning_rate": 1.9017058718084012e-05, | |
| "loss": 11.9371, | |
| "num_input_tokens_seen": 6986137600, | |
| "step": 26650 | |
| }, | |
| { | |
| "epoch": 0.1795975199398651, | |
| "grad_norm": 0.6183739900588989, | |
| "learning_rate": 1.8459473385071865e-05, | |
| "loss": 11.9123, | |
| "num_input_tokens_seen": 6999244800, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 0.17993384488357272, | |
| "grad_norm": 0.6221346259117126, | |
| "learning_rate": 1.7909924885701145e-05, | |
| "loss": 11.9004, | |
| "num_input_tokens_seen": 7012352000, | |
| "step": 26750 | |
| }, | |
| { | |
| "epoch": 0.18027016982728034, | |
| "grad_norm": 0.607341468334198, | |
| "learning_rate": 1.7368428907339983e-05, | |
| "loss": 11.9286, | |
| "num_input_tokens_seen": 7025459200, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 0.18060649477098795, | |
| "grad_norm": 0.6302104592323303, | |
| "learning_rate": 1.6835000907489728e-05, | |
| "loss": 11.9551, | |
| "num_input_tokens_seen": 7038566400, | |
| "step": 26850 | |
| }, | |
| { | |
| "epoch": 0.18094281971469556, | |
| "grad_norm": 0.6029033064842224, | |
| "learning_rate": 1.6309656113344017e-05, | |
| "loss": 11.8979, | |
| "num_input_tokens_seen": 7051673600, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 0.18127914465840317, | |
| "grad_norm": 0.6170194149017334, | |
| "learning_rate": 1.5792409521353732e-05, | |
| "loss": 11.9503, | |
| "num_input_tokens_seen": 7064780800, | |
| "step": 26950 | |
| }, | |
| { | |
| "epoch": 0.18161546960211078, | |
| "grad_norm": 0.6190406084060669, | |
| "learning_rate": 1.5283275896799407e-05, | |
| "loss": 11.945, | |
| "num_input_tokens_seen": 7077888000, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.18161546960211078, | |
| "eval_loss": 2.88728404045105, | |
| "eval_runtime": 143.9093, | |
| "eval_samples_per_second": 34.744, | |
| "eval_steps_per_second": 8.686, | |
| "num_input_tokens_seen": 7077888000, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.1819517945458184, | |
| "grad_norm": 0.6403325796127319, | |
| "learning_rate": 1.478226977336916e-05, | |
| "loss": 11.8936, | |
| "num_input_tokens_seen": 7090995200, | |
| "step": 27050 | |
| }, | |
| { | |
| "epoch": 0.182288119489526, | |
| "grad_norm": 0.6248791813850403, | |
| "learning_rate": 1.428940545274433e-05, | |
| "loss": 11.9114, | |
| "num_input_tokens_seen": 7104102400, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 0.1826244444332336, | |
| "grad_norm": 0.6114192605018616, | |
| "learning_rate": 1.3804697004190869e-05, | |
| "loss": 11.9281, | |
| "num_input_tokens_seen": 7117209600, | |
| "step": 27150 | |
| }, | |
| { | |
| "epoch": 0.18296076937694122, | |
| "grad_norm": 0.6320353746414185, | |
| "learning_rate": 1.3328158264157762e-05, | |
| "loss": 11.9141, | |
| "num_input_tokens_seen": 7130316800, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 0.18329709432064883, | |
| "grad_norm": 0.6097228527069092, | |
| "learning_rate": 1.2859802835882416e-05, | |
| "loss": 11.8966, | |
| "num_input_tokens_seen": 7143424000, | |
| "step": 27250 | |
| }, | |
| { | |
| "epoch": 0.18363341926435645, | |
| "grad_norm": 0.6205602288246155, | |
| "learning_rate": 1.2399644089001825e-05, | |
| "loss": 11.9154, | |
| "num_input_tokens_seen": 7156531200, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 0.18396974420806406, | |
| "grad_norm": 0.6151401996612549, | |
| "learning_rate": 1.1947695159171256e-05, | |
| "loss": 11.8856, | |
| "num_input_tokens_seen": 7169638400, | |
| "step": 27350 | |
| }, | |
| { | |
| "epoch": 0.18430606915177167, | |
| "grad_norm": 0.6325812935829163, | |
| "learning_rate": 1.1503968947689135e-05, | |
| "loss": 11.9602, | |
| "num_input_tokens_seen": 7182745600, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 0.18464239409547928, | |
| "grad_norm": 0.6651480197906494, | |
| "learning_rate": 1.106847812112892e-05, | |
| "loss": 11.8962, | |
| "num_input_tokens_seen": 7195852800, | |
| "step": 27450 | |
| }, | |
| { | |
| "epoch": 0.1849787190391869, | |
| "grad_norm": 0.6203281283378601, | |
| "learning_rate": 1.0641235110977286e-05, | |
| "loss": 11.9267, | |
| "num_input_tokens_seen": 7208960000, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.1849787190391869, | |
| "eval_loss": 2.8867011070251465, | |
| "eval_runtime": 143.8431, | |
| "eval_samples_per_second": 34.76, | |
| "eval_steps_per_second": 8.69, | |
| "num_input_tokens_seen": 7208960000, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.1853150439828945, | |
| "grad_norm": 0.628180205821991, | |
| "learning_rate": 1.022225211327954e-05, | |
| "loss": 11.9684, | |
| "num_input_tokens_seen": 7222067200, | |
| "step": 27550 | |
| }, | |
| { | |
| "epoch": 0.18565136892660214, | |
| "grad_norm": 0.6240800023078918, | |
| "learning_rate": 9.811541088291163e-06, | |
| "loss": 11.9017, | |
| "num_input_tokens_seen": 7235174400, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 0.18598769387030975, | |
| "grad_norm": 0.6192197799682617, | |
| "learning_rate": 9.409113760136766e-06, | |
| "loss": 11.9137, | |
| "num_input_tokens_seen": 7248281600, | |
| "step": 27650 | |
| }, | |
| { | |
| "epoch": 0.18632401881401736, | |
| "grad_norm": 0.6189801096916199, | |
| "learning_rate": 9.014981616474937e-06, | |
| "loss": 11.9493, | |
| "num_input_tokens_seen": 7261388800, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 0.18666034375772497, | |
| "grad_norm": 0.6035293340682983, | |
| "learning_rate": 8.629155908170881e-06, | |
| "loss": 11.9083, | |
| "num_input_tokens_seen": 7274496000, | |
| "step": 27750 | |
| }, | |
| { | |
| "epoch": 0.18699666870143258, | |
| "grad_norm": 0.6316511034965515, | |
| "learning_rate": 8.25164764897468e-06, | |
| "loss": 11.9187, | |
| "num_input_tokens_seen": 7287603200, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 0.1873329936451402, | |
| "grad_norm": 0.6229190826416016, | |
| "learning_rate": 7.882467615207334e-06, | |
| "loss": 11.8842, | |
| "num_input_tokens_seen": 7300710400, | |
| "step": 27850 | |
| }, | |
| { | |
| "epoch": 0.1876693185888478, | |
| "grad_norm": 0.6222130656242371, | |
| "learning_rate": 7.521626345452914e-06, | |
| "loss": 11.9228, | |
| "num_input_tokens_seen": 7313817600, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 0.18800564353255542, | |
| "grad_norm": 0.6076390743255615, | |
| "learning_rate": 7.169134140257871e-06, | |
| "loss": 11.9038, | |
| "num_input_tokens_seen": 7326924800, | |
| "step": 27950 | |
| }, | |
| { | |
| "epoch": 0.18834196847626303, | |
| "grad_norm": 0.6207023859024048, | |
| "learning_rate": 6.825001061836799e-06, | |
| "loss": 11.9013, | |
| "num_input_tokens_seen": 7340032000, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.18834196847626303, | |
| "eval_loss": 2.885740280151367, | |
| "eval_runtime": 143.4494, | |
| "eval_samples_per_second": 34.855, | |
| "eval_steps_per_second": 8.714, | |
| "num_input_tokens_seen": 7340032000, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.18867829341997064, | |
| "grad_norm": 0.6160932779312134, | |
| "learning_rate": 6.4892369337854025e-06, | |
| "loss": 11.9279, | |
| "num_input_tokens_seen": 7353139200, | |
| "step": 28050 | |
| }, | |
| { | |
| "epoch": 0.18901461836367825, | |
| "grad_norm": 0.6192066669464111, | |
| "learning_rate": 6.161851340799984e-06, | |
| "loss": 11.8922, | |
| "num_input_tokens_seen": 7366246400, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 0.18935094330738586, | |
| "grad_norm": 0.6136648654937744, | |
| "learning_rate": 5.842853628403799e-06, | |
| "loss": 11.906, | |
| "num_input_tokens_seen": 7379353600, | |
| "step": 28150 | |
| }, | |
| { | |
| "epoch": 0.18968726825109347, | |
| "grad_norm": 0.621473491191864, | |
| "learning_rate": 5.532252902680367e-06, | |
| "loss": 11.8603, | |
| "num_input_tokens_seen": 7392460800, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 0.19002359319480108, | |
| "grad_norm": 0.6140876412391663, | |
| "learning_rate": 5.2300580300135175e-06, | |
| "loss": 11.8953, | |
| "num_input_tokens_seen": 7405568000, | |
| "step": 28250 | |
| }, | |
| { | |
| "epoch": 0.1903599181385087, | |
| "grad_norm": 0.6015214323997498, | |
| "learning_rate": 4.9362776368341846e-06, | |
| "loss": 11.8874, | |
| "num_input_tokens_seen": 7418675200, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 0.1906962430822163, | |
| "grad_norm": 0.6086856126785278, | |
| "learning_rate": 4.650920109374279e-06, | |
| "loss": 11.9015, | |
| "num_input_tokens_seen": 7431782400, | |
| "step": 28350 | |
| }, | |
| { | |
| "epoch": 0.19103256802592392, | |
| "grad_norm": 0.6232919692993164, | |
| "learning_rate": 4.373993593427238e-06, | |
| "loss": 11.9252, | |
| "num_input_tokens_seen": 7444889600, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 0.19136889296963153, | |
| "grad_norm": 0.6096498966217041, | |
| "learning_rate": 4.105505994115521e-06, | |
| "loss": 11.9018, | |
| "num_input_tokens_seen": 7457996800, | |
| "step": 28450 | |
| }, | |
| { | |
| "epoch": 0.19170521791333917, | |
| "grad_norm": 0.6317954659461975, | |
| "learning_rate": 3.845464975664947e-06, | |
| "loss": 11.9102, | |
| "num_input_tokens_seen": 7471104000, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.19170521791333917, | |
| "eval_loss": 2.8853116035461426, | |
| "eval_runtime": 143.468, | |
| "eval_samples_per_second": 34.851, | |
| "eval_steps_per_second": 8.713, | |
| "num_input_tokens_seen": 7471104000, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.19204154285704678, | |
| "grad_norm": 0.6087967753410339, | |
| "learning_rate": 3.5938779611859093e-06, | |
| "loss": 11.9431, | |
| "num_input_tokens_seen": 7484211200, | |
| "step": 28550 | |
| }, | |
| { | |
| "epoch": 0.1923778678007544, | |
| "grad_norm": 0.614473283290863, | |
| "learning_rate": 3.350752132461443e-06, | |
| "loss": 11.9548, | |
| "num_input_tokens_seen": 7497318400, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 0.192714192744462, | |
| "grad_norm": 0.6257823705673218, | |
| "learning_rate": 3.116094429742222e-06, | |
| "loss": 11.9179, | |
| "num_input_tokens_seen": 7510425600, | |
| "step": 28650 | |
| }, | |
| { | |
| "epoch": 0.1930505176881696, | |
| "grad_norm": 0.6351081728935242, | |
| "learning_rate": 2.889911551548585e-06, | |
| "loss": 11.9183, | |
| "num_input_tokens_seen": 7523532800, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 0.19338684263187722, | |
| "grad_norm": 0.6371856331825256, | |
| "learning_rate": 2.672209954479021e-06, | |
| "loss": 11.9169, | |
| "num_input_tokens_seen": 7536640000, | |
| "step": 28750 | |
| }, | |
| { | |
| "epoch": 0.19372316757558483, | |
| "grad_norm": 0.622117280960083, | |
| "learning_rate": 2.462995853026184e-06, | |
| "loss": 11.9404, | |
| "num_input_tokens_seen": 7549747200, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 0.19405949251929244, | |
| "grad_norm": 0.6010422110557556, | |
| "learning_rate": 2.2622752193992675e-06, | |
| "loss": 11.9441, | |
| "num_input_tokens_seen": 7562854400, | |
| "step": 28850 | |
| }, | |
| { | |
| "epoch": 0.19439581746300005, | |
| "grad_norm": 0.6092264652252197, | |
| "learning_rate": 2.0700537833536422e-06, | |
| "loss": 11.893, | |
| "num_input_tokens_seen": 7575961600, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 0.19473214240670766, | |
| "grad_norm": 0.6216610670089722, | |
| "learning_rate": 1.8863370320272187e-06, | |
| "loss": 11.9201, | |
| "num_input_tokens_seen": 7589068800, | |
| "step": 28950 | |
| }, | |
| { | |
| "epoch": 0.19506846735041528, | |
| "grad_norm": 0.615051805973053, | |
| "learning_rate": 1.7111302097839396e-06, | |
| "loss": 11.9402, | |
| "num_input_tokens_seen": 7602176000, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.19506846735041528, | |
| "eval_loss": 2.885068655014038, | |
| "eval_runtime": 142.9832, | |
| "eval_samples_per_second": 34.969, | |
| "eval_steps_per_second": 8.742, | |
| "num_input_tokens_seen": 7602176000, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.1954047922941229, | |
| "grad_norm": 0.6069262027740479, | |
| "learning_rate": 1.5444383180638342e-06, | |
| "loss": 11.9314, | |
| "num_input_tokens_seen": 7615283200, | |
| "step": 29050 | |
| }, | |
| { | |
| "epoch": 0.1957411172378305, | |
| "grad_norm": 0.628108561038971, | |
| "learning_rate": 1.3862661152405309e-06, | |
| "loss": 11.9151, | |
| "num_input_tokens_seen": 7628390400, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 0.1960774421815381, | |
| "grad_norm": 0.6232333779335022, | |
| "learning_rate": 1.236618116485233e-06, | |
| "loss": 11.8887, | |
| "num_input_tokens_seen": 7641497600, | |
| "step": 29150 | |
| }, | |
| { | |
| "epoch": 0.19641376712524572, | |
| "grad_norm": 0.6372972726821899, | |
| "learning_rate": 1.0954985936379223e-06, | |
| "loss": 11.8873, | |
| "num_input_tokens_seen": 7654604800, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 0.19675009206895333, | |
| "grad_norm": 0.5991822481155396, | |
| "learning_rate": 9.6291157508529e-07, | |
| "loss": 11.9405, | |
| "num_input_tokens_seen": 7667712000, | |
| "step": 29250 | |
| }, | |
| { | |
| "epoch": 0.19708641701266094, | |
| "grad_norm": 0.6108511686325073, | |
| "learning_rate": 8.388608456459612e-07, | |
| "loss": 11.9085, | |
| "num_input_tokens_seen": 7680819200, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 0.19742274195636855, | |
| "grad_norm": 0.6104913949966431, | |
| "learning_rate": 7.23349946462215e-07, | |
| "loss": 11.8859, | |
| "num_input_tokens_seen": 7693926400, | |
| "step": 29350 | |
| }, | |
| { | |
| "epoch": 0.1977590669000762, | |
| "grad_norm": 0.6084222197532654, | |
| "learning_rate": 6.163821748990994e-07, | |
| "loss": 11.9059, | |
| "num_input_tokens_seen": 7707033600, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 0.1980953918437838, | |
| "grad_norm": 0.633105993270874, | |
| "learning_rate": 5.179605844501388e-07, | |
| "loss": 11.9174, | |
| "num_input_tokens_seen": 7720140800, | |
| "step": 29450 | |
| }, | |
| { | |
| "epoch": 0.1984317167874914, | |
| "grad_norm": 0.6088514924049377, | |
| "learning_rate": 4.280879846503049e-07, | |
| "loss": 11.9125, | |
| "num_input_tokens_seen": 7733248000, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.1984317167874914, | |
| "eval_loss": 2.8849411010742188, | |
| "eval_runtime": 143.8146, | |
| "eval_samples_per_second": 34.767, | |
| "eval_steps_per_second": 8.692, | |
| "num_input_tokens_seen": 7733248000, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.19876804173119902, | |
| "grad_norm": 0.6054402589797974, | |
| "learning_rate": 3.467669409957463e-07, | |
| "loss": 11.9468, | |
| "num_input_tokens_seen": 7746355200, | |
| "step": 29550 | |
| }, | |
| { | |
| "epoch": 0.19910436667490664, | |
| "grad_norm": 0.6133595705032349, | |
| "learning_rate": 2.7399977487051473e-07, | |
| "loss": 11.9368, | |
| "num_input_tokens_seen": 7759462400, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 0.19944069161861425, | |
| "grad_norm": 0.6098650693893433, | |
| "learning_rate": 2.097885634804175e-07, | |
| "loss": 11.8971, | |
| "num_input_tokens_seen": 7772569600, | |
| "step": 29650 | |
| }, | |
| { | |
| "epoch": 0.19977701656232186, | |
| "grad_norm": 0.6231054663658142, | |
| "learning_rate": 1.541351397936319e-07, | |
| "loss": 11.9546, | |
| "num_input_tokens_seen": 7785676800, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 0.20011334150602947, | |
| "grad_norm": 0.6323234438896179, | |
| "learning_rate": 1.0704109248838022e-07, | |
| "loss": 11.8848, | |
| "num_input_tokens_seen": 7798784000, | |
| "step": 29750 | |
| }, | |
| { | |
| "epoch": 0.20044966644973708, | |
| "grad_norm": 0.6294256448745728, | |
| "learning_rate": 6.850776590763274e-08, | |
| "loss": 11.9027, | |
| "num_input_tokens_seen": 7811891200, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 0.2007859913934447, | |
| "grad_norm": 0.6184135675430298, | |
| "learning_rate": 3.853626002063848e-08, | |
| "loss": 11.9454, | |
| "num_input_tokens_seen": 7824998400, | |
| "step": 29850 | |
| }, | |
| { | |
| "epoch": 0.2011223163371523, | |
| "grad_norm": 0.6376939415931702, | |
| "learning_rate": 1.7127430391683516e-08, | |
| "loss": 11.8928, | |
| "num_input_tokens_seen": 7838105600, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 0.2014586412808599, | |
| "grad_norm": 0.6745944619178772, | |
| "learning_rate": 4.281888155543978e-09, | |
| "loss": 11.9315, | |
| "num_input_tokens_seen": 7851212800, | |
| "step": 29950 | |
| }, | |
| { | |
| "epoch": 0.20179496622456752, | |
| "grad_norm": 0.6381050944328308, | |
| "learning_rate": 0.0, | |
| "loss": 11.9242, | |
| "num_input_tokens_seen": 7864320000, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.20179496622456752, | |
| "eval_loss": 2.8848958015441895, | |
| "eval_runtime": 142.697, | |
| "eval_samples_per_second": 35.039, | |
| "eval_steps_per_second": 8.76, | |
| "num_input_tokens_seen": 7864320000, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.20179496622456752, | |
| "num_input_tokens_seen": 7864320000, | |
| "step": 30000, | |
| "total_flos": 5.0112805994496e+18, | |
| "train_loss": 13.025330790201822, | |
| "train_runtime": 93774.0829, | |
| "train_samples_per_second": 81.899, | |
| "train_steps_per_second": 0.32, | |
| "train_tokens_per_second": 83864.536 | |
| } | |
| ], | |
| "logging_steps": 50, | |
| "max_steps": 30000, | |
| "num_input_tokens_seen": 7864320000, | |
| "num_train_epochs": 1, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.0112805994496e+18, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |