| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 21487, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0046539768231954205, | |
| "grad_norm": 0.2191164344549179, | |
| "learning_rate": 9.31098696461825e-07, | |
| "loss": 1.8284, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.009307953646390841, | |
| "grad_norm": 0.23232442140579224, | |
| "learning_rate": 1.86219739292365e-06, | |
| "loss": 1.8324, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.013961930469586261, | |
| "grad_norm": 0.27771520614624023, | |
| "learning_rate": 2.793296089385475e-06, | |
| "loss": 1.8126, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.018615907292781682, | |
| "grad_norm": 0.46955519914627075, | |
| "learning_rate": 3.7243947858473e-06, | |
| "loss": 1.7894, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.023269884115977104, | |
| "grad_norm": 0.4999992847442627, | |
| "learning_rate": 4.655493482309125e-06, | |
| "loss": 1.7438, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.027923860939172523, | |
| "grad_norm": 0.6398751139640808, | |
| "learning_rate": 5.58659217877095e-06, | |
| "loss": 1.6969, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.032577837762367945, | |
| "grad_norm": 0.7858927249908447, | |
| "learning_rate": 6.517690875232775e-06, | |
| "loss": 1.7084, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.037231814585563364, | |
| "grad_norm": 0.7117326259613037, | |
| "learning_rate": 7.4487895716946e-06, | |
| "loss": 1.709, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.04188579140875878, | |
| "grad_norm": 0.7050628066062927, | |
| "learning_rate": 8.379888268156426e-06, | |
| "loss": 1.6655, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.04653976823195421, | |
| "grad_norm": 0.9369891285896301, | |
| "learning_rate": 9.31098696461825e-06, | |
| "loss": 1.6765, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.05119374505514963, | |
| "grad_norm": 0.7342795729637146, | |
| "learning_rate": 1.0242085661080074e-05, | |
| "loss": 1.6398, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.055847721878345045, | |
| "grad_norm": 1.5446929931640625, | |
| "learning_rate": 1.11731843575419e-05, | |
| "loss": 1.6714, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.060501698701540464, | |
| "grad_norm": 0.8926479816436768, | |
| "learning_rate": 1.2104283054003724e-05, | |
| "loss": 1.6355, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.06515567552473589, | |
| "grad_norm": 0.9157938361167908, | |
| "learning_rate": 1.303538175046555e-05, | |
| "loss": 1.5984, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.06980965234793131, | |
| "grad_norm": 0.9163572192192078, | |
| "learning_rate": 1.3966480446927374e-05, | |
| "loss": 1.6372, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.07446362917112673, | |
| "grad_norm": 1.3793407678604126, | |
| "learning_rate": 1.48975791433892e-05, | |
| "loss": 1.6025, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.07911760599432215, | |
| "grad_norm": 1.2203880548477173, | |
| "learning_rate": 1.5828677839851026e-05, | |
| "loss": 1.6359, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.08377158281751756, | |
| "grad_norm": 1.18509840965271, | |
| "learning_rate": 1.6759776536312852e-05, | |
| "loss": 1.6273, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.08842555964071298, | |
| "grad_norm": 1.1019757986068726, | |
| "learning_rate": 1.7690875232774675e-05, | |
| "loss": 1.6286, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.09307953646390842, | |
| "grad_norm": 0.6310858726501465, | |
| "learning_rate": 1.86219739292365e-05, | |
| "loss": 1.6132, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.09773351328710383, | |
| "grad_norm": 1.2102677822113037, | |
| "learning_rate": 1.9553072625698326e-05, | |
| "loss": 1.6125, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.10238749011029925, | |
| "grad_norm": 1.16068696975708, | |
| "learning_rate": 1.999964321562415e-05, | |
| "loss": 1.618, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.10704146693349467, | |
| "grad_norm": 1.1368638277053833, | |
| "learning_rate": 1.999695163593931e-05, | |
| "loss": 1.5991, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.11169544375669009, | |
| "grad_norm": 0.9747415781021118, | |
| "learning_rate": 1.9991621965335746e-05, | |
| "loss": 1.5878, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.11634942057988551, | |
| "grad_norm": 0.9197840690612793, | |
| "learning_rate": 1.998365561025775e-05, | |
| "loss": 1.6189, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.12100339740308093, | |
| "grad_norm": 0.9155735373497009, | |
| "learning_rate": 1.997305467294307e-05, | |
| "loss": 1.587, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.12565737422627635, | |
| "grad_norm": 0.7722734212875366, | |
| "learning_rate": 1.9959821950868117e-05, | |
| "loss": 1.5681, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.13031135104947178, | |
| "grad_norm": 1.044062852859497, | |
| "learning_rate": 1.994396093600977e-05, | |
| "loss": 1.5845, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.13496532787266718, | |
| "grad_norm": 0.7750725150108337, | |
| "learning_rate": 1.9925475813923858e-05, | |
| "loss": 1.5936, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.13961930469586262, | |
| "grad_norm": 0.9063745141029358, | |
| "learning_rate": 1.990437146264066e-05, | |
| "loss": 1.5788, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.14427328151905802, | |
| "grad_norm": 0.655881404876709, | |
| "learning_rate": 1.988065345137762e-05, | |
| "loss": 1.5927, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.14892725834225345, | |
| "grad_norm": 1.1087632179260254, | |
| "learning_rate": 1.98543280390697e-05, | |
| "loss": 1.5607, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.1535812351654489, | |
| "grad_norm": 0.7348114848136902, | |
| "learning_rate": 1.982540217271772e-05, | |
| "loss": 1.5344, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.1582352119886443, | |
| "grad_norm": 0.8793938755989075, | |
| "learning_rate": 1.979388348555509e-05, | |
| "loss": 1.5576, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.16288918881183972, | |
| "grad_norm": 0.9496509432792664, | |
| "learning_rate": 1.975978029503352e-05, | |
| "loss": 1.5439, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.16754316563503513, | |
| "grad_norm": 0.9296020269393921, | |
| "learning_rate": 1.9723101600628095e-05, | |
| "loss": 1.6044, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.17219714245823056, | |
| "grad_norm": 0.7520443797111511, | |
| "learning_rate": 1.9683857081462423e-05, | |
| "loss": 1.5319, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.17685111928142597, | |
| "grad_norm": 0.9139896631240845, | |
| "learning_rate": 1.9642057093754426e-05, | |
| "loss": 1.5457, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.1815050961046214, | |
| "grad_norm": 0.781461775302887, | |
| "learning_rate": 1.9597712668083413e-05, | |
| "loss": 1.5604, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.18615907292781683, | |
| "grad_norm": 0.9205677509307861, | |
| "learning_rate": 1.9550835506479266e-05, | |
| "loss": 1.5255, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.19081304975101224, | |
| "grad_norm": 0.7341404557228088, | |
| "learning_rate": 1.9501437979334358e-05, | |
| "loss": 1.5876, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.19546702657420767, | |
| "grad_norm": 0.8709468245506287, | |
| "learning_rate": 1.9449533122139157e-05, | |
| "loss": 1.5469, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.20012100339740307, | |
| "grad_norm": 0.9768949151039124, | |
| "learning_rate": 1.939513463204229e-05, | |
| "loss": 1.5417, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.2047749802205985, | |
| "grad_norm": 0.7858554720878601, | |
| "learning_rate": 1.9338256864236024e-05, | |
| "loss": 1.5672, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.2094289570437939, | |
| "grad_norm": 0.7643204927444458, | |
| "learning_rate": 1.927891482816805e-05, | |
| "loss": 1.5032, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.21408293386698934, | |
| "grad_norm": 1.2041771411895752, | |
| "learning_rate": 1.9217124183580674e-05, | |
| "loss": 1.51, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.21873691069018475, | |
| "grad_norm": 0.9453679323196411, | |
| "learning_rate": 1.9152901236378364e-05, | |
| "loss": 1.5662, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.22339088751338018, | |
| "grad_norm": 0.6478219628334045, | |
| "learning_rate": 1.908626293432479e-05, | |
| "loss": 1.5932, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.22804486433657561, | |
| "grad_norm": 0.838474690914154, | |
| "learning_rate": 1.901722686257051e-05, | |
| "loss": 1.567, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.23269884115977102, | |
| "grad_norm": 1.1010560989379883, | |
| "learning_rate": 1.894581123901238e-05, | |
| "loss": 1.4849, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.23735281798296645, | |
| "grad_norm": 0.8271942734718323, | |
| "learning_rate": 1.88720349094861e-05, | |
| "loss": 1.56, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.24200679480616186, | |
| "grad_norm": 0.9592469334602356, | |
| "learning_rate": 1.8795917342792967e-05, | |
| "loss": 1.6055, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.2466607716293573, | |
| "grad_norm": 0.7423321008682251, | |
| "learning_rate": 1.8717478625562256e-05, | |
| "loss": 1.5268, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.2513147484525527, | |
| "grad_norm": 1.0170488357543945, | |
| "learning_rate": 1.8636739456950593e-05, | |
| "loss": 1.5765, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.2559687252757481, | |
| "grad_norm": 0.7664517760276794, | |
| "learning_rate": 1.8553721143179626e-05, | |
| "loss": 1.5625, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.26062270209894356, | |
| "grad_norm": 0.712568461894989, | |
| "learning_rate": 1.8468445591913587e-05, | |
| "loss": 1.5853, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.26527667892213896, | |
| "grad_norm": 0.9137941002845764, | |
| "learning_rate": 1.8380935306478036e-05, | |
| "loss": 1.5628, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.26993065574533437, | |
| "grad_norm": 0.9478745460510254, | |
| "learning_rate": 1.829121337992151e-05, | |
| "loss": 1.4895, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.27458463256852983, | |
| "grad_norm": 0.5866535305976868, | |
| "learning_rate": 1.8199303488921493e-05, | |
| "loss": 1.5484, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.27923860939172523, | |
| "grad_norm": 0.9393543004989624, | |
| "learning_rate": 1.810522988753638e-05, | |
| "loss": 1.5307, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.28389258621492064, | |
| "grad_norm": 0.6128756403923035, | |
| "learning_rate": 1.8009017400805094e-05, | |
| "loss": 1.5821, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.28854656303811604, | |
| "grad_norm": 0.8191483020782471, | |
| "learning_rate": 1.7910691418196022e-05, | |
| "loss": 1.5206, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.2932005398613115, | |
| "grad_norm": 0.8590126633644104, | |
| "learning_rate": 1.7810277886906998e-05, | |
| "loss": 1.5488, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.2978545166845069, | |
| "grad_norm": 0.6708762645721436, | |
| "learning_rate": 1.7707803305018097e-05, | |
| "loss": 1.5383, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.3025084935077023, | |
| "grad_norm": 0.6897620558738708, | |
| "learning_rate": 1.760329471449909e-05, | |
| "loss": 1.5497, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.3071624703308978, | |
| "grad_norm": 0.7875939607620239, | |
| "learning_rate": 1.7496779694073333e-05, | |
| "loss": 1.5251, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.3118164471540932, | |
| "grad_norm": 0.9739288091659546, | |
| "learning_rate": 1.738828635194003e-05, | |
| "loss": 1.5235, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.3164704239772886, | |
| "grad_norm": 0.9243332743644714, | |
| "learning_rate": 1.7277843318356798e-05, | |
| "loss": 1.5256, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.321124400800484, | |
| "grad_norm": 1.0979468822479248, | |
| "learning_rate": 1.7165479738084405e-05, | |
| "loss": 1.5841, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.32577837762367945, | |
| "grad_norm": 0.931398868560791, | |
| "learning_rate": 1.7051225262695812e-05, | |
| "loss": 1.5408, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.33043235444687485, | |
| "grad_norm": 0.7612720131874084, | |
| "learning_rate": 1.693511004275139e-05, | |
| "loss": 1.4606, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.33508633127007026, | |
| "grad_norm": 1.0624606609344482, | |
| "learning_rate": 1.6817164719842548e-05, | |
| "loss": 1.5322, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.3397403080932657, | |
| "grad_norm": 0.6526899933815002, | |
| "learning_rate": 1.66974204185057e-05, | |
| "loss": 1.5801, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.3443942849164611, | |
| "grad_norm": 0.7007707953453064, | |
| "learning_rate": 1.6575908738008847e-05, | |
| "loss": 1.4867, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.34904826173965653, | |
| "grad_norm": 0.8433778882026672, | |
| "learning_rate": 1.645266174401285e-05, | |
| "loss": 1.5193, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.35370223856285193, | |
| "grad_norm": 1.0169028043746948, | |
| "learning_rate": 1.632771196010965e-05, | |
| "loss": 1.53, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.3583562153860474, | |
| "grad_norm": 0.9545369744300842, | |
| "learning_rate": 1.6201092359239624e-05, | |
| "loss": 1.4862, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.3630101922092428, | |
| "grad_norm": 1.2505319118499756, | |
| "learning_rate": 1.607283635499036e-05, | |
| "loss": 1.5349, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.3676641690324382, | |
| "grad_norm": 0.8437191843986511, | |
| "learning_rate": 1.5942977792779193e-05, | |
| "loss": 1.5245, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.37231814585563366, | |
| "grad_norm": 0.8392757773399353, | |
| "learning_rate": 1.581155094092171e-05, | |
| "loss": 1.519, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.37697212267882907, | |
| "grad_norm": 0.7851161956787109, | |
| "learning_rate": 1.5678590481588743e-05, | |
| "loss": 1.5301, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.3816260995020245, | |
| "grad_norm": 0.8554267883300781, | |
| "learning_rate": 1.5544131501654078e-05, | |
| "loss": 1.5143, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.3862800763252199, | |
| "grad_norm": 0.9319170117378235, | |
| "learning_rate": 1.5408209483435402e-05, | |
| "loss": 1.5079, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.39093405314841534, | |
| "grad_norm": 0.9828823804855347, | |
| "learning_rate": 1.5270860295330912e-05, | |
| "loss": 1.5248, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.39558802997161074, | |
| "grad_norm": 0.8951559662818909, | |
| "learning_rate": 1.5132120182353984e-05, | |
| "loss": 1.4728, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.40024200679480615, | |
| "grad_norm": 0.8046073317527771, | |
| "learning_rate": 1.499202575656852e-05, | |
| "loss": 1.5096, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.4048959836180016, | |
| "grad_norm": 0.7216867208480835, | |
| "learning_rate": 1.4850613987427398e-05, | |
| "loss": 1.538, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.409549960441197, | |
| "grad_norm": 0.5751676559448242, | |
| "learning_rate": 1.470792219201662e-05, | |
| "loss": 1.508, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.4142039372643924, | |
| "grad_norm": 0.7514901161193848, | |
| "learning_rate": 1.4563988025207726e-05, | |
| "loss": 1.5365, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.4188579140875878, | |
| "grad_norm": 0.8872545957565308, | |
| "learning_rate": 1.4418849469721073e-05, | |
| "loss": 1.4705, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.4235118909107833, | |
| "grad_norm": 0.772324800491333, | |
| "learning_rate": 1.4272544826102576e-05, | |
| "loss": 1.521, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.4281658677339787, | |
| "grad_norm": 0.8108826279640198, | |
| "learning_rate": 1.4125112702616607e-05, | |
| "loss": 1.4958, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.4328198445571741, | |
| "grad_norm": 0.8969531655311584, | |
| "learning_rate": 1.3976592005057684e-05, | |
| "loss": 1.4596, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.4374738213803695, | |
| "grad_norm": 0.93816739320755, | |
| "learning_rate": 1.3827021926483622e-05, | |
| "loss": 1.5404, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.44212779820356496, | |
| "grad_norm": 0.5850194096565247, | |
| "learning_rate": 1.3676441936872904e-05, | |
| "loss": 1.5342, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.44678177502676036, | |
| "grad_norm": 0.903218686580658, | |
| "learning_rate": 1.352489177270898e-05, | |
| "loss": 1.5315, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.45143575184995577, | |
| "grad_norm": 0.9102717041969299, | |
| "learning_rate": 1.337241142649419e-05, | |
| "loss": 1.5105, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.45608972867315123, | |
| "grad_norm": 0.5675022006034851, | |
| "learning_rate": 1.3219041136196202e-05, | |
| "loss": 1.465, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.46074370549634663, | |
| "grad_norm": 1.007855772972107, | |
| "learning_rate": 1.3064821374629606e-05, | |
| "loss": 1.5131, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.46539768231954204, | |
| "grad_norm": 0.8553173542022705, | |
| "learning_rate": 1.2909792838775571e-05, | |
| "loss": 1.5054, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.47005165914273744, | |
| "grad_norm": 0.9074007868766785, | |
| "learning_rate": 1.2755558069133832e-05, | |
| "loss": 1.5582, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.4747056359659329, | |
| "grad_norm": 0.7873587608337402, | |
| "learning_rate": 1.2599041981920994e-05, | |
| "loss": 1.5109, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.4793596127891283, | |
| "grad_norm": 0.9822585582733154, | |
| "learning_rate": 1.2441840034728076e-05, | |
| "loss": 1.5067, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.4840135896123237, | |
| "grad_norm": 1.0133826732635498, | |
| "learning_rate": 1.2283993711503824e-05, | |
| "loss": 1.4886, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.4886675664355192, | |
| "grad_norm": 1.1799577474594116, | |
| "learning_rate": 1.2125544666241078e-05, | |
| "loss": 1.4655, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.4933215432587146, | |
| "grad_norm": 0.927507221698761, | |
| "learning_rate": 1.1966534711984717e-05, | |
| "loss": 1.4748, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.49797552008191, | |
| "grad_norm": 0.8527470827102661, | |
| "learning_rate": 1.1808603529490986e-05, | |
| "loss": 1.4607, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.5026294969051054, | |
| "grad_norm": 0.8991233706474304, | |
| "learning_rate": 1.1648602337072668e-05, | |
| "loss": 1.5432, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.5072834737283008, | |
| "grad_norm": 0.7587746381759644, | |
| "learning_rate": 1.1488166095750207e-05, | |
| "loss": 1.5053, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.5119374505514962, | |
| "grad_norm": 0.6698244214057922, | |
| "learning_rate": 1.132733714296872e-05, | |
| "loss": 1.4281, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.5165914273746917, | |
| "grad_norm": 0.848193883895874, | |
| "learning_rate": 1.1166157919805772e-05, | |
| "loss": 1.5043, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.5212454041978871, | |
| "grad_norm": 1.0735975503921509, | |
| "learning_rate": 1.1004670959771624e-05, | |
| "loss": 1.5072, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.5258993810210825, | |
| "grad_norm": 1.1171404123306274, | |
| "learning_rate": 1.0842918877585058e-05, | |
| "loss": 1.4864, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.5305533578442779, | |
| "grad_norm": 0.8309624195098877, | |
| "learning_rate": 1.0680944357927836e-05, | |
| "loss": 1.5373, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.5352073346674734, | |
| "grad_norm": 0.8913981318473816, | |
| "learning_rate": 1.0518790144180635e-05, | |
| "loss": 1.4432, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.5398613114906687, | |
| "grad_norm": 0.8174418807029724, | |
| "learning_rate": 1.0356499027143519e-05, | |
| "loss": 1.4852, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.5445152883138642, | |
| "grad_norm": 0.9098989367485046, | |
| "learning_rate": 1.0194113833743897e-05, | |
| "loss": 1.5405, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.5491692651370597, | |
| "grad_norm": 0.983168363571167, | |
| "learning_rate": 1.0031677415734946e-05, | |
| "loss": 1.5014, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.553823241960255, | |
| "grad_norm": 0.793428361415863, | |
| "learning_rate": 9.869232638387492e-06, | |
| "loss": 1.5047, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.5584772187834505, | |
| "grad_norm": 0.9047326445579529, | |
| "learning_rate": 9.706822369178289e-06, | |
| "loss": 1.4991, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.5631311956066459, | |
| "grad_norm": 0.825706422328949, | |
| "learning_rate": 9.54448946647775e-06, | |
| "loss": 1.5175, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.5677851724298413, | |
| "grad_norm": 0.9917203187942505, | |
| "learning_rate": 9.382276768240104e-06, | |
| "loss": 1.4916, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.5724391492530367, | |
| "grad_norm": 0.8516975045204163, | |
| "learning_rate": 9.220227080698847e-06, | |
| "loss": 1.5003, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.5770931260762321, | |
| "grad_norm": 0.7885426878929138, | |
| "learning_rate": 9.058383167070676e-06, | |
| "loss": 1.5427, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.5817471028994275, | |
| "grad_norm": 0.9626107215881348, | |
| "learning_rate": 8.89678773627071e-06, | |
| "loss": 1.4672, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.586401079722623, | |
| "grad_norm": 1.023768663406372, | |
| "learning_rate": 8.735483431642033e-06, | |
| "loss": 1.5236, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.5910550565458184, | |
| "grad_norm": 0.9998337626457214, | |
| "learning_rate": 8.574512819702594e-06, | |
| "loss": 1.4932, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.5957090333690138, | |
| "grad_norm": 0.9876322746276855, | |
| "learning_rate": 8.413918378912325e-06, | |
| "loss": 1.4849, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.6003630101922093, | |
| "grad_norm": 0.9000487923622131, | |
| "learning_rate": 8.253742488463531e-06, | |
| "loss": 1.4532, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.6050169870154046, | |
| "grad_norm": 0.9810521006584167, | |
| "learning_rate": 8.094027417097468e-06, | |
| "loss": 1.4818, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.6096709638386001, | |
| "grad_norm": 0.8786018490791321, | |
| "learning_rate": 7.936404805248885e-06, | |
| "loss": 1.524, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.6143249406617955, | |
| "grad_norm": 0.8514164686203003, | |
| "learning_rate": 7.777732023413952e-06, | |
| "loss": 1.4763, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.6189789174849909, | |
| "grad_norm": 0.603767991065979, | |
| "learning_rate": 7.619645674842026e-06, | |
| "loss": 1.4977, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.6236328943081864, | |
| "grad_norm": 0.8721168637275696, | |
| "learning_rate": 7.462187476866129e-06, | |
| "loss": 1.4339, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.6282868711313818, | |
| "grad_norm": 0.8393872380256653, | |
| "learning_rate": 7.305398981056899e-06, | |
| "loss": 1.4888, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.6329408479545772, | |
| "grad_norm": 1.0900713205337524, | |
| "learning_rate": 7.149321562257598e-06, | |
| "loss": 1.4612, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.6375948247777726, | |
| "grad_norm": 0.857639491558075, | |
| "learning_rate": 6.9939964076657e-06, | |
| "loss": 1.4866, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.642248801600968, | |
| "grad_norm": 0.7077500820159912, | |
| "learning_rate": 6.839464505964017e-06, | |
| "loss": 1.4587, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.6469027784241634, | |
| "grad_norm": 0.9254826307296753, | |
| "learning_rate": 6.6857666365042354e-06, | |
| "loss": 1.463, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.6515567552473589, | |
| "grad_norm": 0.8199195265769958, | |
| "learning_rate": 6.5329433585456625e-06, | |
| "loss": 1.5347, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.6562107320705542, | |
| "grad_norm": 1.0524635314941406, | |
| "learning_rate": 6.381035000552058e-06, | |
| "loss": 1.5253, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.6608647088937497, | |
| "grad_norm": 0.9598236083984375, | |
| "learning_rate": 6.230081649549371e-06, | |
| "loss": 1.5064, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.6655186857169452, | |
| "grad_norm": 0.8717284202575684, | |
| "learning_rate": 6.080123140547213e-06, | |
| "loss": 1.5284, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.6701726625401405, | |
| "grad_norm": 0.7019008994102478, | |
| "learning_rate": 5.931199046026791e-06, | |
| "loss": 1.5057, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.674826639363336, | |
| "grad_norm": 0.7632008790969849, | |
| "learning_rate": 5.78334866549816e-06, | |
| "loss": 1.4382, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.6794806161865314, | |
| "grad_norm": 0.9148026704788208, | |
| "learning_rate": 5.63661101512949e-06, | |
| "loss": 1.4656, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.6841345930097268, | |
| "grad_norm": 1.0867414474487305, | |
| "learning_rate": 5.491024817451108e-06, | |
| "loss": 1.4611, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.6887885698329222, | |
| "grad_norm": 0.9941554665565491, | |
| "learning_rate": 5.346628491137016e-06, | |
| "loss": 1.471, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.6934425466561176, | |
| "grad_norm": 0.756461501121521, | |
| "learning_rate": 5.203460140866625e-06, | |
| "loss": 1.4691, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.6980965234793131, | |
| "grad_norm": 0.9088025093078613, | |
| "learning_rate": 5.061557547269322e-06, | |
| "loss": 1.4712, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.7027505003025085, | |
| "grad_norm": 0.850143313407898, | |
| "learning_rate": 4.920958156954544e-06, | |
| "loss": 1.4434, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.7074044771257039, | |
| "grad_norm": 1.072135090827942, | |
| "learning_rate": 4.781699072630018e-06, | |
| "loss": 1.5095, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.7120584539488993, | |
| "grad_norm": 0.8175047039985657, | |
| "learning_rate": 4.643817043310759e-06, | |
| "loss": 1.455, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.7167124307720948, | |
| "grad_norm": 0.8658547401428223, | |
| "learning_rate": 4.5073484546213705e-06, | |
| "loss": 1.4545, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.7213664075952901, | |
| "grad_norm": 0.9122436046600342, | |
| "learning_rate": 4.372329319194262e-06, | |
| "loss": 1.5157, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.7260203844184856, | |
| "grad_norm": 1.0552470684051514, | |
| "learning_rate": 4.238795267166298e-06, | |
| "loss": 1.4664, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.7306743612416811, | |
| "grad_norm": 0.8493475914001465, | |
| "learning_rate": 4.106781536776383e-06, | |
| "loss": 1.5032, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.7353283380648764, | |
| "grad_norm": 0.9446743130683899, | |
| "learning_rate": 3.976322965066436e-06, | |
| "loss": 1.4747, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.7399823148880719, | |
| "grad_norm": 0.7051513195037842, | |
| "learning_rate": 3.847453978688287e-06, | |
| "loss": 1.4649, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.7446362917112673, | |
| "grad_norm": 0.941286563873291, | |
| "learning_rate": 3.720208584818833e-06, | |
| "loss": 1.4755, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.7492902685344627, | |
| "grad_norm": 0.9868441820144653, | |
| "learning_rate": 3.5946203621858867e-06, | |
| "loss": 1.5239, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.7539442453576581, | |
| "grad_norm": 1.0158665180206299, | |
| "learning_rate": 3.4707224522071294e-06, | |
| "loss": 1.4625, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.7585982221808535, | |
| "grad_norm": 1.0703004598617554, | |
| "learning_rate": 3.3485475502444055e-06, | |
| "loss": 1.4627, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.763252199004049, | |
| "grad_norm": 0.9671617746353149, | |
| "learning_rate": 3.2281278969757913e-06, | |
| "loss": 1.4777, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.7679061758272444, | |
| "grad_norm": 1.0113691091537476, | |
| "learning_rate": 3.109495269887587e-06, | |
| "loss": 1.4954, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.7725601526504398, | |
| "grad_norm": 1.0288777351379395, | |
| "learning_rate": 2.9926809748885964e-06, | |
| "loss": 1.5105, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.7772141294736352, | |
| "grad_norm": 0.851613461971283, | |
| "learning_rate": 2.877715838048811e-06, | |
| "loss": 1.4307, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.7818681062968307, | |
| "grad_norm": 0.7523232102394104, | |
| "learning_rate": 2.764630197464729e-06, | |
| "loss": 1.4429, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.786522083120026, | |
| "grad_norm": 1.0177912712097168, | |
| "learning_rate": 2.653453895253466e-06, | |
| "loss": 1.4585, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.7911760599432215, | |
| "grad_norm": 0.7985556721687317, | |
| "learning_rate": 2.5442162696777095e-06, | |
| "loss": 1.4912, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.795830036766417, | |
| "grad_norm": 1.0856115818023682, | |
| "learning_rate": 2.4369461474036825e-06, | |
| "loss": 1.4824, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.8004840135896123, | |
| "grad_norm": 1.0469813346862793, | |
| "learning_rate": 2.332714608087351e-06, | |
| "loss": 1.4442, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.8051379904128078, | |
| "grad_norm": 0.7078351378440857, | |
| "learning_rate": 2.2294435164482266e-06, | |
| "loss": 1.5291, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.8097919672360032, | |
| "grad_norm": 0.9295824766159058, | |
| "learning_rate": 2.1282229933463116e-06, | |
| "loss": 1.4083, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.8144459440591986, | |
| "grad_norm": 0.7632860541343689, | |
| "learning_rate": 2.0290797498183858e-06, | |
| "loss": 1.4367, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.819099920882394, | |
| "grad_norm": 0.9528563022613525, | |
| "learning_rate": 1.9320399487288832e-06, | |
| "loss": 1.445, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.8237538977055894, | |
| "grad_norm": 0.8745253682136536, | |
| "learning_rate": 1.8371291978657857e-06, | |
| "loss": 1.4694, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.8284078745287848, | |
| "grad_norm": 0.9276142120361328, | |
| "learning_rate": 1.7443725431830127e-06, | |
| "loss": 1.4682, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.8330618513519803, | |
| "grad_norm": 0.8589171171188354, | |
| "learning_rate": 1.6537944621910307e-06, | |
| "loss": 1.4999, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.8377158281751756, | |
| "grad_norm": 0.8758667707443237, | |
| "learning_rate": 1.5654188574975215e-06, | |
| "loss": 1.4621, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.8423698049983711, | |
| "grad_norm": 0.9397442936897278, | |
| "learning_rate": 1.4792690504997087e-06, | |
| "loss": 1.4598, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.8470237818215666, | |
| "grad_norm": 1.0151734352111816, | |
| "learning_rate": 1.3953677752301032e-06, | |
| "loss": 1.4698, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.8516777586447619, | |
| "grad_norm": 0.9862604737281799, | |
| "learning_rate": 1.3137371723572257e-06, | |
| "loss": 1.4478, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.8563317354679574, | |
| "grad_norm": 1.16762113571167, | |
| "learning_rate": 1.2343987833429062e-06, | |
| "loss": 1.4806, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.8609857122911528, | |
| "grad_norm": 1.06341552734375, | |
| "learning_rate": 1.1573735447577327e-06, | |
| "loss": 1.4555, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.8656396891143482, | |
| "grad_norm": 1.0032464265823364, | |
| "learning_rate": 1.082681782756102e-06, | |
| "loss": 1.4471, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.8702936659375436, | |
| "grad_norm": 0.9144641160964966, | |
| "learning_rate": 1.0103432077123532e-06, | |
| "loss": 1.4721, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.874947642760739, | |
| "grad_norm": 1.0318922996520996, | |
| "learning_rate": 9.403769090194192e-07, | |
| "loss": 1.4542, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.8796016195839345, | |
| "grad_norm": 1.0590611696243286, | |
| "learning_rate": 8.72801350051331e-07, | |
| "loss": 1.497, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.8842555964071299, | |
| "grad_norm": 0.5952686667442322, | |
| "learning_rate": 8.076343632909211e-07, | |
| "loss": 1.4613, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.8889095732303253, | |
| "grad_norm": 0.8993837237358093, | |
| "learning_rate": 7.448931456240205e-07, | |
| "loss": 1.5027, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.8935635500535207, | |
| "grad_norm": 0.8999748229980469, | |
| "learning_rate": 6.845942538013894e-07, | |
| "loss": 1.4677, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.8982175268767162, | |
| "grad_norm": 1.1852108240127563, | |
| "learning_rate": 6.273197876706216e-07, | |
| "loss": 1.4537, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.9028715036999115, | |
| "grad_norm": 0.8120535016059875, | |
| "learning_rate": 5.71927827141906e-07, | |
| "loss": 1.4965, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.907525480523107, | |
| "grad_norm": 0.8027576804161072, | |
| "learning_rate": 5.190238361949229e-07, | |
| "loss": 1.5109, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.9121794573463025, | |
| "grad_norm": 1.2424169778823853, | |
| "learning_rate": 4.686217756392464e-07, | |
| "loss": 1.4614, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.9168334341694978, | |
| "grad_norm": 1.248057246208191, | |
| "learning_rate": 4.207349460512022e-07, | |
| "loss": 1.4786, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.9214874109926933, | |
| "grad_norm": 0.8463609218597412, | |
| "learning_rate": 3.7537598426397925e-07, | |
| "loss": 1.4714, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.9261413878158887, | |
| "grad_norm": 0.6483246684074402, | |
| "learning_rate": 3.325568600329143e-07, | |
| "loss": 1.5209, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.9307953646390841, | |
| "grad_norm": 0.6988236904144287, | |
| "learning_rate": 2.922888728767903e-07, | |
| "loss": 1.485, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.9354493414622795, | |
| "grad_norm": 0.9654485583305359, | |
| "learning_rate": 2.545826490960146e-07, | |
| "loss": 1.4532, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.9401033182854749, | |
| "grad_norm": 1.0685368776321411, | |
| "learning_rate": 2.194481389684433e-07, | |
| "loss": 1.4925, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.9447572951086703, | |
| "grad_norm": 1.161594033241272, | |
| "learning_rate": 1.8689461412362164e-07, | |
| "loss": 1.4376, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.9494112719318658, | |
| "grad_norm": 1.1302499771118164, | |
| "learning_rate": 1.5693066509608333e-07, | |
| "loss": 1.4776, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.9540652487550612, | |
| "grad_norm": 1.0777382850646973, | |
| "learning_rate": 1.295641990584007e-07, | |
| "loss": 1.4702, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.9587192255782566, | |
| "grad_norm": 1.027956247329712, | |
| "learning_rate": 1.0503714003311738e-07, | |
| "loss": 1.4699, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.9633732024014521, | |
| "grad_norm": 0.7537710666656494, | |
| "learning_rate": 8.286047531811392e-08, | |
| "loss": 1.489, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.9680271792246474, | |
| "grad_norm": 0.7544810175895691, | |
| "learning_rate": 6.33008399409063e-08, | |
| "loss": 1.4563, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.9726811560478429, | |
| "grad_norm": 0.9680439829826355, | |
| "learning_rate": 4.636339548457991e-08, | |
| "loss": 1.4345, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.9773351328710383, | |
| "grad_norm": 0.773059606552124, | |
| "learning_rate": 3.205261156347783e-08, | |
| "loss": 1.4591, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.9819891096942337, | |
| "grad_norm": 0.8769505023956299, | |
| "learning_rate": 2.0372264643713223e-08, | |
| "loss": 1.4756, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.9866430865174292, | |
| "grad_norm": 0.9209620356559753, | |
| "learning_rate": 1.132543704660316e-08, | |
| "loss": 1.4957, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.9912970633406246, | |
| "grad_norm": 0.9970301389694214, | |
| "learning_rate": 4.914516135275937e-09, | |
| "loss": 1.4321, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.99595104016382, | |
| "grad_norm": 0.6997882127761841, | |
| "learning_rate": 1.1411936846705828e-09, | |
| "loss": 1.4656, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 21487, | |
| "total_flos": 3.101747713212678e+18, | |
| "train_loss": 1.5223615952344096, | |
| "train_runtime": 9293.396, | |
| "train_samples_per_second": 4.624, | |
| "train_steps_per_second": 2.312 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 21487, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.101747713212678e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |