| { |
| "best_metric": 3.4021902084350586, |
| "best_model_checkpoint": "/scratch/cl5625/exceptions/models/100M__6910/checkpoint-50000", |
| "epoch": 5.390835579514825, |
| "eval_steps": 1000, |
| "global_step": 50000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.005390835579514825, |
| "grad_norm": 2.385856866836548, |
| "learning_rate": 0.000276, |
| "loss": 9.0223, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.01078167115902965, |
| "grad_norm": 2.4063053131103516, |
| "learning_rate": 0.0005759999999999999, |
| "loss": 6.944, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.016172506738544475, |
| "grad_norm": 1.2234184741973877, |
| "learning_rate": 0.000599702104695089, |
| "loss": 6.4879, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.0215633423180593, |
| "grad_norm": 1.0247325897216797, |
| "learning_rate": 0.0005993783054506205, |
| "loss": 6.2173, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.026954177897574125, |
| "grad_norm": 1.8562301397323608, |
| "learning_rate": 0.0005990545062061521, |
| "loss": 6.0821, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.03234501347708895, |
| "grad_norm": 1.6646568775177002, |
| "learning_rate": 0.0005987307069616836, |
| "loss": 5.957, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.03773584905660377, |
| "grad_norm": 0.8979423642158508, |
| "learning_rate": 0.0005984069077172153, |
| "loss": 5.8884, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.0431266846361186, |
| "grad_norm": 1.5213276147842407, |
| "learning_rate": 0.0005980831084727469, |
| "loss": 5.8265, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.04851752021563342, |
| "grad_norm": 1.5135802030563354, |
| "learning_rate": 0.0005977593092282784, |
| "loss": 5.7629, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.05390835579514825, |
| "grad_norm": 1.3815815448760986, |
| "learning_rate": 0.00059743550998381, |
| "loss": 5.6521, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.05929919137466307, |
| "grad_norm": 0.8758553266525269, |
| "learning_rate": 0.0005971117107393416, |
| "loss": 5.602, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.0646900269541779, |
| "grad_norm": 1.5150190591812134, |
| "learning_rate": 0.0005967879114948732, |
| "loss": 5.5349, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.07008086253369272, |
| "grad_norm": 1.2760694026947021, |
| "learning_rate": 0.0005964641122504047, |
| "loss": 5.4673, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.07547169811320754, |
| "grad_norm": 1.3729991912841797, |
| "learning_rate": 0.0005961403130059363, |
| "loss": 5.3942, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.08086253369272237, |
| "grad_norm": 1.3321388959884644, |
| "learning_rate": 0.0005958165137614678, |
| "loss": 5.3261, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.0862533692722372, |
| "grad_norm": 1.0924150943756104, |
| "learning_rate": 0.0005954927145169995, |
| "loss": 5.2547, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.09164420485175202, |
| "grad_norm": 1.0701996088027954, |
| "learning_rate": 0.0005951689152725309, |
| "loss": 5.2296, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.09703504043126684, |
| "grad_norm": 1.1670644283294678, |
| "learning_rate": 0.0005948451160280626, |
| "loss": 5.1749, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.10242587601078167, |
| "grad_norm": 0.833484411239624, |
| "learning_rate": 0.0005945213167835941, |
| "loss": 5.1112, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.1078167115902965, |
| "grad_norm": 1.1527302265167236, |
| "learning_rate": 0.0005941975175391257, |
| "loss": 5.1167, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.1078167115902965, |
| "eval_accuracy": 0.22651139154904357, |
| "eval_loss": 5.03156852722168, |
| "eval_runtime": 184.3163, |
| "eval_samples_per_second": 97.718, |
| "eval_steps_per_second": 6.109, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.11320754716981132, |
| "grad_norm": 0.9724206924438477, |
| "learning_rate": 0.0005938737182946572, |
| "loss": 5.0469, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.11859838274932614, |
| "grad_norm": 1.3919302225112915, |
| "learning_rate": 0.0005935499190501888, |
| "loss": 5.0082, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.12398921832884097, |
| "grad_norm": 1.0906733274459839, |
| "learning_rate": 0.0005932261198057204, |
| "loss": 5.0148, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.1293800539083558, |
| "grad_norm": 0.9858213663101196, |
| "learning_rate": 0.000592902320561252, |
| "loss": 4.9728, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.1347708894878706, |
| "grad_norm": 0.9534234404563904, |
| "learning_rate": 0.0005925785213167835, |
| "loss": 4.9499, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.14016172506738545, |
| "grad_norm": 0.9149211049079895, |
| "learning_rate": 0.0005922547220723151, |
| "loss": 4.9035, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.14555256064690028, |
| "grad_norm": 1.0197161436080933, |
| "learning_rate": 0.0005919309228278468, |
| "loss": 4.8524, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.1509433962264151, |
| "grad_norm": 1.1303359270095825, |
| "learning_rate": 0.0005916071235833783, |
| "loss": 4.8606, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.15633423180592992, |
| "grad_norm": 1.1484166383743286, |
| "learning_rate": 0.0005912833243389097, |
| "loss": 4.8134, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.16172506738544473, |
| "grad_norm": 0.8566966652870178, |
| "learning_rate": 0.0005909595250944414, |
| "loss": 4.8164, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.16711590296495957, |
| "grad_norm": 1.603456735610962, |
| "learning_rate": 0.000590635725849973, |
| "loss": 4.8012, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.1725067385444744, |
| "grad_norm": 1.0490790605545044, |
| "learning_rate": 0.0005903119266055045, |
| "loss": 4.7774, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.1778975741239892, |
| "grad_norm": 1.2788410186767578, |
| "learning_rate": 0.0005899881273610361, |
| "loss": 4.7472, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.18328840970350405, |
| "grad_norm": 1.3455950021743774, |
| "learning_rate": 0.0005896643281165677, |
| "loss": 4.7321, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.18867924528301888, |
| "grad_norm": 1.0504180192947388, |
| "learning_rate": 0.0005893405288720993, |
| "loss": 4.6779, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.1940700808625337, |
| "grad_norm": 0.8884590268135071, |
| "learning_rate": 0.0005890167296276308, |
| "loss": 4.7042, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.19946091644204852, |
| "grad_norm": 1.235971212387085, |
| "learning_rate": 0.0005886929303831624, |
| "loss": 4.6545, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.20485175202156333, |
| "grad_norm": 0.8869856595993042, |
| "learning_rate": 0.0005883691311386939, |
| "loss": 4.6304, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.21024258760107817, |
| "grad_norm": 0.8559747338294983, |
| "learning_rate": 0.0005880453318942256, |
| "loss": 4.6107, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.215633423180593, |
| "grad_norm": 0.8249241709709167, |
| "learning_rate": 0.0005877215326497571, |
| "loss": 4.5939, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.215633423180593, |
| "eval_accuracy": 0.27012444341200054, |
| "eval_loss": 4.512342929840088, |
| "eval_runtime": 183.3017, |
| "eval_samples_per_second": 98.259, |
| "eval_steps_per_second": 6.143, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.2210242587601078, |
| "grad_norm": 1.0064095258712769, |
| "learning_rate": 0.0005873977334052887, |
| "loss": 4.5762, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.22641509433962265, |
| "grad_norm": 0.8109007477760315, |
| "learning_rate": 0.0005870739341608202, |
| "loss": 4.5586, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.23180592991913745, |
| "grad_norm": 0.8962969183921814, |
| "learning_rate": 0.0005867501349163519, |
| "loss": 4.5323, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.2371967654986523, |
| "grad_norm": 0.8573135733604431, |
| "learning_rate": 0.0005864263356718833, |
| "loss": 4.5247, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.24258760107816713, |
| "grad_norm": 0.8410763740539551, |
| "learning_rate": 0.000586102536427415, |
| "loss": 4.5036, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.24797843665768193, |
| "grad_norm": 0.7351256012916565, |
| "learning_rate": 0.0005857787371829465, |
| "loss": 4.4966, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.25336927223719674, |
| "grad_norm": 1.2595137357711792, |
| "learning_rate": 0.0005854549379384781, |
| "loss": 4.4755, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.2587601078167116, |
| "grad_norm": 0.8755801320075989, |
| "learning_rate": 0.0005851311386940096, |
| "loss": 4.4587, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.2641509433962264, |
| "grad_norm": 0.7965346574783325, |
| "learning_rate": 0.0005848073394495412, |
| "loss": 4.4459, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.2695417789757412, |
| "grad_norm": 0.9087278842926025, |
| "learning_rate": 0.0005844835402050728, |
| "loss": 4.4255, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.2749326145552561, |
| "grad_norm": 0.754734218120575, |
| "learning_rate": 0.0005841597409606044, |
| "loss": 4.3955, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.2803234501347709, |
| "grad_norm": 0.9010357856750488, |
| "learning_rate": 0.000583835941716136, |
| "loss": 4.4084, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.2857142857142857, |
| "grad_norm": 0.8619740009307861, |
| "learning_rate": 0.0005835121424716675, |
| "loss": 4.3899, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.29110512129380056, |
| "grad_norm": 0.9053239226341248, |
| "learning_rate": 0.0005831883432271992, |
| "loss": 4.3775, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.29649595687331537, |
| "grad_norm": 0.8653854727745056, |
| "learning_rate": 0.0005828645439827307, |
| "loss": 4.3731, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.3018867924528302, |
| "grad_norm": 0.7907358407974243, |
| "learning_rate": 0.0005825407447382622, |
| "loss": 4.3277, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.30727762803234504, |
| "grad_norm": 0.692861795425415, |
| "learning_rate": 0.0005822169454937938, |
| "loss": 4.3532, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.31266846361185985, |
| "grad_norm": 0.7582786083221436, |
| "learning_rate": 0.0005818931462493254, |
| "loss": 4.3195, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.31805929919137466, |
| "grad_norm": 0.8971874117851257, |
| "learning_rate": 0.0005815693470048569, |
| "loss": 4.3277, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.32345013477088946, |
| "grad_norm": 0.7669051885604858, |
| "learning_rate": 0.0005812455477603885, |
| "loss": 4.3174, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.32345013477088946, |
| "eval_accuracy": 0.29804573656390243, |
| "eval_loss": 4.239420413970947, |
| "eval_runtime": 184.1932, |
| "eval_samples_per_second": 97.783, |
| "eval_steps_per_second": 6.113, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.3288409703504043, |
| "grad_norm": 0.6557288765907288, |
| "learning_rate": 0.0005809217485159201, |
| "loss": 4.3175, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.33423180592991913, |
| "grad_norm": 0.678659975528717, |
| "learning_rate": 0.0005805979492714517, |
| "loss": 4.2974, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.33962264150943394, |
| "grad_norm": 0.8476620316505432, |
| "learning_rate": 0.0005802741500269832, |
| "loss": 4.2748, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.3450134770889488, |
| "grad_norm": 0.6797805428504944, |
| "learning_rate": 0.0005799503507825148, |
| "loss": 4.2811, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.3504043126684636, |
| "grad_norm": 0.6718155145645142, |
| "learning_rate": 0.0005796265515380463, |
| "loss": 4.2701, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.3557951482479784, |
| "grad_norm": 0.9261127710342407, |
| "learning_rate": 0.000579302752293578, |
| "loss": 4.2727, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.3611859838274933, |
| "grad_norm": 0.8913077116012573, |
| "learning_rate": 0.0005789789530491095, |
| "loss": 4.2472, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.3665768194070081, |
| "grad_norm": 0.7431011199951172, |
| "learning_rate": 0.0005786551538046411, |
| "loss": 4.2378, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.3719676549865229, |
| "grad_norm": 0.6269802451133728, |
| "learning_rate": 0.0005783313545601726, |
| "loss": 4.2305, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.37735849056603776, |
| "grad_norm": 0.8824395537376404, |
| "learning_rate": 0.0005780075553157043, |
| "loss": 4.2307, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.38274932614555257, |
| "grad_norm": 0.6562435626983643, |
| "learning_rate": 0.0005776837560712357, |
| "loss": 4.2244, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.3881401617250674, |
| "grad_norm": 0.8994774222373962, |
| "learning_rate": 0.0005773599568267673, |
| "loss": 4.2167, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.3935309973045822, |
| "grad_norm": 0.7905921339988708, |
| "learning_rate": 0.0005770361575822989, |
| "loss": 4.2048, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.39892183288409705, |
| "grad_norm": 0.7321567535400391, |
| "learning_rate": 0.0005767123583378305, |
| "loss": 4.1956, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.40431266846361186, |
| "grad_norm": 0.697921872138977, |
| "learning_rate": 0.000576388559093362, |
| "loss": 4.1971, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.40970350404312667, |
| "grad_norm": 0.6758502125740051, |
| "learning_rate": 0.0005760647598488936, |
| "loss": 4.1617, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.41509433962264153, |
| "grad_norm": 0.7216224670410156, |
| "learning_rate": 0.0005757409606044253, |
| "loss": 4.1734, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.42048517520215634, |
| "grad_norm": 0.7215520143508911, |
| "learning_rate": 0.0005754171613599568, |
| "loss": 4.1892, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.42587601078167114, |
| "grad_norm": 1.9430748224258423, |
| "learning_rate": 0.0005750933621154884, |
| "loss": 4.1593, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.431266846361186, |
| "grad_norm": 0.6709734797477722, |
| "learning_rate": 0.0005747695628710199, |
| "loss": 4.1616, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.431266846361186, |
| "eval_accuracy": 0.31166461548123287, |
| "eval_loss": 4.095009803771973, |
| "eval_runtime": 184.2427, |
| "eval_samples_per_second": 97.757, |
| "eval_steps_per_second": 6.112, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.4366576819407008, |
| "grad_norm": 0.8147162795066833, |
| "learning_rate": 0.0005744457636265515, |
| "loss": 4.1673, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.4420485175202156, |
| "grad_norm": 0.6667271256446838, |
| "learning_rate": 0.0005741219643820831, |
| "loss": 4.1623, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.4474393530997305, |
| "grad_norm": 0.6467653512954712, |
| "learning_rate": 0.0005737981651376146, |
| "loss": 4.1408, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.4528301886792453, |
| "grad_norm": 0.6948724985122681, |
| "learning_rate": 0.0005734743658931462, |
| "loss": 4.1498, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.4582210242587601, |
| "grad_norm": 0.59138023853302, |
| "learning_rate": 0.0005731505666486778, |
| "loss": 4.1371, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.4636118598382749, |
| "grad_norm": 0.8744353652000427, |
| "learning_rate": 0.0005728267674042093, |
| "loss": 4.1414, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.46900269541778977, |
| "grad_norm": 0.7583370804786682, |
| "learning_rate": 0.0005725029681597409, |
| "loss": 4.1251, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.4743935309973046, |
| "grad_norm": 0.6753732562065125, |
| "learning_rate": 0.0005721791689152725, |
| "loss": 4.1139, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.4797843665768194, |
| "grad_norm": 0.7481412291526794, |
| "learning_rate": 0.0005718553696708041, |
| "loss": 4.1187, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.48517520215633425, |
| "grad_norm": 0.6845588088035583, |
| "learning_rate": 0.0005715315704263356, |
| "loss": 4.1127, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.49056603773584906, |
| "grad_norm": 0.8330910205841064, |
| "learning_rate": 0.0005712077711818672, |
| "loss": 4.1234, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.49595687331536387, |
| "grad_norm": 0.8156750202178955, |
| "learning_rate": 0.0005708839719373987, |
| "loss": 4.117, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.5013477088948787, |
| "grad_norm": 0.6442606449127197, |
| "learning_rate": 0.0005705601726929304, |
| "loss": 4.0742, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.5067385444743935, |
| "grad_norm": 0.7438291907310486, |
| "learning_rate": 0.0005702363734484619, |
| "loss": 4.0767, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.5121293800539084, |
| "grad_norm": 0.6073052883148193, |
| "learning_rate": 0.0005699125742039935, |
| "loss": 4.0814, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.5175202156334232, |
| "grad_norm": 0.7255820631980896, |
| "learning_rate": 0.000569588774959525, |
| "loss": 4.0777, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.522911051212938, |
| "grad_norm": 0.6645416617393494, |
| "learning_rate": 0.0005692649757150567, |
| "loss": 4.0733, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.5283018867924528, |
| "grad_norm": 0.7203142642974854, |
| "learning_rate": 0.0005689411764705881, |
| "loss": 4.0753, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.5336927223719676, |
| "grad_norm": 0.6812610626220703, |
| "learning_rate": 0.0005686173772261197, |
| "loss": 4.0645, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.5390835579514824, |
| "grad_norm": 0.7987631559371948, |
| "learning_rate": 0.0005682935779816514, |
| "loss": 4.0631, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5390835579514824, |
| "eval_accuracy": 0.32099018418946196, |
| "eval_loss": 3.9943337440490723, |
| "eval_runtime": 184.2999, |
| "eval_samples_per_second": 97.727, |
| "eval_steps_per_second": 6.11, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5444743935309974, |
| "grad_norm": 0.6587278246879578, |
| "learning_rate": 0.0005679697787371829, |
| "loss": 4.0704, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.5498652291105122, |
| "grad_norm": 0.6426385641098022, |
| "learning_rate": 0.0005676459794927145, |
| "loss": 4.0577, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.555256064690027, |
| "grad_norm": 0.7008543014526367, |
| "learning_rate": 0.000567322180248246, |
| "loss": 4.0488, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.5606469002695418, |
| "grad_norm": 0.6497129797935486, |
| "learning_rate": 0.0005669983810037777, |
| "loss": 4.0616, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.5660377358490566, |
| "grad_norm": 0.6053639650344849, |
| "learning_rate": 0.0005666745817593092, |
| "loss": 4.0349, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.5714285714285714, |
| "grad_norm": 0.5914347767829895, |
| "learning_rate": 0.0005663507825148408, |
| "loss": 4.035, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.5768194070080862, |
| "grad_norm": 0.6652584671974182, |
| "learning_rate": 0.0005660269832703723, |
| "loss": 4.0262, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.5822102425876011, |
| "grad_norm": 0.6383451223373413, |
| "learning_rate": 0.0005657031840259039, |
| "loss": 4.031, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.5876010781671159, |
| "grad_norm": 0.6813966035842896, |
| "learning_rate": 0.0005653793847814355, |
| "loss": 4.033, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.5929919137466307, |
| "grad_norm": 0.7794296145439148, |
| "learning_rate": 0.000565055585536967, |
| "loss": 4.0085, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.5983827493261455, |
| "grad_norm": 0.6553782820701599, |
| "learning_rate": 0.0005647317862924986, |
| "loss": 4.0318, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.6037735849056604, |
| "grad_norm": 0.6803719401359558, |
| "learning_rate": 0.0005644079870480302, |
| "loss": 3.9794, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.6091644204851752, |
| "grad_norm": 0.6251174211502075, |
| "learning_rate": 0.0005640841878035617, |
| "loss": 4.0158, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.6145552560646901, |
| "grad_norm": 0.7014779448509216, |
| "learning_rate": 0.0005637603885590933, |
| "loss": 4.0177, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.6199460916442049, |
| "grad_norm": 0.6140213012695312, |
| "learning_rate": 0.0005634365893146248, |
| "loss": 4.017, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.6253369272237197, |
| "grad_norm": 0.6971213221549988, |
| "learning_rate": 0.0005631127900701565, |
| "loss": 4.0065, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.6307277628032345, |
| "grad_norm": 0.6409980654716492, |
| "learning_rate": 0.000562788990825688, |
| "loss": 3.9797, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.6361185983827493, |
| "grad_norm": 0.6414069533348083, |
| "learning_rate": 0.0005624651915812196, |
| "loss": 3.9951, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.6415094339622641, |
| "grad_norm": 0.7267135381698608, |
| "learning_rate": 0.0005621413923367511, |
| "loss": 4.0005, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.6469002695417789, |
| "grad_norm": 0.5980563759803772, |
| "learning_rate": 0.0005618175930922828, |
| "loss": 4.0063, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.6469002695417789, |
| "eval_accuracy": 0.3277966356937123, |
| "eval_loss": 3.9200706481933594, |
| "eval_runtime": 184.1116, |
| "eval_samples_per_second": 97.827, |
| "eval_steps_per_second": 6.116, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.6522911051212938, |
| "grad_norm": 0.7149510979652405, |
| "learning_rate": 0.0005614937938478143, |
| "loss": 3.9946, |
| "step": 6050 |
| }, |
| { |
| "epoch": 0.6576819407008087, |
| "grad_norm": 0.691727340221405, |
| "learning_rate": 0.0005611699946033459, |
| "loss": 3.9905, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.6630727762803235, |
| "grad_norm": 0.5647642612457275, |
| "learning_rate": 0.0005608461953588774, |
| "loss": 3.9812, |
| "step": 6150 |
| }, |
| { |
| "epoch": 0.6684636118598383, |
| "grad_norm": 0.5823507308959961, |
| "learning_rate": 0.000560522396114409, |
| "loss": 3.9752, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.6738544474393531, |
| "grad_norm": 0.6531707048416138, |
| "learning_rate": 0.0005601985968699405, |
| "loss": 3.9557, |
| "step": 6250 |
| }, |
| { |
| "epoch": 0.6792452830188679, |
| "grad_norm": 0.6895961761474609, |
| "learning_rate": 0.0005598747976254721, |
| "loss": 3.9751, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.6846361185983828, |
| "grad_norm": 0.5271945595741272, |
| "learning_rate": 0.0005595509983810038, |
| "loss": 3.9466, |
| "step": 6350 |
| }, |
| { |
| "epoch": 0.6900269541778976, |
| "grad_norm": 0.5650665760040283, |
| "learning_rate": 0.0005592271991365353, |
| "loss": 3.9575, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.6954177897574124, |
| "grad_norm": 0.629801332950592, |
| "learning_rate": 0.0005589033998920669, |
| "loss": 3.9759, |
| "step": 6450 |
| }, |
| { |
| "epoch": 0.7008086253369272, |
| "grad_norm": 0.630355954170227, |
| "learning_rate": 0.0005585796006475984, |
| "loss": 3.9447, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.706199460916442, |
| "grad_norm": 0.6709086894989014, |
| "learning_rate": 0.0005582558014031301, |
| "loss": 3.9707, |
| "step": 6550 |
| }, |
| { |
| "epoch": 0.7115902964959568, |
| "grad_norm": 0.5377687811851501, |
| "learning_rate": 0.0005579320021586616, |
| "loss": 3.9539, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.7169811320754716, |
| "grad_norm": 0.7604367733001709, |
| "learning_rate": 0.0005576082029141932, |
| "loss": 3.9489, |
| "step": 6650 |
| }, |
| { |
| "epoch": 0.7223719676549866, |
| "grad_norm": 0.5272060632705688, |
| "learning_rate": 0.0005572844036697247, |
| "loss": 3.9361, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.7277628032345014, |
| "grad_norm": 0.5893604159355164, |
| "learning_rate": 0.0005569606044252563, |
| "loss": 3.9425, |
| "step": 6750 |
| }, |
| { |
| "epoch": 0.7331536388140162, |
| "grad_norm": 0.8032468557357788, |
| "learning_rate": 0.0005566368051807879, |
| "loss": 3.9585, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.738544474393531, |
| "grad_norm": 0.6482828855514526, |
| "learning_rate": 0.0005563130059363194, |
| "loss": 3.9433, |
| "step": 6850 |
| }, |
| { |
| "epoch": 0.7439353099730458, |
| "grad_norm": 0.532512903213501, |
| "learning_rate": 0.000555989206691851, |
| "loss": 3.9431, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.7493261455525606, |
| "grad_norm": 0.5721356868743896, |
| "learning_rate": 0.0005556654074473826, |
| "loss": 3.9487, |
| "step": 6950 |
| }, |
| { |
| "epoch": 0.7547169811320755, |
| "grad_norm": 0.7197826504707336, |
| "learning_rate": 0.0005553416082029141, |
| "loss": 3.9296, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7547169811320755, |
| "eval_accuracy": 0.3329282029323024, |
| "eval_loss": 3.8656320571899414, |
| "eval_runtime": 184.4615, |
| "eval_samples_per_second": 97.641, |
| "eval_steps_per_second": 6.104, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7601078167115903, |
| "grad_norm": 0.6058018803596497, |
| "learning_rate": 0.0005550178089584457, |
| "loss": 3.9224, |
| "step": 7050 |
| }, |
| { |
| "epoch": 0.7654986522911051, |
| "grad_norm": 0.5473665595054626, |
| "learning_rate": 0.0005546940097139772, |
| "loss": 3.9212, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.77088948787062, |
| "grad_norm": 0.6420671939849854, |
| "learning_rate": 0.0005543702104695089, |
| "loss": 3.9266, |
| "step": 7150 |
| }, |
| { |
| "epoch": 0.7762803234501348, |
| "grad_norm": 0.5873612761497498, |
| "learning_rate": 0.0005540464112250404, |
| "loss": 3.9235, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.7816711590296496, |
| "grad_norm": 0.6888593435287476, |
| "learning_rate": 0.000553722611980572, |
| "loss": 3.9344, |
| "step": 7250 |
| }, |
| { |
| "epoch": 0.7870619946091644, |
| "grad_norm": 0.5773025155067444, |
| "learning_rate": 0.0005533988127361035, |
| "loss": 3.9178, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.7924528301886793, |
| "grad_norm": 0.5834557414054871, |
| "learning_rate": 0.0005530750134916352, |
| "loss": 3.9313, |
| "step": 7350 |
| }, |
| { |
| "epoch": 0.7978436657681941, |
| "grad_norm": 0.7065815329551697, |
| "learning_rate": 0.0005527512142471668, |
| "loss": 3.9078, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.8032345013477089, |
| "grad_norm": 0.5983999371528625, |
| "learning_rate": 0.0005524274150026982, |
| "loss": 3.9125, |
| "step": 7450 |
| }, |
| { |
| "epoch": 0.8086253369272237, |
| "grad_norm": 0.6277061104774475, |
| "learning_rate": 0.0005521036157582299, |
| "loss": 3.9004, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.8140161725067385, |
| "grad_norm": 0.7020032405853271, |
| "learning_rate": 0.0005517798165137614, |
| "loss": 3.9029, |
| "step": 7550 |
| }, |
| { |
| "epoch": 0.8194070080862533, |
| "grad_norm": 0.6681187152862549, |
| "learning_rate": 0.000551456017269293, |
| "loss": 3.8801, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.8247978436657682, |
| "grad_norm": 0.6278699636459351, |
| "learning_rate": 0.0005511322180248245, |
| "loss": 3.8967, |
| "step": 7650 |
| }, |
| { |
| "epoch": 0.8301886792452831, |
| "grad_norm": 0.6532415747642517, |
| "learning_rate": 0.0005508084187803562, |
| "loss": 3.9083, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.8355795148247979, |
| "grad_norm": 0.5810382962226868, |
| "learning_rate": 0.0005504846195358877, |
| "loss": 3.903, |
| "step": 7750 |
| }, |
| { |
| "epoch": 0.8409703504043127, |
| "grad_norm": 0.6141226291656494, |
| "learning_rate": 0.0005501608202914193, |
| "loss": 3.8898, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.8463611859838275, |
| "grad_norm": 0.6052567362785339, |
| "learning_rate": 0.0005498370210469508, |
| "loss": 3.8946, |
| "step": 7850 |
| }, |
| { |
| "epoch": 0.8517520215633423, |
| "grad_norm": 0.619564950466156, |
| "learning_rate": 0.0005495132218024824, |
| "loss": 3.8788, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.8571428571428571, |
| "grad_norm": 0.620516836643219, |
| "learning_rate": 0.000549189422558014, |
| "loss": 3.8888, |
| "step": 7950 |
| }, |
| { |
| "epoch": 0.862533692722372, |
| "grad_norm": 0.5579094290733337, |
| "learning_rate": 0.0005488656233135456, |
| "loss": 3.8882, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.862533692722372, |
| "eval_accuracy": 0.3372453081238352, |
| "eval_loss": 3.8190271854400635, |
| "eval_runtime": 183.5032, |
| "eval_samples_per_second": 98.151, |
| "eval_steps_per_second": 6.136, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.8679245283018868, |
| "grad_norm": 0.5879581570625305, |
| "learning_rate": 0.0005485418240690771, |
| "loss": 3.8678, |
| "step": 8050 |
| }, |
| { |
| "epoch": 0.8733153638814016, |
| "grad_norm": 0.6095612049102783, |
| "learning_rate": 0.0005482180248246087, |
| "loss": 3.8844, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.8787061994609164, |
| "grad_norm": 0.5836382508277893, |
| "learning_rate": 0.0005478942255801403, |
| "loss": 3.8812, |
| "step": 8150 |
| }, |
| { |
| "epoch": 0.8840970350404312, |
| "grad_norm": 0.5505040884017944, |
| "learning_rate": 0.0005475704263356718, |
| "loss": 3.8725, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.889487870619946, |
| "grad_norm": 0.5962595343589783, |
| "learning_rate": 0.0005472466270912034, |
| "loss": 3.8804, |
| "step": 8250 |
| }, |
| { |
| "epoch": 0.894878706199461, |
| "grad_norm": 0.6875723600387573, |
| "learning_rate": 0.000546922827846735, |
| "loss": 3.8692, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.9002695417789758, |
| "grad_norm": 0.5387910008430481, |
| "learning_rate": 0.0005465990286022665, |
| "loss": 3.8689, |
| "step": 8350 |
| }, |
| { |
| "epoch": 0.9056603773584906, |
| "grad_norm": 0.6388599276542664, |
| "learning_rate": 0.0005462752293577981, |
| "loss": 3.8823, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.9110512129380054, |
| "grad_norm": 0.559060275554657, |
| "learning_rate": 0.0005459514301133296, |
| "loss": 3.8783, |
| "step": 8450 |
| }, |
| { |
| "epoch": 0.9164420485175202, |
| "grad_norm": 0.5374552011489868, |
| "learning_rate": 0.0005456276308688613, |
| "loss": 3.8729, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.921832884097035, |
| "grad_norm": 0.5611122250556946, |
| "learning_rate": 0.0005453038316243929, |
| "loss": 3.8668, |
| "step": 8550 |
| }, |
| { |
| "epoch": 0.9272237196765498, |
| "grad_norm": 0.5464676022529602, |
| "learning_rate": 0.0005449800323799244, |
| "loss": 3.8786, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.9326145552560647, |
| "grad_norm": 0.6095228791236877, |
| "learning_rate": 0.000544656233135456, |
| "loss": 3.8565, |
| "step": 8650 |
| }, |
| { |
| "epoch": 0.9380053908355795, |
| "grad_norm": 0.5888999104499817, |
| "learning_rate": 0.0005443324338909875, |
| "loss": 3.8615, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.9433962264150944, |
| "grad_norm": 0.5549721121788025, |
| "learning_rate": 0.0005440086346465192, |
| "loss": 3.8709, |
| "step": 8750 |
| }, |
| { |
| "epoch": 0.9487870619946092, |
| "grad_norm": 0.650879979133606, |
| "learning_rate": 0.0005436848354020506, |
| "loss": 3.8519, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.954177897574124, |
| "grad_norm": 0.5871933102607727, |
| "learning_rate": 0.0005433610361575823, |
| "loss": 3.8592, |
| "step": 8850 |
| }, |
| { |
| "epoch": 0.9595687331536388, |
| "grad_norm": 0.5539306402206421, |
| "learning_rate": 0.0005430372369131138, |
| "loss": 3.8421, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.9649595687331537, |
| "grad_norm": 0.5857457518577576, |
| "learning_rate": 0.0005427134376686454, |
| "loss": 3.8551, |
| "step": 8950 |
| }, |
| { |
| "epoch": 0.9703504043126685, |
| "grad_norm": 0.5824806690216064, |
| "learning_rate": 0.0005423896384241769, |
| "loss": 3.8432, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.9703504043126685, |
| "eval_accuracy": 0.3406982968985794, |
| "eval_loss": 3.780831813812256, |
| "eval_runtime": 183.3642, |
| "eval_samples_per_second": 98.225, |
| "eval_steps_per_second": 6.141, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.9757412398921833, |
| "grad_norm": 0.8111798167228699, |
| "learning_rate": 0.0005420658391797086, |
| "loss": 3.8374, |
| "step": 9050 |
| }, |
| { |
| "epoch": 0.9811320754716981, |
| "grad_norm": 0.5787947177886963, |
| "learning_rate": 0.0005417420399352401, |
| "loss": 3.8494, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.9865229110512129, |
| "grad_norm": 0.6467134356498718, |
| "learning_rate": 0.0005414182406907717, |
| "loss": 3.8425, |
| "step": 9150 |
| }, |
| { |
| "epoch": 0.9919137466307277, |
| "grad_norm": 0.6325974464416504, |
| "learning_rate": 0.0005410944414463032, |
| "loss": 3.8555, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.9973045822102425, |
| "grad_norm": 0.5801452994346619, |
| "learning_rate": 0.0005407706422018348, |
| "loss": 3.854, |
| "step": 9250 |
| }, |
| { |
| "epoch": 1.0026954177897573, |
| "grad_norm": 0.621691107749939, |
| "learning_rate": 0.0005404468429573664, |
| "loss": 3.8021, |
| "step": 9300 |
| }, |
| { |
| "epoch": 1.0080862533692723, |
| "grad_norm": 0.6182531118392944, |
| "learning_rate": 0.000540123043712898, |
| "loss": 3.771, |
| "step": 9350 |
| }, |
| { |
| "epoch": 1.013477088948787, |
| "grad_norm": 0.5976972579956055, |
| "learning_rate": 0.0005397992444684295, |
| "loss": 3.7536, |
| "step": 9400 |
| }, |
| { |
| "epoch": 1.0188679245283019, |
| "grad_norm": 0.6204501390457153, |
| "learning_rate": 0.0005394754452239611, |
| "loss": 3.7671, |
| "step": 9450 |
| }, |
| { |
| "epoch": 1.0242587601078168, |
| "grad_norm": 0.532163679599762, |
| "learning_rate": 0.0005391516459794927, |
| "loss": 3.7787, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.0296495956873315, |
| "grad_norm": 0.5977817177772522, |
| "learning_rate": 0.0005388278467350242, |
| "loss": 3.7871, |
| "step": 9550 |
| }, |
| { |
| "epoch": 1.0350404312668464, |
| "grad_norm": 0.5936277508735657, |
| "learning_rate": 0.0005385040474905557, |
| "loss": 3.7799, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.0404312668463611, |
| "grad_norm": 0.5855807662010193, |
| "learning_rate": 0.0005381802482460874, |
| "loss": 3.7879, |
| "step": 9650 |
| }, |
| { |
| "epoch": 1.045822102425876, |
| "grad_norm": 0.531672477722168, |
| "learning_rate": 0.000537856449001619, |
| "loss": 3.7905, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.0512129380053907, |
| "grad_norm": 0.6275017261505127, |
| "learning_rate": 0.0005375326497571505, |
| "loss": 3.761, |
| "step": 9750 |
| }, |
| { |
| "epoch": 1.0566037735849056, |
| "grad_norm": 0.6375147104263306, |
| "learning_rate": 0.000537208850512682, |
| "loss": 3.7682, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.0619946091644206, |
| "grad_norm": 0.5759090781211853, |
| "learning_rate": 0.0005368850512682137, |
| "loss": 3.7676, |
| "step": 9850 |
| }, |
| { |
| "epoch": 1.0673854447439353, |
| "grad_norm": 0.5603178143501282, |
| "learning_rate": 0.0005365612520237453, |
| "loss": 3.779, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.0727762803234502, |
| "grad_norm": 0.5935394763946533, |
| "learning_rate": 0.0005362374527792768, |
| "loss": 3.7782, |
| "step": 9950 |
| }, |
| { |
| "epoch": 1.0781671159029649, |
| "grad_norm": 0.5596011281013489, |
| "learning_rate": 0.0005359136535348084, |
| "loss": 3.7782, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.0781671159029649, |
| "eval_accuracy": 0.3441405290374299, |
| "eval_loss": 3.7568960189819336, |
| "eval_runtime": 183.0624, |
| "eval_samples_per_second": 98.387, |
| "eval_steps_per_second": 6.151, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.0835579514824798, |
| "grad_norm": 0.5355740785598755, |
| "learning_rate": 0.0005355898542903399, |
| "loss": 3.7564, |
| "step": 10050 |
| }, |
| { |
| "epoch": 1.0889487870619945, |
| "grad_norm": 0.5919591188430786, |
| "learning_rate": 0.0005352660550458716, |
| "loss": 3.76, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.0943396226415094, |
| "grad_norm": 0.6239626407623291, |
| "learning_rate": 0.000534942255801403, |
| "loss": 3.7471, |
| "step": 10150 |
| }, |
| { |
| "epoch": 1.0997304582210243, |
| "grad_norm": 0.6291561722755432, |
| "learning_rate": 0.0005346184565569347, |
| "loss": 3.7655, |
| "step": 10200 |
| }, |
| { |
| "epoch": 1.105121293800539, |
| "grad_norm": 0.5858433842658997, |
| "learning_rate": 0.0005342946573124662, |
| "loss": 3.7832, |
| "step": 10250 |
| }, |
| { |
| "epoch": 1.110512129380054, |
| "grad_norm": 0.636974573135376, |
| "learning_rate": 0.0005339708580679978, |
| "loss": 3.7845, |
| "step": 10300 |
| }, |
| { |
| "epoch": 1.1159029649595686, |
| "grad_norm": 0.5367472171783447, |
| "learning_rate": 0.0005336470588235293, |
| "loss": 3.76, |
| "step": 10350 |
| }, |
| { |
| "epoch": 1.1212938005390836, |
| "grad_norm": 0.5671682357788086, |
| "learning_rate": 0.000533323259579061, |
| "loss": 3.7657, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.1266846361185983, |
| "grad_norm": 0.5502925515174866, |
| "learning_rate": 0.0005329994603345925, |
| "loss": 3.7594, |
| "step": 10450 |
| }, |
| { |
| "epoch": 1.1320754716981132, |
| "grad_norm": 0.5960683822631836, |
| "learning_rate": 0.0005326756610901241, |
| "loss": 3.7594, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.137466307277628, |
| "grad_norm": 0.5544126033782959, |
| "learning_rate": 0.0005323518618456556, |
| "loss": 3.7582, |
| "step": 10550 |
| }, |
| { |
| "epoch": 1.1428571428571428, |
| "grad_norm": 0.6969329714775085, |
| "learning_rate": 0.0005320280626011872, |
| "loss": 3.7705, |
| "step": 10600 |
| }, |
| { |
| "epoch": 1.1482479784366577, |
| "grad_norm": 0.6323273181915283, |
| "learning_rate": 0.0005317042633567188, |
| "loss": 3.7581, |
| "step": 10650 |
| }, |
| { |
| "epoch": 1.1536388140161726, |
| "grad_norm": 0.6453703045845032, |
| "learning_rate": 0.0005313804641122504, |
| "loss": 3.763, |
| "step": 10700 |
| }, |
| { |
| "epoch": 1.1590296495956873, |
| "grad_norm": 0.5417402386665344, |
| "learning_rate": 0.0005310566648677819, |
| "loss": 3.7579, |
| "step": 10750 |
| }, |
| { |
| "epoch": 1.1644204851752022, |
| "grad_norm": 0.5967243313789368, |
| "learning_rate": 0.0005307328656233135, |
| "loss": 3.7471, |
| "step": 10800 |
| }, |
| { |
| "epoch": 1.169811320754717, |
| "grad_norm": 0.5780621767044067, |
| "learning_rate": 0.000530409066378845, |
| "loss": 3.7432, |
| "step": 10850 |
| }, |
| { |
| "epoch": 1.1752021563342319, |
| "grad_norm": 1.0681740045547485, |
| "learning_rate": 0.0005300852671343766, |
| "loss": 3.759, |
| "step": 10900 |
| }, |
| { |
| "epoch": 1.1805929919137466, |
| "grad_norm": 0.6521495580673218, |
| "learning_rate": 0.0005297614678899081, |
| "loss": 3.7513, |
| "step": 10950 |
| }, |
| { |
| "epoch": 1.1859838274932615, |
| "grad_norm": 0.5670123100280762, |
| "learning_rate": 0.0005294376686454398, |
| "loss": 3.7405, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.1859838274932615, |
| "eval_accuracy": 0.347109251891185, |
| "eval_loss": 3.7269060611724854, |
| "eval_runtime": 183.025, |
| "eval_samples_per_second": 98.407, |
| "eval_steps_per_second": 6.152, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.1913746630727764, |
| "grad_norm": 0.5789473056793213, |
| "learning_rate": 0.0005291138694009714, |
| "loss": 3.7509, |
| "step": 11050 |
| }, |
| { |
| "epoch": 1.196765498652291, |
| "grad_norm": 0.6155396699905396, |
| "learning_rate": 0.0005287900701565029, |
| "loss": 3.7419, |
| "step": 11100 |
| }, |
| { |
| "epoch": 1.202156334231806, |
| "grad_norm": 0.6248864531517029, |
| "learning_rate": 0.0005284662709120345, |
| "loss": 3.7432, |
| "step": 11150 |
| }, |
| { |
| "epoch": 1.2075471698113207, |
| "grad_norm": 0.6895485520362854, |
| "learning_rate": 0.0005281424716675661, |
| "loss": 3.7428, |
| "step": 11200 |
| }, |
| { |
| "epoch": 1.2129380053908356, |
| "grad_norm": 0.6388883590698242, |
| "learning_rate": 0.0005278186724230977, |
| "loss": 3.7363, |
| "step": 11250 |
| }, |
| { |
| "epoch": 1.2183288409703503, |
| "grad_norm": 0.5322048664093018, |
| "learning_rate": 0.0005274948731786292, |
| "loss": 3.7326, |
| "step": 11300 |
| }, |
| { |
| "epoch": 1.2237196765498652, |
| "grad_norm": 0.6221327781677246, |
| "learning_rate": 0.0005271710739341608, |
| "loss": 3.7505, |
| "step": 11350 |
| }, |
| { |
| "epoch": 1.2291105121293802, |
| "grad_norm": 0.6232865452766418, |
| "learning_rate": 0.0005268472746896923, |
| "loss": 3.7446, |
| "step": 11400 |
| }, |
| { |
| "epoch": 1.2345013477088949, |
| "grad_norm": 0.6101617813110352, |
| "learning_rate": 0.000526523475445224, |
| "loss": 3.7366, |
| "step": 11450 |
| }, |
| { |
| "epoch": 1.2398921832884098, |
| "grad_norm": 0.5458541512489319, |
| "learning_rate": 0.0005261996762007554, |
| "loss": 3.7232, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.2452830188679245, |
| "grad_norm": 0.5346110463142395, |
| "learning_rate": 0.0005258758769562871, |
| "loss": 3.7439, |
| "step": 11550 |
| }, |
| { |
| "epoch": 1.2506738544474394, |
| "grad_norm": 0.5948981046676636, |
| "learning_rate": 0.0005255520777118186, |
| "loss": 3.7435, |
| "step": 11600 |
| }, |
| { |
| "epoch": 1.256064690026954, |
| "grad_norm": 0.6083604693412781, |
| "learning_rate": 0.0005252282784673502, |
| "loss": 3.7356, |
| "step": 11650 |
| }, |
| { |
| "epoch": 1.261455525606469, |
| "grad_norm": 0.6157535314559937, |
| "learning_rate": 0.0005249044792228817, |
| "loss": 3.7373, |
| "step": 11700 |
| }, |
| { |
| "epoch": 1.266846361185984, |
| "grad_norm": 0.5235822200775146, |
| "learning_rate": 0.0005245806799784133, |
| "loss": 3.7324, |
| "step": 11750 |
| }, |
| { |
| "epoch": 1.2722371967654986, |
| "grad_norm": 0.6096104383468628, |
| "learning_rate": 0.0005242568807339449, |
| "loss": 3.7291, |
| "step": 11800 |
| }, |
| { |
| "epoch": 1.2776280323450135, |
| "grad_norm": 0.6221447587013245, |
| "learning_rate": 0.0005239330814894765, |
| "loss": 3.7236, |
| "step": 11850 |
| }, |
| { |
| "epoch": 1.2830188679245282, |
| "grad_norm": 0.5434000492095947, |
| "learning_rate": 0.000523609282245008, |
| "loss": 3.7454, |
| "step": 11900 |
| }, |
| { |
| "epoch": 1.2884097035040432, |
| "grad_norm": 0.6391717195510864, |
| "learning_rate": 0.0005232854830005396, |
| "loss": 3.741, |
| "step": 11950 |
| }, |
| { |
| "epoch": 1.2938005390835579, |
| "grad_norm": 0.5883181691169739, |
| "learning_rate": 0.0005229616837560712, |
| "loss": 3.7342, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.2938005390835579, |
| "eval_accuracy": 0.3493015412086178, |
| "eval_loss": 3.702024221420288, |
| "eval_runtime": 183.3785, |
| "eval_samples_per_second": 98.218, |
| "eval_steps_per_second": 6.14, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.2991913746630728, |
| "grad_norm": 0.5594079494476318, |
| "learning_rate": 0.0005226378845116028, |
| "loss": 3.7174, |
| "step": 12050 |
| }, |
| { |
| "epoch": 1.3045822102425877, |
| "grad_norm": 0.6001787781715393, |
| "learning_rate": 0.0005223140852671344, |
| "loss": 3.733, |
| "step": 12100 |
| }, |
| { |
| "epoch": 1.3099730458221024, |
| "grad_norm": 0.623263955116272, |
| "learning_rate": 0.0005219902860226659, |
| "loss": 3.7411, |
| "step": 12150 |
| }, |
| { |
| "epoch": 1.3153638814016173, |
| "grad_norm": 0.6152437329292297, |
| "learning_rate": 0.0005216664867781975, |
| "loss": 3.7405, |
| "step": 12200 |
| }, |
| { |
| "epoch": 1.320754716981132, |
| "grad_norm": 0.6849799752235413, |
| "learning_rate": 0.000521342687533729, |
| "loss": 3.7211, |
| "step": 12250 |
| }, |
| { |
| "epoch": 1.326145552560647, |
| "grad_norm": 0.5973236560821533, |
| "learning_rate": 0.0005210188882892606, |
| "loss": 3.7314, |
| "step": 12300 |
| }, |
| { |
| "epoch": 1.3315363881401616, |
| "grad_norm": 0.5735817551612854, |
| "learning_rate": 0.0005206950890447922, |
| "loss": 3.7194, |
| "step": 12350 |
| }, |
| { |
| "epoch": 1.3369272237196765, |
| "grad_norm": 0.5044131278991699, |
| "learning_rate": 0.0005203712898003238, |
| "loss": 3.7348, |
| "step": 12400 |
| }, |
| { |
| "epoch": 1.3423180592991915, |
| "grad_norm": 0.6093466281890869, |
| "learning_rate": 0.0005200474905558553, |
| "loss": 3.7264, |
| "step": 12450 |
| }, |
| { |
| "epoch": 1.3477088948787062, |
| "grad_norm": 0.7663679122924805, |
| "learning_rate": 0.0005197236913113869, |
| "loss": 3.7262, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.353099730458221, |
| "grad_norm": 0.5749175548553467, |
| "learning_rate": 0.0005193998920669184, |
| "loss": 3.7343, |
| "step": 12550 |
| }, |
| { |
| "epoch": 1.3584905660377358, |
| "grad_norm": 0.6117640733718872, |
| "learning_rate": 0.0005190760928224501, |
| "loss": 3.7167, |
| "step": 12600 |
| }, |
| { |
| "epoch": 1.3638814016172507, |
| "grad_norm": 0.5556171536445618, |
| "learning_rate": 0.0005187522935779816, |
| "loss": 3.7244, |
| "step": 12650 |
| }, |
| { |
| "epoch": 1.3692722371967654, |
| "grad_norm": 0.59083092212677, |
| "learning_rate": 0.0005184284943335132, |
| "loss": 3.7366, |
| "step": 12700 |
| }, |
| { |
| "epoch": 1.3746630727762803, |
| "grad_norm": 0.5685399174690247, |
| "learning_rate": 0.0005181046950890447, |
| "loss": 3.7232, |
| "step": 12750 |
| }, |
| { |
| "epoch": 1.3800539083557952, |
| "grad_norm": 0.5734025239944458, |
| "learning_rate": 0.0005177808958445764, |
| "loss": 3.7146, |
| "step": 12800 |
| }, |
| { |
| "epoch": 1.38544474393531, |
| "grad_norm": 0.613262414932251, |
| "learning_rate": 0.0005174570966001078, |
| "loss": 3.7124, |
| "step": 12850 |
| }, |
| { |
| "epoch": 1.3908355795148248, |
| "grad_norm": 0.6126392483711243, |
| "learning_rate": 0.0005171332973556395, |
| "loss": 3.7171, |
| "step": 12900 |
| }, |
| { |
| "epoch": 1.3962264150943398, |
| "grad_norm": 0.5627440810203552, |
| "learning_rate": 0.000516809498111171, |
| "loss": 3.7154, |
| "step": 12950 |
| }, |
| { |
| "epoch": 1.4016172506738545, |
| "grad_norm": 0.6159223318099976, |
| "learning_rate": 0.0005164856988667026, |
| "loss": 3.7098, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.4016172506738545, |
| "eval_accuracy": 0.3513598615153753, |
| "eval_loss": 3.6815733909606934, |
| "eval_runtime": 182.9927, |
| "eval_samples_per_second": 98.425, |
| "eval_steps_per_second": 6.153, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.4070080862533692, |
| "grad_norm": 0.5440373420715332, |
| "learning_rate": 0.0005161618996222341, |
| "loss": 3.7223, |
| "step": 13050 |
| }, |
| { |
| "epoch": 1.412398921832884, |
| "grad_norm": 0.6110854148864746, |
| "learning_rate": 0.0005158381003777657, |
| "loss": 3.7332, |
| "step": 13100 |
| }, |
| { |
| "epoch": 1.417789757412399, |
| "grad_norm": 0.634506106376648, |
| "learning_rate": 0.0005155143011332973, |
| "loss": 3.714, |
| "step": 13150 |
| }, |
| { |
| "epoch": 1.4231805929919137, |
| "grad_norm": 0.5525026321411133, |
| "learning_rate": 0.0005151905018888289, |
| "loss": 3.7169, |
| "step": 13200 |
| }, |
| { |
| "epoch": 1.4285714285714286, |
| "grad_norm": 0.5637493133544922, |
| "learning_rate": 0.0005148731786292498, |
| "loss": 3.712, |
| "step": 13250 |
| }, |
| { |
| "epoch": 1.4339622641509435, |
| "grad_norm": 0.5339338779449463, |
| "learning_rate": 0.0005145493793847814, |
| "loss": 3.7101, |
| "step": 13300 |
| }, |
| { |
| "epoch": 1.4393530997304582, |
| "grad_norm": 0.613170862197876, |
| "learning_rate": 0.000514225580140313, |
| "loss": 3.6965, |
| "step": 13350 |
| }, |
| { |
| "epoch": 1.444743935309973, |
| "grad_norm": 0.589502215385437, |
| "learning_rate": 0.0005139017808958445, |
| "loss": 3.7058, |
| "step": 13400 |
| }, |
| { |
| "epoch": 1.4501347708894878, |
| "grad_norm": 0.5917065143585205, |
| "learning_rate": 0.0005135779816513762, |
| "loss": 3.7144, |
| "step": 13450 |
| }, |
| { |
| "epoch": 1.4555256064690028, |
| "grad_norm": 0.5540453195571899, |
| "learning_rate": 0.0005132541824069076, |
| "loss": 3.7001, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.4609164420485174, |
| "grad_norm": 0.5671173930168152, |
| "learning_rate": 0.0005129303831624393, |
| "loss": 3.7114, |
| "step": 13550 |
| }, |
| { |
| "epoch": 1.4663072776280324, |
| "grad_norm": 0.5867645740509033, |
| "learning_rate": 0.0005126065839179708, |
| "loss": 3.715, |
| "step": 13600 |
| }, |
| { |
| "epoch": 1.4716981132075473, |
| "grad_norm": 0.5680153369903564, |
| "learning_rate": 0.0005122827846735024, |
| "loss": 3.6852, |
| "step": 13650 |
| }, |
| { |
| "epoch": 1.477088948787062, |
| "grad_norm": 0.5795393586158752, |
| "learning_rate": 0.0005119589854290339, |
| "loss": 3.6843, |
| "step": 13700 |
| }, |
| { |
| "epoch": 1.482479784366577, |
| "grad_norm": 0.5783411860466003, |
| "learning_rate": 0.0005116351861845655, |
| "loss": 3.7033, |
| "step": 13750 |
| }, |
| { |
| "epoch": 1.4878706199460916, |
| "grad_norm": 0.6578272581100464, |
| "learning_rate": 0.0005113113869400971, |
| "loss": 3.7184, |
| "step": 13800 |
| }, |
| { |
| "epoch": 1.4932614555256065, |
| "grad_norm": 0.5720994472503662, |
| "learning_rate": 0.0005109875876956287, |
| "loss": 3.711, |
| "step": 13850 |
| }, |
| { |
| "epoch": 1.4986522911051212, |
| "grad_norm": 0.5809938311576843, |
| "learning_rate": 0.0005106637884511602, |
| "loss": 3.7267, |
| "step": 13900 |
| }, |
| { |
| "epoch": 1.5040431266846361, |
| "grad_norm": 0.5434926748275757, |
| "learning_rate": 0.0005103399892066918, |
| "loss": 3.7006, |
| "step": 13950 |
| }, |
| { |
| "epoch": 1.509433962264151, |
| "grad_norm": 0.6142950057983398, |
| "learning_rate": 0.0005100161899622234, |
| "loss": 3.7016, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.509433962264151, |
| "eval_accuracy": 0.35360506478917375, |
| "eval_loss": 3.658677816390991, |
| "eval_runtime": 183.8031, |
| "eval_samples_per_second": 97.991, |
| "eval_steps_per_second": 6.126, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.5148247978436657, |
| "grad_norm": 0.5697357654571533, |
| "learning_rate": 0.000509692390717755, |
| "loss": 3.6912, |
| "step": 14050 |
| }, |
| { |
| "epoch": 1.5202156334231804, |
| "grad_norm": 0.7325167059898376, |
| "learning_rate": 0.0005093685914732865, |
| "loss": 3.6894, |
| "step": 14100 |
| }, |
| { |
| "epoch": 1.5256064690026954, |
| "grad_norm": 0.5748705863952637, |
| "learning_rate": 0.0005090447922288181, |
| "loss": 3.6853, |
| "step": 14150 |
| }, |
| { |
| "epoch": 1.5309973045822103, |
| "grad_norm": 0.626307487487793, |
| "learning_rate": 0.0005087209929843496, |
| "loss": 3.7191, |
| "step": 14200 |
| }, |
| { |
| "epoch": 1.536388140161725, |
| "grad_norm": 0.6443688273429871, |
| "learning_rate": 0.0005083971937398812, |
| "loss": 3.6925, |
| "step": 14250 |
| }, |
| { |
| "epoch": 1.54177897574124, |
| "grad_norm": 0.6409396529197693, |
| "learning_rate": 0.0005080733944954127, |
| "loss": 3.6974, |
| "step": 14300 |
| }, |
| { |
| "epoch": 1.5471698113207548, |
| "grad_norm": 0.5955445766448975, |
| "learning_rate": 0.0005077495952509444, |
| "loss": 3.6789, |
| "step": 14350 |
| }, |
| { |
| "epoch": 1.5525606469002695, |
| "grad_norm": 0.5558004379272461, |
| "learning_rate": 0.0005074257960064759, |
| "loss": 3.6912, |
| "step": 14400 |
| }, |
| { |
| "epoch": 1.5579514824797842, |
| "grad_norm": 0.589004635810852, |
| "learning_rate": 0.0005071019967620075, |
| "loss": 3.6969, |
| "step": 14450 |
| }, |
| { |
| "epoch": 1.5633423180592994, |
| "grad_norm": 0.6211824417114258, |
| "learning_rate": 0.000506778197517539, |
| "loss": 3.6912, |
| "step": 14500 |
| }, |
| { |
| "epoch": 1.568733153638814, |
| "grad_norm": 0.5962303280830383, |
| "learning_rate": 0.0005064543982730707, |
| "loss": 3.6995, |
| "step": 14550 |
| }, |
| { |
| "epoch": 1.5741239892183287, |
| "grad_norm": 0.6277245283126831, |
| "learning_rate": 0.0005061305990286023, |
| "loss": 3.6931, |
| "step": 14600 |
| }, |
| { |
| "epoch": 1.5795148247978437, |
| "grad_norm": 0.6255839467048645, |
| "learning_rate": 0.0005058067997841338, |
| "loss": 3.7005, |
| "step": 14650 |
| }, |
| { |
| "epoch": 1.5849056603773586, |
| "grad_norm": 0.5719764828681946, |
| "learning_rate": 0.0005054830005396654, |
| "loss": 3.6871, |
| "step": 14700 |
| }, |
| { |
| "epoch": 1.5902964959568733, |
| "grad_norm": 0.6281840801239014, |
| "learning_rate": 0.0005051592012951969, |
| "loss": 3.6767, |
| "step": 14750 |
| }, |
| { |
| "epoch": 1.595687331536388, |
| "grad_norm": 0.5663508772850037, |
| "learning_rate": 0.0005048354020507286, |
| "loss": 3.6987, |
| "step": 14800 |
| }, |
| { |
| "epoch": 1.6010781671159031, |
| "grad_norm": 0.5636190176010132, |
| "learning_rate": 0.00050451160280626, |
| "loss": 3.7, |
| "step": 14850 |
| }, |
| { |
| "epoch": 1.6064690026954178, |
| "grad_norm": 0.5662776827812195, |
| "learning_rate": 0.0005041878035617917, |
| "loss": 3.6897, |
| "step": 14900 |
| }, |
| { |
| "epoch": 1.6118598382749325, |
| "grad_norm": 0.5471853613853455, |
| "learning_rate": 0.0005038640043173232, |
| "loss": 3.6842, |
| "step": 14950 |
| }, |
| { |
| "epoch": 1.6172506738544474, |
| "grad_norm": 0.704903781414032, |
| "learning_rate": 0.0005035402050728548, |
| "loss": 3.6781, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.6172506738544474, |
| "eval_accuracy": 0.3554610734188207, |
| "eval_loss": 3.6408092975616455, |
| "eval_runtime": 183.2026, |
| "eval_samples_per_second": 98.312, |
| "eval_steps_per_second": 6.146, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.6226415094339623, |
| "grad_norm": 0.5391875505447388, |
| "learning_rate": 0.0005032164058283863, |
| "loss": 3.6821, |
| "step": 15050 |
| }, |
| { |
| "epoch": 1.628032345013477, |
| "grad_norm": 0.5684933662414551, |
| "learning_rate": 0.0005028926065839179, |
| "loss": 3.6782, |
| "step": 15100 |
| }, |
| { |
| "epoch": 1.633423180592992, |
| "grad_norm": 0.5583056807518005, |
| "learning_rate": 0.0005025688073394495, |
| "loss": 3.6809, |
| "step": 15150 |
| }, |
| { |
| "epoch": 1.6388140161725069, |
| "grad_norm": 0.5500176548957825, |
| "learning_rate": 0.0005022450080949811, |
| "loss": 3.6846, |
| "step": 15200 |
| }, |
| { |
| "epoch": 1.6442048517520216, |
| "grad_norm": 0.6347390413284302, |
| "learning_rate": 0.0005019212088505126, |
| "loss": 3.6879, |
| "step": 15250 |
| }, |
| { |
| "epoch": 1.6495956873315363, |
| "grad_norm": 0.5185267925262451, |
| "learning_rate": 0.0005015974096060442, |
| "loss": 3.6893, |
| "step": 15300 |
| }, |
| { |
| "epoch": 1.6549865229110512, |
| "grad_norm": 0.6127241253852844, |
| "learning_rate": 0.0005012736103615758, |
| "loss": 3.6717, |
| "step": 15350 |
| }, |
| { |
| "epoch": 1.6603773584905661, |
| "grad_norm": 0.6135085821151733, |
| "learning_rate": 0.0005009498111171074, |
| "loss": 3.6772, |
| "step": 15400 |
| }, |
| { |
| "epoch": 1.6657681940700808, |
| "grad_norm": 0.5636995434761047, |
| "learning_rate": 0.0005006260118726389, |
| "loss": 3.6764, |
| "step": 15450 |
| }, |
| { |
| "epoch": 1.6711590296495957, |
| "grad_norm": 0.5827999114990234, |
| "learning_rate": 0.0005003022126281705, |
| "loss": 3.6785, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.6765498652291106, |
| "grad_norm": 0.7018887400627136, |
| "learning_rate": 0.000499978413383702, |
| "loss": 3.6721, |
| "step": 15550 |
| }, |
| { |
| "epoch": 1.6819407008086253, |
| "grad_norm": 0.564325749874115, |
| "learning_rate": 0.0004996546141392336, |
| "loss": 3.6671, |
| "step": 15600 |
| }, |
| { |
| "epoch": 1.68733153638814, |
| "grad_norm": 0.5548321008682251, |
| "learning_rate": 0.0004993308148947651, |
| "loss": 3.6766, |
| "step": 15650 |
| }, |
| { |
| "epoch": 1.692722371967655, |
| "grad_norm": 0.5533665418624878, |
| "learning_rate": 0.0004990070156502968, |
| "loss": 3.6657, |
| "step": 15700 |
| }, |
| { |
| "epoch": 1.6981132075471699, |
| "grad_norm": 0.564487636089325, |
| "learning_rate": 0.0004986832164058284, |
| "loss": 3.6623, |
| "step": 15750 |
| }, |
| { |
| "epoch": 1.7035040431266846, |
| "grad_norm": 0.6166607141494751, |
| "learning_rate": 0.0004983594171613599, |
| "loss": 3.676, |
| "step": 15800 |
| }, |
| { |
| "epoch": 1.7088948787061995, |
| "grad_norm": 0.6325422525405884, |
| "learning_rate": 0.0004980356179168915, |
| "loss": 3.6815, |
| "step": 15850 |
| }, |
| { |
| "epoch": 1.7142857142857144, |
| "grad_norm": 0.6531565189361572, |
| "learning_rate": 0.000497711818672423, |
| "loss": 3.704, |
| "step": 15900 |
| }, |
| { |
| "epoch": 1.719676549865229, |
| "grad_norm": 0.5142133831977844, |
| "learning_rate": 0.0004973880194279547, |
| "loss": 3.6707, |
| "step": 15950 |
| }, |
| { |
| "epoch": 1.7250673854447438, |
| "grad_norm": 0.5847429633140564, |
| "learning_rate": 0.0004970642201834862, |
| "loss": 3.6798, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.7250673854447438, |
| "eval_accuracy": 0.35691093755381714, |
| "eval_loss": 3.620922088623047, |
| "eval_runtime": 183.1348, |
| "eval_samples_per_second": 98.348, |
| "eval_steps_per_second": 6.148, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.7304582210242587, |
| "grad_norm": 0.5682497620582581, |
| "learning_rate": 0.0004967404209390178, |
| "loss": 3.6472, |
| "step": 16050 |
| }, |
| { |
| "epoch": 1.7358490566037736, |
| "grad_norm": 0.5600143671035767, |
| "learning_rate": 0.0004964166216945493, |
| "loss": 3.6561, |
| "step": 16100 |
| }, |
| { |
| "epoch": 1.7412398921832883, |
| "grad_norm": 0.5824739933013916, |
| "learning_rate": 0.000496092822450081, |
| "loss": 3.6703, |
| "step": 16150 |
| }, |
| { |
| "epoch": 1.7466307277628033, |
| "grad_norm": 0.5835148096084595, |
| "learning_rate": 0.0004957690232056125, |
| "loss": 3.6583, |
| "step": 16200 |
| }, |
| { |
| "epoch": 1.7520215633423182, |
| "grad_norm": 0.5758132338523865, |
| "learning_rate": 0.0004954452239611441, |
| "loss": 3.6676, |
| "step": 16250 |
| }, |
| { |
| "epoch": 1.7574123989218329, |
| "grad_norm": 0.542614758014679, |
| "learning_rate": 0.0004951214247166756, |
| "loss": 3.6587, |
| "step": 16300 |
| }, |
| { |
| "epoch": 1.7628032345013476, |
| "grad_norm": 0.5760119557380676, |
| "learning_rate": 0.0004947976254722072, |
| "loss": 3.6519, |
| "step": 16350 |
| }, |
| { |
| "epoch": 1.7681940700808625, |
| "grad_norm": 0.5551959872245789, |
| "learning_rate": 0.0004944738262277387, |
| "loss": 3.6661, |
| "step": 16400 |
| }, |
| { |
| "epoch": 1.7735849056603774, |
| "grad_norm": 0.5755851864814758, |
| "learning_rate": 0.0004941500269832703, |
| "loss": 3.6653, |
| "step": 16450 |
| }, |
| { |
| "epoch": 1.778975741239892, |
| "grad_norm": 0.5640000104904175, |
| "learning_rate": 0.0004938262277388019, |
| "loss": 3.6512, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.784366576819407, |
| "grad_norm": 0.5701029896736145, |
| "learning_rate": 0.0004935024284943335, |
| "loss": 3.647, |
| "step": 16550 |
| }, |
| { |
| "epoch": 1.789757412398922, |
| "grad_norm": 0.5458316802978516, |
| "learning_rate": 0.000493178629249865, |
| "loss": 3.6534, |
| "step": 16600 |
| }, |
| { |
| "epoch": 1.7951482479784366, |
| "grad_norm": 0.5290755033493042, |
| "learning_rate": 0.0004928548300053966, |
| "loss": 3.6619, |
| "step": 16650 |
| }, |
| { |
| "epoch": 1.8005390835579513, |
| "grad_norm": 0.56656813621521, |
| "learning_rate": 0.0004925310307609282, |
| "loss": 3.6537, |
| "step": 16700 |
| }, |
| { |
| "epoch": 1.8059299191374663, |
| "grad_norm": 0.544216513633728, |
| "learning_rate": 0.0004922072315164598, |
| "loss": 3.6468, |
| "step": 16750 |
| }, |
| { |
| "epoch": 1.8113207547169812, |
| "grad_norm": 0.5542647242546082, |
| "learning_rate": 0.0004918834322719913, |
| "loss": 3.6546, |
| "step": 16800 |
| }, |
| { |
| "epoch": 1.8167115902964959, |
| "grad_norm": 0.5754142999649048, |
| "learning_rate": 0.0004915596330275229, |
| "loss": 3.6427, |
| "step": 16850 |
| }, |
| { |
| "epoch": 1.8221024258760108, |
| "grad_norm": 0.5976019501686096, |
| "learning_rate": 0.0004912358337830544, |
| "loss": 3.6485, |
| "step": 16900 |
| }, |
| { |
| "epoch": 1.8274932614555257, |
| "grad_norm": 0.5452532768249512, |
| "learning_rate": 0.000490912034538586, |
| "loss": 3.655, |
| "step": 16950 |
| }, |
| { |
| "epoch": 1.8328840970350404, |
| "grad_norm": 0.5066123008728027, |
| "learning_rate": 0.0004905882352941175, |
| "loss": 3.6639, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.8328840970350404, |
| "eval_accuracy": 0.3585854958608139, |
| "eval_loss": 3.6066739559173584, |
| "eval_runtime": 183.4798, |
| "eval_samples_per_second": 98.163, |
| "eval_steps_per_second": 6.137, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.838274932614555, |
| "grad_norm": 0.580244779586792, |
| "learning_rate": 0.0004902644360496492, |
| "loss": 3.6551, |
| "step": 17050 |
| }, |
| { |
| "epoch": 1.8436657681940702, |
| "grad_norm": 0.5668180584907532, |
| "learning_rate": 0.0004899406368051808, |
| "loss": 3.6445, |
| "step": 17100 |
| }, |
| { |
| "epoch": 1.849056603773585, |
| "grad_norm": 0.5920048356056213, |
| "learning_rate": 0.0004896168375607123, |
| "loss": 3.6445, |
| "step": 17150 |
| }, |
| { |
| "epoch": 1.8544474393530996, |
| "grad_norm": 0.5886459946632385, |
| "learning_rate": 0.0004892930383162439, |
| "loss": 3.6436, |
| "step": 17200 |
| }, |
| { |
| "epoch": 1.8598382749326146, |
| "grad_norm": 0.544743001461029, |
| "learning_rate": 0.0004889692390717754, |
| "loss": 3.6386, |
| "step": 17250 |
| }, |
| { |
| "epoch": 1.8652291105121295, |
| "grad_norm": 0.5967076420783997, |
| "learning_rate": 0.0004886519158121964, |
| "loss": 3.6638, |
| "step": 17300 |
| }, |
| { |
| "epoch": 1.8706199460916442, |
| "grad_norm": 0.575445294380188, |
| "learning_rate": 0.000488328116567728, |
| "loss": 3.6653, |
| "step": 17350 |
| }, |
| { |
| "epoch": 1.8760107816711589, |
| "grad_norm": 0.5783690810203552, |
| "learning_rate": 0.0004880043173232595, |
| "loss": 3.6453, |
| "step": 17400 |
| }, |
| { |
| "epoch": 1.881401617250674, |
| "grad_norm": 0.5519454479217529, |
| "learning_rate": 0.0004876805180787911, |
| "loss": 3.6561, |
| "step": 17450 |
| }, |
| { |
| "epoch": 1.8867924528301887, |
| "grad_norm": 0.5622634887695312, |
| "learning_rate": 0.0004873567188343227, |
| "loss": 3.6404, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.8921832884097034, |
| "grad_norm": 0.5808458924293518, |
| "learning_rate": 0.0004870329195898542, |
| "loss": 3.655, |
| "step": 17550 |
| }, |
| { |
| "epoch": 1.8975741239892183, |
| "grad_norm": 0.5541762113571167, |
| "learning_rate": 0.00048670912034538583, |
| "loss": 3.6409, |
| "step": 17600 |
| }, |
| { |
| "epoch": 1.9029649595687332, |
| "grad_norm": 0.5877305865287781, |
| "learning_rate": 0.0004863853211009174, |
| "loss": 3.6382, |
| "step": 17650 |
| }, |
| { |
| "epoch": 1.908355795148248, |
| "grad_norm": 0.5746439695358276, |
| "learning_rate": 0.000486061521856449, |
| "loss": 3.6504, |
| "step": 17700 |
| }, |
| { |
| "epoch": 1.9137466307277629, |
| "grad_norm": 0.5914478898048401, |
| "learning_rate": 0.00048573772261198054, |
| "loss": 3.6397, |
| "step": 17750 |
| }, |
| { |
| "epoch": 1.9191374663072778, |
| "grad_norm": 0.5980408191680908, |
| "learning_rate": 0.00048541392336751214, |
| "loss": 3.655, |
| "step": 17800 |
| }, |
| { |
| "epoch": 1.9245283018867925, |
| "grad_norm": 0.5955798625946045, |
| "learning_rate": 0.0004850901241230437, |
| "loss": 3.6641, |
| "step": 17850 |
| }, |
| { |
| "epoch": 1.9299191374663072, |
| "grad_norm": 0.6046438813209534, |
| "learning_rate": 0.0004847663248785753, |
| "loss": 3.639, |
| "step": 17900 |
| }, |
| { |
| "epoch": 1.935309973045822, |
| "grad_norm": 0.6008595824241638, |
| "learning_rate": 0.0004844425256341068, |
| "loss": 3.6477, |
| "step": 17950 |
| }, |
| { |
| "epoch": 1.940700808625337, |
| "grad_norm": 0.5790556073188782, |
| "learning_rate": 0.00048411872638963834, |
| "loss": 3.6461, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.940700808625337, |
| "eval_accuracy": 0.3601660694198512, |
| "eval_loss": 3.5904529094696045, |
| "eval_runtime": 183.0994, |
| "eval_samples_per_second": 98.367, |
| "eval_steps_per_second": 6.15, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.9460916442048517, |
| "grad_norm": 0.5579012036323547, |
| "learning_rate": 0.00048379492714516995, |
| "loss": 3.6269, |
| "step": 18050 |
| }, |
| { |
| "epoch": 1.9514824797843666, |
| "grad_norm": 0.5438567399978638, |
| "learning_rate": 0.0004834711279007015, |
| "loss": 3.6563, |
| "step": 18100 |
| }, |
| { |
| "epoch": 1.9568733153638815, |
| "grad_norm": 0.5683955550193787, |
| "learning_rate": 0.0004831473286562331, |
| "loss": 3.6579, |
| "step": 18150 |
| }, |
| { |
| "epoch": 1.9622641509433962, |
| "grad_norm": 0.5797238349914551, |
| "learning_rate": 0.00048282352941176465, |
| "loss": 3.6462, |
| "step": 18200 |
| }, |
| { |
| "epoch": 1.967654986522911, |
| "grad_norm": 0.5775489211082458, |
| "learning_rate": 0.00048249973016729626, |
| "loss": 3.6476, |
| "step": 18250 |
| }, |
| { |
| "epoch": 1.9730458221024259, |
| "grad_norm": 0.5357772707939148, |
| "learning_rate": 0.0004821759309228278, |
| "loss": 3.6281, |
| "step": 18300 |
| }, |
| { |
| "epoch": 1.9784366576819408, |
| "grad_norm": 0.7332191467285156, |
| "learning_rate": 0.00048185213167835936, |
| "loss": 3.6523, |
| "step": 18350 |
| }, |
| { |
| "epoch": 1.9838274932614555, |
| "grad_norm": 0.5149871110916138, |
| "learning_rate": 0.00048152833243389096, |
| "loss": 3.6294, |
| "step": 18400 |
| }, |
| { |
| "epoch": 1.9892183288409704, |
| "grad_norm": 0.5842679142951965, |
| "learning_rate": 0.0004812045331894225, |
| "loss": 3.6318, |
| "step": 18450 |
| }, |
| { |
| "epoch": 1.9946091644204853, |
| "grad_norm": 0.5898716449737549, |
| "learning_rate": 0.0004808807339449541, |
| "loss": 3.6217, |
| "step": 18500 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 1.104094386100769, |
| "learning_rate": 0.0004805569347004856, |
| "loss": 3.6377, |
| "step": 18550 |
| }, |
| { |
| "epoch": 2.0053908355795147, |
| "grad_norm": 0.5916197299957275, |
| "learning_rate": 0.0004802331354560173, |
| "loss": 3.552, |
| "step": 18600 |
| }, |
| { |
| "epoch": 2.01078167115903, |
| "grad_norm": 0.5588786005973816, |
| "learning_rate": 0.00047990933621154877, |
| "loss": 3.5459, |
| "step": 18650 |
| }, |
| { |
| "epoch": 2.0161725067385445, |
| "grad_norm": 0.5954398512840271, |
| "learning_rate": 0.0004795855369670804, |
| "loss": 3.5515, |
| "step": 18700 |
| }, |
| { |
| "epoch": 2.0215633423180592, |
| "grad_norm": 0.5872916579246521, |
| "learning_rate": 0.0004792617377226119, |
| "loss": 3.5445, |
| "step": 18750 |
| }, |
| { |
| "epoch": 2.026954177897574, |
| "grad_norm": 0.5799959301948547, |
| "learning_rate": 0.0004789379384781435, |
| "loss": 3.5516, |
| "step": 18800 |
| }, |
| { |
| "epoch": 2.032345013477089, |
| "grad_norm": 0.5512281060218811, |
| "learning_rate": 0.0004786141392336751, |
| "loss": 3.5498, |
| "step": 18850 |
| }, |
| { |
| "epoch": 2.0377358490566038, |
| "grad_norm": 0.6156049370765686, |
| "learning_rate": 0.00047829033998920663, |
| "loss": 3.5626, |
| "step": 18900 |
| }, |
| { |
| "epoch": 2.0431266846361185, |
| "grad_norm": 0.5738759636878967, |
| "learning_rate": 0.00047796654074473824, |
| "loss": 3.5591, |
| "step": 18950 |
| }, |
| { |
| "epoch": 2.0485175202156336, |
| "grad_norm": 0.5836319923400879, |
| "learning_rate": 0.0004776427415002698, |
| "loss": 3.5609, |
| "step": 19000 |
| }, |
| { |
| "epoch": 2.0485175202156336, |
| "eval_accuracy": 0.36152336129442963, |
| "eval_loss": 3.581784725189209, |
| "eval_runtime": 183.3107, |
| "eval_samples_per_second": 98.254, |
| "eval_steps_per_second": 6.143, |
| "step": 19000 |
| }, |
| { |
| "epoch": 2.0539083557951483, |
| "grad_norm": 0.5089203119277954, |
| "learning_rate": 0.0004773189422558014, |
| "loss": 3.5693, |
| "step": 19050 |
| }, |
| { |
| "epoch": 2.059299191374663, |
| "grad_norm": 0.5695809721946716, |
| "learning_rate": 0.00047699514301133294, |
| "loss": 3.5613, |
| "step": 19100 |
| }, |
| { |
| "epoch": 2.0646900269541777, |
| "grad_norm": 0.5608691573143005, |
| "learning_rate": 0.00047667134376686455, |
| "loss": 3.5554, |
| "step": 19150 |
| }, |
| { |
| "epoch": 2.070080862533693, |
| "grad_norm": 0.5364721417427063, |
| "learning_rate": 0.0004763475445223961, |
| "loss": 3.5751, |
| "step": 19200 |
| }, |
| { |
| "epoch": 2.0754716981132075, |
| "grad_norm": 0.5620273351669312, |
| "learning_rate": 0.0004760237452779276, |
| "loss": 3.5632, |
| "step": 19250 |
| }, |
| { |
| "epoch": 2.0808625336927222, |
| "grad_norm": 0.5621458888053894, |
| "learning_rate": 0.00047570642201834855, |
| "loss": 3.5559, |
| "step": 19300 |
| }, |
| { |
| "epoch": 2.0862533692722374, |
| "grad_norm": 0.6203056573867798, |
| "learning_rate": 0.00047538262277388015, |
| "loss": 3.5737, |
| "step": 19350 |
| }, |
| { |
| "epoch": 2.091644204851752, |
| "grad_norm": 0.5451525449752808, |
| "learning_rate": 0.0004750588235294117, |
| "loss": 3.5642, |
| "step": 19400 |
| }, |
| { |
| "epoch": 2.0970350404312668, |
| "grad_norm": 0.6249733567237854, |
| "learning_rate": 0.0004747350242849433, |
| "loss": 3.5484, |
| "step": 19450 |
| }, |
| { |
| "epoch": 2.1024258760107815, |
| "grad_norm": 0.5598724484443665, |
| "learning_rate": 0.00047441122504047486, |
| "loss": 3.5713, |
| "step": 19500 |
| }, |
| { |
| "epoch": 2.1078167115902966, |
| "grad_norm": 0.5830626487731934, |
| "learning_rate": 0.0004740874257960064, |
| "loss": 3.562, |
| "step": 19550 |
| }, |
| { |
| "epoch": 2.1132075471698113, |
| "grad_norm": 0.5715135931968689, |
| "learning_rate": 0.000473763626551538, |
| "loss": 3.5612, |
| "step": 19600 |
| }, |
| { |
| "epoch": 2.118598382749326, |
| "grad_norm": 0.6303309798240662, |
| "learning_rate": 0.00047343982730706956, |
| "loss": 3.5628, |
| "step": 19650 |
| }, |
| { |
| "epoch": 2.123989218328841, |
| "grad_norm": 0.5836952328681946, |
| "learning_rate": 0.00047311602806260117, |
| "loss": 3.5534, |
| "step": 19700 |
| }, |
| { |
| "epoch": 2.129380053908356, |
| "grad_norm": 0.6122373938560486, |
| "learning_rate": 0.0004727922288181327, |
| "loss": 3.5445, |
| "step": 19750 |
| }, |
| { |
| "epoch": 2.1347708894878705, |
| "grad_norm": 0.5842995047569275, |
| "learning_rate": 0.0004724684295736643, |
| "loss": 3.566, |
| "step": 19800 |
| }, |
| { |
| "epoch": 2.1401617250673857, |
| "grad_norm": 0.585098147392273, |
| "learning_rate": 0.0004721446303291959, |
| "loss": 3.5585, |
| "step": 19850 |
| }, |
| { |
| "epoch": 2.1455525606469004, |
| "grad_norm": 0.5444265604019165, |
| "learning_rate": 0.0004718208310847275, |
| "loss": 3.5622, |
| "step": 19900 |
| }, |
| { |
| "epoch": 2.150943396226415, |
| "grad_norm": 0.5859130620956421, |
| "learning_rate": 0.000471497031840259, |
| "loss": 3.5544, |
| "step": 19950 |
| }, |
| { |
| "epoch": 2.1563342318059298, |
| "grad_norm": 0.6213077306747437, |
| "learning_rate": 0.0004711732325957905, |
| "loss": 3.5583, |
| "step": 20000 |
| }, |
| { |
| "epoch": 2.1563342318059298, |
| "eval_accuracy": 0.3627601571164219, |
| "eval_loss": 3.572903871536255, |
| "eval_runtime": 182.8426, |
| "eval_samples_per_second": 98.505, |
| "eval_steps_per_second": 6.158, |
| "step": 20000 |
| }, |
| { |
| "epoch": 2.161725067385445, |
| "grad_norm": 0.570515513420105, |
| "learning_rate": 0.00047084943335132213, |
| "loss": 3.5599, |
| "step": 20050 |
| }, |
| { |
| "epoch": 2.1671159029649596, |
| "grad_norm": 0.5801568031311035, |
| "learning_rate": 0.0004705256341068537, |
| "loss": 3.5563, |
| "step": 20100 |
| }, |
| { |
| "epoch": 2.1725067385444743, |
| "grad_norm": 0.5625453591346741, |
| "learning_rate": 0.0004702018348623853, |
| "loss": 3.5663, |
| "step": 20150 |
| }, |
| { |
| "epoch": 2.177897574123989, |
| "grad_norm": 0.5203284025192261, |
| "learning_rate": 0.00046987803561791684, |
| "loss": 3.5694, |
| "step": 20200 |
| }, |
| { |
| "epoch": 2.183288409703504, |
| "grad_norm": 0.5144570469856262, |
| "learning_rate": 0.00046955423637344844, |
| "loss": 3.5731, |
| "step": 20250 |
| }, |
| { |
| "epoch": 2.188679245283019, |
| "grad_norm": 0.6565627455711365, |
| "learning_rate": 0.00046923043712898, |
| "loss": 3.5553, |
| "step": 20300 |
| }, |
| { |
| "epoch": 2.1940700808625335, |
| "grad_norm": 0.5670105814933777, |
| "learning_rate": 0.0004689066378845116, |
| "loss": 3.5481, |
| "step": 20350 |
| }, |
| { |
| "epoch": 2.1994609164420487, |
| "grad_norm": 0.6016785502433777, |
| "learning_rate": 0.00046858283864004315, |
| "loss": 3.5625, |
| "step": 20400 |
| }, |
| { |
| "epoch": 2.2048517520215634, |
| "grad_norm": 0.7234423756599426, |
| "learning_rate": 0.0004682590393955747, |
| "loss": 3.5573, |
| "step": 20450 |
| }, |
| { |
| "epoch": 2.210242587601078, |
| "grad_norm": 0.5361531972885132, |
| "learning_rate": 0.0004679352401511063, |
| "loss": 3.5735, |
| "step": 20500 |
| }, |
| { |
| "epoch": 2.215633423180593, |
| "grad_norm": 0.5746288299560547, |
| "learning_rate": 0.0004676114409066378, |
| "loss": 3.569, |
| "step": 20550 |
| }, |
| { |
| "epoch": 2.221024258760108, |
| "grad_norm": 0.5691829919815063, |
| "learning_rate": 0.00046728764166216946, |
| "loss": 3.564, |
| "step": 20600 |
| }, |
| { |
| "epoch": 2.2264150943396226, |
| "grad_norm": 0.5995154976844788, |
| "learning_rate": 0.00046696384241770095, |
| "loss": 3.5719, |
| "step": 20650 |
| }, |
| { |
| "epoch": 2.2318059299191373, |
| "grad_norm": 0.6250969767570496, |
| "learning_rate": 0.00046664004317323256, |
| "loss": 3.5569, |
| "step": 20700 |
| }, |
| { |
| "epoch": 2.2371967654986524, |
| "grad_norm": 0.5564571619033813, |
| "learning_rate": 0.0004663162439287641, |
| "loss": 3.5747, |
| "step": 20750 |
| }, |
| { |
| "epoch": 2.242587601078167, |
| "grad_norm": 0.5897787809371948, |
| "learning_rate": 0.00046599244468429566, |
| "loss": 3.5692, |
| "step": 20800 |
| }, |
| { |
| "epoch": 2.247978436657682, |
| "grad_norm": 0.5856361985206604, |
| "learning_rate": 0.00046566864543982726, |
| "loss": 3.573, |
| "step": 20850 |
| }, |
| { |
| "epoch": 2.2533692722371965, |
| "grad_norm": 0.5496835708618164, |
| "learning_rate": 0.0004653448461953588, |
| "loss": 3.5611, |
| "step": 20900 |
| }, |
| { |
| "epoch": 2.2587601078167117, |
| "grad_norm": 0.6221925616264343, |
| "learning_rate": 0.0004650210469508904, |
| "loss": 3.5581, |
| "step": 20950 |
| }, |
| { |
| "epoch": 2.2641509433962264, |
| "grad_norm": 0.5658513307571411, |
| "learning_rate": 0.00046469724770642197, |
| "loss": 3.5592, |
| "step": 21000 |
| }, |
| { |
| "epoch": 2.2641509433962264, |
| "eval_accuracy": 0.3634160946001579, |
| "eval_loss": 3.564391851425171, |
| "eval_runtime": 183.2391, |
| "eval_samples_per_second": 98.292, |
| "eval_steps_per_second": 6.145, |
| "step": 21000 |
| }, |
| { |
| "epoch": 2.269541778975741, |
| "grad_norm": 0.5886167883872986, |
| "learning_rate": 0.0004643734484619536, |
| "loss": 3.5563, |
| "step": 21050 |
| }, |
| { |
| "epoch": 2.274932614555256, |
| "grad_norm": 0.6223853230476379, |
| "learning_rate": 0.0004640496492174851, |
| "loss": 3.545, |
| "step": 21100 |
| }, |
| { |
| "epoch": 2.280323450134771, |
| "grad_norm": 0.6089233756065369, |
| "learning_rate": 0.00046372584997301673, |
| "loss": 3.5734, |
| "step": 21150 |
| }, |
| { |
| "epoch": 2.2857142857142856, |
| "grad_norm": 0.5425137877464294, |
| "learning_rate": 0.0004634020507285483, |
| "loss": 3.5528, |
| "step": 21200 |
| }, |
| { |
| "epoch": 2.2911051212938007, |
| "grad_norm": 0.5784063935279846, |
| "learning_rate": 0.0004630782514840798, |
| "loss": 3.5513, |
| "step": 21250 |
| }, |
| { |
| "epoch": 2.2964959568733154, |
| "grad_norm": 0.5775195360183716, |
| "learning_rate": 0.0004627544522396114, |
| "loss": 3.5616, |
| "step": 21300 |
| }, |
| { |
| "epoch": 2.30188679245283, |
| "grad_norm": 0.5808774828910828, |
| "learning_rate": 0.00046243712898003233, |
| "loss": 3.5398, |
| "step": 21350 |
| }, |
| { |
| "epoch": 2.3072776280323453, |
| "grad_norm": 0.5682329535484314, |
| "learning_rate": 0.0004621133297355639, |
| "loss": 3.5759, |
| "step": 21400 |
| }, |
| { |
| "epoch": 2.31266846361186, |
| "grad_norm": 0.6007602214813232, |
| "learning_rate": 0.0004617895304910955, |
| "loss": 3.5335, |
| "step": 21450 |
| }, |
| { |
| "epoch": 2.3180592991913747, |
| "grad_norm": 0.6135632991790771, |
| "learning_rate": 0.00046146573124662704, |
| "loss": 3.5656, |
| "step": 21500 |
| }, |
| { |
| "epoch": 2.3234501347708894, |
| "grad_norm": 0.6100749969482422, |
| "learning_rate": 0.00046114193200215864, |
| "loss": 3.5654, |
| "step": 21550 |
| }, |
| { |
| "epoch": 2.3288409703504045, |
| "grad_norm": 0.5744963884353638, |
| "learning_rate": 0.0004608181327576902, |
| "loss": 3.5383, |
| "step": 21600 |
| }, |
| { |
| "epoch": 2.334231805929919, |
| "grad_norm": 0.6241092085838318, |
| "learning_rate": 0.00046049433351322175, |
| "loss": 3.5591, |
| "step": 21650 |
| }, |
| { |
| "epoch": 2.339622641509434, |
| "grad_norm": 0.5725473761558533, |
| "learning_rate": 0.00046017053426875335, |
| "loss": 3.5715, |
| "step": 21700 |
| }, |
| { |
| "epoch": 2.3450134770889486, |
| "grad_norm": 0.6116576790809631, |
| "learning_rate": 0.0004598467350242849, |
| "loss": 3.5538, |
| "step": 21750 |
| }, |
| { |
| "epoch": 2.3504043126684637, |
| "grad_norm": 0.5640816688537598, |
| "learning_rate": 0.0004595229357798165, |
| "loss": 3.5645, |
| "step": 21800 |
| }, |
| { |
| "epoch": 2.3557951482479784, |
| "grad_norm": 0.6325236558914185, |
| "learning_rate": 0.00045919913653534806, |
| "loss": 3.5661, |
| "step": 21850 |
| }, |
| { |
| "epoch": 2.361185983827493, |
| "grad_norm": 0.5464061498641968, |
| "learning_rate": 0.00045887533729087966, |
| "loss": 3.5576, |
| "step": 21900 |
| }, |
| { |
| "epoch": 2.3665768194070083, |
| "grad_norm": 0.5492430329322815, |
| "learning_rate": 0.00045855153804641116, |
| "loss": 3.5492, |
| "step": 21950 |
| }, |
| { |
| "epoch": 2.371967654986523, |
| "grad_norm": 0.5400915145874023, |
| "learning_rate": 0.0004582277388019427, |
| "loss": 3.5447, |
| "step": 22000 |
| }, |
| { |
| "epoch": 2.371967654986523, |
| "eval_accuracy": 0.36474970014519287, |
| "eval_loss": 3.5518150329589844, |
| "eval_runtime": 183.1051, |
| "eval_samples_per_second": 98.364, |
| "eval_steps_per_second": 6.149, |
| "step": 22000 |
| }, |
| { |
| "epoch": 2.3773584905660377, |
| "grad_norm": 0.5935810804367065, |
| "learning_rate": 0.0004579039395574743, |
| "loss": 3.5459, |
| "step": 22050 |
| }, |
| { |
| "epoch": 2.382749326145553, |
| "grad_norm": 0.5556415319442749, |
| "learning_rate": 0.00045758014031300586, |
| "loss": 3.5638, |
| "step": 22100 |
| }, |
| { |
| "epoch": 2.3881401617250675, |
| "grad_norm": 0.5712567567825317, |
| "learning_rate": 0.0004572628170534268, |
| "loss": 3.5564, |
| "step": 22150 |
| }, |
| { |
| "epoch": 2.393530997304582, |
| "grad_norm": 0.5914583206176758, |
| "learning_rate": 0.0004569390178089584, |
| "loss": 3.5452, |
| "step": 22200 |
| }, |
| { |
| "epoch": 2.398921832884097, |
| "grad_norm": 0.7909786701202393, |
| "learning_rate": 0.00045661521856448997, |
| "loss": 3.539, |
| "step": 22250 |
| }, |
| { |
| "epoch": 2.404312668463612, |
| "grad_norm": 0.6220540404319763, |
| "learning_rate": 0.0004562914193200216, |
| "loss": 3.5651, |
| "step": 22300 |
| }, |
| { |
| "epoch": 2.4097035040431267, |
| "grad_norm": 0.5887907147407532, |
| "learning_rate": 0.0004559676200755531, |
| "loss": 3.5751, |
| "step": 22350 |
| }, |
| { |
| "epoch": 2.4150943396226414, |
| "grad_norm": 0.6566967964172363, |
| "learning_rate": 0.0004556438208310847, |
| "loss": 3.5596, |
| "step": 22400 |
| }, |
| { |
| "epoch": 2.420485175202156, |
| "grad_norm": 0.5419936776161194, |
| "learning_rate": 0.0004553200215866163, |
| "loss": 3.5656, |
| "step": 22450 |
| }, |
| { |
| "epoch": 2.4258760107816713, |
| "grad_norm": 0.5918455719947815, |
| "learning_rate": 0.00045499622234214783, |
| "loss": 3.5622, |
| "step": 22500 |
| }, |
| { |
| "epoch": 2.431266846361186, |
| "grad_norm": 0.5798980593681335, |
| "learning_rate": 0.00045467242309767944, |
| "loss": 3.5416, |
| "step": 22550 |
| }, |
| { |
| "epoch": 2.4366576819407006, |
| "grad_norm": 0.623996376991272, |
| "learning_rate": 0.00045434862385321093, |
| "loss": 3.5592, |
| "step": 22600 |
| }, |
| { |
| "epoch": 2.442048517520216, |
| "grad_norm": 0.550102710723877, |
| "learning_rate": 0.0004540248246087426, |
| "loss": 3.5557, |
| "step": 22650 |
| }, |
| { |
| "epoch": 2.4474393530997305, |
| "grad_norm": 0.6159896850585938, |
| "learning_rate": 0.0004537010253642741, |
| "loss": 3.5641, |
| "step": 22700 |
| }, |
| { |
| "epoch": 2.452830188679245, |
| "grad_norm": 0.5841001272201538, |
| "learning_rate": 0.00045337722611980564, |
| "loss": 3.5525, |
| "step": 22750 |
| }, |
| { |
| "epoch": 2.4582210242587603, |
| "grad_norm": 0.5647518038749695, |
| "learning_rate": 0.00045305342687533724, |
| "loss": 3.5495, |
| "step": 22800 |
| }, |
| { |
| "epoch": 2.463611859838275, |
| "grad_norm": 0.5946810245513916, |
| "learning_rate": 0.0004527296276308688, |
| "loss": 3.5387, |
| "step": 22850 |
| }, |
| { |
| "epoch": 2.4690026954177897, |
| "grad_norm": 0.5658169984817505, |
| "learning_rate": 0.0004524058283864004, |
| "loss": 3.5527, |
| "step": 22900 |
| }, |
| { |
| "epoch": 2.4743935309973044, |
| "grad_norm": 0.6262547373771667, |
| "learning_rate": 0.00045208202914193195, |
| "loss": 3.5385, |
| "step": 22950 |
| }, |
| { |
| "epoch": 2.4797843665768196, |
| "grad_norm": 0.5860013365745544, |
| "learning_rate": 0.00045175822989746355, |
| "loss": 3.5527, |
| "step": 23000 |
| }, |
| { |
| "epoch": 2.4797843665768196, |
| "eval_accuracy": 0.36564021921372036, |
| "eval_loss": 3.5422823429107666, |
| "eval_runtime": 183.1023, |
| "eval_samples_per_second": 98.366, |
| "eval_steps_per_second": 6.15, |
| "step": 23000 |
| }, |
| { |
| "epoch": 2.4851752021563343, |
| "grad_norm": 0.6349847912788391, |
| "learning_rate": 0.0004514344306529951, |
| "loss": 3.5606, |
| "step": 23050 |
| }, |
| { |
| "epoch": 2.490566037735849, |
| "grad_norm": 0.5899332761764526, |
| "learning_rate": 0.0004511106314085267, |
| "loss": 3.5561, |
| "step": 23100 |
| }, |
| { |
| "epoch": 2.4959568733153636, |
| "grad_norm": 0.7006366848945618, |
| "learning_rate": 0.00045078683216405826, |
| "loss": 3.5688, |
| "step": 23150 |
| }, |
| { |
| "epoch": 2.501347708894879, |
| "grad_norm": 0.6054086685180664, |
| "learning_rate": 0.00045046303291958976, |
| "loss": 3.5605, |
| "step": 23200 |
| }, |
| { |
| "epoch": 2.5067385444743935, |
| "grad_norm": 0.5730805397033691, |
| "learning_rate": 0.0004501392336751214, |
| "loss": 3.5609, |
| "step": 23250 |
| }, |
| { |
| "epoch": 2.512129380053908, |
| "grad_norm": 0.5215578675270081, |
| "learning_rate": 0.0004498154344306529, |
| "loss": 3.5586, |
| "step": 23300 |
| }, |
| { |
| "epoch": 2.5175202156334233, |
| "grad_norm": 0.5702964067459106, |
| "learning_rate": 0.0004494916351861845, |
| "loss": 3.5549, |
| "step": 23350 |
| }, |
| { |
| "epoch": 2.522911051212938, |
| "grad_norm": 0.6069727540016174, |
| "learning_rate": 0.00044916783594171607, |
| "loss": 3.534, |
| "step": 23400 |
| }, |
| { |
| "epoch": 2.5283018867924527, |
| "grad_norm": 0.5799237489700317, |
| "learning_rate": 0.00044884403669724767, |
| "loss": 3.5509, |
| "step": 23450 |
| }, |
| { |
| "epoch": 2.533692722371968, |
| "grad_norm": 0.6026593446731567, |
| "learning_rate": 0.0004485202374527792, |
| "loss": 3.5454, |
| "step": 23500 |
| }, |
| { |
| "epoch": 2.5390835579514826, |
| "grad_norm": 0.557666003704071, |
| "learning_rate": 0.00044819643820831083, |
| "loss": 3.548, |
| "step": 23550 |
| }, |
| { |
| "epoch": 2.5444743935309972, |
| "grad_norm": 0.6157044768333435, |
| "learning_rate": 0.0004478726389638424, |
| "loss": 3.5527, |
| "step": 23600 |
| }, |
| { |
| "epoch": 2.5498652291105124, |
| "grad_norm": 0.5705444812774658, |
| "learning_rate": 0.00044754883971937393, |
| "loss": 3.5384, |
| "step": 23650 |
| }, |
| { |
| "epoch": 2.555256064690027, |
| "grad_norm": 0.5903183221817017, |
| "learning_rate": 0.00044722504047490553, |
| "loss": 3.529, |
| "step": 23700 |
| }, |
| { |
| "epoch": 2.560646900269542, |
| "grad_norm": 0.5920628905296326, |
| "learning_rate": 0.0004469012412304371, |
| "loss": 3.5377, |
| "step": 23750 |
| }, |
| { |
| "epoch": 2.5660377358490565, |
| "grad_norm": 0.5568437576293945, |
| "learning_rate": 0.0004465774419859687, |
| "loss": 3.5558, |
| "step": 23800 |
| }, |
| { |
| "epoch": 2.571428571428571, |
| "grad_norm": 0.5784709453582764, |
| "learning_rate": 0.00044625364274150024, |
| "loss": 3.5412, |
| "step": 23850 |
| }, |
| { |
| "epoch": 2.5768194070080863, |
| "grad_norm": 0.5762706398963928, |
| "learning_rate": 0.00044592984349703184, |
| "loss": 3.5437, |
| "step": 23900 |
| }, |
| { |
| "epoch": 2.582210242587601, |
| "grad_norm": 0.6410912871360779, |
| "learning_rate": 0.00044560604425256334, |
| "loss": 3.5356, |
| "step": 23950 |
| }, |
| { |
| "epoch": 2.5876010781671157, |
| "grad_norm": 0.579069197177887, |
| "learning_rate": 0.000445282245008095, |
| "loss": 3.541, |
| "step": 24000 |
| }, |
| { |
| "epoch": 2.5876010781671157, |
| "eval_accuracy": 0.3668237751206835, |
| "eval_loss": 3.531728982925415, |
| "eval_runtime": 183.177, |
| "eval_samples_per_second": 98.326, |
| "eval_steps_per_second": 6.147, |
| "step": 24000 |
| }, |
| { |
| "epoch": 2.592991913746631, |
| "grad_norm": 0.5918624401092529, |
| "learning_rate": 0.0004449584457636265, |
| "loss": 3.5656, |
| "step": 24050 |
| }, |
| { |
| "epoch": 2.5983827493261455, |
| "grad_norm": 0.6068958640098572, |
| "learning_rate": 0.00044463464651915805, |
| "loss": 3.5392, |
| "step": 24100 |
| }, |
| { |
| "epoch": 2.6037735849056602, |
| "grad_norm": 0.5975248217582703, |
| "learning_rate": 0.00044431084727468965, |
| "loss": 3.5544, |
| "step": 24150 |
| }, |
| { |
| "epoch": 2.6091644204851754, |
| "grad_norm": 0.5652947425842285, |
| "learning_rate": 0.0004439870480302212, |
| "loss": 3.5379, |
| "step": 24200 |
| }, |
| { |
| "epoch": 2.61455525606469, |
| "grad_norm": 0.5702681541442871, |
| "learning_rate": 0.0004436632487857528, |
| "loss": 3.5553, |
| "step": 24250 |
| }, |
| { |
| "epoch": 2.6199460916442048, |
| "grad_norm": 0.5995779037475586, |
| "learning_rate": 0.00044333944954128436, |
| "loss": 3.5496, |
| "step": 24300 |
| }, |
| { |
| "epoch": 2.62533692722372, |
| "grad_norm": 0.5882319211959839, |
| "learning_rate": 0.00044301565029681596, |
| "loss": 3.541, |
| "step": 24350 |
| }, |
| { |
| "epoch": 2.6307277628032346, |
| "grad_norm": 0.5981907844543457, |
| "learning_rate": 0.0004426918510523475, |
| "loss": 3.5461, |
| "step": 24400 |
| }, |
| { |
| "epoch": 2.6361185983827493, |
| "grad_norm": 0.5818279385566711, |
| "learning_rate": 0.0004423680518078791, |
| "loss": 3.5445, |
| "step": 24450 |
| }, |
| { |
| "epoch": 2.641509433962264, |
| "grad_norm": 0.5762891173362732, |
| "learning_rate": 0.00044204425256341067, |
| "loss": 3.5515, |
| "step": 24500 |
| }, |
| { |
| "epoch": 2.6469002695417787, |
| "grad_norm": 0.5954058766365051, |
| "learning_rate": 0.0004417204533189422, |
| "loss": 3.5369, |
| "step": 24550 |
| }, |
| { |
| "epoch": 2.652291105121294, |
| "grad_norm": 0.577364444732666, |
| "learning_rate": 0.0004413966540744738, |
| "loss": 3.544, |
| "step": 24600 |
| }, |
| { |
| "epoch": 2.6576819407008085, |
| "grad_norm": 0.5351988077163696, |
| "learning_rate": 0.0004410728548300053, |
| "loss": 3.5396, |
| "step": 24650 |
| }, |
| { |
| "epoch": 2.6630727762803232, |
| "grad_norm": 0.5411438345909119, |
| "learning_rate": 0.0004407490555855369, |
| "loss": 3.5456, |
| "step": 24700 |
| }, |
| { |
| "epoch": 2.6684636118598384, |
| "grad_norm": 0.5812211632728577, |
| "learning_rate": 0.0004404252563410685, |
| "loss": 3.5628, |
| "step": 24750 |
| }, |
| { |
| "epoch": 2.673854447439353, |
| "grad_norm": 0.5671088099479675, |
| "learning_rate": 0.0004401014570966001, |
| "loss": 3.5564, |
| "step": 24800 |
| }, |
| { |
| "epoch": 2.6792452830188678, |
| "grad_norm": 0.5471763610839844, |
| "learning_rate": 0.00043977765785213163, |
| "loss": 3.558, |
| "step": 24850 |
| }, |
| { |
| "epoch": 2.684636118598383, |
| "grad_norm": 0.6299545764923096, |
| "learning_rate": 0.0004394538586076632, |
| "loss": 3.5286, |
| "step": 24900 |
| }, |
| { |
| "epoch": 2.6900269541778976, |
| "grad_norm": 0.584722101688385, |
| "learning_rate": 0.0004391300593631948, |
| "loss": 3.5491, |
| "step": 24950 |
| }, |
| { |
| "epoch": 2.6954177897574123, |
| "grad_norm": 0.574306309223175, |
| "learning_rate": 0.00043880626011872634, |
| "loss": 3.5277, |
| "step": 25000 |
| }, |
| { |
| "epoch": 2.6954177897574123, |
| "eval_accuracy": 0.3673932249057192, |
| "eval_loss": 3.5235400199890137, |
| "eval_runtime": 183.145, |
| "eval_samples_per_second": 98.343, |
| "eval_steps_per_second": 6.148, |
| "step": 25000 |
| }, |
| { |
| "epoch": 2.7008086253369274, |
| "grad_norm": 0.5931011438369751, |
| "learning_rate": 0.00043848246087425794, |
| "loss": 3.5396, |
| "step": 25050 |
| }, |
| { |
| "epoch": 2.706199460916442, |
| "grad_norm": 0.5480889081954956, |
| "learning_rate": 0.0004381586616297895, |
| "loss": 3.5412, |
| "step": 25100 |
| }, |
| { |
| "epoch": 2.711590296495957, |
| "grad_norm": 0.5581277012825012, |
| "learning_rate": 0.0004378348623853211, |
| "loss": 3.5415, |
| "step": 25150 |
| }, |
| { |
| "epoch": 2.7169811320754715, |
| "grad_norm": 0.5917746424674988, |
| "learning_rate": 0.00043751106314085265, |
| "loss": 3.5425, |
| "step": 25200 |
| }, |
| { |
| "epoch": 2.7223719676549867, |
| "grad_norm": 0.641676664352417, |
| "learning_rate": 0.00043718726389638425, |
| "loss": 3.5354, |
| "step": 25250 |
| }, |
| { |
| "epoch": 2.7277628032345014, |
| "grad_norm": 0.5899236798286438, |
| "learning_rate": 0.00043686346465191575, |
| "loss": 3.5433, |
| "step": 25300 |
| }, |
| { |
| "epoch": 2.733153638814016, |
| "grad_norm": 0.638559103012085, |
| "learning_rate": 0.0004365396654074473, |
| "loss": 3.5397, |
| "step": 25350 |
| }, |
| { |
| "epoch": 2.7385444743935308, |
| "grad_norm": 0.5510205626487732, |
| "learning_rate": 0.0004362158661629789, |
| "loss": 3.5326, |
| "step": 25400 |
| }, |
| { |
| "epoch": 2.743935309973046, |
| "grad_norm": 0.6142581105232239, |
| "learning_rate": 0.00043589206691851045, |
| "loss": 3.55, |
| "step": 25450 |
| }, |
| { |
| "epoch": 2.7493261455525606, |
| "grad_norm": 0.5508300065994263, |
| "learning_rate": 0.00043556826767404206, |
| "loss": 3.5277, |
| "step": 25500 |
| }, |
| { |
| "epoch": 2.7547169811320753, |
| "grad_norm": 0.647884726524353, |
| "learning_rate": 0.0004352444684295736, |
| "loss": 3.5193, |
| "step": 25550 |
| }, |
| { |
| "epoch": 2.7601078167115904, |
| "grad_norm": 0.5603880882263184, |
| "learning_rate": 0.0004349206691851052, |
| "loss": 3.5248, |
| "step": 25600 |
| }, |
| { |
| "epoch": 2.765498652291105, |
| "grad_norm": 0.57041335105896, |
| "learning_rate": 0.00043459686994063676, |
| "loss": 3.5298, |
| "step": 25650 |
| }, |
| { |
| "epoch": 2.77088948787062, |
| "grad_norm": 0.5930079221725464, |
| "learning_rate": 0.00043427307069616837, |
| "loss": 3.5188, |
| "step": 25700 |
| }, |
| { |
| "epoch": 2.776280323450135, |
| "grad_norm": 0.5885521769523621, |
| "learning_rate": 0.0004339492714516999, |
| "loss": 3.5224, |
| "step": 25750 |
| }, |
| { |
| "epoch": 2.7816711590296497, |
| "grad_norm": 0.617006778717041, |
| "learning_rate": 0.00043362547220723147, |
| "loss": 3.5361, |
| "step": 25800 |
| }, |
| { |
| "epoch": 2.7870619946091644, |
| "grad_norm": 0.5954247713088989, |
| "learning_rate": 0.0004333016729627631, |
| "loss": 3.5215, |
| "step": 25850 |
| }, |
| { |
| "epoch": 2.7924528301886795, |
| "grad_norm": 0.5700495839118958, |
| "learning_rate": 0.0004329778737182946, |
| "loss": 3.5271, |
| "step": 25900 |
| }, |
| { |
| "epoch": 2.797843665768194, |
| "grad_norm": 0.5143706202507019, |
| "learning_rate": 0.00043265407447382623, |
| "loss": 3.5362, |
| "step": 25950 |
| }, |
| { |
| "epoch": 2.803234501347709, |
| "grad_norm": 0.6275197267532349, |
| "learning_rate": 0.0004323302752293577, |
| "loss": 3.5395, |
| "step": 26000 |
| }, |
| { |
| "epoch": 2.803234501347709, |
| "eval_accuracy": 0.3686822827667502, |
| "eval_loss": 3.5142149925231934, |
| "eval_runtime": 183.2956, |
| "eval_samples_per_second": 98.262, |
| "eval_steps_per_second": 6.143, |
| "step": 26000 |
| }, |
| { |
| "epoch": 2.8086253369272236, |
| "grad_norm": 0.6447458267211914, |
| "learning_rate": 0.00043200647598488933, |
| "loss": 3.5427, |
| "step": 26050 |
| }, |
| { |
| "epoch": 2.8140161725067383, |
| "grad_norm": 0.5703551173210144, |
| "learning_rate": 0.0004316826767404209, |
| "loss": 3.5212, |
| "step": 26100 |
| }, |
| { |
| "epoch": 2.8194070080862534, |
| "grad_norm": 0.6103937029838562, |
| "learning_rate": 0.00043136535348084183, |
| "loss": 3.53, |
| "step": 26150 |
| }, |
| { |
| "epoch": 2.824797843665768, |
| "grad_norm": 0.6109429001808167, |
| "learning_rate": 0.0004310415542363734, |
| "loss": 3.5268, |
| "step": 26200 |
| }, |
| { |
| "epoch": 2.830188679245283, |
| "grad_norm": 0.616741418838501, |
| "learning_rate": 0.000430717754991905, |
| "loss": 3.5213, |
| "step": 26250 |
| }, |
| { |
| "epoch": 2.835579514824798, |
| "grad_norm": 0.5909930467605591, |
| "learning_rate": 0.00043039395574743654, |
| "loss": 3.5342, |
| "step": 26300 |
| }, |
| { |
| "epoch": 2.8409703504043127, |
| "grad_norm": 0.6086243391036987, |
| "learning_rate": 0.00043007015650296814, |
| "loss": 3.54, |
| "step": 26350 |
| }, |
| { |
| "epoch": 2.8463611859838274, |
| "grad_norm": 0.6133831739425659, |
| "learning_rate": 0.0004297463572584997, |
| "loss": 3.5273, |
| "step": 26400 |
| }, |
| { |
| "epoch": 2.8517520215633425, |
| "grad_norm": 0.5661080479621887, |
| "learning_rate": 0.0004294225580140313, |
| "loss": 3.5292, |
| "step": 26450 |
| }, |
| { |
| "epoch": 2.857142857142857, |
| "grad_norm": 0.623001217842102, |
| "learning_rate": 0.00042909875876956285, |
| "loss": 3.5253, |
| "step": 26500 |
| }, |
| { |
| "epoch": 2.862533692722372, |
| "grad_norm": 0.5956571102142334, |
| "learning_rate": 0.0004287749595250944, |
| "loss": 3.517, |
| "step": 26550 |
| }, |
| { |
| "epoch": 2.867924528301887, |
| "grad_norm": 0.6012107729911804, |
| "learning_rate": 0.000428451160280626, |
| "loss": 3.5283, |
| "step": 26600 |
| }, |
| { |
| "epoch": 2.8733153638814017, |
| "grad_norm": 0.5730655789375305, |
| "learning_rate": 0.0004281273610361575, |
| "loss": 3.5232, |
| "step": 26650 |
| }, |
| { |
| "epoch": 2.8787061994609164, |
| "grad_norm": 0.5700306296348572, |
| "learning_rate": 0.0004278035617916891, |
| "loss": 3.5222, |
| "step": 26700 |
| }, |
| { |
| "epoch": 2.884097035040431, |
| "grad_norm": 0.5891921520233154, |
| "learning_rate": 0.00042747976254722066, |
| "loss": 3.5415, |
| "step": 26750 |
| }, |
| { |
| "epoch": 2.889487870619946, |
| "grad_norm": 0.5979741215705872, |
| "learning_rate": 0.00042715596330275226, |
| "loss": 3.5262, |
| "step": 26800 |
| }, |
| { |
| "epoch": 2.894878706199461, |
| "grad_norm": 0.5651412010192871, |
| "learning_rate": 0.0004268321640582838, |
| "loss": 3.5226, |
| "step": 26850 |
| }, |
| { |
| "epoch": 2.9002695417789757, |
| "grad_norm": 0.6269994974136353, |
| "learning_rate": 0.0004265083648138154, |
| "loss": 3.5465, |
| "step": 26900 |
| }, |
| { |
| "epoch": 2.9056603773584904, |
| "grad_norm": 0.5697436928749084, |
| "learning_rate": 0.00042618456556934697, |
| "loss": 3.5394, |
| "step": 26950 |
| }, |
| { |
| "epoch": 2.9110512129380055, |
| "grad_norm": 0.5543894171714783, |
| "learning_rate": 0.0004258607663248785, |
| "loss": 3.522, |
| "step": 27000 |
| }, |
| { |
| "epoch": 2.9110512129380055, |
| "eval_accuracy": 0.3693975447272329, |
| "eval_loss": 3.504847764968872, |
| "eval_runtime": 183.0242, |
| "eval_samples_per_second": 98.408, |
| "eval_steps_per_second": 6.152, |
| "step": 27000 |
| }, |
| { |
| "epoch": 2.91644204851752, |
| "grad_norm": 0.6050317883491516, |
| "learning_rate": 0.0004255369670804101, |
| "loss": 3.5397, |
| "step": 27050 |
| }, |
| { |
| "epoch": 2.921832884097035, |
| "grad_norm": 0.538213312625885, |
| "learning_rate": 0.0004252131678359417, |
| "loss": 3.5442, |
| "step": 27100 |
| }, |
| { |
| "epoch": 2.92722371967655, |
| "grad_norm": 0.6090800762176514, |
| "learning_rate": 0.0004248893685914733, |
| "loss": 3.5282, |
| "step": 27150 |
| }, |
| { |
| "epoch": 2.9326145552560647, |
| "grad_norm": 0.6525107026100159, |
| "learning_rate": 0.00042456556934700483, |
| "loss": 3.519, |
| "step": 27200 |
| }, |
| { |
| "epoch": 2.9380053908355794, |
| "grad_norm": 0.5758708119392395, |
| "learning_rate": 0.00042424177010253643, |
| "loss": 3.5319, |
| "step": 27250 |
| }, |
| { |
| "epoch": 2.9433962264150946, |
| "grad_norm": 0.6072579026222229, |
| "learning_rate": 0.00042391797085806793, |
| "loss": 3.5283, |
| "step": 27300 |
| }, |
| { |
| "epoch": 2.9487870619946093, |
| "grad_norm": 0.5943818092346191, |
| "learning_rate": 0.0004235941716135995, |
| "loss": 3.5201, |
| "step": 27350 |
| }, |
| { |
| "epoch": 2.954177897574124, |
| "grad_norm": 0.5442832112312317, |
| "learning_rate": 0.0004232703723691311, |
| "loss": 3.5261, |
| "step": 27400 |
| }, |
| { |
| "epoch": 2.9595687331536387, |
| "grad_norm": 0.5325363874435425, |
| "learning_rate": 0.00042294657312466264, |
| "loss": 3.5216, |
| "step": 27450 |
| }, |
| { |
| "epoch": 2.964959568733154, |
| "grad_norm": 0.5526369214057922, |
| "learning_rate": 0.00042262277388019424, |
| "loss": 3.5286, |
| "step": 27500 |
| }, |
| { |
| "epoch": 2.9703504043126685, |
| "grad_norm": 0.5498561263084412, |
| "learning_rate": 0.0004222989746357258, |
| "loss": 3.5312, |
| "step": 27550 |
| }, |
| { |
| "epoch": 2.975741239892183, |
| "grad_norm": 0.5401663780212402, |
| "learning_rate": 0.0004219751753912574, |
| "loss": 3.5316, |
| "step": 27600 |
| }, |
| { |
| "epoch": 2.981132075471698, |
| "grad_norm": 0.5896415114402771, |
| "learning_rate": 0.00042165137614678895, |
| "loss": 3.5335, |
| "step": 27650 |
| }, |
| { |
| "epoch": 2.986522911051213, |
| "grad_norm": 0.6067506670951843, |
| "learning_rate": 0.00042132757690232055, |
| "loss": 3.5246, |
| "step": 27700 |
| }, |
| { |
| "epoch": 2.9919137466307277, |
| "grad_norm": 0.5816178321838379, |
| "learning_rate": 0.0004210037776578521, |
| "loss": 3.5349, |
| "step": 27750 |
| }, |
| { |
| "epoch": 2.9973045822102424, |
| "grad_norm": 0.6244997382164001, |
| "learning_rate": 0.00042067997841338365, |
| "loss": 3.5308, |
| "step": 27800 |
| }, |
| { |
| "epoch": 3.0026954177897576, |
| "grad_norm": 0.602767825126648, |
| "learning_rate": 0.00042035617916891526, |
| "loss": 3.4765, |
| "step": 27850 |
| }, |
| { |
| "epoch": 3.0080862533692723, |
| "grad_norm": 0.6553490161895752, |
| "learning_rate": 0.0004200323799244468, |
| "loss": 3.4268, |
| "step": 27900 |
| }, |
| { |
| "epoch": 3.013477088948787, |
| "grad_norm": 0.6417362689971924, |
| "learning_rate": 0.0004197085806799784, |
| "loss": 3.4265, |
| "step": 27950 |
| }, |
| { |
| "epoch": 3.018867924528302, |
| "grad_norm": 0.6044764518737793, |
| "learning_rate": 0.0004193847814355099, |
| "loss": 3.4455, |
| "step": 28000 |
| }, |
| { |
| "epoch": 3.018867924528302, |
| "eval_accuracy": 0.37052916455382073, |
| "eval_loss": 3.5026047229766846, |
| "eval_runtime": 183.178, |
| "eval_samples_per_second": 98.325, |
| "eval_steps_per_second": 6.147, |
| "step": 28000 |
| }, |
| { |
| "epoch": 3.024258760107817, |
| "grad_norm": 0.6100830435752869, |
| "learning_rate": 0.0004190609821910415, |
| "loss": 3.4397, |
| "step": 28050 |
| }, |
| { |
| "epoch": 3.0296495956873315, |
| "grad_norm": 0.5654774904251099, |
| "learning_rate": 0.00041873718294657306, |
| "loss": 3.4469, |
| "step": 28100 |
| }, |
| { |
| "epoch": 3.035040431266846, |
| "grad_norm": 0.5853551030158997, |
| "learning_rate": 0.000418419859686994, |
| "loss": 3.4362, |
| "step": 28150 |
| }, |
| { |
| "epoch": 3.0404312668463613, |
| "grad_norm": 0.6212331652641296, |
| "learning_rate": 0.00041809606044252557, |
| "loss": 3.442, |
| "step": 28200 |
| }, |
| { |
| "epoch": 3.045822102425876, |
| "grad_norm": 0.5993689894676208, |
| "learning_rate": 0.00041777226119805717, |
| "loss": 3.4229, |
| "step": 28250 |
| }, |
| { |
| "epoch": 3.0512129380053907, |
| "grad_norm": 0.6013529300689697, |
| "learning_rate": 0.0004174484619535887, |
| "loss": 3.4456, |
| "step": 28300 |
| }, |
| { |
| "epoch": 3.056603773584906, |
| "grad_norm": 0.599606990814209, |
| "learning_rate": 0.00041712466270912033, |
| "loss": 3.4189, |
| "step": 28350 |
| }, |
| { |
| "epoch": 3.0619946091644206, |
| "grad_norm": 0.5922053456306458, |
| "learning_rate": 0.0004168008634646519, |
| "loss": 3.4486, |
| "step": 28400 |
| }, |
| { |
| "epoch": 3.0673854447439353, |
| "grad_norm": 0.6180664896965027, |
| "learning_rate": 0.0004164770642201835, |
| "loss": 3.4173, |
| "step": 28450 |
| }, |
| { |
| "epoch": 3.07277628032345, |
| "grad_norm": 0.6056652665138245, |
| "learning_rate": 0.00041615326497571503, |
| "loss": 3.4427, |
| "step": 28500 |
| }, |
| { |
| "epoch": 3.078167115902965, |
| "grad_norm": 0.6129570603370667, |
| "learning_rate": 0.0004158294657312466, |
| "loss": 3.4542, |
| "step": 28550 |
| }, |
| { |
| "epoch": 3.08355795148248, |
| "grad_norm": 0.6358380317687988, |
| "learning_rate": 0.0004155056664867782, |
| "loss": 3.4405, |
| "step": 28600 |
| }, |
| { |
| "epoch": 3.0889487870619945, |
| "grad_norm": 0.5838945508003235, |
| "learning_rate": 0.0004151818672423097, |
| "loss": 3.4519, |
| "step": 28650 |
| }, |
| { |
| "epoch": 3.0943396226415096, |
| "grad_norm": 0.5569682121276855, |
| "learning_rate": 0.0004148580679978413, |
| "loss": 3.4472, |
| "step": 28700 |
| }, |
| { |
| "epoch": 3.0997304582210243, |
| "grad_norm": 0.6249170899391174, |
| "learning_rate": 0.00041453426875337284, |
| "loss": 3.4719, |
| "step": 28750 |
| }, |
| { |
| "epoch": 3.105121293800539, |
| "grad_norm": 0.5985543727874756, |
| "learning_rate": 0.00041421046950890445, |
| "loss": 3.433, |
| "step": 28800 |
| }, |
| { |
| "epoch": 3.1105121293800537, |
| "grad_norm": 0.6323471665382385, |
| "learning_rate": 0.000413886670264436, |
| "loss": 3.4428, |
| "step": 28850 |
| }, |
| { |
| "epoch": 3.115902964959569, |
| "grad_norm": 0.5465901494026184, |
| "learning_rate": 0.0004135628710199676, |
| "loss": 3.4372, |
| "step": 28900 |
| }, |
| { |
| "epoch": 3.1212938005390836, |
| "grad_norm": 0.6596279144287109, |
| "learning_rate": 0.00041323907177549915, |
| "loss": 3.4682, |
| "step": 28950 |
| }, |
| { |
| "epoch": 3.1266846361185983, |
| "grad_norm": 0.6128025054931641, |
| "learning_rate": 0.0004129152725310307, |
| "loss": 3.4399, |
| "step": 29000 |
| }, |
| { |
| "epoch": 3.1266846361185983, |
| "eval_accuracy": 0.37060902442636434, |
| "eval_loss": 3.4956202507019043, |
| "eval_runtime": 183.1389, |
| "eval_samples_per_second": 98.346, |
| "eval_steps_per_second": 6.148, |
| "step": 29000 |
| }, |
| { |
| "epoch": 3.1320754716981134, |
| "grad_norm": 0.621014416217804, |
| "learning_rate": 0.0004125914732865623, |
| "loss": 3.4579, |
| "step": 29050 |
| }, |
| { |
| "epoch": 3.137466307277628, |
| "grad_norm": 0.576972246170044, |
| "learning_rate": 0.00041226767404209386, |
| "loss": 3.4598, |
| "step": 29100 |
| }, |
| { |
| "epoch": 3.142857142857143, |
| "grad_norm": 0.6160667538642883, |
| "learning_rate": 0.00041194387479762546, |
| "loss": 3.4504, |
| "step": 29150 |
| }, |
| { |
| "epoch": 3.1482479784366575, |
| "grad_norm": 0.5812036395072937, |
| "learning_rate": 0.000411620075553157, |
| "loss": 3.4588, |
| "step": 29200 |
| }, |
| { |
| "epoch": 3.1536388140161726, |
| "grad_norm": 0.5818783640861511, |
| "learning_rate": 0.0004112962763086886, |
| "loss": 3.4629, |
| "step": 29250 |
| }, |
| { |
| "epoch": 3.1590296495956873, |
| "grad_norm": 0.6303808689117432, |
| "learning_rate": 0.0004109724770642201, |
| "loss": 3.4655, |
| "step": 29300 |
| }, |
| { |
| "epoch": 3.164420485175202, |
| "grad_norm": 0.6596410870552063, |
| "learning_rate": 0.00041064867781975177, |
| "loss": 3.4566, |
| "step": 29350 |
| }, |
| { |
| "epoch": 3.169811320754717, |
| "grad_norm": 0.581000030040741, |
| "learning_rate": 0.00041032487857528327, |
| "loss": 3.4488, |
| "step": 29400 |
| }, |
| { |
| "epoch": 3.175202156334232, |
| "grad_norm": 0.6004059314727783, |
| "learning_rate": 0.0004100010793308148, |
| "loss": 3.4341, |
| "step": 29450 |
| }, |
| { |
| "epoch": 3.1805929919137466, |
| "grad_norm": 0.5809293985366821, |
| "learning_rate": 0.0004096772800863464, |
| "loss": 3.4601, |
| "step": 29500 |
| }, |
| { |
| "epoch": 3.1859838274932613, |
| "grad_norm": 0.6131229400634766, |
| "learning_rate": 0.000409353480841878, |
| "loss": 3.4449, |
| "step": 29550 |
| }, |
| { |
| "epoch": 3.1913746630727764, |
| "grad_norm": 0.5700172185897827, |
| "learning_rate": 0.0004090296815974096, |
| "loss": 3.4522, |
| "step": 29600 |
| }, |
| { |
| "epoch": 3.196765498652291, |
| "grad_norm": 0.6279261112213135, |
| "learning_rate": 0.00040870588235294113, |
| "loss": 3.468, |
| "step": 29650 |
| }, |
| { |
| "epoch": 3.202156334231806, |
| "grad_norm": 0.6200495362281799, |
| "learning_rate": 0.00040838208310847273, |
| "loss": 3.4512, |
| "step": 29700 |
| }, |
| { |
| "epoch": 3.207547169811321, |
| "grad_norm": 0.5797140598297119, |
| "learning_rate": 0.0004080582838640043, |
| "loss": 3.4619, |
| "step": 29750 |
| }, |
| { |
| "epoch": 3.2129380053908356, |
| "grad_norm": 0.6226954460144043, |
| "learning_rate": 0.0004077344846195359, |
| "loss": 3.4709, |
| "step": 29800 |
| }, |
| { |
| "epoch": 3.2183288409703503, |
| "grad_norm": 0.5723651647567749, |
| "learning_rate": 0.00040741068537506744, |
| "loss": 3.4755, |
| "step": 29850 |
| }, |
| { |
| "epoch": 3.223719676549865, |
| "grad_norm": 0.5869478583335876, |
| "learning_rate": 0.000407086886130599, |
| "loss": 3.4414, |
| "step": 29900 |
| }, |
| { |
| "epoch": 3.22911051212938, |
| "grad_norm": 0.5735659599304199, |
| "learning_rate": 0.0004067630868861306, |
| "loss": 3.4457, |
| "step": 29950 |
| }, |
| { |
| "epoch": 3.234501347708895, |
| "grad_norm": 0.5913712978363037, |
| "learning_rate": 0.0004064392876416621, |
| "loss": 3.4688, |
| "step": 30000 |
| }, |
| { |
| "epoch": 3.234501347708895, |
| "eval_accuracy": 0.37179301494487876, |
| "eval_loss": 3.491060256958008, |
| "eval_runtime": 183.4022, |
| "eval_samples_per_second": 98.205, |
| "eval_steps_per_second": 6.14, |
| "step": 30000 |
| }, |
| { |
| "epoch": 3.2398921832884096, |
| "grad_norm": 0.5873755812644958, |
| "learning_rate": 0.0004061154883971937, |
| "loss": 3.4555, |
| "step": 30050 |
| }, |
| { |
| "epoch": 3.2452830188679247, |
| "grad_norm": 0.5795559883117676, |
| "learning_rate": 0.00040579168915272525, |
| "loss": 3.4611, |
| "step": 30100 |
| }, |
| { |
| "epoch": 3.2506738544474394, |
| "grad_norm": 0.5841328501701355, |
| "learning_rate": 0.0004054743658931462, |
| "loss": 3.468, |
| "step": 30150 |
| }, |
| { |
| "epoch": 3.256064690026954, |
| "grad_norm": 0.6134114265441895, |
| "learning_rate": 0.00040515056664867775, |
| "loss": 3.4516, |
| "step": 30200 |
| }, |
| { |
| "epoch": 3.2614555256064692, |
| "grad_norm": 0.6624637842178345, |
| "learning_rate": 0.00040482676740420935, |
| "loss": 3.4571, |
| "step": 30250 |
| }, |
| { |
| "epoch": 3.266846361185984, |
| "grad_norm": 0.61444491147995, |
| "learning_rate": 0.0004045029681597409, |
| "loss": 3.4429, |
| "step": 30300 |
| }, |
| { |
| "epoch": 3.2722371967654986, |
| "grad_norm": 0.6057273745536804, |
| "learning_rate": 0.0004041791689152725, |
| "loss": 3.4574, |
| "step": 30350 |
| }, |
| { |
| "epoch": 3.2776280323450133, |
| "grad_norm": 0.5823548436164856, |
| "learning_rate": 0.00040385536967080406, |
| "loss": 3.468, |
| "step": 30400 |
| }, |
| { |
| "epoch": 3.2830188679245285, |
| "grad_norm": 0.5908448696136475, |
| "learning_rate": 0.00040353157042633567, |
| "loss": 3.4597, |
| "step": 30450 |
| }, |
| { |
| "epoch": 3.288409703504043, |
| "grad_norm": 0.6395969986915588, |
| "learning_rate": 0.0004032077711818672, |
| "loss": 3.4795, |
| "step": 30500 |
| }, |
| { |
| "epoch": 3.293800539083558, |
| "grad_norm": 0.5815283060073853, |
| "learning_rate": 0.0004028839719373988, |
| "loss": 3.4691, |
| "step": 30550 |
| }, |
| { |
| "epoch": 3.2991913746630726, |
| "grad_norm": 0.5855560302734375, |
| "learning_rate": 0.00040256017269293037, |
| "loss": 3.4623, |
| "step": 30600 |
| }, |
| { |
| "epoch": 3.3045822102425877, |
| "grad_norm": 0.5978612899780273, |
| "learning_rate": 0.00040223637344846187, |
| "loss": 3.4608, |
| "step": 30650 |
| }, |
| { |
| "epoch": 3.3099730458221024, |
| "grad_norm": 0.635847270488739, |
| "learning_rate": 0.00040191257420399347, |
| "loss": 3.4523, |
| "step": 30700 |
| }, |
| { |
| "epoch": 3.315363881401617, |
| "grad_norm": 0.6061700582504272, |
| "learning_rate": 0.000401588774959525, |
| "loss": 3.4582, |
| "step": 30750 |
| }, |
| { |
| "epoch": 3.3207547169811322, |
| "grad_norm": 0.6008456945419312, |
| "learning_rate": 0.00040126497571505663, |
| "loss": 3.46, |
| "step": 30800 |
| }, |
| { |
| "epoch": 3.326145552560647, |
| "grad_norm": 0.5615379214286804, |
| "learning_rate": 0.0004009411764705882, |
| "loss": 3.4519, |
| "step": 30850 |
| }, |
| { |
| "epoch": 3.3315363881401616, |
| "grad_norm": 0.6048398017883301, |
| "learning_rate": 0.0004006173772261198, |
| "loss": 3.4741, |
| "step": 30900 |
| }, |
| { |
| "epoch": 3.3369272237196768, |
| "grad_norm": 0.5585691928863525, |
| "learning_rate": 0.00040029357798165133, |
| "loss": 3.4571, |
| "step": 30950 |
| }, |
| { |
| "epoch": 3.3423180592991915, |
| "grad_norm": 0.5774768590927124, |
| "learning_rate": 0.00039996977873718294, |
| "loss": 3.4748, |
| "step": 31000 |
| }, |
| { |
| "epoch": 3.3423180592991915, |
| "eval_accuracy": 0.37259172232320303, |
| "eval_loss": 3.483452320098877, |
| "eval_runtime": 183.6331, |
| "eval_samples_per_second": 98.081, |
| "eval_steps_per_second": 6.132, |
| "step": 31000 |
| }, |
| { |
| "epoch": 3.347708894878706, |
| "grad_norm": 0.6331956386566162, |
| "learning_rate": 0.0003996459794927145, |
| "loss": 3.4485, |
| "step": 31050 |
| }, |
| { |
| "epoch": 3.353099730458221, |
| "grad_norm": 0.6012781262397766, |
| "learning_rate": 0.00039932218024824604, |
| "loss": 3.4705, |
| "step": 31100 |
| }, |
| { |
| "epoch": 3.358490566037736, |
| "grad_norm": 0.6027830839157104, |
| "learning_rate": 0.00039899838100377764, |
| "loss": 3.4493, |
| "step": 31150 |
| }, |
| { |
| "epoch": 3.3638814016172507, |
| "grad_norm": 0.5623087286949158, |
| "learning_rate": 0.0003986745817593092, |
| "loss": 3.4515, |
| "step": 31200 |
| }, |
| { |
| "epoch": 3.3692722371967654, |
| "grad_norm": 0.6207774877548218, |
| "learning_rate": 0.0003983507825148408, |
| "loss": 3.4675, |
| "step": 31250 |
| }, |
| { |
| "epoch": 3.37466307277628, |
| "grad_norm": 0.5530596375465393, |
| "learning_rate": 0.00039802698327037235, |
| "loss": 3.4544, |
| "step": 31300 |
| }, |
| { |
| "epoch": 3.3800539083557952, |
| "grad_norm": 0.600220799446106, |
| "learning_rate": 0.00039770318402590396, |
| "loss": 3.4418, |
| "step": 31350 |
| }, |
| { |
| "epoch": 3.38544474393531, |
| "grad_norm": 0.5809213519096375, |
| "learning_rate": 0.00039737938478143545, |
| "loss": 3.4483, |
| "step": 31400 |
| }, |
| { |
| "epoch": 3.3908355795148246, |
| "grad_norm": 0.5642971992492676, |
| "learning_rate": 0.000397055585536967, |
| "loss": 3.4688, |
| "step": 31450 |
| }, |
| { |
| "epoch": 3.3962264150943398, |
| "grad_norm": 0.6100688576698303, |
| "learning_rate": 0.0003967317862924986, |
| "loss": 3.4687, |
| "step": 31500 |
| }, |
| { |
| "epoch": 3.4016172506738545, |
| "grad_norm": 0.6187987327575684, |
| "learning_rate": 0.00039641446303291956, |
| "loss": 3.455, |
| "step": 31550 |
| }, |
| { |
| "epoch": 3.407008086253369, |
| "grad_norm": 0.5841171145439148, |
| "learning_rate": 0.0003960906637884511, |
| "loss": 3.442, |
| "step": 31600 |
| }, |
| { |
| "epoch": 3.4123989218328843, |
| "grad_norm": 0.6319450736045837, |
| "learning_rate": 0.0003957668645439827, |
| "loss": 3.4622, |
| "step": 31650 |
| }, |
| { |
| "epoch": 3.417789757412399, |
| "grad_norm": 0.6497536897659302, |
| "learning_rate": 0.00039544306529951426, |
| "loss": 3.4676, |
| "step": 31700 |
| }, |
| { |
| "epoch": 3.4231805929919137, |
| "grad_norm": 0.567626416683197, |
| "learning_rate": 0.00039511926605504587, |
| "loss": 3.4645, |
| "step": 31750 |
| }, |
| { |
| "epoch": 3.4285714285714284, |
| "grad_norm": 0.6872853636741638, |
| "learning_rate": 0.0003947954668105774, |
| "loss": 3.4514, |
| "step": 31800 |
| }, |
| { |
| "epoch": 3.4339622641509435, |
| "grad_norm": 0.5771532654762268, |
| "learning_rate": 0.00039447166756610897, |
| "loss": 3.4551, |
| "step": 31850 |
| }, |
| { |
| "epoch": 3.439353099730458, |
| "grad_norm": 0.6318898797035217, |
| "learning_rate": 0.0003941478683216406, |
| "loss": 3.4651, |
| "step": 31900 |
| }, |
| { |
| "epoch": 3.444743935309973, |
| "grad_norm": 0.5701496601104736, |
| "learning_rate": 0.0003938240690771721, |
| "loss": 3.4507, |
| "step": 31950 |
| }, |
| { |
| "epoch": 3.450134770889488, |
| "grad_norm": 0.6465626955032349, |
| "learning_rate": 0.00039350026983270373, |
| "loss": 3.4633, |
| "step": 32000 |
| }, |
| { |
| "epoch": 3.450134770889488, |
| "eval_accuracy": 0.37326808654984817, |
| "eval_loss": 3.476710319519043, |
| "eval_runtime": 183.4161, |
| "eval_samples_per_second": 98.197, |
| "eval_steps_per_second": 6.139, |
| "step": 32000 |
| }, |
| { |
| "epoch": 3.4555256064690028, |
| "grad_norm": 0.5992223024368286, |
| "learning_rate": 0.00039317647058823523, |
| "loss": 3.4851, |
| "step": 32050 |
| }, |
| { |
| "epoch": 3.4609164420485174, |
| "grad_norm": 0.6115779876708984, |
| "learning_rate": 0.00039285267134376683, |
| "loss": 3.4755, |
| "step": 32100 |
| }, |
| { |
| "epoch": 3.466307277628032, |
| "grad_norm": 0.5862306952476501, |
| "learning_rate": 0.0003925288720992984, |
| "loss": 3.4469, |
| "step": 32150 |
| }, |
| { |
| "epoch": 3.4716981132075473, |
| "grad_norm": 0.5586680769920349, |
| "learning_rate": 0.00039220507285482993, |
| "loss": 3.4614, |
| "step": 32200 |
| }, |
| { |
| "epoch": 3.477088948787062, |
| "grad_norm": 0.5931172370910645, |
| "learning_rate": 0.00039188127361036154, |
| "loss": 3.4505, |
| "step": 32250 |
| }, |
| { |
| "epoch": 3.4824797843665767, |
| "grad_norm": 0.6547418236732483, |
| "learning_rate": 0.0003915574743658931, |
| "loss": 3.4545, |
| "step": 32300 |
| }, |
| { |
| "epoch": 3.487870619946092, |
| "grad_norm": 0.5955836176872253, |
| "learning_rate": 0.0003912336751214247, |
| "loss": 3.4467, |
| "step": 32350 |
| }, |
| { |
| "epoch": 3.4932614555256065, |
| "grad_norm": 0.6000773906707764, |
| "learning_rate": 0.00039090987587695624, |
| "loss": 3.4527, |
| "step": 32400 |
| }, |
| { |
| "epoch": 3.498652291105121, |
| "grad_norm": 0.6324712038040161, |
| "learning_rate": 0.00039058607663248785, |
| "loss": 3.4604, |
| "step": 32450 |
| }, |
| { |
| "epoch": 3.5040431266846364, |
| "grad_norm": 0.6126226782798767, |
| "learning_rate": 0.0003902622773880194, |
| "loss": 3.4673, |
| "step": 32500 |
| }, |
| { |
| "epoch": 3.509433962264151, |
| "grad_norm": 0.5712622404098511, |
| "learning_rate": 0.000389938478143551, |
| "loss": 3.4675, |
| "step": 32550 |
| }, |
| { |
| "epoch": 3.5148247978436657, |
| "grad_norm": 0.5935483574867249, |
| "learning_rate": 0.00038961467889908255, |
| "loss": 3.4556, |
| "step": 32600 |
| }, |
| { |
| "epoch": 3.5202156334231804, |
| "grad_norm": 0.5569974184036255, |
| "learning_rate": 0.00038929087965461405, |
| "loss": 3.4606, |
| "step": 32650 |
| }, |
| { |
| "epoch": 3.525606469002695, |
| "grad_norm": 0.6109350919723511, |
| "learning_rate": 0.00038896708041014566, |
| "loss": 3.4762, |
| "step": 32700 |
| }, |
| { |
| "epoch": 3.5309973045822103, |
| "grad_norm": 0.6272343397140503, |
| "learning_rate": 0.0003886432811656772, |
| "loss": 3.4494, |
| "step": 32750 |
| }, |
| { |
| "epoch": 3.536388140161725, |
| "grad_norm": 0.5934761166572571, |
| "learning_rate": 0.0003883194819212088, |
| "loss": 3.4775, |
| "step": 32800 |
| }, |
| { |
| "epoch": 3.5417789757412397, |
| "grad_norm": 0.5385004281997681, |
| "learning_rate": 0.00038799568267674036, |
| "loss": 3.4781, |
| "step": 32850 |
| }, |
| { |
| "epoch": 3.547169811320755, |
| "grad_norm": 0.5813574194908142, |
| "learning_rate": 0.00038767188343227197, |
| "loss": 3.4708, |
| "step": 32900 |
| }, |
| { |
| "epoch": 3.5525606469002695, |
| "grad_norm": 0.614341676235199, |
| "learning_rate": 0.0003873480841878035, |
| "loss": 3.4577, |
| "step": 32950 |
| }, |
| { |
| "epoch": 3.557951482479784, |
| "grad_norm": 0.6198434829711914, |
| "learning_rate": 0.0003870242849433351, |
| "loss": 3.4541, |
| "step": 33000 |
| }, |
| { |
| "epoch": 3.557951482479784, |
| "eval_accuracy": 0.3734748529953591, |
| "eval_loss": 3.4712295532226562, |
| "eval_runtime": 183.2674, |
| "eval_samples_per_second": 98.277, |
| "eval_steps_per_second": 6.144, |
| "step": 33000 |
| }, |
| { |
| "epoch": 3.5633423180592994, |
| "grad_norm": 0.6407135725021362, |
| "learning_rate": 0.00038670048569886667, |
| "loss": 3.4659, |
| "step": 33050 |
| }, |
| { |
| "epoch": 3.568733153638814, |
| "grad_norm": 0.6006309986114502, |
| "learning_rate": 0.0003863766864543982, |
| "loss": 3.4664, |
| "step": 33100 |
| }, |
| { |
| "epoch": 3.5741239892183287, |
| "grad_norm": 0.6235687136650085, |
| "learning_rate": 0.00038605288720992983, |
| "loss": 3.4704, |
| "step": 33150 |
| }, |
| { |
| "epoch": 3.579514824797844, |
| "grad_norm": 0.6336178183555603, |
| "learning_rate": 0.0003857290879654614, |
| "loss": 3.465, |
| "step": 33200 |
| }, |
| { |
| "epoch": 3.5849056603773586, |
| "grad_norm": 0.571612536907196, |
| "learning_rate": 0.000385405288720993, |
| "loss": 3.4488, |
| "step": 33250 |
| }, |
| { |
| "epoch": 3.5902964959568733, |
| "grad_norm": 0.6101464033126831, |
| "learning_rate": 0.00038508148947652453, |
| "loss": 3.4767, |
| "step": 33300 |
| }, |
| { |
| "epoch": 3.595687331536388, |
| "grad_norm": 0.6197885870933533, |
| "learning_rate": 0.00038475769023205614, |
| "loss": 3.4633, |
| "step": 33350 |
| }, |
| { |
| "epoch": 3.601078167115903, |
| "grad_norm": 0.6503936648368835, |
| "learning_rate": 0.00038443389098758763, |
| "loss": 3.4601, |
| "step": 33400 |
| }, |
| { |
| "epoch": 3.606469002695418, |
| "grad_norm": 0.6410123705863953, |
| "learning_rate": 0.00038411009174311924, |
| "loss": 3.4757, |
| "step": 33450 |
| }, |
| { |
| "epoch": 3.6118598382749325, |
| "grad_norm": 0.5860520601272583, |
| "learning_rate": 0.0003837862924986508, |
| "loss": 3.4528, |
| "step": 33500 |
| }, |
| { |
| "epoch": 3.617250673854447, |
| "grad_norm": 0.6135208606719971, |
| "learning_rate": 0.00038346249325418234, |
| "loss": 3.4515, |
| "step": 33550 |
| }, |
| { |
| "epoch": 3.6226415094339623, |
| "grad_norm": 0.6425701379776001, |
| "learning_rate": 0.00038313869400971395, |
| "loss": 3.447, |
| "step": 33600 |
| }, |
| { |
| "epoch": 3.628032345013477, |
| "grad_norm": 0.6353036761283875, |
| "learning_rate": 0.0003828148947652455, |
| "loss": 3.4633, |
| "step": 33650 |
| }, |
| { |
| "epoch": 3.6334231805929917, |
| "grad_norm": 0.621608555316925, |
| "learning_rate": 0.0003824910955207771, |
| "loss": 3.4674, |
| "step": 33700 |
| }, |
| { |
| "epoch": 3.638814016172507, |
| "grad_norm": 0.6142898797988892, |
| "learning_rate": 0.00038216729627630865, |
| "loss": 3.4484, |
| "step": 33750 |
| }, |
| { |
| "epoch": 3.6442048517520216, |
| "grad_norm": 0.6292492151260376, |
| "learning_rate": 0.00038184349703184026, |
| "loss": 3.4527, |
| "step": 33800 |
| }, |
| { |
| "epoch": 3.6495956873315363, |
| "grad_norm": 0.6098408699035645, |
| "learning_rate": 0.0003815196977873718, |
| "loss": 3.4381, |
| "step": 33850 |
| }, |
| { |
| "epoch": 3.6549865229110514, |
| "grad_norm": 0.6431405544281006, |
| "learning_rate": 0.0003811958985429034, |
| "loss": 3.4306, |
| "step": 33900 |
| }, |
| { |
| "epoch": 3.660377358490566, |
| "grad_norm": 0.6275612711906433, |
| "learning_rate": 0.00038087209929843496, |
| "loss": 3.4561, |
| "step": 33950 |
| }, |
| { |
| "epoch": 3.665768194070081, |
| "grad_norm": 0.6220306754112244, |
| "learning_rate": 0.00038054830005396646, |
| "loss": 3.4513, |
| "step": 34000 |
| }, |
| { |
| "epoch": 3.665768194070081, |
| "eval_accuracy": 0.37447750184411116, |
| "eval_loss": 3.4639649391174316, |
| "eval_runtime": 183.3787, |
| "eval_samples_per_second": 98.218, |
| "eval_steps_per_second": 6.14, |
| "step": 34000 |
| }, |
| { |
| "epoch": 3.671159029649596, |
| "grad_norm": 0.6661257147789001, |
| "learning_rate": 0.00038022450080949806, |
| "loss": 3.456, |
| "step": 34050 |
| }, |
| { |
| "epoch": 3.6765498652291106, |
| "grad_norm": 0.6361551880836487, |
| "learning_rate": 0.0003799007015650296, |
| "loss": 3.4529, |
| "step": 34100 |
| }, |
| { |
| "epoch": 3.6819407008086253, |
| "grad_norm": 0.5787711143493652, |
| "learning_rate": 0.0003795769023205612, |
| "loss": 3.4494, |
| "step": 34150 |
| }, |
| { |
| "epoch": 3.68733153638814, |
| "grad_norm": 0.691253125667572, |
| "learning_rate": 0.00037925310307609277, |
| "loss": 3.4608, |
| "step": 34200 |
| }, |
| { |
| "epoch": 3.6927223719676547, |
| "grad_norm": 0.6856685280799866, |
| "learning_rate": 0.0003789293038316244, |
| "loss": 3.4495, |
| "step": 34250 |
| }, |
| { |
| "epoch": 3.69811320754717, |
| "grad_norm": 0.6148468255996704, |
| "learning_rate": 0.0003786055045871559, |
| "loss": 3.455, |
| "step": 34300 |
| }, |
| { |
| "epoch": 3.7035040431266846, |
| "grad_norm": 0.6273311376571655, |
| "learning_rate": 0.0003782817053426875, |
| "loss": 3.4524, |
| "step": 34350 |
| }, |
| { |
| "epoch": 3.7088948787061993, |
| "grad_norm": 0.6452623009681702, |
| "learning_rate": 0.0003779579060982191, |
| "loss": 3.4595, |
| "step": 34400 |
| }, |
| { |
| "epoch": 3.7142857142857144, |
| "grad_norm": 0.6160526871681213, |
| "learning_rate": 0.00037763410685375063, |
| "loss": 3.4324, |
| "step": 34450 |
| }, |
| { |
| "epoch": 3.719676549865229, |
| "grad_norm": 0.6273506879806519, |
| "learning_rate": 0.00037731030760928223, |
| "loss": 3.473, |
| "step": 34500 |
| }, |
| { |
| "epoch": 3.725067385444744, |
| "grad_norm": 0.6229552626609802, |
| "learning_rate": 0.0003769865083648138, |
| "loss": 3.4495, |
| "step": 34550 |
| }, |
| { |
| "epoch": 3.730458221024259, |
| "grad_norm": 0.6476499438285828, |
| "learning_rate": 0.0003766627091203454, |
| "loss": 3.455, |
| "step": 34600 |
| }, |
| { |
| "epoch": 3.7358490566037736, |
| "grad_norm": 0.603552520275116, |
| "learning_rate": 0.00037633890987587694, |
| "loss": 3.4588, |
| "step": 34650 |
| }, |
| { |
| "epoch": 3.7412398921832883, |
| "grad_norm": 0.611316978931427, |
| "learning_rate": 0.00037601511063140855, |
| "loss": 3.4654, |
| "step": 34700 |
| }, |
| { |
| "epoch": 3.7466307277628035, |
| "grad_norm": 0.6224241852760315, |
| "learning_rate": 0.00037569131138694004, |
| "loss": 3.4704, |
| "step": 34750 |
| }, |
| { |
| "epoch": 3.752021563342318, |
| "grad_norm": 0.5943573117256165, |
| "learning_rate": 0.0003753675121424716, |
| "loss": 3.4547, |
| "step": 34800 |
| }, |
| { |
| "epoch": 3.757412398921833, |
| "grad_norm": 0.6184179782867432, |
| "learning_rate": 0.0003750437128980032, |
| "loss": 3.4624, |
| "step": 34850 |
| }, |
| { |
| "epoch": 3.7628032345013476, |
| "grad_norm": 0.6081331968307495, |
| "learning_rate": 0.00037471991365353475, |
| "loss": 3.4541, |
| "step": 34900 |
| }, |
| { |
| "epoch": 3.7681940700808623, |
| "grad_norm": 0.6248754262924194, |
| "learning_rate": 0.00037439611440906635, |
| "loss": 3.4487, |
| "step": 34950 |
| }, |
| { |
| "epoch": 3.7735849056603774, |
| "grad_norm": 0.6937313079833984, |
| "learning_rate": 0.0003740723151645979, |
| "loss": 3.4707, |
| "step": 35000 |
| }, |
| { |
| "epoch": 3.7735849056603774, |
| "eval_accuracy": 0.3750037077797967, |
| "eval_loss": 3.4571778774261475, |
| "eval_runtime": 183.6889, |
| "eval_samples_per_second": 98.052, |
| "eval_steps_per_second": 6.13, |
| "step": 35000 |
| }, |
| { |
| "epoch": 3.778975741239892, |
| "grad_norm": 0.6209198236465454, |
| "learning_rate": 0.0003737485159201295, |
| "loss": 3.4533, |
| "step": 35050 |
| }, |
| { |
| "epoch": 3.784366576819407, |
| "grad_norm": 0.6297261118888855, |
| "learning_rate": 0.00037342471667566106, |
| "loss": 3.4573, |
| "step": 35100 |
| }, |
| { |
| "epoch": 3.789757412398922, |
| "grad_norm": 0.6521171927452087, |
| "learning_rate": 0.00037310091743119266, |
| "loss": 3.4427, |
| "step": 35150 |
| }, |
| { |
| "epoch": 3.7951482479784366, |
| "grad_norm": 0.6195915341377258, |
| "learning_rate": 0.0003727771181867242, |
| "loss": 3.4488, |
| "step": 35200 |
| }, |
| { |
| "epoch": 3.8005390835579513, |
| "grad_norm": 0.6207437515258789, |
| "learning_rate": 0.00037245331894225576, |
| "loss": 3.455, |
| "step": 35250 |
| }, |
| { |
| "epoch": 3.8059299191374665, |
| "grad_norm": 0.6468321681022644, |
| "learning_rate": 0.00037212951969778737, |
| "loss": 3.4479, |
| "step": 35300 |
| }, |
| { |
| "epoch": 3.811320754716981, |
| "grad_norm": 0.6130130290985107, |
| "learning_rate": 0.00037180572045331887, |
| "loss": 3.469, |
| "step": 35350 |
| }, |
| { |
| "epoch": 3.816711590296496, |
| "grad_norm": 0.5929135084152222, |
| "learning_rate": 0.00037148192120885047, |
| "loss": 3.4505, |
| "step": 35400 |
| }, |
| { |
| "epoch": 3.822102425876011, |
| "grad_norm": 0.5916996002197266, |
| "learning_rate": 0.000371158121964382, |
| "loss": 3.4559, |
| "step": 35450 |
| }, |
| { |
| "epoch": 3.8274932614555257, |
| "grad_norm": 0.6095645427703857, |
| "learning_rate": 0.0003708343227199136, |
| "loss": 3.4465, |
| "step": 35500 |
| }, |
| { |
| "epoch": 3.8328840970350404, |
| "grad_norm": 0.561920702457428, |
| "learning_rate": 0.0003705105234754452, |
| "loss": 3.4501, |
| "step": 35550 |
| }, |
| { |
| "epoch": 3.838274932614555, |
| "grad_norm": 0.602867066860199, |
| "learning_rate": 0.0003701867242309768, |
| "loss": 3.4602, |
| "step": 35600 |
| }, |
| { |
| "epoch": 3.8436657681940702, |
| "grad_norm": 0.6007760763168335, |
| "learning_rate": 0.00036986292498650833, |
| "loss": 3.4449, |
| "step": 35650 |
| }, |
| { |
| "epoch": 3.849056603773585, |
| "grad_norm": 0.600691020488739, |
| "learning_rate": 0.0003695456017269293, |
| "loss": 3.4464, |
| "step": 35700 |
| }, |
| { |
| "epoch": 3.8544474393530996, |
| "grad_norm": 0.6398903727531433, |
| "learning_rate": 0.00036922180248246083, |
| "loss": 3.4557, |
| "step": 35750 |
| }, |
| { |
| "epoch": 3.8598382749326143, |
| "grad_norm": 0.6188865303993225, |
| "learning_rate": 0.00036889800323799244, |
| "loss": 3.4427, |
| "step": 35800 |
| }, |
| { |
| "epoch": 3.8652291105121295, |
| "grad_norm": 0.621609091758728, |
| "learning_rate": 0.000368574203993524, |
| "loss": 3.4525, |
| "step": 35850 |
| }, |
| { |
| "epoch": 3.870619946091644, |
| "grad_norm": 0.5898281931877136, |
| "learning_rate": 0.0003682504047490556, |
| "loss": 3.466, |
| "step": 35900 |
| }, |
| { |
| "epoch": 3.876010781671159, |
| "grad_norm": 0.6150808930397034, |
| "learning_rate": 0.00036792660550458714, |
| "loss": 3.4687, |
| "step": 35950 |
| }, |
| { |
| "epoch": 3.881401617250674, |
| "grad_norm": 0.6602946519851685, |
| "learning_rate": 0.00036760280626011864, |
| "loss": 3.4347, |
| "step": 36000 |
| }, |
| { |
| "epoch": 3.881401617250674, |
| "eval_accuracy": 0.3755471895246447, |
| "eval_loss": 3.4529879093170166, |
| "eval_runtime": 183.2643, |
| "eval_samples_per_second": 98.279, |
| "eval_steps_per_second": 6.144, |
| "step": 36000 |
| }, |
| { |
| "epoch": 3.8867924528301887, |
| "grad_norm": 0.6273593306541443, |
| "learning_rate": 0.0003672790070156503, |
| "loss": 3.4721, |
| "step": 36050 |
| }, |
| { |
| "epoch": 3.8921832884097034, |
| "grad_norm": 0.5950552225112915, |
| "learning_rate": 0.0003669552077711818, |
| "loss": 3.4363, |
| "step": 36100 |
| }, |
| { |
| "epoch": 3.8975741239892185, |
| "grad_norm": 0.6331326961517334, |
| "learning_rate": 0.0003666314085267134, |
| "loss": 3.4437, |
| "step": 36150 |
| }, |
| { |
| "epoch": 3.9029649595687332, |
| "grad_norm": 0.624070405960083, |
| "learning_rate": 0.00036630760928224495, |
| "loss": 3.4499, |
| "step": 36200 |
| }, |
| { |
| "epoch": 3.908355795148248, |
| "grad_norm": 0.5709901452064514, |
| "learning_rate": 0.00036598381003777656, |
| "loss": 3.4563, |
| "step": 36250 |
| }, |
| { |
| "epoch": 3.913746630727763, |
| "grad_norm": 0.5700583457946777, |
| "learning_rate": 0.0003656600107933081, |
| "loss": 3.4594, |
| "step": 36300 |
| }, |
| { |
| "epoch": 3.9191374663072778, |
| "grad_norm": 0.607256293296814, |
| "learning_rate": 0.0003653362115488397, |
| "loss": 3.4724, |
| "step": 36350 |
| }, |
| { |
| "epoch": 3.9245283018867925, |
| "grad_norm": 0.5817025303840637, |
| "learning_rate": 0.00036501241230437126, |
| "loss": 3.4393, |
| "step": 36400 |
| }, |
| { |
| "epoch": 3.929919137466307, |
| "grad_norm": 0.6234590411186218, |
| "learning_rate": 0.0003646886130599028, |
| "loss": 3.445, |
| "step": 36450 |
| }, |
| { |
| "epoch": 3.935309973045822, |
| "grad_norm": 0.5800814032554626, |
| "learning_rate": 0.0003643648138154344, |
| "loss": 3.4585, |
| "step": 36500 |
| }, |
| { |
| "epoch": 3.940700808625337, |
| "grad_norm": 0.5907812118530273, |
| "learning_rate": 0.00036404101457096597, |
| "loss": 3.4736, |
| "step": 36550 |
| }, |
| { |
| "epoch": 3.9460916442048517, |
| "grad_norm": 0.6417782306671143, |
| "learning_rate": 0.00036371721532649757, |
| "loss": 3.4261, |
| "step": 36600 |
| }, |
| { |
| "epoch": 3.9514824797843664, |
| "grad_norm": 0.6414207816123962, |
| "learning_rate": 0.0003633934160820291, |
| "loss": 3.4434, |
| "step": 36650 |
| }, |
| { |
| "epoch": 3.9568733153638815, |
| "grad_norm": 0.6781249642372131, |
| "learning_rate": 0.00036306961683756073, |
| "loss": 3.4484, |
| "step": 36700 |
| }, |
| { |
| "epoch": 3.9622641509433962, |
| "grad_norm": 0.6436617374420166, |
| "learning_rate": 0.0003627458175930922, |
| "loss": 3.4289, |
| "step": 36750 |
| }, |
| { |
| "epoch": 3.967654986522911, |
| "grad_norm": 0.5609354972839355, |
| "learning_rate": 0.0003624220183486238, |
| "loss": 3.4438, |
| "step": 36800 |
| }, |
| { |
| "epoch": 3.973045822102426, |
| "grad_norm": 0.5995511412620544, |
| "learning_rate": 0.0003620982191041554, |
| "loss": 3.4516, |
| "step": 36850 |
| }, |
| { |
| "epoch": 3.9784366576819408, |
| "grad_norm": 0.6280034184455872, |
| "learning_rate": 0.00036177441985968693, |
| "loss": 3.45, |
| "step": 36900 |
| }, |
| { |
| "epoch": 3.9838274932614555, |
| "grad_norm": 0.6208735704421997, |
| "learning_rate": 0.00036145062061521854, |
| "loss": 3.4302, |
| "step": 36950 |
| }, |
| { |
| "epoch": 3.9892183288409706, |
| "grad_norm": 0.6470046639442444, |
| "learning_rate": 0.0003611268213707501, |
| "loss": 3.417, |
| "step": 37000 |
| }, |
| { |
| "epoch": 3.9892183288409706, |
| "eval_accuracy": 0.3759129151450282, |
| "eval_loss": 3.4461848735809326, |
| "eval_runtime": 183.7064, |
| "eval_samples_per_second": 98.042, |
| "eval_steps_per_second": 6.129, |
| "step": 37000 |
| }, |
| { |
| "epoch": 3.9946091644204853, |
| "grad_norm": 0.6202898621559143, |
| "learning_rate": 0.0003608030221262817, |
| "loss": 3.4704, |
| "step": 37050 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 1.2718347311019897, |
| "learning_rate": 0.00036047922288181324, |
| "loss": 3.4332, |
| "step": 37100 |
| }, |
| { |
| "epoch": 4.005390835579515, |
| "grad_norm": 0.6157583594322205, |
| "learning_rate": 0.00036015542363734485, |
| "loss": 3.3631, |
| "step": 37150 |
| }, |
| { |
| "epoch": 4.010781671159029, |
| "grad_norm": 0.607601523399353, |
| "learning_rate": 0.0003598316243928764, |
| "loss": 3.3591, |
| "step": 37200 |
| }, |
| { |
| "epoch": 4.0161725067385445, |
| "grad_norm": 0.6501696109771729, |
| "learning_rate": 0.00035950782514840795, |
| "loss": 3.355, |
| "step": 37250 |
| }, |
| { |
| "epoch": 4.02156334231806, |
| "grad_norm": 0.6216338872909546, |
| "learning_rate": 0.00035918402590393955, |
| "loss": 3.3606, |
| "step": 37300 |
| }, |
| { |
| "epoch": 4.026954177897574, |
| "grad_norm": 0.6452534198760986, |
| "learning_rate": 0.00035886022665947105, |
| "loss": 3.3414, |
| "step": 37350 |
| }, |
| { |
| "epoch": 4.032345013477089, |
| "grad_norm": 0.5642078518867493, |
| "learning_rate": 0.0003585364274150027, |
| "loss": 3.346, |
| "step": 37400 |
| }, |
| { |
| "epoch": 4.037735849056604, |
| "grad_norm": 0.6294705271720886, |
| "learning_rate": 0.0003582126281705342, |
| "loss": 3.3541, |
| "step": 37450 |
| }, |
| { |
| "epoch": 4.0431266846361185, |
| "grad_norm": 0.7128744721412659, |
| "learning_rate": 0.0003578888289260658, |
| "loss": 3.3548, |
| "step": 37500 |
| }, |
| { |
| "epoch": 4.048517520215634, |
| "grad_norm": 0.6176431179046631, |
| "learning_rate": 0.00035756502968159736, |
| "loss": 3.3643, |
| "step": 37550 |
| }, |
| { |
| "epoch": 4.053908355795148, |
| "grad_norm": 0.6452947854995728, |
| "learning_rate": 0.00035724123043712896, |
| "loss": 3.3591, |
| "step": 37600 |
| }, |
| { |
| "epoch": 4.059299191374663, |
| "grad_norm": 0.6863994598388672, |
| "learning_rate": 0.0003569174311926605, |
| "loss": 3.3658, |
| "step": 37650 |
| }, |
| { |
| "epoch": 4.064690026954178, |
| "grad_norm": 0.5968023538589478, |
| "learning_rate": 0.00035659363194819206, |
| "loss": 3.3681, |
| "step": 37700 |
| }, |
| { |
| "epoch": 4.070080862533692, |
| "grad_norm": 0.5858391523361206, |
| "learning_rate": 0.00035626983270372367, |
| "loss": 3.3728, |
| "step": 37750 |
| }, |
| { |
| "epoch": 4.0754716981132075, |
| "grad_norm": 0.6459882259368896, |
| "learning_rate": 0.0003559525094441446, |
| "loss": 3.3647, |
| "step": 37800 |
| }, |
| { |
| "epoch": 4.080862533692723, |
| "grad_norm": 0.6718738079071045, |
| "learning_rate": 0.00035562871019967617, |
| "loss": 3.3699, |
| "step": 37850 |
| }, |
| { |
| "epoch": 4.086253369272237, |
| "grad_norm": 0.6096557378768921, |
| "learning_rate": 0.0003553049109552078, |
| "loss": 3.3794, |
| "step": 37900 |
| }, |
| { |
| "epoch": 4.091644204851752, |
| "grad_norm": 0.6182774901390076, |
| "learning_rate": 0.00035498111171073933, |
| "loss": 3.3785, |
| "step": 37950 |
| }, |
| { |
| "epoch": 4.097035040431267, |
| "grad_norm": 0.6097174286842346, |
| "learning_rate": 0.0003546573124662708, |
| "loss": 3.3602, |
| "step": 38000 |
| }, |
| { |
| "epoch": 4.097035040431267, |
| "eval_accuracy": 0.3763951166611489, |
| "eval_loss": 3.448361396789551, |
| "eval_runtime": 183.0845, |
| "eval_samples_per_second": 98.375, |
| "eval_steps_per_second": 6.15, |
| "step": 38000 |
| }, |
| { |
| "epoch": 4.1024258760107815, |
| "grad_norm": 0.6500292420387268, |
| "learning_rate": 0.0003543335132218025, |
| "loss": 3.3779, |
| "step": 38050 |
| }, |
| { |
| "epoch": 4.107816711590297, |
| "grad_norm": 0.6115676164627075, |
| "learning_rate": 0.000354009713977334, |
| "loss": 3.3628, |
| "step": 38100 |
| }, |
| { |
| "epoch": 4.113207547169812, |
| "grad_norm": 0.6318033337593079, |
| "learning_rate": 0.0003536859147328656, |
| "loss": 3.377, |
| "step": 38150 |
| }, |
| { |
| "epoch": 4.118598382749326, |
| "grad_norm": 0.6544750332832336, |
| "learning_rate": 0.00035336211548839713, |
| "loss": 3.381, |
| "step": 38200 |
| }, |
| { |
| "epoch": 4.123989218328841, |
| "grad_norm": 0.8020888566970825, |
| "learning_rate": 0.00035303831624392874, |
| "loss": 3.3684, |
| "step": 38250 |
| }, |
| { |
| "epoch": 4.129380053908355, |
| "grad_norm": 0.6313894391059875, |
| "learning_rate": 0.0003527145169994603, |
| "loss": 3.3717, |
| "step": 38300 |
| }, |
| { |
| "epoch": 4.1347708894878705, |
| "grad_norm": 0.6045191287994385, |
| "learning_rate": 0.0003523907177549919, |
| "loss": 3.3555, |
| "step": 38350 |
| }, |
| { |
| "epoch": 4.140161725067386, |
| "grad_norm": 0.6183720231056213, |
| "learning_rate": 0.00035206691851052345, |
| "loss": 3.3793, |
| "step": 38400 |
| }, |
| { |
| "epoch": 4.1455525606469, |
| "grad_norm": 0.62325519323349, |
| "learning_rate": 0.000351743119266055, |
| "loss": 3.3754, |
| "step": 38450 |
| }, |
| { |
| "epoch": 4.150943396226415, |
| "grad_norm": 0.6222301125526428, |
| "learning_rate": 0.0003514193200215866, |
| "loss": 3.3679, |
| "step": 38500 |
| }, |
| { |
| "epoch": 4.15633423180593, |
| "grad_norm": 0.6233283877372742, |
| "learning_rate": 0.00035109552077711815, |
| "loss": 3.3677, |
| "step": 38550 |
| }, |
| { |
| "epoch": 4.1617250673854445, |
| "grad_norm": 0.6315566301345825, |
| "learning_rate": 0.00035077172153264976, |
| "loss": 3.3666, |
| "step": 38600 |
| }, |
| { |
| "epoch": 4.16711590296496, |
| "grad_norm": 0.6729894876480103, |
| "learning_rate": 0.0003504479222881813, |
| "loss": 3.3734, |
| "step": 38650 |
| }, |
| { |
| "epoch": 4.172506738544475, |
| "grad_norm": 0.62404465675354, |
| "learning_rate": 0.0003501241230437129, |
| "loss": 3.3882, |
| "step": 38700 |
| }, |
| { |
| "epoch": 4.177897574123989, |
| "grad_norm": 0.6314374208450317, |
| "learning_rate": 0.0003498003237992444, |
| "loss": 3.3631, |
| "step": 38750 |
| }, |
| { |
| "epoch": 4.183288409703504, |
| "grad_norm": 0.5996955037117004, |
| "learning_rate": 0.000349476524554776, |
| "loss": 3.379, |
| "step": 38800 |
| }, |
| { |
| "epoch": 4.188679245283019, |
| "grad_norm": 0.6228339076042175, |
| "learning_rate": 0.00034915272531030756, |
| "loss": 3.3869, |
| "step": 38850 |
| }, |
| { |
| "epoch": 4.1940700808625335, |
| "grad_norm": 0.6187623143196106, |
| "learning_rate": 0.0003488289260658391, |
| "loss": 3.4018, |
| "step": 38900 |
| }, |
| { |
| "epoch": 4.199460916442049, |
| "grad_norm": 0.68119215965271, |
| "learning_rate": 0.0003485051268213707, |
| "loss": 3.3689, |
| "step": 38950 |
| }, |
| { |
| "epoch": 4.204851752021563, |
| "grad_norm": 0.6336060762405396, |
| "learning_rate": 0.00034818132757690227, |
| "loss": 3.384, |
| "step": 39000 |
| }, |
| { |
| "epoch": 4.204851752021563, |
| "eval_accuracy": 0.37696337126441865, |
| "eval_loss": 3.4456894397735596, |
| "eval_runtime": 183.3606, |
| "eval_samples_per_second": 98.227, |
| "eval_steps_per_second": 6.141, |
| "step": 39000 |
| }, |
| { |
| "epoch": 4.210242587601078, |
| "grad_norm": 0.6565260887145996, |
| "learning_rate": 0.0003478575283324339, |
| "loss": 3.3906, |
| "step": 39050 |
| }, |
| { |
| "epoch": 4.215633423180593, |
| "grad_norm": 0.6345798969268799, |
| "learning_rate": 0.0003475337290879654, |
| "loss": 3.3934, |
| "step": 39100 |
| }, |
| { |
| "epoch": 4.2210242587601075, |
| "grad_norm": 0.628085732460022, |
| "learning_rate": 0.00034720992984349703, |
| "loss": 3.3796, |
| "step": 39150 |
| }, |
| { |
| "epoch": 4.226415094339623, |
| "grad_norm": 0.698582649230957, |
| "learning_rate": 0.0003468861305990286, |
| "loss": 3.371, |
| "step": 39200 |
| }, |
| { |
| "epoch": 4.231805929919138, |
| "grad_norm": 0.6223801970481873, |
| "learning_rate": 0.0003465623313545602, |
| "loss": 3.3729, |
| "step": 39250 |
| }, |
| { |
| "epoch": 4.237196765498652, |
| "grad_norm": 0.6224102973937988, |
| "learning_rate": 0.00034623853211009173, |
| "loss": 3.3776, |
| "step": 39300 |
| }, |
| { |
| "epoch": 4.242587601078167, |
| "grad_norm": 0.621496856212616, |
| "learning_rate": 0.00034591473286562323, |
| "loss": 3.3878, |
| "step": 39350 |
| }, |
| { |
| "epoch": 4.247978436657682, |
| "grad_norm": 0.6796809434890747, |
| "learning_rate": 0.0003455909336211549, |
| "loss": 3.3817, |
| "step": 39400 |
| }, |
| { |
| "epoch": 4.2533692722371965, |
| "grad_norm": 0.6903562545776367, |
| "learning_rate": 0.0003452671343766864, |
| "loss": 3.3942, |
| "step": 39450 |
| }, |
| { |
| "epoch": 4.258760107816712, |
| "grad_norm": 0.6326731443405151, |
| "learning_rate": 0.000344943335132218, |
| "loss": 3.3807, |
| "step": 39500 |
| }, |
| { |
| "epoch": 4.264150943396227, |
| "grad_norm": 0.6289458274841309, |
| "learning_rate": 0.00034461953588774954, |
| "loss": 3.3671, |
| "step": 39550 |
| }, |
| { |
| "epoch": 4.269541778975741, |
| "grad_norm": 0.6308515071868896, |
| "learning_rate": 0.00034429573664328115, |
| "loss": 3.3838, |
| "step": 39600 |
| }, |
| { |
| "epoch": 4.274932614555256, |
| "grad_norm": 0.5939794182777405, |
| "learning_rate": 0.0003439719373988127, |
| "loss": 3.3924, |
| "step": 39650 |
| }, |
| { |
| "epoch": 4.280323450134771, |
| "grad_norm": 0.6445115208625793, |
| "learning_rate": 0.00034364813815434425, |
| "loss": 3.3743, |
| "step": 39700 |
| }, |
| { |
| "epoch": 4.285714285714286, |
| "grad_norm": 0.6573401093482971, |
| "learning_rate": 0.00034332433890987585, |
| "loss": 3.3976, |
| "step": 39750 |
| }, |
| { |
| "epoch": 4.291105121293801, |
| "grad_norm": 0.637039840221405, |
| "learning_rate": 0.0003430005396654074, |
| "loss": 3.3844, |
| "step": 39800 |
| }, |
| { |
| "epoch": 4.296495956873315, |
| "grad_norm": 0.6390613913536072, |
| "learning_rate": 0.000342676740420939, |
| "loss": 3.385, |
| "step": 39850 |
| }, |
| { |
| "epoch": 4.30188679245283, |
| "grad_norm": 0.6816765666007996, |
| "learning_rate": 0.00034235294117647056, |
| "loss": 3.3858, |
| "step": 39900 |
| }, |
| { |
| "epoch": 4.307277628032345, |
| "grad_norm": 0.7338299751281738, |
| "learning_rate": 0.00034202914193200216, |
| "loss": 3.3951, |
| "step": 39950 |
| }, |
| { |
| "epoch": 4.3126684636118595, |
| "grad_norm": 0.6921063661575317, |
| "learning_rate": 0.0003417118186724231, |
| "loss": 3.4014, |
| "step": 40000 |
| }, |
| { |
| "epoch": 4.3126684636118595, |
| "eval_accuracy": 0.3773256199923921, |
| "eval_loss": 3.4401845932006836, |
| "eval_runtime": 183.335, |
| "eval_samples_per_second": 98.241, |
| "eval_steps_per_second": 6.142, |
| "step": 40000 |
| }, |
| { |
| "epoch": 4.318059299191375, |
| "grad_norm": 0.6596004366874695, |
| "learning_rate": 0.00034138801942795467, |
| "loss": 3.388, |
| "step": 40050 |
| }, |
| { |
| "epoch": 4.32345013477089, |
| "grad_norm": 0.6345239281654358, |
| "learning_rate": 0.00034106422018348616, |
| "loss": 3.3935, |
| "step": 40100 |
| }, |
| { |
| "epoch": 4.328840970350404, |
| "grad_norm": 0.630919873714447, |
| "learning_rate": 0.00034074042093901777, |
| "loss": 3.3751, |
| "step": 40150 |
| }, |
| { |
| "epoch": 4.334231805929919, |
| "grad_norm": 0.6389809846878052, |
| "learning_rate": 0.0003404166216945493, |
| "loss": 3.4071, |
| "step": 40200 |
| }, |
| { |
| "epoch": 4.339622641509434, |
| "grad_norm": 0.6863662004470825, |
| "learning_rate": 0.0003400928224500809, |
| "loss": 3.4102, |
| "step": 40250 |
| }, |
| { |
| "epoch": 4.345013477088949, |
| "grad_norm": 0.6372359991073608, |
| "learning_rate": 0.00033976902320561247, |
| "loss": 3.3826, |
| "step": 40300 |
| }, |
| { |
| "epoch": 4.350404312668464, |
| "grad_norm": 0.6467205882072449, |
| "learning_rate": 0.0003394452239611441, |
| "loss": 3.392, |
| "step": 40350 |
| }, |
| { |
| "epoch": 4.355795148247978, |
| "grad_norm": 0.6186798214912415, |
| "learning_rate": 0.00033912142471667563, |
| "loss": 3.3892, |
| "step": 40400 |
| }, |
| { |
| "epoch": 4.361185983827493, |
| "grad_norm": 0.6394029259681702, |
| "learning_rate": 0.00033879762547220723, |
| "loss": 3.3908, |
| "step": 40450 |
| }, |
| { |
| "epoch": 4.366576819407008, |
| "grad_norm": 0.6361256837844849, |
| "learning_rate": 0.0003384738262277388, |
| "loss": 3.3858, |
| "step": 40500 |
| }, |
| { |
| "epoch": 4.3719676549865225, |
| "grad_norm": 0.6550248265266418, |
| "learning_rate": 0.00033815002698327033, |
| "loss": 3.3896, |
| "step": 40550 |
| }, |
| { |
| "epoch": 4.377358490566038, |
| "grad_norm": 0.6077722311019897, |
| "learning_rate": 0.00033782622773880194, |
| "loss": 3.3989, |
| "step": 40600 |
| }, |
| { |
| "epoch": 4.382749326145553, |
| "grad_norm": 0.6376055479049683, |
| "learning_rate": 0.0003375024284943335, |
| "loss": 3.3889, |
| "step": 40650 |
| }, |
| { |
| "epoch": 4.388140161725067, |
| "grad_norm": 0.6313409209251404, |
| "learning_rate": 0.0003371786292498651, |
| "loss": 3.4015, |
| "step": 40700 |
| }, |
| { |
| "epoch": 4.393530997304582, |
| "grad_norm": 0.6659762263298035, |
| "learning_rate": 0.0003368548300053966, |
| "loss": 3.403, |
| "step": 40750 |
| }, |
| { |
| "epoch": 4.398921832884097, |
| "grad_norm": 0.6614190340042114, |
| "learning_rate": 0.0003365310307609282, |
| "loss": 3.3948, |
| "step": 40800 |
| }, |
| { |
| "epoch": 4.404312668463612, |
| "grad_norm": 0.6322287321090698, |
| "learning_rate": 0.00033620723151645975, |
| "loss": 3.383, |
| "step": 40850 |
| }, |
| { |
| "epoch": 4.409703504043127, |
| "grad_norm": 0.6170388460159302, |
| "learning_rate": 0.0003358834322719913, |
| "loss": 3.383, |
| "step": 40900 |
| }, |
| { |
| "epoch": 4.415094339622642, |
| "grad_norm": 0.896038830280304, |
| "learning_rate": 0.0003355596330275229, |
| "loss": 3.38, |
| "step": 40950 |
| }, |
| { |
| "epoch": 4.420485175202156, |
| "grad_norm": 0.6698582768440247, |
| "learning_rate": 0.00033523583378305445, |
| "loss": 3.4152, |
| "step": 41000 |
| }, |
| { |
| "epoch": 4.420485175202156, |
| "eval_accuracy": 0.37785747587824403, |
| "eval_loss": 3.4369399547576904, |
| "eval_runtime": 183.3239, |
| "eval_samples_per_second": 98.247, |
| "eval_steps_per_second": 6.142, |
| "step": 41000 |
| }, |
| { |
| "epoch": 4.425876010781671, |
| "grad_norm": 0.6632938981056213, |
| "learning_rate": 0.00033491203453858606, |
| "loss": 3.3909, |
| "step": 41050 |
| }, |
| { |
| "epoch": 4.431266846361186, |
| "grad_norm": 0.6375352144241333, |
| "learning_rate": 0.0003345882352941176, |
| "loss": 3.3986, |
| "step": 41100 |
| }, |
| { |
| "epoch": 4.436657681940701, |
| "grad_norm": 0.6697904467582703, |
| "learning_rate": 0.0003342644360496492, |
| "loss": 3.4054, |
| "step": 41150 |
| }, |
| { |
| "epoch": 4.442048517520216, |
| "grad_norm": 0.6245919466018677, |
| "learning_rate": 0.00033394063680518076, |
| "loss": 3.3906, |
| "step": 41200 |
| }, |
| { |
| "epoch": 4.44743935309973, |
| "grad_norm": 0.662598729133606, |
| "learning_rate": 0.00033361683756071237, |
| "loss": 3.4072, |
| "step": 41250 |
| }, |
| { |
| "epoch": 4.452830188679245, |
| "grad_norm": 0.6050485372543335, |
| "learning_rate": 0.0003332930383162439, |
| "loss": 3.3794, |
| "step": 41300 |
| }, |
| { |
| "epoch": 4.45822102425876, |
| "grad_norm": 0.6476341485977173, |
| "learning_rate": 0.0003329692390717754, |
| "loss": 3.385, |
| "step": 41350 |
| }, |
| { |
| "epoch": 4.463611859838275, |
| "grad_norm": 0.5851879715919495, |
| "learning_rate": 0.0003326454398273071, |
| "loss": 3.4083, |
| "step": 41400 |
| }, |
| { |
| "epoch": 4.46900269541779, |
| "grad_norm": 0.6802077293395996, |
| "learning_rate": 0.00033232164058283857, |
| "loss": 3.3837, |
| "step": 41450 |
| }, |
| { |
| "epoch": 4.474393530997305, |
| "grad_norm": 0.8038954734802246, |
| "learning_rate": 0.0003319978413383702, |
| "loss": 3.3939, |
| "step": 41500 |
| }, |
| { |
| "epoch": 4.479784366576819, |
| "grad_norm": 0.6305577158927917, |
| "learning_rate": 0.0003316740420939017, |
| "loss": 3.3885, |
| "step": 41550 |
| }, |
| { |
| "epoch": 4.485175202156334, |
| "grad_norm": 0.6377895474433899, |
| "learning_rate": 0.00033135024284943333, |
| "loss": 3.3988, |
| "step": 41600 |
| }, |
| { |
| "epoch": 4.490566037735849, |
| "grad_norm": 0.5902746319770813, |
| "learning_rate": 0.0003310264436049649, |
| "loss": 3.3829, |
| "step": 41650 |
| }, |
| { |
| "epoch": 4.495956873315364, |
| "grad_norm": 0.5873847603797913, |
| "learning_rate": 0.0003307026443604965, |
| "loss": 3.4045, |
| "step": 41700 |
| }, |
| { |
| "epoch": 4.501347708894879, |
| "grad_norm": 0.6426920890808105, |
| "learning_rate": 0.00033037884511602804, |
| "loss": 3.389, |
| "step": 41750 |
| }, |
| { |
| "epoch": 4.506738544474393, |
| "grad_norm": 0.7010065913200378, |
| "learning_rate": 0.0003300550458715596, |
| "loss": 3.3945, |
| "step": 41800 |
| }, |
| { |
| "epoch": 4.512129380053908, |
| "grad_norm": 0.6423835158348083, |
| "learning_rate": 0.0003297312466270912, |
| "loss": 3.383, |
| "step": 41850 |
| }, |
| { |
| "epoch": 4.517520215633423, |
| "grad_norm": 0.6504900455474854, |
| "learning_rate": 0.00032940744738262274, |
| "loss": 3.3921, |
| "step": 41900 |
| }, |
| { |
| "epoch": 4.5229110512129385, |
| "grad_norm": 0.6463404893875122, |
| "learning_rate": 0.00032908364813815435, |
| "loss": 3.3769, |
| "step": 41950 |
| }, |
| { |
| "epoch": 4.528301886792453, |
| "grad_norm": 0.6280781030654907, |
| "learning_rate": 0.0003287598488936859, |
| "loss": 3.3867, |
| "step": 42000 |
| }, |
| { |
| "epoch": 4.528301886792453, |
| "eval_accuracy": 0.37868834451136135, |
| "eval_loss": 3.430631399154663, |
| "eval_runtime": 183.346, |
| "eval_samples_per_second": 98.235, |
| "eval_steps_per_second": 6.141, |
| "step": 42000 |
| }, |
| { |
| "epoch": 4.533692722371968, |
| "grad_norm": 0.6555848717689514, |
| "learning_rate": 0.0003284360496492175, |
| "loss": 3.3855, |
| "step": 42050 |
| }, |
| { |
| "epoch": 4.539083557951482, |
| "grad_norm": 0.6349523067474365, |
| "learning_rate": 0.00032811872638963834, |
| "loss": 3.3933, |
| "step": 42100 |
| }, |
| { |
| "epoch": 4.544474393530997, |
| "grad_norm": 0.6653229594230652, |
| "learning_rate": 0.00032779492714516995, |
| "loss": 3.4217, |
| "step": 42150 |
| }, |
| { |
| "epoch": 4.549865229110512, |
| "grad_norm": 0.622232973575592, |
| "learning_rate": 0.0003274711279007015, |
| "loss": 3.3876, |
| "step": 42200 |
| }, |
| { |
| "epoch": 4.555256064690027, |
| "grad_norm": 0.6820586323738098, |
| "learning_rate": 0.0003271473286562331, |
| "loss": 3.3862, |
| "step": 42250 |
| }, |
| { |
| "epoch": 4.560646900269542, |
| "grad_norm": 0.6591930985450745, |
| "learning_rate": 0.00032682352941176466, |
| "loss": 3.39, |
| "step": 42300 |
| }, |
| { |
| "epoch": 4.566037735849057, |
| "grad_norm": 0.6004623174667358, |
| "learning_rate": 0.00032649973016729626, |
| "loss": 3.3902, |
| "step": 42350 |
| }, |
| { |
| "epoch": 4.571428571428571, |
| "grad_norm": 0.6339703798294067, |
| "learning_rate": 0.0003261759309228278, |
| "loss": 3.3813, |
| "step": 42400 |
| }, |
| { |
| "epoch": 4.576819407008086, |
| "grad_norm": 0.6326310038566589, |
| "learning_rate": 0.0003258521316783594, |
| "loss": 3.3934, |
| "step": 42450 |
| }, |
| { |
| "epoch": 4.5822102425876015, |
| "grad_norm": 0.6333923935890198, |
| "learning_rate": 0.00032552833243389097, |
| "loss": 3.3914, |
| "step": 42500 |
| }, |
| { |
| "epoch": 4.587601078167116, |
| "grad_norm": 0.6541619300842285, |
| "learning_rate": 0.0003252045331894225, |
| "loss": 3.3918, |
| "step": 42550 |
| }, |
| { |
| "epoch": 4.592991913746631, |
| "grad_norm": 0.6211109161376953, |
| "learning_rate": 0.0003248807339449541, |
| "loss": 3.3753, |
| "step": 42600 |
| }, |
| { |
| "epoch": 4.598382749326145, |
| "grad_norm": 0.7737997174263, |
| "learning_rate": 0.00032455693470048567, |
| "loss": 3.3756, |
| "step": 42650 |
| }, |
| { |
| "epoch": 4.60377358490566, |
| "grad_norm": 0.6622357368469238, |
| "learning_rate": 0.0003242331354560173, |
| "loss": 3.3862, |
| "step": 42700 |
| }, |
| { |
| "epoch": 4.609164420485175, |
| "grad_norm": 0.6532875299453735, |
| "learning_rate": 0.0003239093362115488, |
| "loss": 3.4049, |
| "step": 42750 |
| }, |
| { |
| "epoch": 4.6145552560646905, |
| "grad_norm": 0.6364518404006958, |
| "learning_rate": 0.00032358553696708043, |
| "loss": 3.3759, |
| "step": 42800 |
| }, |
| { |
| "epoch": 4.619946091644205, |
| "grad_norm": 0.6832470893859863, |
| "learning_rate": 0.00032326173772261193, |
| "loss": 3.3814, |
| "step": 42850 |
| }, |
| { |
| "epoch": 4.62533692722372, |
| "grad_norm": 0.677706241607666, |
| "learning_rate": 0.00032293793847814353, |
| "loss": 3.3801, |
| "step": 42900 |
| }, |
| { |
| "epoch": 4.630727762803234, |
| "grad_norm": 0.6813549995422363, |
| "learning_rate": 0.0003226141392336751, |
| "loss": 3.403, |
| "step": 42950 |
| }, |
| { |
| "epoch": 4.636118598382749, |
| "grad_norm": 0.7086654901504517, |
| "learning_rate": 0.00032229033998920663, |
| "loss": 3.4083, |
| "step": 43000 |
| }, |
| { |
| "epoch": 4.636118598382749, |
| "eval_accuracy": 0.37909992165040257, |
| "eval_loss": 3.4262819290161133, |
| "eval_runtime": 183.3323, |
| "eval_samples_per_second": 98.242, |
| "eval_steps_per_second": 6.142, |
| "step": 43000 |
| }, |
| { |
| "epoch": 4.6415094339622645, |
| "grad_norm": 0.6429668664932251, |
| "learning_rate": 0.00032196654074473824, |
| "loss": 3.3947, |
| "step": 43050 |
| }, |
| { |
| "epoch": 4.646900269541779, |
| "grad_norm": 0.610312283039093, |
| "learning_rate": 0.0003216427415002698, |
| "loss": 3.3979, |
| "step": 43100 |
| }, |
| { |
| "epoch": 4.652291105121294, |
| "grad_norm": 0.6420636177062988, |
| "learning_rate": 0.0003213189422558014, |
| "loss": 3.4103, |
| "step": 43150 |
| }, |
| { |
| "epoch": 4.657681940700809, |
| "grad_norm": 0.6642614603042603, |
| "learning_rate": 0.00032099514301133295, |
| "loss": 3.383, |
| "step": 43200 |
| }, |
| { |
| "epoch": 4.663072776280323, |
| "grad_norm": 0.622520923614502, |
| "learning_rate": 0.00032067134376686455, |
| "loss": 3.3825, |
| "step": 43250 |
| }, |
| { |
| "epoch": 4.668463611859838, |
| "grad_norm": 0.6294994950294495, |
| "learning_rate": 0.0003203475445223961, |
| "loss": 3.394, |
| "step": 43300 |
| }, |
| { |
| "epoch": 4.6738544474393535, |
| "grad_norm": 0.6598164439201355, |
| "learning_rate": 0.0003200237452779277, |
| "loss": 3.3937, |
| "step": 43350 |
| }, |
| { |
| "epoch": 4.679245283018868, |
| "grad_norm": 0.6587915420532227, |
| "learning_rate": 0.00031969994603345926, |
| "loss": 3.405, |
| "step": 43400 |
| }, |
| { |
| "epoch": 4.684636118598383, |
| "grad_norm": 0.6498952507972717, |
| "learning_rate": 0.00031937614678899075, |
| "loss": 3.3921, |
| "step": 43450 |
| }, |
| { |
| "epoch": 4.690026954177897, |
| "grad_norm": 0.7558854818344116, |
| "learning_rate": 0.00031905234754452236, |
| "loss": 3.4051, |
| "step": 43500 |
| }, |
| { |
| "epoch": 4.695417789757412, |
| "grad_norm": 0.6848349571228027, |
| "learning_rate": 0.0003187285483000539, |
| "loss": 3.3883, |
| "step": 43550 |
| }, |
| { |
| "epoch": 4.7008086253369274, |
| "grad_norm": 0.6309284567832947, |
| "learning_rate": 0.0003184047490555855, |
| "loss": 3.3793, |
| "step": 43600 |
| }, |
| { |
| "epoch": 4.706199460916442, |
| "grad_norm": 0.6401345729827881, |
| "learning_rate": 0.00031808094981111706, |
| "loss": 3.3948, |
| "step": 43650 |
| }, |
| { |
| "epoch": 4.711590296495957, |
| "grad_norm": 0.6336686611175537, |
| "learning_rate": 0.00031775715056664867, |
| "loss": 3.3802, |
| "step": 43700 |
| }, |
| { |
| "epoch": 4.716981132075472, |
| "grad_norm": 0.6566705703735352, |
| "learning_rate": 0.0003174333513221802, |
| "loss": 3.3997, |
| "step": 43750 |
| }, |
| { |
| "epoch": 4.722371967654986, |
| "grad_norm": 0.641469419002533, |
| "learning_rate": 0.00031710955207771177, |
| "loss": 3.3929, |
| "step": 43800 |
| }, |
| { |
| "epoch": 4.727762803234501, |
| "grad_norm": 0.6639798879623413, |
| "learning_rate": 0.0003167857528332434, |
| "loss": 3.3948, |
| "step": 43850 |
| }, |
| { |
| "epoch": 4.7331536388140165, |
| "grad_norm": 0.6445428729057312, |
| "learning_rate": 0.0003164619535887749, |
| "loss": 3.3984, |
| "step": 43900 |
| }, |
| { |
| "epoch": 4.738544474393531, |
| "grad_norm": 0.6511913537979126, |
| "learning_rate": 0.00031613815434430653, |
| "loss": 3.3889, |
| "step": 43950 |
| }, |
| { |
| "epoch": 4.743935309973046, |
| "grad_norm": 0.6922048330307007, |
| "learning_rate": 0.0003158143550998381, |
| "loss": 3.3929, |
| "step": 44000 |
| }, |
| { |
| "epoch": 4.743935309973046, |
| "eval_accuracy": 0.3794368542555153, |
| "eval_loss": 3.420168399810791, |
| "eval_runtime": 183.4637, |
| "eval_samples_per_second": 98.172, |
| "eval_steps_per_second": 6.137, |
| "step": 44000 |
| }, |
| { |
| "epoch": 4.74932614555256, |
| "grad_norm": 0.8792603015899658, |
| "learning_rate": 0.0003154905558553697, |
| "loss": 3.403, |
| "step": 44050 |
| }, |
| { |
| "epoch": 4.754716981132075, |
| "grad_norm": 0.676875650882721, |
| "learning_rate": 0.0003151667566109012, |
| "loss": 3.3966, |
| "step": 44100 |
| }, |
| { |
| "epoch": 4.7601078167115904, |
| "grad_norm": 0.675866425037384, |
| "learning_rate": 0.00031484943335132213, |
| "loss": 3.4062, |
| "step": 44150 |
| }, |
| { |
| "epoch": 4.765498652291106, |
| "grad_norm": 0.6132005453109741, |
| "learning_rate": 0.0003145256341068537, |
| "loss": 3.4029, |
| "step": 44200 |
| }, |
| { |
| "epoch": 4.77088948787062, |
| "grad_norm": 0.6132225394248962, |
| "learning_rate": 0.0003142018348623853, |
| "loss": 3.3648, |
| "step": 44250 |
| }, |
| { |
| "epoch": 4.776280323450135, |
| "grad_norm": 0.6344752907752991, |
| "learning_rate": 0.00031387803561791684, |
| "loss": 3.3899, |
| "step": 44300 |
| }, |
| { |
| "epoch": 4.781671159029649, |
| "grad_norm": 0.6366958022117615, |
| "learning_rate": 0.00031355423637344844, |
| "loss": 3.3942, |
| "step": 44350 |
| }, |
| { |
| "epoch": 4.787061994609164, |
| "grad_norm": 0.6339268088340759, |
| "learning_rate": 0.00031323043712898, |
| "loss": 3.3764, |
| "step": 44400 |
| }, |
| { |
| "epoch": 4.7924528301886795, |
| "grad_norm": 0.6333193182945251, |
| "learning_rate": 0.0003129066378845116, |
| "loss": 3.4021, |
| "step": 44450 |
| }, |
| { |
| "epoch": 4.797843665768194, |
| "grad_norm": 0.6859214305877686, |
| "learning_rate": 0.00031258283864004315, |
| "loss": 3.3991, |
| "step": 44500 |
| }, |
| { |
| "epoch": 4.803234501347709, |
| "grad_norm": 0.6658663153648376, |
| "learning_rate": 0.0003122590393955747, |
| "loss": 3.3747, |
| "step": 44550 |
| }, |
| { |
| "epoch": 4.808625336927224, |
| "grad_norm": 0.658751904964447, |
| "learning_rate": 0.0003119352401511063, |
| "loss": 3.3873, |
| "step": 44600 |
| }, |
| { |
| "epoch": 4.814016172506738, |
| "grad_norm": 0.6287225484848022, |
| "learning_rate": 0.00031161144090663786, |
| "loss": 3.3881, |
| "step": 44650 |
| }, |
| { |
| "epoch": 4.819407008086253, |
| "grad_norm": 0.6663537621498108, |
| "learning_rate": 0.00031128764166216946, |
| "loss": 3.389, |
| "step": 44700 |
| }, |
| { |
| "epoch": 4.824797843665769, |
| "grad_norm": 0.6953670382499695, |
| "learning_rate": 0.00031096384241770096, |
| "loss": 3.4036, |
| "step": 44750 |
| }, |
| { |
| "epoch": 4.830188679245283, |
| "grad_norm": 0.6840183734893799, |
| "learning_rate": 0.0003106400431732326, |
| "loss": 3.374, |
| "step": 44800 |
| }, |
| { |
| "epoch": 4.835579514824798, |
| "grad_norm": 0.6771771311759949, |
| "learning_rate": 0.0003103162439287641, |
| "loss": 3.3881, |
| "step": 44850 |
| }, |
| { |
| "epoch": 4.840970350404312, |
| "grad_norm": 0.6342368721961975, |
| "learning_rate": 0.0003099924446842957, |
| "loss": 3.3658, |
| "step": 44900 |
| }, |
| { |
| "epoch": 4.846361185983827, |
| "grad_norm": 0.6187852621078491, |
| "learning_rate": 0.00030966864543982727, |
| "loss": 3.4036, |
| "step": 44950 |
| }, |
| { |
| "epoch": 4.8517520215633425, |
| "grad_norm": 0.6666272282600403, |
| "learning_rate": 0.0003093448461953588, |
| "loss": 3.389, |
| "step": 45000 |
| }, |
| { |
| "epoch": 4.8517520215633425, |
| "eval_accuracy": 0.38026935268194983, |
| "eval_loss": 3.4122982025146484, |
| "eval_runtime": 183.2917, |
| "eval_samples_per_second": 98.264, |
| "eval_steps_per_second": 6.143, |
| "step": 45000 |
| }, |
| { |
| "epoch": 4.857142857142857, |
| "grad_norm": 0.7002741694450378, |
| "learning_rate": 0.0003090210469508904, |
| "loss": 3.4004, |
| "step": 45050 |
| }, |
| { |
| "epoch": 4.862533692722372, |
| "grad_norm": 0.648239016532898, |
| "learning_rate": 0.00030869724770642197, |
| "loss": 3.4038, |
| "step": 45100 |
| }, |
| { |
| "epoch": 4.867924528301887, |
| "grad_norm": 0.6664896607398987, |
| "learning_rate": 0.0003083734484619536, |
| "loss": 3.3746, |
| "step": 45150 |
| }, |
| { |
| "epoch": 4.873315363881401, |
| "grad_norm": 0.6408718824386597, |
| "learning_rate": 0.00030804964921748513, |
| "loss": 3.3832, |
| "step": 45200 |
| }, |
| { |
| "epoch": 4.878706199460916, |
| "grad_norm": 0.6719059944152832, |
| "learning_rate": 0.00030772584997301673, |
| "loss": 3.3973, |
| "step": 45250 |
| }, |
| { |
| "epoch": 4.884097035040432, |
| "grad_norm": 0.638828456401825, |
| "learning_rate": 0.0003074020507285483, |
| "loss": 3.4089, |
| "step": 45300 |
| }, |
| { |
| "epoch": 4.889487870619946, |
| "grad_norm": 0.6789811849594116, |
| "learning_rate": 0.0003070782514840799, |
| "loss": 3.3876, |
| "step": 45350 |
| }, |
| { |
| "epoch": 4.894878706199461, |
| "grad_norm": 0.6199337840080261, |
| "learning_rate": 0.00030675445223961144, |
| "loss": 3.3971, |
| "step": 45400 |
| }, |
| { |
| "epoch": 4.900269541778976, |
| "grad_norm": 0.6504912376403809, |
| "learning_rate": 0.00030643065299514294, |
| "loss": 3.3937, |
| "step": 45450 |
| }, |
| { |
| "epoch": 4.90566037735849, |
| "grad_norm": 0.675342857837677, |
| "learning_rate": 0.00030610685375067454, |
| "loss": 3.3929, |
| "step": 45500 |
| }, |
| { |
| "epoch": 4.9110512129380055, |
| "grad_norm": 0.6166520714759827, |
| "learning_rate": 0.0003057830545062061, |
| "loss": 3.3939, |
| "step": 45550 |
| }, |
| { |
| "epoch": 4.916442048517521, |
| "grad_norm": 0.6337319016456604, |
| "learning_rate": 0.0003054592552617377, |
| "loss": 3.3898, |
| "step": 45600 |
| }, |
| { |
| "epoch": 4.921832884097035, |
| "grad_norm": 0.6474262475967407, |
| "learning_rate": 0.00030513545601726925, |
| "loss": 3.3924, |
| "step": 45650 |
| }, |
| { |
| "epoch": 4.92722371967655, |
| "grad_norm": 0.687142014503479, |
| "learning_rate": 0.00030481165677280085, |
| "loss": 3.3675, |
| "step": 45700 |
| }, |
| { |
| "epoch": 4.932614555256064, |
| "grad_norm": 0.692780613899231, |
| "learning_rate": 0.0003044878575283324, |
| "loss": 3.3842, |
| "step": 45750 |
| }, |
| { |
| "epoch": 4.938005390835579, |
| "grad_norm": 0.6651083827018738, |
| "learning_rate": 0.000304164058283864, |
| "loss": 3.4037, |
| "step": 45800 |
| }, |
| { |
| "epoch": 4.943396226415095, |
| "grad_norm": 0.7113507390022278, |
| "learning_rate": 0.00030384025903939556, |
| "loss": 3.3832, |
| "step": 45850 |
| }, |
| { |
| "epoch": 4.948787061994609, |
| "grad_norm": 0.6864883303642273, |
| "learning_rate": 0.0003035164597949271, |
| "loss": 3.3896, |
| "step": 45900 |
| }, |
| { |
| "epoch": 4.954177897574124, |
| "grad_norm": 0.5930712223052979, |
| "learning_rate": 0.0003031926605504587, |
| "loss": 3.3812, |
| "step": 45950 |
| }, |
| { |
| "epoch": 4.959568733153639, |
| "grad_norm": 0.60821133852005, |
| "learning_rate": 0.00030286886130599026, |
| "loss": 3.3876, |
| "step": 46000 |
| }, |
| { |
| "epoch": 4.959568733153639, |
| "eval_accuracy": 0.380540441637047, |
| "eval_loss": 3.4126551151275635, |
| "eval_runtime": 183.4652, |
| "eval_samples_per_second": 98.171, |
| "eval_steps_per_second": 6.137, |
| "step": 46000 |
| }, |
| { |
| "epoch": 4.964959568733153, |
| "grad_norm": 0.6533971428871155, |
| "learning_rate": 0.00030254506206152187, |
| "loss": 3.398, |
| "step": 46050 |
| }, |
| { |
| "epoch": 4.9703504043126685, |
| "grad_norm": 0.633718729019165, |
| "learning_rate": 0.00030222126281705336, |
| "loss": 3.368, |
| "step": 46100 |
| }, |
| { |
| "epoch": 4.975741239892184, |
| "grad_norm": 0.6307101845741272, |
| "learning_rate": 0.0003019039395574743, |
| "loss": 3.3747, |
| "step": 46150 |
| }, |
| { |
| "epoch": 4.981132075471698, |
| "grad_norm": 0.6983139514923096, |
| "learning_rate": 0.00030158014031300587, |
| "loss": 3.3861, |
| "step": 46200 |
| }, |
| { |
| "epoch": 4.986522911051213, |
| "grad_norm": 0.6607715487480164, |
| "learning_rate": 0.00030125634106853747, |
| "loss": 3.4037, |
| "step": 46250 |
| }, |
| { |
| "epoch": 4.991913746630727, |
| "grad_norm": 0.6260371804237366, |
| "learning_rate": 0.000300932541824069, |
| "loss": 3.4017, |
| "step": 46300 |
| }, |
| { |
| "epoch": 4.997304582210242, |
| "grad_norm": 0.6400908827781677, |
| "learning_rate": 0.0003006087425796006, |
| "loss": 3.3789, |
| "step": 46350 |
| }, |
| { |
| "epoch": 5.002695417789758, |
| "grad_norm": 0.7368932366371155, |
| "learning_rate": 0.0003002849433351322, |
| "loss": 3.33, |
| "step": 46400 |
| }, |
| { |
| "epoch": 5.008086253369272, |
| "grad_norm": 0.7171549201011658, |
| "learning_rate": 0.00029996114409066373, |
| "loss": 3.3049, |
| "step": 46450 |
| }, |
| { |
| "epoch": 5.013477088948787, |
| "grad_norm": 0.6396157145500183, |
| "learning_rate": 0.00029963734484619533, |
| "loss": 3.307, |
| "step": 46500 |
| }, |
| { |
| "epoch": 5.018867924528302, |
| "grad_norm": 0.6665785312652588, |
| "learning_rate": 0.0002993135456017269, |
| "loss": 3.3074, |
| "step": 46550 |
| }, |
| { |
| "epoch": 5.024258760107816, |
| "grad_norm": 0.6311437487602234, |
| "learning_rate": 0.0002989897463572585, |
| "loss": 3.2912, |
| "step": 46600 |
| }, |
| { |
| "epoch": 5.0296495956873315, |
| "grad_norm": 0.6656121611595154, |
| "learning_rate": 0.00029866594711279004, |
| "loss": 3.2958, |
| "step": 46650 |
| }, |
| { |
| "epoch": 5.035040431266847, |
| "grad_norm": 0.6491976380348206, |
| "learning_rate": 0.00029834214786832164, |
| "loss": 3.2989, |
| "step": 46700 |
| }, |
| { |
| "epoch": 5.040431266846361, |
| "grad_norm": 0.6659978628158569, |
| "learning_rate": 0.0002980183486238532, |
| "loss": 3.3037, |
| "step": 46750 |
| }, |
| { |
| "epoch": 5.045822102425876, |
| "grad_norm": 0.7180138826370239, |
| "learning_rate": 0.00029769454937938474, |
| "loss": 3.3127, |
| "step": 46800 |
| }, |
| { |
| "epoch": 5.051212938005391, |
| "grad_norm": 0.6873078346252441, |
| "learning_rate": 0.0002973707501349163, |
| "loss": 3.3127, |
| "step": 46850 |
| }, |
| { |
| "epoch": 5.056603773584905, |
| "grad_norm": 0.6599047183990479, |
| "learning_rate": 0.0002970469508904479, |
| "loss": 3.3074, |
| "step": 46900 |
| }, |
| { |
| "epoch": 5.061994609164421, |
| "grad_norm": 0.6925784349441528, |
| "learning_rate": 0.00029672315164597945, |
| "loss": 3.3031, |
| "step": 46950 |
| }, |
| { |
| "epoch": 5.067385444743936, |
| "grad_norm": 0.6558621525764465, |
| "learning_rate": 0.00029639935240151105, |
| "loss": 3.3161, |
| "step": 47000 |
| }, |
| { |
| "epoch": 5.067385444743936, |
| "eval_accuracy": 0.38096636095727976, |
| "eval_loss": 3.4122540950775146, |
| "eval_runtime": 183.2457, |
| "eval_samples_per_second": 98.289, |
| "eval_steps_per_second": 6.145, |
| "step": 47000 |
| }, |
| { |
| "epoch": 5.07277628032345, |
| "grad_norm": 0.67520672082901, |
| "learning_rate": 0.0002960755531570426, |
| "loss": 3.3086, |
| "step": 47050 |
| }, |
| { |
| "epoch": 5.078167115902965, |
| "grad_norm": 0.6475718021392822, |
| "learning_rate": 0.0002957517539125742, |
| "loss": 3.3043, |
| "step": 47100 |
| }, |
| { |
| "epoch": 5.083557951482479, |
| "grad_norm": 0.6276121735572815, |
| "learning_rate": 0.00029542795466810576, |
| "loss": 3.3149, |
| "step": 47150 |
| }, |
| { |
| "epoch": 5.0889487870619945, |
| "grad_norm": 0.6585389375686646, |
| "learning_rate": 0.0002951041554236373, |
| "loss": 3.3141, |
| "step": 47200 |
| }, |
| { |
| "epoch": 5.09433962264151, |
| "grad_norm": 0.654176652431488, |
| "learning_rate": 0.00029478035617916886, |
| "loss": 3.2993, |
| "step": 47250 |
| }, |
| { |
| "epoch": 5.099730458221024, |
| "grad_norm": 0.6916609406471252, |
| "learning_rate": 0.00029445655693470047, |
| "loss": 3.3228, |
| "step": 47300 |
| }, |
| { |
| "epoch": 5.105121293800539, |
| "grad_norm": 0.6311585903167725, |
| "learning_rate": 0.000294132757690232, |
| "loss": 3.3229, |
| "step": 47350 |
| }, |
| { |
| "epoch": 5.110512129380054, |
| "grad_norm": 0.6877700090408325, |
| "learning_rate": 0.0002938089584457636, |
| "loss": 3.3043, |
| "step": 47400 |
| }, |
| { |
| "epoch": 5.115902964959568, |
| "grad_norm": 0.7259128093719482, |
| "learning_rate": 0.00029348515920129517, |
| "loss": 3.3085, |
| "step": 47450 |
| }, |
| { |
| "epoch": 5.121293800539084, |
| "grad_norm": 0.6702103018760681, |
| "learning_rate": 0.0002931613599568267, |
| "loss": 3.2989, |
| "step": 47500 |
| }, |
| { |
| "epoch": 5.126684636118599, |
| "grad_norm": 0.6595789790153503, |
| "learning_rate": 0.00029283756071235833, |
| "loss": 3.3049, |
| "step": 47550 |
| }, |
| { |
| "epoch": 5.132075471698113, |
| "grad_norm": 0.746311366558075, |
| "learning_rate": 0.0002925137614678899, |
| "loss": 3.2956, |
| "step": 47600 |
| }, |
| { |
| "epoch": 5.137466307277628, |
| "grad_norm": 0.714484691619873, |
| "learning_rate": 0.00029218996222342143, |
| "loss": 3.3097, |
| "step": 47650 |
| }, |
| { |
| "epoch": 5.142857142857143, |
| "grad_norm": 0.6939281821250916, |
| "learning_rate": 0.00029186616297895303, |
| "loss": 3.3248, |
| "step": 47700 |
| }, |
| { |
| "epoch": 5.1482479784366575, |
| "grad_norm": 0.7023195028305054, |
| "learning_rate": 0.0002915423637344846, |
| "loss": 3.3148, |
| "step": 47750 |
| }, |
| { |
| "epoch": 5.153638814016173, |
| "grad_norm": 0.633255660533905, |
| "learning_rate": 0.00029121856449001613, |
| "loss": 3.3001, |
| "step": 47800 |
| }, |
| { |
| "epoch": 5.159029649595688, |
| "grad_norm": 0.7074955105781555, |
| "learning_rate": 0.00029089476524554774, |
| "loss": 3.3097, |
| "step": 47850 |
| }, |
| { |
| "epoch": 5.164420485175202, |
| "grad_norm": 0.7120252251625061, |
| "learning_rate": 0.0002905709660010793, |
| "loss": 3.3325, |
| "step": 47900 |
| }, |
| { |
| "epoch": 5.169811320754717, |
| "grad_norm": 0.6911315321922302, |
| "learning_rate": 0.0002902471667566109, |
| "loss": 3.3164, |
| "step": 47950 |
| }, |
| { |
| "epoch": 5.175202156334231, |
| "grad_norm": 0.689180850982666, |
| "learning_rate": 0.00028992336751214245, |
| "loss": 3.3084, |
| "step": 48000 |
| }, |
| { |
| "epoch": 5.175202156334231, |
| "eval_accuracy": 0.3809488678423416, |
| "eval_loss": 3.4123737812042236, |
| "eval_runtime": 183.5311, |
| "eval_samples_per_second": 98.136, |
| "eval_steps_per_second": 6.135, |
| "step": 48000 |
| }, |
| { |
| "epoch": 5.180592991913747, |
| "grad_norm": 0.6789995431900024, |
| "learning_rate": 0.00028959956826767405, |
| "loss": 3.3138, |
| "step": 48050 |
| }, |
| { |
| "epoch": 5.185983827493262, |
| "grad_norm": 0.6542081832885742, |
| "learning_rate": 0.00028927576902320555, |
| "loss": 3.305, |
| "step": 48100 |
| }, |
| { |
| "epoch": 5.191374663072776, |
| "grad_norm": 0.6842049956321716, |
| "learning_rate": 0.00028895196977873715, |
| "loss": 3.3137, |
| "step": 48150 |
| }, |
| { |
| "epoch": 5.196765498652291, |
| "grad_norm": 0.6928696632385254, |
| "learning_rate": 0.0002886346465191581, |
| "loss": 3.3279, |
| "step": 48200 |
| }, |
| { |
| "epoch": 5.202156334231806, |
| "grad_norm": 0.6878511905670166, |
| "learning_rate": 0.00028831084727468965, |
| "loss": 3.32, |
| "step": 48250 |
| }, |
| { |
| "epoch": 5.2075471698113205, |
| "grad_norm": 0.7233080267906189, |
| "learning_rate": 0.00028798704803022126, |
| "loss": 3.3229, |
| "step": 48300 |
| }, |
| { |
| "epoch": 5.212938005390836, |
| "grad_norm": 0.7335208058357239, |
| "learning_rate": 0.0002876632487857528, |
| "loss": 3.3286, |
| "step": 48350 |
| }, |
| { |
| "epoch": 5.218328840970351, |
| "grad_norm": 0.7119630575180054, |
| "learning_rate": 0.0002873394495412844, |
| "loss": 3.3241, |
| "step": 48400 |
| }, |
| { |
| "epoch": 5.223719676549865, |
| "grad_norm": 0.6260629892349243, |
| "learning_rate": 0.0002870156502968159, |
| "loss": 3.3137, |
| "step": 48450 |
| }, |
| { |
| "epoch": 5.22911051212938, |
| "grad_norm": 0.6813670992851257, |
| "learning_rate": 0.0002866918510523475, |
| "loss": 3.3479, |
| "step": 48500 |
| }, |
| { |
| "epoch": 5.234501347708895, |
| "grad_norm": 0.6499662399291992, |
| "learning_rate": 0.00028636805180787907, |
| "loss": 3.3214, |
| "step": 48550 |
| }, |
| { |
| "epoch": 5.2398921832884096, |
| "grad_norm": 0.6558360457420349, |
| "learning_rate": 0.00028604425256341067, |
| "loss": 3.3194, |
| "step": 48600 |
| }, |
| { |
| "epoch": 5.245283018867925, |
| "grad_norm": 0.6462336778640747, |
| "learning_rate": 0.0002857204533189422, |
| "loss": 3.3194, |
| "step": 48650 |
| }, |
| { |
| "epoch": 5.250673854447439, |
| "grad_norm": 0.7324293851852417, |
| "learning_rate": 0.0002853966540744738, |
| "loss": 3.3235, |
| "step": 48700 |
| }, |
| { |
| "epoch": 5.256064690026954, |
| "grad_norm": 0.6889777779579163, |
| "learning_rate": 0.0002850728548300054, |
| "loss": 3.3417, |
| "step": 48750 |
| }, |
| { |
| "epoch": 5.261455525606469, |
| "grad_norm": 0.6718788743019104, |
| "learning_rate": 0.000284749055585537, |
| "loss": 3.3417, |
| "step": 48800 |
| }, |
| { |
| "epoch": 5.2668463611859835, |
| "grad_norm": 0.6515158414840698, |
| "learning_rate": 0.0002844252563410685, |
| "loss": 3.3182, |
| "step": 48850 |
| }, |
| { |
| "epoch": 5.272237196765499, |
| "grad_norm": 0.6929355263710022, |
| "learning_rate": 0.0002841014570966001, |
| "loss": 3.3162, |
| "step": 48900 |
| }, |
| { |
| "epoch": 5.277628032345014, |
| "grad_norm": 0.6629584431648254, |
| "learning_rate": 0.00028377765785213163, |
| "loss": 3.3398, |
| "step": 48950 |
| }, |
| { |
| "epoch": 5.283018867924528, |
| "grad_norm": 0.6952082514762878, |
| "learning_rate": 0.00028345385860766324, |
| "loss": 3.329, |
| "step": 49000 |
| }, |
| { |
| "epoch": 5.283018867924528, |
| "eval_accuracy": 0.3814610575554991, |
| "eval_loss": 3.408320665359497, |
| "eval_runtime": 183.4228, |
| "eval_samples_per_second": 98.194, |
| "eval_steps_per_second": 6.139, |
| "step": 49000 |
| }, |
| { |
| "epoch": 5.288409703504043, |
| "grad_norm": 0.6505891680717468, |
| "learning_rate": 0.0002831300593631948, |
| "loss": 3.3372, |
| "step": 49050 |
| }, |
| { |
| "epoch": 5.293800539083558, |
| "grad_norm": 0.6779699325561523, |
| "learning_rate": 0.0002828062601187264, |
| "loss": 3.3338, |
| "step": 49100 |
| }, |
| { |
| "epoch": 5.2991913746630726, |
| "grad_norm": 0.6804792881011963, |
| "learning_rate": 0.00028248246087425794, |
| "loss": 3.3319, |
| "step": 49150 |
| }, |
| { |
| "epoch": 5.304582210242588, |
| "grad_norm": 0.7583218812942505, |
| "learning_rate": 0.0002821586616297895, |
| "loss": 3.3396, |
| "step": 49200 |
| }, |
| { |
| "epoch": 5.309973045822103, |
| "grad_norm": 0.6410008668899536, |
| "learning_rate": 0.0002818348623853211, |
| "loss": 3.3272, |
| "step": 49250 |
| }, |
| { |
| "epoch": 5.315363881401617, |
| "grad_norm": 0.7008423805236816, |
| "learning_rate": 0.00028151106314085265, |
| "loss": 3.3286, |
| "step": 49300 |
| }, |
| { |
| "epoch": 5.320754716981132, |
| "grad_norm": 0.7405087351799011, |
| "learning_rate": 0.0002811872638963842, |
| "loss": 3.34, |
| "step": 49350 |
| }, |
| { |
| "epoch": 5.3261455525606465, |
| "grad_norm": 0.7460556030273438, |
| "learning_rate": 0.0002808634646519158, |
| "loss": 3.3254, |
| "step": 49400 |
| }, |
| { |
| "epoch": 5.331536388140162, |
| "grad_norm": 0.6831420660018921, |
| "learning_rate": 0.00028053966540744736, |
| "loss": 3.3177, |
| "step": 49450 |
| }, |
| { |
| "epoch": 5.336927223719677, |
| "grad_norm": 0.6601572036743164, |
| "learning_rate": 0.0002802158661629789, |
| "loss": 3.3444, |
| "step": 49500 |
| }, |
| { |
| "epoch": 5.342318059299191, |
| "grad_norm": 0.667539119720459, |
| "learning_rate": 0.0002798920669185105, |
| "loss": 3.3185, |
| "step": 49550 |
| }, |
| { |
| "epoch": 5.347708894878706, |
| "grad_norm": 0.69802325963974, |
| "learning_rate": 0.00027956826767404206, |
| "loss": 3.3355, |
| "step": 49600 |
| }, |
| { |
| "epoch": 5.353099730458221, |
| "grad_norm": 0.668803334236145, |
| "learning_rate": 0.00027924446842957367, |
| "loss": 3.3321, |
| "step": 49650 |
| }, |
| { |
| "epoch": 5.3584905660377355, |
| "grad_norm": 0.6525208950042725, |
| "learning_rate": 0.0002789206691851052, |
| "loss": 3.3391, |
| "step": 49700 |
| }, |
| { |
| "epoch": 5.363881401617251, |
| "grad_norm": 0.7439292669296265, |
| "learning_rate": 0.00027859686994063677, |
| "loss": 3.3249, |
| "step": 49750 |
| }, |
| { |
| "epoch": 5.369272237196766, |
| "grad_norm": 0.669330894947052, |
| "learning_rate": 0.0002782795466810577, |
| "loss": 3.3276, |
| "step": 49800 |
| }, |
| { |
| "epoch": 5.37466307277628, |
| "grad_norm": 0.6631773710250854, |
| "learning_rate": 0.00027795574743658927, |
| "loss": 3.3419, |
| "step": 49850 |
| }, |
| { |
| "epoch": 5.380053908355795, |
| "grad_norm": 0.6949469447135925, |
| "learning_rate": 0.0002776319481921209, |
| "loss": 3.336, |
| "step": 49900 |
| }, |
| { |
| "epoch": 5.38544474393531, |
| "grad_norm": 0.6591998934745789, |
| "learning_rate": 0.0002773081489476524, |
| "loss": 3.3223, |
| "step": 49950 |
| }, |
| { |
| "epoch": 5.390835579514825, |
| "grad_norm": 0.6729640960693359, |
| "learning_rate": 0.00027698434970318403, |
| "loss": 3.3415, |
| "step": 50000 |
| }, |
| { |
| "epoch": 5.390835579514825, |
| "eval_accuracy": 0.3817132409081165, |
| "eval_loss": 3.4021902084350586, |
| "eval_runtime": 183.2074, |
| "eval_samples_per_second": 98.309, |
| "eval_steps_per_second": 6.146, |
| "step": 50000 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 92750, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 10000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 4.18034589696e+17, |
| "train_batch_size": 32, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|