| { | |
| "best_metric": 0.08146847784519196, | |
| "best_model_checkpoint": "./fine-tuned/checkpoint-25000", | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 25000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004, | |
| "grad_norm": 0.24035809934139252, | |
| "learning_rate": 2.99412e-05, | |
| "loss": 0.3973, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.008, | |
| "grad_norm": 0.39832672476768494, | |
| "learning_rate": 2.98812e-05, | |
| "loss": 0.1126, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.012, | |
| "grad_norm": 0.2938326299190521, | |
| "learning_rate": 2.9821200000000002e-05, | |
| "loss": 0.0932, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.016, | |
| "grad_norm": 0.18936102092266083, | |
| "learning_rate": 2.9761200000000002e-05, | |
| "loss": 0.0897, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.18386273086071014, | |
| "learning_rate": 2.9701200000000003e-05, | |
| "loss": 0.092, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.024, | |
| "grad_norm": 0.1968618482351303, | |
| "learning_rate": 2.96412e-05, | |
| "loss": 0.0845, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.028, | |
| "grad_norm": 0.18028958141803741, | |
| "learning_rate": 2.95812e-05, | |
| "loss": 0.0792, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.032, | |
| "grad_norm": 0.2615596652030945, | |
| "learning_rate": 2.95212e-05, | |
| "loss": 0.0796, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.036, | |
| "grad_norm": 0.1913922131061554, | |
| "learning_rate": 2.9461200000000002e-05, | |
| "loss": 0.0798, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.2351102977991104, | |
| "learning_rate": 2.9401200000000002e-05, | |
| "loss": 0.081, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "eval_loss": 0.09441258758306503, | |
| "eval_runtime": 88.0185, | |
| "eval_samples_per_second": 22.722, | |
| "eval_steps_per_second": 5.681, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.044, | |
| "grad_norm": 0.1859619915485382, | |
| "learning_rate": 2.9341200000000003e-05, | |
| "loss": 0.077, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.048, | |
| "grad_norm": 0.11936317384243011, | |
| "learning_rate": 2.92812e-05, | |
| "loss": 0.0727, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.052, | |
| "grad_norm": 0.2207396775484085, | |
| "learning_rate": 2.92212e-05, | |
| "loss": 0.0743, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.056, | |
| "grad_norm": 0.18488994240760803, | |
| "learning_rate": 2.91612e-05, | |
| "loss": 0.0824, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.22228538990020752, | |
| "learning_rate": 2.9101200000000002e-05, | |
| "loss": 0.0716, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.064, | |
| "grad_norm": 0.24138867855072021, | |
| "learning_rate": 2.9041200000000002e-05, | |
| "loss": 0.0814, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.068, | |
| "grad_norm": 0.25113552808761597, | |
| "learning_rate": 2.89812e-05, | |
| "loss": 0.076, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.072, | |
| "grad_norm": 0.8853724598884583, | |
| "learning_rate": 2.89212e-05, | |
| "loss": 0.0781, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.076, | |
| "grad_norm": 0.1753206104040146, | |
| "learning_rate": 2.88612e-05, | |
| "loss": 0.084, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.1704334318637848, | |
| "learning_rate": 2.88012e-05, | |
| "loss": 0.0769, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "eval_loss": 0.09160277992486954, | |
| "eval_runtime": 88.0495, | |
| "eval_samples_per_second": 22.714, | |
| "eval_steps_per_second": 5.679, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.084, | |
| "grad_norm": 0.16729697585105896, | |
| "learning_rate": 2.8741200000000002e-05, | |
| "loss": 0.0823, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.088, | |
| "grad_norm": 0.14851506054401398, | |
| "learning_rate": 2.86812e-05, | |
| "loss": 0.0785, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.092, | |
| "grad_norm": 0.22481797635555267, | |
| "learning_rate": 2.86212e-05, | |
| "loss": 0.0786, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.096, | |
| "grad_norm": 0.13808289170265198, | |
| "learning_rate": 2.85612e-05, | |
| "loss": 0.0785, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.17833128571510315, | |
| "learning_rate": 2.85012e-05, | |
| "loss": 0.0737, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.104, | |
| "grad_norm": 0.14926594495773315, | |
| "learning_rate": 2.84412e-05, | |
| "loss": 0.0767, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.108, | |
| "grad_norm": 0.19346196949481964, | |
| "learning_rate": 2.8381200000000002e-05, | |
| "loss": 0.077, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.112, | |
| "grad_norm": 0.2675027847290039, | |
| "learning_rate": 2.83212e-05, | |
| "loss": 0.0806, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.116, | |
| "grad_norm": 0.19048169255256653, | |
| "learning_rate": 2.82612e-05, | |
| "loss": 0.0756, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.1875162124633789, | |
| "learning_rate": 2.82012e-05, | |
| "loss": 0.0823, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "eval_loss": 0.09038107097148895, | |
| "eval_runtime": 87.8777, | |
| "eval_samples_per_second": 22.759, | |
| "eval_steps_per_second": 5.69, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.124, | |
| "grad_norm": 0.19083499908447266, | |
| "learning_rate": 2.81412e-05, | |
| "loss": 0.0736, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.128, | |
| "grad_norm": 0.19123569130897522, | |
| "learning_rate": 2.80812e-05, | |
| "loss": 0.0766, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.132, | |
| "grad_norm": 0.24691827595233917, | |
| "learning_rate": 2.80212e-05, | |
| "loss": 0.0798, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.136, | |
| "grad_norm": 0.17910048365592957, | |
| "learning_rate": 2.79612e-05, | |
| "loss": 0.0687, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.1740667223930359, | |
| "learning_rate": 2.79012e-05, | |
| "loss": 0.0758, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.144, | |
| "grad_norm": 0.15178219974040985, | |
| "learning_rate": 2.78412e-05, | |
| "loss": 0.0732, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.148, | |
| "grad_norm": 0.1904926896095276, | |
| "learning_rate": 2.77812e-05, | |
| "loss": 0.0734, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.152, | |
| "grad_norm": 0.2795208990573883, | |
| "learning_rate": 2.77212e-05, | |
| "loss": 0.076, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.156, | |
| "grad_norm": 0.18160228431224823, | |
| "learning_rate": 2.76612e-05, | |
| "loss": 0.0716, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.15877611935138702, | |
| "learning_rate": 2.76012e-05, | |
| "loss": 0.0773, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_loss": 0.08910445868968964, | |
| "eval_runtime": 87.8635, | |
| "eval_samples_per_second": 22.763, | |
| "eval_steps_per_second": 5.691, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.164, | |
| "grad_norm": 0.2411368191242218, | |
| "learning_rate": 2.75412e-05, | |
| "loss": 0.0786, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.168, | |
| "grad_norm": 0.16663742065429688, | |
| "learning_rate": 2.74812e-05, | |
| "loss": 0.0724, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.172, | |
| "grad_norm": 0.23420193791389465, | |
| "learning_rate": 2.74212e-05, | |
| "loss": 0.0653, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.176, | |
| "grad_norm": 0.1807372272014618, | |
| "learning_rate": 2.7361199999999998e-05, | |
| "loss": 0.0676, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.16474364697933197, | |
| "learning_rate": 2.73012e-05, | |
| "loss": 0.0767, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.184, | |
| "grad_norm": 0.17184095084667206, | |
| "learning_rate": 2.72412e-05, | |
| "loss": 0.0658, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.188, | |
| "grad_norm": 0.16993258893489838, | |
| "learning_rate": 2.71812e-05, | |
| "loss": 0.0755, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.192, | |
| "grad_norm": 0.1555277407169342, | |
| "learning_rate": 2.71212e-05, | |
| "loss": 0.0698, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.196, | |
| "grad_norm": 0.09040562808513641, | |
| "learning_rate": 2.7061199999999998e-05, | |
| "loss": 0.0757, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.12910398840904236, | |
| "learning_rate": 2.7001199999999998e-05, | |
| "loss": 0.0688, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "eval_loss": 0.08841572701931, | |
| "eval_runtime": 87.7555, | |
| "eval_samples_per_second": 22.791, | |
| "eval_steps_per_second": 5.698, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.204, | |
| "grad_norm": 0.14937053620815277, | |
| "learning_rate": 2.69412e-05, | |
| "loss": 0.0727, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.208, | |
| "grad_norm": 0.15660254657268524, | |
| "learning_rate": 2.68812e-05, | |
| "loss": 0.0656, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.212, | |
| "grad_norm": 0.09593763947486877, | |
| "learning_rate": 2.68212e-05, | |
| "loss": 0.0726, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.216, | |
| "grad_norm": 0.25192323327064514, | |
| "learning_rate": 2.67624e-05, | |
| "loss": 0.0753, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.1267642378807068, | |
| "learning_rate": 2.67024e-05, | |
| "loss": 0.0707, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.224, | |
| "grad_norm": 0.13844658434391022, | |
| "learning_rate": 2.66424e-05, | |
| "loss": 0.0711, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.228, | |
| "grad_norm": 0.15095186233520508, | |
| "learning_rate": 2.65824e-05, | |
| "loss": 0.0696, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.232, | |
| "grad_norm": 0.09553442895412445, | |
| "learning_rate": 2.65224e-05, | |
| "loss": 0.0709, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.236, | |
| "grad_norm": 0.21425922214984894, | |
| "learning_rate": 2.64624e-05, | |
| "loss": 0.0662, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.10706017166376114, | |
| "learning_rate": 2.64024e-05, | |
| "loss": 0.0721, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "eval_loss": 0.08755213767290115, | |
| "eval_runtime": 87.7619, | |
| "eval_samples_per_second": 22.789, | |
| "eval_steps_per_second": 5.697, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.244, | |
| "grad_norm": 0.18366344273090363, | |
| "learning_rate": 2.63424e-05, | |
| "loss": 0.0781, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.248, | |
| "grad_norm": 0.15975314378738403, | |
| "learning_rate": 2.62824e-05, | |
| "loss": 0.0771, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.252, | |
| "grad_norm": 0.14510446786880493, | |
| "learning_rate": 2.6222399999999998e-05, | |
| "loss": 0.0734, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.256, | |
| "grad_norm": 0.10040156543254852, | |
| "learning_rate": 2.61624e-05, | |
| "loss": 0.0624, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.1453912854194641, | |
| "learning_rate": 2.61024e-05, | |
| "loss": 0.0661, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.264, | |
| "grad_norm": 0.13999666273593903, | |
| "learning_rate": 2.60424e-05, | |
| "loss": 0.0694, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.268, | |
| "grad_norm": 0.13396582007408142, | |
| "learning_rate": 2.59824e-05, | |
| "loss": 0.0692, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.272, | |
| "grad_norm": 0.1334969699382782, | |
| "learning_rate": 2.59224e-05, | |
| "loss": 0.0629, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.276, | |
| "grad_norm": 0.16296976804733276, | |
| "learning_rate": 2.5862399999999998e-05, | |
| "loss": 0.0679, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.1321544647216797, | |
| "learning_rate": 2.58024e-05, | |
| "loss": 0.0764, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "eval_loss": 0.0869474709033966, | |
| "eval_runtime": 87.6679, | |
| "eval_samples_per_second": 22.813, | |
| "eval_steps_per_second": 5.703, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.284, | |
| "grad_norm": 0.13956592977046967, | |
| "learning_rate": 2.57424e-05, | |
| "loss": 0.0749, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.288, | |
| "grad_norm": 0.2318839579820633, | |
| "learning_rate": 2.56824e-05, | |
| "loss": 0.0658, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.292, | |
| "grad_norm": 0.10267651081085205, | |
| "learning_rate": 2.56224e-05, | |
| "loss": 0.0651, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.296, | |
| "grad_norm": 0.13903647661209106, | |
| "learning_rate": 2.5562399999999998e-05, | |
| "loss": 0.0684, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.07369179278612137, | |
| "learning_rate": 2.5502399999999998e-05, | |
| "loss": 0.0643, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.304, | |
| "grad_norm": 0.18283645808696747, | |
| "learning_rate": 2.54424e-05, | |
| "loss": 0.0745, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.308, | |
| "grad_norm": 0.1031743511557579, | |
| "learning_rate": 2.5382400000000003e-05, | |
| "loss": 0.0678, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.312, | |
| "grad_norm": 0.2246129959821701, | |
| "learning_rate": 2.5322400000000003e-05, | |
| "loss": 0.0701, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.316, | |
| "grad_norm": 0.1434032917022705, | |
| "learning_rate": 2.52624e-05, | |
| "loss": 0.071, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.11730857193470001, | |
| "learning_rate": 2.52024e-05, | |
| "loss": 0.0696, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "eval_loss": 0.0867869183421135, | |
| "eval_runtime": 87.7801, | |
| "eval_samples_per_second": 22.784, | |
| "eval_steps_per_second": 5.696, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.324, | |
| "grad_norm": 0.07968447357416153, | |
| "learning_rate": 2.5142400000000002e-05, | |
| "loss": 0.0635, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.328, | |
| "grad_norm": 0.1641731858253479, | |
| "learning_rate": 2.5082400000000002e-05, | |
| "loss": 0.0667, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.332, | |
| "grad_norm": 0.15083415806293488, | |
| "learning_rate": 2.5022400000000003e-05, | |
| "loss": 0.071, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.336, | |
| "grad_norm": 0.16534006595611572, | |
| "learning_rate": 2.4962400000000003e-05, | |
| "loss": 0.0735, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.1629945933818817, | |
| "learning_rate": 2.49024e-05, | |
| "loss": 0.0641, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.344, | |
| "grad_norm": 0.11393357813358307, | |
| "learning_rate": 2.48424e-05, | |
| "loss": 0.0679, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.348, | |
| "grad_norm": 0.10581399500370026, | |
| "learning_rate": 2.4782400000000002e-05, | |
| "loss": 0.0666, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.352, | |
| "grad_norm": 0.1571837067604065, | |
| "learning_rate": 2.4722400000000002e-05, | |
| "loss": 0.0619, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.356, | |
| "grad_norm": 0.11944606155157089, | |
| "learning_rate": 2.4662400000000003e-05, | |
| "loss": 0.0666, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.17318391799926758, | |
| "learning_rate": 2.46024e-05, | |
| "loss": 0.0684, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "eval_loss": 0.08653330057859421, | |
| "eval_runtime": 87.7823, | |
| "eval_samples_per_second": 22.784, | |
| "eval_steps_per_second": 5.696, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.364, | |
| "grad_norm": 0.09011202305555344, | |
| "learning_rate": 2.45424e-05, | |
| "loss": 0.0699, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.368, | |
| "grad_norm": 0.1441943347454071, | |
| "learning_rate": 2.44824e-05, | |
| "loss": 0.0634, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.372, | |
| "grad_norm": 0.17736917734146118, | |
| "learning_rate": 2.4422400000000002e-05, | |
| "loss": 0.0688, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.376, | |
| "grad_norm": 0.16143649816513062, | |
| "learning_rate": 2.4362400000000002e-05, | |
| "loss": 0.0575, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.1789257675409317, | |
| "learning_rate": 2.43024e-05, | |
| "loss": 0.067, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.384, | |
| "grad_norm": 0.13589045405387878, | |
| "learning_rate": 2.42424e-05, | |
| "loss": 0.0713, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.388, | |
| "grad_norm": 0.12760789692401886, | |
| "learning_rate": 2.41824e-05, | |
| "loss": 0.0644, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.392, | |
| "grad_norm": 0.17507490515708923, | |
| "learning_rate": 2.41224e-05, | |
| "loss": 0.0634, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.396, | |
| "grad_norm": 0.15503354370594025, | |
| "learning_rate": 2.4062400000000002e-05, | |
| "loss": 0.0654, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.12309867143630981, | |
| "learning_rate": 2.4002400000000002e-05, | |
| "loss": 0.0698, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "eval_loss": 0.0860014408826828, | |
| "eval_runtime": 87.7287, | |
| "eval_samples_per_second": 22.798, | |
| "eval_steps_per_second": 5.699, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.404, | |
| "grad_norm": 0.1547604650259018, | |
| "learning_rate": 2.39424e-05, | |
| "loss": 0.0691, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.408, | |
| "grad_norm": 0.17712494730949402, | |
| "learning_rate": 2.38824e-05, | |
| "loss": 0.064, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.412, | |
| "grad_norm": 0.16606800258159637, | |
| "learning_rate": 2.38224e-05, | |
| "loss": 0.0687, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.416, | |
| "grad_norm": 0.18862195312976837, | |
| "learning_rate": 2.37624e-05, | |
| "loss": 0.0662, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.1778397411108017, | |
| "learning_rate": 2.3702400000000002e-05, | |
| "loss": 0.0632, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.424, | |
| "grad_norm": 0.15363118052482605, | |
| "learning_rate": 2.36424e-05, | |
| "loss": 0.0617, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.428, | |
| "grad_norm": 0.1690954864025116, | |
| "learning_rate": 2.35824e-05, | |
| "loss": 0.0717, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 0.432, | |
| "grad_norm": 0.19384829699993134, | |
| "learning_rate": 2.35224e-05, | |
| "loss": 0.0669, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.436, | |
| "grad_norm": 0.20759643614292145, | |
| "learning_rate": 2.34624e-05, | |
| "loss": 0.0662, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.16528765857219696, | |
| "learning_rate": 2.34024e-05, | |
| "loss": 0.0737, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "eval_loss": 0.08539459854364395, | |
| "eval_runtime": 87.7109, | |
| "eval_samples_per_second": 22.802, | |
| "eval_steps_per_second": 5.701, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.444, | |
| "grad_norm": 0.09632231295108795, | |
| "learning_rate": 2.3342400000000002e-05, | |
| "loss": 0.0621, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 0.448, | |
| "grad_norm": 0.15080232918262482, | |
| "learning_rate": 2.32824e-05, | |
| "loss": 0.0674, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.452, | |
| "grad_norm": 0.14794333279132843, | |
| "learning_rate": 2.32224e-05, | |
| "loss": 0.0662, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 0.456, | |
| "grad_norm": 0.15014077723026276, | |
| "learning_rate": 2.31624e-05, | |
| "loss": 0.0574, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.10953331738710403, | |
| "learning_rate": 2.31024e-05, | |
| "loss": 0.0684, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 0.464, | |
| "grad_norm": 0.17434537410736084, | |
| "learning_rate": 2.30424e-05, | |
| "loss": 0.0687, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.468, | |
| "grad_norm": 0.12245655059814453, | |
| "learning_rate": 2.29824e-05, | |
| "loss": 0.0689, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 0.472, | |
| "grad_norm": 0.10118559747934341, | |
| "learning_rate": 2.29224e-05, | |
| "loss": 0.0609, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.476, | |
| "grad_norm": 0.14768172800540924, | |
| "learning_rate": 2.28624e-05, | |
| "loss": 0.0717, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.18510164320468903, | |
| "learning_rate": 2.28024e-05, | |
| "loss": 0.0716, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "eval_loss": 0.08568704128265381, | |
| "eval_runtime": 87.7082, | |
| "eval_samples_per_second": 22.803, | |
| "eval_steps_per_second": 5.701, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.484, | |
| "grad_norm": 0.09803249686956406, | |
| "learning_rate": 2.27424e-05, | |
| "loss": 0.0693, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 0.488, | |
| "grad_norm": 0.1030198410153389, | |
| "learning_rate": 2.2682399999999998e-05, | |
| "loss": 0.065, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.492, | |
| "grad_norm": 0.15723823010921478, | |
| "learning_rate": 2.26224e-05, | |
| "loss": 0.063, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 0.496, | |
| "grad_norm": 0.1828642040491104, | |
| "learning_rate": 2.25624e-05, | |
| "loss": 0.0665, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.11824677884578705, | |
| "learning_rate": 2.25024e-05, | |
| "loss": 0.0656, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 0.504, | |
| "grad_norm": 0.07376304268836975, | |
| "learning_rate": 2.24424e-05, | |
| "loss": 0.0664, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.508, | |
| "grad_norm": 0.09230540692806244, | |
| "learning_rate": 2.23824e-05, | |
| "loss": 0.0668, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 0.512, | |
| "grad_norm": 0.1208396852016449, | |
| "learning_rate": 2.2322399999999998e-05, | |
| "loss": 0.0642, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.516, | |
| "grad_norm": 0.12650001049041748, | |
| "learning_rate": 2.22624e-05, | |
| "loss": 0.0656, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.11461616307497025, | |
| "learning_rate": 2.22024e-05, | |
| "loss": 0.0676, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "eval_loss": 0.08497656136751175, | |
| "eval_runtime": 87.668, | |
| "eval_samples_per_second": 22.813, | |
| "eval_steps_per_second": 5.703, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.524, | |
| "grad_norm": 0.24468739330768585, | |
| "learning_rate": 2.21424e-05, | |
| "loss": 0.0688, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 0.528, | |
| "grad_norm": 0.17887485027313232, | |
| "learning_rate": 2.20824e-05, | |
| "loss": 0.0645, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.532, | |
| "grad_norm": 0.12986980378627777, | |
| "learning_rate": 2.2022399999999998e-05, | |
| "loss": 0.0609, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 0.536, | |
| "grad_norm": 0.25361281633377075, | |
| "learning_rate": 2.1962399999999998e-05, | |
| "loss": 0.0603, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.1815791130065918, | |
| "learning_rate": 2.19024e-05, | |
| "loss": 0.0659, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 0.544, | |
| "grad_norm": 0.12782719731330872, | |
| "learning_rate": 2.18424e-05, | |
| "loss": 0.0641, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.548, | |
| "grad_norm": 0.1801528036594391, | |
| "learning_rate": 2.17824e-05, | |
| "loss": 0.0666, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 0.552, | |
| "grad_norm": 0.1247314065694809, | |
| "learning_rate": 2.17224e-05, | |
| "loss": 0.0592, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.556, | |
| "grad_norm": 0.19411933422088623, | |
| "learning_rate": 2.16624e-05, | |
| "loss": 0.0688, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.11316727846860886, | |
| "learning_rate": 2.1602400000000002e-05, | |
| "loss": 0.0635, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "eval_loss": 0.08500248938798904, | |
| "eval_runtime": 87.7238, | |
| "eval_samples_per_second": 22.799, | |
| "eval_steps_per_second": 5.7, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.564, | |
| "grad_norm": 0.14090943336486816, | |
| "learning_rate": 2.1542400000000002e-05, | |
| "loss": 0.0648, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 0.568, | |
| "grad_norm": 0.16401338577270508, | |
| "learning_rate": 2.1482400000000003e-05, | |
| "loss": 0.064, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.572, | |
| "grad_norm": 0.14420969784259796, | |
| "learning_rate": 2.1422400000000003e-05, | |
| "loss": 0.0627, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 0.576, | |
| "grad_norm": 0.3247956931591034, | |
| "learning_rate": 2.13624e-05, | |
| "loss": 0.0621, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.1672086864709854, | |
| "learning_rate": 2.13024e-05, | |
| "loss": 0.0673, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 0.584, | |
| "grad_norm": 0.16130150854587555, | |
| "learning_rate": 2.1242400000000002e-05, | |
| "loss": 0.0604, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.588, | |
| "grad_norm": 0.201412633061409, | |
| "learning_rate": 2.1182400000000002e-05, | |
| "loss": 0.0687, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 0.592, | |
| "grad_norm": 0.17173218727111816, | |
| "learning_rate": 2.1122400000000003e-05, | |
| "loss": 0.0626, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.596, | |
| "grad_norm": 0.14102816581726074, | |
| "learning_rate": 2.10624e-05, | |
| "loss": 0.0662, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.12381123006343842, | |
| "learning_rate": 2.10024e-05, | |
| "loss": 0.0676, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "eval_loss": 0.08412499725818634, | |
| "eval_runtime": 87.8551, | |
| "eval_samples_per_second": 22.765, | |
| "eval_steps_per_second": 5.691, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.604, | |
| "grad_norm": 0.14278633892536163, | |
| "learning_rate": 2.09424e-05, | |
| "loss": 0.0665, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 0.608, | |
| "grad_norm": 0.19127364456653595, | |
| "learning_rate": 2.0882400000000002e-05, | |
| "loss": 0.0652, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.612, | |
| "grad_norm": 0.18026478588581085, | |
| "learning_rate": 2.0822400000000002e-05, | |
| "loss": 0.0673, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 0.616, | |
| "grad_norm": 0.1525663435459137, | |
| "learning_rate": 2.0762400000000003e-05, | |
| "loss": 0.0596, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.1647537797689438, | |
| "learning_rate": 2.07024e-05, | |
| "loss": 0.0638, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 0.624, | |
| "grad_norm": 0.14861586689949036, | |
| "learning_rate": 2.06424e-05, | |
| "loss": 0.0616, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.628, | |
| "grad_norm": 0.13468614220619202, | |
| "learning_rate": 2.05824e-05, | |
| "loss": 0.0645, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 0.632, | |
| "grad_norm": 0.22103475034236908, | |
| "learning_rate": 2.0522400000000002e-05, | |
| "loss": 0.0618, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.636, | |
| "grad_norm": 0.21467621624469757, | |
| "learning_rate": 2.0462400000000002e-05, | |
| "loss": 0.0609, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.09938893467187881, | |
| "learning_rate": 2.04024e-05, | |
| "loss": 0.0671, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "eval_loss": 0.08401757478713989, | |
| "eval_runtime": 88.0793, | |
| "eval_samples_per_second": 22.707, | |
| "eval_steps_per_second": 5.677, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.644, | |
| "grad_norm": 0.09193145483732224, | |
| "learning_rate": 2.03424e-05, | |
| "loss": 0.0592, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 0.648, | |
| "grad_norm": 0.13023436069488525, | |
| "learning_rate": 2.02824e-05, | |
| "loss": 0.0627, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.652, | |
| "grad_norm": 0.12572939693927765, | |
| "learning_rate": 2.02224e-05, | |
| "loss": 0.0696, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 0.656, | |
| "grad_norm": 0.08949209004640579, | |
| "learning_rate": 2.0162400000000002e-05, | |
| "loss": 0.0633, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.18614652752876282, | |
| "learning_rate": 2.01024e-05, | |
| "loss": 0.0683, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 0.664, | |
| "grad_norm": 0.1969350129365921, | |
| "learning_rate": 2.00424e-05, | |
| "loss": 0.0629, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.668, | |
| "grad_norm": 0.18870118260383606, | |
| "learning_rate": 1.99824e-05, | |
| "loss": 0.0697, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 0.672, | |
| "grad_norm": 0.13722488284111023, | |
| "learning_rate": 1.99224e-05, | |
| "loss": 0.0654, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.676, | |
| "grad_norm": 0.12897425889968872, | |
| "learning_rate": 1.98624e-05, | |
| "loss": 0.0632, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.16306981444358826, | |
| "learning_rate": 1.9802400000000002e-05, | |
| "loss": 0.0606, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "eval_loss": 0.08417751640081406, | |
| "eval_runtime": 88.0586, | |
| "eval_samples_per_second": 22.712, | |
| "eval_steps_per_second": 5.678, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.684, | |
| "grad_norm": 0.113871268928051, | |
| "learning_rate": 1.97424e-05, | |
| "loss": 0.0651, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 0.688, | |
| "grad_norm": 0.12488531321287155, | |
| "learning_rate": 1.96824e-05, | |
| "loss": 0.0624, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.692, | |
| "grad_norm": 0.16593888401985168, | |
| "learning_rate": 1.96224e-05, | |
| "loss": 0.0619, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 0.696, | |
| "grad_norm": 0.14381399750709534, | |
| "learning_rate": 1.95624e-05, | |
| "loss": 0.0631, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.19177569448947906, | |
| "learning_rate": 1.95024e-05, | |
| "loss": 0.0584, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 0.704, | |
| "grad_norm": 0.18918590247631073, | |
| "learning_rate": 1.94424e-05, | |
| "loss": 0.0563, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.708, | |
| "grad_norm": 0.1318553388118744, | |
| "learning_rate": 1.93824e-05, | |
| "loss": 0.0644, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 0.712, | |
| "grad_norm": 0.11326153576374054, | |
| "learning_rate": 1.93224e-05, | |
| "loss": 0.0678, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.716, | |
| "grad_norm": 0.1167350560426712, | |
| "learning_rate": 1.92624e-05, | |
| "loss": 0.0558, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.10959301143884659, | |
| "learning_rate": 1.92024e-05, | |
| "loss": 0.0618, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "eval_loss": 0.0838891863822937, | |
| "eval_runtime": 88.0828, | |
| "eval_samples_per_second": 22.706, | |
| "eval_steps_per_second": 5.676, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.724, | |
| "grad_norm": 0.152841717004776, | |
| "learning_rate": 1.91424e-05, | |
| "loss": 0.0611, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 0.728, | |
| "grad_norm": 0.16737787425518036, | |
| "learning_rate": 1.90824e-05, | |
| "loss": 0.0626, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.732, | |
| "grad_norm": 0.1456371396780014, | |
| "learning_rate": 1.90224e-05, | |
| "loss": 0.0599, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 0.736, | |
| "grad_norm": 0.1772635579109192, | |
| "learning_rate": 1.89624e-05, | |
| "loss": 0.0577, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.11802179366350174, | |
| "learning_rate": 1.89024e-05, | |
| "loss": 0.0645, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 0.744, | |
| "grad_norm": 0.1235092505812645, | |
| "learning_rate": 1.88424e-05, | |
| "loss": 0.0584, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.748, | |
| "grad_norm": 0.11901592463254929, | |
| "learning_rate": 1.8782399999999998e-05, | |
| "loss": 0.06, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 0.752, | |
| "grad_norm": 0.09477788209915161, | |
| "learning_rate": 1.87224e-05, | |
| "loss": 0.0637, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.756, | |
| "grad_norm": 0.12917232513427734, | |
| "learning_rate": 1.86624e-05, | |
| "loss": 0.0648, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.17290474474430084, | |
| "learning_rate": 1.86024e-05, | |
| "loss": 0.0638, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "eval_loss": 0.08371420204639435, | |
| "eval_runtime": 88.0684, | |
| "eval_samples_per_second": 22.71, | |
| "eval_steps_per_second": 5.677, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.764, | |
| "grad_norm": 0.07568900287151337, | |
| "learning_rate": 1.85424e-05, | |
| "loss": 0.0648, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 0.768, | |
| "grad_norm": 0.11715036630630493, | |
| "learning_rate": 1.8482399999999998e-05, | |
| "loss": 0.0644, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.772, | |
| "grad_norm": 0.18466413021087646, | |
| "learning_rate": 1.84224e-05, | |
| "loss": 0.0641, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 0.776, | |
| "grad_norm": 0.15552838146686554, | |
| "learning_rate": 1.83624e-05, | |
| "loss": 0.0538, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.1408630758523941, | |
| "learning_rate": 1.83024e-05, | |
| "loss": 0.0589, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 0.784, | |
| "grad_norm": 0.1492939293384552, | |
| "learning_rate": 1.82424e-05, | |
| "loss": 0.0599, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.788, | |
| "grad_norm": 0.08888087421655655, | |
| "learning_rate": 1.81824e-05, | |
| "loss": 0.0591, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 0.792, | |
| "grad_norm": 0.11865179240703583, | |
| "learning_rate": 1.8122399999999998e-05, | |
| "loss": 0.0581, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.796, | |
| "grad_norm": 0.18105269968509674, | |
| "learning_rate": 1.80624e-05, | |
| "loss": 0.0592, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.11952503770589828, | |
| "learning_rate": 1.80024e-05, | |
| "loss": 0.0567, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "eval_loss": 0.08371368050575256, | |
| "eval_runtime": 88.0598, | |
| "eval_samples_per_second": 22.712, | |
| "eval_steps_per_second": 5.678, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.804, | |
| "grad_norm": 0.16905057430267334, | |
| "learning_rate": 1.79424e-05, | |
| "loss": 0.065, | |
| "step": 10050 | |
| }, | |
| { | |
| "epoch": 0.808, | |
| "grad_norm": 0.11961635947227478, | |
| "learning_rate": 1.7882400000000003e-05, | |
| "loss": 0.0677, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.812, | |
| "grad_norm": 0.13067485392093658, | |
| "learning_rate": 1.78224e-05, | |
| "loss": 0.0611, | |
| "step": 10150 | |
| }, | |
| { | |
| "epoch": 0.816, | |
| "grad_norm": 0.12020887434482574, | |
| "learning_rate": 1.77624e-05, | |
| "loss": 0.0582, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.1354004591703415, | |
| "learning_rate": 1.7702400000000002e-05, | |
| "loss": 0.0592, | |
| "step": 10250 | |
| }, | |
| { | |
| "epoch": 0.824, | |
| "grad_norm": 0.22322942316532135, | |
| "learning_rate": 1.7642400000000002e-05, | |
| "loss": 0.0644, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.828, | |
| "grad_norm": 0.12816545367240906, | |
| "learning_rate": 1.7582400000000003e-05, | |
| "loss": 0.065, | |
| "step": 10350 | |
| }, | |
| { | |
| "epoch": 0.832, | |
| "grad_norm": 0.07579780369997025, | |
| "learning_rate": 1.75224e-05, | |
| "loss": 0.057, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.836, | |
| "grad_norm": 0.10635198652744293, | |
| "learning_rate": 1.74624e-05, | |
| "loss": 0.062, | |
| "step": 10450 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.10341114550828934, | |
| "learning_rate": 1.74024e-05, | |
| "loss": 0.0593, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "eval_loss": 0.08319947868585587, | |
| "eval_runtime": 88.0121, | |
| "eval_samples_per_second": 22.724, | |
| "eval_steps_per_second": 5.681, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.844, | |
| "grad_norm": 0.16710689663887024, | |
| "learning_rate": 1.7342400000000002e-05, | |
| "loss": 0.0616, | |
| "step": 10550 | |
| }, | |
| { | |
| "epoch": 0.848, | |
| "grad_norm": 0.2513883411884308, | |
| "learning_rate": 1.7282400000000002e-05, | |
| "loss": 0.0638, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.852, | |
| "grad_norm": 0.182452991604805, | |
| "learning_rate": 1.7222400000000003e-05, | |
| "loss": 0.0588, | |
| "step": 10650 | |
| }, | |
| { | |
| "epoch": 0.856, | |
| "grad_norm": 0.16390974819660187, | |
| "learning_rate": 1.71624e-05, | |
| "loss": 0.059, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.13543175160884857, | |
| "learning_rate": 1.71024e-05, | |
| "loss": 0.0648, | |
| "step": 10750 | |
| }, | |
| { | |
| "epoch": 0.864, | |
| "grad_norm": 0.12370152771472931, | |
| "learning_rate": 1.70424e-05, | |
| "loss": 0.0625, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.868, | |
| "grad_norm": 0.23662041127681732, | |
| "learning_rate": 1.6982400000000002e-05, | |
| "loss": 0.0667, | |
| "step": 10850 | |
| }, | |
| { | |
| "epoch": 0.872, | |
| "grad_norm": 0.1502092331647873, | |
| "learning_rate": 1.6922400000000003e-05, | |
| "loss": 0.0652, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.876, | |
| "grad_norm": 0.1745540052652359, | |
| "learning_rate": 1.68624e-05, | |
| "loss": 0.063, | |
| "step": 10950 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.20543019473552704, | |
| "learning_rate": 1.68024e-05, | |
| "loss": 0.0605, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "eval_loss": 0.08301527053117752, | |
| "eval_runtime": 88.0655, | |
| "eval_samples_per_second": 22.71, | |
| "eval_steps_per_second": 5.678, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.884, | |
| "grad_norm": 0.22617511451244354, | |
| "learning_rate": 1.67424e-05, | |
| "loss": 0.066, | |
| "step": 11050 | |
| }, | |
| { | |
| "epoch": 0.888, | |
| "grad_norm": 0.15583598613739014, | |
| "learning_rate": 1.66824e-05, | |
| "loss": 0.0663, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.892, | |
| "grad_norm": 0.06502597779035568, | |
| "learning_rate": 1.6622400000000002e-05, | |
| "loss": 0.0611, | |
| "step": 11150 | |
| }, | |
| { | |
| "epoch": 0.896, | |
| "grad_norm": 0.1777944415807724, | |
| "learning_rate": 1.6562400000000003e-05, | |
| "loss": 0.0618, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.14777645468711853, | |
| "learning_rate": 1.65024e-05, | |
| "loss": 0.0616, | |
| "step": 11250 | |
| }, | |
| { | |
| "epoch": 0.904, | |
| "grad_norm": 0.1988598108291626, | |
| "learning_rate": 1.64424e-05, | |
| "loss": 0.0683, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.908, | |
| "grad_norm": 0.16424456238746643, | |
| "learning_rate": 1.63824e-05, | |
| "loss": 0.0616, | |
| "step": 11350 | |
| }, | |
| { | |
| "epoch": 0.912, | |
| "grad_norm": 0.17275363206863403, | |
| "learning_rate": 1.63224e-05, | |
| "loss": 0.0603, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.916, | |
| "grad_norm": 0.1208202987909317, | |
| "learning_rate": 1.6262400000000002e-05, | |
| "loss": 0.0623, | |
| "step": 11450 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.12853899598121643, | |
| "learning_rate": 1.62024e-05, | |
| "loss": 0.0597, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "eval_loss": 0.08293735980987549, | |
| "eval_runtime": 88.063, | |
| "eval_samples_per_second": 22.711, | |
| "eval_steps_per_second": 5.678, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.924, | |
| "grad_norm": 0.12364531308412552, | |
| "learning_rate": 1.61424e-05, | |
| "loss": 0.0624, | |
| "step": 11550 | |
| }, | |
| { | |
| "epoch": 0.928, | |
| "grad_norm": 0.15913207828998566, | |
| "learning_rate": 1.60824e-05, | |
| "loss": 0.061, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.932, | |
| "grad_norm": 0.16585472226142883, | |
| "learning_rate": 1.60224e-05, | |
| "loss": 0.066, | |
| "step": 11650 | |
| }, | |
| { | |
| "epoch": 0.936, | |
| "grad_norm": 0.13831599056720734, | |
| "learning_rate": 1.59624e-05, | |
| "loss": 0.0578, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.1819075047969818, | |
| "learning_rate": 1.59024e-05, | |
| "loss": 0.0593, | |
| "step": 11750 | |
| }, | |
| { | |
| "epoch": 0.944, | |
| "grad_norm": 0.16260235011577606, | |
| "learning_rate": 1.58424e-05, | |
| "loss": 0.0647, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.948, | |
| "grad_norm": 0.1420051008462906, | |
| "learning_rate": 1.57824e-05, | |
| "loss": 0.0556, | |
| "step": 11850 | |
| }, | |
| { | |
| "epoch": 0.952, | |
| "grad_norm": 0.17357395589351654, | |
| "learning_rate": 1.57224e-05, | |
| "loss": 0.0585, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.956, | |
| "grad_norm": 0.1465808004140854, | |
| "learning_rate": 1.56624e-05, | |
| "loss": 0.0633, | |
| "step": 11950 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.18425939977169037, | |
| "learning_rate": 1.56024e-05, | |
| "loss": 0.061, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "eval_loss": 0.082674041390419, | |
| "eval_runtime": 88.0509, | |
| "eval_samples_per_second": 22.714, | |
| "eval_steps_per_second": 5.679, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.964, | |
| "grad_norm": 0.1296168565750122, | |
| "learning_rate": 1.55424e-05, | |
| "loss": 0.0638, | |
| "step": 12050 | |
| }, | |
| { | |
| "epoch": 0.968, | |
| "grad_norm": 0.14450989663600922, | |
| "learning_rate": 1.54824e-05, | |
| "loss": 0.0623, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.972, | |
| "grad_norm": 0.09367354214191437, | |
| "learning_rate": 1.54224e-05, | |
| "loss": 0.0618, | |
| "step": 12150 | |
| }, | |
| { | |
| "epoch": 0.976, | |
| "grad_norm": 0.13332900404930115, | |
| "learning_rate": 1.53624e-05, | |
| "loss": 0.0608, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.14300012588500977, | |
| "learning_rate": 1.53024e-05, | |
| "loss": 0.0695, | |
| "step": 12250 | |
| }, | |
| { | |
| "epoch": 0.984, | |
| "grad_norm": 0.14829818904399872, | |
| "learning_rate": 1.52424e-05, | |
| "loss": 0.0589, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.988, | |
| "grad_norm": 0.08889272063970566, | |
| "learning_rate": 1.5182399999999999e-05, | |
| "loss": 0.0578, | |
| "step": 12350 | |
| }, | |
| { | |
| "epoch": 0.992, | |
| "grad_norm": 0.18994523584842682, | |
| "learning_rate": 1.51224e-05, | |
| "loss": 0.0667, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.996, | |
| "grad_norm": 0.16152743995189667, | |
| "learning_rate": 1.50624e-05, | |
| "loss": 0.0552, | |
| "step": 12450 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.2399597018957138, | |
| "learning_rate": 1.5002399999999999e-05, | |
| "loss": 0.0632, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.0824647843837738, | |
| "eval_runtime": 88.0514, | |
| "eval_samples_per_second": 22.714, | |
| "eval_steps_per_second": 5.678, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.004, | |
| "grad_norm": 0.12758083641529083, | |
| "learning_rate": 1.4942400000000001e-05, | |
| "loss": 0.0553, | |
| "step": 12550 | |
| }, | |
| { | |
| "epoch": 1.008, | |
| "grad_norm": 0.14093191921710968, | |
| "learning_rate": 1.4882400000000002e-05, | |
| "loss": 0.0572, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 1.012, | |
| "grad_norm": 0.20115911960601807, | |
| "learning_rate": 1.48224e-05, | |
| "loss": 0.055, | |
| "step": 12650 | |
| }, | |
| { | |
| "epoch": 1.016, | |
| "grad_norm": 0.2974820137023926, | |
| "learning_rate": 1.4762400000000001e-05, | |
| "loss": 0.0548, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "grad_norm": 0.06170056387782097, | |
| "learning_rate": 1.47024e-05, | |
| "loss": 0.0528, | |
| "step": 12750 | |
| }, | |
| { | |
| "epoch": 1.024, | |
| "grad_norm": 0.18193961679935455, | |
| "learning_rate": 1.46424e-05, | |
| "loss": 0.0551, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 1.028, | |
| "grad_norm": 0.11086717993021011, | |
| "learning_rate": 1.4582400000000001e-05, | |
| "loss": 0.0533, | |
| "step": 12850 | |
| }, | |
| { | |
| "epoch": 1.032, | |
| "grad_norm": 0.1452319473028183, | |
| "learning_rate": 1.45224e-05, | |
| "loss": 0.0525, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 1.036, | |
| "grad_norm": 0.16397640109062195, | |
| "learning_rate": 1.44624e-05, | |
| "loss": 0.0622, | |
| "step": 12950 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "grad_norm": 0.09709367156028748, | |
| "learning_rate": 1.4402400000000001e-05, | |
| "loss": 0.0559, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "eval_loss": 0.08302706480026245, | |
| "eval_runtime": 88.0122, | |
| "eval_samples_per_second": 22.724, | |
| "eval_steps_per_second": 5.681, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.044, | |
| "grad_norm": 0.09841930866241455, | |
| "learning_rate": 1.43424e-05, | |
| "loss": 0.0534, | |
| "step": 13050 | |
| }, | |
| { | |
| "epoch": 1.048, | |
| "grad_norm": 0.1224198266863823, | |
| "learning_rate": 1.42824e-05, | |
| "loss": 0.0564, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 1.052, | |
| "grad_norm": 0.17220191657543182, | |
| "learning_rate": 1.42224e-05, | |
| "loss": 0.0567, | |
| "step": 13150 | |
| }, | |
| { | |
| "epoch": 1.056, | |
| "grad_norm": 0.08674409985542297, | |
| "learning_rate": 1.41636e-05, | |
| "loss": 0.0508, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "grad_norm": 0.12293367087841034, | |
| "learning_rate": 1.41036e-05, | |
| "loss": 0.0541, | |
| "step": 13250 | |
| }, | |
| { | |
| "epoch": 1.064, | |
| "grad_norm": 0.12341846525669098, | |
| "learning_rate": 1.40436e-05, | |
| "loss": 0.055, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 1.068, | |
| "grad_norm": 0.10165009647607803, | |
| "learning_rate": 1.39836e-05, | |
| "loss": 0.0589, | |
| "step": 13350 | |
| }, | |
| { | |
| "epoch": 1.072, | |
| "grad_norm": 0.18138067424297333, | |
| "learning_rate": 1.39236e-05, | |
| "loss": 0.053, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 1.076, | |
| "grad_norm": 0.18615098297595978, | |
| "learning_rate": 1.3863599999999999e-05, | |
| "loss": 0.0558, | |
| "step": 13450 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "grad_norm": 0.06511889398097992, | |
| "learning_rate": 1.38036e-05, | |
| "loss": 0.062, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "eval_loss": 0.08285626024007797, | |
| "eval_runtime": 88.0463, | |
| "eval_samples_per_second": 22.715, | |
| "eval_steps_per_second": 5.679, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.084, | |
| "grad_norm": 0.1615404486656189, | |
| "learning_rate": 1.3743600000000002e-05, | |
| "loss": 0.0572, | |
| "step": 13550 | |
| }, | |
| { | |
| "epoch": 1.088, | |
| "grad_norm": 0.1500886082649231, | |
| "learning_rate": 1.36836e-05, | |
| "loss": 0.0557, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 1.092, | |
| "grad_norm": 0.08288303017616272, | |
| "learning_rate": 1.3623600000000001e-05, | |
| "loss": 0.0567, | |
| "step": 13650 | |
| }, | |
| { | |
| "epoch": 1.096, | |
| "grad_norm": 0.12978018820285797, | |
| "learning_rate": 1.3563600000000002e-05, | |
| "loss": 0.051, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "grad_norm": 0.17383359372615814, | |
| "learning_rate": 1.35036e-05, | |
| "loss": 0.0527, | |
| "step": 13750 | |
| }, | |
| { | |
| "epoch": 1.104, | |
| "grad_norm": 0.21195685863494873, | |
| "learning_rate": 1.3443600000000001e-05, | |
| "loss": 0.0526, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 1.108, | |
| "grad_norm": 0.15831385552883148, | |
| "learning_rate": 1.33836e-05, | |
| "loss": 0.0595, | |
| "step": 13850 | |
| }, | |
| { | |
| "epoch": 1.112, | |
| "grad_norm": 0.23136693239212036, | |
| "learning_rate": 1.33236e-05, | |
| "loss": 0.0523, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 1.116, | |
| "grad_norm": 0.12520194053649902, | |
| "learning_rate": 1.3263600000000001e-05, | |
| "loss": 0.0603, | |
| "step": 13950 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "grad_norm": 0.10329103469848633, | |
| "learning_rate": 1.32036e-05, | |
| "loss": 0.0494, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "eval_loss": 0.083070769906044, | |
| "eval_runtime": 88.0459, | |
| "eval_samples_per_second": 22.715, | |
| "eval_steps_per_second": 5.679, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.124, | |
| "grad_norm": 0.17533883452415466, | |
| "learning_rate": 1.31436e-05, | |
| "loss": 0.0552, | |
| "step": 14050 | |
| }, | |
| { | |
| "epoch": 1.1280000000000001, | |
| "grad_norm": 0.10817945748567581, | |
| "learning_rate": 1.3083600000000001e-05, | |
| "loss": 0.049, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 1.1320000000000001, | |
| "grad_norm": 0.09337913990020752, | |
| "learning_rate": 1.30236e-05, | |
| "loss": 0.0573, | |
| "step": 14150 | |
| }, | |
| { | |
| "epoch": 1.1360000000000001, | |
| "grad_norm": 0.15710942447185516, | |
| "learning_rate": 1.29636e-05, | |
| "loss": 0.0605, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 1.1400000000000001, | |
| "grad_norm": 0.10915792733430862, | |
| "learning_rate": 1.29036e-05, | |
| "loss": 0.0581, | |
| "step": 14250 | |
| }, | |
| { | |
| "epoch": 1.144, | |
| "grad_norm": 0.10125772655010223, | |
| "learning_rate": 1.28436e-05, | |
| "loss": 0.0599, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 1.148, | |
| "grad_norm": 0.0998956710100174, | |
| "learning_rate": 1.27836e-05, | |
| "loss": 0.0479, | |
| "step": 14350 | |
| }, | |
| { | |
| "epoch": 1.152, | |
| "grad_norm": 0.13762612640857697, | |
| "learning_rate": 1.27236e-05, | |
| "loss": 0.0589, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 1.156, | |
| "grad_norm": 0.11048023402690887, | |
| "learning_rate": 1.26636e-05, | |
| "loss": 0.0591, | |
| "step": 14450 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "grad_norm": 0.14803436398506165, | |
| "learning_rate": 1.26036e-05, | |
| "loss": 0.0553, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "eval_loss": 0.08270228654146194, | |
| "eval_runtime": 88.0514, | |
| "eval_samples_per_second": 22.714, | |
| "eval_steps_per_second": 5.678, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.164, | |
| "grad_norm": 0.06833196431398392, | |
| "learning_rate": 1.25436e-05, | |
| "loss": 0.0586, | |
| "step": 14550 | |
| }, | |
| { | |
| "epoch": 1.168, | |
| "grad_norm": 0.16051365435123444, | |
| "learning_rate": 1.24836e-05, | |
| "loss": 0.056, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 1.172, | |
| "grad_norm": 0.18909616768360138, | |
| "learning_rate": 1.24236e-05, | |
| "loss": 0.057, | |
| "step": 14650 | |
| }, | |
| { | |
| "epoch": 1.176, | |
| "grad_norm": 0.17333486676216125, | |
| "learning_rate": 1.23636e-05, | |
| "loss": 0.0562, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "grad_norm": 0.0701974630355835, | |
| "learning_rate": 1.23036e-05, | |
| "loss": 0.0545, | |
| "step": 14750 | |
| }, | |
| { | |
| "epoch": 1.184, | |
| "grad_norm": 0.1582074612379074, | |
| "learning_rate": 1.2243599999999999e-05, | |
| "loss": 0.06, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 1.188, | |
| "grad_norm": 0.13948781788349152, | |
| "learning_rate": 1.21836e-05, | |
| "loss": 0.0529, | |
| "step": 14850 | |
| }, | |
| { | |
| "epoch": 1.192, | |
| "grad_norm": 0.0946699008345604, | |
| "learning_rate": 1.21236e-05, | |
| "loss": 0.0668, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 1.196, | |
| "grad_norm": 0.1089014783501625, | |
| "learning_rate": 1.20636e-05, | |
| "loss": 0.0562, | |
| "step": 14950 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 0.09682592004537582, | |
| "learning_rate": 1.20036e-05, | |
| "loss": 0.0548, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "eval_loss": 0.08243728429079056, | |
| "eval_runtime": 88.0302, | |
| "eval_samples_per_second": 22.719, | |
| "eval_steps_per_second": 5.68, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.204, | |
| "grad_norm": 0.15492039918899536, | |
| "learning_rate": 1.19436e-05, | |
| "loss": 0.0535, | |
| "step": 15050 | |
| }, | |
| { | |
| "epoch": 1.208, | |
| "grad_norm": 0.0744326263666153, | |
| "learning_rate": 1.1883600000000001e-05, | |
| "loss": 0.0572, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 1.212, | |
| "grad_norm": 0.20596753060817719, | |
| "learning_rate": 1.1823600000000001e-05, | |
| "loss": 0.0539, | |
| "step": 15150 | |
| }, | |
| { | |
| "epoch": 1.216, | |
| "grad_norm": 0.15914830565452576, | |
| "learning_rate": 1.17636e-05, | |
| "loss": 0.0583, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "grad_norm": 0.16685660183429718, | |
| "learning_rate": 1.1703600000000001e-05, | |
| "loss": 0.0566, | |
| "step": 15250 | |
| }, | |
| { | |
| "epoch": 1.224, | |
| "grad_norm": 0.11385693401098251, | |
| "learning_rate": 1.1643600000000001e-05, | |
| "loss": 0.0542, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 1.228, | |
| "grad_norm": 0.12850815057754517, | |
| "learning_rate": 1.15836e-05, | |
| "loss": 0.0622, | |
| "step": 15350 | |
| }, | |
| { | |
| "epoch": 1.232, | |
| "grad_norm": 0.17093044519424438, | |
| "learning_rate": 1.1523600000000001e-05, | |
| "loss": 0.0536, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 1.236, | |
| "grad_norm": 0.18593502044677734, | |
| "learning_rate": 1.1463600000000002e-05, | |
| "loss": 0.0541, | |
| "step": 15450 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "grad_norm": 0.16509443521499634, | |
| "learning_rate": 1.14036e-05, | |
| "loss": 0.0501, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "eval_loss": 0.08269675076007843, | |
| "eval_runtime": 88.0811, | |
| "eval_samples_per_second": 22.706, | |
| "eval_steps_per_second": 5.677, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.244, | |
| "grad_norm": 0.15668867528438568, | |
| "learning_rate": 1.1343600000000001e-05, | |
| "loss": 0.0534, | |
| "step": 15550 | |
| }, | |
| { | |
| "epoch": 1.248, | |
| "grad_norm": 0.08703252673149109, | |
| "learning_rate": 1.12836e-05, | |
| "loss": 0.0534, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 1.252, | |
| "grad_norm": 0.1255914270877838, | |
| "learning_rate": 1.12248e-05, | |
| "loss": 0.0514, | |
| "step": 15650 | |
| }, | |
| { | |
| "epoch": 1.256, | |
| "grad_norm": 0.11858811974525452, | |
| "learning_rate": 1.11648e-05, | |
| "loss": 0.0527, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "grad_norm": 0.13111431896686554, | |
| "learning_rate": 1.1104799999999999e-05, | |
| "loss": 0.0494, | |
| "step": 15750 | |
| }, | |
| { | |
| "epoch": 1.264, | |
| "grad_norm": 0.17452576756477356, | |
| "learning_rate": 1.10448e-05, | |
| "loss": 0.0562, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 1.268, | |
| "grad_norm": 0.13699005544185638, | |
| "learning_rate": 1.09848e-05, | |
| "loss": 0.051, | |
| "step": 15850 | |
| }, | |
| { | |
| "epoch": 1.272, | |
| "grad_norm": 0.18750061094760895, | |
| "learning_rate": 1.0924799999999999e-05, | |
| "loss": 0.0545, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 1.276, | |
| "grad_norm": 0.1259056031703949, | |
| "learning_rate": 1.08648e-05, | |
| "loss": 0.0529, | |
| "step": 15950 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "grad_norm": 0.10538329184055328, | |
| "learning_rate": 1.08048e-05, | |
| "loss": 0.058, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "eval_loss": 0.08262528479099274, | |
| "eval_runtime": 88.074, | |
| "eval_samples_per_second": 22.708, | |
| "eval_steps_per_second": 5.677, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.284, | |
| "grad_norm": 0.14214898645877838, | |
| "learning_rate": 1.07448e-05, | |
| "loss": 0.056, | |
| "step": 16050 | |
| }, | |
| { | |
| "epoch": 1.288, | |
| "grad_norm": 0.13236215710639954, | |
| "learning_rate": 1.06848e-05, | |
| "loss": 0.0566, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 1.292, | |
| "grad_norm": 0.1475580334663391, | |
| "learning_rate": 1.06248e-05, | |
| "loss": 0.0507, | |
| "step": 16150 | |
| }, | |
| { | |
| "epoch": 1.296, | |
| "grad_norm": 0.13150113821029663, | |
| "learning_rate": 1.05648e-05, | |
| "loss": 0.0488, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "grad_norm": 0.0869784876704216, | |
| "learning_rate": 1.0504800000000001e-05, | |
| "loss": 0.0498, | |
| "step": 16250 | |
| }, | |
| { | |
| "epoch": 1.304, | |
| "grad_norm": 0.15695451200008392, | |
| "learning_rate": 1.04448e-05, | |
| "loss": 0.0517, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 1.308, | |
| "grad_norm": 0.1383635401725769, | |
| "learning_rate": 1.0384800000000001e-05, | |
| "loss": 0.0509, | |
| "step": 16350 | |
| }, | |
| { | |
| "epoch": 1.312, | |
| "grad_norm": 0.17300955951213837, | |
| "learning_rate": 1.0324800000000002e-05, | |
| "loss": 0.0587, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 1.316, | |
| "grad_norm": 0.1608356237411499, | |
| "learning_rate": 1.02648e-05, | |
| "loss": 0.0523, | |
| "step": 16450 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "grad_norm": 0.1598045974969864, | |
| "learning_rate": 1.0204800000000001e-05, | |
| "loss": 0.0524, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "eval_loss": 0.08258219808340073, | |
| "eval_runtime": 88.0858, | |
| "eval_samples_per_second": 22.705, | |
| "eval_steps_per_second": 5.676, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.324, | |
| "grad_norm": 0.10573418438434601, | |
| "learning_rate": 1.0144800000000002e-05, | |
| "loss": 0.058, | |
| "step": 16550 | |
| }, | |
| { | |
| "epoch": 1.328, | |
| "grad_norm": 0.1370289921760559, | |
| "learning_rate": 1.00848e-05, | |
| "loss": 0.0552, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 1.332, | |
| "grad_norm": 0.15583133697509766, | |
| "learning_rate": 1.0024800000000001e-05, | |
| "loss": 0.0505, | |
| "step": 16650 | |
| }, | |
| { | |
| "epoch": 1.336, | |
| "grad_norm": 0.12616191804409027, | |
| "learning_rate": 9.9648e-06, | |
| "loss": 0.0563, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "grad_norm": 0.1533484160900116, | |
| "learning_rate": 9.9048e-06, | |
| "loss": 0.0564, | |
| "step": 16750 | |
| }, | |
| { | |
| "epoch": 1.3439999999999999, | |
| "grad_norm": 0.11609365046024323, | |
| "learning_rate": 9.844800000000001e-06, | |
| "loss": 0.0538, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 1.3479999999999999, | |
| "grad_norm": 0.12423556298017502, | |
| "learning_rate": 9.7848e-06, | |
| "loss": 0.0519, | |
| "step": 16850 | |
| }, | |
| { | |
| "epoch": 1.3519999999999999, | |
| "grad_norm": 0.13122333586215973, | |
| "learning_rate": 9.7248e-06, | |
| "loss": 0.0518, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 1.3559999999999999, | |
| "grad_norm": 0.17409804463386536, | |
| "learning_rate": 9.664800000000001e-06, | |
| "loss": 0.0544, | |
| "step": 16950 | |
| }, | |
| { | |
| "epoch": 1.3599999999999999, | |
| "grad_norm": 0.13961808383464813, | |
| "learning_rate": 9.6048e-06, | |
| "loss": 0.0523, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.3599999999999999, | |
| "eval_loss": 0.08219126611948013, | |
| "eval_runtime": 88.051, | |
| "eval_samples_per_second": 22.714, | |
| "eval_steps_per_second": 5.679, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.3639999999999999, | |
| "grad_norm": 0.19439847767353058, | |
| "learning_rate": 9.5448e-06, | |
| "loss": 0.0566, | |
| "step": 17050 | |
| }, | |
| { | |
| "epoch": 1.3679999999999999, | |
| "grad_norm": 0.1604561060667038, | |
| "learning_rate": 9.4848e-06, | |
| "loss": 0.0532, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 1.3719999999999999, | |
| "grad_norm": 0.13188685476779938, | |
| "learning_rate": 9.4248e-06, | |
| "loss": 0.0547, | |
| "step": 17150 | |
| }, | |
| { | |
| "epoch": 1.376, | |
| "grad_norm": 0.09787939488887787, | |
| "learning_rate": 9.3648e-06, | |
| "loss": 0.0476, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "grad_norm": 0.3960016369819641, | |
| "learning_rate": 9.3048e-06, | |
| "loss": 0.0607, | |
| "step": 17250 | |
| }, | |
| { | |
| "epoch": 1.384, | |
| "grad_norm": 0.17494530975818634, | |
| "learning_rate": 9.2448e-06, | |
| "loss": 0.0579, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 1.388, | |
| "grad_norm": 0.17870362102985382, | |
| "learning_rate": 9.1848e-06, | |
| "loss": 0.0481, | |
| "step": 17350 | |
| }, | |
| { | |
| "epoch": 1.392, | |
| "grad_norm": 0.1152658611536026, | |
| "learning_rate": 9.1248e-06, | |
| "loss": 0.0507, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 1.396, | |
| "grad_norm": 0.10560191422700882, | |
| "learning_rate": 9.0648e-06, | |
| "loss": 0.0509, | |
| "step": 17450 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "grad_norm": 0.09202779084444046, | |
| "learning_rate": 9.004799999999999e-06, | |
| "loss": 0.0571, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "eval_loss": 0.0821109265089035, | |
| "eval_runtime": 88.0678, | |
| "eval_samples_per_second": 22.71, | |
| "eval_steps_per_second": 5.677, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.404, | |
| "grad_norm": 0.1634778082370758, | |
| "learning_rate": 8.9448e-06, | |
| "loss": 0.0532, | |
| "step": 17550 | |
| }, | |
| { | |
| "epoch": 1.408, | |
| "grad_norm": 0.12161219865083694, | |
| "learning_rate": 8.8848e-06, | |
| "loss": 0.0541, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 1.412, | |
| "grad_norm": 0.15347328782081604, | |
| "learning_rate": 8.824799999999999e-06, | |
| "loss": 0.0535, | |
| "step": 17650 | |
| }, | |
| { | |
| "epoch": 1.416, | |
| "grad_norm": 0.07917541265487671, | |
| "learning_rate": 8.7648e-06, | |
| "loss": 0.0552, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "grad_norm": 0.15406325459480286, | |
| "learning_rate": 8.7048e-06, | |
| "loss": 0.0584, | |
| "step": 17750 | |
| }, | |
| { | |
| "epoch": 1.424, | |
| "grad_norm": 0.21300417184829712, | |
| "learning_rate": 8.6448e-06, | |
| "loss": 0.0565, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 1.428, | |
| "grad_norm": 0.20691223442554474, | |
| "learning_rate": 8.584800000000001e-06, | |
| "loss": 0.0504, | |
| "step": 17850 | |
| }, | |
| { | |
| "epoch": 1.432, | |
| "grad_norm": 0.1366143375635147, | |
| "learning_rate": 8.5248e-06, | |
| "loss": 0.0573, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 1.436, | |
| "grad_norm": 0.11994505673646927, | |
| "learning_rate": 8.4648e-06, | |
| "loss": 0.0576, | |
| "step": 17950 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "grad_norm": 0.06461376696825027, | |
| "learning_rate": 8.404800000000001e-06, | |
| "loss": 0.0595, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "eval_loss": 0.08217783272266388, | |
| "eval_runtime": 88.0706, | |
| "eval_samples_per_second": 22.709, | |
| "eval_steps_per_second": 5.677, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.444, | |
| "grad_norm": 0.150528684258461, | |
| "learning_rate": 8.3448e-06, | |
| "loss": 0.0519, | |
| "step": 18050 | |
| }, | |
| { | |
| "epoch": 1.448, | |
| "grad_norm": 0.08878117799758911, | |
| "learning_rate": 8.2848e-06, | |
| "loss": 0.0544, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 1.452, | |
| "grad_norm": 0.12337563931941986, | |
| "learning_rate": 8.224800000000001e-06, | |
| "loss": 0.0509, | |
| "step": 18150 | |
| }, | |
| { | |
| "epoch": 1.456, | |
| "grad_norm": 0.12858565151691437, | |
| "learning_rate": 8.1648e-06, | |
| "loss": 0.0561, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "grad_norm": 0.14318782091140747, | |
| "learning_rate": 8.1048e-06, | |
| "loss": 0.0539, | |
| "step": 18250 | |
| }, | |
| { | |
| "epoch": 1.464, | |
| "grad_norm": 0.16946159303188324, | |
| "learning_rate": 8.044800000000001e-06, | |
| "loss": 0.0545, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 1.468, | |
| "grad_norm": 0.11605024337768555, | |
| "learning_rate": 7.9848e-06, | |
| "loss": 0.0563, | |
| "step": 18350 | |
| }, | |
| { | |
| "epoch": 1.472, | |
| "grad_norm": 0.20185694098472595, | |
| "learning_rate": 7.9248e-06, | |
| "loss": 0.0536, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 1.476, | |
| "grad_norm": 0.16189858317375183, | |
| "learning_rate": 7.8648e-06, | |
| "loss": 0.0523, | |
| "step": 18450 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "grad_norm": 0.24366913735866547, | |
| "learning_rate": 7.8048e-06, | |
| "loss": 0.057, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "eval_loss": 0.08215340971946716, | |
| "eval_runtime": 88.0726, | |
| "eval_samples_per_second": 22.709, | |
| "eval_steps_per_second": 5.677, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.484, | |
| "grad_norm": 0.13516181707382202, | |
| "learning_rate": 7.7448e-06, | |
| "loss": 0.0516, | |
| "step": 18550 | |
| }, | |
| { | |
| "epoch": 1.488, | |
| "grad_norm": 0.14996372163295746, | |
| "learning_rate": 7.6848e-06, | |
| "loss": 0.0545, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 1.492, | |
| "grad_norm": 0.11676355451345444, | |
| "learning_rate": 7.6248e-06, | |
| "loss": 0.0546, | |
| "step": 18650 | |
| }, | |
| { | |
| "epoch": 1.496, | |
| "grad_norm": 0.0986240953207016, | |
| "learning_rate": 7.5648e-06, | |
| "loss": 0.054, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 0.16591283679008484, | |
| "learning_rate": 7.5048e-06, | |
| "loss": 0.0554, | |
| "step": 18750 | |
| }, | |
| { | |
| "epoch": 1.504, | |
| "grad_norm": 0.13870297372341156, | |
| "learning_rate": 7.4448e-06, | |
| "loss": 0.0486, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 1.508, | |
| "grad_norm": 0.1334107369184494, | |
| "learning_rate": 7.3848e-06, | |
| "loss": 0.0532, | |
| "step": 18850 | |
| }, | |
| { | |
| "epoch": 1.512, | |
| "grad_norm": 0.1280679702758789, | |
| "learning_rate": 7.3248e-06, | |
| "loss": 0.0543, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 1.516, | |
| "grad_norm": 0.08572965115308762, | |
| "learning_rate": 7.2647999999999995e-06, | |
| "loss": 0.0516, | |
| "step": 18950 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "grad_norm": 0.10845118761062622, | |
| "learning_rate": 7.204800000000001e-06, | |
| "loss": 0.0554, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "eval_loss": 0.08186879754066467, | |
| "eval_runtime": 88.0768, | |
| "eval_samples_per_second": 22.707, | |
| "eval_steps_per_second": 5.677, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.524, | |
| "grad_norm": 0.09976433962583542, | |
| "learning_rate": 7.144800000000001e-06, | |
| "loss": 0.0496, | |
| "step": 19050 | |
| }, | |
| { | |
| "epoch": 1.528, | |
| "grad_norm": 0.19336478412151337, | |
| "learning_rate": 7.0848e-06, | |
| "loss": 0.0528, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 1.532, | |
| "grad_norm": 0.13555702567100525, | |
| "learning_rate": 7.0248e-06, | |
| "loss": 0.0559, | |
| "step": 19150 | |
| }, | |
| { | |
| "epoch": 1.536, | |
| "grad_norm": 0.2016674280166626, | |
| "learning_rate": 6.964800000000001e-06, | |
| "loss": 0.0568, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "grad_norm": 0.8713797330856323, | |
| "learning_rate": 6.9048e-06, | |
| "loss": 0.0522, | |
| "step": 19250 | |
| }, | |
| { | |
| "epoch": 1.544, | |
| "grad_norm": 0.13156233727931976, | |
| "learning_rate": 6.8448e-06, | |
| "loss": 0.0557, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 1.548, | |
| "grad_norm": 0.1713368147611618, | |
| "learning_rate": 6.7848e-06, | |
| "loss": 0.0574, | |
| "step": 19350 | |
| }, | |
| { | |
| "epoch": 1.552, | |
| "grad_norm": 0.13423492014408112, | |
| "learning_rate": 6.7248e-06, | |
| "loss": 0.0494, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 1.556, | |
| "grad_norm": 0.1513233631849289, | |
| "learning_rate": 6.6648e-06, | |
| "loss": 0.0528, | |
| "step": 19450 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "grad_norm": 0.1609751433134079, | |
| "learning_rate": 6.606000000000001e-06, | |
| "loss": 0.0512, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "eval_loss": 0.08218736946582794, | |
| "eval_runtime": 88.094, | |
| "eval_samples_per_second": 22.703, | |
| "eval_steps_per_second": 5.676, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.564, | |
| "grad_norm": 0.22682276368141174, | |
| "learning_rate": 6.5472e-06, | |
| "loss": 0.0603, | |
| "step": 19550 | |
| }, | |
| { | |
| "epoch": 1.568, | |
| "grad_norm": 0.13181114196777344, | |
| "learning_rate": 6.4871999999999995e-06, | |
| "loss": 0.0497, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 1.572, | |
| "grad_norm": 0.1349440962076187, | |
| "learning_rate": 6.427200000000001e-06, | |
| "loss": 0.0507, | |
| "step": 19650 | |
| }, | |
| { | |
| "epoch": 1.576, | |
| "grad_norm": 0.1361471712589264, | |
| "learning_rate": 6.367200000000001e-06, | |
| "loss": 0.0501, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "grad_norm": 0.25193363428115845, | |
| "learning_rate": 6.3072e-06, | |
| "loss": 0.0565, | |
| "step": 19750 | |
| }, | |
| { | |
| "epoch": 1.584, | |
| "grad_norm": 0.12969471514225006, | |
| "learning_rate": 6.2472e-06, | |
| "loss": 0.0616, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 1.588, | |
| "grad_norm": 0.18332916498184204, | |
| "learning_rate": 6.187200000000001e-06, | |
| "loss": 0.0513, | |
| "step": 19850 | |
| }, | |
| { | |
| "epoch": 1.592, | |
| "grad_norm": 0.19354714453220367, | |
| "learning_rate": 6.1272e-06, | |
| "loss": 0.0612, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 1.596, | |
| "grad_norm": 0.17420926690101624, | |
| "learning_rate": 6.0672e-06, | |
| "loss": 0.0604, | |
| "step": 19950 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 0.09988817572593689, | |
| "learning_rate": 6.0072e-06, | |
| "loss": 0.0523, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "eval_loss": 0.08189179003238678, | |
| "eval_runtime": 88.0418, | |
| "eval_samples_per_second": 22.716, | |
| "eval_steps_per_second": 5.679, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.604, | |
| "grad_norm": 0.18323849141597748, | |
| "learning_rate": 5.9472e-06, | |
| "loss": 0.0555, | |
| "step": 20050 | |
| }, | |
| { | |
| "epoch": 1.608, | |
| "grad_norm": 0.10411707311868668, | |
| "learning_rate": 5.8872e-06, | |
| "loss": 0.0506, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 1.612, | |
| "grad_norm": 0.18998867273330688, | |
| "learning_rate": 5.8272e-06, | |
| "loss": 0.0522, | |
| "step": 20150 | |
| }, | |
| { | |
| "epoch": 1.616, | |
| "grad_norm": 0.159571573138237, | |
| "learning_rate": 5.7672e-06, | |
| "loss": 0.0527, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "grad_norm": 0.10246822983026505, | |
| "learning_rate": 5.7072e-06, | |
| "loss": 0.0517, | |
| "step": 20250 | |
| }, | |
| { | |
| "epoch": 1.624, | |
| "grad_norm": 0.17232735455036163, | |
| "learning_rate": 5.6472e-06, | |
| "loss": 0.0559, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 1.6280000000000001, | |
| "grad_norm": 0.13078981637954712, | |
| "learning_rate": 5.5872e-06, | |
| "loss": 0.0497, | |
| "step": 20350 | |
| }, | |
| { | |
| "epoch": 1.6320000000000001, | |
| "grad_norm": 0.06083545461297035, | |
| "learning_rate": 5.527199999999999e-06, | |
| "loss": 0.0502, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 1.6360000000000001, | |
| "grad_norm": 0.1311408132314682, | |
| "learning_rate": 5.467200000000001e-06, | |
| "loss": 0.0572, | |
| "step": 20450 | |
| }, | |
| { | |
| "epoch": 1.6400000000000001, | |
| "grad_norm": 0.16154557466506958, | |
| "learning_rate": 5.4072000000000005e-06, | |
| "loss": 0.0545, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 1.6400000000000001, | |
| "eval_loss": 0.08202797174453735, | |
| "eval_runtime": 87.9943, | |
| "eval_samples_per_second": 22.729, | |
| "eval_steps_per_second": 5.682, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 1.6440000000000001, | |
| "grad_norm": 0.13027295470237732, | |
| "learning_rate": 5.3472e-06, | |
| "loss": 0.0551, | |
| "step": 20550 | |
| }, | |
| { | |
| "epoch": 1.6480000000000001, | |
| "grad_norm": 0.1394919753074646, | |
| "learning_rate": 5.2872e-06, | |
| "loss": 0.054, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 1.6520000000000001, | |
| "grad_norm": 0.16753709316253662, | |
| "learning_rate": 5.2272000000000005e-06, | |
| "loss": 0.0501, | |
| "step": 20650 | |
| }, | |
| { | |
| "epoch": 1.6560000000000001, | |
| "grad_norm": 0.1509876549243927, | |
| "learning_rate": 5.1672e-06, | |
| "loss": 0.0527, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 1.6600000000000001, | |
| "grad_norm": 0.13625292479991913, | |
| "learning_rate": 5.1072e-06, | |
| "loss": 0.0508, | |
| "step": 20750 | |
| }, | |
| { | |
| "epoch": 1.6640000000000001, | |
| "grad_norm": 0.1552583873271942, | |
| "learning_rate": 5.0472000000000006e-06, | |
| "loss": 0.0548, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 1.6680000000000001, | |
| "grad_norm": 0.1763962060213089, | |
| "learning_rate": 4.9872e-06, | |
| "loss": 0.0585, | |
| "step": 20850 | |
| }, | |
| { | |
| "epoch": 1.6720000000000002, | |
| "grad_norm": 0.11216771602630615, | |
| "learning_rate": 4.9272e-06, | |
| "loss": 0.0567, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 1.6760000000000002, | |
| "grad_norm": 0.08550629019737244, | |
| "learning_rate": 4.8672e-06, | |
| "loss": 0.0523, | |
| "step": 20950 | |
| }, | |
| { | |
| "epoch": 1.6800000000000002, | |
| "grad_norm": 0.11488083750009537, | |
| "learning_rate": 4.8072e-06, | |
| "loss": 0.0503, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 1.6800000000000002, | |
| "eval_loss": 0.0817028358578682, | |
| "eval_runtime": 88.097, | |
| "eval_samples_per_second": 22.702, | |
| "eval_steps_per_second": 5.676, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 1.6840000000000002, | |
| "grad_norm": 0.1464158296585083, | |
| "learning_rate": 4.7472e-06, | |
| "loss": 0.0557, | |
| "step": 21050 | |
| }, | |
| { | |
| "epoch": 1.688, | |
| "grad_norm": 0.1558839976787567, | |
| "learning_rate": 4.6872e-06, | |
| "loss": 0.0527, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 1.692, | |
| "grad_norm": 0.093449167907238, | |
| "learning_rate": 4.6271999999999995e-06, | |
| "loss": 0.0544, | |
| "step": 21150 | |
| }, | |
| { | |
| "epoch": 1.696, | |
| "grad_norm": 0.1413930058479309, | |
| "learning_rate": 4.5672e-06, | |
| "loss": 0.0571, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "grad_norm": 0.19774900376796722, | |
| "learning_rate": 4.507200000000001e-06, | |
| "loss": 0.0474, | |
| "step": 21250 | |
| }, | |
| { | |
| "epoch": 1.704, | |
| "grad_norm": 0.092808298766613, | |
| "learning_rate": 4.4472e-06, | |
| "loss": 0.05, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 1.708, | |
| "grad_norm": 0.09530337899923325, | |
| "learning_rate": 4.3872e-06, | |
| "loss": 0.0564, | |
| "step": 21350 | |
| }, | |
| { | |
| "epoch": 1.712, | |
| "grad_norm": 0.1850968301296234, | |
| "learning_rate": 4.327200000000001e-06, | |
| "loss": 0.0541, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 1.716, | |
| "grad_norm": 0.23416727781295776, | |
| "learning_rate": 4.2672e-06, | |
| "loss": 0.0576, | |
| "step": 21450 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "grad_norm": 0.11343374848365784, | |
| "learning_rate": 4.2072e-06, | |
| "loss": 0.0529, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "eval_loss": 0.08182183653116226, | |
| "eval_runtime": 88.0562, | |
| "eval_samples_per_second": 22.713, | |
| "eval_steps_per_second": 5.678, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 1.724, | |
| "grad_norm": 0.14097870886325836, | |
| "learning_rate": 4.1472e-06, | |
| "loss": 0.0488, | |
| "step": 21550 | |
| }, | |
| { | |
| "epoch": 1.728, | |
| "grad_norm": 0.1706978976726532, | |
| "learning_rate": 4.0872000000000004e-06, | |
| "loss": 0.0496, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 1.732, | |
| "grad_norm": 0.14371682703495026, | |
| "learning_rate": 4.0272e-06, | |
| "loss": 0.0557, | |
| "step": 21650 | |
| }, | |
| { | |
| "epoch": 1.736, | |
| "grad_norm": 0.1176629364490509, | |
| "learning_rate": 3.9672e-06, | |
| "loss": 0.0531, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "grad_norm": 0.17272049188613892, | |
| "learning_rate": 3.9072e-06, | |
| "loss": 0.0604, | |
| "step": 21750 | |
| }, | |
| { | |
| "epoch": 1.744, | |
| "grad_norm": 0.15085135400295258, | |
| "learning_rate": 3.8472e-06, | |
| "loss": 0.0537, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 1.748, | |
| "grad_norm": 0.11613863706588745, | |
| "learning_rate": 3.7884e-06, | |
| "loss": 0.0544, | |
| "step": 21850 | |
| }, | |
| { | |
| "epoch": 1.752, | |
| "grad_norm": 0.13247713446617126, | |
| "learning_rate": 3.7284e-06, | |
| "loss": 0.0556, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 1.756, | |
| "grad_norm": 0.1755180060863495, | |
| "learning_rate": 3.6684e-06, | |
| "loss": 0.0563, | |
| "step": 21950 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "grad_norm": 0.07918363809585571, | |
| "learning_rate": 3.6084e-06, | |
| "loss": 0.0508, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "eval_loss": 0.08166228979825974, | |
| "eval_runtime": 88.0895, | |
| "eval_samples_per_second": 22.704, | |
| "eval_steps_per_second": 5.676, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 1.764, | |
| "grad_norm": 0.18784619867801666, | |
| "learning_rate": 3.5483999999999998e-06, | |
| "loss": 0.0559, | |
| "step": 22050 | |
| }, | |
| { | |
| "epoch": 1.768, | |
| "grad_norm": 0.1378210335969925, | |
| "learning_rate": 3.4884e-06, | |
| "loss": 0.0528, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 1.772, | |
| "grad_norm": 0.1482810080051422, | |
| "learning_rate": 3.4284e-06, | |
| "loss": 0.0535, | |
| "step": 22150 | |
| }, | |
| { | |
| "epoch": 1.776, | |
| "grad_norm": 0.2096211314201355, | |
| "learning_rate": 3.3684000000000002e-06, | |
| "loss": 0.0455, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "grad_norm": 0.16895908117294312, | |
| "learning_rate": 3.3084000000000004e-06, | |
| "loss": 0.0527, | |
| "step": 22250 | |
| }, | |
| { | |
| "epoch": 1.784, | |
| "grad_norm": 0.11279798299074173, | |
| "learning_rate": 3.2484e-06, | |
| "loss": 0.0524, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 1.788, | |
| "grad_norm": 0.1444096565246582, | |
| "learning_rate": 3.1884000000000003e-06, | |
| "loss": 0.0539, | |
| "step": 22350 | |
| }, | |
| { | |
| "epoch": 1.792, | |
| "grad_norm": 0.16793648898601532, | |
| "learning_rate": 3.1284e-06, | |
| "loss": 0.0578, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 1.796, | |
| "grad_norm": 0.1865730732679367, | |
| "learning_rate": 3.0684e-06, | |
| "loss": 0.0538, | |
| "step": 22450 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 0.11493191868066788, | |
| "learning_rate": 3.0084e-06, | |
| "loss": 0.0515, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "eval_loss": 0.0815029963850975, | |
| "eval_runtime": 88.0709, | |
| "eval_samples_per_second": 22.709, | |
| "eval_steps_per_second": 5.677, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 1.804, | |
| "grad_norm": 0.10228476673364639, | |
| "learning_rate": 2.9496e-06, | |
| "loss": 0.0534, | |
| "step": 22550 | |
| }, | |
| { | |
| "epoch": 1.808, | |
| "grad_norm": 0.12501686811447144, | |
| "learning_rate": 2.8896000000000003e-06, | |
| "loss": 0.0529, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 1.812, | |
| "grad_norm": 0.10296665877103806, | |
| "learning_rate": 2.8296e-06, | |
| "loss": 0.0575, | |
| "step": 22650 | |
| }, | |
| { | |
| "epoch": 1.8159999999999998, | |
| "grad_norm": 0.15849712491035461, | |
| "learning_rate": 2.7696e-06, | |
| "loss": 0.054, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 1.8199999999999998, | |
| "grad_norm": 0.0910249873995781, | |
| "learning_rate": 2.7096e-06, | |
| "loss": 0.0533, | |
| "step": 22750 | |
| }, | |
| { | |
| "epoch": 1.8239999999999998, | |
| "grad_norm": 0.14466793835163116, | |
| "learning_rate": 2.6496e-06, | |
| "loss": 0.0512, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 1.8279999999999998, | |
| "grad_norm": 0.21741582453250885, | |
| "learning_rate": 2.5895999999999997e-06, | |
| "loss": 0.0574, | |
| "step": 22850 | |
| }, | |
| { | |
| "epoch": 1.8319999999999999, | |
| "grad_norm": 0.14835171401500702, | |
| "learning_rate": 2.5296000000000003e-06, | |
| "loss": 0.0569, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 1.8359999999999999, | |
| "grad_norm": 0.13074947893619537, | |
| "learning_rate": 2.4696e-06, | |
| "loss": 0.0555, | |
| "step": 22950 | |
| }, | |
| { | |
| "epoch": 1.8399999999999999, | |
| "grad_norm": 0.14285966753959656, | |
| "learning_rate": 2.4096e-06, | |
| "loss": 0.0531, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.8399999999999999, | |
| "eval_loss": 0.08154193311929703, | |
| "eval_runtime": 88.0742, | |
| "eval_samples_per_second": 22.708, | |
| "eval_steps_per_second": 5.677, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.8439999999999999, | |
| "grad_norm": 0.2026013433933258, | |
| "learning_rate": 2.3496e-06, | |
| "loss": 0.0526, | |
| "step": 23050 | |
| }, | |
| { | |
| "epoch": 1.8479999999999999, | |
| "grad_norm": 0.1372475028038025, | |
| "learning_rate": 2.2896e-06, | |
| "loss": 0.057, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 1.8519999999999999, | |
| "grad_norm": 0.13629719614982605, | |
| "learning_rate": 2.2296e-06, | |
| "loss": 0.0576, | |
| "step": 23150 | |
| }, | |
| { | |
| "epoch": 1.8559999999999999, | |
| "grad_norm": 0.2135656774044037, | |
| "learning_rate": 2.1696e-06, | |
| "loss": 0.0571, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 1.8599999999999999, | |
| "grad_norm": 0.1738504022359848, | |
| "learning_rate": 2.1095999999999997e-06, | |
| "loss": 0.054, | |
| "step": 23250 | |
| }, | |
| { | |
| "epoch": 1.8639999999999999, | |
| "grad_norm": 0.17974117398262024, | |
| "learning_rate": 2.0496000000000002e-06, | |
| "loss": 0.0551, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 1.8679999999999999, | |
| "grad_norm": 0.16323037445545197, | |
| "learning_rate": 1.9896e-06, | |
| "loss": 0.0515, | |
| "step": 23350 | |
| }, | |
| { | |
| "epoch": 1.8719999999999999, | |
| "grad_norm": 0.11057537794113159, | |
| "learning_rate": 1.9296e-06, | |
| "loss": 0.0564, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 1.876, | |
| "grad_norm": 0.1545117348432541, | |
| "learning_rate": 1.8696e-06, | |
| "loss": 0.0508, | |
| "step": 23450 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "grad_norm": 0.25704920291900635, | |
| "learning_rate": 1.8096e-06, | |
| "loss": 0.0521, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "eval_loss": 0.08151672035455704, | |
| "eval_runtime": 88.0934, | |
| "eval_samples_per_second": 22.703, | |
| "eval_steps_per_second": 5.676, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 1.884, | |
| "grad_norm": 0.10899360477924347, | |
| "learning_rate": 1.7496e-06, | |
| "loss": 0.0558, | |
| "step": 23550 | |
| }, | |
| { | |
| "epoch": 1.888, | |
| "grad_norm": 0.09025511890649796, | |
| "learning_rate": 1.6896e-06, | |
| "loss": 0.0582, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 1.892, | |
| "grad_norm": 0.11647246032953262, | |
| "learning_rate": 1.6296e-06, | |
| "loss": 0.0518, | |
| "step": 23650 | |
| }, | |
| { | |
| "epoch": 1.896, | |
| "grad_norm": 0.15982107818126678, | |
| "learning_rate": 1.5696e-06, | |
| "loss": 0.0535, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "grad_norm": 0.15808264911174774, | |
| "learning_rate": 1.5096e-06, | |
| "loss": 0.0557, | |
| "step": 23750 | |
| }, | |
| { | |
| "epoch": 1.904, | |
| "grad_norm": 0.12610608339309692, | |
| "learning_rate": 1.4496e-06, | |
| "loss": 0.0547, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 1.908, | |
| "grad_norm": 0.10674013942480087, | |
| "learning_rate": 1.3896e-06, | |
| "loss": 0.0506, | |
| "step": 23850 | |
| }, | |
| { | |
| "epoch": 1.912, | |
| "grad_norm": 0.15550707280635834, | |
| "learning_rate": 1.3296e-06, | |
| "loss": 0.0576, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 1.916, | |
| "grad_norm": 0.12588605284690857, | |
| "learning_rate": 1.2696e-06, | |
| "loss": 0.0566, | |
| "step": 23950 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "grad_norm": 0.1209346354007721, | |
| "learning_rate": 1.2096e-06, | |
| "loss": 0.0564, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "eval_loss": 0.08148019760847092, | |
| "eval_runtime": 88.0412, | |
| "eval_samples_per_second": 22.717, | |
| "eval_steps_per_second": 5.679, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 1.924, | |
| "grad_norm": 0.18958355486392975, | |
| "learning_rate": 1.1496e-06, | |
| "loss": 0.054, | |
| "step": 24050 | |
| }, | |
| { | |
| "epoch": 1.928, | |
| "grad_norm": 0.12351219356060028, | |
| "learning_rate": 1.0896e-06, | |
| "loss": 0.0614, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 1.932, | |
| "grad_norm": 0.175857812166214, | |
| "learning_rate": 1.0296e-06, | |
| "loss": 0.0541, | |
| "step": 24150 | |
| }, | |
| { | |
| "epoch": 1.936, | |
| "grad_norm": 0.15968987345695496, | |
| "learning_rate": 9.696e-07, | |
| "loss": 0.0568, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "grad_norm": 0.15781116485595703, | |
| "learning_rate": 9.096e-07, | |
| "loss": 0.055, | |
| "step": 24250 | |
| }, | |
| { | |
| "epoch": 1.944, | |
| "grad_norm": 0.17276370525360107, | |
| "learning_rate": 8.496000000000001e-07, | |
| "loss": 0.0475, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 1.948, | |
| "grad_norm": 0.09511862695217133, | |
| "learning_rate": 7.896e-07, | |
| "loss": 0.0509, | |
| "step": 24350 | |
| }, | |
| { | |
| "epoch": 1.952, | |
| "grad_norm": 0.1636885106563568, | |
| "learning_rate": 7.296000000000001e-07, | |
| "loss": 0.0508, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 1.956, | |
| "grad_norm": 0.12255977094173431, | |
| "learning_rate": 6.696e-07, | |
| "loss": 0.0467, | |
| "step": 24450 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "grad_norm": 0.18492284417152405, | |
| "learning_rate": 6.096000000000001e-07, | |
| "loss": 0.0559, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "eval_loss": 0.08147666603326797, | |
| "eval_runtime": 88.0584, | |
| "eval_samples_per_second": 22.712, | |
| "eval_steps_per_second": 5.678, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 1.964, | |
| "grad_norm": 0.1472010761499405, | |
| "learning_rate": 5.496e-07, | |
| "loss": 0.0485, | |
| "step": 24550 | |
| }, | |
| { | |
| "epoch": 1.968, | |
| "grad_norm": 0.14501795172691345, | |
| "learning_rate": 4.896e-07, | |
| "loss": 0.0554, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 1.972, | |
| "grad_norm": 0.09672766178846359, | |
| "learning_rate": 4.296e-07, | |
| "loss": 0.0534, | |
| "step": 24650 | |
| }, | |
| { | |
| "epoch": 1.976, | |
| "grad_norm": 0.1332007199525833, | |
| "learning_rate": 3.696e-07, | |
| "loss": 0.0513, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "grad_norm": 0.0856068879365921, | |
| "learning_rate": 3.0959999999999997e-07, | |
| "loss": 0.0531, | |
| "step": 24750 | |
| }, | |
| { | |
| "epoch": 1.984, | |
| "grad_norm": 0.07875853031873703, | |
| "learning_rate": 2.4959999999999996e-07, | |
| "loss": 0.0511, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 1.988, | |
| "grad_norm": 0.1319677084684372, | |
| "learning_rate": 1.896e-07, | |
| "loss": 0.0493, | |
| "step": 24850 | |
| }, | |
| { | |
| "epoch": 1.992, | |
| "grad_norm": 0.20934289693832397, | |
| "learning_rate": 1.296e-07, | |
| "loss": 0.0559, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 1.996, | |
| "grad_norm": 0.12190008908510208, | |
| "learning_rate": 6.96e-08, | |
| "loss": 0.0589, | |
| "step": 24950 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.2332906574010849, | |
| "learning_rate": 9.600000000000002e-09, | |
| "loss": 0.0557, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.08146847784519196, | |
| "eval_runtime": 88.0931, | |
| "eval_samples_per_second": 22.703, | |
| "eval_steps_per_second": 5.676, | |
| "step": 25000 | |
| } | |
| ], | |
| "logging_steps": 50, | |
| "max_steps": 25000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 6.0895789056e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |