| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 64.0, | |
| "eval_steps": 500, | |
| "global_step": 168000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 0.004999955274106163, | |
| "loss": 9.6382, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 0.004999821098024974, | |
| "loss": 9.5189, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.00022625107915810782, | |
| "eval_loss": 9.40969181060791, | |
| "eval_runtime": 6168.4952, | |
| "eval_samples_per_second": 27.228, | |
| "eval_steps_per_second": 1.702, | |
| "step": 1313 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 0.0049995974765573515, | |
| "loss": 9.363, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 0.00499928441770463, | |
| "loss": 9.2724, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 0.00499888193266828, | |
| "loss": 9.1635, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.0006430293828704117, | |
| "eval_loss": 9.041797637939453, | |
| "eval_runtime": 4719.6922, | |
| "eval_samples_per_second": 35.586, | |
| "eval_steps_per_second": 2.224, | |
| "step": 2626 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 0.004998390035849502, | |
| "loss": 8.8459, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 0.0049978087448487194, | |
| "loss": 8.3432, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.006466017683308029, | |
| "eval_loss": 7.845377445220947, | |
| "eval_runtime": 2624.5595, | |
| "eval_samples_per_second": 63.994, | |
| "eval_steps_per_second": 4.0, | |
| "step": 3939 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 0.004997138080464938, | |
| "loss": 7.8952, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 0.004996378066695007, | |
| "loss": 7.4467, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 0.004995528730732766, | |
| "loss": 6.9913, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_accuracy": 0.04887023309815129, | |
| "eval_loss": 6.311844825744629, | |
| "eval_runtime": 2630.4579, | |
| "eval_samples_per_second": 63.85, | |
| "eval_steps_per_second": 3.991, | |
| "step": 5252 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 0.0049945901029680604, | |
| "loss": 6.4357, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "learning_rate": 0.004993562216985665, | |
| "loss": 5.9619, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 0.004992445109564073, | |
| "loss": 5.5048, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_accuracy": 0.1423000208389152, | |
| "eval_loss": 4.949326515197754, | |
| "eval_runtime": 2650.8854, | |
| "eval_samples_per_second": 63.358, | |
| "eval_steps_per_second": 3.96, | |
| "step": 6565 | |
| }, | |
| { | |
| "epoch": 5.33, | |
| "learning_rate": 0.004991238820674189, | |
| "loss": 4.982, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 0.00498994339347789, | |
| "loss": 4.6895, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_accuracy": 0.24502396475246346, | |
| "eval_loss": 3.960129499435425, | |
| "eval_runtime": 2617.6761, | |
| "eval_samples_per_second": 64.162, | |
| "eval_steps_per_second": 4.01, | |
| "step": 7878 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 0.004988558874326489, | |
| "loss": 4.3479, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 6.47, | |
| "learning_rate": 0.004987085312759071, | |
| "loss": 3.9972, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 0.004985522761500722, | |
| "loss": 3.8881, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_accuracy": 0.31358994968890475, | |
| "eval_loss": 3.4186415672302246, | |
| "eval_runtime": 2616.4405, | |
| "eval_samples_per_second": 64.192, | |
| "eval_steps_per_second": 4.012, | |
| "step": 9191 | |
| }, | |
| { | |
| "epoch": 7.24, | |
| "learning_rate": 0.004983871276460643, | |
| "loss": 3.6025, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "learning_rate": 0.0049821309167301505, | |
| "loss": 3.4698, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 0.004980301744580561, | |
| "loss": 3.391, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_accuracy": 0.37655919740406657, | |
| "eval_loss": 2.9798192977905273, | |
| "eval_runtime": 2644.3802, | |
| "eval_samples_per_second": 63.514, | |
| "eval_steps_per_second": 3.97, | |
| "step": 10504 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 0.004978383825460959, | |
| "loss": 3.1113, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 8.76, | |
| "learning_rate": 0.004976377227995865, | |
| "loss": 3.0887, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_accuracy": 0.422124973951356, | |
| "eval_loss": 2.705430030822754, | |
| "eval_runtime": 2622.7588, | |
| "eval_samples_per_second": 64.038, | |
| "eval_steps_per_second": 4.003, | |
| "step": 11817 | |
| }, | |
| { | |
| "epoch": 9.14, | |
| "learning_rate": 0.00497428202398277, | |
| "loss": 2.925, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 9.52, | |
| "learning_rate": 0.004972098288389572, | |
| "loss": 2.7977, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 9.9, | |
| "learning_rate": 0.004969826099351891, | |
| "loss": 2.7935, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_accuracy": 0.45515167753267244, | |
| "eval_loss": 2.501314401626587, | |
| "eval_runtime": 2679.1466, | |
| "eval_samples_per_second": 62.69, | |
| "eval_steps_per_second": 3.918, | |
| "step": 13130 | |
| }, | |
| { | |
| "epoch": 10.28, | |
| "learning_rate": 0.004967465538170278, | |
| "loss": 2.5817, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 10.66, | |
| "learning_rate": 0.004965016689307298, | |
| "loss": 2.5629, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_accuracy": 0.4803846268345688, | |
| "eval_loss": 2.3581297397613525, | |
| "eval_runtime": 2615.9499, | |
| "eval_samples_per_second": 64.204, | |
| "eval_steps_per_second": 4.013, | |
| "step": 14443 | |
| }, | |
| { | |
| "epoch": 11.04, | |
| "learning_rate": 0.004962479640384515, | |
| "loss": 2.5549, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 11.42, | |
| "learning_rate": 0.004959854482179353, | |
| "loss": 2.3342, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 11.81, | |
| "learning_rate": 0.004957141308621851, | |
| "loss": 2.3777, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_accuracy": 0.48086689887172157, | |
| "eval_loss": 2.3543388843536377, | |
| "eval_runtime": 2613.6118, | |
| "eval_samples_per_second": 64.262, | |
| "eval_steps_per_second": 4.017, | |
| "step": 15756 | |
| }, | |
| { | |
| "epoch": 12.19, | |
| "learning_rate": 0.004954340216791298, | |
| "loss": 2.1546, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 12.57, | |
| "learning_rate": 0.004951451306912763, | |
| "loss": 2.2162, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 12.95, | |
| "learning_rate": 0.0049484746823535075, | |
| "loss": 2.2264, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_accuracy": 0.5178887201929088, | |
| "eval_loss": 2.163175344467163, | |
| "eval_runtime": 5970.7626, | |
| "eval_samples_per_second": 28.13, | |
| "eval_steps_per_second": 1.758, | |
| "step": 17069 | |
| }, | |
| { | |
| "epoch": 13.33, | |
| "learning_rate": 0.004945410449619286, | |
| "loss": 2.035, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 13.71, | |
| "learning_rate": 0.0049422587183505385, | |
| "loss": 2.0932, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_accuracy": 0.5219076538358489, | |
| "eval_loss": 2.136173725128174, | |
| "eval_runtime": 2759.3303, | |
| "eval_samples_per_second": 60.868, | |
| "eval_steps_per_second": 3.805, | |
| "step": 18382 | |
| }, | |
| { | |
| "epoch": 14.09, | |
| "learning_rate": 0.004939019601318461, | |
| "loss": 2.0545, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 14.47, | |
| "learning_rate": 0.004935693214420979, | |
| "loss": 1.9224, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 14.85, | |
| "learning_rate": 0.004932279676678594, | |
| "loss": 1.9667, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_accuracy": 0.5591200023815903, | |
| "eval_loss": 1.956689715385437, | |
| "eval_runtime": 2781.1993, | |
| "eval_samples_per_second": 60.389, | |
| "eval_steps_per_second": 3.775, | |
| "step": 19695 | |
| }, | |
| { | |
| "epoch": 15.23, | |
| "learning_rate": 0.004928779110230126, | |
| "loss": 1.8635, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 15.61, | |
| "learning_rate": 0.004925191640328348, | |
| "loss": 1.8396, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 15.99, | |
| "learning_rate": 0.004921517395335496, | |
| "loss": 1.8788, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_accuracy": 0.5609538269179244, | |
| "eval_loss": 1.9347097873687744, | |
| "eval_runtime": 2696.6196, | |
| "eval_samples_per_second": 62.284, | |
| "eval_steps_per_second": 3.893, | |
| "step": 21008 | |
| }, | |
| { | |
| "epoch": 16.37, | |
| "learning_rate": 0.004917756506718685, | |
| "loss": 1.7225, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 16.76, | |
| "learning_rate": 0.004913909109045198, | |
| "loss": 1.7705, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_accuracy": 0.5684439284332112, | |
| "eval_loss": 1.948270559310913, | |
| "eval_runtime": 2619.7155, | |
| "eval_samples_per_second": 64.112, | |
| "eval_steps_per_second": 4.007, | |
| "step": 22321 | |
| }, | |
| { | |
| "epoch": 17.14, | |
| "learning_rate": 0.004909975339977676, | |
| "loss": 1.7215, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 17.52, | |
| "learning_rate": 0.004905955340269188, | |
| "loss": 1.6354, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 17.9, | |
| "learning_rate": 0.004901849253758198, | |
| "loss": 1.7089, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_accuracy": 0.5790598672263404, | |
| "eval_loss": 1.8927983045578003, | |
| "eval_runtime": 2625.6089, | |
| "eval_samples_per_second": 63.968, | |
| "eval_steps_per_second": 3.998, | |
| "step": 23634 | |
| }, | |
| { | |
| "epoch": 18.28, | |
| "learning_rate": 0.004897657227363419, | |
| "loss": 1.579, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 18.66, | |
| "learning_rate": 0.004893379411078551, | |
| "loss": 1.6068, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_accuracy": 0.5855497008127177, | |
| "eval_loss": 1.843549370765686, | |
| "eval_runtime": 2690.4067, | |
| "eval_samples_per_second": 62.427, | |
| "eval_steps_per_second": 3.902, | |
| "step": 24947 | |
| }, | |
| { | |
| "epoch": 19.04, | |
| "learning_rate": 0.004889015957966921, | |
| "loss": 1.6159, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 19.42, | |
| "learning_rate": 0.004884567024156001, | |
| "loss": 1.4949, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 19.8, | |
| "learning_rate": 0.004880032768831825, | |
| "loss": 1.5572, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_accuracy": 0.5879670149742491, | |
| "eval_loss": 1.840835690498352, | |
| "eval_runtime": 2628.5762, | |
| "eval_samples_per_second": 63.896, | |
| "eval_steps_per_second": 3.994, | |
| "step": 26260 | |
| }, | |
| { | |
| "epoch": 20.18, | |
| "learning_rate": 0.004875413354233289, | |
| "loss": 1.4928, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 20.56, | |
| "learning_rate": 0.0048707089456463525, | |
| "loss": 1.4466, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 20.94, | |
| "learning_rate": 0.004865919711398117, | |
| "loss": 1.4938, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_accuracy": 0.6109672233634009, | |
| "eval_loss": 1.7412927150726318, | |
| "eval_runtime": 2614.3175, | |
| "eval_samples_per_second": 64.244, | |
| "eval_steps_per_second": 4.016, | |
| "step": 27573 | |
| }, | |
| { | |
| "epoch": 21.33, | |
| "learning_rate": 0.0048610458228508105, | |
| "loss": 1.3796, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 21.71, | |
| "learning_rate": 0.004856087454395648, | |
| "loss": 1.4182, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_accuracy": 0.6154565210919591, | |
| "eval_loss": 1.7196179628372192, | |
| "eval_runtime": 2640.6742, | |
| "eval_samples_per_second": 63.603, | |
| "eval_steps_per_second": 3.975, | |
| "step": 28886 | |
| }, | |
| { | |
| "epoch": 22.09, | |
| "learning_rate": 0.0048510447834466, | |
| "loss": 1.4096, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 22.47, | |
| "learning_rate": 0.0048459179904340385, | |
| "loss": 1.3308, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 22.85, | |
| "learning_rate": 0.004840707258798284, | |
| "loss": 1.3784, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_accuracy": 0.6238218570450418, | |
| "eval_loss": 1.7105211019515991, | |
| "eval_runtime": 2885.4428, | |
| "eval_samples_per_second": 58.208, | |
| "eval_steps_per_second": 3.638, | |
| "step": 30199 | |
| }, | |
| { | |
| "epoch": 23.23, | |
| "learning_rate": 0.004835412774983039, | |
| "loss": 1.3104, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 23.61, | |
| "learning_rate": 0.004830034728428723, | |
| "loss": 1.301, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 23.99, | |
| "learning_rate": 0.004824573311565684, | |
| "loss": 1.3578, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_accuracy": 0.6176059063439612, | |
| "eval_loss": 1.7758694887161255, | |
| "eval_runtime": 2617.3025, | |
| "eval_samples_per_second": 64.171, | |
| "eval_steps_per_second": 4.011, | |
| "step": 31512 | |
| }, | |
| { | |
| "epoch": 24.37, | |
| "learning_rate": 0.004819028719807326, | |
| "loss": 1.2176, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 24.75, | |
| "learning_rate": 0.004813401151543105, | |
| "loss": 1.2763, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_accuracy": 0.6218749069691286, | |
| "eval_loss": 1.7364755868911743, | |
| "eval_runtime": 5925.4842, | |
| "eval_samples_per_second": 28.345, | |
| "eval_steps_per_second": 1.772, | |
| "step": 32825 | |
| }, | |
| { | |
| "epoch": 25.13, | |
| "learning_rate": 0.004807690808131437, | |
| "loss": 1.2551, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 25.51, | |
| "learning_rate": 0.004801897893892493, | |
| "loss": 1.1861, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 25.89, | |
| "learning_rate": 0.004796022616100888, | |
| "loss": 1.2484, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_accuracy": 0.6199220029174481, | |
| "eval_loss": 1.7483400106430054, | |
| "eval_runtime": 2684.7282, | |
| "eval_samples_per_second": 62.559, | |
| "eval_steps_per_second": 3.91, | |
| "step": 34138 | |
| }, | |
| { | |
| "epoch": 26.28, | |
| "learning_rate": 0.004790065184978262, | |
| "loss": 1.1726, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 26.66, | |
| "learning_rate": 0.004784025813685761, | |
| "loss": 1.1936, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_accuracy": 0.6313953142210711, | |
| "eval_loss": 1.700264573097229, | |
| "eval_runtime": 2640.722, | |
| "eval_samples_per_second": 63.602, | |
| "eval_steps_per_second": 3.975, | |
| "step": 35451 | |
| }, | |
| { | |
| "epoch": 27.04, | |
| "learning_rate": 0.004777904718316407, | |
| "loss": 1.2231, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 27.42, | |
| "learning_rate": 0.00477170211788737, | |
| "loss": 1.1121, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 27.8, | |
| "learning_rate": 0.00476541823433213, | |
| "loss": 1.1499, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_accuracy": 0.6247209073859069, | |
| "eval_loss": 1.7398767471313477, | |
| "eval_runtime": 2691.471, | |
| "eval_samples_per_second": 62.403, | |
| "eval_steps_per_second": 3.9, | |
| "step": 36764 | |
| }, | |
| { | |
| "epoch": 28.18, | |
| "learning_rate": 0.004759053292492532, | |
| "loss": 1.1329, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 28.56, | |
| "learning_rate": 0.004752607520110748, | |
| "loss": 1.1121, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 28.94, | |
| "learning_rate": 0.004746081147821121, | |
| "loss": 1.1418, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_accuracy": 0.6316751510821351, | |
| "eval_loss": 1.7091107368469238, | |
| "eval_runtime": 2632.2426, | |
| "eval_samples_per_second": 63.807, | |
| "eval_steps_per_second": 3.988, | |
| "step": 38077 | |
| }, | |
| { | |
| "epoch": 29.32, | |
| "learning_rate": 0.0047394744091419205, | |
| "loss": 1.0493, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 29.7, | |
| "learning_rate": 0.004732787540466979, | |
| "loss": 1.0895, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_accuracy": 0.6383078800869281, | |
| "eval_loss": 1.7165930271148682, | |
| "eval_runtime": 2645.0605, | |
| "eval_samples_per_second": 63.498, | |
| "eval_steps_per_second": 3.969, | |
| "step": 39390 | |
| }, | |
| { | |
| "epoch": 30.08, | |
| "learning_rate": 0.004726020781057241, | |
| "loss": 1.0905, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 30.46, | |
| "learning_rate": 0.004719174373032195, | |
| "loss": 1.0191, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 30.85, | |
| "learning_rate": 0.0047122485613612175, | |
| "loss": 1.0706, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "eval_accuracy": 0.6373611979399244, | |
| "eval_loss": 1.7384090423583984, | |
| "eval_runtime": 2658.4828, | |
| "eval_samples_per_second": 63.177, | |
| "eval_steps_per_second": 3.949, | |
| "step": 40703 | |
| }, | |
| { | |
| "epoch": 31.23, | |
| "learning_rate": 0.004705243593854802, | |
| "loss": 1.0205, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 31.61, | |
| "learning_rate": 0.004698159721155694, | |
| "loss": 1.0072, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 31.99, | |
| "learning_rate": 0.004690997196729925, | |
| "loss": 1.0541, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_accuracy": 0.640868089666875, | |
| "eval_loss": 1.7335774898529053, | |
| "eval_runtime": 2595.2091, | |
| "eval_samples_per_second": 64.717, | |
| "eval_steps_per_second": 4.045, | |
| "step": 42016 | |
| }, | |
| { | |
| "epoch": 32.37, | |
| "learning_rate": 0.004683756276857741, | |
| "loss": 0.9544, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 32.75, | |
| "learning_rate": 0.004676437220624432, | |
| "loss": 1.0013, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "eval_accuracy": 0.6451489982435772, | |
| "eval_loss": 1.7185142040252686, | |
| "eval_runtime": 2605.1367, | |
| "eval_samples_per_second": 64.471, | |
| "eval_steps_per_second": 4.03, | |
| "step": 43329 | |
| }, | |
| { | |
| "epoch": 33.13, | |
| "learning_rate": 0.004669040289911064, | |
| "loss": 0.9986, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 33.51, | |
| "learning_rate": 0.004661565749385107, | |
| "loss": 0.9399, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 33.89, | |
| "learning_rate": 0.004654013866490966, | |
| "loss": 0.9811, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "eval_accuracy": 0.6478640111934745, | |
| "eval_loss": 1.7246044874191284, | |
| "eval_runtime": 2595.1453, | |
| "eval_samples_per_second": 64.719, | |
| "eval_steps_per_second": 4.045, | |
| "step": 44642 | |
| }, | |
| { | |
| "epoch": 34.27, | |
| "learning_rate": 0.004646384911440411, | |
| "loss": 0.914, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 34.65, | |
| "learning_rate": 0.004638679157202909, | |
| "loss": 0.9447, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "eval_accuracy": 0.6540263761126492, | |
| "eval_loss": 1.7245060205459595, | |
| "eval_runtime": 2605.4725, | |
| "eval_samples_per_second": 64.462, | |
| "eval_steps_per_second": 4.029, | |
| "step": 45955 | |
| }, | |
| { | |
| "epoch": 35.03, | |
| "learning_rate": 0.0005810494104962897, | |
| "loss": 0.8424, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 35.42, | |
| "learning_rate": 0.0005391326120998649, | |
| "loss": 0.692, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 35.8, | |
| "learning_rate": 0.0004986016371937035, | |
| "loss": 0.6587, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "eval_accuracy": 0.7018546634515197, | |
| "eval_loss": 1.5848796367645264, | |
| "eval_runtime": 4946.5461, | |
| "eval_samples_per_second": 33.954, | |
| "eval_steps_per_second": 2.122, | |
| "step": 47268 | |
| }, | |
| { | |
| "epoch": 36.18, | |
| "learning_rate": 0.0004594851306405101, | |
| "loss": 0.6234, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 36.56, | |
| "learning_rate": 0.0004218107376415739, | |
| "loss": 0.6053, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 36.94, | |
| "learning_rate": 0.00038560508419879514, | |
| "loss": 0.6044, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "eval_accuracy": 0.706195111785895, | |
| "eval_loss": 1.6146132946014404, | |
| "eval_runtime": 2680.1847, | |
| "eval_samples_per_second": 62.665, | |
| "eval_steps_per_second": 3.917, | |
| "step": 48581 | |
| }, | |
| { | |
| "epoch": 37.32, | |
| "learning_rate": 0.0003508937582970281, | |
| "loss": 0.5763, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 37.7, | |
| "learning_rate": 0.0003177012918200253, | |
| "loss": 0.572, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "eval_accuracy": 0.7081241999345063, | |
| "eval_loss": 1.6583054065704346, | |
| "eval_runtime": 2714.09, | |
| "eval_samples_per_second": 61.883, | |
| "eval_steps_per_second": 3.868, | |
| "step": 49894 | |
| }, | |
| { | |
| "epoch": 38.08, | |
| "learning_rate": 0.0002860511432127716, | |
| "loss": 0.5693, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 38.46, | |
| "learning_rate": 0.00025596568090246547, | |
| "loss": 0.5492, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 38.84, | |
| "learning_rate": 0.00022746616748985177, | |
| "loss": 0.545, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "eval_accuracy": 0.7086898276324015, | |
| "eval_loss": 1.6993166208267212, | |
| "eval_runtime": 8190.9805, | |
| "eval_samples_per_second": 20.505, | |
| "eval_steps_per_second": 1.282, | |
| "step": 51207 | |
| }, | |
| { | |
| "epoch": 39.22, | |
| "learning_rate": 0.0002005727447220948, | |
| "loss": 0.5409, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 39.6, | |
| "learning_rate": 0.00017530441925779654, | |
| "loss": 0.5351, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 39.98, | |
| "learning_rate": 0.00015167904923422986, | |
| "loss": 0.5341, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_accuracy": 0.7105593760233396, | |
| "eval_loss": 1.7077844142913818, | |
| "eval_runtime": 33225.1715, | |
| "eval_samples_per_second": 5.055, | |
| "eval_steps_per_second": 0.316, | |
| "step": 52520 | |
| }, | |
| { | |
| "epoch": 40.37, | |
| "learning_rate": 0.0001297133316462798, | |
| "loss": 0.5184, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 40.75, | |
| "learning_rate": 0.00010942279054600657, | |
| "loss": 0.5284, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "eval_accuracy": 0.7104641124110624, | |
| "eval_loss": 1.7240744829177856, | |
| "eval_runtime": 3401.4075, | |
| "eval_samples_per_second": 49.378, | |
| "eval_steps_per_second": 3.086, | |
| "step": 53833 | |
| }, | |
| { | |
| "epoch": 41.13, | |
| "learning_rate": 9.082176607117799e-05, | |
| "loss": 0.5239, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 41.51, | |
| "learning_rate": 7.392340431051969e-05, | |
| "loss": 0.5152, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 41.89, | |
| "learning_rate": 5.87396480128477e-05, | |
| "loss": 0.5186, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "eval_accuracy": 0.7111666815516061, | |
| "eval_loss": 1.7408066987991333, | |
| "eval_runtime": 4964.1178, | |
| "eval_samples_per_second": 33.834, | |
| "eval_steps_per_second": 2.115, | |
| "step": 55146 | |
| }, | |
| { | |
| "epoch": 42.27, | |
| "learning_rate": 4.528122814664942e-05, | |
| "loss": 0.5113, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 42.65, | |
| "learning_rate": 3.35576563160786e-05, | |
| "loss": 0.506, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 43.0, | |
| "eval_accuracy": 0.7105772379506415, | |
| "eval_loss": 1.7486644983291626, | |
| "eval_runtime": 6355.0593, | |
| "eval_samples_per_second": 26.429, | |
| "eval_steps_per_second": 1.652, | |
| "step": 56459 | |
| }, | |
| { | |
| "epoch": 43.03, | |
| "learning_rate": 2.357721803872476e-05, | |
| "loss": 0.5137, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 43.41, | |
| "learning_rate": 1.5346966889905334e-05, | |
| "loss": 0.5062, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 43.79, | |
| "learning_rate": 8.872719517621642e-06, | |
| "loss": 0.5043, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "eval_accuracy": 0.7108689827632402, | |
| "eval_loss": 1.7547060251235962, | |
| "eval_runtime": 4684.6858, | |
| "eval_samples_per_second": 35.852, | |
| "eval_steps_per_second": 2.241, | |
| "step": 57772 | |
| }, | |
| { | |
| "epoch": 44.17, | |
| "learning_rate": 4.159051531700819e-06, | |
| "loss": 0.5084, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 44.55, | |
| "learning_rate": 1.2092942700298037e-06, | |
| "loss": 0.5043, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 44.94, | |
| "learning_rate": 2.5532444165032865e-08, | |
| "loss": 0.5094, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "eval_accuracy": 0.711053556012027, | |
| "eval_loss": 1.7535769939422607, | |
| "eval_runtime": 2625.3482, | |
| "eval_samples_per_second": 63.974, | |
| "eval_steps_per_second": 3.999, | |
| "step": 59085 | |
| }, | |
| { | |
| "epoch": 45.32, | |
| "learning_rate": 0.0007032209920746526, | |
| "loss": 0.5386, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 45.7, | |
| "learning_rate": 0.0006689210221610515, | |
| "loss": 0.5547, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 46.0, | |
| "eval_accuracy": 0.7068619570718347, | |
| "eval_loss": 1.7073622941970825, | |
| "eval_runtime": 6607.5308, | |
| "eval_samples_per_second": 25.419, | |
| "eval_steps_per_second": 1.589, | |
| "step": 60398 | |
| }, | |
| { | |
| "epoch": 46.08, | |
| "learning_rate": 0.0006353490004690119, | |
| "loss": 0.553, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 46.46, | |
| "learning_rate": 0.0006025182736066154, | |
| "loss": 0.5329, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 46.84, | |
| "learning_rate": 0.000570441893479062, | |
| "loss": 0.5391, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 47.0, | |
| "eval_accuracy": 0.7089637105176982, | |
| "eval_loss": 1.740086317062378, | |
| "eval_runtime": 5830.6066, | |
| "eval_samples_per_second": 28.806, | |
| "eval_steps_per_second": 1.8, | |
| "step": 61711 | |
| }, | |
| { | |
| "epoch": 47.22, | |
| "learning_rate": 0.0005391326120998649, | |
| "loss": 0.5284, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 47.6, | |
| "learning_rate": 0.0005086028765212663, | |
| "loss": 0.5223, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 47.98, | |
| "learning_rate": 0.00047886482388589615, | |
| "loss": 0.5253, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "eval_accuracy": 0.7092852252091334, | |
| "eval_loss": 1.7769759893417358, | |
| "eval_runtime": 2617.8614, | |
| "eval_samples_per_second": 64.157, | |
| "eval_steps_per_second": 4.01, | |
| "step": 63024 | |
| }, | |
| { | |
| "epoch": 48.36, | |
| "learning_rate": 0.0004499302766016297, | |
| "loss": 0.5041, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 48.74, | |
| "learning_rate": 0.0004218107376415739, | |
| "loss": 0.5066, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 49.0, | |
| "eval_accuracy": 0.7102319073561371, | |
| "eval_loss": 1.813459873199463, | |
| "eval_runtime": 2608.5743, | |
| "eval_samples_per_second": 64.386, | |
| "eval_steps_per_second": 4.024, | |
| "step": 64337 | |
| }, | |
| { | |
| "epoch": 49.12, | |
| "learning_rate": 0.00039451738597103935, | |
| "loss": 0.5057, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 49.5, | |
| "learning_rate": 0.00036806107210333034, | |
| "loss": 0.4897, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 49.89, | |
| "learning_rate": 0.00034245231378610494, | |
| "loss": 0.495, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "eval_accuracy": 0.7110476020362597, | |
| "eval_loss": 1.8451627492904663, | |
| "eval_runtime": 2646.4398, | |
| "eval_samples_per_second": 63.465, | |
| "eval_steps_per_second": 3.967, | |
| "step": 65650 | |
| }, | |
| { | |
| "epoch": 50.27, | |
| "learning_rate": 0.0003177012918200253, | |
| "loss": 0.4867, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 50.65, | |
| "learning_rate": 0.0002938178460113711, | |
| "loss": 0.4813, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 51.0, | |
| "eval_accuracy": 0.7107320413205919, | |
| "eval_loss": 1.8846126794815063, | |
| "eval_runtime": 5243.8791, | |
| "eval_samples_per_second": 32.029, | |
| "eval_steps_per_second": 2.002, | |
| "step": 66963 | |
| }, | |
| { | |
| "epoch": 51.03, | |
| "learning_rate": 0.00027081147126020445, | |
| "loss": 0.4829, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 51.41, | |
| "learning_rate": 0.00024869131378566, | |
| "loss": 0.466, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 51.79, | |
| "learning_rate": 0.00022746616748985177, | |
| "loss": 0.4704, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "eval_accuracy": 0.7123693846566045, | |
| "eval_loss": 1.8989039659500122, | |
| "eval_runtime": 2607.2421, | |
| "eval_samples_per_second": 64.419, | |
| "eval_steps_per_second": 4.026, | |
| "step": 68276 | |
| }, | |
| { | |
| "epoch": 52.17, | |
| "learning_rate": 0.00020714447046184416, | |
| "loss": 0.47, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 52.55, | |
| "learning_rate": 0.0001877343016230762, | |
| "loss": 0.465, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 52.93, | |
| "learning_rate": 0.0001692433775155744, | |
| "loss": 0.4689, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 53.0, | |
| "eval_accuracy": 0.7131850793367271, | |
| "eval_loss": 1.931063175201416, | |
| "eval_runtime": 2614.8439, | |
| "eval_samples_per_second": 64.231, | |
| "eval_steps_per_second": 4.015, | |
| "step": 69589 | |
| }, | |
| { | |
| "epoch": 53.31, | |
| "learning_rate": 0.00015167904923422986, | |
| "loss": 0.4572, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 53.69, | |
| "learning_rate": 0.0001350482995043595, | |
| "loss": 0.4611, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 54.0, | |
| "eval_accuracy": 0.7130659998213807, | |
| "eval_loss": 1.93539559841156, | |
| "eval_runtime": 2604.5499, | |
| "eval_samples_per_second": 64.485, | |
| "eval_steps_per_second": 4.031, | |
| "step": 70902 | |
| }, | |
| { | |
| "epoch": 54.07, | |
| "learning_rate": 0.00011935773990571164, | |
| "loss": 0.4579, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 54.46, | |
| "learning_rate": 0.00010461360824402521, | |
| "loss": 0.4554, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 54.84, | |
| "learning_rate": 9.082176607117799e-05, | |
| "loss": 0.4547, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 55.0, | |
| "eval_accuracy": 0.7133160668036082, | |
| "eval_loss": 1.974073052406311, | |
| "eval_runtime": 2574.37, | |
| "eval_samples_per_second": 65.241, | |
| "eval_steps_per_second": 4.078, | |
| "step": 72215 | |
| }, | |
| { | |
| "epoch": 55.22, | |
| "learning_rate": 7.798769635491976e-05, | |
| "loss": 0.4497, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 55.6, | |
| "learning_rate": 6.611650129910801e-05, | |
| "loss": 0.4458, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 55.98, | |
| "learning_rate": 5.521290031532178e-05, | |
| "loss": 0.4481, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "eval_accuracy": 0.713071953797148, | |
| "eval_loss": 1.9898993968963623, | |
| "eval_runtime": 2622.4841, | |
| "eval_samples_per_second": 64.044, | |
| "eval_steps_per_second": 4.003, | |
| "step": 73528 | |
| }, | |
| { | |
| "epoch": 56.36, | |
| "learning_rate": 0.00045401414969093167, | |
| "loss": 0.4607, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 56.74, | |
| "learning_rate": 0.0004297610695623985, | |
| "loss": 0.4709, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 57.0, | |
| "eval_accuracy": 0.7104283885564586, | |
| "eval_loss": 1.9411875009536743, | |
| "eval_runtime": 3269.1549, | |
| "eval_samples_per_second": 51.376, | |
| "eval_steps_per_second": 3.211, | |
| "step": 74841 | |
| }, | |
| { | |
| "epoch": 57.12, | |
| "learning_rate": 0.0004061126673698029, | |
| "loss": 0.4736, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 57.5, | |
| "learning_rate": 0.0003830758503675777, | |
| "loss": 0.4601, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 57.88, | |
| "learning_rate": 0.00036065734717759815, | |
| "loss": 0.4647, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 58.0, | |
| "eval_accuracy": 0.7097913131493555, | |
| "eval_loss": 1.9707293510437012, | |
| "eval_runtime": 2595.9104, | |
| "eval_samples_per_second": 64.7, | |
| "eval_steps_per_second": 4.044, | |
| "step": 76154 | |
| }, | |
| { | |
| "epoch": 58.26, | |
| "learning_rate": 0.00033886370582388, | |
| "loss": 0.4562, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 58.64, | |
| "learning_rate": 0.0003177012918200253, | |
| "loss": 0.4566, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 59.0, | |
| "eval_accuracy": 0.7116251376856896, | |
| "eval_loss": 2.0150792598724365, | |
| "eval_runtime": 2589.1949, | |
| "eval_samples_per_second": 64.868, | |
| "eval_steps_per_second": 4.055, | |
| "step": 77467 | |
| }, | |
| { | |
| "epoch": 59.03, | |
| "learning_rate": 0.0002971762863099739, | |
| "loss": 0.4535, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 59.41, | |
| "learning_rate": 0.0002772946842626087, | |
| "loss": 0.448, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 59.79, | |
| "learning_rate": 0.000258062292720736, | |
| "loss": 0.4511, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "eval_accuracy": 0.7113929326307642, | |
| "eval_loss": 2.0363075733184814, | |
| "eval_runtime": 2640.6575, | |
| "eval_samples_per_second": 63.603, | |
| "eval_steps_per_second": 3.976, | |
| "step": 78780 | |
| }, | |
| { | |
| "epoch": 60.17, | |
| "learning_rate": 0.00023948472910495817, | |
| "loss": 0.4446, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 60.55, | |
| "learning_rate": 0.0002215674195729317, | |
| "loss": 0.4403, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 60.93, | |
| "learning_rate": 0.0002043155974344868, | |
| "loss": 0.4423, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 61.0, | |
| "eval_accuracy": 0.7112143133577447, | |
| "eval_loss": 2.0710268020629883, | |
| "eval_runtime": 2593.0457, | |
| "eval_samples_per_second": 64.771, | |
| "eval_steps_per_second": 4.049, | |
| "step": 80093 | |
| }, | |
| { | |
| "epoch": 61.31, | |
| "learning_rate": 0.0001877343016230762, | |
| "loss": 0.4304, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 61.69, | |
| "learning_rate": 0.00017182837522399742, | |
| "loss": 0.4356, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 62.0, | |
| "eval_accuracy": 0.7116251376856896, | |
| "eval_loss": 2.061058521270752, | |
| "eval_runtime": 2582.6346, | |
| "eval_samples_per_second": 65.032, | |
| "eval_steps_per_second": 4.065, | |
| "step": 81406 | |
| }, | |
| { | |
| "epoch": 62.07, | |
| "learning_rate": 0.00015660246405981793, | |
| "loss": 0.4332, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 62.45, | |
| "learning_rate": 0.00014206101533341926, | |
| "loss": 0.43, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 62.83, | |
| "learning_rate": 0.0001282082763290507, | |
| "loss": 0.4272, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 63.0, | |
| "eval_accuracy": 0.7117918490071745, | |
| "eval_loss": 2.089148998260498, | |
| "eval_runtime": 2586.4124, | |
| "eval_samples_per_second": 64.937, | |
| "eval_steps_per_second": 4.059, | |
| "step": 82719 | |
| }, | |
| { | |
| "epoch": 63.21, | |
| "learning_rate": 0.00011504829317177723, | |
| "loss": 0.4259, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 63.59, | |
| "learning_rate": 0.00010258490964568406, | |
| "loss": 0.4271, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 63.98, | |
| "learning_rate": 9.082176607117799e-05, | |
| "loss": 0.4254, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "eval_accuracy": 0.7123812926081391, | |
| "eval_loss": 2.0879101753234863, | |
| "eval_runtime": 2610.5128, | |
| "eval_samples_per_second": 64.338, | |
| "eval_steps_per_second": 4.021, | |
| "step": 84032 | |
| }, | |
| { | |
| "epoch": 64.36, | |
| "learning_rate": 7.976229824171872e-05, | |
| "loss": 0.4249, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 64.74, | |
| "learning_rate": 6.940973642028724e-05, | |
| "loss": 0.4221, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 65.0, | |
| "eval_accuracy": 0.7130957697002174, | |
| "eval_loss": 2.1167192459106445, | |
| "eval_runtime": 2625.6926, | |
| "eval_samples_per_second": 63.966, | |
| "eval_steps_per_second": 3.998, | |
| "step": 85345 | |
| }, | |
| { | |
| "epoch": 65.12, | |
| "learning_rate": 5.976710439588779e-05, | |
| "loss": 0.4227, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 65.5, | |
| "learning_rate": 5.083721860035529e-05, | |
| "loss": 0.421, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 65.88, | |
| "learning_rate": 4.262268728572882e-05, | |
| "loss": 0.4189, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 66.0, | |
| "eval_accuracy": 0.7129111964514304, | |
| "eval_loss": 2.1362690925598145, | |
| "eval_runtime": 2775.8163, | |
| "eval_samples_per_second": 60.507, | |
| "eval_steps_per_second": 3.782, | |
| "step": 86658 | |
| }, | |
| { | |
| "epoch": 66.26, | |
| "learning_rate": 3.512590976242902e-05, | |
| "loss": 0.4192, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 66.64, | |
| "learning_rate": 2.834907569846612e-05, | |
| "loss": 0.4219, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 67.0, | |
| "eval_accuracy": 0.712958828257569, | |
| "eval_loss": 2.1355273723602295, | |
| "eval_runtime": 2581.0691, | |
| "eval_samples_per_second": 65.072, | |
| "eval_steps_per_second": 4.067, | |
| "step": 87971 | |
| }, | |
| { | |
| "epoch": 67.02, | |
| "learning_rate": 2.2294164479878255e-05, | |
| "loss": 0.4197, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 67.4, | |
| "learning_rate": 1.6962944632589183e-05, | |
| "loss": 0.4157, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 67.78, | |
| "learning_rate": 1.2356973305856779e-05, | |
| "loss": 0.4149, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "eval_accuracy": 0.7132267571670984, | |
| "eval_loss": 2.1466283798217773, | |
| "eval_runtime": 2769.7004, | |
| "eval_samples_per_second": 60.64, | |
| "eval_steps_per_second": 3.79, | |
| "step": 89284 | |
| }, | |
| { | |
| "epoch": 68.16, | |
| "learning_rate": 8.477595817457106e-06, | |
| "loss": 0.4138, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 68.55, | |
| "learning_rate": 5.325945260744314e-06, | |
| "loss": 0.4147, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 68.93, | |
| "learning_rate": 2.90294217369369e-06, | |
| "loss": 0.4125, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 69.0, | |
| "eval_accuracy": 0.7131434015063559, | |
| "eval_loss": 2.1478145122528076, | |
| "eval_runtime": 2585.5905, | |
| "eval_samples_per_second": 64.958, | |
| "eval_steps_per_second": 4.06, | |
| "step": 90597 | |
| }, | |
| { | |
| "epoch": 69.31, | |
| "learning_rate": 1.2092942700298037e-06, | |
| "loss": 0.4146, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 69.69, | |
| "learning_rate": 2.4549623251851615e-07, | |
| "loss": 0.4162, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 70.0, | |
| "eval_accuracy": 0.7132267571670984, | |
| "eval_loss": 2.1484131813049316, | |
| "eval_runtime": 2581.5568, | |
| "eval_samples_per_second": 65.06, | |
| "eval_steps_per_second": 4.067, | |
| "step": 91910 | |
| }, | |
| { | |
| "epoch": 35.05, | |
| "learning_rate": 0.003247619047619048, | |
| "loss": 0.5379, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 35.24, | |
| "learning_rate": 0.0032380952380952383, | |
| "loss": 0.6972, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 35.43, | |
| "learning_rate": 0.0032285714285714283, | |
| "loss": 0.7947, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 35.62, | |
| "learning_rate": 0.003219047619047619, | |
| "loss": 0.8301, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 35.81, | |
| "learning_rate": 0.0032095238095238092, | |
| "loss": 0.8519, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "learning_rate": 0.0032, | |
| "loss": 0.8802, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "eval_accuracy": 0.6566997112321753, | |
| "eval_loss": 1.7587852478027344, | |
| "eval_runtime": 4381.2628, | |
| "eval_samples_per_second": 38.335, | |
| "eval_steps_per_second": 4.792, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 36.19, | |
| "learning_rate": 0.00319047619047619, | |
| "loss": 0.7776, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 36.38, | |
| "learning_rate": 0.003180952380952381, | |
| "loss": 0.8104, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 36.57, | |
| "learning_rate": 0.003171428571428571, | |
| "loss": 0.8249, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 36.76, | |
| "learning_rate": 0.003161904761904762, | |
| "loss": 0.8548, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 36.95, | |
| "learning_rate": 0.0031523809523809525, | |
| "loss": 0.8772, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "eval_accuracy": 0.6668571938912209, | |
| "eval_loss": 1.6901220083236694, | |
| "eval_runtime": 2781.7361, | |
| "eval_samples_per_second": 60.378, | |
| "eval_steps_per_second": 7.547, | |
| "step": 97125 | |
| }, | |
| { | |
| "epoch": 37.14, | |
| "learning_rate": 0.003142857142857143, | |
| "loss": 0.793, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 37.33, | |
| "learning_rate": 0.0031333333333333335, | |
| "loss": 0.7985, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 37.52, | |
| "learning_rate": 0.003123809523809524, | |
| "loss": 0.8144, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 37.71, | |
| "learning_rate": 0.0031142857142857144, | |
| "loss": 0.8338, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 37.9, | |
| "learning_rate": 0.003104761904761905, | |
| "loss": 0.847, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "eval_accuracy": 0.6682742401238427, | |
| "eval_loss": 1.7208112478256226, | |
| "eval_runtime": 2692.0576, | |
| "eval_samples_per_second": 62.389, | |
| "eval_steps_per_second": 7.799, | |
| "step": 99750 | |
| }, | |
| { | |
| "epoch": 38.1, | |
| "learning_rate": 0.0030952380952380953, | |
| "loss": 0.8028, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 38.29, | |
| "learning_rate": 0.0030857142857142854, | |
| "loss": 0.764, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 38.48, | |
| "learning_rate": 0.0030761904761904763, | |
| "loss": 0.7868, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 38.67, | |
| "learning_rate": 0.0030666666666666663, | |
| "loss": 0.8082, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 38.86, | |
| "learning_rate": 0.0030571428571428572, | |
| "loss": 0.8349, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "eval_accuracy": 0.6679705873597095, | |
| "eval_loss": 1.7477226257324219, | |
| "eval_runtime": 2679.7183, | |
| "eval_samples_per_second": 62.676, | |
| "eval_steps_per_second": 7.835, | |
| "step": 102375 | |
| }, | |
| { | |
| "epoch": 39.05, | |
| "learning_rate": 0.003047619047619048, | |
| "loss": 0.8031, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 39.24, | |
| "learning_rate": 0.003038095238095238, | |
| "loss": 0.7409, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 39.43, | |
| "learning_rate": 0.003028571428571429, | |
| "loss": 0.7646, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 39.62, | |
| "learning_rate": 0.003019047619047619, | |
| "loss": 0.7781, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 39.81, | |
| "learning_rate": 0.0030095238095238096, | |
| "loss": 0.8094, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "learning_rate": 0.003, | |
| "loss": 0.8159, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_accuracy": 0.6640588252805811, | |
| "eval_loss": 1.7669143676757812, | |
| "eval_runtime": 2681.934, | |
| "eval_samples_per_second": 62.625, | |
| "eval_steps_per_second": 7.828, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 40.19, | |
| "learning_rate": 0.0029904761904761905, | |
| "loss": 0.7129, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 40.38, | |
| "learning_rate": 0.002980952380952381, | |
| "loss": 0.7458, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 40.57, | |
| "learning_rate": 0.0029714285714285715, | |
| "loss": 0.7784, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 40.76, | |
| "learning_rate": 0.002961904761904762, | |
| "loss": 0.7782, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 40.95, | |
| "learning_rate": 0.0029523809523809524, | |
| "loss": 0.7894, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "eval_accuracy": 0.669774642017207, | |
| "eval_loss": 1.79474675655365, | |
| "eval_runtime": 2697.7877, | |
| "eval_samples_per_second": 62.257, | |
| "eval_steps_per_second": 7.782, | |
| "step": 107625 | |
| }, | |
| { | |
| "epoch": 41.14, | |
| "learning_rate": 0.002942857142857143, | |
| "loss": 0.7193, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 41.33, | |
| "learning_rate": 0.0029333333333333334, | |
| "loss": 0.7162, | |
| "step": 108500 | |
| }, | |
| { | |
| "epoch": 41.52, | |
| "learning_rate": 0.0029238095238095243, | |
| "loss": 0.7387, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 41.71, | |
| "learning_rate": 0.0029142857142857143, | |
| "loss": 0.7592, | |
| "step": 109500 | |
| }, | |
| { | |
| "epoch": 41.9, | |
| "learning_rate": 0.002904761904761905, | |
| "loss": 0.765, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "eval_accuracy": 0.6673335119526064, | |
| "eval_loss": 1.7770148515701294, | |
| "eval_runtime": 2711.4758, | |
| "eval_samples_per_second": 61.942, | |
| "eval_steps_per_second": 7.743, | |
| "step": 110250 | |
| }, | |
| { | |
| "epoch": 42.1, | |
| "learning_rate": 0.0028952380952380953, | |
| "loss": 0.7184, | |
| "step": 110500 | |
| }, | |
| { | |
| "epoch": 42.29, | |
| "learning_rate": 0.002885714285714286, | |
| "loss": 0.6956, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 42.48, | |
| "learning_rate": 0.002876190476190476, | |
| "loss": 0.713, | |
| "step": 111500 | |
| }, | |
| { | |
| "epoch": 42.67, | |
| "learning_rate": 0.0028666666666666667, | |
| "loss": 0.7328, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 42.86, | |
| "learning_rate": 0.002857142857142857, | |
| "loss": 0.7417, | |
| "step": 112500 | |
| }, | |
| { | |
| "epoch": 43.0, | |
| "eval_accuracy": 0.6685659849364413, | |
| "eval_loss": 1.8335988521575928, | |
| "eval_runtime": 4071.9751, | |
| "eval_samples_per_second": 41.247, | |
| "eval_steps_per_second": 5.156, | |
| "step": 112875 | |
| }, | |
| { | |
| "epoch": 43.05, | |
| "learning_rate": 0.0028476190476190476, | |
| "loss": 0.7272, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 43.24, | |
| "learning_rate": 0.002838095238095238, | |
| "loss": 0.6728, | |
| "step": 113500 | |
| }, | |
| { | |
| "epoch": 43.43, | |
| "learning_rate": 0.0028285714285714286, | |
| "loss": 0.6954, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 43.62, | |
| "learning_rate": 0.002819047619047619, | |
| "loss": 0.7025, | |
| "step": 114500 | |
| }, | |
| { | |
| "epoch": 43.81, | |
| "learning_rate": 0.0028095238095238095, | |
| "loss": 0.7208, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "learning_rate": 0.0028000000000000004, | |
| "loss": 0.7214, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "eval_accuracy": 0.6755142746569022, | |
| "eval_loss": 1.7522037029266357, | |
| "eval_runtime": 3540.85, | |
| "eval_samples_per_second": 47.434, | |
| "eval_steps_per_second": 5.929, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 44.19, | |
| "learning_rate": 0.0027904761904761904, | |
| "loss": 0.6433, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 44.38, | |
| "learning_rate": 0.0027809523809523813, | |
| "loss": 0.6634, | |
| "step": 116500 | |
| }, | |
| { | |
| "epoch": 44.57, | |
| "learning_rate": 0.0027714285714285714, | |
| "loss": 0.6854, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 44.76, | |
| "learning_rate": 0.0027619047619047623, | |
| "loss": 0.7026, | |
| "step": 117500 | |
| }, | |
| { | |
| "epoch": 44.95, | |
| "learning_rate": 0.0027523809523809523, | |
| "loss": 0.7113, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "eval_accuracy": 0.6773778690720729, | |
| "eval_loss": 1.7851730585098267, | |
| "eval_runtime": 2781.7496, | |
| "eval_samples_per_second": 60.377, | |
| "eval_steps_per_second": 7.547, | |
| "step": 118125 | |
| }, | |
| { | |
| "epoch": 45.14, | |
| "learning_rate": 0.0027428571428571432, | |
| "loss": 0.6189, | |
| "step": 118500 | |
| }, | |
| { | |
| "epoch": 45.33, | |
| "learning_rate": 0.0027333333333333333, | |
| "loss": 0.6497, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 45.52, | |
| "learning_rate": 0.0027238095238095237, | |
| "loss": 0.666, | |
| "step": 119500 | |
| }, | |
| { | |
| "epoch": 45.71, | |
| "learning_rate": 0.0027142857142857142, | |
| "loss": 0.6832, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 45.9, | |
| "learning_rate": 0.0027047619047619047, | |
| "loss": 0.6954, | |
| "step": 120500 | |
| }, | |
| { | |
| "epoch": 46.0, | |
| "eval_accuracy": 0.6773719150963056, | |
| "eval_loss": 1.7557201385498047, | |
| "eval_runtime": 6130.4043, | |
| "eval_samples_per_second": 27.397, | |
| "eval_steps_per_second": 3.425, | |
| "step": 120750 | |
| }, | |
| { | |
| "epoch": 46.1, | |
| "learning_rate": 0.002695238095238095, | |
| "loss": 0.6582, | |
| "step": 121000 | |
| }, | |
| { | |
| "epoch": 46.29, | |
| "learning_rate": 0.0026857142857142856, | |
| "loss": 0.6174, | |
| "step": 121500 | |
| }, | |
| { | |
| "epoch": 46.48, | |
| "learning_rate": 0.0026761904761904765, | |
| "loss": 0.6356, | |
| "step": 122000 | |
| }, | |
| { | |
| "epoch": 46.67, | |
| "learning_rate": 0.0026666666666666666, | |
| "loss": 0.6565, | |
| "step": 122500 | |
| }, | |
| { | |
| "epoch": 46.86, | |
| "learning_rate": 0.0026571428571428575, | |
| "loss": 0.6658, | |
| "step": 123000 | |
| }, | |
| { | |
| "epoch": 47.0, | |
| "eval_accuracy": 0.678836593135066, | |
| "eval_loss": 1.8116456270217896, | |
| "eval_runtime": 5170.8601, | |
| "eval_samples_per_second": 32.481, | |
| "eval_steps_per_second": 4.06, | |
| "step": 123375 | |
| }, | |
| { | |
| "epoch": 47.05, | |
| "learning_rate": 0.0026476190476190475, | |
| "loss": 0.6552, | |
| "step": 123500 | |
| }, | |
| { | |
| "epoch": 47.24, | |
| "learning_rate": 0.0026380952380952384, | |
| "loss": 0.5994, | |
| "step": 124000 | |
| }, | |
| { | |
| "epoch": 47.43, | |
| "learning_rate": 0.0026285714285714285, | |
| "loss": 0.6221, | |
| "step": 124500 | |
| }, | |
| { | |
| "epoch": 47.62, | |
| "learning_rate": 0.0026190476190476194, | |
| "loss": 0.635, | |
| "step": 125000 | |
| }, | |
| { | |
| "epoch": 47.81, | |
| "learning_rate": 0.0026095238095238094, | |
| "loss": 0.648, | |
| "step": 125500 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "learning_rate": 0.0026000000000000003, | |
| "loss": 0.6593, | |
| "step": 126000 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "eval_accuracy": 0.6829388824387485, | |
| "eval_loss": 1.8153612613677979, | |
| "eval_runtime": 3984.2658, | |
| "eval_samples_per_second": 42.155, | |
| "eval_steps_per_second": 5.269, | |
| "step": 126000 | |
| }, | |
| { | |
| "epoch": 48.19, | |
| "learning_rate": 0.0025904761904761904, | |
| "loss": 0.58, | |
| "step": 126500 | |
| }, | |
| { | |
| "epoch": 48.38, | |
| "learning_rate": 0.0025809523809523813, | |
| "loss": 0.5966, | |
| "step": 127000 | |
| }, | |
| { | |
| "epoch": 48.57, | |
| "learning_rate": 0.0025714285714285713, | |
| "loss": 0.6184, | |
| "step": 127500 | |
| }, | |
| { | |
| "epoch": 48.76, | |
| "learning_rate": 0.0025619047619047618, | |
| "loss": 0.6257, | |
| "step": 128000 | |
| }, | |
| { | |
| "epoch": 48.95, | |
| "learning_rate": 0.0025523809523809527, | |
| "loss": 0.6384, | |
| "step": 128500 | |
| }, | |
| { | |
| "epoch": 49.0, | |
| "eval_accuracy": 0.6795213003483076, | |
| "eval_loss": 1.7874691486358643, | |
| "eval_runtime": 3000.761, | |
| "eval_samples_per_second": 55.971, | |
| "eval_steps_per_second": 6.997, | |
| "step": 128625 | |
| }, | |
| { | |
| "epoch": 49.14, | |
| "learning_rate": 0.0025428571428571427, | |
| "loss": 0.5834, | |
| "step": 129000 | |
| }, | |
| { | |
| "epoch": 49.33, | |
| "learning_rate": 0.0025333333333333336, | |
| "loss": 0.5847, | |
| "step": 129500 | |
| }, | |
| { | |
| "epoch": 49.52, | |
| "learning_rate": 0.0025238095238095237, | |
| "loss": 0.5981, | |
| "step": 130000 | |
| }, | |
| { | |
| "epoch": 49.71, | |
| "learning_rate": 0.0025142857142857146, | |
| "loss": 0.6141, | |
| "step": 130500 | |
| }, | |
| { | |
| "epoch": 49.9, | |
| "learning_rate": 0.0025047619047619046, | |
| "loss": 0.6257, | |
| "step": 131000 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "eval_accuracy": 0.6811050579024144, | |
| "eval_loss": 1.882097840309143, | |
| "eval_runtime": 2757.9073, | |
| "eval_samples_per_second": 60.899, | |
| "eval_steps_per_second": 7.613, | |
| "step": 131250 | |
| }, | |
| { | |
| "epoch": 50.1, | |
| "learning_rate": 0.0024952380952380955, | |
| "loss": 0.5799, | |
| "step": 131500 | |
| }, | |
| { | |
| "epoch": 50.29, | |
| "learning_rate": 0.002485714285714286, | |
| "loss": 0.5715, | |
| "step": 132000 | |
| }, | |
| { | |
| "epoch": 50.48, | |
| "learning_rate": 0.0024761904761904764, | |
| "loss": 0.5847, | |
| "step": 132500 | |
| }, | |
| { | |
| "epoch": 50.67, | |
| "learning_rate": 0.002466666666666667, | |
| "loss": 0.5942, | |
| "step": 133000 | |
| }, | |
| { | |
| "epoch": 50.86, | |
| "learning_rate": 0.0024571428571428574, | |
| "loss": 0.5999, | |
| "step": 133500 | |
| }, | |
| { | |
| "epoch": 51.0, | |
| "eval_accuracy": 0.6871840671608467, | |
| "eval_loss": 1.8406310081481934, | |
| "eval_runtime": 2763.7082, | |
| "eval_samples_per_second": 60.772, | |
| "eval_steps_per_second": 7.597, | |
| "step": 133875 | |
| }, | |
| { | |
| "epoch": 51.05, | |
| "learning_rate": 0.002447619047619048, | |
| "loss": 0.5872, | |
| "step": 134000 | |
| }, | |
| { | |
| "epoch": 51.24, | |
| "learning_rate": 0.0024380952380952383, | |
| "loss": 0.5511, | |
| "step": 134500 | |
| }, | |
| { | |
| "epoch": 51.43, | |
| "learning_rate": 0.002428571428571429, | |
| "loss": 0.5592, | |
| "step": 135000 | |
| }, | |
| { | |
| "epoch": 51.62, | |
| "learning_rate": 0.002419047619047619, | |
| "loss": 0.5658, | |
| "step": 135500 | |
| }, | |
| { | |
| "epoch": 51.81, | |
| "learning_rate": 0.0024095238095238093, | |
| "loss": 0.5887, | |
| "step": 136000 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "learning_rate": 0.0024, | |
| "loss": 0.5924, | |
| "step": 136500 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "eval_accuracy": 0.6764728647554404, | |
| "eval_loss": 1.9697257280349731, | |
| "eval_runtime": 2760.1617, | |
| "eval_samples_per_second": 60.85, | |
| "eval_steps_per_second": 7.606, | |
| "step": 136500 | |
| }, | |
| { | |
| "epoch": 52.19, | |
| "learning_rate": 0.0023904761904761903, | |
| "loss": 0.5313, | |
| "step": 137000 | |
| }, | |
| { | |
| "epoch": 52.38, | |
| "learning_rate": 0.0023809523809523807, | |
| "loss": 0.5451, | |
| "step": 137500 | |
| }, | |
| { | |
| "epoch": 52.57, | |
| "learning_rate": 0.0023714285714285716, | |
| "loss": 0.56, | |
| "step": 138000 | |
| }, | |
| { | |
| "epoch": 52.76, | |
| "learning_rate": 0.002361904761904762, | |
| "loss": 0.5676, | |
| "step": 138500 | |
| }, | |
| { | |
| "epoch": 52.95, | |
| "learning_rate": 0.0023523809523809526, | |
| "loss": 0.5812, | |
| "step": 139000 | |
| }, | |
| { | |
| "epoch": 53.0, | |
| "eval_accuracy": 0.6818195349944925, | |
| "eval_loss": 1.9344266653060913, | |
| "eval_runtime": 2852.1671, | |
| "eval_samples_per_second": 58.887, | |
| "eval_steps_per_second": 7.361, | |
| "step": 139125 | |
| }, | |
| { | |
| "epoch": 53.14, | |
| "learning_rate": 0.002342857142857143, | |
| "loss": 0.506, | |
| "step": 139500 | |
| }, | |
| { | |
| "epoch": 53.33, | |
| "learning_rate": 0.0023333333333333335, | |
| "loss": 0.541, | |
| "step": 140000 | |
| }, | |
| { | |
| "epoch": 53.52, | |
| "learning_rate": 0.002323809523809524, | |
| "loss": 0.5445, | |
| "step": 140500 | |
| }, | |
| { | |
| "epoch": 53.71, | |
| "learning_rate": 0.0023142857142857145, | |
| "loss": 0.5547, | |
| "step": 141000 | |
| }, | |
| { | |
| "epoch": 53.9, | |
| "learning_rate": 0.002304761904761905, | |
| "loss": 0.5521, | |
| "step": 141500 | |
| }, | |
| { | |
| "epoch": 54.0, | |
| "eval_accuracy": 0.6802060075615493, | |
| "eval_loss": 1.9845408201217651, | |
| "eval_runtime": 3383.9959, | |
| "eval_samples_per_second": 49.632, | |
| "eval_steps_per_second": 6.204, | |
| "step": 141750 | |
| }, | |
| { | |
| "epoch": 54.1, | |
| "learning_rate": 0.0022952380952380954, | |
| "loss": 0.5328, | |
| "step": 142000 | |
| }, | |
| { | |
| "epoch": 54.29, | |
| "learning_rate": 0.002285714285714286, | |
| "loss": 0.5141, | |
| "step": 142500 | |
| }, | |
| { | |
| "epoch": 54.48, | |
| "learning_rate": 0.0022761904761904764, | |
| "loss": 0.5266, | |
| "step": 143000 | |
| }, | |
| { | |
| "epoch": 54.67, | |
| "learning_rate": 0.0022666666666666664, | |
| "loss": 0.5354, | |
| "step": 143500 | |
| }, | |
| { | |
| "epoch": 54.86, | |
| "learning_rate": 0.002257142857142857, | |
| "loss": 0.556, | |
| "step": 144000 | |
| }, | |
| { | |
| "epoch": 55.0, | |
| "eval_accuracy": 0.6826173677473133, | |
| "eval_loss": 2.0039005279541016, | |
| "eval_runtime": 2747.9149, | |
| "eval_samples_per_second": 61.121, | |
| "eval_steps_per_second": 7.64, | |
| "step": 144375 | |
| }, | |
| { | |
| "epoch": 55.05, | |
| "learning_rate": 0.0022476190476190478, | |
| "loss": 0.5317, | |
| "step": 144500 | |
| }, | |
| { | |
| "epoch": 55.24, | |
| "learning_rate": 0.0022380952380952382, | |
| "loss": 0.4956, | |
| "step": 145000 | |
| }, | |
| { | |
| "epoch": 55.43, | |
| "learning_rate": 0.0022285714285714287, | |
| "loss": 0.512, | |
| "step": 145500 | |
| }, | |
| { | |
| "epoch": 55.62, | |
| "learning_rate": 0.002219047619047619, | |
| "loss": 0.5243, | |
| "step": 146000 | |
| }, | |
| { | |
| "epoch": 55.81, | |
| "learning_rate": 0.0022095238095238097, | |
| "loss": 0.5283, | |
| "step": 146500 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "learning_rate": 0.0022, | |
| "loss": 0.5412, | |
| "step": 147000 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "eval_accuracy": 0.6856896192432497, | |
| "eval_loss": 1.9339253902435303, | |
| "eval_runtime": 3227.3643, | |
| "eval_samples_per_second": 52.041, | |
| "eval_steps_per_second": 6.505, | |
| "step": 147000 | |
| }, | |
| { | |
| "epoch": 56.19, | |
| "learning_rate": 0.0021904761904761906, | |
| "loss": 0.4902, | |
| "step": 147500 | |
| }, | |
| { | |
| "epoch": 56.38, | |
| "learning_rate": 0.002180952380952381, | |
| "loss": 0.4989, | |
| "step": 148000 | |
| }, | |
| { | |
| "epoch": 56.57, | |
| "learning_rate": 0.0021714285714285715, | |
| "loss": 0.4979, | |
| "step": 148500 | |
| }, | |
| { | |
| "epoch": 56.76, | |
| "learning_rate": 0.002161904761904762, | |
| "loss": 0.5194, | |
| "step": 149000 | |
| }, | |
| { | |
| "epoch": 56.95, | |
| "learning_rate": 0.0021523809523809525, | |
| "loss": 0.5204, | |
| "step": 149500 | |
| }, | |
| { | |
| "epoch": 57.0, | |
| "eval_accuracy": 0.6872436069185198, | |
| "eval_loss": 2.0444183349609375, | |
| "eval_runtime": 2731.7184, | |
| "eval_samples_per_second": 61.483, | |
| "eval_steps_per_second": 7.686, | |
| "step": 149625 | |
| }, | |
| { | |
| "epoch": 57.14, | |
| "learning_rate": 0.002142857142857143, | |
| "loss": 0.4842, | |
| "step": 150000 | |
| }, | |
| { | |
| "epoch": 57.33, | |
| "learning_rate": 0.0021333333333333334, | |
| "loss": 0.4815, | |
| "step": 150500 | |
| }, | |
| { | |
| "epoch": 57.52, | |
| "learning_rate": 0.002123809523809524, | |
| "loss": 0.4911, | |
| "step": 151000 | |
| }, | |
| { | |
| "epoch": 57.71, | |
| "learning_rate": 0.0021142857142857144, | |
| "loss": 0.4954, | |
| "step": 151500 | |
| }, | |
| { | |
| "epoch": 57.9, | |
| "learning_rate": 0.002104761904761905, | |
| "loss": 0.5051, | |
| "step": 152000 | |
| }, | |
| { | |
| "epoch": 58.0, | |
| "eval_accuracy": 0.689601381322378, | |
| "eval_loss": 1.9677560329437256, | |
| "eval_runtime": 2701.9047, | |
| "eval_samples_per_second": 62.162, | |
| "eval_steps_per_second": 7.77, | |
| "step": 152250 | |
| }, | |
| { | |
| "epoch": 58.1, | |
| "learning_rate": 0.0020952380952380953, | |
| "loss": 0.4901, | |
| "step": 152500 | |
| }, | |
| { | |
| "epoch": 58.29, | |
| "learning_rate": 0.002085714285714286, | |
| "loss": 0.4763, | |
| "step": 153000 | |
| }, | |
| { | |
| "epoch": 58.48, | |
| "learning_rate": 0.0020761904761904763, | |
| "loss": 0.4806, | |
| "step": 153500 | |
| }, | |
| { | |
| "epoch": 58.67, | |
| "learning_rate": 0.0020666666666666667, | |
| "loss": 0.4858, | |
| "step": 154000 | |
| }, | |
| { | |
| "epoch": 58.86, | |
| "learning_rate": 0.002057142857142857, | |
| "loss": 0.4977, | |
| "step": 154500 | |
| }, | |
| { | |
| "epoch": 59.0, | |
| "eval_accuracy": 0.6877735107618111, | |
| "eval_loss": 2.0384891033172607, | |
| "eval_runtime": 2728.0396, | |
| "eval_samples_per_second": 61.566, | |
| "eval_steps_per_second": 7.696, | |
| "step": 154875 | |
| }, | |
| { | |
| "epoch": 59.05, | |
| "learning_rate": 0.0020476190476190477, | |
| "loss": 0.483, | |
| "step": 155000 | |
| }, | |
| { | |
| "epoch": 59.24, | |
| "learning_rate": 0.002038095238095238, | |
| "loss": 0.4575, | |
| "step": 155500 | |
| }, | |
| { | |
| "epoch": 59.43, | |
| "learning_rate": 0.0020285714285714286, | |
| "loss": 0.4705, | |
| "step": 156000 | |
| }, | |
| { | |
| "epoch": 59.62, | |
| "learning_rate": 0.002019047619047619, | |
| "loss": 0.4846, | |
| "step": 156500 | |
| }, | |
| { | |
| "epoch": 59.81, | |
| "learning_rate": 0.0020095238095238096, | |
| "loss": 0.4872, | |
| "step": 157000 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "learning_rate": 0.002, | |
| "loss": 0.4932, | |
| "step": 157500 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "eval_accuracy": 0.6890178916971808, | |
| "eval_loss": 2.05938720703125, | |
| "eval_runtime": 2847.0911, | |
| "eval_samples_per_second": 58.992, | |
| "eval_steps_per_second": 7.374, | |
| "step": 157500 | |
| }, | |
| { | |
| "epoch": 60.19, | |
| "learning_rate": 0.0019904761904761905, | |
| "loss": 0.4454, | |
| "step": 158000 | |
| }, | |
| { | |
| "epoch": 60.38, | |
| "learning_rate": 0.001980952380952381, | |
| "loss": 0.4559, | |
| "step": 158500 | |
| }, | |
| { | |
| "epoch": 60.57, | |
| "learning_rate": 0.0019714285714285715, | |
| "loss": 0.46, | |
| "step": 159000 | |
| }, | |
| { | |
| "epoch": 60.76, | |
| "learning_rate": 0.001961904761904762, | |
| "loss": 0.472, | |
| "step": 159500 | |
| }, | |
| { | |
| "epoch": 60.95, | |
| "learning_rate": 0.0019523809523809524, | |
| "loss": 0.4689, | |
| "step": 160000 | |
| }, | |
| { | |
| "epoch": 61.0, | |
| "eval_accuracy": 0.6834806942335745, | |
| "eval_loss": 2.1333200931549072, | |
| "eval_runtime": 3489.1293, | |
| "eval_samples_per_second": 48.137, | |
| "eval_steps_per_second": 6.017, | |
| "step": 160125 | |
| }, | |
| { | |
| "epoch": 61.14, | |
| "learning_rate": 0.0019428571428571429, | |
| "loss": 0.4402, | |
| "step": 160500 | |
| }, | |
| { | |
| "epoch": 61.33, | |
| "learning_rate": 0.0019333333333333333, | |
| "loss": 0.445, | |
| "step": 161000 | |
| }, | |
| { | |
| "epoch": 61.52, | |
| "learning_rate": 0.0019238095238095238, | |
| "loss": 0.4504, | |
| "step": 161500 | |
| }, | |
| { | |
| "epoch": 61.71, | |
| "learning_rate": 0.0019142857142857143, | |
| "loss": 0.4577, | |
| "step": 162000 | |
| }, | |
| { | |
| "epoch": 61.9, | |
| "learning_rate": 0.0019047619047619048, | |
| "loss": 0.4652, | |
| "step": 162500 | |
| }, | |
| { | |
| "epoch": 62.0, | |
| "eval_accuracy": 0.6899228960138132, | |
| "eval_loss": 2.1436057090759277, | |
| "eval_runtime": 2832.828, | |
| "eval_samples_per_second": 59.289, | |
| "eval_steps_per_second": 7.411, | |
| "step": 162750 | |
| }, | |
| { | |
| "epoch": 62.1, | |
| "learning_rate": 0.0018952380952380952, | |
| "loss": 0.4362, | |
| "step": 163000 | |
| }, | |
| { | |
| "epoch": 62.29, | |
| "learning_rate": 0.001885714285714286, | |
| "loss": 0.4351, | |
| "step": 163500 | |
| }, | |
| { | |
| "epoch": 62.48, | |
| "learning_rate": 0.0018761904761904764, | |
| "loss": 0.4404, | |
| "step": 164000 | |
| }, | |
| { | |
| "epoch": 62.67, | |
| "learning_rate": 0.0018666666666666669, | |
| "loss": 0.4443, | |
| "step": 164500 | |
| }, | |
| { | |
| "epoch": 62.86, | |
| "learning_rate": 0.0018571428571428573, | |
| "loss": 0.4515, | |
| "step": 165000 | |
| }, | |
| { | |
| "epoch": 63.0, | |
| "eval_accuracy": 0.6923283022238099, | |
| "eval_loss": 2.1558289527893066, | |
| "eval_runtime": 2793.0236, | |
| "eval_samples_per_second": 60.134, | |
| "eval_steps_per_second": 7.517, | |
| "step": 165375 | |
| }, | |
| { | |
| "epoch": 63.05, | |
| "learning_rate": 0.0018476190476190476, | |
| "loss": 0.4448, | |
| "step": 165500 | |
| }, | |
| { | |
| "epoch": 63.24, | |
| "learning_rate": 0.001838095238095238, | |
| "loss": 0.4217, | |
| "step": 166000 | |
| }, | |
| { | |
| "epoch": 63.43, | |
| "learning_rate": 0.0018285714285714285, | |
| "loss": 0.4293, | |
| "step": 166500 | |
| }, | |
| { | |
| "epoch": 63.62, | |
| "learning_rate": 0.001819047619047619, | |
| "loss": 0.4295, | |
| "step": 167000 | |
| }, | |
| { | |
| "epoch": 63.81, | |
| "learning_rate": 0.0018095238095238095, | |
| "loss": 0.4383, | |
| "step": 167500 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "learning_rate": 0.0018, | |
| "loss": 0.4542, | |
| "step": 168000 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "eval_accuracy": 0.6934059718376946, | |
| "eval_loss": 2.0637614727020264, | |
| "eval_runtime": 2998.8188, | |
| "eval_samples_per_second": 56.007, | |
| "eval_steps_per_second": 7.001, | |
| "step": 168000 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "step": 168000, | |
| "total_flos": 4.5354579076109147e+21, | |
| "train_loss": 0.0, | |
| "train_runtime": 758.3182, | |
| "train_samples_per_second": 53155.811, | |
| "train_steps_per_second": 207.696 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 157500, | |
| "num_train_epochs": 60, | |
| "save_steps": 500, | |
| "total_flos": 4.5354579076109147e+21, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |