| { | |
| "best_global_step": 40000, | |
| "best_metric": 0.198301300406456, | |
| "best_model_checkpoint": "/media/user/Expansion1/multilingual-e5-small-aligned-v2-fineweb2hq-vs-c4-classifier-run2/checkpoint-40000", | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 120000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0125, | |
| "grad_norm": 5.874658584594727, | |
| "learning_rate": 4.979208333333333e-05, | |
| "loss": 0.3436, | |
| "num_input_tokens_seen": 512000, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.025, | |
| "grad_norm": 0.6375065445899963, | |
| "learning_rate": 4.958375e-05, | |
| "loss": 0.281, | |
| "num_input_tokens_seen": 1024000, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.0375, | |
| "grad_norm": 0.48116546869277954, | |
| "learning_rate": 4.937541666666667e-05, | |
| "loss": 0.2595, | |
| "num_input_tokens_seen": 1536000, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.6305193305015564, | |
| "learning_rate": 4.9167083333333336e-05, | |
| "loss": 0.2466, | |
| "num_input_tokens_seen": 2048000, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.0625, | |
| "grad_norm": 3.560100555419922, | |
| "learning_rate": 4.8958750000000006e-05, | |
| "loss": 0.2283, | |
| "num_input_tokens_seen": 2560000, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.075, | |
| "grad_norm": 43.007347106933594, | |
| "learning_rate": 4.875041666666667e-05, | |
| "loss": 0.2106, | |
| "num_input_tokens_seen": 3072000, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.0875, | |
| "grad_norm": 27.75348472595215, | |
| "learning_rate": 4.854208333333334e-05, | |
| "loss": 0.2198, | |
| "num_input_tokens_seen": 3584000, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.11089582741260529, | |
| "learning_rate": 4.833375e-05, | |
| "loss": 0.2145, | |
| "num_input_tokens_seen": 4096000, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.1125, | |
| "grad_norm": 0.08869020640850067, | |
| "learning_rate": 4.812541666666667e-05, | |
| "loss": 0.2033, | |
| "num_input_tokens_seen": 4608000, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.125, | |
| "grad_norm": 22.97330665588379, | |
| "learning_rate": 4.791708333333333e-05, | |
| "loss": 0.2007, | |
| "num_input_tokens_seen": 5120000, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.1375, | |
| "grad_norm": 0.12722039222717285, | |
| "learning_rate": 4.770875e-05, | |
| "loss": 0.1893, | |
| "num_input_tokens_seen": 5632000, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 24.923276901245117, | |
| "learning_rate": 4.750041666666667e-05, | |
| "loss": 0.2073, | |
| "num_input_tokens_seen": 6144000, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.1625, | |
| "grad_norm": 0.17447435855865479, | |
| "learning_rate": 4.729208333333334e-05, | |
| "loss": 0.2024, | |
| "num_input_tokens_seen": 6656000, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.175, | |
| "grad_norm": 5.2057671546936035, | |
| "learning_rate": 4.708375e-05, | |
| "loss": 0.2168, | |
| "num_input_tokens_seen": 7168000, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.1875, | |
| "grad_norm": 0.35154542326927185, | |
| "learning_rate": 4.687541666666667e-05, | |
| "loss": 0.2002, | |
| "num_input_tokens_seen": 7680000, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.17855411767959595, | |
| "learning_rate": 4.6667083333333336e-05, | |
| "loss": 0.1863, | |
| "num_input_tokens_seen": 8192000, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.2125, | |
| "grad_norm": 0.26302626729011536, | |
| "learning_rate": 4.645875e-05, | |
| "loss": 0.2023, | |
| "num_input_tokens_seen": 8704000, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.225, | |
| "grad_norm": 1.0355998277664185, | |
| "learning_rate": 4.625041666666667e-05, | |
| "loss": 0.1996, | |
| "num_input_tokens_seen": 9216000, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.2375, | |
| "grad_norm": 3.5437514781951904, | |
| "learning_rate": 4.6042083333333334e-05, | |
| "loss": 0.1928, | |
| "num_input_tokens_seen": 9728000, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 18.394441604614258, | |
| "learning_rate": 4.5833750000000005e-05, | |
| "loss": 0.1968, | |
| "num_input_tokens_seen": 10240000, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.2625, | |
| "grad_norm": 2.9275922775268555, | |
| "learning_rate": 4.562541666666667e-05, | |
| "loss": 0.2081, | |
| "num_input_tokens_seen": 10752000, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.275, | |
| "grad_norm": 3.922402858734131, | |
| "learning_rate": 4.541708333333334e-05, | |
| "loss": 0.1769, | |
| "num_input_tokens_seen": 11264000, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.2875, | |
| "grad_norm": 0.25219565629959106, | |
| "learning_rate": 4.5208749999999996e-05, | |
| "loss": 0.1722, | |
| "num_input_tokens_seen": 11776000, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.10917416214942932, | |
| "learning_rate": 4.500041666666667e-05, | |
| "loss": 0.1638, | |
| "num_input_tokens_seen": 12288000, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.3125, | |
| "grad_norm": 0.1928466558456421, | |
| "learning_rate": 4.479208333333334e-05, | |
| "loss": 0.1826, | |
| "num_input_tokens_seen": 12800000, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.325, | |
| "grad_norm": 1.1688095331192017, | |
| "learning_rate": 4.458375e-05, | |
| "loss": 0.186, | |
| "num_input_tokens_seen": 13312000, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.3375, | |
| "grad_norm": 0.07879551500082016, | |
| "learning_rate": 4.437541666666667e-05, | |
| "loss": 0.1638, | |
| "num_input_tokens_seen": 13824000, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 1.294481635093689, | |
| "learning_rate": 4.4167083333333336e-05, | |
| "loss": 0.1695, | |
| "num_input_tokens_seen": 14336000, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.3625, | |
| "grad_norm": 21.836345672607422, | |
| "learning_rate": 4.395875000000001e-05, | |
| "loss": 0.1713, | |
| "num_input_tokens_seen": 14848000, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.375, | |
| "grad_norm": 0.1329026073217392, | |
| "learning_rate": 4.375041666666667e-05, | |
| "loss": 0.178, | |
| "num_input_tokens_seen": 15360000, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.3875, | |
| "grad_norm": 2.196962594985962, | |
| "learning_rate": 4.3542083333333334e-05, | |
| "loss": 0.1616, | |
| "num_input_tokens_seen": 15872000, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.06707575172185898, | |
| "learning_rate": 4.333375e-05, | |
| "loss": 0.1563, | |
| "num_input_tokens_seen": 16384000, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.4125, | |
| "grad_norm": 4.03335428237915, | |
| "learning_rate": 4.312541666666667e-05, | |
| "loss": 0.1644, | |
| "num_input_tokens_seen": 16896000, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.425, | |
| "grad_norm": 0.7723323702812195, | |
| "learning_rate": 4.291708333333333e-05, | |
| "loss": 0.1695, | |
| "num_input_tokens_seen": 17408000, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.4375, | |
| "grad_norm": 0.14005227386951447, | |
| "learning_rate": 4.270875e-05, | |
| "loss": 0.1771, | |
| "num_input_tokens_seen": 17920000, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.22879253327846527, | |
| "learning_rate": 4.250041666666667e-05, | |
| "loss": 0.1652, | |
| "num_input_tokens_seen": 18432000, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.4625, | |
| "grad_norm": 41.407352447509766, | |
| "learning_rate": 4.229208333333334e-05, | |
| "loss": 0.1697, | |
| "num_input_tokens_seen": 18944000, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.475, | |
| "grad_norm": 0.2552178204059601, | |
| "learning_rate": 4.208375e-05, | |
| "loss": 0.169, | |
| "num_input_tokens_seen": 19456000, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.4875, | |
| "grad_norm": 4.208723545074463, | |
| "learning_rate": 4.1875416666666666e-05, | |
| "loss": 0.1887, | |
| "num_input_tokens_seen": 19968000, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.6019179821014404, | |
| "learning_rate": 4.1667083333333336e-05, | |
| "loss": 0.149, | |
| "num_input_tokens_seen": 20480000, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.5125, | |
| "grad_norm": 11.652756690979004, | |
| "learning_rate": 4.145875e-05, | |
| "loss": 0.1485, | |
| "num_input_tokens_seen": 20992000, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.525, | |
| "grad_norm": 3.790767192840576, | |
| "learning_rate": 4.125041666666667e-05, | |
| "loss": 0.1613, | |
| "num_input_tokens_seen": 21504000, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.5375, | |
| "grad_norm": 0.0643579512834549, | |
| "learning_rate": 4.1042083333333335e-05, | |
| "loss": 0.156, | |
| "num_input_tokens_seen": 22016000, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 6.347803115844727, | |
| "learning_rate": 4.0833750000000005e-05, | |
| "loss": 0.1743, | |
| "num_input_tokens_seen": 22528000, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.5625, | |
| "grad_norm": 0.45395129919052124, | |
| "learning_rate": 4.062541666666667e-05, | |
| "loss": 0.1566, | |
| "num_input_tokens_seen": 23040000, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.575, | |
| "grad_norm": 4.8587470054626465, | |
| "learning_rate": 4.041708333333333e-05, | |
| "loss": 0.156, | |
| "num_input_tokens_seen": 23552000, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.5875, | |
| "grad_norm": 0.8771836757659912, | |
| "learning_rate": 4.0208750000000004e-05, | |
| "loss": 0.1473, | |
| "num_input_tokens_seen": 24064000, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 7.632213592529297, | |
| "learning_rate": 4.000041666666667e-05, | |
| "loss": 0.1523, | |
| "num_input_tokens_seen": 24576000, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.6125, | |
| "grad_norm": 4.013466835021973, | |
| "learning_rate": 3.979208333333334e-05, | |
| "loss": 0.1646, | |
| "num_input_tokens_seen": 25088000, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.625, | |
| "grad_norm": 0.08471076935529709, | |
| "learning_rate": 3.958375e-05, | |
| "loss": 0.1653, | |
| "num_input_tokens_seen": 25600000, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.6375, | |
| "grad_norm": 4.263392448425293, | |
| "learning_rate": 3.937541666666667e-05, | |
| "loss": 0.1513, | |
| "num_input_tokens_seen": 26112000, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 17.002439498901367, | |
| "learning_rate": 3.9167083333333336e-05, | |
| "loss": 0.144, | |
| "num_input_tokens_seen": 26624000, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.6625, | |
| "grad_norm": 0.056964244693517685, | |
| "learning_rate": 3.895875e-05, | |
| "loss": 0.1517, | |
| "num_input_tokens_seen": 27136000, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.675, | |
| "grad_norm": 0.4403606653213501, | |
| "learning_rate": 3.8750416666666664e-05, | |
| "loss": 0.1501, | |
| "num_input_tokens_seen": 27648000, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.6875, | |
| "grad_norm": 0.11431742459535599, | |
| "learning_rate": 3.8542083333333335e-05, | |
| "loss": 0.1597, | |
| "num_input_tokens_seen": 28160000, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 7.810299396514893, | |
| "learning_rate": 3.833375e-05, | |
| "loss": 0.1437, | |
| "num_input_tokens_seen": 28672000, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.7125, | |
| "grad_norm": 5.501016616821289, | |
| "learning_rate": 3.812541666666667e-05, | |
| "loss": 0.1415, | |
| "num_input_tokens_seen": 29184000, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.725, | |
| "grad_norm": 0.20471607148647308, | |
| "learning_rate": 3.791708333333333e-05, | |
| "loss": 0.1545, | |
| "num_input_tokens_seen": 29696000, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.7375, | |
| "grad_norm": 4.6759748458862305, | |
| "learning_rate": 3.7708750000000004e-05, | |
| "loss": 0.1431, | |
| "num_input_tokens_seen": 30208000, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.12430200725793839, | |
| "learning_rate": 3.7500416666666674e-05, | |
| "loss": 0.1337, | |
| "num_input_tokens_seen": 30720000, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.7625, | |
| "grad_norm": 0.19309231638908386, | |
| "learning_rate": 3.729208333333333e-05, | |
| "loss": 0.1338, | |
| "num_input_tokens_seen": 31232000, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.775, | |
| "grad_norm": 0.18272100389003754, | |
| "learning_rate": 3.708375e-05, | |
| "loss": 0.1504, | |
| "num_input_tokens_seen": 31744000, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.7875, | |
| "grad_norm": 0.08077079057693481, | |
| "learning_rate": 3.6875416666666666e-05, | |
| "loss": 0.13, | |
| "num_input_tokens_seen": 32256000, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.3213728368282318, | |
| "learning_rate": 3.666708333333334e-05, | |
| "loss": 0.1406, | |
| "num_input_tokens_seen": 32768000, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.8125, | |
| "grad_norm": 2.6233737468719482, | |
| "learning_rate": 3.645875e-05, | |
| "loss": 0.1435, | |
| "num_input_tokens_seen": 33280000, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.825, | |
| "grad_norm": 0.1083766371011734, | |
| "learning_rate": 3.625041666666667e-05, | |
| "loss": 0.1361, | |
| "num_input_tokens_seen": 33792000, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.8375, | |
| "grad_norm": 0.08282948285341263, | |
| "learning_rate": 3.6042083333333335e-05, | |
| "loss": 0.1483, | |
| "num_input_tokens_seen": 34304000, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.10612857341766357, | |
| "learning_rate": 3.583375e-05, | |
| "loss": 0.1419, | |
| "num_input_tokens_seen": 34816000, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 0.8625, | |
| "grad_norm": 1.134366750717163, | |
| "learning_rate": 3.562541666666667e-05, | |
| "loss": 0.1381, | |
| "num_input_tokens_seen": 35328000, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 0.875, | |
| "grad_norm": 32.71923065185547, | |
| "learning_rate": 3.5417083333333333e-05, | |
| "loss": 0.1395, | |
| "num_input_tokens_seen": 35840000, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 0.8875, | |
| "grad_norm": 0.14211738109588623, | |
| "learning_rate": 3.5208750000000004e-05, | |
| "loss": 0.1543, | |
| "num_input_tokens_seen": 36352000, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.04908756539225578, | |
| "learning_rate": 3.500041666666667e-05, | |
| "loss": 0.1229, | |
| "num_input_tokens_seen": 36864000, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 0.9125, | |
| "grad_norm": 0.47926560044288635, | |
| "learning_rate": 3.479208333333334e-05, | |
| "loss": 0.1251, | |
| "num_input_tokens_seen": 37376000, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 0.925, | |
| "grad_norm": 1.389073371887207, | |
| "learning_rate": 3.458375e-05, | |
| "loss": 0.1718, | |
| "num_input_tokens_seen": 37888000, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 0.9375, | |
| "grad_norm": 0.14962467551231384, | |
| "learning_rate": 3.437541666666667e-05, | |
| "loss": 0.1579, | |
| "num_input_tokens_seen": 38400000, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 3.8970487117767334, | |
| "learning_rate": 3.416708333333333e-05, | |
| "loss": 0.1294, | |
| "num_input_tokens_seen": 38912000, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 0.9625, | |
| "grad_norm": 0.1765402853488922, | |
| "learning_rate": 3.395875e-05, | |
| "loss": 0.1612, | |
| "num_input_tokens_seen": 39424000, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 0.975, | |
| "grad_norm": 0.1407538652420044, | |
| "learning_rate": 3.3750416666666665e-05, | |
| "loss": 0.1336, | |
| "num_input_tokens_seen": 39936000, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 0.9875, | |
| "grad_norm": 8.721837997436523, | |
| "learning_rate": 3.3542083333333335e-05, | |
| "loss": 0.1268, | |
| "num_input_tokens_seen": 40448000, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.15711814165115356, | |
| "learning_rate": 3.333375e-05, | |
| "loss": 0.1387, | |
| "num_input_tokens_seen": 40960000, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.951525, | |
| "eval_combined_score": 1.3493705490187269, | |
| "eval_loss": 0.198301300406456, | |
| "eval_runtime": 39.36, | |
| "eval_samples_per_second": 2032.519, | |
| "eval_steps_per_second": 254.065, | |
| "num_input_tokens_seen": 40960000, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 1.0125, | |
| "grad_norm": 0.02203565090894699, | |
| "learning_rate": 3.312541666666667e-05, | |
| "loss": 0.1092, | |
| "num_input_tokens_seen": 41472000, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 1.025, | |
| "grad_norm": 328.5529479980469, | |
| "learning_rate": 3.291708333333334e-05, | |
| "loss": 0.0844, | |
| "num_input_tokens_seen": 41984000, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 1.0375, | |
| "grad_norm": 0.11391662806272507, | |
| "learning_rate": 3.270875e-05, | |
| "loss": 0.1115, | |
| "num_input_tokens_seen": 42496000, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "grad_norm": 0.07919144630432129, | |
| "learning_rate": 3.250041666666667e-05, | |
| "loss": 0.0878, | |
| "num_input_tokens_seen": 43008000, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 1.0625, | |
| "grad_norm": 0.029733452945947647, | |
| "learning_rate": 3.229208333333333e-05, | |
| "loss": 0.0914, | |
| "num_input_tokens_seen": 43520000, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 1.075, | |
| "grad_norm": 0.277174711227417, | |
| "learning_rate": 3.208375e-05, | |
| "loss": 0.1028, | |
| "num_input_tokens_seen": 44032000, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 1.0875, | |
| "grad_norm": 0.11592718958854675, | |
| "learning_rate": 3.1875416666666666e-05, | |
| "loss": 0.1032, | |
| "num_input_tokens_seen": 44544000, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "grad_norm": 0.2565874755382538, | |
| "learning_rate": 3.166708333333334e-05, | |
| "loss": 0.0912, | |
| "num_input_tokens_seen": 45056000, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 1.1125, | |
| "grad_norm": 0.03557795658707619, | |
| "learning_rate": 3.145875e-05, | |
| "loss": 0.0882, | |
| "num_input_tokens_seen": 45568000, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 1.125, | |
| "grad_norm": 4.72824764251709, | |
| "learning_rate": 3.125041666666667e-05, | |
| "loss": 0.0918, | |
| "num_input_tokens_seen": 46080000, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 1.1375, | |
| "grad_norm": 0.10075237601995468, | |
| "learning_rate": 3.1042083333333335e-05, | |
| "loss": 0.1117, | |
| "num_input_tokens_seen": 46592000, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "grad_norm": 0.07608671486377716, | |
| "learning_rate": 3.083375e-05, | |
| "loss": 0.1383, | |
| "num_input_tokens_seen": 47104000, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 1.1625, | |
| "grad_norm": 14.187911033630371, | |
| "learning_rate": 3.062541666666667e-05, | |
| "loss": 0.1062, | |
| "num_input_tokens_seen": 47616000, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 1.175, | |
| "grad_norm": 0.0889461487531662, | |
| "learning_rate": 3.0417083333333334e-05, | |
| "loss": 0.1255, | |
| "num_input_tokens_seen": 48128000, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 1.1875, | |
| "grad_norm": 0.10722101479768753, | |
| "learning_rate": 3.020875e-05, | |
| "loss": 0.0987, | |
| "num_input_tokens_seen": 48640000, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 42.01722717285156, | |
| "learning_rate": 3.000041666666667e-05, | |
| "loss": 0.1047, | |
| "num_input_tokens_seen": 49152000, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 1.2125, | |
| "grad_norm": 0.08817047625780106, | |
| "learning_rate": 2.9792083333333336e-05, | |
| "loss": 0.0931, | |
| "num_input_tokens_seen": 49664000, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 1.225, | |
| "grad_norm": 0.058988332748413086, | |
| "learning_rate": 2.958375e-05, | |
| "loss": 0.1048, | |
| "num_input_tokens_seen": 50176000, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 1.2375, | |
| "grad_norm": 0.025500474497675896, | |
| "learning_rate": 2.9375416666666667e-05, | |
| "loss": 0.0845, | |
| "num_input_tokens_seen": 50688000, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 0.07621040940284729, | |
| "learning_rate": 2.9167083333333334e-05, | |
| "loss": 0.0893, | |
| "num_input_tokens_seen": 51200000, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 1.2625, | |
| "grad_norm": 0.14059029519557953, | |
| "learning_rate": 2.895875e-05, | |
| "loss": 0.0911, | |
| "num_input_tokens_seen": 51712000, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 1.275, | |
| "grad_norm": 0.04900716617703438, | |
| "learning_rate": 2.875041666666667e-05, | |
| "loss": 0.0984, | |
| "num_input_tokens_seen": 52224000, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 1.2875, | |
| "grad_norm": 0.0568259134888649, | |
| "learning_rate": 2.8542083333333336e-05, | |
| "loss": 0.0995, | |
| "num_input_tokens_seen": 52736000, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "grad_norm": 0.052453652024269104, | |
| "learning_rate": 2.8333750000000003e-05, | |
| "loss": 0.0875, | |
| "num_input_tokens_seen": 53248000, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 1.3125, | |
| "grad_norm": 0.05766982212662697, | |
| "learning_rate": 2.812541666666667e-05, | |
| "loss": 0.0772, | |
| "num_input_tokens_seen": 53760000, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 1.325, | |
| "grad_norm": 0.06809753179550171, | |
| "learning_rate": 2.7917083333333334e-05, | |
| "loss": 0.0849, | |
| "num_input_tokens_seen": 54272000, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 1.3375, | |
| "grad_norm": 0.5180730223655701, | |
| "learning_rate": 2.770875e-05, | |
| "loss": 0.089, | |
| "num_input_tokens_seen": 54784000, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "grad_norm": 12.439111709594727, | |
| "learning_rate": 2.750041666666667e-05, | |
| "loss": 0.0927, | |
| "num_input_tokens_seen": 55296000, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 1.3625, | |
| "grad_norm": 0.12473966181278229, | |
| "learning_rate": 2.7292083333333336e-05, | |
| "loss": 0.1109, | |
| "num_input_tokens_seen": 55808000, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 1.375, | |
| "grad_norm": 1.5988309383392334, | |
| "learning_rate": 2.7083750000000003e-05, | |
| "loss": 0.0962, | |
| "num_input_tokens_seen": 56320000, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 1.3875, | |
| "grad_norm": 0.09568974375724792, | |
| "learning_rate": 2.687541666666667e-05, | |
| "loss": 0.1041, | |
| "num_input_tokens_seen": 56832000, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "grad_norm": 105.98346710205078, | |
| "learning_rate": 2.6667083333333338e-05, | |
| "loss": 0.0895, | |
| "num_input_tokens_seen": 57344000, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 1.4125, | |
| "grad_norm": 0.04284173250198364, | |
| "learning_rate": 2.6458749999999998e-05, | |
| "loss": 0.09, | |
| "num_input_tokens_seen": 57856000, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 1.425, | |
| "grad_norm": 0.06827585399150848, | |
| "learning_rate": 2.6250416666666665e-05, | |
| "loss": 0.1064, | |
| "num_input_tokens_seen": 58368000, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 1.4375, | |
| "grad_norm": 0.10431079566478729, | |
| "learning_rate": 2.6042083333333333e-05, | |
| "loss": 0.1033, | |
| "num_input_tokens_seen": 58880000, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "grad_norm": 0.08796288073062897, | |
| "learning_rate": 2.583375e-05, | |
| "loss": 0.0781, | |
| "num_input_tokens_seen": 59392000, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 1.4625, | |
| "grad_norm": 1.883025884628296, | |
| "learning_rate": 2.5625416666666667e-05, | |
| "loss": 0.1001, | |
| "num_input_tokens_seen": 59904000, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 1.475, | |
| "grad_norm": 0.2901429533958435, | |
| "learning_rate": 2.5417083333333334e-05, | |
| "loss": 0.0965, | |
| "num_input_tokens_seen": 60416000, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 1.4875, | |
| "grad_norm": 0.05163797363638878, | |
| "learning_rate": 2.520875e-05, | |
| "loss": 0.1064, | |
| "num_input_tokens_seen": 60928000, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 0.03811231628060341, | |
| "learning_rate": 2.5000416666666672e-05, | |
| "loss": 0.0865, | |
| "num_input_tokens_seen": 61440000, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 1.5125, | |
| "grad_norm": 8.308381080627441, | |
| "learning_rate": 2.4792083333333336e-05, | |
| "loss": 0.093, | |
| "num_input_tokens_seen": 61952000, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 1.525, | |
| "grad_norm": 0.03729177638888359, | |
| "learning_rate": 2.458375e-05, | |
| "loss": 0.077, | |
| "num_input_tokens_seen": 62464000, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 1.5375, | |
| "grad_norm": 5.803088188171387, | |
| "learning_rate": 2.4375416666666667e-05, | |
| "loss": 0.101, | |
| "num_input_tokens_seen": 62976000, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "grad_norm": 2.017042636871338, | |
| "learning_rate": 2.4167083333333334e-05, | |
| "loss": 0.0896, | |
| "num_input_tokens_seen": 63488000, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 1.5625, | |
| "grad_norm": 32.310630798339844, | |
| "learning_rate": 2.395875e-05, | |
| "loss": 0.1145, | |
| "num_input_tokens_seen": 64000000, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 1.575, | |
| "grad_norm": 0.37863266468048096, | |
| "learning_rate": 2.3750416666666665e-05, | |
| "loss": 0.1017, | |
| "num_input_tokens_seen": 64512000, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 1.5875, | |
| "grad_norm": 0.05939273163676262, | |
| "learning_rate": 2.3542083333333333e-05, | |
| "loss": 0.0962, | |
| "num_input_tokens_seen": 65024000, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 0.045398563146591187, | |
| "learning_rate": 2.333375e-05, | |
| "loss": 0.0895, | |
| "num_input_tokens_seen": 65536000, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 1.6125, | |
| "grad_norm": 1.5717942714691162, | |
| "learning_rate": 2.3125416666666667e-05, | |
| "loss": 0.0788, | |
| "num_input_tokens_seen": 66048000, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 1.625, | |
| "grad_norm": 6.278552532196045, | |
| "learning_rate": 2.2917083333333334e-05, | |
| "loss": 0.0876, | |
| "num_input_tokens_seen": 66560000, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 1.6375, | |
| "grad_norm": 0.048090457916259766, | |
| "learning_rate": 2.2708750000000002e-05, | |
| "loss": 0.0709, | |
| "num_input_tokens_seen": 67072000, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "grad_norm": 0.5121225714683533, | |
| "learning_rate": 2.250041666666667e-05, | |
| "loss": 0.0906, | |
| "num_input_tokens_seen": 67584000, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 1.6625, | |
| "grad_norm": 0.04399065673351288, | |
| "learning_rate": 2.2292083333333336e-05, | |
| "loss": 0.097, | |
| "num_input_tokens_seen": 68096000, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 1.675, | |
| "grad_norm": 0.051211412996053696, | |
| "learning_rate": 2.208375e-05, | |
| "loss": 0.0931, | |
| "num_input_tokens_seen": 68608000, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 1.6875, | |
| "grad_norm": 58.19650650024414, | |
| "learning_rate": 2.1875416666666667e-05, | |
| "loss": 0.0794, | |
| "num_input_tokens_seen": 69120000, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "grad_norm": 0.07303386926651001, | |
| "learning_rate": 2.1667083333333335e-05, | |
| "loss": 0.1015, | |
| "num_input_tokens_seen": 69632000, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 1.7125, | |
| "grad_norm": 0.02853875607252121, | |
| "learning_rate": 2.145875e-05, | |
| "loss": 0.0684, | |
| "num_input_tokens_seen": 70144000, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 1.725, | |
| "grad_norm": 0.06810135394334793, | |
| "learning_rate": 2.1250416666666666e-05, | |
| "loss": 0.101, | |
| "num_input_tokens_seen": 70656000, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 1.7375, | |
| "grad_norm": 0.019835174083709717, | |
| "learning_rate": 2.1042083333333333e-05, | |
| "loss": 0.0719, | |
| "num_input_tokens_seen": 71168000, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "grad_norm": 0.07850214093923569, | |
| "learning_rate": 2.083375e-05, | |
| "loss": 0.0808, | |
| "num_input_tokens_seen": 71680000, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 1.7625, | |
| "grad_norm": 0.050091035664081573, | |
| "learning_rate": 2.0625416666666667e-05, | |
| "loss": 0.0835, | |
| "num_input_tokens_seen": 72192000, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 1.775, | |
| "grad_norm": 0.01498348731547594, | |
| "learning_rate": 2.0417083333333335e-05, | |
| "loss": 0.0918, | |
| "num_input_tokens_seen": 72704000, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 1.7875, | |
| "grad_norm": 0.034435465931892395, | |
| "learning_rate": 2.0208750000000002e-05, | |
| "loss": 0.0777, | |
| "num_input_tokens_seen": 73216000, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 0.045340586453676224, | |
| "learning_rate": 2.000041666666667e-05, | |
| "loss": 0.0894, | |
| "num_input_tokens_seen": 73728000, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 1.8125, | |
| "grad_norm": 0.28705134987831116, | |
| "learning_rate": 1.9792083333333333e-05, | |
| "loss": 0.0741, | |
| "num_input_tokens_seen": 74240000, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 1.825, | |
| "grad_norm": 155.12445068359375, | |
| "learning_rate": 1.958375e-05, | |
| "loss": 0.0813, | |
| "num_input_tokens_seen": 74752000, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 1.8375, | |
| "grad_norm": 0.08065774291753769, | |
| "learning_rate": 1.9375416666666668e-05, | |
| "loss": 0.0819, | |
| "num_input_tokens_seen": 75264000, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "grad_norm": 0.40842482447624207, | |
| "learning_rate": 1.9167083333333335e-05, | |
| "loss": 0.0669, | |
| "num_input_tokens_seen": 75776000, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 1.8625, | |
| "grad_norm": 0.09376771003007889, | |
| "learning_rate": 1.895875e-05, | |
| "loss": 0.0749, | |
| "num_input_tokens_seen": 76288000, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 1.875, | |
| "grad_norm": 0.035654786974191666, | |
| "learning_rate": 1.8750416666666666e-05, | |
| "loss": 0.0757, | |
| "num_input_tokens_seen": 76800000, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 1.8875, | |
| "grad_norm": 0.04763418436050415, | |
| "learning_rate": 1.8542083333333337e-05, | |
| "loss": 0.0652, | |
| "num_input_tokens_seen": 77312000, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "grad_norm": 0.04799911379814148, | |
| "learning_rate": 1.833375e-05, | |
| "loss": 0.0688, | |
| "num_input_tokens_seen": 77824000, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 1.9125, | |
| "grad_norm": 0.22011104226112366, | |
| "learning_rate": 1.8125416666666668e-05, | |
| "loss": 0.0797, | |
| "num_input_tokens_seen": 78336000, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 1.925, | |
| "grad_norm": 0.05009845644235611, | |
| "learning_rate": 1.7917083333333335e-05, | |
| "loss": 0.0542, | |
| "num_input_tokens_seen": 78848000, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 1.9375, | |
| "grad_norm": 1.8440918922424316, | |
| "learning_rate": 1.7708750000000002e-05, | |
| "loss": 0.0773, | |
| "num_input_tokens_seen": 79360000, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "grad_norm": 0.025103362277150154, | |
| "learning_rate": 1.7500416666666666e-05, | |
| "loss": 0.0692, | |
| "num_input_tokens_seen": 79872000, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 1.9625, | |
| "grad_norm": 14.473067283630371, | |
| "learning_rate": 1.7292083333333333e-05, | |
| "loss": 0.0658, | |
| "num_input_tokens_seen": 80384000, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 1.975, | |
| "grad_norm": 0.051201172173023224, | |
| "learning_rate": 1.708375e-05, | |
| "loss": 0.0648, | |
| "num_input_tokens_seen": 80896000, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 1.9875, | |
| "grad_norm": 0.05888315662741661, | |
| "learning_rate": 1.6875416666666668e-05, | |
| "loss": 0.0792, | |
| "num_input_tokens_seen": 81408000, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.08364333212375641, | |
| "learning_rate": 1.666708333333333e-05, | |
| "loss": 0.0682, | |
| "num_input_tokens_seen": 81920000, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.9528, | |
| "eval_combined_score": 1.327030219983165, | |
| "eval_loss": 0.22639968991279602, | |
| "eval_runtime": 39.5018, | |
| "eval_samples_per_second": 2025.225, | |
| "eval_steps_per_second": 253.153, | |
| "num_input_tokens_seen": 81920000, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 2.0125, | |
| "grad_norm": 1.8050274848937988, | |
| "learning_rate": 1.645875e-05, | |
| "loss": 0.0508, | |
| "num_input_tokens_seen": 82432000, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 2.025, | |
| "grad_norm": 0.02845793031156063, | |
| "learning_rate": 1.625041666666667e-05, | |
| "loss": 0.0525, | |
| "num_input_tokens_seen": 82944000, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 2.0375, | |
| "grad_norm": 0.025743141770362854, | |
| "learning_rate": 1.6042083333333337e-05, | |
| "loss": 0.0483, | |
| "num_input_tokens_seen": 83456000, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "grad_norm": 8.399374008178711, | |
| "learning_rate": 1.583375e-05, | |
| "loss": 0.0652, | |
| "num_input_tokens_seen": 83968000, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 2.0625, | |
| "grad_norm": 0.017835861071944237, | |
| "learning_rate": 1.5625416666666668e-05, | |
| "loss": 0.0482, | |
| "num_input_tokens_seen": 84480000, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 2.075, | |
| "grad_norm": 0.06692575663328171, | |
| "learning_rate": 1.5417083333333335e-05, | |
| "loss": 0.0521, | |
| "num_input_tokens_seen": 84992000, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 2.0875, | |
| "grad_norm": 0.035685233771800995, | |
| "learning_rate": 1.5208749999999999e-05, | |
| "loss": 0.0582, | |
| "num_input_tokens_seen": 85504000, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "grad_norm": 0.13481773436069489, | |
| "learning_rate": 1.5000416666666666e-05, | |
| "loss": 0.0468, | |
| "num_input_tokens_seen": 86016000, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 2.1125, | |
| "grad_norm": 17.486289978027344, | |
| "learning_rate": 1.4792083333333334e-05, | |
| "loss": 0.0559, | |
| "num_input_tokens_seen": 86528000, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 2.125, | |
| "grad_norm": 0.03201691806316376, | |
| "learning_rate": 1.458375e-05, | |
| "loss": 0.0425, | |
| "num_input_tokens_seen": 87040000, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 2.1375, | |
| "grad_norm": 0.021227147430181503, | |
| "learning_rate": 1.4375416666666666e-05, | |
| "loss": 0.05, | |
| "num_input_tokens_seen": 87552000, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "grad_norm": 0.040053412318229675, | |
| "learning_rate": 1.4167083333333334e-05, | |
| "loss": 0.048, | |
| "num_input_tokens_seen": 88064000, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 2.1625, | |
| "grad_norm": 0.01835712045431137, | |
| "learning_rate": 1.395875e-05, | |
| "loss": 0.038, | |
| "num_input_tokens_seen": 88576000, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 2.175, | |
| "grad_norm": 0.03529110550880432, | |
| "learning_rate": 1.3750416666666668e-05, | |
| "loss": 0.051, | |
| "num_input_tokens_seen": 89088000, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 2.1875, | |
| "grad_norm": 97.47898864746094, | |
| "learning_rate": 1.3542083333333334e-05, | |
| "loss": 0.0373, | |
| "num_input_tokens_seen": 89600000, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "grad_norm": 0.014346601441502571, | |
| "learning_rate": 1.3333750000000001e-05, | |
| "loss": 0.0557, | |
| "num_input_tokens_seen": 90112000, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 2.2125, | |
| "grad_norm": 0.01878521591424942, | |
| "learning_rate": 1.3125416666666668e-05, | |
| "loss": 0.0526, | |
| "num_input_tokens_seen": 90624000, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 2.225, | |
| "grad_norm": 0.04132438451051712, | |
| "learning_rate": 1.2917083333333335e-05, | |
| "loss": 0.0464, | |
| "num_input_tokens_seen": 91136000, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 2.2375, | |
| "grad_norm": 0.02642699144780636, | |
| "learning_rate": 1.270875e-05, | |
| "loss": 0.0464, | |
| "num_input_tokens_seen": 91648000, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "grad_norm": 877.631591796875, | |
| "learning_rate": 1.2500416666666666e-05, | |
| "loss": 0.0374, | |
| "num_input_tokens_seen": 92160000, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 2.2625, | |
| "grad_norm": 0.4528743326663971, | |
| "learning_rate": 1.2292083333333334e-05, | |
| "loss": 0.059, | |
| "num_input_tokens_seen": 92672000, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 2.275, | |
| "grad_norm": 0.1183973178267479, | |
| "learning_rate": 1.2083750000000001e-05, | |
| "loss": 0.0505, | |
| "num_input_tokens_seen": 93184000, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 2.2875, | |
| "grad_norm": 0.04196188971400261, | |
| "learning_rate": 1.1875416666666667e-05, | |
| "loss": 0.0517, | |
| "num_input_tokens_seen": 93696000, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "grad_norm": 0.03194092586636543, | |
| "learning_rate": 1.1667083333333334e-05, | |
| "loss": 0.0511, | |
| "num_input_tokens_seen": 94208000, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 2.3125, | |
| "grad_norm": 0.0184203889220953, | |
| "learning_rate": 1.145875e-05, | |
| "loss": 0.0501, | |
| "num_input_tokens_seen": 94720000, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 2.325, | |
| "grad_norm": 0.053758785128593445, | |
| "learning_rate": 1.1250416666666667e-05, | |
| "loss": 0.0412, | |
| "num_input_tokens_seen": 95232000, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 2.3375, | |
| "grad_norm": 0.036872394382953644, | |
| "learning_rate": 1.1042083333333334e-05, | |
| "loss": 0.0418, | |
| "num_input_tokens_seen": 95744000, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "grad_norm": 0.014922083355486393, | |
| "learning_rate": 1.0833750000000001e-05, | |
| "loss": 0.0507, | |
| "num_input_tokens_seen": 96256000, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 2.3625, | |
| "grad_norm": 0.020171664655208588, | |
| "learning_rate": 1.0625416666666667e-05, | |
| "loss": 0.0444, | |
| "num_input_tokens_seen": 96768000, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 2.375, | |
| "grad_norm": 1.7357385158538818, | |
| "learning_rate": 1.0417083333333334e-05, | |
| "loss": 0.0545, | |
| "num_input_tokens_seen": 97280000, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 2.3875, | |
| "grad_norm": 0.023304857313632965, | |
| "learning_rate": 1.020875e-05, | |
| "loss": 0.0515, | |
| "num_input_tokens_seen": 97792000, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 0.01939631998538971, | |
| "learning_rate": 1.0000416666666668e-05, | |
| "loss": 0.0495, | |
| "num_input_tokens_seen": 98304000, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 2.4125, | |
| "grad_norm": 0.019845569506287575, | |
| "learning_rate": 9.792083333333334e-06, | |
| "loss": 0.0411, | |
| "num_input_tokens_seen": 98816000, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 2.425, | |
| "grad_norm": 0.014959324151277542, | |
| "learning_rate": 9.583750000000001e-06, | |
| "loss": 0.0342, | |
| "num_input_tokens_seen": 99328000, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 2.4375, | |
| "grad_norm": 0.01649474911391735, | |
| "learning_rate": 9.375416666666667e-06, | |
| "loss": 0.0472, | |
| "num_input_tokens_seen": 99840000, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "grad_norm": 0.031929146498441696, | |
| "learning_rate": 9.167083333333332e-06, | |
| "loss": 0.0384, | |
| "num_input_tokens_seen": 100352000, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 2.4625, | |
| "grad_norm": 0.10980285704135895, | |
| "learning_rate": 8.958750000000001e-06, | |
| "loss": 0.0513, | |
| "num_input_tokens_seen": 100864000, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 2.475, | |
| "grad_norm": 0.020933426916599274, | |
| "learning_rate": 8.750416666666667e-06, | |
| "loss": 0.0478, | |
| "num_input_tokens_seen": 101376000, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 2.4875, | |
| "grad_norm": 0.017404716461896896, | |
| "learning_rate": 8.542083333333334e-06, | |
| "loss": 0.039, | |
| "num_input_tokens_seen": 101888000, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 0.11211936920881271, | |
| "learning_rate": 8.33375e-06, | |
| "loss": 0.0473, | |
| "num_input_tokens_seen": 102400000, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 2.5125, | |
| "grad_norm": 0.023433908820152283, | |
| "learning_rate": 8.125416666666667e-06, | |
| "loss": 0.0417, | |
| "num_input_tokens_seen": 102912000, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 2.525, | |
| "grad_norm": 0.011504637077450752, | |
| "learning_rate": 7.917083333333334e-06, | |
| "loss": 0.0396, | |
| "num_input_tokens_seen": 103424000, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 2.5375, | |
| "grad_norm": 0.012843768112361431, | |
| "learning_rate": 7.708750000000001e-06, | |
| "loss": 0.0475, | |
| "num_input_tokens_seen": 103936000, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "grad_norm": 0.020524220541119576, | |
| "learning_rate": 7.500416666666667e-06, | |
| "loss": 0.0472, | |
| "num_input_tokens_seen": 104448000, | |
| "step": 102000 | |
| }, | |
| { | |
| "epoch": 2.5625, | |
| "grad_norm": 0.01861303672194481, | |
| "learning_rate": 7.292083333333334e-06, | |
| "loss": 0.0294, | |
| "num_input_tokens_seen": 104960000, | |
| "step": 102500 | |
| }, | |
| { | |
| "epoch": 2.575, | |
| "grad_norm": 0.021343663334846497, | |
| "learning_rate": 7.08375e-06, | |
| "loss": 0.0376, | |
| "num_input_tokens_seen": 105472000, | |
| "step": 103000 | |
| }, | |
| { | |
| "epoch": 2.5875, | |
| "grad_norm": 0.8902124166488647, | |
| "learning_rate": 6.875416666666668e-06, | |
| "loss": 0.0456, | |
| "num_input_tokens_seen": 105984000, | |
| "step": 103500 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "grad_norm": 0.06912536174058914, | |
| "learning_rate": 6.667083333333333e-06, | |
| "loss": 0.0365, | |
| "num_input_tokens_seen": 106496000, | |
| "step": 104000 | |
| }, | |
| { | |
| "epoch": 2.6125, | |
| "grad_norm": 0.17968295514583588, | |
| "learning_rate": 6.458750000000001e-06, | |
| "loss": 0.0372, | |
| "num_input_tokens_seen": 107008000, | |
| "step": 104500 | |
| }, | |
| { | |
| "epoch": 2.625, | |
| "grad_norm": 0.01591988280415535, | |
| "learning_rate": 6.250416666666667e-06, | |
| "loss": 0.0334, | |
| "num_input_tokens_seen": 107520000, | |
| "step": 105000 | |
| }, | |
| { | |
| "epoch": 2.6375, | |
| "grad_norm": 0.015928415581583977, | |
| "learning_rate": 6.0420833333333334e-06, | |
| "loss": 0.0425, | |
| "num_input_tokens_seen": 108032000, | |
| "step": 105500 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "grad_norm": 0.04025963693857193, | |
| "learning_rate": 5.833750000000001e-06, | |
| "loss": 0.0371, | |
| "num_input_tokens_seen": 108544000, | |
| "step": 106000 | |
| }, | |
| { | |
| "epoch": 2.6625, | |
| "grad_norm": 0.07384547591209412, | |
| "learning_rate": 5.625416666666667e-06, | |
| "loss": 0.043, | |
| "num_input_tokens_seen": 109056000, | |
| "step": 106500 | |
| }, | |
| { | |
| "epoch": 2.675, | |
| "grad_norm": 0.0424518883228302, | |
| "learning_rate": 5.4170833333333335e-06, | |
| "loss": 0.0349, | |
| "num_input_tokens_seen": 109568000, | |
| "step": 107000 | |
| }, | |
| { | |
| "epoch": 2.6875, | |
| "grad_norm": 0.01949002780020237, | |
| "learning_rate": 5.208750000000001e-06, | |
| "loss": 0.04, | |
| "num_input_tokens_seen": 110080000, | |
| "step": 107500 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "grad_norm": 0.011860487051308155, | |
| "learning_rate": 5.000416666666667e-06, | |
| "loss": 0.0274, | |
| "num_input_tokens_seen": 110592000, | |
| "step": 108000 | |
| }, | |
| { | |
| "epoch": 2.7125, | |
| "grad_norm": 0.010354108177125454, | |
| "learning_rate": 4.7920833333333335e-06, | |
| "loss": 0.0305, | |
| "num_input_tokens_seen": 111104000, | |
| "step": 108500 | |
| }, | |
| { | |
| "epoch": 2.725, | |
| "grad_norm": 13.753798484802246, | |
| "learning_rate": 4.583750000000001e-06, | |
| "loss": 0.0446, | |
| "num_input_tokens_seen": 111616000, | |
| "step": 109000 | |
| }, | |
| { | |
| "epoch": 2.7375, | |
| "grad_norm": 0.017097918316721916, | |
| "learning_rate": 4.375416666666666e-06, | |
| "loss": 0.0398, | |
| "num_input_tokens_seen": 112128000, | |
| "step": 109500 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "grad_norm": 0.01976764015853405, | |
| "learning_rate": 4.167083333333334e-06, | |
| "loss": 0.0318, | |
| "num_input_tokens_seen": 112640000, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 2.7625, | |
| "grad_norm": 6.2214531898498535, | |
| "learning_rate": 3.95875e-06, | |
| "loss": 0.0424, | |
| "num_input_tokens_seen": 113152000, | |
| "step": 110500 | |
| }, | |
| { | |
| "epoch": 2.775, | |
| "grad_norm": 0.007995002903044224, | |
| "learning_rate": 3.750416666666667e-06, | |
| "loss": 0.0442, | |
| "num_input_tokens_seen": 113664000, | |
| "step": 111000 | |
| }, | |
| { | |
| "epoch": 2.7875, | |
| "grad_norm": 0.05575885996222496, | |
| "learning_rate": 3.5420833333333332e-06, | |
| "loss": 0.037, | |
| "num_input_tokens_seen": 114176000, | |
| "step": 111500 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "grad_norm": 0.1336035579442978, | |
| "learning_rate": 3.33375e-06, | |
| "loss": 0.0385, | |
| "num_input_tokens_seen": 114688000, | |
| "step": 112000 | |
| }, | |
| { | |
| "epoch": 2.8125, | |
| "grad_norm": 0.019037237390875816, | |
| "learning_rate": 3.125416666666667e-06, | |
| "loss": 0.0326, | |
| "num_input_tokens_seen": 115200000, | |
| "step": 112500 | |
| }, | |
| { | |
| "epoch": 2.825, | |
| "grad_norm": 0.00875825248658657, | |
| "learning_rate": 2.9170833333333333e-06, | |
| "loss": 0.0277, | |
| "num_input_tokens_seen": 115712000, | |
| "step": 113000 | |
| }, | |
| { | |
| "epoch": 2.8375, | |
| "grad_norm": 0.052571795880794525, | |
| "learning_rate": 2.70875e-06, | |
| "loss": 0.037, | |
| "num_input_tokens_seen": 116224000, | |
| "step": 113500 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "grad_norm": 0.019845254719257355, | |
| "learning_rate": 2.500416666666667e-06, | |
| "loss": 0.0445, | |
| "num_input_tokens_seen": 116736000, | |
| "step": 114000 | |
| }, | |
| { | |
| "epoch": 2.8625, | |
| "grad_norm": 0.01046211551874876, | |
| "learning_rate": 2.2920833333333338e-06, | |
| "loss": 0.0294, | |
| "num_input_tokens_seen": 117248000, | |
| "step": 114500 | |
| }, | |
| { | |
| "epoch": 2.875, | |
| "grad_norm": 0.012272953987121582, | |
| "learning_rate": 2.0837499999999997e-06, | |
| "loss": 0.0396, | |
| "num_input_tokens_seen": 117760000, | |
| "step": 115000 | |
| }, | |
| { | |
| "epoch": 2.8875, | |
| "grad_norm": 0.022472262382507324, | |
| "learning_rate": 1.8754166666666666e-06, | |
| "loss": 0.047, | |
| "num_input_tokens_seen": 118272000, | |
| "step": 115500 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "grad_norm": 0.07120255380868912, | |
| "learning_rate": 1.6670833333333334e-06, | |
| "loss": 0.0437, | |
| "num_input_tokens_seen": 118784000, | |
| "step": 116000 | |
| }, | |
| { | |
| "epoch": 2.9125, | |
| "grad_norm": 0.060723673552274704, | |
| "learning_rate": 1.45875e-06, | |
| "loss": 0.035, | |
| "num_input_tokens_seen": 119296000, | |
| "step": 116500 | |
| }, | |
| { | |
| "epoch": 2.925, | |
| "grad_norm": 0.012350406497716904, | |
| "learning_rate": 1.2504166666666668e-06, | |
| "loss": 0.03, | |
| "num_input_tokens_seen": 119808000, | |
| "step": 117000 | |
| }, | |
| { | |
| "epoch": 2.9375, | |
| "grad_norm": 0.18025244772434235, | |
| "learning_rate": 1.0420833333333334e-06, | |
| "loss": 0.0306, | |
| "num_input_tokens_seen": 120320000, | |
| "step": 117500 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "grad_norm": 0.007316856179386377, | |
| "learning_rate": 8.3375e-07, | |
| "loss": 0.0328, | |
| "num_input_tokens_seen": 120832000, | |
| "step": 118000 | |
| }, | |
| { | |
| "epoch": 2.9625, | |
| "grad_norm": 0.06496240198612213, | |
| "learning_rate": 6.254166666666667e-07, | |
| "loss": 0.039, | |
| "num_input_tokens_seen": 121344000, | |
| "step": 118500 | |
| }, | |
| { | |
| "epoch": 2.975, | |
| "grad_norm": 0.0057182470336556435, | |
| "learning_rate": 4.170833333333334e-07, | |
| "loss": 0.0367, | |
| "num_input_tokens_seen": 121856000, | |
| "step": 119000 | |
| }, | |
| { | |
| "epoch": 2.9875, | |
| "grad_norm": 0.017466630786657333, | |
| "learning_rate": 2.0875e-07, | |
| "loss": 0.0381, | |
| "num_input_tokens_seen": 122368000, | |
| "step": 119500 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 0.013719202019274235, | |
| "learning_rate": 4.1666666666666673e-10, | |
| "loss": 0.0424, | |
| "num_input_tokens_seen": 122880000, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.955225, | |
| "eval_combined_score": 1.2845397902488602, | |
| "eval_loss": 0.2597787082195282, | |
| "eval_runtime": 39.5285, | |
| "eval_samples_per_second": 2023.856, | |
| "eval_steps_per_second": 252.982, | |
| "num_input_tokens_seen": 122880000, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "num_input_tokens_seen": 122880000, | |
| "step": 120000, | |
| "total_flos": 1.580945522688e+16, | |
| "train_loss": 0.10143799341519674, | |
| "train_runtime": 4258.2582, | |
| "train_samples_per_second": 225.444, | |
| "train_steps_per_second": 28.181, | |
| "train_tokens_per_second": 28856.869 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 120000, | |
| "num_input_tokens_seen": 122880000, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.580945522688e+16, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |