{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 50000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05, "grad_norm": 2.37485408782959, "learning_rate": 0.0002, "loss": 0.7725, "step": 500 }, { "epoch": 0.1, "grad_norm": 2.3556101322174072, "learning_rate": 0.0004, "loss": 0.73, "step": 1000 }, { "epoch": 0.15, "grad_norm": 4.192414283752441, "learning_rate": 0.0006, "loss": 0.7414, "step": 1500 }, { "epoch": 0.2, "grad_norm": 3.5939464569091797, "learning_rate": 0.0008, "loss": 0.7544, "step": 2000 }, { "epoch": 0.25, "grad_norm": 15.336837768554688, "learning_rate": 0.001, "loss": 0.7723, "step": 2500 }, { "epoch": 0.3, "grad_norm": 3.893497943878174, "learning_rate": 0.0009894736842105264, "loss": 0.7893, "step": 3000 }, { "epoch": 0.35, "grad_norm": 11.713598251342773, "learning_rate": 0.0009789473684210528, "loss": 0.7843, "step": 3500 }, { "epoch": 0.4, "grad_norm": 13.350496292114258, "learning_rate": 0.000968421052631579, "loss": 0.7894, "step": 4000 }, { "epoch": 0.45, "grad_norm": 2.8563292026519775, "learning_rate": 0.0009578947368421053, "loss": 0.7843, "step": 4500 }, { "epoch": 0.5, "grad_norm": 9.680468559265137, "learning_rate": 0.0009473684210526315, "loss": 0.8511, "step": 5000 }, { "epoch": 0.55, "grad_norm": 4.680222034454346, "learning_rate": 0.0009368421052631579, "loss": 0.8686, "step": 5500 }, { "epoch": 0.6, "grad_norm": 12.037607192993164, "learning_rate": 0.0009263157894736843, "loss": 0.8393, "step": 6000 }, { "epoch": 0.65, "grad_norm": 8.154277801513672, "learning_rate": 0.0009157894736842105, "loss": 0.8252, "step": 6500 }, { "epoch": 0.7, "grad_norm": 3.256535530090332, "learning_rate": 0.0009052631578947369, "loss": 0.7854, "step": 7000 }, { "epoch": 0.75, "grad_norm": 4.168020725250244, "learning_rate": 0.0008947368421052632, "loss": 0.8059, "step": 7500 }, { "epoch": 0.8, "grad_norm": 2.775940418243408, "learning_rate": 0.0008842105263157894, "loss": 0.8272, "step": 8000 }, { "epoch": 0.85, "grad_norm": 14.476820945739746, "learning_rate": 0.0008736842105263159, "loss": 0.8006, "step": 8500 }, { "epoch": 0.9, "grad_norm": 14.232259750366211, "learning_rate": 0.0008631578947368422, "loss": 0.7537, "step": 9000 }, { "epoch": 0.95, "grad_norm": 5.338253021240234, "learning_rate": 0.0008526315789473684, "loss": 0.7831, "step": 9500 }, { "epoch": 1.0, "grad_norm": 5.454962730407715, "learning_rate": 0.0008421052631578947, "loss": 0.7633, "step": 10000 }, { "epoch": 1.05, "grad_norm": 8.250731468200684, "learning_rate": 0.0008315789473684212, "loss": 0.7565, "step": 10500 }, { "epoch": 1.1, "grad_norm": 8.233278274536133, "learning_rate": 0.0008210526315789474, "loss": 0.7724, "step": 11000 }, { "epoch": 1.15, "grad_norm": 10.603631019592285, "learning_rate": 0.0008105263157894737, "loss": 0.782, "step": 11500 }, { "epoch": 1.2, "grad_norm": 3.75233793258667, "learning_rate": 0.0008, "loss": 0.7519, "step": 12000 }, { "epoch": 1.25, "grad_norm": 4.677459239959717, "learning_rate": 0.0007894736842105263, "loss": 0.7556, "step": 12500 }, { "epoch": 1.3, "grad_norm": 4.583637714385986, "learning_rate": 0.0007789473684210527, "loss": 0.7459, "step": 13000 }, { "epoch": 1.35, "grad_norm": 9.343656539916992, "learning_rate": 0.0007684210526315789, "loss": 0.7802, "step": 13500 }, { "epoch": 1.4, "grad_norm": 1.4162571430206299, "learning_rate": 0.0007578947368421053, "loss": 0.7384, "step": 14000 }, { "epoch": 1.45, "grad_norm": 6.586658477783203, "learning_rate": 0.0007473684210526316, "loss": 0.7795, "step": 14500 }, { "epoch": 1.5, "grad_norm": 5.946737289428711, "learning_rate": 0.0007368421052631579, "loss": 0.7381, "step": 15000 }, { "epoch": 1.55, "grad_norm": 5.825309753417969, "learning_rate": 0.0007263157894736843, "loss": 0.7482, "step": 15500 }, { "epoch": 1.6, "grad_norm": 2.1794614791870117, "learning_rate": 0.0007157894736842105, "loss": 0.7381, "step": 16000 }, { "epoch": 1.65, "grad_norm": 7.837281227111816, "learning_rate": 0.0007052631578947368, "loss": 0.7314, "step": 16500 }, { "epoch": 1.7, "grad_norm": 4.612753868103027, "learning_rate": 0.0006947368421052632, "loss": 0.7384, "step": 17000 }, { "epoch": 1.75, "grad_norm": 6.038923263549805, "learning_rate": 0.0006842105263157895, "loss": 0.752, "step": 17500 }, { "epoch": 1.8, "grad_norm": 5.17457389831543, "learning_rate": 0.0006736842105263158, "loss": 0.7421, "step": 18000 }, { "epoch": 1.85, "grad_norm": 4.990045547485352, "learning_rate": 0.0006631578947368421, "loss": 0.7253, "step": 18500 }, { "epoch": 1.9, "grad_norm": 3.1730449199676514, "learning_rate": 0.0006526315789473684, "loss": 0.7202, "step": 19000 }, { "epoch": 1.95, "grad_norm": 1.5063399076461792, "learning_rate": 0.0006421052631578948, "loss": 0.7488, "step": 19500 }, { "epoch": 2.0, "grad_norm": 3.6247060298919678, "learning_rate": 0.0006315789473684211, "loss": 0.7263, "step": 20000 }, { "epoch": 2.05, "grad_norm": 1.4821553230285645, "learning_rate": 0.0006210526315789474, "loss": 0.7176, "step": 20500 }, { "epoch": 2.1, "grad_norm": 2.641312837600708, "learning_rate": 0.0006105263157894737, "loss": 0.7182, "step": 21000 }, { "epoch": 2.15, "grad_norm": 3.3185555934906006, "learning_rate": 0.0006, "loss": 0.7257, "step": 21500 }, { "epoch": 2.2, "grad_norm": 2.4133715629577637, "learning_rate": 0.0005894736842105263, "loss": 0.7202, "step": 22000 }, { "epoch": 2.25, "grad_norm": 1.8694628477096558, "learning_rate": 0.0005789473684210527, "loss": 0.7172, "step": 22500 }, { "epoch": 2.3, "grad_norm": 5.248977184295654, "learning_rate": 0.0005684210526315789, "loss": 0.7148, "step": 23000 }, { "epoch": 2.35, "grad_norm": 6.179994106292725, "learning_rate": 0.0005578947368421052, "loss": 0.7308, "step": 23500 }, { "epoch": 2.4, "grad_norm": 1.9322715997695923, "learning_rate": 0.0005473684210526317, "loss": 0.7163, "step": 24000 }, { "epoch": 2.45, "grad_norm": 1.6366279125213623, "learning_rate": 0.0005368421052631579, "loss": 0.7262, "step": 24500 }, { "epoch": 2.5, "grad_norm": 5.179625034332275, "learning_rate": 0.0005263157894736842, "loss": 0.7054, "step": 25000 }, { "epoch": 2.55, "grad_norm": 1.7737842798233032, "learning_rate": 0.0005157894736842106, "loss": 0.7104, "step": 25500 }, { "epoch": 2.6, "grad_norm": 1.6575433015823364, "learning_rate": 0.0005052631578947368, "loss": 0.7114, "step": 26000 }, { "epoch": 2.65, "grad_norm": 3.6518235206604004, "learning_rate": 0.0004947368421052632, "loss": 0.7087, "step": 26500 }, { "epoch": 2.7, "grad_norm": 1.3658417463302612, "learning_rate": 0.0004842105263157895, "loss": 0.7106, "step": 27000 }, { "epoch": 2.75, "grad_norm": 1.312829852104187, "learning_rate": 0.00047368421052631577, "loss": 0.7118, "step": 27500 }, { "epoch": 2.8, "grad_norm": 2.94315767288208, "learning_rate": 0.00046315789473684214, "loss": 0.7102, "step": 28000 }, { "epoch": 2.85, "grad_norm": 1.349211573600769, "learning_rate": 0.00045263157894736845, "loss": 0.7058, "step": 28500 }, { "epoch": 2.9, "grad_norm": 1.3551281690597534, "learning_rate": 0.0004421052631578947, "loss": 0.708, "step": 29000 }, { "epoch": 2.95, "grad_norm": 1.3995367288589478, "learning_rate": 0.0004315789473684211, "loss": 0.7106, "step": 29500 }, { "epoch": 3.0, "grad_norm": 0.9616995453834534, "learning_rate": 0.00042105263157894734, "loss": 0.7021, "step": 30000 }, { "epoch": 3.05, "grad_norm": 2.390662908554077, "learning_rate": 0.0004105263157894737, "loss": 0.7061, "step": 30500 }, { "epoch": 3.1, "grad_norm": 1.2397950887680054, "learning_rate": 0.0004, "loss": 0.705, "step": 31000 }, { "epoch": 3.15, "grad_norm": 3.110950231552124, "learning_rate": 0.00038947368421052633, "loss": 0.7031, "step": 31500 }, { "epoch": 3.2, "grad_norm": 1.1811543703079224, "learning_rate": 0.00037894736842105265, "loss": 0.7013, "step": 32000 }, { "epoch": 3.25, "grad_norm": 1.1221048831939697, "learning_rate": 0.00036842105263157896, "loss": 0.7026, "step": 32500 }, { "epoch": 3.3, "grad_norm": 0.9882568717002869, "learning_rate": 0.0003578947368421053, "loss": 0.7012, "step": 33000 }, { "epoch": 3.35, "grad_norm": 0.8155710697174072, "learning_rate": 0.0003473684210526316, "loss": 0.7033, "step": 33500 }, { "epoch": 3.4, "grad_norm": 1.1295320987701416, "learning_rate": 0.0003368421052631579, "loss": 0.701, "step": 34000 }, { "epoch": 3.45, "grad_norm": 2.0776407718658447, "learning_rate": 0.0003263157894736842, "loss": 0.6986, "step": 34500 }, { "epoch": 3.5, "grad_norm": 1.7907339334487915, "learning_rate": 0.00031578947368421053, "loss": 0.698, "step": 35000 }, { "epoch": 3.55, "grad_norm": 1.978267788887024, "learning_rate": 0.00030526315789473684, "loss": 0.7031, "step": 35500 }, { "epoch": 3.6, "grad_norm": 3.065232515335083, "learning_rate": 0.00029473684210526316, "loss": 0.697, "step": 36000 }, { "epoch": 3.65, "grad_norm": 0.9516103267669678, "learning_rate": 0.00028421052631578947, "loss": 0.7006, "step": 36500 }, { "epoch": 3.7, "grad_norm": 1.0045952796936035, "learning_rate": 0.00027368421052631584, "loss": 0.6986, "step": 37000 }, { "epoch": 3.75, "grad_norm": 1.6901030540466309, "learning_rate": 0.0002631578947368421, "loss": 0.6995, "step": 37500 }, { "epoch": 3.8, "grad_norm": 1.8813843727111816, "learning_rate": 0.0002526315789473684, "loss": 0.6987, "step": 38000 }, { "epoch": 3.85, "grad_norm": 1.8174165487289429, "learning_rate": 0.00024210526315789475, "loss": 0.6999, "step": 38500 }, { "epoch": 3.9, "grad_norm": 1.7451245784759521, "learning_rate": 0.00023157894736842107, "loss": 0.6981, "step": 39000 }, { "epoch": 3.95, "grad_norm": 0.9422154426574707, "learning_rate": 0.00022105263157894735, "loss": 0.6962, "step": 39500 }, { "epoch": 4.0, "grad_norm": 2.196399688720703, "learning_rate": 0.00021052631578947367, "loss": 0.6984, "step": 40000 }, { "epoch": 4.05, "grad_norm": 1.940382719039917, "learning_rate": 0.0002, "loss": 0.6945, "step": 40500 }, { "epoch": 4.1, "grad_norm": 1.2243621349334717, "learning_rate": 0.00018947368421052632, "loss": 0.6965, "step": 41000 }, { "epoch": 4.15, "grad_norm": 2.004392623901367, "learning_rate": 0.00017894736842105264, "loss": 0.6992, "step": 41500 }, { "epoch": 4.2, "grad_norm": 1.8801628351211548, "learning_rate": 0.00016842105263157895, "loss": 0.6944, "step": 42000 }, { "epoch": 4.25, "grad_norm": 0.9306873679161072, "learning_rate": 0.00015789473684210527, "loss": 0.693, "step": 42500 }, { "epoch": 4.3, "grad_norm": 1.9730675220489502, "learning_rate": 0.00014736842105263158, "loss": 0.6957, "step": 43000 }, { "epoch": 4.35, "grad_norm": 1.0207428932189941, "learning_rate": 0.00013684210526315792, "loss": 0.7004, "step": 43500 }, { "epoch": 4.4, "grad_norm": 0.9524587988853455, "learning_rate": 0.0001263157894736842, "loss": 0.6927, "step": 44000 }, { "epoch": 4.45, "grad_norm": 0.9829258322715759, "learning_rate": 0.00011578947368421053, "loss": 0.6962, "step": 44500 }, { "epoch": 4.5, "grad_norm": 1.317120909690857, "learning_rate": 0.00010526315789473683, "loss": 0.6943, "step": 45000 }, { "epoch": 4.55, "grad_norm": 1.8932749032974243, "learning_rate": 9.473684210526316e-05, "loss": 0.6933, "step": 45500 }, { "epoch": 4.6, "grad_norm": 0.8536391854286194, "learning_rate": 8.421052631578948e-05, "loss": 0.693, "step": 46000 }, { "epoch": 4.65, "grad_norm": 1.9397382736206055, "learning_rate": 7.368421052631579e-05, "loss": 0.6942, "step": 46500 }, { "epoch": 4.7, "grad_norm": 1.846257209777832, "learning_rate": 6.31578947368421e-05, "loss": 0.6941, "step": 47000 }, { "epoch": 4.75, "grad_norm": 0.937384307384491, "learning_rate": 5.263157894736842e-05, "loss": 0.6956, "step": 47500 }, { "epoch": 4.8, "grad_norm": 0.9019428491592407, "learning_rate": 4.210526315789474e-05, "loss": 0.6935, "step": 48000 }, { "epoch": 4.85, "grad_norm": 1.892066478729248, "learning_rate": 3.157894736842105e-05, "loss": 0.6945, "step": 48500 }, { "epoch": 4.9, "grad_norm": 1.764987826347351, "learning_rate": 2.105263157894737e-05, "loss": 0.695, "step": 49000 }, { "epoch": 4.95, "grad_norm": 1.626028060913086, "learning_rate": 1.0526315789473684e-05, "loss": 0.6941, "step": 49500 }, { "epoch": 5.0, "grad_norm": 1.6325174570083618, "learning_rate": 0.0, "loss": 0.6935, "step": 50000 } ], "logging_steps": 500, "max_steps": 50000, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "total_flos": 2877777168000000.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }