{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.0648249186987478, "eval_steps": 500, "global_step": 6000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005402076558228983, "grad_norm": 4.3125, "learning_rate": 1.7644940583363342e-06, "loss": 1.7943, "step": 50 }, { "epoch": 0.0010804153116457966, "grad_norm": 4.28125, "learning_rate": 3.5649981994958587e-06, "loss": 1.7796, "step": 100 }, { "epoch": 0.001620622967468695, "grad_norm": 3.640625, "learning_rate": 5.365502340655383e-06, "loss": 1.7311, "step": 150 }, { "epoch": 0.002160830623291593, "grad_norm": 1.390625, "learning_rate": 7.166006481814909e-06, "loss": 1.7021, "step": 200 }, { "epoch": 0.0027010382791144917, "grad_norm": 2.890625, "learning_rate": 8.966510622974434e-06, "loss": 1.7806, "step": 250 }, { "epoch": 0.00324124593493739, "grad_norm": 1.15625, "learning_rate": 1.0767014764133958e-05, "loss": 1.7074, "step": 300 }, { "epoch": 0.003781453590760288, "grad_norm": 0.5390625, "learning_rate": 1.2567518905293481e-05, "loss": 1.6765, "step": 350 }, { "epoch": 0.004321661246583186, "grad_norm": 1.078125, "learning_rate": 1.4368023046453008e-05, "loss": 1.6671, "step": 400 }, { "epoch": 0.004861868902406085, "grad_norm": 0.455078125, "learning_rate": 1.6168527187612533e-05, "loss": 1.6235, "step": 450 }, { "epoch": 0.0054020765582289835, "grad_norm": 0.2890625, "learning_rate": 1.7969031328772056e-05, "loss": 1.6529, "step": 500 }, { "epoch": 0.005942284214051881, "grad_norm": 0.345703125, "learning_rate": 1.9769535469931582e-05, "loss": 1.559, "step": 550 }, { "epoch": 0.00648249186987478, "grad_norm": 0.306640625, "learning_rate": 2.1570039611091105e-05, "loss": 1.5842, "step": 600 }, { "epoch": 0.007022699525697678, "grad_norm": 0.359375, "learning_rate": 2.337054375225063e-05, "loss": 1.5504, "step": 650 }, { "epoch": 0.007562907181520576, "grad_norm": 0.30859375, "learning_rate": 2.5171047893410155e-05, "loss": 1.5611, "step": 700 }, { "epoch": 0.008103114837343476, "grad_norm": 0.91015625, "learning_rate": 2.697155203456968e-05, "loss": 1.5447, "step": 750 }, { "epoch": 0.008643322493166372, "grad_norm": 0.359375, "learning_rate": 2.8772056175729207e-05, "loss": 1.546, "step": 800 }, { "epoch": 0.009183530148989271, "grad_norm": 0.328125, "learning_rate": 3.057256031688873e-05, "loss": 1.5344, "step": 850 }, { "epoch": 0.00972373780481217, "grad_norm": 0.29296875, "learning_rate": 3.237306445804826e-05, "loss": 1.4914, "step": 900 }, { "epoch": 0.010263945460635068, "grad_norm": 0.29296875, "learning_rate": 3.4173568599207777e-05, "loss": 1.559, "step": 950 }, { "epoch": 0.010804153116457967, "grad_norm": 0.283203125, "learning_rate": 3.59740727403673e-05, "loss": 1.485, "step": 1000 }, { "epoch": 0.011344360772280866, "grad_norm": 0.263671875, "learning_rate": 3.777457688152683e-05, "loss": 1.4972, "step": 1050 }, { "epoch": 0.011884568428103762, "grad_norm": 0.439453125, "learning_rate": 3.9575081022686356e-05, "loss": 1.45, "step": 1100 }, { "epoch": 0.012424776083926661, "grad_norm": 0.3046875, "learning_rate": 4.137558516384588e-05, "loss": 1.4232, "step": 1150 }, { "epoch": 0.01296498373974956, "grad_norm": 0.35546875, "learning_rate": 4.31760893050054e-05, "loss": 1.4098, "step": 1200 }, { "epoch": 0.013505191395572458, "grad_norm": 0.330078125, "learning_rate": 4.497659344616493e-05, "loss": 1.3447, "step": 1250 }, { "epoch": 0.014045399051395357, "grad_norm": 0.31640625, "learning_rate": 4.6777097587324455e-05, "loss": 1.35, "step": 1300 }, { "epoch": 0.014585606707218255, "grad_norm": 0.37109375, "learning_rate": 4.857760172848398e-05, "loss": 1.3509, "step": 1350 }, { "epoch": 0.015125814363041152, "grad_norm": 0.3515625, "learning_rate": 5.03781058696435e-05, "loss": 1.3161, "step": 1400 }, { "epoch": 0.01566602201886405, "grad_norm": 0.27734375, "learning_rate": 5.2178610010803034e-05, "loss": 1.3277, "step": 1450 }, { "epoch": 0.01620622967468695, "grad_norm": 0.37890625, "learning_rate": 5.3979114151962554e-05, "loss": 1.2814, "step": 1500 }, { "epoch": 0.016746437330509848, "grad_norm": 0.31640625, "learning_rate": 5.577961829312208e-05, "loss": 1.2944, "step": 1550 }, { "epoch": 0.017286644986332745, "grad_norm": 0.32421875, "learning_rate": 5.75801224342816e-05, "loss": 1.2714, "step": 1600 }, { "epoch": 0.017826852642155645, "grad_norm": 0.3203125, "learning_rate": 5.9380626575441126e-05, "loss": 1.2847, "step": 1650 }, { "epoch": 0.018367060297978542, "grad_norm": 0.390625, "learning_rate": 6.118113071660065e-05, "loss": 1.2828, "step": 1700 }, { "epoch": 0.018907267953801443, "grad_norm": 0.33984375, "learning_rate": 6.298163485776018e-05, "loss": 1.252, "step": 1750 }, { "epoch": 0.01944747560962434, "grad_norm": 0.35546875, "learning_rate": 6.47821389989197e-05, "loss": 1.2659, "step": 1800 }, { "epoch": 0.019987683265447236, "grad_norm": 0.375, "learning_rate": 6.658264314007923e-05, "loss": 1.2429, "step": 1850 }, { "epoch": 0.020527890921270137, "grad_norm": 0.357421875, "learning_rate": 6.838314728123874e-05, "loss": 1.2492, "step": 1900 }, { "epoch": 0.021068098577093033, "grad_norm": 0.39453125, "learning_rate": 7.018365142239827e-05, "loss": 1.2083, "step": 1950 }, { "epoch": 0.021608306232915934, "grad_norm": 0.3515625, "learning_rate": 7.19841555635578e-05, "loss": 1.2459, "step": 2000 }, { "epoch": 0.02214851388873883, "grad_norm": 0.35546875, "learning_rate": 7.378465970471732e-05, "loss": 1.2101, "step": 2050 }, { "epoch": 0.02268872154456173, "grad_norm": 0.34765625, "learning_rate": 7.558516384587685e-05, "loss": 1.1946, "step": 2100 }, { "epoch": 0.023228929200384628, "grad_norm": 0.3671875, "learning_rate": 7.738566798703636e-05, "loss": 1.2097, "step": 2150 }, { "epoch": 0.023769136856207525, "grad_norm": 0.400390625, "learning_rate": 7.91861721281959e-05, "loss": 1.1788, "step": 2200 }, { "epoch": 0.024309344512030425, "grad_norm": 0.419921875, "learning_rate": 8.098667626935542e-05, "loss": 1.1804, "step": 2250 }, { "epoch": 0.024849552167853322, "grad_norm": 0.419921875, "learning_rate": 8.278718041051496e-05, "loss": 1.2001, "step": 2300 }, { "epoch": 0.025389759823676222, "grad_norm": 0.3828125, "learning_rate": 8.458768455167447e-05, "loss": 1.181, "step": 2350 }, { "epoch": 0.02592996747949912, "grad_norm": 0.408203125, "learning_rate": 8.6388188692834e-05, "loss": 1.1763, "step": 2400 }, { "epoch": 0.02647017513532202, "grad_norm": 0.431640625, "learning_rate": 8.818869283399352e-05, "loss": 1.1643, "step": 2450 }, { "epoch": 0.027010382791144916, "grad_norm": 0.412109375, "learning_rate": 8.998919697515305e-05, "loss": 1.1801, "step": 2500 }, { "epoch": 0.027550590446967813, "grad_norm": 0.42578125, "learning_rate": 9.178970111631257e-05, "loss": 1.1554, "step": 2550 }, { "epoch": 0.028090798102790714, "grad_norm": 0.435546875, "learning_rate": 9.35902052574721e-05, "loss": 1.1682, "step": 2600 }, { "epoch": 0.02863100575861361, "grad_norm": 0.400390625, "learning_rate": 9.539070939863161e-05, "loss": 1.1518, "step": 2650 }, { "epoch": 0.02917121341443651, "grad_norm": 0.43359375, "learning_rate": 9.719121353979115e-05, "loss": 1.1331, "step": 2700 }, { "epoch": 0.029711421070259408, "grad_norm": 0.44140625, "learning_rate": 9.899171768095067e-05, "loss": 1.1256, "step": 2750 }, { "epoch": 0.030251628726082305, "grad_norm": 0.474609375, "learning_rate": 0.00010079222182211019, "loss": 1.1367, "step": 2800 }, { "epoch": 0.030791836381905205, "grad_norm": 0.4765625, "learning_rate": 0.00010259272596326972, "loss": 1.1206, "step": 2850 }, { "epoch": 0.0313320440377281, "grad_norm": 0.396484375, "learning_rate": 0.00010439323010442925, "loss": 1.1086, "step": 2900 }, { "epoch": 0.031872251693551, "grad_norm": 0.451171875, "learning_rate": 0.00010619373424558876, "loss": 1.1455, "step": 2950 }, { "epoch": 0.0324124593493739, "grad_norm": 0.5390625, "learning_rate": 0.0001079942383867483, "loss": 1.1079, "step": 3000 }, { "epoch": 0.0329526670051968, "grad_norm": 0.46875, "learning_rate": 0.00010979474252790783, "loss": 1.0991, "step": 3050 }, { "epoch": 0.033492874661019696, "grad_norm": 0.453125, "learning_rate": 0.00011159524666906735, "loss": 1.1154, "step": 3100 }, { "epoch": 0.03403308231684259, "grad_norm": 0.439453125, "learning_rate": 0.00011339575081022686, "loss": 1.1002, "step": 3150 }, { "epoch": 0.03457328997266549, "grad_norm": 0.462890625, "learning_rate": 0.00011519625495138639, "loss": 1.0997, "step": 3200 }, { "epoch": 0.035113497628488394, "grad_norm": 0.412109375, "learning_rate": 0.00011699675909254593, "loss": 1.0754, "step": 3250 }, { "epoch": 0.03565370528431129, "grad_norm": 0.486328125, "learning_rate": 0.00011879726323370543, "loss": 1.0933, "step": 3300 }, { "epoch": 0.03619391294013419, "grad_norm": 0.408203125, "learning_rate": 0.00012059776737486497, "loss": 1.0764, "step": 3350 }, { "epoch": 0.036734120595957084, "grad_norm": 0.453125, "learning_rate": 0.0001223982715160245, "loss": 1.1249, "step": 3400 }, { "epoch": 0.03727432825177998, "grad_norm": 0.40234375, "learning_rate": 0.000124198775657184, "loss": 1.0746, "step": 3450 }, { "epoch": 0.037814535907602885, "grad_norm": 0.51953125, "learning_rate": 0.00012599927979834355, "loss": 1.0747, "step": 3500 }, { "epoch": 0.03835474356342578, "grad_norm": 0.44140625, "learning_rate": 0.00012779978393950308, "loss": 1.0814, "step": 3550 }, { "epoch": 0.03889495121924868, "grad_norm": 0.5, "learning_rate": 0.0001296002880806626, "loss": 1.064, "step": 3600 }, { "epoch": 0.039435158875071576, "grad_norm": 0.4453125, "learning_rate": 0.0001314007922218221, "loss": 1.0867, "step": 3650 }, { "epoch": 0.03997536653089447, "grad_norm": 0.4296875, "learning_rate": 0.00013320129636298166, "loss": 1.066, "step": 3700 }, { "epoch": 0.040515574186717376, "grad_norm": 0.51953125, "learning_rate": 0.00013500180050414118, "loss": 1.0721, "step": 3750 }, { "epoch": 0.04105578184254027, "grad_norm": 0.482421875, "learning_rate": 0.00013680230464530068, "loss": 1.0518, "step": 3800 }, { "epoch": 0.04159598949836317, "grad_norm": 0.466796875, "learning_rate": 0.0001386028087864602, "loss": 1.0495, "step": 3850 }, { "epoch": 0.04213619715418607, "grad_norm": 0.39453125, "learning_rate": 0.00014040331292761973, "loss": 1.0292, "step": 3900 }, { "epoch": 0.04267640481000897, "grad_norm": 0.4375, "learning_rate": 0.00014220381706877926, "loss": 1.031, "step": 3950 }, { "epoch": 0.04321661246583187, "grad_norm": 0.48828125, "learning_rate": 0.0001440043212099388, "loss": 1.026, "step": 4000 }, { "epoch": 0.043756820121654764, "grad_norm": 0.46875, "learning_rate": 0.0001458048253510983, "loss": 1.0372, "step": 4050 }, { "epoch": 0.04429702777747766, "grad_norm": 0.396484375, "learning_rate": 0.00014760532949225784, "loss": 1.0388, "step": 4100 }, { "epoch": 0.04483723543330056, "grad_norm": 0.4296875, "learning_rate": 0.00014940583363341737, "loss": 1.0258, "step": 4150 }, { "epoch": 0.04537744308912346, "grad_norm": 0.478515625, "learning_rate": 0.0001512063377745769, "loss": 1.0111, "step": 4200 }, { "epoch": 0.04591765074494636, "grad_norm": 0.46484375, "learning_rate": 0.00015300684191573642, "loss": 1.0316, "step": 4250 }, { "epoch": 0.046457858400769256, "grad_norm": 0.5546875, "learning_rate": 0.00015480734605689592, "loss": 1.0375, "step": 4300 }, { "epoch": 0.04699806605659215, "grad_norm": 0.408203125, "learning_rate": 0.00015660785019805547, "loss": 1.037, "step": 4350 }, { "epoch": 0.04753827371241505, "grad_norm": 0.458984375, "learning_rate": 0.000158408354339215, "loss": 1.0172, "step": 4400 }, { "epoch": 0.04807848136823795, "grad_norm": 0.455078125, "learning_rate": 0.0001602088584803745, "loss": 1.0482, "step": 4450 }, { "epoch": 0.04861868902406085, "grad_norm": 0.451171875, "learning_rate": 0.00016200936262153402, "loss": 1.0072, "step": 4500 }, { "epoch": 0.04915889667988375, "grad_norm": 0.63671875, "learning_rate": 0.00016380986676269358, "loss": 0.9988, "step": 4550 }, { "epoch": 0.049699104335706644, "grad_norm": 0.42578125, "learning_rate": 0.0001656103709038531, "loss": 0.9998, "step": 4600 }, { "epoch": 0.05023931199152954, "grad_norm": 0.5234375, "learning_rate": 0.0001674108750450126, "loss": 1.0084, "step": 4650 }, { "epoch": 0.050779519647352445, "grad_norm": 0.498046875, "learning_rate": 0.00016921137918617213, "loss": 1.0152, "step": 4700 }, { "epoch": 0.05131972730317534, "grad_norm": 0.48046875, "learning_rate": 0.00017101188332733166, "loss": 0.982, "step": 4750 }, { "epoch": 0.05185993495899824, "grad_norm": 0.53515625, "learning_rate": 0.00017281238746849118, "loss": 1.0199, "step": 4800 }, { "epoch": 0.052400142614821135, "grad_norm": 0.48828125, "learning_rate": 0.0001746128916096507, "loss": 1.0044, "step": 4850 }, { "epoch": 0.05294035027064404, "grad_norm": 0.5078125, "learning_rate": 0.00017641339575081024, "loss": 0.9955, "step": 4900 }, { "epoch": 0.053480557926466936, "grad_norm": 0.498046875, "learning_rate": 0.00017821389989196976, "loss": 0.9803, "step": 4950 }, { "epoch": 0.05402076558228983, "grad_norm": 0.443359375, "learning_rate": 0.0001800144040331293, "loss": 1.0071, "step": 5000 }, { "epoch": 0.05456097323811273, "grad_norm": 0.447265625, "learning_rate": 0.00018181490817428881, "loss": 0.9752, "step": 5050 }, { "epoch": 0.055101180893935627, "grad_norm": 0.474609375, "learning_rate": 0.00018361541231544834, "loss": 0.9567, "step": 5100 }, { "epoch": 0.05564138854975853, "grad_norm": 0.6171875, "learning_rate": 0.00018541591645660784, "loss": 0.9918, "step": 5150 }, { "epoch": 0.05618159620558143, "grad_norm": 0.458984375, "learning_rate": 0.0001872164205977674, "loss": 0.9632, "step": 5200 }, { "epoch": 0.056721803861404324, "grad_norm": 0.46484375, "learning_rate": 0.00018901692473892692, "loss": 0.9607, "step": 5250 }, { "epoch": 0.05726201151722722, "grad_norm": 0.51171875, "learning_rate": 0.00019081742888008642, "loss": 0.96, "step": 5300 }, { "epoch": 0.05780221917305012, "grad_norm": 0.474609375, "learning_rate": 0.00019261793302124595, "loss": 0.9752, "step": 5350 }, { "epoch": 0.05834242682887302, "grad_norm": 0.45703125, "learning_rate": 0.0001944184371624055, "loss": 0.9782, "step": 5400 }, { "epoch": 0.05888263448469592, "grad_norm": 0.498046875, "learning_rate": 0.00019621894130356503, "loss": 0.956, "step": 5450 }, { "epoch": 0.059422842140518815, "grad_norm": 0.466796875, "learning_rate": 0.00019801944544472453, "loss": 0.9549, "step": 5500 }, { "epoch": 0.05996304979634171, "grad_norm": 0.6171875, "learning_rate": 0.00019981994958588405, "loss": 0.9739, "step": 5550 }, { "epoch": 0.06050325745216461, "grad_norm": 0.4765625, "learning_rate": 0.00019999996900614756, "loss": 0.9687, "step": 5600 }, { "epoch": 0.06104346510798751, "grad_norm": 0.6171875, "learning_rate": 0.0001999998618669291, "loss": 0.97, "step": 5650 }, { "epoch": 0.06158367276381041, "grad_norm": 0.451171875, "learning_rate": 0.00019999967819978658, "loss": 0.9493, "step": 5700 }, { "epoch": 0.06212388041963331, "grad_norm": 0.447265625, "learning_rate": 0.00019999941800486043, "loss": 0.9411, "step": 5750 }, { "epoch": 0.0626640880754562, "grad_norm": 0.640625, "learning_rate": 0.00019999908128234985, "loss": 0.9662, "step": 5800 }, { "epoch": 0.0632042957312791, "grad_norm": 0.443359375, "learning_rate": 0.0001999986680325125, "loss": 0.9541, "step": 5850 }, { "epoch": 0.063744503387102, "grad_norm": 0.44921875, "learning_rate": 0.00019999817825566463, "loss": 0.9312, "step": 5900 }, { "epoch": 0.0642847110429249, "grad_norm": 0.45703125, "learning_rate": 0.00019999761195218106, "loss": 0.9492, "step": 5950 }, { "epoch": 0.0648249186987478, "grad_norm": 0.466796875, "learning_rate": 0.00019999696912249515, "loss": 0.9508, "step": 6000 } ], "logging_steps": 50, "max_steps": 185114, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.01130777853952e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }