| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.08347245409015025, |
| "eval_steps": 500, |
| "global_step": 5000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0008347245409015025, |
| "grad_norm": 5.03125, |
| "learning_rate": 0.00019800000000000002, |
| "loss": 1.2979, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.001669449081803005, |
| "grad_norm": 3.734375, |
| "learning_rate": 0.000196, |
| "loss": 0.8508, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.0025041736227045075, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.000194, |
| "loss": 0.7959, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.00333889816360601, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.000192, |
| "loss": 0.726, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.004173622704507512, |
| "grad_norm": 0.328125, |
| "learning_rate": 0.00019, |
| "loss": 0.9015, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.005008347245409015, |
| "grad_norm": 69.5, |
| "learning_rate": 0.000188, |
| "loss": 0.4003, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.005843071786310518, |
| "grad_norm": 3.015625, |
| "learning_rate": 0.00018600000000000002, |
| "loss": 0.5518, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.00667779632721202, |
| "grad_norm": 0.034912109375, |
| "learning_rate": 0.00018400000000000003, |
| "loss": 0.4198, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.007512520868113523, |
| "grad_norm": 5.90625, |
| "learning_rate": 0.000182, |
| "loss": 0.8151, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.008347245409015025, |
| "grad_norm": 17.125, |
| "learning_rate": 0.00018, |
| "loss": 0.5379, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.009181969949916527, |
| "grad_norm": 0.0654296875, |
| "learning_rate": 0.00017800000000000002, |
| "loss": 0.5603, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.01001669449081803, |
| "grad_norm": 0.10400390625, |
| "learning_rate": 0.00017600000000000002, |
| "loss": 0.7589, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.010851419031719533, |
| "grad_norm": 0.265625, |
| "learning_rate": 0.000174, |
| "loss": 0.3531, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.011686143572621035, |
| "grad_norm": 4.59375, |
| "learning_rate": 0.000172, |
| "loss": 0.2717, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.012520868113522538, |
| "grad_norm": 0.25390625, |
| "learning_rate": 0.00017, |
| "loss": 0.3344, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.01335559265442404, |
| "grad_norm": 4.6875, |
| "learning_rate": 0.000168, |
| "loss": 0.3585, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.014190317195325543, |
| "grad_norm": 0.04736328125, |
| "learning_rate": 0.000166, |
| "loss": 0.3218, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.015025041736227046, |
| "grad_norm": 0.046875, |
| "learning_rate": 0.000164, |
| "loss": 0.3344, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.015859766277128547, |
| "grad_norm": 0.015625, |
| "learning_rate": 0.000162, |
| "loss": 0.4233, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.01669449081803005, |
| "grad_norm": 0.08056640625, |
| "learning_rate": 0.00016, |
| "loss": 0.6014, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.017529215358931552, |
| "grad_norm": 3.390625, |
| "learning_rate": 0.00015800000000000002, |
| "loss": 0.4044, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.018363939899833055, |
| "grad_norm": 0.31640625, |
| "learning_rate": 0.00015600000000000002, |
| "loss": 0.9183, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.019198664440734557, |
| "grad_norm": 0.076171875, |
| "learning_rate": 0.000154, |
| "loss": 0.2508, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.02003338898163606, |
| "grad_norm": 0.01055908203125, |
| "learning_rate": 0.000152, |
| "loss": 0.3563, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.020868113522537562, |
| "grad_norm": 54.0, |
| "learning_rate": 0.00015000000000000001, |
| "loss": 0.4328, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.021702838063439065, |
| "grad_norm": 19.5, |
| "learning_rate": 0.000148, |
| "loss": 0.312, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.022537562604340568, |
| "grad_norm": 0.0380859375, |
| "learning_rate": 0.000146, |
| "loss": 0.3068, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.02337228714524207, |
| "grad_norm": 0.006622314453125, |
| "learning_rate": 0.000144, |
| "loss": 0.3987, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.024207011686143573, |
| "grad_norm": 8.0625, |
| "learning_rate": 0.000142, |
| "loss": 0.4942, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.025041736227045076, |
| "grad_norm": 26.125, |
| "learning_rate": 0.00014, |
| "loss": 0.4011, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.02587646076794658, |
| "grad_norm": 0.58203125, |
| "learning_rate": 0.000138, |
| "loss": 0.5129, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.02671118530884808, |
| "grad_norm": 0.0267333984375, |
| "learning_rate": 0.00013600000000000003, |
| "loss": 0.1984, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.027545909849749584, |
| "grad_norm": 0.0157470703125, |
| "learning_rate": 0.000134, |
| "loss": 0.4052, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.028380634390651086, |
| "grad_norm": 4.5625, |
| "learning_rate": 0.000132, |
| "loss": 0.4439, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.02921535893155259, |
| "grad_norm": 21.125, |
| "learning_rate": 0.00013000000000000002, |
| "loss": 0.5354, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.03005008347245409, |
| "grad_norm": 15.375, |
| "learning_rate": 0.00012800000000000002, |
| "loss": 0.5931, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.030884808013355594, |
| "grad_norm": 1.3515625, |
| "learning_rate": 0.000126, |
| "loss": 0.1363, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.03171953255425709, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.000124, |
| "loss": 0.4885, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.0325542570951586, |
| "grad_norm": 0.0908203125, |
| "learning_rate": 0.000122, |
| "loss": 0.265, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.0333889816360601, |
| "grad_norm": 0.0142822265625, |
| "learning_rate": 0.00012, |
| "loss": 0.265, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.034223706176961605, |
| "grad_norm": 3.25, |
| "learning_rate": 0.000118, |
| "loss": 0.5515, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.035058430717863104, |
| "grad_norm": 3.5, |
| "learning_rate": 0.000116, |
| "loss": 0.3973, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.03589315525876461, |
| "grad_norm": 1.0703125, |
| "learning_rate": 0.00011399999999999999, |
| "loss": 0.4916, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.03672787979966611, |
| "grad_norm": 0.26171875, |
| "learning_rate": 0.00011200000000000001, |
| "loss": 0.4639, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.037562604340567615, |
| "grad_norm": 0.10693359375, |
| "learning_rate": 0.00011000000000000002, |
| "loss": 0.4581, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.038397328881469114, |
| "grad_norm": 5.75, |
| "learning_rate": 0.00010800000000000001, |
| "loss": 0.3952, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.03923205342237062, |
| "grad_norm": 0.058837890625, |
| "learning_rate": 0.00010600000000000002, |
| "loss": 0.5916, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.04006677796327212, |
| "grad_norm": 69.5, |
| "learning_rate": 0.00010400000000000001, |
| "loss": 0.394, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.040901502504173626, |
| "grad_norm": 0.04833984375, |
| "learning_rate": 0.00010200000000000001, |
| "loss": 0.2667, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.041736227045075125, |
| "grad_norm": 0.005615234375, |
| "learning_rate": 0.0001, |
| "loss": 0.4901, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.04257095158597663, |
| "grad_norm": 0.004180908203125, |
| "learning_rate": 9.8e-05, |
| "loss": 0.3322, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.04340567612687813, |
| "grad_norm": 0.048095703125, |
| "learning_rate": 9.6e-05, |
| "loss": 0.3777, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.04424040066777963, |
| "grad_norm": 0.08447265625, |
| "learning_rate": 9.4e-05, |
| "loss": 0.4049, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.045075125208681135, |
| "grad_norm": 0.0712890625, |
| "learning_rate": 9.200000000000001e-05, |
| "loss": 0.5967, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.045909849749582635, |
| "grad_norm": 5.78125, |
| "learning_rate": 9e-05, |
| "loss": 0.2451, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.04674457429048414, |
| "grad_norm": 0.0081787109375, |
| "learning_rate": 8.800000000000001e-05, |
| "loss": 0.4497, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.04757929883138564, |
| "grad_norm": 0.0260009765625, |
| "learning_rate": 8.6e-05, |
| "loss": 0.4576, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.048414023372287146, |
| "grad_norm": 0.01025390625, |
| "learning_rate": 8.4e-05, |
| "loss": 0.4427, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.049248747913188645, |
| "grad_norm": 0.04296875, |
| "learning_rate": 8.2e-05, |
| "loss": 0.4576, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.05008347245409015, |
| "grad_norm": 0.0634765625, |
| "learning_rate": 8e-05, |
| "loss": 0.3468, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.05091819699499165, |
| "grad_norm": 14.0625, |
| "learning_rate": 7.800000000000001e-05, |
| "loss": 0.491, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.05175292153589316, |
| "grad_norm": 0.0277099609375, |
| "learning_rate": 7.6e-05, |
| "loss": 0.3008, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.052587646076794656, |
| "grad_norm": 0.03466796875, |
| "learning_rate": 7.4e-05, |
| "loss": 0.3269, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.05342237061769616, |
| "grad_norm": 0.024169921875, |
| "learning_rate": 7.2e-05, |
| "loss": 0.4641, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.05425709515859766, |
| "grad_norm": 0.02099609375, |
| "learning_rate": 7e-05, |
| "loss": 0.2215, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.05509181969949917, |
| "grad_norm": 0.04833984375, |
| "learning_rate": 6.800000000000001e-05, |
| "loss": 0.3736, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.055926544240400666, |
| "grad_norm": 0.040283203125, |
| "learning_rate": 6.6e-05, |
| "loss": 0.114, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.05676126878130217, |
| "grad_norm": 0.45703125, |
| "learning_rate": 6.400000000000001e-05, |
| "loss": 0.2316, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.05759599332220367, |
| "grad_norm": 0.06982421875, |
| "learning_rate": 6.2e-05, |
| "loss": 0.6308, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.05843071786310518, |
| "grad_norm": 2.765625, |
| "learning_rate": 6e-05, |
| "loss": 0.2929, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.05926544240400668, |
| "grad_norm": 0.06396484375, |
| "learning_rate": 5.8e-05, |
| "loss": 0.5138, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.06010016694490818, |
| "grad_norm": 0.07861328125, |
| "learning_rate": 5.6000000000000006e-05, |
| "loss": 0.5067, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.06093489148580968, |
| "grad_norm": 3.875, |
| "learning_rate": 5.4000000000000005e-05, |
| "loss": 0.6174, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.06176961602671119, |
| "grad_norm": 0.17578125, |
| "learning_rate": 5.2000000000000004e-05, |
| "loss": 0.3295, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.06260434056761269, |
| "grad_norm": 0.08251953125, |
| "learning_rate": 5e-05, |
| "loss": 0.3084, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.06343906510851419, |
| "grad_norm": 0.33203125, |
| "learning_rate": 4.8e-05, |
| "loss": 0.3046, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.0642737896494157, |
| "grad_norm": 0.126953125, |
| "learning_rate": 4.600000000000001e-05, |
| "loss": 0.1521, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.0651085141903172, |
| "grad_norm": 0.08447265625, |
| "learning_rate": 4.4000000000000006e-05, |
| "loss": 0.4219, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.0659432387312187, |
| "grad_norm": 105.5, |
| "learning_rate": 4.2e-05, |
| "loss": 0.2136, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.0667779632721202, |
| "grad_norm": 0.0830078125, |
| "learning_rate": 4e-05, |
| "loss": 0.1255, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.0676126878130217, |
| "grad_norm": 3.84375, |
| "learning_rate": 3.8e-05, |
| "loss": 0.3222, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.06844741235392321, |
| "grad_norm": 0.04833984375, |
| "learning_rate": 3.6e-05, |
| "loss": 0.4844, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.06928213689482471, |
| "grad_norm": 3.65625, |
| "learning_rate": 3.4000000000000007e-05, |
| "loss": 0.2207, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.07011686143572621, |
| "grad_norm": 0.0478515625, |
| "learning_rate": 3.2000000000000005e-05, |
| "loss": 0.2303, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.0709515859766277, |
| "grad_norm": 0.10595703125, |
| "learning_rate": 3e-05, |
| "loss": 0.3358, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.07178631051752922, |
| "grad_norm": 0.08349609375, |
| "learning_rate": 2.8000000000000003e-05, |
| "loss": 0.6106, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.07262103505843072, |
| "grad_norm": 0.09130859375, |
| "learning_rate": 2.6000000000000002e-05, |
| "loss": 0.4787, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.07345575959933222, |
| "grad_norm": 0.042236328125, |
| "learning_rate": 2.4e-05, |
| "loss": 0.145, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.07429048414023372, |
| "grad_norm": 0.020263671875, |
| "learning_rate": 2.2000000000000003e-05, |
| "loss": 0.3278, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.07512520868113523, |
| "grad_norm": 0.00701904296875, |
| "learning_rate": 2e-05, |
| "loss": 0.3238, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.07595993322203673, |
| "grad_norm": 0.01080322265625, |
| "learning_rate": 1.8e-05, |
| "loss": 0.3776, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.07679465776293823, |
| "grad_norm": 0.049072265625, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 0.3402, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.07762938230383973, |
| "grad_norm": 0.0712890625, |
| "learning_rate": 1.4000000000000001e-05, |
| "loss": 0.429, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.07846410684474124, |
| "grad_norm": 0.53125, |
| "learning_rate": 1.2e-05, |
| "loss": 0.5878, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.07929883138564274, |
| "grad_norm": 3.828125, |
| "learning_rate": 1e-05, |
| "loss": 0.4823, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.08013355592654424, |
| "grad_norm": 0.059326171875, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 0.1802, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.08096828046744574, |
| "grad_norm": 0.00628662109375, |
| "learning_rate": 6e-06, |
| "loss": 0.4154, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.08180300500834725, |
| "grad_norm": 0.06005859375, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 0.295, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.08263772954924875, |
| "grad_norm": 0.01287841796875, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 0.3699, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.08347245409015025, |
| "grad_norm": 0.283203125, |
| "learning_rate": 0.0, |
| "loss": 0.3094, |
| "step": 5000 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 5000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.145082707738624e+16, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|