| { | |
| "best_metric": 1.7699204683303833, | |
| "best_model_checkpoint": "gen_test/checkpoint-31250", | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 31250, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 173470.46875, | |
| "learning_rate": 5.9040000000000004e-05, | |
| "loss": 2.0402, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 170031.4375, | |
| "learning_rate": 5.808e-05, | |
| "loss": 1.8985, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 113037.2421875, | |
| "learning_rate": 5.712e-05, | |
| "loss": 1.8577, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 116889.296875, | |
| "learning_rate": 5.6160000000000004e-05, | |
| "loss": 1.8269, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 96221.234375, | |
| "learning_rate": 5.520000000000001e-05, | |
| "loss": 1.8069, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 81969.984375, | |
| "learning_rate": 5.424e-05, | |
| "loss": 1.7914, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 90032.578125, | |
| "learning_rate": 5.3280000000000005e-05, | |
| "loss": 1.7792, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 77830.828125, | |
| "learning_rate": 5.232e-05, | |
| "loss": 1.7666, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 65369.70703125, | |
| "learning_rate": 5.136e-05, | |
| "loss": 1.7696, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 77884.03125, | |
| "learning_rate": 5.04e-05, | |
| "loss": 1.7607, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 57507.2421875, | |
| "learning_rate": 4.944e-05, | |
| "loss": 1.7514, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 65232.98828125, | |
| "learning_rate": 4.8480000000000003e-05, | |
| "loss": 1.7409, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 58623.87109375, | |
| "learning_rate": 4.7520000000000006e-05, | |
| "loss": 1.7378, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 57649.65625, | |
| "learning_rate": 4.656e-05, | |
| "loss": 1.7372, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 61488.49609375, | |
| "learning_rate": 4.5600000000000004e-05, | |
| "loss": 1.7237, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 57711.09765625, | |
| "learning_rate": 4.464e-05, | |
| "loss": 1.7231, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 50364.1796875, | |
| "learning_rate": 4.368e-05, | |
| "loss": 1.7217, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 50957.51171875, | |
| "learning_rate": 4.272e-05, | |
| "loss": 1.7173, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 72140.7578125, | |
| "learning_rate": 4.176e-05, | |
| "loss": 1.7154, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 49330.06640625, | |
| "learning_rate": 4.08e-05, | |
| "loss": 1.7086, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 50961.6015625, | |
| "learning_rate": 3.9840000000000005e-05, | |
| "loss": 1.7096, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 48586.94921875, | |
| "learning_rate": 3.888e-05, | |
| "loss": 1.7041, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 55556.12890625, | |
| "learning_rate": 3.792e-05, | |
| "loss": 1.7032, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 52399.66796875, | |
| "learning_rate": 3.696e-05, | |
| "loss": 1.6997, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 48360.08203125, | |
| "learning_rate": 3.6e-05, | |
| "loss": 1.6936, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 44732.66015625, | |
| "learning_rate": 3.5039999999999997e-05, | |
| "loss": 1.6921, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 54025.0390625, | |
| "learning_rate": 3.408e-05, | |
| "loss": 1.6905, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 54268.91796875, | |
| "learning_rate": 3.312e-05, | |
| "loss": 1.6876, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 46711.734375, | |
| "learning_rate": 3.2160000000000004e-05, | |
| "loss": 1.688, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 47818.2265625, | |
| "learning_rate": 3.12e-05, | |
| "loss": 1.6817, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 49987.921875, | |
| "learning_rate": 3.0240000000000002e-05, | |
| "loss": 1.6786, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 46060.59375, | |
| "learning_rate": 2.928e-05, | |
| "loss": 1.6736, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 44150.3671875, | |
| "learning_rate": 2.832e-05, | |
| "loss": 1.6753, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 43254.55859375, | |
| "learning_rate": 2.7360000000000002e-05, | |
| "loss": 1.6732, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 50895.7421875, | |
| "learning_rate": 2.64e-05, | |
| "loss": 1.6727, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 49795.7421875, | |
| "learning_rate": 2.544e-05, | |
| "loss": 1.6682, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 57421.609375, | |
| "learning_rate": 2.448e-05, | |
| "loss": 1.6673, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 44257.9296875, | |
| "learning_rate": 2.3520000000000002e-05, | |
| "loss": 1.6651, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 48974.69921875, | |
| "learning_rate": 2.256e-05, | |
| "loss": 1.6608, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 61821.53125, | |
| "learning_rate": 2.16e-05, | |
| "loss": 1.6637, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 51516.53515625, | |
| "learning_rate": 2.064e-05, | |
| "loss": 1.6628, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 50740.87890625, | |
| "learning_rate": 1.968e-05, | |
| "loss": 1.6651, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 49265.2890625, | |
| "learning_rate": 1.872e-05, | |
| "loss": 1.6604, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 44264.171875, | |
| "learning_rate": 1.776e-05, | |
| "loss": 1.6557, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 50852.16015625, | |
| "learning_rate": 1.6800000000000002e-05, | |
| "loss": 1.6579, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 58869.0625, | |
| "learning_rate": 1.584e-05, | |
| "loss": 1.655, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 41634.01953125, | |
| "learning_rate": 1.488e-05, | |
| "loss": 1.6533, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 42209.58984375, | |
| "learning_rate": 1.392e-05, | |
| "loss": 1.6506, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 48772.890625, | |
| "learning_rate": 1.296e-05, | |
| "loss": 1.6501, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 47330.41796875, | |
| "learning_rate": 1.2e-05, | |
| "loss": 1.65, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 46274.1953125, | |
| "learning_rate": 1.104e-05, | |
| "loss": 1.6465, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 48066.18359375, | |
| "learning_rate": 1.008e-05, | |
| "loss": 1.6462, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 44796.63671875, | |
| "learning_rate": 9.12e-06, | |
| "loss": 1.6491, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 51596.578125, | |
| "learning_rate": 8.160000000000001e-06, | |
| "loss": 1.6485, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 42668.67578125, | |
| "learning_rate": 7.2e-06, | |
| "loss": 1.6413, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 48424.36328125, | |
| "learning_rate": 6.2399999999999995e-06, | |
| "loss": 1.6477, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 50913.546875, | |
| "learning_rate": 5.279999999999999e-06, | |
| "loss": 1.6455, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 49008.09375, | |
| "learning_rate": 4.32e-06, | |
| "loss": 1.644, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 43177.625, | |
| "learning_rate": 3.36e-06, | |
| "loss": 1.6416, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 49836.08984375, | |
| "learning_rate": 2.4000000000000003e-06, | |
| "loss": 1.6419, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 49725.71484375, | |
| "learning_rate": 1.44e-06, | |
| "loss": 1.6402, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 44681.953125, | |
| "learning_rate": 4.800000000000001e-07, | |
| "loss": 1.6425, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.7699204683303833, | |
| "eval_runtime": 56.1676, | |
| "eval_samples_per_second": 178.039, | |
| "eval_steps_per_second": 2.795, | |
| "step": 31250 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 31250, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "total_flos": 1.045167082831872e+18, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |