{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.976, "eval_steps": 500, "global_step": 93, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.032, "grad_norm": 6.017088211518867, "learning_rate": 1.0000000000000002e-06, "loss": 0.8466, "step": 1 }, { "epoch": 0.064, "grad_norm": 5.951857736161836, "learning_rate": 2.0000000000000003e-06, "loss": 0.8716, "step": 2 }, { "epoch": 0.096, "grad_norm": 5.7077662516941885, "learning_rate": 3e-06, "loss": 0.8566, "step": 3 }, { "epoch": 0.128, "grad_norm": 5.833985690256031, "learning_rate": 4.000000000000001e-06, "loss": 0.8807, "step": 4 }, { "epoch": 0.16, "grad_norm": 4.601869768823213, "learning_rate": 5e-06, "loss": 0.7739, "step": 5 }, { "epoch": 0.192, "grad_norm": 2.656272883782473, "learning_rate": 6e-06, "loss": 0.8025, "step": 6 }, { "epoch": 0.224, "grad_norm": 2.191913269298967, "learning_rate": 7e-06, "loss": 0.7636, "step": 7 }, { "epoch": 0.256, "grad_norm": 4.059469817937578, "learning_rate": 8.000000000000001e-06, "loss": 0.7703, "step": 8 }, { "epoch": 0.288, "grad_norm": 4.482509294996795, "learning_rate": 9e-06, "loss": 0.7323, "step": 9 }, { "epoch": 0.32, "grad_norm": 5.2896734327529975, "learning_rate": 1e-05, "loss": 0.7473, "step": 10 }, { "epoch": 0.352, "grad_norm": 5.648561404857537, "learning_rate": 9.996418774081658e-06, "loss": 0.7653, "step": 11 }, { "epoch": 0.384, "grad_norm": 4.49874465390001, "learning_rate": 9.985680226398261e-06, "loss": 0.7378, "step": 12 }, { "epoch": 0.416, "grad_norm": 2.5886351134391057, "learning_rate": 9.967799739815925e-06, "loss": 0.6586, "step": 13 }, { "epoch": 0.448, "grad_norm": 2.170217772669203, "learning_rate": 9.942802927959444e-06, "loss": 0.7204, "step": 14 }, { "epoch": 0.48, "grad_norm": 2.5221677834113967, "learning_rate": 9.910725598521014e-06, "loss": 0.6305, "step": 15 }, { "epoch": 0.512, "grad_norm": 2.4169452505699134, "learning_rate": 9.871613701966067e-06, "loss": 0.6628, "step": 16 }, { "epoch": 0.544, "grad_norm": 2.1361025544158485, "learning_rate": 9.825523265709667e-06, "loss": 0.6755, "step": 17 }, { "epoch": 0.576, "grad_norm": 1.256643746613112, "learning_rate": 9.772520313857777e-06, "loss": 0.6011, "step": 18 }, { "epoch": 0.608, "grad_norm": 1.5131065345982715, "learning_rate": 9.712680772628365e-06, "loss": 0.6595, "step": 19 }, { "epoch": 0.64, "grad_norm": 1.5554495037101939, "learning_rate": 9.646090361587828e-06, "loss": 0.5806, "step": 20 }, { "epoch": 0.672, "grad_norm": 1.479197692013354, "learning_rate": 9.572844470858537e-06, "loss": 0.6069, "step": 21 }, { "epoch": 0.704, "grad_norm": 1.09952905385751, "learning_rate": 9.493048024473413e-06, "loss": 0.6094, "step": 22 }, { "epoch": 0.736, "grad_norm": 1.0257031667969818, "learning_rate": 9.406815330073244e-06, "loss": 0.5873, "step": 23 }, { "epoch": 0.768, "grad_norm": 1.3397890175148242, "learning_rate": 9.314269915162115e-06, "loss": 0.6475, "step": 24 }, { "epoch": 0.8, "grad_norm": 1.1987888823222785, "learning_rate": 9.215544350155423e-06, "loss": 0.5892, "step": 25 }, { "epoch": 0.832, "grad_norm": 0.987671863849975, "learning_rate": 9.110780058474052e-06, "loss": 0.6205, "step": 26 }, { "epoch": 0.864, "grad_norm": 1.0538210475650935, "learning_rate": 9.000127113956673e-06, "loss": 0.635, "step": 27 }, { "epoch": 0.896, "grad_norm": 0.9785904516879229, "learning_rate": 8.883744025880429e-06, "loss": 0.592, "step": 28 }, { "epoch": 0.928, "grad_norm": 1.0498722982137005, "learning_rate": 8.761797511897907e-06, "loss": 0.6359, "step": 29 }, { "epoch": 0.96, "grad_norm": 0.9942472305105629, "learning_rate": 8.634462259215719e-06, "loss": 0.614, "step": 30 }, { "epoch": 0.992, "grad_norm": 1.047896159694581, "learning_rate": 8.501920674356755e-06, "loss": 0.5893, "step": 31 }, { "epoch": 1.024, "grad_norm": 0.8393117657377732, "learning_rate": 8.364362621864595e-06, "loss": 0.5501, "step": 32 }, { "epoch": 1.056, "grad_norm": 0.8368148848259993, "learning_rate": 8.221985152324385e-06, "loss": 0.5004, "step": 33 }, { "epoch": 1.088, "grad_norm": 1.0449083783998578, "learning_rate": 8.07499222008977e-06, "loss": 0.5262, "step": 34 }, { "epoch": 1.12, "grad_norm": 0.9498527409166139, "learning_rate": 7.923594391120237e-06, "loss": 0.5087, "step": 35 }, { "epoch": 1.152, "grad_norm": 0.9204978138749593, "learning_rate": 7.768008541347423e-06, "loss": 0.5083, "step": 36 }, { "epoch": 1.184, "grad_norm": 0.9742025167645811, "learning_rate": 7.608457546002423e-06, "loss": 0.5109, "step": 37 }, { "epoch": 1.216, "grad_norm": 0.9820066122143393, "learning_rate": 7.445169960349167e-06, "loss": 0.542, "step": 38 }, { "epoch": 1.248, "grad_norm": 0.8731760697551055, "learning_rate": 7.278379692281209e-06, "loss": 0.4646, "step": 39 }, { "epoch": 1.28, "grad_norm": 0.8552117765440119, "learning_rate": 7.10832566725092e-06, "loss": 0.4691, "step": 40 }, { "epoch": 1.312, "grad_norm": 1.0485886921584444, "learning_rate": 6.9352514860110876e-06, "loss": 0.4867, "step": 41 }, { "epoch": 1.3439999999999999, "grad_norm": 0.8054399255100043, "learning_rate": 6.759405075659165e-06, "loss": 0.4988, "step": 42 }, { "epoch": 1.376, "grad_norm": 0.8137660425234108, "learning_rate": 6.58103833448412e-06, "loss": 0.4943, "step": 43 }, { "epoch": 1.408, "grad_norm": 0.9191572909893893, "learning_rate": 6.4004067711245366e-06, "loss": 0.5099, "step": 44 }, { "epoch": 1.44, "grad_norm": 0.8579377944173575, "learning_rate": 6.2177691385549595e-06, "loss": 0.4974, "step": 45 }, { "epoch": 1.472, "grad_norm": 0.8482605744927274, "learning_rate": 6.033387063424765e-06, "loss": 0.4619, "step": 46 }, { "epoch": 1.504, "grad_norm": 0.7598606051104131, "learning_rate": 5.8475246712804845e-06, "loss": 0.4977, "step": 47 }, { "epoch": 1.536, "grad_norm": 0.7248290008384539, "learning_rate": 5.660448208208513e-06, "loss": 0.4893, "step": 48 }, { "epoch": 1.568, "grad_norm": 0.8290652695699064, "learning_rate": 5.472425659440157e-06, "loss": 0.4997, "step": 49 }, { "epoch": 1.6, "grad_norm": 0.8445778166933251, "learning_rate": 5.2837263654653715e-06, "loss": 0.4802, "step": 50 }, { "epoch": 1.6320000000000001, "grad_norm": 0.8222728417363387, "learning_rate": 5.094620636205096e-06, "loss": 0.5016, "step": 51 }, { "epoch": 1.6640000000000001, "grad_norm": 0.7237424288555369, "learning_rate": 4.905379363794907e-06, "loss": 0.4943, "step": 52 }, { "epoch": 1.696, "grad_norm": 0.8333802083768971, "learning_rate": 4.71627363453463e-06, "loss": 0.491, "step": 53 }, { "epoch": 1.728, "grad_norm": 0.6818057424511519, "learning_rate": 4.527574340559844e-06, "loss": 0.4378, "step": 54 }, { "epoch": 1.76, "grad_norm": 0.8299127275023862, "learning_rate": 4.33955179179149e-06, "loss": 0.4658, "step": 55 }, { "epoch": 1.792, "grad_norm": 0.8405326607955101, "learning_rate": 4.152475328719517e-06, "loss": 0.5099, "step": 56 }, { "epoch": 1.8239999999999998, "grad_norm": 0.7492090483975039, "learning_rate": 3.966612936575235e-06, "loss": 0.4825, "step": 57 }, { "epoch": 1.8559999999999999, "grad_norm": 0.7030190039796792, "learning_rate": 3.782230861445041e-06, "loss": 0.4644, "step": 58 }, { "epoch": 1.888, "grad_norm": 0.6811613244860312, "learning_rate": 3.5995932288754655e-06, "loss": 0.4257, "step": 59 }, { "epoch": 1.92, "grad_norm": 0.7603335654736988, "learning_rate": 3.4189616655158803e-06, "loss": 0.4788, "step": 60 }, { "epoch": 1.952, "grad_norm": 0.7077292083579106, "learning_rate": 3.240594924340835e-06, "loss": 0.4679, "step": 61 }, { "epoch": 1.984, "grad_norm": 0.8338120865067291, "learning_rate": 3.0647485139889145e-06, "loss": 0.5145, "step": 62 }, { "epoch": 2.016, "grad_norm": 0.6135927406398123, "learning_rate": 2.89167433274908e-06, "loss": 0.4013, "step": 63 }, { "epoch": 2.048, "grad_norm": 0.7706602310876762, "learning_rate": 2.721620307718793e-06, "loss": 0.4279, "step": 64 }, { "epoch": 2.08, "grad_norm": 0.6693415795433767, "learning_rate": 2.554830039650834e-06, "loss": 0.4074, "step": 65 }, { "epoch": 2.112, "grad_norm": 0.6646712904057168, "learning_rate": 2.391542453997578e-06, "loss": 0.3972, "step": 66 }, { "epoch": 2.144, "grad_norm": 0.6277893548891085, "learning_rate": 2.2319914586525776e-06, "loss": 0.4317, "step": 67 }, { "epoch": 2.176, "grad_norm": 0.6861006207092378, "learning_rate": 2.0764056088797646e-06, "loss": 0.3826, "step": 68 }, { "epoch": 2.208, "grad_norm": 0.640005258117782, "learning_rate": 1.9250077799102323e-06, "loss": 0.4041, "step": 69 }, { "epoch": 2.24, "grad_norm": 0.6248654175061756, "learning_rate": 1.7780148476756148e-06, "loss": 0.4387, "step": 70 }, { "epoch": 2.2720000000000002, "grad_norm": 0.6670800065284683, "learning_rate": 1.6356373781354058e-06, "loss": 0.4334, "step": 71 }, { "epoch": 2.304, "grad_norm": 0.6112441982790366, "learning_rate": 1.4980793256432474e-06, "loss": 0.4368, "step": 72 }, { "epoch": 2.336, "grad_norm": 0.6820462674363659, "learning_rate": 1.3655377407842813e-06, "loss": 0.4073, "step": 73 }, { "epoch": 2.368, "grad_norm": 0.5637152749781639, "learning_rate": 1.2382024881020937e-06, "loss": 0.4026, "step": 74 }, { "epoch": 2.4, "grad_norm": 0.62501378989424, "learning_rate": 1.1162559741195733e-06, "loss": 0.3788, "step": 75 }, { "epoch": 2.432, "grad_norm": 0.5909257721151473, "learning_rate": 9.998728860433277e-07, "loss": 0.414, "step": 76 }, { "epoch": 2.464, "grad_norm": 0.59961756687512, "learning_rate": 8.892199415259501e-07, "loss": 0.3815, "step": 77 }, { "epoch": 2.496, "grad_norm": 0.6327449480704133, "learning_rate": 7.844556498445788e-07, "loss": 0.3731, "step": 78 }, { "epoch": 2.528, "grad_norm": 0.6163817718459151, "learning_rate": 6.857300848378857e-07, "loss": 0.4096, "step": 79 }, { "epoch": 2.56, "grad_norm": 0.614006560513663, "learning_rate": 5.931846699267558e-07, "loss": 0.4437, "step": 80 }, { "epoch": 2.592, "grad_norm": 0.6432304424213249, "learning_rate": 5.0695197552659e-07, "loss": 0.4076, "step": 81 }, { "epoch": 2.624, "grad_norm": 0.6064212963275805, "learning_rate": 4.271555291414636e-07, "loss": 0.3967, "step": 82 }, { "epoch": 2.656, "grad_norm": 0.6049906222701497, "learning_rate": 3.539096384121743e-07, "loss": 0.3843, "step": 83 }, { "epoch": 2.6879999999999997, "grad_norm": 0.5757860975137236, "learning_rate": 2.873192273716369e-07, "loss": 0.3836, "step": 84 }, { "epoch": 2.7199999999999998, "grad_norm": 0.6129350159111668, "learning_rate": 2.274796861422246e-07, "loss": 0.4543, "step": 85 }, { "epoch": 2.752, "grad_norm": 0.6388122099182162, "learning_rate": 1.7447673429033361e-07, "loss": 0.3527, "step": 86 }, { "epoch": 2.784, "grad_norm": 0.6016188801263589, "learning_rate": 1.2838629803393343e-07, "loss": 0.3922, "step": 87 }, { "epoch": 2.816, "grad_norm": 0.5819716193194474, "learning_rate": 8.927440147898703e-08, "loss": 0.3972, "step": 88 }, { "epoch": 2.848, "grad_norm": 0.5860665357952272, "learning_rate": 5.7197072040557356e-08, "loss": 0.4262, "step": 89 }, { "epoch": 2.88, "grad_norm": 0.6249229708634069, "learning_rate": 3.220026018407541e-08, "loss": 0.4841, "step": 90 }, { "epoch": 2.912, "grad_norm": 0.5591758817715189, "learning_rate": 1.431977360173975e-08, "loss": 0.3779, "step": 91 }, { "epoch": 2.944, "grad_norm": 0.6402184136731239, "learning_rate": 3.5812259183426457e-09, "loss": 0.4023, "step": 92 }, { "epoch": 2.976, "grad_norm": 0.6090640200163872, "learning_rate": 0.0, "loss": 0.3697, "step": 93 }, { "epoch": 2.976, "step": 93, "total_flos": 5.107931021941146e+16, "train_loss": 0.5300623857846825, "train_runtime": 4609.5522, "train_samples_per_second": 0.648, "train_steps_per_second": 0.02 } ], "logging_steps": 1, "max_steps": 93, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.107931021941146e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }