| { | |
| "best_metric": 0.8724454641342163, | |
| "best_model_checkpoint": "./results/checkpoint-2700", | |
| "epoch": 7.0, | |
| "eval_steps": 500, | |
| "global_step": 18900, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.18518518518518517, | |
| "grad_norm": 2.5104269981384277, | |
| "learning_rate": 4.94212962962963e-05, | |
| "loss": 0.9092, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.37037037037037035, | |
| "grad_norm": 1.030969262123108, | |
| "learning_rate": 4.8842592592592595e-05, | |
| "loss": 0.869, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.5555555555555556, | |
| "grad_norm": 0.49540138244628906, | |
| "learning_rate": 4.8263888888888895e-05, | |
| "loss": 0.8595, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.7407407407407407, | |
| "grad_norm": 0.6758684515953064, | |
| "learning_rate": 4.768518518518519e-05, | |
| "loss": 0.8446, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.9259259259259259, | |
| "grad_norm": 1.0114092826843262, | |
| "learning_rate": 4.710648148148149e-05, | |
| "loss": 0.8403, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.8724454641342163, | |
| "eval_runtime": 30.4341, | |
| "eval_samples_per_second": 39.43, | |
| "eval_steps_per_second": 9.857, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.1111111111111112, | |
| "grad_norm": 0.8358873128890991, | |
| "learning_rate": 4.652777777777778e-05, | |
| "loss": 0.8109, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.2962962962962963, | |
| "grad_norm": 0.8434183597564697, | |
| "learning_rate": 4.594907407407408e-05, | |
| "loss": 0.7912, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.4814814814814814, | |
| "grad_norm": 1.2122215032577515, | |
| "learning_rate": 4.5370370370370374e-05, | |
| "loss": 0.7891, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 0.7747897505760193, | |
| "learning_rate": 4.4791666666666673e-05, | |
| "loss": 0.8132, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.8518518518518519, | |
| "grad_norm": 0.9554975032806396, | |
| "learning_rate": 4.4212962962962966e-05, | |
| "loss": 0.8047, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.8918996453285217, | |
| "eval_runtime": 30.4698, | |
| "eval_samples_per_second": 39.383, | |
| "eval_steps_per_second": 9.846, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 2.037037037037037, | |
| "grad_norm": 0.6646651029586792, | |
| "learning_rate": 4.3634259259259266e-05, | |
| "loss": 0.8091, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.2222222222222223, | |
| "grad_norm": 4.084255218505859, | |
| "learning_rate": 4.305555555555556e-05, | |
| "loss": 0.7432, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.4074074074074074, | |
| "grad_norm": 2.1203970909118652, | |
| "learning_rate": 4.247685185185186e-05, | |
| "loss": 0.7355, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.5925925925925926, | |
| "grad_norm": 4.367093086242676, | |
| "learning_rate": 4.1898148148148145e-05, | |
| "loss": 0.751, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 2.7777777777777777, | |
| "grad_norm": 1.1563506126403809, | |
| "learning_rate": 4.1319444444444445e-05, | |
| "loss": 0.7632, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 2.962962962962963, | |
| "grad_norm": 0.9595785140991211, | |
| "learning_rate": 4.074074074074074e-05, | |
| "loss": 0.7714, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.9504669308662415, | |
| "eval_runtime": 30.5135, | |
| "eval_samples_per_second": 39.327, | |
| "eval_steps_per_second": 9.832, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 3.148148148148148, | |
| "grad_norm": 0.6189069747924805, | |
| "learning_rate": 4.016203703703704e-05, | |
| "loss": 0.7277, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 3.3333333333333335, | |
| "grad_norm": 0.6079156994819641, | |
| "learning_rate": 3.958333333333333e-05, | |
| "loss": 0.7373, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 3.5185185185185186, | |
| "grad_norm": 0.4996514320373535, | |
| "learning_rate": 3.900462962962963e-05, | |
| "loss": 0.7206, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 3.7037037037037037, | |
| "grad_norm": 0.9442146420478821, | |
| "learning_rate": 3.8425925925925924e-05, | |
| "loss": 0.7215, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 3.888888888888889, | |
| "grad_norm": 0.46321621537208557, | |
| "learning_rate": 3.7847222222222224e-05, | |
| "loss": 0.7238, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 1.0164023637771606, | |
| "eval_runtime": 30.4613, | |
| "eval_samples_per_second": 39.394, | |
| "eval_steps_per_second": 9.849, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 4.074074074074074, | |
| "grad_norm": 0.808857798576355, | |
| "learning_rate": 3.726851851851852e-05, | |
| "loss": 0.7143, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 4.2592592592592595, | |
| "grad_norm": 0.2266809195280075, | |
| "learning_rate": 3.6689814814814816e-05, | |
| "loss": 0.7017, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 4.444444444444445, | |
| "grad_norm": 0.8129966259002686, | |
| "learning_rate": 3.611111111111111e-05, | |
| "loss": 0.7165, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 4.62962962962963, | |
| "grad_norm": 0.7154943943023682, | |
| "learning_rate": 3.553240740740741e-05, | |
| "loss": 0.6892, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 4.814814814814815, | |
| "grad_norm": 0.823897659778595, | |
| "learning_rate": 3.49537037037037e-05, | |
| "loss": 0.7026, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 0.7548332810401917, | |
| "learning_rate": 3.4375e-05, | |
| "loss": 0.7019, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 1.056677222251892, | |
| "eval_runtime": 30.3702, | |
| "eval_samples_per_second": 39.512, | |
| "eval_steps_per_second": 9.878, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 5.185185185185185, | |
| "grad_norm": 0.6250707507133484, | |
| "learning_rate": 3.3796296296296295e-05, | |
| "loss": 0.7107, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 5.37037037037037, | |
| "grad_norm": 0.7014070749282837, | |
| "learning_rate": 3.3217592592592595e-05, | |
| "loss": 0.696, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 5.555555555555555, | |
| "grad_norm": 0.8305183053016663, | |
| "learning_rate": 3.263888888888889e-05, | |
| "loss": 0.6858, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 5.7407407407407405, | |
| "grad_norm": 0.5459818840026855, | |
| "learning_rate": 3.206018518518519e-05, | |
| "loss": 0.6828, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 5.925925925925926, | |
| "grad_norm": 0.40176087617874146, | |
| "learning_rate": 3.148148148148148e-05, | |
| "loss": 0.6841, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 1.0900229215621948, | |
| "eval_runtime": 30.4123, | |
| "eval_samples_per_second": 39.458, | |
| "eval_steps_per_second": 9.864, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 6.111111111111111, | |
| "grad_norm": 0.4161689281463623, | |
| "learning_rate": 3.090277777777778e-05, | |
| "loss": 0.6615, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 6.296296296296296, | |
| "grad_norm": 0.7132428288459778, | |
| "learning_rate": 3.0324074074074077e-05, | |
| "loss": 0.683, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 6.481481481481482, | |
| "grad_norm": 0.6823524236679077, | |
| "learning_rate": 2.9745370370370373e-05, | |
| "loss": 0.6692, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 6.666666666666667, | |
| "grad_norm": 0.3051627278327942, | |
| "learning_rate": 2.916666666666667e-05, | |
| "loss": 0.6771, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 6.851851851851852, | |
| "grad_norm": 0.5912793278694153, | |
| "learning_rate": 2.8587962962962966e-05, | |
| "loss": 0.6959, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 1.12686288356781, | |
| "eval_runtime": 30.4381, | |
| "eval_samples_per_second": 39.424, | |
| "eval_steps_per_second": 9.856, | |
| "step": 18900 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 43200, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 16, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.97536776192e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |