Model1-v1-Rival / checkpoint-24806 /trainer_state.json
alrivalda's picture
Upload folder using huggingface_hub
109c8cb verified
{
"best_global_step": 24806,
"best_metric": 0.13123206794261932,
"best_model_checkpoint": "Model1-v1-Rival/checkpoint-24806",
"epoch": 2.0,
"eval_steps": 500,
"global_step": 24806,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.04031282754172378,
"grad_norm": 35.80685043334961,
"learning_rate": 2.011285771866183e-07,
"loss": 0.7492,
"step": 500
},
{
"epoch": 0.08062565508344756,
"grad_norm": 27.830657958984375,
"learning_rate": 4.0266021765417167e-07,
"loss": 0.564,
"step": 1000
},
{
"epoch": 0.12093848262517133,
"grad_norm": 22.993253707885742,
"learning_rate": 6.041918581217252e-07,
"loss": 0.497,
"step": 1500
},
{
"epoch": 0.16125131016689512,
"grad_norm": 31.184377670288086,
"learning_rate": 8.057234985892785e-07,
"loss": 0.4362,
"step": 2000
},
{
"epoch": 0.2015641377086189,
"grad_norm": 31.91282844543457,
"learning_rate": 9.991937290033595e-07,
"loss": 0.41,
"step": 2500
},
{
"epoch": 0.24187696525034266,
"grad_norm": 42.18364715576172,
"learning_rate": 9.767973124300112e-07,
"loss": 0.3815,
"step": 3000
},
{
"epoch": 0.28218979279206646,
"grad_norm": 31.624061584472656,
"learning_rate": 9.544008958566627e-07,
"loss": 0.3568,
"step": 3500
},
{
"epoch": 0.32250262033379024,
"grad_norm": 36.55485534667969,
"learning_rate": 9.320044792833146e-07,
"loss": 0.341,
"step": 4000
},
{
"epoch": 0.362815447875514,
"grad_norm": 34.942264556884766,
"learning_rate": 9.096080627099664e-07,
"loss": 0.3293,
"step": 4500
},
{
"epoch": 0.4031282754172378,
"grad_norm": 30.162443161010742,
"learning_rate": 8.872116461366181e-07,
"loss": 0.3184,
"step": 5000
},
{
"epoch": 0.44344110295896155,
"grad_norm": 42.151607513427734,
"learning_rate": 8.648152295632698e-07,
"loss": 0.3149,
"step": 5500
},
{
"epoch": 0.4837539305006853,
"grad_norm": 21.629465103149414,
"learning_rate": 8.424188129899216e-07,
"loss": 0.2969,
"step": 6000
},
{
"epoch": 0.5240667580424091,
"grad_norm": 45.252803802490234,
"learning_rate": 8.200223964165734e-07,
"loss": 0.2962,
"step": 6500
},
{
"epoch": 0.5643795855841329,
"grad_norm": 31.193838119506836,
"learning_rate": 7.97625979843225e-07,
"loss": 0.2858,
"step": 7000
},
{
"epoch": 0.6046924131258566,
"grad_norm": 96.9237289428711,
"learning_rate": 7.752295632698768e-07,
"loss": 0.268,
"step": 7500
},
{
"epoch": 0.6450052406675805,
"grad_norm": 21.470144271850586,
"learning_rate": 7.528331466965285e-07,
"loss": 0.2622,
"step": 8000
},
{
"epoch": 0.6853180682093042,
"grad_norm": 35.70467758178711,
"learning_rate": 7.304367301231803e-07,
"loss": 0.2741,
"step": 8500
},
{
"epoch": 0.725630895751028,
"grad_norm": 19.73318862915039,
"learning_rate": 7.080403135498319e-07,
"loss": 0.2626,
"step": 9000
},
{
"epoch": 0.7659437232927517,
"grad_norm": 23.529918670654297,
"learning_rate": 6.856438969764838e-07,
"loss": 0.2518,
"step": 9500
},
{
"epoch": 0.8062565508344756,
"grad_norm": 41.51838302612305,
"learning_rate": 6.632474804031355e-07,
"loss": 0.2495,
"step": 10000
},
{
"epoch": 0.8465693783761993,
"grad_norm": 14.217453956604004,
"learning_rate": 6.408510638297872e-07,
"loss": 0.2437,
"step": 10500
},
{
"epoch": 0.8868822059179231,
"grad_norm": 18.972307205200195,
"learning_rate": 6.184546472564389e-07,
"loss": 0.2482,
"step": 11000
},
{
"epoch": 0.9271950334596468,
"grad_norm": 47.28855514526367,
"learning_rate": 5.960582306830907e-07,
"loss": 0.2419,
"step": 11500
},
{
"epoch": 0.9675078610013707,
"grad_norm": 48.309165954589844,
"learning_rate": 5.736618141097424e-07,
"loss": 0.2421,
"step": 12000
},
{
"epoch": 1.0,
"eval_accuracy": 0.9340860036526648,
"eval_loss": 0.17090356349945068,
"eval_model_preparation_time": 0.0026,
"eval_runtime": 109.2506,
"eval_samples_per_second": 441.087,
"eval_steps_per_second": 55.139,
"step": 12403
},
{
"epoch": 1.0078206885430945,
"grad_norm": 47.087913513183594,
"learning_rate": 5.512653975363942e-07,
"loss": 0.2372,
"step": 12500
},
{
"epoch": 1.0481335160848182,
"grad_norm": 40.675697326660156,
"learning_rate": 5.288689809630459e-07,
"loss": 0.2249,
"step": 13000
},
{
"epoch": 1.088446343626542,
"grad_norm": 46.77396774291992,
"learning_rate": 5.064725643896977e-07,
"loss": 0.2207,
"step": 13500
},
{
"epoch": 1.1287591711682659,
"grad_norm": 69.7470703125,
"learning_rate": 4.840761478163493e-07,
"loss": 0.2276,
"step": 14000
},
{
"epoch": 1.1690719987099896,
"grad_norm": 35.32638168334961,
"learning_rate": 4.6167973124300107e-07,
"loss": 0.2235,
"step": 14500
},
{
"epoch": 1.2093848262517133,
"grad_norm": 63.080291748046875,
"learning_rate": 4.3928331466965287e-07,
"loss": 0.2219,
"step": 15000
},
{
"epoch": 1.249697653793437,
"grad_norm": 45.029239654541016,
"learning_rate": 4.1688689809630457e-07,
"loss": 0.2184,
"step": 15500
},
{
"epoch": 1.2900104813351607,
"grad_norm": 20.416336059570312,
"learning_rate": 3.944904815229563e-07,
"loss": 0.2081,
"step": 16000
},
{
"epoch": 1.3303233088768847,
"grad_norm": 28.367952346801758,
"learning_rate": 3.72094064949608e-07,
"loss": 0.2157,
"step": 16500
},
{
"epoch": 1.3706361364186084,
"grad_norm": 13.548885345458984,
"learning_rate": 3.4969764837625977e-07,
"loss": 0.212,
"step": 17000
},
{
"epoch": 1.4109489639603323,
"grad_norm": 94.12673950195312,
"learning_rate": 3.273012318029115e-07,
"loss": 0.2075,
"step": 17500
},
{
"epoch": 1.451261791502056,
"grad_norm": 38.12338638305664,
"learning_rate": 3.049048152295632e-07,
"loss": 0.2148,
"step": 18000
},
{
"epoch": 1.4915746190437797,
"grad_norm": 45.156455993652344,
"learning_rate": 2.82508398656215e-07,
"loss": 0.2075,
"step": 18500
},
{
"epoch": 1.5318874465855035,
"grad_norm": 18.083568572998047,
"learning_rate": 2.601119820828667e-07,
"loss": 0.2092,
"step": 19000
},
{
"epoch": 1.5722002741272272,
"grad_norm": 27.93409538269043,
"learning_rate": 2.3771556550951847e-07,
"loss": 0.205,
"step": 19500
},
{
"epoch": 1.612513101668951,
"grad_norm": 39.768192291259766,
"learning_rate": 2.153191489361702e-07,
"loss": 0.2039,
"step": 20000
},
{
"epoch": 1.6528259292106748,
"grad_norm": 28.78324317932129,
"learning_rate": 1.9292273236282195e-07,
"loss": 0.204,
"step": 20500
},
{
"epoch": 1.6931387567523988,
"grad_norm": 94.95115661621094,
"learning_rate": 1.7052631578947368e-07,
"loss": 0.2154,
"step": 21000
},
{
"epoch": 1.7334515842941225,
"grad_norm": 51.76600646972656,
"learning_rate": 1.4812989921612543e-07,
"loss": 0.208,
"step": 21500
},
{
"epoch": 1.7737644118358462,
"grad_norm": 24.089155197143555,
"learning_rate": 1.2573348264277713e-07,
"loss": 0.2071,
"step": 22000
},
{
"epoch": 1.81407723937757,
"grad_norm": 46.96974182128906,
"learning_rate": 1.0333706606942888e-07,
"loss": 0.198,
"step": 22500
},
{
"epoch": 1.8543900669192936,
"grad_norm": 66.37223815917969,
"learning_rate": 8.094064949608062e-08,
"loss": 0.2015,
"step": 23000
},
{
"epoch": 1.8947028944610174,
"grad_norm": 42.39664077758789,
"learning_rate": 5.854423292273236e-08,
"loss": 0.1982,
"step": 23500
},
{
"epoch": 1.9350157220027413,
"grad_norm": 42.68307113647461,
"learning_rate": 3.61478163493841e-08,
"loss": 0.1958,
"step": 24000
},
{
"epoch": 1.975328549544465,
"grad_norm": 17.30171012878418,
"learning_rate": 1.3751399776035833e-08,
"loss": 0.2024,
"step": 24500
},
{
"epoch": 2.0,
"eval_accuracy": 0.9520795284741823,
"eval_loss": 0.13123206794261932,
"eval_model_preparation_time": 0.0026,
"eval_runtime": 100.8699,
"eval_samples_per_second": 477.734,
"eval_steps_per_second": 59.72,
"step": 24806
}
],
"logging_steps": 500,
"max_steps": 24806,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6.6484442538990895e+19,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}