simplewiki_upsampled_tom / trainer_state.json
ptsv's picture
ptsv/gpt2-m_simplewiki_upsampled_tom
c152191 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.98989898989899,
"eval_steps": 20,
"global_step": 111,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.05387205387205387,
"grad_norm": 30.91269874572754,
"learning_rate": 0.0001,
"loss": 9.4247,
"step": 2
},
{
"epoch": 0.10774410774410774,
"grad_norm": 17.899776458740234,
"learning_rate": 9.818181818181818e-05,
"loss": 8.2334,
"step": 4
},
{
"epoch": 0.16161616161616163,
"grad_norm": 11.289294242858887,
"learning_rate": 9.636363636363637e-05,
"loss": 6.7049,
"step": 6
},
{
"epoch": 0.21548821548821548,
"grad_norm": 13.52355670928955,
"learning_rate": 9.454545454545455e-05,
"loss": 6.0078,
"step": 8
},
{
"epoch": 0.26936026936026936,
"grad_norm": 6.83944845199585,
"learning_rate": 9.272727272727273e-05,
"loss": 5.4206,
"step": 10
},
{
"epoch": 0.32323232323232326,
"grad_norm": 5.749248504638672,
"learning_rate": 9.090909090909092e-05,
"loss": 5.0018,
"step": 12
},
{
"epoch": 0.3771043771043771,
"grad_norm": 6.119119644165039,
"learning_rate": 8.90909090909091e-05,
"loss": 4.7585,
"step": 14
},
{
"epoch": 0.43097643097643096,
"grad_norm": 5.070394992828369,
"learning_rate": 8.727272727272727e-05,
"loss": 4.4387,
"step": 16
},
{
"epoch": 0.48484848484848486,
"grad_norm": 5.378049850463867,
"learning_rate": 8.545454545454545e-05,
"loss": 4.1869,
"step": 18
},
{
"epoch": 0.5387205387205387,
"grad_norm": 6.204185962677002,
"learning_rate": 8.363636363636364e-05,
"loss": 4.0268,
"step": 20
},
{
"epoch": 0.5387205387205387,
"eval_loss": 3.4558703899383545,
"eval_runtime": 8.1602,
"eval_samples_per_second": 145.708,
"eval_steps_per_second": 18.259,
"step": 20
},
{
"epoch": 0.5925925925925926,
"grad_norm": 4.224447727203369,
"learning_rate": 8.181818181818183e-05,
"loss": 3.8953,
"step": 22
},
{
"epoch": 0.6464646464646465,
"grad_norm": 3.6069717407226562,
"learning_rate": 8e-05,
"loss": 3.8231,
"step": 24
},
{
"epoch": 0.7003367003367004,
"grad_norm": 2.707709550857544,
"learning_rate": 7.818181818181818e-05,
"loss": 3.7116,
"step": 26
},
{
"epoch": 0.7542087542087542,
"grad_norm": 3.31360125541687,
"learning_rate": 7.636363636363637e-05,
"loss": 3.5999,
"step": 28
},
{
"epoch": 0.8080808080808081,
"grad_norm": 2.9818968772888184,
"learning_rate": 7.454545454545455e-05,
"loss": 3.5958,
"step": 30
},
{
"epoch": 0.8619528619528619,
"grad_norm": 2.68215274810791,
"learning_rate": 7.272727272727273e-05,
"loss": 3.5518,
"step": 32
},
{
"epoch": 0.9158249158249159,
"grad_norm": 1.9956828355789185,
"learning_rate": 7.090909090909092e-05,
"loss": 3.4949,
"step": 34
},
{
"epoch": 0.9696969696969697,
"grad_norm": 1.9735620021820068,
"learning_rate": 6.90909090909091e-05,
"loss": 3.4185,
"step": 36
},
{
"epoch": 1.0235690235690236,
"grad_norm": 1.8688853979110718,
"learning_rate": 6.727272727272727e-05,
"loss": 3.4284,
"step": 38
},
{
"epoch": 1.0774410774410774,
"grad_norm": 1.9668887853622437,
"learning_rate": 6.545454545454546e-05,
"loss": 3.3888,
"step": 40
},
{
"epoch": 1.0774410774410774,
"eval_loss": 3.2203786373138428,
"eval_runtime": 8.1727,
"eval_samples_per_second": 145.484,
"eval_steps_per_second": 18.231,
"step": 40
},
{
"epoch": 1.1313131313131313,
"grad_norm": 1.8594480752944946,
"learning_rate": 6.363636363636364e-05,
"loss": 3.3016,
"step": 42
},
{
"epoch": 1.1851851851851851,
"grad_norm": 2.0851259231567383,
"learning_rate": 6.181818181818182e-05,
"loss": 3.2937,
"step": 44
},
{
"epoch": 1.239057239057239,
"grad_norm": 2.022775411605835,
"learning_rate": 6e-05,
"loss": 3.3097,
"step": 46
},
{
"epoch": 1.2929292929292928,
"grad_norm": 1.7249592542648315,
"learning_rate": 5.818181818181818e-05,
"loss": 3.2856,
"step": 48
},
{
"epoch": 1.3468013468013469,
"grad_norm": 1.894808292388916,
"learning_rate": 5.636363636363636e-05,
"loss": 3.2579,
"step": 50
},
{
"epoch": 1.4006734006734007,
"grad_norm": 2.0816519260406494,
"learning_rate": 5.4545454545454546e-05,
"loss": 3.2534,
"step": 52
},
{
"epoch": 1.4545454545454546,
"grad_norm": 1.8632744550704956,
"learning_rate": 5.272727272727272e-05,
"loss": 3.2358,
"step": 54
},
{
"epoch": 1.5084175084175084,
"grad_norm": 1.8851394653320312,
"learning_rate": 5.090909090909091e-05,
"loss": 3.2256,
"step": 56
},
{
"epoch": 1.5622895622895623,
"grad_norm": 1.7529875040054321,
"learning_rate": 4.909090909090909e-05,
"loss": 3.2148,
"step": 58
},
{
"epoch": 1.6161616161616161,
"grad_norm": 1.8536018133163452,
"learning_rate": 4.7272727272727275e-05,
"loss": 3.2608,
"step": 60
},
{
"epoch": 1.6161616161616161,
"eval_loss": 3.1798043251037598,
"eval_runtime": 8.1729,
"eval_samples_per_second": 145.48,
"eval_steps_per_second": 18.231,
"step": 60
},
{
"epoch": 1.67003367003367,
"grad_norm": 1.9621059894561768,
"learning_rate": 4.545454545454546e-05,
"loss": 3.2202,
"step": 62
},
{
"epoch": 1.723905723905724,
"grad_norm": 1.9148567914962769,
"learning_rate": 4.3636363636363636e-05,
"loss": 3.2282,
"step": 64
},
{
"epoch": 1.7777777777777777,
"grad_norm": 1.7804056406021118,
"learning_rate": 4.181818181818182e-05,
"loss": 3.2654,
"step": 66
},
{
"epoch": 1.8316498316498318,
"grad_norm": 1.8891396522521973,
"learning_rate": 4e-05,
"loss": 3.1924,
"step": 68
},
{
"epoch": 1.8855218855218854,
"grad_norm": 1.8429640531539917,
"learning_rate": 3.818181818181819e-05,
"loss": 3.1756,
"step": 70
},
{
"epoch": 1.9393939393939394,
"grad_norm": 1.795774221420288,
"learning_rate": 3.6363636363636364e-05,
"loss": 3.2097,
"step": 72
},
{
"epoch": 1.9932659932659933,
"grad_norm": 1.8442139625549316,
"learning_rate": 3.454545454545455e-05,
"loss": 3.2271,
"step": 74
},
{
"epoch": 2.047138047138047,
"grad_norm": 1.6865243911743164,
"learning_rate": 3.272727272727273e-05,
"loss": 3.1261,
"step": 76
},
{
"epoch": 2.101010101010101,
"grad_norm": 1.6551978588104248,
"learning_rate": 3.090909090909091e-05,
"loss": 3.101,
"step": 78
},
{
"epoch": 2.154882154882155,
"grad_norm": 1.7607771158218384,
"learning_rate": 2.909090909090909e-05,
"loss": 3.0763,
"step": 80
},
{
"epoch": 2.154882154882155,
"eval_loss": 3.162412405014038,
"eval_runtime": 8.1804,
"eval_samples_per_second": 145.348,
"eval_steps_per_second": 18.214,
"step": 80
},
{
"epoch": 2.208754208754209,
"grad_norm": 1.720595359802246,
"learning_rate": 2.7272727272727273e-05,
"loss": 3.1138,
"step": 82
},
{
"epoch": 2.2626262626262625,
"grad_norm": 1.7866605520248413,
"learning_rate": 2.5454545454545454e-05,
"loss": 3.1239,
"step": 84
},
{
"epoch": 2.3164983164983166,
"grad_norm": 1.7226550579071045,
"learning_rate": 2.3636363636363637e-05,
"loss": 3.0789,
"step": 86
},
{
"epoch": 2.3703703703703702,
"grad_norm": 1.8532516956329346,
"learning_rate": 2.1818181818181818e-05,
"loss": 3.1286,
"step": 88
},
{
"epoch": 2.4242424242424243,
"grad_norm": 1.7198731899261475,
"learning_rate": 2e-05,
"loss": 3.1123,
"step": 90
},
{
"epoch": 2.478114478114478,
"grad_norm": 1.8436646461486816,
"learning_rate": 1.8181818181818182e-05,
"loss": 3.0712,
"step": 92
},
{
"epoch": 2.531986531986532,
"grad_norm": 1.811488151550293,
"learning_rate": 1.6363636363636366e-05,
"loss": 3.0727,
"step": 94
},
{
"epoch": 2.5858585858585856,
"grad_norm": 1.7093764543533325,
"learning_rate": 1.4545454545454545e-05,
"loss": 3.0966,
"step": 96
},
{
"epoch": 2.6397306397306397,
"grad_norm": 1.8103761672973633,
"learning_rate": 1.2727272727272727e-05,
"loss": 3.084,
"step": 98
},
{
"epoch": 2.6936026936026938,
"grad_norm": 1.8916672468185425,
"learning_rate": 1.0909090909090909e-05,
"loss": 3.0912,
"step": 100
},
{
"epoch": 2.6936026936026938,
"eval_loss": 3.153109550476074,
"eval_runtime": 8.1741,
"eval_samples_per_second": 145.459,
"eval_steps_per_second": 18.228,
"step": 100
},
{
"epoch": 2.7474747474747474,
"grad_norm": 1.7815495729446411,
"learning_rate": 9.090909090909091e-06,
"loss": 3.0743,
"step": 102
},
{
"epoch": 2.8013468013468015,
"grad_norm": 1.8348023891448975,
"learning_rate": 7.272727272727272e-06,
"loss": 3.092,
"step": 104
},
{
"epoch": 2.855218855218855,
"grad_norm": 1.807790756225586,
"learning_rate": 5.4545454545454545e-06,
"loss": 3.0913,
"step": 106
},
{
"epoch": 2.909090909090909,
"grad_norm": 1.7084001302719116,
"learning_rate": 3.636363636363636e-06,
"loss": 3.07,
"step": 108
},
{
"epoch": 2.962962962962963,
"grad_norm": 1.719460129737854,
"learning_rate": 1.818181818181818e-06,
"loss": 3.0812,
"step": 110
},
{
"epoch": 2.98989898989899,
"step": 111,
"total_flos": 1.389220151427072e+16,
"train_loss": 3.7213772073522344,
"train_runtime": 747.8744,
"train_samples_per_second": 38.144,
"train_steps_per_second": 0.148
},
{
"epoch": 2.98989898989899,
"eval_loss": 3.1522939205169678,
"eval_runtime": 8.1573,
"eval_samples_per_second": 145.759,
"eval_steps_per_second": 18.266,
"step": 111
},
{
"epoch": 2.98989898989899,
"eval_loss": 3.1559650897979736,
"eval_runtime": 8.1292,
"eval_samples_per_second": 146.262,
"eval_steps_per_second": 18.329,
"step": 111
}
],
"logging_steps": 2,
"max_steps": 111,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.389220151427072e+16,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}