oh_v1.3_evol_instruct_x.25 / trainer_state.json
sedrickkeh's picture
End of training
8feb18f verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 822,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0364963503649635,
"grad_norm": 3.2654903919327523,
"learning_rate": 5e-06,
"loss": 1.0419,
"step": 10
},
{
"epoch": 0.072992700729927,
"grad_norm": 1.5482985205580984,
"learning_rate": 5e-06,
"loss": 0.9117,
"step": 20
},
{
"epoch": 0.10948905109489052,
"grad_norm": 1.2069486973191534,
"learning_rate": 5e-06,
"loss": 0.8752,
"step": 30
},
{
"epoch": 0.145985401459854,
"grad_norm": 1.2169049310832125,
"learning_rate": 5e-06,
"loss": 0.8479,
"step": 40
},
{
"epoch": 0.18248175182481752,
"grad_norm": 1.0461259998073966,
"learning_rate": 5e-06,
"loss": 0.8276,
"step": 50
},
{
"epoch": 0.21897810218978103,
"grad_norm": 0.9136364083656688,
"learning_rate": 5e-06,
"loss": 0.8144,
"step": 60
},
{
"epoch": 0.25547445255474455,
"grad_norm": 1.0016252258412965,
"learning_rate": 5e-06,
"loss": 0.8029,
"step": 70
},
{
"epoch": 0.291970802919708,
"grad_norm": 0.7256070366121048,
"learning_rate": 5e-06,
"loss": 0.7984,
"step": 80
},
{
"epoch": 0.3284671532846715,
"grad_norm": 0.8059560568156144,
"learning_rate": 5e-06,
"loss": 0.7845,
"step": 90
},
{
"epoch": 0.36496350364963503,
"grad_norm": 1.0375230892576275,
"learning_rate": 5e-06,
"loss": 0.784,
"step": 100
},
{
"epoch": 0.40145985401459855,
"grad_norm": 0.6429193748930873,
"learning_rate": 5e-06,
"loss": 0.7786,
"step": 110
},
{
"epoch": 0.43795620437956206,
"grad_norm": 0.6387684294515616,
"learning_rate": 5e-06,
"loss": 0.7741,
"step": 120
},
{
"epoch": 0.4744525547445255,
"grad_norm": 0.7647783012503196,
"learning_rate": 5e-06,
"loss": 0.7706,
"step": 130
},
{
"epoch": 0.5109489051094891,
"grad_norm": 0.6954646892322683,
"learning_rate": 5e-06,
"loss": 0.7693,
"step": 140
},
{
"epoch": 0.5474452554744526,
"grad_norm": 0.7090301419858328,
"learning_rate": 5e-06,
"loss": 0.7698,
"step": 150
},
{
"epoch": 0.583941605839416,
"grad_norm": 0.8080543613629317,
"learning_rate": 5e-06,
"loss": 0.7595,
"step": 160
},
{
"epoch": 0.6204379562043796,
"grad_norm": 0.6419738319193917,
"learning_rate": 5e-06,
"loss": 0.7614,
"step": 170
},
{
"epoch": 0.656934306569343,
"grad_norm": 0.5918340383797932,
"learning_rate": 5e-06,
"loss": 0.7541,
"step": 180
},
{
"epoch": 0.6934306569343066,
"grad_norm": 1.152710341207945,
"learning_rate": 5e-06,
"loss": 0.7573,
"step": 190
},
{
"epoch": 0.7299270072992701,
"grad_norm": 0.6391926393585599,
"learning_rate": 5e-06,
"loss": 0.7582,
"step": 200
},
{
"epoch": 0.7664233576642335,
"grad_norm": 0.5988044301788384,
"learning_rate": 5e-06,
"loss": 0.7582,
"step": 210
},
{
"epoch": 0.8029197080291971,
"grad_norm": 0.6020411724470022,
"learning_rate": 5e-06,
"loss": 0.7519,
"step": 220
},
{
"epoch": 0.8394160583941606,
"grad_norm": 0.6220550687844844,
"learning_rate": 5e-06,
"loss": 0.7551,
"step": 230
},
{
"epoch": 0.8759124087591241,
"grad_norm": 0.6082664111510806,
"learning_rate": 5e-06,
"loss": 0.7519,
"step": 240
},
{
"epoch": 0.9124087591240876,
"grad_norm": 0.7929440801500307,
"learning_rate": 5e-06,
"loss": 0.7471,
"step": 250
},
{
"epoch": 0.948905109489051,
"grad_norm": 0.8557482541404012,
"learning_rate": 5e-06,
"loss": 0.7471,
"step": 260
},
{
"epoch": 0.9854014598540146,
"grad_norm": 0.5725008171409565,
"learning_rate": 5e-06,
"loss": 0.7448,
"step": 270
},
{
"epoch": 1.0,
"eval_loss": 0.7464940547943115,
"eval_runtime": 193.5595,
"eval_samples_per_second": 38.143,
"eval_steps_per_second": 0.599,
"step": 274
},
{
"epoch": 1.0218978102189782,
"grad_norm": 0.8391272846123186,
"learning_rate": 5e-06,
"loss": 0.7201,
"step": 280
},
{
"epoch": 1.0583941605839415,
"grad_norm": 0.8538310326832037,
"learning_rate": 5e-06,
"loss": 0.6987,
"step": 290
},
{
"epoch": 1.094890510948905,
"grad_norm": 0.6827445863139939,
"learning_rate": 5e-06,
"loss": 0.695,
"step": 300
},
{
"epoch": 1.1313868613138687,
"grad_norm": 0.5882300886638387,
"learning_rate": 5e-06,
"loss": 0.6922,
"step": 310
},
{
"epoch": 1.167883211678832,
"grad_norm": 0.9284544165624958,
"learning_rate": 5e-06,
"loss": 0.7005,
"step": 320
},
{
"epoch": 1.2043795620437956,
"grad_norm": 0.7696038972626205,
"learning_rate": 5e-06,
"loss": 0.696,
"step": 330
},
{
"epoch": 1.2408759124087592,
"grad_norm": 0.626331783091493,
"learning_rate": 5e-06,
"loss": 0.6961,
"step": 340
},
{
"epoch": 1.2773722627737225,
"grad_norm": 0.6263624432151553,
"learning_rate": 5e-06,
"loss": 0.6955,
"step": 350
},
{
"epoch": 1.313868613138686,
"grad_norm": 0.8006137682650242,
"learning_rate": 5e-06,
"loss": 0.6967,
"step": 360
},
{
"epoch": 1.3503649635036497,
"grad_norm": 0.5880134638823885,
"learning_rate": 5e-06,
"loss": 0.702,
"step": 370
},
{
"epoch": 1.3868613138686132,
"grad_norm": 0.647442922917404,
"learning_rate": 5e-06,
"loss": 0.7,
"step": 380
},
{
"epoch": 1.4233576642335766,
"grad_norm": 0.6033654206417786,
"learning_rate": 5e-06,
"loss": 0.6935,
"step": 390
},
{
"epoch": 1.4598540145985401,
"grad_norm": 0.842826594179501,
"learning_rate": 5e-06,
"loss": 0.6925,
"step": 400
},
{
"epoch": 1.4963503649635037,
"grad_norm": 0.6487088398169627,
"learning_rate": 5e-06,
"loss": 0.6987,
"step": 410
},
{
"epoch": 1.5328467153284673,
"grad_norm": 0.644019689180568,
"learning_rate": 5e-06,
"loss": 0.6975,
"step": 420
},
{
"epoch": 1.5693430656934306,
"grad_norm": 0.7554928160990546,
"learning_rate": 5e-06,
"loss": 0.6955,
"step": 430
},
{
"epoch": 1.6058394160583942,
"grad_norm": 0.6360155091593863,
"learning_rate": 5e-06,
"loss": 0.6947,
"step": 440
},
{
"epoch": 1.6423357664233578,
"grad_norm": 0.7649060131558615,
"learning_rate": 5e-06,
"loss": 0.6979,
"step": 450
},
{
"epoch": 1.6788321167883211,
"grad_norm": 0.7060249710510591,
"learning_rate": 5e-06,
"loss": 0.6887,
"step": 460
},
{
"epoch": 1.7153284671532847,
"grad_norm": 0.614705856783836,
"learning_rate": 5e-06,
"loss": 0.6925,
"step": 470
},
{
"epoch": 1.7518248175182483,
"grad_norm": 0.6665801961837532,
"learning_rate": 5e-06,
"loss": 0.6939,
"step": 480
},
{
"epoch": 1.7883211678832116,
"grad_norm": 0.6137773686568165,
"learning_rate": 5e-06,
"loss": 0.6968,
"step": 490
},
{
"epoch": 1.8248175182481752,
"grad_norm": 0.597915675609789,
"learning_rate": 5e-06,
"loss": 0.6903,
"step": 500
},
{
"epoch": 1.8613138686131387,
"grad_norm": 0.5853039235207544,
"learning_rate": 5e-06,
"loss": 0.6919,
"step": 510
},
{
"epoch": 1.897810218978102,
"grad_norm": 0.6252196584141081,
"learning_rate": 5e-06,
"loss": 0.6917,
"step": 520
},
{
"epoch": 1.9343065693430657,
"grad_norm": 0.6355616628686108,
"learning_rate": 5e-06,
"loss": 0.6947,
"step": 530
},
{
"epoch": 1.9708029197080292,
"grad_norm": 0.7084920207276291,
"learning_rate": 5e-06,
"loss": 0.6909,
"step": 540
},
{
"epoch": 2.0,
"eval_loss": 0.7327279448509216,
"eval_runtime": 184.0166,
"eval_samples_per_second": 40.121,
"eval_steps_per_second": 0.63,
"step": 548
},
{
"epoch": 2.0072992700729926,
"grad_norm": 0.8657992423171681,
"learning_rate": 5e-06,
"loss": 0.6891,
"step": 550
},
{
"epoch": 2.0437956204379564,
"grad_norm": 0.7941631908501828,
"learning_rate": 5e-06,
"loss": 0.6428,
"step": 560
},
{
"epoch": 2.0802919708029197,
"grad_norm": 0.9214323090929812,
"learning_rate": 5e-06,
"loss": 0.6418,
"step": 570
},
{
"epoch": 2.116788321167883,
"grad_norm": 0.7759251168034266,
"learning_rate": 5e-06,
"loss": 0.6411,
"step": 580
},
{
"epoch": 2.153284671532847,
"grad_norm": 0.7159794483510337,
"learning_rate": 5e-06,
"loss": 0.6429,
"step": 590
},
{
"epoch": 2.18978102189781,
"grad_norm": 0.7275049985326262,
"learning_rate": 5e-06,
"loss": 0.6418,
"step": 600
},
{
"epoch": 2.2262773722627736,
"grad_norm": 0.5798767855808342,
"learning_rate": 5e-06,
"loss": 0.6455,
"step": 610
},
{
"epoch": 2.2627737226277373,
"grad_norm": 0.5604983214077892,
"learning_rate": 5e-06,
"loss": 0.64,
"step": 620
},
{
"epoch": 2.2992700729927007,
"grad_norm": 0.6059873724806449,
"learning_rate": 5e-06,
"loss": 0.6424,
"step": 630
},
{
"epoch": 2.335766423357664,
"grad_norm": 0.8067784530621691,
"learning_rate": 5e-06,
"loss": 0.6422,
"step": 640
},
{
"epoch": 2.372262773722628,
"grad_norm": 0.7136802882114688,
"learning_rate": 5e-06,
"loss": 0.6437,
"step": 650
},
{
"epoch": 2.408759124087591,
"grad_norm": 0.6102999029962267,
"learning_rate": 5e-06,
"loss": 0.6458,
"step": 660
},
{
"epoch": 2.445255474452555,
"grad_norm": 0.614293225765085,
"learning_rate": 5e-06,
"loss": 0.6452,
"step": 670
},
{
"epoch": 2.4817518248175183,
"grad_norm": 0.600198323252072,
"learning_rate": 5e-06,
"loss": 0.644,
"step": 680
},
{
"epoch": 2.5182481751824817,
"grad_norm": 0.7644298597279312,
"learning_rate": 5e-06,
"loss": 0.6491,
"step": 690
},
{
"epoch": 2.554744525547445,
"grad_norm": 0.778090825027965,
"learning_rate": 5e-06,
"loss": 0.6431,
"step": 700
},
{
"epoch": 2.591240875912409,
"grad_norm": 0.6476324319228641,
"learning_rate": 5e-06,
"loss": 0.6473,
"step": 710
},
{
"epoch": 2.627737226277372,
"grad_norm": 0.5998032151877339,
"learning_rate": 5e-06,
"loss": 0.6465,
"step": 720
},
{
"epoch": 2.664233576642336,
"grad_norm": 0.5955651290954712,
"learning_rate": 5e-06,
"loss": 0.6456,
"step": 730
},
{
"epoch": 2.7007299270072993,
"grad_norm": 0.6010226042821943,
"learning_rate": 5e-06,
"loss": 0.6442,
"step": 740
},
{
"epoch": 2.7372262773722627,
"grad_norm": 0.821931399421341,
"learning_rate": 5e-06,
"loss": 0.6459,
"step": 750
},
{
"epoch": 2.7737226277372264,
"grad_norm": 0.6668282736525819,
"learning_rate": 5e-06,
"loss": 0.6503,
"step": 760
},
{
"epoch": 2.81021897810219,
"grad_norm": 0.7099179596842483,
"learning_rate": 5e-06,
"loss": 0.6473,
"step": 770
},
{
"epoch": 2.846715328467153,
"grad_norm": 0.6671120507448849,
"learning_rate": 5e-06,
"loss": 0.6426,
"step": 780
},
{
"epoch": 2.883211678832117,
"grad_norm": 0.6360799683340312,
"learning_rate": 5e-06,
"loss": 0.6473,
"step": 790
},
{
"epoch": 2.9197080291970803,
"grad_norm": 0.7340811126545655,
"learning_rate": 5e-06,
"loss": 0.651,
"step": 800
},
{
"epoch": 2.9562043795620436,
"grad_norm": 0.618110632863654,
"learning_rate": 5e-06,
"loss": 0.6448,
"step": 810
},
{
"epoch": 2.9927007299270074,
"grad_norm": 0.7521256358398457,
"learning_rate": 5e-06,
"loss": 0.6423,
"step": 820
},
{
"epoch": 3.0,
"eval_loss": 0.7341417074203491,
"eval_runtime": 183.6218,
"eval_samples_per_second": 40.208,
"eval_steps_per_second": 0.632,
"step": 822
},
{
"epoch": 3.0,
"step": 822,
"total_flos": 1376671236096000.0,
"train_loss": 0.7106172601962032,
"train_runtime": 27349.7837,
"train_samples_per_second": 15.387,
"train_steps_per_second": 0.03
}
],
"logging_steps": 10,
"max_steps": 822,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1376671236096000.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}