klora_2000_skill / 114 /trainer_state.json
RayDu0010's picture
Upload folder using huggingface_hub
32415af verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 372,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.013458950201884253,
"grad_norm": 1.00712251663208,
"learning_rate": 1.2903225806451614e-06,
"loss": 1.264,
"step": 5
},
{
"epoch": 0.026917900403768506,
"grad_norm": 0.9430832266807556,
"learning_rate": 2.9032258064516128e-06,
"loss": 1.2378,
"step": 10
},
{
"epoch": 0.040376850605652756,
"grad_norm": 0.6684543490409851,
"learning_rate": 4.516129032258065e-06,
"loss": 1.239,
"step": 15
},
{
"epoch": 0.05383580080753701,
"grad_norm": 0.7191868424415588,
"learning_rate": 6.129032258064517e-06,
"loss": 1.2218,
"step": 20
},
{
"epoch": 0.06729475100942127,
"grad_norm": 0.6867708563804626,
"learning_rate": 7.741935483870968e-06,
"loss": 1.2301,
"step": 25
},
{
"epoch": 0.08075370121130551,
"grad_norm": 0.6379647254943848,
"learning_rate": 9.35483870967742e-06,
"loss": 1.2194,
"step": 30
},
{
"epoch": 0.09421265141318977,
"grad_norm": 0.5798360109329224,
"learning_rate": 1.0967741935483872e-05,
"loss": 1.2002,
"step": 35
},
{
"epoch": 0.10767160161507403,
"grad_norm": 0.5719746351242065,
"learning_rate": 1.2580645161290324e-05,
"loss": 1.1982,
"step": 40
},
{
"epoch": 0.12113055181695828,
"grad_norm": 0.5450358390808105,
"learning_rate": 1.4193548387096776e-05,
"loss": 1.1986,
"step": 45
},
{
"epoch": 0.13458950201884254,
"grad_norm": 0.5413178205490112,
"learning_rate": 1.5806451612903226e-05,
"loss": 1.1955,
"step": 50
},
{
"epoch": 0.1480484522207268,
"grad_norm": 0.5677768588066101,
"learning_rate": 1.741935483870968e-05,
"loss": 1.1496,
"step": 55
},
{
"epoch": 0.16150740242261102,
"grad_norm": 0.5705103278160095,
"learning_rate": 1.903225806451613e-05,
"loss": 1.1377,
"step": 60
},
{
"epoch": 0.17496635262449528,
"grad_norm": 0.6191664934158325,
"learning_rate": 2.0645161290322582e-05,
"loss": 1.1242,
"step": 65
},
{
"epoch": 0.18842530282637954,
"grad_norm": 0.41417160630226135,
"learning_rate": 2.2258064516129034e-05,
"loss": 1.0923,
"step": 70
},
{
"epoch": 0.2018842530282638,
"grad_norm": 0.5877333283424377,
"learning_rate": 2.3870967741935483e-05,
"loss": 1.1575,
"step": 75
},
{
"epoch": 0.21534320323014805,
"grad_norm": 0.5006933808326721,
"learning_rate": 2.548387096774194e-05,
"loss": 1.0873,
"step": 80
},
{
"epoch": 0.2288021534320323,
"grad_norm": 0.5454208254814148,
"learning_rate": 2.7096774193548387e-05,
"loss": 1.073,
"step": 85
},
{
"epoch": 0.24226110363391656,
"grad_norm": 0.4708370268344879,
"learning_rate": 2.870967741935484e-05,
"loss": 1.1246,
"step": 90
},
{
"epoch": 0.2557200538358008,
"grad_norm": 0.7446239590644836,
"learning_rate": 2.9999976292388853e-05,
"loss": 1.0893,
"step": 95
},
{
"epoch": 0.2691790040376851,
"grad_norm": 0.5205251574516296,
"learning_rate": 2.999914653386729e-05,
"loss": 1.0233,
"step": 100
},
{
"epoch": 0.28263795423956933,
"grad_norm": 0.5542618632316589,
"learning_rate": 2.9997131469727287e-05,
"loss": 1.0534,
"step": 105
},
{
"epoch": 0.2960969044414536,
"grad_norm": 0.5413870811462402,
"learning_rate": 2.9993931259209015e-05,
"loss": 1.0217,
"step": 110
},
{
"epoch": 0.30955585464333785,
"grad_norm": 0.5619310736656189,
"learning_rate": 2.998954615520871e-05,
"loss": 1.0357,
"step": 115
},
{
"epoch": 0.32301480484522205,
"grad_norm": 0.5382521152496338,
"learning_rate": 2.998397650425863e-05,
"loss": 1.0228,
"step": 120
},
{
"epoch": 0.3364737550471063,
"grad_norm": 0.6172555088996887,
"learning_rate": 2.997722274649974e-05,
"loss": 0.9711,
"step": 125
},
{
"epoch": 0.34993270524899056,
"grad_norm": 0.6441250443458557,
"learning_rate": 2.9969285415646874e-05,
"loss": 0.9977,
"step": 130
},
{
"epoch": 0.3633916554508748,
"grad_norm": 0.5927826166152954,
"learning_rate": 2.9960165138946572e-05,
"loss": 0.9778,
"step": 135
},
{
"epoch": 0.3768506056527591,
"grad_norm": 0.6550005078315735,
"learning_rate": 2.9949862637127523e-05,
"loss": 0.9335,
"step": 140
},
{
"epoch": 0.39030955585464333,
"grad_norm": 0.5886467099189758,
"learning_rate": 2.9938378724343604e-05,
"loss": 0.9681,
"step": 145
},
{
"epoch": 0.4037685060565276,
"grad_norm": 0.7246305346488953,
"learning_rate": 2.992571430810954e-05,
"loss": 0.9701,
"step": 150
},
{
"epoch": 0.41722745625841184,
"grad_norm": 0.7467737197875977,
"learning_rate": 2.9911870389229192e-05,
"loss": 0.9141,
"step": 155
},
{
"epoch": 0.4306864064602961,
"grad_norm": 0.7680078744888306,
"learning_rate": 2.9896848061716456e-05,
"loss": 0.8685,
"step": 160
},
{
"epoch": 0.44414535666218036,
"grad_norm": 0.7801365256309509,
"learning_rate": 2.9880648512708834e-05,
"loss": 0.9138,
"step": 165
},
{
"epoch": 0.4576043068640646,
"grad_norm": 0.6462291479110718,
"learning_rate": 2.9863273022373585e-05,
"loss": 0.9223,
"step": 170
},
{
"epoch": 0.47106325706594887,
"grad_norm": 0.6398594975471497,
"learning_rate": 2.9844722963806592e-05,
"loss": 0.9021,
"step": 175
},
{
"epoch": 0.4845222072678331,
"grad_norm": 0.683367133140564,
"learning_rate": 2.9824999802923846e-05,
"loss": 0.8599,
"step": 180
},
{
"epoch": 0.4979811574697174,
"grad_norm": 0.7672361731529236,
"learning_rate": 2.9804105098345575e-05,
"loss": 0.9282,
"step": 185
},
{
"epoch": 0.5114401076716016,
"grad_norm": 0.6266528367996216,
"learning_rate": 2.9782040501273126e-05,
"loss": 0.8664,
"step": 190
},
{
"epoch": 0.5248990578734859,
"grad_norm": 0.7748366594314575,
"learning_rate": 2.975880775535843e-05,
"loss": 0.8588,
"step": 195
},
{
"epoch": 0.5383580080753702,
"grad_norm": 0.892792820930481,
"learning_rate": 2.9734408696566224e-05,
"loss": 0.8517,
"step": 200
},
{
"epoch": 0.5518169582772544,
"grad_norm": 0.8755563497543335,
"learning_rate": 2.9708845253028988e-05,
"loss": 0.9136,
"step": 205
},
{
"epoch": 0.5652759084791387,
"grad_norm": 0.7859370708465576,
"learning_rate": 2.9682119444894545e-05,
"loss": 0.7955,
"step": 210
},
{
"epoch": 0.5787348586810229,
"grad_norm": 0.7947808504104614,
"learning_rate": 2.9654233384166435e-05,
"loss": 0.8695,
"step": 215
},
{
"epoch": 0.5921938088829072,
"grad_norm": 0.7226110696792603,
"learning_rate": 2.9625189274537004e-05,
"loss": 0.7672,
"step": 220
},
{
"epoch": 0.6056527590847914,
"grad_norm": 1.0677162408828735,
"learning_rate": 2.9594989411213266e-05,
"loss": 0.816,
"step": 225
},
{
"epoch": 0.6191117092866757,
"grad_norm": 0.8201079964637756,
"learning_rate": 2.9563636180735524e-05,
"loss": 0.7877,
"step": 230
},
{
"epoch": 0.6325706594885598,
"grad_norm": 0.961162805557251,
"learning_rate": 2.9531132060788763e-05,
"loss": 0.7566,
"step": 235
},
{
"epoch": 0.6460296096904441,
"grad_norm": 0.9038267731666565,
"learning_rate": 2.9497479620006873e-05,
"loss": 0.8061,
"step": 240
},
{
"epoch": 0.6594885598923284,
"grad_norm": 0.9068340063095093,
"learning_rate": 2.9462681517769638e-05,
"loss": 0.787,
"step": 245
},
{
"epoch": 0.6729475100942126,
"grad_norm": 0.8361156582832336,
"learning_rate": 2.94267405039926e-05,
"loss": 0.7835,
"step": 250
},
{
"epoch": 0.6864064602960969,
"grad_norm": 0.8174508213996887,
"learning_rate": 2.938965941890972e-05,
"loss": 0.7641,
"step": 255
},
{
"epoch": 0.6998654104979811,
"grad_norm": 0.8101845383644104,
"learning_rate": 2.9351441192848972e-05,
"loss": 0.76,
"step": 260
},
{
"epoch": 0.7133243606998654,
"grad_norm": 0.8047847747802734,
"learning_rate": 2.9312088846000733e-05,
"loss": 0.7404,
"step": 265
},
{
"epoch": 0.7267833109017496,
"grad_norm": 0.8787263035774231,
"learning_rate": 2.927160548817913e-05,
"loss": 0.7335,
"step": 270
},
{
"epoch": 0.7402422611036339,
"grad_norm": 0.9566526412963867,
"learning_rate": 2.9229994318576295e-05,
"loss": 0.7546,
"step": 275
},
{
"epoch": 0.7537012113055181,
"grad_norm": 0.9340652227401733,
"learning_rate": 2.9187258625509518e-05,
"loss": 0.6814,
"step": 280
},
{
"epoch": 0.7671601615074024,
"grad_norm": 1.0249980688095093,
"learning_rate": 2.9143401786161445e-05,
"loss": 0.7116,
"step": 285
},
{
"epoch": 0.7806191117092867,
"grad_norm": 0.8947973847389221,
"learning_rate": 2.9098427266313138e-05,
"loss": 0.6845,
"step": 290
},
{
"epoch": 0.7940780619111709,
"grad_norm": 0.9918004274368286,
"learning_rate": 2.9052338620070215e-05,
"loss": 0.7287,
"step": 295
},
{
"epoch": 0.8075370121130552,
"grad_norm": 0.9002223610877991,
"learning_rate": 2.9005139489582002e-05,
"loss": 0.7129,
"step": 300
},
{
"epoch": 0.8209959623149394,
"grad_norm": 0.8825509548187256,
"learning_rate": 2.8956833604753688e-05,
"loss": 0.6633,
"step": 305
},
{
"epoch": 0.8344549125168237,
"grad_norm": 0.8505289554595947,
"learning_rate": 2.8907424782951587e-05,
"loss": 0.6826,
"step": 310
},
{
"epoch": 0.847913862718708,
"grad_norm": 1.0375418663024902,
"learning_rate": 2.8856916928701467e-05,
"loss": 0.7003,
"step": 315
},
{
"epoch": 0.8613728129205922,
"grad_norm": 1.0597079992294312,
"learning_rate": 2.8805314033379985e-05,
"loss": 0.7151,
"step": 320
},
{
"epoch": 0.8748317631224765,
"grad_norm": 1.1800124645233154,
"learning_rate": 2.8752620174899284e-05,
"loss": 0.6594,
"step": 325
},
{
"epoch": 0.8882907133243607,
"grad_norm": 0.9276929497718811,
"learning_rate": 2.8698839517384726e-05,
"loss": 0.651,
"step": 330
},
{
"epoch": 0.901749663526245,
"grad_norm": 1.094962239265442,
"learning_rate": 2.8643976310845828e-05,
"loss": 0.6412,
"step": 335
},
{
"epoch": 0.9152086137281292,
"grad_norm": 0.9263754487037659,
"learning_rate": 2.8588034890840408e-05,
"loss": 0.6269,
"step": 340
},
{
"epoch": 0.9286675639300135,
"grad_norm": 0.9816908240318298,
"learning_rate": 2.8531019678131952e-05,
"loss": 0.6577,
"step": 345
},
{
"epoch": 0.9421265141318977,
"grad_norm": 1.0188283920288086,
"learning_rate": 2.847293517834029e-05,
"loss": 0.66,
"step": 350
},
{
"epoch": 0.955585464333782,
"grad_norm": 1.0503355264663696,
"learning_rate": 2.8413785981585518e-05,
"loss": 0.621,
"step": 355
},
{
"epoch": 0.9690444145356663,
"grad_norm": 0.9092134237289429,
"learning_rate": 2.8353576762125274e-05,
"loss": 0.6343,
"step": 360
},
{
"epoch": 0.9825033647375505,
"grad_norm": 0.9565970301628113,
"learning_rate": 2.8292312277985354e-05,
"loss": 0.6538,
"step": 365
},
{
"epoch": 0.9959623149394348,
"grad_norm": 1.014952301979065,
"learning_rate": 2.82299973705837e-05,
"loss": 0.6081,
"step": 370
}
],
"logging_steps": 5,
"max_steps": 1860,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 2000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.29235739645313e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}