final_score1 / trainer_state.json
LHL3341's picture
upload
25c89e7 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 837,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.03590664272890485,
"grad_norm": 2.185314827906911,
"learning_rate": 1.0714285714285714e-06,
"loss": 0.5255,
"step": 10
},
{
"epoch": 0.0718132854578097,
"grad_norm": 0.9772670854647908,
"learning_rate": 2.261904761904762e-06,
"loss": 0.4864,
"step": 20
},
{
"epoch": 0.10771992818671454,
"grad_norm": 0.5510959017484935,
"learning_rate": 3.4523809523809528e-06,
"loss": 0.4451,
"step": 30
},
{
"epoch": 0.1436265709156194,
"grad_norm": 0.36493152532802847,
"learning_rate": 4.642857142857144e-06,
"loss": 0.4149,
"step": 40
},
{
"epoch": 0.17953321364452424,
"grad_norm": 0.2869233938041407,
"learning_rate": 5.833333333333334e-06,
"loss": 0.3955,
"step": 50
},
{
"epoch": 0.21543985637342908,
"grad_norm": 0.2374087352719959,
"learning_rate": 7.023809523809524e-06,
"loss": 0.3807,
"step": 60
},
{
"epoch": 0.2513464991023339,
"grad_norm": 0.19085750578643032,
"learning_rate": 8.214285714285714e-06,
"loss": 0.3719,
"step": 70
},
{
"epoch": 0.2872531418312388,
"grad_norm": 0.18588659208121952,
"learning_rate": 9.404761904761905e-06,
"loss": 0.3638,
"step": 80
},
{
"epoch": 0.3231597845601436,
"grad_norm": 0.19078133253606472,
"learning_rate": 9.998912137362367e-06,
"loss": 0.3602,
"step": 90
},
{
"epoch": 0.3590664272890485,
"grad_norm": 0.18585754435457444,
"learning_rate": 9.990212076323587e-06,
"loss": 0.357,
"step": 100
},
{
"epoch": 0.39497307001795334,
"grad_norm": 0.1879818930694756,
"learning_rate": 9.97282709575282e-06,
"loss": 0.3511,
"step": 110
},
{
"epoch": 0.43087971274685816,
"grad_norm": 0.17595485058415583,
"learning_rate": 9.946787452311507e-06,
"loss": 0.3474,
"step": 120
},
{
"epoch": 0.466786355475763,
"grad_norm": 0.18168043823227237,
"learning_rate": 9.912138465157325e-06,
"loss": 0.3466,
"step": 130
},
{
"epoch": 0.5026929982046678,
"grad_norm": 0.19246537631691138,
"learning_rate": 9.86894043707114e-06,
"loss": 0.3439,
"step": 140
},
{
"epoch": 0.5385996409335727,
"grad_norm": 0.20913120656523562,
"learning_rate": 9.81726854950659e-06,
"loss": 0.3392,
"step": 150
},
{
"epoch": 0.5745062836624776,
"grad_norm": 0.1943925752607722,
"learning_rate": 9.757212731744973e-06,
"loss": 0.3382,
"step": 160
},
{
"epoch": 0.6104129263913824,
"grad_norm": 0.22384055375769804,
"learning_rate": 9.688877504383158e-06,
"loss": 0.3386,
"step": 170
},
{
"epoch": 0.6463195691202872,
"grad_norm": 0.19509930129044936,
"learning_rate": 9.612381797426874e-06,
"loss": 0.3378,
"step": 180
},
{
"epoch": 0.6822262118491921,
"grad_norm": 0.23177190223223015,
"learning_rate": 9.52785874330602e-06,
"loss": 0.333,
"step": 190
},
{
"epoch": 0.718132854578097,
"grad_norm": 0.2006246018340954,
"learning_rate": 9.435455445172214e-06,
"loss": 0.3311,
"step": 200
},
{
"epoch": 0.7540394973070018,
"grad_norm": 0.2042999512470892,
"learning_rate": 9.33533272088179e-06,
"loss": 0.3313,
"step": 210
},
{
"epoch": 0.7899461400359067,
"grad_norm": 0.21930626870236752,
"learning_rate": 9.227664823109884e-06,
"loss": 0.3325,
"step": 220
},
{
"epoch": 0.8258527827648114,
"grad_norm": 0.23685553683202806,
"learning_rate": 9.11263913608266e-06,
"loss": 0.3308,
"step": 230
},
{
"epoch": 0.8617594254937163,
"grad_norm": 0.21203795699103023,
"learning_rate": 8.990455849455522e-06,
"loss": 0.3306,
"step": 240
},
{
"epoch": 0.8976660682226212,
"grad_norm": 0.2617908535020091,
"learning_rate": 8.861327609904859e-06,
"loss": 0.328,
"step": 250
},
{
"epoch": 0.933572710951526,
"grad_norm": 0.20428110286245318,
"learning_rate": 8.725479151039714e-06,
"loss": 0.3249,
"step": 260
},
{
"epoch": 0.9694793536804309,
"grad_norm": 0.23276915935689596,
"learning_rate": 8.583146902277464e-06,
"loss": 0.3287,
"step": 270
},
{
"epoch": 1.0035906642728905,
"grad_norm": 0.2445973206934874,
"learning_rate": 8.434578577364218e-06,
"loss": 0.3246,
"step": 280
},
{
"epoch": 1.0394973070017954,
"grad_norm": 0.21984748841500337,
"learning_rate": 8.28003274325608e-06,
"loss": 0.3196,
"step": 290
},
{
"epoch": 1.0754039497307002,
"grad_norm": 0.21007463727725653,
"learning_rate": 8.119778370111566e-06,
"loss": 0.3129,
"step": 300
},
{
"epoch": 1.111310592459605,
"grad_norm": 0.22501623459913725,
"learning_rate": 7.954094363178421e-06,
"loss": 0.3145,
"step": 310
},
{
"epoch": 1.14721723518851,
"grad_norm": 0.20904438470673392,
"learning_rate": 7.783269077389447e-06,
"loss": 0.3132,
"step": 320
},
{
"epoch": 1.1831238779174147,
"grad_norm": 0.1929405792634159,
"learning_rate": 7.607599815512226e-06,
"loss": 0.3119,
"step": 330
},
{
"epoch": 1.2190305206463194,
"grad_norm": 0.20576378490371444,
"learning_rate": 7.427392310726088e-06,
"loss": 0.3145,
"step": 340
},
{
"epoch": 1.2549371633752244,
"grad_norm": 0.22314438193898609,
"learning_rate": 7.242960194526893e-06,
"loss": 0.314,
"step": 350
},
{
"epoch": 1.2908438061041292,
"grad_norm": 0.18526094334512155,
"learning_rate": 7.054624450885621e-06,
"loss": 0.3147,
"step": 360
},
{
"epoch": 1.3267504488330342,
"grad_norm": 0.20944522582021186,
"learning_rate": 6.862712857610812e-06,
"loss": 0.3125,
"step": 370
},
{
"epoch": 1.362657091561939,
"grad_norm": 0.20970244825805384,
"learning_rate": 6.667559415887055e-06,
"loss": 0.3119,
"step": 380
},
{
"epoch": 1.398563734290844,
"grad_norm": 0.18929491036558882,
"learning_rate": 6.469503768982379e-06,
"loss": 0.3144,
"step": 390
},
{
"epoch": 1.4344703770197487,
"grad_norm": 0.18469452603163652,
"learning_rate": 6.2688906111362115e-06,
"loss": 0.3131,
"step": 400
},
{
"epoch": 1.4703770197486534,
"grad_norm": 0.18021217056496155,
"learning_rate": 6.066069087656665e-06,
"loss": 0.3128,
"step": 410
},
{
"epoch": 1.5062836624775584,
"grad_norm": 0.31157448102210894,
"learning_rate": 5.8613921872712446e-06,
"loss": 0.3115,
"step": 420
},
{
"epoch": 1.5421903052064632,
"grad_norm": 0.19847526072216384,
"learning_rate": 5.655216127788472e-06,
"loss": 0.3088,
"step": 430
},
{
"epoch": 1.578096947935368,
"grad_norm": 0.19215915923161053,
"learning_rate": 5.447899736139676e-06,
"loss": 0.3063,
"step": 440
},
{
"epoch": 1.614003590664273,
"grad_norm": 0.18221395113654418,
"learning_rate": 5.239803823879878e-06,
"loss": 0.3082,
"step": 450
},
{
"epoch": 1.6499102333931779,
"grad_norm": 0.1749484556213854,
"learning_rate": 5.03129055923465e-06,
"loss": 0.3089,
"step": 460
},
{
"epoch": 1.6858168761220824,
"grad_norm": 0.17613434487474727,
"learning_rate": 4.822722836785842e-06,
"loss": 0.3068,
"step": 470
},
{
"epoch": 1.7217235188509874,
"grad_norm": 0.1751185479216866,
"learning_rate": 4.614463645893175e-06,
"loss": 0.3068,
"step": 480
},
{
"epoch": 1.7576301615798924,
"grad_norm": 0.1748572053302032,
"learning_rate": 4.4068754389508616e-06,
"loss": 0.3093,
"step": 490
},
{
"epoch": 1.7935368043087971,
"grad_norm": 0.16507354789633685,
"learning_rate": 4.2003195005787745e-06,
"loss": 0.3099,
"step": 500
},
{
"epoch": 1.829443447037702,
"grad_norm": 0.170392010993257,
"learning_rate": 3.995155318845994e-06,
"loss": 0.3074,
"step": 510
},
{
"epoch": 1.8653500897666069,
"grad_norm": 0.18627037895031676,
"learning_rate": 3.791739959621054e-06,
"loss": 0.3091,
"step": 520
},
{
"epoch": 1.9012567324955116,
"grad_norm": 0.17081506891008494,
"learning_rate": 3.5904274451377634e-06,
"loss": 0.3071,
"step": 530
},
{
"epoch": 1.9371633752244164,
"grad_norm": 0.1697011202581245,
"learning_rate": 3.3915681378581416e-06,
"loss": 0.3093,
"step": 540
},
{
"epoch": 1.9730700179533214,
"grad_norm": 0.17582268731336928,
"learning_rate": 3.195508130704795e-06,
"loss": 0.3072,
"step": 550
},
{
"epoch": 2.007181328545781,
"grad_norm": 0.16800018749204126,
"learning_rate": 3.0025886447239474e-06,
"loss": 0.3027,
"step": 560
},
{
"epoch": 2.043087971274686,
"grad_norm": 0.17959400944310325,
"learning_rate": 2.8131454352274525e-06,
"loss": 0.2975,
"step": 570
},
{
"epoch": 2.078994614003591,
"grad_norm": 0.1785065928144257,
"learning_rate": 2.627508207447308e-06,
"loss": 0.2986,
"step": 580
},
{
"epoch": 2.1149012567324954,
"grad_norm": 0.15727463247343326,
"learning_rate": 2.4460000427196916e-06,
"loss": 0.2949,
"step": 590
},
{
"epoch": 2.1508078994614004,
"grad_norm": 0.15805698099204835,
"learning_rate": 2.268936836197144e-06,
"loss": 0.2952,
"step": 600
},
{
"epoch": 2.1867145421903054,
"grad_norm": 0.16920663238784117,
"learning_rate": 2.0966267470675273e-06,
"loss": 0.2985,
"step": 610
},
{
"epoch": 2.22262118491921,
"grad_norm": 0.16646936573182844,
"learning_rate": 1.929369662236604e-06,
"loss": 0.2979,
"step": 620
},
{
"epoch": 2.258527827648115,
"grad_norm": 0.3421880613089778,
"learning_rate": 1.7674566744075882e-06,
"loss": 0.298,
"step": 630
},
{
"epoch": 2.29443447037702,
"grad_norm": 0.15025140288330655,
"learning_rate": 1.6111695754660667e-06,
"loss": 0.2942,
"step": 640
},
{
"epoch": 2.3303411131059244,
"grad_norm": 0.17473486663498805,
"learning_rate": 1.4607803660519803e-06,
"loss": 0.2952,
"step": 650
},
{
"epoch": 2.3662477558348294,
"grad_norm": 0.1589150679388623,
"learning_rate": 1.3165507821721906e-06,
"loss": 0.2983,
"step": 660
},
{
"epoch": 2.4021543985637344,
"grad_norm": 0.1703479004667421,
"learning_rate": 1.1787318396775188e-06,
"loss": 0.298,
"step": 670
},
{
"epoch": 2.438061041292639,
"grad_norm": 0.1513614749600422,
"learning_rate": 1.0475633973970573e-06,
"loss": 0.2959,
"step": 680
},
{
"epoch": 2.473967684021544,
"grad_norm": 0.1876551965771974,
"learning_rate": 9.232737396900543e-07,
"loss": 0.2996,
"step": 690
},
{
"epoch": 2.509874326750449,
"grad_norm": 0.15388873373154505,
"learning_rate": 8.060791791418887e-07,
"loss": 0.2966,
"step": 700
},
{
"epoch": 2.545780969479354,
"grad_norm": 0.15535759307395802,
"learning_rate": 6.96183680095639e-07,
"loss": 0.2991,
"step": 710
},
{
"epoch": 2.5816876122082584,
"grad_norm": 0.14993765716059407,
"learning_rate": 5.937785036743893e-07,
"loss": 0.298,
"step": 720
},
{
"epoch": 2.6175942549371634,
"grad_norm": 0.14735943617696928,
"learning_rate": 4.990418749121179e-07,
"loss": 0.2966,
"step": 730
},
{
"epoch": 2.6535008976660683,
"grad_norm": 0.1548863032393796,
"learning_rate": 4.121386725724835e-07,
"loss": 0.2978,
"step": 740
},
{
"epoch": 2.6894075403949733,
"grad_norm": 0.15810495711820732,
"learning_rate": 3.3322014219532575e-07,
"loss": 0.2976,
"step": 750
},
{
"epoch": 2.725314183123878,
"grad_norm": 0.15072810300321354,
"learning_rate": 2.6242363287030617e-07,
"loss": 0.296,
"step": 760
},
{
"epoch": 2.761220825852783,
"grad_norm": 0.2424841654880629,
"learning_rate": 1.9987235819581118e-07,
"loss": 0.2969,
"step": 770
},
{
"epoch": 2.797127468581688,
"grad_norm": 0.22116921613107723,
"learning_rate": 1.4567518183912887e-07,
"loss": 0.2964,
"step": 780
},
{
"epoch": 2.8330341113105924,
"grad_norm": 0.14618317771038028,
"learning_rate": 9.992642807111486e-08,
"loss": 0.2978,
"step": 790
},
{
"epoch": 2.8689407540394973,
"grad_norm": 0.1636178141701447,
"learning_rate": 6.270571760509547e-08,
"loss": 0.2991,
"step": 800
},
{
"epoch": 2.9048473967684023,
"grad_norm": 0.14508080741570623,
"learning_rate": 3.4077829025703226e-08,
"loss": 0.2942,
"step": 810
},
{
"epoch": 2.940754039497307,
"grad_norm": 0.1398003635276662,
"learning_rate": 1.4092586048820578e-08,
"loss": 0.2962,
"step": 820
},
{
"epoch": 2.976660682226212,
"grad_norm": 0.1587618807983056,
"learning_rate": 2.784770808839654e-09,
"loss": 0.2982,
"step": 830
}
],
"logging_steps": 10,
"max_steps": 837,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 10000000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8093209209077760.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}