klora_2000_skill / 110 /trainer_state.json
RayDu0010's picture
Upload folder using huggingface_hub
86c6999 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 266,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.018832391713747645,
"grad_norm": 1.1089932918548584,
"learning_rate": 1.791044776119403e-06,
"loss": 1.365,
"step": 5
},
{
"epoch": 0.03766478342749529,
"grad_norm": 0.8269509673118591,
"learning_rate": 4.029850746268657e-06,
"loss": 1.3307,
"step": 10
},
{
"epoch": 0.05649717514124294,
"grad_norm": 0.6606281995773315,
"learning_rate": 6.268656716417911e-06,
"loss": 1.2869,
"step": 15
},
{
"epoch": 0.07532956685499058,
"grad_norm": 0.5472633242607117,
"learning_rate": 8.507462686567164e-06,
"loss": 1.2774,
"step": 20
},
{
"epoch": 0.09416195856873823,
"grad_norm": 0.6873501539230347,
"learning_rate": 1.0746268656716418e-05,
"loss": 1.2985,
"step": 25
},
{
"epoch": 0.11299435028248588,
"grad_norm": 0.5600975751876831,
"learning_rate": 1.2985074626865672e-05,
"loss": 1.217,
"step": 30
},
{
"epoch": 0.1318267419962335,
"grad_norm": 0.7250403761863708,
"learning_rate": 1.5223880597014927e-05,
"loss": 1.233,
"step": 35
},
{
"epoch": 0.15065913370998116,
"grad_norm": 0.5497453808784485,
"learning_rate": 1.746268656716418e-05,
"loss": 1.2169,
"step": 40
},
{
"epoch": 0.1694915254237288,
"grad_norm": 0.6821231842041016,
"learning_rate": 1.9701492537313435e-05,
"loss": 1.1701,
"step": 45
},
{
"epoch": 0.18832391713747645,
"grad_norm": 0.5772640705108643,
"learning_rate": 2.194029850746269e-05,
"loss": 1.1552,
"step": 50
},
{
"epoch": 0.2071563088512241,
"grad_norm": 0.5170062780380249,
"learning_rate": 2.417910447761194e-05,
"loss": 1.1947,
"step": 55
},
{
"epoch": 0.22598870056497175,
"grad_norm": 0.5546678304672241,
"learning_rate": 2.6417910447761193e-05,
"loss": 1.0869,
"step": 60
},
{
"epoch": 0.2448210922787194,
"grad_norm": 0.6221190690994263,
"learning_rate": 2.8656716417910447e-05,
"loss": 1.1683,
"step": 65
},
{
"epoch": 0.263653483992467,
"grad_norm": 0.7900596857070923,
"learning_rate": 2.9999814384857414e-05,
"loss": 1.1623,
"step": 70
},
{
"epoch": 0.2824858757062147,
"grad_norm": 0.5770227909088135,
"learning_rate": 2.99977262672591e-05,
"loss": 1.1033,
"step": 75
},
{
"epoch": 0.3013182674199623,
"grad_norm": 0.5991471409797668,
"learning_rate": 2.9993318337195934e-05,
"loss": 1.1221,
"step": 80
},
{
"epoch": 0.32015065913371,
"grad_norm": 0.6063827872276306,
"learning_rate": 2.998659127647601e-05,
"loss": 1.0186,
"step": 85
},
{
"epoch": 0.3389830508474576,
"grad_norm": 0.690599262714386,
"learning_rate": 2.9977546125625006e-05,
"loss": 1.0455,
"step": 90
},
{
"epoch": 0.3578154425612053,
"grad_norm": 0.681769073009491,
"learning_rate": 2.9966184283725243e-05,
"loss": 1.014,
"step": 95
},
{
"epoch": 0.3766478342749529,
"grad_norm": 0.6798896789550781,
"learning_rate": 2.9952507508199284e-05,
"loss": 0.9816,
"step": 100
},
{
"epoch": 0.3954802259887006,
"grad_norm": 0.6588433384895325,
"learning_rate": 2.9936517914538085e-05,
"loss": 1.0095,
"step": 105
},
{
"epoch": 0.4143126177024482,
"grad_norm": 0.7236828207969666,
"learning_rate": 2.991821797597379e-05,
"loss": 0.9913,
"step": 110
},
{
"epoch": 0.4331450094161959,
"grad_norm": 0.6340996026992798,
"learning_rate": 2.989761052309717e-05,
"loss": 0.9752,
"step": 115
},
{
"epoch": 0.4519774011299435,
"grad_norm": 0.7232135534286499,
"learning_rate": 2.98746987434198e-05,
"loss": 0.965,
"step": 120
},
{
"epoch": 0.4708097928436911,
"grad_norm": 0.7878739237785339,
"learning_rate": 2.9849486180881017e-05,
"loss": 0.9779,
"step": 125
},
{
"epoch": 0.4896421845574388,
"grad_norm": 0.7223129868507385,
"learning_rate": 2.982197673529976e-05,
"loss": 0.9159,
"step": 130
},
{
"epoch": 0.5084745762711864,
"grad_norm": 0.7257573008537292,
"learning_rate": 2.9792174661771332e-05,
"loss": 0.9403,
"step": 135
},
{
"epoch": 0.527306967984934,
"grad_norm": 0.8936464190483093,
"learning_rate": 2.976008457000927e-05,
"loss": 0.8969,
"step": 140
},
{
"epoch": 0.5461393596986818,
"grad_norm": 0.9090555906295776,
"learning_rate": 2.972571142363229e-05,
"loss": 0.8963,
"step": 145
},
{
"epoch": 0.5649717514124294,
"grad_norm": 0.8844392895698547,
"learning_rate": 2.9689060539396545e-05,
"loss": 0.8657,
"step": 150
},
{
"epoch": 0.583804143126177,
"grad_norm": 0.8749438524246216,
"learning_rate": 2.9650137586373236e-05,
"loss": 0.8104,
"step": 155
},
{
"epoch": 0.6026365348399246,
"grad_norm": 0.9088373184204102,
"learning_rate": 2.9608948585071747e-05,
"loss": 0.842,
"step": 160
},
{
"epoch": 0.6214689265536724,
"grad_norm": 0.8506441116333008,
"learning_rate": 2.9565499906508376e-05,
"loss": 0.8132,
"step": 165
},
{
"epoch": 0.64030131826742,
"grad_norm": 0.9221051335334778,
"learning_rate": 2.9519798271220915e-05,
"loss": 0.7725,
"step": 170
},
{
"epoch": 0.6591337099811676,
"grad_norm": 1.0185768604278564,
"learning_rate": 2.9471850748229117e-05,
"loss": 0.7874,
"step": 175
},
{
"epoch": 0.6779661016949152,
"grad_norm": 0.8905505537986755,
"learning_rate": 2.9421664753941277e-05,
"loss": 0.7491,
"step": 180
},
{
"epoch": 0.696798493408663,
"grad_norm": 0.9472977519035339,
"learning_rate": 2.936924805100708e-05,
"loss": 0.7542,
"step": 185
},
{
"epoch": 0.7156308851224106,
"grad_norm": 1.000342845916748,
"learning_rate": 2.9314608747116894e-05,
"loss": 0.7364,
"step": 190
},
{
"epoch": 0.7344632768361582,
"grad_norm": 0.8593543171882629,
"learning_rate": 2.9257755293747702e-05,
"loss": 0.7712,
"step": 195
},
{
"epoch": 0.7532956685499058,
"grad_norm": 0.9145693778991699,
"learning_rate": 2.919869648485582e-05,
"loss": 0.7399,
"step": 200
},
{
"epoch": 0.7721280602636534,
"grad_norm": 0.9073010683059692,
"learning_rate": 2.9137441455516706e-05,
"loss": 0.6409,
"step": 205
},
{
"epoch": 0.7909604519774012,
"grad_norm": 0.8353155851364136,
"learning_rate": 2.9073999680511935e-05,
"loss": 0.738,
"step": 210
},
{
"epoch": 0.8097928436911488,
"grad_norm": 1.0048359632492065,
"learning_rate": 2.9008380972863693e-05,
"loss": 0.7345,
"step": 215
},
{
"epoch": 0.8286252354048964,
"grad_norm": 1.1321660280227661,
"learning_rate": 2.8940595482316883e-05,
"loss": 0.6374,
"step": 220
},
{
"epoch": 0.847457627118644,
"grad_norm": 0.9749424457550049,
"learning_rate": 2.8870653693769233e-05,
"loss": 0.6937,
"step": 225
},
{
"epoch": 0.8662900188323918,
"grad_norm": 0.9839215874671936,
"learning_rate": 2.8798566425649478e-05,
"loss": 0.6885,
"step": 230
},
{
"epoch": 0.8851224105461394,
"grad_norm": 1.0036780834197998,
"learning_rate": 2.8724344828244014e-05,
"loss": 0.6463,
"step": 235
},
{
"epoch": 0.903954802259887,
"grad_norm": 0.9374542832374573,
"learning_rate": 2.864800038197219e-05,
"loss": 0.6853,
"step": 240
},
{
"epoch": 0.9227871939736346,
"grad_norm": 1.1694684028625488,
"learning_rate": 2.8569544895610536e-05,
"loss": 0.6399,
"step": 245
},
{
"epoch": 0.9416195856873822,
"grad_norm": 1.0283533334732056,
"learning_rate": 2.848899050446624e-05,
"loss": 0.6486,
"step": 250
},
{
"epoch": 0.96045197740113,
"grad_norm": 0.9280611276626587,
"learning_rate": 2.840634966850006e-05,
"loss": 0.6377,
"step": 255
},
{
"epoch": 0.9792843691148776,
"grad_norm": 0.8847799897193909,
"learning_rate": 2.832163517039903e-05,
"loss": 0.6585,
"step": 260
},
{
"epoch": 0.9981167608286252,
"grad_norm": 1.08964204788208,
"learning_rate": 2.8234860113599316e-05,
"loss": 0.6149,
"step": 265
}
],
"logging_steps": 5,
"max_steps": 1330,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 2000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.639319352668324e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}