SeproLM-Adapters / trainer_state.json
Hariharan05's picture
Upload 2 files
81c4a6d verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.1907032181168058,
"eval_steps": 500,
"global_step": 500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.023837902264600714,
"grad_norm": 8.34217357635498,
"learning_rate": 1.6e-07,
"loss": 3.2614,
"step": 10
},
{
"epoch": 0.04767580452920143,
"grad_norm": 10.162466049194336,
"learning_rate": 3.6e-07,
"loss": 3.2005,
"step": 20
},
{
"epoch": 0.07151370679380215,
"grad_norm": 8.322413444519043,
"learning_rate": 5.6e-07,
"loss": 3.3002,
"step": 30
},
{
"epoch": 0.09535160905840286,
"grad_norm": 6.742789268493652,
"learning_rate": 7.599999999999999e-07,
"loss": 3.1868,
"step": 40
},
{
"epoch": 0.11918951132300358,
"grad_norm": 6.748029708862305,
"learning_rate": 9.6e-07,
"loss": 3.1789,
"step": 50
},
{
"epoch": 0.1430274135876043,
"grad_norm": 5.005261421203613,
"learning_rate": 9.992203820909905e-07,
"loss": 3.0969,
"step": 60
},
{
"epoch": 0.16686531585220502,
"grad_norm": 6.2596893310546875,
"learning_rate": 9.960573506572389e-07,
"loss": 3.021,
"step": 70
},
{
"epoch": 0.1907032181168057,
"grad_norm": 5.589771747589111,
"learning_rate": 9.904775776745956e-07,
"loss": 3.1221,
"step": 80
},
{
"epoch": 0.21454112038140644,
"grad_norm": 7.189596176147461,
"learning_rate": 9.825082472361556e-07,
"loss": 3.0537,
"step": 90
},
{
"epoch": 0.23837902264600716,
"grad_norm": 5.899924278259277,
"learning_rate": 9.721881851187405e-07,
"loss": 3.0154,
"step": 100
},
{
"epoch": 0.2622169249106079,
"grad_norm": 4.845108509063721,
"learning_rate": 9.595676696276171e-07,
"loss": 3.0747,
"step": 110
},
{
"epoch": 0.2860548271752086,
"grad_norm": 5.564320087432861,
"learning_rate": 9.447081866456487e-07,
"loss": 2.9519,
"step": 120
},
{
"epoch": 0.3098927294398093,
"grad_norm": 6.226819038391113,
"learning_rate": 9.276821300802533e-07,
"loss": 3.0256,
"step": 130
},
{
"epoch": 0.33373063170441003,
"grad_norm": 5.26694393157959,
"learning_rate": 9.085724491675642e-07,
"loss": 2.843,
"step": 140
},
{
"epoch": 0.3575685339690107,
"grad_norm": 5.471522808074951,
"learning_rate": 8.874722443520898e-07,
"loss": 2.8078,
"step": 150
},
{
"epoch": 0.3814064362336114,
"grad_norm": 5.0895819664001465,
"learning_rate": 8.644843137107057e-07,
"loss": 2.8523,
"step": 160
},
{
"epoch": 0.4052443384982122,
"grad_norm": 4.7197794914245605,
"learning_rate": 8.397206521307583e-07,
"loss": 2.754,
"step": 170
},
{
"epoch": 0.42908224076281287,
"grad_norm": 5.758249282836914,
"learning_rate": 8.133019056822302e-07,
"loss": 2.7133,
"step": 180
},
{
"epoch": 0.45292014302741357,
"grad_norm": 4.340641021728516,
"learning_rate": 7.853567838422159e-07,
"loss": 2.6577,
"step": 190
},
{
"epoch": 0.4767580452920143,
"grad_norm": 4.795559406280518,
"learning_rate": 7.560214324352858e-07,
"loss": 2.7179,
"step": 200
},
{
"epoch": 0.5005959475566151,
"grad_norm": 5.540432929992676,
"learning_rate": 7.254387703447153e-07,
"loss": 2.6652,
"step": 210
},
{
"epoch": 0.5244338498212158,
"grad_norm": 4.245532035827637,
"learning_rate": 6.937577932260514e-07,
"loss": 2.6331,
"step": 220
},
{
"epoch": 0.5482717520858165,
"grad_norm": 4.427391052246094,
"learning_rate": 6.611328476152556e-07,
"loss": 2.6846,
"step": 230
},
{
"epoch": 0.5721096543504172,
"grad_norm": 5.0143842697143555,
"learning_rate": 6.277228789678953e-07,
"loss": 2.605,
"step": 240
},
{
"epoch": 0.5959475566150179,
"grad_norm": 5.8499603271484375,
"learning_rate": 5.936906572928624e-07,
"loss": 2.4651,
"step": 250
},
{
"epoch": 0.6197854588796186,
"grad_norm": 4.577277183532715,
"learning_rate": 5.592019841532506e-07,
"loss": 2.4897,
"step": 260
},
{
"epoch": 0.6436233611442194,
"grad_norm": 4.542954444885254,
"learning_rate": 5.244248848978067e-07,
"loss": 2.4942,
"step": 270
},
{
"epoch": 0.6674612634088201,
"grad_norm": 5.428004741668701,
"learning_rate": 4.895287900583216e-07,
"loss": 2.5005,
"step": 280
},
{
"epoch": 0.6912991656734208,
"grad_norm": 4.4296956062316895,
"learning_rate": 4.5468370990110997e-07,
"loss": 2.3997,
"step": 290
},
{
"epoch": 0.7151370679380215,
"grad_norm": 4.045319557189941,
"learning_rate": 4.200594061540826e-07,
"loss": 2.3467,
"step": 300
},
{
"epoch": 0.7389749702026222,
"grad_norm": 4.404695510864258,
"learning_rate": 3.8582456494467206e-07,
"loss": 2.4801,
"step": 310
},
{
"epoch": 0.7628128724672228,
"grad_norm": 5.092370510101318,
"learning_rate": 3.521459749779768e-07,
"loss": 2.4103,
"step": 320
},
{
"epoch": 0.7866507747318237,
"grad_norm": 5.434021472930908,
"learning_rate": 3.191877149589539e-07,
"loss": 2.401,
"step": 330
},
{
"epoch": 0.8104886769964244,
"grad_norm": 4.2743353843688965,
"learning_rate": 2.8711035421746363e-07,
"loss": 2.4399,
"step": 340
},
{
"epoch": 0.834326579261025,
"grad_norm": 4.966851711273193,
"learning_rate": 2.5607017043063353e-07,
"loss": 2.3473,
"step": 350
},
{
"epoch": 0.8581644815256257,
"grad_norm": 5.600205898284912,
"learning_rate": 2.262183882537249e-07,
"loss": 2.4595,
"step": 360
},
{
"epoch": 0.8820023837902264,
"grad_norm": 4.966123580932617,
"learning_rate": 1.9770044256881258e-07,
"loss": 2.4161,
"step": 370
},
{
"epoch": 0.9058402860548271,
"grad_norm": 4.942025184631348,
"learning_rate": 1.7065526994065972e-07,
"loss": 2.3971,
"step": 380
},
{
"epoch": 0.929678188319428,
"grad_norm": 5.2323102951049805,
"learning_rate": 1.4521463173173965e-07,
"loss": 2.3786,
"step": 390
},
{
"epoch": 0.9535160905840286,
"grad_norm": 5.717218399047852,
"learning_rate": 1.2150247217412185e-07,
"loss": 2.3473,
"step": 400
},
{
"epoch": 0.9773539928486293,
"grad_norm": 5.0170416831970215,
"learning_rate": 9.963431452563331e-08,
"loss": 2.401,
"step": 410
},
{
"epoch": 1.0,
"grad_norm": 7.488749980926514,
"learning_rate": 7.971669825215787e-08,
"loss": 2.3823,
"step": 420
},
{
"epoch": 1.0238379022646007,
"grad_norm": 6.092284202575684,
"learning_rate": 6.184665997806831e-08,
"loss": 2.3511,
"step": 430
},
{
"epoch": 1.0476758045292014,
"grad_norm": 5.5308709144592285,
"learning_rate": 4.611126073354571e-08,
"loss": 2.4092,
"step": 440
},
{
"epoch": 1.071513706793802,
"grad_norm": 5.2230119705200195,
"learning_rate": 3.258716180199278e-08,
"loss": 2.3573,
"step": 450
},
{
"epoch": 1.0953516090584028,
"grad_norm": 4.9154510498046875,
"learning_rate": 2.1340251233966377e-08,
"loss": 2.3055,
"step": 460
},
{
"epoch": 1.1191895113230035,
"grad_norm": 4.627120494842529,
"learning_rate": 1.2425322847218367e-08,
"loss": 2.3723,
"step": 470
},
{
"epoch": 1.1430274135876042,
"grad_norm": 6.616613864898682,
"learning_rate": 5.8858092767236076e-09,
"loss": 2.2949,
"step": 480
},
{
"epoch": 1.166865315852205,
"grad_norm": 5.475731372833252,
"learning_rate": 1.7535703752478147e-09,
"loss": 2.4503,
"step": 490
},
{
"epoch": 1.1907032181168058,
"grad_norm": 5.741678237915039,
"learning_rate": 4.873799534788059e-11,
"loss": 2.2666,
"step": 500
}
],
"logging_steps": 10,
"max_steps": 500,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7169104310525952.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}