copywriting / trainer_state.json
user1ai's picture
Upload 17 files
c71b46f verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9942004971002486,
"eval_steps": 500,
"global_step": 1200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02,
"grad_norm": 0.7059583067893982,
"learning_rate": 0.0002,
"loss": 1.1809,
"step": 25
},
{
"epoch": 0.04,
"grad_norm": 0.6619198322296143,
"learning_rate": 0.0002,
"loss": 0.5511,
"step": 50
},
{
"epoch": 0.06,
"grad_norm": 0.37209564447402954,
"learning_rate": 0.0002,
"loss": 0.833,
"step": 75
},
{
"epoch": 0.08,
"grad_norm": 0.5066412091255188,
"learning_rate": 0.0002,
"loss": 0.497,
"step": 100
},
{
"epoch": 0.1,
"grad_norm": 0.5964956283569336,
"learning_rate": 0.0002,
"loss": 0.77,
"step": 125
},
{
"epoch": 0.12,
"grad_norm": 0.6351748108863831,
"learning_rate": 0.0002,
"loss": 0.4792,
"step": 150
},
{
"epoch": 0.14,
"grad_norm": 0.33565786480903625,
"learning_rate": 0.0002,
"loss": 0.6867,
"step": 175
},
{
"epoch": 0.17,
"grad_norm": 0.5637449622154236,
"learning_rate": 0.0002,
"loss": 0.4725,
"step": 200
},
{
"epoch": 0.19,
"grad_norm": 0.29314109683036804,
"learning_rate": 0.0002,
"loss": 0.6842,
"step": 225
},
{
"epoch": 0.21,
"grad_norm": 0.5340172052383423,
"learning_rate": 0.0002,
"loss": 0.4269,
"step": 250
},
{
"epoch": 0.23,
"grad_norm": 0.38599228858947754,
"learning_rate": 0.0002,
"loss": 0.6939,
"step": 275
},
{
"epoch": 0.25,
"grad_norm": 0.5987310409545898,
"learning_rate": 0.0002,
"loss": 0.4435,
"step": 300
},
{
"epoch": 0.27,
"grad_norm": 0.3571007549762726,
"learning_rate": 0.0002,
"loss": 0.6976,
"step": 325
},
{
"epoch": 0.29,
"grad_norm": 0.5330923199653625,
"learning_rate": 0.0002,
"loss": 0.415,
"step": 350
},
{
"epoch": 0.31,
"grad_norm": 0.4177188277244568,
"learning_rate": 0.0002,
"loss": 0.6663,
"step": 375
},
{
"epoch": 0.33,
"grad_norm": 0.5220038890838623,
"learning_rate": 0.0002,
"loss": 0.4305,
"step": 400
},
{
"epoch": 0.35,
"grad_norm": 0.40320834517478943,
"learning_rate": 0.0002,
"loss": 0.6978,
"step": 425
},
{
"epoch": 0.37,
"grad_norm": 0.3986828625202179,
"learning_rate": 0.0002,
"loss": 0.392,
"step": 450
},
{
"epoch": 0.39,
"grad_norm": 0.3662169873714447,
"learning_rate": 0.0002,
"loss": 0.6647,
"step": 475
},
{
"epoch": 0.41,
"grad_norm": 0.4708826541900635,
"learning_rate": 0.0002,
"loss": 0.4119,
"step": 500
},
{
"epoch": 0.43,
"grad_norm": 0.3558133840560913,
"learning_rate": 0.0002,
"loss": 0.6364,
"step": 525
},
{
"epoch": 0.46,
"grad_norm": 0.5110255479812622,
"learning_rate": 0.0002,
"loss": 0.3949,
"step": 550
},
{
"epoch": 0.48,
"grad_norm": 0.4202500581741333,
"learning_rate": 0.0002,
"loss": 0.6034,
"step": 575
},
{
"epoch": 0.5,
"grad_norm": 0.7778056263923645,
"learning_rate": 0.0002,
"loss": 0.389,
"step": 600
},
{
"epoch": 0.52,
"grad_norm": 0.42499253153800964,
"learning_rate": 0.0002,
"loss": 0.6453,
"step": 625
},
{
"epoch": 0.54,
"grad_norm": 0.4130938947200775,
"learning_rate": 0.0002,
"loss": 0.3776,
"step": 650
},
{
"epoch": 0.56,
"grad_norm": 0.38256368041038513,
"learning_rate": 0.0002,
"loss": 0.5631,
"step": 675
},
{
"epoch": 0.58,
"grad_norm": 0.6242409944534302,
"learning_rate": 0.0002,
"loss": 0.3771,
"step": 700
},
{
"epoch": 0.6,
"grad_norm": 0.3617693781852722,
"learning_rate": 0.0002,
"loss": 0.5804,
"step": 725
},
{
"epoch": 0.62,
"grad_norm": 0.3799505829811096,
"learning_rate": 0.0002,
"loss": 0.3513,
"step": 750
},
{
"epoch": 0.64,
"grad_norm": 0.32906627655029297,
"learning_rate": 0.0002,
"loss": 0.5393,
"step": 775
},
{
"epoch": 0.66,
"grad_norm": 0.4239194989204407,
"learning_rate": 0.0002,
"loss": 0.3636,
"step": 800
},
{
"epoch": 0.68,
"grad_norm": 0.4123442769050598,
"learning_rate": 0.0002,
"loss": 0.5279,
"step": 825
},
{
"epoch": 0.7,
"grad_norm": 0.4623079001903534,
"learning_rate": 0.0002,
"loss": 0.3821,
"step": 850
},
{
"epoch": 0.72,
"grad_norm": 0.3831407427787781,
"learning_rate": 0.0002,
"loss": 0.5286,
"step": 875
},
{
"epoch": 0.75,
"grad_norm": 0.48481571674346924,
"learning_rate": 0.0002,
"loss": 0.3588,
"step": 900
},
{
"epoch": 0.77,
"grad_norm": 0.36328190565109253,
"learning_rate": 0.0002,
"loss": 0.5519,
"step": 925
},
{
"epoch": 0.79,
"grad_norm": 0.4488459527492523,
"learning_rate": 0.0002,
"loss": 0.3833,
"step": 950
},
{
"epoch": 0.81,
"grad_norm": 0.39116260409355164,
"learning_rate": 0.0002,
"loss": 0.5035,
"step": 975
},
{
"epoch": 0.83,
"grad_norm": 0.4923652410507202,
"learning_rate": 0.0002,
"loss": 0.3178,
"step": 1000
},
{
"epoch": 0.85,
"grad_norm": 0.38706672191619873,
"learning_rate": 0.0002,
"loss": 0.5357,
"step": 1025
},
{
"epoch": 0.87,
"grad_norm": 0.3919411897659302,
"learning_rate": 0.0002,
"loss": 0.3329,
"step": 1050
},
{
"epoch": 0.89,
"grad_norm": 0.44446566700935364,
"learning_rate": 0.0002,
"loss": 0.4754,
"step": 1075
},
{
"epoch": 0.91,
"grad_norm": 0.3222676217556,
"learning_rate": 0.0002,
"loss": 0.3128,
"step": 1100
},
{
"epoch": 0.93,
"grad_norm": 0.3832922577857971,
"learning_rate": 0.0002,
"loss": 0.5655,
"step": 1125
},
{
"epoch": 0.95,
"grad_norm": 0.5052505135536194,
"learning_rate": 0.0002,
"loss": 0.3673,
"step": 1150
},
{
"epoch": 0.97,
"grad_norm": 0.3186209797859192,
"learning_rate": 0.0002,
"loss": 0.4993,
"step": 1175
},
{
"epoch": 0.99,
"grad_norm": 0.5416242480278015,
"learning_rate": 0.0002,
"loss": 0.3399,
"step": 1200
}
],
"logging_steps": 25,
"max_steps": 1207,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 25,
"total_flos": 6.735879205561958e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}