ada_model / checkpoint-228 /trainer_state.json
nlouis's picture
Upload folder using huggingface_hub
1684e24 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 228,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"entropy": 0.3891168855130672,
"epoch": 0.043859649122807015,
"grad_norm": 18.536834716796875,
"learning_rate": 2e-05,
"loss": 4.281,
"mean_token_accuracy": 0.5703265547752381,
"num_tokens": 29513.0,
"step": 10
},
{
"entropy": 0.749189381301403,
"epoch": 0.08771929824561403,
"grad_norm": 3.8398051261901855,
"learning_rate": 2e-05,
"loss": 1.9625,
"mean_token_accuracy": 0.6897866070270539,
"num_tokens": 58932.0,
"step": 20
},
{
"entropy": 1.169826951622963,
"epoch": 0.13157894736842105,
"grad_norm": 2.5061917304992676,
"learning_rate": 2e-05,
"loss": 1.4255,
"mean_token_accuracy": 0.7357227891683579,
"num_tokens": 88352.0,
"step": 30
},
{
"entropy": 1.1334790736436844,
"epoch": 0.17543859649122806,
"grad_norm": 1.9473565816879272,
"learning_rate": 2e-05,
"loss": 1.1315,
"mean_token_accuracy": 0.7862708762288093,
"num_tokens": 117877.0,
"step": 40
},
{
"entropy": 0.8736076682806015,
"epoch": 0.21929824561403508,
"grad_norm": 1.7826578617095947,
"learning_rate": 2e-05,
"loss": 0.8549,
"mean_token_accuracy": 0.8343304082751274,
"num_tokens": 147392.0,
"step": 50
},
{
"entropy": 0.5955807730555535,
"epoch": 0.2631578947368421,
"grad_norm": 1.6632131338119507,
"learning_rate": 2e-05,
"loss": 0.58,
"mean_token_accuracy": 0.8876266479492188,
"num_tokens": 176728.0,
"step": 60
},
{
"entropy": 0.34382262006402015,
"epoch": 0.30701754385964913,
"grad_norm": 1.4843252897262573,
"learning_rate": 2e-05,
"loss": 0.3445,
"mean_token_accuracy": 0.9318736225366593,
"num_tokens": 206184.0,
"step": 70
},
{
"entropy": 0.19373956136405468,
"epoch": 0.3508771929824561,
"grad_norm": 0.904003918170929,
"learning_rate": 2e-05,
"loss": 0.189,
"mean_token_accuracy": 0.9710342198610306,
"num_tokens": 235688.0,
"step": 80
},
{
"entropy": 0.12052652426064014,
"epoch": 0.39473684210526316,
"grad_norm": 0.7406700849533081,
"learning_rate": 2e-05,
"loss": 0.1056,
"mean_token_accuracy": 0.9886028110980988,
"num_tokens": 265172.0,
"step": 90
},
{
"entropy": 0.0992697212845087,
"epoch": 0.43859649122807015,
"grad_norm": 0.606019139289856,
"learning_rate": 2e-05,
"loss": 0.0817,
"mean_token_accuracy": 0.989163076877594,
"num_tokens": 294687.0,
"step": 100
},
{
"entropy": 0.08367474023252726,
"epoch": 0.4824561403508772,
"grad_norm": 0.4705955386161804,
"learning_rate": 2e-05,
"loss": 0.0737,
"mean_token_accuracy": 0.9905414953827858,
"num_tokens": 324088.0,
"step": 110
},
{
"entropy": 0.07386345528066159,
"epoch": 0.5263157894736842,
"grad_norm": 0.4300910234451294,
"learning_rate": 2e-05,
"loss": 0.0631,
"mean_token_accuracy": 0.9910139158368111,
"num_tokens": 353582.0,
"step": 120
},
{
"entropy": 0.06888655042275786,
"epoch": 0.5701754385964912,
"grad_norm": 0.45082923769950867,
"learning_rate": 2e-05,
"loss": 0.0661,
"mean_token_accuracy": 0.9910464301705361,
"num_tokens": 383080.0,
"step": 130
},
{
"entropy": 0.06553246006369591,
"epoch": 0.6140350877192983,
"grad_norm": 0.4258216619491577,
"learning_rate": 2e-05,
"loss": 0.0615,
"mean_token_accuracy": 0.9914803236722947,
"num_tokens": 412559.0,
"step": 140
},
{
"entropy": 0.06352790119126439,
"epoch": 0.6578947368421053,
"grad_norm": 0.37360644340515137,
"learning_rate": 2e-05,
"loss": 0.0579,
"mean_token_accuracy": 0.9916400715708733,
"num_tokens": 442008.0,
"step": 150
},
{
"entropy": 0.060972847137600185,
"epoch": 0.7017543859649122,
"grad_norm": 0.3406882584095001,
"learning_rate": 2e-05,
"loss": 0.0576,
"mean_token_accuracy": 0.9914845108985901,
"num_tokens": 471495.0,
"step": 160
},
{
"entropy": 0.05891128294169903,
"epoch": 0.7456140350877193,
"grad_norm": 0.2854160964488983,
"learning_rate": 2e-05,
"loss": 0.057,
"mean_token_accuracy": 0.9915445119142532,
"num_tokens": 500954.0,
"step": 170
},
{
"entropy": 0.0579888011328876,
"epoch": 0.7894736842105263,
"grad_norm": 0.40168505907058716,
"learning_rate": 2e-05,
"loss": 0.0507,
"mean_token_accuracy": 0.9923512250185013,
"num_tokens": 530401.0,
"step": 180
},
{
"entropy": 0.0573225624859333,
"epoch": 0.8333333333333334,
"grad_norm": 0.4589792490005493,
"learning_rate": 2e-05,
"loss": 0.0561,
"mean_token_accuracy": 0.9916056364774704,
"num_tokens": 559940.0,
"step": 190
},
{
"entropy": 0.05970530286431312,
"epoch": 0.8771929824561403,
"grad_norm": 0.5275149941444397,
"learning_rate": 2e-05,
"loss": 0.0573,
"mean_token_accuracy": 0.9915869951248169,
"num_tokens": 589427.0,
"step": 200
},
{
"entropy": 0.059507152158766986,
"epoch": 0.9210526315789473,
"grad_norm": 0.2661688029766083,
"learning_rate": 2e-05,
"loss": 0.0515,
"mean_token_accuracy": 0.9923170626163482,
"num_tokens": 618838.0,
"step": 210
},
{
"entropy": 0.05687206219881773,
"epoch": 0.9649122807017544,
"grad_norm": 0.35179731249809265,
"learning_rate": 2e-05,
"loss": 0.0506,
"mean_token_accuracy": 0.9918064430356026,
"num_tokens": 648258.0,
"step": 220
}
],
"logging_steps": 10,
"max_steps": 1824,
"num_input_tokens_seen": 0,
"num_train_epochs": 8,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 419741502833664.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}