sindhi-bert-base / checkpoint-3924 /trainer_state.json
hellosindh's picture
Upload folder using huggingface_hub
8424953 verified
{
"best_global_step": 3924,
"best_metric": 3.390385389328003,
"best_model_checkpoint": "sindhibert_session5/checkpoint-3924",
"epoch": 2.0,
"eval_steps": 1962,
"global_step": 3924,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.05098139179199592,
"grad_norm": 5.245308876037598,
"learning_rate": 2.1063829787234044e-06,
"loss": 14.123677978515625,
"step": 100
},
{
"epoch": 0.10196278358399184,
"grad_norm": 5.646098613739014,
"learning_rate": 4.23404255319149e-06,
"loss": 14.066470947265625,
"step": 200
},
{
"epoch": 0.15294417537598776,
"grad_norm": 5.477426528930664,
"learning_rate": 4.99628768445363e-06,
"loss": 14.01585693359375,
"step": 300
},
{
"epoch": 0.20392556716798368,
"grad_norm": 5.2961344718933105,
"learning_rate": 4.975657001043815e-06,
"loss": 13.965966796875,
"step": 400
},
{
"epoch": 0.2549069589599796,
"grad_norm": 5.533503532409668,
"learning_rate": 4.937082682936824e-06,
"loss": 13.97388671875,
"step": 500
},
{
"epoch": 0.3058883507519755,
"grad_norm": 5.314619064331055,
"learning_rate": 4.8808443179239025e-06,
"loss": 13.95988525390625,
"step": 600
},
{
"epoch": 0.35686974254397147,
"grad_norm": 5.3324384689331055,
"learning_rate": 4.807349523327375e-06,
"loss": 13.953388671875,
"step": 700
},
{
"epoch": 0.40785113433596737,
"grad_norm": 5.467737674713135,
"learning_rate": 4.717130991578312e-06,
"loss": 13.9411962890625,
"step": 800
},
{
"epoch": 0.45883252612796327,
"grad_norm": 5.45510721206665,
"learning_rate": 4.610842629246138e-06,
"loss": 13.934332275390625,
"step": 900
},
{
"epoch": 0.5098139179199592,
"grad_norm": 5.6341962814331055,
"learning_rate": 4.489254817504595e-06,
"loss": 13.926766357421876,
"step": 1000
},
{
"epoch": 0.5607953097119551,
"grad_norm": 5.584266662597656,
"learning_rate": 4.353248828386359e-06,
"loss": 13.91165283203125,
"step": 1100
},
{
"epoch": 0.611776701503951,
"grad_norm": 5.35775899887085,
"learning_rate": 4.203810437297447e-06,
"loss": 13.915504150390625,
"step": 1200
},
{
"epoch": 0.6627580932959469,
"grad_norm": 5.549437999725342,
"learning_rate": 4.042022778088111e-06,
"loss": 13.945462646484375,
"step": 1300
},
{
"epoch": 0.7137394850879429,
"grad_norm": 5.455720901489258,
"learning_rate": 3.869058492466847e-06,
"loss": 13.89906494140625,
"step": 1400
},
{
"epoch": 0.7647208768799388,
"grad_norm": 5.856232166290283,
"learning_rate": 3.6861712306588434e-06,
"loss": 13.905211181640626,
"step": 1500
},
{
"epoch": 0.8157022686719347,
"grad_norm": 5.481319904327393,
"learning_rate": 3.4946865649123025e-06,
"loss": 13.896219482421875,
"step": 1600
},
{
"epoch": 0.8666836604639306,
"grad_norm": 5.278250217437744,
"learning_rate": 3.2959923817118296e-06,
"loss": 13.906019287109375,
"step": 1700
},
{
"epoch": 0.9176650522559265,
"grad_norm": 5.768893718719482,
"learning_rate": 3.091528822336405e-06,
"loss": 13.93161376953125,
"step": 1800
},
{
"epoch": 0.9686464440479226,
"grad_norm": 5.366046905517578,
"learning_rate": 2.8827778446730764e-06,
"loss": 13.906478271484374,
"step": 1900
},
{
"epoch": 1.0,
"eval_loss": 3.396049737930298,
"eval_runtime": 7.952,
"eval_samples_per_second": 637.827,
"eval_steps_per_second": 10.06,
"step": 1962
},
{
"epoch": 1.0193729288809585,
"grad_norm": 5.584351062774658,
"learning_rate": 2.6712524819426355e-06,
"loss": 13.823892822265625,
"step": 2000
},
{
"epoch": 1.0703543206729544,
"grad_norm": 5.338160037994385,
"learning_rate": 2.458485876190375e-06,
"loss": 13.90267333984375,
"step": 2100
},
{
"epoch": 1.1213357124649503,
"grad_norm": 5.378915309906006,
"learning_rate": 2.246020166027488e-06,
"loss": 13.9045458984375,
"step": 2200
},
{
"epoch": 1.1723171042569462,
"grad_norm": 5.420534133911133,
"learning_rate": 2.0353953091650912e-06,
"loss": 13.91886474609375,
"step": 2300
},
{
"epoch": 1.2232984960489421,
"grad_norm": 5.299575328826904,
"learning_rate": 1.82813792075548e-06,
"loss": 13.879603271484376,
"step": 2400
},
{
"epoch": 1.274279887840938,
"grad_norm": 5.256579875946045,
"learning_rate": 1.6257502084406458e-06,
"loss": 13.92048828125,
"step": 2500
},
{
"epoch": 1.325261279632934,
"grad_norm": 5.846710205078125,
"learning_rate": 1.4296990843071351e-06,
"loss": 13.912366943359375,
"step": 2600
},
{
"epoch": 1.3762426714249298,
"grad_norm": 5.547558307647705,
"learning_rate": 1.2414055326641378e-06,
"loss": 13.929051513671874,
"step": 2700
},
{
"epoch": 1.4272240632169257,
"grad_norm": 5.693734169006348,
"learning_rate": 1.062234310707458e-06,
"loss": 13.929571533203125,
"step": 2800
},
{
"epoch": 1.4782054550089216,
"grad_norm": 5.414492607116699,
"learning_rate": 8.934840567192496e-07,
"loss": 13.91981201171875,
"step": 2900
},
{
"epoch": 1.5291868468009175,
"grad_norm": 5.501009941101074,
"learning_rate": 7.363778774995834e-07,
"loss": 13.91615478515625,
"step": 3000
},
{
"epoch": 1.5801682385929134,
"grad_norm": 5.355737686157227,
"learning_rate": 5.920544832524166e-07,
"loss": 13.9007421875,
"step": 3100
},
{
"epoch": 1.6311496303849093,
"grad_norm": 5.376833915710449,
"learning_rate": 4.61559934180571e-07,
"loss": 13.915396728515624,
"step": 3200
},
{
"epoch": 1.6821310221769055,
"grad_norm": 5.302165985107422,
"learning_rate": 3.4584005861063974e-07,
"loss": 13.94998046875,
"step": 3300
},
{
"epoch": 1.7331124139689014,
"grad_norm": 5.571377277374268,
"learning_rate": 2.4573359760148354e-07,
"loss": 13.98015380859375,
"step": 3400
},
{
"epoch": 1.7840938057608973,
"grad_norm": 5.618070125579834,
"learning_rate": 1.619661257243843e-07,
"loss": 13.96041015625,
"step": 3500
},
{
"epoch": 1.8350751975528932,
"grad_norm": 5.216390132904053,
"learning_rate": 9.514479207723276e-08,
"loss": 13.9774560546875,
"step": 3600
},
{
"epoch": 1.886056589344889,
"grad_norm": 5.458727836608887,
"learning_rate": 4.575391965002174e-08,
"loss": 14.00027587890625,
"step": 3700
},
{
"epoch": 1.9370379811368852,
"grad_norm": 5.513509273529053,
"learning_rate": 1.4151494937582056e-08,
"loss": 14.0021728515625,
"step": 3800
},
{
"epoch": 1.988019372928881,
"grad_norm": 5.200069904327393,
"learning_rate": 5.665732429496751e-10,
"loss": 13.960343017578126,
"step": 3900
},
{
"epoch": 2.0,
"eval_loss": 3.390385389328003,
"eval_runtime": 7.9843,
"eval_samples_per_second": 635.249,
"eval_steps_per_second": 10.02,
"step": 3924
}
],
"logging_steps": 100,
"max_steps": 3924,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 1962,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 3,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.643322074019246e+17,
"train_batch_size": 64,
"trial_name": null,
"trial_params": null
}