be-0.6B / trainer_state.json
arynkiewicz's picture
Upload model
a51ad32 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 411,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.072992700729927,
"grad_norm": 7.62998519999138,
"learning_rate": 3.4615384615384617e-06,
"loss": 2.1931,
"num_tokens": 2861698.0,
"step": 10
},
{
"epoch": 0.145985401459854,
"grad_norm": 2.0378838645741393,
"learning_rate": 4.997196729961109e-06,
"loss": 2.0271,
"num_tokens": 5725093.0,
"step": 20
},
{
"epoch": 0.21897810218978103,
"grad_norm": 1.1687767688324222,
"learning_rate": 4.980088396483145e-06,
"loss": 1.9229,
"num_tokens": 8584884.0,
"step": 30
},
{
"epoch": 0.291970802919708,
"grad_norm": 1.5019234268738317,
"learning_rate": 4.947535513144287e-06,
"loss": 1.899,
"num_tokens": 11447171.0,
"step": 40
},
{
"epoch": 0.36496350364963503,
"grad_norm": 1.3274194647163335,
"learning_rate": 4.899740800392143e-06,
"loss": 1.8861,
"num_tokens": 14305298.0,
"step": 50
},
{
"epoch": 0.43795620437956206,
"grad_norm": 1.3351216353908748,
"learning_rate": 4.837001895933996e-06,
"loss": 1.8724,
"num_tokens": 17165371.0,
"step": 60
},
{
"epoch": 0.5109489051094891,
"grad_norm": 1.3694945668597924,
"learning_rate": 4.75970950122206e-06,
"loss": 1.864,
"num_tokens": 20024286.0,
"step": 70
},
{
"epoch": 0.583941605839416,
"grad_norm": 1.3674205399654282,
"learning_rate": 4.668344948391787e-06,
"loss": 1.8498,
"num_tokens": 22880217.0,
"step": 80
},
{
"epoch": 0.656934306569343,
"grad_norm": 1.377613524812094,
"learning_rate": 4.563477202804941e-06,
"loss": 1.8502,
"num_tokens": 25733499.0,
"step": 90
},
{
"epoch": 0.7299270072992701,
"grad_norm": 1.4034954365891135,
"learning_rate": 4.445759319863825e-06,
"loss": 1.848,
"num_tokens": 28583434.0,
"step": 100
},
{
"epoch": 0.8029197080291971,
"grad_norm": 1.412142777635451,
"learning_rate": 4.315924378161599e-06,
"loss": 1.8352,
"num_tokens": 31446481.0,
"step": 110
},
{
"epoch": 0.8759124087591241,
"grad_norm": 1.3940616426918506,
"learning_rate": 4.174780914294638e-06,
"loss": 1.8398,
"num_tokens": 34303388.0,
"step": 120
},
{
"epoch": 0.948905109489051,
"grad_norm": 1.3677454956298989,
"learning_rate": 4.023207887766292e-06,
"loss": 1.8312,
"num_tokens": 37155157.0,
"step": 130
},
{
"epoch": 1.0218978102189782,
"grad_norm": 1.370013187149163,
"learning_rate": 3.862149207337677e-06,
"loss": 1.8195,
"num_tokens": 40016949.0,
"step": 140
},
{
"epoch": 1.094890510948905,
"grad_norm": 1.3453807513004084,
"learning_rate": 3.6926078529122245e-06,
"loss": 1.8123,
"num_tokens": 42870291.0,
"step": 150
},
{
"epoch": 1.167883211678832,
"grad_norm": 1.3275053674064692,
"learning_rate": 3.5156396295594732e-06,
"loss": 1.7984,
"num_tokens": 45727372.0,
"step": 160
},
{
"epoch": 1.2408759124087592,
"grad_norm": 1.2934233224716136,
"learning_rate": 3.332346592574404e-06,
"loss": 1.7999,
"num_tokens": 48577368.0,
"step": 170
},
{
"epoch": 1.313868613138686,
"grad_norm": 1.2551658629308613,
"learning_rate": 3.1438701845172427e-06,
"loss": 1.7963,
"num_tokens": 51434586.0,
"step": 180
},
{
"epoch": 1.3868613138686132,
"grad_norm": 1.213032797696571,
"learning_rate": 2.9513841269722554e-06,
"loss": 1.7881,
"num_tokens": 54297424.0,
"step": 190
},
{
"epoch": 1.4598540145985401,
"grad_norm": 1.156968122738453,
"learning_rate": 2.756087111291533e-06,
"loss": 1.7897,
"num_tokens": 57162459.0,
"step": 200
},
{
"epoch": 1.5328467153284673,
"grad_norm": 1.1249392481057738,
"learning_rate": 2.5591953338415597e-06,
"loss": 1.7849,
"num_tokens": 60018675.0,
"step": 210
},
{
"epoch": 1.6058394160583942,
"grad_norm": 1.0994749587493056,
"learning_rate": 2.36193492223873e-06,
"loss": 1.7775,
"num_tokens": 62878668.0,
"step": 220
},
{
"epoch": 1.6788321167883211,
"grad_norm": 1.0637208182057718,
"learning_rate": 2.1655342997388027e-06,
"loss": 1.7909,
"num_tokens": 65740651.0,
"step": 230
},
{
"epoch": 1.7518248175182483,
"grad_norm": 1.044107008859833,
"learning_rate": 1.9712165353304528e-06,
"loss": 1.7756,
"num_tokens": 68601510.0,
"step": 240
},
{
"epoch": 1.8248175182481752,
"grad_norm": 0.9925194390180851,
"learning_rate": 1.7801917271720841e-06,
"loss": 1.7832,
"num_tokens": 71458963.0,
"step": 250
},
{
"epoch": 1.897810218978102,
"grad_norm": 0.955995300227254,
"learning_rate": 1.593649466803439e-06,
"loss": 1.7844,
"num_tokens": 74316993.0,
"step": 260
},
{
"epoch": 1.9708029197080292,
"grad_norm": 0.917851333269758,
"learning_rate": 1.412751431060518e-06,
"loss": 1.7762,
"num_tokens": 77175911.0,
"step": 270
},
{
"epoch": 2.0437956204379564,
"grad_norm": 0.8904733879800006,
"learning_rate": 1.2386241478270566e-06,
"loss": 1.7703,
"num_tokens": 80032620.0,
"step": 280
},
{
"epoch": 2.116788321167883,
"grad_norm": 0.8877855846920832,
"learning_rate": 1.0723519806732633e-06,
"loss": 1.7644,
"num_tokens": 82889866.0,
"step": 290
},
{
"epoch": 2.18978102189781,
"grad_norm": 0.8648908773622401,
"learning_rate": 9.149703760694034e-07,
"loss": 1.7676,
"num_tokens": 85743417.0,
"step": 300
},
{
"epoch": 2.2627737226277373,
"grad_norm": 0.8480142852639106,
"learning_rate": 7.674594152266707e-07,
"loss": 1.7678,
"num_tokens": 88603527.0,
"step": 310
},
{
"epoch": 2.335766423357664,
"grad_norm": 0.8220641591692829,
"learning_rate": 6.307377107207275e-07,
"loss": 1.7687,
"num_tokens": 91459160.0,
"step": 320
},
{
"epoch": 2.408759124087591,
"grad_norm": 0.8343743852690421,
"learning_rate": 5.056566859062018e-07,
"loss": 1.7634,
"num_tokens": 94315375.0,
"step": 330
},
{
"epoch": 2.4817518248175183,
"grad_norm": 0.8234329687161832,
"learning_rate": 3.9299527274662353e-07,
"loss": 1.7634,
"num_tokens": 97180408.0,
"step": 340
},
{
"epoch": 2.554744525547445,
"grad_norm": 0.8056748923441135,
"learning_rate": 2.934550610786327e-07,
"loss": 1.7609,
"num_tokens": 100041451.0,
"step": 350
},
{
"epoch": 2.627737226277372,
"grad_norm": 0.8157145557692418,
"learning_rate": 2.0765592951802804e-07,
"loss": 1.7654,
"num_tokens": 102901398.0,
"step": 360
},
{
"epoch": 2.7007299270072993,
"grad_norm": 0.8132927418871645,
"learning_rate": 1.361321852158326e-07,
"loss": 1.7625,
"num_tokens": 105764937.0,
"step": 370
},
{
"epoch": 2.7737226277372264,
"grad_norm": 0.7925420271726725,
"learning_rate": 7.932923650373209e-08,
"loss": 1.7645,
"num_tokens": 108617497.0,
"step": 380
},
{
"epoch": 2.846715328467153,
"grad_norm": 0.8078926286548843,
"learning_rate": 3.7600819149724024e-08,
"loss": 1.7616,
"num_tokens": 111469621.0,
"step": 390
},
{
"epoch": 2.9197080291970803,
"grad_norm": 0.8084587727297771,
"learning_rate": 1.1206793497235413e-08,
"loss": 1.7543,
"num_tokens": 114331353.0,
"step": 400
},
{
"epoch": 2.9927007299270074,
"grad_norm": 0.8140377250304603,
"learning_rate": 3.11526205856727e-10,
"loss": 1.7653,
"num_tokens": 117187018.0,
"step": 410
},
{
"epoch": 3.0,
"num_tokens": 117474320.0,
"step": 411,
"total_flos": 56147696418816.0,
"train_loss": 1.8167425843630973,
"train_runtime": 1364.5021,
"train_samples_per_second": 153.902,
"train_steps_per_second": 0.301
}
],
"logging_steps": 10,
"max_steps": 411,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 56147696418816.0,
"train_batch_size": 64,
"trial_name": null,
"trial_params": null
}