chimcis's picture
Upload folder using huggingface_hub
dd6805d verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 19.05,
"eval_steps": 500,
"global_step": 2000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.005,
"grad_norm": 5.616273880004883,
"learning_rate": 1.8000000000000001e-06,
"loss": 1.9513,
"step": 10
},
{
"epoch": 0.01,
"grad_norm": 5.260721206665039,
"learning_rate": 3.8000000000000005e-06,
"loss": 1.7789,
"step": 20
},
{
"epoch": 0.015,
"grad_norm": 5.373858451843262,
"learning_rate": 5.8e-06,
"loss": 1.6535,
"step": 30
},
{
"epoch": 0.02,
"grad_norm": 4.073526859283447,
"learning_rate": 7.800000000000002e-06,
"loss": 1.4488,
"step": 40
},
{
"epoch": 0.025,
"grad_norm": 4.743659973144531,
"learning_rate": 9.800000000000001e-06,
"loss": 1.4239,
"step": 50
},
{
"epoch": 0.03,
"grad_norm": 5.532003879547119,
"learning_rate": 1.18e-05,
"loss": 1.4531,
"step": 60
},
{
"epoch": 0.035,
"grad_norm": 4.564126491546631,
"learning_rate": 1.38e-05,
"loss": 1.4279,
"step": 70
},
{
"epoch": 0.04,
"grad_norm": 4.471712112426758,
"learning_rate": 1.58e-05,
"loss": 1.3619,
"step": 80
},
{
"epoch": 0.045,
"grad_norm": 5.032183647155762,
"learning_rate": 1.7800000000000002e-05,
"loss": 1.3614,
"step": 90
},
{
"epoch": 0.05,
"grad_norm": 4.518551349639893,
"learning_rate": 1.98e-05,
"loss": 1.3551,
"step": 100
},
{
"epoch": 1.005,
"grad_norm": 6.518773555755615,
"learning_rate": 1.990526315789474e-05,
"loss": 0.8287,
"step": 110
},
{
"epoch": 1.01,
"grad_norm": 3.6112170219421387,
"learning_rate": 1.98e-05,
"loss": 0.7035,
"step": 120
},
{
"epoch": 1.015,
"grad_norm": 5.7912726402282715,
"learning_rate": 1.9694736842105263e-05,
"loss": 0.8074,
"step": 130
},
{
"epoch": 1.02,
"grad_norm": 3.595136880874634,
"learning_rate": 1.9589473684210527e-05,
"loss": 0.7915,
"step": 140
},
{
"epoch": 1.025,
"grad_norm": 4.417491912841797,
"learning_rate": 1.9484210526315793e-05,
"loss": 0.8106,
"step": 150
},
{
"epoch": 1.03,
"grad_norm": 4.246488094329834,
"learning_rate": 1.9378947368421053e-05,
"loss": 0.8249,
"step": 160
},
{
"epoch": 1.035,
"grad_norm": 5.055039882659912,
"learning_rate": 1.9273684210526317e-05,
"loss": 0.8426,
"step": 170
},
{
"epoch": 1.04,
"grad_norm": 4.677482604980469,
"learning_rate": 1.916842105263158e-05,
"loss": 0.8494,
"step": 180
},
{
"epoch": 1.045,
"grad_norm": 4.659126281738281,
"learning_rate": 1.9063157894736843e-05,
"loss": 0.8388,
"step": 190
},
{
"epoch": 1.05,
"grad_norm": 4.688161849975586,
"learning_rate": 1.8957894736842106e-05,
"loss": 0.8079,
"step": 200
},
{
"epoch": 2.005,
"grad_norm": 5.431162357330322,
"learning_rate": 1.885263157894737e-05,
"loss": 0.3679,
"step": 210
},
{
"epoch": 2.01,
"grad_norm": 3.8187713623046875,
"learning_rate": 1.8747368421052633e-05,
"loss": 0.358,
"step": 220
},
{
"epoch": 2.015,
"grad_norm": 3.1807172298431396,
"learning_rate": 1.8642105263157896e-05,
"loss": 0.3387,
"step": 230
},
{
"epoch": 2.02,
"grad_norm": 4.693820953369141,
"learning_rate": 1.853684210526316e-05,
"loss": 0.3047,
"step": 240
},
{
"epoch": 2.025,
"grad_norm": 4.142691135406494,
"learning_rate": 1.8431578947368423e-05,
"loss": 0.3039,
"step": 250
},
{
"epoch": 2.03,
"grad_norm": 3.9841394424438477,
"learning_rate": 1.8326315789473686e-05,
"loss": 0.3011,
"step": 260
},
{
"epoch": 2.035,
"grad_norm": 4.5235676765441895,
"learning_rate": 1.822105263157895e-05,
"loss": 0.3537,
"step": 270
},
{
"epoch": 2.04,
"grad_norm": 4.285941123962402,
"learning_rate": 1.8115789473684213e-05,
"loss": 0.3105,
"step": 280
},
{
"epoch": 2.045,
"grad_norm": 3.71958327293396,
"learning_rate": 1.8010526315789476e-05,
"loss": 0.3346,
"step": 290
},
{
"epoch": 2.05,
"grad_norm": 3.477006196975708,
"learning_rate": 1.790526315789474e-05,
"loss": 0.3125,
"step": 300
},
{
"epoch": 3.005,
"grad_norm": 2.9285550117492676,
"learning_rate": 1.7800000000000002e-05,
"loss": 0.1397,
"step": 310
},
{
"epoch": 3.01,
"grad_norm": 2.792161703109741,
"learning_rate": 1.7694736842105266e-05,
"loss": 0.1566,
"step": 320
},
{
"epoch": 3.015,
"grad_norm": 2.6446938514709473,
"learning_rate": 1.758947368421053e-05,
"loss": 0.1583,
"step": 330
},
{
"epoch": 3.02,
"grad_norm": 3.135241985321045,
"learning_rate": 1.748421052631579e-05,
"loss": 0.1536,
"step": 340
},
{
"epoch": 3.025,
"grad_norm": 2.987708568572998,
"learning_rate": 1.7378947368421052e-05,
"loss": 0.1542,
"step": 350
},
{
"epoch": 3.03,
"grad_norm": 2.5585477352142334,
"learning_rate": 1.727368421052632e-05,
"loss": 0.1742,
"step": 360
},
{
"epoch": 3.035,
"grad_norm": 2.899765729904175,
"learning_rate": 1.7168421052631582e-05,
"loss": 0.1611,
"step": 370
},
{
"epoch": 3.04,
"grad_norm": 2.752307176589966,
"learning_rate": 1.7063157894736845e-05,
"loss": 0.2076,
"step": 380
},
{
"epoch": 3.045,
"grad_norm": 2.4184489250183105,
"learning_rate": 1.6957894736842105e-05,
"loss": 0.1744,
"step": 390
},
{
"epoch": 3.05,
"grad_norm": 2.342548131942749,
"learning_rate": 1.685263157894737e-05,
"loss": 0.1657,
"step": 400
},
{
"epoch": 4.005,
"grad_norm": 3.1114871501922607,
"learning_rate": 1.6747368421052632e-05,
"loss": 0.0931,
"step": 410
},
{
"epoch": 4.01,
"grad_norm": 2.3490405082702637,
"learning_rate": 1.66421052631579e-05,
"loss": 0.097,
"step": 420
},
{
"epoch": 4.015,
"grad_norm": 2.8796679973602295,
"learning_rate": 1.653684210526316e-05,
"loss": 0.1063,
"step": 430
},
{
"epoch": 4.02,
"grad_norm": 2.324800491333008,
"learning_rate": 1.643157894736842e-05,
"loss": 0.1073,
"step": 440
},
{
"epoch": 4.025,
"grad_norm": 2.4009687900543213,
"learning_rate": 1.6326315789473685e-05,
"loss": 0.0968,
"step": 450
},
{
"epoch": 4.03,
"grad_norm": 2.9474356174468994,
"learning_rate": 1.6221052631578948e-05,
"loss": 0.0925,
"step": 460
},
{
"epoch": 4.035,
"grad_norm": 2.1847965717315674,
"learning_rate": 1.611578947368421e-05,
"loss": 0.114,
"step": 470
},
{
"epoch": 4.04,
"grad_norm": 1.7918593883514404,
"learning_rate": 1.6010526315789475e-05,
"loss": 0.1036,
"step": 480
},
{
"epoch": 4.045,
"grad_norm": 2.242753505706787,
"learning_rate": 1.5905263157894738e-05,
"loss": 0.0939,
"step": 490
},
{
"epoch": 4.05,
"grad_norm": 4.071936130523682,
"learning_rate": 1.58e-05,
"loss": 0.1138,
"step": 500
},
{
"epoch": 5.005,
"grad_norm": 1.5614343881607056,
"learning_rate": 1.5694736842105264e-05,
"loss": 0.0469,
"step": 510
},
{
"epoch": 5.01,
"grad_norm": 2.1511878967285156,
"learning_rate": 1.5589473684210528e-05,
"loss": 0.0544,
"step": 520
},
{
"epoch": 5.015,
"grad_norm": 1.4617661237716675,
"learning_rate": 1.548421052631579e-05,
"loss": 0.0565,
"step": 530
},
{
"epoch": 5.02,
"grad_norm": 2.533205270767212,
"learning_rate": 1.5378947368421054e-05,
"loss": 0.0582,
"step": 540
},
{
"epoch": 5.025,
"grad_norm": 2.398082733154297,
"learning_rate": 1.5273684210526318e-05,
"loss": 0.065,
"step": 550
},
{
"epoch": 5.03,
"grad_norm": 1.9306875467300415,
"learning_rate": 1.516842105263158e-05,
"loss": 0.0631,
"step": 560
},
{
"epoch": 5.035,
"grad_norm": 2.3642289638519287,
"learning_rate": 1.5063157894736844e-05,
"loss": 0.0641,
"step": 570
},
{
"epoch": 5.04,
"grad_norm": 1.4011757373809814,
"learning_rate": 1.4957894736842107e-05,
"loss": 0.0773,
"step": 580
},
{
"epoch": 5.045,
"grad_norm": 2.026855945587158,
"learning_rate": 1.4852631578947369e-05,
"loss": 0.0664,
"step": 590
},
{
"epoch": 5.05,
"grad_norm": 2.1766583919525146,
"learning_rate": 1.4747368421052632e-05,
"loss": 0.0545,
"step": 600
},
{
"epoch": 6.005,
"grad_norm": 1.652336835861206,
"learning_rate": 1.4642105263157896e-05,
"loss": 0.0482,
"step": 610
},
{
"epoch": 6.01,
"grad_norm": 1.5396881103515625,
"learning_rate": 1.4536842105263159e-05,
"loss": 0.0349,
"step": 620
},
{
"epoch": 6.015,
"grad_norm": 2.64931321144104,
"learning_rate": 1.4431578947368424e-05,
"loss": 0.0491,
"step": 630
},
{
"epoch": 6.02,
"grad_norm": 1.5272787809371948,
"learning_rate": 1.4326315789473685e-05,
"loss": 0.0471,
"step": 640
},
{
"epoch": 6.025,
"grad_norm": 1.5077707767486572,
"learning_rate": 1.4221052631578949e-05,
"loss": 0.0413,
"step": 650
},
{
"epoch": 6.03,
"grad_norm": 1.2149569988250732,
"learning_rate": 1.4115789473684212e-05,
"loss": 0.0442,
"step": 660
},
{
"epoch": 6.035,
"grad_norm": 1.6021332740783691,
"learning_rate": 1.4010526315789475e-05,
"loss": 0.0386,
"step": 670
},
{
"epoch": 6.04,
"grad_norm": 1.8967393636703491,
"learning_rate": 1.3905263157894737e-05,
"loss": 0.047,
"step": 680
},
{
"epoch": 6.045,
"grad_norm": 1.5336661338806152,
"learning_rate": 1.38e-05,
"loss": 0.0381,
"step": 690
},
{
"epoch": 6.05,
"grad_norm": 2.3286962509155273,
"learning_rate": 1.3694736842105265e-05,
"loss": 0.0479,
"step": 700
},
{
"epoch": 7.005,
"grad_norm": 1.4702962636947632,
"learning_rate": 1.3589473684210528e-05,
"loss": 0.0231,
"step": 710
},
{
"epoch": 7.01,
"grad_norm": 1.7746834754943848,
"learning_rate": 1.3484210526315792e-05,
"loss": 0.0291,
"step": 720
},
{
"epoch": 7.015,
"grad_norm": 1.37478768825531,
"learning_rate": 1.3378947368421053e-05,
"loss": 0.0368,
"step": 730
},
{
"epoch": 7.02,
"grad_norm": 1.2662794589996338,
"learning_rate": 1.3273684210526316e-05,
"loss": 0.034,
"step": 740
},
{
"epoch": 7.025,
"grad_norm": 0.6514212489128113,
"learning_rate": 1.316842105263158e-05,
"loss": 0.0295,
"step": 750
},
{
"epoch": 7.03,
"grad_norm": 0.9872649908065796,
"learning_rate": 1.3063157894736845e-05,
"loss": 0.0248,
"step": 760
},
{
"epoch": 7.035,
"grad_norm": 0.9525852203369141,
"learning_rate": 1.2957894736842108e-05,
"loss": 0.0254,
"step": 770
},
{
"epoch": 7.04,
"grad_norm": 1.5137120485305786,
"learning_rate": 1.285263157894737e-05,
"loss": 0.0238,
"step": 780
},
{
"epoch": 7.045,
"grad_norm": 3.0998730659484863,
"learning_rate": 1.2747368421052633e-05,
"loss": 0.0337,
"step": 790
},
{
"epoch": 7.05,
"grad_norm": 1.361088752746582,
"learning_rate": 1.2642105263157896e-05,
"loss": 0.0399,
"step": 800
},
{
"epoch": 8.005,
"grad_norm": 0.9619424343109131,
"learning_rate": 1.2536842105263158e-05,
"loss": 0.0215,
"step": 810
},
{
"epoch": 8.01,
"grad_norm": 1.0369452238082886,
"learning_rate": 1.2431578947368421e-05,
"loss": 0.0247,
"step": 820
},
{
"epoch": 8.015,
"grad_norm": 1.344269871711731,
"learning_rate": 1.2326315789473686e-05,
"loss": 0.025,
"step": 830
},
{
"epoch": 8.02,
"grad_norm": 1.2112865447998047,
"learning_rate": 1.2221052631578949e-05,
"loss": 0.0216,
"step": 840
},
{
"epoch": 8.025,
"grad_norm": 1.7333956956863403,
"learning_rate": 1.2115789473684212e-05,
"loss": 0.0183,
"step": 850
},
{
"epoch": 8.03,
"grad_norm": 2.0652589797973633,
"learning_rate": 1.2010526315789474e-05,
"loss": 0.0241,
"step": 860
},
{
"epoch": 8.035,
"grad_norm": 0.624588131904602,
"learning_rate": 1.1905263157894737e-05,
"loss": 0.0202,
"step": 870
},
{
"epoch": 8.04,
"grad_norm": 1.0641587972640991,
"learning_rate": 1.18e-05,
"loss": 0.019,
"step": 880
},
{
"epoch": 8.045,
"grad_norm": 1.1550590991973877,
"learning_rate": 1.1694736842105264e-05,
"loss": 0.0214,
"step": 890
},
{
"epoch": 8.05,
"grad_norm": 1.3287566900253296,
"learning_rate": 1.1589473684210529e-05,
"loss": 0.0292,
"step": 900
},
{
"epoch": 9.005,
"grad_norm": 1.3364150524139404,
"learning_rate": 1.148421052631579e-05,
"loss": 0.0213,
"step": 910
},
{
"epoch": 9.01,
"grad_norm": 0.7193664908409119,
"learning_rate": 1.1378947368421054e-05,
"loss": 0.0082,
"step": 920
},
{
"epoch": 9.015,
"grad_norm": 1.0872008800506592,
"learning_rate": 1.1273684210526317e-05,
"loss": 0.0151,
"step": 930
},
{
"epoch": 9.02,
"grad_norm": 0.22245873510837555,
"learning_rate": 1.116842105263158e-05,
"loss": 0.0128,
"step": 940
},
{
"epoch": 9.025,
"grad_norm": 0.9914424419403076,
"learning_rate": 1.1063157894736842e-05,
"loss": 0.0174,
"step": 950
},
{
"epoch": 9.03,
"grad_norm": 0.9509072303771973,
"learning_rate": 1.0957894736842105e-05,
"loss": 0.0181,
"step": 960
},
{
"epoch": 9.035,
"grad_norm": 0.40795469284057617,
"learning_rate": 1.085263157894737e-05,
"loss": 0.0202,
"step": 970
},
{
"epoch": 9.04,
"grad_norm": 1.3863106966018677,
"learning_rate": 1.0747368421052633e-05,
"loss": 0.0172,
"step": 980
},
{
"epoch": 9.045,
"grad_norm": 0.6388317942619324,
"learning_rate": 1.0642105263157897e-05,
"loss": 0.0185,
"step": 990
},
{
"epoch": 9.05,
"grad_norm": 1.0237617492675781,
"learning_rate": 1.0536842105263158e-05,
"loss": 0.0193,
"step": 1000
},
{
"epoch": 10.005,
"grad_norm": 0.2478121817111969,
"learning_rate": 1.0431578947368421e-05,
"loss": 0.0111,
"step": 1010
},
{
"epoch": 10.01,
"grad_norm": 1.0051342248916626,
"learning_rate": 1.0326315789473685e-05,
"loss": 0.0102,
"step": 1020
},
{
"epoch": 10.015,
"grad_norm": 0.9585472345352173,
"learning_rate": 1.0221052631578948e-05,
"loss": 0.0141,
"step": 1030
},
{
"epoch": 10.02,
"grad_norm": 1.248799204826355,
"learning_rate": 1.0115789473684213e-05,
"loss": 0.0088,
"step": 1040
},
{
"epoch": 10.025,
"grad_norm": 1.050134301185608,
"learning_rate": 1.0010526315789474e-05,
"loss": 0.0097,
"step": 1050
},
{
"epoch": 10.03,
"grad_norm": 0.2019995152950287,
"learning_rate": 9.905263157894738e-06,
"loss": 0.011,
"step": 1060
},
{
"epoch": 10.035,
"grad_norm": 0.9100256562232971,
"learning_rate": 9.800000000000001e-06,
"loss": 0.0149,
"step": 1070
},
{
"epoch": 10.04,
"grad_norm": 0.8645588159561157,
"learning_rate": 9.694736842105263e-06,
"loss": 0.011,
"step": 1080
},
{
"epoch": 10.045,
"grad_norm": 0.9465664029121399,
"learning_rate": 9.589473684210528e-06,
"loss": 0.0116,
"step": 1090
},
{
"epoch": 10.05,
"grad_norm": 0.7357518076896667,
"learning_rate": 9.484210526315791e-06,
"loss": 0.0132,
"step": 1100
},
{
"epoch": 11.005,
"grad_norm": 0.0960564911365509,
"learning_rate": 9.378947368421052e-06,
"loss": 0.0079,
"step": 1110
},
{
"epoch": 11.01,
"grad_norm": 0.07488369941711426,
"learning_rate": 9.273684210526317e-06,
"loss": 0.0079,
"step": 1120
},
{
"epoch": 11.015,
"grad_norm": 0.4596630334854126,
"learning_rate": 9.168421052631579e-06,
"loss": 0.0073,
"step": 1130
},
{
"epoch": 11.02,
"grad_norm": 0.915014386177063,
"learning_rate": 9.063157894736842e-06,
"loss": 0.0058,
"step": 1140
},
{
"epoch": 11.025,
"grad_norm": 0.4357207119464874,
"learning_rate": 8.957894736842107e-06,
"loss": 0.0082,
"step": 1150
},
{
"epoch": 11.03,
"grad_norm": 0.6345843076705933,
"learning_rate": 8.852631578947369e-06,
"loss": 0.009,
"step": 1160
},
{
"epoch": 11.035,
"grad_norm": 0.2351154237985611,
"learning_rate": 8.747368421052632e-06,
"loss": 0.0071,
"step": 1170
},
{
"epoch": 11.04,
"grad_norm": 0.5750226378440857,
"learning_rate": 8.642105263157895e-06,
"loss": 0.0065,
"step": 1180
},
{
"epoch": 11.045,
"grad_norm": 0.49880605936050415,
"learning_rate": 8.536842105263159e-06,
"loss": 0.0061,
"step": 1190
},
{
"epoch": 11.05,
"grad_norm": 0.6473588347434998,
"learning_rate": 8.431578947368422e-06,
"loss": 0.0091,
"step": 1200
},
{
"epoch": 12.005,
"grad_norm": 0.02745138294994831,
"learning_rate": 8.326315789473685e-06,
"loss": 0.0027,
"step": 1210
},
{
"epoch": 12.01,
"grad_norm": 0.6922830939292908,
"learning_rate": 8.221052631578948e-06,
"loss": 0.0058,
"step": 1220
},
{
"epoch": 12.015,
"grad_norm": 0.6273651719093323,
"learning_rate": 8.115789473684212e-06,
"loss": 0.0034,
"step": 1230
},
{
"epoch": 12.02,
"grad_norm": 1.0830379724502563,
"learning_rate": 8.010526315789473e-06,
"loss": 0.0063,
"step": 1240
},
{
"epoch": 12.025,
"grad_norm": 0.824835479259491,
"learning_rate": 7.905263157894737e-06,
"loss": 0.0064,
"step": 1250
},
{
"epoch": 12.03,
"grad_norm": 0.5051027536392212,
"learning_rate": 7.800000000000002e-06,
"loss": 0.0029,
"step": 1260
},
{
"epoch": 12.035,
"grad_norm": 0.6191169619560242,
"learning_rate": 7.694736842105263e-06,
"loss": 0.0036,
"step": 1270
},
{
"epoch": 12.04,
"grad_norm": 0.03212520107626915,
"learning_rate": 7.589473684210526e-06,
"loss": 0.001,
"step": 1280
},
{
"epoch": 12.045,
"grad_norm": 0.027301745489239693,
"learning_rate": 7.4842105263157905e-06,
"loss": 0.0024,
"step": 1290
},
{
"epoch": 12.05,
"grad_norm": 0.6917107701301575,
"learning_rate": 7.378947368421053e-06,
"loss": 0.002,
"step": 1300
},
{
"epoch": 13.005,
"grad_norm": 0.009404812939465046,
"learning_rate": 7.273684210526316e-06,
"loss": 0.0023,
"step": 1310
},
{
"epoch": 13.01,
"grad_norm": 0.5634567737579346,
"learning_rate": 7.16842105263158e-06,
"loss": 0.001,
"step": 1320
},
{
"epoch": 13.015,
"grad_norm": 0.02036663331091404,
"learning_rate": 7.063157894736843e-06,
"loss": 0.0025,
"step": 1330
},
{
"epoch": 13.02,
"grad_norm": 0.05707384645938873,
"learning_rate": 6.957894736842106e-06,
"loss": 0.0012,
"step": 1340
},
{
"epoch": 13.025,
"grad_norm": 0.022448807954788208,
"learning_rate": 6.8526315789473685e-06,
"loss": 0.0008,
"step": 1350
},
{
"epoch": 13.03,
"grad_norm": 0.012089096009731293,
"learning_rate": 6.747368421052633e-06,
"loss": 0.0039,
"step": 1360
},
{
"epoch": 13.035,
"grad_norm": 0.14159785211086273,
"learning_rate": 6.642105263157895e-06,
"loss": 0.0005,
"step": 1370
},
{
"epoch": 13.04,
"grad_norm": 0.05650542676448822,
"learning_rate": 6.536842105263158e-06,
"loss": 0.0011,
"step": 1380
},
{
"epoch": 13.045,
"grad_norm": 0.030941806733608246,
"learning_rate": 6.431578947368422e-06,
"loss": 0.0024,
"step": 1390
},
{
"epoch": 13.05,
"grad_norm": 0.022141387686133385,
"learning_rate": 6.326315789473685e-06,
"loss": 0.0035,
"step": 1400
},
{
"epoch": 14.005,
"grad_norm": 0.010789740830659866,
"learning_rate": 6.221052631578947e-06,
"loss": 0.0002,
"step": 1410
},
{
"epoch": 14.01,
"grad_norm": 0.008089865557849407,
"learning_rate": 6.1157894736842106e-06,
"loss": 0.0002,
"step": 1420
},
{
"epoch": 14.015,
"grad_norm": 0.009247499518096447,
"learning_rate": 6.010526315789475e-06,
"loss": 0.0004,
"step": 1430
},
{
"epoch": 14.02,
"grad_norm": 0.005500028375536203,
"learning_rate": 5.905263157894737e-06,
"loss": 0.0019,
"step": 1440
},
{
"epoch": 14.025,
"grad_norm": 0.04416996240615845,
"learning_rate": 5.8e-06,
"loss": 0.0013,
"step": 1450
},
{
"epoch": 14.03,
"grad_norm": 0.10680264979600906,
"learning_rate": 5.694736842105264e-06,
"loss": 0.0012,
"step": 1460
},
{
"epoch": 14.035,
"grad_norm": 0.008884921669960022,
"learning_rate": 5.589473684210527e-06,
"loss": 0.0002,
"step": 1470
},
{
"epoch": 14.04,
"grad_norm": 0.179422989487648,
"learning_rate": 5.484210526315789e-06,
"loss": 0.0009,
"step": 1480
},
{
"epoch": 14.045,
"grad_norm": 0.7773061990737915,
"learning_rate": 5.3789473684210535e-06,
"loss": 0.0003,
"step": 1490
},
{
"epoch": 14.05,
"grad_norm": 0.04718916490674019,
"learning_rate": 5.273684210526317e-06,
"loss": 0.0009,
"step": 1500
},
{
"epoch": 15.005,
"grad_norm": 0.0084745017811656,
"learning_rate": 5.168421052631579e-06,
"loss": 0.0014,
"step": 1510
},
{
"epoch": 15.01,
"grad_norm": 0.006006155628710985,
"learning_rate": 5.0631578947368424e-06,
"loss": 0.0008,
"step": 1520
},
{
"epoch": 15.015,
"grad_norm": 0.007198017556220293,
"learning_rate": 4.957894736842106e-06,
"loss": 0.0002,
"step": 1530
},
{
"epoch": 15.02,
"grad_norm": 0.004182496573776007,
"learning_rate": 4.852631578947369e-06,
"loss": 0.0003,
"step": 1540
},
{
"epoch": 15.025,
"grad_norm": 0.00469869002699852,
"learning_rate": 4.747368421052632e-06,
"loss": 0.0001,
"step": 1550
},
{
"epoch": 15.03,
"grad_norm": 0.005362135358154774,
"learning_rate": 4.642105263157895e-06,
"loss": 0.0002,
"step": 1560
},
{
"epoch": 15.035,
"grad_norm": 0.005833903793245554,
"learning_rate": 4.536842105263158e-06,
"loss": 0.0003,
"step": 1570
},
{
"epoch": 15.04,
"grad_norm": 0.006048844661563635,
"learning_rate": 4.431578947368421e-06,
"loss": 0.0002,
"step": 1580
},
{
"epoch": 15.045,
"grad_norm": 0.005061362404376268,
"learning_rate": 4.3263157894736845e-06,
"loss": 0.0001,
"step": 1590
},
{
"epoch": 15.05,
"grad_norm": 0.007132957689464092,
"learning_rate": 4.221052631578948e-06,
"loss": 0.0004,
"step": 1600
},
{
"epoch": 16.005,
"grad_norm": 0.004316328559070826,
"learning_rate": 4.115789473684211e-06,
"loss": 0.0001,
"step": 1610
},
{
"epoch": 16.01,
"grad_norm": 0.0065157352946698666,
"learning_rate": 4.010526315789474e-06,
"loss": 0.0001,
"step": 1620
},
{
"epoch": 16.015,
"grad_norm": 0.004985947627574205,
"learning_rate": 3.905263157894737e-06,
"loss": 0.0001,
"step": 1630
},
{
"epoch": 16.02,
"grad_norm": 0.0039834328927099705,
"learning_rate": 3.8000000000000005e-06,
"loss": 0.0003,
"step": 1640
},
{
"epoch": 16.025,
"grad_norm": 0.15245139598846436,
"learning_rate": 3.6947368421052637e-06,
"loss": 0.0001,
"step": 1650
},
{
"epoch": 16.03,
"grad_norm": 0.13511842489242554,
"learning_rate": 3.5894736842105266e-06,
"loss": 0.0003,
"step": 1660
},
{
"epoch": 16.035,
"grad_norm": 0.0028432030230760574,
"learning_rate": 3.48421052631579e-06,
"loss": 0.0001,
"step": 1670
},
{
"epoch": 16.04,
"grad_norm": 0.003842442063614726,
"learning_rate": 3.3789473684210527e-06,
"loss": 0.0003,
"step": 1680
},
{
"epoch": 16.045,
"grad_norm": 0.005007196217775345,
"learning_rate": 3.273684210526316e-06,
"loss": 0.0002,
"step": 1690
},
{
"epoch": 16.05,
"grad_norm": 0.003745683468878269,
"learning_rate": 3.168421052631579e-06,
"loss": 0.0003,
"step": 1700
},
{
"epoch": 17.005,
"grad_norm": 0.0036488028708845377,
"learning_rate": 3.0631578947368425e-06,
"loss": 0.0002,
"step": 1710
},
{
"epoch": 17.01,
"grad_norm": 0.0031844789627939463,
"learning_rate": 2.957894736842106e-06,
"loss": 0.0001,
"step": 1720
},
{
"epoch": 17.015,
"grad_norm": 0.008146918378770351,
"learning_rate": 2.8526315789473687e-06,
"loss": 0.0002,
"step": 1730
},
{
"epoch": 17.02,
"grad_norm": 0.0034727228339761496,
"learning_rate": 2.747368421052632e-06,
"loss": 0.0001,
"step": 1740
},
{
"epoch": 17.025,
"grad_norm": 0.00454725231975317,
"learning_rate": 2.6421052631578948e-06,
"loss": 0.0001,
"step": 1750
},
{
"epoch": 17.03,
"grad_norm": 0.005313561297953129,
"learning_rate": 2.536842105263158e-06,
"loss": 0.0002,
"step": 1760
},
{
"epoch": 17.035,
"grad_norm": 0.1343342512845993,
"learning_rate": 2.4315789473684213e-06,
"loss": 0.0002,
"step": 1770
},
{
"epoch": 17.04,
"grad_norm": 0.004856111016124487,
"learning_rate": 2.326315789473684e-06,
"loss": 0.0001,
"step": 1780
},
{
"epoch": 17.045,
"grad_norm": 0.10679910331964493,
"learning_rate": 2.221052631578948e-06,
"loss": 0.0002,
"step": 1790
},
{
"epoch": 17.05,
"grad_norm": 0.003918065223842859,
"learning_rate": 2.1157894736842107e-06,
"loss": 0.0001,
"step": 1800
},
{
"epoch": 18.005,
"grad_norm": 0.0038473308086395264,
"learning_rate": 2.010526315789474e-06,
"loss": 0.0001,
"step": 1810
},
{
"epoch": 18.01,
"grad_norm": 0.0027081076987087727,
"learning_rate": 1.905263157894737e-06,
"loss": 0.0002,
"step": 1820
},
{
"epoch": 18.015,
"grad_norm": 0.003243462648242712,
"learning_rate": 1.8000000000000001e-06,
"loss": 0.0002,
"step": 1830
},
{
"epoch": 18.02,
"grad_norm": 0.004891443997621536,
"learning_rate": 1.6947368421052632e-06,
"loss": 0.0001,
"step": 1840
},
{
"epoch": 18.025,
"grad_norm": 0.00323393102735281,
"learning_rate": 1.5894736842105265e-06,
"loss": 0.0001,
"step": 1850
},
{
"epoch": 18.03,
"grad_norm": 0.00286379037424922,
"learning_rate": 1.4842105263157897e-06,
"loss": 0.0001,
"step": 1860
},
{
"epoch": 18.035,
"grad_norm": 0.0032987131271511316,
"learning_rate": 1.3789473684210528e-06,
"loss": 0.0001,
"step": 1870
},
{
"epoch": 18.04,
"grad_norm": 0.003127966308966279,
"learning_rate": 1.2736842105263159e-06,
"loss": 0.0001,
"step": 1880
},
{
"epoch": 18.045,
"grad_norm": 0.0026523026172071695,
"learning_rate": 1.1684210526315791e-06,
"loss": 0.0002,
"step": 1890
},
{
"epoch": 18.05,
"grad_norm": 0.0035338886082172394,
"learning_rate": 1.0631578947368422e-06,
"loss": 0.0002,
"step": 1900
},
{
"epoch": 19.005,
"grad_norm": 0.0032154133077710867,
"learning_rate": 9.578947368421053e-07,
"loss": 0.0001,
"step": 1910
},
{
"epoch": 19.01,
"grad_norm": 0.0040828268975019455,
"learning_rate": 8.526315789473685e-07,
"loss": 0.0001,
"step": 1920
},
{
"epoch": 19.015,
"grad_norm": 0.002462048316374421,
"learning_rate": 7.473684210526316e-07,
"loss": 0.0001,
"step": 1930
},
{
"epoch": 19.02,
"grad_norm": 0.00201555248349905,
"learning_rate": 6.421052631578948e-07,
"loss": 0.0003,
"step": 1940
},
{
"epoch": 19.025,
"grad_norm": 0.0026351658161729574,
"learning_rate": 5.368421052631579e-07,
"loss": 0.0001,
"step": 1950
},
{
"epoch": 19.03,
"grad_norm": 0.002890300936996937,
"learning_rate": 4.3157894736842105e-07,
"loss": 0.0001,
"step": 1960
},
{
"epoch": 19.035,
"grad_norm": 0.002067763591185212,
"learning_rate": 3.263157894736842e-07,
"loss": 0.0001,
"step": 1970
},
{
"epoch": 19.04,
"grad_norm": 0.1408688873052597,
"learning_rate": 2.2105263157894736e-07,
"loss": 0.0003,
"step": 1980
},
{
"epoch": 19.045,
"grad_norm": 0.002841346897184849,
"learning_rate": 1.1578947368421054e-07,
"loss": 0.0001,
"step": 1990
},
{
"epoch": 19.05,
"grad_norm": 0.0034785745665431023,
"learning_rate": 1.0526315789473684e-08,
"loss": 0.0001,
"step": 2000
}
],
"logging_steps": 10,
"max_steps": 2000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 9.010451846896026e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}