emdb-summary-generator / trainer_state.json
amudhakumari's picture
Upload folder using huggingface_hub
22ac760 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 1890,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.010582010582010581,
"grad_norm": 39.13801574707031,
"learning_rate": 1.9952380952380953e-05,
"loss": 5.4194,
"step": 10
},
{
"epoch": 0.021164021164021163,
"grad_norm": 24.028564453125,
"learning_rate": 1.98994708994709e-05,
"loss": 1.773,
"step": 20
},
{
"epoch": 0.031746031746031744,
"grad_norm": 16.863679885864258,
"learning_rate": 1.9846560846560846e-05,
"loss": 0.7073,
"step": 30
},
{
"epoch": 0.042328042328042326,
"grad_norm": 10.789055824279785,
"learning_rate": 1.9793650793650796e-05,
"loss": 0.3436,
"step": 40
},
{
"epoch": 0.05291005291005291,
"grad_norm": 8.038490295410156,
"learning_rate": 1.9740740740740743e-05,
"loss": 0.2155,
"step": 50
},
{
"epoch": 0.06349206349206349,
"grad_norm": 4.677557945251465,
"learning_rate": 1.968783068783069e-05,
"loss": 0.1287,
"step": 60
},
{
"epoch": 0.07407407407407407,
"grad_norm": 3.4516615867614746,
"learning_rate": 1.9634920634920636e-05,
"loss": 0.0824,
"step": 70
},
{
"epoch": 0.08465608465608465,
"grad_norm": 1.6580854654312134,
"learning_rate": 1.9582010582010586e-05,
"loss": 0.0501,
"step": 80
},
{
"epoch": 0.09523809523809523,
"grad_norm": 1.1410599946975708,
"learning_rate": 1.9529100529100532e-05,
"loss": 0.0316,
"step": 90
},
{
"epoch": 0.10582010582010581,
"grad_norm": 1.6213871240615845,
"learning_rate": 1.947619047619048e-05,
"loss": 0.025,
"step": 100
},
{
"epoch": 0.1164021164021164,
"grad_norm": 0.5977617502212524,
"learning_rate": 1.9423280423280425e-05,
"loss": 0.0153,
"step": 110
},
{
"epoch": 0.12698412698412698,
"grad_norm": 2.150684118270874,
"learning_rate": 1.9370370370370372e-05,
"loss": 0.0162,
"step": 120
},
{
"epoch": 0.13756613756613756,
"grad_norm": 0.3688974678516388,
"learning_rate": 1.931746031746032e-05,
"loss": 0.0116,
"step": 130
},
{
"epoch": 0.14814814814814814,
"grad_norm": 10.174367904663086,
"learning_rate": 1.9264550264550265e-05,
"loss": 0.0149,
"step": 140
},
{
"epoch": 0.15873015873015872,
"grad_norm": 0.2287941724061966,
"learning_rate": 1.921164021164021e-05,
"loss": 0.0068,
"step": 150
},
{
"epoch": 0.1693121693121693,
"grad_norm": 0.7957000136375427,
"learning_rate": 1.915873015873016e-05,
"loss": 0.01,
"step": 160
},
{
"epoch": 0.17989417989417988,
"grad_norm": 0.24841707944869995,
"learning_rate": 1.9105820105820108e-05,
"loss": 0.0161,
"step": 170
},
{
"epoch": 0.19047619047619047,
"grad_norm": 1.3141834735870361,
"learning_rate": 1.9052910052910055e-05,
"loss": 0.0057,
"step": 180
},
{
"epoch": 0.20105820105820105,
"grad_norm": 0.17035521566867828,
"learning_rate": 1.9e-05,
"loss": 0.012,
"step": 190
},
{
"epoch": 0.21164021164021163,
"grad_norm": 3.853283405303955,
"learning_rate": 1.894708994708995e-05,
"loss": 0.0148,
"step": 200
},
{
"epoch": 0.2222222222222222,
"grad_norm": 0.13835227489471436,
"learning_rate": 1.8894179894179898e-05,
"loss": 0.0139,
"step": 210
},
{
"epoch": 0.2328042328042328,
"grad_norm": 0.3647116422653198,
"learning_rate": 1.8841269841269844e-05,
"loss": 0.0117,
"step": 220
},
{
"epoch": 0.24338624338624337,
"grad_norm": 4.365599632263184,
"learning_rate": 1.878835978835979e-05,
"loss": 0.0102,
"step": 230
},
{
"epoch": 0.25396825396825395,
"grad_norm": 0.1187911406159401,
"learning_rate": 1.8735449735449737e-05,
"loss": 0.0053,
"step": 240
},
{
"epoch": 0.26455026455026454,
"grad_norm": 5.611083507537842,
"learning_rate": 1.8682539682539684e-05,
"loss": 0.0052,
"step": 250
},
{
"epoch": 0.2751322751322751,
"grad_norm": 1.2097302675247192,
"learning_rate": 1.862962962962963e-05,
"loss": 0.0066,
"step": 260
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.08928469568490982,
"learning_rate": 1.8576719576719577e-05,
"loss": 0.0183,
"step": 270
},
{
"epoch": 0.2962962962962963,
"grad_norm": 1.7315622568130493,
"learning_rate": 1.8523809523809527e-05,
"loss": 0.0049,
"step": 280
},
{
"epoch": 0.30687830687830686,
"grad_norm": 0.10508707910776138,
"learning_rate": 1.8470899470899473e-05,
"loss": 0.0174,
"step": 290
},
{
"epoch": 0.31746031746031744,
"grad_norm": 0.08319881558418274,
"learning_rate": 1.841798941798942e-05,
"loss": 0.0068,
"step": 300
},
{
"epoch": 0.328042328042328,
"grad_norm": 3.9523701667785645,
"learning_rate": 1.8365079365079366e-05,
"loss": 0.0545,
"step": 310
},
{
"epoch": 0.3386243386243386,
"grad_norm": 0.11226499825716019,
"learning_rate": 1.8312169312169313e-05,
"loss": 0.0066,
"step": 320
},
{
"epoch": 0.3492063492063492,
"grad_norm": 1.6739437580108643,
"learning_rate": 1.825925925925926e-05,
"loss": 0.0075,
"step": 330
},
{
"epoch": 0.35978835978835977,
"grad_norm": 4.503218173980713,
"learning_rate": 1.8206349206349206e-05,
"loss": 0.0051,
"step": 340
},
{
"epoch": 0.37037037037037035,
"grad_norm": 0.07483673095703125,
"learning_rate": 1.8153439153439152e-05,
"loss": 0.0035,
"step": 350
},
{
"epoch": 0.38095238095238093,
"grad_norm": 0.06276049464941025,
"learning_rate": 1.8100529100529102e-05,
"loss": 0.0368,
"step": 360
},
{
"epoch": 0.3915343915343915,
"grad_norm": 3.8753137588500977,
"learning_rate": 1.804761904761905e-05,
"loss": 0.0103,
"step": 370
},
{
"epoch": 0.4021164021164021,
"grad_norm": 0.8073577880859375,
"learning_rate": 1.7994708994708995e-05,
"loss": 0.0112,
"step": 380
},
{
"epoch": 0.4126984126984127,
"grad_norm": 5.514145374298096,
"learning_rate": 1.7941798941798942e-05,
"loss": 0.0038,
"step": 390
},
{
"epoch": 0.42328042328042326,
"grad_norm": 0.3400741517543793,
"learning_rate": 1.7888888888888892e-05,
"loss": 0.0027,
"step": 400
},
{
"epoch": 0.43386243386243384,
"grad_norm": 6.407121181488037,
"learning_rate": 1.783597883597884e-05,
"loss": 0.0061,
"step": 410
},
{
"epoch": 0.4444444444444444,
"grad_norm": 0.0525721050798893,
"learning_rate": 1.7783068783068785e-05,
"loss": 0.0031,
"step": 420
},
{
"epoch": 0.455026455026455,
"grad_norm": 9.535446166992188,
"learning_rate": 1.773015873015873e-05,
"loss": 0.006,
"step": 430
},
{
"epoch": 0.4656084656084656,
"grad_norm": 0.08640318363904953,
"learning_rate": 1.7677248677248678e-05,
"loss": 0.0062,
"step": 440
},
{
"epoch": 0.47619047619047616,
"grad_norm": 0.06083179637789726,
"learning_rate": 1.7624338624338625e-05,
"loss": 0.0036,
"step": 450
},
{
"epoch": 0.48677248677248675,
"grad_norm": 0.0371096171438694,
"learning_rate": 1.757142857142857e-05,
"loss": 0.0009,
"step": 460
},
{
"epoch": 0.4973544973544973,
"grad_norm": 0.7781650424003601,
"learning_rate": 1.7518518518518518e-05,
"loss": 0.0113,
"step": 470
},
{
"epoch": 0.5079365079365079,
"grad_norm": 0.3946099877357483,
"learning_rate": 1.7465608465608468e-05,
"loss": 0.004,
"step": 480
},
{
"epoch": 0.5185185185185185,
"grad_norm": 0.039193294942379,
"learning_rate": 1.7412698412698414e-05,
"loss": 0.0013,
"step": 490
},
{
"epoch": 0.5291005291005291,
"grad_norm": 0.037502456456422806,
"learning_rate": 1.735978835978836e-05,
"loss": 0.003,
"step": 500
},
{
"epoch": 0.5396825396825397,
"grad_norm": 0.03162345662713051,
"learning_rate": 1.7306878306878307e-05,
"loss": 0.0011,
"step": 510
},
{
"epoch": 0.5502645502645502,
"grad_norm": 0.026782995089888573,
"learning_rate": 1.7253968253968257e-05,
"loss": 0.0006,
"step": 520
},
{
"epoch": 0.5608465608465608,
"grad_norm": 0.027304725721478462,
"learning_rate": 1.7201058201058204e-05,
"loss": 0.0024,
"step": 530
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.045405562967061996,
"learning_rate": 1.714814814814815e-05,
"loss": 0.0079,
"step": 540
},
{
"epoch": 0.582010582010582,
"grad_norm": 0.026607122272253036,
"learning_rate": 1.7095238095238097e-05,
"loss": 0.0024,
"step": 550
},
{
"epoch": 0.5925925925925926,
"grad_norm": 0.1648360937833786,
"learning_rate": 1.7042328042328043e-05,
"loss": 0.0038,
"step": 560
},
{
"epoch": 0.6031746031746031,
"grad_norm": 0.2880645990371704,
"learning_rate": 1.698941798941799e-05,
"loss": 0.0069,
"step": 570
},
{
"epoch": 0.6137566137566137,
"grad_norm": 0.08049669116735458,
"learning_rate": 1.6936507936507936e-05,
"loss": 0.0091,
"step": 580
},
{
"epoch": 0.6243386243386243,
"grad_norm": 0.16391621530056,
"learning_rate": 1.6883597883597883e-05,
"loss": 0.003,
"step": 590
},
{
"epoch": 0.6349206349206349,
"grad_norm": 0.31218671798706055,
"learning_rate": 1.6830687830687833e-05,
"loss": 0.0009,
"step": 600
},
{
"epoch": 0.6455026455026455,
"grad_norm": 0.02184475213289261,
"learning_rate": 1.677777777777778e-05,
"loss": 0.0133,
"step": 610
},
{
"epoch": 0.656084656084656,
"grad_norm": 0.025240115821361542,
"learning_rate": 1.6724867724867726e-05,
"loss": 0.0013,
"step": 620
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.021431762725114822,
"learning_rate": 1.6671957671957672e-05,
"loss": 0.0164,
"step": 630
},
{
"epoch": 0.6772486772486772,
"grad_norm": 0.3141108751296997,
"learning_rate": 1.6619047619047622e-05,
"loss": 0.0026,
"step": 640
},
{
"epoch": 0.6878306878306878,
"grad_norm": 0.021496519446372986,
"learning_rate": 1.656613756613757e-05,
"loss": 0.0023,
"step": 650
},
{
"epoch": 0.6984126984126984,
"grad_norm": 0.06723565608263016,
"learning_rate": 1.6513227513227515e-05,
"loss": 0.0019,
"step": 660
},
{
"epoch": 0.708994708994709,
"grad_norm": 0.02188393473625183,
"learning_rate": 1.6460317460317462e-05,
"loss": 0.0036,
"step": 670
},
{
"epoch": 0.7195767195767195,
"grad_norm": 0.07254693657159805,
"learning_rate": 1.640740740740741e-05,
"loss": 0.0061,
"step": 680
},
{
"epoch": 0.7301587301587301,
"grad_norm": 3.8045756816864014,
"learning_rate": 1.6354497354497355e-05,
"loss": 0.0041,
"step": 690
},
{
"epoch": 0.7407407407407407,
"grad_norm": 0.019957900047302246,
"learning_rate": 1.63015873015873e-05,
"loss": 0.0005,
"step": 700
},
{
"epoch": 0.7513227513227513,
"grad_norm": 0.017905883491039276,
"learning_rate": 1.6248677248677248e-05,
"loss": 0.0099,
"step": 710
},
{
"epoch": 0.7619047619047619,
"grad_norm": 0.580219030380249,
"learning_rate": 1.6195767195767198e-05,
"loss": 0.0033,
"step": 720
},
{
"epoch": 0.7724867724867724,
"grad_norm": 0.0343511626124382,
"learning_rate": 1.6142857142857145e-05,
"loss": 0.0057,
"step": 730
},
{
"epoch": 0.783068783068783,
"grad_norm": 0.6010801196098328,
"learning_rate": 1.608994708994709e-05,
"loss": 0.0009,
"step": 740
},
{
"epoch": 0.7936507936507936,
"grad_norm": 0.020727235823869705,
"learning_rate": 1.6037037037037038e-05,
"loss": 0.0024,
"step": 750
},
{
"epoch": 0.8042328042328042,
"grad_norm": 5.676093101501465,
"learning_rate": 1.5984126984126988e-05,
"loss": 0.0052,
"step": 760
},
{
"epoch": 0.8148148148148148,
"grad_norm": 0.22317780554294586,
"learning_rate": 1.5931216931216934e-05,
"loss": 0.0063,
"step": 770
},
{
"epoch": 0.8253968253968254,
"grad_norm": 0.983028769493103,
"learning_rate": 1.587830687830688e-05,
"loss": 0.0046,
"step": 780
},
{
"epoch": 0.8359788359788359,
"grad_norm": 0.018990248441696167,
"learning_rate": 1.5825396825396827e-05,
"loss": 0.0044,
"step": 790
},
{
"epoch": 0.8465608465608465,
"grad_norm": 6.899256706237793,
"learning_rate": 1.5772486772486774e-05,
"loss": 0.004,
"step": 800
},
{
"epoch": 0.8571428571428571,
"grad_norm": 4.200356483459473,
"learning_rate": 1.571957671957672e-05,
"loss": 0.0046,
"step": 810
},
{
"epoch": 0.8677248677248677,
"grad_norm": 2.085758686065674,
"learning_rate": 1.5666666666666667e-05,
"loss": 0.0029,
"step": 820
},
{
"epoch": 0.8783068783068783,
"grad_norm": 0.07632625102996826,
"learning_rate": 1.5613756613756613e-05,
"loss": 0.0054,
"step": 830
},
{
"epoch": 0.8888888888888888,
"grad_norm": 5.698914051055908,
"learning_rate": 1.5560846560846563e-05,
"loss": 0.0037,
"step": 840
},
{
"epoch": 0.8994708994708994,
"grad_norm": 0.16661055386066437,
"learning_rate": 1.550793650793651e-05,
"loss": 0.0033,
"step": 850
},
{
"epoch": 0.91005291005291,
"grad_norm": 0.061884596943855286,
"learning_rate": 1.5455026455026456e-05,
"loss": 0.0021,
"step": 860
},
{
"epoch": 0.9206349206349206,
"grad_norm": 1.1297595500946045,
"learning_rate": 1.5402116402116403e-05,
"loss": 0.0022,
"step": 870
},
{
"epoch": 0.9312169312169312,
"grad_norm": 0.3115411698818207,
"learning_rate": 1.5349206349206353e-05,
"loss": 0.0052,
"step": 880
},
{
"epoch": 0.9417989417989417,
"grad_norm": 2.6712639331817627,
"learning_rate": 1.52962962962963e-05,
"loss": 0.0029,
"step": 890
},
{
"epoch": 0.9523809523809523,
"grad_norm": 0.01280083879828453,
"learning_rate": 1.5243386243386244e-05,
"loss": 0.0005,
"step": 900
},
{
"epoch": 0.9629629629629629,
"grad_norm": 5.4594645500183105,
"learning_rate": 1.519047619047619e-05,
"loss": 0.007,
"step": 910
},
{
"epoch": 0.9735449735449735,
"grad_norm": 0.01234627328813076,
"learning_rate": 1.5137566137566139e-05,
"loss": 0.007,
"step": 920
},
{
"epoch": 0.9841269841269841,
"grad_norm": 0.12818680703639984,
"learning_rate": 1.5084656084656086e-05,
"loss": 0.0005,
"step": 930
},
{
"epoch": 0.9947089947089947,
"grad_norm": 0.010110250674188137,
"learning_rate": 1.5031746031746032e-05,
"loss": 0.0011,
"step": 940
},
{
"epoch": 1.0,
"eval_loss": 0.03604938089847565,
"eval_runtime": 462.3812,
"eval_samples_per_second": 0.227,
"eval_steps_per_second": 0.115,
"step": 945
},
{
"epoch": 1.0052910052910053,
"grad_norm": 0.012420996092259884,
"learning_rate": 1.4978835978835979e-05,
"loss": 0.0003,
"step": 950
},
{
"epoch": 1.0158730158730158,
"grad_norm": 0.010969682596623898,
"learning_rate": 1.4925925925925929e-05,
"loss": 0.0002,
"step": 960
},
{
"epoch": 1.0264550264550265,
"grad_norm": 0.009813076816499233,
"learning_rate": 1.4873015873015875e-05,
"loss": 0.0064,
"step": 970
},
{
"epoch": 1.037037037037037,
"grad_norm": 0.08147811144590378,
"learning_rate": 1.4820105820105822e-05,
"loss": 0.0009,
"step": 980
},
{
"epoch": 1.0476190476190477,
"grad_norm": 0.010766289196908474,
"learning_rate": 1.4767195767195768e-05,
"loss": 0.0025,
"step": 990
},
{
"epoch": 1.0582010582010581,
"grad_norm": 0.009957353584468365,
"learning_rate": 1.4714285714285716e-05,
"loss": 0.0032,
"step": 1000
},
{
"epoch": 1.0687830687830688,
"grad_norm": 7.969147682189941,
"learning_rate": 1.4661375661375663e-05,
"loss": 0.0028,
"step": 1010
},
{
"epoch": 1.0793650793650793,
"grad_norm": 4.326008319854736,
"learning_rate": 1.460846560846561e-05,
"loss": 0.0048,
"step": 1020
},
{
"epoch": 1.08994708994709,
"grad_norm": 0.00999133475124836,
"learning_rate": 1.4555555555555556e-05,
"loss": 0.0026,
"step": 1030
},
{
"epoch": 1.1005291005291005,
"grad_norm": 0.010022982954978943,
"learning_rate": 1.4502645502645504e-05,
"loss": 0.0084,
"step": 1040
},
{
"epoch": 1.1111111111111112,
"grad_norm": 0.013834713026881218,
"learning_rate": 1.444973544973545e-05,
"loss": 0.0043,
"step": 1050
},
{
"epoch": 1.1216931216931216,
"grad_norm": 0.011380772106349468,
"learning_rate": 1.4396825396825397e-05,
"loss": 0.0352,
"step": 1060
},
{
"epoch": 1.1322751322751323,
"grad_norm": 0.07020097970962524,
"learning_rate": 1.4343915343915344e-05,
"loss": 0.0018,
"step": 1070
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.02143331617116928,
"learning_rate": 1.4291005291005294e-05,
"loss": 0.0017,
"step": 1080
},
{
"epoch": 1.1534391534391535,
"grad_norm": 3.8044309616088867,
"learning_rate": 1.423809523809524e-05,
"loss": 0.0009,
"step": 1090
},
{
"epoch": 1.164021164021164,
"grad_norm": 0.010620811954140663,
"learning_rate": 1.4185185185185187e-05,
"loss": 0.0022,
"step": 1100
},
{
"epoch": 1.1746031746031746,
"grad_norm": 0.013597151264548302,
"learning_rate": 1.4132275132275133e-05,
"loss": 0.0002,
"step": 1110
},
{
"epoch": 1.1851851851851851,
"grad_norm": 0.009412454441189766,
"learning_rate": 1.4079365079365082e-05,
"loss": 0.003,
"step": 1120
},
{
"epoch": 1.1957671957671958,
"grad_norm": 9.870640754699707,
"learning_rate": 1.4026455026455028e-05,
"loss": 0.0051,
"step": 1130
},
{
"epoch": 1.2063492063492063,
"grad_norm": 0.15585137903690338,
"learning_rate": 1.3973544973544975e-05,
"loss": 0.0005,
"step": 1140
},
{
"epoch": 1.216931216931217,
"grad_norm": 0.011549165472388268,
"learning_rate": 1.3920634920634921e-05,
"loss": 0.0017,
"step": 1150
},
{
"epoch": 1.2275132275132274,
"grad_norm": 0.022659868001937866,
"learning_rate": 1.386772486772487e-05,
"loss": 0.0076,
"step": 1160
},
{
"epoch": 1.2380952380952381,
"grad_norm": 1.041445255279541,
"learning_rate": 1.3814814814814816e-05,
"loss": 0.0011,
"step": 1170
},
{
"epoch": 1.2486772486772486,
"grad_norm": 0.00932463351637125,
"learning_rate": 1.3761904761904763e-05,
"loss": 0.0024,
"step": 1180
},
{
"epoch": 1.2592592592592593,
"grad_norm": 0.0319746658205986,
"learning_rate": 1.3708994708994709e-05,
"loss": 0.0012,
"step": 1190
},
{
"epoch": 1.2698412698412698,
"grad_norm": 0.0502803809940815,
"learning_rate": 1.3656084656084659e-05,
"loss": 0.0173,
"step": 1200
},
{
"epoch": 1.2804232804232805,
"grad_norm": 3.397770881652832,
"learning_rate": 1.3603174603174606e-05,
"loss": 0.0135,
"step": 1210
},
{
"epoch": 1.291005291005291,
"grad_norm": 0.014460274949669838,
"learning_rate": 1.3550264550264552e-05,
"loss": 0.0008,
"step": 1220
},
{
"epoch": 1.3015873015873016,
"grad_norm": 0.04834635928273201,
"learning_rate": 1.3497354497354499e-05,
"loss": 0.0053,
"step": 1230
},
{
"epoch": 1.312169312169312,
"grad_norm": 0.09482118487358093,
"learning_rate": 1.3444444444444447e-05,
"loss": 0.0017,
"step": 1240
},
{
"epoch": 1.3227513227513228,
"grad_norm": 0.007359776180237532,
"learning_rate": 1.3391534391534393e-05,
"loss": 0.0003,
"step": 1250
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.013864404521882534,
"learning_rate": 1.333862433862434e-05,
"loss": 0.0003,
"step": 1260
},
{
"epoch": 1.343915343915344,
"grad_norm": 1.8579610586166382,
"learning_rate": 1.3285714285714287e-05,
"loss": 0.0014,
"step": 1270
},
{
"epoch": 1.3544973544973544,
"grad_norm": 0.007739638909697533,
"learning_rate": 1.3232804232804235e-05,
"loss": 0.0005,
"step": 1280
},
{
"epoch": 1.3650793650793651,
"grad_norm": 0.010888740420341492,
"learning_rate": 1.3179894179894181e-05,
"loss": 0.0003,
"step": 1290
},
{
"epoch": 1.3756613756613756,
"grad_norm": 0.007155210245400667,
"learning_rate": 1.3126984126984128e-05,
"loss": 0.0093,
"step": 1300
},
{
"epoch": 1.3862433862433863,
"grad_norm": 0.014670108444988728,
"learning_rate": 1.3074074074074074e-05,
"loss": 0.0056,
"step": 1310
},
{
"epoch": 1.3968253968253967,
"grad_norm": 0.007492662873119116,
"learning_rate": 1.3021164021164023e-05,
"loss": 0.0031,
"step": 1320
},
{
"epoch": 1.4074074074074074,
"grad_norm": 4.942596912384033,
"learning_rate": 1.296825396825397e-05,
"loss": 0.0026,
"step": 1330
},
{
"epoch": 1.417989417989418,
"grad_norm": 0.007958627305924892,
"learning_rate": 1.2915343915343916e-05,
"loss": 0.0089,
"step": 1340
},
{
"epoch": 1.4285714285714286,
"grad_norm": 2.9070186614990234,
"learning_rate": 1.2862433862433862e-05,
"loss": 0.0021,
"step": 1350
},
{
"epoch": 1.439153439153439,
"grad_norm": 0.9513811469078064,
"learning_rate": 1.2809523809523809e-05,
"loss": 0.0011,
"step": 1360
},
{
"epoch": 1.4497354497354498,
"grad_norm": 0.0836050733923912,
"learning_rate": 1.2756613756613759e-05,
"loss": 0.0089,
"step": 1370
},
{
"epoch": 1.4603174603174602,
"grad_norm": 0.03550349175930023,
"learning_rate": 1.2703703703703705e-05,
"loss": 0.003,
"step": 1380
},
{
"epoch": 1.470899470899471,
"grad_norm": 0.8934294581413269,
"learning_rate": 1.2650793650793652e-05,
"loss": 0.0032,
"step": 1390
},
{
"epoch": 1.4814814814814814,
"grad_norm": 0.006408127024769783,
"learning_rate": 1.2597883597883598e-05,
"loss": 0.0088,
"step": 1400
},
{
"epoch": 1.492063492063492,
"grad_norm": 0.15951858460903168,
"learning_rate": 1.2544973544973547e-05,
"loss": 0.0008,
"step": 1410
},
{
"epoch": 1.5026455026455028,
"grad_norm": 0.011819873005151749,
"learning_rate": 1.2492063492063493e-05,
"loss": 0.0285,
"step": 1420
},
{
"epoch": 1.5132275132275133,
"grad_norm": 0.005755511112511158,
"learning_rate": 1.243915343915344e-05,
"loss": 0.0006,
"step": 1430
},
{
"epoch": 1.5238095238095237,
"grad_norm": 0.0067981211468577385,
"learning_rate": 1.2386243386243386e-05,
"loss": 0.0023,
"step": 1440
},
{
"epoch": 1.5343915343915344,
"grad_norm": 0.0054528010077774525,
"learning_rate": 1.2333333333333334e-05,
"loss": 0.0002,
"step": 1450
},
{
"epoch": 1.544973544973545,
"grad_norm": 0.00653385603800416,
"learning_rate": 1.2280423280423281e-05,
"loss": 0.0057,
"step": 1460
},
{
"epoch": 1.5555555555555556,
"grad_norm": 0.0077536082826554775,
"learning_rate": 1.2227513227513227e-05,
"loss": 0.0002,
"step": 1470
},
{
"epoch": 1.566137566137566,
"grad_norm": 0.0057713366113603115,
"learning_rate": 1.2174603174603174e-05,
"loss": 0.0004,
"step": 1480
},
{
"epoch": 1.5767195767195767,
"grad_norm": 0.019029010087251663,
"learning_rate": 1.2121693121693124e-05,
"loss": 0.0053,
"step": 1490
},
{
"epoch": 1.5873015873015874,
"grad_norm": 0.783202052116394,
"learning_rate": 1.206878306878307e-05,
"loss": 0.0014,
"step": 1500
},
{
"epoch": 1.597883597883598,
"grad_norm": 0.18570241332054138,
"learning_rate": 1.2015873015873017e-05,
"loss": 0.0019,
"step": 1510
},
{
"epoch": 1.6084656084656084,
"grad_norm": 3.245318651199341,
"learning_rate": 1.1962962962962964e-05,
"loss": 0.0061,
"step": 1520
},
{
"epoch": 1.619047619047619,
"grad_norm": 0.009805947542190552,
"learning_rate": 1.1910052910052912e-05,
"loss": 0.0008,
"step": 1530
},
{
"epoch": 1.6296296296296298,
"grad_norm": 0.1963416039943695,
"learning_rate": 1.1857142857142858e-05,
"loss": 0.0008,
"step": 1540
},
{
"epoch": 1.6402116402116402,
"grad_norm": 0.03911126032471657,
"learning_rate": 1.1804232804232805e-05,
"loss": 0.0015,
"step": 1550
},
{
"epoch": 1.6507936507936507,
"grad_norm": 0.12514500319957733,
"learning_rate": 1.1751322751322751e-05,
"loss": 0.0002,
"step": 1560
},
{
"epoch": 1.6613756613756614,
"grad_norm": 0.005236633587628603,
"learning_rate": 1.16984126984127e-05,
"loss": 0.0037,
"step": 1570
},
{
"epoch": 1.671957671957672,
"grad_norm": 0.004985412582755089,
"learning_rate": 1.1645502645502646e-05,
"loss": 0.0025,
"step": 1580
},
{
"epoch": 1.6825396825396826,
"grad_norm": 0.008958135731518269,
"learning_rate": 1.1592592592592593e-05,
"loss": 0.0001,
"step": 1590
},
{
"epoch": 1.693121693121693,
"grad_norm": 0.050911322236061096,
"learning_rate": 1.153968253968254e-05,
"loss": 0.0018,
"step": 1600
},
{
"epoch": 1.7037037037037037,
"grad_norm": 0.8171196579933167,
"learning_rate": 1.148677248677249e-05,
"loss": 0.0006,
"step": 1610
},
{
"epoch": 1.7142857142857144,
"grad_norm": 0.00528077082708478,
"learning_rate": 1.1433862433862436e-05,
"loss": 0.0001,
"step": 1620
},
{
"epoch": 1.7248677248677249,
"grad_norm": 1.8344557285308838,
"learning_rate": 1.1380952380952382e-05,
"loss": 0.0012,
"step": 1630
},
{
"epoch": 1.7354497354497354,
"grad_norm": 0.1585739254951477,
"learning_rate": 1.1328042328042327e-05,
"loss": 0.0002,
"step": 1640
},
{
"epoch": 1.746031746031746,
"grad_norm": 12.171996116638184,
"learning_rate": 1.1275132275132277e-05,
"loss": 0.0022,
"step": 1650
},
{
"epoch": 1.7566137566137567,
"grad_norm": 2.625952959060669,
"learning_rate": 1.1222222222222224e-05,
"loss": 0.0063,
"step": 1660
},
{
"epoch": 1.7671957671957672,
"grad_norm": 0.4654531180858612,
"learning_rate": 1.116931216931217e-05,
"loss": 0.0025,
"step": 1670
},
{
"epoch": 1.7777777777777777,
"grad_norm": 3.298426628112793,
"learning_rate": 1.1116402116402117e-05,
"loss": 0.0042,
"step": 1680
},
{
"epoch": 1.7883597883597884,
"grad_norm": 0.06672941148281097,
"learning_rate": 1.1063492063492065e-05,
"loss": 0.0002,
"step": 1690
},
{
"epoch": 1.798941798941799,
"grad_norm": 0.9771575927734375,
"learning_rate": 1.1010582010582011e-05,
"loss": 0.0019,
"step": 1700
},
{
"epoch": 1.8095238095238095,
"grad_norm": 0.1068960428237915,
"learning_rate": 1.0957671957671958e-05,
"loss": 0.0039,
"step": 1710
},
{
"epoch": 1.82010582010582,
"grad_norm": 0.0055111791007220745,
"learning_rate": 1.0904761904761905e-05,
"loss": 0.0014,
"step": 1720
},
{
"epoch": 1.8306878306878307,
"grad_norm": 0.34078916907310486,
"learning_rate": 1.0851851851851853e-05,
"loss": 0.0007,
"step": 1730
},
{
"epoch": 1.8412698412698414,
"grad_norm": 0.02196904458105564,
"learning_rate": 1.07989417989418e-05,
"loss": 0.0007,
"step": 1740
},
{
"epoch": 1.8518518518518519,
"grad_norm": 0.004520466551184654,
"learning_rate": 1.0746031746031746e-05,
"loss": 0.0007,
"step": 1750
},
{
"epoch": 1.8624338624338623,
"grad_norm": 0.13541559875011444,
"learning_rate": 1.0693121693121692e-05,
"loss": 0.0023,
"step": 1760
},
{
"epoch": 1.873015873015873,
"grad_norm": 0.6529462337493896,
"learning_rate": 1.0640211640211642e-05,
"loss": 0.0012,
"step": 1770
},
{
"epoch": 1.8835978835978837,
"grad_norm": 0.04629633575677872,
"learning_rate": 1.0587301587301589e-05,
"loss": 0.0482,
"step": 1780
},
{
"epoch": 1.8941798941798942,
"grad_norm": 0.2736707925796509,
"learning_rate": 1.0534391534391535e-05,
"loss": 0.0045,
"step": 1790
},
{
"epoch": 1.9047619047619047,
"grad_norm": 0.010693887248635292,
"learning_rate": 1.0481481481481482e-05,
"loss": 0.0118,
"step": 1800
},
{
"epoch": 1.9153439153439153,
"grad_norm": 0.31277382373809814,
"learning_rate": 1.042857142857143e-05,
"loss": 0.0004,
"step": 1810
},
{
"epoch": 1.925925925925926,
"grad_norm": 0.00564979063346982,
"learning_rate": 1.0375661375661377e-05,
"loss": 0.0038,
"step": 1820
},
{
"epoch": 1.9365079365079365,
"grad_norm": 0.01478109136223793,
"learning_rate": 1.0322751322751323e-05,
"loss": 0.0002,
"step": 1830
},
{
"epoch": 1.947089947089947,
"grad_norm": 0.1157880499958992,
"learning_rate": 1.026984126984127e-05,
"loss": 0.0007,
"step": 1840
},
{
"epoch": 1.9576719576719577,
"grad_norm": 0.011665510945022106,
"learning_rate": 1.0216931216931218e-05,
"loss": 0.0003,
"step": 1850
},
{
"epoch": 1.9682539682539684,
"grad_norm": 0.005074594635516405,
"learning_rate": 1.0164021164021165e-05,
"loss": 0.0002,
"step": 1860
},
{
"epoch": 1.9788359788359788,
"grad_norm": 0.004943141248077154,
"learning_rate": 1.0111111111111111e-05,
"loss": 0.0003,
"step": 1870
},
{
"epoch": 1.9894179894179893,
"grad_norm": 0.004243962932378054,
"learning_rate": 1.0058201058201058e-05,
"loss": 0.0027,
"step": 1880
},
{
"epoch": 2.0,
"grad_norm": 0.2805013060569763,
"learning_rate": 1.0005291005291008e-05,
"loss": 0.0015,
"step": 1890
},
{
"epoch": 2.0,
"eval_loss": 0.016391731798648834,
"eval_runtime": 459.5859,
"eval_samples_per_second": 0.228,
"eval_steps_per_second": 0.115,
"step": 1890
}
],
"logging_steps": 10,
"max_steps": 3780,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1023956924497920.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}