appvoid's picture
Upload folder using huggingface_hub
89c1cc5 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 5000,
"global_step": 2250,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0044444444444444444,
"grad_norm": 172.13922119140625,
"learning_rate": 5.000000000000001e-07,
"loss": 11.5329,
"step": 10
},
{
"epoch": 0.008888888888888889,
"grad_norm": 179.96237182617188,
"learning_rate": 1.5e-06,
"loss": 11.3904,
"step": 20
},
{
"epoch": 0.013333333333333334,
"grad_norm": 1364.69921875,
"learning_rate": 2.5e-06,
"loss": 10.7917,
"step": 30
},
{
"epoch": 0.017777777777777778,
"grad_norm": 8219.5224609375,
"learning_rate": 3.3e-06,
"loss": 9.1512,
"step": 40
},
{
"epoch": 0.022222222222222223,
"grad_norm": 12965.9892578125,
"learning_rate": 4.2999999999999995e-06,
"loss": 7.8213,
"step": 50
},
{
"epoch": 0.02666666666666667,
"grad_norm": 31872.74609375,
"learning_rate": 5.3e-06,
"loss": 6.6272,
"step": 60
},
{
"epoch": 0.03111111111111111,
"grad_norm": 788.8919677734375,
"learning_rate": 6.300000000000001e-06,
"loss": 6.3266,
"step": 70
},
{
"epoch": 0.035555555555555556,
"grad_norm": 106.24903869628906,
"learning_rate": 7.2999999999999996e-06,
"loss": 5.7356,
"step": 80
},
{
"epoch": 0.04,
"grad_norm": 409.2347412109375,
"learning_rate": 8.3e-06,
"loss": 5.018,
"step": 90
},
{
"epoch": 0.044444444444444446,
"grad_norm": 817.6410522460938,
"learning_rate": 9.3e-06,
"loss": 5.422,
"step": 100
},
{
"epoch": 0.04888888888888889,
"grad_norm": 589.453125,
"learning_rate": 1.03e-05,
"loss": 3.6104,
"step": 110
},
{
"epoch": 0.05333333333333334,
"grad_norm": 299.5021667480469,
"learning_rate": 1.13e-05,
"loss": 3.3433,
"step": 120
},
{
"epoch": 0.057777777777777775,
"grad_norm": 105.66629791259766,
"learning_rate": 1.23e-05,
"loss": 2.1184,
"step": 130
},
{
"epoch": 0.06222222222222222,
"grad_norm": 94.36296844482422,
"learning_rate": 1.3300000000000001e-05,
"loss": 1.6441,
"step": 140
},
{
"epoch": 0.06666666666666667,
"grad_norm": 99.46455383300781,
"learning_rate": 1.43e-05,
"loss": 1.0691,
"step": 150
},
{
"epoch": 0.07111111111111111,
"grad_norm": 128.905517578125,
"learning_rate": 1.53e-05,
"loss": 1.347,
"step": 160
},
{
"epoch": 0.07555555555555556,
"grad_norm": 10.669831275939941,
"learning_rate": 1.63e-05,
"loss": 0.8163,
"step": 170
},
{
"epoch": 0.08,
"grad_norm": 17.684833526611328,
"learning_rate": 1.73e-05,
"loss": 1.0315,
"step": 180
},
{
"epoch": 0.08444444444444445,
"grad_norm": 35.8447151184082,
"learning_rate": 1.83e-05,
"loss": 0.6946,
"step": 190
},
{
"epoch": 0.08888888888888889,
"grad_norm": 88.678466796875,
"learning_rate": 1.93e-05,
"loss": 1.0834,
"step": 200
},
{
"epoch": 0.09333333333333334,
"grad_norm": 50.5704231262207,
"learning_rate": 2.0300000000000002e-05,
"loss": 0.9062,
"step": 210
},
{
"epoch": 0.09777777777777778,
"grad_norm": 5.56904935836792,
"learning_rate": 2.13e-05,
"loss": 0.9961,
"step": 220
},
{
"epoch": 0.10222222222222223,
"grad_norm": 2.848574638366699,
"learning_rate": 2.23e-05,
"loss": 0.7547,
"step": 230
},
{
"epoch": 0.10666666666666667,
"grad_norm": 51.646854400634766,
"learning_rate": 2.3300000000000004e-05,
"loss": 0.9434,
"step": 240
},
{
"epoch": 0.1111111111111111,
"grad_norm": 4.968932151794434,
"learning_rate": 2.43e-05,
"loss": 0.9282,
"step": 250
},
{
"epoch": 0.11555555555555555,
"grad_norm": 8.864336967468262,
"learning_rate": 2.5300000000000002e-05,
"loss": 1.1923,
"step": 260
},
{
"epoch": 0.12,
"grad_norm": 6.225351333618164,
"learning_rate": 2.6300000000000002e-05,
"loss": 0.7519,
"step": 270
},
{
"epoch": 0.12444444444444444,
"grad_norm": 16.043298721313477,
"learning_rate": 2.7300000000000003e-05,
"loss": 0.8965,
"step": 280
},
{
"epoch": 0.1288888888888889,
"grad_norm": 13.324910163879395,
"learning_rate": 2.83e-05,
"loss": 0.8238,
"step": 290
},
{
"epoch": 0.13333333333333333,
"grad_norm": 3.661782741546631,
"learning_rate": 2.93e-05,
"loss": 0.6619,
"step": 300
},
{
"epoch": 0.13777777777777778,
"grad_norm": 24.01588249206543,
"learning_rate": 3.03e-05,
"loss": 0.9874,
"step": 310
},
{
"epoch": 0.14222222222222222,
"grad_norm": 3.4116015434265137,
"learning_rate": 3.13e-05,
"loss": 0.7866,
"step": 320
},
{
"epoch": 0.14666666666666667,
"grad_norm": 5.378733158111572,
"learning_rate": 3.2300000000000006e-05,
"loss": 0.784,
"step": 330
},
{
"epoch": 0.1511111111111111,
"grad_norm": 3.0666747093200684,
"learning_rate": 3.33e-05,
"loss": 0.8068,
"step": 340
},
{
"epoch": 0.15555555555555556,
"grad_norm": 5.233002662658691,
"learning_rate": 3.430000000000001e-05,
"loss": 0.8338,
"step": 350
},
{
"epoch": 0.16,
"grad_norm": 12.038925170898438,
"learning_rate": 3.53e-05,
"loss": 0.9978,
"step": 360
},
{
"epoch": 0.16444444444444445,
"grad_norm": 4.42678165435791,
"learning_rate": 3.63e-05,
"loss": 0.9186,
"step": 370
},
{
"epoch": 0.1688888888888889,
"grad_norm": 3.2533979415893555,
"learning_rate": 3.73e-05,
"loss": 0.853,
"step": 380
},
{
"epoch": 0.17333333333333334,
"grad_norm": 6.089888095855713,
"learning_rate": 3.83e-05,
"loss": 0.6885,
"step": 390
},
{
"epoch": 0.17777777777777778,
"grad_norm": 2.185434341430664,
"learning_rate": 3.9300000000000007e-05,
"loss": 0.7051,
"step": 400
},
{
"epoch": 0.18222222222222223,
"grad_norm": 3.463395118713379,
"learning_rate": 4.0300000000000004e-05,
"loss": 0.7594,
"step": 410
},
{
"epoch": 0.18666666666666668,
"grad_norm": 111.5975341796875,
"learning_rate": 4.13e-05,
"loss": 0.7679,
"step": 420
},
{
"epoch": 0.19111111111111112,
"grad_norm": 2.420258045196533,
"learning_rate": 4.23e-05,
"loss": 0.9897,
"step": 430
},
{
"epoch": 0.19555555555555557,
"grad_norm": 3.507974863052368,
"learning_rate": 4.33e-05,
"loss": 0.7764,
"step": 440
},
{
"epoch": 0.2,
"grad_norm": 1.6020970344543457,
"learning_rate": 4.43e-05,
"loss": 0.6163,
"step": 450
},
{
"epoch": 0.20444444444444446,
"grad_norm": 1.9605637788772583,
"learning_rate": 4.53e-05,
"loss": 0.9071,
"step": 460
},
{
"epoch": 0.2088888888888889,
"grad_norm": 6.756528377532959,
"learning_rate": 4.630000000000001e-05,
"loss": 0.939,
"step": 470
},
{
"epoch": 0.21333333333333335,
"grad_norm": 6.34450101852417,
"learning_rate": 4.73e-05,
"loss": 0.832,
"step": 480
},
{
"epoch": 0.21777777777777776,
"grad_norm": 3.8294358253479004,
"learning_rate": 4.83e-05,
"loss": 0.9205,
"step": 490
},
{
"epoch": 0.2222222222222222,
"grad_norm": 2.1834263801574707,
"learning_rate": 4.93e-05,
"loss": 0.6335,
"step": 500
},
{
"epoch": 0.22666666666666666,
"grad_norm": 2.837082624435425,
"learning_rate": 4.9914285714285717e-05,
"loss": 0.6933,
"step": 510
},
{
"epoch": 0.2311111111111111,
"grad_norm": 1.8396414518356323,
"learning_rate": 4.962857142857143e-05,
"loss": 1.0322,
"step": 520
},
{
"epoch": 0.23555555555555555,
"grad_norm": 4.330628871917725,
"learning_rate": 4.934285714285715e-05,
"loss": 0.7124,
"step": 530
},
{
"epoch": 0.24,
"grad_norm": 2.490299701690674,
"learning_rate": 4.905714285714286e-05,
"loss": 0.7426,
"step": 540
},
{
"epoch": 0.24444444444444444,
"grad_norm": 1.341838002204895,
"learning_rate": 4.8771428571428574e-05,
"loss": 0.7588,
"step": 550
},
{
"epoch": 0.24888888888888888,
"grad_norm": 2.234884262084961,
"learning_rate": 4.848571428571429e-05,
"loss": 0.8314,
"step": 560
},
{
"epoch": 0.25333333333333335,
"grad_norm": 2.8187923431396484,
"learning_rate": 4.82e-05,
"loss": 0.8109,
"step": 570
},
{
"epoch": 0.2577777777777778,
"grad_norm": 0.9691933393478394,
"learning_rate": 4.7914285714285715e-05,
"loss": 0.6785,
"step": 580
},
{
"epoch": 0.26222222222222225,
"grad_norm": 1.8897961378097534,
"learning_rate": 4.762857142857143e-05,
"loss": 0.9523,
"step": 590
},
{
"epoch": 0.26666666666666666,
"grad_norm": 7.316945552825928,
"learning_rate": 4.734285714285715e-05,
"loss": 0.5996,
"step": 600
},
{
"epoch": 0.27111111111111114,
"grad_norm": 2.491607427597046,
"learning_rate": 4.7057142857142864e-05,
"loss": 0.5426,
"step": 610
},
{
"epoch": 0.27555555555555555,
"grad_norm": 2.7054076194763184,
"learning_rate": 4.677142857142857e-05,
"loss": 0.6294,
"step": 620
},
{
"epoch": 0.28,
"grad_norm": 1.5034236907958984,
"learning_rate": 4.648571428571429e-05,
"loss": 0.4571,
"step": 630
},
{
"epoch": 0.28444444444444444,
"grad_norm": 5.013509750366211,
"learning_rate": 4.6200000000000005e-05,
"loss": 0.9696,
"step": 640
},
{
"epoch": 0.28888888888888886,
"grad_norm": 1.4784796237945557,
"learning_rate": 4.5914285714285714e-05,
"loss": 0.79,
"step": 650
},
{
"epoch": 0.29333333333333333,
"grad_norm": 1.679485559463501,
"learning_rate": 4.562857142857143e-05,
"loss": 0.5445,
"step": 660
},
{
"epoch": 0.29777777777777775,
"grad_norm": 11.93566608428955,
"learning_rate": 4.534285714285714e-05,
"loss": 0.8077,
"step": 670
},
{
"epoch": 0.3022222222222222,
"grad_norm": 2.100651264190674,
"learning_rate": 4.5057142857142856e-05,
"loss": 0.5643,
"step": 680
},
{
"epoch": 0.30666666666666664,
"grad_norm": 2.368565797805786,
"learning_rate": 4.477142857142858e-05,
"loss": 1.0012,
"step": 690
},
{
"epoch": 0.3111111111111111,
"grad_norm": 2.4255175590515137,
"learning_rate": 4.448571428571429e-05,
"loss": 0.6344,
"step": 700
},
{
"epoch": 0.31555555555555553,
"grad_norm": 3.4809205532073975,
"learning_rate": 4.4200000000000004e-05,
"loss": 0.87,
"step": 710
},
{
"epoch": 0.32,
"grad_norm": 12.76645565032959,
"learning_rate": 4.391428571428572e-05,
"loss": 0.673,
"step": 720
},
{
"epoch": 0.3244444444444444,
"grad_norm": 3.417945623397827,
"learning_rate": 4.362857142857143e-05,
"loss": 0.6753,
"step": 730
},
{
"epoch": 0.3288888888888889,
"grad_norm": 2.9550516605377197,
"learning_rate": 4.3342857142857145e-05,
"loss": 0.6621,
"step": 740
},
{
"epoch": 0.3333333333333333,
"grad_norm": 1.9887003898620605,
"learning_rate": 4.3057142857142854e-05,
"loss": 0.671,
"step": 750
},
{
"epoch": 0.3377777777777778,
"grad_norm": 1.826278567314148,
"learning_rate": 4.277142857142857e-05,
"loss": 0.7536,
"step": 760
},
{
"epoch": 0.3422222222222222,
"grad_norm": 4.41416597366333,
"learning_rate": 4.2485714285714286e-05,
"loss": 0.7106,
"step": 770
},
{
"epoch": 0.3466666666666667,
"grad_norm": 1.8409727811813354,
"learning_rate": 4.22e-05,
"loss": 0.7142,
"step": 780
},
{
"epoch": 0.3511111111111111,
"grad_norm": 2.143974781036377,
"learning_rate": 4.191428571428572e-05,
"loss": 0.5623,
"step": 790
},
{
"epoch": 0.35555555555555557,
"grad_norm": 1.1921825408935547,
"learning_rate": 4.162857142857143e-05,
"loss": 0.6638,
"step": 800
},
{
"epoch": 0.36,
"grad_norm": 2.4289398193359375,
"learning_rate": 4.1342857142857144e-05,
"loss": 0.503,
"step": 810
},
{
"epoch": 0.36444444444444446,
"grad_norm": 1.718072533607483,
"learning_rate": 4.105714285714286e-05,
"loss": 0.4197,
"step": 820
},
{
"epoch": 0.3688888888888889,
"grad_norm": 2.278563976287842,
"learning_rate": 4.077142857142857e-05,
"loss": 0.611,
"step": 830
},
{
"epoch": 0.37333333333333335,
"grad_norm": 2.060332775115967,
"learning_rate": 4.0485714285714285e-05,
"loss": 0.5883,
"step": 840
},
{
"epoch": 0.37777777777777777,
"grad_norm": 2.5066239833831787,
"learning_rate": 4.02e-05,
"loss": 0.6179,
"step": 850
},
{
"epoch": 0.38222222222222224,
"grad_norm": 5.195372104644775,
"learning_rate": 3.991428571428572e-05,
"loss": 0.7709,
"step": 860
},
{
"epoch": 0.38666666666666666,
"grad_norm": 3.0029399394989014,
"learning_rate": 3.9628571428571433e-05,
"loss": 0.584,
"step": 870
},
{
"epoch": 0.39111111111111113,
"grad_norm": 6.307543754577637,
"learning_rate": 3.934285714285714e-05,
"loss": 0.6633,
"step": 880
},
{
"epoch": 0.39555555555555555,
"grad_norm": 6.635744571685791,
"learning_rate": 3.905714285714286e-05,
"loss": 0.85,
"step": 890
},
{
"epoch": 0.4,
"grad_norm": 4.6865434646606445,
"learning_rate": 3.8771428571428575e-05,
"loss": 0.6508,
"step": 900
},
{
"epoch": 0.40444444444444444,
"grad_norm": 1.8700509071350098,
"learning_rate": 3.8485714285714284e-05,
"loss": 0.5405,
"step": 910
},
{
"epoch": 0.4088888888888889,
"grad_norm": 3.749453067779541,
"learning_rate": 3.82e-05,
"loss": 0.6869,
"step": 920
},
{
"epoch": 0.41333333333333333,
"grad_norm": 2.443854570388794,
"learning_rate": 3.7914285714285716e-05,
"loss": 0.5013,
"step": 930
},
{
"epoch": 0.4177777777777778,
"grad_norm": 1.066603183746338,
"learning_rate": 3.762857142857143e-05,
"loss": 0.5993,
"step": 940
},
{
"epoch": 0.4222222222222222,
"grad_norm": 5.453496932983398,
"learning_rate": 3.734285714285715e-05,
"loss": 0.633,
"step": 950
},
{
"epoch": 0.4266666666666667,
"grad_norm": 3.3490309715270996,
"learning_rate": 3.705714285714286e-05,
"loss": 0.7452,
"step": 960
},
{
"epoch": 0.4311111111111111,
"grad_norm": 2.070544719696045,
"learning_rate": 3.6771428571428574e-05,
"loss": 0.7698,
"step": 970
},
{
"epoch": 0.43555555555555553,
"grad_norm": 3.1262307167053223,
"learning_rate": 3.648571428571429e-05,
"loss": 0.7202,
"step": 980
},
{
"epoch": 0.44,
"grad_norm": 4.966836929321289,
"learning_rate": 3.62e-05,
"loss": 0.6881,
"step": 990
},
{
"epoch": 0.4444444444444444,
"grad_norm": 0.9398611783981323,
"learning_rate": 3.5914285714285715e-05,
"loss": 0.3937,
"step": 1000
},
{
"epoch": 0.4488888888888889,
"grad_norm": 2.2783327102661133,
"learning_rate": 3.562857142857143e-05,
"loss": 0.5421,
"step": 1010
},
{
"epoch": 0.4533333333333333,
"grad_norm": 1.5350534915924072,
"learning_rate": 3.534285714285715e-05,
"loss": 1.1952,
"step": 1020
},
{
"epoch": 0.4577777777777778,
"grad_norm": 2.5785741806030273,
"learning_rate": 3.505714285714286e-05,
"loss": 0.7421,
"step": 1030
},
{
"epoch": 0.4622222222222222,
"grad_norm": 3.4625842571258545,
"learning_rate": 3.477142857142857e-05,
"loss": 0.6769,
"step": 1040
},
{
"epoch": 0.4666666666666667,
"grad_norm": 1.2348237037658691,
"learning_rate": 3.448571428571429e-05,
"loss": 0.4613,
"step": 1050
},
{
"epoch": 0.4711111111111111,
"grad_norm": 1.4216806888580322,
"learning_rate": 3.4200000000000005e-05,
"loss": 0.6116,
"step": 1060
},
{
"epoch": 0.47555555555555556,
"grad_norm": 1.089032769203186,
"learning_rate": 3.3914285714285714e-05,
"loss": 0.5686,
"step": 1070
},
{
"epoch": 0.48,
"grad_norm": 2.135321855545044,
"learning_rate": 3.362857142857143e-05,
"loss": 0.5668,
"step": 1080
},
{
"epoch": 0.48444444444444446,
"grad_norm": 1.846582055091858,
"learning_rate": 3.334285714285714e-05,
"loss": 0.62,
"step": 1090
},
{
"epoch": 0.4888888888888889,
"grad_norm": 2.3308606147766113,
"learning_rate": 3.305714285714286e-05,
"loss": 0.7323,
"step": 1100
},
{
"epoch": 0.49333333333333335,
"grad_norm": 10.256339073181152,
"learning_rate": 3.277142857142858e-05,
"loss": 0.6122,
"step": 1110
},
{
"epoch": 0.49777777777777776,
"grad_norm": 5.60221529006958,
"learning_rate": 3.248571428571429e-05,
"loss": 0.7023,
"step": 1120
},
{
"epoch": 0.5022222222222222,
"grad_norm": 1.3172476291656494,
"learning_rate": 3.2200000000000003e-05,
"loss": 0.6019,
"step": 1130
},
{
"epoch": 0.5066666666666667,
"grad_norm": 2.984369993209839,
"learning_rate": 3.191428571428571e-05,
"loss": 0.8876,
"step": 1140
},
{
"epoch": 0.5111111111111111,
"grad_norm": 2.645580291748047,
"learning_rate": 3.162857142857143e-05,
"loss": 0.6198,
"step": 1150
},
{
"epoch": 0.5155555555555555,
"grad_norm": 6.525484085083008,
"learning_rate": 3.1342857142857145e-05,
"loss": 0.7378,
"step": 1160
},
{
"epoch": 0.52,
"grad_norm": 2.520108938217163,
"learning_rate": 3.1057142857142854e-05,
"loss": 0.6144,
"step": 1170
},
{
"epoch": 0.5244444444444445,
"grad_norm": 1.4671144485473633,
"learning_rate": 3.077142857142857e-05,
"loss": 0.5624,
"step": 1180
},
{
"epoch": 0.5288888888888889,
"grad_norm": 1.6521360874176025,
"learning_rate": 3.048571428571429e-05,
"loss": 0.8754,
"step": 1190
},
{
"epoch": 0.5333333333333333,
"grad_norm": 1.7824636697769165,
"learning_rate": 3.02e-05,
"loss": 0.6161,
"step": 1200
},
{
"epoch": 0.5377777777777778,
"grad_norm": 3.1892054080963135,
"learning_rate": 2.9914285714285718e-05,
"loss": 0.6746,
"step": 1210
},
{
"epoch": 0.5422222222222223,
"grad_norm": 2.471684217453003,
"learning_rate": 2.9628571428571428e-05,
"loss": 0.7047,
"step": 1220
},
{
"epoch": 0.5466666666666666,
"grad_norm": 1.1938486099243164,
"learning_rate": 2.9342857142857144e-05,
"loss": 0.7613,
"step": 1230
},
{
"epoch": 0.5511111111111111,
"grad_norm": 4.988212585449219,
"learning_rate": 2.905714285714286e-05,
"loss": 0.7503,
"step": 1240
},
{
"epoch": 0.5555555555555556,
"grad_norm": 6.780579566955566,
"learning_rate": 2.8771428571428572e-05,
"loss": 0.5375,
"step": 1250
},
{
"epoch": 0.56,
"grad_norm": 2.5735678672790527,
"learning_rate": 2.848571428571429e-05,
"loss": 0.703,
"step": 1260
},
{
"epoch": 0.5644444444444444,
"grad_norm": 3.509777069091797,
"learning_rate": 2.8199999999999998e-05,
"loss": 0.5273,
"step": 1270
},
{
"epoch": 0.5688888888888889,
"grad_norm": 1.5865542888641357,
"learning_rate": 2.7914285714285714e-05,
"loss": 0.4821,
"step": 1280
},
{
"epoch": 0.5733333333333334,
"grad_norm": 1.2859346866607666,
"learning_rate": 2.762857142857143e-05,
"loss": 0.6347,
"step": 1290
},
{
"epoch": 0.5777777777777777,
"grad_norm": 2.0483293533325195,
"learning_rate": 2.7342857142857142e-05,
"loss": 0.4798,
"step": 1300
},
{
"epoch": 0.5822222222222222,
"grad_norm": 1.2751628160476685,
"learning_rate": 2.705714285714286e-05,
"loss": 0.4283,
"step": 1310
},
{
"epoch": 0.5866666666666667,
"grad_norm": 1.7461509704589844,
"learning_rate": 2.6771428571428575e-05,
"loss": 0.577,
"step": 1320
},
{
"epoch": 0.5911111111111111,
"grad_norm": 2.4847190380096436,
"learning_rate": 2.6485714285714287e-05,
"loss": 0.6608,
"step": 1330
},
{
"epoch": 0.5955555555555555,
"grad_norm": 7.046814441680908,
"learning_rate": 2.6200000000000003e-05,
"loss": 0.7501,
"step": 1340
},
{
"epoch": 0.6,
"grad_norm": 1.8862992525100708,
"learning_rate": 2.5914285714285713e-05,
"loss": 0.5347,
"step": 1350
},
{
"epoch": 0.6044444444444445,
"grad_norm": 1.941525936126709,
"learning_rate": 2.562857142857143e-05,
"loss": 0.6533,
"step": 1360
},
{
"epoch": 0.6088888888888889,
"grad_norm": 1.9246207475662231,
"learning_rate": 2.5342857142857145e-05,
"loss": 0.4619,
"step": 1370
},
{
"epoch": 0.6133333333333333,
"grad_norm": 2.3284366130828857,
"learning_rate": 2.5057142857142857e-05,
"loss": 0.5442,
"step": 1380
},
{
"epoch": 0.6177777777777778,
"grad_norm": 3.2495412826538086,
"learning_rate": 2.4771428571428573e-05,
"loss": 0.5538,
"step": 1390
},
{
"epoch": 0.6222222222222222,
"grad_norm": 1.9646495580673218,
"learning_rate": 2.4485714285714286e-05,
"loss": 0.6764,
"step": 1400
},
{
"epoch": 0.6266666666666667,
"grad_norm": 2.3741135597229004,
"learning_rate": 2.4200000000000002e-05,
"loss": 0.4923,
"step": 1410
},
{
"epoch": 0.6311111111111111,
"grad_norm": 3.175323486328125,
"learning_rate": 2.3914285714285715e-05,
"loss": 0.5318,
"step": 1420
},
{
"epoch": 0.6355555555555555,
"grad_norm": 4.018131732940674,
"learning_rate": 2.362857142857143e-05,
"loss": 0.6042,
"step": 1430
},
{
"epoch": 0.64,
"grad_norm": 5.065487384796143,
"learning_rate": 2.3342857142857143e-05,
"loss": 0.4944,
"step": 1440
},
{
"epoch": 0.6444444444444445,
"grad_norm": 4.733597278594971,
"learning_rate": 2.3057142857142856e-05,
"loss": 0.5209,
"step": 1450
},
{
"epoch": 0.6488888888888888,
"grad_norm": 2.8453125953674316,
"learning_rate": 2.2771428571428572e-05,
"loss": 0.6981,
"step": 1460
},
{
"epoch": 0.6533333333333333,
"grad_norm": 5.391908168792725,
"learning_rate": 2.2485714285714288e-05,
"loss": 0.6173,
"step": 1470
},
{
"epoch": 0.6577777777777778,
"grad_norm": 3.4642038345336914,
"learning_rate": 2.22e-05,
"loss": 0.5644,
"step": 1480
},
{
"epoch": 0.6622222222222223,
"grad_norm": 5.350740909576416,
"learning_rate": 2.1914285714285714e-05,
"loss": 0.6647,
"step": 1490
},
{
"epoch": 0.6666666666666666,
"grad_norm": 1.7532992362976074,
"learning_rate": 2.162857142857143e-05,
"loss": 0.5771,
"step": 1500
},
{
"epoch": 0.6711111111111111,
"grad_norm": 1.6923083066940308,
"learning_rate": 2.1342857142857146e-05,
"loss": 0.4541,
"step": 1510
},
{
"epoch": 0.6755555555555556,
"grad_norm": 3.4142796993255615,
"learning_rate": 2.105714285714286e-05,
"loss": 0.6353,
"step": 1520
},
{
"epoch": 0.68,
"grad_norm": 1.2692922353744507,
"learning_rate": 2.077142857142857e-05,
"loss": 0.6471,
"step": 1530
},
{
"epoch": 0.6844444444444444,
"grad_norm": 6.867006778717041,
"learning_rate": 2.0485714285714287e-05,
"loss": 0.5668,
"step": 1540
},
{
"epoch": 0.6888888888888889,
"grad_norm": 1.1940428018569946,
"learning_rate": 2.0200000000000003e-05,
"loss": 0.7113,
"step": 1550
},
{
"epoch": 0.6933333333333334,
"grad_norm": 0.9745553731918335,
"learning_rate": 1.9914285714285716e-05,
"loss": 0.6865,
"step": 1560
},
{
"epoch": 0.6977777777777778,
"grad_norm": 1.4725788831710815,
"learning_rate": 1.962857142857143e-05,
"loss": 0.7809,
"step": 1570
},
{
"epoch": 0.7022222222222222,
"grad_norm": 2.4500861167907715,
"learning_rate": 1.9342857142857144e-05,
"loss": 0.5468,
"step": 1580
},
{
"epoch": 0.7066666666666667,
"grad_norm": 13.177860260009766,
"learning_rate": 1.9057142857142857e-05,
"loss": 0.5427,
"step": 1590
},
{
"epoch": 0.7111111111111111,
"grad_norm": 5.718009948730469,
"learning_rate": 1.8771428571428573e-05,
"loss": 0.6003,
"step": 1600
},
{
"epoch": 0.7155555555555555,
"grad_norm": 2.4186694622039795,
"learning_rate": 1.8485714285714286e-05,
"loss": 0.4843,
"step": 1610
},
{
"epoch": 0.72,
"grad_norm": 1.150259256362915,
"learning_rate": 1.8200000000000002e-05,
"loss": 0.4339,
"step": 1620
},
{
"epoch": 0.7244444444444444,
"grad_norm": 2.5581448078155518,
"learning_rate": 1.7914285714285715e-05,
"loss": 0.7765,
"step": 1630
},
{
"epoch": 0.7288888888888889,
"grad_norm": 2.4859399795532227,
"learning_rate": 1.762857142857143e-05,
"loss": 0.5869,
"step": 1640
},
{
"epoch": 0.7333333333333333,
"grad_norm": 1.6887503862380981,
"learning_rate": 1.7342857142857143e-05,
"loss": 0.4651,
"step": 1650
},
{
"epoch": 0.7377777777777778,
"grad_norm": 2.0939149856567383,
"learning_rate": 1.7057142857142856e-05,
"loss": 0.7603,
"step": 1660
},
{
"epoch": 0.7422222222222222,
"grad_norm": 1.9502744674682617,
"learning_rate": 1.6771428571428572e-05,
"loss": 0.5475,
"step": 1670
},
{
"epoch": 0.7466666666666667,
"grad_norm": 0.8679394721984863,
"learning_rate": 1.6485714285714288e-05,
"loss": 0.4752,
"step": 1680
},
{
"epoch": 0.7511111111111111,
"grad_norm": 1.2583119869232178,
"learning_rate": 1.62e-05,
"loss": 0.5677,
"step": 1690
},
{
"epoch": 0.7555555555555555,
"grad_norm": 2.252943754196167,
"learning_rate": 1.5914285714285713e-05,
"loss": 0.4604,
"step": 1700
},
{
"epoch": 0.76,
"grad_norm": 1.150745153427124,
"learning_rate": 1.562857142857143e-05,
"loss": 0.5713,
"step": 1710
},
{
"epoch": 0.7644444444444445,
"grad_norm": 1.7483420372009277,
"learning_rate": 1.5342857142857146e-05,
"loss": 0.5989,
"step": 1720
},
{
"epoch": 0.7688888888888888,
"grad_norm": 0.8522917628288269,
"learning_rate": 1.5057142857142858e-05,
"loss": 0.5211,
"step": 1730
},
{
"epoch": 0.7733333333333333,
"grad_norm": 2.067723035812378,
"learning_rate": 1.4771428571428573e-05,
"loss": 0.6904,
"step": 1740
},
{
"epoch": 0.7777777777777778,
"grad_norm": 1.3494257926940918,
"learning_rate": 1.4485714285714285e-05,
"loss": 0.6667,
"step": 1750
},
{
"epoch": 0.7822222222222223,
"grad_norm": 1.3629013299942017,
"learning_rate": 1.42e-05,
"loss": 0.3367,
"step": 1760
},
{
"epoch": 0.7866666666666666,
"grad_norm": 3.2202422618865967,
"learning_rate": 1.3914285714285716e-05,
"loss": 0.6033,
"step": 1770
},
{
"epoch": 0.7911111111111111,
"grad_norm": 1.038692593574524,
"learning_rate": 1.362857142857143e-05,
"loss": 0.5305,
"step": 1780
},
{
"epoch": 0.7955555555555556,
"grad_norm": 2.57888126373291,
"learning_rate": 1.3342857142857143e-05,
"loss": 0.5989,
"step": 1790
},
{
"epoch": 0.8,
"grad_norm": 1.118715524673462,
"learning_rate": 1.3057142857142857e-05,
"loss": 0.4963,
"step": 1800
},
{
"epoch": 0.8044444444444444,
"grad_norm": 2.586069107055664,
"learning_rate": 1.2771428571428573e-05,
"loss": 0.6829,
"step": 1810
},
{
"epoch": 0.8088888888888889,
"grad_norm": 2.27518630027771,
"learning_rate": 1.2485714285714287e-05,
"loss": 0.7121,
"step": 1820
},
{
"epoch": 0.8133333333333334,
"grad_norm": 1.1266392469406128,
"learning_rate": 1.22e-05,
"loss": 0.5293,
"step": 1830
},
{
"epoch": 0.8177777777777778,
"grad_norm": 1.4908055067062378,
"learning_rate": 1.1914285714285716e-05,
"loss": 0.5562,
"step": 1840
},
{
"epoch": 0.8222222222222222,
"grad_norm": 1.4283602237701416,
"learning_rate": 1.1628571428571429e-05,
"loss": 0.4831,
"step": 1850
},
{
"epoch": 0.8266666666666667,
"grad_norm": 1.0552829504013062,
"learning_rate": 1.1342857142857143e-05,
"loss": 0.5761,
"step": 1860
},
{
"epoch": 0.8311111111111111,
"grad_norm": 6.45361852645874,
"learning_rate": 1.1057142857142858e-05,
"loss": 0.6874,
"step": 1870
},
{
"epoch": 0.8355555555555556,
"grad_norm": 1.207909345626831,
"learning_rate": 1.0771428571428572e-05,
"loss": 0.6289,
"step": 1880
},
{
"epoch": 0.84,
"grad_norm": 3.6151106357574463,
"learning_rate": 1.0485714285714286e-05,
"loss": 0.78,
"step": 1890
},
{
"epoch": 0.8444444444444444,
"grad_norm": 2.560347318649292,
"learning_rate": 1.02e-05,
"loss": 0.7657,
"step": 1900
},
{
"epoch": 0.8488888888888889,
"grad_norm": 1.1826066970825195,
"learning_rate": 9.914285714285715e-06,
"loss": 0.5419,
"step": 1910
},
{
"epoch": 0.8533333333333334,
"grad_norm": 2.4970545768737793,
"learning_rate": 9.628571428571428e-06,
"loss": 0.5137,
"step": 1920
},
{
"epoch": 0.8577777777777778,
"grad_norm": 1.6849817037582397,
"learning_rate": 9.342857142857144e-06,
"loss": 0.4476,
"step": 1930
},
{
"epoch": 0.8622222222222222,
"grad_norm": 3.5557351112365723,
"learning_rate": 9.057142857142856e-06,
"loss": 0.6422,
"step": 1940
},
{
"epoch": 0.8666666666666667,
"grad_norm": 1.7327756881713867,
"learning_rate": 8.771428571428572e-06,
"loss": 0.656,
"step": 1950
},
{
"epoch": 0.8711111111111111,
"grad_norm": 1.3222990036010742,
"learning_rate": 8.485714285714285e-06,
"loss": 0.6862,
"step": 1960
},
{
"epoch": 0.8755555555555555,
"grad_norm": 1.3732144832611084,
"learning_rate": 8.200000000000001e-06,
"loss": 0.5173,
"step": 1970
},
{
"epoch": 0.88,
"grad_norm": 5.491960048675537,
"learning_rate": 7.914285714285714e-06,
"loss": 0.5735,
"step": 1980
},
{
"epoch": 0.8844444444444445,
"grad_norm": 1.9830958843231201,
"learning_rate": 7.628571428571429e-06,
"loss": 0.4936,
"step": 1990
},
{
"epoch": 0.8888888888888888,
"grad_norm": 3.4833927154541016,
"learning_rate": 7.342857142857143e-06,
"loss": 0.5099,
"step": 2000
},
{
"epoch": 0.8933333333333333,
"grad_norm": 1.197811484336853,
"learning_rate": 7.057142857142858e-06,
"loss": 0.4711,
"step": 2010
},
{
"epoch": 0.8977777777777778,
"grad_norm": 2.15141224861145,
"learning_rate": 6.771428571428571e-06,
"loss": 0.5555,
"step": 2020
},
{
"epoch": 0.9022222222222223,
"grad_norm": 1.7254974842071533,
"learning_rate": 6.485714285714286e-06,
"loss": 0.5582,
"step": 2030
},
{
"epoch": 0.9066666666666666,
"grad_norm": 9.579049110412598,
"learning_rate": 6.2e-06,
"loss": 0.4571,
"step": 2040
},
{
"epoch": 0.9111111111111111,
"grad_norm": 2.882373094558716,
"learning_rate": 5.914285714285714e-06,
"loss": 0.6287,
"step": 2050
},
{
"epoch": 0.9155555555555556,
"grad_norm": 5.913229942321777,
"learning_rate": 5.628571428571429e-06,
"loss": 0.5779,
"step": 2060
},
{
"epoch": 0.92,
"grad_norm": 4.352343559265137,
"learning_rate": 5.342857142857143e-06,
"loss": 0.5565,
"step": 2070
},
{
"epoch": 0.9244444444444444,
"grad_norm": 2.8510098457336426,
"learning_rate": 5.057142857142857e-06,
"loss": 0.5999,
"step": 2080
},
{
"epoch": 0.9288888888888889,
"grad_norm": 2.46824049949646,
"learning_rate": 4.771428571428572e-06,
"loss": 0.4779,
"step": 2090
},
{
"epoch": 0.9333333333333333,
"grad_norm": 1.5886437892913818,
"learning_rate": 4.485714285714286e-06,
"loss": 0.5774,
"step": 2100
},
{
"epoch": 0.9377777777777778,
"grad_norm": 1.5184223651885986,
"learning_rate": 4.2000000000000004e-06,
"loss": 0.4278,
"step": 2110
},
{
"epoch": 0.9422222222222222,
"grad_norm": 4.225131511688232,
"learning_rate": 3.914285714285715e-06,
"loss": 0.6479,
"step": 2120
},
{
"epoch": 0.9466666666666667,
"grad_norm": 2.665025234222412,
"learning_rate": 3.6285714285714283e-06,
"loss": 0.4907,
"step": 2130
},
{
"epoch": 0.9511111111111111,
"grad_norm": 1.4222490787506104,
"learning_rate": 3.3428571428571427e-06,
"loss": 0.4263,
"step": 2140
},
{
"epoch": 0.9555555555555556,
"grad_norm": 1.8098088502883911,
"learning_rate": 3.0571428571428575e-06,
"loss": 0.512,
"step": 2150
},
{
"epoch": 0.96,
"grad_norm": 1.5185695886611938,
"learning_rate": 2.771428571428572e-06,
"loss": 0.4707,
"step": 2160
},
{
"epoch": 0.9644444444444444,
"grad_norm": 3.510244846343994,
"learning_rate": 2.4857142857142858e-06,
"loss": 0.5985,
"step": 2170
},
{
"epoch": 0.9688888888888889,
"grad_norm": 2.4858155250549316,
"learning_rate": 2.2e-06,
"loss": 0.5148,
"step": 2180
},
{
"epoch": 0.9733333333333334,
"grad_norm": 2.5503909587860107,
"learning_rate": 1.9142857142857145e-06,
"loss": 0.5905,
"step": 2190
},
{
"epoch": 0.9777777777777777,
"grad_norm": 1.640555739402771,
"learning_rate": 1.6285714285714286e-06,
"loss": 0.4983,
"step": 2200
},
{
"epoch": 0.9822222222222222,
"grad_norm": 3.124906539916992,
"learning_rate": 1.342857142857143e-06,
"loss": 0.4201,
"step": 2210
},
{
"epoch": 0.9866666666666667,
"grad_norm": 2.2014262676239014,
"learning_rate": 1.0571428571428573e-06,
"loss": 0.5991,
"step": 2220
},
{
"epoch": 0.9911111111111112,
"grad_norm": 1.412158727645874,
"learning_rate": 7.714285714285715e-07,
"loss": 0.5062,
"step": 2230
},
{
"epoch": 0.9955555555555555,
"grad_norm": 2.74906325340271,
"learning_rate": 4.857142857142857e-07,
"loss": 0.6499,
"step": 2240
},
{
"epoch": 1.0,
"grad_norm": 1.1791956424713135,
"learning_rate": 2.0000000000000002e-07,
"loss": 0.521,
"step": 2250
}
],
"logging_steps": 10,
"max_steps": 2250,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 5000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.20270921367552e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}