sst2-BERT-base / trainer_state.json
ung30's picture
Upload 8 files
a565f8d verified
{
"best_global_step": 876,
"best_metric": 0.9277850539806721,
"best_model_checkpoint": "./saved_best/sst2_BERT-20000/full_text/checkpoint-876",
"epoch": 3.0,
"eval_steps": 500,
"global_step": 1314,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0228310502283105,
"grad_norm": 127786.96875,
"learning_rate": 1.9863013698630137e-05,
"loss": 0.7128,
"step": 10
},
{
"epoch": 0.045662100456621,
"grad_norm": 288305.125,
"learning_rate": 1.971080669710807e-05,
"loss": 0.6404,
"step": 20
},
{
"epoch": 0.0684931506849315,
"grad_norm": 372462.9375,
"learning_rate": 1.9558599695585997e-05,
"loss": 0.5327,
"step": 30
},
{
"epoch": 0.091324200913242,
"grad_norm": 602445.375,
"learning_rate": 1.940639269406393e-05,
"loss": 0.432,
"step": 40
},
{
"epoch": 0.1141552511415525,
"grad_norm": 358026.8125,
"learning_rate": 1.9254185692541858e-05,
"loss": 0.3758,
"step": 50
},
{
"epoch": 0.136986301369863,
"grad_norm": 402071.21875,
"learning_rate": 1.910197869101979e-05,
"loss": 0.3234,
"step": 60
},
{
"epoch": 0.1598173515981735,
"grad_norm": 352748.53125,
"learning_rate": 1.8949771689497718e-05,
"loss": 0.3151,
"step": 70
},
{
"epoch": 0.182648401826484,
"grad_norm": 905835.5625,
"learning_rate": 1.879756468797565e-05,
"loss": 0.3166,
"step": 80
},
{
"epoch": 0.2054794520547945,
"grad_norm": 466547.9375,
"learning_rate": 1.8645357686453578e-05,
"loss": 0.3419,
"step": 90
},
{
"epoch": 0.228310502283105,
"grad_norm": 423154.84375,
"learning_rate": 1.849315068493151e-05,
"loss": 0.3481,
"step": 100
},
{
"epoch": 0.2511415525114155,
"grad_norm": 297082.96875,
"learning_rate": 1.8340943683409438e-05,
"loss": 0.3115,
"step": 110
},
{
"epoch": 0.273972602739726,
"grad_norm": 369912.21875,
"learning_rate": 1.818873668188737e-05,
"loss": 0.316,
"step": 120
},
{
"epoch": 0.2968036529680365,
"grad_norm": 197489.21875,
"learning_rate": 1.80365296803653e-05,
"loss": 0.295,
"step": 130
},
{
"epoch": 0.319634703196347,
"grad_norm": 557747.25,
"learning_rate": 1.788432267884323e-05,
"loss": 0.2646,
"step": 140
},
{
"epoch": 0.3424657534246575,
"grad_norm": 1081537.125,
"learning_rate": 1.773211567732116e-05,
"loss": 0.2154,
"step": 150
},
{
"epoch": 0.365296803652968,
"grad_norm": 468382.625,
"learning_rate": 1.757990867579909e-05,
"loss": 0.3291,
"step": 160
},
{
"epoch": 0.3881278538812785,
"grad_norm": 1200675.625,
"learning_rate": 1.742770167427702e-05,
"loss": 0.3015,
"step": 170
},
{
"epoch": 0.410958904109589,
"grad_norm": 290634.03125,
"learning_rate": 1.727549467275495e-05,
"loss": 0.2645,
"step": 180
},
{
"epoch": 0.4337899543378995,
"grad_norm": 785782.75,
"learning_rate": 1.712328767123288e-05,
"loss": 0.3038,
"step": 190
},
{
"epoch": 0.45662100456621,
"grad_norm": 454525.84375,
"learning_rate": 1.6971080669710807e-05,
"loss": 0.2706,
"step": 200
},
{
"epoch": 0.4794520547945205,
"grad_norm": 366387.75,
"learning_rate": 1.681887366818874e-05,
"loss": 0.2778,
"step": 210
},
{
"epoch": 0.502283105022831,
"grad_norm": 213485.28125,
"learning_rate": 1.6666666666666667e-05,
"loss": 0.2431,
"step": 220
},
{
"epoch": 0.5251141552511416,
"grad_norm": 434280.5625,
"learning_rate": 1.65144596651446e-05,
"loss": 0.2712,
"step": 230
},
{
"epoch": 0.547945205479452,
"grad_norm": 444368.0625,
"learning_rate": 1.6362252663622528e-05,
"loss": 0.2257,
"step": 240
},
{
"epoch": 0.5707762557077626,
"grad_norm": 325988.9375,
"learning_rate": 1.6210045662100456e-05,
"loss": 0.2046,
"step": 250
},
{
"epoch": 0.593607305936073,
"grad_norm": 376345.3125,
"learning_rate": 1.6057838660578388e-05,
"loss": 0.2685,
"step": 260
},
{
"epoch": 0.6164383561643836,
"grad_norm": 596176.125,
"learning_rate": 1.5905631659056316e-05,
"loss": 0.2729,
"step": 270
},
{
"epoch": 0.639269406392694,
"grad_norm": 650678.25,
"learning_rate": 1.5753424657534248e-05,
"loss": 0.2352,
"step": 280
},
{
"epoch": 0.6621004566210046,
"grad_norm": 688517.5,
"learning_rate": 1.5601217656012176e-05,
"loss": 0.2395,
"step": 290
},
{
"epoch": 0.684931506849315,
"grad_norm": 550269.5625,
"learning_rate": 1.5449010654490108e-05,
"loss": 0.2338,
"step": 300
},
{
"epoch": 0.7077625570776256,
"grad_norm": 382038.0625,
"learning_rate": 1.5296803652968037e-05,
"loss": 0.2819,
"step": 310
},
{
"epoch": 0.730593607305936,
"grad_norm": 413848.71875,
"learning_rate": 1.5144596651445968e-05,
"loss": 0.182,
"step": 320
},
{
"epoch": 0.7534246575342466,
"grad_norm": 833068.375,
"learning_rate": 1.4992389649923897e-05,
"loss": 0.237,
"step": 330
},
{
"epoch": 0.776255707762557,
"grad_norm": 249349.609375,
"learning_rate": 1.4840182648401829e-05,
"loss": 0.2295,
"step": 340
},
{
"epoch": 0.7990867579908676,
"grad_norm": 195666.703125,
"learning_rate": 1.4687975646879757e-05,
"loss": 0.1964,
"step": 350
},
{
"epoch": 0.821917808219178,
"grad_norm": 526547.625,
"learning_rate": 1.4535768645357689e-05,
"loss": 0.3156,
"step": 360
},
{
"epoch": 0.8447488584474886,
"grad_norm": 733883.0,
"learning_rate": 1.4383561643835617e-05,
"loss": 0.2027,
"step": 370
},
{
"epoch": 0.867579908675799,
"grad_norm": 299320.78125,
"learning_rate": 1.4231354642313549e-05,
"loss": 0.2713,
"step": 380
},
{
"epoch": 0.8904109589041096,
"grad_norm": 440281.78125,
"learning_rate": 1.4079147640791477e-05,
"loss": 0.2047,
"step": 390
},
{
"epoch": 0.91324200913242,
"grad_norm": 462590.125,
"learning_rate": 1.3926940639269409e-05,
"loss": 0.2395,
"step": 400
},
{
"epoch": 0.9360730593607306,
"grad_norm": 1140073.75,
"learning_rate": 1.3774733637747338e-05,
"loss": 0.216,
"step": 410
},
{
"epoch": 0.958904109589041,
"grad_norm": 197822.625,
"learning_rate": 1.3622526636225268e-05,
"loss": 0.2449,
"step": 420
},
{
"epoch": 0.9817351598173516,
"grad_norm": 638955.5625,
"learning_rate": 1.3470319634703198e-05,
"loss": 0.253,
"step": 430
},
{
"epoch": 1.0,
"eval_accuracy": 0.9223333333333333,
"eval_f1_macro": 0.9212113479103987,
"eval_f1_weighted": 0.9221891676085987,
"eval_loss": 0.2171599268913269,
"eval_runtime": 14.7894,
"eval_samples_per_second": 405.695,
"eval_steps_per_second": 12.712,
"step": 438
},
{
"epoch": 1.004566210045662,
"grad_norm": 654569.875,
"learning_rate": 1.3318112633181128e-05,
"loss": 0.18,
"step": 440
},
{
"epoch": 1.0273972602739727,
"grad_norm": 182483.90625,
"learning_rate": 1.3165905631659058e-05,
"loss": 0.1107,
"step": 450
},
{
"epoch": 1.0502283105022832,
"grad_norm": 185298.140625,
"learning_rate": 1.3013698630136988e-05,
"loss": 0.1458,
"step": 460
},
{
"epoch": 1.0730593607305936,
"grad_norm": 21770.05078125,
"learning_rate": 1.2861491628614916e-05,
"loss": 0.1592,
"step": 470
},
{
"epoch": 1.095890410958904,
"grad_norm": 851453.125,
"learning_rate": 1.2709284627092848e-05,
"loss": 0.1383,
"step": 480
},
{
"epoch": 1.1187214611872145,
"grad_norm": 348289.34375,
"learning_rate": 1.2557077625570777e-05,
"loss": 0.1205,
"step": 490
},
{
"epoch": 1.1415525114155252,
"grad_norm": 510975.78125,
"learning_rate": 1.2404870624048708e-05,
"loss": 0.1321,
"step": 500
},
{
"epoch": 1.1643835616438356,
"grad_norm": 1570861.375,
"learning_rate": 1.2252663622526637e-05,
"loss": 0.221,
"step": 510
},
{
"epoch": 1.187214611872146,
"grad_norm": 948502.5625,
"learning_rate": 1.2100456621004569e-05,
"loss": 0.1814,
"step": 520
},
{
"epoch": 1.2100456621004567,
"grad_norm": 597450.25,
"learning_rate": 1.1948249619482497e-05,
"loss": 0.0911,
"step": 530
},
{
"epoch": 1.2328767123287672,
"grad_norm": 397320.8125,
"learning_rate": 1.1796042617960429e-05,
"loss": 0.1649,
"step": 540
},
{
"epoch": 1.2557077625570776,
"grad_norm": 513346.84375,
"learning_rate": 1.1643835616438357e-05,
"loss": 0.1953,
"step": 550
},
{
"epoch": 1.278538812785388,
"grad_norm": 504690.0625,
"learning_rate": 1.1491628614916289e-05,
"loss": 0.14,
"step": 560
},
{
"epoch": 1.3013698630136985,
"grad_norm": 1110943.625,
"learning_rate": 1.1339421613394217e-05,
"loss": 0.1191,
"step": 570
},
{
"epoch": 1.3242009132420092,
"grad_norm": 200145.484375,
"learning_rate": 1.1187214611872147e-05,
"loss": 0.1339,
"step": 580
},
{
"epoch": 1.3470319634703196,
"grad_norm": 221766.21875,
"learning_rate": 1.1035007610350078e-05,
"loss": 0.1427,
"step": 590
},
{
"epoch": 1.36986301369863,
"grad_norm": 881168.3125,
"learning_rate": 1.0882800608828008e-05,
"loss": 0.1089,
"step": 600
},
{
"epoch": 1.3926940639269407,
"grad_norm": 854495.5625,
"learning_rate": 1.0730593607305938e-05,
"loss": 0.1368,
"step": 610
},
{
"epoch": 1.4155251141552512,
"grad_norm": 115748.078125,
"learning_rate": 1.0578386605783868e-05,
"loss": 0.15,
"step": 620
},
{
"epoch": 1.4383561643835616,
"grad_norm": 334710.25,
"learning_rate": 1.0426179604261796e-05,
"loss": 0.2423,
"step": 630
},
{
"epoch": 1.461187214611872,
"grad_norm": 972609.8125,
"learning_rate": 1.0273972602739728e-05,
"loss": 0.093,
"step": 640
},
{
"epoch": 1.4840182648401825,
"grad_norm": 519533.71875,
"learning_rate": 1.0121765601217656e-05,
"loss": 0.217,
"step": 650
},
{
"epoch": 1.5068493150684932,
"grad_norm": 156004.296875,
"learning_rate": 9.969558599695586e-06,
"loss": 0.1424,
"step": 660
},
{
"epoch": 1.5296803652968036,
"grad_norm": 398576.53125,
"learning_rate": 9.817351598173517e-06,
"loss": 0.1404,
"step": 670
},
{
"epoch": 1.5525114155251143,
"grad_norm": 648208.25,
"learning_rate": 9.665144596651447e-06,
"loss": 0.1277,
"step": 680
},
{
"epoch": 1.5753424657534247,
"grad_norm": 71873.3046875,
"learning_rate": 9.512937595129377e-06,
"loss": 0.1247,
"step": 690
},
{
"epoch": 1.5981735159817352,
"grad_norm": 349757.6875,
"learning_rate": 9.360730593607307e-06,
"loss": 0.1786,
"step": 700
},
{
"epoch": 1.6210045662100456,
"grad_norm": 395071.59375,
"learning_rate": 9.208523592085237e-06,
"loss": 0.1687,
"step": 710
},
{
"epoch": 1.643835616438356,
"grad_norm": 689678.75,
"learning_rate": 9.056316590563167e-06,
"loss": 0.1239,
"step": 720
},
{
"epoch": 1.6666666666666665,
"grad_norm": 609217.75,
"learning_rate": 8.904109589041097e-06,
"loss": 0.1193,
"step": 730
},
{
"epoch": 1.6894977168949772,
"grad_norm": 429054.15625,
"learning_rate": 8.751902587519027e-06,
"loss": 0.0972,
"step": 740
},
{
"epoch": 1.7123287671232876,
"grad_norm": 658631.0,
"learning_rate": 8.599695585996957e-06,
"loss": 0.2026,
"step": 750
},
{
"epoch": 1.7351598173515983,
"grad_norm": 696978.75,
"learning_rate": 8.447488584474887e-06,
"loss": 0.2093,
"step": 760
},
{
"epoch": 1.7579908675799087,
"grad_norm": 1273698.5,
"learning_rate": 8.295281582952816e-06,
"loss": 0.131,
"step": 770
},
{
"epoch": 1.7808219178082192,
"grad_norm": 381059.125,
"learning_rate": 8.143074581430746e-06,
"loss": 0.0753,
"step": 780
},
{
"epoch": 1.8036529680365296,
"grad_norm": 760526.8125,
"learning_rate": 7.990867579908676e-06,
"loss": 0.1004,
"step": 790
},
{
"epoch": 1.82648401826484,
"grad_norm": 197580.625,
"learning_rate": 7.838660578386606e-06,
"loss": 0.1652,
"step": 800
},
{
"epoch": 1.8493150684931505,
"grad_norm": 1663549.75,
"learning_rate": 7.686453576864536e-06,
"loss": 0.1465,
"step": 810
},
{
"epoch": 1.8721461187214612,
"grad_norm": 517745.75,
"learning_rate": 7.534246575342466e-06,
"loss": 0.1703,
"step": 820
},
{
"epoch": 1.8949771689497716,
"grad_norm": 680767.5625,
"learning_rate": 7.382039573820396e-06,
"loss": 0.1623,
"step": 830
},
{
"epoch": 1.9178082191780823,
"grad_norm": 697195.1875,
"learning_rate": 7.2298325722983265e-06,
"loss": 0.1733,
"step": 840
},
{
"epoch": 1.9406392694063928,
"grad_norm": 263220.8125,
"learning_rate": 7.077625570776257e-06,
"loss": 0.116,
"step": 850
},
{
"epoch": 1.9634703196347032,
"grad_norm": 336279.6875,
"learning_rate": 6.925418569254187e-06,
"loss": 0.1565,
"step": 860
},
{
"epoch": 1.9863013698630136,
"grad_norm": 230299.015625,
"learning_rate": 6.773211567732117e-06,
"loss": 0.1184,
"step": 870
},
{
"epoch": 2.0,
"eval_accuracy": 0.9285,
"eval_f1_macro": 0.9277850539806721,
"eval_f1_weighted": 0.92853233424208,
"eval_loss": 0.2250971645116806,
"eval_runtime": 15.009,
"eval_samples_per_second": 399.76,
"eval_steps_per_second": 12.526,
"step": 876
},
{
"epoch": 2.009132420091324,
"grad_norm": 105916.703125,
"learning_rate": 6.621004566210046e-06,
"loss": 0.1208,
"step": 880
},
{
"epoch": 2.0319634703196345,
"grad_norm": 231579.703125,
"learning_rate": 6.468797564687976e-06,
"loss": 0.0994,
"step": 890
},
{
"epoch": 2.0547945205479454,
"grad_norm": 226497.375,
"learning_rate": 6.316590563165906e-06,
"loss": 0.0713,
"step": 900
},
{
"epoch": 2.077625570776256,
"grad_norm": 275743.15625,
"learning_rate": 6.164383561643836e-06,
"loss": 0.0873,
"step": 910
},
{
"epoch": 2.1004566210045663,
"grad_norm": 28719.689453125,
"learning_rate": 6.012176560121766e-06,
"loss": 0.1131,
"step": 920
},
{
"epoch": 2.1232876712328768,
"grad_norm": 947971.3125,
"learning_rate": 5.8599695585996965e-06,
"loss": 0.0514,
"step": 930
},
{
"epoch": 2.146118721461187,
"grad_norm": 768273.125,
"learning_rate": 5.7077625570776266e-06,
"loss": 0.0813,
"step": 940
},
{
"epoch": 2.1689497716894977,
"grad_norm": 202898.15625,
"learning_rate": 5.555555555555557e-06,
"loss": 0.0813,
"step": 950
},
{
"epoch": 2.191780821917808,
"grad_norm": 481656.03125,
"learning_rate": 5.403348554033486e-06,
"loss": 0.1004,
"step": 960
},
{
"epoch": 2.2146118721461185,
"grad_norm": 250803.953125,
"learning_rate": 5.251141552511416e-06,
"loss": 0.1432,
"step": 970
},
{
"epoch": 2.237442922374429,
"grad_norm": 12995.201171875,
"learning_rate": 5.098934550989346e-06,
"loss": 0.0464,
"step": 980
},
{
"epoch": 2.26027397260274,
"grad_norm": 902957.0625,
"learning_rate": 4.946727549467275e-06,
"loss": 0.137,
"step": 990
},
{
"epoch": 2.2831050228310503,
"grad_norm": 534876.0,
"learning_rate": 4.7945205479452054e-06,
"loss": 0.0922,
"step": 1000
},
{
"epoch": 2.3059360730593608,
"grad_norm": 178206.484375,
"learning_rate": 4.6423135464231355e-06,
"loss": 0.0705,
"step": 1010
},
{
"epoch": 2.328767123287671,
"grad_norm": 40234.16796875,
"learning_rate": 4.490106544901066e-06,
"loss": 0.0372,
"step": 1020
},
{
"epoch": 2.3515981735159817,
"grad_norm": 251605.671875,
"learning_rate": 4.337899543378996e-06,
"loss": 0.0483,
"step": 1030
},
{
"epoch": 2.374429223744292,
"grad_norm": 1325790.375,
"learning_rate": 4.185692541856926e-06,
"loss": 0.0806,
"step": 1040
},
{
"epoch": 2.3972602739726026,
"grad_norm": 793383.375,
"learning_rate": 4.033485540334856e-06,
"loss": 0.123,
"step": 1050
},
{
"epoch": 2.4200913242009134,
"grad_norm": 626307.3125,
"learning_rate": 3.881278538812785e-06,
"loss": 0.117,
"step": 1060
},
{
"epoch": 2.442922374429224,
"grad_norm": 390102.96875,
"learning_rate": 3.7290715372907157e-06,
"loss": 0.1038,
"step": 1070
},
{
"epoch": 2.4657534246575343,
"grad_norm": 965051.0625,
"learning_rate": 3.5768645357686453e-06,
"loss": 0.0869,
"step": 1080
},
{
"epoch": 2.4885844748858448,
"grad_norm": 81258.328125,
"learning_rate": 3.4246575342465754e-06,
"loss": 0.0941,
"step": 1090
},
{
"epoch": 2.5114155251141552,
"grad_norm": 707149.125,
"learning_rate": 3.2724505327245055e-06,
"loss": 0.0825,
"step": 1100
},
{
"epoch": 2.5342465753424657,
"grad_norm": 10573.8984375,
"learning_rate": 3.120243531202435e-06,
"loss": 0.0806,
"step": 1110
},
{
"epoch": 2.557077625570776,
"grad_norm": 2018027.5,
"learning_rate": 2.9680365296803653e-06,
"loss": 0.1719,
"step": 1120
},
{
"epoch": 2.5799086757990866,
"grad_norm": 706195.4375,
"learning_rate": 2.8158295281582954e-06,
"loss": 0.0954,
"step": 1130
},
{
"epoch": 2.602739726027397,
"grad_norm": 230678.828125,
"learning_rate": 2.6636225266362255e-06,
"loss": 0.103,
"step": 1140
},
{
"epoch": 2.625570776255708,
"grad_norm": 455807.78125,
"learning_rate": 2.511415525114155e-06,
"loss": 0.0965,
"step": 1150
},
{
"epoch": 2.6484018264840183,
"grad_norm": 38607.32421875,
"learning_rate": 2.3592085235920852e-06,
"loss": 0.0381,
"step": 1160
},
{
"epoch": 2.671232876712329,
"grad_norm": 449129.09375,
"learning_rate": 2.2070015220700153e-06,
"loss": 0.0836,
"step": 1170
},
{
"epoch": 2.6940639269406392,
"grad_norm": 334416.4375,
"learning_rate": 2.0547945205479454e-06,
"loss": 0.0511,
"step": 1180
},
{
"epoch": 2.7168949771689497,
"grad_norm": 490780.5625,
"learning_rate": 1.9025875190258753e-06,
"loss": 0.1427,
"step": 1190
},
{
"epoch": 2.73972602739726,
"grad_norm": 6038.24072265625,
"learning_rate": 1.7503805175038052e-06,
"loss": 0.1099,
"step": 1200
},
{
"epoch": 2.762557077625571,
"grad_norm": 202150.234375,
"learning_rate": 1.5981735159817353e-06,
"loss": 0.119,
"step": 1210
},
{
"epoch": 2.7853881278538815,
"grad_norm": 232674.578125,
"learning_rate": 1.4459665144596652e-06,
"loss": 0.181,
"step": 1220
},
{
"epoch": 2.808219178082192,
"grad_norm": 1792436.625,
"learning_rate": 1.2937595129375953e-06,
"loss": 0.0718,
"step": 1230
},
{
"epoch": 2.8310502283105023,
"grad_norm": 2906556.0,
"learning_rate": 1.1415525114155251e-06,
"loss": 0.097,
"step": 1240
},
{
"epoch": 2.853881278538813,
"grad_norm": 82039.2734375,
"learning_rate": 9.89345509893455e-07,
"loss": 0.1275,
"step": 1250
},
{
"epoch": 2.8767123287671232,
"grad_norm": 271356.25,
"learning_rate": 8.371385083713851e-07,
"loss": 0.1401,
"step": 1260
},
{
"epoch": 2.8995433789954337,
"grad_norm": 9588.64453125,
"learning_rate": 6.849315068493151e-07,
"loss": 0.1627,
"step": 1270
},
{
"epoch": 2.922374429223744,
"grad_norm": 1104557.125,
"learning_rate": 5.327245053272451e-07,
"loss": 0.0942,
"step": 1280
},
{
"epoch": 2.9452054794520546,
"grad_norm": 31101.583984375,
"learning_rate": 3.8051750380517503e-07,
"loss": 0.0714,
"step": 1290
},
{
"epoch": 2.968036529680365,
"grad_norm": 740889.8125,
"learning_rate": 2.2831050228310502e-07,
"loss": 0.1083,
"step": 1300
},
{
"epoch": 2.990867579908676,
"grad_norm": 7040.435546875,
"learning_rate": 7.6103500761035e-08,
"loss": 0.1593,
"step": 1310
},
{
"epoch": 3.0,
"eval_accuracy": 0.9261666666666667,
"eval_f1_macro": 0.9252761556296953,
"eval_f1_weighted": 0.9261245203427795,
"eval_loss": 0.37234973907470703,
"eval_runtime": 15.0862,
"eval_samples_per_second": 397.714,
"eval_steps_per_second": 12.462,
"step": 1314
},
{
"epoch": 3.0,
"step": 1314,
"total_flos": 1537621007523840.0,
"train_loss": 0.18048278951118707,
"train_runtime": 317.0714,
"train_samples_per_second": 132.462,
"train_steps_per_second": 4.144
},
{
"epoch": 3.0,
"eval_accuracy": 0.9285,
"eval_f1_macro": 0.9277850539806721,
"eval_f1_weighted": 0.92853233424208,
"eval_loss": 0.2250971645116806,
"eval_runtime": 14.8809,
"eval_samples_per_second": 403.201,
"eval_steps_per_second": 12.634,
"step": 1314
}
],
"logging_steps": 10,
"max_steps": 1314,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1537621007523840.0,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}