ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5821
  • Qwk: 0.6640
  • Mse: 0.5821
  • Rmse: 0.7629

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0488 2 5.2949 -0.0281 5.2949 2.3011
No log 0.0976 4 3.3291 0.0547 3.3291 1.8246
No log 0.1463 6 2.2332 -0.0667 2.2332 1.4944
No log 0.1951 8 1.8250 0.0701 1.8250 1.3509
No log 0.2439 10 1.3034 0.0749 1.3034 1.1417
No log 0.2927 12 1.2128 0.1868 1.2128 1.1013
No log 0.3415 14 1.6204 0.0821 1.6204 1.2730
No log 0.3902 16 1.8958 0.0821 1.8958 1.3769
No log 0.4390 18 1.6833 0.0698 1.6833 1.2974
No log 0.4878 20 1.4217 0.1450 1.4217 1.1923
No log 0.5366 22 1.1328 0.2983 1.1328 1.0643
No log 0.5854 24 1.0451 0.3658 1.0451 1.0223
No log 0.6341 26 0.9814 0.3852 0.9814 0.9906
No log 0.6829 28 1.1384 0.3542 1.1384 1.0670
No log 0.7317 30 1.7242 0.2411 1.7242 1.3131
No log 0.7805 32 2.3626 0.2034 2.3626 1.5371
No log 0.8293 34 2.4915 0.2071 2.4915 1.5784
No log 0.8780 36 2.1486 0.2265 2.1486 1.4658
No log 0.9268 38 1.4911 0.2191 1.4911 1.2211
No log 0.9756 40 1.0645 0.3987 1.0645 1.0317
No log 1.0244 42 0.7235 0.5587 0.7235 0.8506
No log 1.0732 44 0.7067 0.6049 0.7067 0.8406
No log 1.1220 46 0.7539 0.5303 0.7539 0.8682
No log 1.1707 48 0.8367 0.4964 0.8367 0.9147
No log 1.2195 50 0.8067 0.5442 0.8067 0.8982
No log 1.2683 52 0.6545 0.6418 0.6545 0.8090
No log 1.3171 54 0.6791 0.6566 0.6791 0.8240
No log 1.3659 56 0.9758 0.5134 0.9758 0.9878
No log 1.4146 58 1.4443 0.4265 1.4443 1.2018
No log 1.4634 60 1.3656 0.4225 1.3656 1.1686
No log 1.5122 62 0.9822 0.5706 0.9822 0.9910
No log 1.5610 64 0.6663 0.6559 0.6663 0.8163
No log 1.6098 66 0.6563 0.6729 0.6563 0.8101
No log 1.6585 68 0.6672 0.6660 0.6672 0.8168
No log 1.7073 70 0.8247 0.5096 0.8247 0.9081
No log 1.7561 72 1.1973 0.3683 1.1973 1.0942
No log 1.8049 74 1.2173 0.3283 1.2173 1.1033
No log 1.8537 76 1.0714 0.4008 1.0714 1.0351
No log 1.9024 78 0.7379 0.5712 0.7379 0.8590
No log 1.9512 80 0.6341 0.6799 0.6341 0.7963
No log 2.0 82 0.6004 0.6979 0.6004 0.7749
No log 2.0488 84 0.6349 0.6736 0.6349 0.7968
No log 2.0976 86 0.6896 0.6441 0.6896 0.8304
No log 2.1463 88 0.7522 0.6436 0.7522 0.8673
No log 2.1951 90 0.7137 0.6440 0.7137 0.8448
No log 2.2439 92 0.5807 0.7106 0.5807 0.7620
No log 2.2927 94 0.5715 0.7165 0.5715 0.7560
No log 2.3415 96 0.6002 0.7188 0.6002 0.7747
No log 2.3902 98 0.6714 0.7017 0.6714 0.8194
No log 2.4390 100 0.7528 0.6962 0.7528 0.8676
No log 2.4878 102 0.7265 0.7298 0.7265 0.8523
No log 2.5366 104 0.8431 0.6294 0.8431 0.9182
No log 2.5854 106 1.0448 0.5396 1.0448 1.0222
No log 2.6341 108 0.8484 0.6534 0.8484 0.9211
No log 2.6829 110 0.6948 0.7089 0.6948 0.8335
No log 2.7317 112 0.9052 0.6535 0.9052 0.9514
No log 2.7805 114 0.8307 0.6712 0.8307 0.9114
No log 2.8293 116 0.5932 0.7579 0.5932 0.7702
No log 2.8780 118 0.7218 0.6732 0.7218 0.8496
No log 2.9268 120 0.7083 0.6634 0.7083 0.8416
No log 2.9756 122 0.5784 0.7029 0.5784 0.7605
No log 3.0244 124 0.5988 0.6890 0.5988 0.7738
No log 3.0732 126 0.5751 0.7242 0.5751 0.7584
No log 3.1220 128 0.5791 0.7480 0.5791 0.7610
No log 3.1707 130 0.6005 0.7424 0.6005 0.7749
No log 3.2195 132 0.6123 0.7164 0.6123 0.7825
No log 3.2683 134 0.5800 0.7274 0.5800 0.7616
No log 3.3171 136 0.5835 0.6945 0.5835 0.7638
No log 3.3659 138 0.5663 0.7137 0.5663 0.7525
No log 3.4146 140 0.5669 0.7243 0.5669 0.7529
No log 3.4634 142 0.5844 0.6846 0.5844 0.7644
No log 3.5122 144 0.5864 0.7174 0.5864 0.7658
No log 3.5610 146 0.6352 0.7380 0.6352 0.7970
No log 3.6098 148 0.6456 0.7382 0.6456 0.8035
No log 3.6585 150 0.6405 0.7432 0.6405 0.8003
No log 3.7073 152 0.6360 0.7353 0.6360 0.7975
No log 3.7561 154 0.6082 0.7207 0.6082 0.7799
No log 3.8049 156 0.6072 0.6862 0.6072 0.7792
No log 3.8537 158 0.6247 0.6759 0.6247 0.7904
No log 3.9024 160 0.6044 0.7162 0.6044 0.7775
No log 3.9512 162 0.6297 0.6899 0.6297 0.7936
No log 4.0 164 0.6440 0.7026 0.6440 0.8025
No log 4.0488 166 0.5988 0.7290 0.5988 0.7738
No log 4.0976 168 0.7022 0.6796 0.7022 0.8380
No log 4.1463 170 0.6794 0.6940 0.6794 0.8242
No log 4.1951 172 0.6368 0.7310 0.6368 0.7980
No log 4.2439 174 0.6258 0.7192 0.6258 0.7911
No log 4.2927 176 0.6906 0.7212 0.6906 0.8310
No log 4.3415 178 0.6205 0.7287 0.6205 0.7877
No log 4.3902 180 0.5986 0.7192 0.5986 0.7737
No log 4.4390 182 0.6724 0.7107 0.6724 0.8200
No log 4.4878 184 0.5948 0.6900 0.5948 0.7713
No log 4.5366 186 0.5760 0.6947 0.5760 0.7589
No log 4.5854 188 0.5940 0.7097 0.5940 0.7707
No log 4.6341 190 0.6193 0.6745 0.6193 0.7870
No log 4.6829 192 0.5925 0.7157 0.5925 0.7697
No log 4.7317 194 0.5979 0.6792 0.5979 0.7732
No log 4.7805 196 0.5831 0.7042 0.5831 0.7636
No log 4.8293 198 0.6485 0.7278 0.6485 0.8053
No log 4.8780 200 0.6850 0.7113 0.6850 0.8277
No log 4.9268 202 0.6231 0.7170 0.6231 0.7893
No log 4.9756 204 0.5987 0.7288 0.5987 0.7738
No log 5.0244 206 0.5996 0.7092 0.5996 0.7743
No log 5.0732 208 0.5764 0.7097 0.5764 0.7592
No log 5.1220 210 0.5751 0.7083 0.5751 0.7584
No log 5.1707 212 0.5625 0.6910 0.5625 0.7500
No log 5.2195 214 0.5880 0.6785 0.5880 0.7668
No log 5.2683 216 0.5721 0.6989 0.5721 0.7564
No log 5.3171 218 0.5809 0.6723 0.5809 0.7622
No log 5.3659 220 0.6093 0.6885 0.6093 0.7806
No log 5.4146 222 0.5806 0.7154 0.5806 0.7620
No log 5.4634 224 0.5826 0.6940 0.5826 0.7633
No log 5.5122 226 0.5699 0.7169 0.5699 0.7549
No log 5.5610 228 0.6206 0.7120 0.6206 0.7878
No log 5.6098 230 0.6657 0.7080 0.6657 0.8159
No log 5.6585 232 0.6031 0.7479 0.6031 0.7766
No log 5.7073 234 0.5863 0.7174 0.5863 0.7657
No log 5.7561 236 0.5892 0.7149 0.5892 0.7676
No log 5.8049 238 0.5998 0.6801 0.5998 0.7745
No log 5.8537 240 0.6071 0.6801 0.6071 0.7792
No log 5.9024 242 0.5982 0.7274 0.5982 0.7734
No log 5.9512 244 0.6139 0.7223 0.6139 0.7835
No log 6.0 246 0.5954 0.7258 0.5954 0.7716
No log 6.0488 248 0.6068 0.6963 0.6068 0.7789
No log 6.0976 250 0.6182 0.6651 0.6182 0.7862
No log 6.1463 252 0.6161 0.6967 0.6161 0.7849
No log 6.1951 254 0.6174 0.6831 0.6174 0.7857
No log 6.2439 256 0.6686 0.6412 0.6686 0.8177
No log 6.2927 258 0.6234 0.6812 0.6234 0.7895
No log 6.3415 260 0.6052 0.7264 0.6052 0.7780
No log 6.3902 262 0.6173 0.7197 0.6173 0.7857
No log 6.4390 264 0.7195 0.6712 0.7195 0.8482
No log 6.4878 266 0.8032 0.6415 0.8032 0.8962
No log 6.5366 268 0.6697 0.6823 0.6697 0.8183
No log 6.5854 270 0.5913 0.7140 0.5913 0.7689
No log 6.6341 272 0.6216 0.6611 0.6216 0.7884
No log 6.6829 274 0.6066 0.6820 0.6066 0.7789
No log 6.7317 276 0.6041 0.6925 0.6041 0.7772
No log 6.7805 278 0.6102 0.6714 0.6102 0.7811
No log 6.8293 280 0.6142 0.6866 0.6142 0.7837
No log 6.8780 282 0.6140 0.6682 0.6140 0.7836
No log 6.9268 284 0.6328 0.6581 0.6328 0.7955
No log 6.9756 286 0.6360 0.6614 0.6360 0.7975
No log 7.0244 288 0.6251 0.7123 0.6251 0.7906
No log 7.0732 290 0.6761 0.7040 0.6761 0.8223
No log 7.1220 292 0.6572 0.6959 0.6572 0.8107
No log 7.1707 294 0.6244 0.7009 0.6244 0.7902
No log 7.2195 296 0.6089 0.6990 0.6089 0.7803
No log 7.2683 298 0.7300 0.6294 0.7300 0.8544
No log 7.3171 300 0.7874 0.5928 0.7874 0.8874
No log 7.3659 302 0.6712 0.6471 0.6712 0.8193
No log 7.4146 304 0.6155 0.6649 0.6155 0.7845
No log 7.4634 306 0.6196 0.6657 0.6196 0.7872
No log 7.5122 308 0.6077 0.6578 0.6077 0.7795
No log 7.5610 310 0.5976 0.6757 0.5976 0.7730
No log 7.6098 312 0.5972 0.6757 0.5972 0.7728
No log 7.6585 314 0.5949 0.7319 0.5949 0.7713
No log 7.7073 316 0.6106 0.7066 0.6106 0.7814
No log 7.7561 318 0.6201 0.7072 0.6201 0.7875
No log 7.8049 320 0.5935 0.6984 0.5935 0.7704
No log 7.8537 322 0.6038 0.6682 0.6038 0.7770
No log 7.9024 324 0.6010 0.7200 0.6010 0.7752
No log 7.9512 326 0.6134 0.7045 0.6134 0.7832
No log 8.0 328 0.6430 0.7078 0.6430 0.8018
No log 8.0488 330 0.7472 0.6866 0.7472 0.8644
No log 8.0976 332 0.6772 0.6742 0.6772 0.8229
No log 8.1463 334 0.6080 0.6921 0.6080 0.7798
No log 8.1951 336 0.6116 0.6640 0.6116 0.7820
No log 8.2439 338 0.6093 0.6639 0.6093 0.7806
No log 8.2927 340 0.6223 0.6715 0.6223 0.7889
No log 8.3415 342 0.6182 0.6694 0.6182 0.7862
No log 8.3902 344 0.6057 0.6750 0.6057 0.7783
No log 8.4390 346 0.6049 0.6596 0.6049 0.7778
No log 8.4878 348 0.6158 0.6821 0.6158 0.7847
No log 8.5366 350 0.5752 0.6891 0.5752 0.7584
No log 8.5854 352 0.6018 0.6962 0.6018 0.7758
No log 8.6341 354 0.6973 0.6632 0.6973 0.8350
No log 8.6829 356 0.6685 0.6820 0.6685 0.8176
No log 8.7317 358 0.6270 0.6842 0.6270 0.7918
No log 8.7805 360 0.5705 0.7243 0.5705 0.7553
No log 8.8293 362 0.6321 0.6915 0.6321 0.7951
No log 8.8780 364 0.6278 0.6816 0.6278 0.7923
No log 8.9268 366 0.5868 0.6852 0.5868 0.7661
No log 8.9756 368 0.6275 0.6833 0.6275 0.7922
No log 9.0244 370 0.6679 0.6727 0.6679 0.8172
No log 9.0732 372 0.6076 0.7004 0.6076 0.7795
No log 9.1220 374 0.5734 0.7389 0.5734 0.7573
No log 9.1707 376 0.5910 0.7024 0.5910 0.7688
No log 9.2195 378 0.6474 0.7036 0.6474 0.8046
No log 9.2683 380 0.6814 0.7064 0.6814 0.8255
No log 9.3171 382 0.6038 0.7296 0.6038 0.7770
No log 9.3659 384 0.7024 0.6911 0.7024 0.8381
No log 9.4146 386 0.8627 0.6532 0.8627 0.9288
No log 9.4634 388 0.8187 0.6540 0.8187 0.9048
No log 9.5122 390 0.7076 0.6721 0.7076 0.8412
No log 9.5610 392 0.5976 0.7320 0.5976 0.7730
No log 9.6098 394 0.5628 0.7259 0.5628 0.7502
No log 9.6585 396 0.5585 0.7494 0.5585 0.7473
No log 9.7073 398 0.5730 0.7336 0.5730 0.7570
No log 9.7561 400 0.5826 0.7280 0.5826 0.7633
No log 9.8049 402 0.5574 0.7414 0.5574 0.7466
No log 9.8537 404 0.5737 0.7464 0.5737 0.7574
No log 9.9024 406 0.6792 0.7416 0.6792 0.8241
No log 9.9512 408 0.6632 0.7391 0.6632 0.8144
No log 10.0 410 0.5777 0.7649 0.5777 0.7601
No log 10.0488 412 0.6381 0.7020 0.6381 0.7988
No log 10.0976 414 0.7777 0.6181 0.7777 0.8819
No log 10.1463 416 0.7458 0.6172 0.7458 0.8636
No log 10.1951 418 0.6262 0.6752 0.6262 0.7913
No log 10.2439 420 0.6231 0.6957 0.6231 0.7893
No log 10.2927 422 0.6503 0.6793 0.6503 0.8064
No log 10.3415 424 0.6120 0.6950 0.6120 0.7823
No log 10.3902 426 0.6111 0.7114 0.6111 0.7817
No log 10.4390 428 0.7042 0.7025 0.7042 0.8392
No log 10.4878 430 0.6745 0.7120 0.6745 0.8213
No log 10.5366 432 0.6198 0.7228 0.6198 0.7873
No log 10.5854 434 0.6136 0.7203 0.6136 0.7833
No log 10.6341 436 0.6132 0.7347 0.6132 0.7831
No log 10.6829 438 0.6218 0.6794 0.6218 0.7886
No log 10.7317 440 0.6288 0.6799 0.6288 0.7930
No log 10.7805 442 0.6287 0.6680 0.6287 0.7929
No log 10.8293 444 0.6237 0.6680 0.6237 0.7898
No log 10.8780 446 0.6122 0.7180 0.6122 0.7824
No log 10.9268 448 0.6274 0.6598 0.6274 0.7921
No log 10.9756 450 0.6426 0.6482 0.6426 0.8017
No log 11.0244 452 0.6013 0.7052 0.6013 0.7754
No log 11.0732 454 0.5932 0.7124 0.5932 0.7702
No log 11.1220 456 0.6211 0.6518 0.6211 0.7881
No log 11.1707 458 0.6080 0.6921 0.6080 0.7797
No log 11.2195 460 0.5897 0.7208 0.5897 0.7679
No log 11.2683 462 0.6415 0.6664 0.6415 0.8010
No log 11.3171 464 0.7156 0.6086 0.7156 0.8459
No log 11.3659 466 0.6786 0.6260 0.6786 0.8238
No log 11.4146 468 0.6108 0.7143 0.6108 0.7816
No log 11.4634 470 0.6052 0.7105 0.6052 0.7779
No log 11.5122 472 0.6033 0.7433 0.6033 0.7767
No log 11.5610 474 0.6036 0.7078 0.6036 0.7769
No log 11.6098 476 0.6994 0.6418 0.6994 0.8363
No log 11.6585 478 0.7407 0.6238 0.7407 0.8607
No log 11.7073 480 0.6811 0.6751 0.6811 0.8253
No log 11.7561 482 0.6334 0.6725 0.6334 0.7958
No log 11.8049 484 0.6365 0.6352 0.6365 0.7978
No log 11.8537 486 0.6297 0.6852 0.6297 0.7935
No log 11.9024 488 0.6578 0.6747 0.6578 0.8110
No log 11.9512 490 0.7490 0.6465 0.7490 0.8655
No log 12.0 492 0.7266 0.6496 0.7266 0.8524
No log 12.0488 494 0.6397 0.6823 0.6397 0.7998
No log 12.0976 496 0.6104 0.6797 0.6104 0.7813
No log 12.1463 498 0.6140 0.6548 0.6140 0.7836
0.3833 12.1951 500 0.6004 0.6658 0.6004 0.7748
0.3833 12.2439 502 0.5791 0.7275 0.5791 0.7610
0.3833 12.2927 504 0.6199 0.7576 0.6199 0.7873
0.3833 12.3415 506 0.6802 0.7015 0.6802 0.8247
0.3833 12.3902 508 0.6921 0.7032 0.6921 0.8319
0.3833 12.4390 510 0.6252 0.7032 0.6252 0.7907
0.3833 12.4878 512 0.5907 0.7291 0.5907 0.7686
0.3833 12.5366 514 0.6168 0.6890 0.6168 0.7853
0.3833 12.5854 516 0.6216 0.6710 0.6216 0.7884
0.3833 12.6341 518 0.6298 0.6561 0.6298 0.7936
0.3833 12.6829 520 0.6831 0.6323 0.6831 0.8265
0.3833 12.7317 522 0.6697 0.6418 0.6697 0.8184
0.3833 12.7805 524 0.6253 0.6594 0.6253 0.7907
0.3833 12.8293 526 0.6167 0.7222 0.6167 0.7853
0.3833 12.8780 528 0.6116 0.7406 0.6116 0.7821
0.3833 12.9268 530 0.6013 0.7282 0.6013 0.7754
0.3833 12.9756 532 0.6192 0.7204 0.6192 0.7869
0.3833 13.0244 534 0.7202 0.6459 0.7202 0.8486
0.3833 13.0732 536 0.8016 0.6295 0.8016 0.8953
0.3833 13.1220 538 0.7451 0.6115 0.7451 0.8632
0.3833 13.1707 540 0.6763 0.6542 0.6763 0.8224
0.3833 13.2195 542 0.6674 0.6685 0.6674 0.8169
0.3833 13.2683 544 0.6667 0.6603 0.6667 0.8165
0.3833 13.3171 546 0.6830 0.6361 0.6830 0.8264
0.3833 13.3659 548 0.6346 0.6715 0.6346 0.7966
0.3833 13.4146 550 0.6114 0.6803 0.6114 0.7819
0.3833 13.4634 552 0.6128 0.7027 0.6128 0.7828
0.3833 13.5122 554 0.6414 0.6847 0.6414 0.8009
0.3833 13.5610 556 0.6668 0.6701 0.6668 0.8166
0.3833 13.6098 558 0.6853 0.6778 0.6853 0.8278
0.3833 13.6585 560 0.7007 0.6778 0.7007 0.8371
0.3833 13.7073 562 0.6574 0.6856 0.6574 0.8108
0.3833 13.7561 564 0.6057 0.7360 0.6057 0.7783
0.3833 13.8049 566 0.5985 0.7477 0.5985 0.7736
0.3833 13.8537 568 0.5887 0.7436 0.5887 0.7672
0.3833 13.9024 570 0.6084 0.6806 0.6084 0.7800
0.3833 13.9512 572 0.6901 0.6418 0.6901 0.8307
0.3833 14.0 574 0.7425 0.6251 0.7425 0.8617
0.3833 14.0488 576 0.6945 0.6366 0.6945 0.8334
0.3833 14.0976 578 0.6357 0.6444 0.6357 0.7973
0.3833 14.1463 580 0.5970 0.7451 0.5970 0.7726
0.3833 14.1951 582 0.6022 0.7416 0.6022 0.7760
0.3833 14.2439 584 0.6024 0.7491 0.6024 0.7761
0.3833 14.2927 586 0.6005 0.7416 0.6005 0.7749
0.3833 14.3415 588 0.6171 0.7090 0.6171 0.7855
0.3833 14.3902 590 0.6802 0.6260 0.6802 0.8247
0.3833 14.4390 592 0.7783 0.6251 0.7783 0.8822
0.3833 14.4878 594 0.7751 0.6113 0.7751 0.8804
0.3833 14.5366 596 0.6709 0.6411 0.6709 0.8191
0.3833 14.5854 598 0.5821 0.6640 0.5821 0.7629

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k8_task1_organization

Finetuned
(4023)
this model