ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k2_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7782
  • Qwk: 0.5763
  • Mse: 0.7782
  • Rmse: 0.8821

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 3.9433 -0.0150 3.9433 1.9858
No log 0.3077 4 2.0174 0.0329 2.0174 1.4204
No log 0.4615 6 1.2015 0.1324 1.2015 1.0961
No log 0.6154 8 1.2318 0.0363 1.2318 1.1099
No log 0.7692 10 1.1835 0.1810 1.1835 1.0879
No log 0.9231 12 1.1415 0.1810 1.1415 1.0684
No log 1.0769 14 1.1009 0.1189 1.1009 1.0493
No log 1.2308 16 1.2798 0.0694 1.2798 1.1313
No log 1.3846 18 1.6053 0.1394 1.6053 1.2670
No log 1.5385 20 1.5745 0.1394 1.5745 1.2548
No log 1.6923 22 1.1129 0.1725 1.1129 1.0549
No log 1.8462 24 0.8077 0.3392 0.8077 0.8987
No log 2.0 26 0.7931 0.4697 0.7931 0.8905
No log 2.1538 28 0.7373 0.4527 0.7373 0.8586
No log 2.3077 30 0.8641 0.3954 0.8641 0.9296
No log 2.4615 32 1.2770 0.1002 1.2770 1.1300
No log 2.6154 34 1.3255 0.1241 1.3255 1.1513
No log 2.7692 36 1.0508 0.2372 1.0508 1.0251
No log 2.9231 38 0.7311 0.4838 0.7311 0.8550
No log 3.0769 40 0.7110 0.5120 0.7110 0.8432
No log 3.2308 42 0.7404 0.4482 0.7404 0.8604
No log 3.3846 44 0.7525 0.5430 0.7525 0.8675
No log 3.5385 46 0.7089 0.5742 0.7089 0.8419
No log 3.6923 48 0.7314 0.6363 0.7314 0.8552
No log 3.8462 50 0.8740 0.5672 0.8740 0.9349
No log 4.0 52 1.2555 0.4148 1.2555 1.1205
No log 4.1538 54 1.2750 0.4005 1.2750 1.1292
No log 4.3077 56 0.8023 0.5899 0.8023 0.8957
No log 4.4615 58 0.6830 0.6440 0.6830 0.8264
No log 4.6154 60 0.6699 0.6350 0.6699 0.8184
No log 4.7692 62 0.6799 0.6945 0.6799 0.8245
No log 4.9231 64 0.7310 0.6065 0.7310 0.8550
No log 5.0769 66 0.8261 0.5927 0.8261 0.9089
No log 5.2308 68 0.7537 0.6517 0.7537 0.8681
No log 5.3846 70 0.7817 0.5360 0.7817 0.8841
No log 5.5385 72 0.8166 0.5361 0.8166 0.9037
No log 5.6923 74 0.7814 0.5413 0.7814 0.8840
No log 5.8462 76 0.7626 0.5813 0.7626 0.8732
No log 6.0 78 0.8204 0.5763 0.8204 0.9058
No log 6.1538 80 0.8326 0.5943 0.8326 0.9125
No log 6.3077 82 0.8071 0.6159 0.8071 0.8984
No log 6.4615 84 0.8315 0.6019 0.8315 0.9118
No log 6.6154 86 0.7688 0.6043 0.7688 0.8768
No log 6.7692 88 0.7447 0.6358 0.7447 0.8630
No log 6.9231 90 0.8031 0.5847 0.8031 0.8961
No log 7.0769 92 0.9035 0.5347 0.9035 0.9505
No log 7.2308 94 1.2263 0.4206 1.2263 1.1074
No log 7.3846 96 1.1926 0.4284 1.1926 1.0921
No log 7.5385 98 0.7808 0.6029 0.7808 0.8836
No log 7.6923 100 0.6620 0.5614 0.6620 0.8136
No log 7.8462 102 0.6996 0.5766 0.6996 0.8364
No log 8.0 104 0.7110 0.6417 0.7110 0.8432
No log 8.1538 106 0.8076 0.5852 0.8076 0.8987
No log 8.3077 108 0.8782 0.5221 0.8782 0.9371
No log 8.4615 110 0.7980 0.5847 0.7980 0.8933
No log 8.6154 112 0.7809 0.5756 0.7809 0.8837
No log 8.7692 114 0.7276 0.5756 0.7276 0.8530
No log 8.9231 116 0.6872 0.5250 0.6872 0.8290
No log 9.0769 118 0.7082 0.5759 0.7082 0.8415
No log 9.2308 120 0.7955 0.5805 0.7955 0.8919
No log 9.3846 122 0.8134 0.5793 0.8134 0.9019
No log 9.5385 124 0.7143 0.6358 0.7143 0.8451
No log 9.6923 126 0.7087 0.5773 0.7087 0.8418
No log 9.8462 128 0.7387 0.6339 0.7387 0.8595
No log 10.0 130 0.8070 0.6135 0.8070 0.8983
No log 10.1538 132 0.7933 0.6318 0.7933 0.8907
No log 10.3077 134 0.7865 0.6288 0.7865 0.8868
No log 10.4615 136 0.8061 0.5749 0.8061 0.8979
No log 10.6154 138 0.7493 0.5578 0.7493 0.8656
No log 10.7692 140 0.6618 0.5542 0.6618 0.8135
No log 10.9231 142 0.6590 0.5542 0.6590 0.8118
No log 11.0769 144 0.7453 0.5997 0.7453 0.8633
No log 11.2308 146 0.9885 0.5183 0.9885 0.9942
No log 11.3846 148 0.9470 0.5174 0.9470 0.9731
No log 11.5385 150 0.7086 0.6578 0.7086 0.8418
No log 11.6923 152 0.6808 0.6102 0.6808 0.8251
No log 11.8462 154 0.6816 0.5752 0.6816 0.8256
No log 12.0 156 0.6956 0.5969 0.6956 0.8340
No log 12.1538 158 0.7399 0.6482 0.7399 0.8602
No log 12.3077 160 0.7906 0.5710 0.7906 0.8891
No log 12.4615 162 0.7185 0.5817 0.7185 0.8476
No log 12.6154 164 0.6874 0.5902 0.6874 0.8291
No log 12.7692 166 0.6700 0.5902 0.6700 0.8185
No log 12.9231 168 0.7171 0.5737 0.7171 0.8468
No log 13.0769 170 0.8690 0.5595 0.8690 0.9322
No log 13.2308 172 0.8605 0.5148 0.8605 0.9276
No log 13.3846 174 0.7822 0.5818 0.7822 0.8844
No log 13.5385 176 0.6976 0.5704 0.6976 0.8352
No log 13.6923 178 0.7052 0.6075 0.7052 0.8397
No log 13.8462 180 0.7113 0.5702 0.7113 0.8434
No log 14.0 182 0.7404 0.6166 0.7404 0.8605
No log 14.1538 184 0.7830 0.5994 0.7830 0.8849
No log 14.3077 186 0.7876 0.6321 0.7876 0.8875
No log 14.4615 188 0.7138 0.5806 0.7138 0.8448
No log 14.6154 190 0.6919 0.5393 0.6919 0.8318
No log 14.7692 192 0.7369 0.5554 0.7369 0.8584
No log 14.9231 194 0.7852 0.5455 0.7852 0.8861
No log 15.0769 196 0.7573 0.6164 0.7573 0.8702
No log 15.2308 198 0.7217 0.6125 0.7217 0.8495
No log 15.3846 200 0.7241 0.6125 0.7241 0.8509
No log 15.5385 202 0.7027 0.6042 0.7027 0.8383
No log 15.6923 204 0.6833 0.5409 0.6833 0.8266
No log 15.8462 206 0.7123 0.5542 0.7123 0.8440
No log 16.0 208 0.7498 0.5566 0.7498 0.8659
No log 16.1538 210 0.7382 0.5966 0.7382 0.8592
No log 16.3077 212 0.7130 0.5841 0.7130 0.8444
No log 16.4615 214 0.6887 0.6058 0.6887 0.8299
No log 16.6154 216 0.6957 0.5662 0.6957 0.8341
No log 16.7692 218 0.7091 0.6215 0.7091 0.8421
No log 16.9231 220 0.7069 0.6215 0.7069 0.8408
No log 17.0769 222 0.7149 0.6535 0.7149 0.8455
No log 17.2308 224 0.7104 0.6590 0.7104 0.8429
No log 17.3846 226 0.7050 0.6377 0.7050 0.8397
No log 17.5385 228 0.7072 0.5839 0.7072 0.8410
No log 17.6923 230 0.6750 0.5421 0.6750 0.8216
No log 17.8462 232 0.6682 0.5421 0.6682 0.8174
No log 18.0 234 0.6610 0.5421 0.6610 0.8130
No log 18.1538 236 0.6795 0.5313 0.6795 0.8243
No log 18.3077 238 0.6783 0.5530 0.6783 0.8236
No log 18.4615 240 0.6884 0.6025 0.6884 0.8297
No log 18.6154 242 0.6651 0.6195 0.6651 0.8156
No log 18.7692 244 0.7292 0.5895 0.7292 0.8539
No log 18.9231 246 0.7620 0.6295 0.7620 0.8729
No log 19.0769 248 0.7041 0.5629 0.7041 0.8391
No log 19.2308 250 0.6603 0.6015 0.6603 0.8126
No log 19.3846 252 0.6297 0.5830 0.6297 0.7935
No log 19.5385 254 0.6361 0.6022 0.6361 0.7976
No log 19.6923 256 0.6713 0.6516 0.6713 0.8193
No log 19.8462 258 0.6934 0.6307 0.6934 0.8327
No log 20.0 260 0.6670 0.5984 0.6670 0.8167
No log 20.1538 262 0.6488 0.6602 0.6488 0.8055
No log 20.3077 264 0.6403 0.6275 0.6403 0.8002
No log 20.4615 266 0.6457 0.6392 0.6457 0.8035
No log 20.6154 268 0.6459 0.6850 0.6459 0.8037
No log 20.7692 270 0.6897 0.6249 0.6897 0.8305
No log 20.9231 272 0.7448 0.5821 0.7448 0.8630
No log 21.0769 274 0.7694 0.5810 0.7694 0.8771
No log 21.2308 276 0.7255 0.6599 0.7255 0.8517
No log 21.3846 278 0.6646 0.6854 0.6646 0.8153
No log 21.5385 280 0.6304 0.6918 0.6304 0.7940
No log 21.6923 282 0.6318 0.7196 0.6318 0.7948
No log 21.8462 284 0.6282 0.7061 0.6282 0.7926
No log 22.0 286 0.6560 0.6851 0.6560 0.8100
No log 22.1538 288 0.6443 0.6967 0.6443 0.8027
No log 22.3077 290 0.6430 0.7118 0.6430 0.8019
No log 22.4615 292 0.6785 0.6802 0.6785 0.8237
No log 22.6154 294 0.6869 0.6404 0.6869 0.8288
No log 22.7692 296 0.6647 0.6404 0.6647 0.8153
No log 22.9231 298 0.6024 0.6756 0.6024 0.7762
No log 23.0769 300 0.6089 0.7440 0.6089 0.7803
No log 23.2308 302 0.6373 0.7440 0.6373 0.7983
No log 23.3846 304 0.6757 0.6299 0.6757 0.8220
No log 23.5385 306 0.6649 0.6922 0.6649 0.8154
No log 23.6923 308 0.6424 0.6598 0.6424 0.8015
No log 23.8462 310 0.6341 0.6717 0.6341 0.7963
No log 24.0 312 0.6136 0.6249 0.6136 0.7833
No log 24.1538 314 0.6086 0.6039 0.6086 0.7801
No log 24.3077 316 0.6035 0.6237 0.6035 0.7769
No log 24.4615 318 0.6137 0.6176 0.6137 0.7834
No log 24.6154 320 0.6437 0.6697 0.6437 0.8023
No log 24.7692 322 0.6394 0.6697 0.6394 0.7996
No log 24.9231 324 0.6339 0.6697 0.6339 0.7962
No log 25.0769 326 0.6557 0.6873 0.6557 0.8098
No log 25.2308 328 0.6375 0.6411 0.6375 0.7985
No log 25.3846 330 0.6226 0.6080 0.6226 0.7890
No log 25.5385 332 0.6033 0.5988 0.6033 0.7767
No log 25.6923 334 0.5922 0.6067 0.5922 0.7696
No log 25.8462 336 0.5972 0.5659 0.5972 0.7728
No log 26.0 338 0.6141 0.6262 0.6141 0.7837
No log 26.1538 340 0.6488 0.6183 0.6488 0.8055
No log 26.3077 342 0.6191 0.6058 0.6191 0.7869
No log 26.4615 344 0.5933 0.6207 0.5933 0.7703
No log 26.6154 346 0.5934 0.6249 0.5934 0.7703
No log 26.7692 348 0.5993 0.6249 0.5993 0.7741
No log 26.9231 350 0.6096 0.6770 0.6096 0.7808
No log 27.0769 352 0.6290 0.6164 0.6290 0.7931
No log 27.2308 354 0.6664 0.6144 0.6664 0.8163
No log 27.3846 356 0.6794 0.6144 0.6794 0.8242
No log 27.5385 358 0.6455 0.6164 0.6455 0.8034
No log 27.6923 360 0.6224 0.6584 0.6224 0.7889
No log 27.8462 362 0.6229 0.6297 0.6229 0.7892
No log 28.0 364 0.6181 0.6297 0.6181 0.7862
No log 28.1538 366 0.6053 0.6442 0.6053 0.7780
No log 28.3077 368 0.6196 0.6675 0.6196 0.7872
No log 28.4615 370 0.6607 0.6567 0.6607 0.8128
No log 28.6154 372 0.6605 0.6918 0.6605 0.8127
No log 28.7692 374 0.6466 0.6918 0.6466 0.8041
No log 28.9231 376 0.6233 0.6952 0.6233 0.7895
No log 29.0769 378 0.6242 0.6952 0.6242 0.7900
No log 29.2308 380 0.6288 0.6918 0.6288 0.7930
No log 29.3846 382 0.6225 0.6756 0.6225 0.7890
No log 29.5385 384 0.6088 0.6426 0.6088 0.7802
No log 29.6923 386 0.6141 0.6606 0.6141 0.7836
No log 29.8462 388 0.6260 0.6578 0.6260 0.7912
No log 30.0 390 0.6510 0.6756 0.6510 0.8068
No log 30.1538 392 0.7362 0.6054 0.7362 0.8580
No log 30.3077 394 0.7826 0.6151 0.7826 0.8846
No log 30.4615 396 0.7424 0.6054 0.7424 0.8616
No log 30.6154 398 0.6581 0.6756 0.6581 0.8112
No log 30.7692 400 0.6126 0.6426 0.6126 0.7827
No log 30.9231 402 0.5954 0.5831 0.5954 0.7716
No log 31.0769 404 0.5952 0.5831 0.5952 0.7715
No log 31.2308 406 0.5819 0.5831 0.5819 0.7629
No log 31.3846 408 0.5833 0.5950 0.5833 0.7638
No log 31.5385 410 0.5758 0.6461 0.5758 0.7588
No log 31.6923 412 0.5768 0.6572 0.5768 0.7595
No log 31.8462 414 0.5981 0.7352 0.5981 0.7734
No log 32.0 416 0.6563 0.6818 0.6563 0.8101
No log 32.1538 418 0.6408 0.6818 0.6408 0.8005
No log 32.3077 420 0.5834 0.7092 0.5834 0.7638
No log 32.4615 422 0.5753 0.6926 0.5753 0.7585
No log 32.6154 424 0.5701 0.6684 0.5701 0.7550
No log 32.7692 426 0.5778 0.6884 0.5778 0.7601
No log 32.9231 428 0.5996 0.6758 0.5996 0.7743
No log 33.0769 430 0.6269 0.6730 0.6269 0.7918
No log 33.2308 432 0.6672 0.7193 0.6672 0.8168
No log 33.3846 434 0.6885 0.6720 0.6885 0.8298
No log 33.5385 436 0.6665 0.6973 0.6665 0.8164
No log 33.6923 438 0.6213 0.6730 0.6213 0.7882
No log 33.8462 440 0.5952 0.6795 0.5952 0.7715
No log 34.0 442 0.5800 0.6952 0.5800 0.7616
No log 34.1538 444 0.5587 0.6644 0.5587 0.7475
No log 34.3077 446 0.5586 0.6078 0.5586 0.7474
No log 34.4615 448 0.5684 0.6438 0.5684 0.7539
No log 34.6154 450 0.5987 0.6811 0.5987 0.7737
No log 34.7692 452 0.5991 0.6885 0.5991 0.7740
No log 34.9231 454 0.5734 0.6926 0.5734 0.7572
No log 35.0769 456 0.5592 0.6644 0.5592 0.7478
No log 35.2308 458 0.5710 0.6644 0.5710 0.7556
No log 35.3846 460 0.5781 0.6988 0.5781 0.7603
No log 35.5385 462 0.6086 0.6795 0.6086 0.7801
No log 35.6923 464 0.6670 0.6838 0.6670 0.8167
No log 35.8462 466 0.6981 0.6727 0.6981 0.8355
No log 36.0 468 0.6641 0.6727 0.6641 0.8149
No log 36.1538 470 0.6010 0.6765 0.6010 0.7752
No log 36.3077 472 0.5742 0.6884 0.5742 0.7577
No log 36.4615 474 0.5729 0.6884 0.5729 0.7569
No log 36.6154 476 0.5885 0.7138 0.5885 0.7671
No log 36.7692 478 0.5760 0.6884 0.5760 0.7590
No log 36.9231 480 0.5658 0.6606 0.5658 0.7522
No log 37.0769 482 0.5701 0.6606 0.5701 0.7551
No log 37.2308 484 0.5928 0.6788 0.5928 0.7699
No log 37.3846 486 0.6213 0.7099 0.6213 0.7882
No log 37.5385 488 0.6668 0.6838 0.6668 0.8166
No log 37.6923 490 0.6631 0.7052 0.6631 0.8143
No log 37.8462 492 0.6305 0.7005 0.6305 0.7940
No log 38.0 494 0.5932 0.6689 0.5932 0.7702
No log 38.1538 496 0.5892 0.7026 0.5892 0.7676
No log 38.3077 498 0.5901 0.7026 0.5901 0.7682
0.2135 38.4615 500 0.6005 0.7026 0.6005 0.7750
0.2135 38.6154 502 0.6283 0.7005 0.6283 0.7927
0.2135 38.7692 504 0.7101 0.6258 0.7101 0.8427
0.2135 38.9231 506 0.8032 0.5599 0.8032 0.8962
0.2135 39.0769 508 0.8233 0.5599 0.8233 0.9074
0.2135 39.2308 510 0.7782 0.5763 0.7782 0.8821

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k2_task5_organization

Finetuned
(4019)
this model