ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k1_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5882
  • Qwk: 0.5880
  • Mse: 0.5882
  • Rmse: 0.7669

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.4 2 3.8579 -0.0092 3.8579 1.9642
No log 0.8 4 2.4617 0.0155 2.4617 1.5690
No log 1.2 6 1.5181 -0.0219 1.5181 1.2321
No log 1.6 8 1.0182 0.2818 1.0182 1.0090
No log 2.0 10 0.9859 0.2035 0.9859 0.9929
No log 2.4 12 0.9606 0.2365 0.9606 0.9801
No log 2.8 14 0.9664 0.3304 0.9664 0.9830
No log 3.2 16 1.0047 0.3795 1.0047 1.0023
No log 3.6 18 0.9898 0.2991 0.9898 0.9949
No log 4.0 20 0.9554 0.3709 0.9554 0.9774
No log 4.4 22 0.9239 0.3770 0.9239 0.9612
No log 4.8 24 0.8888 0.4167 0.8888 0.9427
No log 5.2 26 0.8056 0.4275 0.8056 0.8975
No log 5.6 28 0.7769 0.4365 0.7769 0.8814
No log 6.0 30 0.7329 0.4365 0.7329 0.8561
No log 6.4 32 0.7060 0.5197 0.7060 0.8402
No log 6.8 34 0.7266 0.4853 0.7266 0.8524
No log 7.2 36 0.6731 0.4864 0.6731 0.8204
No log 7.6 38 0.6504 0.6390 0.6504 0.8065
No log 8.0 40 0.7812 0.56 0.7812 0.8839
No log 8.4 42 0.8054 0.5480 0.8054 0.8975
No log 8.8 44 0.6282 0.6716 0.6282 0.7926
No log 9.2 46 0.6105 0.5879 0.6105 0.7813
No log 9.6 48 0.6198 0.6383 0.6198 0.7873
No log 10.0 50 0.5686 0.6455 0.5686 0.7540
No log 10.4 52 0.5801 0.6617 0.5801 0.7616
No log 10.8 54 0.6279 0.7131 0.6279 0.7924
No log 11.2 56 0.5569 0.6822 0.5569 0.7463
No log 11.6 58 0.5423 0.6439 0.5423 0.7364
No log 12.0 60 0.5448 0.6966 0.5448 0.7381
No log 12.4 62 0.5577 0.6349 0.5577 0.7468
No log 12.8 64 0.5394 0.7026 0.5394 0.7344
No log 13.2 66 0.6213 0.6301 0.6213 0.7882
No log 13.6 68 0.6854 0.6181 0.6854 0.8279
No log 14.0 70 0.5732 0.6340 0.5732 0.7571
No log 14.4 72 0.5365 0.6729 0.5365 0.7325
No log 14.8 74 0.5385 0.6619 0.5385 0.7338
No log 15.2 76 0.5468 0.6493 0.5468 0.7395
No log 15.6 78 0.5763 0.6245 0.5763 0.7591
No log 16.0 80 0.5840 0.6245 0.5840 0.7642
No log 16.4 82 0.5637 0.6649 0.5637 0.7508
No log 16.8 84 0.5981 0.6368 0.5981 0.7733
No log 17.2 86 0.5673 0.6144 0.5673 0.7532
No log 17.6 88 0.5799 0.6357 0.5799 0.7615
No log 18.0 90 0.6525 0.6746 0.6525 0.8077
No log 18.4 92 0.6561 0.6269 0.6561 0.8100
No log 18.8 94 0.6204 0.6269 0.6204 0.7876
No log 19.2 96 0.6275 0.5771 0.6275 0.7921
No log 19.6 98 0.5988 0.6018 0.5988 0.7738
No log 20.0 100 0.5772 0.6239 0.5772 0.7597
No log 20.4 102 0.5845 0.6345 0.5845 0.7645
No log 20.8 104 0.5877 0.6345 0.5877 0.7666
No log 21.2 106 0.5811 0.6123 0.5811 0.7623
No log 21.6 108 0.5860 0.6485 0.5860 0.7655
No log 22.0 110 0.5932 0.6388 0.5932 0.7702
No log 22.4 112 0.5838 0.6075 0.5838 0.7640
No log 22.8 114 0.6387 0.6244 0.6387 0.7992
No log 23.2 116 0.6076 0.6729 0.6076 0.7795
No log 23.6 118 0.6017 0.6451 0.6017 0.7757
No log 24.0 120 0.6333 0.6727 0.6333 0.7958
No log 24.4 122 0.5906 0.6511 0.5906 0.7685
No log 24.8 124 0.5772 0.6144 0.5772 0.7598
No log 25.2 126 0.6268 0.5902 0.6268 0.7917
No log 25.6 128 0.5902 0.6644 0.5902 0.7683
No log 26.0 130 0.5681 0.6460 0.5681 0.7537
No log 26.4 132 0.6230 0.6520 0.6230 0.7893
No log 26.8 134 0.6167 0.6240 0.6167 0.7853
No log 27.2 136 0.5685 0.6460 0.5685 0.7540
No log 27.6 138 0.6025 0.5879 0.6025 0.7762
No log 28.0 140 0.6015 0.5905 0.6015 0.7756
No log 28.4 142 0.5855 0.6207 0.5855 0.7652
No log 28.8 144 0.5833 0.6491 0.5833 0.7638
No log 29.2 146 0.6345 0.6395 0.6345 0.7966
No log 29.6 148 0.6430 0.6564 0.6430 0.8018
No log 30.0 150 0.5961 0.6555 0.5961 0.7721
No log 30.4 152 0.5821 0.5856 0.5821 0.7629
No log 30.8 154 0.6053 0.5890 0.6053 0.7780
No log 31.2 156 0.5787 0.6097 0.5787 0.7607
No log 31.6 158 0.5669 0.6123 0.5669 0.7529
No log 32.0 160 0.5941 0.6647 0.5941 0.7708
No log 32.4 162 0.6388 0.6520 0.6388 0.7993
No log 32.8 164 0.6240 0.6319 0.6240 0.7899
No log 33.2 166 0.5970 0.6007 0.5970 0.7726
No log 33.6 168 0.6016 0.5950 0.6016 0.7756
No log 34.0 170 0.5997 0.6114 0.5997 0.7744
No log 34.4 172 0.6002 0.5975 0.6002 0.7747
No log 34.8 174 0.5994 0.5975 0.5994 0.7742
No log 35.2 176 0.6040 0.6035 0.6040 0.7772
No log 35.6 178 0.6208 0.6147 0.6208 0.7879
No log 36.0 180 0.6307 0.5821 0.6307 0.7942
No log 36.4 182 0.6083 0.6055 0.6083 0.7799
No log 36.8 184 0.5964 0.6046 0.5964 0.7722
No log 37.2 186 0.5927 0.6028 0.5927 0.7699
No log 37.6 188 0.5794 0.5990 0.5794 0.7612
No log 38.0 190 0.5914 0.6347 0.5914 0.7691
No log 38.4 192 0.6027 0.6347 0.6027 0.7763
No log 38.8 194 0.5869 0.6247 0.5869 0.7661
No log 39.2 196 0.5936 0.6048 0.5936 0.7704
No log 39.6 198 0.6098 0.6048 0.6098 0.7809
No log 40.0 200 0.5945 0.6157 0.5945 0.7711
No log 40.4 202 0.5798 0.6288 0.5798 0.7615
No log 40.8 204 0.5717 0.6857 0.5717 0.7561
No log 41.2 206 0.5695 0.6167 0.5695 0.7547
No log 41.6 208 0.5748 0.5969 0.5748 0.7581
No log 42.0 210 0.5891 0.6045 0.5891 0.7675
No log 42.4 212 0.5938 0.5842 0.5938 0.7706
No log 42.8 214 0.5969 0.5752 0.5969 0.7726
No log 43.2 216 0.5968 0.6186 0.5968 0.7725
No log 43.6 218 0.5886 0.6537 0.5886 0.7672
No log 44.0 220 0.5915 0.5534 0.5915 0.7691
No log 44.4 222 0.6314 0.5973 0.6314 0.7946
No log 44.8 224 0.6412 0.6247 0.6412 0.8008
No log 45.2 226 0.6008 0.6639 0.6008 0.7751
No log 45.6 228 0.5719 0.6345 0.5719 0.7562
No log 46.0 230 0.5832 0.6733 0.5832 0.7637
No log 46.4 232 0.5831 0.6919 0.5831 0.7636
No log 46.8 234 0.5828 0.6143 0.5828 0.7634
No log 47.2 236 0.6046 0.5855 0.6046 0.7776
No log 47.6 238 0.6163 0.5719 0.6163 0.7851
No log 48.0 240 0.6185 0.5534 0.6185 0.7865
No log 48.4 242 0.6263 0.5534 0.6263 0.7914
No log 48.8 244 0.6293 0.5797 0.6293 0.7933
No log 49.2 246 0.6352 0.5880 0.6352 0.7970
No log 49.6 248 0.6598 0.4853 0.6598 0.8123
No log 50.0 250 0.6701 0.5175 0.6701 0.8186
No log 50.4 252 0.6555 0.5425 0.6555 0.8096
No log 50.8 254 0.6317 0.6014 0.6317 0.7948
No log 51.2 256 0.6146 0.5917 0.6146 0.7840
No log 51.6 258 0.6123 0.5917 0.6123 0.7825
No log 52.0 260 0.6153 0.5917 0.6153 0.7844
No log 52.4 262 0.6339 0.6092 0.6339 0.7962
No log 52.8 264 0.6607 0.6137 0.6607 0.8128
No log 53.2 266 0.6703 0.6137 0.6703 0.8187
No log 53.6 268 0.6518 0.5404 0.6518 0.8073
No log 54.0 270 0.6337 0.5842 0.6337 0.7960
No log 54.4 272 0.6262 0.5741 0.6262 0.7913
No log 54.8 274 0.6199 0.5785 0.6199 0.7874
No log 55.2 276 0.6158 0.5809 0.6158 0.7847
No log 55.6 278 0.6137 0.5785 0.6137 0.7834
No log 56.0 280 0.6079 0.5866 0.6079 0.7797
No log 56.4 282 0.6013 0.5752 0.6013 0.7754
No log 56.8 284 0.5893 0.6102 0.5893 0.7677
No log 57.2 286 0.5870 0.6102 0.5870 0.7662
No log 57.6 288 0.6020 0.5948 0.6020 0.7759
No log 58.0 290 0.6204 0.6319 0.6204 0.7877
No log 58.4 292 0.6169 0.5948 0.6169 0.7854
No log 58.8 294 0.5939 0.6102 0.5939 0.7706
No log 59.2 296 0.5847 0.5917 0.5847 0.7647
No log 59.6 298 0.5810 0.6035 0.5810 0.7622
No log 60.0 300 0.5766 0.5891 0.5766 0.7593
No log 60.4 302 0.5802 0.5855 0.5802 0.7617
No log 60.8 304 0.5944 0.6137 0.5944 0.7710
No log 61.2 306 0.6052 0.6137 0.6052 0.7780
No log 61.6 308 0.6214 0.6319 0.6214 0.7883
No log 62.0 310 0.6118 0.6137 0.6118 0.7821
No log 62.4 312 0.5902 0.6122 0.5902 0.7682
No log 62.8 314 0.5805 0.6252 0.5805 0.7619
No log 63.2 316 0.5784 0.6252 0.5784 0.7606
No log 63.6 318 0.5799 0.6252 0.5799 0.7615
No log 64.0 320 0.5908 0.5855 0.5908 0.7686
No log 64.4 322 0.6226 0.6137 0.6226 0.7890
No log 64.8 324 0.6575 0.5677 0.6575 0.8108
No log 65.2 326 0.6683 0.5864 0.6683 0.8175
No log 65.6 328 0.6380 0.6109 0.6380 0.7987
No log 66.0 330 0.6021 0.6137 0.6021 0.7759
No log 66.4 332 0.5762 0.6380 0.5762 0.7591
No log 66.8 334 0.5716 0.6288 0.5716 0.7561
No log 67.2 336 0.5719 0.6217 0.5719 0.7563
No log 67.6 338 0.5707 0.5990 0.5707 0.7555
No log 68.0 340 0.5743 0.5990 0.5743 0.7579
No log 68.4 342 0.5783 0.5990 0.5783 0.7605
No log 68.8 344 0.5855 0.6301 0.5855 0.7652
No log 69.2 346 0.5940 0.5740 0.5940 0.7707
No log 69.6 348 0.6023 0.5740 0.6023 0.7761
No log 70.0 350 0.6146 0.6167 0.6146 0.7839
No log 70.4 352 0.6168 0.6167 0.6168 0.7853
No log 70.8 354 0.6096 0.6167 0.6096 0.7808
No log 71.2 356 0.5958 0.5740 0.5958 0.7719
No log 71.6 358 0.5904 0.6187 0.5904 0.7684
No log 72.0 360 0.5907 0.6187 0.5907 0.7686
No log 72.4 362 0.5947 0.6157 0.5947 0.7712
No log 72.8 364 0.5990 0.5842 0.5990 0.7740
No log 73.2 366 0.6030 0.5842 0.6030 0.7765
No log 73.6 368 0.6072 0.6147 0.6072 0.7792
No log 74.0 370 0.6036 0.5740 0.6036 0.7769
No log 74.4 372 0.6028 0.5740 0.6028 0.7764
No log 74.8 374 0.6003 0.5740 0.6003 0.7748
No log 75.2 376 0.6001 0.6035 0.6001 0.7747
No log 75.6 378 0.6075 0.6035 0.6075 0.7794
No log 76.0 380 0.6007 0.6035 0.6007 0.7750
No log 76.4 382 0.5876 0.6035 0.5876 0.7666
No log 76.8 384 0.5797 0.6361 0.5797 0.7614
No log 77.2 386 0.5757 0.6028 0.5757 0.7588
No log 77.6 388 0.5735 0.6337 0.5735 0.7573
No log 78.0 390 0.5776 0.6035 0.5776 0.7600
No log 78.4 392 0.5827 0.6147 0.5827 0.7633
No log 78.8 394 0.5844 0.6147 0.5844 0.7645
No log 79.2 396 0.5797 0.6473 0.5797 0.7614
No log 79.6 398 0.5741 0.6038 0.5741 0.7577
No log 80.0 400 0.5734 0.6133 0.5734 0.7572
No log 80.4 402 0.5770 0.6364 0.5770 0.7596
No log 80.8 404 0.5846 0.5950 0.5846 0.7646
No log 81.2 406 0.5882 0.5725 0.5882 0.7669
No log 81.6 408 0.5837 0.6046 0.5837 0.7640
No log 82.0 410 0.5784 0.6262 0.5784 0.7605
No log 82.4 412 0.5804 0.5905 0.5804 0.7619
No log 82.8 414 0.5890 0.6063 0.5890 0.7675
No log 83.2 416 0.6012 0.6035 0.6012 0.7754
No log 83.6 418 0.6124 0.6137 0.6124 0.7825
No log 84.0 420 0.6128 0.6137 0.6128 0.7828
No log 84.4 422 0.6034 0.6137 0.6034 0.7768
No log 84.8 424 0.5920 0.6063 0.5920 0.7694
No log 85.2 426 0.5865 0.6063 0.5865 0.7659
No log 85.6 428 0.5839 0.5880 0.5839 0.7641
No log 86.0 430 0.5819 0.5880 0.5819 0.7628
No log 86.4 432 0.5814 0.5880 0.5814 0.7625
No log 86.8 434 0.5798 0.5797 0.5798 0.7615
No log 87.2 436 0.5814 0.5797 0.5814 0.7625
No log 87.6 438 0.5833 0.5797 0.5833 0.7637
No log 88.0 440 0.5855 0.5822 0.5855 0.7652
No log 88.4 442 0.5877 0.5822 0.5877 0.7666
No log 88.8 444 0.5895 0.5822 0.5895 0.7678
No log 89.2 446 0.5904 0.5822 0.5904 0.7684
No log 89.6 448 0.5910 0.6165 0.5910 0.7688
No log 90.0 450 0.5932 0.6165 0.5932 0.7702
No log 90.4 452 0.5939 0.6133 0.5939 0.7707
No log 90.8 454 0.5944 0.6133 0.5944 0.7709
No log 91.2 456 0.5944 0.5786 0.5944 0.7710
No log 91.6 458 0.5958 0.5786 0.5958 0.7719
No log 92.0 460 0.5986 0.5540 0.5986 0.7737
No log 92.4 462 0.6018 0.5540 0.6018 0.7757
No log 92.8 464 0.6016 0.5540 0.6016 0.7757
No log 93.2 466 0.6011 0.5540 0.6011 0.7753
No log 93.6 468 0.6003 0.5540 0.6003 0.7748
No log 94.0 470 0.6000 0.5540 0.6000 0.7746
No log 94.4 472 0.5981 0.5540 0.5981 0.7733
No log 94.8 474 0.5961 0.5540 0.5961 0.7721
No log 95.2 476 0.5942 0.5786 0.5942 0.7709
No log 95.6 478 0.5930 0.5786 0.5930 0.7701
No log 96.0 480 0.5918 0.5880 0.5918 0.7693
No log 96.4 482 0.5911 0.5880 0.5911 0.7688
No log 96.8 484 0.5907 0.5880 0.5907 0.7686
No log 97.2 486 0.5903 0.5880 0.5903 0.7683
No log 97.6 488 0.5898 0.5880 0.5898 0.7680
No log 98.0 490 0.5893 0.5880 0.5893 0.7676
No log 98.4 492 0.5890 0.5880 0.5890 0.7674
No log 98.8 494 0.5887 0.5880 0.5887 0.7673
No log 99.2 496 0.5884 0.5880 0.5884 0.7671
No log 99.6 498 0.5883 0.5880 0.5883 0.7670
0.1675 100.0 500 0.5882 0.5880 0.5882 0.7669

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k1_task5_organization

Finetuned
(4019)
this model