ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k1_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5895
  • Qwk: 0.6177
  • Mse: 0.5895
  • Rmse: 0.7678

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.4 2 4.1975 -0.0177 4.1975 2.0488
No log 0.8 4 2.9092 0.0005 2.9092 1.7056
No log 1.2 6 1.4592 0.0380 1.4592 1.2080
No log 1.6 8 1.1385 0.0938 1.1385 1.0670
No log 2.0 10 1.0784 0.2888 1.0784 1.0384
No log 2.4 12 1.0752 0.3498 1.0752 1.0369
No log 2.8 14 1.1446 0.1361 1.1446 1.0699
No log 3.2 16 1.1615 0.1119 1.1615 1.0777
No log 3.6 18 1.1452 0.1361 1.1452 1.0701
No log 4.0 20 1.0380 0.2492 1.0380 1.0188
No log 4.4 22 1.0519 0.2161 1.0519 1.0256
No log 4.8 24 0.9591 0.2740 0.9591 0.9793
No log 5.2 26 1.0284 0.1685 1.0284 1.0141
No log 5.6 28 0.9156 0.3397 0.9156 0.9569
No log 6.0 30 1.0077 0.4224 1.0077 1.0038
No log 6.4 32 1.1904 0.3336 1.1904 1.0910
No log 6.8 34 0.9858 0.4139 0.9858 0.9928
No log 7.2 36 0.8126 0.3874 0.8126 0.9015
No log 7.6 38 0.8421 0.4034 0.8421 0.9177
No log 8.0 40 0.7515 0.5129 0.7515 0.8669
No log 8.4 42 0.8299 0.5811 0.8299 0.9110
No log 8.8 44 0.8882 0.5543 0.8882 0.9424
No log 9.2 46 0.8030 0.5610 0.8030 0.8961
No log 9.6 48 0.7102 0.4813 0.7102 0.8427
No log 10.0 50 0.7364 0.5300 0.7364 0.8581
No log 10.4 52 0.7372 0.5063 0.7372 0.8586
No log 10.8 54 0.7204 0.5009 0.7204 0.8488
No log 11.2 56 0.7517 0.5614 0.7517 0.8670
No log 11.6 58 0.7691 0.5988 0.7691 0.8770
No log 12.0 60 0.6550 0.5330 0.6550 0.8093
No log 12.4 62 0.7105 0.5656 0.7105 0.8429
No log 12.8 64 0.6936 0.5839 0.6936 0.8328
No log 13.2 66 0.7174 0.5718 0.7174 0.8470
No log 13.6 68 0.8329 0.5988 0.8329 0.9127
No log 14.0 70 0.7440 0.5688 0.7440 0.8625
No log 14.4 72 0.6873 0.5352 0.6873 0.8290
No log 14.8 74 0.6527 0.5459 0.6527 0.8079
No log 15.2 76 0.6701 0.5678 0.6701 0.8186
No log 15.6 78 0.6740 0.5785 0.6740 0.8210
No log 16.0 80 0.6653 0.5643 0.6653 0.8157
No log 16.4 82 0.6079 0.5925 0.6079 0.7797
No log 16.8 84 0.5717 0.6207 0.5717 0.7561
No log 17.2 86 0.5908 0.6137 0.5908 0.7687
No log 17.6 88 0.5694 0.6147 0.5694 0.7546
No log 18.0 90 0.5917 0.6244 0.5917 0.7692
No log 18.4 92 0.6720 0.5963 0.6720 0.8198
No log 18.8 94 0.6919 0.5666 0.6919 0.8318
No log 19.2 96 0.6227 0.6122 0.6227 0.7891
No log 19.6 98 0.6480 0.6143 0.6480 0.8050
No log 20.0 100 0.7136 0.5964 0.7136 0.8447
No log 20.4 102 0.8097 0.5745 0.8097 0.8998
No log 20.8 104 0.7875 0.5029 0.7875 0.8874
No log 21.2 106 0.8227 0.5250 0.8227 0.9071
No log 21.6 108 0.8246 0.5250 0.8246 0.9081
No log 22.0 110 0.8113 0.5462 0.8113 0.9007
No log 22.4 112 0.8206 0.5431 0.8206 0.9059
No log 22.8 114 0.7156 0.5799 0.7156 0.8459
No log 23.2 116 0.6134 0.5880 0.6134 0.7832
No log 23.6 118 0.6003 0.6415 0.6003 0.7748
No log 24.0 120 0.6244 0.6291 0.6244 0.7902
No log 24.4 122 0.6445 0.6664 0.6445 0.8028
No log 24.8 124 0.6547 0.6254 0.6547 0.8091
No log 25.2 126 0.7651 0.5862 0.7651 0.8747
No log 25.6 128 0.7848 0.5873 0.7848 0.8859
No log 26.0 130 0.6588 0.6157 0.6588 0.8117
No log 26.4 132 0.6694 0.6815 0.6694 0.8182
No log 26.8 134 0.6808 0.5864 0.6808 0.8251
No log 27.2 136 0.6301 0.6352 0.6301 0.7938
No log 27.6 138 0.5812 0.6451 0.5812 0.7623
No log 28.0 140 0.5759 0.6028 0.5759 0.7589
No log 28.4 142 0.5792 0.5701 0.5792 0.7610
No log 28.8 144 0.6152 0.6177 0.6152 0.7843
No log 29.2 146 0.6301 0.7035 0.6301 0.7938
No log 29.6 148 0.6185 0.6414 0.6185 0.7865
No log 30.0 150 0.6036 0.6244 0.6036 0.7769
No log 30.4 152 0.6110 0.5817 0.6110 0.7817
No log 30.8 154 0.6452 0.5873 0.6452 0.8032
No log 31.2 156 0.6174 0.5645 0.6174 0.7857
No log 31.6 158 0.5849 0.5995 0.5849 0.7648
No log 32.0 160 0.5950 0.6740 0.5950 0.7714
No log 32.4 162 0.5871 0.7042 0.5871 0.7662
No log 32.8 164 0.5703 0.7049 0.5703 0.7552
No log 33.2 166 0.5699 0.7049 0.5699 0.7549
No log 33.6 168 0.5649 0.7049 0.5649 0.7516
No log 34.0 170 0.5727 0.6284 0.5727 0.7568
No log 34.4 172 0.6471 0.6363 0.6471 0.8044
No log 34.8 174 0.6806 0.6175 0.6806 0.8250
No log 35.2 176 0.6404 0.6157 0.6404 0.8002
No log 35.6 178 0.6530 0.6333 0.6530 0.8081
No log 36.0 180 0.6597 0.6626 0.6597 0.8122
No log 36.4 182 0.6309 0.6564 0.6309 0.7943
No log 36.8 184 0.6105 0.5843 0.6105 0.7813
No log 37.2 186 0.6384 0.6099 0.6384 0.7990
No log 37.6 188 0.6312 0.6070 0.6312 0.7945
No log 38.0 190 0.6109 0.5495 0.6109 0.7816
No log 38.4 192 0.6522 0.5740 0.6522 0.8076
No log 38.8 194 0.7303 0.5938 0.7303 0.8546
No log 39.2 196 0.7370 0.6045 0.7370 0.8585
No log 39.6 198 0.6596 0.6249 0.6596 0.8121
No log 40.0 200 0.6087 0.5990 0.6087 0.7802
No log 40.4 202 0.6101 0.5500 0.6101 0.7811
No log 40.8 204 0.6112 0.5747 0.6112 0.7818
No log 41.2 206 0.6353 0.5948 0.6353 0.7971
No log 41.6 208 0.6490 0.5948 0.6490 0.8056
No log 42.0 210 0.6351 0.6575 0.6351 0.7970
No log 42.4 212 0.6119 0.6084 0.6119 0.7823
No log 42.8 214 0.5947 0.6500 0.5947 0.7711
No log 43.2 216 0.5929 0.6500 0.5929 0.7700
No log 43.6 218 0.5874 0.6500 0.5874 0.7664
No log 44.0 220 0.5825 0.6500 0.5825 0.7632
No log 44.4 222 0.5900 0.6632 0.5900 0.7681
No log 44.8 224 0.6080 0.6405 0.6080 0.7797
No log 45.2 226 0.6201 0.6073 0.6201 0.7874
No log 45.6 228 0.6085 0.6736 0.6085 0.7801
No log 46.0 230 0.6109 0.6311 0.6109 0.7816
No log 46.4 232 0.6223 0.6311 0.6223 0.7889
No log 46.8 234 0.6400 0.5969 0.6400 0.8000
No log 47.2 236 0.6667 0.5969 0.6667 0.8165
No log 47.6 238 0.6635 0.6301 0.6635 0.8145
No log 48.0 240 0.6569 0.6094 0.6569 0.8105
No log 48.4 242 0.6432 0.5990 0.6432 0.8020
No log 48.8 244 0.6374 0.5599 0.6374 0.7984
No log 49.2 246 0.6305 0.6217 0.6305 0.7941
No log 49.6 248 0.6258 0.5808 0.6258 0.7911
No log 50.0 250 0.6214 0.6325 0.6214 0.7883
No log 50.4 252 0.6093 0.6962 0.6093 0.7806
No log 50.8 254 0.5990 0.6962 0.5990 0.7740
No log 51.2 256 0.5975 0.6584 0.5975 0.7730
No log 51.6 258 0.5884 0.6916 0.5884 0.7671
No log 52.0 260 0.5840 0.6814 0.5840 0.7642
No log 52.4 262 0.5914 0.6482 0.5914 0.7690
No log 52.8 264 0.5960 0.6482 0.5960 0.7720
No log 53.2 266 0.5823 0.6491 0.5823 0.7631
No log 53.6 268 0.5816 0.6528 0.5816 0.7626
No log 54.0 270 0.5917 0.6291 0.5917 0.7692
No log 54.4 272 0.6096 0.6396 0.6096 0.7808
No log 54.8 274 0.6104 0.6396 0.6104 0.7813
No log 55.2 276 0.6069 0.6389 0.6069 0.7790
No log 55.6 278 0.6218 0.5933 0.6218 0.7886
No log 56.0 280 0.6353 0.5517 0.6353 0.7971
No log 56.4 282 0.6560 0.5304 0.6560 0.8100
No log 56.8 284 0.6571 0.5304 0.6571 0.8106
No log 57.2 286 0.6383 0.6167 0.6383 0.7990
No log 57.6 288 0.6273 0.5887 0.6273 0.7920
No log 58.0 290 0.6314 0.5909 0.6314 0.7946
No log 58.4 292 0.6273 0.5909 0.6273 0.7920
No log 58.8 294 0.6191 0.5913 0.6191 0.7868
No log 59.2 296 0.6124 0.5887 0.6124 0.7825
No log 59.6 298 0.6089 0.5990 0.6089 0.7803
No log 60.0 300 0.6091 0.6187 0.6091 0.7804
No log 60.4 302 0.5998 0.6389 0.5998 0.7745
No log 60.8 304 0.5966 0.6389 0.5966 0.7724
No log 61.2 306 0.5992 0.6623 0.5992 0.7741
No log 61.6 308 0.6097 0.5959 0.6097 0.7808
No log 62.0 310 0.6259 0.5948 0.6259 0.7912
No log 62.4 312 0.6361 0.5948 0.6361 0.7975
No log 62.8 314 0.6139 0.5948 0.6139 0.7835
No log 63.2 316 0.6067 0.5948 0.6067 0.7789
No log 63.6 318 0.6122 0.5948 0.6122 0.7825
No log 64.0 320 0.6071 0.5844 0.6071 0.7792
No log 64.4 322 0.5947 0.5880 0.5947 0.7712
No log 64.8 324 0.5945 0.5880 0.5945 0.7711
No log 65.2 326 0.6108 0.5844 0.6108 0.7815
No log 65.6 328 0.6217 0.5844 0.6217 0.7885
No log 66.0 330 0.6051 0.5844 0.6051 0.7779
No log 66.4 332 0.5882 0.6252 0.5882 0.7669
No log 66.8 334 0.5857 0.6262 0.5857 0.7653
No log 67.2 336 0.5845 0.6262 0.5845 0.7645
No log 67.6 338 0.5900 0.5763 0.5900 0.7681
No log 68.0 340 0.5958 0.5763 0.5958 0.7719
No log 68.4 342 0.5997 0.5844 0.5997 0.7744
No log 68.8 344 0.6019 0.6177 0.6019 0.7758
No log 69.2 346 0.5943 0.6291 0.5943 0.7709
No log 69.6 348 0.5919 0.6291 0.5919 0.7693
No log 70.0 350 0.5841 0.6632 0.5841 0.7643
No log 70.4 352 0.5768 0.5902 0.5768 0.7595
No log 70.8 354 0.5751 0.5902 0.5751 0.7583
No log 71.2 356 0.5766 0.6229 0.5766 0.7594
No log 71.6 358 0.5802 0.6025 0.5802 0.7617
No log 72.0 360 0.5803 0.6025 0.5803 0.7618
No log 72.4 362 0.5877 0.6025 0.5877 0.7666
No log 72.8 364 0.5947 0.6396 0.5947 0.7712
No log 73.2 366 0.6075 0.5948 0.6075 0.7794
No log 73.6 368 0.6072 0.5948 0.6072 0.7793
No log 74.0 370 0.5942 0.6510 0.5942 0.7708
No log 74.4 372 0.5779 0.6405 0.5779 0.7602
No log 74.8 374 0.5740 0.6370 0.5740 0.7576
No log 75.2 376 0.5736 0.6025 0.5736 0.7574
No log 75.6 378 0.5670 0.6025 0.5670 0.7530
No log 76.0 380 0.5650 0.6441 0.5650 0.7517
No log 76.4 382 0.5598 0.6025 0.5598 0.7482
No log 76.8 384 0.5530 0.6460 0.5530 0.7436
No log 77.2 386 0.5535 0.6649 0.5535 0.7440
No log 77.6 388 0.5548 0.6649 0.5548 0.7448
No log 78.0 390 0.5590 0.6460 0.5590 0.7477
No log 78.4 392 0.5645 0.5891 0.5645 0.7513
No log 78.8 394 0.5686 0.5996 0.5686 0.7541
No log 79.2 396 0.5730 0.5996 0.5730 0.7570
No log 79.6 398 0.5747 0.5996 0.5747 0.7581
No log 80.0 400 0.5815 0.5996 0.5815 0.7625
No log 80.4 402 0.5872 0.5996 0.5872 0.7663
No log 80.8 404 0.5929 0.5996 0.5929 0.7700
No log 81.2 406 0.6079 0.5959 0.6079 0.7797
No log 81.6 408 0.6140 0.6063 0.6140 0.7836
No log 82.0 410 0.6067 0.5540 0.6067 0.7789
No log 82.4 412 0.6027 0.5540 0.6027 0.7764
No log 82.8 414 0.5967 0.5880 0.5967 0.7725
No log 83.2 416 0.5970 0.5880 0.5970 0.7726
No log 83.6 418 0.5913 0.5905 0.5913 0.7690
No log 84.0 420 0.5820 0.6025 0.5820 0.7629
No log 84.4 422 0.5762 0.6025 0.5762 0.7591
No log 84.8 424 0.5711 0.6025 0.5711 0.7557
No log 85.2 426 0.5706 0.6025 0.5706 0.7554
No log 85.6 428 0.5727 0.6405 0.5727 0.7567
No log 86.0 430 0.5726 0.6405 0.5726 0.7567
No log 86.4 432 0.5737 0.6405 0.5737 0.7574
No log 86.8 434 0.5743 0.6405 0.5743 0.7578
No log 87.2 436 0.5740 0.6405 0.5740 0.7577
No log 87.6 438 0.5734 0.6405 0.5734 0.7572
No log 88.0 440 0.5723 0.6405 0.5723 0.7565
No log 88.4 442 0.5728 0.6405 0.5728 0.7568
No log 88.8 444 0.5721 0.6025 0.5721 0.7564
No log 89.2 446 0.5699 0.6025 0.5699 0.7549
No log 89.6 448 0.5686 0.6025 0.5686 0.7540
No log 90.0 450 0.5687 0.6405 0.5687 0.7541
No log 90.4 452 0.5715 0.6405 0.5715 0.7560
No log 90.8 454 0.5737 0.6405 0.5737 0.7574
No log 91.2 456 0.5729 0.6405 0.5729 0.7569
No log 91.6 458 0.5740 0.6405 0.5740 0.7577
No log 92.0 460 0.5771 0.6405 0.5771 0.7597
No log 92.4 462 0.5804 0.6405 0.5804 0.7618
No log 92.8 464 0.5811 0.6405 0.5811 0.7623
No log 93.2 466 0.5800 0.6405 0.5800 0.7616
No log 93.6 468 0.5792 0.6405 0.5792 0.7611
No log 94.0 470 0.5798 0.6405 0.5798 0.7615
No log 94.4 472 0.5795 0.6405 0.5795 0.7612
No log 94.8 474 0.5803 0.6291 0.5803 0.7618
No log 95.2 476 0.5806 0.6291 0.5806 0.7620
No log 95.6 478 0.5818 0.6291 0.5818 0.7628
No log 96.0 480 0.5832 0.6177 0.5832 0.7637
No log 96.4 482 0.5849 0.6177 0.5849 0.7648
No log 96.8 484 0.5878 0.6177 0.5878 0.7667
No log 97.2 486 0.5896 0.6177 0.5896 0.7678
No log 97.6 488 0.5908 0.6177 0.5908 0.7687
No log 98.0 490 0.5917 0.6177 0.5917 0.7692
No log 98.4 492 0.5918 0.6177 0.5918 0.7693
No log 98.8 494 0.5907 0.6177 0.5907 0.7686
No log 99.2 496 0.5898 0.6177 0.5898 0.7680
No log 99.6 498 0.5896 0.6177 0.5896 0.7678
0.1784 100.0 500 0.5895 0.6177 0.5895 0.7678

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k1_task5_organization

Finetuned
(4019)
this model