ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k15_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5974
  • Qwk: 0.5960
  • Mse: 0.5974
  • Rmse: 0.7729

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0267 2 4.0669 0.0070 4.0669 2.0167
No log 0.0533 4 2.9090 0.0152 2.9090 1.7056
No log 0.08 6 1.2960 0.0 1.2960 1.1384
No log 0.1067 8 1.1147 0.3048 1.1147 1.0558
No log 0.1333 10 1.0430 0.2135 1.0430 1.0213
No log 0.16 12 1.1226 0.1727 1.1226 1.0595
No log 0.1867 14 1.1844 0.0970 1.1844 1.0883
No log 0.2133 16 1.3299 0.0 1.3299 1.1532
No log 0.24 18 1.3558 0.0 1.3558 1.1644
No log 0.2667 20 1.3790 0.0 1.3790 1.1743
No log 0.2933 22 1.2796 -0.0148 1.2796 1.1312
No log 0.32 24 1.1070 0.1261 1.1070 1.0521
No log 0.3467 26 0.9581 0.3207 0.9581 0.9788
No log 0.3733 28 0.9127 0.2770 0.9127 0.9554
No log 0.4 30 0.8811 0.2770 0.8811 0.9387
No log 0.4267 32 0.9003 0.3251 0.9003 0.9489
No log 0.4533 34 1.0816 0.1764 1.0816 1.0400
No log 0.48 36 1.1365 0.2119 1.1365 1.0661
No log 0.5067 38 0.9442 0.3026 0.9442 0.9717
No log 0.5333 40 0.7593 0.4585 0.7593 0.8714
No log 0.56 42 0.7200 0.4019 0.7200 0.8485
No log 0.5867 44 0.7143 0.4516 0.7143 0.8451
No log 0.6133 46 0.7583 0.5033 0.7583 0.8708
No log 0.64 48 0.7634 0.5503 0.7634 0.8737
No log 0.6667 50 0.7299 0.4444 0.7299 0.8543
No log 0.6933 52 0.6961 0.5115 0.6961 0.8343
No log 0.72 54 0.6817 0.6022 0.6817 0.8256
No log 0.7467 56 0.6949 0.4849 0.6949 0.8336
No log 0.7733 58 0.8157 0.5005 0.8157 0.9032
No log 0.8 60 0.7276 0.4971 0.7276 0.8530
No log 0.8267 62 0.7046 0.5732 0.7046 0.8394
No log 0.8533 64 0.8089 0.5411 0.8089 0.8994
No log 0.88 66 0.9645 0.5507 0.9645 0.9821
No log 0.9067 68 0.9393 0.5526 0.9393 0.9692
No log 0.9333 70 0.7803 0.5329 0.7803 0.8834
No log 0.96 72 0.7684 0.6598 0.7684 0.8766
No log 0.9867 74 1.0090 0.5679 1.0090 1.0045
No log 1.0133 76 1.3310 0.3650 1.3310 1.1537
No log 1.04 78 1.2523 0.4182 1.2523 1.1191
No log 1.0667 80 1.0030 0.3813 1.0030 1.0015
No log 1.0933 82 0.7379 0.5773 0.7379 0.8590
No log 1.12 84 0.7610 0.5998 0.7610 0.8723
No log 1.1467 86 0.6616 0.6198 0.6616 0.8134
No log 1.1733 88 0.6368 0.5662 0.6368 0.7980
No log 1.2 90 0.6859 0.6194 0.6859 0.8282
No log 1.2267 92 0.5913 0.6269 0.5913 0.7690
No log 1.2533 94 0.5983 0.5987 0.5983 0.7735
No log 1.28 96 0.6370 0.6451 0.6370 0.7981
No log 1.3067 98 0.6789 0.6630 0.6789 0.8239
No log 1.3333 100 0.7093 0.6239 0.7093 0.8422
No log 1.3600 102 0.7108 0.6398 0.7108 0.8431
No log 1.3867 104 0.7100 0.6104 0.7100 0.8426
No log 1.4133 106 0.7270 0.6689 0.7270 0.8527
No log 1.44 108 0.6990 0.6354 0.6990 0.8361
No log 1.4667 110 0.6416 0.6196 0.6416 0.8010
No log 1.4933 112 0.6311 0.6266 0.6311 0.7944
No log 1.52 114 0.6964 0.6549 0.6964 0.8345
No log 1.5467 116 0.6615 0.6367 0.6615 0.8133
No log 1.5733 118 0.5832 0.6432 0.5832 0.7637
No log 1.6 120 0.5382 0.6437 0.5382 0.7336
No log 1.6267 122 0.7503 0.6278 0.7503 0.8662
No log 1.6533 124 0.8662 0.5367 0.8662 0.9307
No log 1.6800 126 0.6713 0.6730 0.6713 0.8193
No log 1.7067 128 0.5035 0.6956 0.5035 0.7096
No log 1.7333 130 0.7095 0.6624 0.7095 0.8423
No log 1.76 132 0.8668 0.5817 0.8668 0.9310
No log 1.7867 134 0.7866 0.6005 0.7866 0.8869
No log 1.8133 136 0.6394 0.6609 0.6394 0.7996
No log 1.8400 138 0.5454 0.6894 0.5454 0.7385
No log 1.8667 140 0.5180 0.6947 0.5180 0.7197
No log 1.8933 142 0.5995 0.6438 0.5995 0.7743
No log 1.92 144 0.6529 0.6061 0.6529 0.8080
No log 1.9467 146 0.6217 0.5704 0.6217 0.7885
No log 1.9733 148 0.6156 0.5644 0.6156 0.7846
No log 2.0 150 0.6873 0.6328 0.6873 0.8290
No log 2.0267 152 0.7673 0.5307 0.7673 0.8759
No log 2.0533 154 0.7397 0.5725 0.7397 0.8600
No log 2.08 156 0.6382 0.6328 0.6382 0.7989
No log 2.1067 158 0.5611 0.5887 0.5611 0.7491
No log 2.1333 160 0.5623 0.6854 0.5623 0.7499
No log 2.16 162 0.5442 0.6812 0.5442 0.7377
No log 2.1867 164 0.5384 0.6406 0.5384 0.7338
No log 2.2133 166 0.6526 0.6646 0.6526 0.8078
No log 2.24 168 0.8951 0.5781 0.8951 0.9461
No log 2.2667 170 0.9365 0.5501 0.9365 0.9677
No log 2.2933 172 0.7618 0.6296 0.7618 0.8728
No log 2.32 174 0.6305 0.6111 0.6305 0.7941
No log 2.3467 176 0.6448 0.6054 0.6448 0.8030
No log 2.3733 178 0.6728 0.5629 0.6728 0.8203
No log 2.4 180 0.6530 0.5085 0.6530 0.8081
No log 2.4267 182 0.6541 0.5644 0.6541 0.8088
No log 2.4533 184 0.6690 0.6360 0.6690 0.8179
No log 2.48 186 0.6419 0.6018 0.6419 0.8012
No log 2.5067 188 0.6316 0.6345 0.6316 0.7947
No log 2.5333 190 0.6336 0.6219 0.6336 0.7960
No log 2.56 192 0.6483 0.6325 0.6483 0.8052
No log 2.5867 194 0.6338 0.6528 0.6338 0.7961
No log 2.6133 196 0.6261 0.5950 0.6261 0.7912
No log 2.64 198 0.6127 0.6272 0.6127 0.7827
No log 2.6667 200 0.6032 0.6568 0.6032 0.7767
No log 2.6933 202 0.6049 0.6492 0.6049 0.7777
No log 2.7200 204 0.6182 0.6689 0.6182 0.7862
No log 2.7467 206 0.6236 0.6689 0.6236 0.7897
No log 2.7733 208 0.6393 0.6662 0.6393 0.7996
No log 2.8 210 0.6144 0.6251 0.6144 0.7838
No log 2.8267 212 0.6459 0.6120 0.6459 0.8036
No log 2.8533 214 0.6740 0.6041 0.6740 0.8210
No log 2.88 216 0.6985 0.5603 0.6985 0.8358
No log 2.9067 218 0.7028 0.5720 0.7028 0.8383
No log 2.9333 220 0.6802 0.5895 0.6802 0.8247
No log 2.96 222 0.7473 0.6488 0.7473 0.8645
No log 2.9867 224 0.7473 0.6795 0.7473 0.8645
No log 3.0133 226 0.7062 0.6112 0.7062 0.8404
No log 3.04 228 0.6952 0.5921 0.6952 0.8338
No log 3.0667 230 0.6836 0.6259 0.6836 0.8268
No log 3.0933 232 0.6812 0.6067 0.6812 0.8253
No log 3.12 234 0.6530 0.6076 0.6530 0.8081
No log 3.1467 236 0.6424 0.5891 0.6424 0.8015
No log 3.1733 238 0.6588 0.5552 0.6588 0.8117
No log 3.2 240 0.6222 0.5891 0.6222 0.7888
No log 3.2267 242 0.5722 0.6046 0.5722 0.7565
No log 3.2533 244 0.5694 0.6129 0.5694 0.7546
No log 3.2800 246 0.5581 0.6251 0.5581 0.7470
No log 3.3067 248 0.5419 0.6243 0.5419 0.7361
No log 3.3333 250 0.6015 0.6188 0.6015 0.7755
No log 3.36 252 0.6570 0.6333 0.6570 0.8106
No log 3.3867 254 0.6817 0.6476 0.6817 0.8256
No log 3.4133 256 0.6670 0.6238 0.6670 0.8167
No log 3.44 258 0.6334 0.5948 0.6334 0.7959
No log 3.4667 260 0.6681 0.5688 0.6681 0.8174
No log 3.4933 262 0.7076 0.5577 0.7076 0.8412
No log 3.52 264 0.7215 0.5032 0.7215 0.8494
No log 3.5467 266 0.6778 0.5410 0.6778 0.8233
No log 3.5733 268 0.6198 0.5678 0.6198 0.7873
No log 3.6 270 0.5839 0.6473 0.5839 0.7641
No log 3.6267 272 0.5978 0.6901 0.5978 0.7732
No log 3.6533 274 0.6383 0.6340 0.6383 0.7989
No log 3.68 276 0.7059 0.5842 0.7059 0.8402
No log 3.7067 278 0.7757 0.5895 0.7757 0.8807
No log 3.7333 280 0.7295 0.5811 0.7295 0.8541
No log 3.76 282 0.6273 0.5642 0.6273 0.7920
No log 3.7867 284 0.5586 0.5490 0.5586 0.7474
No log 3.8133 286 0.5163 0.6673 0.5163 0.7185
No log 3.84 288 0.5323 0.7101 0.5323 0.7296
No log 3.8667 290 0.5561 0.7168 0.5561 0.7457
No log 3.8933 292 0.5813 0.6928 0.5813 0.7624
No log 3.92 294 0.6090 0.7161 0.6090 0.7804
No log 3.9467 296 0.6111 0.6589 0.6111 0.7818
No log 3.9733 298 0.5996 0.6528 0.5996 0.7743
No log 4.0 300 0.6367 0.6022 0.6367 0.7979
No log 4.0267 302 0.7019 0.5524 0.7019 0.8378
No log 4.0533 304 0.7130 0.5703 0.7130 0.8444
No log 4.08 306 0.6977 0.6028 0.6977 0.8353
No log 4.1067 308 0.7310 0.5486 0.7310 0.8550
No log 4.1333 310 0.6541 0.5832 0.6541 0.8088
No log 4.16 312 0.5684 0.5729 0.5684 0.7539
No log 4.1867 314 0.5062 0.6625 0.5062 0.7115
No log 4.2133 316 0.4951 0.6993 0.4951 0.7036
No log 4.24 318 0.5098 0.6724 0.5098 0.7140
No log 4.2667 320 0.5521 0.5975 0.5521 0.7430
No log 4.2933 322 0.6206 0.5666 0.6206 0.7878
No log 4.32 324 0.7545 0.6122 0.7545 0.8686
No log 4.3467 326 0.7913 0.56 0.7913 0.8896
No log 4.3733 328 0.6909 0.5668 0.6909 0.8312
No log 4.4 330 0.6274 0.6186 0.6274 0.7921
No log 4.4267 332 0.6303 0.5647 0.6303 0.7939
No log 4.4533 334 0.6299 0.6057 0.6299 0.7937
No log 4.48 336 0.6671 0.5731 0.6671 0.8168
No log 4.5067 338 0.6824 0.6160 0.6824 0.8261
No log 4.5333 340 0.6592 0.5731 0.6592 0.8119
No log 4.5600 342 0.6267 0.6456 0.6267 0.7916
No log 4.5867 344 0.6185 0.6570 0.6185 0.7864
No log 4.6133 346 0.6111 0.6667 0.6111 0.7817
No log 4.64 348 0.6259 0.6119 0.6259 0.7912
No log 4.6667 350 0.7116 0.5864 0.7116 0.8436
No log 4.6933 352 0.7122 0.6061 0.7122 0.8439
No log 4.72 354 0.6173 0.6291 0.6173 0.7857
No log 4.7467 356 0.5731 0.6507 0.5731 0.7570
No log 4.7733 358 0.5741 0.6699 0.5741 0.7577
No log 4.8 360 0.5664 0.6460 0.5664 0.7526
No log 4.8267 362 0.5984 0.6664 0.5984 0.7736
No log 4.8533 364 0.6289 0.6555 0.6289 0.7930
No log 4.88 366 0.6402 0.6455 0.6402 0.8001
No log 4.9067 368 0.6321 0.6278 0.6321 0.7950
No log 4.9333 370 0.6554 0.6048 0.6554 0.8096
No log 4.96 372 0.6899 0.5507 0.6899 0.8306
No log 4.9867 374 0.7076 0.5621 0.7076 0.8412
No log 5.0133 376 0.7106 0.5975 0.7106 0.8430
No log 5.04 378 0.7248 0.5492 0.7248 0.8513
No log 5.0667 380 0.7617 0.5509 0.7617 0.8728
No log 5.0933 382 0.7639 0.5813 0.7639 0.8740
No log 5.12 384 0.7232 0.5287 0.7232 0.8504
No log 5.1467 386 0.7326 0.4893 0.7326 0.8559
No log 5.1733 388 0.8029 0.5160 0.8029 0.8961
No log 5.2 390 0.8113 0.4921 0.8113 0.9007
No log 5.2267 392 0.7460 0.5173 0.7460 0.8637
No log 5.2533 394 0.6605 0.5345 0.6605 0.8127
No log 5.28 396 0.6372 0.5910 0.6372 0.7983
No log 5.3067 398 0.6344 0.5548 0.6344 0.7965
No log 5.3333 400 0.6146 0.5843 0.6146 0.7839
No log 5.36 402 0.6146 0.6380 0.6146 0.7839
No log 5.3867 404 0.6396 0.6380 0.6396 0.7998
No log 5.4133 406 0.7257 0.5462 0.7257 0.8519
No log 5.44 408 0.7759 0.5135 0.7759 0.8809
No log 5.4667 410 0.7755 0.5329 0.7755 0.8806
No log 5.4933 412 0.7023 0.5510 0.7023 0.8380
No log 5.52 414 0.6546 0.5503 0.6546 0.8091
No log 5.5467 416 0.6376 0.5874 0.6376 0.7985
No log 5.5733 418 0.6313 0.5874 0.6313 0.7946
No log 5.6 420 0.6218 0.6252 0.6218 0.7885
No log 5.6267 422 0.6019 0.6424 0.6019 0.7758
No log 5.6533 424 0.6025 0.6491 0.6025 0.7762
No log 5.68 426 0.6274 0.6110 0.6274 0.7921
No log 5.7067 428 0.6098 0.6347 0.6098 0.7809
No log 5.7333 430 0.5965 0.6252 0.5965 0.7724
No log 5.76 432 0.6243 0.5862 0.6243 0.7901
No log 5.7867 434 0.6342 0.5874 0.6342 0.7964
No log 5.8133 436 0.6350 0.6065 0.6350 0.7969
No log 5.84 438 0.6105 0.6035 0.6105 0.7814
No log 5.8667 440 0.5910 0.6229 0.5910 0.7688
No log 5.8933 442 0.5699 0.6597 0.5699 0.7549
No log 5.92 444 0.5504 0.6380 0.5504 0.7419
No log 5.9467 446 0.5475 0.6186 0.5475 0.7399
No log 5.9733 448 0.5549 0.6488 0.5549 0.7449
No log 6.0 450 0.6306 0.5833 0.6306 0.7941
No log 6.0267 452 0.7780 0.5756 0.7780 0.8820
No log 6.0533 454 0.8182 0.5745 0.8182 0.9045
No log 6.08 456 0.7531 0.5358 0.7531 0.8678
No log 6.1067 458 0.7121 0.5062 0.7121 0.8439
No log 6.1333 460 0.6498 0.6143 0.6498 0.8061
No log 6.16 462 0.6260 0.6390 0.6260 0.7912
No log 6.1867 464 0.6169 0.6537 0.6169 0.7854
No log 6.2133 466 0.6144 0.6572 0.6144 0.7838
No log 6.24 468 0.5965 0.6572 0.5965 0.7724
No log 6.2667 470 0.5728 0.6762 0.5728 0.7568
No log 6.2933 472 0.5680 0.6762 0.5680 0.7536
No log 6.32 474 0.5661 0.6720 0.5661 0.7524
No log 6.3467 476 0.5629 0.6680 0.5629 0.7503
No log 6.3733 478 0.5645 0.6431 0.5645 0.7513
No log 6.4 480 0.5947 0.6811 0.5947 0.7711
No log 6.4267 482 0.6013 0.6811 0.6013 0.7754
No log 6.4533 484 0.5851 0.6766 0.5851 0.7649
No log 6.48 486 0.5726 0.6689 0.5726 0.7567
No log 6.5067 488 0.6061 0.6833 0.6061 0.7785
No log 6.5333 490 0.6206 0.5889 0.6206 0.7878
No log 6.5600 492 0.6189 0.5771 0.6189 0.7867
No log 6.5867 494 0.5978 0.5902 0.5978 0.7731
No log 6.6133 496 0.5855 0.6276 0.5855 0.7652
No log 6.64 498 0.6264 0.6282 0.6264 0.7915
0.3023 6.6667 500 0.6734 0.6450 0.6734 0.8206
0.3023 6.6933 502 0.6382 0.6774 0.6382 0.7989
0.3023 6.72 504 0.5794 0.6695 0.5794 0.7612
0.3023 6.7467 506 0.5631 0.6510 0.5631 0.7504
0.3023 6.7733 508 0.5737 0.6719 0.5737 0.7575
0.3023 6.8 510 0.6108 0.5902 0.6108 0.7815
0.3023 6.8267 512 0.6008 0.5846 0.6008 0.7751
0.3023 6.8533 514 0.5907 0.6176 0.5907 0.7685
0.3023 6.88 516 0.5830 0.6539 0.5830 0.7635
0.3023 6.9067 518 0.5838 0.6539 0.5838 0.7640
0.3023 6.9333 520 0.5778 0.6736 0.5778 0.7601
0.3023 6.96 522 0.5766 0.6415 0.5766 0.7593
0.3023 6.9867 524 0.5867 0.6672 0.5867 0.7660
0.3023 7.0133 526 0.5857 0.6813 0.5857 0.7653
0.3023 7.04 528 0.5811 0.6861 0.5811 0.7623
0.3023 7.0667 530 0.5800 0.6689 0.5800 0.7616
0.3023 7.0933 532 0.5896 0.6708 0.5896 0.7678
0.3023 7.12 534 0.5974 0.6108 0.5974 0.7729
0.3023 7.1467 536 0.5869 0.6770 0.5869 0.7661
0.3023 7.1733 538 0.5951 0.6712 0.5951 0.7714
0.3023 7.2 540 0.6379 0.6157 0.6379 0.7987
0.3023 7.2267 542 0.6373 0.6282 0.6373 0.7983
0.3023 7.2533 544 0.5875 0.6269 0.5875 0.7665
0.3023 7.28 546 0.5378 0.6788 0.5378 0.7334
0.3023 7.3067 548 0.5388 0.6804 0.5388 0.7340
0.3023 7.3333 550 0.5634 0.7071 0.5634 0.7506
0.3023 7.36 552 0.5636 0.7071 0.5636 0.7508
0.3023 7.3867 554 0.5388 0.6589 0.5388 0.7341
0.3023 7.4133 556 0.5376 0.6689 0.5376 0.7332
0.3023 7.44 558 0.5413 0.6709 0.5413 0.7357
0.3023 7.4667 560 0.5503 0.6709 0.5503 0.7418
0.3023 7.4933 562 0.5475 0.6545 0.5475 0.7399
0.3023 7.52 564 0.5693 0.6551 0.5693 0.7545
0.3023 7.5467 566 0.6082 0.6330 0.6082 0.7799
0.3023 7.5733 568 0.6071 0.6322 0.6071 0.7792
0.3023 7.6 570 0.5900 0.5647 0.5900 0.7681
0.3023 7.6267 572 0.5914 0.6013 0.5914 0.7690
0.3023 7.6533 574 0.5974 0.5960 0.5974 0.7729

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k15_task5_organization

Finetuned
(4019)
this model