ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k9_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7149
  • Qwk: 0.4484
  • Mse: 0.7149
  • Rmse: 0.8455

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 3.8408 0.0 3.8408 1.9598
No log 0.1333 4 2.1332 0.0715 2.1332 1.4606
No log 0.2 6 1.3969 0.0 1.3969 1.1819
No log 0.2667 8 1.2735 0.0 1.2735 1.1285
No log 0.3333 10 1.3683 0.0 1.3683 1.1697
No log 0.4 12 1.7986 0.0530 1.7986 1.3411
No log 0.4667 14 1.8238 0.0530 1.8238 1.3505
No log 0.5333 16 1.6420 0.0399 1.6420 1.2814
No log 0.6 18 1.3345 0.0 1.3345 1.1552
No log 0.6667 20 1.2164 0.0496 1.2164 1.1029
No log 0.7333 22 1.1806 0.0849 1.1806 1.0866
No log 0.8 24 1.1869 0.1205 1.1869 1.0895
No log 0.8667 26 1.0963 0.1482 1.0963 1.0471
No log 0.9333 28 1.2086 0.1268 1.2086 1.0994
No log 1.0 30 1.1744 0.1119 1.1744 1.0837
No log 1.0667 32 1.1026 0.1003 1.1026 1.0501
No log 1.1333 34 1.1920 0.1091 1.1920 1.0918
No log 1.2 36 1.5365 0.1184 1.5365 1.2396
No log 1.2667 38 1.5078 0.1512 1.5078 1.2279
No log 1.3333 40 1.2656 0.1952 1.2656 1.1250
No log 1.4 42 1.0498 0.1837 1.0498 1.0246
No log 1.4667 44 1.0354 0.2647 1.0354 1.0176
No log 1.5333 46 0.9933 0.2788 0.9933 0.9966
No log 1.6 48 1.0865 0.2893 1.0865 1.0423
No log 1.6667 50 1.0635 0.2662 1.0635 1.0313
No log 1.7333 52 1.0341 0.3165 1.0341 1.0169
No log 1.8 54 0.9740 0.2935 0.9740 0.9869
No log 1.8667 56 0.9309 0.3407 0.9309 0.9648
No log 1.9333 58 0.9850 0.2925 0.9850 0.9925
No log 2.0 60 1.0006 0.2734 1.0006 1.0003
No log 2.0667 62 0.8874 0.3478 0.8874 0.9420
No log 2.1333 64 0.8732 0.2696 0.8732 0.9345
No log 2.2 66 0.8958 0.4359 0.8958 0.9465
No log 2.2667 68 0.9045 0.4599 0.9045 0.9510
No log 2.3333 70 0.8762 0.4854 0.8762 0.9361
No log 2.4 72 0.8626 0.5098 0.8626 0.9288
No log 2.4667 74 0.8372 0.4691 0.8372 0.9150
No log 2.5333 76 0.8814 0.5340 0.8814 0.9388
No log 2.6 78 1.0583 0.3952 1.0583 1.0288
No log 2.6667 80 1.0508 0.3780 1.0508 1.0251
No log 2.7333 82 0.8774 0.4780 0.8774 0.9367
No log 2.8 84 0.9392 0.4084 0.9392 0.9691
No log 2.8667 86 1.1985 0.3609 1.1985 1.0948
No log 2.9333 88 1.2201 0.3642 1.2201 1.1046
No log 3.0 90 1.1558 0.3575 1.1558 1.0751
No log 3.0667 92 0.8908 0.4741 0.8908 0.9438
No log 3.1333 94 0.7078 0.5555 0.7078 0.8413
No log 3.2 96 0.6894 0.5415 0.6894 0.8303
No log 3.2667 98 0.6802 0.5945 0.6802 0.8248
No log 3.3333 100 0.8917 0.5526 0.8917 0.9443
No log 3.4 102 1.1283 0.3992 1.1283 1.0622
No log 3.4667 104 0.9826 0.5170 0.9826 0.9913
No log 3.5333 106 0.6836 0.6225 0.6836 0.8268
No log 3.6 108 0.6644 0.5840 0.6644 0.8151
No log 3.6667 110 0.6609 0.6164 0.6609 0.8129
No log 3.7333 112 0.8282 0.5005 0.8282 0.9100
No log 3.8 114 1.0148 0.3295 1.0148 1.0074
No log 3.8667 116 1.1228 0.3185 1.1228 1.0596
No log 3.9333 118 0.9172 0.3812 0.9172 0.9577
No log 4.0 120 0.7285 0.5555 0.7285 0.8535
No log 4.0667 122 0.6728 0.5017 0.6728 0.8202
No log 4.1333 124 0.6614 0.5373 0.6614 0.8133
No log 4.2 126 0.6551 0.5735 0.6551 0.8094
No log 4.2667 128 0.6674 0.6207 0.6674 0.8169
No log 4.3333 130 0.6623 0.6185 0.6623 0.8138
No log 4.4 132 0.6511 0.5720 0.6511 0.8069
No log 4.4667 134 0.6782 0.5710 0.6782 0.8235
No log 4.5333 136 0.6654 0.5469 0.6654 0.8157
No log 4.6 138 0.6400 0.6115 0.6400 0.8000
No log 4.6667 140 0.7238 0.6449 0.7238 0.8508
No log 4.7333 142 0.6752 0.6334 0.6752 0.8217
No log 4.8 144 0.6380 0.6227 0.6380 0.7988
No log 4.8667 146 0.6396 0.6555 0.6396 0.7997
No log 4.9333 148 0.6469 0.6365 0.6469 0.8043
No log 5.0 150 0.6811 0.5917 0.6811 0.8253
No log 5.0667 152 0.6630 0.6197 0.6630 0.8142
No log 5.1333 154 0.6768 0.6409 0.6768 0.8227
No log 5.2 156 0.6911 0.6067 0.6911 0.8313
No log 5.2667 158 0.7096 0.5582 0.7096 0.8424
No log 5.3333 160 0.7157 0.4660 0.7157 0.8460
No log 5.4 162 0.7095 0.5199 0.7095 0.8423
No log 5.4667 164 0.7183 0.4998 0.7183 0.8475
No log 5.5333 166 0.6991 0.5428 0.6991 0.8361
No log 5.6 168 0.6929 0.5659 0.6929 0.8324
No log 5.6667 170 0.6974 0.5185 0.6974 0.8351
No log 5.7333 172 0.7428 0.4615 0.7428 0.8618
No log 5.8 174 0.7497 0.4615 0.7497 0.8658
No log 5.8667 176 0.7126 0.5373 0.7126 0.8441
No log 5.9333 178 0.6926 0.6259 0.6926 0.8322
No log 6.0 180 0.7108 0.5951 0.7108 0.8431
No log 6.0667 182 0.7080 0.5951 0.7080 0.8414
No log 6.1333 184 0.7303 0.5825 0.7303 0.8546
No log 6.2 186 0.7723 0.5632 0.7723 0.8788
No log 6.2667 188 0.7376 0.5966 0.7376 0.8588
No log 6.3333 190 0.7002 0.5287 0.7002 0.8368
No log 6.4 192 0.7170 0.5066 0.7170 0.8467
No log 6.4667 194 0.7586 0.4864 0.7586 0.8710
No log 6.5333 196 0.7561 0.5217 0.7561 0.8695
No log 6.6 198 0.7500 0.5339 0.7500 0.8660
No log 6.6667 200 0.7137 0.5536 0.7137 0.8448
No log 6.7333 202 0.6679 0.5913 0.6679 0.8172
No log 6.8 204 0.6746 0.5785 0.6746 0.8214
No log 6.8667 206 0.7170 0.5626 0.7170 0.8468
No log 6.9333 208 0.7019 0.6021 0.7019 0.8378
No log 7.0 210 0.6627 0.5274 0.6627 0.8140
No log 7.0667 212 0.6600 0.5862 0.6600 0.8124
No log 7.1333 214 0.6476 0.6221 0.6476 0.8047
No log 7.2 216 0.6740 0.6003 0.6740 0.8209
No log 7.2667 218 0.6231 0.6824 0.6231 0.7894
No log 7.3333 220 0.6114 0.6133 0.6114 0.7819
No log 7.4 222 0.6083 0.6330 0.6083 0.7799
No log 7.4667 224 0.6321 0.6060 0.6321 0.7951
No log 7.5333 226 0.6500 0.5662 0.6500 0.8062
No log 7.6 228 0.6366 0.5641 0.6366 0.7979
No log 7.6667 230 0.6537 0.5215 0.6537 0.8085
No log 7.7333 232 0.6439 0.5054 0.6439 0.8025
No log 7.8 234 0.6961 0.4983 0.6961 0.8344
No log 7.8667 236 0.6997 0.4754 0.6997 0.8365
No log 7.9333 238 0.7052 0.4754 0.7052 0.8398
No log 8.0 240 0.6651 0.4970 0.6651 0.8155
No log 8.0667 242 0.6679 0.5346 0.6679 0.8173
No log 8.1333 244 0.6590 0.5386 0.6590 0.8118
No log 8.2 246 0.6904 0.5229 0.6904 0.8309
No log 8.2667 248 0.8110 0.5387 0.8110 0.9006
No log 8.3333 250 0.7897 0.5711 0.7897 0.8887
No log 8.4 252 0.6621 0.6052 0.6621 0.8137
No log 8.4667 254 0.6908 0.5614 0.6908 0.8312
No log 8.5333 256 0.8933 0.4681 0.8933 0.9451
No log 8.6 258 0.9395 0.4064 0.9395 0.9693
No log 8.6667 260 0.8105 0.5370 0.8105 0.9003
No log 8.7333 262 0.6606 0.5505 0.6606 0.8128
No log 8.8 264 0.7076 0.5699 0.7076 0.8412
No log 8.8667 266 0.7387 0.5699 0.7387 0.8595
No log 8.9333 268 0.7009 0.5255 0.7009 0.8372
No log 9.0 270 0.6758 0.4914 0.6758 0.8221
No log 9.0667 272 0.6889 0.5003 0.6889 0.8300
No log 9.1333 274 0.6572 0.5810 0.6572 0.8107
No log 9.2 276 0.6309 0.5357 0.6309 0.7943
No log 9.2667 278 0.6676 0.5917 0.6676 0.8171
No log 9.3333 280 0.6832 0.5686 0.6832 0.8265
No log 9.4 282 0.6459 0.5917 0.6459 0.8037
No log 9.4667 284 0.6240 0.5523 0.6240 0.7899
No log 9.5333 286 0.6296 0.5171 0.6296 0.7935
No log 9.6 288 0.6325 0.5503 0.6325 0.7953
No log 9.6667 290 0.6313 0.5388 0.6313 0.7946
No log 9.7333 292 0.6278 0.5622 0.6278 0.7923
No log 9.8 294 0.6160 0.5939 0.6160 0.7848
No log 9.8667 296 0.6381 0.5929 0.6381 0.7988
No log 9.9333 298 0.7006 0.6410 0.7006 0.8370
No log 10.0 300 0.6573 0.6581 0.6573 0.8108
No log 10.0667 302 0.6161 0.6039 0.6161 0.7849
No log 10.1333 304 0.6421 0.5748 0.6421 0.8013
No log 10.2 306 0.6389 0.5879 0.6389 0.7993
No log 10.2667 308 0.6467 0.5909 0.6467 0.8042
No log 10.3333 310 0.7123 0.5647 0.7123 0.8440
No log 10.4 312 0.7049 0.6269 0.7049 0.8396
No log 10.4667 314 0.6346 0.6476 0.6346 0.7966
No log 10.5333 316 0.6051 0.6433 0.6051 0.7779
No log 10.6 318 0.5947 0.6433 0.5947 0.7712
No log 10.6667 320 0.6163 0.6886 0.6163 0.7850
No log 10.7333 322 0.6956 0.6089 0.6956 0.8340
No log 10.8 324 0.6946 0.6089 0.6946 0.8334
No log 10.8667 326 0.6594 0.6464 0.6594 0.8121
No log 10.9333 328 0.6280 0.6252 0.6280 0.7924
No log 11.0 330 0.6331 0.5797 0.6331 0.7956
No log 11.0667 332 0.6423 0.6539 0.6423 0.8014
No log 11.1333 334 0.6789 0.6340 0.6789 0.8240
No log 11.2 336 0.6689 0.5937 0.6689 0.8179
No log 11.2667 338 0.6707 0.5597 0.6707 0.8189
No log 11.3333 340 0.6597 0.6133 0.6597 0.8122
No log 11.4 342 0.6491 0.5656 0.6491 0.8057
No log 11.4667 344 0.6407 0.5771 0.6407 0.8004
No log 11.5333 346 0.6360 0.6032 0.6360 0.7975
No log 11.6 348 0.6380 0.6007 0.6380 0.7988
No log 11.6667 350 0.6455 0.6370 0.6455 0.8034
No log 11.7333 352 0.6544 0.5641 0.6544 0.8089
No log 11.8 354 0.6691 0.5339 0.6691 0.8180
No log 11.8667 356 0.6626 0.5871 0.6626 0.8140
No log 11.9333 358 0.6511 0.5644 0.6511 0.8069
No log 12.0 360 0.6225 0.6479 0.6225 0.7890
No log 12.0667 362 0.6105 0.6125 0.6105 0.7814
No log 12.1333 364 0.6171 0.6144 0.6171 0.7856
No log 12.2 366 0.6233 0.5530 0.6233 0.7895
No log 12.2667 368 0.6158 0.5961 0.6158 0.7847
No log 12.3333 370 0.6375 0.6664 0.6375 0.7984
No log 12.4 372 0.7203 0.5677 0.7203 0.8487
No log 12.4667 374 0.7178 0.6025 0.7178 0.8472
No log 12.5333 376 0.6536 0.6133 0.6536 0.8084
No log 12.6 378 0.6353 0.5982 0.6353 0.7971
No log 12.6667 380 0.6126 0.6089 0.6126 0.7827
No log 12.7333 382 0.6013 0.6297 0.6013 0.7755
No log 12.8 384 0.6563 0.6705 0.6563 0.8101
No log 12.8667 386 0.6818 0.6605 0.6818 0.8257
No log 12.9333 388 0.6581 0.6446 0.6581 0.8112
No log 13.0 390 0.6488 0.6035 0.6488 0.8055
No log 13.0667 392 0.6491 0.5301 0.6491 0.8057
No log 13.1333 394 0.6364 0.5747 0.6364 0.7977
No log 13.2 396 0.6417 0.6482 0.6417 0.8010
No log 13.2667 398 0.7489 0.5328 0.7489 0.8654
No log 13.3333 400 0.7909 0.5213 0.7909 0.8893
No log 13.4 402 0.7287 0.5973 0.7287 0.8537
No log 13.4667 404 0.6338 0.6415 0.6338 0.7961
No log 13.5333 406 0.6255 0.5684 0.6255 0.7909
No log 13.6 408 0.6433 0.5230 0.6433 0.8021
No log 13.6667 410 0.6332 0.5684 0.6332 0.7957
No log 13.7333 412 0.6233 0.6978 0.6233 0.7895
No log 13.8 414 0.6754 0.6236 0.6754 0.8218
No log 13.8667 416 0.7357 0.5980 0.7357 0.8577
No log 13.9333 418 0.6953 0.6634 0.6953 0.8339
No log 14.0 420 0.5893 0.6709 0.5893 0.7677
No log 14.0667 422 0.5914 0.6675 0.5914 0.7690
No log 14.1333 424 0.7142 0.5803 0.7142 0.8451
No log 14.2 426 0.7262 0.5614 0.7262 0.8521
No log 14.2667 428 0.6471 0.5476 0.6471 0.8044
No log 14.3333 430 0.6515 0.5996 0.6515 0.8071
No log 14.4 432 0.7185 0.6063 0.7185 0.8476
No log 14.4667 434 0.7281 0.6063 0.7281 0.8533
No log 14.5333 436 0.6923 0.5880 0.6923 0.8321
No log 14.6 438 0.6741 0.5459 0.6741 0.8210
No log 14.6667 440 0.7007 0.5151 0.7007 0.8371
No log 14.7333 442 0.6901 0.5245 0.6901 0.8307
No log 14.8 444 0.6634 0.6269 0.6634 0.8145
No log 14.8667 446 0.6858 0.5797 0.6858 0.8282
No log 14.9333 448 0.6931 0.5797 0.6931 0.8325
No log 15.0 450 0.6976 0.5969 0.6976 0.8353
No log 15.0667 452 0.6932 0.6291 0.6932 0.8326
No log 15.1333 454 0.6975 0.5699 0.6975 0.8352
No log 15.2 456 0.6834 0.6291 0.6834 0.8267
No log 15.2667 458 0.6745 0.6177 0.6745 0.8213
No log 15.3333 460 0.6661 0.5606 0.6661 0.8162
No log 15.4 462 0.6765 0.6060 0.6765 0.8225
No log 15.4667 464 0.6793 0.5650 0.6793 0.8242
No log 15.5333 466 0.6686 0.5820 0.6686 0.8177
No log 15.6 468 0.6826 0.5909 0.6826 0.8262
No log 15.6667 470 0.6824 0.5909 0.6824 0.8261
No log 15.7333 472 0.6685 0.5919 0.6685 0.8176
No log 15.8 474 0.6636 0.5909 0.6636 0.8146
No log 15.8667 476 0.6443 0.6128 0.6443 0.8027
No log 15.9333 478 0.6317 0.6207 0.6317 0.7948
No log 16.0 480 0.6174 0.6247 0.6174 0.7858
No log 16.0667 482 0.6162 0.7151 0.6162 0.7850
No log 16.1333 484 0.6212 0.6805 0.6212 0.7881
No log 16.2 486 0.6263 0.6528 0.6263 0.7914
No log 16.2667 488 0.6339 0.6335 0.6339 0.7962
No log 16.3333 490 0.6419 0.6325 0.6419 0.8012
No log 16.4 492 0.6612 0.6043 0.6612 0.8131
No log 16.4667 494 0.6604 0.6198 0.6604 0.8126
No log 16.5333 496 0.6565 0.6014 0.6565 0.8102
No log 16.6 498 0.6443 0.6177 0.6443 0.8027
0.2808 16.6667 500 0.6332 0.6157 0.6332 0.7957
0.2808 16.7333 502 0.6190 0.7074 0.6190 0.7868
0.2808 16.8 504 0.6262 0.5783 0.6262 0.7914
0.2808 16.8667 506 0.6299 0.5229 0.6299 0.7937
0.2808 16.9333 508 0.6121 0.5841 0.6121 0.7824
0.2808 17.0 510 0.5994 0.7041 0.5994 0.7742
0.2808 17.0667 512 0.6029 0.6610 0.6029 0.7765
0.2808 17.1333 514 0.6132 0.6745 0.6132 0.7830
0.2808 17.2 516 0.6274 0.6390 0.6274 0.7921
0.2808 17.2667 518 0.6781 0.4604 0.6781 0.8235
0.2808 17.3333 520 0.7543 0.4109 0.7543 0.8685
0.2808 17.4 522 0.7820 0.4008 0.7820 0.8843
0.2808 17.4667 524 0.7633 0.3990 0.7633 0.8737
0.2808 17.5333 526 0.7149 0.4484 0.7149 0.8455

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k9_task5_organization

Finetuned
(4019)
this model