ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k4_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7980
  • Qwk: 0.4273
  • Mse: 0.7980
  • Rmse: 0.8933

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 4.5806 0.0010 4.5806 2.1402
No log 0.2667 4 2.7165 0.0309 2.7165 1.6482
No log 0.4 6 1.9886 0.1075 1.9886 1.4102
No log 0.5333 8 1.4802 0.0372 1.4801 1.2166
No log 0.6667 10 1.3326 -0.0451 1.3326 1.1544
No log 0.8 12 1.3212 0.0473 1.3212 1.1495
No log 0.9333 14 1.3024 0.0768 1.3024 1.1412
No log 1.0667 16 1.5182 0.0936 1.5182 1.2321
No log 1.2 18 1.2614 0.0741 1.2614 1.1231
No log 1.3333 20 1.2609 0.1557 1.2609 1.1229
No log 1.4667 22 1.1827 0.3042 1.1827 1.0875
No log 1.6 24 1.1701 0.2231 1.1701 1.0817
No log 1.7333 26 1.1554 0.2966 1.1554 1.0749
No log 1.8667 28 1.2300 0.1862 1.2300 1.1091
No log 2.0 30 1.2613 0.2240 1.2613 1.1231
No log 2.1333 32 1.0967 0.4235 1.0967 1.0472
No log 2.2667 34 1.0763 0.4235 1.0763 1.0374
No log 2.4 36 1.0382 0.4671 1.0382 1.0189
No log 2.5333 38 1.2155 0.2325 1.2155 1.1025
No log 2.6667 40 1.1364 0.2143 1.1364 1.0660
No log 2.8 42 0.9996 0.4975 0.9996 0.9998
No log 2.9333 44 1.1045 0.3835 1.1045 1.0510
No log 3.0667 46 1.1828 0.3210 1.1828 1.0876
No log 3.2 48 1.4773 0.2943 1.4773 1.2154
No log 3.3333 50 1.5401 0.2815 1.5401 1.2410
No log 3.4667 52 1.0327 0.4685 1.0327 1.0162
No log 3.6 54 0.9170 0.5350 0.9170 0.9576
No log 3.7333 56 0.9371 0.5525 0.9371 0.9680
No log 3.8667 58 1.0525 0.3292 1.0525 1.0259
No log 4.0 60 1.4900 0.3571 1.4900 1.2207
No log 4.1333 62 1.8674 0.1863 1.8674 1.3665
No log 4.2667 64 1.7342 0.1887 1.7342 1.3169
No log 4.4 66 1.3467 0.3067 1.3467 1.1605
No log 4.5333 68 1.0266 0.3395 1.0266 1.0132
No log 4.6667 70 1.1634 0.3778 1.1634 1.0786
No log 4.8 72 1.2846 0.4157 1.2846 1.1334
No log 4.9333 74 1.1120 0.3884 1.1120 1.0545
No log 5.0667 76 0.9569 0.3902 0.9569 0.9782
No log 5.2 78 1.0791 0.2890 1.0791 1.0388
No log 5.3333 80 1.2384 0.2763 1.2384 1.1128
No log 5.4667 82 1.2146 0.2918 1.2146 1.1021
No log 5.6 84 0.9745 0.4516 0.9745 0.9872
No log 5.7333 86 0.9731 0.5066 0.9731 0.9865
No log 5.8667 88 1.1241 0.5125 1.1241 1.0602
No log 6.0 90 1.0850 0.5125 1.0850 1.0416
No log 6.1333 92 0.9214 0.5556 0.9214 0.9599
No log 6.2667 94 1.0205 0.4032 1.0205 1.0102
No log 6.4 96 1.3222 0.2752 1.3222 1.1499
No log 6.5333 98 1.3128 0.2752 1.3128 1.1458
No log 6.6667 100 1.0159 0.3873 1.0159 1.0079
No log 6.8 102 0.9503 0.5192 0.9503 0.9748
No log 6.9333 104 1.0602 0.4505 1.0602 1.0296
No log 7.0667 106 1.1304 0.5105 1.1304 1.0632
No log 7.2 108 1.0102 0.4973 1.0102 1.0051
No log 7.3333 110 0.8690 0.3998 0.8690 0.9322
No log 7.4667 112 1.0353 0.4574 1.0353 1.0175
No log 7.6 114 1.3306 0.4139 1.3306 1.1535
No log 7.7333 116 1.2693 0.4182 1.2693 1.1266
No log 7.8667 118 0.9472 0.4584 0.9472 0.9732
No log 8.0 120 0.8856 0.5871 0.8856 0.9411
No log 8.1333 122 1.1440 0.5341 1.1440 1.0696
No log 8.2667 124 1.2036 0.5167 1.2036 1.0971
No log 8.4 126 1.0722 0.5304 1.0722 1.0355
No log 8.5333 128 0.9229 0.5209 0.9229 0.9607
No log 8.6667 130 0.8108 0.5754 0.8108 0.9005
No log 8.8 132 0.8043 0.6366 0.8043 0.8968
No log 8.9333 134 0.8368 0.5848 0.8368 0.9148
No log 9.0667 136 0.8405 0.5955 0.8405 0.9168
No log 9.2 138 0.7394 0.6622 0.7394 0.8599
No log 9.3333 140 0.7965 0.5537 0.7965 0.8924
No log 9.4667 142 0.9012 0.5384 0.9012 0.9493
No log 9.6 144 0.8974 0.5384 0.8974 0.9473
No log 9.7333 146 0.7901 0.5500 0.7901 0.8889
No log 9.8667 148 0.7504 0.5993 0.7504 0.8663
No log 10.0 150 0.7858 0.6160 0.7858 0.8865
No log 10.1333 152 0.8059 0.5632 0.8059 0.8977
No log 10.2667 154 0.8013 0.5621 0.8013 0.8952
No log 10.4 156 0.8443 0.5068 0.8443 0.9189
No log 10.5333 158 0.8562 0.5014 0.8562 0.9253
No log 10.6667 160 0.8498 0.5039 0.8498 0.9218
No log 10.8 162 0.8089 0.5587 0.8089 0.8994
No log 10.9333 164 0.8152 0.5542 0.8152 0.9029
No log 11.0667 166 0.8802 0.5572 0.8802 0.9382
No log 11.2 168 0.7848 0.5806 0.7848 0.8859
No log 11.3333 170 0.8109 0.5321 0.8109 0.9005
No log 11.4667 172 1.1304 0.4482 1.1304 1.0632
No log 11.6 174 1.2739 0.3836 1.2739 1.1287
No log 11.7333 176 1.0301 0.4458 1.0301 1.0149
No log 11.8667 178 0.8373 0.4941 0.8373 0.9151
No log 12.0 180 0.8310 0.5245 0.8310 0.9116
No log 12.1333 182 0.8259 0.5359 0.8259 0.9088
No log 12.2667 184 0.8300 0.4916 0.8300 0.9110
No log 12.4 186 0.8356 0.4672 0.8356 0.9141
No log 12.5333 188 0.8249 0.4488 0.8249 0.9083
No log 12.6667 190 0.8400 0.4624 0.8400 0.9165
No log 12.8 192 0.9691 0.4849 0.9691 0.9844
No log 12.9333 194 0.8982 0.4191 0.8982 0.9477
No log 13.0667 196 0.8001 0.6203 0.8001 0.8945
No log 13.2 198 0.8924 0.5173 0.8924 0.9447
No log 13.3333 200 0.9221 0.4985 0.9221 0.9603
No log 13.4667 202 0.8691 0.4340 0.8691 0.9323
No log 13.6 204 0.8339 0.4872 0.8339 0.9132
No log 13.7333 206 0.9478 0.4518 0.9478 0.9736
No log 13.8667 208 1.0188 0.4088 1.0188 1.0093
No log 14.0 210 0.9140 0.4882 0.9140 0.9560
No log 14.1333 212 0.8457 0.4837 0.8457 0.9196
No log 14.2667 214 0.8533 0.5094 0.8533 0.9237
No log 14.4 216 0.7905 0.5964 0.7905 0.8891
No log 14.5333 218 0.7359 0.5432 0.7359 0.8579
No log 14.6667 220 0.7835 0.5451 0.7835 0.8851
No log 14.8 222 0.7970 0.5029 0.7970 0.8928
No log 14.9333 224 0.7848 0.5044 0.7848 0.8859
No log 15.0667 226 0.7634 0.5662 0.7634 0.8737
No log 15.2 228 0.7809 0.5326 0.7809 0.8837
No log 15.3333 230 0.7781 0.6100 0.7781 0.8821
No log 15.4667 232 0.7926 0.5740 0.7926 0.8903
No log 15.6 234 0.8124 0.5766 0.8124 0.9014
No log 15.7333 236 0.8093 0.5836 0.8093 0.8996
No log 15.8667 238 0.8540 0.5493 0.8540 0.9241
No log 16.0 240 0.9007 0.4617 0.9007 0.9491
No log 16.1333 242 0.8749 0.5156 0.8749 0.9354
No log 16.2667 244 0.8019 0.5239 0.8019 0.8955
No log 16.4 246 0.7636 0.4962 0.7636 0.8738
No log 16.5333 248 0.7866 0.5190 0.7866 0.8869
No log 16.6667 250 0.8055 0.5766 0.8055 0.8975
No log 16.8 252 0.7699 0.6067 0.7699 0.8774
No log 16.9333 254 0.7541 0.5691 0.7541 0.8684
No log 17.0667 256 0.8054 0.5357 0.8054 0.8974
No log 17.2 258 0.8376 0.5065 0.8376 0.9152
No log 17.3333 260 0.7934 0.5110 0.7934 0.8907
No log 17.4667 262 0.7872 0.5102 0.7872 0.8872
No log 17.6 264 0.8194 0.4757 0.8194 0.9052
No log 17.7333 266 0.8313 0.4726 0.8313 0.9117
No log 17.8667 268 0.8374 0.5098 0.8374 0.9151
No log 18.0 270 0.8153 0.5611 0.8153 0.9029
No log 18.1333 272 0.8169 0.6035 0.8169 0.9038
No log 18.2667 274 0.8393 0.6037 0.8393 0.9161
No log 18.4 276 0.8354 0.5620 0.8354 0.9140
No log 18.5333 278 0.8563 0.5470 0.8563 0.9254
No log 18.6667 280 0.8438 0.5425 0.8438 0.9186
No log 18.8 282 0.7901 0.5607 0.7901 0.8889
No log 18.9333 284 0.7493 0.6246 0.7493 0.8656
No log 19.0667 286 0.7588 0.5543 0.7588 0.8711
No log 19.2 288 0.7809 0.4915 0.7809 0.8837
No log 19.3333 290 0.8217 0.4733 0.8217 0.9065
No log 19.4667 292 0.8331 0.4765 0.8331 0.9128
No log 19.6 294 0.8075 0.5219 0.8075 0.8986
No log 19.7333 296 0.7362 0.6544 0.7362 0.8580
No log 19.8667 298 0.7220 0.6289 0.7220 0.8497
No log 20.0 300 0.8434 0.5455 0.8434 0.9184
No log 20.1333 302 0.9850 0.5388 0.9850 0.9925
No log 20.2667 304 0.9802 0.5246 0.9802 0.9900
No log 20.4 306 0.8588 0.4811 0.8588 0.9267
No log 20.5333 308 0.8129 0.5406 0.8129 0.9016
No log 20.6667 310 0.7790 0.5828 0.7790 0.8826
No log 20.8 312 0.7685 0.6211 0.7685 0.8766
No log 20.9333 314 0.7993 0.5718 0.7993 0.8941
No log 21.0667 316 0.8091 0.5784 0.8091 0.8995
No log 21.2 318 0.7462 0.6127 0.7462 0.8638
No log 21.3333 320 0.7209 0.5949 0.7209 0.8491
No log 21.4667 322 0.8145 0.6151 0.8145 0.9025
No log 21.6 324 0.8463 0.5649 0.8463 0.9200
No log 21.7333 326 0.8031 0.5447 0.8031 0.8962
No log 21.8667 328 0.8076 0.5560 0.8076 0.8987
No log 22.0 330 0.8184 0.5536 0.8184 0.9047
No log 22.1333 332 0.8735 0.5384 0.8735 0.9346
No log 22.2667 334 0.8822 0.5468 0.8822 0.9393
No log 22.4 336 0.8154 0.6029 0.8154 0.9030
No log 22.5333 338 0.7374 0.6151 0.7374 0.8587
No log 22.6667 340 0.7293 0.6137 0.7293 0.8540
No log 22.8 342 0.7392 0.6137 0.7392 0.8598
No log 22.9333 344 0.7300 0.6517 0.7300 0.8544
No log 23.0667 346 0.7306 0.6001 0.7306 0.8547
No log 23.2 348 0.7476 0.5979 0.7476 0.8646
No log 23.3333 350 0.8112 0.5661 0.8112 0.9007
No log 23.4667 352 0.8286 0.5443 0.8286 0.9103
No log 23.6 354 0.8117 0.4858 0.8117 0.9010
No log 23.7333 356 0.7804 0.4711 0.7804 0.8834
No log 23.8667 358 0.7799 0.4962 0.7799 0.8831
No log 24.0 360 0.8611 0.4715 0.8611 0.9280
No log 24.1333 362 0.9256 0.5098 0.9256 0.9621
No log 24.2667 364 0.8939 0.4764 0.8939 0.9455
No log 24.4 366 0.8018 0.5611 0.8018 0.8954
No log 24.5333 368 0.8202 0.5988 0.8202 0.9057
No log 24.6667 370 0.8666 0.5766 0.8666 0.9309
No log 24.8 372 0.9163 0.5340 0.9163 0.9573
No log 24.9333 374 0.9642 0.4933 0.9642 0.9819
No log 25.0667 376 1.0587 0.5206 1.0587 1.0289
No log 25.2 378 1.0648 0.5206 1.0648 1.0319
No log 25.3333 380 0.9722 0.4933 0.9722 0.9860
No log 25.4667 382 0.8730 0.5624 0.8730 0.9343
No log 25.6 384 0.8257 0.5673 0.8257 0.9087
No log 25.7333 386 0.8259 0.5040 0.8259 0.9088
No log 25.8667 388 0.8313 0.5232 0.8313 0.9117
No log 26.0 390 0.8334 0.5886 0.8334 0.9129
No log 26.1333 392 0.8323 0.5886 0.8323 0.9123
No log 26.2667 394 0.8402 0.5858 0.8402 0.9166
No log 26.4 396 0.8244 0.5858 0.8244 0.9080
No log 26.5333 398 0.8017 0.5571 0.8017 0.8954
No log 26.6667 400 0.7975 0.4738 0.7975 0.8930
No log 26.8 402 0.8093 0.4485 0.8093 0.8996
No log 26.9333 404 0.8084 0.4219 0.8084 0.8991
No log 27.0667 406 0.7975 0.4609 0.7975 0.8930
No log 27.2 408 0.7976 0.4634 0.7976 0.8931
No log 27.3333 410 0.8049 0.4841 0.8049 0.8972
No log 27.4667 412 0.8177 0.5413 0.8177 0.9043
No log 27.6 414 0.8232 0.5238 0.8232 0.9073
No log 27.7333 416 0.8276 0.5887 0.8276 0.9097
No log 27.8667 418 0.8228 0.6047 0.8228 0.9071
No log 28.0 420 0.8259 0.5621 0.8259 0.9088
No log 28.1333 422 0.8224 0.5316 0.8224 0.9068
No log 28.2667 424 0.8277 0.5562 0.8277 0.9098
No log 28.4 426 0.8215 0.5359 0.8215 0.9064
No log 28.5333 428 0.8164 0.4983 0.8164 0.9036
No log 28.6667 430 0.8257 0.5359 0.8257 0.9087
No log 28.8 432 0.8507 0.5365 0.8507 0.9223
No log 28.9333 434 0.8545 0.5556 0.8545 0.9244
No log 29.0667 436 0.8066 0.5860 0.8066 0.8981
No log 29.2 438 0.7668 0.5858 0.7668 0.8757
No log 29.3333 440 0.7654 0.5858 0.7654 0.8748
No log 29.4667 442 0.7711 0.5564 0.7711 0.8781
No log 29.6 444 0.7689 0.5564 0.7689 0.8768
No log 29.7333 446 0.7639 0.5611 0.7639 0.8740
No log 29.8667 448 0.7611 0.5914 0.7611 0.8724
No log 30.0 450 0.7654 0.5458 0.7654 0.8749
No log 30.1333 452 0.7723 0.5043 0.7723 0.8788
No log 30.2667 454 0.7681 0.4181 0.7681 0.8764
No log 30.4 456 0.7524 0.4444 0.7524 0.8674
No log 30.5333 458 0.7391 0.5596 0.7391 0.8597
No log 30.6667 460 0.7392 0.6088 0.7392 0.8597
No log 30.8 462 0.7438 0.6151 0.7438 0.8625
No log 30.9333 464 0.7753 0.6151 0.7753 0.8805
No log 31.0667 466 0.7618 0.6151 0.7618 0.8728
No log 31.2 468 0.7562 0.6151 0.7562 0.8696
No log 31.3333 470 0.7536 0.6252 0.7536 0.8681
No log 31.4667 472 0.7628 0.6079 0.7628 0.8734
No log 31.6 474 0.7814 0.5683 0.7814 0.8840
No log 31.7333 476 0.7823 0.5975 0.7823 0.8845
No log 31.8667 478 0.7503 0.5975 0.7503 0.8662
No log 32.0 480 0.7044 0.5971 0.7044 0.8393
No log 32.1333 482 0.6848 0.6517 0.6848 0.8276
No log 32.2667 484 0.7035 0.5811 0.7035 0.8388
No log 32.4 486 0.7132 0.5632 0.7132 0.8445
No log 32.5333 488 0.6921 0.5489 0.6921 0.8319
No log 32.6667 490 0.6951 0.5722 0.6951 0.8338
No log 32.8 492 0.7530 0.5706 0.7530 0.8678
No log 32.9333 494 0.8267 0.6008 0.8267 0.9092
No log 33.0667 496 0.8651 0.5753 0.8651 0.9301
No log 33.2 498 0.8168 0.6008 0.8168 0.9038
0.2995 33.3333 500 0.7311 0.5920 0.7311 0.8551
0.2995 33.4667 502 0.6867 0.5741 0.6867 0.8287
0.2995 33.6 504 0.6991 0.5660 0.6991 0.8361
0.2995 33.7333 506 0.7149 0.5660 0.7149 0.8455
0.2995 33.8667 508 0.6991 0.6061 0.6991 0.8361
0.2995 34.0 510 0.6925 0.6287 0.6925 0.8322
0.2995 34.1333 512 0.7019 0.6565 0.7019 0.8378
0.2995 34.2667 514 0.7245 0.5836 0.7245 0.8512
0.2995 34.4 516 0.7496 0.5539 0.7496 0.8658
0.2995 34.5333 518 0.7667 0.5562 0.7667 0.8756
0.2995 34.6667 520 0.7773 0.4482 0.7773 0.8816
0.2995 34.8 522 0.7980 0.4273 0.7980 0.8933

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k4_task2_organization

Finetuned
(4023)
this model