ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9054
  • Qwk: 0.2184
  • Mse: 0.9054
  • Rmse: 0.9515

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0435 2 3.3798 -0.0160 3.3798 1.8384
No log 0.0870 4 1.6559 -0.0370 1.6559 1.2868
No log 0.1304 6 1.2114 0.0376 1.2114 1.1006
No log 0.1739 8 0.5823 0.0909 0.5823 0.7631
No log 0.2174 10 0.6536 -0.0853 0.6536 0.8084
No log 0.2609 12 0.7214 -0.1556 0.7214 0.8494
No log 0.3043 14 0.8116 -0.2327 0.8116 0.9009
No log 0.3478 16 0.6862 -0.0732 0.6862 0.8284
No log 0.3913 18 0.6793 -0.0732 0.6793 0.8242
No log 0.4348 20 0.6145 0.0 0.6145 0.7839
No log 0.4783 22 0.5901 0.0 0.5901 0.7682
No log 0.5217 24 0.6121 0.0 0.6121 0.7824
No log 0.5652 26 0.6341 0.0 0.6341 0.7963
No log 0.6087 28 0.6205 0.0 0.6205 0.7877
No log 0.6522 30 0.6141 0.0 0.6141 0.7836
No log 0.6957 32 0.6581 -0.0853 0.6581 0.8112
No log 0.7391 34 0.6647 -0.0909 0.6647 0.8153
No log 0.7826 36 0.6006 0.0222 0.6006 0.7750
No log 0.8261 38 0.6973 0.2000 0.6973 0.8350
No log 0.8696 40 0.7037 0.1746 0.7037 0.8388
No log 0.9130 42 0.7602 0.1590 0.7602 0.8719
No log 0.9565 44 0.6436 0.1905 0.6436 0.8022
No log 1.0 46 0.6218 -0.0233 0.6218 0.7885
No log 1.0435 48 0.7242 0.0 0.7242 0.8510
No log 1.0870 50 0.7466 -0.2148 0.7466 0.8641
No log 1.1304 52 0.7839 -0.2267 0.7839 0.8854
No log 1.1739 54 0.8242 -0.1462 0.8242 0.9078
No log 1.2174 56 0.7673 -0.0864 0.7673 0.8760
No log 1.2609 58 0.6726 0.0933 0.6726 0.8201
No log 1.3043 60 0.6470 0.1484 0.6470 0.8044
No log 1.3478 62 0.6218 -0.0286 0.6218 0.7885
No log 1.3913 64 0.6303 0.0638 0.6303 0.7939
No log 1.4348 66 0.6654 0.0811 0.6654 0.8157
No log 1.4783 68 0.7195 0.2000 0.7195 0.8482
No log 1.5217 70 0.5637 0.0949 0.5637 0.7508
No log 1.5652 72 0.7042 0.2258 0.7042 0.8392
No log 1.6087 74 0.6696 0.2967 0.6696 0.8183
No log 1.6522 76 0.6270 0.2105 0.6270 0.7919
No log 1.6957 78 0.7817 0.0576 0.7817 0.8841
No log 1.7391 80 1.2736 0.0541 1.2736 1.1285
No log 1.7826 82 1.2521 0.1304 1.2521 1.1190
No log 1.8261 84 0.7707 0.1209 0.7707 0.8779
No log 1.8696 86 0.7506 0.0734 0.7506 0.8664
No log 1.9130 88 0.7243 0.0222 0.7243 0.8510
No log 1.9565 90 0.6323 0.0345 0.6323 0.7952
No log 2.0 92 0.8472 0.1429 0.8472 0.9205
No log 2.0435 94 1.4412 0.1662 1.4412 1.2005
No log 2.0870 96 1.1836 0.2054 1.1836 1.0879
No log 2.1304 98 0.7167 0.1045 0.7167 0.8466
No log 2.1739 100 0.6476 0.3073 0.6476 0.8048
No log 2.2174 102 0.6312 0.3708 0.6312 0.7945
No log 2.2609 104 0.8327 0.1765 0.8327 0.9125
No log 2.3043 106 1.4630 0.1342 1.4630 1.2096
No log 2.3478 108 1.2677 0.1373 1.2677 1.1259
No log 2.3913 110 0.8106 0.1753 0.8106 0.9003
No log 2.4348 112 0.8331 0.1373 0.8331 0.9127
No log 2.4783 114 0.7866 0.1005 0.7866 0.8869
No log 2.5217 116 0.6726 0.2941 0.6726 0.8201
No log 2.5652 118 0.6858 0.2079 0.6858 0.8281
No log 2.6087 120 0.6430 0.3297 0.6430 0.8019
No log 2.6522 122 0.7481 0.2982 0.7481 0.8649
No log 2.6957 124 1.5113 0.1395 1.5113 1.2294
No log 2.7391 126 1.8583 0.1190 1.8583 1.3632
No log 2.7826 128 1.2393 0.2258 1.2393 1.1132
No log 2.8261 130 0.7306 0.2618 0.7306 0.8548
No log 2.8696 132 0.6231 0.2542 0.6231 0.7894
No log 2.9130 134 0.6303 0.28 0.6303 0.7939
No log 2.9565 136 0.9497 0.2000 0.9497 0.9745
No log 3.0 138 1.3473 0.2050 1.3473 1.1608
No log 3.0435 140 1.1519 0.2323 1.1519 1.0733
No log 3.0870 142 0.7102 0.1746 0.7102 0.8428
No log 3.1304 144 0.5922 0.0728 0.5922 0.7696
No log 3.1739 146 0.6389 0.2265 0.6389 0.7993
No log 3.2174 148 0.6720 0.2842 0.6720 0.8198
No log 3.2609 150 0.6618 0.2609 0.6618 0.8135
No log 3.3043 152 0.7047 0.2965 0.7047 0.8394
No log 3.3478 154 0.7866 0.25 0.7866 0.8869
No log 3.3913 156 0.6739 0.2941 0.6739 0.8209
No log 3.4348 158 0.5497 0.3735 0.5497 0.7414
No log 3.4783 160 0.5792 0.4220 0.5792 0.7611
No log 3.5217 162 0.5712 0.4286 0.5712 0.7558
No log 3.5652 164 0.7663 0.3305 0.7663 0.8754
No log 3.6087 166 1.0160 0.1729 1.0160 1.0080
No log 3.6522 168 0.8785 0.1942 0.8785 0.9373
No log 3.6957 170 0.6233 0.4175 0.6233 0.7895
No log 3.7391 172 0.5809 0.3927 0.5809 0.7622
No log 3.7826 174 0.6874 0.3628 0.6874 0.8291
No log 3.8261 176 0.6483 0.4123 0.6483 0.8051
No log 3.8696 178 0.7650 0.3277 0.7650 0.8747
No log 3.9130 180 1.1678 0.2000 1.1678 1.0806
No log 3.9565 182 1.3130 0.1315 1.3130 1.1459
No log 4.0 184 1.0851 0.2000 1.0851 1.0417
No log 4.0435 186 0.7655 0.2000 0.7655 0.8749
No log 4.0870 188 0.7157 0.3333 0.7157 0.8460
No log 4.1304 190 0.8450 0.1746 0.8450 0.9192
No log 4.1739 192 1.1014 0.2258 1.1014 1.0495
No log 4.2174 194 1.0495 0.2258 1.0495 1.0245
No log 4.2609 196 1.0567 0.2258 1.0567 1.0280
No log 4.3043 198 1.1787 0.2254 1.1787 1.0857
No log 4.3478 200 1.1649 0.2308 1.1649 1.0793
No log 4.3913 202 1.0014 0.2058 1.0014 1.0007
No log 4.4348 204 1.0247 0.2276 1.0247 1.0123
No log 4.4783 206 0.8145 0.1464 0.8145 0.9025
No log 4.5217 208 0.7980 0.2239 0.7980 0.8933
No log 4.5652 210 0.8271 0.2527 0.8271 0.9095
No log 4.6087 212 1.1268 0.2000 1.1268 1.0615
No log 4.6522 214 1.2547 0.1656 1.2547 1.1201
No log 4.6957 216 1.4511 0.1236 1.4511 1.2046
No log 4.7391 218 1.2066 0.1656 1.2066 1.0985
No log 4.7826 220 0.7520 0.1071 0.7520 0.8672
No log 4.8261 222 0.6014 0.35 0.6014 0.7755
No log 4.8696 224 0.6527 0.2165 0.6527 0.8079
No log 4.9130 226 0.7175 0.2153 0.7175 0.8471
No log 4.9565 228 0.7278 0.3116 0.7278 0.8531
No log 5.0 230 0.9865 0.1773 0.9865 0.9932
No log 5.0435 232 1.6259 0.0983 1.6259 1.2751
No log 5.0870 234 1.9078 0.1186 1.9078 1.3812
No log 5.1304 236 1.6031 0.0997 1.6031 1.2661
No log 5.1739 238 1.0740 0.1270 1.0740 1.0364
No log 5.2174 240 0.8437 0.2000 0.8437 0.9185
No log 5.2609 242 0.8591 0.2000 0.8591 0.9269
No log 5.3043 244 0.9978 0.1579 0.9978 0.9989
No log 5.3478 246 1.3427 0.1429 1.3427 1.1587
No log 5.3913 248 1.5583 0.0983 1.5583 1.2483
No log 5.4348 250 1.3259 0.1429 1.3259 1.1515
No log 5.4783 252 0.9486 0.1880 0.9486 0.9740
No log 5.5217 254 0.7221 0.1388 0.7221 0.8498
No log 5.5652 256 0.6401 0.2842 0.6401 0.8001
No log 5.6087 258 0.7104 0.1388 0.7104 0.8429
No log 5.6522 260 1.0153 0.2329 1.0153 1.0076
No log 5.6957 262 1.2327 0.2099 1.2327 1.1103
No log 5.7391 264 1.1095 0.2051 1.1095 1.0533
No log 5.7826 266 0.8131 0.2333 0.8131 0.9017
No log 5.8261 268 0.7090 0.2000 0.7090 0.8420
No log 5.8696 270 0.8043 0.1366 0.8043 0.8968
No log 5.9130 272 1.0384 0.2329 1.0384 1.0190
No log 5.9565 274 1.0438 0.2115 1.0438 1.0216
No log 6.0 276 0.9420 0.1760 0.9420 0.9705
No log 6.0435 278 0.9641 0.2000 0.9641 0.9819
No log 6.0870 280 1.0941 0.2000 1.0941 1.0460
No log 6.1304 282 1.1779 0.1801 1.1779 1.0853
No log 6.1739 284 1.1056 0.1531 1.1056 1.0515
No log 6.2174 286 1.0953 0.1268 1.0953 1.0466
No log 6.2609 288 0.9153 0.1751 0.9153 0.9567
No log 6.3043 290 0.8577 0.1746 0.8577 0.9261
No log 6.3478 292 0.8844 0.1746 0.8844 0.9404
No log 6.3913 294 0.8737 0.1741 0.8737 0.9347
No log 6.4348 296 0.9689 0.1818 0.9689 0.9843
No log 6.4783 298 0.9813 0.1533 0.9813 0.9906
No log 6.5217 300 0.9394 0.1461 0.9394 0.9692
No log 6.5652 302 0.9333 0.1496 0.9333 0.9661
No log 6.6087 304 1.0276 0.1515 1.0276 1.0137
No log 6.6522 306 1.0088 0.1818 1.0088 1.0044
No log 6.6957 308 0.9528 0.1811 0.9528 0.9761
No log 6.7391 310 0.9261 0.1873 0.9261 0.9623
No log 6.7826 312 0.9429 0.1875 0.9429 0.9710
No log 6.8261 314 1.0043 0.2239 1.0043 1.0022
No log 6.8696 316 1.1052 0.1709 1.1052 1.0513
No log 6.9130 318 1.0393 0.2234 1.0393 1.0194
No log 6.9565 320 0.9056 0.1756 0.9056 0.9516
No log 7.0 322 0.8063 0.2066 0.8063 0.8979
No log 7.0435 324 0.8872 0.1506 0.8872 0.9419
No log 7.0870 326 1.0194 0.1943 1.0194 1.0097
No log 7.1304 328 0.9828 0.1942 0.9828 0.9914
No log 7.1739 330 0.8225 0.1486 0.8225 0.9069
No log 7.2174 332 0.7820 0.2066 0.7820 0.8843
No log 7.2609 334 0.8323 0.1877 0.8323 0.9123
No log 7.3043 336 0.8223 0.2187 0.8223 0.9068
No log 7.3478 338 0.7775 0.2137 0.7775 0.8818
No log 7.3913 340 0.7592 0.1790 0.7592 0.8713
No log 7.4348 342 0.7310 0.1781 0.7310 0.8550
No log 7.4783 344 0.8472 0.2353 0.8472 0.9204
No log 7.5217 346 1.1084 0.2412 1.1084 1.0528
No log 7.5652 348 1.3589 0.1667 1.3589 1.1657
No log 7.6087 350 1.3508 0.1667 1.3508 1.1623
No log 7.6522 352 1.1633 0.2150 1.1633 1.0786
No log 7.6957 354 0.8978 0.2472 0.8978 0.9475
No log 7.7391 356 0.7872 0.1464 0.7872 0.8872
No log 7.7826 358 0.8014 0.2129 0.8014 0.8952
No log 7.8261 360 0.8789 0.2188 0.8789 0.9375
No log 7.8696 362 0.8980 0.2184 0.8980 0.9476
No log 7.9130 364 0.9170 0.2184 0.9170 0.9576
No log 7.9565 366 1.0202 0.2165 1.0202 1.0101
No log 8.0 368 1.0614 0.1945 1.0614 1.0302
No log 8.0435 370 0.9976 0.2168 0.9976 0.9988
No log 8.0870 372 1.0180 0.2165 1.0180 1.0090
No log 8.1304 374 0.9974 0.2111 0.9974 0.9987
No log 8.1739 376 0.8899 0.2416 0.8899 0.9433
No log 8.2174 378 0.7634 0.2469 0.7634 0.8737
No log 8.2609 380 0.6807 0.2727 0.6807 0.8250
No log 8.3043 382 0.6847 0.3067 0.6847 0.8274
No log 8.3478 384 0.7429 0.2137 0.7429 0.8619
No log 8.3913 386 0.8738 0.2188 0.8738 0.9347
No log 8.4348 388 0.9932 0.1945 0.9932 0.9966
No log 8.4783 390 1.0498 0.1678 1.0498 1.0246
No log 8.5217 392 1.0180 0.1945 1.0180 1.0089
No log 8.5652 394 0.9120 0.1882 0.9120 0.9550
No log 8.6087 396 0.8026 0.2129 0.8026 0.8959
No log 8.6522 398 0.7605 0.1790 0.7605 0.8721
No log 8.6957 400 0.7819 0.2134 0.7819 0.8842
No log 8.7391 402 0.8383 0.2195 0.8383 0.9156
No log 8.7826 404 0.9152 0.2180 0.9152 0.9566
No log 8.8261 406 0.9759 0.1884 0.9759 0.9879
No log 8.8696 408 0.9924 0.1886 0.9924 0.9962
No log 8.9130 410 0.9847 0.1886 0.9847 0.9923
No log 8.9565 412 0.9520 0.2180 0.9520 0.9757
No log 9.0 414 0.9252 0.2188 0.9252 0.9619
No log 9.0435 416 0.9149 0.2188 0.9149 0.9565
No log 9.0870 418 0.9490 0.2184 0.9490 0.9742
No log 9.1304 420 1.0249 0.1886 1.0249 1.0124
No log 9.1739 422 1.0583 0.1608 1.0583 1.0288
No log 9.2174 424 1.0373 0.1608 1.0373 1.0185
No log 9.2609 426 0.9951 0.2174 0.9951 0.9976
No log 9.3043 428 0.9510 0.2184 0.9510 0.9752
No log 9.3478 430 0.8966 0.2188 0.8966 0.9469
No log 9.3913 432 0.8626 0.2191 0.8626 0.9287
No log 9.4348 434 0.8479 0.2191 0.8479 0.9208
No log 9.4783 436 0.8429 0.2191 0.8429 0.9181
No log 9.5217 438 0.8497 0.2191 0.8497 0.9218
No log 9.5652 440 0.8637 0.2191 0.8637 0.9294
No log 9.6087 442 0.8761 0.2191 0.8761 0.9360
No log 9.6522 444 0.8888 0.2188 0.8888 0.9428
No log 9.6957 446 0.8981 0.2184 0.8981 0.9477
No log 9.7391 448 0.9021 0.2184 0.9021 0.9498
No log 9.7826 450 0.9033 0.2184 0.9033 0.9504
No log 9.8261 452 0.9075 0.2184 0.9075 0.9526
No log 9.8696 454 0.9100 0.2184 0.9100 0.9539
No log 9.9130 456 0.9095 0.2184 0.9095 0.9537
No log 9.9565 458 0.9065 0.2184 0.9065 0.9521
No log 10.0 460 0.9054 0.2184 0.9054 0.9515

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k9_task3_organization

Finetuned
(4023)
this model