ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k12_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8385
  • Qwk: 0.2661
  • Mse: 0.8385
  • Rmse: 0.9157

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0351 2 3.4914 -0.0066 3.4914 1.8685
No log 0.0702 4 2.4748 -0.0563 2.4748 1.5732
No log 0.1053 6 1.5214 0.0255 1.5214 1.2335
No log 0.1404 8 1.0574 0.0588 1.0574 1.0283
No log 0.1754 10 0.6325 0.0476 0.6325 0.7953
No log 0.2105 12 0.6009 0.0569 0.6009 0.7752
No log 0.2456 14 0.6117 0.0 0.6117 0.7821
No log 0.2807 16 0.6115 0.0 0.6115 0.7820
No log 0.3158 18 0.6416 -0.0732 0.6416 0.8010
No log 0.3509 20 0.7242 0.0899 0.7242 0.8510
No log 0.3860 22 0.6870 0.1813 0.6870 0.8289
No log 0.4211 24 0.6002 0.0569 0.6002 0.7747
No log 0.4561 26 0.6133 0.1304 0.6133 0.7832
No log 0.4912 28 0.8517 0.1289 0.8517 0.9229
No log 0.5263 30 1.0885 -0.0268 1.0885 1.0433
No log 0.5614 32 0.7407 -0.0476 0.7407 0.8606
No log 0.5965 34 0.6795 0.0071 0.6795 0.8243
No log 0.6316 36 0.6240 0.0569 0.6240 0.7900
No log 0.6667 38 0.6038 0.0 0.6038 0.7770
No log 0.7018 40 0.6557 0.1111 0.6557 0.8098
No log 0.7368 42 0.7670 0.1228 0.7670 0.8758
No log 0.7719 44 0.9197 0.0345 0.9197 0.9590
No log 0.8070 46 0.8976 0.0745 0.8976 0.9474
No log 0.8421 48 0.8831 0.0745 0.8831 0.9397
No log 0.8772 50 0.7750 0.1392 0.7750 0.8803
No log 0.9123 52 0.6850 0.2000 0.6850 0.8276
No log 0.9474 54 0.5938 -0.0303 0.5938 0.7706
No log 0.9825 56 0.5879 -0.0303 0.5879 0.7667
No log 1.0175 58 0.6004 -0.0303 0.6004 0.7749
No log 1.0526 60 0.6180 0.0222 0.6180 0.7861
No log 1.0877 62 0.6525 -0.0435 0.6525 0.8078
No log 1.1228 64 0.7838 0.0359 0.7838 0.8853
No log 1.1579 66 1.1660 0.0617 1.1660 1.0798
No log 1.1930 68 1.2749 0.0078 1.2749 1.1291
No log 1.2281 70 0.8813 0.0991 0.8813 0.9388
No log 1.2632 72 0.6739 -0.0314 0.6739 0.8209
No log 1.2982 74 0.6091 0.0725 0.6091 0.7805
No log 1.3333 76 0.6082 0.1111 0.6082 0.7799
No log 1.3684 78 0.6699 0.1475 0.6699 0.8185
No log 1.4035 80 0.9641 0.1347 0.9641 0.9819
No log 1.4386 82 0.9683 0.1405 0.9683 0.9840
No log 1.4737 84 0.7157 0.0476 0.7157 0.8460
No log 1.5088 86 0.6085 -0.0233 0.6085 0.7800
No log 1.5439 88 0.6595 -0.0081 0.6595 0.8121
No log 1.5789 90 0.6198 -0.0081 0.6198 0.7872
No log 1.6140 92 0.5617 -0.0233 0.5617 0.7495
No log 1.6491 94 0.6643 0.1323 0.6643 0.8150
No log 1.6842 96 0.8065 0.1515 0.8065 0.8980
No log 1.7193 98 0.7710 0.1579 0.7710 0.8781
No log 1.7544 100 0.5980 0.1282 0.5980 0.7733
No log 1.7895 102 0.5464 0.0625 0.5464 0.7392
No log 1.8246 104 0.6060 0.1329 0.6060 0.7785
No log 1.8596 106 0.6091 0.1329 0.6091 0.7804
No log 1.8947 108 0.5622 0.1045 0.5622 0.7498
No log 1.9298 110 0.6028 0.1445 0.6028 0.7764
No log 1.9649 112 0.6033 0.1807 0.6033 0.7767
No log 2.0 114 0.6699 0.1605 0.6699 0.8185
No log 2.0351 116 0.8348 0.1071 0.8348 0.9137
No log 2.0702 118 0.7496 0.2527 0.7496 0.8658
No log 2.1053 120 0.6342 0.2941 0.6342 0.7963
No log 2.1404 122 0.6133 0.1628 0.6133 0.7831
No log 2.1754 124 0.6226 0.2832 0.6226 0.7891
No log 2.2105 126 0.5940 0.2000 0.5940 0.7707
No log 2.2456 128 0.7683 0.3118 0.7683 0.8765
No log 2.2807 130 0.7932 0.2315 0.7932 0.8906
No log 2.3158 132 0.6436 0.1807 0.6436 0.8022
No log 2.3509 134 0.6808 0.2281 0.6808 0.8251
No log 2.3860 136 0.6891 0.2265 0.6891 0.8301
No log 2.4211 138 0.6364 0.2265 0.6364 0.7978
No log 2.4561 140 0.7353 0.2637 0.7353 0.8575
No log 2.4912 142 1.0845 0.1389 1.0845 1.0414
No log 2.5263 144 1.1685 0.0621 1.1685 1.0810
No log 2.5614 146 0.8321 0.2212 0.8321 0.9122
No log 2.5965 148 0.6193 0.3623 0.6193 0.7870
No log 2.6316 150 0.7550 0.2356 0.7550 0.8689
No log 2.6667 152 0.6398 0.3797 0.6398 0.7999
No log 2.7018 154 0.6526 0.1138 0.6526 0.8078
No log 2.7368 156 0.8819 0.2315 0.8819 0.9391
No log 2.7719 158 1.0662 0.1318 1.0662 1.0326
No log 2.8070 160 0.9174 0.0901 0.9174 0.9578
No log 2.8421 162 0.6784 0.2174 0.6784 0.8236
No log 2.8772 164 0.7425 0.2372 0.7425 0.8617
No log 2.9123 166 0.7707 0.2881 0.7707 0.8779
No log 2.9474 168 0.7433 0.2222 0.7433 0.8621
No log 2.9825 170 0.8464 0.1636 0.8464 0.9200
No log 3.0175 172 0.7138 0.2523 0.7138 0.8448
No log 3.0526 174 0.7386 0.1636 0.7386 0.8594
No log 3.0877 176 0.8627 0.1864 0.8627 0.9288
No log 3.1228 178 0.9679 0.1875 0.9679 0.9838
No log 3.1579 180 1.0356 0.1880 1.0356 1.0176
No log 3.1930 182 0.7648 0.2074 0.7648 0.8745
No log 3.2281 184 0.6108 0.3407 0.6108 0.7815
No log 3.2632 186 0.6049 0.3407 0.6049 0.7778
No log 3.2982 188 0.6535 0.1917 0.6535 0.8084
No log 3.3333 190 0.7753 0.1845 0.7753 0.8805
No log 3.3684 192 1.1331 0.1027 1.1331 1.0645
No log 3.4035 194 1.1397 0.1886 1.1397 1.0675
No log 3.4386 196 0.8156 0.2759 0.8156 0.9031
No log 3.4737 198 0.7618 0.2212 0.7618 0.8728
No log 3.5088 200 0.8144 0.2696 0.8144 0.9024
No log 3.5439 202 0.8019 0.2653 0.8019 0.8955
No log 3.5789 204 0.8847 0.2569 0.8847 0.9406
No log 3.6140 206 0.9945 0.1811 0.9945 0.9972
No log 3.6491 208 0.8415 0.2829 0.8415 0.9173
No log 3.6842 210 0.7279 0.3103 0.7279 0.8532
No log 3.7193 212 0.6309 0.3978 0.6309 0.7943
No log 3.7544 214 0.6165 0.5080 0.6165 0.7852
No log 3.7895 216 0.7415 0.1481 0.7415 0.8611
No log 3.8246 218 0.9397 0.2195 0.9397 0.9694
No log 3.8596 220 0.8955 0.2520 0.8955 0.9463
No log 3.8947 222 0.7779 0.2566 0.7779 0.8820
No log 3.9298 224 0.8088 0.2129 0.8088 0.8993
No log 3.9649 226 1.1309 0.1888 1.1309 1.0634
No log 4.0 228 1.2103 0.2162 1.2103 1.1001
No log 4.0351 230 0.8493 0.1870 0.8493 0.9216
No log 4.0702 232 0.7685 0.2000 0.7685 0.8766
No log 4.1053 234 0.8315 0.1867 0.8315 0.9118
No log 4.1404 236 0.8707 0.1870 0.8707 0.9331
No log 4.1754 238 1.0659 0.2174 1.0659 1.0324
No log 4.2105 240 1.0964 0.2174 1.0964 1.0471
No log 4.2456 242 0.8200 0.2479 0.8200 0.9055
No log 4.2807 244 0.8069 0.2469 0.8069 0.8983
No log 4.3158 246 1.0436 0.1587 1.0436 1.0216
No log 4.3509 248 1.1411 0.0853 1.1411 1.0682
No log 4.3860 250 1.1177 0.0556 1.1177 1.0572
No log 4.4211 252 0.8255 0.3010 0.8255 0.9086
No log 4.4561 254 0.6892 0.2941 0.6892 0.8302
No log 4.4912 256 0.7202 0.3333 0.7202 0.8486
No log 4.5263 258 0.8599 0.2000 0.8599 0.9273
No log 4.5614 260 1.2490 0.0307 1.2490 1.1176
No log 4.5965 262 1.6969 0.0545 1.6969 1.3027
No log 4.6316 264 1.7133 0.0545 1.7133 1.3089
No log 4.6667 266 1.3756 0.0868 1.3756 1.1729
No log 4.7018 268 0.9400 0.1660 0.9400 0.9695
No log 4.7368 270 0.8435 0.2579 0.8435 0.9184
No log 4.7719 272 0.9466 0.2490 0.9466 0.9729
No log 4.8070 274 1.1084 0.1942 1.1084 1.0528
No log 4.8421 276 1.3996 0.0588 1.3996 1.1830
No log 4.8772 278 1.6024 0.0893 1.6024 1.2659
No log 4.9123 280 1.5517 0.0877 1.5517 1.2457
No log 4.9474 282 1.1531 0.0949 1.1531 1.0738
No log 4.9825 284 0.8142 0.2780 0.8142 0.9023
No log 5.0175 286 0.7920 0.2442 0.7920 0.8900
No log 5.0526 288 0.7607 0.2410 0.7607 0.8722
No log 5.0877 290 0.8249 0.2676 0.8249 0.9082
No log 5.1228 292 1.0373 0.0303 1.0373 1.0185
No log 5.1579 294 1.1401 0.0435 1.1401 1.0678
No log 5.1930 296 1.0288 0.1165 1.0288 1.0143
No log 5.2281 298 0.8749 0.1928 0.8749 0.9353
No log 5.2632 300 0.8437 0.1776 0.8437 0.9185
No log 5.2982 302 0.9017 0.1560 0.9017 0.9496
No log 5.3333 304 1.1003 0.1450 1.1003 1.0490
No log 5.3684 306 1.4351 0.0868 1.4351 1.1980
No log 5.4035 308 1.4310 0.1068 1.4310 1.1963
No log 5.4386 310 1.1101 0.1506 1.1101 1.0536
No log 5.4737 312 0.8714 0.2681 0.8714 0.9335
No log 5.5088 314 0.8776 0.2397 0.8776 0.9368
No log 5.5439 316 0.7974 0.2963 0.7974 0.8930
No log 5.5789 318 0.7649 0.3116 0.7649 0.8746
No log 5.6140 320 0.8818 0.1736 0.8818 0.9391
No log 5.6491 322 0.9795 0.1212 0.9795 0.9897
No log 5.6842 324 1.0277 0.1331 1.0277 1.0137
No log 5.7193 326 0.8927 0.1111 0.8927 0.9448
No log 5.7544 328 0.7945 0.2661 0.7945 0.8913
No log 5.7895 330 0.7975 0.1549 0.7975 0.8931
No log 5.8246 332 0.9074 0.1034 0.9074 0.9526
No log 5.8596 334 0.9986 0.1148 0.9986 0.9993
No log 5.8947 336 1.1003 0.0909 1.1003 1.0489
No log 5.9298 338 1.2011 0.0365 1.2011 1.0959
No log 5.9649 340 1.2082 0.0394 1.2082 1.0992
No log 6.0 342 1.0910 0.1440 1.0910 1.0445
No log 6.0351 344 0.9555 0.1525 0.9555 0.9775
No log 6.0702 346 0.8973 0.1131 0.8973 0.9473
No log 6.1053 348 0.9365 0.0847 0.9365 0.9678
No log 6.1404 350 1.0078 0.1148 1.0078 1.0039
No log 6.1754 352 1.0758 0.0312 1.0758 1.0372
No log 6.2105 354 1.0003 0.0795 1.0003 1.0002
No log 6.2456 356 0.9585 0.1071 0.9585 0.9790
No log 6.2807 358 0.9466 0.0685 0.9466 0.9729
No log 6.3158 360 0.9683 0.0357 0.9683 0.9840
No log 6.3509 362 0.9978 0.0044 0.9978 0.9989
No log 6.3860 364 0.9290 0.0631 0.9290 0.9638
No log 6.4211 366 0.8757 0.1154 0.8757 0.9358
No log 6.4561 368 0.8533 0.1765 0.8533 0.9237
No log 6.4912 370 0.9021 0.1504 0.9021 0.9498
No log 6.5263 372 1.0403 0.0690 1.0403 1.0199
No log 6.5614 374 1.1452 0.1227 1.1452 1.0702
No log 6.5965 376 1.1870 0.0986 1.1870 1.0895
No log 6.6316 378 1.0818 0.1165 1.0818 1.0401
No log 6.6667 380 0.9738 0.0769 0.9738 0.9868
No log 6.7018 382 0.9556 0.0638 0.9556 0.9775
No log 6.7368 384 0.9423 0.1261 0.9423 0.9707
No log 6.7719 386 0.9093 0.1730 0.9093 0.9536
No log 6.8070 388 0.9409 0.1464 0.9409 0.9700
No log 6.8421 390 1.0458 0.1385 1.0458 1.0227
No log 6.8772 392 1.0860 0.0840 1.0860 1.0421
No log 6.9123 394 1.0408 0.1077 1.0408 1.0202
No log 6.9474 396 0.9668 0.1867 0.9668 0.9832
No log 6.9825 398 0.9225 0.1937 0.9225 0.9605
No log 7.0175 400 0.9112 0.1937 0.9112 0.9546
No log 7.0526 402 0.9545 0.2203 0.9545 0.9770
No log 7.0877 404 1.0436 0.0445 1.0436 1.0216
No log 7.1228 406 1.0783 0.0445 1.0783 1.0384
No log 7.1579 408 1.0148 0.1020 1.0148 1.0074
No log 7.1930 410 0.9107 0.2140 0.9107 0.9543
No log 7.2281 412 0.8865 0.1570 0.8865 0.9415
No log 7.2632 414 0.9109 0.1781 0.9109 0.9544
No log 7.2982 416 1.0093 0.1203 1.0093 1.0047
No log 7.3333 418 1.1546 0.1227 1.1546 1.0745
No log 7.3684 420 1.2745 0.0811 1.2745 1.1289
No log 7.4035 422 1.2284 0.1065 1.2284 1.1083
No log 7.4386 424 1.0921 0.0551 1.0921 1.0450
No log 7.4737 426 0.9407 0.1864 0.9407 0.9699
No log 7.5088 428 0.8405 0.1238 0.8405 0.9168
No log 7.5439 430 0.8155 0.1090 0.8155 0.9030
No log 7.5789 432 0.8091 0.1304 0.8091 0.8995
No log 7.6140 434 0.8460 0.1416 0.8460 0.9198
No log 7.6491 436 0.9299 0.1092 0.9299 0.9643
No log 7.6842 438 0.9525 0.1092 0.9525 0.9760
No log 7.7193 440 0.9353 0.1092 0.9353 0.9671
No log 7.7544 442 0.8984 0.2618 0.8984 0.9478
No log 7.7895 444 0.8879 0.2618 0.8879 0.9423
No log 7.8246 446 0.9464 0.1730 0.9464 0.9728
No log 7.8596 448 1.0122 0.1475 1.0122 1.0061
No log 7.8947 450 1.0677 0.1008 1.0677 1.0333
No log 7.9298 452 1.0377 0.1811 1.0377 1.0187
No log 7.9649 454 0.9789 0.1741 0.9789 0.9894
No log 8.0 456 0.9361 0.1736 0.9361 0.9675
No log 8.0351 458 0.9573 0.1392 0.9573 0.9784
No log 8.0702 460 0.9970 0.1220 0.9970 0.9985
No log 8.1053 462 0.9639 0.1220 0.9639 0.9818
No log 8.1404 464 0.9470 0.1186 0.9470 0.9731
No log 8.1754 466 0.8983 0.1034 0.8983 0.9478
No log 8.2105 468 0.8810 0.1034 0.8810 0.9386
No log 8.2456 470 0.8394 0.1698 0.8394 0.9162
No log 8.2807 472 0.8233 0.2709 0.8233 0.9074
No log 8.3158 474 0.8297 0.2676 0.8297 0.9109
No log 8.3509 476 0.8628 0.1712 0.8628 0.9289
No log 8.3860 478 0.8775 0.1712 0.8775 0.9367
No log 8.4211 480 0.8936 0.1034 0.8936 0.9453
No log 8.4561 482 0.8583 0.1304 0.8583 0.9265
No log 8.4912 484 0.8059 0.1698 0.8059 0.8977
No log 8.5263 486 0.7932 0.1698 0.7932 0.8906
No log 8.5614 488 0.8049 0.1698 0.8049 0.8972
No log 8.5965 490 0.8115 0.1705 0.8115 0.9008
No log 8.6316 492 0.8283 0.1781 0.8283 0.9101
No log 8.6667 494 0.8726 0.2000 0.8726 0.9341
No log 8.7018 496 0.8955 0.1321 0.8955 0.9463
No log 8.7368 498 0.8998 0.1336 0.8998 0.9486
0.4443 8.7719 500 0.8758 0.2000 0.8758 0.9358
0.4443 8.8070 502 0.8566 0.1781 0.8566 0.9255
0.4443 8.8421 504 0.8446 0.1781 0.8446 0.9190
0.4443 8.8772 506 0.8518 0.1781 0.8518 0.9229
0.4443 8.9123 508 0.8644 0.1781 0.8644 0.9297
0.4443 8.9474 510 0.8768 0.1852 0.8768 0.9364
0.4443 8.9825 512 0.8826 0.2661 0.8826 0.9395
0.4443 9.0175 514 0.8722 0.2661 0.8722 0.9339
0.4443 9.0526 516 0.8761 0.2000 0.8761 0.9360
0.4443 9.0877 518 0.8905 0.2000 0.8905 0.9437
0.4443 9.1228 520 0.9265 0.1730 0.9265 0.9625
0.4443 9.1579 522 0.9365 0.1453 0.9365 0.9677
0.4443 9.1930 524 0.9189 0.1453 0.9189 0.9586
0.4443 9.2281 526 0.8904 0.1730 0.8904 0.9436
0.4443 9.2632 528 0.8710 0.2340 0.8710 0.9333
0.4443 9.2982 530 0.8488 0.2000 0.8488 0.9213
0.4443 9.3333 532 0.8233 0.2661 0.8233 0.9073
0.4443 9.3684 534 0.8098 0.2661 0.8098 0.8999
0.4443 9.4035 536 0.7984 0.2227 0.7984 0.8935
0.4443 9.4386 538 0.7976 0.1845 0.7976 0.8931
0.4443 9.4737 540 0.8100 0.2661 0.8100 0.9000
0.4443 9.5088 542 0.8303 0.2661 0.8303 0.9112
0.4443 9.5439 544 0.8475 0.2000 0.8475 0.9206
0.4443 9.5789 546 0.8490 0.2000 0.8490 0.9214
0.4443 9.6140 548 0.8420 0.2372 0.8420 0.9176
0.4443 9.6491 550 0.8318 0.2661 0.8318 0.9121
0.4443 9.6842 552 0.8192 0.2300 0.8192 0.9051
0.4443 9.7193 554 0.8173 0.2300 0.8173 0.9040
0.4443 9.7544 556 0.8195 0.2300 0.8195 0.9052
0.4443 9.7895 558 0.8229 0.2300 0.8229 0.9071
0.4443 9.8246 560 0.8238 0.2300 0.8238 0.9076
0.4443 9.8596 562 0.8275 0.2300 0.8275 0.9097
0.4443 9.8947 564 0.8315 0.2300 0.8315 0.9119
0.4443 9.9298 566 0.8352 0.2300 0.8352 0.9139
0.4443 9.9649 568 0.8373 0.2661 0.8373 0.9151
0.4443 10.0 570 0.8385 0.2661 0.8385 0.9157

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k12_task3_organization

Finetuned
(4023)
this model