ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k20_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7909
  • Qwk: 0.6944
  • Mse: 0.7909
  • Rmse: 0.8893

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0222 2 6.6255 0.0188 6.6255 2.5740
No log 0.0444 4 4.0678 0.0702 4.0678 2.0169
No log 0.0667 6 3.5641 -0.0410 3.5641 1.8879
No log 0.0889 8 3.0662 -0.0125 3.0662 1.7511
No log 0.1111 10 2.6223 0.0857 2.6223 1.6194
No log 0.1333 12 2.1814 0.1575 2.1814 1.4770
No log 0.1556 14 2.0039 0.0840 2.0039 1.4156
No log 0.1778 16 2.0383 0.1416 2.0383 1.4277
No log 0.2 18 2.2822 -0.0645 2.2822 1.5107
No log 0.2222 20 2.2410 -0.0163 2.2410 1.4970
No log 0.2444 22 2.1194 0.0855 2.1194 1.4558
No log 0.2667 24 2.1727 0.0333 2.1727 1.4740
No log 0.2889 26 2.0742 0.1626 2.0742 1.4402
No log 0.3111 28 2.0329 0.2047 2.0329 1.4258
No log 0.3333 30 1.9789 0.2154 1.9789 1.4067
No log 0.3556 32 2.0677 0.1654 2.0677 1.4380
No log 0.3778 34 2.0904 0.1159 2.0904 1.4458
No log 0.4 36 1.9902 0.2105 1.9902 1.4107
No log 0.4222 38 1.6787 0.3390 1.6787 1.2956
No log 0.4444 40 1.4915 0.2056 1.4915 1.2213
No log 0.4667 42 1.4805 0.3393 1.4805 1.2167
No log 0.4889 44 1.6735 0.2586 1.6735 1.2936
No log 0.5111 46 2.0698 0.1077 2.0698 1.4387
No log 0.5333 48 1.9046 0.2063 1.9046 1.3801
No log 0.5556 50 1.5135 0.2909 1.5135 1.2302
No log 0.5778 52 1.4884 0.2883 1.4884 1.2200
No log 0.6 54 1.4064 0.2545 1.4064 1.1859
No log 0.6222 56 1.3929 0.2523 1.3929 1.1802
No log 0.6444 58 1.4302 0.2857 1.4302 1.1959
No log 0.6667 60 1.4693 0.3333 1.4693 1.2122
No log 0.6889 62 1.4019 0.3186 1.4019 1.1840
No log 0.7111 64 1.4333 0.4538 1.4333 1.1972
No log 0.7333 66 1.4674 0.3906 1.4674 1.2114
No log 0.7556 68 1.4228 0.4480 1.4228 1.1928
No log 0.7778 70 1.3082 0.4103 1.3082 1.1437
No log 0.8 72 1.3254 0.3793 1.3254 1.1513
No log 0.8222 74 1.3720 0.4516 1.3720 1.1713
No log 0.8444 76 1.4795 0.3876 1.4795 1.2163
No log 0.8667 78 1.3807 0.4320 1.3807 1.1750
No log 0.8889 80 1.2254 0.4921 1.2254 1.1070
No log 0.9111 82 1.3265 0.4762 1.3265 1.1517
No log 0.9333 84 1.7365 0.2222 1.7365 1.3178
No log 0.9556 86 2.1084 0.0465 2.1084 1.4520
No log 0.9778 88 1.9472 0.1395 1.9472 1.3954
No log 1.0 90 1.8154 0.2090 1.8154 1.3474
No log 1.0222 92 1.8056 0.2090 1.8056 1.3437
No log 1.0444 94 1.4864 0.3212 1.4864 1.2192
No log 1.0667 96 1.3500 0.4462 1.3500 1.1619
No log 1.0889 98 1.5844 0.2941 1.5844 1.2587
No log 1.1111 100 1.5787 0.3259 1.5787 1.2564
No log 1.1333 102 1.3178 0.5197 1.3178 1.1480
No log 1.1556 104 1.2853 0.512 1.2853 1.1337
No log 1.1778 106 1.3672 0.4179 1.3672 1.1693
No log 1.2 108 1.3182 0.4179 1.3182 1.1481
No log 1.2222 110 1.3714 0.4265 1.3714 1.1711
No log 1.2444 112 1.2287 0.4748 1.2287 1.1085
No log 1.2667 114 1.2513 0.4965 1.2513 1.1186
No log 1.2889 116 1.3187 0.4748 1.3187 1.1483
No log 1.3111 118 1.5470 0.4234 1.5470 1.2438
No log 1.3333 120 1.4374 0.4265 1.4374 1.1989
No log 1.3556 122 1.2454 0.5156 1.2454 1.1160
No log 1.3778 124 1.2449 0.5385 1.2449 1.1158
No log 1.4 126 1.1296 0.5323 1.1296 1.0628
No log 1.4222 128 1.0778 0.6032 1.0778 1.0382
No log 1.4444 130 1.1421 0.5312 1.1421 1.0687
No log 1.4667 132 1.5262 0.4143 1.5262 1.2354
No log 1.4889 134 1.6230 0.4113 1.6230 1.2740
No log 1.5111 136 1.5758 0.3913 1.5758 1.2553
No log 1.5333 138 1.2836 0.5075 1.2836 1.1330
No log 1.5556 140 1.0182 0.576 1.0182 1.0091
No log 1.5778 142 0.9548 0.5738 0.9548 0.9771
No log 1.6 144 0.9503 0.5410 0.9503 0.9748
No log 1.6222 146 0.9929 0.6299 0.9929 0.9964
No log 1.6444 148 0.8894 0.6066 0.8894 0.9431
No log 1.6667 150 0.8147 0.6418 0.8147 0.9026
No log 1.6889 152 0.9444 0.5957 0.9444 0.9718
No log 1.7111 154 0.8337 0.6338 0.8337 0.9130
No log 1.7333 156 0.6600 0.7376 0.6600 0.8124
No log 1.7556 158 0.8006 0.6809 0.8006 0.8948
No log 1.7778 160 1.0160 0.6099 1.0160 1.0080
No log 1.8 162 0.8951 0.6294 0.8951 0.9461
No log 1.8222 164 0.8329 0.6714 0.8329 0.9126
No log 1.8444 166 0.8544 0.6412 0.8544 0.9243
No log 1.8667 168 0.8431 0.6870 0.8431 0.9182
No log 1.8889 170 0.8928 0.6094 0.8928 0.9449
No log 1.9111 172 0.9041 0.6349 0.9041 0.9509
No log 1.9333 174 0.8602 0.6667 0.8602 0.9274
No log 1.9556 176 0.8393 0.5385 0.8393 0.9161
No log 1.9778 178 0.8158 0.6970 0.8158 0.9032
No log 2.0 180 0.9188 0.6512 0.9188 0.9585
No log 2.0222 182 0.9124 0.6614 0.9124 0.9552
No log 2.0444 184 0.8751 0.6299 0.8751 0.9355
No log 2.0667 186 0.8979 0.5354 0.8979 0.9476
No log 2.0889 188 0.9059 0.5528 0.9059 0.9518
No log 2.1111 190 1.0256 0.6614 1.0256 1.0127
No log 2.1333 192 1.0584 0.6202 1.0584 1.0288
No log 2.1556 194 0.9830 0.6508 0.9830 0.9915
No log 2.1778 196 0.9652 0.5 0.9652 0.9825
No log 2.2 198 0.9613 0.4407 0.9613 0.9805
No log 2.2222 200 0.9022 0.4959 0.9022 0.9498
No log 2.2444 202 0.8263 0.6015 0.8263 0.9090
No log 2.2667 204 0.8090 0.6131 0.8090 0.8994
No log 2.2889 206 0.8130 0.5865 0.8130 0.9017
No log 2.3111 208 0.8693 0.5865 0.8693 0.9324
No log 2.3333 210 0.9652 0.5970 0.9652 0.9825
No log 2.3556 212 0.9363 0.5606 0.9363 0.9676
No log 2.3778 214 0.8678 0.624 0.8678 0.9316
No log 2.4 216 0.8909 0.6406 0.8909 0.9439
No log 2.4222 218 0.8461 0.6349 0.8461 0.9198
No log 2.4444 220 0.8336 0.6061 0.8336 0.9130
No log 2.4667 222 0.8249 0.6471 0.8249 0.9083
No log 2.4889 224 0.7796 0.6619 0.7796 0.8829
No log 2.5111 226 0.7630 0.7015 0.7630 0.8735
No log 2.5333 228 0.8961 0.5954 0.8961 0.9466
No log 2.5556 230 0.9907 0.5606 0.9907 0.9953
No log 2.5778 232 0.9063 0.6119 0.9063 0.9520
No log 2.6 234 0.8926 0.6667 0.8926 0.9448
No log 2.6222 236 0.8659 0.6667 0.8659 0.9305
No log 2.6444 238 0.8801 0.6769 0.8801 0.9381
No log 2.6667 240 0.8517 0.7164 0.8517 0.9229
No log 2.6889 242 0.8679 0.6870 0.8679 0.9316
No log 2.7111 244 0.9648 0.6418 0.9648 0.9822
No log 2.7333 246 1.0215 0.5846 1.0215 1.0107
No log 2.7556 248 0.9187 0.6522 0.9187 0.9585
No log 2.7778 250 0.7329 0.7286 0.7329 0.8561
No log 2.8 252 0.6943 0.7391 0.6943 0.8333
No log 2.8222 254 0.7360 0.7482 0.7360 0.8579
No log 2.8444 256 0.8057 0.6906 0.8057 0.8976
No log 2.8667 258 0.8122 0.6963 0.8122 0.9012
No log 2.8889 260 0.8266 0.7206 0.8266 0.9092
No log 2.9111 262 0.8669 0.6765 0.8669 0.9311
No log 2.9333 264 0.8915 0.6667 0.8915 0.9442
No log 2.9556 266 0.9734 0.6107 0.9734 0.9866
No log 2.9778 268 0.9679 0.6107 0.9679 0.9838
No log 3.0 270 0.8767 0.6212 0.8767 0.9363
No log 3.0222 272 0.8751 0.6212 0.8751 0.9355
No log 3.0444 274 0.8732 0.6212 0.8732 0.9345
No log 3.0667 276 0.8763 0.6154 0.8763 0.9361
No log 3.0889 278 0.9164 0.6400 0.9164 0.9573
No log 3.1111 280 0.9197 0.6508 0.9197 0.9590
No log 3.1333 282 0.8881 0.6094 0.8881 0.9424
No log 3.1556 284 0.8420 0.6370 0.8420 0.9176
No log 3.1778 286 0.7580 0.6667 0.7580 0.8706
No log 3.2 288 0.6794 0.7692 0.6794 0.8242
No log 3.2222 290 0.6567 0.7586 0.6567 0.8104
No log 3.2444 292 0.6355 0.7862 0.6355 0.7972
No log 3.2667 294 0.7521 0.6901 0.7521 0.8673
No log 3.2889 296 0.9588 0.6187 0.9588 0.9792
No log 3.3111 298 0.9023 0.6423 0.9023 0.9499
No log 3.3333 300 0.7527 0.7092 0.7527 0.8676
No log 3.3556 302 0.7808 0.6912 0.7808 0.8836
No log 3.3778 304 0.8854 0.5821 0.8854 0.9409
No log 3.4 306 0.8586 0.6466 0.8586 0.9266
No log 3.4222 308 0.8115 0.6866 0.8115 0.9009
No log 3.4444 310 0.8494 0.6519 0.8494 0.9216
No log 3.4667 312 0.8729 0.6522 0.8729 0.9343
No log 3.4889 314 0.8009 0.6569 0.8009 0.8949
No log 3.5111 316 0.7696 0.6569 0.7696 0.8773
No log 3.5333 318 0.7596 0.6906 0.7596 0.8715
No log 3.5556 320 0.7084 0.7050 0.7084 0.8417
No log 3.5778 322 0.6740 0.7536 0.6740 0.8210
No log 3.6 324 0.6913 0.6912 0.6913 0.8314
No log 3.6222 326 0.8059 0.6377 0.8059 0.8977
No log 3.6444 328 0.8405 0.6667 0.8405 0.9168
No log 3.6667 330 0.7267 0.6714 0.7267 0.8525
No log 3.6889 332 0.5970 0.7206 0.5970 0.7727
No log 3.7111 334 0.5912 0.7571 0.5912 0.7689
No log 3.7333 336 0.6504 0.7353 0.6504 0.8064
No log 3.7556 338 0.6589 0.7111 0.6589 0.8117
No log 3.7778 340 0.6867 0.6963 0.6867 0.8287
No log 3.8 342 0.7327 0.6715 0.7327 0.8560
No log 3.8222 344 0.7483 0.6619 0.7483 0.8651
No log 3.8444 346 0.6974 0.7042 0.6974 0.8351
No log 3.8667 348 0.6177 0.6950 0.6177 0.7859
No log 3.8889 350 0.5719 0.7746 0.5719 0.7562
No log 3.9111 352 0.6019 0.7746 0.6019 0.7758
No log 3.9333 354 0.6164 0.7429 0.6164 0.7851
No log 3.9556 356 0.6486 0.7194 0.6486 0.8054
No log 3.9778 358 0.6904 0.6765 0.6904 0.8309
No log 4.0 360 0.6979 0.7059 0.6979 0.8354
No log 4.0222 362 0.7004 0.6963 0.7004 0.8369
No log 4.0444 364 0.6827 0.7324 0.6827 0.8262
No log 4.0667 366 0.6840 0.7724 0.6840 0.8270
No log 4.0889 368 0.6491 0.7413 0.6491 0.8057
No log 4.1111 370 0.6575 0.7153 0.6575 0.8109
No log 4.1333 372 0.7195 0.6912 0.7195 0.8482
No log 4.1556 374 0.7574 0.6617 0.7574 0.8703
No log 4.1778 376 0.7922 0.7286 0.7922 0.8900
No log 4.2 378 0.7655 0.7606 0.7655 0.8749
No log 4.2222 380 0.6651 0.7194 0.6651 0.8155
No log 4.2444 382 0.6331 0.7222 0.6331 0.7957
No log 4.2667 384 0.6333 0.7222 0.6333 0.7958
No log 4.2889 386 0.6017 0.7755 0.6017 0.7757
No log 4.3111 388 0.6257 0.7862 0.6257 0.7910
No log 4.3333 390 0.6375 0.7891 0.6375 0.7984
No log 4.3556 392 0.6767 0.7361 0.6767 0.8226
No log 4.3778 394 0.7475 0.6906 0.7475 0.8646
No log 4.4 396 0.7362 0.7234 0.7362 0.8580
No log 4.4222 398 0.7084 0.7778 0.7084 0.8416
No log 4.4444 400 0.6761 0.7 0.6761 0.8222
No log 4.4667 402 0.7721 0.6957 0.7721 0.8787
No log 4.4889 404 0.8983 0.6423 0.8983 0.9478
No log 4.5111 406 0.8696 0.6571 0.8696 0.9325
No log 4.5333 408 0.7361 0.7092 0.7361 0.8579
No log 4.5556 410 0.7186 0.7234 0.7186 0.8477
No log 4.5778 412 0.7683 0.7194 0.7683 0.8765
No log 4.6 414 0.8423 0.6767 0.8423 0.9178
No log 4.6222 416 0.9103 0.6260 0.9103 0.9541
No log 4.6444 418 0.9180 0.6154 0.9180 0.9581
No log 4.6667 420 0.8930 0.6562 0.8930 0.9450
No log 4.6889 422 0.8692 0.6299 0.8692 0.9323
No log 4.7111 424 0.8541 0.6412 0.8541 0.9242
No log 4.7333 426 0.8358 0.6412 0.8358 0.9142
No log 4.7556 428 0.7928 0.6615 0.7928 0.8904
No log 4.7778 430 0.8303 0.6569 0.8303 0.9112
No log 4.8 432 0.9872 0.6883 0.9872 0.9936
No log 4.8222 434 1.0980 0.6533 1.0980 1.0479
No log 4.8444 436 0.9860 0.6338 0.9860 0.9930
No log 4.8667 438 0.7785 0.6377 0.7785 0.8823
No log 4.8889 440 0.7014 0.7194 0.7014 0.8375
No log 4.9111 442 0.7042 0.6912 0.7042 0.8391
No log 4.9333 444 0.7285 0.7007 0.7285 0.8535
No log 4.9556 446 0.7304 0.6906 0.7304 0.8546
No log 4.9778 448 0.7656 0.6714 0.7656 0.8750
No log 5.0 450 0.8182 0.6619 0.8182 0.9045
No log 5.0222 452 0.8277 0.6619 0.8277 0.9098
No log 5.0444 454 0.7557 0.6812 0.7557 0.8693
No log 5.0667 456 0.6990 0.7482 0.6990 0.8361
No log 5.0889 458 0.6999 0.7606 0.6999 0.8366
No log 5.1111 460 0.6981 0.7660 0.6981 0.8355
No log 5.1333 462 0.7398 0.7092 0.7398 0.8601
No log 5.1556 464 0.8140 0.6619 0.8140 0.9022
No log 5.1778 466 0.8220 0.6619 0.8220 0.9067
No log 5.2 468 0.7885 0.6715 0.7885 0.8880
No log 5.2222 470 0.7753 0.7007 0.7753 0.8805
No log 5.2444 472 0.7745 0.7007 0.7745 0.8801
No log 5.2667 474 0.7571 0.6963 0.7571 0.8701
No log 5.2889 476 0.7431 0.7101 0.7431 0.8620
No log 5.3111 478 0.7518 0.6812 0.7518 0.8671
No log 5.3333 480 0.8179 0.6377 0.8179 0.9044
No log 5.3556 482 0.7831 0.6857 0.7831 0.8849
No log 5.3778 484 0.7113 0.6857 0.7113 0.8434
No log 5.4 486 0.6322 0.7945 0.6322 0.7951
No log 5.4222 488 0.6245 0.8163 0.6245 0.7903
No log 5.4444 490 0.5706 0.84 0.5706 0.7554
No log 5.4667 492 0.5311 0.7808 0.5311 0.7288
No log 5.4889 494 0.5539 0.75 0.5539 0.7443
No log 5.5111 496 0.5820 0.75 0.5820 0.7629
No log 5.5333 498 0.6170 0.75 0.6170 0.7855
0.4077 5.5556 500 0.6307 0.7534 0.6307 0.7942
0.4077 5.5778 502 0.6531 0.7552 0.6531 0.8082
0.4077 5.6 504 0.6855 0.7571 0.6855 0.8280
0.4077 5.6222 506 0.6782 0.7639 0.6782 0.8235
0.4077 5.6444 508 0.6716 0.7838 0.6716 0.8195
0.4077 5.6667 510 0.6438 0.7755 0.6438 0.8024
0.4077 5.6889 512 0.6211 0.7682 0.6211 0.7881
0.4077 5.7111 514 0.6976 0.7285 0.6976 0.8352
0.4077 5.7333 516 0.7955 0.6842 0.7955 0.8919
0.4077 5.7556 518 0.8008 0.6842 0.8008 0.8948
0.4077 5.7778 520 0.7249 0.7067 0.7249 0.8514
0.4077 5.8 522 0.6356 0.8 0.6356 0.7972
0.4077 5.8222 524 0.6372 0.8235 0.6372 0.7982
0.4077 5.8444 526 0.6759 0.7733 0.6759 0.8221
0.4077 5.8667 528 0.6772 0.7891 0.6772 0.8229
0.4077 5.8889 530 0.7253 0.7448 0.7253 0.8517
0.4077 5.9111 532 0.8106 0.6713 0.8106 0.9003
0.4077 5.9333 534 0.8492 0.6713 0.8492 0.9215
0.4077 5.9556 536 0.8341 0.6713 0.8341 0.9133
0.4077 5.9778 538 0.7909 0.6944 0.7909 0.8893

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k20_task1_organization

Finetuned
(4023)
this model