ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k13_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8535
  • Qwk: 0.7123
  • Mse: 0.8535
  • Rmse: 0.9238

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0333 2 6.9725 0.0176 6.9725 2.6405
No log 0.0667 4 4.9299 0.0821 4.9299 2.2203
No log 0.1 6 3.8401 -0.0398 3.8401 1.9596
No log 0.1333 8 2.4413 0.0784 2.4413 1.5625
No log 0.1667 10 1.8946 0.2791 1.8946 1.3764
No log 0.2 12 2.5538 0.1420 2.5538 1.5981
No log 0.2333 14 2.6523 0.1170 2.6523 1.6286
No log 0.2667 16 1.9520 0.2742 1.9520 1.3971
No log 0.3 18 1.4088 0.2931 1.4088 1.1869
No log 0.3333 20 1.3845 0.4167 1.3845 1.1766
No log 0.3667 22 1.3129 0.4167 1.3129 1.1458
No log 0.4 24 1.2568 0.3967 1.2568 1.1211
No log 0.4333 26 1.4170 0.4496 1.4170 1.1904
No log 0.4667 28 1.2508 0.5038 1.2508 1.1184
No log 0.5 30 1.2263 0.4463 1.2263 1.1074
No log 0.5333 32 1.3497 0.3590 1.3497 1.1617
No log 0.5667 34 1.3873 0.3932 1.3873 1.1778
No log 0.6 36 1.1210 0.5397 1.1210 1.0588
No log 0.6333 38 1.2063 0.5312 1.2063 1.0983
No log 0.6667 40 1.5092 0.375 1.5092 1.2285
No log 0.7 42 1.3494 0.4567 1.3494 1.1616
No log 0.7333 44 1.2413 0.4793 1.2413 1.1141
No log 0.7667 46 1.2353 0.5 1.2353 1.1114
No log 0.8 48 1.1991 0.4833 1.1991 1.0950
No log 0.8333 50 1.1030 0.5625 1.1030 1.0502
No log 0.8667 52 1.0287 0.6047 1.0287 1.0143
No log 0.9 54 0.9299 0.6094 0.9299 0.9643
No log 0.9333 56 0.8683 0.6716 0.8683 0.9318
No log 0.9667 58 0.8775 0.6667 0.8775 0.9367
No log 1.0 60 0.9222 0.6316 0.9222 0.9603
No log 1.0333 62 0.7718 0.6569 0.7718 0.8785
No log 1.0667 64 0.7272 0.7286 0.7272 0.8528
No log 1.1 66 0.7260 0.6912 0.7260 0.8520
No log 1.1333 68 0.8346 0.6812 0.8346 0.9136
No log 1.1667 70 0.8715 0.6471 0.8715 0.9336
No log 1.2 72 0.7459 0.7143 0.7459 0.8637
No log 1.2333 74 0.7454 0.7194 0.7454 0.8634
No log 1.2667 76 0.6757 0.7153 0.6757 0.8220
No log 1.3 78 0.7330 0.7153 0.7330 0.8561
No log 1.3333 80 0.8185 0.6308 0.8185 0.9047
No log 1.3667 82 0.8225 0.6466 0.8225 0.9069
No log 1.4 84 0.8004 0.7059 0.8004 0.8947
No log 1.4333 86 0.7587 0.7299 0.7587 0.8710
No log 1.4667 88 0.6486 0.7534 0.6486 0.8053
No log 1.5 90 0.7290 0.7561 0.7290 0.8538
No log 1.5333 92 0.6506 0.7662 0.6506 0.8066
No log 1.5667 94 0.5678 0.7619 0.5678 0.7535
No log 1.6 96 0.5724 0.7534 0.5724 0.7566
No log 1.6333 98 0.6281 0.7183 0.6281 0.7925
No log 1.6667 100 0.5870 0.7483 0.5870 0.7661
No log 1.7 102 0.5959 0.7808 0.5959 0.7719
No log 1.7333 104 0.7175 0.7632 0.7175 0.8471
No log 1.7667 106 0.6979 0.7778 0.6979 0.8354
No log 1.8 108 0.7539 0.6667 0.7539 0.8683
No log 1.8333 110 0.8332 0.6522 0.8332 0.9128
No log 1.8667 112 0.6811 0.7172 0.6811 0.8253
No log 1.9 114 0.5743 0.8221 0.5743 0.7578
No log 1.9333 116 0.6619 0.7746 0.6619 0.8135
No log 1.9667 118 0.5439 0.8118 0.5439 0.7375
No log 2.0 120 0.5654 0.7692 0.5654 0.7519
No log 2.0333 122 0.6786 0.7226 0.6786 0.8238
No log 2.0667 124 0.5686 0.7821 0.5686 0.7541
No log 2.1 126 0.6649 0.7771 0.6649 0.8154
No log 2.1333 128 0.9542 0.6988 0.9542 0.9768
No log 2.1667 130 0.9612 0.7135 0.9612 0.9804
No log 2.2 132 0.6712 0.8075 0.6712 0.8193
No log 2.2333 134 0.7185 0.75 0.7185 0.8476
No log 2.2667 136 0.9173 0.6197 0.9173 0.9577
No log 2.3 138 0.8689 0.6533 0.8689 0.9322
No log 2.3333 140 0.6753 0.7613 0.6753 0.8217
No log 2.3667 142 0.7419 0.7931 0.7419 0.8614
No log 2.4 144 0.7439 0.8 0.7439 0.8625
No log 2.4333 146 0.6282 0.8095 0.6282 0.7926
No log 2.4667 148 0.6244 0.7925 0.6244 0.7902
No log 2.5 150 0.7283 0.7261 0.7283 0.8534
No log 2.5333 152 0.8001 0.6839 0.8000 0.8945
No log 2.5667 154 0.7291 0.7632 0.7291 0.8539
No log 2.6 156 0.6409 0.76 0.6409 0.8006
No log 2.6333 158 0.6461 0.7733 0.6461 0.8038
No log 2.6667 160 0.6796 0.7742 0.6796 0.8244
No log 2.7 162 0.7245 0.72 0.7245 0.8512
No log 2.7333 164 0.7729 0.7034 0.7729 0.8792
No log 2.7667 166 0.7484 0.7260 0.7484 0.8651
No log 2.8 168 0.7327 0.7651 0.7327 0.8560
No log 2.8333 170 0.7385 0.7417 0.7385 0.8593
No log 2.8667 172 0.6311 0.7771 0.6311 0.7944
No log 2.9 174 0.6039 0.7799 0.6039 0.7771
No log 2.9333 176 0.6212 0.7799 0.6212 0.7882
No log 2.9667 178 0.6677 0.7898 0.6677 0.8172
No log 3.0 180 0.7620 0.7582 0.7620 0.8729
No log 3.0333 182 0.7678 0.7248 0.7678 0.8762
No log 3.0667 184 0.7418 0.7248 0.7418 0.8613
No log 3.1 186 0.6600 0.7763 0.6600 0.8124
No log 3.1333 188 0.6405 0.7975 0.6405 0.8003
No log 3.1667 190 0.6319 0.7898 0.6319 0.7949
No log 3.2 192 0.5872 0.7838 0.5872 0.7663
No log 3.2333 194 0.5697 0.7838 0.5697 0.7548
No log 3.2667 196 0.5279 0.8228 0.5279 0.7266
No log 3.3 198 0.5418 0.8280 0.5418 0.7361
No log 3.3333 200 0.5404 0.8205 0.5404 0.7351
No log 3.3667 202 0.5542 0.7895 0.5542 0.7444
No log 3.4 204 0.6000 0.7662 0.6000 0.7746
No log 3.4333 206 0.8875 0.7586 0.8875 0.9421
No log 3.4667 208 1.0106 0.6705 1.0106 1.0053
No log 3.5 210 0.7216 0.7738 0.7216 0.8495
No log 3.5333 212 0.5044 0.7975 0.5044 0.7102
No log 3.5667 214 0.6851 0.72 0.6851 0.8277
No log 3.6 216 0.8553 0.6883 0.8553 0.9248
No log 3.6333 218 0.8276 0.7059 0.8276 0.9097
No log 3.6667 220 0.8184 0.7020 0.8184 0.9046
No log 3.7 222 0.6765 0.7483 0.6765 0.8225
No log 3.7333 224 0.6934 0.7448 0.6934 0.8327
No log 3.7667 226 0.7736 0.7448 0.7736 0.8795
No log 3.8 228 0.9303 0.6667 0.9303 0.9645
No log 3.8333 230 0.9525 0.6667 0.9525 0.9760
No log 3.8667 232 0.8335 0.6944 0.8335 0.9129
No log 3.9 234 0.7401 0.7586 0.7401 0.8603
No log 3.9333 236 0.6946 0.7792 0.6946 0.8335
No log 3.9667 238 0.6603 0.7742 0.6603 0.8126
No log 4.0 240 0.7353 0.7333 0.7353 0.8575
No log 4.0333 242 0.7259 0.7211 0.7259 0.8520
No log 4.0667 244 0.6729 0.7413 0.6729 0.8203
No log 4.1 246 0.8099 0.6711 0.8099 0.8999
No log 4.1333 248 1.0260 0.5833 1.0260 1.0129
No log 4.1667 250 1.0135 0.6056 1.0135 1.0067
No log 4.2 252 0.8182 0.6619 0.8182 0.9046
No log 4.2333 254 0.7073 0.7273 0.7073 0.8410
No log 4.2667 256 0.8289 0.7042 0.8289 0.9105
No log 4.3 258 0.8900 0.7042 0.8900 0.9434
No log 4.3333 260 0.7668 0.6950 0.7668 0.8757
No log 4.3667 262 0.6622 0.7123 0.6622 0.8138
No log 4.4 264 0.6597 0.7260 0.6597 0.8122
No log 4.4333 266 0.6576 0.7211 0.6576 0.8109
No log 4.4667 268 0.6929 0.7172 0.6929 0.8324
No log 4.5 270 0.7779 0.7083 0.7779 0.8820
No log 4.5333 272 0.9082 0.6716 0.9082 0.9530
No log 4.5667 274 0.8845 0.6861 0.8845 0.9405
No log 4.6 276 0.7233 0.7083 0.7233 0.8505
No log 4.6333 278 0.5788 0.7763 0.5788 0.7608
No log 4.6667 280 0.5583 0.7895 0.5583 0.7472
No log 4.7 282 0.5555 0.7632 0.5555 0.7453
No log 4.7333 284 0.7397 0.6986 0.7397 0.8600
No log 4.7667 286 0.8134 0.6853 0.8134 0.9019
No log 4.8 288 0.6824 0.7376 0.6824 0.8261
No log 4.8333 290 0.6293 0.7808 0.6293 0.7933
No log 4.8667 292 0.7605 0.7114 0.7605 0.8721
No log 4.9 294 0.7465 0.7550 0.7465 0.8640
No log 4.9333 296 0.6361 0.8 0.6361 0.7975
No log 4.9667 298 0.6704 0.7483 0.6704 0.8188
No log 5.0 300 0.7426 0.7432 0.7426 0.8617
No log 5.0333 302 0.7591 0.7347 0.7591 0.8713
No log 5.0667 304 0.7334 0.7397 0.7334 0.8564
No log 5.1 306 0.6988 0.7785 0.6988 0.8359
No log 5.1333 308 0.7118 0.7483 0.7118 0.8437
No log 5.1667 310 0.7093 0.7483 0.7093 0.8422
No log 5.2 312 0.7518 0.7123 0.7518 0.8671
No log 5.2333 314 0.8443 0.6713 0.8443 0.9188
No log 5.2667 316 0.8661 0.6620 0.8661 0.9306
No log 5.3 318 0.8541 0.6620 0.8541 0.9242
No log 5.3333 320 0.8203 0.6620 0.8203 0.9057
No log 5.3667 322 0.8018 0.6713 0.8018 0.8954
No log 5.4 324 0.7746 0.6806 0.7746 0.8801
No log 5.4333 326 0.7622 0.6986 0.7622 0.8731
No log 5.4667 328 0.7961 0.6944 0.7961 0.8922
No log 5.5 330 0.7814 0.6897 0.7814 0.8840
No log 5.5333 332 0.7937 0.6763 0.7937 0.8909
No log 5.5667 334 0.7793 0.6857 0.7793 0.8828
No log 5.6 336 0.7656 0.7 0.7656 0.8750
No log 5.6333 338 0.7199 0.7448 0.7199 0.8485
No log 5.6667 340 0.7256 0.7211 0.7256 0.8518
No log 5.7 342 0.6714 0.7211 0.6714 0.8194
No log 5.7333 344 0.6214 0.8212 0.6214 0.7883
No log 5.7667 346 0.5982 0.8158 0.5982 0.7734
No log 5.8 348 0.6135 0.7867 0.6135 0.7833
No log 5.8333 350 0.6274 0.7867 0.6274 0.7921
No log 5.8667 352 0.6373 0.7867 0.6373 0.7983
No log 5.9 354 0.6366 0.8212 0.6366 0.7979
No log 5.9333 356 0.6237 0.7973 0.6237 0.7897
No log 5.9667 358 0.5895 0.8258 0.5895 0.7678
No log 6.0 360 0.5860 0.8158 0.5860 0.7655
No log 6.0333 362 0.6278 0.8054 0.6278 0.7923
No log 6.0667 364 0.6848 0.7534 0.6848 0.8275
No log 6.1 366 0.6700 0.7586 0.6700 0.8186
No log 6.1333 368 0.6074 0.8133 0.6074 0.7794
No log 6.1667 370 0.5663 0.8205 0.5663 0.7526
No log 6.2 372 0.5592 0.8447 0.5592 0.7478
No log 6.2333 374 0.5356 0.8447 0.5356 0.7318
No log 6.2667 376 0.5371 0.8447 0.5371 0.7329
No log 6.3 378 0.5459 0.8447 0.5459 0.7389
No log 6.3333 380 0.5507 0.8153 0.5507 0.7421
No log 6.3667 382 0.5394 0.8182 0.5394 0.7344
No log 6.4 384 0.5375 0.8079 0.5375 0.7332
No log 6.4333 386 0.5327 0.8312 0.5327 0.7299
No log 6.4667 388 0.5664 0.8205 0.5664 0.7526
No log 6.5 390 0.5547 0.8235 0.5547 0.7448
No log 6.5333 392 0.5643 0.7808 0.5643 0.7512
No log 6.5667 394 0.5824 0.7586 0.5824 0.7632
No log 6.6 396 0.5433 0.7973 0.5433 0.7371
No log 6.6333 398 0.5226 0.8205 0.5226 0.7229
No log 6.6667 400 0.5627 0.7976 0.5627 0.7501
No log 6.7 402 0.6083 0.7929 0.6083 0.7800
No log 6.7333 404 0.5697 0.8302 0.5697 0.7548
No log 6.7667 406 0.6006 0.7919 0.6006 0.7750
No log 6.8 408 0.6542 0.7671 0.6542 0.8088
No log 6.8333 410 0.6883 0.7808 0.6883 0.8296
No log 6.8667 412 0.6738 0.7671 0.6738 0.8208
No log 6.9 414 0.6428 0.7568 0.6428 0.8017
No log 6.9333 416 0.6628 0.7333 0.6628 0.8141
No log 6.9667 418 0.6660 0.7333 0.6660 0.8161
No log 7.0 420 0.6310 0.7619 0.6310 0.7944
No log 7.0333 422 0.6373 0.8052 0.6373 0.7983
No log 7.0667 424 0.6199 0.8101 0.6199 0.7873
No log 7.1 426 0.6172 0.7703 0.6172 0.7856
No log 7.1333 428 0.6650 0.7383 0.6650 0.8155
No log 7.1667 430 0.7464 0.7105 0.7464 0.8640
No log 7.2 432 0.7089 0.72 0.7089 0.8419
No log 7.2333 434 0.6115 0.7838 0.6115 0.7820
No log 7.2667 436 0.5693 0.8205 0.5693 0.7545
No log 7.3 438 0.5541 0.8176 0.5541 0.7444
No log 7.3333 440 0.5830 0.8025 0.5830 0.7636
No log 7.3667 442 0.6622 0.7134 0.6622 0.8138
No log 7.4 444 0.7236 0.7013 0.7236 0.8507
No log 7.4333 446 0.6641 0.7297 0.6641 0.8149
No log 7.4667 448 0.6503 0.7973 0.6503 0.8064
No log 7.5 450 0.6903 0.7785 0.6903 0.8308
No log 7.5333 452 0.6714 0.7785 0.6714 0.8194
No log 7.5667 454 0.6563 0.7891 0.6563 0.8101
No log 7.6 456 0.7153 0.7483 0.7153 0.8458
No log 7.6333 458 0.7896 0.6980 0.7896 0.8886
No log 7.6667 460 0.7732 0.7397 0.7732 0.8793
No log 7.7 462 0.6980 0.7413 0.6980 0.8354
No log 7.7333 464 0.6337 0.7724 0.6337 0.7960
No log 7.7667 466 0.6016 0.7922 0.6016 0.7757
No log 7.8 468 0.5918 0.7843 0.5918 0.7693
No log 7.8333 470 0.6011 0.7692 0.6011 0.7753
No log 7.8667 472 0.6081 0.7799 0.6081 0.7798
No log 7.9 474 0.6426 0.7550 0.6426 0.8016
No log 7.9333 476 0.6993 0.7383 0.6993 0.8362
No log 7.9667 478 0.7650 0.7297 0.7650 0.8746
No log 8.0 480 0.7514 0.7413 0.7514 0.8668
No log 8.0333 482 0.7348 0.7413 0.7348 0.8572
No log 8.0667 484 0.7065 0.7413 0.7065 0.8405
No log 8.1 486 0.6443 0.7413 0.6443 0.8027
No log 8.1333 488 0.6299 0.7451 0.6299 0.7937
No log 8.1667 490 0.6077 0.7361 0.6077 0.7795
No log 8.2 492 0.6060 0.7448 0.6060 0.7785
No log 8.2333 494 0.6295 0.7448 0.6295 0.7934
No log 8.2667 496 0.6486 0.7361 0.6486 0.8054
No log 8.3 498 0.6494 0.7310 0.6494 0.8058
0.3723 8.3333 500 0.6287 0.7310 0.6287 0.7929
0.3723 8.3667 502 0.5700 0.7671 0.5700 0.7550
0.3723 8.4 504 0.5627 0.8108 0.5627 0.7501
0.3723 8.4333 506 0.5892 0.8054 0.5892 0.7676
0.3723 8.4667 508 0.5841 0.8027 0.5841 0.7643
0.3723 8.5 510 0.6411 0.7324 0.6411 0.8007
0.3723 8.5333 512 0.6912 0.7324 0.6912 0.8314
0.3723 8.5667 514 0.7064 0.7413 0.7064 0.8405
0.3723 8.6 516 0.6974 0.7724 0.6974 0.8351
0.3723 8.6333 518 0.7175 0.7586 0.7175 0.8471
0.3723 8.6667 520 0.7566 0.6993 0.7566 0.8698
0.3723 8.7 522 0.7336 0.7448 0.7336 0.8565
0.3723 8.7333 524 0.6805 0.7755 0.6805 0.8249
0.3723 8.7667 526 0.6689 0.7703 0.6689 0.8178
0.3723 8.8 528 0.6477 0.7703 0.6477 0.8048
0.3723 8.8333 530 0.6195 0.7867 0.6195 0.7871
0.3723 8.8667 532 0.6406 0.7785 0.6406 0.8004
0.3723 8.9 534 0.6495 0.7838 0.6495 0.8059
0.3723 8.9333 536 0.6440 0.7815 0.6440 0.8025
0.3723 8.9667 538 0.6459 0.7867 0.6459 0.8037
0.3723 9.0 540 0.6780 0.7867 0.6780 0.8234
0.3723 9.0333 542 0.7184 0.7347 0.7184 0.8476
0.3723 9.0667 544 0.7841 0.7042 0.7841 0.8855
0.3723 9.1 546 0.8851 0.6486 0.8851 0.9408
0.3723 9.1333 548 0.9709 0.6345 0.9709 0.9853
0.3723 9.1667 550 0.9635 0.6345 0.9635 0.9816
0.3723 9.2 552 0.8535 0.7123 0.8535 0.9238

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k13_task1_organization

Finetuned
(4023)
this model