ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k13_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8592
  • Qwk: 0.6269
  • Mse: 0.8592
  • Rmse: 0.9270

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0333 2 6.8595 0.0058 6.8595 2.6191
No log 0.0667 4 4.4832 0.0732 4.4832 2.1174
No log 0.1 6 3.3616 0.0904 3.3616 1.8335
No log 0.1333 8 2.5808 0.0976 2.5808 1.6065
No log 0.1667 10 2.2015 0.2817 2.2015 1.4837
No log 0.2 12 2.0604 0.1575 2.0604 1.4354
No log 0.2333 14 2.1739 0.0763 2.1739 1.4744
No log 0.2667 16 1.7907 0.1947 1.7907 1.3382
No log 0.3 18 1.7029 0.1802 1.7029 1.3049
No log 0.3333 20 1.9092 0.2034 1.9092 1.3818
No log 0.3667 22 1.9784 0.1626 1.9784 1.4066
No log 0.4 24 1.6824 0.2105 1.6824 1.2971
No log 0.4333 26 1.6235 0.2759 1.6235 1.2742
No log 0.4667 28 1.6531 0.3770 1.6531 1.2857
No log 0.5 30 1.8602 0.3175 1.8602 1.3639
No log 0.5333 32 1.8594 0.3438 1.8594 1.3636
No log 0.5667 34 1.6361 0.3710 1.6361 1.2791
No log 0.6 36 1.5251 0.3740 1.5251 1.2349
No log 0.6333 38 1.3421 0.3304 1.3421 1.1585
No log 0.6667 40 1.2746 0.3894 1.2746 1.1290
No log 0.7 42 1.2807 0.4655 1.2807 1.1317
No log 0.7333 44 1.3087 0.4386 1.3087 1.1440
No log 0.7667 46 1.3821 0.2936 1.3821 1.1756
No log 0.8 48 1.9644 0.3200 1.9644 1.4016
No log 0.8333 50 2.4192 0.1000 2.4192 1.5554
No log 0.8667 52 2.0766 0.2923 2.0766 1.4410
No log 0.9 54 1.4518 0.3051 1.4518 1.2049
No log 0.9333 56 1.3558 0.4174 1.3558 1.1644
No log 0.9667 58 1.4906 0.3091 1.4906 1.2209
No log 1.0 60 1.6596 0.2545 1.6596 1.2882
No log 1.0333 62 1.5446 0.2752 1.5446 1.2428
No log 1.0667 64 1.2600 0.4248 1.2600 1.1225
No log 1.1 66 1.3160 0.3902 1.3160 1.1472
No log 1.1333 68 1.4590 0.3710 1.4590 1.2079
No log 1.1667 70 1.3406 0.4160 1.3406 1.1578
No log 1.2 72 1.2314 0.4878 1.2314 1.1097
No log 1.2333 74 1.1939 0.4715 1.1939 1.0926
No log 1.2667 76 1.1219 0.4959 1.1219 1.0592
No log 1.3 78 1.0882 0.5645 1.0882 1.0432
No log 1.3333 80 1.1825 0.5891 1.1825 1.0874
No log 1.3667 82 1.1511 0.5581 1.1511 1.0729
No log 1.4 84 1.0908 0.5366 1.0908 1.0444
No log 1.4333 86 1.2822 0.5161 1.2822 1.1324
No log 1.4667 88 1.2501 0.4878 1.2501 1.1181
No log 1.5 90 1.0625 0.5484 1.0625 1.0308
No log 1.5333 92 1.0854 0.6357 1.0854 1.0418
No log 1.5667 94 1.2489 0.5263 1.2489 1.1175
No log 1.6 96 1.3093 0.5185 1.3093 1.1443
No log 1.6333 98 1.2067 0.5455 1.2067 1.0985
No log 1.6667 100 1.0947 0.6364 1.0947 1.0463
No log 1.7 102 1.1458 0.5970 1.1458 1.0704
No log 1.7333 104 1.3316 0.5072 1.3316 1.1539
No log 1.7667 106 1.3996 0.4857 1.3996 1.1830
No log 1.8 108 1.2482 0.5217 1.2482 1.1172
No log 1.8333 110 1.0346 0.5564 1.0346 1.0172
No log 1.8667 112 0.9585 0.6308 0.9585 0.9790
No log 1.9 114 0.9880 0.5649 0.9880 0.9940
No log 1.9333 116 0.9443 0.6107 0.9443 0.9717
No log 1.9667 118 0.9163 0.5909 0.9163 0.9572
No log 2.0 120 0.8722 0.5909 0.8722 0.9339
No log 2.0333 122 0.9261 0.6015 0.9261 0.9624
No log 2.0667 124 1.0555 0.5735 1.0555 1.0274
No log 2.1 126 0.9626 0.6119 0.9626 0.9811
No log 2.1333 128 0.8571 0.6165 0.8571 0.9258
No log 2.1667 130 0.8414 0.5985 0.8414 0.9173
No log 2.2 132 1.0041 0.5775 1.0041 1.0021
No log 2.2333 134 1.1022 0.5833 1.1022 1.0498
No log 2.2667 136 0.9818 0.6207 0.9818 0.9908
No log 2.3 138 0.9108 0.6187 0.9108 0.9544
No log 2.3333 140 1.0405 0.6434 1.0405 1.0201
No log 2.3667 142 1.6322 0.4472 1.6322 1.2776
No log 2.4 144 1.3872 0.4387 1.3872 1.1778
No log 2.4333 146 0.9413 0.625 0.9413 0.9702
No log 2.4667 148 0.9379 0.6197 0.9379 0.9685
No log 2.5 150 1.0445 0.5957 1.0445 1.0220
No log 2.5333 152 1.1319 0.5755 1.1319 1.0639
No log 2.5667 154 1.0381 0.6423 1.0381 1.0189
No log 2.6 156 1.0202 0.6331 1.0202 1.0100
No log 2.6333 158 0.9673 0.6667 0.9673 0.9835
No log 2.6667 160 0.9748 0.6241 0.9748 0.9873
No log 2.7 162 1.0099 0.6099 1.0099 1.0049
No log 2.7333 164 0.9835 0.5942 0.9835 0.9917
No log 2.7667 166 1.0129 0.5970 1.0129 1.0064
No log 2.8 168 1.1617 0.5429 1.1617 1.0778
No log 2.8333 170 1.0892 0.5373 1.0892 1.0437
No log 2.8667 172 0.9850 0.6029 0.9850 0.9925
No log 2.9 174 1.1514 0.5775 1.1514 1.0730
No log 2.9333 176 1.2168 0.5362 1.2168 1.1031
No log 2.9667 178 1.0754 0.5652 1.0754 1.0370
No log 3.0 180 0.7717 0.6618 0.7717 0.8784
No log 3.0333 182 0.6653 0.7246 0.6653 0.8156
No log 3.0667 184 0.6940 0.7206 0.6940 0.8331
No log 3.1 186 0.8245 0.6519 0.8245 0.9080
No log 3.1333 188 0.9666 0.5926 0.9666 0.9832
No log 3.1667 190 1.0762 0.5652 1.0762 1.0374
No log 3.2 192 1.1819 0.5547 1.1819 1.0871
No log 3.2333 194 1.0368 0.5693 1.0368 1.0182
No log 3.2667 196 0.8351 0.6519 0.8351 0.9138
No log 3.3 198 0.7969 0.6767 0.7969 0.8927
No log 3.3333 200 0.7903 0.6308 0.7903 0.8890
No log 3.3667 202 0.7943 0.6716 0.7943 0.8912
No log 3.4 204 0.8507 0.6423 0.8507 0.9223
No log 3.4333 206 0.9189 0.5985 0.9189 0.9586
No log 3.4667 208 0.9019 0.6029 0.9019 0.9497
No log 3.5 210 0.8219 0.7015 0.8219 0.9066
No log 3.5333 212 0.7807 0.6970 0.7807 0.8836
No log 3.5667 214 0.7814 0.6963 0.7814 0.8840
No log 3.6 216 0.7803 0.6716 0.7803 0.8833
No log 3.6333 218 0.7613 0.6667 0.7613 0.8725
No log 3.6667 220 0.7941 0.6716 0.7941 0.8911
No log 3.7 222 0.8517 0.6324 0.8517 0.9229
No log 3.7333 224 0.8907 0.6324 0.8907 0.9438
No log 3.7667 226 0.8550 0.6667 0.8550 0.9246
No log 3.8 228 0.8634 0.6617 0.8634 0.9292
No log 3.8333 230 0.9133 0.6515 0.9133 0.9557
No log 3.8667 232 0.9477 0.6212 0.9477 0.9735
No log 3.9 234 0.9666 0.6412 0.9666 0.9832
No log 3.9333 236 0.9901 0.6212 0.9901 0.9950
No log 3.9667 238 0.9928 0.6316 0.9928 0.9964
No log 4.0 240 0.9840 0.6119 0.9840 0.9920
No log 4.0333 242 0.9369 0.6260 0.9369 0.9679
No log 4.0667 244 0.9038 0.6412 0.9038 0.9507
No log 4.1 246 0.8715 0.6412 0.8715 0.9335
No log 4.1333 248 0.8397 0.6357 0.8397 0.9164
No log 4.1667 250 0.8113 0.6462 0.8113 0.9007
No log 4.2 252 0.8041 0.6462 0.8041 0.8967
No log 4.2333 254 0.7856 0.6565 0.7856 0.8864
No log 4.2667 256 0.7615 0.6815 0.7615 0.8727
No log 4.3 258 0.8100 0.6377 0.8100 0.9000
No log 4.3333 260 1.0123 0.64 1.0123 1.0062
No log 4.3667 262 1.3749 0.5802 1.3749 1.1726
No log 4.4 264 1.5284 0.4636 1.5284 1.2363
No log 4.4333 266 1.5662 0.4161 1.5662 1.2515
No log 4.4667 268 1.3930 0.5278 1.3930 1.1802
No log 4.5 270 1.2532 0.5429 1.2532 1.1195
No log 4.5333 272 1.1423 0.6043 1.1423 1.0688
No log 4.5667 274 1.1130 0.6043 1.1130 1.0550
No log 4.6 276 1.1136 0.6143 1.1136 1.0553
No log 4.6333 278 0.9761 0.6277 0.9761 0.9880
No log 4.6667 280 0.9017 0.6061 0.9017 0.9496
No log 4.7 282 0.8776 0.6061 0.8776 0.9368
No log 4.7333 284 0.9497 0.6269 0.9497 0.9745
No log 4.7667 286 0.9870 0.5985 0.9870 0.9935
No log 4.8 288 0.9125 0.6418 0.9125 0.9552
No log 4.8333 290 0.8957 0.6222 0.8957 0.9464
No log 4.8667 292 0.9287 0.6324 0.9287 0.9637
No log 4.9 294 1.0007 0.6029 1.0007 1.0003
No log 4.9333 296 1.0283 0.5839 1.0283 1.0140
No log 4.9667 298 1.0200 0.5839 1.0200 1.0099
No log 5.0 300 0.9553 0.6412 0.9553 0.9774
No log 5.0333 302 0.9231 0.5556 0.9231 0.9608
No log 5.0667 304 0.9555 0.5556 0.9555 0.9775
No log 5.1 306 0.9073 0.5846 0.9073 0.9525
No log 5.1333 308 0.9395 0.6418 0.9395 0.9693
No log 5.1667 310 0.9906 0.5797 0.9906 0.9953
No log 5.2 312 0.9928 0.5839 0.9928 0.9964
No log 5.2333 314 0.9777 0.6087 0.9777 0.9888
No log 5.2667 316 0.9193 0.6970 0.9193 0.9588
No log 5.3 318 0.8956 0.6129 0.8956 0.9463
No log 5.3333 320 0.8941 0.64 0.8941 0.9456
No log 5.3667 322 0.9061 0.6870 0.9061 0.9519
No log 5.4 324 0.9199 0.6475 0.9199 0.9591
No log 5.4333 326 0.8460 0.6571 0.8460 0.9198
No log 5.4667 328 0.8020 0.7 0.8020 0.8955
No log 5.5 330 0.7650 0.7 0.7650 0.8746
No log 5.5333 332 0.7622 0.6906 0.7622 0.8730
No log 5.5667 334 0.7814 0.6957 0.7814 0.8839
No log 5.6 336 0.8215 0.6769 0.8215 0.9063
No log 5.6333 338 0.8264 0.6769 0.8264 0.9091
No log 5.6667 340 0.8042 0.6567 0.8042 0.8968
No log 5.7 342 0.7947 0.6667 0.7947 0.8915
No log 5.7333 344 0.7705 0.6906 0.7705 0.8778
No log 5.7667 346 0.7609 0.6950 0.7609 0.8723
No log 5.8 348 0.7509 0.6950 0.7509 0.8665
No log 5.8333 350 0.7259 0.7092 0.7259 0.8520
No log 5.8667 352 0.7185 0.7194 0.7185 0.8477
No log 5.9 354 0.7003 0.7324 0.7003 0.8368
No log 5.9333 356 0.6788 0.7413 0.6788 0.8239
No log 5.9667 358 0.6801 0.7465 0.6801 0.8247
No log 6.0 360 0.7100 0.7059 0.7100 0.8426
No log 6.0333 362 0.7095 0.7259 0.7095 0.8423
No log 6.0667 364 0.7285 0.7068 0.7285 0.8535
No log 6.1 366 0.7438 0.7068 0.7438 0.8625
No log 6.1333 368 0.7366 0.7111 0.7366 0.8583
No log 6.1667 370 0.7067 0.7286 0.7067 0.8406
No log 6.2 372 0.6910 0.7465 0.6910 0.8313
No log 6.2333 374 0.7178 0.7083 0.7178 0.8472
No log 6.2667 376 0.7766 0.6486 0.7766 0.8812
No log 6.3 378 0.7880 0.6345 0.7880 0.8877
No log 6.3333 380 0.7615 0.7286 0.7615 0.8726
No log 6.3667 382 0.7649 0.6667 0.7649 0.8746
No log 6.4 384 0.7835 0.6519 0.7835 0.8852
No log 6.4333 386 0.8009 0.7246 0.8009 0.8949
No log 6.4667 388 0.8169 0.6809 0.8169 0.9038
No log 6.5 390 0.7841 0.7 0.7841 0.8855
No log 6.5333 392 0.7341 0.7153 0.7341 0.8568
No log 6.5667 394 0.7500 0.6957 0.7500 0.8660
No log 6.6 396 0.7960 0.6765 0.7960 0.8922
No log 6.6333 398 0.8221 0.6515 0.8221 0.9067
No log 6.6667 400 0.8351 0.6142 0.8351 0.9138
No log 6.7 402 0.8244 0.6870 0.8244 0.9080
No log 6.7333 404 0.7746 0.6912 0.7746 0.8801
No log 6.7667 406 0.7435 0.6519 0.7435 0.8623
No log 6.8 408 0.7392 0.6957 0.7392 0.8598
No log 6.8333 410 0.7400 0.6957 0.7400 0.8602
No log 6.8667 412 0.7330 0.6957 0.7330 0.8561
No log 6.9 414 0.7270 0.7101 0.7270 0.8527
No log 6.9333 416 0.7658 0.6917 0.7658 0.8751
No log 6.9667 418 0.7930 0.6970 0.7930 0.8905
No log 7.0 420 0.7595 0.7153 0.7595 0.8715
No log 7.0333 422 0.7582 0.6815 0.7582 0.8707
No log 7.0667 424 0.7672 0.6815 0.7672 0.8759
No log 7.1 426 0.8042 0.6718 0.8042 0.8968
No log 7.1333 428 0.8194 0.6565 0.8194 0.9052
No log 7.1667 430 0.8643 0.6615 0.8643 0.9297
No log 7.2 432 0.8784 0.6667 0.8784 0.9372
No log 7.2333 434 0.8742 0.6418 0.8742 0.9350
No log 7.2667 436 0.8865 0.6015 0.8865 0.9415
No log 7.3 438 0.8704 0.6269 0.8704 0.9329
No log 7.3333 440 0.8033 0.6569 0.8033 0.8963
No log 7.3667 442 0.7873 0.6957 0.7873 0.8873
No log 7.4 444 0.8539 0.6716 0.8539 0.9241
No log 7.4333 446 0.8554 0.6567 0.8554 0.9249
No log 7.4667 448 0.7957 0.6917 0.7957 0.8920
No log 7.5 450 0.7548 0.6765 0.7548 0.8688
No log 7.5333 452 0.7329 0.6861 0.7329 0.8561
No log 7.5667 454 0.7261 0.6906 0.7261 0.8521
No log 7.6 456 0.7324 0.6861 0.7324 0.8558
No log 7.6333 458 0.7571 0.7015 0.7571 0.8701
No log 7.6667 460 0.8070 0.6667 0.8070 0.8984
No log 7.7 462 0.7798 0.6715 0.7798 0.8831
No log 7.7333 464 0.7268 0.6857 0.7268 0.8525
No log 7.7667 466 0.7210 0.6857 0.7210 0.8491
No log 7.8 468 0.6906 0.7007 0.6906 0.8310
No log 7.8333 470 0.7135 0.7050 0.7135 0.8447
No log 7.8667 472 0.7050 0.7007 0.7050 0.8397
No log 7.9 474 0.7211 0.7050 0.7211 0.8492
No log 7.9333 476 0.7819 0.6715 0.7819 0.8843
No log 7.9667 478 0.8663 0.6475 0.8663 0.9308
No log 8.0 480 0.8670 0.6475 0.8670 0.9311
No log 8.0333 482 0.8287 0.6912 0.8287 0.9103
No log 8.0667 484 0.8309 0.6667 0.8309 0.9115
No log 8.1 486 0.8147 0.6617 0.8147 0.9026
No log 8.1333 488 0.7752 0.6906 0.7752 0.8804
No log 8.1667 490 0.7524 0.7273 0.7524 0.8674
No log 8.2 492 0.7748 0.7183 0.7748 0.8802
No log 8.2333 494 0.8013 0.6957 0.8013 0.8951
No log 8.2667 496 0.8264 0.6861 0.8264 0.9091
No log 8.3 498 0.8519 0.6618 0.8519 0.9230
0.4549 8.3333 500 0.8927 0.6471 0.8927 0.9448
0.4549 8.3667 502 0.8855 0.6715 0.8855 0.9410
0.4549 8.4 504 0.8823 0.6364 0.8823 0.9393
0.4549 8.4333 506 0.8405 0.6667 0.8405 0.9168
0.4549 8.4667 508 0.8152 0.6912 0.8152 0.9029
0.4549 8.5 510 0.8049 0.7234 0.8049 0.8972
0.4549 8.5333 512 0.8123 0.6957 0.8123 0.9013
0.4549 8.5667 514 0.8280 0.6957 0.8280 0.9099
0.4549 8.6 516 0.8498 0.6812 0.8498 0.9218
0.4549 8.6333 518 0.8995 0.6277 0.8995 0.9484
0.4549 8.6667 520 0.8880 0.6222 0.8880 0.9423
0.4549 8.7 522 0.8634 0.6423 0.8634 0.9292
0.4549 8.7333 524 0.8392 0.6567 0.8392 0.9161
0.4549 8.7667 526 0.8035 0.6567 0.8035 0.8964
0.4549 8.8 528 0.7889 0.6569 0.7889 0.8882
0.4549 8.8333 530 0.8655 0.6429 0.8655 0.9303
0.4549 8.8667 532 1.0093 0.6289 1.0093 1.0046
0.4549 8.9 534 1.0353 0.5976 1.0353 1.0175
0.4549 8.9333 536 0.9635 0.6483 0.9635 0.9816
0.4549 8.9667 538 0.8616 0.6713 0.8616 0.9282
0.4549 9.0 540 0.8331 0.6806 0.8331 0.9127
0.4549 9.0333 542 0.8159 0.6667 0.8159 0.9033
0.4549 9.0667 544 0.8372 0.6423 0.8372 0.9150
0.4549 9.1 546 0.9496 0.6377 0.9496 0.9745
0.4549 9.1333 548 0.9935 0.6286 0.9935 0.9967
0.4549 9.1667 550 0.9469 0.6222 0.9469 0.9731
0.4549 9.2 552 0.8592 0.6269 0.8592 0.9270

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k13_task1_organization

Finetuned
(4023)
this model