ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k5_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6890
  • Qwk: 0.6143
  • Mse: 0.6890
  • Rmse: 0.8301

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1176 2 4.1453 0.0079 4.1454 2.0360
No log 0.2353 4 2.2261 0.0454 2.2261 1.4920
No log 0.3529 6 1.3581 0.0 1.3581 1.1654
No log 0.4706 8 1.1039 0.2125 1.1039 1.0507
No log 0.5882 10 1.2760 -0.0560 1.2760 1.1296
No log 0.7059 12 1.2482 0.0085 1.2482 1.1172
No log 0.8235 14 1.2633 0.0085 1.2633 1.1240
No log 0.9412 16 1.1228 0.1176 1.1228 1.0596
No log 1.0588 18 1.0810 0.1874 1.0810 1.0397
No log 1.1765 20 1.1194 0.1389 1.1194 1.0580
No log 1.2941 22 1.2283 0.0967 1.2283 1.1083
No log 1.4118 24 1.3610 0.0085 1.3610 1.1666
No log 1.5294 26 1.4481 0.0085 1.4481 1.2034
No log 1.6471 28 1.3528 0.0970 1.3528 1.1631
No log 1.7647 30 1.0162 0.2192 1.0162 1.0081
No log 1.8824 32 0.9561 0.2517 0.9561 0.9778
No log 2.0 34 0.9692 0.2818 0.9692 0.9845
No log 2.1176 36 1.0197 0.2496 1.0197 1.0098
No log 2.2353 38 1.1990 0.2177 1.1990 1.0950
No log 2.3529 40 1.0467 0.3015 1.0467 1.0231
No log 2.4706 42 0.8285 0.4810 0.8285 0.9102
No log 2.5882 44 0.7934 0.4759 0.7934 0.8907
No log 2.7059 46 0.8084 0.4822 0.8084 0.8991
No log 2.8235 48 0.8352 0.4513 0.8352 0.9139
No log 2.9412 50 0.8232 0.4021 0.8232 0.9073
No log 3.0588 52 0.8379 0.3852 0.8379 0.9154
No log 3.1765 54 0.7466 0.5331 0.7466 0.8641
No log 3.2941 56 0.9498 0.3943 0.9498 0.9746
No log 3.4118 58 1.0788 0.3941 1.0788 1.0386
No log 3.5294 60 0.8242 0.5707 0.8242 0.9079
No log 3.6471 62 0.7122 0.5618 0.7122 0.8439
No log 3.7647 64 0.6714 0.5712 0.6714 0.8194
No log 3.8824 66 0.6585 0.6464 0.6585 0.8115
No log 4.0 68 0.8052 0.5705 0.8052 0.8973
No log 4.1176 70 0.7546 0.6064 0.7546 0.8687
No log 4.2353 72 0.6587 0.5909 0.6587 0.8116
No log 4.3529 74 0.6741 0.6125 0.6741 0.8211
No log 4.4706 76 0.7403 0.6003 0.7403 0.8604
No log 4.5882 78 0.7351 0.5884 0.7351 0.8574
No log 4.7059 80 0.6428 0.5796 0.6428 0.8017
No log 4.8235 82 0.6697 0.5317 0.6697 0.8183
No log 4.9412 84 0.6490 0.5859 0.6490 0.8056
No log 5.0588 86 0.6891 0.5243 0.6891 0.8301
No log 5.1765 88 0.8608 0.5270 0.8608 0.9278
No log 5.2941 90 0.7080 0.5366 0.7080 0.8414
No log 5.4118 92 0.6650 0.5835 0.6650 0.8155
No log 5.5294 94 0.6478 0.6241 0.6478 0.8049
No log 5.6471 96 0.6261 0.6078 0.6261 0.7913
No log 5.7647 98 0.6269 0.6060 0.6269 0.7918
No log 5.8824 100 0.6061 0.6167 0.6061 0.7785
No log 6.0 102 0.7356 0.6568 0.7356 0.8576
No log 6.1176 104 0.6662 0.5984 0.6662 0.8162
No log 6.2353 106 0.6395 0.6433 0.6395 0.7997
No log 6.3529 108 0.6770 0.6712 0.6770 0.8228
No log 6.4706 110 0.5875 0.6399 0.5875 0.7665
No log 6.5882 112 0.5987 0.6853 0.5987 0.7738
No log 6.7059 114 0.5839 0.6510 0.5839 0.7642
No log 6.8235 116 0.6038 0.6876 0.6038 0.7771
No log 6.9412 118 0.5934 0.6627 0.5934 0.7703
No log 7.0588 120 0.5813 0.6888 0.5813 0.7624
No log 7.1765 122 0.5877 0.6555 0.5877 0.7666
No log 7.2941 124 0.5928 0.6737 0.5928 0.7699
No log 7.4118 126 0.5965 0.6737 0.5965 0.7723
No log 7.5294 128 0.5803 0.6903 0.5803 0.7618
No log 7.6471 130 0.5816 0.6820 0.5816 0.7626
No log 7.7647 132 0.6386 0.6573 0.6386 0.7991
No log 7.8824 134 0.6173 0.6715 0.6173 0.7857
No log 8.0 136 0.5776 0.7458 0.5776 0.7600
No log 8.1176 138 0.5623 0.6840 0.5623 0.7498
No log 8.2353 140 0.5551 0.6903 0.5551 0.7451
No log 8.3529 142 0.5702 0.6820 0.5702 0.7551
No log 8.4706 144 0.5552 0.6584 0.5552 0.7451
No log 8.5882 146 0.5753 0.6606 0.5753 0.7585
No log 8.7059 148 0.5858 0.6207 0.5858 0.7654
No log 8.8235 150 0.6028 0.5328 0.6028 0.7764
No log 8.9412 152 0.5659 0.6546 0.5659 0.7523
No log 9.0588 154 0.6254 0.6774 0.6254 0.7908
No log 9.1765 156 0.5890 0.6519 0.5890 0.7675
No log 9.2941 158 0.5679 0.6896 0.5679 0.7536
No log 9.4118 160 0.5659 0.6659 0.5659 0.7522
No log 9.5294 162 0.5664 0.6581 0.5664 0.7526
No log 9.6471 164 0.5706 0.6455 0.5706 0.7554
No log 9.7647 166 0.5982 0.6157 0.5982 0.7734
No log 9.8824 168 0.5735 0.6606 0.5735 0.7573
No log 10.0 170 0.6061 0.6119 0.6061 0.7785
No log 10.1176 172 0.5961 0.6228 0.5961 0.7721
No log 10.2353 174 0.6164 0.6043 0.6164 0.7851
No log 10.3529 176 0.7055 0.6230 0.7055 0.8399
No log 10.4706 178 0.6685 0.6230 0.6685 0.8176
No log 10.5882 180 0.5829 0.6641 0.5829 0.7635
No log 10.7059 182 0.6061 0.6288 0.6061 0.7785
No log 10.8235 184 0.5954 0.6497 0.5954 0.7716
No log 10.9412 186 0.5857 0.6497 0.5857 0.7653
No log 11.0588 188 0.5849 0.6566 0.5849 0.7648
No log 11.1765 190 0.6239 0.6241 0.6239 0.7899
No log 11.2941 192 0.6665 0.5833 0.6665 0.8164
No log 11.4118 194 0.6595 0.6082 0.6595 0.8121
No log 11.5294 196 0.6144 0.6350 0.6144 0.7838
No log 11.6471 198 0.5666 0.6729 0.5666 0.7527
No log 11.7647 200 0.5733 0.6667 0.5733 0.7571
No log 11.8824 202 0.5695 0.6788 0.5695 0.7547
No log 12.0 204 0.6088 0.6656 0.6088 0.7803
No log 12.1176 206 0.6309 0.6656 0.6309 0.7943
No log 12.2353 208 0.5926 0.6838 0.5926 0.7698
No log 12.3529 210 0.7935 0.5666 0.7935 0.8908
No log 12.4706 212 0.7472 0.5759 0.7472 0.8644
No log 12.5882 214 0.5794 0.6795 0.5794 0.7612
No log 12.7059 216 0.5855 0.6743 0.5855 0.7652
No log 12.8235 218 0.7412 0.5856 0.7412 0.8609
No log 12.9412 220 0.6913 0.5968 0.6913 0.8314
No log 13.0588 222 0.6017 0.6488 0.6017 0.7757
No log 13.1765 224 0.6131 0.5782 0.6131 0.7830
No log 13.2941 226 0.6438 0.5977 0.6438 0.8023
No log 13.4118 228 0.6013 0.6643 0.6013 0.7754
No log 13.5294 230 0.5835 0.6455 0.5835 0.7639
No log 13.6471 232 0.7137 0.6199 0.7137 0.8448
No log 13.7647 234 0.7222 0.6019 0.7222 0.8498
No log 13.8824 236 0.6215 0.6361 0.6215 0.7884
No log 14.0 238 0.6089 0.6185 0.6089 0.7803
No log 14.1176 240 0.6910 0.5565 0.6910 0.8313
No log 14.2353 242 0.6500 0.5860 0.6500 0.8063
No log 14.3529 244 0.5764 0.6729 0.5764 0.7592
No log 14.4706 246 0.7279 0.6363 0.7279 0.8532
No log 14.5882 248 0.8363 0.5499 0.8363 0.9145
No log 14.7059 250 0.7337 0.5655 0.7337 0.8566
No log 14.8235 252 0.6042 0.6933 0.6042 0.7773
No log 14.9412 254 0.6522 0.4888 0.6522 0.8076
No log 15.0588 256 0.7187 0.5359 0.7187 0.8478
No log 15.1765 258 0.6698 0.5132 0.6698 0.8184
No log 15.2941 260 0.6004 0.6024 0.6004 0.7748
No log 15.4118 262 0.6814 0.5775 0.6814 0.8255
No log 15.5294 264 0.7403 0.5515 0.7403 0.8604
No log 15.6471 266 0.6798 0.5463 0.6798 0.8245
No log 15.7647 268 0.6298 0.5432 0.6298 0.7936
No log 15.8824 270 0.6108 0.5555 0.6108 0.7815
No log 16.0 272 0.5878 0.6282 0.5878 0.7667
No log 16.1176 274 0.6125 0.6328 0.6125 0.7826
No log 16.2353 276 0.6130 0.6147 0.6130 0.7829
No log 16.3529 278 0.6524 0.6230 0.6524 0.8077
No log 16.4706 280 0.6317 0.6249 0.6317 0.7948
No log 16.5882 282 0.5934 0.6123 0.5934 0.7703
No log 16.7059 284 0.6089 0.6113 0.6089 0.7803
No log 16.8235 286 0.6040 0.6065 0.6040 0.7772
No log 16.9412 288 0.5912 0.6307 0.5912 0.7689
No log 17.0588 290 0.5853 0.6657 0.5853 0.7651
No log 17.1765 292 0.5792 0.6307 0.5792 0.7611
No log 17.2941 294 0.5672 0.6547 0.5672 0.7532
No log 17.4118 296 0.5836 0.6684 0.5836 0.7639
No log 17.5294 298 0.5819 0.6602 0.5819 0.7628
No log 17.6471 300 0.5690 0.6282 0.5690 0.7543
No log 17.7647 302 0.5761 0.6491 0.5761 0.7590
No log 17.8824 304 0.6327 0.6556 0.6327 0.7954
No log 18.0 306 0.6197 0.6556 0.6197 0.7872
No log 18.1176 308 0.5496 0.6712 0.5496 0.7414
No log 18.2353 310 0.5552 0.6427 0.5552 0.7451
No log 18.3529 312 0.5897 0.5783 0.5897 0.7679
No log 18.4706 314 0.5965 0.6307 0.5965 0.7723
No log 18.5882 316 0.6153 0.6578 0.6153 0.7844
No log 18.7059 318 0.6374 0.6319 0.6374 0.7984
No log 18.8235 320 0.6097 0.6511 0.6097 0.7808
No log 18.9412 322 0.5881 0.6324 0.5881 0.7669
No log 19.0588 324 0.5980 0.6589 0.5980 0.7733
No log 19.1765 326 0.6375 0.6774 0.6375 0.7985
No log 19.2941 328 0.6353 0.6502 0.6353 0.7971
No log 19.4118 330 0.6195 0.6451 0.6195 0.7871
No log 19.5294 332 0.6196 0.6451 0.6196 0.7871
No log 19.6471 334 0.6131 0.5969 0.6131 0.7830
No log 19.7647 336 0.5958 0.6314 0.5958 0.7719
No log 19.8824 338 0.5751 0.6526 0.5751 0.7584
No log 20.0 340 0.6037 0.6288 0.6037 0.7770
No log 20.1176 342 0.5946 0.6174 0.5946 0.7711
No log 20.2353 344 0.6233 0.6237 0.6233 0.7895
No log 20.3529 346 0.7215 0.5968 0.7215 0.8494
No log 20.4706 348 0.6896 0.6015 0.6896 0.8304
No log 20.5882 350 0.6140 0.5972 0.6140 0.7836
No log 20.7059 352 0.6083 0.5315 0.6083 0.7799
No log 20.8235 354 0.6073 0.5202 0.6073 0.7793
No log 20.9412 356 0.6049 0.5983 0.6049 0.7778
No log 21.0588 358 0.6892 0.6032 0.6892 0.8302
No log 21.1765 360 0.7245 0.6079 0.7245 0.8511
No log 21.2941 362 0.6814 0.6377 0.6814 0.8254
No log 21.4118 364 0.6109 0.6511 0.6109 0.7816
No log 21.5294 366 0.5988 0.6617 0.5988 0.7738
No log 21.6471 368 0.6510 0.6823 0.6510 0.8068
No log 21.7647 370 0.6533 0.6720 0.6533 0.8082
No log 21.8824 372 0.5851 0.6177 0.5851 0.7649
No log 22.0 374 0.5748 0.6476 0.5748 0.7582
No log 22.1176 376 0.5785 0.6476 0.5785 0.7606
No log 22.2353 378 0.5823 0.5990 0.5823 0.7631
No log 22.3529 380 0.6705 0.6247 0.6705 0.8188
No log 22.4706 382 0.6899 0.6331 0.6899 0.8306
No log 22.5882 384 0.6270 0.6284 0.6270 0.7919
No log 22.7059 386 0.5760 0.5959 0.5760 0.7589
No log 22.8235 388 0.5840 0.6636 0.5840 0.7642
No log 22.9412 390 0.5774 0.6452 0.5774 0.7598
No log 23.0588 392 0.5904 0.6356 0.5904 0.7683
No log 23.1765 394 0.6305 0.6380 0.6305 0.7941
No log 23.2941 396 0.6321 0.6207 0.6321 0.7951
No log 23.4118 398 0.6016 0.6460 0.6016 0.7756
No log 23.5294 400 0.5722 0.6479 0.5722 0.7564
No log 23.6471 402 0.5627 0.6095 0.5627 0.7501
No log 23.7647 404 0.5568 0.6391 0.5568 0.7462
No log 23.8824 406 0.5651 0.6672 0.5651 0.7517
No log 24.0 408 0.5646 0.6760 0.5646 0.7514
No log 24.1176 410 0.5543 0.6224 0.5543 0.7445
No log 24.2353 412 0.5806 0.6644 0.5806 0.7620
No log 24.3529 414 0.5900 0.6507 0.5900 0.7681
No log 24.4706 416 0.5582 0.6164 0.5582 0.7472
No log 24.5882 418 0.5743 0.6335 0.5743 0.7578
No log 24.7059 420 0.6045 0.6678 0.6045 0.7775
No log 24.8235 422 0.6087 0.6766 0.6087 0.7802
No log 24.9412 424 0.5901 0.6198 0.5901 0.7682
No log 25.0588 426 0.5928 0.6581 0.5928 0.7699
No log 25.1765 428 0.6219 0.6564 0.6219 0.7886
No log 25.2941 430 0.6404 0.6170 0.6404 0.8003
No log 25.4118 432 0.6552 0.5978 0.6552 0.8095
No log 25.5294 434 0.6107 0.6217 0.6107 0.7815
No log 25.6471 436 0.5900 0.6762 0.5900 0.7681
No log 25.7647 438 0.5923 0.6625 0.5923 0.7696
No log 25.8824 440 0.5959 0.6470 0.5959 0.7720
No log 26.0 442 0.6146 0.6347 0.6146 0.7840
No log 26.1176 444 0.5998 0.6564 0.5998 0.7744
No log 26.2353 446 0.5903 0.6433 0.5903 0.7683
No log 26.3529 448 0.5930 0.6667 0.5930 0.7701
No log 26.4706 450 0.5942 0.6317 0.5942 0.7708
No log 26.5882 452 0.5980 0.6196 0.5980 0.7733
No log 26.7059 454 0.6011 0.6076 0.6011 0.7753
No log 26.8235 456 0.6009 0.5669 0.6009 0.7752
No log 26.9412 458 0.5992 0.6036 0.5992 0.7741
No log 27.0588 460 0.5912 0.6317 0.5912 0.7689
No log 27.1765 462 0.5804 0.6272 0.5804 0.7618
No log 27.2941 464 0.5622 0.6364 0.5622 0.7498
No log 27.4118 466 0.5690 0.6312 0.5690 0.7543
No log 27.5294 468 0.6114 0.5987 0.6114 0.7819
No log 27.6471 470 0.5990 0.5566 0.5990 0.7739
No log 27.7647 472 0.5704 0.6476 0.5704 0.7552
No log 27.8824 474 0.5956 0.6704 0.5956 0.7718
No log 28.0 476 0.6557 0.6377 0.6557 0.8097
No log 28.1176 478 0.6888 0.6220 0.6888 0.8299
No log 28.2353 480 0.6639 0.6252 0.6639 0.8148
No log 28.3529 482 0.6044 0.6593 0.6044 0.7774
No log 28.4706 484 0.5580 0.6610 0.5580 0.7470
No log 28.5882 486 0.5393 0.6610 0.5393 0.7344
No log 28.7059 488 0.5402 0.6247 0.5402 0.7350
No log 28.8235 490 0.5452 0.6421 0.5452 0.7384
No log 28.9412 492 0.5694 0.6357 0.5694 0.7546
No log 29.0588 494 0.5662 0.6247 0.5662 0.7525
No log 29.1765 496 0.5509 0.6424 0.5509 0.7422
No log 29.2941 498 0.5566 0.6311 0.5566 0.7461
0.2214 29.4118 500 0.5968 0.6325 0.5968 0.7725
0.2214 29.5294 502 0.6290 0.5751 0.6290 0.7931
0.2214 29.6471 504 0.6324 0.5775 0.6324 0.7952
0.2214 29.7647 506 0.6042 0.6432 0.6042 0.7773
0.2214 29.8824 508 0.5648 0.6424 0.5648 0.7515
0.2214 30.0 510 0.5449 0.6354 0.5449 0.7382
0.2214 30.1176 512 0.5449 0.6564 0.5449 0.7382
0.2214 30.2353 514 0.5497 0.6919 0.5497 0.7414
0.2214 30.3529 516 0.5341 0.6911 0.5341 0.7308
0.2214 30.4706 518 0.5400 0.6491 0.5400 0.7349
0.2214 30.5882 520 0.5505 0.6482 0.5505 0.7419
0.2214 30.7059 522 0.5460 0.6415 0.5460 0.7389
0.2214 30.8235 524 0.5413 0.6433 0.5413 0.7357
0.2214 30.9412 526 0.5430 0.6813 0.5430 0.7369
0.2214 31.0588 528 0.5424 0.6770 0.5424 0.7365
0.2214 31.1765 530 0.5376 0.6874 0.5376 0.7332
0.2214 31.2941 532 0.5386 0.6874 0.5386 0.7339
0.2214 31.4118 534 0.5381 0.6649 0.5381 0.7336
0.2214 31.5294 536 0.5409 0.6649 0.5409 0.7355
0.2214 31.6471 538 0.5426 0.6649 0.5426 0.7366
0.2214 31.7647 540 0.5471 0.6690 0.5471 0.7397
0.2214 31.8824 542 0.5645 0.6327 0.5645 0.7513
0.2214 32.0 544 0.5834 0.5894 0.5834 0.7638
0.2214 32.1176 546 0.5756 0.5894 0.5756 0.7587
0.2214 32.2353 548 0.5661 0.6164 0.5661 0.7524
0.2214 32.3529 550 0.5775 0.6424 0.5775 0.7600
0.2214 32.4706 552 0.5704 0.6320 0.5704 0.7552
0.2214 32.5882 554 0.5782 0.6424 0.5782 0.7604
0.2214 32.7059 556 0.5843 0.6196 0.5843 0.7644
0.2214 32.8235 558 0.5994 0.5542 0.5994 0.7742
0.2214 32.9412 560 0.6063 0.5542 0.6063 0.7787
0.2214 33.0588 562 0.6098 0.5542 0.6098 0.7809
0.2214 33.1765 564 0.6452 0.6335 0.6452 0.8032
0.2214 33.2941 566 0.6751 0.6241 0.6751 0.8217
0.2214 33.4118 568 0.6958 0.6143 0.6958 0.8341
0.2214 33.5294 570 0.7069 0.6143 0.7069 0.8408
0.2214 33.6471 572 0.6921 0.6143 0.6921 0.8320
0.2214 33.7647 574 0.6890 0.6143 0.6890 0.8301

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k5_task5_organization

Finetuned
(4019)
this model