ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7230
  • Qwk: 0.7027
  • Mse: 0.7230
  • Rmse: 0.8503

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 6.7074 0.0188 6.7074 2.5899
No log 0.1739 4 4.3105 0.0779 4.3105 2.0762
No log 0.2609 6 3.0986 0.0585 3.0986 1.7603
No log 0.3478 8 2.6383 0.1227 2.6383 1.6243
No log 0.4348 10 2.1775 0.1385 2.1775 1.4756
No log 0.5217 12 1.9261 0.1008 1.9261 1.3878
No log 0.6087 14 1.6065 0.3051 1.6065 1.2675
No log 0.6957 16 1.5411 0.3866 1.5411 1.2414
No log 0.7826 18 1.5933 0.384 1.5933 1.2622
No log 0.8696 20 1.8375 0.3433 1.8375 1.3556
No log 0.9565 22 1.5974 0.4127 1.5974 1.2639
No log 1.0435 24 1.6175 0.3866 1.6175 1.2718
No log 1.1304 26 1.4327 0.4333 1.4327 1.1970
No log 1.2174 28 1.6745 0.4328 1.6745 1.2940
No log 1.3043 30 1.8622 0.3597 1.8622 1.3646
No log 1.3913 32 2.0158 0.2857 2.0158 1.4198
No log 1.4783 34 2.1250 0.2632 2.1250 1.4577
No log 1.5652 36 1.8860 0.4110 1.8860 1.3733
No log 1.6522 38 1.8951 0.4106 1.8951 1.3766
No log 1.7391 40 2.0633 0.4025 2.0633 1.4364
No log 1.8261 42 1.6501 0.4342 1.6501 1.2846
No log 1.9130 44 1.2329 0.6027 1.2329 1.1103
No log 2.0 46 1.2056 0.6027 1.2056 1.0980
No log 2.0870 48 1.4134 0.4810 1.4134 1.1889
No log 2.1739 50 1.3269 0.4832 1.3269 1.1519
No log 2.2609 52 1.2456 0.5286 1.2456 1.1161
No log 2.3478 54 1.1645 0.5152 1.1645 1.0791
No log 2.4348 56 1.0081 0.6286 1.0081 1.0041
No log 2.5217 58 1.1613 0.5578 1.1613 1.0777
No log 2.6087 60 1.6965 0.3893 1.6965 1.3025
No log 2.6957 62 1.8432 0.3922 1.8432 1.3576
No log 2.7826 64 1.2619 0.5405 1.2619 1.1234
No log 2.8696 66 0.8592 0.7172 0.8592 0.9269
No log 2.9565 68 0.8090 0.6806 0.8090 0.8995
No log 3.0435 70 0.8417 0.7448 0.8417 0.9175
No log 3.1304 72 0.9469 0.6892 0.9469 0.9731
No log 3.2174 74 1.2137 0.56 1.2137 1.1017
No log 3.3043 76 1.0752 0.5714 1.0752 1.0369
No log 3.3913 78 0.8220 0.7194 0.8220 0.9066
No log 3.4783 80 0.8817 0.6667 0.8817 0.9390
No log 3.5652 82 0.8167 0.6993 0.8167 0.9037
No log 3.6522 84 0.7904 0.7123 0.7904 0.8891
No log 3.7391 86 0.8338 0.7143 0.8338 0.9131
No log 3.8261 88 0.7947 0.7451 0.7947 0.8915
No log 3.9130 90 0.7485 0.7067 0.7485 0.8651
No log 4.0 92 0.8002 0.7285 0.8002 0.8946
No log 4.0870 94 0.7786 0.7152 0.7786 0.8824
No log 4.1739 96 0.7921 0.7333 0.7921 0.8900
No log 4.2609 98 1.0459 0.6548 1.0459 1.0227
No log 4.3478 100 1.3655 0.5902 1.3655 1.1685
No log 4.4348 102 1.2201 0.6180 1.2201 1.1046
No log 4.5217 104 0.8873 0.75 0.8873 0.9420
No log 4.6087 106 0.7342 0.7333 0.7342 0.8568
No log 4.6957 108 0.6964 0.7550 0.6964 0.8345
No log 4.7826 110 0.7813 0.7738 0.7813 0.8839
No log 4.8696 112 1.1284 0.6595 1.1284 1.0622
No log 4.9565 114 1.1088 0.6595 1.1088 1.0530
No log 5.0435 116 0.8575 0.7701 0.8575 0.9260
No log 5.1304 118 0.6566 0.7114 0.6566 0.8103
No log 5.2174 120 0.6793 0.7417 0.6793 0.8242
No log 5.3043 122 0.6932 0.7285 0.6932 0.8326
No log 5.3913 124 0.6824 0.7368 0.6824 0.8261
No log 5.4783 126 0.6829 0.7673 0.6829 0.8264
No log 5.5652 128 0.7375 0.7886 0.7375 0.8588
No log 5.6522 130 0.8322 0.7650 0.8322 0.9123
No log 5.7391 132 0.8082 0.7598 0.8082 0.8990
No log 5.8261 134 0.7605 0.8095 0.7605 0.8720
No log 5.9130 136 0.7326 0.8125 0.7326 0.8559
No log 6.0 138 0.7410 0.7568 0.7410 0.8608
No log 6.0870 140 0.8142 0.6429 0.8142 0.9023
No log 6.1739 142 0.8438 0.6667 0.8438 0.9186
No log 6.2609 144 0.8457 0.7006 0.8457 0.9196
No log 6.3478 146 0.6991 0.7582 0.6991 0.8361
No log 6.4348 148 0.6156 0.7755 0.6156 0.7846
No log 6.5217 150 0.6051 0.7692 0.6051 0.7779
No log 6.6087 152 0.6500 0.7771 0.6500 0.8062
No log 6.6957 154 0.7385 0.7853 0.7385 0.8594
No log 6.7826 156 0.8019 0.7879 0.8019 0.8955
No log 6.8696 158 0.7187 0.7595 0.7187 0.8478
No log 6.9565 160 0.6384 0.7742 0.6384 0.7990
No log 7.0435 162 0.6775 0.7368 0.6775 0.8231
No log 7.1304 164 0.6353 0.7632 0.6353 0.7970
No log 7.2174 166 0.6485 0.8114 0.6485 0.8053
No log 7.3043 168 0.8616 0.7182 0.8616 0.9282
No log 7.3913 170 0.9268 0.7111 0.9268 0.9627
No log 7.4783 172 0.7626 0.7778 0.7626 0.8733
No log 7.5652 174 0.7413 0.7347 0.7413 0.8610
No log 7.6522 176 0.9008 0.7007 0.9008 0.9491
No log 7.7391 178 0.9586 0.6917 0.9586 0.9791
No log 7.8261 180 0.9198 0.6260 0.9198 0.9591
No log 7.9130 182 0.8236 0.6567 0.8236 0.9075
No log 8.0 184 0.7901 0.6765 0.7901 0.8889
No log 8.0870 186 0.7373 0.7310 0.7373 0.8587
No log 8.1739 188 0.6846 0.7867 0.6846 0.8274
No log 8.2609 190 0.6526 0.7483 0.6526 0.8079
No log 8.3478 192 0.6433 0.7867 0.6433 0.8020
No log 8.4348 194 0.7210 0.7922 0.7210 0.8491
No log 8.5217 196 0.7914 0.6892 0.7914 0.8896
No log 8.6087 198 0.7862 0.6331 0.7862 0.8867
No log 8.6957 200 0.7415 0.7273 0.7415 0.8611
No log 8.7826 202 0.8183 0.7206 0.8183 0.9046
No log 8.8696 204 0.8654 0.6767 0.8654 0.9303
No log 8.9565 206 0.8295 0.6471 0.8295 0.9108
No log 9.0435 208 0.7660 0.6525 0.7660 0.8752
No log 9.1304 210 0.7954 0.7261 0.7954 0.8918
No log 9.2174 212 0.8268 0.7152 0.8268 0.9093
No log 9.3043 214 0.7279 0.7826 0.7279 0.8531
No log 9.3913 216 0.6359 0.7222 0.6359 0.7974
No log 9.4783 218 0.7564 0.7338 0.7564 0.8697
No log 9.5652 220 0.8322 0.7050 0.8322 0.9122
No log 9.6522 222 0.7428 0.7222 0.7428 0.8619
No log 9.7391 224 0.6360 0.7347 0.6360 0.7975
No log 9.8261 226 0.6607 0.8171 0.6607 0.8128
No log 9.9130 228 0.6765 0.8193 0.6765 0.8225
No log 10.0 230 0.6708 0.7898 0.6708 0.8190
No log 10.0870 232 0.7879 0.7123 0.7879 0.8877
No log 10.1739 234 0.8111 0.7172 0.8111 0.9006
No log 10.2609 236 0.7367 0.6897 0.7367 0.8583
No log 10.3478 238 0.6957 0.7260 0.6957 0.8341
No log 10.4348 240 0.7836 0.6759 0.7836 0.8852
No log 10.5217 242 0.8224 0.6621 0.8224 0.9069
No log 10.6087 244 0.7297 0.7297 0.7297 0.8542
No log 10.6957 246 0.6854 0.7075 0.6854 0.8279
No log 10.7826 248 0.7238 0.7027 0.7238 0.8508
No log 10.8696 250 0.7610 0.6933 0.7610 0.8723
No log 10.9565 252 0.7527 0.6933 0.7527 0.8676
No log 11.0435 254 0.7397 0.7329 0.7397 0.8601
No log 11.1304 256 0.7863 0.7342 0.7863 0.8868
No log 11.2174 258 0.8035 0.7152 0.8035 0.8964
No log 11.3043 260 0.7865 0.7027 0.7865 0.8869
No log 11.3913 262 0.7925 0.6939 0.7925 0.8902
No log 11.4783 264 0.7814 0.7114 0.7814 0.8840
No log 11.5652 266 0.7682 0.7432 0.7682 0.8765
No log 11.6522 268 0.7388 0.7211 0.7388 0.8596
No log 11.7391 270 0.7184 0.7651 0.7184 0.8476
No log 11.8261 272 0.7072 0.7763 0.7072 0.8410
No log 11.9130 274 0.7154 0.7467 0.7154 0.8458
No log 12.0 276 0.7256 0.7534 0.7256 0.8518
No log 12.0870 278 0.7634 0.7361 0.7634 0.8737
No log 12.1739 280 0.7794 0.7194 0.7794 0.8829
No log 12.2609 282 0.7966 0.7246 0.7966 0.8925
No log 12.3478 284 0.7406 0.7517 0.7406 0.8606
No log 12.4348 286 0.7307 0.7368 0.7307 0.8548
No log 12.5217 288 0.7561 0.7564 0.7561 0.8695
No log 12.6087 290 0.7638 0.7771 0.7638 0.8740
No log 12.6957 292 0.7439 0.7484 0.7439 0.8625
No log 12.7826 294 0.7570 0.7273 0.7570 0.8701
No log 12.8696 296 0.7293 0.7273 0.7293 0.8540
No log 12.9565 298 0.6963 0.7550 0.6963 0.8345
No log 13.0435 300 0.7539 0.7613 0.7539 0.8683
No log 13.1304 302 0.7485 0.7467 0.7485 0.8652
No log 13.2174 304 0.7028 0.7550 0.7028 0.8383
No log 13.3043 306 0.6984 0.75 0.6984 0.8357
No log 13.3913 308 0.7116 0.75 0.7116 0.8436
No log 13.4783 310 0.7444 0.75 0.7444 0.8628
No log 13.5652 312 0.7738 0.7237 0.7738 0.8797
No log 13.6522 314 0.8090 0.7059 0.8090 0.8995
No log 13.7391 316 0.8466 0.6887 0.8466 0.9201
No log 13.8261 318 0.8228 0.7114 0.8228 0.9071
No log 13.9130 320 0.7931 0.6853 0.7931 0.8905
No log 14.0 322 0.7473 0.7355 0.7473 0.8644
No log 14.0870 324 0.7287 0.7451 0.7287 0.8536
No log 14.1739 326 0.7333 0.7075 0.7333 0.8563
No log 14.2609 328 0.7340 0.7075 0.7340 0.8567
No log 14.3478 330 0.7063 0.7248 0.7063 0.8404
No log 14.4348 332 0.6688 0.7632 0.6688 0.8178
No log 14.5217 334 0.6572 0.7742 0.6572 0.8107
No log 14.6087 336 0.6515 0.7742 0.6515 0.8072
No log 14.6957 338 0.6708 0.7950 0.6708 0.8190
No log 14.7826 340 0.6731 0.7875 0.6731 0.8204
No log 14.8696 342 0.6613 0.8025 0.6613 0.8132
No log 14.9565 344 0.6783 0.7792 0.6783 0.8236
No log 15.0435 346 0.7174 0.8101 0.7174 0.8470
No log 15.1304 348 0.6994 0.7733 0.6994 0.8363
No log 15.2174 350 0.7050 0.7383 0.7050 0.8397
No log 15.3043 352 0.7196 0.7550 0.7196 0.8483
No log 15.3913 354 0.7175 0.7792 0.7175 0.8471
No log 15.4783 356 0.6944 0.7673 0.6944 0.8333
No log 15.5652 358 0.6820 0.7799 0.6820 0.8258
No log 15.6522 360 0.7772 0.6986 0.7772 0.8816
No log 15.7391 362 0.8735 0.7042 0.8735 0.9346
No log 15.8261 364 0.8611 0.6571 0.8611 0.9280
No log 15.9130 366 0.8234 0.6853 0.8234 0.9074
No log 16.0 368 0.8315 0.6667 0.8315 0.9118
No log 16.0870 370 0.8010 0.6667 0.8010 0.8950
No log 16.1739 372 0.7512 0.7114 0.7512 0.8667
No log 16.2609 374 0.7291 0.7248 0.7291 0.8538
No log 16.3478 376 0.7731 0.7162 0.7731 0.8793
No log 16.4348 378 0.8129 0.6944 0.8129 0.9016
No log 16.5217 380 0.8660 0.7092 0.8660 0.9306
No log 16.6087 382 0.8852 0.6906 0.8852 0.9408
No log 16.6957 384 0.8671 0.6667 0.8671 0.9312
No log 16.7826 386 0.8511 0.6241 0.8511 0.9226
No log 16.8696 388 0.8076 0.6800 0.8076 0.8987
No log 16.9565 390 0.7484 0.7403 0.7484 0.8651
No log 17.0435 392 0.7094 0.775 0.7094 0.8423
No log 17.1304 394 0.6763 0.7722 0.6763 0.8224
No log 17.2174 396 0.6710 0.7722 0.6710 0.8192
No log 17.3043 398 0.6756 0.7673 0.6756 0.8220
No log 17.3913 400 0.6722 0.7643 0.6722 0.8198
No log 17.4783 402 0.6709 0.7692 0.6709 0.8191
No log 17.5652 404 0.6959 0.7564 0.6959 0.8342
No log 17.6522 406 0.7070 0.7451 0.7070 0.8408
No log 17.7391 408 0.6759 0.7662 0.6759 0.8221
No log 17.8261 410 0.6199 0.7898 0.6199 0.7873
No log 17.9130 412 0.6317 0.7733 0.6317 0.7948
No log 18.0 414 0.6409 0.7898 0.6409 0.8006
No log 18.0870 416 0.6707 0.7898 0.6707 0.8190
No log 18.1739 418 0.6953 0.7848 0.6953 0.8338
No log 18.2609 420 0.7149 0.7421 0.7149 0.8455
No log 18.3478 422 0.7506 0.7421 0.7506 0.8664
No log 18.4348 424 0.7629 0.75 0.7629 0.8735
No log 18.5217 426 0.7390 0.7421 0.7390 0.8596
No log 18.6087 428 0.7188 0.7421 0.7188 0.8478
No log 18.6957 430 0.7047 0.7826 0.7047 0.8395
No log 18.7826 432 0.7186 0.7778 0.7186 0.8477
No log 18.8696 434 0.7587 0.7561 0.7587 0.8711
No log 18.9565 436 0.8188 0.7602 0.8188 0.9049
No log 19.0435 438 0.8237 0.7528 0.8237 0.9076
No log 19.1304 440 0.7605 0.7746 0.7605 0.8721
No log 19.2174 442 0.6845 0.7975 0.6845 0.8274
No log 19.3043 444 0.6547 0.8049 0.6547 0.8092
No log 19.3913 446 0.6561 0.7643 0.6561 0.8100
No log 19.4783 448 0.6711 0.7368 0.6711 0.8192
No log 19.5652 450 0.6655 0.7368 0.6655 0.8158
No log 19.6522 452 0.6445 0.7643 0.6445 0.8028
No log 19.7391 454 0.6408 0.7848 0.6408 0.8005
No log 19.8261 456 0.6565 0.7692 0.6565 0.8103
No log 19.9130 458 0.6452 0.7613 0.6452 0.8033
No log 20.0 460 0.6286 0.7848 0.6286 0.7928
No log 20.0870 462 0.6317 0.7550 0.6317 0.7948
No log 20.1739 464 0.6758 0.7234 0.6758 0.8221
No log 20.2609 466 0.7183 0.6861 0.7183 0.8475
No log 20.3478 468 0.7121 0.7050 0.7121 0.8438
No log 20.4348 470 0.6941 0.75 0.6941 0.8331
No log 20.5217 472 0.6699 0.7361 0.6699 0.8185
No log 20.6087 474 0.6618 0.7361 0.6618 0.8135
No log 20.6957 476 0.6719 0.7347 0.6719 0.8197
No log 20.7826 478 0.6660 0.7532 0.6660 0.8161
No log 20.8696 480 0.6606 0.7532 0.6606 0.8127
No log 20.9565 482 0.6814 0.7320 0.6814 0.8255
No log 21.0435 484 0.7075 0.7368 0.7075 0.8411
No log 21.1304 486 0.7117 0.7248 0.7117 0.8436
No log 21.2174 488 0.6958 0.7417 0.6958 0.8342
No log 21.3043 490 0.6913 0.7532 0.6913 0.8314
No log 21.3913 492 0.6969 0.7532 0.6969 0.8348
No log 21.4783 494 0.6956 0.7532 0.6956 0.8340
No log 21.5652 496 0.7039 0.7826 0.7039 0.8390
No log 21.6522 498 0.7193 0.7547 0.7193 0.8481
0.3108 21.7391 500 0.7137 0.7451 0.7137 0.8448
0.3108 21.8261 502 0.7181 0.7451 0.7181 0.8474
0.3108 21.9130 504 0.7057 0.7114 0.7057 0.8401
0.3108 22.0 506 0.6946 0.7211 0.6946 0.8334
0.3108 22.0870 508 0.6925 0.7451 0.6925 0.8322
0.3108 22.1739 510 0.7164 0.7643 0.7164 0.8464
0.3108 22.2609 512 0.7546 0.7848 0.7546 0.8687
0.3108 22.3478 514 0.7537 0.7848 0.7537 0.8682
0.3108 22.4348 516 0.7161 0.7848 0.7161 0.8462
0.3108 22.5217 518 0.6743 0.7925 0.6743 0.8211
0.3108 22.6087 520 0.6537 0.7643 0.6537 0.8085
0.3108 22.6957 522 0.6482 0.7925 0.6482 0.8051
0.3108 22.7826 524 0.6618 0.7925 0.6618 0.8135
0.3108 22.8696 526 0.7053 0.7848 0.7053 0.8398
0.3108 22.9565 528 0.7269 0.7799 0.7269 0.8526
0.3108 23.0435 530 0.6993 0.7848 0.6993 0.8362
0.3108 23.1304 532 0.6825 0.7643 0.6825 0.8261
0.3108 23.2174 534 0.6947 0.7417 0.6947 0.8335
0.3108 23.3043 536 0.6951 0.7083 0.6951 0.8338
0.3108 23.3913 538 0.6978 0.7034 0.6978 0.8354
0.3108 23.4783 540 0.7121 0.6761 0.7121 0.8438
0.3108 23.5652 542 0.7230 0.7027 0.7230 0.8503

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k5_task1_organization

Finetuned
(4023)
this model