ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k14_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8688
  • Qwk: 0.3025
  • Mse: 0.8688
  • Rmse: 0.9321

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0303 2 4.2727 -0.0033 4.2727 2.0671
No log 0.0606 4 2.5006 0.0462 2.5006 1.5813
No log 0.0909 6 2.1371 0.0034 2.1371 1.4619
No log 0.1212 8 1.6659 -0.0168 1.6659 1.2907
No log 0.1515 10 1.1213 0.2466 1.1213 1.0589
No log 0.1818 12 1.0636 0.1891 1.0636 1.0313
No log 0.2121 14 1.0575 0.3229 1.0575 1.0284
No log 0.2424 16 1.1516 0.2100 1.1516 1.0731
No log 0.2727 18 1.0442 0.3770 1.0442 1.0219
No log 0.3030 20 1.0964 0.2611 1.0964 1.0471
No log 0.3333 22 1.1052 0.2465 1.1052 1.0513
No log 0.3636 24 1.0751 0.1848 1.0751 1.0369
No log 0.3939 26 1.0451 0.1864 1.0451 1.0223
No log 0.4242 28 1.0488 0.1971 1.0488 1.0241
No log 0.4545 30 1.0913 0.1979 1.0913 1.0447
No log 0.4848 32 1.2357 0.0466 1.2357 1.1116
No log 0.5152 34 1.3523 0.0613 1.3523 1.1629
No log 0.5455 36 1.3741 0.0613 1.3741 1.1722
No log 0.5758 38 1.2770 0.0613 1.2770 1.1301
No log 0.6061 40 1.1512 0.0466 1.1512 1.0729
No log 0.6364 42 1.0263 0.2880 1.0263 1.0131
No log 0.6667 44 1.0187 0.2880 1.0187 1.0093
No log 0.6970 46 1.1219 0.1086 1.1219 1.0592
No log 0.7273 48 1.2368 0.0760 1.2368 1.1121
No log 0.7576 50 1.1751 0.1086 1.1751 1.0840
No log 0.7879 52 1.1150 0.1205 1.1150 1.0559
No log 0.8182 54 0.9461 0.3631 0.9461 0.9727
No log 0.8485 56 0.9375 0.3548 0.9375 0.9683
No log 0.8788 58 1.0077 0.2376 1.0077 1.0038
No log 0.9091 60 1.1416 0.1304 1.1416 1.0685
No log 0.9394 62 1.1857 0.2827 1.1857 1.0889
No log 0.9697 64 1.1724 0.2070 1.1724 1.0828
No log 1.0 66 1.0297 0.2713 1.0297 1.0147
No log 1.0303 68 1.1213 0.0650 1.1213 1.0589
No log 1.0606 70 1.2322 0.0862 1.2322 1.1101
No log 1.0909 72 1.1926 0.0862 1.1926 1.0921
No log 1.1212 74 1.0343 0.2424 1.0343 1.0170
No log 1.1515 76 0.9724 0.3446 0.9724 0.9861
No log 1.1818 78 0.9449 0.3508 0.9449 0.9721
No log 1.2121 80 0.9759 0.2478 0.9759 0.9879
No log 1.2424 82 0.9357 0.3340 0.9357 0.9673
No log 1.2727 84 0.9208 0.3438 0.9208 0.9596
No log 1.3030 86 0.9248 0.3299 0.9248 0.9617
No log 1.3333 88 0.9031 0.3814 0.9031 0.9503
No log 1.3636 90 0.8873 0.4573 0.8873 0.9420
No log 1.3939 92 0.9922 0.3680 0.9922 0.9961
No log 1.4242 94 1.0563 0.3381 1.0563 1.0278
No log 1.4545 96 0.9827 0.3161 0.9827 0.9913
No log 1.4848 98 0.8168 0.4871 0.8168 0.9037
No log 1.5152 100 0.8376 0.4871 0.8376 0.9152
No log 1.5455 102 0.9741 0.3943 0.9741 0.9870
No log 1.5758 104 0.9403 0.3747 0.9403 0.9697
No log 1.6061 106 0.8315 0.4782 0.8315 0.9119
No log 1.6364 108 0.8966 0.4307 0.8966 0.9469
No log 1.6667 110 0.8241 0.5594 0.8241 0.9078
No log 1.6970 112 0.9443 0.4176 0.9443 0.9717
No log 1.7273 114 1.0530 0.3043 1.0530 1.0261
No log 1.7576 116 1.0252 0.3043 1.0252 1.0125
No log 1.7879 118 0.9762 0.4119 0.9762 0.9881
No log 1.8182 120 1.0917 0.3380 1.0917 1.0448
No log 1.8485 122 1.1154 0.3344 1.1154 1.0561
No log 1.8788 124 1.0366 0.3333 1.0366 1.0181
No log 1.9091 126 0.8865 0.3474 0.8865 0.9415
No log 1.9394 128 0.7981 0.4554 0.7981 0.8934
No log 1.9697 130 0.7968 0.5093 0.7968 0.8926
No log 2.0 132 0.8217 0.4138 0.8217 0.9065
No log 2.0303 134 0.8444 0.3188 0.8444 0.9189
No log 2.0606 136 0.9035 0.4455 0.9035 0.9505
No log 2.0909 138 0.9484 0.4157 0.9484 0.9738
No log 2.1212 140 0.9771 0.3785 0.9771 0.9885
No log 2.1515 142 0.9348 0.3782 0.9348 0.9669
No log 2.1818 144 0.9696 0.3743 0.9696 0.9847
No log 2.2121 146 0.9161 0.4023 0.9161 0.9571
No log 2.2424 148 0.8901 0.3623 0.8901 0.9434
No log 2.2727 150 0.9950 0.3790 0.9950 0.9975
No log 2.3030 152 1.1421 0.2837 1.1421 1.0687
No log 2.3333 154 1.0594 0.3881 1.0594 1.0293
No log 2.3636 156 0.9274 0.4444 0.9274 0.9630
No log 2.3939 158 0.9477 0.4197 0.9477 0.9735
No log 2.4242 160 0.9426 0.3663 0.9426 0.9709
No log 2.4545 162 0.8752 0.3958 0.8752 0.9355
No log 2.4848 164 0.8031 0.5040 0.8031 0.8962
No log 2.5152 166 0.8111 0.5040 0.8111 0.9006
No log 2.5455 168 0.8415 0.4247 0.8415 0.9174
No log 2.5758 170 0.9610 0.3671 0.9610 0.9803
No log 2.6061 172 1.0220 0.3275 1.0220 1.0109
No log 2.6364 174 0.8742 0.3634 0.8742 0.9350
No log 2.6667 176 0.8678 0.4396 0.8678 0.9316
No log 2.6970 178 0.8504 0.4498 0.8504 0.9222
No log 2.7273 180 0.8699 0.4871 0.8699 0.9327
No log 2.7576 182 0.9314 0.3809 0.9314 0.9651
No log 2.7879 184 0.8975 0.3809 0.8975 0.9473
No log 2.8182 186 0.7991 0.6123 0.7991 0.8939
No log 2.8485 188 0.7977 0.5314 0.7977 0.8932
No log 2.8788 190 0.7688 0.6186 0.7688 0.8768
No log 2.9091 192 0.8083 0.4491 0.8083 0.8991
No log 2.9394 194 0.9012 0.3845 0.9012 0.9493
No log 2.9697 196 0.8188 0.4078 0.8188 0.9049
No log 3.0 198 0.8343 0.3089 0.8343 0.9134
No log 3.0303 200 0.9213 0.3483 0.9213 0.9598
No log 3.0606 202 0.8816 0.3067 0.8816 0.9389
No log 3.0909 204 0.8143 0.3774 0.8143 0.9024
No log 3.1212 206 0.7891 0.4524 0.7891 0.8883
No log 3.1515 208 0.7998 0.5455 0.7998 0.8943
No log 3.1818 210 0.9418 0.4943 0.9418 0.9705
No log 3.2121 212 0.9318 0.4513 0.9318 0.9653
No log 3.2424 214 0.8285 0.5977 0.8285 0.9102
No log 3.2727 216 0.7603 0.5732 0.7603 0.8720
No log 3.3030 218 0.7760 0.5818 0.7760 0.8809
No log 3.3333 220 0.8560 0.5267 0.8560 0.9252
No log 3.3636 222 0.8819 0.4938 0.8819 0.9391
No log 3.3939 224 0.8653 0.5175 0.8653 0.9302
No log 3.4242 226 0.8879 0.4920 0.8879 0.9423
No log 3.4545 228 1.0070 0.4583 1.0070 1.0035
No log 3.4848 230 1.0184 0.4583 1.0184 1.0092
No log 3.5152 232 0.9969 0.4583 0.9969 0.9984
No log 3.5455 234 0.8974 0.4708 0.8974 0.9473
No log 3.5758 236 0.9186 0.4371 0.9186 0.9585
No log 3.6061 238 0.9945 0.4202 0.9945 0.9973
No log 3.6364 240 0.9581 0.4341 0.9581 0.9788
No log 3.6667 242 0.8147 0.5522 0.8147 0.9026
No log 3.6970 244 0.7906 0.5540 0.7906 0.8892
No log 3.7273 246 0.7672 0.5833 0.7672 0.8759
No log 3.7576 248 0.7575 0.5614 0.7575 0.8704
No log 3.7879 250 0.8015 0.5304 0.8015 0.8953
No log 3.8182 252 0.9710 0.4064 0.9710 0.9854
No log 3.8485 254 1.1373 0.4059 1.1373 1.0664
No log 3.8788 256 1.0716 0.3631 1.0716 1.0352
No log 3.9091 258 0.8624 0.4335 0.8624 0.9287
No log 3.9394 260 0.7831 0.3616 0.7831 0.8850
No log 3.9697 262 0.7883 0.3544 0.7883 0.8879
No log 4.0 264 0.8388 0.3236 0.8388 0.9159
No log 4.0303 266 0.9502 0.3005 0.9502 0.9748
No log 4.0606 268 0.9459 0.4214 0.9459 0.9726
No log 4.0909 270 0.8913 0.5046 0.8913 0.9441
No log 4.1212 272 0.8503 0.5046 0.8503 0.9221
No log 4.1515 274 0.8281 0.4742 0.8281 0.9100
No log 4.1818 276 0.8443 0.3860 0.8443 0.9189
No log 4.2121 278 0.8524 0.3445 0.8524 0.9233
No log 4.2424 280 0.8558 0.3861 0.8558 0.9251
No log 4.2727 282 0.9427 0.5231 0.9427 0.9709
No log 4.3030 284 0.9294 0.5025 0.9294 0.9641
No log 4.3333 286 0.8148 0.5089 0.8148 0.9026
No log 4.3636 288 0.7776 0.4661 0.7776 0.8818
No log 4.3939 290 0.8060 0.5089 0.8060 0.8978
No log 4.4242 292 0.8529 0.5044 0.8529 0.9235
No log 4.4545 294 0.8809 0.4932 0.8809 0.9386
No log 4.4848 296 0.9021 0.4500 0.9021 0.9498
No log 4.5152 298 0.8129 0.5208 0.8129 0.9016
No log 4.5455 300 0.7826 0.4757 0.7826 0.8847
No log 4.5758 302 0.8373 0.5183 0.8373 0.9150
No log 4.6061 304 0.8623 0.4829 0.8623 0.9286
No log 4.6364 306 0.7857 0.4280 0.7857 0.8864
No log 4.6667 308 0.7721 0.4544 0.7721 0.8787
No log 4.6970 310 0.8355 0.4724 0.8355 0.9140
No log 4.7273 312 0.8993 0.4573 0.8993 0.9483
No log 4.7576 314 0.9688 0.3596 0.9688 0.9843
No log 4.7879 316 0.9045 0.4098 0.9045 0.9511
No log 4.8182 318 0.8339 0.3840 0.8339 0.9132
No log 4.8485 320 0.8224 0.3859 0.8224 0.9069
No log 4.8788 322 0.8150 0.4128 0.8150 0.9028
No log 4.9091 324 0.8737 0.3844 0.8737 0.9347
No log 4.9394 326 1.0147 0.4318 1.0147 1.0073
No log 4.9697 328 0.9441 0.3970 0.9441 0.9716
No log 5.0 330 0.8025 0.3576 0.8025 0.8959
No log 5.0303 332 0.7817 0.4537 0.7817 0.8842
No log 5.0606 334 0.7807 0.4409 0.7807 0.8836
No log 5.0909 336 0.7903 0.4119 0.7903 0.8890
No log 5.1212 338 0.8915 0.3921 0.8915 0.9442
No log 5.1515 340 0.9848 0.3847 0.9848 0.9924
No log 5.1818 342 1.0663 0.3624 1.0663 1.0326
No log 5.2121 344 1.0627 0.3865 1.0627 1.0309
No log 5.2424 346 0.9275 0.4450 0.9275 0.9631
No log 5.2727 348 0.8086 0.5107 0.8086 0.8992
No log 5.3030 350 0.8153 0.4988 0.8153 0.9029
No log 5.3333 352 0.8815 0.3551 0.8815 0.9389
No log 5.3636 354 0.9536 0.3003 0.9536 0.9765
No log 5.3939 356 0.9853 0.3156 0.9853 0.9926
No log 5.4242 358 0.9421 0.2886 0.9421 0.9706
No log 5.4545 360 0.8477 0.3941 0.8477 0.9207
No log 5.4848 362 0.8393 0.3941 0.8393 0.9161
No log 5.5152 364 0.8496 0.4727 0.8496 0.9217
No log 5.5455 366 0.8738 0.4836 0.8738 0.9348
No log 5.5758 368 0.8380 0.4960 0.8380 0.9154
No log 5.6061 370 0.8112 0.4216 0.8112 0.9007
No log 5.6364 372 0.8095 0.4490 0.8095 0.8997
No log 5.6667 374 0.9016 0.4460 0.9016 0.9495
No log 5.6970 376 1.0701 0.3707 1.0701 1.0345
No log 5.7273 378 1.1372 0.3452 1.1372 1.0664
No log 5.7576 380 1.0703 0.2448 1.0703 1.0345
No log 5.7879 382 1.0176 0.2066 1.0176 1.0088
No log 5.8182 384 0.8879 0.3536 0.8879 0.9423
No log 5.8485 386 0.8291 0.4516 0.8291 0.9105
No log 5.8788 388 0.8408 0.4086 0.8408 0.9170
No log 5.9091 390 0.9218 0.4366 0.9218 0.9601
No log 5.9394 392 0.9804 0.4091 0.9804 0.9902
No log 5.9697 394 0.9563 0.4098 0.9563 0.9779
No log 6.0 396 0.9267 0.3510 0.9267 0.9627
No log 6.0303 398 0.9020 0.3414 0.9020 0.9497
No log 6.0606 400 0.8814 0.3528 0.8814 0.9388
No log 6.0909 402 0.9145 0.2643 0.9145 0.9563
No log 6.1212 404 0.9537 0.1727 0.9537 0.9766
No log 6.1515 406 0.9402 0.1727 0.9402 0.9697
No log 6.1818 408 0.8908 0.2618 0.8908 0.9438
No log 6.2121 410 0.8500 0.3498 0.8500 0.9220
No log 6.2424 412 0.8193 0.3795 0.8193 0.9052
No log 6.2727 414 0.8082 0.4381 0.8082 0.8990
No log 6.3030 416 0.8027 0.3876 0.8027 0.8960
No log 6.3333 418 0.7994 0.4537 0.7994 0.8941
No log 6.3636 420 0.8235 0.4223 0.8235 0.9075
No log 6.3939 422 0.8813 0.3424 0.8813 0.9388
No log 6.4242 424 0.8579 0.4223 0.8579 0.9262
No log 6.4545 426 0.8213 0.4267 0.8213 0.9062
No log 6.4848 428 0.8141 0.4267 0.8141 0.9023
No log 6.5152 430 0.8050 0.4794 0.8050 0.8972
No log 6.5455 432 0.8017 0.4810 0.8017 0.8954
No log 6.5758 434 0.8054 0.4826 0.8054 0.8975
No log 6.6061 436 0.8134 0.4537 0.8134 0.9019
No log 6.6364 438 0.8373 0.4363 0.8373 0.9151
No log 6.6667 440 0.8397 0.4363 0.8397 0.9164
No log 6.6970 442 0.8357 0.4363 0.8357 0.9141
No log 6.7273 444 0.8079 0.4794 0.8079 0.8988
No log 6.7576 446 0.7889 0.5419 0.7889 0.8882
No log 6.7879 448 0.7753 0.5656 0.7753 0.8805
No log 6.8182 450 0.7743 0.5040 0.7743 0.8799
No log 6.8485 452 0.8947 0.4573 0.8947 0.9459
No log 6.8788 454 1.0859 0.3683 1.0859 1.0420
No log 6.9091 456 1.0488 0.3641 1.0488 1.0241
No log 6.9394 458 0.8747 0.4586 0.8747 0.9352
No log 6.9697 460 0.7961 0.5131 0.7961 0.8923
No log 7.0 462 0.8117 0.3859 0.8117 0.9010
No log 7.0303 464 0.8948 0.3637 0.8948 0.9459
No log 7.0606 466 0.9198 0.3317 0.9198 0.9591
No log 7.0909 468 0.9170 0.3785 0.9170 0.9576
No log 7.1212 470 0.8846 0.3668 0.8846 0.9405
No log 7.1515 472 0.8585 0.3351 0.8585 0.9266
No log 7.1818 474 0.8503 0.3339 0.8503 0.9221
No log 7.2121 476 0.8183 0.3631 0.8183 0.9046
No log 7.2424 478 0.8204 0.3631 0.8204 0.9058
No log 7.2727 480 0.8027 0.4086 0.8027 0.8959
No log 7.3030 482 0.7763 0.4641 0.7763 0.8811
No log 7.3333 484 0.7670 0.4919 0.7670 0.8758
No log 7.3636 486 0.7754 0.5168 0.7754 0.8805
No log 7.3939 488 0.8285 0.4123 0.8285 0.9102
No log 7.4242 490 0.8900 0.3957 0.8900 0.9434
No log 7.4545 492 0.8717 0.4060 0.8717 0.9336
No log 7.4848 494 0.8514 0.4075 0.8514 0.9227
No log 7.5152 496 0.8194 0.3556 0.8194 0.9052
No log 7.5455 498 0.8033 0.3673 0.8033 0.8963
0.2903 7.5758 500 0.8179 0.3959 0.8179 0.9044
0.2903 7.6061 502 0.8612 0.3957 0.8612 0.9280
0.2903 7.6364 504 0.8868 0.4570 0.8868 0.9417
0.2903 7.6667 506 0.8439 0.3959 0.8439 0.9186
0.2903 7.6970 508 0.8041 0.4110 0.8041 0.8967
0.2903 7.7273 510 0.8249 0.3625 0.8249 0.9083
0.2903 7.7576 512 0.8387 0.3935 0.8387 0.9158
0.2903 7.7879 514 0.7938 0.3709 0.7938 0.8909
0.2903 7.8182 516 0.8248 0.4067 0.8248 0.9082
0.2903 7.8485 518 0.9048 0.4326 0.9048 0.9512
0.2903 7.8788 520 0.8966 0.3921 0.8966 0.9469
0.2903 7.9091 522 0.8825 0.3317 0.8825 0.9394
0.2903 7.9394 524 0.8513 0.3175 0.8513 0.9226
0.2903 7.9697 526 0.8429 0.3476 0.8429 0.9181
0.2903 8.0 528 0.8688 0.3025 0.8688 0.9321

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k14_task5_organization

Finetuned
(4019)
this model