ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7041
  • Qwk: 0.5089
  • Mse: 0.7041
  • Rmse: 0.8391

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 4.0618 -0.0033 4.0618 2.0154
No log 0.0952 4 2.4065 -0.0512 2.4065 1.5513
No log 0.1429 6 2.0256 -0.0252 2.0256 1.4232
No log 0.1905 8 2.2553 0.0970 2.2553 1.5018
No log 0.2381 10 1.4107 0.0030 1.4107 1.1877
No log 0.2857 12 1.2059 -0.0032 1.2059 1.0981
No log 0.3333 14 1.1957 0.0612 1.1957 1.0935
No log 0.3810 16 1.2872 -0.0245 1.2872 1.1345
No log 0.4286 18 1.5933 0.0513 1.5933 1.2623
No log 0.4762 20 1.4603 -0.0175 1.4603 1.2084
No log 0.5238 22 1.2300 0.0256 1.2300 1.1090
No log 0.5714 24 1.0515 0.2466 1.0515 1.0254
No log 0.6190 26 1.0128 0.1837 1.0128 1.0064
No log 0.6667 28 1.0751 0.1997 1.0751 1.0369
No log 0.7143 30 1.2843 0.0827 1.2843 1.1333
No log 0.7619 32 1.4145 0.1245 1.4145 1.1893
No log 0.8095 34 1.7937 0.1845 1.7937 1.3393
No log 0.8571 36 1.9256 0.1941 1.9256 1.3877
No log 0.9048 38 1.6911 0.1109 1.6911 1.3004
No log 0.9524 40 1.3902 0.0399 1.3902 1.1791
No log 1.0 42 1.1196 0.1738 1.1196 1.0581
No log 1.0476 44 0.9718 0.3094 0.9718 0.9858
No log 1.0952 46 0.9691 0.3175 0.9691 0.9844
No log 1.1429 48 0.9700 0.3221 0.9700 0.9849
No log 1.1905 50 0.9893 0.2588 0.9893 0.9946
No log 1.2381 52 0.9962 0.2074 0.9962 0.9981
No log 1.2857 54 0.9165 0.3856 0.9165 0.9573
No log 1.3333 56 0.9210 0.2919 0.9210 0.9597
No log 1.3810 58 0.9533 0.2771 0.9533 0.9763
No log 1.4286 60 0.9179 0.2698 0.9179 0.9581
No log 1.4762 62 0.8512 0.4381 0.8512 0.9226
No log 1.5238 64 0.8897 0.3557 0.8897 0.9432
No log 1.5714 66 1.0532 0.2843 1.0532 1.0262
No log 1.6190 68 0.9486 0.3381 0.9486 0.9740
No log 1.6667 70 0.8241 0.3817 0.8241 0.9078
No log 1.7143 72 0.8441 0.3631 0.8441 0.9187
No log 1.7619 74 0.8140 0.4097 0.8140 0.9022
No log 1.8095 76 0.7946 0.5146 0.7946 0.8914
No log 1.8571 78 0.7942 0.4691 0.7942 0.8912
No log 1.9048 80 0.8293 0.5204 0.8293 0.9107
No log 1.9524 82 0.7957 0.5315 0.7957 0.8920
No log 2.0 84 0.7843 0.4645 0.7843 0.8856
No log 2.0476 86 0.8665 0.4943 0.8665 0.9309
No log 2.0952 88 0.8249 0.4738 0.8249 0.9083
No log 2.1429 90 0.8564 0.4592 0.8564 0.9254
No log 2.1905 92 0.8094 0.5938 0.8094 0.8997
No log 2.2381 94 0.7905 0.5912 0.7905 0.8891
No log 2.2857 96 0.7215 0.5089 0.7215 0.8494
No log 2.3333 98 0.7045 0.5713 0.7045 0.8394
No log 2.3810 100 0.6987 0.5843 0.6987 0.8359
No log 2.4286 102 0.6950 0.4925 0.6950 0.8336
No log 2.4762 104 0.6733 0.5843 0.6733 0.8206
No log 2.5238 106 0.6906 0.4898 0.6906 0.8310
No log 2.5714 108 0.8552 0.4465 0.8552 0.9248
No log 2.6190 110 0.9368 0.3506 0.9368 0.9679
No log 2.6667 112 0.8465 0.5147 0.8465 0.9201
No log 2.7143 114 0.7560 0.4503 0.7560 0.8695
No log 2.7619 116 0.7983 0.4057 0.7983 0.8935
No log 2.8095 118 0.8306 0.3990 0.8306 0.9114
No log 2.8571 120 0.8025 0.4505 0.8025 0.8958
No log 2.9048 122 0.7465 0.4261 0.7465 0.8640
No log 2.9524 124 0.8609 0.5253 0.8609 0.9279
No log 3.0 126 0.9015 0.4454 0.9015 0.9495
No log 3.0476 128 0.7561 0.5610 0.7561 0.8696
No log 3.0952 130 0.7011 0.5905 0.7011 0.8373
No log 3.1429 132 0.6789 0.5562 0.6789 0.8240
No log 3.1905 134 0.6849 0.5939 0.6849 0.8276
No log 3.2381 136 0.7669 0.5948 0.7669 0.8757
No log 3.2857 138 0.8819 0.4574 0.8819 0.9391
No log 3.3333 140 0.7791 0.5140 0.7791 0.8827
No log 3.3810 142 0.6886 0.5783 0.6886 0.8298
No log 3.4286 144 0.7081 0.5681 0.7081 0.8415
No log 3.4762 146 0.7113 0.5160 0.7113 0.8434
No log 3.5238 148 0.7512 0.5127 0.7512 0.8667
No log 3.5714 150 0.8076 0.5068 0.8076 0.8986
No log 3.6190 152 0.8601 0.4934 0.8601 0.9274
No log 3.6667 154 0.8192 0.3485 0.8192 0.9051
No log 3.7143 156 0.7936 0.4277 0.7936 0.8909
No log 3.7619 158 0.7978 0.5084 0.7978 0.8932
No log 3.8095 160 0.7625 0.4707 0.7625 0.8732
No log 3.8571 162 0.7570 0.5149 0.7570 0.8700
No log 3.9048 164 0.8290 0.4947 0.8290 0.9105
No log 3.9524 166 0.9452 0.5196 0.9452 0.9722
No log 4.0 168 1.0264 0.5705 1.0264 1.0131
No log 4.0476 170 0.9560 0.5196 0.9560 0.9777
No log 4.0952 172 0.8064 0.4983 0.8064 0.8980
No log 4.1429 174 0.7854 0.4594 0.7854 0.8862
No log 4.1905 176 0.8127 0.4461 0.8127 0.9015
No log 4.2381 178 0.8188 0.4304 0.8188 0.9049
No log 4.2857 180 0.8352 0.3576 0.8352 0.9139
No log 4.3333 182 0.8624 0.3067 0.8624 0.9287
No log 4.3810 184 0.8398 0.3697 0.8398 0.9164
No log 4.4286 186 0.7945 0.4313 0.7945 0.8914
No log 4.4762 188 0.8079 0.4874 0.8079 0.8989
No log 4.5238 190 0.8101 0.5442 0.8101 0.9001
No log 4.5714 192 0.7860 0.5442 0.7860 0.8866
No log 4.6190 194 0.7416 0.5641 0.7416 0.8611
No log 4.6667 196 0.7164 0.5057 0.7164 0.8464
No log 4.7143 198 0.7711 0.5894 0.7711 0.8781
No log 4.7619 200 0.8519 0.5527 0.8519 0.9230
No log 4.8095 202 0.7877 0.4589 0.7877 0.8875
No log 4.8571 204 0.7520 0.4186 0.7520 0.8672
No log 4.9048 206 0.8443 0.4893 0.8443 0.9189
No log 4.9524 208 0.8519 0.4754 0.8519 0.9230
No log 5.0 210 0.7796 0.4708 0.7796 0.8829
No log 5.0476 212 0.7375 0.4524 0.7375 0.8588
No log 5.0952 214 0.8048 0.5204 0.8048 0.8971
No log 5.1429 216 0.8683 0.4889 0.8683 0.9318
No log 5.1905 218 0.8226 0.5387 0.8226 0.9070
No log 5.2381 220 0.7723 0.4727 0.7723 0.8788
No log 5.2857 222 0.7802 0.4466 0.7802 0.8833
No log 5.3333 224 0.7677 0.5333 0.7677 0.8762
No log 5.3810 226 0.7601 0.5872 0.7601 0.8718
No log 5.4286 228 0.7631 0.5799 0.7631 0.8736
No log 5.4762 230 0.7668 0.5677 0.7668 0.8757
No log 5.5238 232 0.7254 0.5345 0.7254 0.8517
No log 5.5714 234 0.7259 0.4984 0.7259 0.8520
No log 5.6190 236 0.7612 0.5011 0.7612 0.8725
No log 5.6667 238 0.7140 0.5068 0.7140 0.8450
No log 5.7143 240 0.8098 0.5658 0.8098 0.8999
No log 5.7619 242 0.9107 0.5543 0.9107 0.9543
No log 5.8095 244 0.8705 0.5137 0.8705 0.9330
No log 5.8571 246 0.7914 0.4128 0.7914 0.8896
No log 5.9048 248 0.7771 0.4576 0.7771 0.8815
No log 5.9524 250 0.7790 0.4576 0.7790 0.8826
No log 6.0 252 0.7876 0.3820 0.7876 0.8875
No log 6.0476 254 0.8491 0.3657 0.8491 0.9215
No log 6.0952 256 0.9382 0.4444 0.9382 0.9686
No log 6.1429 258 0.9849 0.4454 0.9849 0.9924
No log 6.1905 260 0.9445 0.3913 0.9445 0.9718
No log 6.2381 262 0.8631 0.4406 0.8631 0.9291
No log 6.2857 264 0.8638 0.4620 0.8638 0.9294
No log 6.3333 266 0.8968 0.4921 0.8968 0.9470
No log 6.3810 268 0.9703 0.4444 0.9703 0.9851
No log 6.4286 270 1.0656 0.4668 1.0656 1.0323
No log 6.4762 272 1.1437 0.4478 1.1437 1.0694
No log 6.5238 274 1.0764 0.4668 1.0764 1.0375
No log 6.5714 276 0.8906 0.5006 0.8906 0.9437
No log 6.6190 278 0.7746 0.4763 0.7746 0.8801
No log 6.6667 280 0.7457 0.4721 0.7457 0.8635
No log 6.7143 282 0.7572 0.5446 0.7572 0.8702
No log 6.7619 284 0.7398 0.4596 0.7398 0.8601
No log 6.8095 286 0.7563 0.4529 0.7563 0.8696
No log 6.8571 288 0.7691 0.5175 0.7691 0.8770
No log 6.9048 290 0.7964 0.4503 0.7964 0.8924
No log 6.9524 292 0.7551 0.4499 0.7551 0.8689
No log 7.0 294 0.7070 0.4995 0.7070 0.8408
No log 7.0476 296 0.6945 0.5236 0.6945 0.8334
No log 7.0952 298 0.7401 0.5165 0.7401 0.8603
No log 7.1429 300 0.7679 0.4924 0.7679 0.8763
No log 7.1905 302 0.7212 0.4097 0.7212 0.8492
No log 7.2381 304 0.7273 0.4626 0.7273 0.8528
No log 7.2857 306 0.7666 0.3922 0.7666 0.8755
No log 7.3333 308 0.7869 0.4301 0.7869 0.8871
No log 7.3810 310 0.7516 0.5065 0.7516 0.8669
No log 7.4286 312 0.7383 0.4946 0.7383 0.8592
No log 7.4762 314 0.7824 0.5318 0.7824 0.8845
No log 7.5238 316 0.7493 0.5070 0.7493 0.8656
No log 7.5714 318 0.7131 0.4105 0.7131 0.8445
No log 7.6190 320 0.7339 0.4537 0.7339 0.8567
No log 7.6667 322 0.7573 0.4537 0.7573 0.8702
No log 7.7143 324 0.7748 0.3860 0.7748 0.8802
No log 7.7619 326 0.7805 0.3879 0.7805 0.8835
No log 7.8095 328 0.7607 0.4953 0.7607 0.8722
No log 7.8571 330 0.8354 0.5256 0.8354 0.9140
No log 7.9048 332 0.7985 0.5686 0.7985 0.8936
No log 7.9524 334 0.7315 0.5194 0.7315 0.8552
No log 8.0 336 0.6999 0.5959 0.6999 0.8366
No log 8.0476 338 0.7309 0.6254 0.7309 0.8549
No log 8.0952 340 0.7493 0.5660 0.7493 0.8656
No log 8.1429 342 0.6910 0.5969 0.6910 0.8313
No log 8.1905 344 0.7226 0.4879 0.7226 0.8501
No log 8.2381 346 0.7981 0.4557 0.7981 0.8934
No log 8.2857 348 0.7999 0.4290 0.7999 0.8944
No log 8.3333 350 0.7372 0.4829 0.7372 0.8586
No log 8.3810 352 0.7468 0.4198 0.7468 0.8642
No log 8.4286 354 0.8668 0.4885 0.8668 0.9310
No log 8.4762 356 0.9581 0.5184 0.9581 0.9788
No log 8.5238 358 0.8867 0.5295 0.8867 0.9416
No log 8.5714 360 0.7390 0.5352 0.7390 0.8596
No log 8.6190 362 0.6865 0.5797 0.6865 0.8286
No log 8.6667 364 0.6861 0.5712 0.6861 0.8283
No log 8.7143 366 0.7067 0.5235 0.7067 0.8406
No log 8.7619 368 0.7252 0.5678 0.7252 0.8516
No log 8.8095 370 0.7710 0.4969 0.7710 0.8780
No log 8.8571 372 0.8542 0.5255 0.8542 0.9242
No log 8.9048 374 0.8619 0.4929 0.8619 0.9284
No log 8.9524 376 0.8353 0.5292 0.8353 0.9140
No log 9.0 378 0.8761 0.4916 0.8761 0.9360
No log 9.0476 380 0.9426 0.4560 0.9426 0.9709
No log 9.0952 382 0.9797 0.4560 0.9797 0.9898
No log 9.1429 384 0.9086 0.4801 0.9086 0.9532
No log 9.1905 386 0.8071 0.5425 0.8071 0.8984
No log 9.2381 388 0.7703 0.5719 0.7703 0.8776
No log 9.2857 390 0.7851 0.5571 0.7851 0.8861
No log 9.3333 392 0.8302 0.5924 0.8302 0.9112
No log 9.3810 394 0.7883 0.5756 0.7883 0.8879
No log 9.4286 396 0.7193 0.5594 0.7193 0.8481
No log 9.4762 398 0.7181 0.4482 0.7181 0.8474
No log 9.5238 400 0.7248 0.4113 0.7248 0.8513
No log 9.5714 402 0.7278 0.4563 0.7278 0.8531
No log 9.6190 404 0.7717 0.4537 0.7717 0.8785
No log 9.6667 406 0.8902 0.5332 0.8902 0.9435
No log 9.7143 408 0.9100 0.5307 0.9100 0.9539
No log 9.7619 410 0.8310 0.5019 0.8310 0.9116
No log 9.8095 412 0.7732 0.5069 0.7732 0.8793
No log 9.8571 414 0.7601 0.4863 0.7601 0.8719
No log 9.9048 416 0.7470 0.4661 0.7470 0.8643
No log 9.9524 418 0.7562 0.4534 0.7562 0.8696
No log 10.0 420 0.8021 0.4106 0.8021 0.8956
No log 10.0476 422 0.8700 0.4111 0.8700 0.9327
No log 10.0952 424 0.8651 0.3844 0.8651 0.9301
No log 10.1429 426 0.8229 0.3780 0.8229 0.9071
No log 10.1905 428 0.7863 0.4234 0.7863 0.8867
No log 10.2381 430 0.7499 0.4642 0.7499 0.8660
No log 10.2857 432 0.7263 0.4537 0.7263 0.8523
No log 10.3333 434 0.7158 0.4794 0.7158 0.8460
No log 10.3810 436 0.7530 0.4491 0.7530 0.8678
No log 10.4286 438 0.8449 0.5332 0.8449 0.9192
No log 10.4762 440 0.8325 0.5124 0.8325 0.9124
No log 10.5238 442 0.7521 0.5072 0.7521 0.8673
No log 10.5714 444 0.7054 0.4524 0.7054 0.8399
No log 10.6190 446 0.7090 0.4557 0.7090 0.8420
No log 10.6667 448 0.7184 0.4590 0.7184 0.8476
No log 10.7143 450 0.7173 0.4269 0.7173 0.8470
No log 10.7619 452 0.7504 0.5577 0.7504 0.8663
No log 10.8095 454 0.7623 0.5291 0.7623 0.8731
No log 10.8571 456 0.7354 0.5730 0.7354 0.8576
No log 10.9048 458 0.7064 0.4792 0.7064 0.8405
No log 10.9524 460 0.7036 0.4923 0.7036 0.8388
No log 11.0 462 0.7132 0.5996 0.7132 0.8445
No log 11.0476 464 0.7439 0.5466 0.7439 0.8625
No log 11.0952 466 0.7532 0.5451 0.7532 0.8678
No log 11.1429 468 0.7623 0.5735 0.7623 0.8731
No log 11.1905 470 0.7622 0.6080 0.7622 0.8730
No log 11.2381 472 0.7188 0.6142 0.7188 0.8478
No log 11.2857 474 0.7121 0.6275 0.7121 0.8438
No log 11.3333 476 0.7179 0.6169 0.7179 0.8473
No log 11.3810 478 0.7470 0.6132 0.7470 0.8643
No log 11.4286 480 0.7832 0.5516 0.7832 0.8850
No log 11.4762 482 0.7568 0.5864 0.7568 0.8699
No log 11.5238 484 0.7060 0.5877 0.7060 0.8403
No log 11.5714 486 0.6950 0.4466 0.6950 0.8337
No log 11.6190 488 0.6880 0.4354 0.6880 0.8295
No log 11.6667 490 0.6763 0.4706 0.6763 0.8224
No log 11.7143 492 0.6802 0.5475 0.6802 0.8247
No log 11.7619 494 0.6943 0.5833 0.6943 0.8333
No log 11.8095 496 0.7293 0.5343 0.7293 0.8540
No log 11.8571 498 0.7429 0.5317 0.7429 0.8619
0.2915 11.9048 500 0.7448 0.5317 0.7448 0.8630
0.2915 11.9524 502 0.7719 0.5266 0.7719 0.8786
0.2915 12.0 504 0.7475 0.5397 0.7475 0.8646
0.2915 12.0476 506 0.7358 0.5291 0.7358 0.8578
0.2915 12.0952 508 0.7598 0.5163 0.7598 0.8717
0.2915 12.1429 510 0.8137 0.5330 0.8137 0.9020
0.2915 12.1905 512 0.8641 0.5318 0.8641 0.9296
0.2915 12.2381 514 0.8124 0.5532 0.8124 0.9013
0.2915 12.2857 516 0.7330 0.5342 0.7330 0.8561
0.2915 12.3333 518 0.7091 0.5953 0.7091 0.8421
0.2915 12.3810 520 0.6794 0.6035 0.6794 0.8243
0.2915 12.4286 522 0.6916 0.5785 0.6916 0.8316
0.2915 12.4762 524 0.6887 0.6035 0.6887 0.8299
0.2915 12.5238 526 0.6970 0.5314 0.6970 0.8348
0.2915 12.5714 528 0.7042 0.5314 0.7042 0.8392
0.2915 12.6190 530 0.7079 0.5425 0.7079 0.8414
0.2915 12.6667 532 0.7203 0.5859 0.7203 0.8487
0.2915 12.7143 534 0.7720 0.4875 0.7720 0.8786
0.2915 12.7619 536 0.8221 0.5279 0.8221 0.9067
0.2915 12.8095 538 0.8048 0.4982 0.8048 0.8971
0.2915 12.8571 540 0.7318 0.4660 0.7318 0.8555
0.2915 12.9048 542 0.7041 0.5089 0.7041 0.8391

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task5_organization

Finetuned
(4019)
this model