ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6862
  • Qwk: 0.5472
  • Mse: 0.6862
  • Rmse: 0.8284

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0435 2 3.8865 -0.0294 3.8865 1.9714
No log 0.0870 4 2.0120 0.0142 2.0120 1.4184
No log 0.1304 6 1.7353 -0.0458 1.7353 1.3173
No log 0.1739 8 1.2631 0.0380 1.2631 1.1239
No log 0.2174 10 1.1664 -0.0032 1.1664 1.0800
No log 0.2609 12 1.1818 0.0380 1.1818 1.0871
No log 0.3043 14 1.2235 0.0380 1.2235 1.1061
No log 0.3478 16 1.2637 0.0380 1.2637 1.1242
No log 0.3913 18 1.2607 0.0760 1.2607 1.1228
No log 0.4348 20 1.2646 0.1142 1.2646 1.1246
No log 0.4783 22 1.1968 0.1910 1.1968 1.0940
No log 0.5217 24 1.1106 0.1805 1.1106 1.0539
No log 0.5652 26 1.1018 0.1493 1.1018 1.0497
No log 0.6087 28 1.1995 0.0999 1.1995 1.0952
No log 0.6522 30 1.1395 0.1832 1.1395 1.0675
No log 0.6957 32 1.1605 0.0436 1.1605 1.0772
No log 0.7391 34 1.2550 0.0883 1.2550 1.1203
No log 0.7826 36 1.3800 0.0639 1.3800 1.1747
No log 0.8261 38 1.2540 0.0998 1.2540 1.1198
No log 0.8696 40 1.1479 0.1389 1.1479 1.0714
No log 0.9130 42 1.2217 0.1028 1.2217 1.1053
No log 0.9565 44 1.5665 0.0389 1.5665 1.2516
No log 1.0 46 1.6655 0.0516 1.6655 1.2905
No log 1.0435 48 1.4103 0.0598 1.4103 1.1876
No log 1.0870 50 1.0909 0.1821 1.0909 1.0444
No log 1.1304 52 0.9701 0.3117 0.9701 0.9849
No log 1.1739 54 0.9537 0.3414 0.9537 0.9766
No log 1.2174 56 0.9618 0.3562 0.9618 0.9807
No log 1.2609 58 0.9734 0.3557 0.9734 0.9866
No log 1.3043 60 0.9510 0.3562 0.9510 0.9752
No log 1.3478 62 0.9381 0.3733 0.9381 0.9686
No log 1.3913 64 0.9339 0.3014 0.9339 0.9664
No log 1.4348 66 0.9494 0.2935 0.9494 0.9743
No log 1.4783 68 0.9506 0.2391 0.9506 0.9750
No log 1.5217 70 0.9589 0.2114 0.9589 0.9792
No log 1.5652 72 0.9002 0.3014 0.9002 0.9488
No log 1.6087 74 0.8838 0.3414 0.8838 0.9401
No log 1.6522 76 0.8590 0.3817 0.8590 0.9268
No log 1.6957 78 0.8336 0.3519 0.8336 0.9130
No log 1.7391 80 0.8431 0.4 0.8431 0.9182
No log 1.7826 82 0.8430 0.3981 0.8430 0.9181
No log 1.8261 84 0.8405 0.3537 0.8405 0.9168
No log 1.8696 86 0.8561 0.5155 0.8561 0.9252
No log 1.9130 88 0.8821 0.5174 0.8821 0.9392
No log 1.9565 90 0.8903 0.4459 0.8903 0.9435
No log 2.0 92 1.0073 0.3972 1.0073 1.0036
No log 2.0435 94 1.0262 0.3734 1.0262 1.0130
No log 2.0870 96 0.8925 0.5107 0.8925 0.9447
No log 2.1304 98 0.8793 0.4979 0.8793 0.9377
No log 2.1739 100 1.0192 0.4416 1.0192 1.0096
No log 2.2174 102 0.9693 0.4318 0.9693 0.9845
No log 2.2609 104 0.8455 0.5363 0.8455 0.9195
No log 2.3043 106 0.8263 0.5450 0.8263 0.9090
No log 2.3478 108 0.7896 0.4813 0.7896 0.8886
No log 2.3913 110 0.7701 0.4733 0.7701 0.8776
No log 2.4348 112 0.7638 0.4984 0.7638 0.8740
No log 2.4783 114 0.7713 0.5128 0.7713 0.8782
No log 2.5217 116 0.7180 0.5635 0.7180 0.8474
No log 2.5652 118 0.7116 0.4787 0.7116 0.8436
No log 2.6087 120 0.7141 0.5044 0.7141 0.8450
No log 2.6522 122 0.7626 0.6386 0.7626 0.8733
No log 2.6957 124 0.7743 0.6435 0.7743 0.8800
No log 2.7391 126 0.7699 0.5796 0.7699 0.8774
No log 2.7826 128 0.8402 0.4283 0.8402 0.9166
No log 2.8261 130 1.0326 0.3363 1.0326 1.0161
No log 2.8696 132 1.0287 0.3761 1.0287 1.0143
No log 2.9130 134 0.8748 0.4470 0.8748 0.9353
No log 2.9565 136 0.7884 0.5009 0.7884 0.8879
No log 3.0 138 0.7676 0.5570 0.7676 0.8761
No log 3.0435 140 0.7662 0.5797 0.7662 0.8753
No log 3.0870 142 0.7939 0.5103 0.7939 0.8910
No log 3.1304 144 0.8795 0.4455 0.8795 0.9378
No log 3.1739 146 0.8895 0.4935 0.8895 0.9431
No log 3.2174 148 0.8184 0.5528 0.8184 0.9046
No log 3.2609 150 0.8132 0.4819 0.8132 0.9018
No log 3.3043 152 0.8196 0.4615 0.8196 0.9053
No log 3.3478 154 0.8178 0.4419 0.8178 0.9043
No log 3.3913 156 0.8523 0.4570 0.8523 0.9232
No log 3.4348 158 0.8042 0.4158 0.8042 0.8968
No log 3.4783 160 0.8430 0.3160 0.8430 0.9182
No log 3.5217 162 0.9556 0.3401 0.9556 0.9776
No log 3.5652 164 0.8477 0.3160 0.8477 0.9207
No log 3.6087 166 0.7804 0.5345 0.7804 0.8834
No log 3.6522 168 0.8543 0.4806 0.8543 0.9243
No log 3.6957 170 0.9146 0.2865 0.9146 0.9563
No log 3.7391 172 0.9504 0.2291 0.9504 0.9749
No log 3.7826 174 0.9597 0.2591 0.9597 0.9796
No log 3.8261 176 0.9069 0.3445 0.9069 0.9523
No log 3.8696 178 0.9335 0.3811 0.9335 0.9662
No log 3.9130 180 0.8494 0.3804 0.8494 0.9216
No log 3.9565 182 0.8333 0.4576 0.8333 0.9129
No log 4.0 184 0.8048 0.4730 0.8048 0.8971
No log 4.0435 186 0.7876 0.4461 0.7876 0.8875
No log 4.0870 188 0.8073 0.3719 0.8073 0.8985
No log 4.1304 190 0.7829 0.4128 0.7829 0.8848
No log 4.1739 192 0.7779 0.3996 0.7779 0.8820
No log 4.2174 194 0.7958 0.3184 0.7958 0.8921
No log 4.2609 196 0.7889 0.3676 0.7889 0.8882
No log 4.3043 198 0.7770 0.5163 0.7770 0.8815
No log 4.3478 200 0.7898 0.5766 0.7898 0.8887
No log 4.3913 202 0.7020 0.5874 0.7020 0.8378
No log 4.4348 204 0.7554 0.5618 0.7554 0.8692
No log 4.4783 206 0.7159 0.5379 0.7159 0.8461
No log 4.5217 208 0.7211 0.5330 0.7211 0.8492
No log 4.5652 210 0.8239 0.4902 0.8239 0.9077
No log 4.6087 212 0.7487 0.5498 0.7487 0.8653
No log 4.6522 214 0.7070 0.4898 0.7070 0.8409
No log 4.6957 216 0.7224 0.5213 0.7224 0.8500
No log 4.7391 218 0.8452 0.5414 0.8452 0.9194
No log 4.7826 220 0.9413 0.4574 0.9413 0.9702
No log 4.8261 222 0.8087 0.5231 0.8087 0.8993
No log 4.8696 224 0.7436 0.6007 0.7436 0.8623
No log 4.9130 226 0.7462 0.5902 0.7462 0.8638
No log 4.9565 228 0.7683 0.5763 0.7683 0.8765
No log 5.0 230 0.8623 0.5020 0.8623 0.9286
No log 5.0435 232 0.8875 0.4681 0.8875 0.9421
No log 5.0870 234 0.8004 0.5234 0.8004 0.8947
No log 5.1304 236 0.7856 0.4691 0.7856 0.8864
No log 5.1739 238 0.7886 0.4918 0.7886 0.8880
No log 5.2174 240 0.7940 0.5117 0.7940 0.8911
No log 5.2609 242 0.8967 0.4560 0.8967 0.9469
No log 5.3043 244 0.9099 0.4987 0.9099 0.9539
No log 5.3478 246 0.7790 0.4410 0.7790 0.8826
No log 5.3913 248 0.7415 0.5248 0.7415 0.8611
No log 5.4348 250 0.7559 0.4565 0.7559 0.8694
No log 5.4783 252 0.7639 0.4261 0.7639 0.8740
No log 5.5217 254 0.7792 0.4494 0.7792 0.8827
No log 5.5652 256 0.7704 0.4251 0.7704 0.8777
No log 5.6087 258 0.7568 0.4269 0.7568 0.8700
No log 5.6522 260 0.7523 0.4313 0.7523 0.8674
No log 5.6957 262 0.7222 0.4787 0.7222 0.8498
No log 5.7391 264 0.7149 0.5463 0.7149 0.8455
No log 5.7826 266 0.7193 0.6160 0.7193 0.8481
No log 5.8261 268 0.7691 0.5439 0.7691 0.8770
No log 5.8696 270 0.7113 0.6617 0.7113 0.8434
No log 5.9130 272 0.6821 0.6476 0.6821 0.8259
No log 5.9565 274 0.6982 0.6528 0.6982 0.8356
No log 6.0 276 0.7793 0.5318 0.7793 0.8828
No log 6.0435 278 0.7484 0.5439 0.7484 0.8651
No log 6.0870 280 0.7456 0.5470 0.7456 0.8635
No log 6.1304 282 0.7213 0.5740 0.7213 0.8493
No log 6.1739 284 0.7244 0.4873 0.7244 0.8511
No log 6.2174 286 0.7118 0.5644 0.7118 0.8437
No log 6.2609 288 0.7316 0.4135 0.7316 0.8553
No log 6.3043 290 0.7192 0.4984 0.7192 0.8481
No log 6.3478 292 0.6941 0.5163 0.6941 0.8331
No log 6.3913 294 0.7154 0.6128 0.7154 0.8458
No log 6.4348 296 0.7172 0.5618 0.7172 0.8469
No log 6.4783 298 0.6939 0.5680 0.6939 0.8330
No log 6.5217 300 0.6980 0.5877 0.6980 0.8354
No log 6.5652 302 0.7417 0.5052 0.7417 0.8612
No log 6.6087 304 0.7484 0.5067 0.7484 0.8651
No log 6.6522 306 0.7230 0.4879 0.7230 0.8503
No log 6.6957 308 0.7159 0.5002 0.7159 0.8461
No log 6.7391 310 0.7043 0.5357 0.7043 0.8392
No log 6.7826 312 0.6960 0.5060 0.6960 0.8343
No log 6.8261 314 0.6992 0.5066 0.6992 0.8362
No log 6.8696 316 0.6768 0.5809 0.6768 0.8227
No log 6.9130 318 0.6808 0.5329 0.6808 0.8251
No log 6.9565 320 0.6908 0.5671 0.6908 0.8312
No log 7.0 322 0.7242 0.5263 0.7242 0.8510
No log 7.0435 324 0.7106 0.5459 0.7106 0.8430
No log 7.0870 326 0.8062 0.4708 0.8062 0.8979
No log 7.1304 328 0.8057 0.4708 0.8057 0.8976
No log 7.1739 330 0.7553 0.5046 0.7553 0.8691
No log 7.2174 332 0.7268 0.5094 0.7268 0.8525
No log 7.2609 334 0.7145 0.5582 0.7145 0.8453
No log 7.3043 336 0.7106 0.5475 0.7106 0.8430
No log 7.3478 338 0.7219 0.5446 0.7219 0.8497
No log 7.3913 340 0.7267 0.5129 0.7267 0.8525
No log 7.4348 342 0.7339 0.4461 0.7339 0.8567
No log 7.4783 344 0.7684 0.4641 0.7684 0.8766
No log 7.5217 346 0.7471 0.4641 0.7471 0.8644
No log 7.5652 348 0.7191 0.5352 0.7191 0.8480
No log 7.6087 350 0.7407 0.5953 0.7407 0.8606
No log 7.6522 352 0.7185 0.5683 0.7185 0.8477
No log 7.6957 354 0.7478 0.5476 0.7478 0.8647
No log 7.7391 356 0.7586 0.5197 0.7586 0.8710
No log 7.7826 358 0.7467 0.4996 0.7467 0.8641
No log 7.8261 360 0.8009 0.4388 0.8009 0.8949
No log 7.8696 362 0.8052 0.3269 0.8052 0.8973
No log 7.9130 364 0.7747 0.4279 0.7747 0.8802
No log 7.9565 366 0.7503 0.4660 0.7503 0.8662
No log 8.0 368 0.7182 0.4918 0.7182 0.8475
No log 8.0435 370 0.6882 0.5120 0.6882 0.8296
No log 8.0870 372 0.6722 0.6187 0.6722 0.8199
No log 8.1304 374 0.7175 0.6081 0.7175 0.8470
No log 8.1739 376 0.7358 0.6071 0.7358 0.8578
No log 8.2174 378 0.7982 0.5398 0.7982 0.8934
No log 8.2609 380 0.7700 0.6218 0.7700 0.8775
No log 8.3043 382 0.6887 0.5463 0.6887 0.8299
No log 8.3478 384 0.8329 0.4508 0.8329 0.9127
No log 8.3913 386 0.9849 0.5184 0.9849 0.9924
No log 8.4348 388 0.9149 0.4854 0.9149 0.9565
No log 8.4783 390 0.7384 0.5012 0.7384 0.8593
No log 8.5217 392 0.7104 0.5473 0.7104 0.8428
No log 8.5652 394 0.8558 0.4894 0.8558 0.9251
No log 8.6087 396 0.8252 0.4894 0.8252 0.9084
No log 8.6522 398 0.7126 0.5186 0.7126 0.8441
No log 8.6957 400 0.6932 0.5432 0.6932 0.8326
No log 8.7391 402 0.7075 0.4968 0.7075 0.8412
No log 8.7826 404 0.7033 0.5432 0.7033 0.8387
No log 8.8261 406 0.7082 0.5536 0.7082 0.8415
No log 8.8696 408 0.7080 0.4760 0.7080 0.8414
No log 8.9130 410 0.6967 0.4760 0.6967 0.8347
No log 8.9565 412 0.6681 0.5822 0.6681 0.8174
No log 9.0 414 0.6506 0.6046 0.6506 0.8066
No log 9.0435 416 0.6532 0.6219 0.6532 0.8082
No log 9.0870 418 0.6514 0.6219 0.6514 0.8071
No log 9.1304 420 0.6505 0.5644 0.6505 0.8066
No log 9.1739 422 0.6544 0.5886 0.6544 0.8090
No log 9.2174 424 0.6797 0.5597 0.6797 0.8245
No log 9.2609 426 0.7490 0.5137 0.7490 0.8655
No log 9.3043 428 0.7794 0.4898 0.7794 0.8828
No log 9.3478 430 0.7201 0.5400 0.7201 0.8486
No log 9.3913 432 0.6806 0.5432 0.6806 0.8250
No log 9.4348 434 0.6772 0.5432 0.6772 0.8229
No log 9.4783 436 0.6891 0.5089 0.6891 0.8301
No log 9.5217 438 0.7677 0.5428 0.7677 0.8762
No log 9.5652 440 0.7609 0.5451 0.7609 0.8723
No log 9.6087 442 0.6901 0.5570 0.6901 0.8307
No log 9.6522 444 0.6739 0.5188 0.6739 0.8209
No log 9.6957 446 0.6775 0.5074 0.6775 0.8231
No log 9.7391 448 0.6750 0.5516 0.6750 0.8216
No log 9.7826 450 0.7319 0.5279 0.7319 0.8555
No log 9.8261 452 0.7593 0.5331 0.7593 0.8714
No log 9.8696 454 0.6942 0.5098 0.6942 0.8332
No log 9.9130 456 0.6676 0.5771 0.6676 0.8171
No log 9.9565 458 0.6685 0.5783 0.6685 0.8176
No log 10.0 460 0.6652 0.6307 0.6652 0.8156
No log 10.0435 462 0.7046 0.5395 0.7046 0.8394
No log 10.0870 464 0.7331 0.5470 0.7331 0.8562
No log 10.1304 466 0.7124 0.4494 0.7124 0.8440
No log 10.1739 468 0.7124 0.4893 0.7124 0.8441
No log 10.2174 470 0.7132 0.4923 0.7132 0.8445
No log 10.2609 472 0.7016 0.5415 0.7016 0.8376
No log 10.3043 474 0.6915 0.5554 0.6915 0.8316
No log 10.3478 476 0.6783 0.6177 0.6783 0.8236
No log 10.3913 478 0.7239 0.5862 0.7239 0.8508
No log 10.4348 480 0.7312 0.5958 0.7312 0.8551
No log 10.4783 482 0.6749 0.6198 0.6749 0.8215
No log 10.5217 484 0.6417 0.6335 0.6417 0.8011
No log 10.5652 486 0.6504 0.5084 0.6504 0.8065
No log 10.6087 488 0.6547 0.4968 0.6547 0.8091
No log 10.6522 490 0.6450 0.5529 0.6450 0.8031
No log 10.6957 492 0.6673 0.6073 0.6673 0.8169
No log 10.7391 494 0.6924 0.5973 0.6924 0.8321
No log 10.7826 496 0.6905 0.6109 0.6905 0.8310
No log 10.8261 498 0.6965 0.5833 0.6965 0.8346
0.2902 10.8696 500 0.7077 0.6209 0.7077 0.8412
0.2902 10.9130 502 0.6775 0.6147 0.6775 0.8231
0.2902 10.9565 504 0.6453 0.6500 0.6453 0.8033
0.2902 11.0 506 0.6393 0.6175 0.6393 0.7996
0.2902 11.0435 508 0.6427 0.6057 0.6427 0.8017
0.2902 11.0870 510 0.6740 0.5932 0.6740 0.8209
0.2902 11.1304 512 0.7058 0.5654 0.7058 0.8401
0.2902 11.1739 514 0.6862 0.5472 0.6862 0.8284

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task5_organization

Finetuned
(4019)
this model