ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8513
  • Qwk: 0.6839
  • Mse: 0.8513
  • Rmse: 0.9227

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0328 2 6.6505 0.0121 6.6505 2.5789
No log 0.0656 4 4.5091 0.0598 4.5091 2.1235
No log 0.0984 6 3.5863 -0.0521 3.5863 1.8938
No log 0.1311 8 2.6713 0.0882 2.6713 1.6344
No log 0.1639 10 2.7091 0.0 2.7091 1.6459
No log 0.1967 12 2.2775 0.0161 2.2775 1.5091
No log 0.2295 14 2.0967 0.0806 2.0967 1.4480
No log 0.2623 16 2.3620 0.1972 2.3620 1.5369
No log 0.2951 18 2.5764 0.2384 2.5764 1.6051
No log 0.3279 20 2.7388 0.2264 2.7388 1.6549
No log 0.3607 22 3.5261 0.0636 3.5261 1.8778
No log 0.3934 24 3.7582 0.1021 3.7582 1.9386
No log 0.4262 26 3.3004 0.1502 3.3004 1.8167
No log 0.4590 28 2.6921 0.2249 2.6921 1.6408
No log 0.4918 30 2.1541 0.25 2.1541 1.4677
No log 0.5246 32 2.4622 0.3169 2.4622 1.5691
No log 0.5574 34 3.4944 0.2202 3.4944 1.8693
No log 0.5902 36 3.0153 0.25 3.0153 1.7365
No log 0.6230 38 2.2163 0.4186 2.2163 1.4887
No log 0.6557 40 1.6896 0.4265 1.6896 1.2999
No log 0.6885 42 1.6295 0.3664 1.6295 1.2765
No log 0.7213 44 1.7049 0.4179 1.7049 1.3057
No log 0.7541 46 2.1135 0.4327 2.1135 1.4538
No log 0.7869 48 2.9832 0.2723 2.9832 1.7272
No log 0.8197 50 2.5573 0.3350 2.5573 1.5991
No log 0.8525 52 1.7652 0.4634 1.7652 1.3286
No log 0.8852 54 1.4973 0.3968 1.4973 1.2236
No log 0.9180 56 1.5573 0.3559 1.5573 1.2479
No log 0.9508 58 1.4894 0.3729 1.4894 1.2204
No log 0.9836 60 1.6401 0.4148 1.6401 1.2807
No log 1.0164 62 2.1981 0.3605 2.1981 1.4826
No log 1.0492 64 2.5396 0.3168 2.5396 1.5936
No log 1.0820 66 2.1340 0.4066 2.1340 1.4608
No log 1.1148 68 1.7269 0.4371 1.7269 1.3141
No log 1.1475 70 1.5646 0.3969 1.5646 1.2508
No log 1.1803 72 1.4242 0.4429 1.4242 1.1934
No log 1.2131 74 1.4469 0.4634 1.4469 1.2029
No log 1.2459 76 2.0447 0.4569 2.0447 1.4299
No log 1.2787 78 3.4649 0.2620 3.4649 1.8614
No log 1.3115 80 3.2112 0.2703 3.2112 1.7920
No log 1.3443 82 1.8903 0.4530 1.8903 1.3749
No log 1.3770 84 1.2470 0.4444 1.2470 1.1167
No log 1.4098 86 1.2285 0.3932 1.2285 1.1084
No log 1.4426 88 1.3894 0.2342 1.3894 1.1787
No log 1.4754 90 1.3385 0.2679 1.3385 1.1569
No log 1.5082 92 1.2942 0.4298 1.2942 1.1376
No log 1.5410 94 1.6667 0.4780 1.6667 1.2910
No log 1.5738 96 1.9410 0.4358 1.9410 1.3932
No log 1.6066 98 1.9482 0.4375 1.9482 1.3958
No log 1.6393 100 1.6526 0.5057 1.6526 1.2855
No log 1.6721 102 1.4633 0.5294 1.4633 1.2097
No log 1.7049 104 1.4504 0.5697 1.4504 1.2043
No log 1.7377 106 1.5232 0.5301 1.5232 1.2342
No log 1.7705 108 1.6203 0.4790 1.6203 1.2729
No log 1.8033 110 1.9536 0.4138 1.9536 1.3977
No log 1.8361 112 2.5063 0.4141 2.5063 1.5831
No log 1.8689 114 2.8120 0.3158 2.8120 1.6769
No log 1.9016 116 3.1706 0.2586 3.1706 1.7806
No log 1.9344 118 2.9763 0.2710 2.9763 1.7252
No log 1.9672 120 2.5152 0.3353 2.5152 1.5859
No log 2.0 122 2.4455 0.3095 2.4455 1.5638
No log 2.0328 124 2.4419 0.3095 2.4419 1.5627
No log 2.0656 126 2.2137 0.3145 2.2137 1.4879
No log 2.0984 128 1.6165 0.3624 1.6165 1.2714
No log 2.1311 130 1.3632 0.4627 1.3632 1.1676
No log 2.1639 132 1.3092 0.4818 1.3092 1.1442
No log 2.1967 134 1.3562 0.4490 1.3562 1.1645
No log 2.2295 136 1.7301 0.4815 1.7301 1.3153
No log 2.2623 138 2.1815 0.3864 2.1815 1.4770
No log 2.2951 140 2.1066 0.3931 2.1066 1.4514
No log 2.3279 142 1.8686 0.4607 1.8686 1.3670
No log 2.3607 144 1.3216 0.5939 1.3216 1.1496
No log 2.3934 146 0.8721 0.7134 0.8721 0.9339
No log 2.4262 148 0.8113 0.7534 0.8113 0.9007
No log 2.4590 150 0.8409 0.6849 0.8409 0.9170
No log 2.4918 152 0.8582 0.6897 0.8582 0.9264
No log 2.5246 154 0.8850 0.6479 0.8850 0.9407
No log 2.5574 156 1.0916 0.6316 1.0916 1.0448
No log 2.5902 158 1.1891 0.6424 1.1891 1.0904
No log 2.6230 160 0.9346 0.6832 0.9346 0.9667
No log 2.6557 162 0.7759 0.7417 0.7759 0.8809
No log 2.6885 164 0.7585 0.7417 0.7585 0.8709
No log 2.7213 166 0.7579 0.7417 0.7579 0.8706
No log 2.7541 168 0.7741 0.7568 0.7741 0.8798
No log 2.7869 170 0.8009 0.6944 0.8009 0.8949
No log 2.8197 172 0.7950 0.7172 0.7950 0.8916
No log 2.8525 174 0.7840 0.7517 0.7840 0.8854
No log 2.8852 176 0.7740 0.7582 0.7740 0.8798
No log 2.9180 178 0.7801 0.7226 0.7801 0.8833
No log 2.9508 180 0.8070 0.7436 0.8070 0.8983
No log 2.9836 182 0.8219 0.7436 0.8219 0.9066
No log 3.0164 184 0.7724 0.7436 0.7724 0.8789
No log 3.0492 186 0.7318 0.7643 0.7318 0.8554
No log 3.0820 188 0.7204 0.7607 0.7204 0.8488
No log 3.1148 190 0.6993 0.7547 0.6993 0.8363
No log 3.1475 192 0.7397 0.7484 0.7397 0.8601
No log 3.1803 194 0.7708 0.7436 0.7708 0.8779
No log 3.2131 196 0.7584 0.7436 0.7584 0.8709
No log 3.2459 198 0.7938 0.7261 0.7938 0.8909
No log 3.2787 200 0.7107 0.7436 0.7107 0.8430
No log 3.3115 202 0.7305 0.7758 0.7305 0.8547
No log 3.3443 204 0.8891 0.7333 0.8891 0.9429
No log 3.3770 206 0.8677 0.7586 0.8677 0.9315
No log 3.4098 208 0.8010 0.7485 0.8010 0.8950
No log 3.4426 210 0.8202 0.7006 0.8202 0.9056
No log 3.4754 212 0.9645 0.7059 0.9645 0.9821
No log 3.5082 214 1.2257 0.5641 1.2257 1.1071
No log 3.5410 216 1.0935 0.5957 1.0935 1.0457
No log 3.5738 218 0.9897 0.6986 0.9897 0.9948
No log 3.6066 220 0.8622 0.7083 0.8622 0.9285
No log 3.6393 222 0.8610 0.6763 0.8610 0.9279
No log 3.6721 224 0.8570 0.7092 0.8570 0.9257
No log 3.7049 226 0.8249 0.7183 0.8249 0.9082
No log 3.7377 228 0.7800 0.7297 0.7800 0.8832
No log 3.7705 230 0.7575 0.7368 0.7575 0.8703
No log 3.8033 232 0.6860 0.7742 0.6860 0.8283
No log 3.8361 234 0.6453 0.7895 0.6453 0.8033
No log 3.8689 236 0.6426 0.7895 0.6426 0.8016
No log 3.9016 238 0.6478 0.7974 0.6478 0.8049
No log 3.9344 240 0.6902 0.7564 0.6902 0.8308
No log 3.9672 242 0.7673 0.7329 0.7673 0.8760
No log 4.0 244 0.8216 0.6792 0.8216 0.9064
No log 4.0328 246 0.7627 0.7226 0.7627 0.8733
No log 4.0656 248 0.6921 0.7673 0.6921 0.8319
No log 4.0984 250 0.7095 0.7362 0.7095 0.8423
No log 4.1311 252 0.7033 0.7362 0.7033 0.8386
No log 4.1639 254 0.6813 0.7758 0.6813 0.8254
No log 4.1967 256 0.6718 0.7329 0.6718 0.8196
No log 4.2295 258 0.6674 0.7329 0.6674 0.8169
No log 4.2623 260 0.6588 0.7738 0.6588 0.8117
No log 4.2951 262 0.6871 0.7836 0.6871 0.8289
No log 4.3279 264 0.7133 0.8023 0.7133 0.8446
No log 4.3607 266 0.6720 0.7595 0.6720 0.8198
No log 4.3934 268 0.7612 0.7324 0.7612 0.8725
No log 4.4262 270 0.9074 0.6950 0.9074 0.9526
No log 4.4590 272 1.0239 0.6260 1.0239 1.0119
No log 4.4918 274 0.9547 0.6466 0.9547 0.9771
No log 4.5246 276 0.8487 0.7286 0.8487 0.9213
No log 4.5574 278 0.7501 0.7183 0.7501 0.8661
No log 4.5902 280 0.6740 0.7432 0.6740 0.8210
No log 4.6230 282 0.6301 0.7785 0.6301 0.7938
No log 4.6557 284 0.6149 0.7742 0.6149 0.7842
No log 4.6885 286 0.5951 0.7784 0.5951 0.7714
No log 4.7213 288 0.6488 0.8047 0.6488 0.8055
No log 4.7541 290 0.6098 0.7857 0.6098 0.7809
No log 4.7869 292 0.6272 0.7692 0.6272 0.7919
No log 4.8197 294 0.6524 0.7692 0.6524 0.8077
No log 4.8525 296 0.7073 0.7342 0.7073 0.8410
No log 4.8852 298 0.8413 0.6667 0.8413 0.9172
No log 4.9180 300 0.8732 0.6525 0.8732 0.9344
No log 4.9508 302 0.7616 0.7183 0.7616 0.8727
No log 4.9836 304 0.7314 0.7383 0.7314 0.8552
No log 5.0164 306 0.7516 0.7792 0.7516 0.8670
No log 5.0492 308 0.7412 0.7383 0.7412 0.8610
No log 5.0820 310 0.7345 0.7432 0.7345 0.8570
No log 5.1148 312 0.7222 0.7389 0.7222 0.8498
No log 5.1475 314 0.7265 0.7515 0.7265 0.8524
No log 5.1803 316 0.7239 0.7799 0.7239 0.8508
No log 5.2131 318 0.7178 0.7712 0.7178 0.8472
No log 5.2459 320 0.7103 0.7682 0.7103 0.8428
No log 5.2787 322 0.7131 0.7531 0.7131 0.8444
No log 5.3115 324 0.8119 0.6788 0.8119 0.9010
No log 5.3443 326 0.7076 0.7561 0.7076 0.8412
No log 5.3770 328 0.6346 0.7975 0.6346 0.7966
No log 5.4098 330 0.6598 0.7397 0.6598 0.8123
No log 5.4426 332 0.6851 0.7260 0.6851 0.8277
No log 5.4754 334 0.6833 0.7324 0.6833 0.8266
No log 5.5082 336 0.6964 0.7606 0.6964 0.8345
No log 5.5410 338 0.7565 0.6716 0.7565 0.8698
No log 5.5738 340 0.7590 0.6963 0.7590 0.8712
No log 5.6066 342 0.7077 0.7338 0.7077 0.8413
No log 5.6393 344 0.6610 0.7518 0.6610 0.8130
No log 5.6721 346 0.6530 0.7297 0.6530 0.8081
No log 5.7049 348 0.6835 0.7114 0.6835 0.8267
No log 5.7377 350 0.6681 0.7484 0.6681 0.8173
No log 5.7705 352 0.6374 0.7683 0.6374 0.7984
No log 5.8033 354 0.6784 0.8 0.6784 0.8236
No log 5.8361 356 0.7980 0.7598 0.7980 0.8933
No log 5.8689 358 0.7924 0.7458 0.7924 0.8901
No log 5.9016 360 0.6930 0.7805 0.6930 0.8325
No log 5.9344 362 0.6713 0.7975 0.6713 0.8194
No log 5.9672 364 0.7051 0.7162 0.7051 0.8397
No log 6.0 366 0.6963 0.7211 0.6963 0.8345
No log 6.0328 368 0.6783 0.7534 0.6783 0.8236
No log 6.0656 370 0.6603 0.7639 0.6603 0.8126
No log 6.0984 372 0.6227 0.7815 0.6227 0.7891
No log 6.1311 374 0.6212 0.7898 0.6212 0.7882
No log 6.1639 376 0.6026 0.7949 0.6026 0.7763
No log 6.1967 378 0.5972 0.8050 0.5972 0.7728
No log 6.2295 380 0.6078 0.8075 0.6078 0.7796
No log 6.2623 382 0.6221 0.7927 0.6221 0.7887
No log 6.2951 384 0.6770 0.7901 0.6770 0.8228
No log 6.3279 386 0.6999 0.7738 0.6999 0.8366
No log 6.3607 388 0.6814 0.7952 0.6814 0.8255
No log 6.3934 390 0.6622 0.8024 0.6622 0.8137
No log 6.4262 392 0.6622 0.7950 0.6622 0.8138
No log 6.4590 394 0.6841 0.8 0.6841 0.8271
No log 6.4918 396 0.7045 0.8105 0.7045 0.8394
No log 6.5246 398 0.7391 0.7815 0.7391 0.8597
No log 6.5574 400 0.8001 0.6968 0.8001 0.8945
No log 6.5902 402 0.7816 0.7403 0.7816 0.8841
No log 6.6230 404 0.7048 0.7947 0.7048 0.8395
No log 6.6557 406 0.6833 0.7919 0.6833 0.8266
No log 6.6885 408 0.6648 0.7867 0.6648 0.8153
No log 6.7213 410 0.6597 0.7875 0.6597 0.8122
No log 6.7541 412 0.6391 0.8024 0.6391 0.7994
No log 6.7869 414 0.6304 0.8024 0.6304 0.7940
No log 6.8197 416 0.6699 0.7647 0.6699 0.8185
No log 6.8525 418 0.6948 0.7590 0.6948 0.8336
No log 6.8852 420 0.6817 0.7799 0.6817 0.8257
No log 6.9180 422 0.6476 0.7724 0.6476 0.8048
No log 6.9508 424 0.6873 0.7234 0.6873 0.8290
No log 6.9836 426 0.7054 0.7083 0.7054 0.8399
No log 7.0164 428 0.6754 0.7123 0.6754 0.8218
No log 7.0492 430 0.6401 0.7248 0.6401 0.8001
No log 7.0820 432 0.6414 0.7248 0.6414 0.8009
No log 7.1148 434 0.6244 0.7550 0.6244 0.7902
No log 7.1475 436 0.6412 0.8025 0.6412 0.8007
No log 7.1803 438 0.7165 0.7702 0.7165 0.8465
No log 7.2131 440 0.7401 0.7547 0.7401 0.8603
No log 7.2459 442 0.6723 0.7949 0.6723 0.8199
No log 7.2787 444 0.6569 0.7310 0.6569 0.8105
No log 7.3115 446 0.7655 0.7042 0.7655 0.8749
No log 7.3443 448 0.7873 0.6944 0.7873 0.8873
No log 7.3770 450 0.7072 0.7183 0.7072 0.8410
No log 7.4098 452 0.6512 0.7413 0.6512 0.8070
No log 7.4426 454 0.6396 0.7413 0.6396 0.7998
No log 7.4754 456 0.7060 0.7429 0.7060 0.8402
No log 7.5082 458 0.7453 0.7324 0.7453 0.8633
No log 7.5410 460 0.7042 0.7376 0.7042 0.8391
No log 7.5738 462 0.6463 0.7733 0.6463 0.8039
No log 7.6066 464 0.6553 0.7733 0.6553 0.8095
No log 7.6393 466 0.6573 0.75 0.6573 0.8108
No log 7.6721 468 0.6552 0.7465 0.6552 0.8094
No log 7.7049 470 0.6748 0.7552 0.6748 0.8214
No log 7.7377 472 0.6383 0.7660 0.6383 0.7989
No log 7.7705 474 0.6188 0.7703 0.6188 0.7866
No log 7.8033 476 0.6082 0.7867 0.6082 0.7799
No log 7.8361 478 0.6030 0.7815 0.6030 0.7765
No log 7.8689 480 0.6306 0.7619 0.6306 0.7941
No log 7.9016 482 0.6527 0.7383 0.6527 0.8079
No log 7.9344 484 0.6260 0.7763 0.6260 0.7912
No log 7.9672 486 0.6291 0.7682 0.6291 0.7932
No log 8.0 488 0.6564 0.7853 0.6564 0.8102
No log 8.0328 490 0.6962 0.7683 0.6962 0.8344
No log 8.0656 492 0.6737 0.8 0.6737 0.8208
No log 8.0984 494 0.6664 0.7792 0.6664 0.8163
No log 8.1311 496 0.6890 0.7682 0.6890 0.8300
No log 8.1639 498 0.7049 0.7733 0.7049 0.8396
0.4642 8.1967 500 0.7020 0.7712 0.7020 0.8379
0.4642 8.2295 502 0.6819 0.7843 0.6819 0.8258
0.4642 8.2623 504 0.6638 0.8050 0.6638 0.8147
0.4642 8.2951 506 0.6525 0.8 0.6525 0.8078
0.4642 8.3279 508 0.6394 0.7952 0.6394 0.7996
0.4642 8.3607 510 0.6204 0.8047 0.6204 0.7877
0.4642 8.3934 512 0.6090 0.7953 0.6090 0.7804
0.4642 8.4262 514 0.6168 0.8118 0.6168 0.7853
0.4642 8.4590 516 0.6220 0.7953 0.6220 0.7887
0.4642 8.4918 518 0.6154 0.7602 0.6154 0.7845
0.4642 8.5246 520 0.6438 0.7619 0.6438 0.8023
0.4642 8.5574 522 0.6478 0.7758 0.6478 0.8049
0.4642 8.5902 524 0.6612 0.7826 0.6612 0.8132
0.4642 8.6230 526 0.7056 0.7421 0.7056 0.8400
0.4642 8.6557 528 0.7299 0.7320 0.7299 0.8544
0.4642 8.6885 530 0.7301 0.7550 0.7301 0.8545
0.4642 8.7213 532 0.7164 0.7582 0.7164 0.8464
0.4642 8.7541 534 0.7051 0.7843 0.7051 0.8397
0.4642 8.7869 536 0.7000 0.7815 0.7000 0.8366
0.4642 8.8197 538 0.6900 0.7733 0.6900 0.8307
0.4642 8.8525 540 0.6673 0.7682 0.6673 0.8169
0.4642 8.8852 542 0.6772 0.7651 0.6772 0.8229
0.4642 8.9180 544 0.7243 0.7467 0.7243 0.8510
0.4642 8.9508 546 0.8313 0.6839 0.8313 0.9117
0.4642 8.9836 548 0.8513 0.6839 0.8513 0.9227

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task1_organization

Finetuned
(4019)
this model