ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8716
  • Qwk: 0.6259
  • Mse: 0.8716
  • Rmse: 0.9336

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0328 2 6.8370 0.0242 6.8370 2.6148
No log 0.0656 4 4.5285 0.0714 4.5285 2.1280
No log 0.0984 6 3.4724 0.0108 3.4724 1.8634
No log 0.1311 8 3.4691 -0.0437 3.4691 1.8626
No log 0.1639 10 2.9540 0.0133 2.9540 1.7187
No log 0.1967 12 3.1706 -0.0848 3.1706 1.7806
No log 0.2295 14 3.9163 -0.0201 3.9163 1.9790
No log 0.2623 16 3.3056 -0.0230 3.3056 1.8181
No log 0.2951 18 2.2264 0.128 2.2264 1.4921
No log 0.3279 20 2.0182 0.0862 2.0182 1.4206
No log 0.3607 22 2.0906 0.1333 2.0906 1.4459
No log 0.3934 24 2.2196 0.1958 2.2196 1.4898
No log 0.4262 26 2.7602 0.0892 2.7602 1.6614
No log 0.4590 28 3.2743 0.0936 3.2743 1.8095
No log 0.4918 30 3.0239 0.0958 3.0239 1.7389
No log 0.5246 32 2.3818 0.1053 2.3818 1.5433
No log 0.5574 34 1.8666 0.3158 1.8666 1.3662
No log 0.5902 36 1.7671 0.3577 1.7671 1.3293
No log 0.6230 38 1.7836 0.3721 1.7836 1.3355
No log 0.6557 40 1.7440 0.3622 1.7440 1.3206
No log 0.6885 42 1.8532 0.3111 1.8532 1.3613
No log 0.7213 44 1.8462 0.3043 1.8462 1.3588
No log 0.7541 46 1.7443 0.3111 1.7443 1.3207
No log 0.7869 48 1.5541 0.3511 1.5541 1.2467
No log 0.8197 50 1.5515 0.3511 1.5515 1.2456
No log 0.8525 52 1.7073 0.3008 1.7073 1.3066
No log 0.8852 54 1.9226 0.25 1.9226 1.3866
No log 0.9180 56 2.7961 0.1778 2.7961 1.6722
No log 0.9508 58 2.9017 0.2011 2.9017 1.7034
No log 0.9836 60 2.1900 0.2614 2.1900 1.4799
No log 1.0164 62 2.1579 0.2914 2.1579 1.4690
No log 1.0492 64 2.2993 0.2338 2.2993 1.5163
No log 1.0820 66 2.7593 0.2235 2.7593 1.6611
No log 1.1148 68 2.7572 0.2135 2.7572 1.6605
No log 1.1475 70 2.2282 0.2308 2.2282 1.4927
No log 1.1803 72 1.7880 0.3497 1.7880 1.3372
No log 1.2131 74 1.4809 0.4361 1.4809 1.2169
No log 1.2459 76 1.5013 0.4559 1.5013 1.2253
No log 1.2787 78 1.6155 0.3830 1.6155 1.2710
No log 1.3115 80 1.4866 0.4571 1.4866 1.2193
No log 1.3443 82 1.4822 0.5211 1.4822 1.2175
No log 1.3770 84 1.3868 0.4857 1.3868 1.1776
No log 1.4098 86 1.3617 0.5035 1.3617 1.1669
No log 1.4426 88 1.2369 0.4964 1.2369 1.1122
No log 1.4754 90 1.3051 0.5 1.3051 1.1424
No log 1.5082 92 1.2094 0.5147 1.2094 1.0997
No log 1.5410 94 1.0813 0.5271 1.0813 1.0398
No log 1.5738 96 1.0606 0.5854 1.0606 1.0299
No log 1.6066 98 1.0681 0.5203 1.0681 1.0335
No log 1.6393 100 1.0855 0.5496 1.0855 1.0419
No log 1.6721 102 1.1423 0.5564 1.1423 1.0688
No log 1.7049 104 1.1769 0.5109 1.1769 1.0848
No log 1.7377 106 1.1516 0.5180 1.1516 1.0731
No log 1.7705 108 1.1335 0.5455 1.1335 1.0646
No log 1.8033 110 1.3847 0.5629 1.3847 1.1767
No log 1.8361 112 1.5539 0.5207 1.5539 1.2466
No log 1.8689 114 1.1775 0.5960 1.1775 1.0851
No log 1.9016 116 0.9908 0.6338 0.9908 0.9954
No log 1.9344 118 1.0169 0.5755 1.0169 1.0084
No log 1.9672 120 1.1491 0.4928 1.1491 1.0720
No log 2.0 122 1.2010 0.5106 1.2010 1.0959
No log 2.0328 124 1.2138 0.5175 1.2138 1.1017
No log 2.0656 126 1.0468 0.5857 1.0468 1.0231
No log 2.0984 128 0.9477 0.6338 0.9477 0.9735
No log 2.1311 130 0.9357 0.6483 0.9357 0.9673
No log 2.1639 132 0.8998 0.6575 0.8998 0.9486
No log 2.1967 134 0.8879 0.6759 0.8879 0.9423
No log 2.2295 136 0.8595 0.6993 0.8595 0.9271
No log 2.2623 138 0.8711 0.6993 0.8711 0.9333
No log 2.2951 140 1.0294 0.5946 1.0294 1.0146
No log 2.3279 142 1.3584 0.5660 1.3584 1.1655
No log 2.3607 144 1.3625 0.5663 1.3625 1.1673
No log 2.3934 146 1.2319 0.5767 1.2319 1.1099
No log 2.4262 148 0.9602 0.6883 0.9602 0.9799
No log 2.4590 150 0.9241 0.6857 0.9241 0.9613
No log 2.4918 152 0.9472 0.6809 0.9472 0.9732
No log 2.5246 154 1.0583 0.5833 1.0583 1.0287
No log 2.5574 156 1.1127 0.5676 1.1127 1.0549
No log 2.5902 158 1.0451 0.5833 1.0451 1.0223
No log 2.6230 160 0.9164 0.7075 0.9164 0.9573
No log 2.6557 162 0.9002 0.6853 0.9002 0.9488
No log 2.6885 164 0.9963 0.6579 0.9963 0.9981
No log 2.7213 166 1.2817 0.5590 1.2817 1.1321
No log 2.7541 168 1.2951 0.5590 1.2951 1.1380
No log 2.7869 170 0.9944 0.6176 0.9944 0.9972
No log 2.8197 172 0.9834 0.6154 0.9834 0.9917
No log 2.8525 174 1.0832 0.5440 1.0832 1.0408
No log 2.8852 176 1.1076 0.5440 1.1076 1.0524
No log 2.9180 178 1.0290 0.5736 1.0290 1.0144
No log 2.9508 180 1.0065 0.5954 1.0065 1.0032
No log 2.9836 182 1.0762 0.5147 1.0762 1.0374
No log 3.0164 184 1.0886 0.5333 1.0886 1.0434
No log 3.0492 186 0.9892 0.5616 0.9892 0.9946
No log 3.0820 188 0.9302 0.6429 0.9302 0.9645
No log 3.1148 190 0.9146 0.6479 0.9146 0.9563
No log 3.1475 192 0.9039 0.6761 0.9039 0.9507
No log 3.1803 194 0.9146 0.6761 0.9146 0.9563
No log 3.2131 196 1.0526 0.5468 1.0526 1.0260
No log 3.2459 198 1.2804 0.4795 1.2804 1.1316
No log 3.2787 200 1.1821 0.5468 1.1821 1.0872
No log 3.3115 202 0.9848 0.6119 0.9848 0.9924
No log 3.3443 204 1.0391 0.6522 1.0391 1.0194
No log 3.3770 206 1.1689 0.5564 1.1689 1.0812
No log 3.4098 208 1.1613 0.5736 1.1613 1.0777
No log 3.4426 210 1.1603 0.5079 1.1603 1.0772
No log 3.4754 212 1.1637 0.4640 1.1637 1.0787
No log 3.5082 214 1.1363 0.5156 1.1363 1.0660
No log 3.5410 216 1.0916 0.5496 1.0916 1.0448
No log 3.5738 218 1.0583 0.6015 1.0583 1.0287
No log 3.6066 220 1.0137 0.6029 1.0137 1.0068
No log 3.6393 222 1.0168 0.6187 1.0168 1.0084
No log 3.6721 224 0.9693 0.6619 0.9693 0.9845
No log 3.7049 226 0.9376 0.6897 0.9376 0.9683
No log 3.7377 228 0.9523 0.6434 0.9523 0.9759
No log 3.7705 230 0.8893 0.7143 0.8893 0.9430
No log 3.8033 232 0.8991 0.6853 0.8991 0.9482
No log 3.8361 234 1.0980 0.5390 1.0980 1.0478
No log 3.8689 236 1.0823 0.5390 1.0823 1.0403
No log 3.9016 238 0.9026 0.6712 0.9026 0.9500
No log 3.9344 240 0.8536 0.7368 0.8536 0.9239
No log 3.9672 242 0.9469 0.6289 0.9469 0.9731
No log 4.0 244 1.0376 0.5987 1.0376 1.0186
No log 4.0328 246 1.1161 0.575 1.1161 1.0565
No log 4.0656 248 1.1742 0.5854 1.1742 1.0836
No log 4.0984 250 1.1537 0.5939 1.1537 1.0741
No log 4.1311 252 0.9979 0.6258 0.9979 0.9990
No log 4.1639 254 1.1047 0.6228 1.1047 1.0511
No log 4.1967 256 1.2043 0.6199 1.2043 1.0974
No log 4.2295 258 1.0648 0.6135 1.0648 1.0319
No log 4.2623 260 1.0753 0.6211 1.0753 1.0370
No log 4.2951 262 1.2577 0.6098 1.2577 1.1215
No log 4.3279 264 1.2785 0.6098 1.2785 1.1307
No log 4.3607 266 1.4629 0.5939 1.4629 1.2095
No log 4.3934 268 1.3259 0.5988 1.3259 1.1515
No log 4.4262 270 1.0207 0.6098 1.0207 1.0103
No log 4.4590 272 0.8820 0.7205 0.8820 0.9392
No log 4.4918 274 0.8305 0.7342 0.8305 0.9113
No log 4.5246 276 0.8144 0.7532 0.8144 0.9024
No log 4.5574 278 0.8487 0.7248 0.8487 0.9213
No log 4.5902 280 0.9270 0.7111 0.9270 0.9628
No log 4.6230 282 0.9795 0.7015 0.9795 0.9897
No log 4.6557 284 0.9751 0.5385 0.9751 0.9875
No log 4.6885 286 1.0172 0.5333 1.0172 1.0086
No log 4.7213 288 1.0216 0.5333 1.0216 1.0108
No log 4.7541 290 0.9625 0.5797 0.9625 0.9811
No log 4.7869 292 0.8877 0.6901 0.8877 0.9422
No log 4.8197 294 0.8395 0.6993 0.8395 0.9162
No log 4.8525 296 0.8460 0.6857 0.8460 0.9198
No log 4.8852 298 0.8959 0.6286 0.8959 0.9465
No log 4.9180 300 0.9901 0.5594 0.9901 0.9950
No log 4.9508 302 0.9000 0.6043 0.9000 0.9487
No log 4.9836 304 0.8618 0.6993 0.8618 0.9283
No log 5.0164 306 0.8627 0.6866 0.8627 0.9288
No log 5.0492 308 0.8977 0.7111 0.8977 0.9475
No log 5.0820 310 0.8817 0.6963 0.8817 0.9390
No log 5.1148 312 0.8659 0.6412 0.8659 0.9306
No log 5.1475 314 0.8812 0.6087 0.8812 0.9387
No log 5.1803 316 0.8857 0.5942 0.8857 0.9411
No log 5.2131 318 0.8845 0.6525 0.8845 0.9405
No log 5.2459 320 0.8913 0.6957 0.8913 0.9441
No log 5.2787 322 0.9033 0.6957 0.9033 0.9504
No log 5.3115 324 0.8889 0.6713 0.8889 0.9428
No log 5.3443 326 0.9512 0.6323 0.9512 0.9753
No log 5.3770 328 0.9529 0.6625 0.9529 0.9762
No log 5.4098 330 0.8816 0.6839 0.8816 0.9389
No log 5.4426 332 0.8645 0.6906 0.8645 0.9298
No log 5.4754 334 0.8757 0.6906 0.8757 0.9358
No log 5.5082 336 0.8600 0.6906 0.8600 0.9274
No log 5.5410 338 0.8371 0.6620 0.8371 0.9149
No log 5.5738 340 0.8579 0.6533 0.8579 0.9262
No log 5.6066 342 0.8622 0.6486 0.8622 0.9285
No log 5.6393 344 0.8395 0.6528 0.8395 0.9162
No log 5.6721 346 0.8075 0.7222 0.8075 0.8986
No log 5.7049 348 0.8290 0.6917 0.8290 0.9105
No log 5.7377 350 0.8776 0.6667 0.8776 0.9368
No log 5.7705 352 0.8971 0.6970 0.8971 0.9472
No log 5.8033 354 0.8736 0.6429 0.8736 0.9347
No log 5.8361 356 0.9702 0.5946 0.9702 0.9850
No log 5.8689 358 1.0580 0.6115 1.0580 1.0286
No log 5.9016 360 0.9732 0.6329 0.9732 0.9865
No log 5.9344 362 0.8413 0.6713 0.8413 0.9172
No log 5.9672 364 0.8213 0.7194 0.8213 0.9063
No log 6.0 366 0.8298 0.7194 0.8298 0.9109
No log 6.0328 368 0.8305 0.7101 0.8305 0.9113
No log 6.0656 370 0.8698 0.5839 0.8698 0.9326
No log 6.0984 372 0.8725 0.5882 0.8725 0.9341
No log 6.1311 374 0.8362 0.6815 0.8362 0.9144
No log 6.1639 376 0.8409 0.6866 0.8409 0.9170
No log 6.1967 378 0.8750 0.5942 0.8750 0.9354
No log 6.2295 380 1.0698 0.6026 1.0698 1.0343
No log 6.2623 382 1.2150 0.5732 1.2150 1.1023
No log 6.2951 384 1.1571 0.5949 1.1571 1.0757
No log 6.3279 386 0.9590 0.6625 0.9590 0.9793
No log 6.3607 388 0.8267 0.7237 0.8267 0.9092
No log 6.3934 390 0.8041 0.7320 0.8041 0.8967
No log 6.4262 392 0.7973 0.7320 0.7973 0.8929
No log 6.4590 394 0.7961 0.7547 0.7961 0.8922
No log 6.4918 396 0.8508 0.7425 0.8508 0.9224
No log 6.5246 398 0.8786 0.6748 0.8786 0.9373
No log 6.5574 400 0.8499 0.7407 0.8499 0.9219
No log 6.5902 402 0.8278 0.7516 0.8278 0.9098
No log 6.6230 404 0.8473 0.6944 0.8473 0.9205
No log 6.6557 406 0.8935 0.6861 0.8935 0.9452
No log 6.6885 408 0.9263 0.5970 0.9263 0.9625
No log 6.7213 410 0.9344 0.5606 0.9344 0.9667
No log 6.7541 412 0.9460 0.5672 0.9460 0.9726
No log 6.7869 414 0.9048 0.6087 0.9048 0.9512
No log 6.8197 416 0.8559 0.7183 0.8559 0.9251
No log 6.8525 418 0.8613 0.7059 0.8613 0.9281
No log 6.8852 420 0.9018 0.7068 0.9018 0.9496
No log 6.9180 422 0.8939 0.7068 0.8939 0.9454
No log 6.9508 424 0.9044 0.6870 0.9044 0.9510
No log 6.9836 426 0.9435 0.6212 0.9435 0.9714
No log 7.0164 428 0.9555 0.5846 0.9555 0.9775
No log 7.0492 430 0.9221 0.6316 0.9221 0.9603
No log 7.0820 432 0.9074 0.7101 0.9074 0.9526
No log 7.1148 434 0.8794 0.6667 0.8794 0.9378
No log 7.1475 436 0.9034 0.6111 0.9034 0.9505
No log 7.1803 438 0.9383 0.6014 0.9383 0.9686
No log 7.2131 440 0.9597 0.6111 0.9597 0.9797
No log 7.2459 442 0.9023 0.6143 0.9023 0.9499
No log 7.2787 444 0.8587 0.7034 0.8587 0.9267
No log 7.3115 446 0.8390 0.7034 0.8390 0.9160
No log 7.3443 448 0.8528 0.7034 0.8528 0.9235
No log 7.3770 450 0.9193 0.6027 0.9193 0.9588
No log 7.4098 452 0.9286 0.6410 0.9286 0.9636
No log 7.4426 454 0.8886 0.6624 0.8886 0.9426
No log 7.4754 456 0.8492 0.7089 0.8492 0.9215
No log 7.5082 458 0.8305 0.7468 0.8305 0.9113
No log 7.5410 460 0.8314 0.7403 0.8314 0.9118
No log 7.5738 462 0.8659 0.7013 0.8659 0.9305
No log 7.6066 464 0.9566 0.6364 0.9566 0.9780
No log 7.6393 466 1.1011 0.6391 1.1011 1.0494
No log 7.6721 468 1.2133 0.5952 1.2133 1.1015
No log 7.7049 470 1.1266 0.6303 1.1266 1.0614
No log 7.7377 472 0.9364 0.6497 0.9364 0.9677
No log 7.7705 474 0.8051 0.7013 0.8051 0.8973
No log 7.8033 476 0.7617 0.6950 0.7617 0.8728
No log 7.8361 478 0.7952 0.7101 0.7952 0.8918
No log 7.8689 480 0.8161 0.7246 0.8161 0.9034
No log 7.9016 482 0.8247 0.7246 0.8247 0.9081
No log 7.9344 484 0.8082 0.7376 0.8082 0.8990
No log 7.9672 486 0.8081 0.7183 0.8081 0.8989
No log 8.0 488 0.8070 0.6857 0.8070 0.8983
No log 8.0328 490 0.8185 0.6857 0.8185 0.9047
No log 8.0656 492 0.7702 0.6757 0.7702 0.8776
No log 8.0984 494 0.7198 0.7383 0.7198 0.8484
No log 8.1311 496 0.7337 0.7595 0.7337 0.8566
No log 8.1639 498 0.7388 0.7532 0.7388 0.8595
0.4149 8.1967 500 0.7720 0.7123 0.7720 0.8787
0.4149 8.2295 502 0.8333 0.6667 0.8333 0.9129
0.4149 8.2623 504 0.8022 0.6853 0.8022 0.8957
0.4149 8.2951 506 0.7341 0.7083 0.7341 0.8568
0.4149 8.3279 508 0.7458 0.7436 0.7458 0.8636
0.4149 8.3607 510 0.8929 0.6667 0.8929 0.9449
0.4149 8.3934 512 0.9415 0.6708 0.9415 0.9703
0.4149 8.4262 514 0.8695 0.6957 0.8695 0.9325
0.4149 8.4590 516 0.7562 0.7927 0.7562 0.8696
0.4149 8.4918 518 0.7078 0.7226 0.7078 0.8413
0.4149 8.5246 520 0.7219 0.7342 0.7219 0.8496
0.4149 8.5574 522 0.7687 0.7826 0.7687 0.8768
0.4149 8.5902 524 0.7880 0.72 0.7880 0.8877
0.4149 8.6230 526 0.8238 0.6757 0.8238 0.9076
0.4149 8.6557 528 0.8737 0.6531 0.8737 0.9347
0.4149 8.6885 530 0.9371 0.6438 0.9371 0.9680
0.4149 8.7213 532 0.9474 0.6301 0.9474 0.9733
0.4149 8.7541 534 0.8716 0.6259 0.8716 0.9336

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task1_organization

Finetuned
(4023)
this model