ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8284
  • Qwk: 0.3797
  • Mse: 0.8284
  • Rmse: 0.9102

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0328 2 4.3445 -0.0355 4.3445 2.0843
No log 0.0656 4 2.7341 -0.0708 2.7341 1.6535
No log 0.0984 6 1.7311 -0.0233 1.7311 1.3157
No log 0.1311 8 1.4971 0.0881 1.4971 1.2235
No log 0.1639 10 1.3522 0.0320 1.3522 1.1628
No log 0.1967 12 1.1987 0.1532 1.1987 1.0948
No log 0.2295 14 1.1654 0.1431 1.1654 1.0795
No log 0.2623 16 1.2601 0.2227 1.2601 1.1226
No log 0.2951 18 1.0740 0.2087 1.0740 1.0363
No log 0.3279 20 1.1933 0.2915 1.1933 1.0924
No log 0.3607 22 1.2720 0.2785 1.2720 1.1278
No log 0.3934 24 1.1104 0.3410 1.1104 1.0537
No log 0.4262 26 1.1071 0.4211 1.1071 1.0522
No log 0.4590 28 1.0762 0.4335 1.0762 1.0374
No log 0.4918 30 1.2213 0.3405 1.2213 1.1051
No log 0.5246 32 1.5591 0.2170 1.5591 1.2486
No log 0.5574 34 1.1592 0.3363 1.1592 1.0767
No log 0.5902 36 1.0109 0.3525 1.0109 1.0054
No log 0.6230 38 1.0350 0.3358 1.0350 1.0173
No log 0.6557 40 1.0013 0.3229 1.0013 1.0007
No log 0.6885 42 1.1378 0.2925 1.1378 1.0667
No log 0.7213 44 1.3172 0.1778 1.3172 1.1477
No log 0.7541 46 1.2753 0.2011 1.2753 1.1293
No log 0.7869 48 1.1310 0.2980 1.1310 1.0635
No log 0.8197 50 1.1181 0.2968 1.1181 1.0574
No log 0.8525 52 1.0830 0.2904 1.0830 1.0407
No log 0.8852 54 1.0686 0.2904 1.0686 1.0337
No log 0.9180 56 1.1136 0.2754 1.1136 1.0553
No log 0.9508 58 1.2630 0.2767 1.2630 1.1238
No log 0.9836 60 2.1446 0.1595 2.1446 1.4644
No log 1.0164 62 3.4359 0.0580 3.4359 1.8536
No log 1.0492 64 3.4284 0.0580 3.4284 1.8516
No log 1.0820 66 2.4659 0.1285 2.4659 1.5703
No log 1.1148 68 1.2730 0.2993 1.2730 1.1283
No log 1.1475 70 1.0917 0.3070 1.0917 1.0449
No log 1.1803 72 1.1218 0.2047 1.1218 1.0592
No log 1.2131 74 1.0767 0.2300 1.0767 1.0376
No log 1.2459 76 1.0892 0.1961 1.0892 1.0437
No log 1.2787 78 1.4057 0.2204 1.4057 1.1856
No log 1.3115 80 1.6652 0.1791 1.6652 1.2904
No log 1.3443 82 1.5222 0.2065 1.5222 1.2338
No log 1.3770 84 1.6220 0.1806 1.6220 1.2736
No log 1.4098 86 1.9016 0.1780 1.9016 1.3790
No log 1.4426 88 1.7483 0.1612 1.7483 1.3222
No log 1.4754 90 1.3883 0.2052 1.3883 1.1783
No log 1.5082 92 1.0747 0.3066 1.0747 1.0367
No log 1.5410 94 1.0272 0.3891 1.0272 1.0135
No log 1.5738 96 1.1612 0.2864 1.1612 1.0776
No log 1.6066 98 1.2041 0.2998 1.2041 1.0973
No log 1.6393 100 1.3617 0.1774 1.3617 1.1669
No log 1.6721 102 1.2298 0.2206 1.2298 1.1090
No log 1.7049 104 0.9769 0.2995 0.9769 0.9884
No log 1.7377 106 1.0226 0.4254 1.0226 1.0113
No log 1.7705 108 1.0370 0.4042 1.0370 1.0183
No log 1.8033 110 1.0559 0.4313 1.0559 1.0276
No log 1.8361 112 1.1233 0.2931 1.1233 1.0599
No log 1.8689 114 1.1345 0.4778 1.1345 1.0651
No log 1.9016 116 1.2753 0.3208 1.2753 1.1293
No log 1.9344 118 1.4080 0.3175 1.4080 1.1866
No log 1.9672 120 1.3435 0.3206 1.3435 1.1591
No log 2.0 122 1.1917 0.3313 1.1917 1.0916
No log 2.0328 124 1.1598 0.3753 1.1598 1.0769
No log 2.0656 126 1.1666 0.3578 1.1666 1.0801
No log 2.0984 128 1.1196 0.3996 1.1196 1.0581
No log 2.1311 130 1.0684 0.4328 1.0684 1.0337
No log 2.1639 132 1.0743 0.3371 1.0743 1.0365
No log 2.1967 134 1.0812 0.3218 1.0812 1.0398
No log 2.2295 136 1.0362 0.3689 1.0362 1.0179
No log 2.2623 138 1.0444 0.3169 1.0444 1.0220
No log 2.2951 140 1.1100 0.2757 1.1100 1.0536
No log 2.3279 142 1.0229 0.2808 1.0229 1.0114
No log 2.3607 144 1.0008 0.2857 1.0008 1.0004
No log 2.3934 146 0.9744 0.3779 0.9744 0.9871
No log 2.4262 148 1.0273 0.3372 1.0273 1.0136
No log 2.4590 150 0.9912 0.4164 0.9912 0.9956
No log 2.4918 152 0.9156 0.4646 0.9156 0.9568
No log 2.5246 154 0.9720 0.4266 0.9720 0.9859
No log 2.5574 156 0.9571 0.5327 0.9571 0.9783
No log 2.5902 158 1.0497 0.4248 1.0497 1.0246
No log 2.6230 160 1.2034 0.3961 1.2034 1.0970
No log 2.6557 162 1.1537 0.3921 1.1537 1.0741
No log 2.6885 164 0.9438 0.4407 0.9438 0.9715
No log 2.7213 166 0.9255 0.5224 0.9255 0.9620
No log 2.7541 168 0.9364 0.3972 0.9364 0.9677
No log 2.7869 170 1.0604 0.3740 1.0604 1.0298
No log 2.8197 172 0.9829 0.3346 0.9829 0.9914
No log 2.8525 174 0.8984 0.3874 0.8984 0.9478
No log 2.8852 176 0.9113 0.3857 0.9113 0.9546
No log 2.9180 178 0.9470 0.3578 0.9470 0.9732
No log 2.9508 180 0.9371 0.3519 0.9371 0.9680
No log 2.9836 182 0.9559 0.3857 0.9559 0.9777
No log 3.0164 184 0.9889 0.4546 0.9889 0.9944
No log 3.0492 186 0.9800 0.3733 0.9800 0.9900
No log 3.0820 188 0.9880 0.3430 0.9880 0.9940
No log 3.1148 190 1.0405 0.2903 1.0405 1.0200
No log 3.1475 192 1.0542 0.3137 1.0542 1.0267
No log 3.1803 194 0.9808 0.3711 0.9808 0.9904
No log 3.2131 196 1.0861 0.4032 1.0861 1.0421
No log 3.2459 198 1.0910 0.4032 1.0910 1.0445
No log 3.2787 200 0.9884 0.4181 0.9884 0.9942
No log 3.3115 202 0.9982 0.3474 0.9982 0.9991
No log 3.3443 204 1.0306 0.4042 1.0306 1.0152
No log 3.3770 206 1.0767 0.3319 1.0767 1.0376
No log 3.4098 208 1.1596 0.3134 1.1596 1.0768
No log 3.4426 210 1.1522 0.3267 1.1522 1.0734
No log 3.4754 212 1.1058 0.3285 1.1058 1.0516
No log 3.5082 214 1.0428 0.4511 1.0428 1.0212
No log 3.5410 216 0.9855 0.3638 0.9855 0.9927
No log 3.5738 218 1.0740 0.4367 1.0740 1.0363
No log 3.6066 220 1.0364 0.4367 1.0364 1.0181
No log 3.6393 222 0.9500 0.3354 0.9500 0.9747
No log 3.6721 224 0.9661 0.4413 0.9661 0.9829
No log 3.7049 226 0.9399 0.3931 0.9399 0.9695
No log 3.7377 228 0.9088 0.3951 0.9088 0.9533
No log 3.7705 230 0.9721 0.4004 0.9721 0.9860
No log 3.8033 232 0.9365 0.3661 0.9365 0.9677
No log 3.8361 234 0.9871 0.3993 0.9871 0.9935
No log 3.8689 236 1.0654 0.4201 1.0654 1.0322
No log 3.9016 238 1.0417 0.3793 1.0417 1.0207
No log 3.9344 240 1.0047 0.4372 1.0047 1.0024
No log 3.9672 242 1.0143 0.4541 1.0143 1.0071
No log 4.0 244 1.0476 0.4469 1.0476 1.0235
No log 4.0328 246 1.0144 0.4067 1.0144 1.0072
No log 4.0656 248 0.9540 0.2976 0.9540 0.9767
No log 4.0984 250 0.9858 0.3474 0.9858 0.9929
No log 4.1311 252 1.0134 0.3198 1.0134 1.0067
No log 4.1639 254 1.0117 0.2786 1.0117 1.0058
No log 4.1967 256 1.0183 0.3365 1.0183 1.0091
No log 4.2295 258 1.0182 0.3608 1.0182 1.0091
No log 4.2623 260 1.0125 0.3980 1.0125 1.0062
No log 4.2951 262 1.0014 0.4300 1.0014 1.0007
No log 4.3279 264 0.9978 0.4572 0.9978 0.9989
No log 4.3607 266 0.9333 0.4235 0.9333 0.9661
No log 4.3934 268 0.9349 0.3711 0.9349 0.9669
No log 4.4262 270 1.0136 0.2863 1.0136 1.0068
No log 4.4590 272 1.0127 0.2863 1.0127 1.0063
No log 4.4918 274 0.9994 0.3697 0.9994 0.9997
No log 4.5246 276 1.0509 0.3367 1.0509 1.0251
No log 4.5574 278 1.0618 0.3498 1.0618 1.0304
No log 4.5902 280 1.0878 0.2783 1.0878 1.0430
No log 4.6230 282 1.0238 0.3914 1.0238 1.0118
No log 4.6557 284 0.9669 0.3839 0.9669 0.9833
No log 4.6885 286 0.9604 0.4093 0.9604 0.9800
No log 4.7213 288 0.9613 0.3821 0.9613 0.9804
No log 4.7541 290 0.9770 0.3711 0.9770 0.9884
No log 4.7869 292 1.0069 0.3699 1.0069 1.0034
No log 4.8197 294 1.0747 0.4454 1.0747 1.0367
No log 4.8525 296 1.0565 0.4459 1.0565 1.0279
No log 4.8852 298 1.0228 0.3618 1.0228 1.0113
No log 4.9180 300 1.0442 0.3261 1.0442 1.0219
No log 4.9508 302 1.0692 0.3268 1.0692 1.0340
No log 4.9836 304 1.1013 0.3101 1.1013 1.0494
No log 5.0164 306 1.0971 0.2864 1.0971 1.0474
No log 5.0492 308 1.0896 0.3758 1.0896 1.0438
No log 5.0820 310 1.0986 0.3902 1.0986 1.0481
No log 5.1148 312 1.1297 0.3019 1.1297 1.0629
No log 5.1475 314 1.1759 0.3580 1.1759 1.0844
No log 5.1803 316 1.1194 0.3575 1.1194 1.0580
No log 5.2131 318 1.0340 0.3715 1.0340 1.0169
No log 5.2459 320 1.0062 0.3808 1.0062 1.0031
No log 5.2787 322 1.0352 0.4843 1.0352 1.0174
No log 5.3115 324 0.9906 0.4722 0.9906 0.9953
No log 5.3443 326 0.9008 0.3820 0.9008 0.9491
No log 5.3770 328 0.9518 0.4067 0.9518 0.9756
No log 5.4098 330 0.9717 0.3959 0.9717 0.9858
No log 5.4426 332 0.9255 0.3685 0.9255 0.9620
No log 5.4754 334 1.0373 0.4004 1.0373 1.0185
No log 5.5082 336 1.2074 0.3396 1.2074 1.0988
No log 5.5410 338 1.2000 0.3437 1.2000 1.0954
No log 5.5738 340 1.1016 0.3959 1.1016 1.0496
No log 5.6066 342 0.9871 0.3762 0.9871 0.9935
No log 5.6393 344 0.9990 0.3286 0.9990 0.9995
No log 5.6721 346 0.9806 0.3678 0.9806 0.9902
No log 5.7049 348 0.9816 0.4271 0.9816 0.9908
No log 5.7377 350 1.0603 0.4139 1.0603 1.0297
No log 5.7705 352 1.0936 0.4695 1.0936 1.0458
No log 5.8033 354 1.0062 0.4373 1.0062 1.0031
No log 5.8361 356 0.9212 0.3398 0.9212 0.9598
No log 5.8689 358 0.9215 0.3818 0.9215 0.9600
No log 5.9016 360 0.9118 0.3552 0.9118 0.9549
No log 5.9344 362 0.9467 0.3243 0.9467 0.9730
No log 5.9672 364 1.0719 0.4694 1.0719 1.0353
No log 6.0 366 1.1633 0.4247 1.1633 1.0785
No log 6.0328 368 1.0800 0.3989 1.0800 1.0392
No log 6.0656 370 0.9732 0.3091 0.9732 0.9865
No log 6.0984 372 0.9875 0.3891 0.9875 0.9938
No log 6.1311 374 0.9863 0.4245 0.9863 0.9931
No log 6.1639 376 0.9880 0.3733 0.9880 0.9940
No log 6.1967 378 1.0035 0.2878 1.0035 1.0017
No log 6.2295 380 0.9897 0.3256 0.9897 0.9948
No log 6.2623 382 0.9438 0.3784 0.9438 0.9715
No log 6.2951 384 0.9155 0.3818 0.9155 0.9568
No log 6.3279 386 0.9083 0.3818 0.9083 0.9530
No log 6.3607 388 0.9028 0.3896 0.9028 0.9501
No log 6.3934 390 0.9457 0.4164 0.9457 0.9725
No log 6.4262 392 0.9526 0.3914 0.9526 0.9760
No log 6.4590 394 0.9421 0.3369 0.9421 0.9706
No log 6.4918 396 0.8994 0.3455 0.8994 0.9483
No log 6.5246 398 0.9096 0.3455 0.9096 0.9537
No log 6.5574 400 0.9352 0.3840 0.9352 0.9670
No log 6.5902 402 1.0012 0.3560 1.0012 1.0006
No log 6.6230 404 1.0164 0.4529 1.0164 1.0081
No log 6.6557 406 1.0446 0.4343 1.0446 1.0220
No log 6.6885 408 1.0522 0.4563 1.0522 1.0258
No log 6.7213 410 1.0510 0.4565 1.0510 1.0252
No log 6.7541 412 1.0147 0.3787 1.0147 1.0073
No log 6.7869 414 0.9981 0.3839 0.9981 0.9991
No log 6.8197 416 0.9732 0.3356 0.9732 0.9865
No log 6.8525 418 0.9620 0.3496 0.9620 0.9808
No log 6.8852 420 0.9683 0.3496 0.9683 0.9840
No log 6.9180 422 0.9569 0.3496 0.9569 0.9782
No log 6.9508 424 0.9378 0.3642 0.9378 0.9684
No log 6.9836 426 0.9431 0.3513 0.9431 0.9711
No log 7.0164 428 0.9621 0.4264 0.9621 0.9809
No log 7.0492 430 0.9952 0.4650 0.9952 0.9976
No log 7.0820 432 1.0123 0.4538 1.0123 1.0061
No log 7.1148 434 1.0316 0.5052 1.0316 1.0157
No log 7.1475 436 1.0718 0.4723 1.0718 1.0353
No log 7.1803 438 1.0830 0.4590 1.0830 1.0407
No log 7.2131 440 1.0229 0.4847 1.0229 1.0114
No log 7.2459 442 0.9485 0.3367 0.9485 0.9739
No log 7.2787 444 0.9478 0.3323 0.9478 0.9736
No log 7.3115 446 0.9396 0.3367 0.9396 0.9693
No log 7.3443 448 0.9572 0.2842 0.9572 0.9784
No log 7.3770 450 0.9583 0.2842 0.9583 0.9789
No log 7.4098 452 0.9756 0.3306 0.9756 0.9877
No log 7.4426 454 0.9598 0.3243 0.9598 0.9797
No log 7.4754 456 0.9507 0.3378 0.9507 0.9751
No log 7.5082 458 0.9443 0.3802 0.9443 0.9718
No log 7.5410 460 0.9274 0.4062 0.9274 0.9630
No log 7.5738 462 0.9372 0.3932 0.9372 0.9681
No log 7.6066 464 0.9496 0.3914 0.9496 0.9745
No log 7.6393 466 0.9565 0.3540 0.9565 0.9780
No log 7.6721 468 1.0105 0.3578 1.0105 1.0052
No log 7.7049 470 0.9920 0.3914 0.9920 0.9960
No log 7.7377 472 0.9509 0.3476 0.9509 0.9751
No log 7.7705 474 0.9452 0.3388 0.9452 0.9722
No log 7.8033 476 0.9666 0.3408 0.9666 0.9832
No log 7.8361 478 1.0196 0.3662 1.0196 1.0097
No log 7.8689 480 1.0130 0.3701 1.0130 1.0065
No log 7.9016 482 0.9939 0.3437 0.9939 0.9969
No log 7.9344 484 1.0318 0.2919 1.0318 1.0158
No log 7.9672 486 1.0951 0.3897 1.0951 1.0464
No log 8.0 488 1.0970 0.4234 1.0970 1.0474
No log 8.0328 490 1.0434 0.3930 1.0434 1.0214
No log 8.0656 492 0.9804 0.3609 0.9804 0.9902
No log 8.0984 494 0.9596 0.3800 0.9596 0.9796
No log 8.1311 496 0.9708 0.3474 0.9708 0.9853
No log 8.1639 498 0.9868 0.4250 0.9868 0.9934
0.3141 8.1967 500 1.0114 0.4350 1.0114 1.0057
0.3141 8.2295 502 0.9839 0.3728 0.9839 0.9919
0.3141 8.2623 504 0.9334 0.3634 0.9334 0.9661
0.3141 8.2951 506 0.9081 0.3994 0.9081 0.9530
0.3141 8.3279 508 0.8528 0.4033 0.8528 0.9235
0.3141 8.3607 510 0.8284 0.3797 0.8284 0.9102

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task5_organization

Finetuned
(4019)
this model