ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8914
  • Qwk: 0.2770
  • Mse: 0.8914
  • Rmse: 0.9441

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0351 2 4.0677 0.0034 4.0677 2.0169
No log 0.0702 4 2.3787 -0.0811 2.3787 1.5423
No log 0.1053 6 1.8034 -0.0557 1.8034 1.3429
No log 0.1404 8 1.1913 0.0608 1.1913 1.0915
No log 0.1754 10 1.1144 0.1476 1.1144 1.0557
No log 0.2105 12 1.2480 0.0852 1.2480 1.1172
No log 0.2456 14 1.1462 0.1740 1.1462 1.0706
No log 0.2807 16 1.1632 0.1576 1.1632 1.0785
No log 0.3158 18 1.1225 0.1482 1.1225 1.0595
No log 0.3509 20 0.9984 0.2764 0.9984 0.9992
No log 0.3860 22 0.9966 0.2366 0.9966 0.9983
No log 0.4211 24 1.0466 0.1901 1.0466 1.0230
No log 0.4561 26 1.0720 0.2196 1.0720 1.0354
No log 0.4912 28 1.2404 0.1379 1.2404 1.1137
No log 0.5263 30 1.2868 0.1525 1.2868 1.1344
No log 0.5614 32 1.1057 0.2659 1.1057 1.0515
No log 0.5965 34 1.0041 0.1989 1.0041 1.0020
No log 0.6316 36 1.0757 0.1601 1.0757 1.0372
No log 0.6667 38 1.0737 0.0823 1.0737 1.0362
No log 0.7018 40 1.1300 0.2352 1.1300 1.0630
No log 0.7368 42 1.1143 0.2616 1.1143 1.0556
No log 0.7719 44 1.0159 0.2068 1.0159 1.0079
No log 0.8070 46 0.9971 0.1218 0.9971 0.9985
No log 0.8421 48 1.0264 0.2145 1.0264 1.0131
No log 0.8772 50 1.1379 0.2590 1.1379 1.0667
No log 0.9123 52 1.1959 0.2455 1.1959 1.0936
No log 0.9474 54 1.0810 0.3832 1.0810 1.0397
No log 0.9825 56 0.9725 0.2879 0.9725 0.9861
No log 1.0175 58 0.9714 0.2739 0.9714 0.9856
No log 1.0526 60 0.9867 0.2879 0.9867 0.9933
No log 1.0877 62 1.0083 0.3759 1.0083 1.0041
No log 1.1228 64 1.0017 0.3603 1.0017 1.0008
No log 1.1579 66 0.9449 0.3631 0.9449 0.9721
No log 1.1930 68 0.8940 0.4054 0.8940 0.9455
No log 1.2281 70 0.9087 0.3458 0.9087 0.9533
No log 1.2632 72 0.8743 0.4122 0.8743 0.9350
No log 1.2982 74 0.7956 0.4707 0.7956 0.8919
No log 1.3333 76 0.8100 0.4526 0.8100 0.9000
No log 1.3684 78 0.9517 0.3885 0.9517 0.9756
No log 1.4035 80 1.0992 0.3355 1.0992 1.0484
No log 1.4386 82 0.8422 0.4479 0.8422 0.9177
No log 1.4737 84 0.7822 0.4843 0.7822 0.8844
No log 1.5088 86 0.7972 0.4948 0.7972 0.8929
No log 1.5439 88 0.7710 0.5463 0.7710 0.8780
No log 1.5789 90 0.8098 0.4548 0.8098 0.8999
No log 1.6140 92 0.8778 0.3919 0.8778 0.9369
No log 1.6491 94 0.8956 0.3919 0.8956 0.9464
No log 1.6842 96 0.8018 0.4337 0.8018 0.8954
No log 1.7193 98 0.8147 0.5455 0.8147 0.9026
No log 1.7544 100 0.8256 0.4416 0.8256 0.9086
No log 1.7895 102 0.9035 0.3935 0.9035 0.9505
No log 1.8246 104 0.8296 0.3915 0.8296 0.9108
No log 1.8596 106 0.8322 0.5279 0.8322 0.9122
No log 1.8947 108 0.8748 0.4738 0.8748 0.9353
No log 1.9298 110 0.8430 0.4475 0.8430 0.9182
No log 1.9649 112 0.8438 0.4405 0.8438 0.9186
No log 2.0 114 0.8338 0.4889 0.8338 0.9131
No log 2.0351 116 1.0261 0.4152 1.0261 1.0129
No log 2.0702 118 1.0787 0.3405 1.0787 1.0386
No log 2.1053 120 0.8481 0.5413 0.8481 0.9209
No log 2.1404 122 0.7744 0.5516 0.7744 0.8800
No log 2.1754 124 0.7759 0.4354 0.7759 0.8809
No log 2.2105 126 0.8148 0.5410 0.8148 0.9027
No log 2.2456 128 0.9815 0.4790 0.9815 0.9907
No log 2.2807 130 0.9221 0.4048 0.9221 0.9603
No log 2.3158 132 0.8431 0.4907 0.8431 0.9182
No log 2.3509 134 0.8964 0.4763 0.8964 0.9468
No log 2.3860 136 0.8672 0.4517 0.8672 0.9313
No log 2.4211 138 0.8332 0.4883 0.8332 0.9128
No log 2.4561 140 1.0467 0.4012 1.0467 1.0231
No log 2.4912 142 1.2017 0.2731 1.2017 1.0962
No log 2.5263 144 1.1170 0.3326 1.1170 1.0569
No log 2.5614 146 0.9571 0.3864 0.9571 0.9783
No log 2.5965 148 0.8924 0.4131 0.8924 0.9447
No log 2.6316 150 0.9192 0.4111 0.9192 0.9588
No log 2.6667 152 0.8864 0.4244 0.8864 0.9415
No log 2.7018 154 0.8716 0.4411 0.8716 0.9336
No log 2.7368 156 0.8817 0.4401 0.8817 0.9390
No log 2.7719 158 0.9175 0.4712 0.9175 0.9579
No log 2.8070 160 0.9693 0.4032 0.9693 0.9845
No log 2.8421 162 0.9778 0.3828 0.9778 0.9888
No log 2.8772 164 0.8994 0.4480 0.8994 0.9484
No log 2.9123 166 0.8458 0.4297 0.8458 0.9197
No log 2.9474 168 0.8795 0.4388 0.8795 0.9378
No log 2.9825 170 0.9185 0.4480 0.9185 0.9584
No log 3.0175 172 0.9273 0.4357 0.9273 0.9630
No log 3.0526 174 0.9128 0.4357 0.9128 0.9554
No log 3.0877 176 0.9205 0.4474 0.9205 0.9594
No log 3.1228 178 0.8979 0.4465 0.8979 0.9476
No log 3.1579 180 0.8795 0.4159 0.8795 0.9378
No log 3.1930 182 0.8470 0.4401 0.8470 0.9203
No log 3.2281 184 0.8504 0.4428 0.8504 0.9222
No log 3.2632 186 0.8662 0.4051 0.8662 0.9307
No log 3.2982 188 0.8994 0.3414 0.8994 0.9484
No log 3.3333 190 0.9502 0.3663 0.9502 0.9748
No log 3.3684 192 0.9845 0.4433 0.9845 0.9922
No log 3.4035 194 0.9164 0.3314 0.9164 0.9573
No log 3.4386 196 0.9351 0.3200 0.9351 0.9670
No log 3.4737 198 0.9405 0.3301 0.9405 0.9698
No log 3.5088 200 0.9752 0.3048 0.9752 0.9875
No log 3.5439 202 0.9789 0.3167 0.9789 0.9894
No log 3.5789 204 0.9199 0.3485 0.9199 0.9591
No log 3.6140 206 0.8954 0.4163 0.8954 0.9463
No log 3.6491 208 0.9023 0.4141 0.9023 0.9499
No log 3.6842 210 0.9751 0.4111 0.9751 0.9875
No log 3.7193 212 0.9424 0.4450 0.9424 0.9708
No log 3.7544 214 0.8954 0.3414 0.8954 0.9463
No log 3.7895 216 0.9300 0.4168 0.9300 0.9643
No log 3.8246 218 0.9191 0.3485 0.9191 0.9587
No log 3.8596 220 0.9126 0.2794 0.9126 0.9553
No log 3.8947 222 0.9165 0.3702 0.9165 0.9573
No log 3.9298 224 0.9033 0.3721 0.9033 0.9504
No log 3.9649 226 0.9027 0.4078 0.9027 0.9501
No log 4.0 228 0.9161 0.4310 0.9161 0.9571
No log 4.0351 230 0.9284 0.4310 0.9284 0.9635
No log 4.0702 232 0.9342 0.4174 0.9342 0.9665
No log 4.1053 234 0.9469 0.3902 0.9469 0.9731
No log 4.1404 236 0.9720 0.4051 0.9720 0.9859
No log 4.1754 238 0.9039 0.3902 0.9039 0.9507
No log 4.2105 240 0.8593 0.4227 0.8593 0.9270
No log 4.2456 242 0.8733 0.4966 0.8733 0.9345
No log 4.2807 244 0.8825 0.4843 0.8825 0.9394
No log 4.3158 246 0.9532 0.4318 0.9532 0.9763
No log 4.3509 248 1.0377 0.4151 1.0377 1.0187
No log 4.3860 250 1.0243 0.4036 1.0243 1.0121
No log 4.4211 252 0.9968 0.3883 0.9968 0.9984
No log 4.4561 254 0.9105 0.3209 0.9105 0.9542
No log 4.4912 256 0.8910 0.3485 0.8910 0.9439
No log 4.5263 258 0.9334 0.4471 0.9334 0.9661
No log 4.5614 260 1.0938 0.3596 1.0938 1.0459
No log 4.5965 262 1.3108 0.2449 1.3108 1.1449
No log 4.6316 264 1.3421 0.2123 1.3421 1.1585
No log 4.6667 266 1.1874 0.1961 1.1874 1.0897
No log 4.7018 268 1.0286 0.2135 1.0286 1.0142
No log 4.7368 270 0.9436 0.2351 0.9436 0.9714
No log 4.7719 272 0.9146 0.3200 0.9146 0.9564
No log 4.8070 274 0.8902 0.2670 0.8902 0.9435
No log 4.8421 276 0.9215 0.3822 0.9215 0.9600
No log 4.8772 278 1.0392 0.4548 1.0392 1.0194
No log 4.9123 280 1.0773 0.4137 1.0773 1.0379
No log 4.9474 282 0.9704 0.4027 0.9704 0.9851
No log 4.9825 284 0.9150 0.4565 0.9150 0.9566
No log 5.0175 286 0.8419 0.5042 0.8419 0.9176
No log 5.0526 288 0.8339 0.4203 0.8339 0.9132
No log 5.0877 290 0.9246 0.4157 0.9246 0.9616
No log 5.1228 292 1.0892 0.3810 1.0892 1.0436
No log 5.1579 294 1.1060 0.3810 1.1060 1.0517
No log 5.1930 296 0.9850 0.3629 0.9850 0.9925
No log 5.2281 298 0.8856 0.3326 0.8856 0.9411
No log 5.2632 300 0.8598 0.3548 0.8598 0.9272
No log 5.2982 302 0.8406 0.3314 0.8406 0.9168
No log 5.3333 304 0.8787 0.3740 0.8787 0.9374
No log 5.3684 306 0.9613 0.4529 0.9613 0.9804
No log 5.4035 308 1.0720 0.3363 1.0720 1.0354
No log 5.4386 310 1.1104 0.3619 1.1104 1.0537
No log 5.4737 312 1.0730 0.4002 1.0730 1.0359
No log 5.5088 314 1.0020 0.2697 1.0020 1.0010
No log 5.5439 316 0.9934 0.2842 0.9934 0.9967
No log 5.5789 318 0.9827 0.2865 0.9827 0.9913
No log 5.6140 320 0.9726 0.2818 0.9726 0.9862
No log 5.6491 322 1.0258 0.3787 1.0258 1.0128
No log 5.6842 324 1.1675 0.3578 1.1675 1.0805
No log 5.7193 326 1.1925 0.3208 1.1925 1.0920
No log 5.7544 328 1.0501 0.4284 1.0501 1.0247
No log 5.7895 330 0.8972 0.3280 0.8972 0.9472
No log 5.8246 332 0.9080 0.3278 0.9080 0.9529
No log 5.8596 334 0.9150 0.3078 0.9150 0.9566
No log 5.8947 336 0.9492 0.2424 0.9492 0.9743
No log 5.9298 338 1.0932 0.3250 1.0932 1.0456
No log 5.9649 340 1.1563 0.3553 1.1563 1.0753
No log 6.0 342 1.0733 0.3354 1.0733 1.0360
No log 6.0351 344 0.9611 0.3883 0.9611 0.9804
No log 6.0702 346 0.9386 0.3740 0.9386 0.9688
No log 6.1053 348 0.9553 0.3597 0.9553 0.9774
No log 6.1404 350 1.0487 0.3492 1.0487 1.0241
No log 6.1754 352 1.0858 0.3609 1.0858 1.0420
No log 6.2105 354 1.0285 0.4275 1.0285 1.0142
No log 6.2456 356 0.9665 0.3883 0.9665 0.9831
No log 6.2807 358 0.9530 0.3883 0.9530 0.9762
No log 6.3158 360 0.8946 0.2541 0.8946 0.9458
No log 6.3509 362 0.8741 0.2888 0.8741 0.9350
No log 6.3860 364 0.8659 0.3485 0.8659 0.9305
No log 6.4211 366 0.9146 0.4203 0.9146 0.9563
No log 6.4561 368 1.0961 0.4207 1.0961 1.0469
No log 6.4912 370 1.2532 0.3152 1.2532 1.1195
No log 6.5263 372 1.1785 0.3646 1.1785 1.0856
No log 6.5614 374 0.9939 0.3902 0.9939 0.9969
No log 6.5965 376 0.9179 0.3383 0.9179 0.9581
No log 6.6316 378 0.8905 0.2919 0.8905 0.9437
No log 6.6667 380 0.9273 0.3169 0.9273 0.9630
No log 6.7018 382 0.9950 0.2724 0.9950 0.9975
No log 6.7368 384 1.0428 0.1998 1.0428 1.0212
No log 6.7719 386 1.0732 0.1998 1.0732 1.0360
No log 6.8070 388 1.0419 0.2335 1.0419 1.0207
No log 6.8421 390 0.9744 0.3743 0.9744 0.9871
No log 6.8772 392 0.9633 0.3902 0.9633 0.9815
No log 6.9123 394 0.9839 0.4157 0.9839 0.9919
No log 6.9474 396 0.9824 0.4169 0.9824 0.9912
No log 6.9825 398 1.0981 0.4527 1.0981 1.0479
No log 7.0175 400 1.0979 0.3920 1.0979 1.0478
No log 7.0526 402 0.9861 0.3474 0.9861 0.9930
No log 7.0877 404 0.9084 0.3637 0.9084 0.9531
No log 7.1228 406 0.9071 0.3637 0.9071 0.9524
No log 7.1579 408 0.9285 0.3863 0.9285 0.9636
No log 7.1930 410 0.9143 0.3643 0.9143 0.9562
No log 7.2281 412 0.8993 0.3802 0.8993 0.9483
No log 7.2632 414 0.8953 0.3822 0.8953 0.9462
No log 7.2982 416 0.8865 0.3424 0.8865 0.9415
No log 7.3333 418 0.8903 0.3583 0.8903 0.9435
No log 7.3684 420 0.8911 0.2818 0.8911 0.9440
No log 7.4035 422 0.8941 0.2818 0.8941 0.9456
No log 7.4386 424 0.8945 0.2818 0.8945 0.9458
No log 7.4737 426 0.8822 0.3293 0.8822 0.9393
No log 7.5088 428 0.9912 0.3672 0.9912 0.9956
No log 7.5439 430 1.1022 0.4527 1.1022 1.0499
No log 7.5789 432 1.0453 0.4407 1.0453 1.0224
No log 7.6140 434 0.9220 0.3503 0.9220 0.9602
No log 7.6491 436 0.8783 0.3288 0.8783 0.9372
No log 7.6842 438 0.9153 0.2958 0.9153 0.9567
No log 7.7193 440 0.8901 0.2935 0.8901 0.9435
No log 7.7544 442 0.8484 0.3721 0.8484 0.9211
No log 7.7895 444 0.9022 0.3663 0.9022 0.9498
No log 7.8246 446 0.9219 0.4192 0.9219 0.9602
No log 7.8596 448 0.8956 0.3424 0.8956 0.9464
No log 7.8947 450 0.8987 0.3424 0.8987 0.9480
No log 7.9298 452 0.8838 0.3052 0.8838 0.9401
No log 7.9649 454 0.8874 0.3172 0.8874 0.9420
No log 8.0 456 0.9536 0.3285 0.9536 0.9765
No log 8.0351 458 1.0692 0.3954 1.0692 1.0340
No log 8.0702 460 1.2287 0.3478 1.2287 1.1085
No log 8.1053 462 1.1945 0.3185 1.1945 1.0929
No log 8.1404 464 1.0462 0.3103 1.0462 1.0228
No log 8.1754 466 0.9244 0.3250 0.9244 0.9614
No log 8.2105 468 0.8869 0.2988 0.8869 0.9418
No log 8.2456 470 0.8747 0.3293 0.8747 0.9352
No log 8.2807 472 0.8983 0.3543 0.8983 0.9478
No log 8.3158 474 0.9963 0.3483 0.9963 0.9981
No log 8.3509 476 1.0055 0.4036 1.0055 1.0027
No log 8.3860 478 0.9367 0.3103 0.9367 0.9678
No log 8.4211 480 0.9102 0.3782 0.9102 0.9540
No log 8.4561 482 0.8686 0.3543 0.8686 0.9320
No log 8.4912 484 0.8835 0.3782 0.8835 0.9400
No log 8.5263 486 0.9406 0.3483 0.9406 0.9698
No log 8.5614 488 0.9710 0.3483 0.9710 0.9854
No log 8.5965 490 0.9377 0.3483 0.9377 0.9684
No log 8.6316 492 0.8764 0.3304 0.8764 0.9362
No log 8.6667 494 0.8511 0.2667 0.8511 0.9225
No log 8.7018 496 0.8481 0.2794 0.8481 0.9209
No log 8.7368 498 0.8754 0.3250 0.8754 0.9356
0.3101 8.7719 500 0.9326 0.4162 0.9326 0.9657
0.3101 8.8070 502 1.0317 0.3374 1.0317 1.0157
0.3101 8.8421 504 1.1023 0.3492 1.1023 1.0499
0.3101 8.8772 506 1.0614 0.3342 1.0614 1.0302
0.3101 8.9123 508 0.9653 0.2873 0.9653 0.9825
0.3101 8.9474 510 0.8914 0.2770 0.8914 0.9441

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task5_organization

Finetuned
(4019)
this model