ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k1_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9093
  • Qwk: 0.2781
  • Mse: 0.9093
  • Rmse: 0.9536

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.4 2 2.5845 -0.0729 2.5845 1.6076
No log 0.8 4 1.3911 0.0278 1.3911 1.1794
No log 1.2 6 0.7718 0.0893 0.7718 0.8785
No log 1.6 8 1.0642 0.0922 1.0642 1.0316
No log 2.0 10 1.3836 -0.1679 1.3836 1.1763
No log 2.4 12 1.5122 -0.0630 1.5122 1.2297
No log 2.8 14 1.4156 0.0333 1.4156 1.1898
No log 3.2 16 1.4132 0.0517 1.4132 1.1888
No log 3.6 18 1.3126 0.0776 1.3126 1.1457
No log 4.0 20 0.9229 0.3193 0.9229 0.9607
No log 4.4 22 0.7327 0.3111 0.7327 0.8560
No log 4.8 24 0.8448 0.3940 0.8448 0.9191
No log 5.2 26 0.9854 0.2457 0.9854 0.9927
No log 5.6 28 0.9627 0.3170 0.9627 0.9812
No log 6.0 30 0.8277 0.3169 0.8277 0.9098
No log 6.4 32 1.0722 0.2092 1.0722 1.0355
No log 6.8 34 1.0484 0.2802 1.0484 1.0239
No log 7.2 36 0.8457 0.3219 0.8457 0.9196
No log 7.6 38 0.9807 0.2730 0.9807 0.9903
No log 8.0 40 1.4626 0.1942 1.4626 1.2094
No log 8.4 42 1.4332 0.1838 1.4332 1.1972
No log 8.8 44 1.3776 0.1896 1.3776 1.1737
No log 9.2 46 0.9944 0.2416 0.9944 0.9972
No log 9.6 48 1.0059 0.2988 1.0059 1.0029
No log 10.0 50 1.1640 0.2732 1.1640 1.0789
No log 10.4 52 0.9265 0.2564 0.9265 0.9625
No log 10.8 54 0.8401 0.3425 0.8401 0.9166
No log 11.2 56 0.8935 0.2871 0.8935 0.9452
No log 11.6 58 1.0650 0.2017 1.0650 1.0320
No log 12.0 60 1.2727 0.1984 1.2727 1.1281
No log 12.4 62 1.1810 0.0925 1.1810 1.0868
No log 12.8 64 1.0946 0.0594 1.0946 1.0462
No log 13.2 66 1.3075 0.1086 1.3075 1.1434
No log 13.6 68 1.5251 0.1246 1.5251 1.2349
No log 14.0 70 1.6570 0.1691 1.6570 1.2872
No log 14.4 72 1.3099 0.1388 1.3099 1.1445
No log 14.8 74 1.1125 0.2102 1.1125 1.0547
No log 15.2 76 1.1053 0.2440 1.1053 1.0513
No log 15.6 78 1.1496 0.2421 1.1496 1.0722
No log 16.0 80 1.0754 0.2416 1.0754 1.0370
No log 16.4 82 0.9019 0.2836 0.9019 0.9497
No log 16.8 84 0.8625 0.3194 0.8625 0.9287
No log 17.2 86 0.8365 0.3333 0.8365 0.9146
No log 17.6 88 1.0669 0.3059 1.0669 1.0329
No log 18.0 90 1.0098 0.2613 1.0098 1.0049
No log 18.4 92 0.8199 0.2632 0.8199 0.9055
No log 18.8 94 0.8474 0.2754 0.8474 0.9205
No log 19.2 96 1.0457 0.2602 1.0457 1.0226
No log 19.6 98 1.1493 0.1174 1.1493 1.0720
No log 20.0 100 1.3030 0.1392 1.3030 1.1415
No log 20.4 102 1.2811 0.1805 1.2811 1.1319
No log 20.8 104 1.1682 0.1739 1.1682 1.0809
No log 21.2 106 0.9032 0.2193 0.9032 0.9504
No log 21.6 108 0.9325 0.2702 0.9325 0.9657
No log 22.0 110 1.1082 0.2864 1.1082 1.0527
No log 22.4 112 1.0312 0.3357 1.0312 1.0155
No log 22.8 114 0.9672 0.2971 0.9672 0.9835
No log 23.2 116 0.9380 0.2437 0.9380 0.9685
No log 23.6 118 1.0198 0.1827 1.0198 1.0099
No log 24.0 120 1.0866 0.1747 1.0866 1.0424
No log 24.4 122 1.2614 0.1641 1.2614 1.1231
No log 24.8 124 1.0970 0.1626 1.0970 1.0474
No log 25.2 126 0.8937 0.2012 0.8937 0.9454
No log 25.6 128 0.8871 0.2063 0.8871 0.9418
No log 26.0 130 0.9994 0.1955 0.9994 0.9997
No log 26.4 132 0.9881 0.1692 0.9881 0.9940
No log 26.8 134 0.9604 0.1867 0.9604 0.9800
No log 27.2 136 1.0508 0.1955 1.0508 1.0251
No log 27.6 138 1.1745 0.1729 1.1745 1.0838
No log 28.0 140 1.1007 0.1775 1.1007 1.0491
No log 28.4 142 0.9878 0.1734 0.9878 0.9939
No log 28.8 144 1.0333 0.1911 1.0333 1.0165
No log 29.2 146 1.2442 0.1531 1.2442 1.1154
No log 29.6 148 1.3989 0.1823 1.3989 1.1827
No log 30.0 150 1.3543 0.1823 1.3543 1.1638
No log 30.4 152 1.1461 0.1484 1.1461 1.0706
No log 30.8 154 0.9666 0.1612 0.9666 0.9832
No log 31.2 156 0.9907 0.1955 0.9907 0.9954
No log 31.6 158 1.2121 0.2191 1.2121 1.1009
No log 32.0 160 1.5000 0.1829 1.5000 1.2247
No log 32.4 162 1.4881 0.1829 1.4881 1.2199
No log 32.8 164 1.2200 0.1260 1.2200 1.1046
No log 33.2 166 0.9945 0.2193 0.9945 0.9972
No log 33.6 168 0.9867 0.2193 0.9867 0.9933
No log 34.0 170 1.0754 0.1312 1.0754 1.0370
No log 34.4 172 1.2904 0.0419 1.2904 1.1360
No log 34.8 174 1.3140 0.0633 1.3140 1.1463
No log 35.2 176 1.2099 0.0654 1.2099 1.1000
No log 35.6 178 1.0113 0.2358 1.0113 1.0057
No log 36.0 180 0.9907 0.2410 0.9907 0.9953
No log 36.4 182 1.0768 0.2554 1.0768 1.0377
No log 36.8 184 1.0243 0.2554 1.0243 1.0121
No log 37.2 186 0.8966 0.2781 0.8966 0.9469
No log 37.6 188 0.9275 0.2975 0.9275 0.9631
No log 38.0 190 1.0305 0.2881 1.0305 1.0151
No log 38.4 192 1.0883 0.2070 1.0883 1.0432
No log 38.8 194 0.9808 0.2728 0.9808 0.9904
No log 39.2 196 0.9057 0.2308 0.9057 0.9517
No log 39.6 198 0.9386 0.2308 0.9386 0.9688
No log 40.0 200 1.0475 0.2070 1.0475 1.0235
No log 40.4 202 1.1581 0.2499 1.1581 1.0761
No log 40.8 204 1.1079 0.2481 1.1079 1.0526
No log 41.2 206 0.9163 0.2259 0.9163 0.9572
No log 41.6 208 0.7841 0.3032 0.7841 0.8855
No log 42.0 210 0.7628 0.3032 0.7628 0.8734
No log 42.4 212 0.7785 0.2632 0.7785 0.8823
No log 42.8 214 0.8864 0.2463 0.8864 0.9415
No log 43.2 216 1.0029 0.3022 1.0029 1.0015
No log 43.6 218 1.1403 0.3154 1.1403 1.0679
No log 44.0 220 1.1097 0.2956 1.1097 1.0534
No log 44.4 222 0.9608 0.2905 0.9608 0.9802
No log 44.8 224 0.9334 0.2513 0.9334 0.9661
No log 45.2 226 0.9643 0.2046 0.9643 0.9820
No log 45.6 228 1.0199 0.2109 1.0199 1.0099
No log 46.0 230 1.1253 0.1453 1.1253 1.0608
No log 46.4 232 1.0793 0.1774 1.0793 1.0389
No log 46.8 234 1.0102 0.2297 1.0102 1.0051
No log 47.2 236 0.9826 0.2437 0.9826 0.9913
No log 47.6 238 0.9632 0.2336 0.9632 0.9814
No log 48.0 240 0.9582 0.2336 0.9582 0.9789
No log 48.4 242 0.9115 0.2336 0.9115 0.9547
No log 48.8 244 0.9452 0.2513 0.9452 0.9722
No log 49.2 246 0.9616 0.2754 0.9616 0.9806
No log 49.6 248 0.8858 0.3131 0.8858 0.9412
No log 50.0 250 0.8852 0.3069 0.8852 0.9409
No log 50.4 252 0.9542 0.2754 0.9542 0.9768
No log 50.8 254 0.9574 0.2416 0.9574 0.9785
No log 51.2 256 0.9516 0.2589 0.9516 0.9755
No log 51.6 258 0.9890 0.2988 0.9890 0.9945
No log 52.0 260 0.9991 0.2487 0.9991 0.9996
No log 52.4 262 1.0509 0.1699 1.0509 1.0251
No log 52.8 264 1.0176 0.2651 1.0176 1.0087
No log 53.2 266 0.9419 0.2754 0.9419 0.9705
No log 53.6 268 0.9876 0.1734 0.9876 0.9938
No log 54.0 270 1.1165 0.1774 1.1165 1.0567
No log 54.4 272 1.2048 0.1839 1.2048 1.0976
No log 54.8 274 1.2297 0.1995 1.2297 1.1089
No log 55.2 276 1.0883 0.1821 1.0883 1.0432
No log 55.6 278 0.8854 0.2518 0.8854 0.9410
No log 56.0 280 0.7951 0.3287 0.7951 0.8917
No log 56.4 282 0.7730 0.3355 0.7730 0.8792
No log 56.8 284 0.7985 0.3329 0.7985 0.8936
No log 57.2 286 0.8875 0.2518 0.8875 0.9421
No log 57.6 288 1.0315 0.2271 1.0315 1.0156
No log 58.0 290 1.1540 0.1422 1.1540 1.0743
No log 58.4 292 1.1981 0.1995 1.1981 1.0946
No log 58.8 294 1.1329 0.1839 1.1329 1.0644
No log 59.2 296 0.9768 0.2651 0.9768 0.9883
No log 59.6 298 0.8890 0.2892 0.8890 0.9429
No log 60.0 300 0.8655 0.2949 0.8655 0.9303
No log 60.4 302 0.8776 0.3076 0.8776 0.9368
No log 60.8 304 0.9848 0.2125 0.9848 0.9924
No log 61.2 306 1.1677 0.1638 1.1677 1.0806
No log 61.6 308 1.2829 0.1928 1.2829 1.1326
No log 62.0 310 1.2477 0.1928 1.2477 1.1170
No log 62.4 312 1.1009 0.2207 1.1009 1.0492
No log 62.8 314 0.9234 0.3251 0.9234 0.9609
No log 63.2 316 0.7674 0.3544 0.7674 0.8760
No log 63.6 318 0.7164 0.3713 0.7164 0.8464
No log 64.0 320 0.7134 0.3788 0.7134 0.8447
No log 64.4 322 0.7431 0.3817 0.7431 0.8620
No log 64.8 324 0.8329 0.2917 0.8329 0.9126
No log 65.2 326 0.9450 0.2934 0.9450 0.9721
No log 65.6 328 0.9966 0.2209 0.9966 0.9983
No log 66.0 330 0.9693 0.2343 0.9693 0.9845
No log 66.4 332 0.9041 0.2702 0.9041 0.9508
No log 66.8 334 0.8340 0.3069 0.8340 0.9132
No log 67.2 336 0.8189 0.3069 0.8189 0.9049
No log 67.6 338 0.8526 0.3319 0.8526 0.9234
No log 68.0 340 0.9406 0.2677 0.9406 0.9698
No log 68.4 342 1.0553 0.1557 1.0553 1.0273
No log 68.8 344 1.1069 0.1671 1.1069 1.0521
No log 69.2 346 1.0809 0.1557 1.0809 1.0397
No log 69.6 348 1.0149 0.1896 1.0149 1.0074
No log 70.0 350 0.9517 0.2343 0.9517 0.9755
No log 70.4 352 0.9498 0.2059 0.9498 0.9746
No log 70.8 354 0.9972 0.1935 0.9972 0.9986
No log 71.2 356 1.0572 0.1821 1.0572 1.0282
No log 71.6 358 1.0815 0.1784 1.0815 1.0400
No log 72.0 360 1.0975 0.1784 1.0975 1.0476
No log 72.4 362 1.0691 0.1557 1.0691 1.0340
No log 72.8 364 1.0226 0.1858 1.0226 1.0113
No log 73.2 366 0.9720 0.2343 0.9720 0.9859
No log 73.6 368 0.9160 0.2836 0.9160 0.9571
No log 74.0 370 0.8981 0.2410 0.8981 0.9477
No log 74.4 372 0.9002 0.2410 0.9002 0.9488
No log 74.8 374 0.9446 0.2147 0.9446 0.9719
No log 75.2 376 1.0217 0.1896 1.0217 1.0108
No log 75.6 378 1.0608 0.1557 1.0608 1.0300
No log 76.0 380 1.0411 0.1858 1.0411 1.0204
No log 76.4 382 0.9707 0.2577 0.9707 0.9852
No log 76.8 384 0.9232 0.2343 0.9232 0.9609
No log 77.2 386 0.9263 0.2343 0.9263 0.9624
No log 77.6 388 0.9301 0.2343 0.9301 0.9644
No log 78.0 390 0.9742 0.2577 0.9742 0.9870
No log 78.4 392 1.0656 0.2635 1.0656 1.0323
No log 78.8 394 1.1215 0.1993 1.1215 1.0590
No log 79.2 396 1.1143 0.1993 1.1143 1.0556
No log 79.6 398 1.0665 0.2635 1.0665 1.0327
No log 80.0 400 0.9867 0.2343 0.9867 0.9933
No log 80.4 402 0.9285 0.2102 0.9285 0.9636
No log 80.8 404 0.8729 0.2904 0.8729 0.9343
No log 81.2 406 0.8362 0.3195 0.8362 0.9144
No log 81.6 408 0.8392 0.3195 0.8392 0.9161
No log 82.0 410 0.8619 0.2812 0.8619 0.9284
No log 82.4 412 0.9010 0.2781 0.9010 0.9492
No log 82.8 414 0.9603 0.2437 0.9603 0.9799
No log 83.2 416 1.0129 0.2297 1.0129 1.0064
No log 83.6 418 1.0272 0.1976 1.0272 1.0135
No log 84.0 420 1.0169 0.1976 1.0169 1.0084
No log 84.4 422 0.9967 0.2102 0.9967 0.9984
No log 84.8 424 0.9624 0.2437 0.9624 0.9810
No log 85.2 426 0.9206 0.3043 0.9206 0.9595
No log 85.6 428 0.9143 0.3043 0.9143 0.9562
No log 86.0 430 0.9170 0.2702 0.9170 0.9576
No log 86.4 432 0.9284 0.2702 0.9284 0.9635
No log 86.8 434 0.9343 0.2702 0.9343 0.9666
No log 87.2 436 0.9549 0.2702 0.9549 0.9772
No log 87.6 438 0.9723 0.2702 0.9723 0.9860
No log 88.0 440 0.9907 0.2437 0.9907 0.9953
No log 88.4 442 0.9991 0.2677 0.9991 0.9996
No log 88.8 444 0.9935 0.2677 0.9935 0.9967
No log 89.2 446 0.9641 0.2437 0.9641 0.9819
No log 89.6 448 0.9316 0.2437 0.9316 0.9652
No log 90.0 450 0.9140 0.2781 0.9140 0.9560
No log 90.4 452 0.9173 0.3043 0.9173 0.9578
No log 90.8 454 0.9197 0.2702 0.9197 0.9590
No log 91.2 456 0.9300 0.2702 0.9300 0.9644
No log 91.6 458 0.9325 0.2702 0.9325 0.9657
No log 92.0 460 0.9328 0.2702 0.9328 0.9658
No log 92.4 462 0.9283 0.2702 0.9283 0.9635
No log 92.8 464 0.9270 0.2702 0.9270 0.9628
No log 93.2 466 0.9192 0.2702 0.9192 0.9588
No log 93.6 468 0.9157 0.2702 0.9157 0.9569
No log 94.0 470 0.9170 0.2702 0.9170 0.9576
No log 94.4 472 0.9201 0.2702 0.9201 0.9592
No log 94.8 474 0.9223 0.2702 0.9223 0.9604
No log 95.2 476 0.9191 0.3043 0.9191 0.9587
No log 95.6 478 0.9120 0.3043 0.9120 0.9550
No log 96.0 480 0.9035 0.3043 0.9035 0.9505
No log 96.4 482 0.8974 0.3043 0.8974 0.9473
No log 96.8 484 0.8955 0.3043 0.8955 0.9463
No log 97.2 486 0.8958 0.2781 0.8958 0.9465
No log 97.6 488 0.8955 0.2781 0.8955 0.9463
No log 98.0 490 0.8982 0.2781 0.8982 0.9477
No log 98.4 492 0.9017 0.2781 0.9017 0.9496
No log 98.8 494 0.9042 0.2781 0.9042 0.9509
No log 99.2 496 0.9069 0.2781 0.9069 0.9523
No log 99.6 498 0.9087 0.2781 0.9087 0.9533
0.1632 100.0 500 0.9093 0.2781 0.9093 0.9536

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k1_task7_organization

Finetuned
(4019)
this model