ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8779
  • Qwk: 0.1577
  • Mse: 0.8779
  • Rmse: 0.9370

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.025 2 2.6354 -0.0593 2.6354 1.6234
No log 0.05 4 1.3927 0.1259 1.3927 1.1801
No log 0.075 6 1.0319 -0.1866 1.0319 1.0158
No log 0.1 8 1.0227 0.1179 1.0227 1.0113
No log 0.125 10 1.0801 0.0108 1.0801 1.0393
No log 0.15 12 1.2865 0.0838 1.2865 1.1343
No log 0.175 14 1.0905 0.1840 1.0905 1.0443
No log 0.2 16 0.9768 0.1100 0.9768 0.9883
No log 0.225 18 0.9530 0.1479 0.9530 0.9762
No log 0.25 20 0.9492 -0.0487 0.9492 0.9743
No log 0.275 22 0.9533 0.0441 0.9533 0.9764
No log 0.3 24 0.9146 -0.0491 0.9146 0.9564
No log 0.325 26 0.9027 -0.0475 0.9027 0.9501
No log 0.35 28 0.9146 -0.0259 0.9146 0.9563
No log 0.375 30 0.8965 -0.0117 0.8965 0.9468
No log 0.4 32 0.9979 0.0570 0.9979 0.9990
No log 0.425 34 1.0554 0.1323 1.0554 1.0273
No log 0.45 36 0.9215 0.1379 0.9215 0.9600
No log 0.475 38 1.1193 0.0609 1.1193 1.0580
No log 0.5 40 1.0266 0.1110 1.0266 1.0132
No log 0.525 42 0.8579 0.1091 0.8579 0.9262
No log 0.55 44 1.0485 0.2233 1.0485 1.0239
No log 0.575 46 1.1711 -0.0104 1.1711 1.0822
No log 0.6 48 1.0191 0.2584 1.0191 1.0095
No log 0.625 50 0.8159 0.1529 0.8159 0.9033
No log 0.65 52 0.7881 0.1407 0.7881 0.8877
No log 0.675 54 0.8215 0.2285 0.8215 0.9064
No log 0.7 56 0.8112 0.2027 0.8112 0.9007
No log 0.725 58 0.7676 0.0327 0.7676 0.8761
No log 0.75 60 0.7826 0.0 0.7826 0.8847
No log 0.775 62 0.8481 0.0509 0.8481 0.9209
No log 0.8 64 0.9091 0.0535 0.9091 0.9535
No log 0.825 66 0.9455 0.1697 0.9455 0.9724
No log 0.85 68 0.8838 0.0827 0.8838 0.9401
No log 0.875 70 0.8740 0.1298 0.8740 0.9349
No log 0.9 72 0.9040 0.1331 0.9040 0.9508
No log 0.925 74 0.9333 0.2303 0.9333 0.9661
No log 0.95 76 0.9567 0.2735 0.9567 0.9781
No log 0.975 78 0.8984 0.2273 0.8984 0.9479
No log 1.0 80 0.8751 0.2252 0.8751 0.9355
No log 1.025 82 0.8373 0.2535 0.8373 0.9150
No log 1.05 84 0.8305 0.2593 0.8305 0.9113
No log 1.075 86 0.8334 0.2334 0.8334 0.9129
No log 1.1 88 0.8941 0.2414 0.8941 0.9456
No log 1.125 90 0.9154 0.2419 0.9154 0.9568
No log 1.15 92 0.8555 0.2456 0.8555 0.9249
No log 1.175 94 0.8483 0.1835 0.8483 0.9211
No log 1.2 96 0.8647 0.1786 0.8647 0.9299
No log 1.225 98 0.9126 0.1650 0.9126 0.9553
No log 1.25 100 0.9803 0.2267 0.9803 0.9901
No log 1.275 102 1.0802 0.1689 1.0802 1.0393
No log 1.3 104 1.1722 0.2094 1.1722 1.0827
No log 1.325 106 1.1394 0.2249 1.1394 1.0674
No log 1.35 108 1.1978 0.0302 1.1978 1.0945
No log 1.375 110 1.2800 0.0453 1.2800 1.1314
No log 1.4 112 1.1290 0.1364 1.1290 1.0625
No log 1.425 114 1.1021 0.2256 1.1021 1.0498
No log 1.45 116 0.9918 0.1825 0.9918 0.9959
No log 1.475 118 0.9618 0.0985 0.9618 0.9807
No log 1.5 120 0.9665 0.1346 0.9665 0.9831
No log 1.525 122 0.9953 0.1439 0.9953 0.9977
No log 1.55 124 0.9809 0.1433 0.9809 0.9904
No log 1.575 126 0.8766 0.1935 0.8766 0.9363
No log 1.6 128 0.8651 0.2606 0.8651 0.9301
No log 1.625 130 0.8443 0.2114 0.8443 0.9189
No log 1.65 132 0.9908 0.1180 0.9908 0.9954
No log 1.675 134 1.2228 0.0967 1.2228 1.1058
No log 1.7 136 1.1055 0.1173 1.1055 1.0514
No log 1.725 138 0.9238 0.2434 0.9238 0.9611
No log 1.75 140 0.9037 0.2291 0.9037 0.9506
No log 1.775 142 1.1050 0.1370 1.1050 1.0512
No log 1.8 144 1.0487 0.1570 1.0487 1.0241
No log 1.825 146 0.9543 0.2078 0.9543 0.9769
No log 1.85 148 1.1494 0.1141 1.1494 1.0721
No log 1.875 150 1.1234 0.1439 1.1234 1.0599
No log 1.9 152 0.9059 0.2273 0.9059 0.9518
No log 1.925 154 0.8829 0.2342 0.8829 0.9396
No log 1.95 156 0.9205 0.2605 0.9205 0.9594
No log 1.975 158 0.8980 0.3106 0.8980 0.9476
No log 2.0 160 1.0110 0.1793 1.0110 1.0055
No log 2.025 162 1.0071 0.1458 1.0071 1.0035
No log 2.05 164 0.9563 0.2342 0.9563 0.9779
No log 2.075 166 1.0332 0.1710 1.0332 1.0165
No log 2.1 168 0.9734 0.2342 0.9734 0.9866
No log 2.125 170 0.9732 0.1965 0.9732 0.9865
No log 2.15 172 1.0240 0.1489 1.0240 1.0119
No log 2.175 174 0.9394 0.2222 0.9394 0.9692
No log 2.2 176 0.9596 0.2633 0.9596 0.9796
No log 2.225 178 0.9807 0.0639 0.9807 0.9903
No log 2.25 180 0.9597 0.3007 0.9597 0.9796
No log 2.275 182 0.9694 0.1531 0.9694 0.9846
No log 2.3 184 0.9823 0.1600 0.9823 0.9911
No log 2.325 186 0.9618 0.1124 0.9618 0.9807
No log 2.35 188 0.8914 0.1404 0.8914 0.9441
No log 2.375 190 0.8518 0.1871 0.8518 0.9229
No log 2.4 192 0.8810 0.2342 0.8810 0.9386
No log 2.425 194 0.9647 0.0741 0.9647 0.9822
No log 2.45 196 0.9595 0.1582 0.9595 0.9795
No log 2.475 198 1.0330 0.1652 1.0330 1.0164
No log 2.5 200 1.3227 0.1715 1.3227 1.1501
No log 2.525 202 1.3620 0.1486 1.3620 1.1671
No log 2.55 204 1.1061 0.1900 1.1061 1.0517
No log 2.575 206 1.1012 0.1266 1.1012 1.0494
No log 2.6 208 1.1154 0.1619 1.1154 1.0561
No log 2.625 210 1.0741 0.0835 1.0741 1.0364
No log 2.65 212 1.0452 0.1589 1.0452 1.0224
No log 2.675 214 0.9565 0.0583 0.9565 0.9780
No log 2.7 216 0.9890 0.1379 0.9890 0.9945
No log 2.725 218 1.0053 0.1379 1.0053 1.0026
No log 2.75 220 0.9957 0.0792 0.9957 0.9978
No log 2.775 222 1.2053 0.1696 1.2053 1.0979
No log 2.8 224 1.2202 0.1704 1.2202 1.1046
No log 2.825 226 0.9867 0.1612 0.9866 0.9933
No log 2.85 228 0.7981 0.2819 0.7981 0.8934
No log 2.875 230 0.7756 0.3552 0.7756 0.8807
No log 2.9 232 0.7875 0.3552 0.7875 0.8874
No log 2.925 234 0.8217 0.2445 0.8217 0.9065
No log 2.95 236 0.9916 0.1680 0.9916 0.9958
No log 2.975 238 1.0659 0.1860 1.0659 1.0324
No log 3.0 240 0.9399 0.1623 0.9399 0.9695
No log 3.025 242 0.8601 0.2936 0.8601 0.9274
No log 3.05 244 0.9130 0.2796 0.9130 0.9555
No log 3.075 246 0.9729 0.1622 0.9729 0.9864
No log 3.1 248 0.9667 0.2345 0.9667 0.9832
No log 3.125 250 1.0234 0.2551 1.0234 1.0116
No log 3.15 252 1.0913 0.1394 1.0913 1.0447
No log 3.175 254 1.0315 0.2526 1.0315 1.0156
No log 3.2 256 0.9759 0.2526 0.9759 0.9879
No log 3.225 258 0.8984 0.2564 0.8984 0.9479
No log 3.25 260 0.8635 0.2342 0.8635 0.9293
No log 3.275 262 0.8173 0.2392 0.8173 0.9041
No log 3.3 264 0.8486 0.2414 0.8486 0.9212
No log 3.325 266 0.8684 0.2253 0.8684 0.9319
No log 3.35 268 0.8167 0.2237 0.8167 0.9037
No log 3.375 270 0.8191 0.3129 0.8191 0.9050
No log 3.4 272 0.8957 0.2434 0.8957 0.9464
No log 3.425 274 0.8676 0.2318 0.8676 0.9315
No log 3.45 276 0.9201 0.2808 0.9201 0.9592
No log 3.475 278 1.1514 0.1909 1.1514 1.0730
No log 3.5 280 1.1356 0.2041 1.1356 1.0657
No log 3.525 282 1.0499 0.1603 1.0499 1.0246
No log 3.55 284 0.9585 0.1416 0.9585 0.9790
No log 3.575 286 0.8940 0.1522 0.8940 0.9455
No log 3.6 288 0.9267 0.1683 0.9267 0.9627
No log 3.625 290 0.9888 0.0863 0.9888 0.9944
No log 3.65 292 1.0616 0.1412 1.0616 1.0303
No log 3.675 294 1.0525 0.1680 1.0525 1.0259
No log 3.7 296 0.9389 0.1642 0.9389 0.9690
No log 3.725 298 0.7743 0.2751 0.7743 0.8799
No log 3.75 300 0.7404 0.3939 0.7404 0.8605
No log 3.775 302 0.7366 0.3939 0.7366 0.8583
No log 3.8 304 0.7460 0.3667 0.7460 0.8637
No log 3.825 306 0.7970 0.3068 0.7970 0.8928
No log 3.85 308 0.8393 0.3052 0.8393 0.9162
No log 3.875 310 0.9165 0.3664 0.9165 0.9573
No log 3.9 312 0.9575 0.3168 0.9575 0.9785
No log 3.925 314 0.9057 0.2254 0.9057 0.9517
No log 3.95 316 0.8612 0.2773 0.8612 0.9280
No log 3.975 318 0.9909 0.2777 0.9909 0.9955
No log 4.0 320 1.0716 0.2134 1.0716 1.0352
No log 4.025 322 0.9857 0.2129 0.9857 0.9928
No log 4.05 324 0.8167 0.2577 0.8167 0.9037
No log 4.075 326 0.8296 0.2522 0.8296 0.9108
No log 4.1 328 0.8557 0.2469 0.8557 0.9250
No log 4.125 330 0.8142 0.2516 0.8142 0.9023
No log 4.15 332 0.8289 0.2577 0.8289 0.9104
No log 4.175 334 0.8915 0.1839 0.8915 0.9442
No log 4.2 336 0.8407 0.1607 0.8407 0.9169
No log 4.225 338 0.7802 0.3111 0.7802 0.8833
No log 4.25 340 0.8230 0.1941 0.8230 0.9072
No log 4.275 342 0.8540 0.1537 0.8540 0.9241
No log 4.3 344 0.8292 0.2837 0.8292 0.9106
No log 4.325 346 0.8212 0.2901 0.8212 0.9062
No log 4.35 348 0.9369 0.2156 0.9369 0.9679
No log 4.375 350 1.0151 0.1707 1.0151 1.0075
No log 4.4 352 0.9483 0.1831 0.9483 0.9738
No log 4.425 354 0.8431 0.2604 0.8431 0.9182
No log 4.45 356 0.7848 0.3352 0.7848 0.8859
No log 4.475 358 0.8066 0.2872 0.8066 0.8981
No log 4.5 360 0.8132 0.2434 0.8132 0.9018
No log 4.525 362 0.8152 0.3006 0.8152 0.9029
No log 4.55 364 0.7846 0.3961 0.7846 0.8858
No log 4.575 366 0.8333 0.2808 0.8333 0.9128
No log 4.6 368 0.8442 0.2968 0.8442 0.9188
No log 4.625 370 0.7851 0.3011 0.7851 0.8860
No log 4.65 372 0.7769 0.3409 0.7769 0.8814
No log 4.675 374 0.8173 0.2652 0.8173 0.9040
No log 4.7 376 0.8010 0.2679 0.8010 0.8950
No log 4.725 378 0.7701 0.3762 0.7701 0.8775
No log 4.75 380 0.8517 0.1940 0.8517 0.9229
No log 4.775 382 0.9654 0.2343 0.9654 0.9825
No log 4.8 384 0.9273 0.1912 0.9273 0.9630
No log 4.825 386 0.8212 0.2694 0.8212 0.9062
No log 4.85 388 0.8252 0.3616 0.8252 0.9084
No log 4.875 390 0.8550 0.3255 0.8550 0.9246
No log 4.9 392 0.8849 0.3454 0.8849 0.9407
No log 4.925 394 0.8980 0.3552 0.8980 0.9476
No log 4.95 396 0.9077 0.3176 0.9077 0.9528
No log 4.975 398 0.8920 0.3494 0.8920 0.9445
No log 5.0 400 0.8477 0.3616 0.8477 0.9207
No log 5.025 402 0.8237 0.3126 0.8237 0.9076
No log 5.05 404 0.7956 0.3691 0.7956 0.8920
No log 5.075 406 0.7843 0.3556 0.7843 0.8856
No log 5.1 408 0.7744 0.3934 0.7744 0.8800
No log 5.125 410 0.7871 0.3347 0.7871 0.8872
No log 5.15 412 0.7775 0.2777 0.7775 0.8817
No log 5.175 414 0.7943 0.3106 0.7943 0.8912
No log 5.2 416 0.8123 0.3183 0.8123 0.9013
No log 5.225 418 0.8036 0.2809 0.8036 0.8964
No log 5.25 420 0.7996 0.3171 0.7996 0.8942
No log 5.275 422 0.8109 0.3106 0.8109 0.9005
No log 5.3 424 0.8044 0.3106 0.8044 0.8969
No log 5.325 426 0.7878 0.3198 0.7878 0.8876
No log 5.35 428 0.8034 0.3391 0.8034 0.8964
No log 5.375 430 0.7913 0.3391 0.7913 0.8895
No log 5.4 432 0.7815 0.4179 0.7815 0.8840
No log 5.425 434 0.8200 0.3189 0.8200 0.9055
No log 5.45 436 0.8038 0.3739 0.8038 0.8965
No log 5.475 438 0.7513 0.4105 0.7513 0.8668
No log 5.5 440 0.8180 0.2370 0.8180 0.9045
No log 5.525 442 0.9311 0.2964 0.9311 0.9649
No log 5.55 444 0.8707 0.3688 0.8707 0.9331
No log 5.575 446 0.7381 0.3314 0.7381 0.8592
No log 5.6 448 0.6941 0.3811 0.6941 0.8331
No log 5.625 450 0.7005 0.3811 0.7005 0.8369
No log 5.65 452 0.7129 0.3984 0.7129 0.8443
No log 5.675 454 0.7413 0.2888 0.7413 0.8610
No log 5.7 456 0.7719 0.3393 0.7719 0.8786
No log 5.725 458 0.8274 0.3207 0.8274 0.9096
No log 5.75 460 0.9092 0.2513 0.9092 0.9535
No log 5.775 462 1.0045 0.1591 1.0045 1.0023
No log 5.8 464 1.0252 0.1591 1.0252 1.0125
No log 5.825 466 0.9489 0.1594 0.9489 0.9741
No log 5.85 468 0.8934 0.2643 0.8934 0.9452
No log 5.875 470 0.8831 0.3085 0.8831 0.9397
No log 5.9 472 0.8820 0.2551 0.8820 0.9391
No log 5.925 474 0.8962 0.2053 0.8962 0.9467
No log 5.95 476 0.9257 0.2551 0.9257 0.9621
No log 5.975 478 0.9298 0.2551 0.9298 0.9643
No log 6.0 480 0.9275 0.1704 0.9275 0.9631
No log 6.025 482 0.9178 0.1464 0.9178 0.9580
No log 6.05 484 0.9084 0.1952 0.9084 0.9531
No log 6.075 486 0.8811 0.0583 0.8811 0.9387
No log 6.1 488 0.8960 0.2359 0.8960 0.9466
No log 6.125 490 0.9514 0.1682 0.9514 0.9754
No log 6.15 492 0.8862 0.2173 0.8862 0.9414
No log 6.175 494 0.8288 0.1264 0.8288 0.9104
No log 6.2 496 0.8067 0.1341 0.8067 0.8982
No log 6.225 498 0.8098 0.1341 0.8098 0.8999
0.3561 6.25 500 0.8428 0.2247 0.8428 0.9180
0.3561 6.275 502 0.9144 0.1839 0.9144 0.9563
0.3561 6.3 504 0.9141 0.1362 0.9141 0.9561
0.3561 6.325 506 0.8879 0.2193 0.8879 0.9423
0.3561 6.35 508 0.8856 0.1531 0.8856 0.9411
0.3561 6.375 510 0.8779 0.1577 0.8779 0.9370

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task7_organization

Finetuned
(4019)
this model