ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k3_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8664
  • Qwk: 0.0451
  • Mse: 0.8664
  • Rmse: 0.9308

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 3.6084 0.0042 3.6084 1.8996
No log 0.25 4 1.7272 0.0624 1.7272 1.3142
No log 0.375 6 1.0442 0.0378 1.0442 1.0218
No log 0.5 8 0.8080 -0.0371 0.8080 0.8989
No log 0.625 10 1.1554 0.0300 1.1554 1.0749
No log 0.75 12 1.2243 0.0840 1.2243 1.1065
No log 0.875 14 0.8466 0.0676 0.8466 0.9201
No log 1.0 16 0.9184 0.0182 0.9184 0.9583
No log 1.125 18 0.9028 -0.0143 0.9028 0.9501
No log 1.25 20 0.8016 -0.1244 0.8016 0.8953
No log 1.375 22 0.7395 0.0 0.7395 0.8599
No log 1.5 24 0.8459 -0.0861 0.8459 0.9197
No log 1.625 26 1.0416 0.0196 1.0416 1.0206
No log 1.75 28 0.9237 0.0089 0.9237 0.9611
No log 1.875 30 0.9507 0.0618 0.9507 0.9751
No log 2.0 32 0.8154 0.0456 0.8154 0.9030
No log 2.125 34 0.7651 0.1096 0.7651 0.8747
No log 2.25 36 0.9526 0.0659 0.9526 0.9760
No log 2.375 38 0.8865 0.1065 0.8865 0.9416
No log 2.5 40 0.7541 0.0670 0.7541 0.8684
No log 2.625 42 0.8170 -0.0056 0.8170 0.9039
No log 2.75 44 0.8184 0.1716 0.8184 0.9046
No log 2.875 46 0.7598 0.0909 0.7598 0.8716
No log 3.0 48 0.7426 0.0334 0.7426 0.8617
No log 3.125 50 0.7501 -0.0711 0.7501 0.8661
No log 3.25 52 0.7921 -0.0812 0.7921 0.8900
No log 3.375 54 0.7606 -0.0331 0.7606 0.8721
No log 3.5 56 0.7273 0.0555 0.7273 0.8528
No log 3.625 58 0.8312 -0.0237 0.8312 0.9117
No log 3.75 60 0.8257 0.0135 0.8257 0.9087
No log 3.875 62 1.1314 0.0492 1.1314 1.0637
No log 4.0 64 1.2331 0.0379 1.2331 1.1105
No log 4.125 66 0.9102 0.0639 0.9102 0.9540
No log 4.25 68 0.9205 -0.0257 0.9205 0.9594
No log 4.375 70 1.0121 0.0379 1.0121 1.0060
No log 4.5 72 1.0977 0.0130 1.0977 1.0477
No log 4.625 74 1.0532 -0.0324 1.0532 1.0262
No log 4.75 76 0.9632 0.0311 0.9632 0.9814
No log 4.875 78 0.8921 0.0101 0.8921 0.9445
No log 5.0 80 1.0744 -0.0394 1.0744 1.0365
No log 5.125 82 1.1068 0.0267 1.1068 1.0521
No log 5.25 84 0.9204 0.0330 0.9204 0.9594
No log 5.375 86 1.0512 0.1731 1.0512 1.0253
No log 5.5 88 1.0820 0.1441 1.0820 1.0402
No log 5.625 90 0.9659 0.1044 0.9659 0.9828
No log 5.75 92 0.8549 0.1558 0.8549 0.9246
No log 5.875 94 1.1497 0.0083 1.1497 1.0722
No log 6.0 96 1.4649 0.0712 1.4649 1.2103
No log 6.125 98 1.2468 0.0367 1.2468 1.1166
No log 6.25 100 0.8752 0.1758 0.8752 0.9355
No log 6.375 102 0.7905 0.0376 0.7905 0.8891
No log 6.5 104 0.8495 0.1037 0.8495 0.9217
No log 6.625 106 0.9798 0.1746 0.9798 0.9898
No log 6.75 108 0.9527 0.0769 0.9527 0.9761
No log 6.875 110 0.7801 0.0791 0.7801 0.8833
No log 7.0 112 0.9438 0.0556 0.9438 0.9715
No log 7.125 114 0.9145 0.1065 0.9145 0.9563
No log 7.25 116 0.7392 0.1644 0.7392 0.8597
No log 7.375 118 0.8560 0.0048 0.8560 0.9252
No log 7.5 120 0.9805 -0.0504 0.9805 0.9902
No log 7.625 122 1.0419 -0.0165 1.0419 1.0208
No log 7.75 124 1.0104 -0.0334 1.0104 1.0052
No log 7.875 126 1.0832 0.0245 1.0832 1.0408
No log 8.0 128 1.1474 0.0553 1.1474 1.0711
No log 8.125 130 1.1063 0.1146 1.1063 1.0518
No log 8.25 132 1.0821 0.0130 1.0821 1.0402
No log 8.375 134 0.9657 0.0578 0.9657 0.9827
No log 8.5 136 0.9569 0.0319 0.9569 0.9782
No log 8.625 138 0.9983 0.0061 0.9983 0.9991
No log 8.75 140 0.9643 0.0950 0.9643 0.9820
No log 8.875 142 0.9381 0.1520 0.9381 0.9685
No log 9.0 144 1.0305 -0.0325 1.0305 1.0152
No log 9.125 146 1.0604 0.0378 1.0604 1.0297
No log 9.25 148 0.9971 0.1453 0.9971 0.9986
No log 9.375 150 1.1950 0.0685 1.1950 1.0932
No log 9.5 152 1.2265 0.0442 1.2265 1.1075
No log 9.625 154 1.1290 -0.0122 1.1290 1.0626
No log 9.75 156 0.9774 -0.0071 0.9774 0.9886
No log 9.875 158 0.9620 -0.0115 0.9620 0.9808
No log 10.0 160 0.9780 0.0231 0.9780 0.9889
No log 10.125 162 1.0110 0.0619 1.0110 1.0055
No log 10.25 164 1.0273 0.0707 1.0273 1.0136
No log 10.375 166 1.0695 0.0478 1.0695 1.0341
No log 10.5 168 0.9398 0.0026 0.9398 0.9694
No log 10.625 170 0.8998 0.1269 0.8998 0.9486
No log 10.75 172 0.9569 0.1559 0.9569 0.9782
No log 10.875 174 0.9722 0.1329 0.9722 0.9860
No log 11.0 176 0.8792 0.1546 0.8792 0.9376
No log 11.125 178 0.8950 0.1891 0.8950 0.9460
No log 11.25 180 0.9871 0.0943 0.9871 0.9935
No log 11.375 182 1.0391 0.0985 1.0391 1.0194
No log 11.5 184 0.9239 0.0710 0.9239 0.9612
No log 11.625 186 0.8922 0.0378 0.8922 0.9446
No log 11.75 188 0.9717 0.0142 0.9717 0.9858
No log 11.875 190 0.9061 0.0353 0.9061 0.9519
No log 12.0 192 0.8840 0.1168 0.8840 0.9402
No log 12.125 194 0.9701 0.0945 0.9701 0.9849
No log 12.25 196 1.0599 0.0673 1.0599 1.0295
No log 12.375 198 0.9700 -0.0260 0.9700 0.9849
No log 12.5 200 0.9313 0.0063 0.9313 0.9650
No log 12.625 202 0.9244 0.0366 0.9244 0.9614
No log 12.75 204 0.8582 0.0999 0.8582 0.9264
No log 12.875 206 0.7988 0.1232 0.7988 0.8938
No log 13.0 208 0.8235 0.1277 0.8235 0.9074
No log 13.125 210 0.8799 0.1626 0.8799 0.9380
No log 13.25 212 1.0418 -0.0154 1.0418 1.0207
No log 13.375 214 0.9398 0.0687 0.9398 0.9694
No log 13.5 216 0.7892 0.1387 0.7892 0.8884
No log 13.625 218 0.7795 0.0791 0.7795 0.8829
No log 13.75 220 0.8862 0.0026 0.8862 0.9414
No log 13.875 222 1.1243 0.0468 1.1243 1.0603
No log 14.0 224 1.2887 0.0253 1.2887 1.1352
No log 14.125 226 1.1727 0.0478 1.1727 1.0829
No log 14.25 228 0.9322 0.0183 0.9322 0.9655
No log 14.375 230 0.8554 0.0827 0.8554 0.9249
No log 14.5 232 0.8726 0.0239 0.8726 0.9341
No log 14.625 234 0.9287 -0.0595 0.9287 0.9637
No log 14.75 236 0.9214 0.0048 0.9214 0.9599
No log 14.875 238 0.8217 -0.0277 0.8217 0.9065
No log 15.0 240 0.8151 0.0426 0.8151 0.9028
No log 15.125 242 0.8672 -0.0238 0.8672 0.9312
No log 15.25 244 1.0324 0.0980 1.0324 1.0160
No log 15.375 246 1.1149 0.0784 1.1149 1.0559
No log 15.5 248 0.9749 0.1292 0.9749 0.9874
No log 15.625 250 0.8915 -0.0137 0.8915 0.9442
No log 15.75 252 0.8488 0.0506 0.8488 0.9213
No log 15.875 254 0.8896 0.0970 0.8896 0.9432
No log 16.0 256 0.8302 0.1323 0.8302 0.9112
No log 16.125 258 0.8040 0.1170 0.8040 0.8966
No log 16.25 260 0.8309 0.0717 0.8309 0.9116
No log 16.375 262 0.8609 0.1247 0.8609 0.9278
No log 16.5 264 0.9414 0.0909 0.9414 0.9702
No log 16.625 266 0.9674 0.0402 0.9674 0.9836
No log 16.75 268 0.8812 0.0206 0.8812 0.9387
No log 16.875 270 0.8694 0.0518 0.8694 0.9324
No log 17.0 272 0.9071 0.0206 0.9071 0.9524
No log 17.125 274 1.0391 0.0458 1.0391 1.0194
No log 17.25 276 1.1153 0.0233 1.1153 1.0561
No log 17.375 278 1.0223 0.0169 1.0223 1.0111
No log 17.5 280 0.8745 -0.0055 0.8745 0.9352
No log 17.625 282 0.8027 0.1267 0.8027 0.8959
No log 17.75 284 0.8514 0.0600 0.8514 0.9227
No log 17.875 286 0.9023 -0.0395 0.9023 0.9499
No log 18.0 288 0.8807 0.0627 0.8807 0.9385
No log 18.125 290 0.8478 0.1686 0.8478 0.9208
No log 18.25 292 0.8411 0.1299 0.8411 0.9171
No log 18.375 294 0.8277 0.1630 0.8277 0.9098
No log 18.5 296 0.8377 0.0778 0.8377 0.9153
No log 18.625 298 0.8790 -0.0123 0.8790 0.9375
No log 18.75 300 1.0344 0.0205 1.0344 1.0171
No log 18.875 302 1.1345 -0.0006 1.1345 1.0651
No log 19.0 304 1.0140 -0.0137 1.0140 1.0070
No log 19.125 306 0.9153 -0.0438 0.9153 0.9567
No log 19.25 308 0.8760 0.0049 0.8760 0.9360
No log 19.375 310 0.8960 0.0870 0.8960 0.9466
No log 19.5 312 0.9363 0.0578 0.9363 0.9676
No log 19.625 314 0.9095 0.0897 0.9095 0.9537
No log 19.75 316 0.9267 0.1252 0.9267 0.9627
No log 19.875 318 0.8683 0.1605 0.8683 0.9318
No log 20.0 320 0.7728 0.1287 0.7728 0.8791
No log 20.125 322 0.7867 0.0118 0.7867 0.8870
No log 20.25 324 0.7766 0.0611 0.7766 0.8812
No log 20.375 326 0.7942 0.1687 0.7942 0.8912
No log 20.5 328 0.9543 0.0692 0.9543 0.9769
No log 20.625 330 1.0297 0.0673 1.0297 1.0147
No log 20.75 332 0.9452 0.0268 0.9452 0.9722
No log 20.875 334 0.9359 0.0268 0.9359 0.9674
No log 21.0 336 1.0422 0.1206 1.0422 1.0209
No log 21.125 338 1.0880 0.0169 1.0880 1.0431
No log 21.25 340 1.0212 0.0078 1.0212 1.0105
No log 21.375 342 0.8887 0.0875 0.8887 0.9427
No log 21.5 344 0.8148 0.0341 0.8148 0.9027
No log 21.625 346 0.8126 0.1244 0.8126 0.9014
No log 21.75 348 0.8203 0.0798 0.8203 0.9057
No log 21.875 350 0.8482 0.0469 0.8482 0.9210
No log 22.0 352 0.8230 -0.0391 0.8230 0.9072
No log 22.125 354 0.8215 -0.0391 0.8215 0.9064
No log 22.25 356 0.8452 0.0469 0.8452 0.9194
No log 22.375 358 0.8990 0.0897 0.8990 0.9482
No log 22.5 360 0.8894 0.0550 0.8894 0.9431
No log 22.625 362 0.8182 0.0846 0.8182 0.9045
No log 22.75 364 0.8096 0.1277 0.8096 0.8998
No log 22.875 366 0.8868 0.0286 0.8868 0.9417
No log 23.0 368 0.9467 0.0415 0.9467 0.9730
No log 23.125 370 0.8604 0.0265 0.8604 0.9276
No log 23.25 372 0.8043 0.2057 0.8043 0.8968
No log 23.375 374 0.7725 0.0357 0.7725 0.8789
No log 23.5 376 0.7815 0.0357 0.7815 0.8840
No log 23.625 378 0.7886 0.0357 0.7886 0.8880
No log 23.75 380 0.8479 0.1251 0.8479 0.9208
No log 23.875 382 0.8696 0.0886 0.8696 0.9325
No log 24.0 384 0.8637 0.0883 0.8637 0.9294
No log 24.125 386 0.8323 0.0469 0.8323 0.9123
No log 24.25 388 0.8832 0.0529 0.8832 0.9398
No log 24.375 390 0.9528 -0.0051 0.9528 0.9761
No log 24.5 392 0.9128 0.0893 0.9128 0.9554
No log 24.625 394 0.9099 0.1240 0.9099 0.9539
No log 24.75 396 0.9349 0.0893 0.9349 0.9669
No log 24.875 398 1.0729 0.0445 1.0729 1.0358
No log 25.0 400 1.0558 0.0468 1.0558 1.0275
No log 25.125 402 0.8865 0.0890 0.8865 0.9415
No log 25.25 404 0.7940 0.1646 0.7940 0.8910
No log 25.375 406 0.7723 0.1232 0.7723 0.8788
No log 25.5 408 0.7977 0.0856 0.7977 0.8931
No log 25.625 410 0.9454 0.0138 0.9454 0.9723
No log 25.75 412 1.0602 0.0260 1.0602 1.0297
No log 25.875 414 0.9966 0.0458 0.9966 0.9983
No log 26.0 416 0.8331 0.1660 0.8331 0.9128
No log 26.125 418 0.7658 0.1292 0.7658 0.8751
No log 26.25 420 0.7427 0.1758 0.7427 0.8618
No log 26.375 422 0.7190 0.1758 0.7190 0.8479
No log 26.5 424 0.7024 0.1318 0.7024 0.8381
No log 26.625 426 0.7229 -0.0059 0.7229 0.8503
No log 26.75 428 0.7346 0.0394 0.7346 0.8571
No log 26.875 430 0.7418 0.1249 0.7418 0.8613
No log 27.0 432 0.7843 0.1240 0.7843 0.8856
No log 27.125 434 0.8409 0.1646 0.8409 0.9170
No log 27.25 436 0.9489 0.1860 0.9489 0.9741
No log 27.375 438 0.9922 0.2172 0.9922 0.9961
No log 27.5 440 0.9612 0.2172 0.9612 0.9804
No log 27.625 442 0.9669 0.1544 0.9669 0.9833
No log 27.75 444 0.9208 0.1581 0.9208 0.9596
No log 27.875 446 0.8595 0.1203 0.8595 0.9271
No log 28.0 448 0.8727 0.0886 0.8727 0.9342
No log 28.125 450 0.9741 0.0356 0.9741 0.9870
No log 28.25 452 0.9790 0.0078 0.9790 0.9894
No log 28.375 454 0.8755 -0.0193 0.8755 0.9357
No log 28.5 456 0.8433 0.0049 0.8433 0.9183
No log 28.625 458 0.8685 0.0114 0.8685 0.9319
No log 28.75 460 0.8438 0.0469 0.8438 0.9186
No log 28.875 462 0.8184 -0.0851 0.8184 0.9046
No log 29.0 464 0.8189 -0.0851 0.8189 0.9049
No log 29.125 466 0.8284 0.0 0.8284 0.9102
No log 29.25 468 0.8871 -0.0142 0.8871 0.9419
No log 29.375 470 0.9829 0.1249 0.9829 0.9914
No log 29.5 472 1.0332 0.0401 1.0332 1.0164
No log 29.625 474 0.9273 0.0246 0.9273 0.9630
No log 29.75 476 0.8551 0.0 0.8551 0.9247
No log 29.875 478 0.8585 0.0393 0.8585 0.9265
No log 30.0 480 0.8927 -0.0313 0.8927 0.9448
No log 30.125 482 1.0073 0.0596 1.0073 1.0036
No log 30.25 484 1.0714 0.0922 1.0714 1.0351
No log 30.375 486 1.0045 0.0268 1.0045 1.0022
No log 30.5 488 0.9728 0.0215 0.9728 0.9863
No log 30.625 490 0.9314 0.0532 0.9314 0.9651
No log 30.75 492 0.9381 0.0231 0.9381 0.9685
No log 30.875 494 0.8774 0.0540 0.8774 0.9367
No log 31.0 496 0.8033 0.0423 0.8033 0.8963
No log 31.125 498 0.7697 0.0783 0.7697 0.8773
0.2979 31.25 500 0.7807 0.0791 0.7807 0.8836
0.2979 31.375 502 0.8462 -0.0212 0.8462 0.9199
0.2979 31.5 504 0.9347 -0.0051 0.9347 0.9668
0.2979 31.625 506 1.0144 0.0660 1.0144 1.0072
0.2979 31.75 508 0.9423 0.0560 0.9423 0.9707
0.2979 31.875 510 0.8664 0.0451 0.8664 0.9308

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k3_task3_organization

Finetuned
(4019)
this model