ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k17_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8346
  • Qwk: 0.1047
  • Mse: 0.8346
  • Rmse: 0.9135

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0227 2 3.5405 -0.0154 3.5405 1.8816
No log 0.0455 4 2.0321 -0.0284 2.0321 1.4255
No log 0.0682 6 0.9992 -0.0695 0.9992 0.9996
No log 0.0909 8 1.2121 -0.0149 1.2121 1.1010
No log 0.1136 10 0.7817 0.0670 0.7817 0.8842
No log 0.1364 12 0.6893 0.0 0.6893 0.8302
No log 0.1591 14 0.7224 0.0 0.7224 0.8500
No log 0.1818 16 0.6951 0.0 0.6951 0.8337
No log 0.2045 18 0.7325 0.0460 0.7325 0.8558
No log 0.2273 20 1.1123 -0.0457 1.1123 1.0547
No log 0.25 22 1.6146 -0.0041 1.6146 1.2707
No log 0.2727 24 1.3684 0.0247 1.3684 1.1698
No log 0.2955 26 1.0104 0.0810 1.0104 1.0052
No log 0.3182 28 0.9554 0.0618 0.9554 0.9775
No log 0.3409 30 0.8899 0.1106 0.8899 0.9434
No log 0.3636 32 1.0248 0.0152 1.0248 1.0123
No log 0.3864 34 0.7685 -0.0287 0.7685 0.8766
No log 0.4091 36 0.7157 0.0555 0.7157 0.8460
No log 0.4318 38 0.7842 0.0909 0.7842 0.8856
No log 0.4545 40 1.0608 -0.0097 1.0608 1.0300
No log 0.4773 42 1.3557 0.0923 1.3557 1.1643
No log 0.5 44 1.0553 -0.0049 1.0553 1.0273
No log 0.5227 46 0.8956 0.0968 0.8956 0.9464
No log 0.5455 48 0.8763 -0.0306 0.8763 0.9361
No log 0.5682 50 0.8479 0.0406 0.8479 0.9208
No log 0.5909 52 0.7391 0.0983 0.7391 0.8597
No log 0.6136 54 0.7923 0.0826 0.7923 0.8901
No log 0.6364 56 0.8284 0.0711 0.8284 0.9102
No log 0.6591 58 0.7134 0.1787 0.7134 0.8446
No log 0.6818 60 0.7541 0.1996 0.7541 0.8684
No log 0.7045 62 0.7895 0.1996 0.7895 0.8886
No log 0.7273 64 1.0386 0.0482 1.0386 1.0191
No log 0.75 66 1.3014 0.1003 1.3014 1.1408
No log 0.7727 68 1.2419 0.0620 1.2419 1.1144
No log 0.7955 70 0.9896 0.1876 0.9896 0.9948
No log 0.8182 72 0.9115 0.1213 0.9115 0.9547
No log 0.8409 74 0.8706 0.1228 0.8706 0.9331
No log 0.8636 76 0.8838 0.2099 0.8838 0.9401
No log 0.8864 78 1.1724 0.0342 1.1724 1.0828
No log 0.9091 80 1.0761 -0.0492 1.0761 1.0373
No log 0.9318 82 0.9022 0.1198 0.9022 0.9498
No log 0.9545 84 1.1635 0.1476 1.1635 1.0787
No log 0.9773 86 1.1765 0.1521 1.1765 1.0846
No log 1.0 88 1.1246 0.1370 1.1246 1.0605
No log 1.0227 90 1.6583 0.1147 1.6583 1.2877
No log 1.0455 92 1.7024 0.0899 1.7024 1.3048
No log 1.0682 94 1.1680 0.0634 1.1680 1.0807
No log 1.0909 96 0.9114 0.1161 0.9114 0.9547
No log 1.1136 98 0.9780 0.1565 0.9780 0.9890
No log 1.1364 100 0.8596 0.1234 0.8596 0.9271
No log 1.1591 102 0.8383 0.1185 0.8383 0.9156
No log 1.1818 104 0.9207 0.0180 0.9207 0.9595
No log 1.2045 106 0.8820 0.1459 0.8820 0.9392
No log 1.2273 108 0.9047 0.1156 0.9047 0.9511
No log 1.25 110 1.0296 0.1265 1.0296 1.0147
No log 1.2727 112 0.9989 0.1492 0.9989 0.9994
No log 1.2955 114 0.9203 0.1737 0.9203 0.9593
No log 1.3182 116 0.9991 0.0639 0.9991 0.9995
No log 1.3409 118 0.9828 0.0623 0.9828 0.9913
No log 1.3636 120 0.8389 0.1489 0.8389 0.9159
No log 1.3864 122 0.8805 0.1609 0.8805 0.9383
No log 1.4091 124 0.9043 0.1292 0.9043 0.9510
No log 1.4318 126 0.9019 0.0544 0.9019 0.9497
No log 1.4545 128 1.0776 0.0138 1.0776 1.0381
No log 1.4773 130 1.0420 0.0186 1.0420 1.0208
No log 1.5 132 0.9250 0.1871 0.9250 0.9618
No log 1.5227 134 0.9686 0.1228 0.9686 0.9842
No log 1.5455 136 0.9037 0.1483 0.9037 0.9506
No log 1.5682 138 0.8449 0.0586 0.8449 0.9192
No log 1.5909 140 0.8458 0.0679 0.8458 0.9197
No log 1.6136 142 0.8437 0.0733 0.8437 0.9185
No log 1.6364 144 0.9194 0.1001 0.9194 0.9589
No log 1.6591 146 0.8814 0.1782 0.8814 0.9388
No log 1.6818 148 1.0224 0.0852 1.0224 1.0111
No log 1.7045 150 1.0330 0.0820 1.0330 1.0164
No log 1.7273 152 0.8168 0.0574 0.8168 0.9038
No log 1.75 154 0.8025 0.1425 0.8025 0.8958
No log 1.7727 156 0.7735 0.0983 0.7735 0.8795
No log 1.7955 158 0.8558 0.0837 0.8558 0.9251
No log 1.8182 160 0.9492 0.0627 0.9492 0.9742
No log 1.8409 162 1.0437 0.0147 1.0437 1.0216
No log 1.8636 164 0.9836 0.0951 0.9836 0.9917
No log 1.8864 166 0.9049 0.0933 0.9049 0.9512
No log 1.9091 168 0.9060 0.1638 0.9060 0.9518
No log 1.9318 170 0.8612 0.1011 0.8612 0.9280
No log 1.9545 172 0.8276 0.1310 0.8276 0.9097
No log 1.9773 174 0.7855 -0.0054 0.7855 0.8863
No log 2.0 176 0.8069 0.0159 0.8069 0.8983
No log 2.0227 178 0.8329 0.1342 0.8329 0.9126
No log 2.0455 180 0.9311 0.0684 0.9311 0.9649
No log 2.0682 182 0.9121 0.1235 0.9121 0.9550
No log 2.0909 184 0.8974 0.1212 0.8974 0.9473
No log 2.1136 186 1.0484 0.1184 1.0484 1.0239
No log 2.1364 188 0.9813 0.0847 0.9813 0.9906
No log 2.1591 190 0.9805 0.0639 0.9805 0.9902
No log 2.1818 192 0.9852 -0.0151 0.9852 0.9926
No log 2.2045 194 0.9417 0.0207 0.9417 0.9704
No log 2.2273 196 0.7449 0.0723 0.7449 0.8631
No log 2.25 198 0.7393 0.0056 0.7393 0.8598
No log 2.2727 200 0.8057 0.0664 0.8057 0.8976
No log 2.2955 202 0.7999 0.0341 0.7999 0.8944
No log 2.3182 204 1.0013 0.0692 1.0013 1.0006
No log 2.3409 206 1.1822 0.0098 1.1822 1.0873
No log 2.3636 208 0.9796 0.0419 0.9796 0.9898
No log 2.3864 210 0.9257 0.0271 0.9257 0.9621
No log 2.4091 212 0.9084 0.0734 0.9084 0.9531
No log 2.4318 214 0.7702 0.0481 0.7702 0.8776
No log 2.4545 216 0.7886 0.1691 0.7886 0.8880
No log 2.4773 218 0.8244 0.1495 0.8244 0.9079
No log 2.5 220 0.8391 0.1187 0.8391 0.9160
No log 2.5227 222 0.8305 0.1187 0.8305 0.9113
No log 2.5455 224 0.8595 0.1393 0.8595 0.9271
No log 2.5682 226 0.9872 0.0257 0.9872 0.9936
No log 2.5909 228 0.9472 0.0781 0.9472 0.9732
No log 2.6136 230 0.9323 0.0563 0.9323 0.9656
No log 2.6364 232 0.9779 0.0900 0.9779 0.9889
No log 2.6591 234 1.1405 0.0448 1.1405 1.0679
No log 2.6818 236 1.0040 0.0415 1.0040 1.0020
No log 2.7045 238 0.8464 -0.0230 0.8464 0.9200
No log 2.7273 240 1.0032 0.0147 1.0032 1.0016
No log 2.75 242 0.9513 0.0881 0.9513 0.9754
No log 2.7727 244 0.8273 0.0875 0.8273 0.9095
No log 2.7955 246 0.8786 0.1277 0.8786 0.9374
No log 2.8182 248 0.8450 0.1923 0.8450 0.9192
No log 2.8409 250 0.9120 0.0805 0.9120 0.9550
No log 2.8636 252 0.8857 0.1633 0.8857 0.9411
No log 2.8864 254 0.8068 0.1529 0.8068 0.8982
No log 2.9091 256 0.8238 0.1352 0.8238 0.9076
No log 2.9318 258 0.8306 0.1770 0.8306 0.9114
No log 2.9545 260 0.8601 0.1144 0.8601 0.9274
No log 2.9773 262 0.9133 0.1522 0.9133 0.9556
No log 3.0 264 0.8087 0.1440 0.8087 0.8993
No log 3.0227 266 0.7684 0.0428 0.7684 0.8766
No log 3.0455 268 0.8338 0.1686 0.8338 0.9131
No log 3.0682 270 0.7937 0.1315 0.7937 0.8909
No log 3.0909 272 0.7628 0.1691 0.7628 0.8734
No log 3.1136 274 0.8494 0.1395 0.8494 0.9216
No log 3.1364 276 0.9000 0.1065 0.9000 0.9487
No log 3.1591 278 0.8100 0.1691 0.8100 0.9000
No log 3.1818 280 0.7814 0.0412 0.7814 0.8840
No log 3.2045 282 0.7751 0.1196 0.7751 0.8804
No log 3.2273 284 0.8189 0.2009 0.8189 0.9049
No log 3.25 286 0.9204 0.0283 0.9204 0.9594
No log 3.2727 288 0.8337 0.0588 0.8337 0.9131
No log 3.2955 290 0.7454 0.1254 0.7454 0.8634
No log 3.3182 292 0.7126 0.0964 0.7126 0.8442
No log 3.3409 294 0.7086 0.1318 0.7086 0.8418
No log 3.3636 296 0.8149 0.1633 0.8149 0.9027
No log 3.3864 298 0.9299 0.1065 0.9299 0.9643
No log 3.4091 300 0.8673 0.1522 0.8673 0.9313
No log 3.4318 302 0.8071 0.1691 0.8071 0.8984
No log 3.4545 304 0.7967 0.1691 0.7967 0.8926
No log 3.4773 306 0.7466 0.1691 0.7466 0.8641
No log 3.5 308 0.7454 0.1691 0.7454 0.8633
No log 3.5227 310 0.7162 0.1259 0.7162 0.8463
No log 3.5455 312 0.6714 0.0964 0.6714 0.8194
No log 3.5682 314 0.6712 0.0964 0.6712 0.8193
No log 3.5909 316 0.6998 0.0914 0.6998 0.8366
No log 3.6136 318 0.7638 0.0879 0.7638 0.8740
No log 3.6364 320 0.8857 0.1264 0.8857 0.9411
No log 3.6591 322 0.9157 0.0779 0.9157 0.9569
No log 3.6818 324 0.9403 0.0277 0.9403 0.9697
No log 3.7045 326 0.8744 0.0761 0.8744 0.9351
No log 3.7273 328 0.8297 0.0670 0.8297 0.9109
No log 3.75 330 0.8045 0.0660 0.8045 0.8969
No log 3.7727 332 0.8321 0.1431 0.8321 0.9122
No log 3.7955 334 0.8102 0.0741 0.8102 0.9001
No log 3.8182 336 0.8459 0.0514 0.8459 0.9197
No log 3.8409 338 0.8436 0.0069 0.8436 0.9185
No log 3.8636 340 0.8458 0.0660 0.8458 0.9197
No log 3.8864 342 0.8578 0.0257 0.8578 0.9262
No log 3.9091 344 0.8444 0.0633 0.8444 0.9189
No log 3.9318 346 0.7975 0.0323 0.7975 0.8930
No log 3.9545 348 0.7855 0.0 0.7855 0.8863
No log 3.9773 350 0.7736 0.0930 0.7736 0.8796
No log 4.0 352 0.7296 0.0918 0.7296 0.8542
No log 4.0227 354 0.7514 0.1097 0.7514 0.8669
No log 4.0455 356 0.9130 0.1025 0.9130 0.9555
No log 4.0682 358 0.8653 0.1395 0.8653 0.9302
No log 4.0909 360 0.7804 0.1047 0.7804 0.8834
No log 4.1136 362 0.7585 0.1565 0.7585 0.8709
No log 4.1364 364 0.7730 0.1565 0.7730 0.8792
No log 4.1591 366 0.7920 0.1232 0.7920 0.8900
No log 4.1818 368 0.8305 0.0771 0.8305 0.9113
No log 4.2045 370 0.8656 0.0652 0.8656 0.9304
No log 4.2273 372 0.9195 0.0802 0.9195 0.9589
No log 4.25 374 0.8823 0.0796 0.8823 0.9393
No log 4.2727 376 0.8760 0.0876 0.8760 0.9360
No log 4.2955 378 0.8336 0.1047 0.8336 0.9130
No log 4.3182 380 0.7776 0.2181 0.7776 0.8818
No log 4.3409 382 0.8755 -0.0056 0.8755 0.9357
No log 4.3636 384 0.8415 0.0230 0.8415 0.9173
No log 4.3864 386 0.7471 0.1404 0.7471 0.8643
No log 4.4091 388 0.7562 0.1675 0.7562 0.8696
No log 4.4318 390 0.8926 0.1190 0.8926 0.9448
No log 4.4545 392 0.9902 0.0557 0.9902 0.9951
No log 4.4773 394 0.9243 0.1475 0.9243 0.9614
No log 4.5 396 0.7969 0.1049 0.7969 0.8927
No log 4.5227 398 0.8467 0.1299 0.8467 0.9202
No log 4.5455 400 0.9184 0.0964 0.9184 0.9583
No log 4.5682 402 0.8917 0.1209 0.8917 0.9443
No log 4.5909 404 0.8588 0.0930 0.8588 0.9267
No log 4.6136 406 0.9018 0.0734 0.9018 0.9496
No log 4.6364 408 0.8351 0.1329 0.8351 0.9138
No log 4.6591 410 0.8090 0.1329 0.8090 0.8994
No log 4.6818 412 0.7248 0.2339 0.7248 0.8513
No log 4.7045 414 0.7108 0.1902 0.7108 0.8431
No log 4.7273 416 0.6987 0.2339 0.6987 0.8359
No log 4.75 418 0.7938 0.1097 0.7938 0.8910
No log 4.7727 420 0.9750 -0.0094 0.9750 0.9874
No log 4.7955 422 0.9731 0.0293 0.9731 0.9865
No log 4.8182 424 0.8464 0.1096 0.8464 0.9200
No log 4.8409 426 0.8154 0.0361 0.8154 0.9030
No log 4.8636 428 0.8041 0.0757 0.8041 0.8967
No log 4.8864 430 0.8632 0.1506 0.8632 0.9291
No log 4.9091 432 1.0148 0.0353 1.0148 1.0074
No log 4.9318 434 1.1761 -0.0151 1.1761 1.0845
No log 4.9545 436 1.0820 0.0224 1.0820 1.0402
No log 4.9773 438 0.8541 0.0953 0.8541 0.9242
No log 5.0 440 0.7556 0.1148 0.7556 0.8692
No log 5.0227 442 0.7568 0.1192 0.7568 0.8699
No log 5.0455 444 0.7970 0.0377 0.7970 0.8928
No log 5.0682 446 0.8249 0.0362 0.8249 0.9083
No log 5.0909 448 0.8241 0.0301 0.8241 0.9078
No log 5.1136 450 0.8541 0.0953 0.8541 0.9242
No log 5.1364 452 0.8876 0.0867 0.8876 0.9421
No log 5.1591 454 0.8376 0.0909 0.8376 0.9152
No log 5.1818 456 0.7954 -0.0023 0.7954 0.8918
No log 5.2045 458 0.9194 0.0651 0.9194 0.9589
No log 5.2273 460 0.9282 0.0651 0.9282 0.9634
No log 5.25 462 0.8374 0.0827 0.8374 0.9151
No log 5.2727 464 0.8041 0.1096 0.8041 0.8967
No log 5.2955 466 0.8328 0.1105 0.8328 0.9126
No log 5.3182 468 0.7753 0.1097 0.7753 0.8805
No log 5.3409 470 0.7120 0.1691 0.7120 0.8438
No log 5.3636 472 0.7075 0.1444 0.7075 0.8411
No log 5.3864 474 0.7575 0.0930 0.7575 0.8703
No log 5.4091 476 0.8017 0.0078 0.8017 0.8954
No log 5.4318 478 0.7919 0.0884 0.7919 0.8899
No log 5.4545 480 0.8380 0.1183 0.8380 0.9154
No log 5.4773 482 0.8737 0.1718 0.8737 0.9347
No log 5.5 484 0.8197 0.0964 0.8197 0.9054
No log 5.5227 486 0.7585 0.0791 0.7585 0.8709
No log 5.5455 488 0.7253 0.0922 0.7253 0.8516
No log 5.5682 490 0.7143 0.1444 0.7143 0.8451
No log 5.5909 492 0.7260 0.1691 0.7260 0.8521
No log 5.6136 494 0.7693 0.1148 0.7693 0.8771
No log 5.6364 496 0.8151 0.1148 0.8151 0.9028
No log 5.6591 498 0.8028 0.1148 0.8028 0.8960
0.3314 5.6818 500 0.7706 0.1691 0.7706 0.8779
0.3314 5.7045 502 0.7790 0.1192 0.7790 0.8826
0.3314 5.7273 504 0.8151 0.1189 0.8151 0.9029
0.3314 5.75 506 0.8774 0.0525 0.8774 0.9367
0.3314 5.7727 508 0.9252 -0.0008 0.9252 0.9619
0.3314 5.7955 510 0.8346 0.1047 0.8346 0.9135

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k17_task3_organization

Finetuned
(4019)
this model