ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9180
  • Qwk: -0.1057
  • Mse: 0.9180
  • Rmse: 0.9581

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0244 2 3.6975 0.0029 3.6975 1.9229
No log 0.0488 4 2.2004 -0.0105 2.2004 1.4834
No log 0.0732 6 1.3888 -0.0494 1.3888 1.1785
No log 0.0976 8 1.0420 0.1316 1.0420 1.0208
No log 0.1220 10 0.8344 -0.1233 0.8344 0.9134
No log 0.1463 12 0.8559 -0.0778 0.8559 0.9251
No log 0.1707 14 0.8360 -0.0371 0.8360 0.9143
No log 0.1951 16 0.7363 0.0759 0.7363 0.8581
No log 0.2195 18 0.7330 0.1202 0.7330 0.8561
No log 0.2439 20 0.8740 0.0515 0.8740 0.9349
No log 0.2683 22 0.9916 0.0046 0.9916 0.9958
No log 0.2927 24 0.8871 0.0346 0.8871 0.9418
No log 0.3171 26 0.7248 0.1202 0.7248 0.8513
No log 0.3415 28 0.6813 0.1582 0.6813 0.8254
No log 0.3659 30 0.6650 0.0416 0.6650 0.8155
No log 0.3902 32 0.6780 0.0857 0.6780 0.8234
No log 0.4146 34 0.8420 0.0642 0.8420 0.9176
No log 0.4390 36 0.9931 -0.0049 0.9931 0.9965
No log 0.4634 38 0.9617 -0.0316 0.9617 0.9807
No log 0.4878 40 0.7947 0.1342 0.7947 0.8915
No log 0.5122 42 0.6503 0.1902 0.6503 0.8064
No log 0.5366 44 0.7029 0.0714 0.7029 0.8384
No log 0.5610 46 0.7313 0.0129 0.7313 0.8552
No log 0.5854 48 0.7828 0.0071 0.7828 0.8847
No log 0.6098 50 0.6885 0.1379 0.6885 0.8298
No log 0.6341 52 0.6972 0.0964 0.6972 0.8350
No log 0.6585 54 0.7837 0.1259 0.7837 0.8853
No log 0.6829 56 0.8231 0.0260 0.8231 0.9073
No log 0.7073 58 0.7978 -0.0152 0.7978 0.8932
No log 0.7317 60 0.8666 -0.0089 0.8666 0.9309
No log 0.7561 62 0.9411 0.1111 0.9411 0.9701
No log 0.7805 64 0.8603 -0.0116 0.8603 0.9275
No log 0.8049 66 0.7686 0.0874 0.7686 0.8767
No log 0.8293 68 0.7659 -0.0138 0.7659 0.8751
No log 0.8537 70 0.7288 0.1240 0.7288 0.8537
No log 0.8780 72 0.7454 0.0481 0.7454 0.8634
No log 0.9024 74 0.7602 0.0930 0.7602 0.8719
No log 0.9268 76 0.7913 0.0444 0.7913 0.8896
No log 0.9512 78 0.8275 0.0444 0.8275 0.9097
No log 0.9756 80 0.8891 0.0378 0.8891 0.9429
No log 1.0 82 0.9566 0.0685 0.9566 0.9781
No log 1.0244 84 1.0434 0.0028 1.0434 1.0215
No log 1.0488 86 1.1202 0.0046 1.1202 1.0584
No log 1.0732 88 1.0515 0.0529 1.0515 1.0254
No log 1.0976 90 1.0663 0.0779 1.0663 1.0326
No log 1.1220 92 1.0213 -0.1896 1.0213 1.0106
No log 1.1463 94 1.0241 -0.0350 1.0241 1.0120
No log 1.1707 96 1.0136 -0.1027 1.0136 1.0068
No log 1.1951 98 1.0033 -0.0439 1.0033 1.0016
No log 1.2195 100 1.0097 -0.1255 1.0097 1.0048
No log 1.2439 102 0.9749 -0.1355 0.9749 0.9874
No log 1.2683 104 0.9821 -0.1462 0.9821 0.9910
No log 1.2927 106 1.0894 -0.1371 1.0894 1.0437
No log 1.3171 108 1.1636 0.0098 1.1636 1.0787
No log 1.3415 110 1.1278 0.0729 1.1278 1.0620
No log 1.3659 112 1.0123 -0.1306 1.0123 1.0061
No log 1.3902 114 0.9968 -0.1229 0.9968 0.9984
No log 1.4146 116 1.0301 -0.1115 1.0301 1.0149
No log 1.4390 118 1.0923 0.1926 1.0923 1.0451
No log 1.4634 120 1.0782 0.0462 1.0782 1.0384
No log 1.4878 122 0.9933 0.0323 0.9933 0.9966
No log 1.5122 124 1.0004 -0.0247 1.0004 1.0002
No log 1.5366 126 1.0542 -0.0133 1.0542 1.0267
No log 1.5610 128 1.2677 -0.0297 1.2677 1.1259
No log 1.5854 130 1.3608 0.0581 1.3608 1.1665
No log 1.6098 132 1.0900 -0.0096 1.0900 1.0440
No log 1.6341 134 1.1030 0.0039 1.1030 1.0502
No log 1.6585 136 1.1865 -0.0689 1.1865 1.0893
No log 1.6829 138 1.4270 -0.0022 1.4270 1.1946
No log 1.7073 140 1.2341 -0.0207 1.2341 1.1109
No log 1.7317 142 1.1111 -0.0074 1.1111 1.0541
No log 1.7561 144 1.1082 -0.0160 1.1082 1.0527
No log 1.7805 146 1.0327 0.0602 1.0327 1.0162
No log 1.8049 148 1.0136 -0.0368 1.0136 1.0068
No log 1.8293 150 1.1816 -0.1302 1.1816 1.0870
No log 1.8537 152 1.0332 -0.1721 1.0332 1.0164
No log 1.8780 154 0.9009 -0.0108 0.9009 0.9492
No log 1.9024 156 0.9007 -0.0334 0.9007 0.9491
No log 1.9268 158 1.0615 -0.1624 1.0615 1.0303
No log 1.9512 160 1.0853 -0.1429 1.0853 1.0418
No log 1.9756 162 1.1878 -0.0632 1.1878 1.0898
No log 2.0 164 1.1235 0.0164 1.1235 1.0599
No log 2.0244 166 1.0834 -0.0550 1.0834 1.0409
No log 2.0488 168 1.1514 -0.0550 1.1514 1.0730
No log 2.0732 170 1.3303 -0.0716 1.3303 1.1534
No log 2.0976 172 1.2045 -0.0550 1.2045 1.0975
No log 2.1220 174 1.2253 -0.0523 1.2253 1.1069
No log 2.1463 176 1.4236 -0.0249 1.4236 1.1932
No log 2.1707 178 1.2540 -0.1115 1.2540 1.1198
No log 2.1951 180 1.1216 -0.0908 1.1216 1.0591
No log 2.2195 182 1.0989 -0.1307 1.0989 1.0483
No log 2.2439 184 1.2296 -0.0802 1.2296 1.1089
No log 2.2683 186 1.3271 -0.0270 1.3271 1.1520
No log 2.2927 188 1.1032 -0.0981 1.1032 1.0503
No log 2.3171 190 1.0289 0.0218 1.0289 1.0143
No log 2.3415 192 1.0562 -0.0655 1.0562 1.0277
No log 2.3659 194 1.1070 -0.1517 1.1070 1.0521
No log 2.3902 196 1.4623 -0.0172 1.4623 1.2093
No log 2.4146 198 1.3008 -0.0587 1.3008 1.1405
No log 2.4390 200 0.9401 -0.1166 0.9401 0.9696
No log 2.4634 202 1.0053 -0.0735 1.0053 1.0026
No log 2.4878 204 0.9144 -0.0889 0.9144 0.9562
No log 2.5122 206 0.9391 -0.0217 0.9391 0.9691
No log 2.5366 208 1.0656 -0.1733 1.0656 1.0323
No log 2.5610 210 1.1110 -0.1791 1.1110 1.0540
No log 2.5854 212 1.1618 -0.0586 1.1618 1.0779
No log 2.6098 214 1.2536 -0.1151 1.2536 1.1196
No log 2.6341 216 1.0429 -0.0512 1.0429 1.0212
No log 2.6585 218 1.1389 -0.0828 1.1389 1.0672
No log 2.6829 220 1.1562 -0.1528 1.1562 1.0752
No log 2.7073 222 0.9060 0.0175 0.9060 0.9518
No log 2.7317 224 0.9233 -0.1642 0.9233 0.9609
No log 2.7561 226 1.0653 -0.1108 1.0653 1.0321
No log 2.7805 228 0.9360 -0.1703 0.9360 0.9675
No log 2.8049 230 0.8672 0.0700 0.8672 0.9312
No log 2.8293 232 0.9010 0.0690 0.9010 0.9492
No log 2.8537 234 0.9219 -0.1062 0.9219 0.9601
No log 2.8780 236 1.0355 -0.0712 1.0355 1.0176
No log 2.9024 238 1.1195 -0.0013 1.1195 1.0581
No log 2.9268 240 1.1055 0.0521 1.1055 1.0514
No log 2.9512 242 1.1809 0.0042 1.1809 1.0867
No log 2.9756 244 1.2287 -0.0107 1.2287 1.1085
No log 3.0 246 1.0832 -0.0648 1.0832 1.0408
No log 3.0244 248 1.0310 -0.0931 1.0310 1.0154
No log 3.0488 250 0.9598 -0.0893 0.9598 0.9797
No log 3.0732 252 0.9608 -0.1208 0.9608 0.9802
No log 3.0976 254 0.9595 -0.1420 0.9595 0.9796
No log 3.1220 256 1.0473 -0.1032 1.0473 1.0234
No log 3.1463 258 0.9266 -0.1249 0.9266 0.9626
No log 3.1707 260 0.9126 -0.0878 0.9126 0.9553
No log 3.1951 262 0.9007 -0.1021 0.9007 0.9491
No log 3.2195 264 0.8877 -0.1408 0.8877 0.9422
No log 3.2439 266 0.8839 -0.1355 0.8839 0.9402
No log 3.2683 268 0.9224 -0.0870 0.9224 0.9604
No log 3.2927 270 1.1023 -0.0436 1.1023 1.0499
No log 3.3171 272 1.4483 0.0610 1.4483 1.2035
No log 3.3415 274 1.3302 0.0603 1.3302 1.1533
No log 3.3659 276 0.9706 -0.2323 0.9706 0.9852
No log 3.3902 278 0.8984 -0.0567 0.8984 0.9478
No log 3.4146 280 0.8950 -0.1597 0.8950 0.9461
No log 3.4390 282 1.0204 -0.1422 1.0204 1.0102
No log 3.4634 284 1.0019 -0.2372 1.0019 1.0009
No log 3.4878 286 0.8912 0.0 0.8912 0.9440
No log 3.5122 288 0.9143 0.0791 0.9143 0.9562
No log 3.5366 290 0.9828 -0.2413 0.9828 0.9914
No log 3.5610 292 1.1006 -0.0855 1.1006 1.0491
No log 3.5854 294 1.0246 -0.1777 1.0246 1.0122
No log 3.6098 296 0.9701 -0.0099 0.9701 0.9849
No log 3.6341 298 0.9667 0.0071 0.9667 0.9832
No log 3.6585 300 1.1105 -0.1620 1.1105 1.0538
No log 3.6829 302 1.1824 -0.1429 1.1824 1.0874
No log 3.7073 304 1.0644 -0.1624 1.0644 1.0317
No log 3.7317 306 0.9166 -0.0730 0.9166 0.9574
No log 3.7561 308 0.9058 0.0732 0.9058 0.9517
No log 3.7805 310 0.8775 -0.0462 0.8775 0.9367
No log 3.8049 312 0.8867 -0.1967 0.8867 0.9417
No log 3.8293 314 1.0295 -0.1971 1.0295 1.0146
No log 3.8537 316 0.9877 -0.1099 0.9877 0.9938
No log 3.8780 318 0.8972 0.0359 0.8972 0.9472
No log 3.9024 320 0.9501 0.0956 0.9501 0.9747
No log 3.9268 322 0.9354 -0.0295 0.9354 0.9672
No log 3.9512 324 1.0657 -0.1962 1.0657 1.0323
No log 3.9756 326 1.0470 -0.0181 1.0470 1.0232
No log 4.0 328 0.8931 -0.0522 0.8931 0.9451
No log 4.0244 330 0.8904 0.0469 0.8904 0.9436
No log 4.0488 332 0.9135 0.1192 0.9135 0.9558
No log 4.0732 334 0.9244 0.0732 0.9244 0.9614
No log 4.0976 336 0.9332 -0.1686 0.9332 0.9660
No log 4.1220 338 1.1167 -0.1803 1.1167 1.0567
No log 4.1463 340 1.0853 -0.1232 1.0853 1.0418
No log 4.1707 342 0.9677 -0.1572 0.9677 0.9837
No log 4.1951 344 1.1687 -0.1541 1.1687 1.0810
No log 4.2195 346 1.1212 -0.1107 1.1212 1.0589
No log 4.2439 348 0.9677 -0.1520 0.9677 0.9837
No log 4.2683 350 1.1967 -0.1069 1.1967 1.0940
No log 4.2927 352 1.2180 -0.0661 1.2180 1.1036
No log 4.3171 354 1.0003 -0.1249 1.0003 1.0002
No log 4.3415 356 0.9173 0.0323 0.9173 0.9578
No log 4.3659 358 0.9274 -0.0672 0.9274 0.9630
No log 4.3902 360 0.9455 -0.0730 0.9455 0.9724
No log 4.4146 362 1.1187 -0.0820 1.1187 1.0577
No log 4.4390 364 1.0865 -0.1189 1.0865 1.0423
No log 4.4634 366 0.9353 -0.1045 0.9353 0.9671
No log 4.4878 368 0.9094 -0.1155 0.9094 0.9536
No log 4.5122 370 0.9584 -0.0622 0.9584 0.9790
No log 4.5366 372 0.9859 -0.0585 0.9859 0.9929
No log 4.5610 374 1.0407 -0.1068 1.0407 1.0201
No log 4.5854 376 1.1558 -0.0855 1.1558 1.0751
No log 4.6098 378 1.0445 -0.1032 1.0445 1.0220
No log 4.6341 380 0.9435 -0.1222 0.9435 0.9714
No log 4.6585 382 0.9565 -0.0816 0.9565 0.9780
No log 4.6829 384 1.0226 -0.0454 1.0226 1.0112
No log 4.7073 386 0.9947 -0.0829 0.9947 0.9973
No log 4.7317 388 0.9786 -0.1637 0.9786 0.9892
No log 4.7561 390 0.8825 -0.2043 0.8825 0.9394
No log 4.7805 392 0.8242 0.0375 0.8242 0.9079
No log 4.8049 394 0.8305 0.1199 0.8305 0.9113
No log 4.8293 396 0.8572 -0.1333 0.8572 0.9259
No log 4.8537 398 0.9345 -0.0963 0.9345 0.9667
No log 4.8780 400 0.9551 -0.0920 0.9551 0.9773
No log 4.9024 402 0.9578 -0.0920 0.9578 0.9787
No log 4.9268 404 0.9796 -0.1896 0.9796 0.9898
No log 4.9512 406 0.9054 -0.1532 0.9054 0.9515
No log 4.9756 408 0.8442 0.1254 0.8442 0.9188
No log 5.0 410 0.8748 0.1001 0.8748 0.9353
No log 5.0244 412 0.8264 0.2009 0.8264 0.9090
No log 5.0488 414 0.8099 -0.0032 0.8099 0.9000
No log 5.0732 416 0.9641 -0.1323 0.9641 0.9819
No log 5.0976 418 1.1998 -0.0661 1.1998 1.0954
No log 5.1220 420 1.1197 -0.0464 1.1197 1.0582
No log 5.1463 422 0.9066 -0.0598 0.9066 0.9522
No log 5.1707 424 0.9515 0.0134 0.9515 0.9755
No log 5.1951 426 0.9057 -0.0923 0.9057 0.9517
No log 5.2195 428 0.9002 -0.0488 0.9002 0.9488
No log 5.2439 430 0.8407 -0.0248 0.8407 0.9169
No log 5.2683 432 0.7591 0.1486 0.7591 0.8712
No log 5.2927 434 0.7536 0.0967 0.7536 0.8681
No log 5.3171 436 0.7571 0.0 0.7571 0.8701
No log 5.3415 438 0.8300 -0.0314 0.8300 0.9110
No log 5.3659 440 0.9558 -0.0488 0.9558 0.9776
No log 5.3902 442 1.0438 -0.0470 1.0438 1.0217
No log 5.4146 444 1.0592 0.0153 1.0592 1.0292
No log 5.4390 446 1.0305 0.0066 1.0305 1.0151
No log 5.4634 448 1.0219 -0.0326 1.0219 1.0109
No log 5.4878 450 1.0990 -0.1603 1.0990 1.0483
No log 5.5122 452 1.0474 -0.1073 1.0474 1.0234
No log 5.5366 454 0.9199 -0.1969 0.9199 0.9591
No log 5.5610 456 0.8842 0.0732 0.8842 0.9403
No log 5.5854 458 0.9508 -0.0391 0.9508 0.9751
No log 5.6098 460 0.9143 0.1003 0.9143 0.9562
No log 5.6341 462 0.9270 -0.1102 0.9270 0.9628
No log 5.6585 464 1.0179 -0.1152 1.0179 1.0089
No log 5.6829 466 0.9679 -0.2094 0.9679 0.9838
No log 5.7073 468 0.9018 -0.1057 0.9018 0.9496
No log 5.7317 470 0.8640 -0.1217 0.8640 0.9295
No log 5.7561 472 0.8303 -0.0029 0.8303 0.9112
No log 5.7805 474 0.8211 0.0030 0.8211 0.9061
No log 5.8049 476 0.8567 -0.1102 0.8567 0.9256
No log 5.8293 478 0.8719 -0.0687 0.8719 0.9337
No log 5.8537 480 0.8970 -0.0622 0.8970 0.9471
No log 5.8780 482 0.9465 -0.1159 0.9465 0.9729
No log 5.9024 484 0.9481 -0.1304 0.9481 0.9737
No log 5.9268 486 0.9746 -0.1305 0.9746 0.9872
No log 5.9512 488 0.9379 -0.1027 0.9379 0.9685
No log 5.9756 490 0.8960 -0.1355 0.8960 0.9466
No log 6.0 492 0.8753 -0.1409 0.8753 0.9356
No log 6.0244 494 0.8415 -0.1159 0.8415 0.9174
No log 6.0488 496 0.8189 -0.0407 0.8189 0.9049
No log 6.0732 498 0.8194 -0.0446 0.8194 0.9052
0.348 6.0976 500 0.8319 -0.0334 0.8319 0.9121
0.348 6.1220 502 0.8697 -0.1106 0.8697 0.9326
0.348 6.1463 504 0.9339 -0.1301 0.9339 0.9664
0.348 6.1707 506 0.9924 -0.0492 0.9924 0.9962
0.348 6.1951 508 0.8842 -0.0558 0.8842 0.9403
0.348 6.2195 510 0.8099 0.0503 0.8099 0.8999
0.348 6.2439 512 0.7826 0.0914 0.7826 0.8846
0.348 6.2683 514 0.7757 0.0967 0.7757 0.8807
0.348 6.2927 516 0.8411 -0.1201 0.8411 0.9171
0.348 6.3171 518 0.8693 -0.0982 0.8693 0.9324
0.348 6.3415 520 0.8363 -0.0334 0.8363 0.9145
0.348 6.3659 522 0.8783 0.0211 0.8783 0.9372
0.348 6.3902 524 0.9491 -0.0820 0.9491 0.9742
0.348 6.4146 526 0.9051 -0.0536 0.9051 0.9514
0.348 6.4390 528 0.9054 -0.0511 0.9054 0.9515
0.348 6.4634 530 0.9076 -0.0806 0.9076 0.9527
0.348 6.4878 532 0.9180 -0.1057 0.9180 0.9581

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task3_organization

Finetuned
(4019)
this model