ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7756
  • Qwk: 0.0481
  • Mse: 0.7756
  • Rmse: 0.8807

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 3.4792 -0.0047 3.4792 1.8653
No log 0.0851 4 1.5870 -0.0029 1.5870 1.2598
No log 0.1277 6 1.0639 -0.0987 1.0639 1.0315
No log 0.1702 8 1.0540 -0.0359 1.0540 1.0267
No log 0.2128 10 0.8407 0.0512 0.8407 0.9169
No log 0.2553 12 0.8655 0.1107 0.8655 0.9303
No log 0.2979 14 0.7198 -0.0069 0.7198 0.8484
No log 0.3404 16 0.7528 -0.0626 0.7528 0.8676
No log 0.3830 18 0.7752 0.0 0.7752 0.8805
No log 0.4255 20 0.7477 -0.0069 0.7477 0.8647
No log 0.4681 22 1.2878 0.0309 1.2878 1.1348
No log 0.5106 24 1.0506 0.0083 1.0506 1.0250
No log 0.5532 26 0.7554 -0.0101 0.7554 0.8691
No log 0.5957 28 0.7563 -0.0069 0.7563 0.8697
No log 0.6383 30 0.7548 -0.1223 0.7548 0.8688
No log 0.6809 32 0.7648 -0.1233 0.7648 0.8745
No log 0.7234 34 0.8259 -0.0351 0.8259 0.9088
No log 0.7660 36 1.0244 0.0169 1.0244 1.0121
No log 0.8085 38 1.4251 0.0993 1.4251 1.1938
No log 0.8511 40 1.0803 0.0789 1.0803 1.0394
No log 0.8936 42 0.8129 0.0159 0.8129 0.9016
No log 0.9362 44 0.8291 -0.0152 0.8291 0.9106
No log 0.9787 46 0.8344 0.0282 0.8344 0.9134
No log 1.0213 48 0.8362 0.0318 0.8362 0.9144
No log 1.0638 50 0.8291 -0.0331 0.8291 0.9106
No log 1.1064 52 1.0148 0.0200 1.0148 1.0074
No log 1.1489 54 1.0825 0.0403 1.0825 1.0405
No log 1.1915 56 0.8204 0.1823 0.8204 0.9058
No log 1.2340 58 0.8642 0.0093 0.8642 0.9296
No log 1.2766 60 0.8202 0.0344 0.8202 0.9057
No log 1.3191 62 0.8392 0.1727 0.8392 0.9161
No log 1.3617 64 0.8221 0.0246 0.8221 0.9067
No log 1.4043 66 0.9766 0.0596 0.9766 0.9882
No log 1.4468 68 0.9651 0.1162 0.9651 0.9824
No log 1.4894 70 0.9680 -0.0813 0.9680 0.9839
No log 1.5319 72 1.2280 0.1241 1.2280 1.1082
No log 1.5745 74 1.3255 0.0746 1.3255 1.1513
No log 1.6170 76 0.9957 -0.0439 0.9957 0.9978
No log 1.6596 78 1.1233 0.0247 1.1233 1.0598
No log 1.7021 80 1.2618 0.0865 1.2618 1.1233
No log 1.7447 82 1.0304 0.1712 1.0304 1.0151
No log 1.7872 84 0.7679 0.0733 0.7679 0.8763
No log 1.8298 86 0.8097 0.1395 0.8097 0.8998
No log 1.8723 88 0.7657 0.1983 0.7657 0.8750
No log 1.9149 90 0.7669 0.1599 0.7669 0.8757
No log 1.9574 92 0.7759 0.2181 0.7759 0.8809
No log 2.0 94 0.8083 0.1687 0.8083 0.8991
No log 2.0426 96 0.8579 0.0964 0.8579 0.9262
No log 2.0851 98 0.8275 0.0600 0.8275 0.9097
No log 2.1277 100 0.7782 0.2270 0.7782 0.8821
No log 2.1702 102 0.8326 0.0880 0.8326 0.9125
No log 2.2128 104 0.9467 0.1744 0.9467 0.9730
No log 2.2553 106 0.9287 0.0807 0.9287 0.9637
No log 2.2979 108 1.3184 0.0513 1.3184 1.1482
No log 2.3404 110 1.4474 0.0330 1.4474 1.2031
No log 2.3830 112 1.1852 0.0802 1.1852 1.0887
No log 2.4255 114 0.8321 0.1558 0.8321 0.9122
No log 2.4681 116 0.9899 0.1231 0.9899 0.9949
No log 2.5106 118 0.9822 0.1273 0.9822 0.9911
No log 2.5532 120 0.7866 0.0549 0.7866 0.8869
No log 2.5957 122 0.7822 0.0834 0.7822 0.8844
No log 2.6383 124 0.9392 0.1262 0.9392 0.9691
No log 2.6809 126 1.1245 0.1312 1.1245 1.0604
No log 2.7234 128 1.0579 0.0301 1.0579 1.0286
No log 2.7660 130 0.9515 0.0310 0.9515 0.9755
No log 2.8085 132 1.0064 0.0701 1.0064 1.0032
No log 2.8511 134 1.2744 0.0603 1.2744 1.1289
No log 2.8936 136 1.0019 0.0142 1.0019 1.0009
No log 2.9362 138 0.7974 0.0303 0.7974 0.8930
No log 2.9787 140 0.8751 0.0017 0.8751 0.9355
No log 3.0213 142 0.8432 0.1449 0.8432 0.9183
No log 3.0638 144 0.8937 -0.1057 0.8937 0.9453
No log 3.1064 146 1.2044 0.0832 1.2044 1.0975
No log 3.1489 148 1.2593 0.0726 1.2593 1.1222
No log 3.1915 150 1.0646 0.0579 1.0646 1.0318
No log 3.2340 152 1.0373 0.0589 1.0373 1.0185
No log 3.2766 154 1.1391 0.0789 1.1391 1.0673
No log 3.3191 156 1.0087 0.0758 1.0087 1.0044
No log 3.3617 158 0.8171 -0.0054 0.8171 0.9039
No log 3.4043 160 0.9522 0.0711 0.9522 0.9758
No log 3.4468 162 0.8916 0.0438 0.8916 0.9443
No log 3.4894 164 0.8569 -0.0357 0.8569 0.9257
No log 3.5319 166 0.9931 0.0717 0.9931 0.9965
No log 3.5745 168 0.9944 0.0378 0.9944 0.9972
No log 3.6170 170 0.8849 0.0540 0.8849 0.9407
No log 3.6596 172 0.8952 -0.0099 0.8952 0.9462
No log 3.7021 174 0.9381 0.0455 0.9381 0.9686
No log 3.7447 176 0.9965 0.0687 0.9965 0.9983
No log 3.7872 178 0.9954 0.0912 0.9954 0.9977
No log 3.8298 180 0.9281 -0.0062 0.9281 0.9634
No log 3.8723 182 0.9298 -0.0152 0.9298 0.9643
No log 3.9149 184 0.8952 0.0218 0.8952 0.9462
No log 3.9574 186 0.8839 0.0847 0.8839 0.9402
No log 4.0 188 0.9541 0.0651 0.9541 0.9768
No log 4.0426 190 0.8812 0.0627 0.8812 0.9387
No log 4.0851 192 0.7855 0.0 0.7855 0.8863
No log 4.1277 194 0.8027 0.0249 0.8027 0.8959
No log 4.1702 196 0.7983 0.0297 0.7983 0.8935
No log 4.2128 198 0.7738 -0.0705 0.7738 0.8796
No log 4.2553 200 0.7589 0.1249 0.7589 0.8711
No log 4.2979 202 0.8485 0.1239 0.8485 0.9211
No log 4.3404 204 0.8607 0.0847 0.8607 0.9277
No log 4.3830 206 0.8454 0.1558 0.8454 0.9195
No log 4.4255 208 0.8929 0.0657 0.8929 0.9449
No log 4.4681 210 0.9670 0.0735 0.9670 0.9834
No log 4.5106 212 0.8615 0.0919 0.8615 0.9282
No log 4.5532 214 0.8336 0.1095 0.8336 0.9130
No log 4.5957 216 0.8764 0.0719 0.8764 0.9362
No log 4.6383 218 0.8308 0.0700 0.8308 0.9115
No log 4.6809 220 0.8321 0.1136 0.8321 0.9122
No log 4.7234 222 0.8848 0.0500 0.8848 0.9407
No log 4.7660 224 0.9607 0.0063 0.9607 0.9802
No log 4.8085 226 0.8523 0.0590 0.8523 0.9232
No log 4.8511 228 0.7522 0.0741 0.7522 0.8673
No log 4.8936 230 0.7816 0.0574 0.7816 0.8841
No log 4.9362 232 0.7660 0.1644 0.7660 0.8752
No log 4.9787 234 0.8160 0.0628 0.8160 0.9033
No log 5.0213 236 0.8728 0.0364 0.8728 0.9342
No log 5.0638 238 0.8779 0.0048 0.8779 0.9370
No log 5.1064 240 0.8424 0.1649 0.8424 0.9178
No log 5.1489 242 0.8302 0.1673 0.8302 0.9111
No log 5.1915 244 0.8173 0.1251 0.8173 0.9040
No log 5.2340 246 0.8246 0.0239 0.8246 0.9081
No log 5.2766 248 0.9567 0.1077 0.9567 0.9781
No log 5.3191 250 1.0700 0.1701 1.0700 1.0344
No log 5.3617 252 0.9201 0.0975 0.9201 0.9592
No log 5.4043 254 0.7928 0.1228 0.7928 0.8904
No log 5.4468 256 0.7520 0.0393 0.7520 0.8672
No log 5.4894 258 0.7384 0.0884 0.7384 0.8593
No log 5.5319 260 0.8155 0.1120 0.8155 0.9030
No log 5.5745 262 1.0187 0.1144 1.0187 1.0093
No log 5.6170 264 0.9886 0.1144 0.9886 0.9943
No log 5.6596 266 0.7924 0.1036 0.7924 0.8902
No log 5.7021 268 0.7774 -0.0316 0.7774 0.8817
No log 5.7447 270 0.9155 0.0755 0.9155 0.9568
No log 5.7872 272 0.8471 0.0880 0.8471 0.9204
No log 5.8298 274 0.8201 0.1528 0.8201 0.9056
No log 5.8723 276 0.9987 0.1077 0.9987 0.9993
No log 5.9149 278 1.0143 0.1109 1.0143 1.0071
No log 5.9574 280 0.8831 -0.0035 0.8831 0.9397
No log 6.0 282 0.7963 0.1333 0.7963 0.8923
No log 6.0426 284 0.8072 0.0514 0.8072 0.8984
No log 6.0851 286 0.8641 0.0569 0.8641 0.9296
No log 6.1277 288 0.9249 0.0301 0.9249 0.9617
No log 6.1702 290 0.8810 -0.0099 0.8810 0.9386
No log 6.2128 292 0.8460 -0.0161 0.8460 0.9198
No log 6.2553 294 0.8058 0.1539 0.8058 0.8977
No log 6.2979 296 0.8140 0.1184 0.8140 0.9022
No log 6.3404 298 0.8556 0.0159 0.8556 0.9250
No log 6.3830 300 0.9962 0.0440 0.9962 0.9981
No log 6.4255 302 1.1560 0.1076 1.1560 1.0752
No log 6.4681 304 1.0579 0.1926 1.0579 1.0285
No log 6.5106 306 0.8873 -0.0646 0.8873 0.9420
No log 6.5532 308 0.8605 0.1050 0.8605 0.9276
No log 6.5957 310 0.8263 -0.0734 0.8263 0.9090
No log 6.6383 312 0.9064 0.0378 0.9064 0.9520
No log 6.6809 314 0.8863 0.0275 0.8863 0.9414
No log 6.7234 316 0.8103 0.0056 0.8103 0.9001
No log 6.7660 318 0.7771 0.1675 0.7771 0.8815
No log 6.8085 320 0.8225 0.1660 0.8225 0.9069
No log 6.8511 322 0.8764 0.0793 0.8764 0.9362
No log 6.8936 324 0.7953 0.1901 0.7953 0.8918
No log 6.9362 326 0.7657 0.0437 0.7657 0.8750
No log 6.9787 328 0.9034 0.0767 0.9034 0.9505
No log 7.0213 330 0.9560 0.0783 0.9560 0.9777
No log 7.0638 332 0.8702 -0.0103 0.8702 0.9328
No log 7.1064 334 0.7766 0.1267 0.7766 0.8812
No log 7.1489 336 0.7679 0.1573 0.7679 0.8763
No log 7.1915 338 0.7793 0.1179 0.7793 0.8828
No log 7.2340 340 0.7968 0.0514 0.7968 0.8926
No log 7.2766 342 0.8739 -0.0200 0.8739 0.9348
No log 7.3191 344 0.9055 -0.0484 0.9055 0.9516
No log 7.3617 346 0.8405 -0.0228 0.8405 0.9168
No log 7.4043 348 0.8250 -0.0277 0.8250 0.9083
No log 7.4468 350 0.7819 0.1244 0.7819 0.8842
No log 7.4894 352 0.8158 0.1095 0.8158 0.9032
No log 7.5319 354 0.8422 0.1095 0.8422 0.9177
No log 7.5745 356 0.8114 0.0717 0.8114 0.9008
No log 7.6170 358 0.9499 -0.0381 0.9499 0.9746
No log 7.6596 360 1.0321 0.1312 1.0321 1.0159
No log 7.7021 362 0.9147 -0.0808 0.9147 0.9564
No log 7.7447 364 0.7698 0.1287 0.7698 0.8774
No log 7.7872 366 0.7670 0.1196 0.7670 0.8758
No log 7.8298 368 0.7907 0.1095 0.7907 0.8892
No log 7.8723 370 0.7677 0.1192 0.7677 0.8762
No log 7.9149 372 0.7624 0.1333 0.7624 0.8731
No log 7.9574 374 0.7986 0.0501 0.7986 0.8936
No log 8.0 376 0.7913 0.0898 0.7913 0.8896
No log 8.0426 378 0.7656 0.1187 0.7656 0.8750
No log 8.0851 380 0.8045 0.0408 0.8045 0.8969
No log 8.1277 382 0.8729 0.0230 0.8729 0.9343
No log 8.1702 384 0.8417 0.0110 0.8417 0.9175
No log 8.2128 386 0.8020 0.0725 0.8020 0.8955
No log 8.2553 388 0.8159 0.1431 0.8159 0.9033
No log 8.2979 390 0.8070 0.1518 0.8070 0.8984
No log 8.3404 392 0.8005 0.1141 0.8005 0.8947
No log 8.3830 394 0.7817 0.1141 0.7817 0.8841
No log 8.4255 396 0.7748 0.1141 0.7748 0.8802
No log 8.4681 398 0.7814 0.1141 0.7814 0.8839
No log 8.5106 400 0.8095 0.0377 0.8095 0.8997
No log 8.5532 402 0.8318 -0.1210 0.8318 0.9120
No log 8.5957 404 0.8375 -0.1210 0.8375 0.9152
No log 8.6383 406 0.7797 0.0361 0.7797 0.8830
No log 8.6809 408 0.7827 0.1143 0.7827 0.8847
No log 8.7234 410 0.7897 0.1541 0.7897 0.8887
No log 8.7660 412 0.8135 0.1529 0.8135 0.9019
No log 8.8085 414 0.8697 -0.0634 0.8697 0.9326
No log 8.8511 416 0.9505 -0.0452 0.9505 0.9749
No log 8.8936 418 0.8955 -0.0239 0.8955 0.9463
No log 8.9362 420 0.8813 0.0295 0.8813 0.9388
No log 8.9787 422 0.8451 0.0643 0.8451 0.9193
No log 9.0213 424 0.8328 -0.1224 0.8328 0.9126
No log 9.0638 426 0.8591 0.0185 0.8591 0.9269
No log 9.1064 428 0.8750 -0.0424 0.8750 0.9354
No log 9.1489 430 0.8208 -0.0257 0.8208 0.9060
No log 9.1915 432 0.8081 0.1095 0.8081 0.8990
No log 9.2340 434 0.7913 0.0783 0.7913 0.8896
No log 9.2766 436 0.7935 -0.0268 0.7935 0.8908
No log 9.3191 438 0.8688 -0.0365 0.8688 0.9321
No log 9.3617 440 0.8074 -0.1746 0.8074 0.8985
No log 9.4043 442 0.7509 0.2096 0.7509 0.8665
No log 9.4468 444 0.8486 0.2015 0.8486 0.9212
No log 9.4894 446 0.8622 0.1553 0.8622 0.9286
No log 9.5319 448 0.7912 0.1983 0.7912 0.8895
No log 9.5745 450 0.8533 -0.1021 0.8533 0.9237
No log 9.6170 452 0.9435 0.0286 0.9435 0.9714
No log 9.6596 454 0.9502 0.0900 0.9502 0.9748
No log 9.7021 456 0.8603 -0.0313 0.8603 0.9275
No log 9.7447 458 0.8486 0.0377 0.8486 0.9212
No log 9.7872 460 0.8824 -0.0137 0.8824 0.9393
No log 9.8298 462 0.8838 -0.0846 0.8838 0.9401
No log 9.8723 464 0.8280 -0.0646 0.8280 0.9100
No log 9.9149 466 0.8057 0.0764 0.8057 0.8976
No log 9.9574 468 0.8115 0.0733 0.8115 0.9008
No log 10.0 470 0.8168 0.1475 0.8168 0.9037
No log 10.0426 472 0.8191 0.1465 0.8191 0.9050
No log 10.0851 474 0.8262 0.0741 0.8262 0.9090
No log 10.1277 476 0.8501 0.0087 0.8501 0.9220
No log 10.1702 478 0.8358 0.0434 0.8358 0.9142
No log 10.2128 480 0.8013 0.1139 0.8013 0.8952
No log 10.2553 482 0.8341 0.1395 0.8341 0.9133
No log 10.2979 484 0.8567 0.1243 0.8567 0.9256
No log 10.3404 486 0.7981 0.1395 0.7981 0.8934
No log 10.3830 488 0.7511 0.2547 0.7511 0.8666
No log 10.4255 490 0.7752 -0.0320 0.7752 0.8805
No log 10.4681 492 0.7685 -0.0426 0.7685 0.8766
No log 10.5106 494 0.7596 0.2195 0.7596 0.8716
No log 10.5532 496 0.7943 0.1395 0.7943 0.8913
No log 10.5957 498 0.8205 0.1395 0.8205 0.9058
0.2926 10.6383 500 0.7981 0.1964 0.7981 0.8934
0.2926 10.6809 502 0.8060 0.0810 0.8060 0.8978
0.2926 10.7234 504 0.8438 -0.1355 0.8438 0.9186
0.2926 10.7660 506 0.8251 -0.1464 0.8251 0.9084
0.2926 10.8085 508 0.8167 -0.1464 0.8167 0.9037
0.2926 10.8511 510 0.7756 0.0481 0.7756 0.8807

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k9_task3_organization

Finetuned
(4019)
this model