ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k8_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9619
  • Qwk: -0.1224
  • Mse: 0.9619
  • Rmse: 0.9808

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1 2 3.5026 -0.0354 3.5026 1.8715
No log 0.2 4 1.8202 0.0737 1.8202 1.3492
No log 0.3 6 1.1473 -0.1015 1.1473 1.0711
No log 0.4 8 1.0014 -0.0595 1.0014 1.0007
No log 0.5 10 0.7662 0.0159 0.7662 0.8753
No log 0.6 12 0.9721 -0.0606 0.9721 0.9859
No log 0.7 14 1.3685 0.0 1.3685 1.1698
No log 0.8 16 1.4891 0.0 1.4891 1.2203
No log 0.9 18 1.4558 0.0 1.4558 1.2065
No log 1.0 20 1.3516 -0.0247 1.3516 1.1626
No log 1.1 22 1.2995 -0.0221 1.2995 1.1400
No log 1.2 24 1.0748 -0.0961 1.0748 1.0367
No log 1.3 26 0.9892 -0.2364 0.9892 0.9946
No log 1.4 28 1.0093 -0.1609 1.0093 1.0047
No log 1.5 30 1.0983 -0.1292 1.0983 1.0480
No log 1.6 32 1.1614 -0.0457 1.1614 1.0777
No log 1.7 34 1.0374 -0.1287 1.0374 1.0185
No log 1.8 36 1.1010 -0.0712 1.1010 1.0493
No log 1.9 38 1.2095 -0.0457 1.2095 1.0998
No log 2.0 40 1.1977 -0.0149 1.1977 1.0944
No log 2.1 42 1.0194 -0.0285 1.0194 1.0097
No log 2.2 44 0.9584 -0.0861 0.9584 0.9790
No log 2.3 46 0.9547 0.0676 0.9547 0.9771
No log 2.4 48 1.0534 -0.1281 1.0534 1.0263
No log 2.5 50 1.1992 -0.0695 1.1992 1.0951
No log 2.6 52 1.2839 0.0152 1.2839 1.1331
No log 2.7 54 1.5710 0.0690 1.5710 1.2534
No log 2.8 56 1.1618 -0.0628 1.1618 1.0779
No log 2.9 58 1.1595 -0.0628 1.1595 1.0768
No log 3.0 60 0.9127 0.0409 0.9127 0.9554
No log 3.1 62 0.8508 0.0909 0.8508 0.9224
No log 3.2 64 0.8096 0.1291 0.8096 0.8998
No log 3.3 66 0.8012 0.0549 0.8012 0.8951
No log 3.4 68 1.0609 0.0735 1.0609 1.0300
No log 3.5 70 0.9832 0.1145 0.9832 0.9916
No log 3.6 72 0.8177 0.0257 0.8177 0.9042
No log 3.7 74 1.0919 0.0707 1.0919 1.0449
No log 3.8 76 1.2515 -0.0353 1.2515 1.1187
No log 3.9 78 1.1727 -0.0288 1.1727 1.0829
No log 4.0 80 0.8557 -0.0560 0.8557 0.9251
No log 4.1 82 0.8672 -0.0560 0.8672 0.9312
No log 4.2 84 1.0731 -0.0862 1.0731 1.0359
No log 4.3 86 1.2800 -0.0629 1.2800 1.1314
No log 4.4 88 1.0361 -0.1589 1.0361 1.0179
No log 4.5 90 1.0570 -0.1214 1.0570 1.0281
No log 4.6 92 1.2468 -0.1468 1.2468 1.1166
No log 4.7 94 1.1077 -0.1224 1.1077 1.0525
No log 4.8 96 0.9528 -0.1255 0.9528 0.9761
No log 4.9 98 0.8477 -0.0541 0.8477 0.9207
No log 5.0 100 0.8328 0.0374 0.8328 0.9126
No log 5.1 102 1.2754 0.0488 1.2754 1.1293
No log 5.2 104 1.6608 0.0502 1.6608 1.2887
No log 5.3 106 1.1961 -0.0953 1.1961 1.0937
No log 5.4 108 0.8787 -0.1668 0.8787 0.9374
No log 5.5 110 0.9075 -0.1191 0.9075 0.9526
No log 5.6 112 1.0789 -0.1230 1.0789 1.0387
No log 5.7 114 0.9707 -0.1618 0.9707 0.9853
No log 5.8 116 0.8886 -0.0560 0.8886 0.9427
No log 5.9 118 0.9173 -0.1542 0.9173 0.9578
No log 6.0 120 0.9518 -0.2511 0.9518 0.9756
No log 6.1 122 1.1052 -0.0777 1.1052 1.0513
No log 6.2 124 0.9347 -0.1121 0.9347 0.9668
No log 6.3 126 0.9285 -0.0560 0.9285 0.9636
No log 6.4 128 0.9069 -0.0578 0.9069 0.9523
No log 6.5 130 1.1948 -0.1905 1.1948 1.0931
No log 6.6 132 1.3958 0.0199 1.3958 1.1814
No log 6.7 134 0.9849 -0.2448 0.9849 0.9924
No log 6.8 136 0.8507 -0.0179 0.8507 0.9223
No log 6.9 138 0.9272 -0.2468 0.9272 0.9629
No log 7.0 140 1.0051 -0.2021 1.0051 1.0025
No log 7.1 142 1.0243 -0.2021 1.0243 1.0121
No log 7.2 144 1.0701 -0.2351 1.0701 1.0345
No log 7.3 146 1.1915 -0.1988 1.1915 1.0916
No log 7.4 148 1.0139 -0.2386 1.0139 1.0069
No log 7.5 150 1.1031 -0.2810 1.1031 1.0503
No log 7.6 152 1.0085 -0.1695 1.0085 1.0042
No log 7.7 154 0.8440 -0.1094 0.8440 0.9187
No log 7.8 156 0.8369 -0.1088 0.8369 0.9148
No log 7.9 158 0.9448 -0.1715 0.9448 0.9720
No log 8.0 160 1.4682 -0.0927 1.4682 1.2117
No log 8.1 162 1.3773 -0.0942 1.3773 1.1736
No log 8.2 164 0.9293 -0.2116 0.9293 0.9640
No log 8.3 166 0.9073 -0.0949 0.9073 0.9525
No log 8.4 168 0.8871 -0.0949 0.8871 0.9419
No log 8.5 170 0.8737 -0.1594 0.8737 0.9347
No log 8.6 172 1.0454 -0.1977 1.0454 1.0225
No log 8.7 174 0.9607 -0.1194 0.9607 0.9801
No log 8.8 176 0.8901 -0.0252 0.8901 0.9435
No log 8.9 178 0.7941 -0.0152 0.7941 0.8911
No log 9.0 180 0.8055 -0.0240 0.8055 0.8975
No log 9.1 182 0.8461 -0.1187 0.8461 0.9198
No log 9.2 184 0.9098 -0.1253 0.9098 0.9538
No log 9.3 186 1.0147 -0.1267 1.0147 1.0073
No log 9.4 188 0.8699 -0.0331 0.8699 0.9327
No log 9.5 190 0.7930 0.0541 0.7930 0.8905
No log 9.6 192 0.7766 -0.1153 0.7766 0.8813
No log 9.7 194 0.8516 0.0191 0.8516 0.9228
No log 9.8 196 0.9313 -0.1255 0.9313 0.9650
No log 9.9 198 1.1797 0.0045 1.1797 1.0861
No log 10.0 200 0.9809 -0.1259 0.9809 0.9904
No log 10.1 202 0.8022 -0.1163 0.8022 0.8956
No log 10.2 204 0.8028 -0.1163 0.8028 0.8960
No log 10.3 206 0.8813 -0.0801 0.8813 0.9388
No log 10.4 208 1.1895 -0.0306 1.1895 1.0907
No log 10.5 210 1.0984 -0.1274 1.0984 1.0480
No log 10.6 212 0.8789 -0.1158 0.8789 0.9375
No log 10.7 214 1.0165 -0.0801 1.0165 1.0082
No log 10.8 216 0.9739 -0.0678 0.9739 0.9869
No log 10.9 218 0.8834 -0.0675 0.8834 0.9399
No log 11.0 220 1.5505 -0.0443 1.5505 1.2452
No log 11.1 222 2.0996 -0.0490 2.0996 1.4490
No log 11.2 224 1.8107 -0.0480 1.8107 1.3456
No log 11.3 226 1.0764 0.0404 1.0764 1.0375
No log 11.4 228 0.7737 -0.0675 0.7737 0.8796
No log 11.5 230 0.7566 -0.0065 0.7566 0.8698
No log 11.6 232 0.7654 -0.0660 0.7654 0.8749
No log 11.7 234 0.8473 -0.1200 0.8473 0.9205
No log 11.8 236 0.9937 0.1114 0.9937 0.9968
No log 11.9 238 0.9204 -0.0545 0.9204 0.9594
No log 12.0 240 0.8627 -0.2033 0.8627 0.9288
No log 12.1 242 0.8137 -0.1180 0.8137 0.9020
No log 12.2 244 0.8662 -0.1644 0.8662 0.9307
No log 12.3 246 0.9594 -0.1601 0.9594 0.9795
No log 12.4 248 0.9388 -0.1605 0.9388 0.9689
No log 12.5 250 0.8592 -0.1623 0.8592 0.9269
No log 12.6 252 0.7720 -0.0179 0.7720 0.8786
No log 12.7 254 0.7600 -0.0179 0.7600 0.8718
No log 12.8 256 0.8460 -0.0799 0.8460 0.9198
No log 12.9 258 0.9784 -0.1228 0.9784 0.9892
No log 13.0 260 0.9677 -0.0845 0.9677 0.9837
No log 13.1 262 0.8887 -0.1206 0.8887 0.9427
No log 13.2 264 0.8272 -0.0704 0.8272 0.9095
No log 13.3 266 0.8652 -0.1197 0.8652 0.9302
No log 13.4 268 0.9738 -0.1214 0.9738 0.9868
No log 13.5 270 0.9615 -0.1605 0.9615 0.9806
No log 13.6 272 0.8420 -0.0179 0.8420 0.9176
No log 13.7 274 0.8558 0.0 0.8558 0.9251
No log 13.8 276 0.8637 0.0 0.8637 0.9293
No log 13.9 278 0.8974 -0.1184 0.8974 0.9473
No log 14.0 280 0.9745 -0.0390 0.9745 0.9871
No log 14.1 282 0.9567 -0.0842 0.9567 0.9781
No log 14.2 284 0.8547 -0.1184 0.8547 0.9245
No log 14.3 286 0.8384 -0.1172 0.8384 0.9156
No log 14.4 288 0.8753 -0.0295 0.8753 0.9356
No log 14.5 290 0.9589 -0.1265 0.9589 0.9792
No log 14.6 292 0.9432 -0.1601 0.9432 0.9712
No log 14.7 294 0.9182 -0.1605 0.9182 0.9582
No log 14.8 296 0.8898 -0.0295 0.8898 0.9433
No log 14.9 298 0.8746 -0.1656 0.8746 0.9352
No log 15.0 300 0.8813 -0.0731 0.8813 0.9388
No log 15.1 302 0.9196 -0.1200 0.9196 0.9589
No log 15.2 304 0.8952 -0.0295 0.8952 0.9461
No log 15.3 306 0.8869 -0.0295 0.8869 0.9418
No log 15.4 308 0.8921 0.0099 0.8921 0.9445
No log 15.5 310 0.9049 -0.0316 0.9049 0.9512
No log 15.6 312 0.8920 -0.0690 0.8920 0.9445
No log 15.7 314 0.9263 -0.0316 0.9263 0.9625
No log 15.8 316 0.9419 -0.0295 0.9419 0.9705
No log 15.9 318 1.0363 -0.1988 1.0363 1.0180
No log 16.0 320 1.2354 -0.1245 1.2354 1.1115
No log 16.1 322 1.2467 -0.0942 1.2467 1.1166
No log 16.2 324 1.0794 -0.1921 1.0794 1.0390
No log 16.3 326 0.9624 -0.1998 0.9624 0.9810
No log 16.4 328 0.9833 -0.1913 0.9833 0.9916
No log 16.5 330 1.0693 -0.0668 1.0693 1.0341
No log 16.6 332 1.0543 -0.0372 1.0543 1.0268
No log 16.7 334 0.8671 -0.1270 0.8671 0.9312
No log 16.8 336 0.7371 0.0857 0.7371 0.8586
No log 16.9 338 0.8539 -0.0717 0.8539 0.9241
No log 17.0 340 0.8976 -0.1159 0.8976 0.9474
No log 17.1 342 0.9026 -0.0686 0.9026 0.9500
No log 17.2 344 1.0366 -0.1541 1.0366 1.0181
No log 17.3 346 1.0784 -0.1541 1.0784 1.0385
No log 17.4 348 0.9585 -0.0699 0.9585 0.9790
No log 17.5 350 0.9290 -0.2077 0.9290 0.9639
No log 17.6 352 0.8993 -0.1094 0.8993 0.9483
No log 17.7 354 0.8928 -0.1623 0.8928 0.9449
No log 17.8 356 1.0980 -0.0686 1.0980 1.0479
No log 17.9 358 1.2145 -0.0435 1.2145 1.1020
No log 18.0 360 1.1151 -0.0117 1.1151 1.0560
No log 18.1 362 0.8778 -0.0870 0.8778 0.9369
No log 18.2 364 0.7991 -0.0204 0.7991 0.8939
No log 18.3 366 0.7953 -0.0644 0.7953 0.8918
No log 18.4 368 0.8183 -0.0675 0.8183 0.9046
No log 18.5 370 0.8258 -0.0675 0.8258 0.9088
No log 18.6 372 0.8315 -0.1172 0.8315 0.9118
No log 18.7 374 0.8369 -0.0731 0.8369 0.9148
No log 18.8 376 0.8744 -0.0778 0.8744 0.9351
No log 18.9 378 0.9263 -0.1593 0.9263 0.9624
No log 19.0 380 0.8946 -0.1206 0.8946 0.9459
No log 19.1 382 0.8602 -0.2093 0.8602 0.9275
No log 19.2 384 0.8703 -0.1604 0.8703 0.9329
No log 19.3 386 0.8623 -0.1604 0.8623 0.9286
No log 19.4 388 0.8659 -0.1172 0.8659 0.9305
No log 19.5 390 0.9083 -0.1206 0.9083 0.9530
No log 19.6 392 1.0777 -0.1279 1.0777 1.0381
No log 19.7 394 1.1298 -0.0966 1.1298 1.0629
No log 19.8 396 1.0345 -0.1921 1.0345 1.0171
No log 19.9 398 0.9853 -0.1597 0.9853 0.9926
No log 20.0 400 1.0069 -0.1967 1.0069 1.0035
No log 20.1 402 1.0841 -0.1569 1.0841 1.0412
No log 20.2 404 1.0277 -0.1572 1.0277 1.0138
No log 20.3 406 0.9518 -0.1998 0.9518 0.9756
No log 20.4 408 0.9429 -0.1998 0.9429 0.9710
No log 20.5 410 0.9375 -0.1671 0.9375 0.9683
No log 20.6 412 0.8945 -0.1638 0.8945 0.9458
No log 20.7 414 0.8909 -0.0690 0.8909 0.9439
No log 20.8 416 0.8941 -0.1610 0.8941 0.9456
No log 20.9 418 0.9065 -0.0704 0.9065 0.9521
No log 21.0 420 0.9824 -0.1605 0.9824 0.9911
No log 21.1 422 1.0377 -0.0943 1.0377 1.0187
No log 21.2 424 0.9951 -0.0930 0.9951 0.9976
No log 21.3 426 0.8615 -0.1676 0.8615 0.9282
No log 21.4 428 0.8000 0.0914 0.8000 0.8944
No log 21.5 430 0.8058 0.0914 0.8058 0.8976
No log 21.6 432 0.8651 -0.1671 0.8651 0.9301
No log 21.7 434 0.9241 -0.1655 0.9241 0.9613
No log 21.8 436 0.9396 -0.1655 0.9396 0.9693
No log 21.9 438 0.9685 -0.1589 0.9685 0.9841
No log 22.0 440 0.9778 -0.1977 0.9778 0.9889
No log 22.1 442 0.9798 -0.1977 0.9798 0.9898
No log 22.2 444 1.0024 -0.1601 1.0024 1.0012
No log 22.3 446 1.0149 -0.1605 1.0149 1.0074
No log 22.4 448 1.0276 -0.1605 1.0276 1.0137
No log 22.5 450 1.0193 -0.1214 1.0193 1.0096
No log 22.6 452 1.0060 -0.1589 1.0060 1.0030
No log 22.7 454 0.9345 -0.1593 0.9345 0.9667
No log 22.8 456 0.8978 -0.0861 0.8978 0.9475
No log 22.9 458 0.8507 -0.1257 0.8507 0.9223
No log 23.0 460 0.8251 -0.0274 0.8251 0.9083
No log 23.1 462 0.8326 -0.0274 0.8326 0.9125
No log 23.2 464 0.9210 -0.0828 0.9210 0.9597
No log 23.3 466 0.9748 -0.1224 0.9748 0.9873
No log 23.4 468 0.9645 -0.1212 0.9645 0.9821
No log 23.5 470 0.9479 -0.1623 0.9479 0.9736
No log 23.6 472 0.9042 -0.1187 0.9042 0.9509
No log 23.7 474 0.9100 -0.1628 0.9100 0.9539
No log 23.8 476 0.9448 -0.1209 0.9448 0.9720
No log 23.9 478 1.0447 -0.1230 1.0447 1.0221
No log 24.0 480 1.0990 -0.0931 1.0990 1.0483
No log 24.1 482 1.0914 -0.0657 1.0914 1.0447
No log 24.2 484 0.9898 -0.1238 0.9898 0.9949
No log 24.3 486 0.9424 -0.1589 0.9424 0.9708
No log 24.4 488 0.9125 -0.1605 0.9125 0.9552
No log 24.5 490 0.9173 -0.1605 0.9173 0.9578
No log 24.6 492 0.9262 -0.2021 0.9262 0.9624
No log 24.7 494 1.0065 -0.1273 1.0065 1.0033
No log 24.8 496 1.1876 -0.0982 1.1876 1.0898
No log 24.9 498 1.2856 -0.0423 1.2856 1.1338
0.2331 25.0 500 1.2859 -0.0728 1.2859 1.1340
0.2331 25.1 502 1.2544 -0.0728 1.2544 1.1200
0.2331 25.2 504 1.1484 -0.0972 1.1484 1.0716
0.2331 25.3 506 0.9963 -0.1224 0.9963 0.9982
0.2331 25.4 508 0.9563 -0.1593 0.9563 0.9779
0.2331 25.5 510 0.9619 -0.1224 0.9619 0.9808

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k8_task3_organization

Finetuned
(4019)
this model