ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k20_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9675
  • Qwk: -0.0878
  • Mse: 0.9675
  • Rmse: 0.9836

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 3.6495 -0.0051 3.6495 1.9104
No log 0.08 4 2.1092 0.0643 2.1092 1.4523
No log 0.12 6 2.2727 0.0076 2.2727 1.5076
No log 0.16 8 1.4068 0.0045 1.4068 1.1861
No log 0.2 10 1.7743 -0.0431 1.7743 1.3320
No log 0.24 12 1.9622 -0.0241 1.9622 1.4008
No log 0.28 14 0.9517 -0.0143 0.9517 0.9755
No log 0.32 16 0.7105 -0.0695 0.7105 0.8429
No log 0.36 18 0.7815 0.0159 0.7815 0.8840
No log 0.4 20 1.1852 -0.0468 1.1852 1.0887
No log 0.44 22 1.8096 -0.0265 1.8096 1.3452
No log 0.48 24 1.7371 -0.0265 1.7371 1.3180
No log 0.52 26 1.1175 0.0 1.1175 1.0571
No log 0.56 28 0.8423 -0.0852 0.8423 0.9178
No log 0.6 30 0.8201 0.0129 0.8201 0.9056
No log 0.64 32 0.9250 -0.0638 0.9250 0.9618
No log 0.68 34 0.8575 0.0287 0.8575 0.9260
No log 0.72 36 1.0037 -0.0398 1.0037 1.0019
No log 0.76 38 1.0246 -0.0411 1.0246 1.0122
No log 0.8 40 1.2544 -0.0468 1.2544 1.1200
No log 0.84 42 1.3908 -0.0468 1.3908 1.1793
No log 0.88 44 1.4291 0.0048 1.4291 1.1955
No log 0.92 46 1.1245 -0.0446 1.1245 1.0604
No log 0.96 48 0.8784 0.0346 0.8784 0.9372
No log 1.0 50 0.7699 -0.0160 0.7699 0.8774
No log 1.04 52 0.8468 0.0442 0.8468 0.9202
No log 1.08 54 0.8950 0.0233 0.8950 0.9460
No log 1.12 56 0.9382 -0.0545 0.9382 0.9686
No log 1.16 58 1.0168 -0.0930 1.0168 1.0084
No log 1.2 60 1.8268 -0.0241 1.8268 1.3516
No log 1.24 62 2.2176 -0.0263 2.2176 1.4892
No log 1.28 64 1.9999 -0.0247 1.9999 1.4142
No log 1.32 66 1.5452 -0.0221 1.5452 1.2431
No log 1.3600 68 0.8990 -0.0240 0.8990 0.9481
No log 1.4 70 0.8560 -0.0892 0.8560 0.9252
No log 1.44 72 0.8460 -0.0958 0.8460 0.9198
No log 1.48 74 1.0325 -0.1279 1.0325 1.0161
No log 1.52 76 1.7710 -0.0014 1.7710 1.3308
No log 1.56 78 1.8453 -0.0264 1.8453 1.3584
No log 1.6 80 1.2716 -0.0133 1.2716 1.1277
No log 1.6400 82 0.8392 -0.0240 0.8392 0.9161
No log 1.6800 84 0.7946 -0.0520 0.7946 0.8914
No log 1.72 86 0.7793 -0.0035 0.7793 0.8828
No log 1.76 88 0.8384 -0.0331 0.8384 0.9156
No log 1.8 90 1.0952 -0.0677 1.0952 1.0465
No log 1.8400 92 1.2323 -0.0145 1.2323 1.1101
No log 1.88 94 0.9739 0.0233 0.9739 0.9869
No log 1.92 96 0.7798 0.0479 0.7798 0.8831
No log 1.96 98 0.8266 -0.0644 0.8266 0.9092
No log 2.0 100 1.0447 0.0217 1.0447 1.0221
No log 2.04 102 1.3981 -0.0629 1.3981 1.1824
No log 2.08 104 1.1779 -0.1196 1.1779 1.0853
No log 2.12 106 0.9919 -0.0440 0.9919 0.9959
No log 2.16 108 0.9620 -0.1159 0.9620 0.9808
No log 2.2 110 0.8875 0.0303 0.8875 0.9421
No log 2.24 112 0.9224 -0.0699 0.9224 0.9604
No log 2.2800 114 1.2124 -0.1192 1.2124 1.1011
No log 2.32 116 2.0412 -0.0638 2.0412 1.4287
No log 2.36 118 1.9102 -0.1544 1.9102 1.3821
No log 2.4 120 1.2123 -0.0804 1.2123 1.1010
No log 2.44 122 0.9696 -0.1755 0.9696 0.9847
No log 2.48 124 0.9891 -0.0186 0.9891 0.9945
No log 2.52 126 1.1474 -0.0832 1.1474 1.0712
No log 2.56 128 1.0627 -0.0376 1.0627 1.0309
No log 2.6 130 0.9790 -0.1280 0.9790 0.9895
No log 2.64 132 1.0015 -0.1632 1.0015 1.0007
No log 2.68 134 0.9992 -0.0764 0.9992 0.9996
No log 2.7200 136 1.4096 -0.0817 1.4096 1.1873
No log 2.76 138 2.4075 -0.0777 2.4075 1.5516
No log 2.8 140 2.2937 -0.0428 2.2937 1.5145
No log 2.84 142 1.5408 -0.0613 1.5408 1.2413
No log 2.88 144 0.9140 0.0 0.9140 0.9560
No log 2.92 146 0.9453 -0.0648 0.9453 0.9723
No log 2.96 148 0.9584 -0.0989 0.9584 0.9790
No log 3.0 150 0.8693 -0.1331 0.8693 0.9324
No log 3.04 152 0.9807 -0.1671 0.9807 0.9903
No log 3.08 154 1.2231 -0.0285 1.2231 1.1059
No log 3.12 156 1.1537 -0.0269 1.1537 1.0741
No log 3.16 158 0.9401 -0.0801 0.9401 0.9696
No log 3.2 160 0.8492 0.0318 0.8492 0.9215
No log 3.24 162 0.8893 -0.0252 0.8893 0.9430
No log 3.2800 164 0.9604 -0.1148 0.9604 0.9800
No log 3.32 166 0.9267 -0.0723 0.9267 0.9626
No log 3.36 168 0.9277 -0.0723 0.9277 0.9632
No log 3.4 170 0.9176 -0.1121 0.9176 0.9579
No log 3.44 172 0.9266 -0.0731 0.9266 0.9626
No log 3.48 174 1.2122 -0.0175 1.2122 1.1010
No log 3.52 176 1.3550 -0.0561 1.3550 1.1641
No log 3.56 178 1.0982 -0.1175 1.0982 1.0480
No log 3.6 180 0.9053 -0.0274 0.9053 0.9515
No log 3.64 182 0.8024 -0.0493 0.8024 0.8958
No log 3.68 184 0.7654 -0.0033 0.7654 0.8749
No log 3.7200 186 0.8394 0.0129 0.8394 0.9162
No log 3.76 188 1.1515 -0.0345 1.1515 1.0731
No log 3.8 190 1.2446 -0.0348 1.2446 1.1156
No log 3.84 192 0.9898 -0.0474 0.9898 0.9949
No log 3.88 194 0.8088 -0.0125 0.8088 0.8993
No log 3.92 196 0.8252 -0.1111 0.8252 0.9084
No log 3.96 198 0.9458 -0.0767 0.9458 0.9725
No log 4.0 200 0.9540 -0.0767 0.9540 0.9767
No log 4.04 202 0.9412 -0.2270 0.9412 0.9702
No log 4.08 204 1.0059 -0.1026 1.0059 1.0029
No log 4.12 206 1.0500 -0.1088 1.0500 1.0247
No log 4.16 208 1.0118 -0.0495 1.0118 1.0059
No log 4.2 210 0.9713 -0.0915 0.9713 0.9856
No log 4.24 212 0.9251 -0.0879 0.9251 0.9618
No log 4.28 214 0.9659 0.0017 0.9659 0.9828
No log 4.32 216 0.9894 -0.0008 0.9894 0.9947
No log 4.36 218 0.9693 -0.0230 0.9693 0.9845
No log 4.4 220 0.9259 -0.2295 0.9259 0.9622
No log 4.44 222 0.9035 -0.1047 0.9035 0.9505
No log 4.48 224 1.0033 -0.1209 1.0033 1.0017
No log 4.52 226 0.8818 0.0214 0.8818 0.9391
No log 4.5600 228 0.7699 -0.1697 0.7699 0.8775
No log 4.6 230 0.7517 -0.0591 0.7517 0.8670
No log 4.64 232 0.7897 -0.0695 0.7897 0.8886
No log 4.68 234 0.7865 -0.0679 0.7865 0.8868
No log 4.72 236 0.8094 -0.1616 0.8094 0.8996
No log 4.76 238 0.8884 -0.1585 0.8884 0.9426
No log 4.8 240 1.0917 -0.0079 1.0917 1.0449
No log 4.84 242 1.0408 0.0016 1.0408 1.0202
No log 4.88 244 0.9841 -0.0373 0.9841 0.9920
No log 4.92 246 1.0120 -0.0163 1.0120 1.0060
No log 4.96 248 0.9546 0.0377 0.9546 0.9770
No log 5.0 250 0.8735 0.0071 0.8735 0.9346
No log 5.04 252 0.8480 -0.0488 0.8480 0.9209
No log 5.08 254 0.8994 -0.0956 0.8994 0.9483
No log 5.12 256 0.9022 -0.1155 0.9022 0.9498
No log 5.16 258 0.9130 -0.0082 0.9130 0.9555
No log 5.2 260 0.9319 -0.0054 0.9319 0.9653
No log 5.24 262 0.9313 -0.1214 0.9313 0.9650
No log 5.28 264 0.9548 -0.0230 0.9548 0.9771
No log 5.32 266 1.1374 -0.0163 1.1374 1.0665
No log 5.36 268 1.1704 -0.0936 1.1704 1.0818
No log 5.4 270 1.0388 -0.0809 1.0388 1.0192
No log 5.44 272 0.9424 -0.0930 0.9424 0.9708
No log 5.48 274 0.9841 -0.0240 0.9841 0.9920
No log 5.52 276 0.9766 -0.0240 0.9766 0.9882
No log 5.5600 278 0.9714 -0.0240 0.9714 0.9856
No log 5.6 280 0.9883 0.0562 0.9883 0.9941
No log 5.64 282 0.9996 -0.0788 0.9996 0.9998
No log 5.68 284 1.0031 -0.0833 1.0031 1.0016
No log 5.72 286 0.9274 0.0159 0.9274 0.9630
No log 5.76 288 0.9509 -0.0812 0.9509 0.9751
No log 5.8 290 0.9876 -0.0441 0.9876 0.9938
No log 5.84 292 0.8703 -0.0612 0.8703 0.9329
No log 5.88 294 0.8629 -0.0612 0.8629 0.9289
No log 5.92 296 0.8666 0.0282 0.8666 0.9309
No log 5.96 298 0.8930 0.0099 0.8930 0.9450
No log 6.0 300 0.9379 -0.0833 0.9379 0.9684
No log 6.04 302 0.8313 0.0807 0.8313 0.9118
No log 6.08 304 0.8141 -0.1088 0.8141 0.9023
No log 6.12 306 0.8155 -0.0493 0.8155 0.9030
No log 6.16 308 0.8261 0.0260 0.8261 0.9089
No log 6.2 310 0.8889 -0.0371 0.8889 0.9428
No log 6.24 312 0.9383 -0.0823 0.9383 0.9687
No log 6.28 314 0.8998 -0.0295 0.8998 0.9486
No log 6.32 316 0.9676 -0.0833 0.9676 0.9837
No log 6.36 318 1.0628 -0.0163 1.0628 1.0309
No log 6.4 320 0.9380 -0.0371 0.9380 0.9685
No log 6.44 322 0.9040 -0.1795 0.9040 0.9508
No log 6.48 324 0.9428 -0.0811 0.9428 0.9710
No log 6.52 326 0.9231 -0.1146 0.9231 0.9608
No log 6.5600 328 0.9547 -0.0113 0.9547 0.9771
No log 6.6 330 1.0290 0.0038 1.0290 1.0144
No log 6.64 332 0.9702 -0.0138 0.9702 0.9850
No log 6.68 334 0.9949 -0.1303 0.9949 0.9974
No log 6.72 336 0.9961 -0.0956 0.9961 0.9980
No log 6.76 338 0.9377 -0.1939 0.9377 0.9683
No log 6.8 340 1.0063 -0.0442 1.0063 1.0031
No log 6.84 342 1.1540 0.0353 1.1540 1.0742
No log 6.88 344 1.1446 -0.0345 1.1446 1.0699
No log 6.92 346 1.0164 0.0515 1.0164 1.0082
No log 6.96 348 0.9169 -0.1708 0.9169 0.9576
No log 7.0 350 0.9024 -0.1246 0.9024 0.9500
No log 7.04 352 0.9324 -0.0408 0.9324 0.9656
No log 7.08 354 1.0037 -0.0182 1.0037 1.0018
No log 7.12 356 0.9234 -0.0861 0.9234 0.9609
No log 7.16 358 0.8064 -0.0152 0.8064 0.8980
No log 7.2 360 0.8477 -0.0837 0.8477 0.9207
No log 7.24 362 0.8443 -0.0949 0.8443 0.9188
No log 7.28 364 0.7846 -0.1153 0.7846 0.8858
No log 7.32 366 0.7837 -0.1227 0.7837 0.8853
No log 7.36 368 0.8155 0.0260 0.8155 0.9030
No log 7.4 370 0.8622 -0.0309 0.8622 0.9285
No log 7.44 372 0.9390 -0.1261 0.9390 0.9690
No log 7.48 374 0.8917 -0.0351 0.8917 0.9443
No log 7.52 376 0.9025 -0.0351 0.9025 0.9500
No log 7.5600 378 0.8939 -0.0218 0.8939 0.9455
No log 7.6 380 0.9066 -0.1197 0.9066 0.9522
No log 7.64 382 0.9774 0.0207 0.9774 0.9886
No log 7.68 384 1.1487 -0.0067 1.1487 1.0718
No log 7.72 386 1.1468 -0.0084 1.1468 1.0709
No log 7.76 388 0.9294 0.0182 0.9294 0.9640
No log 7.8 390 0.7993 -0.0032 0.7993 0.8941
No log 7.84 392 0.8702 -0.2411 0.8702 0.9328
No log 7.88 394 0.8860 -0.1833 0.8860 0.9413
No log 7.92 396 0.8164 -0.0493 0.8164 0.9036
No log 7.96 398 0.8246 0.0225 0.8246 0.9081
No log 8.0 400 0.9537 -0.0490 0.9537 0.9766
No log 8.04 402 0.9146 -0.0390 0.9146 0.9563
No log 8.08 404 0.8039 -0.1100 0.8039 0.8966
No log 8.12 406 0.8244 -0.0493 0.8244 0.9080
No log 8.16 408 0.8587 -0.0958 0.8587 0.9267
No log 8.2 410 0.8370 -0.0062 0.8370 0.9149
No log 8.24 412 0.8431 -0.1168 0.8431 0.9182
No log 8.28 414 0.9076 0.0071 0.9076 0.9527
No log 8.32 416 0.8998 0.0043 0.8998 0.9486
No log 8.36 418 0.8275 -0.0179 0.8275 0.9096
No log 8.4 420 0.7927 -0.1100 0.7927 0.8903
No log 8.44 422 0.7889 -0.0152 0.7889 0.8882
No log 8.48 424 0.8087 -0.0264 0.8087 0.8993
No log 8.52 426 0.8346 -0.0778 0.8346 0.9136
No log 8.56 428 0.8916 -0.0442 0.8916 0.9442
No log 8.6 430 0.9303 -0.0122 0.9303 0.9645
No log 8.64 432 0.8834 -0.1197 0.8834 0.9399
No log 8.68 434 0.8952 -0.1748 0.8952 0.9462
No log 8.72 436 0.9262 -0.0608 0.9262 0.9624
No log 8.76 438 0.9223 -0.1120 0.9223 0.9604
No log 8.8 440 0.9072 -0.0573 0.9072 0.9525
No log 8.84 442 0.8854 -0.0138 0.8854 0.9410
No log 8.88 444 0.8674 -0.0228 0.8674 0.9313
No log 8.92 446 0.9067 -0.0842 0.9067 0.9522
No log 8.96 448 0.9057 -0.0474 0.9057 0.9517
No log 9.0 450 0.9052 -0.0518 0.9052 0.9514
No log 9.04 452 0.9039 -0.0143 0.9039 0.9507
No log 9.08 454 0.8434 -0.0079 0.8434 0.9184
No log 9.12 456 0.8063 -0.0801 0.8063 0.8979
No log 9.16 458 0.7807 0.0214 0.7807 0.8836
No log 9.2 460 0.7938 0.0600 0.7938 0.8910
No log 9.24 462 0.8782 0.0260 0.8782 0.9371
No log 9.28 464 0.9027 0.0486 0.9027 0.9501
No log 9.32 466 0.9263 0.0458 0.9263 0.9625
No log 9.36 468 0.8694 0.0233 0.8694 0.9324
No log 9.4 470 0.8048 -0.0390 0.8048 0.8971
No log 9.44 472 0.8146 -0.0704 0.8146 0.9026
No log 9.48 474 0.8111 0.0863 0.8111 0.9006
No log 9.52 476 0.7948 -0.0240 0.7948 0.8915
No log 9.56 478 0.8730 -0.0033 0.8730 0.9344
No log 9.6 480 0.9432 -0.0079 0.9432 0.9712
No log 9.64 482 0.8810 -0.0459 0.8810 0.9386
No log 9.68 484 0.7997 -0.0240 0.7997 0.8943
No log 9.72 486 0.7835 -0.0188 0.7835 0.8851
No log 9.76 488 0.8042 -0.0287 0.8042 0.8968
No log 9.8 490 0.8778 -0.0474 0.8778 0.9369
No log 9.84 492 0.8701 -0.0474 0.8701 0.9328
No log 9.88 494 0.7879 0.0159 0.7879 0.8876
No log 9.92 496 0.7623 0.0436 0.7623 0.8731
No log 9.96 498 0.8188 -0.0428 0.8188 0.9049
0.339 10.0 500 0.8238 -0.0428 0.8238 0.9076
0.339 10.04 502 0.7815 0.0436 0.7815 0.8840
0.339 10.08 504 0.7825 -0.0188 0.7825 0.8846
0.339 10.12 506 0.8220 0.0043 0.8220 0.9066
0.339 10.16 508 0.8374 -0.0033 0.8374 0.9151
0.339 10.2 510 0.8287 -0.0474 0.8287 0.9103
0.339 10.24 512 0.7752 0.1202 0.7752 0.8805
0.339 10.28 514 0.7384 -0.0160 0.7384 0.8593
0.339 10.32 516 0.7308 0.0964 0.7308 0.8549
0.339 10.36 518 0.7230 0.0334 0.7230 0.8503
0.339 10.4 520 0.7807 -0.0459 0.7807 0.8836
0.339 10.44 522 0.9597 0.0305 0.9597 0.9796
0.339 10.48 524 1.0517 0.0446 1.0517 1.0255
0.339 10.52 526 0.9465 0.0623 0.9465 0.9729
0.339 10.56 528 0.7741 -0.0408 0.7741 0.8798
0.339 10.6 530 0.7801 -0.0091 0.7801 0.8832
0.339 10.64 532 0.7995 0.0058 0.7995 0.8942
0.339 10.68 534 0.7805 0.0318 0.7805 0.8834
0.339 10.72 536 0.7987 -0.0309 0.7987 0.8937
0.339 10.76 538 0.8678 -0.0101 0.8678 0.9316
0.339 10.8 540 0.8944 -0.0143 0.8944 0.9457
0.339 10.84 542 0.9077 -0.0571 0.9077 0.9527
0.339 10.88 544 0.8493 -0.0079 0.8493 0.9216
0.339 10.92 546 0.8144 0.0017 0.8144 0.9025
0.339 10.96 548 0.7896 0.0043 0.7896 0.8886
0.339 11.0 550 0.7774 -0.0371 0.7774 0.8817
0.339 11.04 552 0.7769 0.0562 0.7769 0.8814
0.339 11.08 554 0.7949 -0.0283 0.7949 0.8916
0.339 11.12 556 0.8214 -0.1088 0.8214 0.9063
0.339 11.16 558 0.8345 -0.0303 0.8345 0.9135
0.339 11.2 560 0.8279 0.0260 0.8279 0.9099
0.339 11.24 562 0.8256 -0.0351 0.8256 0.9086
0.339 11.28 564 0.8209 0.0017 0.8209 0.9061
0.339 11.32 566 0.8525 -0.0861 0.8525 0.9233
0.339 11.36 568 0.8289 -0.0425 0.8289 0.9104
0.339 11.4 570 0.7860 -0.0264 0.7860 0.8865
0.339 11.44 572 0.8081 -0.0188 0.8081 0.8989
0.339 11.48 574 0.8449 -0.1249 0.8449 0.9192
0.339 11.52 576 0.9324 0.0250 0.9324 0.9656
0.339 11.56 578 1.0842 0.0329 1.0842 1.0413
0.339 11.6 580 1.0656 -0.0269 1.0656 1.0323
0.339 11.64 582 0.9374 0.0017 0.9374 0.9682
0.339 11.68 584 0.8703 -0.1033 0.8703 0.9329
0.339 11.72 586 0.8656 -0.0493 0.8656 0.9304
0.339 11.76 588 0.8897 -0.0506 0.8897 0.9432
0.339 11.8 590 0.9305 -0.1847 0.9305 0.9646
0.339 11.84 592 0.9675 -0.0878 0.9675 0.9836

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k20_task3_organization

Finetuned
(4019)
this model