ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7963
  • Qwk: 0.0783
  • Mse: 0.7963
  • Rmse: 0.8923

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0769 2 4.0471 0.0011 4.0471 2.0117
No log 0.1538 4 2.1387 -0.0296 2.1387 1.4624
No log 0.2308 6 1.0899 -0.0446 1.0899 1.0440
No log 0.3077 8 1.0769 -0.0704 1.0769 1.0377
No log 0.3846 10 0.9415 -0.1980 0.9415 0.9703
No log 0.4615 12 0.9453 -0.0617 0.9453 0.9723
No log 0.5385 14 0.9054 -0.1632 0.9054 0.9515
No log 0.6154 16 0.9593 -0.1612 0.9593 0.9794
No log 0.6923 18 1.0944 -0.0966 1.0944 1.0461
No log 0.7692 20 1.0813 -0.0218 1.0813 1.0398
No log 0.8462 22 0.9571 -0.0799 0.9571 0.9783
No log 0.9231 24 0.7835 0.0191 0.7835 0.8852
No log 1.0 26 0.7803 0.0191 0.7803 0.8834
No log 1.0769 28 0.8061 0.0152 0.8061 0.8979
No log 1.1538 30 0.8561 -0.0295 0.8561 0.9252
No log 1.2308 32 1.1247 0.0557 1.1247 1.0605
No log 1.3077 34 1.6436 0.0558 1.6436 1.2820
No log 1.3846 36 1.5267 -0.0066 1.5267 1.2356
No log 1.4615 38 0.9419 -0.1334 0.9419 0.9705
No log 1.5385 40 0.8990 -0.0912 0.8990 0.9482
No log 1.6154 42 0.8836 -0.0187 0.8836 0.9400
No log 1.6923 44 0.8665 -0.0718 0.8665 0.9309
No log 1.7692 46 1.2125 -0.0228 1.2125 1.1011
No log 1.8462 48 1.1537 -0.0526 1.1537 1.0741
No log 1.9231 50 0.8484 -0.0274 0.8484 0.9211
No log 2.0 52 0.8088 -0.0976 0.8088 0.8994
No log 2.0769 54 0.7956 0.0318 0.7956 0.8920
No log 2.1538 56 0.8725 0.1196 0.8725 0.9341
No log 2.2308 58 0.9474 0.0676 0.9474 0.9734
No log 2.3077 60 0.8368 0.0123 0.8368 0.9148
No log 2.3846 62 0.9160 0.0104 0.9160 0.9571
No log 2.4615 64 0.9748 -0.0204 0.9748 0.9873
No log 2.5385 66 0.9259 0.0514 0.9259 0.9623
No log 2.6154 68 0.9445 0.0733 0.9445 0.9719
No log 2.6923 70 0.9482 -0.0670 0.9482 0.9737
No log 2.7692 72 0.9635 -0.0670 0.9635 0.9816
No log 2.8462 74 1.0172 -0.0306 1.0172 1.0086
No log 2.9231 76 1.0917 0.0116 1.0917 1.0448
No log 3.0 78 1.1604 -0.0045 1.1604 1.0772
No log 3.0769 80 1.0939 -0.0090 1.0939 1.0459
No log 3.1538 82 1.3069 -0.0145 1.3069 1.1432
No log 3.2308 84 1.2012 -0.0550 1.2012 1.0960
No log 3.3077 86 1.0211 0.0036 1.0211 1.0105
No log 3.3846 88 1.0376 -0.0768 1.0376 1.0186
No log 3.4615 90 0.9213 0.0488 0.9213 0.9598
No log 3.5385 92 0.9695 -0.0440 0.9695 0.9846
No log 3.6154 94 0.8765 0.2063 0.8765 0.9362
No log 3.6923 96 0.8648 0.0934 0.8648 0.9299
No log 3.7692 98 0.8890 -0.0209 0.8890 0.9429
No log 3.8462 100 0.9054 0.2024 0.9054 0.9515
No log 3.9231 102 1.2510 0.0028 1.2510 1.1185
No log 4.0 104 1.6450 0.0408 1.6450 1.2826
No log 4.0769 106 1.5976 0.0642 1.5976 1.2640
No log 4.1538 108 1.1726 -0.0651 1.1726 1.0829
No log 4.2308 110 1.0868 -0.1388 1.0868 1.0425
No log 4.3077 112 1.1833 -0.1131 1.1833 1.0878
No log 4.3846 114 0.9503 0.1093 0.9503 0.9748
No log 4.4615 116 1.0344 -0.0269 1.0344 1.0171
No log 4.5385 118 0.9588 -0.0409 0.9588 0.9792
No log 4.6154 120 0.8718 0.0236 0.8718 0.9337
No log 4.6923 122 1.1816 -0.0516 1.1816 1.0870
No log 4.7692 124 1.1898 -0.0492 1.1898 1.0908
No log 4.8462 126 0.9229 0.1440 0.9229 0.9607
No log 4.9231 128 1.1027 0.0824 1.1027 1.0501
No log 5.0 130 1.1050 0.0558 1.1050 1.0512
No log 5.0769 132 0.8667 0.0827 0.8667 0.9309
No log 5.1538 134 1.0520 -0.0486 1.0520 1.0257
No log 5.2308 136 1.0901 -0.0845 1.0901 1.0441
No log 5.3077 138 0.8718 -0.0643 0.8718 0.9337
No log 5.3846 140 0.8398 0.0359 0.8398 0.9164
No log 5.4615 142 0.8863 0.0600 0.8863 0.9414
No log 5.5385 144 1.0203 -0.0828 1.0203 1.0101
No log 5.6154 146 0.9898 -0.0828 0.9898 0.9949
No log 5.6923 148 0.9091 0.0318 0.9091 0.9535
No log 5.7692 150 0.9887 0.1504 0.9887 0.9943
No log 5.8462 152 1.1743 0.0401 1.1743 1.0837
No log 5.9231 154 1.1400 0.0390 1.1400 1.0677
No log 6.0 156 0.9850 0.1367 0.9850 0.9925
No log 6.0769 158 1.0908 0.0257 1.0908 1.0444
No log 6.1538 160 1.0546 -0.0118 1.0546 1.0269
No log 6.2308 162 0.8958 0.0733 0.8958 0.9465
No log 6.3077 164 0.8957 0.0101 0.8957 0.9464
No log 6.3846 166 0.8421 0.1561 0.8421 0.9176
No log 6.4615 168 1.0316 -0.0862 1.0316 1.0157
No log 6.5385 170 1.1644 0.0578 1.1644 1.0791
No log 6.6154 172 0.9229 -0.0425 0.9229 0.9607
No log 6.6923 174 0.8569 0.1034 0.8569 0.9257
No log 6.7692 176 0.9955 -0.0533 0.9955 0.9977
No log 6.8462 178 0.9302 0.0351 0.9302 0.9645
No log 6.9231 180 0.9157 0.0161 0.9157 0.9569
No log 7.0 182 1.0730 -0.0124 1.0730 1.0359
No log 7.0769 184 0.9562 0.0060 0.9562 0.9779
No log 7.1538 186 0.9324 0.0559 0.9324 0.9656
No log 7.2308 188 0.9485 0.0007 0.9485 0.9739
No log 7.3077 190 0.8411 0.0518 0.8411 0.9171
No log 7.3846 192 0.8408 0.0214 0.8408 0.9169
No log 7.4615 194 0.9018 -0.0355 0.9018 0.9496
No log 7.5385 196 0.8601 -0.0316 0.8601 0.9274
No log 7.6154 198 0.8354 0.0183 0.8354 0.9140
No log 7.6923 200 0.8154 0.0269 0.8154 0.9030
No log 7.7692 202 0.8408 0.0757 0.8408 0.9169
No log 7.8462 204 0.8542 0.1236 0.8542 0.9242
No log 7.9231 206 0.8633 0.1139 0.8633 0.9291
No log 8.0 208 0.8865 0.1591 0.8865 0.9415
No log 8.0769 210 0.8951 0.1475 0.8951 0.9461
No log 8.1538 212 0.8789 0.0670 0.8789 0.9375
No log 8.2308 214 0.9091 0.0856 0.9091 0.9535
No log 8.3077 216 0.8809 0.0376 0.8809 0.9386
No log 8.3846 218 0.8828 0.0846 0.8828 0.9396
No log 8.4615 220 0.8970 0.0764 0.8970 0.9471
No log 8.5385 222 0.9504 0.0870 0.9504 0.9749
No log 8.6154 224 0.9710 0.0529 0.9710 0.9854
No log 8.6923 226 0.9157 0.0916 0.9157 0.9569
No log 8.7692 228 0.9050 0.0613 0.9050 0.9513
No log 8.8462 230 0.8422 -0.0103 0.8422 0.9177
No log 8.9231 232 0.8397 -0.0103 0.8397 0.9163
No log 9.0 234 0.8674 -0.0375 0.8674 0.9313
No log 9.0769 236 0.8572 0.0257 0.8572 0.9258
No log 9.1538 238 0.8418 -0.0218 0.8418 0.9175
No log 9.2308 240 0.7946 -0.0030 0.7946 0.8914
No log 9.3077 242 0.7864 0.0513 0.7864 0.8868
No log 9.3846 244 0.7780 -0.0228 0.7780 0.8820
No log 9.4615 246 0.9831 0.0515 0.9831 0.9915
No log 9.5385 248 0.9547 0.0515 0.9547 0.9771
No log 9.6154 250 0.8281 0.0512 0.8281 0.9100
No log 9.6923 252 0.7947 -0.0170 0.7947 0.8915
No log 9.7692 254 0.8371 0.0856 0.8371 0.9149
No log 9.8462 256 0.8353 0.0749 0.8353 0.9139
No log 9.9231 258 0.8859 0.0068 0.8859 0.9412
No log 10.0 260 0.8613 0.0095 0.8613 0.9281
No log 10.0769 262 0.7919 0.0289 0.7919 0.8899
No log 10.1538 264 0.8128 0.1327 0.8128 0.9016
No log 10.2308 266 0.8166 0.0764 0.8166 0.9037
No log 10.3077 268 0.8398 0.1561 0.8398 0.9164
No log 10.3846 270 0.9963 -0.0378 0.9963 0.9982
No log 10.4615 272 1.2418 0.0585 1.2418 1.1144
No log 10.5385 274 1.0671 -0.0128 1.0671 1.0330
No log 10.6154 276 0.8702 0.2290 0.8702 0.9328
No log 10.6923 278 0.8867 0.1891 0.8867 0.9417
No log 10.7692 280 0.8194 0.1138 0.8194 0.9052
No log 10.8462 282 0.9148 -0.0440 0.9148 0.9564
No log 10.9231 284 0.9920 -0.0840 0.9920 0.9960
No log 11.0 286 0.8616 0.0490 0.8616 0.9282
No log 11.0769 288 0.8059 0.1604 0.8059 0.8977
No log 11.1538 290 0.8376 0.0989 0.8376 0.9152
No log 11.2308 292 0.8012 0.1604 0.8012 0.8951
No log 11.3077 294 0.8448 0.1095 0.8448 0.9191
No log 11.3846 296 0.8687 0.0562 0.8687 0.9320
No log 11.4615 298 0.9355 -0.0471 0.9355 0.9672
No log 11.5385 300 0.9037 -0.0054 0.9037 0.9507
No log 11.6154 302 0.8065 0.1146 0.8065 0.8980
No log 11.6923 304 0.7678 0.0776 0.7678 0.8762
No log 11.7692 306 0.7640 -0.0595 0.7640 0.8741
No log 11.8462 308 0.7862 0.0798 0.7862 0.8867
No log 11.9231 310 0.8386 0.1187 0.8386 0.9157
No log 12.0 312 0.9128 0.0652 0.9128 0.9554
No log 12.0769 314 1.0031 -0.0008 1.0031 1.0016
No log 12.1538 316 0.9307 0.0966 0.9307 0.9647
No log 12.2308 318 0.9055 0.1212 0.9055 0.9516
No log 12.3077 320 0.9071 0.1212 0.9071 0.9524
No log 12.3846 322 0.8642 0.1232 0.8642 0.9296
No log 12.4615 324 0.8376 0.0690 0.8376 0.9152
No log 12.5385 326 0.8113 0.0776 0.8113 0.9007
No log 12.6154 328 0.8152 0.0783 0.8152 0.9029
No log 12.6923 330 0.8178 0.0783 0.8178 0.9043
No log 12.7692 332 0.8539 0.1196 0.8539 0.9241
No log 12.8462 334 0.9553 -0.0362 0.9553 0.9774
No log 12.9231 336 0.9509 -0.0315 0.9509 0.9751
No log 13.0 338 0.9161 0.1308 0.9161 0.9571
No log 13.0769 340 0.8849 0.2412 0.8849 0.9407
No log 13.1538 342 0.8681 0.2709 0.8681 0.9317
No log 13.2308 344 0.8403 0.0622 0.8403 0.9167
No log 13.3077 346 0.8170 0.0611 0.8170 0.9039
No log 13.3846 348 0.7860 0.1244 0.7860 0.8866
No log 13.4615 350 0.7759 0.1244 0.7759 0.8809
No log 13.5385 352 0.8069 0.0157 0.8069 0.8983
No log 13.6154 354 0.8033 0.0985 0.8033 0.8963
No log 13.6923 356 0.8109 0.0650 0.8109 0.9005
No log 13.7692 358 1.0277 -0.0496 1.0277 1.0137
No log 13.8462 360 1.1590 0.0673 1.1590 1.0766
No log 13.9231 362 0.9881 0.0267 0.9881 0.9940
No log 14.0 364 0.8426 0.1179 0.8426 0.9179
No log 14.0769 366 1.0286 0.0764 1.0286 1.0142
No log 14.1538 368 1.0608 0.0462 1.0608 1.0299
No log 14.2308 370 0.8701 0.0934 0.8701 0.9328
No log 14.3077 372 0.7604 0.1722 0.7604 0.8720
No log 14.3846 374 0.7489 0.1722 0.7489 0.8654
No log 14.4615 376 0.7609 0.1612 0.7609 0.8723
No log 14.5385 378 0.7431 0.1612 0.7431 0.8620
No log 14.6154 380 0.7221 0.1304 0.7221 0.8497
No log 14.6923 382 0.7189 0.1304 0.7189 0.8479
No log 14.7692 384 0.7327 0.1612 0.7327 0.8560
No log 14.8462 386 0.7730 0.0525 0.7730 0.8792
No log 14.9231 388 0.8681 0.0333 0.8681 0.9317
No log 15.0 390 0.8687 0.0333 0.8687 0.9321
No log 15.0769 392 0.8337 0.0424 0.8337 0.9131
No log 15.1538 394 0.7998 0.0118 0.7998 0.8943
No log 15.2308 396 0.7719 0.0650 0.7719 0.8786
No log 15.3077 398 0.7682 0.1644 0.7682 0.8765
No log 15.3846 400 0.7785 0.0690 0.7785 0.8823
No log 15.4615 402 0.8379 0.0490 0.8379 0.9154
No log 15.5385 404 0.8353 0.0490 0.8353 0.9140
No log 15.6154 406 0.7860 0.1187 0.7860 0.8866
No log 15.6923 408 0.7950 0.1187 0.7950 0.8916
No log 15.7692 410 0.7850 0.1189 0.7850 0.8860
No log 15.8462 412 0.7757 0.1189 0.7757 0.8808
No log 15.9231 414 0.7771 0.0741 0.7771 0.8815
No log 16.0 416 0.7825 0.2078 0.7825 0.8846
No log 16.0769 418 0.7909 0.0741 0.7909 0.8893
No log 16.1538 420 0.8255 0.0822 0.8255 0.9086
No log 16.2308 422 0.8047 0.1189 0.8047 0.8970
No log 16.3077 424 0.8049 0.1599 0.8049 0.8971
No log 16.3846 426 0.8170 0.1599 0.8170 0.9039
No log 16.4615 428 0.8309 0.1599 0.8309 0.9115
No log 16.5385 430 0.8357 0.1591 0.8357 0.9142
No log 16.6154 432 0.8505 0.1942 0.8505 0.9222
No log 16.6923 434 0.8469 0.1591 0.8469 0.9202
No log 16.7692 436 0.8504 0.0741 0.8504 0.9222
No log 16.8462 438 0.8323 0.0741 0.8323 0.9123
No log 16.9231 440 0.8485 0.0119 0.8485 0.9211
No log 17.0 442 0.8501 0.0172 0.8501 0.9220
No log 17.0769 444 0.8208 0.0488 0.8208 0.9060
No log 17.1538 446 0.7819 -0.0567 0.7819 0.8842
No log 17.2308 448 0.7742 0.1675 0.7742 0.8799
No log 17.3077 450 0.8208 0.1146 0.8208 0.9060
No log 17.3846 452 0.8235 0.1146 0.8235 0.9075
No log 17.4615 454 0.8131 0.0341 0.8131 0.9017
No log 17.5385 456 0.8209 -0.0163 0.8209 0.9060
No log 17.6154 458 0.8561 0.0123 0.8561 0.9253
No log 17.6923 460 0.9074 0.0346 0.9074 0.9526
No log 17.7692 462 0.8160 -0.0295 0.8160 0.9033
No log 17.8462 464 0.7634 0.0828 0.7634 0.8737
No log 17.9231 466 0.8193 0.1079 0.8193 0.9052
No log 18.0 468 0.8494 0.1800 0.8494 0.9216
No log 18.0769 470 0.8078 0.0846 0.8078 0.8988
No log 18.1538 472 0.8056 -0.0573 0.8056 0.8976
No log 18.2308 474 0.8216 -0.0573 0.8216 0.9064
No log 18.3077 476 0.8248 -0.0108 0.8248 0.9082
No log 18.3846 478 0.8207 -0.0108 0.8207 0.9059
No log 18.4615 480 0.8037 0.0247 0.8037 0.8965
No log 18.5385 482 0.7554 0.0821 0.7554 0.8691
No log 18.6154 484 0.7490 0.0432 0.7490 0.8654
No log 18.6923 486 0.7752 -0.0366 0.7752 0.8805
No log 18.7692 488 0.7594 -0.0029 0.7594 0.8714
No log 18.8462 490 0.7716 0.1254 0.7716 0.8784
No log 18.9231 492 0.8528 0.0512 0.8528 0.9235
No log 19.0 494 0.8728 0.0490 0.8728 0.9342
No log 19.0769 496 0.8461 0.0341 0.8461 0.9198
No log 19.1538 498 0.9576 0.1259 0.9576 0.9786
0.288 19.2308 500 0.9959 0.0947 0.9959 0.9980
0.288 19.3077 502 0.8993 0.1623 0.8993 0.9483
0.288 19.3846 504 0.8426 0.0341 0.8426 0.9179
0.288 19.4615 506 0.8230 0.1254 0.8230 0.9072
0.288 19.5385 508 0.8039 0.1254 0.8039 0.8966
0.288 19.6154 510 0.7758 0.1254 0.7758 0.8808
0.288 19.6923 512 0.7562 0.1254 0.7562 0.8696
0.288 19.7692 514 0.7657 0.0783 0.7657 0.8750
0.288 19.8462 516 0.7824 0.0783 0.7824 0.8845
0.288 19.9231 518 0.7883 0.0783 0.7883 0.8879
0.288 20.0 520 0.7881 0.0783 0.7881 0.8877
0.288 20.0769 522 0.7963 0.0783 0.7963 0.8923

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task3_organization

Finetuned
(4019)
this model