ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k2_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7984
  • Qwk: 0.0851
  • Mse: 0.7984
  • Rmse: 0.8935

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.25 2 3.7209 0.0118 3.7209 1.9290
No log 0.5 4 1.9475 0.0560 1.9475 1.3955
No log 0.75 6 1.6570 0.0136 1.6570 1.2872
No log 1.0 8 1.2115 0.0100 1.2115 1.1007
No log 1.25 10 0.9932 -0.1624 0.9932 0.9966
No log 1.5 12 1.1768 -0.0101 1.1768 1.0848
No log 1.75 14 0.7510 0.0628 0.7510 0.8666
No log 2.0 16 0.7351 -0.1233 0.7351 0.8574
No log 2.25 18 0.9550 0.0545 0.9550 0.9773
No log 2.5 20 0.9581 0.0026 0.9581 0.9788
No log 2.75 22 0.9605 -0.0423 0.9605 0.9800
No log 3.0 24 1.1308 -0.0247 1.1308 1.0634
No log 3.25 26 1.1750 0.0065 1.1750 1.0840
No log 3.5 28 0.8064 0.1899 0.8064 0.8980
No log 3.75 30 0.7501 -0.0160 0.7501 0.8661
No log 4.0 32 1.2162 -0.0468 1.2162 1.1028
No log 4.25 34 1.7732 -0.0029 1.7732 1.3316
No log 4.5 36 1.9219 -0.0029 1.9219 1.3863
No log 4.75 38 1.6727 0.0 1.6727 1.2933
No log 5.0 40 1.2529 0.0 1.2529 1.1193
No log 5.25 42 0.8982 -0.0936 0.8982 0.9477
No log 5.5 44 0.7598 0.0964 0.7598 0.8716
No log 5.75 46 0.7585 0.0460 0.7585 0.8709
No log 6.0 48 0.7838 -0.0131 0.7838 0.8853
No log 6.25 50 0.8136 -0.0287 0.8136 0.9020
No log 6.5 52 1.0144 -0.0398 1.0144 1.0072
No log 6.75 54 1.1666 -0.0193 1.1666 1.0801
No log 7.0 56 1.0772 -0.0133 1.0772 1.0379
No log 7.25 58 0.8408 -0.1676 0.8408 0.9169
No log 7.5 60 0.7085 -0.0035 0.7085 0.8417
No log 7.75 62 0.6858 0.0 0.6858 0.8281
No log 8.0 64 0.6958 -0.0035 0.6958 0.8341
No log 8.25 66 0.7584 0.1379 0.7584 0.8708
No log 8.5 68 0.9424 -0.0949 0.9424 0.9708
No log 8.75 70 1.0934 -0.0234 1.0934 1.0457
No log 9.0 72 1.0323 -0.0117 1.0323 1.0160
No log 9.25 74 0.8805 -0.1263 0.8805 0.9383
No log 9.5 76 0.7684 0.0334 0.7684 0.8766
No log 9.75 78 0.8021 0.0159 0.8021 0.8956
No log 10.0 80 0.8153 0.0159 0.8153 0.9029
No log 10.25 82 0.7930 0.0670 0.7930 0.8905
No log 10.5 84 0.7641 -0.0188 0.7641 0.8741
No log 10.75 86 0.8218 0.0588 0.8218 0.9065
No log 11.0 88 0.9160 -0.0870 0.9160 0.9571
No log 11.25 90 0.8596 -0.0442 0.8596 0.9271
No log 11.5 92 0.8484 0.0909 0.8484 0.9211
No log 11.75 94 0.9317 -0.0049 0.9317 0.9653
No log 12.0 96 1.0658 0.0192 1.0658 1.0324
No log 12.25 98 1.4523 -0.1016 1.4523 1.2051
No log 12.5 100 1.3229 -0.0971 1.3229 1.1502
No log 12.75 102 1.1497 0.0378 1.1497 1.0722
No log 13.0 104 1.1049 -0.0331 1.1049 1.0511
No log 13.25 106 0.9343 0.1094 0.9343 0.9666
No log 13.5 108 1.0832 -0.0159 1.0832 1.0408
No log 13.75 110 1.0502 -0.0028 1.0502 1.0248
No log 14.0 112 1.0042 0.0755 1.0042 1.0021
No log 14.25 114 1.0826 -0.1013 1.0826 1.0405
No log 14.5 116 0.9842 -0.0047 0.9842 0.9920
No log 14.75 118 1.0163 0.0392 1.0163 1.0081
No log 15.0 120 0.9586 0.1048 0.9586 0.9791
No log 15.25 122 1.0940 -0.0370 1.0940 1.0459
No log 15.5 124 1.1408 -0.0291 1.1408 1.0681
No log 15.75 126 1.0289 -0.0425 1.0289 1.0144
No log 16.0 128 1.0754 0.0227 1.0754 1.0370
No log 16.25 130 1.0882 -0.0524 1.0882 1.0432
No log 16.5 132 1.0718 -0.0611 1.0718 1.0353
No log 16.75 134 1.0432 -0.0152 1.0432 1.0214
No log 17.0 136 1.1566 -0.0842 1.1566 1.0755
No log 17.25 138 1.0554 -0.0777 1.0554 1.0273
No log 17.5 140 0.9147 -0.1166 0.9147 0.9564
No log 17.75 142 1.0328 -0.0014 1.0328 1.0163
No log 18.0 144 0.9872 -0.0148 0.9872 0.9936
No log 18.25 146 0.9142 0.0327 0.9142 0.9562
No log 18.5 148 0.9805 0.0391 0.9805 0.9902
No log 18.75 150 0.9857 0.0023 0.9857 0.9928
No log 19.0 152 1.0408 -0.0137 1.0408 1.0202
No log 19.25 154 1.1461 0.0455 1.1461 1.0706
No log 19.5 156 1.0635 -0.0842 1.0635 1.0313
No log 19.75 158 0.9379 -0.0623 0.9379 0.9684
No log 20.0 160 0.9704 0.0336 0.9704 0.9851
No log 20.25 162 0.9976 0.0989 0.9976 0.9988
No log 20.5 164 0.9073 -0.0024 0.9073 0.9525
No log 20.75 166 0.9797 -0.1073 0.9797 0.9898
No log 21.0 168 0.9440 -0.0439 0.9440 0.9716
No log 21.25 170 0.8577 0.0834 0.8577 0.9261
No log 21.5 172 1.0175 -0.0056 1.0175 1.0087
No log 21.75 174 1.0382 -0.0440 1.0382 1.0189
No log 22.0 176 0.8718 0.0680 0.8718 0.9337
No log 22.25 178 0.8719 0.0192 0.8719 0.9337
No log 22.5 180 0.8718 -0.0208 0.8718 0.9337
No log 22.75 182 0.8442 0.0874 0.8442 0.9188
No log 23.0 184 0.9744 -0.0408 0.9744 0.9871
No log 23.25 186 1.0773 -0.0466 1.0773 1.0379
No log 23.5 188 0.9333 0.0961 0.9333 0.9660
No log 23.75 190 0.9293 0.0134 0.9293 0.9640
No log 24.0 192 0.9390 0.0134 0.9390 0.9690
No log 24.25 194 0.8858 0.1135 0.8858 0.9411
No log 24.5 196 0.9032 0.0956 0.9032 0.9504
No log 24.75 198 0.9099 0.0871 0.9099 0.9539
No log 25.0 200 0.8183 0.1199 0.8183 0.9046
No log 25.25 202 0.7902 0.1347 0.7902 0.8889
No log 25.5 204 0.7601 0.1769 0.7601 0.8718
No log 25.75 206 0.7810 0.1627 0.7810 0.8838
No log 26.0 208 0.8594 0.0588 0.8594 0.9270
No log 26.25 210 0.8508 0.1612 0.8508 0.9224
No log 26.5 212 0.8640 0.0810 0.8640 0.9295
No log 26.75 214 0.8924 0.0024 0.8924 0.9446
No log 27.0 216 0.8734 0.0764 0.8734 0.9345
No log 27.25 218 0.8646 0.0764 0.8646 0.9298
No log 27.5 220 0.8489 -0.0026 0.8489 0.9213
No log 27.75 222 0.9088 -0.0886 0.9088 0.9533
No log 28.0 224 0.8918 -0.1015 0.8918 0.9443
No log 28.25 226 0.8120 0.0 0.8120 0.9011
No log 28.5 228 0.7749 0.1244 0.7749 0.8803
No log 28.75 230 0.7886 0.1244 0.7886 0.8880
No log 29.0 232 0.8318 0.1244 0.8318 0.9121
No log 29.25 234 0.8773 0.1456 0.8773 0.9366
No log 29.5 236 0.8814 0.1573 0.8814 0.9388
No log 29.75 238 0.9168 0.0071 0.9168 0.9575
No log 30.0 240 0.9765 -0.0186 0.9765 0.9882
No log 30.25 242 0.9462 0.0129 0.9462 0.9727
No log 30.5 244 0.9076 0.1236 0.9076 0.9527
No log 30.75 246 0.8885 0.1254 0.8885 0.9426
No log 31.0 248 0.8475 0.1254 0.8475 0.9206
No log 31.25 250 0.8323 0.1240 0.8323 0.9123
No log 31.5 252 0.8770 0.0090 0.8770 0.9365
No log 31.75 254 0.8972 0.0090 0.8972 0.9472
No log 32.0 256 0.9142 -0.0256 0.9142 0.9561
No log 32.25 258 0.9201 0.1232 0.9201 0.9592
No log 32.5 260 0.9367 0.1094 0.9367 0.9678
No log 32.75 262 0.9611 0.0856 0.9611 0.9803
No log 33.0 264 0.9425 0.0362 0.9425 0.9708
No log 33.25 266 0.9708 -0.0543 0.9708 0.9853
No log 33.5 268 0.9442 0.0153 0.9442 0.9717
No log 33.75 270 0.8840 0.0093 0.8840 0.9402
No log 34.0 272 0.8592 0.1769 0.8592 0.9269
No log 34.25 274 0.8354 0.1787 0.8354 0.9140
No log 34.5 276 0.8502 0.0455 0.8502 0.9221
No log 34.75 278 0.8847 0.0506 0.8847 0.9406
No log 35.0 280 0.9518 0.0540 0.9518 0.9756
No log 35.25 282 0.9934 -0.0093 0.9934 0.9967
No log 35.5 284 1.0069 0.0540 1.0069 1.0035
No log 35.75 286 0.9852 0.0879 0.9852 0.9926
No log 36.0 288 0.9418 0.0661 0.9418 0.9705
No log 36.25 290 0.9077 0.1282 0.9077 0.9527
No log 36.5 292 0.9103 0.0134 0.9103 0.9541
No log 36.75 294 0.9406 0.0928 0.9406 0.9699
No log 37.0 296 0.9213 0.0569 0.9213 0.9598
No log 37.25 298 0.9119 0.0153 0.9119 0.9549
No log 37.5 300 0.9307 0.0172 0.9307 0.9647
No log 37.75 302 1.0012 -0.0137 1.0012 1.0006
No log 38.0 304 1.0524 -0.0137 1.0524 1.0259
No log 38.25 306 1.0196 -0.0512 1.0196 1.0097
No log 38.5 308 0.9844 -0.0256 0.9844 0.9922
No log 38.75 310 0.9410 0.0802 0.9410 0.9700
No log 39.0 312 0.9407 -0.0585 0.9407 0.9699
No log 39.25 314 0.9737 -0.0470 0.9737 0.9868
No log 39.5 316 0.9904 -0.0099 0.9904 0.9952
No log 39.75 318 1.0101 -0.0168 1.0101 1.0050
No log 40.0 320 1.0384 -0.0146 1.0384 1.0190
No log 40.25 322 1.0274 -0.0519 1.0274 1.0136
No log 40.5 324 0.9898 0.0404 0.9898 0.9949
No log 40.75 326 0.9356 0.1093 0.9356 0.9673
No log 41.0 328 0.9266 0.1456 0.9266 0.9626
No log 41.25 330 0.9149 0.1093 0.9149 0.9565
No log 41.5 332 0.9237 0.0071 0.9237 0.9611
No log 41.75 334 0.9474 -0.0563 0.9474 0.9733
No log 42.0 336 0.9346 -0.0893 0.9346 0.9667
No log 42.25 338 0.8997 -0.0295 0.8997 0.9485
No log 42.5 340 0.8760 0.0856 0.8760 0.9359
No log 42.75 342 0.8802 0.1298 0.8802 0.9382
No log 43.0 344 0.8857 0.0846 0.8857 0.9411
No log 43.25 346 0.8827 0.0455 0.8827 0.9395
No log 43.5 348 0.9047 -0.0295 0.9047 0.9511
No log 43.75 350 0.9188 0.0071 0.9188 0.9586
No log 44.0 352 0.9325 0.0822 0.9325 0.9657
No log 44.25 354 0.9449 0.0847 0.9449 0.9721
No log 44.5 356 0.9706 0.0847 0.9706 0.9852
No log 44.75 358 0.9938 0.0183 0.9938 0.9969
No log 45.0 360 1.0296 -0.0093 1.0296 1.0147
No log 45.25 362 0.9838 -0.0137 0.9838 0.9919
No log 45.5 364 0.9661 -0.0137 0.9661 0.9829
No log 45.75 366 0.9317 0.0540 0.9317 0.9652
No log 46.0 368 0.8617 0.0851 0.8617 0.9283
No log 46.25 370 0.8505 0.1705 0.8505 0.9222
No log 46.5 372 0.8404 0.1298 0.8404 0.9167
No log 46.75 374 0.8623 0.0455 0.8623 0.9286
No log 47.0 376 0.9092 0.0153 0.9092 0.9535
No log 47.25 378 0.9019 0.0093 0.9019 0.9497
No log 47.5 380 0.8653 0.0834 0.8653 0.9302
No log 47.75 382 0.8624 0.1787 0.8624 0.9287
No log 48.0 384 0.8592 0.0393 0.8592 0.9269
No log 48.25 386 0.8728 0.0455 0.8728 0.9342
No log 48.5 388 0.8815 0.0529 0.8815 0.9389
No log 48.75 390 0.8680 0.0093 0.8680 0.9317
No log 49.0 392 0.8823 0.0426 0.8823 0.9393
No log 49.25 394 0.8936 0.0410 0.8936 0.9453
No log 49.5 396 0.8973 0.0410 0.8973 0.9472
No log 49.75 398 0.8703 0.0410 0.8703 0.9329
No log 50.0 400 0.8743 0.0025 0.8743 0.9350
No log 50.25 402 0.8544 0.0840 0.8544 0.9243
No log 50.5 404 0.8523 0.1292 0.8523 0.9232
No log 50.75 406 0.8502 0.1292 0.8502 0.9221
No log 51.0 408 0.8924 -0.0307 0.8924 0.9447
No log 51.25 410 0.9473 -0.0563 0.9473 0.9733
No log 51.5 412 0.9644 -0.0238 0.9644 0.9821
No log 51.75 414 0.9828 -0.0941 0.9828 0.9914
No log 52.0 416 1.0029 -0.0646 1.0029 1.0014
No log 52.25 418 1.0243 -0.0603 1.0243 1.0121
No log 52.5 420 1.0754 -0.0589 1.0754 1.0370
No log 52.75 422 1.0930 -0.0941 1.0930 1.0455
No log 53.0 424 1.0315 -0.0589 1.0315 1.0156
No log 53.25 426 0.9732 -0.0589 0.9732 0.9865
No log 53.5 428 0.9054 0.0437 0.9054 0.9515
No log 53.75 430 0.8596 0.1240 0.8596 0.9272
No log 54.0 432 0.8516 0.1244 0.8516 0.9228
No log 54.25 434 0.8737 0.1240 0.8737 0.9347
No log 54.5 436 0.9018 0.1240 0.9018 0.9496
No log 54.75 438 0.9195 0.1752 0.9195 0.9589
No log 55.0 440 0.9097 0.1244 0.9097 0.9538
No log 55.25 442 0.8886 0.0810 0.8886 0.9426
No log 55.5 444 0.8555 0.0810 0.8555 0.9249
No log 55.75 446 0.8155 0.1244 0.8155 0.9030
No log 56.0 448 0.8117 0.1740 0.8117 0.9010
No log 56.25 450 0.8207 0.1740 0.8207 0.9059
No log 56.5 452 0.8351 0.1740 0.8351 0.9138
No log 56.75 454 0.8558 0.1740 0.8558 0.9251
No log 57.0 456 0.8647 0.1244 0.8647 0.9299
No log 57.25 458 0.8887 0.1608 0.8887 0.9427
No log 57.5 460 0.9138 0.1215 0.9138 0.9559
No log 57.75 462 0.9161 0.2019 0.9161 0.9571
No log 58.0 464 0.9231 0.1232 0.9231 0.9608
No log 58.25 466 0.9174 0.1221 0.9174 0.9578
No log 58.5 468 0.9070 0.1232 0.9070 0.9523
No log 58.75 470 0.8900 0.1674 0.8900 0.9434
No log 59.0 472 0.8726 0.1674 0.8726 0.9341
No log 59.25 474 0.8578 0.1287 0.8578 0.9262
No log 59.5 476 0.8584 0.1141 0.8584 0.9265
No log 59.75 478 0.8760 0.1415 0.8760 0.9359
No log 60.0 480 0.8747 0.0717 0.8747 0.9353
No log 60.25 482 0.8853 0.0330 0.8853 0.9409
No log 60.5 484 0.9003 0.0778 0.9003 0.9488
No log 60.75 486 0.9034 0.0784 0.9034 0.9505
No log 61.0 488 0.9086 0.0301 0.9086 0.9532
No log 61.25 490 0.9370 0.0541 0.9370 0.9680
No log 61.5 492 0.9271 0.1365 0.9271 0.9629
No log 61.75 494 0.9130 0.1415 0.9130 0.9555
No log 62.0 496 0.8881 0.1228 0.8881 0.9424
No log 62.25 498 0.8835 0.0861 0.8835 0.9399
0.3056 62.5 500 0.8666 0.0851 0.8666 0.9309
0.3056 62.75 502 0.8510 0.0851 0.8510 0.9225
0.3056 63.0 504 0.8480 0.1674 0.8480 0.9208
0.3056 63.25 506 0.8502 0.1674 0.8502 0.9221
0.3056 63.5 508 0.8542 0.1674 0.8542 0.9242
0.3056 63.75 510 0.8533 0.1674 0.8533 0.9237
0.3056 64.0 512 0.8496 0.1232 0.8496 0.9217
0.3056 64.25 514 0.8570 0.0851 0.8570 0.9257
0.3056 64.5 516 0.8635 0.1228 0.8635 0.9293
0.3056 64.75 518 0.8598 0.1228 0.8598 0.9272
0.3056 65.0 520 0.8497 0.2195 0.8497 0.9218
0.3056 65.25 522 0.8361 0.2195 0.8361 0.9144
0.3056 65.5 524 0.8162 0.1228 0.8162 0.9034
0.3056 65.75 526 0.8130 0.0851 0.8130 0.9016
0.3056 66.0 528 0.7960 0.0846 0.7960 0.8922
0.3056 66.25 530 0.7862 0.2195 0.7862 0.8867
0.3056 66.5 532 0.7887 0.2195 0.7887 0.8881
0.3056 66.75 534 0.7989 0.2195 0.7989 0.8938
0.3056 67.0 536 0.8029 0.2195 0.8029 0.8961
0.3056 67.25 538 0.8147 0.1660 0.8147 0.9026
0.3056 67.5 540 0.8365 0.0861 0.8365 0.9146
0.3056 67.75 542 0.8783 0.0114 0.8783 0.9372
0.3056 68.0 544 0.8886 0.0114 0.8886 0.9426
0.3056 68.25 546 0.8764 0.0469 0.8764 0.9361
0.3056 68.5 548 0.8754 0.1133 0.8754 0.9356
0.3056 68.75 550 0.8843 0.1550 0.8843 0.9404
0.3056 69.0 552 0.8809 0.1136 0.8809 0.9386
0.3056 69.25 554 0.8677 0.0822 0.8677 0.9315
0.3056 69.5 556 0.8547 0.0455 0.8547 0.9245
0.3056 69.75 558 0.8303 0.0455 0.8303 0.9112
0.3056 70.0 560 0.8080 0.0455 0.8080 0.8989
0.3056 70.25 562 0.7984 0.0851 0.7984 0.8935

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k2_task3_organization

Finetuned
(4023)
this model