ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k7_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7700
  • Qwk: 0.7368
  • Mse: 0.7700
  • Rmse: 0.8775

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0588 2 6.8670 0.0116 6.8670 2.6205
No log 0.1176 4 4.9159 0.0444 4.9159 2.2172
No log 0.1765 6 3.0127 0.0848 3.0127 1.7357
No log 0.2353 8 2.7505 0.0625 2.7505 1.6585
No log 0.2941 10 2.3014 0.1304 2.3014 1.5170
No log 0.3529 12 1.7381 0.1887 1.7381 1.3184
No log 0.4118 14 1.7320 0.1905 1.7320 1.3161
No log 0.4706 16 1.9616 0.1321 1.9616 1.4006
No log 0.5294 18 2.0860 0.2435 2.0860 1.4443
No log 0.5882 20 1.8681 0.1651 1.8681 1.3668
No log 0.6471 22 1.5896 0.1905 1.5896 1.2608
No log 0.7059 24 1.4676 0.1714 1.4676 1.2115
No log 0.7647 26 1.6078 0.1165 1.6078 1.2680
No log 0.8235 28 1.5776 0.1165 1.5776 1.2560
No log 0.8824 30 1.5450 0.3540 1.5450 1.2430
No log 0.9412 32 1.7121 0.4252 1.7121 1.3085
No log 1.0 34 1.7538 0.3906 1.7538 1.3243
No log 1.0588 36 1.5625 0.3636 1.5625 1.2500
No log 1.1176 38 1.3117 0.2936 1.3117 1.1453
No log 1.1765 40 1.4223 0.3273 1.4223 1.1926
No log 1.2353 42 1.4171 0.3273 1.4171 1.1904
No log 1.2941 44 1.3358 0.3604 1.3358 1.1558
No log 1.3529 46 1.3198 0.4138 1.3198 1.1488
No log 1.4118 48 1.3072 0.3894 1.3072 1.1433
No log 1.4706 50 1.2791 0.3571 1.2791 1.1310
No log 1.5294 52 1.2395 0.3540 1.2395 1.1133
No log 1.5882 54 1.0471 0.5938 1.0471 1.0233
No log 1.6471 56 1.0628 0.5672 1.0628 1.0309
No log 1.7059 58 1.1568 0.5075 1.1568 1.0756
No log 1.7647 60 1.2997 0.5507 1.2997 1.1401
No log 1.8235 62 1.0943 0.5839 1.0943 1.0461
No log 1.8824 64 0.7864 0.7059 0.7864 0.8868
No log 1.9412 66 1.3532 0.4252 1.3532 1.1633
No log 2.0 68 1.8550 0.2185 1.8550 1.3620
No log 2.0588 70 1.6723 0.2903 1.6723 1.2932
No log 2.1176 72 1.1895 0.5116 1.1895 1.0907
No log 2.1765 74 0.9489 0.6519 0.9489 0.9741
No log 2.2353 76 0.9123 0.6618 0.9123 0.9552
No log 2.2941 78 0.8967 0.6906 0.8967 0.9469
No log 2.3529 80 0.8741 0.6569 0.8741 0.9350
No log 2.4118 82 0.8753 0.6475 0.8753 0.9355
No log 2.4706 84 0.8640 0.6618 0.8640 0.9295
No log 2.5294 86 0.9295 0.6338 0.9295 0.9641
No log 2.5882 88 1.0026 0.64 1.0026 1.0013
No log 2.6471 90 1.0637 0.6405 1.0637 1.0314
No log 2.7059 92 0.9877 0.6536 0.9877 0.9938
No log 2.7647 94 0.7562 0.7389 0.7562 0.8696
No log 2.8235 96 0.7265 0.7329 0.7265 0.8524
No log 2.8824 98 0.7132 0.725 0.7132 0.8445
No log 2.9412 100 0.6967 0.7613 0.6967 0.8347
No log 3.0 102 0.6855 0.7532 0.6855 0.8280
No log 3.0588 104 0.8325 0.7285 0.8325 0.9124
No log 3.1176 106 0.8103 0.7211 0.8103 0.9002
No log 3.1765 108 0.6691 0.75 0.6691 0.8180
No log 3.2353 110 0.9978 0.6184 0.9978 0.9989
No log 3.2941 112 1.0566 0.5946 1.0566 1.0279
No log 3.3529 114 0.8577 0.7105 0.8577 0.9261
No log 3.4118 116 0.8659 0.7383 0.8659 0.9306
No log 3.4706 118 0.9405 0.5816 0.9405 0.9698
No log 3.5294 120 0.8509 0.7211 0.8509 0.9225
No log 3.5882 122 0.8778 0.7092 0.8778 0.9369
No log 3.6471 124 0.8805 0.6901 0.8805 0.9384
No log 3.7059 126 0.8094 0.6901 0.8094 0.8997
No log 3.7647 128 0.7271 0.7413 0.7271 0.8527
No log 3.8235 130 0.6719 0.7534 0.6719 0.8197
No log 3.8824 132 0.6067 0.7733 0.6067 0.7789
No log 3.9412 134 0.5817 0.775 0.5817 0.7627
No log 4.0 136 0.5621 0.7673 0.5621 0.7497
No log 4.0588 138 0.5536 0.7730 0.5536 0.7440
No log 4.1176 140 0.5402 0.7927 0.5402 0.7350
No log 4.1765 142 0.6658 0.7879 0.6658 0.8159
No log 4.2353 144 0.6670 0.7578 0.6670 0.8167
No log 4.2941 146 0.5851 0.7682 0.5851 0.7649
No log 4.3529 148 0.5992 0.7448 0.5992 0.7741
No log 4.4118 150 0.6606 0.7891 0.6606 0.8128
No log 4.4706 152 0.6188 0.7448 0.6188 0.7867
No log 4.5294 154 0.8082 0.6667 0.8082 0.8990
No log 4.5882 156 0.8698 0.6573 0.8698 0.9326
No log 4.6471 158 0.7315 0.7034 0.7315 0.8553
No log 4.7059 160 0.6351 0.7534 0.6351 0.7969
No log 4.7647 162 0.6717 0.7310 0.6717 0.8196
No log 4.8235 164 0.7832 0.7083 0.7832 0.8850
No log 4.8824 166 0.7752 0.7083 0.7752 0.8804
No log 4.9412 168 0.6317 0.7682 0.6317 0.7948
No log 5.0 170 0.5783 0.8272 0.5783 0.7605
No log 5.0588 172 0.6492 0.8047 0.6492 0.8057
No log 5.1176 174 0.5665 0.8402 0.5665 0.7527
No log 5.1765 176 0.5582 0.7925 0.5582 0.7471
No log 5.2353 178 0.7390 0.7114 0.7390 0.8597
No log 5.2941 180 0.8141 0.7034 0.8141 0.9023
No log 5.3529 182 0.7204 0.7083 0.7204 0.8488
No log 5.4118 184 0.6615 0.7692 0.6615 0.8133
No log 5.4706 186 0.6645 0.7606 0.6645 0.8152
No log 5.5294 188 0.6969 0.76 0.6969 0.8348
No log 5.5882 190 0.8680 0.6986 0.8680 0.9316
No log 5.6471 192 0.7348 0.72 0.7348 0.8572
No log 5.7059 194 0.5586 0.8075 0.5586 0.7474
No log 5.7647 196 0.8443 0.7456 0.8443 0.9188
No log 5.8235 198 0.9025 0.7186 0.9025 0.9500
No log 5.8824 200 0.6668 0.7730 0.6668 0.8166
No log 5.9412 202 0.6074 0.7947 0.6074 0.7793
No log 6.0 204 0.6336 0.7755 0.6336 0.7960
No log 6.0588 206 0.7015 0.7467 0.7015 0.8376
No log 6.1176 208 0.8248 0.6711 0.8248 0.9082
No log 6.1765 210 0.7550 0.7190 0.7550 0.8689
No log 6.2353 212 0.7098 0.7821 0.7098 0.8425
No log 6.2941 214 0.7310 0.7703 0.7310 0.8550
No log 6.3529 216 0.7742 0.7237 0.7742 0.8799
No log 6.4118 218 0.8008 0.7067 0.8008 0.8949
No log 6.4706 220 0.7765 0.7383 0.7765 0.8812
No log 6.5294 222 0.7559 0.7755 0.7559 0.8694
No log 6.5882 224 0.7443 0.7703 0.7443 0.8627
No log 6.6471 226 0.7675 0.7333 0.7675 0.8761
No log 6.7059 228 0.7500 0.7383 0.7500 0.8660
No log 6.7647 230 0.7424 0.7383 0.7424 0.8616
No log 6.8235 232 0.7085 0.7919 0.7085 0.8417
No log 6.8824 234 0.6935 0.7651 0.6935 0.8328
No log 6.9412 236 0.6936 0.7651 0.6936 0.8328
No log 7.0 238 0.6627 0.7651 0.6627 0.8141
No log 7.0588 240 0.6396 0.7682 0.6396 0.7998
No log 7.1176 242 0.6261 0.7919 0.6261 0.7913
No log 7.1765 244 0.5947 0.8228 0.5947 0.7711
No log 7.2353 246 0.5640 0.8302 0.5640 0.7510
No log 7.2941 248 0.6316 0.775 0.6316 0.7947
No log 7.3529 250 0.6866 0.7453 0.6866 0.8286
No log 7.4118 252 0.7981 0.7170 0.7981 0.8934
No log 7.4706 254 0.7854 0.7355 0.7854 0.8863
No log 7.5294 256 0.7257 0.7484 0.7257 0.8519
No log 7.5882 258 0.7964 0.7484 0.7964 0.8924
No log 7.6471 260 0.8233 0.7436 0.8233 0.9074
No log 7.7059 262 0.6996 0.7607 0.6996 0.8364
No log 7.7647 264 0.5718 0.8095 0.5718 0.7562
No log 7.8235 266 0.5594 0.8114 0.5594 0.7479
No log 7.8824 268 0.5697 0.8171 0.5697 0.7548
No log 7.9412 270 0.5804 0.8072 0.5804 0.7618
No log 8.0 272 0.5556 0.8095 0.5556 0.7454
No log 8.0588 274 0.6127 0.8023 0.6127 0.7827
No log 8.1176 276 0.6450 0.7784 0.6450 0.8031
No log 8.1765 278 0.6046 0.8024 0.6046 0.7776
No log 8.2353 280 0.6251 0.7853 0.6251 0.7906
No log 8.2941 282 0.6589 0.8077 0.6589 0.8118
No log 8.3529 284 0.6872 0.7792 0.6872 0.8290
No log 8.4118 286 0.7034 0.7564 0.7034 0.8387
No log 8.4706 288 0.6751 0.7843 0.6751 0.8216
No log 8.5294 290 0.6570 0.8075 0.6570 0.8105
No log 8.5882 292 0.7664 0.7485 0.7664 0.8754
No log 8.6471 294 0.8412 0.7314 0.8412 0.9172
No log 8.7059 296 0.7713 0.7683 0.7713 0.8783
No log 8.7647 298 0.7690 0.7261 0.7690 0.8769
No log 8.8235 300 0.8023 0.7211 0.8023 0.8957
No log 8.8824 302 0.8257 0.7397 0.8257 0.9087
No log 8.9412 304 0.8079 0.7397 0.8079 0.8988
No log 9.0 306 0.7901 0.7619 0.7901 0.8889
No log 9.0588 308 0.7543 0.7517 0.7543 0.8685
No log 9.1176 310 0.7228 0.7568 0.7228 0.8502
No log 9.1765 312 0.7150 0.7568 0.7150 0.8456
No log 9.2353 314 0.7375 0.7517 0.7375 0.8588
No log 9.2941 316 0.7165 0.7651 0.7165 0.8465
No log 9.3529 318 0.6828 0.7651 0.6828 0.8263
No log 9.4118 320 0.6829 0.76 0.6829 0.8264
No log 9.4706 322 0.7510 0.7421 0.7510 0.8666
No log 9.5294 324 0.7463 0.75 0.7463 0.8639
No log 9.5882 326 0.6646 0.7898 0.6646 0.8152
No log 9.6471 328 0.6486 0.7561 0.6486 0.8053
No log 9.7059 330 0.6833 0.7654 0.6833 0.8266
No log 9.7647 332 0.6917 0.7682 0.6917 0.8317
No log 9.8235 334 0.7415 0.7568 0.7415 0.8611
No log 9.8824 336 0.8306 0.6853 0.8306 0.9114
No log 9.9412 338 0.8442 0.6713 0.8442 0.9188
No log 10.0 340 0.7988 0.6950 0.7988 0.8937
No log 10.0588 342 0.8224 0.6667 0.8224 0.9069
No log 10.1176 344 0.8224 0.6892 0.8224 0.9069
No log 10.1765 346 0.7970 0.6849 0.7970 0.8927
No log 10.2353 348 0.7180 0.7383 0.7180 0.8474
No log 10.2941 350 0.6475 0.7712 0.6475 0.8047
No log 10.3529 352 0.6423 0.7662 0.6423 0.8014
No log 10.4118 354 0.6581 0.7662 0.6581 0.8112
No log 10.4706 356 0.7080 0.7692 0.7080 0.8414
No log 10.5294 358 0.7660 0.7143 0.7660 0.8752
No log 10.5882 360 0.8160 0.7143 0.8160 0.9033
No log 10.6471 362 0.8029 0.7273 0.8029 0.8961
No log 10.7059 364 0.7738 0.7632 0.7738 0.8796
No log 10.7647 366 0.7853 0.7632 0.7853 0.8862
No log 10.8235 368 0.7370 0.7613 0.7370 0.8585
No log 10.8824 370 0.6837 0.7927 0.6837 0.8268
No log 10.9412 372 0.6495 0.7791 0.6495 0.8059
No log 11.0 374 0.6197 0.7953 0.6197 0.7872
No log 11.0588 376 0.6105 0.7910 0.6105 0.7814
No log 11.1176 378 0.6254 0.8 0.6254 0.7908
No log 11.1765 380 0.6216 0.7816 0.6216 0.7884
No log 11.2353 382 0.6333 0.8 0.6333 0.7958
No log 11.2941 384 0.6662 0.7879 0.6662 0.8162
No log 11.3529 386 0.6908 0.7702 0.6908 0.8312
No log 11.4118 388 0.6717 0.7758 0.6717 0.8196
No log 11.4706 390 0.6779 0.8024 0.6779 0.8233
No log 11.5294 392 0.6589 0.8221 0.6589 0.8117
No log 11.5882 394 0.6362 0.8049 0.6362 0.7976
No log 11.6471 396 0.6390 0.7799 0.6390 0.7994
No log 11.7059 398 0.6511 0.7625 0.6511 0.8069
No log 11.7647 400 0.6246 0.8024 0.6246 0.7903
No log 11.8235 402 0.5962 0.7882 0.5962 0.7721
No log 11.8824 404 0.5699 0.8095 0.5699 0.7549
No log 11.9412 406 0.5909 0.7977 0.5909 0.7687
No log 12.0 408 0.6109 0.8 0.6109 0.7816
No log 12.0588 410 0.6338 0.7389 0.6338 0.7961
No log 12.1176 412 0.6383 0.7682 0.6383 0.7990
No log 12.1765 414 0.6048 0.7632 0.6048 0.7777
No log 12.2353 416 0.5676 0.8228 0.5676 0.7534
No log 12.2941 418 0.5383 0.8199 0.5383 0.7337
No log 12.3529 420 0.5321 0.8095 0.5321 0.7294
No log 12.4118 422 0.5561 0.8229 0.5561 0.7457
No log 12.4706 424 0.5882 0.8118 0.5882 0.7670
No log 12.5294 426 0.6388 0.7805 0.6388 0.7993
No log 12.5882 428 0.6256 0.7643 0.6256 0.7910
No log 12.6471 430 0.5787 0.7792 0.5787 0.7607
No log 12.7059 432 0.5731 0.8182 0.5731 0.7570
No log 12.7647 434 0.5876 0.8129 0.5876 0.7665
No log 12.8235 436 0.5848 0.8302 0.5848 0.7647
No log 12.8824 438 0.5558 0.8050 0.5558 0.7455
No log 12.9412 440 0.5649 0.8171 0.5649 0.7516
No log 13.0 442 0.6168 0.7976 0.6168 0.7854
No log 13.0588 444 0.7864 0.7375 0.7864 0.8868
No log 13.1176 446 0.8548 0.7125 0.8548 0.9246
No log 13.1765 448 0.8068 0.6853 0.8068 0.8982
No log 13.2353 450 0.7807 0.7101 0.7807 0.8836
No log 13.2941 452 0.7684 0.7153 0.7684 0.8766
No log 13.3529 454 0.6997 0.7808 0.6997 0.8365
No log 13.4118 456 0.6577 0.7815 0.6577 0.8110
No log 13.4706 458 0.7798 0.6971 0.7798 0.8830
No log 13.5294 460 0.8684 0.7358 0.8684 0.9319
No log 13.5882 462 0.8387 0.7283 0.8387 0.9158
No log 13.6471 464 0.8586 0.7093 0.8586 0.9266
No log 13.7059 466 0.8123 0.6918 0.8123 0.9013
No log 13.7647 468 0.7290 0.7467 0.7290 0.8538
No log 13.8235 470 0.6846 0.7843 0.6846 0.8274
No log 13.8824 472 0.6823 0.7898 0.6823 0.8260
No log 13.9412 474 0.6597 0.7904 0.6597 0.8122
No log 14.0 476 0.6194 0.8228 0.6194 0.7870
No log 14.0588 478 0.6550 0.7805 0.6550 0.8093
No log 14.1176 480 0.6845 0.7654 0.6845 0.8273
No log 14.1765 482 0.6875 0.7564 0.6875 0.8291
No log 14.2353 484 0.6491 0.7632 0.6491 0.8057
No log 14.2941 486 0.6658 0.7947 0.6658 0.8160
No log 14.3529 488 0.6706 0.8129 0.6706 0.8189
No log 14.4118 490 0.6389 0.8205 0.6389 0.7993
No log 14.4706 492 0.6092 0.8205 0.6092 0.7805
No log 14.5294 494 0.6008 0.8121 0.6008 0.7751
No log 14.5882 496 0.5879 0.8193 0.5879 0.7667
No log 14.6471 498 0.5953 0.8025 0.5953 0.7716
0.3985 14.7059 500 0.6350 0.7898 0.6350 0.7968
0.3985 14.7647 502 0.6370 0.7898 0.6370 0.7981
0.3985 14.8235 504 0.6419 0.8025 0.6419 0.8012
0.3985 14.8824 506 0.6657 0.7949 0.6657 0.8159
0.3985 14.9412 508 0.6980 0.7692 0.6980 0.8355
0.3985 15.0 510 0.7495 0.7368 0.7495 0.8658
0.3985 15.0588 512 0.8280 0.7285 0.8280 0.9100
0.3985 15.1176 514 0.8359 0.7097 0.8359 0.9143
0.3985 15.1765 516 0.7700 0.7368 0.7700 0.8775

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k7_task1_organization

Finetuned
(4023)
this model