ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k8_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7943
  • Qwk: 0.5031
  • Mse: 0.7943
  • Rmse: 0.8912

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 4.0070 -0.0078 4.0070 2.0017
No log 0.0851 4 2.0946 0.0366 2.0946 1.4473
No log 0.1277 6 1.4504 0.0174 1.4504 1.2043
No log 0.1702 8 0.8582 -0.0144 0.8582 0.9264
No log 0.2128 10 0.6674 0.3216 0.6674 0.8170
No log 0.2553 12 0.6956 0.2053 0.6956 0.8340
No log 0.2979 14 0.6781 0.2459 0.6781 0.8235
No log 0.3404 16 0.6602 0.2588 0.6602 0.8125
No log 0.3830 18 0.6574 0.2413 0.6574 0.8108
No log 0.4255 20 0.7627 0.1079 0.7627 0.8733
No log 0.4681 22 0.6795 0.2681 0.6795 0.8243
No log 0.5106 24 0.6269 0.3314 0.6269 0.7918
No log 0.5532 26 0.6183 0.2868 0.6183 0.7863
No log 0.5957 28 0.6049 0.2806 0.6049 0.7778
No log 0.6383 30 0.6019 0.3859 0.6019 0.7758
No log 0.6809 32 0.6933 0.3676 0.6933 0.8326
No log 0.7234 34 0.6470 0.3683 0.6470 0.8043
No log 0.7660 36 0.6404 0.3839 0.6404 0.8003
No log 0.8085 38 0.6124 0.4614 0.6124 0.7826
No log 0.8511 40 0.6371 0.4840 0.6371 0.7982
No log 0.8936 42 0.6172 0.5155 0.6172 0.7856
No log 0.9362 44 0.6427 0.5483 0.6427 0.8017
No log 0.9787 46 0.6739 0.5208 0.6739 0.8209
No log 1.0213 48 0.7289 0.4919 0.7289 0.8537
No log 1.0638 50 0.7688 0.4533 0.7688 0.8768
No log 1.1064 52 0.7738 0.4313 0.7738 0.8797
No log 1.1489 54 0.7672 0.4314 0.7672 0.8759
No log 1.1915 56 0.7689 0.4511 0.7689 0.8769
No log 1.2340 58 0.6478 0.5281 0.6478 0.8048
No log 1.2766 60 0.6115 0.5302 0.6115 0.7820
No log 1.3191 62 0.6282 0.5446 0.6282 0.7926
No log 1.3617 64 0.5980 0.5251 0.5980 0.7733
No log 1.4043 66 0.6821 0.3834 0.6821 0.8259
No log 1.4468 68 0.6217 0.4773 0.6217 0.7885
No log 1.4894 70 0.5982 0.5600 0.5982 0.7735
No log 1.5319 72 0.6521 0.5580 0.6521 0.8075
No log 1.5745 74 0.7503 0.5607 0.7503 0.8662
No log 1.6170 76 0.9035 0.5427 0.9035 0.9505
No log 1.6596 78 0.9226 0.5063 0.9226 0.9605
No log 1.7021 80 0.8519 0.5457 0.8519 0.9230
No log 1.7447 82 0.8928 0.5071 0.8928 0.9449
No log 1.7872 84 0.7074 0.5413 0.7074 0.8410
No log 1.8298 86 0.6553 0.5133 0.6553 0.8095
No log 1.8723 88 0.6757 0.5253 0.6757 0.8220
No log 1.9149 90 0.8409 0.4459 0.8409 0.9170
No log 1.9574 92 0.7560 0.5028 0.7560 0.8695
No log 2.0 94 0.8598 0.5531 0.8598 0.9272
No log 2.0426 96 1.1439 0.4167 1.1439 1.0695
No log 2.0851 98 1.1012 0.4487 1.1012 1.0494
No log 2.1277 100 0.8752 0.5192 0.8752 0.9355
No log 2.1702 102 0.7932 0.5108 0.7932 0.8906
No log 2.2128 104 0.8167 0.4899 0.8167 0.9037
No log 2.2553 106 0.7873 0.4974 0.7873 0.8873
No log 2.2979 108 0.7814 0.5036 0.7814 0.8840
No log 2.3404 110 0.8546 0.4686 0.8546 0.9244
No log 2.3830 112 0.8945 0.5149 0.8945 0.9458
No log 2.4255 114 0.9327 0.5184 0.9327 0.9658
No log 2.4681 116 0.9583 0.5382 0.9583 0.9789
No log 2.5106 118 0.9887 0.5150 0.9887 0.9943
No log 2.5532 120 1.0016 0.5007 1.0016 1.0008
No log 2.5957 122 0.8999 0.5098 0.8999 0.9486
No log 2.6383 124 0.9490 0.4967 0.9490 0.9742
No log 2.6809 126 0.8707 0.5066 0.8707 0.9331
No log 2.7234 128 0.7662 0.5672 0.7662 0.8754
No log 2.7660 130 0.8204 0.5466 0.8204 0.9058
No log 2.8085 132 0.7419 0.5755 0.7419 0.8613
No log 2.8511 134 0.6812 0.5462 0.6812 0.8253
No log 2.8936 136 0.7396 0.5140 0.7396 0.8600
No log 2.9362 138 0.7563 0.4922 0.7563 0.8697
No log 2.9787 140 0.7849 0.4891 0.7849 0.8860
No log 3.0213 142 0.7407 0.5382 0.7407 0.8607
No log 3.0638 144 0.8347 0.5220 0.8347 0.9136
No log 3.1064 146 0.7833 0.5402 0.7833 0.8850
No log 3.1489 148 0.7359 0.5468 0.7359 0.8579
No log 3.1915 150 0.8052 0.5133 0.8052 0.8973
No log 3.2340 152 0.8645 0.5095 0.8645 0.9298
No log 3.2766 154 0.9823 0.4862 0.9823 0.9911
No log 3.3191 156 1.0680 0.4949 1.0680 1.0334
No log 3.3617 158 1.0257 0.4857 1.0257 1.0128
No log 3.4043 160 1.0008 0.4872 1.0008 1.0004
No log 3.4468 162 1.0012 0.4878 1.0012 1.0006
No log 3.4894 164 1.0019 0.4750 1.0019 1.0009
No log 3.5319 166 1.0106 0.4868 1.0106 1.0053
No log 3.5745 168 0.9405 0.4827 0.9405 0.9698
No log 3.6170 170 0.8527 0.4972 0.8527 0.9234
No log 3.6596 172 0.8516 0.5256 0.8516 0.9228
No log 3.7021 174 0.8899 0.5089 0.8899 0.9434
No log 3.7447 176 0.9151 0.4904 0.9151 0.9566
No log 3.7872 178 0.8859 0.4942 0.8859 0.9412
No log 3.8298 180 0.9199 0.4888 0.9199 0.9591
No log 3.8723 182 0.9139 0.4788 0.9139 0.9560
No log 3.9149 184 0.8779 0.4732 0.8779 0.9369
No log 3.9574 186 0.9198 0.4992 0.9198 0.9590
No log 4.0 188 0.9362 0.5140 0.9362 0.9676
No log 4.0426 190 0.9505 0.4751 0.9505 0.9750
No log 4.0851 192 1.0274 0.4752 1.0274 1.0136
No log 4.1277 194 1.0031 0.4801 1.0031 1.0016
No log 4.1702 196 0.8882 0.4873 0.8882 0.9424
No log 4.2128 198 0.8637 0.4898 0.8637 0.9294
No log 4.2553 200 0.8083 0.5196 0.8083 0.8990
No log 4.2979 202 0.7808 0.5177 0.7808 0.8836
No log 4.3404 204 0.8578 0.4805 0.8578 0.9262
No log 4.3830 206 1.0220 0.4761 1.0220 1.0109
No log 4.4255 208 0.9685 0.4780 0.9685 0.9841
No log 4.4681 210 0.8309 0.4990 0.8309 0.9115
No log 4.5106 212 0.8463 0.5146 0.8463 0.9199
No log 4.5532 214 1.0318 0.4449 1.0318 1.0158
No log 4.5957 216 1.0383 0.4709 1.0383 1.0190
No log 4.6383 218 0.9199 0.4980 0.9199 0.9591
No log 4.6809 220 0.9543 0.4977 0.9543 0.9769
No log 4.7234 222 1.0813 0.4709 1.0813 1.0399
No log 4.7660 224 1.0666 0.4702 1.0666 1.0327
No log 4.8085 226 0.9783 0.4966 0.9783 0.9891
No log 4.8511 228 0.9411 0.4719 0.9411 0.9701
No log 4.8936 230 0.9420 0.4855 0.9420 0.9706
No log 4.9362 232 0.9832 0.4784 0.9832 0.9916
No log 4.9787 234 1.0356 0.4496 1.0356 1.0176
No log 5.0213 236 0.9948 0.4818 0.9948 0.9974
No log 5.0638 238 0.9096 0.5047 0.9096 0.9538
No log 5.1064 240 0.9367 0.4927 0.9367 0.9678
No log 5.1489 242 0.9808 0.4716 0.9808 0.9903
No log 5.1915 244 0.8986 0.4976 0.8986 0.9479
No log 5.2340 246 0.8184 0.5304 0.8184 0.9046
No log 5.2766 248 0.8164 0.5373 0.8164 0.9035
No log 5.3191 250 0.8299 0.4677 0.8299 0.9110
No log 5.3617 252 0.8545 0.4726 0.8545 0.9244
No log 5.4043 254 0.8608 0.4726 0.8608 0.9278
No log 5.4468 256 0.8156 0.5072 0.8156 0.9031
No log 5.4894 258 0.8040 0.5037 0.8040 0.8967
No log 5.5319 260 0.8314 0.5012 0.8314 0.9118
No log 5.5745 262 0.8429 0.5100 0.8429 0.9181
No log 5.6170 264 0.8536 0.5027 0.8536 0.9239
No log 5.6596 266 0.8400 0.5262 0.8400 0.9165
No log 5.7021 268 0.8138 0.5381 0.8138 0.9021
No log 5.7447 270 0.7965 0.5241 0.7965 0.8924
No log 5.7872 272 0.8012 0.5096 0.8012 0.8951
No log 5.8298 274 0.8166 0.4992 0.8166 0.9037
No log 5.8723 276 0.8628 0.4677 0.8628 0.9289
No log 5.9149 278 0.8527 0.4796 0.8527 0.9234
No log 5.9574 280 0.8295 0.4858 0.8295 0.9107
No log 6.0 282 0.7813 0.5166 0.7813 0.8839
No log 6.0426 284 0.7610 0.5048 0.7610 0.8724
No log 6.0851 286 0.7304 0.4960 0.7304 0.8547
No log 6.1277 288 0.6986 0.5523 0.6986 0.8358
No log 6.1702 290 0.6793 0.5035 0.6793 0.8242
No log 6.2128 292 0.6652 0.4950 0.6652 0.8156
No log 6.2553 294 0.6920 0.4932 0.6920 0.8319
No log 6.2979 296 0.7605 0.5114 0.7605 0.8721
No log 6.3404 298 0.8007 0.5085 0.8007 0.8948
No log 6.3830 300 0.8062 0.5049 0.8062 0.8979
No log 6.4255 302 0.8252 0.4931 0.8252 0.9084
No log 6.4681 304 0.8460 0.4931 0.8460 0.9198
No log 6.5106 306 0.8491 0.5260 0.8491 0.9214
No log 6.5532 308 0.8349 0.5192 0.8349 0.9137
No log 6.5957 310 0.8362 0.5326 0.8362 0.9144
No log 6.6383 312 0.8256 0.5211 0.8256 0.9086
No log 6.6809 314 0.8335 0.5157 0.8335 0.9130
No log 6.7234 316 0.8014 0.5094 0.8014 0.8952
No log 6.7660 318 0.7417 0.5097 0.7417 0.8612
No log 6.8085 320 0.7082 0.5387 0.7082 0.8415
No log 6.8511 322 0.7096 0.5521 0.7096 0.8424
No log 6.8936 324 0.7143 0.5258 0.7143 0.8452
No log 6.9362 326 0.7433 0.5506 0.7433 0.8622
No log 6.9787 328 0.8102 0.5199 0.8102 0.9001
No log 7.0213 330 0.8170 0.5180 0.8170 0.9039
No log 7.0638 332 0.7725 0.5450 0.7725 0.8789
No log 7.1064 334 0.7258 0.5211 0.7258 0.8519
No log 7.1489 336 0.7178 0.5241 0.7178 0.8473
No log 7.1915 338 0.7330 0.5193 0.7330 0.8562
No log 7.2340 340 0.7433 0.5157 0.7433 0.8621
No log 7.2766 342 0.7639 0.5715 0.7639 0.8740
No log 7.3191 344 0.7893 0.5453 0.7893 0.8884
No log 7.3617 346 0.8286 0.5336 0.8286 0.9103
No log 7.4043 348 0.8477 0.5213 0.8477 0.9207
No log 7.4468 350 0.8620 0.5081 0.8620 0.9284
No log 7.4894 352 0.9073 0.4783 0.9073 0.9525
No log 7.5319 354 0.9085 0.4783 0.9085 0.9532
No log 7.5745 356 0.8450 0.5037 0.8450 0.9192
No log 7.6170 358 0.7959 0.5213 0.7959 0.8921
No log 7.6596 360 0.7870 0.5175 0.7870 0.8871
No log 7.7021 362 0.7739 0.5154 0.7739 0.8797
No log 7.7447 364 0.7730 0.5154 0.7730 0.8792
No log 7.7872 366 0.7756 0.5144 0.7756 0.8807
No log 7.8298 368 0.7933 0.5124 0.7933 0.8907
No log 7.8723 370 0.8198 0.5170 0.8198 0.9054
No log 7.9149 372 0.8019 0.5124 0.8019 0.8955
No log 7.9574 374 0.7914 0.5160 0.7914 0.8896
No log 8.0 376 0.7867 0.5276 0.7867 0.8870
No log 8.0426 378 0.7834 0.5027 0.7834 0.8851
No log 8.0851 380 0.7896 0.5226 0.7896 0.8886
No log 8.1277 382 0.7917 0.5226 0.7917 0.8898
No log 8.1702 384 0.7874 0.5081 0.7874 0.8873
No log 8.2128 386 0.7967 0.5550 0.7968 0.8926
No log 8.2553 388 0.8184 0.5286 0.8184 0.9047
No log 8.2979 390 0.8238 0.5306 0.8238 0.9076
No log 8.3404 392 0.8398 0.5284 0.8398 0.9164
No log 8.3830 394 0.8577 0.5099 0.8577 0.9261
No log 8.4255 396 0.8440 0.5194 0.8440 0.9187
No log 8.4681 398 0.8120 0.5302 0.8120 0.9011
No log 8.5106 400 0.7717 0.5608 0.7717 0.8785
No log 8.5532 402 0.7503 0.5548 0.7503 0.8662
No log 8.5957 404 0.7459 0.5548 0.7459 0.8637
No log 8.6383 406 0.7435 0.5549 0.7435 0.8623
No log 8.6809 408 0.7555 0.5608 0.7555 0.8692
No log 8.7234 410 0.7676 0.5250 0.7676 0.8762
No log 8.7660 412 0.7758 0.5239 0.7758 0.8808
No log 8.8085 414 0.7857 0.5285 0.7857 0.8864
No log 8.8511 416 0.8042 0.5203 0.8042 0.8968
No log 8.8936 418 0.8080 0.5203 0.8080 0.8989
No log 8.9362 420 0.8028 0.5149 0.8028 0.8960
No log 8.9787 422 0.8075 0.5149 0.8075 0.8986
No log 9.0213 424 0.8101 0.5159 0.8101 0.9001
No log 9.0638 426 0.8045 0.5190 0.8045 0.8970
No log 9.1064 428 0.7961 0.5031 0.7961 0.8922
No log 9.1489 430 0.7881 0.5031 0.7881 0.8877
No log 9.1915 432 0.7737 0.5255 0.7737 0.8796
No log 9.2340 434 0.7602 0.5607 0.7602 0.8719
No log 9.2766 436 0.7524 0.5549 0.7524 0.8674
No log 9.3191 438 0.7475 0.5549 0.7475 0.8646
No log 9.3617 440 0.7505 0.5549 0.7505 0.8663
No log 9.4043 442 0.7576 0.5549 0.7576 0.8704
No log 9.4468 444 0.7657 0.5197 0.7657 0.8750
No log 9.4894 446 0.7779 0.5116 0.7779 0.8820
No log 9.5319 448 0.7931 0.5144 0.7931 0.8906
No log 9.5745 450 0.8030 0.5190 0.8030 0.8961
No log 9.6170 452 0.8091 0.5169 0.8091 0.8995
No log 9.6596 454 0.8123 0.5169 0.8123 0.9013
No log 9.7021 456 0.8133 0.5169 0.8133 0.9018
No log 9.7447 458 0.8107 0.5240 0.8107 0.9004
No log 9.7872 460 0.8060 0.5134 0.8060 0.8977
No log 9.8298 462 0.8001 0.5140 0.8001 0.8945
No log 9.8723 464 0.7963 0.5031 0.7963 0.8924
No log 9.9149 466 0.7951 0.5031 0.7951 0.8917
No log 9.9574 468 0.7945 0.5031 0.7945 0.8914
No log 10.0 470 0.7943 0.5031 0.7943 0.8912

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k8_task2_organization

Finetuned
(4023)
this model