ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8707
  • Qwk: 0.4930
  • Mse: 0.8707
  • Rmse: 0.9331

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 4.0291 0.0024 4.0291 2.0073
No log 0.0784 4 1.9731 0.0648 1.9731 1.4047
No log 0.1176 6 1.6716 0.0181 1.6716 1.2929
No log 0.1569 8 1.2219 0.0291 1.2219 1.1054
No log 0.1961 10 0.7579 0.1770 0.7579 0.8706
No log 0.2353 12 0.7333 0.2281 0.7333 0.8563
No log 0.2745 14 0.7684 0.2149 0.7684 0.8766
No log 0.3137 16 0.8231 0.0918 0.8231 0.9073
No log 0.3529 18 0.8600 0.0535 0.8600 0.9274
No log 0.3922 20 0.7646 0.0884 0.7646 0.8744
No log 0.4314 22 0.8071 0.1045 0.8071 0.8984
No log 0.4706 24 0.9517 0.1387 0.9517 0.9755
No log 0.5098 26 1.0394 0.1444 1.0394 1.0195
No log 0.5490 28 0.9179 0.2666 0.9179 0.9581
No log 0.5882 30 0.7819 0.2288 0.7819 0.8843
No log 0.6275 32 0.6986 0.2969 0.6986 0.8358
No log 0.6667 34 0.6333 0.2702 0.6333 0.7958
No log 0.7059 36 0.6133 0.3911 0.6133 0.7831
No log 0.7451 38 0.6080 0.3810 0.6080 0.7797
No log 0.7843 40 0.6262 0.3740 0.6262 0.7913
No log 0.8235 42 0.6131 0.4315 0.6131 0.7830
No log 0.8627 44 0.8177 0.4002 0.8177 0.9043
No log 0.9020 46 1.0846 0.3252 1.0846 1.0415
No log 0.9412 48 1.1445 0.2463 1.1445 1.0698
No log 0.9804 50 0.8134 0.3934 0.8134 0.9019
No log 1.0196 52 0.6331 0.4813 0.6331 0.7957
No log 1.0588 54 0.6727 0.4381 0.6727 0.8202
No log 1.0980 56 0.6975 0.4444 0.6975 0.8352
No log 1.1373 58 0.5731 0.4831 0.5731 0.7570
No log 1.1765 60 0.5573 0.5257 0.5573 0.7465
No log 1.2157 62 0.5496 0.5230 0.5496 0.7414
No log 1.2549 64 0.7153 0.4016 0.7153 0.8458
No log 1.2941 66 0.8435 0.4492 0.8435 0.9184
No log 1.3333 68 0.7059 0.4642 0.7059 0.8402
No log 1.3725 70 0.6331 0.4928 0.6331 0.7957
No log 1.4118 72 0.7388 0.5114 0.7388 0.8595
No log 1.4510 74 0.6902 0.4821 0.6902 0.8308
No log 1.4902 76 0.6877 0.4641 0.6877 0.8292
No log 1.5294 78 1.0625 0.4451 1.0625 1.0308
No log 1.5686 80 1.2185 0.4232 1.2185 1.1039
No log 1.6078 82 0.8803 0.4335 0.8803 0.9383
No log 1.6471 84 0.5889 0.4398 0.5889 0.7674
No log 1.6863 86 0.5722 0.4236 0.5722 0.7565
No log 1.7255 88 0.6050 0.3869 0.6050 0.7778
No log 1.7647 90 0.5978 0.4346 0.5978 0.7732
No log 1.8039 92 0.5796 0.5244 0.5796 0.7613
No log 1.8431 94 0.7222 0.5535 0.7222 0.8498
No log 1.8824 96 0.8227 0.5232 0.8227 0.9070
No log 1.9216 98 0.7983 0.5585 0.7983 0.8935
No log 1.9608 100 0.7388 0.5916 0.7388 0.8595
No log 2.0 102 0.7455 0.6045 0.7455 0.8634
No log 2.0392 104 0.7906 0.5388 0.7906 0.8891
No log 2.0784 106 0.7549 0.6147 0.7549 0.8688
No log 2.1176 108 0.7378 0.5963 0.7378 0.8590
No log 2.1569 110 0.7873 0.5737 0.7873 0.8873
No log 2.1961 112 0.8115 0.5760 0.8115 0.9008
No log 2.2353 114 0.7943 0.5717 0.7943 0.8912
No log 2.2745 116 0.8017 0.5786 0.8017 0.8954
No log 2.3137 118 0.8018 0.5698 0.8018 0.8954
No log 2.3529 120 0.8250 0.5747 0.8250 0.9083
No log 2.3922 122 0.8534 0.5486 0.8534 0.9238
No log 2.4314 124 0.9022 0.5074 0.9022 0.9499
No log 2.4706 126 0.8519 0.5108 0.8519 0.9230
No log 2.5098 128 0.8163 0.4852 0.8163 0.9035
No log 2.5490 130 0.7682 0.5468 0.7682 0.8765
No log 2.5882 132 0.7643 0.5697 0.7643 0.8742
No log 2.6275 134 0.8093 0.5519 0.8093 0.8996
No log 2.6667 136 0.8915 0.5071 0.8915 0.9442
No log 2.7059 138 0.8832 0.5591 0.8832 0.9398
No log 2.7451 140 0.9129 0.5198 0.9129 0.9555
No log 2.7843 142 0.9090 0.5253 0.9090 0.9534
No log 2.8235 144 0.8453 0.5318 0.8453 0.9194
No log 2.8627 146 0.8282 0.5467 0.8282 0.9100
No log 2.9020 148 0.8442 0.5318 0.8442 0.9188
No log 2.9412 150 1.0006 0.4908 1.0006 1.0003
No log 2.9804 152 1.3014 0.4456 1.3014 1.1408
No log 3.0196 154 1.2417 0.4397 1.2417 1.1143
No log 3.0588 156 0.9161 0.4704 0.9161 0.9571
No log 3.0980 158 0.7994 0.5431 0.7994 0.8941
No log 3.1373 160 0.7757 0.5251 0.7757 0.8807
No log 3.1765 162 0.7875 0.5226 0.7875 0.8874
No log 3.2157 164 0.8416 0.4873 0.8416 0.9174
No log 3.2549 166 0.9046 0.4836 0.9046 0.9511
No log 3.2941 168 0.8622 0.5464 0.8622 0.9286
No log 3.3333 170 0.8496 0.5755 0.8496 0.9218
No log 3.3725 172 0.9055 0.5408 0.9055 0.9516
No log 3.4118 174 0.9275 0.5383 0.9275 0.9631
No log 3.4510 176 0.9768 0.5274 0.9768 0.9883
No log 3.4902 178 0.9907 0.5453 0.9907 0.9953
No log 3.5294 180 1.0216 0.4979 1.0216 1.0107
No log 3.5686 182 1.0763 0.4282 1.0763 1.0374
No log 3.6078 184 1.0773 0.4493 1.0773 1.0379
No log 3.6471 186 0.9152 0.4581 0.9152 0.9567
No log 3.6863 188 0.7929 0.5114 0.7929 0.8905
No log 3.7255 190 0.7903 0.5701 0.7903 0.8890
No log 3.7647 192 0.7892 0.5621 0.7892 0.8883
No log 3.8039 194 0.8333 0.5052 0.8333 0.9128
No log 3.8431 196 0.9594 0.4934 0.9594 0.9795
No log 3.8824 198 1.0030 0.4830 1.0030 1.0015
No log 3.9216 200 0.9976 0.4908 0.9976 0.9988
No log 3.9608 202 0.9868 0.5018 0.9868 0.9934
No log 4.0 204 0.9568 0.5098 0.9568 0.9781
No log 4.0392 206 0.9830 0.4960 0.9830 0.9915
No log 4.0784 208 0.9578 0.5002 0.9578 0.9787
No log 4.1176 210 0.9456 0.5147 0.9456 0.9724
No log 4.1569 212 0.9461 0.4899 0.9461 0.9727
No log 4.1961 214 1.0815 0.4994 1.0815 1.0399
No log 4.2353 216 1.1922 0.4684 1.1922 1.0919
No log 4.2745 218 1.1137 0.5028 1.1137 1.0553
No log 4.3137 220 0.9702 0.4823 0.9702 0.9850
No log 4.3529 222 0.9153 0.5031 0.9153 0.9567
No log 4.3922 224 0.9339 0.5228 0.9339 0.9664
No log 4.4314 226 0.9437 0.5179 0.9437 0.9714
No log 4.4706 228 0.9183 0.5172 0.9183 0.9583
No log 4.5098 230 0.9024 0.4925 0.9024 0.9499
No log 4.5490 232 0.9725 0.5138 0.9725 0.9861
No log 4.5882 234 1.0327 0.5198 1.0327 1.0162
No log 4.6275 236 0.9930 0.4791 0.9930 0.9965
No log 4.6667 238 0.9793 0.5063 0.9793 0.9896
No log 4.7059 240 0.9788 0.5025 0.9788 0.9894
No log 4.7451 242 0.9390 0.5270 0.9390 0.9690
No log 4.7843 244 0.9263 0.5257 0.9263 0.9625
No log 4.8235 246 0.9163 0.5065 0.9163 0.9573
No log 4.8627 248 0.8615 0.5069 0.8615 0.9282
No log 4.9020 250 0.8200 0.5446 0.8200 0.9056
No log 4.9412 252 0.8433 0.4864 0.8433 0.9183
No log 4.9804 254 0.9228 0.4704 0.9228 0.9606
No log 5.0196 256 0.9324 0.4806 0.9324 0.9656
No log 5.0588 258 0.8504 0.4909 0.8504 0.9222
No log 5.0980 260 0.8557 0.5041 0.8557 0.9250
No log 5.1373 262 0.8585 0.5402 0.8585 0.9265
No log 5.1765 264 0.8957 0.5265 0.8957 0.9464
No log 5.2157 266 0.9324 0.5326 0.9324 0.9656
No log 5.2549 268 0.9675 0.5023 0.9675 0.9836
No log 5.2941 270 0.9901 0.4903 0.9901 0.9950
No log 5.3333 272 0.9760 0.4981 0.9760 0.9879
No log 5.3725 274 0.9825 0.4748 0.9825 0.9912
No log 5.4118 276 1.0092 0.4942 1.0092 1.0046
No log 5.4510 278 1.0162 0.4971 1.0162 1.0081
No log 5.4902 280 0.9615 0.4853 0.9615 0.9806
No log 5.5294 282 0.9130 0.4940 0.9130 0.9555
No log 5.5686 284 0.8985 0.4948 0.8985 0.9479
No log 5.6078 286 0.8704 0.5152 0.8704 0.9330
No log 5.6471 288 0.8418 0.5540 0.8418 0.9175
No log 5.6863 290 0.8320 0.5540 0.8320 0.9121
No log 5.7255 292 0.8431 0.5380 0.8431 0.9182
No log 5.7647 294 0.8971 0.4887 0.8971 0.9472
No log 5.8039 296 0.9338 0.4761 0.9338 0.9663
No log 5.8431 298 0.9475 0.4632 0.9475 0.9734
No log 5.8824 300 0.9127 0.4926 0.9127 0.9553
No log 5.9216 302 0.8813 0.5298 0.8813 0.9388
No log 5.9608 304 0.8898 0.5592 0.8898 0.9433
No log 6.0 306 0.9329 0.5269 0.9329 0.9658
No log 6.0392 308 0.9279 0.5269 0.9279 0.9633
No log 6.0784 310 0.8847 0.5631 0.8847 0.9406
No log 6.1176 312 0.8594 0.5579 0.8594 0.9270
No log 6.1569 314 0.8457 0.5251 0.8457 0.9196
No log 6.1961 316 0.8402 0.4986 0.8402 0.9166
No log 6.2353 318 0.8215 0.4986 0.8215 0.9063
No log 6.2745 320 0.8189 0.5404 0.8189 0.9049
No log 6.3137 322 0.8374 0.5356 0.8374 0.9151
No log 6.3529 324 0.8717 0.5429 0.8717 0.9336
No log 6.3922 326 0.8948 0.5404 0.8948 0.9459
No log 6.4314 328 0.9119 0.5262 0.9119 0.9549
No log 6.4706 330 0.9432 0.5197 0.9432 0.9712
No log 6.5098 332 0.9465 0.5130 0.9465 0.9729
No log 6.5490 334 0.9627 0.4921 0.9627 0.9812
No log 6.5882 336 0.9682 0.4788 0.9682 0.9840
No log 6.6275 338 0.9473 0.4745 0.9473 0.9733
No log 6.6667 340 0.9028 0.4944 0.9028 0.9502
No log 6.7059 342 0.8477 0.4896 0.8477 0.9207
No log 6.7451 344 0.8151 0.5413 0.8151 0.9028
No log 6.7843 346 0.8279 0.5514 0.8279 0.9099
No log 6.8235 348 0.8477 0.5369 0.8477 0.9207
No log 6.8627 350 0.8706 0.5369 0.8706 0.9331
No log 6.9020 352 0.8896 0.5182 0.8896 0.9432
No log 6.9412 354 0.8744 0.5074 0.8744 0.9351
No log 6.9804 356 0.8414 0.5391 0.8414 0.9173
No log 7.0196 358 0.8386 0.5391 0.8386 0.9158
No log 7.0588 360 0.8524 0.5068 0.8524 0.9232
No log 7.0980 362 0.8805 0.5362 0.8805 0.9383
No log 7.1373 364 0.9334 0.4810 0.9334 0.9661
No log 7.1765 366 0.9469 0.4646 0.9469 0.9731
No log 7.2157 368 0.9042 0.4876 0.9042 0.9509
No log 7.2549 370 0.8678 0.5182 0.8678 0.9316
No log 7.2941 372 0.8343 0.5163 0.8343 0.9134
No log 7.3333 374 0.8148 0.5290 0.8148 0.9026
No log 7.3725 376 0.8083 0.5364 0.8083 0.8991
No log 7.4118 378 0.8082 0.5358 0.8082 0.8990
No log 7.4510 380 0.8281 0.4967 0.8281 0.9100
No log 7.4902 382 0.8647 0.4993 0.8647 0.9299
No log 7.5294 384 0.9055 0.4814 0.9055 0.9516
No log 7.5686 386 0.9311 0.4814 0.9311 0.9649
No log 7.6078 388 0.9479 0.4872 0.9479 0.9736
No log 7.6471 390 0.9728 0.4998 0.9728 0.9863
No log 7.6863 392 1.0251 0.4933 1.0251 1.0125
No log 7.7255 394 1.0656 0.4851 1.0656 1.0323
No log 7.7647 396 1.0426 0.4851 1.0426 1.0211
No log 7.8039 398 0.9817 0.4934 0.9817 0.9908
No log 7.8431 400 0.9155 0.4912 0.9155 0.9568
No log 7.8824 402 0.8850 0.4878 0.8850 0.9407
No log 7.9216 404 0.8741 0.4878 0.8741 0.9349
No log 7.9608 406 0.8716 0.4864 0.8716 0.9336
No log 8.0 408 0.8802 0.4911 0.8802 0.9382
No log 8.0392 410 0.8953 0.4898 0.8953 0.9462
No log 8.0784 412 0.9122 0.4898 0.9122 0.9551
No log 8.1176 414 0.9327 0.4932 0.9327 0.9658
No log 8.1569 416 0.9361 0.4932 0.9361 0.9675
No log 8.1961 418 0.9385 0.4873 0.9385 0.9688
No log 8.2353 420 0.9445 0.4710 0.9445 0.9719
No log 8.2745 422 0.9420 0.4768 0.9420 0.9706
No log 8.3137 424 0.9434 0.4724 0.9434 0.9713
No log 8.3529 426 0.9549 0.4806 0.9549 0.9772
No log 8.3922 428 0.9554 0.4947 0.9554 0.9774
No log 8.4314 430 0.9518 0.4983 0.9518 0.9756
No log 8.4706 432 0.9265 0.4867 0.9265 0.9626
No log 8.5098 434 0.8824 0.4778 0.8824 0.9393
No log 8.5490 436 0.8385 0.4946 0.8385 0.9157
No log 8.5882 438 0.8146 0.4952 0.8146 0.9025
No log 8.6275 440 0.7863 0.4973 0.7863 0.8868
No log 8.6667 442 0.7702 0.5135 0.7702 0.8776
No log 8.7059 444 0.7614 0.5007 0.7614 0.8726
No log 8.7451 446 0.7610 0.5080 0.7610 0.8724
No log 8.7843 448 0.7740 0.5 0.7740 0.8798
No log 8.8235 450 0.7945 0.5088 0.7945 0.8913
No log 8.8627 452 0.8107 0.5088 0.8107 0.9004
No log 8.9020 454 0.8308 0.5088 0.8308 0.9115
No log 8.9412 456 0.8490 0.5088 0.8490 0.9214
No log 8.9804 458 0.8739 0.5065 0.8739 0.9348
No log 9.0196 460 0.8931 0.5065 0.8931 0.9450
No log 9.0588 462 0.8973 0.5065 0.8973 0.9473
No log 9.0980 464 0.8907 0.5065 0.8907 0.9437
No log 9.1373 466 0.8743 0.4933 0.8743 0.9351
No log 9.1765 468 0.8635 0.5061 0.8635 0.9292
No log 9.2157 470 0.8611 0.5121 0.8611 0.9280
No log 9.2549 472 0.8615 0.5121 0.8615 0.9282
No log 9.2941 474 0.8558 0.5076 0.8558 0.9251
No log 9.3333 476 0.8504 0.5014 0.8504 0.9222
No log 9.3725 478 0.8498 0.4952 0.8498 0.9218
No log 9.4118 480 0.8535 0.4925 0.8535 0.9238
No log 9.4510 482 0.8622 0.5065 0.8622 0.9285
No log 9.4902 484 0.8775 0.4930 0.8775 0.9368
No log 9.5294 486 0.8892 0.4864 0.8892 0.9430
No log 9.5686 488 0.9002 0.4804 0.9002 0.9488
No log 9.6078 490 0.9036 0.4799 0.9036 0.9506
No log 9.6471 492 0.9005 0.4804 0.9005 0.9489
No log 9.6863 494 0.8958 0.4804 0.8958 0.9465
No log 9.7255 496 0.8923 0.4804 0.8923 0.9446
No log 9.7647 498 0.8893 0.4864 0.8893 0.9430
0.439 9.8039 500 0.8855 0.4864 0.8855 0.9410
0.439 9.8431 502 0.8821 0.4864 0.8821 0.9392
0.439 9.8824 504 0.8782 0.4930 0.8782 0.9371
0.439 9.9216 506 0.8747 0.4930 0.8747 0.9352
0.439 9.9608 508 0.8719 0.4930 0.8719 0.9338
0.439 10.0 510 0.8707 0.4930 0.8707 0.9331

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
4
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k9_task2_organization

Finetuned
(4023)
this model