ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1245
  • Qwk: 0.4728
  • Mse: 1.1245
  • Rmse: 1.0604

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.9052 -0.0043 3.9052 1.9761
No log 0.1111 4 1.7841 0.0632 1.7841 1.3357
No log 0.1667 6 1.0055 0.0508 1.0055 1.0027
No log 0.2222 8 0.7077 0.2161 0.7077 0.8413
No log 0.2778 10 0.6855 0.2562 0.6855 0.8279
No log 0.3333 12 0.6464 0.2742 0.6464 0.8040
No log 0.3889 14 0.6531 0.3235 0.6531 0.8082
No log 0.4444 16 1.0476 0.2029 1.0476 1.0235
No log 0.5 18 0.9695 0.3030 0.9695 0.9846
No log 0.5556 20 1.0238 0.2364 1.0238 1.0118
No log 0.6111 22 1.2950 0.1626 1.2950 1.1380
No log 0.6667 24 1.2178 0.2006 1.2178 1.1035
No log 0.7222 26 0.7423 0.4298 0.7423 0.8616
No log 0.7778 28 0.6375 0.4565 0.6375 0.7984
No log 0.8333 30 1.0019 0.3193 1.0019 1.0010
No log 0.8889 32 1.2241 0.2613 1.2241 1.1064
No log 0.9444 34 0.8866 0.4340 0.8866 0.9416
No log 1.0 36 0.6091 0.5221 0.6091 0.7805
No log 1.0556 38 0.6357 0.5066 0.6357 0.7973
No log 1.1111 40 0.5808 0.4726 0.5808 0.7621
No log 1.1667 42 0.6040 0.4512 0.6040 0.7772
No log 1.2222 44 0.6984 0.4532 0.6984 0.8357
No log 1.2778 46 0.7018 0.4947 0.7018 0.8378
No log 1.3333 48 0.6586 0.5397 0.6586 0.8116
No log 1.3889 50 0.7679 0.4812 0.7679 0.8763
No log 1.4444 52 0.8936 0.4413 0.8936 0.9453
No log 1.5 54 0.8789 0.4151 0.8789 0.9375
No log 1.5556 56 0.8041 0.4987 0.8041 0.8967
No log 1.6111 58 0.9666 0.4961 0.9666 0.9832
No log 1.6667 60 1.0558 0.4841 1.0558 1.0275
No log 1.7222 62 1.1333 0.4557 1.1333 1.0646
No log 1.7778 64 1.1798 0.4522 1.1798 1.0862
No log 1.8333 66 1.1116 0.4918 1.1116 1.0543
No log 1.8889 68 1.1358 0.4656 1.1358 1.0658
No log 1.9444 70 1.1689 0.4833 1.1689 1.0812
No log 2.0 72 1.1277 0.4604 1.1277 1.0619
No log 2.0556 74 0.9618 0.4856 0.9618 0.9807
No log 2.1111 76 0.8747 0.5183 0.8747 0.9353
No log 2.1667 78 0.7862 0.5750 0.7862 0.8867
No log 2.2222 80 0.7508 0.5643 0.7508 0.8665
No log 2.2778 82 0.8913 0.5078 0.8913 0.9441
No log 2.3333 84 0.8267 0.5384 0.8267 0.9092
No log 2.3889 86 0.7962 0.5270 0.7962 0.8923
No log 2.4444 88 0.8734 0.4760 0.8734 0.9346
No log 2.5 90 0.8828 0.5295 0.8828 0.9396
No log 2.5556 92 0.9813 0.5027 0.9813 0.9906
No log 2.6111 94 1.0550 0.5137 1.0550 1.0271
No log 2.6667 96 1.0664 0.5261 1.0664 1.0327
No log 2.7222 98 1.1514 0.4739 1.1514 1.0730
No log 2.7778 100 1.2776 0.4622 1.2776 1.1303
No log 2.8333 102 1.3190 0.4556 1.3190 1.1485
No log 2.8889 104 1.2600 0.4600 1.2600 1.1225
No log 2.9444 106 1.2192 0.4571 1.2192 1.1042
No log 3.0 108 1.1476 0.4689 1.1476 1.0713
No log 3.0556 110 1.1103 0.4649 1.1103 1.0537
No log 3.1111 112 1.0624 0.4836 1.0624 1.0307
No log 3.1667 114 1.0086 0.5066 1.0086 1.0043
No log 3.2222 116 0.9596 0.4836 0.9596 0.9796
No log 3.2778 118 1.0021 0.5151 1.0021 1.0010
No log 3.3333 120 1.1332 0.5144 1.1332 1.0645
No log 3.3889 122 1.2682 0.4552 1.2682 1.1262
No log 3.4444 124 1.2003 0.4777 1.2003 1.0956
No log 3.5 126 1.1069 0.4865 1.1069 1.0521
No log 3.5556 128 0.9881 0.5142 0.9881 0.9940
No log 3.6111 130 0.9824 0.5088 0.9824 0.9911
No log 3.6667 132 1.1805 0.4547 1.1805 1.0865
No log 3.7222 134 1.3401 0.4180 1.3401 1.1576
No log 3.7778 136 1.2598 0.4533 1.2598 1.1224
No log 3.8333 138 1.0679 0.5088 1.0679 1.0334
No log 3.8889 140 1.0256 0.5085 1.0256 1.0127
No log 3.9444 142 1.0780 0.4937 1.0780 1.0383
No log 4.0 144 1.2382 0.5024 1.2382 1.1128
No log 4.0556 146 1.3297 0.4846 1.3297 1.1531
No log 4.1111 148 1.3485 0.5126 1.3485 1.1612
No log 4.1667 150 1.3249 0.5033 1.3249 1.1510
No log 4.2222 152 1.2474 0.5195 1.2474 1.1169
No log 4.2778 154 1.1783 0.5215 1.1783 1.0855
No log 4.3333 156 1.0837 0.5259 1.0837 1.0410
No log 4.3889 158 0.9750 0.5153 0.9750 0.9874
No log 4.4444 160 0.9387 0.5184 0.9387 0.9689
No log 4.5 162 0.9696 0.5437 0.9696 0.9847
No log 4.5556 164 1.0677 0.4729 1.0677 1.0333
No log 4.6111 166 1.2037 0.4374 1.2037 1.0971
No log 4.6667 168 1.2345 0.4246 1.2345 1.1111
No log 4.7222 170 1.1793 0.4365 1.1793 1.0860
No log 4.7778 172 1.0345 0.4875 1.0345 1.0171
No log 4.8333 174 0.9714 0.4819 0.9714 0.9856
No log 4.8889 176 1.0222 0.4615 1.0222 1.0110
No log 4.9444 178 1.0641 0.4846 1.0641 1.0315
No log 5.0 180 1.0791 0.4747 1.0791 1.0388
No log 5.0556 182 1.1332 0.5075 1.1332 1.0645
No log 5.1111 184 1.2026 0.4836 1.2026 1.0967
No log 5.1667 186 1.2219 0.4918 1.2219 1.1054
No log 5.2222 188 1.1957 0.4926 1.1957 1.0935
No log 5.2778 190 1.1514 0.4894 1.1514 1.0730
No log 5.3333 192 1.1522 0.4887 1.1522 1.0734
No log 5.3889 194 1.2014 0.4631 1.2014 1.0961
No log 5.4444 196 1.1978 0.4630 1.1978 1.0944
No log 5.5 198 1.1330 0.4743 1.1330 1.0644
No log 5.5556 200 1.1102 0.4585 1.1102 1.0537
No log 5.6111 202 1.1694 0.4465 1.1694 1.0814
No log 5.6667 204 1.1317 0.4297 1.1317 1.0638
No log 5.7222 206 1.1148 0.4284 1.1148 1.0558
No log 5.7778 208 1.0839 0.4519 1.0839 1.0411
No log 5.8333 210 1.1026 0.4490 1.1026 1.0501
No log 5.8889 212 1.1318 0.4996 1.1318 1.0639
No log 5.9444 214 1.1450 0.5008 1.1450 1.0700
No log 6.0 216 1.1449 0.4869 1.1449 1.0700
No log 6.0556 218 1.1795 0.4733 1.1795 1.0861
No log 6.1111 220 1.2330 0.4407 1.2330 1.1104
No log 6.1667 222 1.2480 0.4469 1.2480 1.1171
No log 6.2222 224 1.1893 0.4687 1.1893 1.0906
No log 6.2778 226 1.1265 0.4729 1.1265 1.0614
No log 6.3333 228 1.1268 0.4880 1.1268 1.0615
No log 6.3889 230 1.1397 0.4849 1.1397 1.0676
No log 6.4444 232 1.1501 0.4838 1.1501 1.0724
No log 6.5 234 1.1875 0.4753 1.1875 1.0897
No log 6.5556 236 1.2673 0.4842 1.2673 1.1257
No log 6.6111 238 1.4039 0.4353 1.4039 1.1849
No log 6.6667 240 1.4423 0.4315 1.4423 1.2010
No log 6.7222 242 1.3598 0.4271 1.3598 1.1661
No log 6.7778 244 1.2463 0.4769 1.2463 1.1164
No log 6.8333 246 1.1564 0.4864 1.1564 1.0754
No log 6.8889 248 1.1127 0.4837 1.1127 1.0549
No log 6.9444 250 1.0673 0.4896 1.0673 1.0331
No log 7.0 252 1.0650 0.4860 1.0650 1.0320
No log 7.0556 254 1.0709 0.4896 1.0709 1.0348
No log 7.1111 256 1.0967 0.5085 1.0967 1.0472
No log 7.1667 258 1.1586 0.4642 1.1586 1.0764
No log 7.2222 260 1.2086 0.4541 1.2086 1.0994
No log 7.2778 262 1.2687 0.4397 1.2687 1.1264
No log 7.3333 264 1.2773 0.4688 1.2773 1.1302
No log 7.3889 266 1.2573 0.4968 1.2573 1.1213
No log 7.4444 268 1.2212 0.4959 1.2212 1.1051
No log 7.5 270 1.1777 0.4670 1.1777 1.0852
No log 7.5556 272 1.1497 0.4552 1.1497 1.0722
No log 7.6111 274 1.1285 0.4650 1.1285 1.0623
No log 7.6667 276 1.1270 0.4650 1.1270 1.0616
No log 7.7222 278 1.1189 0.4650 1.1189 1.0578
No log 7.7778 280 1.1085 0.4582 1.1085 1.0528
No log 7.8333 282 1.0950 0.4814 1.0950 1.0464
No log 7.8889 284 1.1125 0.5195 1.1125 1.0548
No log 7.9444 286 1.1535 0.4777 1.1535 1.0740
No log 8.0 288 1.1827 0.4592 1.1827 1.0875
No log 8.0556 290 1.1851 0.4437 1.1851 1.0886
No log 8.1111 292 1.1466 0.4643 1.1466 1.0708
No log 8.1667 294 1.1005 0.4954 1.1005 1.0490
No log 8.2222 296 1.0745 0.4988 1.0745 1.0366
No log 8.2778 298 1.0787 0.4905 1.0787 1.0386
No log 8.3333 300 1.0988 0.4880 1.0988 1.0482
No log 8.3889 302 1.1285 0.4740 1.1285 1.0623
No log 8.4444 304 1.1599 0.4866 1.1599 1.0770
No log 8.5 306 1.1848 0.4826 1.1848 1.0885
No log 8.5556 308 1.2173 0.4636 1.2173 1.1033
No log 8.6111 310 1.2421 0.4677 1.2421 1.1145
No log 8.6667 312 1.2362 0.4775 1.2362 1.1119
No log 8.7222 314 1.2185 0.4857 1.2185 1.1039
No log 8.7778 316 1.1964 0.4914 1.1964 1.0938
No log 8.8333 318 1.1757 0.4690 1.1757 1.0843
No log 8.8889 320 1.1580 0.4690 1.1580 1.0761
No log 8.9444 322 1.1399 0.4765 1.1399 1.0677
No log 9.0 324 1.1246 0.4868 1.1246 1.0605
No log 9.0556 326 1.1139 0.4659 1.1139 1.0554
No log 9.1111 328 1.1096 0.4660 1.1096 1.0534
No log 9.1667 330 1.1081 0.4614 1.1081 1.0527
No log 9.2222 332 1.1078 0.4524 1.1078 1.0525
No log 9.2778 334 1.1044 0.4660 1.1044 1.0509
No log 9.3333 336 1.1020 0.4733 1.1020 1.0498
No log 9.3889 338 1.1042 0.4803 1.1042 1.0508
No log 9.4444 340 1.1088 0.4721 1.1088 1.0530
No log 9.5 342 1.1148 0.4803 1.1148 1.0558
No log 9.5556 344 1.1219 0.4647 1.1219 1.0592
No log 9.6111 346 1.1246 0.4728 1.1246 1.0605
No log 9.6667 348 1.1265 0.4728 1.1265 1.0614
No log 9.7222 350 1.1248 0.4728 1.1248 1.0606
No log 9.7778 352 1.1231 0.4728 1.1231 1.0598
No log 9.8333 354 1.1226 0.4728 1.1226 1.0595
No log 9.8889 356 1.1232 0.4728 1.1232 1.0598
No log 9.9444 358 1.1240 0.4728 1.1240 1.0602
No log 10.0 360 1.1245 0.4728 1.1245 1.0604

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k6_task2_organization

Finetuned
(4023)
this model