ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k7_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0251
  • Qwk: 0.2340
  • Mse: 1.0251
  • Rmse: 1.0125

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.3839 -0.0149 3.3839 1.8395
No log 0.1111 4 1.7337 -0.0370 1.7337 1.3167
No log 0.1667 6 1.3518 0.0294 1.3518 1.1627
No log 0.2222 8 0.8496 0.1736 0.8496 0.9217
No log 0.2778 10 0.5541 0.0569 0.5541 0.7444
No log 0.3333 12 0.5377 0.0569 0.5377 0.7333
No log 0.3889 14 0.5627 0.0388 0.5627 0.7502
No log 0.4444 16 0.5904 0.0909 0.5904 0.7684
No log 0.5 18 0.6082 0.0909 0.6082 0.7798
No log 0.5556 20 0.6494 0.1467 0.6494 0.8058
No log 0.6111 22 0.6164 0.0222 0.6164 0.7851
No log 0.6667 24 0.5875 0.1111 0.5875 0.7665
No log 0.7222 26 0.5591 0.0815 0.5591 0.7477
No log 0.7778 28 0.5405 0.1220 0.5405 0.7352
No log 0.8333 30 0.5262 0.1220 0.5262 0.7254
No log 0.8889 32 0.5061 0.1111 0.5061 0.7114
No log 0.9444 34 0.5117 0.2533 0.5117 0.7153
No log 1.0 36 0.5136 0.2821 0.5136 0.7167
No log 1.0556 38 0.5390 0.1045 0.5390 0.7342
No log 1.1111 40 0.6040 0.1429 0.6040 0.7772
No log 1.1667 42 0.9759 0.1464 0.9759 0.9879
No log 1.2222 44 0.8086 0.0798 0.8086 0.8992
No log 1.2778 46 0.7071 0.0 0.7071 0.8409
No log 1.3333 48 0.7413 0.0968 0.7413 0.8610
No log 1.3889 50 0.6354 0.1698 0.6354 0.7971
No log 1.4444 52 0.6073 0.0897 0.6073 0.7793
No log 1.5 54 0.5846 0.1688 0.5846 0.7646
No log 1.5556 56 0.5716 0.3136 0.5716 0.7560
No log 1.6111 58 0.6840 0.3077 0.6840 0.8271
No log 1.6667 60 0.5338 0.2883 0.5338 0.7306
No log 1.7222 62 0.5585 0.2911 0.5585 0.7473
No log 1.7778 64 0.6039 0.2513 0.6039 0.7771
No log 1.8333 66 0.6861 0.2487 0.6861 0.8283
No log 1.8889 68 0.8359 0.1169 0.8359 0.9143
No log 1.9444 70 0.9238 0.1008 0.9238 0.9611
No log 2.0 72 0.8453 0.1799 0.8453 0.9194
No log 2.0556 74 0.6573 0.3242 0.6573 0.8107
No log 2.1111 76 0.7585 0.2137 0.7585 0.8709
No log 2.1667 78 1.0748 0.1062 1.0748 1.0367
No log 2.2222 80 1.0941 0.0769 1.0941 1.0460
No log 2.2778 82 0.8204 0.2381 0.8204 0.9057
No log 2.3333 84 0.7686 0.2356 0.7686 0.8767
No log 2.3889 86 0.9916 0.1880 0.9916 0.9958
No log 2.4444 88 1.1178 0.1594 1.1178 1.0572
No log 2.5 90 0.8955 0.2258 0.8955 0.9463
No log 2.5556 92 1.1695 0.1882 1.1695 1.0814
No log 2.6111 94 1.5833 0.1304 1.5833 1.2583
No log 2.6667 96 1.6181 0.1086 1.6181 1.2720
No log 2.7222 98 1.1779 0.2119 1.1779 1.0853
No log 2.7778 100 0.7848 0.4595 0.7848 0.8859
No log 2.8333 102 0.7964 0.3684 0.7964 0.8924
No log 2.8889 104 1.0068 0.1939 1.0068 1.0034
No log 2.9444 106 1.3267 0.0307 1.3267 1.1518
No log 3.0 108 1.4314 0.0627 1.4314 1.1964
No log 3.0556 110 1.2832 0.0811 1.2832 1.1328
No log 3.1111 112 1.0704 0.1886 1.0704 1.0346
No log 3.1667 114 1.2016 0.1773 1.2016 1.0962
No log 3.2222 116 1.1233 0.2276 1.1233 1.0599
No log 3.2778 118 0.9945 0.2243 0.9945 0.9973
No log 3.3333 120 1.2657 0.1572 1.2657 1.1250
No log 3.3889 122 1.2792 0.1572 1.2792 1.1310
No log 3.4444 124 1.3335 0.1141 1.3335 1.1548
No log 3.5 126 1.2887 0.0853 1.2887 1.1352
No log 3.5556 128 1.2148 0.1304 1.2148 1.1022
No log 3.6111 130 1.2027 0.1186 1.2027 1.0967
No log 3.6667 132 1.1005 0.1507 1.1005 1.0490
No log 3.7222 134 1.1792 0.1340 1.1792 1.0859
No log 3.7778 136 1.0827 0.2218 1.0827 1.0405
No log 3.8333 138 0.9344 0.2941 0.9344 0.9667
No log 3.8889 140 1.0568 0.2621 1.0568 1.0280
No log 3.9444 142 0.9821 0.2941 0.9821 0.9910
No log 4.0 144 0.7454 0.3834 0.7454 0.8634
No log 4.0556 146 0.7809 0.4062 0.7809 0.8837
No log 4.1111 148 0.9220 0.3083 0.9220 0.9602
No log 4.1667 150 1.1775 0.0769 1.1775 1.0851
No log 4.2222 152 0.8922 0.3623 0.8922 0.9446
No log 4.2778 154 0.5652 0.3488 0.5652 0.7518
No log 4.3333 156 0.5674 0.4737 0.5674 0.7532
No log 4.3889 158 0.6174 0.4386 0.6174 0.7858
No log 4.4444 160 1.1086 0.1565 1.1086 1.0529
No log 4.5 162 1.2015 0.0853 1.2015 1.0961
No log 4.5556 164 0.8565 0.3358 0.8565 0.9255
No log 4.6111 166 0.6063 0.4027 0.6063 0.7787
No log 4.6667 168 0.6305 0.4286 0.6305 0.7941
No log 4.7222 170 0.9365 0.2889 0.9365 0.9678
No log 4.7778 172 1.4655 0.1409 1.4655 1.2106
No log 4.8333 174 1.5287 0.0909 1.5287 1.2364
No log 4.8889 176 1.2053 0.1378 1.2053 1.0978
No log 4.9444 178 0.8151 0.2922 0.8151 0.9028
No log 5.0 180 0.6943 0.3761 0.6943 0.8333
No log 5.0556 182 0.7695 0.3448 0.7695 0.8772
No log 5.1111 184 1.1361 0.2000 1.1361 1.0659
No log 5.1667 186 1.5674 0.1373 1.5674 1.2519
No log 5.2222 188 1.5200 0.1125 1.5200 1.2329
No log 5.2778 190 1.1111 0.1781 1.1111 1.0541
No log 5.3333 192 0.7782 0.3846 0.7782 0.8822
No log 5.3889 194 0.6966 0.4236 0.6966 0.8346
No log 5.4444 196 0.7071 0.4236 0.7071 0.8409
No log 5.5 198 0.8758 0.2653 0.8758 0.9358
No log 5.5556 200 1.1028 0.1317 1.1028 1.0502
No log 5.6111 202 1.0771 0.1317 1.0771 1.0378
No log 5.6667 204 0.9410 0.2366 0.9410 0.9701
No log 5.7222 206 0.8209 0.2269 0.8209 0.9060
No log 5.7778 208 0.7967 0.3220 0.7967 0.8926
No log 5.8333 210 0.9554 0.2000 0.9554 0.9774
No log 5.8889 212 1.1750 0.1785 1.1750 1.0840
No log 5.9444 214 1.2412 0.1565 1.2412 1.1141
No log 6.0 216 1.1665 0.1785 1.1665 1.0800
No log 6.0556 218 1.0370 0.1704 1.0370 1.0183
No log 6.1111 220 0.9719 0.1698 0.9719 0.9859
No log 6.1667 222 0.9883 0.1698 0.9883 0.9941
No log 6.2222 224 1.0630 0.1777 1.0630 1.0310
No log 6.2778 226 1.1168 0.1781 1.1168 1.0568
No log 6.3333 228 1.1964 0.1340 1.1964 1.0938
No log 6.3889 230 1.1232 0.1329 1.1232 1.0598
No log 6.4444 232 0.9782 0.2360 0.9782 0.9890
No log 6.5 234 0.8699 0.2374 0.8699 0.9327
No log 6.5556 236 0.9424 0.2360 0.9424 0.9708
No log 6.6111 238 0.9421 0.2360 0.9421 0.9706
No log 6.6667 240 1.0550 0.2409 1.0550 1.0271
No log 6.7222 242 1.2221 0.1894 1.2221 1.1055
No log 6.7778 244 1.1227 0.1888 1.1227 1.0596
No log 6.8333 246 0.8576 0.3185 0.8576 0.9260
No log 6.8889 248 0.7313 0.3668 0.7313 0.8551
No log 6.9444 250 0.7742 0.36 0.7742 0.8799
No log 7.0 252 0.9862 0.2416 0.9862 0.9931
No log 7.0556 254 1.3092 0.1895 1.3092 1.1442
No log 7.1111 256 1.3938 0.1392 1.3938 1.1806
No log 7.1667 258 1.2574 0.1894 1.2574 1.1213
No log 7.2222 260 1.0371 0.2117 1.0371 1.0184
No log 7.2778 262 1.0017 0.2647 1.0017 1.0008
No log 7.3333 264 1.1020 0.2113 1.1020 1.0498
No log 7.3889 266 1.1586 0.1572 1.1586 1.0764
No log 7.4444 268 1.1119 0.2054 1.1119 1.0545
No log 7.5 270 1.0785 0.2055 1.0785 1.0385
No log 7.5556 272 0.9792 0.2347 0.9792 0.9895
No log 7.6111 274 1.0358 0.2058 1.0358 1.0177
No log 7.6667 276 1.0895 0.1049 1.0895 1.0438
No log 7.7222 278 1.0032 0.1304 1.0032 1.0016
No log 7.7778 280 0.8823 0.2698 0.8823 0.9393
No log 7.8333 282 0.8622 0.3016 0.8622 0.9286
No log 7.8889 284 0.9224 0.2360 0.9224 0.9604
No log 7.9444 286 1.0494 0.1549 1.0494 1.0244
No log 8.0 288 1.0692 0.2055 1.0692 1.0340
No log 8.0556 290 0.9682 0.3811 0.9682 0.9840
No log 8.1111 292 0.9217 0.3410 0.9217 0.9600
No log 8.1667 294 0.9926 0.3764 0.9926 0.9963
No log 8.2222 296 1.1858 0.1952 1.1858 1.0889
No log 8.2778 298 1.4040 0.1573 1.4040 1.1849
No log 8.3333 300 1.5635 0.1579 1.5635 1.2504
No log 8.3889 302 1.5645 0.1579 1.5645 1.2508
No log 8.4444 304 1.4450 0.1605 1.4450 1.2021
No log 8.5 306 1.2426 0.1798 1.2426 1.1147
No log 8.5556 308 1.0965 0.2055 1.0965 1.0472
No log 8.6111 310 0.9934 0.3114 0.9934 0.9967
No log 8.6667 312 0.9476 0.3653 0.9476 0.9734
No log 8.7222 314 0.9359 0.3653 0.9359 0.9674
No log 8.7778 316 0.9529 0.3653 0.9529 0.9762
No log 8.8333 318 1.0183 0.2340 1.0183 1.0091
No log 8.8889 320 1.1292 0.2054 1.1292 1.0626
No log 8.9444 322 1.2139 0.1678 1.2139 1.1018
No log 9.0 324 1.2329 0.1678 1.2329 1.1103
No log 9.0556 326 1.1734 0.1892 1.1734 1.0832
No log 9.1111 328 1.0820 0.1831 1.0820 1.0402
No log 9.1667 330 1.0017 0.2340 1.0017 1.0009
No log 9.2222 332 0.9513 0.3208 0.9513 0.9754
No log 9.2778 334 0.9430 0.3208 0.9430 0.9711
No log 9.3333 336 0.9433 0.3208 0.9433 0.9712
No log 9.3889 338 0.9317 0.3731 0.9317 0.9653
No log 9.4444 340 0.9547 0.3282 0.9547 0.9771
No log 9.5 342 0.9793 0.2659 0.9793 0.9896
No log 9.5556 344 1.0096 0.2635 1.0096 1.0048
No log 9.6111 346 1.0292 0.2340 1.0292 1.0145
No log 9.6667 348 1.0439 0.2340 1.0439 1.0217
No log 9.7222 350 1.0570 0.1834 1.0570 1.0281
No log 9.7778 352 1.0505 0.2115 1.0505 1.0250
No log 9.8333 354 1.0402 0.2340 1.0402 1.0199
No log 9.8889 356 1.0314 0.2340 1.0314 1.0156
No log 9.9444 358 1.0276 0.2340 1.0276 1.0137
No log 10.0 360 1.0251 0.2340 1.0251 1.0125

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k7_task3_organization

Finetuned
(4023)
this model