ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k8_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6947
  • Qwk: 0.1675
  • Mse: 0.6947
  • Rmse: 0.8335

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.7111 0.0 3.7111 1.9264
No log 0.1111 4 2.4117 -0.0163 2.4117 1.5530
No log 0.1667 6 1.4622 0.0255 1.4622 1.2092
No log 0.2222 8 1.0519 0.0632 1.0519 1.0256
No log 0.2778 10 0.6496 0.1020 0.6496 0.8060
No log 0.3333 12 0.6119 0.0569 0.6119 0.7822
No log 0.3889 14 0.6217 0.0569 0.6217 0.7884
No log 0.4444 16 0.7361 0.1475 0.7361 0.8579
No log 0.5 18 0.6389 0.1030 0.6389 0.7993
No log 0.5556 20 0.6311 0.0569 0.6311 0.7944
No log 0.6111 22 0.6515 0.0 0.6515 0.8072
No log 0.6667 24 0.5680 0.0569 0.5680 0.7537
No log 0.7222 26 0.8891 0.0823 0.8891 0.9429
No log 0.7778 28 0.9588 0.0617 0.9588 0.9792
No log 0.8333 30 0.7530 0.1712 0.7530 0.8678
No log 0.8889 32 0.5959 0.0303 0.5959 0.7720
No log 0.9444 34 0.6484 0.0 0.6484 0.8052
No log 1.0 36 0.7096 0.0 0.7096 0.8424
No log 1.0556 38 0.6402 0.0 0.6402 0.8001
No log 1.1111 40 0.6079 0.0222 0.6079 0.7797
No log 1.1667 42 0.7106 0.1638 0.7106 0.8430
No log 1.2222 44 0.7749 0.0918 0.7749 0.8803
No log 1.2778 46 0.6087 0.1111 0.6087 0.7802
No log 1.3333 48 0.6242 0.0 0.6242 0.7901
No log 1.3889 50 0.6590 0.0 0.6590 0.8118
No log 1.4444 52 0.6369 0.0909 0.6369 0.7981
No log 1.5 54 0.7216 0.0409 0.7216 0.8495
No log 1.5556 56 1.1468 0.0888 1.1468 1.0709
No log 1.6111 58 1.0264 0.0357 1.0264 1.0131
No log 1.6667 60 0.7527 0.1195 0.7527 0.8676
No log 1.7222 62 0.8996 0.0045 0.8996 0.9485
No log 1.7778 64 0.9511 -0.0396 0.9511 0.9753
No log 1.8333 66 0.8171 0.1186 0.8171 0.9039
No log 1.8889 68 0.8384 0.0417 0.8384 0.9156
No log 1.9444 70 0.7068 -0.0115 0.7068 0.8407
No log 2.0 72 1.1652 0.0040 1.1652 1.0794
No log 2.0556 74 1.5367 -0.0323 1.5367 1.2396
No log 2.1111 76 0.9921 0.0442 0.9921 0.9960
No log 2.1667 78 0.6773 0.2169 0.6773 0.8230
No log 2.2222 80 0.6838 0.1919 0.6838 0.8269
No log 2.2778 82 0.7907 0.1579 0.7907 0.8892
No log 2.3333 84 0.9282 0.0044 0.9282 0.9634
No log 2.3889 86 1.3364 0.1049 1.3364 1.1560
No log 2.4444 88 1.3189 0.1304 1.3189 1.1484
No log 2.5 90 0.6844 0.3371 0.6844 0.8273
No log 2.5556 92 0.6332 0.3333 0.6332 0.7957
No log 2.6111 94 0.7316 0.2332 0.7316 0.8554
No log 2.6667 96 1.5324 0.1084 1.5324 1.2379
No log 2.7222 98 1.7160 0.0659 1.7160 1.3100
No log 2.7778 100 0.8899 0.1111 0.8899 0.9433
No log 2.8333 102 0.8783 0.1712 0.8783 0.9372
No log 2.8889 104 1.1429 0.1571 1.1429 1.0691
No log 2.9444 106 0.7111 0.1675 0.7111 0.8433
No log 3.0 108 1.0955 0.0406 1.0955 1.0467
No log 3.0556 110 1.5839 0.0788 1.5839 1.2585
No log 3.1111 112 1.3248 0.0278 1.3248 1.1510
No log 3.1667 114 0.7078 0.2195 0.7078 0.8413
No log 3.2222 116 0.7698 0.0833 0.7698 0.8774
No log 3.2778 118 0.8155 0.0680 0.8155 0.9030
No log 3.3333 120 0.6633 0.1801 0.6633 0.8145
No log 3.3889 122 0.9323 0.1504 0.9323 0.9655
No log 3.4444 124 0.9197 0.1504 0.9197 0.9590
No log 3.5 126 0.7194 0.1345 0.7194 0.8482
No log 3.5556 128 0.9082 0.0685 0.9082 0.9530
No log 3.6111 130 0.9033 0.0631 0.9033 0.9504
No log 3.6667 132 0.7376 0.1732 0.7376 0.8589
No log 3.7222 134 0.9603 0.0769 0.9603 0.9799
No log 3.7778 136 0.8652 0.1287 0.8652 0.9301
No log 3.8333 138 0.7199 0.1732 0.7199 0.8485
No log 3.8889 140 0.7234 0.1364 0.7234 0.8505
No log 3.9444 142 0.7667 0.1828 0.7667 0.8756
No log 4.0 144 0.7989 0.1429 0.7989 0.8938
No log 4.0556 146 0.8139 0.1443 0.8139 0.9022
No log 4.1111 148 0.8245 0.1527 0.8245 0.9080
No log 4.1667 150 0.8026 0.1515 0.8026 0.8959
No log 4.2222 152 0.7831 0.2350 0.7831 0.8849
No log 4.2778 154 0.9611 0.0744 0.9611 0.9804
No log 4.3333 156 1.3294 0.1351 1.3294 1.1530
No log 4.3889 158 1.1538 0.0929 1.1538 1.0742
No log 4.4444 160 0.7771 0.2897 0.7771 0.8815
No log 4.5 162 0.7825 0.2838 0.7825 0.8846
No log 4.5556 164 0.9562 0.2069 0.9562 0.9778
No log 4.6111 166 1.3205 0.1560 1.3205 1.1491
No log 4.6667 168 1.1807 0.1572 1.1807 1.0866
No log 4.7222 170 0.7151 0.2251 0.7151 0.8456
No log 4.7778 172 0.6254 0.3023 0.6254 0.7908
No log 4.8333 174 0.6918 0.2360 0.6918 0.8318
No log 4.8889 176 0.8978 0.1864 0.8978 0.9475
No log 4.9444 178 0.7463 0.2941 0.7463 0.8639
No log 5.0 180 0.6126 0.3103 0.6126 0.7827
No log 5.0556 182 0.5966 0.3103 0.5966 0.7724
No log 5.1111 184 0.6572 0.2350 0.6572 0.8107
No log 5.1667 186 0.7763 0.2453 0.7763 0.8811
No log 5.2222 188 1.1121 0.1882 1.1121 1.0545
No log 5.2778 190 0.9522 0.2129 0.9522 0.9758
No log 5.3333 192 0.7115 0.1921 0.7115 0.8435
No log 5.3889 194 0.6371 0.3043 0.6371 0.7982
No log 5.4444 196 0.6441 0.3297 0.6441 0.8026
No log 5.5 198 0.7819 0.1373 0.7819 0.8842
No log 5.5556 200 1.0166 0.2119 1.0166 1.0083
No log 5.6111 202 0.9517 0.2126 0.9517 0.9755
No log 5.6667 204 0.7567 0.1373 0.7567 0.8699
No log 5.7222 206 0.7618 0.1269 0.7618 0.8728
No log 5.7778 208 0.8680 0.1781 0.8680 0.9317
No log 5.8333 210 0.9054 0.1781 0.9054 0.9515
No log 5.8889 212 0.7296 0.1915 0.7296 0.8542
No log 5.9444 214 0.7102 0.1398 0.7102 0.8427
No log 6.0 216 0.8339 0.1287 0.8339 0.9132
No log 6.0556 218 1.1690 0.1601 1.1690 1.0812
No log 6.1111 220 1.3049 0.1634 1.3049 1.1423
No log 6.1667 222 1.0334 0.1815 1.0334 1.0166
No log 6.2222 224 0.7347 0.2410 0.7347 0.8571
No log 6.2778 226 0.6372 0.2558 0.6372 0.7982
No log 6.3333 228 0.6287 0.3455 0.6287 0.7929
No log 6.3889 230 0.6358 0.1902 0.6358 0.7974
No log 6.4444 232 0.7067 0.125 0.7067 0.8407
No log 6.5 234 0.7805 0.1373 0.7805 0.8834
No log 6.5556 236 0.8392 0.1388 0.8392 0.9161
No log 6.6111 238 0.8276 0.1321 0.8276 0.9098
No log 6.6667 240 0.8603 0.1321 0.8603 0.9275
No log 6.7222 242 0.9161 0.1781 0.9161 0.9571
No log 6.7778 244 0.8036 0.1321 0.8036 0.8965
No log 6.8333 246 0.7088 0.1475 0.7088 0.8419
No log 6.8889 248 0.7146 0.1088 0.7146 0.8453
No log 6.9444 250 0.7385 0.1340 0.7385 0.8594
No log 7.0 252 0.7377 0.1340 0.7377 0.8589
No log 7.0556 254 0.7251 0.1269 0.7251 0.8515
No log 7.1111 256 0.6848 0.1828 0.6848 0.8275
No log 7.1667 258 0.6769 0.2688 0.6769 0.8228
No log 7.2222 260 0.6956 0.3333 0.6956 0.8340
No log 7.2778 262 0.7275 0.1340 0.7275 0.8529
No log 7.3333 264 0.8131 0.1456 0.8131 0.9017
No log 7.3889 266 0.9727 0.1795 0.9727 0.9863
No log 7.4444 268 0.9668 0.1795 0.9668 0.9833
No log 7.5 270 0.8493 0.0189 0.8493 0.9216
No log 7.5556 272 0.7374 0.1269 0.7374 0.8587
No log 7.6111 274 0.7209 0.3016 0.7209 0.8491
No log 7.6667 276 0.7249 0.3089 0.7249 0.8514
No log 7.7222 278 0.7048 0.2766 0.7048 0.8395
No log 7.7778 280 0.6818 0.3369 0.6818 0.8257
No log 7.8333 282 0.6630 0.2179 0.6630 0.8143
No log 7.8889 284 0.7102 0.1675 0.7102 0.8427
No log 7.9444 286 0.7658 0.0980 0.7658 0.8751
No log 8.0 288 0.7375 0.1287 0.7375 0.8588
No log 8.0556 290 0.6749 0.2165 0.6749 0.8215
No log 8.1111 292 0.6572 0.2179 0.6572 0.8107
No log 8.1667 294 0.6607 0.2179 0.6607 0.8128
No log 8.2222 296 0.6851 0.2536 0.6851 0.8277
No log 8.2778 298 0.7691 0.1402 0.7691 0.8770
No log 8.3333 300 0.8190 0.1416 0.8190 0.9050
No log 8.3889 302 0.7985 0.1402 0.7985 0.8936
No log 8.4444 304 0.7887 0.1402 0.7887 0.8881
No log 8.5 306 0.7871 0.1402 0.7871 0.8872
No log 8.5556 308 0.7521 0.0980 0.7521 0.8672
No log 8.6111 310 0.7039 0.2536 0.7039 0.8390
No log 8.6667 312 0.6633 0.2563 0.6633 0.8144
No log 8.7222 314 0.6456 0.2542 0.6456 0.8035
No log 8.7778 316 0.6435 0.2542 0.6435 0.8022
No log 8.8333 318 0.6594 0.2563 0.6594 0.8120
No log 8.8889 320 0.6989 0.2233 0.6989 0.8360
No log 8.9444 322 0.7628 0.0943 0.7628 0.8734
No log 9.0 324 0.8416 0.1429 0.8416 0.9174
No log 9.0556 326 0.8984 0.1504 0.8984 0.9478
No log 9.1111 328 0.9080 0.1515 0.9080 0.9529
No log 9.1667 330 0.8691 0.1855 0.8691 0.9323
No log 9.2222 332 0.7984 0.1416 0.7984 0.8936
No log 9.2778 334 0.7296 0.1304 0.7296 0.8542
No log 9.3333 336 0.6901 0.1675 0.6901 0.8307
No log 9.3889 338 0.6645 0.1503 0.6645 0.8152
No log 9.4444 340 0.6491 0.2563 0.6491 0.8057
No log 9.5 342 0.6451 0.2563 0.6451 0.8032
No log 9.5556 344 0.6510 0.2563 0.6510 0.8069
No log 9.6111 346 0.6587 0.1503 0.6587 0.8116
No log 9.6667 348 0.6668 0.1503 0.6668 0.8166
No log 9.7222 350 0.6755 0.1919 0.6755 0.8219
No log 9.7778 352 0.6869 0.1675 0.6869 0.8288
No log 9.8333 354 0.6929 0.1675 0.6929 0.8324
No log 9.8889 356 0.6949 0.1675 0.6949 0.8336
No log 9.9444 358 0.6947 0.1675 0.6947 0.8335
No log 10.0 360 0.6947 0.1675 0.6947 0.8335

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k8_task3_organization

Finetuned
(4023)
this model