ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k7_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9533
  • Qwk: 0.6298
  • Mse: 0.9533
  • Rmse: 0.9764

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 2.2980 0.0485 2.2980 1.5159
No log 0.16 4 1.4638 0.2102 1.4638 1.2099
No log 0.24 6 1.3461 0.1811 1.3461 1.1602
No log 0.32 8 1.4253 0.1788 1.4253 1.1939
No log 0.4 10 1.4584 0.1469 1.4584 1.2076
No log 0.48 12 1.5363 0.2859 1.5363 1.2395
No log 0.56 14 1.9154 0.2997 1.9154 1.3840
No log 0.64 16 2.2542 0.2219 2.2542 1.5014
No log 0.72 18 2.4478 0.1800 2.4478 1.5646
No log 0.8 20 2.3170 0.2435 2.3170 1.5222
No log 0.88 22 1.7271 0.2519 1.7271 1.3142
No log 0.96 24 1.7270 0.3201 1.7270 1.3141
No log 1.04 26 1.5881 0.2847 1.5881 1.2602
No log 1.12 28 1.6525 0.3213 1.6525 1.2855
No log 1.2 30 1.6781 0.3344 1.6781 1.2954
No log 1.28 32 1.7143 0.3344 1.7143 1.3093
No log 1.3600 34 1.7732 0.3257 1.7732 1.3316
No log 1.44 36 1.8122 0.3262 1.8122 1.3462
No log 1.52 38 1.7981 0.3262 1.7981 1.3409
No log 1.6 40 1.6100 0.3587 1.6100 1.2689
No log 1.6800 42 1.2922 0.4106 1.2922 1.1368
No log 1.76 44 1.1648 0.4618 1.1648 1.0793
No log 1.8400 46 1.1859 0.4696 1.1859 1.0890
No log 1.92 48 1.2562 0.4604 1.2562 1.1208
No log 2.0 50 1.4539 0.4038 1.4539 1.2058
No log 2.08 52 1.4547 0.4178 1.4547 1.2061
No log 2.16 54 1.3811 0.4635 1.3811 1.1752
No log 2.24 56 1.2009 0.5226 1.2009 1.0959
No log 2.32 58 1.0945 0.5817 1.0945 1.0462
No log 2.4 60 1.1179 0.5757 1.1179 1.0573
No log 2.48 62 1.4276 0.4877 1.4276 1.1948
No log 2.56 64 1.9512 0.3692 1.9512 1.3968
No log 2.64 66 2.1661 0.3364 2.1661 1.4718
No log 2.7200 68 2.1496 0.3793 2.1496 1.4662
No log 2.8 70 1.7506 0.4648 1.7506 1.3231
No log 2.88 72 1.2768 0.5313 1.2768 1.1300
No log 2.96 74 1.1092 0.5758 1.1092 1.0532
No log 3.04 76 1.1873 0.5621 1.1873 1.0896
No log 3.12 78 1.4974 0.5190 1.4974 1.2237
No log 3.2 80 1.8220 0.4463 1.8220 1.3498
No log 3.2800 82 1.8637 0.4252 1.8637 1.3652
No log 3.36 84 1.8321 0.4557 1.8321 1.3536
No log 3.44 86 1.7455 0.5020 1.7455 1.3212
No log 3.52 88 1.8166 0.4778 1.8166 1.3478
No log 3.6 90 1.6687 0.4807 1.6687 1.2918
No log 3.68 92 1.4959 0.5107 1.4959 1.2231
No log 3.76 94 1.2248 0.5641 1.2248 1.1067
No log 3.84 96 1.1880 0.5990 1.1880 1.0900
No log 3.92 98 1.2818 0.5498 1.2818 1.1322
No log 4.0 100 1.3837 0.5474 1.3837 1.1763
No log 4.08 102 1.4083 0.5365 1.4083 1.1867
No log 4.16 104 1.2970 0.5556 1.2970 1.1388
No log 4.24 106 1.3029 0.5670 1.3029 1.1415
No log 4.32 108 1.3623 0.5728 1.3623 1.1672
No log 4.4 110 1.1476 0.6496 1.1476 1.0713
No log 4.48 112 0.9073 0.6552 0.9073 0.9525
No log 4.5600 114 0.8698 0.6351 0.8698 0.9326
No log 4.64 116 0.9028 0.6672 0.9028 0.9502
No log 4.72 118 1.1343 0.6457 1.1343 1.0650
No log 4.8 120 1.5265 0.5591 1.5265 1.2355
No log 4.88 122 1.5606 0.5583 1.5606 1.2493
No log 4.96 124 1.3124 0.5728 1.3124 1.1456
No log 5.04 126 1.0187 0.6275 1.0187 1.0093
No log 5.12 128 0.9462 0.6411 0.9462 0.9727
No log 5.2 130 1.0489 0.6035 1.0489 1.0242
No log 5.28 132 1.1524 0.6060 1.1524 1.0735
No log 5.36 134 1.3237 0.5399 1.3237 1.1505
No log 5.44 136 1.3841 0.5452 1.3841 1.1765
No log 5.52 138 1.4322 0.5344 1.4322 1.1968
No log 5.6 140 1.3607 0.5507 1.3607 1.1665
No log 5.68 142 1.1963 0.5990 1.1963 1.0937
No log 5.76 144 1.0914 0.6051 1.0914 1.0447
No log 5.84 146 1.0736 0.6051 1.0736 1.0362
No log 5.92 148 1.0133 0.6238 1.0133 1.0066
No log 6.0 150 1.0427 0.6144 1.0427 1.0211
No log 6.08 152 1.1764 0.5984 1.1764 1.0846
No log 6.16 154 1.2062 0.5930 1.2062 1.0983
No log 6.24 156 1.1681 0.6054 1.1681 1.0808
No log 6.32 158 1.0897 0.6054 1.0897 1.0439
No log 6.4 160 0.9762 0.6003 0.9762 0.9880
No log 6.48 162 0.9631 0.6003 0.9631 0.9814
No log 6.5600 164 0.9620 0.5869 0.9620 0.9808
No log 6.64 166 0.9039 0.6153 0.9039 0.9507
No log 6.72 168 0.8867 0.6140 0.8867 0.9417
No log 6.8 170 0.8824 0.6140 0.8824 0.9394
No log 6.88 172 0.9015 0.6175 0.9015 0.9495
No log 6.96 174 0.9963 0.5847 0.9963 0.9981
No log 7.04 176 1.1182 0.5947 1.1182 1.0575
No log 7.12 178 1.2152 0.5835 1.2152 1.1023
No log 7.2 180 1.2231 0.5939 1.2231 1.1059
No log 7.28 182 1.1215 0.5967 1.1215 1.0590
No log 7.36 184 0.9730 0.6254 0.9730 0.9864
No log 7.44 186 0.8690 0.6291 0.8690 0.9322
No log 7.52 188 0.8328 0.6356 0.8328 0.9126
No log 7.6 190 0.8502 0.6345 0.8502 0.9221
No log 7.68 192 0.8955 0.6435 0.8955 0.9463
No log 7.76 194 0.9764 0.6451 0.9764 0.9881
No log 7.84 196 1.1028 0.6285 1.1028 1.0502
No log 7.92 198 1.1773 0.6101 1.1773 1.0851
No log 8.0 200 1.2402 0.5958 1.2402 1.1136
No log 8.08 202 1.2368 0.5958 1.2368 1.1121
No log 8.16 204 1.1552 0.6101 1.1552 1.0748
No log 8.24 206 1.0323 0.6419 1.0323 1.0160
No log 8.32 208 0.9411 0.6304 0.9411 0.9701
No log 8.4 210 0.8691 0.6516 0.8691 0.9323
No log 8.48 212 0.8517 0.6408 0.8517 0.9229
No log 8.56 214 0.8746 0.6563 0.8746 0.9352
No log 8.64 216 0.9054 0.6652 0.9054 0.9515
No log 8.72 218 0.9719 0.6427 0.9719 0.9859
No log 8.8 220 1.0421 0.6399 1.0421 1.0208
No log 8.88 222 1.0813 0.6201 1.0813 1.0398
No log 8.96 224 1.1142 0.6118 1.1142 1.0556
No log 9.04 226 1.1307 0.6178 1.1307 1.0633
No log 9.12 228 1.1329 0.6178 1.1329 1.0644
No log 9.2 230 1.1162 0.6133 1.1162 1.0565
No log 9.28 232 1.0754 0.6325 1.0754 1.0370
No log 9.36 234 1.0230 0.6409 1.0230 1.0114
No log 9.44 236 0.9975 0.6408 0.9975 0.9988
No log 9.52 238 0.9728 0.6451 0.9728 0.9863
No log 9.6 240 0.9590 0.6298 0.9590 0.9793
No log 9.68 242 0.9539 0.6298 0.9539 0.9767
No log 9.76 244 0.9521 0.6298 0.9521 0.9758
No log 9.84 246 0.9499 0.6298 0.9499 0.9747
No log 9.92 248 0.9517 0.6298 0.9517 0.9755
No log 10.0 250 0.9533 0.6298 0.9533 0.9764

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k7_task5_organization

Finetuned
(4023)
this model