ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k7_task5_organization
This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.9533
- Qwk: 0.6298
- Mse: 0.9533
- Rmse: 0.9764
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Qwk | Mse | Rmse |
|---|---|---|---|---|---|---|
| No log | 0.08 | 2 | 2.2980 | 0.0485 | 2.2980 | 1.5159 |
| No log | 0.16 | 4 | 1.4638 | 0.2102 | 1.4638 | 1.2099 |
| No log | 0.24 | 6 | 1.3461 | 0.1811 | 1.3461 | 1.1602 |
| No log | 0.32 | 8 | 1.4253 | 0.1788 | 1.4253 | 1.1939 |
| No log | 0.4 | 10 | 1.4584 | 0.1469 | 1.4584 | 1.2076 |
| No log | 0.48 | 12 | 1.5363 | 0.2859 | 1.5363 | 1.2395 |
| No log | 0.56 | 14 | 1.9154 | 0.2997 | 1.9154 | 1.3840 |
| No log | 0.64 | 16 | 2.2542 | 0.2219 | 2.2542 | 1.5014 |
| No log | 0.72 | 18 | 2.4478 | 0.1800 | 2.4478 | 1.5646 |
| No log | 0.8 | 20 | 2.3170 | 0.2435 | 2.3170 | 1.5222 |
| No log | 0.88 | 22 | 1.7271 | 0.2519 | 1.7271 | 1.3142 |
| No log | 0.96 | 24 | 1.7270 | 0.3201 | 1.7270 | 1.3141 |
| No log | 1.04 | 26 | 1.5881 | 0.2847 | 1.5881 | 1.2602 |
| No log | 1.12 | 28 | 1.6525 | 0.3213 | 1.6525 | 1.2855 |
| No log | 1.2 | 30 | 1.6781 | 0.3344 | 1.6781 | 1.2954 |
| No log | 1.28 | 32 | 1.7143 | 0.3344 | 1.7143 | 1.3093 |
| No log | 1.3600 | 34 | 1.7732 | 0.3257 | 1.7732 | 1.3316 |
| No log | 1.44 | 36 | 1.8122 | 0.3262 | 1.8122 | 1.3462 |
| No log | 1.52 | 38 | 1.7981 | 0.3262 | 1.7981 | 1.3409 |
| No log | 1.6 | 40 | 1.6100 | 0.3587 | 1.6100 | 1.2689 |
| No log | 1.6800 | 42 | 1.2922 | 0.4106 | 1.2922 | 1.1368 |
| No log | 1.76 | 44 | 1.1648 | 0.4618 | 1.1648 | 1.0793 |
| No log | 1.8400 | 46 | 1.1859 | 0.4696 | 1.1859 | 1.0890 |
| No log | 1.92 | 48 | 1.2562 | 0.4604 | 1.2562 | 1.1208 |
| No log | 2.0 | 50 | 1.4539 | 0.4038 | 1.4539 | 1.2058 |
| No log | 2.08 | 52 | 1.4547 | 0.4178 | 1.4547 | 1.2061 |
| No log | 2.16 | 54 | 1.3811 | 0.4635 | 1.3811 | 1.1752 |
| No log | 2.24 | 56 | 1.2009 | 0.5226 | 1.2009 | 1.0959 |
| No log | 2.32 | 58 | 1.0945 | 0.5817 | 1.0945 | 1.0462 |
| No log | 2.4 | 60 | 1.1179 | 0.5757 | 1.1179 | 1.0573 |
| No log | 2.48 | 62 | 1.4276 | 0.4877 | 1.4276 | 1.1948 |
| No log | 2.56 | 64 | 1.9512 | 0.3692 | 1.9512 | 1.3968 |
| No log | 2.64 | 66 | 2.1661 | 0.3364 | 2.1661 | 1.4718 |
| No log | 2.7200 | 68 | 2.1496 | 0.3793 | 2.1496 | 1.4662 |
| No log | 2.8 | 70 | 1.7506 | 0.4648 | 1.7506 | 1.3231 |
| No log | 2.88 | 72 | 1.2768 | 0.5313 | 1.2768 | 1.1300 |
| No log | 2.96 | 74 | 1.1092 | 0.5758 | 1.1092 | 1.0532 |
| No log | 3.04 | 76 | 1.1873 | 0.5621 | 1.1873 | 1.0896 |
| No log | 3.12 | 78 | 1.4974 | 0.5190 | 1.4974 | 1.2237 |
| No log | 3.2 | 80 | 1.8220 | 0.4463 | 1.8220 | 1.3498 |
| No log | 3.2800 | 82 | 1.8637 | 0.4252 | 1.8637 | 1.3652 |
| No log | 3.36 | 84 | 1.8321 | 0.4557 | 1.8321 | 1.3536 |
| No log | 3.44 | 86 | 1.7455 | 0.5020 | 1.7455 | 1.3212 |
| No log | 3.52 | 88 | 1.8166 | 0.4778 | 1.8166 | 1.3478 |
| No log | 3.6 | 90 | 1.6687 | 0.4807 | 1.6687 | 1.2918 |
| No log | 3.68 | 92 | 1.4959 | 0.5107 | 1.4959 | 1.2231 |
| No log | 3.76 | 94 | 1.2248 | 0.5641 | 1.2248 | 1.1067 |
| No log | 3.84 | 96 | 1.1880 | 0.5990 | 1.1880 | 1.0900 |
| No log | 3.92 | 98 | 1.2818 | 0.5498 | 1.2818 | 1.1322 |
| No log | 4.0 | 100 | 1.3837 | 0.5474 | 1.3837 | 1.1763 |
| No log | 4.08 | 102 | 1.4083 | 0.5365 | 1.4083 | 1.1867 |
| No log | 4.16 | 104 | 1.2970 | 0.5556 | 1.2970 | 1.1388 |
| No log | 4.24 | 106 | 1.3029 | 0.5670 | 1.3029 | 1.1415 |
| No log | 4.32 | 108 | 1.3623 | 0.5728 | 1.3623 | 1.1672 |
| No log | 4.4 | 110 | 1.1476 | 0.6496 | 1.1476 | 1.0713 |
| No log | 4.48 | 112 | 0.9073 | 0.6552 | 0.9073 | 0.9525 |
| No log | 4.5600 | 114 | 0.8698 | 0.6351 | 0.8698 | 0.9326 |
| No log | 4.64 | 116 | 0.9028 | 0.6672 | 0.9028 | 0.9502 |
| No log | 4.72 | 118 | 1.1343 | 0.6457 | 1.1343 | 1.0650 |
| No log | 4.8 | 120 | 1.5265 | 0.5591 | 1.5265 | 1.2355 |
| No log | 4.88 | 122 | 1.5606 | 0.5583 | 1.5606 | 1.2493 |
| No log | 4.96 | 124 | 1.3124 | 0.5728 | 1.3124 | 1.1456 |
| No log | 5.04 | 126 | 1.0187 | 0.6275 | 1.0187 | 1.0093 |
| No log | 5.12 | 128 | 0.9462 | 0.6411 | 0.9462 | 0.9727 |
| No log | 5.2 | 130 | 1.0489 | 0.6035 | 1.0489 | 1.0242 |
| No log | 5.28 | 132 | 1.1524 | 0.6060 | 1.1524 | 1.0735 |
| No log | 5.36 | 134 | 1.3237 | 0.5399 | 1.3237 | 1.1505 |
| No log | 5.44 | 136 | 1.3841 | 0.5452 | 1.3841 | 1.1765 |
| No log | 5.52 | 138 | 1.4322 | 0.5344 | 1.4322 | 1.1968 |
| No log | 5.6 | 140 | 1.3607 | 0.5507 | 1.3607 | 1.1665 |
| No log | 5.68 | 142 | 1.1963 | 0.5990 | 1.1963 | 1.0937 |
| No log | 5.76 | 144 | 1.0914 | 0.6051 | 1.0914 | 1.0447 |
| No log | 5.84 | 146 | 1.0736 | 0.6051 | 1.0736 | 1.0362 |
| No log | 5.92 | 148 | 1.0133 | 0.6238 | 1.0133 | 1.0066 |
| No log | 6.0 | 150 | 1.0427 | 0.6144 | 1.0427 | 1.0211 |
| No log | 6.08 | 152 | 1.1764 | 0.5984 | 1.1764 | 1.0846 |
| No log | 6.16 | 154 | 1.2062 | 0.5930 | 1.2062 | 1.0983 |
| No log | 6.24 | 156 | 1.1681 | 0.6054 | 1.1681 | 1.0808 |
| No log | 6.32 | 158 | 1.0897 | 0.6054 | 1.0897 | 1.0439 |
| No log | 6.4 | 160 | 0.9762 | 0.6003 | 0.9762 | 0.9880 |
| No log | 6.48 | 162 | 0.9631 | 0.6003 | 0.9631 | 0.9814 |
| No log | 6.5600 | 164 | 0.9620 | 0.5869 | 0.9620 | 0.9808 |
| No log | 6.64 | 166 | 0.9039 | 0.6153 | 0.9039 | 0.9507 |
| No log | 6.72 | 168 | 0.8867 | 0.6140 | 0.8867 | 0.9417 |
| No log | 6.8 | 170 | 0.8824 | 0.6140 | 0.8824 | 0.9394 |
| No log | 6.88 | 172 | 0.9015 | 0.6175 | 0.9015 | 0.9495 |
| No log | 6.96 | 174 | 0.9963 | 0.5847 | 0.9963 | 0.9981 |
| No log | 7.04 | 176 | 1.1182 | 0.5947 | 1.1182 | 1.0575 |
| No log | 7.12 | 178 | 1.2152 | 0.5835 | 1.2152 | 1.1023 |
| No log | 7.2 | 180 | 1.2231 | 0.5939 | 1.2231 | 1.1059 |
| No log | 7.28 | 182 | 1.1215 | 0.5967 | 1.1215 | 1.0590 |
| No log | 7.36 | 184 | 0.9730 | 0.6254 | 0.9730 | 0.9864 |
| No log | 7.44 | 186 | 0.8690 | 0.6291 | 0.8690 | 0.9322 |
| No log | 7.52 | 188 | 0.8328 | 0.6356 | 0.8328 | 0.9126 |
| No log | 7.6 | 190 | 0.8502 | 0.6345 | 0.8502 | 0.9221 |
| No log | 7.68 | 192 | 0.8955 | 0.6435 | 0.8955 | 0.9463 |
| No log | 7.76 | 194 | 0.9764 | 0.6451 | 0.9764 | 0.9881 |
| No log | 7.84 | 196 | 1.1028 | 0.6285 | 1.1028 | 1.0502 |
| No log | 7.92 | 198 | 1.1773 | 0.6101 | 1.1773 | 1.0851 |
| No log | 8.0 | 200 | 1.2402 | 0.5958 | 1.2402 | 1.1136 |
| No log | 8.08 | 202 | 1.2368 | 0.5958 | 1.2368 | 1.1121 |
| No log | 8.16 | 204 | 1.1552 | 0.6101 | 1.1552 | 1.0748 |
| No log | 8.24 | 206 | 1.0323 | 0.6419 | 1.0323 | 1.0160 |
| No log | 8.32 | 208 | 0.9411 | 0.6304 | 0.9411 | 0.9701 |
| No log | 8.4 | 210 | 0.8691 | 0.6516 | 0.8691 | 0.9323 |
| No log | 8.48 | 212 | 0.8517 | 0.6408 | 0.8517 | 0.9229 |
| No log | 8.56 | 214 | 0.8746 | 0.6563 | 0.8746 | 0.9352 |
| No log | 8.64 | 216 | 0.9054 | 0.6652 | 0.9054 | 0.9515 |
| No log | 8.72 | 218 | 0.9719 | 0.6427 | 0.9719 | 0.9859 |
| No log | 8.8 | 220 | 1.0421 | 0.6399 | 1.0421 | 1.0208 |
| No log | 8.88 | 222 | 1.0813 | 0.6201 | 1.0813 | 1.0398 |
| No log | 8.96 | 224 | 1.1142 | 0.6118 | 1.1142 | 1.0556 |
| No log | 9.04 | 226 | 1.1307 | 0.6178 | 1.1307 | 1.0633 |
| No log | 9.12 | 228 | 1.1329 | 0.6178 | 1.1329 | 1.0644 |
| No log | 9.2 | 230 | 1.1162 | 0.6133 | 1.1162 | 1.0565 |
| No log | 9.28 | 232 | 1.0754 | 0.6325 | 1.0754 | 1.0370 |
| No log | 9.36 | 234 | 1.0230 | 0.6409 | 1.0230 | 1.0114 |
| No log | 9.44 | 236 | 0.9975 | 0.6408 | 0.9975 | 0.9988 |
| No log | 9.52 | 238 | 0.9728 | 0.6451 | 0.9728 | 0.9863 |
| No log | 9.6 | 240 | 0.9590 | 0.6298 | 0.9590 | 0.9793 |
| No log | 9.68 | 242 | 0.9539 | 0.6298 | 0.9539 | 0.9767 |
| No log | 9.76 | 244 | 0.9521 | 0.6298 | 0.9521 | 0.9758 |
| No log | 9.84 | 246 | 0.9499 | 0.6298 | 0.9499 | 0.9747 |
| No log | 9.92 | 248 | 0.9517 | 0.6298 | 0.9517 | 0.9755 |
| No log | 10.0 | 250 | 0.9533 | 0.6298 | 0.9533 | 0.9764 |
Framework versions
- Transformers 4.44.2
- Pytorch 2.4.0+cu118
- Datasets 2.21.0
- Tokenizers 0.19.1
- Downloads last month
- 1
Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k7_task5_organization
Base model
aubmindlab/bert-base-arabertv02