ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k6_task5_organization
This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:
- Loss: 1.1952
- Qwk: 0.6156
- Mse: 1.1952
- Rmse: 1.0932
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Qwk | Mse | Rmse |
|---|---|---|---|---|---|---|
| No log | 0.08 | 2 | 2.4318 | 0.0010 | 2.4318 | 1.5594 |
| No log | 0.16 | 4 | 1.6702 | 0.1238 | 1.6702 | 1.2924 |
| No log | 0.24 | 6 | 1.3137 | 0.1528 | 1.3137 | 1.1462 |
| No log | 0.32 | 8 | 1.5339 | 0.1550 | 1.5339 | 1.2385 |
| No log | 0.4 | 10 | 1.6410 | 0.0393 | 1.6410 | 1.2810 |
| No log | 0.48 | 12 | 1.5840 | 0.1940 | 1.5840 | 1.2586 |
| No log | 0.56 | 14 | 1.5331 | 0.2132 | 1.5331 | 1.2382 |
| No log | 0.64 | 16 | 1.6071 | 0.2310 | 1.6071 | 1.2677 |
| No log | 0.72 | 18 | 1.7480 | 0.2373 | 1.7480 | 1.3221 |
| No log | 0.8 | 20 | 1.6493 | 0.3072 | 1.6493 | 1.2842 |
| No log | 0.88 | 22 | 1.6963 | 0.3175 | 1.6963 | 1.3024 |
| No log | 0.96 | 24 | 1.6261 | 0.3295 | 1.6261 | 1.2752 |
| No log | 1.04 | 26 | 1.4373 | 0.3743 | 1.4373 | 1.1989 |
| No log | 1.12 | 28 | 1.3960 | 0.4291 | 1.3960 | 1.1815 |
| No log | 1.2 | 30 | 1.2818 | 0.4622 | 1.2818 | 1.1322 |
| No log | 1.28 | 32 | 1.3464 | 0.4400 | 1.3464 | 1.1603 |
| No log | 1.3600 | 34 | 1.5466 | 0.4565 | 1.5466 | 1.2436 |
| No log | 1.44 | 36 | 1.5268 | 0.4565 | 1.5268 | 1.2356 |
| No log | 1.52 | 38 | 1.3652 | 0.4705 | 1.3652 | 1.1684 |
| No log | 1.6 | 40 | 1.1954 | 0.4580 | 1.1954 | 1.0933 |
| No log | 1.6800 | 42 | 1.0662 | 0.4454 | 1.0662 | 1.0326 |
| No log | 1.76 | 44 | 1.0232 | 0.4710 | 1.0232 | 1.0115 |
| No log | 1.8400 | 46 | 1.0087 | 0.4548 | 1.0087 | 1.0043 |
| No log | 1.92 | 48 | 1.0057 | 0.4569 | 1.0057 | 1.0029 |
| No log | 2.0 | 50 | 1.1358 | 0.4712 | 1.1358 | 1.0658 |
| No log | 2.08 | 52 | 1.2720 | 0.4675 | 1.2720 | 1.1279 |
| No log | 2.16 | 54 | 1.2885 | 0.4977 | 1.2885 | 1.1351 |
| No log | 2.24 | 56 | 1.3784 | 0.4817 | 1.3784 | 1.1740 |
| No log | 2.32 | 58 | 1.3184 | 0.5067 | 1.3184 | 1.1482 |
| No log | 2.4 | 60 | 1.1251 | 0.4693 | 1.1251 | 1.0607 |
| No log | 2.48 | 62 | 1.1514 | 0.4634 | 1.1514 | 1.0730 |
| No log | 2.56 | 64 | 1.4748 | 0.4973 | 1.4748 | 1.2144 |
| No log | 2.64 | 66 | 1.5275 | 0.4963 | 1.5275 | 1.2359 |
| No log | 2.7200 | 68 | 1.5623 | 0.5094 | 1.5623 | 1.2499 |
| No log | 2.8 | 70 | 1.2982 | 0.5354 | 1.2982 | 1.1394 |
| No log | 2.88 | 72 | 1.0562 | 0.5025 | 1.0562 | 1.0277 |
| No log | 2.96 | 74 | 1.0919 | 0.5234 | 1.0919 | 1.0449 |
| No log | 3.04 | 76 | 1.5255 | 0.5239 | 1.5255 | 1.2351 |
| No log | 3.12 | 78 | 1.7981 | 0.4848 | 1.7981 | 1.3409 |
| No log | 3.2 | 80 | 1.6918 | 0.5031 | 1.6918 | 1.3007 |
| No log | 3.2800 | 82 | 1.1956 | 0.6008 | 1.1956 | 1.0934 |
| No log | 3.36 | 84 | 0.9030 | 0.6235 | 0.9030 | 0.9502 |
| No log | 3.44 | 86 | 0.8909 | 0.6350 | 0.8909 | 0.9438 |
| No log | 3.52 | 88 | 1.0038 | 0.5739 | 1.0038 | 1.0019 |
| No log | 3.6 | 90 | 1.4083 | 0.5280 | 1.4083 | 1.1867 |
| No log | 3.68 | 92 | 1.6602 | 0.4998 | 1.6602 | 1.2885 |
| No log | 3.76 | 94 | 1.5378 | 0.5199 | 1.5378 | 1.2401 |
| No log | 3.84 | 96 | 1.3449 | 0.5453 | 1.3449 | 1.1597 |
| No log | 3.92 | 98 | 1.3108 | 0.5564 | 1.3108 | 1.1449 |
| No log | 4.0 | 100 | 1.3871 | 0.5638 | 1.3871 | 1.1777 |
| No log | 4.08 | 102 | 1.3878 | 0.5638 | 1.3878 | 1.1780 |
| No log | 4.16 | 104 | 1.3663 | 0.5880 | 1.3663 | 1.1689 |
| No log | 4.24 | 106 | 1.3619 | 0.5874 | 1.3619 | 1.1670 |
| No log | 4.32 | 108 | 1.2841 | 0.6096 | 1.2841 | 1.1332 |
| No log | 4.4 | 110 | 1.2495 | 0.6096 | 1.2495 | 1.1178 |
| No log | 4.48 | 112 | 1.4535 | 0.5840 | 1.4535 | 1.2056 |
| No log | 4.5600 | 114 | 1.6155 | 0.5812 | 1.6155 | 1.2710 |
| No log | 4.64 | 116 | 1.5852 | 0.5812 | 1.5852 | 1.2590 |
| No log | 4.72 | 118 | 1.4995 | 0.5674 | 1.4995 | 1.2245 |
| No log | 4.8 | 120 | 1.2242 | 0.6151 | 1.2242 | 1.1064 |
| No log | 4.88 | 122 | 1.0272 | 0.6397 | 1.0272 | 1.0135 |
| No log | 4.96 | 124 | 0.9846 | 0.6293 | 0.9846 | 0.9923 |
| No log | 5.04 | 126 | 0.9252 | 0.6163 | 0.9252 | 0.9619 |
| No log | 5.12 | 128 | 0.9781 | 0.6105 | 0.9781 | 0.9890 |
| No log | 5.2 | 130 | 1.0533 | 0.6064 | 1.0533 | 1.0263 |
| No log | 5.28 | 132 | 1.1317 | 0.5617 | 1.1317 | 1.0638 |
| No log | 5.36 | 134 | 1.2297 | 0.6022 | 1.2297 | 1.1089 |
| No log | 5.44 | 136 | 1.3741 | 0.5569 | 1.3741 | 1.1722 |
| No log | 5.52 | 138 | 1.3062 | 0.5726 | 1.3062 | 1.1429 |
| No log | 5.6 | 140 | 1.1382 | 0.5471 | 1.1382 | 1.0668 |
| No log | 5.68 | 142 | 1.0213 | 0.5751 | 1.0213 | 1.0106 |
| No log | 5.76 | 144 | 1.0852 | 0.5864 | 1.0852 | 1.0417 |
| No log | 5.84 | 146 | 1.3130 | 0.5879 | 1.3130 | 1.1459 |
| No log | 5.92 | 148 | 1.6371 | 0.5109 | 1.6371 | 1.2795 |
| No log | 6.0 | 150 | 1.6530 | 0.5109 | 1.6530 | 1.2857 |
| No log | 6.08 | 152 | 1.4481 | 0.5473 | 1.4481 | 1.2034 |
| No log | 6.16 | 154 | 1.1785 | 0.5990 | 1.1785 | 1.0856 |
| No log | 6.24 | 156 | 1.0236 | 0.5987 | 1.0236 | 1.0117 |
| No log | 6.32 | 158 | 1.0104 | 0.5987 | 1.0104 | 1.0052 |
| No log | 6.4 | 160 | 1.0748 | 0.6004 | 1.0748 | 1.0367 |
| No log | 6.48 | 162 | 1.2348 | 0.5797 | 1.2348 | 1.1112 |
| No log | 6.5600 | 164 | 1.4122 | 0.5547 | 1.4122 | 1.1884 |
| No log | 6.64 | 166 | 1.4189 | 0.5547 | 1.4189 | 1.1912 |
| No log | 6.72 | 168 | 1.4205 | 0.5609 | 1.4205 | 1.1919 |
| No log | 6.8 | 170 | 1.3790 | 0.5687 | 1.3790 | 1.1743 |
| No log | 6.88 | 172 | 1.3199 | 0.5786 | 1.3199 | 1.1489 |
| No log | 6.96 | 174 | 1.2592 | 0.5693 | 1.2592 | 1.1221 |
| No log | 7.04 | 176 | 1.3199 | 0.5786 | 1.3199 | 1.1489 |
| No log | 7.12 | 178 | 1.5071 | 0.5383 | 1.5071 | 1.2276 |
| No log | 7.2 | 180 | 1.7504 | 0.4919 | 1.7504 | 1.3230 |
| No log | 7.28 | 182 | 1.8340 | 0.4869 | 1.8340 | 1.3543 |
| No log | 7.36 | 184 | 1.7631 | 0.4989 | 1.7631 | 1.3278 |
| No log | 7.44 | 186 | 1.5568 | 0.5304 | 1.5568 | 1.2477 |
| No log | 7.52 | 188 | 1.2829 | 0.5837 | 1.2829 | 1.1326 |
| No log | 7.6 | 190 | 1.1311 | 0.6042 | 1.1311 | 1.0635 |
| No log | 7.68 | 192 | 1.0340 | 0.6086 | 1.0340 | 1.0169 |
| No log | 7.76 | 194 | 1.0348 | 0.6103 | 1.0348 | 1.0173 |
| No log | 7.84 | 196 | 1.1023 | 0.6022 | 1.1023 | 1.0499 |
| No log | 7.92 | 198 | 1.1858 | 0.6035 | 1.1858 | 1.0889 |
| No log | 8.0 | 200 | 1.3107 | 0.5765 | 1.3107 | 1.1449 |
| No log | 8.08 | 202 | 1.4087 | 0.5333 | 1.4087 | 1.1869 |
| No log | 8.16 | 204 | 1.3942 | 0.5280 | 1.3942 | 1.1808 |
| No log | 8.24 | 206 | 1.2923 | 0.5664 | 1.2923 | 1.1368 |
| No log | 8.32 | 208 | 1.1652 | 0.5956 | 1.1652 | 1.0794 |
| No log | 8.4 | 210 | 1.0756 | 0.6039 | 1.0756 | 1.0371 |
| No log | 8.48 | 212 | 1.0379 | 0.6129 | 1.0379 | 1.0188 |
| No log | 8.56 | 214 | 1.0403 | 0.6129 | 1.0403 | 1.0200 |
| No log | 8.64 | 216 | 1.0810 | 0.5983 | 1.0810 | 1.0397 |
| No log | 8.72 | 218 | 1.1612 | 0.6072 | 1.1612 | 1.0776 |
| No log | 8.8 | 220 | 1.2419 | 0.5745 | 1.2419 | 1.1144 |
| No log | 8.88 | 222 | 1.2898 | 0.5615 | 1.2898 | 1.1357 |
| No log | 8.96 | 224 | 1.3030 | 0.5615 | 1.3030 | 1.1415 |
| No log | 9.04 | 226 | 1.3107 | 0.5615 | 1.3107 | 1.1449 |
| No log | 9.12 | 228 | 1.3335 | 0.5615 | 1.3335 | 1.1548 |
| No log | 9.2 | 230 | 1.3326 | 0.5615 | 1.3326 | 1.1544 |
| No log | 9.28 | 232 | 1.3049 | 0.5725 | 1.3049 | 1.1423 |
| No log | 9.36 | 234 | 1.2707 | 0.5834 | 1.2707 | 1.1272 |
| No log | 9.44 | 236 | 1.2200 | 0.6108 | 1.2200 | 1.1045 |
| No log | 9.52 | 238 | 1.1930 | 0.6089 | 1.1930 | 1.0922 |
| No log | 9.6 | 240 | 1.1834 | 0.6089 | 1.1834 | 1.0878 |
| No log | 9.68 | 242 | 1.1736 | 0.6141 | 1.1736 | 1.0833 |
| No log | 9.76 | 244 | 1.1756 | 0.6141 | 1.1756 | 1.0843 |
| No log | 9.84 | 246 | 1.1855 | 0.6156 | 1.1855 | 1.0888 |
| No log | 9.92 | 248 | 1.1915 | 0.6156 | 1.1915 | 1.0916 |
| No log | 10.0 | 250 | 1.1952 | 0.6156 | 1.1952 | 1.0932 |
Framework versions
- Transformers 4.44.2
- Pytorch 2.4.0+cu118
- Datasets 2.21.0
- Tokenizers 0.19.1
- Downloads last month
- 2
Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k6_task5_organization
Base model
aubmindlab/bert-base-arabertv02