ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k5_task2_organization
This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:
- Loss: 0.9364
- Qwk: 0.4216
- Mse: 0.9364
- Rmse: 0.9677
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Qwk | Mse | Rmse |
|---|---|---|---|---|---|---|
| No log | 0.0588 | 2 | 4.0665 | -0.0151 | 4.0665 | 2.0166 |
| No log | 0.1176 | 4 | 2.0305 | 0.0284 | 2.0305 | 1.4250 |
| No log | 0.1765 | 6 | 1.1765 | 0.0592 | 1.1765 | 1.0846 |
| No log | 0.2353 | 8 | 0.8194 | 0.0770 | 0.8194 | 0.9052 |
| No log | 0.2941 | 10 | 0.9559 | 0.1514 | 0.9559 | 0.9777 |
| No log | 0.3529 | 12 | 1.1527 | 0.1040 | 1.1527 | 1.0736 |
| No log | 0.4118 | 14 | 0.8714 | 0.1748 | 0.8714 | 0.9335 |
| No log | 0.4706 | 16 | 0.6749 | 0.2880 | 0.6749 | 0.8215 |
| No log | 0.5294 | 18 | 0.6440 | 0.3611 | 0.6440 | 0.8025 |
| No log | 0.5882 | 20 | 0.8069 | 0.1337 | 0.8069 | 0.8983 |
| No log | 0.6471 | 22 | 0.9382 | 0.1106 | 0.9382 | 0.9686 |
| No log | 0.7059 | 24 | 0.9534 | 0.1590 | 0.9534 | 0.9764 |
| No log | 0.7647 | 26 | 1.1871 | 0.1583 | 1.1871 | 1.0895 |
| No log | 0.8235 | 28 | 1.5034 | 0.1392 | 1.5034 | 1.2261 |
| No log | 0.8824 | 30 | 1.8216 | 0.1717 | 1.8216 | 1.3497 |
| No log | 0.9412 | 32 | 1.7859 | 0.1794 | 1.7859 | 1.3364 |
| No log | 1.0 | 34 | 1.4960 | 0.1639 | 1.4960 | 1.2231 |
| No log | 1.0588 | 36 | 1.7451 | 0.1670 | 1.7451 | 1.3210 |
| No log | 1.1176 | 38 | 1.6910 | 0.1884 | 1.6910 | 1.3004 |
| No log | 1.1765 | 40 | 1.3136 | 0.1579 | 1.3136 | 1.1461 |
| No log | 1.2353 | 42 | 1.0181 | 0.1786 | 1.0181 | 1.0090 |
| No log | 1.2941 | 44 | 0.9802 | 0.2295 | 0.9802 | 0.9900 |
| No log | 1.3529 | 46 | 0.9417 | 0.2887 | 0.9417 | 0.9704 |
| No log | 1.4118 | 48 | 0.8933 | 0.2487 | 0.8933 | 0.9451 |
| No log | 1.4706 | 50 | 0.8008 | 0.2508 | 0.8008 | 0.8949 |
| No log | 1.5294 | 52 | 0.7552 | 0.3298 | 0.7552 | 0.8690 |
| No log | 1.5882 | 54 | 0.9062 | 0.2864 | 0.9062 | 0.9520 |
| No log | 1.6471 | 56 | 1.0425 | 0.1967 | 1.0425 | 1.0210 |
| No log | 1.7059 | 58 | 1.0478 | 0.2196 | 1.0478 | 1.0236 |
| No log | 1.7647 | 60 | 0.9543 | 0.2626 | 0.9543 | 0.9769 |
| No log | 1.8235 | 62 | 0.8922 | 0.2739 | 0.8922 | 0.9446 |
| No log | 1.8824 | 64 | 0.8045 | 0.3342 | 0.8045 | 0.8969 |
| No log | 1.9412 | 66 | 0.7612 | 0.3828 | 0.7612 | 0.8724 |
| No log | 2.0 | 68 | 0.7807 | 0.3835 | 0.7807 | 0.8836 |
| No log | 2.0588 | 70 | 0.8746 | 0.3877 | 0.8746 | 0.9352 |
| No log | 2.1176 | 72 | 1.0062 | 0.3381 | 1.0062 | 1.0031 |
| No log | 2.1765 | 74 | 1.2001 | 0.3087 | 1.2001 | 1.0955 |
| No log | 2.2353 | 76 | 1.1859 | 0.3109 | 1.1859 | 1.0890 |
| No log | 2.2941 | 78 | 0.8804 | 0.4553 | 0.8804 | 0.9383 |
| No log | 2.3529 | 80 | 0.7851 | 0.4756 | 0.7851 | 0.8860 |
| No log | 2.4118 | 82 | 0.7514 | 0.4407 | 0.7514 | 0.8668 |
| No log | 2.4706 | 84 | 0.7220 | 0.4470 | 0.7220 | 0.8497 |
| No log | 2.5294 | 86 | 0.6965 | 0.4655 | 0.6965 | 0.8346 |
| No log | 2.5882 | 88 | 0.6990 | 0.4789 | 0.6990 | 0.8360 |
| No log | 2.6471 | 90 | 0.8786 | 0.4517 | 0.8786 | 0.9373 |
| No log | 2.7059 | 92 | 1.2702 | 0.2916 | 1.2702 | 1.1270 |
| No log | 2.7647 | 94 | 1.4689 | 0.2385 | 1.4689 | 1.2120 |
| No log | 2.8235 | 96 | 1.3729 | 0.2553 | 1.3729 | 1.1717 |
| No log | 2.8824 | 98 | 1.0170 | 0.3935 | 1.0170 | 1.0085 |
| No log | 2.9412 | 100 | 0.6981 | 0.5465 | 0.6981 | 0.8355 |
| No log | 3.0 | 102 | 0.7003 | 0.5664 | 0.7003 | 0.8369 |
| No log | 3.0588 | 104 | 0.7187 | 0.5061 | 0.7187 | 0.8478 |
| No log | 3.1176 | 106 | 0.7027 | 0.5024 | 0.7027 | 0.8383 |
| No log | 3.1765 | 108 | 0.6615 | 0.5748 | 0.6615 | 0.8133 |
| No log | 3.2353 | 110 | 0.6902 | 0.5839 | 0.6902 | 0.8308 |
| No log | 3.2941 | 112 | 0.7647 | 0.5466 | 0.7647 | 0.8745 |
| No log | 3.3529 | 114 | 0.7679 | 0.5368 | 0.7679 | 0.8763 |
| No log | 3.4118 | 116 | 0.7913 | 0.5440 | 0.7913 | 0.8895 |
| No log | 3.4706 | 118 | 0.8293 | 0.5099 | 0.8293 | 0.9107 |
| No log | 3.5294 | 120 | 0.8751 | 0.5291 | 0.8751 | 0.9355 |
| No log | 3.5882 | 122 | 0.9250 | 0.5036 | 0.9250 | 0.9618 |
| No log | 3.6471 | 124 | 0.9585 | 0.4668 | 0.9585 | 0.9790 |
| No log | 3.7059 | 126 | 1.0456 | 0.4595 | 1.0456 | 1.0225 |
| No log | 3.7647 | 128 | 1.0706 | 0.4356 | 1.0706 | 1.0347 |
| No log | 3.8235 | 130 | 1.0680 | 0.4522 | 1.0680 | 1.0335 |
| No log | 3.8824 | 132 | 1.0563 | 0.4662 | 1.0563 | 1.0278 |
| No log | 3.9412 | 134 | 1.0663 | 0.4737 | 1.0663 | 1.0326 |
| No log | 4.0 | 136 | 1.1037 | 0.4532 | 1.1037 | 1.0506 |
| No log | 4.0588 | 138 | 1.1187 | 0.4426 | 1.1187 | 1.0577 |
| No log | 4.1176 | 140 | 1.1449 | 0.4759 | 1.1449 | 1.0700 |
| No log | 4.1765 | 142 | 1.1760 | 0.4452 | 1.1760 | 1.0844 |
| No log | 4.2353 | 144 | 1.1950 | 0.4615 | 1.1950 | 1.0932 |
| No log | 4.2941 | 146 | 1.1842 | 0.4350 | 1.1842 | 1.0882 |
| No log | 4.3529 | 148 | 1.2260 | 0.4234 | 1.2260 | 1.1072 |
| No log | 4.4118 | 150 | 1.2154 | 0.3912 | 1.2154 | 1.1025 |
| No log | 4.4706 | 152 | 1.1700 | 0.3772 | 1.1700 | 1.0816 |
| No log | 4.5294 | 154 | 1.1192 | 0.4257 | 1.1192 | 1.0579 |
| No log | 4.5882 | 156 | 1.0561 | 0.4358 | 1.0561 | 1.0277 |
| No log | 4.6471 | 158 | 1.0440 | 0.4691 | 1.0440 | 1.0218 |
| No log | 4.7059 | 160 | 1.0321 | 0.4311 | 1.0321 | 1.0159 |
| No log | 4.7647 | 162 | 0.9661 | 0.4310 | 0.9661 | 0.9829 |
| No log | 4.8235 | 164 | 0.9112 | 0.5018 | 0.9112 | 0.9546 |
| No log | 4.8824 | 166 | 0.8906 | 0.4318 | 0.8906 | 0.9437 |
| No log | 4.9412 | 168 | 0.9176 | 0.3801 | 0.9176 | 0.9579 |
| No log | 5.0 | 170 | 0.9519 | 0.4179 | 0.9519 | 0.9756 |
| No log | 5.0588 | 172 | 0.9834 | 0.4340 | 0.9834 | 0.9917 |
| No log | 5.1176 | 174 | 1.0008 | 0.5042 | 1.0008 | 1.0004 |
| No log | 5.1765 | 176 | 1.0241 | 0.4682 | 1.0241 | 1.0120 |
| No log | 5.2353 | 178 | 1.0493 | 0.4440 | 1.0493 | 1.0244 |
| No log | 5.2941 | 180 | 1.0713 | 0.4496 | 1.0713 | 1.0351 |
| No log | 5.3529 | 182 | 1.0770 | 0.4695 | 1.0770 | 1.0378 |
| No log | 5.4118 | 184 | 1.1351 | 0.4451 | 1.1351 | 1.0654 |
| No log | 5.4706 | 186 | 1.2191 | 0.4129 | 1.2191 | 1.1041 |
| No log | 5.5294 | 188 | 1.2173 | 0.4129 | 1.2173 | 1.1033 |
| No log | 5.5882 | 190 | 1.1616 | 0.4623 | 1.1616 | 1.0778 |
| No log | 5.6471 | 192 | 1.0955 | 0.4381 | 1.0955 | 1.0466 |
| No log | 5.7059 | 194 | 1.0688 | 0.4671 | 1.0688 | 1.0338 |
| No log | 5.7647 | 196 | 1.0831 | 0.4452 | 1.0831 | 1.0407 |
| No log | 5.8235 | 198 | 1.1109 | 0.4654 | 1.1109 | 1.0540 |
| No log | 5.8824 | 200 | 1.1308 | 0.4701 | 1.1308 | 1.0634 |
| No log | 5.9412 | 202 | 1.1264 | 0.3850 | 1.1264 | 1.0613 |
| No log | 6.0 | 204 | 1.1111 | 0.3996 | 1.1111 | 1.0541 |
| No log | 6.0588 | 206 | 1.0877 | 0.4082 | 1.0877 | 1.0429 |
| No log | 6.1176 | 208 | 1.0474 | 0.4279 | 1.0474 | 1.0234 |
| No log | 6.1765 | 210 | 1.0091 | 0.4641 | 1.0091 | 1.0046 |
| No log | 6.2353 | 212 | 0.9861 | 0.4747 | 0.9861 | 0.9930 |
| No log | 6.2941 | 214 | 0.9879 | 0.4471 | 0.9879 | 0.9939 |
| No log | 6.3529 | 216 | 1.0024 | 0.4347 | 1.0024 | 1.0012 |
| No log | 6.4118 | 218 | 0.9924 | 0.4262 | 0.9924 | 0.9962 |
| No log | 6.4706 | 220 | 0.9852 | 0.4476 | 0.9852 | 0.9926 |
| No log | 6.5294 | 222 | 0.9965 | 0.4613 | 0.9965 | 0.9983 |
| No log | 6.5882 | 224 | 1.0144 | 0.4613 | 1.0144 | 1.0072 |
| No log | 6.6471 | 226 | 1.0307 | 0.4457 | 1.0307 | 1.0152 |
| No log | 6.7059 | 228 | 1.0625 | 0.4382 | 1.0625 | 1.0308 |
| No log | 6.7647 | 230 | 1.0921 | 0.4420 | 1.0921 | 1.0450 |
| No log | 6.8235 | 232 | 1.1039 | 0.4517 | 1.1039 | 1.0507 |
| No log | 6.8824 | 234 | 1.0963 | 0.4582 | 1.0963 | 1.0470 |
| No log | 6.9412 | 236 | 1.0767 | 0.4517 | 1.0767 | 1.0376 |
| No log | 7.0 | 238 | 1.0565 | 0.4595 | 1.0565 | 1.0279 |
| No log | 7.0588 | 240 | 1.0241 | 0.4474 | 1.0241 | 1.0120 |
| No log | 7.1176 | 242 | 0.9961 | 0.4433 | 0.9961 | 0.9981 |
| No log | 7.1765 | 244 | 0.9827 | 0.4777 | 0.9827 | 0.9913 |
| No log | 7.2353 | 246 | 0.9821 | 0.4770 | 0.9821 | 0.9910 |
| No log | 7.2941 | 248 | 0.9922 | 0.4681 | 0.9922 | 0.9961 |
| No log | 7.3529 | 250 | 0.9982 | 0.4804 | 0.9982 | 0.9991 |
| No log | 7.4118 | 252 | 1.0067 | 0.4759 | 1.0067 | 1.0033 |
| No log | 7.4706 | 254 | 1.0110 | 0.4726 | 1.0110 | 1.0055 |
| No log | 7.5294 | 256 | 1.0059 | 0.4607 | 1.0059 | 1.0029 |
| No log | 7.5882 | 258 | 0.9957 | 0.4854 | 0.9957 | 0.9979 |
| No log | 7.6471 | 260 | 0.9883 | 0.4880 | 0.9883 | 0.9941 |
| No log | 7.7059 | 262 | 0.9862 | 0.4980 | 0.9862 | 0.9931 |
| No log | 7.7647 | 264 | 0.9978 | 0.4702 | 0.9978 | 0.9989 |
| No log | 7.8235 | 266 | 1.0106 | 0.4616 | 1.0106 | 1.0053 |
| No log | 7.8824 | 268 | 1.0101 | 0.4745 | 1.0101 | 1.0050 |
| No log | 7.9412 | 270 | 1.0034 | 0.4628 | 1.0034 | 1.0017 |
| No log | 8.0 | 272 | 0.9934 | 0.5186 | 0.9934 | 0.9967 |
| No log | 8.0588 | 274 | 0.9773 | 0.4827 | 0.9773 | 0.9886 |
| No log | 8.1176 | 276 | 0.9647 | 0.4833 | 0.9647 | 0.9822 |
| No log | 8.1765 | 278 | 0.9683 | 0.4833 | 0.9683 | 0.9840 |
| No log | 8.2353 | 280 | 0.9718 | 0.4828 | 0.9718 | 0.9858 |
| No log | 8.2941 | 282 | 0.9815 | 0.4859 | 0.9815 | 0.9907 |
| No log | 8.3529 | 284 | 0.9952 | 0.4675 | 0.9952 | 0.9976 |
| No log | 8.4118 | 286 | 0.9986 | 0.4675 | 0.9986 | 0.9993 |
| No log | 8.4706 | 288 | 0.9968 | 0.4625 | 0.9968 | 0.9984 |
| No log | 8.5294 | 290 | 0.9935 | 0.4625 | 0.9935 | 0.9967 |
| No log | 8.5882 | 292 | 0.9856 | 0.4787 | 0.9856 | 0.9928 |
| No log | 8.6471 | 294 | 0.9756 | 0.4625 | 0.9756 | 0.9877 |
| No log | 8.7059 | 296 | 0.9656 | 0.4718 | 0.9656 | 0.9827 |
| No log | 8.7647 | 298 | 0.9577 | 0.4575 | 0.9577 | 0.9786 |
| No log | 8.8235 | 300 | 0.9488 | 0.4471 | 0.9488 | 0.9741 |
| No log | 8.8824 | 302 | 0.9476 | 0.4471 | 0.9476 | 0.9735 |
| No log | 8.9412 | 304 | 0.9479 | 0.4471 | 0.9479 | 0.9736 |
| No log | 9.0 | 306 | 0.9458 | 0.4471 | 0.9458 | 0.9725 |
| No log | 9.0588 | 308 | 0.9418 | 0.4585 | 0.9418 | 0.9705 |
| No log | 9.1176 | 310 | 0.9353 | 0.4585 | 0.9353 | 0.9671 |
| No log | 9.1765 | 312 | 0.9310 | 0.4475 | 0.9310 | 0.9649 |
| No log | 9.2353 | 314 | 0.9316 | 0.4475 | 0.9316 | 0.9652 |
| No log | 9.2941 | 316 | 0.9324 | 0.4475 | 0.9324 | 0.9656 |
| No log | 9.3529 | 318 | 0.9351 | 0.4346 | 0.9351 | 0.9670 |
| No log | 9.4118 | 320 | 0.9350 | 0.4329 | 0.9350 | 0.9670 |
| No log | 9.4706 | 322 | 0.9328 | 0.4329 | 0.9328 | 0.9658 |
| No log | 9.5294 | 324 | 0.9320 | 0.4329 | 0.9320 | 0.9654 |
| No log | 9.5882 | 326 | 0.9326 | 0.4543 | 0.9326 | 0.9657 |
| No log | 9.6471 | 328 | 0.9347 | 0.4543 | 0.9347 | 0.9668 |
| No log | 9.7059 | 330 | 0.9352 | 0.4543 | 0.9352 | 0.9670 |
| No log | 9.7647 | 332 | 0.9355 | 0.4543 | 0.9355 | 0.9672 |
| No log | 9.8235 | 334 | 0.9358 | 0.4216 | 0.9358 | 0.9674 |
| No log | 9.8824 | 336 | 0.9361 | 0.4216 | 0.9361 | 0.9675 |
| No log | 9.9412 | 338 | 0.9363 | 0.4216 | 0.9363 | 0.9676 |
| No log | 10.0 | 340 | 0.9364 | 0.4216 | 0.9364 | 0.9677 |
Framework versions
- Transformers 4.44.2
- Pytorch 2.4.0+cu118
- Datasets 2.21.0
- Tokenizers 0.19.1
- Downloads last month
- 1
Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k5_task2_organization
Base model
aubmindlab/bert-base-arabertv02