ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k8_task5_organization
This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:
- Loss: 1.0692
- Qwk: 0.6292
- Mse: 1.0692
- Rmse: 1.0340
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
Training results
| Training Loss | Epoch | Step | Validation Loss | Qwk | Mse | Rmse |
|---|---|---|---|---|---|---|
| No log | 0.0625 | 2 | 2.4368 | 0.0170 | 2.4368 | 1.5610 |
| No log | 0.125 | 4 | 1.6639 | 0.1265 | 1.6639 | 1.2899 |
| No log | 0.1875 | 6 | 1.7030 | -0.0044 | 1.7030 | 1.3050 |
| No log | 0.25 | 8 | 1.7409 | 0.1332 | 1.7409 | 1.3194 |
| No log | 0.3125 | 10 | 1.7665 | 0.2389 | 1.7665 | 1.3291 |
| No log | 0.375 | 12 | 1.6905 | 0.2851 | 1.6905 | 1.3002 |
| No log | 0.4375 | 14 | 1.6569 | 0.2924 | 1.6569 | 1.2872 |
| No log | 0.5 | 16 | 1.6187 | 0.3467 | 1.6187 | 1.2723 |
| No log | 0.5625 | 18 | 1.3933 | 0.2476 | 1.3933 | 1.1804 |
| No log | 0.625 | 20 | 1.2790 | 0.2217 | 1.2790 | 1.1309 |
| No log | 0.6875 | 22 | 1.2802 | 0.2846 | 1.2802 | 1.1315 |
| No log | 0.75 | 24 | 1.2515 | 0.2495 | 1.2515 | 1.1187 |
| No log | 0.8125 | 26 | 1.2441 | 0.2330 | 1.2441 | 1.1154 |
| No log | 0.875 | 28 | 1.2869 | 0.3264 | 1.2869 | 1.1344 |
| No log | 0.9375 | 30 | 1.2641 | 0.3676 | 1.2641 | 1.1243 |
| No log | 1.0 | 32 | 1.2120 | 0.3931 | 1.2120 | 1.1009 |
| No log | 1.0625 | 34 | 1.1730 | 0.3814 | 1.1730 | 1.0831 |
| No log | 1.125 | 36 | 1.1501 | 0.4272 | 1.1501 | 1.0724 |
| No log | 1.1875 | 38 | 1.1131 | 0.4075 | 1.1131 | 1.0551 |
| No log | 1.25 | 40 | 1.1009 | 0.4293 | 1.1009 | 1.0492 |
| No log | 1.3125 | 42 | 1.0895 | 0.4331 | 1.0895 | 1.0438 |
| No log | 1.375 | 44 | 1.1371 | 0.4953 | 1.1371 | 1.0664 |
| No log | 1.4375 | 46 | 1.1822 | 0.4904 | 1.1822 | 1.0873 |
| No log | 1.5 | 48 | 1.2365 | 0.4521 | 1.2365 | 1.1120 |
| No log | 1.5625 | 50 | 1.2205 | 0.4838 | 1.2205 | 1.1047 |
| No log | 1.625 | 52 | 1.2152 | 0.4680 | 1.2152 | 1.1023 |
| No log | 1.6875 | 54 | 1.0994 | 0.4738 | 1.0994 | 1.0485 |
| No log | 1.75 | 56 | 1.0871 | 0.5106 | 1.0871 | 1.0426 |
| No log | 1.8125 | 58 | 1.0915 | 0.4787 | 1.0915 | 1.0448 |
| No log | 1.875 | 60 | 1.0637 | 0.4684 | 1.0637 | 1.0313 |
| No log | 1.9375 | 62 | 1.0795 | 0.4794 | 1.0795 | 1.0390 |
| No log | 2.0 | 64 | 1.2050 | 0.4482 | 1.2050 | 1.0977 |
| No log | 2.0625 | 66 | 1.2526 | 0.4649 | 1.2526 | 1.1192 |
| No log | 2.125 | 68 | 1.1777 | 0.4840 | 1.1777 | 1.0852 |
| No log | 2.1875 | 70 | 1.0680 | 0.4846 | 1.0680 | 1.0334 |
| No log | 2.25 | 72 | 0.9713 | 0.5468 | 0.9713 | 0.9855 |
| No log | 2.3125 | 74 | 0.9510 | 0.5981 | 0.9510 | 0.9752 |
| No log | 2.375 | 76 | 1.0052 | 0.5533 | 1.0052 | 1.0026 |
| No log | 2.4375 | 78 | 1.3063 | 0.5200 | 1.3063 | 1.1429 |
| No log | 2.5 | 80 | 1.3657 | 0.4702 | 1.3657 | 1.1686 |
| No log | 2.5625 | 82 | 1.2633 | 0.4605 | 1.2633 | 1.1240 |
| No log | 2.625 | 84 | 1.1759 | 0.5081 | 1.1759 | 1.0844 |
| No log | 2.6875 | 86 | 1.1162 | 0.5339 | 1.1162 | 1.0565 |
| No log | 2.75 | 88 | 1.0818 | 0.5579 | 1.0818 | 1.0401 |
| No log | 2.8125 | 90 | 0.9999 | 0.5805 | 0.9999 | 0.9999 |
| No log | 2.875 | 92 | 0.9831 | 0.5716 | 0.9831 | 0.9915 |
| No log | 2.9375 | 94 | 1.0717 | 0.5754 | 1.0717 | 1.0353 |
| No log | 3.0 | 96 | 1.1018 | 0.5925 | 1.1018 | 1.0497 |
| No log | 3.0625 | 98 | 0.9779 | 0.5982 | 0.9779 | 0.9889 |
| No log | 3.125 | 100 | 0.9138 | 0.6255 | 0.9138 | 0.9559 |
| No log | 3.1875 | 102 | 0.8914 | 0.6129 | 0.8914 | 0.9442 |
| No log | 3.25 | 104 | 0.8946 | 0.6140 | 0.8946 | 0.9458 |
| No log | 3.3125 | 106 | 0.8964 | 0.6076 | 0.8964 | 0.9468 |
| No log | 3.375 | 108 | 0.9153 | 0.6187 | 0.9153 | 0.9567 |
| No log | 3.4375 | 110 | 1.0031 | 0.6086 | 1.0031 | 1.0015 |
| No log | 3.5 | 112 | 1.2576 | 0.5545 | 1.2576 | 1.1214 |
| No log | 3.5625 | 114 | 1.5144 | 0.5217 | 1.5144 | 1.2306 |
| No log | 3.625 | 116 | 1.5392 | 0.5223 | 1.5392 | 1.2407 |
| No log | 3.6875 | 118 | 1.3254 | 0.5370 | 1.3254 | 1.1513 |
| No log | 3.75 | 120 | 1.1411 | 0.5912 | 1.1411 | 1.0682 |
| No log | 3.8125 | 122 | 1.0610 | 0.6112 | 1.0610 | 1.0301 |
| No log | 3.875 | 124 | 1.0553 | 0.5956 | 1.0553 | 1.0273 |
| No log | 3.9375 | 126 | 1.0057 | 0.5742 | 1.0057 | 1.0029 |
| No log | 4.0 | 128 | 0.9320 | 0.5591 | 0.9320 | 0.9654 |
| No log | 4.0625 | 130 | 1.0088 | 0.5496 | 1.0088 | 1.0044 |
| No log | 4.125 | 132 | 1.1129 | 0.5455 | 1.1129 | 1.0549 |
| No log | 4.1875 | 134 | 1.1682 | 0.5428 | 1.1682 | 1.0809 |
| No log | 4.25 | 136 | 1.1659 | 0.5472 | 1.1659 | 1.0798 |
| No log | 4.3125 | 138 | 1.2829 | 0.5495 | 1.2829 | 1.1326 |
| No log | 4.375 | 140 | 1.3074 | 0.5542 | 1.3074 | 1.1434 |
| No log | 4.4375 | 142 | 1.3250 | 0.5512 | 1.3250 | 1.1511 |
| No log | 4.5 | 144 | 1.3414 | 0.5408 | 1.3414 | 1.1582 |
| No log | 4.5625 | 146 | 1.2082 | 0.5589 | 1.2082 | 1.0992 |
| No log | 4.625 | 148 | 1.0410 | 0.5887 | 1.0410 | 1.0203 |
| No log | 4.6875 | 150 | 0.9627 | 0.6414 | 0.9627 | 0.9812 |
| No log | 4.75 | 152 | 1.0040 | 0.6358 | 1.0040 | 1.0020 |
| No log | 4.8125 | 154 | 1.0924 | 0.6246 | 1.0924 | 1.0452 |
| No log | 4.875 | 156 | 1.0736 | 0.6374 | 1.0736 | 1.0361 |
| No log | 4.9375 | 158 | 1.0044 | 0.6722 | 1.0044 | 1.0022 |
| No log | 5.0 | 160 | 1.0023 | 0.6722 | 1.0023 | 1.0012 |
| No log | 5.0625 | 162 | 1.1328 | 0.6305 | 1.1328 | 1.0643 |
| No log | 5.125 | 164 | 1.2473 | 0.5861 | 1.2473 | 1.1168 |
| No log | 5.1875 | 166 | 1.2044 | 0.5965 | 1.2044 | 1.0975 |
| No log | 5.25 | 168 | 1.0631 | 0.6342 | 1.0631 | 1.0311 |
| No log | 5.3125 | 170 | 0.9741 | 0.6602 | 0.9741 | 0.9869 |
| No log | 5.375 | 172 | 0.9521 | 0.6541 | 0.9521 | 0.9758 |
| No log | 5.4375 | 174 | 0.9749 | 0.6572 | 0.9749 | 0.9874 |
| No log | 5.5 | 176 | 1.0751 | 0.6202 | 1.0751 | 1.0369 |
| No log | 5.5625 | 178 | 1.1060 | 0.6133 | 1.1060 | 1.0516 |
| No log | 5.625 | 180 | 1.1643 | 0.5942 | 1.1643 | 1.0790 |
| No log | 5.6875 | 182 | 1.2327 | 0.5766 | 1.2327 | 1.1103 |
| No log | 5.75 | 184 | 1.1998 | 0.5654 | 1.1998 | 1.0953 |
| No log | 5.8125 | 186 | 1.1536 | 0.5805 | 1.1536 | 1.0741 |
| No log | 5.875 | 188 | 1.0810 | 0.6310 | 1.0810 | 1.0397 |
| No log | 5.9375 | 190 | 0.9935 | 0.6314 | 0.9935 | 0.9967 |
| No log | 6.0 | 192 | 0.9686 | 0.6638 | 0.9686 | 0.9842 |
| No log | 6.0625 | 194 | 1.0367 | 0.6181 | 1.0367 | 1.0182 |
| No log | 6.125 | 196 | 1.2126 | 0.6085 | 1.2126 | 1.1012 |
| No log | 6.1875 | 198 | 1.4077 | 0.5672 | 1.4077 | 1.1865 |
| No log | 6.25 | 200 | 1.4245 | 0.5601 | 1.4245 | 1.1935 |
| No log | 6.3125 | 202 | 1.2742 | 0.6063 | 1.2742 | 1.1288 |
| No log | 6.375 | 204 | 1.0690 | 0.6139 | 1.0690 | 1.0339 |
| No log | 6.4375 | 206 | 0.9859 | 0.6465 | 0.9859 | 0.9929 |
| No log | 6.5 | 208 | 0.9466 | 0.6559 | 0.9466 | 0.9729 |
| No log | 6.5625 | 210 | 0.9683 | 0.6501 | 0.9683 | 0.9840 |
| No log | 6.625 | 212 | 1.0415 | 0.6371 | 1.0415 | 1.0206 |
| No log | 6.6875 | 214 | 1.1107 | 0.6124 | 1.1107 | 1.0539 |
| No log | 6.75 | 216 | 1.0933 | 0.6279 | 1.0933 | 1.0456 |
| No log | 6.8125 | 218 | 0.9874 | 0.5999 | 0.9874 | 0.9937 |
| No log | 6.875 | 220 | 0.8998 | 0.6058 | 0.8998 | 0.9486 |
| No log | 6.9375 | 222 | 0.8714 | 0.6209 | 0.8714 | 0.9335 |
| No log | 7.0 | 224 | 0.8803 | 0.6067 | 0.8803 | 0.9382 |
| No log | 7.0625 | 226 | 0.9361 | 0.6066 | 0.9361 | 0.9675 |
| No log | 7.125 | 228 | 1.0598 | 0.5903 | 1.0598 | 1.0294 |
| No log | 7.1875 | 230 | 1.1462 | 0.5931 | 1.1462 | 1.0706 |
| No log | 7.25 | 232 | 1.1211 | 0.5681 | 1.1211 | 1.0588 |
| No log | 7.3125 | 234 | 1.0315 | 0.6097 | 1.0315 | 1.0157 |
| No log | 7.375 | 236 | 1.0021 | 0.6241 | 1.0021 | 1.0010 |
| No log | 7.4375 | 238 | 0.9794 | 0.6241 | 0.9794 | 0.9897 |
| No log | 7.5 | 240 | 1.0154 | 0.6152 | 1.0154 | 1.0077 |
| No log | 7.5625 | 242 | 1.0255 | 0.6106 | 1.0255 | 1.0127 |
| No log | 7.625 | 244 | 0.9824 | 0.6306 | 0.9824 | 0.9912 |
| No log | 7.6875 | 246 | 0.9709 | 0.6273 | 0.9709 | 0.9853 |
| No log | 7.75 | 248 | 0.9572 | 0.6254 | 0.9572 | 0.9784 |
| No log | 7.8125 | 250 | 0.9881 | 0.6273 | 0.9881 | 0.9940 |
| No log | 7.875 | 252 | 0.9812 | 0.6273 | 0.9812 | 0.9905 |
| No log | 7.9375 | 254 | 0.9454 | 0.6254 | 0.9454 | 0.9723 |
| No log | 8.0 | 256 | 0.9107 | 0.6296 | 0.9107 | 0.9543 |
| No log | 8.0625 | 258 | 0.9022 | 0.6234 | 0.9022 | 0.9498 |
| No log | 8.125 | 260 | 0.9239 | 0.6193 | 0.9239 | 0.9612 |
| No log | 8.1875 | 262 | 0.9914 | 0.6359 | 0.9914 | 0.9957 |
| No log | 8.25 | 264 | 1.0406 | 0.6391 | 1.0406 | 1.0201 |
| No log | 8.3125 | 266 | 1.0300 | 0.6391 | 1.0300 | 1.0149 |
| No log | 8.375 | 268 | 0.9831 | 0.6434 | 0.9831 | 0.9915 |
| No log | 8.4375 | 270 | 0.9185 | 0.6358 | 0.9185 | 0.9584 |
| No log | 8.5 | 272 | 0.8801 | 0.6402 | 0.8801 | 0.9381 |
| No log | 8.5625 | 274 | 0.8735 | 0.6445 | 0.8735 | 0.9346 |
| No log | 8.625 | 276 | 0.8839 | 0.6464 | 0.8839 | 0.9402 |
| No log | 8.6875 | 278 | 0.9150 | 0.6358 | 0.9150 | 0.9566 |
| No log | 8.75 | 280 | 0.9303 | 0.6391 | 0.9303 | 0.9645 |
| No log | 8.8125 | 282 | 0.9466 | 0.6348 | 0.9466 | 0.9729 |
| No log | 8.875 | 284 | 0.9786 | 0.6434 | 0.9786 | 0.9893 |
| No log | 8.9375 | 286 | 0.9855 | 0.6434 | 0.9855 | 0.9927 |
| No log | 9.0 | 288 | 0.9987 | 0.6347 | 0.9987 | 0.9994 |
| No log | 9.0625 | 290 | 1.0152 | 0.6347 | 1.0152 | 1.0076 |
| No log | 9.125 | 292 | 1.0183 | 0.6347 | 1.0183 | 1.0091 |
| No log | 9.1875 | 294 | 1.0112 | 0.6347 | 1.0112 | 1.0056 |
| No log | 9.25 | 296 | 1.0182 | 0.6347 | 1.0182 | 1.0090 |
| No log | 9.3125 | 298 | 1.0264 | 0.6431 | 1.0264 | 1.0131 |
| No log | 9.375 | 300 | 1.0340 | 0.6431 | 1.0340 | 1.0169 |
| No log | 9.4375 | 302 | 1.0450 | 0.6431 | 1.0450 | 1.0223 |
| No log | 9.5 | 304 | 1.0507 | 0.6274 | 1.0507 | 1.0250 |
| No log | 9.5625 | 306 | 1.0519 | 0.6274 | 1.0519 | 1.0256 |
| No log | 9.625 | 308 | 1.0602 | 0.6274 | 1.0602 | 1.0297 |
| No log | 9.6875 | 310 | 1.0728 | 0.6292 | 1.0728 | 1.0358 |
| No log | 9.75 | 312 | 1.0769 | 0.6292 | 1.0769 | 1.0377 |
| No log | 9.8125 | 314 | 1.0772 | 0.6292 | 1.0772 | 1.0379 |
| No log | 9.875 | 316 | 1.0745 | 0.6292 | 1.0745 | 1.0366 |
| No log | 9.9375 | 318 | 1.0712 | 0.6292 | 1.0712 | 1.0350 |
| No log | 10.0 | 320 | 1.0692 | 0.6292 | 1.0692 | 1.0340 |
Framework versions
- Transformers 4.44.2
- Pytorch 2.4.0+cu118
- Datasets 2.21.0
- Tokenizers 0.19.1
- Downloads last month
- -
Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k8_task5_organization
Base model
aubmindlab/bert-base-arabertv02