ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k5_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9364
  • Qwk: 0.4216
  • Mse: 0.9364
  • Rmse: 0.9677

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0588 2 4.0665 -0.0151 4.0665 2.0166
No log 0.1176 4 2.0305 0.0284 2.0305 1.4250
No log 0.1765 6 1.1765 0.0592 1.1765 1.0846
No log 0.2353 8 0.8194 0.0770 0.8194 0.9052
No log 0.2941 10 0.9559 0.1514 0.9559 0.9777
No log 0.3529 12 1.1527 0.1040 1.1527 1.0736
No log 0.4118 14 0.8714 0.1748 0.8714 0.9335
No log 0.4706 16 0.6749 0.2880 0.6749 0.8215
No log 0.5294 18 0.6440 0.3611 0.6440 0.8025
No log 0.5882 20 0.8069 0.1337 0.8069 0.8983
No log 0.6471 22 0.9382 0.1106 0.9382 0.9686
No log 0.7059 24 0.9534 0.1590 0.9534 0.9764
No log 0.7647 26 1.1871 0.1583 1.1871 1.0895
No log 0.8235 28 1.5034 0.1392 1.5034 1.2261
No log 0.8824 30 1.8216 0.1717 1.8216 1.3497
No log 0.9412 32 1.7859 0.1794 1.7859 1.3364
No log 1.0 34 1.4960 0.1639 1.4960 1.2231
No log 1.0588 36 1.7451 0.1670 1.7451 1.3210
No log 1.1176 38 1.6910 0.1884 1.6910 1.3004
No log 1.1765 40 1.3136 0.1579 1.3136 1.1461
No log 1.2353 42 1.0181 0.1786 1.0181 1.0090
No log 1.2941 44 0.9802 0.2295 0.9802 0.9900
No log 1.3529 46 0.9417 0.2887 0.9417 0.9704
No log 1.4118 48 0.8933 0.2487 0.8933 0.9451
No log 1.4706 50 0.8008 0.2508 0.8008 0.8949
No log 1.5294 52 0.7552 0.3298 0.7552 0.8690
No log 1.5882 54 0.9062 0.2864 0.9062 0.9520
No log 1.6471 56 1.0425 0.1967 1.0425 1.0210
No log 1.7059 58 1.0478 0.2196 1.0478 1.0236
No log 1.7647 60 0.9543 0.2626 0.9543 0.9769
No log 1.8235 62 0.8922 0.2739 0.8922 0.9446
No log 1.8824 64 0.8045 0.3342 0.8045 0.8969
No log 1.9412 66 0.7612 0.3828 0.7612 0.8724
No log 2.0 68 0.7807 0.3835 0.7807 0.8836
No log 2.0588 70 0.8746 0.3877 0.8746 0.9352
No log 2.1176 72 1.0062 0.3381 1.0062 1.0031
No log 2.1765 74 1.2001 0.3087 1.2001 1.0955
No log 2.2353 76 1.1859 0.3109 1.1859 1.0890
No log 2.2941 78 0.8804 0.4553 0.8804 0.9383
No log 2.3529 80 0.7851 0.4756 0.7851 0.8860
No log 2.4118 82 0.7514 0.4407 0.7514 0.8668
No log 2.4706 84 0.7220 0.4470 0.7220 0.8497
No log 2.5294 86 0.6965 0.4655 0.6965 0.8346
No log 2.5882 88 0.6990 0.4789 0.6990 0.8360
No log 2.6471 90 0.8786 0.4517 0.8786 0.9373
No log 2.7059 92 1.2702 0.2916 1.2702 1.1270
No log 2.7647 94 1.4689 0.2385 1.4689 1.2120
No log 2.8235 96 1.3729 0.2553 1.3729 1.1717
No log 2.8824 98 1.0170 0.3935 1.0170 1.0085
No log 2.9412 100 0.6981 0.5465 0.6981 0.8355
No log 3.0 102 0.7003 0.5664 0.7003 0.8369
No log 3.0588 104 0.7187 0.5061 0.7187 0.8478
No log 3.1176 106 0.7027 0.5024 0.7027 0.8383
No log 3.1765 108 0.6615 0.5748 0.6615 0.8133
No log 3.2353 110 0.6902 0.5839 0.6902 0.8308
No log 3.2941 112 0.7647 0.5466 0.7647 0.8745
No log 3.3529 114 0.7679 0.5368 0.7679 0.8763
No log 3.4118 116 0.7913 0.5440 0.7913 0.8895
No log 3.4706 118 0.8293 0.5099 0.8293 0.9107
No log 3.5294 120 0.8751 0.5291 0.8751 0.9355
No log 3.5882 122 0.9250 0.5036 0.9250 0.9618
No log 3.6471 124 0.9585 0.4668 0.9585 0.9790
No log 3.7059 126 1.0456 0.4595 1.0456 1.0225
No log 3.7647 128 1.0706 0.4356 1.0706 1.0347
No log 3.8235 130 1.0680 0.4522 1.0680 1.0335
No log 3.8824 132 1.0563 0.4662 1.0563 1.0278
No log 3.9412 134 1.0663 0.4737 1.0663 1.0326
No log 4.0 136 1.1037 0.4532 1.1037 1.0506
No log 4.0588 138 1.1187 0.4426 1.1187 1.0577
No log 4.1176 140 1.1449 0.4759 1.1449 1.0700
No log 4.1765 142 1.1760 0.4452 1.1760 1.0844
No log 4.2353 144 1.1950 0.4615 1.1950 1.0932
No log 4.2941 146 1.1842 0.4350 1.1842 1.0882
No log 4.3529 148 1.2260 0.4234 1.2260 1.1072
No log 4.4118 150 1.2154 0.3912 1.2154 1.1025
No log 4.4706 152 1.1700 0.3772 1.1700 1.0816
No log 4.5294 154 1.1192 0.4257 1.1192 1.0579
No log 4.5882 156 1.0561 0.4358 1.0561 1.0277
No log 4.6471 158 1.0440 0.4691 1.0440 1.0218
No log 4.7059 160 1.0321 0.4311 1.0321 1.0159
No log 4.7647 162 0.9661 0.4310 0.9661 0.9829
No log 4.8235 164 0.9112 0.5018 0.9112 0.9546
No log 4.8824 166 0.8906 0.4318 0.8906 0.9437
No log 4.9412 168 0.9176 0.3801 0.9176 0.9579
No log 5.0 170 0.9519 0.4179 0.9519 0.9756
No log 5.0588 172 0.9834 0.4340 0.9834 0.9917
No log 5.1176 174 1.0008 0.5042 1.0008 1.0004
No log 5.1765 176 1.0241 0.4682 1.0241 1.0120
No log 5.2353 178 1.0493 0.4440 1.0493 1.0244
No log 5.2941 180 1.0713 0.4496 1.0713 1.0351
No log 5.3529 182 1.0770 0.4695 1.0770 1.0378
No log 5.4118 184 1.1351 0.4451 1.1351 1.0654
No log 5.4706 186 1.2191 0.4129 1.2191 1.1041
No log 5.5294 188 1.2173 0.4129 1.2173 1.1033
No log 5.5882 190 1.1616 0.4623 1.1616 1.0778
No log 5.6471 192 1.0955 0.4381 1.0955 1.0466
No log 5.7059 194 1.0688 0.4671 1.0688 1.0338
No log 5.7647 196 1.0831 0.4452 1.0831 1.0407
No log 5.8235 198 1.1109 0.4654 1.1109 1.0540
No log 5.8824 200 1.1308 0.4701 1.1308 1.0634
No log 5.9412 202 1.1264 0.3850 1.1264 1.0613
No log 6.0 204 1.1111 0.3996 1.1111 1.0541
No log 6.0588 206 1.0877 0.4082 1.0877 1.0429
No log 6.1176 208 1.0474 0.4279 1.0474 1.0234
No log 6.1765 210 1.0091 0.4641 1.0091 1.0046
No log 6.2353 212 0.9861 0.4747 0.9861 0.9930
No log 6.2941 214 0.9879 0.4471 0.9879 0.9939
No log 6.3529 216 1.0024 0.4347 1.0024 1.0012
No log 6.4118 218 0.9924 0.4262 0.9924 0.9962
No log 6.4706 220 0.9852 0.4476 0.9852 0.9926
No log 6.5294 222 0.9965 0.4613 0.9965 0.9983
No log 6.5882 224 1.0144 0.4613 1.0144 1.0072
No log 6.6471 226 1.0307 0.4457 1.0307 1.0152
No log 6.7059 228 1.0625 0.4382 1.0625 1.0308
No log 6.7647 230 1.0921 0.4420 1.0921 1.0450
No log 6.8235 232 1.1039 0.4517 1.1039 1.0507
No log 6.8824 234 1.0963 0.4582 1.0963 1.0470
No log 6.9412 236 1.0767 0.4517 1.0767 1.0376
No log 7.0 238 1.0565 0.4595 1.0565 1.0279
No log 7.0588 240 1.0241 0.4474 1.0241 1.0120
No log 7.1176 242 0.9961 0.4433 0.9961 0.9981
No log 7.1765 244 0.9827 0.4777 0.9827 0.9913
No log 7.2353 246 0.9821 0.4770 0.9821 0.9910
No log 7.2941 248 0.9922 0.4681 0.9922 0.9961
No log 7.3529 250 0.9982 0.4804 0.9982 0.9991
No log 7.4118 252 1.0067 0.4759 1.0067 1.0033
No log 7.4706 254 1.0110 0.4726 1.0110 1.0055
No log 7.5294 256 1.0059 0.4607 1.0059 1.0029
No log 7.5882 258 0.9957 0.4854 0.9957 0.9979
No log 7.6471 260 0.9883 0.4880 0.9883 0.9941
No log 7.7059 262 0.9862 0.4980 0.9862 0.9931
No log 7.7647 264 0.9978 0.4702 0.9978 0.9989
No log 7.8235 266 1.0106 0.4616 1.0106 1.0053
No log 7.8824 268 1.0101 0.4745 1.0101 1.0050
No log 7.9412 270 1.0034 0.4628 1.0034 1.0017
No log 8.0 272 0.9934 0.5186 0.9934 0.9967
No log 8.0588 274 0.9773 0.4827 0.9773 0.9886
No log 8.1176 276 0.9647 0.4833 0.9647 0.9822
No log 8.1765 278 0.9683 0.4833 0.9683 0.9840
No log 8.2353 280 0.9718 0.4828 0.9718 0.9858
No log 8.2941 282 0.9815 0.4859 0.9815 0.9907
No log 8.3529 284 0.9952 0.4675 0.9952 0.9976
No log 8.4118 286 0.9986 0.4675 0.9986 0.9993
No log 8.4706 288 0.9968 0.4625 0.9968 0.9984
No log 8.5294 290 0.9935 0.4625 0.9935 0.9967
No log 8.5882 292 0.9856 0.4787 0.9856 0.9928
No log 8.6471 294 0.9756 0.4625 0.9756 0.9877
No log 8.7059 296 0.9656 0.4718 0.9656 0.9827
No log 8.7647 298 0.9577 0.4575 0.9577 0.9786
No log 8.8235 300 0.9488 0.4471 0.9488 0.9741
No log 8.8824 302 0.9476 0.4471 0.9476 0.9735
No log 8.9412 304 0.9479 0.4471 0.9479 0.9736
No log 9.0 306 0.9458 0.4471 0.9458 0.9725
No log 9.0588 308 0.9418 0.4585 0.9418 0.9705
No log 9.1176 310 0.9353 0.4585 0.9353 0.9671
No log 9.1765 312 0.9310 0.4475 0.9310 0.9649
No log 9.2353 314 0.9316 0.4475 0.9316 0.9652
No log 9.2941 316 0.9324 0.4475 0.9324 0.9656
No log 9.3529 318 0.9351 0.4346 0.9351 0.9670
No log 9.4118 320 0.9350 0.4329 0.9350 0.9670
No log 9.4706 322 0.9328 0.4329 0.9328 0.9658
No log 9.5294 324 0.9320 0.4329 0.9320 0.9654
No log 9.5882 326 0.9326 0.4543 0.9326 0.9657
No log 9.6471 328 0.9347 0.4543 0.9347 0.9668
No log 9.7059 330 0.9352 0.4543 0.9352 0.9670
No log 9.7647 332 0.9355 0.4543 0.9355 0.9672
No log 9.8235 334 0.9358 0.4216 0.9358 0.9674
No log 9.8824 336 0.9361 0.4216 0.9361 0.9675
No log 9.9412 338 0.9363 0.4216 0.9363 0.9676
No log 10.0 340 0.9364 0.4216 0.9364 0.9677

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k5_task2_organization

Finetuned
(4023)
this model