ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k6_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1952
  • Qwk: 0.6156
  • Mse: 1.1952
  • Rmse: 1.0932

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 2.4318 0.0010 2.4318 1.5594
No log 0.16 4 1.6702 0.1238 1.6702 1.2924
No log 0.24 6 1.3137 0.1528 1.3137 1.1462
No log 0.32 8 1.5339 0.1550 1.5339 1.2385
No log 0.4 10 1.6410 0.0393 1.6410 1.2810
No log 0.48 12 1.5840 0.1940 1.5840 1.2586
No log 0.56 14 1.5331 0.2132 1.5331 1.2382
No log 0.64 16 1.6071 0.2310 1.6071 1.2677
No log 0.72 18 1.7480 0.2373 1.7480 1.3221
No log 0.8 20 1.6493 0.3072 1.6493 1.2842
No log 0.88 22 1.6963 0.3175 1.6963 1.3024
No log 0.96 24 1.6261 0.3295 1.6261 1.2752
No log 1.04 26 1.4373 0.3743 1.4373 1.1989
No log 1.12 28 1.3960 0.4291 1.3960 1.1815
No log 1.2 30 1.2818 0.4622 1.2818 1.1322
No log 1.28 32 1.3464 0.4400 1.3464 1.1603
No log 1.3600 34 1.5466 0.4565 1.5466 1.2436
No log 1.44 36 1.5268 0.4565 1.5268 1.2356
No log 1.52 38 1.3652 0.4705 1.3652 1.1684
No log 1.6 40 1.1954 0.4580 1.1954 1.0933
No log 1.6800 42 1.0662 0.4454 1.0662 1.0326
No log 1.76 44 1.0232 0.4710 1.0232 1.0115
No log 1.8400 46 1.0087 0.4548 1.0087 1.0043
No log 1.92 48 1.0057 0.4569 1.0057 1.0029
No log 2.0 50 1.1358 0.4712 1.1358 1.0658
No log 2.08 52 1.2720 0.4675 1.2720 1.1279
No log 2.16 54 1.2885 0.4977 1.2885 1.1351
No log 2.24 56 1.3784 0.4817 1.3784 1.1740
No log 2.32 58 1.3184 0.5067 1.3184 1.1482
No log 2.4 60 1.1251 0.4693 1.1251 1.0607
No log 2.48 62 1.1514 0.4634 1.1514 1.0730
No log 2.56 64 1.4748 0.4973 1.4748 1.2144
No log 2.64 66 1.5275 0.4963 1.5275 1.2359
No log 2.7200 68 1.5623 0.5094 1.5623 1.2499
No log 2.8 70 1.2982 0.5354 1.2982 1.1394
No log 2.88 72 1.0562 0.5025 1.0562 1.0277
No log 2.96 74 1.0919 0.5234 1.0919 1.0449
No log 3.04 76 1.5255 0.5239 1.5255 1.2351
No log 3.12 78 1.7981 0.4848 1.7981 1.3409
No log 3.2 80 1.6918 0.5031 1.6918 1.3007
No log 3.2800 82 1.1956 0.6008 1.1956 1.0934
No log 3.36 84 0.9030 0.6235 0.9030 0.9502
No log 3.44 86 0.8909 0.6350 0.8909 0.9438
No log 3.52 88 1.0038 0.5739 1.0038 1.0019
No log 3.6 90 1.4083 0.5280 1.4083 1.1867
No log 3.68 92 1.6602 0.4998 1.6602 1.2885
No log 3.76 94 1.5378 0.5199 1.5378 1.2401
No log 3.84 96 1.3449 0.5453 1.3449 1.1597
No log 3.92 98 1.3108 0.5564 1.3108 1.1449
No log 4.0 100 1.3871 0.5638 1.3871 1.1777
No log 4.08 102 1.3878 0.5638 1.3878 1.1780
No log 4.16 104 1.3663 0.5880 1.3663 1.1689
No log 4.24 106 1.3619 0.5874 1.3619 1.1670
No log 4.32 108 1.2841 0.6096 1.2841 1.1332
No log 4.4 110 1.2495 0.6096 1.2495 1.1178
No log 4.48 112 1.4535 0.5840 1.4535 1.2056
No log 4.5600 114 1.6155 0.5812 1.6155 1.2710
No log 4.64 116 1.5852 0.5812 1.5852 1.2590
No log 4.72 118 1.4995 0.5674 1.4995 1.2245
No log 4.8 120 1.2242 0.6151 1.2242 1.1064
No log 4.88 122 1.0272 0.6397 1.0272 1.0135
No log 4.96 124 0.9846 0.6293 0.9846 0.9923
No log 5.04 126 0.9252 0.6163 0.9252 0.9619
No log 5.12 128 0.9781 0.6105 0.9781 0.9890
No log 5.2 130 1.0533 0.6064 1.0533 1.0263
No log 5.28 132 1.1317 0.5617 1.1317 1.0638
No log 5.36 134 1.2297 0.6022 1.2297 1.1089
No log 5.44 136 1.3741 0.5569 1.3741 1.1722
No log 5.52 138 1.3062 0.5726 1.3062 1.1429
No log 5.6 140 1.1382 0.5471 1.1382 1.0668
No log 5.68 142 1.0213 0.5751 1.0213 1.0106
No log 5.76 144 1.0852 0.5864 1.0852 1.0417
No log 5.84 146 1.3130 0.5879 1.3130 1.1459
No log 5.92 148 1.6371 0.5109 1.6371 1.2795
No log 6.0 150 1.6530 0.5109 1.6530 1.2857
No log 6.08 152 1.4481 0.5473 1.4481 1.2034
No log 6.16 154 1.1785 0.5990 1.1785 1.0856
No log 6.24 156 1.0236 0.5987 1.0236 1.0117
No log 6.32 158 1.0104 0.5987 1.0104 1.0052
No log 6.4 160 1.0748 0.6004 1.0748 1.0367
No log 6.48 162 1.2348 0.5797 1.2348 1.1112
No log 6.5600 164 1.4122 0.5547 1.4122 1.1884
No log 6.64 166 1.4189 0.5547 1.4189 1.1912
No log 6.72 168 1.4205 0.5609 1.4205 1.1919
No log 6.8 170 1.3790 0.5687 1.3790 1.1743
No log 6.88 172 1.3199 0.5786 1.3199 1.1489
No log 6.96 174 1.2592 0.5693 1.2592 1.1221
No log 7.04 176 1.3199 0.5786 1.3199 1.1489
No log 7.12 178 1.5071 0.5383 1.5071 1.2276
No log 7.2 180 1.7504 0.4919 1.7504 1.3230
No log 7.28 182 1.8340 0.4869 1.8340 1.3543
No log 7.36 184 1.7631 0.4989 1.7631 1.3278
No log 7.44 186 1.5568 0.5304 1.5568 1.2477
No log 7.52 188 1.2829 0.5837 1.2829 1.1326
No log 7.6 190 1.1311 0.6042 1.1311 1.0635
No log 7.68 192 1.0340 0.6086 1.0340 1.0169
No log 7.76 194 1.0348 0.6103 1.0348 1.0173
No log 7.84 196 1.1023 0.6022 1.1023 1.0499
No log 7.92 198 1.1858 0.6035 1.1858 1.0889
No log 8.0 200 1.3107 0.5765 1.3107 1.1449
No log 8.08 202 1.4087 0.5333 1.4087 1.1869
No log 8.16 204 1.3942 0.5280 1.3942 1.1808
No log 8.24 206 1.2923 0.5664 1.2923 1.1368
No log 8.32 208 1.1652 0.5956 1.1652 1.0794
No log 8.4 210 1.0756 0.6039 1.0756 1.0371
No log 8.48 212 1.0379 0.6129 1.0379 1.0188
No log 8.56 214 1.0403 0.6129 1.0403 1.0200
No log 8.64 216 1.0810 0.5983 1.0810 1.0397
No log 8.72 218 1.1612 0.6072 1.1612 1.0776
No log 8.8 220 1.2419 0.5745 1.2419 1.1144
No log 8.88 222 1.2898 0.5615 1.2898 1.1357
No log 8.96 224 1.3030 0.5615 1.3030 1.1415
No log 9.04 226 1.3107 0.5615 1.3107 1.1449
No log 9.12 228 1.3335 0.5615 1.3335 1.1548
No log 9.2 230 1.3326 0.5615 1.3326 1.1544
No log 9.28 232 1.3049 0.5725 1.3049 1.1423
No log 9.36 234 1.2707 0.5834 1.2707 1.1272
No log 9.44 236 1.2200 0.6108 1.2200 1.1045
No log 9.52 238 1.1930 0.6089 1.1930 1.0922
No log 9.6 240 1.1834 0.6089 1.1834 1.0878
No log 9.68 242 1.1736 0.6141 1.1736 1.0833
No log 9.76 244 1.1756 0.6141 1.1756 1.0843
No log 9.84 246 1.1855 0.6156 1.1855 1.0888
No log 9.92 248 1.1915 0.6156 1.1915 1.0916
No log 10.0 250 1.1952 0.6156 1.1952 1.0932

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k6_task5_organization

Finetuned
(4023)
this model