ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k5_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0013
  • Qwk: 0.4730
  • Mse: 1.0013
  • Rmse: 1.0006

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 3.7363 0.0038 3.7363 1.9329
No log 0.1429 4 1.8671 0.0287 1.8671 1.3664
No log 0.2143 6 1.0654 0.0602 1.0654 1.0322
No log 0.2857 8 0.7212 0.1989 0.7212 0.8493
No log 0.3571 10 0.8901 0.1385 0.8901 0.9435
No log 0.4286 12 0.9571 0.0352 0.9571 0.9783
No log 0.5 14 0.8028 0.1762 0.8028 0.8960
No log 0.5714 16 0.9145 0.1917 0.9145 0.9563
No log 0.6429 18 1.0497 0.2279 1.0497 1.0245
No log 0.7143 20 0.8953 0.0909 0.8953 0.9462
No log 0.7857 22 0.8852 0.0387 0.8852 0.9408
No log 0.8571 24 0.7669 0.0535 0.7669 0.8757
No log 0.9286 26 0.7063 0.1208 0.7063 0.8404
No log 1.0 28 0.6904 0.1063 0.6904 0.8309
No log 1.0714 30 0.7313 0.1063 0.7313 0.8551
No log 1.1429 32 0.7734 0.1640 0.7734 0.8794
No log 1.2143 34 0.8398 0.2212 0.8398 0.9164
No log 1.2857 36 0.6964 0.1834 0.6964 0.8345
No log 1.3571 38 0.6776 0.2530 0.6776 0.8231
No log 1.4286 40 0.7171 0.3019 0.7171 0.8468
No log 1.5 42 0.6605 0.1534 0.6605 0.8127
No log 1.5714 44 0.6969 0.1599 0.6969 0.8348
No log 1.6429 46 0.7212 0.2087 0.7212 0.8493
No log 1.7143 48 0.7904 0.2360 0.7904 0.8890
No log 1.7857 50 0.9215 0.3138 0.9215 0.9599
No log 1.8571 52 0.8383 0.3049 0.8383 0.9156
No log 1.9286 54 0.6258 0.3552 0.6258 0.7910
No log 2.0 56 0.5766 0.4255 0.5766 0.7593
No log 2.0714 58 0.5790 0.4338 0.5790 0.7610
No log 2.1429 60 0.5700 0.4204 0.5700 0.7550
No log 2.2143 62 0.5770 0.4359 0.5770 0.7596
No log 2.2857 64 0.6295 0.4795 0.6295 0.7934
No log 2.3571 66 0.6886 0.4763 0.6886 0.8298
No log 2.4286 68 0.7792 0.4437 0.7792 0.8827
No log 2.5 70 0.7256 0.4533 0.7256 0.8518
No log 2.5714 72 0.7524 0.4454 0.7524 0.8674
No log 2.6429 74 0.8502 0.3647 0.8502 0.9221
No log 2.7143 76 0.9450 0.3787 0.9450 0.9721
No log 2.7857 78 0.9760 0.3792 0.9760 0.9879
No log 2.8571 80 0.9642 0.3849 0.9642 0.9820
No log 2.9286 82 0.9658 0.3921 0.9658 0.9827
No log 3.0 84 0.8425 0.4262 0.8425 0.9179
No log 3.0714 86 0.7273 0.5459 0.7273 0.8528
No log 3.1429 88 0.6948 0.5418 0.6948 0.8336
No log 3.2143 90 0.7605 0.5407 0.7605 0.8720
No log 3.2857 92 0.9194 0.4857 0.9194 0.9589
No log 3.3571 94 1.1331 0.4159 1.1331 1.0645
No log 3.4286 96 1.1660 0.4062 1.1660 1.0798
No log 3.5 98 1.0913 0.4233 1.0913 1.0446
No log 3.5714 100 0.9862 0.5208 0.9862 0.9931
No log 3.6429 102 0.9517 0.5026 0.9517 0.9755
No log 3.7143 104 1.0828 0.5235 1.0828 1.0406
No log 3.7857 106 1.3057 0.4177 1.3057 1.1427
No log 3.8571 108 1.1554 0.5104 1.1554 1.0749
No log 3.9286 110 0.8942 0.5148 0.8942 0.9456
No log 4.0 112 0.8265 0.4827 0.8265 0.9091
No log 4.0714 114 0.8953 0.4916 0.8953 0.9462
No log 4.1429 116 1.1000 0.5177 1.1000 1.0488
No log 4.2143 118 1.0718 0.4958 1.0718 1.0353
No log 4.2857 120 0.9629 0.5137 0.9629 0.9813
No log 4.3571 122 0.9037 0.5204 0.9037 0.9506
No log 4.4286 124 0.8414 0.5002 0.8414 0.9173
No log 4.5 126 0.8672 0.4889 0.8672 0.9312
No log 4.5714 128 0.8825 0.5159 0.8825 0.9394
No log 4.6429 130 0.8990 0.5025 0.8990 0.9482
No log 4.7143 132 0.8774 0.4881 0.8774 0.9367
No log 4.7857 134 0.9513 0.4856 0.9513 0.9753
No log 4.8571 136 1.2584 0.4699 1.2584 1.1218
No log 4.9286 138 1.4071 0.4064 1.4071 1.1862
No log 5.0 140 1.2609 0.4712 1.2609 1.1229
No log 5.0714 142 1.0649 0.4714 1.0649 1.0320
No log 5.1429 144 0.9511 0.4724 0.9511 0.9753
No log 5.2143 146 0.9368 0.4752 0.9368 0.9679
No log 5.2857 148 1.0156 0.4534 1.0156 1.0078
No log 5.3571 150 1.3247 0.3713 1.3247 1.1510
No log 5.4286 152 1.6439 0.3275 1.6439 1.2821
No log 5.5 154 1.6433 0.3261 1.6433 1.2819
No log 5.5714 156 1.4442 0.3507 1.4442 1.2017
No log 5.6429 158 1.1259 0.4173 1.1259 1.0611
No log 5.7143 160 0.9697 0.4751 0.9697 0.9847
No log 5.7857 162 0.9320 0.4895 0.9320 0.9654
No log 5.8571 164 0.9893 0.4683 0.9893 0.9946
No log 5.9286 166 1.1586 0.4565 1.1586 1.0764
No log 6.0 168 1.2124 0.4317 1.2124 1.1011
No log 6.0714 170 1.1725 0.4600 1.1725 1.0828
No log 6.1429 172 1.1008 0.4698 1.1008 1.0492
No log 6.2143 174 1.0172 0.4780 1.0172 1.0086
No log 6.2857 176 0.9142 0.5161 0.9142 0.9562
No log 6.3571 178 0.8674 0.5145 0.8674 0.9313
No log 6.4286 180 0.8702 0.5145 0.8702 0.9328
No log 6.5 182 0.8619 0.5220 0.8619 0.9284
No log 6.5714 184 0.9053 0.5379 0.9053 0.9515
No log 6.6429 186 1.0045 0.4871 1.0045 1.0022
No log 6.7143 188 1.1299 0.4717 1.1299 1.0630
No log 6.7857 190 1.1717 0.4709 1.1717 1.0825
No log 6.8571 192 1.0848 0.4790 1.0848 1.0415
No log 6.9286 194 0.9763 0.5374 0.9763 0.9881
No log 7.0 196 0.9658 0.5384 0.9658 0.9828
No log 7.0714 198 1.0374 0.4688 1.0374 1.0185
No log 7.1429 200 1.0540 0.4721 1.0540 1.0266
No log 7.2143 202 1.0425 0.4728 1.0425 1.0210
No log 7.2857 204 0.9725 0.4903 0.9725 0.9862
No log 7.3571 206 0.8801 0.5044 0.8801 0.9381
No log 7.4286 208 0.7969 0.5688 0.7969 0.8927
No log 7.5 210 0.7761 0.5503 0.7761 0.8810
No log 7.5714 212 0.7852 0.5688 0.7852 0.8861
No log 7.6429 214 0.8341 0.4897 0.8341 0.9133
No log 7.7143 216 0.9502 0.4719 0.9502 0.9748
No log 7.7857 218 1.0990 0.4323 1.0990 1.0483
No log 7.8571 220 1.2200 0.4092 1.2200 1.1045
No log 7.9286 222 1.2246 0.4053 1.2246 1.1066
No log 8.0 224 1.1372 0.4443 1.1372 1.0664
No log 8.0714 226 1.0269 0.4882 1.0269 1.0133
No log 8.1429 228 0.9140 0.5071 0.9140 0.9560
No log 8.2143 230 0.8544 0.5544 0.8544 0.9243
No log 8.2857 232 0.8475 0.5520 0.8475 0.9206
No log 8.3571 234 0.8654 0.5520 0.8654 0.9302
No log 8.4286 236 0.9031 0.5379 0.9031 0.9503
No log 8.5 238 0.9732 0.4888 0.9732 0.9865
No log 8.5714 240 1.0607 0.4739 1.0607 1.0299
No log 8.6429 242 1.0967 0.4575 1.0967 1.0472
No log 8.7143 244 1.0932 0.4575 1.0932 1.0455
No log 8.7857 246 1.0631 0.4581 1.0631 1.0311
No log 8.8571 248 1.0395 0.4581 1.0395 1.0196
No log 8.9286 250 1.0217 0.4578 1.0217 1.0108
No log 9.0 252 1.0002 0.4581 1.0002 1.0001
No log 9.0714 254 0.9693 0.4634 0.9693 0.9845
No log 9.1429 256 0.9565 0.4529 0.9565 0.9780
No log 9.2143 258 0.9610 0.4699 0.9610 0.9803
No log 9.2857 260 0.9600 0.4812 0.9600 0.9798
No log 9.3571 262 0.9716 0.4756 0.9716 0.9857
No log 9.4286 264 0.9775 0.4730 0.9775 0.9887
No log 9.5 266 0.9839 0.4730 0.9839 0.9919
No log 9.5714 268 0.9916 0.4730 0.9916 0.9958
No log 9.6429 270 0.9991 0.4730 0.9991 0.9995
No log 9.7143 272 0.9976 0.4730 0.9976 0.9988
No log 9.7857 274 0.9987 0.4730 0.9987 0.9993
No log 9.8571 276 0.9990 0.4730 0.9990 0.9995
No log 9.9286 278 1.0007 0.4730 1.0007 1.0003
No log 10.0 280 1.0013 0.4730 1.0013 1.0006

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k5_task2_organization

Finetuned
(4019)
this model