ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k4_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8226
  • Qwk: 0.6881
  • Mse: 0.8226
  • Rmse: 0.9070

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0952 2 2.3356 0.0562 2.3356 1.5283
No log 0.1905 4 1.6075 0.1917 1.6075 1.2679
No log 0.2857 6 1.3963 0.1914 1.3963 1.1816
No log 0.3810 8 1.3428 0.2092 1.3428 1.1588
No log 0.4762 10 1.4381 0.1757 1.4381 1.1992
No log 0.5714 12 1.3845 0.2535 1.3845 1.1766
No log 0.6667 14 1.2096 0.3256 1.2096 1.0998
No log 0.7619 16 1.2696 0.4150 1.2696 1.1267
No log 0.8571 18 1.2827 0.2115 1.2827 1.1326
No log 0.9524 20 1.3187 0.2542 1.3187 1.1483
No log 1.0476 22 1.3004 0.2695 1.3004 1.1404
No log 1.1429 24 1.2279 0.2991 1.2279 1.1081
No log 1.2381 26 1.1328 0.2991 1.1328 1.0643
No log 1.3333 28 1.0992 0.3342 1.0992 1.0484
No log 1.4286 30 1.0866 0.3538 1.0866 1.0424
No log 1.5238 32 1.0473 0.3344 1.0473 1.0234
No log 1.6190 34 1.0018 0.3920 1.0018 1.0009
No log 1.7143 36 0.9730 0.4786 0.9730 0.9864
No log 1.8095 38 0.9185 0.5027 0.9185 0.9584
No log 1.9048 40 0.8852 0.5342 0.8852 0.9408
No log 2.0 42 0.8663 0.5585 0.8663 0.9307
No log 2.0952 44 0.8868 0.5313 0.8868 0.9417
No log 2.1905 46 0.9275 0.5503 0.9275 0.9631
No log 2.2857 48 0.9614 0.5428 0.9614 0.9805
No log 2.3810 50 0.8632 0.5363 0.8632 0.9291
No log 2.4762 52 0.8310 0.5603 0.8310 0.9116
No log 2.5714 54 0.8587 0.5806 0.8587 0.9267
No log 2.6667 56 0.8643 0.5876 0.8643 0.9297
No log 2.7619 58 0.8517 0.6246 0.8517 0.9229
No log 2.8571 60 0.8241 0.6545 0.8241 0.9078
No log 2.9524 62 0.7932 0.5805 0.7932 0.8906
No log 3.0476 64 0.8481 0.5426 0.8481 0.9209
No log 3.1429 66 0.8272 0.5398 0.8272 0.9095
No log 3.2381 68 0.7982 0.6441 0.7982 0.8934
No log 3.3333 70 0.9506 0.5910 0.9506 0.9750
No log 3.4286 72 1.0133 0.5745 1.0133 1.0066
No log 3.5238 74 1.0623 0.5561 1.0623 1.0307
No log 3.6190 76 0.9367 0.6226 0.9367 0.9679
No log 3.7143 78 0.8342 0.6517 0.8342 0.9134
No log 3.8095 80 0.7344 0.6554 0.7344 0.8570
No log 3.9048 82 0.7209 0.6636 0.7209 0.8491
No log 4.0 84 0.7255 0.6743 0.7255 0.8518
No log 4.0952 86 0.7626 0.6663 0.7626 0.8733
No log 4.1905 88 0.8657 0.6535 0.8657 0.9305
No log 4.2857 90 1.0346 0.5899 1.0346 1.0172
No log 4.3810 92 1.1264 0.5697 1.1264 1.0613
No log 4.4762 94 1.0858 0.5697 1.0858 1.0420
No log 4.5714 96 0.9184 0.6483 0.9184 0.9583
No log 4.6667 98 0.7354 0.6822 0.7354 0.8575
No log 4.7619 100 0.6951 0.6847 0.6951 0.8338
No log 4.8571 102 0.6824 0.6968 0.6824 0.8261
No log 4.9524 104 0.7326 0.6732 0.7326 0.8559
No log 5.0476 106 0.8846 0.6759 0.8846 0.9405
No log 5.1429 108 1.0194 0.6474 1.0194 1.0096
No log 5.2381 110 0.9861 0.6541 0.9861 0.9930
No log 5.3333 112 0.9347 0.6530 0.9347 0.9668
No log 5.4286 114 0.9441 0.6458 0.9441 0.9716
No log 5.5238 116 0.9118 0.6573 0.9118 0.9549
No log 5.6190 118 0.9727 0.6561 0.9727 0.9863
No log 5.7143 120 1.0022 0.6471 1.0022 1.0011
No log 5.8095 122 0.9299 0.6687 0.9299 0.9643
No log 5.9048 124 0.8760 0.6559 0.8760 0.9360
No log 6.0 126 0.8259 0.6990 0.8259 0.9088
No log 6.0952 128 0.7652 0.6888 0.7652 0.8748
No log 6.1905 130 0.7354 0.6961 0.7354 0.8575
No log 6.2857 132 0.7532 0.7062 0.7532 0.8679
No log 6.3810 134 0.8266 0.6979 0.8266 0.9092
No log 6.4762 136 0.9212 0.6606 0.9212 0.9598
No log 6.5714 138 0.9031 0.6654 0.9031 0.9503
No log 6.6667 140 0.8545 0.6886 0.8545 0.9244
No log 6.7619 142 0.7787 0.7108 0.7787 0.8825
No log 6.8571 144 0.7030 0.7270 0.7030 0.8384
No log 6.9524 146 0.6768 0.7177 0.6768 0.8227
No log 7.0476 148 0.6743 0.7177 0.6743 0.8212
No log 7.1429 150 0.7197 0.7272 0.7197 0.8483
No log 7.2381 152 0.8165 0.6800 0.8165 0.9036
No log 7.3333 154 0.9181 0.6671 0.9181 0.9582
No log 7.4286 156 1.0155 0.6265 1.0155 1.0077
No log 7.5238 158 1.0252 0.6418 1.0252 1.0125
No log 7.6190 160 0.9684 0.6525 0.9684 0.9841
No log 7.7143 162 0.8914 0.6456 0.8914 0.9441
No log 7.8095 164 0.8216 0.6806 0.8216 0.9064
No log 7.9048 166 0.7952 0.6814 0.7952 0.8917
No log 8.0 168 0.8119 0.6739 0.8119 0.9011
No log 8.0952 170 0.8447 0.6722 0.8447 0.9191
No log 8.1905 172 0.8836 0.6408 0.8836 0.9400
No log 8.2857 174 0.9357 0.6441 0.9357 0.9673
No log 8.3810 176 0.9503 0.6545 0.9503 0.9748
No log 8.4762 178 0.9361 0.6577 0.9361 0.9675
No log 8.5714 180 0.8954 0.6764 0.8954 0.9463
No log 8.6667 182 0.8567 0.6820 0.8567 0.9256
No log 8.7619 184 0.8146 0.7141 0.8146 0.9026
No log 8.8571 186 0.7700 0.7146 0.7700 0.8775
No log 8.9524 188 0.7302 0.7229 0.7302 0.8545
No log 9.0476 190 0.7138 0.7352 0.7138 0.8448
No log 9.1429 192 0.7036 0.7352 0.7036 0.8388
No log 9.2381 194 0.7100 0.7352 0.7100 0.8426
No log 9.3333 196 0.7303 0.7286 0.7303 0.8546
No log 9.4286 198 0.7554 0.7186 0.7554 0.8691
No log 9.5238 200 0.7804 0.7053 0.7804 0.8834
No log 9.6190 202 0.7921 0.7086 0.7921 0.8900
No log 9.7143 204 0.8014 0.7086 0.8014 0.8952
No log 9.8095 206 0.8133 0.7086 0.8133 0.9018
No log 9.9048 208 0.8197 0.6921 0.8197 0.9054
No log 10.0 210 0.8226 0.6881 0.8226 0.9070

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k4_task5_organization

Finetuned
(4023)
this model