ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k5_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7588
  • Qwk: 0.7004
  • Mse: 0.7588
  • Rmse: 0.8711

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1 2 2.3515 -0.0201 2.3515 1.5334
No log 0.2 4 1.5743 0.1804 1.5743 1.2547
No log 0.3 6 1.4911 0.0962 1.4911 1.2211
No log 0.4 8 1.5553 0.1955 1.5553 1.2471
No log 0.5 10 1.6274 0.3369 1.6274 1.2757
No log 0.6 12 1.4857 0.1276 1.4857 1.2189
No log 0.7 14 1.4648 0.1009 1.4648 1.2103
No log 0.8 16 1.4577 0.0811 1.4577 1.2074
No log 0.9 18 1.4115 0.1057 1.4115 1.1881
No log 1.0 20 1.4379 0.0859 1.4379 1.1991
No log 1.1 22 1.5012 0.2174 1.5012 1.2252
No log 1.2 24 1.5528 0.2731 1.5528 1.2461
No log 1.3 26 1.6686 0.3214 1.6686 1.2917
No log 1.4 28 1.8822 0.2788 1.8822 1.3719
No log 1.5 30 1.8667 0.2934 1.8667 1.3663
No log 1.6 32 1.6847 0.3642 1.6847 1.2980
No log 1.7 34 1.4215 0.4073 1.4215 1.1923
No log 1.8 36 1.2429 0.2717 1.2429 1.1148
No log 1.9 38 1.1813 0.2389 1.1813 1.0869
No log 2.0 40 1.1926 0.2452 1.1926 1.0920
No log 2.1 42 1.1651 0.2718 1.1651 1.0794
No log 2.2 44 1.1074 0.3281 1.1074 1.0523
No log 2.3 46 1.1226 0.4217 1.1226 1.0595
No log 2.4 48 1.2471 0.4149 1.2471 1.1167
No log 2.5 50 1.2194 0.4286 1.2194 1.1042
No log 2.6 52 1.1923 0.4073 1.1923 1.0919
No log 2.7 54 1.3077 0.4128 1.3077 1.1436
No log 2.8 56 1.3472 0.3891 1.3472 1.1607
No log 2.9 58 1.3315 0.4126 1.3315 1.1539
No log 3.0 60 1.1433 0.4032 1.1433 1.0693
No log 3.1 62 1.0340 0.4439 1.0340 1.0168
No log 3.2 64 1.0286 0.4327 1.0286 1.0142
No log 3.3 66 1.0400 0.4574 1.0400 1.0198
No log 3.4 68 1.0216 0.4730 1.0216 1.0107
No log 3.5 70 0.9581 0.5057 0.9581 0.9788
No log 3.6 72 1.0332 0.4856 1.0332 1.0165
No log 3.7 74 1.3527 0.4817 1.3527 1.1631
No log 3.8 76 1.5324 0.4636 1.5324 1.2379
No log 3.9 78 1.5564 0.4596 1.5564 1.2476
No log 4.0 80 1.3033 0.4817 1.3033 1.1416
No log 4.1 82 1.0266 0.5126 1.0266 1.0132
No log 4.2 84 0.9423 0.5172 0.9423 0.9707
No log 4.3 86 0.9304 0.5398 0.9304 0.9646
No log 4.4 88 0.9557 0.5490 0.9557 0.9776
No log 4.5 90 0.9740 0.5425 0.9740 0.9869
No log 4.6 92 1.0209 0.5455 1.0209 1.0104
No log 4.7 94 1.1036 0.5256 1.1036 1.0505
No log 4.8 96 1.0424 0.5374 1.0424 1.0210
No log 4.9 98 0.9124 0.6298 0.9124 0.9552
No log 5.0 100 0.8661 0.6385 0.8661 0.9306
No log 5.1 102 0.8333 0.6469 0.8333 0.9129
No log 5.2 104 0.8442 0.6451 0.8442 0.9188
No log 5.3 106 0.9257 0.6301 0.9257 0.9621
No log 5.4 108 1.0467 0.6099 1.0467 1.0231
No log 5.5 110 0.9822 0.6300 0.9822 0.9911
No log 5.6 112 0.9660 0.6286 0.9660 0.9828
No log 5.7 114 0.8785 0.6523 0.8785 0.9373
No log 5.8 116 0.7643 0.6827 0.7643 0.8742
No log 5.9 118 0.7212 0.6999 0.7212 0.8492
No log 6.0 120 0.7327 0.6985 0.7327 0.8560
No log 6.1 122 0.8022 0.6622 0.8022 0.8956
No log 6.2 124 0.9035 0.6282 0.9035 0.9505
No log 6.3 126 1.0753 0.5703 1.0753 1.0370
No log 6.4 128 1.1506 0.5641 1.1506 1.0727
No log 6.5 130 1.0722 0.5714 1.0722 1.0355
No log 6.6 132 0.9181 0.6298 0.9181 0.9582
No log 6.7 134 0.8269 0.6264 0.8269 0.9093
No log 6.8 136 0.7880 0.6656 0.7880 0.8877
No log 6.9 138 0.7953 0.6667 0.7953 0.8918
No log 7.0 140 0.8240 0.6447 0.8240 0.9077
No log 7.1 142 0.8631 0.6459 0.8631 0.9291
No log 7.2 144 0.8408 0.6531 0.8408 0.9169
No log 7.3 146 0.8036 0.6659 0.8036 0.8964
No log 7.4 148 0.8183 0.6632 0.8183 0.9046
No log 7.5 150 0.8124 0.6613 0.8124 0.9013
No log 7.6 152 0.8293 0.6742 0.8293 0.9107
No log 7.7 154 0.8371 0.6679 0.8371 0.9149
No log 7.8 156 0.8110 0.6734 0.8110 0.9005
No log 7.9 158 0.7885 0.6781 0.7885 0.8880
No log 8.0 160 0.7880 0.6781 0.7880 0.8877
No log 8.1 162 0.7689 0.6762 0.7689 0.8769
No log 8.2 164 0.7546 0.6914 0.7546 0.8687
No log 8.3 166 0.7296 0.7224 0.7296 0.8542
No log 8.4 168 0.7191 0.7207 0.7191 0.8480
No log 8.5 170 0.7092 0.7262 0.7092 0.8422
No log 8.6 172 0.6987 0.7310 0.6987 0.8359
No log 8.7 174 0.6963 0.7310 0.6963 0.8344
No log 8.8 176 0.6995 0.7222 0.6995 0.8364
No log 8.9 178 0.7148 0.7188 0.7148 0.8454
No log 9.0 180 0.7417 0.7180 0.7417 0.8612
No log 9.1 182 0.7664 0.7033 0.7664 0.8754
No log 9.2 184 0.7825 0.6923 0.7825 0.8846
No log 9.3 186 0.7932 0.6878 0.7932 0.8906
No log 9.4 188 0.7932 0.6878 0.7932 0.8906
No log 9.5 190 0.7877 0.6860 0.7877 0.8875
No log 9.6 192 0.7779 0.6906 0.7779 0.8820
No log 9.7 194 0.7690 0.6906 0.7690 0.8769
No log 9.8 196 0.7623 0.7004 0.7623 0.8731
No log 9.9 198 0.7594 0.7004 0.7594 0.8714
No log 10.0 200 0.7588 0.7004 0.7588 0.8711

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k5_task5_organization

Finetuned
(4023)
this model