ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k4_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0043
  • Qwk: 0.4610
  • Mse: 1.0043
  • Rmse: 1.0021

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 4.2326 -0.0051 4.2326 2.0573
No log 0.16 4 2.1027 0.0578 2.1027 1.4501
No log 0.24 6 1.3092 0.1474 1.3092 1.1442
No log 0.32 8 0.8169 0.1395 0.8169 0.9038
No log 0.4 10 0.6840 0.1882 0.6840 0.8270
No log 0.48 12 0.7687 0.1028 0.7687 0.8768
No log 0.56 14 0.8230 0.0918 0.8230 0.9072
No log 0.64 16 0.7966 0.1447 0.7966 0.8925
No log 0.72 18 0.6775 0.1241 0.6775 0.8231
No log 0.8 20 0.7279 0.2901 0.7279 0.8532
No log 0.88 22 0.7059 0.3030 0.7059 0.8402
No log 0.96 24 0.6909 0.3342 0.6909 0.8312
No log 1.04 26 0.6091 0.375 0.6091 0.7804
No log 1.12 28 0.6099 0.3195 0.6099 0.7810
No log 1.2 30 0.5851 0.3394 0.5851 0.7649
No log 1.28 32 0.5610 0.4556 0.5610 0.7490
No log 1.3600 34 0.5654 0.4431 0.5654 0.7519
No log 1.44 36 0.6382 0.4827 0.6382 0.7989
No log 1.52 38 0.8306 0.3981 0.8306 0.9114
No log 1.6 40 0.8404 0.4247 0.8404 0.9167
No log 1.6800 42 0.6982 0.5190 0.6982 0.8356
No log 1.76 44 1.0030 0.3898 1.0030 1.0015
No log 1.8400 46 1.1895 0.2890 1.1895 1.0906
No log 1.92 48 0.9583 0.4380 0.9583 0.9789
No log 2.0 50 0.7778 0.4460 0.7778 0.8819
No log 2.08 52 0.7495 0.4987 0.7495 0.8657
No log 2.16 54 0.7300 0.5051 0.7300 0.8544
No log 2.24 56 0.7198 0.4906 0.7198 0.8484
No log 2.32 58 0.7395 0.4983 0.7395 0.8599
No log 2.4 60 0.7321 0.4954 0.7321 0.8556
No log 2.48 62 0.7379 0.5039 0.7379 0.8590
No log 2.56 64 0.7942 0.4992 0.7942 0.8912
No log 2.64 66 1.0320 0.4592 1.0320 1.0159
No log 2.7200 68 0.9575 0.4598 0.9575 0.9785
No log 2.8 70 0.8996 0.4935 0.8996 0.9485
No log 2.88 72 0.9453 0.4856 0.9453 0.9723
No log 2.96 74 0.9263 0.4805 0.9263 0.9625
No log 3.04 76 0.9090 0.5104 0.9090 0.9534
No log 3.12 78 0.9228 0.4685 0.9228 0.9606
No log 3.2 80 0.8833 0.5124 0.8833 0.9398
No log 3.2800 82 0.8541 0.4971 0.8541 0.9242
No log 3.36 84 0.8429 0.5147 0.8429 0.9181
No log 3.44 86 0.8452 0.5286 0.8452 0.9194
No log 3.52 88 0.7993 0.5357 0.7993 0.8940
No log 3.6 90 0.8277 0.4411 0.8277 0.9098
No log 3.68 92 0.8156 0.4350 0.8156 0.9031
No log 3.76 94 0.8137 0.4842 0.8137 0.9021
No log 3.84 96 0.8645 0.5090 0.8645 0.9298
No log 3.92 98 0.8947 0.4964 0.8947 0.9459
No log 4.0 100 0.9527 0.4453 0.9527 0.9760
No log 4.08 102 1.0212 0.4228 1.0212 1.0106
No log 4.16 104 1.0004 0.4614 1.0004 1.0002
No log 4.24 106 0.9908 0.4753 0.9908 0.9954
No log 4.32 108 0.9804 0.4753 0.9804 0.9902
No log 4.4 110 0.9759 0.4753 0.9759 0.9879
No log 4.48 112 0.9476 0.4478 0.9476 0.9734
No log 4.5600 114 0.9493 0.4887 0.9493 0.9743
No log 4.64 116 1.0895 0.4361 1.0895 1.0438
No log 4.72 118 1.1371 0.3963 1.1371 1.0663
No log 4.8 120 0.9869 0.4650 0.9869 0.9934
No log 4.88 122 0.8731 0.4927 0.8731 0.9344
No log 4.96 124 0.8623 0.4787 0.8623 0.9286
No log 5.04 126 0.8562 0.4860 0.8562 0.9253
No log 5.12 128 0.8523 0.5092 0.8523 0.9232
No log 5.2 130 0.8657 0.5353 0.8657 0.9304
No log 5.28 132 0.9492 0.4630 0.9492 0.9743
No log 5.36 134 1.0265 0.4313 1.0265 1.0132
No log 5.44 136 0.9505 0.4532 0.9505 0.9749
No log 5.52 138 0.8281 0.5224 0.8281 0.9100
No log 5.6 140 0.8200 0.4379 0.8200 0.9055
No log 5.68 142 0.8292 0.4343 0.8292 0.9106
No log 5.76 144 0.8289 0.4796 0.8289 0.9104
No log 5.84 146 0.9124 0.4748 0.9124 0.9552
No log 5.92 148 1.1171 0.4113 1.1171 1.0569
No log 6.0 150 1.1839 0.3846 1.1839 1.0880
No log 6.08 152 1.0839 0.4657 1.0839 1.0411
No log 6.16 154 1.0077 0.4567 1.0077 1.0038
No log 6.24 156 1.0301 0.4435 1.0301 1.0149
No log 6.32 158 1.0606 0.4422 1.0606 1.0298
No log 6.4 160 1.0564 0.4496 1.0564 1.0278
No log 6.48 162 1.0266 0.4554 1.0266 1.0132
No log 6.5600 164 1.0117 0.4599 1.0117 1.0059
No log 6.64 166 0.9940 0.4478 0.9940 0.9970
No log 6.72 168 0.9740 0.4416 0.9740 0.9869
No log 6.8 170 0.9512 0.4543 0.9512 0.9753
No log 6.88 172 0.9393 0.4699 0.9393 0.9692
No log 6.96 174 0.9273 0.4648 0.9273 0.9630
No log 7.04 176 0.9109 0.4620 0.9109 0.9544
No log 7.12 178 0.9425 0.4635 0.9425 0.9708
No log 7.2 180 0.9776 0.4239 0.9776 0.9887
No log 7.28 182 0.9689 0.4516 0.9689 0.9843
No log 7.36 184 0.9792 0.4727 0.9792 0.9895
No log 7.44 186 1.0011 0.4482 1.0011 1.0005
No log 7.52 188 1.0194 0.4250 1.0194 1.0097
No log 7.6 190 1.0323 0.4299 1.0323 1.0160
No log 7.68 192 1.0453 0.4487 1.0453 1.0224
No log 7.76 194 1.0526 0.4299 1.0526 1.0260
No log 7.84 196 1.0696 0.4281 1.0696 1.0342
No log 7.92 198 1.0846 0.3983 1.0846 1.0414
No log 8.0 200 1.0866 0.4087 1.0866 1.0424
No log 8.08 202 1.0820 0.4281 1.0820 1.0402
No log 8.16 204 1.0864 0.4369 1.0864 1.0423
No log 8.24 206 1.0972 0.4193 1.0972 1.0475
No log 8.32 208 1.0999 0.4262 1.0999 1.0487
No log 8.4 210 1.0891 0.4191 1.0891 1.0436
No log 8.48 212 1.0891 0.4235 1.0891 1.0436
No log 8.56 214 1.0922 0.3932 1.0922 1.0451
No log 8.64 216 1.1021 0.3976 1.1021 1.0498
No log 8.72 218 1.0865 0.3932 1.0865 1.0424
No log 8.8 220 1.0708 0.4368 1.0708 1.0348
No log 8.88 222 1.0635 0.4190 1.0635 1.0313
No log 8.96 224 1.0560 0.4190 1.0560 1.0276
No log 9.04 226 1.0498 0.4097 1.0498 1.0246
No log 9.12 228 1.0376 0.4097 1.0376 1.0186
No log 9.2 230 1.0254 0.4094 1.0254 1.0126
No log 9.28 232 1.0185 0.4171 1.0185 1.0092
No log 9.36 234 1.0148 0.4343 1.0148 1.0074
No log 9.44 236 1.0106 0.4641 1.0106 1.0053
No log 9.52 238 1.0094 0.4550 1.0094 1.0047
No log 9.6 240 1.0090 0.4311 1.0090 1.0045
No log 9.68 242 1.0075 0.4311 1.0075 1.0037
No log 9.76 244 1.0054 0.4437 1.0054 1.0027
No log 9.84 246 1.0037 0.4610 1.0037 1.0018
No log 9.92 248 1.0039 0.4610 1.0039 1.0020
No log 10.0 250 1.0043 0.4610 1.0043 1.0021

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k4_task2_organization

Finetuned
(4019)
this model