ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k4_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1117
  • Qwk: 0.4652
  • Mse: 1.1117
  • Rmse: 1.0544

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 4.0224 -0.0062 4.0224 2.0056
No log 0.16 4 2.0194 0.0383 2.0194 1.4211
No log 0.24 6 1.1727 0.0864 1.1727 1.0829
No log 0.32 8 0.8729 0.0130 0.8729 0.9343
No log 0.4 10 0.8119 0.0184 0.8119 0.9011
No log 0.48 12 0.7030 0.2464 0.7030 0.8385
No log 0.56 14 0.7509 0.0864 0.7509 0.8665
No log 0.64 16 0.7880 0.1863 0.7880 0.8877
No log 0.72 18 0.6923 0.3002 0.6923 0.8321
No log 0.8 20 0.7053 0.1395 0.7053 0.8398
No log 0.88 22 0.7063 0.1358 0.7063 0.8404
No log 0.96 24 0.7350 0.1502 0.7350 0.8573
No log 1.04 26 0.6827 0.1778 0.6827 0.8263
No log 1.12 28 0.6552 0.2149 0.6552 0.8095
No log 1.2 30 0.6597 0.1964 0.6597 0.8122
No log 1.28 32 0.8595 0.2581 0.8595 0.9271
No log 1.3600 34 0.8731 0.2746 0.8731 0.9344
No log 1.44 36 0.6802 0.2692 0.6802 0.8248
No log 1.52 38 0.5866 0.3958 0.5866 0.7659
No log 1.6 40 0.6464 0.2742 0.6464 0.8040
No log 1.6800 42 0.6601 0.2781 0.6601 0.8124
No log 1.76 44 0.6699 0.2532 0.6699 0.8185
No log 1.8400 46 0.6419 0.3342 0.6419 0.8012
No log 1.92 48 0.5997 0.3405 0.5997 0.7744
No log 2.0 50 0.5642 0.3357 0.5642 0.7511
No log 2.08 52 0.6078 0.3426 0.6078 0.7796
No log 2.16 54 0.7458 0.3249 0.7458 0.8636
No log 2.24 56 0.7929 0.3444 0.7929 0.8905
No log 2.32 58 0.6949 0.3642 0.6949 0.8336
No log 2.4 60 0.6286 0.4122 0.6286 0.7928
No log 2.48 62 0.8762 0.3429 0.8762 0.9361
No log 2.56 64 1.0111 0.2933 1.0111 1.0056
No log 2.64 66 0.8197 0.3763 0.8197 0.9054
No log 2.7200 68 0.6387 0.4961 0.6387 0.7992
No log 2.8 70 0.6455 0.4949 0.6455 0.8035
No log 2.88 72 0.7669 0.4166 0.7669 0.8757
No log 2.96 74 0.8834 0.4227 0.8834 0.9399
No log 3.04 76 0.9140 0.4508 0.9140 0.9560
No log 3.12 78 0.8293 0.4524 0.8293 0.9106
No log 3.2 80 0.8136 0.4896 0.8136 0.9020
No log 3.2800 82 0.8971 0.4382 0.8971 0.9472
No log 3.36 84 0.9990 0.4302 0.9990 0.9995
No log 3.44 86 0.9584 0.4271 0.9584 0.9790
No log 3.52 88 0.9417 0.4952 0.9417 0.9704
No log 3.6 90 0.9229 0.4903 0.9229 0.9607
No log 3.68 92 0.9485 0.5160 0.9485 0.9739
No log 3.76 94 0.9337 0.4780 0.9337 0.9663
No log 3.84 96 0.9340 0.5121 0.9340 0.9664
No log 3.92 98 0.9151 0.5118 0.9151 0.9566
No log 4.0 100 0.8727 0.4982 0.8727 0.9342
No log 4.08 102 0.8447 0.4801 0.8447 0.9191
No log 4.16 104 0.8237 0.4957 0.8237 0.9076
No log 4.24 106 0.7980 0.4873 0.7980 0.8933
No log 4.32 108 0.8507 0.4784 0.8507 0.9223
No log 4.4 110 0.9171 0.4650 0.9171 0.9577
No log 4.48 112 0.9024 0.4730 0.9024 0.9499
No log 4.5600 114 0.9851 0.4609 0.9851 0.9925
No log 4.64 116 1.0968 0.4163 1.0968 1.0473
No log 4.72 118 1.2476 0.3914 1.2476 1.1170
No log 4.8 120 1.1740 0.3894 1.1740 1.0835
No log 4.88 122 0.9959 0.4801 0.9959 0.9979
No log 4.96 124 0.8876 0.4581 0.8876 0.9421
No log 5.04 126 0.9326 0.4507 0.9326 0.9657
No log 5.12 128 1.0700 0.4253 1.0700 1.0344
No log 5.2 130 1.1195 0.4183 1.1195 1.0580
No log 5.28 132 1.1035 0.4398 1.1035 1.0505
No log 5.36 134 1.0685 0.4941 1.0685 1.0337
No log 5.44 136 1.0504 0.4968 1.0504 1.0249
No log 5.52 138 1.0883 0.4856 1.0883 1.0432
No log 5.6 140 1.1930 0.4282 1.1930 1.0922
No log 5.68 142 1.1929 0.4257 1.1929 1.0922
No log 5.76 144 1.1291 0.4896 1.1291 1.0626
No log 5.84 146 1.0257 0.5051 1.0257 1.0128
No log 5.92 148 0.9913 0.4787 0.9913 0.9956
No log 6.0 150 0.9810 0.4814 0.9810 0.9905
No log 6.08 152 0.9565 0.4440 0.9565 0.9780
No log 6.16 154 0.9405 0.4707 0.9405 0.9698
No log 6.24 156 0.9809 0.4445 0.9809 0.9904
No log 6.32 158 1.0755 0.4268 1.0755 1.0371
No log 6.4 160 1.1245 0.4324 1.1245 1.0604
No log 6.48 162 1.1437 0.4378 1.1437 1.0694
No log 6.5600 164 1.1292 0.4351 1.1292 1.0626
No log 6.64 166 1.0843 0.4371 1.0843 1.0413
No log 6.72 168 1.0570 0.4531 1.0570 1.0281
No log 6.8 170 1.0595 0.4562 1.0595 1.0293
No log 6.88 172 1.0758 0.4770 1.0758 1.0372
No log 6.96 174 1.0578 0.4705 1.0578 1.0285
No log 7.04 176 1.0535 0.4593 1.0535 1.0264
No log 7.12 178 1.0459 0.4581 1.0459 1.0227
No log 7.2 180 1.0542 0.4521 1.0542 1.0267
No log 7.28 182 1.0740 0.4393 1.0740 1.0363
No log 7.36 184 1.0904 0.4408 1.0904 1.0442
No log 7.44 186 1.0565 0.4271 1.0565 1.0278
No log 7.52 188 1.0267 0.4145 1.0267 1.0133
No log 7.6 190 0.9725 0.4272 0.9725 0.9861
No log 7.68 192 0.9266 0.4290 0.9266 0.9626
No log 7.76 194 0.9172 0.4439 0.9172 0.9577
No log 7.84 196 0.9292 0.4424 0.9292 0.9640
No log 7.92 198 0.9463 0.4564 0.9463 0.9728
No log 8.0 200 0.9940 0.4521 0.9940 0.9970
No log 8.08 202 1.0600 0.4303 1.0600 1.0295
No log 8.16 204 1.1199 0.4313 1.1199 1.0583
No log 8.24 206 1.1796 0.4232 1.1796 1.0861
No log 8.32 208 1.1984 0.4246 1.1984 1.0947
No log 8.4 210 1.2067 0.4246 1.2067 1.0985
No log 8.48 212 1.1969 0.4424 1.1969 1.0940
No log 8.56 214 1.1834 0.4435 1.1834 1.0878
No log 8.64 216 1.1740 0.4435 1.1740 1.0835
No log 8.72 218 1.1707 0.4584 1.1707 1.0820
No log 8.8 220 1.1705 0.4584 1.1705 1.0819
No log 8.88 222 1.1782 0.4584 1.1782 1.0855
No log 8.96 224 1.1956 0.4534 1.1956 1.0934
No log 9.04 226 1.1897 0.4625 1.1897 1.0907
No log 9.12 228 1.1707 0.4561 1.1707 1.0820
No log 9.2 230 1.1482 0.4530 1.1482 1.0715
No log 9.28 232 1.1345 0.4486 1.1345 1.0651
No log 9.36 234 1.1229 0.4486 1.1229 1.0597
No log 9.44 236 1.1159 0.4582 1.1159 1.0564
No log 9.52 238 1.1136 0.4582 1.1136 1.0553
No log 9.6 240 1.1117 0.4582 1.1117 1.0544
No log 9.68 242 1.1097 0.4612 1.1097 1.0534
No log 9.76 244 1.1075 0.4612 1.1075 1.0524
No log 9.84 246 1.1086 0.4652 1.1086 1.0529
No log 9.92 248 1.1106 0.4652 1.1106 1.0538
No log 10.0 250 1.1117 0.4652 1.1117 1.0544

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k4_task2_organization

Finetuned
(4023)
this model