ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k5_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9475
  • Qwk: 0.4310
  • Mse: 0.9475
  • Rmse: 0.9734

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 4.1597 -0.0116 4.1597 2.0395
No log 0.1333 4 2.5080 0.0358 2.5080 1.5837
No log 0.2 6 1.2416 0.0864 1.2416 1.1143
No log 0.2667 8 0.9489 -0.0046 0.9489 0.9741
No log 0.3333 10 0.7799 0.0359 0.7799 0.8831
No log 0.4 12 0.8802 0.0086 0.8802 0.9382
No log 0.4667 14 0.8336 0.0763 0.8336 0.9130
No log 0.5333 16 0.7976 0.1308 0.7976 0.8931
No log 0.6 18 0.7446 0.2024 0.7446 0.8629
No log 0.6667 20 0.7042 0.2103 0.7042 0.8392
No log 0.7333 22 0.6787 0.3235 0.6787 0.8238
No log 0.8 24 0.6741 0.2980 0.6741 0.8211
No log 0.8667 26 0.6767 0.2690 0.6767 0.8226
No log 0.9333 28 0.6517 0.3123 0.6517 0.8073
No log 1.0 30 0.6625 0.2705 0.6625 0.8139
No log 1.0667 32 0.6409 0.4105 0.6409 0.8006
No log 1.1333 34 0.6027 0.4332 0.6027 0.7763
No log 1.2 36 0.6123 0.4715 0.6123 0.7825
No log 1.2667 38 0.6779 0.4864 0.6779 0.8234
No log 1.3333 40 0.6342 0.5050 0.6342 0.7964
No log 1.4 42 0.6569 0.4112 0.6569 0.8105
No log 1.4667 44 0.6491 0.5061 0.6491 0.8056
No log 1.5333 46 0.6763 0.5323 0.6763 0.8224
No log 1.6 48 0.7094 0.5315 0.7094 0.8423
No log 1.6667 50 0.6930 0.5229 0.6930 0.8325
No log 1.7333 52 1.0749 0.3387 1.0749 1.0368
No log 1.8 54 0.9787 0.3800 0.9787 0.9893
No log 1.8667 56 0.7230 0.4869 0.7230 0.8503
No log 1.9333 58 0.9999 0.5099 0.9999 0.9999
No log 2.0 60 1.1302 0.3971 1.1302 1.0631
No log 2.0667 62 0.8802 0.5027 0.8802 0.9382
No log 2.1333 64 0.8607 0.5145 0.8607 0.9277
No log 2.2 66 0.8359 0.5322 0.8359 0.9143
No log 2.2667 68 0.9635 0.5365 0.9635 0.9816
No log 2.3333 70 1.0217 0.4957 1.0217 1.0108
No log 2.4 72 0.9469 0.4820 0.9469 0.9731
No log 2.4667 74 0.9690 0.5250 0.9690 0.9844
No log 2.5333 76 1.0663 0.4857 1.0663 1.0326
No log 2.6 78 1.0205 0.4947 1.0205 1.0102
No log 2.6667 80 0.9439 0.4438 0.9439 0.9716
No log 2.7333 82 0.9455 0.4875 0.9455 0.9724
No log 2.8 84 0.9604 0.4696 0.9604 0.9800
No log 2.8667 86 1.1668 0.4533 1.1668 1.0802
No log 2.9333 88 1.0385 0.4789 1.0385 1.0191
No log 3.0 90 0.8093 0.5069 0.8093 0.8996
No log 3.0667 92 0.7671 0.4773 0.7671 0.8759
No log 3.1333 94 0.6894 0.4383 0.6894 0.8303
No log 3.2 96 0.6820 0.4675 0.6820 0.8258
No log 3.2667 98 0.6876 0.4937 0.6876 0.8292
No log 3.3333 100 0.7417 0.4863 0.7417 0.8612
No log 3.4 102 0.9646 0.4968 0.9646 0.9821
No log 3.4667 104 0.9332 0.4965 0.9332 0.9660
No log 3.5333 106 0.8792 0.4499 0.8792 0.9376
No log 3.6 108 0.9572 0.4668 0.9572 0.9783
No log 3.6667 110 1.2718 0.4548 1.2718 1.1277
No log 3.7333 112 1.4120 0.4437 1.4120 1.1883
No log 3.8 114 1.1778 0.4440 1.1778 1.0852
No log 3.8667 116 0.9454 0.4431 0.9454 0.9723
No log 3.9333 118 0.8944 0.4629 0.8944 0.9457
No log 4.0 120 0.8626 0.4449 0.8626 0.9288
No log 4.0667 122 0.8298 0.4188 0.8298 0.9110
No log 4.1333 124 0.7509 0.4514 0.7509 0.8665
No log 4.2 126 0.7220 0.4537 0.7220 0.8497
No log 4.2667 128 0.7333 0.4564 0.7333 0.8563
No log 4.3333 130 0.7786 0.4834 0.7786 0.8824
No log 4.4 132 0.7789 0.4836 0.7789 0.8825
No log 4.4667 134 0.8399 0.4685 0.8399 0.9165
No log 4.5333 136 0.8316 0.4654 0.8316 0.9119
No log 4.6 138 0.7795 0.4734 0.7795 0.8829
No log 4.6667 140 0.8336 0.5170 0.8336 0.9130
No log 4.7333 142 0.8708 0.5103 0.8708 0.9332
No log 4.8 144 0.8636 0.4622 0.8636 0.9293
No log 4.8667 146 0.8812 0.4681 0.8812 0.9387
No log 4.9333 148 0.9719 0.5387 0.9719 0.9858
No log 5.0 150 1.0049 0.5131 1.0049 1.0025
No log 5.0667 152 1.0445 0.5004 1.0445 1.0220
No log 5.1333 154 0.9861 0.5018 0.9861 0.9930
No log 5.2 156 0.8861 0.5018 0.8861 0.9413
No log 5.2667 158 0.8027 0.4263 0.8027 0.8959
No log 5.3333 160 0.8338 0.4339 0.8338 0.9131
No log 5.4 162 0.8629 0.4319 0.8629 0.9289
No log 5.4667 164 0.8843 0.4318 0.8843 0.9404
No log 5.5333 166 0.9130 0.4420 0.9130 0.9555
No log 5.6 168 0.9818 0.4684 0.9818 0.9909
No log 5.6667 170 1.0623 0.4832 1.0623 1.0307
No log 5.7333 172 1.1084 0.4700 1.1084 1.0528
No log 5.8 174 1.0742 0.4609 1.0742 1.0364
No log 5.8667 176 1.0266 0.4145 1.0266 1.0132
No log 5.9333 178 1.0083 0.4138 1.0083 1.0042
No log 6.0 180 0.9823 0.4286 0.9823 0.9911
No log 6.0667 182 0.9468 0.4135 0.9468 0.9730
No log 6.1333 184 0.9801 0.4585 0.9801 0.9900
No log 6.2 186 0.9983 0.4849 0.9983 0.9992
No log 6.2667 188 0.9302 0.4547 0.9302 0.9644
No log 6.3333 190 0.9258 0.4424 0.9258 0.9622
No log 6.4 192 0.9326 0.4730 0.9326 0.9657
No log 6.4667 194 0.9002 0.4371 0.9002 0.9488
No log 6.5333 196 0.9308 0.4416 0.9308 0.9648
No log 6.6 198 0.9909 0.4846 0.9909 0.9954
No log 6.6667 200 0.9711 0.4726 0.9711 0.9854
No log 6.7333 202 0.9365 0.4422 0.9365 0.9677
No log 6.8 204 0.9650 0.4450 0.9650 0.9823
No log 6.8667 206 0.9674 0.4467 0.9674 0.9836
No log 6.9333 208 0.9407 0.4399 0.9407 0.9699
No log 7.0 210 0.9493 0.4313 0.9493 0.9743
No log 7.0667 212 0.9673 0.4489 0.9673 0.9835
No log 7.1333 214 1.0339 0.4631 1.0339 1.0168
No log 7.2 216 1.0583 0.4704 1.0583 1.0287
No log 7.2667 218 1.0481 0.4786 1.0481 1.0238
No log 7.3333 220 1.0095 0.4487 1.0095 1.0047
No log 7.4 222 1.0123 0.4777 1.0123 1.0061
No log 7.4667 224 0.9783 0.4833 0.9783 0.9891
No log 7.5333 226 0.9280 0.4758 0.9280 0.9633
No log 7.6 228 0.9025 0.4692 0.9025 0.9500
No log 7.6667 230 0.8961 0.4719 0.8961 0.9466
No log 7.7333 232 0.8872 0.4617 0.8872 0.9419
No log 7.8 234 0.9038 0.4707 0.9038 0.9507
No log 7.8667 236 0.9328 0.5009 0.9328 0.9658
No log 7.9333 238 0.9336 0.4930 0.9336 0.9662
No log 8.0 240 0.9031 0.4557 0.9031 0.9503
No log 8.0667 242 0.8810 0.4379 0.8810 0.9386
No log 8.1333 244 0.8739 0.4441 0.8739 0.9348
No log 8.2 246 0.8873 0.4380 0.8873 0.9420
No log 8.2667 248 0.8892 0.4379 0.8892 0.9430
No log 8.3333 250 0.8890 0.4379 0.8890 0.9429
No log 8.4 252 0.8879 0.4379 0.8879 0.9423
No log 8.4667 254 0.8790 0.4725 0.8790 0.9376
No log 8.5333 256 0.8835 0.4689 0.8835 0.9399
No log 8.6 258 0.8969 0.4671 0.8969 0.9470
No log 8.6667 260 0.9197 0.4524 0.9197 0.9590
No log 8.7333 262 0.9386 0.4491 0.9386 0.9688
No log 8.8 264 0.9445 0.4375 0.9445 0.9719
No log 8.8667 266 0.9383 0.4429 0.9383 0.9686
No log 8.9333 268 0.9414 0.4311 0.9414 0.9703
No log 9.0 270 0.9547 0.4436 0.9547 0.9771
No log 9.0667 272 0.9633 0.4629 0.9633 0.9815
No log 9.1333 274 0.9641 0.4629 0.9641 0.9819
No log 9.2 276 0.9675 0.4629 0.9675 0.9836
No log 9.2667 278 0.9591 0.4573 0.9591 0.9793
No log 9.3333 280 0.9528 0.4518 0.9528 0.9761
No log 9.4 282 0.9506 0.4458 0.9506 0.9750
No log 9.4667 284 0.9572 0.4573 0.9572 0.9784
No log 9.5333 286 0.9589 0.4511 0.9589 0.9792
No log 9.6 288 0.9630 0.4555 0.9630 0.9813
No log 9.6667 290 0.9668 0.4555 0.9668 0.9832
No log 9.7333 292 0.9652 0.4555 0.9652 0.9824
No log 9.8 294 0.9593 0.4573 0.9593 0.9794
No log 9.8667 296 0.9533 0.4458 0.9533 0.9764
No log 9.9333 298 0.9490 0.4424 0.9490 0.9742
No log 10.0 300 0.9475 0.4310 0.9475 0.9734

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k5_task2_organization

Finetuned
(4023)
this model