ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5767
  • Qwk: 0.3725
  • Mse: 0.5767
  • Rmse: 0.7594

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.0480 0.0243 3.0480 1.7459
No log 0.1 4 2.0274 0.0 2.0274 1.4239
No log 0.15 6 1.3139 0.0255 1.3139 1.1462
No log 0.2 8 1.2496 -0.0327 1.2496 1.1178
No log 0.25 10 0.7973 0.0370 0.7973 0.8929
No log 0.3 12 0.7432 0.1765 0.7432 0.8621
No log 0.35 14 1.0186 0.0388 1.0186 1.0093
No log 0.4 16 1.1159 0.0388 1.1159 1.0564
No log 0.45 18 0.8029 0.1579 0.8029 0.8960
No log 0.5 20 0.6351 0.0815 0.6351 0.7969
No log 0.55 22 0.5943 -0.0081 0.5943 0.7709
No log 0.6 24 0.5801 0.0 0.5801 0.7616
No log 0.65 26 0.5615 0.0388 0.5615 0.7493
No log 0.7 28 1.2011 0.0817 1.2011 1.0959
No log 0.75 30 0.9884 0.1111 0.9884 0.9942
No log 0.8 32 0.5902 0.0256 0.5902 0.7683
No log 0.85 34 0.6365 0.0720 0.6365 0.7978
No log 0.9 36 0.6860 0.2000 0.6860 0.8283
No log 0.95 38 0.6061 0.0 0.6061 0.7785
No log 1.0 40 0.5664 0.0 0.5664 0.7526
No log 1.05 42 0.5758 0.0 0.5758 0.7588
No log 1.1 44 0.5739 0.0 0.5739 0.7576
No log 1.15 46 0.5298 0.0 0.5298 0.7278
No log 1.2 48 0.5289 0.0 0.5289 0.7273
No log 1.25 50 0.5288 0.0 0.5288 0.7272
No log 1.3 52 0.5855 0.0400 0.5855 0.7652
No log 1.35 54 0.6567 0.1913 0.6567 0.8104
No log 1.4 56 0.8137 0.1453 0.8137 0.9021
No log 1.45 58 0.6289 0.1590 0.6289 0.7930
No log 1.5 60 0.5590 0.1448 0.5590 0.7477
No log 1.55 62 0.6107 0.2000 0.6107 0.7815
No log 1.6 64 0.5629 0.1688 0.5629 0.7503
No log 1.65 66 0.5922 0.2370 0.5922 0.7695
No log 1.7 68 0.6366 0.3073 0.6366 0.7979
No log 1.75 70 0.6554 0.1813 0.6554 0.8096
No log 1.8 72 0.8734 0.1628 0.8734 0.9345
No log 1.85 74 0.5683 0.3086 0.5683 0.7539
No log 1.9 76 0.7412 0.2919 0.7412 0.8610
No log 1.95 78 0.6788 0.3402 0.6788 0.8239
No log 2.0 80 0.5690 0.2663 0.5690 0.7543
No log 2.05 82 0.6099 0.3548 0.6099 0.7809
No log 2.1 84 0.5704 0.2298 0.5704 0.7552
No log 2.15 86 0.6876 0.2184 0.6876 0.8292
No log 2.2 88 0.9162 0.1861 0.9162 0.9572
No log 2.25 90 0.7006 0.2093 0.7006 0.8370
No log 2.3 92 0.6548 0.1823 0.6548 0.8092
No log 2.35 94 0.6914 0.2251 0.6914 0.8315
No log 2.4 96 0.5920 0.1304 0.5920 0.7694
No log 2.45 98 0.8022 0.2233 0.8022 0.8957
No log 2.5 100 0.6253 0.2088 0.6253 0.7907
No log 2.55 102 0.7109 0.3103 0.7109 0.8432
No log 2.6 104 0.6160 0.1556 0.6160 0.7848
No log 2.65 106 1.1635 0.1367 1.1635 1.0787
No log 2.7 108 1.4410 0.1656 1.4410 1.2004
No log 2.75 110 0.8380 0.1776 0.8380 0.9154
No log 2.8 112 0.6822 0.2941 0.6822 0.8259
No log 2.85 114 0.6317 0.3016 0.6317 0.7948
No log 2.9 116 0.6146 0.2746 0.6146 0.7840
No log 2.95 118 0.8986 0.1261 0.8986 0.9480
No log 3.0 120 1.2285 0.1943 1.2285 1.1084
No log 3.05 122 0.7410 0.2239 0.7410 0.8608
No log 3.1 124 0.5985 0.2432 0.5985 0.7736
No log 3.15 126 0.5890 0.3182 0.5890 0.7675
No log 3.2 128 0.6306 0.2527 0.6306 0.7941
No log 3.25 130 0.6086 0.1617 0.6086 0.7801
No log 3.3 132 0.9076 0.1366 0.9076 0.9527
No log 3.35 134 1.0742 0.1008 1.0742 1.0364
No log 3.4 136 0.7710 0.1908 0.7710 0.8781
No log 3.45 138 0.8522 0.0884 0.8522 0.9231
No log 3.5 140 1.2307 0.1062 1.2307 1.1094
No log 3.55 142 0.9685 0.0847 0.9685 0.9841
No log 3.6 144 0.6888 0.1086 0.6888 0.8299
No log 3.65 146 0.6631 0.1807 0.6631 0.8143
No log 3.7 148 0.7592 0.1340 0.7592 0.8713
No log 3.75 150 0.6799 0.2432 0.6799 0.8246
No log 3.8 152 0.6370 0.2626 0.6370 0.7981
No log 3.85 154 0.7618 0.2410 0.7618 0.8728
No log 3.9 156 0.6331 0.2917 0.6331 0.7957
No log 3.95 158 0.6507 0.2893 0.6507 0.8067
No log 4.0 160 0.7210 0.2453 0.7210 0.8491
No log 4.05 162 0.8110 0.2222 0.8110 0.9006
No log 4.1 164 0.7866 0.2146 0.7866 0.8869
No log 4.15 166 0.7948 0.2222 0.7948 0.8915
No log 4.2 168 0.7123 0.3028 0.7123 0.8440
No log 4.25 170 0.6928 0.2315 0.6928 0.8324
No log 4.3 172 0.6562 0.2487 0.6562 0.8101
No log 4.35 174 0.6138 0.4019 0.6138 0.7835
No log 4.4 176 0.7349 0.2523 0.7349 0.8573
No log 4.45 178 0.6593 0.1675 0.6593 0.8120
No log 4.5 180 0.5582 0.3371 0.5582 0.7471
No log 4.55 182 0.5873 0.3073 0.5873 0.7664
No log 4.6 184 0.5343 0.3829 0.5343 0.7309
No log 4.65 186 0.8398 0.2281 0.8398 0.9164
No log 4.7 188 1.1123 0.2226 1.1123 1.0547
No log 4.75 190 0.8585 0.2203 0.8585 0.9266
No log 4.8 192 0.5911 0.3641 0.5911 0.7688
No log 4.85 194 0.5892 0.3831 0.5892 0.7676
No log 4.9 196 0.7271 0.2000 0.7271 0.8527
No log 4.95 198 1.1137 0.2109 1.1137 1.0553
No log 5.0 200 1.0426 0.2115 1.0426 1.0211
No log 5.05 202 0.7081 0.3301 0.7081 0.8415
No log 5.1 204 0.6532 0.4178 0.6532 0.8082
No log 5.15 206 0.6490 0.4074 0.6490 0.8056
No log 5.2 208 0.6751 0.3488 0.6751 0.8217
No log 5.25 210 0.9602 0.1181 0.9602 0.9799
No log 5.3 212 0.9370 0.1148 0.9370 0.9680
No log 5.35 214 0.6520 0.3398 0.6520 0.8075
No log 5.4 216 0.6045 0.4444 0.6045 0.7775
No log 5.45 218 0.7076 0.2727 0.7076 0.8412
No log 5.5 220 0.8876 0.1535 0.8876 0.9421
No log 5.55 222 1.0566 0.2174 1.0566 1.0279
No log 5.6 224 0.8229 0.1535 0.8229 0.9071
No log 5.65 226 0.5609 0.4343 0.5609 0.7489
No log 5.7 228 0.5824 0.3469 0.5824 0.7631
No log 5.75 230 0.5605 0.4518 0.5605 0.7487
No log 5.8 232 0.6749 0.2727 0.6749 0.8215
No log 5.85 234 1.0728 0.1888 1.0728 1.0358
No log 5.9 236 1.1176 0.2165 1.1176 1.0571
No log 5.95 238 0.8432 0.1790 0.8432 0.9183
No log 6.0 240 0.6230 0.3962 0.6230 0.7893
No log 6.05 242 0.6124 0.4010 0.6124 0.7826
No log 6.1 244 0.6964 0.2157 0.6964 0.8345
No log 6.15 246 0.7895 0.1790 0.7895 0.8885
No log 6.2 248 0.9153 0.2450 0.9153 0.9567
No log 6.25 250 0.7975 0.1790 0.7975 0.8931
No log 6.3 252 0.5991 0.3297 0.5991 0.7740
No log 6.35 254 0.5631 0.3591 0.5631 0.7504
No log 6.4 256 0.5768 0.3333 0.5768 0.7595
No log 6.45 258 0.6395 0.2174 0.6395 0.7997
No log 6.5 260 0.7758 0.1402 0.7758 0.8808
No log 6.55 262 0.8126 0.2212 0.8126 0.9015
No log 6.6 264 0.7287 0.2161 0.7287 0.8537
No log 6.65 266 0.6414 0.2727 0.6414 0.8009
No log 6.7 268 0.5848 0.3708 0.5848 0.7647
No log 6.75 270 0.5719 0.3591 0.5719 0.7562
No log 6.8 272 0.5625 0.4098 0.5625 0.7500
No log 6.85 274 0.6016 0.3035 0.6016 0.7756
No log 6.9 276 0.6288 0.28 0.6288 0.7930
No log 6.95 278 0.6037 0.3663 0.6037 0.7770
No log 7.0 280 0.6086 0.3237 0.6086 0.7801
No log 7.05 282 0.6105 0.3333 0.6105 0.7813
No log 7.1 284 0.5776 0.3575 0.5776 0.7600
No log 7.15 286 0.5716 0.4286 0.5716 0.7560
No log 7.2 288 0.5662 0.4286 0.5662 0.7525
No log 7.25 290 0.5619 0.4409 0.5619 0.7496
No log 7.3 292 0.5972 0.3769 0.5972 0.7728
No log 7.35 294 0.6115 0.2842 0.6115 0.7820
No log 7.4 296 0.5950 0.2842 0.5950 0.7714
No log 7.45 298 0.5875 0.2842 0.5875 0.7665
No log 7.5 300 0.5713 0.2787 0.5713 0.7558
No log 7.55 302 0.5760 0.3191 0.5760 0.7589
No log 7.6 304 0.5697 0.3508 0.5697 0.7548
No log 7.65 306 0.5926 0.3469 0.5926 0.7698
No log 7.7 308 0.6195 0.3498 0.6195 0.7871
No log 7.75 310 0.7211 0.1402 0.7211 0.8492
No log 7.8 312 0.8011 0.1786 0.8011 0.8950
No log 7.85 314 0.7613 0.1786 0.7613 0.8725
No log 7.9 316 0.6695 0.3744 0.6695 0.8182
No log 7.95 318 0.5838 0.3951 0.5838 0.7641
No log 8.0 320 0.5696 0.4227 0.5696 0.7547
No log 8.05 322 0.5699 0.3862 0.5699 0.7549
No log 8.1 324 0.5610 0.4573 0.5610 0.7490
No log 8.15 326 0.6040 0.3684 0.6040 0.7772
No log 8.2 328 0.6739 0.3143 0.6739 0.8209
No log 8.25 330 0.6748 0.3143 0.6748 0.8214
No log 8.3 332 0.6143 0.3684 0.6143 0.7838
No log 8.35 334 0.5688 0.4059 0.5688 0.7542
No log 8.4 336 0.5475 0.3769 0.5475 0.7399
No log 8.45 338 0.5468 0.4518 0.5468 0.7395
No log 8.5 340 0.5456 0.3575 0.5456 0.7387
No log 8.55 342 0.5553 0.4167 0.5553 0.7452
No log 8.6 344 0.5726 0.3769 0.5726 0.7567
No log 8.65 346 0.5933 0.3725 0.5933 0.7703
No log 8.7 348 0.5876 0.3725 0.5876 0.7665
No log 8.75 350 0.5748 0.3769 0.5748 0.7581
No log 8.8 352 0.5721 0.3769 0.5721 0.7564
No log 8.85 354 0.5769 0.3769 0.5769 0.7595
No log 8.9 356 0.5801 0.3769 0.5801 0.7616
No log 8.95 358 0.6031 0.3131 0.6031 0.7766
No log 9.0 360 0.6325 0.2780 0.6325 0.7953
No log 9.05 362 0.6687 0.2536 0.6687 0.8178
No log 9.1 364 0.6696 0.2830 0.6696 0.8183
No log 9.15 366 0.6557 0.3143 0.6557 0.8098
No log 9.2 368 0.6312 0.3143 0.6312 0.7945
No log 9.25 370 0.6030 0.3398 0.6030 0.7765
No log 9.3 372 0.5741 0.3725 0.5741 0.7577
No log 9.35 374 0.5634 0.3814 0.5634 0.7506
No log 9.4 376 0.5564 0.4112 0.5564 0.7459
No log 9.45 378 0.5545 0.4112 0.5545 0.7446
No log 9.5 380 0.5552 0.4112 0.5552 0.7451
No log 9.55 382 0.5572 0.4112 0.5572 0.7464
No log 9.6 384 0.5629 0.3769 0.5629 0.7502
No log 9.65 386 0.5687 0.3725 0.5687 0.7541
No log 9.7 388 0.5679 0.3725 0.5679 0.7536
No log 9.75 390 0.5700 0.3725 0.5700 0.7550
No log 9.8 392 0.5723 0.3725 0.5723 0.7565
No log 9.85 394 0.5752 0.3725 0.5752 0.7584
No log 9.9 396 0.5765 0.3725 0.5765 0.7593
No log 9.95 398 0.5767 0.3725 0.5767 0.7594
No log 10.0 400 0.5767 0.3725 0.5767 0.7594

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k9_task3_organization

Finetuned
(4023)
this model