ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5770
  • Qwk: 0.3725
  • Mse: 0.5770
  • Rmse: 0.7596

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.0480 0.0243 3.0480 1.7459
No log 0.1 4 2.0274 0.0 2.0274 1.4239
No log 0.15 6 1.3139 0.0255 1.3139 1.1462
No log 0.2 8 1.2496 -0.0327 1.2496 1.1178
No log 0.25 10 0.7973 0.0370 0.7973 0.8929
No log 0.3 12 0.7432 0.1765 0.7432 0.8621
No log 0.35 14 1.0186 0.0388 1.0186 1.0093
No log 0.4 16 1.1159 0.0388 1.1159 1.0564
No log 0.45 18 0.8029 0.1579 0.8029 0.8960
No log 0.5 20 0.6351 0.0815 0.6351 0.7969
No log 0.55 22 0.5943 -0.0081 0.5943 0.7709
No log 0.6 24 0.5801 0.0 0.5801 0.7616
No log 0.65 26 0.5615 0.0388 0.5615 0.7493
No log 0.7 28 1.2011 0.0817 1.2011 1.0959
No log 0.75 30 0.9884 0.1111 0.9884 0.9942
No log 0.8 32 0.5902 0.0256 0.5902 0.7683
No log 0.85 34 0.6365 0.0720 0.6365 0.7978
No log 0.9 36 0.6860 0.2000 0.6860 0.8283
No log 0.95 38 0.6061 0.0 0.6061 0.7785
No log 1.0 40 0.5664 0.0 0.5664 0.7526
No log 1.05 42 0.5758 0.0 0.5758 0.7588
No log 1.1 44 0.5739 0.0 0.5739 0.7576
No log 1.15 46 0.5298 0.0 0.5298 0.7278
No log 1.2 48 0.5289 0.0 0.5289 0.7273
No log 1.25 50 0.5288 0.0 0.5288 0.7272
No log 1.3 52 0.5855 0.0400 0.5855 0.7652
No log 1.35 54 0.6567 0.1913 0.6567 0.8104
No log 1.4 56 0.8137 0.1453 0.8137 0.9021
No log 1.45 58 0.6289 0.1590 0.6289 0.7930
No log 1.5 60 0.5590 0.1448 0.5590 0.7477
No log 1.55 62 0.6107 0.2000 0.6107 0.7815
No log 1.6 64 0.5629 0.1688 0.5629 0.7503
No log 1.65 66 0.5922 0.2370 0.5922 0.7695
No log 1.7 68 0.6366 0.3073 0.6366 0.7979
No log 1.75 70 0.6554 0.1813 0.6554 0.8096
No log 1.8 72 0.8733 0.1628 0.8733 0.9345
No log 1.85 74 0.5684 0.3086 0.5684 0.7539
No log 1.9 76 0.7413 0.2919 0.7413 0.8610
No log 1.95 78 0.6788 0.3402 0.6788 0.8239
No log 2.0 80 0.5691 0.2663 0.5691 0.7544
No log 2.05 82 0.6100 0.3548 0.6100 0.7810
No log 2.1 84 0.5704 0.2298 0.5704 0.7553
No log 2.15 86 0.6877 0.2184 0.6877 0.8293
No log 2.2 88 0.9163 0.1861 0.9163 0.9572
No log 2.25 90 0.7006 0.2093 0.7006 0.8370
No log 2.3 92 0.6530 0.1823 0.6530 0.8081
No log 2.35 94 0.6887 0.2251 0.6887 0.8299
No log 2.4 96 0.5915 0.1304 0.5915 0.7691
No log 2.45 98 0.7965 0.2233 0.7965 0.8925
No log 2.5 100 0.6219 0.2088 0.6219 0.7886
No log 2.55 102 0.7129 0.3103 0.7129 0.8443
No log 2.6 104 0.6158 0.1556 0.6158 0.7847
No log 2.65 106 1.1662 0.1367 1.1662 1.0799
No log 2.7 108 1.4432 0.1656 1.4432 1.2013
No log 2.75 110 0.8365 0.1776 0.8365 0.9146
No log 2.8 112 0.6859 0.2941 0.6859 0.8282
No log 2.85 114 0.6344 0.3016 0.6344 0.7965
No log 2.9 116 0.6163 0.2746 0.6163 0.7850
No log 2.95 118 0.8967 0.1261 0.8967 0.9470
No log 3.0 120 1.2317 0.1943 1.2317 1.1098
No log 3.05 122 0.7447 0.2239 0.7447 0.8629
No log 3.1 124 0.5993 0.2513 0.5993 0.7742
No log 3.15 126 0.5897 0.3182 0.5897 0.7680
No log 3.2 128 0.6284 0.2527 0.6284 0.7927
No log 3.25 130 0.6091 0.1617 0.6091 0.7804
No log 3.3 132 0.9069 0.1366 0.9069 0.9523
No log 3.35 134 1.0766 0.1008 1.0766 1.0376
No log 3.4 136 0.7742 0.2086 0.7742 0.8799
No log 3.45 138 0.8473 0.0884 0.8473 0.9205
No log 3.5 140 1.2256 0.1062 1.2256 1.1071
No log 3.55 142 0.9679 0.0847 0.9679 0.9838
No log 3.6 144 0.6911 0.1086 0.6911 0.8313
No log 3.65 146 0.6657 0.1807 0.6657 0.8159
No log 3.7 148 0.7596 0.1340 0.7596 0.8716
No log 3.75 150 0.6767 0.2432 0.6767 0.8226
No log 3.8 152 0.6437 0.2527 0.6437 0.8023
No log 3.85 154 0.7713 0.2727 0.7713 0.8783
No log 3.9 156 0.6352 0.2917 0.6352 0.7970
No log 3.95 158 0.6515 0.2893 0.6515 0.8072
No log 4.0 160 0.7233 0.2453 0.7233 0.8504
No log 4.05 162 0.8106 0.2222 0.8106 0.9003
No log 4.1 164 0.7828 0.2146 0.7828 0.8847
No log 4.15 166 0.7931 0.2222 0.7931 0.8906
No log 4.2 168 0.7123 0.3028 0.7123 0.8440
No log 4.25 170 0.6915 0.2315 0.6915 0.8316
No log 4.3 172 0.6546 0.2487 0.6546 0.8091
No log 4.35 174 0.6136 0.4019 0.6136 0.7833
No log 4.4 176 0.7302 0.2523 0.7302 0.8545
No log 4.45 178 0.6549 0.1675 0.6549 0.8093
No log 4.5 180 0.5597 0.3371 0.5597 0.7481
No log 4.55 182 0.5904 0.2967 0.5904 0.7684
No log 4.6 184 0.5336 0.3829 0.5336 0.7305
No log 4.65 186 0.8367 0.2281 0.8367 0.9147
No log 4.7 188 1.1086 0.2226 1.1086 1.0529
No log 4.75 190 0.8551 0.2203 0.8551 0.9247
No log 4.8 192 0.5892 0.3641 0.5892 0.7676
No log 4.85 194 0.5887 0.3831 0.5887 0.7673
No log 4.9 196 0.7272 0.2000 0.7272 0.8528
No log 4.95 198 1.1132 0.2109 1.1132 1.0551
No log 5.0 200 1.0424 0.1880 1.0424 1.0210
No log 5.05 202 0.7086 0.3301 0.7086 0.8418
No log 5.1 204 0.6538 0.3951 0.6538 0.8086
No log 5.15 206 0.6494 0.4074 0.6494 0.8059
No log 5.2 208 0.6771 0.2830 0.6771 0.8228
No log 5.25 210 0.9598 0.1181 0.9598 0.9797
No log 5.3 212 0.9317 0.1148 0.9317 0.9652
No log 5.35 214 0.6484 0.3684 0.6484 0.8052
No log 5.4 216 0.6019 0.4444 0.6019 0.7758
No log 5.45 218 0.7013 0.2727 0.7013 0.8375
No log 5.5 220 0.8873 0.1535 0.8873 0.9420
No log 5.55 222 1.0638 0.2174 1.0638 1.0314
No log 5.6 224 0.8325 0.1535 0.8325 0.9124
No log 5.65 226 0.5612 0.4343 0.5612 0.7491
No log 5.7 228 0.5844 0.3508 0.5844 0.7645
No log 5.75 230 0.5591 0.4518 0.5591 0.7477
No log 5.8 232 0.6701 0.3116 0.6701 0.8186
No log 5.85 234 1.0799 0.1888 1.0799 1.0392
No log 5.9 236 1.1351 0.2165 1.1351 1.0654
No log 5.95 238 0.8605 0.2469 0.8605 0.9277
No log 6.0 240 0.6298 0.4067 0.6298 0.7936
No log 6.05 242 0.6130 0.4010 0.6130 0.7829
No log 6.1 244 0.6884 0.2464 0.6884 0.8297
No log 6.15 246 0.7689 0.1416 0.7689 0.8769
No log 6.2 248 0.8926 0.2208 0.8926 0.9448
No log 6.25 250 0.7817 0.1790 0.7817 0.8842
No log 6.3 252 0.5944 0.3617 0.5944 0.7710
No log 6.35 254 0.5638 0.3591 0.5638 0.7509
No log 6.4 256 0.5815 0.3797 0.5815 0.7626
No log 6.45 258 0.6476 0.2593 0.6476 0.8047
No log 6.5 260 0.7705 0.1402 0.7705 0.8778
No log 6.55 262 0.8005 0.1429 0.8005 0.8947
No log 6.6 264 0.7251 0.2079 0.7251 0.8515
No log 6.65 266 0.6486 0.3103 0.6486 0.8053
No log 6.7 268 0.5899 0.4098 0.5899 0.7680
No log 6.75 270 0.5702 0.3708 0.5702 0.7551
No log 6.8 272 0.5673 0.4098 0.5673 0.7532
No log 6.85 274 0.5986 0.3061 0.5986 0.7737
No log 6.9 276 0.6107 0.3035 0.6107 0.7815
No log 6.95 278 0.5946 0.4 0.5946 0.7711
No log 7.0 280 0.6050 0.3237 0.6050 0.7778
No log 7.05 282 0.6146 0.3333 0.6146 0.7840
No log 7.1 284 0.5821 0.3641 0.5821 0.7630
No log 7.15 286 0.5712 0.4286 0.5712 0.7558
No log 7.2 288 0.5671 0.4409 0.5671 0.7531
No log 7.25 290 0.5629 0.4409 0.5629 0.7503
No log 7.3 292 0.5953 0.3769 0.5953 0.7715
No log 7.35 294 0.6077 0.3161 0.6077 0.7796
No log 7.4 296 0.5991 0.3161 0.5991 0.7740
No log 7.45 298 0.5910 0.2842 0.5910 0.7687
No log 7.5 300 0.5730 0.3508 0.5730 0.7570
No log 7.55 302 0.5764 0.3191 0.5764 0.7592
No log 7.6 304 0.5699 0.3508 0.5699 0.7549
No log 7.65 306 0.5961 0.3469 0.5961 0.7720
No log 7.7 308 0.6263 0.3744 0.6263 0.7914
No log 7.75 310 0.7291 0.1416 0.7291 0.8539
No log 7.8 312 0.7967 0.1786 0.7967 0.8926
No log 7.85 314 0.7489 0.2423 0.7489 0.8654
No log 7.9 316 0.6577 0.3744 0.6577 0.8110
No log 7.95 318 0.5781 0.4286 0.5781 0.7604
No log 8.0 320 0.5679 0.3862 0.5679 0.7536
No log 8.05 322 0.5668 0.3862 0.5668 0.7529
No log 8.1 324 0.5593 0.4694 0.5593 0.7479
No log 8.15 326 0.6089 0.3684 0.6089 0.7803
No log 8.2 328 0.6759 0.3143 0.6759 0.8221
No log 8.25 330 0.6709 0.3427 0.6709 0.8191
No log 8.3 332 0.6081 0.3684 0.6081 0.7798
No log 8.35 334 0.5649 0.4112 0.5649 0.7516
No log 8.4 336 0.5466 0.3769 0.5466 0.7393
No log 8.45 338 0.5457 0.4518 0.5457 0.7387
No log 8.5 340 0.5451 0.3684 0.5451 0.7383
No log 8.55 342 0.5574 0.4167 0.5574 0.7466
No log 8.6 344 0.5769 0.3725 0.5769 0.7595
No log 8.65 346 0.5980 0.3725 0.5980 0.7733
No log 8.7 348 0.5901 0.3725 0.5901 0.7682
No log 8.75 350 0.5757 0.3769 0.5757 0.7587
No log 8.8 352 0.5719 0.3769 0.5719 0.7562
No log 8.85 354 0.5762 0.3769 0.5762 0.7591
No log 8.9 356 0.5794 0.3769 0.5794 0.7612
No log 8.95 358 0.6024 0.3433 0.6024 0.7761
No log 9.0 360 0.6324 0.2780 0.6324 0.7952
No log 9.05 362 0.6697 0.2830 0.6697 0.8183
No log 9.1 364 0.6715 0.3488 0.6715 0.8194
No log 9.15 366 0.6574 0.3143 0.6574 0.8108
No log 9.2 368 0.6315 0.3143 0.6315 0.7947
No log 9.25 370 0.6023 0.3398 0.6023 0.7761
No log 9.3 372 0.5729 0.3725 0.5729 0.7569
No log 9.35 374 0.5623 0.3814 0.5623 0.7499
No log 9.4 376 0.5559 0.4112 0.5559 0.7456
No log 9.45 378 0.5542 0.4167 0.5542 0.7444
No log 9.5 380 0.5552 0.4112 0.5552 0.7451
No log 9.55 382 0.5573 0.4112 0.5573 0.7465
No log 9.6 384 0.5630 0.4010 0.5630 0.7503
No log 9.65 386 0.5690 0.3725 0.5690 0.7543
No log 9.7 388 0.5684 0.3725 0.5684 0.7539
No log 9.75 390 0.5706 0.3725 0.5706 0.7554
No log 9.8 392 0.5729 0.3725 0.5729 0.7569
No log 9.85 394 0.5757 0.3725 0.5757 0.7587
No log 9.9 396 0.5769 0.3725 0.5769 0.7596
No log 9.95 398 0.5771 0.3725 0.5771 0.7596
No log 10.0 400 0.5770 0.3725 0.5770 0.7596

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k9_task3_organization

Finetuned
(4023)
this model