ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k7_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8332
  • Qwk: 0.5023
  • Mse: 0.8332
  • Rmse: 0.9128

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.7240 0.0017 3.7240 1.9298
No log 0.1 4 1.6841 0.1188 1.6841 1.2977
No log 0.15 6 1.1225 0.0355 1.1225 1.0595
No log 0.2 8 0.9851 -0.0472 0.9851 0.9925
No log 0.25 10 0.7383 0.1647 0.7383 0.8593
No log 0.3 12 0.7414 0.1777 0.7414 0.8610
No log 0.35 14 0.7633 0.1352 0.7633 0.8736
No log 0.4 16 0.7846 0.0918 0.7846 0.8858
No log 0.45 18 0.7271 0.1101 0.7271 0.8527
No log 0.5 20 0.7451 0.2080 0.7451 0.8632
No log 0.55 22 0.9395 0.0701 0.9395 0.9693
No log 0.6 24 0.8387 0.1169 0.8387 0.9158
No log 0.65 26 0.6921 0.2376 0.6921 0.8319
No log 0.7 28 0.7440 0.0619 0.7440 0.8626
No log 0.75 30 0.8854 0.1171 0.8854 0.9409
No log 0.8 32 0.8453 0.1039 0.8453 0.9194
No log 0.85 34 0.7122 0.1786 0.7122 0.8439
No log 0.9 36 0.6421 0.2765 0.6421 0.8013
No log 0.95 38 0.6607 0.2909 0.6607 0.8128
No log 1.0 40 0.7203 0.2454 0.7203 0.8487
No log 1.05 42 0.7083 0.2819 0.7083 0.8416
No log 1.1 44 0.6891 0.2737 0.6891 0.8301
No log 1.15 46 0.6386 0.2863 0.6386 0.7991
No log 1.2 48 0.5999 0.3217 0.5999 0.7745
No log 1.25 50 0.5498 0.3634 0.5498 0.7415
No log 1.3 52 0.5566 0.4162 0.5566 0.7460
No log 1.35 54 0.6665 0.4703 0.6665 0.8164
No log 1.4 56 0.7665 0.3830 0.7665 0.8755
No log 1.45 58 0.7083 0.5019 0.7083 0.8416
No log 1.5 60 0.6157 0.5321 0.6157 0.7847
No log 1.55 62 0.6996 0.4994 0.6996 0.8364
No log 1.6 64 0.6748 0.5124 0.6748 0.8215
No log 1.65 66 0.6316 0.4997 0.6316 0.7947
No log 1.7 68 0.6633 0.4860 0.6633 0.8145
No log 1.75 70 0.7355 0.5003 0.7355 0.8576
No log 1.8 72 0.7153 0.4898 0.7153 0.8458
No log 1.85 74 0.6845 0.5072 0.6845 0.8273
No log 1.9 76 0.7245 0.5128 0.7245 0.8512
No log 1.95 78 0.6934 0.5246 0.6934 0.8327
No log 2.0 80 0.7200 0.5123 0.7200 0.8485
No log 2.05 82 0.7275 0.5501 0.7275 0.8529
No log 2.1 84 0.7416 0.5229 0.7416 0.8612
No log 2.15 86 0.7800 0.5408 0.7800 0.8832
No log 2.2 88 0.7791 0.5 0.7791 0.8827
No log 2.25 90 0.9256 0.4337 0.9256 0.9621
No log 2.3 92 1.2160 0.3761 1.2160 1.1027
No log 2.35 94 1.3492 0.3516 1.3492 1.1615
No log 2.4 96 1.1608 0.4005 1.1608 1.0774
No log 2.45 98 0.8837 0.4600 0.8837 0.9401
No log 2.5 100 0.7690 0.5387 0.7690 0.8769
No log 2.55 102 0.7621 0.5200 0.7621 0.8730
No log 2.6 104 0.7994 0.5253 0.7994 0.8941
No log 2.65 106 0.8025 0.5503 0.8025 0.8958
No log 2.7 108 0.8508 0.5024 0.8508 0.9224
No log 2.75 110 0.9329 0.4770 0.9329 0.9659
No log 2.8 112 1.0021 0.4355 1.0021 1.0010
No log 2.85 114 0.9260 0.4264 0.9260 0.9623
No log 2.9 116 0.7388 0.5527 0.7388 0.8595
No log 2.95 118 0.6083 0.5563 0.6083 0.7799
No log 3.0 120 0.5963 0.5554 0.5963 0.7722
No log 3.05 122 0.5949 0.5537 0.5949 0.7713
No log 3.1 124 0.6119 0.5548 0.6119 0.7822
No log 3.15 126 0.7646 0.5076 0.7646 0.8744
No log 3.2 128 1.0293 0.3950 1.0293 1.0145
No log 3.25 130 1.0720 0.3959 1.0720 1.0354
No log 3.3 132 0.8929 0.4755 0.8929 0.9449
No log 3.35 134 0.7705 0.5398 0.7705 0.8778
No log 3.4 136 0.7617 0.5610 0.7617 0.8728
No log 3.45 138 0.7556 0.5355 0.7556 0.8693
No log 3.5 140 0.7815 0.5304 0.7815 0.8840
No log 3.55 142 0.9004 0.4690 0.9004 0.9489
No log 3.6 144 0.8823 0.4901 0.8823 0.9393
No log 3.65 146 0.7197 0.5650 0.7197 0.8484
No log 3.7 148 0.6480 0.5473 0.6480 0.8050
No log 3.75 150 0.6421 0.6084 0.6421 0.8013
No log 3.8 152 0.6583 0.5946 0.6583 0.8114
No log 3.85 154 0.7211 0.5491 0.7211 0.8491
No log 3.9 156 0.9255 0.4566 0.9255 0.9620
No log 3.95 158 1.1967 0.4010 1.1967 1.0939
No log 4.0 160 1.2771 0.4233 1.2771 1.1301
No log 4.05 162 1.1234 0.4478 1.1234 1.0599
No log 4.1 164 0.9834 0.4891 0.9834 0.9916
No log 4.15 166 0.9571 0.5009 0.9571 0.9783
No log 4.2 168 1.0276 0.4546 1.0276 1.0137
No log 4.25 170 1.0605 0.4079 1.0605 1.0298
No log 4.3 172 1.0493 0.3855 1.0493 1.0244
No log 4.35 174 0.9572 0.4212 0.9572 0.9783
No log 4.4 176 0.9488 0.4211 0.9488 0.9741
No log 4.45 178 0.9587 0.4295 0.9587 0.9791
No log 4.5 180 0.8829 0.4682 0.8829 0.9396
No log 4.55 182 0.8207 0.4550 0.8207 0.9059
No log 4.6 184 0.8320 0.4490 0.8320 0.9121
No log 4.65 186 0.8375 0.4589 0.8375 0.9151
No log 4.7 188 0.8210 0.4833 0.8210 0.9061
No log 4.75 190 0.8014 0.5175 0.8014 0.8952
No log 4.8 192 0.8336 0.5307 0.8336 0.9130
No log 4.85 194 0.9024 0.4955 0.9024 0.9500
No log 4.9 196 1.0686 0.4850 1.0686 1.0337
No log 4.95 198 1.2286 0.4787 1.2286 1.1084
No log 5.0 200 1.1859 0.4610 1.1859 1.0890
No log 5.05 202 1.0040 0.4766 1.0040 1.0020
No log 5.1 204 0.8418 0.5024 0.8418 0.9175
No log 5.15 206 0.8202 0.5017 0.8202 0.9056
No log 5.2 208 0.8807 0.4835 0.8807 0.9384
No log 5.25 210 1.0987 0.4590 1.0987 1.0482
No log 5.3 212 1.2131 0.4373 1.2131 1.1014
No log 5.35 214 1.1158 0.4464 1.1158 1.0563
No log 5.4 216 1.0358 0.4549 1.0358 1.0177
No log 5.45 218 0.9242 0.4572 0.9242 0.9614
No log 5.5 220 0.9459 0.4617 0.9459 0.9726
No log 5.55 222 1.0418 0.4247 1.0418 1.0207
No log 5.6 224 1.1932 0.4319 1.1932 1.0923
No log 5.65 226 1.2772 0.4218 1.2772 1.1302
No log 5.7 228 1.1590 0.4474 1.1590 1.0766
No log 5.75 230 0.9904 0.4768 0.9904 0.9952
No log 5.8 232 0.9025 0.4598 0.9025 0.9500
No log 5.85 234 0.8181 0.4842 0.8181 0.9045
No log 5.9 236 0.7898 0.4847 0.7898 0.8887
No log 5.95 238 0.8050 0.4864 0.8050 0.8972
No log 6.0 240 0.9233 0.4754 0.9233 0.9609
No log 6.05 242 0.9782 0.4550 0.9782 0.9890
No log 6.1 244 0.9271 0.4445 0.9271 0.9629
No log 6.15 246 0.8617 0.4636 0.8617 0.9283
No log 6.2 248 0.8182 0.4665 0.8182 0.9045
No log 6.25 250 0.8149 0.4665 0.8149 0.9027
No log 6.3 252 0.8583 0.4759 0.8583 0.9264
No log 6.35 254 0.8604 0.4953 0.8604 0.9276
No log 6.4 256 0.8336 0.5053 0.8336 0.9130
No log 6.45 258 0.7930 0.4930 0.7930 0.8905
No log 6.5 260 0.7546 0.5094 0.7546 0.8687
No log 6.55 262 0.7374 0.4899 0.7374 0.8587
No log 6.6 264 0.7736 0.4826 0.7736 0.8796
No log 6.65 266 0.7938 0.4986 0.7938 0.8910
No log 6.7 268 0.8244 0.5128 0.8244 0.9079
No log 6.75 270 0.8542 0.4839 0.8542 0.9242
No log 6.8 272 0.8509 0.4472 0.8509 0.9224
No log 6.85 274 0.8613 0.4334 0.8613 0.9281
No log 6.9 276 0.8508 0.4194 0.8508 0.9224
No log 6.95 278 0.8265 0.4194 0.8265 0.9091
No log 7.0 280 0.7713 0.4193 0.7713 0.8782
No log 7.05 282 0.7259 0.4779 0.7259 0.8520
No log 7.1 284 0.7180 0.4962 0.7180 0.8474
No log 7.15 286 0.7501 0.5042 0.7501 0.8661
No log 7.2 288 0.7927 0.4915 0.7927 0.8903
No log 7.25 290 0.8827 0.4453 0.8827 0.9395
No log 7.3 292 1.0001 0.4280 1.0001 1.0000
No log 7.35 294 1.0248 0.4432 1.0248 1.0123
No log 7.4 296 1.0016 0.4434 1.0016 1.0008
No log 7.45 298 0.9511 0.4634 0.9511 0.9752
No log 7.5 300 0.8590 0.4552 0.8590 0.9268
No log 7.55 302 0.7918 0.4462 0.7918 0.8898
No log 7.6 304 0.7919 0.4462 0.7919 0.8899
No log 7.65 306 0.8357 0.4803 0.8357 0.9142
No log 7.7 308 0.9176 0.4740 0.9176 0.9579
No log 7.75 310 0.9752 0.4434 0.9752 0.9875
No log 7.8 312 1.0507 0.4437 1.0507 1.0250
No log 7.85 314 1.0483 0.4381 1.0483 1.0238
No log 7.9 316 1.0100 0.4549 1.0100 1.0050
No log 7.95 318 0.9673 0.4572 0.9673 0.9835
No log 8.0 320 0.9489 0.4637 0.9489 0.9741
No log 8.05 322 0.9831 0.4581 0.9831 0.9915
No log 8.1 324 1.0580 0.4581 1.0580 1.0286
No log 8.15 326 1.0964 0.4507 1.0964 1.0471
No log 8.2 328 1.0679 0.4479 1.0679 1.0334
No log 8.25 330 0.9913 0.4581 0.9913 0.9956
No log 8.3 332 0.9281 0.4641 0.9281 0.9634
No log 8.35 334 0.8678 0.4878 0.8678 0.9315
No log 8.4 336 0.8258 0.4910 0.8258 0.9087
No log 8.45 338 0.8062 0.4864 0.8062 0.8979
No log 8.5 340 0.7963 0.4864 0.7963 0.8923
No log 8.55 342 0.7982 0.4864 0.7982 0.8934
No log 8.6 344 0.8178 0.4838 0.8178 0.9043
No log 8.65 346 0.8551 0.4917 0.8551 0.9247
No log 8.7 348 0.8939 0.4843 0.8939 0.9455
No log 8.75 350 0.9131 0.471 0.9131 0.9556
No log 8.8 352 0.8998 0.4781 0.8998 0.9486
No log 8.85 354 0.8673 0.4917 0.8673 0.9313
No log 8.9 356 0.8279 0.4773 0.8279 0.9099
No log 8.95 358 0.8137 0.4901 0.8137 0.9020
No log 9.0 360 0.8016 0.4606 0.8016 0.8953
No log 9.05 362 0.7963 0.4606 0.7963 0.8923
No log 9.1 364 0.7945 0.4606 0.7945 0.8913
No log 9.15 366 0.8030 0.4838 0.8030 0.8961
No log 9.2 368 0.8225 0.4621 0.8225 0.9069
No log 9.25 370 0.8437 0.4792 0.8437 0.9185
No log 9.3 372 0.8558 0.4792 0.8558 0.9251
No log 9.35 374 0.8526 0.4682 0.8526 0.9234
No log 9.4 376 0.8451 0.4682 0.8451 0.9193
No log 9.45 378 0.8432 0.4813 0.8432 0.9183
No log 9.5 380 0.8440 0.4813 0.8440 0.9187
No log 9.55 382 0.8435 0.4813 0.8435 0.9184
No log 9.6 384 0.8378 0.5074 0.8378 0.9153
No log 9.65 386 0.8328 0.5023 0.8328 0.9126
No log 9.7 388 0.8289 0.4913 0.8289 0.9104
No log 9.75 390 0.8287 0.4913 0.8287 0.9103
No log 9.8 392 0.8270 0.4913 0.8270 0.9094
No log 9.85 394 0.8284 0.4913 0.8284 0.9102
No log 9.9 396 0.8310 0.5023 0.8310 0.9116
No log 9.95 398 0.8327 0.5023 0.8327 0.9125
No log 10.0 400 0.8332 0.5023 0.8332 0.9128

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k7_task2_organization

Finetuned
(4019)
this model