ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k8_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5009
  • Qwk: 0.5459
  • Mse: 0.5009
  • Rmse: 0.7077

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 3.3864 -0.0149 3.3864 1.8402
No log 0.0952 4 1.9182 -0.0070 1.9182 1.3850
No log 0.1429 6 1.0033 0.0 1.0033 1.0016
No log 0.1905 8 0.7451 0.2077 0.7451 0.8632
No log 0.2381 10 0.7688 0.2222 0.7688 0.8768
No log 0.2857 12 0.5747 0.1895 0.5747 0.7581
No log 0.3333 14 0.6479 0.3898 0.6479 0.8049
No log 0.3810 16 0.6476 0.3778 0.6476 0.8047
No log 0.4286 18 0.6928 0.3118 0.6928 0.8324
No log 0.4762 20 0.6042 0.3086 0.6042 0.7773
No log 0.5238 22 0.5718 0.1206 0.5718 0.7562
No log 0.5714 24 0.6035 0.1895 0.6035 0.7769
No log 0.6190 26 0.6356 0.1905 0.6356 0.7972
No log 0.6667 28 0.7227 0.2169 0.7227 0.8501
No log 0.7143 30 0.6249 0.2281 0.6249 0.7905
No log 0.7619 32 0.6317 0.2542 0.6317 0.7948
No log 0.8095 34 0.7780 0.1691 0.7780 0.8821
No log 0.8571 36 0.6317 0.2967 0.6317 0.7948
No log 0.9048 38 0.5995 0.4033 0.5995 0.7743
No log 0.9524 40 0.6789 0.2577 0.6789 0.8239
No log 1.0 42 0.7099 0.2577 0.7099 0.8426
No log 1.0476 44 0.7268 0.2536 0.7268 0.8525
No log 1.0952 46 0.6305 0.2090 0.6305 0.7940
No log 1.1429 48 0.9944 0.0938 0.9944 0.9972
No log 1.1905 50 1.0930 0.0977 1.0930 1.0455
No log 1.2381 52 0.6766 0.2593 0.6766 0.8226
No log 1.2857 54 1.1769 0.1839 1.1769 1.0849
No log 1.3333 56 1.0270 0.1746 1.0270 1.0134
No log 1.3810 58 0.6120 0.36 0.6120 0.7823
No log 1.4286 60 0.5889 0.375 0.5889 0.7674
No log 1.4762 62 0.5608 0.4043 0.5608 0.7489
No log 1.5238 64 0.5740 0.3369 0.5740 0.7576
No log 1.5714 66 0.5844 0.3439 0.5844 0.7645
No log 1.6190 68 0.5850 0.3191 0.5850 0.7648
No log 1.6667 70 0.8016 0.2212 0.8016 0.8953
No log 1.7143 72 0.8578 0.2000 0.8578 0.9262
No log 1.7619 74 0.5914 0.3978 0.5914 0.7690
No log 1.8095 76 0.7632 0.2332 0.7632 0.8736
No log 1.8571 78 0.8214 0.2780 0.8214 0.9063
No log 1.9048 80 0.5531 0.1329 0.5531 0.7437
No log 1.9524 82 0.6147 0.3439 0.6147 0.7840
No log 2.0 84 0.7900 0.2000 0.7900 0.8888
No log 2.0476 86 0.7191 0.2432 0.7191 0.8480
No log 2.0952 88 0.4931 0.4947 0.4931 0.7022
No log 2.1429 90 0.4743 0.3913 0.4743 0.6887
No log 2.1905 92 1.0849 0.2715 1.0849 1.0416
No log 2.2381 94 0.8659 0.3125 0.8659 0.9306
No log 2.2857 96 0.5085 0.3769 0.5085 0.7131
No log 2.3333 98 0.5795 0.3514 0.5795 0.7612
No log 2.3810 100 0.5326 0.3398 0.5326 0.7298
No log 2.4286 102 0.5650 0.2821 0.5650 0.7517
No log 2.4762 104 0.9988 0.2688 0.9988 0.9994
No log 2.5238 106 1.0137 0.2892 1.0137 1.0068
No log 2.5714 108 0.7450 0.3226 0.7450 0.8631
No log 2.6190 110 0.5898 0.3973 0.5898 0.7680
No log 2.6667 112 0.6494 0.4182 0.6494 0.8059
No log 2.7143 114 0.5594 0.2692 0.5594 0.7479
No log 2.7619 116 0.6799 0.2323 0.6799 0.8245
No log 2.8095 118 1.0080 0.2243 1.0080 1.0040
No log 2.8571 120 0.6587 0.1919 0.6587 0.8116
No log 2.9048 122 0.4809 0.4732 0.4809 0.6934
No log 2.9524 124 0.4974 0.4518 0.4974 0.7053
No log 3.0 126 0.4788 0.4229 0.4788 0.6920
No log 3.0476 128 0.5097 0.4286 0.5097 0.7139
No log 3.0952 130 0.5177 0.4112 0.5177 0.7195
No log 3.1429 132 0.4895 0.4563 0.4895 0.6997
No log 3.1905 134 0.4998 0.4455 0.4998 0.7069
No log 3.2381 136 0.7176 0.2146 0.7176 0.8471
No log 3.2857 138 0.8351 0.2450 0.8351 0.9138
No log 3.3333 140 0.9100 0.2432 0.9100 0.9539
No log 3.3810 142 0.8441 0.2996 0.8441 0.9188
No log 3.4286 144 0.6286 0.3651 0.6286 0.7929
No log 3.4762 146 0.5858 0.4182 0.5858 0.7654
No log 3.5238 148 0.5713 0.5169 0.5713 0.7558
No log 3.5714 150 0.7011 0.3214 0.7011 0.8373
No log 3.6190 152 0.6917 0.3571 0.6917 0.8317
No log 3.6667 154 0.5358 0.3333 0.5358 0.7320
No log 3.7143 156 0.8883 0.1931 0.8883 0.9425
No log 3.7619 158 1.1318 0.0815 1.1318 1.0639
No log 3.8095 160 0.7639 0.2212 0.7639 0.8740
No log 3.8571 162 0.5377 0.3231 0.5377 0.7333
No log 3.9048 164 0.5474 0.3367 0.5474 0.7399
No log 3.9524 166 0.5562 0.4105 0.5562 0.7458
No log 4.0 168 0.5801 0.4167 0.5801 0.7616
No log 4.0476 170 0.5743 0.4694 0.5743 0.7578
No log 4.0952 172 0.5966 0.3200 0.5966 0.7724
No log 4.1429 174 0.5927 0.3641 0.5927 0.7699
No log 4.1905 176 0.7286 0.2000 0.7286 0.8536
No log 4.2381 178 0.8087 0.2143 0.8087 0.8993
No log 4.2857 180 0.8872 0.2137 0.8872 0.9419
No log 4.3333 182 0.6610 0.2842 0.6610 0.8130
No log 4.3810 184 0.6314 0.4234 0.6314 0.7946
No log 4.4286 186 0.6172 0.4234 0.6172 0.7856
No log 4.4762 188 0.6223 0.2842 0.6223 0.7889
No log 4.5238 190 0.8516 0.2143 0.8516 0.9228
No log 4.5714 192 0.7381 0.2146 0.7381 0.8592
No log 4.6190 194 0.5837 0.36 0.5837 0.7640
No log 4.6667 196 0.5944 0.3299 0.5944 0.7710
No log 4.7143 198 0.5978 0.3367 0.5978 0.7732
No log 4.7619 200 0.5813 0.3398 0.5813 0.7624
No log 4.8095 202 0.6844 0.2563 0.6844 0.8273
No log 4.8571 204 0.8157 0.2140 0.8157 0.9031
No log 4.9048 206 0.6575 0.2549 0.6575 0.8108
No log 4.9524 208 0.4919 0.4680 0.4919 0.7013
No log 5.0 210 0.5277 0.5068 0.5277 0.7264
No log 5.0476 212 0.5001 0.4341 0.5001 0.7072
No log 5.0952 214 0.4941 0.4413 0.4941 0.7029
No log 5.1429 216 0.5680 0.3118 0.5680 0.7537
No log 5.1905 218 0.5092 0.4286 0.5092 0.7136
No log 5.2381 220 0.4783 0.4348 0.4783 0.6916
No log 5.2857 222 0.4948 0.375 0.4948 0.7034
No log 5.3333 224 0.4860 0.4627 0.4860 0.6971
No log 5.3810 226 0.4897 0.4105 0.4897 0.6998
No log 5.4286 228 0.4840 0.4468 0.4840 0.6957
No log 5.4762 230 0.5038 0.3927 0.5038 0.7098
No log 5.5238 232 0.4660 0.4468 0.4660 0.6827
No log 5.5714 234 0.5111 0.4450 0.5111 0.7149
No log 5.6190 236 0.5359 0.4171 0.5359 0.7320
No log 5.6667 238 0.4600 0.6 0.4600 0.6782
No log 5.7143 240 0.4935 0.3333 0.4935 0.7025
No log 5.7619 242 0.5532 0.3439 0.5532 0.7438
No log 5.8095 244 0.5450 0.3299 0.5450 0.7383
No log 5.8571 246 0.4689 0.52 0.4689 0.6848
No log 5.9048 248 0.5079 0.4680 0.5079 0.7127
No log 5.9524 250 0.6265 0.3929 0.6265 0.7915
No log 6.0 252 0.6251 0.3929 0.6251 0.7906
No log 6.0476 254 0.5179 0.48 0.5179 0.7197
No log 6.0952 256 0.4678 0.5862 0.4678 0.6840
No log 6.1429 258 0.4856 0.4595 0.4856 0.6968
No log 6.1905 260 0.4671 0.6 0.4671 0.6834
No log 6.2381 262 0.5270 0.4343 0.5270 0.7259
No log 6.2857 264 0.5272 0.3990 0.5272 0.7261
No log 6.3333 266 0.4739 0.5468 0.4739 0.6884
No log 6.3810 268 0.4734 0.5468 0.4734 0.6880
No log 6.4286 270 0.4828 0.4924 0.4828 0.6949
No log 6.4762 272 0.5298 0.3814 0.5298 0.7279
No log 6.5238 274 0.5028 0.3939 0.5028 0.7091
No log 6.5714 276 0.4934 0.4694 0.4934 0.7024
No log 6.6190 278 0.5195 0.4400 0.5195 0.7208
No log 6.6667 280 0.5064 0.4518 0.5064 0.7116
No log 6.7143 282 0.4833 0.4409 0.4833 0.6952
No log 6.7619 284 0.5024 0.4545 0.5024 0.7088
No log 6.8095 286 0.4932 0.4413 0.4932 0.7023
No log 6.8571 288 0.4790 0.4595 0.4790 0.6921
No log 6.9048 290 0.5273 0.3333 0.5273 0.7262
No log 6.9524 292 0.5503 0.3224 0.5503 0.7418
No log 7.0 294 0.5261 0.3446 0.5261 0.7253
No log 7.0476 296 0.4971 0.3966 0.4971 0.7051
No log 7.0952 298 0.4771 0.4033 0.4771 0.6907
No log 7.1429 300 0.4783 0.4894 0.4783 0.6916
No log 7.1905 302 0.4816 0.3966 0.4816 0.6940
No log 7.2381 304 0.5055 0.4152 0.5055 0.7110
No log 7.2857 306 0.5531 0.2889 0.5531 0.7437
No log 7.3333 308 0.5329 0.2889 0.5329 0.7300
No log 7.3810 310 0.5029 0.5152 0.5029 0.7091
No log 7.4286 312 0.5001 0.3966 0.5001 0.7072
No log 7.4762 314 0.5268 0.3829 0.5268 0.7258
No log 7.5238 316 0.5201 0.4091 0.5201 0.7212
No log 7.5714 318 0.5046 0.4468 0.5046 0.7104
No log 7.6190 320 0.5269 0.4639 0.5269 0.7259
No log 7.6667 322 0.5395 0.4231 0.5395 0.7345
No log 7.7143 324 0.5139 0.4694 0.5139 0.7169
No log 7.7619 326 0.5107 0.4526 0.5107 0.7146
No log 7.8095 328 0.5241 0.4033 0.5241 0.7240
No log 7.8571 330 0.5342 0.4033 0.5342 0.7309
No log 7.9048 332 0.5189 0.4348 0.5189 0.7204
No log 7.9524 334 0.5149 0.5468 0.5149 0.7176
No log 8.0 336 0.5479 0.3892 0.5479 0.7402
No log 8.0476 338 0.5528 0.3892 0.5528 0.7435
No log 8.0952 340 0.5493 0.4 0.5493 0.7412
No log 8.1429 342 0.5272 0.4955 0.5272 0.7261
No log 8.1905 344 0.5479 0.4033 0.5479 0.7402
No log 8.2381 346 0.6348 0.2563 0.6348 0.7968
No log 8.2857 348 0.6662 0.2536 0.6662 0.8162
No log 8.3333 350 0.6264 0.2990 0.6264 0.7914
No log 8.3810 352 0.5644 0.3617 0.5644 0.7512
No log 8.4286 354 0.5359 0.4975 0.5359 0.7321
No log 8.4762 356 0.5633 0.3398 0.5633 0.7505
No log 8.5238 358 0.5968 0.3237 0.5968 0.7725
No log 8.5714 360 0.5892 0.3333 0.5892 0.7676
No log 8.6190 362 0.5549 0.3892 0.5549 0.7449
No log 8.6667 364 0.5248 0.4033 0.5248 0.7244
No log 8.7143 366 0.5219 0.3708 0.5219 0.7224
No log 8.7619 368 0.5285 0.3708 0.5285 0.7270
No log 8.8095 370 0.5329 0.3708 0.5329 0.7300
No log 8.8571 372 0.5299 0.3708 0.5299 0.7280
No log 8.9048 374 0.5288 0.3708 0.5288 0.7272
No log 8.9524 376 0.5129 0.3708 0.5129 0.7162
No log 9.0 378 0.4988 0.4652 0.4988 0.7063
No log 9.0476 380 0.4968 0.4639 0.4968 0.7048
No log 9.0952 382 0.4953 0.4652 0.4953 0.7038
No log 9.1429 384 0.4952 0.4819 0.4952 0.7037
No log 9.1905 386 0.4920 0.5330 0.4920 0.7014
No log 9.2381 388 0.4920 0.4725 0.4920 0.7014
No log 9.2857 390 0.4939 0.4348 0.4939 0.7028
No log 9.3333 392 0.4987 0.4348 0.4987 0.7062
No log 9.3810 394 0.5082 0.3708 0.5082 0.7129
No log 9.4286 396 0.5181 0.3708 0.5181 0.7198
No log 9.4762 398 0.5209 0.3708 0.5209 0.7218
No log 9.5238 400 0.5218 0.3708 0.5218 0.7223
No log 9.5714 402 0.5165 0.4033 0.5165 0.7187
No log 9.6190 404 0.5087 0.4033 0.5087 0.7133
No log 9.6667 406 0.5037 0.4652 0.5037 0.7097
No log 9.7143 408 0.5013 0.4725 0.5013 0.7080
No log 9.7619 410 0.5006 0.4725 0.5006 0.7075
No log 9.8095 412 0.5006 0.5459 0.5006 0.7075
No log 9.8571 414 0.5007 0.5459 0.5007 0.7076
No log 9.9048 416 0.5008 0.5459 0.5008 0.7077
No log 9.9524 418 0.5009 0.5459 0.5009 0.7077
No log 10.0 420 0.5009 0.5459 0.5009 0.7077

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k8_task3_organization

Finetuned
(4023)
this model