ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k6_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6113
  • Qwk: 0.4027
  • Mse: 0.6113
  • Rmse: 0.7819

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 3.3307 0.0026 3.3307 1.8250
No log 0.1429 4 1.6942 -0.0070 1.6942 1.3016
No log 0.2143 6 0.8104 0.1169 0.8104 0.9002
No log 0.2857 8 0.6546 0.2749 0.6546 0.8091
No log 0.3571 10 0.5471 0.0638 0.5471 0.7396
No log 0.4286 12 0.5640 0.0569 0.5640 0.7510
No log 0.5 14 0.5255 0.0 0.5255 0.7249
No log 0.5714 16 0.5531 0.0 0.5531 0.7437
No log 0.6429 18 0.5623 0.0 0.5623 0.7499
No log 0.7143 20 0.5242 0.0569 0.5242 0.7240
No log 0.7857 22 0.5452 0.3333 0.5452 0.7383
No log 0.8571 24 0.6195 0.25 0.6195 0.7871
No log 0.9286 26 0.5119 0.2941 0.5119 0.7155
No log 1.0 28 0.6617 0.2000 0.6617 0.8135
No log 1.0714 30 0.8820 0.2000 0.8820 0.9391
No log 1.1429 32 0.8419 0.0210 0.8419 0.9175
No log 1.2143 34 0.7764 0.0720 0.7764 0.8811
No log 1.2857 36 0.6149 0.0 0.6149 0.7841
No log 1.3571 38 0.5197 0.0 0.5197 0.7209
No log 1.4286 40 0.5628 0.3475 0.5628 0.7502
No log 1.5 42 0.5616 0.3043 0.5616 0.7494
No log 1.5714 44 0.5103 0.0 0.5103 0.7143
No log 1.6429 46 0.5016 0.0 0.5016 0.7082
No log 1.7143 48 0.5838 0.0720 0.5838 0.7640
No log 1.7857 50 0.5222 0.1278 0.5222 0.7226
No log 1.8571 52 0.7219 0.2464 0.7219 0.8497
No log 1.9286 54 2.0526 0.0239 2.0526 1.4327
No log 2.0 56 2.1196 0.0649 2.1196 1.4559
No log 2.0714 58 1.1593 0.0929 1.1593 1.0767
No log 2.1429 60 0.7751 0.1644 0.7751 0.8804
No log 2.2143 62 0.4934 0.1429 0.4934 0.7024
No log 2.2857 64 0.5960 0.2533 0.5960 0.7720
No log 2.3571 66 0.6080 0.3032 0.6080 0.7797
No log 2.4286 68 0.5155 0.0986 0.5155 0.7180
No log 2.5 70 0.6164 0.2410 0.6164 0.7851
No log 2.5714 72 0.7871 0.2300 0.7871 0.8872
No log 2.6429 74 0.7138 0.1919 0.7138 0.8448
No log 2.7143 76 0.5540 0.2105 0.5540 0.7443
No log 2.7857 78 0.5761 0.2704 0.5761 0.7590
No log 2.8571 80 0.7009 0.2621 0.7009 0.8372
No log 2.9286 82 0.6112 0.3563 0.6112 0.7818
No log 3.0 84 0.5519 0.1141 0.5519 0.7429
No log 3.0714 86 0.7834 0.2676 0.7834 0.8851
No log 3.1429 88 0.7942 0.2579 0.7942 0.8912
No log 3.2143 90 0.5650 0.2444 0.5650 0.7517
No log 3.2857 92 0.5593 0.3295 0.5593 0.7479
No log 3.3571 94 0.5624 0.2444 0.5624 0.7499
No log 3.4286 96 0.5556 0.2688 0.5556 0.7454
No log 3.5 98 0.5363 0.3446 0.5363 0.7323
No log 3.5714 100 0.5019 0.3208 0.5019 0.7085
No log 3.6429 102 0.6094 0.2871 0.6094 0.7806
No log 3.7143 104 0.8255 0.1867 0.8255 0.9086
No log 3.7857 106 0.6381 0.4286 0.6381 0.7988
No log 3.8571 108 0.5761 0.3607 0.5761 0.7590
No log 3.9286 110 0.5953 0.3607 0.5953 0.7716
No log 4.0 112 0.6619 0.4338 0.6619 0.8136
No log 4.0714 114 0.8145 0.2253 0.8145 0.9025
No log 4.1429 116 0.8520 0.2253 0.8520 0.9231
No log 4.2143 118 0.5986 0.3769 0.5986 0.7737
No log 4.2857 120 0.6269 0.3702 0.6269 0.7918
No log 4.3571 122 0.5947 0.4396 0.5947 0.7711
No log 4.4286 124 0.5843 0.3874 0.5843 0.7644
No log 4.5 126 0.5908 0.4851 0.5908 0.7687
No log 4.5714 128 0.5712 0.4518 0.5712 0.7558
No log 4.6429 130 0.6753 0.3593 0.6753 0.8218
No log 4.7143 132 0.6735 0.3761 0.6735 0.8207
No log 4.7857 134 0.5934 0.4882 0.5934 0.7703
No log 4.8571 136 0.5821 0.4882 0.5821 0.7630
No log 4.9286 138 0.5613 0.5102 0.5613 0.7492
No log 5.0 140 0.5776 0.5074 0.5776 0.7600
No log 5.0714 142 0.7155 0.25 0.7155 0.8459
No log 5.1429 144 0.6373 0.4286 0.6373 0.7983
No log 5.2143 146 0.6027 0.5 0.6027 0.7763
No log 5.2857 148 0.7826 0.2441 0.7826 0.8846
No log 5.3571 150 0.8473 0.25 0.8473 0.9205
No log 5.4286 152 0.7450 0.3414 0.7450 0.8631
No log 5.5 154 0.6618 0.4386 0.6618 0.8135
No log 5.5714 156 0.5541 0.4545 0.5541 0.7444
No log 5.6429 158 0.6759 0.4237 0.6759 0.8221
No log 5.7143 160 0.7898 0.3588 0.7898 0.8887
No log 5.7857 162 0.7115 0.3548 0.7115 0.8435
No log 5.8571 164 0.8465 0.3030 0.8465 0.9200
No log 5.9286 166 1.0238 0.1888 1.0238 1.0118
No log 6.0 168 0.8432 0.3359 0.8432 0.9182
No log 6.0714 170 0.5610 0.4286 0.5610 0.7490
No log 6.1429 172 0.5462 0.4455 0.5462 0.7390
No log 6.2143 174 0.6685 0.4087 0.6685 0.8176
No log 6.2857 176 0.9246 0.1884 0.9246 0.9616
No log 6.3571 178 0.8901 0.1882 0.8901 0.9434
No log 6.4286 180 0.7435 0.3360 0.7435 0.8623
No log 6.5 182 0.5322 0.4924 0.5322 0.7295
No log 6.5714 184 0.5110 0.4400 0.5110 0.7149
No log 6.6429 186 0.5250 0.5025 0.5250 0.7246
No log 6.7143 188 0.6396 0.3722 0.6396 0.7998
No log 6.7857 190 0.8359 0.2450 0.8359 0.9143
No log 6.8571 192 0.9146 0.1524 0.9146 0.9564
No log 6.9286 194 0.7468 0.3021 0.7468 0.8642
No log 7.0 196 0.5876 0.4234 0.5876 0.7665
No log 7.0714 198 0.5308 0.4627 0.5308 0.7286
No log 7.1429 200 0.5464 0.5122 0.5464 0.7392
No log 7.2143 202 0.5914 0.4286 0.5914 0.7690
No log 7.2857 204 0.6458 0.3982 0.6458 0.8036
No log 7.3571 206 0.5774 0.4783 0.5774 0.7598
No log 7.4286 208 0.5762 0.4783 0.5762 0.7591
No log 7.5 210 0.6081 0.3929 0.6081 0.7798
No log 7.5714 212 0.6406 0.3665 0.6406 0.8004
No log 7.6429 214 0.7121 0.3778 0.7121 0.8439
No log 7.7143 216 0.7080 0.4087 0.7080 0.8414
No log 7.7857 218 0.8176 0.2424 0.8176 0.9042
No log 7.8571 220 0.8084 0.2756 0.8084 0.8991
No log 7.9286 222 0.6650 0.3761 0.6650 0.8155
No log 8.0 224 0.5281 0.4233 0.5281 0.7267
No log 8.0714 226 0.5096 0.4732 0.5096 0.7139
No log 8.1429 228 0.5328 0.4233 0.5328 0.7299
No log 8.2143 230 0.5714 0.4234 0.5714 0.7559
No log 8.2857 232 0.6407 0.4027 0.6407 0.8004
No log 8.3571 234 0.6780 0.4237 0.6780 0.8234
No log 8.4286 236 0.6628 0.4027 0.6628 0.8141
No log 8.5 238 0.5750 0.4393 0.5750 0.7583
No log 8.5714 240 0.5297 0.4233 0.5297 0.7278
No log 8.6429 242 0.5039 0.5025 0.5039 0.7098
No log 8.7143 244 0.5042 0.4059 0.5042 0.7101
No log 8.7857 246 0.5062 0.4171 0.5062 0.7115
No log 8.8571 248 0.5078 0.5025 0.5078 0.7126
No log 8.9286 250 0.5264 0.5330 0.5264 0.7255
No log 9.0 252 0.5616 0.4340 0.5616 0.7494
No log 9.0714 254 0.6030 0.4185 0.6030 0.7766
No log 9.1429 256 0.6198 0.4027 0.6198 0.7873
No log 9.2143 258 0.6393 0.4027 0.6393 0.7996
No log 9.2857 260 0.6228 0.4027 0.6228 0.7892
No log 9.3571 262 0.6089 0.4027 0.6089 0.7803
No log 9.4286 264 0.6207 0.4027 0.6207 0.7878
No log 9.5 266 0.6218 0.4027 0.6218 0.7885
No log 9.5714 268 0.6290 0.4027 0.6290 0.7931
No log 9.6429 270 0.6195 0.4027 0.6195 0.7871
No log 9.7143 272 0.6016 0.4286 0.6016 0.7756
No log 9.7857 274 0.5959 0.4286 0.5959 0.7720
No log 9.8571 276 0.6002 0.4286 0.6002 0.7747
No log 9.9286 278 0.6069 0.4027 0.6069 0.7790
No log 10.0 280 0.6113 0.4027 0.6113 0.7819

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k6_task3_organization

Finetuned
(4023)
this model