ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k6_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6096
  • Qwk: 0.4027
  • Mse: 0.6096
  • Rmse: 0.7807

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 3.3307 0.0026 3.3307 1.8250
No log 0.1429 4 1.6942 -0.0070 1.6942 1.3016
No log 0.2143 6 0.8104 0.1169 0.8104 0.9002
No log 0.2857 8 0.6546 0.2749 0.6546 0.8091
No log 0.3571 10 0.5471 0.0638 0.5471 0.7396
No log 0.4286 12 0.5640 0.0569 0.5640 0.7510
No log 0.5 14 0.5255 0.0 0.5255 0.7249
No log 0.5714 16 0.5531 0.0 0.5531 0.7437
No log 0.6429 18 0.5623 0.0 0.5623 0.7499
No log 0.7143 20 0.5242 0.0569 0.5242 0.7240
No log 0.7857 22 0.5452 0.3333 0.5452 0.7383
No log 0.8571 24 0.6195 0.25 0.6195 0.7871
No log 0.9286 26 0.5119 0.2941 0.5119 0.7155
No log 1.0 28 0.6617 0.2000 0.6617 0.8135
No log 1.0714 30 0.8820 0.2000 0.8820 0.9391
No log 1.1429 32 0.8419 0.0210 0.8419 0.9175
No log 1.2143 34 0.7764 0.0720 0.7764 0.8811
No log 1.2857 36 0.6148 0.0 0.6148 0.7841
No log 1.3571 38 0.5197 0.0 0.5197 0.7209
No log 1.4286 40 0.5628 0.3475 0.5628 0.7502
No log 1.5 42 0.5615 0.3043 0.5615 0.7494
No log 1.5714 44 0.5103 0.0 0.5103 0.7143
No log 1.6429 46 0.5016 0.0 0.5016 0.7082
No log 1.7143 48 0.5838 0.0720 0.5838 0.7640
No log 1.7857 50 0.5221 0.1278 0.5221 0.7226
No log 1.8571 52 0.7221 0.2464 0.7221 0.8498
No log 1.9286 54 2.0530 0.0239 2.0530 1.4328
No log 2.0 56 2.1200 0.0649 2.1200 1.4560
No log 2.0714 58 1.1596 0.0929 1.1596 1.0768
No log 2.1429 60 0.7752 0.1644 0.7752 0.8805
No log 2.2143 62 0.4934 0.1429 0.4934 0.7024
No log 2.2857 64 0.5961 0.2533 0.5961 0.7721
No log 2.3571 66 0.6081 0.3032 0.6081 0.7798
No log 2.4286 68 0.5156 0.0986 0.5156 0.7180
No log 2.5 70 0.6162 0.2410 0.6162 0.7850
No log 2.5714 72 0.7869 0.2300 0.7869 0.8871
No log 2.6429 74 0.7139 0.1919 0.7139 0.8449
No log 2.7143 76 0.5540 0.2105 0.5540 0.7443
No log 2.7857 78 0.5760 0.2704 0.5760 0.7589
No log 2.8571 80 0.7008 0.2621 0.7008 0.8372
No log 2.9286 82 0.6113 0.3563 0.6113 0.7818
No log 3.0 84 0.5518 0.1141 0.5518 0.7429
No log 3.0714 86 0.7833 0.2676 0.7833 0.8850
No log 3.1429 88 0.7943 0.2579 0.7943 0.8912
No log 3.2143 90 0.5651 0.2444 0.5651 0.7517
No log 3.2857 92 0.5593 0.3295 0.5593 0.7478
No log 3.3571 94 0.5624 0.2444 0.5624 0.7499
No log 3.4286 96 0.5555 0.2688 0.5555 0.7453
No log 3.5 98 0.5363 0.3446 0.5363 0.7323
No log 3.5714 100 0.5019 0.3208 0.5019 0.7085
No log 3.6429 102 0.6091 0.2871 0.6091 0.7804
No log 3.7143 104 0.8252 0.1867 0.8252 0.9084
No log 3.7857 106 0.6380 0.4286 0.6380 0.7988
No log 3.8571 108 0.5760 0.3607 0.5760 0.7590
No log 3.9286 110 0.5952 0.3607 0.5952 0.7715
No log 4.0 112 0.6618 0.4338 0.6618 0.8135
No log 4.0714 114 0.8140 0.2253 0.8140 0.9022
No log 4.1429 116 0.8522 0.2253 0.8522 0.9232
No log 4.2143 118 0.5986 0.3769 0.5986 0.7737
No log 4.2857 120 0.6268 0.3702 0.6268 0.7917
No log 4.3571 122 0.5947 0.4396 0.5947 0.7712
No log 4.4286 124 0.5845 0.3874 0.5845 0.7645
No log 4.5 126 0.5912 0.4851 0.5912 0.7689
No log 4.5714 128 0.5715 0.4518 0.5715 0.7560
No log 4.6429 130 0.6758 0.3593 0.6758 0.8221
No log 4.7143 132 0.6735 0.3761 0.6735 0.8207
No log 4.7857 134 0.5932 0.4882 0.5932 0.7702
No log 4.8571 136 0.5830 0.4882 0.5830 0.7635
No log 4.9286 138 0.5617 0.5423 0.5617 0.7495
No log 5.0 140 0.5765 0.5074 0.5765 0.7593
No log 5.0714 142 0.7124 0.25 0.7124 0.8440
No log 5.1429 144 0.6397 0.4233 0.6397 0.7998
No log 5.2143 146 0.6054 0.5 0.6054 0.7781
No log 5.2857 148 0.7828 0.2441 0.7828 0.8847
No log 5.3571 150 0.8429 0.25 0.8429 0.9181
No log 5.4286 152 0.7466 0.3414 0.7466 0.8641
No log 5.5 154 0.6662 0.4386 0.6662 0.8162
No log 5.5714 156 0.5554 0.4545 0.5554 0.7453
No log 5.6429 158 0.6733 0.4237 0.6733 0.8205
No log 5.7143 160 0.7877 0.3588 0.7877 0.8875
No log 5.7857 162 0.7121 0.3548 0.7121 0.8439
No log 5.8571 164 0.8512 0.3030 0.8512 0.9226
No log 5.9286 166 1.0298 0.1888 1.0298 1.0148
No log 6.0 168 0.8456 0.3359 0.8456 0.9196
No log 6.0714 170 0.5614 0.4286 0.5614 0.7492
No log 6.1429 172 0.5457 0.4400 0.5457 0.7387
No log 6.2143 174 0.6683 0.4087 0.6683 0.8175
No log 6.2857 176 0.9301 0.1884 0.9301 0.9644
No log 6.3571 178 0.9002 0.1882 0.9002 0.9488
No log 6.4286 180 0.7472 0.3016 0.7472 0.8644
No log 6.5 182 0.5311 0.48 0.5311 0.7288
No log 6.5714 184 0.5115 0.4400 0.5115 0.7152
No log 6.6429 186 0.5228 0.4902 0.5228 0.7230
No log 6.7143 188 0.6335 0.3982 0.6335 0.7959
No log 6.7857 190 0.8339 0.2450 0.8339 0.9132
No log 6.8571 192 0.9182 0.1524 0.9182 0.9582
No log 6.9286 194 0.7520 0.3021 0.7520 0.8672
No log 7.0 196 0.5891 0.4234 0.5891 0.7675
No log 7.0714 198 0.5301 0.4627 0.5301 0.7281
No log 7.1429 200 0.5434 0.5122 0.5434 0.7372
No log 7.2143 202 0.5891 0.4286 0.5891 0.7675
No log 7.2857 204 0.6482 0.3982 0.6482 0.8051
No log 7.3571 206 0.5799 0.4783 0.5799 0.7615
No log 7.4286 208 0.5772 0.4717 0.5772 0.7597
No log 7.5 210 0.6071 0.3929 0.6071 0.7792
No log 7.5714 212 0.6384 0.3665 0.6384 0.7990
No log 7.6429 214 0.7085 0.3684 0.7085 0.8417
No log 7.7143 216 0.7060 0.3684 0.7060 0.8402
No log 7.7857 218 0.8186 0.2424 0.8186 0.9048
No log 7.8571 220 0.8129 0.2756 0.8129 0.9016
No log 7.9286 222 0.6706 0.4035 0.6706 0.8189
No log 8.0 224 0.5313 0.4233 0.5313 0.7289
No log 8.0714 226 0.5120 0.4732 0.5120 0.7155
No log 8.1429 228 0.5359 0.4233 0.5359 0.7320
No log 8.2143 230 0.5737 0.4338 0.5737 0.7574
No log 8.2857 232 0.6386 0.4027 0.6386 0.7991
No log 8.3571 234 0.6708 0.3982 0.6708 0.8190
No log 8.4286 236 0.6569 0.4027 0.6569 0.8105
No log 8.5 238 0.5732 0.4654 0.5732 0.7571
No log 8.5714 240 0.5299 0.4233 0.5299 0.7279
No log 8.6429 242 0.5042 0.5025 0.5042 0.7101
No log 8.7143 244 0.5045 0.4059 0.5045 0.7102
No log 8.7857 246 0.5065 0.4171 0.5065 0.7117
No log 8.8571 248 0.5078 0.5025 0.5078 0.7126
No log 8.9286 250 0.5262 0.5330 0.5262 0.7254
No log 9.0 252 0.5616 0.4340 0.5616 0.7494
No log 9.0714 254 0.6034 0.4185 0.6034 0.7768
No log 9.1429 256 0.6204 0.4027 0.6204 0.7877
No log 9.2143 258 0.6401 0.4027 0.6401 0.8001
No log 9.2857 260 0.6238 0.4027 0.6238 0.7898
No log 9.3571 262 0.6098 0.4027 0.6098 0.7809
No log 9.4286 264 0.6214 0.4027 0.6214 0.7883
No log 9.5 266 0.6216 0.4027 0.6216 0.7884
No log 9.5714 268 0.6282 0.4027 0.6282 0.7926
No log 9.6429 270 0.6183 0.4027 0.6183 0.7863
No log 9.7143 272 0.6001 0.4286 0.6001 0.7747
No log 9.7857 274 0.5944 0.4286 0.5944 0.7710
No log 9.8571 276 0.5986 0.4286 0.5986 0.7737
No log 9.9286 278 0.6052 0.4027 0.6052 0.7779
No log 10.0 280 0.6096 0.4027 0.6096 0.7807

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k6_task3_organization

Finetuned
(4023)
this model