ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k8_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1986
  • Qwk: 0.5909
  • Mse: 1.1986
  • Rmse: 1.0948

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 2.4144 -0.0350 2.4144 1.5538
No log 0.1429 4 1.6143 0.0760 1.6143 1.2706
No log 0.2143 6 1.5262 0.1172 1.5262 1.2354
No log 0.2857 8 1.4181 0.2585 1.4181 1.1909
No log 0.3571 10 1.5937 0.2998 1.5937 1.2624
No log 0.4286 12 2.0439 0.2339 2.0439 1.4297
No log 0.5 14 1.9472 0.1355 1.9472 1.3954
No log 0.5714 16 1.8158 0.2905 1.8158 1.3475
No log 0.6429 18 1.6146 0.3350 1.6146 1.2707
No log 0.7143 20 1.4844 0.3599 1.4844 1.2184
No log 0.7857 22 1.4286 0.3576 1.4286 1.1952
No log 0.8571 24 1.4535 0.3967 1.4535 1.2056
No log 0.9286 26 1.4781 0.3968 1.4781 1.2158
No log 1.0 28 1.6605 0.3670 1.6605 1.2886
No log 1.0714 30 1.9520 0.3424 1.9520 1.3971
No log 1.1429 32 2.1558 0.2557 2.1558 1.4683
No log 1.2143 34 2.0933 0.2268 2.0933 1.4468
No log 1.2857 36 1.9270 0.1936 1.9270 1.3882
No log 1.3571 38 1.7828 0.2439 1.7828 1.3352
No log 1.4286 40 1.7126 0.3001 1.7126 1.3087
No log 1.5 42 1.6923 0.3811 1.6923 1.3009
No log 1.5714 44 1.6446 0.3701 1.6446 1.2824
No log 1.6429 46 1.6751 0.3778 1.6751 1.2943
No log 1.7143 48 1.6044 0.4074 1.6044 1.2666
No log 1.7857 50 1.6828 0.3511 1.6828 1.2972
No log 1.8571 52 1.7433 0.3285 1.7433 1.3203
No log 1.9286 54 1.7678 0.3625 1.7678 1.3296
No log 2.0 56 1.7213 0.3875 1.7213 1.3120
No log 2.0714 58 1.6606 0.3874 1.6606 1.2886
No log 2.1429 60 1.6907 0.3934 1.6907 1.3003
No log 2.2143 62 1.7492 0.3804 1.7492 1.3226
No log 2.2857 64 1.7602 0.3863 1.7602 1.3267
No log 2.3571 66 1.7281 0.3816 1.7281 1.3146
No log 2.4286 68 1.7211 0.4181 1.7211 1.3119
No log 2.5 70 1.7148 0.4263 1.7148 1.3095
No log 2.5714 72 1.8587 0.4395 1.8587 1.3633
No log 2.6429 74 1.9887 0.4328 1.9887 1.4102
No log 2.7143 76 2.1046 0.4191 2.1046 1.4507
No log 2.7857 78 1.8530 0.4367 1.8530 1.3612
No log 2.8571 80 1.5199 0.4732 1.5199 1.2328
No log 2.9286 82 1.4150 0.5329 1.4150 1.1895
No log 3.0 84 1.4803 0.5391 1.4803 1.2167
No log 3.0714 86 1.4862 0.5636 1.4862 1.2191
No log 3.1429 88 1.5934 0.5179 1.5934 1.2623
No log 3.2143 90 1.5575 0.5371 1.5575 1.2480
No log 3.2857 92 1.3749 0.5615 1.3749 1.1726
No log 3.3571 94 1.2188 0.5865 1.2188 1.1040
No log 3.4286 96 1.1231 0.6279 1.1231 1.0598
No log 3.5 98 1.0690 0.6308 1.0690 1.0339
No log 3.5714 100 1.0025 0.6606 1.0025 1.0013
No log 3.6429 102 1.0362 0.6490 1.0362 1.0180
No log 3.7143 104 1.2358 0.6298 1.2358 1.1117
No log 3.7857 106 1.5398 0.5591 1.5398 1.2409
No log 3.8571 108 1.6313 0.5376 1.6313 1.2772
No log 3.9286 110 1.7240 0.4768 1.7240 1.3130
No log 4.0 112 1.8626 0.4425 1.8626 1.3648
No log 4.0714 114 2.0248 0.4442 2.0248 1.4230
No log 4.1429 116 2.2227 0.4407 2.2227 1.4909
No log 4.2143 118 2.0574 0.4760 2.0574 1.4344
No log 4.2857 120 1.6506 0.5019 1.6506 1.2847
No log 4.3571 122 1.3131 0.5756 1.3131 1.1459
No log 4.4286 124 1.1154 0.6022 1.1154 1.0561
No log 4.5 126 1.1020 0.5933 1.1020 1.0498
No log 4.5714 128 1.1299 0.6174 1.1299 1.0630
No log 4.6429 130 1.2416 0.6032 1.2416 1.1143
No log 4.7143 132 1.2165 0.6059 1.2165 1.1029
No log 4.7857 134 1.1986 0.6221 1.1986 1.0948
No log 4.8571 136 1.1705 0.6176 1.1705 1.0819
No log 4.9286 138 1.0769 0.6430 1.0769 1.0377
No log 5.0 140 1.0208 0.6469 1.0208 1.0103
No log 5.0714 142 0.9916 0.6484 0.9916 0.9958
No log 5.1429 144 1.0286 0.6395 1.0286 1.0142
No log 5.2143 146 1.2556 0.6224 1.2556 1.1205
No log 5.2857 148 1.4408 0.5918 1.4408 1.2003
No log 5.3571 150 1.3794 0.5880 1.3794 1.1745
No log 5.4286 152 1.2571 0.5920 1.2571 1.1212
No log 5.5 154 1.2289 0.5870 1.2289 1.1086
No log 5.5714 156 1.1360 0.6082 1.1360 1.0658
No log 5.6429 158 1.1240 0.6121 1.1240 1.0602
No log 5.7143 160 1.2120 0.5956 1.2120 1.1009
No log 5.7857 162 1.2542 0.5765 1.2542 1.1199
No log 5.8571 164 1.1573 0.5926 1.1573 1.0758
No log 5.9286 166 1.0488 0.6348 1.0488 1.0241
No log 6.0 168 0.9719 0.5830 0.9719 0.9858
No log 6.0714 170 1.0442 0.6154 1.0442 1.0218
No log 6.1429 172 1.1671 0.6024 1.1671 1.0803
No log 6.2143 174 1.2561 0.5906 1.2561 1.1208
No log 6.2857 176 1.2112 0.6015 1.2112 1.1005
No log 6.3571 178 1.1208 0.6102 1.1208 1.0587
No log 6.4286 180 1.1035 0.6232 1.1035 1.0505
No log 6.5 182 1.1438 0.6152 1.1438 1.0695
No log 6.5714 184 1.2169 0.6193 1.2169 1.1031
No log 6.6429 186 1.2167 0.6193 1.2167 1.1030
No log 6.7143 188 1.1596 0.6193 1.1596 1.0769
No log 6.7857 190 1.0984 0.6218 1.0984 1.0481
No log 6.8571 192 1.0248 0.6218 1.0248 1.0123
No log 6.9286 194 1.0256 0.6373 1.0256 1.0127
No log 7.0 196 1.0632 0.6218 1.0632 1.0311
No log 7.0714 198 1.1013 0.6104 1.1013 1.0494
No log 7.1429 200 1.0784 0.6097 1.0784 1.0384
No log 7.2143 202 1.0558 0.6155 1.0558 1.0275
No log 7.2857 204 0.9846 0.5974 0.9846 0.9923
No log 7.3571 206 0.9585 0.6078 0.9585 0.9790
No log 7.4286 208 0.9781 0.6086 0.9781 0.9890
No log 7.5 210 1.0191 0.6164 1.0191 1.0095
No log 7.5714 212 1.0388 0.6164 1.0388 1.0192
No log 7.6429 214 1.0348 0.6016 1.0348 1.0172
No log 7.7143 216 1.0475 0.6007 1.0475 1.0235
No log 7.7857 218 1.0381 0.6001 1.0381 1.0189
No log 7.8571 220 0.9980 0.5872 0.9980 0.9990
No log 7.9286 222 0.9531 0.5984 0.9531 0.9763
No log 8.0 224 0.9294 0.5841 0.9294 0.9641
No log 8.0714 226 0.9533 0.6018 0.9533 0.9764
No log 8.1429 228 0.9860 0.5885 0.9860 0.9930
No log 8.2143 230 1.0127 0.5865 1.0127 1.0063
No log 8.2857 232 1.0610 0.5933 1.0610 1.0300
No log 8.3571 234 1.0980 0.5917 1.0980 1.0478
No log 8.4286 236 1.1164 0.5917 1.1164 1.0566
No log 8.5 238 1.1381 0.5994 1.1381 1.0668
No log 8.5714 240 1.1253 0.5977 1.1253 1.0608
No log 8.6429 242 1.1005 0.5959 1.1005 1.0490
No log 8.7143 244 1.0740 0.5818 1.0740 1.0364
No log 8.7857 246 1.0854 0.5869 1.0854 1.0418
No log 8.8571 248 1.0939 0.5869 1.0939 1.0459
No log 8.9286 250 1.1173 0.5977 1.1173 1.0570
No log 9.0 252 1.1476 0.5994 1.1476 1.0713
No log 9.0714 254 1.1652 0.5999 1.1652 1.0794
No log 9.1429 256 1.1707 0.5920 1.1707 1.0820
No log 9.2143 258 1.1688 0.5920 1.1688 1.0811
No log 9.2857 260 1.1671 0.5920 1.1671 1.0803
No log 9.3571 262 1.1759 0.5920 1.1759 1.0844
No log 9.4286 264 1.1717 0.5920 1.1717 1.0825
No log 9.5 266 1.1690 0.5920 1.1690 1.0812
No log 9.5714 268 1.1711 0.5920 1.1711 1.0822
No log 9.6429 270 1.1760 0.5920 1.1760 1.0844
No log 9.7143 272 1.1827 0.5909 1.1827 1.0875
No log 9.7857 274 1.1892 0.5909 1.1892 1.0905
No log 9.8571 276 1.1941 0.5909 1.1941 1.0928
No log 9.9286 278 1.1975 0.5909 1.1975 1.0943
No log 10.0 280 1.1986 0.5909 1.1986 1.0948

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k8_task5_organization

Finetuned
(4023)
this model