ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k7_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9662
  • Qwk: 0.6819
  • Mse: 0.9662
  • Rmse: 0.9829

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0571 2 2.2490 0.0081 2.2490 1.4997
No log 0.1143 4 1.4130 0.1915 1.4130 1.1887
No log 0.1714 6 1.5537 0.0639 1.5537 1.2465
No log 0.2286 8 1.5720 0.0905 1.5720 1.2538
No log 0.2857 10 1.5267 0.1134 1.5267 1.2356
No log 0.3429 12 1.3625 0.2058 1.3625 1.1673
No log 0.4 14 1.2899 0.2867 1.2899 1.1357
No log 0.4571 16 1.2404 0.2733 1.2404 1.1137
No log 0.5143 18 1.2122 0.2602 1.2122 1.1010
No log 0.5714 20 1.0804 0.4301 1.0804 1.0394
No log 0.6286 22 1.0621 0.4900 1.0621 1.0306
No log 0.6857 24 1.0153 0.4819 1.0153 1.0076
No log 0.7429 26 1.0078 0.5316 1.0078 1.0039
No log 0.8 28 0.9769 0.5433 0.9769 0.9884
No log 0.8571 30 1.0173 0.5365 1.0173 1.0086
No log 0.9143 32 0.9723 0.5258 0.9723 0.9861
No log 0.9714 34 0.9089 0.6059 0.9089 0.9534
No log 1.0286 36 0.9087 0.5980 0.9087 0.9532
No log 1.0857 38 0.8973 0.6153 0.8973 0.9473
No log 1.1429 40 0.8914 0.6006 0.8914 0.9441
No log 1.2 42 0.9807 0.5826 0.9807 0.9903
No log 1.2571 44 1.2304 0.5653 1.2304 1.1092
No log 1.3143 46 1.4200 0.4748 1.4200 1.1916
No log 1.3714 48 1.2233 0.5547 1.2233 1.1060
No log 1.4286 50 0.9527 0.5890 0.9527 0.9761
No log 1.4857 52 0.9411 0.5953 0.9411 0.9701
No log 1.5429 54 1.1171 0.5604 1.1171 1.0570
No log 1.6 56 1.4986 0.4614 1.4986 1.2242
No log 1.6571 58 1.5859 0.4472 1.5859 1.2593
No log 1.7143 60 1.4430 0.4710 1.4430 1.2013
No log 1.7714 62 1.0928 0.5547 1.0928 1.0454
No log 1.8286 64 0.8641 0.6353 0.8641 0.9296
No log 1.8857 66 0.8243 0.6050 0.8243 0.9079
No log 1.9429 68 0.8214 0.6319 0.8214 0.9063
No log 2.0 70 0.8469 0.6335 0.8469 0.9203
No log 2.0571 72 1.0465 0.6106 1.0465 1.0230
No log 2.1143 74 1.1546 0.5993 1.1546 1.0745
No log 2.1714 76 1.0615 0.6193 1.0615 1.0303
No log 2.2286 78 0.8776 0.6447 0.8776 0.9368
No log 2.2857 80 0.8890 0.6629 0.8890 0.9429
No log 2.3429 82 1.0028 0.6151 1.0028 1.0014
No log 2.4 84 1.1391 0.5654 1.1391 1.0673
No log 2.4571 86 1.2507 0.5220 1.2507 1.1183
No log 2.5143 88 1.2973 0.5338 1.2973 1.1390
No log 2.5714 90 1.5006 0.5302 1.5006 1.2250
No log 2.6286 92 1.5107 0.5261 1.5107 1.2291
No log 2.6857 94 1.2257 0.5669 1.2257 1.1071
No log 2.7429 96 1.0778 0.5931 1.0778 1.0382
No log 2.8 98 1.1752 0.5589 1.1752 1.0841
No log 2.8571 100 1.3097 0.5097 1.3097 1.1444
No log 2.9143 102 1.5294 0.4882 1.5294 1.2367
No log 2.9714 104 1.6401 0.4673 1.6401 1.2807
No log 3.0286 106 1.4243 0.5052 1.4243 1.1934
No log 3.0857 108 0.9593 0.6320 0.9593 0.9794
No log 3.1429 110 0.6844 0.6872 0.6844 0.8273
No log 3.2 112 0.6907 0.6357 0.6907 0.8311
No log 3.2571 114 0.6998 0.6669 0.6998 0.8366
No log 3.3143 116 0.8310 0.6876 0.8310 0.9116
No log 3.3714 118 1.2009 0.5902 1.2009 1.0959
No log 3.4286 120 1.4423 0.5401 1.4423 1.2010
No log 3.4857 122 1.4257 0.5578 1.4257 1.1940
No log 3.5429 124 1.1832 0.6172 1.1832 1.0878
No log 3.6 126 0.8999 0.6888 0.8999 0.9486
No log 3.6571 128 0.8102 0.6654 0.8102 0.9001
No log 3.7143 130 0.8463 0.6663 0.8463 0.9199
No log 3.7714 132 1.0390 0.6243 1.0390 1.0193
No log 3.8286 134 1.3699 0.5501 1.3699 1.1704
No log 3.8857 136 1.5470 0.5021 1.5470 1.2438
No log 3.9429 138 1.5816 0.4996 1.5816 1.2576
No log 4.0 140 1.4849 0.5108 1.4849 1.2186
No log 4.0571 142 1.2185 0.5754 1.2185 1.1038
No log 4.1143 144 0.9473 0.6592 0.9473 0.9733
No log 4.1714 146 0.8279 0.6529 0.8279 0.9099
No log 4.2286 148 0.8550 0.6885 0.8550 0.9246
No log 4.2857 150 1.0383 0.6458 1.0383 1.0190
No log 4.3429 152 1.3441 0.5504 1.3441 1.1594
No log 4.4 154 1.6302 0.5079 1.6302 1.2768
No log 4.4571 156 1.7188 0.5033 1.7188 1.3110
No log 4.5143 158 1.5525 0.5193 1.5525 1.2460
No log 4.5714 160 1.3918 0.5503 1.3918 1.1798
No log 4.6286 162 1.1869 0.5953 1.1869 1.0894
No log 4.6857 164 1.0977 0.6106 1.0977 1.0477
No log 4.7429 166 1.1781 0.5837 1.1781 1.0854
No log 4.8 168 1.2916 0.5436 1.2916 1.1365
No log 4.8571 170 1.3719 0.5584 1.3719 1.1713
No log 4.9143 172 1.4620 0.5109 1.4620 1.2091
No log 4.9714 174 1.5141 0.5175 1.5141 1.2305
No log 5.0286 176 1.5582 0.5193 1.5582 1.2483
No log 5.0857 178 1.4182 0.5418 1.4182 1.1909
No log 5.1429 180 1.1215 0.6323 1.1215 1.0590
No log 5.2 182 0.8942 0.6915 0.8942 0.9456
No log 5.2571 184 0.7824 0.6783 0.7824 0.8845
No log 5.3143 186 0.8148 0.6830 0.8148 0.9027
No log 5.3714 188 0.9955 0.6758 0.9955 0.9977
No log 5.4286 190 1.2329 0.5671 1.2329 1.1104
No log 5.4857 192 1.2631 0.5430 1.2631 1.1239
No log 5.5429 194 1.1647 0.5820 1.1647 1.0792
No log 5.6 196 0.9765 0.6644 0.9765 0.9882
No log 5.6571 198 0.8761 0.6723 0.8761 0.9360
No log 5.7143 200 0.9042 0.6810 0.9042 0.9509
No log 5.7714 202 0.9462 0.6756 0.9462 0.9728
No log 5.8286 204 0.9868 0.6639 0.9868 0.9934
No log 5.8857 206 1.0218 0.6463 1.0218 1.0108
No log 5.9429 208 0.9564 0.6639 0.9564 0.9780
No log 6.0 210 0.9667 0.6647 0.9667 0.9832
No log 6.0571 212 0.9886 0.6647 0.9886 0.9943
No log 6.1143 214 0.9674 0.6837 0.9674 0.9836
No log 6.1714 216 0.9129 0.6979 0.9129 0.9555
No log 6.2286 218 0.8360 0.6822 0.8360 0.9143
No log 6.2857 220 0.8283 0.6759 0.8283 0.9101
No log 6.3429 222 0.7877 0.6838 0.7877 0.8875
No log 6.4 224 0.7741 0.6810 0.7741 0.8798
No log 6.4571 226 0.8214 0.6908 0.8214 0.9063
No log 6.5143 228 0.8495 0.6871 0.8495 0.9217
No log 6.5714 230 0.9258 0.6822 0.9258 0.9622
No log 6.6286 232 1.0356 0.6273 1.0356 1.0176
No log 6.6857 234 1.1225 0.5814 1.1225 1.0595
No log 6.7429 236 1.1504 0.5783 1.1504 1.0726
No log 6.8 238 1.1166 0.5972 1.1166 1.0567
No log 6.8571 240 1.0034 0.6422 1.0034 1.0017
No log 6.9143 242 0.8849 0.6833 0.8849 0.9407
No log 6.9714 244 0.8534 0.6894 0.8534 0.9238
No log 7.0286 246 0.8397 0.6964 0.8397 0.9164
No log 7.0857 248 0.8559 0.6795 0.8559 0.9252
No log 7.1429 250 0.9594 0.6543 0.9594 0.9795
No log 7.2 252 1.0793 0.5995 1.0793 1.0389
No log 7.2571 254 1.1627 0.5956 1.1627 1.0783
No log 7.3143 256 1.1541 0.5783 1.1541 1.0743
No log 7.3714 258 1.0637 0.6118 1.0637 1.0314
No log 7.4286 260 0.9638 0.6660 0.9638 0.9817
No log 7.4857 262 0.8803 0.6845 0.8803 0.9382
No log 7.5429 264 0.8295 0.6879 0.8295 0.9108
No log 7.6 266 0.8018 0.6923 0.8018 0.8954
No log 7.6571 268 0.8197 0.6957 0.8197 0.9054
No log 7.7143 270 0.8737 0.6801 0.8737 0.9347
No log 7.7714 272 0.9026 0.6832 0.9026 0.9501
No log 7.8286 274 0.9408 0.6653 0.9408 0.9700
No log 7.8857 276 0.9868 0.6671 0.9868 0.9934
No log 7.9429 278 1.0002 0.6752 1.0002 1.0001
No log 8.0 280 0.9642 0.6538 0.9642 0.9819
No log 8.0571 282 0.9316 0.6789 0.9316 0.9652
No log 8.1143 284 0.8822 0.6862 0.8822 0.9393
No log 8.1714 286 0.8272 0.7024 0.8272 0.9095
No log 8.2286 288 0.8179 0.7024 0.8179 0.9044
No log 8.2857 290 0.8519 0.6925 0.8519 0.9230
No log 8.3429 292 0.8866 0.6749 0.8866 0.9416
No log 8.4 294 0.9395 0.6635 0.9395 0.9693
No log 8.4571 296 0.9558 0.6635 0.9558 0.9776
No log 8.5143 298 0.9484 0.6635 0.9484 0.9739
No log 8.5714 300 0.9160 0.6511 0.9160 0.9571
No log 8.6286 302 0.8964 0.6785 0.8964 0.9468
No log 8.6857 304 0.8823 0.6785 0.8823 0.9393
No log 8.7429 306 0.8460 0.6794 0.8460 0.9198
No log 8.8 308 0.8181 0.6838 0.8181 0.9045
No log 8.8571 310 0.8019 0.7024 0.8019 0.8955
No log 8.9143 312 0.7949 0.7024 0.7949 0.8915
No log 8.9714 314 0.7905 0.7024 0.7905 0.8891
No log 9.0286 316 0.7939 0.7024 0.7939 0.8910
No log 9.0857 318 0.8140 0.6925 0.8140 0.9022
No log 9.1429 320 0.8502 0.6925 0.8502 0.9221
No log 9.2 322 0.8962 0.6849 0.8962 0.9467
No log 9.2571 324 0.9456 0.6969 0.9456 0.9724
No log 9.3143 326 1.0011 0.6819 1.0011 1.0005
No log 9.3714 328 1.0497 0.6656 1.0497 1.0246
No log 9.4286 330 1.0680 0.6455 1.0680 1.0334
No log 9.4857 332 1.0738 0.6199 1.0738 1.0363
No log 9.5429 334 1.0655 0.6212 1.0655 1.0322
No log 9.6 336 1.0551 0.6455 1.0551 1.0272
No log 9.6571 338 1.0378 0.6656 1.0378 1.0187
No log 9.7143 340 1.0223 0.6808 1.0223 1.0111
No log 9.7714 342 1.0049 0.6802 1.0049 1.0025
No log 9.8286 344 0.9885 0.6819 0.9885 0.9942
No log 9.8857 346 0.9753 0.6819 0.9753 0.9876
No log 9.9429 348 0.9688 0.6819 0.9688 0.9843
No log 10.0 350 0.9662 0.6819 0.9662 0.9829

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k7_task5_organization

Finetuned
(4023)
this model