ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k8_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0692
  • Qwk: 0.6292
  • Mse: 1.0692
  • Rmse: 1.0340

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 2.4368 0.0170 2.4368 1.5610
No log 0.125 4 1.6639 0.1265 1.6639 1.2899
No log 0.1875 6 1.7030 -0.0044 1.7030 1.3050
No log 0.25 8 1.7409 0.1332 1.7409 1.3194
No log 0.3125 10 1.7665 0.2389 1.7665 1.3291
No log 0.375 12 1.6905 0.2851 1.6905 1.3002
No log 0.4375 14 1.6569 0.2924 1.6569 1.2872
No log 0.5 16 1.6187 0.3467 1.6187 1.2723
No log 0.5625 18 1.3933 0.2476 1.3933 1.1804
No log 0.625 20 1.2790 0.2217 1.2790 1.1309
No log 0.6875 22 1.2802 0.2846 1.2802 1.1315
No log 0.75 24 1.2515 0.2495 1.2515 1.1187
No log 0.8125 26 1.2441 0.2330 1.2441 1.1154
No log 0.875 28 1.2869 0.3264 1.2869 1.1344
No log 0.9375 30 1.2641 0.3676 1.2641 1.1243
No log 1.0 32 1.2120 0.3931 1.2120 1.1009
No log 1.0625 34 1.1730 0.3814 1.1730 1.0831
No log 1.125 36 1.1501 0.4272 1.1501 1.0724
No log 1.1875 38 1.1131 0.4075 1.1131 1.0551
No log 1.25 40 1.1009 0.4293 1.1009 1.0492
No log 1.3125 42 1.0895 0.4331 1.0895 1.0438
No log 1.375 44 1.1371 0.4953 1.1371 1.0664
No log 1.4375 46 1.1822 0.4904 1.1822 1.0873
No log 1.5 48 1.2365 0.4521 1.2365 1.1120
No log 1.5625 50 1.2205 0.4838 1.2205 1.1047
No log 1.625 52 1.2152 0.4680 1.2152 1.1023
No log 1.6875 54 1.0994 0.4738 1.0994 1.0485
No log 1.75 56 1.0871 0.5106 1.0871 1.0426
No log 1.8125 58 1.0915 0.4787 1.0915 1.0448
No log 1.875 60 1.0637 0.4684 1.0637 1.0313
No log 1.9375 62 1.0795 0.4794 1.0795 1.0390
No log 2.0 64 1.2050 0.4482 1.2050 1.0977
No log 2.0625 66 1.2526 0.4649 1.2526 1.1192
No log 2.125 68 1.1777 0.4840 1.1777 1.0852
No log 2.1875 70 1.0680 0.4846 1.0680 1.0334
No log 2.25 72 0.9713 0.5468 0.9713 0.9855
No log 2.3125 74 0.9510 0.5981 0.9510 0.9752
No log 2.375 76 1.0052 0.5533 1.0052 1.0026
No log 2.4375 78 1.3063 0.5200 1.3063 1.1429
No log 2.5 80 1.3657 0.4702 1.3657 1.1686
No log 2.5625 82 1.2633 0.4605 1.2633 1.1240
No log 2.625 84 1.1759 0.5081 1.1759 1.0844
No log 2.6875 86 1.1162 0.5339 1.1162 1.0565
No log 2.75 88 1.0818 0.5579 1.0818 1.0401
No log 2.8125 90 0.9999 0.5805 0.9999 0.9999
No log 2.875 92 0.9831 0.5716 0.9831 0.9915
No log 2.9375 94 1.0717 0.5754 1.0717 1.0353
No log 3.0 96 1.1018 0.5925 1.1018 1.0497
No log 3.0625 98 0.9779 0.5982 0.9779 0.9889
No log 3.125 100 0.9138 0.6255 0.9138 0.9559
No log 3.1875 102 0.8914 0.6129 0.8914 0.9442
No log 3.25 104 0.8946 0.6140 0.8946 0.9458
No log 3.3125 106 0.8964 0.6076 0.8964 0.9468
No log 3.375 108 0.9153 0.6187 0.9153 0.9567
No log 3.4375 110 1.0031 0.6086 1.0031 1.0015
No log 3.5 112 1.2576 0.5545 1.2576 1.1214
No log 3.5625 114 1.5144 0.5217 1.5144 1.2306
No log 3.625 116 1.5392 0.5223 1.5392 1.2407
No log 3.6875 118 1.3254 0.5370 1.3254 1.1513
No log 3.75 120 1.1411 0.5912 1.1411 1.0682
No log 3.8125 122 1.0610 0.6112 1.0610 1.0301
No log 3.875 124 1.0553 0.5956 1.0553 1.0273
No log 3.9375 126 1.0057 0.5742 1.0057 1.0029
No log 4.0 128 0.9320 0.5591 0.9320 0.9654
No log 4.0625 130 1.0088 0.5496 1.0088 1.0044
No log 4.125 132 1.1129 0.5455 1.1129 1.0549
No log 4.1875 134 1.1682 0.5428 1.1682 1.0809
No log 4.25 136 1.1659 0.5472 1.1659 1.0798
No log 4.3125 138 1.2829 0.5495 1.2829 1.1326
No log 4.375 140 1.3074 0.5542 1.3074 1.1434
No log 4.4375 142 1.3250 0.5512 1.3250 1.1511
No log 4.5 144 1.3414 0.5408 1.3414 1.1582
No log 4.5625 146 1.2082 0.5589 1.2082 1.0992
No log 4.625 148 1.0410 0.5887 1.0410 1.0203
No log 4.6875 150 0.9627 0.6414 0.9627 0.9812
No log 4.75 152 1.0040 0.6358 1.0040 1.0020
No log 4.8125 154 1.0924 0.6246 1.0924 1.0452
No log 4.875 156 1.0736 0.6374 1.0736 1.0361
No log 4.9375 158 1.0044 0.6722 1.0044 1.0022
No log 5.0 160 1.0023 0.6722 1.0023 1.0012
No log 5.0625 162 1.1328 0.6305 1.1328 1.0643
No log 5.125 164 1.2473 0.5861 1.2473 1.1168
No log 5.1875 166 1.2044 0.5965 1.2044 1.0975
No log 5.25 168 1.0631 0.6342 1.0631 1.0311
No log 5.3125 170 0.9741 0.6602 0.9741 0.9869
No log 5.375 172 0.9521 0.6541 0.9521 0.9758
No log 5.4375 174 0.9749 0.6572 0.9749 0.9874
No log 5.5 176 1.0751 0.6202 1.0751 1.0369
No log 5.5625 178 1.1060 0.6133 1.1060 1.0516
No log 5.625 180 1.1643 0.5942 1.1643 1.0790
No log 5.6875 182 1.2327 0.5766 1.2327 1.1103
No log 5.75 184 1.1998 0.5654 1.1998 1.0953
No log 5.8125 186 1.1536 0.5805 1.1536 1.0741
No log 5.875 188 1.0810 0.6310 1.0810 1.0397
No log 5.9375 190 0.9935 0.6314 0.9935 0.9967
No log 6.0 192 0.9686 0.6638 0.9686 0.9842
No log 6.0625 194 1.0367 0.6181 1.0367 1.0182
No log 6.125 196 1.2126 0.6085 1.2126 1.1012
No log 6.1875 198 1.4077 0.5672 1.4077 1.1865
No log 6.25 200 1.4245 0.5601 1.4245 1.1935
No log 6.3125 202 1.2742 0.6063 1.2742 1.1288
No log 6.375 204 1.0690 0.6139 1.0690 1.0339
No log 6.4375 206 0.9859 0.6465 0.9859 0.9929
No log 6.5 208 0.9466 0.6559 0.9466 0.9729
No log 6.5625 210 0.9683 0.6501 0.9683 0.9840
No log 6.625 212 1.0415 0.6371 1.0415 1.0206
No log 6.6875 214 1.1107 0.6124 1.1107 1.0539
No log 6.75 216 1.0933 0.6279 1.0933 1.0456
No log 6.8125 218 0.9874 0.5999 0.9874 0.9937
No log 6.875 220 0.8998 0.6058 0.8998 0.9486
No log 6.9375 222 0.8714 0.6209 0.8714 0.9335
No log 7.0 224 0.8803 0.6067 0.8803 0.9382
No log 7.0625 226 0.9361 0.6066 0.9361 0.9675
No log 7.125 228 1.0598 0.5903 1.0598 1.0294
No log 7.1875 230 1.1462 0.5931 1.1462 1.0706
No log 7.25 232 1.1211 0.5681 1.1211 1.0588
No log 7.3125 234 1.0315 0.6097 1.0315 1.0157
No log 7.375 236 1.0021 0.6241 1.0021 1.0010
No log 7.4375 238 0.9794 0.6241 0.9794 0.9897
No log 7.5 240 1.0154 0.6152 1.0154 1.0077
No log 7.5625 242 1.0255 0.6106 1.0255 1.0127
No log 7.625 244 0.9824 0.6306 0.9824 0.9912
No log 7.6875 246 0.9709 0.6273 0.9709 0.9853
No log 7.75 248 0.9572 0.6254 0.9572 0.9784
No log 7.8125 250 0.9881 0.6273 0.9881 0.9940
No log 7.875 252 0.9812 0.6273 0.9812 0.9905
No log 7.9375 254 0.9454 0.6254 0.9454 0.9723
No log 8.0 256 0.9107 0.6296 0.9107 0.9543
No log 8.0625 258 0.9022 0.6234 0.9022 0.9498
No log 8.125 260 0.9239 0.6193 0.9239 0.9612
No log 8.1875 262 0.9914 0.6359 0.9914 0.9957
No log 8.25 264 1.0406 0.6391 1.0406 1.0201
No log 8.3125 266 1.0300 0.6391 1.0300 1.0149
No log 8.375 268 0.9831 0.6434 0.9831 0.9915
No log 8.4375 270 0.9185 0.6358 0.9185 0.9584
No log 8.5 272 0.8801 0.6402 0.8801 0.9381
No log 8.5625 274 0.8735 0.6445 0.8735 0.9346
No log 8.625 276 0.8839 0.6464 0.8839 0.9402
No log 8.6875 278 0.9150 0.6358 0.9150 0.9566
No log 8.75 280 0.9303 0.6391 0.9303 0.9645
No log 8.8125 282 0.9466 0.6348 0.9466 0.9729
No log 8.875 284 0.9786 0.6434 0.9786 0.9893
No log 8.9375 286 0.9855 0.6434 0.9855 0.9927
No log 9.0 288 0.9987 0.6347 0.9987 0.9994
No log 9.0625 290 1.0152 0.6347 1.0152 1.0076
No log 9.125 292 1.0183 0.6347 1.0183 1.0091
No log 9.1875 294 1.0112 0.6347 1.0112 1.0056
No log 9.25 296 1.0182 0.6347 1.0182 1.0090
No log 9.3125 298 1.0264 0.6431 1.0264 1.0131
No log 9.375 300 1.0340 0.6431 1.0340 1.0169
No log 9.4375 302 1.0450 0.6431 1.0450 1.0223
No log 9.5 304 1.0507 0.6274 1.0507 1.0250
No log 9.5625 306 1.0519 0.6274 1.0519 1.0256
No log 9.625 308 1.0602 0.6274 1.0602 1.0297
No log 9.6875 310 1.0728 0.6292 1.0728 1.0358
No log 9.75 312 1.0769 0.6292 1.0769 1.0377
No log 9.8125 314 1.0772 0.6292 1.0772 1.0379
No log 9.875 316 1.0745 0.6292 1.0745 1.0366
No log 9.9375 318 1.0712 0.6292 1.0712 1.0350
No log 10.0 320 1.0692 0.6292 1.0692 1.0340

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k8_task5_organization

Finetuned
(4023)
this model