ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k6_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6513
  • Qwk: 0.2577
  • Mse: 0.6513
  • Rmse: 0.8070

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0606 2 3.2121 -0.0149 3.2121 1.7922
No log 0.1212 4 1.7620 -0.0640 1.7620 1.3274
No log 0.1818 6 0.8444 0.0427 0.8444 0.9189
No log 0.2424 8 0.8878 0.1304 0.8878 0.9422
No log 0.3030 10 0.5901 0.0569 0.5901 0.7682
No log 0.3636 12 0.6076 0.0569 0.6076 0.7795
No log 0.4242 14 0.5831 0.0303 0.5831 0.7636
No log 0.4848 16 0.8216 0.0380 0.8216 0.9064
No log 0.5455 18 0.7198 0.1841 0.7198 0.8484
No log 0.6061 20 0.6877 0.1186 0.6877 0.8293
No log 0.6667 22 0.6217 0.1304 0.6217 0.7885
No log 0.7273 24 0.6095 -0.0159 0.6095 0.7807
No log 0.7879 26 0.6340 -0.0159 0.6340 0.7963
No log 0.8485 28 0.6620 -0.0159 0.6620 0.8136
No log 0.9091 30 0.6947 0.0130 0.6947 0.8335
No log 0.9697 32 0.9295 0.1193 0.9295 0.9641
No log 1.0303 34 1.8279 0.0620 1.8279 1.3520
No log 1.0909 36 1.3712 0.0632 1.3712 1.1710
No log 1.1515 38 0.6586 0.1732 0.6586 0.8116
No log 1.2121 40 1.1898 0.0035 1.1898 1.0908
No log 1.2727 42 1.4089 0.0303 1.4089 1.1870
No log 1.3333 44 1.1643 0.0286 1.1643 1.0790
No log 1.3939 46 0.6325 0.1206 0.6325 0.7953
No log 1.4545 48 0.7465 0.0303 0.7465 0.8640
No log 1.5152 50 0.9388 0.0201 0.9388 0.9689
No log 1.5758 52 1.0641 0.0698 1.0641 1.0316
No log 1.6364 54 0.9438 0.0769 0.9438 0.9715
No log 1.6970 56 0.6396 0.0698 0.6396 0.7997
No log 1.7576 58 0.7512 0.1915 0.7512 0.8667
No log 1.8182 60 0.8534 0.1610 0.8534 0.9238
No log 1.8788 62 0.6644 0.2883 0.6644 0.8151
No log 1.9394 64 0.5623 0.2707 0.5623 0.7499
No log 2.0 66 0.8267 0.2137 0.8267 0.9092
No log 2.0606 68 0.6716 0.2549 0.6716 0.8195
No log 2.1212 70 0.4877 0.2000 0.4877 0.6983
No log 2.1818 72 0.7716 0.3103 0.7716 0.8784
No log 2.2424 74 0.7928 0.2676 0.7928 0.8904
No log 2.3030 76 0.5479 0.3292 0.5479 0.7402
No log 2.3636 78 0.5443 0.2994 0.5443 0.7378
No log 2.4242 80 0.6358 0.2821 0.6358 0.7974
No log 2.4848 82 0.4778 0.3810 0.4778 0.6912
No log 2.5455 84 0.5303 0.4023 0.5303 0.7282
No log 2.6061 86 0.5586 0.3913 0.5586 0.7474
No log 2.6667 88 0.4786 0.4409 0.4786 0.6918
No log 2.7273 90 0.4935 0.4409 0.4935 0.7025
No log 2.7879 92 0.5028 0.4409 0.5028 0.7091
No log 2.8485 94 0.6856 0.4019 0.6856 0.8280
No log 2.9091 96 1.1329 0.1079 1.1329 1.0644
No log 2.9697 98 0.9641 0.1621 0.9641 0.9819
No log 3.0303 100 0.8038 0.2566 0.8038 0.8966
No log 3.0909 102 0.5346 0.4098 0.5346 0.7312
No log 3.1515 104 0.5273 0.3073 0.5273 0.7262
No log 3.2121 106 0.5230 0.2970 0.5230 0.7232
No log 3.2727 108 0.5715 0.3446 0.5715 0.7559
No log 3.3333 110 0.6682 0.2990 0.6682 0.8175
No log 3.3939 112 0.8622 0.2441 0.8622 0.9286
No log 3.4545 114 0.6988 0.2941 0.6988 0.8359
No log 3.5152 116 0.5923 0.2832 0.5923 0.7696
No log 3.5758 118 0.6254 0.2809 0.6254 0.7908
No log 3.6364 120 0.6734 0.2917 0.6734 0.8206
No log 3.6970 122 0.8317 0.2479 0.8317 0.9120
No log 3.7576 124 0.6946 0.2563 0.6946 0.8334
No log 3.8182 126 0.5449 0.4167 0.5449 0.7382
No log 3.8788 128 0.6819 0.1864 0.6819 0.8258
No log 3.9394 130 0.5576 0.4225 0.5576 0.7467
No log 4.0 132 0.5944 0.3478 0.5944 0.7710
No log 4.0606 134 0.8139 0.1549 0.8139 0.9022
No log 4.1212 136 0.7446 0.1921 0.7446 0.8629
No log 4.1818 138 0.5401 0.4012 0.5401 0.7349
No log 4.2424 140 0.5256 0.4424 0.5256 0.7250
No log 4.3030 142 0.5499 0.3953 0.5499 0.7416
No log 4.3636 144 0.6258 0.2688 0.6258 0.7911
No log 4.4242 146 0.5873 0.3797 0.5873 0.7664
No log 4.4848 148 0.6700 0.2965 0.6700 0.8185
No log 4.5455 150 0.6268 0.375 0.6268 0.7917
No log 4.6061 152 0.7061 0.3208 0.7061 0.8403
No log 4.6667 154 1.1591 0.2254 1.1591 1.0766
No log 4.7273 156 1.3065 0.2727 1.3065 1.1430
No log 4.7879 158 0.9126 0.1803 0.9126 0.9553
No log 4.8485 160 0.6932 0.2245 0.6932 0.8326
No log 4.9091 162 0.6553 0.2245 0.6553 0.8095
No log 4.9697 164 0.5779 0.3073 0.5779 0.7602
No log 5.0303 166 0.6380 0.2670 0.6380 0.7988
No log 5.0909 168 0.7604 0.2222 0.7604 0.8720
No log 5.1515 170 0.6812 0.2245 0.6812 0.8254
No log 5.2121 172 0.7856 0.2222 0.7856 0.8863
No log 5.2727 174 0.7517 0.1845 0.7517 0.8670
No log 5.3333 176 0.6478 0.2917 0.6478 0.8048
No log 5.3939 178 0.5992 0.4518 0.5992 0.7741
No log 5.4545 180 0.6881 0.2871 0.6881 0.8295
No log 5.5152 182 0.7263 0.2146 0.7263 0.8522
No log 5.5758 184 0.7546 0.2146 0.7546 0.8687
No log 5.6364 186 0.8543 0.2542 0.8543 0.9243
No log 5.6970 188 0.7960 0.1923 0.7960 0.8922
No log 5.7576 190 0.6121 0.2893 0.6121 0.7824
No log 5.8182 192 0.5502 0.4091 0.5502 0.7418
No log 5.8788 194 0.5392 0.3953 0.5392 0.7343
No log 5.9394 196 0.5744 0.3407 0.5744 0.7579
No log 6.0 198 0.6451 0.2239 0.6451 0.8032
No log 6.0606 200 0.6394 0.2653 0.6394 0.7996
No log 6.1212 202 0.5823 0.3797 0.5823 0.7631
No log 6.1818 204 0.5641 0.4286 0.5641 0.7511
No log 6.2424 206 0.6580 0.2965 0.6580 0.8112
No log 6.3030 208 0.7790 0.1927 0.7790 0.8826
No log 6.3636 210 0.7095 0.2233 0.7095 0.8423
No log 6.4242 212 0.5979 0.2893 0.5979 0.7732
No log 6.4848 214 0.5430 0.4526 0.5430 0.7369
No log 6.5455 216 0.5393 0.4917 0.5393 0.7343
No log 6.6061 218 0.5421 0.5056 0.5421 0.7363
No log 6.6667 220 0.5761 0.3446 0.5761 0.7590
No log 6.7273 222 0.6637 0.2251 0.6637 0.8147
No log 6.7879 224 0.7660 0.2300 0.7660 0.8752
No log 6.8485 226 0.7689 0.2300 0.7689 0.8769
No log 6.9091 228 0.6404 0.2577 0.6404 0.8003
No log 6.9697 230 0.5748 0.4171 0.5748 0.7582
No log 7.0303 232 0.5798 0.3704 0.5798 0.7614
No log 7.0909 234 0.6053 0.3548 0.6053 0.7780
No log 7.1515 236 0.6721 0.2577 0.6721 0.8198
No log 7.2121 238 0.7101 0.2239 0.7101 0.8427
No log 7.2727 240 0.6834 0.2563 0.6834 0.8267
No log 7.3333 242 0.6821 0.2871 0.6821 0.8259
No log 7.3939 244 0.6492 0.2893 0.6492 0.8057
No log 7.4545 246 0.6224 0.3263 0.6224 0.7889
No log 7.5152 248 0.6138 0.3778 0.6138 0.7835
No log 7.5758 250 0.6489 0.2893 0.6489 0.8056
No log 7.6364 252 0.7443 0.2607 0.7443 0.8627
No log 7.6970 254 0.7773 0.2593 0.7773 0.8816
No log 7.7576 256 0.8139 0.2300 0.8139 0.9021
No log 7.8182 258 0.7270 0.2607 0.7270 0.8527
No log 7.8788 260 0.6700 0.2577 0.6700 0.8185
No log 7.9394 262 0.6668 0.2577 0.6668 0.8166
No log 8.0 264 0.6921 0.2621 0.6921 0.8319
No log 8.0606 266 0.7005 0.2621 0.7005 0.8369
No log 8.1212 268 0.7583 0.2607 0.7583 0.8708
No log 8.1818 270 0.8231 0.1927 0.8231 0.9072
No log 8.2424 272 0.7909 0.2300 0.7909 0.8893
No log 8.3030 274 0.6856 0.2637 0.6856 0.8280
No log 8.3636 276 0.5974 0.3333 0.5974 0.7729
No log 8.4242 278 0.5744 0.2832 0.5744 0.7579
No log 8.4848 280 0.5655 0.3073 0.5655 0.7520
No log 8.5455 282 0.5862 0.3208 0.5862 0.7656
No log 8.6061 284 0.6109 0.3520 0.6109 0.7816
No log 8.6667 286 0.6414 0.3016 0.6414 0.8009
No log 8.7273 288 0.6960 0.2607 0.6960 0.8343
No log 8.7879 290 0.7180 0.2607 0.7180 0.8473
No log 8.8485 292 0.6861 0.2621 0.6861 0.8283
No log 8.9091 294 0.6524 0.2688 0.6524 0.8077
No log 8.9697 296 0.6202 0.3016 0.6202 0.7876
No log 9.0303 298 0.5842 0.3563 0.5842 0.7643
No log 9.0909 300 0.5646 0.3659 0.5646 0.7514
No log 9.1515 302 0.5694 0.3563 0.5694 0.7546
No log 9.2121 304 0.5902 0.3520 0.5902 0.7682
No log 9.2727 306 0.6126 0.3016 0.6126 0.7827
No log 9.3333 308 0.6340 0.3016 0.6340 0.7963
No log 9.3939 310 0.6389 0.3016 0.6389 0.7993
No log 9.4545 312 0.6521 0.3016 0.6521 0.8075
No log 9.5152 314 0.6693 0.2941 0.6693 0.8181
No log 9.5758 316 0.6672 0.2941 0.6672 0.8168
No log 9.6364 318 0.6532 0.2577 0.6532 0.8082
No log 9.6970 320 0.6468 0.3016 0.6468 0.8043
No log 9.7576 322 0.6446 0.3016 0.6446 0.8029
No log 9.8182 324 0.6425 0.3016 0.6425 0.8016
No log 9.8788 326 0.6463 0.3016 0.6463 0.8039
No log 9.9394 328 0.6492 0.2577 0.6492 0.8057
No log 10.0 330 0.6513 0.2577 0.6513 0.8070

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k6_task3_organization

Finetuned
(4023)
this model