ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k6_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6050
  • Qwk: 0.3161
  • Mse: 0.6050
  • Rmse: 0.7778

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 3.0885 -0.0114 3.0885 1.7574
No log 0.1333 4 1.5351 -0.0070 1.5351 1.2390
No log 0.2 6 0.9492 0.0462 0.9492 0.9743
No log 0.2667 8 0.7810 0.1770 0.7810 0.8837
No log 0.3333 10 0.6009 0.0569 0.6009 0.7752
No log 0.4 12 0.6257 -0.0732 0.6257 0.7910
No log 0.4667 14 0.6285 -0.0732 0.6285 0.7928
No log 0.5333 16 0.6139 0.1008 0.6139 0.7835
No log 0.6 18 0.8207 0.1276 0.8207 0.9059
No log 0.6667 20 0.8019 0.2000 0.8019 0.8955
No log 0.7333 22 0.9251 0.0476 0.9251 0.9618
No log 0.8 24 0.6633 0.1638 0.6633 0.8144
No log 0.8667 26 0.6217 0.0725 0.6217 0.7885
No log 0.9333 28 0.6255 0.1895 0.6255 0.7909
No log 1.0 30 0.6375 0.2201 0.6375 0.7985
No log 1.0667 32 0.5666 0.1008 0.5666 0.7528
No log 1.1333 34 0.5833 0.0815 0.5833 0.7637
No log 1.2 36 0.6911 0.2184 0.6911 0.8314
No log 1.2667 38 0.6013 0.0556 0.6013 0.7754
No log 1.3333 40 0.5894 0.0556 0.5894 0.7677
No log 1.4 42 0.6202 0.1020 0.6202 0.7875
No log 1.4667 44 0.6079 0.0556 0.6079 0.7797
No log 1.5333 46 0.5657 0.1008 0.5657 0.7521
No log 1.6 48 0.5624 0.0476 0.5624 0.7500
No log 1.6667 50 0.5382 0.1008 0.5382 0.7336
No log 1.7333 52 0.5505 0.1407 0.5505 0.7420
No log 1.8 54 0.5360 0.1515 0.5360 0.7321
No log 1.8667 56 0.5506 0.1429 0.5506 0.7420
No log 1.9333 58 0.5845 0.1448 0.5845 0.7645
No log 2.0 60 0.6109 0.0526 0.6109 0.7816
No log 2.0667 62 0.6247 0.1529 0.6247 0.7904
No log 2.1333 64 0.5811 0.1895 0.5811 0.7623
No log 2.2 66 0.6946 0.3023 0.6946 0.8334
No log 2.2667 68 0.7021 0.2626 0.7021 0.8379
No log 2.3333 70 0.6732 0.2189 0.6732 0.8205
No log 2.4 72 0.5593 0.4091 0.5593 0.7478
No log 2.4667 74 0.5674 0.2941 0.5674 0.7533
No log 2.5333 76 0.6960 0.3402 0.6960 0.8343
No log 2.6 78 0.8435 0.1870 0.8435 0.9184
No log 2.6667 80 1.1961 0.1693 1.1961 1.0937
No log 2.7333 82 0.8071 0.2208 0.8071 0.8984
No log 2.8 84 0.6381 0.2157 0.6381 0.7988
No log 2.8667 86 0.8381 0.2681 0.8381 0.9155
No log 2.9333 88 0.6315 0.2222 0.6315 0.7947
No log 3.0 90 0.8999 0.1486 0.8999 0.9486
No log 3.0667 92 1.2449 0.1389 1.2449 1.1157
No log 3.1333 94 1.0013 0.0677 1.0013 1.0007
No log 3.2 96 0.8263 0.2000 0.8263 0.9090
No log 3.2667 98 0.9594 0.0427 0.9594 0.9795
No log 3.3333 100 1.4039 0.0891 1.4039 1.1848
No log 3.4 102 1.5565 0.0732 1.5565 1.2476
No log 3.4667 104 1.2232 0.1367 1.2232 1.1060
No log 3.5333 106 0.7894 0.3043 0.7894 0.8885
No log 3.6 108 0.8273 0.3496 0.8273 0.9096
No log 3.6667 110 1.1843 0.1304 1.1843 1.0882
No log 3.7333 112 1.2679 0.0872 1.2679 1.1260
No log 3.8 114 1.5132 0.0943 1.5132 1.2301
No log 3.8667 116 1.3942 0.1111 1.3942 1.1808
No log 3.9333 118 1.0093 0.1655 1.0093 1.0046
No log 4.0 120 0.9533 0.1939 0.9533 0.9763
No log 4.0667 122 0.8747 0.2581 0.8747 0.9352
No log 4.1333 124 0.7696 0.2661 0.7696 0.8773
No log 4.2 126 0.8337 0.2432 0.8337 0.9130
No log 4.2667 128 0.7974 0.2607 0.7974 0.8930
No log 4.3333 130 0.7150 0.2637 0.7150 0.8456
No log 4.4 132 0.6350 0.3478 0.6350 0.7968
No log 4.4667 134 0.6937 0.2965 0.6937 0.8329
No log 4.5333 136 0.8855 0.1148 0.8855 0.9410
No log 4.6 138 0.8278 0.1416 0.8278 0.9098
No log 4.6667 140 0.6691 0.3131 0.6691 0.8180
No log 4.7333 142 0.6777 0.3103 0.6777 0.8232
No log 4.8 144 0.7845 0.2150 0.7845 0.8857
No log 4.8667 146 0.9391 0.1579 0.9391 0.9691
No log 4.9333 148 1.0481 0.1601 1.0481 1.0238
No log 5.0 150 0.8745 0.1235 0.8745 0.9351
No log 5.0667 152 0.6263 0.3231 0.6263 0.7914
No log 5.1333 154 0.6145 0.3939 0.6145 0.7839
No log 5.2 156 0.7633 0.2143 0.7633 0.8737
No log 5.2667 158 0.9994 0.1882 0.9994 0.9997
No log 5.3333 160 0.8366 0.2140 0.8366 0.9147
No log 5.4 162 0.6509 0.3535 0.6509 0.8068
No log 5.4667 164 0.6577 0.3433 0.6577 0.8110
No log 5.5333 166 0.7429 0.2762 0.7429 0.8619
No log 5.6 168 0.9928 0.0598 0.9928 0.9964
No log 5.6667 170 1.0396 0.0958 1.0396 1.0196
No log 5.7333 172 0.8052 0.2536 0.8052 0.8973
No log 5.8 174 0.7298 0.3143 0.7298 0.8543
No log 5.8667 176 0.8009 0.2233 0.8009 0.8949
No log 5.9333 178 0.9157 0.0871 0.9157 0.9569
No log 6.0 180 0.8822 0.1504 0.8822 0.9393
No log 6.0667 182 0.8586 0.2233 0.8586 0.9266
No log 6.1333 184 0.7923 0.2000 0.7923 0.8901
No log 6.2 186 0.7860 0.2153 0.7860 0.8866
No log 6.2667 188 0.9448 0.1562 0.9448 0.9720
No log 6.3333 190 1.1101 0.1317 1.1101 1.0536
No log 6.4 192 0.9768 0.1880 0.9768 0.9883
No log 6.4667 194 0.7415 0.2881 0.7415 0.8611
No log 6.5333 196 0.6107 0.4231 0.6107 0.7815
No log 6.6 198 0.6137 0.4231 0.6137 0.7834
No log 6.6667 200 0.6944 0.3874 0.6944 0.8333
No log 6.7333 202 0.8485 0.2129 0.8485 0.9212
No log 6.8 204 0.9104 0.2184 0.9104 0.9542
No log 6.8667 206 0.7548 0.2523 0.7548 0.8688
No log 6.9333 208 0.6361 0.3103 0.6361 0.7976
No log 7.0 210 0.5726 0.4573 0.5726 0.7567
No log 7.0667 212 0.5609 0.4573 0.5609 0.7490
No log 7.1333 214 0.5975 0.3684 0.5975 0.7730
No log 7.2 216 0.7122 0.2637 0.7122 0.8439
No log 7.2667 218 0.7380 0.2637 0.7380 0.8591
No log 7.3333 220 0.6516 0.2563 0.6516 0.8072
No log 7.4 222 0.6018 0.3769 0.6018 0.7757
No log 7.4667 224 0.6003 0.3769 0.6003 0.7748
No log 7.5333 226 0.6435 0.2727 0.6435 0.8022
No log 7.6 228 0.6670 0.2780 0.6670 0.8167
No log 7.6667 230 0.6535 0.2727 0.6535 0.8084
No log 7.7333 232 0.6667 0.2709 0.6667 0.8165
No log 7.8 234 0.7118 0.2871 0.7118 0.8437
No log 7.8667 236 0.7734 0.2536 0.7734 0.8794
No log 7.9333 238 0.7652 0.2536 0.7652 0.8747
No log 8.0 240 0.7112 0.2871 0.7112 0.8433
No log 8.0667 242 0.6623 0.2917 0.6623 0.8138
No log 8.1333 244 0.6557 0.2917 0.6557 0.8097
No log 8.2 246 0.6510 0.3016 0.6510 0.8069
No log 8.2667 248 0.6327 0.3016 0.6327 0.7954
No log 8.3333 250 0.6653 0.3016 0.6653 0.8157
No log 8.4 252 0.6734 0.3016 0.6734 0.8206
No log 8.4667 254 0.7075 0.2941 0.7075 0.8411
No log 8.5333 256 0.7821 0.2897 0.7821 0.8844
No log 8.6 258 0.8005 0.2897 0.8005 0.8947
No log 8.6667 260 0.7820 0.2897 0.7820 0.8843
No log 8.7333 262 0.7233 0.28 0.7233 0.8505
No log 8.8 264 0.6689 0.2821 0.6689 0.8178
No log 8.8667 266 0.6627 0.3131 0.6627 0.8141
No log 8.9333 268 0.6664 0.3131 0.6664 0.8163
No log 9.0 270 0.7020 0.2821 0.7020 0.8378
No log 9.0667 272 0.7166 0.28 0.7166 0.8465
No log 9.1333 274 0.7172 0.28 0.7172 0.8469
No log 9.2 276 0.7136 0.28 0.7136 0.8447
No log 9.2667 278 0.7277 0.2780 0.7277 0.8531
No log 9.3333 280 0.7083 0.28 0.7083 0.8416
No log 9.4 282 0.6790 0.3131 0.6790 0.8240
No log 9.4667 284 0.6606 0.3131 0.6606 0.8128
No log 9.5333 286 0.6495 0.3131 0.6495 0.8059
No log 9.6 288 0.6336 0.3131 0.6336 0.7960
No log 9.6667 290 0.6188 0.3131 0.6188 0.7866
No log 9.7333 292 0.6105 0.3161 0.6105 0.7814
No log 9.8 294 0.6048 0.3161 0.6048 0.7777
No log 9.8667 296 0.6026 0.3191 0.6026 0.7763
No log 9.9333 298 0.6037 0.3191 0.6037 0.7770
No log 10.0 300 0.6050 0.3161 0.6050 0.7778

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k6_task3_organization

Finetuned
(4023)
this model