ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k5_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8622
  • Qwk: 0.2381
  • Mse: 0.8622
  • Rmse: 0.9285

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0645 2 3.2191 -0.0138 3.2191 1.7942
No log 0.1290 4 1.7386 -0.0070 1.7386 1.3186
No log 0.1935 6 1.1691 0.0294 1.1691 1.0813
No log 0.2581 8 1.2552 0.0 1.2552 1.1203
No log 0.3226 10 1.3894 0.0 1.3894 1.1787
No log 0.3871 12 1.1443 0.0 1.1443 1.0697
No log 0.4516 14 1.0472 0.0 1.0472 1.0233
No log 0.5161 16 0.7757 -0.0547 0.7757 0.8807
No log 0.5806 18 0.7600 0.0359 0.7600 0.8718
No log 0.6452 20 0.8267 0.0991 0.8267 0.9092
No log 0.7097 22 0.8137 0.1111 0.8137 0.9020
No log 0.7742 24 0.9318 0.0201 0.9318 0.9653
No log 0.8387 26 0.8773 0.0380 0.8773 0.9366
No log 0.9032 28 0.8646 0.0526 0.8646 0.9298
No log 0.9677 30 0.8028 -0.1073 0.8028 0.8960
No log 1.0323 32 0.8444 -0.1217 0.8444 0.9189
No log 1.0968 34 0.9526 0.0617 0.9526 0.9760
No log 1.1613 36 1.1076 0.0388 1.1076 1.0524
No log 1.2258 38 1.0864 0.0345 1.0864 1.0423
No log 1.2903 40 0.8545 0.0717 0.8545 0.9244
No log 1.3548 42 0.7772 0.0857 0.7772 0.8816
No log 1.4194 44 0.6910 0.0256 0.6910 0.8312
No log 1.4839 46 0.6457 0.0400 0.6457 0.8035
No log 1.5484 48 0.6853 0.0061 0.6853 0.8278
No log 1.6129 50 0.7348 0.0877 0.7348 0.8572
No log 1.6774 52 1.0232 0.2397 1.0232 1.0115
No log 1.7419 54 0.8267 0.1691 0.8267 0.9092
No log 1.8065 56 0.6346 0.0222 0.6346 0.7966
No log 1.8710 58 0.6989 -0.1515 0.6989 0.8360
No log 1.9355 60 0.6905 -0.1473 0.6905 0.8310
No log 2.0 62 0.6234 -0.1556 0.6234 0.7896
No log 2.0645 64 0.7015 0.2258 0.7015 0.8375
No log 2.1290 66 1.3157 0.0903 1.3157 1.1471
No log 2.1935 68 1.3669 0.0614 1.3669 1.1691
No log 2.2581 70 0.9912 0.2066 0.9912 0.9956
No log 2.3226 72 0.6649 0.2593 0.6649 0.8154
No log 2.3871 74 0.5977 0.1707 0.5977 0.7731
No log 2.4516 76 0.6064 0.2105 0.6064 0.7787
No log 2.5161 78 0.6180 0.1902 0.6180 0.7861
No log 2.5806 80 0.6577 0.2000 0.6577 0.8110
No log 2.6452 82 0.7030 0.2323 0.7030 0.8385
No log 2.7097 84 0.7140 0.2340 0.7140 0.8450
No log 2.7742 86 0.7482 0.2086 0.7482 0.8650
No log 2.8387 88 0.7046 0.3433 0.7046 0.8394
No log 2.9032 90 0.8478 0.2000 0.8478 0.9208
No log 2.9677 92 0.9828 0.2527 0.9828 0.9914
No log 3.0323 94 0.6721 0.2579 0.6721 0.8198
No log 3.0968 96 0.5711 0.3161 0.5711 0.7557
No log 3.1613 98 0.5300 0.3043 0.5300 0.7280
No log 3.2258 100 0.5343 0.3231 0.5343 0.7309
No log 3.2903 102 0.5304 0.3433 0.5304 0.7283
No log 3.3548 104 0.5680 0.3905 0.5680 0.7537
No log 3.4194 106 0.8299 0.1858 0.8299 0.9110
No log 3.4839 108 0.9447 0.0988 0.9447 0.9719
No log 3.5484 110 0.9173 0.0894 0.9173 0.9578
No log 3.6129 112 0.7026 0.2709 0.7026 0.8382
No log 3.6774 114 0.6845 0.28 0.6845 0.8273
No log 3.7419 116 0.8811 0.1724 0.8811 0.9387
No log 3.8065 118 0.7490 0.1925 0.7490 0.8654
No log 3.8710 120 0.6530 0.3118 0.6530 0.8081
No log 3.9355 122 0.8656 0.0151 0.8656 0.9304
No log 4.0 124 0.9055 -0.0196 0.9055 0.9516
No log 4.0645 126 0.7687 0.0497 0.7687 0.8767
No log 4.1290 128 0.6164 0.3623 0.6164 0.7851
No log 4.1935 130 0.6102 0.3398 0.6102 0.7811
No log 4.2581 132 0.6261 0.3665 0.6261 0.7913
No log 4.3226 134 0.7310 0.3271 0.7310 0.8550
No log 4.3871 136 0.6670 0.3116 0.6670 0.8167
No log 4.4516 138 0.6782 0.3208 0.6782 0.8236
No log 4.5161 140 0.7597 0.3303 0.7597 0.8716
No log 4.5806 142 0.8625 0.1730 0.8625 0.9287
No log 4.6452 144 1.1424 0.0747 1.1424 1.0688
No log 4.7097 146 1.2341 0.0435 1.2341 1.1109
No log 4.7742 148 1.2610 0.0183 1.2610 1.1230
No log 4.8387 150 1.0460 0.0598 1.0460 1.0228
No log 4.9032 152 0.7827 0.2137 0.7827 0.8847
No log 4.9677 154 0.6576 0.2593 0.6576 0.8109
No log 5.0323 156 0.6680 0.2566 0.6680 0.8173
No log 5.0968 158 0.8516 0.1319 0.8516 0.9228
No log 5.1613 160 1.1343 0.0625 1.1343 1.0650
No log 5.2258 162 1.3731 0.0363 1.3731 1.1718
No log 5.2903 164 1.4313 0.0872 1.4313 1.1964
No log 5.3548 166 1.4155 0.0627 1.4155 1.1897
No log 5.4194 168 1.0556 0.1278 1.0556 1.0274
No log 5.4839 170 0.9361 0.1571 0.9361 0.9675
No log 5.5484 172 0.9951 0.1278 0.9951 0.9975
No log 5.6129 174 0.9991 0.1571 0.9991 0.9995
No log 5.6774 176 0.8773 0.2698 0.8773 0.9366
No log 5.7419 178 0.9063 0.2199 0.9063 0.9520
No log 5.8065 180 0.9946 0.1571 0.9946 0.9973
No log 5.8710 182 1.1448 0.1032 1.1448 1.0700
No log 5.9355 184 1.1738 0.1049 1.1738 1.0834
No log 6.0 186 0.8733 0.3255 0.8733 0.9345
No log 6.0645 188 0.7111 0.3770 0.7111 0.8433
No log 6.1290 190 0.6335 0.4386 0.6335 0.7959
No log 6.1935 192 0.6354 0.4678 0.6354 0.7971
No log 6.2581 194 0.6568 0.4050 0.6568 0.8104
No log 6.3226 196 0.8086 0.2868 0.8086 0.8992
No log 6.3871 198 0.9190 0.2062 0.9190 0.9587
No log 6.4516 200 0.8817 0.2062 0.8817 0.9390
No log 6.5161 202 0.7728 0.3613 0.7728 0.8791
No log 6.5806 204 0.6828 0.3846 0.6828 0.8263
No log 6.6452 206 0.6678 0.3571 0.6678 0.8172
No log 6.7097 208 0.7621 0.3613 0.7621 0.8730
No log 6.7742 210 0.8866 0.2698 0.8866 0.9416
No log 6.8387 212 0.8307 0.3333 0.8307 0.9114
No log 6.9032 214 0.8584 0.3306 0.8584 0.9265
No log 6.9677 216 0.9693 0.1562 0.9693 0.9845
No log 7.0323 218 1.0275 0.1264 1.0275 1.0136
No log 7.0968 220 0.9542 0.1818 0.9542 0.9768
No log 7.1613 222 0.8374 0.3333 0.8374 0.9151
No log 7.2258 224 0.7137 0.3665 0.7137 0.8448
No log 7.2903 226 0.6331 0.3665 0.6331 0.7957
No log 7.3548 228 0.6287 0.3333 0.6287 0.7929
No log 7.4194 230 0.6557 0.3761 0.6557 0.8097
No log 7.4839 232 0.7239 0.3665 0.7239 0.8508
No log 7.5484 234 0.7617 0.3684 0.7617 0.8728
No log 7.6129 236 0.8016 0.3153 0.8016 0.8953
No log 7.6774 238 0.7892 0.3422 0.7892 0.8883
No log 7.7419 240 0.7996 0.2511 0.7996 0.8942
No log 7.8065 242 0.7647 0.2982 0.7647 0.8745
No log 7.8710 244 0.7866 0.2793 0.7866 0.8869
No log 7.9355 246 0.8470 0.2137 0.8470 0.9203
No log 8.0 248 0.9113 0.1811 0.9113 0.9546
No log 8.0645 250 0.8883 0.2459 0.8883 0.9425
No log 8.1290 252 0.7801 0.3128 0.7801 0.8833
No log 8.1935 254 0.6758 0.3719 0.6758 0.8221
No log 8.2581 256 0.6379 0.4133 0.6379 0.7987
No log 8.3226 258 0.6506 0.4435 0.6506 0.8066
No log 8.3871 260 0.6714 0.4240 0.6714 0.8194
No log 8.4516 262 0.7337 0.3719 0.7337 0.8565
No log 8.5161 264 0.8042 0.4043 0.8042 0.8968
No log 8.5806 266 0.8765 0.2126 0.8765 0.9362
No log 8.6452 268 0.9347 0.1571 0.9347 0.9668
No log 8.7097 270 0.9328 0.1571 0.9328 0.9658
No log 8.7742 272 0.9070 0.1875 0.9070 0.9524
No log 8.8387 274 0.8560 0.2131 0.8560 0.9252
No log 8.9032 276 0.7888 0.4 0.7888 0.8881
No log 8.9677 278 0.7244 0.3613 0.7244 0.8511
No log 9.0323 280 0.6783 0.4093 0.6783 0.8236
No log 9.0968 282 0.6781 0.4093 0.6781 0.8235
No log 9.1613 284 0.7127 0.4050 0.7127 0.8442
No log 9.2258 286 0.7631 0.3613 0.7631 0.8735
No log 9.2903 288 0.8085 0.2922 0.8085 0.8992
No log 9.3548 290 0.8356 0.2569 0.8356 0.9141
No log 9.4194 292 0.8635 0.2941 0.8635 0.9292
No log 9.4839 294 0.8569 0.2941 0.8569 0.9257
No log 9.5484 296 0.8476 0.2941 0.8476 0.9206
No log 9.6129 298 0.8472 0.2960 0.8472 0.9205
No log 9.6774 300 0.8486 0.2960 0.8486 0.9212
No log 9.7419 302 0.8469 0.2960 0.8469 0.9202
No log 9.8065 304 0.8515 0.2960 0.8515 0.9228
No log 9.8710 306 0.8552 0.2941 0.8552 0.9248
No log 9.9355 308 0.8594 0.2381 0.8594 0.9270
No log 10.0 310 0.8622 0.2381 0.8622 0.9285

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k5_task3_organization

Finetuned
(4023)
this model