ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k5_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7844
  • Qwk: 0.5139
  • Mse: 0.7844
  • Rmse: 0.8857

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 6.3621 -0.0278 6.3621 2.5223
No log 0.3077 4 4.2937 -0.0283 4.2937 2.0721
No log 0.4615 6 2.8830 0.0148 2.8830 1.6979
No log 0.6154 8 1.9814 0.0383 1.9814 1.4076
No log 0.7692 10 1.3362 0.1036 1.3362 1.1560
No log 0.9231 12 0.9734 0.0912 0.9734 0.9866
No log 1.0769 14 0.9058 0.0340 0.9058 0.9518
No log 1.2308 16 0.9885 -0.0344 0.9885 0.9942
No log 1.3846 18 0.8878 0.0023 0.8878 0.9422
No log 1.5385 20 0.8030 0.0094 0.8030 0.8961
No log 1.6923 22 0.7178 0.1273 0.7178 0.8472
No log 1.8462 24 0.6525 0.2200 0.6525 0.8078
No log 2.0 26 0.6477 0.3310 0.6477 0.8048
No log 2.1538 28 0.6822 0.3798 0.6822 0.8260
No log 2.3077 30 0.7293 0.1633 0.7293 0.8540
No log 2.4615 32 0.7344 0.1751 0.7344 0.8570
No log 2.6154 34 0.7303 0.1751 0.7303 0.8546
No log 2.7692 36 0.7113 0.2333 0.7113 0.8434
No log 2.9231 38 0.6908 0.2412 0.6908 0.8311
No log 3.0769 40 0.6796 0.3012 0.6796 0.8244
No log 3.2308 42 0.6825 0.3294 0.6825 0.8261
No log 3.3846 44 0.7281 0.2547 0.7281 0.8533
No log 3.5385 46 0.7533 0.2852 0.7533 0.8679
No log 3.6923 48 0.6811 0.3230 0.6811 0.8253
No log 3.8462 50 0.5992 0.3969 0.5992 0.7741
No log 4.0 52 0.5650 0.3991 0.5650 0.7517
No log 4.1538 54 0.5854 0.4317 0.5854 0.7651
No log 4.3077 56 0.5893 0.4524 0.5893 0.7677
No log 4.4615 58 0.5575 0.4344 0.5575 0.7467
No log 4.6154 60 0.5375 0.3981 0.5375 0.7332
No log 4.7692 62 0.5373 0.4375 0.5373 0.7330
No log 4.9231 64 0.5724 0.4714 0.5724 0.7566
No log 5.0769 66 0.6115 0.4132 0.6115 0.7820
No log 5.2308 68 0.6651 0.4171 0.6651 0.8155
No log 5.3846 70 0.7229 0.3476 0.7229 0.8502
No log 5.5385 72 0.6914 0.3723 0.6914 0.8315
No log 5.6923 74 0.5880 0.4207 0.5880 0.7668
No log 5.8462 76 0.5281 0.5 0.5281 0.7267
No log 6.0 78 0.5219 0.4885 0.5219 0.7224
No log 6.1538 80 0.5302 0.4555 0.5302 0.7281
No log 6.3077 82 0.5304 0.4555 0.5304 0.7283
No log 6.4615 84 0.5280 0.4555 0.5280 0.7267
No log 6.6154 86 0.5224 0.4566 0.5224 0.7228
No log 6.7692 88 0.5333 0.4968 0.5333 0.7303
No log 6.9231 90 0.5647 0.5260 0.5647 0.7514
No log 7.0769 92 0.6269 0.5018 0.6269 0.7917
No log 7.2308 94 0.6175 0.5153 0.6175 0.7858
No log 7.3846 96 0.5944 0.5357 0.5944 0.7710
No log 7.5385 98 0.5462 0.5424 0.5462 0.7391
No log 7.6923 100 0.5073 0.5193 0.5073 0.7122
No log 7.8462 102 0.4948 0.5441 0.4948 0.7034
No log 8.0 104 0.4960 0.5330 0.4960 0.7042
No log 8.1538 106 0.4977 0.5330 0.4977 0.7055
No log 8.3077 108 0.5104 0.5446 0.5104 0.7145
No log 8.4615 110 0.5309 0.5612 0.5309 0.7286
No log 8.6154 112 0.5436 0.5427 0.5436 0.7373
No log 8.7692 114 0.5618 0.5373 0.5618 0.7495
No log 8.9231 116 0.5671 0.5264 0.5671 0.7531
No log 9.0769 118 0.5658 0.5311 0.5658 0.7522
No log 9.2308 120 0.5572 0.5243 0.5572 0.7465
No log 9.3846 122 0.5574 0.5577 0.5574 0.7466
No log 9.5385 124 0.5407 0.5770 0.5407 0.7353
No log 9.6923 126 0.5393 0.5930 0.5393 0.7343
No log 9.8462 128 0.5564 0.5574 0.5564 0.7459
No log 10.0 130 0.5519 0.5692 0.5519 0.7429
No log 10.1538 132 0.5398 0.5694 0.5398 0.7347
No log 10.3077 134 0.5329 0.5333 0.5329 0.7300
No log 10.4615 136 0.5356 0.5449 0.5356 0.7318
No log 10.6154 138 0.5396 0.5586 0.5396 0.7346
No log 10.7692 140 0.5460 0.5649 0.5460 0.7389
No log 10.9231 142 0.5691 0.5607 0.5691 0.7544
No log 11.0769 144 0.6098 0.5523 0.6098 0.7809
No log 11.2308 146 0.6153 0.5609 0.6153 0.7844
No log 11.3846 148 0.5861 0.5707 0.5861 0.7656
No log 11.5385 150 0.5606 0.5504 0.5606 0.7488
No log 11.6923 152 0.5747 0.5403 0.5747 0.7581
No log 11.8462 154 0.5895 0.5408 0.5895 0.7678
No log 12.0 156 0.5823 0.5220 0.5823 0.7631
No log 12.1538 158 0.5629 0.5689 0.5629 0.7502
No log 12.3077 160 0.5564 0.5895 0.5564 0.7460
No log 12.4615 162 0.5592 0.5749 0.5592 0.7478
No log 12.6154 164 0.5560 0.5814 0.5560 0.7457
No log 12.7692 166 0.5553 0.6010 0.5553 0.7452
No log 12.9231 168 0.5625 0.5655 0.5625 0.7500
No log 13.0769 170 0.5692 0.5741 0.5692 0.7544
No log 13.2308 172 0.5847 0.6005 0.5847 0.7647
No log 13.3846 174 0.5995 0.5865 0.5995 0.7743
No log 13.5385 176 0.6021 0.5807 0.6021 0.7759
No log 13.6923 178 0.5962 0.5251 0.5962 0.7721
No log 13.8462 180 0.6018 0.5308 0.6018 0.7758
No log 14.0 182 0.6099 0.5173 0.6099 0.7810
No log 14.1538 184 0.5976 0.5243 0.5976 0.7730
No log 14.3077 186 0.5940 0.5075 0.5940 0.7707
No log 14.4615 188 0.6246 0.5960 0.6246 0.7903
No log 14.6154 190 0.6318 0.5452 0.6318 0.7949
No log 14.7692 192 0.6163 0.5616 0.6163 0.7851
No log 14.9231 194 0.5961 0.5778 0.5961 0.7721
No log 15.0769 196 0.5983 0.5381 0.5983 0.7735
No log 15.2308 198 0.6075 0.5461 0.6075 0.7794
No log 15.3846 200 0.6308 0.5246 0.6308 0.7942
No log 15.5385 202 0.6703 0.5018 0.6703 0.8187
No log 15.6923 204 0.6880 0.5046 0.6880 0.8295
No log 15.8462 206 0.6542 0.5049 0.6542 0.8088
No log 16.0 208 0.6319 0.5579 0.6319 0.7949
No log 16.1538 210 0.6399 0.5214 0.6399 0.8000
No log 16.3077 212 0.6499 0.5248 0.6499 0.8062
No log 16.4615 214 0.6552 0.5120 0.6552 0.8094
No log 16.6154 216 0.6504 0.5639 0.6504 0.8064
No log 16.7692 218 0.6460 0.5289 0.6460 0.8037
No log 16.9231 220 0.6544 0.5569 0.6544 0.8089
No log 17.0769 222 0.6566 0.5509 0.6566 0.8103
No log 17.2308 224 0.6439 0.5475 0.6439 0.8024
No log 17.3846 226 0.6425 0.5364 0.6425 0.8016
No log 17.5385 228 0.6530 0.5297 0.6530 0.8081
No log 17.6923 230 0.6537 0.5649 0.6537 0.8085
No log 17.8462 232 0.6505 0.5248 0.6505 0.8065
No log 18.0 234 0.6500 0.5394 0.6500 0.8062
No log 18.1538 236 0.6595 0.5509 0.6595 0.8121
No log 18.3077 238 0.6512 0.5416 0.6512 0.8070
No log 18.4615 240 0.6356 0.5211 0.6356 0.7972
No log 18.6154 242 0.6380 0.5129 0.6380 0.7988
No log 18.7692 244 0.6483 0.5234 0.6483 0.8051
No log 18.9231 246 0.6559 0.5552 0.6559 0.8099
No log 19.0769 248 0.6581 0.5552 0.6581 0.8112
No log 19.2308 250 0.6665 0.5392 0.6665 0.8164
No log 19.3846 252 0.6697 0.5258 0.6697 0.8184
No log 19.5385 254 0.6790 0.5341 0.6790 0.8240
No log 19.6923 256 0.6716 0.5397 0.6716 0.8195
No log 19.8462 258 0.6572 0.5319 0.6572 0.8107
No log 20.0 260 0.6592 0.5634 0.6592 0.8119
No log 20.1538 262 0.6674 0.5729 0.6674 0.8169
No log 20.3077 264 0.6778 0.5675 0.6778 0.8233
No log 20.4615 266 0.6870 0.5830 0.6870 0.8289
No log 20.6154 268 0.6991 0.5798 0.6991 0.8361
No log 20.7692 270 0.7070 0.5853 0.7070 0.8408
No log 20.9231 272 0.7096 0.5853 0.7096 0.8424
No log 21.0769 274 0.7078 0.5550 0.7078 0.8413
No log 21.2308 276 0.7038 0.5248 0.7038 0.8389
No log 21.3846 278 0.7035 0.5475 0.7035 0.8387
No log 21.5385 280 0.7191 0.5530 0.7191 0.8480
No log 21.6923 282 0.7436 0.5461 0.7436 0.8623
No log 21.8462 284 0.7330 0.5422 0.7330 0.8562
No log 22.0 286 0.6992 0.5519 0.6992 0.8362
No log 22.1538 288 0.6847 0.5635 0.6847 0.8275
No log 22.3077 290 0.6924 0.5594 0.6924 0.8321
No log 22.4615 292 0.6932 0.5594 0.6932 0.8326
No log 22.6154 294 0.6936 0.5594 0.6936 0.8328
No log 22.7692 296 0.6979 0.5481 0.6979 0.8354
No log 22.9231 298 0.7031 0.5367 0.7031 0.8385
No log 23.0769 300 0.7163 0.5092 0.7163 0.8464
No log 23.2308 302 0.7417 0.4896 0.7417 0.8612
No log 23.3846 304 0.7555 0.4907 0.7555 0.8692
No log 23.5385 306 0.7574 0.4804 0.7574 0.8703
No log 23.6923 308 0.7555 0.5035 0.7555 0.8692
No log 23.8462 310 0.7587 0.5235 0.7587 0.8710
No log 24.0 312 0.7599 0.5315 0.7599 0.8717
No log 24.1538 314 0.7554 0.5522 0.7554 0.8691
No log 24.3077 316 0.7587 0.5452 0.7587 0.8710
No log 24.4615 318 0.7699 0.5301 0.7699 0.8774
No log 24.6154 320 0.7688 0.5312 0.7688 0.8768
No log 24.7692 322 0.7664 0.5003 0.7664 0.8755
No log 24.9231 324 0.7711 0.5058 0.7711 0.8781
No log 25.0769 326 0.7749 0.5297 0.7749 0.8803
No log 25.2308 328 0.7793 0.5600 0.7793 0.8828
No log 25.3846 330 0.7632 0.5218 0.7632 0.8736
No log 25.5385 332 0.7507 0.5388 0.7507 0.8664
No log 25.6923 334 0.7426 0.5469 0.7426 0.8617
No log 25.8462 336 0.7585 0.5632 0.7585 0.8709
No log 26.0 338 0.7659 0.5526 0.7659 0.8752
No log 26.1538 340 0.7613 0.5169 0.7613 0.8725
No log 26.3077 342 0.7604 0.4919 0.7604 0.8720
No log 26.4615 344 0.7687 0.4709 0.7687 0.8767
No log 26.6154 346 0.7730 0.5182 0.7730 0.8792
No log 26.7692 348 0.7598 0.5420 0.7598 0.8717
No log 26.9231 350 0.7513 0.5521 0.7513 0.8668
No log 27.0769 352 0.7578 0.5683 0.7578 0.8705
No log 27.2308 354 0.7626 0.5750 0.7626 0.8733
No log 27.3846 356 0.7637 0.5550 0.7637 0.8739
No log 27.5385 358 0.7791 0.5264 0.7791 0.8827
No log 27.6923 360 0.7976 0.5235 0.7976 0.8931
No log 27.8462 362 0.7959 0.5351 0.7959 0.8921
No log 28.0 364 0.7896 0.5281 0.7896 0.8886
No log 28.1538 366 0.7947 0.5026 0.7947 0.8915
No log 28.3077 368 0.7937 0.5043 0.7937 0.8909
No log 28.4615 370 0.8050 0.4941 0.8050 0.8972
No log 28.6154 372 0.7988 0.4941 0.7988 0.8937
No log 28.7692 374 0.7909 0.5183 0.7909 0.8893
No log 28.9231 376 0.8011 0.5377 0.8011 0.8950
No log 29.0769 378 0.8099 0.5420 0.8099 0.9000
No log 29.2308 380 0.8074 0.5304 0.8074 0.8986
No log 29.3846 382 0.8141 0.5304 0.8141 0.9023
No log 29.5385 384 0.8227 0.5201 0.8227 0.9070
No log 29.6923 386 0.8502 0.4952 0.8502 0.9221
No log 29.8462 388 0.8526 0.5151 0.8526 0.9234
No log 30.0 390 0.8234 0.5275 0.8234 0.9074
No log 30.1538 392 0.8091 0.5327 0.8091 0.8995
No log 30.3077 394 0.8037 0.5327 0.8037 0.8965
No log 30.4615 396 0.7973 0.5362 0.7973 0.8929
No log 30.6154 398 0.7852 0.5481 0.7852 0.8861
No log 30.7692 400 0.7677 0.5120 0.7677 0.8762
No log 30.9231 402 0.7539 0.5120 0.7539 0.8683
No log 31.0769 404 0.7556 0.5124 0.7556 0.8693
No log 31.2308 406 0.7528 0.5532 0.7528 0.8676
No log 31.3846 408 0.7322 0.5344 0.7322 0.8557
No log 31.5385 410 0.7217 0.5437 0.7217 0.8495
No log 31.6923 412 0.7153 0.5379 0.7153 0.8458
No log 31.8462 414 0.7131 0.5436 0.7131 0.8445
No log 32.0 416 0.7160 0.5493 0.7160 0.8462
No log 32.1538 418 0.7170 0.5548 0.7170 0.8467
No log 32.3077 420 0.7169 0.5548 0.7169 0.8467
No log 32.4615 422 0.7228 0.5521 0.7228 0.8502
No log 32.6154 424 0.7318 0.5494 0.7318 0.8554
No log 32.7692 426 0.7378 0.5766 0.7378 0.8589
No log 32.9231 428 0.7508 0.5439 0.7508 0.8665
No log 33.0769 430 0.7566 0.5123 0.7566 0.8698
No log 33.2308 432 0.7729 0.5123 0.7729 0.8791
No log 33.3846 434 0.7990 0.5299 0.7990 0.8939
No log 33.5385 436 0.8334 0.5265 0.8334 0.9129
No log 33.6923 438 0.8503 0.5174 0.8503 0.9221
No log 33.8462 440 0.8331 0.5248 0.8331 0.9127
No log 34.0 442 0.8093 0.5623 0.8093 0.8996
No log 34.1538 444 0.8002 0.5586 0.8002 0.8946
No log 34.3077 446 0.7969 0.5559 0.7969 0.8927
No log 34.4615 448 0.7972 0.5547 0.7972 0.8928
No log 34.6154 450 0.7996 0.5620 0.7996 0.8942
No log 34.7692 452 0.7937 0.5670 0.7937 0.8909
No log 34.9231 454 0.7709 0.5546 0.7709 0.8780
No log 35.0769 456 0.7627 0.5546 0.7627 0.8734
No log 35.2308 458 0.7542 0.5546 0.7542 0.8684
No log 35.3846 460 0.7552 0.5546 0.7552 0.8690
No log 35.5385 462 0.7661 0.5546 0.7661 0.8753
No log 35.6923 464 0.7649 0.5470 0.7649 0.8746
No log 35.8462 466 0.7702 0.5434 0.7702 0.8776
No log 36.0 468 0.7819 0.5546 0.7819 0.8843
No log 36.1538 470 0.7875 0.5546 0.7875 0.8874
No log 36.3077 472 0.7780 0.5385 0.7780 0.8821
No log 36.4615 474 0.7723 0.5286 0.7723 0.8788
No log 36.6154 476 0.7728 0.5437 0.7728 0.8791
No log 36.7692 478 0.7864 0.5519 0.7864 0.8868
No log 36.9231 480 0.8208 0.5440 0.8208 0.9060
No log 37.0769 482 0.8157 0.5519 0.8157 0.9032
No log 37.2308 484 0.8034 0.5472 0.8034 0.8963
No log 37.3846 486 0.7869 0.5447 0.7869 0.8871
No log 37.5385 488 0.7872 0.5559 0.7872 0.8872
No log 37.6923 490 0.7812 0.5419 0.7812 0.8838
No log 37.8462 492 0.7937 0.5336 0.7937 0.8909
No log 38.0 494 0.8051 0.5235 0.8051 0.8973
No log 38.1538 496 0.8061 0.5304 0.8061 0.8978
No log 38.3077 498 0.7890 0.5322 0.7890 0.8883
0.5165 38.4615 500 0.7793 0.5384 0.7793 0.8828
0.5165 38.6154 502 0.7826 0.5447 0.7826 0.8846
0.5165 38.7692 504 0.7840 0.5546 0.7840 0.8855
0.5165 38.9231 506 0.7782 0.5382 0.7782 0.8822
0.5165 39.0769 508 0.7729 0.5324 0.7729 0.8792
0.5165 39.2308 510 0.7754 0.4893 0.7754 0.8806
0.5165 39.3846 512 0.7782 0.5061 0.7782 0.8822
0.5165 39.5385 514 0.7844 0.5139 0.7844 0.8857

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run1_AugV5_k5_task2_organization

Finetuned
(4023)
this model