ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k11_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7097
  • Qwk: 0.5344
  • Mse: 0.7097
  • Rmse: 0.8425

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 6.3367 -0.0230 6.3367 2.5173
No log 0.1429 4 4.2674 -0.0201 4.2674 2.0658
No log 0.2143 6 2.9687 -0.0190 2.9687 1.7230
No log 0.2857 8 2.1094 -0.0042 2.1094 1.4524
No log 0.3571 10 1.5889 -0.0188 1.5889 1.2605
No log 0.4286 12 1.4950 -0.0756 1.4950 1.2227
No log 0.5 14 1.6369 -0.0910 1.6369 1.2794
No log 0.5714 16 1.5454 -0.1002 1.5454 1.2431
No log 0.6429 18 1.2052 0.0194 1.2052 1.0978
No log 0.7143 20 0.9291 0.0808 0.9291 0.9639
No log 0.7857 22 0.7580 0.2036 0.7580 0.8706
No log 0.8571 24 0.6977 0.2587 0.6977 0.8353
No log 0.9286 26 0.6993 0.2251 0.6993 0.8363
No log 1.0 28 0.7114 0.2333 0.7114 0.8435
No log 1.0714 30 0.7777 0.2172 0.7777 0.8819
No log 1.1429 32 0.8854 0.1664 0.8854 0.9410
No log 1.2143 34 0.9876 0.0683 0.9876 0.9938
No log 1.2857 36 1.0005 0.0921 1.0005 1.0002
No log 1.3571 38 0.9419 0.1249 0.9419 0.9705
No log 1.4286 40 0.8606 0.1985 0.8606 0.9277
No log 1.5 42 0.7876 0.1827 0.7876 0.8875
No log 1.5714 44 0.7507 0.1868 0.7507 0.8664
No log 1.6429 46 0.7142 0.1863 0.7142 0.8451
No log 1.7143 48 0.6738 0.3176 0.6738 0.8208
No log 1.7857 50 0.6430 0.3538 0.6430 0.8019
No log 1.8571 52 0.6434 0.3024 0.6434 0.8021
No log 1.9286 54 0.7169 0.3356 0.7169 0.8467
No log 2.0 56 0.8847 0.3152 0.8847 0.9406
No log 2.0714 58 0.9221 0.3042 0.9221 0.9603
No log 2.1429 60 0.7942 0.3329 0.7942 0.8912
No log 2.2143 62 0.7232 0.3791 0.7232 0.8504
No log 2.2857 64 0.6634 0.3649 0.6634 0.8145
No log 2.3571 66 0.6503 0.3241 0.6503 0.8064
No log 2.4286 68 0.6281 0.3639 0.6281 0.7925
No log 2.5 70 0.6166 0.3805 0.6166 0.7852
No log 2.5714 72 0.6272 0.4062 0.6272 0.7920
No log 2.6429 74 0.6640 0.3700 0.6640 0.8149
No log 2.7143 76 0.6741 0.3923 0.6741 0.8210
No log 2.7857 78 0.6816 0.3776 0.6816 0.8256
No log 2.8571 80 0.6361 0.4047 0.6361 0.7975
No log 2.9286 82 0.6063 0.4397 0.6063 0.7787
No log 3.0 84 0.5725 0.4122 0.5725 0.7566
No log 3.0714 86 0.5553 0.3973 0.5553 0.7452
No log 3.1429 88 0.5392 0.4140 0.5392 0.7343
No log 3.2143 90 0.5271 0.4595 0.5271 0.7260
No log 3.2857 92 0.5323 0.4011 0.5323 0.7296
No log 3.3571 94 0.5647 0.4677 0.5647 0.7515
No log 3.4286 96 0.6122 0.4823 0.6122 0.7824
No log 3.5 98 0.6603 0.4528 0.6603 0.8126
No log 3.5714 100 0.6685 0.4685 0.6685 0.8176
No log 3.6429 102 0.6231 0.4837 0.6231 0.7894
No log 3.7143 104 0.5716 0.4562 0.5716 0.7560
No log 3.7857 106 0.5521 0.4583 0.5521 0.7431
No log 3.8571 108 0.5230 0.4813 0.5230 0.7232
No log 3.9286 110 0.5176 0.4694 0.5176 0.7195
No log 4.0 112 0.5299 0.5175 0.5299 0.7280
No log 4.0714 114 0.5594 0.4908 0.5594 0.7480
No log 4.1429 116 0.5784 0.5155 0.5784 0.7605
No log 4.2143 118 0.5693 0.5094 0.5693 0.7545
No log 4.2857 120 0.5683 0.4692 0.5683 0.7539
No log 4.3571 122 0.5464 0.4458 0.5464 0.7392
No log 4.4286 124 0.5216 0.4826 0.5216 0.7222
No log 4.5 126 0.5136 0.4895 0.5136 0.7167
No log 4.5714 128 0.5143 0.4653 0.5143 0.7172
No log 4.6429 130 0.5488 0.5096 0.5488 0.7408
No log 4.7143 132 0.5777 0.5011 0.5777 0.7600
No log 4.7857 134 0.5886 0.5285 0.5886 0.7672
No log 4.8571 136 0.6132 0.5010 0.6132 0.7831
No log 4.9286 138 0.6326 0.4909 0.6326 0.7954
No log 5.0 140 0.6257 0.4974 0.6257 0.7910
No log 5.0714 142 0.6038 0.5225 0.6038 0.7770
No log 5.1429 144 0.5827 0.5183 0.5827 0.7633
No log 5.2143 146 0.5634 0.4823 0.5634 0.7506
No log 5.2857 148 0.5527 0.5586 0.5527 0.7434
No log 5.3571 150 0.5422 0.5918 0.5422 0.7364
No log 5.4286 152 0.5389 0.5622 0.5389 0.7341
No log 5.5 154 0.5581 0.5675 0.5581 0.7471
No log 5.5714 156 0.5821 0.5643 0.5821 0.7629
No log 5.6429 158 0.5833 0.5643 0.5833 0.7637
No log 5.7143 160 0.5836 0.5538 0.5836 0.7639
No log 5.7857 162 0.5835 0.5505 0.5835 0.7638
No log 5.8571 164 0.5842 0.5673 0.5842 0.7644
No log 5.9286 166 0.5805 0.5656 0.5805 0.7619
No log 6.0 168 0.5698 0.5686 0.5698 0.7548
No log 6.0714 170 0.5684 0.5367 0.5684 0.7539
No log 6.1429 172 0.5787 0.5519 0.5787 0.7607
No log 6.2143 174 0.5769 0.5756 0.5769 0.7595
No log 6.2857 176 0.5752 0.5603 0.5752 0.7584
No log 6.3571 178 0.5794 0.5846 0.5794 0.7612
No log 6.4286 180 0.5790 0.5765 0.5790 0.7609
No log 6.5 182 0.5811 0.5343 0.5811 0.7623
No log 6.5714 184 0.5988 0.5347 0.5988 0.7739
No log 6.6429 186 0.6061 0.5395 0.6061 0.7785
No log 6.7143 188 0.6017 0.5367 0.6017 0.7757
No log 6.7857 190 0.6023 0.5354 0.6023 0.7761
No log 6.8571 192 0.5981 0.5585 0.5981 0.7734
No log 6.9286 194 0.5982 0.5882 0.5982 0.7735
No log 7.0 196 0.5850 0.5556 0.5850 0.7648
No log 7.0714 198 0.5750 0.5406 0.5750 0.7583
No log 7.1429 200 0.6012 0.5272 0.6012 0.7754
No log 7.2143 202 0.6220 0.5234 0.6220 0.7886
No log 7.2857 204 0.6264 0.5466 0.6264 0.7915
No log 7.3571 206 0.6311 0.5465 0.6311 0.7944
No log 7.4286 208 0.6290 0.5585 0.6290 0.7931
No log 7.5 210 0.6239 0.5582 0.6239 0.7898
No log 7.5714 212 0.6177 0.5746 0.6177 0.7859
No log 7.6429 214 0.6152 0.5746 0.6152 0.7844
No log 7.7143 216 0.6020 0.5428 0.6020 0.7759
No log 7.7857 218 0.5929 0.5569 0.5929 0.7700
No log 7.8571 220 0.5930 0.5649 0.5930 0.7700
No log 7.9286 222 0.5983 0.5265 0.5983 0.7735
No log 8.0 224 0.6065 0.5564 0.6065 0.7788
No log 8.0714 226 0.6300 0.5624 0.6300 0.7937
No log 8.1429 228 0.6396 0.5785 0.6396 0.7997
No log 8.2143 230 0.6261 0.5979 0.6261 0.7913
No log 8.2857 232 0.6056 0.5519 0.6056 0.7782
No log 8.3571 234 0.5993 0.5 0.5993 0.7742
No log 8.4286 236 0.6034 0.5378 0.6034 0.7768
No log 8.5 238 0.6331 0.5444 0.6331 0.7957
No log 8.5714 240 0.6441 0.5769 0.6441 0.8026
No log 8.6429 242 0.6237 0.5676 0.6237 0.7898
No log 8.7143 244 0.6116 0.5628 0.6116 0.7820
No log 8.7857 246 0.6080 0.5733 0.6080 0.7797
No log 8.8571 248 0.6156 0.5823 0.6156 0.7846
No log 8.9286 250 0.6070 0.5970 0.6070 0.7791
No log 9.0 252 0.6104 0.5501 0.6104 0.7813
No log 9.0714 254 0.6263 0.5662 0.6263 0.7914
No log 9.1429 256 0.6482 0.5415 0.6482 0.8051
No log 9.2143 258 0.6543 0.5198 0.6543 0.8089
No log 9.2857 260 0.6397 0.4929 0.6397 0.7998
No log 9.3571 262 0.6277 0.5804 0.6277 0.7923
No log 9.4286 264 0.6435 0.5477 0.6435 0.8022
No log 9.5 266 0.6412 0.5477 0.6412 0.8008
No log 9.5714 268 0.6334 0.6016 0.6334 0.7958
No log 9.6429 270 0.6431 0.5269 0.6431 0.8019
No log 9.7143 272 0.6736 0.5288 0.6736 0.8207
No log 9.7857 274 0.6947 0.4872 0.6947 0.8335
No log 9.8571 276 0.6998 0.4823 0.6998 0.8366
No log 9.9286 278 0.6640 0.5012 0.6640 0.8148
No log 10.0 280 0.6517 0.4997 0.6517 0.8073
No log 10.0714 282 0.6600 0.5413 0.6600 0.8124
No log 10.1429 284 0.6779 0.5518 0.6779 0.8233
No log 10.2143 286 0.6911 0.5690 0.6911 0.8313
No log 10.2857 288 0.7020 0.5501 0.7020 0.8379
No log 10.3571 290 0.6899 0.5277 0.6899 0.8306
No log 10.4286 292 0.6732 0.4985 0.6732 0.8205
No log 10.5 294 0.6608 0.5166 0.6608 0.8129
No log 10.5714 296 0.6600 0.5278 0.6600 0.8124
No log 10.6429 298 0.6446 0.5948 0.6446 0.8029
No log 10.7143 300 0.6332 0.5461 0.6332 0.7958
No log 10.7857 302 0.6316 0.5476 0.6316 0.7947
No log 10.8571 304 0.6332 0.5519 0.6332 0.7957
No log 10.9286 306 0.6581 0.5444 0.6581 0.8112
No log 11.0 308 0.6629 0.5575 0.6629 0.8142
No log 11.0714 310 0.6825 0.5134 0.6825 0.8261
No log 11.1429 312 0.7016 0.5359 0.7016 0.8376
No log 11.2143 314 0.7051 0.5296 0.7051 0.8397
No log 11.2857 316 0.6923 0.5604 0.6923 0.8320
No log 11.3571 318 0.6835 0.5461 0.6835 0.8267
No log 11.4286 320 0.6777 0.5405 0.6777 0.8232
No log 11.5 322 0.6709 0.5636 0.6709 0.8191
No log 11.5714 324 0.6757 0.5435 0.6757 0.8220
No log 11.6429 326 0.6983 0.5445 0.6983 0.8357
No log 11.7143 328 0.7124 0.5360 0.7124 0.8440
No log 11.7857 330 0.7362 0.5363 0.7362 0.8580
No log 11.8571 332 0.7263 0.5388 0.7263 0.8522
No log 11.9286 334 0.6796 0.5297 0.6796 0.8244
No log 12.0 336 0.6463 0.5602 0.6463 0.8039
No log 12.0714 338 0.6262 0.5592 0.6262 0.7913
No log 12.1429 340 0.6392 0.5752 0.6392 0.7995
No log 12.2143 342 0.6806 0.5265 0.6806 0.8250
No log 12.2857 344 0.7115 0.5293 0.7115 0.8435
No log 12.3571 346 0.7069 0.5261 0.7069 0.8408
No log 12.4286 348 0.6844 0.5530 0.6844 0.8273
No log 12.5 350 0.6771 0.5545 0.6771 0.8228
No log 12.5714 352 0.6667 0.5692 0.6667 0.8165
No log 12.6429 354 0.6626 0.5627 0.6626 0.8140
No log 12.7143 356 0.6584 0.5695 0.6584 0.8114
No log 12.7857 358 0.6541 0.5655 0.6541 0.8088
No log 12.8571 360 0.6569 0.5586 0.6569 0.8105
No log 12.9286 362 0.6494 0.5491 0.6494 0.8058
No log 13.0 364 0.6686 0.5651 0.6686 0.8177
No log 13.0714 366 0.6857 0.5584 0.6857 0.8281
No log 13.1429 368 0.6834 0.5530 0.6834 0.8267
No log 13.2143 370 0.6753 0.5780 0.6753 0.8217
No log 13.2857 372 0.6651 0.5785 0.6651 0.8155
No log 13.3571 374 0.6689 0.5895 0.6689 0.8178
No log 13.4286 376 0.6778 0.5764 0.6778 0.8233
No log 13.5 378 0.6991 0.5444 0.6991 0.8361
No log 13.5714 380 0.7189 0.5402 0.7189 0.8479
No log 13.6429 382 0.7324 0.5556 0.7324 0.8558
No log 13.7143 384 0.7306 0.5450 0.7306 0.8548
No log 13.7857 386 0.7116 0.5466 0.7116 0.8436
No log 13.8571 388 0.7060 0.5532 0.7060 0.8402
No log 13.9286 390 0.7072 0.5584 0.7072 0.8410
No log 14.0 392 0.6918 0.5503 0.6918 0.8318
No log 14.0714 394 0.6732 0.5558 0.6732 0.8205
No log 14.1429 396 0.6629 0.5327 0.6629 0.8142
No log 14.2143 398 0.6457 0.5118 0.6457 0.8036
No log 14.2857 400 0.6590 0.5259 0.6590 0.8118
No log 14.3571 402 0.6661 0.5530 0.6661 0.8162
No log 14.4286 404 0.6740 0.5732 0.6740 0.8210
No log 14.5 406 0.6889 0.5368 0.6889 0.8300
No log 14.5714 408 0.6983 0.5480 0.6983 0.8356
No log 14.6429 410 0.7195 0.5377 0.7195 0.8483
No log 14.7143 412 0.7291 0.5235 0.7291 0.8539
No log 14.7857 414 0.7284 0.5348 0.7284 0.8535
No log 14.8571 416 0.7220 0.5664 0.7220 0.8497
No log 14.9286 418 0.7505 0.5277 0.7505 0.8663
No log 15.0 420 0.7899 0.4901 0.7899 0.8888
No log 15.0714 422 0.7942 0.4976 0.7942 0.8912
No log 15.1429 424 0.7569 0.4931 0.7569 0.8700
No log 15.2143 426 0.7472 0.5023 0.7472 0.8644
No log 15.2857 428 0.7620 0.4948 0.7620 0.8729
No log 15.3571 430 0.7867 0.5021 0.7867 0.8870
No log 15.4286 432 0.7989 0.5058 0.7989 0.8938
No log 15.5 434 0.7677 0.4871 0.7677 0.8762
No log 15.5714 436 0.7492 0.4762 0.7492 0.8656
No log 15.6429 438 0.7262 0.4852 0.7262 0.8522
No log 15.7143 440 0.7299 0.4987 0.7299 0.8543
No log 15.7857 442 0.7359 0.4987 0.7359 0.8579
No log 15.8571 444 0.7529 0.5015 0.7529 0.8677
No log 15.9286 446 0.7372 0.5098 0.7372 0.8586
No log 16.0 448 0.7231 0.5226 0.7231 0.8503
No log 16.0714 450 0.7159 0.5477 0.7159 0.8461
No log 16.1429 452 0.7138 0.5211 0.7138 0.8449
No log 16.2143 454 0.7166 0.5082 0.7166 0.8465
No log 16.2857 456 0.7141 0.5074 0.7141 0.8451
No log 16.3571 458 0.7218 0.5119 0.7218 0.8496
No log 16.4286 460 0.7147 0.5316 0.7147 0.8454
No log 16.5 462 0.6913 0.5216 0.6913 0.8315
No log 16.5714 464 0.6671 0.5307 0.6671 0.8168
No log 16.6429 466 0.6615 0.5668 0.6615 0.8134
No log 16.7143 468 0.6750 0.5747 0.6750 0.8216
No log 16.7857 470 0.7034 0.5253 0.7034 0.8387
No log 16.8571 472 0.7246 0.5060 0.7246 0.8513
No log 16.9286 474 0.7066 0.5220 0.7066 0.8406
No log 17.0 476 0.6991 0.5440 0.6991 0.8361
No log 17.0714 478 0.7051 0.5278 0.7051 0.8397
No log 17.1429 480 0.7108 0.5507 0.7108 0.8431
No log 17.2143 482 0.7143 0.5624 0.7143 0.8451
No log 17.2857 484 0.7154 0.5739 0.7154 0.8458
No log 17.3571 486 0.7184 0.5586 0.7184 0.8476
No log 17.4286 488 0.7141 0.5598 0.7141 0.8451
No log 17.5 490 0.6810 0.5642 0.6810 0.8252
No log 17.5714 492 0.6739 0.5518 0.6739 0.8209
No log 17.6429 494 0.6643 0.5725 0.6643 0.8150
No log 17.7143 496 0.6490 0.5993 0.6490 0.8056
No log 17.7857 498 0.6442 0.5993 0.6442 0.8026
0.5657 17.8571 500 0.6526 0.5631 0.6526 0.8078
0.5657 17.9286 502 0.6928 0.5244 0.6928 0.8324
0.5657 18.0 504 0.7058 0.5229 0.7058 0.8401
0.5657 18.0714 506 0.6865 0.5309 0.6865 0.8286
0.5657 18.1429 508 0.6794 0.5501 0.6794 0.8242
0.5657 18.2143 510 0.7097 0.5344 0.7097 0.8425

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k11_task2_organization

Finetuned
(4023)
this model