ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5853
  • Qwk: 0.7733
  • Mse: 0.5853
  • Rmse: 0.7651

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0263 2 6.7815 0.0239 6.7815 2.6041
No log 0.0526 4 4.4512 0.0855 4.4512 2.1098
No log 0.0789 6 4.0728 -0.0676 4.0728 2.0181
No log 0.1053 8 4.3657 -0.0711 4.3657 2.0894
No log 0.1316 10 3.0382 0.0503 3.0382 1.7430
No log 0.1579 12 1.8243 0.2000 1.8243 1.3507
No log 0.1842 14 2.2408 -0.0165 2.2408 1.4969
No log 0.2105 16 2.5161 -0.1439 2.5161 1.5862
No log 0.2368 18 2.8062 0.0 2.8062 1.6752
No log 0.2632 20 3.0956 0.0637 3.0956 1.7594
No log 0.2895 22 3.1541 0.1257 3.1541 1.7760
No log 0.3158 24 2.9782 0.1730 2.9782 1.7257
No log 0.3421 26 2.6822 0.2156 2.6822 1.6378
No log 0.3684 28 2.1783 0.3077 2.1783 1.4759
No log 0.3947 30 1.9780 0.3636 1.9780 1.4064
No log 0.4211 32 2.5944 0.2963 2.5944 1.6107
No log 0.4474 34 2.5974 0.2902 2.5974 1.6116
No log 0.4737 36 2.1491 0.3333 2.1491 1.4660
No log 0.5 38 1.8439 0.3899 1.8439 1.3579
No log 0.5263 40 1.7937 0.4156 1.7937 1.3393
No log 0.5526 42 2.2518 0.3371 2.2518 1.5006
No log 0.5789 44 2.1996 0.3425 2.1996 1.4831
No log 0.6053 46 2.2371 0.3529 2.2371 1.4957
No log 0.6316 48 1.9180 0.3978 1.9180 1.3849
No log 0.6579 50 1.6006 0.4740 1.6006 1.2652
No log 0.6842 52 2.0468 0.4500 2.0468 1.4307
No log 0.7105 54 2.0069 0.4677 2.0069 1.4166
No log 0.7368 56 1.9644 0.4677 1.9644 1.4016
No log 0.7632 58 1.4491 0.5165 1.4491 1.2038
No log 0.7895 60 1.2186 0.6145 1.2186 1.1039
No log 0.8158 62 1.1213 0.6364 1.1213 1.0589
No log 0.8421 64 0.9253 0.6790 0.9253 0.9619
No log 0.8684 66 0.8831 0.7273 0.8831 0.9398
No log 0.8947 68 0.9310 0.6901 0.9310 0.9649
No log 0.9211 70 0.8832 0.75 0.8832 0.9398
No log 0.9474 72 1.0012 0.7514 1.0012 1.0006
No log 0.9737 74 0.9196 0.7574 0.9196 0.9590
No log 1.0 76 0.8085 0.7564 0.8085 0.8991
No log 1.0263 78 0.8637 0.7368 0.8637 0.9294
No log 1.0526 80 0.8684 0.7111 0.8684 0.9319
No log 1.0789 82 1.2123 0.6535 1.2123 1.1011
No log 1.1053 84 1.0368 0.6965 1.0368 1.0183
No log 1.1316 86 0.7793 0.7486 0.7793 0.8828
No log 1.1579 88 0.9146 0.7006 0.9146 0.9564
No log 1.1842 90 0.8122 0.7683 0.8122 0.9012
No log 1.2105 92 0.7257 0.7647 0.7257 0.8519
No log 1.2368 94 0.7479 0.7709 0.7479 0.8648
No log 1.2632 96 0.6725 0.7975 0.6725 0.8201
No log 1.2895 98 0.6888 0.7848 0.6888 0.8299
No log 1.3158 100 0.9398 0.5811 0.9398 0.9694
No log 1.3421 102 1.0964 0.5479 1.0964 1.0471
No log 1.3684 104 0.9492 0.5811 0.9492 0.9743
No log 1.3947 106 0.7118 0.8075 0.7118 0.8437
No log 1.4211 108 0.6680 0.8 0.6680 0.8173
No log 1.4474 110 0.7156 0.8121 0.7156 0.8459
No log 1.4737 112 0.6840 0.8221 0.6840 0.8270
No log 1.5 114 0.7361 0.7816 0.7361 0.8579
No log 1.5263 116 0.6687 0.8353 0.6687 0.8178
No log 1.5526 118 0.6552 0.8025 0.6552 0.8095
No log 1.5789 120 0.6021 0.8221 0.6021 0.7759
No log 1.6053 122 0.6575 0.7976 0.6575 0.8108
No log 1.6316 124 0.8697 0.7312 0.8697 0.9326
No log 1.6579 126 0.7355 0.7598 0.7355 0.8576
No log 1.6842 128 0.5595 0.8364 0.5595 0.7480
No log 1.7105 130 0.6442 0.7568 0.6442 0.8026
No log 1.7368 132 0.6516 0.7619 0.6516 0.8072
No log 1.7632 134 0.6486 0.8 0.6486 0.8053
No log 1.7895 136 1.3925 0.6096 1.3925 1.1800
No log 1.8158 138 2.0824 0.4831 2.0824 1.4430
No log 1.8421 140 1.7956 0.5178 1.7956 1.3400
No log 1.8684 142 0.9566 0.7052 0.9566 0.9781
No log 1.8947 144 0.6195 0.7875 0.6195 0.7871
No log 1.9211 146 0.5987 0.8228 0.5987 0.7738
No log 1.9474 148 0.5803 0.8148 0.5803 0.7618
No log 1.9737 150 0.7703 0.7746 0.7703 0.8777
No log 2.0 152 0.8714 0.7222 0.8714 0.9335
No log 2.0263 154 0.6531 0.7875 0.6531 0.8081
No log 2.0526 156 0.5976 0.8050 0.5976 0.7731
No log 2.0789 158 0.6111 0.7867 0.6111 0.7817
No log 2.1053 160 0.6532 0.7534 0.6532 0.8082
No log 2.1316 162 0.6471 0.7619 0.6471 0.8044
No log 2.1579 164 0.6106 0.7919 0.6106 0.7814
No log 2.1842 166 0.7337 0.7465 0.7337 0.8566
No log 2.2105 168 0.8285 0.7133 0.8285 0.9102
No log 2.2368 170 0.6283 0.7532 0.6283 0.7926
No log 2.2632 172 0.5637 0.7843 0.5637 0.7508
No log 2.2895 174 0.6540 0.7368 0.6540 0.8087
No log 2.3158 176 0.5841 0.7895 0.5841 0.7643
No log 2.3421 178 0.5212 0.8313 0.5212 0.7220
No log 2.3684 180 0.7197 0.8045 0.7197 0.8483
No log 2.3947 182 1.0969 0.6571 1.0969 1.0474
No log 2.4211 184 0.9938 0.6837 0.9938 0.9969
No log 2.4474 186 0.6463 0.8045 0.6463 0.8040
No log 2.4737 188 0.5340 0.8344 0.5340 0.7308
No log 2.5 190 0.5347 0.8344 0.5347 0.7313
No log 2.5263 192 0.5652 0.8452 0.5652 0.7518
No log 2.5526 194 0.6336 0.8383 0.6336 0.7960
No log 2.5789 196 0.6443 0.8199 0.6443 0.8027
No log 2.6053 198 0.6242 0.8 0.6242 0.7901
No log 2.6316 200 0.6193 0.7973 0.6193 0.7870
No log 2.6579 202 0.6418 0.7123 0.6418 0.8011
No log 2.6842 204 0.6262 0.7123 0.6262 0.7913
No log 2.7105 206 0.5419 0.7895 0.5419 0.7361
No log 2.7368 208 0.5070 0.8148 0.5070 0.7120
No log 2.7632 210 0.5185 0.8242 0.5185 0.7200
No log 2.7895 212 0.5331 0.8077 0.5331 0.7301
No log 2.8158 214 0.5379 0.8098 0.5379 0.7334
No log 2.8421 216 0.5983 0.8343 0.5983 0.7735
No log 2.8684 218 0.8684 0.7579 0.8684 0.9319
No log 2.8947 220 1.0082 0.7461 1.0082 1.0041
No log 2.9211 222 0.7854 0.7935 0.7854 0.8862
No log 2.9474 224 0.5940 0.8344 0.5940 0.7707
No log 2.9737 226 0.5922 0.7922 0.5922 0.7696
No log 3.0 228 0.6067 0.8176 0.6067 0.7789
No log 3.0263 230 0.7483 0.8156 0.7483 0.8651
No log 3.0526 232 1.2331 0.6528 1.2331 1.1104
No log 3.0789 234 1.4099 0.5876 1.4099 1.1874
No log 3.1053 236 1.1396 0.6774 1.1396 1.0675
No log 3.1316 238 0.7963 0.75 0.7963 0.8923
No log 3.1579 240 0.7086 0.7532 0.7086 0.8418
No log 3.1842 242 0.6801 0.7949 0.6801 0.8247
No log 3.2105 244 0.6794 0.8098 0.6794 0.8243
No log 3.2368 246 0.7252 0.8046 0.7252 0.8516
No log 3.2632 248 0.7376 0.8045 0.7376 0.8588
No log 3.2895 250 0.7237 0.7841 0.7237 0.8507
No log 3.3158 252 0.6303 0.7901 0.6303 0.7939
No log 3.3421 254 0.5583 0.8077 0.5583 0.7472
No log 3.3684 256 0.5380 0.8077 0.5380 0.7335
No log 3.3947 258 0.5365 0.8101 0.5365 0.7325
No log 3.4211 260 0.5070 0.8182 0.5070 0.7120
No log 3.4474 262 0.5037 0.8129 0.5037 0.7097
No log 3.4737 264 0.4920 0.7922 0.4920 0.7014
No log 3.5 266 0.5631 0.7733 0.5631 0.7504
No log 3.5263 268 0.6375 0.7183 0.6375 0.7985
No log 3.5526 270 0.6298 0.7692 0.6298 0.7936
No log 3.5789 272 0.6225 0.7838 0.6225 0.7890
No log 3.6053 274 0.6627 0.7763 0.6627 0.8141
No log 3.6316 276 0.6849 0.7692 0.6849 0.8276
No log 3.6579 278 0.5999 0.7843 0.5999 0.7746
No log 3.6842 280 0.5693 0.7922 0.5693 0.7545
No log 3.7105 282 0.5374 0.7755 0.5374 0.7331
No log 3.7368 284 0.5140 0.7867 0.5140 0.7169
No log 3.7632 286 0.4899 0.7895 0.4899 0.6999
No log 3.7895 288 0.4925 0.7947 0.4925 0.7018
No log 3.8158 290 0.5109 0.7922 0.5109 0.7148
No log 3.8421 292 0.5378 0.8 0.5378 0.7333
No log 3.8684 294 0.5819 0.8025 0.5819 0.7628
No log 3.8947 296 0.6568 0.7683 0.6568 0.8104
No log 3.9211 298 0.6611 0.7758 0.6611 0.8131
No log 3.9474 300 0.6079 0.8025 0.6079 0.7797
No log 3.9737 302 0.6068 0.8026 0.6068 0.7790
No log 4.0 304 0.6116 0.8 0.6116 0.7821
No log 4.0263 306 0.5901 0.7867 0.5901 0.7682
No log 4.0526 308 0.5676 0.8052 0.5676 0.7534
No log 4.0789 310 0.5466 0.8052 0.5466 0.7393
No log 4.1053 312 0.5194 0.7922 0.5194 0.7207
No log 4.1316 314 0.5478 0.7838 0.5478 0.7401
No log 4.1579 316 0.5934 0.7534 0.5934 0.7704
No log 4.1842 318 0.6053 0.7619 0.6053 0.7780
No log 4.2105 320 0.6278 0.7755 0.6278 0.7924
No log 4.2368 322 0.6479 0.7619 0.6479 0.8050
No log 4.2632 324 0.6603 0.7703 0.6603 0.8126
No log 4.2895 326 0.6693 0.7703 0.6693 0.8181
No log 4.3158 328 0.6775 0.7763 0.6775 0.8231
No log 4.3421 330 0.6779 0.8125 0.6779 0.8233
No log 4.3684 332 0.6381 0.8293 0.6381 0.7988
No log 4.3947 334 0.6371 0.8364 0.6371 0.7982
No log 4.4211 336 0.6439 0.8171 0.6439 0.8024
No log 4.4474 338 0.6811 0.8193 0.6811 0.8253
No log 4.4737 340 0.7471 0.8072 0.7471 0.8643
No log 4.5 342 0.8122 0.7470 0.8122 0.9012
No log 4.5263 344 0.7255 0.7778 0.7255 0.8518
No log 4.5526 346 0.6481 0.7347 0.6481 0.8050
No log 4.5789 348 0.6293 0.7651 0.6293 0.7933
No log 4.6053 350 0.6563 0.7799 0.6563 0.8101
No log 4.6316 352 0.6163 0.8242 0.6163 0.7851
No log 4.6579 354 0.5540 0.7922 0.5540 0.7443
No log 4.6842 356 0.5568 0.8 0.5568 0.7462
No log 4.7105 358 0.5652 0.7733 0.5652 0.7518
No log 4.7368 360 0.6052 0.7586 0.6052 0.7779
No log 4.7632 362 0.6602 0.7376 0.6602 0.8126
No log 4.7895 364 0.6664 0.7361 0.6664 0.8163
No log 4.8158 366 0.7740 0.6803 0.7740 0.8798
No log 4.8421 368 0.8171 0.6483 0.8171 0.9039
No log 4.8684 370 0.7379 0.7027 0.7379 0.8590
No log 4.8947 372 0.6079 0.75 0.6079 0.7797
No log 4.9211 374 0.5776 0.7552 0.5776 0.7600
No log 4.9474 376 0.5574 0.75 0.5574 0.7466
No log 4.9737 378 0.5467 0.7895 0.5467 0.7394
No log 5.0 380 0.6315 0.8049 0.6315 0.7947
No log 5.0263 382 0.7844 0.7647 0.7844 0.8857
No log 5.0526 384 0.7939 0.7702 0.7939 0.8910
No log 5.0789 386 0.6902 0.7550 0.6902 0.8308
No log 5.1053 388 0.6028 0.7606 0.6028 0.7764
No log 5.1316 390 0.6024 0.7606 0.6024 0.7762
No log 5.1579 392 0.6267 0.7808 0.6267 0.7916
No log 5.1842 394 0.6611 0.7397 0.6611 0.8131
No log 5.2105 396 0.6180 0.76 0.6180 0.7861
No log 5.2368 398 0.5561 0.7703 0.5561 0.7457
No log 5.2632 400 0.5473 0.7671 0.5473 0.7398
No log 5.2895 402 0.5486 0.7671 0.5486 0.7407
No log 5.3158 404 0.5720 0.7815 0.5720 0.7563
No log 5.3421 406 0.6888 0.7662 0.6888 0.8299
No log 5.3684 408 0.7720 0.7929 0.7720 0.8786
No log 5.3947 410 0.7161 0.7929 0.7161 0.8463
No log 5.4211 412 0.5798 0.8272 0.5798 0.7614
No log 5.4474 414 0.5195 0.8129 0.5195 0.7207
No log 5.4737 416 0.5145 0.8105 0.5145 0.7173
No log 5.5 418 0.5075 0.8 0.5075 0.7124
No log 5.5263 420 0.5070 0.7838 0.5070 0.7120
No log 5.5526 422 0.5411 0.7919 0.5411 0.7356
No log 5.5789 424 0.5653 0.7808 0.5653 0.7518
No log 5.6053 426 0.5766 0.7445 0.5766 0.7593
No log 5.6316 428 0.5659 0.7606 0.5659 0.7522
No log 5.6579 430 0.5092 0.7838 0.5092 0.7136
No log 5.6842 432 0.5107 0.8153 0.5107 0.7146
No log 5.7105 434 0.6349 0.7831 0.6349 0.7968
No log 5.7368 436 0.6884 0.7665 0.6884 0.8297
No log 5.7632 438 0.5524 0.8144 0.5524 0.7432
No log 5.7895 440 0.4977 0.8228 0.4977 0.7054
No log 5.8158 442 0.5112 0.7947 0.5112 0.7150
No log 5.8421 444 0.5209 0.7838 0.5209 0.7217
No log 5.8684 446 0.5333 0.8101 0.5333 0.7303
No log 5.8947 448 0.7265 0.7758 0.7265 0.8524
No log 5.9211 450 0.9094 0.7273 0.9094 0.9536
No log 5.9474 452 0.8528 0.7470 0.8528 0.9234
No log 5.9737 454 0.6587 0.7875 0.6587 0.8116
No log 6.0 456 0.5167 0.7922 0.5167 0.7188
No log 6.0263 458 0.4879 0.8 0.4879 0.6985
No log 6.0526 460 0.5288 0.8025 0.5288 0.7272
No log 6.0789 462 0.5793 0.8025 0.5793 0.7611
No log 6.1053 464 0.6920 0.7758 0.6920 0.8319
No log 6.1316 466 0.6809 0.7692 0.6809 0.8252
No log 6.1579 468 0.5972 0.76 0.5972 0.7728
No log 6.1842 470 0.5768 0.7808 0.5768 0.7595
No log 6.2105 472 0.5863 0.7801 0.5863 0.7657
No log 6.2368 474 0.5684 0.7808 0.5684 0.7539
No log 6.2632 476 0.5876 0.7815 0.5876 0.7666
No log 6.2895 478 0.6572 0.7632 0.6572 0.8107
No log 6.3158 480 0.6825 0.7632 0.6825 0.8261
No log 6.3421 482 0.6801 0.7397 0.6801 0.8247
No log 6.3684 484 0.6784 0.7397 0.6784 0.8237
No log 6.3947 486 0.6325 0.7703 0.6325 0.7953
No log 6.4211 488 0.5872 0.7895 0.5872 0.7663
No log 6.4474 490 0.5539 0.8077 0.5539 0.7442
No log 6.4737 492 0.5603 0.8272 0.5603 0.7486
No log 6.5 494 0.5829 0.8372 0.5829 0.7635
No log 6.5263 496 0.5510 0.8171 0.5510 0.7423
No log 6.5526 498 0.5347 0.8101 0.5347 0.7312
0.3826 6.5789 500 0.5251 0.8 0.5251 0.7246
0.3826 6.6053 502 0.5053 0.7785 0.5053 0.7109
0.3826 6.6316 504 0.5029 0.7947 0.5029 0.7091
0.3826 6.6579 506 0.5279 0.7785 0.5279 0.7266
0.3826 6.6842 508 0.5690 0.7895 0.5690 0.7543
0.3826 6.7105 510 0.5853 0.7733 0.5853 0.7651

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k16_task1_organization

Finetuned
(4023)
this model