ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6663
  • Qwk: 0.6963
  • Mse: 0.6663
  • Rmse: 0.8163

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 7.1977 0.0110 7.1977 2.6829
No log 0.1053 4 4.9155 0.0375 4.9155 2.2171
No log 0.1579 6 2.9462 0.0848 2.9462 1.7165
No log 0.2105 8 2.4118 0.0699 2.4118 1.5530
No log 0.2632 10 2.2453 0.0305 2.2453 1.4984
No log 0.3158 12 2.5037 0.0 2.5037 1.5823
No log 0.3684 14 2.0324 0.1311 2.0324 1.4256
No log 0.4211 16 1.6133 0.1165 1.6133 1.2701
No log 0.4737 18 1.8282 0.0396 1.8282 1.3521
No log 0.5263 20 2.0048 0.0784 2.0048 1.4159
No log 0.5789 22 2.1402 0.1416 2.1402 1.4629
No log 0.6316 24 2.1755 0.1967 2.1755 1.4750
No log 0.6842 26 2.1798 0.1613 2.1798 1.4764
No log 0.7368 28 2.1239 0.2258 2.1239 1.4573
No log 0.7895 30 1.9313 0.1786 1.9313 1.3897
No log 0.8421 32 1.6493 0.0971 1.6493 1.2843
No log 0.8947 34 1.5376 0.1165 1.5376 1.2400
No log 0.9474 36 1.5807 0.3130 1.5807 1.2572
No log 1.0 38 1.4966 0.3390 1.4966 1.2233
No log 1.0526 40 1.3099 0.3063 1.3099 1.1445
No log 1.1053 42 1.3004 0.375 1.3004 1.1403
No log 1.1579 44 1.3096 0.4957 1.3096 1.1444
No log 1.2105 46 1.1492 0.4828 1.1492 1.0720
No log 1.2632 48 1.0139 0.5271 1.0139 1.0069
No log 1.3158 50 1.4218 0.5298 1.4218 1.1924
No log 1.3684 52 1.4099 0.4967 1.4099 1.1874
No log 1.4211 54 1.1177 0.4724 1.1177 1.0572
No log 1.4737 56 1.3092 0.3729 1.3092 1.1442
No log 1.5263 58 1.2252 0.3621 1.2252 1.1069
No log 1.5789 60 0.9628 0.5397 0.9628 0.9812
No log 1.6316 62 1.0127 0.5581 1.0127 1.0063
No log 1.6842 64 1.0276 0.5312 1.0276 1.0137
No log 1.7368 66 0.9843 0.5079 0.9843 0.9921
No log 1.7895 68 0.9411 0.6142 0.9411 0.9701
No log 1.8421 70 1.0531 0.6202 1.0531 1.0262
No log 1.8947 72 1.2994 0.5191 1.2994 1.1399
No log 1.9474 74 1.4784 0.4444 1.4784 1.2159
No log 2.0 76 1.3328 0.5152 1.3328 1.1545
No log 2.0526 78 0.9948 0.6370 0.9948 0.9974
No log 2.1053 80 0.8246 0.6571 0.8246 0.9081
No log 2.1579 82 0.7681 0.6803 0.7681 0.8764
No log 2.2105 84 0.7569 0.7092 0.7569 0.8700
No log 2.2632 86 0.9507 0.5985 0.9507 0.9751
No log 2.3158 88 1.1160 0.5734 1.1160 1.0564
No log 2.3684 90 0.9682 0.6122 0.9682 0.9840
No log 2.4211 92 0.7199 0.7034 0.7199 0.8485
No log 2.4737 94 0.8362 0.6939 0.8362 0.9145
No log 2.5263 96 0.9331 0.6923 0.9331 0.9660
No log 2.5789 98 0.8249 0.6897 0.8249 0.9082
No log 2.6316 100 0.7949 0.6901 0.7949 0.8916
No log 2.6842 102 0.9500 0.5972 0.9500 0.9747
No log 2.7368 104 0.9447 0.6331 0.9447 0.9719
No log 2.7895 106 0.8363 0.6861 0.8363 0.9145
No log 2.8421 108 0.7605 0.6667 0.7605 0.8721
No log 2.8947 110 0.8265 0.6429 0.8265 0.9091
No log 2.9474 112 0.8178 0.6573 0.8178 0.9043
No log 3.0 114 0.7030 0.7211 0.7030 0.8385
No log 3.0526 116 0.7974 0.6968 0.7974 0.8929
No log 3.1053 118 1.1484 0.6049 1.1484 1.0716
No log 3.1579 120 1.2072 0.5875 1.2072 1.0987
No log 3.2105 122 1.0226 0.5986 1.0226 1.0112
No log 3.2632 124 0.8813 0.6619 0.8813 0.9388
No log 3.3158 126 0.7641 0.6716 0.7641 0.8741
No log 3.3684 128 0.6967 0.7534 0.6967 0.8347
No log 3.4211 130 0.7141 0.7451 0.7141 0.8451
No log 3.4737 132 0.6182 0.8025 0.6182 0.7863
No log 3.5263 134 0.5929 0.7875 0.5929 0.7700
No log 3.5789 136 0.8099 0.7090 0.8099 0.9000
No log 3.6316 138 1.2199 0.6667 1.2199 1.1045
No log 3.6842 140 1.2365 0.6667 1.2365 1.1120
No log 3.7368 142 0.9192 0.6943 0.9192 0.9588
No log 3.7895 144 0.8614 0.7083 0.8614 0.9281
No log 3.8421 146 0.8939 0.6776 0.8939 0.9455
No log 3.8947 148 0.7603 0.7453 0.7603 0.8720
No log 3.9474 150 0.7387 0.7310 0.7387 0.8595
No log 4.0 152 0.8779 0.6081 0.8779 0.9369
No log 4.0526 154 0.9848 0.5694 0.9848 0.9924
No log 4.1053 156 0.8582 0.6471 0.8582 0.9264
No log 4.1579 158 0.7231 0.7092 0.7231 0.8504
No log 4.2105 160 0.7009 0.7183 0.7009 0.8372
No log 4.2632 162 0.7204 0.7299 0.7204 0.8488
No log 4.3158 164 0.9839 0.6164 0.9839 0.9919
No log 4.3684 166 1.2876 0.5033 1.2876 1.1347
No log 4.4211 168 1.2752 0.4930 1.2752 1.1292
No log 4.4737 170 1.0763 0.6131 1.0763 1.0374
No log 4.5263 172 0.9589 0.6260 0.9589 0.9793
No log 4.5789 174 0.9068 0.6567 0.9068 0.9523
No log 4.6316 176 0.8384 0.6519 0.8384 0.9156
No log 4.6842 178 0.7625 0.6957 0.7625 0.8732
No log 4.7368 180 0.7455 0.7170 0.7455 0.8634
No log 4.7895 182 0.8110 0.6957 0.8110 0.9006
No log 4.8421 184 0.7105 0.7545 0.7105 0.8429
No log 4.8947 186 0.5396 0.8193 0.5396 0.7345
No log 4.9474 188 0.5202 0.7975 0.5202 0.7212
No log 5.0 190 0.5286 0.8098 0.5286 0.7270
No log 5.0526 192 0.5239 0.7975 0.5239 0.7238
No log 5.1053 194 0.6081 0.8 0.6081 0.7798
No log 5.1579 196 0.8225 0.7081 0.8225 0.9069
No log 5.2105 198 0.8948 0.6842 0.8948 0.9459
No log 5.2632 200 0.9336 0.6711 0.9336 0.9662
No log 5.3158 202 0.8863 0.6714 0.8863 0.9414
No log 5.3684 204 0.8062 0.7075 0.8062 0.8979
No log 5.4211 206 0.7527 0.7261 0.7527 0.8676
No log 5.4737 208 0.7347 0.7485 0.7347 0.8572
No log 5.5263 210 0.7419 0.7114 0.7419 0.8613
No log 5.5789 212 0.8087 0.6906 0.8087 0.8993
No log 5.6316 214 0.8454 0.6667 0.8454 0.9195
No log 5.6842 216 0.8479 0.6767 0.8479 0.9208
No log 5.7368 218 0.8376 0.6667 0.8376 0.9152
No log 5.7895 220 0.8099 0.6763 0.8099 0.9000
No log 5.8421 222 0.7399 0.6853 0.7399 0.8602
No log 5.8947 224 0.6670 0.72 0.6670 0.8167
No log 5.9474 226 0.6349 0.7875 0.6349 0.7968
No log 6.0 228 0.6297 0.7778 0.6297 0.7935
No log 6.0526 230 0.6824 0.7602 0.6824 0.8261
No log 6.1053 232 0.6828 0.75 0.6828 0.8263
No log 6.1579 234 0.6562 0.7362 0.6562 0.8101
No log 6.2105 236 0.6725 0.7133 0.6725 0.8200
No log 6.2632 238 0.7265 0.6715 0.7265 0.8524
No log 6.3158 240 0.7239 0.6812 0.7239 0.8508
No log 6.3684 242 0.6899 0.7172 0.6899 0.8306
No log 6.4211 244 0.7343 0.7333 0.7343 0.8569
No log 6.4737 246 0.8402 0.7186 0.8402 0.9166
No log 6.5263 248 0.8079 0.7586 0.8079 0.8989
No log 6.5789 250 0.8112 0.7586 0.8112 0.9007
No log 6.6316 252 0.7067 0.7296 0.7067 0.8406
No log 6.6842 254 0.6388 0.7785 0.6388 0.7993
No log 6.7368 256 0.6563 0.7724 0.6563 0.8101
No log 6.7895 258 0.6949 0.7162 0.6949 0.8336
No log 6.8421 260 0.6768 0.7586 0.6768 0.8227
No log 6.8947 262 0.6496 0.75 0.6496 0.8060
No log 6.9474 264 0.6456 0.7619 0.6456 0.8035
No log 7.0 266 0.6557 0.7383 0.6557 0.8098
No log 7.0526 268 0.7305 0.7362 0.7305 0.8547
No log 7.1053 270 0.7921 0.7125 0.7921 0.8900
No log 7.1579 272 0.7687 0.6713 0.7687 0.8767
No log 7.2105 274 0.8097 0.6713 0.8097 0.8998
No log 7.2632 276 0.8667 0.6667 0.8667 0.9310
No log 7.3158 278 0.8331 0.7050 0.8331 0.9128
No log 7.3684 280 0.8140 0.7 0.8140 0.9022
No log 7.4211 282 0.7904 0.7050 0.7904 0.8891
No log 7.4737 284 0.7874 0.6912 0.7874 0.8873
No log 7.5263 286 0.7650 0.6912 0.7650 0.8746
No log 7.5789 288 0.7519 0.6912 0.7519 0.8671
No log 7.6316 290 0.7297 0.6901 0.7297 0.8542
No log 7.6842 292 0.7759 0.7170 0.7759 0.8808
No log 7.7368 294 0.7623 0.7105 0.7623 0.8731
No log 7.7895 296 0.7325 0.6986 0.7325 0.8559
No log 7.8421 298 0.7654 0.7067 0.7654 0.8749
No log 7.8947 300 0.8087 0.7170 0.8087 0.8993
No log 7.9474 302 0.8259 0.7273 0.8259 0.9088
No log 8.0 304 0.7741 0.7190 0.7741 0.8798
No log 8.0526 306 0.7079 0.6806 0.7079 0.8413
No log 8.1053 308 0.6767 0.7299 0.6767 0.8226
No log 8.1579 310 0.6524 0.7338 0.6524 0.8077
No log 8.2105 312 0.6637 0.7391 0.6637 0.8147
No log 8.2632 314 0.7013 0.7172 0.7013 0.8375
No log 8.3158 316 0.8508 0.7349 0.8508 0.9224
No log 8.3684 318 0.9038 0.7305 0.9038 0.9507
No log 8.4211 320 0.8029 0.7317 0.8029 0.8960
No log 8.4737 322 0.6709 0.7134 0.6709 0.8191
No log 8.5263 324 0.6275 0.7285 0.6275 0.7922
No log 8.5789 326 0.6311 0.7383 0.6311 0.7944
No log 8.6316 328 0.6606 0.6857 0.6606 0.8127
No log 8.6842 330 0.7512 0.7134 0.7512 0.8667
No log 8.7368 332 0.8010 0.7362 0.8010 0.8950
No log 8.7895 334 0.7365 0.7211 0.7365 0.8582
No log 8.8421 336 0.6785 0.7042 0.6785 0.8237
No log 8.8947 338 0.6654 0.6957 0.6654 0.8157
No log 8.9474 340 0.6566 0.7153 0.6566 0.8103
No log 9.0 342 0.6563 0.7517 0.6563 0.8101
No log 9.0526 344 0.6725 0.7662 0.6725 0.8201
No log 9.1053 346 0.6874 0.7407 0.6874 0.8291
No log 9.1579 348 0.6496 0.7848 0.6496 0.8060
No log 9.2105 350 0.6155 0.7785 0.6155 0.7845
No log 9.2632 352 0.6084 0.7632 0.6084 0.7800
No log 9.3158 354 0.6621 0.7590 0.6621 0.8137
No log 9.3684 356 0.7390 0.7205 0.7390 0.8597
No log 9.4211 358 0.8129 0.7 0.8129 0.9016
No log 9.4737 360 0.7853 0.6901 0.7853 0.8862
No log 9.5263 362 0.7668 0.6418 0.7668 0.8757
No log 9.5789 364 0.8115 0.6462 0.8115 0.9008
No log 9.6316 366 0.8820 0.6418 0.8820 0.9392
No log 9.6842 368 0.9474 0.6338 0.9474 0.9733
No log 9.7368 370 0.9430 0.6483 0.9430 0.9711
No log 9.7895 372 0.8257 0.6842 0.8257 0.9087
No log 9.8421 374 0.7246 0.6980 0.7246 0.8512
No log 9.8947 376 0.6743 0.6861 0.6743 0.8212
No log 9.9474 378 0.6828 0.6917 0.6828 0.8263
No log 10.0 380 0.7450 0.6667 0.7450 0.8632
No log 10.0526 382 0.8856 0.6712 0.8856 0.9410
No log 10.1053 384 0.8889 0.6618 0.8889 0.9428
No log 10.1579 386 0.8692 0.6617 0.8692 0.9323
No log 10.2105 388 0.7892 0.6763 0.7892 0.8884
No log 10.2632 390 0.7101 0.7050 0.7101 0.8427
No log 10.3158 392 0.6305 0.7518 0.6305 0.7940
No log 10.3684 394 0.6052 0.7808 0.6052 0.7779
No log 10.4211 396 0.6078 0.7432 0.6078 0.7796
No log 10.4737 398 0.6744 0.7468 0.6744 0.8212
No log 10.5263 400 0.7251 0.7226 0.7251 0.8516
No log 10.5789 402 0.7793 0.6892 0.7793 0.8828
No log 10.6316 404 0.7793 0.6475 0.7793 0.8828
No log 10.6842 406 0.7649 0.6569 0.7649 0.8746
No log 10.7368 408 0.7192 0.6917 0.7192 0.8481
No log 10.7895 410 0.6739 0.6815 0.6739 0.8209
No log 10.8421 412 0.6895 0.6716 0.6895 0.8304
No log 10.8947 414 0.7384 0.6716 0.7384 0.8593
No log 10.9474 416 0.7362 0.6617 0.7362 0.8580
No log 11.0 418 0.7772 0.6617 0.7772 0.8816
No log 11.0526 420 0.7762 0.6617 0.7762 0.8810
No log 11.1053 422 0.7371 0.6466 0.7371 0.8585
No log 11.1579 424 0.7141 0.6569 0.7141 0.8451
No log 11.2105 426 0.6820 0.6667 0.6820 0.8258
No log 11.2632 428 0.6465 0.7083 0.6465 0.8040
No log 11.3158 430 0.6475 0.7375 0.6475 0.8047
No log 11.3684 432 0.6760 0.7485 0.6760 0.8222
No log 11.4211 434 0.6712 0.7205 0.6712 0.8193
No log 11.4737 436 0.7076 0.6667 0.7076 0.8412
No log 11.5263 438 0.7364 0.6957 0.7364 0.8581
No log 11.5789 440 0.7709 0.6714 0.7709 0.8780
No log 11.6316 442 0.8027 0.6800 0.8027 0.8959
No log 11.6842 444 0.7878 0.7134 0.7878 0.8876
No log 11.7368 446 0.7101 0.7436 0.7101 0.8427
No log 11.7895 448 0.6807 0.7547 0.6807 0.8250
No log 11.8421 450 0.6893 0.7613 0.6893 0.8303
No log 11.8947 452 0.7124 0.7248 0.7124 0.8441
No log 11.9474 454 0.7391 0.6308 0.7391 0.8597
No log 12.0 456 0.7389 0.6412 0.7389 0.8596
No log 12.0526 458 0.7152 0.6923 0.7152 0.8457
No log 12.1053 460 0.7191 0.6912 0.7191 0.8480
No log 12.1579 462 0.6440 0.7571 0.6440 0.8025
No log 12.2105 464 0.5929 0.7673 0.5929 0.7700
No log 12.2632 466 0.6744 0.7514 0.6744 0.8212
No log 12.3158 468 0.7500 0.75 0.7500 0.8660
No log 12.3684 470 0.6847 0.7709 0.6847 0.8275
No log 12.4211 472 0.6431 0.7456 0.6431 0.8019
No log 12.4737 474 0.6292 0.7261 0.6292 0.7932
No log 12.5263 476 0.5961 0.7692 0.5961 0.7720
No log 12.5789 478 0.6259 0.7338 0.6259 0.7911
No log 12.6316 480 0.6444 0.7338 0.6444 0.8027
No log 12.6842 482 0.6836 0.6866 0.6836 0.8268
No log 12.7368 484 0.7327 0.7114 0.7327 0.8560
No log 12.7895 486 0.7255 0.7226 0.7255 0.8518
No log 12.8421 488 0.6387 0.7237 0.6387 0.7992
No log 12.8947 490 0.5773 0.8098 0.5773 0.7598
No log 12.9474 492 0.5563 0.8148 0.5563 0.7459
No log 13.0 494 0.5666 0.8171 0.5666 0.7527
No log 13.0526 496 0.5775 0.8221 0.5775 0.7600
No log 13.1053 498 0.6305 0.7875 0.6305 0.7941
0.4882 13.1579 500 0.6071 0.8075 0.6071 0.7792
0.4882 13.2105 502 0.5663 0.8025 0.5663 0.7525
0.4882 13.2632 504 0.5659 0.8049 0.5659 0.7522
0.4882 13.3158 506 0.5812 0.8144 0.5812 0.7624
0.4882 13.3684 508 0.6420 0.7702 0.6420 0.8012
0.4882 13.4211 510 0.6996 0.7308 0.6996 0.8364
0.4882 13.4737 512 0.6924 0.7273 0.6924 0.8321
0.4882 13.5263 514 0.6649 0.6993 0.6649 0.8154
0.4882 13.5789 516 0.6663 0.6963 0.6663 0.8163

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task1_organization

Finetuned
(4019)
this model