ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k6_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5464
  • Qwk: 0.5022
  • Mse: 0.5464
  • Rmse: 0.7392

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 2.7056 -0.0262 2.7056 1.6449
No log 0.125 4 1.4289 0.0278 1.4289 1.1954
No log 0.1875 6 1.0822 -0.0927 1.0822 1.0403
No log 0.25 8 1.0991 -0.0727 1.0991 1.0484
No log 0.3125 10 0.9725 0.0864 0.9725 0.9862
No log 0.375 12 0.8443 0.2467 0.8443 0.9188
No log 0.4375 14 0.7929 0.2467 0.7929 0.8905
No log 0.5 16 0.7599 0.1598 0.7599 0.8717
No log 0.5625 18 0.7549 0.1918 0.7549 0.8688
No log 0.625 20 0.7390 0.1268 0.7390 0.8597
No log 0.6875 22 0.7779 0.0154 0.7779 0.8820
No log 0.75 24 0.7204 0.0757 0.7204 0.8488
No log 0.8125 26 0.7021 0.0481 0.7021 0.8379
No log 0.875 28 0.6938 0.0481 0.6938 0.8329
No log 0.9375 30 0.6699 0.0481 0.6699 0.8185
No log 1.0 32 0.6672 0.1236 0.6672 0.8168
No log 1.0625 34 0.6852 0.1617 0.6852 0.8278
No log 1.125 36 0.7292 0.1922 0.7292 0.8539
No log 1.1875 38 0.7756 0.1648 0.7756 0.8807
No log 1.25 40 0.7486 0.0927 0.7486 0.8652
No log 1.3125 42 0.6835 0.1863 0.6835 0.8267
No log 1.375 44 0.6849 0.1277 0.6849 0.8276
No log 1.4375 46 0.7551 0.2109 0.7551 0.8690
No log 1.5 48 0.7039 0.2132 0.7039 0.8390
No log 1.5625 50 0.6442 0.2607 0.6442 0.8026
No log 1.625 52 0.6680 0.1737 0.6680 0.8173
No log 1.6875 54 0.6958 0.1268 0.6958 0.8341
No log 1.75 56 0.7115 0.1268 0.7115 0.8435
No log 1.8125 58 0.6768 0.1699 0.6768 0.8227
No log 1.875 60 0.6825 0.3387 0.6825 0.8261
No log 1.9375 62 0.6279 0.3728 0.6279 0.7924
No log 2.0 64 0.6110 0.3093 0.6110 0.7816
No log 2.0625 66 0.6861 0.2109 0.6861 0.8283
No log 2.125 68 0.6897 0.1979 0.6897 0.8305
No log 2.1875 70 0.7559 0.1834 0.7559 0.8694
No log 2.25 72 0.8295 0.0262 0.8295 0.9108
No log 2.3125 74 0.7926 0.0618 0.7926 0.8903
No log 2.375 76 0.7146 0.1790 0.7146 0.8454
No log 2.4375 78 0.6768 0.2002 0.6768 0.8227
No log 2.5 80 0.6710 0.1604 0.6710 0.8191
No log 2.5625 82 0.7371 0.3399 0.7371 0.8586
No log 2.625 84 0.7869 0.3475 0.7869 0.8871
No log 2.6875 86 0.6627 0.4206 0.6627 0.8140
No log 2.75 88 0.6248 0.2368 0.6248 0.7905
No log 2.8125 90 0.7832 0.2875 0.7832 0.8850
No log 2.875 92 0.8122 0.3938 0.8122 0.9012
No log 2.9375 94 0.6572 0.3257 0.6572 0.8107
No log 3.0 96 0.6456 0.2955 0.6456 0.8035
No log 3.0625 98 0.6316 0.3936 0.6316 0.7947
No log 3.125 100 0.6305 0.3936 0.6305 0.7941
No log 3.1875 102 0.5746 0.4061 0.5746 0.7580
No log 3.25 104 0.5759 0.3974 0.5759 0.7588
No log 3.3125 106 0.5732 0.3153 0.5732 0.7571
No log 3.375 108 0.6304 0.4100 0.6304 0.7940
No log 3.4375 110 0.6312 0.3919 0.6312 0.7945
No log 3.5 112 0.6188 0.3296 0.6188 0.7866
No log 3.5625 114 0.6060 0.3228 0.6060 0.7785
No log 3.625 116 0.6011 0.3688 0.6011 0.7753
No log 3.6875 118 0.5944 0.3183 0.5944 0.7710
No log 3.75 120 0.5984 0.4273 0.5984 0.7735
No log 3.8125 122 0.6486 0.3754 0.6486 0.8054
No log 3.875 124 0.6076 0.4198 0.6076 0.7795
No log 3.9375 126 0.6217 0.4139 0.6217 0.7885
No log 4.0 128 0.7176 0.3594 0.7176 0.8471
No log 4.0625 130 0.6946 0.3343 0.6946 0.8334
No log 4.125 132 0.5518 0.4314 0.5518 0.7428
No log 4.1875 134 0.5677 0.4459 0.5677 0.7535
No log 4.25 136 0.5546 0.4224 0.5546 0.7447
No log 4.3125 138 0.5535 0.4224 0.5535 0.7440
No log 4.375 140 0.5404 0.5044 0.5404 0.7351
No log 4.4375 142 0.5215 0.4504 0.5215 0.7222
No log 4.5 144 0.5091 0.4768 0.5091 0.7135
No log 4.5625 146 0.5111 0.4613 0.5111 0.7149
No log 4.625 148 0.5722 0.4927 0.5722 0.7564
No log 4.6875 150 0.5259 0.5475 0.5259 0.7252
No log 4.75 152 0.5226 0.5022 0.5226 0.7229
No log 4.8125 154 0.5139 0.5617 0.5139 0.7169
No log 4.875 156 0.5713 0.4949 0.5713 0.7558
No log 4.9375 158 0.5825 0.4949 0.5825 0.7632
No log 5.0 160 0.5252 0.4929 0.5252 0.7247
No log 5.0625 162 0.5856 0.4067 0.5856 0.7653
No log 5.125 164 0.7060 0.3776 0.7060 0.8402
No log 5.1875 166 0.6800 0.4270 0.6800 0.8246
No log 5.25 168 0.5992 0.4358 0.5992 0.7741
No log 5.3125 170 0.6102 0.5248 0.6102 0.7811
No log 5.375 172 0.5927 0.5304 0.5927 0.7699
No log 5.4375 174 0.6046 0.3738 0.6046 0.7776
No log 5.5 176 0.6443 0.4375 0.6443 0.8027
No log 5.5625 178 0.6092 0.3865 0.6092 0.7805
No log 5.625 180 0.6037 0.4441 0.6037 0.7770
No log 5.6875 182 0.6195 0.4345 0.6195 0.7871
No log 5.75 184 0.6118 0.4136 0.6118 0.7822
No log 5.8125 186 0.6059 0.3599 0.6059 0.7784
No log 5.875 188 0.6228 0.4618 0.6228 0.7891
No log 5.9375 190 0.6408 0.4534 0.6408 0.8005
No log 6.0 192 0.5990 0.4639 0.5990 0.7740
No log 6.0625 194 0.6527 0.4083 0.6527 0.8079
No log 6.125 196 0.8093 0.3002 0.8093 0.8996
No log 6.1875 198 0.7635 0.2612 0.7635 0.8738
No log 6.25 200 0.6094 0.3289 0.6094 0.7806
No log 6.3125 202 0.6979 0.4329 0.6979 0.8354
No log 6.375 204 0.6815 0.4329 0.6815 0.8255
No log 6.4375 206 0.5538 0.5631 0.5538 0.7442
No log 6.5 208 0.5661 0.4595 0.5661 0.7524
No log 6.5625 210 0.5630 0.4517 0.5630 0.7503
No log 6.625 212 0.5458 0.5800 0.5458 0.7388
No log 6.6875 214 0.5660 0.5404 0.5660 0.7523
No log 6.75 216 0.5768 0.5170 0.5768 0.7595
No log 6.8125 218 0.5839 0.4972 0.5839 0.7641
No log 6.875 220 0.6169 0.4149 0.6169 0.7854
No log 6.9375 222 0.6342 0.3883 0.6342 0.7964
No log 7.0 224 0.6178 0.3703 0.6178 0.7860
No log 7.0625 226 0.6220 0.3890 0.6220 0.7887
No log 7.125 228 0.6410 0.3890 0.6410 0.8006
No log 7.1875 230 0.6563 0.3702 0.6563 0.8101
No log 7.25 232 0.6751 0.2847 0.6751 0.8216
No log 7.3125 234 0.6755 0.2780 0.6755 0.8219
No log 7.375 236 0.6879 0.3622 0.6879 0.8294
No log 7.4375 238 0.6898 0.3416 0.6898 0.8306
No log 7.5 240 0.6507 0.3352 0.6507 0.8067
No log 7.5625 242 0.6297 0.3426 0.6297 0.7935
No log 7.625 244 0.6505 0.3729 0.6505 0.8065
No log 7.6875 246 0.6465 0.3831 0.6465 0.8041
No log 7.75 248 0.6352 0.3806 0.6352 0.7970
No log 7.8125 250 0.6035 0.3558 0.6035 0.7768
No log 7.875 252 0.6577 0.3271 0.6577 0.8110
No log 7.9375 254 0.8824 0.2756 0.8824 0.9394
No log 8.0 256 1.0130 0.2676 1.0130 1.0065
No log 8.0625 258 0.9300 0.2702 0.9300 0.9644
No log 8.125 260 0.7431 0.3060 0.7431 0.8620
No log 8.1875 262 0.6088 0.4215 0.6088 0.7803
No log 8.25 264 0.5918 0.4215 0.5918 0.7693
No log 8.3125 266 0.7277 0.3556 0.7277 0.8531
No log 8.375 268 0.9366 0.3906 0.9366 0.9678
No log 8.4375 270 0.8709 0.3483 0.8709 0.9332
No log 8.5 272 0.6563 0.3970 0.6563 0.8101
No log 8.5625 274 0.5215 0.4527 0.5215 0.7222
No log 8.625 276 0.5009 0.5304 0.5009 0.7077
No log 8.6875 278 0.5848 0.4886 0.5848 0.7647
No log 8.75 280 0.7514 0.4208 0.7514 0.8669
No log 8.8125 282 0.7705 0.3998 0.7705 0.8778
No log 8.875 284 0.6504 0.4170 0.6504 0.8065
No log 8.9375 286 0.5113 0.5750 0.5113 0.7151
No log 9.0 288 0.5341 0.5232 0.5341 0.7308
No log 9.0625 290 0.5460 0.5307 0.5460 0.7390
No log 9.125 292 0.5304 0.5268 0.5304 0.7283
No log 9.1875 294 0.5424 0.4742 0.5424 0.7365
No log 9.25 296 0.5538 0.5152 0.5538 0.7442
No log 9.3125 298 0.5591 0.4972 0.5591 0.7477
No log 9.375 300 0.5766 0.4837 0.5766 0.7593
No log 9.4375 302 0.5803 0.4878 0.5803 0.7618
No log 9.5 304 0.6046 0.4345 0.6046 0.7776
No log 9.5625 306 0.6011 0.3934 0.6011 0.7753
No log 9.625 308 0.6059 0.4345 0.6059 0.7784
No log 9.6875 310 0.5952 0.3988 0.5952 0.7715
No log 9.75 312 0.5816 0.5114 0.5816 0.7626
No log 9.8125 314 0.5923 0.4681 0.5923 0.7696
No log 9.875 316 0.5817 0.5042 0.5817 0.7627
No log 9.9375 318 0.5831 0.5286 0.5831 0.7636
No log 10.0 320 0.5899 0.4991 0.5899 0.7681
No log 10.0625 322 0.5971 0.4364 0.5971 0.7727
No log 10.125 324 0.5988 0.4160 0.5988 0.7738
No log 10.1875 326 0.5918 0.4137 0.5918 0.7693
No log 10.25 328 0.5952 0.4701 0.5952 0.7715
No log 10.3125 330 0.6058 0.4895 0.6058 0.7783
No log 10.375 332 0.6080 0.4895 0.6080 0.7798
No log 10.4375 334 0.6060 0.5141 0.6060 0.7785
No log 10.5 336 0.5821 0.4484 0.5821 0.7630
No log 10.5625 338 0.5851 0.4898 0.5851 0.7649
No log 10.625 340 0.5817 0.4484 0.5817 0.7627
No log 10.6875 342 0.5757 0.4934 0.5757 0.7587
No log 10.75 344 0.5700 0.4299 0.5700 0.7550
No log 10.8125 346 0.5919 0.3489 0.5919 0.7694
No log 10.875 348 0.6256 0.3542 0.6256 0.7909
No log 10.9375 350 0.5982 0.3489 0.5982 0.7734
No log 11.0 352 0.5712 0.3558 0.5712 0.7558
No log 11.0625 354 0.5656 0.5095 0.5656 0.7521
No log 11.125 356 0.5554 0.5326 0.5554 0.7452
No log 11.1875 358 0.5325 0.5095 0.5325 0.7297
No log 11.25 360 0.5369 0.4611 0.5369 0.7327
No log 11.3125 362 0.5945 0.3803 0.5945 0.7711
No log 11.375 364 0.6050 0.3803 0.6050 0.7778
No log 11.4375 366 0.5419 0.5432 0.5419 0.7361
No log 11.5 368 0.5043 0.5782 0.5043 0.7101
No log 11.5625 370 0.5102 0.5326 0.5102 0.7143
No log 11.625 372 0.5064 0.5521 0.5064 0.7116
No log 11.6875 374 0.5119 0.4990 0.5119 0.7155
No log 11.75 376 0.5414 0.4869 0.5414 0.7358
No log 11.8125 378 0.5213 0.4869 0.5213 0.7220
No log 11.875 380 0.5191 0.5095 0.5191 0.7205
No log 11.9375 382 0.5704 0.5345 0.5704 0.7553
No log 12.0 384 0.6354 0.5986 0.6354 0.7971
No log 12.0625 386 0.6277 0.5986 0.6277 0.7922
No log 12.125 388 0.5491 0.5501 0.5491 0.7410
No log 12.1875 390 0.5077 0.5430 0.5077 0.7125
No log 12.25 392 0.5313 0.5046 0.5313 0.7289
No log 12.3125 394 0.5522 0.5046 0.5522 0.7431
No log 12.375 396 0.5293 0.5286 0.5293 0.7275
No log 12.4375 398 0.5591 0.5455 0.5591 0.7477
No log 12.5 400 0.6025 0.4724 0.6025 0.7762
No log 12.5625 402 0.5764 0.5214 0.5764 0.7592
No log 12.625 404 0.5573 0.4126 0.5573 0.7465
No log 12.6875 406 0.6255 0.3343 0.6255 0.7909
No log 12.75 408 0.6489 0.2769 0.6489 0.8055
No log 12.8125 410 0.6174 0.3407 0.6174 0.7857
No log 12.875 412 0.5850 0.4186 0.5850 0.7648
No log 12.9375 414 0.5965 0.4964 0.5965 0.7723
No log 13.0 416 0.5896 0.5123 0.5896 0.7679
No log 13.0625 418 0.5347 0.5344 0.5347 0.7312
No log 13.125 420 0.5260 0.5633 0.5260 0.7252
No log 13.1875 422 0.5987 0.4613 0.5987 0.7738
No log 13.25 424 0.6059 0.5003 0.6059 0.7784
No log 13.3125 426 0.5243 0.5217 0.5243 0.7241
No log 13.375 428 0.4849 0.5765 0.4848 0.6963
No log 13.4375 430 0.4905 0.5656 0.4905 0.7003
No log 13.5 432 0.4917 0.5373 0.4917 0.7012
No log 13.5625 434 0.4965 0.5765 0.4965 0.7046
No log 13.625 436 0.4969 0.5373 0.4969 0.7049
No log 13.6875 438 0.5009 0.5765 0.5009 0.7078
No log 13.75 440 0.5085 0.5248 0.5085 0.7131
No log 13.8125 442 0.5053 0.5457 0.5053 0.7108
No log 13.875 444 0.5084 0.5152 0.5084 0.7130
No log 13.9375 446 0.5254 0.5248 0.5254 0.7248
No log 14.0 448 0.5380 0.5248 0.5380 0.7335
No log 14.0625 450 0.5586 0.4821 0.5586 0.7474
No log 14.125 452 0.5739 0.4756 0.5739 0.7575
No log 14.1875 454 0.5328 0.5266 0.5328 0.7299
No log 14.25 456 0.5248 0.5266 0.5248 0.7244
No log 14.3125 458 0.5614 0.5283 0.5614 0.7493
No log 14.375 460 0.6004 0.4610 0.6004 0.7749
No log 14.4375 462 0.6044 0.4610 0.6044 0.7774
No log 14.5 464 0.5784 0.5050 0.5784 0.7605
No log 14.5625 466 0.5625 0.4817 0.5625 0.7500
No log 14.625 468 0.5485 0.4837 0.5485 0.7406
No log 14.6875 470 0.5649 0.5307 0.5649 0.7516
No log 14.75 472 0.5532 0.4837 0.5532 0.7437
No log 14.8125 474 0.5517 0.4299 0.5517 0.7428
No log 14.875 476 0.5786 0.5117 0.5786 0.7607
No log 14.9375 478 0.5618 0.5098 0.5618 0.7495
No log 15.0 480 0.5608 0.5028 0.5608 0.7489
No log 15.0625 482 0.5488 0.5028 0.5488 0.7408
No log 15.125 484 0.5315 0.4991 0.5315 0.7290
No log 15.1875 486 0.5357 0.5344 0.5357 0.7319
No log 15.25 488 0.5365 0.5344 0.5365 0.7325
No log 15.3125 490 0.5365 0.4788 0.5365 0.7325
No log 15.375 492 0.5553 0.5028 0.5553 0.7452
No log 15.4375 494 0.5560 0.4635 0.5560 0.7457
No log 15.5 496 0.5365 0.4788 0.5365 0.7325
No log 15.5625 498 0.5527 0.5095 0.5527 0.7434
0.3531 15.625 500 0.5454 0.5095 0.5454 0.7385
0.3531 15.6875 502 0.5532 0.4635 0.5532 0.7438
0.3531 15.75 504 0.6112 0.4528 0.6112 0.7818
0.3531 15.8125 506 0.6180 0.4293 0.6180 0.7861
0.3531 15.875 508 0.5664 0.4635 0.5664 0.7526
0.3531 15.9375 510 0.5464 0.5022 0.5464 0.7392

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k6_task7_organization

Finetuned
(4019)
this model