ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7490
  • Qwk: 0.6322
  • Mse: 0.7490
  • Rmse: 0.8655

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0294 2 5.1784 -0.0207 5.1784 2.2756
No log 0.0588 4 2.9136 0.0857 2.9136 1.7069
No log 0.0882 6 1.9192 0.0236 1.9192 1.3854
No log 0.1176 8 1.4595 0.0984 1.4595 1.2081
No log 0.1471 10 1.2047 0.1621 1.2047 1.0976
No log 0.1765 12 1.3215 0.1494 1.3215 1.1496
No log 0.2059 14 2.6948 0.0790 2.6948 1.6416
No log 0.2353 16 3.0062 -0.0129 3.0062 1.7338
No log 0.2647 18 3.0468 -0.0165 3.0468 1.7455
No log 0.2941 20 2.4170 0.1080 2.4170 1.5547
No log 0.3235 22 1.7457 0.0376 1.7457 1.3213
No log 0.3529 24 1.3741 0.0403 1.3741 1.1722
No log 0.3824 26 1.3456 0.0539 1.3456 1.1600
No log 0.4118 28 1.3713 0.0533 1.3713 1.1710
No log 0.4412 30 1.2732 0.2512 1.2732 1.1283
No log 0.4706 32 1.1768 0.2722 1.1768 1.0848
No log 0.5 34 1.1962 0.1762 1.1962 1.0937
No log 0.5294 36 1.1917 0.2456 1.1917 1.0916
No log 0.5588 38 1.2290 0.2150 1.2290 1.1086
No log 0.5882 40 1.3514 0.1492 1.3514 1.1625
No log 0.6176 42 1.7797 0.0430 1.7797 1.3341
No log 0.6471 44 1.9130 0.0156 1.9130 1.3831
No log 0.6765 46 1.6550 0.0156 1.6550 1.2865
No log 0.7059 48 1.4258 0.0430 1.4258 1.1941
No log 0.7353 50 1.2460 0.2811 1.2460 1.1163
No log 0.7647 52 1.1155 0.1858 1.1155 1.0562
No log 0.7941 54 1.1142 0.1245 1.1142 1.0556
No log 0.8235 56 1.1018 0.1560 1.1018 1.0497
No log 0.8529 58 1.0991 0.3110 1.0991 1.0484
No log 0.8824 60 1.1914 0.3183 1.1914 1.0915
No log 0.9118 62 1.2893 0.1538 1.2893 1.1355
No log 0.9412 64 1.2114 0.2694 1.2114 1.1006
No log 0.9706 66 1.0258 0.3619 1.0258 1.0128
No log 1.0 68 1.0253 0.3197 1.0253 1.0126
No log 1.0294 70 1.2684 0.1828 1.2684 1.1262
No log 1.0588 72 1.1884 0.1838 1.1884 1.0901
No log 1.0882 74 0.9745 0.3423 0.9745 0.9872
No log 1.1176 76 0.9689 0.3826 0.9689 0.9843
No log 1.1471 78 1.0086 0.3611 1.0086 1.0043
No log 1.1765 80 0.9527 0.4066 0.9527 0.9761
No log 1.2059 82 0.8412 0.5572 0.8412 0.9172
No log 1.2353 84 1.0458 0.3090 1.0458 1.0226
No log 1.2647 86 1.4305 0.1269 1.4305 1.1960
No log 1.2941 88 1.2549 0.3213 1.2549 1.1202
No log 1.3235 90 0.9098 0.5386 0.9098 0.9538
No log 1.3529 92 0.8906 0.5318 0.8906 0.9437
No log 1.3824 94 0.9725 0.5380 0.9725 0.9862
No log 1.4118 96 0.9739 0.5917 0.9739 0.9869
No log 1.4412 98 0.9671 0.5429 0.9671 0.9834
No log 1.4706 100 1.1081 0.5199 1.1081 1.0526
No log 1.5 102 1.2398 0.4341 1.2398 1.1135
No log 1.5294 104 1.1402 0.4946 1.1402 1.0678
No log 1.5588 106 1.0947 0.4867 1.0947 1.0463
No log 1.5882 108 1.0192 0.5308 1.0192 1.0096
No log 1.6176 110 1.0873 0.4953 1.0873 1.0427
No log 1.6471 112 1.1152 0.4894 1.1152 1.0560
No log 1.6765 114 1.0857 0.4999 1.0857 1.0420
No log 1.7059 116 1.1194 0.4572 1.1194 1.0580
No log 1.7353 118 1.1447 0.4144 1.1447 1.0699
No log 1.7647 120 1.1796 0.3582 1.1796 1.0861
No log 1.7941 122 1.2433 0.3708 1.2433 1.1150
No log 1.8235 124 1.1862 0.3776 1.1862 1.0891
No log 1.8529 126 1.1676 0.4254 1.1676 1.0806
No log 1.8824 128 1.0388 0.4703 1.0388 1.0192
No log 1.9118 130 0.9522 0.4906 0.9522 0.9758
No log 1.9412 132 0.9314 0.5278 0.9314 0.9651
No log 1.9706 134 0.8755 0.5958 0.8755 0.9357
No log 2.0 136 0.9288 0.5636 0.9288 0.9638
No log 2.0294 138 1.1312 0.5198 1.1312 1.0636
No log 2.0588 140 1.2692 0.4564 1.2692 1.1266
No log 2.0882 142 1.0489 0.5769 1.0489 1.0242
No log 2.1176 144 0.7740 0.6408 0.7740 0.8798
No log 2.1471 146 0.7262 0.6761 0.7262 0.8522
No log 2.1765 148 0.7274 0.6730 0.7274 0.8529
No log 2.2059 150 0.7704 0.6476 0.7704 0.8777
No log 2.2353 152 1.0191 0.5473 1.0191 1.0095
No log 2.2647 154 1.1480 0.4873 1.1480 1.0715
No log 2.2941 156 0.9455 0.5533 0.9455 0.9724
No log 2.3235 158 0.7323 0.6180 0.7323 0.8558
No log 2.3529 160 0.7346 0.6180 0.7346 0.8571
No log 2.3824 162 0.9152 0.5600 0.9152 0.9567
No log 2.4118 164 0.9944 0.5376 0.9944 0.9972
No log 2.4412 166 1.0520 0.5178 1.0520 1.0257
No log 2.4706 168 1.0750 0.5242 1.0750 1.0368
No log 2.5 170 1.0496 0.5164 1.0496 1.0245
No log 2.5294 172 0.9947 0.5366 0.9947 0.9974
No log 2.5588 174 0.8463 0.6099 0.8463 0.9200
No log 2.5882 176 0.8013 0.6238 0.8013 0.8952
No log 2.6176 178 0.8353 0.5921 0.8353 0.9140
No log 2.6471 180 0.8474 0.5829 0.8474 0.9206
No log 2.6765 182 0.8332 0.6092 0.8332 0.9128
No log 2.7059 184 0.8414 0.6393 0.8414 0.9173
No log 2.7353 186 0.8391 0.6239 0.8391 0.9160
No log 2.7647 188 0.8157 0.6886 0.8157 0.9031
No log 2.7941 190 0.7529 0.6891 0.7529 0.8677
No log 2.8235 192 0.7874 0.6441 0.7874 0.8874
No log 2.8529 194 0.8337 0.6358 0.8337 0.9131
No log 2.8824 196 0.7590 0.6929 0.7590 0.8712
No log 2.9118 198 0.7826 0.6720 0.7826 0.8846
No log 2.9412 200 0.7599 0.7121 0.7599 0.8717
No log 2.9706 202 0.7373 0.6584 0.7373 0.8587
No log 3.0 204 0.7673 0.6407 0.7673 0.8759
No log 3.0294 206 0.7610 0.6341 0.7610 0.8724
No log 3.0588 208 0.9115 0.6208 0.9115 0.9547
No log 3.0882 210 1.1835 0.5022 1.1835 1.0879
No log 3.1176 212 1.3984 0.4064 1.3984 1.1825
No log 3.1471 214 1.3973 0.4430 1.3973 1.1821
No log 3.1765 216 1.1794 0.4732 1.1794 1.0860
No log 3.2059 218 0.9001 0.5534 0.9001 0.9488
No log 3.2353 220 0.7920 0.5798 0.7920 0.8899
No log 3.2647 222 0.7873 0.6125 0.7873 0.8873
No log 3.2941 224 0.8792 0.5894 0.8792 0.9377
No log 3.3235 226 0.9294 0.5803 0.9294 0.9641
No log 3.3529 228 0.8209 0.6369 0.8209 0.9060
No log 3.3824 230 0.7385 0.6664 0.7385 0.8594
No log 3.4118 232 0.7244 0.6711 0.7244 0.8511
No log 3.4412 234 0.7374 0.6576 0.7374 0.8587
No log 3.4706 236 0.8218 0.5928 0.8218 0.9066
No log 3.5 238 0.7952 0.5919 0.7952 0.8918
No log 3.5294 240 0.6972 0.6246 0.6972 0.8350
No log 3.5588 242 0.6723 0.6442 0.6723 0.8200
No log 3.5882 244 0.6790 0.6162 0.6790 0.8240
No log 3.6176 246 0.6776 0.6576 0.6776 0.8231
No log 3.6471 248 0.7143 0.6165 0.7143 0.8452
No log 3.6765 250 0.7118 0.6235 0.7118 0.8437
No log 3.7059 252 0.6937 0.6123 0.6937 0.8329
No log 3.7353 254 0.6913 0.6068 0.6913 0.8315
No log 3.7647 256 0.7062 0.5949 0.7062 0.8404
No log 3.7941 258 0.7080 0.6191 0.7080 0.8415
No log 3.8235 260 0.7110 0.6460 0.7110 0.8432
No log 3.8529 262 0.7420 0.6594 0.7420 0.8614
No log 3.8824 264 0.7431 0.6722 0.7431 0.8620
No log 3.9118 266 0.7581 0.6684 0.7581 0.8707
No log 3.9412 268 0.7766 0.6637 0.7766 0.8813
No log 3.9706 270 0.7309 0.6851 0.7309 0.8549
No log 4.0 272 0.6960 0.6920 0.6960 0.8342
No log 4.0294 274 0.6687 0.6774 0.6687 0.8178
No log 4.0588 276 0.6681 0.6136 0.6681 0.8174
No log 4.0882 278 0.6654 0.6397 0.6654 0.8157
No log 4.1176 280 0.6922 0.6338 0.6922 0.8320
No log 4.1471 282 0.7429 0.6412 0.7429 0.8619
No log 4.1765 284 0.7069 0.6743 0.7069 0.8408
No log 4.2059 286 0.6898 0.6962 0.6898 0.8305
No log 4.2353 288 0.6880 0.6787 0.6880 0.8295
No log 4.2647 290 0.6934 0.7163 0.6934 0.8327
No log 4.2941 292 0.6959 0.7107 0.6959 0.8342
No log 4.3235 294 0.6880 0.7113 0.6880 0.8295
No log 4.3529 296 0.7020 0.6594 0.7020 0.8379
No log 4.3824 298 0.7417 0.6337 0.7417 0.8612
No log 4.4118 300 0.7658 0.6361 0.7658 0.8751
No log 4.4412 302 0.7304 0.6523 0.7304 0.8547
No log 4.4706 304 0.6835 0.6736 0.6835 0.8267
No log 4.5 306 0.7060 0.6901 0.7060 0.8403
No log 4.5294 308 0.7018 0.6968 0.7018 0.8377
No log 4.5588 310 0.6577 0.6850 0.6577 0.8110
No log 4.5882 312 0.6574 0.6642 0.6574 0.8108
No log 4.6176 314 0.6861 0.6643 0.6861 0.8283
No log 4.6471 316 0.6749 0.6762 0.6749 0.8215
No log 4.6765 318 0.6578 0.6847 0.6578 0.8110
No log 4.7059 320 0.6466 0.6965 0.6466 0.8041
No log 4.7353 322 0.6635 0.6956 0.6635 0.8146
No log 4.7647 324 0.6728 0.6893 0.6728 0.8202
No log 4.7941 326 0.7095 0.6537 0.7095 0.8423
No log 4.8235 328 0.7733 0.5826 0.7733 0.8794
No log 4.8529 330 0.8025 0.5826 0.8025 0.8958
No log 4.8824 332 0.7872 0.6258 0.7872 0.8873
No log 4.9118 334 0.7553 0.6788 0.7553 0.8691
No log 4.9412 336 0.7064 0.6754 0.7064 0.8405
No log 4.9706 338 0.6927 0.6890 0.6927 0.8323
No log 5.0 340 0.6977 0.6783 0.6977 0.8353
No log 5.0294 342 0.7149 0.6645 0.7149 0.8455
No log 5.0588 344 0.7022 0.6665 0.7022 0.8380
No log 5.0882 346 0.6745 0.6857 0.6745 0.8213
No log 5.1176 348 0.7214 0.6575 0.7214 0.8494
No log 5.1471 350 0.7764 0.6577 0.7764 0.8811
No log 5.1765 352 0.8510 0.6295 0.8510 0.9225
No log 5.2059 354 0.9105 0.5705 0.9105 0.9542
No log 5.2353 356 0.8608 0.5969 0.8608 0.9278
No log 5.2647 358 0.7759 0.6478 0.7759 0.8808
No log 5.2941 360 0.7786 0.6540 0.7786 0.8824
No log 5.3235 362 0.8068 0.6401 0.8068 0.8982
No log 5.3529 364 0.8212 0.6262 0.8212 0.9062
No log 5.3824 366 0.7968 0.6648 0.7968 0.8926
No log 5.4118 368 0.7062 0.6911 0.7062 0.8403
No log 5.4412 370 0.6594 0.7193 0.6594 0.8121
No log 5.4706 372 0.6716 0.6937 0.6716 0.8195
No log 5.5 374 0.6524 0.6973 0.6524 0.8077
No log 5.5294 376 0.6869 0.7143 0.6869 0.8288
No log 5.5588 378 0.7292 0.6763 0.7292 0.8539
No log 5.5882 380 0.7464 0.6523 0.7464 0.8639
No log 5.6176 382 0.7217 0.6776 0.7217 0.8495
No log 5.6471 384 0.7162 0.6693 0.7162 0.8463
No log 5.6765 386 0.7217 0.6733 0.7217 0.8495
No log 5.7059 388 0.7160 0.7014 0.7160 0.8462
No log 5.7353 390 0.7440 0.7036 0.7440 0.8625
No log 5.7647 392 0.7778 0.6932 0.7778 0.8819
No log 5.7941 394 0.8342 0.6827 0.8342 0.9133
No log 5.8235 396 0.8633 0.6807 0.8633 0.9292
No log 5.8529 398 0.7553 0.7023 0.7553 0.8691
No log 5.8824 400 0.6701 0.7168 0.6701 0.8186
No log 5.9118 402 0.6646 0.7277 0.6646 0.8152
No log 5.9412 404 0.7387 0.6863 0.7387 0.8595
No log 5.9706 406 0.7808 0.6647 0.7808 0.8837
No log 6.0 408 0.8205 0.6148 0.8205 0.9058
No log 6.0294 410 0.7773 0.6313 0.7773 0.8817
No log 6.0588 412 0.7102 0.6624 0.7102 0.8427
No log 6.0882 414 0.6801 0.7171 0.6801 0.8247
No log 6.1176 416 0.6885 0.7122 0.6885 0.8298
No log 6.1471 418 0.7657 0.6675 0.7657 0.8751
No log 6.1765 420 0.8716 0.6208 0.8716 0.9336
No log 6.2059 422 0.8924 0.5964 0.8924 0.9447
No log 6.2353 424 0.8303 0.6505 0.8303 0.9112
No log 6.2647 426 0.7480 0.6534 0.7480 0.8649
No log 6.2941 428 0.7170 0.6684 0.7170 0.8467
No log 6.3235 430 0.7192 0.6050 0.7192 0.8480
No log 6.3529 432 0.7444 0.5423 0.7444 0.8628
No log 6.3824 434 0.8048 0.6029 0.8048 0.8971
No log 6.4118 436 0.9196 0.5771 0.9196 0.9590
No log 6.4412 438 0.9461 0.5731 0.9461 0.9727
No log 6.4706 440 0.8591 0.6256 0.8591 0.9269
No log 6.5 442 0.8141 0.6358 0.8141 0.9023
No log 6.5294 444 0.7764 0.6375 0.7764 0.8811
No log 6.5588 446 0.7011 0.6782 0.7011 0.8373
No log 6.5882 448 0.7092 0.6694 0.7092 0.8421
No log 6.6176 450 0.7301 0.6694 0.7301 0.8544
No log 6.6471 452 0.6723 0.6763 0.6723 0.8199
No log 6.6765 454 0.6227 0.7145 0.6227 0.7891
No log 6.7059 456 0.6305 0.7213 0.6305 0.7941
No log 6.7353 458 0.7036 0.6694 0.7036 0.8388
No log 6.7647 460 0.8380 0.6246 0.8380 0.9154
No log 6.7941 462 0.8748 0.5997 0.8748 0.9353
No log 6.8235 464 0.8056 0.6302 0.8056 0.8976
No log 6.8529 466 0.7969 0.6157 0.7969 0.8927
No log 6.8824 468 0.8024 0.6120 0.8024 0.8958
No log 6.9118 470 0.8029 0.6104 0.8029 0.8960
No log 6.9412 472 0.8052 0.6072 0.8052 0.8973
No log 6.9706 474 0.8506 0.6041 0.8506 0.9223
No log 7.0 476 0.9704 0.5954 0.9704 0.9851
No log 7.0294 478 1.2546 0.5219 1.2546 1.1201
No log 7.0588 480 1.3364 0.5211 1.3364 1.1560
No log 7.0882 482 1.1720 0.5354 1.1720 1.0826
No log 7.1176 484 0.9373 0.5655 0.9373 0.9681
No log 7.1471 486 0.7503 0.6261 0.7503 0.8662
No log 7.1765 488 0.7029 0.6717 0.7029 0.8384
No log 7.2059 490 0.7224 0.6578 0.7224 0.8499
No log 7.2353 492 0.7932 0.6226 0.7932 0.8906
No log 7.2647 494 0.9159 0.5782 0.9159 0.9570
No log 7.2941 496 0.9358 0.5859 0.9358 0.9674
No log 7.3235 498 0.7903 0.6348 0.7903 0.8890
0.505 7.3529 500 0.7154 0.6471 0.7154 0.8458
0.505 7.3824 502 0.7117 0.6230 0.7117 0.8436
0.505 7.4118 504 0.8115 0.6083 0.8115 0.9008
0.505 7.4412 506 0.8770 0.5854 0.8770 0.9365
0.505 7.4706 508 0.8207 0.6046 0.8207 0.9059
0.505 7.5 510 0.7490 0.6322 0.7490 0.8655

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k14_task1_organization

Finetuned
(4023)
this model