ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7870
  • Qwk: 0.6626
  • Mse: 0.7870
  • Rmse: 0.8871

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0741 2 5.6380 -0.0097 5.6380 2.3744
No log 0.1481 4 3.5427 0.0532 3.5427 1.8822
No log 0.2222 6 2.1919 0.0362 2.1919 1.4805
No log 0.2963 8 1.5661 0.1236 1.5661 1.2514
No log 0.3704 10 1.2526 0.2542 1.2526 1.1192
No log 0.4444 12 1.3778 0.2951 1.3778 1.1738
No log 0.5185 14 1.1267 0.2714 1.1267 1.0615
No log 0.5926 16 1.0877 0.1947 1.0877 1.0429
No log 0.6667 18 1.1850 0.1502 1.1850 1.0886
No log 0.7407 20 1.1222 0.2836 1.1222 1.0593
No log 0.8148 22 1.0837 0.2274 1.0837 1.0410
No log 0.8889 24 1.3161 0.2269 1.3161 1.1472
No log 0.9630 26 1.2731 0.2597 1.2731 1.1283
No log 1.0370 28 1.0464 0.3267 1.0464 1.0229
No log 1.1111 30 1.0252 0.3916 1.0252 1.0125
No log 1.1852 32 1.0098 0.3820 1.0098 1.0049
No log 1.2593 34 1.0357 0.4314 1.0357 1.0177
No log 1.3333 36 1.0017 0.4497 1.0017 1.0008
No log 1.4074 38 0.9017 0.4658 0.9017 0.9496
No log 1.4815 40 0.9321 0.3327 0.9321 0.9655
No log 1.5556 42 0.9500 0.3605 0.9500 0.9747
No log 1.6296 44 0.9107 0.4458 0.9107 0.9543
No log 1.7037 46 0.8848 0.5079 0.8848 0.9407
No log 1.7778 48 0.8349 0.5038 0.8349 0.9137
No log 1.8519 50 0.8820 0.4779 0.8820 0.9391
No log 1.9259 52 0.9445 0.4266 0.9445 0.9719
No log 2.0 54 0.9390 0.4733 0.9390 0.9690
No log 2.0741 56 0.9160 0.5116 0.9160 0.9571
No log 2.1481 58 0.9391 0.4807 0.9391 0.9691
No log 2.2222 60 0.8981 0.5353 0.8981 0.9477
No log 2.2963 62 0.8935 0.4377 0.8935 0.9452
No log 2.3704 64 0.8804 0.4090 0.8804 0.9383
No log 2.4444 66 0.9018 0.4230 0.9018 0.9496
No log 2.5185 68 0.9927 0.4272 0.9927 0.9963
No log 2.5926 70 0.8994 0.4723 0.8994 0.9483
No log 2.6667 72 0.8109 0.5109 0.8109 0.9005
No log 2.7407 74 0.8905 0.5174 0.8905 0.9437
No log 2.8148 76 0.8998 0.5769 0.8998 0.9486
No log 2.8889 78 0.8360 0.6349 0.8360 0.9143
No log 2.9630 80 0.8653 0.6350 0.8653 0.9302
No log 3.0370 82 1.0340 0.5063 1.0340 1.0169
No log 3.1111 84 1.2018 0.4787 1.2018 1.0963
No log 3.1852 86 1.2169 0.4460 1.2169 1.1031
No log 3.2593 88 0.9518 0.5518 0.9518 0.9756
No log 3.3333 90 0.7870 0.5876 0.7870 0.8871
No log 3.4074 92 0.9249 0.4617 0.9249 0.9617
No log 3.4815 94 0.8521 0.4701 0.8521 0.9231
No log 3.5556 96 0.8414 0.5530 0.8414 0.9173
No log 3.6296 98 0.9649 0.5031 0.9649 0.9823
No log 3.7037 100 0.9703 0.5194 0.9703 0.9850
No log 3.7778 102 0.8715 0.5691 0.8715 0.9335
No log 3.8519 104 0.7952 0.6232 0.7952 0.8918
No log 3.9259 106 0.7991 0.6225 0.7991 0.8939
No log 4.0 108 0.8237 0.6119 0.8237 0.9076
No log 4.0741 110 0.7943 0.6184 0.7943 0.8912
No log 4.1481 112 0.8612 0.5996 0.8612 0.9280
No log 4.2222 114 0.9626 0.5668 0.9626 0.9811
No log 4.2963 116 0.9939 0.5658 0.9939 0.9970
No log 4.3704 118 0.9425 0.5852 0.9425 0.9708
No log 4.4444 120 0.8749 0.6327 0.8749 0.9354
No log 4.5185 122 0.8572 0.5577 0.8572 0.9259
No log 4.5926 124 0.8935 0.4970 0.8935 0.9453
No log 4.6667 126 0.9593 0.5071 0.9593 0.9794
No log 4.7407 128 0.8575 0.5239 0.8575 0.9260
No log 4.8148 130 0.7406 0.5492 0.7406 0.8606
No log 4.8889 132 0.7179 0.5851 0.7179 0.8473
No log 4.9630 134 0.7106 0.6031 0.7106 0.8430
No log 5.0370 136 0.7240 0.5927 0.7240 0.8509
No log 5.1111 138 0.7372 0.6028 0.7372 0.8586
No log 5.1852 140 0.6780 0.6603 0.6780 0.8234
No log 5.2593 142 0.6860 0.6441 0.6860 0.8282
No log 5.3333 144 0.6939 0.6045 0.6939 0.8330
No log 5.4074 146 0.6940 0.6098 0.6940 0.8330
No log 5.4815 148 0.6906 0.6243 0.6906 0.8310
No log 5.5556 150 0.7250 0.6421 0.7250 0.8515
No log 5.6296 152 0.7471 0.6038 0.7471 0.8643
No log 5.7037 154 0.7044 0.6265 0.7044 0.8393
No log 5.7778 156 0.6891 0.6270 0.6891 0.8301
No log 5.8519 158 0.7329 0.6137 0.7329 0.8561
No log 5.9259 160 0.8251 0.5590 0.8251 0.9084
No log 6.0 162 0.8885 0.5595 0.8885 0.9426
No log 6.0741 164 0.9174 0.5399 0.9174 0.9578
No log 6.1481 166 0.8000 0.5729 0.8000 0.8945
No log 6.2222 168 0.6901 0.6704 0.6901 0.8307
No log 6.2963 170 0.6810 0.6555 0.6810 0.8252
No log 6.3704 172 0.6903 0.6295 0.6903 0.8308
No log 6.4444 174 0.7523 0.5812 0.7523 0.8674
No log 6.5185 176 0.8734 0.5918 0.8734 0.9346
No log 6.5926 178 0.8940 0.6060 0.8940 0.9455
No log 6.6667 180 0.7747 0.6619 0.7747 0.8802
No log 6.7407 182 0.7607 0.6339 0.7607 0.8722
No log 6.8148 184 0.6906 0.6257 0.6906 0.8310
No log 6.8889 186 0.6599 0.6661 0.6599 0.8123
No log 6.9630 188 0.6695 0.6423 0.6695 0.8183
No log 7.0370 190 0.7226 0.5659 0.7226 0.8500
No log 7.1111 192 0.7325 0.5800 0.7325 0.8559
No log 7.1852 194 0.6920 0.5758 0.6920 0.8319
No log 7.2593 196 0.6828 0.6271 0.6828 0.8263
No log 7.3333 198 0.7751 0.6104 0.7751 0.8804
No log 7.4074 200 0.9025 0.6211 0.9025 0.9500
No log 7.4815 202 0.8559 0.6540 0.8559 0.9251
No log 7.5556 204 0.7716 0.6487 0.7716 0.8784
No log 7.6296 206 0.7687 0.6551 0.7687 0.8767
No log 7.7037 208 0.8314 0.6216 0.8314 0.9118
No log 7.7778 210 0.8920 0.6097 0.8920 0.9444
No log 7.8519 212 0.9882 0.5354 0.9882 0.9941
No log 7.9259 214 0.9979 0.5347 0.9979 0.9990
No log 8.0 216 0.9420 0.4967 0.9420 0.9706
No log 8.0741 218 0.8897 0.5184 0.8897 0.9433
No log 8.1481 220 0.8020 0.5727 0.8020 0.8955
No log 8.2222 222 0.7201 0.6377 0.7201 0.8486
No log 8.2963 224 0.7001 0.6505 0.7001 0.8367
No log 8.3704 226 0.7499 0.6675 0.7499 0.8660
No log 8.4444 228 0.8097 0.6830 0.8097 0.8999
No log 8.5185 230 0.8642 0.6438 0.8642 0.9296
No log 8.5926 232 0.8270 0.6438 0.8270 0.9094
No log 8.6667 234 0.7491 0.6835 0.7491 0.8655
No log 8.7407 236 0.7984 0.6438 0.7984 0.8935
No log 8.8148 238 0.8615 0.5826 0.8615 0.9281
No log 8.8889 240 0.8361 0.5917 0.8361 0.9144
No log 8.9630 242 0.7059 0.6615 0.7059 0.8402
No log 9.0370 244 0.6498 0.6765 0.6498 0.8061
No log 9.1111 246 0.6413 0.6589 0.6413 0.8008
No log 9.1852 248 0.6571 0.6722 0.6571 0.8106
No log 9.2593 250 0.6916 0.6783 0.6916 0.8316
No log 9.3333 252 0.8144 0.5867 0.8144 0.9024
No log 9.4074 254 0.8888 0.5759 0.8888 0.9428
No log 9.4815 256 0.8495 0.5813 0.8495 0.9217
No log 9.5556 258 0.8362 0.5731 0.8362 0.9145
No log 9.6296 260 0.7693 0.6033 0.7693 0.8771
No log 9.7037 262 0.7058 0.6210 0.7058 0.8401
No log 9.7778 264 0.6872 0.6288 0.6872 0.8290
No log 9.8519 266 0.7181 0.6385 0.7181 0.8474
No log 9.9259 268 0.8988 0.5638 0.8988 0.9480
No log 10.0 270 1.0375 0.5309 1.0375 1.0186
No log 10.0741 272 0.9434 0.5519 0.9434 0.9713
No log 10.1481 274 0.8060 0.5629 0.8060 0.8978
No log 10.2222 276 0.7228 0.5797 0.7228 0.8502
No log 10.2963 278 0.7044 0.6106 0.7044 0.8393
No log 10.3704 280 0.7055 0.5787 0.7055 0.8399
No log 10.4444 282 0.7283 0.5541 0.7283 0.8534
No log 10.5185 284 0.7216 0.5529 0.7216 0.8495
No log 10.5926 286 0.6904 0.5981 0.6904 0.8309
No log 10.6667 288 0.7233 0.5686 0.7233 0.8505
No log 10.7407 290 0.7389 0.6083 0.7389 0.8596
No log 10.8148 292 0.7423 0.6544 0.7423 0.8616
No log 10.8889 294 0.7324 0.6585 0.7324 0.8558
No log 10.9630 296 0.6886 0.6652 0.6886 0.8298
No log 11.0370 298 0.6971 0.6682 0.6971 0.8349
No log 11.1111 300 0.7259 0.6508 0.7259 0.8520
No log 11.1852 302 0.7962 0.6357 0.7962 0.8923
No log 11.2593 304 0.8165 0.5852 0.8165 0.9036
No log 11.3333 306 0.7356 0.5771 0.7356 0.8576
No log 11.4074 308 0.6791 0.6560 0.6791 0.8241
No log 11.4815 310 0.6787 0.6302 0.6787 0.8238
No log 11.5556 312 0.6892 0.6499 0.6892 0.8302
No log 11.6296 314 0.7282 0.6779 0.7282 0.8534
No log 11.7037 316 0.7018 0.6555 0.7018 0.8377
No log 11.7778 318 0.6797 0.6191 0.6797 0.8244
No log 11.8519 320 0.7578 0.5572 0.7578 0.8705
No log 11.9259 322 0.7925 0.5974 0.7925 0.8902
No log 12.0 324 0.7378 0.5686 0.7378 0.8589
No log 12.0741 326 0.7028 0.6164 0.7028 0.8383
No log 12.1481 328 0.7144 0.6021 0.7144 0.8452
No log 12.2222 330 0.7646 0.5782 0.7646 0.8744
No log 12.2963 332 0.9093 0.5977 0.9093 0.9536
No log 12.3704 334 0.9947 0.5920 0.9947 0.9974
No log 12.4444 336 0.9642 0.5967 0.9642 0.9819
No log 12.5185 338 0.8687 0.6215 0.8687 0.9320
No log 12.5926 340 0.8148 0.6123 0.8148 0.9027
No log 12.6667 342 0.8114 0.6002 0.8114 0.9008
No log 12.7407 344 0.8386 0.5602 0.8386 0.9157
No log 12.8148 346 0.8601 0.5538 0.8601 0.9274
No log 12.8889 348 0.8675 0.5330 0.8675 0.9314
No log 12.9630 350 0.8661 0.5422 0.8661 0.9306
No log 13.0370 352 0.8932 0.5564 0.8932 0.9451
No log 13.1111 354 0.9880 0.5586 0.9880 0.9940
No log 13.1852 356 1.0882 0.5485 1.0882 1.0432
No log 13.2593 358 1.1503 0.5194 1.1503 1.0725
No log 13.3333 360 1.1009 0.5388 1.1009 1.0493
No log 13.4074 362 0.9265 0.5632 0.9265 0.9625
No log 13.4815 364 0.7939 0.6537 0.7939 0.8910
No log 13.5556 366 0.7783 0.6415 0.7783 0.8822
No log 13.6296 368 0.8349 0.5896 0.8349 0.9137
No log 13.7037 370 0.8271 0.5730 0.8271 0.9095
No log 13.7778 372 0.7538 0.5598 0.7538 0.8682
No log 13.8519 374 0.6715 0.6606 0.6715 0.8195
No log 13.9259 376 0.6576 0.6514 0.6576 0.8109
No log 14.0 378 0.7079 0.6535 0.7079 0.8414
No log 14.0741 380 0.7007 0.6627 0.7007 0.8371
No log 14.1481 382 0.6572 0.6568 0.6572 0.8107
No log 14.2222 384 0.6700 0.6663 0.6700 0.8185
No log 14.2963 386 0.7346 0.6321 0.7346 0.8571
No log 14.3704 388 0.7328 0.6244 0.7328 0.8560
No log 14.4444 390 0.7006 0.6212 0.7006 0.8370
No log 14.5185 392 0.6814 0.6642 0.6814 0.8255
No log 14.5926 394 0.6720 0.6499 0.6720 0.8198
No log 14.6667 396 0.6852 0.6632 0.6852 0.8278
No log 14.7407 398 0.7178 0.5728 0.7178 0.8472
No log 14.8148 400 0.8076 0.5863 0.8076 0.8987
No log 14.8889 402 0.8281 0.5744 0.8281 0.9100
No log 14.9630 404 0.7810 0.5494 0.7810 0.8838
No log 15.0370 406 0.7787 0.5630 0.7787 0.8824
No log 15.1111 408 0.7969 0.5811 0.7969 0.8927
No log 15.1852 410 0.8528 0.5851 0.8528 0.9235
No log 15.2593 412 0.8618 0.5744 0.8618 0.9283
No log 15.3333 414 0.9010 0.5745 0.9010 0.9492
No log 15.4074 416 0.8726 0.6000 0.8726 0.9341
No log 15.4815 418 0.8276 0.6130 0.8276 0.9097
No log 15.5556 420 0.7811 0.6120 0.7811 0.8838
No log 15.6296 422 0.7917 0.5962 0.7917 0.8898
No log 15.7037 424 0.8765 0.5624 0.8765 0.9362
No log 15.7778 426 0.9493 0.5268 0.9493 0.9743
No log 15.8519 428 0.9382 0.5627 0.9382 0.9686
No log 15.9259 430 0.8152 0.6274 0.8152 0.9029
No log 16.0 432 0.7862 0.6757 0.7862 0.8867
No log 16.0741 434 0.7911 0.6677 0.7911 0.8895
No log 16.1481 436 0.8194 0.6387 0.8194 0.9052
No log 16.2222 438 0.8748 0.5996 0.8748 0.9353
No log 16.2963 440 0.9614 0.5785 0.9614 0.9805
No log 16.3704 442 0.9787 0.5824 0.9787 0.9893
No log 16.4444 444 0.8889 0.5829 0.8889 0.9428
No log 16.5185 446 0.7808 0.6227 0.7808 0.8836
No log 16.5926 448 0.7433 0.6346 0.7433 0.8622
No log 16.6667 450 0.7489 0.6355 0.7489 0.8654
No log 16.7407 452 0.7935 0.6658 0.7935 0.8908
No log 16.8148 454 0.9004 0.6053 0.9004 0.9489
No log 16.8889 456 0.9501 0.6039 0.9501 0.9747
No log 16.9630 458 0.9403 0.5630 0.9403 0.9697
No log 17.0370 460 0.8443 0.5731 0.8443 0.9189
No log 17.1111 462 0.7475 0.6062 0.7475 0.8646
No log 17.1852 464 0.7145 0.6081 0.7145 0.8453
No log 17.2593 466 0.7316 0.5698 0.7316 0.8553
No log 17.3333 468 0.7602 0.5770 0.7602 0.8719
No log 17.4074 470 0.7581 0.5742 0.7581 0.8707
No log 17.4815 472 0.7758 0.6149 0.7758 0.8808
No log 17.5556 474 0.8103 0.6262 0.8103 0.9002
No log 17.6296 476 0.8232 0.6588 0.8232 0.9073
No log 17.7037 478 0.8670 0.6657 0.8670 0.9311
No log 17.7778 480 0.8579 0.6315 0.8579 0.9262
No log 17.8519 482 0.8324 0.6173 0.8324 0.9124
No log 17.9259 484 0.7704 0.6047 0.7704 0.8777
No log 18.0 486 0.7211 0.5920 0.7211 0.8491
No log 18.0741 488 0.7235 0.5800 0.7235 0.8506
No log 18.1481 490 0.7359 0.6194 0.7359 0.8578
No log 18.2222 492 0.7339 0.6348 0.7339 0.8567
No log 18.2963 494 0.6980 0.6224 0.6980 0.8354
No log 18.3704 496 0.6885 0.6632 0.6885 0.8297
No log 18.4444 498 0.7206 0.6613 0.7206 0.8489
0.3349 18.5185 500 0.7740 0.6808 0.7740 0.8798
0.3349 18.5926 502 0.8257 0.6751 0.8257 0.9087
0.3349 18.6667 504 0.8623 0.6250 0.8623 0.9286
0.3349 18.7407 506 0.8453 0.6493 0.8453 0.9194
0.3349 18.8148 508 0.7794 0.6675 0.7794 0.8828
0.3349 18.8889 510 0.7870 0.6626 0.7870 0.8871

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k5_task1_organization

Finetuned
(4023)
this model