ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8371
  • Qwk: 0.4470
  • Mse: 0.8371
  • Rmse: 0.9149

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0222 2 4.6481 0.0042 4.6481 2.1560
No log 0.0444 4 2.7766 -0.0144 2.7766 1.6663
No log 0.0667 6 1.6420 0.0372 1.6420 1.2814
No log 0.0889 8 1.2712 0.0 1.2712 1.1275
No log 0.1111 10 1.1925 0.1927 1.1925 1.0920
No log 0.1333 12 1.1866 0.1247 1.1866 1.0893
No log 0.1556 14 1.2291 0.1076 1.2291 1.1086
No log 0.1778 16 1.2216 0.1482 1.2216 1.1053
No log 0.2 18 1.2603 0.1237 1.2603 1.1227
No log 0.2222 20 1.5055 0.0210 1.5055 1.2270
No log 0.2444 22 1.4686 0.0210 1.4686 1.2119
No log 0.2667 24 1.1593 0.1990 1.1593 1.0767
No log 0.2889 26 0.9927 0.2813 0.9927 0.9964
No log 0.3111 28 0.9690 0.3250 0.9690 0.9844
No log 0.3333 30 0.9636 0.3421 0.9636 0.9816
No log 0.3556 32 0.9606 0.3421 0.9606 0.9801
No log 0.3778 34 1.0148 0.2155 1.0148 1.0074
No log 0.4 36 1.2398 0.1748 1.2398 1.1135
No log 0.4222 38 1.2501 0.2654 1.2501 1.1181
No log 0.4444 40 1.1428 0.3651 1.1428 1.0690
No log 0.4667 42 1.0686 0.3579 1.0686 1.0337
No log 0.4889 44 0.9718 0.3577 0.9718 0.9858
No log 0.5111 46 0.9830 0.3772 0.9830 0.9914
No log 0.5333 48 1.0760 0.3412 1.0760 1.0373
No log 0.5556 50 1.0675 0.3103 1.0675 1.0332
No log 0.5778 52 1.0547 0.3703 1.0547 1.0270
No log 0.6 54 1.1280 0.3742 1.1280 1.0621
No log 0.6222 56 1.1456 0.3993 1.1456 1.0703
No log 0.6444 58 1.0893 0.4149 1.0893 1.0437
No log 0.6667 60 1.0268 0.4302 1.0268 1.0133
No log 0.6889 62 0.9643 0.4597 0.9643 0.9820
No log 0.7111 64 0.9402 0.4517 0.9402 0.9696
No log 0.7333 66 0.9926 0.5522 0.9926 0.9963
No log 0.7556 68 1.0236 0.5421 1.0236 1.0117
No log 0.7778 70 1.1205 0.4425 1.1205 1.0585
No log 0.8 72 1.1524 0.5122 1.1524 1.0735
No log 0.8222 74 1.1927 0.4452 1.1927 1.0921
No log 0.8444 76 1.2278 0.3856 1.2278 1.1081
No log 0.8667 78 1.1127 0.3864 1.1127 1.0548
No log 0.8889 80 1.0070 0.2757 1.0070 1.0035
No log 0.9111 82 0.8779 0.4366 0.8779 0.9370
No log 0.9333 84 0.8332 0.5184 0.8332 0.9128
No log 0.9556 86 0.8619 0.5562 0.8619 0.9284
No log 0.9778 88 0.9037 0.5025 0.9037 0.9506
No log 1.0 90 0.7887 0.5661 0.7887 0.8881
No log 1.0222 92 0.8138 0.4942 0.8138 0.9021
No log 1.0444 94 1.2119 0.4992 1.2119 1.1009
No log 1.0667 96 1.3989 0.4474 1.3989 1.1828
No log 1.0889 98 1.2630 0.5249 1.2630 1.1238
No log 1.1111 100 0.9547 0.5406 0.9547 0.9771
No log 1.1333 102 0.8993 0.5192 0.8993 0.9483
No log 1.1556 104 0.8576 0.4925 0.8576 0.9260
No log 1.1778 106 0.7692 0.5589 0.7692 0.8770
No log 1.2 108 0.9510 0.5375 0.9510 0.9752
No log 1.2222 110 1.1012 0.4758 1.1012 1.0494
No log 1.2444 112 1.0767 0.4856 1.0767 1.0376
No log 1.2667 114 0.8914 0.5474 0.8914 0.9442
No log 1.2889 116 0.7507 0.5571 0.7507 0.8664
No log 1.3111 118 0.7597 0.5624 0.7597 0.8716
No log 1.3333 120 0.8881 0.5002 0.8881 0.9424
No log 1.3556 122 1.0167 0.4828 1.0167 1.0083
No log 1.3778 124 0.9362 0.5395 0.9362 0.9676
No log 1.4 126 0.7762 0.5479 0.7762 0.8810
No log 1.4222 128 0.7083 0.6199 0.7083 0.8416
No log 1.4444 130 0.7033 0.6186 0.7033 0.8386
No log 1.4667 132 0.6792 0.6333 0.6792 0.8241
No log 1.4889 134 0.7499 0.6140 0.7499 0.8660
No log 1.5111 136 0.7921 0.6201 0.7921 0.8900
No log 1.5333 138 0.9814 0.4808 0.9814 0.9906
No log 1.5556 140 1.0355 0.5027 1.0355 1.0176
No log 1.5778 142 0.8345 0.5743 0.8345 0.9135
No log 1.6 144 0.7072 0.5262 0.7072 0.8410
No log 1.6222 146 0.7186 0.6622 0.7186 0.8477
No log 1.6444 148 0.6998 0.6578 0.6998 0.8366
No log 1.6667 150 0.7130 0.6067 0.7130 0.8444
No log 1.6889 152 0.8901 0.5626 0.8901 0.9434
No log 1.7111 154 1.0428 0.5647 1.0428 1.0212
No log 1.7333 156 0.9526 0.5798 0.9526 0.9760
No log 1.7556 158 0.7541 0.5896 0.7541 0.8684
No log 1.7778 160 0.7375 0.6352 0.7375 0.8588
No log 1.8 162 0.7297 0.6160 0.7297 0.8542
No log 1.8222 164 0.7659 0.5647 0.7659 0.8752
No log 1.8444 166 0.9321 0.5253 0.9321 0.9655
No log 1.8667 168 1.0018 0.5015 1.0018 1.0009
No log 1.8889 170 0.9563 0.5177 0.9563 0.9779
No log 1.9111 172 0.8731 0.5408 0.8731 0.9344
No log 1.9333 174 0.9032 0.5705 0.9032 0.9504
No log 1.9556 176 0.9368 0.5583 0.9368 0.9679
No log 1.9778 178 0.9037 0.5262 0.9037 0.9506
No log 2.0 180 0.8453 0.5105 0.8453 0.9194
No log 2.0222 182 0.8843 0.5094 0.8843 0.9404
No log 2.0444 184 0.8478 0.4946 0.8478 0.9208
No log 2.0667 186 0.8228 0.5086 0.8228 0.9071
No log 2.0889 188 0.7906 0.4940 0.7906 0.8892
No log 2.1111 190 0.8633 0.5570 0.8633 0.9291
No log 2.1333 192 1.1160 0.5073 1.1160 1.0564
No log 2.1556 194 1.3259 0.4211 1.3259 1.1515
No log 2.1778 196 1.3175 0.4186 1.3175 1.1478
No log 2.2 198 1.2211 0.3839 1.2211 1.1050
No log 2.2222 200 1.0741 0.5053 1.0741 1.0364
No log 2.2444 202 0.8963 0.5125 0.8963 0.9467
No log 2.2667 204 0.8107 0.6035 0.8107 0.9004
No log 2.2889 206 0.7921 0.5426 0.7921 0.8900
No log 2.3111 208 0.7874 0.4858 0.7874 0.8874
No log 2.3333 210 0.8601 0.4507 0.8601 0.9274
No log 2.3556 212 0.9838 0.5109 0.9838 0.9919
No log 2.3778 214 1.0661 0.5388 1.0661 1.0325
No log 2.4 216 1.0417 0.5015 1.0417 1.0206
No log 2.4222 218 0.8830 0.5547 0.8830 0.9397
No log 2.4444 220 0.7883 0.5245 0.7883 0.8879
No log 2.4667 222 0.8062 0.4968 0.8062 0.8979
No log 2.4889 224 0.8141 0.4639 0.8141 0.9023
No log 2.5111 226 0.9224 0.4430 0.9224 0.9604
No log 2.5333 228 0.9639 0.4856 0.9639 0.9818
No log 2.5556 230 0.8655 0.4879 0.8655 0.9303
No log 2.5778 232 0.8073 0.5788 0.8073 0.8985
No log 2.6 234 0.7960 0.5578 0.7960 0.8922
No log 2.6222 236 0.8748 0.5232 0.8748 0.9353
No log 2.6444 238 0.8588 0.4987 0.8588 0.9267
No log 2.6667 240 0.7459 0.5157 0.7459 0.8637
No log 2.6889 242 0.6821 0.6305 0.6821 0.8259
No log 2.7111 244 0.6847 0.6324 0.6847 0.8275
No log 2.7333 246 0.8022 0.5176 0.8022 0.8957
No log 2.7556 248 1.0744 0.4919 1.0744 1.0365
No log 2.7778 250 1.2172 0.4980 1.2172 1.1033
No log 2.8 252 1.2932 0.4884 1.2932 1.1372
No log 2.8222 254 1.1848 0.4884 1.1848 1.0885
No log 2.8444 256 0.9421 0.5246 0.9421 0.9706
No log 2.8667 258 0.7933 0.5556 0.7933 0.8907
No log 2.8889 260 0.7850 0.5683 0.7850 0.8860
No log 2.9111 262 0.8790 0.6004 0.8790 0.9376
No log 2.9333 264 0.8868 0.6004 0.8868 0.9417
No log 2.9556 266 0.8053 0.5166 0.8053 0.8974
No log 2.9778 268 0.6807 0.6060 0.6807 0.8251
No log 3.0 270 0.6530 0.6181 0.6530 0.8081
No log 3.0222 272 0.6543 0.6204 0.6543 0.8089
No log 3.0444 274 0.7449 0.5294 0.7449 0.8631
No log 3.0667 276 0.8928 0.5342 0.8928 0.9449
No log 3.0889 278 0.8723 0.5672 0.8723 0.9340
No log 3.1111 280 0.7540 0.5073 0.7540 0.8683
No log 3.1333 282 0.6959 0.5763 0.6959 0.8342
No log 3.1556 284 0.7241 0.5213 0.7241 0.8509
No log 3.1778 286 0.7939 0.4916 0.7939 0.8910
No log 3.2 288 0.8024 0.5124 0.8024 0.8958
No log 3.2222 290 0.7608 0.5089 0.7608 0.8722
No log 3.2444 292 0.7623 0.5073 0.7623 0.8731
No log 3.2667 294 0.7937 0.5548 0.7937 0.8909
No log 3.2889 296 0.8563 0.5505 0.8563 0.9253
No log 3.3111 298 0.8914 0.6075 0.8914 0.9442
No log 3.3333 300 0.9615 0.6146 0.9615 0.9805
No log 3.3556 302 1.1325 0.5388 1.1325 1.0642
No log 3.3778 304 1.2992 0.4668 1.2992 1.1398
No log 3.4 306 1.2033 0.4934 1.2033 1.0970
No log 3.4222 308 1.0794 0.5215 1.0794 1.0389
No log 3.4444 310 0.8975 0.5857 0.8975 0.9474
No log 3.4667 312 0.7988 0.5917 0.7988 0.8938
No log 3.4889 314 0.7826 0.5514 0.7826 0.8846
No log 3.5111 316 0.7886 0.5094 0.7886 0.8880
No log 3.5333 318 0.8639 0.4869 0.8639 0.9295
No log 3.5556 320 1.0485 0.5375 1.0485 1.0240
No log 3.5778 322 1.1973 0.4833 1.1973 1.0942
No log 3.6 324 1.1944 0.5200 1.1944 1.0929
No log 3.6222 326 1.0397 0.5202 1.0397 1.0196
No log 3.6444 328 0.9352 0.5675 0.9352 0.9671
No log 3.6667 330 1.1016 0.4788 1.1016 1.0496
No log 3.6889 332 1.0221 0.4955 1.0221 1.0110
No log 3.7111 334 0.8224 0.5183 0.8224 0.9069
No log 3.7333 336 0.8024 0.4853 0.8024 0.8957
No log 3.7556 338 0.9164 0.5166 0.9164 0.9573
No log 3.7778 340 0.9781 0.4733 0.9781 0.9890
No log 3.8 342 0.9420 0.4527 0.9420 0.9706
No log 3.8222 344 0.8312 0.4420 0.8312 0.9117
No log 3.8444 346 0.7938 0.4937 0.7938 0.8910
No log 3.8667 348 0.8073 0.4923 0.8073 0.8985
No log 3.8889 350 0.8884 0.5451 0.8884 0.9425
No log 3.9111 352 0.9534 0.5208 0.9534 0.9764
No log 3.9333 354 0.9385 0.5201 0.9385 0.9688
No log 3.9556 356 0.8920 0.5054 0.8920 0.9445
No log 3.9778 358 0.9041 0.5056 0.9041 0.9509
No log 4.0 360 0.8645 0.4822 0.8645 0.9298
No log 4.0222 362 0.7968 0.4587 0.7968 0.8926
No log 4.0444 364 0.7445 0.4767 0.7445 0.8628
No log 4.0667 366 0.6868 0.5811 0.6868 0.8287
No log 4.0889 368 0.6780 0.5121 0.6780 0.8234
No log 4.1111 370 0.6768 0.6139 0.6768 0.8227
No log 4.1333 372 0.6970 0.6077 0.6970 0.8348
No log 4.1556 374 0.7939 0.5913 0.7939 0.8910
No log 4.1778 376 0.8659 0.5774 0.8659 0.9305
No log 4.2 378 0.8970 0.5392 0.8970 0.9471
No log 4.2222 380 0.8638 0.4948 0.8638 0.9294
No log 4.2444 382 0.8047 0.5359 0.8047 0.8970
No log 4.2667 384 0.7925 0.5381 0.7925 0.8902
No log 4.2889 386 0.7843 0.5510 0.7843 0.8856
No log 4.3111 388 0.7761 0.5510 0.7761 0.8810
No log 4.3333 390 0.7889 0.4840 0.7889 0.8882
No log 4.3556 392 0.8530 0.5086 0.8530 0.9236
No log 4.3778 394 0.8952 0.5044 0.8952 0.9461
No log 4.4 396 0.8971 0.5250 0.8971 0.9471
No log 4.4222 398 0.8696 0.5814 0.8696 0.9325
No log 4.4444 400 0.8451 0.5587 0.8451 0.9193
No log 4.4667 402 0.8236 0.5040 0.8236 0.9075
No log 4.4889 404 0.8325 0.5395 0.8325 0.9124
No log 4.5111 406 0.8407 0.5648 0.8407 0.9169
No log 4.5333 408 0.8224 0.5561 0.8224 0.9069
No log 4.5556 410 0.7673 0.5239 0.7673 0.8759
No log 4.5778 412 0.7460 0.4976 0.7460 0.8637
No log 4.6 414 0.7495 0.4912 0.7495 0.8658
No log 4.6222 416 0.7807 0.4681 0.7807 0.8836
No log 4.6444 418 0.8703 0.5086 0.8703 0.9329
No log 4.6667 420 1.0421 0.5273 1.0421 1.0208
No log 4.6889 422 1.0988 0.5409 1.0988 1.0482
No log 4.7111 424 1.0431 0.5562 1.0431 1.0213
No log 4.7333 426 1.0034 0.5406 1.0034 1.0017
No log 4.7556 428 0.8937 0.5135 0.8937 0.9453
No log 4.7778 430 0.8524 0.5029 0.8524 0.9233
No log 4.8 432 0.8710 0.4973 0.8710 0.9333
No log 4.8222 434 0.8792 0.4973 0.8792 0.9377
No log 4.8444 436 0.8683 0.4973 0.8683 0.9318
No log 4.8667 438 0.8459 0.4940 0.8459 0.9197
No log 4.8889 440 0.8757 0.5118 0.8757 0.9358
No log 4.9111 442 0.9407 0.4957 0.9407 0.9699
No log 4.9333 444 0.9400 0.5180 0.9400 0.9695
No log 4.9556 446 0.8659 0.4743 0.8659 0.9306
No log 4.9778 448 0.8041 0.4879 0.8041 0.8967
No log 5.0 450 0.8031 0.5195 0.8031 0.8962
No log 5.0222 452 0.8047 0.5143 0.8047 0.8970
No log 5.0444 454 0.8691 0.4429 0.8691 0.9322
No log 5.0667 456 1.0733 0.3537 1.0733 1.0360
No log 5.0889 458 1.2104 0.3867 1.2104 1.1002
No log 5.1111 460 1.1630 0.3872 1.1630 1.0784
No log 5.1333 462 0.9863 0.4817 0.9863 0.9931
No log 5.1556 464 0.7918 0.4889 0.7918 0.8898
No log 5.1778 466 0.7527 0.4548 0.7527 0.8676
No log 5.2 468 0.7419 0.5076 0.7419 0.8614
No log 5.2222 470 0.7647 0.4983 0.7647 0.8745
No log 5.2444 472 0.8059 0.5625 0.8059 0.8977
No log 5.2667 474 0.7824 0.5521 0.7824 0.8845
No log 5.2889 476 0.7627 0.5287 0.7627 0.8733
No log 5.3111 478 0.7502 0.5287 0.7502 0.8661
No log 5.3333 480 0.7416 0.4965 0.7416 0.8612
No log 5.3556 482 0.8083 0.5710 0.8083 0.8991
No log 5.3778 484 0.9211 0.5174 0.9211 0.9597
No log 5.4 486 0.9299 0.4444 0.9299 0.9643
No log 5.4222 488 0.8939 0.4972 0.8939 0.9455
No log 5.4444 490 0.8464 0.4754 0.8464 0.9200
No log 5.4667 492 0.8459 0.4297 0.8459 0.9197
No log 5.4889 494 0.8774 0.4261 0.8774 0.9367
No log 5.5111 496 0.9507 0.4339 0.9507 0.9750
No log 5.5333 498 1.0271 0.4511 1.0271 1.0135
0.3479 5.5556 500 1.0747 0.4841 1.0747 1.0367
0.3479 5.5778 502 0.9966 0.4841 0.9966 0.9983
0.3479 5.6 504 0.8701 0.4655 0.8701 0.9328
0.3479 5.6222 506 0.8282 0.4337 0.8282 0.9101
0.3479 5.6444 508 0.8161 0.4832 0.8161 0.9034
0.3479 5.6667 510 0.8371 0.4470 0.8371 0.9149

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task2_organization

Finetuned
(4019)
this model