ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9864
  • Qwk: 0.3486
  • Mse: 0.9864
  • Rmse: 0.9932

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.025 2 2.5252 -0.0084 2.5252 1.5891
No log 0.05 4 1.8987 0.0182 1.8987 1.3779
No log 0.075 6 0.9584 0.2222 0.9584 0.9790
No log 0.1 8 0.7997 -0.0027 0.7997 0.8943
No log 0.125 10 1.0554 -0.0796 1.0554 1.0273
No log 0.15 12 1.5820 -0.1503 1.5820 1.2578
No log 0.175 14 1.2063 0.0401 1.2063 1.0983
No log 0.2 16 0.7981 0.1617 0.7981 0.8934
No log 0.225 18 1.0316 0.1709 1.0316 1.0157
No log 0.25 20 1.0739 0.2100 1.0739 1.0363
No log 0.275 22 0.8656 0.1407 0.8656 0.9304
No log 0.3 24 0.7604 0.0 0.7604 0.8720
No log 0.325 26 0.7139 0.1983 0.7139 0.8449
No log 0.35 28 0.7641 0.3387 0.7641 0.8741
No log 0.375 30 0.9344 0.2939 0.9344 0.9667
No log 0.4 32 0.9883 0.1795 0.9883 0.9941
No log 0.425 34 0.8183 0.3347 0.8183 0.9046
No log 0.45 36 0.6845 0.3746 0.6845 0.8273
No log 0.475 38 0.7440 0.4684 0.7440 0.8626
No log 0.5 40 0.7747 0.4464 0.7747 0.8802
No log 0.525 42 0.7151 0.4743 0.7151 0.8456
No log 0.55 44 0.6879 0.3677 0.6879 0.8294
No log 0.575 46 0.6885 0.2545 0.6885 0.8298
No log 0.6 48 0.6951 0.1548 0.6951 0.8337
No log 0.625 50 0.7130 0.2264 0.7130 0.8444
No log 0.65 52 0.6774 0.2947 0.6774 0.8230
No log 0.675 54 0.6489 0.3481 0.6489 0.8055
No log 0.7 56 0.6692 0.4895 0.6692 0.8180
No log 0.725 58 0.8168 0.2756 0.8168 0.9038
No log 0.75 60 0.8265 0.2866 0.8265 0.9091
No log 0.775 62 0.8331 0.3688 0.8331 0.9127
No log 0.8 64 0.8181 0.3889 0.8181 0.9045
No log 0.825 66 0.7846 0.3866 0.7846 0.8858
No log 0.85 68 0.7405 0.3486 0.7405 0.8605
No log 0.875 70 0.6954 0.2962 0.6954 0.8339
No log 0.9 72 0.6656 0.4029 0.6656 0.8158
No log 0.925 74 0.6034 0.3887 0.6034 0.7768
No log 0.95 76 0.6013 0.4037 0.6013 0.7755
No log 0.975 78 0.6751 0.4322 0.6751 0.8216
No log 1.0 80 0.8338 0.3285 0.8338 0.9131
No log 1.025 82 1.0429 0.3430 1.0429 1.0212
No log 1.05 84 0.8773 0.4026 0.8773 0.9367
No log 1.075 86 0.7567 0.4263 0.7567 0.8699
No log 1.1 88 0.6855 0.4740 0.6855 0.8280
No log 1.125 90 0.6672 0.4582 0.6672 0.8168
No log 1.15 92 0.6548 0.4434 0.6548 0.8092
No log 1.175 94 0.6808 0.5195 0.6808 0.8251
No log 1.2 96 0.9923 0.3884 0.9923 0.9961
No log 1.225 98 1.3593 0.2609 1.3593 1.1659
No log 1.25 100 1.1351 0.2733 1.1351 1.0654
No log 1.275 102 0.8094 0.4018 0.8094 0.8997
No log 1.3 104 0.6745 0.5045 0.6745 0.8213
No log 1.325 106 0.7770 0.3439 0.7770 0.8815
No log 1.35 108 1.0691 0.3134 1.0691 1.0340
No log 1.375 110 0.9991 0.3922 0.9991 0.9995
No log 1.4 112 0.8627 0.4032 0.8627 0.9288
No log 1.425 114 0.7764 0.4385 0.7764 0.8811
No log 1.45 116 0.8964 0.3810 0.8964 0.9468
No log 1.475 118 0.8481 0.3889 0.8481 0.9209
No log 1.5 120 0.8046 0.3934 0.8046 0.8970
No log 1.525 122 0.8026 0.3115 0.8026 0.8959
No log 1.55 124 0.7802 0.4207 0.7802 0.8833
No log 1.575 126 0.7358 0.4135 0.7358 0.8578
No log 1.6 128 0.8287 0.5023 0.8287 0.9103
No log 1.625 130 1.0737 0.3935 1.0737 1.0362
No log 1.65 132 1.4205 0.3332 1.4205 1.1918
No log 1.675 134 1.3532 0.2495 1.3532 1.1633
No log 1.7 136 1.1357 0.3506 1.1357 1.0657
No log 1.725 138 0.7570 0.4937 0.7570 0.8701
No log 1.75 140 0.6197 0.5272 0.6197 0.7872
No log 1.775 142 0.6288 0.5030 0.6288 0.7930
No log 1.8 144 0.6899 0.4864 0.6899 0.8306
No log 1.825 146 0.8498 0.4142 0.8498 0.9218
No log 1.85 148 1.2018 0.2650 1.2018 1.0962
No log 1.875 150 1.1747 0.2459 1.1747 1.0838
No log 1.9 152 0.9695 0.3079 0.9695 0.9846
No log 1.925 154 0.7473 0.4556 0.7473 0.8645
No log 1.95 156 0.7411 0.5117 0.7411 0.8609
No log 1.975 158 0.8739 0.4304 0.8739 0.9348
No log 2.0 160 1.1522 0.2772 1.1522 1.0734
No log 2.025 162 1.5588 0.1529 1.5588 1.2485
No log 2.05 164 1.6385 0.2139 1.6385 1.2800
No log 2.075 166 1.4412 0.2040 1.4412 1.2005
No log 2.1 168 0.8507 0.3933 0.8507 0.9224
No log 2.125 170 0.6758 0.3302 0.6758 0.8221
No log 2.15 172 0.7215 0.3494 0.7215 0.8494
No log 2.175 174 0.8381 0.3998 0.8381 0.9155
No log 2.2 176 0.8758 0.3538 0.8758 0.9358
No log 2.225 178 0.7438 0.4574 0.7438 0.8624
No log 2.25 180 0.7478 0.4432 0.7478 0.8648
No log 2.275 182 0.7108 0.4133 0.7108 0.8431
No log 2.3 184 0.8215 0.4208 0.8215 0.9064
No log 2.325 186 1.0082 0.3161 1.0082 1.0041
No log 2.35 188 0.9980 0.3316 0.9980 0.9990
No log 2.375 190 0.8213 0.4268 0.8213 0.9063
No log 2.4 192 0.8069 0.4400 0.8069 0.8983
No log 2.425 194 0.8596 0.4601 0.8596 0.9271
No log 2.45 196 1.0893 0.5285 1.0893 1.0437
No log 2.475 198 1.3747 0.3255 1.3747 1.1725
No log 2.5 200 1.3379 0.2971 1.3379 1.1567
No log 2.525 202 1.0751 0.4090 1.0751 1.0369
No log 2.55 204 0.8018 0.4597 0.8018 0.8954
No log 2.575 206 0.6503 0.4044 0.6503 0.8064
No log 2.6 208 0.6588 0.3798 0.6588 0.8117
No log 2.625 210 0.7328 0.4223 0.7328 0.8560
No log 2.65 212 0.8308 0.4396 0.8308 0.9115
No log 2.675 214 0.9404 0.4608 0.9404 0.9697
No log 2.7 216 0.7848 0.5339 0.7848 0.8859
No log 2.725 218 0.6888 0.5179 0.6888 0.8299
No log 2.75 220 0.6855 0.5179 0.6855 0.8279
No log 2.775 222 0.6799 0.5179 0.6799 0.8245
No log 2.8 224 0.6388 0.4336 0.6388 0.7993
No log 2.825 226 0.6146 0.4081 0.6146 0.7840
No log 2.85 228 0.5858 0.4081 0.5858 0.7654
No log 2.875 230 0.6787 0.5179 0.6787 0.8238
No log 2.9 232 0.8407 0.4444 0.8407 0.9169
No log 2.925 234 0.8648 0.4124 0.8648 0.9299
No log 2.95 236 0.7801 0.5249 0.7801 0.8832
No log 2.975 238 0.6166 0.5231 0.6166 0.7853
No log 3.0 240 0.5320 0.5308 0.5320 0.7294
No log 3.025 242 0.4767 0.4763 0.4767 0.6904
No log 3.05 244 0.4743 0.5022 0.4743 0.6887
No log 3.075 246 0.5496 0.5098 0.5496 0.7414
No log 3.1 248 0.7737 0.3707 0.7737 0.8796
No log 3.125 250 0.9850 0.3791 0.9850 0.9925
No log 3.15 252 1.1305 0.2691 1.1305 1.0632
No log 3.175 254 1.1364 0.2997 1.1364 1.0660
No log 3.2 256 0.9338 0.3688 0.9338 0.9663
No log 3.225 258 0.6607 0.5373 0.6607 0.8129
No log 3.25 260 0.5597 0.5528 0.5597 0.7481
No log 3.275 262 0.5402 0.5256 0.5402 0.7350
No log 3.3 264 0.5115 0.5304 0.5115 0.7152
No log 3.325 266 0.5038 0.5304 0.5038 0.7098
No log 3.35 268 0.5941 0.5568 0.5941 0.7708
No log 3.375 270 0.8129 0.4731 0.8129 0.9016
No log 3.4 272 1.1308 0.3278 1.1308 1.0634
No log 3.425 274 1.3353 0.2809 1.3353 1.1555
No log 3.45 276 1.2392 0.3127 1.2392 1.1132
No log 3.475 278 0.8992 0.4132 0.8992 0.9482
No log 3.5 280 0.6016 0.4933 0.6016 0.7756
No log 3.525 282 0.5712 0.5324 0.5712 0.7558
No log 3.55 284 0.6167 0.5056 0.6167 0.7853
No log 3.575 286 0.6798 0.4700 0.6798 0.8245
No log 3.6 288 0.7821 0.5029 0.7821 0.8844
No log 3.625 290 0.8194 0.4558 0.8194 0.9052
No log 3.65 292 0.8152 0.4803 0.8152 0.9029
No log 3.675 294 0.7740 0.5077 0.7740 0.8798
No log 3.7 296 0.7691 0.5077 0.7691 0.8770
No log 3.725 298 0.7883 0.5333 0.7883 0.8879
No log 3.75 300 0.7167 0.5152 0.7167 0.8466
No log 3.775 302 0.6703 0.5073 0.6703 0.8187
No log 3.8 304 0.6560 0.5293 0.6560 0.8099
No log 3.825 306 0.8198 0.5426 0.8198 0.9055
No log 3.85 308 1.0490 0.4532 1.0490 1.0242
No log 3.875 310 1.1220 0.4537 1.1220 1.0592
No log 3.9 312 0.8682 0.5512 0.8682 0.9317
No log 3.925 314 0.6110 0.4772 0.6110 0.7817
No log 3.95 316 0.5642 0.5190 0.5642 0.7512
No log 3.975 318 0.5643 0.5190 0.5643 0.7512
No log 4.0 320 0.6192 0.5133 0.6192 0.7869
No log 4.025 322 0.8950 0.4851 0.8950 0.9460
No log 4.05 324 1.2730 0.2662 1.2730 1.1283
No log 4.075 326 1.5920 0.2834 1.5920 1.2617
No log 4.1 328 1.5260 0.2286 1.5260 1.2353
No log 4.125 330 1.2383 0.2961 1.2383 1.1128
No log 4.15 332 1.0230 0.3939 1.0230 1.0114
No log 4.175 334 0.8912 0.4622 0.8912 0.9440
No log 4.2 336 0.7674 0.4447 0.7674 0.8760
No log 4.225 338 0.7055 0.4717 0.7055 0.8400
No log 4.25 340 0.7576 0.3884 0.7576 0.8704
No log 4.275 342 0.9273 0.3867 0.9273 0.9630
No log 4.3 344 1.1331 0.3601 1.1331 1.0645
No log 4.325 346 1.1666 0.3295 1.1666 1.0801
No log 4.35 348 1.0249 0.3269 1.0249 1.0124
No log 4.375 350 0.8364 0.3439 0.8364 0.9146
No log 4.4 352 0.6571 0.4197 0.6571 0.8106
No log 4.425 354 0.6060 0.4622 0.6060 0.7785
No log 4.45 356 0.6233 0.4845 0.6233 0.7895
No log 4.475 358 0.7357 0.4610 0.7357 0.8577
No log 4.5 360 0.8514 0.4126 0.8514 0.9227
No log 4.525 362 0.8937 0.3848 0.8937 0.9454
No log 4.55 364 0.8447 0.3094 0.8447 0.9191
No log 4.575 366 0.7890 0.3152 0.7890 0.8883
No log 4.6 368 0.7738 0.3377 0.7738 0.8796
No log 4.625 370 0.8136 0.3377 0.8136 0.9020
No log 4.65 372 0.7734 0.3747 0.7734 0.8795
No log 4.675 374 0.7740 0.4286 0.7740 0.8797
No log 4.7 376 0.7298 0.4286 0.7298 0.8543
No log 4.725 378 0.6342 0.4550 0.6342 0.7964
No log 4.75 380 0.6438 0.5059 0.6438 0.8024
No log 4.775 382 0.6759 0.4701 0.6759 0.8221
No log 4.8 384 0.7573 0.4504 0.7573 0.8702
No log 4.825 386 0.9304 0.3781 0.9304 0.9646
No log 4.85 388 1.1002 0.3781 1.1002 1.0489
No log 4.875 390 1.2461 0.3193 1.2461 1.1163
No log 4.9 392 1.3226 0.3418 1.3226 1.1500
No log 4.925 394 1.2215 0.3367 1.2215 1.1052
No log 4.95 396 1.1275 0.3506 1.1275 1.0619
No log 4.975 398 1.0550 0.3689 1.0550 1.0271
No log 5.0 400 0.9438 0.3977 0.9438 0.9715
No log 5.025 402 0.7925 0.4081 0.7925 0.8903
No log 5.05 404 0.7395 0.4418 0.7395 0.8599
No log 5.075 406 0.8245 0.4246 0.8245 0.9080
No log 5.1 408 0.8870 0.4305 0.8870 0.9418
No log 5.125 410 0.8834 0.4305 0.8834 0.9399
No log 5.15 412 0.8054 0.3747 0.8054 0.8974
No log 5.175 414 0.7824 0.3806 0.7824 0.8846
No log 5.2 416 0.8554 0.3946 0.8554 0.9249
No log 5.225 418 1.0415 0.4172 1.0415 1.0206
No log 5.25 420 1.1283 0.3412 1.1283 1.0622
No log 5.275 422 1.0753 0.3358 1.0753 1.0370
No log 5.3 424 0.9016 0.3849 0.9016 0.9495
No log 5.325 426 0.8101 0.3868 0.8101 0.9000
No log 5.35 428 0.7085 0.3992 0.7085 0.8417
No log 5.375 430 0.7276 0.3992 0.7276 0.8530
No log 5.4 432 0.7825 0.4304 0.7825 0.8846
No log 5.425 434 0.9852 0.4465 0.9852 0.9926
No log 5.45 436 1.0442 0.4262 1.0442 1.0218
No log 5.475 438 0.9391 0.4639 0.9391 0.9690
No log 5.5 440 0.7317 0.4952 0.7317 0.8554
No log 5.525 442 0.6522 0.5342 0.6522 0.8076
No log 5.55 444 0.6775 0.4759 0.6775 0.8231
No log 5.575 446 0.7479 0.4133 0.7479 0.8648
No log 5.6 448 0.8781 0.3977 0.8781 0.9371
No log 5.625 450 1.0145 0.3977 1.0145 1.0072
No log 5.65 452 1.0677 0.3689 1.0677 1.0333
No log 5.675 454 1.0928 0.3358 1.0928 1.0454
No log 5.7 456 1.1244 0.3681 1.1244 1.0604
No log 5.725 458 0.9854 0.3689 0.9854 0.9926
No log 5.75 460 0.7972 0.4481 0.7972 0.8928
No log 5.775 462 0.7016 0.4545 0.7016 0.8376
No log 5.8 464 0.6506 0.4772 0.6506 0.8066
No log 5.825 466 0.6095 0.5357 0.6095 0.7807
No log 5.85 468 0.6528 0.4556 0.6528 0.8079
No log 5.875 470 0.7594 0.4142 0.7594 0.8715
No log 5.9 472 0.9320 0.3697 0.9320 0.9654
No log 5.925 474 1.0911 0.3587 1.0911 1.0445
No log 5.95 476 1.1227 0.3587 1.1227 1.0596
No log 5.975 478 1.0116 0.3587 1.0116 1.0058
No log 6.0 480 0.8418 0.3807 0.8418 0.9175
No log 6.025 482 0.7030 0.4887 0.7030 0.8385
No log 6.05 484 0.6447 0.4568 0.6447 0.8029
No log 6.075 486 0.7015 0.4614 0.7015 0.8376
No log 6.1 488 0.8362 0.4481 0.8362 0.9144
No log 6.125 490 0.9549 0.4363 0.9549 0.9772
No log 6.15 492 0.9882 0.3987 0.9882 0.9941
No log 6.175 494 0.8973 0.4421 0.8973 0.9473
No log 6.2 496 0.8591 0.4050 0.8591 0.9269
No log 6.225 498 0.7846 0.4444 0.7846 0.8858
0.3935 6.25 500 0.8741 0.4003 0.8741 0.9349
0.3935 6.275 502 0.9339 0.3945 0.9339 0.9664
0.3935 6.3 504 1.0910 0.2943 1.0910 1.0445
0.3935 6.325 506 1.1445 0.2754 1.1445 1.0698
0.3935 6.35 508 0.9742 0.3505 0.9742 0.9870
0.3935 6.375 510 0.8585 0.4246 0.8585 0.9265
0.3935 6.4 512 0.7583 0.4243 0.7583 0.8708
0.3935 6.425 514 0.7164 0.4351 0.7164 0.8464
0.3935 6.45 516 0.7292 0.3439 0.7292 0.8540
0.3935 6.475 518 0.7681 0.3381 0.7681 0.8764
0.3935 6.5 520 0.8212 0.3247 0.8212 0.9062
0.3935 6.525 522 0.8902 0.3484 0.8902 0.9435
0.3935 6.55 524 0.9864 0.3486 0.9864 0.9932

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task7_organization

Finetuned
(4019)
this model