ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9363
  • Qwk: 0.4857
  • Mse: 0.9363
  • Rmse: 0.9676

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0588 2 4.8073 -0.0075 4.8073 2.1925
No log 0.1176 4 2.5545 -0.0594 2.5545 1.5983
No log 0.1765 6 1.5434 0.0504 1.5434 1.2423
No log 0.2353 8 1.2968 0.0635 1.2968 1.1388
No log 0.2941 10 1.2389 0.1483 1.2389 1.1130
No log 0.3529 12 1.1839 0.1482 1.1839 1.0880
No log 0.4118 14 1.1813 0.1585 1.1813 1.0869
No log 0.4706 16 1.1448 0.1649 1.1448 1.0700
No log 0.5294 18 1.1924 0.1772 1.1924 1.0920
No log 0.5882 20 1.2469 0.0714 1.2469 1.1166
No log 0.6471 22 1.2969 0.0625 1.2969 1.1388
No log 0.7059 24 1.1896 0.0456 1.1896 1.0907
No log 0.7647 26 1.1012 0.1856 1.1012 1.0494
No log 0.8235 28 1.0987 0.1845 1.0987 1.0482
No log 0.8824 30 1.0988 0.1895 1.0988 1.0482
No log 0.9412 32 1.0035 0.1646 1.0035 1.0017
No log 1.0 34 1.0315 0.3086 1.0315 1.0157
No log 1.0588 36 1.1165 0.3165 1.1165 1.0566
No log 1.1176 38 1.2016 0.3220 1.2016 1.0962
No log 1.1765 40 1.0641 0.4347 1.0641 1.0316
No log 1.2353 42 0.9868 0.4056 0.9868 0.9934
No log 1.2941 44 1.0230 0.3720 1.0230 1.0114
No log 1.3529 46 1.1405 0.3689 1.1405 1.0679
No log 1.4118 48 1.4039 0.2636 1.4039 1.1849
No log 1.4706 50 1.7321 0.2780 1.7321 1.3161
No log 1.5294 52 1.4118 0.3407 1.4118 1.1882
No log 1.5882 54 1.0776 0.5147 1.0776 1.0381
No log 1.6471 56 1.3043 0.4312 1.3043 1.1420
No log 1.7059 58 1.3774 0.3744 1.3774 1.1736
No log 1.7647 60 1.1497 0.4806 1.1497 1.0723
No log 1.8235 62 0.9181 0.4200 0.9181 0.9582
No log 1.8824 64 0.9433 0.4867 0.9433 0.9712
No log 1.9412 66 1.0860 0.3833 1.0860 1.0421
No log 2.0 68 1.0100 0.4861 1.0100 1.0050
No log 2.0588 70 0.9800 0.4771 0.9800 0.9899
No log 2.1176 72 0.9562 0.5448 0.9562 0.9779
No log 2.1765 74 0.9972 0.5253 0.9972 0.9986
No log 2.2353 76 1.0175 0.5024 1.0175 1.0087
No log 2.2941 78 0.9892 0.4959 0.9892 0.9946
No log 2.3529 80 0.9601 0.5399 0.9601 0.9799
No log 2.4118 82 0.9414 0.5377 0.9414 0.9702
No log 2.4706 84 0.9290 0.5769 0.9290 0.9639
No log 2.5294 86 0.8964 0.5221 0.8964 0.9468
No log 2.5882 88 0.8416 0.5343 0.8416 0.9174
No log 2.6471 90 0.7869 0.5658 0.7869 0.8871
No log 2.7059 92 0.7773 0.5229 0.7773 0.8817
No log 2.7647 94 0.7829 0.4654 0.7829 0.8848
No log 2.8235 96 0.8221 0.5250 0.8221 0.9067
No log 2.8824 98 0.9051 0.5825 0.9051 0.9514
No log 2.9412 100 0.8431 0.6212 0.8431 0.9182
No log 3.0 102 0.8162 0.6034 0.8162 0.9034
No log 3.0588 104 0.8209 0.6057 0.8209 0.9060
No log 3.1176 106 0.7949 0.6029 0.7949 0.8916
No log 3.1765 108 0.7654 0.6448 0.7654 0.8749
No log 3.2353 110 0.7822 0.6414 0.7822 0.8844
No log 3.2941 112 0.8589 0.6047 0.8589 0.9268
No log 3.3529 114 1.0838 0.5752 1.0838 1.0411
No log 3.4118 116 1.0218 0.6173 1.0218 1.0109
No log 3.4706 118 1.0056 0.6103 1.0056 1.0028
No log 3.5294 120 1.0571 0.5318 1.0571 1.0282
No log 3.5882 122 1.0519 0.5222 1.0519 1.0256
No log 3.6471 124 1.0174 0.4918 1.0174 1.0087
No log 3.7059 126 0.9177 0.4677 0.9177 0.9579
No log 3.7647 128 0.8333 0.4956 0.8333 0.9128
No log 3.8235 130 0.9265 0.5648 0.9265 0.9625
No log 3.8824 132 1.2090 0.4780 1.2090 1.0996
No log 3.9412 134 1.2520 0.4504 1.2520 1.1189
No log 4.0 136 0.9556 0.4887 0.9556 0.9776
No log 4.0588 138 0.8191 0.5415 0.8191 0.9050
No log 4.1176 140 0.7976 0.5263 0.7976 0.8931
No log 4.1765 142 0.7563 0.5721 0.7563 0.8696
No log 4.2353 144 0.7445 0.5815 0.7445 0.8628
No log 4.2941 146 0.8453 0.5714 0.8453 0.9194
No log 4.3529 148 0.9320 0.5868 0.9320 0.9654
No log 4.4118 150 0.7972 0.5229 0.7972 0.8929
No log 4.4706 152 0.8510 0.5489 0.8510 0.9225
No log 4.5294 154 0.7697 0.5838 0.7697 0.8773
No log 4.5882 156 0.8295 0.5578 0.8295 0.9108
No log 4.6471 158 0.8703 0.5375 0.8703 0.9329
No log 4.7059 160 0.9492 0.5777 0.9492 0.9743
No log 4.7647 162 0.9810 0.5347 0.9810 0.9904
No log 4.8235 164 0.9033 0.4668 0.9033 0.9504
No log 4.8824 166 0.8938 0.4798 0.8938 0.9454
No log 4.9412 168 0.9649 0.5370 0.9649 0.9823
No log 5.0 170 0.9984 0.5002 0.9984 0.9992
No log 5.0588 172 1.1131 0.4528 1.1131 1.0550
No log 5.1176 174 1.0323 0.4936 1.0323 1.0160
No log 5.1765 176 0.8974 0.5042 0.8974 0.9473
No log 5.2353 178 0.9264 0.4960 0.9264 0.9625
No log 5.2941 180 1.0547 0.4913 1.0547 1.0270
No log 5.3529 182 1.0315 0.4760 1.0315 1.0156
No log 5.4118 184 0.9232 0.5349 0.9232 0.9608
No log 5.4706 186 1.0073 0.5217 1.0073 1.0037
No log 5.5294 188 0.9929 0.5146 0.9929 0.9965
No log 5.5882 190 1.0666 0.5108 1.0666 1.0328
No log 5.6471 192 1.2388 0.5148 1.2388 1.1130
No log 5.7059 194 1.1747 0.3970 1.1747 1.0838
No log 5.7647 196 1.0035 0.4458 1.0035 1.0018
No log 5.8235 198 0.9178 0.4946 0.9178 0.9580
No log 5.8824 200 0.8918 0.5030 0.8918 0.9444
No log 5.9412 202 0.9883 0.4511 0.9883 0.9941
No log 6.0 204 1.1818 0.3976 1.1818 1.0871
No log 6.0588 206 1.3538 0.3990 1.3538 1.1635
No log 6.1176 208 1.2778 0.4237 1.2778 1.1304
No log 6.1765 210 0.9557 0.5358 0.9557 0.9776
No log 6.2353 212 0.7748 0.6300 0.7748 0.8802
No log 6.2941 214 0.7756 0.5783 0.7756 0.8807
No log 6.3529 216 0.8790 0.5015 0.8790 0.9376
No log 6.4118 218 1.1155 0.4260 1.1155 1.0562
No log 6.4706 220 1.2222 0.4258 1.2222 1.1055
No log 6.5294 222 1.1216 0.5003 1.1216 1.0591
No log 6.5882 224 0.8959 0.5283 0.8959 0.9465
No log 6.6471 226 0.7421 0.6687 0.7421 0.8614
No log 6.7059 228 0.7455 0.6006 0.7455 0.8634
No log 6.7647 230 0.7401 0.6633 0.7401 0.8603
No log 6.8235 232 0.8204 0.5818 0.8204 0.9058
No log 6.8824 234 0.9285 0.5293 0.9285 0.9636
No log 6.9412 236 1.0132 0.5308 1.0132 1.0066
No log 7.0 238 1.0125 0.5292 1.0125 1.0062
No log 7.0588 240 0.9543 0.5414 0.9543 0.9769
No log 7.1176 242 0.8118 0.5578 0.8118 0.9010
No log 7.1765 244 0.7730 0.5165 0.7730 0.8792
No log 7.2353 246 0.8026 0.5182 0.8026 0.8959
No log 7.2941 248 0.8112 0.3920 0.8112 0.9007
No log 7.3529 250 0.8761 0.4398 0.8761 0.9360
No log 7.4118 252 0.9571 0.3095 0.9571 0.9783
No log 7.4706 254 1.0451 0.3387 1.0451 1.0223
No log 7.5294 256 1.0735 0.4352 1.0735 1.0361
No log 7.5882 258 1.0608 0.4291 1.0608 1.0300
No log 7.6471 260 0.9099 0.4726 0.9099 0.9539
No log 7.7059 262 0.8271 0.5214 0.8271 0.9095
No log 7.7647 264 0.8186 0.5382 0.8186 0.9048
No log 7.8235 266 0.8076 0.5098 0.8076 0.8987
No log 7.8824 268 0.8680 0.4902 0.8680 0.9317
No log 7.9412 270 0.9831 0.4407 0.9831 0.9915
No log 8.0 272 0.9947 0.4806 0.9947 0.9973
No log 8.0588 274 0.9782 0.4702 0.9782 0.9891
No log 8.1176 276 0.9330 0.4007 0.9330 0.9659
No log 8.1765 278 0.8718 0.4865 0.8718 0.9337
No log 8.2353 280 0.8713 0.4966 0.8713 0.9334
No log 8.2941 282 0.9194 0.4545 0.9194 0.9589
No log 8.3529 284 1.0445 0.3952 1.0445 1.0220
No log 8.4118 286 1.0593 0.3630 1.0593 1.0292
No log 8.4706 288 0.9801 0.4545 0.9801 0.9900
No log 8.5294 290 0.8859 0.4982 0.8859 0.9412
No log 8.5882 292 0.8652 0.5471 0.8652 0.9302
No log 8.6471 294 0.8860 0.4876 0.8860 0.9413
No log 8.7059 296 0.9593 0.4494 0.9593 0.9794
No log 8.7647 298 1.1084 0.4598 1.1084 1.0528
No log 8.8235 300 1.2444 0.4290 1.2444 1.1155
No log 8.8824 302 1.2129 0.4154 1.2129 1.1013
No log 8.9412 304 1.0751 0.3571 1.0751 1.0369
No log 9.0 306 0.9779 0.3519 0.9779 0.9889
No log 9.0588 308 0.9194 0.4203 0.9194 0.9589
No log 9.1176 310 0.8765 0.4482 0.8765 0.9362
No log 9.1765 312 0.8702 0.5220 0.8702 0.9329
No log 9.2353 314 1.0166 0.5238 1.0166 1.0083
No log 9.2941 316 1.2063 0.4569 1.2063 1.0983
No log 9.3529 318 1.1624 0.4758 1.1624 1.0781
No log 9.4118 320 0.9525 0.5028 0.9525 0.9760
No log 9.4706 322 0.8250 0.5250 0.8250 0.9083
No log 9.5294 324 0.8140 0.5462 0.8140 0.9022
No log 9.5882 326 0.8228 0.5462 0.8228 0.9071
No log 9.6471 328 0.8498 0.5858 0.8498 0.9219
No log 9.7059 330 0.8486 0.6009 0.8486 0.9212
No log 9.7647 332 0.8452 0.5548 0.8452 0.9193
No log 9.8235 334 0.8896 0.5433 0.8896 0.9432
No log 9.8824 336 0.8586 0.5602 0.8586 0.9266
No log 9.9412 338 0.8246 0.5902 0.8246 0.9080
No log 10.0 340 0.8058 0.5875 0.8058 0.8977
No log 10.0588 342 0.8457 0.5523 0.8457 0.9196
No log 10.1176 344 0.9574 0.6102 0.9574 0.9785
No log 10.1765 346 0.9846 0.6047 0.9846 0.9923
No log 10.2353 348 0.8664 0.5505 0.8664 0.9308
No log 10.2941 350 0.8362 0.5819 0.8362 0.9144
No log 10.3529 352 0.8155 0.5323 0.8155 0.9031
No log 10.4118 354 0.8155 0.5223 0.8155 0.9031
No log 10.4706 356 0.8601 0.5455 0.8601 0.9274
No log 10.5294 358 0.9218 0.5635 0.9218 0.9601
No log 10.5882 360 0.9265 0.5297 0.9265 0.9625
No log 10.6471 362 0.9111 0.4906 0.9111 0.9545
No log 10.7059 364 0.9038 0.4894 0.9038 0.9507
No log 10.7647 366 0.9135 0.5185 0.9135 0.9558
No log 10.8235 368 0.9191 0.4711 0.9191 0.9587
No log 10.8824 370 0.8989 0.4906 0.8989 0.9481
No log 10.9412 372 0.8667 0.5209 0.8667 0.9309
No log 11.0 374 0.8879 0.5227 0.8879 0.9423
No log 11.0588 376 0.9208 0.5209 0.9208 0.9596
No log 11.1176 378 0.8844 0.4454 0.8844 0.9405
No log 11.1765 380 0.8818 0.4774 0.8818 0.9390
No log 11.2353 382 0.8871 0.4998 0.8871 0.9419
No log 11.2941 384 0.8887 0.4772 0.8887 0.9427
No log 11.3529 386 0.8918 0.4763 0.8918 0.9443
No log 11.4118 388 0.8408 0.5294 0.8408 0.9170
No log 11.4706 390 0.8165 0.5251 0.8165 0.9036
No log 11.5294 392 0.8228 0.5530 0.8228 0.9071
No log 11.5882 394 0.9010 0.5558 0.9010 0.9492
No log 11.6471 396 1.0061 0.4927 1.0061 1.0030
No log 11.7059 398 0.9851 0.4921 0.9851 0.9925
No log 11.7647 400 0.9548 0.4742 0.9548 0.9771
No log 11.8235 402 0.8631 0.4820 0.8631 0.9290
No log 11.8824 404 0.8331 0.4979 0.8331 0.9127
No log 11.9412 406 0.8442 0.4611 0.8442 0.9188
No log 12.0 408 0.9193 0.4911 0.9193 0.9588
No log 12.0588 410 0.9717 0.4989 0.9717 0.9858
No log 12.1176 412 0.9289 0.5408 0.9289 0.9638
No log 12.1765 414 0.7957 0.5979 0.7957 0.8920
No log 12.2353 416 0.7618 0.6109 0.7618 0.8728
No log 12.2941 418 0.7863 0.6131 0.7863 0.8868
No log 12.3529 420 0.8430 0.5738 0.8430 0.9182
No log 12.4118 422 0.8415 0.5738 0.8415 0.9173
No log 12.4706 424 0.8278 0.5869 0.8278 0.9099
No log 12.5294 426 0.8158 0.5983 0.8158 0.9032
No log 12.5882 428 0.8148 0.5550 0.8148 0.9027
No log 12.6471 430 0.8363 0.5674 0.8363 0.9145
No log 12.7059 432 0.8917 0.5661 0.8917 0.9443
No log 12.7647 434 0.9555 0.5801 0.9555 0.9775
No log 12.8235 436 0.9174 0.5637 0.9174 0.9578
No log 12.8824 438 0.8757 0.5637 0.8757 0.9358
No log 12.9412 440 0.8956 0.5509 0.8956 0.9464
No log 13.0 442 0.9253 0.5509 0.9253 0.9619
No log 13.0588 444 1.0133 0.5002 1.0133 1.0066
No log 13.1176 446 1.0696 0.4894 1.0696 1.0342
No log 13.1765 448 1.0745 0.5002 1.0745 1.0366
No log 13.2353 450 1.0689 0.4921 1.0689 1.0339
No log 13.2941 452 1.0145 0.5509 1.0145 1.0072
No log 13.3529 454 0.9539 0.5509 0.9539 0.9767
No log 13.4118 456 0.8530 0.5359 0.8530 0.9236
No log 13.4706 458 0.8397 0.5519 0.8397 0.9163
No log 13.5294 460 0.8329 0.4832 0.8329 0.9126
No log 13.5882 462 0.8536 0.5124 0.8536 0.9239
No log 13.6471 464 0.9047 0.5575 0.9047 0.9511
No log 13.7059 466 1.0288 0.4482 1.0288 1.0143
No log 13.7647 468 1.0931 0.4604 1.0931 1.0455
No log 13.8235 470 1.0227 0.5393 1.0227 1.0113
No log 13.8824 472 0.9183 0.5352 0.9183 0.9583
No log 13.9412 474 0.8399 0.5680 0.8399 0.9165
No log 14.0 476 0.8296 0.5704 0.8296 0.9108
No log 14.0588 478 0.8459 0.5706 0.8459 0.9197
No log 14.1176 480 0.9344 0.5040 0.9344 0.9667
No log 14.1765 482 1.0940 0.4583 1.0940 1.0460
No log 14.2353 484 1.1797 0.3791 1.1797 1.0861
No log 14.2941 486 1.1145 0.4075 1.1145 1.0557
No log 14.3529 488 0.9840 0.4647 0.9840 0.9919
No log 14.4118 490 0.9040 0.4425 0.9040 0.9508
No log 14.4706 492 0.8835 0.5136 0.8835 0.9399
No log 14.5294 494 0.9373 0.4962 0.9373 0.9681
No log 14.5882 496 1.0882 0.5177 1.0882 1.0432
No log 14.6471 498 1.1370 0.5489 1.1370 1.0663
0.3316 14.7059 500 1.0293 0.5115 1.0293 1.0145
0.3316 14.7647 502 0.9039 0.4883 0.9039 0.9507
0.3316 14.8235 504 0.8290 0.4724 0.8290 0.9105
0.3316 14.8824 506 0.8272 0.4724 0.8272 0.9095
0.3316 14.9412 508 0.8598 0.4429 0.8598 0.9272
0.3316 15.0 510 0.9363 0.4857 0.9363 0.9676

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task2_organization

Finetuned
(4019)
this model