ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k5_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6931
  • Qwk: 0.4592
  • Mse: 0.6931
  • Rmse: 0.8325

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 2.5881 0.0052 2.5881 1.6088
No log 0.16 4 1.4276 0.0547 1.4276 1.1948
No log 0.24 6 0.8136 0.0509 0.8136 0.9020
No log 0.32 8 0.9098 0.0715 0.9098 0.9538
No log 0.4 10 0.9846 0.3066 0.9846 0.9923
No log 0.48 12 0.7534 0.3538 0.7534 0.8680
No log 0.56 14 0.6428 0.3166 0.6428 0.8018
No log 0.64 16 0.7146 0.3991 0.7146 0.8453
No log 0.72 18 0.8478 0.3516 0.8478 0.9208
No log 0.8 20 0.7647 0.3938 0.7647 0.8745
No log 0.88 22 0.6610 0.4624 0.6610 0.8130
No log 0.96 24 0.5752 0.4124 0.5752 0.7584
No log 1.04 26 0.5608 0.4378 0.5608 0.7489
No log 1.12 28 0.7491 0.4646 0.7491 0.8655
No log 1.2 30 0.9540 0.2756 0.9540 0.9768
No log 1.28 32 0.8396 0.3657 0.8396 0.9163
No log 1.3600 34 0.5506 0.4124 0.5506 0.7420
No log 1.44 36 0.5588 0.4475 0.5588 0.7475
No log 1.52 38 0.5685 0.4514 0.5685 0.7540
No log 1.6 40 0.6887 0.4808 0.6887 0.8299
No log 1.6800 42 0.9237 0.4003 0.9237 0.9611
No log 1.76 44 1.0499 0.3689 1.0499 1.0246
No log 1.8400 46 0.8730 0.4032 0.8730 0.9343
No log 1.92 48 0.6103 0.4788 0.6103 0.7812
No log 2.0 50 0.5954 0.4438 0.5954 0.7717
No log 2.08 52 0.5527 0.4257 0.5527 0.7434
No log 2.16 54 0.6231 0.4482 0.6231 0.7894
No log 2.24 56 0.6802 0.4574 0.6802 0.8247
No log 2.32 58 0.5971 0.4745 0.5971 0.7727
No log 2.4 60 0.5527 0.5075 0.5527 0.7434
No log 2.48 62 0.5780 0.4813 0.5780 0.7603
No log 2.56 64 0.5634 0.4681 0.5634 0.7506
No log 2.64 66 0.6465 0.5139 0.6465 0.8040
No log 2.7200 68 0.8086 0.4511 0.8086 0.8992
No log 2.8 70 0.9270 0.3450 0.9270 0.9628
No log 2.88 72 0.9196 0.4032 0.9196 0.9589
No log 2.96 74 0.8886 0.4287 0.8886 0.9427
No log 3.04 76 0.8288 0.3645 0.8288 0.9104
No log 3.12 78 0.9177 0.4652 0.9177 0.9579
No log 3.2 80 0.9250 0.4270 0.9250 0.9617
No log 3.2800 82 0.9289 0.4124 0.9289 0.9638
No log 3.36 84 0.9107 0.4124 0.9107 0.9543
No log 3.44 86 0.8706 0.4837 0.8706 0.9330
No log 3.52 88 0.8164 0.5146 0.8164 0.9035
No log 3.6 90 0.8725 0.4455 0.8725 0.9341
No log 3.68 92 0.8850 0.4701 0.8850 0.9407
No log 3.76 94 0.8400 0.4568 0.8400 0.9165
No log 3.84 96 0.8667 0.4703 0.8667 0.9310
No log 3.92 98 0.7959 0.3503 0.7959 0.8922
No log 4.0 100 0.8302 0.3525 0.8302 0.9112
No log 4.08 102 1.0053 0.2343 1.0053 1.0026
No log 4.16 104 1.4739 0.3044 1.4739 1.2140
No log 4.24 106 1.3607 0.2523 1.3607 1.1665
No log 4.32 108 1.1740 0.2650 1.1740 1.0835
No log 4.4 110 0.8322 0.3978 0.8322 0.9123
No log 4.48 112 0.6772 0.4788 0.6772 0.8229
No log 4.5600 114 0.6430 0.4197 0.6430 0.8018
No log 4.64 116 0.7687 0.4096 0.7687 0.8767
No log 4.72 118 0.8610 0.4270 0.8610 0.9279
No log 4.8 120 0.8336 0.5181 0.8336 0.9130
No log 4.88 122 0.6882 0.5582 0.6882 0.8296
No log 4.96 124 0.6101 0.4624 0.6101 0.7811
No log 5.04 126 0.5971 0.5259 0.5971 0.7727
No log 5.12 128 0.6823 0.4795 0.6823 0.8260
No log 5.2 130 0.7661 0.4632 0.7661 0.8753
No log 5.28 132 0.9291 0.5100 0.9291 0.9639
No log 5.36 134 0.9661 0.4819 0.9661 0.9829
No log 5.44 136 0.8234 0.5229 0.8234 0.9074
No log 5.52 138 0.6461 0.4845 0.6461 0.8038
No log 5.6 140 0.5360 0.5098 0.5360 0.7321
No log 5.68 142 0.5240 0.4933 0.5240 0.7239
No log 5.76 144 0.5711 0.5107 0.5711 0.7557
No log 5.84 146 0.5940 0.4783 0.5940 0.7707
No log 5.92 148 0.5435 0.4933 0.5435 0.7372
No log 6.0 150 0.5519 0.4819 0.5519 0.7429
No log 6.08 152 0.5910 0.5093 0.5910 0.7687
No log 6.16 154 0.7270 0.4735 0.7270 0.8526
No log 6.24 156 0.8356 0.4568 0.8356 0.9141
No log 6.32 158 0.8208 0.4346 0.8208 0.9060
No log 6.4 160 0.7693 0.4208 0.7693 0.8771
No log 6.48 162 0.8319 0.4346 0.8319 0.9121
No log 6.5600 164 0.7844 0.4760 0.7844 0.8857
No log 6.64 166 0.6959 0.4853 0.6959 0.8342
No log 6.72 168 0.7506 0.5160 0.7506 0.8664
No log 6.8 170 0.7488 0.5296 0.7488 0.8653
No log 6.88 172 0.7076 0.5296 0.7076 0.8412
No log 6.96 174 0.6541 0.5720 0.6541 0.8088
No log 7.04 176 0.5915 0.4646 0.5915 0.7691
No log 7.12 178 0.5910 0.4726 0.5910 0.7688
No log 7.2 180 0.6618 0.4687 0.6618 0.8135
No log 7.28 182 0.7645 0.4545 0.7645 0.8743
No log 7.36 184 0.9115 0.3830 0.9115 0.9547
No log 7.44 186 0.8645 0.4545 0.8645 0.9298
No log 7.52 188 0.6470 0.5009 0.6470 0.8044
No log 7.6 190 0.5909 0.4272 0.5909 0.7687
No log 7.68 192 0.6430 0.4769 0.6430 0.8019
No log 7.76 194 0.7047 0.4308 0.7047 0.8395
No log 7.84 196 0.8778 0.4189 0.8778 0.9369
No log 7.92 198 1.0907 0.3723 1.0907 1.0444
No log 8.0 200 0.9898 0.3630 0.9898 0.9949
No log 8.08 202 0.8144 0.4400 0.8144 0.9024
No log 8.16 204 0.7062 0.4197 0.7062 0.8403
No log 8.24 206 0.7720 0.4328 0.7720 0.8786
No log 8.32 208 0.9503 0.4045 0.9503 0.9748
No log 8.4 210 0.8920 0.4450 0.8920 0.9445
No log 8.48 212 0.8004 0.4618 0.8004 0.8946
No log 8.56 214 0.7571 0.5146 0.7571 0.8701
No log 8.64 216 0.7189 0.4721 0.7189 0.8479
No log 8.72 218 0.6692 0.4795 0.6692 0.8180
No log 8.8 220 0.6761 0.4275 0.6761 0.8223
No log 8.88 222 0.6384 0.4218 0.6384 0.7990
No log 8.96 224 0.6764 0.4424 0.6764 0.8224
No log 9.04 226 0.8217 0.4538 0.8217 0.9065
No log 9.12 228 0.8604 0.4454 0.8604 0.9276
No log 9.2 230 0.7712 0.3918 0.7712 0.8782
No log 9.28 232 0.8133 0.4400 0.8133 0.9018
No log 9.36 234 0.7539 0.4870 0.7539 0.8683
No log 9.44 236 0.6622 0.5388 0.6622 0.8137
No log 9.52 238 0.6231 0.5758 0.6231 0.7894
No log 9.6 240 0.6206 0.5368 0.6206 0.7878
No log 9.68 242 0.6326 0.5368 0.6326 0.7954
No log 9.76 244 0.6708 0.5278 0.6708 0.8191
No log 9.84 246 0.6359 0.4741 0.6359 0.7974
No log 9.92 248 0.6040 0.4819 0.6040 0.7772
No log 10.0 250 0.5856 0.4997 0.5856 0.7653
No log 10.08 252 0.5838 0.5468 0.5838 0.7640
No log 10.16 254 0.6196 0.5140 0.6196 0.7872
No log 10.24 256 0.7477 0.4574 0.7477 0.8647
No log 10.32 258 0.9512 0.4233 0.9512 0.9753
No log 10.4 260 0.9778 0.4161 0.9778 0.9888
No log 10.48 262 0.8131 0.4704 0.8131 0.9017
No log 10.56 264 0.6497 0.4845 0.6497 0.8060
No log 10.64 266 0.5941 0.4920 0.5941 0.7708
No log 10.72 268 0.6133 0.4315 0.6133 0.7831
No log 10.8 270 0.7092 0.4769 0.7092 0.8422
No log 10.88 272 0.8899 0.3870 0.8899 0.9434
No log 10.96 274 1.0215 0.3233 1.0215 1.0107
No log 11.04 276 1.1429 0.2775 1.1429 1.0691
No log 11.12 278 1.0016 0.3486 1.0016 1.0008
No log 11.2 280 0.7471 0.4502 0.7471 0.8644
No log 11.28 282 0.6711 0.4438 0.6711 0.8192
No log 11.36 284 0.6312 0.4295 0.6312 0.7945
No log 11.44 286 0.5988 0.4663 0.5988 0.7738
No log 11.52 288 0.6003 0.4663 0.6003 0.7748
No log 11.6 290 0.6941 0.4549 0.6941 0.8331
No log 11.68 292 0.7219 0.3960 0.7219 0.8496
No log 11.76 294 0.6841 0.4624 0.6841 0.8271
No log 11.84 296 0.6587 0.4387 0.6587 0.8116
No log 11.92 298 0.6506 0.4230 0.6506 0.8066
No log 12.0 300 0.5794 0.4997 0.5794 0.7612
No log 12.08 302 0.5404 0.4919 0.5404 0.7351
No log 12.16 304 0.5570 0.4907 0.5570 0.7463
No log 12.24 306 0.6482 0.5354 0.6482 0.8051
No log 12.32 308 0.7694 0.5377 0.7694 0.8772
No log 12.4 310 0.8443 0.4936 0.8443 0.9189
No log 12.48 312 0.8177 0.4573 0.8177 0.9043
No log 12.56 314 0.8558 0.4439 0.8558 0.9251
No log 12.64 316 0.8797 0.4618 0.8797 0.9379
No log 12.72 318 0.8521 0.4258 0.8521 0.9231
No log 12.8 320 0.8324 0.4258 0.8324 0.9124
No log 12.88 322 0.8308 0.4013 0.8308 0.9115
No log 12.96 324 0.8369 0.3913 0.8369 0.9148
No log 13.04 326 0.7691 0.4632 0.7691 0.8770
No log 13.12 328 0.7565 0.4347 0.7565 0.8697
No log 13.2 330 0.7701 0.4760 0.7701 0.8776
No log 13.28 332 0.7160 0.4419 0.7160 0.8462
No log 13.36 334 0.6181 0.4568 0.6181 0.7862
No log 13.44 336 0.5264 0.5989 0.5264 0.7256
No log 13.52 338 0.5102 0.5989 0.5102 0.7143
No log 13.6 340 0.5319 0.5544 0.5319 0.7293
No log 13.68 342 0.6156 0.4493 0.6156 0.7846
No log 13.76 344 0.6651 0.4183 0.6651 0.8156
No log 13.84 346 0.6332 0.4030 0.6332 0.7957
No log 13.92 348 0.5853 0.4979 0.5853 0.7650
No log 14.0 350 0.5656 0.4749 0.5656 0.7520
No log 14.08 352 0.5776 0.4014 0.5776 0.7600
No log 14.16 354 0.6349 0.4404 0.6349 0.7968
No log 14.24 356 0.6822 0.4549 0.6822 0.8260
No log 14.32 358 0.6881 0.4630 0.6881 0.8295
No log 14.4 360 0.6538 0.4835 0.6538 0.8086
No log 14.48 362 0.6426 0.5138 0.6426 0.8016
No log 14.56 364 0.6273 0.4255 0.6273 0.7920
No log 14.64 366 0.6319 0.4275 0.6319 0.7949
No log 14.72 368 0.6799 0.4400 0.6799 0.8245
No log 14.8 370 0.7295 0.4346 0.7295 0.8541
No log 14.88 372 0.7058 0.4328 0.7058 0.8401
No log 14.96 374 0.6404 0.4154 0.6404 0.8002
No log 15.04 376 0.6238 0.4482 0.6238 0.7898
No log 15.12 378 0.5763 0.5310 0.5763 0.7591
No log 15.2 380 0.5616 0.5310 0.5616 0.7494
No log 15.28 382 0.5837 0.4898 0.5837 0.7640
No log 15.36 384 0.6473 0.4400 0.6473 0.8046
No log 15.44 386 0.7749 0.4511 0.7749 0.8803
No log 15.52 388 0.7812 0.4814 0.7812 0.8839
No log 15.6 390 0.6786 0.4484 0.6786 0.8238
No log 15.68 392 0.5669 0.3894 0.5669 0.7529
No log 15.76 394 0.5330 0.4569 0.5330 0.7300
No log 15.84 396 0.5550 0.3972 0.5550 0.7450
No log 15.92 398 0.6137 0.3425 0.6137 0.7834
No log 16.0 400 0.7182 0.4113 0.7182 0.8475
No log 16.08 402 0.7819 0.4044 0.7819 0.8842
No log 16.16 404 0.7449 0.4044 0.7449 0.8631
No log 16.24 406 0.7169 0.4308 0.7169 0.8467
No log 16.32 408 0.6790 0.4308 0.6790 0.8240
No log 16.4 410 0.6218 0.4328 0.6218 0.7886
No log 16.48 412 0.5580 0.4684 0.5580 0.7470
No log 16.56 414 0.5407 0.5254 0.5407 0.7353
No log 16.64 416 0.5636 0.4684 0.5636 0.7507
No log 16.72 418 0.6610 0.4630 0.6610 0.8130
No log 16.8 420 0.7533 0.4494 0.7533 0.8679
No log 16.88 422 0.8658 0.4621 0.8658 0.9305
No log 16.96 424 0.8892 0.4298 0.8892 0.9430
No log 17.04 426 0.8783 0.4214 0.8783 0.9372
No log 17.12 428 0.7901 0.3892 0.7901 0.8889
No log 17.2 430 0.7375 0.3960 0.7375 0.8588
No log 17.28 432 0.7158 0.3799 0.7158 0.8461
No log 17.36 434 0.6777 0.3799 0.6777 0.8233
No log 17.44 436 0.6594 0.3799 0.6594 0.8121
No log 17.52 438 0.6971 0.4203 0.6971 0.8349
No log 17.6 440 0.7645 0.5312 0.7645 0.8744
No log 17.68 442 0.8768 0.4926 0.8768 0.9364
No log 17.76 444 0.8783 0.4926 0.8783 0.9372
No log 17.84 446 0.8392 0.4701 0.8392 0.9161
No log 17.92 448 0.7423 0.4347 0.7423 0.8615
No log 18.0 450 0.6484 0.3822 0.6484 0.8052
No log 18.08 452 0.6177 0.4272 0.6177 0.7859
No log 18.16 454 0.6194 0.4272 0.6194 0.7870
No log 18.24 456 0.6757 0.3754 0.6757 0.8220
No log 18.32 458 0.7680 0.4347 0.7680 0.8764
No log 18.4 460 0.8638 0.4568 0.8638 0.9294
No log 18.48 462 0.9217 0.4503 0.9217 0.9600
No log 18.56 464 0.8603 0.4494 0.8603 0.9275
No log 18.64 466 0.7765 0.4556 0.7765 0.8812
No log 18.72 468 0.7050 0.3623 0.7050 0.8396
No log 18.8 470 0.6557 0.3868 0.6557 0.8098
No log 18.88 472 0.6474 0.4104 0.6474 0.8046
No log 18.96 474 0.6627 0.4104 0.6627 0.8141
No log 19.04 476 0.6613 0.4085 0.6613 0.8132
No log 19.12 478 0.6947 0.3981 0.6947 0.8335
No log 19.2 480 0.7361 0.3665 0.7361 0.8579
No log 19.28 482 0.7353 0.3665 0.7353 0.8575
No log 19.36 484 0.7447 0.3665 0.7447 0.8630
No log 19.44 486 0.7599 0.3892 0.7599 0.8717
No log 19.52 488 0.7437 0.3665 0.7437 0.8624
No log 19.6 490 0.7340 0.3665 0.7340 0.8567
No log 19.68 492 0.7829 0.4113 0.7829 0.8848
No log 19.76 494 0.8854 0.4922 0.8854 0.9409
No log 19.84 496 0.8827 0.4214 0.8827 0.9395
No log 19.92 498 0.7930 0.4427 0.7930 0.8905
0.2751 20.0 500 0.7612 0.4635 0.7612 0.8725
0.2751 20.08 502 0.7433 0.4906 0.7433 0.8621
0.2751 20.16 504 0.6994 0.5050 0.6994 0.8363
0.2751 20.24 506 0.6922 0.4721 0.6922 0.8320
0.2751 20.32 508 0.6955 0.4104 0.6955 0.8340
0.2751 20.4 510 0.6675 0.4243 0.6675 0.8170
0.2751 20.48 512 0.6760 0.4144 0.6760 0.8222
0.2751 20.56 514 0.7078 0.4521 0.7078 0.8413
0.2751 20.64 516 0.6931 0.4592 0.6931 0.8325

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k5_task7_organization

Finetuned
(4019)
this model