ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7606
  • Qwk: 0.4556
  • Mse: 0.7606
  • Rmse: 0.8721

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 2.5165 -0.0449 2.5165 1.5864
No log 0.08 4 1.5184 0.1230 1.5184 1.2322
No log 0.12 6 0.7524 0.1754 0.7524 0.8674
No log 0.16 8 0.7832 0.1359 0.7832 0.8850
No log 0.2 10 0.8425 0.1225 0.8425 0.9179
No log 0.24 12 0.9179 0.3228 0.9179 0.9581
No log 0.28 14 0.8057 0.1183 0.8057 0.8976
No log 0.32 16 0.7644 0.0428 0.7644 0.8743
No log 0.36 18 0.7438 0.0798 0.7438 0.8624
No log 0.4 20 0.8863 0.2510 0.8863 0.9414
No log 0.44 22 0.9742 0.3118 0.9742 0.9870
No log 0.48 24 0.8669 0.3169 0.8669 0.9311
No log 0.52 26 0.7308 0.3084 0.7308 0.8549
No log 0.56 28 0.6695 0.0 0.6695 0.8182
No log 0.6 30 0.6630 0.0481 0.6630 0.8142
No log 0.64 32 0.5998 0.2963 0.5998 0.7745
No log 0.68 34 0.5574 0.4148 0.5574 0.7466
No log 0.72 36 0.6585 0.4167 0.6585 0.8115
No log 0.76 38 0.6337 0.3867 0.6337 0.7960
No log 0.8 40 0.5503 0.4934 0.5503 0.7418
No log 0.84 42 0.5516 0.5003 0.5516 0.7427
No log 0.88 44 0.5214 0.4829 0.5214 0.7221
No log 0.92 46 0.6051 0.4315 0.6051 0.7779
No log 0.96 48 0.5769 0.4474 0.5769 0.7595
No log 1.0 50 0.5083 0.4857 0.5083 0.7130
No log 1.04 52 0.5448 0.5614 0.5448 0.7381
No log 1.08 54 0.6666 0.4218 0.6666 0.8164
No log 1.12 56 0.6387 0.4703 0.6387 0.7992
No log 1.16 58 0.5049 0.5476 0.5049 0.7106
No log 1.2 60 0.7212 0.4977 0.7212 0.8492
No log 1.24 62 0.7701 0.4479 0.7701 0.8776
No log 1.28 64 0.5566 0.5687 0.5566 0.7461
No log 1.32 66 0.4988 0.5884 0.4988 0.7062
No log 1.3600 68 0.4829 0.5583 0.4829 0.6949
No log 1.4 70 0.5672 0.4825 0.5672 0.7531
No log 1.44 72 0.6863 0.4092 0.6863 0.8284
No log 1.48 74 0.5752 0.4860 0.5752 0.7584
No log 1.52 76 0.5037 0.5649 0.5037 0.7097
No log 1.56 78 0.4905 0.6439 0.4905 0.7003
No log 1.6 80 0.5324 0.5642 0.5324 0.7296
No log 1.6400 82 0.6295 0.4652 0.6295 0.7934
No log 1.6800 84 0.5478 0.6156 0.5478 0.7402
No log 1.72 86 0.5384 0.5826 0.5384 0.7338
No log 1.76 88 0.5346 0.6001 0.5346 0.7311
No log 1.8 90 0.5371 0.6060 0.5371 0.7329
No log 1.8400 92 0.5561 0.5248 0.5561 0.7457
No log 1.88 94 0.5474 0.5926 0.5474 0.7399
No log 1.92 96 0.5652 0.5463 0.5652 0.7518
No log 1.96 98 0.5947 0.5495 0.5947 0.7712
No log 2.0 100 0.5407 0.6156 0.5407 0.7353
No log 2.04 102 0.6541 0.4504 0.6541 0.8087
No log 2.08 104 0.5544 0.5306 0.5544 0.7446
No log 2.12 106 0.5682 0.5756 0.5682 0.7538
No log 2.16 108 0.6611 0.5096 0.6611 0.8131
No log 2.2 110 0.5702 0.5668 0.5702 0.7551
No log 2.24 112 0.5817 0.5200 0.5817 0.7627
No log 2.2800 114 0.7175 0.4828 0.7175 0.8470
No log 2.32 116 0.6118 0.5513 0.6118 0.7822
No log 2.36 118 0.5114 0.5089 0.5114 0.7151
No log 2.4 120 0.5748 0.5275 0.5748 0.7581
No log 2.44 122 0.6089 0.5190 0.6089 0.7803
No log 2.48 124 0.4862 0.5437 0.4862 0.6973
No log 2.52 126 0.5149 0.4997 0.5149 0.7176
No log 2.56 128 0.5073 0.5212 0.5073 0.7122
No log 2.6 130 0.4586 0.5995 0.4586 0.6772
No log 2.64 132 0.5935 0.4961 0.5935 0.7704
No log 2.68 134 0.5470 0.5086 0.5470 0.7396
No log 2.7200 136 0.4711 0.5232 0.4711 0.6864
No log 2.76 138 0.4880 0.5356 0.4880 0.6986
No log 2.8 140 0.4723 0.5357 0.4723 0.6873
No log 2.84 142 0.4704 0.5768 0.4704 0.6859
No log 2.88 144 0.5150 0.4997 0.5150 0.7176
No log 2.92 146 0.5141 0.4925 0.5141 0.7170
No log 2.96 148 0.4970 0.5852 0.4970 0.7050
No log 3.0 150 0.4980 0.5289 0.4980 0.7057
No log 3.04 152 0.5268 0.5322 0.5268 0.7258
No log 3.08 154 0.6290 0.5259 0.6290 0.7931
No log 3.12 156 0.6849 0.4966 0.6849 0.8276
No log 3.16 158 0.5606 0.5150 0.5606 0.7487
No log 3.2 160 0.5995 0.5657 0.5995 0.7742
No log 3.24 162 0.6190 0.4359 0.6190 0.7868
No log 3.2800 164 0.5770 0.5612 0.5770 0.7596
No log 3.32 166 0.8111 0.4757 0.8111 0.9006
No log 3.36 168 0.8707 0.4803 0.8707 0.9331
No log 3.4 170 0.7527 0.4338 0.7527 0.8676
No log 3.44 172 0.6076 0.5076 0.6076 0.7795
No log 3.48 174 0.5521 0.4820 0.5521 0.7430
No log 3.52 176 0.5526 0.4262 0.5526 0.7434
No log 3.56 178 0.5526 0.4124 0.5526 0.7433
No log 3.6 180 0.6186 0.4925 0.6186 0.7865
No log 3.64 182 0.7768 0.3665 0.7768 0.8814
No log 3.68 184 0.8336 0.3868 0.8336 0.9130
No log 3.7200 186 0.6541 0.4749 0.6541 0.8088
No log 3.76 188 0.5183 0.4171 0.5183 0.7200
No log 3.8 190 0.5937 0.5015 0.5937 0.7705
No log 3.84 192 0.5623 0.4569 0.5623 0.7498
No log 3.88 194 0.5582 0.4945 0.5582 0.7471
No log 3.92 196 0.6321 0.4089 0.6321 0.7951
No log 3.96 198 0.6256 0.4491 0.6256 0.7909
No log 4.0 200 0.5602 0.3398 0.5602 0.7485
No log 4.04 202 0.5788 0.3352 0.5788 0.7608
No log 4.08 204 0.5786 0.3352 0.5786 0.7606
No log 4.12 206 0.5859 0.3788 0.5859 0.7654
No log 4.16 208 0.7061 0.4424 0.7061 0.8403
No log 4.2 210 0.7806 0.4255 0.7806 0.8835
No log 4.24 212 0.6844 0.4329 0.6844 0.8273
No log 4.28 214 0.6518 0.4329 0.6518 0.8073
No log 4.32 216 0.5730 0.4375 0.5730 0.7570
No log 4.36 218 0.5702 0.3689 0.5702 0.7551
No log 4.4 220 0.6004 0.4073 0.6004 0.7749
No log 4.44 222 0.6355 0.3775 0.6355 0.7972
No log 4.48 224 0.6649 0.4263 0.6649 0.8154
No log 4.52 226 0.6911 0.4369 0.6911 0.8313
No log 4.5600 228 0.6645 0.4444 0.6645 0.8152
No log 4.6 230 0.6504 0.4051 0.6504 0.8065
No log 4.64 232 0.6078 0.4867 0.6078 0.7796
No log 4.68 234 0.5770 0.3380 0.5770 0.7596
No log 4.72 236 0.5707 0.3561 0.5707 0.7554
No log 4.76 238 0.5414 0.4160 0.5414 0.7358
No log 4.8 240 0.5516 0.5256 0.5516 0.7427
No log 4.84 242 0.6040 0.5185 0.6040 0.7772
No log 4.88 244 0.6018 0.4937 0.6018 0.7757
No log 4.92 246 0.5980 0.4227 0.5980 0.7733
No log 4.96 248 0.5763 0.3887 0.5763 0.7591
No log 5.0 250 0.5642 0.3380 0.5642 0.7512
No log 5.04 252 0.5579 0.3808 0.5579 0.7469
No log 5.08 254 0.5748 0.4397 0.5748 0.7582
No log 5.12 256 0.6504 0.4272 0.6504 0.8065
No log 5.16 258 0.6329 0.4272 0.6329 0.7956
No log 5.2 260 0.5733 0.4855 0.5733 0.7572
No log 5.24 262 0.5292 0.5022 0.5292 0.7274
No log 5.28 264 0.5246 0.4402 0.5246 0.7243
No log 5.32 266 0.5263 0.4938 0.5263 0.7255
No log 5.36 268 0.5216 0.4299 0.5216 0.7222
No log 5.4 270 0.5822 0.4663 0.5822 0.7630
No log 5.44 272 0.6700 0.4123 0.6700 0.8185
No log 5.48 274 0.6340 0.4089 0.6340 0.7963
No log 5.52 276 0.5347 0.5516 0.5347 0.7312
No log 5.5600 278 0.5055 0.5321 0.5055 0.7110
No log 5.6 280 0.5017 0.5321 0.5017 0.7083
No log 5.64 282 0.5200 0.5388 0.5200 0.7211
No log 5.68 284 0.5832 0.4625 0.5832 0.7637
No log 5.72 286 0.5823 0.4997 0.5823 0.7631
No log 5.76 288 0.5647 0.4929 0.5647 0.7514
No log 5.8 290 0.5114 0.5214 0.5114 0.7151
No log 5.84 292 0.5271 0.4705 0.5271 0.7260
No log 5.88 294 0.5263 0.4705 0.5263 0.7255
No log 5.92 296 0.5860 0.3918 0.5860 0.7655
No log 5.96 298 0.6407 0.4409 0.6407 0.8004
No log 6.0 300 0.5985 0.3843 0.5985 0.7736
No log 6.04 302 0.5332 0.4985 0.5332 0.7302
No log 6.08 304 0.5168 0.5177 0.5168 0.7189
No log 6.12 306 0.5372 0.5218 0.5372 0.7329
No log 6.16 308 0.5861 0.4330 0.5861 0.7656
No log 6.2 310 0.6327 0.3918 0.6327 0.7954
No log 6.24 312 0.5783 0.4414 0.5783 0.7604
No log 6.28 314 0.5251 0.4752 0.5251 0.7246
No log 6.32 316 0.5159 0.5010 0.5159 0.7183
No log 6.36 318 0.5101 0.4423 0.5101 0.7142
No log 6.4 320 0.5934 0.4239 0.5934 0.7703
No log 6.44 322 0.7650 0.4704 0.7650 0.8747
No log 6.48 324 0.8029 0.4364 0.8029 0.8960
No log 6.52 326 0.6890 0.4568 0.6890 0.8301
No log 6.5600 328 0.5816 0.3867 0.5816 0.7626
No log 6.6 330 0.5596 0.4212 0.5596 0.7481
No log 6.64 332 0.5473 0.4234 0.5473 0.7398
No log 6.68 334 0.5789 0.4081 0.5789 0.7609
No log 6.72 336 0.6166 0.4764 0.6166 0.7852
No log 6.76 338 0.6567 0.4726 0.6567 0.8103
No log 6.8 340 0.7090 0.4568 0.7090 0.8420
No log 6.84 342 0.6467 0.4582 0.6467 0.8042
No log 6.88 344 0.5784 0.3814 0.5784 0.7605
No log 6.92 346 0.5426 0.4504 0.5426 0.7366
No log 6.96 348 0.5460 0.4493 0.5460 0.7389
No log 7.0 350 0.5628 0.4234 0.5628 0.7502
No log 7.04 352 0.6041 0.4602 0.6041 0.7772
No log 7.08 354 0.6119 0.4522 0.6119 0.7822
No log 7.12 356 0.5566 0.4523 0.5566 0.7461
No log 7.16 358 0.5299 0.4657 0.5299 0.7279
No log 7.2 360 0.5404 0.4858 0.5404 0.7351
No log 7.24 362 0.5279 0.4839 0.5279 0.7266
No log 7.28 364 0.5651 0.4582 0.5651 0.7517
No log 7.32 366 0.6565 0.4966 0.6565 0.8103
No log 7.36 368 0.6824 0.4877 0.6824 0.8261
No log 7.4 370 0.6386 0.4602 0.6386 0.7991
No log 7.44 372 0.5542 0.4555 0.5542 0.7444
No log 7.48 374 0.4999 0.5286 0.4999 0.7071
No log 7.52 376 0.4891 0.5304 0.4891 0.6993
No log 7.5600 378 0.4883 0.5340 0.4883 0.6988
No log 7.6 380 0.4873 0.5252 0.4873 0.6981
No log 7.64 382 0.4778 0.5734 0.4778 0.6912
No log 7.68 384 0.4746 0.5734 0.4746 0.6889
No log 7.72 386 0.4766 0.5815 0.4766 0.6903
No log 7.76 388 0.5106 0.4555 0.5106 0.7146
No log 7.8 390 0.5060 0.4964 0.5060 0.7114
No log 7.84 392 0.4977 0.5386 0.4977 0.7055
No log 7.88 394 0.4651 0.5846 0.4651 0.6820
No log 7.92 396 0.4620 0.5956 0.4620 0.6797
No log 7.96 398 0.4708 0.5160 0.4708 0.6861
No log 8.0 400 0.4612 0.5941 0.4612 0.6791
No log 8.04 402 0.4948 0.5550 0.4948 0.7035
No log 8.08 404 0.5469 0.4270 0.5469 0.7395
No log 8.12 406 0.5541 0.4076 0.5541 0.7444
No log 8.16 408 0.5609 0.4076 0.5609 0.7489
No log 8.2 410 0.5491 0.4076 0.5491 0.7410
No log 8.24 412 0.5185 0.5248 0.5185 0.7200
No log 8.28 414 0.5133 0.5600 0.5133 0.7164
No log 8.32 416 0.5248 0.5413 0.5248 0.7245
No log 8.36 418 0.5507 0.5071 0.5507 0.7421
No log 8.4 420 0.6037 0.4270 0.6037 0.7770
No log 8.44 422 0.6415 0.3996 0.6415 0.8009
No log 8.48 424 0.6432 0.4076 0.6432 0.8020
No log 8.52 426 0.6149 0.3545 0.6149 0.7842
No log 8.56 428 0.5932 0.3840 0.5932 0.7702
No log 8.6 430 0.5499 0.5022 0.5499 0.7416
No log 8.64 432 0.5434 0.4904 0.5434 0.7371
No log 8.68 434 0.5386 0.4738 0.5386 0.7339
No log 8.72 436 0.5486 0.5003 0.5486 0.7407
No log 8.76 438 0.5603 0.5133 0.5603 0.7485
No log 8.8 440 0.5278 0.5084 0.5278 0.7265
No log 8.84 442 0.5340 0.5234 0.5340 0.7307
No log 8.88 444 0.5330 0.5168 0.5330 0.7301
No log 8.92 446 0.5097 0.4955 0.5097 0.7139
No log 8.96 448 0.5090 0.4738 0.5090 0.7134
No log 9.0 450 0.5078 0.4839 0.5078 0.7126
No log 9.04 452 0.5067 0.5231 0.5067 0.7118
No log 9.08 454 0.5031 0.4801 0.5031 0.7093
No log 9.12 456 0.5304 0.4576 0.5304 0.7283
No log 9.16 458 0.5347 0.4555 0.5347 0.7312
No log 9.2 460 0.5559 0.4618 0.5559 0.7456
No log 9.24 462 0.5774 0.4911 0.5774 0.7599
No log 9.28 464 0.5679 0.4451 0.5679 0.7536
No log 9.32 466 0.5295 0.5195 0.5295 0.7277
No log 9.36 468 0.5199 0.5131 0.5199 0.7210
No log 9.4 470 0.5180 0.5131 0.5180 0.7197
No log 9.44 472 0.4934 0.5056 0.4934 0.7024
No log 9.48 474 0.4677 0.5567 0.4677 0.6838
No log 9.52 476 0.4602 0.5941 0.4602 0.6784
No log 9.56 478 0.4542 0.6377 0.4542 0.6740
No log 9.6 480 0.4654 0.5826 0.4654 0.6822
No log 9.64 482 0.4630 0.6491 0.4630 0.6805
No log 9.68 484 0.4720 0.5084 0.4720 0.6870
No log 9.72 486 0.5886 0.5362 0.5886 0.7672
No log 9.76 488 0.6302 0.5045 0.6302 0.7938
No log 9.8 490 0.6034 0.4749 0.6034 0.7768
No log 9.84 492 0.6203 0.4873 0.6203 0.7876
No log 9.88 494 0.6268 0.4873 0.6268 0.7917
No log 9.92 496 0.5531 0.4393 0.5531 0.7437
No log 9.96 498 0.5042 0.4576 0.5042 0.7101
0.3112 10.0 500 0.4965 0.5505 0.4965 0.7046
0.3112 10.04 502 0.5041 0.5505 0.5041 0.7100
0.3112 10.08 504 0.5207 0.4397 0.5207 0.7216
0.3112 10.12 506 0.5871 0.4190 0.5871 0.7662
0.3112 10.16 508 0.7323 0.4844 0.7323 0.8557
0.3112 10.2 510 0.7606 0.4556 0.7606 0.8721

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

Finetuned
(4019)
this model