ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6481
  • Qwk: 0.5500
  • Mse: 0.6481
  • Rmse: 0.8050

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0312 2 2.5993 -0.0593 2.5993 1.6122
No log 0.0625 4 1.2493 0.0982 1.2493 1.1177
No log 0.0938 6 0.8911 0.0145 0.8911 0.9440
No log 0.125 8 0.7589 0.0376 0.7589 0.8711
No log 0.1562 10 0.8983 0.1808 0.8983 0.9478
No log 0.1875 12 1.0233 0.1277 1.0233 1.0116
No log 0.2188 14 0.8984 0.2308 0.8984 0.9478
No log 0.25 16 0.7303 0.2621 0.7303 0.8546
No log 0.2812 18 0.7291 0.2490 0.7291 0.8539
No log 0.3125 20 0.7961 0.2705 0.7961 0.8922
No log 0.3438 22 0.7005 0.2325 0.7005 0.8370
No log 0.375 24 0.6840 0.3127 0.6840 0.8271
No log 0.4062 26 0.7937 0.3606 0.7937 0.8909
No log 0.4375 28 0.6765 0.3425 0.6765 0.8225
No log 0.4688 30 0.8270 0.3204 0.8270 0.9094
No log 0.5 32 1.0347 0.2009 1.0347 1.0172
No log 0.5312 34 1.1963 0.1710 1.1963 1.0938
No log 0.5625 36 1.2997 0.1694 1.2997 1.1401
No log 0.5938 38 1.1482 0.2199 1.1482 1.0715
No log 0.625 40 0.9135 0.2508 0.9135 0.9558
No log 0.6562 42 0.6874 0.2361 0.6874 0.8291
No log 0.6875 44 0.6776 0.2540 0.6776 0.8232
No log 0.7188 46 0.7047 0.2095 0.7047 0.8394
No log 0.75 48 0.7640 0.1673 0.7640 0.8741
No log 0.7812 50 0.8609 0.1697 0.8609 0.9279
No log 0.8125 52 1.0069 0.2394 1.0069 1.0034
No log 0.8438 54 0.8118 0.2109 0.8118 0.9010
No log 0.875 56 0.6465 0.1604 0.6465 0.8041
No log 0.9062 58 0.7707 0.3042 0.7707 0.8779
No log 0.9375 60 0.7380 0.3675 0.7380 0.8591
No log 0.9688 62 0.6079 0.3523 0.6079 0.7797
No log 1.0 64 0.8162 0.3917 0.8162 0.9034
No log 1.0312 66 1.0775 0.3024 1.0775 1.0380
No log 1.0625 68 1.0460 0.3864 1.0460 1.0228
No log 1.0938 70 0.8708 0.4296 0.8708 0.9332
No log 1.125 72 0.7819 0.4092 0.7819 0.8842
No log 1.1562 74 0.7410 0.4289 0.7410 0.8608
No log 1.1875 76 0.7586 0.3917 0.7586 0.8710
No log 1.2188 78 0.6834 0.4051 0.6834 0.8267
No log 1.25 80 0.5966 0.4126 0.5966 0.7724
No log 1.2812 82 0.6260 0.3023 0.6260 0.7912
No log 1.3125 84 0.6298 0.3183 0.6298 0.7936
No log 1.3438 86 0.5904 0.3915 0.5904 0.7684
No log 1.375 88 0.6777 0.4610 0.6777 0.8232
No log 1.4062 90 0.7898 0.4114 0.7898 0.8887
No log 1.4375 92 0.8664 0.3766 0.8664 0.9308
No log 1.4688 94 0.8693 0.3766 0.8693 0.9324
No log 1.5 96 0.7934 0.4114 0.7934 0.8907
No log 1.5312 98 0.6839 0.5095 0.6839 0.8270
No log 1.5625 100 0.7143 0.5184 0.7143 0.8452
No log 1.5938 102 0.8292 0.2555 0.8292 0.9106
No log 1.625 104 0.7728 0.3657 0.7728 0.8791
No log 1.6562 106 0.6547 0.4719 0.6547 0.8091
No log 1.6875 108 0.6695 0.4908 0.6695 0.8182
No log 1.7188 110 0.6467 0.5189 0.6467 0.8042
No log 1.75 112 0.6436 0.4938 0.6436 0.8022
No log 1.7812 114 0.6821 0.5046 0.6821 0.8259
No log 1.8125 116 0.7281 0.3876 0.7281 0.8533
No log 1.8438 118 0.7593 0.4451 0.7593 0.8714
No log 1.875 120 0.7406 0.3661 0.7406 0.8606
No log 1.9062 122 0.6137 0.4345 0.6137 0.7834
No log 1.9375 124 0.6405 0.4330 0.6405 0.8003
No log 1.9688 126 0.6588 0.4554 0.6588 0.8117
No log 2.0 128 0.5914 0.3572 0.5914 0.7690
No log 2.0312 130 0.6341 0.4452 0.6341 0.7963
No log 2.0625 132 0.6194 0.4452 0.6194 0.7870
No log 2.0938 134 0.5759 0.4463 0.5759 0.7589
No log 2.125 136 0.6222 0.4587 0.6222 0.7888
No log 2.1562 138 0.6027 0.5134 0.6027 0.7763
No log 2.1875 140 0.7245 0.5407 0.7245 0.8512
No log 2.2188 142 1.0430 0.1727 1.0430 1.0213
No log 2.25 144 0.9410 0.2594 0.9410 0.9701
No log 2.2812 146 0.6663 0.5348 0.6663 0.8162
No log 2.3125 148 0.5948 0.4907 0.5948 0.7712
No log 2.3438 150 0.6123 0.4543 0.6123 0.7825
No log 2.375 152 0.6175 0.4036 0.6175 0.7858
No log 2.4062 154 0.6193 0.4722 0.6193 0.7870
No log 2.4375 156 0.6513 0.5195 0.6513 0.8070
No log 2.4688 158 0.6884 0.4373 0.6884 0.8297
No log 2.5 160 0.6663 0.4727 0.6663 0.8163
No log 2.5312 162 0.6231 0.4824 0.6231 0.7893
No log 2.5625 164 0.6359 0.5144 0.6359 0.7974
No log 2.5938 166 0.6206 0.5212 0.6206 0.7878
No log 2.625 168 0.6025 0.4910 0.6025 0.7762
No log 2.6562 170 0.6074 0.4975 0.6074 0.7793
No log 2.6875 172 0.6177 0.4752 0.6177 0.7859
No log 2.7188 174 0.7066 0.4545 0.7066 0.8406
No log 2.75 176 0.6944 0.4598 0.6944 0.8333
No log 2.7812 178 0.6280 0.4370 0.6280 0.7925
No log 2.8125 180 0.6117 0.4427 0.6117 0.7821
No log 2.8438 182 0.6646 0.4634 0.6646 0.8152
No log 2.875 184 0.6903 0.5251 0.6903 0.8308
No log 2.9062 186 0.6559 0.4689 0.6559 0.8099
No log 2.9375 188 0.6573 0.4997 0.6573 0.8107
No log 2.9688 190 0.6378 0.4673 0.6378 0.7986
No log 3.0 192 0.5875 0.4725 0.5875 0.7665
No log 3.0312 194 0.5556 0.5336 0.5556 0.7454
No log 3.0625 196 0.5502 0.5436 0.5502 0.7418
No log 3.0938 198 0.6311 0.5046 0.6311 0.7944
No log 3.125 200 0.6856 0.5018 0.6856 0.8280
No log 3.1562 202 0.6377 0.5190 0.6377 0.7986
No log 3.1875 204 0.5446 0.5519 0.5446 0.7380
No log 3.2188 206 0.5876 0.5425 0.5876 0.7666
No log 3.25 208 0.5602 0.5929 0.5602 0.7485
No log 3.2812 210 0.7025 0.5326 0.7025 0.8382
No log 3.3125 212 0.8812 0.3699 0.8812 0.9387
No log 3.3438 214 0.7698 0.5209 0.7698 0.8774
No log 3.375 216 0.5718 0.5052 0.5718 0.7562
No log 3.4062 218 0.5706 0.6187 0.5706 0.7554
No log 3.4375 220 0.5506 0.6084 0.5506 0.7420
No log 3.4688 222 0.5226 0.5741 0.5226 0.7229
No log 3.5 224 0.6153 0.5190 0.6153 0.7844
No log 3.5312 226 0.9597 0.3497 0.9597 0.9796
No log 3.5625 228 1.2603 0.2637 1.2603 1.1227
No log 3.5938 230 1.1957 0.2658 1.1957 1.0935
No log 3.625 232 0.8754 0.3547 0.8754 0.9356
No log 3.6562 234 0.5807 0.4834 0.5807 0.7620
No log 3.6875 236 0.5572 0.5299 0.5572 0.7465
No log 3.7188 238 0.6547 0.4756 0.6547 0.8091
No log 3.75 240 0.8095 0.4383 0.8095 0.8997
No log 3.7812 242 0.8077 0.4240 0.8077 0.8987
No log 3.8125 244 0.6316 0.4634 0.6316 0.7948
No log 3.8438 246 0.5430 0.5248 0.5430 0.7369
No log 3.875 248 0.5377 0.5009 0.5377 0.7333
No log 3.9062 250 0.5768 0.5406 0.5768 0.7595
No log 3.9375 252 0.6137 0.5136 0.6137 0.7834
No log 3.9688 254 0.6130 0.5136 0.6130 0.7830
No log 4.0 256 0.5808 0.5319 0.5808 0.7621
No log 4.0312 258 0.6123 0.4767 0.6123 0.7825
No log 4.0625 260 0.6405 0.4574 0.6405 0.8003
No log 4.0938 262 0.6168 0.4735 0.6168 0.7854
No log 4.125 264 0.6763 0.4527 0.6763 0.8224
No log 4.1562 266 0.8600 0.3176 0.8600 0.9274
No log 4.1875 268 0.8174 0.4039 0.8174 0.9041
No log 4.2188 270 0.6388 0.4636 0.6388 0.7993
No log 4.25 272 0.5812 0.4502 0.5812 0.7624
No log 4.2812 274 0.7066 0.3675 0.7066 0.8406
No log 4.3125 276 0.7645 0.3433 0.7645 0.8744
No log 4.3438 278 0.6581 0.3894 0.6581 0.8112
No log 4.375 280 0.5565 0.4397 0.5565 0.7460
No log 4.4062 282 0.6676 0.4038 0.6676 0.8171
No log 4.4375 284 1.0538 0.2457 1.0538 1.0265
No log 4.4688 286 1.2749 0.1686 1.2749 1.1291
No log 4.5 288 1.2700 0.1686 1.2700 1.1269
No log 4.5312 290 1.0474 0.3046 1.0474 1.0234
No log 4.5625 292 0.7401 0.4472 0.7401 0.8603
No log 4.5938 294 0.5855 0.6130 0.5855 0.7652
No log 4.625 296 0.5736 0.5288 0.5736 0.7574
No log 4.6562 298 0.5702 0.5840 0.5702 0.7551
No log 4.6875 300 0.5984 0.5058 0.5984 0.7736
No log 4.7188 302 0.7866 0.4668 0.7866 0.8869
No log 4.75 304 0.9077 0.3846 0.9077 0.9527
No log 4.7812 306 0.8615 0.3882 0.8615 0.9282
No log 4.8125 308 0.7102 0.4290 0.7102 0.8427
No log 4.8438 310 0.5876 0.4813 0.5876 0.7665
No log 4.875 312 0.5744 0.5053 0.5744 0.7579
No log 4.9062 314 0.5698 0.5289 0.5698 0.7549
No log 4.9375 316 0.5759 0.5234 0.5759 0.7589
No log 4.9688 318 0.6642 0.5030 0.6642 0.8150
No log 5.0 320 0.6709 0.4987 0.6709 0.8191
No log 5.0312 322 0.5903 0.5058 0.5903 0.7683
No log 5.0625 324 0.5459 0.5377 0.5459 0.7388
No log 5.0938 326 0.5523 0.5211 0.5523 0.7431
No log 5.125 328 0.5514 0.4622 0.5514 0.7425
No log 5.1562 330 0.5077 0.5195 0.5077 0.7125
No log 5.1875 332 0.5070 0.5970 0.5070 0.7120
No log 5.2188 334 0.5390 0.5335 0.5390 0.7341
No log 5.25 336 0.5444 0.4876 0.5444 0.7379
No log 5.2812 338 0.5266 0.5286 0.5266 0.7257
No log 5.3125 340 0.5214 0.5533 0.5214 0.7221
No log 5.3438 342 0.5352 0.5053 0.5352 0.7315
No log 5.375 344 0.5404 0.5320 0.5404 0.7351
No log 5.4062 346 0.5942 0.5392 0.5942 0.7708
No log 5.4375 348 0.8024 0.4717 0.8024 0.8958
No log 5.4688 350 1.0719 0.3160 1.0719 1.0353
No log 5.5 352 1.1116 0.2744 1.1116 1.0543
No log 5.5312 354 0.9415 0.3480 0.9415 0.9703
No log 5.5625 356 0.6929 0.4859 0.6929 0.8324
No log 5.5938 358 0.5354 0.5797 0.5354 0.7317
No log 5.625 360 0.5301 0.4856 0.5301 0.7281
No log 5.6562 362 0.5245 0.5235 0.5245 0.7242
No log 5.6875 364 0.5696 0.5267 0.5696 0.7547
No log 5.7188 366 0.7182 0.4477 0.7182 0.8475
No log 5.75 368 0.7751 0.4332 0.7751 0.8804
No log 5.7812 370 0.7202 0.4477 0.7202 0.8487
No log 5.8125 372 0.6433 0.4729 0.6433 0.8021
No log 5.8438 374 0.5695 0.4569 0.5695 0.7546
No log 5.875 376 0.5541 0.4489 0.5541 0.7444
No log 5.9062 378 0.5594 0.4035 0.5594 0.7479
No log 5.9375 380 0.5429 0.4586 0.5429 0.7368
No log 5.9688 382 0.5591 0.3323 0.5591 0.7477
No log 6.0 384 0.6410 0.3496 0.6410 0.8007
No log 6.0312 386 0.6865 0.3617 0.6865 0.8285
No log 6.0625 388 0.6732 0.4079 0.6732 0.8205
No log 6.0938 390 0.5998 0.5061 0.5998 0.7745
No log 6.125 392 0.5582 0.5918 0.5582 0.7472
No log 6.1562 394 0.5845 0.4935 0.5845 0.7646
No log 6.1875 396 0.6514 0.4971 0.6514 0.8071
No log 6.2188 398 0.7657 0.5073 0.7657 0.8751
No log 6.25 400 0.7803 0.4536 0.7803 0.8833
No log 6.2812 402 0.6439 0.5175 0.6439 0.8024
No log 6.3125 404 0.5618 0.5283 0.5618 0.7495
No log 6.3438 406 0.5224 0.5413 0.5224 0.7228
No log 6.375 408 0.5337 0.5979 0.5337 0.7306
No log 6.4062 410 0.5827 0.5486 0.5827 0.7633
No log 6.4375 412 0.7204 0.5058 0.7204 0.8487
No log 6.4688 414 0.8924 0.4458 0.8924 0.9447
No log 6.5 416 0.8553 0.4683 0.8553 0.9248
No log 6.5312 418 0.7202 0.5058 0.7202 0.8486
No log 6.5625 420 0.5744 0.6362 0.5744 0.7579
No log 6.5938 422 0.5289 0.64 0.5289 0.7272
No log 6.625 424 0.5148 0.6210 0.5148 0.7175
No log 6.6562 426 0.5055 0.6210 0.5055 0.7110
No log 6.6875 428 0.5014 0.6210 0.5014 0.7081
No log 6.7188 430 0.5146 0.6688 0.5146 0.7174
No log 6.75 432 0.5605 0.5030 0.5605 0.7487
No log 6.7812 434 0.6080 0.5436 0.6080 0.7798
No log 6.8125 436 0.6442 0.5373 0.6442 0.8026
No log 6.8438 438 0.6705 0.5190 0.6705 0.8189
No log 6.875 440 0.6497 0.4634 0.6497 0.8060
No log 6.9062 442 0.5764 0.5127 0.5764 0.7592
No log 6.9375 444 0.5415 0.4497 0.5415 0.7358
No log 6.9688 446 0.5292 0.4591 0.5292 0.7275
No log 7.0 448 0.5270 0.4224 0.5270 0.7260
No log 7.0312 450 0.5354 0.5065 0.5354 0.7317
No log 7.0625 452 0.5637 0.4737 0.5637 0.7508
No log 7.0938 454 0.6118 0.4250 0.6118 0.7821
No log 7.125 456 0.6183 0.4250 0.6183 0.7863
No log 7.1562 458 0.5662 0.4875 0.5662 0.7525
No log 7.1875 460 0.5124 0.6020 0.5124 0.7159
No log 7.2188 462 0.5197 0.6053 0.5197 0.7209
No log 7.25 464 0.5763 0.5283 0.5763 0.7591
No log 7.2812 466 0.6553 0.5662 0.6553 0.8095
No log 7.3125 468 0.6559 0.5662 0.6559 0.8098
No log 7.3438 470 0.6061 0.6135 0.6061 0.7785
No log 7.375 472 0.5440 0.6195 0.5440 0.7376
No log 7.4062 474 0.5216 0.6526 0.5216 0.7222
No log 7.4375 476 0.4774 0.7184 0.4774 0.6910
No log 7.4688 478 0.4687 0.6124 0.4687 0.6846
No log 7.5 480 0.4571 0.5965 0.4571 0.6761
No log 7.5312 482 0.4550 0.6479 0.4550 0.6745
No log 7.5625 484 0.4820 0.5868 0.4820 0.6943
No log 7.5938 486 0.5780 0.5636 0.5780 0.7602
No log 7.625 488 0.6413 0.5436 0.6413 0.8008
No log 7.6562 490 0.6572 0.5436 0.6572 0.8107
No log 7.6875 492 0.6561 0.5500 0.6561 0.8100
No log 7.7188 494 0.6400 0.5500 0.6400 0.8000
No log 7.75 496 0.5746 0.5457 0.5746 0.7580
No log 7.7812 498 0.5341 0.5767 0.5341 0.7308
0.3479 7.8125 500 0.5238 0.5619 0.5238 0.7237
0.3479 7.8438 502 0.5227 0.5929 0.5227 0.7230
0.3479 7.875 504 0.5786 0.5429 0.5786 0.7606
0.3479 7.9062 506 0.6413 0.5090 0.6413 0.8008
0.3479 7.9375 508 0.6639 0.5313 0.6639 0.8148
0.3479 7.9688 510 0.6766 0.5190 0.6766 0.8226
0.3479 8.0 512 0.6481 0.5500 0.6481 0.8050

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task7_organization

Finetuned
(4019)
this model