ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5219
  • Qwk: 0.4437
  • Mse: 0.5219
  • Rmse: 0.7224

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 2.7236 -0.0829 2.7236 1.6503
No log 0.1111 4 1.6801 -0.0216 1.6801 1.2962
No log 0.1667 6 1.4297 -0.2040 1.4297 1.1957
No log 0.2222 8 1.1354 -0.0142 1.1354 1.0655
No log 0.2778 10 1.1852 0.0566 1.1852 1.0887
No log 0.3333 12 1.2509 -0.0025 1.2509 1.1184
No log 0.3889 14 1.0550 0.1029 1.0550 1.0271
No log 0.4444 16 0.8806 0.1400 0.8806 0.9384
No log 0.5 18 0.7622 -0.0054 0.7622 0.8730
No log 0.5556 20 0.7264 0.1770 0.7264 0.8523
No log 0.6111 22 0.7896 0.2804 0.7896 0.8886
No log 0.6667 24 0.7253 0.2181 0.7253 0.8516
No log 0.7222 26 0.7346 0.0481 0.7346 0.8571
No log 0.7778 28 0.7819 0.2558 0.7819 0.8843
No log 0.8333 30 0.7496 0.1714 0.7496 0.8658
No log 0.8889 32 0.6853 0.1942 0.6853 0.8278
No log 0.9444 34 0.7576 0.2769 0.7576 0.8704
No log 1.0 36 0.9897 0.2627 0.9897 0.9948
No log 1.0556 38 0.9323 0.3206 0.9323 0.9655
No log 1.1111 40 0.8054 0.2790 0.8054 0.8974
No log 1.1667 42 0.8721 0.3263 0.8721 0.9339
No log 1.2222 44 0.9258 0.2040 0.9258 0.9622
No log 1.2778 46 0.7929 0.3035 0.7929 0.8905
No log 1.3333 48 0.7769 0.3978 0.7769 0.8814
No log 1.3889 50 0.6653 0.3754 0.6653 0.8157
No log 1.4444 52 0.6196 0.3806 0.6196 0.7872
No log 1.5 54 0.6716 0.3287 0.6716 0.8195
No log 1.5556 56 0.6642 0.3615 0.6642 0.8150
No log 1.6111 58 0.6716 0.3467 0.6716 0.8195
No log 1.6667 60 0.6887 0.3175 0.6887 0.8299
No log 1.7222 62 0.7544 0.3590 0.7544 0.8686
No log 1.7778 64 0.7420 0.3629 0.7420 0.8614
No log 1.8333 66 0.7493 0.3962 0.7493 0.8656
No log 1.8889 68 0.8247 0.3787 0.8247 0.9081
No log 1.9444 70 0.7944 0.3872 0.7944 0.8913
No log 2.0 72 0.6905 0.4562 0.6905 0.8310
No log 2.0556 74 0.6854 0.4562 0.6854 0.8279
No log 2.1111 76 0.7520 0.3916 0.7520 0.8672
No log 2.1667 78 0.7369 0.3916 0.7369 0.8584
No log 2.2222 80 0.6967 0.4230 0.6967 0.8347
No log 2.2778 82 0.6725 0.3988 0.6725 0.8201
No log 2.3333 84 0.6215 0.4182 0.6215 0.7884
No log 2.3889 86 0.7363 0.4223 0.7363 0.8581
No log 2.4444 88 0.6753 0.4562 0.6753 0.8218
No log 2.5 90 0.6607 0.3964 0.6607 0.8128
No log 2.5556 92 0.6986 0.3890 0.6986 0.8358
No log 2.6111 94 1.0016 0.3043 1.0016 1.0008
No log 2.6667 96 1.3607 0.0354 1.3607 1.1665
No log 2.7222 98 1.0995 0.1789 1.0995 1.0486
No log 2.7778 100 0.7030 0.4895 0.7030 0.8385
No log 2.8333 102 0.7151 0.4322 0.7151 0.8457
No log 2.8889 104 0.6972 0.3786 0.6972 0.8350
No log 2.9444 106 0.7121 0.4596 0.7121 0.8439
No log 3.0 108 0.7315 0.4714 0.7315 0.8553
No log 3.0556 110 0.6576 0.5159 0.6576 0.8109
No log 3.1111 112 0.8770 0.4032 0.8770 0.9365
No log 3.1667 114 1.2570 0.1641 1.2570 1.1212
No log 3.2222 116 1.3219 0.0789 1.3219 1.1497
No log 3.2778 118 1.1286 0.1871 1.1286 1.0624
No log 3.3333 120 0.7899 0.4562 0.7899 0.8888
No log 3.3889 122 0.6047 0.5117 0.6047 0.7777
No log 3.4444 124 0.5840 0.4186 0.5840 0.7642
No log 3.5 126 0.5831 0.3834 0.5831 0.7636
No log 3.5556 128 0.5947 0.4717 0.5947 0.7712
No log 3.6111 130 0.6396 0.4360 0.6396 0.7997
No log 3.6667 132 0.7972 0.3305 0.7972 0.8929
No log 3.7222 134 0.8723 0.3008 0.8723 0.9340
No log 3.7778 136 0.7013 0.4597 0.7013 0.8374
No log 3.8333 138 0.5773 0.5057 0.5773 0.7598
No log 3.8889 140 0.5724 0.4832 0.5724 0.7566
No log 3.9444 142 0.5495 0.5089 0.5495 0.7413
No log 4.0 144 0.6571 0.4502 0.6571 0.8106
No log 4.0556 146 0.6469 0.4892 0.6469 0.8043
No log 4.1111 148 0.5328 0.5017 0.5328 0.7300
No log 4.1667 150 0.5094 0.4698 0.5094 0.7137
No log 4.2222 152 0.5021 0.5098 0.5021 0.7086
No log 4.2778 154 0.5000 0.5214 0.5000 0.7071
No log 4.3333 156 0.5263 0.5485 0.5263 0.7255
No log 4.3889 158 0.5243 0.5405 0.5243 0.7241
No log 4.4444 160 0.6800 0.4562 0.6800 0.8246
No log 4.5 162 0.7998 0.4177 0.7998 0.8943
No log 4.5556 164 0.7577 0.4315 0.7577 0.8705
No log 4.6111 166 0.6159 0.4652 0.6159 0.7848
No log 4.6667 168 0.5318 0.4617 0.5318 0.7293
No log 4.7222 170 0.5924 0.5252 0.5924 0.7697
No log 4.7778 172 0.6350 0.5385 0.6350 0.7969
No log 4.8333 174 0.6155 0.5385 0.6155 0.7846
No log 4.8889 176 0.5049 0.5323 0.5049 0.7106
No log 4.9444 178 0.5192 0.5918 0.5192 0.7206
No log 5.0 180 0.6148 0.4961 0.6148 0.7841
No log 5.0556 182 0.6888 0.4719 0.6888 0.8299
No log 5.1111 184 0.5977 0.5003 0.5977 0.7731
No log 5.1667 186 0.5642 0.5500 0.5642 0.7511
No log 5.2222 188 0.5575 0.5168 0.5575 0.7466
No log 5.2778 190 0.5384 0.5339 0.5384 0.7337
No log 5.3333 192 0.5124 0.5189 0.5124 0.7158
No log 5.3889 194 0.5263 0.4828 0.5263 0.7255
No log 5.4444 196 0.5201 0.5114 0.5201 0.7212
No log 5.5 198 0.5761 0.5836 0.5761 0.7590
No log 5.5556 200 0.6459 0.5077 0.6459 0.8037
No log 5.6111 202 0.6137 0.4708 0.6137 0.7834
No log 5.6667 204 0.5366 0.5379 0.5366 0.7325
No log 5.7222 206 0.5148 0.4569 0.5148 0.7175
No log 5.7778 208 0.5015 0.4929 0.5015 0.7082
No log 5.8333 210 0.4904 0.4857 0.4904 0.7003
No log 5.8889 212 0.5118 0.5028 0.5118 0.7154
No log 5.9444 214 0.5068 0.4555 0.5068 0.7119
No log 6.0 216 0.5277 0.5560 0.5277 0.7264
No log 6.0556 218 0.6800 0.5124 0.6800 0.8247
No log 6.1111 220 0.6516 0.5124 0.6516 0.8072
No log 6.1667 222 0.5371 0.5617 0.5371 0.7329
No log 6.2222 224 0.5359 0.5378 0.5359 0.7321
No log 6.2778 226 0.6456 0.4961 0.6456 0.8035
No log 6.3333 228 0.6559 0.4961 0.6559 0.8099
No log 6.3889 230 0.5770 0.5515 0.5770 0.7596
No log 6.4444 232 0.5256 0.4575 0.5256 0.7250
No log 6.5 234 0.5240 0.4380 0.5240 0.7239
No log 6.5556 236 0.5494 0.4752 0.5494 0.7412
No log 6.6111 238 0.6313 0.4327 0.6313 0.7945
No log 6.6667 240 0.7100 0.4961 0.7100 0.8426
No log 6.7222 242 0.6350 0.5017 0.6350 0.7969
No log 6.7778 244 0.5342 0.5741 0.5342 0.7309
No log 6.8333 246 0.5575 0.6011 0.5575 0.7467
No log 6.8889 248 0.5628 0.5712 0.5628 0.7502
No log 6.9444 250 0.5085 0.5533 0.5085 0.7131
No log 7.0 252 0.6250 0.4279 0.6250 0.7906
No log 7.0556 254 0.7560 0.4286 0.7560 0.8695
No log 7.1111 256 0.6965 0.4364 0.6965 0.8346
No log 7.1667 258 0.5923 0.4969 0.5923 0.7696
No log 7.2222 260 0.5296 0.4914 0.5296 0.7278
No log 7.2778 262 0.6237 0.4664 0.6237 0.7897
No log 7.3333 264 0.6160 0.4745 0.6160 0.7848
No log 7.3889 266 0.5250 0.4816 0.5250 0.7246
No log 7.4444 268 0.6081 0.4103 0.6081 0.7798
No log 7.5 270 0.7988 0.4479 0.7988 0.8937
No log 7.5556 272 0.7806 0.4479 0.7806 0.8835
No log 7.6111 274 0.6450 0.5692 0.6450 0.8031
No log 7.6667 276 0.5597 0.5505 0.5597 0.7481
No log 7.7222 278 0.6587 0.5308 0.6587 0.8116
No log 7.7778 280 0.6609 0.5259 0.6609 0.8130
No log 7.8333 282 0.5821 0.5601 0.5821 0.7629
No log 7.8889 284 0.5195 0.5252 0.5195 0.7207
No log 7.9444 286 0.5411 0.6194 0.5411 0.7356
No log 8.0 288 0.5700 0.6144 0.5700 0.7550
No log 8.0556 290 0.5634 0.5778 0.5634 0.7506
No log 8.1111 292 0.5236 0.5509 0.5236 0.7236
No log 8.1667 294 0.5144 0.5421 0.5144 0.7172
No log 8.2222 296 0.5245 0.5421 0.5245 0.7242
No log 8.2778 298 0.5448 0.4883 0.5448 0.7381
No log 8.3333 300 0.5734 0.4866 0.5734 0.7572
No log 8.3889 302 0.5623 0.4824 0.5623 0.7498
No log 8.4444 304 0.5492 0.4953 0.5492 0.7411
No log 8.5 306 0.5538 0.4701 0.5538 0.7442
No log 8.5556 308 0.5504 0.4953 0.5504 0.7419
No log 8.6111 310 0.5518 0.4953 0.5518 0.7429
No log 8.6667 312 0.5552 0.4953 0.5552 0.7451
No log 8.7222 314 0.5731 0.5178 0.5731 0.7571
No log 8.7778 316 0.5846 0.5512 0.5846 0.7646
No log 8.8333 318 0.5922 0.5078 0.5922 0.7695
No log 8.8889 320 0.5685 0.4832 0.5685 0.7540
No log 8.9444 322 0.5489 0.4678 0.5489 0.7409
No log 9.0 324 0.5636 0.4655 0.5636 0.7507
No log 9.0556 326 0.5766 0.4979 0.5766 0.7593
No log 9.1111 328 0.5737 0.5283 0.5737 0.7574
No log 9.1667 330 0.6599 0.3556 0.6599 0.8123
No log 9.2222 332 0.7263 0.3719 0.7263 0.8522
No log 9.2778 334 0.6681 0.3804 0.6681 0.8174
No log 9.3333 336 0.5467 0.4838 0.5467 0.7394
No log 9.3889 338 0.5194 0.4482 0.5194 0.7207
No log 9.4444 340 0.5857 0.5523 0.5857 0.7653
No log 9.5 342 0.5674 0.6189 0.5674 0.7533
No log 9.5556 344 0.5360 0.5813 0.5360 0.7321
No log 9.6111 346 0.5732 0.5137 0.5732 0.7571
No log 9.6667 348 0.6115 0.5160 0.6115 0.7820
No log 9.7222 350 0.5991 0.5160 0.5991 0.7740
No log 9.7778 352 0.5241 0.5120 0.5241 0.7239
No log 9.8333 354 0.5039 0.5538 0.5039 0.7099
No log 9.8889 356 0.5031 0.5687 0.5031 0.7093
No log 9.9444 358 0.5030 0.5877 0.5030 0.7092
No log 10.0 360 0.4897 0.5379 0.4897 0.6998
No log 10.0556 362 0.4897 0.5397 0.4897 0.6998
No log 10.1111 364 0.4789 0.5800 0.4789 0.6920
No log 10.1667 366 0.4795 0.5076 0.4795 0.6925
No log 10.2222 368 0.5147 0.4704 0.5147 0.7174
No log 10.2778 370 0.5285 0.4618 0.5285 0.7270
No log 10.3333 372 0.4884 0.4895 0.4884 0.6989
No log 10.3889 374 0.4792 0.5698 0.4792 0.6922
No log 10.4444 376 0.4971 0.6096 0.4971 0.7050
No log 10.5 378 0.5113 0.5682 0.5113 0.7151
No log 10.5556 380 0.4823 0.5658 0.4823 0.6945
No log 10.6111 382 0.4839 0.5405 0.4839 0.6956
No log 10.6667 384 0.5192 0.5449 0.5192 0.7206
No log 10.7222 386 0.5865 0.4860 0.5865 0.7658
No log 10.7778 388 0.5685 0.5129 0.5685 0.7540
No log 10.8333 390 0.4851 0.6032 0.4851 0.6965
No log 10.8889 392 0.4853 0.4883 0.4853 0.6967
No log 10.9444 394 0.5059 0.5098 0.5059 0.7113
No log 11.0 396 0.4847 0.4397 0.4847 0.6962
No log 11.0556 398 0.4655 0.6091 0.4655 0.6823
No log 11.1111 400 0.5064 0.5544 0.5064 0.7116
No log 11.1667 402 0.5141 0.5470 0.5141 0.7170
No log 11.2222 404 0.4716 0.6047 0.4716 0.6867
No log 11.2778 406 0.4581 0.5658 0.4581 0.6768
No log 11.3333 408 0.4546 0.5869 0.4546 0.6743
No log 11.3889 410 0.4572 0.5344 0.4572 0.6761
No log 11.4444 412 0.4569 0.6370 0.4569 0.6759
No log 11.5 414 0.4694 0.6096 0.4694 0.6851
No log 11.5556 416 0.5127 0.5706 0.5127 0.7160
No log 11.6111 418 0.6347 0.5281 0.6347 0.7967
No log 11.6667 420 0.6228 0.5281 0.6228 0.7892
No log 11.7222 422 0.5278 0.5581 0.5278 0.7265
No log 11.7778 424 0.4613 0.5796 0.4613 0.6792
No log 11.8333 426 0.4529 0.5960 0.4529 0.6730
No log 11.8889 428 0.4496 0.5731 0.4496 0.6705
No log 11.9444 430 0.4411 0.5324 0.4411 0.6642
No log 12.0 432 0.4404 0.5324 0.4404 0.6636
No log 12.0556 434 0.4466 0.5414 0.4466 0.6683
No log 12.1111 436 0.4458 0.5414 0.4458 0.6676
No log 12.1667 438 0.4791 0.5855 0.4791 0.6922
No log 12.2222 440 0.4923 0.5706 0.4923 0.7016
No log 12.2778 442 0.4460 0.5414 0.4460 0.6678
No log 12.3333 444 0.4401 0.5475 0.4401 0.6634
No log 12.3889 446 0.4444 0.5414 0.4444 0.6666
No log 12.4444 448 0.4626 0.5571 0.4626 0.6801
No log 12.5 450 0.4993 0.5445 0.4993 0.7066
No log 12.5556 452 0.4776 0.6013 0.4776 0.6911
No log 12.6111 454 0.4647 0.5649 0.4647 0.6817
No log 12.6667 456 0.4700 0.5649 0.4700 0.6856
No log 12.7222 458 0.4706 0.5414 0.4706 0.6860
No log 12.7778 460 0.4830 0.5432 0.4830 0.6950
No log 12.8333 462 0.5144 0.5892 0.5144 0.7172
No log 12.8889 464 0.5352 0.5373 0.5352 0.7316
No log 12.9444 466 0.5217 0.5373 0.5217 0.7223
No log 13.0 468 0.5441 0.5300 0.5441 0.7376
No log 13.0556 470 0.5332 0.5300 0.5332 0.7302
No log 13.1111 472 0.4862 0.5930 0.4862 0.6973
No log 13.1667 474 0.4839 0.5765 0.4839 0.6956
No log 13.2222 476 0.4924 0.5267 0.4924 0.7017
No log 13.2778 478 0.4994 0.5203 0.4994 0.7067
No log 13.3333 480 0.4979 0.4634 0.4979 0.7056
No log 13.3889 482 0.5067 0.4547 0.5067 0.7118
No log 13.4444 484 0.5185 0.4292 0.5185 0.7201
No log 13.5 486 0.4874 0.5182 0.4874 0.6981
No log 13.5556 488 0.5008 0.5339 0.5008 0.7077
No log 13.6111 490 0.5587 0.5015 0.5587 0.7474
No log 13.6667 492 0.5393 0.5230 0.5393 0.7344
No log 13.7222 494 0.4893 0.5414 0.4893 0.6995
No log 13.7778 496 0.4749 0.5641 0.4749 0.6892
No log 13.8333 498 0.4915 0.4637 0.4915 0.7011
0.305 13.8889 500 0.4910 0.4637 0.4910 0.7007
0.305 13.9444 502 0.4860 0.4637 0.4860 0.6971
0.305 14.0 504 0.4834 0.5386 0.4834 0.6952
0.305 14.0556 506 0.4896 0.4289 0.4896 0.6997
0.305 14.1111 508 0.4958 0.4382 0.4958 0.7041
0.305 14.1667 510 0.5069 0.3953 0.5069 0.7120
0.305 14.2222 512 0.5334 0.4060 0.5334 0.7303
0.305 14.2778 514 0.5219 0.4437 0.5219 0.7224

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task7_organization

Finetuned
(4019)
this model