ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k4_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5694
  • Qwk: 0.5423
  • Mse: 0.5694
  • Rmse: 0.7546

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 2.6243 -0.0788 2.6243 1.6200
No log 0.25 4 1.3092 0.1234 1.3092 1.1442
No log 0.375 6 0.7322 0.1372 0.7322 0.8557
No log 0.5 8 0.6718 0.2181 0.6718 0.8196
No log 0.625 10 0.6131 0.3123 0.6131 0.7830
No log 0.75 12 0.7630 0.3547 0.7630 0.8735
No log 0.875 14 0.8550 0.3094 0.8550 0.9247
No log 1.0 16 0.7916 0.3719 0.7916 0.8897
No log 1.125 18 0.6373 0.3156 0.6373 0.7983
No log 1.25 20 0.5708 0.3499 0.5708 0.7555
No log 1.375 22 0.5692 0.3416 0.5692 0.7545
No log 1.5 24 0.6705 0.3187 0.6705 0.8188
No log 1.625 26 0.7776 0.3620 0.7776 0.8818
No log 1.75 28 0.8041 0.3802 0.8041 0.8967
No log 1.875 30 0.7037 0.3187 0.7037 0.8389
No log 2.0 32 0.6011 0.3137 0.6011 0.7753
No log 2.125 34 0.5754 0.3477 0.5754 0.7585
No log 2.25 36 0.5489 0.4322 0.5489 0.7409
No log 2.375 38 0.6412 0.4814 0.6412 0.8008
No log 2.5 40 0.8703 0.4045 0.8703 0.9329
No log 2.625 42 0.8144 0.4240 0.8144 0.9024
No log 2.75 44 0.6370 0.5351 0.6370 0.7981
No log 2.875 46 0.6117 0.4946 0.6117 0.7821
No log 3.0 48 0.5920 0.4946 0.5920 0.7694
No log 3.125 50 0.4819 0.5267 0.4819 0.6942
No log 3.25 52 0.4881 0.5250 0.4881 0.6986
No log 3.375 54 0.6003 0.5220 0.6003 0.7748
No log 3.5 56 0.6443 0.5421 0.6443 0.8027
No log 3.625 58 0.5708 0.5110 0.5708 0.7555
No log 3.75 60 0.5287 0.5213 0.5287 0.7271
No log 3.875 62 0.5231 0.5671 0.5231 0.7233
No log 4.0 64 0.5006 0.5687 0.5006 0.7076
No log 4.125 66 0.5817 0.5247 0.5817 0.7627
No log 4.25 68 0.9862 0.2679 0.9862 0.9931
No log 4.375 70 0.9474 0.2975 0.9474 0.9733
No log 4.5 72 0.5427 0.4892 0.5427 0.7367
No log 4.625 74 0.4537 0.6024 0.4537 0.6735
No log 4.75 76 0.4542 0.6667 0.4542 0.6740
No log 4.875 78 0.7711 0.5249 0.7711 0.8781
No log 5.0 80 1.0714 0.2147 1.0714 1.0351
No log 5.125 82 0.9356 0.3330 0.9356 0.9673
No log 5.25 84 0.5822 0.5065 0.5822 0.7630
No log 5.375 86 0.4954 0.5758 0.4954 0.7038
No log 5.5 88 0.5011 0.5553 0.5011 0.7079
No log 5.625 90 0.4699 0.5662 0.4699 0.6855
No log 5.75 92 0.5552 0.5095 0.5552 0.7451
No log 5.875 94 0.5710 0.5328 0.5710 0.7556
No log 6.0 96 0.6294 0.5464 0.6294 0.7933
No log 6.125 98 0.5992 0.5643 0.5992 0.7741
No log 6.25 100 0.5407 0.6026 0.5407 0.7353
No log 6.375 102 0.5785 0.5943 0.5785 0.7606
No log 6.5 104 0.5862 0.5675 0.5862 0.7657
No log 6.625 106 0.4416 0.6280 0.4416 0.6645
No log 6.75 108 0.4067 0.6643 0.4067 0.6377
No log 6.875 110 0.4188 0.6830 0.4188 0.6471
No log 7.0 112 0.4917 0.5738 0.4917 0.7012
No log 7.125 114 0.5166 0.5767 0.5166 0.7188
No log 7.25 116 0.4677 0.6275 0.4677 0.6839
No log 7.375 118 0.4720 0.6786 0.4720 0.6870
No log 7.5 120 0.4719 0.6052 0.4719 0.6869
No log 7.625 122 0.4731 0.5231 0.4731 0.6878
No log 7.75 124 0.4637 0.4991 0.4637 0.6809
No log 7.875 126 0.4820 0.5171 0.4820 0.6943
No log 8.0 128 0.4412 0.6242 0.4412 0.6642
No log 8.125 130 0.4490 0.5812 0.4490 0.6701
No log 8.25 132 0.5509 0.5230 0.5509 0.7422
No log 8.375 134 0.6525 0.5034 0.6525 0.8078
No log 8.5 136 0.6134 0.5408 0.6134 0.7832
No log 8.625 138 0.4974 0.6148 0.4974 0.7053
No log 8.75 140 0.4917 0.6867 0.4917 0.7012
No log 8.875 142 0.4843 0.6198 0.4843 0.6959
No log 9.0 144 0.5154 0.5452 0.5154 0.7179
No log 9.125 146 0.5193 0.5230 0.5193 0.7206
No log 9.25 148 0.4611 0.5658 0.4611 0.6790
No log 9.375 150 0.4450 0.5782 0.4450 0.6671
No log 9.5 152 0.4608 0.5836 0.4608 0.6788
No log 9.625 154 0.4765 0.6286 0.4765 0.6903
No log 9.75 156 0.4280 0.6942 0.4280 0.6542
No log 9.875 158 0.6419 0.4669 0.6419 0.8012
No log 10.0 160 0.9024 0.3424 0.9024 0.9500
No log 10.125 162 0.7381 0.5050 0.7381 0.8591
No log 10.25 164 0.4433 0.6434 0.4433 0.6658
No log 10.375 166 0.5372 0.5904 0.5372 0.7329
No log 10.5 168 0.5383 0.5264 0.5383 0.7337
No log 10.625 170 0.4290 0.6745 0.4290 0.6550
No log 10.75 172 0.5232 0.4795 0.5232 0.7233
No log 10.875 174 0.6850 0.3917 0.6850 0.8277
No log 11.0 176 0.6465 0.4193 0.6465 0.8041
No log 11.125 178 0.4964 0.5461 0.4964 0.7045
No log 11.25 180 0.4692 0.5675 0.4692 0.6849
No log 11.375 182 0.4977 0.4703 0.4977 0.7055
No log 11.5 184 0.5173 0.5056 0.5173 0.7192
No log 11.625 186 0.4882 0.6198 0.4882 0.6987
No log 11.75 188 0.4913 0.6298 0.4913 0.7009
No log 11.875 190 0.4752 0.6108 0.4752 0.6893
No log 12.0 192 0.5152 0.6311 0.5152 0.7178
No log 12.125 194 0.5366 0.6132 0.5366 0.7325
No log 12.25 196 0.4757 0.6423 0.4757 0.6897
No log 12.375 198 0.4540 0.5707 0.4540 0.6738
No log 12.5 200 0.4820 0.6169 0.4820 0.6942
No log 12.625 202 0.4846 0.5657 0.4846 0.6962
No log 12.75 204 0.4543 0.5246 0.4543 0.6740
No log 12.875 206 0.4602 0.5343 0.4602 0.6784
No log 13.0 208 0.4743 0.5343 0.4743 0.6887
No log 13.125 210 0.4838 0.5397 0.4838 0.6955
No log 13.25 212 0.4796 0.5983 0.4796 0.6925
No log 13.375 214 0.5246 0.6137 0.5246 0.7243
No log 13.5 216 0.4969 0.5983 0.4969 0.7049
No log 13.625 218 0.5287 0.5661 0.5287 0.7271
No log 13.75 220 0.6631 0.5034 0.6631 0.8143
No log 13.875 222 0.6399 0.5093 0.6399 0.7999
No log 14.0 224 0.5031 0.5510 0.5031 0.7093
No log 14.125 226 0.4612 0.6417 0.4612 0.6792
No log 14.25 228 0.4625 0.6417 0.4625 0.6801
No log 14.375 230 0.4754 0.5343 0.4754 0.6895
No log 14.5 232 0.5481 0.5449 0.5481 0.7403
No log 14.625 234 0.5297 0.5356 0.5297 0.7278
No log 14.75 236 0.4771 0.5475 0.4771 0.6907
No log 14.875 238 0.4839 0.5397 0.4839 0.6956
No log 15.0 240 0.5089 0.5127 0.5089 0.7134
No log 15.125 242 0.4931 0.5178 0.4931 0.7022
No log 15.25 244 0.5470 0.4875 0.5470 0.7396
No log 15.375 246 0.6549 0.5251 0.6549 0.8093
No log 15.5 248 0.5759 0.4997 0.5759 0.7589
No log 15.625 250 0.5196 0.4832 0.5196 0.7208
No log 15.75 252 0.5673 0.5078 0.5673 0.7532
No log 15.875 254 0.6007 0.5095 0.6007 0.7751
No log 16.0 256 0.5218 0.5117 0.5218 0.7224
No log 16.125 258 0.4885 0.4774 0.4885 0.6990
No log 16.25 260 0.4865 0.4774 0.4865 0.6975
No log 16.375 262 0.4906 0.5085 0.4906 0.7005
No log 16.5 264 0.4801 0.5171 0.4801 0.6929
No log 16.625 266 0.4761 0.5141 0.4761 0.6900
No log 16.75 268 0.5018 0.5957 0.5018 0.7084
No log 16.875 270 0.4842 0.6251 0.4842 0.6958
No log 17.0 272 0.4567 0.5768 0.4567 0.6758
No log 17.125 274 0.5166 0.5348 0.5166 0.7188
No log 17.25 276 0.5804 0.5393 0.5804 0.7618
No log 17.375 278 0.5198 0.4931 0.5198 0.7210
No log 17.5 280 0.4617 0.5930 0.4617 0.6795
No log 17.625 282 0.4620 0.5784 0.4620 0.6797
No log 17.75 284 0.4564 0.6017 0.4564 0.6756
No log 17.875 286 0.5135 0.5299 0.5135 0.7166
No log 18.0 288 0.6467 0.5153 0.6467 0.8042
No log 18.125 290 0.6914 0.4959 0.6914 0.8315
No log 18.25 292 0.5798 0.5178 0.5798 0.7614
No log 18.375 294 0.4937 0.5010 0.4937 0.7027
No log 18.5 296 0.4831 0.5085 0.4831 0.6950
No log 18.625 298 0.4758 0.5556 0.4758 0.6898
No log 18.75 300 0.4865 0.4960 0.4865 0.6975
No log 18.875 302 0.4799 0.5702 0.4799 0.6927
No log 19.0 304 0.4717 0.5722 0.4717 0.6868
No log 19.125 306 0.4758 0.5663 0.4758 0.6898
No log 19.25 308 0.4839 0.5283 0.4839 0.6957
No log 19.375 310 0.4766 0.5765 0.4766 0.6903
No log 19.5 312 0.4702 0.5926 0.4702 0.6857
No log 19.625 314 0.4732 0.5718 0.4732 0.6879
No log 19.75 316 0.4852 0.6146 0.4852 0.6966
No log 19.875 318 0.5330 0.5630 0.5330 0.7301
No log 20.0 320 0.5463 0.5630 0.5463 0.7391
No log 20.125 322 0.5287 0.6123 0.5287 0.7271
No log 20.25 324 0.5081 0.6158 0.5081 0.7128
No log 20.375 326 0.5341 0.5696 0.5341 0.7308
No log 20.5 328 0.5502 0.5710 0.5502 0.7417
No log 20.625 330 0.5146 0.5513 0.5146 0.7174
No log 20.75 332 0.4866 0.5065 0.4866 0.6976
No log 20.875 334 0.4671 0.4960 0.4671 0.6835
No log 21.0 336 0.4570 0.5046 0.4570 0.6760
No log 21.125 338 0.5133 0.5015 0.5133 0.7165
No log 21.25 340 0.5977 0.5489 0.5977 0.7731
No log 21.375 342 0.6127 0.5169 0.6127 0.7828
No log 21.5 344 0.5085 0.4997 0.5085 0.7131
No log 21.625 346 0.4369 0.6643 0.4369 0.6610
No log 21.75 348 0.4785 0.6063 0.4785 0.6917
No log 21.875 350 0.4768 0.6063 0.4768 0.6905
No log 22.0 352 0.4395 0.6736 0.4395 0.6630
No log 22.125 354 0.4665 0.5065 0.4665 0.6830
No log 22.25 356 0.5254 0.5735 0.5254 0.7248
No log 22.375 358 0.4974 0.5455 0.4974 0.7052
No log 22.5 360 0.4478 0.6170 0.4478 0.6692
No log 22.625 362 0.5409 0.6267 0.5409 0.7355
No log 22.75 364 0.6670 0.6103 0.6670 0.8167
No log 22.875 366 0.6418 0.6275 0.6418 0.8011
No log 23.0 368 0.5289 0.6231 0.5289 0.7273
No log 23.125 370 0.5128 0.5151 0.5128 0.7161
No log 23.25 372 0.5484 0.5552 0.5484 0.7405
No log 23.375 374 0.5561 0.5595 0.5561 0.7457
No log 23.5 376 0.5006 0.5368 0.5006 0.7075
No log 23.625 378 0.4666 0.5985 0.4666 0.6831
No log 23.75 380 0.4729 0.6083 0.4729 0.6877
No log 23.875 382 0.4812 0.6100 0.4812 0.6937
No log 24.0 384 0.4713 0.5783 0.4713 0.6865
No log 24.125 386 0.5132 0.5513 0.5132 0.7164
No log 24.25 388 0.5934 0.5343 0.5934 0.7703
No log 24.375 390 0.5788 0.5266 0.5788 0.7608
No log 24.5 392 0.5079 0.5457 0.5079 0.7126
No log 24.625 394 0.4859 0.4913 0.4859 0.6971
No log 24.75 396 0.4914 0.4694 0.4914 0.7010
No log 24.875 398 0.4741 0.4958 0.4741 0.6886
No log 25.0 400 0.4631 0.5010 0.4631 0.6805
No log 25.125 402 0.4631 0.5010 0.4631 0.6805
No log 25.25 404 0.4582 0.5079 0.4582 0.6769
No log 25.375 406 0.4626 0.5168 0.4626 0.6801
No log 25.5 408 0.4740 0.5159 0.4740 0.6885
No log 25.625 410 0.4742 0.5039 0.4742 0.6886
No log 25.75 412 0.4939 0.5373 0.4939 0.7028
No log 25.875 414 0.4977 0.5560 0.4977 0.7055
No log 26.0 416 0.5247 0.5898 0.5247 0.7244
No log 26.125 418 0.5535 0.5643 0.5535 0.7440
No log 26.25 420 0.5273 0.5543 0.5273 0.7262
No log 26.375 422 0.4607 0.5110 0.4607 0.6788
No log 26.5 424 0.4407 0.5248 0.4407 0.6638
No log 26.625 426 0.4497 0.6797 0.4497 0.6706
No log 26.75 428 0.4597 0.6797 0.4597 0.6780
No log 26.875 430 0.4492 0.6129 0.4492 0.6702
No log 27.0 432 0.4444 0.6129 0.4444 0.6666
No log 27.125 434 0.4401 0.5143 0.4401 0.6634
No log 27.25 436 0.4472 0.5178 0.4472 0.6687
No log 27.375 438 0.4541 0.5104 0.4541 0.6739
No log 27.5 440 0.4716 0.5780 0.4716 0.6867
No log 27.625 442 0.5056 0.6129 0.5056 0.7111
No log 27.75 444 0.4823 0.5849 0.4823 0.6945
No log 27.875 446 0.4774 0.5922 0.4774 0.6909
No log 28.0 448 0.4732 0.5672 0.4732 0.6879
No log 28.125 450 0.4579 0.5230 0.4579 0.6767
No log 28.25 452 0.4561 0.4756 0.4561 0.6753
No log 28.375 454 0.4388 0.5265 0.4388 0.6624
No log 28.5 456 0.4357 0.5305 0.4357 0.6601
No log 28.625 458 0.4394 0.5800 0.4394 0.6629
No log 28.75 460 0.4412 0.5361 0.4412 0.6642
No log 28.875 462 0.4406 0.5269 0.4406 0.6637
No log 29.0 464 0.4439 0.5254 0.4439 0.6663
No log 29.125 466 0.4590 0.5533 0.4590 0.6775
No log 29.25 468 0.4610 0.5505 0.4610 0.6790
No log 29.375 470 0.4738 0.6021 0.4738 0.6883
No log 29.5 472 0.4947 0.5348 0.4947 0.7034
No log 29.625 474 0.5380 0.5152 0.5380 0.7335
No log 29.75 476 0.5601 0.5249 0.5601 0.7484
No log 29.875 478 0.5411 0.5249 0.5411 0.7356
No log 30.0 480 0.5006 0.5524 0.5006 0.7076
No log 30.125 482 0.4709 0.5392 0.4709 0.6862
No log 30.25 484 0.4370 0.6388 0.4370 0.6611
No log 30.375 486 0.4285 0.6198 0.4285 0.6546
No log 30.5 488 0.4426 0.5467 0.4426 0.6653
No log 30.625 490 0.4707 0.5470 0.4707 0.6861
No log 30.75 492 0.4933 0.5195 0.4933 0.7024
No log 30.875 494 0.4711 0.5470 0.4711 0.6864
No log 31.0 496 0.4203 0.5831 0.4203 0.6483
No log 31.125 498 0.4088 0.6140 0.4088 0.6394
0.2293 31.25 500 0.4130 0.6257 0.4130 0.6426
0.2293 31.375 502 0.4518 0.5599 0.4518 0.6722
0.2293 31.5 504 0.5474 0.5948 0.5474 0.7399
0.2293 31.625 506 0.5922 0.5609 0.5922 0.7695
0.2293 31.75 508 0.6267 0.5358 0.6267 0.7916
0.2293 31.875 510 0.5694 0.5423 0.5694 0.7546

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k4_task7_organization

Finetuned
(4019)
this model