ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k3_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4413
  • Qwk: 0.5600
  • Mse: 0.4413
  • Rmse: 0.6643

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 2.4893 -0.0788 2.4893 1.5778
No log 0.2667 4 1.1296 0.0983 1.1296 1.0628
No log 0.4 6 0.7618 0.0937 0.7618 0.8728
No log 0.5333 8 0.8008 0.2574 0.8008 0.8949
No log 0.6667 10 0.8432 0.2756 0.8432 0.9183
No log 0.8 12 0.9972 0.2589 0.9972 0.9986
No log 0.9333 14 0.8926 0.2703 0.8926 0.9448
No log 1.0667 16 0.6849 0.1646 0.6849 0.8276
No log 1.2 18 0.7589 0.2804 0.7589 0.8711
No log 1.3333 20 0.9308 0.3347 0.9308 0.9648
No log 1.4667 22 0.8305 0.4088 0.8305 0.9113
No log 1.6 24 0.5339 0.4634 0.5339 0.7307
No log 1.7333 26 0.7801 0.3716 0.7801 0.8832
No log 1.8667 28 0.7159 0.4367 0.7159 0.8461
No log 2.0 30 0.5663 0.4337 0.5663 0.7525
No log 2.1333 32 0.6322 0.4412 0.6322 0.7951
No log 2.2667 34 1.2334 0.1678 1.2334 1.1106
No log 2.4 36 1.4639 0.0595 1.4639 1.2099
No log 2.5333 38 1.3420 0.1005 1.3420 1.1585
No log 2.6667 40 0.9122 0.2627 0.9122 0.9551
No log 2.8 42 0.5432 0.4229 0.5432 0.7370
No log 2.9333 44 0.7369 0.3739 0.7369 0.8584
No log 3.0667 46 0.7417 0.3676 0.7417 0.8612
No log 3.2 48 0.6526 0.3409 0.6526 0.8078
No log 3.3333 50 0.6735 0.3347 0.6735 0.8207
No log 3.4667 52 0.5685 0.4816 0.5685 0.7540
No log 3.6 54 0.5725 0.4452 0.5725 0.7567
No log 3.7333 56 0.6759 0.4416 0.6759 0.8221
No log 3.8667 58 0.6871 0.4357 0.6871 0.8289
No log 4.0 60 0.7039 0.4609 0.7039 0.8390
No log 4.1333 62 0.6321 0.3931 0.6321 0.7950
No log 4.2667 64 0.5901 0.3643 0.5901 0.7682
No log 4.4 66 0.5859 0.3693 0.5859 0.7655
No log 4.5333 68 0.5420 0.3719 0.5420 0.7362
No log 4.6667 70 0.5090 0.3585 0.5090 0.7134
No log 4.8 72 0.5506 0.5498 0.5506 0.7420
No log 4.9333 74 0.5614 0.5252 0.5614 0.7493
No log 5.0667 76 0.5362 0.5736 0.5362 0.7323
No log 5.2 78 0.5137 0.4726 0.5137 0.7167
No log 5.3333 80 0.5774 0.4394 0.5774 0.7599
No log 5.4667 82 0.6445 0.3926 0.6445 0.8028
No log 5.6 84 0.5842 0.4306 0.5842 0.7644
No log 5.7333 86 0.5351 0.4026 0.5351 0.7315
No log 5.8667 88 0.5043 0.5250 0.5043 0.7102
No log 6.0 90 0.4918 0.5687 0.4918 0.7013
No log 6.1333 92 0.4790 0.4991 0.4790 0.6921
No log 6.2667 94 0.6282 0.5232 0.6282 0.7926
No log 6.4 96 0.9231 0.3444 0.9231 0.9608
No log 6.5333 98 0.8470 0.3766 0.8470 0.9203
No log 6.6667 100 0.5479 0.5614 0.5479 0.7402
No log 6.8 102 0.4910 0.5587 0.4910 0.7007
No log 6.9333 104 0.4702 0.5373 0.4702 0.6857
No log 7.0667 106 0.5109 0.5422 0.5109 0.7148
No log 7.2 108 0.5011 0.5736 0.5011 0.7079
No log 7.3333 110 0.4858 0.5860 0.4858 0.6970
No log 7.4667 112 0.5279 0.6188 0.5279 0.7265
No log 7.6 114 0.5384 0.6175 0.5384 0.7338
No log 7.7333 116 0.4780 0.5936 0.4780 0.6914
No log 7.8667 118 0.5510 0.5997 0.5510 0.7423
No log 8.0 120 0.5578 0.6193 0.5578 0.7469
No log 8.1333 122 0.5507 0.6261 0.5507 0.7421
No log 8.2667 124 0.4937 0.6251 0.4937 0.7026
No log 8.4 126 0.4615 0.6706 0.4615 0.6793
No log 8.5333 128 0.4625 0.6263 0.4625 0.6801
No log 8.6667 130 0.4477 0.5687 0.4477 0.6691
No log 8.8 132 0.4949 0.5855 0.4949 0.7035
No log 8.9333 134 0.4394 0.6053 0.4394 0.6629
No log 9.0667 136 0.4708 0.5845 0.4708 0.6861
No log 9.2 138 0.4438 0.6032 0.4438 0.6662
No log 9.3333 140 0.4421 0.5985 0.4421 0.6649
No log 9.4667 142 0.4319 0.5985 0.4319 0.6572
No log 9.6 144 0.4474 0.6210 0.4474 0.6689
No log 9.7333 146 0.4946 0.5445 0.4946 0.7033
No log 9.8667 148 0.4720 0.5933 0.4720 0.6870
No log 10.0 150 0.4345 0.6402 0.4345 0.6592
No log 10.1333 152 0.4201 0.6530 0.4201 0.6481
No log 10.2667 154 0.4693 0.6518 0.4693 0.6851
No log 10.4 156 0.4885 0.6776 0.4885 0.6989
No log 10.5333 158 0.4255 0.5874 0.4255 0.6523
No log 10.6667 160 0.4609 0.5933 0.4609 0.6789
No log 10.8 162 0.5106 0.5794 0.5106 0.7146
No log 10.9333 164 0.4803 0.5983 0.4803 0.6930
No log 11.0667 166 0.4433 0.6096 0.4433 0.6658
No log 11.2 168 0.4615 0.6617 0.4615 0.6793
No log 11.3333 170 0.4699 0.6427 0.4699 0.6855
No log 11.4667 172 0.5215 0.6065 0.5215 0.7221
No log 11.6 174 0.6588 0.5529 0.6588 0.8117
No log 11.7333 176 0.6293 0.5455 0.6293 0.7933
No log 11.8667 178 0.4712 0.6257 0.4712 0.6864
No log 12.0 180 0.4503 0.6108 0.4503 0.6710
No log 12.1333 182 0.4459 0.6078 0.4459 0.6677
No log 12.2667 184 0.4488 0.5463 0.4488 0.6700
No log 12.4 186 0.4967 0.5718 0.4967 0.7048
No log 12.5333 188 0.4625 0.5555 0.4625 0.6801
No log 12.6667 190 0.4730 0.6227 0.4730 0.6877
No log 12.8 192 0.6376 0.5649 0.6376 0.7985
No log 12.9333 194 0.6764 0.5337 0.6764 0.8224
No log 13.0667 196 0.5694 0.5803 0.5694 0.7546
No log 13.2 198 0.4467 0.6039 0.4467 0.6684
No log 13.3333 200 0.5406 0.5378 0.5406 0.7353
No log 13.4667 202 0.5622 0.5581 0.5622 0.7498
No log 13.6 204 0.4747 0.5195 0.4747 0.6890
No log 13.7333 206 0.4566 0.5951 0.4566 0.6757
No log 13.8667 208 0.5460 0.6314 0.5460 0.7389
No log 14.0 210 0.5615 0.6322 0.5615 0.7493
No log 14.1333 212 0.4699 0.6702 0.4699 0.6855
No log 14.2667 214 0.4474 0.6515 0.4474 0.6689
No log 14.4 216 0.4885 0.5970 0.4885 0.6990
No log 14.5333 218 0.4522 0.5765 0.4522 0.6724
No log 14.6667 220 0.4369 0.6634 0.4369 0.6610
No log 14.8 222 0.4500 0.6620 0.4500 0.6708
No log 14.9333 224 0.4510 0.6463 0.4510 0.6716
No log 15.0667 226 0.4855 0.6702 0.4855 0.6968
No log 15.2 228 0.5654 0.5696 0.5654 0.7520
No log 15.3333 230 0.6014 0.5373 0.6014 0.7755
No log 15.4667 232 0.5358 0.6104 0.5358 0.7320
No log 15.6 234 0.5137 0.6454 0.5137 0.7167
No log 15.7333 236 0.5531 0.5922 0.5531 0.7437
No log 15.8667 238 0.5553 0.5922 0.5553 0.7452
No log 16.0 240 0.5005 0.6454 0.5005 0.7075
No log 16.1333 242 0.5334 0.5735 0.5334 0.7304
No log 16.2667 244 0.5181 0.6547 0.5181 0.7198
No log 16.4 246 0.4435 0.6692 0.4435 0.6660
No log 16.5333 248 0.4263 0.6310 0.4263 0.6529
No log 16.6667 250 0.4129 0.6310 0.4129 0.6426
No log 16.8 252 0.4111 0.6101 0.4111 0.6412
No log 16.9333 254 0.4474 0.5811 0.4474 0.6689
No log 17.0667 256 0.4625 0.5811 0.4625 0.6801
No log 17.2 258 0.4369 0.5736 0.4369 0.6610
No log 17.3333 260 0.4422 0.5736 0.4422 0.6650
No log 17.4667 262 0.4656 0.6235 0.4656 0.6823
No log 17.6 264 0.5309 0.5975 0.5309 0.7286
No log 17.7333 266 0.5659 0.5749 0.5659 0.7522
No log 17.8667 268 0.4982 0.5512 0.4982 0.7058
No log 18.0 270 0.4494 0.6198 0.4494 0.6704
No log 18.1333 272 0.4691 0.5970 0.4691 0.6849
No log 18.2667 274 0.4571 0.5227 0.4571 0.6761
No log 18.4 276 0.4976 0.5373 0.4976 0.7054
No log 18.5333 278 0.5499 0.5083 0.5499 0.7416
No log 18.6667 280 0.5111 0.5373 0.5111 0.7149
No log 18.8 282 0.4672 0.5939 0.4672 0.6835
No log 18.9333 284 0.4794 0.6314 0.4794 0.6924
No log 19.0667 286 0.4826 0.6148 0.4826 0.6947
No log 19.2 288 0.4843 0.6235 0.4843 0.6959
No log 19.3333 290 0.4938 0.6158 0.4938 0.7027
No log 19.4667 292 0.4814 0.6222 0.4814 0.6938
No log 19.6 294 0.4702 0.6310 0.4702 0.6857
No log 19.7333 296 0.4615 0.5840 0.4615 0.6793
No log 19.8667 298 0.4630 0.5840 0.4630 0.6804
No log 20.0 300 0.4441 0.6105 0.4441 0.6664
No log 20.1333 302 0.4699 0.5495 0.4699 0.6855
No log 20.2667 304 0.4963 0.5587 0.4963 0.7045
No log 20.4 306 0.4825 0.5587 0.4825 0.6946
No log 20.5333 308 0.4461 0.5555 0.4461 0.6679
No log 20.6667 310 0.4547 0.6145 0.4547 0.6743
No log 20.8 312 0.4810 0.5715 0.4810 0.6935
No log 20.9333 314 0.4735 0.6108 0.4735 0.6881
No log 21.0667 316 0.4748 0.6108 0.4748 0.6891
No log 21.2 318 0.4677 0.6185 0.4677 0.6839
No log 21.3333 320 0.4547 0.5765 0.4547 0.6743
No log 21.4667 322 0.4579 0.5781 0.4579 0.6767
No log 21.6 324 0.4719 0.5495 0.4719 0.6869
No log 21.7333 326 0.4665 0.5495 0.4665 0.6830
No log 21.8667 328 0.4543 0.5781 0.4543 0.6740
No log 22.0 330 0.4400 0.6105 0.4400 0.6634
No log 22.1333 332 0.4438 0.6105 0.4438 0.6662
No log 22.2667 334 0.4447 0.6087 0.4447 0.6669
No log 22.4 336 0.4678 0.5781 0.4678 0.6839
No log 22.5333 338 0.4909 0.5718 0.4909 0.7006
No log 22.6667 340 0.4598 0.6292 0.4598 0.6781
No log 22.8 342 0.4685 0.6286 0.4685 0.6845
No log 22.9333 344 0.4735 0.5813 0.4735 0.6881
No log 23.0667 346 0.4652 0.5813 0.4652 0.6821
No log 23.2 348 0.4519 0.6087 0.4519 0.6723
No log 23.3333 350 0.4563 0.5782 0.4563 0.6755
No log 23.4667 352 0.4819 0.6067 0.4819 0.6942
No log 23.6 354 0.4866 0.5965 0.4866 0.6976
No log 23.7333 356 0.4880 0.5836 0.4880 0.6986
No log 23.8667 358 0.4681 0.5189 0.4681 0.6841
No log 24.0 360 0.4724 0.5414 0.4724 0.6873
No log 24.1333 362 0.4882 0.5117 0.4882 0.6987
No log 24.2667 364 0.4807 0.5189 0.4807 0.6933
No log 24.4 366 0.4665 0.5493 0.4665 0.6830
No log 24.5333 368 0.4520 0.6017 0.4520 0.6723
No log 24.6667 370 0.4462 0.6017 0.4462 0.6680
No log 24.8 372 0.4388 0.6344 0.4388 0.6624
No log 24.9333 374 0.4752 0.5666 0.4752 0.6894
No log 25.0667 376 0.5111 0.5373 0.5111 0.7149
No log 25.2 378 0.5472 0.5373 0.5472 0.7397
No log 25.3333 380 0.5187 0.5373 0.5187 0.7202
No log 25.4667 382 0.4485 0.5798 0.4485 0.6697
No log 25.6 384 0.4582 0.6493 0.4582 0.6769
No log 25.7333 386 0.4848 0.5920 0.4848 0.6962
No log 25.8667 388 0.4645 0.6492 0.4645 0.6815
No log 26.0 390 0.4386 0.6076 0.4386 0.6623
No log 26.1333 392 0.4436 0.5386 0.4436 0.6661
No log 26.2667 394 0.4496 0.5493 0.4496 0.6705
No log 26.4 396 0.4475 0.5731 0.4475 0.6689
No log 26.5333 398 0.4597 0.5208 0.4597 0.6780
No log 26.6667 400 0.4501 0.5432 0.4501 0.6709
No log 26.8 402 0.4396 0.5405 0.4396 0.6630
No log 26.9333 404 0.4481 0.6809 0.4481 0.6694
No log 27.0667 406 0.4424 0.7227 0.4424 0.6651
No log 27.2 408 0.4364 0.5798 0.4364 0.6606
No log 27.3333 410 0.4529 0.5432 0.4529 0.6730
No log 27.4667 412 0.4618 0.5495 0.4618 0.6796
No log 27.6 414 0.4660 0.5495 0.4660 0.6827
No log 27.7333 416 0.4666 0.5538 0.4666 0.6831
No log 27.8667 418 0.4647 0.5538 0.4647 0.6817
No log 28.0 420 0.4626 0.5915 0.4626 0.6802
No log 28.1333 422 0.4734 0.5702 0.4734 0.6880
No log 28.2667 424 0.4786 0.5781 0.4786 0.6918
No log 28.4 426 0.4709 0.5765 0.4709 0.6862
No log 28.5333 428 0.4620 0.6156 0.4620 0.6797
No log 28.6667 430 0.4534 0.6171 0.4534 0.6733
No log 28.8 432 0.4473 0.6171 0.4473 0.6688
No log 28.9333 434 0.4526 0.6154 0.4526 0.6728
No log 29.0667 436 0.4628 0.5781 0.4628 0.6803
No log 29.2 438 0.4851 0.5657 0.4851 0.6965
No log 29.3333 440 0.4824 0.5855 0.4824 0.6946
No log 29.4667 442 0.4460 0.5985 0.4460 0.6679
No log 29.6 444 0.4323 0.6068 0.4323 0.6575
No log 29.7333 446 0.4278 0.6344 0.4278 0.6540
No log 29.8667 448 0.4295 0.6636 0.4295 0.6554
No log 30.0 450 0.4296 0.6435 0.4296 0.6554
No log 30.1333 452 0.4288 0.5798 0.4288 0.6548
No log 30.2667 454 0.4482 0.5747 0.4482 0.6695
No log 30.4 456 0.4641 0.5603 0.4641 0.6812
No log 30.5333 458 0.4553 0.5718 0.4553 0.6748
No log 30.6667 460 0.4358 0.6282 0.4358 0.6601
No log 30.8 462 0.4300 0.6068 0.4300 0.6557
No log 30.9333 464 0.4345 0.6154 0.4345 0.6592
No log 31.0667 466 0.4611 0.5510 0.4611 0.6791
No log 31.2 468 0.4791 0.5544 0.4791 0.6922
No log 31.3333 470 0.4682 0.5461 0.4682 0.6843
No log 31.4667 472 0.4591 0.5105 0.4591 0.6776
No log 31.6 474 0.4410 0.6017 0.4410 0.6641
No log 31.7333 476 0.4394 0.6039 0.4394 0.6629
No log 31.8667 478 0.4337 0.5930 0.4337 0.6586
No log 32.0 480 0.4436 0.5587 0.4436 0.6661
No log 32.1333 482 0.4771 0.5528 0.4771 0.6907
No log 32.2667 484 0.4763 0.5438 0.4763 0.6902
No log 32.4 486 0.4421 0.5796 0.4421 0.6649
No log 32.5333 488 0.4352 0.5796 0.4352 0.6597
No log 32.6667 490 0.4164 0.6210 0.4164 0.6453
No log 32.8 492 0.4288 0.5999 0.4288 0.6548
No log 32.9333 494 0.4339 0.5796 0.4339 0.6587
No log 33.0667 496 0.4328 0.5796 0.4328 0.6579
No log 33.2 498 0.4232 0.5781 0.4232 0.6505
0.29 33.3333 500 0.4215 0.5956 0.4215 0.6493
0.29 33.4667 502 0.4267 0.6254 0.4267 0.6532
0.29 33.6 504 0.4266 0.6452 0.4266 0.6531
0.29 33.7333 506 0.4303 0.5414 0.4303 0.6560
0.29 33.8667 508 0.4353 0.5731 0.4353 0.6598
0.29 34.0 510 0.4352 0.5930 0.4352 0.6597
0.29 34.1333 512 0.4369 0.6255 0.4369 0.6610
0.29 34.2667 514 0.4385 0.6255 0.4385 0.6622
0.29 34.4 516 0.4536 0.5587 0.4536 0.6735
0.29 34.5333 518 0.4929 0.5438 0.4929 0.7021
0.29 34.6667 520 0.5030 0.5438 0.5030 0.7092
0.29 34.8 522 0.4726 0.6034 0.4726 0.6875
0.29 34.9333 524 0.4586 0.6574 0.4586 0.6772
0.29 35.0667 526 0.4664 0.6797 0.4664 0.6830
0.29 35.2 528 0.4724 0.6777 0.4724 0.6873
0.29 35.3333 530 0.4510 0.6807 0.4510 0.6715
0.29 35.4667 532 0.4452 0.5945 0.4452 0.6672
0.29 35.6 534 0.4794 0.5587 0.4794 0.6924
0.29 35.7333 536 0.4879 0.5512 0.4879 0.6985
0.29 35.8667 538 0.4531 0.5587 0.4531 0.6731
0.29 36.0 540 0.4221 0.6032 0.4221 0.6497
0.29 36.1333 542 0.4296 0.6830 0.4296 0.6555
0.29 36.2667 544 0.4345 0.7002 0.4345 0.6592
0.29 36.4 546 0.4263 0.6830 0.4263 0.6529
0.29 36.5333 548 0.4249 0.6032 0.4249 0.6519
0.29 36.6667 550 0.4557 0.5587 0.4557 0.6750
0.29 36.8 552 0.4709 0.5373 0.4709 0.6863
0.29 36.9333 554 0.4532 0.6114 0.4532 0.6732
0.29 37.0667 556 0.4401 0.6007 0.4401 0.6634
0.29 37.2 558 0.4380 0.6723 0.4380 0.6618
0.29 37.3333 560 0.4460 0.7126 0.4460 0.6679
0.29 37.4667 562 0.4373 0.7126 0.4373 0.6613
0.29 37.6 564 0.4315 0.6469 0.4315 0.6569
0.29 37.7333 566 0.4315 0.6289 0.4315 0.6569
0.29 37.8667 568 0.4351 0.6087 0.4351 0.6596
0.29 38.0 570 0.4470 0.5611 0.4470 0.6685
0.29 38.1333 572 0.4561 0.5301 0.4561 0.6754
0.29 38.2667 574 0.4489 0.5283 0.4489 0.6700
0.29 38.4 576 0.4423 0.5714 0.4423 0.6650
0.29 38.5333 578 0.4413 0.5600 0.4413 0.6643

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k3_task7_organization

Finetuned
(4019)
this model