ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6529
  • Qwk: 0.4704
  • Mse: 0.6529
  • Rmse: 0.8080

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1667 2 2.4203 -0.0136 2.4203 1.5557
No log 0.3333 4 1.2344 0.0487 1.2344 1.1110
No log 0.5 6 0.8379 0.0937 0.8379 0.9154
No log 0.6667 8 0.9440 -0.0511 0.9440 0.9716
No log 0.8333 10 0.9495 0.1869 0.9495 0.9744
No log 1.0 12 0.9860 0.2504 0.9860 0.9930
No log 1.1667 14 0.9714 0.2504 0.9714 0.9856
No log 1.3333 16 1.0289 0.1795 1.0289 1.0143
No log 1.5 18 0.9392 0.3110 0.9392 0.9691
No log 1.6667 20 0.9112 0.1918 0.9112 0.9546
No log 1.8333 22 0.8928 -0.0483 0.8928 0.9449
No log 2.0 24 1.0556 -0.0785 1.0556 1.0274
No log 2.1667 26 0.9367 0.0053 0.9367 0.9678
No log 2.3333 28 0.8338 0.0053 0.8338 0.9131
No log 2.5 30 0.8608 0.0053 0.8608 0.9278
No log 2.6667 32 0.8317 0.0 0.8317 0.9120
No log 2.8333 34 0.7743 0.0 0.7743 0.8799
No log 3.0 36 0.7558 0.0 0.7558 0.8694
No log 3.1667 38 0.7315 0.1007 0.7315 0.8552
No log 3.3333 40 0.7671 0.2817 0.7671 0.8758
No log 3.5 42 0.7384 0.2085 0.7384 0.8593
No log 3.6667 44 0.7325 0.0840 0.7325 0.8559
No log 3.8333 46 0.7585 0.0840 0.7585 0.8709
No log 4.0 48 0.7519 0.0840 0.7519 0.8671
No log 4.1667 50 0.7033 0.0428 0.7033 0.8386
No log 4.3333 52 0.6914 0.0428 0.6914 0.8315
No log 4.5 54 0.6767 0.1327 0.6767 0.8226
No log 4.6667 56 0.6413 0.1604 0.6413 0.8008
No log 4.8333 58 0.6414 0.2002 0.6414 0.8008
No log 5.0 60 0.6290 0.2374 0.6290 0.7931
No log 5.1667 62 0.6587 0.1646 0.6587 0.8116
No log 5.3333 64 0.6825 0.1702 0.6825 0.8261
No log 5.5 66 0.6464 0.1702 0.6464 0.8040
No log 5.6667 68 0.5854 0.3745 0.5854 0.7651
No log 5.8333 70 0.6351 0.3032 0.6351 0.7969
No log 6.0 72 0.6510 0.3302 0.6510 0.8069
No log 6.1667 74 0.5977 0.3267 0.5977 0.7731
No log 6.3333 76 0.5839 0.4034 0.5839 0.7642
No log 6.5 78 0.5717 0.4634 0.5717 0.7561
No log 6.6667 80 0.5667 0.4354 0.5667 0.7528
No log 6.8333 82 0.5616 0.4908 0.5616 0.7494
No log 7.0 84 0.5783 0.5405 0.5783 0.7604
No log 7.1667 86 0.5706 0.5728 0.5706 0.7554
No log 7.3333 88 0.5650 0.5521 0.5650 0.7517
No log 7.5 90 0.5678 0.5095 0.5678 0.7535
No log 7.6667 92 0.5805 0.5095 0.5805 0.7619
No log 7.8333 94 0.5936 0.6001 0.5936 0.7704
No log 8.0 96 0.6588 0.4329 0.6588 0.8116
No log 8.1667 98 0.5898 0.5763 0.5898 0.7680
No log 8.3333 100 0.5752 0.4837 0.5752 0.7584
No log 8.5 102 0.6297 0.4134 0.6297 0.7935
No log 8.6667 104 0.6748 0.3544 0.6748 0.8214
No log 8.8333 106 0.6565 0.4007 0.6565 0.8103
No log 9.0 108 0.6681 0.4007 0.6681 0.8174
No log 9.1667 110 0.6724 0.4642 0.6724 0.8200
No log 9.3333 112 0.6656 0.5272 0.6656 0.8159
No log 9.5 114 0.6449 0.5289 0.6449 0.8031
No log 9.6667 116 0.6606 0.4663 0.6606 0.8128
No log 9.8333 118 0.6667 0.4663 0.6667 0.8165
No log 10.0 120 0.6230 0.5322 0.6230 0.7893
No log 10.1667 122 0.6256 0.5125 0.6256 0.7909
No log 10.3333 124 0.6397 0.5125 0.6397 0.7998
No log 10.5 126 0.6115 0.5753 0.6115 0.7820
No log 10.6667 128 0.6398 0.4663 0.6398 0.7999
No log 10.8333 130 0.6148 0.4872 0.6148 0.7841
No log 11.0 132 0.6157 0.4937 0.6157 0.7847
No log 11.1667 134 0.6388 0.4800 0.6388 0.7992
No log 11.3333 136 0.5981 0.4582 0.5981 0.7734
No log 11.5 138 0.6746 0.4447 0.6746 0.8213
No log 11.6667 140 0.6052 0.4740 0.6052 0.7779
No log 11.8333 142 0.5820 0.5714 0.5820 0.7629
No log 12.0 144 0.5936 0.4958 0.5936 0.7704
No log 12.1667 146 0.5709 0.5508 0.5709 0.7556
No log 12.3333 148 0.5671 0.5003 0.5671 0.7530
No log 12.5 150 0.6155 0.4441 0.6155 0.7845
No log 12.6667 152 0.6557 0.5126 0.6557 0.8097
No log 12.8333 154 0.5576 0.5324 0.5576 0.7468
No log 13.0 156 0.5220 0.5874 0.5220 0.7225
No log 13.1667 158 0.5374 0.5812 0.5374 0.7331
No log 13.3333 160 0.5203 0.6170 0.5203 0.7213
No log 13.5 162 0.5851 0.5521 0.5851 0.7649
No log 13.6667 164 0.7356 0.4577 0.7356 0.8577
No log 13.8333 166 0.6129 0.5508 0.6129 0.7829
No log 14.0 168 0.5204 0.5897 0.5204 0.7214
No log 14.1667 170 0.5271 0.5923 0.5271 0.7260
No log 14.3333 172 0.6438 0.4888 0.6438 0.8024
No log 14.5 174 0.6971 0.5043 0.6971 0.8349
No log 14.6667 176 0.6157 0.5670 0.6157 0.7847
No log 14.8333 178 0.5806 0.5813 0.5806 0.7620
No log 15.0 180 0.6069 0.5304 0.6069 0.7790
No log 15.1667 182 0.6538 0.5406 0.6538 0.8086
No log 15.3333 184 0.6308 0.5286 0.6308 0.7942
No log 15.5 186 0.6663 0.5437 0.6663 0.8163
No log 15.6667 188 0.7702 0.4474 0.7702 0.8776
No log 15.8333 190 0.7726 0.4536 0.7726 0.8790
No log 16.0 192 0.6429 0.5552 0.6429 0.8018
No log 16.1667 194 0.6160 0.4955 0.6160 0.7848
No log 16.3333 196 0.6187 0.5357 0.6187 0.7866
No log 16.5 198 0.6124 0.5021 0.6124 0.7826
No log 16.6667 200 0.6219 0.5003 0.6219 0.7886
No log 16.8333 202 0.6503 0.5324 0.6503 0.8064
No log 17.0 204 0.7593 0.4088 0.7593 0.8714
No log 17.1667 206 0.7886 0.3867 0.7886 0.8880
No log 17.3333 208 0.6724 0.4933 0.6724 0.8200
No log 17.5 210 0.6308 0.5488 0.6308 0.7942
No log 17.6667 212 0.6037 0.5042 0.6037 0.7770
No log 17.8333 214 0.6017 0.4632 0.6017 0.7757
No log 18.0 216 0.5705 0.5675 0.5705 0.7553
No log 18.1667 218 0.5613 0.4681 0.5613 0.7492
No log 18.3333 220 0.5654 0.4504 0.5654 0.7519
No log 18.5 222 0.5782 0.4504 0.5782 0.7604
No log 18.6667 224 0.5867 0.4229 0.5867 0.7660
No log 18.8333 226 0.5835 0.4229 0.5835 0.7639
No log 19.0 228 0.6416 0.5034 0.6416 0.8010
No log 19.1667 230 0.6759 0.4408 0.6759 0.8221
No log 19.3333 232 0.6189 0.4337 0.6189 0.7867
No log 19.5 234 0.5750 0.5133 0.5750 0.7583
No log 19.6667 236 0.5832 0.5304 0.5832 0.7637
No log 19.8333 238 0.5879 0.5213 0.5879 0.7667
No log 20.0 240 0.6007 0.5238 0.6007 0.7751
No log 20.1667 242 0.6105 0.5238 0.6105 0.7813
No log 20.3333 244 0.6301 0.5397 0.6301 0.7938
No log 20.5 246 0.6367 0.5397 0.6367 0.7979
No log 20.6667 248 0.6687 0.4997 0.6687 0.8177
No log 20.8333 250 0.6851 0.4464 0.6851 0.8277
No log 21.0 252 0.6685 0.4353 0.6685 0.8176
No log 21.1667 254 0.6941 0.3781 0.6941 0.8331
No log 21.3333 256 0.6742 0.4165 0.6742 0.8211
No log 21.5 258 0.6908 0.4093 0.6908 0.8311
No log 21.6667 260 0.6251 0.4408 0.6251 0.7906
No log 21.8333 262 0.5899 0.5703 0.5899 0.7681
No log 22.0 264 0.5899 0.5272 0.5899 0.7681
No log 22.1667 266 0.6273 0.5088 0.6274 0.7921
No log 22.3333 268 0.6849 0.5176 0.6849 0.8276
No log 22.5 270 0.6397 0.5290 0.6397 0.7998
No log 22.6667 272 0.6069 0.5574 0.6069 0.7790
No log 22.8333 274 0.6192 0.5475 0.6192 0.7869
No log 23.0 276 0.6183 0.4838 0.6183 0.7863
No log 23.1667 278 0.5932 0.5424 0.5932 0.7702
No log 23.3333 280 0.6161 0.4437 0.6161 0.7849
No log 23.5 282 0.7169 0.3824 0.7169 0.8467
No log 23.6667 284 0.6986 0.3891 0.6986 0.8358
No log 23.8333 286 0.7040 0.3824 0.7040 0.8391
No log 24.0 288 0.6610 0.4112 0.6610 0.8130
No log 24.1667 290 0.5977 0.4182 0.5977 0.7731
No log 24.3333 292 0.5744 0.4888 0.5744 0.7579
No log 24.5 294 0.5765 0.4459 0.5765 0.7593
No log 24.6667 296 0.5997 0.4375 0.5997 0.7744
No log 24.8333 298 0.6587 0.3667 0.6587 0.8116
No log 25.0 300 0.7098 0.3886 0.7098 0.8425
No log 25.1667 302 0.6703 0.4664 0.6703 0.8187
No log 25.3333 304 0.5882 0.5404 0.5882 0.7670
No log 25.5 306 0.5798 0.6154 0.5798 0.7614
No log 25.6667 308 0.5851 0.6068 0.5851 0.7649
No log 25.8333 310 0.5779 0.6344 0.5779 0.7602
No log 26.0 312 0.5783 0.6344 0.5783 0.7605
No log 26.1667 314 0.6040 0.4885 0.6040 0.7772
No log 26.3333 316 0.6169 0.4681 0.6169 0.7854
No log 26.5 318 0.6359 0.4681 0.6359 0.7975
No log 26.6667 320 0.6198 0.4496 0.6198 0.7873
No log 26.8333 322 0.6118 0.5899 0.6118 0.7821
No log 27.0 324 0.6499 0.5361 0.6499 0.8062
No log 27.1667 326 0.6464 0.5442 0.6464 0.8040
No log 27.3333 328 0.6106 0.5869 0.6106 0.7814
No log 27.5 330 0.6821 0.4568 0.6821 0.8259
No log 27.6667 332 0.8108 0.3542 0.8108 0.9005
No log 27.8333 334 0.8054 0.3747 0.8054 0.8975
No log 28.0 336 0.6948 0.4353 0.6948 0.8335
No log 28.1667 338 0.6089 0.4743 0.6089 0.7803
No log 28.3333 340 0.6055 0.5272 0.6055 0.7781
No log 28.5 342 0.6230 0.5357 0.6230 0.7893
No log 28.6667 344 0.6474 0.5133 0.6474 0.8046
No log 28.8333 346 0.6430 0.5133 0.6430 0.8019
No log 29.0 348 0.5983 0.5428 0.5983 0.7735
No log 29.1667 350 0.5846 0.5505 0.5846 0.7646
No log 29.3333 352 0.5876 0.4990 0.5876 0.7666
No log 29.5 354 0.5845 0.5556 0.5845 0.7645
No log 29.6667 356 0.5886 0.5306 0.5886 0.7672
No log 29.8333 358 0.6405 0.4239 0.6405 0.8003
No log 30.0 360 0.7283 0.3781 0.7283 0.8534
No log 30.1667 362 0.7223 0.3710 0.7223 0.8499
No log 30.3333 364 0.6465 0.4622 0.6465 0.8040
No log 30.5 366 0.6068 0.4289 0.6068 0.7790
No log 30.6667 368 0.6068 0.4364 0.6068 0.7790
No log 30.8333 370 0.6146 0.4654 0.6146 0.7840
No log 31.0 372 0.5963 0.4912 0.5963 0.7722
No log 31.1667 374 0.5820 0.5208 0.5820 0.7629
No log 31.3333 376 0.5953 0.5133 0.5953 0.7716
No log 31.5 378 0.5881 0.5357 0.5881 0.7669
No log 31.6667 380 0.5824 0.5305 0.5824 0.7632
No log 31.8333 382 0.5751 0.5305 0.5751 0.7584
No log 32.0 384 0.5720 0.5367 0.5720 0.7563
No log 32.1667 386 0.6004 0.5034 0.6004 0.7748
No log 32.3333 388 0.6661 0.4550 0.6661 0.8161
No log 32.5 390 0.6967 0.3699 0.6967 0.8347
No log 32.6667 392 0.6907 0.3822 0.6907 0.8311
No log 32.8333 394 0.6197 0.4704 0.6197 0.7872
No log 33.0 396 0.5546 0.4637 0.5546 0.7447
No log 33.1667 398 0.5455 0.4631 0.5455 0.7386
No log 33.3333 400 0.5602 0.5361 0.5602 0.7485
No log 33.5 402 0.5462 0.5246 0.5462 0.7391
No log 33.6667 404 0.5468 0.5912 0.5468 0.7395
No log 33.8333 406 0.5695 0.5411 0.5695 0.7547
No log 34.0 408 0.6029 0.5744 0.6029 0.7765
No log 34.1667 410 0.6128 0.5149 0.6128 0.7828
No log 34.3333 412 0.6026 0.4929 0.6026 0.7763
No log 34.5 414 0.5864 0.4788 0.5864 0.7657
No log 34.6667 416 0.5990 0.5016 0.5990 0.7740
No log 34.8333 418 0.6051 0.4724 0.6051 0.7779
No log 35.0 420 0.6226 0.4782 0.6226 0.7891
No log 35.1667 422 0.6119 0.5254 0.6119 0.7822
No log 35.3333 424 0.5951 0.4681 0.5951 0.7714
No log 35.5 426 0.5932 0.5089 0.5932 0.7702
No log 35.6667 428 0.6058 0.5517 0.6058 0.7783
No log 35.8333 430 0.6413 0.4933 0.6413 0.8008
No log 36.0 432 0.6403 0.5157 0.6403 0.8002
No log 36.1667 434 0.6021 0.5723 0.6021 0.7760
No log 36.3333 436 0.5837 0.5344 0.5837 0.7640
No log 36.5 438 0.5793 0.5344 0.5793 0.7611
No log 36.6667 440 0.5772 0.5250 0.5772 0.7598
No log 36.8333 442 0.5834 0.5923 0.5834 0.7638
No log 37.0 444 0.5762 0.5897 0.5762 0.7591
No log 37.1667 446 0.5670 0.5472 0.5670 0.7530
No log 37.3333 448 0.5691 0.5897 0.5691 0.7544
No log 37.5 450 0.5820 0.5801 0.5820 0.7629
No log 37.6667 452 0.6193 0.5184 0.6193 0.7870
No log 37.8333 454 0.6322 0.5098 0.6322 0.7951
No log 38.0 456 0.6044 0.5411 0.6044 0.7774
No log 38.1667 458 0.5757 0.5022 0.5757 0.7587
No log 38.3333 460 0.5808 0.4724 0.5808 0.7621
No log 38.5 462 0.5802 0.4724 0.5802 0.7617
No log 38.6667 464 0.5858 0.5801 0.5858 0.7654
No log 38.8333 466 0.6192 0.4704 0.6192 0.7869
No log 39.0 468 0.6463 0.4315 0.6463 0.8039
No log 39.1667 470 0.6368 0.4704 0.6368 0.7980
No log 39.3333 472 0.6240 0.4704 0.6240 0.7900
No log 39.5 474 0.6091 0.4576 0.6091 0.7804
No log 39.6667 476 0.6064 0.4828 0.6064 0.7787
No log 39.8333 478 0.6167 0.5184 0.6167 0.7853
No log 40.0 480 0.6153 0.5201 0.6153 0.7844
No log 40.1667 482 0.6348 0.4872 0.6348 0.7967
No log 40.3333 484 0.6714 0.4664 0.6714 0.8194
No log 40.5 486 0.6660 0.4664 0.6660 0.8161
No log 40.6667 488 0.6458 0.4872 0.6458 0.8036
No log 40.8333 490 0.6090 0.4801 0.6090 0.7804
No log 41.0 492 0.5775 0.5201 0.5775 0.7599
No log 41.1667 494 0.5680 0.5379 0.5680 0.7536
No log 41.3333 496 0.5695 0.5379 0.5695 0.7547
No log 41.5 498 0.5691 0.5379 0.5691 0.7544
0.2872 41.6667 500 0.5837 0.5184 0.5837 0.7640
0.2872 41.8333 502 0.6250 0.4864 0.6250 0.7906
0.2872 42.0 504 0.6430 0.4777 0.6430 0.8019
0.2872 42.1667 506 0.6182 0.5157 0.6182 0.7863
0.2872 42.3333 508 0.5884 0.5801 0.5884 0.7671
0.2872 42.5 510 0.5779 0.5472 0.5779 0.7602
0.2872 42.6667 512 0.5846 0.5517 0.5846 0.7646
0.2872 42.8333 514 0.6017 0.5923 0.6017 0.7757
0.2872 43.0 516 0.6368 0.5157 0.6368 0.7980
0.2872 43.1667 518 0.6495 0.4335 0.6495 0.8059
0.2872 43.3333 520 0.6358 0.4933 0.6358 0.7974
0.2872 43.5 522 0.6508 0.4144 0.6508 0.8068
0.2872 43.6667 524 0.6351 0.5239 0.6351 0.7970
0.2872 43.8333 526 0.6068 0.5428 0.6068 0.7790
0.2872 44.0 528 0.5821 0.4701 0.5821 0.7630
0.2872 44.1667 530 0.5801 0.5114 0.5801 0.7617
0.2872 44.3333 532 0.5952 0.5159 0.5952 0.7715
0.2872 44.5 534 0.6320 0.4704 0.6320 0.7950
0.2872 44.6667 536 0.6529 0.4704 0.6529 0.8080

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task7_organization

Finetuned
(4019)
this model