ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6059
  • Qwk: 0.4659
  • Mse: 0.6059
  • Rmse: 0.7784

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0571 2 2.5203 -0.0109 2.5203 1.5875
No log 0.1143 4 1.2409 0.0991 1.2409 1.1140
No log 0.1714 6 1.0130 -0.0970 1.0130 1.0065
No log 0.2286 8 1.0163 -0.1163 1.0163 1.0081
No log 0.2857 10 1.0262 -0.0269 1.0262 1.0130
No log 0.3429 12 1.0610 -0.1278 1.0610 1.0301
No log 0.4 14 1.0005 -0.0907 1.0005 1.0003
No log 0.4571 16 0.8764 0.0 0.8764 0.9361
No log 0.5143 18 0.8490 0.0 0.8490 0.9214
No log 0.5714 20 0.8951 0.0 0.8951 0.9461
No log 0.6286 22 0.9469 -0.0426 0.9469 0.9731
No log 0.6857 24 1.0005 0.0547 1.0005 1.0003
No log 0.7429 26 0.9631 -0.0228 0.9631 0.9814
No log 0.8 28 0.8388 0.0053 0.8388 0.9158
No log 0.8571 30 0.7682 0.1983 0.7682 0.8764
No log 0.9143 32 0.8276 0.0688 0.8276 0.9097
No log 0.9714 34 0.8928 0.0460 0.8928 0.9449
No log 1.0286 36 0.8261 0.0889 0.8261 0.9089
No log 1.0857 38 0.7507 0.0717 0.7507 0.8664
No log 1.1429 40 0.7058 0.0327 0.7058 0.8401
No log 1.2 42 0.7488 0.0889 0.7488 0.8653
No log 1.2571 44 0.7746 0.0846 0.7746 0.8801
No log 1.3143 46 0.7402 0.1139 0.7402 0.8603
No log 1.3714 48 0.7431 0.1660 0.7431 0.8620
No log 1.4286 50 0.7937 0.1327 0.7937 0.8909
No log 1.4857 52 0.8251 0.0944 0.8251 0.9083
No log 1.5429 54 0.9484 0.0600 0.9484 0.9739
No log 1.6 56 0.9898 -0.0076 0.9898 0.9949
No log 1.6571 58 0.9457 -0.0148 0.9457 0.9725
No log 1.7143 60 0.8791 0.2109 0.8791 0.9376
No log 1.7714 62 0.8863 0.2408 0.8863 0.9414
No log 1.8286 64 0.8285 0.1739 0.8285 0.9102
No log 1.8857 66 0.7029 0.2087 0.7029 0.8384
No log 1.9429 68 0.6871 0.1903 0.6871 0.8289
No log 2.0 70 0.6790 0.1903 0.6790 0.8240
No log 2.0571 72 0.6874 0.1903 0.6874 0.8291
No log 2.1143 74 0.6966 0.1903 0.6966 0.8346
No log 2.1714 76 0.7093 0.1903 0.7093 0.8422
No log 2.2286 78 0.7715 0.2205 0.7715 0.8783
No log 2.2857 80 0.9874 0.1609 0.9874 0.9937
No log 2.3429 82 1.0042 0.1843 1.0042 1.0021
No log 2.4 84 1.0392 0.2316 1.0392 1.0194
No log 2.4571 86 0.8782 0.2942 0.8782 0.9371
No log 2.5143 88 0.7224 0.3745 0.7224 0.8499
No log 2.5714 90 0.7118 0.3072 0.7118 0.8437
No log 2.6286 92 0.7171 0.3039 0.7171 0.8468
No log 2.6857 94 0.6861 0.2930 0.6861 0.8283
No log 2.7429 96 0.6773 0.2930 0.6773 0.8230
No log 2.8 98 0.6664 0.3243 0.6664 0.8163
No log 2.8571 100 0.6230 0.3039 0.6230 0.7893
No log 2.9143 102 0.6526 0.3477 0.6526 0.8079
No log 2.9714 104 0.6853 0.4259 0.6853 0.8279
No log 3.0286 106 0.7183 0.3754 0.7183 0.8475
No log 3.0857 108 0.6427 0.3729 0.6427 0.8017
No log 3.1429 110 0.6464 0.3728 0.6464 0.8040
No log 3.2 112 0.7152 0.3996 0.7152 0.8457
No log 3.2571 114 0.6876 0.3594 0.6876 0.8292
No log 3.3143 116 0.6660 0.1923 0.6660 0.8161
No log 3.3714 118 0.6675 0.2488 0.6675 0.8170
No log 3.4286 120 0.6188 0.3225 0.6188 0.7866
No log 3.4857 122 0.6835 0.4602 0.6835 0.8268
No log 3.5429 124 0.8915 0.3790 0.8915 0.9442
No log 3.6 126 0.9186 0.3946 0.9186 0.9584
No log 3.6571 128 0.7155 0.3452 0.7155 0.8459
No log 3.7143 130 0.7285 0.2998 0.7285 0.8535
No log 3.7714 132 0.8036 0.3227 0.8036 0.8965
No log 3.8286 134 0.7582 0.2652 0.7582 0.8708
No log 3.8857 136 0.8586 0.3137 0.8586 0.9266
No log 3.9429 138 0.7542 0.2926 0.7542 0.8684
No log 4.0 140 0.6479 0.3478 0.6479 0.8049
No log 4.0571 142 0.7575 0.4014 0.7575 0.8703
No log 4.1143 144 0.8712 0.3822 0.8712 0.9334
No log 4.1714 146 0.7859 0.3891 0.7859 0.8865
No log 4.2286 148 0.7016 0.3530 0.7016 0.8376
No log 4.2857 150 0.8330 0.3686 0.8330 0.9127
No log 4.3429 152 0.9218 0.2971 0.9218 0.9601
No log 4.4 154 0.8066 0.3686 0.8066 0.8981
No log 4.4571 156 0.7033 0.3902 0.7033 0.8386
No log 4.5143 158 0.6925 0.4514 0.6925 0.8321
No log 4.5714 160 0.7586 0.4486 0.7586 0.8710
No log 4.6286 162 0.7251 0.4631 0.7251 0.8515
No log 4.6857 164 0.6368 0.3340 0.6368 0.7980
No log 4.7429 166 0.6251 0.2878 0.6251 0.7907
No log 4.8 168 0.6148 0.3426 0.6148 0.7841
No log 4.8571 170 0.6351 0.4163 0.6351 0.7969
No log 4.9143 172 0.7486 0.4993 0.7486 0.8652
No log 4.9714 174 0.7301 0.5632 0.7301 0.8544
No log 5.0286 176 0.6070 0.5786 0.6070 0.7791
No log 5.0857 178 0.5813 0.5238 0.5813 0.7624
No log 5.1429 180 0.5910 0.48 0.5910 0.7687
No log 5.2 182 0.6047 0.4568 0.6047 0.7776
No log 5.2571 184 0.5928 0.5020 0.5928 0.7699
No log 5.3143 186 0.5661 0.5133 0.5661 0.7524
No log 5.3714 188 0.5878 0.5081 0.5878 0.7667
No log 5.4286 190 0.6089 0.4703 0.6089 0.7803
No log 5.4857 192 0.5843 0.4661 0.5843 0.7644
No log 5.5429 194 0.6047 0.4830 0.6047 0.7776
No log 5.6 196 0.6013 0.5142 0.6013 0.7754
No log 5.6571 198 0.6052 0.4659 0.6052 0.7779
No log 5.7143 200 0.6348 0.4845 0.6348 0.7967
No log 5.7714 202 0.6500 0.4916 0.6500 0.8062
No log 5.8286 204 0.6249 0.4393 0.6249 0.7905
No log 5.8857 206 0.6087 0.4444 0.6087 0.7802
No log 5.9429 208 0.6083 0.4361 0.6083 0.7799
No log 6.0 210 0.6326 0.3320 0.6326 0.7953
No log 6.0571 212 0.6706 0.4091 0.6706 0.8189
No log 6.1143 214 0.6637 0.3747 0.6637 0.8147
No log 6.1714 216 0.6168 0.4114 0.6168 0.7853
No log 6.2286 218 0.6002 0.3809 0.6002 0.7747
No log 6.2857 220 0.6131 0.3945 0.6131 0.7830
No log 6.3429 222 0.6488 0.4158 0.6488 0.8055
No log 6.4 224 0.6282 0.3894 0.6282 0.7926
No log 6.4571 226 0.5892 0.3837 0.5892 0.7676
No log 6.5143 228 0.5778 0.3995 0.5778 0.7601
No log 6.5714 230 0.5706 0.4322 0.5706 0.7554
No log 6.6286 232 0.5883 0.4128 0.5883 0.7670
No log 6.6857 234 0.5904 0.4348 0.5904 0.7684
No log 6.7429 236 0.6196 0.4486 0.6196 0.7872
No log 6.8 238 0.6330 0.4907 0.6330 0.7956
No log 6.8571 240 0.6151 0.5167 0.6151 0.7843
No log 6.9143 242 0.6126 0.4581 0.6126 0.7827
No log 6.9714 244 0.7749 0.4222 0.7749 0.8803
No log 7.0286 246 0.8190 0.4177 0.8190 0.9050
No log 7.0857 248 0.6720 0.4761 0.6720 0.8197
No log 7.1429 250 0.5928 0.4820 0.5928 0.7700
No log 7.2 252 0.6922 0.5293 0.6922 0.8320
No log 7.2571 254 0.6849 0.5293 0.6849 0.8276
No log 7.3143 256 0.6302 0.5032 0.6302 0.7938
No log 7.3714 258 0.5889 0.4719 0.5889 0.7674
No log 7.4286 260 0.5888 0.4801 0.5888 0.7673
No log 7.4857 262 0.6048 0.4719 0.6048 0.7777
No log 7.5429 264 0.6089 0.4619 0.6089 0.7803
No log 7.6 266 0.6409 0.4731 0.6409 0.8006
No log 7.6571 268 0.6391 0.4696 0.6391 0.7995
No log 7.7143 270 0.7468 0.5185 0.7468 0.8642
No log 7.7714 272 0.8344 0.4752 0.8344 0.9135
No log 7.8286 274 0.8087 0.4334 0.8087 0.8993
No log 7.8857 276 0.6803 0.4582 0.6803 0.8248
No log 7.9429 278 0.6555 0.5111 0.6555 0.8096
No log 8.0 280 0.7005 0.5444 0.7005 0.8369
No log 8.0571 282 0.7615 0.4785 0.7615 0.8726
No log 8.1143 284 0.8201 0.5377 0.8201 0.9056
No log 8.1714 286 0.8075 0.4888 0.8075 0.8986
No log 8.2286 288 0.7282 0.4966 0.7282 0.8533
No log 8.2857 290 0.6085 0.4789 0.6085 0.7801
No log 8.3429 292 0.6071 0.4599 0.6071 0.7792
No log 8.4 294 0.6192 0.4189 0.6192 0.7869
No log 8.4571 296 0.5783 0.4992 0.5783 0.7605
No log 8.5143 298 0.6210 0.5342 0.6210 0.7881
No log 8.5714 300 0.6881 0.5008 0.6881 0.8295
No log 8.6286 302 0.6487 0.5495 0.6487 0.8054
No log 8.6857 304 0.6041 0.5363 0.6041 0.7772
No log 8.7429 306 0.6430 0.4895 0.6430 0.8019
No log 8.8 308 0.6354 0.4833 0.6354 0.7971
No log 8.8571 310 0.6063 0.5190 0.6063 0.7786
No log 8.9143 312 0.6187 0.5262 0.6187 0.7866
No log 8.9714 314 0.5996 0.5256 0.5996 0.7744
No log 9.0286 316 0.5794 0.4562 0.5794 0.7612
No log 9.0857 318 0.5861 0.5034 0.5861 0.7656
No log 9.1429 320 0.6355 0.5495 0.6355 0.7972
No log 9.2 322 0.6312 0.5495 0.6312 0.7945
No log 9.2571 324 0.5912 0.5617 0.5912 0.7689
No log 9.3143 326 0.5684 0.5254 0.5684 0.7539
No log 9.3714 328 0.5705 0.4789 0.5705 0.7553
No log 9.4286 330 0.5767 0.4932 0.5767 0.7594
No log 9.4857 332 0.5892 0.5640 0.5892 0.7676
No log 9.5429 334 0.5989 0.5899 0.5989 0.7739
No log 9.6 336 0.5944 0.5725 0.5944 0.7709
No log 9.6571 338 0.5877 0.5067 0.5877 0.7666
No log 9.7143 340 0.5802 0.4731 0.5802 0.7617
No log 9.7714 342 0.5681 0.4789 0.5681 0.7537
No log 9.8286 344 0.5658 0.5662 0.5658 0.7522
No log 9.8857 346 0.6083 0.4782 0.6083 0.7799
No log 9.9429 348 0.6008 0.4951 0.6008 0.7751
No log 10.0 350 0.5851 0.5593 0.5851 0.7649
No log 10.0571 352 0.5835 0.5444 0.5835 0.7639
No log 10.1143 354 0.5994 0.4581 0.5994 0.7742
No log 10.1714 356 0.6068 0.4806 0.6068 0.7789
No log 10.2286 358 0.5941 0.5035 0.5941 0.7708
No log 10.2857 360 0.5973 0.5019 0.5973 0.7728
No log 10.3429 362 0.5904 0.5225 0.5904 0.7684
No log 10.4 364 0.5711 0.5238 0.5711 0.7557
No log 10.4571 366 0.5615 0.4771 0.5615 0.7493
No log 10.5143 368 0.5823 0.4406 0.5823 0.7631
No log 10.5714 370 0.6248 0.5230 0.6248 0.7904
No log 10.6286 372 0.6200 0.5352 0.6200 0.7874
No log 10.6857 374 0.5753 0.5092 0.5753 0.7585
No log 10.7429 376 0.5779 0.5419 0.5779 0.7602
No log 10.8 378 0.6176 0.4997 0.6176 0.7859
No log 10.8571 380 0.6007 0.4879 0.6007 0.7751
No log 10.9143 382 0.5877 0.5353 0.5877 0.7666
No log 10.9714 384 0.5955 0.5008 0.5955 0.7717
No log 11.0286 386 0.6271 0.5058 0.6271 0.7919
No log 11.0857 388 0.6051 0.4675 0.6051 0.7779
No log 11.1429 390 0.5838 0.4743 0.5838 0.7640
No log 11.2 392 0.5984 0.4555 0.5984 0.7736
No log 11.2571 394 0.6147 0.5081 0.6147 0.7841
No log 11.3143 396 0.5995 0.5019 0.5995 0.7742
No log 11.3714 398 0.6006 0.5476 0.6006 0.7750
No log 11.4286 400 0.5938 0.5174 0.5938 0.7706
No log 11.4857 402 0.5870 0.5238 0.5870 0.7661
No log 11.5429 404 0.5874 0.5067 0.5874 0.7664
No log 11.6 406 0.5847 0.5084 0.5847 0.7647
No log 11.6571 408 0.5792 0.4856 0.5792 0.7611
No log 11.7143 410 0.5692 0.5450 0.5692 0.7545
No log 11.7714 412 0.5825 0.5578 0.5825 0.7632
No log 11.8286 414 0.5749 0.5578 0.5749 0.7582
No log 11.8857 416 0.5496 0.5184 0.5496 0.7413
No log 11.9429 418 0.5617 0.5816 0.5617 0.7494
No log 12.0 420 0.5738 0.5501 0.5738 0.7575
No log 12.0571 422 0.5669 0.4719 0.5669 0.7529
No log 12.1143 424 0.5658 0.5143 0.5658 0.7522
No log 12.1714 426 0.5732 0.5075 0.5732 0.7571
No log 12.2286 428 0.5788 0.5075 0.5788 0.7608
No log 12.2857 430 0.5687 0.5075 0.5687 0.7541
No log 12.3429 432 0.5578 0.5286 0.5578 0.7469
No log 12.4 434 0.5551 0.5707 0.5551 0.7451
No log 12.4571 436 0.5624 0.5352 0.5624 0.7500
No log 12.5143 438 0.5794 0.5368 0.5794 0.7612
No log 12.5714 440 0.5671 0.6034 0.5671 0.7531
No log 12.6286 442 0.6305 0.5323 0.6305 0.7940
No log 12.6857 444 0.7019 0.4805 0.7019 0.8378
No log 12.7429 446 0.6757 0.4812 0.6757 0.8220
No log 12.8 448 0.5860 0.5998 0.5860 0.7655
No log 12.8571 450 0.5626 0.5093 0.5626 0.7501
No log 12.9143 452 0.5661 0.5756 0.5661 0.7524
No log 12.9714 454 0.5748 0.5742 0.5748 0.7581
No log 13.0286 456 0.6085 0.5088 0.6085 0.7801
No log 13.0857 458 0.6186 0.5013 0.6186 0.7865
No log 13.1429 460 0.5877 0.5103 0.5877 0.7666
No log 13.2 462 0.5783 0.5770 0.5783 0.7605
No log 13.2571 464 0.5764 0.4828 0.5764 0.7592
No log 13.3143 466 0.5855 0.5036 0.5855 0.7652
No log 13.3714 468 0.6076 0.4618 0.6076 0.7795
No log 13.4286 470 0.6273 0.4393 0.6273 0.7920
No log 13.4857 472 0.6440 0.4864 0.6440 0.8025
No log 13.5429 474 0.6163 0.3980 0.6163 0.7851
No log 13.6 476 0.5991 0.5619 0.5991 0.7740
No log 13.6571 478 0.6080 0.5286 0.6080 0.7798
No log 13.7143 480 0.6091 0.4472 0.6091 0.7804
No log 13.7714 482 0.6180 0.4434 0.6180 0.7861
No log 13.8286 484 0.7059 0.5179 0.7059 0.8402
No log 13.8857 486 0.7109 0.5179 0.7109 0.8432
No log 13.9429 488 0.6238 0.4281 0.6238 0.7898
No log 14.0 490 0.6041 0.4514 0.6041 0.7772
No log 14.0571 492 0.5966 0.4044 0.5966 0.7724
No log 14.1143 494 0.5962 0.4379 0.5962 0.7721
No log 14.1714 496 0.6233 0.4523 0.6233 0.7895
No log 14.2286 498 0.6524 0.4663 0.6524 0.8077
0.3527 14.2857 500 0.6262 0.4754 0.6262 0.7913
0.3527 14.3429 502 0.5926 0.5465 0.5926 0.7698
0.3527 14.4 504 0.5920 0.5634 0.5920 0.7694
0.3527 14.4571 506 0.5913 0.4813 0.5913 0.7690
0.3527 14.5143 508 0.6030 0.4581 0.6030 0.7765
0.3527 14.5714 510 0.6242 0.4377 0.6242 0.7901
0.3527 14.6286 512 0.6298 0.4029 0.6298 0.7936
0.3527 14.6857 514 0.6203 0.4705 0.6203 0.7876
0.3527 14.7429 516 0.6059 0.4659 0.6059 0.7784

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

Finetuned
(4019)
this model