ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k11_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7459
  • Qwk: 0.6939
  • Mse: 0.7459
  • Rmse: 0.8637

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 7.2633 -0.0211 7.2633 2.6951
No log 0.0784 4 4.3897 0.0735 4.3897 2.0952
No log 0.1176 6 3.0293 0.0952 3.0293 1.7405
No log 0.1569 8 2.1153 0.1926 2.1153 1.4544
No log 0.1961 10 1.8169 0.1579 1.8169 1.3479
No log 0.2353 12 1.8198 0.2435 1.8198 1.3490
No log 0.2745 14 3.2015 0.0787 3.2015 1.7893
No log 0.3137 16 4.9245 0.0215 4.9245 2.2191
No log 0.3529 18 4.7299 0.0444 4.7299 2.1748
No log 0.3922 20 3.7723 0.1587 3.7723 1.9422
No log 0.4314 22 2.9146 0.0261 2.9146 1.7072
No log 0.4706 24 2.1950 0.1111 2.1950 1.4816
No log 0.5098 26 1.9037 0.2769 1.9037 1.3798
No log 0.5490 28 1.7331 0.3040 1.7331 1.3165
No log 0.5882 30 1.8497 0.2946 1.8497 1.3600
No log 0.6275 32 1.9451 0.3030 1.9451 1.3947
No log 0.6667 34 1.8258 0.3030 1.8258 1.3512
No log 0.7059 36 1.4989 0.3471 1.4989 1.2243
No log 0.7451 38 1.4095 0.3667 1.4095 1.1872
No log 0.7843 40 1.4722 0.4 1.4722 1.2133
No log 0.8235 42 1.7517 0.3556 1.7517 1.3235
No log 0.8627 44 1.9383 0.2609 1.9383 1.3922
No log 0.9020 46 1.8025 0.2920 1.8025 1.3426
No log 0.9412 48 1.2538 0.5075 1.2538 1.1197
No log 0.9804 50 1.0106 0.5581 1.0106 1.0053
No log 1.0196 52 0.8971 0.6107 0.8971 0.9471
No log 1.0588 54 0.9862 0.6423 0.9862 0.9931
No log 1.0980 56 1.1151 0.5693 1.1151 1.0560
No log 1.1373 58 1.0576 0.5507 1.0576 1.0284
No log 1.1765 60 0.8552 0.6308 0.8552 0.9247
No log 1.2157 62 0.8816 0.6047 0.8816 0.9389
No log 1.2549 64 0.9808 0.625 0.9808 0.9904
No log 1.2941 66 1.5718 0.4967 1.5718 1.2537
No log 1.3333 68 1.6301 0.4503 1.6301 1.2767
No log 1.3725 70 1.3498 0.4828 1.3498 1.1618
No log 1.4118 72 1.0120 0.5926 1.0120 1.0060
No log 1.4510 74 0.8712 0.6569 0.8712 0.9334
No log 1.4902 76 0.8257 0.6471 0.8257 0.9087
No log 1.5294 78 0.9052 0.6331 0.9052 0.9514
No log 1.5686 80 1.2040 0.5594 1.2040 1.0973
No log 1.6078 82 1.3931 0.5405 1.3931 1.1803
No log 1.6471 84 1.2966 0.5405 1.2966 1.1387
No log 1.6863 86 1.0260 0.6259 1.0260 1.0129
No log 1.7255 88 1.1786 0.5946 1.1786 1.0856
No log 1.7647 90 1.5945 0.5375 1.5945 1.2627
No log 1.8039 92 1.2474 0.5638 1.2474 1.1169
No log 1.8431 94 0.9332 0.6383 0.9332 0.9660
No log 1.8824 96 1.0054 0.5469 1.0054 1.0027
No log 1.9216 98 0.9853 0.5954 0.9853 0.9926
No log 1.9608 100 1.1725 0.5612 1.1725 1.0828
No log 2.0 102 1.6378 0.4675 1.6378 1.2798
No log 2.0392 104 1.7123 0.4615 1.7123 1.3086
No log 2.0784 106 1.4480 0.5806 1.4480 1.2033
No log 2.1176 108 1.1044 0.6 1.1044 1.0509
No log 2.1569 110 0.9386 0.6490 0.9386 0.9688
No log 2.1961 112 0.8535 0.7013 0.8535 0.9239
No log 2.2353 114 0.9418 0.6792 0.9418 0.9705
No log 2.2745 116 0.9127 0.7170 0.9127 0.9553
No log 2.3137 118 1.0376 0.6584 1.0376 1.0186
No log 2.3529 120 1.0542 0.675 1.0542 1.0268
No log 2.3922 122 0.9377 0.6928 0.9377 0.9683
No log 2.4314 124 0.9514 0.6974 0.9514 0.9754
No log 2.4706 126 0.7992 0.7162 0.7992 0.8940
No log 2.5098 128 0.7172 0.7007 0.7172 0.8469
No log 2.5490 130 0.7944 0.6370 0.7944 0.8913
No log 2.5882 132 0.7754 0.6963 0.7754 0.8806
No log 2.6275 134 0.9347 0.6074 0.9347 0.9668
No log 2.6667 136 1.1792 0.6 1.1792 1.0859
No log 2.7059 138 1.3622 0.5912 1.3622 1.1671
No log 2.7451 140 1.0857 0.6225 1.0857 1.0420
No log 2.7843 142 0.7818 0.7105 0.7818 0.8842
No log 2.8235 144 0.7408 0.7368 0.7408 0.8607
No log 2.8627 146 0.9231 0.7407 0.9231 0.9608
No log 2.9020 148 0.9970 0.7030 0.9970 0.9985
No log 2.9412 150 0.8551 0.7421 0.8551 0.9247
No log 2.9804 152 0.6889 0.7383 0.6889 0.8300
No log 3.0196 154 0.7310 0.6761 0.7310 0.8550
No log 3.0588 156 0.7344 0.7162 0.7344 0.8570
No log 3.0980 158 0.8525 0.7114 0.8525 0.9233
No log 3.1373 160 0.8831 0.7625 0.8831 0.9397
No log 3.1765 162 0.7284 0.7826 0.7284 0.8534
No log 3.2157 164 0.6440 0.7162 0.6440 0.8025
No log 3.2549 166 0.6379 0.75 0.6379 0.7987
No log 3.2941 168 0.6601 0.7821 0.6601 0.8125
No log 3.3333 170 0.6960 0.775 0.6960 0.8343
No log 3.3725 172 0.8307 0.7453 0.8307 0.9114
No log 3.4118 174 1.0147 0.6784 1.0147 1.0073
No log 3.4510 176 0.9460 0.7170 0.9460 0.9726
No log 3.4902 178 0.8472 0.7468 0.8472 0.9204
No log 3.5294 180 0.8250 0.7673 0.8250 0.9083
No log 3.5686 182 0.7710 0.7432 0.7710 0.8781
No log 3.6078 184 0.7932 0.7075 0.7932 0.8906
No log 3.6471 186 0.7461 0.6809 0.7461 0.8638
No log 3.6863 188 0.7616 0.6809 0.7616 0.8727
No log 3.7255 190 0.8517 0.7152 0.8517 0.9229
No log 3.7647 192 1.0971 0.6335 1.0971 1.0474
No log 3.8039 194 1.5198 0.5474 1.5198 1.2328
No log 3.8431 196 1.4266 0.5957 1.4266 1.1944
No log 3.8824 198 0.9417 0.7066 0.9417 0.9704
No log 3.9216 200 0.8214 0.7468 0.8214 0.9063
No log 3.9608 202 0.8993 0.7073 0.8993 0.9483
No log 4.0 204 1.2755 0.6102 1.2755 1.1294
No log 4.0392 206 1.4388 0.6054 1.4388 1.1995
No log 4.0784 208 1.2021 0.6358 1.2021 1.0964
No log 4.1176 210 0.8681 0.6883 0.8681 0.9317
No log 4.1569 212 0.7850 0.7632 0.7850 0.8860
No log 4.1961 214 0.7786 0.7742 0.7786 0.8824
No log 4.2353 216 0.8220 0.6797 0.8220 0.9067
No log 4.2745 218 0.8272 0.6923 0.8272 0.9095
No log 4.3137 220 0.7158 0.7123 0.7158 0.8461
No log 4.3529 222 0.6223 0.7432 0.6223 0.7889
No log 4.3922 224 0.6008 0.7310 0.6008 0.7751
No log 4.4314 226 0.6264 0.7550 0.6264 0.7914
No log 4.4706 228 0.6342 0.7632 0.6342 0.7964
No log 4.5098 230 0.6701 0.7248 0.6701 0.8186
No log 4.5490 232 0.7009 0.7763 0.7009 0.8372
No log 4.5882 234 0.7883 0.7654 0.7883 0.8879
No log 4.6275 236 0.8542 0.7619 0.8542 0.9242
No log 4.6667 238 0.7561 0.7730 0.7561 0.8695
No log 4.7059 240 0.7092 0.7730 0.7092 0.8421
No log 4.7451 242 0.6615 0.7742 0.6615 0.8134
No log 4.7843 244 0.7385 0.7595 0.7385 0.8594
No log 4.8235 246 0.7942 0.7389 0.7942 0.8912
No log 4.8627 248 0.8188 0.7296 0.8188 0.9049
No log 4.9020 250 0.7643 0.7027 0.7643 0.8742
No log 4.9412 252 0.7210 0.6892 0.7210 0.8491
No log 4.9804 254 0.7085 0.7114 0.7085 0.8417
No log 5.0196 256 0.7005 0.7075 0.7005 0.8369
No log 5.0588 258 0.7924 0.7296 0.7924 0.8902
No log 5.0980 260 1.0667 0.625 1.0667 1.0328
No log 5.1373 262 1.4186 0.6108 1.4186 1.1911
No log 5.1765 264 1.3882 0.6108 1.3882 1.1782
No log 5.2157 266 1.0139 0.6369 1.0139 1.0069
No log 5.2549 268 0.8770 0.6225 0.8770 0.9365
No log 5.2941 270 0.9454 0.6369 0.9454 0.9723
No log 5.3333 272 1.3212 0.625 1.3212 1.1494
No log 5.3725 274 1.4892 0.5851 1.4892 1.2203
No log 5.4118 276 1.3037 0.6145 1.3037 1.1418
No log 5.4510 278 0.9875 0.6087 0.9875 0.9937
No log 5.4902 280 0.8333 0.6759 0.8333 0.9128
No log 5.5294 282 0.8216 0.6571 0.8216 0.9064
No log 5.5686 284 0.8512 0.6232 0.8512 0.9226
No log 5.6078 286 0.9092 0.6531 0.9092 0.9535
No log 5.6471 288 0.8942 0.6486 0.8942 0.9456
No log 5.6863 290 0.8129 0.6377 0.8129 0.9016
No log 5.7255 292 0.7546 0.7007 0.7546 0.8687
No log 5.7647 294 0.7649 0.6861 0.7649 0.8746
No log 5.8039 296 0.7716 0.6906 0.7716 0.8784
No log 5.8431 298 0.8874 0.6099 0.8874 0.9420
No log 5.8824 300 0.9944 0.6395 0.9944 0.9972
No log 5.9216 302 0.9201 0.6207 0.9201 0.9592
No log 5.9608 304 0.8151 0.6429 0.8151 0.9028
No log 6.0 306 0.7953 0.6853 0.7953 0.8918
No log 6.0392 308 0.8014 0.7034 0.8014 0.8952
No log 6.0784 310 0.8794 0.6056 0.8794 0.9378
No log 6.1176 312 0.8797 0.6667 0.8797 0.9379
No log 6.1569 314 0.9260 0.6506 0.9260 0.9623
No log 6.1961 316 0.7984 0.7349 0.7984 0.8936
No log 6.2353 318 0.7017 0.7702 0.7017 0.8377
No log 6.2745 320 0.6723 0.7389 0.6723 0.8199
No log 6.3137 322 0.7668 0.7561 0.7668 0.8757
No log 6.3529 324 0.9996 0.6705 0.9996 0.9998
No log 6.3922 326 1.0631 0.6391 1.0631 1.0310
No log 6.4314 328 0.8862 0.6829 0.8862 0.9414
No log 6.4706 330 0.6973 0.7211 0.6973 0.8350
No log 6.5098 332 0.6771 0.7123 0.6771 0.8229
No log 6.5490 334 0.7061 0.7222 0.7061 0.8403
No log 6.5882 336 0.7040 0.7172 0.7040 0.8390
No log 6.6275 338 0.7217 0.7368 0.7217 0.8495
No log 6.6667 340 0.7069 0.7516 0.7069 0.8407
No log 6.7059 342 0.6371 0.7333 0.6371 0.7982
No log 6.7451 344 0.6478 0.7310 0.6478 0.8049
No log 6.7843 346 0.6515 0.7183 0.6515 0.8072
No log 6.8235 348 0.6148 0.7183 0.6148 0.7841
No log 6.8627 350 0.6029 0.7682 0.6029 0.7765
No log 6.9020 352 0.7243 0.725 0.7243 0.8511
No log 6.9412 354 0.7166 0.7453 0.7166 0.8465
No log 6.9804 356 0.6456 0.7799 0.6456 0.8035
No log 7.0196 358 0.5853 0.7347 0.5853 0.7650
No log 7.0588 360 0.6383 0.7194 0.6383 0.7990
No log 7.0980 362 0.7100 0.6818 0.7100 0.8426
No log 7.1373 364 0.7845 0.6567 0.7845 0.8857
No log 7.1765 366 0.8101 0.6462 0.8101 0.9000
No log 7.2157 368 0.7911 0.6512 0.7911 0.8894
No log 7.2549 370 0.7404 0.6718 0.7404 0.8605
No log 7.2941 372 0.6466 0.7324 0.6466 0.8041
No log 7.3333 374 0.6242 0.7222 0.6242 0.7900
No log 7.3725 376 0.6829 0.7226 0.6829 0.8264
No log 7.4118 378 0.7557 0.7143 0.7557 0.8693
No log 7.4510 380 0.8701 0.7143 0.8701 0.9328
No log 7.4902 382 0.8063 0.6892 0.8063 0.8980
No log 7.5294 384 0.6907 0.7042 0.6907 0.8311
No log 7.5686 386 0.6725 0.7143 0.6725 0.8200
No log 7.6078 388 0.6705 0.7133 0.6705 0.8188
No log 7.6471 390 0.6675 0.7297 0.6675 0.8170
No log 7.6863 392 0.7103 0.7417 0.7103 0.8428
No log 7.7255 394 0.7604 0.72 0.7604 0.8720
No log 7.7647 396 0.8024 0.7114 0.8024 0.8957
No log 7.8039 398 0.7507 0.6849 0.7507 0.8664
No log 7.8431 400 0.7405 0.7067 0.7405 0.8605
No log 7.8824 402 0.6304 0.7172 0.6304 0.7940
No log 7.9216 404 0.5743 0.7324 0.5743 0.7578
No log 7.9608 406 0.5775 0.7183 0.5775 0.7599
No log 8.0 408 0.5657 0.7534 0.5657 0.7521
No log 8.0392 410 0.6915 0.7059 0.6915 0.8316
No log 8.0784 412 0.8162 0.7226 0.8162 0.9034
No log 8.1176 414 0.7860 0.7105 0.7860 0.8865
No log 8.1569 416 0.6727 0.6892 0.6727 0.8202
No log 8.1961 418 0.6504 0.7448 0.6504 0.8065
No log 8.2353 420 0.6580 0.7050 0.6580 0.8112
No log 8.2745 422 0.6800 0.7050 0.6800 0.8246
No log 8.3137 424 0.6952 0.6857 0.6952 0.8338
No log 8.3529 426 0.7042 0.6475 0.7042 0.8392
No log 8.3922 428 0.7065 0.6933 0.7065 0.8406
No log 8.4314 430 0.6698 0.7226 0.6698 0.8184
No log 8.4706 432 0.5907 0.7632 0.5907 0.7686
No log 8.5098 434 0.5787 0.7550 0.5787 0.7608
No log 8.5490 436 0.6106 0.7662 0.6106 0.7814
No log 8.5882 438 0.6700 0.7059 0.6700 0.8185
No log 8.6275 440 0.7426 0.7485 0.7426 0.8617
No log 8.6667 442 0.6960 0.7190 0.6960 0.8343
No log 8.7059 444 0.6837 0.6892 0.6837 0.8269
No log 8.7451 446 0.6893 0.7067 0.6893 0.8302
No log 8.7843 448 0.6376 0.7582 0.6376 0.7985
No log 8.8235 450 0.6194 0.7763 0.6194 0.7870
No log 8.8627 452 0.6058 0.7712 0.6058 0.7783
No log 8.9020 454 0.5877 0.7712 0.5877 0.7666
No log 8.9412 456 0.5980 0.7771 0.5980 0.7733
No log 8.9804 458 0.6038 0.7564 0.6038 0.7770
No log 9.0196 460 0.6718 0.7403 0.6718 0.8196
No log 9.0588 462 0.7925 0.7134 0.7925 0.8902
No log 9.0980 464 0.7528 0.7421 0.7528 0.8676
No log 9.1373 466 0.7685 0.7421 0.7685 0.8766
No log 9.1765 468 0.9625 0.6585 0.9625 0.9810
No log 9.2157 470 0.9552 0.6667 0.9552 0.9774
No log 9.2549 472 0.7853 0.7105 0.7853 0.8862
No log 9.2941 474 0.7039 0.72 0.7039 0.8390
No log 9.3333 476 0.6752 0.7347 0.6752 0.8217
No log 9.3725 478 0.6504 0.7211 0.6504 0.8065
No log 9.4118 480 0.6186 0.7517 0.6186 0.7865
No log 9.4510 482 0.6012 0.7517 0.6012 0.7754
No log 9.4902 484 0.6247 0.7582 0.6247 0.7904
No log 9.5294 486 0.6548 0.7403 0.6548 0.8092
No log 9.5686 488 0.6074 0.7613 0.6074 0.7794
No log 9.6078 490 0.5740 0.7632 0.5740 0.7576
No log 9.6471 492 0.6311 0.7517 0.6311 0.7944
No log 9.6863 494 0.6394 0.7517 0.6394 0.7996
No log 9.7255 496 0.5805 0.7843 0.5805 0.7619
No log 9.7647 498 0.6578 0.7436 0.6578 0.8111
0.4198 9.8039 500 0.8876 0.7195 0.8876 0.9421
0.4198 9.8431 502 1.0356 0.6506 1.0356 1.0176
0.4198 9.8824 504 0.9060 0.7 0.9060 0.9519
0.4198 9.9216 506 0.7059 0.7285 0.7059 0.8402
0.4198 9.9608 508 0.6518 0.7172 0.6518 0.8074
0.4198 10.0 510 0.6757 0.6944 0.6757 0.8220
0.4198 10.0392 512 0.7422 0.7027 0.7422 0.8615
0.4198 10.0784 514 0.7830 0.7114 0.7830 0.8849
0.4198 10.1176 516 0.7459 0.6939 0.7459 0.8637

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k11_task1_organization

Finetuned
(4019)
this model