ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k18_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8512
  • Qwk: 0.3548
  • Mse: 0.8512
  • Rmse: 0.9226

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 3.9660 -0.0215 3.9660 1.9915
No log 0.0690 4 2.3962 -0.0062 2.3962 1.5480
No log 0.1034 6 2.4980 0.0034 2.4980 1.5805
No log 0.1379 8 1.4351 0.0513 1.4351 1.1979
No log 0.1724 10 1.0780 0.2588 1.0780 1.0383
No log 0.2069 12 1.1004 0.2366 1.1004 1.0490
No log 0.2414 14 1.2216 0.0731 1.2216 1.1053
No log 0.2759 16 1.4671 0.0201 1.4671 1.2112
No log 0.3103 18 1.4431 0.0201 1.4431 1.2013
No log 0.3448 20 1.2182 0.1509 1.2182 1.1037
No log 0.3793 22 1.4026 0.0970 1.4026 1.1843
No log 0.4138 24 1.4763 0.0976 1.4763 1.2150
No log 0.4483 26 1.1538 0.1416 1.1538 1.0741
No log 0.4828 28 1.1172 0.2611 1.1172 1.0570
No log 0.5172 30 1.2735 0.2337 1.2735 1.1285
No log 0.5517 32 1.1822 0.2555 1.1822 1.0873
No log 0.5862 34 1.1599 0.3497 1.1599 1.0770
No log 0.6207 36 1.3276 0.2049 1.3276 1.1522
No log 0.6552 38 1.2135 0.2534 1.2135 1.1016
No log 0.6897 40 1.0129 0.3059 1.0129 1.0064
No log 0.7241 42 0.9704 0.2643 0.9704 0.9851
No log 0.7586 44 0.9517 0.2566 0.9517 0.9756
No log 0.7931 46 0.9888 0.2958 0.9888 0.9944
No log 0.8276 48 0.9252 0.3392 0.9252 0.9619
No log 0.8621 50 1.0116 0.2384 1.0116 1.0058
No log 0.8966 52 1.1066 0.1842 1.1066 1.0520
No log 0.9310 54 0.9942 0.3044 0.9942 0.9971
No log 0.9655 56 0.9219 0.3117 0.9219 0.9601
No log 1.0 58 0.9781 0.2564 0.9781 0.9890
No log 1.0345 60 0.9483 0.3059 0.9483 0.9738
No log 1.0690 62 0.9154 0.2988 0.9154 0.9567
No log 1.1034 64 0.9971 0.2467 0.9971 0.9985
No log 1.1379 66 0.9982 0.2795 0.9982 0.9991
No log 1.1724 68 0.9272 0.2765 0.9272 0.9629
No log 1.2069 70 1.0855 0.3183 1.0855 1.0419
No log 1.2414 72 1.2231 0.2602 1.2231 1.1059
No log 1.2759 74 1.1840 0.2512 1.1840 1.0881
No log 1.3103 76 1.0501 0.3004 1.0501 1.0248
No log 1.3448 78 0.9968 0.3271 0.9968 0.9984
No log 1.3793 80 0.9466 0.4097 0.9466 0.9729
No log 1.4138 82 0.8884 0.4261 0.8884 0.9426
No log 1.4483 84 0.8641 0.5052 0.8641 0.9296
No log 1.4828 86 1.0004 0.3595 1.0004 1.0002
No log 1.5172 88 0.9277 0.4367 0.9277 0.9632
No log 1.5517 90 0.8445 0.4717 0.8445 0.9190
No log 1.5862 92 0.8230 0.4894 0.8230 0.9072
No log 1.6207 94 0.8192 0.5657 0.8192 0.9051
No log 1.6552 96 0.8157 0.5390 0.8157 0.9032
No log 1.6897 98 0.8736 0.4686 0.8736 0.9347
No log 1.7241 100 0.9547 0.4397 0.9547 0.9771
No log 1.7586 102 0.9884 0.3849 0.9884 0.9942
No log 1.7931 104 0.8574 0.4984 0.8574 0.9260
No log 1.8276 106 0.8569 0.5057 0.8569 0.9257
No log 1.8621 108 0.9082 0.4392 0.9082 0.9530
No log 1.8966 110 0.8263 0.4883 0.8263 0.9090
No log 1.9310 112 0.9126 0.4282 0.9126 0.9553
No log 1.9655 114 1.0445 0.3650 1.0445 1.0220
No log 2.0 116 1.0500 0.3497 1.0500 1.0247
No log 2.0345 118 1.0473 0.3450 1.0473 1.0234
No log 2.0690 120 0.8660 0.4152 0.8660 0.9306
No log 2.1034 122 0.8012 0.5131 0.8012 0.8951
No log 2.1379 124 0.8067 0.5246 0.8067 0.8981
No log 2.1724 126 0.8229 0.5054 0.8229 0.9071
No log 2.2069 128 0.8705 0.4513 0.8705 0.9330
No log 2.2414 130 0.8986 0.4396 0.8986 0.9479
No log 2.2759 132 0.8306 0.4691 0.8306 0.9114
No log 2.3103 134 0.8716 0.3802 0.8716 0.9336
No log 2.3448 136 0.9368 0.4085 0.9368 0.9679
No log 2.3793 138 0.8882 0.3859 0.8882 0.9425
No log 2.4138 140 0.8981 0.4422 0.8981 0.9477
No log 2.4483 142 0.9115 0.4279 0.9115 0.9547
No log 2.4828 144 0.9152 0.3784 0.9152 0.9566
No log 2.5172 146 0.8766 0.4014 0.8766 0.9363
No log 2.5517 148 0.8592 0.4014 0.8592 0.9269
No log 2.5862 150 0.8675 0.4422 0.8675 0.9314
No log 2.6207 152 0.8337 0.4422 0.8337 0.9131
No log 2.6552 154 0.8394 0.4204 0.8394 0.9162
No log 2.6897 156 0.8537 0.3777 0.8537 0.9240
No log 2.7241 158 0.9787 0.4149 0.9787 0.9893
No log 2.7586 160 0.9691 0.4353 0.9691 0.9844
No log 2.7931 162 0.8265 0.5313 0.8265 0.9091
No log 2.8276 164 0.8266 0.5421 0.8266 0.9092
No log 2.8621 166 0.8637 0.5205 0.8637 0.9294
No log 2.8966 168 1.0319 0.4273 1.0319 1.0158
No log 2.9310 170 1.0057 0.4497 1.0057 1.0028
No log 2.9655 172 0.8669 0.4657 0.8669 0.9311
No log 3.0 174 0.8412 0.4983 0.8412 0.9172
No log 3.0345 176 0.8451 0.4505 0.8451 0.9193
No log 3.0690 178 0.9834 0.4324 0.9834 0.9917
No log 3.1034 180 1.4302 0.1929 1.4302 1.1959
No log 3.1379 182 1.4239 0.2354 1.4239 1.1933
No log 3.1724 184 1.0497 0.3288 1.0497 1.0245
No log 3.2069 186 0.8595 0.4152 0.8595 0.9271
No log 3.2414 188 0.8216 0.4656 0.8216 0.9064
No log 3.2759 190 0.7930 0.5171 0.7930 0.8905
No log 3.3103 192 0.7891 0.5478 0.7891 0.8883
No log 3.3448 194 0.8080 0.5406 0.8080 0.8989
No log 3.3793 196 0.8090 0.5213 0.8090 0.8995
No log 3.4138 198 0.8158 0.5329 0.8158 0.9032
No log 3.4483 200 0.8228 0.4988 0.8228 0.9071
No log 3.4828 202 0.8186 0.4873 0.8186 0.9047
No log 3.5172 204 0.8321 0.4614 0.8321 0.9122
No log 3.5517 206 0.8125 0.4660 0.8125 0.9014
No log 3.5862 208 0.8322 0.4483 0.8322 0.9122
No log 3.6207 210 0.8285 0.4483 0.8285 0.9102
No log 3.6552 212 0.8029 0.4395 0.8029 0.8960
No log 3.6897 214 0.8187 0.4932 0.8187 0.9048
No log 3.7241 216 0.8059 0.4279 0.8059 0.8977
No log 3.7586 218 0.8232 0.4097 0.8232 0.9073
No log 3.7931 220 0.8434 0.4216 0.8434 0.9183
No log 3.8276 222 0.8498 0.3280 0.8498 0.9219
No log 3.8621 224 0.8835 0.4038 0.8835 0.9400
No log 3.8966 226 0.8900 0.3529 0.8900 0.9434
No log 3.9310 228 0.9056 0.3229 0.9056 0.9517
No log 3.9655 230 0.8856 0.3651 0.8856 0.9411
No log 4.0 232 0.8651 0.4180 0.8651 0.9301
No log 4.0345 234 0.8728 0.5131 0.8728 0.9342
No log 4.0690 236 0.8262 0.4874 0.8262 0.9089
No log 4.1034 238 0.8152 0.4611 0.8152 0.9029
No log 4.1379 240 0.8909 0.4584 0.8909 0.9439
No log 4.1724 242 0.9539 0.4430 0.9539 0.9767
No log 4.2069 244 0.8740 0.4300 0.8740 0.9349
No log 4.2414 246 0.8308 0.3977 0.8308 0.9115
No log 4.2759 248 0.8547 0.4344 0.8547 0.9245
No log 4.3103 250 0.9213 0.3939 0.9213 0.9599
No log 4.3448 252 0.9662 0.3939 0.9662 0.9830
No log 4.3793 254 0.9260 0.3939 0.9260 0.9623
No log 4.4138 256 0.8684 0.4465 0.8684 0.9319
No log 4.4483 258 0.8608 0.4366 0.8608 0.9278
No log 4.4828 260 0.8862 0.4742 0.8862 0.9414
No log 4.5172 262 0.8477 0.4875 0.8477 0.9207
No log 4.5517 264 0.8025 0.4718 0.8025 0.8958
No log 4.5862 266 0.7902 0.5044 0.7902 0.8889
No log 4.6207 268 0.8109 0.4883 0.8109 0.9005
No log 4.6552 270 0.8076 0.4772 0.8076 0.8987
No log 4.6897 272 0.8014 0.4941 0.8014 0.8952
No log 4.7241 274 0.8646 0.5134 0.8646 0.9298
No log 4.7586 276 0.8594 0.5005 0.8594 0.9270
No log 4.7931 278 0.8070 0.5066 0.8070 0.8983
No log 4.8276 280 0.7947 0.4802 0.7947 0.8915
No log 4.8621 282 0.8066 0.5542 0.8066 0.8981
No log 4.8966 284 1.0058 0.5041 1.0058 1.0029
No log 4.9310 286 1.0243 0.4835 1.0243 1.0121
No log 4.9655 288 0.8584 0.5548 0.8584 0.9265
No log 5.0 290 0.7974 0.4770 0.7974 0.8930
No log 5.0345 292 0.8874 0.4588 0.8874 0.9420
No log 5.0690 294 0.8563 0.4712 0.8563 0.9254
No log 5.1034 296 0.8040 0.4938 0.8040 0.8967
No log 5.1379 298 0.9755 0.4240 0.9755 0.9877
No log 5.1724 300 0.9991 0.4160 0.9991 0.9995
No log 5.2069 302 0.8358 0.5326 0.8358 0.9142
No log 5.2414 304 0.7625 0.4544 0.7625 0.8732
No log 5.2759 306 0.7800 0.4216 0.7800 0.8832
No log 5.3103 308 0.8068 0.4478 0.8068 0.8982
No log 5.3448 310 0.8017 0.4216 0.8017 0.8954
No log 5.3793 312 0.7836 0.4269 0.7836 0.8852
No log 5.4138 314 0.7754 0.4269 0.7754 0.8805
No log 5.4483 316 0.7684 0.4269 0.7684 0.8766
No log 5.4828 318 0.7748 0.4628 0.7748 0.8802
No log 5.5172 320 0.7877 0.5195 0.7877 0.8875
No log 5.5517 322 0.8066 0.4839 0.8066 0.8981
No log 5.5862 324 0.8459 0.4433 0.8459 0.9197
No log 5.6207 326 0.8123 0.4044 0.8123 0.9013
No log 5.6552 328 0.7744 0.4133 0.7744 0.8800
No log 5.6897 330 0.8372 0.4321 0.8372 0.9150
No log 5.7241 332 0.8432 0.4198 0.8432 0.9183
No log 5.7586 334 0.8423 0.4540 0.8423 0.9178
No log 5.7931 336 0.8741 0.4327 0.8741 0.9349
No log 5.8276 338 0.8919 0.4344 0.8919 0.9444
No log 5.8621 340 0.8877 0.3802 0.8877 0.9422
No log 5.8966 342 0.8680 0.3859 0.8680 0.9317
No log 5.9310 344 0.9160 0.4225 0.9160 0.9571
No log 5.9655 346 1.0638 0.4288 1.0638 1.0314
No log 6.0 348 1.0280 0.4265 1.0280 1.0139
No log 6.0345 350 0.8654 0.4826 0.8654 0.9302
No log 6.0690 352 0.8159 0.3603 0.8159 0.9033
No log 6.1034 354 0.8230 0.4093 0.8230 0.9072
No log 6.1379 356 0.7947 0.3996 0.7947 0.8914
No log 6.1724 358 0.7901 0.4676 0.7901 0.8889
No log 6.2069 360 0.8880 0.4806 0.8880 0.9423
No log 6.2414 362 0.8839 0.4912 0.8839 0.9402
No log 6.2759 364 0.8026 0.4596 0.8026 0.8959
No log 6.3103 366 0.8051 0.4726 0.8051 0.8973
No log 6.3448 368 0.8669 0.5150 0.8669 0.9311
No log 6.3793 370 0.8324 0.4726 0.8324 0.9124
No log 6.4138 372 0.8314 0.4017 0.8314 0.9118
No log 6.4483 374 0.9137 0.3834 0.9137 0.9559
No log 6.4828 376 0.9119 0.2713 0.9119 0.9549
No log 6.5172 378 0.8796 0.4017 0.8796 0.9379
No log 6.5517 380 0.8588 0.3737 0.8588 0.9267
No log 6.5862 382 0.8600 0.3552 0.8600 0.9274
No log 6.6207 384 0.8562 0.3878 0.8562 0.9253
No log 6.6552 386 0.8556 0.3996 0.8556 0.9250
No log 6.6897 388 0.8515 0.3571 0.8515 0.9228
No log 6.7241 390 0.8464 0.4109 0.8464 0.9200
No log 6.7586 392 0.8768 0.4165 0.8768 0.9364
No log 6.7931 394 0.8766 0.4368 0.8766 0.9362
No log 6.8276 396 0.8629 0.4229 0.8629 0.9289
No log 6.8621 398 0.8421 0.4759 0.8421 0.9176
No log 6.8966 400 0.8378 0.4216 0.8378 0.9153
No log 6.9310 402 0.8051 0.3979 0.8051 0.8973
No log 6.9655 404 0.8294 0.4826 0.8294 0.9107
No log 7.0 406 0.8522 0.4738 0.8522 0.9232
No log 7.0345 408 0.7877 0.4692 0.7877 0.8875
No log 7.0690 410 0.7932 0.4760 0.7932 0.8906
No log 7.1034 412 0.9170 0.5006 0.9170 0.9576
No log 7.1379 414 0.9325 0.4906 0.9325 0.9657
No log 7.1724 416 0.8142 0.5331 0.8142 0.9023
No log 7.2069 418 0.7705 0.3552 0.7705 0.8778
No log 7.2414 420 0.7732 0.4151 0.7732 0.8793
No log 7.2759 422 0.7925 0.3996 0.7925 0.8902
No log 7.3103 424 0.8325 0.3757 0.8325 0.9124
No log 7.3448 426 0.8358 0.3757 0.8358 0.9142
No log 7.3793 428 0.8083 0.4048 0.8083 0.8991
No log 7.4138 430 0.7606 0.4269 0.7606 0.8721
No log 7.4483 432 0.7413 0.4428 0.7413 0.8610
No log 7.4828 434 0.7381 0.5017 0.7381 0.8591
No log 7.5172 436 0.7804 0.5317 0.7804 0.8834
No log 7.5517 438 0.7659 0.5331 0.7659 0.8752
No log 7.5862 440 0.7333 0.4988 0.7333 0.8563
No log 7.6207 442 0.7204 0.4938 0.7204 0.8488
No log 7.6552 444 0.7402 0.5794 0.7402 0.8603
No log 7.6897 446 0.8059 0.5257 0.8059 0.8977
No log 7.7241 448 0.8005 0.5019 0.8005 0.8947
No log 7.7586 450 0.7585 0.4984 0.7585 0.8709
No log 7.7931 452 0.7515 0.5202 0.7515 0.8669
No log 7.8276 454 0.7637 0.5202 0.7637 0.8739
No log 7.8621 456 0.7685 0.5202 0.7685 0.8766
No log 7.8966 458 0.7582 0.4932 0.7582 0.8707
No log 7.9310 460 0.7577 0.5602 0.7577 0.8705
No log 7.9655 462 0.7722 0.5329 0.7722 0.8788
No log 8.0 464 0.7801 0.5459 0.7801 0.8832
No log 8.0345 466 0.7859 0.5459 0.7859 0.8865
No log 8.0690 468 0.7966 0.5431 0.7966 0.8925
No log 8.1034 470 0.7842 0.5342 0.7842 0.8855
No log 8.1379 472 0.7690 0.5480 0.7690 0.8769
No log 8.1724 474 0.7618 0.4804 0.7618 0.8728
No log 8.2069 476 0.7650 0.4834 0.7650 0.8746
No log 8.2414 478 0.7835 0.5327 0.7835 0.8852
No log 8.2759 480 0.7689 0.5202 0.7689 0.8769
No log 8.3103 482 0.7789 0.5188 0.7789 0.8826
No log 8.3448 484 0.8135 0.4984 0.8135 0.9020
No log 8.3793 486 0.8067 0.5074 0.8067 0.8982
No log 8.4138 488 0.8138 0.4787 0.8138 0.9021
No log 8.4483 490 0.8398 0.3922 0.8398 0.9164
No log 8.4828 492 0.8234 0.4204 0.8234 0.9074
No log 8.5172 494 0.8183 0.4706 0.8183 0.9046
No log 8.5517 496 0.8322 0.4609 0.8322 0.9123
No log 8.5862 498 0.8111 0.4594 0.8111 0.9006
0.2927 8.6207 500 0.8244 0.4768 0.8244 0.9080
0.2927 8.6552 502 0.9357 0.4464 0.9357 0.9673
0.2927 8.6897 504 0.9840 0.4387 0.9840 0.9920
0.2927 8.7241 506 0.9039 0.5147 0.9039 0.9508
0.2927 8.7586 508 0.8327 0.4648 0.8327 0.9125
0.2927 8.7931 510 0.8139 0.4145 0.8139 0.9022
0.2927 8.8276 512 0.8159 0.4296 0.8159 0.9033
0.2927 8.8621 514 0.8374 0.3877 0.8374 0.9151
0.2927 8.8966 516 0.8512 0.3548 0.8512 0.9226

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k18_task5_organization

Finetuned
(4019)
this model