ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k17_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8513
  • Qwk: 0.3787
  • Mse: 0.8513
  • Rmse: 0.9227

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0235 2 4.6978 -0.0020 4.6978 2.1674
No log 0.0471 4 2.6845 -0.0030 2.6845 1.6384
No log 0.0706 6 2.0855 -0.0303 2.0855 1.4441
No log 0.0941 8 1.6451 -0.0801 1.6451 1.2826
No log 0.1176 10 1.2830 0.0452 1.2830 1.1327
No log 0.1412 12 1.4011 -0.0182 1.4011 1.1837
No log 0.1647 14 1.6973 0.0385 1.6973 1.3028
No log 0.1882 16 1.4450 -0.0029 1.4450 1.2021
No log 0.2118 18 1.1391 0.2408 1.1391 1.0673
No log 0.2353 20 1.1452 0.1752 1.1452 1.0701
No log 0.2588 22 1.1059 0.2025 1.1059 1.0516
No log 0.2824 24 1.3816 0.1427 1.3816 1.1754
No log 0.3059 26 1.5301 0.0477 1.5301 1.2370
No log 0.3294 28 1.4768 0.0317 1.4768 1.2152
No log 0.3529 30 1.3300 0.0662 1.3300 1.1533
No log 0.3765 32 1.2329 0.0750 1.2329 1.1104
No log 0.4 34 1.2585 0.0898 1.2585 1.1218
No log 0.4235 36 1.4999 0.1418 1.4999 1.2247
No log 0.4471 38 1.6755 0.1540 1.6755 1.2944
No log 0.4706 40 1.6172 0.1702 1.6172 1.2717
No log 0.4941 42 1.3334 0.1364 1.3334 1.1547
No log 0.5176 44 1.0979 0.3491 1.0979 1.0478
No log 0.5412 46 1.1120 0.3705 1.1120 1.0545
No log 0.5647 48 1.3838 0.3453 1.3838 1.1763
No log 0.5882 50 1.9076 0.2209 1.9076 1.3812
No log 0.6118 52 2.5446 0.0991 2.5446 1.5952
No log 0.6353 54 2.5761 0.1061 2.5761 1.6050
No log 0.6588 56 2.2703 0.0991 2.2703 1.5067
No log 0.6824 58 1.8026 0.3216 1.8026 1.3426
No log 0.7059 60 1.4631 0.2564 1.4631 1.2096
No log 0.7294 62 1.1054 0.3077 1.1054 1.0514
No log 0.7529 64 1.0038 0.3796 1.0038 1.0019
No log 0.7765 66 1.0853 0.3033 1.0853 1.0418
No log 0.8 68 1.3362 0.2153 1.3362 1.1559
No log 0.8235 70 1.7341 0.2245 1.7341 1.3168
No log 0.8471 72 1.8396 0.2002 1.8396 1.3563
No log 0.8706 74 1.7149 0.2245 1.7149 1.3095
No log 0.8941 76 1.3625 0.1713 1.3625 1.1673
No log 0.9176 78 1.0775 0.4045 1.0775 1.0380
No log 0.9412 80 1.0729 0.3678 1.0729 1.0358
No log 0.9647 82 1.0855 0.3678 1.0855 1.0419
No log 0.9882 84 1.0805 0.3829 1.0805 1.0395
No log 1.0118 86 1.1872 0.2131 1.1872 1.0896
No log 1.0353 88 1.2672 0.2839 1.2672 1.1257
No log 1.0588 90 1.2523 0.3050 1.2523 1.1191
No log 1.0824 92 0.9520 0.4460 0.9520 0.9757
No log 1.1059 94 0.8982 0.3652 0.8982 0.9477
No log 1.1294 96 0.9434 0.3139 0.9434 0.9713
No log 1.1529 98 1.0477 0.3354 1.0477 1.0236
No log 1.1765 100 1.1406 0.3294 1.1406 1.0680
No log 1.2 102 1.1698 0.3056 1.1698 1.0816
No log 1.2235 104 1.2104 0.2876 1.2104 1.1002
No log 1.2471 106 1.0814 0.3750 1.0815 1.0399
No log 1.2706 108 1.0630 0.4022 1.0630 1.0310
No log 1.2941 110 0.9960 0.3496 0.9960 0.9980
No log 1.3176 112 0.9925 0.3913 0.9925 0.9962
No log 1.3412 114 0.9357 0.3555 0.9357 0.9673
No log 1.3647 116 0.8679 0.3705 0.8679 0.9316
No log 1.3882 118 0.8420 0.4653 0.8420 0.9176
No log 1.4118 120 0.8590 0.5380 0.8590 0.9268
No log 1.4353 122 0.8527 0.5164 0.8527 0.9234
No log 1.4588 124 1.1251 0.5157 1.1251 1.0607
No log 1.4824 126 1.0803 0.5062 1.0803 1.0394
No log 1.5059 128 0.9184 0.4493 0.9184 0.9583
No log 1.5294 130 1.0107 0.4817 1.0107 1.0054
No log 1.5529 132 1.0242 0.4978 1.0242 1.0120
No log 1.5765 134 0.9512 0.4913 0.9512 0.9753
No log 1.6 136 0.8296 0.5392 0.8296 0.9108
No log 1.6235 138 0.7871 0.5892 0.7871 0.8872
No log 1.6471 140 0.7866 0.5684 0.7866 0.8869
No log 1.6706 142 0.7699 0.5921 0.7699 0.8774
No log 1.6941 144 0.7896 0.6010 0.7896 0.8886
No log 1.7176 146 0.7775 0.5749 0.7775 0.8817
No log 1.7412 148 0.8372 0.5685 0.8372 0.9150
No log 1.7647 150 0.8481 0.5409 0.8481 0.9209
No log 1.7882 152 0.7659 0.4966 0.7659 0.8751
No log 1.8118 154 0.8482 0.5339 0.8482 0.9210
No log 1.8353 156 0.8650 0.4621 0.8650 0.9300
No log 1.8588 158 0.7737 0.6025 0.7737 0.8796
No log 1.8824 160 0.8146 0.5411 0.8146 0.9026
No log 1.9059 162 0.8044 0.5136 0.8044 0.8969
No log 1.9294 164 0.7680 0.5759 0.7680 0.8763
No log 1.9529 166 0.8363 0.5209 0.8363 0.9145
No log 1.9765 168 0.9673 0.4005 0.9673 0.9835
No log 2.0 170 0.8997 0.4574 0.8997 0.9485
No log 2.0235 172 0.8758 0.4842 0.8758 0.9358
No log 2.0471 174 0.9189 0.4556 0.9189 0.9586
No log 2.0706 176 0.9031 0.4042 0.9031 0.9503
No log 2.0941 178 0.8768 0.4197 0.8768 0.9364
No log 2.1176 180 0.8625 0.4534 0.8625 0.9287
No log 2.1412 182 0.8316 0.5905 0.8316 0.9119
No log 2.1647 184 0.8293 0.6032 0.8293 0.9107
No log 2.1882 186 0.8343 0.5606 0.8343 0.9134
No log 2.2118 188 0.8328 0.5028 0.8328 0.9126
No log 2.2353 190 0.8461 0.4519 0.8461 0.9198
No log 2.2588 192 0.8535 0.4804 0.8535 0.9239
No log 2.2824 194 0.8627 0.4571 0.8627 0.9288
No log 2.3059 196 0.8921 0.4299 0.8921 0.9445
No log 2.3294 198 0.8768 0.4498 0.8768 0.9364
No log 2.3529 200 0.8628 0.4603 0.8628 0.9289
No log 2.3765 202 0.9039 0.4553 0.9039 0.9507
No log 2.4 204 0.8851 0.4459 0.8851 0.9408
No log 2.4235 206 0.8651 0.4995 0.8651 0.9301
No log 2.4471 208 0.8931 0.4016 0.8931 0.9451
No log 2.4706 210 0.8819 0.3787 0.8819 0.9391
No log 2.4941 212 0.9899 0.4128 0.9899 0.9949
No log 2.5176 214 0.9355 0.3231 0.9355 0.9672
No log 2.5412 216 0.9751 0.3133 0.9751 0.9875
No log 2.5647 218 1.4594 0.3184 1.4594 1.2081
No log 2.5882 220 1.5700 0.2885 1.5700 1.2530
No log 2.6118 222 1.2437 0.3772 1.2437 1.1152
No log 2.6353 224 0.9551 0.3225 0.9551 0.9773
No log 2.6588 226 1.0093 0.4166 1.0093 1.0046
No log 2.6824 228 0.9681 0.3613 0.9681 0.9839
No log 2.7059 230 0.9293 0.3613 0.9293 0.9640
No log 2.7294 232 0.9147 0.4599 0.9147 0.9564
No log 2.7529 234 0.9239 0.4256 0.9239 0.9612
No log 2.7765 236 0.9571 0.3875 0.9571 0.9783
No log 2.8 238 1.0132 0.3433 1.0132 1.0066
No log 2.8235 240 1.0609 0.3149 1.0609 1.0300
No log 2.8471 242 0.9750 0.4098 0.9750 0.9874
No log 2.8706 244 0.9346 0.4474 0.9346 0.9668
No log 2.8941 246 0.9648 0.4477 0.9648 0.9822
No log 2.9176 248 1.0334 0.4200 1.0334 1.0166
No log 2.9412 250 1.0681 0.4668 1.0681 1.0335
No log 2.9647 252 1.0462 0.4131 1.0462 1.0228
No log 2.9882 254 1.0106 0.4164 1.0106 1.0053
No log 3.0118 256 0.9536 0.3298 0.9536 0.9765
No log 3.0353 258 0.9139 0.3119 0.9139 0.9560
No log 3.0588 260 0.9264 0.3590 0.9264 0.9625
No log 3.0824 262 0.9000 0.4056 0.9000 0.9487
No log 3.1059 264 0.8665 0.4512 0.8665 0.9309
No log 3.1294 266 0.8972 0.3802 0.8972 0.9472
No log 3.1529 268 0.9108 0.4113 0.9108 0.9543
No log 3.1765 270 0.8682 0.4734 0.8682 0.9318
No log 3.2 272 0.9041 0.3472 0.9041 0.9508
No log 3.2235 274 0.9253 0.3086 0.9253 0.9619
No log 3.2471 276 0.9022 0.3090 0.9022 0.9498
No log 3.2706 278 0.9063 0.3392 0.9063 0.9520
No log 3.2941 280 0.8984 0.3243 0.8984 0.9479
No log 3.3176 282 0.9208 0.4439 0.9208 0.9596
No log 3.3412 284 1.0361 0.4056 1.0361 1.0179
No log 3.3647 286 1.0575 0.4282 1.0575 1.0283
No log 3.3882 288 1.0114 0.4186 1.0114 1.0057
No log 3.4118 290 0.9534 0.4186 0.9534 0.9764
No log 3.4353 292 0.8667 0.3734 0.8667 0.9310
No log 3.4588 294 0.8403 0.4115 0.8403 0.9167
No log 3.4824 296 0.8125 0.4260 0.8125 0.9014
No log 3.5059 298 0.7728 0.5635 0.7728 0.8791
No log 3.5294 300 0.7611 0.5195 0.7611 0.8724
No log 3.5529 302 0.7497 0.5837 0.7497 0.8659
No log 3.5765 304 0.7584 0.6032 0.7584 0.8708
No log 3.6 306 0.8409 0.4295 0.8409 0.9170
No log 3.6235 308 0.8978 0.4295 0.8978 0.9475
No log 3.6471 310 0.8312 0.4526 0.8312 0.9117
No log 3.6706 312 0.7703 0.5062 0.7703 0.8777
No log 3.6941 314 0.7749 0.4526 0.7749 0.8803
No log 3.7176 316 0.7980 0.4611 0.7980 0.8933
No log 3.7412 318 0.8146 0.5112 0.8146 0.9026
No log 3.7647 320 0.8271 0.4016 0.8271 0.9095
No log 3.7882 322 0.8331 0.4498 0.8331 0.9128
No log 3.8118 324 0.8330 0.5012 0.8330 0.9127
No log 3.8353 326 0.8660 0.4657 0.8660 0.9306
No log 3.8588 328 0.9121 0.4072 0.9121 0.9550
No log 3.8824 330 0.9074 0.4062 0.9074 0.9526
No log 3.9059 332 0.9121 0.3298 0.9121 0.9550
No log 3.9294 334 0.9375 0.3174 0.9375 0.9682
No log 3.9529 336 0.9614 0.3151 0.9614 0.9805
No log 3.9765 338 0.9895 0.2812 0.9895 0.9948
No log 4.0 340 1.0317 0.3939 1.0317 1.0157
No log 4.0235 342 1.0360 0.4661 1.0360 1.0179
No log 4.0471 344 0.9430 0.4596 0.9430 0.9711
No log 4.0706 346 0.9047 0.3961 0.9047 0.9512
No log 4.0941 348 0.9098 0.4067 0.9098 0.9538
No log 4.1176 350 0.9106 0.4203 0.9106 0.9542
No log 4.1412 352 0.9308 0.4503 0.9308 0.9648
No log 4.1647 354 0.9251 0.4163 0.9251 0.9618
No log 4.1882 356 0.9489 0.4226 0.9489 0.9741
No log 4.2118 358 0.9592 0.4466 0.9592 0.9794
No log 4.2353 360 0.9950 0.4797 0.9950 0.9975
No log 4.2588 362 0.9485 0.4125 0.9485 0.9739
No log 4.2824 364 0.9028 0.3437 0.9028 0.9502
No log 4.3059 366 1.0167 0.3602 1.0167 1.0083
No log 4.3294 368 1.0456 0.3613 1.0456 1.0225
No log 4.3529 370 0.9421 0.3626 0.9421 0.9706
No log 4.3765 372 0.9156 0.3979 0.9156 0.9569
No log 4.4 374 0.9955 0.4040 0.9955 0.9978
No log 4.4235 376 0.9971 0.4040 0.9971 0.9985
No log 4.4471 378 0.9799 0.4165 0.9799 0.9899
No log 4.4706 380 0.9872 0.4025 0.9872 0.9936
No log 4.4941 382 0.9670 0.3820 0.9670 0.9834
No log 4.5176 384 0.9591 0.3142 0.9591 0.9793
No log 4.5412 386 0.9563 0.2993 0.9563 0.9779
No log 4.5647 388 0.9468 0.3437 0.9468 0.9730
No log 4.5882 390 0.9320 0.3337 0.9320 0.9654
No log 4.6118 392 0.9227 0.3671 0.9227 0.9606
No log 4.6353 394 0.9102 0.3671 0.9102 0.9541
No log 4.6588 396 0.8647 0.4617 0.8647 0.9299
No log 4.6824 398 0.9032 0.4526 0.9032 0.9504
No log 4.7059 400 0.9811 0.3902 0.9811 0.9905
No log 4.7294 402 0.9734 0.3517 0.9734 0.9866
No log 4.7529 404 0.9172 0.3298 0.9172 0.9577
No log 4.7765 406 0.8760 0.3539 0.8760 0.9360
No log 4.8 408 0.8719 0.3685 0.8719 0.9338
No log 4.8235 410 0.8790 0.3539 0.8790 0.9375
No log 4.8471 412 0.9251 0.3564 0.9251 0.9618
No log 4.8706 414 1.0060 0.4186 1.0060 1.0030
No log 4.8941 416 1.0580 0.3730 1.0580 1.0286
No log 4.9176 418 1.0161 0.4186 1.0161 1.0080
No log 4.9412 420 0.9235 0.3448 0.9235 0.9610
No log 4.9647 422 0.8744 0.4466 0.8744 0.9351
No log 4.9882 424 0.8958 0.3590 0.8958 0.9465
No log 5.0118 426 0.8838 0.3693 0.8838 0.9401
No log 5.0353 428 0.8576 0.3685 0.8576 0.9261
No log 5.0588 430 0.9193 0.4439 0.9193 0.9588
No log 5.0824 432 1.0046 0.3730 1.0046 1.0023
No log 5.1059 434 1.0129 0.3730 1.0129 1.0065
No log 5.1294 436 0.9701 0.3730 0.9701 0.9850
No log 5.1529 438 0.9200 0.3645 0.9200 0.9592
No log 5.1765 440 0.8522 0.3787 0.8522 0.9231
No log 5.2 442 0.8231 0.4608 0.8231 0.9072
No log 5.2235 444 0.8316 0.4297 0.8316 0.9119
No log 5.2471 446 0.8319 0.4220 0.8319 0.9121
No log 5.2706 448 0.8576 0.3938 0.8576 0.9261
No log 5.2941 450 0.9710 0.4423 0.9710 0.9854
No log 5.3176 452 1.0298 0.4423 1.0298 1.0148
No log 5.3412 454 1.0066 0.4423 1.0066 1.0033
No log 5.3647 456 0.9514 0.3306 0.9514 0.9754
No log 5.3882 458 0.9205 0.3278 0.9205 0.9594
No log 5.4118 460 0.8924 0.3404 0.8924 0.9447
No log 5.4353 462 0.8816 0.3539 0.8816 0.9389
No log 5.4588 464 0.8735 0.3998 0.8735 0.9346
No log 5.4824 466 0.8256 0.4418 0.8256 0.9087
No log 5.5059 468 0.8055 0.4599 0.8055 0.8975
No log 5.5294 470 0.8191 0.3938 0.8191 0.9050
No log 5.5529 472 0.8216 0.4077 0.8216 0.9064
No log 5.5765 474 0.8145 0.4116 0.8145 0.9025
No log 5.6 476 0.8045 0.4116 0.8045 0.8969
No log 5.6235 478 0.7956 0.4737 0.7956 0.8919
No log 5.6471 480 0.7844 0.5061 0.7844 0.8856
No log 5.6706 482 0.8016 0.5094 0.8016 0.8953
No log 5.6941 484 0.8267 0.5192 0.8267 0.9093
No log 5.7176 486 0.8268 0.5192 0.8268 0.9093
No log 5.7412 488 0.8193 0.4519 0.8193 0.9052
No log 5.7647 490 0.8290 0.3938 0.8290 0.9105
No log 5.7882 492 0.8353 0.4701 0.8353 0.9139
No log 5.8118 494 0.8474 0.4261 0.8474 0.9205
No log 5.8353 496 0.8685 0.4198 0.8685 0.9319
No log 5.8588 498 0.8690 0.4406 0.8690 0.9322
0.3797 5.8824 500 0.8576 0.4221 0.8576 0.9261
0.3797 5.9059 502 0.8810 0.3525 0.8810 0.9386
0.3797 5.9294 504 0.8945 0.3583 0.8945 0.9458
0.3797 5.9529 506 0.8986 0.3689 0.8986 0.9479
0.3797 5.9765 508 0.8912 0.3804 0.8912 0.9440
0.3797 6.0 510 0.8513 0.3787 0.8513 0.9227

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k17_task2_organization

Finetuned
(4019)
this model