ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k3_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0690
  • Qwk: 0.4327
  • Mse: 1.0690
  • Rmse: 1.0339

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1667 2 4.6909 0.0010 4.6909 2.1659
No log 0.3333 4 2.9220 -0.0233 2.9220 1.7094
No log 0.5 6 1.5661 0.0504 1.5661 1.2515
No log 0.6667 8 1.4775 -0.0211 1.4775 1.2155
No log 0.8333 10 1.4456 -0.0911 1.4456 1.2023
No log 1.0 12 1.3068 0.0704 1.3068 1.1431
No log 1.1667 14 1.2509 0.0883 1.2509 1.1184
No log 1.3333 16 1.3051 0.0370 1.3051 1.1424
No log 1.5 18 1.3225 0.0276 1.3225 1.1500
No log 1.6667 20 1.2392 0.1144 1.2392 1.1132
No log 1.8333 22 1.2313 0.0771 1.2313 1.1097
No log 2.0 24 1.2331 0.0722 1.2331 1.1104
No log 2.1667 26 1.2356 0.0414 1.2356 1.1116
No log 2.3333 28 1.1910 0.1616 1.1910 1.0913
No log 2.5 30 1.1836 0.2025 1.1836 1.0879
No log 2.6667 32 1.1406 0.1556 1.1406 1.0680
No log 2.8333 34 1.0829 0.2369 1.0829 1.0406
No log 3.0 36 1.0741 0.3216 1.0741 1.0364
No log 3.1667 38 1.1065 0.3418 1.1065 1.0519
No log 3.3333 40 1.0270 0.3821 1.0270 1.0134
No log 3.5 42 1.0123 0.2995 1.0123 1.0061
No log 3.6667 44 1.1710 0.3844 1.1710 1.0821
No log 3.8333 46 1.1656 0.4647 1.1656 1.0796
No log 4.0 48 1.1314 0.3893 1.1314 1.0637
No log 4.1667 50 1.1796 0.4064 1.1796 1.0861
No log 4.3333 52 1.1176 0.3857 1.1176 1.0572
No log 4.5 54 1.2374 0.3930 1.2374 1.1124
No log 4.6667 56 1.2709 0.3953 1.2709 1.1273
No log 4.8333 58 1.2951 0.3798 1.2951 1.1380
No log 5.0 60 1.0875 0.3913 1.0875 1.0428
No log 5.1667 62 1.2853 0.2316 1.2853 1.1337
No log 5.3333 64 1.2153 0.2356 1.2153 1.1024
No log 5.5 66 0.9978 0.5089 0.9978 0.9989
No log 5.6667 68 0.9915 0.4314 0.9915 0.9957
No log 5.8333 70 0.9981 0.4780 0.9981 0.9990
No log 6.0 72 0.8981 0.4681 0.8981 0.9477
No log 6.1667 74 0.8726 0.4241 0.8726 0.9341
No log 6.3333 76 0.8723 0.4690 0.8723 0.9340
No log 6.5 78 0.8731 0.4672 0.8731 0.9344
No log 6.6667 80 0.9013 0.4852 0.9013 0.9494
No log 6.8333 82 0.9860 0.5246 0.9860 0.9930
No log 7.0 84 0.9521 0.5091 0.9521 0.9758
No log 7.1667 86 0.9018 0.5262 0.9018 0.9496
No log 7.3333 88 0.9635 0.5146 0.9635 0.9816
No log 7.5 90 0.9149 0.5492 0.9149 0.9565
No log 7.6667 92 0.8867 0.5492 0.8867 0.9417
No log 7.8333 94 0.8719 0.5223 0.8719 0.9338
No log 8.0 96 0.8935 0.4700 0.8935 0.9452
No log 8.1667 98 1.0199 0.4492 1.0199 1.0099
No log 8.3333 100 0.9480 0.4096 0.9480 0.9737
No log 8.5 102 0.8285 0.4439 0.8285 0.9102
No log 8.6667 104 0.8018 0.4845 0.8018 0.8954
No log 8.8333 106 0.7994 0.4519 0.7994 0.8941
No log 9.0 108 0.8190 0.4979 0.8190 0.9050
No log 9.1667 110 0.9266 0.4510 0.9266 0.9626
No log 9.3333 112 0.9631 0.4623 0.9631 0.9814
No log 9.5 114 1.0103 0.4623 1.0103 1.0051
No log 9.6667 116 0.9010 0.4907 0.9010 0.9492
No log 9.8333 118 0.8803 0.5131 0.8803 0.9383
No log 10.0 120 0.8665 0.4318 0.8665 0.9309
No log 10.1667 122 0.8696 0.4611 0.8696 0.9325
No log 10.3333 124 0.8873 0.4202 0.8873 0.9420
No log 10.5 126 0.8920 0.4202 0.8920 0.9444
No log 10.6667 128 0.9027 0.4074 0.9027 0.9501
No log 10.8333 130 0.9326 0.4115 0.9326 0.9657
No log 11.0 132 0.8965 0.4114 0.8965 0.9468
No log 11.1667 134 0.8694 0.4792 0.8694 0.9324
No log 11.3333 136 0.8737 0.5023 0.8737 0.9347
No log 11.5 138 0.9395 0.5210 0.9395 0.9693
No log 11.6667 140 0.9613 0.4524 0.9613 0.9804
No log 11.8333 142 0.9131 0.4483 0.9131 0.9556
No log 12.0 144 0.9203 0.4591 0.9203 0.9593
No log 12.1667 146 0.9187 0.4540 0.9187 0.9585
No log 12.3333 148 0.9047 0.4277 0.9047 0.9511
No log 12.5 150 0.9912 0.3722 0.9912 0.9956
No log 12.6667 152 1.0094 0.4133 1.0094 1.0047
No log 12.8333 154 1.0495 0.3932 1.0495 1.0244
No log 13.0 156 0.9917 0.4515 0.9917 0.9958
No log 13.1667 158 1.0213 0.4231 1.0213 1.0106
No log 13.3333 160 1.0823 0.4489 1.0823 1.0403
No log 13.5 162 1.2149 0.4625 1.2149 1.1022
No log 13.6667 164 1.2469 0.4320 1.2469 1.1167
No log 13.8333 166 1.0683 0.4681 1.0683 1.0336
No log 14.0 168 0.9104 0.5038 0.9104 0.9542
No log 14.1667 170 0.8799 0.5082 0.8799 0.9380
No log 14.3333 172 0.8638 0.4894 0.8638 0.9294
No log 14.5 174 0.8665 0.4146 0.8665 0.9308
No log 14.6667 176 0.8558 0.4493 0.8558 0.9251
No log 14.8333 178 0.8326 0.5735 0.8326 0.9125
No log 15.0 180 0.8320 0.5483 0.8320 0.9121
No log 15.1667 182 0.8791 0.4527 0.8791 0.9376
No log 15.3333 184 0.9152 0.4200 0.9152 0.9566
No log 15.5 186 0.8980 0.4080 0.8980 0.9476
No log 15.6667 188 0.8632 0.4681 0.8632 0.9291
No log 15.8333 190 0.8782 0.4546 0.8782 0.9371
No log 16.0 192 0.8630 0.4889 0.8630 0.9290
No log 16.1667 194 0.8521 0.5365 0.8521 0.9231
No log 16.3333 196 0.9024 0.4038 0.9024 0.9500
No log 16.5 198 1.0091 0.4665 1.0091 1.0045
No log 16.6667 200 1.0007 0.4665 1.0007 1.0003
No log 16.8333 202 0.9808 0.4359 0.9808 0.9904
No log 17.0 204 1.0171 0.4359 1.0171 1.0085
No log 17.1667 206 1.0732 0.5080 1.0732 1.0359
No log 17.3333 208 0.9534 0.4359 0.9534 0.9764
No log 17.5 210 0.8266 0.4907 0.8266 0.9091
No log 17.6667 212 0.8006 0.6084 0.8006 0.8948
No log 17.8333 214 0.8189 0.4984 0.8189 0.9049
No log 18.0 216 0.8903 0.4314 0.8903 0.9435
No log 18.1667 218 0.8905 0.4400 0.8905 0.9436
No log 18.3333 220 0.8856 0.4232 0.8856 0.9411
No log 18.5 222 0.9048 0.4232 0.9048 0.9512
No log 18.6667 224 0.8601 0.4559 0.8601 0.9274
No log 18.8333 226 0.8264 0.5528 0.8264 0.9091
No log 19.0 228 0.8363 0.5403 0.8363 0.9145
No log 19.1667 230 0.8909 0.4412 0.8909 0.9439
No log 19.3333 232 0.9656 0.4479 0.9656 0.9826
No log 19.5 234 0.9390 0.4510 0.9390 0.9690
No log 19.6667 236 0.8526 0.5254 0.8526 0.9233
No log 19.8333 238 0.9029 0.4655 0.9029 0.9502
No log 20.0 240 0.8918 0.4655 0.8918 0.9444
No log 20.1667 242 0.8399 0.5054 0.8399 0.9165
No log 20.3333 244 0.9068 0.4347 0.9068 0.9523
No log 20.5 246 1.0833 0.4950 1.0833 1.0408
No log 20.6667 248 1.1897 0.4083 1.1897 1.0907
No log 20.8333 250 1.1216 0.4261 1.1216 1.0591
No log 21.0 252 0.9729 0.4717 0.9729 0.9864
No log 21.1667 254 0.8597 0.4685 0.8597 0.9272
No log 21.3333 256 0.8288 0.5248 0.8288 0.9104
No log 21.5 258 0.8855 0.4534 0.8855 0.9410
No log 21.6667 260 0.9482 0.5040 0.9482 0.9738
No log 21.8333 262 1.0786 0.5055 1.0786 1.0385
No log 22.0 264 1.0560 0.4813 1.0560 1.0276
No log 22.1667 266 0.9365 0.4284 0.9365 0.9677
No log 22.3333 268 0.8957 0.4724 0.8957 0.9464
No log 22.5 270 0.9285 0.4378 0.9285 0.9636
No log 22.6667 272 0.9124 0.4648 0.9124 0.9552
No log 22.8333 274 0.9123 0.3872 0.9123 0.9551
No log 23.0 276 0.9295 0.4284 0.9295 0.9641
No log 23.1667 278 0.9103 0.4510 0.9103 0.9541
No log 23.3333 280 0.8987 0.4909 0.8987 0.9480
No log 23.5 282 0.8685 0.5024 0.8685 0.9320
No log 23.6667 284 0.8230 0.5671 0.8230 0.9072
No log 23.8333 286 0.8279 0.5443 0.8279 0.9099
No log 24.0 288 0.8380 0.5438 0.8380 0.9154
No log 24.1667 290 0.8577 0.4934 0.8577 0.9261
No log 24.3333 292 0.9754 0.3758 0.9754 0.9876
No log 24.5 294 1.0474 0.4329 1.0474 1.0234
No log 24.6667 296 0.9742 0.4677 0.9742 0.9870
No log 24.8333 298 0.8713 0.4579 0.8713 0.9334
No log 25.0 300 0.8438 0.4934 0.8438 0.9186
No log 25.1667 302 0.8481 0.5043 0.8481 0.9209
No log 25.3333 304 0.9313 0.4638 0.9313 0.9651
No log 25.5 306 1.0298 0.4169 1.0298 1.0148
No log 25.6667 308 1.0661 0.4295 1.0661 1.0325
No log 25.8333 310 1.1055 0.3693 1.1055 1.0514
No log 26.0 312 1.1162 0.3995 1.1162 1.0565
No log 26.1667 314 1.1188 0.3995 1.1188 1.0578
No log 26.3333 316 1.1081 0.3798 1.1081 1.0527
No log 26.5 318 1.0214 0.4186 1.0214 1.0106
No log 26.6667 320 0.9072 0.4583 0.9072 0.9525
No log 26.8333 322 0.7996 0.5596 0.7996 0.8942
No log 27.0 324 0.7639 0.5732 0.7639 0.8740
No log 27.1667 326 0.7669 0.5732 0.7669 0.8757
No log 27.3333 328 0.7933 0.5741 0.7933 0.8907
No log 27.5 330 0.9024 0.5040 0.9024 0.9500
No log 27.6667 332 0.9555 0.5040 0.9555 0.9775
No log 27.8333 334 0.9320 0.5040 0.9320 0.9654
No log 28.0 336 0.8291 0.5864 0.8291 0.9105
No log 28.1667 338 0.8006 0.5302 0.8006 0.8948
No log 28.3333 340 0.7997 0.5489 0.7997 0.8943
No log 28.5 342 0.8061 0.5321 0.8061 0.8978
No log 28.6667 344 0.8076 0.5333 0.8076 0.8986
No log 28.8333 346 0.7977 0.5268 0.7977 0.8931
No log 29.0 348 0.8040 0.5079 0.8040 0.8967
No log 29.1667 350 0.8064 0.5655 0.8064 0.8980
No log 29.3333 352 0.8325 0.5047 0.8325 0.9124
No log 29.5 354 0.8497 0.5321 0.8497 0.9218
No log 29.6667 356 0.8534 0.5458 0.8534 0.9238
No log 29.8333 358 0.8522 0.5158 0.8522 0.9231
No log 30.0 360 0.8651 0.4914 0.8651 0.9301
No log 30.1667 362 0.9479 0.4987 0.9479 0.9736
No log 30.3333 364 0.9761 0.4681 0.9761 0.9880
No log 30.5 366 0.9405 0.4504 0.9405 0.9698
No log 30.6667 368 0.8937 0.4869 0.8937 0.9454
No log 30.8333 370 0.8935 0.4590 0.8935 0.9452
No log 31.0 372 0.8926 0.4898 0.8926 0.9448
No log 31.1667 374 0.9435 0.5345 0.9435 0.9713
No log 31.3333 376 1.0895 0.4658 1.0895 1.0438
No log 31.5 378 1.1835 0.4805 1.1835 1.0879
No log 31.6667 380 1.0971 0.4743 1.0971 1.0474
No log 31.8333 382 0.9369 0.4745 0.9369 0.9679
No log 32.0 384 0.8791 0.5696 0.8791 0.9376
No log 32.1667 386 0.8595 0.5344 0.8595 0.9271
No log 32.3333 388 0.8595 0.4984 0.8595 0.9271
No log 32.5 390 0.9148 0.4600 0.9148 0.9565
No log 32.6667 392 0.9443 0.4958 0.9443 0.9717
No log 32.8333 394 0.9260 0.4790 0.9260 0.9623
No log 33.0 396 0.8533 0.4764 0.8533 0.9237
No log 33.1667 398 0.8136 0.5315 0.8136 0.9020
No log 33.3333 400 0.8046 0.5315 0.8046 0.8970
No log 33.5 402 0.8173 0.5211 0.8173 0.9041
No log 33.6667 404 0.8950 0.5066 0.8950 0.9460
No log 33.8333 406 0.9343 0.4924 0.9343 0.9666
No log 34.0 408 0.9034 0.4770 0.9034 0.9505
No log 34.1667 410 0.8383 0.4829 0.8383 0.9156
No log 34.3333 412 0.8062 0.4969 0.8062 0.8979
No log 34.5 414 0.8278 0.5058 0.8278 0.9098
No log 34.6667 416 0.8743 0.5114 0.8743 0.9350
No log 34.8333 418 0.9541 0.5403 0.9541 0.9768
No log 35.0 420 0.9682 0.5384 0.9682 0.9839
No log 35.1667 422 0.9252 0.5166 0.9252 0.9619
No log 35.3333 424 0.9049 0.4986 0.9049 0.9512
No log 35.5 426 0.9166 0.4986 0.9166 0.9574
No log 35.6667 428 0.8822 0.4902 0.8822 0.9393
No log 35.8333 430 0.8385 0.5041 0.8385 0.9157
No log 36.0 432 0.8144 0.5102 0.8144 0.9024
No log 36.1667 434 0.7944 0.5896 0.7944 0.8913
No log 36.3333 436 0.7884 0.5972 0.7884 0.8879
No log 36.5 438 0.7917 0.5669 0.7917 0.8898
No log 36.6667 440 0.7981 0.5895 0.7981 0.8934
No log 36.8333 442 0.8172 0.5318 0.8172 0.9040
No log 37.0 444 0.8565 0.4754 0.8565 0.9255
No log 37.1667 446 0.8736 0.4754 0.8736 0.9347
No log 37.3333 448 0.8543 0.4754 0.8543 0.9243
No log 37.5 450 0.8189 0.5358 0.8189 0.9049
No log 37.6667 452 0.8163 0.5877 0.8163 0.9035
No log 37.8333 454 0.8238 0.6117 0.8238 0.9076
No log 38.0 456 0.8307 0.5785 0.8307 0.9114
No log 38.1667 458 0.8403 0.4681 0.8403 0.9167
No log 38.3333 460 0.8732 0.4334 0.8732 0.9345
No log 38.5 462 0.9011 0.4639 0.9011 0.9492
No log 38.6667 464 0.8851 0.4764 0.8851 0.9408
No log 38.8333 466 0.8709 0.4337 0.8709 0.9332
No log 39.0 468 0.8564 0.4242 0.8564 0.9254
No log 39.1667 470 0.8449 0.3998 0.8449 0.9192
No log 39.3333 472 0.8431 0.4242 0.8431 0.9182
No log 39.5 474 0.8807 0.4425 0.8807 0.9385
No log 39.6667 476 0.9680 0.4563 0.9680 0.9839
No log 39.8333 478 1.0129 0.5029 1.0129 1.0064
No log 40.0 480 0.9661 0.4563 0.9661 0.9829
No log 40.1667 482 0.8763 0.4412 0.8763 0.9361
No log 40.3333 484 0.8177 0.4916 0.8177 0.9043
No log 40.5 486 0.8151 0.4916 0.8151 0.9029
No log 40.6667 488 0.8430 0.4775 0.8430 0.9181
No log 40.8333 490 0.9049 0.4811 0.9049 0.9513
No log 41.0 492 0.9410 0.4771 0.9410 0.9701
No log 41.1667 494 0.9582 0.4251 0.9582 0.9789
No log 41.3333 496 0.9441 0.4576 0.9441 0.9716
No log 41.5 498 0.8886 0.4845 0.8886 0.9427
0.2691 41.6667 500 0.8238 0.5013 0.8238 0.9076
0.2691 41.8333 502 0.7791 0.4998 0.7791 0.8827
0.2691 42.0 504 0.7585 0.5660 0.7585 0.8709
0.2691 42.1667 506 0.7604 0.6023 0.7604 0.8720
0.2691 42.3333 508 0.7652 0.6324 0.7652 0.8747
0.2691 42.5 510 0.7701 0.6142 0.7701 0.8776
0.2691 42.6667 512 0.7940 0.5365 0.7940 0.8910
0.2691 42.8333 514 0.8126 0.5426 0.8126 0.9014
0.2691 43.0 516 0.8025 0.5447 0.8025 0.8958
0.2691 43.1667 518 0.7828 0.5073 0.7828 0.8848
0.2691 43.3333 520 0.7663 0.5112 0.7663 0.8754
0.2691 43.5 522 0.7679 0.5012 0.7679 0.8763
0.2691 43.6667 524 0.7688 0.5012 0.7688 0.8768
0.2691 43.8333 526 0.7676 0.5012 0.7676 0.8761
0.2691 44.0 528 0.7777 0.5543 0.7777 0.8818
0.2691 44.1667 530 0.8080 0.5447 0.8080 0.8989
0.2691 44.3333 532 0.8138 0.5426 0.8138 0.9021
0.2691 44.5 534 0.7813 0.5673 0.7813 0.8839
0.2691 44.6667 536 0.7606 0.6108 0.7606 0.8721
0.2691 44.8333 538 0.7947 0.5749 0.7947 0.8915
0.2691 45.0 540 0.8237 0.6130 0.8237 0.9076
0.2691 45.1667 542 0.8337 0.5811 0.8337 0.9130
0.2691 45.3333 544 0.8238 0.6143 0.8238 0.9076
0.2691 45.5 546 0.8122 0.5143 0.8122 0.9012
0.2691 45.6667 548 0.8227 0.4852 0.8227 0.9071
0.2691 45.8333 550 0.8563 0.4864 0.8563 0.9254
0.2691 46.0 552 0.9267 0.4790 0.9267 0.9626
0.2691 46.1667 554 1.0321 0.4632 1.0321 1.0159
0.2691 46.3333 556 1.0690 0.4327 1.0690 1.0339

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k3_task2_organization

Finetuned
(4023)
this model