ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k20_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9029
  • Qwk: 0.4051
  • Mse: 0.9029
  • Rmse: 0.9502

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0328 2 4.7482 0.0018 4.7482 2.1790
No log 0.0656 4 2.9283 -0.0029 2.9283 1.7112
No log 0.0984 6 2.0087 0.0062 2.0087 1.4173
No log 0.1311 8 1.6917 0.0062 1.6917 1.3006
No log 0.1639 10 1.4433 -0.0143 1.4433 1.2014
No log 0.1967 12 1.2982 0.0509 1.2982 1.1394
No log 0.2295 14 1.3475 0.0362 1.3475 1.1608
No log 0.2623 16 1.4387 -0.0149 1.4387 1.1994
No log 0.2951 18 1.3303 0.0169 1.3303 1.1534
No log 0.3279 20 1.2713 0.0600 1.2713 1.1275
No log 0.3607 22 1.1749 0.2632 1.1749 1.0840
No log 0.3934 24 1.1438 0.3250 1.1438 1.0695
No log 0.4262 26 1.1708 0.2485 1.1708 1.0820
No log 0.4590 28 1.3060 0.0512 1.3060 1.1428
No log 0.4918 30 1.6309 0.0403 1.6309 1.2771
No log 0.5246 32 1.8393 0.0 1.8393 1.3562
No log 0.5574 34 1.7742 0.0 1.7742 1.3320
No log 0.5902 36 1.4105 0.0811 1.4105 1.1876
No log 0.6230 38 1.1393 0.3140 1.1393 1.0674
No log 0.6557 40 1.0924 0.2674 1.0924 1.0452
No log 0.6885 42 1.1492 0.3095 1.1492 1.0720
No log 0.7213 44 1.6081 0.1051 1.6081 1.2681
No log 0.7541 46 1.8438 0.0284 1.8438 1.3579
No log 0.7869 48 1.5327 0.0697 1.5327 1.2380
No log 0.8197 50 1.2531 0.2053 1.2531 1.1194
No log 0.8525 52 1.2827 0.1110 1.2827 1.1326
No log 0.8852 54 1.2907 0.0499 1.2907 1.1361
No log 0.9180 56 1.2108 0.2142 1.2108 1.1004
No log 0.9508 58 1.1372 0.2678 1.1372 1.0664
No log 0.9836 60 1.1162 0.2733 1.1162 1.0565
No log 1.0164 62 1.2270 0.1587 1.2270 1.1077
No log 1.0492 64 1.2791 0.1076 1.2791 1.1310
No log 1.0820 66 1.1667 0.1587 1.1667 1.0801
No log 1.1148 68 1.0519 0.4702 1.0519 1.0256
No log 1.1475 70 1.0999 0.2678 1.0999 1.0487
No log 1.1803 72 1.1727 0.2212 1.1727 1.0829
No log 1.2131 74 1.0526 0.3965 1.0526 1.0259
No log 1.2459 76 1.0432 0.3737 1.0432 1.0214
No log 1.2787 78 1.1253 0.3083 1.1253 1.0608
No log 1.3115 80 1.0926 0.3035 1.0926 1.0453
No log 1.3443 82 0.9657 0.4810 0.9657 0.9827
No log 1.3770 84 0.9527 0.2915 0.9527 0.9761
No log 1.4098 86 0.8973 0.3663 0.8973 0.9473
No log 1.4426 88 0.8519 0.4008 0.8519 0.9230
No log 1.4754 90 0.9001 0.4854 0.9001 0.9487
No log 1.5082 92 1.0238 0.4570 1.0238 1.0118
No log 1.5410 94 0.9190 0.4606 0.9190 0.9586
No log 1.5738 96 0.7963 0.4977 0.7963 0.8924
No log 1.6066 98 0.8162 0.4681 0.8162 0.9034
No log 1.6393 100 0.8282 0.4197 0.8282 0.9101
No log 1.6721 102 1.0973 0.4478 1.0973 1.0475
No log 1.7049 104 1.3827 0.2442 1.3827 1.1759
No log 1.7377 106 1.1998 0.3939 1.1998 1.0954
No log 1.7705 108 0.8255 0.4334 0.8255 0.9086
No log 1.8033 110 0.8221 0.5727 0.8221 0.9067
No log 1.8361 112 1.0022 0.4453 1.0022 1.0011
No log 1.8689 114 0.9301 0.5507 0.9301 0.9644
No log 1.9016 116 0.8048 0.4801 0.8048 0.8971
No log 1.9344 118 0.9330 0.4507 0.9330 0.9659
No log 1.9672 120 1.0035 0.5023 1.0035 1.0017
No log 2.0 122 0.9255 0.4570 0.9255 0.9621
No log 2.0328 124 0.8648 0.4603 0.8648 0.9300
No log 2.0656 126 1.0283 0.4430 1.0283 1.0140
No log 2.0984 128 1.0318 0.4221 1.0318 1.0158
No log 2.1311 130 0.9145 0.3275 0.9145 0.9563
No log 2.1639 132 0.9016 0.3902 0.9016 0.9495
No log 2.1967 134 0.9215 0.3902 0.9215 0.9599
No log 2.2295 136 0.9552 0.3427 0.9552 0.9773
No log 2.2623 138 0.9545 0.3779 0.9545 0.9770
No log 2.2951 140 1.0403 0.4048 1.0403 1.0199
No log 2.3279 142 1.1542 0.4186 1.1542 1.0743
No log 2.3607 144 1.0535 0.4219 1.0535 1.0264
No log 2.3934 146 0.9060 0.3896 0.9060 0.9519
No log 2.4262 148 0.8785 0.4079 0.8785 0.9373
No log 2.4590 150 0.9087 0.4164 0.9087 0.9533
No log 2.4918 152 0.9232 0.3517 0.9232 0.9608
No log 2.5246 154 0.8809 0.4302 0.8809 0.9386
No log 2.5574 156 0.8847 0.3660 0.8847 0.9406
No log 2.5902 158 0.8933 0.3291 0.8933 0.9452
No log 2.6230 160 0.8518 0.4039 0.8518 0.9229
No log 2.6557 162 0.8549 0.4002 0.8549 0.9246
No log 2.6885 164 0.8448 0.3970 0.8448 0.9191
No log 2.7213 166 0.9247 0.3646 0.9247 0.9616
No log 2.7541 168 1.1068 0.3822 1.1068 1.0520
No log 2.7869 170 0.9467 0.4175 0.9467 0.9730
No log 2.8197 172 0.8374 0.4628 0.8374 0.9151
No log 2.8525 174 0.9696 0.5507 0.9696 0.9847
No log 2.8852 176 1.1439 0.4666 1.1439 1.0695
No log 2.9180 178 1.2789 0.3839 1.2789 1.1309
No log 2.9508 180 1.1188 0.4516 1.1188 1.0577
No log 2.9836 182 0.8989 0.4644 0.8989 0.9481
No log 3.0164 184 0.9329 0.3806 0.9329 0.9658
No log 3.0492 186 0.9649 0.3709 0.9649 0.9823
No log 3.0820 188 0.8881 0.4054 0.8881 0.9424
No log 3.1148 190 0.8551 0.4780 0.8551 0.9247
No log 3.1475 192 0.8852 0.4439 0.8852 0.9409
No log 3.1803 194 0.9246 0.4302 0.9246 0.9616
No log 3.2131 196 0.8763 0.4300 0.8763 0.9361
No log 3.2459 198 0.8348 0.4845 0.8348 0.9137
No log 3.2787 200 0.8063 0.4220 0.8063 0.8979
No log 3.3115 202 0.8089 0.3821 0.8089 0.8994
No log 3.3443 204 0.8255 0.4598 0.8255 0.9085
No log 3.3770 206 0.8066 0.4159 0.8066 0.8981
No log 3.4098 208 0.8268 0.4711 0.8268 0.9093
No log 3.4426 210 0.9171 0.4402 0.9171 0.9576
No log 3.4754 212 0.9142 0.4402 0.9142 0.9561
No log 3.5082 214 0.8073 0.5028 0.8073 0.8985
No log 3.5410 216 0.8635 0.4379 0.8635 0.9293
No log 3.5738 218 0.9031 0.4285 0.9031 0.9503
No log 3.6066 220 0.9333 0.3955 0.9333 0.9661
No log 3.6393 222 0.8304 0.4363 0.8304 0.9113
No log 3.6721 224 0.8749 0.4420 0.8749 0.9354
No log 3.7049 226 0.8613 0.4104 0.8613 0.9281
No log 3.7377 228 0.8447 0.4324 0.8447 0.9191
No log 3.7705 230 0.8403 0.4221 0.8403 0.9167
No log 3.8033 232 0.8526 0.3822 0.8526 0.9234
No log 3.8361 234 0.8647 0.3271 0.8647 0.9299
No log 3.8689 236 0.9380 0.4386 0.9380 0.9685
No log 3.9016 238 1.0855 0.3763 1.0855 1.0419
No log 3.9344 240 1.1335 0.3763 1.1335 1.0647
No log 3.9672 242 1.1513 0.4358 1.1513 1.0730
No log 4.0 244 1.0418 0.4430 1.0418 1.0207
No log 4.0328 246 0.8732 0.3847 0.8732 0.9345
No log 4.0656 248 0.8510 0.3603 0.8510 0.9225
No log 4.0984 250 0.9452 0.3927 0.9452 0.9722
No log 4.1311 252 0.9136 0.3884 0.9136 0.9558
No log 4.1639 254 0.8612 0.3498 0.8612 0.9280
No log 4.1967 256 0.8899 0.3700 0.8899 0.9433
No log 4.2295 258 0.9125 0.3961 0.9125 0.9552
No log 4.2623 260 0.8741 0.4282 0.8741 0.9350
No log 4.2951 262 0.8552 0.4180 0.8552 0.9248
No log 4.3279 264 0.8337 0.3873 0.8337 0.9131
No log 4.3607 266 0.8185 0.4220 0.8185 0.9047
No log 4.3934 268 0.8194 0.3787 0.8194 0.9052
No log 4.4262 270 0.9148 0.4297 0.9148 0.9564
No log 4.4590 272 1.0858 0.3814 1.0858 1.0420
No log 4.4918 274 1.0210 0.3460 1.0210 1.0104
No log 4.5246 276 0.8664 0.3326 0.8664 0.9308
No log 4.5574 278 0.8431 0.4051 0.8431 0.9182
No log 4.5902 280 0.8506 0.4379 0.8506 0.9223
No log 4.6230 282 0.8025 0.4406 0.8025 0.8958
No log 4.6557 284 0.7983 0.4282 0.7983 0.8935
No log 4.6885 286 0.8417 0.4202 0.8417 0.9174
No log 4.7213 288 0.8403 0.4202 0.8403 0.9167
No log 4.7541 290 0.8059 0.3769 0.8059 0.8977
No log 4.7869 292 0.8522 0.4833 0.8522 0.9231
No log 4.8197 294 0.8476 0.4263 0.8476 0.9206
No log 4.8525 296 0.8505 0.2782 0.8505 0.9222
No log 4.8852 298 0.9129 0.2944 0.9129 0.9554
No log 4.9180 300 1.0908 0.275 1.0908 1.0444
No log 4.9508 302 1.1219 0.3115 1.1219 1.0592
No log 4.9836 304 0.9818 0.3389 0.9818 0.9908
No log 5.0164 306 0.8500 0.3094 0.8500 0.9220
No log 5.0492 308 0.7933 0.4541 0.7933 0.8907
No log 5.0820 310 0.7735 0.4772 0.7735 0.8795
No log 5.1148 312 0.7688 0.4142 0.7688 0.8768
No log 5.1475 314 0.9599 0.5002 0.9599 0.9797
No log 5.1803 316 1.1219 0.5016 1.1219 1.0592
No log 5.2131 318 1.0137 0.5174 1.0137 1.0068
No log 5.2459 320 0.7953 0.5220 0.7953 0.8918
No log 5.2787 322 0.7530 0.5676 0.7530 0.8678
No log 5.3115 324 0.9303 0.4449 0.9303 0.9645
No log 5.3443 326 0.9041 0.4516 0.9041 0.9508
No log 5.3770 328 0.7793 0.4990 0.7793 0.8828
No log 5.4098 330 0.7640 0.4423 0.7640 0.8741
No log 5.4426 332 0.8488 0.5124 0.8488 0.9213
No log 5.4754 334 0.8445 0.4916 0.8445 0.9189
No log 5.5082 336 0.7767 0.4853 0.7767 0.8813
No log 5.5410 338 0.8028 0.4728 0.8028 0.8960
No log 5.5738 340 0.7971 0.4728 0.7971 0.8928
No log 5.6066 342 0.7768 0.4466 0.7768 0.8814
No log 5.6393 344 0.8284 0.4811 0.8284 0.9102
No log 5.6721 346 0.8537 0.5014 0.8537 0.9240
No log 5.7049 348 0.8211 0.4243 0.8211 0.9061
No log 5.7377 350 0.7803 0.4286 0.7803 0.8833
No log 5.7705 352 0.7713 0.3822 0.7713 0.8782
No log 5.8033 354 0.8034 0.4488 0.8034 0.8963
No log 5.8361 356 0.9002 0.4681 0.9002 0.9488
No log 5.8689 358 0.9890 0.4221 0.9890 0.9945
No log 5.9016 360 0.9788 0.3299 0.9788 0.9893
No log 5.9344 362 0.9202 0.3434 0.9202 0.9593
No log 5.9672 364 0.8597 0.3616 0.8597 0.9272
No log 6.0 366 0.8776 0.3909 0.8776 0.9368
No log 6.0328 368 0.8767 0.3909 0.8767 0.9363
No log 6.0656 370 0.8655 0.3478 0.8655 0.9303
No log 6.0984 372 0.9553 0.3186 0.9553 0.9774
No log 6.1311 374 1.0362 0.3657 1.0362 1.0179
No log 6.1639 376 1.0097 0.3625 1.0097 1.0049
No log 6.1967 378 0.9107 0.3147 0.9107 0.9543
No log 6.2295 380 0.8668 0.3674 0.8668 0.9310
No log 6.2623 382 0.8466 0.3859 0.8466 0.9201
No log 6.2951 384 0.8180 0.4261 0.8180 0.9045
No log 6.3279 386 0.7929 0.4075 0.7929 0.8905
No log 6.3607 388 0.7960 0.3641 0.7960 0.8922
No log 6.3934 390 0.8004 0.3641 0.8004 0.8947
No log 6.4262 392 0.8065 0.4221 0.8065 0.8980
No log 6.4590 394 0.8465 0.4198 0.8465 0.9200
No log 6.4918 396 0.8653 0.4051 0.8653 0.9302
No log 6.5246 398 0.9018 0.3946 0.9018 0.9496
No log 6.5574 400 0.9744 0.3635 0.9744 0.9871
No log 6.5902 402 0.9314 0.3648 0.9314 0.9651
No log 6.6230 404 0.8951 0.3195 0.8951 0.9461
No log 6.6557 406 0.8836 0.3744 0.8836 0.9400
No log 6.6885 408 0.8773 0.3861 0.8773 0.9367
No log 6.7213 410 0.8814 0.4202 0.8814 0.9388
No log 6.7541 412 0.8869 0.4743 0.8869 0.9417
No log 6.7869 414 0.8250 0.5308 0.8250 0.9083
No log 6.8197 416 0.7910 0.4930 0.7910 0.8894
No log 6.8525 418 0.7931 0.4930 0.7931 0.8906
No log 6.8852 420 0.8100 0.4916 0.8100 0.9000
No log 6.9180 422 0.8789 0.4774 0.8789 0.9375
No log 6.9508 424 0.8925 0.4165 0.8925 0.9447
No log 6.9836 426 0.8661 0.3093 0.8661 0.9307
No log 7.0164 428 0.8630 0.3191 0.8630 0.9290
No log 7.0492 430 0.8236 0.4656 0.8236 0.9075
No log 7.0820 432 0.8130 0.3708 0.8130 0.9016
No log 7.1148 434 0.8200 0.3753 0.8200 0.9056
No log 7.1475 436 0.8138 0.3970 0.8138 0.9021
No log 7.1803 438 0.8127 0.3970 0.8127 0.9015
No log 7.2131 440 0.8155 0.3708 0.8155 0.9031
No log 7.2459 442 0.8636 0.4051 0.8636 0.9293
No log 7.2787 444 0.9151 0.4051 0.9151 0.9566
No log 7.3115 446 0.8660 0.4051 0.8660 0.9306
No log 7.3443 448 0.8318 0.3708 0.8318 0.9120
No log 7.3770 450 0.8439 0.3674 0.8439 0.9186
No log 7.4098 452 0.8522 0.3674 0.8522 0.9232
No log 7.4426 454 0.8683 0.3902 0.8683 0.9318
No log 7.4754 456 0.9018 0.4483 0.9018 0.9497
No log 7.5082 458 0.8721 0.3708 0.8721 0.9339
No log 7.5410 460 0.8802 0.2605 0.8802 0.9382
No log 7.5738 462 0.9525 0.2850 0.9525 0.9760
No log 7.6066 464 0.9934 0.2955 0.9934 0.9967
No log 7.6393 466 0.9888 0.3201 0.9888 0.9944
No log 7.6721 468 0.9090 0.2409 0.9090 0.9534
No log 7.7049 470 0.8521 0.2944 0.8521 0.9231
No log 7.7377 472 0.8470 0.3674 0.8470 0.9203
No log 7.7705 474 0.8474 0.3674 0.8474 0.9206
No log 7.8033 476 0.8475 0.3933 0.8475 0.9206
No log 7.8361 478 0.8823 0.3256 0.8823 0.9393
No log 7.8689 480 0.9662 0.3478 0.9662 0.9830
No log 7.9016 482 0.9692 0.3478 0.9692 0.9845
No log 7.9344 484 0.9085 0.2608 0.9085 0.9532
No log 7.9672 486 0.8833 0.3117 0.8833 0.9398
No log 8.0 488 0.8677 0.4444 0.8677 0.9315
No log 8.0328 490 0.8837 0.4094 0.8837 0.9401
No log 8.0656 492 0.8915 0.4237 0.8915 0.9442
No log 8.0984 494 0.8641 0.4587 0.8641 0.9296
No log 8.1311 496 0.8585 0.4587 0.8585 0.9266
No log 8.1639 498 0.8553 0.4587 0.8553 0.9248
0.42 8.1967 500 0.8360 0.4321 0.8360 0.9143
0.42 8.2295 502 0.8985 0.2997 0.8985 0.9479
0.42 8.2623 504 0.8806 0.3506 0.8806 0.9384
0.42 8.2951 506 0.8301 0.4461 0.8301 0.9111
0.42 8.3279 508 0.8233 0.4363 0.8233 0.9074
0.42 8.3607 510 0.8201 0.4115 0.8201 0.9056
0.42 8.3934 512 0.7984 0.4461 0.7984 0.8935
0.42 8.4262 514 0.8065 0.4792 0.8065 0.8981
0.42 8.4590 516 0.8260 0.5013 0.8260 0.9088
0.42 8.4918 518 0.7818 0.5426 0.7818 0.8842
0.42 8.5246 520 0.7615 0.4860 0.7615 0.8726
0.42 8.5574 522 0.7994 0.4752 0.7994 0.8941
0.42 8.5902 524 0.8759 0.4604 0.8759 0.9359
0.42 8.6230 526 0.9054 0.4550 0.9054 0.9515
0.42 8.6557 528 0.8975 0.4413 0.8975 0.9473
0.42 8.6885 530 0.8571 0.4406 0.8571 0.9258
0.42 8.7213 532 0.8736 0.3510 0.8736 0.9347
0.42 8.7541 534 0.9045 0.2813 0.9045 0.9510
0.42 8.7869 536 0.9069 0.2520 0.9069 0.9523
0.42 8.8197 538 0.8823 0.3921 0.8823 0.9393
0.42 8.8525 540 0.9029 0.4051 0.9029 0.9502

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k20_task2_organization

Finetuned
(4023)
this model