ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k6_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1299
  • Qwk: 0.1775
  • Mse: 1.1299
  • Rmse: 1.0630

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1176 2 2.7268 -0.0262 2.7268 1.6513
No log 0.2353 4 1.4160 0.0771 1.4160 1.1900
No log 0.3529 6 1.0447 -0.1304 1.0447 1.0221
No log 0.4706 8 1.2835 -0.2437 1.2835 1.1329
No log 0.5882 10 1.5149 -0.2448 1.5149 1.2308
No log 0.7059 12 1.6208 -0.1307 1.6208 1.2731
No log 0.8235 14 1.3323 -0.1803 1.3323 1.1542
No log 0.9412 16 1.0674 0.0327 1.0674 1.0331
No log 1.0588 18 0.9372 0.2227 0.9372 0.9681
No log 1.1765 20 0.9146 0.2504 0.9146 0.9564
No log 1.2941 22 0.8599 0.3425 0.8599 0.9273
No log 1.4118 24 0.8638 0.1407 0.8638 0.9294
No log 1.5294 26 0.9672 0.0 0.9672 0.9834
No log 1.6471 28 1.0828 -0.0660 1.0828 1.0406
No log 1.7647 30 1.0018 -0.0426 1.0018 1.0009
No log 1.8824 32 0.8917 0.0 0.8917 0.9443
No log 2.0 34 0.8040 0.0327 0.8040 0.8967
No log 2.1176 36 0.8083 0.2407 0.8083 0.8990
No log 2.2353 38 0.8258 0.2463 0.8258 0.9087
No log 2.3529 40 0.8547 0.3294 0.8547 0.9245
No log 2.4706 42 0.8759 0.3099 0.8759 0.9359
No log 2.5882 44 0.8845 0.0840 0.8845 0.9405
No log 2.7059 46 0.8996 0.1327 0.8996 0.9485
No log 2.8235 48 0.8628 0.1754 0.8628 0.9288
No log 2.9412 50 0.8310 0.1372 0.8310 0.9116
No log 3.0588 52 0.7444 0.0428 0.7444 0.8628
No log 3.1765 54 0.7285 0.1508 0.7285 0.8535
No log 3.2941 56 0.7495 0.1863 0.7495 0.8657
No log 3.4118 58 0.7724 0.1407 0.7724 0.8788
No log 3.5294 60 0.8269 -0.0511 0.8269 0.9093
No log 3.6471 62 0.9254 -0.0392 0.9254 0.9620
No log 3.7647 64 1.1030 -0.0264 1.1030 1.0502
No log 3.8824 66 1.2248 0.0081 1.2248 1.1067
No log 4.0 68 1.0680 -0.0112 1.0680 1.0335
No log 4.1176 70 1.0519 0.2173 1.0519 1.0256
No log 4.2353 72 1.0236 0.2076 1.0236 1.0117
No log 4.3529 74 1.0578 0.0120 1.0578 1.0285
No log 4.4706 76 1.0491 -0.0497 1.0491 1.0243
No log 4.5882 78 1.0570 0.1101 1.0570 1.0281
No log 4.7059 80 1.0229 0.1289 1.0229 1.0114
No log 4.8235 82 1.0278 0.1827 1.0278 1.0138
No log 4.9412 84 1.0840 0.1709 1.0840 1.0411
No log 5.0588 86 1.0419 0.2537 1.0419 1.0208
No log 5.1765 88 1.0101 0.0244 1.0101 1.0050
No log 5.2941 90 0.9974 -0.0108 0.9974 0.9987
No log 5.4118 92 1.0215 0.2124 1.0215 1.0107
No log 5.5294 94 1.2083 0.0948 1.2083 1.0992
No log 5.6471 96 1.2811 0.0852 1.2811 1.1319
No log 5.7647 98 1.1840 0.0881 1.1840 1.0881
No log 5.8824 100 1.0397 0.2728 1.0397 1.0197
No log 6.0 102 0.8852 0.2149 0.8852 0.9408
No log 6.1176 104 0.8421 0.1353 0.8421 0.9177
No log 6.2353 106 0.8525 0.2270 0.8525 0.9233
No log 6.3529 108 0.8804 0.3817 0.8804 0.9383
No log 6.4706 110 0.9761 0.3398 0.9761 0.9880
No log 6.5882 112 1.0359 0.2708 1.0359 1.0178
No log 6.7059 114 0.9867 0.28 0.9867 0.9933
No log 6.8235 116 0.9716 0.1843 0.9716 0.9857
No log 6.9412 118 1.0156 0.3043 1.0156 1.0078
No log 7.0588 120 1.2545 0.1409 1.2545 1.1201
No log 7.1765 122 1.3188 0.1357 1.3188 1.1484
No log 7.2941 124 1.1722 0.2231 1.1722 1.0827
No log 7.4118 126 0.9740 0.2076 0.9740 0.9869
No log 7.5294 128 0.9138 0.1373 0.9138 0.9559
No log 7.6471 130 0.9126 0.1867 0.9126 0.9553
No log 7.7647 132 0.9566 0.3110 0.9566 0.9780
No log 7.8824 134 0.8954 0.3069 0.8954 0.9463
No log 8.0 136 0.8675 0.2547 0.8675 0.9314
No log 8.1176 138 0.8378 0.2661 0.8378 0.9153
No log 8.2353 140 0.8697 0.2670 0.8697 0.9326
No log 8.3529 142 0.9027 0.4085 0.9027 0.9501
No log 8.4706 144 0.8102 0.4464 0.8102 0.9001
No log 8.5882 146 0.8692 0.3760 0.8692 0.9323
No log 8.7059 148 1.0551 0.2613 1.0551 1.0272
No log 8.8235 150 1.2279 0.2299 1.2279 1.1081
No log 8.9412 152 1.0097 0.3123 1.0097 1.0049
No log 9.0588 154 0.8426 0.3011 0.8426 0.9179
No log 9.1765 156 0.9170 0.4321 0.9170 0.9576
No log 9.2941 158 1.2483 0.1841 1.2483 1.1173
No log 9.4118 160 1.5483 0.1965 1.5483 1.2443
No log 9.5294 162 1.4330 0.1965 1.4330 1.1971
No log 9.6471 164 1.2412 0.1805 1.2412 1.1141
No log 9.7647 166 0.9849 0.3461 0.9849 0.9924
No log 9.8824 168 0.8784 0.3255 0.8784 0.9372
No log 10.0 170 0.8658 0.2402 0.8658 0.9305
No log 10.1176 172 0.8826 0.2993 0.8826 0.9395
No log 10.2353 174 0.9103 0.2937 0.9103 0.9541
No log 10.3529 176 0.9526 0.3149 0.9526 0.9760
No log 10.4706 178 0.9896 0.3582 0.9896 0.9948
No log 10.5882 180 0.9768 0.3425 0.9768 0.9883
No log 10.7059 182 0.9128 0.3125 0.9128 0.9554
No log 10.8235 184 0.8358 0.3609 0.8358 0.9142
No log 10.9412 186 0.8132 0.2172 0.8132 0.9018
No log 11.0588 188 0.8547 0.2988 0.8547 0.9245
No log 11.1765 190 0.9638 0.2981 0.9638 0.9817
No log 11.2941 192 1.0044 0.2905 1.0044 1.0022
No log 11.4118 194 0.8985 0.3645 0.8985 0.9479
No log 11.5294 196 0.8472 0.3520 0.8472 0.9204
No log 11.6471 198 0.8329 0.3520 0.8329 0.9126
No log 11.7647 200 0.8679 0.3520 0.8679 0.9316
No log 11.8824 202 0.8890 0.3520 0.8890 0.9429
No log 12.0 204 0.9563 0.3456 0.9563 0.9779
No log 12.1176 206 1.0002 0.3128 1.0002 1.0001
No log 12.2353 208 1.1117 0.2568 1.1117 1.0544
No log 12.3529 210 1.0345 0.2577 1.0345 1.0171
No log 12.4706 212 0.9111 0.3100 0.9111 0.9545
No log 12.5882 214 0.8839 0.3100 0.8839 0.9402
No log 12.7059 216 0.9808 0.2602 0.9808 0.9904
No log 12.8235 218 1.1467 0.2622 1.1467 1.0708
No log 12.9412 220 1.0584 0.2568 1.0584 1.0288
No log 13.0588 222 0.8061 0.3231 0.8061 0.8978
No log 13.1765 224 0.7892 0.3316 0.7892 0.8884
No log 13.2941 226 0.8005 0.3239 0.8005 0.8947
No log 13.4118 228 0.7286 0.3239 0.7286 0.8536
No log 13.5294 230 0.8072 0.3699 0.8072 0.8985
No log 13.6471 232 0.9795 0.2547 0.9795 0.9897
No log 13.7647 234 0.9308 0.2894 0.9308 0.9648
No log 13.8824 236 0.8059 0.3565 0.8059 0.8977
No log 14.0 238 0.8116 0.2867 0.8116 0.9009
No log 14.1176 240 0.8460 0.2508 0.8460 0.9198
No log 14.2353 242 0.9114 0.2725 0.9114 0.9547
No log 14.3529 244 1.0428 0.2636 1.0428 1.0212
No log 14.4706 246 1.1791 0.2524 1.1791 1.0859
No log 14.5882 248 1.1057 0.2372 1.1057 1.0515
No log 14.7059 250 0.9941 0.2626 0.9941 0.9971
No log 14.8235 252 0.9349 0.1884 0.9349 0.9669
No log 14.9412 254 0.8950 0.2616 0.8950 0.9460
No log 15.0588 256 0.9057 0.2616 0.9057 0.9517
No log 15.1765 258 0.9757 0.2220 0.9757 0.9878
No log 15.2941 260 1.0184 0.2659 1.0184 1.0091
No log 15.4118 262 1.1376 0.2312 1.1376 1.0666
No log 15.5294 264 0.9745 0.3022 0.9745 0.9872
No log 15.6471 266 0.8096 0.2832 0.8096 0.8998
No log 15.7647 268 0.7996 0.3710 0.7996 0.8942
No log 15.8824 270 0.9065 0.3346 0.9065 0.9521
No log 16.0 272 1.0541 0.2777 1.0541 1.0267
No log 16.1176 274 1.1364 0.2481 1.1364 1.0660
No log 16.2353 276 1.0103 0.3080 1.0103 1.0052
No log 16.3529 278 0.8159 0.2861 0.8159 0.9032
No log 16.4706 280 0.7298 0.3524 0.7298 0.8543
No log 16.5882 282 0.7215 0.3183 0.7215 0.8494
No log 16.7059 284 0.7433 0.3195 0.7433 0.8621
No log 16.8235 286 0.8522 0.3310 0.8522 0.9232
No log 16.9412 288 0.9333 0.2964 0.9333 0.9661
No log 17.0588 290 0.8740 0.3310 0.8740 0.9349
No log 17.1765 292 0.7518 0.3520 0.7518 0.8671
No log 17.2941 294 0.7313 0.3340 0.7313 0.8552
No log 17.4118 296 0.7497 0.3700 0.7497 0.8659
No log 17.5294 298 0.8428 0.3100 0.8428 0.9180
No log 17.6471 300 1.1370 0.2956 1.1370 1.0663
No log 17.7647 302 1.4161 0.3177 1.4161 1.1900
No log 17.8824 304 1.3695 0.2880 1.3695 1.1703
No log 18.0 306 1.1179 0.3003 1.1179 1.0573
No log 18.1176 308 0.8695 0.3092 0.8695 0.9325
No log 18.2353 310 0.7828 0.3482 0.7828 0.8848
No log 18.3529 312 0.7489 0.3700 0.7489 0.8654
No log 18.4706 314 0.7356 0.4369 0.7356 0.8576
No log 18.5882 316 0.7641 0.3981 0.7641 0.8741
No log 18.7059 318 0.7772 0.3981 0.7772 0.8816
No log 18.8235 320 0.7786 0.3731 0.7786 0.8824
No log 18.9412 322 0.7743 0.3623 0.7743 0.8800
No log 19.0588 324 0.7556 0.4104 0.7556 0.8692
No log 19.1765 326 0.7104 0.4093 0.7104 0.8428
No log 19.2941 328 0.7145 0.4023 0.7145 0.8453
No log 19.4118 330 0.7883 0.3699 0.7883 0.8879
No log 19.5294 332 0.7953 0.3699 0.7953 0.8918
No log 19.6471 334 0.7413 0.3955 0.7413 0.8610
No log 19.7647 336 0.7792 0.3699 0.7792 0.8827
No log 19.8824 338 0.8309 0.3461 0.8309 0.9115
No log 20.0 340 0.8468 0.2958 0.8468 0.9202
No log 20.1176 342 0.8889 0.3217 0.8889 0.9428
No log 20.2353 344 0.9399 0.2881 0.9399 0.9695
No log 20.3529 346 0.9329 0.3217 0.9329 0.9659
No log 20.4706 348 0.8325 0.2958 0.8325 0.9124
No log 20.5882 350 0.7942 0.2958 0.7942 0.8912
No log 20.7059 352 0.7621 0.3543 0.7621 0.8730
No log 20.8235 354 0.7518 0.4334 0.7518 0.8670
No log 20.9412 356 0.7600 0.4104 0.7600 0.8718
No log 21.0588 358 0.7326 0.4404 0.7326 0.8559
No log 21.1765 360 0.6829 0.4476 0.6829 0.8264
No log 21.2941 362 0.6539 0.3888 0.6539 0.8087
No log 21.4118 364 0.6483 0.4260 0.6483 0.8052
No log 21.5294 366 0.6354 0.4227 0.6354 0.7971
No log 21.6471 368 0.6492 0.4393 0.6492 0.8057
No log 21.7647 370 0.6610 0.4413 0.6610 0.8130
No log 21.8824 372 0.7028 0.4017 0.7028 0.8383
No log 22.0 374 0.8054 0.4265 0.8054 0.8974
No log 22.1176 376 0.8046 0.4199 0.8046 0.8970
No log 22.2353 378 0.7372 0.3183 0.7372 0.8586
No log 22.3529 380 0.7133 0.3798 0.7133 0.8446
No log 22.4706 382 0.6964 0.4044 0.6964 0.8345
No log 22.5882 384 0.6870 0.3990 0.6870 0.8289
No log 22.7059 386 0.7074 0.3613 0.7074 0.8411
No log 22.8235 388 0.7800 0.3909 0.7800 0.8832
No log 22.9412 390 0.9000 0.3403 0.9000 0.9487
No log 23.0588 392 0.9239 0.2830 0.9239 0.9612
No log 23.1765 394 0.8343 0.3274 0.8343 0.9134
No log 23.2941 396 0.7481 0.3417 0.7481 0.8649
No log 23.4118 398 0.7175 0.3567 0.7175 0.8471
No log 23.5294 400 0.7203 0.3567 0.7203 0.8487
No log 23.6471 402 0.7941 0.3667 0.7941 0.8911
No log 23.7647 404 0.8685 0.3699 0.8685 0.9320
No log 23.8824 406 0.8363 0.3760 0.8363 0.9145
No log 24.0 408 0.8155 0.3262 0.8155 0.9030
No log 24.1176 410 0.8211 0.3760 0.8211 0.9061
No log 24.2353 412 0.7396 0.3732 0.7396 0.8600
No log 24.3529 414 0.6927 0.3754 0.6927 0.8323
No log 24.4706 416 0.6589 0.4081 0.6589 0.8117
No log 24.5882 418 0.6387 0.3665 0.6387 0.7992
No log 24.7059 420 0.6273 0.4482 0.6273 0.7920
No log 24.8235 422 0.6391 0.4029 0.6391 0.7994
No log 24.9412 424 0.7224 0.3889 0.7224 0.8499
No log 25.0588 426 0.7751 0.3824 0.7751 0.8804
No log 25.1765 428 0.7297 0.3732 0.7297 0.8542
No log 25.2941 430 0.6965 0.4093 0.6965 0.8346
No log 25.4118 432 0.6474 0.4336 0.6474 0.8046
No log 25.5294 434 0.6268 0.3961 0.6268 0.7917
No log 25.6471 436 0.6265 0.4253 0.6265 0.7915
No log 25.7647 438 0.6381 0.4724 0.6381 0.7988
No log 25.8824 440 0.7220 0.4023 0.7220 0.8497
No log 26.0 442 0.8814 0.4426 0.8814 0.9388
No log 26.1176 444 0.9367 0.4225 0.9367 0.9678
No log 26.2353 446 0.8408 0.4222 0.8408 0.9169
No log 26.3529 448 0.7113 0.4404 0.7113 0.8434
No log 26.4706 450 0.6752 0.3936 0.6752 0.8217
No log 26.5882 452 0.6850 0.3936 0.6850 0.8276
No log 26.7059 454 0.7238 0.3866 0.7238 0.8508
No log 26.8235 456 0.8541 0.2905 0.8541 0.9242
No log 26.9412 458 1.0946 0.2799 1.0946 1.0462
No log 27.0588 460 1.2750 0.3255 1.2750 1.1291
No log 27.1765 462 1.2941 0.3255 1.2941 1.1376
No log 27.2941 464 1.1443 0.3003 1.1443 1.0697
No log 27.4118 466 0.9177 0.2651 0.9177 0.9579
No log 27.5294 468 0.7517 0.3609 0.7517 0.8670
No log 27.6471 470 0.7091 0.2099 0.7091 0.8421
No log 27.7647 472 0.7056 0.1813 0.7056 0.8400
No log 27.8824 474 0.7069 0.2872 0.7069 0.8408
No log 28.0 476 0.7308 0.3633 0.7308 0.8549
No log 28.1176 478 0.7710 0.3776 0.7710 0.8780
No log 28.2353 480 0.8296 0.3803 0.8296 0.9108
No log 28.3529 482 0.9358 0.3204 0.9358 0.9674
No log 28.4706 484 0.9827 0.3574 0.9827 0.9913
No log 28.5882 486 0.9004 0.3806 0.9004 0.9489
No log 28.7059 488 0.7607 0.3844 0.7607 0.8722
No log 28.8235 490 0.7104 0.4239 0.7104 0.8428
No log 28.9412 492 0.7054 0.4239 0.7054 0.8399
No log 29.0588 494 0.7133 0.3866 0.7133 0.8445
No log 29.1765 496 0.7619 0.3520 0.7619 0.8729
No log 29.2941 498 0.8223 0.3623 0.8223 0.9068
0.3298 29.4118 500 0.8126 0.3159 0.8126 0.9014
0.3298 29.5294 502 0.8103 0.3012 0.8103 0.9002
0.3298 29.6471 504 0.7970 0.3356 0.7970 0.8928
0.3298 29.7647 506 0.7674 0.4023 0.7674 0.8760
0.3298 29.8824 508 0.7534 0.3976 0.7534 0.8680
0.3298 30.0 510 0.7638 0.3976 0.7638 0.8739
0.3298 30.1176 512 0.7785 0.3976 0.7785 0.8824
0.3298 30.2353 514 0.7944 0.4023 0.7944 0.8913
0.3298 30.3529 516 0.8079 0.4023 0.8079 0.8988
0.3298 30.4706 518 0.8221 0.3503 0.8221 0.9067
0.3298 30.5882 520 0.8389 0.3503 0.8389 0.9159
0.3298 30.7059 522 0.8922 0.2832 0.8922 0.9446
0.3298 30.8235 524 1.0238 0.2239 1.0238 1.0118
0.3298 30.9412 526 1.1045 0.2626 1.1045 1.0510
0.3298 31.0588 528 1.0971 0.2437 1.0971 1.0474
0.3298 31.1765 530 1.1299 0.1775 1.1299 1.0630

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k6_task7_organization

Finetuned
(4023)
this model