ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0122
  • Qwk: 0.2777
  • Mse: 1.0122
  • Rmse: 1.0061

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0488 2 2.4815 -0.0449 2.4815 1.5753
No log 0.0976 4 1.3059 0.0401 1.3059 1.1428
No log 0.1463 6 0.9850 -0.1408 0.9850 0.9925
No log 0.1951 8 1.0396 0.0870 1.0396 1.0196
No log 0.2439 10 1.2502 0.0338 1.2502 1.1181
No log 0.2927 12 1.3839 -0.1276 1.3839 1.1764
No log 0.3415 14 1.2962 -0.0971 1.2962 1.1385
No log 0.3902 16 1.1915 -0.0861 1.1915 1.0916
No log 0.4390 18 1.0231 -0.0511 1.0231 1.0115
No log 0.4878 20 1.0139 0.0448 1.0139 1.0069
No log 0.5366 22 1.0111 -0.1285 1.0111 1.0055
No log 0.5854 24 0.9419 0.0236 0.9419 0.9705
No log 0.6341 26 0.8744 0.0573 0.8744 0.9351
No log 0.6829 28 0.9231 0.0419 0.9231 0.9608
No log 0.7317 30 0.9251 0.0448 0.9251 0.9618
No log 0.7805 32 0.9190 0.0255 0.9190 0.9586
No log 0.8293 34 0.9629 0.0774 0.9629 0.9813
No log 0.8780 36 1.0923 0.0260 1.0923 1.0451
No log 0.9268 38 1.2640 0.0169 1.2640 1.1243
No log 0.9756 40 1.2137 0.0986 1.2137 1.1017
No log 1.0244 42 1.3328 0.0470 1.3328 1.1545
No log 1.0732 44 1.1418 0.0980 1.1418 1.0685
No log 1.1220 46 0.9232 0.2712 0.9232 0.9608
No log 1.1707 48 0.9377 0.1328 0.9377 0.9683
No log 1.2195 50 0.9590 0.1332 0.9590 0.9793
No log 1.2683 52 1.0831 0.2002 1.0831 1.0407
No log 1.3171 54 1.1087 0.1293 1.1087 1.0529
No log 1.3659 56 0.8897 0.2871 0.8897 0.9432
No log 1.4146 58 0.8150 0.1737 0.8150 0.9028
No log 1.4634 60 0.8025 0.1092 0.8025 0.8959
No log 1.5122 62 0.8125 0.1492 0.8125 0.9014
No log 1.5610 64 0.7955 0.1131 0.7955 0.8919
No log 1.6098 66 0.7964 0.1823 0.7964 0.8924
No log 1.6585 68 0.7853 0.1941 0.7853 0.8862
No log 1.7073 70 0.7693 0.0784 0.7693 0.8771
No log 1.7561 72 0.7446 0.1498 0.7446 0.8629
No log 1.8049 74 0.7727 0.1710 0.7727 0.8791
No log 1.8537 76 0.7906 0.1567 0.7906 0.8891
No log 1.9024 78 0.7737 0.0748 0.7737 0.8796
No log 1.9512 80 0.8815 0.1323 0.8815 0.9389
No log 2.0 82 0.9939 0.2263 0.9939 0.9970
No log 2.0488 84 0.9604 0.1628 0.9604 0.9800
No log 2.0976 86 0.8694 0.1581 0.8694 0.9324
No log 2.1463 88 0.7641 0.1492 0.7641 0.8741
No log 2.1951 90 0.8261 0.2847 0.8261 0.9089
No log 2.2439 92 0.9497 0.1101 0.9497 0.9745
No log 2.2927 94 0.9542 0.1101 0.9542 0.9768
No log 2.3415 96 0.8970 0.1603 0.8970 0.9471
No log 2.3902 98 0.8470 0.1716 0.8470 0.9203
No log 2.4390 100 0.9057 0.2056 0.9057 0.9517
No log 2.4878 102 0.9011 0.2122 0.9011 0.9493
No log 2.5366 104 0.8577 0.2182 0.8577 0.9261
No log 2.5854 106 0.7936 0.1797 0.7936 0.8908
No log 2.6341 108 0.8035 0.1737 0.8035 0.8964
No log 2.6829 110 0.8455 0.2204 0.8455 0.9195
No log 2.7317 112 0.8680 0.3127 0.8680 0.9317
No log 2.7805 114 0.8416 0.2847 0.8416 0.9174
No log 2.8293 116 0.8236 0.2847 0.8236 0.9075
No log 2.8780 118 0.8269 0.2847 0.8269 0.9094
No log 2.9268 120 0.8463 0.2440 0.8463 0.9199
No log 2.9756 122 0.8347 0.2498 0.8347 0.9136
No log 3.0244 124 0.8419 0.2784 0.8419 0.9175
No log 3.0732 126 0.8240 0.2913 0.8240 0.9077
No log 3.1220 128 0.8294 0.2498 0.8294 0.9107
No log 3.1707 130 0.7894 0.2113 0.7894 0.8885
No log 3.2195 132 0.8256 0.2741 0.8256 0.9086
No log 3.2683 134 0.9098 0.1458 0.9098 0.9538
No log 3.3171 136 0.8494 0.2409 0.8494 0.9216
No log 3.3659 138 0.8293 0.2279 0.8293 0.9107
No log 3.4146 140 0.8698 0.3060 0.8698 0.9326
No log 3.4634 142 0.8762 0.3305 0.8762 0.9360
No log 3.5122 144 0.8735 0.2809 0.8735 0.9346
No log 3.5610 146 0.8728 0.1649 0.8728 0.9343
No log 3.6098 148 0.8414 0.2691 0.8414 0.9173
No log 3.6585 150 0.7579 0.3763 0.7579 0.8706
No log 3.7073 152 0.7030 0.2181 0.7030 0.8385
No log 3.7561 154 0.7216 0.2837 0.7216 0.8495
No log 3.8049 156 0.7125 0.2774 0.7125 0.8441
No log 3.8537 158 0.6728 0.1432 0.6728 0.8203
No log 3.9024 160 0.6511 0.3575 0.6511 0.8069
No log 3.9512 162 0.6832 0.3399 0.6832 0.8265
No log 4.0 164 0.6934 0.2932 0.6934 0.8327
No log 4.0488 166 0.7235 0.3475 0.7235 0.8506
No log 4.0976 168 0.7521 0.3314 0.7521 0.8672
No log 4.1463 170 0.7559 0.3186 0.7559 0.8694
No log 4.1951 172 0.7937 0.3107 0.7937 0.8909
No log 4.2439 174 0.7443 0.2842 0.7443 0.8628
No log 4.2927 176 0.7296 0.2751 0.7296 0.8542
No log 4.3415 178 0.7936 0.3723 0.7936 0.8909
No log 4.3902 180 0.8049 0.3157 0.8049 0.8971
No log 4.4390 182 0.7723 0.3144 0.7723 0.8788
No log 4.4878 184 0.8017 0.2303 0.8017 0.8954
No log 4.5366 186 0.8616 0.2345 0.8616 0.9282
No log 4.5854 188 0.8246 0.3523 0.8246 0.9081
No log 4.6341 190 0.7952 0.3605 0.7952 0.8918
No log 4.6829 192 0.7861 0.4149 0.7861 0.8866
No log 4.7317 194 0.7714 0.3305 0.7714 0.8783
No log 4.7805 196 0.7682 0.3561 0.7682 0.8765
No log 4.8293 198 0.7810 0.2857 0.7810 0.8837
No log 4.8780 200 0.7877 0.3086 0.7877 0.8875
No log 4.9268 202 0.7981 0.3433 0.7981 0.8934
No log 4.9756 204 0.7893 0.3393 0.7893 0.8884
No log 5.0244 206 0.7625 0.3859 0.7625 0.8732
No log 5.0732 208 0.7728 0.3500 0.7728 0.8791
No log 5.1220 210 0.7916 0.3409 0.7916 0.8897
No log 5.1707 212 0.7584 0.3475 0.7584 0.8708
No log 5.2195 214 0.7327 0.4547 0.7327 0.8560
No log 5.2683 216 0.7358 0.3787 0.7358 0.8578
No log 5.3171 218 0.7345 0.2386 0.7345 0.8570
No log 5.3659 220 0.7367 0.2386 0.7367 0.8583
No log 5.4146 222 0.7447 0.1455 0.7447 0.8630
No log 5.4634 224 0.7467 0.1373 0.7467 0.8641
No log 5.5122 226 0.7578 0.1417 0.7578 0.8705
No log 5.5610 228 0.7745 0.1410 0.7745 0.8801
No log 5.6098 230 0.7785 0.1417 0.7785 0.8823
No log 5.6585 232 0.7871 0.2652 0.7871 0.8872
No log 5.7073 234 0.8254 0.2847 0.8254 0.9085
No log 5.7561 236 0.8508 0.3399 0.8508 0.9224
No log 5.8049 238 0.8017 0.2007 0.8017 0.8954
No log 5.8537 240 0.7859 0.2319 0.7859 0.8865
No log 5.9024 242 0.8130 0.3355 0.8130 0.9017
No log 5.9512 244 0.8607 0.3157 0.8607 0.9277
No log 6.0 246 0.8802 0.2960 0.8802 0.9382
No log 6.0488 248 0.9455 0.2724 0.9455 0.9724
No log 6.0976 250 0.9840 0.3173 0.9840 0.9920
No log 6.1463 252 0.8529 0.3320 0.8529 0.9235
No log 6.1951 254 0.7496 0.2327 0.7496 0.8658
No log 6.2439 256 0.7763 0.2439 0.7763 0.8811
No log 6.2927 258 0.7719 0.2130 0.7719 0.8786
No log 6.3415 260 0.7632 0.3229 0.7632 0.8736
No log 6.3902 262 0.7431 0.2135 0.7431 0.8620
No log 6.4390 264 0.7775 0.3544 0.7775 0.8818
No log 6.4878 266 0.8851 0.3320 0.8851 0.9408
No log 6.5366 268 1.0608 0.3123 1.0608 1.0299
No log 6.5854 270 1.0974 0.3024 1.0974 1.0476
No log 6.6341 272 1.0274 0.2821 1.0274 1.0136
No log 6.6829 274 0.9414 0.28 0.9414 0.9703
No log 6.7317 276 0.8268 0.3320 0.8268 0.9093
No log 6.7805 278 0.8063 0.3344 0.8063 0.8980
No log 6.8293 280 0.7809 0.3677 0.7809 0.8837
No log 6.8780 282 0.7676 0.3382 0.7676 0.8761
No log 6.9268 284 0.8208 0.3125 0.8208 0.9060
No log 6.9756 286 0.9071 0.2590 0.9071 0.9524
No log 7.0244 288 0.8490 0.2669 0.8490 0.9214
No log 7.0732 290 0.8170 0.3316 0.8170 0.9039
No log 7.1220 292 0.8059 0.3754 0.8059 0.8977
No log 7.1707 294 0.8489 0.3781 0.8489 0.9214
No log 7.2195 296 0.8553 0.3781 0.8553 0.9248
No log 7.2683 298 0.8084 0.3085 0.8084 0.8991
No log 7.3171 300 0.7965 0.3085 0.7965 0.8924
No log 7.3659 302 0.8009 0.3746 0.8009 0.8950
No log 7.4146 304 0.7861 0.3574 0.7861 0.8866
No log 7.4634 306 0.7697 0.2929 0.7697 0.8773
No log 7.5122 308 0.7610 0.2622 0.7610 0.8724
No log 7.5610 310 0.7582 0.2161 0.7582 0.8708
No log 7.6098 312 0.7568 0.2445 0.7568 0.8699
No log 7.6585 314 0.7685 0.3287 0.7685 0.8766
No log 7.7073 316 0.8012 0.3433 0.8012 0.8951
No log 7.7561 318 0.8532 0.2943 0.8532 0.9237
No log 7.8049 320 0.8827 0.2247 0.8827 0.9395
No log 7.8537 322 0.8562 0.2566 0.8562 0.9253
No log 7.9024 324 0.8140 0.3914 0.8140 0.9022
No log 7.9512 326 0.8180 0.3914 0.8180 0.9044
No log 8.0 328 0.9461 0.3226 0.9461 0.9727
No log 8.0488 330 1.0324 0.3123 1.0324 1.0161
No log 8.0976 332 0.9294 0.2627 0.9294 0.9640
No log 8.1463 334 0.7944 0.3195 0.7944 0.8913
No log 8.1951 336 0.7357 0.1806 0.7357 0.8578
No log 8.2439 338 0.7321 0.2505 0.7321 0.8556
No log 8.2927 340 0.7397 0.2720 0.7397 0.8601
No log 8.3415 342 0.7854 0.3060 0.7854 0.8862
No log 8.3902 344 0.8301 0.3131 0.8301 0.9111
No log 8.4390 346 0.8676 0.2861 0.8676 0.9314
No log 8.4878 348 0.8558 0.2392 0.8558 0.9251
No log 8.5366 350 0.8726 0.2723 0.8726 0.9341
No log 8.5854 352 0.8523 0.2893 0.8523 0.9232
No log 8.6341 354 0.8260 0.4171 0.8260 0.9088
No log 8.6829 356 0.7958 0.3625 0.7958 0.8921
No log 8.7317 358 0.7930 0.3171 0.7930 0.8905
No log 8.7805 360 0.8145 0.3478 0.8145 0.9025
No log 8.8293 362 0.8969 0.3377 0.8969 0.9470
No log 8.8780 364 1.0134 0.2020 1.0134 1.0067
No log 8.9268 366 1.0690 0.1979 1.0690 1.0339
No log 8.9756 368 1.0531 0.1979 1.0531 1.0262
No log 9.0244 370 1.0272 0.2147 1.0272 1.0135
No log 9.0732 372 1.0168 0.2954 1.0168 1.0084
No log 9.1220 374 1.0180 0.2192 1.0180 1.0089
No log 9.1707 376 0.9255 0.3077 0.9255 0.9620
No log 9.2195 378 0.8835 0.2920 0.8835 0.9400
No log 9.2683 380 0.8919 0.2862 0.8919 0.9444
No log 9.3171 382 0.8729 0.3144 0.8729 0.9343
No log 9.3659 384 0.8529 0.3837 0.8529 0.9235
No log 9.4146 386 0.8322 0.3714 0.8322 0.9122
No log 9.4634 388 0.8261 0.3482 0.8261 0.9089
No log 9.5122 390 0.8867 0.3489 0.8867 0.9416
No log 9.5610 392 0.9575 0.3115 0.9575 0.9785
No log 9.6098 394 1.1328 0.2292 1.1328 1.0643
No log 9.6585 396 1.1433 0.2075 1.1433 1.0693
No log 9.7073 398 0.9920 0.1549 0.9920 0.9960
No log 9.7561 400 0.8906 0.2294 0.8906 0.9437
No log 9.8049 402 0.8566 0.2832 0.8566 0.9255
No log 9.8537 404 0.8141 0.3001 0.8141 0.9023
No log 9.9024 406 0.7900 0.3095 0.7900 0.8888
No log 9.9512 408 0.7903 0.3475 0.7903 0.8890
No log 10.0 410 0.7743 0.3569 0.7743 0.8800
No log 10.0488 412 0.8078 0.3060 0.8078 0.8988
No log 10.0976 414 0.9446 0.2055 0.9446 0.9719
No log 10.1463 416 1.0374 0.1974 1.0374 1.0185
No log 10.1951 418 1.0207 0.1974 1.0207 1.0103
No log 10.2439 420 0.9737 0.2509 0.9737 0.9868
No log 10.2927 422 0.8885 0.2832 0.8885 0.9426
No log 10.3415 424 0.8153 0.3656 0.8153 0.9029
No log 10.3902 426 0.8346 0.2835 0.8346 0.9136
No log 10.4390 428 0.8459 0.3060 0.8459 0.9197
No log 10.4878 430 0.8471 0.3060 0.8471 0.9204
No log 10.5366 432 0.8740 0.2832 0.8740 0.9349
No log 10.5854 434 0.9135 0.2832 0.9135 0.9558
No log 10.6341 436 0.8849 0.2832 0.8849 0.9407
No log 10.6829 438 0.8930 0.2832 0.8930 0.9450
No log 10.7317 440 0.9553 0.2627 0.9553 0.9774
No log 10.7805 442 1.1051 0.2335 1.1051 1.0512
No log 10.8293 444 1.2390 0.2520 1.2390 1.1131
No log 10.8780 446 1.1763 0.2840 1.1763 1.0846
No log 10.9268 448 0.9811 0.3183 0.9811 0.9905
No log 10.9756 450 0.8129 0.2474 0.8129 0.9016
No log 11.0244 452 0.7788 0.1661 0.7788 0.8825
No log 11.0732 454 0.7708 0.1843 0.7708 0.8779
No log 11.1220 456 0.7910 0.2953 0.7910 0.8894
No log 11.1707 458 0.8244 0.2495 0.8244 0.9080
No log 11.2195 460 0.8672 0.2696 0.8672 0.9312
No log 11.2683 462 0.8645 0.2696 0.8645 0.9298
No log 11.3171 464 0.8312 0.3085 0.8312 0.9117
No log 11.3659 466 0.8052 0.3417 0.8052 0.8973
No log 11.4146 468 0.8075 0.3353 0.8075 0.8986
No log 11.4634 470 0.8085 0.3353 0.8085 0.8992
No log 11.5122 472 0.8410 0.3231 0.8410 0.9171
No log 11.5610 474 0.9134 0.2776 0.9134 0.9557
No log 11.6098 476 1.0172 0.2727 1.0172 1.0086
No log 11.6585 478 1.0527 0.2627 1.0527 1.0260
No log 11.7073 480 1.0091 0.2467 1.0091 1.0045
No log 11.7561 482 1.0077 0.2467 1.0077 1.0039
No log 11.8049 484 1.0328 0.2905 1.0328 1.0163
No log 11.8537 486 0.9451 0.2928 0.9451 0.9721
No log 11.9024 488 0.8947 0.3149 0.8947 0.9459
No log 11.9512 490 0.8126 0.3060 0.8126 0.9014
No log 12.0 492 0.7824 0.3172 0.7824 0.8845
No log 12.0488 494 0.7847 0.2893 0.7847 0.8858
No log 12.0976 496 0.8339 0.2781 0.8339 0.9132
No log 12.1463 498 0.8886 0.2670 0.8886 0.9427
0.3573 12.1951 500 0.9576 0.2861 0.9576 0.9786
0.3573 12.2439 502 0.9600 0.3100 0.9600 0.9798
0.3573 12.2927 504 0.9384 0.3394 0.9384 0.9687
0.3573 12.3415 506 0.8457 0.2670 0.8457 0.9196
0.3573 12.3902 508 0.8048 0.3274 0.8048 0.8971
0.3573 12.4390 510 0.8276 0.3172 0.8276 0.9097
0.3573 12.4878 512 0.8758 0.2134 0.8758 0.9358
0.3573 12.5366 514 0.9947 0.3044 0.9947 0.9974
0.3573 12.5854 516 1.0587 0.2821 1.0587 1.0290
0.3573 12.6341 518 1.0639 0.2821 1.0639 1.0314
0.3573 12.6829 520 0.9720 0.3425 0.9720 0.9859
0.3573 12.7317 522 0.9135 0.2670 0.9135 0.9558
0.3573 12.7805 524 0.9103 0.3034 0.9103 0.9541
0.3573 12.8293 526 0.9347 0.3760 0.9347 0.9668
0.3573 12.8780 528 0.8763 0.3608 0.8763 0.9361
0.3573 12.9268 530 0.8244 0.3391 0.8244 0.9080
0.3573 12.9756 532 0.8197 0.3325 0.8197 0.9054
0.3573 13.0244 534 0.8378 0.2835 0.8378 0.9153
0.3573 13.0732 536 0.8945 0.2315 0.8945 0.9458
0.3573 13.1220 538 0.9583 0.2564 0.9583 0.9789
0.3573 13.1707 540 1.0565 0.2485 1.0565 1.0279
0.3573 13.2195 542 1.0122 0.2777 1.0122 1.0061

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task7_organization

Finetuned
(4019)
this model