ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k15_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6683
  • Qwk: 0.3961
  • Mse: 0.6683
  • Rmse: 0.8175

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0263 2 2.4678 -0.0958 2.4678 1.5709
No log 0.0526 4 1.2661 0.1609 1.2661 1.1252
No log 0.0789 6 1.0793 -0.1866 1.0793 1.0389
No log 0.1053 8 1.1927 -0.1356 1.1927 1.0921
No log 0.1316 10 1.1875 -0.1665 1.1875 1.0897
No log 0.1579 12 1.0744 0.0288 1.0744 1.0366
No log 0.1842 14 1.0978 0.0175 1.0978 1.0477
No log 0.2105 16 0.9805 0.1142 0.9805 0.9902
No log 0.2368 18 0.9673 0.0185 0.9673 0.9835
No log 0.2632 20 1.2086 -0.0122 1.2086 1.0994
No log 0.2895 22 1.0229 0.1618 1.0229 1.0114
No log 0.3158 24 0.8146 0.2811 0.8146 0.9026
No log 0.3421 26 0.8571 0.2118 0.8571 0.9258
No log 0.3684 28 0.9099 0.1537 0.9099 0.9539
No log 0.3947 30 0.9726 0.2212 0.9726 0.9862
No log 0.4211 32 0.8423 0.1215 0.8423 0.9177
No log 0.4474 34 0.9011 0.1459 0.9011 0.9492
No log 0.4737 36 0.9816 0.1867 0.9816 0.9908
No log 0.5 38 0.9068 0.1459 0.9068 0.9523
No log 0.5263 40 0.8703 0.0697 0.8703 0.9329
No log 0.5526 42 1.0654 0.0609 1.0654 1.0322
No log 0.5789 44 1.1998 0.0990 1.1998 1.0954
No log 0.6053 46 1.0457 0.2381 1.0457 1.0226
No log 0.6316 48 0.8262 0.0947 0.8262 0.9089
No log 0.6579 50 0.7360 0.1922 0.7360 0.8579
No log 0.6842 52 0.7562 0.1807 0.7562 0.8696
No log 0.7105 54 0.7381 0.2181 0.7381 0.8591
No log 0.7368 56 0.7063 0.2676 0.7063 0.8404
No log 0.7632 58 0.7305 0.3125 0.7305 0.8547
No log 0.7895 60 0.7675 0.3360 0.7675 0.8761
No log 0.8158 62 0.8193 0.3230 0.8193 0.9052
No log 0.8421 64 0.7649 0.3279 0.7649 0.8746
No log 0.8684 66 0.7852 0.3122 0.7852 0.8861
No log 0.8947 68 0.8202 0.2709 0.8202 0.9057
No log 0.9211 70 0.9736 0.2601 0.9736 0.9867
No log 0.9474 72 1.3211 0.0514 1.3211 1.1494
No log 0.9737 74 1.2969 0.0974 1.2969 1.1388
No log 1.0 76 1.0919 0.1488 1.0919 1.0449
No log 1.0263 78 0.8811 0.3289 0.8811 0.9387
No log 1.0526 80 0.7731 0.1723 0.7731 0.8792
No log 1.0789 82 0.8012 0.2558 0.8012 0.8951
No log 1.1053 84 0.8232 0.2334 0.8232 0.9073
No log 1.1316 86 0.8468 0.2092 0.8468 0.9202
No log 1.1579 88 0.8766 0.2034 0.8766 0.9363
No log 1.1842 90 0.9101 0.2034 0.9101 0.9540
No log 1.2105 92 0.9673 0.1950 0.9673 0.9835
No log 1.2368 94 1.0942 0.1086 1.0942 1.0461
No log 1.2632 96 0.9994 0.1291 0.9994 0.9997
No log 1.2895 98 0.9816 0.1874 0.9816 0.9908
No log 1.3158 100 0.9482 0.1650 0.9482 0.9738
No log 1.3421 102 0.9172 0.1379 0.9172 0.9577
No log 1.3684 104 0.9123 0.1400 0.9123 0.9551
No log 1.3947 106 1.0000 0.0960 1.0000 1.0000
No log 1.4211 108 1.1834 0.1144 1.1834 1.0878
No log 1.4474 110 1.0549 0.1178 1.0549 1.0271
No log 1.4737 112 0.9157 0.1010 0.9157 0.9569
No log 1.5 114 1.0123 0.1562 1.0123 1.0061
No log 1.5263 116 0.9841 0.1544 0.9841 0.9920
No log 1.5526 118 0.8857 0.2623 0.8857 0.9411
No log 1.5789 120 0.8475 0.2511 0.8475 0.9206
No log 1.6053 122 0.8116 0.2404 0.8116 0.9009
No log 1.6316 124 0.8079 0.3097 0.8079 0.8988
No log 1.6579 126 0.8077 0.2611 0.8077 0.8987
No log 1.6842 128 0.7840 0.2308 0.7840 0.8854
No log 1.7105 130 0.7433 0.1136 0.7433 0.8621
No log 1.7368 132 0.7304 0.1829 0.7304 0.8546
No log 1.7632 134 0.7506 0.2294 0.7506 0.8663
No log 1.7895 136 0.7729 0.2237 0.7729 0.8791
No log 1.8158 138 0.7848 0.2994 0.7848 0.8859
No log 1.8421 140 0.8497 0.2884 0.8497 0.9218
No log 1.8684 142 1.0022 0.0917 1.0022 1.0011
No log 1.8947 144 0.9182 0.2844 0.9182 0.9582
No log 1.9211 146 0.8628 0.2718 0.8628 0.9289
No log 1.9474 148 0.9172 0.1998 0.9172 0.9577
No log 1.9737 150 0.8513 0.1740 0.8513 0.9226
No log 2.0 152 0.8287 0.2682 0.8287 0.9104
No log 2.0263 154 1.0580 0.1514 1.0580 1.0286
No log 2.0526 156 1.2151 0.0967 1.2151 1.1023
No log 2.0789 158 1.0996 0.1226 1.0996 1.0486
No log 2.1053 160 0.9008 0.2023 0.9008 0.9491
No log 2.1316 162 0.8581 0.3318 0.8581 0.9264
No log 2.1579 164 0.8978 0.3318 0.8978 0.9475
No log 2.1842 166 0.9626 0.2055 0.9626 0.9811
No log 2.2105 168 0.9850 0.2659 0.9850 0.9925
No log 2.2368 170 0.9015 0.3202 0.9015 0.9495
No log 2.2632 172 0.8785 0.3316 0.8785 0.9373
No log 2.2895 174 0.7655 0.2424 0.7655 0.8750
No log 2.3158 176 0.7414 0.3155 0.7414 0.8610
No log 2.3421 178 0.7149 0.3088 0.7149 0.8455
No log 2.3684 180 0.6795 0.3078 0.6795 0.8243
No log 2.3947 182 0.6896 0.3382 0.6896 0.8304
No log 2.4211 184 0.6835 0.3640 0.6835 0.8268
No log 2.4474 186 0.6646 0.3714 0.6646 0.8152
No log 2.4737 188 0.7816 0.2625 0.7816 0.8841
No log 2.5 190 0.7722 0.2815 0.7722 0.8788
No log 2.5263 192 0.6983 0.4402 0.6983 0.8357
No log 2.5526 194 0.7347 0.3549 0.7347 0.8572
No log 2.5789 196 0.7508 0.3506 0.7508 0.8665
No log 2.6053 198 0.7749 0.4420 0.7749 0.8803
No log 2.6316 200 0.9347 0.2659 0.9347 0.9668
No log 2.6579 202 1.0790 0.1760 1.0790 1.0388
No log 2.6842 204 0.9187 0.2677 0.9187 0.9585
No log 2.7105 206 0.7563 0.3010 0.7563 0.8697
No log 2.7368 208 0.7166 0.2747 0.7166 0.8465
No log 2.7632 210 0.7391 0.3382 0.7391 0.8597
No log 2.7895 212 0.7458 0.3382 0.7458 0.8636
No log 2.8158 214 0.7627 0.2907 0.7627 0.8734
No log 2.8421 216 0.8067 0.2318 0.8067 0.8982
No log 2.8684 218 0.8190 0.2429 0.8190 0.9050
No log 2.8947 220 0.7583 0.2973 0.7583 0.8708
No log 2.9211 222 0.7591 0.3382 0.7591 0.8713
No log 2.9474 224 0.7389 0.3738 0.7389 0.8596
No log 2.9737 226 0.7300 0.3225 0.7300 0.8544
No log 3.0 228 0.8016 0.2737 0.8016 0.8953
No log 3.0263 230 1.0002 0.2316 1.0002 1.0001
No log 3.0526 232 0.9943 0.2337 0.9943 0.9972
No log 3.0789 234 0.8062 0.2013 0.8062 0.8979
No log 3.1053 236 0.8377 0.3121 0.8377 0.9153
No log 3.1316 238 1.0923 0.3608 1.0923 1.0452
No log 3.1579 240 1.0668 0.3418 1.0668 1.0329
No log 3.1842 242 0.8959 0.2643 0.8959 0.9465
No log 3.2105 244 0.7752 0.3015 0.7752 0.8805
No log 3.2368 246 0.8260 0.1823 0.8260 0.9088
No log 3.2632 248 0.9412 0.1875 0.9412 0.9702
No log 3.2895 250 0.9105 0.1891 0.9105 0.9542
No log 3.3158 252 0.8114 0.2092 0.8114 0.9008
No log 3.3421 254 0.8020 0.1884 0.8020 0.8956
No log 3.3684 256 0.8580 0.2988 0.8580 0.9263
No log 3.3947 258 0.8871 0.2370 0.8871 0.9418
No log 3.4211 260 0.8840 0.2962 0.8840 0.9402
No log 3.4474 262 0.8627 0.3290 0.8627 0.9288
No log 3.4737 264 0.8668 0.3209 0.8668 0.9310
No log 3.5 266 0.9194 0.3526 0.9194 0.9589
No log 3.5263 268 0.9846 0.3262 0.9846 0.9923
No log 3.5526 270 0.9140 0.3822 0.9140 0.9560
No log 3.5789 272 0.8424 0.3586 0.8424 0.9178
No log 3.6053 274 0.7950 0.3475 0.7950 0.8916
No log 3.6316 276 0.7508 0.3133 0.7508 0.8665
No log 3.6579 278 0.7716 0.2265 0.7716 0.8784
No log 3.6842 280 0.7988 0.1905 0.7988 0.8937
No log 3.7105 282 0.7901 0.1811 0.7901 0.8888
No log 3.7368 284 0.8244 0.3302 0.8244 0.9080
No log 3.7632 286 0.8324 0.2020 0.8324 0.9123
No log 3.7895 288 0.8332 0.1857 0.8332 0.9128
No log 3.8158 290 0.9264 0.2733 0.9264 0.9625
No log 3.8421 292 1.0006 0.2303 1.0006 1.0003
No log 3.8684 294 0.9318 0.1823 0.9318 0.9653
No log 3.8947 296 0.7971 0.3170 0.7971 0.8928
No log 3.9211 298 0.7619 0.1686 0.7619 0.8729
No log 3.9474 300 0.7867 0.2149 0.7867 0.8869
No log 3.9737 302 0.8118 0.2932 0.8118 0.9010
No log 4.0 304 0.8000 0.2662 0.8000 0.8944
No log 4.0263 306 0.7651 0.2334 0.7651 0.8747
No log 4.0526 308 0.7808 0.2053 0.7808 0.8837
No log 4.0789 310 0.7773 0.2213 0.7773 0.8816
No log 4.1053 312 0.8037 0.3018 0.8037 0.8965
No log 4.1316 314 0.8511 0.2993 0.8511 0.9225
No log 4.1579 316 0.8400 0.3231 0.8400 0.9165
No log 4.1842 318 0.7359 0.3353 0.7359 0.8578
No log 4.2105 320 0.6988 0.2862 0.6988 0.8359
No log 4.2368 322 0.6967 0.2661 0.6967 0.8347
No log 4.2632 324 0.7303 0.3353 0.7303 0.8546
No log 4.2895 326 0.7614 0.3353 0.7614 0.8726
No log 4.3158 328 0.7977 0.3885 0.7977 0.8932
No log 4.3421 330 0.8032 0.3885 0.8032 0.8962
No log 4.3684 332 0.8018 0.3711 0.8018 0.8955
No log 4.3947 334 0.7504 0.3018 0.7504 0.8662
No log 4.4211 336 0.7194 0.3018 0.7194 0.8482
No log 4.4474 338 0.6883 0.3069 0.6883 0.8296
No log 4.4737 340 0.6892 0.3689 0.6892 0.8302
No log 4.5 342 0.6989 0.3638 0.6989 0.8360
No log 4.5263 344 0.7321 0.3363 0.7321 0.8557
No log 4.5526 346 0.7457 0.3631 0.7457 0.8635
No log 4.5789 348 0.7693 0.3631 0.7693 0.8771
No log 4.6053 350 0.7656 0.3417 0.7656 0.8750
No log 4.6316 352 0.7366 0.3906 0.7366 0.8583
No log 4.6579 354 0.7370 0.3906 0.7370 0.8585
No log 4.6842 356 0.7423 0.3978 0.7423 0.8615
No log 4.7105 358 0.7970 0.3239 0.7970 0.8928
No log 4.7368 360 0.7999 0.3331 0.7999 0.8944
No log 4.7632 362 0.7524 0.3715 0.7524 0.8674
No log 4.7895 364 0.7881 0.2633 0.7881 0.8878
No log 4.8158 366 0.8133 0.3146 0.8133 0.9018
No log 4.8421 368 0.7656 0.3574 0.7656 0.8750
No log 4.8684 370 0.7729 0.3201 0.7729 0.8792
No log 4.8947 372 0.8148 0.1875 0.8148 0.9026
No log 4.9211 374 0.7877 0.1555 0.7877 0.8875
No log 4.9474 376 0.7416 0.3628 0.7416 0.8612
No log 4.9737 378 0.7576 0.3299 0.7576 0.8704
No log 5.0 380 0.7897 0.3299 0.7897 0.8886
No log 5.0263 382 0.7818 0.3574 0.7818 0.8842
No log 5.0526 384 0.8273 0.2154 0.8273 0.9096
No log 5.0789 386 0.8421 0.2884 0.8421 0.9177
No log 5.1053 388 0.7908 0.2514 0.7908 0.8893
No log 5.1316 390 0.7719 0.3078 0.7719 0.8786
No log 5.1579 392 0.8603 0.2576 0.8603 0.9275
No log 5.1842 394 0.9588 0.2728 0.9588 0.9792
No log 5.2105 396 0.9056 0.3256 0.9056 0.9516
No log 5.2368 398 0.7830 0.2691 0.7830 0.8849
No log 5.2632 400 0.7535 0.2909 0.7535 0.8680
No log 5.2895 402 0.8337 0.1571 0.8337 0.9131
No log 5.3158 404 0.8445 0.1571 0.8445 0.9190
No log 5.3421 406 0.7758 0.2857 0.7758 0.8808
No log 5.3684 408 0.7887 0.2389 0.7887 0.8881
No log 5.3947 410 0.8547 0.3305 0.8547 0.9245
No log 5.4211 412 0.8279 0.2899 0.8279 0.9099
No log 5.4474 414 0.7656 0.3078 0.7656 0.8750
No log 5.4737 416 0.7857 0.2803 0.7857 0.8864
No log 5.5 418 0.8572 0.1878 0.8572 0.9259
No log 5.5263 420 0.8511 0.1878 0.8511 0.9225
No log 5.5526 422 0.7858 0.3265 0.7858 0.8865
No log 5.5789 424 0.7149 0.3198 0.7149 0.8455
No log 5.6053 426 0.7388 0.3088 0.7388 0.8595
No log 5.6316 428 0.7522 0.2784 0.7522 0.8673
No log 5.6579 430 0.7339 0.2784 0.7339 0.8567
No log 5.6842 432 0.7086 0.2215 0.7086 0.8418
No log 5.7105 434 0.7139 0.3002 0.7139 0.8449
No log 5.7368 436 0.7246 0.3239 0.7246 0.8512
No log 5.7632 438 0.7260 0.2838 0.7260 0.8521
No log 5.7895 440 0.7514 0.3078 0.7514 0.8668
No log 5.8158 442 0.8059 0.3026 0.8059 0.8977
No log 5.8421 444 0.8685 0.3005 0.8685 0.9319
No log 5.8684 446 0.8109 0.3121 0.8109 0.9005
No log 5.8947 448 0.7318 0.3618 0.7318 0.8554
No log 5.9211 450 0.7068 0.4006 0.7068 0.8407
No log 5.9474 452 0.7084 0.3792 0.7084 0.8417
No log 5.9737 454 0.7394 0.3417 0.7394 0.8599
No log 6.0 456 0.7383 0.3701 0.7383 0.8593
No log 6.0263 458 0.7087 0.3769 0.7087 0.8418
No log 6.0526 460 0.7401 0.4599 0.7401 0.8603
No log 6.0789 462 0.7277 0.4675 0.7277 0.8530
No log 6.1053 464 0.6886 0.4615 0.6886 0.8298
No log 6.1316 466 0.6743 0.3299 0.6743 0.8212
No log 6.1579 468 0.7145 0.4029 0.7145 0.8453
No log 6.1842 470 0.7283 0.3221 0.7283 0.8534
No log 6.2105 472 0.6809 0.3549 0.6809 0.8252
No log 6.2368 474 0.6590 0.3762 0.6590 0.8118
No log 6.2632 476 0.6836 0.4364 0.6836 0.8268
No log 6.2895 478 0.7096 0.3972 0.7096 0.8424
No log 6.3158 480 0.6991 0.4458 0.6991 0.8361
No log 6.3421 482 0.7007 0.4137 0.7007 0.8371
No log 6.3684 484 0.7063 0.3598 0.7063 0.8404
No log 6.3947 486 0.7198 0.3482 0.7198 0.8484
No log 6.4211 488 0.7076 0.3689 0.7076 0.8412
No log 6.4474 490 0.7320 0.3888 0.7320 0.8556
No log 6.4737 492 0.7365 0.3888 0.7365 0.8582
No log 6.5 494 0.7550 0.3888 0.7550 0.8689
No log 6.5263 496 0.7985 0.3526 0.7985 0.8936
No log 6.5526 498 0.7498 0.3888 0.7498 0.8659
0.3803 6.5789 500 0.6813 0.4147 0.6813 0.8254
0.3803 6.6053 502 0.6741 0.4402 0.6741 0.8211
0.3803 6.6316 504 0.6727 0.4402 0.6727 0.8202
0.3803 6.6579 506 0.6606 0.4535 0.6606 0.8128
0.3803 6.6842 508 0.6566 0.4278 0.6566 0.8103
0.3803 6.7105 510 0.6626 0.3887 0.6626 0.8140
0.3803 6.7368 512 0.6764 0.3887 0.6764 0.8224
0.3803 6.7632 514 0.6786 0.4147 0.6786 0.8238
0.3803 6.7895 516 0.6958 0.4362 0.6958 0.8342
0.3803 6.8158 518 0.7586 0.3586 0.7586 0.8710
0.3803 6.8421 520 0.7662 0.3873 0.7662 0.8753
0.3803 6.8684 522 0.7223 0.3725 0.7223 0.8499
0.3803 6.8947 524 0.7021 0.4458 0.7021 0.8379
0.3803 6.9211 526 0.7088 0.3961 0.7088 0.8419
0.3803 6.9474 528 0.6976 0.3961 0.6976 0.8352
0.3803 6.9737 530 0.6825 0.3961 0.6825 0.8261
0.3803 7.0 532 0.6641 0.3961 0.6641 0.8149
0.3803 7.0263 534 0.6683 0.3961 0.6683 0.8175

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k15_task7_organization

Finetuned
(4019)
this model