ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0182
  • Qwk: 0.3024
  • Mse: 1.0182
  • Rmse: 1.0090

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0267 2 4.4352 -0.0313 4.4352 2.1060
No log 0.0533 4 2.7589 -0.0047 2.7589 1.6610
No log 0.08 6 1.3516 0.0287 1.3516 1.1626
No log 0.1067 8 1.5089 -0.0684 1.5089 1.2284
No log 0.1333 10 1.0679 0.1810 1.0679 1.0334
No log 0.16 12 1.0460 0.1573 1.0460 1.0227
No log 0.1867 14 1.0517 0.2268 1.0517 1.0255
No log 0.2133 16 1.2093 0.1057 1.2093 1.0997
No log 0.24 18 1.1013 0.1927 1.1013 1.0494
No log 0.2667 20 1.0614 0.2416 1.0614 1.0302
No log 0.2933 22 1.0897 0.1658 1.0897 1.0439
No log 0.32 24 1.4939 0.1315 1.4939 1.2223
No log 0.3467 26 1.6200 0.1019 1.6200 1.2728
No log 0.3733 28 1.3100 0.2490 1.3100 1.1446
No log 0.4 30 1.1928 0.2495 1.1928 1.0922
No log 0.4267 32 0.9521 0.3562 0.9521 0.9757
No log 0.4533 34 1.1172 0.2905 1.1172 1.0570
No log 0.48 36 1.8512 0.1342 1.8512 1.3606
No log 0.5067 38 1.9586 0.1554 1.9586 1.3995
No log 0.5333 40 1.3337 0.2021 1.3337 1.1549
No log 0.56 42 0.9266 0.2841 0.9266 0.9626
No log 0.5867 44 0.9749 0.3782 0.9749 0.9873
No log 0.6133 46 0.9304 0.3506 0.9304 0.9645
No log 0.64 48 0.9552 0.3353 0.9552 0.9773
No log 0.6667 50 1.1707 0.2260 1.1707 1.0820
No log 0.6933 52 1.1972 0.2260 1.1972 1.0942
No log 0.72 54 0.9896 0.3082 0.9896 0.9948
No log 0.7467 56 0.9275 0.2812 0.9275 0.9630
No log 0.7733 58 0.9400 0.2998 0.9400 0.9695
No log 0.8 60 0.9593 0.3676 0.9593 0.9795
No log 0.8267 62 0.9485 0.2424 0.9485 0.9739
No log 0.8533 64 0.9974 0.3167 0.9974 0.9987
No log 0.88 66 1.1816 0.3480 1.1816 1.0870
No log 0.9067 68 1.2096 0.2317 1.2096 1.0998
No log 0.9333 70 1.1610 0.2611 1.1610 1.0775
No log 0.96 72 1.1583 0.2176 1.1583 1.0762
No log 0.9867 74 1.0527 0.3172 1.0527 1.0260
No log 1.0133 76 1.0191 0.2790 1.0191 1.0095
No log 1.04 78 1.0419 0.3556 1.0419 1.0207
No log 1.0667 80 1.0273 0.3284 1.0273 1.0135
No log 1.0933 82 0.9828 0.4215 0.9828 0.9913
No log 1.12 84 1.0146 0.3622 1.0146 1.0073
No log 1.1467 86 1.1808 0.2007 1.1808 1.0866
No log 1.1733 88 1.0537 0.2721 1.0537 1.0265
No log 1.2 90 0.9881 0.4232 0.9881 0.9940
No log 1.2267 92 1.0728 0.3174 1.0728 1.0358
No log 1.2533 94 0.9645 0.3722 0.9645 0.9821
No log 1.28 96 1.0378 0.3551 1.0378 1.0187
No log 1.3067 98 1.1727 0.2461 1.1727 1.0829
No log 1.3333 100 1.1501 0.3135 1.1501 1.0724
No log 1.3600 102 0.9955 0.3107 0.9955 0.9977
No log 1.3867 104 1.0390 0.2352 1.0390 1.0193
No log 1.4133 106 1.1368 0.2995 1.1368 1.0662
No log 1.44 108 1.0013 0.2549 1.0013 1.0007
No log 1.4667 110 0.9460 0.3151 0.9460 0.9726
No log 1.4933 112 1.1357 0.2037 1.1357 1.0657
No log 1.52 114 1.2223 0.1775 1.2223 1.1056
No log 1.5467 116 1.1165 0.2346 1.1165 1.0566
No log 1.5733 118 0.9065 0.4888 0.9065 0.9521
No log 1.6 120 1.0115 0.3347 1.0115 1.0057
No log 1.6267 122 1.0189 0.3392 1.0189 1.0094
No log 1.6533 124 0.8983 0.4180 0.8983 0.9478
No log 1.6800 126 1.1082 0.2520 1.1082 1.0527
No log 1.7067 128 1.4347 0.2398 1.4347 1.1978
No log 1.7333 130 1.7065 0.1394 1.7065 1.3063
No log 1.76 132 1.6082 0.0978 1.6082 1.2682
No log 1.7867 134 1.3322 0.1576 1.3322 1.1542
No log 1.8133 136 1.1428 0.1105 1.1428 1.0690
No log 1.8400 138 1.0585 0.0986 1.0585 1.0288
No log 1.8667 140 1.0631 0.1729 1.0631 1.0311
No log 1.8933 142 1.1645 0.2195 1.1645 1.0791
No log 1.92 144 1.3568 0.1500 1.3568 1.1648
No log 1.9467 146 1.6057 0.1520 1.6057 1.2671
No log 1.9733 148 1.5733 0.1310 1.5733 1.2543
No log 2.0 150 1.4183 0.1322 1.4183 1.1909
No log 2.0267 152 1.2796 0.1622 1.2796 1.1312
No log 2.0533 154 1.2764 0.1622 1.2764 1.1298
No log 2.08 156 1.2377 0.1998 1.2377 1.1125
No log 2.1067 158 1.1040 0.2951 1.1040 1.0507
No log 2.1333 160 1.0192 0.3455 1.0192 1.0096
No log 2.16 162 1.0238 0.3378 1.0238 1.0118
No log 2.1867 164 1.1481 0.2487 1.1481 1.0715
No log 2.2133 166 1.2140 0.2355 1.2140 1.1018
No log 2.24 168 1.1644 0.3578 1.1644 1.0791
No log 2.2667 170 1.0576 0.3115 1.0576 1.0284
No log 2.2933 172 0.9646 0.3973 0.9646 0.9821
No log 2.32 174 0.9967 0.4039 0.9967 0.9984
No log 2.3467 176 1.0438 0.3156 1.0438 1.0217
No log 2.3733 178 1.0186 0.3539 1.0186 1.0093
No log 2.4 180 1.0306 0.3826 1.0306 1.0152
No log 2.4267 182 0.9971 0.4343 0.9971 0.9986
No log 2.4533 184 0.9803 0.3990 0.9803 0.9901
No log 2.48 186 1.0246 0.3938 1.0246 1.0122
No log 2.5067 188 1.0545 0.3954 1.0545 1.0269
No log 2.5333 190 0.9626 0.4343 0.9626 0.9811
No log 2.56 192 0.9040 0.3974 0.9040 0.9508
No log 2.5867 194 0.8853 0.3188 0.8853 0.9409
No log 2.6133 196 0.9218 0.3383 0.9218 0.9601
No log 2.64 198 0.9066 0.2819 0.9066 0.9522
No log 2.6667 200 0.8834 0.2910 0.8834 0.9399
No log 2.6933 202 0.8935 0.2910 0.8935 0.9453
No log 2.7200 204 0.9307 0.2432 0.9307 0.9647
No log 2.7467 206 1.0917 0.3443 1.0917 1.0448
No log 2.7733 208 1.1209 0.3968 1.1209 1.0587
No log 2.8 210 0.9765 0.3806 0.9765 0.9882
No log 2.8267 212 0.9018 0.4352 0.9018 0.9496
No log 2.8533 214 0.9096 0.3804 0.9096 0.9537
No log 2.88 216 0.8683 0.4244 0.8683 0.9318
No log 2.9067 218 0.8458 0.4279 0.8458 0.9197
No log 2.9333 220 0.8718 0.3637 0.8718 0.9337
No log 2.96 222 1.0015 0.3766 1.0015 1.0007
No log 2.9867 224 1.0456 0.3385 1.0456 1.0225
No log 3.0133 226 0.9563 0.3648 0.9563 0.9779
No log 3.04 228 0.8336 0.4234 0.8336 0.9130
No log 3.0667 230 0.8399 0.4381 0.8399 0.9165
No log 3.0933 232 0.8485 0.3838 0.8485 0.9212
No log 3.12 234 0.9186 0.3668 0.9186 0.9584
No log 3.1467 236 1.0031 0.3140 1.0031 1.0016
No log 3.1733 238 1.0467 0.3531 1.0467 1.0231
No log 3.2 240 1.0875 0.2961 1.0875 1.0428
No log 3.2267 242 1.1317 0.3119 1.1317 1.0638
No log 3.2533 244 1.0769 0.3229 1.0769 1.0377
No log 3.2800 246 1.0849 0.3229 1.0849 1.0416
No log 3.3067 248 1.1516 0.2723 1.1516 1.0731
No log 3.3333 250 1.2061 0.2076 1.2061 1.0982
No log 3.36 252 1.1763 0.1259 1.1763 1.0846
No log 3.3867 254 1.0727 0.1418 1.0727 1.0357
No log 3.4133 256 1.0104 0.2214 1.0104 1.0052
No log 3.44 258 1.0321 0.2416 1.0321 1.0159
No log 3.4667 260 1.0423 0.1962 1.0423 1.0209
No log 3.4933 262 1.1367 0.2610 1.1367 1.0662
No log 3.52 264 1.2085 0.2794 1.2085 1.0993
No log 3.5467 266 1.1428 0.2215 1.1428 1.0690
No log 3.5733 268 1.0600 0.1927 1.0600 1.0295
No log 3.6 270 1.0405 0.2061 1.0405 1.0200
No log 3.6267 272 1.0472 0.1962 1.0472 1.0234
No log 3.6533 274 1.0755 0.2695 1.0755 1.0371
No log 3.68 276 1.1347 0.2701 1.1347 1.0652
No log 3.7067 278 1.1559 0.2820 1.1559 1.0751
No log 3.7333 280 1.1615 0.2449 1.1615 1.0777
No log 3.76 282 1.2167 0.1170 1.2167 1.1030
No log 3.7867 284 1.2776 0.1882 1.2776 1.1303
No log 3.8133 286 1.2658 0.1611 1.2658 1.1251
No log 3.84 288 1.2171 0.2076 1.2171 1.1032
No log 3.8667 290 1.1475 0.2091 1.1475 1.0712
No log 3.8933 292 1.0975 0.2108 1.0975 1.0476
No log 3.92 294 1.1308 0.1676 1.1308 1.0634
No log 3.9467 296 1.2002 0.1918 1.2002 1.0955
No log 3.9733 298 1.2069 0.1500 1.2069 1.0986
No log 4.0 300 1.1199 0.1170 1.1199 1.0583
No log 4.0267 302 1.0307 0.2416 1.0307 1.0152
No log 4.0533 304 1.0134 0.2746 1.0134 1.0067
No log 4.08 306 1.0849 0.2505 1.0849 1.0416
No log 4.1067 308 1.1832 0.2812 1.1832 1.0877
No log 4.1333 310 1.1842 0.2282 1.1842 1.0882
No log 4.16 312 1.0614 0.2505 1.0614 1.0303
No log 4.1867 314 1.0588 0.2325 1.0588 1.0290
No log 4.2133 316 1.1807 0.1935 1.1807 1.0866
No log 4.24 318 1.3297 0.2667 1.3297 1.1531
No log 4.2667 320 1.2436 0.2837 1.2436 1.1152
No log 4.2933 322 1.0559 0.2647 1.0559 1.0276
No log 4.32 324 1.0337 0.3011 1.0337 1.0167
No log 4.3467 326 1.0515 0.3112 1.0515 1.0254
No log 4.3733 328 1.1586 0.2680 1.1586 1.0764
No log 4.4 330 1.2469 0.1214 1.2469 1.1166
No log 4.4267 332 1.2903 0.1476 1.2903 1.1359
No log 4.4533 334 1.2027 0.1944 1.2027 1.0967
No log 4.48 336 1.0642 0.2647 1.0642 1.0316
No log 4.5067 338 1.0305 0.3112 1.0305 1.0151
No log 4.5333 340 1.0343 0.3112 1.0343 1.0170
No log 4.5600 342 1.0778 0.2623 1.0778 1.0382
No log 4.5867 344 1.0698 0.3085 1.0698 1.0343
No log 4.6133 346 1.0700 0.2623 1.0700 1.0344
No log 4.64 348 1.0687 0.2623 1.0687 1.0338
No log 4.6667 350 1.0425 0.2400 1.0425 1.0210
No log 4.6933 352 1.0350 0.2695 1.0350 1.0174
No log 4.72 354 1.0340 0.2695 1.0340 1.0169
No log 4.7467 356 1.0931 0.2408 1.0931 1.0455
No log 4.7733 358 1.1333 0.2298 1.1333 1.0646
No log 4.8 360 1.0891 0.2204 1.0891 1.0436
No log 4.8267 362 1.0439 0.3107 1.0439 1.0217
No log 4.8533 364 1.0618 0.2698 1.0618 1.0304
No log 4.88 366 1.1274 0.2651 1.1274 1.0618
No log 4.9067 368 1.0949 0.2325 1.0949 1.0464
No log 4.9333 370 1.0330 0.2467 1.0330 1.0163
No log 4.96 372 1.0149 0.2643 1.0149 1.0074
No log 4.9867 374 1.0185 0.2492 1.0185 1.0092
No log 5.0133 376 1.0588 0.1981 1.0588 1.0290
No log 5.04 378 1.1363 0.1233 1.1363 1.0660
No log 5.0667 380 1.1340 0.2091 1.1340 1.0649
No log 5.0933 382 1.0475 0.2896 1.0475 1.0235
No log 5.12 384 0.9862 0.2671 0.9862 0.9931
No log 5.1467 386 0.9772 0.2921 0.9772 0.9885
No log 5.1733 388 0.9999 0.2771 0.9999 0.9999
No log 5.2 390 1.1077 0.2298 1.1077 1.0525
No log 5.2267 392 1.1834 0.1935 1.1834 1.0878
No log 5.2533 394 1.1821 0.1259 1.1821 1.0873
No log 5.28 396 1.1142 0.1389 1.1142 1.0556
No log 5.3067 398 1.0783 0.1545 1.0783 1.0384
No log 5.3333 400 1.0330 0.2135 1.0330 1.0164
No log 5.36 402 1.0029 0.2919 1.0029 1.0015
No log 5.3867 404 1.0337 0.3383 1.0337 1.0167
No log 5.4133 406 1.0507 0.3207 1.0507 1.0250
No log 5.44 408 1.0147 0.2896 1.0147 1.0073
No log 5.4667 410 1.0211 0.2416 1.0211 1.0105
No log 5.4933 412 1.0625 0.2175 1.0625 1.0308
No log 5.52 414 1.0907 0.2175 1.0907 1.0444
No log 5.5467 416 1.0784 0.2175 1.0784 1.0385
No log 5.5733 418 1.0252 0.2748 1.0252 1.0125
No log 5.6 420 0.9684 0.2492 0.9684 0.9841
No log 5.6267 422 0.9553 0.2492 0.9553 0.9774
No log 5.6533 424 0.9641 0.2492 0.9641 0.9819
No log 5.68 426 1.0230 0.3085 1.0230 1.0114
No log 5.7067 428 1.1380 0.1797 1.1380 1.0668
No log 5.7333 430 1.1766 0.2376 1.1766 1.0847
No log 5.76 432 1.1602 0.1797 1.1602 1.0771
No log 5.7867 434 1.1441 0.2165 1.1441 1.0696
No log 5.8133 436 1.0929 0.2674 1.0929 1.0454
No log 5.84 438 1.0046 0.2594 1.0046 1.0023
No log 5.8667 440 1.0091 0.2897 1.0091 1.0045
No log 5.8933 442 1.0672 0.2161 1.0672 1.0331
No log 5.92 444 1.1296 0.1474 1.1296 1.0628
No log 5.9467 446 1.2071 0.1110 1.2071 1.0987
No log 5.9733 448 1.1724 0.1528 1.1724 1.0828
No log 6.0 450 1.0894 0.2050 1.0894 1.0438
No log 6.0267 452 1.0405 0.2161 1.0405 1.0201
No log 6.0533 454 1.0045 0.2594 1.0045 1.0023
No log 6.08 456 1.0543 0.2748 1.0543 1.0268
No log 6.1067 458 1.1081 0.3290 1.1081 1.0527
No log 6.1333 460 1.0915 0.3290 1.0915 1.0447
No log 6.16 462 1.0533 0.2440 1.0533 1.0263
No log 6.1867 464 1.0100 0.2135 1.0100 1.0050
No log 6.2133 466 1.0117 0.2569 1.0117 1.0058
No log 6.24 468 0.9831 0.2721 0.9831 0.9915
No log 6.2667 470 0.9759 0.2721 0.9759 0.9879
No log 6.2933 472 0.9899 0.2721 0.9899 0.9950
No log 6.32 474 1.0184 0.2135 1.0184 1.0092
No log 6.3467 476 1.0584 0.1900 1.0584 1.0288
No log 6.3733 478 1.1209 0.2522 1.1209 1.0587
No log 6.4 480 1.0769 0.3156 1.0769 1.0377
No log 6.4267 482 0.9760 0.4180 0.9760 0.9879
No log 6.4533 484 0.9913 0.4180 0.9913 0.9956
No log 6.48 486 1.1357 0.3059 1.1357 1.0657
No log 6.5067 488 1.2152 0.2906 1.2152 1.1023
No log 6.5333 490 1.1763 0.1986 1.1763 1.0846
No log 6.5600 492 1.0618 0.1981 1.0618 1.0304
No log 6.5867 494 1.0054 0.2849 1.0054 1.0027
No log 6.6133 496 1.0115 0.2849 1.0115 1.0057
No log 6.64 498 1.0760 0.2748 1.0760 1.0373
0.2831 6.6667 500 1.1436 0.3434 1.1436 1.0694
0.2831 6.6933 502 1.2391 0.3089 1.2391 1.1131
0.2831 6.72 504 1.2376 0.3336 1.2376 1.1125
0.2831 6.7467 506 1.1019 0.3614 1.1019 1.0497
0.2831 6.7733 508 0.9730 0.3974 0.9730 0.9864
0.2831 6.8 510 0.9746 0.3822 0.9746 0.9872
0.2831 6.8267 512 1.0909 0.3079 1.0909 1.0445
0.2831 6.8533 514 1.2200 0.2667 1.2200 1.1045
0.2831 6.88 516 1.2159 0.2623 1.2159 1.1027
0.2831 6.9067 518 1.0994 0.3042 1.0994 1.0485
0.2831 6.9333 520 1.0182 0.3024 1.0182 1.0090

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task5_organization

Finetuned
(4019)
this model