ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.4102
  • Qwk: 0.0401
  • Mse: 1.4102
  • Rmse: 1.1875

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.0498 -0.0019 4.0498 2.0124
No log 0.4 4 2.4761 -0.0040 2.4761 1.5736
No log 0.6 6 1.5345 0.0294 1.5345 1.2388
No log 0.8 8 1.4773 0.0232 1.4773 1.2154
No log 1.0 10 1.2927 0.0642 1.2927 1.1370
No log 1.2 12 1.1116 0.1398 1.1116 1.0543
No log 1.4 14 1.0998 0.1398 1.0998 1.0487
No log 1.6 16 1.1149 0.0944 1.1149 1.0559
No log 1.8 18 1.2037 0.0312 1.2037 1.0971
No log 2.0 20 1.1541 0.1576 1.1541 1.0743
No log 2.2 22 1.1068 0.1011 1.1068 1.0520
No log 2.4 24 1.1367 0.1046 1.1367 1.0662
No log 2.6 26 1.1566 0.2161 1.1566 1.0754
No log 2.8 28 1.2817 0.0520 1.2817 1.1321
No log 3.0 30 1.2800 0.0520 1.2800 1.1314
No log 3.2 32 1.1833 0.1196 1.1833 1.0878
No log 3.4 34 1.2791 0.0987 1.2791 1.1310
No log 3.6 36 1.3068 0.1202 1.3068 1.1431
No log 3.8 38 1.1916 -0.0022 1.1916 1.0916
No log 4.0 40 1.2001 0.1989 1.2001 1.0955
No log 4.2 42 1.1969 0.1195 1.1969 1.0940
No log 4.4 44 1.5278 0.1168 1.5278 1.2360
No log 4.6 46 1.6737 0.0529 1.6737 1.2937
No log 4.8 48 1.7383 -0.1128 1.7383 1.3185
No log 5.0 50 1.6067 -0.1750 1.6067 1.2676
No log 5.2 52 1.3162 -0.0428 1.3162 1.1473
No log 5.4 54 1.1247 0.1137 1.1247 1.0605
No log 5.6 56 1.1473 0.1625 1.1473 1.0711
No log 5.8 58 1.3744 0.0841 1.3744 1.1724
No log 6.0 60 1.6298 -0.0449 1.6298 1.2766
No log 6.2 62 1.7150 -0.0219 1.7150 1.3096
No log 6.4 64 1.7049 0.0 1.7049 1.3057
No log 6.6 66 1.6382 0.1531 1.6382 1.2799
No log 6.8 68 1.6400 0.1166 1.6400 1.2806
No log 7.0 70 1.5793 0.0623 1.5793 1.2567
No log 7.2 72 1.4968 0.0615 1.4968 1.2234
No log 7.4 74 1.4552 0.1053 1.4552 1.2063
No log 7.6 76 1.3814 0.1814 1.3814 1.1753
No log 7.8 78 1.2365 0.1351 1.2365 1.1120
No log 8.0 80 1.2489 0.1351 1.2489 1.1176
No log 8.2 82 1.3662 0.1628 1.3662 1.1689
No log 8.4 84 1.5632 0.1703 1.5632 1.2503
No log 8.6 86 1.5422 0.2126 1.5422 1.2418
No log 8.8 88 1.5549 0.2424 1.5549 1.2470
No log 9.0 90 1.7371 0.1847 1.7371 1.3180
No log 9.2 92 1.7287 0.1141 1.7287 1.3148
No log 9.4 94 1.6153 0.1058 1.6153 1.2709
No log 9.6 96 1.5805 0.1058 1.5805 1.2572
No log 9.8 98 1.4348 0.1142 1.4348 1.1978
No log 10.0 100 1.3654 0.1142 1.3654 1.1685
No log 10.2 102 1.4526 0.1486 1.4526 1.2052
No log 10.4 104 1.6156 0.0946 1.6156 1.2711
No log 10.6 106 1.6302 0.0226 1.6302 1.2768
No log 10.8 108 1.6305 0.1142 1.6305 1.2769
No log 11.0 110 1.5920 0.1486 1.5920 1.2618
No log 11.2 112 1.6886 0.1601 1.6886 1.2995
No log 11.4 114 1.8335 0.2252 1.8335 1.3541
No log 11.6 116 1.8873 0.1559 1.8873 1.3738
No log 11.8 118 1.7781 0.2252 1.7781 1.3335
No log 12.0 120 1.6658 0.1667 1.6658 1.2907
No log 12.2 122 1.6168 0.0786 1.6168 1.2715
No log 12.4 124 1.6373 0.1114 1.6373 1.2796
No log 12.6 126 1.5642 0.0806 1.5642 1.2507
No log 12.8 128 1.3999 0.0931 1.3999 1.1832
No log 13.0 130 1.3650 0.0931 1.3650 1.1683
No log 13.2 132 1.4927 0.0806 1.4927 1.2218
No log 13.4 134 1.5378 0.0806 1.5378 1.2401
No log 13.6 136 1.5138 0.0806 1.5138 1.2304
No log 13.8 138 1.4805 0.1769 1.4805 1.2168
No log 14.0 140 1.5469 0.1498 1.5469 1.2438
No log 14.2 142 1.7746 0.1467 1.7746 1.3321
No log 14.4 144 1.9749 0.1313 1.9749 1.4053
No log 14.6 146 1.9077 0.1323 1.9077 1.3812
No log 14.8 148 1.6668 0.0828 1.6668 1.2911
No log 15.0 150 1.4133 0.0931 1.4133 1.1888
No log 15.2 152 1.3321 0.0833 1.3321 1.1542
No log 15.4 154 1.3478 0.0310 1.3478 1.1610
No log 15.6 156 1.4129 -0.0355 1.4129 1.1886
No log 15.8 158 1.4733 0.0263 1.4733 1.2138
No log 16.0 160 1.6277 0.1462 1.6277 1.2758
No log 16.2 162 1.7238 0.2006 1.7238 1.3129
No log 16.4 164 1.4970 0.2417 1.4970 1.2235
No log 16.6 166 1.4515 0.2690 1.4515 1.2048
No log 16.8 168 1.5985 0.2386 1.5985 1.2643
No log 17.0 170 1.9078 0.1835 1.9078 1.3812
No log 17.2 172 1.9993 0.0981 1.9993 1.4140
No log 17.4 174 1.8276 0.2053 1.8276 1.3519
No log 17.6 176 1.6837 0.2252 1.6837 1.2976
No log 17.8 178 1.6094 0.2566 1.6094 1.2686
No log 18.0 180 1.5513 0.2058 1.5513 1.2455
No log 18.2 182 1.4789 0.1573 1.4789 1.2161
No log 18.4 184 1.5058 0.1288 1.5058 1.2271
No log 18.6 186 1.4586 0.0970 1.4586 1.2077
No log 18.8 188 1.4723 0.0878 1.4723 1.2134
No log 19.0 190 1.4598 0.0510 1.4598 1.2082
No log 19.2 192 1.4299 0.0122 1.4299 1.1958
No log 19.4 194 1.4404 0.0122 1.4404 1.2002
No log 19.6 196 1.4663 0.0510 1.4663 1.2109
No log 19.8 198 1.4560 0.0510 1.4560 1.2066
No log 20.0 200 1.4952 0.0878 1.4952 1.2228
No log 20.2 202 1.5508 0.1462 1.5508 1.2453
No log 20.4 204 1.5710 0.1371 1.5710 1.2534
No log 20.6 206 1.5164 0.0896 1.5164 1.2314
No log 20.8 208 1.4645 0.0806 1.4645 1.2101
No log 21.0 210 1.4549 0.0896 1.4549 1.2062
No log 21.2 212 1.5677 0.1441 1.5677 1.2521
No log 21.4 214 1.7095 0.1902 1.7095 1.3075
No log 21.6 216 1.8639 0.2056 1.8639 1.3653
No log 21.8 218 1.7856 0.2406 1.7856 1.3363
No log 22.0 220 1.5619 0.1729 1.5619 1.2497
No log 22.2 222 1.4111 0.0510 1.4111 1.1879
No log 22.4 224 1.3698 0.0401 1.3698 1.1704
No log 22.6 226 1.5159 0.0401 1.5159 1.2312
No log 22.8 228 1.7175 0.1703 1.7175 1.3105
No log 23.0 230 1.8611 0.2448 1.8611 1.3642
No log 23.2 232 1.8564 0.2488 1.8564 1.3625
No log 23.4 234 1.7135 0.2292 1.7135 1.3090
No log 23.6 236 1.5944 0.0878 1.5944 1.2627
No log 23.8 238 1.5408 0.0510 1.5408 1.2413
No log 24.0 240 1.5522 0.0510 1.5522 1.2459
No log 24.2 242 1.5882 0.0510 1.5882 1.2602
No log 24.4 244 1.6420 0.1703 1.6420 1.2814
No log 24.6 246 1.6790 0.2062 1.6790 1.2957
No log 24.8 248 1.7049 0.2292 1.7049 1.3057
No log 25.0 250 1.7393 0.2292 1.7393 1.3188
No log 25.2 252 1.7825 0.2391 1.7825 1.3351
No log 25.4 254 1.7613 0.1729 1.7613 1.3271
No log 25.6 256 1.7461 0.1729 1.7461 1.3214
No log 25.8 258 1.6906 0.2062 1.6906 1.3002
No log 26.0 260 1.6594 0.2117 1.6594 1.2882
No log 26.2 262 1.7066 0.1950 1.7066 1.3064
No log 26.4 264 1.6797 0.1832 1.6797 1.2960
No log 26.6 266 1.6258 0.2004 1.6258 1.2751
No log 26.8 268 1.7060 0.1729 1.7060 1.3062
No log 27.0 270 1.7249 0.1892 1.7249 1.3134
No log 27.2 272 1.6540 0.1814 1.6540 1.2861
No log 27.4 274 1.5383 0.0878 1.5383 1.2403
No log 27.6 276 1.4535 0.0781 1.4535 1.2056
No log 27.8 278 1.3746 0.0781 1.3746 1.1724
No log 28.0 280 1.4048 0.0781 1.4048 1.1852
No log 28.2 282 1.4214 0.0781 1.4214 1.1922
No log 28.4 284 1.5320 0.1142 1.5320 1.2377
No log 28.6 286 1.7115 0.2004 1.7115 1.3082
No log 28.8 288 1.8136 0.1729 1.8136 1.3467
No log 29.0 290 1.8273 0.1729 1.8273 1.3518
No log 29.2 292 1.7652 0.1441 1.7652 1.3286
No log 29.4 294 1.7647 0.1441 1.7647 1.3284
No log 29.6 296 1.6837 0.2004 1.6837 1.2976
No log 29.8 298 1.5792 0.1486 1.5792 1.2567
No log 30.0 300 1.4932 0.1486 1.4932 1.2220
No log 30.2 302 1.4079 0.0401 1.4079 1.1866
No log 30.4 304 1.3345 0.0 1.3345 1.1552
No log 30.6 306 1.3135 0.0 1.3135 1.1461
No log 30.8 308 1.4074 0.0401 1.4074 1.1863
No log 31.0 310 1.5220 0.1142 1.5220 1.2337
No log 31.2 312 1.6803 0.2474 1.6803 1.2963
No log 31.4 314 1.7273 0.2474 1.7273 1.3143
No log 31.6 316 1.6615 0.2474 1.6615 1.2890
No log 31.8 318 1.5954 0.1486 1.5954 1.2631
No log 32.0 320 1.4951 0.0401 1.4951 1.2228
No log 32.2 322 1.4027 0.0 1.4027 1.1844
No log 32.4 324 1.3964 0.0 1.3964 1.1817
No log 32.6 326 1.4673 0.0781 1.4673 1.2113
No log 32.8 328 1.5876 0.2126 1.5876 1.2600
No log 33.0 330 1.6714 0.2709 1.6714 1.2928
No log 33.2 332 1.6641 0.2126 1.6641 1.2900
No log 33.4 334 1.6247 0.2126 1.6247 1.2746
No log 33.6 336 1.5407 0.1814 1.5407 1.2413
No log 33.8 338 1.4519 0.0878 1.4519 1.2050
No log 34.0 340 1.3393 0.0 1.3393 1.1573
No log 34.2 342 1.3015 0.0445 1.3015 1.1408
No log 34.4 344 1.3298 0.0 1.3298 1.1532
No log 34.6 346 1.3996 0.0 1.3996 1.1830
No log 34.8 348 1.5258 0.0878 1.5258 1.2352
No log 35.0 350 1.7377 0.1533 1.7377 1.3182
No log 35.2 352 1.9137 0.1822 1.9137 1.3833
No log 35.4 354 1.9696 0.2051 1.9696 1.4034
No log 35.6 356 1.9028 0.2193 1.9028 1.3794
No log 35.8 358 1.7479 0.2566 1.7479 1.3221
No log 36.0 360 1.5548 0.1943 1.5548 1.2469
No log 36.2 362 1.4301 0.1814 1.4301 1.1958
No log 36.4 364 1.3956 0.0510 1.3956 1.1814
No log 36.6 366 1.3763 0.0122 1.3763 1.1732
No log 36.8 368 1.3887 0.0 1.3887 1.1784
No log 37.0 370 1.4038 0.0510 1.4038 1.1848
No log 37.2 372 1.4025 0.0510 1.4025 1.1843
No log 37.4 374 1.4473 0.0510 1.4473 1.2030
No log 37.6 376 1.4883 0.1228 1.4883 1.2200
No log 37.8 378 1.5589 0.2126 1.5589 1.2486
No log 38.0 380 1.6037 0.2424 1.6037 1.2664
No log 38.2 382 1.6104 0.2424 1.6104 1.2690
No log 38.4 384 1.5218 0.1228 1.5218 1.2336
No log 38.6 386 1.4294 0.0510 1.4294 1.1956
No log 38.8 388 1.3741 0.0 1.3741 1.1722
No log 39.0 390 1.3618 0.0 1.3618 1.1669
No log 39.2 392 1.3953 0.0 1.3953 1.1812
No log 39.4 394 1.4529 0.0510 1.4529 1.2054
No log 39.6 396 1.5061 0.0878 1.5061 1.2272
No log 39.8 398 1.5802 0.2424 1.5802 1.2571
No log 40.0 400 1.6654 0.2982 1.6654 1.2905
No log 40.2 402 1.6763 0.3086 1.6763 1.2947
No log 40.4 404 1.6473 0.2832 1.6473 1.2835
No log 40.6 406 1.5620 0.2126 1.5620 1.2498
No log 40.8 408 1.5255 0.1814 1.5255 1.2351
No log 41.0 410 1.5802 0.2424 1.5802 1.2570
No log 41.2 412 1.5907 0.2424 1.5907 1.2612
No log 41.4 414 1.5676 0.2424 1.5676 1.2520
No log 41.6 416 1.5632 0.1814 1.5632 1.2503
No log 41.8 418 1.5374 0.0878 1.5374 1.2399
No log 42.0 420 1.4800 0.0510 1.4800 1.2165
No log 42.2 422 1.4926 0.0510 1.4926 1.2217
No log 42.4 424 1.4646 0.0510 1.4646 1.2102
No log 42.6 426 1.4533 0.0401 1.4533 1.2055
No log 42.8 428 1.4761 0.0781 1.4761 1.2150
No log 43.0 430 1.5426 0.0878 1.5426 1.2420
No log 43.2 432 1.6211 0.2424 1.6211 1.2732
No log 43.4 434 1.6103 0.2424 1.6103 1.2690
No log 43.6 436 1.5458 0.1486 1.5458 1.2433
No log 43.8 438 1.5384 0.0781 1.5384 1.2403
No log 44.0 440 1.5025 0.0401 1.5025 1.2258
No log 44.2 442 1.5143 0.0781 1.5143 1.2306
No log 44.4 444 1.5331 0.0878 1.5331 1.2382
No log 44.6 446 1.5535 0.0878 1.5535 1.2464
No log 44.8 448 1.5290 0.0878 1.5290 1.2365
No log 45.0 450 1.4567 0.0401 1.4567 1.2069
No log 45.2 452 1.3745 0.0401 1.3745 1.1724
No log 45.4 454 1.3125 0.0445 1.3125 1.1456
No log 45.6 456 1.3428 0.0401 1.3428 1.1588
No log 45.8 458 1.4199 0.0401 1.4199 1.1916
No log 46.0 460 1.5429 0.0878 1.5429 1.2421
No log 46.2 462 1.6184 0.1486 1.6184 1.2722
No log 46.4 464 1.6063 0.0510 1.6063 1.2674
No log 46.6 466 1.5314 0.0510 1.5314 1.2375
No log 46.8 468 1.4773 0.0401 1.4773 1.2155
No log 47.0 470 1.4701 0.0401 1.4701 1.2125
No log 47.2 472 1.4609 0.0781 1.4609 1.2087
No log 47.4 474 1.4881 0.2372 1.4881 1.2199
No log 47.6 476 1.5905 0.2437 1.5905 1.2611
No log 47.8 478 1.6316 0.2406 1.6316 1.2773
No log 48.0 480 1.6611 0.2448 1.6611 1.2888
No log 48.2 482 1.6178 0.2694 1.6178 1.2719
No log 48.4 484 1.5943 0.3018 1.5943 1.2627
No log 48.6 486 1.5985 0.3018 1.5985 1.2643
No log 48.8 488 1.5590 0.2424 1.5590 1.2486
No log 49.0 490 1.4545 0.1486 1.4545 1.2060
No log 49.2 492 1.4175 0.1886 1.4175 1.1906
No log 49.4 494 1.4205 0.1886 1.4205 1.1918
No log 49.6 496 1.4806 0.2506 1.4806 1.2168
No log 49.8 498 1.5203 0.2424 1.5203 1.2330
0.2164 50.0 500 1.5213 0.2424 1.5213 1.2334
0.2164 50.2 502 1.5012 0.1562 1.5012 1.2252
0.2164 50.4 504 1.4402 0.0401 1.4402 1.2001
0.2164 50.6 506 1.3771 0.0 1.3771 1.1735
0.2164 50.8 508 1.3733 0.0 1.3733 1.1719
0.2164 51.0 510 1.4102 0.0401 1.4102 1.1875

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task5_organization

Finetuned
(4019)
this model