ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k4_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.4519
  • Qwk: 0.0781
  • Mse: 1.4519
  • Rmse: 1.2050

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.0498 -0.0019 4.0498 2.0124
No log 0.4 4 2.4761 -0.0040 2.4761 1.5736
No log 0.6 6 1.5345 0.0294 1.5345 1.2388
No log 0.8 8 1.4773 0.0232 1.4773 1.2154
No log 1.0 10 1.2927 0.0642 1.2927 1.1370
No log 1.2 12 1.1116 0.1398 1.1116 1.0543
No log 1.4 14 1.0998 0.1398 1.0998 1.0487
No log 1.6 16 1.1149 0.0944 1.1149 1.0559
No log 1.8 18 1.2037 0.0312 1.2037 1.0971
No log 2.0 20 1.1541 0.1576 1.1541 1.0743
No log 2.2 22 1.1068 0.1011 1.1068 1.0520
No log 2.4 24 1.1367 0.1046 1.1367 1.0662
No log 2.6 26 1.1566 0.2161 1.1566 1.0754
No log 2.8 28 1.2817 0.0520 1.2817 1.1321
No log 3.0 30 1.2800 0.0520 1.2800 1.1314
No log 3.2 32 1.1833 0.1196 1.1833 1.0878
No log 3.4 34 1.2791 0.0987 1.2791 1.1310
No log 3.6 36 1.3068 0.1202 1.3068 1.1431
No log 3.8 38 1.1916 -0.0022 1.1916 1.0916
No log 4.0 40 1.2001 0.1989 1.2001 1.0955
No log 4.2 42 1.1969 0.1195 1.1969 1.0940
No log 4.4 44 1.5277 0.1168 1.5277 1.2360
No log 4.6 46 1.6736 0.0529 1.6736 1.2937
No log 4.8 48 1.7383 -0.1128 1.7383 1.3184
No log 5.0 50 1.6067 -0.1750 1.6067 1.2676
No log 5.2 52 1.3162 -0.0428 1.3162 1.1473
No log 5.4 54 1.1247 0.1137 1.1247 1.0605
No log 5.6 56 1.1473 0.1625 1.1473 1.0711
No log 5.8 58 1.3745 0.0841 1.3745 1.1724
No log 6.0 60 1.6298 -0.0449 1.6298 1.2766
No log 6.2 62 1.7150 -0.0219 1.7150 1.3096
No log 6.4 64 1.7049 0.0 1.7049 1.3057
No log 6.6 66 1.6382 0.1531 1.6382 1.2799
No log 6.8 68 1.6400 0.1166 1.6400 1.2806
No log 7.0 70 1.5793 0.0623 1.5793 1.2567
No log 7.2 72 1.4967 0.0513 1.4967 1.2234
No log 7.4 74 1.4551 0.1053 1.4551 1.2063
No log 7.6 76 1.3815 0.1814 1.3815 1.1754
No log 7.8 78 1.2365 0.1351 1.2365 1.1120
No log 8.0 80 1.2490 0.1351 1.2490 1.1176
No log 8.2 82 1.3663 0.1628 1.3663 1.1689
No log 8.4 84 1.5632 0.1703 1.5632 1.2503
No log 8.6 86 1.5422 0.2126 1.5422 1.2418
No log 8.8 88 1.5550 0.2424 1.5550 1.2470
No log 9.0 90 1.7372 0.1847 1.7372 1.3180
No log 9.2 92 1.7288 0.1141 1.7288 1.3148
No log 9.4 94 1.6154 0.1058 1.6154 1.2710
No log 9.6 96 1.5806 0.1058 1.5806 1.2572
No log 9.8 98 1.4349 0.1142 1.4349 1.1979
No log 10.0 100 1.3656 0.1142 1.3656 1.1686
No log 10.2 102 1.4528 0.1486 1.4528 1.2053
No log 10.4 104 1.6158 0.0946 1.6158 1.2711
No log 10.6 106 1.6305 0.0226 1.6305 1.2769
No log 10.8 108 1.6307 0.1142 1.6307 1.2770
No log 11.0 110 1.5920 0.1486 1.5920 1.2617
No log 11.2 112 1.6885 0.1601 1.6885 1.2994
No log 11.4 114 1.8334 0.2252 1.8334 1.3540
No log 11.6 116 1.8874 0.1559 1.8874 1.3738
No log 11.8 118 1.7783 0.2252 1.7783 1.3335
No log 12.0 120 1.6659 0.1667 1.6659 1.2907
No log 12.2 122 1.6168 0.0786 1.6168 1.2715
No log 12.4 124 1.6373 0.1114 1.6373 1.2796
No log 12.6 126 1.5642 0.0806 1.5642 1.2507
No log 12.8 128 1.3999 0.0931 1.3999 1.1832
No log 13.0 130 1.3652 0.0931 1.3652 1.1684
No log 13.2 132 1.4930 0.0806 1.4930 1.2219
No log 13.4 134 1.5379 0.0806 1.5379 1.2401
No log 13.6 136 1.5133 0.0806 1.5133 1.2302
No log 13.8 138 1.4794 0.1769 1.4794 1.2163
No log 14.0 140 1.5455 0.1498 1.5455 1.2432
No log 14.2 142 1.7737 0.1467 1.7737 1.3318
No log 14.4 144 1.9747 0.1313 1.9747 1.4053
No log 14.6 146 1.9080 0.1323 1.9080 1.3813
No log 14.8 148 1.6673 0.0911 1.6673 1.2912
No log 15.0 150 1.4134 0.0931 1.4134 1.1889
No log 15.2 152 1.3320 0.0833 1.3320 1.1541
No log 15.4 154 1.3475 0.0310 1.3475 1.1608
No log 15.6 156 1.4127 -0.0355 1.4127 1.1886
No log 15.8 158 1.4732 0.0263 1.4732 1.2138
No log 16.0 160 1.6276 0.1462 1.6276 1.2758
No log 16.2 162 1.7241 0.2006 1.7241 1.3130
No log 16.4 164 1.4980 0.2417 1.4980 1.2239
No log 16.6 166 1.4527 0.2690 1.4527 1.2053
No log 16.8 168 1.5997 0.2386 1.5997 1.2648
No log 17.0 170 1.9087 0.1835 1.9087 1.3816
No log 17.2 172 2.0000 0.0981 2.0000 1.4142
No log 17.4 174 1.8277 0.2053 1.8277 1.3519
No log 17.6 176 1.6832 0.2252 1.6832 1.2974
No log 17.8 178 1.6087 0.2566 1.6087 1.2683
No log 18.0 180 1.5513 0.2058 1.5513 1.2455
No log 18.2 182 1.4801 0.1573 1.4801 1.2166
No log 18.4 184 1.5061 0.1288 1.5061 1.2272
No log 18.6 186 1.4586 0.0970 1.4586 1.2077
No log 18.8 188 1.4722 0.0878 1.4722 1.2134
No log 19.0 190 1.4603 0.0510 1.4603 1.2084
No log 19.2 192 1.4311 0.0122 1.4311 1.1963
No log 19.4 194 1.4412 0.0122 1.4412 1.2005
No log 19.6 196 1.4655 0.0510 1.4655 1.2106
No log 19.8 198 1.4548 0.0510 1.4548 1.2061
No log 20.0 200 1.4937 0.0878 1.4937 1.2222
No log 20.2 202 1.5505 0.1462 1.5505 1.2452
No log 20.4 204 1.5724 0.1371 1.5724 1.2540
No log 20.6 206 1.5182 0.0896 1.5182 1.2322
No log 20.8 208 1.4657 0.0806 1.4657 1.2107
No log 21.0 210 1.4546 0.0896 1.4546 1.2061
No log 21.2 212 1.5659 0.1441 1.5659 1.2514
No log 21.4 214 1.7100 0.1902 1.7100 1.3077
No log 21.6 216 1.8689 0.2296 1.8689 1.3671
No log 21.8 218 1.7922 0.2406 1.7922 1.3387
No log 22.0 220 1.5663 0.1789 1.5663 1.2515
No log 22.2 222 1.4093 0.0510 1.4093 1.1871
No log 22.4 224 1.3654 0.0401 1.3654 1.1685
No log 22.6 226 1.5114 0.0401 1.5114 1.2294
No log 22.8 228 1.7181 0.1703 1.7181 1.3108
No log 23.0 230 1.8659 0.2448 1.8659 1.3660
No log 23.2 232 1.8638 0.2193 1.8638 1.3652
No log 23.4 234 1.7207 0.2568 1.7207 1.3117
No log 23.6 236 1.5962 0.0878 1.5962 1.2634
No log 23.8 238 1.5402 0.0510 1.5402 1.2411
No log 24.0 240 1.5540 0.0510 1.5540 1.2466
No log 24.2 242 1.5903 0.0878 1.5903 1.2611
No log 24.4 244 1.6391 0.1703 1.6391 1.2803
No log 24.6 246 1.6749 0.2062 1.6749 1.2942
No log 24.8 248 1.7072 0.2342 1.7072 1.3066
No log 25.0 250 1.7492 0.2292 1.7492 1.3226
No log 25.2 252 1.7911 0.2005 1.7911 1.3383
No log 25.4 254 1.7652 0.1729 1.7652 1.3286
No log 25.6 256 1.7661 0.1789 1.7661 1.3289
No log 25.8 258 1.7260 0.2270 1.7260 1.3138
No log 26.0 260 1.6997 0.1847 1.6997 1.3037
No log 26.2 262 1.7394 0.1752 1.7394 1.3189
No log 26.4 264 1.7005 0.1832 1.7005 1.3040
No log 26.6 266 1.6076 0.2004 1.6076 1.2679
No log 26.8 268 1.6388 0.2292 1.6388 1.2802
No log 27.0 270 1.6334 0.2424 1.6334 1.2781
No log 27.2 272 1.5477 0.0878 1.5477 1.2441
No log 27.4 274 1.4644 0.0781 1.4644 1.2101
No log 27.6 276 1.4810 0.0781 1.4810 1.2170
No log 27.8 278 1.4909 0.0781 1.4909 1.2210
No log 28.0 280 1.5762 0.0781 1.5762 1.2555
No log 28.2 282 1.6106 0.2065 1.6106 1.2691
No log 28.4 284 1.5868 0.0781 1.5868 1.2597
No log 28.6 286 1.5811 0.0781 1.5811 1.2574
No log 28.8 288 1.5858 0.0781 1.5858 1.2593
No log 29.0 290 1.5949 0.0781 1.5949 1.2629
No log 29.2 292 1.6186 0.0781 1.6186 1.2722
No log 29.4 294 1.7056 0.1222 1.7056 1.3060
No log 29.6 296 1.7375 0.0828 1.7375 1.3181
No log 29.8 298 1.6717 0.1222 1.6717 1.2929
No log 30.0 300 1.5594 0.0781 1.5594 1.2488
No log 30.2 302 1.4603 0.0401 1.4603 1.2084
No log 30.4 304 1.3846 0.0 1.3846 1.1767
No log 30.6 306 1.3558 0.0 1.3558 1.1644
No log 30.8 308 1.4206 0.0401 1.4206 1.1919
No log 31.0 310 1.4929 0.0781 1.4929 1.2218
No log 31.2 312 1.6148 0.2065 1.6148 1.2708
No log 31.4 314 1.6460 0.2474 1.6460 1.2830
No log 31.6 316 1.5591 0.1814 1.5591 1.2486
No log 31.8 318 1.4613 0.0781 1.4613 1.2088
No log 32.0 320 1.4139 0.0781 1.4139 1.1891
No log 32.2 322 1.4506 0.0781 1.4506 1.2044
No log 32.4 324 1.5394 0.2372 1.5394 1.2407
No log 32.6 326 1.6589 0.2424 1.6589 1.2880
No log 32.8 328 1.6808 0.2709 1.6808 1.2965
No log 33.0 330 1.5745 0.2126 1.5745 1.2548
No log 33.2 332 1.4481 0.0781 1.4481 1.2034
No log 33.4 334 1.4090 0.0401 1.4090 1.1870
No log 33.6 336 1.3698 0.0401 1.3698 1.1704
No log 33.8 338 1.3875 0.0781 1.3875 1.1779
No log 34.0 340 1.3704 0.0401 1.3704 1.1706
No log 34.2 342 1.3632 0.0833 1.3632 1.1675
No log 34.4 344 1.3741 0.1202 1.3741 1.1722
No log 34.6 346 1.3583 0.1202 1.3583 1.1654
No log 34.8 348 1.3990 0.1202 1.3990 1.1828
No log 35.0 350 1.5546 0.2424 1.5546 1.2469
No log 35.2 352 1.7763 0.2840 1.7763 1.3328
No log 35.4 354 1.9028 0.2494 1.9028 1.3794
No log 35.6 356 1.8911 0.2193 1.8911 1.3752
No log 35.8 358 1.7854 0.3086 1.7854 1.3362
No log 36.0 360 1.5876 0.2126 1.5876 1.2600
No log 36.2 362 1.4447 0.0878 1.4447 1.2019
No log 36.4 364 1.3986 0.0833 1.3986 1.1826
No log 36.6 366 1.3971 0.0401 1.3971 1.1820
No log 36.8 368 1.4199 0.0401 1.4199 1.1916
No log 37.0 370 1.4241 0.0781 1.4241 1.1934
No log 37.2 372 1.3505 0.1202 1.3505 1.1621
No log 37.4 374 1.3212 0.1202 1.3212 1.1494
No log 37.6 376 1.3856 0.1202 1.3856 1.1771
No log 37.8 378 1.5086 0.2424 1.5086 1.2283
No log 38.0 380 1.6151 0.2424 1.6151 1.2708
No log 38.2 382 1.6507 0.2424 1.6507 1.2848
No log 38.4 384 1.5795 0.1562 1.5795 1.2568
No log 38.6 386 1.4796 0.0878 1.4796 1.2164
No log 38.8 388 1.4047 0.0401 1.4047 1.1852
No log 39.0 390 1.3633 0.0 1.3633 1.1676
No log 39.2 392 1.3801 0.0401 1.3801 1.1748
No log 39.4 394 1.4562 0.0878 1.4562 1.2067
No log 39.6 396 1.5552 0.2424 1.5552 1.2471
No log 39.8 398 1.6742 0.2270 1.6742 1.2939
No log 40.0 400 1.8019 0.3200 1.8019 1.3424
No log 40.2 402 1.8259 0.3200 1.8259 1.3513
No log 40.4 404 1.7788 0.3095 1.7788 1.3337
No log 40.6 406 1.6376 0.2709 1.6376 1.2797
No log 40.8 408 1.4944 0.1814 1.4944 1.2225
No log 41.0 410 1.4326 0.0878 1.4326 1.1969
No log 41.2 412 1.3983 0.0401 1.3983 1.1825
No log 41.4 414 1.3910 0.0401 1.3910 1.1794
No log 41.6 416 1.4307 0.0878 1.4307 1.1961
No log 41.8 418 1.4834 0.0878 1.4834 1.2179
No log 42.0 420 1.5147 0.1562 1.5147 1.2307
No log 42.2 422 1.6019 0.2709 1.6019 1.2657
No log 42.4 424 1.6298 0.2568 1.6298 1.2766
No log 42.6 426 1.6047 0.2004 1.6047 1.2667
No log 42.8 428 1.5678 0.1814 1.5678 1.2521
No log 43.0 430 1.5000 0.1486 1.5000 1.2247
No log 43.2 432 1.4733 0.1486 1.4733 1.2138
No log 43.4 434 1.3846 0.0781 1.3846 1.1767
No log 43.6 436 1.3187 0.0833 1.3187 1.1483
No log 43.8 438 1.3267 0.0401 1.3267 1.1518
No log 44.0 440 1.3451 0.0781 1.3451 1.1598
No log 44.2 442 1.4128 0.1142 1.4128 1.1886
No log 44.4 444 1.5253 0.1814 1.5253 1.2350
No log 44.6 446 1.6350 0.2292 1.6350 1.2787
No log 44.8 448 1.6281 0.1462 1.6281 1.2760
No log 45.0 450 1.5409 0.1562 1.5409 1.2413
No log 45.2 452 1.4348 0.0401 1.4348 1.1978
No log 45.4 454 1.3495 0.0 1.3495 1.1617
No log 45.6 456 1.3266 0.0 1.3266 1.1518
No log 45.8 458 1.3508 0.0 1.3508 1.1623
No log 46.0 460 1.4155 0.0401 1.4155 1.1897
No log 46.2 462 1.4770 0.0781 1.4770 1.2153
No log 46.4 464 1.5437 0.1228 1.5437 1.2425
No log 46.6 466 1.5577 0.0878 1.5577 1.2481
No log 46.8 468 1.5606 0.0878 1.5606 1.2492
No log 47.0 470 1.5842 0.1562 1.5842 1.2586
No log 47.2 472 1.5710 0.2424 1.5710 1.2534
No log 47.4 474 1.5688 0.2709 1.5688 1.2525
No log 47.6 476 1.6057 0.2709 1.6057 1.2672
No log 47.8 478 1.6067 0.2709 1.6067 1.2676
No log 48.0 480 1.6084 0.2709 1.6084 1.2682
No log 48.2 482 1.5623 0.2424 1.5623 1.2499
No log 48.4 484 1.5370 0.2372 1.5370 1.2398
No log 48.6 486 1.5594 0.2424 1.5594 1.2487
No log 48.8 488 1.5362 0.1486 1.5362 1.2395
No log 49.0 490 1.4532 0.0781 1.4532 1.2055
No log 49.2 492 1.4360 0.0781 1.4360 1.1983
No log 49.4 494 1.4381 0.1202 1.4381 1.1992
No log 49.6 496 1.4727 0.1552 1.4727 1.2135
No log 49.8 498 1.4955 0.1552 1.4955 1.2229
0.2145 50.0 500 1.5063 0.1142 1.5063 1.2273
0.2145 50.2 502 1.4989 0.0781 1.4989 1.2243
0.2145 50.4 504 1.4633 0.0781 1.4633 1.2097
0.2145 50.6 506 1.4296 0.0401 1.4296 1.1956
0.2145 50.8 508 1.4278 0.0401 1.4278 1.1949
0.2145 51.0 510 1.4519 0.0781 1.4519 1.2050

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k4_task5_organization

Finetuned
(4019)
this model