ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.4796
  • Qwk: 0.1744
  • Mse: 1.4796
  • Rmse: 1.2164

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 3.9297 -0.0091 3.9297 1.9823
No log 0.1053 4 2.2644 0.0093 2.2644 1.5048
No log 0.1579 6 1.4707 -0.0342 1.4707 1.2127
No log 0.2105 8 1.1646 0.1901 1.1646 1.0792
No log 0.2632 10 1.1048 0.1891 1.1048 1.0511
No log 0.3158 12 1.4166 -0.0720 1.4166 1.1902
No log 0.3684 14 1.5278 -0.0657 1.5278 1.2361
No log 0.4211 16 1.1902 0.2268 1.1902 1.0910
No log 0.4737 18 1.0738 0.2492 1.0738 1.0362
No log 0.5263 20 1.1138 0.1107 1.1138 1.0554
No log 0.5789 22 1.0448 0.2061 1.0448 1.0222
No log 0.6316 24 1.0492 0.3082 1.0492 1.0243
No log 0.6842 26 1.0534 0.2834 1.0534 1.0263
No log 0.7368 28 0.9923 0.3457 0.9923 0.9961
No log 0.7895 30 1.1923 0.0884 1.1923 1.0919
No log 0.8421 32 1.5746 -0.2749 1.5746 1.2548
No log 0.8947 34 1.4644 -0.1478 1.4644 1.2101
No log 0.9474 36 1.2039 0.0293 1.2039 1.0972
No log 1.0 38 1.1307 0.0855 1.1307 1.0633
No log 1.0526 40 1.3227 -0.0057 1.3227 1.1501
No log 1.1053 42 1.6379 -0.0797 1.6379 1.2798
No log 1.1579 44 1.6627 -0.0210 1.6627 1.2895
No log 1.2105 46 1.8100 -0.0904 1.8100 1.3454
No log 1.2632 48 1.8027 -0.0274 1.8027 1.3427
No log 1.3158 50 1.8947 -0.0763 1.8947 1.3765
No log 1.3684 52 1.8409 0.0178 1.8409 1.3568
No log 1.4211 54 1.6383 0.1545 1.6383 1.2800
No log 1.4737 56 1.7825 -0.1017 1.7825 1.3351
No log 1.5263 58 1.7611 -0.1364 1.7611 1.3271
No log 1.5789 60 1.6069 0.0791 1.6069 1.2676
No log 1.6316 62 1.6913 -0.0381 1.6913 1.3005
No log 1.6842 64 1.6415 -0.0192 1.6415 1.2812
No log 1.7368 66 1.4119 0.1462 1.4119 1.1882
No log 1.7895 68 1.3219 0.1952 1.3219 1.1497
No log 1.8421 70 1.4256 0.1462 1.4256 1.1940
No log 1.8947 72 1.5223 0.1298 1.5223 1.2338
No log 1.9474 74 1.4131 0.1538 1.4131 1.1887
No log 2.0 76 1.2267 0.0464 1.2267 1.1076
No log 2.0526 78 1.0881 0.0496 1.0881 1.0431
No log 2.1053 80 1.0730 0.0618 1.0730 1.0358
No log 2.1579 82 1.1872 0.0527 1.1872 1.0896
No log 2.2105 84 1.5001 -0.0468 1.5001 1.2248
No log 2.2632 86 1.7777 -0.1153 1.7777 1.3333
No log 2.3158 88 1.7310 -0.0084 1.7310 1.3157
No log 2.3684 90 1.7927 0.1211 1.7927 1.3389
No log 2.4211 92 1.8897 0.1559 1.8897 1.3747
No log 2.4737 94 1.7848 0.1918 1.7848 1.3360
No log 2.5263 96 1.6566 0.1752 1.6566 1.2871
No log 2.5789 98 1.4749 0.2184 1.4749 1.2145
No log 2.6316 100 1.6390 0.1950 1.6390 1.2802
No log 2.6842 102 1.8270 0.1911 1.8270 1.3517
No log 2.7368 104 1.8310 0.1607 1.8310 1.3531
No log 2.7895 106 1.7616 0.1423 1.7616 1.3272
No log 2.8421 108 1.5839 0.2372 1.5839 1.2585
No log 2.8947 110 1.6115 0.1769 1.6115 1.2694
No log 2.9474 112 1.7426 0.2005 1.7426 1.3201
No log 3.0 114 1.5861 0.1832 1.5861 1.2594
No log 3.0526 116 1.4765 0.2184 1.4765 1.2151
No log 3.1053 118 1.4913 0.2474 1.4913 1.2212
No log 3.1579 120 1.4850 0.2611 1.4850 1.2186
No log 3.2105 122 1.6912 0.2527 1.6912 1.3005
No log 3.2632 124 1.7547 0.2296 1.7547 1.3247
No log 3.3158 126 1.5767 0.2522 1.5767 1.2557
No log 3.3684 128 1.3353 0.1744 1.3353 1.1556
No log 3.4211 130 1.2308 0.1142 1.2308 1.1094
No log 3.4737 132 1.2191 0.1142 1.2191 1.1041
No log 3.5263 134 1.2827 0.1407 1.2827 1.1326
No log 3.5789 136 1.3880 0.1814 1.3880 1.1781
No log 3.6316 138 1.4316 0.1814 1.4316 1.1965
No log 3.6842 140 1.3256 0.1744 1.3256 1.1514
No log 3.7368 142 1.2272 0.1538 1.2272 1.1078
No log 3.7895 144 1.2630 0.1725 1.2630 1.1238
No log 3.8421 146 1.4161 0.1744 1.4161 1.1900
No log 3.8947 148 1.6288 0.1298 1.6288 1.2763
No log 3.9474 150 1.7018 0.0500 1.7018 1.3045
No log 4.0 152 1.6058 0.0694 1.6058 1.2672
No log 4.0526 154 1.4152 0.1228 1.4152 1.1896
No log 4.1053 156 1.1930 0.1142 1.1930 1.0923
No log 4.1579 158 1.0785 0.1886 1.0785 1.0385
No log 4.2105 160 1.0833 0.2506 1.0833 1.0408
No log 4.2632 162 1.1443 0.2372 1.1443 1.0697
No log 4.3158 164 1.2600 0.2372 1.2600 1.1225
No log 4.3684 166 1.4573 0.2752 1.4573 1.2072
No log 4.4211 168 1.6630 0.2252 1.6630 1.2896
No log 4.4737 170 1.6235 0.2363 1.6235 1.2742
No log 4.5263 172 1.4592 0.2126 1.4592 1.2080
No log 4.5789 174 1.3272 0.2126 1.3272 1.1520
No log 4.6316 176 1.2588 0.2372 1.2588 1.1220
No log 4.6842 178 1.2214 0.2372 1.2214 1.1052
No log 4.7368 180 1.2314 0.1744 1.2314 1.1097
No log 4.7895 182 1.2735 0.1744 1.2735 1.1285
No log 4.8421 184 1.2229 0.1142 1.2229 1.1058
No log 4.8947 186 1.1806 0.1407 1.1806 1.0865
No log 4.9474 188 1.2160 0.1744 1.2160 1.1027
No log 5.0 190 1.2665 0.1744 1.2665 1.1254
No log 5.0526 192 1.3711 0.2372 1.3711 1.1709
No log 5.1053 194 1.5440 0.2568 1.5440 1.2426
No log 5.1579 196 1.7498 0.2840 1.7498 1.3228
No log 5.2105 198 1.9141 0.2222 1.9141 1.3835
No log 5.2632 200 2.0083 0.2172 2.0083 1.4172
No log 5.3158 202 1.9577 0.1925 1.9577 1.3992
No log 5.3684 204 1.8021 0.2007 1.8021 1.3424
No log 5.4211 206 1.6258 0.2525 1.6258 1.2751
No log 5.4737 208 1.3262 0.2424 1.3262 1.1516
No log 5.5263 210 1.2135 0.1744 1.2135 1.1016
No log 5.5789 212 1.2696 0.1744 1.2696 1.1268
No log 5.6316 214 1.3835 0.1744 1.3835 1.1762
No log 5.6842 216 1.5323 0.2126 1.5323 1.2379
No log 5.7368 218 1.6369 0.2117 1.6369 1.2794
No log 5.7895 220 1.5869 0.2062 1.5869 1.2597
No log 5.8421 222 1.5572 0.2117 1.5572 1.2479
No log 5.8947 224 1.5213 0.2522 1.5213 1.2334
No log 5.9474 226 1.3987 0.2126 1.3987 1.1827
No log 6.0 228 1.4577 0.2184 1.4577 1.2073
No log 6.0526 230 1.6843 0.2270 1.6843 1.2978
No log 6.1053 232 1.7583 0.3095 1.7583 1.3260
No log 6.1579 234 1.6287 0.2806 1.6287 1.2762
No log 6.2105 236 1.4436 0.2184 1.4436 1.2015
No log 6.2632 238 1.3642 0.2126 1.3642 1.1680
No log 6.3158 240 1.3977 0.2126 1.3977 1.1822
No log 6.3684 242 1.5417 0.2653 1.5417 1.2417
No log 6.4211 244 1.5790 0.2906 1.5790 1.2566
No log 6.4737 246 1.5083 0.2793 1.5083 1.2281
No log 6.5263 248 1.4159 0.2372 1.4159 1.1899
No log 6.5789 250 1.4045 0.1814 1.4045 1.1851
No log 6.6316 252 1.3809 0.2126 1.3809 1.1751
No log 6.6842 254 1.3700 0.2126 1.3700 1.1705
No log 6.7368 256 1.3292 0.1814 1.3292 1.1529
No log 6.7895 258 1.2753 0.1744 1.2753 1.1293
No log 6.8421 260 1.2630 0.1744 1.2630 1.1239
No log 6.8947 262 1.3843 0.1744 1.3843 1.1766
No log 6.9474 264 1.4974 0.2372 1.4974 1.2237
No log 7.0 266 1.5296 0.2065 1.5296 1.2368
No log 7.0526 268 1.4739 0.1744 1.4739 1.2140
No log 7.1053 270 1.3741 0.0781 1.3741 1.1722
No log 7.1579 272 1.3888 0.0781 1.3888 1.1785
No log 7.2105 274 1.4584 0.1744 1.4584 1.2076
No log 7.2632 276 1.5707 0.2611 1.5707 1.2533
No log 7.3158 278 1.5594 0.2611 1.5594 1.2488
No log 7.3684 280 1.5060 0.2391 1.5060 1.2272
No log 7.4211 282 1.5509 0.2653 1.5509 1.2453
No log 7.4737 284 1.6087 0.2525 1.6087 1.2683
No log 7.5263 286 1.5205 0.2653 1.5205 1.2331
No log 7.5789 288 1.3532 0.2126 1.3532 1.1633
No log 7.6316 290 1.2372 0.1407 1.2372 1.1123
No log 7.6842 292 1.2360 0.1407 1.2360 1.1117
No log 7.7368 294 1.3004 0.1814 1.3004 1.1403
No log 7.7895 296 1.4173 0.2752 1.4173 1.1905
No log 7.8421 298 1.4341 0.2752 1.4341 1.1975
No log 7.8947 300 1.4097 0.1744 1.4097 1.1873
No log 7.9474 302 1.4514 0.2424 1.4514 1.2047
No log 8.0 304 1.5678 0.2611 1.5678 1.2521
No log 8.0526 306 1.7697 0.2566 1.7697 1.3303
No log 8.1053 308 1.9838 0.0966 1.9838 1.4085
No log 8.1579 310 2.0136 0.0974 2.0136 1.4190
No log 8.2105 312 1.8972 0.1505 1.8972 1.3774
No log 8.2632 314 1.6155 0.2566 1.6155 1.2710
No log 8.3158 316 1.3311 0.1744 1.3311 1.1538
No log 8.3684 318 1.2252 0.0781 1.2252 1.1069
No log 8.4211 320 1.2126 0.0401 1.2126 1.1012
No log 8.4737 322 1.2642 0.0781 1.2642 1.1244
No log 8.5263 324 1.4155 0.1744 1.4155 1.1898
No log 8.5789 326 1.6375 0.2482 1.6375 1.2797
No log 8.6316 328 1.7811 0.2606 1.7811 1.3346
No log 8.6842 330 1.8331 0.2056 1.8331 1.3539
No log 8.7368 332 1.8766 0.2103 1.8766 1.3699
No log 8.7895 334 1.8602 0.2206 1.8602 1.3639
No log 8.8421 336 1.7522 0.2566 1.7522 1.3237
No log 8.8947 338 1.6155 0.2126 1.6155 1.2710
No log 8.9474 340 1.5343 0.1814 1.5343 1.2387
No log 9.0 342 1.4914 0.1814 1.4914 1.2212
No log 9.0526 344 1.5108 0.1814 1.5108 1.2291
No log 9.1053 346 1.4965 0.2065 1.4965 1.2233
No log 9.1579 348 1.5076 0.2372 1.5076 1.2278
No log 9.2105 350 1.6038 0.2653 1.6038 1.2664
No log 9.2632 352 1.6319 0.2525 1.6319 1.2774
No log 9.3158 354 1.6202 0.2694 1.6202 1.2729
No log 9.3684 356 1.4101 0.2292 1.4101 1.1875
No log 9.4211 358 1.3255 0.2126 1.3255 1.1513
No log 9.4737 360 1.3585 0.2126 1.3585 1.1655
No log 9.5263 362 1.4460 0.2126 1.4460 1.2025
No log 9.5789 364 1.4511 0.1814 1.4511 1.2046
No log 9.6316 366 1.5111 0.2126 1.5111 1.2293
No log 9.6842 368 1.6088 0.2474 1.6088 1.2684
No log 9.7368 370 1.6129 0.2474 1.6129 1.2700
No log 9.7895 372 1.5257 0.2184 1.5257 1.2352
No log 9.8421 374 1.4448 0.2126 1.4448 1.2020
No log 9.8947 376 1.3540 0.0781 1.3540 1.1636
No log 9.9474 378 1.3796 0.0781 1.3796 1.1746
No log 10.0 380 1.4803 0.2126 1.4803 1.2167
No log 10.0526 382 1.5460 0.2522 1.5460 1.2434
No log 10.1053 384 1.5008 0.2522 1.5008 1.2251
No log 10.1579 386 1.3805 0.2065 1.3805 1.1749
No log 10.2105 388 1.3778 0.2065 1.3778 1.1738
No log 10.2632 390 1.4347 0.2832 1.4347 1.1978
No log 10.3158 392 1.5193 0.2694 1.5193 1.2326
No log 10.3684 394 1.4947 0.2653 1.4947 1.2226
No log 10.4211 396 1.4301 0.2522 1.4301 1.1959
No log 10.4737 398 1.4006 0.2126 1.4006 1.1835
No log 10.5263 400 1.3855 0.2126 1.3855 1.1771
No log 10.5789 402 1.2805 0.1744 1.2805 1.1316
No log 10.6316 404 1.2260 0.0401 1.2260 1.1073
No log 10.6842 406 1.2479 0.0781 1.2479 1.1171
No log 10.7368 408 1.2929 0.1142 1.2929 1.1371
No log 10.7895 410 1.3824 0.2065 1.3824 1.1757
No log 10.8421 412 1.4348 0.2126 1.4348 1.1978
No log 10.8947 414 1.3752 0.1744 1.3752 1.1727
No log 10.9474 416 1.2854 0.1142 1.2854 1.1338
No log 11.0 418 1.2648 0.0401 1.2648 1.1246
No log 11.0526 420 1.2810 0.1744 1.2810 1.1318
No log 11.1053 422 1.2908 0.1744 1.2908 1.1361
No log 11.1579 424 1.2946 0.1880 1.2946 1.1378
No log 11.2105 426 1.3243 0.2522 1.3243 1.1508
No log 11.2632 428 1.3926 0.2611 1.3926 1.1801
No log 11.3158 430 1.5053 0.2611 1.5053 1.2269
No log 11.3684 432 1.5397 0.2653 1.5397 1.2409
No log 11.4211 434 1.4423 0.2752 1.4423 1.2010
No log 11.4737 436 1.4041 0.2752 1.4041 1.1849
No log 11.5263 438 1.3216 0.2424 1.3216 1.1496
No log 11.5789 440 1.3020 0.2424 1.3020 1.1410
No log 11.6316 442 1.2782 0.1744 1.2782 1.1306
No log 11.6842 444 1.2705 0.1744 1.2705 1.1272
No log 11.7368 446 1.2250 0.1744 1.2250 1.1068
No log 11.7895 448 1.2044 0.1407 1.2044 1.0975
No log 11.8421 450 1.3074 0.2709 1.3074 1.1434
No log 11.8947 452 1.4268 0.2653 1.4268 1.1945
No log 11.9474 454 1.4278 0.2653 1.4278 1.1949
No log 12.0 456 1.3728 0.2793 1.3728 1.1717
No log 12.0526 458 1.2805 0.2424 1.2805 1.1316
No log 12.1053 460 1.2649 0.1744 1.2649 1.1247
No log 12.1579 462 1.2277 0.1052 1.2277 1.1080
No log 12.2105 464 1.1920 0.1052 1.1920 1.0918
No log 12.2632 466 1.2249 0.1052 1.2249 1.1068
No log 12.3158 468 1.3238 0.1814 1.3238 1.1506
No log 12.3684 470 1.4443 0.2709 1.4443 1.2018
No log 12.4211 472 1.4771 0.2611 1.4771 1.2154
No log 12.4737 474 1.4046 0.2611 1.4046 1.1852
No log 12.5263 476 1.3116 0.2522 1.3116 1.1453
No log 12.5789 478 1.2477 0.2315 1.2477 1.1170
No log 12.6316 480 1.3294 0.2424 1.3294 1.1530
No log 12.6842 482 1.3686 0.2126 1.3686 1.1699
No log 12.7368 484 1.3430 0.2126 1.3430 1.1589
No log 12.7895 486 1.2678 0.1052 1.2678 1.1260
No log 12.8421 488 1.2453 0.0401 1.2453 1.1159
No log 12.8947 490 1.2621 0.1052 1.2621 1.1234
No log 12.9474 492 1.3086 0.1052 1.3086 1.1440
No log 13.0 494 1.3155 0.1052 1.3155 1.1470
No log 13.0526 496 1.3102 0.1744 1.3102 1.1446
No log 13.1053 498 1.3544 0.2474 1.3544 1.1638
0.2443 13.1579 500 1.4682 0.2342 1.4682 1.2117
0.2443 13.2105 502 1.6026 0.2694 1.6026 1.2659
0.2443 13.2632 504 1.6453 0.2653 1.6453 1.2827
0.2443 13.3158 506 1.5832 0.2752 1.5832 1.2583
0.2443 13.3684 508 1.4710 0.2126 1.4710 1.2128
0.2443 13.4211 510 1.4099 0.1052 1.4099 1.1874
0.2443 13.4737 512 1.3994 0.1052 1.3994 1.1829
0.2443 13.5263 514 1.4199 0.1407 1.4199 1.1916
0.2443 13.5789 516 1.4796 0.1744 1.4796 1.2164

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task5_organization

Finetuned
(4019)
this model