ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2996
  • Qwk: 0.1235
  • Mse: 1.2996
  • Rmse: 1.1400

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1053 2 2.4641 -0.0568 2.4641 1.5697
No log 0.2105 4 1.2791 0.1882 1.2791 1.1310
No log 0.3158 6 1.0031 -0.0550 1.0031 1.0015
No log 0.4211 8 1.1737 -0.1355 1.1737 1.0834
No log 0.5263 10 1.2532 -0.1993 1.2532 1.1195
No log 0.6316 12 0.8472 0.0 0.8472 0.9204
No log 0.7368 14 0.6607 0.1232 0.6607 0.8128
No log 0.8421 16 0.6433 0.2676 0.6433 0.8021
No log 0.9474 18 0.7040 0.3019 0.7040 0.8391
No log 1.0526 20 0.7188 0.3019 0.7188 0.8478
No log 1.1579 22 0.6808 0.2676 0.6808 0.8251
No log 1.2632 24 0.6844 0.2676 0.6844 0.8273
No log 1.3684 26 0.6940 0.2676 0.6940 0.8331
No log 1.4737 28 0.7603 0.3125 0.7603 0.8720
No log 1.5789 30 0.8185 0.1660 0.8185 0.9047
No log 1.6842 32 0.7336 0.2748 0.7336 0.8565
No log 1.7895 34 0.8494 0.2358 0.8494 0.9217
No log 1.8947 36 1.0127 0.1955 1.0127 1.0063
No log 2.0 38 0.8270 0.1459 0.8270 0.9094
No log 2.1053 40 0.7232 0.1277 0.7232 0.8504
No log 2.2105 42 0.7874 0.2156 0.7874 0.8874
No log 2.3158 44 0.7714 0.1365 0.7714 0.8783
No log 2.4211 46 0.7528 0.0717 0.7528 0.8677
No log 2.5263 48 0.8233 0.2407 0.8233 0.9073
No log 2.6316 50 0.8316 0.2652 0.8316 0.9119
No log 2.7368 52 0.7702 0.1863 0.7702 0.8776
No log 2.8421 54 0.7735 0.2884 0.7735 0.8795
No log 2.9474 56 0.7935 0.3238 0.7935 0.8908
No log 3.0526 58 1.0933 0.1241 1.0933 1.0456
No log 3.1579 60 1.2316 0.1839 1.2316 1.1098
No log 3.2632 62 1.0365 0.2460 1.0365 1.0181
No log 3.3684 64 0.9874 0.1692 0.9874 0.9937
No log 3.4737 66 1.1109 0.2209 1.1109 1.0540
No log 3.5789 68 1.4546 0.1067 1.4546 1.2061
No log 3.6842 70 1.6366 0.1555 1.6366 1.2793
No log 3.7895 72 1.3934 0.1093 1.3934 1.1804
No log 3.8947 74 1.3039 0.1175 1.3039 1.1419
No log 4.0 76 1.1431 0.1394 1.1431 1.0692
No log 4.1053 78 1.1404 0.1976 1.1404 1.0679
No log 4.2105 80 1.3409 0.1568 1.3409 1.1580
No log 4.3158 82 1.1559 0.1618 1.1559 1.0751
No log 4.4211 84 1.2002 0.1799 1.2002 1.0955
No log 4.5263 86 1.5612 0.1169 1.5612 1.2495
No log 4.6316 88 1.9708 0.0421 1.9708 1.4038
No log 4.7368 90 1.9778 0.0421 1.9778 1.4063
No log 4.8421 92 1.7506 0.0589 1.7506 1.3231
No log 4.9474 94 1.5731 0.1195 1.5731 1.2542
No log 5.0526 96 1.4243 0.1093 1.4243 1.1934
No log 5.1579 98 1.3727 0.1093 1.3727 1.1716
No log 5.2632 100 1.6732 0.0689 1.6732 1.2935
No log 5.3684 102 1.6295 0.0300 1.6295 1.2765
No log 5.4737 104 1.1847 0.1029 1.1847 1.0884
No log 5.5789 106 0.9578 0.1661 0.9578 0.9787
No log 5.6842 108 0.9987 0.1603 0.9987 0.9994
No log 5.7895 110 1.2834 0.1458 1.2834 1.1329
No log 5.8947 112 1.4730 0.0803 1.4730 1.2137
No log 6.0 114 1.3214 0.1458 1.3214 1.1495
No log 6.1053 116 0.9662 0.1651 0.9662 0.9830
No log 6.2105 118 0.9059 0.2076 0.9059 0.9518
No log 6.3158 120 1.0265 0.1210 1.0265 1.0132
No log 6.4211 122 1.4517 0.0361 1.4517 1.2049
No log 6.5263 124 1.8053 0.0350 1.8053 1.3436
No log 6.6316 126 1.7976 0.0350 1.7976 1.3408
No log 6.7368 128 1.5456 0.0447 1.5456 1.2432
No log 6.8421 130 1.2203 0.1262 1.2203 1.1047
No log 6.9474 132 1.1002 0.1356 1.1002 1.0489
No log 7.0526 134 1.1795 0.1293 1.1795 1.0861
No log 7.1579 136 1.3348 0.1174 1.3348 1.1553
No log 7.2632 138 1.2655 0.1293 1.2655 1.1249
No log 7.3684 140 1.2032 0.1293 1.2032 1.0969
No log 7.4737 142 1.3617 0.1175 1.3617 1.1669
No log 7.5789 144 1.5876 0.0283 1.5876 1.2600
No log 7.6842 146 1.5824 0.0283 1.5824 1.2579
No log 7.7895 148 1.3209 0.1175 1.3209 1.1493
No log 7.8947 150 1.0629 0.2119 1.0629 1.0310
No log 8.0 152 1.0602 0.1787 1.0602 1.0297
No log 8.1053 154 1.1824 0.2412 1.1824 1.0874
No log 8.2105 156 1.3902 0.0873 1.3902 1.1791
No log 8.3158 158 1.3196 0.1464 1.3196 1.1488
No log 8.4211 160 1.0933 0.1709 1.0933 1.0456
No log 8.5263 162 0.9094 0.1777 0.9094 0.9536
No log 8.6316 164 0.8668 0.2692 0.8668 0.9310
No log 8.7368 166 0.9346 0.1955 0.9346 0.9667
No log 8.8421 168 1.2596 0.1458 1.2596 1.1223
No log 8.9474 170 1.4876 0.0745 1.4876 1.2197
No log 9.0526 172 1.3800 0.0829 1.3800 1.1747
No log 9.1579 174 1.0758 0.2782 1.0758 1.0372
No log 9.2632 176 0.8217 0.2352 0.8217 0.9065
No log 9.3684 178 0.7643 0.2407 0.7643 0.8743
No log 9.4737 180 0.7747 0.2718 0.7747 0.8802
No log 9.5789 182 0.9130 0.2000 0.9130 0.9555
No log 9.6842 184 1.2059 0.2045 1.2059 1.0981
No log 9.7895 186 1.3614 0.1427 1.3614 1.1668
No log 9.8947 188 1.3851 0.1275 1.3851 1.1769
No log 10.0 190 1.4571 0.1549 1.4571 1.2071
No log 10.1053 192 1.2856 0.1638 1.2856 1.1338
No log 10.2105 194 1.1946 0.1784 1.1946 1.0930
No log 10.3158 196 1.1869 0.1784 1.1869 1.0894
No log 10.4211 198 1.2125 0.1490 1.2125 1.1011
No log 10.5263 200 1.2332 0.1458 1.2332 1.1105
No log 10.6316 202 1.2355 0.1458 1.2355 1.1115
No log 10.7368 204 1.1226 0.1626 1.1226 1.0595
No log 10.8421 206 0.9740 0.1274 0.9740 0.9869
No log 10.9474 208 0.9782 0.1557 0.9782 0.9890
No log 11.0526 210 1.0508 0.0925 1.0508 1.0251
No log 11.1579 212 1.2252 0.1943 1.2252 1.1069
No log 11.2632 214 1.3686 0.1220 1.3686 1.1699
No log 11.3684 216 1.2856 0.1427 1.2856 1.1339
No log 11.4737 218 1.1311 0.2782 1.1311 1.0635
No log 11.5789 220 1.0827 0.2032 1.0827 1.0405
No log 11.6842 222 1.0560 0.1787 1.0560 1.0276
No log 11.7895 224 1.1182 0.1949 1.1182 1.0575
No log 11.8947 226 1.2789 0.0712 1.2789 1.1309
No log 12.0 228 1.3569 0.0419 1.3569 1.1649
No log 12.1053 230 1.3336 0.0694 1.3336 1.1548
No log 12.2105 232 1.5365 0.0832 1.5365 1.2395
No log 12.3158 234 1.7802 0.0932 1.7802 1.3342
No log 12.4211 236 1.6460 0.1577 1.6460 1.2830
No log 12.5263 238 1.3414 0.0952 1.3414 1.1582
No log 12.6316 240 1.1119 0.0448 1.1119 1.0545
No log 12.7368 242 1.0200 0.0774 1.0200 1.0100
No log 12.8421 244 1.0183 0.0799 1.0183 1.0091
No log 12.9474 246 1.1438 0.0561 1.1438 1.0695
No log 13.0526 248 1.2162 0.1205 1.2162 1.1028
No log 13.1579 250 1.1931 0.0761 1.1931 1.0923
No log 13.2632 252 1.0985 0.0982 1.0985 1.0481
No log 13.3684 254 1.1480 0.1243 1.1480 1.0714
No log 13.4737 256 1.3344 0.0694 1.3344 1.1551
No log 13.5789 258 1.4987 0.0086 1.4987 1.2242
No log 13.6842 260 1.4678 0.0343 1.4678 1.2115
No log 13.7895 262 1.2875 0.0459 1.2875 1.1347
No log 13.8947 264 1.2565 0.0761 1.2565 1.1209
No log 14.0 266 1.1461 0.0838 1.1461 1.0706
No log 14.1053 268 1.0270 0.1312 1.0270 1.0134
No log 14.2105 270 1.0162 0.1734 1.0162 1.0081
No log 14.3158 272 1.1608 0.0561 1.1608 1.0774
No log 14.4211 274 1.4093 0.0325 1.4093 1.1871
No log 14.5263 276 1.5299 0.0230 1.5299 1.2369
No log 14.6316 278 1.4362 0.0531 1.4362 1.1984
No log 14.7368 280 1.3022 0.1175 1.3022 1.1411
No log 14.8421 282 1.2217 0.1029 1.2217 1.1053
No log 14.9474 284 1.1848 0.0546 1.1848 1.0885
No log 15.0526 286 1.2586 0.1262 1.2586 1.1219
No log 15.1579 288 1.3400 0.0921 1.3400 1.1576
No log 15.2632 290 1.3906 0.1067 1.3906 1.1792
No log 15.3684 292 1.2660 0.0921 1.2660 1.1252
No log 15.4737 294 1.1009 0.1530 1.1009 1.0492
No log 15.5789 296 1.0032 0.2363 1.0032 1.0016
No log 15.6842 298 0.9403 0.1312 0.9403 0.9697
No log 15.7895 300 0.9077 0.1822 0.9077 0.9527
No log 15.8947 302 0.9571 0.1013 0.9571 0.9783
No log 16.0 304 1.0693 0.1499 1.0693 1.0341
No log 16.1053 306 1.2585 0.1203 1.2585 1.1218
No log 16.2105 308 1.5040 0.0519 1.5040 1.2264
No log 16.3158 310 1.5247 0.0519 1.5247 1.2348
No log 16.4211 312 1.3325 0.1233 1.3325 1.1544
No log 16.5263 314 1.1556 0.1293 1.1556 1.0750
No log 16.6316 316 1.1220 0.1293 1.1220 1.0593
No log 16.7368 318 1.0058 0.1312 1.0058 1.0029
No log 16.8421 320 1.0166 0.1210 1.0166 1.0083
No log 16.9474 322 1.0913 0.1709 1.0913 1.0446
No log 17.0526 324 1.1894 0.1235 1.1894 1.0906
No log 17.1579 326 1.1429 0.1530 1.1429 1.0691
No log 17.2632 328 1.0604 0.1949 1.0604 1.0297
No log 17.3684 330 1.1282 0.1530 1.1282 1.0622
No log 17.4737 332 1.1423 0.1530 1.1423 1.0688
No log 17.5789 334 1.1916 0.1530 1.1916 1.0916
No log 17.6842 336 1.1933 0.1530 1.1933 1.0924
No log 17.7895 338 1.2090 0.1832 1.2090 1.0996
No log 17.8947 340 1.3492 0.0898 1.3492 1.1615
No log 18.0 342 1.2947 0.1490 1.2947 1.1378
No log 18.1053 344 1.1130 0.1626 1.1130 1.0550
No log 18.2105 346 0.9967 0.2411 0.9967 0.9984
No log 18.3158 348 0.8932 0.2259 0.8932 0.9451
No log 18.4211 350 0.8709 0.2000 0.8709 0.9332
No log 18.5263 352 0.9221 0.2211 0.9221 0.9603
No log 18.6316 354 1.0804 0.1990 1.0804 1.0394
No log 18.7368 356 1.2835 0.1758 1.2835 1.1329
No log 18.8421 358 1.4312 0.0849 1.4312 1.1963
No log 18.9474 360 1.5142 0.0519 1.5142 1.2305
No log 19.0526 362 1.4429 0.1042 1.4429 1.2012
No log 19.1579 364 1.3233 0.1205 1.3233 1.1504
No log 19.2632 366 1.3399 0.0620 1.3399 1.1575
No log 19.3684 368 1.3808 0.0544 1.3808 1.1751
No log 19.4737 370 1.4864 0.0923 1.4864 1.2192
No log 19.5789 372 1.4842 0.0992 1.4842 1.2183
No log 19.6842 374 1.3251 0.0333 1.3251 1.1511
No log 19.7895 376 1.1285 0.1671 1.1285 1.0623
No log 19.8947 378 1.0125 0.1651 1.0125 1.0062
No log 20.0 380 1.0529 0.1651 1.0529 1.0261
No log 20.1053 382 1.1272 0.0592 1.1272 1.0617
No log 20.2105 384 1.2052 0.0765 1.2052 1.0978
No log 20.3158 386 1.3905 0.0605 1.3905 1.1792
No log 20.4211 388 1.5084 -0.0165 1.5084 1.2282
No log 20.5263 390 1.4851 -0.0165 1.4851 1.2186
No log 20.6316 392 1.3252 0.0898 1.3252 1.1512
No log 20.7368 394 1.0801 0.0952 1.0801 1.0393
No log 20.8421 396 0.8944 0.1628 0.8944 0.9457
No log 20.9474 398 0.8394 0.2652 0.8394 0.9162
No log 21.0526 400 0.8274 0.2652 0.8274 0.9096
No log 21.1579 402 0.8750 0.2116 0.8750 0.9354
No log 21.2632 404 1.0301 0.0982 1.0301 1.0150
No log 21.3684 406 1.3061 0.0947 1.3061 1.1428
No log 21.4737 408 1.5595 0.0101 1.5595 1.2488
No log 21.5789 410 1.7050 -0.0186 1.7050 1.3058
No log 21.6842 412 1.6549 0.0055 1.6549 1.2864
No log 21.7895 414 1.3882 0.0923 1.3882 1.1782
No log 21.8947 416 1.1467 0.0616 1.1467 1.0708
No log 22.0 418 1.0397 0.0539 1.0397 1.0196
No log 22.1053 420 1.0614 0.0512 1.0614 1.0302
No log 22.2105 422 1.0479 0.0891 1.0479 1.0237
No log 22.3158 424 1.0309 0.0891 1.0309 1.0153
No log 22.4211 426 1.1298 -0.0165 1.1298 1.0629
No log 22.5263 428 1.3419 0.0741 1.3419 1.1584
No log 22.6316 430 1.4391 -0.0100 1.4391 1.1996
No log 22.7368 432 1.3573 0.0166 1.3573 1.1650
No log 22.8421 434 1.2212 0.1394 1.2212 1.1051
No log 22.9474 436 1.1695 0.1147 1.1695 1.0814
No log 23.0526 438 1.1767 0.1709 1.1767 1.0847
No log 23.1579 440 1.1413 0.1709 1.1413 1.0683
No log 23.2632 442 1.1386 0.1463 1.1386 1.0670
No log 23.3684 444 1.1475 0.1428 1.1475 1.0712
No log 23.4737 446 1.1194 0.0982 1.1194 1.0580
No log 23.5789 448 1.0326 0.1348 1.0326 1.0162
No log 23.6842 450 1.0183 0.1348 1.0183 1.0091
No log 23.7895 452 1.0843 0.0982 1.0843 1.0413
No log 23.8947 454 1.1687 0.1210 1.1687 1.0811
No log 24.0 456 1.1701 0.1178 1.1701 1.0817
No log 24.1053 458 1.1358 0.1210 1.1358 1.0657
No log 24.2105 460 1.0002 0.1692 1.0002 1.0001
No log 24.3158 462 0.9057 0.2297 0.9057 0.9517
No log 24.4211 464 0.8573 0.2116 0.8573 0.9259
No log 24.5263 466 0.8890 0.2116 0.8890 0.9429
No log 24.6316 468 0.9877 0.1110 0.9877 0.9938
No log 24.7368 470 1.0737 0.1308 1.0737 1.0362
No log 24.8421 472 1.1142 0.0925 1.1142 1.0556
No log 24.9474 474 1.1482 0.1422 1.1482 1.0715
No log 25.0526 476 1.0793 0.1422 1.0793 1.0389
No log 25.1579 478 1.0190 0.1573 1.0190 1.0095
No log 25.2632 480 1.0339 0.1827 1.0339 1.0168
No log 25.3684 482 1.0465 0.1827 1.0465 1.0230
No log 25.4737 484 1.1158 0.1747 1.1158 1.0563
No log 25.5789 486 1.1277 0.1747 1.1277 1.0619
No log 25.6842 488 1.0563 0.2164 1.0563 1.0278
No log 25.7895 490 1.0323 0.1651 1.0323 1.0160
No log 25.8947 492 1.0232 0.1651 1.0232 1.0115
No log 26.0 494 1.0698 0.2164 1.0698 1.0343
No log 26.1053 496 1.0505 0.1911 1.0505 1.0249
No log 26.2105 498 0.9993 0.1911 0.9993 0.9997
0.2463 26.3158 500 0.9288 0.1734 0.9288 0.9638
0.2463 26.4211 502 0.8990 0.1734 0.8990 0.9482
0.2463 26.5263 504 0.9162 0.1385 0.9162 0.9572
0.2463 26.6316 506 0.9487 0.1911 0.9487 0.9740
0.2463 26.7368 508 0.9808 0.1911 0.9808 0.9904
0.2463 26.8421 510 1.0316 0.1499 1.0316 1.0157
0.2463 26.9474 512 1.1073 0.1709 1.1073 1.0523
0.2463 27.0526 514 1.2262 0.2100 1.2262 1.1073
0.2463 27.1579 516 1.3173 0.1748 1.3173 1.1477
0.2463 27.2632 518 1.3183 0.1458 1.3183 1.1482
0.2463 27.3684 520 1.2479 0.2100 1.2479 1.1171
0.2463 27.4737 522 1.2457 0.2100 1.2457 1.1161
0.2463 27.5789 524 1.2612 0.2100 1.2612 1.1230
0.2463 27.6842 526 1.1119 0.1463 1.1119 1.0544
0.2463 27.7895 528 0.9374 0.1692 0.9374 0.9682
0.2463 27.8947 530 0.8263 0.2012 0.8263 0.9090
0.2463 28.0 532 0.8027 0.2116 0.8027 0.8960
0.2463 28.1053 534 0.8304 0.2116 0.8304 0.9113
0.2463 28.2105 536 0.9080 0.1734 0.9080 0.9529
0.2463 28.3158 538 1.0356 0.1573 1.0356 1.0176
0.2463 28.4211 540 1.1520 0.2316 1.1520 1.0733
0.2463 28.5263 542 1.3547 0.0642 1.3547 1.1639
0.2463 28.6316 544 1.4093 0.0351 1.4093 1.1872
0.2463 28.7368 546 1.3420 0.0642 1.3420 1.1585
0.2463 28.8421 548 1.3353 0.0921 1.3353 1.1555
0.2463 28.9474 550 1.2996 0.1235 1.2996 1.1400

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task7_organization

Finetuned
(4019)
this model